1 /* 2 * (C) Copyright 2011 Michal Simek 3 * 4 * Michal SIMEK <monstr@monstr.eu> 5 * 6 * Based on Xilinx gmac driver: 7 * (C) Copyright 2011 Xilinx 8 * 9 * SPDX-License-Identifier: GPL-2.0+ 10 */ 11 12 #include <common.h> 13 #include <net.h> 14 #include <netdev.h> 15 #include <config.h> 16 #include <fdtdec.h> 17 #include <libfdt.h> 18 #include <malloc.h> 19 #include <asm/io.h> 20 #include <phy.h> 21 #include <miiphy.h> 22 #include <watchdog.h> 23 #include <asm/system.h> 24 #include <asm/arch/hardware.h> 25 #include <asm/arch/sys_proto.h> 26 #include <asm-generic/errno.h> 27 28 #if !defined(CONFIG_PHYLIB) 29 # error XILINX_GEM_ETHERNET requires PHYLIB 30 #endif 31 32 /* Bit/mask specification */ 33 #define ZYNQ_GEM_PHYMNTNC_OP_MASK 0x40020000 /* operation mask bits */ 34 #define ZYNQ_GEM_PHYMNTNC_OP_R_MASK 0x20000000 /* read operation */ 35 #define ZYNQ_GEM_PHYMNTNC_OP_W_MASK 0x10000000 /* write operation */ 36 #define ZYNQ_GEM_PHYMNTNC_PHYAD_SHIFT_MASK 23 /* Shift bits for PHYAD */ 37 #define ZYNQ_GEM_PHYMNTNC_PHREG_SHIFT_MASK 18 /* Shift bits for PHREG */ 38 39 #define ZYNQ_GEM_RXBUF_EOF_MASK 0x00008000 /* End of frame. */ 40 #define ZYNQ_GEM_RXBUF_SOF_MASK 0x00004000 /* Start of frame. */ 41 #define ZYNQ_GEM_RXBUF_LEN_MASK 0x00003FFF /* Mask for length field */ 42 43 #define ZYNQ_GEM_RXBUF_WRAP_MASK 0x00000002 /* Wrap bit, last BD */ 44 #define ZYNQ_GEM_RXBUF_NEW_MASK 0x00000001 /* Used bit.. */ 45 #define ZYNQ_GEM_RXBUF_ADD_MASK 0xFFFFFFFC /* Mask for address */ 46 47 /* Wrap bit, last descriptor */ 48 #define ZYNQ_GEM_TXBUF_WRAP_MASK 0x40000000 49 #define ZYNQ_GEM_TXBUF_LAST_MASK 0x00008000 /* Last buffer */ 50 #define ZYNQ_GEM_TXBUF_USED_MASK 0x80000000 /* Used by Hw */ 51 52 #define ZYNQ_GEM_NWCTRL_TXEN_MASK 0x00000008 /* Enable transmit */ 53 #define ZYNQ_GEM_NWCTRL_RXEN_MASK 0x00000004 /* Enable receive */ 54 #define ZYNQ_GEM_NWCTRL_MDEN_MASK 0x00000010 /* Enable MDIO port */ 55 #define ZYNQ_GEM_NWCTRL_STARTTX_MASK 0x00000200 /* Start tx (tx_go) */ 56 57 #define ZYNQ_GEM_NWCFG_SPEED100 0x000000001 /* 100 Mbps operation */ 58 #define ZYNQ_GEM_NWCFG_SPEED1000 0x000000400 /* 1Gbps operation */ 59 #define ZYNQ_GEM_NWCFG_FDEN 0x000000002 /* Full Duplex mode */ 60 #define ZYNQ_GEM_NWCFG_FSREM 0x000020000 /* FCS removal */ 61 #define ZYNQ_GEM_NWCFG_MDCCLKDIV 0x0000c0000 /* Div pclk by 48, max 120MHz */ 62 63 #ifdef CONFIG_ARM64 64 # define ZYNQ_GEM_DBUS_WIDTH (1 << 21) /* 64 bit bus */ 65 #else 66 # define ZYNQ_GEM_DBUS_WIDTH (0 << 21) /* 32 bit bus */ 67 #endif 68 69 #define ZYNQ_GEM_NWCFG_INIT (ZYNQ_GEM_DBUS_WIDTH | \ 70 ZYNQ_GEM_NWCFG_FDEN | \ 71 ZYNQ_GEM_NWCFG_FSREM | \ 72 ZYNQ_GEM_NWCFG_MDCCLKDIV) 73 74 #define ZYNQ_GEM_NWSR_MDIOIDLE_MASK 0x00000004 /* PHY management idle */ 75 76 #define ZYNQ_GEM_DMACR_BLENGTH 0x00000004 /* INCR4 AHB bursts */ 77 /* Use full configured addressable space (8 Kb) */ 78 #define ZYNQ_GEM_DMACR_RXSIZE 0x00000300 79 /* Use full configured addressable space (4 Kb) */ 80 #define ZYNQ_GEM_DMACR_TXSIZE 0x00000400 81 /* Set with binary 00011000 to use 1536 byte(1*max length frame/buffer) */ 82 #define ZYNQ_GEM_DMACR_RXBUF 0x00180000 83 84 #define ZYNQ_GEM_DMACR_INIT (ZYNQ_GEM_DMACR_BLENGTH | \ 85 ZYNQ_GEM_DMACR_RXSIZE | \ 86 ZYNQ_GEM_DMACR_TXSIZE | \ 87 ZYNQ_GEM_DMACR_RXBUF) 88 89 #define ZYNQ_GEM_TSR_DONE 0x00000020 /* Tx done mask */ 90 91 /* Use MII register 1 (MII status register) to detect PHY */ 92 #define PHY_DETECT_REG 1 93 94 /* Mask used to verify certain PHY features (or register contents) 95 * in the register above: 96 * 0x1000: 10Mbps full duplex support 97 * 0x0800: 10Mbps half duplex support 98 * 0x0008: Auto-negotiation support 99 */ 100 #define PHY_DETECT_MASK 0x1808 101 102 /* TX BD status masks */ 103 #define ZYNQ_GEM_TXBUF_FRMLEN_MASK 0x000007ff 104 #define ZYNQ_GEM_TXBUF_EXHAUSTED 0x08000000 105 #define ZYNQ_GEM_TXBUF_UNDERRUN 0x10000000 106 107 /* Clock frequencies for different speeds */ 108 #define ZYNQ_GEM_FREQUENCY_10 2500000UL 109 #define ZYNQ_GEM_FREQUENCY_100 25000000UL 110 #define ZYNQ_GEM_FREQUENCY_1000 125000000UL 111 112 /* Device registers */ 113 struct zynq_gem_regs { 114 u32 nwctrl; /* 0x0 - Network Control reg */ 115 u32 nwcfg; /* 0x4 - Network Config reg */ 116 u32 nwsr; /* 0x8 - Network Status reg */ 117 u32 reserved1; 118 u32 dmacr; /* 0x10 - DMA Control reg */ 119 u32 txsr; /* 0x14 - TX Status reg */ 120 u32 rxqbase; /* 0x18 - RX Q Base address reg */ 121 u32 txqbase; /* 0x1c - TX Q Base address reg */ 122 u32 rxsr; /* 0x20 - RX Status reg */ 123 u32 reserved2[2]; 124 u32 idr; /* 0x2c - Interrupt Disable reg */ 125 u32 reserved3; 126 u32 phymntnc; /* 0x34 - Phy Maintaince reg */ 127 u32 reserved4[18]; 128 u32 hashl; /* 0x80 - Hash Low address reg */ 129 u32 hashh; /* 0x84 - Hash High address reg */ 130 #define LADDR_LOW 0 131 #define LADDR_HIGH 1 132 u32 laddr[4][LADDR_HIGH + 1]; /* 0x8c - Specific1 addr low/high reg */ 133 u32 match[4]; /* 0xa8 - Type ID1 Match reg */ 134 u32 reserved6[18]; 135 #define STAT_SIZE 44 136 u32 stat[STAT_SIZE]; /* 0x100 - Octects transmitted Low reg */ 137 u32 reserved7[164]; 138 u32 transmit_q1_ptr; /* 0x440 - Transmit priority queue 1 */ 139 u32 reserved8[15]; 140 u32 receive_q1_ptr; /* 0x480 - Receive priority queue 1 */ 141 }; 142 143 /* BD descriptors */ 144 struct emac_bd { 145 u32 addr; /* Next descriptor pointer */ 146 u32 status; 147 }; 148 149 #define RX_BUF 32 150 /* Page table entries are set to 1MB, or multiples of 1MB 151 * (not < 1MB). driver uses less bd's so use 1MB bdspace. 152 */ 153 #define BD_SPACE 0x100000 154 /* BD separation space */ 155 #define BD_SEPRN_SPACE (RX_BUF * sizeof(struct emac_bd)) 156 157 /* Setup the first free TX descriptor */ 158 #define TX_FREE_DESC 2 159 160 /* Initialized, rxbd_current, rx_first_buf must be 0 after init */ 161 struct zynq_gem_priv { 162 struct emac_bd *tx_bd; 163 struct emac_bd *rx_bd; 164 char *rxbuffers; 165 u32 rxbd_current; 166 u32 rx_first_buf; 167 int phyaddr; 168 u32 emio; 169 int init; 170 struct zynq_gem_regs *iobase; 171 phy_interface_t interface; 172 struct phy_device *phydev; 173 struct mii_dev *bus; 174 }; 175 176 static inline int mdio_wait(struct zynq_gem_regs *regs) 177 { 178 u32 timeout = 20000; 179 180 /* Wait till MDIO interface is ready to accept a new transaction. */ 181 while (--timeout) { 182 if (readl(®s->nwsr) & ZYNQ_GEM_NWSR_MDIOIDLE_MASK) 183 break; 184 WATCHDOG_RESET(); 185 } 186 187 if (!timeout) { 188 printf("%s: Timeout\n", __func__); 189 return 1; 190 } 191 192 return 0; 193 } 194 195 static u32 phy_setup_op(struct zynq_gem_priv *priv, u32 phy_addr, u32 regnum, 196 u32 op, u16 *data) 197 { 198 u32 mgtcr; 199 struct zynq_gem_regs *regs = priv->iobase; 200 201 if (mdio_wait(regs)) 202 return 1; 203 204 /* Construct mgtcr mask for the operation */ 205 mgtcr = ZYNQ_GEM_PHYMNTNC_OP_MASK | op | 206 (phy_addr << ZYNQ_GEM_PHYMNTNC_PHYAD_SHIFT_MASK) | 207 (regnum << ZYNQ_GEM_PHYMNTNC_PHREG_SHIFT_MASK) | *data; 208 209 /* Write mgtcr and wait for completion */ 210 writel(mgtcr, ®s->phymntnc); 211 212 if (mdio_wait(regs)) 213 return 1; 214 215 if (op == ZYNQ_GEM_PHYMNTNC_OP_R_MASK) 216 *data = readl(®s->phymntnc); 217 218 return 0; 219 } 220 221 static u32 phyread(struct zynq_gem_priv *priv, u32 phy_addr, 222 u32 regnum, u16 *val) 223 { 224 u32 ret; 225 226 ret = phy_setup_op(priv, phy_addr, regnum, 227 ZYNQ_GEM_PHYMNTNC_OP_R_MASK, val); 228 229 if (!ret) 230 debug("%s: phy_addr %d, regnum 0x%x, val 0x%x\n", __func__, 231 phy_addr, regnum, *val); 232 233 return ret; 234 } 235 236 static u32 phywrite(struct zynq_gem_priv *priv, u32 phy_addr, 237 u32 regnum, u16 data) 238 { 239 debug("%s: phy_addr %d, regnum 0x%x, data 0x%x\n", __func__, phy_addr, 240 regnum, data); 241 242 return phy_setup_op(priv, phy_addr, regnum, 243 ZYNQ_GEM_PHYMNTNC_OP_W_MASK, &data); 244 } 245 246 static int phy_detection(struct eth_device *dev) 247 { 248 int i; 249 u16 phyreg; 250 struct zynq_gem_priv *priv = dev->priv; 251 252 if (priv->phyaddr != -1) { 253 phyread(priv, priv->phyaddr, PHY_DETECT_REG, &phyreg); 254 if ((phyreg != 0xFFFF) && 255 ((phyreg & PHY_DETECT_MASK) == PHY_DETECT_MASK)) { 256 /* Found a valid PHY address */ 257 debug("Default phy address %d is valid\n", 258 priv->phyaddr); 259 return 0; 260 } else { 261 debug("PHY address is not setup correctly %d\n", 262 priv->phyaddr); 263 priv->phyaddr = -1; 264 } 265 } 266 267 debug("detecting phy address\n"); 268 if (priv->phyaddr == -1) { 269 /* detect the PHY address */ 270 for (i = 31; i >= 0; i--) { 271 phyread(priv, i, PHY_DETECT_REG, &phyreg); 272 if ((phyreg != 0xFFFF) && 273 ((phyreg & PHY_DETECT_MASK) == PHY_DETECT_MASK)) { 274 /* Found a valid PHY address */ 275 priv->phyaddr = i; 276 debug("Found valid phy address, %d\n", i); 277 return 0; 278 } 279 } 280 } 281 printf("PHY is not detected\n"); 282 return -1; 283 } 284 285 static int zynq_gem_setup_mac(struct eth_device *dev) 286 { 287 u32 i, macaddrlow, macaddrhigh; 288 struct zynq_gem_regs *regs = (struct zynq_gem_regs *)dev->iobase; 289 290 /* Set the MAC bits [31:0] in BOT */ 291 macaddrlow = dev->enetaddr[0]; 292 macaddrlow |= dev->enetaddr[1] << 8; 293 macaddrlow |= dev->enetaddr[2] << 16; 294 macaddrlow |= dev->enetaddr[3] << 24; 295 296 /* Set MAC bits [47:32] in TOP */ 297 macaddrhigh = dev->enetaddr[4]; 298 macaddrhigh |= dev->enetaddr[5] << 8; 299 300 for (i = 0; i < 4; i++) { 301 writel(0, ®s->laddr[i][LADDR_LOW]); 302 writel(0, ®s->laddr[i][LADDR_HIGH]); 303 /* Do not use MATCHx register */ 304 writel(0, ®s->match[i]); 305 } 306 307 writel(macaddrlow, ®s->laddr[0][LADDR_LOW]); 308 writel(macaddrhigh, ®s->laddr[0][LADDR_HIGH]); 309 310 return 0; 311 } 312 313 static int zynq_gem_init(struct eth_device *dev, bd_t * bis) 314 { 315 u32 i; 316 int ret; 317 unsigned long clk_rate = 0; 318 struct zynq_gem_regs *regs = (struct zynq_gem_regs *)dev->iobase; 319 struct zynq_gem_priv *priv = dev->priv; 320 struct emac_bd *dummy_tx_bd = &priv->tx_bd[TX_FREE_DESC]; 321 struct emac_bd *dummy_rx_bd = &priv->tx_bd[TX_FREE_DESC + 2]; 322 const u32 supported = SUPPORTED_10baseT_Half | 323 SUPPORTED_10baseT_Full | 324 SUPPORTED_100baseT_Half | 325 SUPPORTED_100baseT_Full | 326 SUPPORTED_1000baseT_Half | 327 SUPPORTED_1000baseT_Full; 328 329 if (!priv->init) { 330 /* Disable all interrupts */ 331 writel(0xFFFFFFFF, ®s->idr); 332 333 /* Disable the receiver & transmitter */ 334 writel(0, ®s->nwctrl); 335 writel(0, ®s->txsr); 336 writel(0, ®s->rxsr); 337 writel(0, ®s->phymntnc); 338 339 /* Clear the Hash registers for the mac address 340 * pointed by AddressPtr 341 */ 342 writel(0x0, ®s->hashl); 343 /* Write bits [63:32] in TOP */ 344 writel(0x0, ®s->hashh); 345 346 /* Clear all counters */ 347 for (i = 0; i < STAT_SIZE; i++) 348 readl(®s->stat[i]); 349 350 /* Setup RxBD space */ 351 memset(priv->rx_bd, 0, RX_BUF * sizeof(struct emac_bd)); 352 353 for (i = 0; i < RX_BUF; i++) { 354 priv->rx_bd[i].status = 0xF0000000; 355 priv->rx_bd[i].addr = 356 ((ulong)(priv->rxbuffers) + 357 (i * PKTSIZE_ALIGN)); 358 } 359 /* WRAP bit to last BD */ 360 priv->rx_bd[--i].addr |= ZYNQ_GEM_RXBUF_WRAP_MASK; 361 /* Write RxBDs to IP */ 362 writel((ulong)priv->rx_bd, ®s->rxqbase); 363 364 /* Setup for DMA Configuration register */ 365 writel(ZYNQ_GEM_DMACR_INIT, ®s->dmacr); 366 367 /* Setup for Network Control register, MDIO, Rx and Tx enable */ 368 setbits_le32(®s->nwctrl, ZYNQ_GEM_NWCTRL_MDEN_MASK); 369 370 /* Disable the second priority queue */ 371 dummy_tx_bd->addr = 0; 372 dummy_tx_bd->status = ZYNQ_GEM_TXBUF_WRAP_MASK | 373 ZYNQ_GEM_TXBUF_LAST_MASK| 374 ZYNQ_GEM_TXBUF_USED_MASK; 375 376 dummy_rx_bd->addr = ZYNQ_GEM_RXBUF_WRAP_MASK | 377 ZYNQ_GEM_RXBUF_NEW_MASK; 378 dummy_rx_bd->status = 0; 379 flush_dcache_range((ulong)&dummy_tx_bd, (ulong)&dummy_tx_bd + 380 sizeof(dummy_tx_bd)); 381 flush_dcache_range((ulong)&dummy_rx_bd, (ulong)&dummy_rx_bd + 382 sizeof(dummy_rx_bd)); 383 384 writel((ulong)dummy_tx_bd, ®s->transmit_q1_ptr); 385 writel((ulong)dummy_rx_bd, ®s->receive_q1_ptr); 386 387 priv->init++; 388 } 389 390 ret = phy_detection(dev); 391 if (ret) { 392 printf("GEM PHY init failed\n"); 393 return ret; 394 } 395 396 priv->phydev = phy_connect(priv->bus, priv->phyaddr, dev, 397 priv->interface); 398 399 priv->phydev->supported = supported | ADVERTISED_Pause | 400 ADVERTISED_Asym_Pause; 401 priv->phydev->advertising = priv->phydev->supported; 402 phy_config(priv->phydev); 403 phy_startup(priv->phydev); 404 405 if (!priv->phydev->link) { 406 printf("%s: No link.\n", priv->phydev->dev->name); 407 return -1; 408 } 409 410 switch (priv->phydev->speed) { 411 case SPEED_1000: 412 writel(ZYNQ_GEM_NWCFG_INIT | ZYNQ_GEM_NWCFG_SPEED1000, 413 ®s->nwcfg); 414 clk_rate = ZYNQ_GEM_FREQUENCY_1000; 415 break; 416 case SPEED_100: 417 writel(ZYNQ_GEM_NWCFG_INIT | ZYNQ_GEM_NWCFG_SPEED100, 418 ®s->nwcfg); 419 clk_rate = ZYNQ_GEM_FREQUENCY_100; 420 break; 421 case SPEED_10: 422 clk_rate = ZYNQ_GEM_FREQUENCY_10; 423 break; 424 } 425 426 /* Change the rclk and clk only not using EMIO interface */ 427 if (!priv->emio) 428 zynq_slcr_gem_clk_setup(dev->iobase != 429 ZYNQ_GEM_BASEADDR0, clk_rate); 430 431 setbits_le32(®s->nwctrl, ZYNQ_GEM_NWCTRL_RXEN_MASK | 432 ZYNQ_GEM_NWCTRL_TXEN_MASK); 433 434 return 0; 435 } 436 437 static int wait_for_bit(const char *func, u32 *reg, const u32 mask, 438 bool set, unsigned int timeout) 439 { 440 u32 val; 441 unsigned long start = get_timer(0); 442 443 while (1) { 444 val = readl(reg); 445 446 if (!set) 447 val = ~val; 448 449 if ((val & mask) == mask) 450 return 0; 451 452 if (get_timer(start) > timeout) 453 break; 454 455 udelay(1); 456 } 457 458 debug("%s: Timeout (reg=%p mask=%08x wait_set=%i)\n", 459 func, reg, mask, set); 460 461 return -ETIMEDOUT; 462 } 463 464 static int zynq_gem_send(struct eth_device *dev, void *ptr, int len) 465 { 466 u32 addr, size; 467 struct zynq_gem_priv *priv = dev->priv; 468 struct zynq_gem_regs *regs = (struct zynq_gem_regs *)dev->iobase; 469 struct emac_bd *current_bd = &priv->tx_bd[1]; 470 471 /* Setup Tx BD */ 472 memset(priv->tx_bd, 0, sizeof(struct emac_bd)); 473 474 priv->tx_bd->addr = (ulong)ptr; 475 priv->tx_bd->status = (len & ZYNQ_GEM_TXBUF_FRMLEN_MASK) | 476 ZYNQ_GEM_TXBUF_LAST_MASK; 477 /* Dummy descriptor to mark it as the last in descriptor chain */ 478 current_bd->addr = 0x0; 479 current_bd->status = ZYNQ_GEM_TXBUF_WRAP_MASK | 480 ZYNQ_GEM_TXBUF_LAST_MASK| 481 ZYNQ_GEM_TXBUF_USED_MASK; 482 483 /* setup BD */ 484 writel((ulong)priv->tx_bd, ®s->txqbase); 485 486 addr = (ulong) ptr; 487 addr &= ~(ARCH_DMA_MINALIGN - 1); 488 size = roundup(len, ARCH_DMA_MINALIGN); 489 flush_dcache_range(addr, addr + size); 490 491 addr = (ulong)priv->rxbuffers; 492 addr &= ~(ARCH_DMA_MINALIGN - 1); 493 size = roundup((RX_BUF * PKTSIZE_ALIGN), ARCH_DMA_MINALIGN); 494 flush_dcache_range(addr, addr + size); 495 barrier(); 496 497 /* Start transmit */ 498 setbits_le32(®s->nwctrl, ZYNQ_GEM_NWCTRL_STARTTX_MASK); 499 500 /* Read TX BD status */ 501 if (priv->tx_bd->status & ZYNQ_GEM_TXBUF_EXHAUSTED) 502 printf("TX buffers exhausted in mid frame\n"); 503 504 return wait_for_bit(__func__, ®s->txsr, ZYNQ_GEM_TSR_DONE, 505 true, 20000); 506 } 507 508 /* Do not check frame_recd flag in rx_status register 0x20 - just poll BD */ 509 static int zynq_gem_recv(struct eth_device *dev) 510 { 511 int frame_len; 512 struct zynq_gem_priv *priv = dev->priv; 513 struct emac_bd *current_bd = &priv->rx_bd[priv->rxbd_current]; 514 struct emac_bd *first_bd; 515 516 if (!(current_bd->addr & ZYNQ_GEM_RXBUF_NEW_MASK)) 517 return 0; 518 519 if (!(current_bd->status & 520 (ZYNQ_GEM_RXBUF_SOF_MASK | ZYNQ_GEM_RXBUF_EOF_MASK))) { 521 printf("GEM: SOF or EOF not set for last buffer received!\n"); 522 return 0; 523 } 524 525 frame_len = current_bd->status & ZYNQ_GEM_RXBUF_LEN_MASK; 526 if (frame_len) { 527 u32 addr = current_bd->addr & ZYNQ_GEM_RXBUF_ADD_MASK; 528 addr &= ~(ARCH_DMA_MINALIGN - 1); 529 530 net_process_received_packet((u8 *)(ulong)addr, frame_len); 531 532 if (current_bd->status & ZYNQ_GEM_RXBUF_SOF_MASK) 533 priv->rx_first_buf = priv->rxbd_current; 534 else { 535 current_bd->addr &= ~ZYNQ_GEM_RXBUF_NEW_MASK; 536 current_bd->status = 0xF0000000; /* FIXME */ 537 } 538 539 if (current_bd->status & ZYNQ_GEM_RXBUF_EOF_MASK) { 540 first_bd = &priv->rx_bd[priv->rx_first_buf]; 541 first_bd->addr &= ~ZYNQ_GEM_RXBUF_NEW_MASK; 542 first_bd->status = 0xF0000000; 543 } 544 545 if ((++priv->rxbd_current) >= RX_BUF) 546 priv->rxbd_current = 0; 547 } 548 549 return frame_len; 550 } 551 552 static void zynq_gem_halt(struct eth_device *dev) 553 { 554 struct zynq_gem_regs *regs = (struct zynq_gem_regs *)dev->iobase; 555 556 clrsetbits_le32(®s->nwctrl, ZYNQ_GEM_NWCTRL_RXEN_MASK | 557 ZYNQ_GEM_NWCTRL_TXEN_MASK, 0); 558 } 559 560 static int zynq_gem_miiphyread(const char *devname, uchar addr, 561 uchar reg, ushort *val) 562 { 563 struct eth_device *dev = eth_get_dev(); 564 struct zynq_gem_priv *priv = dev->priv; 565 int ret; 566 567 ret = phyread(priv, addr, reg, val); 568 debug("%s 0x%x, 0x%x, 0x%x\n", __func__, addr, reg, *val); 569 return ret; 570 } 571 572 static int zynq_gem_miiphy_write(const char *devname, uchar addr, 573 uchar reg, ushort val) 574 { 575 struct eth_device *dev = eth_get_dev(); 576 struct zynq_gem_priv *priv = dev->priv; 577 578 debug("%s 0x%x, 0x%x, 0x%x\n", __func__, addr, reg, val); 579 return phywrite(priv, addr, reg, val); 580 } 581 582 int zynq_gem_initialize(bd_t *bis, phys_addr_t base_addr, 583 int phy_addr, u32 emio) 584 { 585 struct eth_device *dev; 586 struct zynq_gem_priv *priv; 587 void *bd_space; 588 589 dev = calloc(1, sizeof(*dev)); 590 if (dev == NULL) 591 return -1; 592 593 dev->priv = calloc(1, sizeof(struct zynq_gem_priv)); 594 if (dev->priv == NULL) { 595 free(dev); 596 return -1; 597 } 598 priv = dev->priv; 599 600 /* Align rxbuffers to ARCH_DMA_MINALIGN */ 601 priv->rxbuffers = memalign(ARCH_DMA_MINALIGN, RX_BUF * PKTSIZE_ALIGN); 602 memset(priv->rxbuffers, 0, RX_BUF * PKTSIZE_ALIGN); 603 604 /* Align bd_space to MMU_SECTION_SHIFT */ 605 bd_space = memalign(1 << MMU_SECTION_SHIFT, BD_SPACE); 606 mmu_set_region_dcache_behaviour((phys_addr_t)bd_space, 607 BD_SPACE, DCACHE_OFF); 608 609 /* Initialize the bd spaces for tx and rx bd's */ 610 priv->tx_bd = (struct emac_bd *)bd_space; 611 priv->rx_bd = (struct emac_bd *)((ulong)bd_space + BD_SEPRN_SPACE); 612 613 priv->phyaddr = phy_addr; 614 priv->emio = emio; 615 616 #ifndef CONFIG_ZYNQ_GEM_INTERFACE 617 priv->interface = PHY_INTERFACE_MODE_MII; 618 #else 619 priv->interface = CONFIG_ZYNQ_GEM_INTERFACE; 620 #endif 621 622 sprintf(dev->name, "Gem.%lx", base_addr); 623 624 dev->iobase = base_addr; 625 priv->iobase = (struct zynq_gem_regs *)base_addr; 626 627 dev->init = zynq_gem_init; 628 dev->halt = zynq_gem_halt; 629 dev->send = zynq_gem_send; 630 dev->recv = zynq_gem_recv; 631 dev->write_hwaddr = zynq_gem_setup_mac; 632 633 eth_register(dev); 634 635 miiphy_register(dev->name, zynq_gem_miiphyread, zynq_gem_miiphy_write); 636 priv->bus = miiphy_get_dev_by_name(dev->name); 637 638 return 1; 639 } 640 641 #if CONFIG_IS_ENABLED(OF_CONTROL) 642 int zynq_gem_of_init(const void *blob) 643 { 644 int offset = 0; 645 u32 ret = 0; 646 u32 reg, phy_reg; 647 648 debug("ZYNQ GEM: Initialization\n"); 649 650 do { 651 offset = fdt_node_offset_by_compatible(blob, offset, 652 "xlnx,ps7-ethernet-1.00.a"); 653 if (offset != -1) { 654 reg = fdtdec_get_addr(blob, offset, "reg"); 655 if (reg != FDT_ADDR_T_NONE) { 656 offset = fdtdec_lookup_phandle(blob, offset, 657 "phy-handle"); 658 if (offset != -1) 659 phy_reg = fdtdec_get_addr(blob, offset, 660 "reg"); 661 else 662 phy_reg = 0; 663 664 debug("ZYNQ GEM: addr %x, phyaddr %x\n", 665 reg, phy_reg); 666 667 ret |= zynq_gem_initialize(NULL, reg, 668 phy_reg, 0); 669 670 } else { 671 debug("ZYNQ GEM: Can't get base address\n"); 672 return -1; 673 } 674 } 675 } while (offset != -1); 676 677 return ret; 678 } 679 #endif 680