1f6569884SThomas Chou /* 2f6569884SThomas Chou * Opencore 10/100 ethernet mac driver 3f6569884SThomas Chou * 4f6569884SThomas Chou * Copyright (C) 2007-2008 Avionic Design Development GmbH 5f6569884SThomas Chou * Copyright (C) 2008-2009 Avionic Design GmbH 6f6569884SThomas Chou * Thierry Reding <thierry.reding@avionic-design.de> 7f6569884SThomas Chou * Copyright (C) 2010 Thomas Chou <thomas@wytron.com.tw> 8*5d43feabSMax Filippov * Copyright (C) 2016 Cadence Design Systems Inc. 9f6569884SThomas Chou * 10*5d43feabSMax Filippov * SPDX-License-Identifier: GPL-2.0 11f6569884SThomas Chou */ 12f6569884SThomas Chou 13f6569884SThomas Chou #include <common.h> 14*5d43feabSMax Filippov #include <dm/device.h> 15*5d43feabSMax Filippov #include <dm/platform_data/net_ethoc.h> 16a84a757aSMax Filippov #include <linux/io.h> 17f6569884SThomas Chou #include <malloc.h> 18f6569884SThomas Chou #include <net.h> 19f6569884SThomas Chou #include <miiphy.h> 20f6569884SThomas Chou #include <asm/cache.h> 21f6569884SThomas Chou 22f6569884SThomas Chou /* register offsets */ 23f6569884SThomas Chou #define MODER 0x00 24f6569884SThomas Chou #define INT_SOURCE 0x04 25f6569884SThomas Chou #define INT_MASK 0x08 26f6569884SThomas Chou #define IPGT 0x0c 27f6569884SThomas Chou #define IPGR1 0x10 28f6569884SThomas Chou #define IPGR2 0x14 29f6569884SThomas Chou #define PACKETLEN 0x18 30f6569884SThomas Chou #define COLLCONF 0x1c 31f6569884SThomas Chou #define TX_BD_NUM 0x20 32f6569884SThomas Chou #define CTRLMODER 0x24 33f6569884SThomas Chou #define MIIMODER 0x28 34f6569884SThomas Chou #define MIICOMMAND 0x2c 35f6569884SThomas Chou #define MIIADDRESS 0x30 36f6569884SThomas Chou #define MIITX_DATA 0x34 37f6569884SThomas Chou #define MIIRX_DATA 0x38 38f6569884SThomas Chou #define MIISTATUS 0x3c 39f6569884SThomas Chou #define MAC_ADDR0 0x40 40f6569884SThomas Chou #define MAC_ADDR1 0x44 41f6569884SThomas Chou #define ETH_HASH0 0x48 42f6569884SThomas Chou #define ETH_HASH1 0x4c 43f6569884SThomas Chou #define ETH_TXCTRL 0x50 44f6569884SThomas Chou 45f6569884SThomas Chou /* mode register */ 46f6569884SThomas Chou #define MODER_RXEN (1 << 0) /* receive enable */ 47f6569884SThomas Chou #define MODER_TXEN (1 << 1) /* transmit enable */ 48f6569884SThomas Chou #define MODER_NOPRE (1 << 2) /* no preamble */ 49f6569884SThomas Chou #define MODER_BRO (1 << 3) /* broadcast address */ 50f6569884SThomas Chou #define MODER_IAM (1 << 4) /* individual address mode */ 51f6569884SThomas Chou #define MODER_PRO (1 << 5) /* promiscuous mode */ 52f6569884SThomas Chou #define MODER_IFG (1 << 6) /* interframe gap for incoming frames */ 53f6569884SThomas Chou #define MODER_LOOP (1 << 7) /* loopback */ 54f6569884SThomas Chou #define MODER_NBO (1 << 8) /* no back-off */ 55f6569884SThomas Chou #define MODER_EDE (1 << 9) /* excess defer enable */ 56f6569884SThomas Chou #define MODER_FULLD (1 << 10) /* full duplex */ 57f6569884SThomas Chou #define MODER_RESET (1 << 11) /* FIXME: reset (undocumented) */ 58f6569884SThomas Chou #define MODER_DCRC (1 << 12) /* delayed CRC enable */ 59f6569884SThomas Chou #define MODER_CRC (1 << 13) /* CRC enable */ 60f6569884SThomas Chou #define MODER_HUGE (1 << 14) /* huge packets enable */ 61f6569884SThomas Chou #define MODER_PAD (1 << 15) /* padding enabled */ 62f6569884SThomas Chou #define MODER_RSM (1 << 16) /* receive small packets */ 63f6569884SThomas Chou 64f6569884SThomas Chou /* interrupt source and mask registers */ 65f6569884SThomas Chou #define INT_MASK_TXF (1 << 0) /* transmit frame */ 66f6569884SThomas Chou #define INT_MASK_TXE (1 << 1) /* transmit error */ 67f6569884SThomas Chou #define INT_MASK_RXF (1 << 2) /* receive frame */ 68f6569884SThomas Chou #define INT_MASK_RXE (1 << 3) /* receive error */ 69f6569884SThomas Chou #define INT_MASK_BUSY (1 << 4) 70f6569884SThomas Chou #define INT_MASK_TXC (1 << 5) /* transmit control frame */ 71f6569884SThomas Chou #define INT_MASK_RXC (1 << 6) /* receive control frame */ 72f6569884SThomas Chou 73f6569884SThomas Chou #define INT_MASK_TX (INT_MASK_TXF | INT_MASK_TXE) 74f6569884SThomas Chou #define INT_MASK_RX (INT_MASK_RXF | INT_MASK_RXE) 75f6569884SThomas Chou 76f6569884SThomas Chou #define INT_MASK_ALL ( \ 77f6569884SThomas Chou INT_MASK_TXF | INT_MASK_TXE | \ 78f6569884SThomas Chou INT_MASK_RXF | INT_MASK_RXE | \ 79f6569884SThomas Chou INT_MASK_TXC | INT_MASK_RXC | \ 80f6569884SThomas Chou INT_MASK_BUSY \ 81f6569884SThomas Chou ) 82f6569884SThomas Chou 83f6569884SThomas Chou /* packet length register */ 84f6569884SThomas Chou #define PACKETLEN_MIN(min) (((min) & 0xffff) << 16) 85f6569884SThomas Chou #define PACKETLEN_MAX(max) (((max) & 0xffff) << 0) 86f6569884SThomas Chou #define PACKETLEN_MIN_MAX(min, max) (PACKETLEN_MIN(min) | \ 87f6569884SThomas Chou PACKETLEN_MAX(max)) 88f6569884SThomas Chou 89f6569884SThomas Chou /* transmit buffer number register */ 90f6569884SThomas Chou #define TX_BD_NUM_VAL(x) (((x) <= 0x80) ? (x) : 0x80) 91f6569884SThomas Chou 92f6569884SThomas Chou /* control module mode register */ 93f6569884SThomas Chou #define CTRLMODER_PASSALL (1 << 0) /* pass all receive frames */ 94f6569884SThomas Chou #define CTRLMODER_RXFLOW (1 << 1) /* receive control flow */ 95f6569884SThomas Chou #define CTRLMODER_TXFLOW (1 << 2) /* transmit control flow */ 96f6569884SThomas Chou 97f6569884SThomas Chou /* MII mode register */ 98f6569884SThomas Chou #define MIIMODER_CLKDIV(x) ((x) & 0xfe) /* needs to be an even number */ 99f6569884SThomas Chou #define MIIMODER_NOPRE (1 << 8) /* no preamble */ 100f6569884SThomas Chou 101f6569884SThomas Chou /* MII command register */ 102f6569884SThomas Chou #define MIICOMMAND_SCAN (1 << 0) /* scan status */ 103f6569884SThomas Chou #define MIICOMMAND_READ (1 << 1) /* read status */ 104f6569884SThomas Chou #define MIICOMMAND_WRITE (1 << 2) /* write control data */ 105f6569884SThomas Chou 106f6569884SThomas Chou /* MII address register */ 107f6569884SThomas Chou #define MIIADDRESS_FIAD(x) (((x) & 0x1f) << 0) 108f6569884SThomas Chou #define MIIADDRESS_RGAD(x) (((x) & 0x1f) << 8) 109f6569884SThomas Chou #define MIIADDRESS_ADDR(phy, reg) (MIIADDRESS_FIAD(phy) | \ 110f6569884SThomas Chou MIIADDRESS_RGAD(reg)) 111f6569884SThomas Chou 112f6569884SThomas Chou /* MII transmit data register */ 113f6569884SThomas Chou #define MIITX_DATA_VAL(x) ((x) & 0xffff) 114f6569884SThomas Chou 115f6569884SThomas Chou /* MII receive data register */ 116f6569884SThomas Chou #define MIIRX_DATA_VAL(x) ((x) & 0xffff) 117f6569884SThomas Chou 118f6569884SThomas Chou /* MII status register */ 119f6569884SThomas Chou #define MIISTATUS_LINKFAIL (1 << 0) 120f6569884SThomas Chou #define MIISTATUS_BUSY (1 << 1) 121f6569884SThomas Chou #define MIISTATUS_INVALID (1 << 2) 122f6569884SThomas Chou 123f6569884SThomas Chou /* TX buffer descriptor */ 124f6569884SThomas Chou #define TX_BD_CS (1 << 0) /* carrier sense lost */ 125f6569884SThomas Chou #define TX_BD_DF (1 << 1) /* defer indication */ 126f6569884SThomas Chou #define TX_BD_LC (1 << 2) /* late collision */ 127f6569884SThomas Chou #define TX_BD_RL (1 << 3) /* retransmission limit */ 128f6569884SThomas Chou #define TX_BD_RETRY_MASK (0x00f0) 129f6569884SThomas Chou #define TX_BD_RETRY(x) (((x) & 0x00f0) >> 4) 130f6569884SThomas Chou #define TX_BD_UR (1 << 8) /* transmitter underrun */ 131f6569884SThomas Chou #define TX_BD_CRC (1 << 11) /* TX CRC enable */ 132f6569884SThomas Chou #define TX_BD_PAD (1 << 12) /* pad enable */ 133f6569884SThomas Chou #define TX_BD_WRAP (1 << 13) 134f6569884SThomas Chou #define TX_BD_IRQ (1 << 14) /* interrupt request enable */ 135f6569884SThomas Chou #define TX_BD_READY (1 << 15) /* TX buffer ready */ 136f6569884SThomas Chou #define TX_BD_LEN(x) (((x) & 0xffff) << 16) 137f6569884SThomas Chou #define TX_BD_LEN_MASK (0xffff << 16) 138f6569884SThomas Chou 139f6569884SThomas Chou #define TX_BD_STATS (TX_BD_CS | TX_BD_DF | TX_BD_LC | \ 140f6569884SThomas Chou TX_BD_RL | TX_BD_RETRY_MASK | TX_BD_UR) 141f6569884SThomas Chou 142f6569884SThomas Chou /* RX buffer descriptor */ 143f6569884SThomas Chou #define RX_BD_LC (1 << 0) /* late collision */ 144f6569884SThomas Chou #define RX_BD_CRC (1 << 1) /* RX CRC error */ 145f6569884SThomas Chou #define RX_BD_SF (1 << 2) /* short frame */ 146f6569884SThomas Chou #define RX_BD_TL (1 << 3) /* too long */ 147f6569884SThomas Chou #define RX_BD_DN (1 << 4) /* dribble nibble */ 148f6569884SThomas Chou #define RX_BD_IS (1 << 5) /* invalid symbol */ 149f6569884SThomas Chou #define RX_BD_OR (1 << 6) /* receiver overrun */ 150f6569884SThomas Chou #define RX_BD_MISS (1 << 7) 151f6569884SThomas Chou #define RX_BD_CF (1 << 8) /* control frame */ 152f6569884SThomas Chou #define RX_BD_WRAP (1 << 13) 153f6569884SThomas Chou #define RX_BD_IRQ (1 << 14) /* interrupt request enable */ 154f6569884SThomas Chou #define RX_BD_EMPTY (1 << 15) 155f6569884SThomas Chou #define RX_BD_LEN(x) (((x) & 0xffff) << 16) 156f6569884SThomas Chou 157f6569884SThomas Chou #define RX_BD_STATS (RX_BD_LC | RX_BD_CRC | RX_BD_SF | RX_BD_TL | \ 158f6569884SThomas Chou RX_BD_DN | RX_BD_IS | RX_BD_OR | RX_BD_MISS) 159f6569884SThomas Chou 160f6569884SThomas Chou #define ETHOC_BUFSIZ 1536 161f6569884SThomas Chou #define ETHOC_ZLEN 64 162f6569884SThomas Chou #define ETHOC_BD_BASE 0x400 163f6569884SThomas Chou #define ETHOC_TIMEOUT (HZ / 2) 164f6569884SThomas Chou #define ETHOC_MII_TIMEOUT (1 + (HZ / 5)) 165a84a757aSMax Filippov #define ETHOC_IOSIZE 0x54 166f6569884SThomas Chou 167f6569884SThomas Chou /** 168f6569884SThomas Chou * struct ethoc - driver-private device structure 169f6569884SThomas Chou * @num_tx: number of send buffers 170f6569884SThomas Chou * @cur_tx: last send buffer written 171f6569884SThomas Chou * @dty_tx: last buffer actually sent 172f6569884SThomas Chou * @num_rx: number of receive buffers 173f6569884SThomas Chou * @cur_rx: current receive buffer 174f6569884SThomas Chou */ 175f6569884SThomas Chou struct ethoc { 176f6569884SThomas Chou u32 num_tx; 177f6569884SThomas Chou u32 cur_tx; 178f6569884SThomas Chou u32 dty_tx; 179f6569884SThomas Chou u32 num_rx; 180f6569884SThomas Chou u32 cur_rx; 181a84a757aSMax Filippov void __iomem *iobase; 182f6569884SThomas Chou }; 183f6569884SThomas Chou 184f6569884SThomas Chou /** 185f6569884SThomas Chou * struct ethoc_bd - buffer descriptor 186f6569884SThomas Chou * @stat: buffer statistics 187f6569884SThomas Chou * @addr: physical memory address 188f6569884SThomas Chou */ 189f6569884SThomas Chou struct ethoc_bd { 190f6569884SThomas Chou u32 stat; 191f6569884SThomas Chou u32 addr; 192f6569884SThomas Chou }; 193f6569884SThomas Chou 194a84a757aSMax Filippov static inline u32 ethoc_read(struct ethoc *priv, size_t offset) 195f6569884SThomas Chou { 196a84a757aSMax Filippov return readl(priv->iobase + offset); 197f6569884SThomas Chou } 198f6569884SThomas Chou 199a84a757aSMax Filippov static inline void ethoc_write(struct ethoc *priv, size_t offset, u32 data) 200f6569884SThomas Chou { 201a84a757aSMax Filippov writel(data, priv->iobase + offset); 202f6569884SThomas Chou } 203f6569884SThomas Chou 204a84a757aSMax Filippov static inline void ethoc_read_bd(struct ethoc *priv, int index, 205f6569884SThomas Chou struct ethoc_bd *bd) 206f6569884SThomas Chou { 2079f680d2dSVasili Galka size_t offset = ETHOC_BD_BASE + (index * sizeof(struct ethoc_bd)); 208a84a757aSMax Filippov bd->stat = ethoc_read(priv, offset + 0); 209a84a757aSMax Filippov bd->addr = ethoc_read(priv, offset + 4); 210f6569884SThomas Chou } 211f6569884SThomas Chou 212a84a757aSMax Filippov static inline void ethoc_write_bd(struct ethoc *priv, int index, 213f6569884SThomas Chou const struct ethoc_bd *bd) 214f6569884SThomas Chou { 2159f680d2dSVasili Galka size_t offset = ETHOC_BD_BASE + (index * sizeof(struct ethoc_bd)); 216a84a757aSMax Filippov ethoc_write(priv, offset + 0, bd->stat); 217a84a757aSMax Filippov ethoc_write(priv, offset + 4, bd->addr); 218f6569884SThomas Chou } 219f6569884SThomas Chou 220*5d43feabSMax Filippov static int ethoc_write_hwaddr_common(struct ethoc *priv, u8 *mac) 221f6569884SThomas Chou { 222a84a757aSMax Filippov ethoc_write(priv, MAC_ADDR0, (mac[2] << 24) | (mac[3] << 16) | 223f6569884SThomas Chou (mac[4] << 8) | (mac[5] << 0)); 224a84a757aSMax Filippov ethoc_write(priv, MAC_ADDR1, (mac[0] << 8) | (mac[1] << 0)); 2253ac9d6c6SThomas Chou return 0; 226f6569884SThomas Chou } 227f6569884SThomas Chou 228a84a757aSMax Filippov static inline void ethoc_ack_irq(struct ethoc *priv, u32 mask) 229f6569884SThomas Chou { 230a84a757aSMax Filippov ethoc_write(priv, INT_SOURCE, mask); 231f6569884SThomas Chou } 232f6569884SThomas Chou 233a84a757aSMax Filippov static inline void ethoc_enable_rx_and_tx(struct ethoc *priv) 234f6569884SThomas Chou { 235a84a757aSMax Filippov u32 mode = ethoc_read(priv, MODER); 236f6569884SThomas Chou mode |= MODER_RXEN | MODER_TXEN; 237a84a757aSMax Filippov ethoc_write(priv, MODER, mode); 238f6569884SThomas Chou } 239f6569884SThomas Chou 240a84a757aSMax Filippov static inline void ethoc_disable_rx_and_tx(struct ethoc *priv) 241f6569884SThomas Chou { 242a84a757aSMax Filippov u32 mode = ethoc_read(priv, MODER); 243f6569884SThomas Chou mode &= ~(MODER_RXEN | MODER_TXEN); 244a84a757aSMax Filippov ethoc_write(priv, MODER, mode); 245f6569884SThomas Chou } 246f6569884SThomas Chou 247a84a757aSMax Filippov static int ethoc_init_ring(struct ethoc *priv) 248f6569884SThomas Chou { 249f6569884SThomas Chou struct ethoc_bd bd; 250f6569884SThomas Chou int i; 251f6569884SThomas Chou 252f6569884SThomas Chou priv->cur_tx = 0; 253f6569884SThomas Chou priv->dty_tx = 0; 254f6569884SThomas Chou priv->cur_rx = 0; 255f6569884SThomas Chou 256f6569884SThomas Chou /* setup transmission buffers */ 257f6569884SThomas Chou bd.stat = TX_BD_IRQ | TX_BD_CRC; 258f6569884SThomas Chou 259f6569884SThomas Chou for (i = 0; i < priv->num_tx; i++) { 260f6569884SThomas Chou if (i == priv->num_tx - 1) 261f6569884SThomas Chou bd.stat |= TX_BD_WRAP; 262f6569884SThomas Chou 263a84a757aSMax Filippov ethoc_write_bd(priv, i, &bd); 264f6569884SThomas Chou } 265f6569884SThomas Chou 266f6569884SThomas Chou bd.stat = RX_BD_EMPTY | RX_BD_IRQ; 267f6569884SThomas Chou 268f6569884SThomas Chou for (i = 0; i < priv->num_rx; i++) { 2691fd92db8SJoe Hershberger bd.addr = (u32)net_rx_packets[i]; 270f6569884SThomas Chou if (i == priv->num_rx - 1) 271f6569884SThomas Chou bd.stat |= RX_BD_WRAP; 272f6569884SThomas Chou 27383ea1308SStefan Kristiansson flush_dcache_range(bd.addr, bd.addr + PKTSIZE_ALIGN); 274a84a757aSMax Filippov ethoc_write_bd(priv, priv->num_tx + i, &bd); 275f6569884SThomas Chou } 276f6569884SThomas Chou 277f6569884SThomas Chou return 0; 278f6569884SThomas Chou } 279f6569884SThomas Chou 280a84a757aSMax Filippov static int ethoc_reset(struct ethoc *priv) 281f6569884SThomas Chou { 282f6569884SThomas Chou u32 mode; 283f6569884SThomas Chou 284f6569884SThomas Chou /* TODO: reset controller? */ 285f6569884SThomas Chou 286a84a757aSMax Filippov ethoc_disable_rx_and_tx(priv); 287f6569884SThomas Chou 288f6569884SThomas Chou /* TODO: setup registers */ 289f6569884SThomas Chou 290f6569884SThomas Chou /* enable FCS generation and automatic padding */ 291a84a757aSMax Filippov mode = ethoc_read(priv, MODER); 292f6569884SThomas Chou mode |= MODER_CRC | MODER_PAD; 293a84a757aSMax Filippov ethoc_write(priv, MODER, mode); 294f6569884SThomas Chou 295f6569884SThomas Chou /* set full-duplex mode */ 296a84a757aSMax Filippov mode = ethoc_read(priv, MODER); 297f6569884SThomas Chou mode |= MODER_FULLD; 298a84a757aSMax Filippov ethoc_write(priv, MODER, mode); 299a84a757aSMax Filippov ethoc_write(priv, IPGT, 0x15); 300f6569884SThomas Chou 301a84a757aSMax Filippov ethoc_ack_irq(priv, INT_MASK_ALL); 302a84a757aSMax Filippov ethoc_enable_rx_and_tx(priv); 303f6569884SThomas Chou return 0; 304f6569884SThomas Chou } 305f6569884SThomas Chou 306*5d43feabSMax Filippov static int ethoc_init_common(struct ethoc *priv) 307f6569884SThomas Chou { 308f6569884SThomas Chou priv->num_tx = 1; 309f6569884SThomas Chou priv->num_rx = PKTBUFSRX; 310a84a757aSMax Filippov ethoc_write(priv, TX_BD_NUM, priv->num_tx); 311a84a757aSMax Filippov ethoc_init_ring(priv); 312a84a757aSMax Filippov ethoc_reset(priv); 313f6569884SThomas Chou 314f6569884SThomas Chou return 0; 315f6569884SThomas Chou } 316f6569884SThomas Chou 317f6569884SThomas Chou static int ethoc_update_rx_stats(struct ethoc_bd *bd) 318f6569884SThomas Chou { 319f6569884SThomas Chou int ret = 0; 320f6569884SThomas Chou 321f6569884SThomas Chou if (bd->stat & RX_BD_TL) { 322f6569884SThomas Chou debug("ETHOC: " "RX: frame too long\n"); 323f6569884SThomas Chou ret++; 324f6569884SThomas Chou } 325f6569884SThomas Chou 326f6569884SThomas Chou if (bd->stat & RX_BD_SF) { 327f6569884SThomas Chou debug("ETHOC: " "RX: frame too short\n"); 328f6569884SThomas Chou ret++; 329f6569884SThomas Chou } 330f6569884SThomas Chou 331f6569884SThomas Chou if (bd->stat & RX_BD_DN) 332f6569884SThomas Chou debug("ETHOC: " "RX: dribble nibble\n"); 333f6569884SThomas Chou 334f6569884SThomas Chou if (bd->stat & RX_BD_CRC) { 335f6569884SThomas Chou debug("ETHOC: " "RX: wrong CRC\n"); 336f6569884SThomas Chou ret++; 337f6569884SThomas Chou } 338f6569884SThomas Chou 339f6569884SThomas Chou if (bd->stat & RX_BD_OR) { 340f6569884SThomas Chou debug("ETHOC: " "RX: overrun\n"); 341f6569884SThomas Chou ret++; 342f6569884SThomas Chou } 343f6569884SThomas Chou 344f6569884SThomas Chou if (bd->stat & RX_BD_LC) { 345f6569884SThomas Chou debug("ETHOC: " "RX: late collision\n"); 346f6569884SThomas Chou ret++; 347f6569884SThomas Chou } 348f6569884SThomas Chou 349f6569884SThomas Chou return ret; 350f6569884SThomas Chou } 351f6569884SThomas Chou 352*5d43feabSMax Filippov static int ethoc_rx_common(struct ethoc *priv, uchar **packetp) 353f6569884SThomas Chou { 354f6569884SThomas Chou u32 entry; 355f6569884SThomas Chou struct ethoc_bd bd; 356f6569884SThomas Chou 357f6569884SThomas Chou entry = priv->num_tx + (priv->cur_rx % priv->num_rx); 358a84a757aSMax Filippov ethoc_read_bd(priv, entry, &bd); 359f6569884SThomas Chou if (bd.stat & RX_BD_EMPTY) 360*5d43feabSMax Filippov return -EAGAIN; 361f6569884SThomas Chou 362f6569884SThomas Chou debug("%s(): RX buffer %d, %x received\n", 363f6569884SThomas Chou __func__, priv->cur_rx, bd.stat); 364f6569884SThomas Chou if (ethoc_update_rx_stats(&bd) == 0) { 365f6569884SThomas Chou int size = bd.stat >> 16; 366*5d43feabSMax Filippov 367f6569884SThomas Chou size -= 4; /* strip the CRC */ 368*5d43feabSMax Filippov *packetp = (void *)bd.addr; 369*5d43feabSMax Filippov return size; 370*5d43feabSMax Filippov } else { 371*5d43feabSMax Filippov return 0; 372*5d43feabSMax Filippov } 373f6569884SThomas Chou } 374f6569884SThomas Chou 375*5d43feabSMax Filippov static int ethoc_is_new_packet_received(struct ethoc *priv) 376*5d43feabSMax Filippov { 377*5d43feabSMax Filippov u32 pending; 378*5d43feabSMax Filippov 379*5d43feabSMax Filippov pending = ethoc_read(priv, INT_SOURCE); 380*5d43feabSMax Filippov ethoc_ack_irq(priv, pending); 381*5d43feabSMax Filippov if (pending & INT_MASK_BUSY) 382*5d43feabSMax Filippov debug("%s(): packet dropped\n", __func__); 383*5d43feabSMax Filippov if (pending & INT_MASK_RX) { 384*5d43feabSMax Filippov debug("%s(): rx irq\n", __func__); 385*5d43feabSMax Filippov return 1; 386f6569884SThomas Chou } 387f6569884SThomas Chou 388*5d43feabSMax Filippov return 0; 389f6569884SThomas Chou } 390f6569884SThomas Chou 391f6569884SThomas Chou static int ethoc_update_tx_stats(struct ethoc_bd *bd) 392f6569884SThomas Chou { 393f6569884SThomas Chou if (bd->stat & TX_BD_LC) 394f6569884SThomas Chou debug("ETHOC: " "TX: late collision\n"); 395f6569884SThomas Chou 396f6569884SThomas Chou if (bd->stat & TX_BD_RL) 397f6569884SThomas Chou debug("ETHOC: " "TX: retransmit limit\n"); 398f6569884SThomas Chou 399f6569884SThomas Chou if (bd->stat & TX_BD_UR) 400f6569884SThomas Chou debug("ETHOC: " "TX: underrun\n"); 401f6569884SThomas Chou 402f6569884SThomas Chou if (bd->stat & TX_BD_CS) 403f6569884SThomas Chou debug("ETHOC: " "TX: carrier sense lost\n"); 404f6569884SThomas Chou 405f6569884SThomas Chou return 0; 406f6569884SThomas Chou } 407f6569884SThomas Chou 408a84a757aSMax Filippov static void ethoc_tx(struct ethoc *priv) 409f6569884SThomas Chou { 410f6569884SThomas Chou u32 entry = priv->dty_tx % priv->num_tx; 411f6569884SThomas Chou struct ethoc_bd bd; 412f6569884SThomas Chou 413a84a757aSMax Filippov ethoc_read_bd(priv, entry, &bd); 414f6569884SThomas Chou if ((bd.stat & TX_BD_READY) == 0) 415f6569884SThomas Chou (void)ethoc_update_tx_stats(&bd); 416f6569884SThomas Chou } 417f6569884SThomas Chou 418*5d43feabSMax Filippov static int ethoc_send_common(struct ethoc *priv, void *packet, int length) 419f6569884SThomas Chou { 420f6569884SThomas Chou struct ethoc_bd bd; 421f6569884SThomas Chou u32 entry; 422f6569884SThomas Chou u32 pending; 423f6569884SThomas Chou int tmo; 424f6569884SThomas Chou 425f6569884SThomas Chou entry = priv->cur_tx % priv->num_tx; 426a84a757aSMax Filippov ethoc_read_bd(priv, entry, &bd); 427f6569884SThomas Chou if (unlikely(length < ETHOC_ZLEN)) 428f6569884SThomas Chou bd.stat |= TX_BD_PAD; 429f6569884SThomas Chou else 430f6569884SThomas Chou bd.stat &= ~TX_BD_PAD; 431f6569884SThomas Chou bd.addr = (u32)packet; 432f6569884SThomas Chou 43383ea1308SStefan Kristiansson flush_dcache_range(bd.addr, bd.addr + length); 434f6569884SThomas Chou bd.stat &= ~(TX_BD_STATS | TX_BD_LEN_MASK); 435f6569884SThomas Chou bd.stat |= TX_BD_LEN(length); 436a84a757aSMax Filippov ethoc_write_bd(priv, entry, &bd); 437f6569884SThomas Chou 438f6569884SThomas Chou /* start transmit */ 439f6569884SThomas Chou bd.stat |= TX_BD_READY; 440a84a757aSMax Filippov ethoc_write_bd(priv, entry, &bd); 441f6569884SThomas Chou 442f6569884SThomas Chou /* wait for transfer to succeed */ 443f6569884SThomas Chou tmo = get_timer(0) + 5 * CONFIG_SYS_HZ; 444f6569884SThomas Chou while (1) { 445a84a757aSMax Filippov pending = ethoc_read(priv, INT_SOURCE); 446a84a757aSMax Filippov ethoc_ack_irq(priv, pending & ~INT_MASK_RX); 447f6569884SThomas Chou if (pending & INT_MASK_BUSY) 448f6569884SThomas Chou debug("%s(): packet dropped\n", __func__); 449f6569884SThomas Chou 450f6569884SThomas Chou if (pending & INT_MASK_TX) { 451a84a757aSMax Filippov ethoc_tx(priv); 452f6569884SThomas Chou break; 453f6569884SThomas Chou } 454f6569884SThomas Chou if (get_timer(0) >= tmo) { 455f6569884SThomas Chou debug("%s(): timed out\n", __func__); 456f6569884SThomas Chou return -1; 457f6569884SThomas Chou } 458f6569884SThomas Chou } 459f6569884SThomas Chou 460f6569884SThomas Chou debug("%s(): packet sent\n", __func__); 461f6569884SThomas Chou return 0; 462f6569884SThomas Chou } 463f6569884SThomas Chou 464*5d43feabSMax Filippov static int ethoc_free_pkt_common(struct ethoc *priv) 465*5d43feabSMax Filippov { 466*5d43feabSMax Filippov u32 entry; 467*5d43feabSMax Filippov struct ethoc_bd bd; 468*5d43feabSMax Filippov 469*5d43feabSMax Filippov entry = priv->num_tx + (priv->cur_rx % priv->num_rx); 470*5d43feabSMax Filippov ethoc_read_bd(priv, entry, &bd); 471*5d43feabSMax Filippov 472*5d43feabSMax Filippov /* clear the buffer descriptor so it can be reused */ 473*5d43feabSMax Filippov flush_dcache_range(bd.addr, bd.addr + PKTSIZE_ALIGN); 474*5d43feabSMax Filippov bd.stat &= ~RX_BD_STATS; 475*5d43feabSMax Filippov bd.stat |= RX_BD_EMPTY; 476*5d43feabSMax Filippov ethoc_write_bd(priv, entry, &bd); 477*5d43feabSMax Filippov priv->cur_rx++; 478*5d43feabSMax Filippov 479*5d43feabSMax Filippov return 0; 480*5d43feabSMax Filippov } 481*5d43feabSMax Filippov 482*5d43feabSMax Filippov #ifdef CONFIG_DM_ETH 483*5d43feabSMax Filippov 484*5d43feabSMax Filippov static int ethoc_write_hwaddr(struct udevice *dev) 485*5d43feabSMax Filippov { 486*5d43feabSMax Filippov struct ethoc_eth_pdata *pdata = dev_get_platdata(dev); 487*5d43feabSMax Filippov struct ethoc *priv = dev_get_priv(dev); 488*5d43feabSMax Filippov u8 *mac = pdata->eth_pdata.enetaddr; 489*5d43feabSMax Filippov 490*5d43feabSMax Filippov return ethoc_write_hwaddr_common(priv, mac); 491*5d43feabSMax Filippov } 492*5d43feabSMax Filippov 493*5d43feabSMax Filippov static int ethoc_send(struct udevice *dev, void *packet, int length) 494*5d43feabSMax Filippov { 495*5d43feabSMax Filippov return ethoc_send_common(dev_get_priv(dev), packet, length); 496*5d43feabSMax Filippov } 497*5d43feabSMax Filippov 498*5d43feabSMax Filippov static int ethoc_free_pkt(struct udevice *dev, uchar *packet, int length) 499*5d43feabSMax Filippov { 500*5d43feabSMax Filippov return ethoc_free_pkt_common(dev_get_priv(dev)); 501*5d43feabSMax Filippov } 502*5d43feabSMax Filippov 503*5d43feabSMax Filippov static int ethoc_recv(struct udevice *dev, int flags, uchar **packetp) 504*5d43feabSMax Filippov { 505*5d43feabSMax Filippov struct ethoc *priv = dev_get_priv(dev); 506*5d43feabSMax Filippov 507*5d43feabSMax Filippov if (flags & ETH_RECV_CHECK_DEVICE) 508*5d43feabSMax Filippov if (!ethoc_is_new_packet_received(priv)) 509*5d43feabSMax Filippov return -EAGAIN; 510*5d43feabSMax Filippov 511*5d43feabSMax Filippov return ethoc_rx_common(priv, packetp); 512*5d43feabSMax Filippov } 513*5d43feabSMax Filippov 514*5d43feabSMax Filippov static int ethoc_start(struct udevice *dev) 515*5d43feabSMax Filippov { 516*5d43feabSMax Filippov return ethoc_init_common(dev_get_priv(dev)); 517*5d43feabSMax Filippov } 518*5d43feabSMax Filippov 519*5d43feabSMax Filippov static void ethoc_stop(struct udevice *dev) 520*5d43feabSMax Filippov { 521*5d43feabSMax Filippov struct ethoc *priv = dev_get_priv(dev); 522*5d43feabSMax Filippov 523*5d43feabSMax Filippov ethoc_disable_rx_and_tx(priv); 524*5d43feabSMax Filippov } 525*5d43feabSMax Filippov 526*5d43feabSMax Filippov static int ethoc_probe(struct udevice *dev) 527*5d43feabSMax Filippov { 528*5d43feabSMax Filippov struct ethoc_eth_pdata *pdata = dev_get_platdata(dev); 529*5d43feabSMax Filippov struct ethoc *priv = dev_get_priv(dev); 530*5d43feabSMax Filippov 531*5d43feabSMax Filippov priv->iobase = ioremap(pdata->eth_pdata.iobase, ETHOC_IOSIZE); 532*5d43feabSMax Filippov return 0; 533*5d43feabSMax Filippov } 534*5d43feabSMax Filippov 535*5d43feabSMax Filippov static int ethoc_remove(struct udevice *dev) 536*5d43feabSMax Filippov { 537*5d43feabSMax Filippov struct ethoc *priv = dev_get_priv(dev); 538*5d43feabSMax Filippov 539*5d43feabSMax Filippov iounmap(priv->iobase); 540*5d43feabSMax Filippov return 0; 541*5d43feabSMax Filippov } 542*5d43feabSMax Filippov 543*5d43feabSMax Filippov static const struct eth_ops ethoc_ops = { 544*5d43feabSMax Filippov .start = ethoc_start, 545*5d43feabSMax Filippov .stop = ethoc_stop, 546*5d43feabSMax Filippov .send = ethoc_send, 547*5d43feabSMax Filippov .recv = ethoc_recv, 548*5d43feabSMax Filippov .free_pkt = ethoc_free_pkt, 549*5d43feabSMax Filippov .write_hwaddr = ethoc_write_hwaddr, 550*5d43feabSMax Filippov }; 551*5d43feabSMax Filippov 552*5d43feabSMax Filippov U_BOOT_DRIVER(ethoc) = { 553*5d43feabSMax Filippov .name = "ethoc", 554*5d43feabSMax Filippov .id = UCLASS_ETH, 555*5d43feabSMax Filippov .probe = ethoc_probe, 556*5d43feabSMax Filippov .remove = ethoc_remove, 557*5d43feabSMax Filippov .ops = ðoc_ops, 558*5d43feabSMax Filippov .priv_auto_alloc_size = sizeof(struct ethoc), 559*5d43feabSMax Filippov .platdata_auto_alloc_size = sizeof(struct ethoc_eth_pdata), 560*5d43feabSMax Filippov }; 561*5d43feabSMax Filippov 562*5d43feabSMax Filippov #else 563*5d43feabSMax Filippov 564*5d43feabSMax Filippov static int ethoc_init(struct eth_device *dev, bd_t *bd) 565*5d43feabSMax Filippov { 566*5d43feabSMax Filippov struct ethoc *priv = (struct ethoc *)dev->priv; 567*5d43feabSMax Filippov 568*5d43feabSMax Filippov return ethoc_init_common(priv); 569*5d43feabSMax Filippov } 570*5d43feabSMax Filippov 571*5d43feabSMax Filippov static int ethoc_write_hwaddr(struct eth_device *dev) 572*5d43feabSMax Filippov { 573*5d43feabSMax Filippov struct ethoc *priv = (struct ethoc *)dev->priv; 574*5d43feabSMax Filippov u8 *mac = dev->enetaddr; 575*5d43feabSMax Filippov 576*5d43feabSMax Filippov return ethoc_write_hwaddr_common(priv, mac); 577*5d43feabSMax Filippov } 578*5d43feabSMax Filippov 579*5d43feabSMax Filippov static int ethoc_send(struct eth_device *dev, void *packet, int length) 580*5d43feabSMax Filippov { 581*5d43feabSMax Filippov return ethoc_send_common(dev->priv, packet, length); 582*5d43feabSMax Filippov } 583*5d43feabSMax Filippov 584f6569884SThomas Chou static void ethoc_halt(struct eth_device *dev) 585f6569884SThomas Chou { 586a84a757aSMax Filippov ethoc_disable_rx_and_tx(dev->priv); 587f6569884SThomas Chou } 588f6569884SThomas Chou 589f6569884SThomas Chou static int ethoc_recv(struct eth_device *dev) 590f6569884SThomas Chou { 591a84a757aSMax Filippov struct ethoc *priv = (struct ethoc *)dev->priv; 592*5d43feabSMax Filippov int count; 593f6569884SThomas Chou 594*5d43feabSMax Filippov if (!ethoc_is_new_packet_received(priv)) 595*5d43feabSMax Filippov return 0; 596*5d43feabSMax Filippov 597*5d43feabSMax Filippov for (count = 0; count < PKTBUFSRX; ++count) { 598*5d43feabSMax Filippov uchar *packetp; 599*5d43feabSMax Filippov int size = ethoc_rx_common(priv, &packetp); 600*5d43feabSMax Filippov 601*5d43feabSMax Filippov if (size < 0) 602*5d43feabSMax Filippov break; 603*5d43feabSMax Filippov if (size > 0) 604*5d43feabSMax Filippov net_process_received_packet(packetp, size); 605*5d43feabSMax Filippov ethoc_free_pkt_common(priv); 606f6569884SThomas Chou } 607f6569884SThomas Chou return 0; 608f6569884SThomas Chou } 609f6569884SThomas Chou 610f6569884SThomas Chou int ethoc_initialize(u8 dev_num, int base_addr) 611f6569884SThomas Chou { 612f6569884SThomas Chou struct ethoc *priv; 613f6569884SThomas Chou struct eth_device *dev; 614f6569884SThomas Chou 615f6569884SThomas Chou priv = malloc(sizeof(*priv)); 616f6569884SThomas Chou if (!priv) 617f6569884SThomas Chou return 0; 618f6569884SThomas Chou dev = malloc(sizeof(*dev)); 619f6569884SThomas Chou if (!dev) { 620f6569884SThomas Chou free(priv); 621f6569884SThomas Chou return 0; 622f6569884SThomas Chou } 623f6569884SThomas Chou 624f6569884SThomas Chou memset(dev, 0, sizeof(*dev)); 625f6569884SThomas Chou dev->priv = priv; 626f6569884SThomas Chou dev->iobase = base_addr; 627f6569884SThomas Chou dev->init = ethoc_init; 628f6569884SThomas Chou dev->halt = ethoc_halt; 629f6569884SThomas Chou dev->send = ethoc_send; 630f6569884SThomas Chou dev->recv = ethoc_recv; 631*5d43feabSMax Filippov dev->write_hwaddr = ethoc_write_hwaddr; 632f6569884SThomas Chou sprintf(dev->name, "%s-%hu", "ETHOC", dev_num); 633a84a757aSMax Filippov priv->iobase = ioremap(dev->iobase, ETHOC_IOSIZE); 634f6569884SThomas Chou 635f6569884SThomas Chou eth_register(dev); 636f6569884SThomas Chou return 1; 637f6569884SThomas Chou } 638*5d43feabSMax Filippov 639*5d43feabSMax Filippov #endif 640