Lines Matching refs:rdata
1114 struct xgbe_ring_data *rdata; in xgbe_free_tx_data() local
1125 rdata = XGBE_GET_DESC_DATA(ring, j); in xgbe_free_tx_data()
1126 desc_if->unmap_rdata(pdata, rdata); in xgbe_free_tx_data()
1137 struct xgbe_ring_data *rdata; in xgbe_free_rx_data() local
1148 rdata = XGBE_GET_DESC_DATA(ring, j); in xgbe_free_rx_data()
1149 desc_if->unmap_rdata(pdata, rdata); in xgbe_free_rx_data()
2317 struct xgbe_ring_data *rdata; in xgbe_rx_refresh() local
2320 rdata = XGBE_GET_DESC_DATA(ring, ring->dirty); in xgbe_rx_refresh()
2323 desc_if->unmap_rdata(pdata, rdata); in xgbe_rx_refresh()
2325 if (desc_if->map_rx_buffer(pdata, ring, rdata)) in xgbe_rx_refresh()
2328 hw_if->rx_desc_reset(pdata, rdata, ring->dirty); in xgbe_rx_refresh()
2338 rdata = XGBE_GET_DESC_DATA(ring, ring->dirty - 1); in xgbe_rx_refresh()
2340 lower_32_bits(rdata->rdesc_dma)); in xgbe_rx_refresh()
2345 struct xgbe_ring_data *rdata, in xgbe_create_skb() argument
2351 skb = napi_alloc_skb(napi, rdata->rx.hdr.dma_len); in xgbe_create_skb()
2358 dma_sync_single_range_for_cpu(pdata->dev, rdata->rx.hdr.dma_base, in xgbe_create_skb()
2359 rdata->rx.hdr.dma_off, in xgbe_create_skb()
2360 rdata->rx.hdr.dma_len, DMA_FROM_DEVICE); in xgbe_create_skb()
2362 packet = page_address(rdata->rx.hdr.pa.pages) + in xgbe_create_skb()
2363 rdata->rx.hdr.pa.pages_offset; in xgbe_create_skb()
2370 static unsigned int xgbe_rx_buf1_len(struct xgbe_ring_data *rdata, in xgbe_rx_buf1_len() argument
2378 if (rdata->rx.hdr_len) in xgbe_rx_buf1_len()
2379 return rdata->rx.hdr_len; in xgbe_rx_buf1_len()
2385 return rdata->rx.hdr.dma_len; in xgbe_rx_buf1_len()
2390 return min_t(unsigned int, rdata->rx.hdr.dma_len, rdata->rx.len); in xgbe_rx_buf1_len()
2393 static unsigned int xgbe_rx_buf2_len(struct xgbe_ring_data *rdata, in xgbe_rx_buf2_len() argument
2399 return rdata->rx.buf.dma_len; in xgbe_rx_buf2_len()
2404 return rdata->rx.len - len; in xgbe_rx_buf2_len()
2413 struct xgbe_ring_data *rdata; in xgbe_tx_poll() local
2436 rdata = XGBE_GET_DESC_DATA(ring, ring->dirty); in xgbe_tx_poll()
2437 rdesc = rdata->rdesc; in xgbe_tx_poll()
2450 tx_packets += rdata->tx.packets; in xgbe_tx_poll()
2451 tx_bytes += rdata->tx.bytes; in xgbe_tx_poll()
2455 desc_if->unmap_rdata(pdata, rdata); in xgbe_tx_poll()
2456 hw_if->tx_desc_reset(rdata); in xgbe_tx_poll()
2483 struct xgbe_ring_data *rdata; in xgbe_rx_poll() local
2505 rdata = XGBE_GET_DESC_DATA(ring, ring->cur); in xgbe_rx_poll()
2511 if (!received && rdata->state_saved) { in xgbe_rx_poll()
2512 skb = rdata->state.skb; in xgbe_rx_poll()
2513 error = rdata->state.error; in xgbe_rx_poll()
2514 len = rdata->state.len; in xgbe_rx_poll()
2523 rdata = XGBE_GET_DESC_DATA(ring, ring->cur); in xgbe_rx_poll()
2557 buf1_len = xgbe_rx_buf1_len(rdata, packet); in xgbe_rx_poll()
2559 buf2_len = xgbe_rx_buf2_len(rdata, packet, len); in xgbe_rx_poll()
2562 if (buf2_len > rdata->rx.buf.dma_len) { in xgbe_rx_poll()
2571 skb = xgbe_create_skb(pdata, napi, rdata, in xgbe_rx_poll()
2581 rdata->rx.buf.dma_base, in xgbe_rx_poll()
2582 rdata->rx.buf.dma_off, in xgbe_rx_poll()
2583 rdata->rx.buf.dma_len, in xgbe_rx_poll()
2587 rdata->rx.buf.pa.pages, in xgbe_rx_poll()
2588 rdata->rx.buf.pa.pages_offset, in xgbe_rx_poll()
2590 rdata->rx.buf.dma_len); in xgbe_rx_poll()
2591 rdata->rx.buf.pa.pages = NULL; in xgbe_rx_poll()
2666 rdata = XGBE_GET_DESC_DATA(ring, ring->cur); in xgbe_rx_poll()
2667 rdata->state_saved = 1; in xgbe_rx_poll()
2668 rdata->state.skb = skb; in xgbe_rx_poll()
2669 rdata->state.len = len; in xgbe_rx_poll()
2670 rdata->state.error = error; in xgbe_rx_poll()
2750 struct xgbe_ring_data *rdata; in xgbe_dump_tx_desc() local
2754 rdata = XGBE_GET_DESC_DATA(ring, idx); in xgbe_dump_tx_desc()
2755 rdesc = rdata->rdesc; in xgbe_dump_tx_desc()
2770 struct xgbe_ring_data *rdata; in xgbe_dump_rx_desc() local
2773 rdata = XGBE_GET_DESC_DATA(ring, idx); in xgbe_dump_rx_desc()
2774 rdesc = rdata->rdesc; in xgbe_dump_rx_desc()