Lines Matching refs:rx_ring

288 struct rx_ring {  struct
492 struct rx_ring rx_ring; member
740 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_enable() local
742 if (rx_ring->fbr[1]->buffsize == 4096) in et131x_rx_dma_enable()
744 else if (rx_ring->fbr[1]->buffsize == 8192) in et131x_rx_dma_enable()
746 else if (rx_ring->fbr[1]->buffsize == 16384) in et131x_rx_dma_enable()
750 if (rx_ring->fbr[0]->buffsize == 256) in et131x_rx_dma_enable()
752 else if (rx_ring->fbr[0]->buffsize == 512) in et131x_rx_dma_enable()
754 else if (rx_ring->fbr[0]->buffsize == 1024) in et131x_rx_dma_enable()
1539 struct rx_ring *rx_local = &adapter->rx_ring; in et131x_config_rx_dma_regs()
1865 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_memory_alloc() local
1869 rx_ring->fbr[0] = kzalloc(sizeof(*fbr), GFP_KERNEL); in et131x_rx_dma_memory_alloc()
1870 if (rx_ring->fbr[0] == NULL) in et131x_rx_dma_memory_alloc()
1872 rx_ring->fbr[1] = kzalloc(sizeof(*fbr), GFP_KERNEL); in et131x_rx_dma_memory_alloc()
1873 if (rx_ring->fbr[1] == NULL) in et131x_rx_dma_memory_alloc()
1894 rx_ring->fbr[0]->buffsize = 256; in et131x_rx_dma_memory_alloc()
1895 rx_ring->fbr[0]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1896 rx_ring->fbr[1]->buffsize = 2048; in et131x_rx_dma_memory_alloc()
1897 rx_ring->fbr[1]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1899 rx_ring->fbr[0]->buffsize = 512; in et131x_rx_dma_memory_alloc()
1900 rx_ring->fbr[0]->num_entries = 1024; in et131x_rx_dma_memory_alloc()
1901 rx_ring->fbr[1]->buffsize = 4096; in et131x_rx_dma_memory_alloc()
1902 rx_ring->fbr[1]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1904 rx_ring->fbr[0]->buffsize = 1024; in et131x_rx_dma_memory_alloc()
1905 rx_ring->fbr[0]->num_entries = 768; in et131x_rx_dma_memory_alloc()
1906 rx_ring->fbr[1]->buffsize = 16384; in et131x_rx_dma_memory_alloc()
1907 rx_ring->fbr[1]->num_entries = 128; in et131x_rx_dma_memory_alloc()
1910 rx_ring->psr_entries = rx_ring->fbr[0]->num_entries + in et131x_rx_dma_memory_alloc()
1911 rx_ring->fbr[1]->num_entries; in et131x_rx_dma_memory_alloc()
1914 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_alloc()
1930 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_alloc()
1970 psr_size = sizeof(struct pkt_stat_desc) * rx_ring->psr_entries; in et131x_rx_dma_memory_alloc()
1972 rx_ring->ps_ring_virtaddr = dma_alloc_coherent(&adapter->pdev->dev, in et131x_rx_dma_memory_alloc()
1974 &rx_ring->ps_ring_physaddr, in et131x_rx_dma_memory_alloc()
1977 if (!rx_ring->ps_ring_virtaddr) { in et131x_rx_dma_memory_alloc()
1984 rx_ring->rx_status_block = dma_alloc_coherent(&adapter->pdev->dev, in et131x_rx_dma_memory_alloc()
1986 &rx_ring->rx_status_bus, in et131x_rx_dma_memory_alloc()
1988 if (!rx_ring->rx_status_block) { in et131x_rx_dma_memory_alloc()
1993 rx_ring->num_rfd = NIC_DEFAULT_NUM_RFD; in et131x_rx_dma_memory_alloc()
1998 INIT_LIST_HEAD(&rx_ring->recv_list); in et131x_rx_dma_memory_alloc()
2009 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_memory_free() local
2013 WARN_ON(rx_ring->num_ready_recv != rx_ring->num_rfd); in et131x_rx_dma_memory_free()
2015 while (!list_empty(&rx_ring->recv_list)) { in et131x_rx_dma_memory_free()
2016 rfd = list_entry(rx_ring->recv_list.next, in et131x_rx_dma_memory_free()
2026 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_free()
2056 if (rx_ring->ps_ring_virtaddr) { in et131x_rx_dma_memory_free()
2057 psr_size = sizeof(struct pkt_stat_desc) * rx_ring->psr_entries; in et131x_rx_dma_memory_free()
2060 rx_ring->ps_ring_virtaddr, in et131x_rx_dma_memory_free()
2061 rx_ring->ps_ring_physaddr); in et131x_rx_dma_memory_free()
2063 rx_ring->ps_ring_virtaddr = NULL; in et131x_rx_dma_memory_free()
2067 if (rx_ring->rx_status_block) { in et131x_rx_dma_memory_free()
2070 rx_ring->rx_status_block, in et131x_rx_dma_memory_free()
2071 rx_ring->rx_status_bus); in et131x_rx_dma_memory_free()
2072 rx_ring->rx_status_block = NULL; in et131x_rx_dma_memory_free()
2076 kfree(rx_ring->fbr[0]); in et131x_rx_dma_memory_free()
2077 kfree(rx_ring->fbr[1]); in et131x_rx_dma_memory_free()
2080 rx_ring->num_ready_recv = 0; in et131x_rx_dma_memory_free()
2088 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_init_recv() local
2091 for (rfdct = 0; rfdct < rx_ring->num_rfd; rfdct++) { in et131x_init_recv()
2099 list_add_tail(&rfd->list_node, &rx_ring->recv_list); in et131x_init_recv()
2102 rx_ring->num_ready_recv++; in et131x_init_recv()
2125 struct rx_ring *rx_local = &adapter->rx_ring; in nic_return_rfd()
2184 struct rx_ring *rx_local = &adapter->rx_ring; in nic_rx_pkts()
2298 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_handle_recv_pkts() local
2305 if (list_empty(&rx_ring->recv_list)) { in et131x_handle_recv_pkts()
2306 WARN_ON(rx_ring->num_ready_recv != 0); in et131x_handle_recv_pkts()
2328 if (rx_ring->num_ready_recv < RFD_LOW_WATER_MARK) in et131x_handle_recv_pkts()
2335 rx_ring->unfinished_receives = true; in et131x_handle_recv_pkts()
2340 rx_ring->unfinished_receives = false; in et131x_handle_recv_pkts()
3364 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_isr() local
3398 if (rx_ring->unfinished_receives) in et131x_isr()