| /OK3568_Linux_fs/kernel/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_xsk.c | 256 xdp = xsk_buff_alloc(rx_ring->xsk_pool); in i40e_alloc_rx_buffers_zc() 387 xsk_buff_dma_sync_for_cpu(*bi, rx_ring->xsk_pool); in i40e_clean_rx_irq_zc() 438 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in i40e_clean_rx_irq_zc() 440 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in i40e_clean_rx_irq_zc() 442 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in i40e_clean_rx_irq_zc() 465 if (!xsk_tx_peek_desc(xdp_ring->xsk_pool, &desc)) in i40e_xmit_zc() 468 dma = xsk_buff_raw_get_dma(xdp_ring->xsk_pool, desc.addr); in i40e_xmit_zc() 469 xsk_buff_raw_dma_sync_for_device(xdp_ring->xsk_pool, dma, in i40e_xmit_zc() 496 xsk_tx_release(xdp_ring->xsk_pool); in i40e_xmit_zc() 528 struct xsk_buff_pool *bp = tx_ring->xsk_pool; in i40e_clean_xdp_tx_irq() [all …]
|
| H A D | i40e_txrx.h | 391 struct xsk_buff_pool *xsk_pool; member
|
| H A D | i40e_txrx.c | 639 if (ring_is_xdp(tx_ring) && tx_ring->xsk_pool) { in i40e_clean_tx_ring() 1330 if (rx_ring->xsk_pool) { in i40e_clean_rx_ring() 1364 if (rx_ring->xsk_pool) in i40e_clean_rx_ring() 2599 bool wd = ring->xsk_pool ? in i40e_napi_poll() 2627 int cleaned = ring->xsk_pool ? in i40e_napi_poll()
|
| H A D | i40e_main.c | 3290 ring->xsk_pool = i40e_xsk_pool(ring); in i40e_configure_tx_ring() 3412 ring->xsk_pool = i40e_xsk_pool(ring); in i40e_configure_rx_ring() 3413 if (ring->xsk_pool) { in i40e_configure_rx_ring() 3415 xsk_pool_get_rx_frame_size(ring->xsk_pool); in i40e_configure_rx_ring() 3495 if (ring->xsk_pool) { in i40e_configure_rx_ring() 3496 xsk_pool_set_rxq_info(ring->xsk_pool, &ring->xdp_rxq); in i40e_configure_rx_ring() 3507 ring->xsk_pool ? "AF_XDP ZC enabled " : "", in i40e_configure_rx_ring() 12716 if (vsi->xdp_rings[i]->xsk_pool) in i40e_xdp_setup()
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
| H A D | rx.h | 25 dma_info->xsk = xsk_buff_alloc(rq->xsk_pool); in mlx5e_xsk_page_alloc_pool() 41 if (!xsk_uses_need_wakeup(rq->xsk_pool)) in mlx5e_xsk_update_rx_wakeup() 45 xsk_set_rx_need_wakeup(rq->xsk_pool); in mlx5e_xsk_update_rx_wakeup() 47 xsk_clear_rx_need_wakeup(rq->xsk_pool); in mlx5e_xsk_update_rx_wakeup()
|
| H A D | tx.h | 18 if (!xsk_uses_need_wakeup(sq->xsk_pool)) in mlx5e_xsk_update_tx_wakeup() 22 xsk_clear_tx_need_wakeup(sq->xsk_pool); in mlx5e_xsk_update_tx_wakeup() 24 xsk_set_tx_need_wakeup(sq->xsk_pool); in mlx5e_xsk_update_tx_wakeup()
|
| H A D | rx.c | 50 xsk_buff_dma_sync_for_cpu(xdp, rq->xsk_pool); in mlx5e_xsk_skb_from_cqe_mpwrq_linear() 96 xsk_buff_dma_sync_for_cpu(xdp, rq->xsk_pool); in mlx5e_xsk_skb_from_cqe_linear()
|
| H A D | tx.c | 69 struct xsk_buff_pool *pool = sq->xsk_pool; in mlx5e_xsk_tx()
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/intel/ice/ |
| H A D | ice_xsk.c | 241 xdp_ring->xsk_pool = ice_xsk_pool(xdp_ring); in ice_qp_ena() 444 rx_buf->xdp = xsk_buff_alloc(rx_ring->xsk_pool); in ice_alloc_rx_bufs_zc() 625 xsk_buff_dma_sync_for_cpu(rx_buf->xdp, rx_ring->xsk_pool); in ice_clean_rx_irq_zc() 675 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in ice_clean_rx_irq_zc() 677 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in ice_clean_rx_irq_zc() 679 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in ice_clean_rx_irq_zc() 712 if (!xsk_tx_peek_desc(xdp_ring->xsk_pool, &desc)) in ice_xmit_zc() 715 dma = xsk_buff_raw_get_dma(xdp_ring->xsk_pool, desc.addr); in ice_xmit_zc() 716 xsk_buff_raw_dma_sync_for_device(xdp_ring->xsk_pool, dma, in ice_xmit_zc() 733 xsk_tx_release(xdp_ring->xsk_pool); in ice_xmit_zc() [all …]
|
| H A D | ice_base.c | 311 ring->xsk_pool = ice_xsk_pool(ring); in ice_setup_rx_ctx() 312 if (ring->xsk_pool) { in ice_setup_rx_ctx() 316 xsk_pool_get_rx_frame_size(ring->xsk_pool); in ice_setup_rx_ctx() 327 xsk_pool_set_rxq_info(ring->xsk_pool, &ring->xdp_rxq); in ice_setup_rx_ctx() 420 if (ring->xsk_pool) { in ice_setup_rx_ctx() 421 if (!xsk_buff_can_alloc(ring->xsk_pool, num_bufs)) { in ice_setup_rx_ctx()
|
| H A D | ice_txrx.h | 298 struct xsk_buff_pool *xsk_pool; member
|
| H A D | ice_txrx.c | 148 if (ice_ring_is_xdp(tx_ring) && tx_ring->xsk_pool) { in ice_clean_tx_ring() 378 if (rx_ring->xsk_pool) { in ice_clean_rx_ring() 1627 bool wd = ring->xsk_pool ? in ice_napi_poll() 1657 cleaned = ring->xsk_pool ? in ice_napi_poll()
|
| H A D | ice_lib.c | 1772 vsi->xdp_rings[i]->xsk_pool = ice_xsk_pool(vsi->xdp_rings[i]); in ice_vsi_cfg_xdp_txqs()
|
| H A D | ice_main.c | 2294 xdp_ring->xsk_pool = ice_xsk_pool(xdp_ring); in ice_xdp_alloc_setup_rings() 2562 if (rx_ring->xsk_pool) in ice_xdp_setup_prog()
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_xsk.c | 161 bi->xdp = xsk_buff_alloc(rx_ring->xsk_pool); in ixgbe_alloc_rx_buffers_zc() 298 xsk_buff_dma_sync_for_cpu(bi->xdp, rx_ring->xsk_pool); in ixgbe_clean_rx_irq_zc() 356 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in ixgbe_clean_rx_irq_zc() 358 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in ixgbe_clean_rx_irq_zc() 360 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in ixgbe_clean_rx_irq_zc() 385 struct xsk_buff_pool *pool = xdp_ring->xsk_pool; in ixgbe_xmit_zc() 453 struct xsk_buff_pool *pool = tx_ring->xsk_pool; in ixgbe_clean_xdp_tx_irq() 525 if (!ring->xsk_pool) in ixgbe_xsk_wakeup() 540 struct xsk_buff_pool *pool = tx_ring->xsk_pool; in ixgbe_xsk_clean_tx_ring()
|
| H A D | ixgbe_main.c | 3165 bool wd = ring->xsk_pool ? in ixgbe_poll() 3185 int cleaned = ring->xsk_pool ? in ixgbe_poll() 3480 ring->xsk_pool = NULL; in ixgbe_configure_tx_ring() 3482 ring->xsk_pool = ixgbe_xsk_pool(adapter, ring); in ixgbe_configure_tx_ring() 3722 if (rx_ring->xsk_pool) { in ixgbe_configure_srrctl() 3723 u32 xsk_buf_len = xsk_pool_get_rx_frame_size(rx_ring->xsk_pool); in ixgbe_configure_srrctl() 4068 ring->xsk_pool = ixgbe_xsk_pool(adapter, ring); in ixgbe_configure_rx_ring() 4069 if (ring->xsk_pool) { in ixgbe_configure_rx_ring() 4073 xsk_pool_set_rxq_info(ring->xsk_pool, &ring->xdp_rxq); in ixgbe_configure_rx_ring() 4128 if (ring->xsk_pool && hw->mac.type != ixgbe_mac_82599EB) { in ixgbe_configure_rx_ring() [all …]
|
| H A D | ixgbe.h | 353 struct xsk_buff_pool *xsk_pool; member
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_main.c | 396 struct xsk_buff_pool *xsk_pool, in mlx5e_alloc_rq() argument 422 rq->xsk_pool = xsk_pool; in mlx5e_alloc_rq() 424 if (rq->xsk_pool) in mlx5e_alloc_rq() 514 xsk_pool_set_rxq_info(rq->xsk_pool, &rq->xdp_rxq); in mlx5e_alloc_rq() 864 struct xsk_buff_pool *xsk_pool, struct mlx5e_rq *rq) in mlx5e_open_rq() argument 868 err = mlx5e_alloc_rq(c, params, xsk, xsk_pool, param, rq); in mlx5e_open_rq() 980 struct xsk_buff_pool *xsk_pool, in mlx5e_alloc_xdpsq() argument 996 sq->xsk_pool = xsk_pool; in mlx5e_alloc_xdpsq() 998 sq->stats = sq->xsk_pool ? in mlx5e_alloc_xdpsq() 1484 struct mlx5e_sq_param *param, struct xsk_buff_pool *xsk_pool, in mlx5e_open_xdpsq() argument [all …]
|
| H A D | en.h | 460 struct xsk_buff_pool *xsk_pool; member 624 struct xsk_buff_pool *xsk_pool; member 913 struct xsk_buff_pool *xsk_pool, struct mlx5e_rq *rq); 920 struct mlx5e_sq_param *param, struct xsk_buff_pool *xsk_pool,
|
| H A D | en_rx.c | 293 if (rq->xsk_pool) in mlx5e_page_alloc() 326 if (rq->xsk_pool) in mlx5e_page_release() 413 if (rq->xsk_pool) { in mlx5e_alloc_rx_wqes() 420 if (unlikely(!xsk_buff_can_alloc(rq->xsk_pool, pages_desired))) in mlx5e_alloc_rx_wqes() 517 if (rq->xsk_pool && in mlx5e_alloc_rx_mpwqe() 518 unlikely(!xsk_buff_can_alloc(rq->xsk_pool, MLX5_MPWRQ_PAGES_PER_WQE))) { in mlx5e_alloc_rx_mpwqe() 767 if (unlikely(alloc_err == -ENOMEM && rq->xsk_pool)) in mlx5e_post_rx_mpwqes()
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | xdp.c | 451 xsk_tx_completed(sq->xsk_pool, xsk_frames); in mlx5e_poll_xdpsq_cq() 481 xsk_tx_completed(sq->xsk_pool, xsk_frames); in mlx5e_free_xdpsq_descs()
|