Lines Matching refs:xdpf
272 struct xdp_frame *xdpf; member
667 bytes += desc->xdpf->len; in netsec_clean_tx_dring()
668 xdp_return_frame(desc->xdpf); in netsec_clean_tx_dring()
807 dring->desc[idx].xdpf = buf; in netsec_set_tx_de()
815 struct xdp_frame *xdpf, bool is_ndo) in netsec_xdp_queue_one() argument
819 struct page *page = virt_to_page(xdpf->data); in netsec_xdp_queue_one()
837 dma_handle = dma_map_single(priv->dev, xdpf->data, xdpf->len, in netsec_xdp_queue_one()
851 dma_handle = page_pool_get_dma_addr(page) + xdpf->headroom + in netsec_xdp_queue_one()
852 sizeof(*xdpf); in netsec_xdp_queue_one()
853 dma_sync_single_for_device(priv->dev, dma_handle, xdpf->len, in netsec_xdp_queue_one()
859 tx_desc.addr = xdpf->data; in netsec_xdp_queue_one()
860 tx_desc.len = xdpf->len; in netsec_xdp_queue_one()
862 netdev_sent_queue(priv->ndev, xdpf->len); in netsec_xdp_queue_one()
863 netsec_set_tx_de(priv, tx_ring, &tx_ctrl, &tx_desc, xdpf); in netsec_xdp_queue_one()
871 struct xdp_frame *xdpf = xdp_convert_buff_to_frame(xdp); in netsec_xdp_xmit_back() local
874 if (unlikely(!xdpf)) in netsec_xdp_xmit_back()
878 ret = netsec_xdp_queue_one(priv, xdpf, false); in netsec_xdp_xmit_back()
1764 struct xdp_frame *xdpf = frames[i]; in netsec_xdp_xmit() local
1767 err = netsec_xdp_queue_one(priv, xdpf, true); in netsec_xdp_xmit()
1769 xdp_return_frame_rx_napi(xdpf); in netsec_xdp_xmit()