Lines Matching refs:umem
45 struct xdp_umem *umem) in xp_create_and_assign_umem() argument
51 pool = kvzalloc(struct_size(pool, free_heads, umem->chunks), in xp_create_and_assign_umem()
56 pool->heads = kvcalloc(umem->chunks, sizeof(*pool->heads), GFP_KERNEL); in xp_create_and_assign_umem()
60 pool->chunk_mask = ~((u64)umem->chunk_size - 1); in xp_create_and_assign_umem()
61 pool->addrs_cnt = umem->size; in xp_create_and_assign_umem()
62 pool->heads_cnt = umem->chunks; in xp_create_and_assign_umem()
63 pool->free_heads_cnt = umem->chunks; in xp_create_and_assign_umem()
64 pool->headroom = umem->headroom; in xp_create_and_assign_umem()
65 pool->chunk_size = umem->chunk_size; in xp_create_and_assign_umem()
66 pool->unaligned = umem->flags & XDP_UMEM_UNALIGNED_CHUNK_FLAG; in xp_create_and_assign_umem()
67 pool->frame_len = umem->chunk_size - umem->headroom - in xp_create_and_assign_umem()
69 pool->umem = umem; in xp_create_and_assign_umem()
70 pool->addrs = umem->addrs; in xp_create_and_assign_umem()
83 xskb->xdp.frame_sz = umem->chunk_size - umem->headroom; in xp_create_and_assign_umem()
110 if (pool->umem->zc) { in xp_disable_drv_zc()
180 pool->umem->zc = true; in __xp_assign_dev()
205 struct xdp_umem *umem = umem_xs->umem; in xp_assign_dev_shared() local
211 flags = umem->zc ? XDP_ZEROCOPY : XDP_COPY; in xp_assign_dev_shared()
248 xdp_put_umem(pool->umem, false); in xp_release_deferred()
275 list_for_each_entry(dma_map, &pool->umem->xsk_dma_list, list) { in xp_find_dma_map()
284 u32 nr_pages, struct xdp_umem *umem) in xp_create_dma_map() argument
303 list_add(&dma_map->list, &umem->xsk_dma_list); in xp_create_dma_map()
400 dma_map = xp_create_dma_map(dev, pool->netdev, nr_pages, pool->umem); in xp_dma_map()