| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/ |
| H A D | mali_kbase_mem_pool.c | 37 #define pool_dbg(pool, format, ...) \ argument 38 dev_dbg(pool->kbdev->dev, "%s-pool [%zu/%zu]: " format, \ 39 (pool->next_pool) ? "kctx" : "kbdev", \ 40 kbase_mem_pool_size(pool), \ 41 kbase_mem_pool_max_size(pool), \ 50 * @pool: Pointer to the memory pool. 73 static inline bool can_alloc_page(struct kbase_mem_pool *pool, struct task_struct *page_owner, in can_alloc_page() argument 80 dev_info(pool->kbdev->dev, "%s : Process %s/%d exiting", in can_alloc_page() 88 static size_t kbase_mem_pool_capacity(struct kbase_mem_pool *pool) in kbase_mem_pool_capacity() argument 90 ssize_t max_size = kbase_mem_pool_max_size(pool); in kbase_mem_pool_capacity() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/ |
| H A D | mali_kbase_mem_pool.c | 27 #define pool_dbg(pool, format, ...) \ argument 28 dev_dbg(pool->kbdev->dev, "%s-pool [%zu/%zu]: " format, \ 29 (pool->next_pool) ? "kctx" : "kbdev", \ 30 kbase_mem_pool_size(pool), \ 31 kbase_mem_pool_max_size(pool), \ 37 static inline void kbase_mem_pool_lock(struct kbase_mem_pool *pool) in kbase_mem_pool_lock() argument 39 spin_lock(&pool->pool_lock); in kbase_mem_pool_lock() 42 static inline void kbase_mem_pool_unlock(struct kbase_mem_pool *pool) in kbase_mem_pool_unlock() argument 44 spin_unlock(&pool->pool_lock); in kbase_mem_pool_unlock() 47 static size_t kbase_mem_pool_capacity(struct kbase_mem_pool *pool) in kbase_mem_pool_capacity() argument [all …]
|
| /OK3568_Linux_fs/kernel/net/xdp/ |
| H A D | xsk_buff_pool.c | 11 void xp_add_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs) in xp_add_xsk() argument 18 spin_lock_irqsave(&pool->xsk_tx_list_lock, flags); in xp_add_xsk() 19 list_add_rcu(&xs->tx_list, &pool->xsk_tx_list); in xp_add_xsk() 20 spin_unlock_irqrestore(&pool->xsk_tx_list_lock, flags); in xp_add_xsk() 23 void xp_del_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs) in xp_del_xsk() argument 30 spin_lock_irqsave(&pool->xsk_tx_list_lock, flags); in xp_del_xsk() 32 spin_unlock_irqrestore(&pool->xsk_tx_list_lock, flags); in xp_del_xsk() 35 void xp_destroy(struct xsk_buff_pool *pool) in xp_destroy() argument 37 if (!pool) in xp_destroy() 40 kvfree(pool->heads); in xp_destroy() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/infiniband/sw/rxe/ |
| H A D | rxe_pool.c | 81 static inline const char *pool_name(struct rxe_pool *pool) in pool_name() argument 83 return rxe_type_info[pool->type].name; in pool_name() 86 static int rxe_pool_init_index(struct rxe_pool *pool, u32 max, u32 min) in rxe_pool_init_index() argument 91 if ((max - min + 1) < pool->max_elem) { in rxe_pool_init_index() 97 pool->max_index = max; in rxe_pool_init_index() 98 pool->min_index = min; in rxe_pool_init_index() 101 pool->table = kmalloc(size, GFP_KERNEL); in rxe_pool_init_index() 102 if (!pool->table) { in rxe_pool_init_index() 107 pool->table_size = size; in rxe_pool_init_index() 108 bitmap_zero(pool->table, max - min + 1); in rxe_pool_init_index() [all …]
|
| /OK3568_Linux_fs/kernel/mm/ |
| H A D | mempool.c | 5 * memory buffer pool support. Such pools are mostly used 25 static void poison_error(mempool_t *pool, void *element, size_t size, in poison_error() argument 28 const int nr = pool->curr_nr; in poison_error() 34 pr_err("Mempool %p size %zu\n", pool, size); in poison_error() 42 static void __check_element(mempool_t *pool, void *element, size_t size) in __check_element() argument 51 poison_error(pool, element, size, i); in __check_element() 58 static void check_element(mempool_t *pool, void *element) in check_element() argument 61 if (pool->free == mempool_free_slab || pool->free == mempool_kfree) { in check_element() 62 __check_element(pool, element, ksize(element)); in check_element() 63 } else if (pool->free == mempool_free_pages) { in check_element() [all …]
|
| H A D | dmapool.c | 3 * DMA Pool allocator 14 * The current design of this allocator is fairly simple. The pool is 42 struct dma_pool { /* the pool */ 71 struct dma_pool *pool; in show_pools() local 81 list_for_each_entry(pool, &dev->dma_pools, pools) { in show_pools() 85 spin_lock_irq(&pool->lock); in show_pools() 86 list_for_each_entry(page, &pool->page_list, page_list) { in show_pools() 90 spin_unlock_irq(&pool->lock); in show_pools() 92 /* per-pool info, no real statistics yet */ in show_pools() 94 pool->name, blocks, in show_pools() [all …]
|
| H A D | zbud.c | 63 * freelists maintained in each pool. NCHUNKS_ORDER of 6 means that the 67 * 63 freelists per pool. 77 * struct zbud_pool - stores metadata for each zbud pool 78 * @lock: protects all pool fields and first|last_chunk fields of any 79 * zbud page in the pool 87 * @pages_nr: number of zbud pages in the pool. 89 * pool creation time. 91 * This structure is allocated at pool creation time and maintains metadata 92 * pertaining to a particular zbud pool. 110 * @buddy: links the zbud page into the unbuddied/buddied lists in the pool [all …]
|
| H A D | z3fold.c | 51 * freelists maintained in each pool. NCHUNKS_ORDER of 6 means that the 56 * be 63, or 62, respectively, freelists per pool. 76 int (*evict)(struct z3fold_pool *pool, unsigned long handle); 93 unsigned long pool; /* back link */ member 102 * pool 107 * @pool: pointer to the containing pool 121 struct z3fold_pool *pool; member 133 * struct z3fold_pool - stores metadata for each z3fold pool 134 * @name: pool name 135 * @lock: protects pool unbuddied/lru lists [all …]
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/ti/ |
| H A D | k3-cppi-desc-pool.c | 2 /* TI K3 CPPI5 descriptors pool API 15 #include "k3-cppi-desc-pool.h" 27 void k3_cppi_desc_pool_destroy(struct k3_cppi_desc_pool *pool) in k3_cppi_desc_pool_destroy() argument 29 if (!pool) in k3_cppi_desc_pool_destroy() 32 WARN(gen_pool_size(pool->gen_pool) != gen_pool_avail(pool->gen_pool), in k3_cppi_desc_pool_destroy() 34 gen_pool_size(pool->gen_pool), in k3_cppi_desc_pool_destroy() 35 gen_pool_avail(pool->gen_pool)); in k3_cppi_desc_pool_destroy() 36 if (pool->cpumem) in k3_cppi_desc_pool_destroy() 37 dma_free_coherent(pool->dev, pool->mem_size, pool->cpumem, in k3_cppi_desc_pool_destroy() 38 pool->dma_addr); in k3_cppi_desc_pool_destroy() [all …]
|
| /OK3568_Linux_fs/kernel/net/core/ |
| H A D | page_pool.c | 24 static int page_pool_init(struct page_pool *pool, in page_pool_init() argument 29 memcpy(&pool->p, params, sizeof(pool->p)); in page_pool_init() 32 if (pool->p.flags & ~(PP_FLAG_ALL)) in page_pool_init() 35 if (pool->p.pool_size) in page_pool_init() 36 ring_qsize = pool->p.pool_size; in page_pool_init() 46 if (pool->p.flags & PP_FLAG_DMA_MAP) { in page_pool_init() 47 if ((pool->p.dma_dir != DMA_FROM_DEVICE) && in page_pool_init() 48 (pool->p.dma_dir != DMA_BIDIRECTIONAL)) in page_pool_init() 52 if (pool->p.flags & PP_FLAG_DMA_SYNC_DEV) { in page_pool_init() 56 if (!(pool->p.flags & PP_FLAG_DMA_MAP)) in page_pool_init() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/md/ |
| H A D | dm-thin.c | 41 * The block size of the device holding pool data must be 191 * A pool device ties together a metadata device and a data device. It 198 * The pool runs in various modes. Ordered in degraded order for comparisons. 229 struct pool { struct 231 struct dm_target *ti; /* Only set if a pool target is bound */ argument 289 static void metadata_operation_failed(struct pool *pool, const char *op, int r); argument 291 static enum pool_mode get_pool_mode(struct pool *pool) in get_pool_mode() argument 293 return pool->pf.mode; in get_pool_mode() 296 static void notify_of_pool_mode_change(struct pool *pool) in notify_of_pool_mode_change() argument 306 enum pool_mode mode = get_pool_mode(pool); in notify_of_pool_mode_change() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/staging/android/ion/heaps/ |
| H A D | ion_page_pool.c | 3 * ION Memory Allocator page pool helpers 15 static inline struct page *ion_page_pool_alloc_pages(struct ion_page_pool *pool) in ion_page_pool_alloc_pages() argument 19 return alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages() 22 static void ion_page_pool_free_pages(struct ion_page_pool *pool, in ion_page_pool_free_pages() argument 25 __free_pages(page, pool->order); in ion_page_pool_free_pages() 28 static void ion_page_pool_add(struct ion_page_pool *pool, struct page *page) in ion_page_pool_add() argument 30 mutex_lock(&pool->mutex); in ion_page_pool_add() 32 list_add_tail(&page->lru, &pool->high_items); in ion_page_pool_add() 33 pool->high_count++; in ion_page_pool_add() 35 list_add_tail(&page->lru, &pool->low_items); in ion_page_pool_add() [all …]
|
| /OK3568_Linux_fs/kernel/sound/core/seq/ |
| H A D | seq_memory.c | 22 static inline int snd_seq_pool_available(struct snd_seq_pool *pool) in snd_seq_pool_available() argument 24 return pool->total_elements - atomic_read(&pool->counter); in snd_seq_pool_available() 27 static inline int snd_seq_output_ok(struct snd_seq_pool *pool) in snd_seq_output_ok() argument 29 return snd_seq_pool_available(pool) >= pool->room; in snd_seq_output_ok() 166 static inline void free_cell(struct snd_seq_pool *pool, in free_cell() argument 169 cell->next = pool->free; in free_cell() 170 pool->free = cell; in free_cell() 171 atomic_dec(&pool->counter); in free_cell() 177 struct snd_seq_pool *pool; in snd_seq_cell_free() local 181 pool = cell->pool; in snd_seq_cell_free() [all …]
|
| /OK3568_Linux_fs/kernel/lib/ |
| H A D | genalloc.c | 16 * available. If new memory is added to the pool a lock has to be 145 * gen_pool_create - create a new special memory pool 147 * @nid: node id of the node the pool structure should be allocated on, or -1 149 * Create a new special memory pool that can be used to manage special purpose 154 struct gen_pool *pool; in gen_pool_create() local 156 pool = kmalloc_node(sizeof(struct gen_pool), GFP_KERNEL, nid); in gen_pool_create() 157 if (pool != NULL) { in gen_pool_create() 158 spin_lock_init(&pool->lock); in gen_pool_create() 159 INIT_LIST_HEAD(&pool->chunks); in gen_pool_create() 160 pool->min_alloc_order = min_alloc_order; in gen_pool_create() [all …]
|
| /OK3568_Linux_fs/kernel/include/net/ |
| H A D | xdp_sock_drv.h | 14 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries); 15 bool xsk_tx_peek_desc(struct xsk_buff_pool *pool, struct xdp_desc *desc); 16 void xsk_tx_release(struct xsk_buff_pool *pool); 19 void xsk_set_rx_need_wakeup(struct xsk_buff_pool *pool); 20 void xsk_set_tx_need_wakeup(struct xsk_buff_pool *pool); 21 void xsk_clear_rx_need_wakeup(struct xsk_buff_pool *pool); 22 void xsk_clear_tx_need_wakeup(struct xsk_buff_pool *pool); 23 bool xsk_uses_need_wakeup(struct xsk_buff_pool *pool); 25 static inline u32 xsk_pool_get_headroom(struct xsk_buff_pool *pool) in xsk_pool_get_headroom() argument 27 return XDP_PACKET_HEADROOM + pool->headroom; in xsk_pool_get_headroom() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/dma-buf/heaps/ |
| H A D | page_pool.c | 3 * DMA BUF page pool system 7 * Based on the ION page pool code 20 struct dmabuf_page_pool pool; member 28 struct page *dmabuf_page_pool_alloc_pages(struct dmabuf_page_pool *pool) in dmabuf_page_pool_alloc_pages() argument 32 return alloc_pages(pool->gfp_mask, pool->order); in dmabuf_page_pool_alloc_pages() 35 static inline void dmabuf_page_pool_free_pages(struct dmabuf_page_pool *pool, in dmabuf_page_pool_free_pages() argument 38 __free_pages(page, pool->order); in dmabuf_page_pool_free_pages() 41 static void dmabuf_page_pool_add(struct dmabuf_page_pool *pool, struct page *page) in dmabuf_page_pool_add() argument 45 container_of(pool, struct dmabuf_page_pool_with_spinlock, pool); in dmabuf_page_pool_add() 53 list_add_tail(&page->lru, &pool->items[index]); in dmabuf_page_pool_add() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/ttm/ |
| H A D | ttm_page_alloc_dma.c | 27 * A simple DMA pool losely based on dmapool.c. It has certain advantages 29 * - Pool collects resently freed pages for reuse (and hooks up to 70 * The pool structure. There are up to nine pools: 80 * @type: Type of the pool 82 * used with irqsave/irqrestore variants because pool allocator maybe called 84 * @free_list: Pool of pages that are free to be used. No order requirements. 89 * @nfrees: Stats when pool is shrinking. 90 * @nrefills: Stats when the pool is grown. 92 * @name: Name of the pool. 117 * huge pool [all …]
|
| /OK3568_Linux_fs/kernel/drivers/staging/media/atomisp/pci/runtime/rmgr/src/ |
| H A D | rmgr_vbuf.c | 31 * @brief VBUF resource pool - refpool 42 * @brief VBUF resource pool - writepool 53 * @brief VBUF resource pool - hmmbufferpool 137 * @brief Initialize the resource pool (host, vbuf) 139 * @param pool The pointer to the pool 141 int ia_css_rmgr_init_vbuf(struct ia_css_rmgr_vbuf_pool *pool) in ia_css_rmgr_init_vbuf() argument 147 assert(pool); in ia_css_rmgr_init_vbuf() 148 if (!pool) in ia_css_rmgr_init_vbuf() 150 /* initialize the recycle pool if used */ in ia_css_rmgr_init_vbuf() 151 if (pool->recycle && pool->size) { in ia_css_rmgr_init_vbuf() [all …]
|
| /OK3568_Linux_fs/kernel/net/ceph/ |
| H A D | msgpool.c | 14 struct ceph_msgpool *pool = arg; in msgpool_alloc() local 17 msg = ceph_msg_new2(pool->type, pool->front_len, pool->max_data_items, in msgpool_alloc() 20 dout("msgpool_alloc %s failed\n", pool->name); in msgpool_alloc() 22 dout("msgpool_alloc %s %p\n", pool->name, msg); in msgpool_alloc() 23 msg->pool = pool; in msgpool_alloc() 30 struct ceph_msgpool *pool = arg; in msgpool_free() local 33 dout("msgpool_release %s %p\n", pool->name, msg); in msgpool_free() 34 msg->pool = NULL; in msgpool_free() 38 int ceph_msgpool_init(struct ceph_msgpool *pool, int type, in ceph_msgpool_init() argument 43 pool->type = type; in ceph_msgpool_init() [all …]
|
| /OK3568_Linux_fs/external/rkwifibt/drivers/bcmdhd/include/ |
| H A D | bcm_mpool_pub.h | 14 * pool manager: A singleton object that acts as a factory for 15 * pool allocators. It also is used for global 17 * in use across all data structures. The pool manager 21 * memory pool: An object for allocating homogenous memory blocks. 24 * bcm_mpm_* Memory pool manager 25 * bcm_mp_* Memory pool 33 * heap: The memory pool allocator uses the heap (malloc/free) for memory. 34 * In this case, the pool allocator is just providing statistics 78 * Opaque type definition for the pool manager handle. This object is used for global 79 * memory pool operations such as obtaining a new pool, deleting a pool, iterating and [all …]
|
| /OK3568_Linux_fs/kernel/drivers/net/wireless/rockchip_wlan/rkwifi/bcmdhd/include/ |
| H A D | bcm_mpool_pub.h | 14 * pool manager: A singleton object that acts as a factory for 15 * pool allocators. It also is used for global 17 * in use across all data structures. The pool manager 21 * memory pool: An object for allocating homogenous memory blocks. 24 * bcm_mpm_* Memory pool manager 25 * bcm_mp_* Memory pool 33 * heap: The memory pool allocator uses the heap (malloc/free) for memory. 34 * In this case, the pool allocator is just providing statistics 78 * Opaque type definition for the pool manager handle. This object is used for global 79 * memory pool operations such as obtaining a new pool, deleting a pool, iterating and [all …]
|
| /OK3568_Linux_fs/kernel/tools/testing/selftests/drivers/net/mlxsw/ |
| H A D | sharedbuffer_configuration.py | 16 objects, pool, tcbind and portpool. Provide an interface to get random 18 1. Pool: 22 - random pool number 30 for pool in pools: 31 self._pools.append(pool) 47 def _get_th(self, pool): argument 50 if pool["thtype"] == "dynamic": 58 for pool in self._pools: 59 if pool["type"] == "ingress": 60 ing_pools.append(pool) [all …]
|
| /OK3568_Linux_fs/kernel/drivers/staging/android/ion/ |
| H A D | ion_protected_heap.c | 35 pr_info("ION: created protected pool at %pa, size %ld MiB\n", in protected_dma_setup() 41 RESERVEDMEM_OF_DECLARE(protected, "protected-dma-pool", protected_dma_setup); 60 static void protected_pool_add(struct ion_page_pool *pool, struct page *page) in protected_pool_add() argument 62 mutex_lock(&pool->mutex); in protected_pool_add() 64 list_add_tail(&page->lru, &pool->high_items); in protected_pool_add() 65 pool->high_count++; in protected_pool_add() 67 list_add_tail(&page->lru, &pool->low_items); in protected_pool_add() 68 pool->low_count++; in protected_pool_add() 70 mutex_unlock(&pool->mutex); in protected_pool_add() 73 static struct page *protected_pool_remove(struct ion_page_pool *pool, in protected_pool_remove() argument [all …]
|
| /OK3568_Linux_fs/kernel/drivers/net/wireless/rockchip_wlan/infineon/bcmdhd/include/ |
| H A D | bcm_mpool_pub.h | 14 * pool manager: A singleton object that acts as a factory for 15 * pool allocators. It also is used for global 17 * in use across all data structures. The pool manager 21 * memory pool: An object for allocating homogenous memory blocks. 24 * bcm_mpm_* Memory pool manager 25 * bcm_mp_* Memory pool 33 * heap: The memory pool allocator uses the heap (malloc/free) for memory. 34 * In this case, the pool allocator is just providing statistics 86 * Opaque type definition for the pool manager handle. This object is used for global 87 * memory pool operations such as obtaining a new pool, deleting a pool, iterating and [all …]
|
| /OK3568_Linux_fs/external/rkwifibt/drivers/infineon/include/ |
| H A D | bcm_mpool_pub.h | 14 * pool manager: A singleton object that acts as a factory for 15 * pool allocators. It also is used for global 17 * in use across all data structures. The pool manager 21 * memory pool: An object for allocating homogenous memory blocks. 24 * bcm_mpm_* Memory pool manager 25 * bcm_mp_* Memory pool 33 * heap: The memory pool allocator uses the heap (malloc/free) for memory. 34 * In this case, the pool allocator is just providing statistics 86 * Opaque type definition for the pool manager handle. This object is used for global 87 * memory pool operations such as obtaining a new pool, deleting a pool, iterating and [all …]
|