| /OK3568_Linux_fs/kernel/tools/testing/selftests/drivers/net/mlxsw/ |
| H A D | sharedbuffer_configuration.py | 28 def __init__(self, pools): argument 30 for pool in pools: 187 pools = PoolList() 190 pools.append(Pool(pooldict)) 191 return pools 194 def do_check_pools(dlname, pools, vp): argument 195 for pool in pools: 219 def check_pools(dlname, pools): argument 221 record_vp = RecordValuePicker(pools) 224 do_check_pools(dlname, pools, RandomValuePicker(pools)) [all …]
|
| /OK3568_Linux_fs/kernel/drivers/staging/android/ion/heaps/ |
| H A D | ion_system_heap.c | 46 struct ion_page_pool *pools[NUM_ORDERS]; member 53 struct ion_page_pool *pool = heap->pools[order_to_index(order)]; in alloc_buffer_page() 70 pool = heap->pools[order_to_index(order)]; in free_buffer_page() 245 pool = sys_heap->pools[i]; in ion_system_heap_shrink() 273 total_pages += ion_page_pool_nr_pages(sys_heap->pools[i]); in ion_system_get_pool_size() 278 static void ion_system_heap_destroy_pools(struct ion_page_pool **pools) in ion_system_heap_destroy_pools() argument 283 if (pools[i]) in ion_system_heap_destroy_pools() 284 ion_page_pool_destroy(pools[i]); in ion_system_heap_destroy_pools() 287 static int ion_system_heap_create_pools(struct ion_page_pool **pools) in ion_system_heap_create_pools() argument 301 pools[i] = pool; in ion_system_heap_create_pools() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
| H A D | pool.c | 25 if (!xsk->pools) { in mlx5e_xsk_get_pools() 26 xsk->pools = kcalloc(MLX5E_MAX_NUM_CHANNELS, in mlx5e_xsk_get_pools() 27 sizeof(*xsk->pools), GFP_KERNEL); in mlx5e_xsk_get_pools() 28 if (unlikely(!xsk->pools)) in mlx5e_xsk_get_pools() 41 kfree(xsk->pools); in mlx5e_xsk_put_pools() 42 xsk->pools = NULL; in mlx5e_xsk_put_pools() 54 xsk->pools[ix] = pool; in mlx5e_xsk_add_pool() 60 xsk->pools[ix] = NULL; in mlx5e_xsk_remove_pool()
|
| H A D | pool.h | 12 if (!xsk || !xsk->pools) in mlx5e_xsk_get_pool() 18 return xsk->pools[ix]; in mlx5e_xsk_get_pool()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/ttm/ |
| H A D | ttm_page_alloc_dma.c | 97 struct list_head pools; /* The 'struct device->dma_pools link */ member 148 struct list_head pools; member 164 struct list_head pools; member 506 list_for_each_entry_reverse(p, &_manager->pools, pools) { in ttm_dma_free_pool() 513 list_del(&p->pools); in ttm_dma_free_pool() 518 list_for_each_entry_reverse(pool, &dev->dma_pools, pools) { in ttm_dma_free_pool() 529 list_del(&pool->pools); in ttm_dma_free_pool() 583 INIT_LIST_HEAD(&sec_pool->pools); in ttm_dma_pool_init() 588 INIT_LIST_HEAD(&pool->pools); in ttm_dma_pool_init() 618 list_add(&sec_pool->pools, &_manager->pools); in ttm_dma_pool_init() [all …]
|
| H A D | ttm_page_alloc.c | 114 struct ttm_page_pool pools[NUM_POOLS]; member 243 return &_manager->pools[pool_index]; in ttm_get_pool() 399 pool = &_manager->pools[(i + pool_offset)%NUM_POOLS]; in ttm_pool_shrink_scan() 422 pool = &_manager->pools[i]; in ttm_pool_shrink_count() 1022 ttm_page_pool_free(&_manager->pools[i], FREE_ALL_PAGES, true); in ttm_page_alloc_fini() 1181 p = &_manager->pools[i]; in ttm_page_alloc_debugfs()
|
| /OK3568_Linux_fs/kernel/drivers/staging/android/ion/ |
| H A D | ion_protected_heap.c | 171 struct ion_page_pool *pools[NUM_ORDERS]; member 195 pool = pheap->pools[order_to_index(order)]; in free_buffer_page() 209 pool = pheap->pools[order_to_index(order)]; in alloc_buffer_page() 376 pool = pheap->pools[i]; in ion_protected_heap_shrink() 416 pool = pheap->pools[i]; in ion_protected_heap_debug_show() 429 static void ion_protected_heap_destroy_pools(struct ion_page_pool **pools) in ion_protected_heap_destroy_pools() argument 434 if (pools[i]) in ion_protected_heap_destroy_pools() 435 protected_pool_destroy(pools[i]); in ion_protected_heap_destroy_pools() 438 static int ion_protected_heap_create_pools(struct ion_page_pool **pools) in ion_protected_heap_create_pools() argument 448 pools[i] = pool; in ion_protected_heap_create_pools() [all …]
|
| /OK3568_Linux_fs/kernel/arch/sparc/kernel/ |
| H A D | iommu-common.c | 82 spin_lock_init(&(iommu->pools[i].lock)); in iommu_tbl_pool_init() 83 iommu->pools[i].start = start; in iommu_tbl_pool_init() 84 iommu->pools[i].hint = start; in iommu_tbl_pool_init() 86 iommu->pools[i].end = start - 1; in iommu_tbl_pool_init() 131 pool = &(iommu->pools[pool_nr]); in iommu_tbl_range_alloc() 161 pool = &(iommu->pools[0]); in iommu_tbl_range_alloc() 193 pool = &(iommu->pools[pool_nr]); in iommu_tbl_range_alloc() 237 p = &tbl->pools[pool_nr]; in get_pool()
|
| /OK3568_Linux_fs/u-boot/include/fsl-mc/ |
| H A D | fsl_dpni.h | 144 MC_CMD_OP(cmd, 0, 8, 1, int, cfg->pools[0].backup_pool); \ 145 MC_CMD_OP(cmd, 0, 9, 1, int, cfg->pools[1].backup_pool); \ 146 MC_CMD_OP(cmd, 0, 10, 1, int, cfg->pools[2].backup_pool); \ 147 MC_CMD_OP(cmd, 0, 11, 1, int, cfg->pools[3].backup_pool); \ 148 MC_CMD_OP(cmd, 0, 12, 1, int, cfg->pools[4].backup_pool); \ 149 MC_CMD_OP(cmd, 0, 13, 1, int, cfg->pools[5].backup_pool); \ 150 MC_CMD_OP(cmd, 0, 14, 1, int, cfg->pools[6].backup_pool); \ 151 MC_CMD_OP(cmd, 0, 15, 1, int, cfg->pools[7].backup_pool); \ 152 MC_CMD_OP(cmd, 0, 32, 32, int, cfg->pools[0].dpbp_id); \ 153 MC_CMD_OP(cmd, 4, 32, 16, uint16_t, cfg->pools[0].buffer_size);\ [all …]
|
| /OK3568_Linux_fs/kernel/drivers/dma-buf/heaps/ |
| H A D | rk_system_heap.c | 48 struct dmabuf_page_pool **pools; member 74 struct dmabuf_page_pool *pools[NUM_ORDERS]; variable 459 dmabuf_page_pool_free(buffer->pools[j], page); in system_heap_buf_free() 544 buffer->pools = strstr(dma_heap_get_name(heap), "dma32") ? dma32_pools : pools; in system_heap_do_allocate() 558 page = system_heap_alloc_largest_available(heap, buffer->pools, in system_heap_do_allocate() 664 pool = strstr(dma_heap_get_name(heap), "dma32") ? dma32_pools : pools; in system_get_pool_size() 758 pools[i] = dmabuf_page_pool_create(order_flags[i], orders[i]); in system_heap_create() 760 if (!pools[i]) { in system_heap_create() 765 dmabuf_page_pool_destroy(pools[j]); in system_heap_create() 836 dmabuf_page_pool_destroy(pools[i]); in system_heap_create()
|
| H A D | system_heap.c | 65 struct dmabuf_page_pool *pools[NUM_ORDERS]; variable 350 dmabuf_page_pool_free(pools[j], page); in system_heap_buf_free() 389 page = dmabuf_page_pool_alloc(pools[i]); in alloc_largest_available() 509 pool = pools; in system_get_pool_size() 578 pools[i] = dmabuf_page_pool_create(order_flags[i], orders[i]); in system_heap_create() 580 if (!pools[i]) { in system_heap_create() 585 dmabuf_page_pool_destroy(pools[j]); in system_heap_create()
|
| /OK3568_Linux_fs/kernel/mm/ |
| H A D | dmapool.c | 50 struct list_head pools; member 81 list_for_each_entry(pool, &dev->dma_pools, pools) { in show_pools() 105 static DEVICE_ATTR(pools, 0444, show_pools, NULL); 169 INIT_LIST_HEAD(&retval->pools); in dma_pool_create() 183 list_add(&retval->pools, &dev->dma_pools); in dma_pool_create() 191 list_del(&retval->pools); in dma_pool_create() 277 list_del(&pool->pools); in dma_pool_destroy()
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/chelsio/libcxgb/ |
| H A D | libcxgb_ppm.c | 348 struct cxgbi_ppm_pool *pools; in ppm_alloc_cpu_pool() local 350 unsigned int max = (PCPU_MIN_UNIT_SIZE - sizeof(*pools)) << 3; in ppm_alloc_cpu_pool() 367 alloc_sz = sizeof(*pools) + sizeof(unsigned long) * bmap; in ppm_alloc_cpu_pool() 368 pools = __alloc_percpu(alloc_sz, __alignof__(struct cxgbi_ppm_pool)); in ppm_alloc_cpu_pool() 370 if (!pools) in ppm_alloc_cpu_pool() 374 struct cxgbi_ppm_pool *ppool = per_cpu_ptr(pools, cpu); in ppm_alloc_cpu_pool() 384 return pools; in ppm_alloc_cpu_pool()
|
| /OK3568_Linux_fs/kernel/drivers/atm/ |
| H A D | zatm.c | 1038 unsigned long pools; in zatm_int() local 1041 pools = zin(RQA); in zatm_int() 1042 EVENT("RQA (0x%08x)\n",pools,0); in zatm_int() 1043 for (i = 0; pools; i++) { in zatm_int() 1044 if (pools & 1) { in zatm_int() 1048 pools >>= 1; in zatm_int() 1052 unsigned long pools; in zatm_int() local 1054 pools = zin(RQU); in zatm_int() 1056 dev->number,pools); in zatm_int() 1058 for (i = 0; pools; i++) { in zatm_int() [all …]
|
| /OK3568_Linux_fs/kernel/Documentation/core-api/ |
| H A D | workqueue.rst | 60 * Use per-CPU unified worker pools shared by all wq to provide 83 called worker-pools. 87 which manages worker-pools and processes the queued work items. 89 There are two worker-pools, one for normal work items and the other 91 worker-pools to serve work items queued on unbound workqueues - the 92 number of these backing pools is dynamic. 132 For unbound workqueues, the number of backing pools is dynamic. 135 backing worker pools matching the attributes. The responsibility of 169 worker-pools which host workers which are not bound to any 172 worker-pools try to start execution of work items as soon as [all …]
|
| H A D | mm-api.rst | 79 Memory pools 85 DMA pools
|
| /OK3568_Linux_fs/kernel/Documentation/ABI/testing/ |
| H A D | sysfs-kernel-ion | 8 internal state of ION memory heaps and pools. 27 memory in Kb is allocated to ION pools.
|
| H A D | sysfs-device-mali | 128 memory pools that the driver can contain. Large pages are of 130 pools and can be used to modify each individual pools as well. 138 pool size for all available pools and we can modify individual 144 for memory pools that the driver can contain. Here small pages 146 available pools and allows us to set max size of 147 individual pools. 155 pool size for all available pools and we can modify individual 167 for all memory pools that the driver can contain. 173 for all the memory pools that the driver can contain.
|
| H A D | sysfs-kernel-dmaheap | 7 memory in Kb is allocated to DMA-BUF heap pools.
|
| /OK3568_Linux_fs/kernel/arch/powerpc/kernel/ |
| H A D | iommu.c | 202 pool = &(tbl->pools[pool_nr]); in iommu_range_alloc() 230 pool = &(tbl->pools[0]); in iommu_range_alloc() 252 pool = &tbl->pools[pool_nr]; in iommu_range_alloc() 370 p = &tbl->pools[pool_nr]; in get_pool() 707 p = &tbl->pools[i]; in iommu_init_table() 1053 spin_lock_nest_lock(&tbl->pools[i].lock, &tbl->large_pool.lock); in iommu_take_ownership() 1068 spin_unlock(&tbl->pools[i].lock); in iommu_take_ownership() 1081 spin_lock_nest_lock(&tbl->pools[i].lock, &tbl->large_pool.lock); in iommu_release_ownership() 1089 spin_unlock(&tbl->pools[i].lock); in iommu_release_ownership()
|
| /OK3568_Linux_fs/kernel/drivers/soc/ti/ |
| H A D | knav_qmss.h | 203 struct list_head pools; member 304 struct list_head pools; member 363 list_for_each_entry(pool, &kdev->pools, list)
|
| /OK3568_Linux_fs/kernel/Documentation/devicetree/bindings/soc/ti/ |
| H A D | keystone-navigator-qmss.txt | 6 processors(PDSP), linking RAM, descriptor pools and infrastructure 49 - queue-pools : child node classifying the queue ranges into pools. 50 Queue ranges are grouped into 3 type of pools: 151 queue-pools {
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/intel/ice/ |
| H A D | ice.h | 518 struct xsk_buff_pool **pools = ring->vsi->xsk_pools; in ice_xsk_pool() local 524 if (qid >= ring->vsi->num_xsk_pools || !pools || !pools[qid] || in ice_xsk_pool() 528 return pools[qid]; in ice_xsk_pool()
|
| /OK3568_Linux_fs/prebuilts/gcc/linux-x86/arm/gcc-arm-10.3-2021.07-x86_64-arm-none-linux-gnueabihf/arm-none-linux-gnueabihf/include/c++/10.3.1/ |
| H A D | memory_resource | 381 * unused, but means pools will need to replenished more frequently. 385 /* @brief Largest block size (in bytes) that should be served from pools. 388 * not from one of the pools managed by the pool resource. 436 /// A thread-safe memory resource that manages pools of fixed-size blocks. 486 // Thread-specific pools (only public for access by implementation details) 496 // Linked list of thread-specific pools. All threads share _M_tpools[0]. 502 /// A non-thread-safe memory resource that manages pools of fixed-size blocks.
|
| /OK3568_Linux_fs/prebuilts/gcc/linux-x86/aarch64/gcc-arm-10.3-2021.07-x86_64-aarch64-none-linux-gnu/aarch64-none-linux-gnu/include/c++/10.3.1/ |
| H A D | memory_resource | 381 * unused, but means pools will need to replenished more frequently. 385 /* @brief Largest block size (in bytes) that should be served from pools. 388 * not from one of the pools managed by the pool resource. 436 /// A thread-safe memory resource that manages pools of fixed-size blocks. 486 // Thread-specific pools (only public for access by implementation details) 496 // Linked list of thread-specific pools. All threads share _M_tpools[0]. 502 /// A non-thread-safe memory resource that manages pools of fixed-size blocks.
|