| /OK3568_Linux_fs/kernel/Documentation/core-api/ |
| H A D | workqueue.rst | 34 number of workers as the number of CPUs. The kernel grew a lot of MT 118 number of the currently runnable workers. Generally, work items are 122 workers on the CPU, the worker-pool doesn't start execution of a new 125 are pending work items. This allows using a minimal number of workers 128 Keeping idle workers around doesn't cost other than the memory space 140 Forward progress guarantee relies on that workers can be created when 142 through the use of rescue workers. All work items which might be used 169 worker-pools which host workers which are not bound to any 178 of mostly unused workers across different CPUs as the issuer 200 each other. Each maintains its separate pool of workers and [all …]
|
| /OK3568_Linux_fs/yocto/poky/meta/lib/oe/ |
| H A D | utils.py | 520 self.workers = [] 524 self.workers.append(worker) 527 for worker in self.workers: 537 for worker in self.workers:
|
| /OK3568_Linux_fs/yocto/poky/meta-selftest/recipes-devtools/python/ |
| H A D | python-async-test.inc | 1 SUMMARY = "Python framework to process interdependent tasks in a pool of workers"
|
| /OK3568_Linux_fs/yocto/poky/meta/recipes-devtools/python/ |
| H A D | python-async.inc | 1 SUMMARY = "Python framework to process interdependent tasks in a pool of workers"
|
| /OK3568_Linux_fs/yocto/poky/documentation/ref-manual/ |
| H A D | release-process.rst | 180 takes the Autobuilder workers several hours. 184 The Autobuilder workers are non-homogeneous, which means regular
|
| /OK3568_Linux_fs/kernel/drivers/md/ |
| H A D | raid5.h | 516 struct r5worker *workers; member
|
| H A D | raid5.c | 197 group->workers[0].working = true; in raid5_wakeup_stripe_thread() 199 queue_work_on(sh->cpu, raid5_wq, &group->workers[0].work); in raid5_wakeup_stripe_thread() 204 if (group->workers[i].working == false) { in raid5_wakeup_stripe_thread() 205 group->workers[i].working = true; in raid5_wakeup_stripe_thread() 207 &group->workers[i].work); in raid5_wakeup_stripe_thread() 6933 kfree(old_groups[0].workers); in raid5_store_group_thread_cnt() 6970 struct r5worker *workers; in alloc_thread_groups() local 6979 workers = kcalloc(size, *group_cnt, GFP_NOIO); in alloc_thread_groups() 6982 if (!*worker_groups || !workers) { in alloc_thread_groups() 6983 kfree(workers); in alloc_thread_groups() [all …]
|
| /OK3568_Linux_fs/kernel/net/l2tp/ |
| H A D | Kconfig | 23 with home workers to connect to their offices.
|
| /OK3568_Linux_fs/kernel/kernel/ |
| H A D | workqueue.c | 176 struct list_head workers; /* A: attached workers */ member 417 list_for_each_entry((worker), &(pool)->workers, node) \ 1887 list_add_tail(&worker->node, &pool->workers); in worker_attach_to_pool() 1911 if (list_empty(&pool->workers)) in worker_detach_from_pool() 3467 INIT_LIST_HEAD(&pool->workers); in init_worker_pool() 3602 if (!list_empty(&pool->workers)) in put_unbound_pool()
|
| /OK3568_Linux_fs/yocto/poky/meta/classes/ |
| H A D | uninative.bbclass | 125 This event handler is called in the workers and is responsible for setting
|
| /OK3568_Linux_fs/kernel/drivers/block/mtip32xx/ |
| H A D | mtip32xx.c | 745 int do_irq_enable = 1, i, workers; in mtip_handle_irq() local 766 for (i = 0, workers = 0; i < MTIP_MAX_SLOT_GROUPS; in mtip_handle_irq() 771 workers++; in mtip_handle_irq() 774 atomic_set(&dd->irq_workers_active, workers); in mtip_handle_irq() 775 if (workers) { in mtip_handle_irq()
|
| /OK3568_Linux_fs/yocto/poky/documentation/test-manual/ |
| H A D | intro.rst | 82 topology that includes a controller and a cluster of workers: 505 workers, consider the following:
|
| /OK3568_Linux_fs/kernel/Documentation/dev-tools/ |
| H A D | kcov.rst | 220 some kernel interface (e.g. vhost workers); as well as from soft
|
| /OK3568_Linux_fs/kernel/fs/btrfs/ |
| H A D | disk-io.c | 774 btrfs_queue_work(fs_info->workers, &async->work); in btrfs_wq_submit_bio() 1995 btrfs_destroy_workqueue(fs_info->workers); in btrfs_stop_all_workers() 2166 fs_info->workers = in btrfs_init_workqueues() 2218 if (!(fs_info->workers && fs_info->delalloc_workers && in btrfs_init_workqueues()
|
| H A D | super.c | 1806 btrfs_workqueue_set_max(fs_info->workers, new_pool_size); in btrfs_resize_thread_pool()
|
| H A D | ctree.h | 768 struct btrfs_workqueue *workers; member
|
| /OK3568_Linux_fs/kernel/Documentation/admin-guide/ |
| H A D | kernel-per-CPU-kthreads.rst | 262 d. As of v3.18, Christoph Lameter's on-demand vmstat workers
|
| /OK3568_Linux_fs/yocto/poky/bitbake/lib/bb/ |
| H A D | runqueue.py | 3106 for workers, name in [(self.rq.worker, "Worker"), (self.rq.fakeworker, "Fakeroot")]: 3107 for worker in workers.values():
|
| /OK3568_Linux_fs/yocto/poky/bitbake/doc/bitbake-user-manual/ |
| H A D | bitbake-user-manual-metadata.rst | 1614 events when each of the workers parse the base configuration or if
|
| /OK3568_Linux_fs/prebuilts/gcc/linux-x86/aarch64/gcc-arm-10.3-2021.07-x86_64-aarch64-none-linux-gnu/share/info/ |
| H A D | gdb.info | 25625 matchers and xmethod workers (*note Xmethods In Python::). Consider the 25679 workers = [] 25684 workers.append(worker) 25686 return workers 25736 workers into 'GDBN', invoking the method 'geta' or using the operator 25766 xmethod workers and xmethod matchers is as follows:
|
| /OK3568_Linux_fs/prebuilts/gcc/linux-x86/arm/gcc-arm-10.3-2021.07-x86_64-arm-none-linux-gnueabihf/share/info/ |
| H A D | gdb.info | 25625 matchers and xmethod workers (*note Xmethods In Python::). Consider the 25679 workers = [] 25684 workers.append(worker) 25686 return workers 25736 workers into 'GDBN', invoking the method 'geta' or using the operator 25766 xmethod workers and xmethod matchers is as follows:
|
| /OK3568_Linux_fs/recovery/ |
| HD | rootfs.cpio.gz | 107070100A8AEB0000041ED0000000000000000000000116841B48100000000000000FD0000000200000000000000000000000200000000.07070100A8B07C000081A40000000000000000000000016841263500000000000000FD0000000200000000000000000000000900000000.gitkeep07070100A8B07D000081A40000000000000000000000016841263500000000000000FD0000000200000000000000000000000B00000000.skip_fsck07070100A8AEB10000A1FF0000000000000000000000016841B0B600000007000000FD0000000200000000000000000000000400000000binusr/bin07070100A8B07E000081A400000000000000000000000168412635000001CC000000FD0000000200000000000000000000001100000000busybox. ... |