| /OK3568_Linux_fs/yocto/poky/bitbake/lib/bb/tests/ |
| H A D | runqueue.py | 45 tasks = [] 49 tasks = [line.rstrip() for line in f] 52 return tasks 58 tasks = self.run_bitbakecmd(cmd, tempdir, sstatevalid) 60 self.assertEqual(set(tasks), set(expected)) 68 tasks = self.run_bitbakecmd(cmd, tempdir, sstatevalid) 72 self.assertEqual(set(tasks), set(expected)) 80 tasks = self.run_bitbakecmd(cmd, tempdir, sstatevalid) 83 self.assertEqual(set(tasks), set(expected)) 91 tasks = self.run_bitbakecmd(cmd, tempdir, sstatevalid) [all …]
|
| /OK3568_Linux_fs/yocto/poky/documentation/ref-manual/ |
| H A D | tasks.rst | 8 tasks to complete configuring, compiling, and packaging software. This 9 chapter provides a reference of the tasks defined in the OpenEmbedded 15 The following sections describe normal tasks associated with building a 16 recipe. For more information on tasks and dependencies, see the 17 ":ref:`bitbake:bitbake-user-manual/bitbake-user-manual-metadata:tasks`" and 21 .. _ref-tasks-build: 27 tasks required to build a recipe. 29 .. _ref-tasks-compile: 41 .. _ref-tasks-compile_ptest_base: 48 .. _ref-tasks-configure: [all …]
|
| /OK3568_Linux_fs/buildroot/package/directfb/ |
| H A D | 0001-fix-missing-davinci-voodoo-header.patch | 76 + c64xTask *tasks; 94 +DFBResult davinci_c64x_tasks_init ( DavinciC64xTasks *tasks, 97 +DFBResult davinci_c64x_tasks_destroy( DavinciC64xTasks *tasks ); 102 + DavinciC64xTasks *tasks, 191 +davinci_c64x_load_block__L( DavinciC64xTasks *tasks, 196 + c64xTask *task = &tasks->tasks[tasks->num_tasks]; 198 + D_ASSERT( tasks->num_tasks < tasks->max_tasks ); 206 + tasks->num_tasks++; 272 +davinci_c64x_put_idct_uyvy_16x16__L( DavinciC64xTasks *tasks, 277 + c64xTask *task = &tasks->tasks[tasks->num_tasks]; [all …]
|
| /OK3568_Linux_fs/kernel/Documentation/admin-guide/cgroup-v1/ |
| H A D | freezer-subsystem.rst | 6 and stop sets of tasks in order to schedule the resources of a machine 9 whole. The cgroup freezer uses cgroups to describe the set of tasks to 11 a means to start and stop the tasks composing the job. 14 of tasks. The freezer allows the checkpoint code to obtain a consistent 15 image of the tasks by attempting to force the tasks in a cgroup into a 16 quiescent state. Once the tasks are quiescent another task can 18 quiesced tasks. Checkpointed tasks can be restarted later should a 19 recoverable error occur. This also allows the checkpointed tasks to be 21 to another node and restarting the tasks there. 24 and resuming tasks in userspace. Both of these signals are observable [all …]
|
| H A D | cpuacct.rst | 5 The CPU accounting controller is used to group tasks using cgroups and 6 account the CPU usage of these groups of tasks. 9 group accumulates the CPU usage of all of its child groups and the tasks 17 visible at /sys/fs/cgroup. At bootup, this group includes all the tasks in 18 the system. /sys/fs/cgroup/tasks lists the tasks in this cgroup. 20 by this group which is essentially the CPU time obtained by all the tasks 27 # echo $$ > g1/tasks 38 user: Time spent by tasks of the cgroup in user mode. 39 system: Time spent by tasks of the cgroup in kernel mode.
|
| H A D | cgroups.rst | 45 tasks, and all their future children, into hierarchical groups with 50 A *cgroup* associates a set of tasks with a set of parameters for one 54 facilities provided by cgroups to treat groups of tasks in 67 cgroups. Each hierarchy is a partition of all tasks in the system. 81 tasks in each cgroup. 100 the division of tasks into cgroups is distinctly different for 102 hierarchy to be a natural division of tasks, without having to handle 103 complex combinations of tasks that would be present if several 114 tasks etc. The resource planning for this server could be along the 123 In addition (system tasks) are attached to topcpuset (so [all …]
|
| /OK3568_Linux_fs/yocto/poky/bitbake/lib/toaster/tests/builds/ |
| H A D | test_core_image_min.py | 61 tasks = Task.objects.filter( 68 for task in tasks: 82 tasks = Task.objects.filter(outcome=2).values('id', 'sstate_result') 84 for task in tasks: 95 tasks = Task.objects.filter( 102 for task in tasks: 116 tasks = Task.objects.filter( 120 for task in tasks: 131 tasks = Task.objects.filter( 135 for task in tasks: [all …]
|
| /OK3568_Linux_fs/yocto/poky/scripts/lib/ |
| H A D | buildstats.py | 126 def __init__(self, tasks=None): argument 127 self._tasks = tasks or [] 159 self.tasks = {} 166 if set(self.tasks.keys()) != set(bsrecipe.tasks.keys()): 170 for taskname, taskdata in bsrecipe.tasks.items(): 171 if not isinstance(self.tasks[taskname], BSTaskAggregate): 172 self.tasks[taskname] = BSTaskAggregate([self.tasks[taskname]]) 173 self.tasks[taskname].append(taskdata) 188 num += len(recipe.tasks) 202 bsrecipe.tasks[task] = BSTask(data) [all …]
|
| /OK3568_Linux_fs/external/mpp/mpp/base/ |
| H A D | mpp_task_impl.cpp | 57 MppTaskImpl *tasks; member 471 mpp_assert(impl->tasks == NULL); in mpp_task_queue_setup() 473 MppTaskImpl *tasks = mpp_calloc(MppTaskImpl, task_count); in mpp_task_queue_setup() local 474 if (NULL == tasks) { in mpp_task_queue_setup() 479 impl->tasks = tasks; in mpp_task_queue_setup() 485 setup_mpp_task_name(&tasks[i]); in mpp_task_queue_setup() 486 INIT_LIST_HEAD(&tasks[i].list); in mpp_task_queue_setup() 487 tasks[i].index = i; in mpp_task_queue_setup() 488 tasks[i].queue = queue; in mpp_task_queue_setup() 489 tasks[i].status = MPP_INPUT_PORT; in mpp_task_queue_setup() [all …]
|
| /OK3568_Linux_fs/kernel/Documentation/scheduler/ |
| H A D | sched-design-CFS.rst | 19 1/nr_running speed. For example: if there are 2 tasks running, then it runs 26 is its actual runtime normalized to the total number of running tasks. 37 [ small detail: on "ideal" hardware, at any time all tasks would have the same 38 p->se.vruntime value --- i.e., tasks would execute simultaneously and no task 44 up CPU time between runnable tasks as close to "ideal multitasking hardware" as 62 increasing value tracking the smallest vruntime among all tasks in the 67 The total number of running tasks in the runqueue is accounted through the 68 rq->cfs.load value, which is the sum of the weights of the tasks queued on the 71 CFS maintains a time-ordered rbtree, where all runnable tasks are sorted by the 73 As the system progresses forwards, the executed tasks are put into the tree [all …]
|
| H A D | sched-deadline.rst | 43 that makes it possible to isolate the behavior of tasks between each other. 53 "deadline", to schedule tasks. A SCHED_DEADLINE task should receive 65 Summing up, the CBS[2,3] algorithm assigns scheduling deadlines to tasks so 67 interference between different tasks (bandwidth isolation), while the EDF[1] 69 to be executed next. Thanks to this feature, tasks that do not strictly comply 74 tasks in the following way: 128 Bandwidth reclaiming for deadline tasks is based on the GRUB (Greedy 132 The following diagram illustrates the state names for tasks handled by GRUB:: 201 tasks in active state (i.e., ActiveContending or ActiveNonContending); 203 - Total bandwidth (this_bw): this is the sum of all tasks "belonging" to the [all …]
|
| H A D | sched-rt-group.rst | 14 2.3 Basis for grouping tasks 44 multiple groups of realtime tasks, each group must be assigned a fixed portion 57 tasks (SCHED_OTHER). Any allocated run time not used will also be picked up by 72 The remaining CPU time will be used for user input and other tasks. Because 73 realtime tasks have explicitly allocated the CPU time they need to perform 74 their tasks, buffer underruns in the graphics or audio can be eliminated. 110 SCHED_OTHER (non-RT tasks). These defaults were chosen so that a run-away 111 realtime tasks will not lock up the machine but leave a little time to recover 120 bandwidth to the group before it will accept realtime tasks. Therefore you will 121 not be able to run realtime tasks as any user other than root until you have [all …]
|
| /OK3568_Linux_fs/kernel/Documentation/power/ |
| H A D | freezing-of-tasks.rst | 2 Freezing of tasks 7 I. What is the freezing of tasks? 10 The freezing of tasks is a mechanism by which user space processes and some 18 and PF_FREEZER_SKIP (the last one is auxiliary). The tasks that have 30 All freezable tasks must react to that by calling try_to_freeze(), which 62 initiated a freezing operation, the freezing of tasks will fail and the entire 69 order to clear the PF_FROZEN flag for each frozen task. Then, the tasks that 73 Rationale behind the functions dealing with freezing and thawing of tasks 77 - freezes only userspace tasks 80 - freezes all tasks (including kernel threads) because we can't freeze [all …]
|
| /OK3568_Linux_fs/yocto/meta-openembedded/meta-python/recipes-extended/python-blivet/python3-blivet/ |
| H A D | 0009-invoking-fsck-with-infinite-timeout.patch | 13 blivet/tasks/fsck.py | 2 +- 16 diff --git a/blivet/tasks/fsck.py b/blivet/tasks/fsck.py 18 --- a/blivet/tasks/fsck.py 19 +++ b/blivet/tasks/fsck.py
|
| H A D | 0010-invoking-mkfs-with-infinite-timeout.patch | 17 blivet/tasks/fsmkfs.py | 2 +- 20 diff --git a/blivet/tasks/fsmkfs.py b/blivet/tasks/fsmkfs.py 22 --- a/blivet/tasks/fsmkfs.py 23 +++ b/blivet/tasks/fsmkfs.py
|
| /OK3568_Linux_fs/kernel/samples/bpf/ |
| H A D | tracex2_user.c | 86 static struct task tasks[1024]; in print_hist() local 94 if (memcmp(&tasks[i], &next_key, SIZE) == 0) in print_hist() 97 memcpy(&tasks[task_cnt++], &next_key, SIZE); in print_hist() 103 (__u32) tasks[i].pid_tgid, in print_hist() 104 tasks[i].comm, in print_hist() 105 (__u32) tasks[i].uid_gid); in print_hist() 106 print_hist_for_pid(fd, &tasks[i]); in print_hist()
|
| H A D | map_perf_test_user.c | 95 static int pre_test_lru_hash_lookup(int tasks) in pre_test_lru_hash_lookup() argument 293 typedef int (*pre_test_func)(int tasks); 313 static int pre_test(int tasks) in pre_test() argument 319 int ret = pre_test_funcs[i](tasks); in pre_test() 344 static void run_perf_test(int tasks) in run_perf_test() argument 346 pid_t pid[tasks]; in run_perf_test() 349 assert(!pre_test(tasks)); in run_perf_test() 351 for (i = 0; i < tasks; i++) { in run_perf_test() 361 for (i = 0; i < tasks; i++) { in run_perf_test()
|
| /OK3568_Linux_fs/external/mpp/mpp/hal/ |
| H A D | mpp_enc_hal.cpp | 53 HalTaskGroup tasks; member 83 ret = hal_task_group_init(&p->tasks, TASK_BUTT, cfg->task_cnt, in mpp_enc_hal_init() 90 cfg->tasks = p->tasks; in mpp_enc_hal_init() 113 if (p->tasks) in mpp_enc_hal_deinit() 114 hal_task_group_deinit(p->tasks); in mpp_enc_hal_deinit()
|
| /OK3568_Linux_fs/kernel/tools/perf/scripts/python/ |
| H A D | sched-migration.py | 100 def __init__(self, tasks = [0], event = RunqueueEventUnknown()): argument 101 self.tasks = tuple(tasks) 107 if taskState(prev_state) == "R" and next in self.tasks \ 108 and prev in self.tasks: 114 next_tasks = list(self.tasks[:]) 115 if prev in self.tasks: 127 if old not in self.tasks: 129 next_tasks = [task for task in self.tasks if task != old] 134 if new in self.tasks: 137 next_tasks = self.tasks[:] + tuple([new]) [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/ |
| H A D | drm_flip_work.c | 117 struct list_head tasks; in flip_worker() local 123 INIT_LIST_HEAD(&tasks); in flip_worker() 125 list_splice_tail(&work->commited, &tasks); in flip_worker() 129 if (list_empty(&tasks)) in flip_worker() 132 list_for_each_entry_safe(task, tmp, &tasks, node) { in flip_worker()
|
| /OK3568_Linux_fs/kernel/kernel/sched/ |
| H A D | psi.c | 226 static bool test_state(unsigned int *tasks, enum psi_states state) in test_state() argument 230 return tasks[NR_IOWAIT]; in test_state() 232 return tasks[NR_IOWAIT] && !tasks[NR_RUNNING]; in test_state() 234 return tasks[NR_MEMSTALL]; in test_state() 236 return tasks[NR_MEMSTALL] && !tasks[NR_RUNNING]; in test_state() 238 return tasks[NR_RUNNING] > tasks[NR_ONCPU]; in test_state() 240 return tasks[NR_IOWAIT] || tasks[NR_MEMSTALL] || in test_state() 241 tasks[NR_RUNNING]; in test_state() 769 if (groupc->tasks[t]) { in psi_group_change() 770 groupc->tasks[t]--; in psi_group_change() [all …]
|
| /OK3568_Linux_fs/yocto/poky/meta/lib/oeqa/selftest/cases/ |
| H A D | bbtests.py | 242 tasks = re.findall(r'task\s+(do_\S+):', ret.output) 244 for task in tasks: 246 'Executed tasks were: %s' % (task, str(tasks))) 255 tasks = re.findall(r'task\s+(do_\S+):', ret.output) 257 for task in tasks: 259 'Executed tasks were: %s' % (task, str(tasks))) 263 tasks = re.findall(r'task\s+(do_\S+):', ret.output) 265 self.assertFalse(tasks, 'Tasks %s ran when they should not have' % (str(tasks))) 272 tasks = re.findall(r'task\s+(do_\S+):', ret.output) 274 for task in tasks: [all …]
|
| /OK3568_Linux_fs/yocto/poky/scripts/ |
| H A D | oe-check-sstate | 26 def translate_virtualfns(tasks): argument 34 for task in tasks: 57 tasks = [] 61 tasks.append(res.group(1)) 62 outtasks = translate_virtualfns(tasks)
|
| /OK3568_Linux_fs/kernel/Documentation/admin-guide/kdump/ |
| H A D | gdbmacros.txt | 17 set $tasks_off=((size_t)&((struct task_struct *)0)->tasks) 20 set $next_t=(((char *)($init_t->tasks).next) - $tasks_off) 51 set $next_t=(char *)($next_t->tasks.next) - $tasks_off 83 set $tasks_off=((size_t)&((struct task_struct *)0)->tasks) 86 set $next_t=(((char *)($init_t->tasks).next) - $tasks_off) 97 set $next_t=(char *)($next_t->tasks.next) - $tasks_off 106 set $tasks_off=((size_t)&((struct task_struct *)0)->tasks) 109 set $next_t=(((char *)($init_t->tasks).next) - $tasks_off) 127 set $next_t=(char *)($next_t->tasks.next) - $tasks_off 139 set $tasks_off=((size_t)&((struct task_struct *)0)->tasks) [all …]
|
| /OK3568_Linux_fs/kernel/tools/perf/Documentation/ |
| H A D | perf-timechart.txt | 48 --tasks-only:: 60 Print task info for at least given number of tasks. 65 Highlight tasks (using different color) that run more than given 66 duration or tasks with given name. If number is given it's interpreted 89 --tasks-only:: 90 Record only tasks-related events 114 then generate timechart and highlight 'gcc' tasks:
|