Lines Matching refs:topo
232 topo.nodes_per_pkg * topo.cores_per_node * topo.threads_per_core) + \
233 ((node_no) * topo.cores_per_node * topo.threads_per_core) + \
234 ((core_no) * topo.threads_per_core) + \
239 ((pkg_no) * topo.nodes_per_pkg * topo.cores_per_node) + \
240 ((node_no) * topo.cores_per_node) + \
411 } topo; variable
438 for (pkg_no = 0; pkg_no < topo.num_packages; ++pkg_no) { in for_all_cpus()
439 for (node_no = 0; node_no < topo.nodes_per_pkg; node_no++) { in for_all_cpus()
440 for (core_no = 0; core_no < topo.cores_per_node; ++core_no) { in for_all_cpus()
442 topo.threads_per_core; ++thread_no) { in for_all_cpus()
1711 average.threads.tsc /= topo.num_cpus; in compute_average()
1712 average.threads.aperf /= topo.num_cpus; in compute_average()
1713 average.threads.mperf /= topo.num_cpus; in compute_average()
1714 average.threads.c1 /= topo.num_cpus; in compute_average()
1719 average.cores.c3 /= topo.num_cores; in compute_average()
1720 average.cores.c6 /= topo.num_cores; in compute_average()
1721 average.cores.c7 /= topo.num_cores; in compute_average()
1722 average.cores.mc6_us /= topo.num_cores; in compute_average()
1725 average.packages.pkg_wtd_core_c0 /= topo.num_packages; in compute_average()
1727 average.packages.pkg_any_core_c0 /= topo.num_packages; in compute_average()
1729 average.packages.pkg_any_gfxe_c0 /= topo.num_packages; in compute_average()
1731 average.packages.pkg_both_core_gfxe_c0 /= topo.num_packages; in compute_average()
1733 average.packages.pc2 /= topo.num_packages; in compute_average()
1735 average.packages.pc3 /= topo.num_packages; in compute_average()
1737 average.packages.pc6 /= topo.num_packages; in compute_average()
1739 average.packages.pc7 /= topo.num_packages; in compute_average()
1741 average.packages.pc8 /= topo.num_packages; in compute_average()
1742 average.packages.pc9 /= topo.num_packages; in compute_average()
1743 average.packages.pc10 /= topo.num_packages; in compute_average()
1753 average.threads.counter[i] /= topo.num_cpus; in compute_average()
1762 average.cores.counter[i] /= topo.num_cores; in compute_average()
1771 average.packages.counter[i] /= topo.num_packages; in compute_average()
2588 for (i = 0; i < topo.max_cpu_num + 1; ++i) { in free_fd_percpu()
2633 for (i = 0; i <= topo.max_cpu_num; ++i) { in free_all_buffers()
2695 for (cpu = 0; cpu <= topo.max_cpu_num; ++cpu) in set_node_data()
2699 for (pkg = 0; pkg < topo.num_packages; pkg++) { in set_node_data()
2701 for (cpu = 0; cpu <= topo.max_cpu_num; ++cpu) { in set_node_data()
2714 for (cpux = cpu; cpux <= topo.max_cpu_num; cpux++) { in set_node_data()
2722 if (lnode > topo.nodes_per_pkg) in set_node_data()
2723 topo.nodes_per_pkg = lnode; in set_node_data()
2725 if (cpu_count >= topo.max_cpu_num) in set_node_data()
2737 for (i = 0; i <= topo.max_cpu_num; i++) { in get_physical_node_id()
2756 int offset = topo.max_cpu_num + 1; in get_thread_siblings()
2760 thiscpu->put_ids = CPU_ALLOC((topo.max_cpu_num + 1)); in get_thread_siblings()
2766 size = CPU_ALLOC_SIZE((topo.max_cpu_num + 1)); in get_thread_siblings()
2814 for (pkg_no = 0; pkg_no < topo.num_packages; ++pkg_no) { in for_all_cpus_2()
2815 for (node_no = 0; node_no < topo.nodes_per_pkg; ++node_no) { in for_all_cpus_2()
2816 for (core_no = 0; core_no < topo.cores_per_node; in for_all_cpus_2()
2819 topo.threads_per_core; ++thread_no) { in for_all_cpus_2()
2889 fprintf(outf, "turbostat: re-initialized with num_cpus %d\n", topo.num_cpus); in re_initialize()
2907 topo.max_cpu_num = 0; in set_max_cpu_num()
2909 topo.max_cpu_num += BITMASK_SIZE; in set_max_cpu_num()
2911 topo.max_cpu_num--; /* 0 based */ in set_max_cpu_num()
2920 topo.num_cpus++; in count_cpus()
2953 for (column = 0; column < topo.num_cpus; ++column) { in snapshot_proc_interrupts()
2960 if (cpu_number > topo.max_cpu_num) { in snapshot_proc_interrupts()
2961 warn("/proc/interrupts: cpu%d: > %d", cpu_number, topo.max_cpu_num); in snapshot_proc_interrupts()
2979 for (column = 0; column < topo.num_cpus; ++column) { in snapshot_proc_interrupts()
3286 per_cpu_msr_sum = calloc(topo.max_cpu_num + 1, sizeof(struct msr_sum_array)); in msr_sum_record()
5364 topo.num_cpus = 0; in topology_probe()
5366 if (!summary_only && topo.num_cpus > 1) in topology_probe()
5370 fprintf(outf, "num_cpus %d max_cpu_num %d\n", topo.num_cpus, topo.max_cpu_num); in topology_probe()
5372 cpus = calloc(1, (topo.max_cpu_num + 1) * sizeof(struct cpu_topology)); in topology_probe()
5379 cpu_present_set = CPU_ALLOC((topo.max_cpu_num + 1)); in topology_probe()
5382 cpu_present_setsize = CPU_ALLOC_SIZE((topo.max_cpu_num + 1)); in topology_probe()
5398 cpu_affinity_set = CPU_ALLOC((topo.max_cpu_num + 1)); in topology_probe()
5401 cpu_affinity_setsize = CPU_ALLOC_SIZE((topo.max_cpu_num + 1)); in topology_probe()
5410 for (i = 0; i <= topo.max_cpu_num; ++i) { in topology_probe()
5433 if (cpus[i].physical_node_id > topo.max_node_num) in topology_probe()
5434 topo.max_node_num = cpus[i].physical_node_id; in topology_probe()
5446 topo.num_cores++; in topology_probe()
5449 topo.cores_per_node = max_core_id + 1; in topology_probe()
5452 max_core_id, topo.cores_per_node); in topology_probe()
5453 if (!summary_only && topo.cores_per_node > 1) in topology_probe()
5456 topo.num_die = max_die_id + 1; in topology_probe()
5459 max_die_id, topo.num_die); in topology_probe()
5460 if (!summary_only && topo.num_die > 1) in topology_probe()
5463 topo.num_packages = max_package_id + 1; in topology_probe()
5466 max_package_id, topo.num_packages); in topology_probe()
5467 if (!summary_only && topo.num_packages > 1) in topology_probe()
5472 fprintf(outf, "nodes_per_pkg %d\n", topo.nodes_per_pkg); in topology_probe()
5473 if (!summary_only && topo.nodes_per_pkg > 1) in topology_probe()
5476 topo.threads_per_core = max_siblings; in topology_probe()
5483 for (i = 0; i <= topo.max_cpu_num; ++i) { in topology_probe()
5502 int num_cores = topo.cores_per_node * topo.nodes_per_pkg * in allocate_counters()
5503 topo.num_packages; in allocate_counters()
5504 int num_threads = topo.threads_per_core * num_cores; in allocate_counters()
5520 *p = calloc(topo.num_packages, sizeof(struct pkg_data)); in allocate_counters()
5524 for (i = 0; i < topo.num_packages; i++) in allocate_counters()
5579 output_buffer = calloc(1, (1 + topo.num_cpus) * 2048); in allocate_output_buffer()
5586 fd_percpu = calloc(topo.max_cpu_num + 1, sizeof(int)); in allocate_fd_percpu()
5592 irq_column_2_cpu = calloc(topo.num_cpus, sizeof(int)); in allocate_irq_buffers()
5594 err(-1, "calloc %d", topo.num_cpus); in allocate_irq_buffers()
5596 irqs_per_cpu = calloc(topo.max_cpu_num + 1, sizeof(int)); in allocate_irq_buffers()
5598 err(-1, "calloc %d", topo.max_cpu_num + 1); in allocate_irq_buffers()