/linux/Documentation/RCU/Design/Expedited-Grace-Periods/ |
H A D | Expedited-Grace-Periods.rst | 150 ``rcu_state`` structure's ``->ncpus`` field. The ``rcu_state`` 160 ``rcu_state`` structure's ``->ncpus`` and ``->ncpus_snap`` fields are
|
/linux/Documentation/bpf/ |
H A D | map_array.rst | 216 int ncpus = libbpf_num_possible_cpus(); 217 long values[ncpus]; 222 for (j = 0; j < ncpus; j++) 238 int ncpus = libbpf_num_possible_cpus(); 240 long values[ncpus]; 247 for (j = 0; j < ncpus; j++) { 259 in userspace, each value is an array with ``ncpus`` elements.
|
/linux/arch/loongarch/kernel/ |
H A D | machine_kexec.c | 194 unsigned int ncpus; in crash_smp_send_stop() local 208 ncpus = num_online_cpus() - 1; in crash_smp_send_stop() 219 while ((cpumask_weight(&cpus_in_crash) < ncpus) && timeout--) { in crash_smp_send_stop()
|
/linux/arch/mips/kernel/ |
H A D | crash.c | 59 unsigned int ncpus; in crash_kexec_prepare_cpus() local 64 ncpus = num_online_cpus() - 1;/* Excluding the panic cpu */ in crash_kexec_prepare_cpus() 75 while ((cpumask_weight(&cpus_in_crash) < ncpus) && (--msecs > 0)) { in crash_kexec_prepare_cpus()
|
/linux/arch/powerpc/kernel/ |
H A D | fadump.c | 699 unsigned int ncpus = num_online_cpus() - 1; in crash_fadump() local 746 while ((atomic_read(&cpus_in_fadump) < ncpus) && (--msecs > 0)) in crash_fadump()
|
/linux/arch/powerpc/kexec/ |
H A D | crash.c | 115 volatile unsigned int ncpus = num_online_cpus() - 1;/* Excluding the panic cpu */ in crash_kexec_prepare_cpus() local 122 ncpus = num_present_cpus() - 1; in crash_kexec_prepare_cpus() 142 while ((atomic_read(&cpus_in_crash) < ncpus) && (--msecs > 0)) in crash_kexec_prepare_cpus() 147 if (atomic_read(&cpus_in_crash) >= ncpus) { in crash_kexec_prepare_cpus() 153 ncpus - atomic_read(&cpus_in_crash)); in crash_kexec_prepare_cpus() 184 while (atomic_read(&cpus_in_crash) < ncpus) in crash_kexec_prepare_cpus()
|
/linux/arch/powerpc/platforms/powermac/ |
H A D | smp.c | 271 int i, ncpus; in smp_psurge_probe() local 297 ncpus = 4; in smp_psurge_probe() 309 ncpus = 2; in smp_psurge_probe() 323 if (ncpus > NR_CPUS) in smp_psurge_probe() 324 ncpus = NR_CPUS; in smp_psurge_probe() 325 for (i = 1; i < ncpus ; ++i) in smp_psurge_probe() 753 for (i = 1; i < ncpus; ++i) in smp_core99_setup() 766 int ncpus = 0; in smp_core99_probe() local 772 ++ncpus; in smp_core99_probe() 777 if (ncpus <= 1) in smp_core99_probe() [all …]
|
/linux/arch/powerpc/xmon/ |
H A D | xmon.c | 464 static bool wait_for_other_cpus(int ncpus) in wait_for_other_cpus() argument 470 if (cpumask_weight(&cpus_in_xmon) >= ncpus) in wait_for_other_cpus() 595 int ncpus = num_online_cpus(); in xmon_core() local 599 if (ncpus > 1) { in xmon_core() 607 if (TRAP(regs) != INTERRUPT_SYSTEM_RESET || !wait_for_other_cpus(ncpus)) in xmon_core() 610 wait_for_other_cpus(ncpus); in xmon_core()
|
/linux/arch/sparc/kernel/ |
H A D | ds.c | 502 tag->num_records = ncpus; in dr_cpu_init_response() 511 BUG_ON(i != ncpus); in dr_cpu_init_response() 524 for (i = 0; i < ncpus; i++) { in dr_cpu_mark() 537 int resp_len, ncpus, cpu; in dr_cpu_configure() local 540 ncpus = cpumask_weight(mask); in dr_cpu_configure() 541 resp_len = dr_cpu_size_response(ncpus); in dr_cpu_configure() 547 resp_len, ncpus, mask, in dr_cpu_configure() 596 int resp_len, ncpus, cpu; in dr_cpu_unconfigure() local 599 ncpus = cpumask_weight(mask); in dr_cpu_unconfigure() 606 resp_len, ncpus, mask, in dr_cpu_unconfigure() [all …]
|
H A D | setup_32.c | 378 int i, ncpus, err; in topology_init() local 384 ncpus = 0; in topology_init() 385 while (!cpu_find_by_instance(ncpus, NULL, NULL)) in topology_init() 386 ncpus++; in topology_init() 387 ncpus_probed = ncpus; in topology_init()
|
H A D | smp_64.c | 1151 int ncpus = num_online_cpus(); in smp_capture() local 1160 while (atomic_read(&smp_capture_registry) != ncpus) in smp_capture()
|
H A D | sun4m_smp.c | 177 register int ncpus = SUN4M_NCPUS; in sun4m_cross_call() local 196 for (i = 0; i < ncpus; i++) { in sun4m_cross_call() 217 } while (++i < ncpus); in sun4m_cross_call() 225 } while (++i < ncpus); in sun4m_cross_call()
|
/linux/arch/um/include/shared/ |
H A D | kern_util.h | 16 extern int ncpus;
|
/linux/arch/um/kernel/ |
H A D | process.c | 55 for (i = 0; i < ncpus; i++) { in pid_to_processor_id()
|
H A D | um_arch.c | 129 int ncpus = 1; variable
|
/linux/arch/x86/include/asm/trace/ |
H A D | hyperv.h | 16 __field(unsigned int, ncpus) 21 TP_fast_assign(__entry->ncpus = cpumask_weight(cpus); 27 __entry->ncpus, __entry->mm, 64 __field(unsigned int, ncpus) 67 TP_fast_assign(__entry->ncpus = cpumask_weight(cpus); 71 __entry->ncpus, __entry->vector)
|
/linux/arch/x86/kernel/cpu/ |
H A D | topology.h | 21 unsigned int shift, unsigned int ncpus); 50 unsigned int shift, unsigned int ncpus) in topology_update_dom() argument 53 tscan->dom_ncpus[dom] = ncpus; in topology_update_dom()
|
H A D | topology_common.c | 19 unsigned int shift, unsigned int ncpus) in topology_set_dom() argument 21 topology_update_dom(tscan, dom, shift, ncpus); in topology_set_dom()
|
/linux/arch/x86/kernel/ |
H A D | kvmclock.c | 204 unsigned long ncpus; in kvmclock_init_mem() local 212 ncpus = num_possible_cpus() - HVC_BOOT_ARRAY_SIZE; in kvmclock_init_mem() 213 order = get_order(ncpus * sizeof(*hvclock_mem)); in kvmclock_init_mem()
|
/linux/arch/x86/platform/uv/ |
H A D | uv_time.c | 52 int ncpus; member 56 } cpu[] __counted_by(ncpus); 158 head->ncpus = uv_blade_nr_possible_cpus(bid); in uv_rtc_allocate_timers() 177 for (c = 0; c < head->ncpus; c++) { in uv_rtc_find_next_timer()
|
/linux/arch/xtensa/kernel/ |
H A D | smp.c | 91 unsigned int ncpus = get_core_count(); in smp_init_cpus() local 94 pr_info("%s: Core Count = %d\n", __func__, ncpus); in smp_init_cpus() 97 if (ncpus > NR_CPUS) { in smp_init_cpus() 98 ncpus = NR_CPUS; in smp_init_cpus() 99 pr_info("%s: limiting core count by %d\n", __func__, ncpus); in smp_init_cpus() 102 for (i = 0; i < ncpus; ++i) in smp_init_cpus()
|
/linux/drivers/block/aoe/ |
H A D | aoecmd.c | 44 static int ncpus; variable 1202 actual_id = f->t->d->aoeminor % ncpus; in ktio() 1275 id = f->t->d->aoeminor % ncpus; in ktcomplete() 1635 for (i = 0; i < ncpus; i++) { in aoe_flush_iocq() 1685 ncpus = num_online_cpus(); in aoecmd_init() 1687 iocq = kcalloc(ncpus, sizeof(struct iocq_ktio), GFP_KERNEL); in aoecmd_init() 1691 kts = kcalloc(ncpus, sizeof(struct ktstate), GFP_KERNEL); in aoecmd_init() 1697 ktiowq = kcalloc(ncpus, sizeof(wait_queue_head_t), GFP_KERNEL); in aoecmd_init() 1703 for (i = 0; i < ncpus; i++) { in aoecmd_init() 1736 for (i = 0; i < ncpus; i++) in aoecmd_exit()
|
/linux/drivers/clk/mvebu/ |
H A D | clk-cpu.c | 171 int ncpus = num_possible_cpus(); in of_cpu_clk_setup() local 184 cpuclk = kcalloc(ncpus, sizeof(*cpuclk), GFP_KERNEL); in of_cpu_clk_setup() 188 clks = kcalloc(ncpus, sizeof(*clks), GFP_KERNEL); in of_cpu_clk_setup() 228 while(ncpus--) in of_cpu_clk_setup() 229 kfree(cpuclk[ncpus].clk_name); in of_cpu_clk_setup()
|
/linux/drivers/dma-buf/ |
H A D | st-dma-fence-chain.c | 445 int ncpus = num_online_cpus(); in find_race() local 455 threads = kmalloc_array(ncpus, sizeof(*threads), GFP_KERNEL); in find_race() 462 for (i = 0; i < ncpus; i++) { in find_race() 465 ncpus = i; in find_race() 476 for (i = 0; i < ncpus; i++) { in find_race()
|
/linux/drivers/gpu/drm/i915/selftests/ |
H A D | i915_request.c | 463 unsigned int ncpus = num_online_cpus(); in mock_breadcrumbs_smoketest() local 492 for (n = 0; n < ncpus; n++) { in mock_breadcrumbs_smoketest() 498 ncpus = n; in mock_breadcrumbs_smoketest() 514 for (n = 0; n < ncpus; n++) { in mock_breadcrumbs_smoketest() 528 ncpus); in mock_breadcrumbs_smoketest() 1801 smoke[idx].max_batch /= ncpus + 1; in live_breadcrumbs_smoketest() 1805 for (n = 0; n < ncpus; n++) { in live_breadcrumbs_smoketest() 1806 unsigned int i = idx * ncpus + n; in live_breadcrumbs_smoketest() 1833 for (n = 0; n < ncpus; n++) { in live_breadcrumbs_smoketest() 1834 unsigned int i = idx * ncpus + n; in live_breadcrumbs_smoketest() [all …]
|