/netbsd/sys/sys/ |
H A D | cpu_data.h | 200 #define ci_index ci_data.cpu_index 203 #define ci_cpuname ci_data.cpu_name 206 #define ci_lockstat ci_data.cpu_lockstat 210 #define ci_kcpuset ci_data.cpu_kcpuset 211 #define ci_ipipend ci_data.cpu_ipipend 215 #define ci_core_id ci_data.cpu_core_id 216 #define ci_smt_id ci_data.cpu_smt_id 217 #define ci_numa_id ci_data.cpu_numa_id 218 #define ci_is_slow ci_data.cpu_is_slow 220 #define ci_sibling ci_data.cpu_sibling [all …]
|
/netbsd/sys/kern/ |
H A D | kern_softint.c | 318 ci->ci_data.cpu_softcpu = sc; in softint_init() 319 ci->ci_data.cpu_softints = 0; in softint_init() 333 scfirst = first->ci_data.cpu_softcpu; in softint_init() 370 sc = curcpu()->ci_data.cpu_softcpu; in softint_establish() 393 sc = curcpu()->ci_data.cpu_softcpu; in softint_establish() 401 sc = ci->ci_data.cpu_softcpu; in softint_establish() 444 sc = curcpu()->ci_data.cpu_softcpu; in softint_disestablish() 466 sc = ci->ci_data.cpu_softcpu; in softint_disestablish() 691 ci->ci_data.cpu_softints |= machdep; in softint_trigger() 701 if (l == ci->ci_data.cpu_idlelwp) { in softint_trigger() [all …]
|
H A D | subr_xcall.c | 245 cv_init(&ci->ci_data.cpu_xcall, "xcall"); in xc_init_cpu() 415 ci->ci_data.cpu_xcall_pending = true; in xc_lowpri() 416 cv_signal(&ci->ci_data.cpu_xcall); in xc_lowpri() 421 ci->ci_data.cpu_xcall_pending = true; in xc_lowpri() 422 cv_signal(&ci->ci_data.cpu_xcall); in xc_lowpri() 448 while (!ci->ci_data.cpu_xcall_pending) { in xc_thread() 452 cv_wait(&ci->ci_data.cpu_xcall, &xc->xc_lock); in xc_thread() 455 ci->ci_data.cpu_xcall_pending = false; in xc_thread()
|
H A D | kern_cpu.c | 101 CTASSERT(offsetof(struct cpu_info, ci_data) == 0); 103 CTASSERT(offsetof(struct cpu_info, ci_data) != 0); 145 TAILQ_INIT(&ci->ci_data.cpu_ld_locks); in mi_cpu_attach() 146 __cpu_simple_lock_init(&ci->ci_data.cpu_ld_lock); in mi_cpu_attach() 149 snprintf(ci->ci_data.cpu_name, sizeof(ci->ci_data.cpu_name), "cpu%d", in mi_cpu_attach() 169 ci->ci_onproc = ci->ci_data.cpu_idlelwp; in mi_cpu_attach() 178 TAILQ_INIT(&ci->ci_data.cpu_biodone); in mi_cpu_attach()
|
H A D | subr_lockdebug.c | 176 __cpu_simple_lock(&ci->ci_data.cpu_ld_lock); in lockdebug_lookup1() 178 __cpu_simple_unlock(&ci->ci_data.cpu_ld_lock); in lockdebug_lookup1() 194 __cpu_simple_lock(&ci->ci_data.cpu_ld_lock); in lockdebug_lock_cpus() 205 __cpu_simple_unlock(&ci->ci_data.cpu_ld_lock); in lockdebug_unlock_cpus() 243 TAILQ_INIT(&curcpu()->ci_data.cpu_ld_locks); in lockdebug_init() 245 __cpu_simple_lock_init(&curcpu()->ci_data.cpu_ld_lock); in lockdebug_init() 536 TAILQ_INSERT_TAIL(&curcpu()->ci_data.cpu_ld_locks, in lockdebug_locked() 609 TAILQ_REMOVE(&curcpu()->ci_data.cpu_ld_locks, ld, in lockdebug_unlocked() 693 __cpu_simple_lock(&ci->ci_data.cpu_ld_lock); in lockdebug_mem_check() 704 __cpu_simple_unlock(&ci->ci_data.cpu_ld_lock); in lockdebug_mem_check() [all …]
|
H A D | kern_timeout.c | 287 KASSERT(curcpu()->ci_data.cpu_callout == NULL); in callout_startup() 294 curcpu()->ci_data.cpu_callout = cc; in callout_startup() 310 if ((cc = ci->ci_data.cpu_callout) == NULL) { in callout_init_cpu() 337 ci->ci_data.cpu_callout = cc; in callout_init_cpu() 356 cc = curcpu()->ci_data.cpu_callout; in callout_init() 437 cc = curcpu()->ci_data.cpu_callout; in callout_schedule_locked() 660 cc = ci->ci_data.cpu_callout; in callout_bind() 781 cc = curcpu()->ci_data.cpu_callout; in callout_hardclock() 822 cc = l->l_cpu->ci_data.cpu_callout; in callout_softclock() 936 offsetof(struct cpu_info, ci_data.cpu_callout), in db_show_callout() [all …]
|
H A D | kern_idle.c | 106 KASSERT(ci->ci_data.cpu_idlelwp == NULL); in create_idle_lwp() 125 ci->ci_data.cpu_idlelwp = l; in create_idle_lwp()
|
/netbsd/sys/arch/powerpc/powerpc/ |
H A D | softint_machdep.c | 58 KASSERT(ci->ci_data.cpu_softints & (1 << ipl)); in softint_deliver() 59 ci->ci_data.cpu_softints ^= 1 << ipl; in softint_deliver() 76 while ((softints = (ci->ci_data.cpu_softints & softint_mask)) != 0) { in powerpc_softint() 100 ci->ci_data.cpu_softints ^= (1 << IPL_NONE); in powerpc_softint() 121 atomic_or_uint(&ci->ci_data.cpu_softints, machdep); in powerpc_softint_trigger()
|
/netbsd/sys/arch/x86/x86/ |
H A D | cpu.c | 772 if (ci->ci_data.cpu_idlelwp == NULL) in cpu_boot_secondary_processors() 795 struct lwp *l = ci->ci_data.cpu_idlelwp; in cpu_init_idle_lwp() 811 if (ci->ci_data.cpu_idlelwp == NULL) in cpu_init_idle_lwps() 890 drift = ci->ci_data.cpu_cc_skew; in cpu_boot_secondary() 895 drift -= ci->ci_data.cpu_cc_skew; in cpu_boot_secondary() 951 ci->ci_data.cpu_cc_freq = cpu_info_primary.ci_data.cpu_cc_freq; in cpu_hatch() 1335 if (ci->ci_data.cpu_idlelwp == NULL) in cpu_stop() 1380 if (ci->ci_data.cpu_idlelwp == NULL) in cpu_resume() 1428 if (ci->ci_data.cpu_cc_freq == 0) in cpu_get_tsc_freq() 1477 if (ci->ci_data.cpu_cc_freq != 0) { in cpu_get_tsc_freq() [all …]
|
H A D | tsc.c | 215 tsc_freq = ci->ci_data.cpu_cc_freq; in tsc_tc_init() 304 ci->ci_data.cpu_cc_skew = val; in tsc_sync_bp() 383 return ci->ci_data.cpu_cc_freq; in cpu_frequency()
|
/netbsd/sys/arch/powerpc/booke/ |
H A D | e500_timer.c | 153 cpu->cpu_ticks_per_clock_intr = (ci->ci_data.cpu_cc_freq + hz/2 - 1) / hz; in cpu_initclocks() 172 ci->ci_data.cpu_cc_freq = board_info_get_number("timebase-frequency"); in calc_delayconst() 173 ticks_per_sec = (uint32_t)ci->ci_data.cpu_cc_freq; in calc_delayconst() 174 ns_per_tick = 1000000000 / (u_int)ci->ci_data.cpu_cc_freq; in calc_delayconst() 225 ppcbooke_timecounter.tc_frequency = curcpu()->ci_data.cpu_cc_freq; in init_ppcbooke_tc()
|
/netbsd/sys/arch/arm/arm32/ |
H A D | cortex_pmc.c | 60 #define COUNTS_PER_USEC (curcpu()->ci_data.cpu_cc_freq / (1000*1000)) 70 if (curcpu()->ci_data.cpu_cc_freq == 0) { in cortex_pmc_ccnt_init() 71 curcpu()->ci_data.cpu_cc_freq = CORTEX_PMC_CCNT_HZ; in cortex_pmc_ccnt_init()
|
/netbsd/sys/arch/riscv/riscv/ |
H A D | cpu_subr.c | 197 if (kcpuset_match(cpus_running, ci->ci_data.cpu_kcpuset)) in cpu_multicast_ipi() 201 kcpuset_remove(kcp2, ci->ci_data.cpu_kcpuset); in cpu_multicast_ipi() 263 if (kcpuset_match(cpus_running, curcpu()->ci_data.cpu_kcpuset)) in cpu_halt_others() 269 kcpuset_remove(kcp, curcpu()->ci_data.cpu_kcpuset); in cpu_halt_others() 326 if (cold || kcpuset_match(cpus_running, ci->ci_data.cpu_kcpuset)) in cpu_pause_others() 332 kcpuset_remove(kcp, ci->ci_data.cpu_kcpuset); in cpu_pause_others()
|
/netbsd/sys/arch/xen/x86/ |
H A D | cpu.c | 415 KASSERT(ci->ci_data.cpu_idlelwp != NULL); in cpu_attach_common() 423 KASSERT(ci->ci_data.cpu_idlelwp != NULL); in cpu_attach_common() 512 struct lwp *l = ci->ci_data.cpu_idlelwp; in cpu_attach_common() 601 if (ci->ci_data.cpu_idlelwp == NULL) in cpu_boot_secondary_processors() 620 struct lwp *l = ci->ci_data.cpu_idlelwp; in cpu_init_idle_lwp() 636 if (ci->ci_data.cpu_idlelwp == NULL) in cpu_init_idle_lwps() 651 ci->ci_curlwp = ci->ci_data.cpu_idlelwp; in cpu_start_secondary() 728 KASSERT(curlwp == ci->ci_data.cpu_idlelwp); in cpu_hatch() 842 l = ci->ci_data.cpu_idlelwp; in xen_init_amd64_vcpuctxt() 939 l = ci->ci_data.cpu_idlelwp; in xen_init_i386_vcpuctxt() [all …]
|
/netbsd/sys/arch/arm/cortex/ |
H A D | a9tmr.c | 216 __func__, ci->ci_data.cpu_name, in a9tmr_init_cpu_clock() 234 __func__, ci->ci_data.cpu_name, in a9tmr_init_cpu_clock() 298 curcpu()->ci_data.cpu_cc_freq / 2; in a9tmr_delay() 335 __func__, cf, ci->ci_data.cpu_name, now, delta); in a9tmr_intr() 339 ci->ci_data.cpu_name, delta); in a9tmr_intr()
|
/netbsd/sys/arch/or1k/include/ |
H A D | cpu_counter.h | 44 #define cpu_hascounter() (curcpu()->ci_data.cpu_cc_freq != 0) 56 return ci->ci_data.cpu_cc_freq; in cpu_frequency()
|
H A D | cpu.h | 51 struct cpu_data ci_data; member 111 && (ci->ci_data.cpu_softints >> ci->ci_cpl) > 0) in cpu_dosoftints()
|
/netbsd/sys/arch/aarch64/include/ |
H A D | cpu_counter.h | 46 #define cpu_hascounter() (curcpu()->ci_data.cpu_cc_freq != 0) 63 return ci->ci_data.cpu_cc_freq; in cpu_frequency()
|
/netbsd/sys/arch/arm/include/ |
H A D | cpu_counter.h | 45 #define cpu_hascounter() (curcpu()->ci_data.cpu_cc_freq != 0) 74 return ci->ci_data.cpu_cc_freq; in cpu_frequency()
|
/netbsd/sys/arch/powerpc/include/ |
H A D | userret.h | 67 || l->l_cpu->ci_data.cpu_pcu_curlwp[PCU_FPU] == l, in userret() 69 tf, tf->tf_srr1, l->l_cpu->ci_data.cpu_pcu_curlwp[PCU_FPU], l); in userret()
|
/netbsd/sys/rump/librump/rumpkern/ |
H A D | intr.c | 153 struct cpu_data *cd = &curcpu()->ci_data; in sithread() 285 struct cpu_data *cd = &ci->ci_data; in softint_init() 379 struct cpu_data *cd = &ci->ci_data; in softint_schedule() 458 struct cpu_data *cd = &ci->ci_data; in rump_softint_run()
|
/netbsd/sys/uvm/ |
H A D | uvm_pgflcache.c | 276 ucpu = ci->ci_data.cpu_uvm; in uvm_pgflcache_init_cpu() 298 ucpu = curcpu()->ci_data.cpu_uvm; in uvm_pgflcache_fini_cpu() 362 ucpu = ci->ci_data.cpu_uvm; in uvm_pgflcache_resume()
|
/netbsd/sys/arch/powerpc/ibm4xx/ |
H A D | clock.c | 97 ci->ci_data.cpu_nintr++; in stat_intr() 138 ci->ci_data.cpu_nintr++; in decr_intr()
|
/netbsd/sys/arch/sun68k/sun68k/ |
H A D | isr.c | 113 curcpu()->ci_data.cpu_nintr++; in isr_autovec() 179 curcpu()->ci_data.cpu_nintr++; in isr_vectored()
|
/netbsd/sys/arch/riscv/include/ |
H A D | cpu.h | 53 struct cpu_data ci_data; member 222 && (ci->ci_data.cpu_softints >> ci->ci_cpl) > 0) in cpu_dosoftints()
|