/dragonfly/sys/vm/ |
H A D | vm_page2.h | 81 globaldata_t gd = mycpu; in vm_paging_severe() local 114 globaldata_t gd = mycpu; in vm_paging_min_dnc() local 175 globaldata_t gd = mycpu; in vm_paging_wait() local 204 globaldata_t gd = mycpu; in vm_paging_start() local 235 globaldata_t gd = mycpu; in vm_paging_target1() local 255 globaldata_t gd = mycpu; in vm_paging_target1_count() local 259 (gd->gd_vmstats.v_free_count + gd->gd_vmstats.v_cache_count); in vm_paging_target1_count() 273 globaldata_t gd = mycpu; in vm_paging_target2() local 297 (gd->gd_vmstats.v_free_count + gd->gd_vmstats.v_cache_count); in vm_paging_target2_count() 348 delta = (gd->gd_vmstats.v_free_min + gd->gd_vmstats.v_inactive_target) - in vm_paging_inactive_count() [all …]
|
H A D | vm_meter.c | 110 globaldata_t gd; in do_vmtotal() local 115 gd = globaldata_find(n); in do_vmtotal() 122 total.t_vm += gd->gd_vmtotal.t_vm; in do_vmtotal() 124 total.t_rm += gd->gd_vmtotal.t_rm; in do_vmtotal() 268 count += gd->gd_cnt.v_intr + gd->gd_cnt.v_ipi + in vcnt_intr() 269 gd->gd_cnt.v_timer; in vcnt_intr() 431 struct globaldata *gd = arg1; in do_vmmeter_pcpu() local 434 vmm = gd->gd_cnt; in do_vmmeter_pcpu() 520 struct globaldata *gd; in vmmeter_init() local 530 gd = globaldata_find(i); in vmmeter_init() [all …]
|
/dragonfly/sys/kern/ |
H A D | lwkt_ipiq.c | 79 #define ipiq_stat(gd) ipiq_stats_percpu[(gd)->gd_cpuid] argument 361 crit_exit_gd(gd); in lwkt_send_ipiq3_passive() 563 ip += gd->gd_cpuid; in lwkt_process_ipiq() 584 if (lwkt_process_ipiq_core(gd, &gd->gd_cpusyncq, NULL, 0)) { in lwkt_process_ipiq() 615 ip += gd->gd_cpuid; in lwkt_process_ipiq_frame() 630 if (gd->gd_cpusyncq.ip_rindex != gd->gd_cpusyncq.ip_windex) { in lwkt_process_ipiq_frame() 631 if (lwkt_process_ipiq_core(gd, &gd->gd_cpusyncq, frame, 0)) { in lwkt_process_ipiq_frame() 676 ip += gd->gd_cpuid; in lwkt_process_ipiq_nested() 691 if (lwkt_process_ipiq_core(gd, &gd->gd_cpusyncq, NULL, 0)) { in lwkt_process_ipiq_nested() 1035 gd->gd_cpuid, in lwkt_cpusync_remote2() [all …]
|
H A D | lwkt_thread.c | 159 --gd->gd_tdrunqcount; in _lwkt_dequeue() 238 ++gd->gd_tdrunqcount; in _lwkt_enqueue() 364 crit_enter_gd(gd); in lwkt_alloc_thread() 374 crit_exit_gd(gd); in lwkt_alloc_thread() 451 td->td_gd = gd; in lwkt_init_thread() 470 if (gd == mygd || td == &gd->gd_idlethread) { in lwkt_init_thread() 719 gd->gd_cpuid, in lwkt_switch() 1380 globaldata_t gd; in lwkt_acquire() local 1384 gd = td->td_gd; in lwkt_acquire() 1622 td->td_gd = gd; in lwkt_setcpu_remote() [all …]
|
H A D | kern_synch.c | 463 globaldata_t gd = mycpu; in _tsleep_remove() local 517 globaldata_t gd; in tsleep() local 553 gd = td->td_gd; in tsleep() 824 KKASSERT(gd == mycpu); in ssleep() 1004 globaldata_t gd; in _wakeup() local 1012 gd = mycpu; in _wakeup() 1215 if (gd == mycpu) { in wakeup_oncpu() 1233 if (gd == mygd) { in wakeup_oncpu_one() 1283 if (gd->gd_delayed_wakeup[0] || gd->gd_delayed_wakeup[1]) { in wakeup_end_delayed() 1471 globaldata_t gd; in sched_setup() local [all …]
|
H A D | kern_systimer.c | 71 globaldata_t gd = mycpu; in systimer_intr() local 75 if (gd->gd_syst_nest) in systimer_intr() 79 ++gd->gd_syst_nest; in systimer_intr() 130 --gd->gd_syst_nest; in systimer_intr() 150 if (info->gd == gd) { in systimer_add() 218 struct globaldata *gd = info->gd; in systimer_del() local 287 info->gd = mycpu; in _systimer_init_periodic() 368 info->gd = mycpu; in systimer_init_oneshot() 380 globaldata_t gd = mycpu; in systimer_changed_pcpu() local 407 globaldata_t gd = mycpu; in systimer_changed() local [all …]
|
H A D | kern_clock.c | 226 * in seconds (gd->gd_time_seconds). 345 struct globaldata *gd = mycpu; in initclocks_pcpu() local 348 if (gd->gd_cpuid == 0) { in initclocks_pcpu() 349 gd->gd_time_seconds = 1; in initclocks_pcpu() 350 gd->gd_cpuclock_base = sys_cputimer->count(); in initclocks_pcpu() 351 hardtime[0].time_second = gd->gd_time_seconds; in initclocks_pcpu() 352 hardtime[0].cpuclock_base = gd->gd_cpuclock_base; in initclocks_pcpu() 354 gd->gd_time_seconds = globaldata_find(0)->gd_time_seconds; in initclocks_pcpu() 355 gd->gd_cpuclock_base = globaldata_find(0)->gd_cpuclock_base; in initclocks_pcpu() 418 struct globaldata *gd; in initclocks_other() local 470 struct globaldata *gd; initclocks_usched() local 556 struct globaldata *gd = mycpu; hardclock() local 795 globaldata_t gd; hardclock() local 892 globaldata_t gd = mycpu; statclock() local 1371 struct globaldata *gd = mycpu; getmicrouptime() local 1393 struct globaldata *gd = mycpu; getnanouptime() local 1411 struct globaldata *gd = mycpu; microuptime() local 1429 struct globaldata *gd = mycpu; nanouptime() local 1507 struct globaldata *gd = mycpu; getnanotime_nbt() local 1533 struct globaldata *gd = mycpu; microtime() local 1561 struct globaldata *gd = mycpu; nanotime() local 1599 struct globaldata *gd = mycpu; get_approximate_time_t() local 1676 struct globaldata *gd; pps_event() local [all...] |
H A D | usched_dummy.c | 151 globaldata_t gd = mycpu; in dummy_acquire_curproc() local 159 dummy_select_curproc(gd); in dummy_acquire_curproc() 191 gd = mycpu; in dummy_acquire_curproc() 192 dd = &dummy_pcpu[gd->gd_cpuid]; in dummy_acquire_curproc() 212 globaldata_t gd = mycpu; in dummy_release_curproc() local 217 dummy_select_curproc(gd); in dummy_release_curproc() 275 globaldata_t gd = mycpu; in dummy_setrunqueue() local 335 globaldata_t gd = mycpu; in dummy_schedulerclock() local 467 globaldata_t gd; in dummy_sched_thread() local 475 gd = mycpu; in dummy_sched_thread() [all …]
|
H A D | kern_sysref.c | 109 globaldata_t gd; in sysref_init() local 111 gd = mycpu; in sysref_init() 112 crit_enter_gd(gd); in sysref_init() 124 crit_exit_gd(gd); in sysref_init() 197 globaldata_t gd; in sysref_ctor() local 214 gd = mycpu; in sysref_ctor() 215 crit_enter_gd(gd); in sysref_ctor() 227 crit_exit_gd(gd); in sysref_ctor() 365 globaldata_t gd = mycpu; in allocsysid() local 368 crit_enter_gd(gd); in allocsysid() [all …]
|
H A D | usched_bsd4.c | 332 globaldata_t gd; in bsd4_acquire_curproc() local 366 gd = mycpu; in bsd4_acquire_curproc() 373 gd->gd_cpuid); in bsd4_acquire_curproc() 447 gd = mycpu; in bsd4_acquire_curproc() 531 crit_enter_gd(gd); in bsd4_select_curproc() 569 crit_exit_gd(gd); 627 globaldata_t gd; in bsd4_setrunqueue() local 982 cpbase = gd->gd_schedclock.time - gd->gd_schedclock.periodic; in bsd4_recalculate_estcpu() 1592 globaldata_t gd; in bsd4_kick_helper() local 1759 gd = mycpu; in sched_thread() [all …]
|
H A D | usched_dfly.c | 327 globaldata_t gd; in dfly_acquire_curproc() local 334 gd = mycpu; in dfly_acquire_curproc() 401 gd = mycpu; in dfly_acquire_curproc() 428 gd = mycpu; in dfly_acquire_curproc() 518 gd = mycpu; in dfly_acquire_curproc() 539 gd = mycpu; in dfly_acquire_curproc() 560 gd = mycpu; in dfly_acquire_curproc() 794 rgd = rdd->gd; in dfly_setrunqueue_dd() 872 if (gd->gd_curthread == &gd->gd_idlethread) { in dfly_schedulerclock() 1012 cpbase = gd->gd_schedclock.time - gd->gd_schedclock.periodic; in dfly_recalculate_estcpu() [all …]
|
H A D | kern_intr.c | 720 globaldata_t gd; in ithread_fast_handler() local 724 gd = mycpu; in ithread_fast_handler() 740 ++gd->gd_cnt.v_intr; in ithread_fast_handler() 761 ++gd->gd_cnt.v_intr; in ithread_fast_handler() 840 globaldata_t gd; in ithread_handler() local 855 gd = mycpu; in ithread_handler() 857 crit_enter_gd(gd); in ithread_handler() 937 crit_exit_gd(gd); in ithread_handler() 938 crit_enter_gd(gd); in ithread_handler() 1034 globaldata_t gd = mycpu; in ithread_emergency() local [all …]
|
/dragonfly/sys/sys/ |
H A D | spinlock2.h | 70 globaldata_t gd = mycpu; in spin_trylock() local 72 crit_enter_raw(gd->gd_curthread); in spin_trylock() 73 ++gd->gd_spinlocks; in spin_trylock() 111 crit_enter_raw(gd->gd_curthread); in _spin_lock_quick() 112 ++gd->gd_spinlocks; in _spin_lock_quick() 173 KKASSERT(gd->gd_spinlocks > 0); in spin_unlock_quick() 176 --gd->gd_spinlocks; in spin_unlock_quick() 208 crit_enter_raw(gd->gd_curthread); in _spin_lock_shared_quick() 209 ++gd->gd_spinlocks; in _spin_lock_shared_quick() 266 KKASSERT(gd->gd_spinlocks > 0); in spin_unlock_shared_quick() [all …]
|
H A D | mutex2.h | 220 globaldata_t gd = mycpu; in mtx_spinlock() local 225 crit_enter_quick(gd->gd_curthread); in mtx_spinlock() 226 ++gd->gd_spinlocks; in mtx_spinlock() 238 mtx->mtx_owner = gd->gd_curthread; in mtx_spinlock() 244 globaldata_t gd = mycpu; in mtx_spinlock_try() local 249 crit_enter_quick(gd->gd_curthread); in mtx_spinlock_try() 250 ++gd->gd_spinlocks; in mtx_spinlock_try() 259 mtx->mtx_owner = gd->gd_curthread; in mtx_spinlock_try() 392 globaldata_t gd = mycpu; in mtx_spinunlock() local 397 --gd->gd_spinlocks; in mtx_spinunlock() [all …]
|
H A D | thread2.h | 189 _crit_enter(globaldata_t gd __DEBUG_CRIT_ADD_ARG__) in _crit_enter() 195 _crit_enter_hard(globaldata_t gd __DEBUG_CRIT_ADD_ARG__) in _crit_enter_hard() 198 ++gd->gd_intr_nesting_level; in _crit_enter_hard() 235 _crit_exit(globaldata_t gd __DEBUG_CRIT_ADD_ARG__) in _crit_exit() 237 _crit_exit_quick(gd->gd_curthread __DEBUG_CRIT_PASS_ARG__); in _crit_exit() 241 _crit_exit_hard(globaldata_t gd __DEBUG_CRIT_ADD_ARG__) in _crit_exit_hard() 244 --gd->gd_intr_nesting_level; in _crit_exit_hard() 245 _crit_exit_quick(gd->gd_curthread __DEBUG_CRIT_PASS_ARG__); in _crit_exit_hard() 362 lwkt_need_ipiq_process(globaldata_t gd) in lwkt_need_ipiq_process() argument 366 if (CPUMASK_TESTNZERO(gd->gd_ipimask)) in lwkt_need_ipiq_process() [all …]
|
/dragonfly/sys/platform/vkernel64/platform/ |
H A D | machintr.c | 124 struct mdglobaldata *gd = mdcpu; in splz() local 125 thread_t td = gd->mi.gd_curthread; in splz() 131 if (gd->mi.gd_reqflags & RQF_IPIQ) { in splz() 133 atomic_swap_int(&gd->mi.gd_npoll, 0); in splz() 150 if (gd->mi.gd_reqflags & RQF_TIMER) { in splz() 171 struct mdglobaldata *gd = mdcpu; in signalintr() local 172 thread_t td = gd->mi.gd_curthread; in signalintr() 177 umtx_wakeup(&gd->mi.gd_reqflags, 0); in signalintr() 221 struct mdglobaldata *gd = mdcpu; in cpu_interrupt_running() local 224 gd->gd_fpending || in cpu_interrupt_running() [all …]
|
H A D | ipl_funcs.c | 51 struct mdglobaldata *gd = mdcpu; \ 53 atomic_set_int(&gd->mi.gd_reqflags, RQF_INTPEND); \ 54 umtx_wakeup(&gd->mi.gd_reqflags, 0); \ 57 DO_SETBITS(setdelayed, &gd->gd_spending, loadandclear(&gd->gd_sdelayed)) 59 DO_SETBITS(setsoftcamnet,&gd->gd_spending, SWI_CAMNET_PENDING) 60 DO_SETBITS(setsoftcambio,&gd->gd_spending, SWI_CAMBIO_PENDING) 63 DO_SETBITS(setsofttty, &gd->gd_spending, SWI_TTY_PENDING) 64 DO_SETBITS(setsoftvm, &gd->gd_spending, SWI_VM_PENDING) 65 DO_SETBITS(setsofttq, &gd->gd_spending, SWI_TQ_PENDING) 66 DO_SETBITS(setsoftcrypto,&gd->gd_spending, SWI_CRYPTO_PENDING) [all …]
|
H A D | globaldata.c | 66 cpu_gdinit(struct mdglobaldata *gd, int cpu) in cpu_gdinit() argument 69 gd->mi.gd_curthread = &gd->mi.gd_idlethread; in cpu_gdinit() 71 lwkt_init_thread(&gd->mi.gd_idlethread, in cpu_gdinit() 72 gd->mi.gd_prvspace->idlestack, in cpu_gdinit() 73 sizeof(gd->mi.gd_prvspace->idlestack), in cpu_gdinit() 74 0, &gd->mi); in cpu_gdinit() 75 lwkt_set_comm(&gd->mi.gd_idlethread, "idle_%d", cpu); in cpu_gdinit() 76 gd->mi.gd_idlethread.td_switch = cpu_lwkt_switch; in cpu_gdinit() 77 gd->mi.gd_idlethread.td_sp -= sizeof(void *); in cpu_gdinit() 78 *(void **)gd->mi.gd_idlethread.td_sp = cpu_idle_restore; in cpu_gdinit()
|
/dragonfly/sys/platform/pc64/x86_64/ |
H A D | ipl_funcs.c | 48 struct mdglobaldata *gd = mdcpu; \ 50 atomic_set_int(&gd->mi.gd_reqflags, RQF_INTPEND); \ 53 DO_SETBITS(setdelayed, &gd->gd_spending, loadandclear(&gd->gd_sdelayed)) 55 DO_SETBITS(setsoftcamnet,&gd->gd_spending, SWI_CAMNET_PENDING) 56 DO_SETBITS(setsoftcambio,&gd->gd_spending, SWI_CAMBIO_PENDING) 59 DO_SETBITS(setsofttty, &gd->gd_spending, SWI_TTY_PENDING) 60 DO_SETBITS(setsoftvm, &gd->gd_spending, SWI_VM_PENDING) 61 DO_SETBITS(setsofttq, &gd->gd_spending, SWI_TQ_PENDING) 62 DO_SETBITS(setsoftcrypto,&gd->gd_spending, SWI_CRYPTO_PENDING) 67 DO_SETBITS(schedsofttty, &gd->gd_sdelayed, SWI_TTY_PENDING)
|
H A D | tls.c | 159 struct mdglobaldata *gd = mdcpu; in set_user_TLS() local 160 thread_t td = gd->mi.gd_curthread; in set_user_TLS() 165 if (gd->gd_user_fs != td->td_pcb->pcb_fsbase) { in set_user_TLS() 166 gd->gd_user_fs = td->td_pcb->pcb_fsbase; in set_user_TLS() 167 wrmsr(MSR_FSBASE, gd->gd_user_fs); in set_user_TLS() 169 if (gd->gd_user_gs != td->td_pcb->pcb_gsbase) { in set_user_TLS() 170 gd->gd_user_gs = td->td_pcb->pcb_gsbase; in set_user_TLS() 171 wrmsr(MSR_KGSBASE, gd->gd_user_gs); in set_user_TLS()
|
/dragonfly/sys/platform/vkernel64/x86_64/ |
H A D | exception.c | 78 globaldata_t gd = mycpu; in ipisig() local 85 ++gd->gd_cnt.v_ipi; in ipisig() 86 ++gd->gd_intr_nesting_level; in ipisig() 89 --gd->gd_intr_nesting_level; in ipisig() 112 globaldata_t gd = mycpu; in stopsig() local 129 ++gd->gd_intr_nesting_level; in stopsig() 133 --gd->gd_intr_nesting_level; in stopsig() 146 globaldata_t gd = mycpu; in kqueuesig() local 153 ++gd->gd_intr_nesting_level; in kqueuesig() 157 --gd->gd_intr_nesting_level; in kqueuesig() [all …]
|
H A D | mp.c | 389 struct mdglobaldata *gd; in start_all_aps() local 422 bzero(gd, sizeof(*gd)); in start_all_aps() 423 gd->mi.gd_prvspace = ps = &CPU_prvspace[x]; in start_all_aps() 426 mi_gdinit(&gd->mi, x); in start_all_aps() 427 cpu_gdinit(gd, x); in start_all_aps() 434 gd->gd_CADDR1 = ps->CPAGE1; in start_all_aps() 435 gd->gd_CADDR2 = ps->CPAGE2; in start_all_aps() 436 gd->gd_CADDR3 = ps->CPAGE3; in start_all_aps() 437 gd->gd_PADDR1 = (vpte_t *)ps->PPAGE1; in start_all_aps() 441 gd->mi.gd_ipiq = (void *)kmem_alloc(kernel_map, ipiq_size, in start_all_aps() [all …]
|
/dragonfly/test/nvmm/demo/smallkern/ |
H A D | main.c | 176 gd->gd_looffset = (uint64_t)func & 0xffff; in setgate() 177 gd->gd_selector = sel; in setgate() 178 gd->gd_ist = ist; in setgate() 179 gd->gd_type = type; in setgate() 180 gd->gd_dpl = dpl; in setgate() 181 gd->gd_p = 1; in setgate() 182 gd->gd_hioffset = (uint64_t)func >> 16; in setgate() 183 gd->gd_xx1 = 0; in setgate() 185 gd->gd_zero = 0; in setgate() 186 gd->gd_xx2 = 0; in setgate() [all …]
|
/dragonfly/test/debug/ |
H A D | slabinfo.c | 93 struct globaldata gd; in main() local 135 kkread(kd, (long)psary[i], &gd, sizeof(gd)); in main() 137 i, gd.gd_slab.NFreeZones); in main() 138 totalfree += gd.gd_slab.NFreeZones; in main() 142 totalzones += slzonedump(kd, &gd.gd_slab.ZoneAry[j]); in main() 147 totalzones += slzonedump(kd, &gd.gd_slab.FreeZones); in main() 151 totalzones += slzonedump(kd, &gd.gd_slab.FreeOvZones); in main()
|
/dragonfly/sys/ddb/ |
H A D | db_ps.c | 105 struct globaldata *gd = globaldata_find(cpuidx); in db_ps() local 112 gd->gd_cpuid, gd->gd_curthread, gd->gd_reqflags); in db_ps() 113 if (gd->gd_curthread && gd->gd_curthread->td_preempted) { in db_ps() 115 gd->gd_curthread->td_preempted); in db_ps() 134 TAILQ_FOREACH(td, &gd->gd_tdrunq, td_threadq) { in db_ps() 157 TAILQ_FOREACH(td, &gd->gd_tdallq, td_allq) { in db_ps()
|