/dragonfly/sys/sys/ |
H A D | spinlock2.h | 72 crit_enter_raw(gd->gd_curthread); in spin_trylock() 81 gd->gd_curthread->td_spinlock_stack_id[i] = 1; in spin_trylock() 82 gd->gd_curthread->td_spinlock_stack[i] = spin; in spin_trylock() 83 gd->gd_curthread->td_spinlock_caller_pc[i] = in spin_trylock() 111 crit_enter_raw(gd->gd_curthread); in _spin_lock_quick() 123 gd->gd_curthread->td_spinlock_stack_id[i] = 1; in _spin_lock_quick() 125 gd->gd_curthread->td_spinlock_caller_pc[i] = in _spin_lock_quick() 177 crit_exit_quick(gd->gd_curthread); in spin_unlock_quick() 208 crit_enter_raw(gd->gd_curthread); in _spin_lock_shared_quick() 225 gd->gd_curthread->td_spinlock_caller_pc[i] = in _spin_lock_shared_quick() [all …]
|
H A D | mutex2.h | 225 crit_enter_quick(gd->gd_curthread); in mtx_spinlock() 238 mtx->mtx_owner = gd->gd_curthread; in mtx_spinlock() 249 crit_enter_quick(gd->gd_curthread); in mtx_spinlock_try() 259 mtx->mtx_owner = gd->gd_curthread; in mtx_spinlock_try() 304 mtx->mtx_owner == gd->gd_curthread); in mtx_downgrade() 349 mtx->mtx_owner == gd->gd_curthread); in mtx_unlock() 369 mtx->mtx_owner == gd->gd_curthread); in mtx_unlock_ex() 398 crit_exit_quick(gd->gd_curthread); in mtx_spinunlock()
|
H A D | thread2.h | 191 _crit_enter_quick(gd->gd_curthread __DEBUG_CRIT_PASS_ARG__); in _crit_enter() 197 _crit_enter_quick(gd->gd_curthread __DEBUG_CRIT_PASS_ARG__); in _crit_enter_hard() 237 _crit_exit_quick(gd->gd_curthread __DEBUG_CRIT_PASS_ARG__); in _crit_exit() 245 _crit_exit_quick(gd->gd_curthread __DEBUG_CRIT_PASS_ARG__); in _crit_exit_hard()
|
H A D | globaldata.h | 131 struct thread *gd_curthread; member
|
/dragonfly/sys/ddb/ |
H A D | db_ps.c | 112 gd->gd_cpuid, gd->gd_curthread, gd->gd_reqflags); in db_ps() 113 if (gd->gd_curthread && gd->gd_curthread->td_preempted) { in db_ps() 115 gd->gd_curthread->td_preempted); in db_ps()
|
/dragonfly/sys/platform/vkernel64/x86_64/ |
H A D | exception.c | 79 thread_t td = gd->gd_curthread; in ipisig() 113 thread_t td = gd->gd_curthread; in stopsig() 147 thread_t td = gd->gd_curthread; in kqueuesig() 170 thread_t td = gd->gd_curthread; in timersig()
|
H A D | global.s | 51 .globl gd_curthread, gd_npxthread, gd_reqflags, gd_common_tss 52 .set gd_curthread,globaldata + GD_CURTHREAD define
|
H A D | genassym.c | 77 ASSYM(GD_CURTHREAD, offsetof(struct mdglobaldata, mi.gd_curthread));
|
/dragonfly/sys/platform/pc64/x86_64/ |
H A D | global.s | 53 .globl gd_curthread, gd_npxthread, gd_reqflags, gd_common_tss 57 .set gd_curthread,globaldata + GD_CURTHREAD define
|
H A D | tls.c | 160 thread_t td = gd->mi.gd_curthread; in set_user_TLS()
|
H A D | mp_flame.c | 121 td = gd->gd_curthread; in hard_sniff()
|
/dragonfly/sys/kern/ |
H A D | lwkt_ipiq.c | 585 if (gd->gd_curthread->td_cscount == 0) in lwkt_process_ipiq() 632 if (gd->gd_curthread->td_cscount == 0) in lwkt_process_ipiq_frame() 692 if (gd->gd_curthread->td_cscount == 0) in lwkt_process_ipiq_nested() 863 ++gd->gd_curthread->td_cscount; in lwkt_cpusync_interlock() 919 --gd->gd_curthread->td_cscount; in lwkt_cpusync_deinterlock() 952 ++gd->gd_curthread->td_cscount; in lwkt_cpusync_quick() 969 --gd->gd_curthread->td_cscount; in lwkt_cpusync_quick()
|
H A D | kern_intr.c | 878 TD_INVARIANTS_GET(gd->gd_curthread); in ithread_handler() 907 TD_INVARIANTS_TEST(gd->gd_curthread, rec->name); in ithread_handler() 963 lwkt_deschedule_self(gd->gd_curthread); in ithread_handler() 989 lwkt_deschedule_self(gd->gd_curthread); in ithread_handler() 1042 TD_INVARIANTS_GET(gd->gd_curthread); in ithread_emergency() 1057 TD_INVARIANTS_TEST(gd->gd_curthread, rec->name); in ithread_emergency() 1061 lwkt_deschedule_self(gd->gd_curthread); in ithread_emergency()
|
H A D | lwkt_thread.c | 568 thread_t td = gd->gd_curthread; in lwkt_switch() 960 td = gd->gd_curthread; in lwkt_preempt() 1071 thread_t td = gd->gd_curthread; in splz_check() 1151 thread_t td = gd->gd_curthread; in lwkt_yield() 1188 thread_t td = gd->gd_curthread; in lwkt_yield_quick() 1219 thread_t td = gd->gd_curthread; in lwkt_user_yield() 1298 if (td == mygd->gd_curthread) { in _lwkt_schedule()
|
H A D | kern_synch.c | 388 thread_t td = gd->gd_curthread; in _tsleep_interlock() 1151 thread_t td = gd->gd_curthread; in wakeup() 1271 gd->gd_curthread->td_flags |= TDF_DELAYED_WAKEUP; in wakeup_start_delayed() 1280 if (gd->gd_curthread->td_flags & TDF_DELAYED_WAKEUP) { in wakeup_end_delayed() 1282 gd->gd_curthread->td_flags &= ~TDF_DELAYED_WAKEUP; in wakeup_end_delayed()
|
H A D | kern_mutex.c | 434 mtx->mtx_owner = gd->gd_curthread; in _mtx_spinlock_try() 438 mtx->mtx_owner == gd->gd_curthread) { in _mtx_spinlock_try() 446 crit_exit_quick(gd->gd_curthread); in _mtx_spinlock_try()
|
H A D | kern_spinlock.c | 137 crit_exit_quick(gd->gd_curthread); in spin_trylock_contested()
|
H A D | usched_dummy.c | 481 lwkt_deschedule_self(gd->gd_curthread); /* interlock */ in dummy_sched_thread()
|
/dragonfly/sys/platform/vkernel64/platform/ |
H A D | machintr.c | 125 thread_t td = gd->mi.gd_curthread; in splz() 172 thread_t td = gd->mi.gd_curthread; in signalintr()
|
H A D | globaldata.c | 69 gd->mi.gd_curthread = &gd->mi.gd_idlethread; in cpu_gdinit()
|
/dragonfly/sys/contrib/dev/acpica/source/components/parser/ |
H A D | psloop.c | 367 #define TDTRACKER mycpu->gd_curthread->td_tracker 368 #define TDTRACKER_INC ++mycpu->gd_curthread->td_tracker 369 #define TDTRACKER_DEC --mycpu->gd_curthread->td_tracker
|
/dragonfly/sys/platform/pc64/include/ |
H A D | thread.h | 80 #define curthread mycpu->gd_curthread
|
/dragonfly/sys/platform/vkernel64/include/ |
H A D | thread.h | 90 #define curthread mycpu->gd_curthread
|
/dragonfly/sys/dev/acpica/Osd/ |
H A D | OsdSynch.c | 383 if (gd->gd_curthread == &gd->gd_idlethread) { in _AcpiOsAcquireLock() 423 if (gd->gd_curthread == &gd->gd_idlethread) { in AcpiOsReleaseLock()
|
/dragonfly/gnu/usr.bin/gdb/kgdb/ |
H A D | kthr.c | 127 dumptid = (CORE_ADDR)gd.mi.gd_curthread; in kgdb_thr_init()
|