/dragonfly/sys/cpu/x86_64/misc/ |
H A D | cputimer_tsc.c | 88 tsc = rdtsc(); in tsc_cputimer_count() 111 return (rdtsc()); in tsc_cpucounter_count_lfence() 118 return (rdtsc()); in tsc_cpucounter_count_mfence()
|
/dragonfly/test/sysperf/ |
H A D | randread.c | 109 tsc2 = rdtsc(); in main() 130 tsc2 = rdtsc(); in main() 143 tsc2 = rdtsc(); in main() 175 tsc2 = rdtsc(); in main()
|
/dragonfly/sys/sys/ |
H A D | indefinite2.h | 94 info->base = rdtsc(); in indefinite_check() 109 delta = rdtsc() - info->base; in indefinite_check() 205 delta = rdtsc() - info->base; in indefinite_done()
|
H A D | microtime_pcpu.h | 62 t->tsc = rdtsc(); in microtime_pcpu_get()
|
/dragonfly/sys/platform/pc64/x86_64/ |
H A D | mp_machdep.c | 590 tsc0_offset = rdtsc(); in start_all_aps() 599 tsc0_offset = rdtsc(); in start_all_aps() 839 ltsc = rdtsc(); in smitest() 841 ntsc = rdtsc(); /* force loop to occur */ in smitest() 945 tsc_uclock_t tsc_base = rdtsc(); in smp_invltlb() 1021 if (tsc_frequency && rdtsc() - tsc_base > tsc_frequency) { in smp_invltlb() 1051 tsc_base = rdtsc(); in smp_invltlb() 1176 tsc_uclock_t tsc_base = rdtsc(); in smp_inval_intr() 1268 tsc_base = rdtsc(); in smp_inval_intr() 1466 tsc_offsets[mycpu->gd_cpuid] = rdtsc() - tsc0_offset; in ap_init() [all …]
|
H A D | pmap_inval.c | 168 tsc = rdtsc(); in loopwdog() 363 info->tsc_target = rdtsc() + (tsc_frequency * LOOPRECOVER_TIMEOUT1); in pmap_inval_smp() 512 info->tsc_target = rdtsc() + (tsc_frequency * LOOPRECOVER_TIMEOUT1); in pmap_inval_smp_cmpset()
|
H A D | trap.c | 1282 tsc_uclock_t tscval = rdtsc(); in syscall2() 1286 tscval = rdtsc() - tscval; in syscall2()
|
H A D | machdep.c | 3630 tsc = rdtsc() - last_tsc[gd->gd_cpuid]; in pcpu_timer_always() 3633 last_tsc[gd->gd_cpuid] = rdtsc(); in pcpu_timer_always()
|
/dragonfly/sys/platform/pc64/isa/ |
H A D | clock.c | 602 old_tsc = rdtsc(); in calibrate_clocks() 634 tsc_frequency = rdtsc() - old_tsc; in calibrate_clocks() 841 new_tsc = rdtsc(); in do_calibrate_cputimer() 1343 return (rdtsc()); 1357 tsc[i] = (u_int32_t)rdtsc(); in _TSTMP() 1383 tscval = rdtsc(); in hw_i8254_timestamp() 1583 last = rdtsc(); in tsc_mpsync_test() 1586 next = rdtsc(); in tsc_mpsync_test() 1605 last = rdtsc(); in tsc_mpsync_test() 1608 next = rdtsc(); in tsc_mpsync_test()
|
/dragonfly/sys/platform/pc64/apic/ |
H A D | lapic.c | 463 uint64_t val = rdtsc() + diff; in lapic_timer_tscdeadline_quick() 682 now = rdtsc(); in lapic_timer_tscdlt_reload() 888 tsc = rdtsc(); in lapic_mem_icr_unpend() 891 if ((tsc_sclock_t)(rdtsc() - in lapic_mem_icr_unpend() 893 tsc = rdtsc(); in lapic_mem_icr_unpend() 998 val += rdtsc(); in set_apic_timer() 1030 now = rdtsc(); in read_apic_timer()
|
/dragonfly/sys/kern/ |
H A D | lwkt_thread.c | 573 uint64_t tsc_base = rdtsc(); in lwkt_switch() 715 if (tsc_frequency && rdtsc() - tsc_base > tsc_frequency) { in lwkt_switch() 723 tsc_base = rdtsc(); in lwkt_switch() 843 uint64_t tsc_base = rdtsc(); in lwkt_switch_return() 903 if (tsc_frequency && rdtsc() - tsc_base > tsc_frequency) { in lwkt_switch_return() 906 tsc_base = rdtsc(); in lwkt_switch_return() 1388 uint64_t tsc_base = rdtsc(); in lwkt_acquire() 1401 if (tsc_frequency && rdtsc() - tsc_base > tsc_frequency) { in lwkt_acquire() 1404 tsc_base = rdtsc(); in lwkt_acquire()
|
H A D | kern_ktr.c | 374 ktr_sync_tsc = rdtsc(); in ktr_resync_callback() 399 ktr_sync_tsc = rdtsc(); in ktr_resync_remote() 401 delta = rdtsc() - ktr_sync_tsc; in ktr_resync_remote() 515 entry->ktr_timestamp = rdtsc() - tsc_offsets[cpu]; in ktr_begin_write_entry()
|
H A D | kern_spinlock.c | 218 if ((rdtsc() >> spin_window_shift) % ncpus != mycpuid) { in _spin_lock_contested() 329 (rdtsc() >> spin_window_shift) % ncpus == mycpuid) { in _spin_lock_shared_contested()
|
H A D | lwkt_ipiq.c | 234 uint64_t tsc_base = rdtsc(); in lwkt_send_ipiq3() 258 if (rdtsc() - tsc_base > tsc_frequency) { in lwkt_send_ipiq3() 274 tsc_base = rdtsc(); in lwkt_send_ipiq3()
|
H A D | kern_nrandom.c | 595 ((int)rdtsc() % 151); in add_interrupt_randomness() 956 nsec ^= (rdtsc() & 255) << 8; in NANOUP_EVENT()
|
H A D | lwkt_token.c | 400 if ((rdtsc() >> token_window_shift) % ncpus != mycpuid) { in _lwkt_trytokref_spin() 413 if ((rdtsc() >> token_window_shift) % ncpus == mycpuid) { in _lwkt_trytokref_spin()
|
H A D | kern_clock.c | 1815 return (rdtsc() + tsc_frequency * ns / (int64_t)1000000000); in tsc_get_target() 1835 if ((int64_t)(target - rdtsc()) <= 0) in tsc_test_target()
|
/dragonfly/sys/cpu/x86_64/include/ |
H A D | cpufunc.h | 545 rdtsc(void) in rdtsc() function 564 return rdtsc(); in rdtsc_ordered() 971 tsc_uclock_t rdtsc(void);
|
/dragonfly/sys/dev/acpica/ |
H A D | acpi_hpet.c | 217 old_tsc = rdtsc(); in acpi_hpet_early_init() 220 new_tsc = rdtsc(); in acpi_hpet_early_init()
|
/dragonfly/sys/dev/virtual/hyperv/x86_64/ |
H A D | hyperv_machdep.c | 109 tsc = rdtsc(); \
|
/dragonfly/sys/dev/misc/kbd/ |
H A D | atkbdc.c | 316 tscval[0] = rdtsc(); in atkbdc_setup() 318 tscval[1] = rdtsc(); in atkbdc_setup() 320 tscval[2] = rdtsc(); in atkbdc_setup()
|
/dragonfly/sys/net/altq/ |
H A D | altq_subr.c | 877 val = rdtsc(); in read_machclk()
|
/dragonfly/sys/dev/virtual/nvmm/x86/ |
H A D | nvmm_x86_svm.c | 1224 cpudata->gtsc_offset = exit->u.wrmsr.val - rdtsc(); in svm_inkernel_handle_msr() 1938 state->msrs[NVMM_X64_MSR_TSC] - rdtsc(); in svm_vcpu_setstate() 2062 state->msrs[NVMM_X64_MSR_TSC] = rdtsc() + cpudata->gtsc_offset; in svm_vcpu_getstate()
|
H A D | nvmm_x86_vmx.c | 1904 cpudata->gtsc_offset = exit->u.wrmsr.val - rdtsc(); in vmx_inkernel_handle_msr() 2702 state->msrs[NVMM_X64_MSR_TSC] - rdtsc(); in vmx_vcpu_setstate() 2834 state->msrs[NVMM_X64_MSR_TSC] = rdtsc() + cpudata->gtsc_offset; in vmx_vcpu_getstate()
|
/dragonfly/contrib/gcc-4.7/gcc/doc/ |
H A D | md.texi | 2178 For example on i386 the following implements @code{rdtsc}: 2181 unsigned long long rdtsc (void) 2184 __asm__ __volatile__("rdtsc":"=A"(tick)); 2193 unsigned long long rdtsc (void) 2196 __asm__ __volatile__("rdtsc":"=a"(tickl),"=d"(tickh));
|