/netbsd/sys/kern/ |
H A D | subr_thmap.c | 271 s = atomic_load_relaxed(&node->state); in lock_node() 430 atomic_load_relaxed(&node->state) + 1); in node_insert() 440 ASSERT(NODE_COUNT(atomic_load_relaxed(&node->state)) > 0); in node_remove() 446 atomic_load_relaxed(&node->state) - 1); in node_remove() 532 if (atomic_load_relaxed(&thmap->root[i])) { in root_try_put() 549 if (atomic_load_relaxed(&thmap->root[i])) { in root_try_put() 601 if (atomic_load_relaxed(&parent->state) & NODE_DELETED) { in find_edge_node() 643 target = atomic_load_relaxed(&node->slots[*slot]); in find_edge_node_locked() 876 atomic_load_relaxed(&node->state) | NODE_DELETED); in thmap_del() 897 atomic_load_relaxed(&thmap->root[rslot]); in thmap_del() [all …]
|
H A D | kern_heartbeat.c | 354 unsigned count = atomic_load_relaxed(&curcpu()->ci_heartbeat_count); in heartbeat_intr() 455 if (atomic_load_relaxed(&ack) || in defibrillate() 456 atomic_load_relaxed(&panicstr) != NULL) in defibrillate() 564 period_ticks = atomic_load_relaxed(&heartbeat_max_period_ticks); in heartbeat() 565 period_secs = atomic_load_relaxed(&heartbeat_max_period_secs); in heartbeat() 583 cache = atomic_load_relaxed(&curcpu()->ci_heartbeat_uptime_cache); in heartbeat() 596 atomic_load_relaxed(&curcpu()->ci_heartbeat_uptime_stamp); in heartbeat() 650 d = uptime - atomic_load_relaxed(&patient->ci_heartbeat_uptime_cache); in heartbeat()
|
H A D | subr_fault.c | 89 if (atomic_load_relaxed(&f->oneshot)) { in fault_inject() 90 if (__predict_true(atomic_load_relaxed(&f->nfaults) > 0)) in fault_inject() 95 if (__predict_false(cnt % atomic_load_relaxed(&f->nth) == 0)) { in fault_inject() 200 args->nfaults = atomic_load_relaxed(&fault_global.nfaults); in fault_ioc_getinfo() 206 args->nfaults = atomic_load_relaxed(&f->nfaults); in fault_ioc_getinfo()
|
H A D | subr_pcq.c | 187 v = atomic_load_relaxed(&pcq->pcq_pc); in pcq_put() 219 const uint32_t v = atomic_load_relaxed(&pcq->pcq_pc); in pcq_peek() 240 v = atomic_load_relaxed(&pcq->pcq_pc); in pcq_get() 278 v = atomic_load_relaxed(&pcq->pcq_pc); in pcq_get()
|
H A D | kern_entropy.c | 727 return atomic_load_relaxed(&E->epoch); in entropy_epoch() 739 return atomic_load_relaxed(&E->bitsneeded) == 0; in entropy_ready() 1035 sih = atomic_load_relaxed(&entropy_sih); in entropy_enter_intr() 1061 atomic_load_relaxed(&entropy_depletion)) && in entropy_enter_intr() 1063 sih = atomic_load_relaxed(&entropy_sih); in entropy_enter_intr() 1682 if (atomic_load_relaxed(&entropy_depletion)) in filt_entropy_read_event() 2107 flags = atomic_load_relaxed(&rs->flags); in rnd_add_data() 2284 urs->flags = atomic_load_relaxed(&rs->flags); in rndsource_to_user() 2321 atomic_load_relaxed(&rc->rc_timesamples)); in rndsource_to_user_est_cpu() 2323 atomic_load_relaxed(&rc->rc_datasamples)); in rndsource_to_user_est_cpu() [all …]
|
H A D | kern_tc.c | 220 hi = atomic_load_relaxed(&time__second32.hi); in getrealtime() 226 lo = atomic_load_relaxed(&time__second32.lo); in getrealtime() 228 } while (hi != atomic_load_relaxed(&time__second32.hi)); in getrealtime() 240 hi = atomic_load_relaxed(&time__uptime32.hi); in getuptime() 246 lo = atomic_load_relaxed(&time__uptime32.lo); in getuptime() 248 } while (hi != atomic_load_relaxed(&time__uptime32.hi)); in getuptime() 264 return atomic_load_relaxed(&time__uptime32.lo); in getuptime32() 688 while ((gen = atomic_load_relaxed(&timebase.gen)) & 1) in getbinboottime() 697 } while (gen != atomic_load_relaxed(&timebase.gen)); in getbinboottime()
|
/netbsd/sys/net/npf/ |
H A D | npf_conn.c | 221 KASSERT(atomic_load_relaxed(&npf->conn_tracking) in npf_conn_load() 223 odb = atomic_load_relaxed(&npf->conn_db); in npf_conn_load() 532 KASSERT(atomic_load_relaxed(&con->c_refcnt) == 0); in npf_conn_destroy() 576 KASSERT(atomic_load_relaxed(&con->c_refcnt) > 0); in npf_conn_setnat() 583 flags = atomic_load_relaxed(&con->c_flags); in npf_conn_setnat() 649 KASSERT(atomic_load_relaxed(&con->c_refcnt) > 0); in npf_conn_pass() 651 mi->mi_retfl = atomic_load_relaxed(&con->c_retfl); in npf_conn_pass() 667 KASSERT(atomic_load_relaxed(&con->c_refcnt) > 0); in npf_conn_setpass() 696 KASSERT(atomic_load_relaxed(&con->c_refcnt) > 0); in npf_conn_release() 787 conn_db = atomic_load_relaxed(&npf->conn_db); in npf_conndb_export() [all …]
|
H A D | npf_alg.c | 230 count = atomic_load_relaxed(&aset->alg_count); in npf_alg_match() 235 match_func = atomic_load_relaxed(&f->match); in npf_alg_match() 263 count = atomic_load_relaxed(&aset->alg_count); in npf_alg_exec() 268 translate_func = atomic_load_relaxed(&f->translate); in npf_alg_exec() 303 count = atomic_load_relaxed(&aset->alg_count); in npf_alg_conn() 308 inspect_func = atomic_load_relaxed(&f->inspect); in npf_alg_conn() 327 if ((destroy_func = atomic_load_relaxed(&f->destroy)) != NULL) { in npf_alg_destroy()
|
H A D | npf_rproc.c | 158 if (atomic_load_relaxed(&ext->ext_refcnt)) { in npf_ext_unregister() 163 if (atomic_load_relaxed(&ext->ext_refcnt)) { in npf_ext_unregister() 331 KASSERT(atomic_load_relaxed(&rp->rp_refcnt) > 0); in npf_rproc_release() 372 KASSERT(atomic_load_relaxed(&rp->rp_refcnt) > 0); in npf_rproc_run() 378 KASSERT(atomic_load_relaxed(&ext->ext_refcnt) > 0); in npf_rproc_run()
|
H A D | npf_conndb.c | 166 npf_conndb_t *cd = atomic_load_relaxed(&npf->conn_db); in npf_conndb_lookup() 251 head = atomic_load_relaxed(&cd->cd_new); in npf_conndb_enqueue() 268 npf_conn_t *next = atomic_load_relaxed(&con->c_next); // union in npf_conndb_update() 443 const unsigned refcnt = atomic_load_relaxed(&con->c_refcnt); in npf_conndb_gc()
|
/netbsd/sys/external/bsd/common/linux/ |
H A D | linux_tasklet.c | 227 KASSERT(atomic_load_relaxed(&tasklet->tl_state) & in tasklet_softintr() 261 KASSERT(atomic_load_relaxed(&tasklet->tl_state) & in tasklet_softintr() 286 ostate = atomic_load_relaxed(&tasklet->tl_state); in tasklet_queue_schedule() 312 KASSERT(atomic_load_relaxed(&tasklet->tl_state) & TASKLET_SCHEDULED); in tasklet_queue_enqueue() 459 while (atomic_load_relaxed(&tasklet->tl_state) & TASKLET_SCHEDULED) in tasklet_kill() 492 return atomic_load_relaxed(&tasklet->tl_state) & TASKLET_RUNNING; in tasklet_is_locked() 509 state = atomic_load_relaxed(&tasklet->tl_state); in tasklet_trylock() 532 KASSERT(atomic_load_relaxed(&tasklet->tl_state) & TASKLET_RUNNING); in tasklet_unlock() 635 disablecount = atomic_load_relaxed(&tasklet->tl_disablecount); in __tasklet_is_enabled() 651 return atomic_load_relaxed(&tasklet->tl_state) & TASKLET_SCHEDULED; in __tasklet_is_scheduled()
|
/netbsd/sys/external/bsd/drm2/include/linux/ |
H A D | kref.h | 69 count = atomic_load_relaxed(&kref->kr_count); in kref_get_unless_zero() 86 old = atomic_load_relaxed(&kref->kr_count); in kref_sub() 110 old = atomic_load_relaxed(&kref->kr_count); in kref_put_lock() 144 old = atomic_load_relaxed(&kref->kr_count); in kref_put_mutex() 166 return atomic_load_relaxed(&kref->kr_count); in kref_read()
|
/netbsd/external/mpl/bind/dist/lib/isc/ |
H A D | quota.c | 70 return (atomic_load_relaxed("a->max)); in isc_quota_getmax() 76 return (atomic_load_relaxed("a->soft)); in isc_quota_getsoft() 82 return (atomic_load_relaxed("a->used)); in isc_quota_getused() 135 if (atomic_load_relaxed("a->waiting) > 0) { in quota_release()
|
/netbsd/external/gpl3/gcc/dist/libsanitizer/tsan/ |
H A D | tsan_sync.h | 86 return atomic_load_relaxed(&flags) & f; in IsFlagSet() 90 atomic_store_relaxed(&flags, atomic_load_relaxed(&flags) | f); in SetFlags() 97 u32 current = atomic_load_relaxed(&flags); in UpdateFlags()
|
/netbsd/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/ |
H A D | tsan_sync.h | 88 return atomic_load_relaxed(&flags) & f; in IsFlagSet() 92 atomic_store_relaxed(&flags, atomic_load_relaxed(&flags) | f); in SetFlags() 99 u32 current = atomic_load_relaxed(&flags); in UpdateFlags()
|
/netbsd/external/gpl3/gcc.old/dist/libsanitizer/tsan/ |
H A D | tsan_sync.h | 86 return atomic_load_relaxed(&flags) & f; in IsFlagSet() 90 atomic_store_relaxed(&flags, atomic_load_relaxed(&flags) | f); in SetFlags() 97 u32 current = atomic_load_relaxed(&flags); in UpdateFlags()
|
/netbsd/external/mpl/bind/dist/lib/dns/ |
H A D | badcache.c | 141 if (atomic_load_relaxed(&bc->count) > bc->size * 8) { in badcache_resize() 143 } else if (atomic_load_relaxed(&bc->count) < bc->size * 2 && in badcache_resize() 186 for (i = 0; atomic_load_relaxed(&bc->count) > 0 && i < bc->size; i++) { in badcache_resize() 314 if (atomic_load_relaxed(&bc->count) == 0) { in dns_badcache_find() 377 for (i = 0; atomic_load_relaxed(&bc->count) > 0 && i < bc->size; i++) { in dns_badcache_flush() 453 for (i = 0; atomic_load_relaxed(&bc->count) > 0 && i < bc->size; i++) { in dns_badcache_flushtree() 498 for (i = 0; atomic_load_relaxed(&bc->count) > 0 && i < bc->size; i++) { in dns_badcache_print()
|
/netbsd/sys/external/bsd/drm2/linux/ |
H A D | linux_dma_resv.c | 484 shared_count = list ? atomic_load_relaxed(&list->shared_count) : 0; in dma_resv_get_shared_reader() 842 shared[i] = atomic_load_relaxed(&list->shared[i]); in dma_resv_get_fences_rcu() 858 if (dma_fence_get_rcu(atomic_load_relaxed(&shared[i])) == NULL) in dma_resv_get_fences_rcu() 951 fence = atomic_load_relaxed(&src_list->shared[i]); in dma_resv_copy_fences() 1077 fence = atomic_load_relaxed(&list->shared[i]); in dma_resv_test_signaled_rcu() 1167 fence = atomic_load_relaxed(&list->shared[i]); in dma_resv_wait_timeout_rcu() 1333 fence = atomic_load_relaxed(&list->shared[i]); in dma_resv_do_poll() 1372 fence = atomic_load_relaxed(&list->shared[i]); in dma_resv_do_poll()
|
/netbsd/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/ |
H A D | sanitizer_quarantine.h | 96 uptr GetSize() const { return atomic_load_relaxed(&max_size_); } in GetSize() 98 return atomic_load_relaxed(&max_cache_size_); in GetCacheSize() 120 Recycle(atomic_load_relaxed(&min_size_), cb); in Drain() 212 return atomic_load_relaxed(&size_); in Size()
|
/netbsd/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/ |
H A D | sanitizer_quarantine.h | 98 uptr GetSize() const { return atomic_load_relaxed(&max_size_); } in GetSize() 100 return atomic_load_relaxed(&max_cache_size_); in GetCacheSize() 122 Recycle(atomic_load_relaxed(&min_size_), cb); in Drain() 214 return atomic_load_relaxed(&size_); in Size()
|
/netbsd/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/ |
H A D | sanitizer_quarantine.h | 96 uptr GetSize() const { return atomic_load_relaxed(&max_size_); } in GetSize() 98 return atomic_load_relaxed(&max_cache_size_); in GetCacheSize() 120 Recycle(atomic_load_relaxed(&min_size_), cb); in Drain() 212 return atomic_load_relaxed(&size_); in Size()
|
/netbsd/sys/arch/x86/x86/ |
H A D | x86_tlb.c | 95 #define TP_GET_DONE(tp) (atomic_load_relaxed(&(tp)->tp_store[TP_DONE]) & 1) 106 uintptr_t v = atomic_load_relaxed(&(tp)->tp_store[TP_DONE]); \ 415 KASSERT(atomic_load_relaxed(&pmap_tlb_packet) != ts); in pmap_tlb_shootnow() 425 } while (atomic_load_relaxed(&pmap_tlb_packet) != NULL); in pmap_tlb_shootnow()
|
/netbsd/sys/dev/pci/ixgbe/ |
H A D | ixgbe_netbsd.h | 61 atomic_load_relaxed(&((evp)->ev_count)) 66 atomic_load_relaxed(&((evp)->ev_count)) + (val))
|
/netbsd/sys/external/bsd/compiler_rt/dist/lib/ubsan_minimal/ |
H A D | ubsan_minimal_handlers.cc | 27 unsigned sz = __sanitizer::atomic_load_relaxed(&caller_pcs_sz); in report_this_error() 34 p = __sanitizer::atomic_load_relaxed(&caller_pcs[i]); in report_this_error()
|
/netbsd/sys/external/bsd/compiler_rt/dist/lib/xray/ |
H A D | xray_profiling.cc | 86 if (atomic_load_relaxed(&ThreadExitingLatch)) in getThreadLocalData() 143 atomic_load_relaxed(&TLD.Allocators))); in getThreadLocalData() 295 auto FCT = reinterpret_cast<FunctionCallTrie *>(atomic_load_relaxed(&T->FCT)); in profilingHandleArg0() 439 if (atomic_load_relaxed(&T->Allocators) == 0) in profilingLoggingInit()
|