Home
last modified time | relevance | path

Searched refs:atomic_load_relaxed (Results 1 – 25 of 173) sorted by relevance

1234567

/netbsd/sys/kern/
H A Dsubr_thmap.c271 s = atomic_load_relaxed(&node->state); in lock_node()
430 atomic_load_relaxed(&node->state) + 1); in node_insert()
440 ASSERT(NODE_COUNT(atomic_load_relaxed(&node->state)) > 0); in node_remove()
446 atomic_load_relaxed(&node->state) - 1); in node_remove()
532 if (atomic_load_relaxed(&thmap->root[i])) { in root_try_put()
549 if (atomic_load_relaxed(&thmap->root[i])) { in root_try_put()
601 if (atomic_load_relaxed(&parent->state) & NODE_DELETED) { in find_edge_node()
643 target = atomic_load_relaxed(&node->slots[*slot]); in find_edge_node_locked()
876 atomic_load_relaxed(&node->state) | NODE_DELETED); in thmap_del()
897 atomic_load_relaxed(&thmap->root[rslot]); in thmap_del()
[all …]
H A Dkern_heartbeat.c354 unsigned count = atomic_load_relaxed(&curcpu()->ci_heartbeat_count); in heartbeat_intr()
455 if (atomic_load_relaxed(&ack) || in defibrillate()
456 atomic_load_relaxed(&panicstr) != NULL) in defibrillate()
564 period_ticks = atomic_load_relaxed(&heartbeat_max_period_ticks); in heartbeat()
565 period_secs = atomic_load_relaxed(&heartbeat_max_period_secs); in heartbeat()
583 cache = atomic_load_relaxed(&curcpu()->ci_heartbeat_uptime_cache); in heartbeat()
596 atomic_load_relaxed(&curcpu()->ci_heartbeat_uptime_stamp); in heartbeat()
650 d = uptime - atomic_load_relaxed(&patient->ci_heartbeat_uptime_cache); in heartbeat()
H A Dsubr_fault.c89 if (atomic_load_relaxed(&f->oneshot)) { in fault_inject()
90 if (__predict_true(atomic_load_relaxed(&f->nfaults) > 0)) in fault_inject()
95 if (__predict_false(cnt % atomic_load_relaxed(&f->nth) == 0)) { in fault_inject()
200 args->nfaults = atomic_load_relaxed(&fault_global.nfaults); in fault_ioc_getinfo()
206 args->nfaults = atomic_load_relaxed(&f->nfaults); in fault_ioc_getinfo()
H A Dsubr_pcq.c187 v = atomic_load_relaxed(&pcq->pcq_pc); in pcq_put()
219 const uint32_t v = atomic_load_relaxed(&pcq->pcq_pc); in pcq_peek()
240 v = atomic_load_relaxed(&pcq->pcq_pc); in pcq_get()
278 v = atomic_load_relaxed(&pcq->pcq_pc); in pcq_get()
H A Dkern_entropy.c727 return atomic_load_relaxed(&E->epoch); in entropy_epoch()
739 return atomic_load_relaxed(&E->bitsneeded) == 0; in entropy_ready()
1035 sih = atomic_load_relaxed(&entropy_sih); in entropy_enter_intr()
1061 atomic_load_relaxed(&entropy_depletion)) && in entropy_enter_intr()
1063 sih = atomic_load_relaxed(&entropy_sih); in entropy_enter_intr()
1682 if (atomic_load_relaxed(&entropy_depletion)) in filt_entropy_read_event()
2107 flags = atomic_load_relaxed(&rs->flags); in rnd_add_data()
2284 urs->flags = atomic_load_relaxed(&rs->flags); in rndsource_to_user()
2321 atomic_load_relaxed(&rc->rc_timesamples)); in rndsource_to_user_est_cpu()
2323 atomic_load_relaxed(&rc->rc_datasamples)); in rndsource_to_user_est_cpu()
[all …]
H A Dkern_tc.c220 hi = atomic_load_relaxed(&time__second32.hi); in getrealtime()
226 lo = atomic_load_relaxed(&time__second32.lo); in getrealtime()
228 } while (hi != atomic_load_relaxed(&time__second32.hi)); in getrealtime()
240 hi = atomic_load_relaxed(&time__uptime32.hi); in getuptime()
246 lo = atomic_load_relaxed(&time__uptime32.lo); in getuptime()
248 } while (hi != atomic_load_relaxed(&time__uptime32.hi)); in getuptime()
264 return atomic_load_relaxed(&time__uptime32.lo); in getuptime32()
688 while ((gen = atomic_load_relaxed(&timebase.gen)) & 1) in getbinboottime()
697 } while (gen != atomic_load_relaxed(&timebase.gen)); in getbinboottime()
/netbsd/sys/net/npf/
H A Dnpf_conn.c221 KASSERT(atomic_load_relaxed(&npf->conn_tracking) in npf_conn_load()
223 odb = atomic_load_relaxed(&npf->conn_db); in npf_conn_load()
532 KASSERT(atomic_load_relaxed(&con->c_refcnt) == 0); in npf_conn_destroy()
576 KASSERT(atomic_load_relaxed(&con->c_refcnt) > 0); in npf_conn_setnat()
583 flags = atomic_load_relaxed(&con->c_flags); in npf_conn_setnat()
649 KASSERT(atomic_load_relaxed(&con->c_refcnt) > 0); in npf_conn_pass()
651 mi->mi_retfl = atomic_load_relaxed(&con->c_retfl); in npf_conn_pass()
667 KASSERT(atomic_load_relaxed(&con->c_refcnt) > 0); in npf_conn_setpass()
696 KASSERT(atomic_load_relaxed(&con->c_refcnt) > 0); in npf_conn_release()
787 conn_db = atomic_load_relaxed(&npf->conn_db); in npf_conndb_export()
[all …]
H A Dnpf_alg.c230 count = atomic_load_relaxed(&aset->alg_count); in npf_alg_match()
235 match_func = atomic_load_relaxed(&f->match); in npf_alg_match()
263 count = atomic_load_relaxed(&aset->alg_count); in npf_alg_exec()
268 translate_func = atomic_load_relaxed(&f->translate); in npf_alg_exec()
303 count = atomic_load_relaxed(&aset->alg_count); in npf_alg_conn()
308 inspect_func = atomic_load_relaxed(&f->inspect); in npf_alg_conn()
327 if ((destroy_func = atomic_load_relaxed(&f->destroy)) != NULL) { in npf_alg_destroy()
H A Dnpf_rproc.c158 if (atomic_load_relaxed(&ext->ext_refcnt)) { in npf_ext_unregister()
163 if (atomic_load_relaxed(&ext->ext_refcnt)) { in npf_ext_unregister()
331 KASSERT(atomic_load_relaxed(&rp->rp_refcnt) > 0); in npf_rproc_release()
372 KASSERT(atomic_load_relaxed(&rp->rp_refcnt) > 0); in npf_rproc_run()
378 KASSERT(atomic_load_relaxed(&ext->ext_refcnt) > 0); in npf_rproc_run()
H A Dnpf_conndb.c166 npf_conndb_t *cd = atomic_load_relaxed(&npf->conn_db); in npf_conndb_lookup()
251 head = atomic_load_relaxed(&cd->cd_new); in npf_conndb_enqueue()
268 npf_conn_t *next = atomic_load_relaxed(&con->c_next); // union in npf_conndb_update()
443 const unsigned refcnt = atomic_load_relaxed(&con->c_refcnt); in npf_conndb_gc()
/netbsd/sys/external/bsd/common/linux/
H A Dlinux_tasklet.c227 KASSERT(atomic_load_relaxed(&tasklet->tl_state) & in tasklet_softintr()
261 KASSERT(atomic_load_relaxed(&tasklet->tl_state) & in tasklet_softintr()
286 ostate = atomic_load_relaxed(&tasklet->tl_state); in tasklet_queue_schedule()
312 KASSERT(atomic_load_relaxed(&tasklet->tl_state) & TASKLET_SCHEDULED); in tasklet_queue_enqueue()
459 while (atomic_load_relaxed(&tasklet->tl_state) & TASKLET_SCHEDULED) in tasklet_kill()
492 return atomic_load_relaxed(&tasklet->tl_state) & TASKLET_RUNNING; in tasklet_is_locked()
509 state = atomic_load_relaxed(&tasklet->tl_state); in tasklet_trylock()
532 KASSERT(atomic_load_relaxed(&tasklet->tl_state) & TASKLET_RUNNING); in tasklet_unlock()
635 disablecount = atomic_load_relaxed(&tasklet->tl_disablecount); in __tasklet_is_enabled()
651 return atomic_load_relaxed(&tasklet->tl_state) & TASKLET_SCHEDULED; in __tasklet_is_scheduled()
/netbsd/sys/external/bsd/drm2/include/linux/
H A Dkref.h69 count = atomic_load_relaxed(&kref->kr_count); in kref_get_unless_zero()
86 old = atomic_load_relaxed(&kref->kr_count); in kref_sub()
110 old = atomic_load_relaxed(&kref->kr_count); in kref_put_lock()
144 old = atomic_load_relaxed(&kref->kr_count); in kref_put_mutex()
166 return atomic_load_relaxed(&kref->kr_count); in kref_read()
/netbsd/external/mpl/bind/dist/lib/isc/
H A Dquota.c70 return (atomic_load_relaxed(&quota->max)); in isc_quota_getmax()
76 return (atomic_load_relaxed(&quota->soft)); in isc_quota_getsoft()
82 return (atomic_load_relaxed(&quota->used)); in isc_quota_getused()
135 if (atomic_load_relaxed(&quota->waiting) > 0) { in quota_release()
/netbsd/external/gpl3/gcc/dist/libsanitizer/tsan/
H A Dtsan_sync.h86 return atomic_load_relaxed(&flags) & f; in IsFlagSet()
90 atomic_store_relaxed(&flags, atomic_load_relaxed(&flags) | f); in SetFlags()
97 u32 current = atomic_load_relaxed(&flags); in UpdateFlags()
/netbsd/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/
H A Dtsan_sync.h88 return atomic_load_relaxed(&flags) & f; in IsFlagSet()
92 atomic_store_relaxed(&flags, atomic_load_relaxed(&flags) | f); in SetFlags()
99 u32 current = atomic_load_relaxed(&flags); in UpdateFlags()
/netbsd/external/gpl3/gcc.old/dist/libsanitizer/tsan/
H A Dtsan_sync.h86 return atomic_load_relaxed(&flags) & f; in IsFlagSet()
90 atomic_store_relaxed(&flags, atomic_load_relaxed(&flags) | f); in SetFlags()
97 u32 current = atomic_load_relaxed(&flags); in UpdateFlags()
/netbsd/external/mpl/bind/dist/lib/dns/
H A Dbadcache.c141 if (atomic_load_relaxed(&bc->count) > bc->size * 8) { in badcache_resize()
143 } else if (atomic_load_relaxed(&bc->count) < bc->size * 2 && in badcache_resize()
186 for (i = 0; atomic_load_relaxed(&bc->count) > 0 && i < bc->size; i++) { in badcache_resize()
314 if (atomic_load_relaxed(&bc->count) == 0) { in dns_badcache_find()
377 for (i = 0; atomic_load_relaxed(&bc->count) > 0 && i < bc->size; i++) { in dns_badcache_flush()
453 for (i = 0; atomic_load_relaxed(&bc->count) > 0 && i < bc->size; i++) { in dns_badcache_flushtree()
498 for (i = 0; atomic_load_relaxed(&bc->count) > 0 && i < bc->size; i++) { in dns_badcache_print()
/netbsd/sys/external/bsd/drm2/linux/
H A Dlinux_dma_resv.c484 shared_count = list ? atomic_load_relaxed(&list->shared_count) : 0; in dma_resv_get_shared_reader()
842 shared[i] = atomic_load_relaxed(&list->shared[i]); in dma_resv_get_fences_rcu()
858 if (dma_fence_get_rcu(atomic_load_relaxed(&shared[i])) == NULL) in dma_resv_get_fences_rcu()
951 fence = atomic_load_relaxed(&src_list->shared[i]); in dma_resv_copy_fences()
1077 fence = atomic_load_relaxed(&list->shared[i]); in dma_resv_test_signaled_rcu()
1167 fence = atomic_load_relaxed(&list->shared[i]); in dma_resv_wait_timeout_rcu()
1333 fence = atomic_load_relaxed(&list->shared[i]); in dma_resv_do_poll()
1372 fence = atomic_load_relaxed(&list->shared[i]); in dma_resv_do_poll()
/netbsd/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/
H A Dsanitizer_quarantine.h96 uptr GetSize() const { return atomic_load_relaxed(&max_size_); } in GetSize()
98 return atomic_load_relaxed(&max_cache_size_); in GetCacheSize()
120 Recycle(atomic_load_relaxed(&min_size_), cb); in Drain()
212 return atomic_load_relaxed(&size_); in Size()
/netbsd/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/
H A Dsanitizer_quarantine.h98 uptr GetSize() const { return atomic_load_relaxed(&max_size_); } in GetSize()
100 return atomic_load_relaxed(&max_cache_size_); in GetCacheSize()
122 Recycle(atomic_load_relaxed(&min_size_), cb); in Drain()
214 return atomic_load_relaxed(&size_); in Size()
/netbsd/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/
H A Dsanitizer_quarantine.h96 uptr GetSize() const { return atomic_load_relaxed(&max_size_); } in GetSize()
98 return atomic_load_relaxed(&max_cache_size_); in GetCacheSize()
120 Recycle(atomic_load_relaxed(&min_size_), cb); in Drain()
212 return atomic_load_relaxed(&size_); in Size()
/netbsd/sys/arch/x86/x86/
H A Dx86_tlb.c95 #define TP_GET_DONE(tp) (atomic_load_relaxed(&(tp)->tp_store[TP_DONE]) & 1)
106 uintptr_t v = atomic_load_relaxed(&(tp)->tp_store[TP_DONE]); \
415 KASSERT(atomic_load_relaxed(&pmap_tlb_packet) != ts); in pmap_tlb_shootnow()
425 } while (atomic_load_relaxed(&pmap_tlb_packet) != NULL); in pmap_tlb_shootnow()
/netbsd/sys/dev/pci/ixgbe/
H A Dixgbe_netbsd.h61 atomic_load_relaxed(&((evp)->ev_count))
66 atomic_load_relaxed(&((evp)->ev_count)) + (val))
/netbsd/sys/external/bsd/compiler_rt/dist/lib/ubsan_minimal/
H A Dubsan_minimal_handlers.cc27 unsigned sz = __sanitizer::atomic_load_relaxed(&caller_pcs_sz); in report_this_error()
34 p = __sanitizer::atomic_load_relaxed(&caller_pcs[i]); in report_this_error()
/netbsd/sys/external/bsd/compiler_rt/dist/lib/xray/
H A Dxray_profiling.cc86 if (atomic_load_relaxed(&ThreadExitingLatch)) in getThreadLocalData()
143 atomic_load_relaxed(&TLD.Allocators))); in getThreadLocalData()
295 auto FCT = reinterpret_cast<FunctionCallTrie *>(atomic_load_relaxed(&T->FCT)); in profilingHandleArg0()
439 if (atomic_load_relaxed(&T->Allocators) == 0) in profilingLoggingInit()

1234567