/dports/games/retroarch/RetroArch-1.9.7/gfx/include/userland/middleware/khronos/common/ |
H A D | khrn_interlock.h | 49 extern void khrn_interlock_init(KHRN_INTERLOCK_T *interlock); 50 extern void khrn_interlock_term(KHRN_INTERLOCK_T *interlock); 54 extern KHRN_INTERLOCK_USER_T khrn_interlock_get_writer(KHRN_INTERLOCK_T *interlock); 55 extern bool khrn_interlock_release(KHRN_INTERLOCK_T *interlock, KHRN_INTERLOCK_USER_T user); 57 extern bool khrn_interlock_write_would_block(KHRN_INTERLOCK_T *interlock); 59 extern void khrn_interlock_invalidate(KHRN_INTERLOCK_T *interlock); 60 extern bool khrn_interlock_is_invalid(KHRN_INTERLOCK_T *interlock); 66 extern void khrn_interlock_extra_init(KHRN_INTERLOCK_T *interlock); 67 extern void khrn_interlock_extra_term(KHRN_INTERLOCK_T *interlock); 69 extern void khrn_interlock_read_immediate(KHRN_INTERLOCK_T *interlock); [all …]
|
/dports/misc/rump/buildrump.sh-b914579/src/sys/external/bsd/drm2/include/drm/ |
H A D | drm_wait_netbsd.h | 64 DRM_WAITERS_P(drm_waitqueue_t *q, struct mutex *interlock) in DRM_WAITERS_P() argument 66 KASSERT(mutex_is_locked(interlock)); in DRM_WAITERS_P() 71 DRM_WAKEUP_ONE(drm_waitqueue_t *q, struct mutex *interlock) in DRM_WAKEUP_ONE() argument 73 KASSERT(mutex_is_locked(interlock)); in DRM_WAKEUP_ONE() 78 DRM_WAKEUP_ALL(drm_waitqueue_t *q, struct mutex *interlock) in DRM_WAKEUP_ALL() argument 80 KASSERT(mutex_is_locked(interlock)); in DRM_WAKEUP_ALL() 85 DRM_SPIN_WAITERS_P(drm_waitqueue_t *q, spinlock_t *interlock) in DRM_SPIN_WAITERS_P() argument 87 KASSERT(spin_is_locked(interlock)); in DRM_SPIN_WAITERS_P() 92 DRM_SPIN_WAKEUP_ONE(drm_waitqueue_t *q, spinlock_t *interlock) in DRM_SPIN_WAKEUP_ONE() argument 94 KASSERT(spin_is_locked(interlock)); in DRM_SPIN_WAKEUP_ONE() [all …]
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/angle/third_party/VK-GL-CTS/src/external/vulkancts/modules/vulkan/fragment_shader_interlock/ |
H A D | vktFragmentShaderInterlockBasic.cpp | 93 Interlock interlock; member 100 return sampleShading || interlock == INT_SAMPLE_ORDERED || interlock == INT_SAMPLE_UNORDERED; in isSampleInterlock() 104 …return interlock == INT_PIXEL_ORDERED || interlock == INT_SAMPLE_ORDERED || interlock == INT_SHADI… in isOrdered() 156 if ((m_data.interlock == INT_SAMPLE_ORDERED || m_data.interlock == INT_SAMPLE_UNORDERED) && in checkSupport() 162 if ((m_data.interlock == INT_PIXEL_ORDERED || m_data.interlock == INT_PIXEL_UNORDERED) && in checkSupport() 168 …if ((m_data.interlock == INT_SHADING_RATE_ORDERED || m_data.interlock == INT_SHADING_RATE_UNORDERE… in checkSupport() 174 …if ((m_data.interlock == INT_SHADING_RATE_ORDERED || m_data.interlock == INT_SHADING_RATE_UNORDERE… in checkSupport() 187 else if (c.interlock == INT_SHADING_RATE_ORDERED || c.interlock == INT_SHADING_RATE_UNORDERED) in bitsPerQuad() 219 switch (m_data.interlock) in initPrograms() 255 …else if (m_data.interlock == INT_SHADING_RATE_ORDERED || m_data.interlock == INT_SHADING_RATE_UNOR… in initPrograms() [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/nouveau/dispnv50/ |
H A D | wimmc37b.c | 32 wimmc37b_update(struct nv50_wndw *wndw, u32 *interlock) in wimmc37b_update() argument 42 !!(interlock[NV50_DISP_INTERLOCK_WNDW] & wndw->interlock.data))); in wimmc37b_update() 86 wndw->interlock.wimm = wndw->interlock.data; in wimmc37b_init_()
|
H A D | wndw.h | 19 struct nv50_disp_interlock interlock; member 45 void nv50_wndw_flush_set(struct nv50_wndw *, u32 *interlock, 47 void nv50_wndw_flush_clr(struct nv50_wndw *, u32 *interlock, bool flush, 82 int (*update)(struct nv50_wndw *, u32 *interlock); 100 int (*update)(struct nv50_wndw *, u32 *interlock);
|
H A D | wndw.c | 125 nv50_wndw_flush_clr(struct nv50_wndw *wndw, u32 *interlock, bool flush, in nv50_wndw_flush_clr() argument 137 interlock[wndw->interlock.type] |= wndw->interlock.data; in nv50_wndw_flush_clr() 141 nv50_wndw_flush_set(struct nv50_wndw *wndw, u32 *interlock, in nv50_wndw_flush_set() argument 144 if (interlock[NV50_DISP_INTERLOCK_CORE]) { in nv50_wndw_flush_set() 167 interlock[wndw->interlock.type] |= wndw->interlock.data; in nv50_wndw_flush_set() 168 interlock[NV50_DISP_INTERLOCK_WIMM] |= wndw->interlock.wimm; in nv50_wndw_flush_set() 171 wndw->immd->update(wndw, interlock); in nv50_wndw_flush_set() 173 interlock[wndw->interlock.type] |= wndw->interlock.data; in nv50_wndw_flush_set() 733 wndw->interlock.type = interlock_type; in nv50_wndw_new_() 734 wndw->interlock.data = interlock_data; in nv50_wndw_new_()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/nouveau/dispnv50/ |
H A D | wimmc37b.c | 32 wimmc37b_update(struct nv50_wndw *wndw, u32 *interlock) in wimmc37b_update() argument 42 !!(interlock[NV50_DISP_INTERLOCK_WNDW] & wndw->interlock.data))); in wimmc37b_update() 86 wndw->interlock.wimm = wndw->interlock.data; in wimmc37b_init_()
|
H A D | wndw.h | 19 struct nv50_disp_interlock interlock; member 45 void nv50_wndw_flush_set(struct nv50_wndw *, u32 *interlock, 47 void nv50_wndw_flush_clr(struct nv50_wndw *, u32 *interlock, bool flush, 82 int (*update)(struct nv50_wndw *, u32 *interlock); 100 int (*update)(struct nv50_wndw *, u32 *interlock);
|
H A D | wndw.c | 125 nv50_wndw_flush_clr(struct nv50_wndw *wndw, u32 *interlock, bool flush, in nv50_wndw_flush_clr() argument 137 interlock[wndw->interlock.type] |= wndw->interlock.data; in nv50_wndw_flush_clr() 141 nv50_wndw_flush_set(struct nv50_wndw *wndw, u32 *interlock, in nv50_wndw_flush_set() argument 144 if (interlock[NV50_DISP_INTERLOCK_CORE]) { in nv50_wndw_flush_set() 167 interlock[wndw->interlock.type] |= wndw->interlock.data; in nv50_wndw_flush_set() 168 interlock[NV50_DISP_INTERLOCK_WIMM] |= wndw->interlock.wimm; in nv50_wndw_flush_set() 171 wndw->immd->update(wndw, interlock); in nv50_wndw_flush_set() 173 interlock[wndw->interlock.type] |= wndw->interlock.data; in nv50_wndw_flush_set() 733 wndw->interlock.type = interlock_type; in nv50_wndw_new_() 734 wndw->interlock.data = interlock_data; in nv50_wndw_new_()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/nouveau/dispnv50/ |
H A D | wimmc37b.c | 32 wimmc37b_update(struct nv50_wndw *wndw, u32 *interlock) in wimmc37b_update() argument 42 !!(interlock[NV50_DISP_INTERLOCK_WNDW] & wndw->interlock.data))); in wimmc37b_update() 86 wndw->interlock.wimm = wndw->interlock.data; in wimmc37b_init_()
|
H A D | wndw.h | 19 struct nv50_disp_interlock interlock; member 45 void nv50_wndw_flush_set(struct nv50_wndw *, u32 *interlock, 47 void nv50_wndw_flush_clr(struct nv50_wndw *, u32 *interlock, bool flush, 82 int (*update)(struct nv50_wndw *, u32 *interlock); 100 int (*update)(struct nv50_wndw *, u32 *interlock);
|
H A D | wndw.c | 125 nv50_wndw_flush_clr(struct nv50_wndw *wndw, u32 *interlock, bool flush, in nv50_wndw_flush_clr() argument 137 interlock[wndw->interlock.type] |= wndw->interlock.data; in nv50_wndw_flush_clr() 141 nv50_wndw_flush_set(struct nv50_wndw *wndw, u32 *interlock, in nv50_wndw_flush_set() argument 144 if (interlock[NV50_DISP_INTERLOCK_CORE]) { in nv50_wndw_flush_set() 167 interlock[wndw->interlock.type] |= wndw->interlock.data; in nv50_wndw_flush_set() 168 interlock[NV50_DISP_INTERLOCK_WIMM] |= wndw->interlock.wimm; in nv50_wndw_flush_set() 171 wndw->immd->update(wndw, interlock); in nv50_wndw_flush_set() 173 interlock[wndw->interlock.type] |= wndw->interlock.data; in nv50_wndw_flush_set() 733 wndw->interlock.type = interlock_type; in nv50_wndw_new_() 734 wndw->interlock.data = interlock_data; in nv50_wndw_new_()
|
/dports/misc/rump/buildrump.sh-b914579/src/lib/librumpuser/ |
H A D | rumpuser_int.h | 37 rumpkern_unsched(int *nlocks, void *interlock) in rumpkern_unsched() argument 40 rumpuser__hyp.hyp_backend_unschedule(0, nlocks, interlock); in rumpkern_unsched() 44 rumpkern_sched(int nlocks, void *interlock) in rumpkern_sched() argument 47 rumpuser__hyp.hyp_backend_schedule(nlocks, interlock); in rumpkern_sched()
|
/dports/misc/rump/buildrump.sh-b914579/src/sys/rump/librump/rumpkern/ |
H A D | scheduler.c | 290 rump_schedule_cpu_interlock(struct lwp *l, void *interlock) in rump_schedule_cpu_interlock() argument 313 if (interlock == rcpu->rcpu_mtx) in rump_schedule_cpu_interlock() 330 if (interlock != rcpu->rcpu_mtx) in rump_schedule_cpu_interlock() 436 rump_unschedule_cpu_interlock(struct lwp *l, void *interlock) in rump_unschedule_cpu_interlock() argument 441 rump_unschedule_cpu1(l, interlock); in rump_unschedule_cpu_interlock() 445 rump_unschedule_cpu1(struct lwp *l, void *interlock) in rump_unschedule_cpu1() argument 469 if (interlock == rcpu->rcpu_mtx) in rump_unschedule_cpu1() 490 if (interlock != rcpu->rcpu_mtx) in rump_unschedule_cpu1() 494 if (interlock != rcpu->rcpu_mtx) in rump_unschedule_cpu1()
|
H A D | klock.c | 149 rump_user_unschedule(int nlocks, int *countp, void *interlock) in rump_user_unschedule() argument 157 rump_unschedule_cpu_interlock(curlwp, interlock); in rump_user_unschedule() 161 rump_user_schedule(int nlocks, void *interlock) in rump_user_schedule() argument 164 rump_schedule_cpu_interlock(curlwp, interlock); in rump_user_schedule()
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/native_client_sdk/src/libraries/third_party/pthreads-win32/ |
H A D | pthread_spin_lock.c | 64 PTW32_INTERLOCKED_COMPARE_EXCHANGE_LONG ((PTW32_INTERLOCKED_LONGPTR) &s->interlock, in pthread_spin_lock() 70 if (s->interlock == PTW32_SPIN_LOCKED) in pthread_spin_lock() 74 else if (s->interlock == PTW32_SPIN_USE_MUTEX) in pthread_spin_lock()
|
H A D | pthread_spin_destroy.c | 54 if (s->interlock == PTW32_SPIN_USE_MUTEX) in pthread_spin_destroy() 59 PTW32_INTERLOCKED_COMPARE_EXCHANGE_LONG ((PTW32_INTERLOCKED_LONGPTR) &s->interlock, in pthread_spin_destroy()
|
/dports/misc/visp/visp-3.4.0/3rdparty/pthreads4w/ |
H A D | pthread_spin_lock.c | 66 __PTW32_INTERLOCKED_COMPARE_EXCHANGE_LONG ((__PTW32_INTERLOCKED_LONGPTR) &s->interlock, in pthread_spin_lock() 72 if (s->interlock == __PTW32_SPIN_LOCKED) in pthread_spin_lock() 76 else if (s->interlock == __PTW32_SPIN_USE_MUTEX) in pthread_spin_lock()
|
H A D | pthread_spin_destroy.c | 56 if (s->interlock == __PTW32_SPIN_USE_MUTEX) in pthread_spin_destroy() 61 __PTW32_INTERLOCKED_COMPARE_EXCHANGE_LONG ((__PTW32_INTERLOCKED_LONGPTR) &s->interlock, in pthread_spin_destroy()
|
/dports/devel/mingw32-pthreads/pthreads-w32-2-8-0-release/ |
H A D | pthread_spin_lock.c | 65 (s->interlock), in pthread_spin_lock() 73 if (s->interlock == PTW32_SPIN_LOCKED) in pthread_spin_lock() 77 else if (s->interlock == PTW32_SPIN_USE_MUTEX) in pthread_spin_lock()
|
H A D | pthread_spin_destroy.c | 54 if (s->interlock == PTW32_SPIN_USE_MUTEX) in pthread_spin_destroy() 60 & (s->interlock), in pthread_spin_destroy()
|
/dports/multimedia/obs-studio/obs-studio-27.1.3/deps/w32-pthreads/ |
H A D | pthread_spin_lock.c | 64 PTW32_INTERLOCKED_COMPARE_EXCHANGE_LONG ((PTW32_INTERLOCKED_LONGPTR) &s->interlock, in pthread_spin_lock() 70 if (s->interlock == PTW32_SPIN_LOCKED) in pthread_spin_lock() 74 else if (s->interlock == PTW32_SPIN_USE_MUTEX) in pthread_spin_lock()
|
H A D | pthread_spin_destroy.c | 54 if (s->interlock == PTW32_SPIN_USE_MUTEX) in pthread_spin_destroy() 59 PTW32_INTERLOCKED_COMPARE_EXCHANGE_LONG ((PTW32_INTERLOCKED_LONGPTR) &s->interlock, in pthread_spin_destroy()
|
/dports/www/gitlab-workhorse/gitlab-foss-0a901d60f8ae4a60c04ae82e6e9c3a03e9321417/lib/backup/ |
H A D | repositories.rb | 37 ActiveSupport::Dependencies.interlock.permit_concurrent_loads do 96 ActiveSupport::Dependencies.interlock.permit_concurrent_loads do 118 ActiveSupport::Dependencies.interlock.permit_concurrent_loads do 208 ActiveSupport::Dependencies.interlock.permit_concurrent_loads do 215 ActiveSupport::Dependencies.interlock.permit_concurrent_loads do
|
/dports/misc/rump/buildrump.sh-b914579/src/sys/external/bsd/drm2/include/linux/ |
H A D | kref.h | 104 struct mutex *interlock) in kref_put_mutex() argument 112 mutex_lock(interlock); in kref_put_mutex() 117 mutex_unlock(interlock); in kref_put_mutex()
|