/dports/lang/mono/mono-5.10.1.57/mono/utils/ |
H A D | memfuncs.c | 33 #define ptr_mask ((sizeof (void*) - 1)) macro 35 #define unaligned_bytes(ptr) (_toi(ptr) & ptr_mask) 36 #define align_down(ptr) ((void*)(_toi(ptr) & ~ptr_mask)) 37 #define align_up(ptr) ((void*) ((_toi(ptr) + ptr_mask) & ~ptr_mask))
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_ih.c | 51 ih->ptr_mask = ih->ring_size - 1; in amdgpu_ih_ring_init() 156 wptr &= ih->ptr_mask; in amdgpu_ih_ring_write() 171 uint32_t cur_rptr = ih->rptr | (*prev_rptr & ~ih->ptr_mask); in amdgpu_ih_has_checkpoint_processed() 175 cur_rptr += ih->ptr_mask + 1; in amdgpu_ih_has_checkpoint_processed() 204 checkpoint_wptr += ih->ptr_mask + 1; in amdgpu_ih_wait_on_checkpoint_process() 238 ih->rptr &= ih->ptr_mask; in amdgpu_ih_process()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_ih.c | 51 ih->ptr_mask = ih->ring_size - 1; in amdgpu_ih_ring_init() 156 wptr &= ih->ptr_mask; in amdgpu_ih_ring_write() 171 uint32_t cur_rptr = ih->rptr | (*prev_rptr & ~ih->ptr_mask); in amdgpu_ih_has_checkpoint_processed() 175 cur_rptr += ih->ptr_mask + 1; in amdgpu_ih_has_checkpoint_processed() 204 checkpoint_wptr += ih->ptr_mask + 1; in amdgpu_ih_wait_on_checkpoint_process() 238 ih->rptr &= ih->ptr_mask; in amdgpu_ih_process()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_ih.c | 51 ih->ptr_mask = ih->ring_size - 1; in amdgpu_ih_ring_init() 156 wptr &= ih->ptr_mask; in amdgpu_ih_ring_write() 171 uint32_t cur_rptr = ih->rptr | (*prev_rptr & ~ih->ptr_mask); in amdgpu_ih_has_checkpoint_processed() 175 cur_rptr += ih->ptr_mask + 1; in amdgpu_ih_has_checkpoint_processed() 204 checkpoint_wptr += ih->ptr_mask + 1; in amdgpu_ih_wait_on_checkpoint_process() 238 ih->rptr &= ih->ptr_mask; in amdgpu_ih_process()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/net/ethernet/sfc/ |
H A D | tx_common.c | 18 return DIV_ROUND_UP(tx_queue->ptr_mask + 1, in efx_tx_cb_page_count() 31 tx_queue->ptr_mask = entries - 1; in efx_probe_tx_queue() 35 tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask); in efx_probe_tx_queue() 111 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_fini_tx_queue() 205 stop_index = (index + 1) & tx_queue->ptr_mask; in efx_dequeue_buffers() 206 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_dequeue_buffers() 222 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_dequeue_buffers() 244 EFX_WARN_ON_ONCE_PARANOID(index > tx_queue->ptr_mask); in efx_xmit_done()
|
H A D | ef100_tx.c | 27 (tx_queue->ptr_mask + 2) * in ef100_tx_probe() 120 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in ef100_notify_tx_desc() 280 write_ptr = new_write_count & tx_queue->ptr_mask; in ef100_tx_make_descriptors() 340 tx_queue->ptr_mask; in ef100_ev_tx() 361 if (!tx_queue->buffer || !tx_queue->ptr_mask) { in ef100_enqueue_skb()
|
H A D | ef100_rx.c | 145 rx_queue->removed_count & rx_queue->ptr_mask); in efx_ef100_ev_rx() 158 idx = rx_queue->notified_count & rx_queue->ptr_mask; in ef100_rx_write() 169 rx_queue->added_count & rx_queue->ptr_mask); in ef100_rx_write()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/net/ethernet/sfc/ |
H A D | tx_common.c | 18 return DIV_ROUND_UP(tx_queue->ptr_mask + 1, in efx_tx_cb_page_count() 31 tx_queue->ptr_mask = entries - 1; in efx_probe_tx_queue() 35 tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask); in efx_probe_tx_queue() 111 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_fini_tx_queue() 205 stop_index = (index + 1) & tx_queue->ptr_mask; in efx_dequeue_buffers() 206 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_dequeue_buffers() 222 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_dequeue_buffers() 244 EFX_WARN_ON_ONCE_PARANOID(index > tx_queue->ptr_mask); in efx_xmit_done()
|
H A D | ef100_tx.c | 27 (tx_queue->ptr_mask + 2) * in ef100_tx_probe() 120 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in ef100_notify_tx_desc() 280 write_ptr = new_write_count & tx_queue->ptr_mask; in ef100_tx_make_descriptors() 340 tx_queue->ptr_mask; in ef100_ev_tx() 361 if (!tx_queue->buffer || !tx_queue->ptr_mask) { in ef100_enqueue_skb()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/net/ethernet/sfc/ |
H A D | tx_common.c | 18 return DIV_ROUND_UP(tx_queue->ptr_mask + 1, in efx_tx_cb_page_count() 31 tx_queue->ptr_mask = entries - 1; in efx_probe_tx_queue() 35 tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask); in efx_probe_tx_queue() 111 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_fini_tx_queue() 205 stop_index = (index + 1) & tx_queue->ptr_mask; in efx_dequeue_buffers() 206 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_dequeue_buffers() 222 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_dequeue_buffers() 244 EFX_WARN_ON_ONCE_PARANOID(index > tx_queue->ptr_mask); in efx_xmit_done()
|
H A D | ef100_tx.c | 27 (tx_queue->ptr_mask + 2) * in ef100_tx_probe() 120 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in ef100_notify_tx_desc() 280 write_ptr = new_write_count & tx_queue->ptr_mask; in ef100_tx_make_descriptors() 340 tx_queue->ptr_mask; in ef100_ev_tx() 361 if (!tx_queue->buffer || !tx_queue->ptr_mask) { in ef100_enqueue_skb()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | radeon_ring.c | 88 ring->ring_free_dw &= ring->ptr_mask; in radeon_ring_free_size() 316 size &= ring->ptr_mask; in radeon_ring_backup() 330 ptr &= ring->ptr_mask; in radeon_ring_backup() 417 ring->ptr_mask = (ring->ring_size / 4) - 1; in radeon_ring_init() 506 i = (rptr + ring->ptr_mask + 1 - 32) & ring->ptr_mask; in radeon_debugfs_ring_info_show() 514 i = (i + 1) & ring->ptr_mask; in radeon_debugfs_ring_info_show()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | radeon_ring.c | 88 ring->ring_free_dw &= ring->ptr_mask; in radeon_ring_free_size() 316 size &= ring->ptr_mask; in radeon_ring_backup() 330 ptr &= ring->ptr_mask; in radeon_ring_backup() 417 ring->ptr_mask = (ring->ring_size / 4) - 1; in radeon_ring_init() 506 i = (rptr + ring->ptr_mask + 1 - 32) & ring->ptr_mask; in radeon_debugfs_ring_info_show() 514 i = (i + 1) & ring->ptr_mask; in radeon_debugfs_ring_info_show()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | radeon_ring.c | 88 ring->ring_free_dw &= ring->ptr_mask; in radeon_ring_free_size() 316 size &= ring->ptr_mask; in radeon_ring_backup() 330 ptr &= ring->ptr_mask; in radeon_ring_backup() 417 ring->ptr_mask = (ring->ring_size / 4) - 1; in radeon_ring_init() 506 i = (rptr + ring->ptr_mask + 1 - 32) & ring->ptr_mask; in radeon_debugfs_ring_info_show() 514 i = (i + 1) & ring->ptr_mask; in radeon_debugfs_ring_info_show()
|
/dports/java/openjdk13/jdk13u-jdk-13.0.10-1-1/src/hotspot/share/jvmci/ |
H A D | metadataHandleBlock.hpp | 87 ptr_mask = ~((intptr_t)ptr_tag) enumerator 127 HandleRecord* handle = (HandleRecord*) (_free_list & ptr_mask); in get_free_handle() 128 _free_list = (ptr_mask & (intptr_t) (handle->value())); in get_free_handle()
|
/dports/java/openjdk15/jdk15u-jdk-15.0.6-1-1/src/hotspot/share/jvmci/ |
H A D | metadataHandleBlock.hpp | 87 ptr_mask = ~((intptr_t)ptr_tag) enumerator 127 HandleRecord* handle = (HandleRecord*) (_free_list & ptr_mask); in get_free_handle() 128 _free_list = (ptr_mask & (intptr_t) (handle->value())); in get_free_handle()
|
/dports/java/openjdk14/jdk14u-jdk-14.0.2-12-1/src/hotspot/share/jvmci/ |
H A D | metadataHandleBlock.hpp | 87 ptr_mask = ~((intptr_t)ptr_tag) enumerator 127 HandleRecord* handle = (HandleRecord*) (_free_list & ptr_mask); in get_free_handle() 128 _free_list = (ptr_mask & (intptr_t) (handle->value())); in get_free_handle()
|
/dports/java/openjdk16/jdk16u-jdk-16.0.2-7-1/src/hotspot/share/jvmci/ |
H A D | metadataHandles.hpp | 123 ptr_mask = ~((intptr_t)ptr_tag) enumerator 135 HandleRecord* handle = (HandleRecord*) (_free_list & ptr_mask); in get_free_handle() 136 _free_list = (ptr_mask & (intptr_t) (handle->value())); in get_free_handle()
|
/dports/java/openjdk17/jdk17u-jdk-17.0.1-12-1/src/hotspot/share/jvmci/ |
H A D | metadataHandles.hpp | 123 ptr_mask = ~((intptr_t)ptr_tag) enumerator 135 HandleRecord* handle = (HandleRecord*) (_free_list & ptr_mask); in get_free_handle() 136 _free_list = (ptr_mask & (intptr_t) (handle->value())); in get_free_handle()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/net/ethernet/sfc/falcon/ |
H A D | tx.c | 362 stop_index = (index + 1) & tx_queue->ptr_mask; in ef4_dequeue_buffers() 363 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in ef4_dequeue_buffers() 380 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in ef4_dequeue_buffers() 498 EF4_BUG_ON_PARANOID(index > tx_queue->ptr_mask); in ef4_xmit_done() 535 return DIV_ROUND_UP(tx_queue->ptr_mask + 1, PAGE_SIZE >> EF4_TX_CB_ORDER); in ef4_tx_cb_page_count() 547 tx_queue->ptr_mask = entries - 1; in ef4_probe_tx_queue() 551 tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask); in ef4_probe_tx_queue() 619 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in ef4_fini_tx_queue()
|
/dports/misc/rump/buildrump.sh-b914579/src/sys/external/bsd/drm2/dist/drm/radeon/ |
H A D | radeon_ring.c | 314 ring->wptr &= ring->ptr_mask; in radeon_ring_write() 357 ring->ring_free_dw &= ring->ptr_mask; in radeon_ring_free_size() 568 size &= ring->ptr_mask; in radeon_ring_backup() 582 ptr &= ring->ptr_mask; in radeon_ring_backup() 668 ring->ptr_mask = (ring->ring_size / 4) - 1; in radeon_ring_init() 762 i = (rptr + ring->ptr_mask + 1 - 32) & ring->ptr_mask; in radeon_debugfs_ring_info() 770 i = (i + 1) & ring->ptr_mask; in radeon_debugfs_ring_info()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/net/ethernet/sfc/falcon/ |
H A D | tx.c | 362 stop_index = (index + 1) & tx_queue->ptr_mask; in ef4_dequeue_buffers() 363 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in ef4_dequeue_buffers() 380 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in ef4_dequeue_buffers() 498 EF4_BUG_ON_PARANOID(index > tx_queue->ptr_mask); in ef4_xmit_done() 535 return DIV_ROUND_UP(tx_queue->ptr_mask + 1, PAGE_SIZE >> EF4_TX_CB_ORDER); in ef4_tx_cb_page_count() 547 tx_queue->ptr_mask = entries - 1; in ef4_probe_tx_queue() 551 tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask); in ef4_probe_tx_queue() 619 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in ef4_fini_tx_queue()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/net/ethernet/sfc/falcon/ |
H A D | tx.c | 362 stop_index = (index + 1) & tx_queue->ptr_mask; in ef4_dequeue_buffers() 363 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in ef4_dequeue_buffers() 380 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in ef4_dequeue_buffers() 498 EF4_BUG_ON_PARANOID(index > tx_queue->ptr_mask); in ef4_xmit_done() 535 return DIV_ROUND_UP(tx_queue->ptr_mask + 1, PAGE_SIZE >> EF4_TX_CB_ORDER); in ef4_tx_cb_page_count() 547 tx_queue->ptr_mask = entries - 1; in ef4_probe_tx_queue() 551 tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask); in ef4_probe_tx_queue() 619 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in ef4_fini_tx_queue()
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/tcmalloc/chromium/src/ |
H A D | free_list.cc | 72 uintptr_t ptr_mask = 0; variable 75 return reinterpret_cast<void*>(reinterpret_cast<uintptr_t>(p) ^ ptr_mask); in MaskPtr() 119 ptr_mask = ~(seed >> 13); in FL_InitPtrMask()
|
/dports/devel/hpx/hpx-1.2.1/hpx/util/lockfree/detail/ |
H A D | tagged_ptr_pair.hpp | 70 HPX_STATIC_CONSTEXPR compressed_ptr_t ptr_mask = 0xffffffffffff; member 76 return reinterpret_cast<Left*>(cu.ptrs[left_ptr_index] & ptr_mask); in extract_left_ptr() 83 return reinterpret_cast<Right*>(cu.ptrs[right_ptr_index] & ptr_mask); in extract_right_ptr() 255 HPX_STATIC_CONSTEXPR compressed_ptr_t ptr_mask = 0xffffffffffff; 258 { return reinterpret_cast<Left*>(i.left & ptr_mask); } 261 { return reinterpret_cast<Right*>(i.right & ptr_mask); }
|