/netbsd/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
H A D | igt_spinner.c | 52 spin->batch = vaddr; in igt_spinner_init() 102 u32 *batch; in igt_spinner_create_request() local 140 batch = spin->batch; in igt_spinner_create_request() 148 *batch++ = 0; in igt_spinner_create_request() 149 *batch++ = hws_address(hws, rq); in igt_spinner_create_request() 152 *batch++ = 0; in igt_spinner_create_request() 153 *batch++ = hws_address(hws, rq); in igt_spinner_create_request() 156 *batch++ = hws_address(hws, rq); in igt_spinner_create_request() 158 *batch++ = rq->fence.seqno; in igt_spinner_create_request() 160 *batch++ = arbitration_command; in igt_spinner_create_request() [all …]
|
H A D | i915_request.c | 691 if (IS_ERR(batch)) in live_empty_request() 750 i915_vma_unpin(batch); in live_empty_request() 751 i915_vma_put(batch); in live_empty_request() 850 if (IS_ERR(batch)) { in live_all_engines() 851 err = PTR_ERR(batch); in live_all_engines() 871 request[idx]->batch = batch; in live_all_engines() 873 i915_vma_lock(batch); in live_all_engines() 930 i915_vma_unpin(batch); in live_all_engines() 931 i915_vma_put(batch); in live_all_engines() 968 if (IS_ERR(batch)) { in live_sequential_engines() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gem/ |
H A D | i915_gem_object_blt.c | 27 struct i915_vma *batch; in intel_emit_vma_fill_blt() local 91 if (IS_ERR(batch)) { in intel_emit_vma_fill_blt() 92 err = PTR_ERR(batch); in intel_emit_vma_fill_blt() 100 batch->private = pool; in intel_emit_vma_fill_blt() 101 return batch; in intel_emit_vma_fill_blt() 156 if (IS_ERR(batch)) { in i915_gem_object_fill_blt() 190 batch->node.start, batch->node.size, in i915_gem_object_fill_blt() 290 if (IS_ERR(batch)) { in intel_emit_vma_copy_blt() 300 return batch; in intel_emit_vma_copy_blt() 347 if (IS_ERR(batch)) { in i915_gem_object_copy_blt() [all …]
|
H A D | i915_gem_execbuffer.c | 534 eb->batch = vma; in eb_add_vma() 1210 rq->batch = batch; in __reloc_gpu_alloc() 1276 u32 *batch; in relocate_entry() local 1313 *batch++ = 0; in relocate_entry() 1314 *batch++ = addr; in relocate_entry() 1318 *batch++ = 0; in relocate_entry() 1319 *batch++ = addr; in relocate_entry() 1323 *batch++ = addr; in relocate_entry() 2069 pw->batch = eb->batch; in eb_parse_pipeline() 2763 eb.batch = vma; in i915_gem_do_execbuffer() [all …]
|
H A D | i915_gem_client_blt.c | 166 struct i915_vma *batch; in clear_pages_worker() local 184 batch = intel_emit_vma_fill_blt(w->ce, vma, w->value); in clear_pages_worker() 185 if (IS_ERR(batch)) { in clear_pages_worker() 186 err = PTR_ERR(batch); in clear_pages_worker() 201 err = intel_emit_vma_mark_active(batch, rq); in clear_pages_worker() 221 batch->node.start, batch->node.size, in clear_pages_worker() 231 intel_emit_vma_release(w->ce, batch); in clear_pages_worker()
|
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
H A D | igt_gem_utils.c | 117 struct i915_vma *batch; in igt_gpu_fill_dw() local 124 batch = igt_emit_store_dw(vma, offset, count, val); in igt_gpu_fill_dw() 125 if (IS_ERR(batch)) in igt_gpu_fill_dw() 126 return PTR_ERR(batch); in igt_gpu_fill_dw() 139 batch->node.start, batch->node.size, in igt_gpu_fill_dw() 144 i915_vma_lock(batch); in igt_gpu_fill_dw() 145 err = i915_request_await_object(rq, batch->obj, false); in igt_gpu_fill_dw() 147 err = i915_vma_move_to_active(batch, rq, 0); in igt_gpu_fill_dw() 148 i915_vma_unlock(batch); in igt_gpu_fill_dw() 162 i915_vma_unpin_and_release(&batch, 0); in igt_gpu_fill_dw() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
H A D | vmwgfx_mob.c | 266 0, false, &batch->otable_bo); in vmw_otable_batch_setup() 280 ttm_bo_unreserve(batch->otable_bo); in vmw_otable_batch_setup() 284 if (!batch->otables[i].enabled) in vmw_otable_batch_setup() 301 if (batch->otables[i].enabled) in vmw_otable_batch_setup() 303 &batch->otables[i]); in vmw_otable_batch_setup() 306 ttm_bo_put(batch->otable_bo); in vmw_otable_batch_setup() 307 batch->otable_bo = NULL; in vmw_otable_batch_setup() 362 if (batch->otables[i].enabled) in vmw_otable_batch_takedown() 364 &batch->otables[i]); in vmw_otable_batch_takedown() 372 ttm_bo_put(batch->otable_bo); in vmw_otable_batch_takedown() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gt/ |
H A D | selftest_hangcheck.c | 57 u32 *batch; member 100 h->batch = vaddr; in hang_init() 147 u32 *batch; in hang_create_request() local 167 h->batch = vaddr; in hang_create_request() 205 batch = h->batch; in hang_create_request() 214 batch += 1024 / sizeof(*batch); in hang_create_request() 222 *batch++ = 0; in hang_create_request() 228 batch += 1024 / sizeof(*batch); in hang_create_request() 235 *batch++ = 0; in hang_create_request() 241 batch += 1024 / sizeof(*batch); in hang_create_request() [all …]
|
H A D | selftest_workarounds.c | 479 struct i915_vma *batch; in check_dirty_whitelist() local 488 if (IS_ERR(batch)) { in check_dirty_whitelist() 489 err = PTR_ERR(batch); in check_dirty_whitelist() 583 i915_vma_lock(batch); in check_dirty_whitelist() 587 i915_vma_unlock(batch); in check_dirty_whitelist() 819 struct i915_vma *batch; in scrub_whitelisted_registers() local 824 batch = create_batch(vm); in scrub_whitelisted_registers() 826 if (IS_ERR(batch)) in scrub_whitelisted_registers() 827 return PTR_ERR(batch); in scrub_whitelisted_registers() 865 i915_vma_lock(batch); in scrub_whitelisted_registers() [all …]
|
H A D | intel_lrc.c | 3116 *batch++ = 0; in gen8_emit_flush_coherentl3_wa() 3122 batch = gen8_emit_pipe_control(batch, in gen8_emit_flush_coherentl3_wa() 3131 *batch++ = 0; in gen8_emit_flush_coherentl3_wa() 3133 return batch; in gen8_emit_flush_coherentl3_wa() 3158 batch = gen8_emit_flush_coherentl3_wa(engine, batch); in gen8_init_indirectctx_bb() 3162 batch = gen8_emit_pipe_control(batch, in gen8_init_indirectctx_bb() 3181 return batch; in gen8_init_indirectctx_bb() 3231 batch = gen8_emit_flush_coherentl3_wa(engine, batch); in gen9_init_indirectctx_bb() 3234 batch = gen8_emit_pipe_control(batch, in gen9_init_indirectctx_bb() 3241 batch = emit_lri(batch, lri, ARRAY_SIZE(lri)); in gen9_init_indirectctx_bb() [all …]
|
H A D | intel_engine.h | 246 static inline u32 *gen8_emit_pipe_control(u32 *batch, u32 flags, u32 offset) in gen8_emit_pipe_control() argument 248 memset(batch, 0, 6 * sizeof(u32)); in gen8_emit_pipe_control() 250 batch[0] = GFX_OP_PIPE_CONTROL(6); in gen8_emit_pipe_control() 251 batch[1] = flags; in gen8_emit_pipe_control() 252 batch[2] = offset; in gen8_emit_pipe_control() 254 return batch + 6; in gen8_emit_pipe_control()
|
/netbsd/sys/external/bsd/drm/dist/shared-core/ |
H A D | i915_dma.c | 490 int nbox = batch->num_cliprects; in i915_dispatch_batchbuffer() 494 if ((batch->start | batch->used) & 0x7) { in i915_dispatch_batchbuffer() 506 batch->DR1, batch->DR4); in i915_dispatch_batchbuffer() 515 OUT_RING(batch->start); in i915_dispatch_batchbuffer() 525 OUT_RING(batch->start + batch->used - 4); in i915_dispatch_batchbuffer() 613 drm_i915_batchbuffer_t *batch = data; in i915_batchbuffer() local 623 batch->start, batch->used, batch->num_cliprects); in i915_batchbuffer() 629 if (batch->num_cliprects && DRM_VERIFYAREA_READ(batch->cliprects, in i915_batchbuffer() 634 if (batch->num_cliprects) { in i915_batchbuffer() 650 if (batch->num_cliprects) in i915_batchbuffer() [all …]
|
/netbsd/usr.bin/mail/ |
H A D | mime_detach.c | 62 int batch; member 71 detach_ctl.batch = value(ENAME_MIME_DETACH_BATCH) != NULL; in mime_detach_control() 72 detach_ctl.ask = detach_ctl.batch ? 0 : 1; in mime_detach_control() 97 if (!detach_ctl.batch) { in detach_get_fname() 146 detach_ctl.batch = 1; in detach_open_core() 155 detach_ctl.batch = 1; in detach_open_core() 237 detach_ctl.batch = 0; in detach_open_target() 242 } while (!detach_ctl.batch); in detach_open_target()
|
/netbsd/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/ |
H A D | sanitizer_quarantine.h | 31 void *batch[kSize]; member 35 batch[0] = ptr; in init() 46 batch[count++] = ptr; in push_back() 59 batch[count + i] = from->batch[i]; in merge() 185 CHECK(kPrefetch <= ARRAY_SIZE(b->batch)); in DoRecycle() 187 PREFETCH(b->batch[i]); in DoRecycle() 190 PREFETCH(b->batch[i + kPrefetch]); in DoRecycle() 191 cb.Recycle((Node*)b->batch[i]); in DoRecycle()
|
H A D | sanitizer_allocator_local_cache.h | 167 void *res = c->batch[--c->count]; in Allocate() 168 PREFETCH(c->batch[c->count - 1]); in Allocate() 182 c->batch[c->count++] = p; in Deallocate() 210 void *batch[2 * TransferBatch::kMaxNumCached]; member 247 b->CopyToArray(c->batch); in Refill() 258 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]); in Drain() 266 b->SetFromArray(&c->batch[first_idx_to_drain], count); in Drain()
|
/netbsd/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/ |
H A D | sanitizer_quarantine.h | 33 void *batch[kSize]; member 37 batch[0] = ptr; in init() 48 batch[count++] = ptr; in push_back() 61 batch[count + i] = from->batch[i]; in merge() 187 CHECK(kPrefetch <= ARRAY_SIZE(b->batch)); in DoRecycle() 189 PREFETCH(b->batch[i]); in DoRecycle() 192 PREFETCH(b->batch[i + kPrefetch]); in DoRecycle() 193 cb.Recycle((Node*)b->batch[i]); in DoRecycle()
|
H A D | sanitizer_allocator_local_cache.h | 169 void *res = c->batch[--c->count]; in Allocate() 170 PREFETCH(c->batch[c->count - 1]); in Allocate() 184 c->batch[c->count++] = p; in Deallocate() 212 void *batch[2 * TransferBatch::kMaxNumCached]; member 249 b->CopyToArray(c->batch); in Refill() 260 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]); in Drain() 268 b->SetFromArray(&c->batch[first_idx_to_drain], count); in Drain()
|
/netbsd/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/ |
H A D | sanitizer_quarantine.h | 31 void *batch[kSize]; member 35 batch[0] = ptr; in init() 46 batch[count++] = ptr; in push_back() 59 batch[count + i] = from->batch[i]; in merge() 185 CHECK(kPrefetch <= ARRAY_SIZE(b->batch)); in DoRecycle() 187 PREFETCH(b->batch[i]); in DoRecycle() 190 PREFETCH(b->batch[i + kPrefetch]); in DoRecycle() 191 cb.Recycle((Node*)b->batch[i]); in DoRecycle()
|
H A D | sanitizer_allocator_local_cache.h | 167 void *res = c->batch[--c->count]; in Allocate() 168 PREFETCH(c->batch[c->count - 1]); in Allocate() 182 c->batch[c->count++] = p; in Deallocate() 210 void *batch[2 * TransferBatch::kMaxNumCached]; member 247 b->CopyToArray(c->batch); in Refill() 258 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]); in Drain() 266 b->SetFromArray(&c->batch[first_idx_to_drain], count); in Drain()
|
/netbsd/external/gpl3/gcc/dist/libsanitizer/tsan/ |
H A D | tsan_dense_alloc.h | 110 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); in Refill() local 114 new(batch + i) T; in Refill() 115 *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size; in Refill() 117 *(IndexT*)(batch + kL2Size - 1) = 0; in Refill() 119 map_[fillpos_++] = batch; in Refill()
|
/netbsd/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/ |
H A D | tsan_dense_alloc.h | 112 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); in Refill() local 116 new(batch + i) T; in Refill() 117 *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size; in Refill() 119 *(IndexT*)(batch + kL2Size - 1) = 0; in Refill() 121 map_[fillpos_++] = batch; in Refill()
|
/netbsd/external/gpl3/gcc.old/dist/libsanitizer/tsan/ |
H A D | tsan_dense_alloc.h | 110 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); in Refill() local 114 new(batch + i) T; in Refill() 115 *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size; in Refill() 117 *(IndexT*)(batch + kL2Size - 1) = 0; in Refill() 119 map_[fillpos_++] = batch; in Refill()
|
/netbsd/external/bsd/openldap/dist/tests/data/tls/ |
H A D | create-crt.sh | 59 -batch > /dev/null 2>&1 63 -batch >/dev/null 2>&1 74 -batch >/dev/null 2>&1 78 -cert ca/certs/testsuiteCA.crt -batch >/dev/null 2>&1
|
/netbsd/tests/usr.bin/gdb/ |
H A D | t_regress.sh | 48 gdb --batch -x test.gdb dig >gdb.out 68 gdb --batch -x test.gdb ./test >gdb.out 2>&1 87 gdb --batch -x test.gdb >gdb.out 2>&1
|
/netbsd/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/tests/ |
H A D | sanitizer_quarantine_test.cc | 39 while (QuarantineBatch *batch = cache->DequeueBatch()) in DeallocateCache() local 40 cb.Deallocate(batch); in DeallocateCache() 53 ASSERT_EQ(into.batch[0], kFakePtr); in TEST() 54 ASSERT_EQ(into.batch[1], kFakePtr); in TEST()
|