/netbsd/sys/external/bsd/drm2/dist/drm/i915/ |
H A D | i915_request.c | 257 rq->ring->head = rq->postfix; in i915_request_retire() 582 if (rq) in request_alloc_slow() 738 rq->head = rq->ring->emit; in __i915_request_create() 740 ret = rq->engine->request_alloc(rq); in __i915_request_create() 747 return rq; in __i915_request_create() 787 return rq; in i915_request_create() 791 return rq; in i915_request_create() 1207 if (rq->infix == rq->postfix) in i915_request_skip() 1221 rq->infix = rq->postfix; in i915_request_skip() 1310 rq->postfix = intel_ring_offset(rq, cs); in __i915_request_commit() [all …]
|
H A D | i915_request.h | 54 #define RQ_TRACE(rq, fmt, ...) do { \ argument 343 dma_fence_put(&rq->fence); in i915_request_put() 404 return READ_ONCE(*rq->hwsp_seqno); in __hwsp_seqno() 425 seqno = __hwsp_seqno(rq); in hwsp_seqno() 433 return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno - 1); in __i915_request_has_started() 464 if (i915_request_signaled(rq)) in i915_request_started() 481 if (!i915_request_is_active(rq)) in i915_request_is_running() 505 return !list_empty(&rq->sched.link); in i915_request_is_ready() 510 if (i915_request_signaled(rq)) in i915_request_completed() 513 return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno); in i915_request_completed() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gt/ |
H A D | selftest_lrc.c | 234 GEM_BUG_ON(rq[1]->postfix <= rq[0]->postfix); in live_unlite_restore() 254 i915_request_await_dma_fence(rq[0], &rq[1]->fence); in live_unlite_restore() 257 GEM_BUG_ON(rq[0]->postfix > rq[1]->postfix); in live_unlite_restore() 448 err = rq->engine->emit_init_breadcrumb(rq); in semaphore_queue() 1772 rq = kzalloc(sizeof(*rq), GFP_KERNEL); in dummy_request() 1780 rq->fence.lock = &rq->lock; in dummy_request() 1788 rq->hwsp_seqno = (u32 *)&rq->fence.seqno + 1; in dummy_request() 1795 rq->fence.lock = &rq->lock; in dummy_request() 1982 ring_size = rq->wa_tail - rq->head; in live_chain_preempt() 2162 err = rq->engine->emit_bb_start(rq, in create_gang() [all …]
|
H A D | intel_breadcrumbs.c | 94 return i915_seqno_passed(__hwsp_seqno(rq), rq->fence.seqno); in __request_completed() 107 rq->fence.seqno)) in check_signal_order() 173 struct i915_request *rq = in signal_irq_work() local 182 &rq->fence.flags)); in signal_irq_work() 193 i915_request_get(rq); in signal_irq_work() 215 struct i915_request *rq = in signal_irq_work() local 222 spin_lock(&rq->lock); in signal_irq_work() 226 spin_unlock(&rq->lock); in signal_irq_work() 229 i915_request_put(rq); in signal_irq_work() 364 list_del(&rq->signal_link); in i915_request_cancel_breadcrumb() [all …]
|
H A D | selftest_hangcheck.c | 192 if (IS_ERR(rq)) { in hang_create_request() 263 err = rq->engine->emit_init_breadcrumb(rq); in hang_create_request() 607 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in __igt_reset_engine() 685 if (!rq) in active_request_put() 859 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in __igt_reset_engines() 878 if (rq) { in __igt_reset_engines() 1062 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in igt_reset_wait() 1246 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in __igt_reset_evict_vma() 1473 prev = rq; in igt_reset_queue() 1534 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in igt_handle_error() [all …]
|
H A D | intel_ring_submission.c | 125 if (IS_G4X(rq->i915) || IS_GEN(rq->i915, 5)) in gen4_render_ring_flush() 329 rq->tail = intel_ring_offset(rq, cs); in gen6_rcs_emit_breadcrumb() 330 assert_ring_tail_valid(rq->ring, rq->tail); in gen6_rcs_emit_breadcrumb() 431 rq->tail = intel_ring_offset(rq, cs); in gen7_rcs_emit_breadcrumb() 432 assert_ring_tail_valid(rq->ring, rq->tail); in gen7_rcs_emit_breadcrumb() 448 rq->tail = intel_ring_offset(rq, cs); in gen6_xcs_emit_breadcrumb() 449 assert_ring_tail_valid(rq->ring, rq->tail); in gen6_xcs_emit_breadcrumb() 480 rq->tail = intel_ring_offset(rq, cs); in gen7_xcs_emit_breadcrumb() 481 assert_ring_tail_valid(rq->ring, rq->tail); in gen7_xcs_emit_breadcrumb() 966 rq->tail = intel_ring_offset(rq, cs); in i9xx_emit_breadcrumb() [all …]
|
H A D | intel_engine_heartbeat.c | 69 struct i915_request *rq; in heartbeat() local 72 if (rq && i915_request_completed(rq)) { in heartbeat() 73 i915_request_put(rq); in heartbeat() 129 if (IS_ERR(rq)) in heartbeat() 132 idle_pulse(engine, rq); in heartbeat() 210 if (IS_ERR(rq)) { in intel_engine_pulse() 211 err = PTR_ERR(rq); in intel_engine_pulse() 216 idle_pulse(engine, rq); in intel_engine_pulse() 240 if (IS_ERR(rq)) { in intel_engine_flush_barriers() 241 err = PTR_ERR(rq); in intel_engine_flush_barriers() [all …]
|
H A D | intel_lrc.c | 1231 rq->fence.context, rq->fence.seqno); in reset_active() 1391 tail = intel_ring_set_tail(rq->ring, rq->tail); in execlists_update_context() 1396 rq->tail = rq->wa_tail; in execlists_update_context() 1559 rq ? execlists_update_context(rq) : 0, in execlists_submit_ports() 1730 if (!rq) in defer_active() 1773 if (!rq || i915_request_completed(rq)) in active_timeslice() 2689 cap->rq = active_request(cap->rq->context->timeline, cap->rq); in execlists_capture() 2690 cap->rq = i915_request_get_rcu(cap->rq); in execlists_capture() 3057 rq->infix = intel_ring_offset(rq, cs); in gen8_emit_init_breadcrumb() 3601 rq = active_request(ce->timeline, rq); in __execlists_reset() [all …]
|
H A D | selftest_mocs.c | 27 i915_request_get(rq); in request_add_sync() 28 i915_request_add(rq); in request_add_sync() 31 i915_request_put(rq); in request_add_sync() 40 i915_request_get(rq); in request_add_spin() 41 i915_request_add(rq); in request_add_spin() 44 i915_request_put(rq); in request_add_spin() 211 struct i915_request *rq; in check_mocs_engine() local 219 if (IS_ERR(rq)) in check_mocs_engine() 220 return PTR_ERR(rq); in check_mocs_engine() 325 if (IS_ERR(rq)) { in active_engine_reset() [all …]
|
H A D | selftest_timeline.c | 462 rq = ERR_PTR(err); in tl_write() 467 if (IS_ERR(rq)) in tl_write() 470 i915_request_get(rq); in tl_write() 476 rq = ERR_PTR(err); in tl_write() 482 if (IS_ERR(rq)) in tl_write() 484 return rq; in tl_write() 545 if (IS_ERR(rq)) { in live_hwsp_engine() 620 if (IS_ERR(rq)) { in live_hwsp_alternate() 684 if (IS_ERR(rq)) { in live_hwsp_wrap() 685 err = PTR_ERR(rq); in live_hwsp_wrap() [all …]
|
H A D | intel_engine_cs.c | 1192 const char *name = rq->fence.ops->get_timeline_name(&rq->fence); in print_request() 1196 x = print_sched_attr(rq->i915, &rq->sched.attr, buf, x, sizeof(buf)); in print_request() 1200 (uint64_t)rq->fence.context, (uint64_t)rq->fence.seqno, in print_request() 1415 hwsp_seqno(rq)); in intel_engine_print_registers() 1444 rq->head, rq->postfix, rq->tail, in print_request_ring() 1445 rq->batch ? upper_32_bits(rq->batch->node.start) : ~0u, in print_request_ring() 1446 rq->batch ? lower_32_bits(rq->batch->node.start) : ~0u); in print_request_ring() 1448 size = rq->tail - rq->head; in print_request_ring() 1449 if (rq->tail < rq->head) in print_request_ring() 1509 if (rq) in intel_engine_dump() [all …]
|
H A D | selftest_workarounds.c | 43 i915_request_get(rq); in request_add_sync() 44 i915_request_add(rq); in request_add_sync() 47 i915_request_put(rq); in request_add_sync() 56 i915_request_get(rq); in request_add_spin() 57 i915_request_add(rq); in request_add_spin() 60 i915_request_put(rq); in request_add_spin() 139 if (IS_ERR(rq)) { in read_nonprivs() 140 err = PTR_ERR(rq); in read_nonprivs() 283 if (IS_ERR(rq)) { in switch_to_scratch_context() 775 if (IS_ERR(rq)) in read_whitelisted_registers() [all …]
|
/netbsd/external/gpl2/lvm2/dist/daemons/cmirrord/ |
H A D | functions.c | 560 if (strlen(rq->data) > rq->data_size) { in clog_ctr() 566 rq->data, (int)strlen(rq->data)); in clog_ctr() 623 struct log_c *lc = get_log(rq->uuid, rq->luid); in clog_dtr() 659 struct log_c *lc = get_log(rq->uuid, rq->luid); in clog_presuspend() 857 lc = get_pending_log(rq->uuid, rq->luid); in local_resume() 892 r = create_cluster_cpg(rq->uuid, rq->luid); in local_resume() 1356 lc = get_pending_log(rq->uuid, rq->luid); in clog_get_sync_count() 1418 lc = get_pending_log(rq->uuid, rq->luid); in clog_status_info() 1473 lc = get_pending_log(rq->uuid, rq->luid); in clog_status_table() 1641 if (rq->u_rq.error && rq->u_rq.data_size) { in do_request() [all …]
|
H A D | cluster.c | 322 memcpy(orig_rq, rq, sizeof(*rq) + rq->u_rq.data_size); in handle_cluster_response() 577 if (!rq) { in export_checkpoint() 581 memset(rq, 0, sizeof(*rq)); in export_checkpoint() 594 free(rq); in export_checkpoint() 860 free(rq); in resend_requests() 932 free(rq); in flush_startup_list() 972 memcpy(tmp_rq, rq, sizeof(*rq) + rq->u_rq.data_size); in cpg_message_callback() 1078 memcpy(tmp_rq, rq, sizeof(*rq) + rq->u_rq.data_size); in cpg_message_callback() 1129 SHORT_UUID(rq->u_rq.uuid), rq->originator); in cpg_message_callback() 1147 rq->u_rq.seq, SHORT_UUID(rq->u_rq.uuid), in cpg_message_callback() [all …]
|
/netbsd/external/bsd/openldap/dist/libraries/libldap/ |
H A D | rq.c | 46 struct runqueue_s* rq, in ldap_pvt_runqueue_insert() argument 73 struct runqueue_s *rq, in ldap_pvt_runqueue_find() argument 89 struct runqueue_s* rq, in ldap_pvt_runqueue_remove() argument 109 struct runqueue_s* rq, in ldap_pvt_runqueue_next_sched() argument 115 entry = LDAP_STAILQ_FIRST( &rq->task_list ); in ldap_pvt_runqueue_next_sched() 126 struct runqueue_s* rq, in ldap_pvt_runqueue_runtask() argument 135 struct runqueue_s* rq, in ldap_pvt_runqueue_stoptask() argument 144 struct runqueue_s* rq, in ldap_pvt_runqueue_isrunning() argument 160 struct runqueue_s* rq, in ldap_pvt_runqueue_resched() argument 183 if ( LDAP_STAILQ_EMPTY( &rq->task_list )) { in ldap_pvt_runqueue_resched() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
H A D | igt_spinner.c | 78 struct i915_request *rq, in move_to_active() argument 99 struct i915_request *rq = NULL; in igt_spinner_create_request() local 127 if (IS_ERR(rq)) { in igt_spinner_create_request() 128 err = PTR_ERR(rq); in igt_spinner_create_request() 158 *batch++ = rq->fence.seqno; in igt_spinner_create_request() 162 if (INTEL_GEN(rq->i915) >= 8) in igt_spinner_create_request() 185 if (INTEL_GEN(rq->i915) <= 5) in igt_spinner_create_request() 191 i915_request_skip(rq, err); in igt_spinner_create_request() 192 i915_request_add(rq); in igt_spinner_create_request() 229 rq->fence.seqno), in igt_wait_for_spinner() [all …]
|
H A D | i915_perf.c | 141 if (IS_ERR(rq)) { in live_noa_delay() 142 err = PTR_ERR(rq); in live_noa_delay() 148 err = rq->engine->emit_init_breadcrumb(rq); in live_noa_delay() 150 i915_request_add(rq); in live_noa_delay() 157 i915_request_add(rq); in live_noa_delay() 161 err = rq->engine->emit_bb_start(rq, in live_noa_delay() 165 i915_request_add(rq); in live_noa_delay() 171 i915_request_add(rq); in live_noa_delay() 175 i915_request_get(rq); in live_noa_delay() 176 i915_request_add(rq); in live_noa_delay() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 226 u32 ring_tail = intel_ring_set_tail(rq->ring, rq->tail) / sizeof(u64); in guc_add_request() 258 guc_add_request(guc, rq); in guc_submit() 281 return i915_request_get(rq); in schedule_in() 286 trace_i915_request_out(rq); in schedule_out() 289 i915_request_put(rq); in schedule_out() 335 last = rq; in __guc_dequeue() 365 schedule_out(rq); in guc_submission_tasklet() 404 schedule_out(rq); in cancel_port_requests() 412 struct i915_request *rq; in guc_reset_rewind() local 421 if (!rq) in guc_reset_rewind() [all …]
|
/netbsd/external/lgpl3/gmp/dist/mini-gmp/tests/ |
H A D | t-mpq_muldiv_2exp.c | 54 mpq_t aq, rq, tq; in testmain() local 62 mpq_init (rq); in testmain() 75 mpq_mul_2exp (rq, aq, e); in testmain() 78 mpq_neg (tq, rq); in testmain() 93 mpq_div_2exp (rq, aq, e); in testmain() 96 mpq_div (aq, aq, rq); in testmain() 112 mpq_set_ui (rq, 6, 7); in testmain() 116 if (!mpq_equal (tq, rq)) in testmain() 122 mpq_set_ui (rq, 7, 6); in testmain() 125 if (!mpq_equal (rq, tq)) in testmain() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/scheduler/ |
H A D | sched_entity.c | 69 entity->rq = NULL; in drm_sched_entity_init() 137 struct drm_sched_rq *rq = NULL; in drm_sched_entity_get_free_sched() local 156 return rq; in drm_sched_entity_get_free_sched() 181 if (!entity->rq) in drm_sched_entity_flush() 184 sched = entity->rq->sched; in drm_sched_entity_flush() 300 if (entity->rq) { in drm_sched_entity_fini() 301 sched = entity->rq->sched; in drm_sched_entity_fini() 375 drm_sched_wakeup(entity->rq->sched); in drm_sched_entity_wakeup() 494 struct drm_sched_rq *rq; in drm_sched_entity_select_rq() local 505 if (rq != entity->rq) { in drm_sched_entity_select_rq() [all …]
|
H A D | sched_main.c | 80 spin_lock_init(&rq->lock); in drm_sched_rq_init() 83 rq->sched = sched; in drm_sched_rq_init() 99 spin_lock(&rq->lock); in drm_sched_rq_add_entity() 102 spin_unlock(&rq->lock); in drm_sched_rq_add_entity() 118 spin_lock(&rq->lock); in drm_sched_rq_remove_entity() 123 spin_unlock(&rq->lock); in drm_sched_rq_remove_entity() 138 spin_lock(&rq->lock); in drm_sched_rq_select_entity() 157 spin_unlock(&rq->lock); in drm_sched_rq_select_entity() 165 spin_unlock(&rq->lock); in drm_sched_rq_select_entity() 363 spin_lock(&rq->lock); in drm_sched_increase_karma() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gem/ |
H A D | i915_gem_object_blt.c | 136 struct i915_request *rq; in i915_gem_object_fill_blt() local 162 if (IS_ERR(rq)) { in i915_gem_object_fill_blt() 163 err = PTR_ERR(rq); in i915_gem_object_fill_blt() 194 i915_request_skip(rq, err); in i915_gem_object_fill_blt() 196 i915_request_add(rq); in i915_gem_object_fill_blt() 327 struct i915_request *rq; in i915_gem_object_copy_blt() local 353 if (IS_ERR(rq)) { in i915_gem_object_copy_blt() 354 err = PTR_ERR(rq); in i915_gem_object_copy_blt() 381 err = rq->engine->emit_init_breadcrumb(rq); in i915_gem_object_copy_blt() 386 err = rq->engine->emit_bb_start(rq, in i915_gem_object_copy_blt() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
H A D | igt_gem_utils.c | 27 struct i915_request *rq; in igt_request_alloc() local 38 rq = intel_context_create_request(ce); in igt_request_alloc() 41 return rq; in igt_request_alloc() 116 struct i915_request *rq; in igt_gpu_fill_dw() local 128 rq = intel_context_create_request(ce); in igt_gpu_fill_dw() 129 if (IS_ERR(rq)) { in igt_gpu_fill_dw() 130 err = PTR_ERR(rq); in igt_gpu_fill_dw() 138 err = rq->engine->emit_bb_start(rq, in igt_gpu_fill_dw() 160 i915_request_add(rq); in igt_gpu_fill_dw() 167 i915_request_skip(rq, err); in igt_gpu_fill_dw() [all …]
|
H A D | i915_gem_context.c | 95 if (rq) { in live_nop_switch() 124 rq = NULL; in live_nop_switch() 156 GEM_BUG_ON(!rq); in live_nop_switch() 210 if (IS_ERR(rq)) { in __live_parallel_switch1() 251 if (IS_ERR(rq)) { in __live_parallel_switchN() 975 if (IS_ERR(rq)) { in emit_rpcs_query() 980 err = rq->engine->emit_bb_start(rq, in emit_rpcs_query() 1049 if (IS_ERR(rq)) { in __sseu_prepare() 1541 if (IS_ERR(rq)) { in write_to_scratch() 1643 if (IS_ERR(rq)) { in read_from_scratch() [all …]
|
/netbsd/external/bsd/openldap/dist/servers/slapd/overlays/ |
H A D | refint.c | 471 rq->attrs = ip; in refint_search_cb() 584 refint_q *rq ) in refint_repair() argument 769 refint_q *rq; in refint_qtask() local 814 rq = id->qhead; in refint_qtask() 815 if ( rq ) { in refint_qtask() 821 if ( !rq ) in refint_qtask() 905 ch_free( rq ); in refint_qtask() 946 refint_q *rq; in refint_response() local 962 rq->db = id->db; in refint_response() 963 rq->rdata = id; in refint_response() [all …]
|