Home
last modified time | relevance | path

Searched refs:rq (Results 1 – 25 of 416) sorted by relevance

12345678910>>...17

/netbsd/sys/external/bsd/drm2/dist/drm/i915/
H A Di915_request.c257 rq->ring->head = rq->postfix; in i915_request_retire()
582 if (rq) in request_alloc_slow()
738 rq->head = rq->ring->emit; in __i915_request_create()
740 ret = rq->engine->request_alloc(rq); in __i915_request_create()
747 return rq; in __i915_request_create()
787 return rq; in i915_request_create()
791 return rq; in i915_request_create()
1207 if (rq->infix == rq->postfix) in i915_request_skip()
1221 rq->infix = rq->postfix; in i915_request_skip()
1310 rq->postfix = intel_ring_offset(rq, cs); in __i915_request_commit()
[all …]
H A Di915_request.h54 #define RQ_TRACE(rq, fmt, ...) do { \ argument
343 dma_fence_put(&rq->fence); in i915_request_put()
404 return READ_ONCE(*rq->hwsp_seqno); in __hwsp_seqno()
425 seqno = __hwsp_seqno(rq); in hwsp_seqno()
433 return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno - 1); in __i915_request_has_started()
464 if (i915_request_signaled(rq)) in i915_request_started()
481 if (!i915_request_is_active(rq)) in i915_request_is_running()
505 return !list_empty(&rq->sched.link); in i915_request_is_ready()
510 if (i915_request_signaled(rq)) in i915_request_completed()
513 return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno); in i915_request_completed()
[all …]
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gt/
H A Dselftest_lrc.c234 GEM_BUG_ON(rq[1]->postfix <= rq[0]->postfix); in live_unlite_restore()
254 i915_request_await_dma_fence(rq[0], &rq[1]->fence); in live_unlite_restore()
257 GEM_BUG_ON(rq[0]->postfix > rq[1]->postfix); in live_unlite_restore()
448 err = rq->engine->emit_init_breadcrumb(rq); in semaphore_queue()
1772 rq = kzalloc(sizeof(*rq), GFP_KERNEL); in dummy_request()
1780 rq->fence.lock = &rq->lock; in dummy_request()
1788 rq->hwsp_seqno = (u32 *)&rq->fence.seqno + 1; in dummy_request()
1795 rq->fence.lock = &rq->lock; in dummy_request()
1982 ring_size = rq->wa_tail - rq->head; in live_chain_preempt()
2162 err = rq->engine->emit_bb_start(rq, in create_gang()
[all …]
H A Dintel_breadcrumbs.c94 return i915_seqno_passed(__hwsp_seqno(rq), rq->fence.seqno); in __request_completed()
107 rq->fence.seqno)) in check_signal_order()
173 struct i915_request *rq = in signal_irq_work() local
182 &rq->fence.flags)); in signal_irq_work()
193 i915_request_get(rq); in signal_irq_work()
215 struct i915_request *rq = in signal_irq_work() local
222 spin_lock(&rq->lock); in signal_irq_work()
226 spin_unlock(&rq->lock); in signal_irq_work()
229 i915_request_put(rq); in signal_irq_work()
364 list_del(&rq->signal_link); in i915_request_cancel_breadcrumb()
[all …]
H A Dselftest_hangcheck.c192 if (IS_ERR(rq)) { in hang_create_request()
263 err = rq->engine->emit_init_breadcrumb(rq); in hang_create_request()
607 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in __igt_reset_engine()
685 if (!rq) in active_request_put()
859 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in __igt_reset_engines()
878 if (rq) { in __igt_reset_engines()
1062 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in igt_reset_wait()
1246 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in __igt_reset_evict_vma()
1473 prev = rq; in igt_reset_queue()
1534 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in igt_handle_error()
[all …]
H A Dintel_ring_submission.c125 if (IS_G4X(rq->i915) || IS_GEN(rq->i915, 5)) in gen4_render_ring_flush()
329 rq->tail = intel_ring_offset(rq, cs); in gen6_rcs_emit_breadcrumb()
330 assert_ring_tail_valid(rq->ring, rq->tail); in gen6_rcs_emit_breadcrumb()
431 rq->tail = intel_ring_offset(rq, cs); in gen7_rcs_emit_breadcrumb()
432 assert_ring_tail_valid(rq->ring, rq->tail); in gen7_rcs_emit_breadcrumb()
448 rq->tail = intel_ring_offset(rq, cs); in gen6_xcs_emit_breadcrumb()
449 assert_ring_tail_valid(rq->ring, rq->tail); in gen6_xcs_emit_breadcrumb()
480 rq->tail = intel_ring_offset(rq, cs); in gen7_xcs_emit_breadcrumb()
481 assert_ring_tail_valid(rq->ring, rq->tail); in gen7_xcs_emit_breadcrumb()
966 rq->tail = intel_ring_offset(rq, cs); in i9xx_emit_breadcrumb()
[all …]
H A Dintel_engine_heartbeat.c69 struct i915_request *rq; in heartbeat() local
72 if (rq && i915_request_completed(rq)) { in heartbeat()
73 i915_request_put(rq); in heartbeat()
129 if (IS_ERR(rq)) in heartbeat()
132 idle_pulse(engine, rq); in heartbeat()
210 if (IS_ERR(rq)) { in intel_engine_pulse()
211 err = PTR_ERR(rq); in intel_engine_pulse()
216 idle_pulse(engine, rq); in intel_engine_pulse()
240 if (IS_ERR(rq)) { in intel_engine_flush_barriers()
241 err = PTR_ERR(rq); in intel_engine_flush_barriers()
[all …]
H A Dintel_lrc.c1231 rq->fence.context, rq->fence.seqno); in reset_active()
1391 tail = intel_ring_set_tail(rq->ring, rq->tail); in execlists_update_context()
1396 rq->tail = rq->wa_tail; in execlists_update_context()
1559 rq ? execlists_update_context(rq) : 0, in execlists_submit_ports()
1730 if (!rq) in defer_active()
1773 if (!rq || i915_request_completed(rq)) in active_timeslice()
2689 cap->rq = active_request(cap->rq->context->timeline, cap->rq); in execlists_capture()
2690 cap->rq = i915_request_get_rcu(cap->rq); in execlists_capture()
3057 rq->infix = intel_ring_offset(rq, cs); in gen8_emit_init_breadcrumb()
3601 rq = active_request(ce->timeline, rq); in __execlists_reset()
[all …]
H A Dselftest_mocs.c27 i915_request_get(rq); in request_add_sync()
28 i915_request_add(rq); in request_add_sync()
31 i915_request_put(rq); in request_add_sync()
40 i915_request_get(rq); in request_add_spin()
41 i915_request_add(rq); in request_add_spin()
44 i915_request_put(rq); in request_add_spin()
211 struct i915_request *rq; in check_mocs_engine() local
219 if (IS_ERR(rq)) in check_mocs_engine()
220 return PTR_ERR(rq); in check_mocs_engine()
325 if (IS_ERR(rq)) { in active_engine_reset()
[all …]
H A Dselftest_timeline.c462 rq = ERR_PTR(err); in tl_write()
467 if (IS_ERR(rq)) in tl_write()
470 i915_request_get(rq); in tl_write()
476 rq = ERR_PTR(err); in tl_write()
482 if (IS_ERR(rq)) in tl_write()
484 return rq; in tl_write()
545 if (IS_ERR(rq)) { in live_hwsp_engine()
620 if (IS_ERR(rq)) { in live_hwsp_alternate()
684 if (IS_ERR(rq)) { in live_hwsp_wrap()
685 err = PTR_ERR(rq); in live_hwsp_wrap()
[all …]
H A Dintel_engine_cs.c1192 const char *name = rq->fence.ops->get_timeline_name(&rq->fence); in print_request()
1196 x = print_sched_attr(rq->i915, &rq->sched.attr, buf, x, sizeof(buf)); in print_request()
1200 (uint64_t)rq->fence.context, (uint64_t)rq->fence.seqno, in print_request()
1415 hwsp_seqno(rq)); in intel_engine_print_registers()
1444 rq->head, rq->postfix, rq->tail, in print_request_ring()
1445 rq->batch ? upper_32_bits(rq->batch->node.start) : ~0u, in print_request_ring()
1446 rq->batch ? lower_32_bits(rq->batch->node.start) : ~0u); in print_request_ring()
1448 size = rq->tail - rq->head; in print_request_ring()
1449 if (rq->tail < rq->head) in print_request_ring()
1509 if (rq) in intel_engine_dump()
[all …]
H A Dselftest_workarounds.c43 i915_request_get(rq); in request_add_sync()
44 i915_request_add(rq); in request_add_sync()
47 i915_request_put(rq); in request_add_sync()
56 i915_request_get(rq); in request_add_spin()
57 i915_request_add(rq); in request_add_spin()
60 i915_request_put(rq); in request_add_spin()
139 if (IS_ERR(rq)) { in read_nonprivs()
140 err = PTR_ERR(rq); in read_nonprivs()
283 if (IS_ERR(rq)) { in switch_to_scratch_context()
775 if (IS_ERR(rq)) in read_whitelisted_registers()
[all …]
/netbsd/external/gpl2/lvm2/dist/daemons/cmirrord/
H A Dfunctions.c560 if (strlen(rq->data) > rq->data_size) { in clog_ctr()
566 rq->data, (int)strlen(rq->data)); in clog_ctr()
623 struct log_c *lc = get_log(rq->uuid, rq->luid); in clog_dtr()
659 struct log_c *lc = get_log(rq->uuid, rq->luid); in clog_presuspend()
857 lc = get_pending_log(rq->uuid, rq->luid); in local_resume()
892 r = create_cluster_cpg(rq->uuid, rq->luid); in local_resume()
1356 lc = get_pending_log(rq->uuid, rq->luid); in clog_get_sync_count()
1418 lc = get_pending_log(rq->uuid, rq->luid); in clog_status_info()
1473 lc = get_pending_log(rq->uuid, rq->luid); in clog_status_table()
1641 if (rq->u_rq.error && rq->u_rq.data_size) { in do_request()
[all …]
H A Dcluster.c322 memcpy(orig_rq, rq, sizeof(*rq) + rq->u_rq.data_size); in handle_cluster_response()
577 if (!rq) { in export_checkpoint()
581 memset(rq, 0, sizeof(*rq)); in export_checkpoint()
594 free(rq); in export_checkpoint()
860 free(rq); in resend_requests()
932 free(rq); in flush_startup_list()
972 memcpy(tmp_rq, rq, sizeof(*rq) + rq->u_rq.data_size); in cpg_message_callback()
1078 memcpy(tmp_rq, rq, sizeof(*rq) + rq->u_rq.data_size); in cpg_message_callback()
1129 SHORT_UUID(rq->u_rq.uuid), rq->originator); in cpg_message_callback()
1147 rq->u_rq.seq, SHORT_UUID(rq->u_rq.uuid), in cpg_message_callback()
[all …]
/netbsd/external/bsd/openldap/dist/libraries/libldap/
H A Drq.c46 struct runqueue_s* rq, in ldap_pvt_runqueue_insert() argument
73 struct runqueue_s *rq, in ldap_pvt_runqueue_find() argument
89 struct runqueue_s* rq, in ldap_pvt_runqueue_remove() argument
109 struct runqueue_s* rq, in ldap_pvt_runqueue_next_sched() argument
115 entry = LDAP_STAILQ_FIRST( &rq->task_list ); in ldap_pvt_runqueue_next_sched()
126 struct runqueue_s* rq, in ldap_pvt_runqueue_runtask() argument
135 struct runqueue_s* rq, in ldap_pvt_runqueue_stoptask() argument
144 struct runqueue_s* rq, in ldap_pvt_runqueue_isrunning() argument
160 struct runqueue_s* rq, in ldap_pvt_runqueue_resched() argument
183 if ( LDAP_STAILQ_EMPTY( &rq->task_list )) { in ldap_pvt_runqueue_resched()
[all …]
/netbsd/sys/external/bsd/drm2/dist/drm/i915/selftests/
H A Digt_spinner.c78 struct i915_request *rq, in move_to_active() argument
99 struct i915_request *rq = NULL; in igt_spinner_create_request() local
127 if (IS_ERR(rq)) { in igt_spinner_create_request()
128 err = PTR_ERR(rq); in igt_spinner_create_request()
158 *batch++ = rq->fence.seqno; in igt_spinner_create_request()
162 if (INTEL_GEN(rq->i915) >= 8) in igt_spinner_create_request()
185 if (INTEL_GEN(rq->i915) <= 5) in igt_spinner_create_request()
191 i915_request_skip(rq, err); in igt_spinner_create_request()
192 i915_request_add(rq); in igt_spinner_create_request()
229 rq->fence.seqno), in igt_wait_for_spinner()
[all …]
H A Di915_perf.c141 if (IS_ERR(rq)) { in live_noa_delay()
142 err = PTR_ERR(rq); in live_noa_delay()
148 err = rq->engine->emit_init_breadcrumb(rq); in live_noa_delay()
150 i915_request_add(rq); in live_noa_delay()
157 i915_request_add(rq); in live_noa_delay()
161 err = rq->engine->emit_bb_start(rq, in live_noa_delay()
165 i915_request_add(rq); in live_noa_delay()
171 i915_request_add(rq); in live_noa_delay()
175 i915_request_get(rq); in live_noa_delay()
176 i915_request_add(rq); in live_noa_delay()
[all …]
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gt/uc/
H A Dintel_guc_submission.c226 u32 ring_tail = intel_ring_set_tail(rq->ring, rq->tail) / sizeof(u64); in guc_add_request()
258 guc_add_request(guc, rq); in guc_submit()
281 return i915_request_get(rq); in schedule_in()
286 trace_i915_request_out(rq); in schedule_out()
289 i915_request_put(rq); in schedule_out()
335 last = rq; in __guc_dequeue()
365 schedule_out(rq); in guc_submission_tasklet()
404 schedule_out(rq); in cancel_port_requests()
412 struct i915_request *rq; in guc_reset_rewind() local
421 if (!rq) in guc_reset_rewind()
[all …]
/netbsd/external/lgpl3/gmp/dist/mini-gmp/tests/
H A Dt-mpq_muldiv_2exp.c54 mpq_t aq, rq, tq; in testmain() local
62 mpq_init (rq); in testmain()
75 mpq_mul_2exp (rq, aq, e); in testmain()
78 mpq_neg (tq, rq); in testmain()
93 mpq_div_2exp (rq, aq, e); in testmain()
96 mpq_div (aq, aq, rq); in testmain()
112 mpq_set_ui (rq, 6, 7); in testmain()
116 if (!mpq_equal (tq, rq)) in testmain()
122 mpq_set_ui (rq, 7, 6); in testmain()
125 if (!mpq_equal (rq, tq)) in testmain()
[all …]
/netbsd/sys/external/bsd/drm2/dist/drm/scheduler/
H A Dsched_entity.c69 entity->rq = NULL; in drm_sched_entity_init()
137 struct drm_sched_rq *rq = NULL; in drm_sched_entity_get_free_sched() local
156 return rq; in drm_sched_entity_get_free_sched()
181 if (!entity->rq) in drm_sched_entity_flush()
184 sched = entity->rq->sched; in drm_sched_entity_flush()
300 if (entity->rq) { in drm_sched_entity_fini()
301 sched = entity->rq->sched; in drm_sched_entity_fini()
375 drm_sched_wakeup(entity->rq->sched); in drm_sched_entity_wakeup()
494 struct drm_sched_rq *rq; in drm_sched_entity_select_rq() local
505 if (rq != entity->rq) { in drm_sched_entity_select_rq()
[all …]
H A Dsched_main.c80 spin_lock_init(&rq->lock); in drm_sched_rq_init()
83 rq->sched = sched; in drm_sched_rq_init()
99 spin_lock(&rq->lock); in drm_sched_rq_add_entity()
102 spin_unlock(&rq->lock); in drm_sched_rq_add_entity()
118 spin_lock(&rq->lock); in drm_sched_rq_remove_entity()
123 spin_unlock(&rq->lock); in drm_sched_rq_remove_entity()
138 spin_lock(&rq->lock); in drm_sched_rq_select_entity()
157 spin_unlock(&rq->lock); in drm_sched_rq_select_entity()
165 spin_unlock(&rq->lock); in drm_sched_rq_select_entity()
363 spin_lock(&rq->lock); in drm_sched_increase_karma()
[all …]
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gem/
H A Di915_gem_object_blt.c136 struct i915_request *rq; in i915_gem_object_fill_blt() local
162 if (IS_ERR(rq)) { in i915_gem_object_fill_blt()
163 err = PTR_ERR(rq); in i915_gem_object_fill_blt()
194 i915_request_skip(rq, err); in i915_gem_object_fill_blt()
196 i915_request_add(rq); in i915_gem_object_fill_blt()
327 struct i915_request *rq; in i915_gem_object_copy_blt() local
353 if (IS_ERR(rq)) { in i915_gem_object_copy_blt()
354 err = PTR_ERR(rq); in i915_gem_object_copy_blt()
381 err = rq->engine->emit_init_breadcrumb(rq); in i915_gem_object_copy_blt()
386 err = rq->engine->emit_bb_start(rq, in i915_gem_object_copy_blt()
[all …]
/netbsd/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/
H A Digt_gem_utils.c27 struct i915_request *rq; in igt_request_alloc() local
38 rq = intel_context_create_request(ce); in igt_request_alloc()
41 return rq; in igt_request_alloc()
116 struct i915_request *rq; in igt_gpu_fill_dw() local
128 rq = intel_context_create_request(ce); in igt_gpu_fill_dw()
129 if (IS_ERR(rq)) { in igt_gpu_fill_dw()
130 err = PTR_ERR(rq); in igt_gpu_fill_dw()
138 err = rq->engine->emit_bb_start(rq, in igt_gpu_fill_dw()
160 i915_request_add(rq); in igt_gpu_fill_dw()
167 i915_request_skip(rq, err); in igt_gpu_fill_dw()
[all …]
H A Di915_gem_context.c95 if (rq) { in live_nop_switch()
124 rq = NULL; in live_nop_switch()
156 GEM_BUG_ON(!rq); in live_nop_switch()
210 if (IS_ERR(rq)) { in __live_parallel_switch1()
251 if (IS_ERR(rq)) { in __live_parallel_switchN()
975 if (IS_ERR(rq)) { in emit_rpcs_query()
980 err = rq->engine->emit_bb_start(rq, in emit_rpcs_query()
1049 if (IS_ERR(rq)) { in __sseu_prepare()
1541 if (IS_ERR(rq)) { in write_to_scratch()
1643 if (IS_ERR(rq)) { in read_from_scratch()
[all …]
/netbsd/external/bsd/openldap/dist/servers/slapd/overlays/
H A Drefint.c471 rq->attrs = ip; in refint_search_cb()
584 refint_q *rq ) in refint_repair() argument
769 refint_q *rq; in refint_qtask() local
814 rq = id->qhead; in refint_qtask()
815 if ( rq ) { in refint_qtask()
821 if ( !rq ) in refint_qtask()
905 ch_free( rq ); in refint_qtask()
946 refint_q *rq; in refint_response() local
962 rq->db = id->db; in refint_response()
963 rq->rdata = id; in refint_response()
[all …]

12345678910>>...17