Home
last modified time | relevance | path

Searched refs:rhp (Results 1 – 25 of 40) sorted by relevance

12

/linux/include/trace/events/
H A Drcu.h520 __field(void *, rhp)
527 __entry->rhp = rhp;
580 __field(void *, rhp)
587 __entry->rhp = rhp;
639 __field(void *, rhp)
645 __entry->rhp = rhp;
668 __field(void *, rhp)
674 __entry->rhp = rhp;
723 __field(void *, rhp)
729 __entry->rhp = rhp;
[all …]
/linux/drivers/infiniband/hw/cxgb4/
H A Dmem.c432 struct c4iw_dev *rhp; in c4iw_get_dma_mr() local
440 rhp = php->rhp; in c4iw_get_dma_mr()
458 mhp->rhp = rhp; in c4iw_get_dma_mr()
498 struct c4iw_dev *rhp; in c4iw_reg_user_mr() local
511 rhp = php->rhp; in c4iw_reg_user_mr()
527 mhp->rhp = rhp; in c4iw_reg_user_mr()
601 struct c4iw_dev *rhp; in c4iw_alloc_mr() local
610 rhp = php->rhp; in c4iw_alloc_mr()
638 mhp->rhp = rhp; in c4iw_alloc_mr()
702 struct c4iw_dev *rhp; in c4iw_dereg_mr() local
[all …]
H A Dprovider.c65 struct c4iw_dev *rhp; in c4iw_dealloc_ucontext() local
196 struct c4iw_dev *rhp; in c4iw_deallocate_pd() local
200 rhp = php->rhp; in c4iw_deallocate_pd()
203 mutex_lock(&rhp->rdev.stats.lock); in c4iw_deallocate_pd()
204 rhp->rdev.stats.pd.cur--; in c4iw_deallocate_pd()
214 struct c4iw_dev *rhp; in c4iw_allocate_pd() local
217 rhp = (struct c4iw_dev *) ibdev; in c4iw_allocate_pd()
223 php->rhp = rhp; in c4iw_allocate_pd()
233 rhp->rdev.stats.pd.cur++; in c4iw_allocate_pd()
234 if (rhp->rdev.stats.pd.cur > rhp->rdev.stats.pd.max) in c4iw_allocate_pd()
[all …]
H A Dqp.c1085 struct c4iw_dev *rhp; in c4iw_post_send() local
1093 rhp = qhp->rhp; in c4iw_post_send()
2072 struct c4iw_dev *rhp; in c4iw_destroy_qp() local
2078 rhp = qhp->rhp; in c4iw_destroy_qp()
2113 struct c4iw_dev *rhp; in c4iw_create_qp() local
2130 rhp = php->rhp; in c4iw_create_qp()
2186 qhp->rhp = rhp; in c4iw_create_qp()
2378 rhp = qhp->rhp; in c4iw_ib_modify_qp()
2682 rhp = php->rhp; in c4iw_create_srq()
2717 srq->rhp = rhp; in c4iw_create_srq()
[all …]
H A Dcq.c351 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); in c4iw_flush_hw_cq()
844 c4iw_invalidate_mr(qhp->rhp, in __c4iw_poll_cq_one()
933 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); in c4iw_poll_cq_one()
984 xa_erase_irq(&chp->rhp->cqs, chp->cq.cqid); in c4iw_destroy_cq()
990 destroy_cq(&chp->rhp->rdev, &chp->cq, in c4iw_destroy_cq()
1020 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq()
1078 ret = create_cq(&rhp->rdev, &chp->cq, in c4iw_create_cq()
1084 chp->rhp = rhp; in c4iw_create_cq()
1105 uresp.qid_mask = rhp->rdev.cqmask; in c4iw_create_cq()
1147 xa_erase_irq(&rhp->cqs, chp->cq.cqid); in c4iw_create_cq()
[all …]
H A Diw_cxgb4.h343 static inline struct c4iw_cq *get_chp(struct c4iw_dev *rhp, u32 cqid) in get_chp() argument
345 return xa_load(&rhp->cqs, cqid); in get_chp()
350 return xa_load(&rhp->qps, qpid); in get_qhp()
363 struct c4iw_dev *rhp; member
392 struct c4iw_dev *rhp; member
410 struct c4iw_dev *rhp; member
424 struct c4iw_dev *rhp; member
481 struct c4iw_dev *rhp; member
504 struct c4iw_dev *rhp; member
597 int c4iw_modify_qp(struct c4iw_dev *rhp,
[all …]
H A Ddevice.c1376 xa_lock_irq(&qp->rhp->qps); in recover_lost_dbs()
1378 ret = cxgb4_sync_txq_pidx(qp->rhp->rdev.lldi.ports[0], in recover_lost_dbs()
1386 xa_unlock_irq(&qp->rhp->qps); in recover_lost_dbs()
1391 ret = cxgb4_sync_txq_pidx(qp->rhp->rdev.lldi.ports[0], in recover_lost_dbs()
1400 xa_unlock_irq(&qp->rhp->qps); in recover_lost_dbs()
1405 xa_unlock_irq(&qp->rhp->qps); in recover_lost_dbs()
1408 while (cxgb4_dbfifo_count(qp->rhp->rdev.lldi.ports[0], 1) > 0) { in recover_lost_dbs()
/linux/kernel/rcu/
H A Drcu_segcblist.c30 *rclp->tail = rhp; in rcu_cblist_enqueue()
31 rclp->tail = &rhp->next; in rcu_cblist_enqueue()
53 if (!rhp) { in rcu_cblist_flush_enqueue()
56 rhp->next = NULL; in rcu_cblist_flush_enqueue()
57 srclp->head = rhp; in rcu_cblist_flush_enqueue()
69 struct rcu_head *rhp; in rcu_cblist_dequeue() local
71 rhp = rclp->head; in rcu_cblist_dequeue()
72 if (!rhp) in rcu_cblist_dequeue()
78 return rhp; in rcu_cblist_dequeue()
345 rhp->next = NULL; in rcu_segcblist_enqueue()
[all …]
H A Dsrcutiny.c119 struct rcu_head *rhp; in srcu_drive_gp() local
148 rhp = lh; in srcu_drive_gp()
150 debug_rcu_head_callback(rhp); in srcu_drive_gp()
152 rhp->func(rhp); in srcu_drive_gp()
195 void call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in call_srcu() argument
200 rhp->func = func; in call_srcu()
201 rhp->next = NULL; in call_srcu()
204 *ssp->srcu_cb_tail = rhp; in call_srcu()
205 ssp->srcu_cb_tail = &rhp->next; in call_srcu()
H A Dupdate.c516 void do_trace_rcu_torture_read(const char *rcutorturename, struct rcu_head *rhp, in do_trace_rcu_torture_read() argument
520 trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c); in do_trace_rcu_torture_read()
524 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
610 struct early_boot_kfree_rcu *rhp; in early_boot_test_call_rcu() local
617 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in early_boot_test_call_rcu()
618 if (!WARN_ON_ONCE(!rhp)) in early_boot_test_call_rcu()
619 kfree_rcu(rhp, rh); in early_boot_test_call_rcu()
H A Drcu_segcblist.h24 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp);
27 struct rcu_head *rhp);
139 struct rcu_head *rhp);
141 struct rcu_head *rhp);
H A Drcutorture.c1722 kfree(rhp); in rcu_torture_timer_cb()
2074 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_NOWAIT); in rcu_torture_timer() local
2076 if (rhp) in rcu_torture_timer()
2330 mem_dump_obj(rhp); in rcu_torture_mem_dump_obj()
2337 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in rcu_torture_mem_dump_obj()
2342 mem_dump_obj(rhp); in rcu_torture_mem_dump_obj()
2345 kfree(rhp); in rcu_torture_mem_dump_obj()
2354 vfree(rhp); in rcu_torture_mem_dump_obj()
3488 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in rcu_test_debug_objects() local
3501 if (rhp) { in rcu_test_debug_objects()
[all …]
H A Dtasks.h336 rhp->next = NULL; in call_rcu_tasks_generic()
337 rhp->func = func; in call_rcu_tasks_generic()
367 rcu_segcblist_enqueue(&rtpcp->cblist, rhp); in call_rcu_tasks_generic()
519 struct rcu_head *rhp; in rcu_tasks_invoke_cbs() local
544 for (rhp = rcu_cblist_dequeue(&rcl); rhp; rhp = rcu_cblist_dequeue(&rcl)) { in rcu_tasks_invoke_cbs()
545 debug_rcu_head_callback(rhp); in rcu_tasks_invoke_cbs()
547 rhp->func(rhp); in rcu_tasks_invoke_cbs()
1112 call_rcu_tasks_generic(rhp, func, &rcu_tasks); in call_rcu_tasks()
1305 call_rcu_tasks_generic(rhp, func, &rcu_tasks_rude); in call_rcu_tasks_rude()
1951 call_rcu_tasks_generic(rhp, func, &rcu_tasks_trace); in call_rcu_tasks_trace()
[all …]
H A Dsrcutree.c1234 if (rhp) in srcu_gp_start_if_needed()
1273 if (rhp) { in srcu_gp_start_if_needed()
1340 if (debug_rcu_head_queue(rhp)) { in __call_srcu()
1346 rhp->func = func; in __call_srcu()
1370 __call_srcu(ssp, rhp, func, true); in call_srcu()
1722 struct rcu_head *rhp; in srcu_invoke_callbacks() local
1750 rhp = rcu_cblist_dequeue(&ready_cbs); in srcu_invoke_callbacks()
1751 for (; rhp != NULL; rhp = rcu_cblist_dequeue(&ready_cbs)) { in srcu_invoke_callbacks()
1752 debug_rcu_head_unqueue(rhp); in srcu_invoke_callbacks()
1753 debug_rcu_head_callback(rhp); in srcu_invoke_callbacks()
[all …]
H A Dsync.c27 static void rcu_sync_func(struct rcu_head *rhp);
57 static void rcu_sync_func(struct rcu_head *rhp) in rcu_sync_func() argument
59 struct rcu_sync *rsp = container_of(rhp, struct rcu_sync, cb_head); in rcu_sync_func()
H A Drcu.h252 static inline void debug_rcu_head_callback(struct rcu_head *rhp) in debug_rcu_head_callback() argument
254 if (unlikely(!rhp->func)) in debug_rcu_head_callback()
255 kmem_dump_obj(rhp); in debug_rcu_head_callback()
568 struct rcu_head *rhp,
581 struct rcu_head *rhp,
586 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
H A Drcuscale.c442 static void rcu_scale_async_cb(struct rcu_head *rhp) in rcu_scale_async_cb() argument
445 kfree(rhp); in rcu_scale_async_cb()
458 struct rcu_head *rhp = NULL; in rcu_scale_writer() local
503 if (!rhp) in rcu_scale_writer()
504 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in rcu_scale_writer()
505 if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) { in rcu_scale_writer()
507 cur_ops->async(rhp, rcu_scale_async_cb); in rcu_scale_writer()
508 rhp = NULL; in rcu_scale_writer()
513 kfree(rhp); /* Because we are stopping. */ in rcu_scale_writer()
H A Dtree_nocb.h334 struct rcu_head *rhp = rhp_in; in rcu_nocb_do_flush_bypass() local
339 if (rhp && !rcu_cblist_n_cbs(&rdp->nocb_bypass)) { in rcu_nocb_do_flush_bypass()
344 if (rhp) in rcu_nocb_do_flush_bypass()
353 if (lazy && rhp) { in rcu_nocb_do_flush_bypass()
354 rcu_cblist_enqueue(&rdp->nocb_bypass, rhp); in rcu_nocb_do_flush_bypass()
355 rhp = NULL; in rcu_nocb_do_flush_bypass()
357 rcu_cblist_flush_enqueue(&rcl, &rdp->nocb_bypass, rhp); in rcu_nocb_do_flush_bypass()
381 return rcu_nocb_do_flush_bypass(rdp, rhp, j, lazy); in rcu_nocb_flush_bypass()
415 static bool rcu_nocb_try_bypass(struct rcu_data *rdp, struct rcu_head *rhp, in rcu_nocb_try_bypass() argument
490 if (!rcu_nocb_flush_bypass(rdp, rhp, j, lazy)) { in rcu_nocb_try_bypass()
[all …]
H A Dtiny.c162 static void tiny_rcu_leak_callback(struct rcu_head *rhp) in tiny_rcu_leak_callback() argument
H A Dtree.c2469 struct rcu_head *rhp; in rcu_do_batch() local
2521 rhp = rcu_cblist_dequeue(&rcl); in rcu_do_batch()
2523 for (; rhp; rhp = rcu_cblist_dequeue(&rcl)) { in rcu_do_batch()
2527 debug_rcu_head_unqueue(rhp); in rcu_do_batch()
2530 trace_rcu_invoke_callback(rcu_state.name, rhp); in rcu_do_batch()
2532 f = rhp->func; in rcu_do_batch()
2533 debug_rcu_head_callback(rhp); in rcu_do_batch()
2534 WRITE_ONCE(rhp->func, (rcu_callback_t)0L); in rcu_do_batch()
2535 f(rhp); in rcu_do_batch()
3000 static void rcu_leak_callback(struct rcu_head *rhp) in rcu_leak_callback() argument
[all …]
/linux/include/linux/
H A Drcupdate_trace.h87 void call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func);
96 static inline void call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func) { BUG(); } in call_rcu_tasks_trace() argument
H A Drcupdate.h1060 static inline void rcu_head_init(struct rcu_head *rhp) in rcu_head_init() argument
1062 rhp->func = (rcu_callback_t)~0L; in rcu_head_init()
1079 rcu_head_after_call_rcu(struct rcu_head *rhp, rcu_callback_t f) in rcu_head_after_call_rcu() argument
1081 rcu_callback_t func = READ_ONCE(rhp->func); in rcu_head_after_call_rcu()
/linux/drivers/media/usb/pvrusb2/
H A Dpvrusb2-v4l2.c42 struct pvr2_ioread *rhp; member
915 if (fhp->rhp) { in pvr2_v4l2_release()
918 sp = pvr2_ioread_get_stream(fhp->rhp); in pvr2_v4l2_release()
920 pvr2_ioread_destroy(fhp->rhp); in pvr2_v4l2_release()
921 fhp->rhp = NULL; in pvr2_v4l2_release()
1048 if (fh->rhp) return 0; in pvr2_v4l2_iosetup()
1065 if (!fh->rhp) { in pvr2_v4l2_iosetup()
1118 if (!fh->rhp) { in pvr2_v4l2_read()
1133 pvr2_ioread_avail(fh->rhp) >= 0); in pvr2_v4l2_read()
1152 if (!fh->rhp) { in pvr2_v4l2_poll()
[all …]
/linux/include/linux/sched/
H A Dmm.h63 static inline void __mmdrop_delayed(struct rcu_head *rhp) in __mmdrop_delayed() argument
65 struct mm_struct *mm = container_of(rhp, struct mm_struct, delayed_drop); in __mmdrop_delayed()
H A Dtask.h123 extern void __put_task_struct_rcu_cb(struct rcu_head *rhp);

12