Home
last modified time | relevance | path

Searched refs:cb_list (Results 1 – 25 of 26) sorted by relevance

12

/linux/drivers/dma-buf/
H A Ddma-fence.c371 struct list_head cb_list; in dma_fence_signal_timestamp_locked() local
380 list_replace(&fence->cb_list, &cb_list); in dma_fence_signal_timestamp_locked()
386 list_for_each_entry_safe(cur, tmp, &cb_list, node) { in dma_fence_signal_timestamp_locked()
538 if (WARN(!list_empty(&fence->cb_list) && in dma_fence_release()
661 list_add_tail(&cb->node, &fence->cb_list); in dma_fence_add_callback()
784 list_add(&cb.base.node, &fence->cb_list); in dma_fence_default_wait()
1016 INIT_LIST_HEAD(&fence->cb_list); in dma_fence_init()
/linux/net/netfilter/
H A Dnf_tables_offload.c199 struct list_head *cb_list) in nft_setup_cb_call() argument
204 list_for_each_entry(block_cb, cb_list, list) { in nft_setup_cb_call()
281 &basechain->flow_block.cb_list); in nft_flow_offload_cmd()
317 list_splice(&bo->cb_list, &basechain->flow_block.cb_list); in nft_flow_offload_bind()
335 nft_setup_cb_call(TC_SETUP_CLSFLOWER, &cls_flow, &bo->cb_list); in nft_flow_offload_unbind()
338 list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) { in nft_flow_offload_unbind()
379 bo->cb_list_head = &basechain->flow_block.cb_list; in nft_flow_block_offload_init()
380 INIT_LIST_HEAD(&bo->cb_list); in nft_flow_block_offload_init()
414 list_move(&block_cb->list, &bo.cb_list); in nft_indr_block_cleanup()
434 if (list_empty(&bo.cb_list)) in nft_indr_block_offload_cmd()
H A Dnf_flow_table_offload.c879 &offload->flowtable->flow_block.cb_list); in flow_offload_tuple_add()
888 &offload->flowtable->flow_block.cb_list); in flow_offload_tuple_del()
942 &offload->flowtable->flow_block.cb_list); in flow_offload_tuple_stats()
1108 list_splice(&bo->cb_list, &flowtable->flow_block.cb_list); in nf_flow_table_block_setup()
1111 list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) { in nf_flow_table_block_setup()
1137 bo->cb_list_head = &flowtable->flow_block.cb_list; in nf_flow_table_block_offload_init()
1138 INIT_LIST_HEAD(&bo->cb_list); in nf_flow_table_block_offload_init()
/linux/drivers/dma/
H A Dbcm2835-dma.c95 struct bcm2835_cb_entry cb_list[]; member
209 dma_pool_free(desc->c->cb_pool, desc->cb_list[i].cb, in bcm2835_dma_free_cb_chain()
210 desc->cb_list[i].paddr); in bcm2835_dma_free_cb_chain()
317 d = kzalloc(struct_size(d, cb_list, frames), gfp); in bcm2835_dma_create_cb_chain()
330 cb_entry = &d->cb_list[frame]; in bcm2835_dma_create_cb_chain()
357 d->cb_list[frame - 1].cb->next = cb_entry->paddr; in bcm2835_dma_create_cb_chain()
370 d->cb_list[d->frames - 1].cb->info |= finalextrainfo; in bcm2835_dma_create_cb_chain()
452 writel(d->cb_list[0].paddr, c->chan_base + BCM2835_DMA_ADDR); in bcm2835_dma_start_desc()
544 struct bcm2835_dma_cb *control_block = d->cb_list[i].cb; in bcm2835_dma_desc_size_pos()
686 bcm2835_dma_fill_cb_chain_with_sg(chan, direction, d->cb_list, in bcm2835_dma_prep_slave_sg()
[all …]
/linux/net/core/
H A Dflow_offload.c294 list_for_each_entry(block_cb, &block->cb_list, list) { in flow_block_cb_lookup()
414 struct list_head *cb_list; member
426 INIT_LIST_HEAD(&bo.cb_list); in existing_qdiscs_register()
428 list_splice(&bo.cb_list, cur->cb_list); in existing_qdiscs_register()
585 info->cb_list = bo->cb_list_head; in indir_dev_add()
630 return (bo && list_empty(&bo->cb_list)) ? -EOPNOTSUPP : count; in flow_indr_dev_setup_offload()
/linux/include/net/
H A Dflow_offload.h548 struct list_head cb_list; member
560 struct list_head cb_list; member
617 list_add_tail(&block_cb->list, &offload->cb_list); in flow_block_cb_add()
623 list_move(&block_cb->list, &offload->cb_list); in flow_block_cb_remove()
630 list_move(&block_cb->list, &offload->cb_list); in flow_indr_block_cb_remove()
692 INIT_LIST_HEAD(&flow_block->cb_list); in flow_block_init()
/linux/drivers/gpu/drm/i915/gt/
H A Dintel_breadcrumbs.c252 struct list_head cb_list; in signal_irq_work() local
258 list_replace(&rq->fence.cb_list, &cb_list); in signal_irq_work()
260 __dma_fence_signal__notify(&rq->fence, &cb_list); in signal_irq_work()
H A Dselftest_hangcheck.c1544 if (wait_for(!list_empty(&rq->fence.cb_list), 10)) { in __igt_reset_evict_vma()
/linux/block/
H A Dblk-core.c1106 INIT_LIST_HEAD(&plug->cb_list); in blk_start_plug_nr_ios()
1148 while (!list_empty(&plug->cb_list)) { in flush_plug_callbacks()
1149 list_splice_init(&plug->cb_list, &callbacks); in flush_plug_callbacks()
1170 list_for_each_entry(cb, &plug->cb_list, list) in blk_check_plugged()
1180 list_add(&cb->list, &plug->cb_list); in blk_check_plugged()
1188 if (!list_empty(&plug->cb_list)) in __blk_flush_plug()
/linux/arch/s390/kernel/
H A Dvtime.c290 LIST_HEAD(cb_list); in virt_timer_expire()
298 list_move_tail(&timer->entry, &cb_list); in virt_timer_expire()
311 list_for_each_entry_safe(timer, tmp, &cb_list, entry) { in virt_timer_expire()
/linux/include/drm/
H A Ddrm_syncobj.h55 struct list_head cb_list; member
/linux/drivers/infiniband/hw/mlx4/
H A Dalias_GUID.c499 &dev->sriov.alias_guid.ports_guid[port - 1].cb_list; in set_guid_rec()
809 while (!list_empty(&det->cb_list)) { in mlx4_ib_destroy_alias_guid_service()
810 cb_ctx = list_entry(det->cb_list.next, in mlx4_ib_destroy_alias_guid_service()
867 INIT_LIST_HEAD(&dev->sriov.alias_guid.ports_guid[i].cb_list); in mlx4_ib_init_alias_guid_service()
H A Dmlx4_ib.h436 struct list_head cb_list; member
/linux/drivers/gpu/drm/
H A Ddrm_syncobj.c281 list_add_tail(&wait->node, &syncobj->cb_list); in drm_syncobj_fence_add_wait()
352 list_for_each_entry_safe(wait_cur, wait_tmp, &syncobj->cb_list, node) in drm_syncobj_add_point()
388 list_for_each_entry_safe(wait_cur, wait_tmp, &syncobj->cb_list, node) in drm_syncobj_replace_fence()
564 INIT_LIST_HEAD(&syncobj->cb_list); in drm_syncobj_create()
/linux/include/linux/
H A Ddma-fence.h86 struct list_head cb_list; member
H A Dblkdev.h980 struct list_head cb_list; /* md requires an unplug callback */ member
/linux/fs/xfs/
H A Dxfs_log.c527 LIST_HEAD(cb_list); in xlog_state_shutdown_callbacks()
535 list_splice_init(&iclog->ic_callbacks, &cb_list); in xlog_state_shutdown_callbacks()
538 xlog_cil_process_committed(&cb_list); in xlog_state_shutdown_callbacks()
2785 LIST_HEAD(cb_list); in xlog_state_do_iclog_callbacks()
2793 list_splice_init(&iclog->ic_callbacks, &cb_list); in xlog_state_do_iclog_callbacks()
2797 xlog_cil_process_committed(&cb_list); in xlog_state_do_iclog_callbacks()
/linux/include/net/netfilter/
H A Dnf_flow_table.h245 list_add_tail(&block_cb->list, &block->cb_list); in nf_flow_table_offload_add_cb()
/linux/net/sched/
H A Dcls_api.c797 bo->cb_list_head = &flow_block->cb_list; in tcf_block_offload_init()
798 INIT_LIST_HEAD(&bo->cb_list); in tcf_block_offload_init()
819 list_move(&block_cb->list, &bo.cb_list); in tc_block_indr_cleanup()
1640 list_for_each_entry(block_cb, &bo->cb_list, list) { in tcf_block_bind()
1652 list_splice(&bo->cb_list, &block->flow_block.cb_list); in tcf_block_bind()
1657 list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) { in tcf_block_bind()
1681 list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) { in tcf_block_unbind()
3577 list_for_each_entry(block_cb, &block->flow_block.cb_list, list) { in __tc_setup_cb_call()
H A Dact_ct.c378 WARN_ON(!list_empty(&block->cb_list)); in tcf_ct_flow_table_cleanup_work()
/linux/drivers/gpu/drm/i915/
H A Di915_active.c1009 list_add_tail((struct list_head *)node, &rq->fence.cb_list); in i915_request_add_active_barriers()
1112 list_add_tail(&active->cb.node, &fence->cb_list); in __i915_active_fence_set()
/linux/drivers/accel/habanalabs/common/
H A Ddebugfs.c118 list_for_each_entry(cb, &dev_entry->cb_list, debugfs_list) { in command_buffers_show()
1752 INIT_LIST_HEAD(&dev_entry->cb_list); in hl_debugfs_device_init()
1820 list_add(&cb->debugfs_list, &dev_entry->cb_list); in hl_debugfs_add_cb()
H A Dhabanalabs.h2393 struct list_head cb_list; member
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_fence.c205 list_add(&cb.base.node, &f->cb_list); in vmw_fence_wait()
/linux/drivers/net/wireless/intel/ipw2x00/
H A Dipw2200.h752 struct command_block cb_list[CB_NUMBER_OF_ELEMENTS_SMALL]; member

12