/linux/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-srv.h | 111 struct page **chunks; member
|
/linux/drivers/accel/habanalabs/common/ |
H A D | command_submission.c | 1409 void __user *chunks, u32 num_chunks, in hl_cs_copy_chunk_array() argument 1435 if (copy_from_user(*cs_chunk_array, chunks, size_to_copy)) { in hl_cs_copy_chunk_array() 1510 rc = hl_cs_copy_chunk_array(hdev, &cs_chunk_array, chunks, num_chunks, in cs_ioctl_default() 1694 void __user *chunks; in hl_cs_ctx_switch() local 1729 chunks = (void __user *) (uintptr_t) args->in.chunks_restore; in hl_cs_ctx_switch() 1737 rc = cs_ioctl_default(hpriv, chunks, num_chunks, in hl_cs_ctx_switch() 2199 void __user *chunks, u32 num_chunks, in cs_ioctl_signal_wait() argument 2563 void __user *chunks; in hl_cs_ioctl() local 2579 chunks = (void __user *) (uintptr_t) args->in.chunks_execute; in hl_cs_ioctl() 2596 rc = cs_ioctl_signal_wait(hpriv, cs_type, chunks, num_chunks, in hl_cs_ioctl() [all …]
|
/linux/include/net/ |
H A D | xdp_sock.h | 28 u32 chunks; member
|
/linux/Documentation/networking/device_drivers/ethernet/microsoft/ |
H A D | netvsc.rst | 86 is probed. The receive area is broken into MTU sized chunks and each may 91 for sending. The send area is broken into chunks, typically of 6144
|
/linux/drivers/gpu/drm/i915/gt/shaders/ |
H A D | README | 5 pre-compiled batch chunks that will clear any residual render cache during
|
/linux/fs/xfs/ |
H A D | xfs_buf_item.c | 833 int chunks; in xfs_buf_item_init() local 866 chunks = DIV_ROUND_UP(BBTOB(bp->b_maps[i].bm_len), in xfs_buf_item_init() 868 map_size = DIV_ROUND_UP(chunks, NBWORD); in xfs_buf_item_init()
|
/linux/fs/xfs/scrub/ |
H A D | reap.c | 753 int chunks; in xreap_buf_loggable() local 756 chunks = DIV_ROUND_UP(BBTOB(bp->b_maps[i].bm_len), in xreap_buf_loggable() 758 map_size = DIV_ROUND_UP(chunks, NBWORD); in xreap_buf_loggable()
|
/linux/fs/smb/server/ |
H A D | vfs.c | 1782 struct srv_copychunk *chunks, in ksmbd_vfs_copy_file_ranges() argument 1813 src_off = le64_to_cpu(chunks[i].SourceOffset); in ksmbd_vfs_copy_file_ranges() 1814 dst_off = le64_to_cpu(chunks[i].TargetOffset); in ksmbd_vfs_copy_file_ranges() 1815 len = le32_to_cpu(chunks[i].Length); in ksmbd_vfs_copy_file_ranges() 1829 src_off = le64_to_cpu(chunks[i].SourceOffset); in ksmbd_vfs_copy_file_ranges() 1830 dst_off = le64_to_cpu(chunks[i].TargetOffset); in ksmbd_vfs_copy_file_ranges() 1831 len = le32_to_cpu(chunks[i].Length); in ksmbd_vfs_copy_file_ranges()
|
H A D | vfs.h | 97 struct srv_copychunk *chunks,
|
/linux/drivers/md/ |
H A D | md-bitmap.h | 191 unsigned long chunks; /* Total number of data member
|
/linux/include/linux/ |
H A D | genalloc.h | 60 struct list_head chunks; /* list of chunks in this pool */ member
|
/linux/net/sctp/ |
H A D | stream_sched_rr.c | 104 ch = list_first_entry(&msg->chunks, struct sctp_chunk, frag_list); in sctp_sched_rr_enqueue()
|
H A D | stream_sched_fc.c | 105 ch = list_first_entry(&msg->chunks, struct sctp_chunk, frag_list); in sctp_sched_fc_enqueue()
|
H A D | stream_sched.c | 237 if (!list_is_last(&ch->frag_list, &ch->msg->chunks) && in sctp_sched_dequeue_done()
|
H A D | stream_sched_prio.c | 222 ch = list_first_entry(&msg->chunks, struct sctp_chunk, frag_list); in sctp_sched_prio_enqueue()
|
/linux/Documentation/admin-guide/ |
H A D | pstore-blk.rst | 81 There are multiple chunks for oops/panic front-end depending on the remaining 84 pstore/blk will log to oops/panic chunks one by one, and always overwrite the 116 Similar to oops front-end, there are multiple chunks for ftrace front-end
|
/linux/drivers/net/wireless/ath/carl9170/ |
H A D | tx.c | 193 unsigned int chunks; in carl9170_alloc_dev_space() local 198 chunks = DIV_ROUND_UP(skb->len, ar->fw.mem_block_size); in carl9170_alloc_dev_space() 199 if (unlikely(atomic_sub_return(chunks, &ar->mem_free_blocks) < 0)) { in carl9170_alloc_dev_space() 200 atomic_add(chunks, &ar->mem_free_blocks); in carl9170_alloc_dev_space() 209 atomic_add(chunks, &ar->mem_free_blocks); in carl9170_alloc_dev_space()
|
/linux/Documentation/devicetree/bindings/iommu/ |
H A D | samsung,sysmmu.yaml | 14 physical memory chunks visible as a contiguous region to DMA-capable peripheral
|
/linux/Documentation/usb/ |
H A D | ehci.rst | 206 In typical situations, a usb_bulk_msg() loop writing out 4 KB chunks is 209 than the I/O. If that same loop used 16 KB chunks, it'd be better; a 210 sequence of 128 KB chunks would waste a lot less.
|
/linux/Documentation/crypto/ |
H A D | devel-algos.rst | 131 hardware needs to be fed separate chunks of the scatterlist which 237 implementation needs to be fed separate chunks of the scatterlist which
|
/linux/scripts/ |
H A D | checkpatch.pl | 1851 my (@chunks); 1857 push(@chunks, [ $condition, $statement ]); 1859 return ($level, $linenr, @chunks); 1873 return ($level, $linenr, @chunks); 6122 my ($level, $endln, @chunks) = 6126 if ($#chunks > 0 && $level == 0) { 6132 for my $chunk (@chunks) { 6193 my ($level, $endln, @chunks) = 6197 my ($cond, $block) = @{$chunks[0]}; 6215 if (defined $chunks[1]) { [all …]
|
/linux/drivers/net/wireless/intel/iwlwifi/fw/ |
H A D | uefi.c | 144 pnvm_data->chunks[pnvm_data->n_chunks].data = desc->data; in iwl_uefi_handle_tlv_mem_desc() 145 pnvm_data->chunks[pnvm_data->n_chunks].len = data_len; in iwl_uefi_handle_tlv_mem_desc()
|
/linux/Documentation/driver-api/mmc/ |
H A D | mmc-dev-attrs.rst | 56 be desirable to do it in smaller chunks for three reasons:
|
H A D | mmc-async-req.rst | 72 request in two chunks, prepare the first chunk and start the request,
|
/linux/Documentation/networking/devlink/ |
H A D | devlink-region.rst | 36 of size 256 bytes or larger will be split into multiple chunks. If atomic
|