Searched refs:subchunk_size (Results 1 – 3 of 3) sorted by relevance
/open-nvidia-gpu/kernel-open/nvidia-uvm/ |
H A D | uvm_pmm_gpu.c | 1055 uvm_chunk_size_t subchunk_size, in uvm_pmm_gpu_split_chunk() argument 1061 .min_size = subchunk_size, in uvm_pmm_gpu_split_chunk() 1068 UVM_ASSERT(is_power_of_2(subchunk_size)); in uvm_pmm_gpu_split_chunk() 1069 UVM_ASSERT(subchunk_size & pmm->chunk_sizes[chunk->type]); in uvm_pmm_gpu_split_chunk() 1070 UVM_ASSERT(subchunk_size < uvm_gpu_chunk_get_size(chunk)); in uvm_pmm_gpu_split_chunk() 2326 uvm_chunk_size_t subchunk_size; in split_gpu_chunk() local 2338 subchunk_size = uvm_chunk_find_prev_size(chunk_sizes, chunk_size); in split_gpu_chunk() 2339 UVM_ASSERT(subchunk_size != UVM_CHUNK_SIZE_INVALID); in split_gpu_chunk() 2341 num_sub = chunk_size / subchunk_size; in split_gpu_chunk() 2364 subchunk->address = chunk->address + i * subchunk_size; in split_gpu_chunk() [all …]
|
H A D | uvm_pmm_gpu.h | 559 uvm_chunk_size_t subchunk_size,
|
H A D | uvm_va_block.c | 9867 uvm_chunk_size_t subchunk_size; in block_presplit_gpu_chunks() local 9920 for_each_chunk_size_rev(subchunk_size, split_sizes) { in block_presplit_gpu_chunks() 9923 status = uvm_pmm_gpu_split_chunk(&gpu->pmm, curr_chunk, subchunk_size, NULL); in block_presplit_gpu_chunks() 9933 subchunk_size); in block_presplit_gpu_chunks() 9938 if (subchunk_size == new_state.chunk_size) in block_presplit_gpu_chunks() 9944 last_index = (size_t)uvm_div_pow2_64(new->start - 1, subchunk_size); in block_presplit_gpu_chunks() 9945 num_subchunks = (size_t)uvm_div_pow2_64(uvm_gpu_chunk_get_size(curr_chunk), subchunk_size); in block_presplit_gpu_chunks() 9950 UVM_ASSERT(uvm_gpu_chunk_get_size(curr_chunk) == subchunk_size); in block_presplit_gpu_chunks() 10166 uvm_chunk_size_t subchunk_size; in block_presplit_cpu_chunks() local 10190 for_each_chunk_size_rev(subchunk_size, split_sizes) { in block_presplit_cpu_chunks()
|