/linux/mm/ |
H A D | page_alloc.c | 2271 alloc_flags); in __rmqueue() 2291 alloc_flags); in rmqueue_bulk() 3283 return alloc_flags; in alloc_flags_nofragment() 3286 return alloc_flags; in alloc_flags_nofragment() 3295 return alloc_flags; in alloc_flags_nofragment() 3299 return alloc_flags; in alloc_flags_nofragment() 3310 return alloc_flags; in gfp_to_alloc_flags_cma() 3550 alloc_flags, ac); in __alloc_pages_cpuset_fallback() 4019 alloc_flags = gfp_to_alloc_flags_cma(gfp_mask, alloc_flags); in gfp_to_alloc_flags() 4021 return alloc_flags; in gfp_to_alloc_flags() [all …]
|
H A D | compaction.c | 2453 int alloc_flags) in compaction_zonelist_suitable() argument 2491 int highest_zoneidx, unsigned int alloc_flags) in compaction_suit_allocation_order() argument 2495 watermark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK); in compaction_suit_allocation_order() 2497 alloc_flags)) in compaction_suit_allocation_order() 2535 cc->alloc_flags); in compact_zone() 2745 unsigned int alloc_flags, int highest_zoneidx, in compact_zone_order() argument 2756 .alloc_flags = alloc_flags, in compact_zone_order() 2809 unsigned int alloc_flags, const struct alloc_context *ac, in try_to_compact_pages() argument 2827 (alloc_flags & ALLOC_CPUSET) && in try_to_compact_pages() 2838 alloc_flags, ac->highest_zoneidx, capture); in try_to_compact_pages()
|
/linux/drivers/base/regmap/ |
H A D | regcache-maple.c | 77 map->alloc_flags); in regcache_maple_write() 95 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_write() 138 map->alloc_flags); in regcache_maple_drop() 151 map->alloc_flags); in regcache_maple_drop() 165 ret = mas_store_gfp(&mas, lower, map->alloc_flags); in regcache_maple_drop() 173 ret = mas_store_gfp(&mas, upper, map->alloc_flags); in regcache_maple_drop() 207 buf = kmalloc(val_bytes * (max - min), map->alloc_flags); in regcache_maple_sync_block() 323 entry = kcalloc(last - first + 1, sizeof(unsigned long), map->alloc_flags); in regcache_maple_insert_block() 334 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_insert_block() 351 mt = kmalloc(sizeof(*mt), map->alloc_flags); in regcache_maple_init()
|
H A D | regcache-rbtree.c | 190 map->cache = kmalloc(sizeof *rbtree_ctx, map->alloc_flags); in regcache_rbtree_init() 280 map->alloc_flags); in regcache_rbtree_insert_to_block() 289 map->alloc_flags); in regcache_rbtree_insert_to_block() 323 rbnode = kzalloc(sizeof(*rbnode), map->alloc_flags); in regcache_rbtree_node_alloc() 349 map->alloc_flags); in regcache_rbtree_node_alloc() 355 map->alloc_flags); in regcache_rbtree_node_alloc()
|
H A D | regcache-flat.c | 30 + 1, sizeof(unsigned int), map->alloc_flags); in regcache_flat_init()
|
H A D | internal.h | 65 gfp_t alloc_flags; member
|
H A D | regmap.c | 755 map->alloc_flags = GFP_ATOMIC; in __regmap_init() 757 map->alloc_flags = GFP_KERNEL; in __regmap_init() 2351 wval = kmemdup_array(val, val_count, val_bytes, map->alloc_flags); in regmap_bulk_write()
|
/linux/fs/xfs/libxfs/ |
H A D | xfs_alloc.c | 1677 uint32_t alloc_flags) in xfs_alloc_ag_vextent_near() argument 1697 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_near() 1762 acur.busy_gen, alloc_flags); in xfs_alloc_ag_vextent_near() 1792 uint32_t alloc_flags) in xfs_alloc_ag_vextent_size() argument 1878 busy_gen, alloc_flags); in xfs_alloc_ag_vextent_size() 1977 busy_gen, alloc_flags); in xfs_alloc_ag_vextent_size() 2816 uint32_t alloc_flags) in xfs_alloc_fix_freelist() argument 3523 uint32_t alloc_flags) in xfs_alloc_vextent_prepare_ag() argument 3654 uint32_t alloc_flags = 0; in xfs_alloc_vextent_this_ag() local 3704 uint32_t alloc_flags) in xfs_alloc_vextent_iterate_ags() argument [all …]
|
H A D | xfs_alloc.h | 199 int xfs_alloc_fix_freelist(struct xfs_alloc_arg *args, uint32_t alloc_flags);
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_amdkfd_gpuvm.c | 296 u32 alloc_flags = bo->kfd_bo->alloc_flags; in amdgpu_amdkfd_release_notify() local 1730 u64 alloc_flags; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() local 1742 alloc_flags = 0; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1756 alloc_flags = 0; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1760 alloc_flags = AMDGPU_GEM_CREATE_PREEMPTIBLE; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1780 alloc_flags |= AMDGPU_GEM_CREATE_COHERENT; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1782 alloc_flags |= AMDGPU_GEM_CREATE_EXT_COHERENT; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1784 alloc_flags |= AMDGPU_GEM_CREATE_UNCACHED; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1803 (*mem)->alloc_flags = flags; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1923 if (mem->alloc_flags & in amdgpu_amdkfd_gpuvm_free_memory_of_gpu() [all …]
|
H A D | amdgpu_amdkfd.h | 81 uint32_t alloc_flags; member
|
/linux/include/linux/ |
H A D | compaction.h | 88 unsigned int order, unsigned int alloc_flags, 99 int alloc_flags);
|
H A D | mmzone.h | 1456 int highest_zoneidx, unsigned int alloc_flags, 1460 unsigned int alloc_flags);
|
/linux/lib/ |
H A D | stackdepot.c | 586 gfp_t alloc_flags, in stack_depot_save_flags() argument 627 page = alloc_pages(gfp_nested_mask(alloc_flags), in stack_depot_save_flags() 676 gfp_t alloc_flags) in stack_depot_save() argument 678 return stack_depot_save_flags(entries, nr_entries, alloc_flags, in stack_depot_save()
|
/linux/drivers/md/ |
H A D | dm-zoned-reclaim.c | 286 int alloc_flags = DMZ_ALLOC_SEQ; in dmz_reclaim_rnd_data() local 292 alloc_flags | DMZ_ALLOC_RECLAIM); in dmz_reclaim_rnd_data() 293 if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) { in dmz_reclaim_rnd_data() 294 alloc_flags = DMZ_ALLOC_RND; in dmz_reclaim_rnd_data()
|
H A D | dm-zoned-metadata.c | 2056 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_mapping() local 2071 dzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_mapping() 2168 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_buffer() local 2177 bzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_buffer()
|
/linux/fs/xfs/ |
H A D | xfs_extent_busy.c | 602 uint32_t alloc_flags) in xfs_extent_busy_flush() argument 613 if (alloc_flags & XFS_ALLOC_FLAG_TRYFLUSH) in xfs_extent_busy_flush() 619 if (alloc_flags & XFS_ALLOC_FLAG_FREEING) in xfs_extent_busy_flush()
|
H A D | xfs_extent_busy.h | 74 unsigned busy_gen, uint32_t alloc_flags);
|
/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_chardev.c | 1937 bo_bucket->alloc_flags = (uint32_t)kgd_mem->alloc_flags; in criu_checkpoint_bos() 1940 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) { in criu_checkpoint_bos() 1948 if (bo_bucket->alloc_flags in criu_checkpoint_bos() 1951 bo_bucket->alloc_flags & in criu_checkpoint_bos() 1960 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) in criu_checkpoint_bos() 1963 else if (bo_bucket->alloc_flags & in criu_checkpoint_bos() 1981 bo_bucket->alloc_flags, in criu_checkpoint_bos() 2296 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) { in criu_restore_memory_of_gpu() 2321 &offset, bo_bucket->alloc_flags, criu_resume); in criu_restore_memory_of_gpu() 2341 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) in criu_restore_memory_of_gpu() [all …]
|
/linux/fs/xfs/scrub/ |
H A D | repair.h | 54 int xrep_fix_freelist(struct xfs_scrub *sc, int alloc_flags);
|
H A D | repair.c | 411 int alloc_flags) in xrep_fix_freelist() argument 421 return xfs_alloc_fix_freelist(&args, alloc_flags); in xrep_fix_freelist()
|
/linux/fs/btrfs/ |
H A D | block-group.c | 2940 u64 alloc_flags; in btrfs_inc_block_group_ro() local 2988 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro() 2989 if (alloc_flags != cache->flags) { in btrfs_inc_block_group_ro() 2990 ret = btrfs_chunk_alloc(trans, alloc_flags, in btrfs_inc_block_group_ro() 3018 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->space_info->flags); in btrfs_inc_block_group_ro() 3019 ret = btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_inc_block_group_ro() 3035 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro() 3037 check_system_chunk(trans, alloc_flags); in btrfs_inc_block_group_ro() 3875 u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type); in btrfs_force_chunk_alloc() local 3877 return btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_force_chunk_alloc()
|
/linux/drivers/iommu/ |
H A D | dma-iommu.c | 915 gfp_t alloc_flags = gfp; in __iommu_dma_alloc_pages() local 919 alloc_flags |= __GFP_NORETRY; in __iommu_dma_alloc_pages() 920 page = alloc_pages_node(nid, alloc_flags, order); in __iommu_dma_alloc_pages()
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_object.h | 49 unsigned alloc_flags);
|
/linux/include/uapi/linux/ |
H A D | kfd_ioctl.h | 715 __u32 alloc_flags; member
|