/linux/include/linux/ |
H A D | mempolicy.h | 46 struct mempolicy { struct 64 extern void __mpol_put(struct mempolicy *pol); 86 extern struct mempolicy *__mpol_dup(struct mempolicy *pol); 87 static inline struct mempolicy *mpol_dup(struct mempolicy *pol) in mpol_dup() 100 extern bool __mpol_equal(struct mempolicy *a, struct mempolicy *b); 101 static inline bool mpol_equal(struct mempolicy *a, struct mempolicy *b) in mpol_equal() 118 struct mempolicy *policy; 183 struct mempolicy {}; struct 190 static inline bool mpol_equal(struct mempolicy *a, struct mempolicy *b) in mpol_equal() 210 struct mempolicy *mpol) in mpol_shared_policy_init() [all …]
|
H A D | gfp.h | 13 struct mempolicy; 304 struct mempolicy *mpol, pgoff_t ilx, int nid); 314 struct mempolicy *mpol, pgoff_t ilx, int nid) in alloc_pages_mpol_noprof()
|
H A D | shmem_fs.h | 71 struct mempolicy *mpol; /* default memory policy for mappings */
|
H A D | sched.h | 62 struct mempolicy; 1268 struct mempolicy *mempolicy; member
|
H A D | mm.h | 35 struct mempolicy; 622 int (*set_policy)(struct vm_area_struct *vma, struct mempolicy *new); 634 struct mempolicy *(*get_policy)(struct vm_area_struct *vma, 3335 struct mempolicy *policy, 3372 struct mempolicy *new_pol) in vma_modify_policy()
|
H A D | mm_types.h | 740 struct mempolicy *vm_policy; /* NUMA policy for the VMA */
|
/linux/mm/ |
H A D | mempolicy.c | 201 struct mempolicy *pol = p->mempolicy; in get_task_policy() 473 struct mempolicy *pol; 791 struct mempolicy *old; in vma_replace_policy() 931 struct mempolicy *pol = current->mempolicy, *pol_refcount = NULL; in do_get_mempolicy() 2133 struct mempolicy *mempolicy; in init_nodemask_of_mempolicy() local 2139 mempolicy = current->mempolicy; in init_nodemask_of_mempolicy() 2175 struct mempolicy *mempolicy; in mempolicy_in_oom_domain() local 2182 mempolicy = tsk->mempolicy; in mempolicy_in_oom_domain() 2183 if (mempolicy && mempolicy->mode == MPOL_BIND) in mempolicy_in_oom_domain() 2573 struct mempolicy *__mpol_dup(struct mempolicy *old) in __mpol_dup() [all …]
|
H A D | swap.h | 5 struct mempolicy; 54 struct mempolicy *mpol, pgoff_t ilx, bool *new_page_allocated, 57 struct mempolicy *mpol, pgoff_t ilx); 85 gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t ilx) in swap_cluster_readahead()
|
H A D | swap_state.c | 430 struct mempolicy *mpol, pgoff_t ilx, bool *new_page_allocated, in __read_swap_cache_async() 557 struct mempolicy *mpol; in read_swap_cache_async() 651 struct mempolicy *mpol, pgoff_t ilx) in swap_cluster_readahead() 816 struct mempolicy *mpol, pgoff_t targ_ilx, struct vm_fault *vmf) in swap_vma_readahead() 897 struct mempolicy *mpol; in swapin_readahead()
|
H A D | shmem.c | 116 struct mempolicy *mpol; 1532 static void shmem_show_mpol(struct seq_file *seq, struct mempolicy *mpol) in shmem_show_mpol() 1544 static struct mempolicy *shmem_get_sbmpol(struct shmem_sb_info *sbinfo) in shmem_get_sbmpol() 1546 struct mempolicy *mpol = NULL; in shmem_get_sbmpol() 1571 struct mempolicy *mpol; in shmem_swapin_cluster() 1609 struct mempolicy *mpol; in shmem_alloc_hugefolio() 1623 struct mempolicy *mpol; in shmem_alloc_folio() 2361 static struct mempolicy *shmem_get_policy(struct vm_area_struct *vma, in shmem_get_policy() 2381 struct mempolicy *mpol; in shmem_get_pgoff_policy() 4152 struct mempolicy *mpol = NULL; in shmem_reconfigure() [all …]
|
H A D | Makefile | 81 obj-$(CONFIG_NUMA) += mempolicy.o
|
H A D | zswap.c | 1034 struct mempolicy *mpol; in zswap_writeback_entry()
|
H A D | hugetlb.c | 1396 struct mempolicy *mpol; in dequeue_hugetlb_folio_vma() 2575 struct mempolicy *mpol; in alloc_buddy_hugetlb_folio_with_mpol() 4882 struct mempolicy *mpol = get_task_policy(current); in policy_mbind_nodemask() 6649 struct mempolicy *mpol; in alloc_hugetlb_folio_vma()
|
H A D | mmap.c | 865 unsigned long vm_flags, pgoff_t pgoff, struct mempolicy *policy, in vma_merge() 2487 struct mempolicy *policy, in vma_modify()
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-kernel-mm-mempolicy-weighted-interleave | 1 What: /sys/kernel/mm/mempolicy/weighted_interleave/ 6 What: /sys/kernel/mm/mempolicy/weighted_interleave/nodeN 12 utilized by tasks which have set their mempolicy to
|
H A D | sysfs-kernel-mm-mempolicy | 1 What: /sys/kernel/mm/mempolicy/
|
/linux/fs/proc/ |
H A D | internal.h | 18 struct mempolicy; 294 struct mempolicy *task_mempolicy;
|
/linux/Documentation/translations/zh_CN/core-api/ |
H A D | mm-api.rst | 117 mm/mempolicy.c
|
/linux/Documentation/admin-guide/mm/ |
H A D | numa_memory_policy.rst | 170 structure, struct mempolicy. Details of this structure will be 207 preferred_node member of struct mempolicy. When the internal 256 /sys/kernel/mm/mempolicy/weighted_interleave/ 269 Without this flag, any time a mempolicy is rebound because of a 301 mempolicy is rebound because of a change in the set of allowed 321 if not already set, sets the node in the mempolicy nodemask. 347 To resolve use/free races, struct mempolicy contains an atomic reference 350 the structure back to the mempolicy kmem cache when the reference count 427 definitions are defined in <linux/mempolicy.h>. 478 mempolicy range. Other address ranges are ignored. A home node is the NUMA node
|
H A D | hugetlbpage.rst | 291 sysctl or attribute. When the ``nr_hugepages`` attribute is used, mempolicy 313 #. Regardless of mempolicy mode [see 316 specified in the mempolicy as if "interleave" had been specified. 332 Any of the other mempolicy modes may be used to specify a single node. 334 #. The nodes allowed mask will be derived from any non-default task mempolicy, 341 #. Any task mempolicy specified--e.g., using numactl--will be constrained by 373 resources exist, regardless of the task's mempolicy or cpuset constraints. 376 as we don't know until fault time, when the faulting task's mempolicy is
|
/linux/Documentation/core-api/ |
H A D | mm-api.rst | 107 .. kernel-doc:: mm/mempolicy.c
|
/linux/kernel/ |
H A D | fork.c | 2321 p->mempolicy = mpol_dup(p->mempolicy); in copy_process() 2322 if (IS_ERR(p->mempolicy)) { in copy_process() 2323 retval = PTR_ERR(p->mempolicy); in copy_process() 2324 p->mempolicy = NULL; in copy_process() 2670 mpol_put(p->mempolicy); in copy_process()
|
/linux/ipc/ |
H A D | shm.c | 566 static int shm_set_policy(struct vm_area_struct *vma, struct mempolicy *mpol) in shm_set_policy() 576 static struct mempolicy *shm_get_policy(struct vm_area_struct *vma, in shm_get_policy() 580 struct mempolicy *mpol = vma->vm_policy; in shm_get_policy()
|
/linux/Documentation/admin-guide/cgroup-v1/ |
H A D | cpusets.rst | 342 except perhaps as modified by the task's NUMA mempolicy or cpuset 349 or slab caches to ignore the task's NUMA mempolicy and be spread 353 is turned off, then the currently specified NUMA mempolicy once again 631 mempolicy MPOL_BIND, and the nodes to which it was bound overlap with
|
/linux/Documentation/filesystems/ |
H A D | proc.rst | 1693 cpuset. If it is due to a mempolicy's node(s) being exhausted, the allowed 1694 memory represents the set of mempolicy nodes. If it is due to a memory 1710 same system, cpuset, mempolicy, or memory controller resources to use at least
|