/linux/arch/arm64/kvm/hyp/nvhe/ |
H A D | page_alloc.c | 127 page_add_to_list(p, &pool->free_area[order]); in __hyp_attach_page() 147 page_add_to_list(buddy, &pool->free_area[buddy->order]); in __hyp_extract_page() 206 while (i <= pool->max_order && list_empty(&pool->free_area[i])) in hyp_alloc_pages() 214 p = node_to_page(pool->free_area[i].next); in hyp_alloc_pages() 234 INIT_LIST_HEAD(&pool->free_area[i]); in hyp_pool_init()
|
/linux/kernel/ |
H A D | vmcore_info.c | 169 VMCOREINFO_STRUCT_SIZE(free_area); in crash_save_vmcoreinfo_init() 187 VMCOREINFO_OFFSET(zone, free_area); in crash_save_vmcoreinfo_init() 190 VMCOREINFO_OFFSET(free_area, free_list); in crash_save_vmcoreinfo_init() 193 VMCOREINFO_LENGTH(zone.free_area, NR_PAGE_ORDERS); in crash_save_vmcoreinfo_init() 195 VMCOREINFO_LENGTH(free_area.free_list, MIGRATE_TYPES); in crash_save_vmcoreinfo_init()
|
/linux/tools/testing/selftests/x86/ |
H A D | test_shadow_stack.c | 561 void *free_area, *shstk, *test_map = (void *)0xFFFFFFFFFFFFFFFF; in test_guard_gap_other_gaps() local 564 free_area = mmap(0, SS_SIZE * 3, PROT_READ | PROT_WRITE, in test_guard_gap_other_gaps() 566 munmap(free_area, SS_SIZE * 3); in test_guard_gap_other_gaps() 568 shstk = create_shstk(free_area + SS_SIZE); in test_guard_gap_other_gaps() 604 void *free_area, *shstk_start, *test_map = (void *)0xFFFFFFFFFFFFFFFF; in test_guard_gap_new_mappings_gaps() local 608 free_area = mmap(0, PAGE_SIZE * 4, PROT_READ | PROT_WRITE, in test_guard_gap_new_mappings_gaps() 610 munmap(free_area, PAGE_SIZE * 4); in test_guard_gap_new_mappings_gaps() 613 shstk_start = mmap(free_area, PAGE_SIZE, PROT_READ | PROT_WRITE, in test_guard_gap_new_mappings_gaps() 615 if (shstk_start == MAP_FAILED || shstk_start != free_area) in test_guard_gap_new_mappings_gaps() 632 if (test_map == free_area + PAGE_SIZE) { in test_guard_gap_new_mappings_gaps()
|
/linux/Documentation/admin-guide/kdump/ |
H A D | vmcoreinfo.rst | 126 free_area section in Common variables 129 The size of a free_area structure. It indicates whether the free_area 160 (zone, free_area|vm_stat|spanned_pages) 169 (free_area, free_list) 175 Each zone has a free_area structure array called free_area[NR_PAGE_ORDERS]. 192 (zone.free_area, NR_PAGE_ORDERS) 196 free_area ranges. NR_PAGE_ORDERS is used by the zone buddy allocator. 316 (free_area.free_list, MIGRATE_TYPES)
|
/linux/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | gfp.h | 19 struct list_head free_area[NR_PAGE_ORDERS]; member
|
/linux/mm/ |
H A D | page_reporting.c | 150 struct free_area *area = &zone->free_area[order]; in page_reporting_cycle()
|
H A D | show_mem.c | 367 struct free_area *area = &zone->free_area[order]; in show_free_areas()
|
H A D | page_alloc.c | 643 struct free_area *area = &zone->free_area[order]; in __add_to_free_list() 664 struct free_area *area = &zone->free_area[order]; in move_to_free_list() 691 zone->free_area[order].nr_free--; in __del_page_from_free_list() 1502 struct free_area *area; in __rmqueue_smallest() 1507 area = &(zone->free_area[current_order]); in __rmqueue_smallest() 1929 int find_suitable_fallback(struct free_area *area, unsigned int order, in find_suitable_fallback() 2037 struct free_area *area = &(zone->free_area[order]); in unreserve_highatomic_pageblock() 2115 struct free_area *area; in __rmqueue_fallback() 2137 area = &(zone->free_area[current_order]); in __rmqueue_fallback() 2162 area = &(zone->free_area[current_order]); in __rmqueue_fallback() [all …]
|
H A D | vmstat.c | 1071 blocks = data_race(zone->free_area[order].nr_free); in fill_contig_page_info() 1487 seq_printf(m, "%6lu ", data_race(zone->free_area[order].nr_free)); in frag_show_print() 1513 struct free_area *area; in pagetypeinfo_showfree_print() 1517 area = &(zone->free_area[order]); in pagetypeinfo_showfree_print()
|
H A D | compaction.c | 1605 struct free_area *area = &cc->zone->free_area[order]; in fast_isolate_freepages() 2039 struct free_area *area = &cc->zone->free_area[order]; in fast_find_migrateblock() 2354 struct free_area *area = &cc->zone->free_area[order]; in __compact_finished()
|
H A D | internal.h | 741 int find_suitable_fallback(struct free_area *area, unsigned int order, 744 static inline bool free_area_empty(struct free_area *area, int migratetype) in free_area_empty()
|
H A D | mm_init.c | 1394 INIT_LIST_HEAD(&zone->free_area[order].free_list[t]); in zone_init_free_lists() 1395 zone->free_area[order].nr_free = 0; in zone_init_free_lists()
|
/linux/drivers/md/ |
H A D | dm-snap-persistent.c | 205 static void free_area(struct pstore *ps) in free_area() function 365 free_area(ps); in read_header() 379 free_area(ps); in read_header() 601 free_area(ps); in persistent_dtr()
|
/linux/include/linux/ |
H A D | mmzone.h | 117 struct free_area { struct 944 struct free_area free_area[NR_PAGE_ORDERS]; member
|
/linux/kernel/events/ |
H A D | uprobes.c | 1490 goto free_area; in __create_xol_area() 1513 free_area: in __create_xol_area()
|
/linux/kernel/power/ |
H A D | snapshot.c | 1279 &zone->free_area[order].free_list[t], buddy_list) { in mark_free_pages()
|