xref: /linux/arch/um/include/asm/pgtable-3level.h (revision 44f57d78)
1 /*
2  * Copyright 2003 PathScale Inc
3  * Derived from include/asm-i386/pgtable.h
4  * Licensed under the GPL
5  */
6 
7 #ifndef __UM_PGTABLE_3LEVEL_H
8 #define __UM_PGTABLE_3LEVEL_H
9 
10 #define __ARCH_USE_5LEVEL_HACK
11 #include <asm-generic/pgtable-nopud.h>
12 
13 /* PGDIR_SHIFT determines what a third-level page table entry can map */
14 
15 #ifdef CONFIG_64BIT
16 #define PGDIR_SHIFT	30
17 #else
18 #define PGDIR_SHIFT	31
19 #endif
20 #define PGDIR_SIZE	(1UL << PGDIR_SHIFT)
21 #define PGDIR_MASK	(~(PGDIR_SIZE-1))
22 
23 /* PMD_SHIFT determines the size of the area a second-level page table can
24  * map
25  */
26 
27 #define PMD_SHIFT	21
28 #define PMD_SIZE	(1UL << PMD_SHIFT)
29 #define PMD_MASK	(~(PMD_SIZE-1))
30 
31 /*
32  * entries per page directory level
33  */
34 
35 #define PTRS_PER_PTE 512
36 #ifdef CONFIG_64BIT
37 #define PTRS_PER_PMD 512
38 #define PTRS_PER_PGD 512
39 #else
40 #define PTRS_PER_PMD 1024
41 #define PTRS_PER_PGD 1024
42 #endif
43 
44 #define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
45 #define FIRST_USER_ADDRESS	0UL
46 
47 #define pte_ERROR(e) \
48         printk("%s:%d: bad pte %p(%016lx).\n", __FILE__, __LINE__, &(e), \
49 	       pte_val(e))
50 #define pmd_ERROR(e) \
51         printk("%s:%d: bad pmd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
52 	       pmd_val(e))
53 #define pgd_ERROR(e) \
54         printk("%s:%d: bad pgd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
55 	       pgd_val(e))
56 
57 #define pud_none(x)	(!(pud_val(x) & ~_PAGE_NEWPAGE))
58 #define	pud_bad(x)	((pud_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
59 #define pud_present(x)	(pud_val(x) & _PAGE_PRESENT)
60 #define pud_populate(mm, pud, pmd) \
61 	set_pud(pud, __pud(_PAGE_TABLE + __pa(pmd)))
62 
63 #ifdef CONFIG_64BIT
64 #define set_pud(pudptr, pudval) set_64bit((u64 *) (pudptr), pud_val(pudval))
65 #else
66 #define set_pud(pudptr, pudval) (*(pudptr) = (pudval))
67 #endif
68 
69 static inline int pgd_newpage(pgd_t pgd)
70 {
71 	return(pgd_val(pgd) & _PAGE_NEWPAGE);
72 }
73 
74 static inline void pgd_mkuptodate(pgd_t pgd) { pgd_val(pgd) &= ~_PAGE_NEWPAGE; }
75 
76 #ifdef CONFIG_64BIT
77 #define set_pmd(pmdptr, pmdval) set_64bit((u64 *) (pmdptr), pmd_val(pmdval))
78 #else
79 #define set_pmd(pmdptr, pmdval) (*(pmdptr) = (pmdval))
80 #endif
81 
82 struct mm_struct;
83 extern pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address);
84 
85 static inline void pud_clear (pud_t *pud)
86 {
87 	set_pud(pud, __pud(_PAGE_NEWPAGE));
88 }
89 
90 #define pud_page(pud) phys_to_page(pud_val(pud) & PAGE_MASK)
91 #define pud_page_vaddr(pud) ((unsigned long) __va(pud_val(pud) & PAGE_MASK))
92 
93 /* Find an entry in the second-level page table.. */
94 #define pmd_offset(pud, address) ((pmd_t *) pud_page_vaddr(*(pud)) + \
95 			pmd_index(address))
96 
97 static inline unsigned long pte_pfn(pte_t pte)
98 {
99 	return phys_to_pfn(pte_val(pte));
100 }
101 
102 static inline pte_t pfn_pte(unsigned long page_nr, pgprot_t pgprot)
103 {
104 	pte_t pte;
105 	phys_t phys = pfn_to_phys(page_nr);
106 
107 	pte_set_val(pte, phys, pgprot);
108 	return pte;
109 }
110 
111 static inline pmd_t pfn_pmd(unsigned long page_nr, pgprot_t pgprot)
112 {
113 	return __pmd((page_nr << PAGE_SHIFT) | pgprot_val(pgprot));
114 }
115 
116 #endif
117 
118