xref: /freebsd/sys/arm64/include/vmparam.h (revision 1edb7116)
1 /*-
2  * Copyright (c) 1990 The Regents of the University of California.
3  * All rights reserved.
4  * Copyright (c) 1994 John S. Dyson
5  * All rights reserved.
6  *
7  * This code is derived from software contributed to Berkeley by
8  * William Jolitz.
9  *
10  * Redistribution and use in source and binary forms, with or without
11  * modification, are permitted provided that the following conditions
12  * are met:
13  * 1. Redistributions of source code must retain the above copyright
14  *    notice, this list of conditions and the following disclaimer.
15  * 2. Redistributions in binary form must reproduce the above copyright
16  *    notice, this list of conditions and the following disclaimer in the
17  *    documentation and/or other materials provided with the distribution.
18  * 3. Neither the name of the University nor the names of its contributors
19  *    may be used to endorse or promote products derived from this software
20  *    without specific prior written permission.
21  *
22  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
23  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
24  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
25  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
26  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
27  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
28  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
29  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
30  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
31  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
32  * SUCH DAMAGE.
33  *	from: FreeBSD: src/sys/i386/include/vmparam.h,v 1.33 2000/03/30
34  */
35 
36 #ifdef __arm__
37 #include <arm/vmparam.h>
38 #else /* !__arm__ */
39 
40 #ifndef	_MACHINE_VMPARAM_H_
41 #define	_MACHINE_VMPARAM_H_
42 
43 /*
44  * Virtual memory related constants, all in bytes
45  */
46 #ifndef MAXTSIZ
47 #define	MAXTSIZ		(1*1024*1024*1024)	/* max text size */
48 #endif
49 #ifndef DFLDSIZ
50 #define	DFLDSIZ		(128*1024*1024)		/* initial data size limit */
51 #endif
52 #ifndef MAXDSIZ
53 #define	MAXDSIZ		(1*1024*1024*1024)	/* max data size */
54 #endif
55 #ifndef DFLSSIZ
56 #define	DFLSSIZ		(128*1024*1024)		/* initial stack size limit */
57 #endif
58 #ifndef MAXSSIZ
59 #define	MAXSSIZ		(1*1024*1024*1024)	/* max stack size */
60 #endif
61 #ifndef SGROWSIZ
62 #define	SGROWSIZ	(128*1024)		/* amount to grow stack */
63 #endif
64 
65 /*
66  * The physical address space is sparsely populated.
67  */
68 #define	VM_PHYSSEG_SPARSE
69 
70 /*
71  * The number of PHYSSEG entries.
72  */
73 #define	VM_PHYSSEG_MAX		64
74 
75 /*
76  * Create two free page pools: VM_FREEPOOL_DEFAULT is the default pool
77  * from which physical pages are allocated and VM_FREEPOOL_DIRECT is
78  * the pool from which physical pages for small UMA objects are
79  * allocated.
80  */
81 #define	VM_NFREEPOOL		2
82 #define	VM_FREEPOOL_DEFAULT	0
83 #define	VM_FREEPOOL_DIRECT	1
84 
85 /*
86  * Create two free page lists: VM_FREELIST_DMA32 is for physical pages that have
87  * physical addresses below 4G, and VM_FREELIST_DEFAULT is for all other
88  * physical pages.
89  */
90 #define	VM_NFREELIST		2
91 #define	VM_FREELIST_DEFAULT	0
92 #define	VM_FREELIST_DMA32	1
93 
94 /*
95  * When PAGE_SIZE is 4KB, an allocation size of 16MB is supported in order
96  * to optimize the use of the direct map by UMA.  Specifically, a 64-byte
97  * cache line contains at most 8 L2 BLOCK entries, collectively mapping 16MB
98  * of physical memory.  By reducing the number of distinct 16MB "pages" that
99  * are used by UMA, the physical memory allocator reduces the likelihood of
100  * both 2MB page TLB misses and cache misses during the page table walk when
101  * a 2MB page TLB miss does occur.
102  */
103 #define	VM_NFREEORDER		13
104 
105 /*
106  * Enable superpage reservations: 1 level.
107  */
108 #ifndef	VM_NRESERVLEVEL
109 #define	VM_NRESERVLEVEL		1
110 #endif
111 
112 /*
113  * Level 0 reservations consist of 512 pages.
114  */
115 #ifndef	VM_LEVEL_0_ORDER
116 #define	VM_LEVEL_0_ORDER	9
117 #endif
118 
119 /**
120  * Address space layout.
121  *
122  * ARMv8 implements up to a 48 bit virtual address space. The address space is
123  * split into 2 regions at each end of the 64 bit address space, with an
124  * out of range "hole" in the middle.
125  *
126  * We use the full 48 bits for each region, however the kernel may only use
127  * a limited range within this space.
128  *
129  * Upper region:    0xffffffffffffffff  Top of virtual memory
130  *
131  *                  0xfffffeffffffffff  End of DMAP
132  *                  0xffffa00000000000  Start of DMAP
133  *
134  *                  0xffff027fffffffff  End of KMSAN origin map
135  *                  0xffff020000000000  Start of KMSAN origin map
136  *
137  *                  0xffff017fffffffff  End of KMSAN shadow map
138  *                  0xffff010000000000  Start of KMSAN shadow map
139  *
140  *                  0xffff009fffffffff  End of KASAN shadow map
141  *                  0xffff008000000000  Start of KASAN shadow map
142  *
143  *                  0xffff007fffffffff  End of KVA
144  *                  0xffff000000000000  Kernel base address & start of KVA
145  *
146  * Hole:            0xfffeffffffffffff
147  *                  0x0001000000000000
148  *
149  * Lower region:    0x0000ffffffffffff End of user address space
150  *                  0x0000000000000000 Start of user address space
151  *
152  * We use the upper region for the kernel, and the lower region for userland.
153  *
154  * We define some interesting address constants:
155  *
156  * VM_MIN_ADDRESS and VM_MAX_ADDRESS define the start and end of the entire
157  * 64 bit address space, mostly just for convenience.
158  *
159  * VM_MIN_KERNEL_ADDRESS and VM_MAX_KERNEL_ADDRESS define the start and end of
160  * mappable kernel virtual address space.
161  *
162  * VM_MIN_USER_ADDRESS and VM_MAX_USER_ADDRESS define the start and end of the
163  * user address space.
164  */
165 #define	VM_MIN_ADDRESS		(0x0000000000000000UL)
166 #define	VM_MAX_ADDRESS		(0xffffffffffffffffUL)
167 
168 /* 512 GiB of kernel addresses */
169 #define	VM_MIN_KERNEL_ADDRESS	(0xffff000000000000UL)
170 #define	VM_MAX_KERNEL_ADDRESS	(0xffff008000000000UL)
171 
172 /* 128 GiB KASAN shadow map */
173 #define	KASAN_MIN_ADDRESS	(0xffff008000000000UL)
174 #define	KASAN_MAX_ADDRESS	(0xffff00a000000000UL)
175 
176 /* 512GiB KMSAN shadow map */
177 #define	KMSAN_SHAD_MIN_ADDRESS	(0xffff010000000000UL)
178 #define	KMSAN_SHAD_MAX_ADDRESS	(0xffff018000000000UL)
179 
180 /* 512GiB KMSAN origin map */
181 #define	KMSAN_ORIG_MIN_ADDRESS	(0xffff020000000000UL)
182 #define	KMSAN_ORIG_MAX_ADDRESS	(0xffff028000000000UL)
183 
184 /* The address bits that hold a pointer authentication code */
185 #define	PAC_ADDR_MASK		(0xff7f000000000000UL)
186 
187 /* If true addr is in the kernel address space */
188 #define	ADDR_IS_KERNEL(addr)	(((addr) & (1ul << 55)) == (1ul << 55))
189 /* If true addr is in its canonical form (i.e. no TBI, PAC, etc.) */
190 #define	ADDR_IS_CANONICAL(addr)	\
191     (((addr) & 0xffff000000000000UL) == 0 || \
192      ((addr) & 0xffff000000000000UL) == 0xffff000000000000UL)
193 #define	ADDR_MAKE_CANONICAL(addr) ({			\
194 	__typeof(addr) _tmp_addr = (addr);		\
195 							\
196 	_tmp_addr &= ~0xffff000000000000UL;		\
197 	if (ADDR_IS_KERNEL(addr))			\
198 		_tmp_addr |= 0xffff000000000000UL;	\
199 							\
200 	_tmp_addr;					\
201 })
202 
203 /* 95 TiB maximum for the direct map region */
204 #define	DMAP_MIN_ADDRESS	(0xffffa00000000000UL)
205 #define	DMAP_MAX_ADDRESS	(0xffffff0000000000UL)
206 
207 #define	DMAP_MIN_PHYSADDR	(dmap_phys_base)
208 #define	DMAP_MAX_PHYSADDR	(dmap_phys_max)
209 
210 /* True if pa is in the dmap range */
211 #define	PHYS_IN_DMAP(pa)	((pa) >= DMAP_MIN_PHYSADDR && \
212     (pa) < DMAP_MAX_PHYSADDR)
213 /* True if va is in the dmap range */
214 #define	VIRT_IN_DMAP(va)	((va) >= DMAP_MIN_ADDRESS && \
215     (va) < (dmap_max_addr))
216 
217 #define	PMAP_HAS_DMAP	1
218 #define	PHYS_TO_DMAP(pa)						\
219 ({									\
220 	KASSERT(PHYS_IN_DMAP(pa),					\
221 	    ("%s: PA out of range, PA: 0x%lx", __func__,		\
222 	    (vm_paddr_t)(pa)));						\
223 	((pa) - dmap_phys_base) + DMAP_MIN_ADDRESS;			\
224 })
225 
226 #define	DMAP_TO_PHYS(va)						\
227 ({									\
228 	KASSERT(VIRT_IN_DMAP(va),					\
229 	    ("%s: VA out of range, VA: 0x%lx", __func__,		\
230 	    (vm_offset_t)(va)));					\
231 	((va) - DMAP_MIN_ADDRESS) + dmap_phys_base;			\
232 })
233 
234 #define	VM_MIN_USER_ADDRESS	(0x0000000000000000UL)
235 #define	VM_MAX_USER_ADDRESS	(0x0001000000000000UL)
236 
237 #define	VM_MINUSER_ADDRESS	(VM_MIN_USER_ADDRESS)
238 #define	VM_MAXUSER_ADDRESS	(VM_MAX_USER_ADDRESS)
239 
240 #define	KERNBASE		(VM_MIN_KERNEL_ADDRESS)
241 #define	SHAREDPAGE		(VM_MAXUSER_ADDRESS - PAGE_SIZE)
242 #define	USRSTACK		SHAREDPAGE
243 
244 /*
245  * How many physical pages per kmem arena virtual page.
246  */
247 #ifndef VM_KMEM_SIZE_SCALE
248 #define	VM_KMEM_SIZE_SCALE	(1)
249 #endif
250 
251 /*
252  * Optional ceiling (in bytes) on the size of the kmem arena: 60% of the
253  * kernel map.
254  */
255 #ifndef VM_KMEM_SIZE_MAX
256 #define	VM_KMEM_SIZE_MAX	((VM_MAX_KERNEL_ADDRESS - \
257     VM_MIN_KERNEL_ADDRESS + 1) * 3 / 5)
258 #endif
259 
260 /*
261  * Initial pagein size of beginning of executable file.
262  */
263 #ifndef	VM_INITIAL_PAGEIN
264 #define	VM_INITIAL_PAGEIN	16
265 #endif
266 
267 #if !defined(KASAN) && !defined(KMSAN)
268 #define	UMA_MD_SMALL_ALLOC
269 #endif
270 
271 #ifndef LOCORE
272 
273 extern vm_paddr_t dmap_phys_base;
274 extern vm_paddr_t dmap_phys_max;
275 extern vm_offset_t dmap_max_addr;
276 extern vm_offset_t vm_max_kernel_address;
277 
278 #endif
279 
280 #define	ZERO_REGION_SIZE	(64 * 1024)	/* 64KB */
281 
282 #define	DEVMAP_MAX_VADDR	VM_MAX_KERNEL_ADDRESS
283 
284 /*
285  * The pmap can create non-transparent large page mappings.
286  */
287 #define	PMAP_HAS_LARGEPAGES	1
288 
289 /*
290  * Need a page dump array for minidump.
291  */
292 #define MINIDUMP_PAGE_TRACKING	1
293 
294 #endif /* !_MACHINE_VMPARAM_H_ */
295 
296 #endif /* !__arm__ */
297