1 /* $NetBSD: uvm_amap.h,v 1.17 2001/06/02 18:09:25 chs Exp $ */ 2 3 /* 4 * 5 * Copyright (c) 1997 Charles D. Cranor and Washington University. 6 * All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions 10 * are met: 11 * 1. Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in the 15 * documentation and/or other materials provided with the distribution. 16 * 3. All advertising materials mentioning features or use of this software 17 * must display the following acknowledgement: 18 * This product includes software developed by Charles D. Cranor and 19 * Washington University. 20 * 4. The name of the author may not be used to endorse or promote products 21 * derived from this software without specific prior written permission. 22 * 23 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 24 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 25 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 26 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 27 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 28 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 29 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 30 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 31 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 32 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 33 */ 34 35 #ifndef _UVM_UVM_AMAP_H_ 36 #define _UVM_UVM_AMAP_H_ 37 38 /* 39 * uvm_amap.h: general amap interface and amap implementation-specific info 40 */ 41 42 /* 43 * an amap structure contains pointers to a set of anons that are 44 * mapped together in virtual memory (an anon is a single page of 45 * anonymous virtual memory -- see uvm_anon.h). in uvm we hide the 46 * details of the implementation of amaps behind a general amap 47 * interface. this allows us to change the amap implementation 48 * without having to touch the rest of the code. this file is divided 49 * into two parts: the definition of the uvm amap interface and the 50 * amap implementation-specific definitions. 51 */ 52 53 #ifdef _KERNEL 54 55 /* 56 * part 1: amap interface 57 */ 58 59 /* 60 * forward definition of vm_amap structure. only amap 61 * implementation-specific code should directly access the fields of 62 * this structure. 63 */ 64 65 struct vm_amap; 66 67 /* 68 * handle inline options... we allow amap ops to be inline, but we also 69 * provide a hook to turn this off. macros can also be used. 70 */ 71 72 #ifdef UVM_AMAP_INLINE /* defined/undef'd in uvm_amap.c */ 73 #define AMAP_INLINE static __inline /* inline enabled */ 74 #else 75 #define AMAP_INLINE /* inline disabled */ 76 #endif /* UVM_AMAP_INLINE */ 77 78 79 /* 80 * prototypes for the amap interface 81 */ 82 83 AMAP_INLINE 84 void amap_add /* add an anon to an amap */ 85 __P((struct vm_aref *, vaddr_t, 86 struct vm_anon *, boolean_t)); 87 struct vm_amap *amap_alloc /* allocate a new amap */ 88 __P((vaddr_t, vaddr_t, int)); 89 void amap_copy /* clear amap needs-copy flag */ 90 __P((struct vm_map *, struct vm_map_entry *, int, 91 boolean_t, vaddr_t, vaddr_t)); 92 void amap_cow_now /* resolve all COW faults now */ 93 __P((struct vm_map *, struct vm_map_entry *)); 94 void amap_extend /* make amap larger */ 95 __P((struct vm_map_entry *, vsize_t)); 96 int amap_flags /* get amap's flags */ 97 __P((struct vm_amap *)); 98 void amap_free /* free amap */ 99 __P((struct vm_amap *)); 100 void amap_init /* init amap module (at boot time) */ 101 __P((void)); 102 void amap_lock /* lock amap */ 103 __P((struct vm_amap *)); 104 AMAP_INLINE 105 struct vm_anon *amap_lookup /* lookup an anon @ offset in amap */ 106 __P((struct vm_aref *, vaddr_t)); 107 AMAP_INLINE 108 void amap_lookups /* lookup multiple anons */ 109 __P((struct vm_aref *, vaddr_t, 110 struct vm_anon **, int)); 111 AMAP_INLINE 112 void amap_ref /* add a reference to an amap */ 113 __P((struct vm_amap *, vaddr_t, vsize_t, int)); 114 int amap_refs /* get number of references of amap */ 115 __P((struct vm_amap *)); 116 void amap_share_protect /* protect pages in a shared amap */ 117 __P((struct vm_map_entry *, vm_prot_t)); 118 void amap_splitref /* split reference to amap into two */ 119 __P((struct vm_aref *, struct vm_aref *, 120 vaddr_t)); 121 AMAP_INLINE 122 void amap_unadd /* remove an anon from an amap */ 123 __P((struct vm_aref *, vaddr_t)); 124 void amap_unlock /* unlock amap */ 125 __P((struct vm_amap *)); 126 AMAP_INLINE 127 void amap_unref /* drop reference to an amap */ 128 __P((struct vm_amap *, vaddr_t, vsize_t, int)); 129 void amap_wipeout /* remove all anons from amap */ 130 __P((struct vm_amap *)); 131 132 /* 133 * amap flag values 134 */ 135 136 #define AMAP_SHARED 0x1 /* amap is shared */ 137 #define AMAP_REFALL 0x2 /* amap_ref: reference entire amap */ 138 139 #endif /* _KERNEL */ 140 141 /**********************************************************************/ 142 143 /* 144 * part 2: amap implementation-specific info 145 */ 146 147 /* 148 * we currently provide an array-based amap implementation. in this 149 * implementation we provide the option of tracking split references 150 * so that we don't lose track of references during partial unmaps 151 * ... this is enabled with the "UVM_AMAP_PPREF" define. 152 */ 153 154 #define UVM_AMAP_PPREF /* track partial references */ 155 156 /* 157 * here is the definition of the vm_amap structure for this implementation. 158 */ 159 160 struct vm_amap { 161 struct simplelock am_l; /* simple lock [locks all vm_amap fields] */ 162 int am_ref; /* reference count */ 163 int am_flags; /* flags */ 164 int am_maxslot; /* max # of slots allocated */ 165 int am_nslot; /* # of slots currently in map ( <= maxslot) */ 166 int am_nused; /* # of slots currently in use */ 167 int *am_slots; /* contig array of active slots */ 168 int *am_bckptr; /* back pointer array to am_slots */ 169 struct vm_anon **am_anon; /* array of anonymous pages */ 170 #ifdef UVM_AMAP_PPREF 171 int *am_ppref; /* per page reference count (if !NULL) */ 172 #endif 173 }; 174 175 /* 176 * note that am_slots, am_bckptr, and am_anon are arrays. this allows 177 * fast lookup of pages based on their virual address at the expense of 178 * some extra memory. in the future we should be smarter about memory 179 * usage and fall back to a non-array based implementation on systems 180 * that are short of memory (XXXCDC). 181 * 182 * the entries in the array are called slots... for example an amap that 183 * covers four pages of virtual memory is said to have four slots. here 184 * is an example of the array usage for a four slot amap. note that only 185 * slots one and three have anons assigned to them. "D/C" means that we 186 * "don't care" about the value. 187 * 188 * 0 1 2 3 189 * am_anon: NULL, anon0, NULL, anon1 (actual pointers to anons) 190 * am_bckptr: D/C, 1, D/C, 0 (points to am_slots entry) 191 * 192 * am_slots: 3, 1, D/C, D/C (says slots 3 and 1 are in use) 193 * 194 * note that am_bckptr is D/C if the slot in am_anon is set to NULL. 195 * to find the entry in am_slots for an anon, look at am_bckptr[slot], 196 * thus the entry for slot 3 in am_slots[] is at am_slots[am_bckptr[3]]. 197 * in general, if am_anon[X] is non-NULL, then the following must be 198 * true: am_slots[am_bckptr[X]] == X 199 * 200 * note that am_slots is always contig-packed. 201 */ 202 203 /* 204 * defines for handling of large sparce amaps: 205 * 206 * one of the problems of array-based amaps is that if you allocate a 207 * large sparcely-used area of virtual memory you end up allocating 208 * large arrays that, for the most part, don't get used. this is a 209 * problem for BSD in that the kernel likes to make these types of 210 * allocations to "reserve" memory for possible future use. 211 * 212 * for example, the kernel allocates (reserves) a large chunk of user 213 * VM for possible stack growth. most of the time only a page or two 214 * of this VM is actually used. since the stack is anonymous memory 215 * it makes sense for it to live in an amap, but if we allocated an 216 * amap for the entire stack range we could end up wasting a large 217 * amount of malloc'd KVM. 218 * 219 * for example, on the i386 at boot time we allocate two amaps for the stack 220 * of /sbin/init: 221 * 1. a 7680 slot amap at protection 0 (reserve space for stack) 222 * 2. a 512 slot amap at protection 7 (top of stack) 223 * 224 * most of the array allocated for the amaps for this is never used. 225 * the amap interface provides a way for us to avoid this problem by 226 * allowing amap_copy() to break larger amaps up into smaller sized 227 * chunks (controlled by the "canchunk" option). we use this feature 228 * to reduce our memory usage with the BSD stack management. if we 229 * are asked to create an amap with more than UVM_AMAP_LARGE slots in it, 230 * we attempt to break it up into a UVM_AMAP_CHUNK sized amap if the 231 * "canchunk" flag is set. 232 * 233 * so, in the i386 example, the 7680 slot area is never referenced so 234 * nothing gets allocated (amap_copy is never called because the protection 235 * is zero). the 512 slot area for the top of the stack is referenced. 236 * the chunking code breaks it up into 16 slot chunks (hopefully a single 237 * 16 slot chunk is enough to handle the whole stack). 238 */ 239 240 #define UVM_AMAP_LARGE 256 /* # of slots in "large" amap */ 241 #define UVM_AMAP_CHUNK 16 /* # of slots to chunk large amaps in */ 242 243 #ifdef _KERNEL 244 245 /* 246 * macros 247 */ 248 249 /* AMAP_B2SLOT: convert byte offset to slot */ 250 #define AMAP_B2SLOT(S,B) { \ 251 KASSERT(((B) & (PAGE_SIZE - 1)) == 0); \ 252 (S) = (B) >> PAGE_SHIFT; \ 253 } 254 255 /* 256 * lock/unlock/refs/flags macros 257 */ 258 259 #define amap_flags(AMAP) ((AMAP)->am_flags) 260 #define amap_lock(AMAP) simple_lock(&(AMAP)->am_l) 261 #define amap_refs(AMAP) ((AMAP)->am_ref) 262 #define amap_unlock(AMAP) simple_unlock(&(AMAP)->am_l) 263 264 /* 265 * if we enable PPREF, then we have a couple of extra functions that 266 * we need to prototype here... 267 */ 268 269 #ifdef UVM_AMAP_PPREF 270 271 #define PPREF_NONE ((int *) -1) /* not using ppref */ 272 273 void amap_pp_adjref /* adjust references */ 274 __P((struct vm_amap *, int, vsize_t, int)); 275 void amap_pp_establish /* establish ppref */ 276 __P((struct vm_amap *)); 277 void amap_wiperange /* wipe part of an amap */ 278 __P((struct vm_amap *, int, int)); 279 #endif /* UVM_AMAP_PPREF */ 280 281 #endif /* _KERNEL */ 282 283 #endif /* _UVM_UVM_AMAP_H_ */ 284