/original-bsd/sys/vm/ |
H A D | vm.h | 21 struct vm_map; 22 typedef struct vm_map *vm_map_t; 58 struct vm_map vm_map; /* VM address map */ member
|
H A D | vm_map.h | 62 struct vm_map *share_map; /* share map */ 63 struct vm_map *sub_map; /* belongs to another map */ 97 struct vm_map { struct 205 void vm_map_init __P((struct vm_map *,
|
H A D | vm_glue.c | 89 rv = vm_map_check_protection(&curproc->p_vmspace->vm_map, 137 vm_map_pageable(&curproc->p_vmspace->vm_map, trunc_page(addr), 150 vm_map_pageable(&curproc->p_vmspace->vm_map, trunc_page(addr), 178 (void)vm_map_inherit(&p1->p_vmspace->vm_map, 224 { u_int addr = UPT_MIN_ADDRESS - UPAGES*NBPG; struct vm_map *vp; 226 vp = &p2->p_vmspace->vm_map; 490 pmap_collect(vm_map_pmap(&p->p_vmspace->vm_map));
|
H A D | vm_unix.c | 49 rv = vm_allocate(&vm->vm_map, &old, diff, FALSE); 57 rv = vm_deallocate(&vm->vm_map, new, diff);
|
H A D | vm_mmap.c | 274 error = vm_mmap(&p->p_vmspace->vm_map, &addr, size, prot, maxprot, 304 map = &p->p_vmspace->vm_map; 394 map = &p->p_vmspace->vm_map; 450 switch (vm_map_protect(&p->p_vmspace->vm_map, addr, addr+size, prot, 518 if (size + ptoa(pmap_wired_count(vm_map_pmap(&p->p_vmspace->vm_map))) > 526 error = vm_map_pageable(&p->p_vmspace->vm_map, addr, addr+size, FALSE); 557 error = vm_map_pageable(&p->p_vmspace->vm_map, addr, addr+size, TRUE);
|
H A D | vm_map.c | 162 vm_map_init(&vm->vm_map, min, max, pageable); 164 vm->vm_map.pmap = &vm->vm_pmap; /* XXX */ 180 vm_map_lock(&vm->vm_map); 181 (void) vm_map_delete(&vm->vm_map, vm->vm_map.min_offset, 182 vm->vm_map.max_offset); 210 MALLOC(result, vm_map_t, sizeof(struct vm_map), 225 register struct vm_map *map; in vm_map_init() 2090 vm_map_t old_map = &vm1->vm_map; 2103 new_map = &vm2->vm_map; /* XXX */
|
H A D | vm_meter.c | 168 for (map = &p->p_vmspace->vm_map, entry = map->header.next;
|
/original-bsd/sys/kern/ |
H A D | kern_exec.c | 576 (void) vm_map_remove(&vm->vm_map, VM_MIN_ADDRESS, 615 if (vm_allocate(&vm->vm_map, &addr, xts + ctob(ds), FALSE)) { 625 if (vm_allocate(&vm->vm_map, &addr, xts, FALSE)) { 632 if (vm_allocate(&vm->vm_map, &addr, round_page(ctob(ds)), FALSE)) { 644 if (vm_allocate(&vm->vm_map, &addr, size, FALSE)) { 650 if (vm_map_protect(&vm->vm_map, addr, addr+size, VM_PROT_NONE, FALSE)) { 672 (void) vm_map_protect(&vm->vm_map, 684 error = vm_mmap(&vm->vm_map, &addr, size, 688 (void) vm_map_protect(&vm->vm_map, addr, addr + xts, 693 error = vm_mmap(&vm->vm_map, &addr, size, [all …]
|
H A D | init_main.c | 162 vm_map_init(&p->p_vmspace->vm_map, round_page(VM_MIN_ADDRESS), 164 vmspace0.vm_map.pmap = &vmspace0.vm_pmap; 321 if (vm_allocate(&p->p_vmspace->vm_map, &addr, PAGE_SIZE, FALSE) != 0)
|
H A D | sys_process.c | 194 rv = vm_map_protect(&p->p_vmspace->vm_map, sa, ea, 198 (void) vm_map_protect(&p->p_vmspace->vm_map,
|
H A D | sysv_shm.c | 319 error = vm_mmap(&p->p_vmspace->vm_map, (vm_offset_t *)&uva, 423 (void) vm_deallocate(&p->p_vmspace->vm_map, shmd->shmd_uva,
|
/original-bsd/sys/miscfs/procfs/ |
H A D | procfs_mem.c | 83 map = &p->p_vmspace->vm_map; 237 map = &p->p_vmspace->vm_map;
|
/original-bsd/sys/luna68k/luna68k/ |
H A D | vm_machdep.c | 243 pa = pmap_extract(vm_map_pmap(&p->p_vmspace->vm_map), 290 upmap = vm_map_pmap(&curproc->p_vmspace->vm_map); 332 upmap = vm_map_pmap(&curproc->p_vmspace->vm_map);
|
H A D | trap.c | 416 map = &vm->vm_map; 732 if (vm_map_check_protection(&p->p_vmspace->vm_map, 735 !vm_map_check_protection(&p->p_vmspace->vm_map,
|
/original-bsd/sys/hp300/hp300/ |
H A D | vm_machdep.c | 252 pa = pmap_extract(vm_map_pmap(&p->p_vmspace->vm_map), 299 upmap = vm_map_pmap(&curproc->p_vmspace->vm_map); 341 upmap = vm_map_pmap(&curproc->p_vmspace->vm_map);
|
H A D | trap.c | 466 map = &vm->vm_map; 795 if (vm_map_check_protection(&p->p_vmspace->vm_map, 798 !vm_map_check_protection(&p->p_vmspace->vm_map,
|
/original-bsd/sys/pmax/pmax/ |
H A D | kadb.s | 222 rv = vm_map_protect(&p->p_vmspace->vm_map, sa, ea, 226 (void) vm_map_protect(&p->p_vmspace->vm_map, 280 rv = vm_map_protect(&p->p_vmspace->vm_map, sa, ea, 284 (void) vm_map_protect(&p->p_vmspace->vm_map,
|
H A D | kadb.c | 224 rv = vm_map_protect(&p->p_vmspace->vm_map, sa, ea, in kdbsetsstep() 228 (void) vm_map_protect(&p->p_vmspace->vm_map, in kdbsetsstep() 279 rv = vm_map_protect(&p->p_vmspace->vm_map, sa, ea, in kdbclrsstep() 283 (void) vm_map_protect(&p->p_vmspace->vm_map, in kdbclrsstep()
|
H A D | vm_machdep.c | 233 pa = pmap_extract(vm_map_pmap(&p->p_vmspace->vm_map),
|
H A D | trap.c | 321 map = &vm->vm_map; 582 rv = vm_map_protect(&p->p_vmspace->vm_map, sa, ea, 586 (void) vm_map_protect(&p->p_vmspace->vm_map, 1577 rv = vm_map_protect(&p->p_vmspace->vm_map, sa, ea, 1581 (void) vm_map_protect(&p->p_vmspace->vm_map,
|
/original-bsd/sys/luna68k/include/ |
H A D | pmap.h | 53 ((pm) == kernel_pmap || (pm) == curproc->p_vmspace->vm_map.pmap)
|
/original-bsd/sys/hp300/include/ |
H A D | pmap.h | 50 ((pm) == kernel_pmap || (pm) == curproc->p_vmspace->vm_map.pmap)
|
/original-bsd/sys/i386/i386/ |
H A D | trap.c | 179 map = &vm->vm_map; 453 if (user_page_fault (curproc, &curproc->p_vmspace->vm_map,
|
/original-bsd/sys/news3400/news3400/ |
H A D | trap.c | 296 map = &vm->vm_map; 558 rv = vm_map_protect(&p->p_vmspace->vm_map, sa, ea, 562 (void) vm_map_protect(&p->p_vmspace->vm_map, 1068 rv = vm_map_protect(&p->p_vmspace->vm_map, sa, ea, 1072 (void) vm_map_protect(&p->p_vmspace->vm_map,
|
/original-bsd/sys/sparc/sparc/ |
H A D | vm_machdep.c | 89 pa = pmap_extract(vm_map_pmap(&p->p_vmspace->vm_map),
|