xref: /freebsd/sys/vm/vm_extern.h (revision 29d481ae)
1 /*-
2  * SPDX-License-Identifier: BSD-3-Clause
3  *
4  * Copyright (c) 1992, 1993
5  *	The Regents of the University of California.  All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  * 3. Neither the name of the University nor the names of its contributors
16  *    may be used to endorse or promote products derived from this software
17  *    without specific prior written permission.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
20  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
23  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
24  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
25  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
26  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
27  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
28  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29  * SUCH DAMAGE.
30  *
31  *	@(#)vm_extern.h	8.2 (Berkeley) 1/12/94
32  * $FreeBSD$
33  */
34 
35 #ifndef _VM_EXTERN_H_
36 #define	_VM_EXTERN_H_
37 
38 struct pmap;
39 struct proc;
40 struct vmspace;
41 struct vnode;
42 struct vmem;
43 
44 #ifdef _KERNEL
45 #include <sys/systm.h>
46 
47 struct cdev;
48 struct cdevsw;
49 struct domainset;
50 
51 /* These operate on kernel virtual addresses only. */
52 vm_offset_t kva_alloc(vm_size_t);
53 void kva_free(vm_offset_t, vm_size_t);
54 
55 /* These operate on pageable virtual addresses. */
56 vm_offset_t kmap_alloc_wait(vm_map_t, vm_size_t);
57 void kmap_free_wakeup(vm_map_t, vm_offset_t, vm_size_t);
58 
59 /* These operate on virtual addresses backed by memory. */
60 vm_offset_t kmem_alloc_attr(vm_size_t size, int flags,
61     vm_paddr_t low, vm_paddr_t high, vm_memattr_t memattr);
62 vm_offset_t kmem_alloc_attr_domainset(struct domainset *ds, vm_size_t size,
63     int flags, vm_paddr_t low, vm_paddr_t high, vm_memattr_t memattr);
64 vm_offset_t kmem_alloc_contig(vm_size_t size, int flags,
65     vm_paddr_t low, vm_paddr_t high, u_long alignment, vm_paddr_t boundary,
66     vm_memattr_t memattr);
67 vm_offset_t kmem_alloc_contig_domainset(struct domainset *ds, vm_size_t size,
68     int flags, vm_paddr_t low, vm_paddr_t high, u_long alignment,
69     vm_paddr_t boundary, vm_memattr_t memattr);
70 vm_offset_t kmem_malloc(vm_size_t size, int flags);
71 vm_offset_t kmem_malloc_domainset(struct domainset *ds, vm_size_t size,
72     int flags);
73 void kmem_free(vm_offset_t addr, vm_size_t size);
74 
75 /* This provides memory for previously allocated address space. */
76 int kmem_back(vm_object_t, vm_offset_t, vm_size_t, int);
77 int kmem_back_domain(int, vm_object_t, vm_offset_t, vm_size_t, int);
78 void kmem_unback(vm_object_t, vm_offset_t, vm_size_t);
79 
80 /* Bootstrapping. */
81 void kmem_bootstrap_free(vm_offset_t, vm_size_t);
82 void kmem_subinit(vm_map_t, vm_map_t, vm_offset_t *, vm_offset_t *, vm_size_t,
83     bool);
84 void kmem_init(vm_offset_t, vm_offset_t);
85 void kmem_init_zero_region(void);
86 void kmeminit(void);
87 
88 int kernacc(void *, int, int);
89 int useracc(void *, int, int);
90 int vm_fault(vm_map_t map, vm_offset_t vaddr, vm_prot_t fault_type,
91     int fault_flags, vm_page_t *m_hold);
92 void vm_fault_copy_entry(vm_map_t, vm_map_t, vm_map_entry_t, vm_map_entry_t,
93     vm_ooffset_t *);
94 int vm_fault_disable_pagefaults(void);
95 void vm_fault_enable_pagefaults(int save);
96 int vm_fault_quick_hold_pages(vm_map_t map, vm_offset_t addr, vm_size_t len,
97     vm_prot_t prot, vm_page_t *ma, int max_count);
98 int vm_fault_trap(vm_map_t map, vm_offset_t vaddr, vm_prot_t fault_type,
99     int fault_flags, int *signo, int *ucode);
100 int vm_forkproc(struct thread *, struct proc *, struct thread *,
101     struct vmspace *, int);
102 void vm_waitproc(struct proc *);
103 int vm_mmap(vm_map_t, vm_offset_t *, vm_size_t, vm_prot_t, vm_prot_t, int,
104     objtype_t, void *, vm_ooffset_t);
105 int vm_mmap_object(vm_map_t, vm_offset_t *, vm_size_t, vm_prot_t,
106     vm_prot_t, int, vm_object_t, vm_ooffset_t, boolean_t, struct thread *);
107 int vm_mmap_to_errno(int rv);
108 int vm_mmap_cdev(struct thread *, vm_size_t, vm_prot_t, vm_prot_t *,
109     int *, struct cdev *, struct cdevsw *, vm_ooffset_t *, vm_object_t *);
110 int vm_mmap_vnode(struct thread *, vm_size_t, vm_prot_t, vm_prot_t *, int *,
111     struct vnode *, vm_ooffset_t *, vm_object_t *, boolean_t *);
112 void vm_set_page_size(void);
113 void vm_sync_icache(vm_map_t, vm_offset_t, vm_size_t);
114 typedef int (*pmap_pinit_t)(struct pmap *pmap);
115 struct vmspace *vmspace_alloc(vm_offset_t, vm_offset_t, pmap_pinit_t);
116 struct vmspace *vmspace_fork(struct vmspace *, vm_ooffset_t *);
117 int vmspace_exec(struct proc *, vm_offset_t, vm_offset_t);
118 int vmspace_unshare(struct proc *);
119 void vmspace_exit(struct thread *);
120 struct vmspace *vmspace_acquire_ref(struct proc *);
121 void vmspace_free(struct vmspace *);
122 void vmspace_exitfree(struct proc *);
123 void vmspace_switch_aio(struct vmspace *);
124 void vnode_pager_setsize(struct vnode *, vm_ooffset_t);
125 void vnode_pager_purge_range(struct vnode *, vm_ooffset_t, vm_ooffset_t);
126 int vslock(void *, size_t);
127 void vsunlock(void *, size_t);
128 struct sf_buf *vm_imgact_map_page(vm_object_t object, vm_ooffset_t offset);
129 void vm_imgact_unmap_page(struct sf_buf *sf);
130 void vm_thread_dispose(struct thread *td);
131 int vm_thread_new(struct thread *td, int pages);
132 void vm_thread_stack_back(struct domainset *ds, vm_offset_t kaddr,
133     vm_page_t ma[], int npages, int req_class);
134 u_int vm_active_count(void);
135 u_int vm_inactive_count(void);
136 u_int vm_laundry_count(void);
137 u_int vm_wait_count(void);
138 
139 /*
140  * Is pa a multiple of alignment, which is a power-of-two?
141  */
142 static inline bool
143 vm_addr_align_ok(vm_paddr_t pa, u_long alignment)
144 {
145 	KASSERT(powerof2(alignment), ("%s: alignment is not a power of 2: %#lx",
146 	    __func__, alignment));
147 	return ((pa & (alignment - 1)) == 0);
148 }
149 
150 /*
151  * Do the first and last addresses of a range match in all bits except the ones
152  * in -boundary (a power-of-two)?  For boundary == 0, all addresses match.
153  */
154 static inline bool
155 vm_addr_bound_ok(vm_paddr_t pa, vm_paddr_t size, vm_paddr_t boundary)
156 {
157 	KASSERT(powerof2(boundary), ("%s: boundary is not a power of 2: %#jx",
158 	    __func__, (uintmax_t)boundary));
159 	return (((pa ^ (pa + size - 1)) & -boundary) == 0);
160 }
161 
162 static inline bool
163 vm_addr_ok(vm_paddr_t pa, vm_paddr_t size, u_long alignment,
164     vm_paddr_t boundary)
165 {
166 	return (vm_addr_align_ok(pa, alignment) &&
167 	    vm_addr_bound_ok(pa, size, boundary));
168 }
169 #endif				/* _KERNEL */
170 #endif				/* !_VM_EXTERN_H_ */
171