1 /*-
2 * Copyright (c) 2010 Isilon Systems, Inc.
3 * Copyright (c) 2010 iX Systems, Inc.
4 * Copyright (c) 2010 Panasas, Inc.
5 * Copyright (c) 2013-2021 Mellanox Technologies, Ltd.
6 * All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 * notice unmodified, this list of conditions, and the following
13 * disclaimer.
14 * 2. Redistributions in binary form must reproduce the above copyright
15 * notice, this list of conditions and the following disclaimer in the
16 * documentation and/or other materials provided with the distribution.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
19 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
20 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
21 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
22 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
23 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
27 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28 */
29 #ifndef _LINUXKPI_LINUX_SLAB_H_
30 #define _LINUXKPI_LINUX_SLAB_H_
31
32 #include <sys/types.h>
33 #include <sys/malloc.h>
34 #include <sys/limits.h>
35
36 #include <linux/compat.h>
37 #include <linux/types.h>
38 #include <linux/gfp.h>
39 #include <linux/llist.h>
40 #include <linux/overflow.h>
41
42 MALLOC_DECLARE(M_KMALLOC);
43
44 #define kmalloc(size, flags) lkpi_kmalloc(size, flags)
45 #define kvmalloc(size, flags) kmalloc(size, flags)
46 #define kvzalloc(size, flags) kmalloc(size, (flags) | __GFP_ZERO)
47 #define kvcalloc(n, size, flags) kvmalloc_array(n, size, (flags) | __GFP_ZERO)
48 #define kzalloc(size, flags) kmalloc(size, (flags) | __GFP_ZERO)
49 #define kzalloc_node(size, flags, node) kmalloc_node(size, (flags) | __GFP_ZERO, node)
50 #define kfree_const(ptr) kfree(ptr)
51 #define vzalloc(size) __vmalloc(size, GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO, 0)
52 #define vfree(arg) kfree(arg)
53 #define kvfree(arg) kfree(arg)
54 #define vmalloc_node(size, node) __vmalloc_node(size, GFP_KERNEL, node)
55 #define vmalloc_user(size) __vmalloc(size, GFP_KERNEL | __GFP_ZERO, 0)
56 #define vmalloc(size) __vmalloc(size, GFP_KERNEL, 0)
57
58 /*
59 * Prefix some functions with linux_ to avoid namespace conflict
60 * with the OpenSolaris code in the kernel.
61 */
62 #define kmem_cache linux_kmem_cache
63 #define kmem_cache_create(...) linux_kmem_cache_create(__VA_ARGS__)
64 #define kmem_cache_alloc(...) lkpi_kmem_cache_alloc(__VA_ARGS__)
65 #define kmem_cache_zalloc(...) lkpi_kmem_cache_zalloc(__VA_ARGS__)
66 #define kmem_cache_free(...) lkpi_kmem_cache_free(__VA_ARGS__)
67 #define kmem_cache_destroy(...) linux_kmem_cache_destroy(__VA_ARGS__)
68 #define kmem_cache_shrink(x) (0)
69
70 #define KMEM_CACHE(__struct, flags) \
71 linux_kmem_cache_create(#__struct, sizeof(struct __struct), \
72 __alignof(struct __struct), (flags), NULL)
73
74 typedef void linux_kmem_ctor_t (void *);
75
76 struct linux_kmem_cache;
77
78 #define SLAB_HWCACHE_ALIGN (1 << 0)
79 #define SLAB_TYPESAFE_BY_RCU (1 << 1)
80 #define SLAB_RECLAIM_ACCOUNT (1 << 2)
81
82 #define SLAB_DESTROY_BY_RCU \
83 SLAB_TYPESAFE_BY_RCU
84
85 #define ARCH_KMALLOC_MINALIGN \
86 __alignof(unsigned long long)
87
88 /* drm-kmod 5.4 compat */
89 #define kfree_async(ptr) kfree(ptr);
90
91 #define ZERO_SIZE_PTR ((void *)16)
92 #define ZERO_OR_NULL_PTR(x) ((x) == NULL || (x) == ZERO_SIZE_PTR)
93
94 static inline gfp_t
linux_check_m_flags(gfp_t flags)95 linux_check_m_flags(gfp_t flags)
96 {
97 const gfp_t m = M_NOWAIT | M_WAITOK;
98
99 /* make sure either M_NOWAIT or M_WAITOK is set */
100 if ((flags & m) == 0)
101 flags |= M_NOWAIT;
102 else if ((flags & m) == m)
103 flags &= ~M_WAITOK;
104
105 /* mask away LinuxKPI specific flags */
106 return (flags & GFP_NATIVE_MASK);
107 }
108
109 static inline void *
__kmalloc(size_t size,gfp_t flags)110 __kmalloc(size_t size, gfp_t flags)
111 {
112 return (malloc(MAX(size, sizeof(struct llist_node)), M_KMALLOC,
113 linux_check_m_flags(flags)));
114 }
115
116 static inline void *
kmalloc_node(size_t size,gfp_t flags,int node)117 kmalloc_node(size_t size, gfp_t flags, int node)
118 {
119 return (malloc_domainset(size, M_KMALLOC,
120 linux_get_vm_domain_set(node), linux_check_m_flags(flags)));
121 }
122
123 static inline void *
kcalloc(size_t n,size_t size,gfp_t flags)124 kcalloc(size_t n, size_t size, gfp_t flags)
125 {
126 flags |= __GFP_ZERO;
127 return (mallocarray(n, size, M_KMALLOC, linux_check_m_flags(flags)));
128 }
129
130 static inline void *
kcalloc_node(size_t n,size_t size,gfp_t flags,int node)131 kcalloc_node(size_t n, size_t size, gfp_t flags, int node)
132 {
133 flags |= __GFP_ZERO;
134 return (mallocarray_domainset(n, size, M_KMALLOC,
135 linux_get_vm_domain_set(node), linux_check_m_flags(flags)));
136 }
137
138 static inline void *
__vmalloc(size_t size,gfp_t flags,int other)139 __vmalloc(size_t size, gfp_t flags, int other)
140 {
141 return (malloc(size, M_KMALLOC, linux_check_m_flags(flags)));
142 }
143
144 static inline void *
__vmalloc_node(size_t size,gfp_t flags,int node)145 __vmalloc_node(size_t size, gfp_t flags, int node)
146 {
147 return (malloc_domainset(size, M_KMALLOC,
148 linux_get_vm_domain_set(node), linux_check_m_flags(flags)));
149 }
150
151 static inline void *
vmalloc_32(size_t size)152 vmalloc_32(size_t size)
153 {
154 return (contigmalloc(size, M_KMALLOC, M_WAITOK, 0, UINT_MAX, 1, 1));
155 }
156
157 static inline void *
kmalloc_array(size_t n,size_t size,gfp_t flags)158 kmalloc_array(size_t n, size_t size, gfp_t flags)
159 {
160 return (mallocarray(n, size, M_KMALLOC, linux_check_m_flags(flags)));
161 }
162
163 static inline void *
kmalloc_array_node(size_t n,size_t size,gfp_t flags,int node)164 kmalloc_array_node(size_t n, size_t size, gfp_t flags, int node)
165 {
166 return (mallocarray_domainset(n, size, M_KMALLOC,
167 linux_get_vm_domain_set(node), linux_check_m_flags(flags)));
168 }
169
170 static inline void *
kvmalloc_array(size_t n,size_t size,gfp_t flags)171 kvmalloc_array(size_t n, size_t size, gfp_t flags)
172 {
173 return (mallocarray(n, size, M_KMALLOC, linux_check_m_flags(flags)));
174 }
175
176 static inline void *
krealloc(void * ptr,size_t size,gfp_t flags)177 krealloc(void *ptr, size_t size, gfp_t flags)
178 {
179 return (realloc(ptr, size, M_KMALLOC, linux_check_m_flags(flags)));
180 }
181
182 static inline void *
krealloc_array(void * ptr,size_t n,size_t size,gfp_t flags)183 krealloc_array(void *ptr, size_t n, size_t size, gfp_t flags)
184 {
185 if (WOULD_OVERFLOW(n, size)) {
186 return NULL;
187 }
188
189 return (realloc(ptr, n * size, M_KMALLOC, linux_check_m_flags(flags)));
190 }
191
192 extern void linux_kfree_async(void *);
193
194 static inline void
kfree(const void * ptr)195 kfree(const void *ptr)
196 {
197 if (ZERO_OR_NULL_PTR(ptr))
198 return;
199
200 if (curthread->td_critnest != 0)
201 linux_kfree_async(__DECONST(void *, ptr));
202 else
203 free(__DECONST(void *, ptr), M_KMALLOC);
204 }
205
206 static __inline void
kfree_sensitive(const void * ptr)207 kfree_sensitive(const void *ptr)
208 {
209 if (ZERO_OR_NULL_PTR(ptr))
210 return;
211
212 zfree(__DECONST(void *, ptr), M_KMALLOC);
213 }
214
215 static inline size_t
ksize(const void * ptr)216 ksize(const void *ptr)
217 {
218 return (malloc_usable_size(ptr));
219 }
220
221 extern void *lkpi_kmalloc(size_t size, gfp_t flags);
222 extern struct linux_kmem_cache *linux_kmem_cache_create(const char *name,
223 size_t size, size_t align, unsigned flags, linux_kmem_ctor_t *ctor);
224 extern void *lkpi_kmem_cache_alloc(struct linux_kmem_cache *, gfp_t);
225 extern void *lkpi_kmem_cache_zalloc(struct linux_kmem_cache *, gfp_t);
226 extern void lkpi_kmem_cache_free(struct linux_kmem_cache *, void *);
227 extern void linux_kmem_cache_destroy(struct linux_kmem_cache *);
228
229 #endif /* _LINUXKPI_LINUX_SLAB_H_ */
230