1 /*	$NetBSD: pool.h,v 1.54 2006/08/20 09:35:25 yamt Exp $	*/
2 
3 /*-
4  * Copyright (c) 1997, 1998, 1999, 2000 The NetBSD Foundation, Inc.
5  * All rights reserved.
6  *
7  * This code is derived from software contributed to The NetBSD Foundation
8  * by Paul Kranenburg; by Jason R. Thorpe of the Numerical Aerospace
9  * Simulation Facility, NASA Ames Research Center.
10  *
11  * Redistribution and use in source and binary forms, with or without
12  * modification, are permitted provided that the following conditions
13  * are met:
14  * 1. Redistributions of source code must retain the above copyright
15  *    notice, this list of conditions and the following disclaimer.
16  * 2. Redistributions in binary form must reproduce the above copyright
17  *    notice, this list of conditions and the following disclaimer in the
18  *    documentation and/or other materials provided with the distribution.
19  * 3. All advertising materials mentioning features or use of this software
20  *    must display the following acknowledgement:
21  *	This product includes software developed by the NetBSD
22  *	Foundation, Inc. and its contributors.
23  * 4. Neither the name of The NetBSD Foundation nor the names of its
24  *    contributors may be used to endorse or promote products derived
25  *    from this software without specific prior written permission.
26  *
27  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
28  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
29  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
30  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
31  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
32  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
33  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
34  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
35  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
36  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
37  * POSSIBILITY OF SUCH DAMAGE.
38  */
39 
40 #ifndef _SYS_POOL_H_
41 #define _SYS_POOL_H_
42 
43 #ifdef _KERNEL
44 #define	__POOL_EXPOSE
45 #endif
46 
47 #if defined(_KERNEL_OPT)
48 #include "opt_pool.h"
49 #endif
50 
51 #ifdef __POOL_EXPOSE
52 #include <sys/lock.h>
53 #include <sys/queue.h>
54 #include <sys/time.h>
55 #include <sys/tree.h>
56 #if defined(_KERNEL)
57 #include <sys/callback.h>
58 #endif /* defined(_KERNEL) */
59 #endif
60 
61 #define	PCG_NOBJECTS		16
62 
63 #define	POOL_PADDR_INVALID	((paddr_t) -1)
64 
65 #ifdef __POOL_EXPOSE
66 /* The pool cache group. */
67 struct pool_cache_group {
68 	LIST_ENTRY(pool_cache_group)
69 		pcg_list;	/* link in the pool cache's group list */
70 	u_int	pcg_avail;	/* # available objects */
71 				/* pointers to the objects */
72 	struct {
73 		void *pcgo_va;	/* cache object virtual address */
74 		paddr_t pcgo_pa;/* cache object physical address */
75 	} pcg_objects[PCG_NOBJECTS];
76 };
77 
78 LIST_HEAD(pool_cache_grouplist,pool_cache_group);
79 struct pool_cache {
80 	LIST_ENTRY(pool_cache)
81 			pc_poollist;	/* entry on pool's group list */
82 	struct pool_cache_grouplist
83 			pc_emptygroups;	/* list of empty cache groups */
84 	struct pool_cache_grouplist
85 			pc_fullgroups;	/* list of full cache groups */
86 	struct pool_cache_grouplist
87 			pc_partgroups;	/* list of partial cache groups */
88 	struct pool	*pc_pool;	/* parent pool */
89 	struct simplelock pc_slock;	/* mutex */
90 
91 	int		(*pc_ctor)(void *, void *, int);
92 	void		(*pc_dtor)(void *, void *);
93 	void		*pc_arg;
94 
95 	/* Statistics. */
96 	unsigned long	pc_hits;	/* cache hits */
97 	unsigned long	pc_misses;	/* cache misses */
98 
99 	unsigned long	pc_ngroups;	/* # cache groups */
100 
101 	unsigned long	pc_nitems;	/* # objects currently in cache */
102 };
103 
104 struct pool_allocator {
105 	void		*(*pa_alloc)(struct pool *, int);
106 	void		(*pa_free)(struct pool *, void *);
107 	unsigned int	pa_pagesz;
108 
109 	/* The following fields are for internal use only. */
110 	struct simplelock pa_slock;
111 	TAILQ_HEAD(, pool) pa_list;	/* list of pools using this allocator */
112 	int		pa_flags;
113 #define	PA_INITIALIZED	0x01
114 	int		pa_pagemask;
115 	int		pa_pageshift;
116 	struct vm_map *pa_backingmap;
117 #if defined(_KERNEL)
118 	struct vm_map **pa_backingmapptr;
119 	SLIST_ENTRY(pool_allocator) pa_q;
120 #endif /* defined(_KERNEL) */
121 };
122 
123 LIST_HEAD(pool_pagelist,pool_item_header);
124 
125 struct pool {
126 	LIST_ENTRY(pool)
127 			pr_poollist;
128 	struct pool_pagelist
129 			pr_emptypages;	/* Empty pages */
130 	struct pool_pagelist
131 			pr_fullpages;	/* Full pages */
132 	struct pool_pagelist
133 			pr_partpages;	/* Partially-allocated pages */
134 	struct pool_item_header	*pr_curpage;
135 	struct pool	*pr_phpool;	/* Pool item header pool */
136 	LIST_HEAD(,pool_cache)
137 			pr_cachelist;	/* Caches for this pool */
138 	unsigned int	pr_size;	/* Size of item */
139 	unsigned int	pr_align;	/* Requested alignment, must be 2^n */
140 	unsigned int	pr_itemoffset;	/* Align this offset in item */
141 	unsigned int	pr_minitems;	/* minimum # of items to keep */
142 	unsigned int	pr_minpages;	/* same in page units */
143 	unsigned int	pr_maxpages;	/* maximum # of pages to keep */
144 	unsigned int	pr_npages;	/* # of pages allocated */
145 	unsigned int	pr_itemsperpage;/* # items that fit in a page */
146 	unsigned int	pr_slack;	/* unused space in a page */
147 	unsigned int	pr_nitems;	/* number of available items in pool */
148 	unsigned int	pr_nout;	/* # items currently allocated */
149 	unsigned int	pr_hardlimit;	/* hard limit to number of allocated
150 					   items */
151 	struct pool_allocator *pr_alloc;/* back-end allocator */
152 	TAILQ_ENTRY(pool) pr_alloc_list;/* link on allocator's pool list */
153 
154 	/* Drain hook. */
155 	void		(*pr_drain_hook)(void *, int);
156 	void		*pr_drain_hook_arg;
157 
158 	const char	*pr_wchan;	/* tsleep(9) identifier */
159 	unsigned int	pr_flags;	/* r/w flags */
160 	unsigned int	pr_roflags;	/* r/o flags */
161 #define	PR_NOWAIT	0x00		/* for symmetry */
162 #define PR_WAITOK	0x02
163 #define PR_WANTED	0x04
164 #define PR_PHINPAGE	0x40
165 #define PR_LOGGING	0x80
166 #define PR_LIMITFAIL	0x100	/* even if waiting, fail if we hit limit */
167 #define PR_RECURSIVE	0x200	/* pool contains pools, for vmstat(8) */
168 #define PR_NOTOUCH	0x400	/* don't use free items to keep internal state*/
169 #define PR_NOALIGN	0x800	/* don't assume backend alignment */
170 
171 	/*
172 	 * `pr_slock' protects the pool's data structures when removing
173 	 * items from or returning items to the pool, or when reading
174 	 * or updating read/write fields in the pool descriptor.
175 	 *
176 	 * We assume back-end page allocators provide their own locking
177 	 * scheme.  They will be called with the pool descriptor _unlocked_,
178 	 * since the page allocators may block.
179 	 */
180 	struct simplelock	pr_slock;
181 
182 	SPLAY_HEAD(phtree, pool_item_header) pr_phtree;
183 
184 	int		pr_maxcolor;	/* Cache colouring */
185 	int		pr_curcolor;
186 	int		pr_phoffset;	/* Offset in page of page header */
187 
188 	/*
189 	 * Warning message to be issued, and a per-time-delta rate cap,
190 	 * if the hard limit is reached.
191 	 */
192 	const char	*pr_hardlimit_warning;
193 	struct timeval	pr_hardlimit_ratecap;
194 	struct timeval	pr_hardlimit_warning_last;
195 
196 	/*
197 	 * Instrumentation
198 	 */
199 	unsigned long	pr_nget;	/* # of successful requests */
200 	unsigned long	pr_nfail;	/* # of unsuccessful requests */
201 	unsigned long	pr_nput;	/* # of releases */
202 	unsigned long	pr_npagealloc;	/* # of pages allocated */
203 	unsigned long	pr_npagefree;	/* # of pages released */
204 	unsigned int	pr_hiwat;	/* max # of pages in pool */
205 	unsigned long	pr_nidle;	/* # of idle pages */
206 
207 	/*
208 	 * Diagnostic aides.
209 	 */
210 	struct pool_log	*pr_log;
211 	int		pr_curlogentry;
212 	int		pr_logsize;
213 
214 	const char	*pr_entered_file; /* reentrancy check */
215 	long		pr_entered_line;
216 
217 #if defined(_KERNEL)
218 	struct callback_entry pr_reclaimerentry;
219 #endif
220 };
221 #endif /* __POOL_EXPOSE */
222 
223 #ifdef _KERNEL
224 /*
225  * pool_allocator_kmem is the default that all pools get unless
226  * otherwise specified.  pool_allocator_nointr is provided for
227  * pools that know they will never be accessed in interrupt
228  * context.
229  */
230 extern struct pool_allocator pool_allocator_kmem;
231 extern struct pool_allocator pool_allocator_nointr;
232 #ifdef POOL_SUBPAGE
233 /* The above are subpage allocators in this case. */
234 extern struct pool_allocator pool_allocator_kmem_fullpage;
235 extern struct pool_allocator pool_allocator_nointr_fullpage;
236 #endif
237 
238 struct link_pool_init {	/* same as args to pool_init() */
239 	struct pool *pp;
240 	size_t size;
241 	u_int align;
242 	u_int align_offset;
243 	int flags;
244 	const char *wchan;
245 	struct pool_allocator *palloc;
246 };
247 #define	POOL_INIT(pp, size, align, align_offset, flags, wchan, palloc)	\
248 struct pool pp;								\
249 static const struct link_pool_init _link_ ## pp[1] = {			\
250 	{ &pp, size, align, align_offset, flags, wchan, palloc }	\
251 };									\
252 __link_set_add_rodata(pools, _link_ ## pp)
253 
254 void		pool_subsystem_init(void);
255 
256 void		pool_init(struct pool *, size_t, u_int, u_int,
257 		    int, const char *, struct pool_allocator *);
258 void		pool_destroy(struct pool *);
259 
260 void		pool_set_drain_hook(struct pool *,
261 		    void (*)(void *, int), void *);
262 
263 void		*pool_get(struct pool *, int);
264 void		pool_put(struct pool *, void *);
265 int		pool_reclaim(struct pool *);
266 
267 #ifdef POOL_DIAGNOSTIC
268 /*
269  * These versions do reentrancy checking.
270  */
271 void		*_pool_get(struct pool *, int, const char *, long);
272 void		_pool_put(struct pool *, void *, const char *, long);
273 int		_pool_reclaim(struct pool *, const char *, long);
274 #define		pool_get(h, f)	_pool_get((h), (f), __FILE__, __LINE__)
275 #define		pool_put(h, v)	_pool_put((h), (v), __FILE__, __LINE__)
276 #define		pool_reclaim(h)	_pool_reclaim((h), __FILE__, __LINE__)
277 #endif /* POOL_DIAGNOSTIC */
278 
279 int		pool_prime(struct pool *, int);
280 void		pool_setlowat(struct pool *, int);
281 void		pool_sethiwat(struct pool *, int);
282 void		pool_sethardlimit(struct pool *, int, const char *, int);
283 void		pool_drain(void *);
284 
285 /*
286  * Debugging and diagnostic aides.
287  */
288 void		pool_print(struct pool *, const char *);
289 void		pool_printit(struct pool *, const char *,
290 		    void (*)(const char *, ...));
291 void		pool_printall(const char *, void (*)(const char *, ...));
292 int		pool_chk(struct pool *, const char *);
293 
294 /*
295  * Pool cache routines.
296  */
297 void		pool_cache_init(struct pool_cache *, struct pool *,
298 		    int (*)(void *, void *, int),
299 		    void (*)(void *, void *),
300 		    void *);
301 void		pool_cache_destroy(struct pool_cache *);
302 void		*pool_cache_get_paddr(struct pool_cache *, int, paddr_t *);
303 #define		pool_cache_get(pc, f) pool_cache_get_paddr((pc), (f), NULL)
304 void		pool_cache_put_paddr(struct pool_cache *, void *, paddr_t);
305 #define		pool_cache_put(pc, o) pool_cache_put_paddr((pc), (o), \
306 				          POOL_PADDR_INVALID)
307 void		pool_cache_destruct_object(struct pool_cache *, void *);
308 void		pool_cache_invalidate(struct pool_cache *);
309 #endif /* _KERNEL */
310 
311 #endif /* _SYS_POOL_H_ */
312