xref: /freebsd/sys/sys/systm.h (revision 42249ef2)
1 /*-
2  * SPDX-License-Identifier: BSD-3-Clause
3  *
4  * Copyright (c) 1982, 1988, 1991, 1993
5  *	The Regents of the University of California.  All rights reserved.
6  * (c) UNIX System Laboratories, Inc.
7  * All or some portions of this file are derived from material licensed
8  * to the University of California by American Telephone and Telegraph
9  * Co. or Unix System Laboratories, Inc. and are reproduced herein with
10  * the permission of UNIX System Laboratories, Inc.
11  *
12  * Redistribution and use in source and binary forms, with or without
13  * modification, are permitted provided that the following conditions
14  * are met:
15  * 1. Redistributions of source code must retain the above copyright
16  *    notice, this list of conditions and the following disclaimer.
17  * 2. Redistributions in binary form must reproduce the above copyright
18  *    notice, this list of conditions and the following disclaimer in the
19  *    documentation and/or other materials provided with the distribution.
20  * 3. Neither the name of the University nor the names of its contributors
21  *    may be used to endorse or promote products derived from this software
22  *    without specific prior written permission.
23  *
24  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
25  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
26  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
27  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
28  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
29  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
30  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
31  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
32  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
33  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
34  * SUCH DAMAGE.
35  *
36  *	@(#)systm.h	8.7 (Berkeley) 3/29/95
37  * $FreeBSD$
38  */
39 
40 #ifndef _SYS_SYSTM_H_
41 #define	_SYS_SYSTM_H_
42 
43 #include <sys/cdefs.h>
44 #include <machine/atomic.h>
45 #include <machine/cpufunc.h>
46 #include <sys/callout.h>
47 #include <sys/queue.h>
48 #include <sys/stdint.h>		/* for people using printf mainly */
49 
50 __NULLABILITY_PRAGMA_PUSH
51 
52 extern int cold;		/* nonzero if we are doing a cold boot */
53 extern int suspend_blocked;	/* block suspend due to pending shutdown */
54 extern int rebooting;		/* kern_reboot() has been called. */
55 extern const char *panicstr;	/* panic message */
56 extern char version[];		/* system version */
57 extern char compiler_version[];	/* compiler version */
58 extern char copyright[];	/* system copyright */
59 extern int kstack_pages;	/* number of kernel stack pages */
60 
61 extern u_long pagesizes[];	/* supported page sizes */
62 extern long physmem;		/* physical memory */
63 extern long realmem;		/* 'real' memory */
64 
65 extern char *rootdevnames[2];	/* names of possible root devices */
66 
67 extern int boothowto;		/* reboot flags, from console subsystem */
68 extern int bootverbose;		/* nonzero to print verbose messages */
69 
70 extern int maxusers;		/* system tune hint */
71 extern int ngroups_max;		/* max # of supplemental groups */
72 extern int vm_guest;		/* Running as virtual machine guest? */
73 
74 /*
75  * Detected virtual machine guest types. The intention is to expand
76  * and/or add to the VM_GUEST_VM type if specific VM functionality is
77  * ever implemented (e.g. vendor-specific paravirtualization features).
78  * Keep in sync with vm_guest_sysctl_names[].
79  */
80 enum VM_GUEST { VM_GUEST_NO = 0, VM_GUEST_VM, VM_GUEST_XEN, VM_GUEST_HV,
81 		VM_GUEST_VMWARE, VM_GUEST_KVM, VM_GUEST_BHYVE, VM_GUEST_VBOX,
82 		VM_GUEST_PARALLELS, VM_LAST };
83 
84 /*
85  * These functions need to be declared before the KASSERT macro is invoked in
86  * !KASSERT_PANIC_OPTIONAL builds, so their declarations are sort of out of
87  * place compared to other function definitions in this header.  On the other
88  * hand, this header is a bit disorganized anyway.
89  */
90 void	panic(const char *, ...) __dead2 __printflike(1, 2);
91 void	vpanic(const char *, __va_list) __dead2 __printflike(1, 0);
92 
93 #if defined(WITNESS) || defined(INVARIANT_SUPPORT)
94 #ifdef KASSERT_PANIC_OPTIONAL
95 void	kassert_panic(const char *fmt, ...)  __printflike(1, 2);
96 #else
97 #define kassert_panic	panic
98 #endif
99 #endif
100 
101 #ifdef	INVARIANTS		/* The option is always available */
102 #define	KASSERT(exp,msg) do {						\
103 	if (__predict_false(!(exp)))					\
104 		kassert_panic msg;					\
105 } while (0)
106 #define	VNASSERT(exp, vp, msg) do {					\
107 	if (__predict_false(!(exp))) {					\
108 		vn_printf(vp, "VNASSERT failed\n");			\
109 		kassert_panic msg;					\
110 	}								\
111 } while (0)
112 #else
113 #define	KASSERT(exp,msg) do { \
114 } while (0)
115 
116 #define	VNASSERT(exp, vp, msg) do { \
117 } while (0)
118 #endif
119 
120 #ifndef CTASSERT	/* Allow lint to override */
121 #define	CTASSERT(x)	_Static_assert(x, "compile-time assertion failed")
122 #endif
123 
124 #if defined(_KERNEL)
125 #include <sys/param.h>		/* MAXCPU */
126 #include <sys/pcpu.h>		/* curthread */
127 #include <sys/kpilite.h>
128 #endif
129 
130 /*
131  * Assert that a pointer can be loaded from memory atomically.
132  *
133  * This assertion enforces stronger alignment than necessary.  For example,
134  * on some architectures, atomicity for unaligned loads will depend on
135  * whether or not the load spans multiple cache lines.
136  */
137 #define	ASSERT_ATOMIC_LOAD_PTR(var, msg)				\
138 	KASSERT(sizeof(var) == sizeof(void *) &&			\
139 	    ((uintptr_t)&(var) & (sizeof(void *) - 1)) == 0, msg)
140 
141 /*
142  * Assert that a thread is in critical(9) section.
143  */
144 #define	CRITICAL_ASSERT(td)						\
145 	KASSERT((td)->td_critnest >= 1, ("Not in critical section"));
146 
147 /*
148  * If we have already panic'd and this is the thread that called
149  * panic(), then don't block on any mutexes but silently succeed.
150  * Otherwise, the kernel will deadlock since the scheduler isn't
151  * going to run the thread that holds any lock we need.
152  */
153 #define	SCHEDULER_STOPPED_TD(td)  ({					\
154 	MPASS((td) == curthread);					\
155 	__predict_false((td)->td_stopsched);				\
156 })
157 #define	SCHEDULER_STOPPED() SCHEDULER_STOPPED_TD(curthread)
158 
159 /*
160  * Align variables.
161  */
162 #define	__read_mostly		__section(".data.read_mostly")
163 #define	__read_frequently	__section(".data.read_frequently")
164 #define	__exclusive_cache_line	__aligned(CACHE_LINE_SIZE) \
165 				    __section(".data.exclusive_cache_line")
166 /*
167  * XXX the hints declarations are even more misplaced than most declarations
168  * in this file, since they are needed in one file (per arch) and only used
169  * in two files.
170  * XXX most of these variables should be const.
171  */
172 extern int osreldate;
173 extern bool dynamic_kenv;
174 extern struct mtx kenv_lock;
175 extern char *kern_envp;
176 extern char *md_envp;
177 extern char static_env[];
178 extern char static_hints[];	/* by config for now */
179 
180 extern char **kenvp;
181 
182 extern const void *zero_region;	/* address space maps to a zeroed page	*/
183 
184 extern int unmapped_buf_allowed;
185 
186 #ifdef __LP64__
187 #define	IOSIZE_MAX		iosize_max()
188 #define	DEVFS_IOSIZE_MAX	devfs_iosize_max()
189 #else
190 #define	IOSIZE_MAX		SSIZE_MAX
191 #define	DEVFS_IOSIZE_MAX	SSIZE_MAX
192 #endif
193 
194 /*
195  * General function declarations.
196  */
197 
198 struct inpcb;
199 struct lock_object;
200 struct malloc_type;
201 struct mtx;
202 struct proc;
203 struct socket;
204 struct thread;
205 struct tty;
206 struct ucred;
207 struct uio;
208 struct _jmp_buf;
209 struct trapframe;
210 struct eventtimer;
211 
212 int	setjmp(struct _jmp_buf *) __returns_twice;
213 void	longjmp(struct _jmp_buf *, int) __dead2;
214 int	dumpstatus(vm_offset_t addr, off_t count);
215 int	nullop(void);
216 int	eopnotsupp(void);
217 int	ureadc(int, struct uio *);
218 void	hashdestroy(void *, struct malloc_type *, u_long);
219 void	*hashinit(int count, struct malloc_type *type, u_long *hashmask);
220 void	*hashinit_flags(int count, struct malloc_type *type,
221     u_long *hashmask, int flags);
222 #define	HASH_NOWAIT	0x00000001
223 #define	HASH_WAITOK	0x00000002
224 
225 void	*phashinit(int count, struct malloc_type *type, u_long *nentries);
226 void	*phashinit_flags(int count, struct malloc_type *type, u_long *nentries,
227     int flags);
228 void	g_waitidle(void);
229 
230 void	cpu_boot(int);
231 void	cpu_flush_dcache(void *, size_t);
232 void	cpu_rootconf(void);
233 void	critical_enter_KBI(void);
234 void	critical_exit_KBI(void);
235 void	critical_exit_preempt(void);
236 void	init_param1(void);
237 void	init_param2(long physpages);
238 void	init_static_kenv(char *, size_t);
239 void	tablefull(const char *);
240 
241 /*
242  * Allocate per-thread "current" state in the linuxkpi
243  */
244 extern int (*lkpi_alloc_current)(struct thread *, int);
245 int linux_alloc_current_noop(struct thread *, int);
246 
247 
248 #if defined(KLD_MODULE) || defined(KTR_CRITICAL) || !defined(_KERNEL) || defined(GENOFFSET)
249 #define critical_enter() critical_enter_KBI()
250 #define critical_exit() critical_exit_KBI()
251 #else
252 static __inline void
253 critical_enter(void)
254 {
255 	struct thread_lite *td;
256 
257 	td = (struct thread_lite *)curthread;
258 	td->td_critnest++;
259 	__compiler_membar();
260 }
261 
262 static __inline void
263 critical_exit(void)
264 {
265 	struct thread_lite *td;
266 
267 	td = (struct thread_lite *)curthread;
268 	KASSERT(td->td_critnest != 0,
269 	    ("critical_exit: td_critnest == 0"));
270 	__compiler_membar();
271 	td->td_critnest--;
272 	__compiler_membar();
273 	if (__predict_false(td->td_owepreempt))
274 		critical_exit_preempt();
275 
276 }
277 #endif
278 
279 
280 #ifdef  EARLY_PRINTF
281 typedef void early_putc_t(int ch);
282 extern early_putc_t *early_putc;
283 #endif
284 int	kvprintf(char const *, void (*)(int, void*), void *, int,
285 	    __va_list) __printflike(1, 0);
286 void	log(int, const char *, ...) __printflike(2, 3);
287 void	log_console(struct uio *);
288 void	vlog(int, const char *, __va_list) __printflike(2, 0);
289 int	asprintf(char **ret, struct malloc_type *mtp, const char *format,
290 	    ...) __printflike(3, 4);
291 int	printf(const char *, ...) __printflike(1, 2);
292 int	snprintf(char *, size_t, const char *, ...) __printflike(3, 4);
293 int	sprintf(char *buf, const char *, ...) __printflike(2, 3);
294 int	uprintf(const char *, ...) __printflike(1, 2);
295 int	vprintf(const char *, __va_list) __printflike(1, 0);
296 int	vasprintf(char **ret, struct malloc_type *mtp, const char *format,
297 	    __va_list ap) __printflike(3, 0);
298 int	vsnprintf(char *, size_t, const char *, __va_list) __printflike(3, 0);
299 int	vsnrprintf(char *, size_t, int, const char *, __va_list) __printflike(4, 0);
300 int	vsprintf(char *buf, const char *, __va_list) __printflike(2, 0);
301 int	sscanf(const char *, char const * _Nonnull, ...) __scanflike(2, 3);
302 int	vsscanf(const char * _Nonnull, char const * _Nonnull, __va_list)  __scanflike(2, 0);
303 long	strtol(const char *, char **, int);
304 u_long	strtoul(const char *, char **, int);
305 quad_t	strtoq(const char *, char **, int);
306 u_quad_t strtouq(const char *, char **, int);
307 void	tprintf(struct proc *p, int pri, const char *, ...) __printflike(3, 4);
308 void	vtprintf(struct proc *, int, const char *, __va_list) __printflike(3, 0);
309 void	hexdump(const void *ptr, int length, const char *hdr, int flags);
310 #define	HD_COLUMN_MASK	0xff
311 #define	HD_DELIM_MASK	0xff00
312 #define	HD_OMIT_COUNT	(1 << 16)
313 #define	HD_OMIT_HEX	(1 << 17)
314 #define	HD_OMIT_CHARS	(1 << 18)
315 
316 #define ovbcopy(f, t, l) bcopy((f), (t), (l))
317 void	bcopy(const void * _Nonnull from, void * _Nonnull to, size_t len);
318 #define bcopy(from, to, len) __builtin_memmove((to), (from), (len))
319 void	bzero(void * _Nonnull buf, size_t len);
320 #define bzero(buf, len) __builtin_memset((buf), 0, (len))
321 void	explicit_bzero(void * _Nonnull, size_t);
322 int	bcmp(const void *b1, const void *b2, size_t len);
323 #define bcmp(b1, b2, len) __builtin_memcmp((b1), (b2), (len))
324 
325 void	*memset(void * _Nonnull buf, int c, size_t len);
326 #define memset(buf, c, len) __builtin_memset((buf), (c), (len))
327 void	*memcpy(void * _Nonnull to, const void * _Nonnull from, size_t len);
328 #define memcpy(to, from, len) __builtin_memcpy((to), (from), (len))
329 void	*memmove(void * _Nonnull dest, const void * _Nonnull src, size_t n);
330 #define memmove(dest, src, n) __builtin_memmove((dest), (src), (n))
331 int	memcmp(const void *b1, const void *b2, size_t len);
332 #define memcmp(b1, b2, len) __builtin_memcmp((b1), (b2), (len))
333 
334 void	*memset_early(void * _Nonnull buf, int c, size_t len);
335 #define bzero_early(buf, len) memset_early((buf), 0, (len))
336 void	*memcpy_early(void * _Nonnull to, const void * _Nonnull from, size_t len);
337 void	*memmove_early(void * _Nonnull dest, const void * _Nonnull src, size_t n);
338 #define bcopy_early(from, to, len) memmove_early((to), (from), (len))
339 
340 int	copystr(const void * _Nonnull __restrict kfaddr,
341 	    void * _Nonnull __restrict kdaddr, size_t len,
342 	    size_t * __restrict lencopied);
343 int	copyinstr(const void * __restrict udaddr,
344 	    void * _Nonnull __restrict kaddr, size_t len,
345 	    size_t * __restrict lencopied);
346 int	copyin(const void * __restrict udaddr,
347 	    void * _Nonnull __restrict kaddr, size_t len);
348 int	copyin_nofault(const void * __restrict udaddr,
349 	    void * _Nonnull __restrict kaddr, size_t len);
350 int	copyout(const void * _Nonnull __restrict kaddr,
351 	    void * __restrict udaddr, size_t len);
352 int	copyout_nofault(const void * _Nonnull __restrict kaddr,
353 	    void * __restrict udaddr, size_t len);
354 
355 int	fubyte(volatile const void *base);
356 long	fuword(volatile const void *base);
357 int	fuword16(volatile const void *base);
358 int32_t	fuword32(volatile const void *base);
359 int64_t	fuword64(volatile const void *base);
360 int	fueword(volatile const void *base, long *val);
361 int	fueword32(volatile const void *base, int32_t *val);
362 int	fueword64(volatile const void *base, int64_t *val);
363 int	subyte(volatile void *base, int byte);
364 int	suword(volatile void *base, long word);
365 int	suword16(volatile void *base, int word);
366 int	suword32(volatile void *base, int32_t word);
367 int	suword64(volatile void *base, int64_t word);
368 uint32_t casuword32(volatile uint32_t *base, uint32_t oldval, uint32_t newval);
369 u_long	casuword(volatile u_long *p, u_long oldval, u_long newval);
370 int	casueword32(volatile uint32_t *base, uint32_t oldval, uint32_t *oldvalp,
371 	    uint32_t newval);
372 int	casueword(volatile u_long *p, u_long oldval, u_long *oldvalp,
373 	    u_long newval);
374 
375 void	realitexpire(void *);
376 
377 int	sysbeep(int hertz, int period);
378 
379 void	hardclock(int cnt, int usermode);
380 void	hardclock_sync(int cpu);
381 void	softclock(void *);
382 void	statclock(int cnt, int usermode);
383 void	profclock(int cnt, int usermode, uintfptr_t pc);
384 
385 int	hardclockintr(void);
386 
387 void	startprofclock(struct proc *);
388 void	stopprofclock(struct proc *);
389 void	cpu_startprofclock(void);
390 void	cpu_stopprofclock(void);
391 void	suspendclock(void);
392 void	resumeclock(void);
393 sbintime_t 	cpu_idleclock(void);
394 void	cpu_activeclock(void);
395 void	cpu_new_callout(int cpu, sbintime_t bt, sbintime_t bt_opt);
396 void	cpu_et_frequency(struct eventtimer *et, uint64_t newfreq);
397 extern int	cpu_disable_c2_sleep;
398 extern int	cpu_disable_c3_sleep;
399 
400 char	*kern_getenv(const char *name);
401 void	freeenv(char *env);
402 int	getenv_int(const char *name, int *data);
403 int	getenv_uint(const char *name, unsigned int *data);
404 int	getenv_long(const char *name, long *data);
405 int	getenv_ulong(const char *name, unsigned long *data);
406 int	getenv_string(const char *name, char *data, int size);
407 int	getenv_int64(const char *name, int64_t *data);
408 int	getenv_uint64(const char *name, uint64_t *data);
409 int	getenv_quad(const char *name, quad_t *data);
410 int	kern_setenv(const char *name, const char *value);
411 int	kern_unsetenv(const char *name);
412 int	testenv(const char *name);
413 
414 int	getenv_array(const char *name, void *data, int size, int *psize,
415     int type_size, bool allow_signed);
416 #define	GETENV_UNSIGNED	false	/* negative numbers not allowed */
417 #define	GETENV_SIGNED	true	/* negative numbers allowed */
418 
419 typedef uint64_t (cpu_tick_f)(void);
420 void set_cputicker(cpu_tick_f *func, uint64_t freq, unsigned var);
421 extern cpu_tick_f *cpu_ticks;
422 uint64_t cpu_tickrate(void);
423 uint64_t cputick2usec(uint64_t tick);
424 
425 #ifdef APM_FIXUP_CALLTODO
426 struct timeval;
427 void	adjust_timeout_calltodo(struct timeval *time_change);
428 #endif /* APM_FIXUP_CALLTODO */
429 
430 #include <sys/libkern.h>
431 
432 /* Initialize the world */
433 void	consinit(void);
434 void	cpu_initclocks(void);
435 void	cpu_initclocks_bsp(void);
436 void	cpu_initclocks_ap(void);
437 void	usrinfoinit(void);
438 
439 /* Finalize the world */
440 void	kern_reboot(int) __dead2;
441 void	shutdown_nice(int);
442 
443 /* Timeouts */
444 typedef void timeout_t(void *);	/* timeout function type */
445 #define CALLOUT_HANDLE_INITIALIZER(handle)	\
446 	{ NULL }
447 
448 void	callout_handle_init(struct callout_handle *);
449 struct	callout_handle timeout(timeout_t *, void *, int);
450 void	untimeout(timeout_t *, void *, struct callout_handle);
451 
452 /* Stubs for obsolete functions that used to be for interrupt management */
453 static __inline intrmask_t	splbio(void)		{ return 0; }
454 static __inline intrmask_t	splcam(void)		{ return 0; }
455 static __inline intrmask_t	splclock(void)		{ return 0; }
456 static __inline intrmask_t	splhigh(void)		{ return 0; }
457 static __inline intrmask_t	splimp(void)		{ return 0; }
458 static __inline intrmask_t	splnet(void)		{ return 0; }
459 static __inline intrmask_t	spltty(void)		{ return 0; }
460 static __inline void		splx(intrmask_t ipl __unused)	{ return; }
461 
462 /*
463  * Common `proc' functions are declared here so that proc.h can be included
464  * less often.
465  */
466 int	_sleep(void * _Nonnull chan, struct lock_object *lock, int pri,
467 	   const char *wmesg, sbintime_t sbt, sbintime_t pr, int flags);
468 #define	msleep(chan, mtx, pri, wmesg, timo)				\
469 	_sleep((chan), &(mtx)->lock_object, (pri), (wmesg),		\
470 	    tick_sbt * (timo), 0, C_HARDCLOCK)
471 #define	msleep_sbt(chan, mtx, pri, wmesg, bt, pr, flags)		\
472 	_sleep((chan), &(mtx)->lock_object, (pri), (wmesg), (bt), (pr),	\
473 	    (flags))
474 int	msleep_spin_sbt(void * _Nonnull chan, struct mtx *mtx,
475 	    const char *wmesg, sbintime_t sbt, sbintime_t pr, int flags);
476 #define	msleep_spin(chan, mtx, wmesg, timo)				\
477 	msleep_spin_sbt((chan), (mtx), (wmesg), tick_sbt * (timo),	\
478 	    0, C_HARDCLOCK)
479 int	pause_sbt(const char *wmesg, sbintime_t sbt, sbintime_t pr,
480 	    int flags);
481 #define	pause(wmesg, timo)						\
482 	pause_sbt((wmesg), tick_sbt * (timo), 0, C_HARDCLOCK)
483 #define	pause_sig(wmesg, timo)						\
484 	pause_sbt((wmesg), tick_sbt * (timo), 0, C_HARDCLOCK | C_CATCH)
485 #define	tsleep(chan, pri, wmesg, timo)					\
486 	_sleep((chan), NULL, (pri), (wmesg), tick_sbt * (timo),		\
487 	    0, C_HARDCLOCK)
488 #define	tsleep_sbt(chan, pri, wmesg, bt, pr, flags)			\
489 	_sleep((chan), NULL, (pri), (wmesg), (bt), (pr), (flags))
490 void	wakeup(void * chan);
491 void	wakeup_one(void * chan);
492 void	wakeup_any(void * chan);
493 
494 /*
495  * Common `struct cdev *' stuff are declared here to avoid #include poisoning
496  */
497 
498 struct cdev;
499 dev_t dev2udev(struct cdev *x);
500 const char *devtoname(struct cdev *cdev);
501 
502 #ifdef __LP64__
503 size_t	devfs_iosize_max(void);
504 size_t	iosize_max(void);
505 #endif
506 
507 int poll_no_poll(int events);
508 
509 /* XXX: Should be void nanodelay(u_int nsec); */
510 void	DELAY(int usec);
511 
512 /* Root mount holdback API */
513 struct root_hold_token;
514 
515 struct root_hold_token *root_mount_hold(const char *identifier);
516 void root_mount_rel(struct root_hold_token *h);
517 int root_mounted(void);
518 
519 
520 /*
521  * Unit number allocation API. (kern/subr_unit.c)
522  */
523 struct unrhdr;
524 struct unrhdr *new_unrhdr(int low, int high, struct mtx *mutex);
525 void init_unrhdr(struct unrhdr *uh, int low, int high, struct mtx *mutex);
526 void delete_unrhdr(struct unrhdr *uh);
527 void clear_unrhdr(struct unrhdr *uh);
528 void clean_unrhdr(struct unrhdr *uh);
529 void clean_unrhdrl(struct unrhdr *uh);
530 int alloc_unr(struct unrhdr *uh);
531 int alloc_unr_specific(struct unrhdr *uh, u_int item);
532 int alloc_unrl(struct unrhdr *uh);
533 void free_unr(struct unrhdr *uh, u_int item);
534 
535 #ifndef __LP64__
536 #define UNR64_LOCKED
537 #endif
538 
539 struct unrhdr64 {
540         uint64_t	counter;
541 };
542 
543 static __inline void
544 new_unrhdr64(struct unrhdr64 *unr64, uint64_t low)
545 {
546 
547 	unr64->counter = low;
548 }
549 
550 #ifdef UNR64_LOCKED
551 uint64_t alloc_unr64(struct unrhdr64 *);
552 #else
553 static __inline uint64_t
554 alloc_unr64(struct unrhdr64 *unr64)
555 {
556 
557 	return (atomic_fetchadd_64(&unr64->counter, 1));
558 }
559 #endif
560 
561 void	intr_prof_stack_use(struct thread *td, struct trapframe *frame);
562 
563 void counted_warning(unsigned *counter, const char *msg);
564 
565 /*
566  * APIs to manage deprecation and obsolescence.
567  */
568 struct device;
569 void _gone_in(int major, const char *msg);
570 void _gone_in_dev(struct device *dev, int major, const char *msg);
571 #ifdef NO_OBSOLETE_CODE
572 #define __gone_ok(m, msg)					 \
573 	_Static_assert(m < P_OSREL_MAJOR(__FreeBSD_version)),	 \
574 	    "Obsolete code" msg);
575 #else
576 #define	__gone_ok(m, msg)
577 #endif
578 #define gone_in(major, msg)		__gone_ok(major, msg) _gone_in(major, msg)
579 #define gone_in_dev(dev, major, msg)	__gone_ok(major, msg) _gone_in_dev(dev, major, msg)
580 
581 __NULLABILITY_PRAGMA_POP
582 
583 #endif /* !_SYS_SYSTM_H_ */
584