xref: /dragonfly/sys/cpu/x86_64/include/cpufunc.h (revision ef54aa85)
1b2b3ffcdSSimon Schubert /*-
2b2b3ffcdSSimon Schubert  * Copyright (c) 2003 Peter Wemm.
3b2b3ffcdSSimon Schubert  * Copyright (c) 1993 The Regents of the University of California.
4b2b3ffcdSSimon Schubert  * Copyright (c) 2008 The DragonFly Project.
5b2b3ffcdSSimon Schubert  * All rights reserved.
6b2b3ffcdSSimon Schubert  *
7b2b3ffcdSSimon Schubert  * Redistribution and use in source and binary forms, with or without
8b2b3ffcdSSimon Schubert  * modification, are permitted provided that the following conditions
9b2b3ffcdSSimon Schubert  * are met:
10b2b3ffcdSSimon Schubert  * 1. Redistributions of source code must retain the above copyright
11b2b3ffcdSSimon Schubert  *    notice, this list of conditions and the following disclaimer.
12b2b3ffcdSSimon Schubert  * 2. Redistributions in binary form must reproduce the above copyright
13b2b3ffcdSSimon Schubert  *    notice, this list of conditions and the following disclaimer in the
14b2b3ffcdSSimon Schubert  *    documentation and/or other materials provided with the distribution.
152c64e990Szrj  * 3. Neither the name of the University nor the names of its contributors
16b2b3ffcdSSimon Schubert  *    may be used to endorse or promote products derived from this software
17b2b3ffcdSSimon Schubert  *    without specific prior written permission.
18b2b3ffcdSSimon Schubert  *
19b2b3ffcdSSimon Schubert  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
20b2b3ffcdSSimon Schubert  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21b2b3ffcdSSimon Schubert  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22b2b3ffcdSSimon Schubert  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
23b2b3ffcdSSimon Schubert  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
24b2b3ffcdSSimon Schubert  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
25b2b3ffcdSSimon Schubert  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
26b2b3ffcdSSimon Schubert  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
27b2b3ffcdSSimon Schubert  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
28b2b3ffcdSSimon Schubert  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29b2b3ffcdSSimon Schubert  * SUCH DAMAGE.
30b2b3ffcdSSimon Schubert  *
31b2b3ffcdSSimon Schubert  * $FreeBSD: src/sys/amd64/include/cpufunc.h,v 1.139 2004/01/28 23:53:04 peter Exp $
32b2b3ffcdSSimon Schubert  */
33b2b3ffcdSSimon Schubert 
34b2b3ffcdSSimon Schubert /*
35b2b3ffcdSSimon Schubert  * Functions to provide access to special i386 instructions.
36b2b3ffcdSSimon Schubert  * This in included in sys/systm.h, and that file should be
37b2b3ffcdSSimon Schubert  * used in preference to this.
38b2b3ffcdSSimon Schubert  */
39b2b3ffcdSSimon Schubert 
40b2b3ffcdSSimon Schubert #ifndef _CPU_CPUFUNC_H_
41b2b3ffcdSSimon Schubert #define	_CPU_CPUFUNC_H_
42b2b3ffcdSSimon Schubert 
43b2b3ffcdSSimon Schubert #include <sys/cdefs.h>
443c8aa76fSFrançois Tigeot #include <sys/thread.h>
455b49787bSMatthew Dillon #include <machine/clock.h>
46b2b3ffcdSSimon Schubert #include <machine/psl.h>
473c8aa76fSFrançois Tigeot #include <machine/smp.h>
48b2b3ffcdSSimon Schubert 
49b2b3ffcdSSimon Schubert struct thread;
50b2b3ffcdSSimon Schubert struct region_descriptor;
5179f2da03SMatthew Dillon struct pmap;
52b2b3ffcdSSimon Schubert 
53b2b3ffcdSSimon Schubert __BEGIN_DECLS
54b2b3ffcdSSimon Schubert #define readb(va)	(*(volatile u_int8_t *) (va))
55b2b3ffcdSSimon Schubert #define readw(va)	(*(volatile u_int16_t *) (va))
56b2b3ffcdSSimon Schubert #define readl(va)	(*(volatile u_int32_t *) (va))
57b2b3ffcdSSimon Schubert #define readq(va)	(*(volatile u_int64_t *) (va))
58b2b3ffcdSSimon Schubert 
59b2b3ffcdSSimon Schubert #define writeb(va, d)	(*(volatile u_int8_t *) (va) = (d))
60b2b3ffcdSSimon Schubert #define writew(va, d)	(*(volatile u_int16_t *) (va) = (d))
61b2b3ffcdSSimon Schubert #define writel(va, d)	(*(volatile u_int32_t *) (va) = (d))
62b2b3ffcdSSimon Schubert #define writeq(va, d)	(*(volatile u_int64_t *) (va) = (d))
63b2b3ffcdSSimon Schubert 
64b2b3ffcdSSimon Schubert #ifdef	__GNUC__
65b2b3ffcdSSimon Schubert 
66b2b3ffcdSSimon Schubert #include <machine/lock.h>		/* XXX */
67b2b3ffcdSSimon Schubert 
6891dc43ddSMatthew Dillon struct trapframe;
6991dc43ddSMatthew Dillon 
70b2b3ffcdSSimon Schubert static __inline void
breakpoint(void)71b2b3ffcdSSimon Schubert breakpoint(void)
72b2b3ffcdSSimon Schubert {
73b2b3ffcdSSimon Schubert 	__asm __volatile("int $3");
74b2b3ffcdSSimon Schubert }
75b2b3ffcdSSimon Schubert 
76b2b3ffcdSSimon Schubert static __inline void
cpu_pause(void)77b2b3ffcdSSimon Schubert cpu_pause(void)
78b2b3ffcdSSimon Schubert {
79106d187cSMatthew Dillon 	__asm __volatile("pause":::"memory");
80b2b3ffcdSSimon Schubert }
81b2b3ffcdSSimon Schubert 
82b2b3ffcdSSimon Schubert static __inline u_int
bsfl(u_int mask)83b2b3ffcdSSimon Schubert bsfl(u_int mask)
84b2b3ffcdSSimon Schubert {
85b2b3ffcdSSimon Schubert 	u_int	result;
86b2b3ffcdSSimon Schubert 
87b2b3ffcdSSimon Schubert 	__asm __volatile("bsfl %1,%0" : "=r" (result) : "rm" (mask));
88b2b3ffcdSSimon Schubert 	return (result);
89b2b3ffcdSSimon Schubert }
90b2b3ffcdSSimon Schubert 
91b2b3ffcdSSimon Schubert static __inline u_long
bsfq(u_long mask)92b2b3ffcdSSimon Schubert bsfq(u_long mask)
93b2b3ffcdSSimon Schubert {
94b2b3ffcdSSimon Schubert 	u_long	result;
95b2b3ffcdSSimon Schubert 
96b2b3ffcdSSimon Schubert 	__asm __volatile("bsfq %1,%0" : "=r" (result) : "rm" (mask));
97b2b3ffcdSSimon Schubert 	return (result);
98b2b3ffcdSSimon Schubert }
99b2b3ffcdSSimon Schubert 
1009c20f2eaSMatthew Dillon static __inline u_long
bsflong(u_long mask)1019c20f2eaSMatthew Dillon bsflong(u_long mask)
1029c20f2eaSMatthew Dillon {
1039c20f2eaSMatthew Dillon 	u_long	result;
1049c20f2eaSMatthew Dillon 
1059c20f2eaSMatthew Dillon 	__asm __volatile("bsfq %1,%0" : "=r" (result) : "rm" (mask));
1069c20f2eaSMatthew Dillon 	return (result);
1079c20f2eaSMatthew Dillon }
1089c20f2eaSMatthew Dillon 
109b2b3ffcdSSimon Schubert static __inline u_int
bsrl(u_int mask)110b2b3ffcdSSimon Schubert bsrl(u_int mask)
111b2b3ffcdSSimon Schubert {
112b2b3ffcdSSimon Schubert 	u_int	result;
113b2b3ffcdSSimon Schubert 
114b2b3ffcdSSimon Schubert 	__asm __volatile("bsrl %1,%0" : "=r" (result) : "rm" (mask));
115b2b3ffcdSSimon Schubert 	return (result);
116b2b3ffcdSSimon Schubert }
117b2b3ffcdSSimon Schubert 
118b2b3ffcdSSimon Schubert static __inline u_long
bsrq(u_long mask)119b2b3ffcdSSimon Schubert bsrq(u_long mask)
120b2b3ffcdSSimon Schubert {
121b2b3ffcdSSimon Schubert 	u_long	result;
122b2b3ffcdSSimon Schubert 
123b2b3ffcdSSimon Schubert 	__asm __volatile("bsrq %1,%0" : "=r" (result) : "rm" (mask));
124b2b3ffcdSSimon Schubert 	return (result);
125b2b3ffcdSSimon Schubert }
126b2b3ffcdSSimon Schubert 
127b2b3ffcdSSimon Schubert static __inline void
clflush(u_long addr)12808771751SVenkatesh Srinivas clflush(u_long addr)
12908771751SVenkatesh Srinivas {
13008771751SVenkatesh Srinivas 	__asm __volatile("clflush %0" : : "m" (*(char *) addr));
13108771751SVenkatesh Srinivas }
13208771751SVenkatesh Srinivas 
13308771751SVenkatesh Srinivas static __inline void
do_cpuid(u_int ax,u_int * p)134b2b3ffcdSSimon Schubert do_cpuid(u_int ax, u_int *p)
135b2b3ffcdSSimon Schubert {
136b2b3ffcdSSimon Schubert 	__asm __volatile("cpuid"
137b2b3ffcdSSimon Schubert 			 : "=a" (p[0]), "=b" (p[1]), "=c" (p[2]), "=d" (p[3])
138b2b3ffcdSSimon Schubert 			 :  "0" (ax));
139b2b3ffcdSSimon Schubert }
140b2b3ffcdSSimon Schubert 
141b2b3ffcdSSimon Schubert static __inline void
cpuid_count(u_int ax,u_int cx,u_int * p)142b2b3ffcdSSimon Schubert cpuid_count(u_int ax, u_int cx, u_int *p)
143b2b3ffcdSSimon Schubert {
144b2b3ffcdSSimon Schubert 	__asm __volatile("cpuid"
145b2b3ffcdSSimon Schubert 			 : "=a" (p[0]), "=b" (p[1]), "=c" (p[2]), "=d" (p[3])
146b2b3ffcdSSimon Schubert 			 :  "0" (ax), "c" (cx));
147b2b3ffcdSSimon Schubert }
148b2b3ffcdSSimon Schubert 
149a293ac08SJordan Gordeev #ifndef _CPU_DISABLE_INTR_DEFINED
150a293ac08SJordan Gordeev 
151a293ac08SJordan Gordeev static __inline void
cpu_disable_intr(void)152a293ac08SJordan Gordeev cpu_disable_intr(void)
153a293ac08SJordan Gordeev {
154a293ac08SJordan Gordeev 	__asm __volatile("cli" : : : "memory");
155a293ac08SJordan Gordeev }
156a293ac08SJordan Gordeev 
157a293ac08SJordan Gordeev #endif
158a293ac08SJordan Gordeev 
159a293ac08SJordan Gordeev #ifndef _CPU_ENABLE_INTR_DEFINED
160a293ac08SJordan Gordeev 
161b2b3ffcdSSimon Schubert static __inline void
cpu_enable_intr(void)162b2b3ffcdSSimon Schubert cpu_enable_intr(void)
163b2b3ffcdSSimon Schubert {
164b2b3ffcdSSimon Schubert 	__asm __volatile("sti");
165b2b3ffcdSSimon Schubert }
166b2b3ffcdSSimon Schubert 
167a293ac08SJordan Gordeev #endif
168a293ac08SJordan Gordeev 
169b2b3ffcdSSimon Schubert /*
170b2b3ffcdSSimon Schubert  * Cpu and compiler memory ordering fence.  mfence ensures strong read and
171b2b3ffcdSSimon Schubert  * write ordering.
172b2b3ffcdSSimon Schubert  *
173b2b3ffcdSSimon Schubert  * A serializing or fence instruction is required here.  A locked bus
174b2b3ffcdSSimon Schubert  * cycle on data for which we already own cache mastership is the most
175b2b3ffcdSSimon Schubert  * portable.
176b2b3ffcdSSimon Schubert  */
177b2b3ffcdSSimon Schubert static __inline void
cpu_mfence(void)178b2b3ffcdSSimon Schubert cpu_mfence(void)
179b2b3ffcdSSimon Schubert {
180b2b3ffcdSSimon Schubert 	__asm __volatile("mfence" : : : "memory");
181b2b3ffcdSSimon Schubert }
182b2b3ffcdSSimon Schubert 
183b2b3ffcdSSimon Schubert /*
184b2b3ffcdSSimon Schubert  * cpu_lfence() ensures strong read ordering for reads issued prior
185b2b3ffcdSSimon Schubert  * to the instruction verses reads issued afterwords.
186b2b3ffcdSSimon Schubert  *
187b2b3ffcdSSimon Schubert  * A serializing or fence instruction is required here.  A locked bus
188b2b3ffcdSSimon Schubert  * cycle on data for which we already own cache mastership is the most
189b2b3ffcdSSimon Schubert  * portable.
190b2b3ffcdSSimon Schubert  */
191b2b3ffcdSSimon Schubert static __inline void
cpu_lfence(void)192b2b3ffcdSSimon Schubert cpu_lfence(void)
193b2b3ffcdSSimon Schubert {
194b2b3ffcdSSimon Schubert 	__asm __volatile("lfence" : : : "memory");
195b2b3ffcdSSimon Schubert }
196b2b3ffcdSSimon Schubert 
197b2b3ffcdSSimon Schubert /*
198b2b3ffcdSSimon Schubert  * cpu_sfence() ensures strong write ordering for writes issued prior
199b2b3ffcdSSimon Schubert  * to the instruction verses writes issued afterwords.  Writes are
200b2b3ffcdSSimon Schubert  * ordered on intel cpus so we do not actually have to do anything.
201b2b3ffcdSSimon Schubert  */
202b2b3ffcdSSimon Schubert static __inline void
cpu_sfence(void)203b2b3ffcdSSimon Schubert cpu_sfence(void)
204b2b3ffcdSSimon Schubert {
20539bedebcSSepherosa Ziehau 	/*
20639bedebcSSepherosa Ziehau 	 * NOTE:
20739bedebcSSepherosa Ziehau 	 * Don't use 'sfence' here, as it will create a lot of
20839bedebcSSepherosa Ziehau 	 * unnecessary stalls.
20939bedebcSSepherosa Ziehau 	 */
210b2b3ffcdSSimon Schubert 	__asm __volatile("" : : : "memory");
211b2b3ffcdSSimon Schubert }
212b2b3ffcdSSimon Schubert 
213b2b3ffcdSSimon Schubert /*
214b2b3ffcdSSimon Schubert  * cpu_ccfence() prevents the compiler from reordering instructions, in
215b2b3ffcdSSimon Schubert  * particular stores, relative to the current cpu.  Use cpu_sfence() if
216b2b3ffcdSSimon Schubert  * you need to guarentee ordering by both the compiler and by the cpu.
217b2b3ffcdSSimon Schubert  *
218b2b3ffcdSSimon Schubert  * This also prevents the compiler from caching memory loads into local
219b2b3ffcdSSimon Schubert  * variables across the routine.
220b2b3ffcdSSimon Schubert  */
221b2b3ffcdSSimon Schubert static __inline void
cpu_ccfence(void)222b2b3ffcdSSimon Schubert cpu_ccfence(void)
223b2b3ffcdSSimon Schubert {
224b2b3ffcdSSimon Schubert 	__asm __volatile("" : : : "memory");
225b2b3ffcdSSimon Schubert }
226b2b3ffcdSSimon Schubert 
2278e32ecc0SMatthew Dillon /*
2288e32ecc0SMatthew Dillon  * This is a horrible, horrible hack that might have to be put at the
2298e32ecc0SMatthew Dillon  * end of certain procedures (on a case by case basis), just before it
2308e32ecc0SMatthew Dillon  * returns to avoid what we believe to be an unreported AMD cpu bug.
2318e32ecc0SMatthew Dillon  * Found to occur on both a Phenom II X4 820 (two of them), as well
2328e32ecc0SMatthew Dillon  * as a 48-core built around an Opteron 6168 (Id = 0x100f91  Stepping = 1).
2338e32ecc0SMatthew Dillon  * The problem does not appear to occur w/Intel cpus.
2348e32ecc0SMatthew Dillon  *
2358e32ecc0SMatthew Dillon  * The bug is likely related to either a write combining issue or the
2368e32ecc0SMatthew Dillon  * Return Address Stack (RAS) hardware cache.
2378e32ecc0SMatthew Dillon  *
2388e32ecc0SMatthew Dillon  * In particular, we had to do this for GCC's fill_sons_in_loop() routine
2398e32ecc0SMatthew Dillon  * which due to its deep recursion and stack flow appears to be able to
2408e32ecc0SMatthew Dillon  * tickle the amd cpu bug (w/ gcc-4.4.7).  Adding a single 'nop' to the
2418e32ecc0SMatthew Dillon  * end of the routine just before it returns works around the bug.
2428e32ecc0SMatthew Dillon  *
2438e32ecc0SMatthew Dillon  * The bug appears to be extremely sensitive to %rip and %rsp values, to
2448e32ecc0SMatthew Dillon  * the point where even just inserting an instruction in an unrelated
2458e32ecc0SMatthew Dillon  * procedure (shifting the entire code base being run) effects the outcome.
2468e32ecc0SMatthew Dillon  * DragonFly is probably able to more readily reproduce the bug due to
2478e32ecc0SMatthew Dillon  * the stackgap randomization code.  We would expect OpenBSD (where we got
2488e32ecc0SMatthew Dillon  * the stackgap randomization code from) to also be able to reproduce the
2498e32ecc0SMatthew Dillon  * issue.  To date we have only reproduced the issue in DragonFly.
2508e32ecc0SMatthew Dillon  */
2518e32ecc0SMatthew Dillon #define __AMDCPUBUG_DFLY01_AVAILABLE__
2528e32ecc0SMatthew Dillon 
2538e32ecc0SMatthew Dillon static __inline void
cpu_amdcpubug_dfly01(void)2548e32ecc0SMatthew Dillon cpu_amdcpubug_dfly01(void)
2558e32ecc0SMatthew Dillon {
2568e32ecc0SMatthew Dillon 	__asm __volatile("nop" : : : "memory");
2578e32ecc0SMatthew Dillon }
2588e32ecc0SMatthew Dillon 
259b2b3ffcdSSimon Schubert #ifdef _KERNEL
260b2b3ffcdSSimon Schubert 
261b2b3ffcdSSimon Schubert #define	HAVE_INLINE_FFS
262b2b3ffcdSSimon Schubert 
263b2b3ffcdSSimon Schubert static __inline int
ffs(int mask)264b2b3ffcdSSimon Schubert ffs(int mask)
265b2b3ffcdSSimon Schubert {
266b2b3ffcdSSimon Schubert 	return (__builtin_ffs(mask));
267b2b3ffcdSSimon Schubert }
268b2b3ffcdSSimon Schubert 
269b2b3ffcdSSimon Schubert #define	HAVE_INLINE_FFSL
270b2b3ffcdSSimon Schubert 
271b2b3ffcdSSimon Schubert static __inline int
ffsl(long mask)272b2b3ffcdSSimon Schubert ffsl(long mask)
273b2b3ffcdSSimon Schubert {
2743dc50e18SMatthew Dillon 	return (__builtin_ffsl(mask));
275b2b3ffcdSSimon Schubert }
276b2b3ffcdSSimon Schubert 
277b2b3ffcdSSimon Schubert #define	HAVE_INLINE_FLS
278b2b3ffcdSSimon Schubert 
279b2b3ffcdSSimon Schubert static __inline int
fls(int mask)280b2b3ffcdSSimon Schubert fls(int mask)
281b2b3ffcdSSimon Schubert {
282b2b3ffcdSSimon Schubert 	return (mask == 0 ? mask : (int)bsrl((u_int)mask) + 1);
283b2b3ffcdSSimon Schubert }
284b2b3ffcdSSimon Schubert 
285b2b3ffcdSSimon Schubert #define	HAVE_INLINE_FLSL
286b2b3ffcdSSimon Schubert 
287b2b3ffcdSSimon Schubert static __inline int
flsl(long mask)288b2b3ffcdSSimon Schubert flsl(long mask)
289b2b3ffcdSSimon Schubert {
290b2b3ffcdSSimon Schubert 	return (mask == 0 ? mask : (int)bsrq((u_long)mask) + 1);
291b2b3ffcdSSimon Schubert }
292b2b3ffcdSSimon Schubert 
2934e70a2e1SSascha Wildner #define	HAVE_INLINE_FLSLL
2944e70a2e1SSascha Wildner 
2954e70a2e1SSascha Wildner static __inline int
flsll(long long mask)2964e70a2e1SSascha Wildner flsll(long long mask)
2974e70a2e1SSascha Wildner {
2984e70a2e1SSascha Wildner 	return (flsl((long)mask));
2994e70a2e1SSascha Wildner }
3004e70a2e1SSascha Wildner 
301b2b3ffcdSSimon Schubert #endif /* _KERNEL */
302b2b3ffcdSSimon Schubert 
303b2b3ffcdSSimon Schubert static __inline void
halt(void)304b2b3ffcdSSimon Schubert halt(void)
305b2b3ffcdSSimon Schubert {
306b2b3ffcdSSimon Schubert 	__asm __volatile("hlt");
307b2b3ffcdSSimon Schubert }
308b2b3ffcdSSimon Schubert 
309b2b3ffcdSSimon Schubert /*
310b2b3ffcdSSimon Schubert  * The following complications are to get around gcc not having a
311b2b3ffcdSSimon Schubert  * constraint letter for the range 0..255.  We still put "d" in the
312b2b3ffcdSSimon Schubert  * constraint because "i" isn't a valid constraint when the port
313b2b3ffcdSSimon Schubert  * isn't constant.  This only matters for -O0 because otherwise
314b2b3ffcdSSimon Schubert  * the non-working version gets optimized away.
315b2b3ffcdSSimon Schubert  *
316b2b3ffcdSSimon Schubert  * Use an expression-statement instead of a conditional expression
317b2b3ffcdSSimon Schubert  * because gcc-2.6.0 would promote the operands of the conditional
318b2b3ffcdSSimon Schubert  * and produce poor code for "if ((inb(var) & const1) == const2)".
319b2b3ffcdSSimon Schubert  *
320b2b3ffcdSSimon Schubert  * The unnecessary test `(port) < 0x10000' is to generate a warning if
321b2b3ffcdSSimon Schubert  * the `port' has type u_short or smaller.  Such types are pessimal.
322b2b3ffcdSSimon Schubert  * This actually only works for signed types.  The range check is
323b2b3ffcdSSimon Schubert  * careful to avoid generating warnings.
324b2b3ffcdSSimon Schubert  */
325b2b3ffcdSSimon Schubert #define	inb(port) __extension__ ({					\
326b2b3ffcdSSimon Schubert 	u_char	_data;							\
327b2b3ffcdSSimon Schubert 	if (__builtin_constant_p(port) && ((port) & 0xffff) < 0x100	\
328b2b3ffcdSSimon Schubert 	    && (port) < 0x10000)					\
329b2b3ffcdSSimon Schubert 		_data = inbc(port);					\
330b2b3ffcdSSimon Schubert 	else								\
331b2b3ffcdSSimon Schubert 		_data = inbv(port);					\
332b2b3ffcdSSimon Schubert 	_data; })
333b2b3ffcdSSimon Schubert 
334b2b3ffcdSSimon Schubert #define	outb(port, data) (						\
335b2b3ffcdSSimon Schubert 	__builtin_constant_p(port) && ((port) & 0xffff) < 0x100		\
336b2b3ffcdSSimon Schubert 	&& (port) < 0x10000						\
337b2b3ffcdSSimon Schubert 	? outbc(port, data) : outbv(port, data))
338b2b3ffcdSSimon Schubert 
339b2b3ffcdSSimon Schubert static __inline u_char
inbc(u_int port)340b2b3ffcdSSimon Schubert inbc(u_int port)
341b2b3ffcdSSimon Schubert {
342b2b3ffcdSSimon Schubert 	u_char	data;
343b2b3ffcdSSimon Schubert 
344b2b3ffcdSSimon Schubert 	__asm __volatile("inb %1,%0" : "=a" (data) : "id" ((u_short)(port)));
345b2b3ffcdSSimon Schubert 	return (data);
346b2b3ffcdSSimon Schubert }
347b2b3ffcdSSimon Schubert 
348b2b3ffcdSSimon Schubert static __inline void
outbc(u_int port,u_char data)349b2b3ffcdSSimon Schubert outbc(u_int port, u_char data)
350b2b3ffcdSSimon Schubert {
351b2b3ffcdSSimon Schubert 	__asm __volatile("outb %0,%1" : : "a" (data), "id" ((u_short)(port)));
352b2b3ffcdSSimon Schubert }
353b2b3ffcdSSimon Schubert 
354b2b3ffcdSSimon Schubert static __inline u_char
inbv(u_int port)355b2b3ffcdSSimon Schubert inbv(u_int port)
356b2b3ffcdSSimon Schubert {
357b2b3ffcdSSimon Schubert 	u_char	data;
358b2b3ffcdSSimon Schubert 	/*
359b2b3ffcdSSimon Schubert 	 * We use %%dx and not %1 here because i/o is done at %dx and not at
360b2b3ffcdSSimon Schubert 	 * %edx, while gcc generates inferior code (movw instead of movl)
361b2b3ffcdSSimon Schubert 	 * if we tell it to load (u_short) port.
362b2b3ffcdSSimon Schubert 	 */
363b2b3ffcdSSimon Schubert 	__asm __volatile("inb %%dx,%0" : "=a" (data) : "d" (port));
364b2b3ffcdSSimon Schubert 	return (data);
365b2b3ffcdSSimon Schubert }
366b2b3ffcdSSimon Schubert 
367b2b3ffcdSSimon Schubert static __inline u_int
inl(u_int port)368b2b3ffcdSSimon Schubert inl(u_int port)
369b2b3ffcdSSimon Schubert {
370b2b3ffcdSSimon Schubert 	u_int	data;
371b2b3ffcdSSimon Schubert 
372b2b3ffcdSSimon Schubert 	__asm __volatile("inl %%dx,%0" : "=a" (data) : "d" (port));
373b2b3ffcdSSimon Schubert 	return (data);
374b2b3ffcdSSimon Schubert }
375b2b3ffcdSSimon Schubert 
376b2b3ffcdSSimon Schubert static __inline void
insb(u_int port,void * addr,size_t cnt)377b2b3ffcdSSimon Schubert insb(u_int port, void *addr, size_t cnt)
378b2b3ffcdSSimon Schubert {
379b2b3ffcdSSimon Schubert 	__asm __volatile("cld; rep; insb"
380b2b3ffcdSSimon Schubert 			 : "+D" (addr), "+c" (cnt)
381b2b3ffcdSSimon Schubert 			 : "d" (port)
382b2b3ffcdSSimon Schubert 			 : "memory");
383b2b3ffcdSSimon Schubert }
384b2b3ffcdSSimon Schubert 
385b2b3ffcdSSimon Schubert static __inline void
insw(u_int port,void * addr,size_t cnt)386b2b3ffcdSSimon Schubert insw(u_int port, void *addr, size_t cnt)
387b2b3ffcdSSimon Schubert {
388b2b3ffcdSSimon Schubert 	__asm __volatile("cld; rep; insw"
389b2b3ffcdSSimon Schubert 			 : "+D" (addr), "+c" (cnt)
390b2b3ffcdSSimon Schubert 			 : "d" (port)
391b2b3ffcdSSimon Schubert 			 : "memory");
392b2b3ffcdSSimon Schubert }
393b2b3ffcdSSimon Schubert 
394b2b3ffcdSSimon Schubert static __inline void
insl(u_int port,void * addr,size_t cnt)395b2b3ffcdSSimon Schubert insl(u_int port, void *addr, size_t cnt)
396b2b3ffcdSSimon Schubert {
397b2b3ffcdSSimon Schubert 	__asm __volatile("cld; rep; insl"
398b2b3ffcdSSimon Schubert 			 : "+D" (addr), "+c" (cnt)
399b2b3ffcdSSimon Schubert 			 : "d" (port)
400b2b3ffcdSSimon Schubert 			 : "memory");
401b2b3ffcdSSimon Schubert }
402b2b3ffcdSSimon Schubert 
403b2b3ffcdSSimon Schubert static __inline void
invd(void)404b2b3ffcdSSimon Schubert invd(void)
405b2b3ffcdSSimon Schubert {
406b2b3ffcdSSimon Schubert 	__asm __volatile("invd");
407b2b3ffcdSSimon Schubert }
408b2b3ffcdSSimon Schubert 
409b2b3ffcdSSimon Schubert #if defined(_KERNEL)
410b2b3ffcdSSimon Schubert 
411b2b3ffcdSSimon Schubert #ifndef _CPU_INVLPG_DEFINED
412b2b3ffcdSSimon Schubert 
413b2b3ffcdSSimon Schubert /*
4143486bb88SFrançois Tigeot  * Invalidate a particular VA on this cpu only
4153486bb88SFrançois Tigeot  *
4163486bb88SFrançois Tigeot  * TLB flush for an individual page (even if it has PG_G).
4173486bb88SFrançois Tigeot  * Only works on 486+ CPUs (i386 does not have PG_G).
418b2b3ffcdSSimon Schubert  */
419b2b3ffcdSSimon Schubert static __inline void
cpu_invlpg(void * addr)420b2b3ffcdSSimon Schubert cpu_invlpg(void *addr)
421b2b3ffcdSSimon Schubert {
422b2b3ffcdSSimon Schubert 	__asm __volatile("invlpg %0" : : "m" (*(char *)addr) : "memory");
423b2b3ffcdSSimon Schubert }
424b2b3ffcdSSimon Schubert 
425b2b3ffcdSSimon Schubert #endif
426b2b3ffcdSSimon Schubert 
427b2b3ffcdSSimon Schubert static __inline void
cpu_nop(void)428b2b3ffcdSSimon Schubert cpu_nop(void)
429b2b3ffcdSSimon Schubert {
430b2b3ffcdSSimon Schubert 	__asm __volatile("rep; nop");
431b2b3ffcdSSimon Schubert }
432b2b3ffcdSSimon Schubert 
433b2b3ffcdSSimon Schubert #endif	/* _KERNEL */
434b2b3ffcdSSimon Schubert 
435b2b3ffcdSSimon Schubert static __inline u_short
inw(u_int port)436b2b3ffcdSSimon Schubert inw(u_int port)
437b2b3ffcdSSimon Schubert {
438b2b3ffcdSSimon Schubert 	u_short	data;
439b2b3ffcdSSimon Schubert 
440b2b3ffcdSSimon Schubert 	__asm __volatile("inw %%dx,%0" : "=a" (data) : "d" (port));
441b2b3ffcdSSimon Schubert 	return (data);
442b2b3ffcdSSimon Schubert }
443b2b3ffcdSSimon Schubert 
444b2b3ffcdSSimon Schubert static __inline u_int
loadandclear(volatile u_int * addr)445b2b3ffcdSSimon Schubert loadandclear(volatile u_int *addr)
446b2b3ffcdSSimon Schubert {
447b2b3ffcdSSimon Schubert 	u_int   result;
448b2b3ffcdSSimon Schubert 
449b2b3ffcdSSimon Schubert 	__asm __volatile("xorl %0,%0; xchgl %1,%0"
450b2b3ffcdSSimon Schubert 			: "=&r" (result) : "m" (*addr));
451b2b3ffcdSSimon Schubert 	return (result);
452b2b3ffcdSSimon Schubert }
453b2b3ffcdSSimon Schubert 
454b2b3ffcdSSimon Schubert static __inline void
outbv(u_int port,u_char data)455b2b3ffcdSSimon Schubert outbv(u_int port, u_char data)
456b2b3ffcdSSimon Schubert {
457b2b3ffcdSSimon Schubert 	u_char	al;
458b2b3ffcdSSimon Schubert 	/*
459b2b3ffcdSSimon Schubert 	 * Use an unnecessary assignment to help gcc's register allocator.
460b2b3ffcdSSimon Schubert 	 * This make a large difference for gcc-1.40 and a tiny difference
461b2b3ffcdSSimon Schubert 	 * for gcc-2.6.0.  For gcc-1.40, al had to be ``asm("ax")'' for
462b2b3ffcdSSimon Schubert 	 * best results.  gcc-2.6.0 can't handle this.
463b2b3ffcdSSimon Schubert 	 */
464b2b3ffcdSSimon Schubert 	al = data;
465b2b3ffcdSSimon Schubert 	__asm __volatile("outb %0,%%dx" : : "a" (al), "d" (port));
466b2b3ffcdSSimon Schubert }
467b2b3ffcdSSimon Schubert 
468b2b3ffcdSSimon Schubert static __inline void
outl(u_int port,u_int data)469b2b3ffcdSSimon Schubert outl(u_int port, u_int data)
470b2b3ffcdSSimon Schubert {
471b2b3ffcdSSimon Schubert 	/*
472b2b3ffcdSSimon Schubert 	 * outl() and outw() aren't used much so we haven't looked at
473b2b3ffcdSSimon Schubert 	 * possible micro-optimizations such as the unnecessary
474b2b3ffcdSSimon Schubert 	 * assignment for them.
475b2b3ffcdSSimon Schubert 	 */
476b2b3ffcdSSimon Schubert 	__asm __volatile("outl %0,%%dx" : : "a" (data), "d" (port));
477b2b3ffcdSSimon Schubert }
478b2b3ffcdSSimon Schubert 
479b2b3ffcdSSimon Schubert static __inline void
outsb(u_int port,const void * addr,size_t cnt)480b2b3ffcdSSimon Schubert outsb(u_int port, const void *addr, size_t cnt)
481b2b3ffcdSSimon Schubert {
482b2b3ffcdSSimon Schubert 	__asm __volatile("cld; rep; outsb"
483b2b3ffcdSSimon Schubert 			 : "+S" (addr), "+c" (cnt)
484b2b3ffcdSSimon Schubert 			 : "d" (port));
485b2b3ffcdSSimon Schubert }
486b2b3ffcdSSimon Schubert 
487b2b3ffcdSSimon Schubert static __inline void
outsw(u_int port,const void * addr,size_t cnt)488b2b3ffcdSSimon Schubert outsw(u_int port, const void *addr, size_t cnt)
489b2b3ffcdSSimon Schubert {
490b2b3ffcdSSimon Schubert 	__asm __volatile("cld; rep; outsw"
491b2b3ffcdSSimon Schubert 			 : "+S" (addr), "+c" (cnt)
492b2b3ffcdSSimon Schubert 			 : "d" (port));
493b2b3ffcdSSimon Schubert }
494b2b3ffcdSSimon Schubert 
495b2b3ffcdSSimon Schubert static __inline void
outsl(u_int port,const void * addr,size_t cnt)496b2b3ffcdSSimon Schubert outsl(u_int port, const void *addr, size_t cnt)
497b2b3ffcdSSimon Schubert {
498b2b3ffcdSSimon Schubert 	__asm __volatile("cld; rep; outsl"
499b2b3ffcdSSimon Schubert 			 : "+S" (addr), "+c" (cnt)
500b2b3ffcdSSimon Schubert 			 : "d" (port));
501b2b3ffcdSSimon Schubert }
502b2b3ffcdSSimon Schubert 
503b2b3ffcdSSimon Schubert static __inline void
outw(u_int port,u_short data)504b2b3ffcdSSimon Schubert outw(u_int port, u_short data)
505b2b3ffcdSSimon Schubert {
506b2b3ffcdSSimon Schubert 	__asm __volatile("outw %0,%%dx" : : "a" (data), "d" (port));
507b2b3ffcdSSimon Schubert }
508b2b3ffcdSSimon Schubert 
509b2b3ffcdSSimon Schubert static __inline void
ia32_pause(void)510b2b3ffcdSSimon Schubert ia32_pause(void)
511b2b3ffcdSSimon Schubert {
512b2b3ffcdSSimon Schubert 	__asm __volatile("pause");
513b2b3ffcdSSimon Schubert }
514b2b3ffcdSSimon Schubert 
515b2b3ffcdSSimon Schubert static __inline u_long
read_rflags(void)516b2b3ffcdSSimon Schubert read_rflags(void)
517b2b3ffcdSSimon Schubert {
518b2b3ffcdSSimon Schubert 	u_long	rf;
519b2b3ffcdSSimon Schubert 
520b2b3ffcdSSimon Schubert 	__asm __volatile("pushfq; popq %0" : "=r" (rf));
521b2b3ffcdSSimon Schubert 	return (rf);
522b2b3ffcdSSimon Schubert }
523b2b3ffcdSSimon Schubert 
524b2b3ffcdSSimon Schubert static __inline u_int64_t
rdmsr(u_int msr)525b2b3ffcdSSimon Schubert rdmsr(u_int msr)
526b2b3ffcdSSimon Schubert {
527b2b3ffcdSSimon Schubert 	u_int32_t low, high;
528b2b3ffcdSSimon Schubert 
529b2b3ffcdSSimon Schubert 	__asm __volatile("rdmsr" : "=a" (low), "=d" (high) : "c" (msr));
530b2b3ffcdSSimon Schubert 	return (low | ((u_int64_t)high << 32));
531b2b3ffcdSSimon Schubert }
532b2b3ffcdSSimon Schubert 
533b2b3ffcdSSimon Schubert static __inline u_int64_t
rdpmc(u_int pmc)534b2b3ffcdSSimon Schubert rdpmc(u_int pmc)
535b2b3ffcdSSimon Schubert {
536b2b3ffcdSSimon Schubert 	u_int32_t low, high;
537b2b3ffcdSSimon Schubert 
538b2b3ffcdSSimon Schubert 	__asm __volatile("rdpmc" : "=a" (low), "=d" (high) : "c" (pmc));
539b2b3ffcdSSimon Schubert 	return (low | ((u_int64_t)high << 32));
540b2b3ffcdSSimon Schubert }
541b2b3ffcdSSimon Schubert 
542b2b3ffcdSSimon Schubert #define _RDTSC_SUPPORTED_
543b2b3ffcdSSimon Schubert 
5445b49787bSMatthew Dillon static __inline tsc_uclock_t
rdtsc(void)545b2b3ffcdSSimon Schubert rdtsc(void)
546b2b3ffcdSSimon Schubert {
547b2b3ffcdSSimon Schubert 	u_int32_t low, high;
548b2b3ffcdSSimon Schubert 
549b2b3ffcdSSimon Schubert 	__asm __volatile("rdtsc" : "=a" (low), "=d" (high));
5505b49787bSMatthew Dillon 	return (low | ((tsc_uclock_t)high << 32));
551b2b3ffcdSSimon Schubert }
552b2b3ffcdSSimon Schubert 
553ea9728caSSepherosa Ziehau #ifdef _KERNEL
554ea9728caSSepherosa Ziehau #include <machine/cputypes.h>
555ea9728caSSepherosa Ziehau #include <machine/md_var.h>
556ea9728caSSepherosa Ziehau 
5575b49787bSMatthew Dillon static __inline tsc_uclock_t
rdtsc_ordered(void)558ea9728caSSepherosa Ziehau rdtsc_ordered(void)
559ea9728caSSepherosa Ziehau {
560ea9728caSSepherosa Ziehau 	if (cpu_vendor_id == CPU_VENDOR_INTEL)
561ea9728caSSepherosa Ziehau 		cpu_lfence();
562ea9728caSSepherosa Ziehau 	else
563ea9728caSSepherosa Ziehau 		cpu_mfence();
564ea9728caSSepherosa Ziehau 	return rdtsc();
565ea9728caSSepherosa Ziehau }
566ea9728caSSepherosa Ziehau #endif
567ea9728caSSepherosa Ziehau 
568b2b3ffcdSSimon Schubert static __inline void
wbinvd(void)569b2b3ffcdSSimon Schubert wbinvd(void)
570b2b3ffcdSSimon Schubert {
571b2b3ffcdSSimon Schubert 	__asm __volatile("wbinvd");
572b2b3ffcdSSimon Schubert }
573b2b3ffcdSSimon Schubert 
5743c8aa76fSFrançois Tigeot #if defined(_KERNEL)
5753c8aa76fSFrançois Tigeot void cpu_wbinvd_on_all_cpus_callback(void *arg);
5763c8aa76fSFrançois Tigeot 
5773c8aa76fSFrançois Tigeot static __inline void
cpu_wbinvd_on_all_cpus(void)5783c8aa76fSFrançois Tigeot cpu_wbinvd_on_all_cpus(void)
5793c8aa76fSFrançois Tigeot {
5803c8aa76fSFrançois Tigeot 	lwkt_cpusync_simple(smp_active_mask, cpu_wbinvd_on_all_cpus_callback, NULL);
5813c8aa76fSFrançois Tigeot }
5823c8aa76fSFrançois Tigeot #endif
5833c8aa76fSFrançois Tigeot 
584b2b3ffcdSSimon Schubert static __inline void
write_rflags(u_long rf)585b2b3ffcdSSimon Schubert write_rflags(u_long rf)
586b2b3ffcdSSimon Schubert {
587b2b3ffcdSSimon Schubert 	__asm __volatile("pushq %0;  popfq" : : "r" (rf));
588b2b3ffcdSSimon Schubert }
589b2b3ffcdSSimon Schubert 
590b2b3ffcdSSimon Schubert static __inline void
wrmsr(u_int msr,u_int64_t newval)591b2b3ffcdSSimon Schubert wrmsr(u_int msr, u_int64_t newval)
592b2b3ffcdSSimon Schubert {
593b2b3ffcdSSimon Schubert 	u_int32_t low, high;
594b2b3ffcdSSimon Schubert 
595b2b3ffcdSSimon Schubert 	low = newval;
596b2b3ffcdSSimon Schubert 	high = newval >> 32;
5979ba9002bSMatthew Dillon 	__asm __volatile("wrmsr"
5989ba9002bSMatthew Dillon 	    :
5999ba9002bSMatthew Dillon 	    : "a" (low), "d" (high), "c" (msr)
6009ba9002bSMatthew Dillon 	    : "memory");
601b2b3ffcdSSimon Schubert }
602b2b3ffcdSSimon Schubert 
603b2b3ffcdSSimon Schubert static __inline void
load_xcr(u_int xcr,uint64_t newval)60423bf518bSAaron LI load_xcr(u_int xcr, uint64_t newval)
6055cf56a8dSAlex Hornung {
6067c656f7bSMatthew Dillon 	uint32_t low, high;
6077c656f7bSMatthew Dillon 
6087c656f7bSMatthew Dillon 	low = newval;
6097c656f7bSMatthew Dillon 	high = newval >> 32;
6107c656f7bSMatthew Dillon 
61123bf518bSAaron LI 	__asm __volatile("xsetbv"
6125cf56a8dSAlex Hornung 	    :
6137c656f7bSMatthew Dillon 	    : "a" (low), "d" (high), "c" (xcr)
6149ba9002bSMatthew Dillon 	    : "memory");
6155cf56a8dSAlex Hornung }
6165cf56a8dSAlex Hornung 
61723bf518bSAaron LI static __inline uint64_t
rxcr(u_int xcr)61823bf518bSAaron LI rxcr(u_int xcr)
61923bf518bSAaron LI {
62023bf518bSAaron LI 	uint32_t low, high;
62123bf518bSAaron LI 
62223bf518bSAaron LI 	__asm __volatile("xgetbv" : "=a" (low), "=d" (high) : "c" (xcr));
62323bf518bSAaron LI 	return (low | ((uint64_t)high << 32));
62423bf518bSAaron LI }
62523bf518bSAaron LI 
6265cf56a8dSAlex Hornung static __inline void
load_cr0(u_long data)627b2b3ffcdSSimon Schubert load_cr0(u_long data)
628b2b3ffcdSSimon Schubert {
6299ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%cr0" : : "r" (data) : "memory");
630b2b3ffcdSSimon Schubert }
631b2b3ffcdSSimon Schubert 
632b2b3ffcdSSimon Schubert static __inline u_long
rcr0(void)633b2b3ffcdSSimon Schubert rcr0(void)
634b2b3ffcdSSimon Schubert {
635b2b3ffcdSSimon Schubert 	u_long	data;
636b2b3ffcdSSimon Schubert 
637b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%cr0,%0" : "=r" (data));
638b2b3ffcdSSimon Schubert 	return (data);
639b2b3ffcdSSimon Schubert }
640b2b3ffcdSSimon Schubert 
641*251f5dcbSAaron LI static __inline void
load_cr2(u_long data)642*251f5dcbSAaron LI load_cr2(u_long data)
643*251f5dcbSAaron LI {
644*251f5dcbSAaron LI 	__asm __volatile("movq %0,%%cr2" : : "r" (data) : "memory");
645*251f5dcbSAaron LI }
646*251f5dcbSAaron LI 
647b2b3ffcdSSimon Schubert static __inline u_long
rcr2(void)648b2b3ffcdSSimon Schubert rcr2(void)
649b2b3ffcdSSimon Schubert {
650b2b3ffcdSSimon Schubert 	u_long	data;
651b2b3ffcdSSimon Schubert 
652b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%cr2,%0" : "=r" (data));
653b2b3ffcdSSimon Schubert 	return (data);
654b2b3ffcdSSimon Schubert }
655b2b3ffcdSSimon Schubert 
656b2b3ffcdSSimon Schubert static __inline void
load_cr3(u_long data)657b2b3ffcdSSimon Schubert load_cr3(u_long data)
658b2b3ffcdSSimon Schubert {
659b2b3ffcdSSimon Schubert 	__asm __volatile("movq %0,%%cr3" : : "r" (data) : "memory");
660b2b3ffcdSSimon Schubert }
661b2b3ffcdSSimon Schubert 
662b2b3ffcdSSimon Schubert static __inline u_long
rcr3(void)663b2b3ffcdSSimon Schubert rcr3(void)
664b2b3ffcdSSimon Schubert {
665b2b3ffcdSSimon Schubert 	u_long	data;
666b2b3ffcdSSimon Schubert 
667b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%cr3,%0" : "=r" (data));
668b2b3ffcdSSimon Schubert 	return (data);
669b2b3ffcdSSimon Schubert }
670b2b3ffcdSSimon Schubert 
671b2b3ffcdSSimon Schubert static __inline void
load_cr4(u_long data)672b2b3ffcdSSimon Schubert load_cr4(u_long data)
673b2b3ffcdSSimon Schubert {
6749ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%cr4" : : "r" (data) : "memory");
675b2b3ffcdSSimon Schubert }
676b2b3ffcdSSimon Schubert 
677b2b3ffcdSSimon Schubert static __inline u_long
rcr4(void)678b2b3ffcdSSimon Schubert rcr4(void)
679b2b3ffcdSSimon Schubert {
680b2b3ffcdSSimon Schubert 	u_long	data;
681b2b3ffcdSSimon Schubert 
682b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%cr4,%0" : "=r" (data));
683b2b3ffcdSSimon Schubert 	return (data);
684b2b3ffcdSSimon Schubert }
685b2b3ffcdSSimon Schubert 
686a293ac08SJordan Gordeev #ifndef _CPU_INVLTLB_DEFINED
687a293ac08SJordan Gordeev 
688b2b3ffcdSSimon Schubert /*
689a293ac08SJordan Gordeev  * Invalidate the TLB on this cpu only
690b2b3ffcdSSimon Schubert  */
691b2b3ffcdSSimon Schubert static __inline void
cpu_invltlb(void)692b2b3ffcdSSimon Schubert cpu_invltlb(void)
693b2b3ffcdSSimon Schubert {
694b2b3ffcdSSimon Schubert 	load_cr3(rcr3());
695a293ac08SJordan Gordeev #if defined(SWTCH_OPTIM_STATS)
696a293ac08SJordan Gordeev 	++tlb_flush_count;
697a293ac08SJordan Gordeev #endif
698b2b3ffcdSSimon Schubert }
699b2b3ffcdSSimon Schubert 
700a293ac08SJordan Gordeev #endif
701a293ac08SJordan Gordeev 
702b5516a55SSascha Wildner void smp_invltlb(void);
703b5516a55SSascha Wildner void smp_sniff(void);
704b5516a55SSascha Wildner void cpu_sniff(int);
705b5516a55SSascha Wildner void hard_sniff(struct trapframe *);
706ccd67bf6SMatthew Dillon 
7077529a5ccSSascha Wildner static __inline u_short
rfs(void)708b2b3ffcdSSimon Schubert rfs(void)
709b2b3ffcdSSimon Schubert {
7107529a5ccSSascha Wildner 	u_short sel;
7117529a5ccSSascha Wildner 	__asm __volatile("movw %%fs,%0" : "=rm" (sel));
712b2b3ffcdSSimon Schubert 	return (sel);
713b2b3ffcdSSimon Schubert }
714b2b3ffcdSSimon Schubert 
7157529a5ccSSascha Wildner static __inline u_short
rgs(void)716b2b3ffcdSSimon Schubert rgs(void)
717b2b3ffcdSSimon Schubert {
7187529a5ccSSascha Wildner 	u_short sel;
7197529a5ccSSascha Wildner 	__asm __volatile("movw %%gs,%0" : "=rm" (sel));
720b2b3ffcdSSimon Schubert 	return (sel);
721b2b3ffcdSSimon Schubert }
722b2b3ffcdSSimon Schubert 
723b2b3ffcdSSimon Schubert static __inline void
load_ds(u_short sel)7247529a5ccSSascha Wildner load_ds(u_short sel)
725b2b3ffcdSSimon Schubert {
7267529a5ccSSascha Wildner 	__asm __volatile("movw %0,%%ds" : : "rm" (sel));
727b2b3ffcdSSimon Schubert }
728b2b3ffcdSSimon Schubert 
729b2b3ffcdSSimon Schubert static __inline void
load_es(u_short sel)7307529a5ccSSascha Wildner load_es(u_short sel)
731b2b3ffcdSSimon Schubert {
7327529a5ccSSascha Wildner 	__asm __volatile("movw %0,%%es" : : "rm" (sel));
733b2b3ffcdSSimon Schubert }
734b2b3ffcdSSimon Schubert 
735b2b3ffcdSSimon Schubert #ifdef _KERNEL
736b2b3ffcdSSimon Schubert /* This is defined in <machine/specialreg.h> but is too painful to get to */
737b2b3ffcdSSimon Schubert #ifndef	MSR_FSBASE
738b2b3ffcdSSimon Schubert #define	MSR_FSBASE	0xc0000100
739b2b3ffcdSSimon Schubert #endif
740b2b3ffcdSSimon Schubert static __inline void
load_fs(u_short sel)7417529a5ccSSascha Wildner load_fs(u_short sel)
742b2b3ffcdSSimon Schubert {
743b2b3ffcdSSimon Schubert 	/* Preserve the fsbase value across the selector load */
7447529a5ccSSascha Wildner 	__asm __volatile("rdmsr; movw %0,%%fs; wrmsr"
745b2b3ffcdSSimon Schubert             : : "rm" (sel), "c" (MSR_FSBASE) : "eax", "edx");
746b2b3ffcdSSimon Schubert }
747b2b3ffcdSSimon Schubert 
748b2b3ffcdSSimon Schubert #ifndef	MSR_GSBASE
749b2b3ffcdSSimon Schubert #define	MSR_GSBASE	0xc0000101
750b2b3ffcdSSimon Schubert #endif
751b2b3ffcdSSimon Schubert static __inline void
load_gs(u_short sel)7527529a5ccSSascha Wildner load_gs(u_short sel)
753b2b3ffcdSSimon Schubert {
754b2b3ffcdSSimon Schubert 	/*
755b2b3ffcdSSimon Schubert 	 * Preserve the gsbase value across the selector load.
756b2b3ffcdSSimon Schubert 	 * Note that we have to disable interrupts because the gsbase
757b2b3ffcdSSimon Schubert 	 * being trashed happens to be the kernel gsbase at the time.
758b2b3ffcdSSimon Schubert 	 */
759b2b3ffcdSSimon Schubert 	__asm __volatile("pushfq; cli; rdmsr; movw %0,%%gs; wrmsr; popfq"
760b2b3ffcdSSimon Schubert             : : "rm" (sel), "c" (MSR_GSBASE) : "eax", "edx");
761b2b3ffcdSSimon Schubert }
762b2b3ffcdSSimon Schubert #else
763b2b3ffcdSSimon Schubert /* Usable by userland */
764b2b3ffcdSSimon Schubert static __inline void
load_fs(u_short sel)7657529a5ccSSascha Wildner load_fs(u_short sel)
766b2b3ffcdSSimon Schubert {
7677529a5ccSSascha Wildner 	__asm __volatile("movw %0,%%fs" : : "rm" (sel));
768b2b3ffcdSSimon Schubert }
769b2b3ffcdSSimon Schubert 
770b2b3ffcdSSimon Schubert static __inline void
load_gs(u_short sel)7717529a5ccSSascha Wildner load_gs(u_short sel)
772b2b3ffcdSSimon Schubert {
7737529a5ccSSascha Wildner 	__asm __volatile("movw %0,%%gs" : : "rm" (sel));
774b2b3ffcdSSimon Schubert }
775b2b3ffcdSSimon Schubert #endif
776b2b3ffcdSSimon Schubert 
777b2b3ffcdSSimon Schubert /* void lidt(struct region_descriptor *addr); */
778b2b3ffcdSSimon Schubert static __inline void
lidt(struct region_descriptor * addr)779b2b3ffcdSSimon Schubert lidt(struct region_descriptor *addr)
780b2b3ffcdSSimon Schubert {
781b2b3ffcdSSimon Schubert 	__asm __volatile("lidt (%0)" : : "r" (addr));
782b2b3ffcdSSimon Schubert }
783b2b3ffcdSSimon Schubert 
784b2b3ffcdSSimon Schubert /* void lldt(u_short sel); */
785b2b3ffcdSSimon Schubert static __inline void
lldt(u_short sel)786b2b3ffcdSSimon Schubert lldt(u_short sel)
787b2b3ffcdSSimon Schubert {
788b2b3ffcdSSimon Schubert 	__asm __volatile("lldt %0" : : "r" (sel));
789b2b3ffcdSSimon Schubert }
790b2b3ffcdSSimon Schubert 
791b2b3ffcdSSimon Schubert /* void ltr(u_short sel); */
792b2b3ffcdSSimon Schubert static __inline void
ltr(u_short sel)793b2b3ffcdSSimon Schubert ltr(u_short sel)
794b2b3ffcdSSimon Schubert {
795b2b3ffcdSSimon Schubert 	__asm __volatile("ltr %0" : : "r" (sel));
796b2b3ffcdSSimon Schubert }
797b2b3ffcdSSimon Schubert 
798b2b3ffcdSSimon Schubert static __inline u_int64_t
rdr0(void)799b2b3ffcdSSimon Schubert rdr0(void)
800b2b3ffcdSSimon Schubert {
801b2b3ffcdSSimon Schubert 	u_int64_t data;
802b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%dr0,%0" : "=r" (data));
803b2b3ffcdSSimon Schubert 	return (data);
804b2b3ffcdSSimon Schubert }
805b2b3ffcdSSimon Schubert 
806b2b3ffcdSSimon Schubert static __inline void
load_dr0(u_int64_t dr0)807b2b3ffcdSSimon Schubert load_dr0(u_int64_t dr0)
808b2b3ffcdSSimon Schubert {
8099ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%dr0" : : "r" (dr0) : "memory");
810b2b3ffcdSSimon Schubert }
811b2b3ffcdSSimon Schubert 
812b2b3ffcdSSimon Schubert static __inline u_int64_t
rdr1(void)813b2b3ffcdSSimon Schubert rdr1(void)
814b2b3ffcdSSimon Schubert {
815b2b3ffcdSSimon Schubert 	u_int64_t data;
816b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%dr1,%0" : "=r" (data));
817b2b3ffcdSSimon Schubert 	return (data);
818b2b3ffcdSSimon Schubert }
819b2b3ffcdSSimon Schubert 
820b2b3ffcdSSimon Schubert static __inline void
load_dr1(u_int64_t dr1)821b2b3ffcdSSimon Schubert load_dr1(u_int64_t dr1)
822b2b3ffcdSSimon Schubert {
8239ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%dr1" : : "r" (dr1) : "memory");
824b2b3ffcdSSimon Schubert }
825b2b3ffcdSSimon Schubert 
826b2b3ffcdSSimon Schubert static __inline u_int64_t
rdr2(void)827b2b3ffcdSSimon Schubert rdr2(void)
828b2b3ffcdSSimon Schubert {
829b2b3ffcdSSimon Schubert 	u_int64_t data;
830b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%dr2,%0" : "=r" (data));
831b2b3ffcdSSimon Schubert 	return (data);
832b2b3ffcdSSimon Schubert }
833b2b3ffcdSSimon Schubert 
834b2b3ffcdSSimon Schubert static __inline void
load_dr2(u_int64_t dr2)835b2b3ffcdSSimon Schubert load_dr2(u_int64_t dr2)
836b2b3ffcdSSimon Schubert {
8379ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%dr2" : : "r" (dr2) : "memory");
838b2b3ffcdSSimon Schubert }
839b2b3ffcdSSimon Schubert 
840b2b3ffcdSSimon Schubert static __inline u_int64_t
rdr3(void)841b2b3ffcdSSimon Schubert rdr3(void)
842b2b3ffcdSSimon Schubert {
843b2b3ffcdSSimon Schubert 	u_int64_t data;
844b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%dr3,%0" : "=r" (data));
845b2b3ffcdSSimon Schubert 	return (data);
846b2b3ffcdSSimon Schubert }
847b2b3ffcdSSimon Schubert 
848b2b3ffcdSSimon Schubert static __inline void
load_dr3(u_int64_t dr3)849b2b3ffcdSSimon Schubert load_dr3(u_int64_t dr3)
850b2b3ffcdSSimon Schubert {
8519ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%dr3" : : "r" (dr3) : "memory");
852b2b3ffcdSSimon Schubert }
853b2b3ffcdSSimon Schubert 
854b2b3ffcdSSimon Schubert static __inline u_int64_t
rdr4(void)855b2b3ffcdSSimon Schubert rdr4(void)
856b2b3ffcdSSimon Schubert {
857b2b3ffcdSSimon Schubert 	u_int64_t data;
858b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%dr4,%0" : "=r" (data));
859b2b3ffcdSSimon Schubert 	return (data);
860b2b3ffcdSSimon Schubert }
861b2b3ffcdSSimon Schubert 
862b2b3ffcdSSimon Schubert static __inline void
load_dr4(u_int64_t dr4)863b2b3ffcdSSimon Schubert load_dr4(u_int64_t dr4)
864b2b3ffcdSSimon Schubert {
8659ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%dr4" : : "r" (dr4) : "memory");
866b2b3ffcdSSimon Schubert }
867b2b3ffcdSSimon Schubert 
868b2b3ffcdSSimon Schubert static __inline u_int64_t
rdr5(void)869b2b3ffcdSSimon Schubert rdr5(void)
870b2b3ffcdSSimon Schubert {
871b2b3ffcdSSimon Schubert 	u_int64_t data;
872b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%dr5,%0" : "=r" (data));
873b2b3ffcdSSimon Schubert 	return (data);
874b2b3ffcdSSimon Schubert }
875b2b3ffcdSSimon Schubert 
876b2b3ffcdSSimon Schubert static __inline void
load_dr5(u_int64_t dr5)877b2b3ffcdSSimon Schubert load_dr5(u_int64_t dr5)
878b2b3ffcdSSimon Schubert {
8799ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%dr5" : : "r" (dr5) : "memory");
880b2b3ffcdSSimon Schubert }
881b2b3ffcdSSimon Schubert 
882b2b3ffcdSSimon Schubert static __inline u_int64_t
rdr6(void)883b2b3ffcdSSimon Schubert rdr6(void)
884b2b3ffcdSSimon Schubert {
885b2b3ffcdSSimon Schubert 	u_int64_t data;
886b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%dr6,%0" : "=r" (data));
887b2b3ffcdSSimon Schubert 	return (data);
888b2b3ffcdSSimon Schubert }
889b2b3ffcdSSimon Schubert 
890b2b3ffcdSSimon Schubert static __inline void
load_dr6(u_int64_t dr6)891b2b3ffcdSSimon Schubert load_dr6(u_int64_t dr6)
892b2b3ffcdSSimon Schubert {
8939ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%dr6" : : "r" (dr6) : "memory");
894b2b3ffcdSSimon Schubert }
895b2b3ffcdSSimon Schubert 
896b2b3ffcdSSimon Schubert static __inline u_int64_t
rdr7(void)897b2b3ffcdSSimon Schubert rdr7(void)
898b2b3ffcdSSimon Schubert {
899b2b3ffcdSSimon Schubert 	u_int64_t data;
900b2b3ffcdSSimon Schubert 	__asm __volatile("movq %%dr7,%0" : "=r" (data));
901b2b3ffcdSSimon Schubert 	return (data);
902b2b3ffcdSSimon Schubert }
903b2b3ffcdSSimon Schubert 
904b2b3ffcdSSimon Schubert static __inline void
load_dr7(u_int64_t dr7)905b2b3ffcdSSimon Schubert load_dr7(u_int64_t dr7)
906b2b3ffcdSSimon Schubert {
9079ba9002bSMatthew Dillon 	__asm __volatile("movq %0,%%dr7" : : "r" (dr7) : "memory");
908b2b3ffcdSSimon Schubert }
909b2b3ffcdSSimon Schubert 
910b2b3ffcdSSimon Schubert static __inline register_t
intr_disable(void)911b2b3ffcdSSimon Schubert intr_disable(void)
912b2b3ffcdSSimon Schubert {
913b2b3ffcdSSimon Schubert 	register_t rflags;
914b2b3ffcdSSimon Schubert 
915b2b3ffcdSSimon Schubert 	rflags = read_rflags();
916b2b3ffcdSSimon Schubert 	cpu_disable_intr();
917b2b3ffcdSSimon Schubert 	return (rflags);
918b2b3ffcdSSimon Schubert }
919b2b3ffcdSSimon Schubert 
920b2b3ffcdSSimon Schubert static __inline void
intr_restore(register_t rflags)921b2b3ffcdSSimon Schubert intr_restore(register_t rflags)
922b2b3ffcdSSimon Schubert {
923b2b3ffcdSSimon Schubert 	write_rflags(rflags);
924b2b3ffcdSSimon Schubert }
925b2b3ffcdSSimon Schubert 
926b2b3ffcdSSimon Schubert #else /* !__GNUC__ */
927b2b3ffcdSSimon Schubert 
928b2b3ffcdSSimon Schubert int	breakpoint(void);
929b2b3ffcdSSimon Schubert void	cpu_pause(void);
930b2b3ffcdSSimon Schubert u_int	bsfl(u_int mask);
931b2b3ffcdSSimon Schubert u_int	bsrl(u_int mask);
932b2b3ffcdSSimon Schubert void	cpu_disable_intr(void);
933b2b3ffcdSSimon Schubert void	cpu_enable_intr(void);
934b2b3ffcdSSimon Schubert void	cpu_invlpg(u_long addr);
935b2b3ffcdSSimon Schubert void	cpu_invlpg_range(u_long start, u_long end);
936b2b3ffcdSSimon Schubert void	do_cpuid(u_int ax, u_int *p);
937b2b3ffcdSSimon Schubert void	halt(void);
938b2b3ffcdSSimon Schubert u_char	inb(u_int port);
939b2b3ffcdSSimon Schubert u_int	inl(u_int port);
940b2b3ffcdSSimon Schubert void	insb(u_int port, void *addr, size_t cnt);
941b2b3ffcdSSimon Schubert void	insl(u_int port, void *addr, size_t cnt);
942b2b3ffcdSSimon Schubert void	insw(u_int port, void *addr, size_t cnt);
943b2b3ffcdSSimon Schubert void	invd(void);
944b2b3ffcdSSimon Schubert void	invlpg_range(u_int start, u_int end);
945b2b3ffcdSSimon Schubert void	cpu_invltlb(void);
946b2b3ffcdSSimon Schubert u_short	inw(u_int port);
947b2b3ffcdSSimon Schubert void	load_cr0(u_int cr0);
948*251f5dcbSAaron LI void	load_cr2(u_int cr2);
949b2b3ffcdSSimon Schubert void	load_cr3(u_int cr3);
950b2b3ffcdSSimon Schubert void	load_cr4(u_int cr4);
951b2b3ffcdSSimon Schubert void	load_fs(u_int sel);
952b2b3ffcdSSimon Schubert void	load_gs(u_int sel);
953b2b3ffcdSSimon Schubert void	lidt(struct region_descriptor *addr);
954b2b3ffcdSSimon Schubert void	lldt(u_short sel);
955b2b3ffcdSSimon Schubert void	ltr(u_short sel);
956b2b3ffcdSSimon Schubert void	outb(u_int port, u_char data);
957b2b3ffcdSSimon Schubert void	outl(u_int port, u_int data);
958b2b3ffcdSSimon Schubert void	outsb(u_int port, void *addr, size_t cnt);
959b2b3ffcdSSimon Schubert void	outsl(u_int port, void *addr, size_t cnt);
960b2b3ffcdSSimon Schubert void	outsw(u_int port, void *addr, size_t cnt);
961b2b3ffcdSSimon Schubert void	outw(u_int port, u_short data);
962b2b3ffcdSSimon Schubert void	ia32_pause(void);
963b2b3ffcdSSimon Schubert u_int	rcr0(void);
964b2b3ffcdSSimon Schubert u_int	rcr2(void);
965b2b3ffcdSSimon Schubert u_int	rcr3(void);
966b2b3ffcdSSimon Schubert u_int	rcr4(void);
9677529a5ccSSascha Wildner u_short	rfs(void);
9687529a5ccSSascha Wildner u_short	rgs(void);
969b2b3ffcdSSimon Schubert u_int64_t rdmsr(u_int msr);
970b2b3ffcdSSimon Schubert u_int64_t rdpmc(u_int pmc);
9715b49787bSMatthew Dillon tsc_uclock_t rdtsc(void);
972b2b3ffcdSSimon Schubert u_int	read_rflags(void);
973b2b3ffcdSSimon Schubert void	wbinvd(void);
974b2b3ffcdSSimon Schubert void	write_rflags(u_int rf);
975b2b3ffcdSSimon Schubert void	wrmsr(u_int msr, u_int64_t newval);
976b2b3ffcdSSimon Schubert u_int64_t	rdr0(void);
977b2b3ffcdSSimon Schubert void	load_dr0(u_int64_t dr0);
978b2b3ffcdSSimon Schubert u_int64_t	rdr1(void);
979b2b3ffcdSSimon Schubert void	load_dr1(u_int64_t dr1);
980b2b3ffcdSSimon Schubert u_int64_t	rdr2(void);
981b2b3ffcdSSimon Schubert void	load_dr2(u_int64_t dr2);
982b2b3ffcdSSimon Schubert u_int64_t	rdr3(void);
983b2b3ffcdSSimon Schubert void	load_dr3(u_int64_t dr3);
984b2b3ffcdSSimon Schubert u_int64_t	rdr4(void);
985b2b3ffcdSSimon Schubert void	load_dr4(u_int64_t dr4);
986b2b3ffcdSSimon Schubert u_int64_t	rdr5(void);
987b2b3ffcdSSimon Schubert void	load_dr5(u_int64_t dr5);
988b2b3ffcdSSimon Schubert u_int64_t	rdr6(void);
989b2b3ffcdSSimon Schubert void	load_dr6(u_int64_t dr6);
990b2b3ffcdSSimon Schubert u_int64_t	rdr7(void);
991b2b3ffcdSSimon Schubert void	load_dr7(u_int64_t dr7);
992b2b3ffcdSSimon Schubert register_t	intr_disable(void);
993b2b3ffcdSSimon Schubert void	intr_restore(register_t rf);
994b2b3ffcdSSimon Schubert 
995b2b3ffcdSSimon Schubert #endif	/* __GNUC__ */
996b2b3ffcdSSimon Schubert 
9970bdfdda1SSascha Wildner int	rdmsr_safe(u_int msr, uint64_t *val);
998d4ef6694SJoris Giovannangeli int wrmsr_safe(u_int msr, uint64_t newval);
999b2b3ffcdSSimon Schubert void	reset_dbregs(void);
10008d2aaeecSMatthew Dillon void	smap_open(void);
10018d2aaeecSMatthew Dillon void	smap_close(void);
1002b2b3ffcdSSimon Schubert 
1003b2b3ffcdSSimon Schubert __END_DECLS
1004b2b3ffcdSSimon Schubert 
1005b2b3ffcdSSimon Schubert #endif /* !_CPU_CPUFUNC_H_ */
1006