xref: /freebsd/sys/cddl/compat/opensolaris/sys/atomic.h (revision 315ee00f)
1 /*-
2  * Copyright (c) 2007 Pawel Jakub Dawidek <pjd@FreeBSD.org>
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' AND
15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHORS OR CONTRIBUTORS BE LIABLE
18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  */
26 
27 #ifndef _OPENSOLARIS_SYS_ATOMIC_H_
28 #define	_OPENSOLARIS_SYS_ATOMIC_H_
29 
30 #include <sys/types.h>
31 #include <machine/atomic.h>
32 
33 #if defined(__i386__) && (defined(_KERNEL) || defined(KLD_MODULE))
34 #define	I386_HAVE_ATOMIC64
35 #endif
36 
37 #if defined(__i386__) || defined(__amd64__) || defined(__arm__)
38 /* No spurious failures from fcmpset. */
39 #define	STRONG_FCMPSET
40 #endif
41 
42 #if !defined(__LP64__) && !defined(__mips_n32) && \
43     !defined(ARM_HAVE_ATOMIC64) && !defined(I386_HAVE_ATOMIC64) && \
44     !defined(HAS_EMULATED_ATOMIC64)
45 extern void atomic_add_64(volatile uint64_t *target, int64_t delta);
46 extern void atomic_dec_64(volatile uint64_t *target);
47 extern uint64_t atomic_swap_64(volatile uint64_t *a, uint64_t value);
48 extern uint64_t atomic_load_64(volatile uint64_t *a);
49 extern uint64_t atomic_add_64_nv(volatile uint64_t *target, int64_t delta);
50 extern uint64_t atomic_cas_64(volatile uint64_t *target, uint64_t cmp,
51     uint64_t newval);
52 #endif
53 
54 #define	membar_producer	atomic_thread_fence_rel
55 
56 static __inline uint32_t
57 atomic_add_32_nv(volatile uint32_t *target, int32_t delta)
58 {
59 	return (atomic_fetchadd_32(target, delta) + delta);
60 }
61 
62 static __inline u_int
63 atomic_add_int_nv(volatile u_int *target, int delta)
64 {
65 	return (atomic_add_32_nv(target, delta));
66 }
67 
68 static __inline void
69 atomic_inc_32(volatile uint32_t *target)
70 {
71 	atomic_add_32(target, 1);
72 }
73 
74 static __inline uint32_t
75 atomic_inc_32_nv(volatile uint32_t *target)
76 {
77 	return (atomic_add_32_nv(target, 1));
78 }
79 
80 static __inline void
81 atomic_dec_32(volatile uint32_t *target)
82 {
83 	atomic_subtract_32(target, 1);
84 }
85 
86 static __inline uint32_t
87 atomic_dec_32_nv(volatile uint32_t *target)
88 {
89 	return (atomic_add_32_nv(target, -1));
90 }
91 
92 static inline uint32_t
93 atomic_cas_32(volatile uint32_t *target, uint32_t cmp, uint32_t newval)
94 {
95 #ifdef STRONG_FCMPSET
96 	(void)atomic_fcmpset_32(target, &cmp, newval);
97 #else
98 	uint32_t expected = cmp;
99 
100 	do {
101 		if (atomic_fcmpset_32(target, &cmp, newval))
102 			break;
103 	} while (cmp == expected);
104 #endif
105 	return (cmp);
106 }
107 
108 #if defined(__LP64__) || defined(__mips_n32) || \
109     defined(ARM_HAVE_ATOMIC64) || defined(I386_HAVE_ATOMIC64) || \
110     defined(HAS_EMULATED_ATOMIC64)
111 static __inline void
112 atomic_dec_64(volatile uint64_t *target)
113 {
114 	atomic_subtract_64(target, 1);
115 }
116 
117 static inline uint64_t
118 atomic_add_64_nv(volatile uint64_t *target, int64_t delta)
119 {
120 	return (atomic_fetchadd_64(target, delta) + delta);
121 }
122 
123 static inline uint64_t
124 atomic_cas_64(volatile uint64_t *target, uint64_t cmp, uint64_t newval)
125 {
126 #ifdef STRONG_FCMPSET
127 	(void)atomic_fcmpset_64(target, &cmp, newval);
128 #else
129 	uint64_t expected = cmp;
130 
131 	do {
132 		if (atomic_fcmpset_64(target, &cmp, newval))
133 			break;
134 	} while (cmp == expected);
135 #endif
136 	return (cmp);
137 }
138 #endif
139 
140 static __inline void
141 atomic_inc_64(volatile uint64_t *target)
142 {
143 	atomic_add_64(target, 1);
144 }
145 
146 static __inline uint64_t
147 atomic_inc_64_nv(volatile uint64_t *target)
148 {
149 	return (atomic_add_64_nv(target, 1));
150 }
151 
152 static __inline uint64_t
153 atomic_dec_64_nv(volatile uint64_t *target)
154 {
155 	return (atomic_add_64_nv(target, -1));
156 }
157 
158 #ifdef __LP64__
159 static __inline void *
160 atomic_cas_ptr(volatile void *target, void *cmp,  void *newval)
161 {
162 	return ((void *)atomic_cas_64((volatile uint64_t *)target,
163 	    (uint64_t)cmp, (uint64_t)newval));
164 }
165 #else
166 static __inline void *
167 atomic_cas_ptr(volatile void *target, void *cmp,  void *newval)
168 {
169 	return ((void *)atomic_cas_32((volatile uint32_t *)target,
170 	    (uint32_t)cmp, (uint32_t)newval));
171 }
172 #endif	/* __LP64__ */
173 
174 #endif	/* !_OPENSOLARIS_SYS_ATOMIC_H_ */
175