xref: /freebsd/sys/contrib/ck/include/ck_rwcohort.h (revision d6b92ffa)
1 /*
2  * Copyright 2013-2015 Samy Al Bahra.
3  * Copyright 2013 Brendon Scheinman.
4  * All rights reserved.
5  *
6  * Redistribution and use in source and binary forms, with or without
7  * modification, are permitted provided that the following conditions
8  * are met:
9  * 1. Redistributions of source code must retain the above copyright
10  *    notice, this list of conditions and the following disclaimer.
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  *
15  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
16  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
19  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25  * SUCH DAMAGE.
26  */
27 
28 #ifndef CK_RWCOHORT_H
29 #define CK_RWCOHORT_H
30 
31 /*
32  * This is an implementation of NUMA-aware reader-writer locks as described in:
33  *     Calciu, I.; Dice, D.; Lev, Y.; Luchangco, V.; Marathe, V.; and Shavit, N. 2014.
34  *     NUMA-Aware Reader-Writer Locks
35  */
36 
37 #include <ck_cc.h>
38 #include <ck_pr.h>
39 #include <ck_stddef.h>
40 #include <ck_cohort.h>
41 
42 #define CK_RWCOHORT_WP_NAME(N) ck_rwcohort_wp_##N
43 #define CK_RWCOHORT_WP_INSTANCE(N) struct CK_RWCOHORT_WP_NAME(N)
44 #define CK_RWCOHORT_WP_INIT(N, RW, WL) ck_rwcohort_wp_##N##_init(RW, WL)
45 #define CK_RWCOHORT_WP_READ_LOCK(N, RW, C, GC, LC)	\
46 	ck_rwcohort_wp_##N##_read_lock(RW, C, GC, LC)
47 #define CK_RWCOHORT_WP_READ_UNLOCK(N, RW, C, GC, LC)	\
48 	ck_rwcohort_wp_##N##_read_unlock(RW)
49 #define CK_RWCOHORT_WP_WRITE_LOCK(N, RW, C, GC, LC)	\
50 	ck_rwcohort_wp_##N##_write_lock(RW, C, GC, LC)
51 #define CK_RWCOHORT_WP_WRITE_UNLOCK(N, RW, C, GC, LC)	\
52 	ck_rwcohort_wp_##N##_write_unlock(RW, C, GC, LC)
53 #define CK_RWCOHORT_WP_DEFAULT_WAIT_LIMIT 1000
54 
55 #define CK_RWCOHORT_WP_PROTOTYPE(N)							\
56 	CK_RWCOHORT_WP_INSTANCE(N) {							\
57 		unsigned int read_counter;						\
58 		unsigned int write_barrier;						\
59 		unsigned int wait_limit;						\
60 	};										\
61 	CK_CC_INLINE static void							\
62 	ck_rwcohort_wp_##N##_init(CK_RWCOHORT_WP_INSTANCE(N) *rw_cohort,		\
63 	    unsigned int wait_limit)							\
64 	{										\
65 											\
66 		rw_cohort->read_counter = 0;						\
67 		rw_cohort->write_barrier = 0;						\
68 		rw_cohort->wait_limit = wait_limit;					\
69 		ck_pr_barrier();							\
70 		return;									\
71 	}										\
72 	CK_CC_INLINE static void							\
73 	ck_rwcohort_wp_##N##_write_lock(CK_RWCOHORT_WP_INSTANCE(N) *rw_cohort,		\
74 	    CK_COHORT_INSTANCE(N) *cohort, void *global_context,			\
75 	    void *local_context)							\
76 	{										\
77 											\
78 		while (ck_pr_load_uint(&rw_cohort->write_barrier) > 0)			\
79 			ck_pr_stall();							\
80 											\
81 		CK_COHORT_LOCK(N, cohort, global_context, local_context);		\
82 											\
83 		while (ck_pr_load_uint(&rw_cohort->read_counter) > 0) 			\
84 			ck_pr_stall();							\
85 											\
86 		return;									\
87 	}										\
88 	CK_CC_INLINE static void							\
89 	ck_rwcohort_wp_##N##_write_unlock(CK_RWCOHORT_WP_INSTANCE(N) *rw_cohort,	\
90 	    CK_COHORT_INSTANCE(N) *cohort, void *global_context,			\
91 	    void *local_context)							\
92 	{										\
93 											\
94 		(void)rw_cohort;							\
95 		CK_COHORT_UNLOCK(N, cohort, global_context, local_context);		\
96 		return;									\
97 	}										\
98 	CK_CC_INLINE static void							\
99 	ck_rwcohort_wp_##N##_read_lock(CK_RWCOHORT_WP_INSTANCE(N) *rw_cohort,		\
100 	    CK_COHORT_INSTANCE(N) *cohort, void *global_context,			\
101 	    void *local_context)							\
102 	{										\
103 		unsigned int wait_count = 0;						\
104 		bool raised = false;							\
105 											\
106 		for (;;) {								\
107 			ck_pr_inc_uint(&rw_cohort->read_counter);			\
108 			ck_pr_fence_atomic_load();					\
109 			if (CK_COHORT_LOCKED(N, cohort, global_context,			\
110 			    local_context) == false)					\
111 				break;							\
112 											\
113 			ck_pr_dec_uint(&rw_cohort->read_counter);			\
114 			while (CK_COHORT_LOCKED(N, cohort, global_context,		\
115 			    local_context) == true) {					\
116 				ck_pr_stall();						\
117 				if (++wait_count > rw_cohort->wait_limit &&		\
118 				    raised == false) {					\
119 					ck_pr_inc_uint(&rw_cohort->write_barrier);	\
120 					raised = true;					\
121 				}							\
122 			}								\
123 		}									\
124 											\
125 		if (raised == true)							\
126 			ck_pr_dec_uint(&rw_cohort->write_barrier);			\
127 											\
128 		ck_pr_fence_load();							\
129 		return;									\
130 	}										\
131 	CK_CC_INLINE static void							\
132 	ck_rwcohort_wp_##N##_read_unlock(CK_RWCOHORT_WP_INSTANCE(N) *cohort)		\
133 	{										\
134 											\
135 		ck_pr_fence_load_atomic();						\
136 		ck_pr_dec_uint(&cohort->read_counter);					\
137 		return;									\
138 	}
139 
140 #define CK_RWCOHORT_WP_INITIALIZER {							\
141 	.read_counter = 0,								\
142 	.write_barrier = 0,								\
143 	.wait_limit = 0									\
144 }
145 
146 #define CK_RWCOHORT_RP_NAME(N) ck_rwcohort_rp_##N
147 #define CK_RWCOHORT_RP_INSTANCE(N) struct CK_RWCOHORT_RP_NAME(N)
148 #define CK_RWCOHORT_RP_INIT(N, RW, WL) ck_rwcohort_rp_##N##_init(RW, WL)
149 #define CK_RWCOHORT_RP_READ_LOCK(N, RW, C, GC, LC)	\
150 	ck_rwcohort_rp_##N##_read_lock(RW, C, GC, LC)
151 #define CK_RWCOHORT_RP_READ_UNLOCK(N, RW, C, GC, LC)	\
152 	ck_rwcohort_rp_##N##_read_unlock(RW)
153 #define CK_RWCOHORT_RP_WRITE_LOCK(N, RW, C, GC, LC)	\
154 	ck_rwcohort_rp_##N##_write_lock(RW, C, GC, LC)
155 #define CK_RWCOHORT_RP_WRITE_UNLOCK(N, RW, C, GC, LC)	\
156 	ck_rwcohort_rp_##N##_write_unlock(RW, C, GC, LC)
157 #define CK_RWCOHORT_RP_DEFAULT_WAIT_LIMIT 1000
158 
159 #define CK_RWCOHORT_RP_PROTOTYPE(N)							\
160 	CK_RWCOHORT_RP_INSTANCE(N) {							\
161 		unsigned int read_counter;						\
162 		unsigned int read_barrier;						\
163 		unsigned int wait_limit;						\
164 	};										\
165 	CK_CC_INLINE static void							\
166 	ck_rwcohort_rp_##N##_init(CK_RWCOHORT_RP_INSTANCE(N) *rw_cohort,		\
167 	    unsigned int wait_limit)							\
168 	{										\
169 											\
170 		rw_cohort->read_counter = 0;						\
171 		rw_cohort->read_barrier = 0;						\
172 		rw_cohort->wait_limit = wait_limit;					\
173 		ck_pr_barrier();							\
174 		return;									\
175 	}										\
176 	CK_CC_INLINE static void							\
177 	ck_rwcohort_rp_##N##_write_lock(CK_RWCOHORT_RP_INSTANCE(N) *rw_cohort,		\
178 	    CK_COHORT_INSTANCE(N) *cohort, void *global_context,			\
179 	    void *local_context)							\
180 	{										\
181 		unsigned int wait_count = 0;						\
182 		bool raised = false;							\
183 											\
184 		for (;;) {								\
185 			CK_COHORT_LOCK(N, cohort, global_context, local_context);	\
186 			if (ck_pr_load_uint(&rw_cohort->read_counter) == 0)		\
187 				break;							\
188 											\
189 			CK_COHORT_UNLOCK(N, cohort, global_context, local_context);	\
190 			while (ck_pr_load_uint(&rw_cohort->read_counter) > 0) {		\
191 				ck_pr_stall();						\
192 				if (++wait_count > rw_cohort->wait_limit &&		\
193 				    raised == false) {					\
194 					ck_pr_inc_uint(&rw_cohort->read_barrier);	\
195 					raised = true;					\
196 				}							\
197 			}								\
198 		}									\
199 											\
200 		if (raised == true)							\
201 			ck_pr_dec_uint(&rw_cohort->read_barrier);			\
202 											\
203 		return;									\
204 	}										\
205 	CK_CC_INLINE static void							\
206 	ck_rwcohort_rp_##N##_write_unlock(CK_RWCOHORT_RP_INSTANCE(N) *rw_cohort,	\
207 	    CK_COHORT_INSTANCE(N) *cohort, void *global_context, void *local_context)	\
208 	{										\
209 											\
210 		(void)rw_cohort;							\
211 		CK_COHORT_UNLOCK(N, cohort, global_context, local_context);		\
212 		return;									\
213 	}										\
214 	CK_CC_INLINE static void							\
215 	ck_rwcohort_rp_##N##_read_lock(CK_RWCOHORT_RP_INSTANCE(N) *rw_cohort,		\
216 	    CK_COHORT_INSTANCE(N) *cohort, void *global_context,			\
217 	    void *local_context)							\
218 	{										\
219 											\
220 		while (ck_pr_load_uint(&rw_cohort->read_barrier) > 0)			\
221 			ck_pr_stall();							\
222 											\
223 		ck_pr_inc_uint(&rw_cohort->read_counter);				\
224 		ck_pr_fence_atomic_load();						\
225 											\
226 		while (CK_COHORT_LOCKED(N, cohort, global_context,			\
227 		    local_context) == true)						\
228 			ck_pr_stall();							\
229 											\
230 		return;									\
231 	}										\
232 	CK_CC_INLINE static void							\
233 	ck_rwcohort_rp_##N##_read_unlock(CK_RWCOHORT_RP_INSTANCE(N) *cohort)		\
234 	{										\
235 											\
236 		ck_pr_fence_load_atomic();						\
237 		ck_pr_dec_uint(&cohort->read_counter);					\
238 		return;									\
239 	}
240 
241 #define CK_RWCOHORT_RP_INITIALIZER {							\
242 	.read_counter = 0,								\
243 	.read_barrier = 0,								\
244 	.wait_limit = 0									\
245 }
246 
247 #define CK_RWCOHORT_NEUTRAL_NAME(N) ck_rwcohort_neutral_##N
248 #define CK_RWCOHORT_NEUTRAL_INSTANCE(N) struct CK_RWCOHORT_NEUTRAL_NAME(N)
249 #define CK_RWCOHORT_NEUTRAL_INIT(N, RW) ck_rwcohort_neutral_##N##_init(RW)
250 #define CK_RWCOHORT_NEUTRAL_READ_LOCK(N, RW, C, GC, LC)		\
251 	ck_rwcohort_neutral_##N##_read_lock(RW, C, GC, LC)
252 #define CK_RWCOHORT_NEUTRAL_READ_UNLOCK(N, RW, C, GC, LC)	\
253 	ck_rwcohort_neutral_##N##_read_unlock(RW)
254 #define CK_RWCOHORT_NEUTRAL_WRITE_LOCK(N, RW, C, GC, LC)	\
255 	ck_rwcohort_neutral_##N##_write_lock(RW, C, GC, LC)
256 #define CK_RWCOHORT_NEUTRAL_WRITE_UNLOCK(N, RW, C, GC, LC)	\
257 	ck_rwcohort_neutral_##N##_write_unlock(RW, C, GC, LC)
258 #define CK_RWCOHORT_NEUTRAL_DEFAULT_WAIT_LIMIT 1000
259 
260 #define CK_RWCOHORT_NEUTRAL_PROTOTYPE(N)						\
261 	CK_RWCOHORT_NEUTRAL_INSTANCE(N) {						\
262 		unsigned int read_counter;						\
263 	};										\
264 	CK_CC_INLINE static void							\
265 	ck_rwcohort_neutral_##N##_init(CK_RWCOHORT_NEUTRAL_INSTANCE(N) *rw_cohort)	\
266 	{										\
267 											\
268 		rw_cohort->read_counter = 0;						\
269 		ck_pr_barrier();							\
270 		return;									\
271 	}										\
272 	CK_CC_INLINE static void							\
273 	ck_rwcohort_neutral_##N##_write_lock(CK_RWCOHORT_NEUTRAL_INSTANCE(N) *rw_cohort,\
274 	    CK_COHORT_INSTANCE(N) *cohort, void *global_context,			\
275 	    void *local_context)							\
276 	{										\
277 											\
278 		CK_COHORT_LOCK(N, cohort, global_context, local_context);		\
279 		while (ck_pr_load_uint(&rw_cohort->read_counter) > 0) {			\
280 			ck_pr_stall();							\
281 		}									\
282 		return;									\
283 	}										\
284 	CK_CC_INLINE static void							\
285 	ck_rwcohort_neutral_##N##_write_unlock(CK_RWCOHORT_NEUTRAL_INSTANCE(N) *rw_cohort,\
286 	    CK_COHORT_INSTANCE(N) *cohort, void *global_context, void *local_context)	\
287 	{										\
288 											\
289 		(void)rw_cohort;							\
290 		CK_COHORT_UNLOCK(N, cohort, global_context, local_context);		\
291 		return;									\
292 	}										\
293 	CK_CC_INLINE static void							\
294 	ck_rwcohort_neutral_##N##_read_lock(CK_RWCOHORT_NEUTRAL_INSTANCE(N) *rw_cohort,	\
295 	    CK_COHORT_INSTANCE(N) *cohort, void *global_context,			\
296 	    void *local_context)							\
297 	{										\
298 											\
299 		CK_COHORT_LOCK(N, cohort, global_context, local_context);		\
300 		ck_pr_inc_uint(&rw_cohort->read_counter);				\
301 		CK_COHORT_UNLOCK(N, cohort, global_context, local_context);		\
302 		return;									\
303 	}										\
304 	CK_CC_INLINE static void							\
305 	ck_rwcohort_neutral_##N##_read_unlock(CK_RWCOHORT_NEUTRAL_INSTANCE(N) *cohort)	\
306 	{										\
307 											\
308 		ck_pr_fence_load_atomic();						\
309 		ck_pr_dec_uint(&cohort->read_counter);					\
310 		return;									\
311 	}
312 
313 #define CK_RWCOHORT_NEUTRAL_INITIALIZER {						\
314 	.read_counter = 0,								\
315 }
316 
317 #endif /* CK_RWCOHORT_H */
318