xref: /freebsd/sys/contrib/ck/src/ck_hs.c (revision 74e9b5f2)
1 /*
2  * Copyright 2012-2015 Samy Al Bahra.
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  */
26 
27 #include <ck_cc.h>
28 #include <ck_hs.h>
29 #include <ck_limits.h>
30 #include <ck_md.h>
31 #include <ck_pr.h>
32 #include <ck_stdint.h>
33 #include <ck_stdbool.h>
34 #include <ck_string.h>
35 
36 #include "ck_internal.h"
37 
38 #ifndef CK_HS_PROBE_L1_SHIFT
39 #define CK_HS_PROBE_L1_SHIFT 3ULL
40 #endif /* CK_HS_PROBE_L1_SHIFT */
41 
42 #define CK_HS_PROBE_L1 (1 << CK_HS_PROBE_L1_SHIFT)
43 #define CK_HS_PROBE_L1_MASK (CK_HS_PROBE_L1 - 1)
44 
45 #ifndef CK_HS_PROBE_L1_DEFAULT
46 #define CK_HS_PROBE_L1_DEFAULT CK_MD_CACHELINE
47 #endif
48 
49 #define CK_HS_VMA_MASK ((uintptr_t)((1ULL << CK_MD_VMA_BITS) - 1))
50 #define CK_HS_VMA(x)	\
51 	((void *)((uintptr_t)(x) & CK_HS_VMA_MASK))
52 
53 #define CK_HS_EMPTY     NULL
54 #define CK_HS_TOMBSTONE ((void *)~(uintptr_t)0)
55 #define CK_HS_G		(2)
56 #define CK_HS_G_MASK	(CK_HS_G - 1)
57 
58 #if defined(CK_F_PR_LOAD_8) && defined(CK_F_PR_STORE_8)
59 #define CK_HS_WORD          uint8_t
60 #define CK_HS_WORD_MAX	    UINT8_MAX
61 #define CK_HS_STORE(x, y)   ck_pr_store_8(x, y)
62 #define CK_HS_LOAD(x)       ck_pr_load_8(x)
63 #elif defined(CK_F_PR_LOAD_16) && defined(CK_F_PR_STORE_16)
64 #define CK_HS_WORD          uint16_t
65 #define CK_HS_WORD_MAX	    UINT16_MAX
66 #define CK_HS_STORE(x, y)   ck_pr_store_16(x, y)
67 #define CK_HS_LOAD(x)       ck_pr_load_16(x)
68 #elif defined(CK_F_PR_LOAD_32) && defined(CK_F_PR_STORE_32)
69 #define CK_HS_WORD          uint32_t
70 #define CK_HS_WORD_MAX	    UINT32_MAX
71 #define CK_HS_STORE(x, y)   ck_pr_store_32(x, y)
72 #define CK_HS_LOAD(x)       ck_pr_load_32(x)
73 #else
74 #error "ck_hs is not supported on your platform."
75 #endif
76 
77 enum ck_hs_probe_behavior {
78 	CK_HS_PROBE = 0,	/* Default behavior. */
79 	CK_HS_PROBE_TOMBSTONE,	/* Short-circuit on tombstone. */
80 	CK_HS_PROBE_INSERT	/* Short-circuit on probe bound if tombstone found. */
81 };
82 
83 struct ck_hs_map {
84 	unsigned int generation[CK_HS_G];
85 	unsigned int probe_maximum;
86 	unsigned long mask;
87 	unsigned long step;
88 	unsigned int probe_limit;
89 	unsigned int tombstones;
90 	unsigned long n_entries;
91 	unsigned long capacity;
92 	unsigned long size;
93 	CK_HS_WORD *probe_bound;
94 	const void **entries;
95 };
96 
97 static inline void
ck_hs_map_signal(struct ck_hs_map * map,unsigned long h)98 ck_hs_map_signal(struct ck_hs_map *map, unsigned long h)
99 {
100 
101 	h &= CK_HS_G_MASK;
102 	ck_pr_store_uint(&map->generation[h],
103 	    map->generation[h] + 1);
104 	ck_pr_fence_store();
105 	return;
106 }
107 
108 static bool
_ck_hs_next(struct ck_hs * hs,struct ck_hs_map * map,struct ck_hs_iterator * i,void ** key)109 _ck_hs_next(struct ck_hs *hs, struct ck_hs_map *map,
110     struct ck_hs_iterator *i, void **key)
111 {
112 	void *value;
113 
114 	if (i->offset >= map->capacity)
115 		return false;
116 
117 	do {
118 		value = CK_CC_DECONST_PTR(map->entries[i->offset]);
119 		if (value != CK_HS_EMPTY && value != CK_HS_TOMBSTONE) {
120 #ifdef CK_HS_PP
121 			if (hs->mode & CK_HS_MODE_OBJECT)
122 				value = CK_HS_VMA(value);
123 #else
124 			(void)hs; /* Avoid unused parameter warning. */
125 #endif
126 			i->offset++;
127 			*key = value;
128 			return true;
129 		}
130 	} while (++i->offset < map->capacity);
131 
132 	return false;
133 }
134 
135 void
ck_hs_iterator_init(struct ck_hs_iterator * iterator)136 ck_hs_iterator_init(struct ck_hs_iterator *iterator)
137 {
138 
139 	iterator->cursor = NULL;
140 	iterator->offset = 0;
141 	iterator->map = NULL;
142 	return;
143 }
144 
145 bool
ck_hs_next(struct ck_hs * hs,struct ck_hs_iterator * i,void ** key)146 ck_hs_next(struct ck_hs *hs, struct ck_hs_iterator *i, void **key)
147 {
148 
149 	return _ck_hs_next(hs, hs->map, i, key);
150 }
151 
152 bool
ck_hs_next_spmc(struct ck_hs * hs,struct ck_hs_iterator * i,void ** key)153 ck_hs_next_spmc(struct ck_hs *hs, struct ck_hs_iterator *i, void **key)
154 {
155 	struct ck_hs_map *m = i->map;
156 
157 	if (m == NULL) {
158 		m = i->map = ck_pr_load_ptr(&hs->map);
159 	}
160 
161 	return _ck_hs_next(hs, m, i, key);
162 }
163 
164 void
ck_hs_stat(struct ck_hs * hs,struct ck_hs_stat * st)165 ck_hs_stat(struct ck_hs *hs, struct ck_hs_stat *st)
166 {
167 	struct ck_hs_map *map = hs->map;
168 
169 	st->n_entries = map->n_entries;
170 	st->tombstones = map->tombstones;
171 	st->probe_maximum = map->probe_maximum;
172 	return;
173 }
174 
175 unsigned long
ck_hs_count(struct ck_hs * hs)176 ck_hs_count(struct ck_hs *hs)
177 {
178 
179 	return hs->map->n_entries;
180 }
181 
182 static void
ck_hs_map_destroy(struct ck_malloc * m,struct ck_hs_map * map,bool defer)183 ck_hs_map_destroy(struct ck_malloc *m, struct ck_hs_map *map, bool defer)
184 {
185 
186 	m->free(map, map->size, defer);
187 	return;
188 }
189 
190 void
ck_hs_destroy(struct ck_hs * hs)191 ck_hs_destroy(struct ck_hs *hs)
192 {
193 
194 	ck_hs_map_destroy(hs->m, hs->map, false);
195 	return;
196 }
197 
198 static struct ck_hs_map *
ck_hs_map_create(struct ck_hs * hs,unsigned long entries)199 ck_hs_map_create(struct ck_hs *hs, unsigned long entries)
200 {
201 	struct ck_hs_map *map;
202 	unsigned long size, n_entries, prefix, limit;
203 
204 	n_entries = ck_internal_power_2(entries);
205 	if (n_entries < CK_HS_PROBE_L1)
206 		n_entries = CK_HS_PROBE_L1;
207 
208 	size = sizeof(struct ck_hs_map) + (sizeof(void *) * n_entries + CK_MD_CACHELINE - 1);
209 
210 	if (hs->mode & CK_HS_MODE_DELETE) {
211 		prefix = sizeof(CK_HS_WORD) * n_entries;
212 		size += prefix;
213 	} else {
214 		prefix = 0;
215 	}
216 
217 	map = hs->m->malloc(size);
218 	if (map == NULL)
219 		return NULL;
220 
221 	map->size = size;
222 
223 	/* We should probably use a more intelligent heuristic for default probe length. */
224 	limit = ck_internal_max(n_entries >> (CK_HS_PROBE_L1_SHIFT + 2), CK_HS_PROBE_L1_DEFAULT);
225 	if (limit > UINT_MAX)
226 		limit = UINT_MAX;
227 
228 	map->probe_limit = (unsigned int)limit;
229 	map->probe_maximum = 0;
230 	map->capacity = n_entries;
231 	map->step = ck_cc_ffsl(n_entries);
232 	map->mask = n_entries - 1;
233 	map->n_entries = 0;
234 
235 	/* Align map allocation to cache line. */
236 	map->entries = (void *)(((uintptr_t)&map[1] + prefix +
237 	    CK_MD_CACHELINE - 1) & ~(CK_MD_CACHELINE - 1));
238 
239 	memset(map->entries, 0, sizeof(void *) * n_entries);
240 	memset(map->generation, 0, sizeof map->generation);
241 
242 	if (hs->mode & CK_HS_MODE_DELETE) {
243 		map->probe_bound = (CK_HS_WORD *)&map[1];
244 		memset(map->probe_bound, 0, prefix);
245 	} else {
246 		map->probe_bound = NULL;
247 	}
248 
249 	/* Commit entries purge with respect to map publication. */
250 	ck_pr_fence_store();
251 	return map;
252 }
253 
254 bool
ck_hs_reset_size(struct ck_hs * hs,unsigned long capacity)255 ck_hs_reset_size(struct ck_hs *hs, unsigned long capacity)
256 {
257 	struct ck_hs_map *map, *previous;
258 
259 	previous = hs->map;
260 	map = ck_hs_map_create(hs, capacity);
261 	if (map == NULL)
262 		return false;
263 
264 	ck_pr_store_ptr(&hs->map, map);
265 	ck_hs_map_destroy(hs->m, previous, true);
266 	return true;
267 }
268 
269 bool
ck_hs_reset(struct ck_hs * hs)270 ck_hs_reset(struct ck_hs *hs)
271 {
272 	struct ck_hs_map *previous;
273 
274 	previous = hs->map;
275 	return ck_hs_reset_size(hs, previous->capacity);
276 }
277 
278 static inline unsigned long
ck_hs_map_probe_next(struct ck_hs_map * map,unsigned long offset,unsigned long h,unsigned long level,unsigned long probes)279 ck_hs_map_probe_next(struct ck_hs_map *map,
280     unsigned long offset,
281     unsigned long h,
282     unsigned long level,
283     unsigned long probes)
284 {
285 	unsigned long r, stride;
286 
287 	r = (h >> map->step) >> level;
288 	stride = (r & ~CK_HS_PROBE_L1_MASK) << 1 | (r & CK_HS_PROBE_L1_MASK);
289 
290 	return (offset + (probes >> CK_HS_PROBE_L1_SHIFT) +
291 	    (stride | CK_HS_PROBE_L1)) & map->mask;
292 }
293 
294 static inline void
ck_hs_map_bound_set(struct ck_hs_map * m,unsigned long h,unsigned long n_probes)295 ck_hs_map_bound_set(struct ck_hs_map *m,
296     unsigned long h,
297     unsigned long n_probes)
298 {
299 	unsigned long offset = h & m->mask;
300 
301 	if (n_probes > m->probe_maximum)
302 		ck_pr_store_uint(&m->probe_maximum, n_probes);
303 
304 	if (m->probe_bound != NULL && m->probe_bound[offset] < n_probes) {
305 		if (n_probes > CK_HS_WORD_MAX)
306 			n_probes = CK_HS_WORD_MAX;
307 
308 		CK_HS_STORE(&m->probe_bound[offset], n_probes);
309 		ck_pr_fence_store();
310 	}
311 
312 	return;
313 }
314 
315 static inline unsigned int
ck_hs_map_bound_get(struct ck_hs_map * m,unsigned long h)316 ck_hs_map_bound_get(struct ck_hs_map *m, unsigned long h)
317 {
318 	unsigned long offset = h & m->mask;
319 	unsigned int r = CK_HS_WORD_MAX;
320 
321 	if (m->probe_bound != NULL) {
322 		r = CK_HS_LOAD(&m->probe_bound[offset]);
323 		if (r == CK_HS_WORD_MAX)
324 			r = ck_pr_load_uint(&m->probe_maximum);
325 	} else {
326 		r = ck_pr_load_uint(&m->probe_maximum);
327 	}
328 
329 	return r;
330 }
331 
332 bool
ck_hs_grow(struct ck_hs * hs,unsigned long capacity)333 ck_hs_grow(struct ck_hs *hs,
334     unsigned long capacity)
335 {
336 	struct ck_hs_map *map, *update;
337 	unsigned long k, i, j, offset, probes;
338 	const void *previous, **bucket;
339 
340 restart:
341 	map = hs->map;
342 	if (map->capacity > capacity)
343 		return false;
344 
345 	update = ck_hs_map_create(hs, capacity);
346 	if (update == NULL)
347 		return false;
348 
349 	for (k = 0; k < map->capacity; k++) {
350 		unsigned long h;
351 
352 		previous = map->entries[k];
353 		if (previous == CK_HS_EMPTY || previous == CK_HS_TOMBSTONE)
354 			continue;
355 
356 #ifdef CK_HS_PP
357 		if (hs->mode & CK_HS_MODE_OBJECT)
358 			previous = CK_HS_VMA(previous);
359 #endif
360 
361 		h = hs->hf(previous, hs->seed);
362 		offset = h & update->mask;
363 		i = probes = 0;
364 
365 		for (;;) {
366 			bucket = (const void **)((uintptr_t)&update->entries[offset] & ~(CK_MD_CACHELINE - 1));
367 
368 			for (j = 0; j < CK_HS_PROBE_L1; j++) {
369 				const void **cursor = bucket + ((j + offset) & (CK_HS_PROBE_L1 - 1));
370 
371 				if (probes++ == update->probe_limit)
372 					break;
373 
374 				if (CK_CC_LIKELY(*cursor == CK_HS_EMPTY)) {
375 					*cursor = map->entries[k];
376 					update->n_entries++;
377 
378 					ck_hs_map_bound_set(update, h, probes);
379 					break;
380 				}
381 			}
382 
383 			if (j < CK_HS_PROBE_L1)
384 				break;
385 
386 			offset = ck_hs_map_probe_next(update, offset, h, i++, probes);
387 		}
388 
389 		if (probes > update->probe_limit) {
390 			/*
391 			 * We have hit the probe limit, map needs to be even larger.
392 			 */
393 			ck_hs_map_destroy(hs->m, update, false);
394 			capacity <<= 1;
395 			goto restart;
396 		}
397 	}
398 
399 	ck_pr_fence_store();
400 	ck_pr_store_ptr(&hs->map, update);
401 	ck_hs_map_destroy(hs->m, map, true);
402 	return true;
403 }
404 
405 static void
ck_hs_map_postinsert(struct ck_hs * hs,struct ck_hs_map * map)406 ck_hs_map_postinsert(struct ck_hs *hs, struct ck_hs_map *map)
407 {
408 
409 	map->n_entries++;
410 	if ((map->n_entries << 1) > map->capacity)
411 		ck_hs_grow(hs, map->capacity << 1);
412 
413 	return;
414 }
415 
416 bool
ck_hs_rebuild(struct ck_hs * hs)417 ck_hs_rebuild(struct ck_hs *hs)
418 {
419 
420 	return ck_hs_grow(hs, hs->map->capacity);
421 }
422 
423 static const void **
ck_hs_map_probe(struct ck_hs * hs,struct ck_hs_map * map,unsigned long * n_probes,const void *** priority,unsigned long h,const void * key,const void ** object,unsigned long probe_limit,enum ck_hs_probe_behavior behavior)424 ck_hs_map_probe(struct ck_hs *hs,
425     struct ck_hs_map *map,
426     unsigned long *n_probes,
427     const void ***priority,
428     unsigned long h,
429     const void *key,
430     const void **object,
431     unsigned long probe_limit,
432     enum ck_hs_probe_behavior behavior)
433 {
434 	const void **bucket, **cursor, *k, *compare;
435 	const void **pr = NULL;
436 	unsigned long offset, j, i, probes, opl;
437 
438 #ifdef CK_HS_PP
439 	/* If we are storing object pointers, then we may leverage pointer packing. */
440 	unsigned long hv = 0;
441 
442 	if (hs->mode & CK_HS_MODE_OBJECT) {
443 		hv = (h >> 25) & CK_HS_KEY_MASK;
444 		compare = CK_HS_VMA(key);
445 	} else {
446 		compare = key;
447 	}
448 #else
449 	compare = key;
450 #endif
451 
452 	offset = h & map->mask;
453 	*object = NULL;
454 	i = probes = 0;
455 
456 	opl = probe_limit;
457 	if (behavior == CK_HS_PROBE_INSERT)
458 		probe_limit = ck_hs_map_bound_get(map, h);
459 
460 	for (;;) {
461 		bucket = (const void **)((uintptr_t)&map->entries[offset] & ~(CK_MD_CACHELINE - 1));
462 
463 		for (j = 0; j < CK_HS_PROBE_L1; j++) {
464 			cursor = bucket + ((j + offset) & (CK_HS_PROBE_L1 - 1));
465 
466 			if (probes++ == probe_limit) {
467 				if (probe_limit == opl || pr != NULL) {
468 					k = CK_HS_EMPTY;
469 					goto leave;
470 				}
471 
472 				/*
473 				 * If no eligible slot has been found yet, continue probe
474 				 * sequence with original probe limit.
475 				 */
476 				probe_limit = opl;
477 			}
478 
479 			k = ck_pr_load_ptr(cursor);
480 			if (k == CK_HS_EMPTY)
481 				goto leave;
482 
483 			if (k == CK_HS_TOMBSTONE) {
484 				if (pr == NULL) {
485 					pr = cursor;
486 					*n_probes = probes;
487 
488 					if (behavior == CK_HS_PROBE_TOMBSTONE) {
489 						k = CK_HS_EMPTY;
490 						goto leave;
491 					}
492 				}
493 
494 				continue;
495 			}
496 
497 #ifdef CK_HS_PP
498 			if (hs->mode & CK_HS_MODE_OBJECT) {
499 				if (((uintptr_t)k >> CK_MD_VMA_BITS) != hv)
500 					continue;
501 
502 				k = CK_HS_VMA(k);
503 			}
504 #endif
505 
506 			if (k == compare)
507 				goto leave;
508 
509 			if (hs->compare == NULL)
510 				continue;
511 
512 			if (hs->compare(k, key) == true)
513 				goto leave;
514 		}
515 
516 		offset = ck_hs_map_probe_next(map, offset, h, i++, probes);
517 	}
518 
519 leave:
520 	if (probes > probe_limit) {
521 		cursor = NULL;
522 	} else {
523 		*object = k;
524 	}
525 
526 	if (pr == NULL)
527 		*n_probes = probes;
528 
529 	*priority = pr;
530 	return cursor;
531 }
532 
533 static inline const void *
ck_hs_marshal(unsigned int mode,const void * key,unsigned long h)534 ck_hs_marshal(unsigned int mode, const void *key, unsigned long h)
535 {
536 #ifdef CK_HS_PP
537 	const void *insert;
538 
539 	if (mode & CK_HS_MODE_OBJECT) {
540 		insert = (void *)((uintptr_t)CK_HS_VMA(key) |
541 		    ((h >> 25) << CK_MD_VMA_BITS));
542 	} else {
543 		insert = key;
544 	}
545 
546 	return insert;
547 #else
548 	(void)mode;
549 	(void)h;
550 
551 	return key;
552 #endif
553 }
554 
555 bool
ck_hs_gc(struct ck_hs * hs,unsigned long cycles,unsigned long seed)556 ck_hs_gc(struct ck_hs *hs, unsigned long cycles, unsigned long seed)
557 {
558 	unsigned long size = 0;
559 	unsigned long i;
560 	struct ck_hs_map *map = hs->map;
561 	unsigned int maximum;
562 	CK_HS_WORD *bounds = NULL;
563 
564 	if (map->n_entries == 0) {
565 		ck_pr_store_uint(&map->probe_maximum, 0);
566 		if (map->probe_bound != NULL)
567 			memset(map->probe_bound, 0, sizeof(CK_HS_WORD) * map->capacity);
568 
569 		return true;
570 	}
571 
572 	if (cycles == 0) {
573 		maximum = 0;
574 
575 		if (map->probe_bound != NULL) {
576 			size = sizeof(CK_HS_WORD) * map->capacity;
577 			bounds = hs->m->malloc(size);
578 			if (bounds == NULL)
579 				return false;
580 
581 			memset(bounds, 0, size);
582 		}
583 	} else {
584 		maximum = map->probe_maximum;
585 	}
586 
587 	for (i = 0; i < map->capacity; i++) {
588 		const void **first, *object, **slot, *entry;
589 		unsigned long n_probes, offset, h;
590 
591 		entry = map->entries[(i + seed) & map->mask];
592 		if (entry == CK_HS_EMPTY || entry == CK_HS_TOMBSTONE)
593 			continue;
594 
595 #ifdef CK_HS_PP
596 		if (hs->mode & CK_HS_MODE_OBJECT)
597 			entry = CK_HS_VMA(entry);
598 #endif
599 
600 		h = hs->hf(entry, hs->seed);
601 		offset = h & map->mask;
602 
603 		slot = ck_hs_map_probe(hs, map, &n_probes, &first, h, entry, &object,
604 		    ck_hs_map_bound_get(map, h), CK_HS_PROBE);
605 
606 		if (first != NULL) {
607 			const void *insert = ck_hs_marshal(hs->mode, entry, h);
608 
609 			ck_pr_store_ptr(first, insert);
610 			ck_hs_map_signal(map, h);
611 			ck_pr_store_ptr(slot, CK_HS_TOMBSTONE);
612 		}
613 
614 		if (cycles == 0) {
615 			if (n_probes > maximum)
616 				maximum = n_probes;
617 
618 			if (n_probes > CK_HS_WORD_MAX)
619 				n_probes = CK_HS_WORD_MAX;
620 
621 			if (bounds != NULL && n_probes > bounds[offset])
622 				bounds[offset] = n_probes;
623 		} else if (--cycles == 0)
624 			break;
625 	}
626 
627 	/*
628 	 * The following only apply to garbage collection involving
629 	 * a full scan of all entries.
630 	 */
631 	if (maximum != map->probe_maximum)
632 		ck_pr_store_uint(&map->probe_maximum, maximum);
633 
634 	if (bounds != NULL) {
635 		for (i = 0; i < map->capacity; i++)
636 			CK_HS_STORE(&map->probe_bound[i], bounds[i]);
637 
638 		hs->m->free(bounds, size, false);
639 	}
640 
641 	return true;
642 }
643 
644 bool
ck_hs_fas(struct ck_hs * hs,unsigned long h,const void * key,void ** previous)645 ck_hs_fas(struct ck_hs *hs,
646     unsigned long h,
647     const void *key,
648     void **previous)
649 {
650 	const void **slot, **first, *object, *insert;
651 	struct ck_hs_map *map = hs->map;
652 	unsigned long n_probes;
653 
654 	*previous = NULL;
655 	slot = ck_hs_map_probe(hs, map, &n_probes, &first, h, key, &object,
656 	    ck_hs_map_bound_get(map, h), CK_HS_PROBE);
657 
658 	/* Replacement semantics presume existence. */
659 	if (object == NULL)
660 		return false;
661 
662 	insert = ck_hs_marshal(hs->mode, key, h);
663 
664 	if (first != NULL) {
665 		ck_pr_store_ptr(first, insert);
666 		ck_hs_map_signal(map, h);
667 		ck_pr_store_ptr(slot, CK_HS_TOMBSTONE);
668 	} else {
669 		ck_pr_store_ptr(slot, insert);
670 	}
671 
672 	*previous = CK_CC_DECONST_PTR(object);
673 	return true;
674 }
675 
676 /*
677  * An apply function takes two arguments. The first argument is a pointer to a
678  * pre-existing object. The second argument is a pointer to the fifth argument
679  * passed to ck_hs_apply. If a non-NULL pointer is passed to the first argument
680  * and the return value of the apply function is NULL, then the pre-existing
681  * value is deleted. If the return pointer is the same as the one passed to the
682  * apply function then no changes are made to the hash table.  If the first
683  * argument is non-NULL and the return pointer is different than that passed to
684  * the apply function, then the pre-existing value is replaced. For
685  * replacement, it is required that the value itself is identical to the
686  * previous value.
687  */
688 bool
ck_hs_apply(struct ck_hs * hs,unsigned long h,const void * key,ck_hs_apply_fn_t * fn,void * cl)689 ck_hs_apply(struct ck_hs *hs,
690     unsigned long h,
691     const void *key,
692     ck_hs_apply_fn_t *fn,
693     void *cl)
694 {
695 	const void **slot, **first, *object, *delta, *insert;
696 	unsigned long n_probes;
697 	struct ck_hs_map *map;
698 
699 restart:
700 	map = hs->map;
701 
702 	slot = ck_hs_map_probe(hs, map, &n_probes, &first, h, key, &object, map->probe_limit, CK_HS_PROBE_INSERT);
703 	if (slot == NULL && first == NULL) {
704 		if (ck_hs_grow(hs, map->capacity << 1) == false)
705 			return false;
706 
707 		goto restart;
708 	}
709 
710 	delta = fn(CK_CC_DECONST_PTR(object), cl);
711 	if (delta == NULL) {
712 		/*
713 		 * The apply function has requested deletion. If the object doesn't exist,
714 		 * then exit early.
715 		 */
716 		if (CK_CC_UNLIKELY(object == NULL))
717 			return true;
718 
719 		/* Otherwise, mark slot as deleted. */
720 		ck_pr_store_ptr(slot, CK_HS_TOMBSTONE);
721 		map->n_entries--;
722 		map->tombstones++;
723 		return true;
724 	}
725 
726 	/* The apply function has not requested hash set modification so exit early. */
727 	if (delta == object)
728 		return true;
729 
730 	/* A modification or insertion has been requested. */
731 	ck_hs_map_bound_set(map, h, n_probes);
732 
733 	insert = ck_hs_marshal(hs->mode, delta, h);
734 	if (first != NULL) {
735 		/*
736 		 * This follows the same semantics as ck_hs_set, please refer to that
737 		 * function for documentation.
738 		 */
739 		ck_pr_store_ptr(first, insert);
740 
741 		if (object != NULL) {
742 			ck_hs_map_signal(map, h);
743 			ck_pr_store_ptr(slot, CK_HS_TOMBSTONE);
744 		}
745 	} else {
746 		/*
747 		 * If we are storing into same slot, then atomic store is sufficient
748 		 * for replacement.
749 		 */
750 		ck_pr_store_ptr(slot, insert);
751 	}
752 
753 	if (object == NULL)
754 		ck_hs_map_postinsert(hs, map);
755 
756 	return true;
757 }
758 
759 bool
ck_hs_set(struct ck_hs * hs,unsigned long h,const void * key,void ** previous)760 ck_hs_set(struct ck_hs *hs,
761     unsigned long h,
762     const void *key,
763     void **previous)
764 {
765 	const void **slot, **first, *object, *insert;
766 	unsigned long n_probes;
767 	struct ck_hs_map *map;
768 
769 	*previous = NULL;
770 
771 restart:
772 	map = hs->map;
773 
774 	slot = ck_hs_map_probe(hs, map, &n_probes, &first, h, key, &object, map->probe_limit, CK_HS_PROBE_INSERT);
775 	if (slot == NULL && first == NULL) {
776 		if (ck_hs_grow(hs, map->capacity << 1) == false)
777 			return false;
778 
779 		goto restart;
780 	}
781 
782 	ck_hs_map_bound_set(map, h, n_probes);
783 	insert = ck_hs_marshal(hs->mode, key, h);
784 
785 	if (first != NULL) {
786 		/* If an earlier bucket was found, then store entry there. */
787 		ck_pr_store_ptr(first, insert);
788 
789 		/*
790 		 * If a duplicate key was found, then delete it after
791 		 * signaling concurrent probes to restart. Optionally,
792 		 * it is possible to install tombstone after grace
793 		 * period if we can guarantee earlier position of
794 		 * duplicate key.
795 		 */
796 		if (object != NULL) {
797 			ck_hs_map_signal(map, h);
798 			ck_pr_store_ptr(slot, CK_HS_TOMBSTONE);
799 		}
800 	} else {
801 		/*
802 		 * If we are storing into same slot, then atomic store is sufficient
803 		 * for replacement.
804 		 */
805 		ck_pr_store_ptr(slot, insert);
806 	}
807 
808 	if (object == NULL)
809 		ck_hs_map_postinsert(hs, map);
810 
811 	*previous = CK_CC_DECONST_PTR(object);
812 	return true;
813 }
814 
815 CK_CC_INLINE static bool
ck_hs_put_internal(struct ck_hs * hs,unsigned long h,const void * key,enum ck_hs_probe_behavior behavior)816 ck_hs_put_internal(struct ck_hs *hs,
817     unsigned long h,
818     const void *key,
819     enum ck_hs_probe_behavior behavior)
820 {
821 	const void **slot, **first, *object, *insert;
822 	unsigned long n_probes;
823 	struct ck_hs_map *map;
824 
825 restart:
826 	map = hs->map;
827 
828 	slot = ck_hs_map_probe(hs, map, &n_probes, &first, h, key, &object,
829 	    map->probe_limit, behavior);
830 
831 	if (slot == NULL && first == NULL) {
832 		if (ck_hs_grow(hs, map->capacity << 1) == false)
833 			return false;
834 
835 		goto restart;
836 	}
837 
838 	/* Fail operation if a match was found. */
839 	if (object != NULL)
840 		return false;
841 
842 	ck_hs_map_bound_set(map, h, n_probes);
843 	insert = ck_hs_marshal(hs->mode, key, h);
844 
845 	if (first != NULL) {
846 		/* Insert key into first bucket in probe sequence. */
847 		ck_pr_store_ptr(first, insert);
848 	} else {
849 		/* An empty slot was found. */
850 		ck_pr_store_ptr(slot, insert);
851 	}
852 
853 	ck_hs_map_postinsert(hs, map);
854 	return true;
855 }
856 
857 bool
ck_hs_put(struct ck_hs * hs,unsigned long h,const void * key)858 ck_hs_put(struct ck_hs *hs,
859     unsigned long h,
860     const void *key)
861 {
862 
863 	return ck_hs_put_internal(hs, h, key, CK_HS_PROBE_INSERT);
864 }
865 
866 bool
ck_hs_put_unique(struct ck_hs * hs,unsigned long h,const void * key)867 ck_hs_put_unique(struct ck_hs *hs,
868     unsigned long h,
869     const void *key)
870 {
871 
872 	return ck_hs_put_internal(hs, h, key, CK_HS_PROBE_TOMBSTONE);
873 }
874 
875 void *
ck_hs_get(struct ck_hs * hs,unsigned long h,const void * key)876 ck_hs_get(struct ck_hs *hs,
877     unsigned long h,
878     const void *key)
879 {
880 	const void **first, *object;
881 	struct ck_hs_map *map;
882 	unsigned long n_probes;
883 	unsigned int g, g_p, probe;
884 	unsigned int *generation;
885 
886 	do {
887 		map = ck_pr_load_ptr(&hs->map);
888 		generation = &map->generation[h & CK_HS_G_MASK];
889 		g = ck_pr_load_uint(generation);
890 		probe  = ck_hs_map_bound_get(map, h);
891 		ck_pr_fence_load();
892 
893 		ck_hs_map_probe(hs, map, &n_probes, &first, h, key, &object, probe, CK_HS_PROBE);
894 
895 		ck_pr_fence_load();
896 		g_p = ck_pr_load_uint(generation);
897 	} while (g != g_p);
898 
899 	return CK_CC_DECONST_PTR(object);
900 }
901 
902 void *
ck_hs_remove(struct ck_hs * hs,unsigned long h,const void * key)903 ck_hs_remove(struct ck_hs *hs,
904     unsigned long h,
905     const void *key)
906 {
907 	const void **slot, **first, *object;
908 	struct ck_hs_map *map = hs->map;
909 	unsigned long n_probes;
910 
911 	slot = ck_hs_map_probe(hs, map, &n_probes, &first, h, key, &object,
912 	    ck_hs_map_bound_get(map, h), CK_HS_PROBE);
913 	if (object == NULL)
914 		return NULL;
915 
916 	ck_pr_store_ptr(slot, CK_HS_TOMBSTONE);
917 	map->n_entries--;
918 	map->tombstones++;
919 	return CK_CC_DECONST_PTR(object);
920 }
921 
922 bool
ck_hs_move(struct ck_hs * hs,struct ck_hs * source,ck_hs_hash_cb_t * hf,ck_hs_compare_cb_t * compare,struct ck_malloc * m)923 ck_hs_move(struct ck_hs *hs,
924     struct ck_hs *source,
925     ck_hs_hash_cb_t *hf,
926     ck_hs_compare_cb_t *compare,
927     struct ck_malloc *m)
928 {
929 
930 	if (m == NULL || m->malloc == NULL || m->free == NULL || hf == NULL)
931 		return false;
932 
933 	hs->mode = source->mode;
934 	hs->seed = source->seed;
935 	hs->map = source->map;
936 	hs->m = m;
937 	hs->hf = hf;
938 	hs->compare = compare;
939 	return true;
940 }
941 
942 bool
ck_hs_init(struct ck_hs * hs,unsigned int mode,ck_hs_hash_cb_t * hf,ck_hs_compare_cb_t * compare,struct ck_malloc * m,unsigned long n_entries,unsigned long seed)943 ck_hs_init(struct ck_hs *hs,
944     unsigned int mode,
945     ck_hs_hash_cb_t *hf,
946     ck_hs_compare_cb_t *compare,
947     struct ck_malloc *m,
948     unsigned long n_entries,
949     unsigned long seed)
950 {
951 
952 	if (m == NULL || m->malloc == NULL || m->free == NULL || hf == NULL)
953 		return false;
954 
955 	hs->m = m;
956 	hs->mode = mode;
957 	hs->seed = seed;
958 	hs->hf = hf;
959 	hs->compare = compare;
960 
961 	hs->map = ck_hs_map_create(hs, n_entries);
962 	return hs->map != NULL;
963 }
964