/linux/mm/ |
H A D | slab.h | 445 SLAB_TYPESAFE_BY_RCU | SLAB_DEBUG_OBJECTS ) 602 if (s->flags & (SLAB_TYPESAFE_BY_RCU | SLAB_STORE_USER)) in slab_ksize() 639 if (c->flags & (SLAB_TYPESAFE_BY_RCU | SLAB_POISON)) in slab_want_init_on_alloc() 651 (c->flags & (SLAB_TYPESAFE_BY_RCU | SLAB_POISON))); in slab_want_init_on_free()
|
H A D | slab_common.c | 52 SLAB_TRACE | SLAB_TYPESAFE_BY_RCU | SLAB_NOLEAKTRACE | \ 461 if (s->flags & SLAB_TYPESAFE_BY_RCU) { in shutdown_cache() 490 rcu_set = s->flags & SLAB_TYPESAFE_BY_RCU; in kmem_cache_destroy()
|
H A D | slub.c | 2166 if (!(s->flags & SLAB_TYPESAFE_BY_RCU)) in slab_free_hook() 2518 if (unlikely(s->flags & SLAB_TYPESAFE_BY_RCU)) in free_slab() 5131 if ((flags & SLAB_POISON) && !(flags & SLAB_TYPESAFE_BY_RCU) && in calculate_sizes() 5154 (flags & (SLAB_TYPESAFE_BY_RCU | SLAB_POISON)) || in calculate_sizes() 6464 return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_TYPESAFE_BY_RCU)); in destroy_by_rcu_show()
|
/linux/Documentation/RCU/ |
H A D | rculist_nulls.rst | 9 objects using SLAB_TYPESAFE_BY_RCU allocations. 20 allocated with SLAB_TYPESAFE_BY_RCU kmem_cache can use the following 124 But thanks to SLAB_TYPESAFE_BY_RCU, beware a deleted object can be reused
|
H A D | whatisRCU.rst | 953 slab cache marked ``SLAB_TYPESAFE_BY_RCU``. RCU operations may yield a 962 unfortunately any spinlock in a ``SLAB_TYPESAFE_BY_RCU`` object must be 965 using ``SLAB_TYPESAFE_BY_RCU``, make proper use of a reference counter. 1176 If so, consider SLAB_TYPESAFE_BY_RCU (which was originally
|
/linux/mm/kmsan/ |
H A D | hooks.c | 60 if (s->ctor || (s->flags & SLAB_TYPESAFE_BY_RCU)) in kmsan_slab_alloc() 79 if (unlikely(s->flags & (SLAB_TYPESAFE_BY_RCU | SLAB_POISON))) in kmsan_slab_free()
|
/linux/lib/ |
H A D | test_meminit.c | 227 want_rcu ? SLAB_TYPESAFE_BY_RCU : 0, in do_kmem_cache_size() 300 c = kmem_cache_create("test_cache", size, size, SLAB_TYPESAFE_BY_RCU, in do_kmem_cache_rcu_persistent()
|
/linux/mm/kasan/ |
H A D | common.c | 187 if (!cache->ctor && !(cache->flags & SLAB_TYPESAFE_BY_RCU)) in assign_tag() 228 if (unlikely(cache->flags & SLAB_TYPESAFE_BY_RCU)) in poison_slab_object()
|
H A D | generic.c | 397 if ((cache->flags & SLAB_TYPESAFE_BY_RCU) || cache->ctor) { in kasan_cache_create()
|
H A D | kasan_test.c | 976 cache = kmem_cache_create("test_cache", size, 0, SLAB_TYPESAFE_BY_RCU, in kmem_cache_invalid_free()
|
/linux/Documentation/filesystems/ |
H A D | files.rst | 109 SLAB_TYPESAFE_BY_RCU instead of call_rcu(). It isn't sufficient anymore 121 in struct file. With SLAB_TYPESAFE_BY_RCU it is necessary that callers
|
/linux/drivers/gpu/drm/i915/ |
H A D | i915_scheduler.c | 498 SLAB_TYPESAFE_BY_RCU); in i915_scheduler_module_init()
|
H A D | i915_request.c | 2294 SLAB_TYPESAFE_BY_RCU, in i915_request_module_init() 2302 SLAB_TYPESAFE_BY_RCU); in i915_request_module_init()
|
/linux/fs/ |
H A D | file_table.c | 516 SLAB_TYPESAFE_BY_RCU | SLAB_HWCACHE_ALIGN | in files_init()
|
/linux/drivers/dma-buf/ |
H A D | st-dma-fence.c | 587 SLAB_TYPESAFE_BY_RCU | in dma_fence()
|
H A D | st-dma-fence-chain.c | 697 SLAB_TYPESAFE_BY_RCU | in dma_fence_chain()
|
/linux/include/linux/ |
H A D | slab.h | 141 #define SLAB_TYPESAFE_BY_RCU __SLAB_FLAG_BIT(_SLAB_TYPESAFE_BY_RCU) macro
|
/linux/mm/kfence/ |
H A D | kfence_test.c | 665 setup_test_cache(test, size, SLAB_TYPESAFE_BY_RCU, NULL); in test_memcache_typesafe_by_rcu()
|
H A D | core.c | 1151 if (unlikely(meta->cache && (meta->cache->flags & SLAB_TYPESAFE_BY_RCU))) in __kfence_free()
|
/linux/kernel/rcu/ |
H A D | refscale.c | 719 SLAB_TYPESAFE_BY_RCU, refscale_typesafe_ctor); in typesafe_init()
|
/linux/net/dccp/ |
H A D | ipv4.c | 980 .slab_flags = SLAB_TYPESAFE_BY_RCU,
|
H A D | ipv6.c | 1062 .slab_flags = SLAB_TYPESAFE_BY_RCU,
|
/linux/net/llc/ |
H A D | af_llc.c | 157 .slab_flags = SLAB_TYPESAFE_BY_RCU,
|
/linux/block/ |
H A D | bio.c | 95 SLAB_HWCACHE_ALIGN | SLAB_TYPESAFE_BY_RCU, NULL); in create_bio_slab()
|
/linux/net/smc/ |
H A D | af_smc.c | 229 .slab_flags = SLAB_TYPESAFE_BY_RCU, 242 .slab_flags = SLAB_TYPESAFE_BY_RCU,
|