xref: /linux/kernel/rcu/refscale.c (revision d6fea1dd)
11fbeb3a8SPaul E. McKenney // SPDX-License-Identifier: GPL-2.0+
21fbeb3a8SPaul E. McKenney //
31fbeb3a8SPaul E. McKenney // Scalability test comparing RCU vs other mechanisms
41fbeb3a8SPaul E. McKenney // for acquiring references on objects.
51fbeb3a8SPaul E. McKenney //
61fbeb3a8SPaul E. McKenney // Copyright (C) Google, 2020.
71fbeb3a8SPaul E. McKenney //
81fbeb3a8SPaul E. McKenney // Author: Joel Fernandes <joel@joelfernandes.org>
91fbeb3a8SPaul E. McKenney 
101fbeb3a8SPaul E. McKenney #define pr_fmt(fmt) fmt
111fbeb3a8SPaul E. McKenney 
121fbeb3a8SPaul E. McKenney #include <linux/atomic.h>
131fbeb3a8SPaul E. McKenney #include <linux/bitops.h>
141fbeb3a8SPaul E. McKenney #include <linux/completion.h>
151fbeb3a8SPaul E. McKenney #include <linux/cpu.h>
161fbeb3a8SPaul E. McKenney #include <linux/delay.h>
171fbeb3a8SPaul E. McKenney #include <linux/err.h>
181fbeb3a8SPaul E. McKenney #include <linux/init.h>
191fbeb3a8SPaul E. McKenney #include <linux/interrupt.h>
201fbeb3a8SPaul E. McKenney #include <linux/kthread.h>
211fbeb3a8SPaul E. McKenney #include <linux/kernel.h>
221fbeb3a8SPaul E. McKenney #include <linux/mm.h>
231fbeb3a8SPaul E. McKenney #include <linux/module.h>
241fbeb3a8SPaul E. McKenney #include <linux/moduleparam.h>
251fbeb3a8SPaul E. McKenney #include <linux/notifier.h>
261fbeb3a8SPaul E. McKenney #include <linux/percpu.h>
271fbeb3a8SPaul E. McKenney #include <linux/rcupdate.h>
281fbeb3a8SPaul E. McKenney #include <linux/rcupdate_trace.h>
291fbeb3a8SPaul E. McKenney #include <linux/reboot.h>
301fbeb3a8SPaul E. McKenney #include <linux/sched.h>
311fbeb3a8SPaul E. McKenney #include <linux/spinlock.h>
321fbeb3a8SPaul E. McKenney #include <linux/smp.h>
331fbeb3a8SPaul E. McKenney #include <linux/stat.h>
341fbeb3a8SPaul E. McKenney #include <linux/srcu.h>
351fbeb3a8SPaul E. McKenney #include <linux/slab.h>
361fbeb3a8SPaul E. McKenney #include <linux/torture.h>
371fbeb3a8SPaul E. McKenney #include <linux/types.h>
381fbeb3a8SPaul E. McKenney 
391fbeb3a8SPaul E. McKenney #include "rcu.h"
401fbeb3a8SPaul E. McKenney 
411fbeb3a8SPaul E. McKenney #define SCALE_FLAG "-ref-scale: "
421fbeb3a8SPaul E. McKenney 
431fbeb3a8SPaul E. McKenney #define SCALEOUT(s, x...) \
441fbeb3a8SPaul E. McKenney 	pr_alert("%s" SCALE_FLAG s, scale_type, ## x)
451fbeb3a8SPaul E. McKenney 
461fbeb3a8SPaul E. McKenney #define VERBOSE_SCALEOUT(s, x...) \
47f71f22b6SLi Zhijian 	do { \
48f71f22b6SLi Zhijian 		if (verbose) \
49f71f22b6SLi Zhijian 			pr_alert("%s" SCALE_FLAG s "\n", scale_type, ## x); \
50f71f22b6SLi Zhijian 	} while (0)
511fbeb3a8SPaul E. McKenney 
52e76506f0SPaul E. McKenney static atomic_t verbose_batch_ctr;
53e76506f0SPaul E. McKenney 
54e76506f0SPaul E. McKenney #define VERBOSE_SCALEOUT_BATCH(s, x...)							\
55e76506f0SPaul E. McKenney do {											\
56e76506f0SPaul E. McKenney 	if (verbose &&									\
57e76506f0SPaul E. McKenney 	    (verbose_batched <= 0 ||							\
58414c116eSPaul E. McKenney 	     !(atomic_inc_return(&verbose_batch_ctr) % verbose_batched))) {		\
59414c116eSPaul E. McKenney 		schedule_timeout_uninterruptible(1);					\
60f71f22b6SLi Zhijian 		pr_alert("%s" SCALE_FLAG s "\n", scale_type, ## x);			\
61414c116eSPaul E. McKenney 	}										\
62e76506f0SPaul E. McKenney } while (0)
63e76506f0SPaul E. McKenney 
64f71f22b6SLi Zhijian #define SCALEOUT_ERRSTRING(s, x...) pr_alert("%s" SCALE_FLAG "!!! " s "\n", scale_type, ## x)
651fbeb3a8SPaul E. McKenney 
661fbeb3a8SPaul E. McKenney MODULE_LICENSE("GPL");
671fbeb3a8SPaul E. McKenney MODULE_AUTHOR("Joel Fernandes (Google) <joel@joelfernandes.org>");
681fbeb3a8SPaul E. McKenney 
691fbeb3a8SPaul E. McKenney static char *scale_type = "rcu";
701fbeb3a8SPaul E. McKenney module_param(scale_type, charp, 0444);
711fbeb3a8SPaul E. McKenney MODULE_PARM_DESC(scale_type, "Type of test (rcu, srcu, refcnt, rwsem, rwlock.");
721fbeb3a8SPaul E. McKenney 
731fbeb3a8SPaul E. McKenney torture_param(int, verbose, 0, "Enable verbose debugging printk()s");
74e76506f0SPaul E. McKenney torture_param(int, verbose_batched, 0, "Batch verbose debugging printk()s");
751fbeb3a8SPaul E. McKenney 
761fbeb3a8SPaul E. McKenney // Wait until there are multiple CPUs before starting test.
771fbeb3a8SPaul E. McKenney torture_param(int, holdoff, IS_BUILTIN(CONFIG_RCU_REF_SCALE_TEST) ? 10 : 0,
781fbeb3a8SPaul E. McKenney 	      "Holdoff time before test start (s)");
79a6889becSPaul E. McKenney // Number of typesafe_lookup structures, that is, the degree of concurrency.
80a6889becSPaul E. McKenney torture_param(long, lookup_instances, 0, "Number of typesafe_lookup structures.");
811fbeb3a8SPaul E. McKenney // Number of loops per experiment, all readers execute operations concurrently.
821fbeb3a8SPaul E. McKenney torture_param(long, loops, 10000, "Number of loops per experiment.");
831fbeb3a8SPaul E. McKenney // Number of readers, with -1 defaulting to about 75% of the CPUs.
841fbeb3a8SPaul E. McKenney torture_param(int, nreaders, -1, "Number of readers, -1 for 75% of CPUs.");
851fbeb3a8SPaul E. McKenney // Number of runs.
861fbeb3a8SPaul E. McKenney torture_param(int, nruns, 30, "Number of experiments to run.");
871fbeb3a8SPaul E. McKenney // Reader delay in nanoseconds, 0 for no delay.
881fbeb3a8SPaul E. McKenney torture_param(int, readdelay, 0, "Read-side delay in nanoseconds.");
891fbeb3a8SPaul E. McKenney 
901fbeb3a8SPaul E. McKenney #ifdef MODULE
911fbeb3a8SPaul E. McKenney # define REFSCALE_SHUTDOWN 0
921fbeb3a8SPaul E. McKenney #else
931fbeb3a8SPaul E. McKenney # define REFSCALE_SHUTDOWN 1
941fbeb3a8SPaul E. McKenney #endif
951fbeb3a8SPaul E. McKenney 
961fbeb3a8SPaul E. McKenney torture_param(bool, shutdown, REFSCALE_SHUTDOWN,
971fbeb3a8SPaul E. McKenney 	      "Shutdown at end of scalability tests.");
981fbeb3a8SPaul E. McKenney 
991fbeb3a8SPaul E. McKenney struct reader_task {
1001fbeb3a8SPaul E. McKenney 	struct task_struct *task;
1011fbeb3a8SPaul E. McKenney 	int start_reader;
1021fbeb3a8SPaul E. McKenney 	wait_queue_head_t wq;
1031fbeb3a8SPaul E. McKenney 	u64 last_duration_ns;
1041fbeb3a8SPaul E. McKenney };
1051fbeb3a8SPaul E. McKenney 
1061fbeb3a8SPaul E. McKenney static struct task_struct *shutdown_task;
1071fbeb3a8SPaul E. McKenney static wait_queue_head_t shutdown_wq;
1081fbeb3a8SPaul E. McKenney 
1091fbeb3a8SPaul E. McKenney static struct task_struct *main_task;
1101fbeb3a8SPaul E. McKenney static wait_queue_head_t main_wq;
1111fbeb3a8SPaul E. McKenney static int shutdown_start;
1121fbeb3a8SPaul E. McKenney 
1131fbeb3a8SPaul E. McKenney static struct reader_task *reader_tasks;
1141fbeb3a8SPaul E. McKenney 
1151fbeb3a8SPaul E. McKenney // Number of readers that are part of the current experiment.
1161fbeb3a8SPaul E. McKenney static atomic_t nreaders_exp;
1171fbeb3a8SPaul E. McKenney 
1181fbeb3a8SPaul E. McKenney // Use to wait for all threads to start.
1191fbeb3a8SPaul E. McKenney static atomic_t n_init;
1201fbeb3a8SPaul E. McKenney static atomic_t n_started;
1211fbeb3a8SPaul E. McKenney static atomic_t n_warmedup;
1221fbeb3a8SPaul E. McKenney static atomic_t n_cooleddown;
1231fbeb3a8SPaul E. McKenney 
1241fbeb3a8SPaul E. McKenney // Track which experiment is currently running.
1251fbeb3a8SPaul E. McKenney static int exp_idx;
1261fbeb3a8SPaul E. McKenney 
1271fbeb3a8SPaul E. McKenney // Operations vector for selecting different types of tests.
1281fbeb3a8SPaul E. McKenney struct ref_scale_ops {
1293c6496c8SPaul E. McKenney 	bool (*init)(void);
1301fbeb3a8SPaul E. McKenney 	void (*cleanup)(void);
1311fbeb3a8SPaul E. McKenney 	void (*readsection)(const int nloops);
1321fbeb3a8SPaul E. McKenney 	void (*delaysection)(const int nloops, const int udl, const int ndl);
1331fbeb3a8SPaul E. McKenney 	const char *name;
1341fbeb3a8SPaul E. McKenney };
1351fbeb3a8SPaul E. McKenney 
1361fbeb3a8SPaul E. McKenney static struct ref_scale_ops *cur_ops;
1371fbeb3a8SPaul E. McKenney 
un_delay(const int udl,const int ndl)1381fbeb3a8SPaul E. McKenney static void un_delay(const int udl, const int ndl)
1391fbeb3a8SPaul E. McKenney {
1401fbeb3a8SPaul E. McKenney 	if (udl)
1411fbeb3a8SPaul E. McKenney 		udelay(udl);
1421fbeb3a8SPaul E. McKenney 	if (ndl)
1431fbeb3a8SPaul E. McKenney 		ndelay(ndl);
1441fbeb3a8SPaul E. McKenney }
1451fbeb3a8SPaul E. McKenney 
ref_rcu_read_section(const int nloops)1461fbeb3a8SPaul E. McKenney static void ref_rcu_read_section(const int nloops)
1471fbeb3a8SPaul E. McKenney {
1481fbeb3a8SPaul E. McKenney 	int i;
1491fbeb3a8SPaul E. McKenney 
1501fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
1511fbeb3a8SPaul E. McKenney 		rcu_read_lock();
1521fbeb3a8SPaul E. McKenney 		rcu_read_unlock();
1531fbeb3a8SPaul E. McKenney 	}
1541fbeb3a8SPaul E. McKenney }
1551fbeb3a8SPaul E. McKenney 
ref_rcu_delay_section(const int nloops,const int udl,const int ndl)1561fbeb3a8SPaul E. McKenney static void ref_rcu_delay_section(const int nloops, const int udl, const int ndl)
1571fbeb3a8SPaul E. McKenney {
1581fbeb3a8SPaul E. McKenney 	int i;
1591fbeb3a8SPaul E. McKenney 
1601fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
1611fbeb3a8SPaul E. McKenney 		rcu_read_lock();
1621fbeb3a8SPaul E. McKenney 		un_delay(udl, ndl);
1631fbeb3a8SPaul E. McKenney 		rcu_read_unlock();
1641fbeb3a8SPaul E. McKenney 	}
1651fbeb3a8SPaul E. McKenney }
1661fbeb3a8SPaul E. McKenney 
rcu_sync_scale_init(void)1673c6496c8SPaul E. McKenney static bool rcu_sync_scale_init(void)
1681fbeb3a8SPaul E. McKenney {
1693c6496c8SPaul E. McKenney 	return true;
1701fbeb3a8SPaul E. McKenney }
1711fbeb3a8SPaul E. McKenney 
1721fbeb3a8SPaul E. McKenney static struct ref_scale_ops rcu_ops = {
1731fbeb3a8SPaul E. McKenney 	.init		= rcu_sync_scale_init,
1741fbeb3a8SPaul E. McKenney 	.readsection	= ref_rcu_read_section,
1751fbeb3a8SPaul E. McKenney 	.delaysection	= ref_rcu_delay_section,
1761fbeb3a8SPaul E. McKenney 	.name		= "rcu"
1771fbeb3a8SPaul E. McKenney };
1781fbeb3a8SPaul E. McKenney 
1791fbeb3a8SPaul E. McKenney // Definitions for SRCU ref scale testing.
1801fbeb3a8SPaul E. McKenney DEFINE_STATIC_SRCU(srcu_refctl_scale);
1811fbeb3a8SPaul E. McKenney static struct srcu_struct *srcu_ctlp = &srcu_refctl_scale;
1821fbeb3a8SPaul E. McKenney 
srcu_ref_scale_read_section(const int nloops)1831fbeb3a8SPaul E. McKenney static void srcu_ref_scale_read_section(const int nloops)
1841fbeb3a8SPaul E. McKenney {
1851fbeb3a8SPaul E. McKenney 	int i;
1861fbeb3a8SPaul E. McKenney 	int idx;
1871fbeb3a8SPaul E. McKenney 
1881fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
1891fbeb3a8SPaul E. McKenney 		idx = srcu_read_lock(srcu_ctlp);
1901fbeb3a8SPaul E. McKenney 		srcu_read_unlock(srcu_ctlp, idx);
1911fbeb3a8SPaul E. McKenney 	}
1921fbeb3a8SPaul E. McKenney }
1931fbeb3a8SPaul E. McKenney 
srcu_ref_scale_delay_section(const int nloops,const int udl,const int ndl)1941fbeb3a8SPaul E. McKenney static void srcu_ref_scale_delay_section(const int nloops, const int udl, const int ndl)
1951fbeb3a8SPaul E. McKenney {
1961fbeb3a8SPaul E. McKenney 	int i;
1971fbeb3a8SPaul E. McKenney 	int idx;
1981fbeb3a8SPaul E. McKenney 
1991fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
2001fbeb3a8SPaul E. McKenney 		idx = srcu_read_lock(srcu_ctlp);
2011fbeb3a8SPaul E. McKenney 		un_delay(udl, ndl);
2021fbeb3a8SPaul E. McKenney 		srcu_read_unlock(srcu_ctlp, idx);
2031fbeb3a8SPaul E. McKenney 	}
2041fbeb3a8SPaul E. McKenney }
2051fbeb3a8SPaul E. McKenney 
2061fbeb3a8SPaul E. McKenney static struct ref_scale_ops srcu_ops = {
2071fbeb3a8SPaul E. McKenney 	.init		= rcu_sync_scale_init,
2081fbeb3a8SPaul E. McKenney 	.readsection	= srcu_ref_scale_read_section,
2091fbeb3a8SPaul E. McKenney 	.delaysection	= srcu_ref_scale_delay_section,
2101fbeb3a8SPaul E. McKenney 	.name		= "srcu"
2111fbeb3a8SPaul E. McKenney };
2121fbeb3a8SPaul E. McKenney 
2135f654af1SPaul E. McKenney #ifdef CONFIG_TASKS_RCU
2145f654af1SPaul E. McKenney 
2151fbeb3a8SPaul E. McKenney // Definitions for RCU Tasks ref scale testing: Empty read markers.
2161fbeb3a8SPaul E. McKenney // These definitions also work for RCU Rude readers.
rcu_tasks_ref_scale_read_section(const int nloops)2171fbeb3a8SPaul E. McKenney static void rcu_tasks_ref_scale_read_section(const int nloops)
2181fbeb3a8SPaul E. McKenney {
2191fbeb3a8SPaul E. McKenney 	int i;
2201fbeb3a8SPaul E. McKenney 
2211fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--)
2221fbeb3a8SPaul E. McKenney 		continue;
2231fbeb3a8SPaul E. McKenney }
2241fbeb3a8SPaul E. McKenney 
rcu_tasks_ref_scale_delay_section(const int nloops,const int udl,const int ndl)2251fbeb3a8SPaul E. McKenney static void rcu_tasks_ref_scale_delay_section(const int nloops, const int udl, const int ndl)
2261fbeb3a8SPaul E. McKenney {
2271fbeb3a8SPaul E. McKenney 	int i;
2281fbeb3a8SPaul E. McKenney 
2291fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--)
2301fbeb3a8SPaul E. McKenney 		un_delay(udl, ndl);
2311fbeb3a8SPaul E. McKenney }
2321fbeb3a8SPaul E. McKenney 
2331fbeb3a8SPaul E. McKenney static struct ref_scale_ops rcu_tasks_ops = {
2341fbeb3a8SPaul E. McKenney 	.init		= rcu_sync_scale_init,
2351fbeb3a8SPaul E. McKenney 	.readsection	= rcu_tasks_ref_scale_read_section,
2361fbeb3a8SPaul E. McKenney 	.delaysection	= rcu_tasks_ref_scale_delay_section,
2371fbeb3a8SPaul E. McKenney 	.name		= "rcu-tasks"
2381fbeb3a8SPaul E. McKenney };
2391fbeb3a8SPaul E. McKenney 
2405f654af1SPaul E. McKenney #define RCU_TASKS_OPS &rcu_tasks_ops,
2415f654af1SPaul E. McKenney 
2425f654af1SPaul E. McKenney #else // #ifdef CONFIG_TASKS_RCU
2435f654af1SPaul E. McKenney 
2445f654af1SPaul E. McKenney #define RCU_TASKS_OPS
2455f654af1SPaul E. McKenney 
2465f654af1SPaul E. McKenney #endif // #else // #ifdef CONFIG_TASKS_RCU
2475f654af1SPaul E. McKenney 
248dec86781SPaul E. McKenney #ifdef CONFIG_TASKS_TRACE_RCU
249dec86781SPaul E. McKenney 
2501fbeb3a8SPaul E. McKenney // Definitions for RCU Tasks Trace ref scale testing.
rcu_trace_ref_scale_read_section(const int nloops)2511fbeb3a8SPaul E. McKenney static void rcu_trace_ref_scale_read_section(const int nloops)
2521fbeb3a8SPaul E. McKenney {
2531fbeb3a8SPaul E. McKenney 	int i;
2541fbeb3a8SPaul E. McKenney 
2551fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
2561fbeb3a8SPaul E. McKenney 		rcu_read_lock_trace();
2571fbeb3a8SPaul E. McKenney 		rcu_read_unlock_trace();
2581fbeb3a8SPaul E. McKenney 	}
2591fbeb3a8SPaul E. McKenney }
2601fbeb3a8SPaul E. McKenney 
rcu_trace_ref_scale_delay_section(const int nloops,const int udl,const int ndl)2611fbeb3a8SPaul E. McKenney static void rcu_trace_ref_scale_delay_section(const int nloops, const int udl, const int ndl)
2621fbeb3a8SPaul E. McKenney {
2631fbeb3a8SPaul E. McKenney 	int i;
2641fbeb3a8SPaul E. McKenney 
2651fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
2661fbeb3a8SPaul E. McKenney 		rcu_read_lock_trace();
2671fbeb3a8SPaul E. McKenney 		un_delay(udl, ndl);
2681fbeb3a8SPaul E. McKenney 		rcu_read_unlock_trace();
2691fbeb3a8SPaul E. McKenney 	}
2701fbeb3a8SPaul E. McKenney }
2711fbeb3a8SPaul E. McKenney 
2721fbeb3a8SPaul E. McKenney static struct ref_scale_ops rcu_trace_ops = {
2731fbeb3a8SPaul E. McKenney 	.init		= rcu_sync_scale_init,
2741fbeb3a8SPaul E. McKenney 	.readsection	= rcu_trace_ref_scale_read_section,
2751fbeb3a8SPaul E. McKenney 	.delaysection	= rcu_trace_ref_scale_delay_section,
2761fbeb3a8SPaul E. McKenney 	.name		= "rcu-trace"
2771fbeb3a8SPaul E. McKenney };
2781fbeb3a8SPaul E. McKenney 
279dec86781SPaul E. McKenney #define RCU_TRACE_OPS &rcu_trace_ops,
280dec86781SPaul E. McKenney 
281dec86781SPaul E. McKenney #else // #ifdef CONFIG_TASKS_TRACE_RCU
282dec86781SPaul E. McKenney 
283dec86781SPaul E. McKenney #define RCU_TRACE_OPS
284dec86781SPaul E. McKenney 
285dec86781SPaul E. McKenney #endif // #else // #ifdef CONFIG_TASKS_TRACE_RCU
286dec86781SPaul E. McKenney 
2871fbeb3a8SPaul E. McKenney // Definitions for reference count
2881fbeb3a8SPaul E. McKenney static atomic_t refcnt;
2891fbeb3a8SPaul E. McKenney 
ref_refcnt_section(const int nloops)2901fbeb3a8SPaul E. McKenney static void ref_refcnt_section(const int nloops)
2911fbeb3a8SPaul E. McKenney {
2921fbeb3a8SPaul E. McKenney 	int i;
2931fbeb3a8SPaul E. McKenney 
2941fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
2951fbeb3a8SPaul E. McKenney 		atomic_inc(&refcnt);
2961fbeb3a8SPaul E. McKenney 		atomic_dec(&refcnt);
2971fbeb3a8SPaul E. McKenney 	}
2981fbeb3a8SPaul E. McKenney }
2991fbeb3a8SPaul E. McKenney 
ref_refcnt_delay_section(const int nloops,const int udl,const int ndl)3001fbeb3a8SPaul E. McKenney static void ref_refcnt_delay_section(const int nloops, const int udl, const int ndl)
3011fbeb3a8SPaul E. McKenney {
3021fbeb3a8SPaul E. McKenney 	int i;
3031fbeb3a8SPaul E. McKenney 
3041fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
3051fbeb3a8SPaul E. McKenney 		atomic_inc(&refcnt);
3061fbeb3a8SPaul E. McKenney 		un_delay(udl, ndl);
3071fbeb3a8SPaul E. McKenney 		atomic_dec(&refcnt);
3081fbeb3a8SPaul E. McKenney 	}
3091fbeb3a8SPaul E. McKenney }
3101fbeb3a8SPaul E. McKenney 
3111fbeb3a8SPaul E. McKenney static struct ref_scale_ops refcnt_ops = {
3121fbeb3a8SPaul E. McKenney 	.init		= rcu_sync_scale_init,
3131fbeb3a8SPaul E. McKenney 	.readsection	= ref_refcnt_section,
3141fbeb3a8SPaul E. McKenney 	.delaysection	= ref_refcnt_delay_section,
3151fbeb3a8SPaul E. McKenney 	.name		= "refcnt"
3161fbeb3a8SPaul E. McKenney };
3171fbeb3a8SPaul E. McKenney 
3181fbeb3a8SPaul E. McKenney // Definitions for rwlock
3191fbeb3a8SPaul E. McKenney static rwlock_t test_rwlock;
3201fbeb3a8SPaul E. McKenney 
ref_rwlock_init(void)3213c6496c8SPaul E. McKenney static bool ref_rwlock_init(void)
3221fbeb3a8SPaul E. McKenney {
3231fbeb3a8SPaul E. McKenney 	rwlock_init(&test_rwlock);
3243c6496c8SPaul E. McKenney 	return true;
3251fbeb3a8SPaul E. McKenney }
3261fbeb3a8SPaul E. McKenney 
ref_rwlock_section(const int nloops)3271fbeb3a8SPaul E. McKenney static void ref_rwlock_section(const int nloops)
3281fbeb3a8SPaul E. McKenney {
3291fbeb3a8SPaul E. McKenney 	int i;
3301fbeb3a8SPaul E. McKenney 
3311fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
3321fbeb3a8SPaul E. McKenney 		read_lock(&test_rwlock);
3331fbeb3a8SPaul E. McKenney 		read_unlock(&test_rwlock);
3341fbeb3a8SPaul E. McKenney 	}
3351fbeb3a8SPaul E. McKenney }
3361fbeb3a8SPaul E. McKenney 
ref_rwlock_delay_section(const int nloops,const int udl,const int ndl)3371fbeb3a8SPaul E. McKenney static void ref_rwlock_delay_section(const int nloops, const int udl, const int ndl)
3381fbeb3a8SPaul E. McKenney {
3391fbeb3a8SPaul E. McKenney 	int i;
3401fbeb3a8SPaul E. McKenney 
3411fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
3421fbeb3a8SPaul E. McKenney 		read_lock(&test_rwlock);
3431fbeb3a8SPaul E. McKenney 		un_delay(udl, ndl);
3441fbeb3a8SPaul E. McKenney 		read_unlock(&test_rwlock);
3451fbeb3a8SPaul E. McKenney 	}
3461fbeb3a8SPaul E. McKenney }
3471fbeb3a8SPaul E. McKenney 
3481fbeb3a8SPaul E. McKenney static struct ref_scale_ops rwlock_ops = {
3491fbeb3a8SPaul E. McKenney 	.init		= ref_rwlock_init,
3501fbeb3a8SPaul E. McKenney 	.readsection	= ref_rwlock_section,
3511fbeb3a8SPaul E. McKenney 	.delaysection	= ref_rwlock_delay_section,
3521fbeb3a8SPaul E. McKenney 	.name		= "rwlock"
3531fbeb3a8SPaul E. McKenney };
3541fbeb3a8SPaul E. McKenney 
3551fbeb3a8SPaul E. McKenney // Definitions for rwsem
3561fbeb3a8SPaul E. McKenney static struct rw_semaphore test_rwsem;
3571fbeb3a8SPaul E. McKenney 
ref_rwsem_init(void)3583c6496c8SPaul E. McKenney static bool ref_rwsem_init(void)
3591fbeb3a8SPaul E. McKenney {
3601fbeb3a8SPaul E. McKenney 	init_rwsem(&test_rwsem);
3613c6496c8SPaul E. McKenney 	return true;
3621fbeb3a8SPaul E. McKenney }
3631fbeb3a8SPaul E. McKenney 
ref_rwsem_section(const int nloops)3641fbeb3a8SPaul E. McKenney static void ref_rwsem_section(const int nloops)
3651fbeb3a8SPaul E. McKenney {
3661fbeb3a8SPaul E. McKenney 	int i;
3671fbeb3a8SPaul E. McKenney 
3681fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
3691fbeb3a8SPaul E. McKenney 		down_read(&test_rwsem);
3701fbeb3a8SPaul E. McKenney 		up_read(&test_rwsem);
3711fbeb3a8SPaul E. McKenney 	}
3721fbeb3a8SPaul E. McKenney }
3731fbeb3a8SPaul E. McKenney 
ref_rwsem_delay_section(const int nloops,const int udl,const int ndl)3741fbeb3a8SPaul E. McKenney static void ref_rwsem_delay_section(const int nloops, const int udl, const int ndl)
3751fbeb3a8SPaul E. McKenney {
3761fbeb3a8SPaul E. McKenney 	int i;
3771fbeb3a8SPaul E. McKenney 
3781fbeb3a8SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
3791fbeb3a8SPaul E. McKenney 		down_read(&test_rwsem);
3801fbeb3a8SPaul E. McKenney 		un_delay(udl, ndl);
3811fbeb3a8SPaul E. McKenney 		up_read(&test_rwsem);
3821fbeb3a8SPaul E. McKenney 	}
3831fbeb3a8SPaul E. McKenney }
3841fbeb3a8SPaul E. McKenney 
3851fbeb3a8SPaul E. McKenney static struct ref_scale_ops rwsem_ops = {
3861fbeb3a8SPaul E. McKenney 	.init		= ref_rwsem_init,
3871fbeb3a8SPaul E. McKenney 	.readsection	= ref_rwsem_section,
3881fbeb3a8SPaul E. McKenney 	.delaysection	= ref_rwsem_delay_section,
3891fbeb3a8SPaul E. McKenney 	.name		= "rwsem"
3901fbeb3a8SPaul E. McKenney };
3911fbeb3a8SPaul E. McKenney 
392e9b800dbSPaul E. McKenney // Definitions for global spinlock
3937bf336fbSZqiang static DEFINE_RAW_SPINLOCK(test_lock);
394e9b800dbSPaul E. McKenney 
ref_lock_section(const int nloops)395e9b800dbSPaul E. McKenney static void ref_lock_section(const int nloops)
396e9b800dbSPaul E. McKenney {
397e9b800dbSPaul E. McKenney 	int i;
398e9b800dbSPaul E. McKenney 
399e9b800dbSPaul E. McKenney 	preempt_disable();
400e9b800dbSPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
4017bf336fbSZqiang 		raw_spin_lock(&test_lock);
4027bf336fbSZqiang 		raw_spin_unlock(&test_lock);
403e9b800dbSPaul E. McKenney 	}
404e9b800dbSPaul E. McKenney 	preempt_enable();
405e9b800dbSPaul E. McKenney }
406e9b800dbSPaul E. McKenney 
ref_lock_delay_section(const int nloops,const int udl,const int ndl)407e9b800dbSPaul E. McKenney static void ref_lock_delay_section(const int nloops, const int udl, const int ndl)
408e9b800dbSPaul E. McKenney {
409e9b800dbSPaul E. McKenney 	int i;
410e9b800dbSPaul E. McKenney 
411e9b800dbSPaul E. McKenney 	preempt_disable();
412e9b800dbSPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
4137bf336fbSZqiang 		raw_spin_lock(&test_lock);
414e9b800dbSPaul E. McKenney 		un_delay(udl, ndl);
4157bf336fbSZqiang 		raw_spin_unlock(&test_lock);
416e9b800dbSPaul E. McKenney 	}
417e9b800dbSPaul E. McKenney 	preempt_enable();
418e9b800dbSPaul E. McKenney }
419e9b800dbSPaul E. McKenney 
420e9b800dbSPaul E. McKenney static struct ref_scale_ops lock_ops = {
421e9b800dbSPaul E. McKenney 	.readsection	= ref_lock_section,
422e9b800dbSPaul E. McKenney 	.delaysection	= ref_lock_delay_section,
423e9b800dbSPaul E. McKenney 	.name		= "lock"
424e9b800dbSPaul E. McKenney };
425e9b800dbSPaul E. McKenney 
426e9b800dbSPaul E. McKenney // Definitions for global irq-save spinlock
427e9b800dbSPaul E. McKenney 
ref_lock_irq_section(const int nloops)428e9b800dbSPaul E. McKenney static void ref_lock_irq_section(const int nloops)
429e9b800dbSPaul E. McKenney {
430e9b800dbSPaul E. McKenney 	unsigned long flags;
431e9b800dbSPaul E. McKenney 	int i;
432e9b800dbSPaul E. McKenney 
433e9b800dbSPaul E. McKenney 	preempt_disable();
434e9b800dbSPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
4357bf336fbSZqiang 		raw_spin_lock_irqsave(&test_lock, flags);
4367bf336fbSZqiang 		raw_spin_unlock_irqrestore(&test_lock, flags);
437e9b800dbSPaul E. McKenney 	}
438e9b800dbSPaul E. McKenney 	preempt_enable();
439e9b800dbSPaul E. McKenney }
440e9b800dbSPaul E. McKenney 
ref_lock_irq_delay_section(const int nloops,const int udl,const int ndl)441e9b800dbSPaul E. McKenney static void ref_lock_irq_delay_section(const int nloops, const int udl, const int ndl)
442e9b800dbSPaul E. McKenney {
443e9b800dbSPaul E. McKenney 	unsigned long flags;
444e9b800dbSPaul E. McKenney 	int i;
445e9b800dbSPaul E. McKenney 
446e9b800dbSPaul E. McKenney 	preempt_disable();
447e9b800dbSPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
4487bf336fbSZqiang 		raw_spin_lock_irqsave(&test_lock, flags);
449e9b800dbSPaul E. McKenney 		un_delay(udl, ndl);
4507bf336fbSZqiang 		raw_spin_unlock_irqrestore(&test_lock, flags);
451e9b800dbSPaul E. McKenney 	}
452e9b800dbSPaul E. McKenney 	preempt_enable();
453e9b800dbSPaul E. McKenney }
454e9b800dbSPaul E. McKenney 
455e9b800dbSPaul E. McKenney static struct ref_scale_ops lock_irq_ops = {
456e9b800dbSPaul E. McKenney 	.readsection	= ref_lock_irq_section,
457e9b800dbSPaul E. McKenney 	.delaysection	= ref_lock_irq_delay_section,
458e9b800dbSPaul E. McKenney 	.name		= "lock-irq"
459e9b800dbSPaul E. McKenney };
460e9b800dbSPaul E. McKenney 
461e9b800dbSPaul E. McKenney // Definitions acquire-release.
462e9b800dbSPaul E. McKenney static DEFINE_PER_CPU(unsigned long, test_acqrel);
463e9b800dbSPaul E. McKenney 
ref_acqrel_section(const int nloops)464e9b800dbSPaul E. McKenney static void ref_acqrel_section(const int nloops)
465e9b800dbSPaul E. McKenney {
466e9b800dbSPaul E. McKenney 	unsigned long x;
467e9b800dbSPaul E. McKenney 	int i;
468e9b800dbSPaul E. McKenney 
469e9b800dbSPaul E. McKenney 	preempt_disable();
470e9b800dbSPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
471e9b800dbSPaul E. McKenney 		x = smp_load_acquire(this_cpu_ptr(&test_acqrel));
472e9b800dbSPaul E. McKenney 		smp_store_release(this_cpu_ptr(&test_acqrel), x + 1);
473e9b800dbSPaul E. McKenney 	}
474e9b800dbSPaul E. McKenney 	preempt_enable();
475e9b800dbSPaul E. McKenney }
476e9b800dbSPaul E. McKenney 
ref_acqrel_delay_section(const int nloops,const int udl,const int ndl)477e9b800dbSPaul E. McKenney static void ref_acqrel_delay_section(const int nloops, const int udl, const int ndl)
478e9b800dbSPaul E. McKenney {
479e9b800dbSPaul E. McKenney 	unsigned long x;
480e9b800dbSPaul E. McKenney 	int i;
481e9b800dbSPaul E. McKenney 
482e9b800dbSPaul E. McKenney 	preempt_disable();
483e9b800dbSPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
484e9b800dbSPaul E. McKenney 		x = smp_load_acquire(this_cpu_ptr(&test_acqrel));
485e9b800dbSPaul E. McKenney 		un_delay(udl, ndl);
486e9b800dbSPaul E. McKenney 		smp_store_release(this_cpu_ptr(&test_acqrel), x + 1);
487e9b800dbSPaul E. McKenney 	}
488e9b800dbSPaul E. McKenney 	preempt_enable();
489e9b800dbSPaul E. McKenney }
490e9b800dbSPaul E. McKenney 
491e9b800dbSPaul E. McKenney static struct ref_scale_ops acqrel_ops = {
492e9b800dbSPaul E. McKenney 	.readsection	= ref_acqrel_section,
493e9b800dbSPaul E. McKenney 	.delaysection	= ref_acqrel_delay_section,
494e9b800dbSPaul E. McKenney 	.name		= "acqrel"
495e9b800dbSPaul E. McKenney };
496e9b800dbSPaul E. McKenney 
49725f6fa53SPaul E. McKenney static volatile u64 stopopts;
49825f6fa53SPaul E. McKenney 
ref_clock_section(const int nloops)49925f6fa53SPaul E. McKenney static void ref_clock_section(const int nloops)
50025f6fa53SPaul E. McKenney {
50125f6fa53SPaul E. McKenney 	u64 x = 0;
50225f6fa53SPaul E. McKenney 	int i;
50325f6fa53SPaul E. McKenney 
50425f6fa53SPaul E. McKenney 	preempt_disable();
50525f6fa53SPaul E. McKenney 	for (i = nloops; i >= 0; i--)
50625f6fa53SPaul E. McKenney 		x += ktime_get_real_fast_ns();
50725f6fa53SPaul E. McKenney 	preempt_enable();
50825f6fa53SPaul E. McKenney 	stopopts = x;
50925f6fa53SPaul E. McKenney }
51025f6fa53SPaul E. McKenney 
ref_clock_delay_section(const int nloops,const int udl,const int ndl)51125f6fa53SPaul E. McKenney static void ref_clock_delay_section(const int nloops, const int udl, const int ndl)
51225f6fa53SPaul E. McKenney {
51325f6fa53SPaul E. McKenney 	u64 x = 0;
51425f6fa53SPaul E. McKenney 	int i;
51525f6fa53SPaul E. McKenney 
51625f6fa53SPaul E. McKenney 	preempt_disable();
51725f6fa53SPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
51825f6fa53SPaul E. McKenney 		x += ktime_get_real_fast_ns();
51925f6fa53SPaul E. McKenney 		un_delay(udl, ndl);
52025f6fa53SPaul E. McKenney 	}
52125f6fa53SPaul E. McKenney 	preempt_enable();
52225f6fa53SPaul E. McKenney 	stopopts = x;
52325f6fa53SPaul E. McKenney }
52425f6fa53SPaul E. McKenney 
52525f6fa53SPaul E. McKenney static struct ref_scale_ops clock_ops = {
52625f6fa53SPaul E. McKenney 	.readsection	= ref_clock_section,
52725f6fa53SPaul E. McKenney 	.delaysection	= ref_clock_delay_section,
52825f6fa53SPaul E. McKenney 	.name		= "clock"
52925f6fa53SPaul E. McKenney };
53025f6fa53SPaul E. McKenney 
ref_jiffies_section(const int nloops)531b5a2801fSPaul E. McKenney static void ref_jiffies_section(const int nloops)
532b5a2801fSPaul E. McKenney {
533b5a2801fSPaul E. McKenney 	u64 x = 0;
534b5a2801fSPaul E. McKenney 	int i;
535b5a2801fSPaul E. McKenney 
536b5a2801fSPaul E. McKenney 	preempt_disable();
537b5a2801fSPaul E. McKenney 	for (i = nloops; i >= 0; i--)
538b5a2801fSPaul E. McKenney 		x += jiffies;
539b5a2801fSPaul E. McKenney 	preempt_enable();
540b5a2801fSPaul E. McKenney 	stopopts = x;
541b5a2801fSPaul E. McKenney }
542b5a2801fSPaul E. McKenney 
ref_jiffies_delay_section(const int nloops,const int udl,const int ndl)543b5a2801fSPaul E. McKenney static void ref_jiffies_delay_section(const int nloops, const int udl, const int ndl)
544b5a2801fSPaul E. McKenney {
545b5a2801fSPaul E. McKenney 	u64 x = 0;
546b5a2801fSPaul E. McKenney 	int i;
547b5a2801fSPaul E. McKenney 
548b5a2801fSPaul E. McKenney 	preempt_disable();
549b5a2801fSPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
550b5a2801fSPaul E. McKenney 		x += jiffies;
551b5a2801fSPaul E. McKenney 		un_delay(udl, ndl);
552b5a2801fSPaul E. McKenney 	}
553b5a2801fSPaul E. McKenney 	preempt_enable();
554b5a2801fSPaul E. McKenney 	stopopts = x;
555b5a2801fSPaul E. McKenney }
556b5a2801fSPaul E. McKenney 
557b5a2801fSPaul E. McKenney static struct ref_scale_ops jiffies_ops = {
558b5a2801fSPaul E. McKenney 	.readsection	= ref_jiffies_section,
559b5a2801fSPaul E. McKenney 	.delaysection	= ref_jiffies_delay_section,
560b5a2801fSPaul E. McKenney 	.name		= "jiffies"
561b5a2801fSPaul E. McKenney };
562b5a2801fSPaul E. McKenney 
563a6889becSPaul E. McKenney ////////////////////////////////////////////////////////////////////////
564a6889becSPaul E. McKenney //
565a6889becSPaul E. McKenney // Methods leveraging SLAB_TYPESAFE_BY_RCU.
566a6889becSPaul E. McKenney //
567a6889becSPaul E. McKenney 
568a6889becSPaul E. McKenney // Item to look up in a typesafe manner.  Array of pointers to these.
569a6889becSPaul E. McKenney struct refscale_typesafe {
570a6889becSPaul E. McKenney 	atomic_t rts_refctr;  // Used by all flavors
571a6889becSPaul E. McKenney 	spinlock_t rts_lock;
572a6889becSPaul E. McKenney 	seqlock_t rts_seqlock;
573a6889becSPaul E. McKenney 	unsigned int a;
574a6889becSPaul E. McKenney 	unsigned int b;
575a6889becSPaul E. McKenney };
576a6889becSPaul E. McKenney 
577a6889becSPaul E. McKenney static struct kmem_cache *typesafe_kmem_cachep;
578a6889becSPaul E. McKenney static struct refscale_typesafe **rtsarray;
579a6889becSPaul E. McKenney static long rtsarray_size;
580a6889becSPaul E. McKenney static DEFINE_TORTURE_RANDOM_PERCPU(refscale_rand);
581a6889becSPaul E. McKenney static bool (*rts_acquire)(struct refscale_typesafe *rtsp, unsigned int *start);
582a6889becSPaul E. McKenney static bool (*rts_release)(struct refscale_typesafe *rtsp, unsigned int start);
583a6889becSPaul E. McKenney 
584a6889becSPaul E. McKenney // Conditionally acquire an explicit in-structure reference count.
typesafe_ref_acquire(struct refscale_typesafe * rtsp,unsigned int * start)585a6889becSPaul E. McKenney static bool typesafe_ref_acquire(struct refscale_typesafe *rtsp, unsigned int *start)
586a6889becSPaul E. McKenney {
587a6889becSPaul E. McKenney 	return atomic_inc_not_zero(&rtsp->rts_refctr);
588a6889becSPaul E. McKenney }
589a6889becSPaul E. McKenney 
590a6889becSPaul E. McKenney // Unconditionally release an explicit in-structure reference count.
typesafe_ref_release(struct refscale_typesafe * rtsp,unsigned int start)591a6889becSPaul E. McKenney static bool typesafe_ref_release(struct refscale_typesafe *rtsp, unsigned int start)
592a6889becSPaul E. McKenney {
593a6889becSPaul E. McKenney 	if (!atomic_dec_return(&rtsp->rts_refctr)) {
594a6889becSPaul E. McKenney 		WRITE_ONCE(rtsp->a, rtsp->a + 1);
595a6889becSPaul E. McKenney 		kmem_cache_free(typesafe_kmem_cachep, rtsp);
596a6889becSPaul E. McKenney 	}
597a6889becSPaul E. McKenney 	return true;
598a6889becSPaul E. McKenney }
599a6889becSPaul E. McKenney 
600a6889becSPaul E. McKenney // Unconditionally acquire an explicit in-structure spinlock.
typesafe_lock_acquire(struct refscale_typesafe * rtsp,unsigned int * start)601a6889becSPaul E. McKenney static bool typesafe_lock_acquire(struct refscale_typesafe *rtsp, unsigned int *start)
602a6889becSPaul E. McKenney {
603a6889becSPaul E. McKenney 	spin_lock(&rtsp->rts_lock);
604a6889becSPaul E. McKenney 	return true;
605a6889becSPaul E. McKenney }
606a6889becSPaul E. McKenney 
607a6889becSPaul E. McKenney // Unconditionally release an explicit in-structure spinlock.
typesafe_lock_release(struct refscale_typesafe * rtsp,unsigned int start)608a6889becSPaul E. McKenney static bool typesafe_lock_release(struct refscale_typesafe *rtsp, unsigned int start)
609a6889becSPaul E. McKenney {
610a6889becSPaul E. McKenney 	spin_unlock(&rtsp->rts_lock);
611a6889becSPaul E. McKenney 	return true;
612a6889becSPaul E. McKenney }
613a6889becSPaul E. McKenney 
614a6889becSPaul E. McKenney // Unconditionally acquire an explicit in-structure sequence lock.
typesafe_seqlock_acquire(struct refscale_typesafe * rtsp,unsigned int * start)615a6889becSPaul E. McKenney static bool typesafe_seqlock_acquire(struct refscale_typesafe *rtsp, unsigned int *start)
616a6889becSPaul E. McKenney {
617a6889becSPaul E. McKenney 	*start = read_seqbegin(&rtsp->rts_seqlock);
618a6889becSPaul E. McKenney 	return true;
619a6889becSPaul E. McKenney }
620a6889becSPaul E. McKenney 
621a6889becSPaul E. McKenney // Conditionally release an explicit in-structure sequence lock.  Return
622a6889becSPaul E. McKenney // true if this release was successful, that is, if no retry is required.
typesafe_seqlock_release(struct refscale_typesafe * rtsp,unsigned int start)623a6889becSPaul E. McKenney static bool typesafe_seqlock_release(struct refscale_typesafe *rtsp, unsigned int start)
624a6889becSPaul E. McKenney {
625a6889becSPaul E. McKenney 	return !read_seqretry(&rtsp->rts_seqlock, start);
626a6889becSPaul E. McKenney }
627a6889becSPaul E. McKenney 
628a6889becSPaul E. McKenney // Do a read-side critical section with the specified delay in
629a6889becSPaul E. McKenney // microseconds and nanoseconds inserted so as to increase probability
630a6889becSPaul E. McKenney // of failure.
typesafe_delay_section(const int nloops,const int udl,const int ndl)631a6889becSPaul E. McKenney static void typesafe_delay_section(const int nloops, const int udl, const int ndl)
632a6889becSPaul E. McKenney {
633a6889becSPaul E. McKenney 	unsigned int a;
634a6889becSPaul E. McKenney 	unsigned int b;
635a6889becSPaul E. McKenney 	int i;
636a6889becSPaul E. McKenney 	long idx;
637a6889becSPaul E. McKenney 	struct refscale_typesafe *rtsp;
638a6889becSPaul E. McKenney 	unsigned int start;
639a6889becSPaul E. McKenney 
640a6889becSPaul E. McKenney 	for (i = nloops; i >= 0; i--) {
641a6889becSPaul E. McKenney 		preempt_disable();
642a6889becSPaul E. McKenney 		idx = torture_random(this_cpu_ptr(&refscale_rand)) % rtsarray_size;
643a6889becSPaul E. McKenney 		preempt_enable();
644a6889becSPaul E. McKenney retry:
645a6889becSPaul E. McKenney 		rcu_read_lock();
646a6889becSPaul E. McKenney 		rtsp = rcu_dereference(rtsarray[idx]);
647a6889becSPaul E. McKenney 		a = READ_ONCE(rtsp->a);
648a6889becSPaul E. McKenney 		if (!rts_acquire(rtsp, &start)) {
649a6889becSPaul E. McKenney 			rcu_read_unlock();
650a6889becSPaul E. McKenney 			goto retry;
651a6889becSPaul E. McKenney 		}
652a6889becSPaul E. McKenney 		if (a != READ_ONCE(rtsp->a)) {
653a6889becSPaul E. McKenney 			(void)rts_release(rtsp, start);
654a6889becSPaul E. McKenney 			rcu_read_unlock();
655a6889becSPaul E. McKenney 			goto retry;
656a6889becSPaul E. McKenney 		}
657a6889becSPaul E. McKenney 		un_delay(udl, ndl);
658730c3ed4SPaul E. McKenney 		b = READ_ONCE(rtsp->a);
659a6889becSPaul E. McKenney 		// Remember, seqlock read-side release can fail.
660a6889becSPaul E. McKenney 		if (!rts_release(rtsp, start)) {
661a6889becSPaul E. McKenney 			rcu_read_unlock();
662a6889becSPaul E. McKenney 			goto retry;
663a6889becSPaul E. McKenney 		}
664a6889becSPaul E. McKenney 		WARN_ONCE(a != b, "Re-read of ->a changed from %u to %u.\n", a, b);
665a6889becSPaul E. McKenney 		b = rtsp->b;
666a6889becSPaul E. McKenney 		rcu_read_unlock();
667a6889becSPaul E. McKenney 		WARN_ON_ONCE(a * a != b);
668a6889becSPaul E. McKenney 	}
669a6889becSPaul E. McKenney }
670a6889becSPaul E. McKenney 
671a6889becSPaul E. McKenney // Because the acquisition and release methods are expensive, there
672a6889becSPaul E. McKenney // is no point in optimizing away the un_delay() function's two checks.
673a6889becSPaul E. McKenney // Thus simply define typesafe_read_section() as a simple wrapper around
674a6889becSPaul E. McKenney // typesafe_delay_section().
typesafe_read_section(const int nloops)675a6889becSPaul E. McKenney static void typesafe_read_section(const int nloops)
676a6889becSPaul E. McKenney {
677a6889becSPaul E. McKenney 	typesafe_delay_section(nloops, 0, 0);
678a6889becSPaul E. McKenney }
679a6889becSPaul E. McKenney 
680a6889becSPaul E. McKenney // Allocate and initialize one refscale_typesafe structure.
typesafe_alloc_one(void)681a6889becSPaul E. McKenney static struct refscale_typesafe *typesafe_alloc_one(void)
682a6889becSPaul E. McKenney {
683a6889becSPaul E. McKenney 	struct refscale_typesafe *rtsp;
684a6889becSPaul E. McKenney 
685a6889becSPaul E. McKenney 	rtsp = kmem_cache_alloc(typesafe_kmem_cachep, GFP_KERNEL);
686a6889becSPaul E. McKenney 	if (!rtsp)
687a6889becSPaul E. McKenney 		return NULL;
688a6889becSPaul E. McKenney 	atomic_set(&rtsp->rts_refctr, 1);
689a6889becSPaul E. McKenney 	WRITE_ONCE(rtsp->a, rtsp->a + 1);
690a6889becSPaul E. McKenney 	WRITE_ONCE(rtsp->b, rtsp->a * rtsp->a);
691a6889becSPaul E. McKenney 	return rtsp;
692a6889becSPaul E. McKenney }
693a6889becSPaul E. McKenney 
694a6889becSPaul E. McKenney // Slab-allocator constructor for refscale_typesafe structures created
695a6889becSPaul E. McKenney // out of a new slab of system memory.
refscale_typesafe_ctor(void * rtsp_in)696a6889becSPaul E. McKenney static void refscale_typesafe_ctor(void *rtsp_in)
697a6889becSPaul E. McKenney {
698a6889becSPaul E. McKenney 	struct refscale_typesafe *rtsp = rtsp_in;
699a6889becSPaul E. McKenney 
700a6889becSPaul E. McKenney 	spin_lock_init(&rtsp->rts_lock);
701a6889becSPaul E. McKenney 	seqlock_init(&rtsp->rts_seqlock);
702a6889becSPaul E. McKenney 	preempt_disable();
703a6889becSPaul E. McKenney 	rtsp->a = torture_random(this_cpu_ptr(&refscale_rand));
704a6889becSPaul E. McKenney 	preempt_enable();
705a6889becSPaul E. McKenney }
706a6889becSPaul E. McKenney 
707a6889becSPaul E. McKenney static struct ref_scale_ops typesafe_ref_ops;
708a6889becSPaul E. McKenney static struct ref_scale_ops typesafe_lock_ops;
709a6889becSPaul E. McKenney static struct ref_scale_ops typesafe_seqlock_ops;
710a6889becSPaul E. McKenney 
711a6889becSPaul E. McKenney // Initialize for a typesafe test.
typesafe_init(void)712a6889becSPaul E. McKenney static bool typesafe_init(void)
713a6889becSPaul E. McKenney {
714a6889becSPaul E. McKenney 	long idx;
715a6889becSPaul E. McKenney 	long si = lookup_instances;
716a6889becSPaul E. McKenney 
717a6889becSPaul E. McKenney 	typesafe_kmem_cachep = kmem_cache_create("refscale_typesafe",
718a6889becSPaul E. McKenney 						 sizeof(struct refscale_typesafe), sizeof(void *),
719a6889becSPaul E. McKenney 						 SLAB_TYPESAFE_BY_RCU, refscale_typesafe_ctor);
720a6889becSPaul E. McKenney 	if (!typesafe_kmem_cachep)
721a6889becSPaul E. McKenney 		return false;
722a6889becSPaul E. McKenney 	if (si < 0)
723a6889becSPaul E. McKenney 		si = -si * nr_cpu_ids;
724a6889becSPaul E. McKenney 	else if (si == 0)
725a6889becSPaul E. McKenney 		si = nr_cpu_ids;
726a6889becSPaul E. McKenney 	rtsarray_size = si;
727a6889becSPaul E. McKenney 	rtsarray = kcalloc(si, sizeof(*rtsarray), GFP_KERNEL);
728a6889becSPaul E. McKenney 	if (!rtsarray)
729a6889becSPaul E. McKenney 		return false;
730a6889becSPaul E. McKenney 	for (idx = 0; idx < rtsarray_size; idx++) {
731a6889becSPaul E. McKenney 		rtsarray[idx] = typesafe_alloc_one();
732a6889becSPaul E. McKenney 		if (!rtsarray[idx])
733a6889becSPaul E. McKenney 			return false;
734a6889becSPaul E. McKenney 	}
735a6889becSPaul E. McKenney 	if (cur_ops == &typesafe_ref_ops) {
736a6889becSPaul E. McKenney 		rts_acquire = typesafe_ref_acquire;
737a6889becSPaul E. McKenney 		rts_release = typesafe_ref_release;
738a6889becSPaul E. McKenney 	} else if (cur_ops == &typesafe_lock_ops) {
739a6889becSPaul E. McKenney 		rts_acquire = typesafe_lock_acquire;
740a6889becSPaul E. McKenney 		rts_release = typesafe_lock_release;
741a6889becSPaul E. McKenney 	} else if (cur_ops == &typesafe_seqlock_ops) {
742a6889becSPaul E. McKenney 		rts_acquire = typesafe_seqlock_acquire;
743a6889becSPaul E. McKenney 		rts_release = typesafe_seqlock_release;
744a6889becSPaul E. McKenney 	} else {
745a6889becSPaul E. McKenney 		WARN_ON_ONCE(1);
746a6889becSPaul E. McKenney 		return false;
747a6889becSPaul E. McKenney 	}
748a6889becSPaul E. McKenney 	return true;
749a6889becSPaul E. McKenney }
750a6889becSPaul E. McKenney 
751a6889becSPaul E. McKenney // Clean up after a typesafe test.
typesafe_cleanup(void)752a6889becSPaul E. McKenney static void typesafe_cleanup(void)
753a6889becSPaul E. McKenney {
754a6889becSPaul E. McKenney 	long idx;
755a6889becSPaul E. McKenney 
756a6889becSPaul E. McKenney 	if (rtsarray) {
757a6889becSPaul E. McKenney 		for (idx = 0; idx < rtsarray_size; idx++)
758a6889becSPaul E. McKenney 			kmem_cache_free(typesafe_kmem_cachep, rtsarray[idx]);
759a6889becSPaul E. McKenney 		kfree(rtsarray);
760a6889becSPaul E. McKenney 		rtsarray = NULL;
761a6889becSPaul E. McKenney 		rtsarray_size = 0;
762a6889becSPaul E. McKenney 	}
763a6889becSPaul E. McKenney 	kmem_cache_destroy(typesafe_kmem_cachep);
764a6889becSPaul E. McKenney 	typesafe_kmem_cachep = NULL;
765a6889becSPaul E. McKenney 	rts_acquire = NULL;
766a6889becSPaul E. McKenney 	rts_release = NULL;
767a6889becSPaul E. McKenney }
768a6889becSPaul E. McKenney 
769a6889becSPaul E. McKenney // The typesafe_init() function distinguishes these structures by address.
770a6889becSPaul E. McKenney static struct ref_scale_ops typesafe_ref_ops = {
771a6889becSPaul E. McKenney 	.init		= typesafe_init,
772a6889becSPaul E. McKenney 	.cleanup	= typesafe_cleanup,
773a6889becSPaul E. McKenney 	.readsection	= typesafe_read_section,
774a6889becSPaul E. McKenney 	.delaysection	= typesafe_delay_section,
775a6889becSPaul E. McKenney 	.name		= "typesafe_ref"
776a6889becSPaul E. McKenney };
777a6889becSPaul E. McKenney 
778a6889becSPaul E. McKenney static struct ref_scale_ops typesafe_lock_ops = {
779a6889becSPaul E. McKenney 	.init		= typesafe_init,
780a6889becSPaul E. McKenney 	.cleanup	= typesafe_cleanup,
781a6889becSPaul E. McKenney 	.readsection	= typesafe_read_section,
782a6889becSPaul E. McKenney 	.delaysection	= typesafe_delay_section,
783a6889becSPaul E. McKenney 	.name		= "typesafe_lock"
784a6889becSPaul E. McKenney };
785a6889becSPaul E. McKenney 
786a6889becSPaul E. McKenney static struct ref_scale_ops typesafe_seqlock_ops = {
787a6889becSPaul E. McKenney 	.init		= typesafe_init,
788a6889becSPaul E. McKenney 	.cleanup	= typesafe_cleanup,
789a6889becSPaul E. McKenney 	.readsection	= typesafe_read_section,
790a6889becSPaul E. McKenney 	.delaysection	= typesafe_delay_section,
791a6889becSPaul E. McKenney 	.name		= "typesafe_seqlock"
792a6889becSPaul E. McKenney };
793a6889becSPaul E. McKenney 
rcu_scale_one_reader(void)7941fbeb3a8SPaul E. McKenney static void rcu_scale_one_reader(void)
7951fbeb3a8SPaul E. McKenney {
7961fbeb3a8SPaul E. McKenney 	if (readdelay <= 0)
7971fbeb3a8SPaul E. McKenney 		cur_ops->readsection(loops);
7981fbeb3a8SPaul E. McKenney 	else
7991fbeb3a8SPaul E. McKenney 		cur_ops->delaysection(loops, readdelay / 1000, readdelay % 1000);
8001fbeb3a8SPaul E. McKenney }
8011fbeb3a8SPaul E. McKenney 
8021fbeb3a8SPaul E. McKenney // Reader kthread.  Repeatedly does empty RCU read-side
8031fbeb3a8SPaul E. McKenney // critical section, minimizing update-side interference.
8041fbeb3a8SPaul E. McKenney static int
ref_scale_reader(void * arg)8051fbeb3a8SPaul E. McKenney ref_scale_reader(void *arg)
8061fbeb3a8SPaul E. McKenney {
8071fbeb3a8SPaul E. McKenney 	unsigned long flags;
8081fbeb3a8SPaul E. McKenney 	long me = (long)arg;
8091fbeb3a8SPaul E. McKenney 	struct reader_task *rt = &(reader_tasks[me]);
8101fbeb3a8SPaul E. McKenney 	u64 start;
8111fbeb3a8SPaul E. McKenney 	s64 duration;
8121fbeb3a8SPaul E. McKenney 
813e76506f0SPaul E. McKenney 	VERBOSE_SCALEOUT_BATCH("ref_scale_reader %ld: task started", me);
81405bc276cSPaul E. McKenney 	WARN_ON_ONCE(set_cpus_allowed_ptr(current, cpumask_of(me % nr_cpu_ids)));
8151fbeb3a8SPaul E. McKenney 	set_user_nice(current, MAX_NICE);
8161fbeb3a8SPaul E. McKenney 	atomic_inc(&n_init);
8171fbeb3a8SPaul E. McKenney 	if (holdoff)
8181fbeb3a8SPaul E. McKenney 		schedule_timeout_interruptible(holdoff * HZ);
8191fbeb3a8SPaul E. McKenney repeat:
82005bc276cSPaul E. McKenney 	VERBOSE_SCALEOUT_BATCH("ref_scale_reader %ld: waiting to start next experiment on cpu %d", me, raw_smp_processor_id());
8211fbeb3a8SPaul E. McKenney 
8221fbeb3a8SPaul E. McKenney 	// Wait for signal that this reader can start.
8231fbeb3a8SPaul E. McKenney 	wait_event(rt->wq, (atomic_read(&nreaders_exp) && smp_load_acquire(&rt->start_reader)) ||
8241fbeb3a8SPaul E. McKenney 			   torture_must_stop());
8251fbeb3a8SPaul E. McKenney 
8261fbeb3a8SPaul E. McKenney 	if (torture_must_stop())
8271fbeb3a8SPaul E. McKenney 		goto end;
8281fbeb3a8SPaul E. McKenney 
8291fbeb3a8SPaul E. McKenney 	// Make sure that the CPU is affinitized appropriately during testing.
83005bc276cSPaul E. McKenney 	WARN_ON_ONCE(raw_smp_processor_id() != me);
8311fbeb3a8SPaul E. McKenney 
8321fbeb3a8SPaul E. McKenney 	WRITE_ONCE(rt->start_reader, 0);
8331fbeb3a8SPaul E. McKenney 	if (!atomic_dec_return(&n_started))
8341fbeb3a8SPaul E. McKenney 		while (atomic_read_acquire(&n_started))
8351fbeb3a8SPaul E. McKenney 			cpu_relax();
8361fbeb3a8SPaul E. McKenney 
837e76506f0SPaul E. McKenney 	VERBOSE_SCALEOUT_BATCH("ref_scale_reader %ld: experiment %d started", me, exp_idx);
8381fbeb3a8SPaul E. McKenney 
8391fbeb3a8SPaul E. McKenney 
8401fbeb3a8SPaul E. McKenney 	// To reduce noise, do an initial cache-warming invocation, check
8411fbeb3a8SPaul E. McKenney 	// in, and then keep warming until everyone has checked in.
8421fbeb3a8SPaul E. McKenney 	rcu_scale_one_reader();
8431fbeb3a8SPaul E. McKenney 	if (!atomic_dec_return(&n_warmedup))
8441fbeb3a8SPaul E. McKenney 		while (atomic_read_acquire(&n_warmedup))
8451fbeb3a8SPaul E. McKenney 			rcu_scale_one_reader();
8461fbeb3a8SPaul E. McKenney 	// Also keep interrupts disabled.  This also has the effect
8471fbeb3a8SPaul E. McKenney 	// of preventing entries into slow path for rcu_read_unlock().
8481fbeb3a8SPaul E. McKenney 	local_irq_save(flags);
8491fbeb3a8SPaul E. McKenney 	start = ktime_get_mono_fast_ns();
8501fbeb3a8SPaul E. McKenney 
8511fbeb3a8SPaul E. McKenney 	rcu_scale_one_reader();
8521fbeb3a8SPaul E. McKenney 
8531fbeb3a8SPaul E. McKenney 	duration = ktime_get_mono_fast_ns() - start;
8541fbeb3a8SPaul E. McKenney 	local_irq_restore(flags);
8551fbeb3a8SPaul E. McKenney 
8561fbeb3a8SPaul E. McKenney 	rt->last_duration_ns = WARN_ON_ONCE(duration < 0) ? 0 : duration;
8571fbeb3a8SPaul E. McKenney 	// To reduce runtime-skew noise, do maintain-load invocations until
8581fbeb3a8SPaul E. McKenney 	// everyone is done.
8591fbeb3a8SPaul E. McKenney 	if (!atomic_dec_return(&n_cooleddown))
8601fbeb3a8SPaul E. McKenney 		while (atomic_read_acquire(&n_cooleddown))
8611fbeb3a8SPaul E. McKenney 			rcu_scale_one_reader();
8621fbeb3a8SPaul E. McKenney 
8631fbeb3a8SPaul E. McKenney 	if (atomic_dec_and_test(&nreaders_exp))
8641fbeb3a8SPaul E. McKenney 		wake_up(&main_wq);
8651fbeb3a8SPaul E. McKenney 
866e76506f0SPaul E. McKenney 	VERBOSE_SCALEOUT_BATCH("ref_scale_reader %ld: experiment %d ended, (readers remaining=%d)",
8671fbeb3a8SPaul E. McKenney 				me, exp_idx, atomic_read(&nreaders_exp));
8681fbeb3a8SPaul E. McKenney 
8691fbeb3a8SPaul E. McKenney 	if (!torture_must_stop())
8701fbeb3a8SPaul E. McKenney 		goto repeat;
8711fbeb3a8SPaul E. McKenney end:
8721fbeb3a8SPaul E. McKenney 	torture_kthread_stopping("ref_scale_reader");
8731fbeb3a8SPaul E. McKenney 	return 0;
8741fbeb3a8SPaul E. McKenney }
8751fbeb3a8SPaul E. McKenney 
reset_readers(void)8761fbeb3a8SPaul E. McKenney static void reset_readers(void)
8771fbeb3a8SPaul E. McKenney {
8781fbeb3a8SPaul E. McKenney 	int i;
8791fbeb3a8SPaul E. McKenney 	struct reader_task *rt;
8801fbeb3a8SPaul E. McKenney 
8811fbeb3a8SPaul E. McKenney 	for (i = 0; i < nreaders; i++) {
8821fbeb3a8SPaul E. McKenney 		rt = &(reader_tasks[i]);
8831fbeb3a8SPaul E. McKenney 
8841fbeb3a8SPaul E. McKenney 		rt->last_duration_ns = 0;
8851fbeb3a8SPaul E. McKenney 	}
8861fbeb3a8SPaul E. McKenney }
8871fbeb3a8SPaul E. McKenney 
8881fbeb3a8SPaul E. McKenney // Print the results of each reader and return the sum of all their durations.
process_durations(int n)8891fbeb3a8SPaul E. McKenney static u64 process_durations(int n)
8901fbeb3a8SPaul E. McKenney {
8911fbeb3a8SPaul E. McKenney 	int i;
8921fbeb3a8SPaul E. McKenney 	struct reader_task *rt;
8931fbeb3a8SPaul E. McKenney 	char buf1[64];
8941fbeb3a8SPaul E. McKenney 	char *buf;
8951fbeb3a8SPaul E. McKenney 	u64 sum = 0;
8961fbeb3a8SPaul E. McKenney 
8979880eb87SLi Zhijian 	buf = kmalloc(800 + 64, GFP_KERNEL);
8981fbeb3a8SPaul E. McKenney 	if (!buf)
8991fbeb3a8SPaul E. McKenney 		return 0;
9001fbeb3a8SPaul E. McKenney 	buf[0] = 0;
9011fbeb3a8SPaul E. McKenney 	sprintf(buf, "Experiment #%d (Format: <THREAD-NUM>:<Total loop time in ns>)",
9021fbeb3a8SPaul E. McKenney 		exp_idx);
9031fbeb3a8SPaul E. McKenney 
9041fbeb3a8SPaul E. McKenney 	for (i = 0; i < n && !torture_must_stop(); i++) {
9051fbeb3a8SPaul E. McKenney 		rt = &(reader_tasks[i]);
9061fbeb3a8SPaul E. McKenney 		sprintf(buf1, "%d: %llu\t", i, rt->last_duration_ns);
9071fbeb3a8SPaul E. McKenney 
9081fbeb3a8SPaul E. McKenney 		if (i % 5 == 0)
9091fbeb3a8SPaul E. McKenney 			strcat(buf, "\n");
9109880eb87SLi Zhijian 		if (strlen(buf) >= 800) {
9119880eb87SLi Zhijian 			pr_alert("%s", buf);
9129880eb87SLi Zhijian 			buf[0] = 0;
9139880eb87SLi Zhijian 		}
9141fbeb3a8SPaul E. McKenney 		strcat(buf, buf1);
9151fbeb3a8SPaul E. McKenney 
9161fbeb3a8SPaul E. McKenney 		sum += rt->last_duration_ns;
9171fbeb3a8SPaul E. McKenney 	}
9189880eb87SLi Zhijian 	pr_alert("%s\n", buf);
9191fbeb3a8SPaul E. McKenney 
9201fbeb3a8SPaul E. McKenney 	kfree(buf);
9211fbeb3a8SPaul E. McKenney 	return sum;
9221fbeb3a8SPaul E. McKenney }
9231fbeb3a8SPaul E. McKenney 
9241fbeb3a8SPaul E. McKenney // The main_func is the main orchestrator, it performs a bunch of
9251fbeb3a8SPaul E. McKenney // experiments.  For every experiment, it orders all the readers
9261fbeb3a8SPaul E. McKenney // involved to start and waits for them to finish the experiment. It
9271fbeb3a8SPaul E. McKenney // then reads their timestamps and starts the next experiment. Each
9281fbeb3a8SPaul E. McKenney // experiment progresses from 1 concurrent reader to N of them at which
9291fbeb3a8SPaul E. McKenney // point all the timestamps are printed.
main_func(void * arg)9301fbeb3a8SPaul E. McKenney static int main_func(void *arg)
9311fbeb3a8SPaul E. McKenney {
9321fbeb3a8SPaul E. McKenney 	int exp, r;
9331fbeb3a8SPaul E. McKenney 	char buf1[64];
9341fbeb3a8SPaul E. McKenney 	char *buf;
9351fbeb3a8SPaul E. McKenney 	u64 *result_avg;
9361fbeb3a8SPaul E. McKenney 
9371fbeb3a8SPaul E. McKenney 	set_cpus_allowed_ptr(current, cpumask_of(nreaders % nr_cpu_ids));
9381fbeb3a8SPaul E. McKenney 	set_user_nice(current, MAX_NICE);
9391fbeb3a8SPaul E. McKenney 
9401fbeb3a8SPaul E. McKenney 	VERBOSE_SCALEOUT("main_func task started");
9411fbeb3a8SPaul E. McKenney 	result_avg = kzalloc(nruns * sizeof(*result_avg), GFP_KERNEL);
9429880eb87SLi Zhijian 	buf = kzalloc(800 + 64, GFP_KERNEL);
9431fbeb3a8SPaul E. McKenney 	if (!result_avg || !buf) {
9444feeb9d5SLi Zhijian 		SCALEOUT_ERRSTRING("out of memory");
945c30c8763SLi Zhijian 		goto oom_exit;
9461fbeb3a8SPaul E. McKenney 	}
9471fbeb3a8SPaul E. McKenney 	if (holdoff)
9481fbeb3a8SPaul E. McKenney 		schedule_timeout_interruptible(holdoff * HZ);
9491fbeb3a8SPaul E. McKenney 
9501fbeb3a8SPaul E. McKenney 	// Wait for all threads to start.
9511fbeb3a8SPaul E. McKenney 	atomic_inc(&n_init);
9521fbeb3a8SPaul E. McKenney 	while (atomic_read(&n_init) < nreaders + 1)
9531fbeb3a8SPaul E. McKenney 		schedule_timeout_uninterruptible(1);
9541fbeb3a8SPaul E. McKenney 
9551fbeb3a8SPaul E. McKenney 	// Start exp readers up per experiment
9561fbeb3a8SPaul E. McKenney 	for (exp = 0; exp < nruns && !torture_must_stop(); exp++) {
9571fbeb3a8SPaul E. McKenney 		if (torture_must_stop())
9581fbeb3a8SPaul E. McKenney 			goto end;
9591fbeb3a8SPaul E. McKenney 
9601fbeb3a8SPaul E. McKenney 		reset_readers();
9611fbeb3a8SPaul E. McKenney 		atomic_set(&nreaders_exp, nreaders);
9621fbeb3a8SPaul E. McKenney 		atomic_set(&n_started, nreaders);
9631fbeb3a8SPaul E. McKenney 		atomic_set(&n_warmedup, nreaders);
9641fbeb3a8SPaul E. McKenney 		atomic_set(&n_cooleddown, nreaders);
9651fbeb3a8SPaul E. McKenney 
9661fbeb3a8SPaul E. McKenney 		exp_idx = exp;
9671fbeb3a8SPaul E. McKenney 
9681fbeb3a8SPaul E. McKenney 		for (r = 0; r < nreaders; r++) {
9691fbeb3a8SPaul E. McKenney 			smp_store_release(&reader_tasks[r].start_reader, 1);
9701fbeb3a8SPaul E. McKenney 			wake_up(&reader_tasks[r].wq);
9711fbeb3a8SPaul E. McKenney 		}
9721fbeb3a8SPaul E. McKenney 
9731fbeb3a8SPaul E. McKenney 		VERBOSE_SCALEOUT("main_func: experiment started, waiting for %d readers",
9741fbeb3a8SPaul E. McKenney 				nreaders);
9751fbeb3a8SPaul E. McKenney 
9761fbeb3a8SPaul E. McKenney 		wait_event(main_wq,
9771fbeb3a8SPaul E. McKenney 			   !atomic_read(&nreaders_exp) || torture_must_stop());
9781fbeb3a8SPaul E. McKenney 
9791fbeb3a8SPaul E. McKenney 		VERBOSE_SCALEOUT("main_func: experiment ended");
9801fbeb3a8SPaul E. McKenney 
9811fbeb3a8SPaul E. McKenney 		if (torture_must_stop())
9821fbeb3a8SPaul E. McKenney 			goto end;
9831fbeb3a8SPaul E. McKenney 
9841fbeb3a8SPaul E. McKenney 		result_avg[exp] = div_u64(1000 * process_durations(nreaders), nreaders * loops);
9851fbeb3a8SPaul E. McKenney 	}
9861fbeb3a8SPaul E. McKenney 
9871fbeb3a8SPaul E. McKenney 	// Print the average of all experiments
9881fbeb3a8SPaul E. McKenney 	SCALEOUT("END OF TEST. Calculating average duration per loop (nanoseconds)...\n");
9891fbeb3a8SPaul E. McKenney 
9909880eb87SLi Zhijian 	pr_alert("Runs\tTime(ns)\n");
9911fbeb3a8SPaul E. McKenney 	for (exp = 0; exp < nruns; exp++) {
9921fbeb3a8SPaul E. McKenney 		u64 avg;
9931fbeb3a8SPaul E. McKenney 		u32 rem;
9941fbeb3a8SPaul E. McKenney 
9951fbeb3a8SPaul E. McKenney 		avg = div_u64_rem(result_avg[exp], 1000, &rem);
9961fbeb3a8SPaul E. McKenney 		sprintf(buf1, "%d\t%llu.%03u\n", exp + 1, avg, rem);
9971fbeb3a8SPaul E. McKenney 		strcat(buf, buf1);
9989880eb87SLi Zhijian 		if (strlen(buf) >= 800) {
9999880eb87SLi Zhijian 			pr_alert("%s", buf);
10009880eb87SLi Zhijian 			buf[0] = 0;
10019880eb87SLi Zhijian 		}
10021fbeb3a8SPaul E. McKenney 	}
10031fbeb3a8SPaul E. McKenney 
10049880eb87SLi Zhijian 	pr_alert("%s", buf);
10051fbeb3a8SPaul E. McKenney 
1006c30c8763SLi Zhijian oom_exit:
10071fbeb3a8SPaul E. McKenney 	// This will shutdown everything including us.
10081fbeb3a8SPaul E. McKenney 	if (shutdown) {
10091fbeb3a8SPaul E. McKenney 		shutdown_start = 1;
10101fbeb3a8SPaul E. McKenney 		wake_up(&shutdown_wq);
10111fbeb3a8SPaul E. McKenney 	}
10121fbeb3a8SPaul E. McKenney 
10131fbeb3a8SPaul E. McKenney 	// Wait for torture to stop us
10141fbeb3a8SPaul E. McKenney 	while (!torture_must_stop())
10151fbeb3a8SPaul E. McKenney 		schedule_timeout_uninterruptible(1);
10161fbeb3a8SPaul E. McKenney 
10171fbeb3a8SPaul E. McKenney end:
10181fbeb3a8SPaul E. McKenney 	torture_kthread_stopping("main_func");
10191fbeb3a8SPaul E. McKenney 	kfree(result_avg);
10201fbeb3a8SPaul E. McKenney 	kfree(buf);
10211fbeb3a8SPaul E. McKenney 	return 0;
10221fbeb3a8SPaul E. McKenney }
10231fbeb3a8SPaul E. McKenney 
10241fbeb3a8SPaul E. McKenney static void
ref_scale_print_module_parms(struct ref_scale_ops * cur_ops,const char * tag)10251fbeb3a8SPaul E. McKenney ref_scale_print_module_parms(struct ref_scale_ops *cur_ops, const char *tag)
10261fbeb3a8SPaul E. McKenney {
10271fbeb3a8SPaul E. McKenney 	pr_alert("%s" SCALE_FLAG
1028*d6fea1ddSPaul E. McKenney 		 "--- %s:  verbose=%d verbose_batched=%d shutdown=%d holdoff=%d lookup_instances=%ld loops=%ld nreaders=%d nruns=%d readdelay=%d\n", scale_type, tag,
1029*d6fea1ddSPaul E. McKenney 		 verbose, verbose_batched, shutdown, holdoff, lookup_instances, loops, nreaders, nruns, readdelay);
10301fbeb3a8SPaul E. McKenney }
10311fbeb3a8SPaul E. McKenney 
10321fbeb3a8SPaul E. McKenney static void
ref_scale_cleanup(void)10331fbeb3a8SPaul E. McKenney ref_scale_cleanup(void)
10341fbeb3a8SPaul E. McKenney {
10351fbeb3a8SPaul E. McKenney 	int i;
10361fbeb3a8SPaul E. McKenney 
10371fbeb3a8SPaul E. McKenney 	if (torture_cleanup_begin())
10381fbeb3a8SPaul E. McKenney 		return;
10391fbeb3a8SPaul E. McKenney 
10401fbeb3a8SPaul E. McKenney 	if (!cur_ops) {
10411fbeb3a8SPaul E. McKenney 		torture_cleanup_end();
10421fbeb3a8SPaul E. McKenney 		return;
10431fbeb3a8SPaul E. McKenney 	}
10441fbeb3a8SPaul E. McKenney 
10451fbeb3a8SPaul E. McKenney 	if (reader_tasks) {
10461fbeb3a8SPaul E. McKenney 		for (i = 0; i < nreaders; i++)
10471fbeb3a8SPaul E. McKenney 			torture_stop_kthread("ref_scale_reader",
10481fbeb3a8SPaul E. McKenney 					     reader_tasks[i].task);
10491fbeb3a8SPaul E. McKenney 	}
10501fbeb3a8SPaul E. McKenney 	kfree(reader_tasks);
10511fbeb3a8SPaul E. McKenney 
10521fbeb3a8SPaul E. McKenney 	torture_stop_kthread("main_task", main_task);
10531fbeb3a8SPaul E. McKenney 	kfree(main_task);
10541fbeb3a8SPaul E. McKenney 
10551fbeb3a8SPaul E. McKenney 	// Do scale-type-specific cleanup operations.
10561fbeb3a8SPaul E. McKenney 	if (cur_ops->cleanup != NULL)
10571fbeb3a8SPaul E. McKenney 		cur_ops->cleanup();
10581fbeb3a8SPaul E. McKenney 
10591fbeb3a8SPaul E. McKenney 	torture_cleanup_end();
10601fbeb3a8SPaul E. McKenney }
10611fbeb3a8SPaul E. McKenney 
10621fbeb3a8SPaul E. McKenney // Shutdown kthread.  Just waits to be awakened, then shuts down system.
10631fbeb3a8SPaul E. McKenney static int
ref_scale_shutdown(void * arg)10641fbeb3a8SPaul E. McKenney ref_scale_shutdown(void *arg)
10651fbeb3a8SPaul E. McKenney {
10666bc6e6b2SPaul E. McKenney 	wait_event_idle(shutdown_wq, shutdown_start);
10671fbeb3a8SPaul E. McKenney 
10681fbeb3a8SPaul E. McKenney 	smp_mb(); // Wake before output.
10691fbeb3a8SPaul E. McKenney 	ref_scale_cleanup();
10701fbeb3a8SPaul E. McKenney 	kernel_power_off();
10711fbeb3a8SPaul E. McKenney 
10721fbeb3a8SPaul E. McKenney 	return -EINVAL;
10731fbeb3a8SPaul E. McKenney }
10741fbeb3a8SPaul E. McKenney 
10751fbeb3a8SPaul E. McKenney static int __init
ref_scale_init(void)10761fbeb3a8SPaul E. McKenney ref_scale_init(void)
10771fbeb3a8SPaul E. McKenney {
10781fbeb3a8SPaul E. McKenney 	long i;
10791fbeb3a8SPaul E. McKenney 	int firsterr = 0;
10801fbeb3a8SPaul E. McKenney 	static struct ref_scale_ops *scale_ops[] = {
1081dec86781SPaul E. McKenney 		&rcu_ops, &srcu_ops, RCU_TRACE_OPS RCU_TASKS_OPS &refcnt_ops, &rwlock_ops,
1082b5a2801fSPaul E. McKenney 		&rwsem_ops, &lock_ops, &lock_irq_ops, &acqrel_ops, &clock_ops, &jiffies_ops,
1083a6889becSPaul E. McKenney 		&typesafe_ref_ops, &typesafe_lock_ops, &typesafe_seqlock_ops,
10841fbeb3a8SPaul E. McKenney 	};
10851fbeb3a8SPaul E. McKenney 
10861fbeb3a8SPaul E. McKenney 	if (!torture_init_begin(scale_type, verbose))
10871fbeb3a8SPaul E. McKenney 		return -EBUSY;
10881fbeb3a8SPaul E. McKenney 
10891fbeb3a8SPaul E. McKenney 	for (i = 0; i < ARRAY_SIZE(scale_ops); i++) {
10901fbeb3a8SPaul E. McKenney 		cur_ops = scale_ops[i];
10911fbeb3a8SPaul E. McKenney 		if (strcmp(scale_type, cur_ops->name) == 0)
10921fbeb3a8SPaul E. McKenney 			break;
10931fbeb3a8SPaul E. McKenney 	}
10941fbeb3a8SPaul E. McKenney 	if (i == ARRAY_SIZE(scale_ops)) {
10951fbeb3a8SPaul E. McKenney 		pr_alert("rcu-scale: invalid scale type: \"%s\"\n", scale_type);
10961fbeb3a8SPaul E. McKenney 		pr_alert("rcu-scale types:");
10971fbeb3a8SPaul E. McKenney 		for (i = 0; i < ARRAY_SIZE(scale_ops); i++)
10981fbeb3a8SPaul E. McKenney 			pr_cont(" %s", scale_ops[i]->name);
10991fbeb3a8SPaul E. McKenney 		pr_cont("\n");
11001fbeb3a8SPaul E. McKenney 		firsterr = -EINVAL;
11011fbeb3a8SPaul E. McKenney 		cur_ops = NULL;
11021fbeb3a8SPaul E. McKenney 		goto unwind;
11031fbeb3a8SPaul E. McKenney 	}
11041fbeb3a8SPaul E. McKenney 	if (cur_ops->init)
11053c6496c8SPaul E. McKenney 		if (!cur_ops->init()) {
11063c6496c8SPaul E. McKenney 			firsterr = -EUCLEAN;
11073c6496c8SPaul E. McKenney 			goto unwind;
11083c6496c8SPaul E. McKenney 		}
11091fbeb3a8SPaul E. McKenney 
11101fbeb3a8SPaul E. McKenney 	ref_scale_print_module_parms(cur_ops, "Start of test");
11111fbeb3a8SPaul E. McKenney 
11121fbeb3a8SPaul E. McKenney 	// Shutdown task
11131fbeb3a8SPaul E. McKenney 	if (shutdown) {
11141fbeb3a8SPaul E. McKenney 		init_waitqueue_head(&shutdown_wq);
11151fbeb3a8SPaul E. McKenney 		firsterr = torture_create_kthread(ref_scale_shutdown, NULL,
11161fbeb3a8SPaul E. McKenney 						  shutdown_task);
1117ed60ad73SPaul E. McKenney 		if (torture_init_error(firsterr))
11181fbeb3a8SPaul E. McKenney 			goto unwind;
11191fbeb3a8SPaul E. McKenney 		schedule_timeout_uninterruptible(1);
11201fbeb3a8SPaul E. McKenney 	}
11211fbeb3a8SPaul E. McKenney 
11221fbeb3a8SPaul E. McKenney 	// Reader tasks (default to ~75% of online CPUs).
11231fbeb3a8SPaul E. McKenney 	if (nreaders < 0)
11241fbeb3a8SPaul E. McKenney 		nreaders = (num_online_cpus() >> 1) + (num_online_cpus() >> 2);
11250c6d18d8SPaul E. McKenney 	if (WARN_ONCE(loops <= 0, "%s: loops = %ld, adjusted to 1\n", __func__, loops))
11260c6d18d8SPaul E. McKenney 		loops = 1;
11270c6d18d8SPaul E. McKenney 	if (WARN_ONCE(nreaders <= 0, "%s: nreaders = %d, adjusted to 1\n", __func__, nreaders))
11280c6d18d8SPaul E. McKenney 		nreaders = 1;
11290c6d18d8SPaul E. McKenney 	if (WARN_ONCE(nruns <= 0, "%s: nruns = %d, adjusted to 1\n", __func__, nruns))
11300c6d18d8SPaul E. McKenney 		nruns = 1;
11311fbeb3a8SPaul E. McKenney 	reader_tasks = kcalloc(nreaders, sizeof(reader_tasks[0]),
11321fbeb3a8SPaul E. McKenney 			       GFP_KERNEL);
11331fbeb3a8SPaul E. McKenney 	if (!reader_tasks) {
11344feeb9d5SLi Zhijian 		SCALEOUT_ERRSTRING("out of memory");
11351fbeb3a8SPaul E. McKenney 		firsterr = -ENOMEM;
11361fbeb3a8SPaul E. McKenney 		goto unwind;
11371fbeb3a8SPaul E. McKenney 	}
11381fbeb3a8SPaul E. McKenney 
1139f71f22b6SLi Zhijian 	VERBOSE_SCALEOUT("Starting %d reader threads", nreaders);
11401fbeb3a8SPaul E. McKenney 
11411fbeb3a8SPaul E. McKenney 	for (i = 0; i < nreaders; i++) {
1142f5063e89SWaiman Long 		init_waitqueue_head(&reader_tasks[i].wq);
11431fbeb3a8SPaul E. McKenney 		firsterr = torture_create_kthread(ref_scale_reader, (void *)i,
11441fbeb3a8SPaul E. McKenney 						  reader_tasks[i].task);
1145ed60ad73SPaul E. McKenney 		if (torture_init_error(firsterr))
11461fbeb3a8SPaul E. McKenney 			goto unwind;
11471fbeb3a8SPaul E. McKenney 	}
11481fbeb3a8SPaul E. McKenney 
11491fbeb3a8SPaul E. McKenney 	// Main Task
11501fbeb3a8SPaul E. McKenney 	init_waitqueue_head(&main_wq);
11511fbeb3a8SPaul E. McKenney 	firsterr = torture_create_kthread(main_func, NULL, main_task);
1152ed60ad73SPaul E. McKenney 	if (torture_init_error(firsterr))
11531fbeb3a8SPaul E. McKenney 		goto unwind;
11541fbeb3a8SPaul E. McKenney 
11551fbeb3a8SPaul E. McKenney 	torture_init_end();
11561fbeb3a8SPaul E. McKenney 	return 0;
11571fbeb3a8SPaul E. McKenney 
11581fbeb3a8SPaul E. McKenney unwind:
11591fbeb3a8SPaul E. McKenney 	torture_init_end();
11601fbeb3a8SPaul E. McKenney 	ref_scale_cleanup();
1161bc80d353SPaul E. McKenney 	if (shutdown) {
1162bc80d353SPaul E. McKenney 		WARN_ON(!IS_MODULE(CONFIG_RCU_REF_SCALE_TEST));
1163bc80d353SPaul E. McKenney 		kernel_power_off();
1164bc80d353SPaul E. McKenney 	}
11651fbeb3a8SPaul E. McKenney 	return firsterr;
11661fbeb3a8SPaul E. McKenney }
11671fbeb3a8SPaul E. McKenney 
11681fbeb3a8SPaul E. McKenney module_init(ref_scale_init);
11691fbeb3a8SPaul E. McKenney module_exit(ref_scale_cleanup);
1170