11f87dcf1SSong Liu // SPDX-License-Identifier: GPL-2.0
21f87dcf1SSong Liu /* Copyright (c) 2021 Facebook */
31f87dcf1SSong Liu 
41f87dcf1SSong Liu #define _GNU_SOURCE         /* See feature_test_macros(7) */
51f87dcf1SSong Liu #include <unistd.h>
60334b4d8SMartin KaFai Lau #include <sched.h>
70334b4d8SMartin KaFai Lau #include <pthread.h>
81f87dcf1SSong Liu #include <sys/syscall.h>   /* For SYS_xxx definitions */
91f87dcf1SSong Liu #include <sys/types.h>
101f87dcf1SSong Liu #include <test_progs.h>
11387b5321SMartin KaFai Lau #include "task_local_storage_helpers.h"
121f87dcf1SSong Liu #include "task_local_storage.skel.h"
131f87dcf1SSong Liu #include "task_local_storage_exit_creds.skel.h"
14c540957aSSong Liu #include "task_ls_recursion.skel.h"
150334b4d8SMartin KaFai Lau #include "task_storage_nodeadlock.skel.h"
161f87dcf1SSong Liu 
test_sys_enter_exit(void)171f87dcf1SSong Liu static void test_sys_enter_exit(void)
181f87dcf1SSong Liu {
191f87dcf1SSong Liu 	struct task_local_storage *skel;
201f87dcf1SSong Liu 	int err;
211f87dcf1SSong Liu 
221f87dcf1SSong Liu 	skel = task_local_storage__open_and_load();
231f87dcf1SSong Liu 	if (!ASSERT_OK_PTR(skel, "skel_open_and_load"))
241f87dcf1SSong Liu 		return;
251f87dcf1SSong Liu 
261f87dcf1SSong Liu 	skel->bss->target_pid = syscall(SYS_gettid);
271f87dcf1SSong Liu 
281f87dcf1SSong Liu 	err = task_local_storage__attach(skel);
291f87dcf1SSong Liu 	if (!ASSERT_OK(err, "skel_attach"))
301f87dcf1SSong Liu 		goto out;
311f87dcf1SSong Liu 
321f87dcf1SSong Liu 	syscall(SYS_gettid);
331f87dcf1SSong Liu 	syscall(SYS_gettid);
341f87dcf1SSong Liu 
351f87dcf1SSong Liu 	/* 3x syscalls: 1x attach and 2x gettid */
361f87dcf1SSong Liu 	ASSERT_EQ(skel->bss->enter_cnt, 3, "enter_cnt");
371f87dcf1SSong Liu 	ASSERT_EQ(skel->bss->exit_cnt, 3, "exit_cnt");
381f87dcf1SSong Liu 	ASSERT_EQ(skel->bss->mismatch_cnt, 0, "mismatch_cnt");
391f87dcf1SSong Liu out:
401f87dcf1SSong Liu 	task_local_storage__destroy(skel);
411f87dcf1SSong Liu }
421f87dcf1SSong Liu 
test_exit_creds(void)431f87dcf1SSong Liu static void test_exit_creds(void)
441f87dcf1SSong Liu {
451f87dcf1SSong Liu 	struct task_local_storage_exit_creds *skel;
46eb814cf1SDelyan Kratunov 	int err, run_count, sync_rcu_calls = 0;
47eb814cf1SDelyan Kratunov 	const int MAX_SYNC_RCU_CALLS = 1000;
481f87dcf1SSong Liu 
491f87dcf1SSong Liu 	skel = task_local_storage_exit_creds__open_and_load();
501f87dcf1SSong Liu 	if (!ASSERT_OK_PTR(skel, "skel_open_and_load"))
511f87dcf1SSong Liu 		return;
521f87dcf1SSong Liu 
531f87dcf1SSong Liu 	err = task_local_storage_exit_creds__attach(skel);
541f87dcf1SSong Liu 	if (!ASSERT_OK(err, "skel_attach"))
551f87dcf1SSong Liu 		goto out;
561f87dcf1SSong Liu 
571f87dcf1SSong Liu 	/* trigger at least one exit_creds() */
581f87dcf1SSong Liu 	if (CHECK_FAIL(system("ls > /dev/null")))
591f87dcf1SSong Liu 		goto out;
601f87dcf1SSong Liu 
61eb814cf1SDelyan Kratunov 	/* kern_sync_rcu is not enough on its own as the read section we want
62eb814cf1SDelyan Kratunov 	 * to wait for may start after we enter synchronize_rcu, so our call
63eb814cf1SDelyan Kratunov 	 * won't wait for the section to finish. Loop on the run counter
64eb814cf1SDelyan Kratunov 	 * as well to ensure the program has run.
65eb814cf1SDelyan Kratunov 	 */
66eb814cf1SDelyan Kratunov 	do {
671f87dcf1SSong Liu 		kern_sync_rcu();
68eb814cf1SDelyan Kratunov 		run_count = __atomic_load_n(&skel->bss->run_count, __ATOMIC_SEQ_CST);
69eb814cf1SDelyan Kratunov 	} while (run_count == 0 && ++sync_rcu_calls < MAX_SYNC_RCU_CALLS);
70eb814cf1SDelyan Kratunov 
71eb814cf1SDelyan Kratunov 	ASSERT_NEQ(sync_rcu_calls, MAX_SYNC_RCU_CALLS,
72eb814cf1SDelyan Kratunov 		   "sync_rcu count too high");
73eb814cf1SDelyan Kratunov 	ASSERT_NEQ(run_count, 0, "run_count");
741f87dcf1SSong Liu 	ASSERT_EQ(skel->bss->valid_ptr_count, 0, "valid_ptr_count");
751f87dcf1SSong Liu 	ASSERT_NEQ(skel->bss->null_ptr_count, 0, "null_ptr_count");
761f87dcf1SSong Liu out:
771f87dcf1SSong Liu 	task_local_storage_exit_creds__destroy(skel);
781f87dcf1SSong Liu }
791f87dcf1SSong Liu 
test_recursion(void)80c540957aSSong Liu static void test_recursion(void)
81c540957aSSong Liu {
82387b5321SMartin KaFai Lau 	int err, map_fd, prog_fd, task_fd;
83c540957aSSong Liu 	struct task_ls_recursion *skel;
84387b5321SMartin KaFai Lau 	struct bpf_prog_info info;
85387b5321SMartin KaFai Lau 	__u32 info_len = sizeof(info);
86387b5321SMartin KaFai Lau 	long value;
87387b5321SMartin KaFai Lau 
88387b5321SMartin KaFai Lau 	task_fd = sys_pidfd_open(getpid(), 0);
89387b5321SMartin KaFai Lau 	if (!ASSERT_NEQ(task_fd, -1, "sys_pidfd_open"))
90387b5321SMartin KaFai Lau 		return;
91c540957aSSong Liu 
92c540957aSSong Liu 	skel = task_ls_recursion__open_and_load();
93c540957aSSong Liu 	if (!ASSERT_OK_PTR(skel, "skel_open_and_load"))
94387b5321SMartin KaFai Lau 		goto out;
95c540957aSSong Liu 
96c540957aSSong Liu 	err = task_ls_recursion__attach(skel);
97c540957aSSong Liu 	if (!ASSERT_OK(err, "skel_attach"))
98c540957aSSong Liu 		goto out;
99c540957aSSong Liu 
100c540957aSSong Liu 	/* trigger sys_enter, make sure it does not cause deadlock */
101387b5321SMartin KaFai Lau 	skel->bss->test_pid = getpid();
102c540957aSSong Liu 	syscall(SYS_gettid);
103387b5321SMartin KaFai Lau 	skel->bss->test_pid = 0;
104387b5321SMartin KaFai Lau 	task_ls_recursion__detach(skel);
105387b5321SMartin KaFai Lau 
106387b5321SMartin KaFai Lau 	/* Refer to the comment in BPF_PROG(on_update) for
107387b5321SMartin KaFai Lau 	 * the explanation on the value 201 and 100.
108387b5321SMartin KaFai Lau 	 */
109387b5321SMartin KaFai Lau 	map_fd = bpf_map__fd(skel->maps.map_a);
110387b5321SMartin KaFai Lau 	err = bpf_map_lookup_elem(map_fd, &task_fd, &value);
111387b5321SMartin KaFai Lau 	ASSERT_OK(err, "lookup map_a");
112387b5321SMartin KaFai Lau 	ASSERT_EQ(value, 201, "map_a value");
113387b5321SMartin KaFai Lau 	ASSERT_EQ(skel->bss->nr_del_errs, 1, "bpf_task_storage_delete busy");
114387b5321SMartin KaFai Lau 
115387b5321SMartin KaFai Lau 	map_fd = bpf_map__fd(skel->maps.map_b);
116387b5321SMartin KaFai Lau 	err = bpf_map_lookup_elem(map_fd, &task_fd, &value);
117387b5321SMartin KaFai Lau 	ASSERT_OK(err, "lookup map_b");
118387b5321SMartin KaFai Lau 	ASSERT_EQ(value, 100, "map_b value");
119387b5321SMartin KaFai Lau 
120387b5321SMartin KaFai Lau 	prog_fd = bpf_program__fd(skel->progs.on_update);
121387b5321SMartin KaFai Lau 	memset(&info, 0, sizeof(info));
122*c5a237a4SIlya Leoshkevich 	err = bpf_prog_get_info_by_fd(prog_fd, &info, &info_len);
123387b5321SMartin KaFai Lau 	ASSERT_OK(err, "get prog info");
124387b5321SMartin KaFai Lau 	ASSERT_EQ(info.recursion_misses, 0, "on_update prog recursion");
125387b5321SMartin KaFai Lau 
126387b5321SMartin KaFai Lau 	prog_fd = bpf_program__fd(skel->progs.on_enter);
127387b5321SMartin KaFai Lau 	memset(&info, 0, sizeof(info));
128*c5a237a4SIlya Leoshkevich 	err = bpf_prog_get_info_by_fd(prog_fd, &info, &info_len);
129387b5321SMartin KaFai Lau 	ASSERT_OK(err, "get prog info");
130387b5321SMartin KaFai Lau 	ASSERT_EQ(info.recursion_misses, 0, "on_enter prog recursion");
131c540957aSSong Liu 
132c540957aSSong Liu out:
133387b5321SMartin KaFai Lau 	close(task_fd);
134c540957aSSong Liu 	task_ls_recursion__destroy(skel);
135c540957aSSong Liu }
136c540957aSSong Liu 
1370334b4d8SMartin KaFai Lau static bool stop;
1380334b4d8SMartin KaFai Lau 
waitall(const pthread_t * tids,int nr)1390334b4d8SMartin KaFai Lau static void waitall(const pthread_t *tids, int nr)
1400334b4d8SMartin KaFai Lau {
1410334b4d8SMartin KaFai Lau 	int i;
1420334b4d8SMartin KaFai Lau 
1430334b4d8SMartin KaFai Lau 	stop = true;
1440334b4d8SMartin KaFai Lau 	for (i = 0; i < nr; i++)
1450334b4d8SMartin KaFai Lau 		pthread_join(tids[i], NULL);
1460334b4d8SMartin KaFai Lau }
1470334b4d8SMartin KaFai Lau 
sock_create_loop(void * arg)1480334b4d8SMartin KaFai Lau static void *sock_create_loop(void *arg)
1490334b4d8SMartin KaFai Lau {
1500334b4d8SMartin KaFai Lau 	struct task_storage_nodeadlock *skel = arg;
1510334b4d8SMartin KaFai Lau 	int fd;
1520334b4d8SMartin KaFai Lau 
1530334b4d8SMartin KaFai Lau 	while (!stop) {
1540334b4d8SMartin KaFai Lau 		fd = socket(AF_INET, SOCK_STREAM, 0);
1550334b4d8SMartin KaFai Lau 		close(fd);
1560334b4d8SMartin KaFai Lau 		if (skel->bss->nr_get_errs || skel->bss->nr_del_errs)
1570334b4d8SMartin KaFai Lau 			stop = true;
1580334b4d8SMartin KaFai Lau 	}
1590334b4d8SMartin KaFai Lau 
1600334b4d8SMartin KaFai Lau 	return NULL;
1610334b4d8SMartin KaFai Lau }
1620334b4d8SMartin KaFai Lau 
test_nodeadlock(void)1630334b4d8SMartin KaFai Lau static void test_nodeadlock(void)
1640334b4d8SMartin KaFai Lau {
1650334b4d8SMartin KaFai Lau 	struct task_storage_nodeadlock *skel;
1660334b4d8SMartin KaFai Lau 	struct bpf_prog_info info = {};
1670334b4d8SMartin KaFai Lau 	__u32 info_len = sizeof(info);
1680334b4d8SMartin KaFai Lau 	const int nr_threads = 32;
1690334b4d8SMartin KaFai Lau 	pthread_t tids[nr_threads];
1700334b4d8SMartin KaFai Lau 	int i, prog_fd, err;
1710334b4d8SMartin KaFai Lau 	cpu_set_t old, new;
1720334b4d8SMartin KaFai Lau 
1730334b4d8SMartin KaFai Lau 	/* Pin all threads to one cpu to increase the chance of preemption
1740334b4d8SMartin KaFai Lau 	 * in a sleepable bpf prog.
1750334b4d8SMartin KaFai Lau 	 */
1760334b4d8SMartin KaFai Lau 	CPU_ZERO(&new);
1770334b4d8SMartin KaFai Lau 	CPU_SET(0, &new);
1780334b4d8SMartin KaFai Lau 	err = sched_getaffinity(getpid(), sizeof(old), &old);
1790334b4d8SMartin KaFai Lau 	if (!ASSERT_OK(err, "getaffinity"))
1800334b4d8SMartin KaFai Lau 		return;
1810334b4d8SMartin KaFai Lau 	err = sched_setaffinity(getpid(), sizeof(new), &new);
1820334b4d8SMartin KaFai Lau 	if (!ASSERT_OK(err, "setaffinity"))
1830334b4d8SMartin KaFai Lau 		return;
1840334b4d8SMartin KaFai Lau 
1850334b4d8SMartin KaFai Lau 	skel = task_storage_nodeadlock__open_and_load();
1860334b4d8SMartin KaFai Lau 	if (!ASSERT_OK_PTR(skel, "open_and_load"))
1870334b4d8SMartin KaFai Lau 		goto done;
1880334b4d8SMartin KaFai Lau 
1890334b4d8SMartin KaFai Lau 	/* Unnecessary recursion and deadlock detection are reproducible
1900334b4d8SMartin KaFai Lau 	 * in the preemptible kernel.
1910334b4d8SMartin KaFai Lau 	 */
1920334b4d8SMartin KaFai Lau 	if (!skel->kconfig->CONFIG_PREEMPT) {
1930334b4d8SMartin KaFai Lau 		test__skip();
1940334b4d8SMartin KaFai Lau 		goto done;
1950334b4d8SMartin KaFai Lau 	}
1960334b4d8SMartin KaFai Lau 
1970334b4d8SMartin KaFai Lau 	err = task_storage_nodeadlock__attach(skel);
1980334b4d8SMartin KaFai Lau 	ASSERT_OK(err, "attach prog");
1990334b4d8SMartin KaFai Lau 
2000334b4d8SMartin KaFai Lau 	for (i = 0; i < nr_threads; i++) {
2010334b4d8SMartin KaFai Lau 		err = pthread_create(&tids[i], NULL, sock_create_loop, skel);
2020334b4d8SMartin KaFai Lau 		if (err) {
2030334b4d8SMartin KaFai Lau 			/* Only assert once here to avoid excessive
2040334b4d8SMartin KaFai Lau 			 * PASS printing during test failure.
2050334b4d8SMartin KaFai Lau 			 */
2060334b4d8SMartin KaFai Lau 			ASSERT_OK(err, "pthread_create");
2070334b4d8SMartin KaFai Lau 			waitall(tids, i);
2080334b4d8SMartin KaFai Lau 			goto done;
2090334b4d8SMartin KaFai Lau 		}
2100334b4d8SMartin KaFai Lau 	}
2110334b4d8SMartin KaFai Lau 
2120334b4d8SMartin KaFai Lau 	/* With 32 threads, 1s is enough to reproduce the issue */
2130334b4d8SMartin KaFai Lau 	sleep(1);
2140334b4d8SMartin KaFai Lau 	waitall(tids, nr_threads);
2150334b4d8SMartin KaFai Lau 
2160334b4d8SMartin KaFai Lau 	info_len = sizeof(info);
2170334b4d8SMartin KaFai Lau 	prog_fd = bpf_program__fd(skel->progs.socket_post_create);
218*c5a237a4SIlya Leoshkevich 	err = bpf_prog_get_info_by_fd(prog_fd, &info, &info_len);
2190334b4d8SMartin KaFai Lau 	ASSERT_OK(err, "get prog info");
2200334b4d8SMartin KaFai Lau 	ASSERT_EQ(info.recursion_misses, 0, "prog recursion");
2210334b4d8SMartin KaFai Lau 
2220334b4d8SMartin KaFai Lau 	ASSERT_EQ(skel->bss->nr_get_errs, 0, "bpf_task_storage_get busy");
2230334b4d8SMartin KaFai Lau 	ASSERT_EQ(skel->bss->nr_del_errs, 0, "bpf_task_storage_delete busy");
2240334b4d8SMartin KaFai Lau 
2250334b4d8SMartin KaFai Lau done:
2260334b4d8SMartin KaFai Lau 	task_storage_nodeadlock__destroy(skel);
2270334b4d8SMartin KaFai Lau 	sched_setaffinity(getpid(), sizeof(old), &old);
2280334b4d8SMartin KaFai Lau }
2290334b4d8SMartin KaFai Lau 
test_task_local_storage(void)2301f87dcf1SSong Liu void test_task_local_storage(void)
2311f87dcf1SSong Liu {
2321f87dcf1SSong Liu 	if (test__start_subtest("sys_enter_exit"))
2331f87dcf1SSong Liu 		test_sys_enter_exit();
2341f87dcf1SSong Liu 	if (test__start_subtest("exit_creds"))
2351f87dcf1SSong Liu 		test_exit_creds();
236c540957aSSong Liu 	if (test__start_subtest("recursion"))
237c540957aSSong Liu 		test_recursion();
2380334b4d8SMartin KaFai Lau 	if (test__start_subtest("nodeadlock"))
2390334b4d8SMartin KaFai Lau 		test_nodeadlock();
2401f87dcf1SSong Liu }
241