xref: /linux/include/linux/thread_info.h (revision af0a76e1)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 /* thread_info.h: common low-level thread information accessors
3  *
4  * Copyright (C) 2002  David Howells (dhowells@redhat.com)
5  * - Incorporating suggestions made by Linus Torvalds
6  */
7 
8 #ifndef _LINUX_THREAD_INFO_H
9 #define _LINUX_THREAD_INFO_H
10 
11 #include <linux/types.h>
12 #include <linux/limits.h>
13 #include <linux/bug.h>
14 #include <linux/restart_block.h>
15 #include <linux/errno.h>
16 
17 #ifdef CONFIG_THREAD_INFO_IN_TASK
18 /*
19  * For CONFIG_THREAD_INFO_IN_TASK kernels we need <asm/current.h> for the
20  * definition of current, but for !CONFIG_THREAD_INFO_IN_TASK kernels,
21  * including <asm/current.h> can cause a circular dependency on some platforms.
22  */
23 #include <asm/current.h>
24 #define current_thread_info() ((struct thread_info *)current)
25 #endif
26 
27 #include <linux/bitops.h>
28 
29 /*
30  * For per-arch arch_within_stack_frames() implementations, defined in
31  * asm/thread_info.h.
32  */
33 enum {
34 	BAD_STACK = -1,
35 	NOT_STACK = 0,
36 	GOOD_FRAME,
37 	GOOD_STACK,
38 };
39 
40 #ifdef CONFIG_GENERIC_ENTRY
41 enum syscall_work_bit {
42 	SYSCALL_WORK_BIT_SECCOMP,
43 	SYSCALL_WORK_BIT_SYSCALL_TRACEPOINT,
44 	SYSCALL_WORK_BIT_SYSCALL_TRACE,
45 	SYSCALL_WORK_BIT_SYSCALL_EMU,
46 	SYSCALL_WORK_BIT_SYSCALL_AUDIT,
47 	SYSCALL_WORK_BIT_SYSCALL_USER_DISPATCH,
48 	SYSCALL_WORK_BIT_SYSCALL_EXIT_TRAP,
49 };
50 
51 #define SYSCALL_WORK_SECCOMP		BIT(SYSCALL_WORK_BIT_SECCOMP)
52 #define SYSCALL_WORK_SYSCALL_TRACEPOINT	BIT(SYSCALL_WORK_BIT_SYSCALL_TRACEPOINT)
53 #define SYSCALL_WORK_SYSCALL_TRACE	BIT(SYSCALL_WORK_BIT_SYSCALL_TRACE)
54 #define SYSCALL_WORK_SYSCALL_EMU	BIT(SYSCALL_WORK_BIT_SYSCALL_EMU)
55 #define SYSCALL_WORK_SYSCALL_AUDIT	BIT(SYSCALL_WORK_BIT_SYSCALL_AUDIT)
56 #define SYSCALL_WORK_SYSCALL_USER_DISPATCH BIT(SYSCALL_WORK_BIT_SYSCALL_USER_DISPATCH)
57 #define SYSCALL_WORK_SYSCALL_EXIT_TRAP	BIT(SYSCALL_WORK_BIT_SYSCALL_EXIT_TRAP)
58 #endif
59 
60 #include <asm/thread_info.h>
61 
62 #ifdef __KERNEL__
63 
64 #ifndef arch_set_restart_data
65 #define arch_set_restart_data(restart) do { } while (0)
66 #endif
67 
set_restart_fn(struct restart_block * restart,long (* fn)(struct restart_block *))68 static inline long set_restart_fn(struct restart_block *restart,
69 					long (*fn)(struct restart_block *))
70 {
71 	restart->fn = fn;
72 	arch_set_restart_data(restart);
73 	return -ERESTART_RESTARTBLOCK;
74 }
75 
76 #ifndef THREAD_ALIGN
77 #define THREAD_ALIGN	THREAD_SIZE
78 #endif
79 
80 #define THREADINFO_GFP		(GFP_KERNEL_ACCOUNT | __GFP_ZERO)
81 
82 /*
83  * flag set/clear/test wrappers
84  * - pass TIF_xxxx constants to these functions
85  */
86 
set_ti_thread_flag(struct thread_info * ti,int flag)87 static inline void set_ti_thread_flag(struct thread_info *ti, int flag)
88 {
89 	set_bit(flag, (unsigned long *)&ti->flags);
90 }
91 
clear_ti_thread_flag(struct thread_info * ti,int flag)92 static inline void clear_ti_thread_flag(struct thread_info *ti, int flag)
93 {
94 	clear_bit(flag, (unsigned long *)&ti->flags);
95 }
96 
update_ti_thread_flag(struct thread_info * ti,int flag,bool value)97 static inline void update_ti_thread_flag(struct thread_info *ti, int flag,
98 					 bool value)
99 {
100 	if (value)
101 		set_ti_thread_flag(ti, flag);
102 	else
103 		clear_ti_thread_flag(ti, flag);
104 }
105 
test_and_set_ti_thread_flag(struct thread_info * ti,int flag)106 static inline int test_and_set_ti_thread_flag(struct thread_info *ti, int flag)
107 {
108 	return test_and_set_bit(flag, (unsigned long *)&ti->flags);
109 }
110 
test_and_clear_ti_thread_flag(struct thread_info * ti,int flag)111 static inline int test_and_clear_ti_thread_flag(struct thread_info *ti, int flag)
112 {
113 	return test_and_clear_bit(flag, (unsigned long *)&ti->flags);
114 }
115 
test_ti_thread_flag(struct thread_info * ti,int flag)116 static inline int test_ti_thread_flag(struct thread_info *ti, int flag)
117 {
118 	return test_bit(flag, (unsigned long *)&ti->flags);
119 }
120 
121 /*
122  * This may be used in noinstr code, and needs to be __always_inline to prevent
123  * inadvertent instrumentation.
124  */
read_ti_thread_flags(struct thread_info * ti)125 static __always_inline unsigned long read_ti_thread_flags(struct thread_info *ti)
126 {
127 	return READ_ONCE(ti->flags);
128 }
129 
130 #define set_thread_flag(flag) \
131 	set_ti_thread_flag(current_thread_info(), flag)
132 #define clear_thread_flag(flag) \
133 	clear_ti_thread_flag(current_thread_info(), flag)
134 #define update_thread_flag(flag, value) \
135 	update_ti_thread_flag(current_thread_info(), flag, value)
136 #define test_and_set_thread_flag(flag) \
137 	test_and_set_ti_thread_flag(current_thread_info(), flag)
138 #define test_and_clear_thread_flag(flag) \
139 	test_and_clear_ti_thread_flag(current_thread_info(), flag)
140 #define test_thread_flag(flag) \
141 	test_ti_thread_flag(current_thread_info(), flag)
142 #define read_thread_flags() \
143 	read_ti_thread_flags(current_thread_info())
144 
145 #define read_task_thread_flags(t) \
146 	read_ti_thread_flags(task_thread_info(t))
147 
148 #ifdef CONFIG_GENERIC_ENTRY
149 #define set_syscall_work(fl) \
150 	set_bit(SYSCALL_WORK_BIT_##fl, &current_thread_info()->syscall_work)
151 #define test_syscall_work(fl) \
152 	test_bit(SYSCALL_WORK_BIT_##fl, &current_thread_info()->syscall_work)
153 #define clear_syscall_work(fl) \
154 	clear_bit(SYSCALL_WORK_BIT_##fl, &current_thread_info()->syscall_work)
155 
156 #define set_task_syscall_work(t, fl) \
157 	set_bit(SYSCALL_WORK_BIT_##fl, &task_thread_info(t)->syscall_work)
158 #define test_task_syscall_work(t, fl) \
159 	test_bit(SYSCALL_WORK_BIT_##fl, &task_thread_info(t)->syscall_work)
160 #define clear_task_syscall_work(t, fl) \
161 	clear_bit(SYSCALL_WORK_BIT_##fl, &task_thread_info(t)->syscall_work)
162 
163 #else /* CONFIG_GENERIC_ENTRY */
164 
165 #define set_syscall_work(fl)						\
166 	set_ti_thread_flag(current_thread_info(), TIF_##fl)
167 #define test_syscall_work(fl) \
168 	test_ti_thread_flag(current_thread_info(), TIF_##fl)
169 #define clear_syscall_work(fl) \
170 	clear_ti_thread_flag(current_thread_info(), TIF_##fl)
171 
172 #define set_task_syscall_work(t, fl) \
173 	set_ti_thread_flag(task_thread_info(t), TIF_##fl)
174 #define test_task_syscall_work(t, fl) \
175 	test_ti_thread_flag(task_thread_info(t), TIF_##fl)
176 #define clear_task_syscall_work(t, fl) \
177 	clear_ti_thread_flag(task_thread_info(t), TIF_##fl)
178 #endif /* !CONFIG_GENERIC_ENTRY */
179 
180 #ifdef _ASM_GENERIC_BITOPS_INSTRUMENTED_NON_ATOMIC_H
181 
tif_need_resched(void)182 static __always_inline bool tif_need_resched(void)
183 {
184 	return arch_test_bit(TIF_NEED_RESCHED,
185 			     (unsigned long *)(&current_thread_info()->flags));
186 }
187 
188 #else
189 
tif_need_resched(void)190 static __always_inline bool tif_need_resched(void)
191 {
192 	return test_bit(TIF_NEED_RESCHED,
193 			(unsigned long *)(&current_thread_info()->flags));
194 }
195 
196 #endif /* _ASM_GENERIC_BITOPS_INSTRUMENTED_NON_ATOMIC_H */
197 
198 #ifndef CONFIG_HAVE_ARCH_WITHIN_STACK_FRAMES
arch_within_stack_frames(const void * const stack,const void * const stackend,const void * obj,unsigned long len)199 static inline int arch_within_stack_frames(const void * const stack,
200 					   const void * const stackend,
201 					   const void *obj, unsigned long len)
202 {
203 	return 0;
204 }
205 #endif
206 
207 #ifdef CONFIG_HARDENED_USERCOPY
208 extern void __check_object_size(const void *ptr, unsigned long n,
209 					bool to_user);
210 
check_object_size(const void * ptr,unsigned long n,bool to_user)211 static __always_inline void check_object_size(const void *ptr, unsigned long n,
212 					      bool to_user)
213 {
214 	if (!__builtin_constant_p(n))
215 		__check_object_size(ptr, n, to_user);
216 }
217 #else
check_object_size(const void * ptr,unsigned long n,bool to_user)218 static inline void check_object_size(const void *ptr, unsigned long n,
219 				     bool to_user)
220 { }
221 #endif /* CONFIG_HARDENED_USERCOPY */
222 
223 extern void __compiletime_error("copy source size is too small")
224 __bad_copy_from(void);
225 extern void __compiletime_error("copy destination size is too small")
226 __bad_copy_to(void);
227 
228 void __copy_overflow(int size, unsigned long count);
229 
copy_overflow(int size,unsigned long count)230 static inline void copy_overflow(int size, unsigned long count)
231 {
232 	if (IS_ENABLED(CONFIG_BUG))
233 		__copy_overflow(size, count);
234 }
235 
236 static __always_inline __must_check bool
check_copy_size(const void * addr,size_t bytes,bool is_source)237 check_copy_size(const void *addr, size_t bytes, bool is_source)
238 {
239 	int sz = __builtin_object_size(addr, 0);
240 	if (unlikely(sz >= 0 && sz < bytes)) {
241 		if (!__builtin_constant_p(bytes))
242 			copy_overflow(sz, bytes);
243 		else if (is_source)
244 			__bad_copy_from();
245 		else
246 			__bad_copy_to();
247 		return false;
248 	}
249 	if (WARN_ON_ONCE(bytes > INT_MAX))
250 		return false;
251 	check_object_size(addr, bytes, is_source);
252 	return true;
253 }
254 
255 #ifndef arch_setup_new_exec
arch_setup_new_exec(void)256 static inline void arch_setup_new_exec(void) { }
257 #endif
258 
259 void arch_task_cache_init(void); /* for CONFIG_SH */
260 void arch_release_task_struct(struct task_struct *tsk);
261 int arch_dup_task_struct(struct task_struct *dst,
262 				struct task_struct *src);
263 
264 #endif	/* __KERNEL__ */
265 
266 #endif /* _LINUX_THREAD_INFO_H */
267