xref: /qemu/util/coroutine-ucontext.c (revision 73b49878)
1 /*
2  * ucontext coroutine initialization code
3  *
4  * Copyright (C) 2006  Anthony Liguori <anthony@codemonkey.ws>
5  * Copyright (C) 2011  Kevin Wolf <kwolf@redhat.com>
6  *
7  * This library is free software; you can redistribute it and/or
8  * modify it under the terms of the GNU Lesser General Public
9  * License as published by the Free Software Foundation; either
10  * version 2.0 of the License, or (at your option) any later version.
11  *
12  * This library is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
15  * Lesser General Public License for more details.
16  *
17  * You should have received a copy of the GNU Lesser General Public
18  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
19  */
20 
21 /* XXX Is there a nicer way to disable glibc's stack check for longjmp? */
22 #undef _FORTIFY_SOURCE
23 #define _FORTIFY_SOURCE 0
24 
25 #include "qemu/osdep.h"
26 #include <ucontext.h>
27 #include "qemu/coroutine_int.h"
28 #include "qemu/coroutine-tls.h"
29 
30 #ifdef CONFIG_VALGRIND_H
31 #include <valgrind/valgrind.h>
32 #endif
33 
34 #ifdef QEMU_SANITIZE_ADDRESS
35 #ifdef CONFIG_ASAN_IFACE_FIBER
36 #define CONFIG_ASAN 1
37 #include <sanitizer/asan_interface.h>
38 #endif
39 #endif
40 
41 #ifdef CONFIG_TSAN
42 #include <sanitizer/tsan_interface.h>
43 #endif
44 
45 typedef struct {
46     Coroutine base;
47     void *stack;
48     size_t stack_size;
49 #ifdef CONFIG_SAFESTACK
50     /* Need an unsafe stack for each coroutine */
51     void *unsafe_stack;
52     size_t unsafe_stack_size;
53 #endif
54     sigjmp_buf env;
55 
56 #ifdef CONFIG_TSAN
57     void *tsan_co_fiber;
58     void *tsan_caller_fiber;
59 #endif
60 
61 #ifdef CONFIG_VALGRIND_H
62     unsigned int valgrind_stack_id;
63 #endif
64 
65 } CoroutineUContext;
66 
67 /**
68  * Per-thread coroutine bookkeeping
69  */
70 QEMU_DEFINE_STATIC_CO_TLS(Coroutine *, current);
71 QEMU_DEFINE_STATIC_CO_TLS(CoroutineUContext, leader);
72 
73 /*
74  * va_args to makecontext() must be type 'int', so passing
75  * the pointer we need may require several int args. This
76  * union is a quick hack to let us do that
77  */
78 union cc_arg {
79     void *p;
80     int i[2];
81 };
82 
83 /*
84  * QEMU_ALWAYS_INLINE only does so if __OPTIMIZE__, so we cannot use it.
85  * always_inline is required to avoid TSan runtime fatal errors.
86  */
87 static inline __attribute__((always_inline))
88 void on_new_fiber(CoroutineUContext *co)
89 {
90 #ifdef CONFIG_TSAN
91     co->tsan_co_fiber = __tsan_create_fiber(0); /* flags: sync on switch */
92     co->tsan_caller_fiber = __tsan_get_current_fiber();
93 #endif
94 }
95 
96 /* always_inline is required to avoid TSan runtime fatal errors. */
97 static inline __attribute__((always_inline))
98 void finish_switch_fiber(void *fake_stack_save)
99 {
100 #ifdef CONFIG_ASAN
101     CoroutineUContext *leaderp = get_ptr_leader();
102     const void *bottom_old;
103     size_t size_old;
104 
105     __sanitizer_finish_switch_fiber(fake_stack_save, &bottom_old, &size_old);
106 
107     if (!leaderp->stack) {
108         leaderp->stack = (void *)bottom_old;
109         leaderp->stack_size = size_old;
110     }
111 #endif
112 #ifdef CONFIG_TSAN
113     if (fake_stack_save) {
114         __tsan_release(fake_stack_save);
115         __tsan_switch_to_fiber(fake_stack_save, 0);  /* 0=synchronize */
116     }
117 #endif
118 }
119 
120 /* always_inline is required to avoid TSan runtime fatal errors. */
121 static inline __attribute__((always_inline))
122 void start_switch_fiber_asan(void **fake_stack_save,
123                              const void *bottom, size_t size)
124 {
125 #ifdef CONFIG_ASAN
126     __sanitizer_start_switch_fiber(fake_stack_save, bottom, size);
127 #endif
128 }
129 
130 /* always_inline is required to avoid TSan runtime fatal errors. */
131 static inline __attribute__((always_inline))
132 void start_switch_fiber_tsan(void **fake_stack_save,
133                              CoroutineUContext *co,
134                              bool caller)
135 {
136 #ifdef CONFIG_TSAN
137     void *new_fiber = caller ?
138                       co->tsan_caller_fiber :
139                       co->tsan_co_fiber;
140     void *curr_fiber = __tsan_get_current_fiber();
141     __tsan_acquire(curr_fiber);
142 
143     *fake_stack_save = curr_fiber;
144     __tsan_switch_to_fiber(new_fiber, 0);  /* 0=synchronize */
145 #endif
146 }
147 
148 static void coroutine_trampoline(int i0, int i1)
149 {
150     union cc_arg arg;
151     CoroutineUContext *self;
152     Coroutine *co;
153     void *fake_stack_save = NULL;
154 
155     finish_switch_fiber(NULL);
156 
157     arg.i[0] = i0;
158     arg.i[1] = i1;
159     self = arg.p;
160     co = &self->base;
161 
162     /* Initialize longjmp environment and switch back the caller */
163     if (!sigsetjmp(self->env, 0)) {
164         CoroutineUContext *leaderp = get_ptr_leader();
165 
166         start_switch_fiber_asan(&fake_stack_save,
167                                 leaderp->stack, leaderp->stack_size);
168         start_switch_fiber_tsan(&fake_stack_save, self, true); /* true=caller */
169         siglongjmp(*(sigjmp_buf *)co->entry_arg, 1);
170     }
171 
172     finish_switch_fiber(fake_stack_save);
173 
174     while (true) {
175         co->entry(co->entry_arg);
176         qemu_coroutine_switch(co, co->caller, COROUTINE_TERMINATE);
177     }
178 }
179 
180 Coroutine *qemu_coroutine_new(void)
181 {
182     CoroutineUContext *co;
183     ucontext_t old_uc, uc;
184     sigjmp_buf old_env;
185     union cc_arg arg = {0};
186     void *fake_stack_save = NULL;
187 
188     /* The ucontext functions preserve signal masks which incurs a
189      * system call overhead.  sigsetjmp(buf, 0)/siglongjmp() does not
190      * preserve signal masks but only works on the current stack.
191      * Since we need a way to create and switch to a new stack, use
192      * the ucontext functions for that but sigsetjmp()/siglongjmp() for
193      * everything else.
194      */
195 
196     if (getcontext(&uc) == -1) {
197         abort();
198     }
199 
200     co = g_malloc0(sizeof(*co));
201     co->stack_size = COROUTINE_STACK_SIZE;
202     co->stack = qemu_alloc_stack(&co->stack_size);
203 #ifdef CONFIG_SAFESTACK
204     co->unsafe_stack_size = COROUTINE_STACK_SIZE;
205     co->unsafe_stack = qemu_alloc_stack(&co->unsafe_stack_size);
206 #endif
207     co->base.entry_arg = &old_env; /* stash away our jmp_buf */
208 
209     uc.uc_link = &old_uc;
210     uc.uc_stack.ss_sp = co->stack;
211     uc.uc_stack.ss_size = co->stack_size;
212     uc.uc_stack.ss_flags = 0;
213 
214 #ifdef CONFIG_VALGRIND_H
215     co->valgrind_stack_id =
216         VALGRIND_STACK_REGISTER(co->stack, co->stack + co->stack_size);
217 #endif
218 
219     arg.p = co;
220 
221     on_new_fiber(co);
222     makecontext(&uc, (void (*)(void))coroutine_trampoline,
223                 2, arg.i[0], arg.i[1]);
224 
225     /* swapcontext() in, siglongjmp() back out */
226     if (!sigsetjmp(old_env, 0)) {
227         start_switch_fiber_asan(&fake_stack_save, co->stack, co->stack_size);
228         start_switch_fiber_tsan(&fake_stack_save,
229                                 co, false); /* false=not caller */
230 
231 #ifdef CONFIG_SAFESTACK
232         /*
233          * Before we swap the context, set the new unsafe stack
234          * The unsafe stack grows just like the normal stack, so start from
235          * the last usable location of the memory area.
236          * NOTE: we don't have to re-set the usp afterwards because we are
237          * coming back to this context through a siglongjmp.
238          * The compiler already wrapped the corresponding sigsetjmp call with
239          * code that saves the usp on the (safe) stack before the call, and
240          * restores it right after (which is where we return with siglongjmp).
241          */
242         void *usp = co->unsafe_stack + co->unsafe_stack_size;
243         __safestack_unsafe_stack_ptr = usp;
244 #endif
245 
246         swapcontext(&old_uc, &uc);
247     }
248 
249     finish_switch_fiber(fake_stack_save);
250 
251     return &co->base;
252 }
253 
254 #ifdef CONFIG_VALGRIND_H
255 /* Work around an unused variable in the valgrind.h macro... */
256 #if !defined(__clang__)
257 #pragma GCC diagnostic push
258 #pragma GCC diagnostic ignored "-Wunused-but-set-variable"
259 #endif
260 static inline void valgrind_stack_deregister(CoroutineUContext *co)
261 {
262     VALGRIND_STACK_DEREGISTER(co->valgrind_stack_id);
263 }
264 #if !defined(__clang__)
265 #pragma GCC diagnostic pop
266 #endif
267 #endif
268 
269 #if defined(CONFIG_ASAN) && defined(CONFIG_COROUTINE_POOL)
270 static void coroutine_fn terminate_asan(void *opaque)
271 {
272     CoroutineUContext *to = DO_UPCAST(CoroutineUContext, base, opaque);
273 
274     set_current(opaque);
275     start_switch_fiber_asan(NULL, to->stack, to->stack_size);
276     G_STATIC_ASSERT(!IS_ENABLED(CONFIG_TSAN));
277     siglongjmp(to->env, COROUTINE_ENTER);
278 }
279 #endif
280 
281 void qemu_coroutine_delete(Coroutine *co_)
282 {
283     CoroutineUContext *co = DO_UPCAST(CoroutineUContext, base, co_);
284 
285 #if defined(CONFIG_ASAN) && defined(CONFIG_COROUTINE_POOL)
286     co_->entry_arg = qemu_coroutine_self();
287     co_->entry = terminate_asan;
288     qemu_coroutine_switch(co_->entry_arg, co_, COROUTINE_ENTER);
289 #endif
290 
291 #ifdef CONFIG_VALGRIND_H
292     valgrind_stack_deregister(co);
293 #endif
294 
295     qemu_free_stack(co->stack, co->stack_size);
296 #ifdef CONFIG_SAFESTACK
297     qemu_free_stack(co->unsafe_stack, co->unsafe_stack_size);
298 #endif
299     g_free(co);
300 }
301 
302 /* This function is marked noinline to prevent GCC from inlining it
303  * into coroutine_trampoline(). If we allow it to do that then it
304  * hoists the code to get the address of the TLS variable "current"
305  * out of the while() loop. This is an invalid transformation because
306  * the sigsetjmp() call may be called when running thread A but
307  * return in thread B, and so we might be in a different thread
308  * context each time round the loop.
309  */
310 CoroutineAction __attribute__((noinline))
311 qemu_coroutine_switch(Coroutine *from_, Coroutine *to_,
312                       CoroutineAction action)
313 {
314     CoroutineUContext *from = DO_UPCAST(CoroutineUContext, base, from_);
315     CoroutineUContext *to = DO_UPCAST(CoroutineUContext, base, to_);
316     int ret;
317     void *fake_stack_save = NULL;
318 
319     set_current(to_);
320 
321     ret = sigsetjmp(from->env, 0);
322     if (ret == 0) {
323         start_switch_fiber_asan(IS_ENABLED(CONFIG_COROUTINE_POOL) ||
324                                 action != COROUTINE_TERMINATE ?
325                                     &fake_stack_save : NULL,
326                                 to->stack, to->stack_size);
327         start_switch_fiber_tsan(&fake_stack_save,
328                                 to, false); /* false=not caller */
329         siglongjmp(to->env, action);
330     }
331 
332     finish_switch_fiber(fake_stack_save);
333 
334     return ret;
335 }
336 
337 Coroutine *qemu_coroutine_self(void)
338 {
339     Coroutine *self = get_current();
340     CoroutineUContext *leaderp = get_ptr_leader();
341 
342     if (!self) {
343         self = &leaderp->base;
344         set_current(self);
345     }
346 #ifdef CONFIG_TSAN
347     if (!leaderp->tsan_co_fiber) {
348         leaderp->tsan_co_fiber = __tsan_get_current_fiber();
349     }
350 #endif
351     return self;
352 }
353 
354 bool qemu_in_coroutine(void)
355 {
356     Coroutine *self = get_current();
357 
358     return self && self->caller;
359 }
360