blob: 8ef603d081ea02b80ecfca27a22a6b19d8a1a49c [file] [log] [blame]
Kevin Wolf00dccaf2011-01-17 16:08:14 +00001/*
2 * ucontext coroutine initialization code
3 *
4 * Copyright (C) 2006 Anthony Liguori <anthony@codemonkey.ws>
5 * Copyright (C) 2011 Kevin Wolf <kwolf@redhat.com>
6 *
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2.0 of the License, or (at your option) any later version.
11 *
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
19 */
20
21/* XXX Is there a nicer way to disable glibc's stack check for longjmp? */
Kevin Wolf00dccaf2011-01-17 16:08:14 +000022#undef _FORTIFY_SOURCE
Daniel P. Berrangé9afa8882023-10-03 10:15:49 +010023#define _FORTIFY_SOURCE 0
24
Peter Maydellaafd7582016-01-29 17:49:55 +000025#include "qemu/osdep.h"
Kevin Wolf00dccaf2011-01-17 16:08:14 +000026#include <ucontext.h>
Daniel P. Berrange10817bf2015-09-01 14:48:02 +010027#include "qemu/coroutine_int.h"
Stefan Hajnoczi34145a32022-03-07 15:38:51 +000028#include "qemu/coroutine-tls.h"
Kevin Wolf00dccaf2011-01-17 16:08:14 +000029
Kevin Wolf3f4349d2012-06-29 13:40:27 +020030#ifdef CONFIG_VALGRIND_H
31#include <valgrind/valgrind.h>
32#endif
33
Marc-André Lureau638466f2022-04-20 17:21:45 +040034#ifdef QEMU_SANITIZE_ADDRESS
Marc-André Lureaud83414e2018-01-16 16:11:52 +010035#ifdef CONFIG_ASAN_IFACE_FIBER
36#define CONFIG_ASAN 1
37#include <sanitizer/asan_interface.h>
38#endif
39#endif
40
Lingfeng Yang0aebab02020-06-12 20:02:23 +010041#ifdef CONFIG_TSAN
42#include <sanitizer/tsan_interface.h>
43#endif
44
Kevin Wolf00dccaf2011-01-17 16:08:14 +000045typedef struct {
46 Coroutine base;
47 void *stack;
Peter Lievenddba1592016-09-27 11:58:43 +020048 size_t stack_size;
Daniele Buono58ebc2c2020-05-29 16:51:19 -040049#ifdef CONFIG_SAFESTACK
50 /* Need an unsafe stack for each coroutine */
51 void *unsafe_stack;
52 size_t unsafe_stack_size;
53#endif
Peter Maydell6ab7e542013-02-20 15:21:09 +000054 sigjmp_buf env;
Kevin Wolf3f4349d2012-06-29 13:40:27 +020055
Robert Foley995f5c32020-07-01 14:56:16 +010056#ifdef CONFIG_TSAN
Lingfeng Yang0aebab02020-06-12 20:02:23 +010057 void *tsan_co_fiber;
58 void *tsan_caller_fiber;
Robert Foley995f5c32020-07-01 14:56:16 +010059#endif
Lingfeng Yang0aebab02020-06-12 20:02:23 +010060
Kevin Wolf3f4349d2012-06-29 13:40:27 +020061#ifdef CONFIG_VALGRIND_H
62 unsigned int valgrind_stack_id;
63#endif
64
Kevin Wolf00dccaf2011-01-17 16:08:14 +000065} CoroutineUContext;
66
67/**
68 * Per-thread coroutine bookkeeping
69 */
Stefan Hajnoczi34145a32022-03-07 15:38:51 +000070QEMU_DEFINE_STATIC_CO_TLS(Coroutine *, current);
71QEMU_DEFINE_STATIC_CO_TLS(CoroutineUContext, leader);
Kevin Wolf00dccaf2011-01-17 16:08:14 +000072
73/*
74 * va_args to makecontext() must be type 'int', so passing
75 * the pointer we need may require several int args. This
76 * union is a quick hack to let us do that
77 */
78union cc_arg {
79 void *p;
80 int i[2];
81};
82
Robert Foley995f5c32020-07-01 14:56:16 +010083/*
84 * QEMU_ALWAYS_INLINE only does so if __OPTIMIZE__, so we cannot use it.
85 * always_inline is required to avoid TSan runtime fatal errors.
86 */
Lingfeng Yang0aebab02020-06-12 20:02:23 +010087static inline __attribute__((always_inline))
88void on_new_fiber(CoroutineUContext *co)
89{
90#ifdef CONFIG_TSAN
91 co->tsan_co_fiber = __tsan_create_fiber(0); /* flags: sync on switch */
92 co->tsan_caller_fiber = __tsan_get_current_fiber();
93#endif
94}
95
Robert Foley995f5c32020-07-01 14:56:16 +010096/* always_inline is required to avoid TSan runtime fatal errors. */
Lingfeng Yang0aebab02020-06-12 20:02:23 +010097static inline __attribute__((always_inline))
98void finish_switch_fiber(void *fake_stack_save)
Marc-André Lureaud83414e2018-01-16 16:11:52 +010099{
100#ifdef CONFIG_ASAN
Stefan Hajnoczi34145a32022-03-07 15:38:51 +0000101 CoroutineUContext *leaderp = get_ptr_leader();
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100102 const void *bottom_old;
103 size_t size_old;
104
105 __sanitizer_finish_switch_fiber(fake_stack_save, &bottom_old, &size_old);
106
Stefan Hajnoczi34145a32022-03-07 15:38:51 +0000107 if (!leaderp->stack) {
108 leaderp->stack = (void *)bottom_old;
109 leaderp->stack_size = size_old;
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100110 }
111#endif
Lingfeng Yang0aebab02020-06-12 20:02:23 +0100112#ifdef CONFIG_TSAN
113 if (fake_stack_save) {
114 __tsan_release(fake_stack_save);
115 __tsan_switch_to_fiber(fake_stack_save, 0); /* 0=synchronize */
116 }
117#endif
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100118}
119
Robert Foley995f5c32020-07-01 14:56:16 +0100120/* always_inline is required to avoid TSan runtime fatal errors. */
121static inline __attribute__((always_inline))
Akihiko Odakid9945cc2024-01-17 16:06:02 +0900122void start_switch_fiber_asan(void **fake_stack_save,
Robert Foley995f5c32020-07-01 14:56:16 +0100123 const void *bottom, size_t size)
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100124{
125#ifdef CONFIG_ASAN
Akihiko Odakid9945cc2024-01-17 16:06:02 +0900126 __sanitizer_start_switch_fiber(fake_stack_save, bottom, size);
Lingfeng Yang0aebab02020-06-12 20:02:23 +0100127#endif
Robert Foley995f5c32020-07-01 14:56:16 +0100128}
129
130/* always_inline is required to avoid TSan runtime fatal errors. */
131static inline __attribute__((always_inline))
132void start_switch_fiber_tsan(void **fake_stack_save,
133 CoroutineUContext *co,
134 bool caller)
135{
Lingfeng Yang0aebab02020-06-12 20:02:23 +0100136#ifdef CONFIG_TSAN
Robert Foley995f5c32020-07-01 14:56:16 +0100137 void *new_fiber = caller ?
138 co->tsan_caller_fiber :
139 co->tsan_co_fiber;
140 void *curr_fiber = __tsan_get_current_fiber();
Lingfeng Yang0aebab02020-06-12 20:02:23 +0100141 __tsan_acquire(curr_fiber);
142
143 *fake_stack_save = curr_fiber;
144 __tsan_switch_to_fiber(new_fiber, 0); /* 0=synchronize */
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100145#endif
146}
147
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000148static void coroutine_trampoline(int i0, int i1)
149{
150 union cc_arg arg;
151 CoroutineUContext *self;
152 Coroutine *co;
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100153 void *fake_stack_save = NULL;
154
155 finish_switch_fiber(NULL);
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000156
157 arg.i[0] = i0;
158 arg.i[1] = i1;
159 self = arg.p;
160 co = &self->base;
161
162 /* Initialize longjmp environment and switch back the caller */
Peter Maydell6ab7e542013-02-20 15:21:09 +0000163 if (!sigsetjmp(self->env, 0)) {
Stefan Hajnoczi34145a32022-03-07 15:38:51 +0000164 CoroutineUContext *leaderp = get_ptr_leader();
165
Akihiko Odakid9945cc2024-01-17 16:06:02 +0900166 start_switch_fiber_asan(&fake_stack_save,
Stefan Hajnoczi34145a32022-03-07 15:38:51 +0000167 leaderp->stack, leaderp->stack_size);
Robert Foley995f5c32020-07-01 14:56:16 +0100168 start_switch_fiber_tsan(&fake_stack_save, self, true); /* true=caller */
Peter Maydell6ab7e542013-02-20 15:21:09 +0000169 siglongjmp(*(sigjmp_buf *)co->entry_arg, 1);
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000170 }
171
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100172 finish_switch_fiber(fake_stack_save);
173
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000174 while (true) {
175 co->entry(co->entry_arg);
176 qemu_coroutine_switch(co, co->caller, COROUTINE_TERMINATE);
177 }
178}
179
Paolo Bonzini40239782013-02-19 11:59:09 +0100180Coroutine *qemu_coroutine_new(void)
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000181{
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000182 CoroutineUContext *co;
183 ucontext_t old_uc, uc;
Peter Maydell6ab7e542013-02-20 15:21:09 +0000184 sigjmp_buf old_env;
malc32b74672011-08-08 13:46:51 +0400185 union cc_arg arg = {0};
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100186 void *fake_stack_save = NULL;
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000187
Peter Maydell6ab7e542013-02-20 15:21:09 +0000188 /* The ucontext functions preserve signal masks which incurs a
189 * system call overhead. sigsetjmp(buf, 0)/siglongjmp() does not
190 * preserve signal masks but only works on the current stack.
191 * Since we need a way to create and switch to a new stack, use
192 * the ucontext functions for that but sigsetjmp()/siglongjmp() for
193 * everything else.
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000194 */
195
196 if (getcontext(&uc) == -1) {
197 abort();
198 }
199
Anthony Liguori7267c092011-08-20 22:09:37 -0500200 co = g_malloc0(sizeof(*co));
Peter Lievenddba1592016-09-27 11:58:43 +0200201 co->stack_size = COROUTINE_STACK_SIZE;
202 co->stack = qemu_alloc_stack(&co->stack_size);
Daniele Buono58ebc2c2020-05-29 16:51:19 -0400203#ifdef CONFIG_SAFESTACK
204 co->unsafe_stack_size = COROUTINE_STACK_SIZE;
205 co->unsafe_stack = qemu_alloc_stack(&co->unsafe_stack_size);
206#endif
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000207 co->base.entry_arg = &old_env; /* stash away our jmp_buf */
208
209 uc.uc_link = &old_uc;
210 uc.uc_stack.ss_sp = co->stack;
Peter Lievenddba1592016-09-27 11:58:43 +0200211 uc.uc_stack.ss_size = co->stack_size;
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000212 uc.uc_stack.ss_flags = 0;
213
Kevin Wolf3f4349d2012-06-29 13:40:27 +0200214#ifdef CONFIG_VALGRIND_H
215 co->valgrind_stack_id =
Peter Lievenddba1592016-09-27 11:58:43 +0200216 VALGRIND_STACK_REGISTER(co->stack, co->stack + co->stack_size);
Kevin Wolf3f4349d2012-06-29 13:40:27 +0200217#endif
218
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000219 arg.p = co;
220
Lingfeng Yang0aebab02020-06-12 20:02:23 +0100221 on_new_fiber(co);
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000222 makecontext(&uc, (void (*)(void))coroutine_trampoline,
223 2, arg.i[0], arg.i[1]);
224
Peter Maydell6ab7e542013-02-20 15:21:09 +0000225 /* swapcontext() in, siglongjmp() back out */
226 if (!sigsetjmp(old_env, 0)) {
Akihiko Odakid9945cc2024-01-17 16:06:02 +0900227 start_switch_fiber_asan(&fake_stack_save, co->stack, co->stack_size);
Robert Foley995f5c32020-07-01 14:56:16 +0100228 start_switch_fiber_tsan(&fake_stack_save,
229 co, false); /* false=not caller */
Daniele Buono58ebc2c2020-05-29 16:51:19 -0400230
231#ifdef CONFIG_SAFESTACK
232 /*
233 * Before we swap the context, set the new unsafe stack
234 * The unsafe stack grows just like the normal stack, so start from
235 * the last usable location of the memory area.
236 * NOTE: we don't have to re-set the usp afterwards because we are
237 * coming back to this context through a siglongjmp.
238 * The compiler already wrapped the corresponding sigsetjmp call with
239 * code that saves the usp on the (safe) stack before the call, and
240 * restores it right after (which is where we return with siglongjmp).
241 */
242 void *usp = co->unsafe_stack + co->unsafe_stack_size;
243 __safestack_unsafe_stack_ptr = usp;
244#endif
245
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000246 swapcontext(&old_uc, &uc);
247 }
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100248
249 finish_switch_fiber(fake_stack_save);
250
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000251 return &co->base;
252}
253
Kevin Wolf3f4349d2012-06-29 13:40:27 +0200254#ifdef CONFIG_VALGRIND_H
255/* Work around an unused variable in the valgrind.h macro... */
Thomas Huth7aa12aa2020-07-08 20:19:44 +0200256#if !defined(__clang__)
Markus Armbrustere6f53fd2013-04-16 13:51:06 +0200257#pragma GCC diagnostic push
Kevin Wolf3f4349d2012-06-29 13:40:27 +0200258#pragma GCC diagnostic ignored "-Wunused-but-set-variable"
Peter Maydell06d71fa2012-07-30 16:13:07 +0100259#endif
Kevin Wolf3f4349d2012-06-29 13:40:27 +0200260static inline void valgrind_stack_deregister(CoroutineUContext *co)
261{
262 VALGRIND_STACK_DEREGISTER(co->valgrind_stack_id);
263}
Thomas Huth7aa12aa2020-07-08 20:19:44 +0200264#if !defined(__clang__)
Markus Armbrustere6f53fd2013-04-16 13:51:06 +0200265#pragma GCC diagnostic pop
Kevin Wolf3f4349d2012-06-29 13:40:27 +0200266#endif
Peter Maydell06d71fa2012-07-30 16:13:07 +0100267#endif
Kevin Wolf3f4349d2012-06-29 13:40:27 +0200268
Akihiko Odakid9945cc2024-01-17 16:06:02 +0900269#if defined(CONFIG_ASAN) && defined(CONFIG_COROUTINE_POOL)
270static void coroutine_fn terminate_asan(void *opaque)
271{
272 CoroutineUContext *to = DO_UPCAST(CoroutineUContext, base, opaque);
273
274 set_current(opaque);
275 start_switch_fiber_asan(NULL, to->stack, to->stack_size);
276 G_STATIC_ASSERT(!IS_ENABLED(CONFIG_TSAN));
277 siglongjmp(to->env, COROUTINE_ENTER);
278}
279#endif
280
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000281void qemu_coroutine_delete(Coroutine *co_)
282{
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000283 CoroutineUContext *co = DO_UPCAST(CoroutineUContext, base, co_);
284
Akihiko Odakid9945cc2024-01-17 16:06:02 +0900285#if defined(CONFIG_ASAN) && defined(CONFIG_COROUTINE_POOL)
286 co_->entry_arg = qemu_coroutine_self();
287 co_->entry = terminate_asan;
288 qemu_coroutine_switch(co_->entry_arg, co_, COROUTINE_ENTER);
289#endif
290
Kevin Wolf3f4349d2012-06-29 13:40:27 +0200291#ifdef CONFIG_VALGRIND_H
292 valgrind_stack_deregister(co);
293#endif
294
Peter Lievenddba1592016-09-27 11:58:43 +0200295 qemu_free_stack(co->stack, co->stack_size);
Daniele Buono58ebc2c2020-05-29 16:51:19 -0400296#ifdef CONFIG_SAFESTACK
297 qemu_free_stack(co->unsafe_stack, co->unsafe_stack_size);
298#endif
Anthony Liguori7267c092011-08-20 22:09:37 -0500299 g_free(co);
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000300}
301
Paolo Bonzinid1d1b202014-12-02 12:05:44 +0100302/* This function is marked noinline to prevent GCC from inlining it
303 * into coroutine_trampoline(). If we allow it to do that then it
304 * hoists the code to get the address of the TLS variable "current"
305 * out of the while() loop. This is an invalid transformation because
306 * the sigsetjmp() call may be called when running thread A but
307 * return in thread B, and so we might be in a different thread
308 * context each time round the loop.
309 */
310CoroutineAction __attribute__((noinline))
311qemu_coroutine_switch(Coroutine *from_, Coroutine *to_,
312 CoroutineAction action)
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000313{
314 CoroutineUContext *from = DO_UPCAST(CoroutineUContext, base, from_);
315 CoroutineUContext *to = DO_UPCAST(CoroutineUContext, base, to_);
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000316 int ret;
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100317 void *fake_stack_save = NULL;
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000318
Stefan Hajnoczi34145a32022-03-07 15:38:51 +0000319 set_current(to_);
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000320
Peter Maydell6ab7e542013-02-20 15:21:09 +0000321 ret = sigsetjmp(from->env, 0);
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000322 if (ret == 0) {
Akihiko Odakid9945cc2024-01-17 16:06:02 +0900323 start_switch_fiber_asan(IS_ENABLED(CONFIG_COROUTINE_POOL) ||
324 action != COROUTINE_TERMINATE ?
325 &fake_stack_save : NULL,
326 to->stack, to->stack_size);
Robert Foley995f5c32020-07-01 14:56:16 +0100327 start_switch_fiber_tsan(&fake_stack_save,
328 to, false); /* false=not caller */
Peter Maydell6ab7e542013-02-20 15:21:09 +0000329 siglongjmp(to->env, action);
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000330 }
Marc-André Lureaud83414e2018-01-16 16:11:52 +0100331
332 finish_switch_fiber(fake_stack_save);
333
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000334 return ret;
335}
336
337Coroutine *qemu_coroutine_self(void)
338{
Stefan Hajnoczi34145a32022-03-07 15:38:51 +0000339 Coroutine *self = get_current();
340 CoroutineUContext *leaderp = get_ptr_leader();
341
342 if (!self) {
343 self = &leaderp->base;
344 set_current(self);
Paolo Bonzinid1d1b202014-12-02 12:05:44 +0100345 }
Lingfeng Yang0aebab02020-06-12 20:02:23 +0100346#ifdef CONFIG_TSAN
Stefan Hajnoczi34145a32022-03-07 15:38:51 +0000347 if (!leaderp->tsan_co_fiber) {
348 leaderp->tsan_co_fiber = __tsan_get_current_fiber();
Lingfeng Yang0aebab02020-06-12 20:02:23 +0100349 }
350#endif
Stefan Hajnoczi34145a32022-03-07 15:38:51 +0000351 return self;
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000352}
353
354bool qemu_in_coroutine(void)
355{
Stefan Hajnoczi34145a32022-03-07 15:38:51 +0000356 Coroutine *self = get_current();
357
358 return self && self->caller;
Kevin Wolf00dccaf2011-01-17 16:08:14 +0000359}