| /* |
| * SPDX-License-Identifier: GPL-2.0-or-later |
| * Compare-and-swap for 128-bit atomic operations, generic version. |
| * |
| * Copyright (C) 2018, 2023 Linaro, Ltd. |
| * |
| * See docs/devel/atomics.rst for discussion about the guarantees each |
| * atomic primitive is meant to provide. |
| */ |
| |
| #ifndef HOST_ATOMIC128_CAS_H |
| #define HOST_ATOMIC128_CAS_H |
| |
| #if defined(CONFIG_ATOMIC128) |
| static inline Int128 ATTRIBUTE_ATOMIC128_OPT |
| atomic16_cmpxchg(Int128 *ptr, Int128 cmp, Int128 new) |
| { |
| __int128_t *ptr_align = __builtin_assume_aligned(ptr, 16); |
| Int128Alias r, c, n; |
| |
| c.s = cmp; |
| n.s = new; |
| r.i = qatomic_cmpxchg__nocheck(ptr_align, c.i, n.i); |
| return r.s; |
| } |
| # define HAVE_CMPXCHG128 1 |
| #elif defined(CONFIG_CMPXCHG128) |
| static inline Int128 ATTRIBUTE_ATOMIC128_OPT |
| atomic16_cmpxchg(Int128 *ptr, Int128 cmp, Int128 new) |
| { |
| Int128Aligned *ptr_align = __builtin_assume_aligned(ptr, 16); |
| Int128Alias r, c, n; |
| |
| c.s = cmp; |
| n.s = new; |
| r.i = __sync_val_compare_and_swap_16(ptr_align, c.i, n.i); |
| return r.s; |
| } |
| # define HAVE_CMPXCHG128 1 |
| #else |
| /* Fallback definition that must be optimized away, or error. */ |
| Int128 QEMU_ERROR("unsupported atomic") |
| atomic16_cmpxchg(Int128 *ptr, Int128 cmp, Int128 new); |
| # define HAVE_CMPXCHG128 0 |
| #endif |
| |
| #endif /* HOST_ATOMIC128_CAS_H */ |