/* Spin locks for communication between threads and signal handlers. Copyright (C) 2020-2023 Free Software Foundation, Inc. This file is free software: you can redistribute it and/or modify it under the terms of the GNU Lesser General Public License as published by the Free Software Foundation; either version 2.1 of the License, or (at your option) any later version. This file is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for more details. You should have received a copy of the GNU Lesser General Public License along with this program. If not, see . */ /* Written by Bruno Haible , 2020. */ #include /* Specification. */ #include "asyncsafe-spin.h" #include #if defined _AIX # include #endif #if 0x590 <= __SUNPRO_C && __STDC__ # define asm __asm #endif #if defined _WIN32 && ! defined __CYGWIN__ /* Use Windows threads. */ void asyncsafe_spin_init (asyncsafe_spinlock_t *lock) { glwthread_spin_init (lock); } static inline void do_lock (asyncsafe_spinlock_t *lock) { glwthread_spin_lock (lock); } static inline void do_unlock (asyncsafe_spinlock_t *lock) { if (glwthread_spin_unlock (lock)) abort (); } void asyncsafe_spin_destroy (asyncsafe_spinlock_t *lock) { glwthread_spin_destroy (lock); } #else # if HAVE_PTHREAD_H /* Use POSIX threads. */ /* We don't use semaphores (although sem_post() is allowed in signal handlers), because it would require to link with -lrt on HP-UX 11, OSF/1, Solaris 10, and also because on macOS only named semaphores work. We don't use the C11 (available in GCC >= 4.9) because it would require to link with -latomic. */ # if (__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 7) \ || __clang_major__ > 3 || (__clang_major__ == 3 && __clang_minor__ >= 1)) \ && !defined __ibmxl__ /* Use GCC built-ins (available in GCC >= 4.7 and clang >= 3.1) that operate on the first byte of the lock. Documentation: */ # if 1 /* An implementation that verifies the unlocks. */ void asyncsafe_spin_init (asyncsafe_spinlock_t *lock) { __atomic_store_n (lock, 0, __ATOMIC_SEQ_CST); } static inline void do_lock (asyncsafe_spinlock_t *lock) { /* Wait until *lock becomes 0, then replace it with 1. */ asyncsafe_spinlock_t zero; while (!(zero = 0, __atomic_compare_exchange_n (lock, &zero, 1, false, __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST))) ; } static inline void do_unlock (asyncsafe_spinlock_t *lock) { /* If *lock is 1, then replace it with 0. */ asyncsafe_spinlock_t one = 1; if (!__atomic_compare_exchange_n (lock, &one, 0, false, __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST)) abort (); } # else /* An implementation that is a little bit more optimized, but does not verify the unlocks. */ void asyncsafe_spin_init (asyncsafe_spinlock_t *lock) { __atomic_clear (lock, __ATOMIC_SEQ_CST); } static inline void do_lock (asyncsafe_spinlock_t *lock) { while (__atomic_test_and_set (lock, __ATOMIC_SEQ_CST)) ; } static inline void do_unlock (asyncsafe_spinlock_t *lock) { __atomic_clear (lock, __ATOMIC_SEQ_CST); } # endif # elif (((__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 1)) \ && !defined __sparc__ && !defined __ANDROID__) \ || __clang_major__ >= 3) \ && !defined __ibmxl__ /* Use GCC built-ins (available in GCC >= 4.1, except on SPARC, and clang >= 3.0). Documentation: */ void asyncsafe_spin_init (asyncsafe_spinlock_t *lock) { volatile unsigned int *vp = lock; *vp = 0; __sync_synchronize (); } static inline void do_lock (asyncsafe_spinlock_t *lock) { /* Wait until *lock becomes 0, then replace it with 1. */ while (__sync_val_compare_and_swap (lock, 0, 1) != 0) ; } static inline void do_unlock (asyncsafe_spinlock_t *lock) { /* If *lock is 1, then replace it with 0. */ if (__sync_val_compare_and_swap (lock, 1, 0) != 1) abort (); } # elif defined _AIX /* AIX */ void asyncsafe_spin_init (asyncsafe_spinlock_t *lock) { atomic_p vp = (int *) lock; _clear_lock (vp, 0); } static inline void do_lock (asyncsafe_spinlock_t *lock) { atomic_p vp = (int *) lock; while (_check_lock (vp, 0, 1)) ; } static inline void do_unlock (asyncsafe_spinlock_t *lock) { atomic_p vp = (int *) lock; if (_check_lock (vp, 1, 0)) abort (); } # elif ((defined __GNUC__ || defined __clang__ || defined __SUNPRO_C) && (defined __sparc || defined __i386 || defined __x86_64__)) || (defined __TINYC__ && (defined __i386 || defined __x86_64__)) /* For older versions of GCC or clang, use inline assembly. GCC, clang, and the Oracle Studio C 12 compiler understand GCC's extended asm syntax, but the plain Oracle Studio C 11 compiler understands only simple asm. */ /* An implementation that verifies the unlocks. */ static void memory_barrier (void) { # if defined __GNUC__ || defined __clang__ || __SUNPRO_C >= 0x590 || defined __TINYC__ # if defined __i386 || defined __x86_64__ # if defined __TINYC__ && defined __i386 /* Cannot use the SSE instruction "mfence" with this compiler. */ asm volatile ("lock orl $0,(%esp)"); # else asm volatile ("mfence"); # endif # endif # if defined __sparc asm volatile ("membar 2"); # endif # else # if defined __i386 || defined __x86_64__ asm ("mfence"); # endif # if defined __sparc asm ("membar 2"); # endif # endif } /* Store NEWVAL in *VP if the old value *VP is == CMP. Return the old value. */ static unsigned int atomic_compare_and_swap (volatile unsigned int *vp, unsigned int cmp, unsigned int newval) { # if defined __GNUC__ || defined __clang__ || __SUNPRO_C >= 0x590 || defined __TINYC__ unsigned int oldval; # if defined __i386 || defined __x86_64__ asm volatile (" lock\n cmpxchgl %3,(%1)" : "=a" (oldval) : "r" (vp), "a" (cmp), "r" (newval) : "memory"); # endif # if defined __sparc asm volatile (" cas [%1],%2,%3\n" " mov %3,%0" : "=r" (oldval) : "r" (vp), "r" (cmp), "r" (newval) : "memory"); # endif return oldval; # else /* __SUNPRO_C */ # if defined __x86_64__ asm (" movl %esi,%eax\n" " lock\n cmpxchgl %edx,(%rdi)"); # elif defined __i386 asm (" movl 16(%ebp),%ecx\n" " movl 12(%ebp),%eax\n" " movl 8(%ebp),%edx\n" " lock\n cmpxchgl %ecx,(%edx)"); # endif # if defined __sparc asm (" cas [%i0],%i1,%i2\n" " mov %i2,%i0"); # endif # endif } void asyncsafe_spin_init (asyncsafe_spinlock_t *lock) { volatile unsigned int *vp = lock; *vp = 0; memory_barrier (); } static inline void do_lock (asyncsafe_spinlock_t *lock) { volatile unsigned int *vp = lock; while (atomic_compare_and_swap (vp, 0, 1) != 0) ; } static inline void do_unlock (asyncsafe_spinlock_t *lock) { volatile unsigned int *vp = lock; if (atomic_compare_and_swap (vp, 1, 0) != 1) abort (); } # else /* Fallback code. It has some race conditions. */ void asyncsafe_spin_init (asyncsafe_spinlock_t *lock) { volatile unsigned int *vp = lock; *vp = 0; } static inline void do_lock (asyncsafe_spinlock_t *lock) { volatile unsigned int *vp = lock; while (*vp) ; *vp = 1; } static inline void do_unlock (asyncsafe_spinlock_t *lock) { volatile unsigned int *vp = lock; *vp = 0; } # endif # else /* Provide a dummy implementation for single-threaded applications. */ void asyncsafe_spin_init (asyncsafe_spinlock_t *lock) { } static inline void do_lock (asyncsafe_spinlock_t *lock) { } static inline void do_unlock (asyncsafe_spinlock_t *lock) { } # endif void asyncsafe_spin_destroy (asyncsafe_spinlock_t *lock) { } #endif void asyncsafe_spin_lock (asyncsafe_spinlock_t *lock, const sigset_t *mask, sigset_t *saved_mask) { sigprocmask (SIG_BLOCK, mask, saved_mask); /* equivalent to pthread_sigmask */ do_lock (lock); } void asyncsafe_spin_unlock (asyncsafe_spinlock_t *lock, const sigset_t *saved_mask) { do_unlock (lock); sigprocmask (SIG_SETMASK, saved_mask, NULL); /* equivalent to pthread_sigmask */ }