summaryrefslogtreecommitdiff
path: root/atomic.h
blob: 714bf23a6a57e8144515f8de64eaaaf0a1d4cb07 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
#ifdef __GNUC__

/*
 * These atomic operations copied from the linux kernel and altered
 * only slightly. I need to get official permission to distribute
 * under the Artistic License.
 */
/* We really need to integrate the atomic typedef with the typedef
 * used by sv_refcnt of an SV. It's possible that for CPUs like alpha
 * where we'd need to up sv_refcnt from 32 to 64 bits, we may be better
 * off sticking with EMULATE_ATOMIC_REFCOUNTS instead.
 */
typedef U32 atomic_t;	/* kludge */

#ifdef i386

#  ifdef NO_SMP
#    define LOCK ""
#  else
#    define LOCK "lock ; "
#  endif

#  define __atomic_fool_gcc(x) (*(struct { int a[100]; } *)x)
static __inline__ void atomic_inc(atomic_t *v)
{
    __asm__ __volatile__(
	    LOCK "incl %0"
	    :"=m" (__atomic_fool_gcc(v))
	    :"m" (__atomic_fool_gcc(v)));
}

static __inline__ int atomic_dec_and_test(atomic_t *v)
{
    unsigned char c;

    __asm__ __volatile__(
	    LOCK "decl %0; sete %1"
	    :"=m" (__atomic_fool_gcc(v)), "=qm" (c)
	    :"m" (__atomic_fool_gcc(v)));
    return c != 0;
}
#  else
/* XXX What symbol does gcc define for sparc64? */
#  ifdef sparc64
#    define __atomic_fool_gcc(x) ((struct { int a[100]; } *)x)
typedef U32 atomic_t;
extern __inline__ void atomic_add(int i, atomic_t *v)
{
        __asm__ __volatile__("
1:      lduw            [%1], %%g5
        add             %%g5, %0, %%g7
        cas             [%1], %%g5, %%g7
        sub             %%g5, %%g7, %%g5
        brnz,pn         %%g5, 1b
         nop"
        : /* No outputs */
        : "HIr" (i), "r" (__atomic_fool_gcc(v))
        : "g5", "g7", "memory");
}

extern __inline__ int atomic_sub_return(int i, atomic_t *v)
{
        unsigned long oldval;
        __asm__ __volatile__("
1:      lduw            [%2], %%g5
        sub             %%g5, %1, %%g7
        cas             [%2], %%g5, %%g7
        sub             %%g5, %%g7, %%g5
        brnz,pn         %%g5, 1b
         sub            %%g7, %1, %0"
        : "=&r" (oldval)
        : "HIr" (i), "r" (__atomic_fool_gcc(v))
        : "g5", "g7", "memory");
        return (int)oldval;
}

#define atomic_inc(v) atomic_add(1,(v))
#define atomic_dec_and_test(v) (atomic_sub_return(1, (v)) == 0)
/* Add further gcc architectures here */
#  endif /* sparc64 */
#endif /* i386 */
#else
/* Add non-gcc native atomic operations here */
#  define EMULATE_ATOMIC_REFCOUNTS
#endif