1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */ 2a439fe51SSam Ravnborg /* atomic.h: Thankfully the V9 is at least reasonable for this 3a439fe51SSam Ravnborg * stuff. 4a439fe51SSam Ravnborg * 5193d2aadSDavid S. Miller * Copyright (C) 1996, 1997, 2000, 2012 David S. Miller (davem@redhat.com) 6a439fe51SSam Ravnborg */ 7a439fe51SSam Ravnborg 8a439fe51SSam Ravnborg #ifndef __ARCH_SPARC64_ATOMIC__ 9a439fe51SSam Ravnborg #define __ARCH_SPARC64_ATOMIC__ 10a439fe51SSam Ravnborg 11a439fe51SSam Ravnborg #include <linux/types.h> 12d550bbd4SDavid Howells #include <asm/cmpxchg.h> 1356d36489SPeter Zijlstra #include <asm/barrier.h> 14a439fe51SSam Ravnborg 15a439fe51SSam Ravnborg #define ATOMIC64_INIT(i) { (i) } 16a439fe51SSam Ravnborg 17ff5b4f1eSMark Rutland #define arch_atomic_read(v) READ_ONCE((v)->counter) 18ff5b4f1eSMark Rutland #define arch_atomic64_read(v) READ_ONCE((v)->counter) 19a439fe51SSam Ravnborg 20ff5b4f1eSMark Rutland #define arch_atomic_set(v, i) WRITE_ONCE(((v)->counter), (i)) 21ff5b4f1eSMark Rutland #define arch_atomic64_set(v, i) WRITE_ONCE(((v)->counter), (i)) 22a439fe51SSam Ravnborg 234f3316c2SPeter Zijlstra #define ATOMIC_OP(op) \ 24ff5b4f1eSMark Rutland void arch_atomic_##op(int, atomic_t *); \ 25ff5b4f1eSMark Rutland void arch_atomic64_##op(s64, atomic64_t *); 26a439fe51SSam Ravnborg 274f3316c2SPeter Zijlstra #define ATOMIC_OP_RETURN(op) \ 28ff5b4f1eSMark Rutland int arch_atomic_##op##_return(int, atomic_t *); \ 29ff5b4f1eSMark Rutland s64 arch_atomic64_##op##_return(s64, atomic64_t *); 30a439fe51SSam Ravnborg 313a1adb23SPeter Zijlstra #define ATOMIC_FETCH_OP(op) \ 32ff5b4f1eSMark Rutland int arch_atomic_fetch_##op(int, atomic_t *); \ 33ff5b4f1eSMark Rutland s64 arch_atomic64_fetch_##op(s64, atomic64_t *); 343a1adb23SPeter Zijlstra 353a1adb23SPeter Zijlstra #define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETURN(op) ATOMIC_FETCH_OP(op) 36a439fe51SSam Ravnborg 374f3316c2SPeter Zijlstra ATOMIC_OPS(add) 384f3316c2SPeter Zijlstra ATOMIC_OPS(sub) 39a439fe51SSam Ravnborg 40*358c449aSMark Rutland #define arch_atomic_add_return arch_atomic_add_return 41*358c449aSMark Rutland #define arch_atomic_sub_return arch_atomic_sub_return 42*358c449aSMark Rutland #define arch_atomic_fetch_add arch_atomic_fetch_add 43*358c449aSMark Rutland #define arch_atomic_fetch_sub arch_atomic_fetch_sub 44*358c449aSMark Rutland 45*358c449aSMark Rutland #define arch_atomic64_add_return arch_atomic64_add_return 46*358c449aSMark Rutland #define arch_atomic64_sub_return arch_atomic64_sub_return 47*358c449aSMark Rutland #define arch_atomic64_fetch_add arch_atomic64_fetch_add 48*358c449aSMark Rutland #define arch_atomic64_fetch_sub arch_atomic64_fetch_sub 49*358c449aSMark Rutland 503a1adb23SPeter Zijlstra #undef ATOMIC_OPS 513a1adb23SPeter Zijlstra #define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_FETCH_OP(op) 523a1adb23SPeter Zijlstra 533a1adb23SPeter Zijlstra ATOMIC_OPS(and) 543a1adb23SPeter Zijlstra ATOMIC_OPS(or) 553a1adb23SPeter Zijlstra ATOMIC_OPS(xor) 56304a0d69SPeter Zijlstra 57*358c449aSMark Rutland #define arch_atomic_fetch_and arch_atomic_fetch_and 58*358c449aSMark Rutland #define arch_atomic_fetch_or arch_atomic_fetch_or 59*358c449aSMark Rutland #define arch_atomic_fetch_xor arch_atomic_fetch_xor 60*358c449aSMark Rutland 61*358c449aSMark Rutland #define arch_atomic64_fetch_and arch_atomic64_fetch_and 62*358c449aSMark Rutland #define arch_atomic64_fetch_or arch_atomic64_fetch_or 63*358c449aSMark Rutland #define arch_atomic64_fetch_xor arch_atomic64_fetch_xor 64*358c449aSMark Rutland 654f3316c2SPeter Zijlstra #undef ATOMIC_OPS 663a1adb23SPeter Zijlstra #undef ATOMIC_FETCH_OP 674f3316c2SPeter Zijlstra #undef ATOMIC_OP_RETURN 684f3316c2SPeter Zijlstra #undef ATOMIC_OP 69a439fe51SSam Ravnborg 70ff5b4f1eSMark Rutland s64 arch_atomic64_dec_if_positive(atomic64_t *v); 71ff5b4f1eSMark Rutland #define arch_atomic64_dec_if_positive arch_atomic64_dec_if_positive 72193d2aadSDavid S. Miller 73a439fe51SSam Ravnborg #endif /* !(__ARCH_SPARC64_ATOMIC__) */ 74