| // Copyright 2014 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // RISC-V's atomic operations have two bits, aq ("acquire") and rl ("release"), |
| // which may be toggled on and off. Their precise semantics are defined in |
| // section 6.3 of the specification, but the basic idea is as follows: |
| // |
| // - If neither aq nor rl is set, the CPU may reorder the atomic arbitrarily. |
| // It guarantees only that it will execute atomically. |
| // |
| // - If aq is set, the CPU may move the instruction backward, but not forward. |
| // |
| // - If rl is set, the CPU may move the instruction forward, but not backward. |
| // |
| // - If both are set, the CPU may not reorder the instruction at all. |
| // |
| // These four modes correspond to other well-known memory models on other CPUs. |
| // On ARM, aq corresponds to a dmb ishst, aq+rl corresponds to a dmb ish. On |
| // Intel, aq corresponds to an lfence, rl to an sfence, and aq+rl to an mfence |
| // (or a lock prefix). |
| // |
| // Go's memory model requires that |
| // - if a read happens after a write, the read must observe the write, and |
| // that |
| // - if a read happens concurrently with a write, the read may observe the |
| // write. |
| // aq is sufficient to guarantee this, so that's what we use here. (This jibes |
| // with ARM, which uses dmb ishst.) |
| |
| #include "textflag.h" |
| |
| // func Cas(ptr *uint64, old, new uint64) bool |
| // Atomically: |
| // if(*val == old){ |
| // *val = new; |
| // return 1; |
| // } else { |
| // return 0; |
| // } |
| TEXT ·Cas(SB), NOSPLIT, $0-17 |
| MOV ptr+0(FP), A0 |
| MOVW old+8(FP), A1 |
| MOVW new+12(FP), A2 |
| cas_again: |
| LRW (A0), A3 |
| BNE A3, A1, cas_fail |
| SCW A2, (A0), A4 |
| BNE A4, ZERO, cas_again |
| MOV $1, A0 |
| MOVB A0, ret+16(FP) |
| RET |
| cas_fail: |
| MOV $0, A0 |
| MOV A0, ret+16(FP) |
| RET |
| |
| // func Cas64(ptr *uint64, old, new uint64) bool |
| TEXT ·Cas64(SB), NOSPLIT, $0-25 |
| MOV ptr+0(FP), A0 |
| MOV old+8(FP), A1 |
| MOV new+16(FP), A2 |
| cas_again: |
| LRD (A0), A3 |
| BNE A3, A1, cas_fail |
| SCD A2, (A0), A4 |
| BNE A4, ZERO, cas_again |
| MOV $1, A0 |
| MOVB A0, ret+24(FP) |
| RET |
| cas_fail: |
| MOVB ZERO, ret+24(FP) |
| RET |
| |
| // func Load(ptr *uint32) uint32 |
| TEXT ·Load(SB),NOSPLIT|NOFRAME,$0-12 |
| MOV ptr+0(FP), A0 |
| LRW (A0), A0 |
| MOVW A0, ret+8(FP) |
| RET |
| |
| // func Load8(ptr *uint8) uint8 |
| TEXT ·Load8(SB),NOSPLIT|NOFRAME,$0-9 |
| MOV ptr+0(FP), A0 |
| FENCE |
| MOVBU (A0), A1 |
| FENCE |
| MOVB A1, ret+8(FP) |
| RET |
| |
| // func Load64(ptr *uint64) uint64 |
| TEXT ·Load64(SB),NOSPLIT|NOFRAME,$0-16 |
| MOV ptr+0(FP), A0 |
| LRD (A0), A0 |
| MOV A0, ret+8(FP) |
| RET |
| |
| // func Store(ptr *uint32, val uint32) |
| TEXT ·Store(SB), NOSPLIT, $0-12 |
| MOV ptr+0(FP), A0 |
| MOVW val+8(FP), A1 |
| AMOSWAPW A1, (A0), ZERO |
| RET |
| |
| // func Store8(ptr *uint8, val uint8) |
| TEXT ·Store8(SB), NOSPLIT, $0-9 |
| MOV ptr+0(FP), A0 |
| MOVBU val+8(FP), A1 |
| FENCE |
| MOVB A1, (A0) |
| FENCE |
| RET |
| |
| // func Store64(ptr *uint64, val uint64) |
| TEXT ·Store64(SB), NOSPLIT, $0-16 |
| MOV ptr+0(FP), A0 |
| MOV val+8(FP), A1 |
| AMOSWAPD A1, (A0), ZERO |
| RET |
| |
| TEXT ·Casp1(SB), NOSPLIT, $0-25 |
| JMP ·Cas64(SB) |
| |
| TEXT ·Casint32(SB),NOSPLIT,$0-17 |
| JMP ·Cas(SB) |
| |
| TEXT ·Casint64(SB),NOSPLIT,$0-25 |
| JMP ·Cas64(SB) |
| |
| TEXT ·Casuintptr(SB),NOSPLIT,$0-25 |
| JMP ·Cas64(SB) |
| |
| TEXT ·CasRel(SB), NOSPLIT, $0-17 |
| JMP ·Cas(SB) |
| |
| TEXT ·Loaduintptr(SB),NOSPLIT,$0-16 |
| JMP ·Load64(SB) |
| |
| TEXT ·Storeint32(SB),NOSPLIT,$0-12 |
| JMP ·Store(SB) |
| |
| TEXT ·Storeint64(SB),NOSPLIT,$0-16 |
| JMP ·Store64(SB) |
| |
| TEXT ·Storeuintptr(SB),NOSPLIT,$0-16 |
| JMP ·Store64(SB) |
| |
| TEXT ·Loaduint(SB),NOSPLIT,$0-16 |
| JMP ·Loaduintptr(SB) |
| |
| TEXT ·Loadint32(SB),NOSPLIT,$0-12 |
| JMP ·Load(SB) |
| |
| TEXT ·Loadint64(SB),NOSPLIT,$0-16 |
| JMP ·Load64(SB) |
| |
| TEXT ·Xaddint32(SB),NOSPLIT,$0-20 |
| JMP ·Xadd(SB) |
| |
| TEXT ·Xaddint64(SB),NOSPLIT,$0-24 |
| MOV ptr+0(FP), A0 |
| MOV delta+8(FP), A1 |
| AMOADDD A1, (A0), A0 |
| ADD A0, A1, A0 |
| MOVW A0, ret+16(FP) |
| RET |
| |
| TEXT ·LoadAcq(SB),NOSPLIT|NOFRAME,$0-12 |
| JMP ·Load(SB) |
| |
| TEXT ·LoadAcq64(SB),NOSPLIT|NOFRAME,$0-16 |
| JMP ·Load64(SB) |
| |
| TEXT ·LoadAcquintptr(SB),NOSPLIT|NOFRAME,$0-16 |
| JMP ·Load64(SB) |
| |
| // func Loadp(ptr unsafe.Pointer) unsafe.Pointer |
| TEXT ·Loadp(SB),NOSPLIT,$0-16 |
| JMP ·Load64(SB) |
| |
| // func StorepNoWB(ptr unsafe.Pointer, val unsafe.Pointer) |
| TEXT ·StorepNoWB(SB), NOSPLIT, $0-16 |
| JMP ·Store64(SB) |
| |
| TEXT ·StoreRel(SB), NOSPLIT, $0-12 |
| JMP ·Store(SB) |
| |
| TEXT ·StoreRel64(SB), NOSPLIT, $0-16 |
| JMP ·Store64(SB) |
| |
| TEXT ·StoreReluintptr(SB), NOSPLIT, $0-16 |
| JMP ·Store64(SB) |
| |
| // func Xchg(ptr *uint32, new uint32) uint32 |
| TEXT ·Xchg(SB), NOSPLIT, $0-20 |
| MOV ptr+0(FP), A0 |
| MOVW new+8(FP), A1 |
| AMOSWAPW A1, (A0), A1 |
| MOVW A1, ret+16(FP) |
| RET |
| |
| // func Xchg64(ptr *uint64, new uint64) uint64 |
| TEXT ·Xchg64(SB), NOSPLIT, $0-24 |
| MOV ptr+0(FP), A0 |
| MOV new+8(FP), A1 |
| AMOSWAPD A1, (A0), A1 |
| MOV A1, ret+16(FP) |
| RET |
| |
| // Atomically: |
| // *val += delta; |
| // return *val; |
| |
| // func Xadd(ptr *uint32, delta int32) uint32 |
| TEXT ·Xadd(SB), NOSPLIT, $0-20 |
| MOV ptr+0(FP), A0 |
| MOVW delta+8(FP), A1 |
| AMOADDW A1, (A0), A2 |
| ADD A2,A1,A0 |
| MOVW A0, ret+16(FP) |
| RET |
| |
| // func Xadd64(ptr *uint64, delta int64) uint64 |
| TEXT ·Xadd64(SB), NOSPLIT, $0-24 |
| MOV ptr+0(FP), A0 |
| MOV delta+8(FP), A1 |
| AMOADDD A1, (A0), A2 |
| ADD A2, A1, A0 |
| MOV A0, ret+16(FP) |
| RET |
| |
| // func Xadduintptr(ptr *uintptr, delta uintptr) uintptr |
| TEXT ·Xadduintptr(SB), NOSPLIT, $0-24 |
| JMP ·Xadd64(SB) |
| |
| // func Xchgint32(ptr *int32, new int32) int32 |
| TEXT ·Xchgint32(SB), NOSPLIT, $0-20 |
| JMP ·Xchg(SB) |
| |
| // func Xchgint64(ptr *int64, new int64) int64 |
| TEXT ·Xchgint64(SB), NOSPLIT, $0-24 |
| JMP ·Xchg64(SB) |
| |
| // func Xchguintptr(ptr *uintptr, new uintptr) uintptr |
| TEXT ·Xchguintptr(SB), NOSPLIT, $0-24 |
| JMP ·Xchg64(SB) |
| |
| // func And8(ptr *uint8, val uint8) |
| TEXT ·And8(SB), NOSPLIT, $0-9 |
| MOV ptr+0(FP), A0 |
| MOVBU val+8(FP), A1 |
| AND $3, A0, A2 |
| AND $-4, A0 |
| SLL $3, A2 |
| XOR $255, A1 |
| SLL A2, A1 |
| XOR $-1, A1 |
| AMOANDW A1, (A0), ZERO |
| RET |
| |
| // func Or8(ptr *uint8, val uint8) |
| TEXT ·Or8(SB), NOSPLIT, $0-9 |
| MOV ptr+0(FP), A0 |
| MOVBU val+8(FP), A1 |
| AND $3, A0, A2 |
| AND $-4, A0 |
| SLL $3, A2 |
| SLL A2, A1 |
| AMOORW A1, (A0), ZERO |
| RET |
| |
| // func And(ptr *uint32, val uint32) |
| TEXT ·And(SB), NOSPLIT, $0-12 |
| MOV ptr+0(FP), A0 |
| MOVW val+8(FP), A1 |
| AMOANDW A1, (A0), ZERO |
| RET |
| |
| // func Or(ptr *uint32, val uint32) |
| TEXT ·Or(SB), NOSPLIT, $0-12 |
| MOV ptr+0(FP), A0 |
| MOVW val+8(FP), A1 |
| AMOORW A1, (A0), ZERO |
| RET |