| // Copyright 2016 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| package sync_test |
| |
| import ( |
| "internal/testenv" |
| "math/rand" |
| "reflect" |
| "runtime" |
| "sync" |
| "sync/atomic" |
| "testing" |
| "testing/quick" |
| ) |
| |
| type mapOp string |
| |
| const ( |
| opLoad = mapOp("Load") |
| opStore = mapOp("Store") |
| opLoadOrStore = mapOp("LoadOrStore") |
| opLoadAndDelete = mapOp("LoadAndDelete") |
| opDelete = mapOp("Delete") |
| opSwap = mapOp("Swap") |
| opCompareAndSwap = mapOp("CompareAndSwap") |
| opCompareAndDelete = mapOp("CompareAndDelete") |
| ) |
| |
| var mapOps = [...]mapOp{ |
| opLoad, |
| opStore, |
| opLoadOrStore, |
| opLoadAndDelete, |
| opDelete, |
| opSwap, |
| opCompareAndSwap, |
| opCompareAndDelete, |
| } |
| |
| // mapCall is a quick.Generator for calls on mapInterface. |
| type mapCall struct { |
| op mapOp |
| k, v any |
| } |
| |
| func (c mapCall) apply(m mapInterface) (any, bool) { |
| switch c.op { |
| case opLoad: |
| return m.Load(c.k) |
| case opStore: |
| m.Store(c.k, c.v) |
| return nil, false |
| case opLoadOrStore: |
| return m.LoadOrStore(c.k, c.v) |
| case opLoadAndDelete: |
| return m.LoadAndDelete(c.k) |
| case opDelete: |
| m.Delete(c.k) |
| return nil, false |
| case opSwap: |
| return m.Swap(c.k, c.v) |
| case opCompareAndSwap: |
| if m.CompareAndSwap(c.k, c.v, rand.Int()) { |
| m.Delete(c.k) |
| return c.v, true |
| } |
| return nil, false |
| case opCompareAndDelete: |
| if m.CompareAndDelete(c.k, c.v) { |
| if _, ok := m.Load(c.k); !ok { |
| return nil, true |
| } |
| } |
| return nil, false |
| default: |
| panic("invalid mapOp") |
| } |
| } |
| |
| type mapResult struct { |
| value any |
| ok bool |
| } |
| |
| func randValue(r *rand.Rand) any { |
| b := make([]byte, r.Intn(4)) |
| for i := range b { |
| b[i] = 'a' + byte(rand.Intn(26)) |
| } |
| return string(b) |
| } |
| |
| func (mapCall) Generate(r *rand.Rand, size int) reflect.Value { |
| c := mapCall{op: mapOps[rand.Intn(len(mapOps))], k: randValue(r)} |
| switch c.op { |
| case opStore, opLoadOrStore: |
| c.v = randValue(r) |
| } |
| return reflect.ValueOf(c) |
| } |
| |
| func applyCalls(m mapInterface, calls []mapCall) (results []mapResult, final map[any]any) { |
| for _, c := range calls { |
| v, ok := c.apply(m) |
| results = append(results, mapResult{v, ok}) |
| } |
| |
| final = make(map[any]any) |
| m.Range(func(k, v any) bool { |
| final[k] = v |
| return true |
| }) |
| |
| return results, final |
| } |
| |
| func applyMap(calls []mapCall) ([]mapResult, map[any]any) { |
| return applyCalls(new(sync.Map), calls) |
| } |
| |
| func applyRWMutexMap(calls []mapCall) ([]mapResult, map[any]any) { |
| return applyCalls(new(RWMutexMap), calls) |
| } |
| |
| func applyDeepCopyMap(calls []mapCall) ([]mapResult, map[any]any) { |
| return applyCalls(new(DeepCopyMap), calls) |
| } |
| |
| func TestMapMatchesRWMutex(t *testing.T) { |
| if err := quick.CheckEqual(applyMap, applyRWMutexMap, nil); err != nil { |
| t.Error(err) |
| } |
| } |
| |
| func TestMapMatchesDeepCopy(t *testing.T) { |
| if err := quick.CheckEqual(applyMap, applyDeepCopyMap, nil); err != nil { |
| t.Error(err) |
| } |
| } |
| |
| func TestConcurrentRange(t *testing.T) { |
| const mapSize = 1 << 10 |
| |
| m := new(sync.Map) |
| for n := int64(1); n <= mapSize; n++ { |
| m.Store(n, int64(n)) |
| } |
| |
| done := make(chan struct{}) |
| var wg sync.WaitGroup |
| defer func() { |
| close(done) |
| wg.Wait() |
| }() |
| for g := int64(runtime.GOMAXPROCS(0)); g > 0; g-- { |
| r := rand.New(rand.NewSource(g)) |
| wg.Add(1) |
| go func(g int64) { |
| defer wg.Done() |
| for i := int64(0); ; i++ { |
| select { |
| case <-done: |
| return |
| default: |
| } |
| for n := int64(1); n < mapSize; n++ { |
| if r.Int63n(mapSize) == 0 { |
| m.Store(n, n*i*g) |
| } else { |
| m.Load(n) |
| } |
| } |
| } |
| }(g) |
| } |
| |
| iters := 1 << 10 |
| if testing.Short() { |
| iters = 16 |
| } |
| for n := iters; n > 0; n-- { |
| seen := make(map[int64]bool, mapSize) |
| |
| m.Range(func(ki, vi any) bool { |
| k, v := ki.(int64), vi.(int64) |
| if v%k != 0 { |
| t.Fatalf("while Storing multiples of %v, Range saw value %v", k, v) |
| } |
| if seen[k] { |
| t.Fatalf("Range visited key %v twice", k) |
| } |
| seen[k] = true |
| return true |
| }) |
| |
| if len(seen) != mapSize { |
| t.Fatalf("Range visited %v elements of %v-element Map", len(seen), mapSize) |
| } |
| } |
| } |
| |
| func TestIssue40999(t *testing.T) { |
| var m sync.Map |
| |
| // Since the miss-counting in missLocked (via Delete) |
| // compares the miss count with len(m.dirty), |
| // add an initial entry to bias len(m.dirty) above the miss count. |
| m.Store(nil, struct{}{}) |
| |
| var finalized uint32 |
| |
| // Set finalizers that count for collected keys. A non-zero count |
| // indicates that keys have not been leaked. |
| for atomic.LoadUint32(&finalized) == 0 { |
| p := new(int) |
| runtime.SetFinalizer(p, func(*int) { |
| atomic.AddUint32(&finalized, 1) |
| }) |
| m.Store(p, struct{}{}) |
| m.Delete(p) |
| runtime.GC() |
| } |
| } |
| |
| func TestMapRangeNestedCall(t *testing.T) { // Issue 46399 |
| var m sync.Map |
| for i, v := range [3]string{"hello", "world", "Go"} { |
| m.Store(i, v) |
| } |
| m.Range(func(key, value any) bool { |
| m.Range(func(key, value any) bool { |
| // We should be able to load the key offered in the Range callback, |
| // because there are no concurrent Delete involved in this tested map. |
| if v, ok := m.Load(key); !ok || !reflect.DeepEqual(v, value) { |
| t.Fatalf("Nested Range loads unexpected value, got %+v want %+v", v, value) |
| } |
| |
| // We didn't keep 42 and a value into the map before, if somehow we loaded |
| // a value from such a key, meaning there must be an internal bug regarding |
| // nested range in the Map. |
| if _, loaded := m.LoadOrStore(42, "dummy"); loaded { |
| t.Fatalf("Nested Range loads unexpected value, want store a new value") |
| } |
| |
| // Try to Store then LoadAndDelete the corresponding value with the key |
| // 42 to the Map. In this case, the key 42 and associated value should be |
| // removed from the Map. Therefore any future range won't observe key 42 |
| // as we checked in above. |
| val := "sync.Map" |
| m.Store(42, val) |
| if v, loaded := m.LoadAndDelete(42); !loaded || !reflect.DeepEqual(v, val) { |
| t.Fatalf("Nested Range loads unexpected value, got %v, want %v", v, val) |
| } |
| return true |
| }) |
| |
| // Remove key from Map on-the-fly. |
| m.Delete(key) |
| return true |
| }) |
| |
| // After a Range of Delete, all keys should be removed and any |
| // further Range won't invoke the callback. Hence length remains 0. |
| length := 0 |
| m.Range(func(key, value any) bool { |
| length++ |
| return true |
| }) |
| |
| if length != 0 { |
| t.Fatalf("Unexpected sync.Map size, got %v want %v", length, 0) |
| } |
| } |
| |
| func TestCompareAndSwap_NonExistingKey(t *testing.T) { |
| m := &sync.Map{} |
| if m.CompareAndSwap(m, nil, 42) { |
| // See https://go.dev/issue/51972#issuecomment-1126408637. |
| t.Fatalf("CompareAndSwap on a non-existing key succeeded") |
| } |
| } |
| |
| func TestMapRangeNoAllocations(t *testing.T) { // Issue 62404 |
| testenv.SkipIfOptimizationOff(t) |
| var m sync.Map |
| allocs := testing.AllocsPerRun(10, func() { |
| m.Range(func(key, value any) bool { |
| return true |
| }) |
| }) |
| if allocs > 0 { |
| t.Errorf("AllocsPerRun of m.Range = %v; want 0", allocs) |
| } |
| } |