|  | // Copyright 2019 The Go Authors. All rights reserved. | 
|  | // Use of this source code is governed by a BSD-style | 
|  | // license that can be found in the LICENSE file. | 
|  |  | 
|  | package runtime_test | 
|  |  | 
|  | import ( | 
|  | . "runtime" | 
|  | "sync" | 
|  | "sync/atomic" | 
|  | "testing" | 
|  | ) | 
|  |  | 
|  | // TestSemaHandoff checks that when semrelease+handoff is | 
|  | // requested, the G that releases the semaphore yields its | 
|  | // P directly to the first waiter in line. | 
|  | // See issue 33747 for discussion. | 
|  | func TestSemaHandoff(t *testing.T) { | 
|  | const iter = 10000 | 
|  | ok := 0 | 
|  | for i := 0; i < iter; i++ { | 
|  | if testSemaHandoff() { | 
|  | ok++ | 
|  | } | 
|  | } | 
|  | // As long as two thirds of handoffs are direct, we | 
|  | // consider the test successful. The scheduler is | 
|  | // nondeterministic, so this test checks that we get the | 
|  | // desired outcome in a significant majority of cases. | 
|  | // The actual ratio of direct handoffs is much higher | 
|  | // (>90%) but we use a lower threshold to minimize the | 
|  | // chances that unrelated changes in the runtime will | 
|  | // cause the test to fail or become flaky. | 
|  | if ok < iter*2/3 { | 
|  | t.Fatal("direct handoff < 2/3:", ok, iter) | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestSemaHandoff1(t *testing.T) { | 
|  | if GOMAXPROCS(-1) <= 1 { | 
|  | t.Skip("GOMAXPROCS <= 1") | 
|  | } | 
|  | defer GOMAXPROCS(GOMAXPROCS(-1)) | 
|  | GOMAXPROCS(1) | 
|  | TestSemaHandoff(t) | 
|  | } | 
|  |  | 
|  | func TestSemaHandoff2(t *testing.T) { | 
|  | if GOMAXPROCS(-1) <= 2 { | 
|  | t.Skip("GOMAXPROCS <= 2") | 
|  | } | 
|  | defer GOMAXPROCS(GOMAXPROCS(-1)) | 
|  | GOMAXPROCS(2) | 
|  | TestSemaHandoff(t) | 
|  | } | 
|  |  | 
|  | func testSemaHandoff() bool { | 
|  | var sema, res uint32 | 
|  | done := make(chan struct{}) | 
|  |  | 
|  | // We're testing that the current goroutine is able to yield its time slice | 
|  | // to another goroutine. Stop the current goroutine from migrating to | 
|  | // another CPU where it can win the race (and appear to have not yielded) by | 
|  | // keeping the CPUs slightly busy. | 
|  | var wg sync.WaitGroup | 
|  | for i := 0; i < GOMAXPROCS(-1); i++ { | 
|  | wg.Add(1) | 
|  | go func() { | 
|  | defer wg.Done() | 
|  | for { | 
|  | select { | 
|  | case <-done: | 
|  | return | 
|  | default: | 
|  | } | 
|  | Gosched() | 
|  | } | 
|  | }() | 
|  | } | 
|  |  | 
|  | wg.Add(1) | 
|  | go func() { | 
|  | defer wg.Done() | 
|  | Semacquire(&sema) | 
|  | atomic.CompareAndSwapUint32(&res, 0, 1) | 
|  |  | 
|  | Semrelease1(&sema, true, 0) | 
|  | close(done) | 
|  | }() | 
|  | for SemNwait(&sema) == 0 { | 
|  | Gosched() // wait for goroutine to block in Semacquire | 
|  | } | 
|  |  | 
|  | // The crux of the test: we release the semaphore with handoff | 
|  | // and immediately perform a CAS both here and in the waiter; we | 
|  | // want the CAS in the waiter to execute first. | 
|  | Semrelease1(&sema, true, 0) | 
|  | atomic.CompareAndSwapUint32(&res, 0, 2) | 
|  |  | 
|  | wg.Wait() // wait for goroutines to finish to avoid data races | 
|  |  | 
|  | return res == 1 // did the waiter run first? | 
|  | } |