| // Copyright 2014 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| package testing_test |
| |
| import ( |
| "bytes" |
| "internal/race" |
| "internal/testenv" |
| "os" |
| "path/filepath" |
| "regexp" |
| "sync" |
| "testing" |
| ) |
| |
| // This is exactly what a test would do without a TestMain. |
| // It's here only so that there is at least one package in the |
| // standard library with a TestMain, so that code is executed. |
| |
| func TestMain(m *testing.M) { |
| if os.Getenv("GO_WANT_RACE_BEFORE_TESTS") == "1" { |
| doRace() |
| } |
| |
| m.Run() |
| |
| // Note: m.Run currently prints the final "PASS" line, so if any race is |
| // reported here (after m.Run but before the process exits), it will print |
| // "PASS", then print the stack traces for the race, then exit with nonzero |
| // status. |
| // |
| // This is a somewhat fundamental race: because the race detector hooks into |
| // the runtime at a very low level, no matter where we put the printing it |
| // would be possible to report a race that occurs afterward. However, we could |
| // theoretically move the printing after TestMain, which would at least do a |
| // better job of diagnosing races in cleanup functions within TestMain itself. |
| } |
| |
| func TestTempDirInCleanup(t *testing.T) { |
| var dir string |
| |
| t.Run("test", func(t *testing.T) { |
| t.Cleanup(func() { |
| dir = t.TempDir() |
| }) |
| _ = t.TempDir() |
| }) |
| |
| fi, err := os.Stat(dir) |
| if fi != nil { |
| t.Fatalf("Directory %q from user Cleanup still exists", dir) |
| } |
| if !os.IsNotExist(err) { |
| t.Fatalf("Unexpected error: %v", err) |
| } |
| } |
| |
| func TestTempDirInBenchmark(t *testing.T) { |
| testing.Benchmark(func(b *testing.B) { |
| if !b.Run("test", func(b *testing.B) { |
| // Add a loop so that the test won't fail. See issue 38677. |
| for i := 0; i < b.N; i++ { |
| _ = b.TempDir() |
| } |
| }) { |
| t.Fatal("Sub test failure in a benchmark") |
| } |
| }) |
| } |
| |
| func TestTempDir(t *testing.T) { |
| testTempDir(t) |
| t.Run("InSubtest", testTempDir) |
| t.Run("test/subtest", testTempDir) |
| t.Run("test\\subtest", testTempDir) |
| t.Run("test:subtest", testTempDir) |
| t.Run("test/..", testTempDir) |
| t.Run("../test", testTempDir) |
| t.Run("test[]", testTempDir) |
| t.Run("test*", testTempDir) |
| t.Run("äöüéè", testTempDir) |
| } |
| |
| func testTempDir(t *testing.T) { |
| dirCh := make(chan string, 1) |
| t.Cleanup(func() { |
| // Verify directory has been removed. |
| select { |
| case dir := <-dirCh: |
| fi, err := os.Stat(dir) |
| if os.IsNotExist(err) { |
| // All good |
| return |
| } |
| if err != nil { |
| t.Fatal(err) |
| } |
| t.Errorf("directory %q still exists: %v, isDir=%v", dir, fi, fi.IsDir()) |
| default: |
| if !t.Failed() { |
| t.Fatal("never received dir channel") |
| } |
| } |
| }) |
| |
| dir := t.TempDir() |
| if dir == "" { |
| t.Fatal("expected dir") |
| } |
| dir2 := t.TempDir() |
| if dir == dir2 { |
| t.Fatal("subsequent calls to TempDir returned the same directory") |
| } |
| if filepath.Dir(dir) != filepath.Dir(dir2) { |
| t.Fatalf("calls to TempDir do not share a parent; got %q, %q", dir, dir2) |
| } |
| dirCh <- dir |
| fi, err := os.Stat(dir) |
| if err != nil { |
| t.Fatal(err) |
| } |
| if !fi.IsDir() { |
| t.Errorf("dir %q is not a dir", dir) |
| } |
| files, err := os.ReadDir(dir) |
| if err != nil { |
| t.Fatal(err) |
| } |
| if len(files) > 0 { |
| t.Errorf("unexpected %d files in TempDir: %v", len(files), files) |
| } |
| |
| glob := filepath.Join(dir, "*.txt") |
| if _, err := filepath.Glob(glob); err != nil { |
| t.Error(err) |
| } |
| } |
| |
| func TestSetenv(t *testing.T) { |
| tests := []struct { |
| name string |
| key string |
| initialValueExists bool |
| initialValue string |
| newValue string |
| }{ |
| { |
| name: "initial value exists", |
| key: "GO_TEST_KEY_1", |
| initialValueExists: true, |
| initialValue: "111", |
| newValue: "222", |
| }, |
| { |
| name: "initial value exists but empty", |
| key: "GO_TEST_KEY_2", |
| initialValueExists: true, |
| initialValue: "", |
| newValue: "222", |
| }, |
| { |
| name: "initial value is not exists", |
| key: "GO_TEST_KEY_3", |
| initialValueExists: false, |
| initialValue: "", |
| newValue: "222", |
| }, |
| } |
| |
| for _, test := range tests { |
| if test.initialValueExists { |
| if err := os.Setenv(test.key, test.initialValue); err != nil { |
| t.Fatalf("unable to set env: got %v", err) |
| } |
| } else { |
| os.Unsetenv(test.key) |
| } |
| |
| t.Run(test.name, func(t *testing.T) { |
| t.Setenv(test.key, test.newValue) |
| if os.Getenv(test.key) != test.newValue { |
| t.Fatalf("unexpected value after t.Setenv: got %s, want %s", os.Getenv(test.key), test.newValue) |
| } |
| }) |
| |
| got, exists := os.LookupEnv(test.key) |
| if got != test.initialValue { |
| t.Fatalf("unexpected value after t.Setenv cleanup: got %s, want %s", got, test.initialValue) |
| } |
| if exists != test.initialValueExists { |
| t.Fatalf("unexpected value after t.Setenv cleanup: got %t, want %t", exists, test.initialValueExists) |
| } |
| } |
| } |
| |
| func TestSetenvWithParallelAfterSetenv(t *testing.T) { |
| defer func() { |
| want := "testing: t.Parallel called after t.Setenv; cannot set environment variables in parallel tests" |
| if got := recover(); got != want { |
| t.Fatalf("expected panic; got %#v want %q", got, want) |
| } |
| }() |
| |
| t.Setenv("GO_TEST_KEY_1", "value") |
| |
| t.Parallel() |
| } |
| |
| func TestSetenvWithParallelBeforeSetenv(t *testing.T) { |
| defer func() { |
| want := "testing: t.Setenv called after t.Parallel; cannot set environment variables in parallel tests" |
| if got := recover(); got != want { |
| t.Fatalf("expected panic; got %#v want %q", got, want) |
| } |
| }() |
| |
| t.Parallel() |
| |
| t.Setenv("GO_TEST_KEY_1", "value") |
| } |
| |
| func TestSetenvWithParallelParentBeforeSetenv(t *testing.T) { |
| t.Parallel() |
| |
| t.Run("child", func(t *testing.T) { |
| defer func() { |
| want := "testing: t.Setenv called after t.Parallel; cannot set environment variables in parallel tests" |
| if got := recover(); got != want { |
| t.Fatalf("expected panic; got %#v want %q", got, want) |
| } |
| }() |
| |
| t.Setenv("GO_TEST_KEY_1", "value") |
| }) |
| } |
| |
| func TestSetenvWithParallelGrandParentBeforeSetenv(t *testing.T) { |
| t.Parallel() |
| |
| t.Run("child", func(t *testing.T) { |
| t.Run("grand-child", func(t *testing.T) { |
| defer func() { |
| want := "testing: t.Setenv called after t.Parallel; cannot set environment variables in parallel tests" |
| if got := recover(); got != want { |
| t.Fatalf("expected panic; got %#v want %q", got, want) |
| } |
| }() |
| |
| t.Setenv("GO_TEST_KEY_1", "value") |
| }) |
| }) |
| } |
| |
| // testingTrueInInit is part of TestTesting. |
| var testingTrueInInit = false |
| |
| // testingTrueInPackageVarInit is part of TestTesting. |
| var testingTrueInPackageVarInit = testing.Testing() |
| |
| // init is part of TestTesting. |
| func init() { |
| if testing.Testing() { |
| testingTrueInInit = true |
| } |
| } |
| |
| var testingProg = ` |
| package main |
| |
| import ( |
| "fmt" |
| "testing" |
| ) |
| |
| func main() { |
| fmt.Println(testing.Testing()) |
| } |
| ` |
| |
| func TestTesting(t *testing.T) { |
| if !testing.Testing() { |
| t.Errorf("testing.Testing() == %t, want %t", testing.Testing(), true) |
| } |
| if !testingTrueInInit { |
| t.Errorf("testing.Testing() called by init function == %t, want %t", testingTrueInInit, true) |
| } |
| if !testingTrueInPackageVarInit { |
| t.Errorf("testing.Testing() variable initialized as %t, want %t", testingTrueInPackageVarInit, true) |
| } |
| |
| if testing.Short() { |
| t.Skip("skipping building a binary in short mode") |
| } |
| testenv.MustHaveGoRun(t) |
| |
| fn := filepath.Join(t.TempDir(), "x.go") |
| if err := os.WriteFile(fn, []byte(testingProg), 0644); err != nil { |
| t.Fatal(err) |
| } |
| |
| cmd := testenv.Command(t, testenv.GoToolPath(t), "run", fn) |
| out, err := cmd.CombinedOutput() |
| if err != nil { |
| t.Fatalf("%v failed: %v\n%s", cmd, err, out) |
| } |
| |
| s := string(bytes.TrimSpace(out)) |
| if s != "false" { |
| t.Errorf("in non-test testing.Test() returned %q, want %q", s, "false") |
| } |
| } |
| |
| // runTest runs a helper test with -test.v, ignoring its exit status. |
| // runTest both logs and returns the test output. |
| func runTest(t *testing.T, test string) []byte { |
| t.Helper() |
| |
| testenv.MustHaveExec(t) |
| |
| exe, err := os.Executable() |
| if err != nil { |
| t.Skipf("can't find test executable: %v", err) |
| } |
| |
| cmd := testenv.Command(t, exe, "-test.run=^"+test+"$", "-test.bench="+test, "-test.v", "-test.parallel=2", "-test.benchtime=2x") |
| cmd = testenv.CleanCmdEnv(cmd) |
| cmd.Env = append(cmd.Env, "GO_WANT_HELPER_PROCESS=1") |
| out, err := cmd.CombinedOutput() |
| t.Logf("%v: %v\n%s", cmd, err, out) |
| |
| return out |
| } |
| |
| // doRace provokes a data race that generates a race detector report if run |
| // under the race detector and is otherwise benign. |
| func doRace() { |
| var x int |
| c1 := make(chan bool) |
| go func() { |
| x = 1 // racy write |
| c1 <- true |
| }() |
| _ = x // racy read |
| <-c1 |
| } |
| |
| func TestRaceReports(t *testing.T) { |
| if os.Getenv("GO_WANT_HELPER_PROCESS") == "1" { |
| // Generate a race detector report in a sub test. |
| t.Run("Sub", func(t *testing.T) { |
| doRace() |
| }) |
| return |
| } |
| |
| out := runTest(t, "TestRaceReports") |
| |
| // We should see at most one race detector report. |
| c := bytes.Count(out, []byte("race detected")) |
| want := 0 |
| if race.Enabled { |
| want = 1 |
| } |
| if c != want { |
| t.Errorf("got %d race reports, want %d", c, want) |
| } |
| } |
| |
| // Issue #60083. This used to fail on the race builder. |
| func TestRaceName(t *testing.T) { |
| if os.Getenv("GO_WANT_HELPER_PROCESS") == "1" { |
| doRace() |
| return |
| } |
| |
| out := runTest(t, "TestRaceName") |
| |
| if regexp.MustCompile(`=== NAME\s*$`).Match(out) { |
| t.Errorf("incorrectly reported test with no name") |
| } |
| } |
| |
| func TestRaceSubReports(t *testing.T) { |
| if os.Getenv("GO_WANT_HELPER_PROCESS") == "1" { |
| t.Parallel() |
| c1 := make(chan bool, 1) |
| t.Run("sub", func(t *testing.T) { |
| t.Run("subsub1", func(t *testing.T) { |
| t.Parallel() |
| doRace() |
| c1 <- true |
| }) |
| t.Run("subsub2", func(t *testing.T) { |
| t.Parallel() |
| doRace() |
| <-c1 |
| }) |
| }) |
| doRace() |
| return |
| } |
| |
| out := runTest(t, "TestRaceSubReports") |
| |
| // There should be three race reports: one for each subtest, and one for the |
| // race after the subtests complete. Note that because the subtests run in |
| // parallel, the race stacks may both be printed in with one or the other |
| // test's logs. |
| cReport := bytes.Count(out, []byte("race detected during execution of test")) |
| wantReport := 0 |
| if race.Enabled { |
| wantReport = 3 |
| } |
| if cReport != wantReport { |
| t.Errorf("got %d race reports, want %d", cReport, wantReport) |
| } |
| |
| // Regardless of when the stacks are printed, we expect each subtest to be |
| // marked as failed, and that failure should propagate up to the parents. |
| cFail := bytes.Count(out, []byte("--- FAIL:")) |
| wantFail := 0 |
| if race.Enabled { |
| wantFail = 4 |
| } |
| if cFail != wantFail { |
| t.Errorf(`got %d "--- FAIL:" lines, want %d`, cReport, wantReport) |
| } |
| } |
| |
| func TestRaceInCleanup(t *testing.T) { |
| if os.Getenv("GO_WANT_HELPER_PROCESS") == "1" { |
| t.Cleanup(doRace) |
| t.Parallel() |
| t.Run("sub", func(t *testing.T) { |
| t.Parallel() |
| // No race should be reported for sub. |
| }) |
| return |
| } |
| |
| out := runTest(t, "TestRaceInCleanup") |
| |
| // There should be one race report, for the parent test only. |
| cReport := bytes.Count(out, []byte("race detected during execution of test")) |
| wantReport := 0 |
| if race.Enabled { |
| wantReport = 1 |
| } |
| if cReport != wantReport { |
| t.Errorf("got %d race reports, want %d", cReport, wantReport) |
| } |
| |
| // Only the parent test should be marked as failed. |
| // (The subtest does not race, and should pass.) |
| cFail := bytes.Count(out, []byte("--- FAIL:")) |
| wantFail := 0 |
| if race.Enabled { |
| wantFail = 1 |
| } |
| if cFail != wantFail { |
| t.Errorf(`got %d "--- FAIL:" lines, want %d`, cReport, wantReport) |
| } |
| } |
| |
| func TestDeepSubtestRace(t *testing.T) { |
| if os.Getenv("GO_WANT_HELPER_PROCESS") == "1" { |
| t.Run("sub", func(t *testing.T) { |
| t.Run("subsub", func(t *testing.T) { |
| t.Run("subsubsub", func(t *testing.T) { |
| doRace() |
| }) |
| }) |
| doRace() |
| }) |
| return |
| } |
| |
| out := runTest(t, "TestDeepSubtestRace") |
| |
| c := bytes.Count(out, []byte("race detected during execution of test")) |
| want := 0 |
| // There should be two race reports. |
| if race.Enabled { |
| want = 2 |
| } |
| if c != want { |
| t.Errorf("got %d race reports, want %d", c, want) |
| } |
| } |
| |
| func TestRaceDuringParallelFailsAllSubtests(t *testing.T) { |
| if os.Getenv("GO_WANT_HELPER_PROCESS") == "1" { |
| var ready sync.WaitGroup |
| ready.Add(2) |
| done := make(chan struct{}) |
| go func() { |
| ready.Wait() |
| doRace() // This race happens while both subtests are running. |
| close(done) |
| }() |
| |
| t.Run("sub", func(t *testing.T) { |
| t.Run("subsub1", func(t *testing.T) { |
| t.Parallel() |
| ready.Done() |
| <-done |
| }) |
| t.Run("subsub2", func(t *testing.T) { |
| t.Parallel() |
| ready.Done() |
| <-done |
| }) |
| }) |
| |
| return |
| } |
| |
| out := runTest(t, "TestRaceDuringParallelFailsAllSubtests") |
| |
| c := bytes.Count(out, []byte("race detected during execution of test")) |
| want := 0 |
| // Each subtest should report the race independently. |
| if race.Enabled { |
| want = 2 |
| } |
| if c != want { |
| t.Errorf("got %d race reports, want %d", c, want) |
| } |
| } |
| |
| func TestRaceBeforeParallel(t *testing.T) { |
| if os.Getenv("GO_WANT_HELPER_PROCESS") == "1" { |
| t.Run("sub", func(t *testing.T) { |
| doRace() |
| t.Parallel() |
| }) |
| return |
| } |
| |
| out := runTest(t, "TestRaceBeforeParallel") |
| |
| c := bytes.Count(out, []byte("race detected during execution of test")) |
| want := 0 |
| // We should see one race detector report. |
| if race.Enabled { |
| want = 1 |
| } |
| if c != want { |
| t.Errorf("got %d race reports, want %d", c, want) |
| } |
| } |
| |
| func TestRaceBeforeTests(t *testing.T) { |
| testenv.MustHaveExec(t) |
| |
| exe, err := os.Executable() |
| if err != nil { |
| t.Skipf("can't find test executable: %v", err) |
| } |
| |
| cmd := testenv.Command(t, exe, "-test.run=^$") |
| cmd = testenv.CleanCmdEnv(cmd) |
| cmd.Env = append(cmd.Env, "GO_WANT_RACE_BEFORE_TESTS=1") |
| out, _ := cmd.CombinedOutput() |
| t.Logf("%s", out) |
| |
| c := bytes.Count(out, []byte("race detected outside of test execution")) |
| |
| want := 0 |
| if race.Enabled { |
| want = 1 |
| } |
| if c != want { |
| t.Errorf("got %d race reports; want %d", c, want) |
| } |
| } |
| |
| func TestBenchmarkRace(t *testing.T) { |
| out := runTest(t, "BenchmarkRacy") |
| c := bytes.Count(out, []byte("race detected during execution of test")) |
| |
| want := 0 |
| // We should see one race detector report. |
| if race.Enabled { |
| want = 1 |
| } |
| if c != want { |
| t.Errorf("got %d race reports; want %d", c, want) |
| } |
| } |
| |
| func BenchmarkRacy(b *testing.B) { |
| if os.Getenv("GO_WANT_HELPER_PROCESS") != "1" { |
| b.Skipf("skipping intentionally-racy benchmark") |
| } |
| for i := 0; i < b.N; i++ { |
| doRace() |
| } |
| } |
| |
| func TestBenchmarkSubRace(t *testing.T) { |
| out := runTest(t, "BenchmarkSubRacy") |
| c := bytes.Count(out, []byte("race detected during execution of test")) |
| |
| want := 0 |
| // We should see two race detector reports: |
| // one in the sub-bencmark, and one in the parent afterward. |
| if race.Enabled { |
| want = 2 |
| } |
| if c != want { |
| t.Errorf("got %d race reports; want %d", c, want) |
| } |
| } |
| |
| func BenchmarkSubRacy(b *testing.B) { |
| if os.Getenv("GO_WANT_HELPER_PROCESS") != "1" { |
| b.Skipf("skipping intentionally-racy benchmark") |
| } |
| |
| b.Run("non-racy", func(b *testing.B) { |
| tot := 0 |
| for i := 0; i < b.N; i++ { |
| tot++ |
| } |
| _ = tot |
| }) |
| |
| b.Run("racy", func(b *testing.B) { |
| for i := 0; i < b.N; i++ { |
| doRace() |
| } |
| }) |
| |
| doRace() // should be reported separately |
| } |