| // Copyright 2013 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| package debug_test |
| |
| import ( |
| "internal/testenv" |
| "os" |
| "runtime" |
| . "runtime/debug" |
| "testing" |
| "time" |
| ) |
| |
| func TestReadGCStats(t *testing.T) { |
| defer SetGCPercent(SetGCPercent(-1)) |
| |
| var stats GCStats |
| var mstats runtime.MemStats |
| var min, max time.Duration |
| |
| // First ReadGCStats will allocate, second should not, |
| // especially if we follow up with an explicit garbage collection. |
| stats.PauseQuantiles = make([]time.Duration, 10) |
| ReadGCStats(&stats) |
| runtime.GC() |
| |
| // Assume these will return same data: no GC during ReadGCStats. |
| ReadGCStats(&stats) |
| runtime.ReadMemStats(&mstats) |
| |
| if stats.NumGC != int64(mstats.NumGC) { |
| t.Errorf("stats.NumGC = %d, but mstats.NumGC = %d", stats.NumGC, mstats.NumGC) |
| } |
| if stats.PauseTotal != time.Duration(mstats.PauseTotalNs) { |
| t.Errorf("stats.PauseTotal = %d, but mstats.PauseTotalNs = %d", stats.PauseTotal, mstats.PauseTotalNs) |
| } |
| if stats.LastGC.UnixNano() != int64(mstats.LastGC) { |
| t.Errorf("stats.LastGC.UnixNano = %d, but mstats.LastGC = %d", stats.LastGC.UnixNano(), mstats.LastGC) |
| } |
| n := int(mstats.NumGC) |
| if n > len(mstats.PauseNs) { |
| n = len(mstats.PauseNs) |
| } |
| if len(stats.Pause) != n { |
| t.Errorf("len(stats.Pause) = %d, want %d", len(stats.Pause), n) |
| } else { |
| off := (int(mstats.NumGC) + len(mstats.PauseNs) - 1) % len(mstats.PauseNs) |
| for i := 0; i < n; i++ { |
| dt := stats.Pause[i] |
| if dt != time.Duration(mstats.PauseNs[off]) { |
| t.Errorf("stats.Pause[%d] = %d, want %d", i, dt, mstats.PauseNs[off]) |
| } |
| if max < dt { |
| max = dt |
| } |
| if min > dt || i == 0 { |
| min = dt |
| } |
| off = (off + len(mstats.PauseNs) - 1) % len(mstats.PauseNs) |
| } |
| } |
| |
| q := stats.PauseQuantiles |
| nq := len(q) |
| if q[0] != min || q[nq-1] != max { |
| t.Errorf("stats.PauseQuantiles = [%d, ..., %d], want [%d, ..., %d]", q[0], q[nq-1], min, max) |
| } |
| |
| for i := 0; i < nq-1; i++ { |
| if q[i] > q[i+1] { |
| t.Errorf("stats.PauseQuantiles[%d]=%d > stats.PauseQuantiles[%d]=%d", i, q[i], i+1, q[i+1]) |
| } |
| } |
| |
| // compare memory stats with gc stats: |
| if len(stats.PauseEnd) != n { |
| t.Fatalf("len(stats.PauseEnd) = %d, want %d", len(stats.PauseEnd), n) |
| } |
| off := (int(mstats.NumGC) + len(mstats.PauseEnd) - 1) % len(mstats.PauseEnd) |
| for i := 0; i < n; i++ { |
| dt := stats.PauseEnd[i] |
| if dt.UnixNano() != int64(mstats.PauseEnd[off]) { |
| t.Errorf("stats.PauseEnd[%d] = %d, want %d", i, dt.UnixNano(), mstats.PauseEnd[off]) |
| } |
| off = (off + len(mstats.PauseEnd) - 1) % len(mstats.PauseEnd) |
| } |
| } |
| |
| var big []byte |
| |
| func TestFreeOSMemory(t *testing.T) { |
| if runtime.Compiler == "gccgo" { |
| t.Skip("conservative GC") |
| } |
| |
| // Tests FreeOSMemory by making big susceptible to collection |
| // and checking that at least that much memory is returned to |
| // the OS after. |
| |
| const bigBytes = 32 << 20 |
| big = make([]byte, bigBytes) |
| |
| // Make sure any in-progress GCs are complete. |
| runtime.GC() |
| |
| var before runtime.MemStats |
| runtime.ReadMemStats(&before) |
| |
| // Clear the last reference to the big allocation, making it |
| // susceptible to collection. |
| big = nil |
| |
| // FreeOSMemory runs a GC cycle before releasing memory, |
| // so it's fine to skip a GC here. |
| // |
| // It's possible the background scavenger runs concurrently |
| // with this function and does most of the work for it. |
| // If that happens, it's OK. What we want is a test that fails |
| // often if FreeOSMemory does not work correctly, and a test |
| // that passes every time if it does. |
| FreeOSMemory() |
| |
| var after runtime.MemStats |
| runtime.ReadMemStats(&after) |
| |
| // Check to make sure that the big allocation (now freed) |
| // had its memory shift into HeapReleased as a result of that |
| // FreeOSMemory. |
| if after.HeapReleased <= before.HeapReleased { |
| t.Fatalf("no memory released: %d -> %d", before.HeapReleased, after.HeapReleased) |
| } |
| |
| // Check to make sure bigBytes was released, plus some slack. Pages may get |
| // allocated in between the two measurements above for a variety for reasons, |
| // most commonly for GC work bufs. Since this can get fairly high, depending |
| // on scheduling and what GOMAXPROCS is, give a lot of slack up-front. |
| // |
| // Add a little more slack too if the page size is bigger than the runtime page size. |
| // "big" could end up unaligned on its ends, forcing the scavenger to skip at worst |
| // 2x pages. |
| slack := uint64(bigBytes / 2) |
| pageSize := uint64(os.Getpagesize()) |
| if pageSize > 8<<10 { |
| slack += pageSize * 2 |
| } |
| if slack > bigBytes { |
| // We basically already checked this. |
| return |
| } |
| if after.HeapReleased-before.HeapReleased < bigBytes-slack { |
| t.Fatalf("less than %d released: %d -> %d", bigBytes, before.HeapReleased, after.HeapReleased) |
| } |
| } |
| |
| var ( |
| setGCPercentBallast any |
| setGCPercentSink any |
| ) |
| |
| func TestSetGCPercent(t *testing.T) { |
| testenv.SkipFlaky(t, 20076) |
| |
| // Test that the variable is being set and returned correctly. |
| old := SetGCPercent(123) |
| new := SetGCPercent(old) |
| if new != 123 { |
| t.Errorf("SetGCPercent(123); SetGCPercent(x) = %d, want 123", new) |
| } |
| |
| // Test that the percentage is implemented correctly. |
| defer func() { |
| SetGCPercent(old) |
| setGCPercentBallast, setGCPercentSink = nil, nil |
| }() |
| SetGCPercent(100) |
| runtime.GC() |
| // Create 100 MB of live heap as a baseline. |
| const baseline = 100 << 20 |
| var ms runtime.MemStats |
| runtime.ReadMemStats(&ms) |
| setGCPercentBallast = make([]byte, baseline-ms.Alloc) |
| runtime.GC() |
| runtime.ReadMemStats(&ms) |
| if abs64(baseline-int64(ms.Alloc)) > 10<<20 { |
| t.Fatalf("failed to set up baseline live heap; got %d MB, want %d MB", ms.Alloc>>20, baseline>>20) |
| } |
| // NextGC should be ~200 MB. |
| const thresh = 20 << 20 // TODO: Figure out why this is so noisy on some builders |
| if want := int64(2 * baseline); abs64(want-int64(ms.NextGC)) > thresh { |
| t.Errorf("NextGC = %d MB, want %d±%d MB", ms.NextGC>>20, want>>20, thresh>>20) |
| } |
| // Create some garbage, but not enough to trigger another GC. |
| for i := 0; float64(i) < 1.2*baseline; i += 1 << 10 { |
| setGCPercentSink = make([]byte, 1<<10) |
| } |
| setGCPercentSink = nil |
| // Adjust GOGC to 50. NextGC should be ~150 MB. |
| SetGCPercent(50) |
| runtime.ReadMemStats(&ms) |
| if want := int64(1.5 * baseline); abs64(want-int64(ms.NextGC)) > thresh { |
| t.Errorf("NextGC = %d MB, want %d±%d MB", ms.NextGC>>20, want>>20, thresh>>20) |
| } |
| |
| // Trigger a GC and get back to 100 MB live with GOGC=100. |
| SetGCPercent(100) |
| runtime.GC() |
| // Raise live to 120 MB. |
| setGCPercentSink = make([]byte, int(0.2*baseline)) |
| // Lower GOGC to 10. This must force a GC. |
| runtime.ReadMemStats(&ms) |
| ngc1 := ms.NumGC |
| SetGCPercent(10) |
| // It may require an allocation to actually force the GC. |
| setGCPercentSink = make([]byte, 1<<20) |
| runtime.ReadMemStats(&ms) |
| ngc2 := ms.NumGC |
| if ngc1 == ngc2 { |
| t.Errorf("expected GC to run but it did not") |
| } |
| } |
| |
| func abs64(a int64) int64 { |
| if a < 0 { |
| return -a |
| } |
| return a |
| } |
| |
| func TestSetMaxThreadsOvf(t *testing.T) { |
| // Verify that a big threads count will not overflow the int32 |
| // maxmcount variable, causing a panic (see Issue 16076). |
| // |
| // This can only happen when ints are 64 bits, since on platforms |
| // with 32 bit ints SetMaxThreads (which takes an int parameter) |
| // cannot be given anything that will overflow an int32. |
| // |
| // Call SetMaxThreads with 1<<31, but only on 64 bit systems. |
| nt := SetMaxThreads(1 << (30 + ^uint(0)>>63)) |
| SetMaxThreads(nt) // restore previous value |
| } |