blob: 89866063dc15197b4e11aa298d10d6fe8c69fc5a [file] [log] [blame]
// Copyright 2013 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
package debug_test
import (
"internal/testenv"
"os"
"runtime"
. "runtime/debug"
"testing"
"time"
)
func TestReadGCStats(t *testing.T) {
defer SetGCPercent(SetGCPercent(-1))
var stats GCStats
var mstats runtime.MemStats
var min, max time.Duration
// First ReadGCStats will allocate, second should not,
// especially if we follow up with an explicit garbage collection.
stats.PauseQuantiles = make([]time.Duration, 10)
ReadGCStats(&stats)
runtime.GC()
// Assume these will return same data: no GC during ReadGCStats.
ReadGCStats(&stats)
runtime.ReadMemStats(&mstats)
if stats.NumGC != int64(mstats.NumGC) {
t.Errorf("stats.NumGC = %d, but mstats.NumGC = %d", stats.NumGC, mstats.NumGC)
}
if stats.PauseTotal != time.Duration(mstats.PauseTotalNs) {
t.Errorf("stats.PauseTotal = %d, but mstats.PauseTotalNs = %d", stats.PauseTotal, mstats.PauseTotalNs)
}
if stats.LastGC.UnixNano() != int64(mstats.LastGC) {
t.Errorf("stats.LastGC.UnixNano = %d, but mstats.LastGC = %d", stats.LastGC.UnixNano(), mstats.LastGC)
}
n := int(mstats.NumGC)
if n > len(mstats.PauseNs) {
n = len(mstats.PauseNs)
}
if len(stats.Pause) != n {
t.Errorf("len(stats.Pause) = %d, want %d", len(stats.Pause), n)
} else {
off := (int(mstats.NumGC) + len(mstats.PauseNs) - 1) % len(mstats.PauseNs)
for i := 0; i < n; i++ {
dt := stats.Pause[i]
if dt != time.Duration(mstats.PauseNs[off]) {
t.Errorf("stats.Pause[%d] = %d, want %d", i, dt, mstats.PauseNs[off])
}
if max < dt {
max = dt
}
if min > dt || i == 0 {
min = dt
}
off = (off + len(mstats.PauseNs) - 1) % len(mstats.PauseNs)
}
}
q := stats.PauseQuantiles
nq := len(q)
if q[0] != min || q[nq-1] != max {
t.Errorf("stats.PauseQuantiles = [%d, ..., %d], want [%d, ..., %d]", q[0], q[nq-1], min, max)
}
for i := 0; i < nq-1; i++ {
if q[i] > q[i+1] {
t.Errorf("stats.PauseQuantiles[%d]=%d > stats.PauseQuantiles[%d]=%d", i, q[i], i+1, q[i+1])
}
}
// compare memory stats with gc stats:
if len(stats.PauseEnd) != n {
t.Fatalf("len(stats.PauseEnd) = %d, want %d", len(stats.PauseEnd), n)
}
off := (int(mstats.NumGC) + len(mstats.PauseEnd) - 1) % len(mstats.PauseEnd)
for i := 0; i < n; i++ {
dt := stats.PauseEnd[i]
if dt.UnixNano() != int64(mstats.PauseEnd[off]) {
t.Errorf("stats.PauseEnd[%d] = %d, want %d", i, dt.UnixNano(), mstats.PauseEnd[off])
}
off = (off + len(mstats.PauseEnd) - 1) % len(mstats.PauseEnd)
}
}
var big []byte
func TestFreeOSMemory(t *testing.T) {
if runtime.Compiler == "gccgo" {
t.Skip("conservative GC")
}
// Tests FreeOSMemory by making big susceptible to collection
// and checking that at least that much memory is returned to
// the OS after.
const bigBytes = 32 << 20
big = make([]byte, bigBytes)
// Make sure any in-progress GCs are complete.
runtime.GC()
var before runtime.MemStats
runtime.ReadMemStats(&before)
// Clear the last reference to the big allocation, making it
// susceptible to collection.
big = nil
// FreeOSMemory runs a GC cycle before releasing memory,
// so it's fine to skip a GC here.
//
// It's possible the background scavenger runs concurrently
// with this function and does most of the work for it.
// If that happens, it's OK. What we want is a test that fails
// often if FreeOSMemory does not work correctly, and a test
// that passes every time if it does.
FreeOSMemory()
var after runtime.MemStats
runtime.ReadMemStats(&after)
// Check to make sure that the big allocation (now freed)
// had its memory shift into HeapReleased as a result of that
// FreeOSMemory.
if after.HeapReleased <= before.HeapReleased {
t.Fatalf("no memory released: %d -> %d", before.HeapReleased, after.HeapReleased)
}
// Check to make sure bigBytes was released, plus some slack. Pages may get
// allocated in between the two measurements above for a variety for reasons,
// most commonly for GC work bufs. Since this can get fairly high, depending
// on scheduling and what GOMAXPROCS is, give a lot of slack up-front.
//
// Add a little more slack too if the page size is bigger than the runtime page size.
// "big" could end up unaligned on its ends, forcing the scavenger to skip at worst
// 2x pages.
slack := uint64(bigBytes / 2)
pageSize := uint64(os.Getpagesize())
if pageSize > 8<<10 {
slack += pageSize * 2
}
if slack > bigBytes {
// We basically already checked this.
return
}
if after.HeapReleased-before.HeapReleased < bigBytes-slack {
t.Fatalf("less than %d released: %d -> %d", bigBytes, before.HeapReleased, after.HeapReleased)
}
}
var (
setGCPercentBallast any
setGCPercentSink any
)
func TestSetGCPercent(t *testing.T) {
testenv.SkipFlaky(t, 20076)
// Test that the variable is being set and returned correctly.
old := SetGCPercent(123)
new := SetGCPercent(old)
if new != 123 {
t.Errorf("SetGCPercent(123); SetGCPercent(x) = %d, want 123", new)
}
// Test that the percentage is implemented correctly.
defer func() {
SetGCPercent(old)
setGCPercentBallast, setGCPercentSink = nil, nil
}()
SetGCPercent(100)
runtime.GC()
// Create 100 MB of live heap as a baseline.
const baseline = 100 << 20
var ms runtime.MemStats
runtime.ReadMemStats(&ms)
setGCPercentBallast = make([]byte, baseline-ms.Alloc)
runtime.GC()
runtime.ReadMemStats(&ms)
if abs64(baseline-int64(ms.Alloc)) > 10<<20 {
t.Fatalf("failed to set up baseline live heap; got %d MB, want %d MB", ms.Alloc>>20, baseline>>20)
}
// NextGC should be ~200 MB.
const thresh = 20 << 20 // TODO: Figure out why this is so noisy on some builders
if want := int64(2 * baseline); abs64(want-int64(ms.NextGC)) > thresh {
t.Errorf("NextGC = %d MB, want %d±%d MB", ms.NextGC>>20, want>>20, thresh>>20)
}
// Create some garbage, but not enough to trigger another GC.
for i := 0; float64(i) < 1.2*baseline; i += 1 << 10 {
setGCPercentSink = make([]byte, 1<<10)
}
setGCPercentSink = nil
// Adjust GOGC to 50. NextGC should be ~150 MB.
SetGCPercent(50)
runtime.ReadMemStats(&ms)
if want := int64(1.5 * baseline); abs64(want-int64(ms.NextGC)) > thresh {
t.Errorf("NextGC = %d MB, want %d±%d MB", ms.NextGC>>20, want>>20, thresh>>20)
}
// Trigger a GC and get back to 100 MB live with GOGC=100.
SetGCPercent(100)
runtime.GC()
// Raise live to 120 MB.
setGCPercentSink = make([]byte, int(0.2*baseline))
// Lower GOGC to 10. This must force a GC.
runtime.ReadMemStats(&ms)
ngc1 := ms.NumGC
SetGCPercent(10)
// It may require an allocation to actually force the GC.
setGCPercentSink = make([]byte, 1<<20)
runtime.ReadMemStats(&ms)
ngc2 := ms.NumGC
if ngc1 == ngc2 {
t.Errorf("expected GC to run but it did not")
}
}
func abs64(a int64) int64 {
if a < 0 {
return -a
}
return a
}
func TestSetMaxThreadsOvf(t *testing.T) {
// Verify that a big threads count will not overflow the int32
// maxmcount variable, causing a panic (see Issue 16076).
//
// This can only happen when ints are 64 bits, since on platforms
// with 32 bit ints SetMaxThreads (which takes an int parameter)
// cannot be given anything that will overflow an int32.
//
// Call SetMaxThreads with 1<<31, but only on 64 bit systems.
nt := SetMaxThreads(1 << (30 + ^uint(0)>>63))
SetMaxThreads(nt) // restore previous value
}