blob: 470c23dd419561af74cfcbb9e03db3c0924fdabd [file] [log] [blame]
// Copyright 2017 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
package runtime_test
import (
"fmt"
"regexp"
"runtime"
. "runtime"
"slices"
"sync"
"testing"
"time"
"unsafe"
)
func TestProfBuf(t *testing.T) {
const hdrSize = 2
write := func(t *testing.T, b *ProfBuf, tag unsafe.Pointer, now int64, hdr []uint64, stk []uintptr) {
b.Write(&tag, now, hdr, stk)
}
read := func(t *testing.T, b *ProfBuf, data []uint64, tags []unsafe.Pointer) {
rdata, rtags, eof := b.Read(ProfBufNonBlocking)
if !slices.Equal(rdata, data) || !slices.Equal(rtags, tags) {
t.Fatalf("unexpected profile read:\nhave data %#x\nwant data %#x\nhave tags %#x\nwant tags %#x", rdata, data, rtags, tags)
}
if eof {
t.Fatalf("unexpected eof")
}
}
readBlock := func(t *testing.T, b *ProfBuf, data []uint64, tags []unsafe.Pointer) func() {
c := make(chan int)
go func() {
eof := data == nil
rdata, rtags, reof := b.Read(ProfBufBlocking)
if !slices.Equal(rdata, data) || !slices.Equal(rtags, tags) || reof != eof {
// Errorf, not Fatalf, because called in goroutine.
t.Errorf("unexpected profile read:\nhave data %#x\nwant data %#x\nhave tags %#x\nwant tags %#x\nhave eof=%v, want %v", rdata, data, rtags, tags, reof, eof)
}
c <- 1
}()
time.Sleep(10 * time.Millisecond) // let goroutine run and block
return func() { <-c }
}
readEOF := func(t *testing.T, b *ProfBuf) {
rdata, rtags, eof := b.Read(ProfBufBlocking)
if rdata != nil || rtags != nil || !eof {
t.Errorf("unexpected profile read: %#x, %#x, eof=%v; want nil, nil, eof=true", rdata, rtags, eof)
}
rdata, rtags, eof = b.Read(ProfBufNonBlocking)
if rdata != nil || rtags != nil || !eof {
t.Errorf("unexpected profile read (non-blocking): %#x, %#x, eof=%v; want nil, nil, eof=true", rdata, rtags, eof)
}
}
myTags := make([]byte, 100)
t.Logf("myTags is %p", &myTags[0])
t.Run("BasicWriteRead", func(t *testing.T) {
b := NewProfBuf(2, 11, 1)
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9}, []unsafe.Pointer{unsafe.Pointer(&myTags[0])})
read(t, b, nil, nil) // release data returned by previous read
write(t, b, unsafe.Pointer(&myTags[2]), 99, []uint64{101, 102}, []uintptr{201, 202, 203, 204})
read(t, b, []uint64{8, 99, 101, 102, 201, 202, 203, 204}, []unsafe.Pointer{unsafe.Pointer(&myTags[2])})
})
t.Run("ReadMany", func(t *testing.T) {
b := NewProfBuf(2, 50, 50)
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
write(t, b, unsafe.Pointer(&myTags[2]), 99, []uint64{101, 102}, []uintptr{201, 202, 203, 204})
write(t, b, unsafe.Pointer(&myTags[1]), 500, []uint64{502, 504}, []uintptr{506})
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9, 8, 99, 101, 102, 201, 202, 203, 204, 5, 500, 502, 504, 506}, []unsafe.Pointer{unsafe.Pointer(&myTags[0]), unsafe.Pointer(&myTags[2]), unsafe.Pointer(&myTags[1])})
})
t.Run("ReadManyShortData", func(t *testing.T) {
b := NewProfBuf(2, 50, 50)
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
write(t, b, unsafe.Pointer(&myTags[2]), 99, []uint64{101, 102}, []uintptr{201, 202, 203, 204})
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9, 8, 99, 101, 102, 201, 202, 203, 204}, []unsafe.Pointer{unsafe.Pointer(&myTags[0]), unsafe.Pointer(&myTags[2])})
})
t.Run("ReadManyShortTags", func(t *testing.T) {
b := NewProfBuf(2, 50, 50)
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
write(t, b, unsafe.Pointer(&myTags[2]), 99, []uint64{101, 102}, []uintptr{201, 202, 203, 204})
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9, 8, 99, 101, 102, 201, 202, 203, 204}, []unsafe.Pointer{unsafe.Pointer(&myTags[0]), unsafe.Pointer(&myTags[2])})
})
t.Run("ReadAfterOverflow1", func(t *testing.T) {
// overflow record synthesized by write
b := NewProfBuf(2, 16, 5)
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9}) // uses 10
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9}, []unsafe.Pointer{unsafe.Pointer(&myTags[0])}) // reads 10 but still in use until next read
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5}) // uses 6
read(t, b, []uint64{6, 1, 2, 3, 4, 5}, []unsafe.Pointer{unsafe.Pointer(&myTags[0])}) // reads 6 but still in use until next read
// now 10 available
write(t, b, unsafe.Pointer(&myTags[2]), 99, []uint64{101, 102}, []uintptr{201, 202, 203, 204, 205, 206, 207, 208, 209}) // no room
for i := 0; i < 299; i++ {
write(t, b, unsafe.Pointer(&myTags[3]), int64(100+i), []uint64{101, 102}, []uintptr{201, 202, 203, 204}) // no room for overflow+this record
}
write(t, b, unsafe.Pointer(&myTags[1]), 500, []uint64{502, 504}, []uintptr{506}) // room for overflow+this record
read(t, b, []uint64{5, 99, 0, 0, 300, 5, 500, 502, 504, 506}, []unsafe.Pointer{nil, unsafe.Pointer(&myTags[1])})
})
t.Run("ReadAfterOverflow2", func(t *testing.T) {
// overflow record synthesized by read
b := NewProfBuf(2, 16, 5)
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
write(t, b, unsafe.Pointer(&myTags[2]), 99, []uint64{101, 102}, []uintptr{201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213})
for i := 0; i < 299; i++ {
write(t, b, unsafe.Pointer(&myTags[3]), 100, []uint64{101, 102}, []uintptr{201, 202, 203, 204})
}
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9}, []unsafe.Pointer{unsafe.Pointer(&myTags[0])}) // reads 10 but still in use until next read
write(t, b, unsafe.Pointer(&myTags[1]), 500, []uint64{502, 504}, []uintptr{}) // still overflow
read(t, b, []uint64{5, 99, 0, 0, 301}, []unsafe.Pointer{nil}) // overflow synthesized by read
write(t, b, unsafe.Pointer(&myTags[1]), 500, []uint64{502, 505}, []uintptr{506}) // written
read(t, b, []uint64{5, 500, 502, 505, 506}, []unsafe.Pointer{unsafe.Pointer(&myTags[1])})
})
t.Run("ReadAtEndAfterOverflow", func(t *testing.T) {
b := NewProfBuf(2, 12, 5)
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
write(t, b, unsafe.Pointer(&myTags[2]), 99, []uint64{101, 102}, []uintptr{201, 202, 203, 204})
for i := 0; i < 299; i++ {
write(t, b, unsafe.Pointer(&myTags[3]), 100, []uint64{101, 102}, []uintptr{201, 202, 203, 204})
}
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9}, []unsafe.Pointer{unsafe.Pointer(&myTags[0])})
read(t, b, []uint64{5, 99, 0, 0, 300}, []unsafe.Pointer{nil})
write(t, b, unsafe.Pointer(&myTags[1]), 500, []uint64{502, 504}, []uintptr{506})
read(t, b, []uint64{5, 500, 502, 504, 506}, []unsafe.Pointer{unsafe.Pointer(&myTags[1])})
})
t.Run("BlockingWriteRead", func(t *testing.T) {
b := NewProfBuf(2, 11, 1)
wait := readBlock(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9}, []unsafe.Pointer{unsafe.Pointer(&myTags[0])})
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
wait()
wait = readBlock(t, b, []uint64{8, 99, 101, 102, 201, 202, 203, 204}, []unsafe.Pointer{unsafe.Pointer(&myTags[2])})
time.Sleep(10 * time.Millisecond)
write(t, b, unsafe.Pointer(&myTags[2]), 99, []uint64{101, 102}, []uintptr{201, 202, 203, 204})
wait()
wait = readBlock(t, b, nil, nil)
b.Close()
wait()
wait = readBlock(t, b, nil, nil)
wait()
readEOF(t, b)
})
t.Run("DataWraparound", func(t *testing.T) {
b := NewProfBuf(2, 16, 1024)
for i := 0; i < 10; i++ {
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9}, []unsafe.Pointer{unsafe.Pointer(&myTags[0])})
read(t, b, nil, nil) // release data returned by previous read
}
})
t.Run("TagWraparound", func(t *testing.T) {
b := NewProfBuf(2, 1024, 2)
for i := 0; i < 10; i++ {
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9}, []unsafe.Pointer{unsafe.Pointer(&myTags[0])})
read(t, b, nil, nil) // release data returned by previous read
}
})
t.Run("BothWraparound", func(t *testing.T) {
b := NewProfBuf(2, 16, 2)
for i := 0; i < 10; i++ {
write(t, b, unsafe.Pointer(&myTags[0]), 1, []uint64{2, 3}, []uintptr{4, 5, 6, 7, 8, 9})
read(t, b, []uint64{10, 1, 2, 3, 4, 5, 6, 7, 8, 9}, []unsafe.Pointer{unsafe.Pointer(&myTags[0])})
read(t, b, nil, nil) // release data returned by previous read
}
})
}
func TestProfBufDoubleWakeup(t *testing.T) {
b := NewProfBuf(2, 16, 2)
go func() {
for range 1000 {
b.Write(nil, 1, []uint64{5, 6}, []uintptr{7, 8})
}
b.Close()
}()
for {
_, _, eof := b.Read(ProfBufBlocking)
if eof {
return
}
}
}
func TestProfBufWakeup(t *testing.T) {
b := NewProfBuf(2, 16, 2)
var wg sync.WaitGroup
wg.Go(func() {
read := 0
for {
rdata, _, eof := b.Read(ProfBufBlocking)
if read == 0 && len(rdata) < 8 {
t.Errorf("first wake up at less than half full, got %x", rdata)
}
read += len(rdata)
if eof {
return
}
}
})
// Under the hood profBuf uses notetsleepg when the reader blocks.
// Different platforms have different implementations, leading to
// different statuses we need to look for to determine whether the
// reader is blocked.
var waitStatus string
switch runtime.GOOS {
case "js":
waitStatus = "waiting"
case "wasip1":
waitStatus = "runnable"
default:
waitStatus = "syscall"
}
// Ensure that the reader is blocked
awaitBlockedGoroutine(waitStatus, "TestProfBufWakeup.func1")
// NB: this writes 6 words not 4. Fine for the test.
// The reader shouldn't wake up for this
b.Write(nil, 1, []uint64{1, 2}, []uintptr{3, 4})
// The reader should still be blocked. The awaitBlockedGoroutine here
// checks that and also gives a buggy implementation a chance to
// actually wake up (it calls Gosched) before the next write. There is a
// small chance that a buggy implementation would have woken up but
// doesn't get scheduled by the time we do the next write. In that case
// the reader will see a more-than-half-full buffer and the test will
// pass. But if the implementation is broken, this test should fail
// regularly, even if not 100% of the time.
awaitBlockedGoroutine(waitStatus, "TestProfBufWakeup.func1")
b.Write(nil, 1, []uint64{5, 6}, []uintptr{7, 8})
b.Close()
// Wait here so we can be sure the reader got the data
wg.Wait()
}
// see also runtime/pprof tests
func awaitBlockedGoroutine(state, fName string) {
// NB: this matches [state] as well as [state, n minutes]
re := fmt.Sprintf(`(?m)^goroutine \d+ \[%s.*\]:\n(?:.+\n\t.+\n)*runtime_test\.%s`, regexp.QuoteMeta(state), fName)
r := regexp.MustCompile(re)
buf := make([]byte, 64<<10)
for {
Gosched()
n := Stack(buf, true)
if n == len(buf) {
// Buffer wasn't large enough for a full goroutine dump.
// Resize it and try again.
buf = make([]byte, 2*len(buf))
continue
}
const count = 1
if len(r.FindAll(buf[:n], -1)) >= count {
return
}
}
}