blob: 376b4a58f239bb9fc55570a7cfc8ca5f594ab1b5 [file] [log] [blame]
// Copyright 2011 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
package runtime_test
import (
"fmt"
"internal/asan"
"internal/msan"
"internal/race"
"internal/testenv"
"runtime"
"testing"
)
const N = 20
func BenchmarkMakeSliceCopy(b *testing.B) {
const length = 32
var bytes = make([]byte, 8*length)
var ints = make([]int, length)
var ptrs = make([]*byte, length)
b.Run("mallocmove", func(b *testing.B) {
b.Run("Byte", func(b *testing.B) {
var x []byte
for i := 0; i < b.N; i++ {
x = make([]byte, len(bytes))
copy(x, bytes)
}
})
b.Run("Int", func(b *testing.B) {
var x []int
for i := 0; i < b.N; i++ {
x = make([]int, len(ints))
copy(x, ints)
}
})
b.Run("Ptr", func(b *testing.B) {
var x []*byte
for i := 0; i < b.N; i++ {
x = make([]*byte, len(ptrs))
copy(x, ptrs)
}
})
})
b.Run("makecopy", func(b *testing.B) {
b.Run("Byte", func(b *testing.B) {
var x []byte
for i := 0; i < b.N; i++ {
x = make([]byte, 8*length)
copy(x, bytes)
}
})
b.Run("Int", func(b *testing.B) {
var x []int
for i := 0; i < b.N; i++ {
x = make([]int, length)
copy(x, ints)
}
})
b.Run("Ptr", func(b *testing.B) {
var x []*byte
for i := 0; i < b.N; i++ {
x = make([]*byte, length)
copy(x, ptrs)
}
})
})
b.Run("nilappend", func(b *testing.B) {
b.Run("Byte", func(b *testing.B) {
var x []byte
for i := 0; i < b.N; i++ {
x = append([]byte(nil), bytes...)
_ = x
}
})
b.Run("Int", func(b *testing.B) {
var x []int
for i := 0; i < b.N; i++ {
x = append([]int(nil), ints...)
_ = x
}
})
b.Run("Ptr", func(b *testing.B) {
var x []*byte
for i := 0; i < b.N; i++ {
x = append([]*byte(nil), ptrs...)
_ = x
}
})
})
}
type (
struct24 struct{ a, b, c int64 }
struct32 struct{ a, b, c, d int64 }
struct40 struct{ a, b, c, d, e int64 }
)
func BenchmarkMakeSlice(b *testing.B) {
const length = 2
b.Run("Byte", func(b *testing.B) {
var x []byte
for i := 0; i < b.N; i++ {
x = make([]byte, length, 2*length)
_ = x
}
})
b.Run("Int16", func(b *testing.B) {
var x []int16
for i := 0; i < b.N; i++ {
x = make([]int16, length, 2*length)
_ = x
}
})
b.Run("Int", func(b *testing.B) {
var x []int
for i := 0; i < b.N; i++ {
x = make([]int, length, 2*length)
_ = x
}
})
b.Run("Ptr", func(b *testing.B) {
var x []*byte
for i := 0; i < b.N; i++ {
x = make([]*byte, length, 2*length)
_ = x
}
})
b.Run("Struct", func(b *testing.B) {
b.Run("24", func(b *testing.B) {
var x []struct24
for i := 0; i < b.N; i++ {
x = make([]struct24, length, 2*length)
_ = x
}
})
b.Run("32", func(b *testing.B) {
var x []struct32
for i := 0; i < b.N; i++ {
x = make([]struct32, length, 2*length)
_ = x
}
})
b.Run("40", func(b *testing.B) {
var x []struct40
for i := 0; i < b.N; i++ {
x = make([]struct40, length, 2*length)
_ = x
}
})
})
}
func BenchmarkGrowSlice(b *testing.B) {
b.Run("Byte", func(b *testing.B) {
x := make([]byte, 9)
for i := 0; i < b.N; i++ {
_ = append([]byte(nil), x...)
}
})
b.Run("Int16", func(b *testing.B) {
x := make([]int16, 9)
for i := 0; i < b.N; i++ {
_ = append([]int16(nil), x...)
}
})
b.Run("Int", func(b *testing.B) {
x := make([]int, 9)
for i := 0; i < b.N; i++ {
_ = append([]int(nil), x...)
}
})
b.Run("Ptr", func(b *testing.B) {
x := make([]*byte, 9)
for i := 0; i < b.N; i++ {
_ = append([]*byte(nil), x...)
}
})
b.Run("Struct", func(b *testing.B) {
b.Run("24", func(b *testing.B) {
x := make([]struct24, 9)
for i := 0; i < b.N; i++ {
_ = append([]struct24(nil), x...)
}
})
b.Run("32", func(b *testing.B) {
x := make([]struct32, 9)
for i := 0; i < b.N; i++ {
_ = append([]struct32(nil), x...)
}
})
b.Run("40", func(b *testing.B) {
x := make([]struct40, 9)
for i := 0; i < b.N; i++ {
_ = append([]struct40(nil), x...)
}
})
})
}
var (
SinkIntSlice []int
SinkIntPointerSlice []*int
)
func BenchmarkExtendSlice(b *testing.B) {
var length = 4 // Use a variable to prevent stack allocation of slices.
b.Run("IntSlice", func(b *testing.B) {
s := make([]int, 0, length)
for i := 0; i < b.N; i++ {
s = append(s[:0:length/2], make([]int, length)...)
}
SinkIntSlice = s
})
b.Run("PointerSlice", func(b *testing.B) {
s := make([]*int, 0, length)
for i := 0; i < b.N; i++ {
s = append(s[:0:length/2], make([]*int, length)...)
}
SinkIntPointerSlice = s
})
b.Run("NoGrow", func(b *testing.B) {
s := make([]int, 0, length)
for i := 0; i < b.N; i++ {
s = append(s[:0:length], make([]int, length)...)
}
SinkIntSlice = s
})
}
func BenchmarkAppend(b *testing.B) {
b.StopTimer()
x := make([]int, 0, N)
b.StartTimer()
for i := 0; i < b.N; i++ {
x = x[0:0]
for j := 0; j < N; j++ {
x = append(x, j)
}
}
}
func BenchmarkAppendGrowByte(b *testing.B) {
for i := 0; i < b.N; i++ {
var x []byte
for j := 0; j < 1<<20; j++ {
x = append(x, byte(j))
}
}
}
func BenchmarkAppendGrowString(b *testing.B) {
var s string
for i := 0; i < b.N; i++ {
var x []string
for j := 0; j < 1<<20; j++ {
x = append(x, s)
}
}
}
func BenchmarkAppendSlice(b *testing.B) {
for _, length := range []int{1, 4, 7, 8, 15, 16, 32} {
b.Run(fmt.Sprint(length, "Bytes"), func(b *testing.B) {
x := make([]byte, 0, N)
y := make([]byte, length)
for i := 0; i < b.N; i++ {
x = x[0:0]
x = append(x, y...)
}
})
}
}
var (
blackhole []byte
)
func BenchmarkAppendSliceLarge(b *testing.B) {
for _, length := range []int{1 << 10, 4 << 10, 16 << 10, 64 << 10, 256 << 10, 1024 << 10} {
y := make([]byte, length)
b.Run(fmt.Sprint(length, "Bytes"), func(b *testing.B) {
for i := 0; i < b.N; i++ {
blackhole = nil
blackhole = append(blackhole, y...)
}
})
}
}
func BenchmarkAppendStr(b *testing.B) {
for _, str := range []string{
"1",
"1234",
"12345678",
"1234567890123456",
"12345678901234567890123456789012",
} {
b.Run(fmt.Sprint(len(str), "Bytes"), func(b *testing.B) {
x := make([]byte, 0, N)
for i := 0; i < b.N; i++ {
x = x[0:0]
x = append(x, str...)
}
})
}
}
func BenchmarkAppendSpecialCase(b *testing.B) {
b.StopTimer()
x := make([]int, 0, N)
b.StartTimer()
for i := 0; i < b.N; i++ {
x = x[0:0]
for j := 0; j < N; j++ {
if len(x) < cap(x) {
x = x[:len(x)+1]
x[len(x)-1] = j
} else {
x = append(x, j)
}
}
}
}
var x []int
func f() int {
x[:1][0] = 3
return 2
}
func TestSideEffectOrder(t *testing.T) {
x = make([]int, 0, 10)
x = append(x, 1, f())
if x[0] != 1 || x[1] != 2 {
t.Error("append failed: ", x[0], x[1])
}
}
func TestAppendOverlap(t *testing.T) {
x := []byte("1234")
x = append(x[1:], x...) // p > q in runtime·appendslice.
got := string(x)
want := "2341234"
if got != want {
t.Errorf("overlap failed: got %q want %q", got, want)
}
}
func BenchmarkCopy(b *testing.B) {
for _, l := range []int{1, 2, 4, 8, 12, 16, 32, 128, 1024} {
buf := make([]byte, 4096)
b.Run(fmt.Sprint(l, "Byte"), func(b *testing.B) {
s := make([]byte, l)
var n int
for i := 0; i < b.N; i++ {
n = copy(buf, s)
}
b.SetBytes(int64(n))
})
b.Run(fmt.Sprint(l, "String"), func(b *testing.B) {
s := string(make([]byte, l))
var n int
for i := 0; i < b.N; i++ {
n = copy(buf, s)
}
b.SetBytes(int64(n))
})
}
}
var (
sByte []byte
s1Ptr []uintptr
s2Ptr [][2]uintptr
s3Ptr [][3]uintptr
s4Ptr [][4]uintptr
)
// BenchmarkAppendInPlace tests the performance of append
// when the result is being written back to the same slice.
// In order for the in-place optimization to occur,
// the slice must be referred to by address;
// using a global is an easy way to trigger that.
// We test the "grow" and "no grow" paths separately,
// but not the "normal" (occasionally grow) path,
// because it is a blend of the other two.
// We use small numbers and small sizes in an attempt
// to avoid benchmarking memory allocation and copying.
// We use scalars instead of pointers in an attempt
// to avoid benchmarking the write barriers.
// We benchmark four common sizes (byte, pointer, string/interface, slice),
// and one larger size.
func BenchmarkAppendInPlace(b *testing.B) {
b.Run("NoGrow", func(b *testing.B) {
const C = 128
b.Run("Byte", func(b *testing.B) {
for i := 0; i < b.N; i++ {
sByte = make([]byte, C)
for j := 0; j < C; j++ {
sByte = append(sByte, 0x77)
}
}
})
b.Run("1Ptr", func(b *testing.B) {
for i := 0; i < b.N; i++ {
s1Ptr = make([]uintptr, C)
for j := 0; j < C; j++ {
s1Ptr = append(s1Ptr, 0x77)
}
}
})
b.Run("2Ptr", func(b *testing.B) {
for i := 0; i < b.N; i++ {
s2Ptr = make([][2]uintptr, C)
for j := 0; j < C; j++ {
s2Ptr = append(s2Ptr, [2]uintptr{0x77, 0x88})
}
}
})
b.Run("3Ptr", func(b *testing.B) {
for i := 0; i < b.N; i++ {
s3Ptr = make([][3]uintptr, C)
for j := 0; j < C; j++ {
s3Ptr = append(s3Ptr, [3]uintptr{0x77, 0x88, 0x99})
}
}
})
b.Run("4Ptr", func(b *testing.B) {
for i := 0; i < b.N; i++ {
s4Ptr = make([][4]uintptr, C)
for j := 0; j < C; j++ {
s4Ptr = append(s4Ptr, [4]uintptr{0x77, 0x88, 0x99, 0xAA})
}
}
})
})
b.Run("Grow", func(b *testing.B) {
const C = 5
b.Run("Byte", func(b *testing.B) {
for i := 0; i < b.N; i++ {
sByte = make([]byte, 0)
for j := 0; j < C; j++ {
sByte = append(sByte, 0x77)
sByte = sByte[:cap(sByte)]
}
}
})
b.Run("1Ptr", func(b *testing.B) {
for i := 0; i < b.N; i++ {
s1Ptr = make([]uintptr, 0)
for j := 0; j < C; j++ {
s1Ptr = append(s1Ptr, 0x77)
s1Ptr = s1Ptr[:cap(s1Ptr)]
}
}
})
b.Run("2Ptr", func(b *testing.B) {
for i := 0; i < b.N; i++ {
s2Ptr = make([][2]uintptr, 0)
for j := 0; j < C; j++ {
s2Ptr = append(s2Ptr, [2]uintptr{0x77, 0x88})
s2Ptr = s2Ptr[:cap(s2Ptr)]
}
}
})
b.Run("3Ptr", func(b *testing.B) {
for i := 0; i < b.N; i++ {
s3Ptr = make([][3]uintptr, 0)
for j := 0; j < C; j++ {
s3Ptr = append(s3Ptr, [3]uintptr{0x77, 0x88, 0x99})
s3Ptr = s3Ptr[:cap(s3Ptr)]
}
}
})
b.Run("4Ptr", func(b *testing.B) {
for i := 0; i < b.N; i++ {
s4Ptr = make([][4]uintptr, 0)
for j := 0; j < C; j++ {
s4Ptr = append(s4Ptr, [4]uintptr{0x77, 0x88, 0x99, 0xAA})
s4Ptr = s4Ptr[:cap(s4Ptr)]
}
}
})
})
}
//go:noinline
func byteSlice(n int) []byte {
var r []byte
for i := range n {
r = append(r, byte(i))
}
return r
}
func TestAppendByteInLoop(t *testing.T) {
testenv.SkipIfOptimizationOff(t)
if race.Enabled {
t.Skip("skipping in -race mode")
}
if asan.Enabled || msan.Enabled {
t.Skip("skipping in sanitizer mode")
}
for _, test := range [][3]int{
{0, 0, 0},
{1, 1, 8},
{2, 1, 8},
{8, 1, 8},
{9, 1, 16},
{16, 1, 16},
{17, 1, 24},
{24, 1, 24},
{25, 1, 32},
{32, 1, 32},
{33, 1, 64}, // If we up the stack buffer size from 32->64, this line and the next would become 48.
{48, 1, 64},
{49, 1, 64},
{64, 1, 64},
{65, 2, 128},
} {
n := test[0]
want := test[1]
wantCap := test[2]
var r []byte
got := testing.AllocsPerRun(10, func() {
r = byteSlice(n)
})
if got != float64(want) {
t.Errorf("for size %d, got %f allocs want %d", n, got, want)
}
if cap(r) != wantCap {
t.Errorf("for size %d, got capacity %d want %d", n, cap(r), wantCap)
}
}
}
//go:noinline
func ptrSlice(n int, p *[]*byte) {
var r []*byte
for range n {
r = append(r, nil)
}
*p = r
}
func TestAppendPtrInLoop(t *testing.T) {
testenv.SkipIfOptimizationOff(t)
if race.Enabled {
t.Skip("skipping in -race mode")
}
if asan.Enabled || msan.Enabled {
t.Skip("skipping in sanitizer mode")
}
var tests [][3]int
if runtime.PtrSize == 8 {
tests = [][3]int{
{0, 0, 0},
{1, 1, 1},
{2, 1, 2},
{3, 1, 3}, // This is the interesting case, allocates 24 bytes when before it was 32.
{4, 1, 4},
{5, 1, 8},
{6, 1, 8},
{7, 1, 8},
{8, 1, 8},
{9, 2, 16},
}
} else {
tests = [][3]int{
{0, 0, 0},
{1, 1, 2},
{2, 1, 2},
{3, 1, 4},
{4, 1, 4},
{5, 1, 6}, // These two are also 24 bytes instead of 32.
{6, 1, 6}, //
{7, 1, 8},
{8, 1, 8},
{9, 1, 16},
{10, 1, 16},
{11, 1, 16},
{12, 1, 16},
{13, 1, 16},
{14, 1, 16},
{15, 1, 16},
{16, 1, 16},
{17, 2, 32},
}
}
for _, test := range tests {
n := test[0]
want := test[1]
wantCap := test[2]
var r []*byte
got := testing.AllocsPerRun(10, func() {
ptrSlice(n, &r)
})
if got != float64(want) {
t.Errorf("for size %d, got %f allocs want %d", n, got, want)
}
if cap(r) != wantCap {
t.Errorf("for size %d, got capacity %d want %d", n, cap(r), wantCap)
}
}
}
//go:noinline
func byteCapSlice(n int) ([]byte, int) {
var r []byte
for i := range n {
r = append(r, byte(i))
}
return r, cap(r)
}
func TestAppendByteCapInLoop(t *testing.T) {
testenv.SkipIfOptimizationOff(t)
if race.Enabled {
t.Skip("skipping in -race mode")
}
if asan.Enabled || msan.Enabled {
t.Skip("skipping in sanitizer mode")
}
for _, test := range [][3]int{
{0, 0, 0},
{1, 1, 8},
{2, 1, 8},
{8, 1, 8},
{9, 1, 16},
{16, 1, 16},
{17, 1, 24},
{24, 1, 24},
{25, 1, 32},
{32, 1, 32},
{33, 1, 64},
{48, 1, 64},
{49, 1, 64},
{64, 1, 64},
{65, 2, 128},
} {
n := test[0]
want := test[1]
wantCap := test[2]
var r []byte
got := testing.AllocsPerRun(10, func() {
r, _ = byteCapSlice(n)
})
if got != float64(want) {
t.Errorf("for size %d, got %f allocs want %d", n, got, want)
}
if cap(r) != wantCap {
t.Errorf("for size %d, got capacity %d want %d", n, cap(r), wantCap)
}
}
}
func TestAppendGeneric(t *testing.T) {
type I *int
r := testAppendGeneric[I](100)
if len(r) != 100 {
t.Errorf("bad length")
}
}
//go:noinline
func testAppendGeneric[E any](n int) []E {
var r []E
var z E
for range n {
r = append(r, z)
}
return r
}
func appendSomeBytes(r []byte, s []byte) []byte {
for _, b := range s {
r = append(r, b)
}
return r
}
func TestAppendOfArg(t *testing.T) {
r := make([]byte, 24)
for i := 0; i < 24; i++ {
r[i] = byte(i)
}
appendSomeBytes(r, []byte{25, 26, 27})
// Do the same thing, trying to overwrite any
// stack-allocated buffers used above.
s := make([]byte, 24)
for i := 0; i < 24; i++ {
s[i] = 99
}
appendSomeBytes(s, []byte{99, 99, 99})
// Check that we still have the right data.
for i, b := range r {
if b != byte(i) {
t.Errorf("r[%d]=%d, want %d", i, b, byte(i))
}
}
}
func BenchmarkAppendInLoop(b *testing.B) {
for _, size := range []int{0, 1, 8, 16, 32, 64, 128} {
b.Run(fmt.Sprintf("%d", size),
func(b *testing.B) {
b.ReportAllocs()
for b.Loop() {
byteSlice(size)
}
})
}
}
func TestMoveToHeapEarly(t *testing.T) {
// Just checking that this compiles.
var x []int
y := x // causes a move2heap in the entry block
for range 5 {
x = append(x, 5)
}
_ = y
}
func TestMoveToHeapCap(t *testing.T) {
var c int
r := func() []byte {
var s []byte
for i := range 10 {
s = append(s, byte(i))
}
c = cap(s)
return s
}()
if c != cap(r) {
t.Errorf("got cap=%d, want %d", c, cap(r))
}
sinkSlice = r
}
//go:noinline
func runit(f func()) {
f()
}
func TestMoveToHeapClosure1(t *testing.T) {
var c int
r := func() []byte {
var s []byte
for i := range 10 {
s = append(s, byte(i))
}
runit(func() {
c = cap(s)
})
return s
}()
if c != cap(r) {
t.Errorf("got cap=%d, want %d", c, cap(r))
}
sinkSlice = r
}
func TestMoveToHeapClosure2(t *testing.T) {
var c int
r := func() []byte {
var s []byte
for i := range 10 {
s = append(s, byte(i))
}
c = func() int {
return cap(s)
}()
return s
}()
if c != cap(r) {
t.Errorf("got cap=%d, want %d", c, cap(r))
}
sinkSlice = r
}
//go:noinline
func buildClosure(t *testing.T) ([]byte, func()) {
var s []byte
for i := range 20 {
s = append(s, byte(i))
}
c := func() {
for i, b := range s {
if b != byte(i) {
t.Errorf("s[%d]=%d, want %d", i, b, i)
}
}
}
return s, c
}
func TestMoveToHeapClosure3(t *testing.T) {
_, f := buildClosure(t)
overwriteStack(0)
f()
}
//go:noinline
func overwriteStack(n int) uint64 {
var x [100]uint64
for i := range x {
x[i] = 0xabcdabcdabcdabcd
}
return x[n]
}
var sinkSlice []byte