|  | // Copyright 2017 The Go Authors. All rights reserved. | 
|  | // Use of this source code is governed by a BSD-style | 
|  | // license that can be found in the LICENSE file. | 
|  |  | 
|  | package bits_test | 
|  |  | 
|  | import ( | 
|  | . "math/bits" | 
|  | "runtime" | 
|  | "testing" | 
|  | "unsafe" | 
|  | ) | 
|  |  | 
|  | func TestUintSize(t *testing.T) { | 
|  | var x uint | 
|  | if want := unsafe.Sizeof(x) * 8; UintSize != want { | 
|  | t.Fatalf("UintSize = %d; want %d", UintSize, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestLeadingZeros(t *testing.T) { | 
|  | for i := 0; i < 256; i++ { | 
|  | nlz := tab[i].nlz | 
|  | for k := 0; k < 64-8; k++ { | 
|  | x := uint64(i) << uint(k) | 
|  | if x <= 1<<8-1 { | 
|  | got := LeadingZeros8(uint8(x)) | 
|  | want := nlz - k + (8 - 8) | 
|  | if x == 0 { | 
|  | want = 8 | 
|  | } | 
|  | if got != want { | 
|  | t.Fatalf("LeadingZeros8(%#02x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<16-1 { | 
|  | got := LeadingZeros16(uint16(x)) | 
|  | want := nlz - k + (16 - 8) | 
|  | if x == 0 { | 
|  | want = 16 | 
|  | } | 
|  | if got != want { | 
|  | t.Fatalf("LeadingZeros16(%#04x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<32-1 { | 
|  | got := LeadingZeros32(uint32(x)) | 
|  | want := nlz - k + (32 - 8) | 
|  | if x == 0 { | 
|  | want = 32 | 
|  | } | 
|  | if got != want { | 
|  | t.Fatalf("LeadingZeros32(%#08x) == %d; want %d", x, got, want) | 
|  | } | 
|  | if UintSize == 32 { | 
|  | got = LeadingZeros(uint(x)) | 
|  | if got != want { | 
|  | t.Fatalf("LeadingZeros(%#08x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<64-1 { | 
|  | got := LeadingZeros64(uint64(x)) | 
|  | want := nlz - k + (64 - 8) | 
|  | if x == 0 { | 
|  | want = 64 | 
|  | } | 
|  | if got != want { | 
|  | t.Fatalf("LeadingZeros64(%#016x) == %d; want %d", x, got, want) | 
|  | } | 
|  | if UintSize == 64 { | 
|  | got = LeadingZeros(uint(x)) | 
|  | if got != want { | 
|  | t.Fatalf("LeadingZeros(%#016x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | // Exported (global) variable serving as input for some | 
|  | // of the benchmarks to ensure side-effect free calls | 
|  | // are not optimized away. | 
|  | var Input uint64 = DeBruijn64 | 
|  |  | 
|  | // Exported (global) variable to store function results | 
|  | // during benchmarking to ensure side-effect free calls | 
|  | // are not optimized away. | 
|  | var Output int | 
|  |  | 
|  | func BenchmarkLeadingZeros(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += LeadingZeros(uint(Input) >> (uint(i) % UintSize)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkLeadingZeros8(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += LeadingZeros8(uint8(Input) >> (uint(i) % 8)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkLeadingZeros16(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += LeadingZeros16(uint16(Input) >> (uint(i) % 16)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkLeadingZeros32(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += LeadingZeros32(uint32(Input) >> (uint(i) % 32)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkLeadingZeros64(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += LeadingZeros64(uint64(Input) >> (uint(i) % 64)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func TestTrailingZeros(t *testing.T) { | 
|  | for i := 0; i < 256; i++ { | 
|  | ntz := tab[i].ntz | 
|  | for k := 0; k < 64-8; k++ { | 
|  | x := uint64(i) << uint(k) | 
|  | want := ntz + k | 
|  | if x <= 1<<8-1 { | 
|  | got := TrailingZeros8(uint8(x)) | 
|  | if x == 0 { | 
|  | want = 8 | 
|  | } | 
|  | if got != want { | 
|  | t.Fatalf("TrailingZeros8(%#02x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<16-1 { | 
|  | got := TrailingZeros16(uint16(x)) | 
|  | if x == 0 { | 
|  | want = 16 | 
|  | } | 
|  | if got != want { | 
|  | t.Fatalf("TrailingZeros16(%#04x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<32-1 { | 
|  | got := TrailingZeros32(uint32(x)) | 
|  | if x == 0 { | 
|  | want = 32 | 
|  | } | 
|  | if got != want { | 
|  | t.Fatalf("TrailingZeros32(%#08x) == %d; want %d", x, got, want) | 
|  | } | 
|  | if UintSize == 32 { | 
|  | got = TrailingZeros(uint(x)) | 
|  | if got != want { | 
|  | t.Fatalf("TrailingZeros(%#08x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<64-1 { | 
|  | got := TrailingZeros64(uint64(x)) | 
|  | if x == 0 { | 
|  | want = 64 | 
|  | } | 
|  | if got != want { | 
|  | t.Fatalf("TrailingZeros64(%#016x) == %d; want %d", x, got, want) | 
|  | } | 
|  | if UintSize == 64 { | 
|  | got = TrailingZeros(uint(x)) | 
|  | if got != want { | 
|  | t.Fatalf("TrailingZeros(%#016x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | func BenchmarkTrailingZeros(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += TrailingZeros(uint(Input) << (uint(i) % UintSize)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkTrailingZeros8(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += TrailingZeros8(uint8(Input) << (uint(i) % 8)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkTrailingZeros16(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += TrailingZeros16(uint16(Input) << (uint(i) % 16)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkTrailingZeros32(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += TrailingZeros32(uint32(Input) << (uint(i) % 32)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkTrailingZeros64(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += TrailingZeros64(uint64(Input) << (uint(i) % 64)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func TestOnesCount(t *testing.T) { | 
|  | var x uint64 | 
|  | for i := 0; i <= 64; i++ { | 
|  | testOnesCount(t, x, i) | 
|  | x = x<<1 | 1 | 
|  | } | 
|  |  | 
|  | for i := 64; i >= 0; i-- { | 
|  | testOnesCount(t, x, i) | 
|  | x = x << 1 | 
|  | } | 
|  |  | 
|  | for i := 0; i < 256; i++ { | 
|  | for k := 0; k < 64-8; k++ { | 
|  | testOnesCount(t, uint64(i)<<uint(k), tab[i].pop) | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | func testOnesCount(t *testing.T, x uint64, want int) { | 
|  | if x <= 1<<8-1 { | 
|  | got := OnesCount8(uint8(x)) | 
|  | if got != want { | 
|  | t.Fatalf("OnesCount8(%#02x) == %d; want %d", uint8(x), got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<16-1 { | 
|  | got := OnesCount16(uint16(x)) | 
|  | if got != want { | 
|  | t.Fatalf("OnesCount16(%#04x) == %d; want %d", uint16(x), got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<32-1 { | 
|  | got := OnesCount32(uint32(x)) | 
|  | if got != want { | 
|  | t.Fatalf("OnesCount32(%#08x) == %d; want %d", uint32(x), got, want) | 
|  | } | 
|  | if UintSize == 32 { | 
|  | got = OnesCount(uint(x)) | 
|  | if got != want { | 
|  | t.Fatalf("OnesCount(%#08x) == %d; want %d", uint32(x), got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<64-1 { | 
|  | got := OnesCount64(uint64(x)) | 
|  | if got != want { | 
|  | t.Fatalf("OnesCount64(%#016x) == %d; want %d", x, got, want) | 
|  | } | 
|  | if UintSize == 64 { | 
|  | got = OnesCount(uint(x)) | 
|  | if got != want { | 
|  | t.Fatalf("OnesCount(%#016x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | func BenchmarkOnesCount(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += OnesCount(uint(Input)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkOnesCount8(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += OnesCount8(uint8(Input)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkOnesCount16(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += OnesCount16(uint16(Input)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkOnesCount32(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += OnesCount32(uint32(Input)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func BenchmarkOnesCount64(b *testing.B) { | 
|  | var s int | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += OnesCount64(uint64(Input)) | 
|  | } | 
|  | Output = s | 
|  | } | 
|  |  | 
|  | func TestRotateLeft(t *testing.T) { | 
|  | var m uint64 = DeBruijn64 | 
|  |  | 
|  | for k := uint(0); k < 128; k++ { | 
|  | x8 := uint8(m) | 
|  | got8 := RotateLeft8(x8, int(k)) | 
|  | want8 := x8<<(k&0x7) | x8>>(8-k&0x7) | 
|  | if got8 != want8 { | 
|  | t.Fatalf("RotateLeft8(%#02x, %d) == %#02x; want %#02x", x8, k, got8, want8) | 
|  | } | 
|  | got8 = RotateLeft8(want8, -int(k)) | 
|  | if got8 != x8 { | 
|  | t.Fatalf("RotateLeft8(%#02x, -%d) == %#02x; want %#02x", want8, k, got8, x8) | 
|  | } | 
|  |  | 
|  | x16 := uint16(m) | 
|  | got16 := RotateLeft16(x16, int(k)) | 
|  | want16 := x16<<(k&0xf) | x16>>(16-k&0xf) | 
|  | if got16 != want16 { | 
|  | t.Fatalf("RotateLeft16(%#04x, %d) == %#04x; want %#04x", x16, k, got16, want16) | 
|  | } | 
|  | got16 = RotateLeft16(want16, -int(k)) | 
|  | if got16 != x16 { | 
|  | t.Fatalf("RotateLeft16(%#04x, -%d) == %#04x; want %#04x", want16, k, got16, x16) | 
|  | } | 
|  |  | 
|  | x32 := uint32(m) | 
|  | got32 := RotateLeft32(x32, int(k)) | 
|  | want32 := x32<<(k&0x1f) | x32>>(32-k&0x1f) | 
|  | if got32 != want32 { | 
|  | t.Fatalf("RotateLeft32(%#08x, %d) == %#08x; want %#08x", x32, k, got32, want32) | 
|  | } | 
|  | got32 = RotateLeft32(want32, -int(k)) | 
|  | if got32 != x32 { | 
|  | t.Fatalf("RotateLeft32(%#08x, -%d) == %#08x; want %#08x", want32, k, got32, x32) | 
|  | } | 
|  | if UintSize == 32 { | 
|  | x := uint(m) | 
|  | got := RotateLeft(x, int(k)) | 
|  | want := x<<(k&0x1f) | x>>(32-k&0x1f) | 
|  | if got != want { | 
|  | t.Fatalf("RotateLeft(%#08x, %d) == %#08x; want %#08x", x, k, got, want) | 
|  | } | 
|  | got = RotateLeft(want, -int(k)) | 
|  | if got != x { | 
|  | t.Fatalf("RotateLeft(%#08x, -%d) == %#08x; want %#08x", want, k, got, x) | 
|  | } | 
|  | } | 
|  |  | 
|  | x64 := uint64(m) | 
|  | got64 := RotateLeft64(x64, int(k)) | 
|  | want64 := x64<<(k&0x3f) | x64>>(64-k&0x3f) | 
|  | if got64 != want64 { | 
|  | t.Fatalf("RotateLeft64(%#016x, %d) == %#016x; want %#016x", x64, k, got64, want64) | 
|  | } | 
|  | got64 = RotateLeft64(want64, -int(k)) | 
|  | if got64 != x64 { | 
|  | t.Fatalf("RotateLeft64(%#016x, -%d) == %#016x; want %#016x", want64, k, got64, x64) | 
|  | } | 
|  | if UintSize == 64 { | 
|  | x := uint(m) | 
|  | got := RotateLeft(x, int(k)) | 
|  | want := x<<(k&0x3f) | x>>(64-k&0x3f) | 
|  | if got != want { | 
|  | t.Fatalf("RotateLeft(%#016x, %d) == %#016x; want %#016x", x, k, got, want) | 
|  | } | 
|  | got = RotateLeft(want, -int(k)) | 
|  | if got != x { | 
|  | t.Fatalf("RotateLeft(%#08x, -%d) == %#08x; want %#08x", want, k, got, x) | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | func BenchmarkRotateLeft(b *testing.B) { | 
|  | var s uint | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += RotateLeft(uint(Input), i) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkRotateLeft8(b *testing.B) { | 
|  | var s uint8 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += RotateLeft8(uint8(Input), i) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkRotateLeft16(b *testing.B) { | 
|  | var s uint16 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += RotateLeft16(uint16(Input), i) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkRotateLeft32(b *testing.B) { | 
|  | var s uint32 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += RotateLeft32(uint32(Input), i) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkRotateLeft64(b *testing.B) { | 
|  | var s uint64 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += RotateLeft64(uint64(Input), i) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func TestReverse(t *testing.T) { | 
|  | // test each bit | 
|  | for i := uint(0); i < 64; i++ { | 
|  | testReverse(t, uint64(1)<<i, uint64(1)<<(63-i)) | 
|  | } | 
|  |  | 
|  | // test a few patterns | 
|  | for _, test := range []struct { | 
|  | x, r uint64 | 
|  | }{ | 
|  | {0, 0}, | 
|  | {0x1, 0x8 << 60}, | 
|  | {0x2, 0x4 << 60}, | 
|  | {0x3, 0xc << 60}, | 
|  | {0x4, 0x2 << 60}, | 
|  | {0x5, 0xa << 60}, | 
|  | {0x6, 0x6 << 60}, | 
|  | {0x7, 0xe << 60}, | 
|  | {0x8, 0x1 << 60}, | 
|  | {0x9, 0x9 << 60}, | 
|  | {0xa, 0x5 << 60}, | 
|  | {0xb, 0xd << 60}, | 
|  | {0xc, 0x3 << 60}, | 
|  | {0xd, 0xb << 60}, | 
|  | {0xe, 0x7 << 60}, | 
|  | {0xf, 0xf << 60}, | 
|  | {0x5686487, 0xe12616a000000000}, | 
|  | {0x0123456789abcdef, 0xf7b3d591e6a2c480}, | 
|  | } { | 
|  | testReverse(t, test.x, test.r) | 
|  | testReverse(t, test.r, test.x) | 
|  | } | 
|  | } | 
|  |  | 
|  | func testReverse(t *testing.T, x64, want64 uint64) { | 
|  | x8 := uint8(x64) | 
|  | got8 := Reverse8(x8) | 
|  | want8 := uint8(want64 >> (64 - 8)) | 
|  | if got8 != want8 { | 
|  | t.Fatalf("Reverse8(%#02x) == %#02x; want %#02x", x8, got8, want8) | 
|  | } | 
|  |  | 
|  | x16 := uint16(x64) | 
|  | got16 := Reverse16(x16) | 
|  | want16 := uint16(want64 >> (64 - 16)) | 
|  | if got16 != want16 { | 
|  | t.Fatalf("Reverse16(%#04x) == %#04x; want %#04x", x16, got16, want16) | 
|  | } | 
|  |  | 
|  | x32 := uint32(x64) | 
|  | got32 := Reverse32(x32) | 
|  | want32 := uint32(want64 >> (64 - 32)) | 
|  | if got32 != want32 { | 
|  | t.Fatalf("Reverse32(%#08x) == %#08x; want %#08x", x32, got32, want32) | 
|  | } | 
|  | if UintSize == 32 { | 
|  | x := uint(x32) | 
|  | got := Reverse(x) | 
|  | want := uint(want32) | 
|  | if got != want { | 
|  | t.Fatalf("Reverse(%#08x) == %#08x; want %#08x", x, got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | got64 := Reverse64(x64) | 
|  | if got64 != want64 { | 
|  | t.Fatalf("Reverse64(%#016x) == %#016x; want %#016x", x64, got64, want64) | 
|  | } | 
|  | if UintSize == 64 { | 
|  | x := uint(x64) | 
|  | got := Reverse(x) | 
|  | want := uint(want64) | 
|  | if got != want { | 
|  | t.Fatalf("Reverse(%#08x) == %#016x; want %#016x", x, got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | func BenchmarkReverse(b *testing.B) { | 
|  | var s uint | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += Reverse(uint(i)) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkReverse8(b *testing.B) { | 
|  | var s uint8 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += Reverse8(uint8(i)) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkReverse16(b *testing.B) { | 
|  | var s uint16 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += Reverse16(uint16(i)) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkReverse32(b *testing.B) { | 
|  | var s uint32 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += Reverse32(uint32(i)) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkReverse64(b *testing.B) { | 
|  | var s uint64 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += Reverse64(uint64(i)) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func TestReverseBytes(t *testing.T) { | 
|  | for _, test := range []struct { | 
|  | x, r uint64 | 
|  | }{ | 
|  | {0, 0}, | 
|  | {0x01, 0x01 << 56}, | 
|  | {0x0123, 0x2301 << 48}, | 
|  | {0x012345, 0x452301 << 40}, | 
|  | {0x01234567, 0x67452301 << 32}, | 
|  | {0x0123456789, 0x8967452301 << 24}, | 
|  | {0x0123456789ab, 0xab8967452301 << 16}, | 
|  | {0x0123456789abcd, 0xcdab8967452301 << 8}, | 
|  | {0x0123456789abcdef, 0xefcdab8967452301 << 0}, | 
|  | } { | 
|  | testReverseBytes(t, test.x, test.r) | 
|  | testReverseBytes(t, test.r, test.x) | 
|  | } | 
|  | } | 
|  |  | 
|  | func testReverseBytes(t *testing.T, x64, want64 uint64) { | 
|  | x16 := uint16(x64) | 
|  | got16 := ReverseBytes16(x16) | 
|  | want16 := uint16(want64 >> (64 - 16)) | 
|  | if got16 != want16 { | 
|  | t.Fatalf("ReverseBytes16(%#04x) == %#04x; want %#04x", x16, got16, want16) | 
|  | } | 
|  |  | 
|  | x32 := uint32(x64) | 
|  | got32 := ReverseBytes32(x32) | 
|  | want32 := uint32(want64 >> (64 - 32)) | 
|  | if got32 != want32 { | 
|  | t.Fatalf("ReverseBytes32(%#08x) == %#08x; want %#08x", x32, got32, want32) | 
|  | } | 
|  | if UintSize == 32 { | 
|  | x := uint(x32) | 
|  | got := ReverseBytes(x) | 
|  | want := uint(want32) | 
|  | if got != want { | 
|  | t.Fatalf("ReverseBytes(%#08x) == %#08x; want %#08x", x, got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | got64 := ReverseBytes64(x64) | 
|  | if got64 != want64 { | 
|  | t.Fatalf("ReverseBytes64(%#016x) == %#016x; want %#016x", x64, got64, want64) | 
|  | } | 
|  | if UintSize == 64 { | 
|  | x := uint(x64) | 
|  | got := ReverseBytes(x) | 
|  | want := uint(want64) | 
|  | if got != want { | 
|  | t.Fatalf("ReverseBytes(%#016x) == %#016x; want %#016x", x, got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | func BenchmarkReverseBytes(b *testing.B) { | 
|  | var s uint | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += ReverseBytes(uint(i)) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkReverseBytes16(b *testing.B) { | 
|  | var s uint16 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += ReverseBytes16(uint16(i)) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkReverseBytes32(b *testing.B) { | 
|  | var s uint32 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += ReverseBytes32(uint32(i)) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func BenchmarkReverseBytes64(b *testing.B) { | 
|  | var s uint64 | 
|  | for i := 0; i < b.N; i++ { | 
|  | s += ReverseBytes64(uint64(i)) | 
|  | } | 
|  | Output = int(s) | 
|  | } | 
|  |  | 
|  | func TestLen(t *testing.T) { | 
|  | for i := 0; i < 256; i++ { | 
|  | len := 8 - tab[i].nlz | 
|  | for k := 0; k < 64-8; k++ { | 
|  | x := uint64(i) << uint(k) | 
|  | want := 0 | 
|  | if x != 0 { | 
|  | want = len + k | 
|  | } | 
|  | if x <= 1<<8-1 { | 
|  | got := Len8(uint8(x)) | 
|  | if got != want { | 
|  | t.Fatalf("Len8(%#02x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<16-1 { | 
|  | got := Len16(uint16(x)) | 
|  | if got != want { | 
|  | t.Fatalf("Len16(%#04x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<32-1 { | 
|  | got := Len32(uint32(x)) | 
|  | if got != want { | 
|  | t.Fatalf("Len32(%#08x) == %d; want %d", x, got, want) | 
|  | } | 
|  | if UintSize == 32 { | 
|  | got := Len(uint(x)) | 
|  | if got != want { | 
|  | t.Fatalf("Len(%#08x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | if x <= 1<<64-1 { | 
|  | got := Len64(uint64(x)) | 
|  | if got != want { | 
|  | t.Fatalf("Len64(%#016x) == %d; want %d", x, got, want) | 
|  | } | 
|  | if UintSize == 64 { | 
|  | got := Len(uint(x)) | 
|  | if got != want { | 
|  | t.Fatalf("Len(%#016x) == %d; want %d", x, got, want) | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | const ( | 
|  | _M   = 1<<UintSize - 1 | 
|  | _M32 = 1<<32 - 1 | 
|  | _M64 = 1<<64 - 1 | 
|  | ) | 
|  |  | 
|  | func TestAddSubUint(t *testing.T) { | 
|  | test := func(msg string, f func(x, y, c uint) (z, cout uint), x, y, c, z, cout uint) { | 
|  | z1, cout1 := f(x, y, c) | 
|  | if z1 != z || cout1 != cout { | 
|  | t.Errorf("%s: got z:cout = %#x:%#x; want %#x:%#x", msg, z1, cout1, z, cout) | 
|  | } | 
|  | } | 
|  | for _, a := range []struct{ x, y, c, z, cout uint }{ | 
|  | {0, 0, 0, 0, 0}, | 
|  | {0, 1, 0, 1, 0}, | 
|  | {0, 0, 1, 1, 0}, | 
|  | {0, 1, 1, 2, 0}, | 
|  | {12345, 67890, 0, 80235, 0}, | 
|  | {12345, 67890, 1, 80236, 0}, | 
|  | {_M, 1, 0, 0, 1}, | 
|  | {_M, 0, 1, 0, 1}, | 
|  | {_M, 1, 1, 1, 1}, | 
|  | {_M, _M, 0, _M - 1, 1}, | 
|  | {_M, _M, 1, _M, 1}, | 
|  | } { | 
|  | test("Add", Add, a.x, a.y, a.c, a.z, a.cout) | 
|  | test("Add symmetric", Add, a.y, a.x, a.c, a.z, a.cout) | 
|  | test("Sub", Sub, a.z, a.x, a.c, a.y, a.cout) | 
|  | test("Sub symmetric", Sub, a.z, a.y, a.c, a.x, a.cout) | 
|  | // The above code can't test intrinsic implementation, because the passed function is not called directly. | 
|  | // The following code uses a closure to test the intrinsic version in case the function is intrinsified. | 
|  | test("Add intrinsic", func(x, y, c uint) (uint, uint) { return Add(x, y, c) }, a.x, a.y, a.c, a.z, a.cout) | 
|  | test("Add intrinsic symmetric", func(x, y, c uint) (uint, uint) { return Add(x, y, c) }, a.y, a.x, a.c, a.z, a.cout) | 
|  | test("Sub intrinsic", func(x, y, c uint) (uint, uint) { return Sub(x, y, c) }, a.z, a.x, a.c, a.y, a.cout) | 
|  | test("Sub intrinsic symmetric", func(x, y, c uint) (uint, uint) { return Sub(x, y, c) }, a.z, a.y, a.c, a.x, a.cout) | 
|  |  | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestAddSubUint32(t *testing.T) { | 
|  | test := func(msg string, f func(x, y, c uint32) (z, cout uint32), x, y, c, z, cout uint32) { | 
|  | z1, cout1 := f(x, y, c) | 
|  | if z1 != z || cout1 != cout { | 
|  | t.Errorf("%s: got z:cout = %#x:%#x; want %#x:%#x", msg, z1, cout1, z, cout) | 
|  | } | 
|  | } | 
|  | for _, a := range []struct{ x, y, c, z, cout uint32 }{ | 
|  | {0, 0, 0, 0, 0}, | 
|  | {0, 1, 0, 1, 0}, | 
|  | {0, 0, 1, 1, 0}, | 
|  | {0, 1, 1, 2, 0}, | 
|  | {12345, 67890, 0, 80235, 0}, | 
|  | {12345, 67890, 1, 80236, 0}, | 
|  | {_M32, 1, 0, 0, 1}, | 
|  | {_M32, 0, 1, 0, 1}, | 
|  | {_M32, 1, 1, 1, 1}, | 
|  | {_M32, _M32, 0, _M32 - 1, 1}, | 
|  | {_M32, _M32, 1, _M32, 1}, | 
|  | } { | 
|  | test("Add32", Add32, a.x, a.y, a.c, a.z, a.cout) | 
|  | test("Add32 symmetric", Add32, a.y, a.x, a.c, a.z, a.cout) | 
|  | test("Sub32", Sub32, a.z, a.x, a.c, a.y, a.cout) | 
|  | test("Sub32 symmetric", Sub32, a.z, a.y, a.c, a.x, a.cout) | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestAddSubUint64(t *testing.T) { | 
|  | test := func(msg string, f func(x, y, c uint64) (z, cout uint64), x, y, c, z, cout uint64) { | 
|  | z1, cout1 := f(x, y, c) | 
|  | if z1 != z || cout1 != cout { | 
|  | t.Errorf("%s: got z:cout = %#x:%#x; want %#x:%#x", msg, z1, cout1, z, cout) | 
|  | } | 
|  | } | 
|  | for _, a := range []struct{ x, y, c, z, cout uint64 }{ | 
|  | {0, 0, 0, 0, 0}, | 
|  | {0, 1, 0, 1, 0}, | 
|  | {0, 0, 1, 1, 0}, | 
|  | {0, 1, 1, 2, 0}, | 
|  | {12345, 67890, 0, 80235, 0}, | 
|  | {12345, 67890, 1, 80236, 0}, | 
|  | {_M64, 1, 0, 0, 1}, | 
|  | {_M64, 0, 1, 0, 1}, | 
|  | {_M64, 1, 1, 1, 1}, | 
|  | {_M64, _M64, 0, _M64 - 1, 1}, | 
|  | {_M64, _M64, 1, _M64, 1}, | 
|  | } { | 
|  | test("Add64", Add64, a.x, a.y, a.c, a.z, a.cout) | 
|  | test("Add64 symmetric", Add64, a.y, a.x, a.c, a.z, a.cout) | 
|  | test("Sub64", Sub64, a.z, a.x, a.c, a.y, a.cout) | 
|  | test("Sub64 symmetric", Sub64, a.z, a.y, a.c, a.x, a.cout) | 
|  | // The above code can't test intrinsic implementation, because the passed function is not called directly. | 
|  | // The following code uses a closure to test the intrinsic version in case the function is intrinsified. | 
|  | test("Add64 intrinsic", func(x, y, c uint64) (uint64, uint64) { return Add64(x, y, c) }, a.x, a.y, a.c, a.z, a.cout) | 
|  | test("Add64 intrinsic symmetric", func(x, y, c uint64) (uint64, uint64) { return Add64(x, y, c) }, a.y, a.x, a.c, a.z, a.cout) | 
|  | test("Sub64 intrinsic", func(x, y, c uint64) (uint64, uint64) { return Sub64(x, y, c) }, a.z, a.x, a.c, a.y, a.cout) | 
|  | test("Sub64 intrinsic symmetric", func(x, y, c uint64) (uint64, uint64) { return Sub64(x, y, c) }, a.z, a.y, a.c, a.x, a.cout) | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestAdd64OverflowPanic(t *testing.T) { | 
|  | // Test that 64-bit overflow panics fire correctly. | 
|  | // These are designed to improve coverage of compiler intrinsics. | 
|  | tests := []func(uint64, uint64) uint64{ | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Add64(a, b, 0) | 
|  | if c > 0 { | 
|  | panic("overflow") | 
|  | } | 
|  | return x | 
|  | }, | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Add64(a, b, 0) | 
|  | if c != 0 { | 
|  | panic("overflow") | 
|  | } | 
|  | return x | 
|  | }, | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Add64(a, b, 0) | 
|  | if c == 1 { | 
|  | panic("overflow") | 
|  | } | 
|  | return x | 
|  | }, | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Add64(a, b, 0) | 
|  | if c != 1 { | 
|  | return x | 
|  | } | 
|  | panic("overflow") | 
|  | }, | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Add64(a, b, 0) | 
|  | if c == 0 { | 
|  | return x | 
|  | } | 
|  | panic("overflow") | 
|  | }, | 
|  | } | 
|  | for _, test := range tests { | 
|  | shouldPanic := func(f func()) { | 
|  | defer func() { | 
|  | if err := recover(); err == nil { | 
|  | t.Fatalf("expected panic") | 
|  | } | 
|  | }() | 
|  | f() | 
|  | } | 
|  |  | 
|  | // overflow | 
|  | shouldPanic(func() { test(_M64, 1) }) | 
|  | shouldPanic(func() { test(1, _M64) }) | 
|  | shouldPanic(func() { test(_M64, _M64) }) | 
|  |  | 
|  | // no overflow | 
|  | test(_M64, 0) | 
|  | test(0, 0) | 
|  | test(1, 1) | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestSub64OverflowPanic(t *testing.T) { | 
|  | // Test that 64-bit overflow panics fire correctly. | 
|  | // These are designed to improve coverage of compiler intrinsics. | 
|  | tests := []func(uint64, uint64) uint64{ | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Sub64(a, b, 0) | 
|  | if c > 0 { | 
|  | panic("overflow") | 
|  | } | 
|  | return x | 
|  | }, | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Sub64(a, b, 0) | 
|  | if c != 0 { | 
|  | panic("overflow") | 
|  | } | 
|  | return x | 
|  | }, | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Sub64(a, b, 0) | 
|  | if c == 1 { | 
|  | panic("overflow") | 
|  | } | 
|  | return x | 
|  | }, | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Sub64(a, b, 0) | 
|  | if c != 1 { | 
|  | return x | 
|  | } | 
|  | panic("overflow") | 
|  | }, | 
|  | func(a, b uint64) uint64 { | 
|  | x, c := Sub64(a, b, 0) | 
|  | if c == 0 { | 
|  | return x | 
|  | } | 
|  | panic("overflow") | 
|  | }, | 
|  | } | 
|  | for _, test := range tests { | 
|  | shouldPanic := func(f func()) { | 
|  | defer func() { | 
|  | if err := recover(); err == nil { | 
|  | t.Fatalf("expected panic") | 
|  | } | 
|  | }() | 
|  | f() | 
|  | } | 
|  |  | 
|  | // overflow | 
|  | shouldPanic(func() { test(0, 1) }) | 
|  | shouldPanic(func() { test(1, _M64) }) | 
|  | shouldPanic(func() { test(_M64-1, _M64) }) | 
|  |  | 
|  | // no overflow | 
|  | test(_M64, 0) | 
|  | test(0, 0) | 
|  | test(1, 1) | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestMulDiv(t *testing.T) { | 
|  | testMul := func(msg string, f func(x, y uint) (hi, lo uint), x, y, hi, lo uint) { | 
|  | hi1, lo1 := f(x, y) | 
|  | if hi1 != hi || lo1 != lo { | 
|  | t.Errorf("%s: got hi:lo = %#x:%#x; want %#x:%#x", msg, hi1, lo1, hi, lo) | 
|  | } | 
|  | } | 
|  | testDiv := func(msg string, f func(hi, lo, y uint) (q, r uint), hi, lo, y, q, r uint) { | 
|  | q1, r1 := f(hi, lo, y) | 
|  | if q1 != q || r1 != r { | 
|  | t.Errorf("%s: got q:r = %#x:%#x; want %#x:%#x", msg, q1, r1, q, r) | 
|  | } | 
|  | } | 
|  | for _, a := range []struct { | 
|  | x, y      uint | 
|  | hi, lo, r uint | 
|  | }{ | 
|  | {1 << (UintSize - 1), 2, 1, 0, 1}, | 
|  | {_M, _M, _M - 1, 1, 42}, | 
|  | } { | 
|  | testMul("Mul", Mul, a.x, a.y, a.hi, a.lo) | 
|  | testMul("Mul symmetric", Mul, a.y, a.x, a.hi, a.lo) | 
|  | testDiv("Div", Div, a.hi, a.lo+a.r, a.y, a.x, a.r) | 
|  | testDiv("Div symmetric", Div, a.hi, a.lo+a.r, a.x, a.y, a.r) | 
|  | // The above code can't test intrinsic implementation, because the passed function is not called directly. | 
|  | // The following code uses a closure to test the intrinsic version in case the function is intrinsified. | 
|  | testMul("Mul intrinsic", func(x, y uint) (uint, uint) { return Mul(x, y) }, a.x, a.y, a.hi, a.lo) | 
|  | testMul("Mul intrinsic symmetric", func(x, y uint) (uint, uint) { return Mul(x, y) }, a.y, a.x, a.hi, a.lo) | 
|  | testDiv("Div intrinsic", func(hi, lo, y uint) (uint, uint) { return Div(hi, lo, y) }, a.hi, a.lo+a.r, a.y, a.x, a.r) | 
|  | testDiv("Div intrinsic symmetric", func(hi, lo, y uint) (uint, uint) { return Div(hi, lo, y) }, a.hi, a.lo+a.r, a.x, a.y, a.r) | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestMulDiv32(t *testing.T) { | 
|  | testMul := func(msg string, f func(x, y uint32) (hi, lo uint32), x, y, hi, lo uint32) { | 
|  | hi1, lo1 := f(x, y) | 
|  | if hi1 != hi || lo1 != lo { | 
|  | t.Errorf("%s: got hi:lo = %#x:%#x; want %#x:%#x", msg, hi1, lo1, hi, lo) | 
|  | } | 
|  | } | 
|  | testDiv := func(msg string, f func(hi, lo, y uint32) (q, r uint32), hi, lo, y, q, r uint32) { | 
|  | q1, r1 := f(hi, lo, y) | 
|  | if q1 != q || r1 != r { | 
|  | t.Errorf("%s: got q:r = %#x:%#x; want %#x:%#x", msg, q1, r1, q, r) | 
|  | } | 
|  | } | 
|  | for _, a := range []struct { | 
|  | x, y      uint32 | 
|  | hi, lo, r uint32 | 
|  | }{ | 
|  | {1 << 31, 2, 1, 0, 1}, | 
|  | {0xc47dfa8c, 50911, 0x98a4, 0x998587f4, 13}, | 
|  | {_M32, _M32, _M32 - 1, 1, 42}, | 
|  | } { | 
|  | testMul("Mul32", Mul32, a.x, a.y, a.hi, a.lo) | 
|  | testMul("Mul32 symmetric", Mul32, a.y, a.x, a.hi, a.lo) | 
|  | testDiv("Div32", Div32, a.hi, a.lo+a.r, a.y, a.x, a.r) | 
|  | testDiv("Div32 symmetric", Div32, a.hi, a.lo+a.r, a.x, a.y, a.r) | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestMulDiv64(t *testing.T) { | 
|  | testMul := func(msg string, f func(x, y uint64) (hi, lo uint64), x, y, hi, lo uint64) { | 
|  | hi1, lo1 := f(x, y) | 
|  | if hi1 != hi || lo1 != lo { | 
|  | t.Errorf("%s: got hi:lo = %#x:%#x; want %#x:%#x", msg, hi1, lo1, hi, lo) | 
|  | } | 
|  | } | 
|  | testDiv := func(msg string, f func(hi, lo, y uint64) (q, r uint64), hi, lo, y, q, r uint64) { | 
|  | q1, r1 := f(hi, lo, y) | 
|  | if q1 != q || r1 != r { | 
|  | t.Errorf("%s: got q:r = %#x:%#x; want %#x:%#x", msg, q1, r1, q, r) | 
|  | } | 
|  | } | 
|  | for _, a := range []struct { | 
|  | x, y      uint64 | 
|  | hi, lo, r uint64 | 
|  | }{ | 
|  | {1 << 63, 2, 1, 0, 1}, | 
|  | {0x3626229738a3b9, 0xd8988a9f1cc4a61, 0x2dd0712657fe8, 0x9dd6a3364c358319, 13}, | 
|  | {_M64, _M64, _M64 - 1, 1, 42}, | 
|  | } { | 
|  | testMul("Mul64", Mul64, a.x, a.y, a.hi, a.lo) | 
|  | testMul("Mul64 symmetric", Mul64, a.y, a.x, a.hi, a.lo) | 
|  | testDiv("Div64", Div64, a.hi, a.lo+a.r, a.y, a.x, a.r) | 
|  | testDiv("Div64 symmetric", Div64, a.hi, a.lo+a.r, a.x, a.y, a.r) | 
|  | // The above code can't test intrinsic implementation, because the passed function is not called directly. | 
|  | // The following code uses a closure to test the intrinsic version in case the function is intrinsified. | 
|  | testMul("Mul64 intrinsic", func(x, y uint64) (uint64, uint64) { return Mul64(x, y) }, a.x, a.y, a.hi, a.lo) | 
|  | testMul("Mul64 intrinsic symmetric", func(x, y uint64) (uint64, uint64) { return Mul64(x, y) }, a.y, a.x, a.hi, a.lo) | 
|  | testDiv("Div64 intrinsic", func(hi, lo, y uint64) (uint64, uint64) { return Div64(hi, lo, y) }, a.hi, a.lo+a.r, a.y, a.x, a.r) | 
|  | testDiv("Div64 intrinsic symmetric", func(hi, lo, y uint64) (uint64, uint64) { return Div64(hi, lo, y) }, a.hi, a.lo+a.r, a.x, a.y, a.r) | 
|  | } | 
|  | } | 
|  |  | 
|  | const ( | 
|  | divZeroError  = "runtime error: integer divide by zero" | 
|  | overflowError = "runtime error: integer overflow" | 
|  | ) | 
|  |  | 
|  | func TestDivPanicOverflow(t *testing.T) { | 
|  | // Expect a panic | 
|  | defer func() { | 
|  | if err := recover(); err == nil { | 
|  | t.Error("Div should have panicked when y<=hi") | 
|  | } else if e, ok := err.(runtime.Error); !ok || e.Error() != overflowError { | 
|  | t.Errorf("Div expected panic: %q, got: %q ", overflowError, e.Error()) | 
|  | } | 
|  | }() | 
|  | q, r := Div(1, 0, 1) | 
|  | t.Errorf("undefined q, r = %v, %v calculated when Div should have panicked", q, r) | 
|  | } | 
|  |  | 
|  | func TestDiv32PanicOverflow(t *testing.T) { | 
|  | // Expect a panic | 
|  | defer func() { | 
|  | if err := recover(); err == nil { | 
|  | t.Error("Div32 should have panicked when y<=hi") | 
|  | } else if e, ok := err.(runtime.Error); !ok || e.Error() != overflowError { | 
|  | t.Errorf("Div32 expected panic: %q, got: %q ", overflowError, e.Error()) | 
|  | } | 
|  | }() | 
|  | q, r := Div32(1, 0, 1) | 
|  | t.Errorf("undefined q, r = %v, %v calculated when Div32 should have panicked", q, r) | 
|  | } | 
|  |  | 
|  | func TestDiv64PanicOverflow(t *testing.T) { | 
|  | // Expect a panic | 
|  | defer func() { | 
|  | if err := recover(); err == nil { | 
|  | t.Error("Div64 should have panicked when y<=hi") | 
|  | } else if e, ok := err.(runtime.Error); !ok || e.Error() != overflowError { | 
|  | t.Errorf("Div64 expected panic: %q, got: %q ", overflowError, e.Error()) | 
|  | } | 
|  | }() | 
|  | q, r := Div64(1, 0, 1) | 
|  | t.Errorf("undefined q, r = %v, %v calculated when Div64 should have panicked", q, r) | 
|  | } | 
|  |  | 
|  | func TestDivPanicZero(t *testing.T) { | 
|  | // Expect a panic | 
|  | defer func() { | 
|  | if err := recover(); err == nil { | 
|  | t.Error("Div should have panicked when y==0") | 
|  | } else if e, ok := err.(runtime.Error); !ok || e.Error() != divZeroError { | 
|  | t.Errorf("Div expected panic: %q, got: %q ", divZeroError, e.Error()) | 
|  | } | 
|  | }() | 
|  | q, r := Div(1, 1, 0) | 
|  | t.Errorf("undefined q, r = %v, %v calculated when Div should have panicked", q, r) | 
|  | } | 
|  |  | 
|  | func TestDiv32PanicZero(t *testing.T) { | 
|  | // Expect a panic | 
|  | defer func() { | 
|  | if err := recover(); err == nil { | 
|  | t.Error("Div32 should have panicked when y==0") | 
|  | } else if e, ok := err.(runtime.Error); !ok || e.Error() != divZeroError { | 
|  | t.Errorf("Div32 expected panic: %q, got: %q ", divZeroError, e.Error()) | 
|  | } | 
|  | }() | 
|  | q, r := Div32(1, 1, 0) | 
|  | t.Errorf("undefined q, r = %v, %v calculated when Div32 should have panicked", q, r) | 
|  | } | 
|  |  | 
|  | func TestDiv64PanicZero(t *testing.T) { | 
|  | // Expect a panic | 
|  | defer func() { | 
|  | if err := recover(); err == nil { | 
|  | t.Error("Div64 should have panicked when y==0") | 
|  | } else if e, ok := err.(runtime.Error); !ok || e.Error() != divZeroError { | 
|  | t.Errorf("Div64 expected panic: %q, got: %q ", divZeroError, e.Error()) | 
|  | } | 
|  | }() | 
|  | q, r := Div64(1, 1, 0) | 
|  | t.Errorf("undefined q, r = %v, %v calculated when Div64 should have panicked", q, r) | 
|  | } | 
|  |  | 
|  | func TestRem32(t *testing.T) { | 
|  | // Sanity check: for non-oveflowing dividends, the result is the | 
|  | // same as the rem returned by Div32 | 
|  | hi, lo, y := uint32(510510), uint32(9699690), uint32(510510+1) // ensure hi < y | 
|  | for i := 0; i < 1000; i++ { | 
|  | r := Rem32(hi, lo, y) | 
|  | _, r2 := Div32(hi, lo, y) | 
|  | if r != r2 { | 
|  | t.Errorf("Rem32(%v, %v, %v) returned %v, but Div32 returned rem %v", hi, lo, y, r, r2) | 
|  | } | 
|  | y += 13 | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestRem32Overflow(t *testing.T) { | 
|  | // To trigger a quotient overflow, we need y <= hi | 
|  | hi, lo, y := uint32(510510), uint32(9699690), uint32(7) | 
|  | for i := 0; i < 1000; i++ { | 
|  | r := Rem32(hi, lo, y) | 
|  | _, r2 := Div64(0, uint64(hi)<<32|uint64(lo), uint64(y)) | 
|  | if r != uint32(r2) { | 
|  | t.Errorf("Rem32(%v, %v, %v) returned %v, but Div64 returned rem %v", hi, lo, y, r, r2) | 
|  | } | 
|  | y += 13 | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestRem64(t *testing.T) { | 
|  | // Sanity check: for non-oveflowing dividends, the result is the | 
|  | // same as the rem returned by Div64 | 
|  | hi, lo, y := uint64(510510), uint64(9699690), uint64(510510+1) // ensure hi < y | 
|  | for i := 0; i < 1000; i++ { | 
|  | r := Rem64(hi, lo, y) | 
|  | _, r2 := Div64(hi, lo, y) | 
|  | if r != r2 { | 
|  | t.Errorf("Rem64(%v, %v, %v) returned %v, but Div64 returned rem %v", hi, lo, y, r, r2) | 
|  | } | 
|  | y += 13 | 
|  | } | 
|  | } | 
|  |  | 
|  | func TestRem64Overflow(t *testing.T) { | 
|  | Rem64Tests := []struct { | 
|  | hi, lo, y uint64 | 
|  | rem       uint64 | 
|  | }{ | 
|  | // Testcases computed using Python 3, as: | 
|  | //   >>> hi = 42; lo = 1119; y = 42 | 
|  | //   >>> ((hi<<64)+lo) % y | 
|  | {42, 1119, 42, 27}, | 
|  | {42, 1119, 38, 9}, | 
|  | {42, 1119, 26, 23}, | 
|  | {469, 0, 467, 271}, | 
|  | {469, 0, 113, 58}, | 
|  | {111111, 111111, 1171, 803}, | 
|  | {3968194946088682615, 3192705705065114702, 1000037, 56067}, | 
|  | } | 
|  |  | 
|  | for _, rt := range Rem64Tests { | 
|  | if rt.hi < rt.y { | 
|  | t.Fatalf("Rem64(%v, %v, %v) is not a test with quo overflow", rt.hi, rt.lo, rt.y) | 
|  | } | 
|  | rem := Rem64(rt.hi, rt.lo, rt.y) | 
|  | if rem != rt.rem { | 
|  | t.Errorf("Rem64(%v, %v, %v) returned %v, wanted %v", | 
|  | rt.hi, rt.lo, rt.y, rem, rt.rem) | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | func BenchmarkAdd(b *testing.B) { | 
|  | var z, c uint | 
|  | for i := 0; i < b.N; i++ { | 
|  | z, c = Add(uint(Input), uint(i), c) | 
|  | } | 
|  | Output = int(z + c) | 
|  | } | 
|  |  | 
|  | func BenchmarkAdd32(b *testing.B) { | 
|  | var z, c uint32 | 
|  | for i := 0; i < b.N; i++ { | 
|  | z, c = Add32(uint32(Input), uint32(i), c) | 
|  | } | 
|  | Output = int(z + c) | 
|  | } | 
|  |  | 
|  | func BenchmarkAdd64(b *testing.B) { | 
|  | var z, c uint64 | 
|  | for i := 0; i < b.N; i++ { | 
|  | z, c = Add64(uint64(Input), uint64(i), c) | 
|  | } | 
|  | Output = int(z + c) | 
|  | } | 
|  |  | 
|  | func BenchmarkAdd64multiple(b *testing.B) { | 
|  | var z0 = uint64(Input) | 
|  | var z1 = uint64(Input) | 
|  | var z2 = uint64(Input) | 
|  | var z3 = uint64(Input) | 
|  | for i := 0; i < b.N; i++ { | 
|  | var c uint64 | 
|  | z0, c = Add64(z0, uint64(i), c) | 
|  | z1, c = Add64(z1, uint64(i), c) | 
|  | z2, c = Add64(z2, uint64(i), c) | 
|  | z3, _ = Add64(z3, uint64(i), c) | 
|  | } | 
|  | Output = int(z0 + z1 + z2 + z3) | 
|  | } | 
|  |  | 
|  | func BenchmarkSub(b *testing.B) { | 
|  | var z, c uint | 
|  | for i := 0; i < b.N; i++ { | 
|  | z, c = Sub(uint(Input), uint(i), c) | 
|  | } | 
|  | Output = int(z + c) | 
|  | } | 
|  |  | 
|  | func BenchmarkSub32(b *testing.B) { | 
|  | var z, c uint32 | 
|  | for i := 0; i < b.N; i++ { | 
|  | z, c = Sub32(uint32(Input), uint32(i), c) | 
|  | } | 
|  | Output = int(z + c) | 
|  | } | 
|  |  | 
|  | func BenchmarkSub64(b *testing.B) { | 
|  | var z, c uint64 | 
|  | for i := 0; i < b.N; i++ { | 
|  | z, c = Sub64(uint64(Input), uint64(i), c) | 
|  | } | 
|  | Output = int(z + c) | 
|  | } | 
|  |  | 
|  | func BenchmarkSub64multiple(b *testing.B) { | 
|  | var z0 = uint64(Input) | 
|  | var z1 = uint64(Input) | 
|  | var z2 = uint64(Input) | 
|  | var z3 = uint64(Input) | 
|  | for i := 0; i < b.N; i++ { | 
|  | var c uint64 | 
|  | z0, c = Sub64(z0, uint64(i), c) | 
|  | z1, c = Sub64(z1, uint64(i), c) | 
|  | z2, c = Sub64(z2, uint64(i), c) | 
|  | z3, _ = Sub64(z3, uint64(i), c) | 
|  | } | 
|  | Output = int(z0 + z1 + z2 + z3) | 
|  | } | 
|  |  | 
|  | func BenchmarkMul(b *testing.B) { | 
|  | var hi, lo uint | 
|  | for i := 0; i < b.N; i++ { | 
|  | hi, lo = Mul(uint(Input), uint(i)) | 
|  | } | 
|  | Output = int(hi + lo) | 
|  | } | 
|  |  | 
|  | func BenchmarkMul32(b *testing.B) { | 
|  | var hi, lo uint32 | 
|  | for i := 0; i < b.N; i++ { | 
|  | hi, lo = Mul32(uint32(Input), uint32(i)) | 
|  | } | 
|  | Output = int(hi + lo) | 
|  | } | 
|  |  | 
|  | func BenchmarkMul64(b *testing.B) { | 
|  | var hi, lo uint64 | 
|  | for i := 0; i < b.N; i++ { | 
|  | hi, lo = Mul64(uint64(Input), uint64(i)) | 
|  | } | 
|  | Output = int(hi + lo) | 
|  | } | 
|  |  | 
|  | func BenchmarkDiv(b *testing.B) { | 
|  | var q, r uint | 
|  | for i := 0; i < b.N; i++ { | 
|  | q, r = Div(1, uint(i), uint(Input)) | 
|  | } | 
|  | Output = int(q + r) | 
|  | } | 
|  |  | 
|  | func BenchmarkDiv32(b *testing.B) { | 
|  | var q, r uint32 | 
|  | for i := 0; i < b.N; i++ { | 
|  | q, r = Div32(1, uint32(i), uint32(Input)) | 
|  | } | 
|  | Output = int(q + r) | 
|  | } | 
|  |  | 
|  | func BenchmarkDiv64(b *testing.B) { | 
|  | var q, r uint64 | 
|  | for i := 0; i < b.N; i++ { | 
|  | q, r = Div64(1, uint64(i), uint64(Input)) | 
|  | } | 
|  | Output = int(q + r) | 
|  | } | 
|  |  | 
|  | // ---------------------------------------------------------------------------- | 
|  | // Testing support | 
|  |  | 
|  | type entry = struct { | 
|  | nlz, ntz, pop int | 
|  | } | 
|  |  | 
|  | // tab contains results for all uint8 values | 
|  | var tab [256]entry | 
|  |  | 
|  | func init() { | 
|  | tab[0] = entry{8, 8, 0} | 
|  | for i := 1; i < len(tab); i++ { | 
|  | // nlz | 
|  | x := i // x != 0 | 
|  | n := 0 | 
|  | for x&0x80 == 0 { | 
|  | n++ | 
|  | x <<= 1 | 
|  | } | 
|  | tab[i].nlz = n | 
|  |  | 
|  | // ntz | 
|  | x = i // x != 0 | 
|  | n = 0 | 
|  | for x&1 == 0 { | 
|  | n++ | 
|  | x >>= 1 | 
|  | } | 
|  | tab[i].ntz = n | 
|  |  | 
|  | // pop | 
|  | x = i // x != 0 | 
|  | n = 0 | 
|  | for x != 0 { | 
|  | n += int(x & 1) | 
|  | x >>= 1 | 
|  | } | 
|  | tab[i].pop = n | 
|  | } | 
|  | } |