// Copyright 2013 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. package runtime_test import ( "crypto/rand" "encoding/binary" "fmt" "internal/race" "internal/testenv" . "runtime" "sync/atomic" "testing" "unsafe" ) func TestMemmove(t *testing.T) { if *flagQuick { t.Skip("-quick") } t.Parallel() size := 256 if testing.Short() { size = 128 + 16 } src := make([]byte, size) dst := make([]byte, size) for i := 0; i < size; i++ { src[i] = byte(128 + (i & 127)) } for i := 0; i < size; i++ { dst[i] = byte(i & 127) } for n := 0; n <= size; n++ { for x := 0; x <= size-n; x++ { // offset in src for y := 0; y <= size-n; y++ { // offset in dst copy(dst[y:y+n], src[x:x+n]) for i := 0; i < y; i++ { if dst[i] != byte(i&127) { t.Fatalf("prefix dst[%d] = %d", i, dst[i]) } } for i := y; i < y+n; i++ { if dst[i] != byte(128+((i-y+x)&127)) { t.Fatalf("copied dst[%d] = %d", i, dst[i]) } dst[i] = byte(i & 127) // reset dst } for i := y + n; i < size; i++ { if dst[i] != byte(i&127) { t.Fatalf("suffix dst[%d] = %d", i, dst[i]) } } } } } } func TestMemmoveAlias(t *testing.T) { if *flagQuick { t.Skip("-quick") } t.Parallel() size := 256 if testing.Short() { size = 128 + 16 } buf := make([]byte, size) for i := 0; i < size; i++ { buf[i] = byte(i) } for n := 0; n <= size; n++ { for x := 0; x <= size-n; x++ { // src offset for y := 0; y <= size-n; y++ { // dst offset copy(buf[y:y+n], buf[x:x+n]) for i := 0; i < y; i++ { if buf[i] != byte(i) { t.Fatalf("prefix buf[%d] = %d", i, buf[i]) } } for i := y; i < y+n; i++ { if buf[i] != byte(i-y+x) { t.Fatalf("copied buf[%d] = %d", i, buf[i]) } buf[i] = byte(i) // reset buf } for i := y + n; i < size; i++ { if buf[i] != byte(i) { t.Fatalf("suffix buf[%d] = %d", i, buf[i]) } } } } } } func TestMemmoveLarge0x180000(t *testing.T) { if testing.Short() && testenv.Builder() == "" { t.Skip("-short") } t.Parallel() if race.Enabled { t.Skip("skipping large memmove test under race detector") } testSize(t, 0x180000) } func TestMemmoveOverlapLarge0x120000(t *testing.T) { if testing.Short() && testenv.Builder() == "" { t.Skip("-short") } t.Parallel() if race.Enabled { t.Skip("skipping large memmove test under race detector") } testOverlap(t, 0x120000) } func testSize(t *testing.T, size int) { src := make([]byte, size) dst := make([]byte, size) _, _ = rand.Read(src) _, _ = rand.Read(dst) ref := make([]byte, size) copyref(ref, dst) for n := size - 50; n > 1; n >>= 1 { for x := 0; x <= size-n; x = x*7 + 1 { // offset in src for y := 0; y <= size-n; y = y*9 + 1 { // offset in dst copy(dst[y:y+n], src[x:x+n]) copyref(ref[y:y+n], src[x:x+n]) p := cmpb(dst, ref) if p >= 0 { t.Fatalf("Copy failed, copying from src[%d:%d] to dst[%d:%d].\nOffset %d is different, %v != %v", x, x+n, y, y+n, p, dst[p], ref[p]) } } } } } func testOverlap(t *testing.T, size int) { src := make([]byte, size) test := make([]byte, size) ref := make([]byte, size) _, _ = rand.Read(src) for n := size - 50; n > 1; n >>= 1 { for x := 0; x <= size-n; x = x*7 + 1 { // offset in src for y := 0; y <= size-n; y = y*9 + 1 { // offset in dst // Reset input copyref(test, src) copyref(ref, src) copy(test[y:y+n], test[x:x+n]) if y <= x { copyref(ref[y:y+n], ref[x:x+n]) } else { copybw(ref[y:y+n], ref[x:x+n]) } p := cmpb(test, ref) if p >= 0 { t.Fatalf("Copy failed, copying from src[%d:%d] to dst[%d:%d].\nOffset %d is different, %v != %v", x, x+n, y, y+n, p, test[p], ref[p]) } } } } } // Forward copy. func copyref(dst, src []byte) { for i, v := range src { dst[i] = v } } // Backwards copy func copybw(dst, src []byte) { if len(src) == 0 { return } for i := len(src) - 1; i >= 0; i-- { dst[i] = src[i] } } // Returns offset of difference func matchLen(a, b []byte, max int) int { a = a[:max] b = b[:max] for i, av := range a { if b[i] != av { return i } } return max } func cmpb(a, b []byte) int { l := matchLen(a, b, len(a)) if l == len(a) { return -1 } return l } // Ensure that memmove writes pointers atomically, so the GC won't // observe a partially updated pointer. func TestMemmoveAtomicity(t *testing.T) { if race.Enabled { t.Skip("skip under the race detector -- this test is intentionally racy") } var x int for _, backward := range []bool{true, false} { for _, n := range []int{3, 4, 5, 6, 7, 8, 9, 10, 15, 25, 49} { n := n // test copying [N]*int. sz := uintptr(n * PtrSize) name := fmt.Sprint(sz) if backward { name += "-backward" } else { name += "-forward" } t.Run(name, func(t *testing.T) { // Use overlapping src and dst to force forward/backward copy. var s [100]*int src := s[n-1 : 2*n-1] dst := s[:n] if backward { src, dst = dst, src } for i := range src { src[i] = &x } for i := range dst { dst[i] = nil } var ready uint32 go func() { sp := unsafe.Pointer(&src[0]) dp := unsafe.Pointer(&dst[0]) atomic.StoreUint32(&ready, 1) for i := 0; i < 10000; i++ { Memmove(dp, sp, sz) MemclrNoHeapPointers(dp, sz) } atomic.StoreUint32(&ready, 2) }() for atomic.LoadUint32(&ready) == 0 { Gosched() } for atomic.LoadUint32(&ready) != 2 { for i := range dst { p := dst[i] if p != nil && p != &x { t.Fatalf("got partially updated pointer %p at dst[%d], want either nil or %p", p, i, &x) } } } }) } } } func benchmarkSizes(b *testing.B, sizes []int, fn func(b *testing.B, n int)) { for _, n := range sizes { b.Run(fmt.Sprint(n), func(b *testing.B) { b.SetBytes(int64(n)) fn(b, n) }) } } var bufSizes = []int{ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, } var bufSizesOverlap = []int{ 32, 64, 128, 256, 512, 1024, 2048, 4096, } func BenchmarkMemmove(b *testing.B) { benchmarkSizes(b, bufSizes, func(b *testing.B, n int) { x := make([]byte, n) y := make([]byte, n) for i := 0; i < b.N; i++ { copy(x, y) } }) } func BenchmarkMemmoveOverlap(b *testing.B) { benchmarkSizes(b, bufSizesOverlap, func(b *testing.B, n int) { x := make([]byte, n+16) for i := 0; i < b.N; i++ { copy(x[16:n+16], x[:n]) } }) } func BenchmarkMemmoveUnalignedDst(b *testing.B) { benchmarkSizes(b, bufSizes, func(b *testing.B, n int) { x := make([]byte, n+1) y := make([]byte, n) for i := 0; i < b.N; i++ { copy(x[1:], y) } }) } func BenchmarkMemmoveUnalignedDstOverlap(b *testing.B) { benchmarkSizes(b, bufSizesOverlap, func(b *testing.B, n int) { x := make([]byte, n+16) for i := 0; i < b.N; i++ { copy(x[16:n+16], x[1:n+1]) } }) } func BenchmarkMemmoveUnalignedSrc(b *testing.B) { benchmarkSizes(b, bufSizes, func(b *testing.B, n int) { x := make([]byte, n) y := make([]byte, n+1) for i := 0; i < b.N; i++ { copy(x, y[1:]) } }) } func BenchmarkMemmoveUnalignedSrcOverlap(b *testing.B) { benchmarkSizes(b, bufSizesOverlap, func(b *testing.B, n int) { x := make([]byte, n+1) for i := 0; i < b.N; i++ { copy(x[1:n+1], x[:n]) } }) } func TestMemclr(t *testing.T) { size := 512 if testing.Short() { size = 128 + 16 } mem := make([]byte, size) for i := 0; i < size; i++ { mem[i] = 0xee } for n := 0; n < size; n++ { for x := 0; x <= size-n; x++ { // offset in mem MemclrBytes(mem[x : x+n]) for i := 0; i < x; i++ { if mem[i] != 0xee { t.Fatalf("overwrite prefix mem[%d] = %d", i, mem[i]) } } for i := x; i < x+n; i++ { if mem[i] != 0 { t.Fatalf("failed clear mem[%d] = %d", i, mem[i]) } mem[i] = 0xee } for i := x + n; i < size; i++ { if mem[i] != 0xee { t.Fatalf("overwrite suffix mem[%d] = %d", i, mem[i]) } } } } } func BenchmarkMemclr(b *testing.B) { for _, n := range []int{5, 16, 64, 256, 4096, 65536} { x := make([]byte, n) b.Run(fmt.Sprint(n), func(b *testing.B) { b.SetBytes(int64(n)) for i := 0; i < b.N; i++ { MemclrBytes(x) } }) } for _, m := range []int{1, 4, 8, 16, 64} { x := make([]byte, m<<20) b.Run(fmt.Sprint(m, "M"), func(b *testing.B) { b.SetBytes(int64(m << 20)) for i := 0; i < b.N; i++ { MemclrBytes(x) } }) } } func BenchmarkGoMemclr(b *testing.B) { benchmarkSizes(b, []int{5, 16, 64, 256}, func(b *testing.B, n int) { x := make([]byte, n) for i := 0; i < b.N; i++ { for j := range x { x[j] = 0 } } }) } func BenchmarkClearFat8(b *testing.B) { for i := 0; i < b.N; i++ { var x [8 / 4]uint32 _ = x } } func BenchmarkClearFat12(b *testing.B) { for i := 0; i < b.N; i++ { var x [12 / 4]uint32 _ = x } } func BenchmarkClearFat16(b *testing.B) { for i := 0; i < b.N; i++ { var x [16 / 4]uint32 _ = x } } func BenchmarkClearFat24(b *testing.B) { for i := 0; i < b.N; i++ { var x [24 / 4]uint32 _ = x } } func BenchmarkClearFat32(b *testing.B) { for i := 0; i < b.N; i++ { var x [32 / 4]uint32 _ = x } } func BenchmarkClearFat40(b *testing.B) { for i := 0; i < b.N; i++ { var x [40 / 4]uint32 _ = x } } func BenchmarkClearFat48(b *testing.B) { for i := 0; i < b.N; i++ { var x [48 / 4]uint32 _ = x } } func BenchmarkClearFat56(b *testing.B) { for i := 0; i < b.N; i++ { var x [56 / 4]uint32 _ = x } } func BenchmarkClearFat64(b *testing.B) { for i := 0; i < b.N; i++ { var x [64 / 4]uint32 _ = x } } func BenchmarkClearFat128(b *testing.B) { for i := 0; i < b.N; i++ { var x [128 / 4]uint32 _ = x } } func BenchmarkClearFat256(b *testing.B) { for i := 0; i < b.N; i++ { var x [256 / 4]uint32 _ = x } } func BenchmarkClearFat512(b *testing.B) { for i := 0; i < b.N; i++ { var x [512 / 4]uint32 _ = x } } func BenchmarkClearFat1024(b *testing.B) { for i := 0; i < b.N; i++ { var x [1024 / 4]uint32 _ = x } } func BenchmarkCopyFat8(b *testing.B) { var x [8 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat12(b *testing.B) { var x [12 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat16(b *testing.B) { var x [16 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat24(b *testing.B) { var x [24 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat32(b *testing.B) { var x [32 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat64(b *testing.B) { var x [64 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat128(b *testing.B) { var x [128 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat256(b *testing.B) { var x [256 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat512(b *testing.B) { var x [512 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat520(b *testing.B) { var x [520 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } func BenchmarkCopyFat1024(b *testing.B) { var x [1024 / 4]uint32 for i := 0; i < b.N; i++ { y := x _ = y } } // BenchmarkIssue18740 ensures that memmove uses 4 and 8 byte load/store to move 4 and 8 bytes. // It used to do 2 2-byte load/stores, which leads to a pipeline stall // when we try to read the result with one 4-byte load. func BenchmarkIssue18740(b *testing.B) { benchmarks := []struct { name string nbyte int f func([]byte) uint64 }{ {"2byte", 2, func(buf []byte) uint64 { return uint64(binary.LittleEndian.Uint16(buf)) }}, {"4byte", 4, func(buf []byte) uint64 { return uint64(binary.LittleEndian.Uint32(buf)) }}, {"8byte", 8, func(buf []byte) uint64 { return binary.LittleEndian.Uint64(buf) }}, } var g [4096]byte for _, bm := range benchmarks { buf := make([]byte, bm.nbyte) b.Run(bm.name, func(b *testing.B) { for j := 0; j < b.N; j++ { for i := 0; i < 4096; i += bm.nbyte { copy(buf[:], g[i:]) sink += bm.f(buf[:]) } } }) } }