| // Copyright 2014 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| //go:build !plan9 |
| |
| #include "go_asm.h" |
| #include "textflag.h" |
| #include "asm_amd64.h" |
| |
| // See memclrNoHeapPointers Go doc for important implementation constraints. |
| |
| // func memclrNoHeapPointers(ptr unsafe.Pointer, n uintptr) |
| // ABIInternal for performance. |
| TEXT runtime·memclrNoHeapPointers<ABIInternal>(SB), NOSPLIT, $0-16 |
| // AX = ptr |
| // BX = n |
| MOVQ AX, DI // DI = ptr |
| XORQ AX, AX |
| |
| // MOVOU seems always faster than REP STOSQ when Enhanced REP STOSQ is not available. |
| tail: |
| // BSR+branch table make almost all memmove/memclr benchmarks worse. Not worth doing. |
| TESTQ BX, BX |
| JEQ _0 |
| CMPQ BX, $2 |
| JBE _1or2 |
| CMPQ BX, $4 |
| JBE _3or4 |
| CMPQ BX, $8 |
| JB _5through7 |
| JE _8 |
| CMPQ BX, $16 |
| JBE _9through16 |
| CMPQ BX, $32 |
| JBE _17through32 |
| CMPQ BX, $64 |
| JBE _33through64 |
| CMPQ BX, $128 |
| JBE _65through128 |
| CMPQ BX, $256 |
| JBE _129through256 |
| |
| CMPB internal∕cpu·X86+const_offsetX86HasERMS(SB), $1 // enhanced REP MOVSB/STOSB |
| JNE skip_erms |
| |
| // If the size is less than 2kb, do not use ERMS as it has a big start-up cost. |
| // Table 3-4. Relative Performance of Memcpy() Using ERMSB Vs. 128-bit AVX |
| // in the Intel Optimization Guide shows better performance for ERMSB starting |
| // from 2KB. Benchmarks show the similar threshold for REP STOS vs AVX. |
| CMPQ BX, $2048 |
| JAE loop_preheader_erms |
| |
| skip_erms: |
| #ifndef hasAVX2 |
| CMPB internal∕cpu·X86+const_offsetX86HasAVX2(SB), $1 |
| JE loop_preheader_avx2 |
| // TODO: for really big clears, use MOVNTDQ, even without AVX2. |
| |
| loop: |
| MOVOU X15, 0(DI) |
| MOVOU X15, 16(DI) |
| MOVOU X15, 32(DI) |
| MOVOU X15, 48(DI) |
| MOVOU X15, 64(DI) |
| MOVOU X15, 80(DI) |
| MOVOU X15, 96(DI) |
| MOVOU X15, 112(DI) |
| MOVOU X15, 128(DI) |
| MOVOU X15, 144(DI) |
| MOVOU X15, 160(DI) |
| MOVOU X15, 176(DI) |
| MOVOU X15, 192(DI) |
| MOVOU X15, 208(DI) |
| MOVOU X15, 224(DI) |
| MOVOU X15, 240(DI) |
| SUBQ $256, BX |
| ADDQ $256, DI |
| CMPQ BX, $256 |
| JAE loop |
| JMP tail |
| #endif |
| |
| loop_preheader_avx2: |
| VPXOR X0, X0, X0 |
| // For smaller sizes MOVNTDQ may be faster or slower depending on hardware. |
| // For larger sizes it is always faster, even on dual Xeons with 30M cache. |
| // TODO take into account actual LLC size. E. g. glibc uses LLC size/2. |
| CMPQ BX, $0x2000000 |
| JAE loop_preheader_avx2_huge |
| |
| loop_avx2: |
| VMOVDQU Y0, 0(DI) |
| VMOVDQU Y0, 32(DI) |
| VMOVDQU Y0, 64(DI) |
| VMOVDQU Y0, 96(DI) |
| SUBQ $128, BX |
| ADDQ $128, DI |
| CMPQ BX, $128 |
| JAE loop_avx2 |
| VMOVDQU Y0, -32(DI)(BX*1) |
| VMOVDQU Y0, -64(DI)(BX*1) |
| VMOVDQU Y0, -96(DI)(BX*1) |
| VMOVDQU Y0, -128(DI)(BX*1) |
| VZEROUPPER |
| RET |
| |
| loop_preheader_erms: |
| #ifndef hasAVX2 |
| CMPB internal∕cpu·X86+const_offsetX86HasAVX2(SB), $1 |
| JNE loop_erms |
| #endif |
| |
| VPXOR X0, X0, X0 |
| // At this point both ERMS and AVX2 is supported. While REP STOS can use a no-RFO |
| // write protocol, ERMS could show the same or slower performance comparing to |
| // Non-Temporal Stores when the size is bigger than LLC depending on hardware. |
| CMPQ BX, $0x2000000 |
| JAE loop_preheader_avx2_huge |
| |
| loop_erms: |
| // STOSQ is used to guarantee that the whole zeroed pointer-sized word is visible |
| // for a memory subsystem as the GC requires this. |
| MOVQ BX, CX |
| SHRQ $3, CX |
| ANDQ $7, BX |
| REP; STOSQ |
| JMP tail |
| |
| loop_preheader_avx2_huge: |
| // Align to 32 byte boundary |
| VMOVDQU Y0, 0(DI) |
| MOVQ DI, SI |
| ADDQ $32, DI |
| ANDQ $~31, DI |
| SUBQ DI, SI |
| ADDQ SI, BX |
| loop_avx2_huge: |
| VMOVNTDQ Y0, 0(DI) |
| VMOVNTDQ Y0, 32(DI) |
| VMOVNTDQ Y0, 64(DI) |
| VMOVNTDQ Y0, 96(DI) |
| SUBQ $128, BX |
| ADDQ $128, DI |
| CMPQ BX, $128 |
| JAE loop_avx2_huge |
| // In the description of MOVNTDQ in [1] |
| // "... fencing operation implemented with the SFENCE or MFENCE instruction |
| // should be used in conjunction with MOVNTDQ instructions..." |
| // [1] 64-ia-32-architectures-software-developer-manual-325462.pdf |
| SFENCE |
| VMOVDQU Y0, -32(DI)(BX*1) |
| VMOVDQU Y0, -64(DI)(BX*1) |
| VMOVDQU Y0, -96(DI)(BX*1) |
| VMOVDQU Y0, -128(DI)(BX*1) |
| VZEROUPPER |
| RET |
| |
| _1or2: |
| MOVB AX, (DI) |
| MOVB AX, -1(DI)(BX*1) |
| RET |
| _0: |
| RET |
| _3or4: |
| MOVW AX, (DI) |
| MOVW AX, -2(DI)(BX*1) |
| RET |
| _5through7: |
| MOVL AX, (DI) |
| MOVL AX, -4(DI)(BX*1) |
| RET |
| _8: |
| // We need a separate case for 8 to make sure we clear pointers atomically. |
| MOVQ AX, (DI) |
| RET |
| _9through16: |
| MOVQ AX, (DI) |
| MOVQ AX, -8(DI)(BX*1) |
| RET |
| _17through32: |
| MOVOU X15, (DI) |
| MOVOU X15, -16(DI)(BX*1) |
| RET |
| _33through64: |
| MOVOU X15, (DI) |
| MOVOU X15, 16(DI) |
| MOVOU X15, -32(DI)(BX*1) |
| MOVOU X15, -16(DI)(BX*1) |
| RET |
| _65through128: |
| MOVOU X15, (DI) |
| MOVOU X15, 16(DI) |
| MOVOU X15, 32(DI) |
| MOVOU X15, 48(DI) |
| MOVOU X15, -64(DI)(BX*1) |
| MOVOU X15, -48(DI)(BX*1) |
| MOVOU X15, -32(DI)(BX*1) |
| MOVOU X15, -16(DI)(BX*1) |
| RET |
| _129through256: |
| MOVOU X15, (DI) |
| MOVOU X15, 16(DI) |
| MOVOU X15, 32(DI) |
| MOVOU X15, 48(DI) |
| MOVOU X15, 64(DI) |
| MOVOU X15, 80(DI) |
| MOVOU X15, 96(DI) |
| MOVOU X15, 112(DI) |
| MOVOU X15, -128(DI)(BX*1) |
| MOVOU X15, -112(DI)(BX*1) |
| MOVOU X15, -96(DI)(BX*1) |
| MOVOU X15, -80(DI)(BX*1) |
| MOVOU X15, -64(DI)(BX*1) |
| MOVOU X15, -48(DI)(BX*1) |
| MOVOU X15, -32(DI)(BX*1) |
| MOVOU X15, -16(DI)(BX*1) |
| RET |