| // Copyright 2019 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| //go:build libfuzzer |
| |
| #include "go_asm.h" |
| #include "go_tls.h" |
| #include "textflag.h" |
| |
| // Based on race_amd64.s; see commentary there. |
| |
| #ifdef GOOS_windows |
| #define RARG0 CX |
| #define RARG1 DX |
| #define RARG2 R8 |
| #define RARG3 R9 |
| #else |
| #define RARG0 DI |
| #define RARG1 SI |
| #define RARG2 DX |
| #define RARG3 CX |
| #endif |
| |
| // void runtime·libfuzzerCall4(fn, hookId int, s1, s2 unsafe.Pointer, result uintptr) |
| // Calls C function fn from libFuzzer and passes 4 arguments to it. |
| TEXT runtime·libfuzzerCall4(SB), NOSPLIT, $0-40 |
| MOVQ fn+0(FP), AX |
| MOVQ hookId+8(FP), RARG0 |
| MOVQ s1+16(FP), RARG1 |
| MOVQ s2+24(FP), RARG2 |
| MOVQ result+32(FP), RARG3 |
| |
| get_tls(R12) |
| MOVQ g(R12), R14 |
| MOVQ g_m(R14), R13 |
| |
| // Switch to g0 stack. |
| MOVQ SP, R12 // callee-saved, preserved across the CALL |
| MOVQ m_g0(R13), R10 |
| CMPQ R10, R14 |
| JE call // already on g0 |
| MOVQ (g_sched+gobuf_sp)(R10), SP |
| call: |
| ANDQ $~15, SP // alignment for gcc ABI |
| CALL AX |
| MOVQ R12, SP |
| RET |
| |
| // void runtime·libfuzzerCallTraceIntCmp(fn, arg0, arg1, fakePC uintptr) |
| // Calls C function fn from libFuzzer and passes 2 arguments to it after |
| // manipulating the return address so that libfuzzer's integer compare hooks |
| // work |
| // libFuzzer's compare hooks obtain the caller's address from the compiler |
| // builtin __builtin_return_adress. Since we invoke the hooks always |
| // from the same native function, this builtin would always return the same |
| // value. Internally, the libFuzzer hooks call through to the always inlined |
| // HandleCmp and thus can't be mimicked without patching libFuzzer. |
| // |
| // We solve this problem via an inline assembly trampoline construction that |
| // translates a runtime argument `fake_pc` in the range [0, 512) into a call to |
| // a hook with a fake return address whose lower 9 bits are `fake_pc` up to a |
| // constant shift. This is achieved by pushing a return address pointing into |
| // 512 ret instructions at offset `fake_pc` onto the stack and then jumping |
| // directly to the address of the hook. |
| // |
| // Note: We only set the lowest 9 bits of the return address since only these |
| // bits are used by the libFuzzer value profiling mode for integer compares, see |
| // https://github.com/llvm/llvm-project/blob/704d92607d26e696daba596b72cb70effe79a872/compiler-rt/lib/fuzzer/FuzzerTracePC.cpp#L390 |
| // as well as |
| // https://github.com/llvm/llvm-project/blob/704d92607d26e696daba596b72cb70effe79a872/compiler-rt/lib/fuzzer/FuzzerValueBitMap.h#L34 |
| // ValueProfileMap.AddValue() truncates its argument to 16 bits and shifts the |
| // PC to the left by log_2(128)=7, which means that only the lowest 16 - 7 bits |
| // of the return address matter. String compare hooks use the lowest 12 bits, |
| // but take the return address as an argument and thus don't require the |
| // indirection through a trampoline. |
| // TODO: Remove the inline assembly trampoline once a PC argument has been added to libfuzzer's int compare hooks. |
| TEXT runtime·libfuzzerCallTraceIntCmp(SB), NOSPLIT, $0-32 |
| MOVQ fn+0(FP), AX |
| MOVQ arg0+8(FP), RARG0 |
| MOVQ arg1+16(FP), RARG1 |
| MOVQ fakePC+24(FP), R8 |
| |
| get_tls(R12) |
| MOVQ g(R12), R14 |
| MOVQ g_m(R14), R13 |
| |
| // Switch to g0 stack. |
| MOVQ SP, R12 // callee-saved, preserved across the CALL |
| MOVQ m_g0(R13), R10 |
| CMPQ R10, R14 |
| JE call // already on g0 |
| MOVQ (g_sched+gobuf_sp)(R10), SP |
| call: |
| ANDQ $~15, SP // alignment for gcc ABI |
| // Load the address of the end of the function and push it into the stack. |
| // This address will be jumped to after executing the return instruction |
| // from the return sled. There we reset the stack pointer and return. |
| MOVQ $end_of_function<>(SB), BX |
| PUSHQ BX |
| // Load the starting address of the return sled into BX. |
| MOVQ $ret_sled<>(SB), BX |
| // Load the address of the i'th return instruction fron the return sled. |
| // The index is given in the fakePC argument. |
| ADDQ R8, BX |
| PUSHQ BX |
| // Call the original function with the fakePC return address on the stack. |
| // Function arguments arg0 and arg1 are passed in the registers specified |
| // by the x64 calling convention. |
| JMP AX |
| // This code will not be executed and is only there to statisfy assembler |
| // check of a balanced stack. |
| not_reachable: |
| POPQ BX |
| POPQ BX |
| RET |
| |
| TEXT end_of_function<>(SB), NOSPLIT, $0-0 |
| MOVQ R12, SP |
| RET |
| |
| #define REPEAT_8(a) a \ |
| a \ |
| a \ |
| a \ |
| a \ |
| a \ |
| a \ |
| a |
| |
| #define REPEAT_512(a) REPEAT_8(REPEAT_8(REPEAT_8(a))) |
| |
| TEXT ret_sled<>(SB), NOSPLIT, $0-0 |
| REPEAT_512(RET) |
| |
| // void runtime·libfuzzerCallWithTwoByteBuffers(fn, start, end *byte) |
| // Calls C function fn from libFuzzer and passes 2 arguments of type *byte to it. |
| TEXT runtime·libfuzzerCallWithTwoByteBuffers(SB), NOSPLIT, $0-24 |
| MOVQ fn+0(FP), AX |
| MOVQ start+8(FP), RARG0 |
| MOVQ end+16(FP), RARG1 |
| |
| get_tls(R12) |
| MOVQ g(R12), R14 |
| MOVQ g_m(R14), R13 |
| |
| // Switch to g0 stack. |
| MOVQ SP, R12 // callee-saved, preserved across the CALL |
| MOVQ m_g0(R13), R10 |
| CMPQ R10, R14 |
| JE call // already on g0 |
| MOVQ (g_sched+gobuf_sp)(R10), SP |
| call: |
| ANDQ $~15, SP // alignment for gcc ABI |
| CALL AX |
| MOVQ R12, SP |
| RET |