| // Copyright 2019 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| //go:build ignore |
| |
| // mkpreempt generates the asyncPreempt functions for each |
| // architecture. |
| package main |
| |
| import ( |
| "bytes" |
| "flag" |
| "fmt" |
| "go/format" |
| "io" |
| "log" |
| "os" |
| "strings" |
| ) |
| |
| // Copied from cmd/compile/internal/ssa/gen/*Ops.go |
| |
| var regNames386 = []string{ |
| "AX", |
| "CX", |
| "DX", |
| "BX", |
| "SP", |
| "BP", |
| "SI", |
| "DI", |
| "X0", |
| "X1", |
| "X2", |
| "X3", |
| "X4", |
| "X5", |
| "X6", |
| "X7", |
| } |
| |
| var regNamesAMD64 = []string{ |
| "AX", |
| "CX", |
| "DX", |
| "BX", |
| "SP", |
| "BP", |
| "SI", |
| "DI", |
| "R8", |
| "R9", |
| "R10", |
| "R11", |
| "R12", |
| "R13", |
| "R14", |
| "R15", |
| "X0", |
| "X1", |
| "X2", |
| "X3", |
| "X4", |
| "X5", |
| "X6", |
| "X7", |
| "X8", |
| "X9", |
| "X10", |
| "X11", |
| "X12", |
| "X13", |
| "X14", |
| "X15", |
| } |
| |
| var arches = map[string]func(g *gen){ |
| "386": gen386, |
| "amd64": genAMD64, |
| "arm": genARM, |
| "arm64": genARM64, |
| "loong64": genLoong64, |
| "mips64x": func(g *gen) { genMIPS(g, true) }, |
| "mipsx": func(g *gen) { genMIPS(g, false) }, |
| "ppc64x": genPPC64, |
| "riscv64": genRISCV64, |
| "s390x": genS390X, |
| "wasm": genWasm, |
| } |
| var beLe = map[string]bool{"mips64x": true, "mipsx": true, "ppc64x": true} |
| |
| func main() { |
| flag.Parse() |
| if flag.NArg() > 0 { |
| for _, arch := range flag.Args() { |
| genFn, ok := arches[arch] |
| if !ok { |
| log.Fatalf("unknown arch %s", arch) |
| } |
| g := gen{os.Stdout, arch} |
| g.asmHeader() |
| genFn(&g) |
| } |
| return |
| } |
| |
| for arch, genFn := range arches { |
| f, err := os.Create(fmt.Sprintf("preempt_%s.s", arch)) |
| if err != nil { |
| log.Fatal(err) |
| } |
| g := gen{f, arch} |
| g.asmHeader() |
| genFn(&g) |
| if err := f.Close(); err != nil { |
| log.Fatal(err) |
| } |
| } |
| } |
| |
| type gen struct { |
| w io.Writer |
| goarch string |
| } |
| |
| func (g *gen) commonHeader() { |
| fmt.Fprintf(g.w, "// Code generated by mkpreempt.go; DO NOT EDIT.\n\n") |
| if beLe[g.goarch] { |
| base := g.goarch[:len(g.goarch)-1] |
| fmt.Fprintf(g.w, "//go:build %s || %sle\n\n", base, base) |
| } |
| } |
| |
| func (g *gen) asmHeader() { |
| g.commonHeader() |
| fmt.Fprintf(g.w, "#include \"go_asm.h\"\n") |
| if g.goarch == "amd64" { |
| fmt.Fprintf(g.w, "#include \"go_tls.h\"\n") |
| fmt.Fprintf(g.w, "#include \"asm_amd64.h\"\n") |
| } |
| fmt.Fprintf(g.w, "#include \"textflag.h\"\n\n") |
| fmt.Fprintf(g.w, "TEXT ·asyncPreempt(SB),NOSPLIT|NOFRAME,$0-0\n") |
| } |
| |
| func (g *gen) p(f string, args ...any) { |
| fmted := fmt.Sprintf(f, args...) |
| fmt.Fprintf(g.w, "\t%s\n", strings.ReplaceAll(fmted, "\n", "\n\t")) |
| } |
| |
| func (g *gen) label(l string) { |
| fmt.Fprintf(g.w, "%s\n", l) |
| } |
| |
| // writeXRegs writes an architecture xregs file. |
| func writeXRegs(arch string, l *layout) { |
| var code bytes.Buffer |
| g := gen{&code, arch} |
| g.commonHeader() |
| fmt.Fprintf(g.w, ` |
| package runtime |
| |
| type xRegs struct { |
| `) |
| pos := 0 |
| for _, reg := range l.regs { |
| if reg.pos != pos { |
| log.Fatalf("padding not implemented") |
| } |
| typ := fmt.Sprintf("[%d]byte", reg.size) |
| switch { |
| case reg.size == 4 && reg.pos%4 == 0: |
| typ = "uint32" |
| case reg.size == 8 && reg.pos%8 == 0: |
| typ = "uint64" |
| } |
| fmt.Fprintf(g.w, "\t%s %s\n", reg.reg, typ) |
| pos += reg.size |
| } |
| fmt.Fprintf(g.w, "}\n") |
| |
| path := fmt.Sprintf("preempt_%s.go", arch) |
| b, err := format.Source(code.Bytes()) |
| if err != nil { |
| log.Fatalf("formatting %s: %s", path, err) |
| } |
| if err := os.WriteFile(path, b, 0666); err != nil { |
| log.Fatal(err) |
| } |
| } |
| |
| type layout struct { |
| stack int |
| regs []regPos |
| sp string // stack pointer register |
| } |
| |
| type regPos struct { |
| pos, size int |
| |
| saveOp string |
| restoreOp string |
| reg string |
| |
| // If this register requires special save and restore, these |
| // give those operations with a %d placeholder for the stack |
| // offset. |
| save, restore string |
| } |
| |
| func (l *layout) add(op, reg string, size int) { |
| l.regs = append(l.regs, regPos{saveOp: op, restoreOp: op, reg: reg, pos: l.stack, size: size}) |
| l.stack += size |
| } |
| |
| func (l *layout) add2(sop, rop, reg string, size int) { |
| l.regs = append(l.regs, regPos{saveOp: sop, restoreOp: rop, reg: reg, pos: l.stack, size: size}) |
| l.stack += size |
| } |
| |
| func (l *layout) addSpecial(save, restore string, size int) { |
| l.regs = append(l.regs, regPos{save: save, restore: restore, pos: l.stack, size: size}) |
| l.stack += size |
| } |
| |
| func (l *layout) save(g *gen) { |
| for _, reg := range l.regs { |
| if reg.save != "" { |
| g.p(reg.save, reg.pos) |
| } else { |
| g.p("%s %s, %d(%s)", reg.saveOp, reg.reg, reg.pos, l.sp) |
| } |
| } |
| } |
| |
| func (l *layout) restore(g *gen) { |
| for i := len(l.regs) - 1; i >= 0; i-- { |
| reg := l.regs[i] |
| if reg.restore != "" { |
| g.p(reg.restore, reg.pos) |
| } else { |
| g.p("%s %d(%s), %s", reg.restoreOp, reg.pos, l.sp, reg.reg) |
| } |
| } |
| } |
| |
| func gen386(g *gen) { |
| p := g.p |
| |
| p("PUSHFL") |
| // Save general purpose registers. |
| var l = layout{sp: "SP"} |
| for _, reg := range regNames386 { |
| if reg == "SP" || strings.HasPrefix(reg, "X") { |
| continue |
| } |
| l.add("MOVL", reg, 4) |
| } |
| |
| softfloat := "GO386_softfloat" |
| |
| // Save SSE state only if supported. |
| lSSE := layout{stack: l.stack, sp: "SP"} |
| for i := 0; i < 8; i++ { |
| lSSE.add("MOVUPS", fmt.Sprintf("X%d", i), 16) |
| } |
| |
| p("ADJSP $%d", lSSE.stack) |
| p("NOP SP") |
| l.save(g) |
| p("#ifndef %s", softfloat) |
| lSSE.save(g) |
| p("#endif") |
| p("CALL ·asyncPreempt2(SB)") |
| p("#ifndef %s", softfloat) |
| lSSE.restore(g) |
| p("#endif") |
| l.restore(g) |
| p("ADJSP $%d", -lSSE.stack) |
| |
| p("POPFL") |
| p("RET") |
| } |
| |
| func genAMD64(g *gen) { |
| const xReg = "AX" // *xRegState |
| |
| p, label := g.p, g.label |
| |
| // Assign stack offsets. |
| var l = layout{sp: "SP"} |
| for _, reg := range regNamesAMD64 { |
| if reg == "SP" || reg == "BP" { |
| continue |
| } |
| if !strings.HasPrefix(reg, "X") { |
| l.add("MOVQ", reg, 8) |
| } |
| } |
| // Create layouts for X, Y, and Z registers. |
| const ( |
| numXRegs = 16 |
| numZRegs = 16 // TODO: If we start using upper registers, change to 32 |
| numKRegs = 8 |
| ) |
| lZRegs := layout{sp: xReg} // Non-GP registers |
| lXRegs, lYRegs := lZRegs, lZRegs |
| for i := range numZRegs { |
| lZRegs.add("VMOVDQU64", fmt.Sprintf("Z%d", i), 512/8) |
| if i < numXRegs { |
| // Use SSE-only instructions for X registers. |
| lXRegs.add("MOVUPS", fmt.Sprintf("X%d", i), 128/8) |
| lYRegs.add("VMOVDQU", fmt.Sprintf("Y%d", i), 256/8) |
| } |
| } |
| for i := range numKRegs { |
| lZRegs.add("KMOVQ", fmt.Sprintf("K%d", i), 8) |
| } |
| // The Z layout is the most general, so we line up the others with that one. |
| // We don't have to do this, but it results in a nice Go type. If we split |
| // this into multiple types, we probably should stop doing this. |
| for i := range lXRegs.regs { |
| lXRegs.regs[i].pos = lZRegs.regs[i].pos |
| lYRegs.regs[i].pos = lZRegs.regs[i].pos |
| } |
| writeXRegs(g.goarch, &lZRegs) |
| |
| p("PUSHQ BP") |
| p("MOVQ SP, BP") |
| p("// Save flags before clobbering them") |
| p("PUSHFQ") |
| p("// obj doesn't understand ADD/SUB on SP, but does understand ADJSP") |
| p("ADJSP $%d", l.stack) |
| p("// But vet doesn't know ADJSP, so suppress vet stack checking") |
| p("NOP SP") |
| |
| p("// Save GPs") |
| l.save(g) |
| |
| // In general, the limitations on asynchronous preemption mean we only |
| // preempt in ABIInternal code. However, there's at least one exception to |
| // this: when we're in an open-coded transition between an ABIInternal |
| // function and an ABI0 call. We could more carefully arrange unsafe points |
| // to avoid ever landing in ABI0, but it's easy to just make this code not |
| // sensitive to the ABI we're preempting. The CALL to asyncPreempt2 will |
| // ensure we're in ABIInternal register state. |
| p("// Save extended register state to p.xRegs.scratch") |
| p("// Don't make assumptions about ABI register state. See mkpreempt.go") |
| p("get_tls(CX)") |
| p("MOVQ g(CX), R14") |
| p("MOVQ g_m(R14), %s", xReg) |
| p("MOVQ m_p(%s), %s", xReg, xReg) |
| p("LEAQ (p_xRegs+xRegPerP_scratch)(%s), %s", xReg, xReg) |
| |
| // Which registers do we need to save? |
| p("#ifdef GOEXPERIMENT_simd") |
| p("CMPB internal∕cpu·X86+const_offsetX86HasAVX512(SB), $1") |
| p("JE saveAVX512") |
| p("CMPB internal∕cpu·X86+const_offsetX86HasAVX2(SB), $1") |
| p("JE saveAVX2") |
| p("#endif") |
| |
| // No features. Assume only SSE. |
| label("saveSSE:") |
| lXRegs.save(g) |
| p("JMP preempt") |
| |
| label("saveAVX2:") |
| lYRegs.save(g) |
| p("JMP preempt") |
| |
| label("saveAVX512:") |
| lZRegs.save(g) |
| p("JMP preempt") |
| |
| label("preempt:") |
| p("CALL ·asyncPreempt2(SB)") |
| |
| p("// Restore non-GPs from *p.xRegs.cache") |
| p("MOVQ g_m(R14), %s", xReg) |
| p("MOVQ m_p(%s), %s", xReg, xReg) |
| p("MOVQ (p_xRegs+xRegPerP_cache)(%s), %s", xReg, xReg) |
| |
| p("#ifdef GOEXPERIMENT_simd") |
| p("CMPB internal∕cpu·X86+const_offsetX86HasAVX512(SB), $1") |
| p("JE restoreAVX512") |
| p("CMPB internal∕cpu·X86+const_offsetX86HasAVX2(SB), $1") |
| p("JE restoreAVX2") |
| p("#endif") |
| |
| label("restoreSSE:") |
| lXRegs.restore(g) |
| p("JMP restoreGPs") |
| |
| label("restoreAVX2:") |
| lYRegs.restore(g) |
| p("JMP restoreGPs") |
| |
| label("restoreAVX512:") |
| lZRegs.restore(g) |
| p("JMP restoreGPs") |
| |
| label("restoreGPs:") |
| p("// Restore GPs") |
| l.restore(g) |
| p("ADJSP $%d", -l.stack) |
| p("POPFQ") |
| p("POPQ BP") |
| p("RET") |
| } |
| |
| func genARM(g *gen) { |
| p := g.p |
| |
| // Add integer registers R0-R12. |
| // R13 (SP), R14 (LR), R15 (PC) are special and not saved here. |
| var l = layout{sp: "R13", stack: 4} // add LR slot |
| for i := 0; i <= 12; i++ { |
| reg := fmt.Sprintf("R%d", i) |
| if i == 10 { |
| continue // R10 is g register, no need to save/restore |
| } |
| l.add("MOVW", reg, 4) |
| } |
| // Add flag register. |
| l.addSpecial( |
| "MOVW CPSR, R0\nMOVW R0, %d(R13)", |
| "MOVW %d(R13), R0\nMOVW R0, CPSR", |
| 4) |
| |
| // Add floating point registers F0-F15 and flag register. |
| var lfp = layout{stack: l.stack, sp: "R13"} |
| lfp.addSpecial( |
| "MOVW FPCR, R0\nMOVW R0, %d(R13)", |
| "MOVW %d(R13), R0\nMOVW R0, FPCR", |
| 4) |
| for i := 0; i <= 15; i++ { |
| reg := fmt.Sprintf("F%d", i) |
| lfp.add("MOVD", reg, 8) |
| } |
| |
| p("MOVW.W R14, -%d(R13)", lfp.stack) // allocate frame, save LR |
| l.save(g) |
| p("MOVB ·goarmsoftfp(SB), R0\nCMP $0, R0\nBNE nofp") // test goarmsoftfp, and skip FP registers if goarmsoftfp!=0. |
| lfp.save(g) |
| g.label("nofp:") |
| p("CALL ·asyncPreempt2(SB)") |
| p("MOVB ·goarmsoftfp(SB), R0\nCMP $0, R0\nBNE nofp2") // test goarmsoftfp, and skip FP registers if goarmsoftfp!=0. |
| lfp.restore(g) |
| g.label("nofp2:") |
| l.restore(g) |
| |
| p("MOVW %d(R13), R14", lfp.stack) // sigctxt.pushCall pushes LR on stack, restore it |
| p("MOVW.P %d(R13), R15", lfp.stack+4) // load PC, pop frame (including the space pushed by sigctxt.pushCall) |
| p("UNDEF") // shouldn't get here |
| } |
| |
| func genARM64(g *gen) { |
| p := g.p |
| // Add integer registers R0-R26 |
| // R27 (REGTMP), R28 (g), R29 (FP), R30 (LR), R31 (SP) are special |
| // and not saved here. |
| var l = layout{sp: "RSP", stack: 8} // add slot to save PC of interrupted instruction |
| for i := 0; i < 26; i += 2 { |
| if i == 18 { |
| i-- |
| continue // R18 is not used, skip |
| } |
| reg := fmt.Sprintf("(R%d, R%d)", i, i+1) |
| l.add2("STP", "LDP", reg, 16) |
| } |
| // Add flag registers. |
| l.addSpecial( |
| "MOVD NZCV, R0\nMOVD R0, %d(RSP)", |
| "MOVD %d(RSP), R0\nMOVD R0, NZCV", |
| 8) |
| l.addSpecial( |
| "MOVD FPSR, R0\nMOVD R0, %d(RSP)", |
| "MOVD %d(RSP), R0\nMOVD R0, FPSR", |
| 8) |
| // TODO: FPCR? I don't think we'll change it, so no need to save. |
| // Add floating point registers F0-F31. |
| for i := 0; i < 31; i += 2 { |
| reg := fmt.Sprintf("(F%d, F%d)", i, i+1) |
| l.add2("FSTPD", "FLDPD", reg, 16) |
| } |
| if l.stack%16 != 0 { |
| l.stack += 8 // SP needs 16-byte alignment |
| } |
| |
| // allocate frame, save PC of interrupted instruction (in LR) |
| p("MOVD R30, %d(RSP)", -l.stack) |
| p("SUB $%d, RSP", l.stack) |
| p("MOVD R29, -8(RSP)") // save frame pointer (only used on Linux) |
| p("SUB $8, RSP, R29") // set up new frame pointer |
| // On iOS, save the LR again after decrementing SP. We run the |
| // signal handler on the G stack (as it doesn't support sigaltstack), |
| // so any writes below SP may be clobbered. |
| p("#ifdef GOOS_ios") |
| p("MOVD R30, (RSP)") |
| p("#endif") |
| |
| l.save(g) |
| p("CALL ·asyncPreempt2(SB)") |
| l.restore(g) |
| |
| p("MOVD %d(RSP), R30", l.stack) // sigctxt.pushCall has pushed LR (at interrupt) on stack, restore it |
| p("MOVD -8(RSP), R29") // restore frame pointer |
| p("MOVD (RSP), R27") // load PC to REGTMP |
| p("ADD $%d, RSP", l.stack+16) // pop frame (including the space pushed by sigctxt.pushCall) |
| p("RET (R27)") |
| } |
| |
| func genMIPS(g *gen, _64bit bool) { |
| p := g.p |
| |
| mov := "MOVW" |
| movf := "MOVF" |
| add := "ADD" |
| sub := "SUB" |
| r28 := "R28" |
| regsize := 4 |
| softfloat := "GOMIPS_softfloat" |
| if _64bit { |
| mov = "MOVV" |
| movf = "MOVD" |
| add = "ADDV" |
| sub = "SUBV" |
| r28 = "RSB" |
| regsize = 8 |
| softfloat = "GOMIPS64_softfloat" |
| } |
| |
| // Add integer registers R1-R22, R24-R25, R28 |
| // R0 (zero), R23 (REGTMP), R29 (SP), R30 (g), R31 (LR) are special, |
| // and not saved here. R26 and R27 are reserved by kernel and not used. |
| var l = layout{sp: "R29", stack: regsize} // add slot to save PC of interrupted instruction (in LR) |
| for i := 1; i <= 25; i++ { |
| if i == 23 { |
| continue // R23 is REGTMP |
| } |
| reg := fmt.Sprintf("R%d", i) |
| l.add(mov, reg, regsize) |
| } |
| l.add(mov, r28, regsize) |
| l.addSpecial( |
| mov+" HI, R1\n"+mov+" R1, %d(R29)", |
| mov+" %d(R29), R1\n"+mov+" R1, HI", |
| regsize) |
| l.addSpecial( |
| mov+" LO, R1\n"+mov+" R1, %d(R29)", |
| mov+" %d(R29), R1\n"+mov+" R1, LO", |
| regsize) |
| |
| // Add floating point control/status register FCR31 (FCR0-FCR30 are irrelevant) |
| var lfp = layout{sp: "R29", stack: l.stack} |
| lfp.addSpecial( |
| mov+" FCR31, R1\n"+mov+" R1, %d(R29)", |
| mov+" %d(R29), R1\n"+mov+" R1, FCR31", |
| regsize) |
| // Add floating point registers F0-F31. |
| for i := 0; i <= 31; i++ { |
| reg := fmt.Sprintf("F%d", i) |
| lfp.add(movf, reg, regsize) |
| } |
| |
| // allocate frame, save PC of interrupted instruction (in LR) |
| p(mov+" R31, -%d(R29)", lfp.stack) |
| p(sub+" $%d, R29", lfp.stack) |
| |
| l.save(g) |
| p("#ifndef %s", softfloat) |
| lfp.save(g) |
| p("#endif") |
| p("CALL ·asyncPreempt2(SB)") |
| p("#ifndef %s", softfloat) |
| lfp.restore(g) |
| p("#endif") |
| l.restore(g) |
| |
| p(mov+" %d(R29), R31", lfp.stack) // sigctxt.pushCall has pushed LR (at interrupt) on stack, restore it |
| p(mov + " (R29), R23") // load PC to REGTMP |
| p(add+" $%d, R29", lfp.stack+regsize) // pop frame (including the space pushed by sigctxt.pushCall) |
| p("JMP (R23)") |
| } |
| |
| func genLoong64(g *gen) { |
| p := g.p |
| |
| mov := "MOVV" |
| movf := "MOVD" |
| add := "ADDV" |
| sub := "SUBV" |
| regsize := 8 |
| |
| // Add integer registers r4-r21 r23-r29 r31 |
| // R0 (zero), R30 (REGTMP), R2 (tp), R3 (SP), R22 (g), R1 (LR) are special, |
| var l = layout{sp: "R3", stack: regsize} // add slot to save PC of interrupted instruction (in LR) |
| for i := 4; i <= 31; i++ { |
| if i == 22 || i == 30 { |
| continue |
| } |
| reg := fmt.Sprintf("R%d", i) |
| l.add(mov, reg, regsize) |
| } |
| |
| // Add floating point registers F0-F31. |
| for i := 0; i <= 31; i++ { |
| reg := fmt.Sprintf("F%d", i) |
| l.add(movf, reg, regsize) |
| } |
| |
| // Add condition flag register fcc0-fcc7 |
| sv := "" |
| rs := "" |
| last := 7 |
| for i := 0; i <= last; i++ { |
| msb := 7 + (i * 8) |
| lsb := 0 + (i * 8) |
| |
| // MOVV FCCx, R4, |
| // BSTRINSV $msb, R4, $lsb, R5 |
| sv += fmt.Sprintf("%s FCC%d, R4\n", mov, i) |
| sv += fmt.Sprintf("BSTRINSV $%d, R4, $%d, R5\n", msb, lsb) |
| |
| // BSTRPICKV $msb, R5, $lsb, R4 |
| // MOVV R4, FCCx |
| rs += fmt.Sprintf("BSTRPICKV $%d, R5, $%d, R4\n", msb, lsb) |
| rs += fmt.Sprintf("%s R4, FCC%d", mov, i) |
| if i != last { |
| rs += fmt.Sprintf("\n") |
| } |
| } |
| l.addSpecial( |
| sv+mov+" R5, %d(R3)", |
| mov+" %d(R3), R5\n"+rs, |
| regsize) |
| |
| // allocate frame, save PC of interrupted instruction (in LR) |
| p(mov+" R1, -%d(R3)", l.stack) |
| p(sub+" $%d, R3", l.stack) |
| |
| l.save(g) |
| p("CALL ·asyncPreempt2(SB)") |
| l.restore(g) |
| |
| p(mov+" %d(R3), R1", l.stack) // sigctxt.pushCall has pushed LR (at interrupt) on stack, restore it |
| p(mov + " (R3), R30") // load PC to REGTMP |
| p(add+" $%d, R3", l.stack+regsize) // pop frame (including the space pushed by sigctxt.pushCall) |
| p("JMP (R30)") |
| } |
| |
| func genPPC64(g *gen) { |
| p := g.p |
| |
| // Add integer registers R3-R29 |
| // R0 (zero), R1 (SP), R30 (g) are special and not saved here. |
| // R2 (TOC pointer in PIC mode), R12 (function entry address in PIC mode) have been saved in sigctxt.pushCall. |
| // R31 (REGTMP) will be saved manually. |
| var l = layout{sp: "R1", stack: 32 + 8} // MinFrameSize on PPC64, plus one word for saving R31 |
| for i := 3; i <= 29; i++ { |
| if i == 12 || i == 13 { |
| // R12 has been saved in sigctxt.pushCall. |
| // R13 is TLS pointer, not used by Go code. we must NOT |
| // restore it, otherwise if we parked and resumed on a |
| // different thread we'll mess up TLS addresses. |
| continue |
| } |
| reg := fmt.Sprintf("R%d", i) |
| l.add("MOVD", reg, 8) |
| } |
| l.addSpecial( |
| "MOVW CR, R31\nMOVW R31, %d(R1)", |
| "MOVW %d(R1), R31\nMOVFL R31, $0xff", // this is MOVW R31, CR |
| 8) // CR is 4-byte wide, but just keep the alignment |
| l.addSpecial( |
| "MOVD XER, R31\nMOVD R31, %d(R1)", |
| "MOVD %d(R1), R31\nMOVD R31, XER", |
| 8) |
| // Add floating point registers F0-F31. |
| for i := 0; i <= 31; i++ { |
| reg := fmt.Sprintf("F%d", i) |
| l.add("FMOVD", reg, 8) |
| } |
| // Add floating point control/status register FPSCR. |
| l.addSpecial( |
| "MOVFL FPSCR, F0\nFMOVD F0, %d(R1)", |
| "FMOVD %d(R1), F0\nMOVFL F0, FPSCR", |
| 8) |
| |
| p("MOVD R31, -%d(R1)", l.stack-32) // save R31 first, we'll use R31 for saving LR |
| p("MOVD LR, R31") |
| p("MOVDU R31, -%d(R1)", l.stack) // allocate frame, save PC of interrupted instruction (in LR) |
| |
| l.save(g) |
| p("CALL ·asyncPreempt2(SB)") |
| l.restore(g) |
| |
| p("MOVD %d(R1), R31", l.stack) // sigctxt.pushCall has pushed LR, R2, R12 (at interrupt) on stack, restore them |
| p("MOVD R31, LR") |
| p("MOVD %d(R1), R2", l.stack+8) |
| p("MOVD %d(R1), R12", l.stack+16) |
| p("MOVD (R1), R31") // load PC to CTR |
| p("MOVD R31, CTR") |
| p("MOVD 32(R1), R31") // restore R31 |
| p("ADD $%d, R1", l.stack+32) // pop frame (including the space pushed by sigctxt.pushCall) |
| p("JMP (CTR)") |
| } |
| |
| func genRISCV64(g *gen) { |
| p := g.p |
| |
| // X0 (zero), X1 (LR), X2 (SP), X3 (GP), X4 (TP), X27 (g), X31 (TMP) are special. |
| var l = layout{sp: "X2", stack: 8} |
| |
| // Add integer registers (X5-X26, X28-30). |
| for i := 5; i < 31; i++ { |
| if i == 27 { |
| continue |
| } |
| reg := fmt.Sprintf("X%d", i) |
| l.add("MOV", reg, 8) |
| } |
| |
| // Add floating point registers (F0-F31). |
| for i := 0; i <= 31; i++ { |
| reg := fmt.Sprintf("F%d", i) |
| l.add("MOVD", reg, 8) |
| } |
| |
| p("MOV X1, -%d(X2)", l.stack) |
| p("SUB $%d, X2", l.stack) |
| l.save(g) |
| p("CALL ·asyncPreempt2(SB)") |
| l.restore(g) |
| p("MOV %d(X2), X1", l.stack) |
| p("MOV (X2), X31") |
| p("ADD $%d, X2", l.stack+8) |
| p("JMP (X31)") |
| } |
| |
| func genS390X(g *gen) { |
| p := g.p |
| |
| // Add integer registers R0-R12 |
| // R13 (g), R14 (LR), R15 (SP) are special, and not saved here. |
| // Saving R10 (REGTMP) is not necessary, but it is saved anyway. |
| var l = layout{sp: "R15", stack: 16} // add slot to save PC of interrupted instruction and flags |
| l.addSpecial( |
| "STMG R0, R12, %d(R15)", |
| "LMG %d(R15), R0, R12", |
| 13*8) |
| // Add floating point registers F0-F31. |
| for i := 0; i <= 15; i++ { |
| reg := fmt.Sprintf("F%d", i) |
| l.add("FMOVD", reg, 8) |
| } |
| |
| // allocate frame, save PC of interrupted instruction (in LR) and flags (condition code) |
| p("IPM R10") // save flags upfront, as ADD will clobber flags |
| p("MOVD R14, -%d(R15)", l.stack) |
| p("ADD $-%d, R15", l.stack) |
| p("MOVW R10, 8(R15)") // save flags |
| |
| l.save(g) |
| p("CALL ·asyncPreempt2(SB)") |
| l.restore(g) |
| |
| p("MOVD %d(R15), R14", l.stack) // sigctxt.pushCall has pushed LR (at interrupt) on stack, restore it |
| p("ADD $%d, R15", l.stack+8) // pop frame (including the space pushed by sigctxt.pushCall) |
| p("MOVWZ -%d(R15), R10", l.stack) // load flags to REGTMP |
| p("TMLH R10, $(3<<12)") // restore flags |
| p("MOVD -%d(R15), R10", l.stack+8) // load PC to REGTMP |
| p("JMP (R10)") |
| } |
| |
| func genWasm(g *gen) { |
| p := g.p |
| p("// No async preemption on wasm") |
| p("UNDEF") |
| } |
| |
| func notImplemented(g *gen) { |
| p := g.p |
| p("// Not implemented yet") |
| p("JMP ·abort(SB)") |
| } |