cmd/compile: make vet happy with ssa code
Fixes #15488
Change-Id: I054eb1e1c859de315e3cdbdef5428682bce693fd
Reviewed-on: https://go-review.googlesource.com/22609
Run-TryBot: David Chase <drchase@google.com>
TryBot-Result: Gobot Gobot <gobot@golang.org>
Reviewed-by: David Chase <drchase@google.com>
diff --git a/src/cmd/compile/internal/amd64/ssa.go b/src/cmd/compile/internal/amd64/ssa.go
index 6557287..1001e14 100644
--- a/src/cmd/compile/internal/amd64/ssa.go
+++ b/src/cmd/compile/internal/amd64/ssa.go
@@ -141,7 +141,6 @@
panic(fmt.Sprintf("bad int register width %d:%s", t.Size(), t))
}
}
- panic("bad register type")
}
// opregreg emits instructions for
@@ -966,12 +965,12 @@
}
var eqfJumps = [2][2]gc.FloatingEQNEJump{
- {{x86.AJNE, 1}, {x86.AJPS, 1}}, // next == b.Succs[0]
- {{x86.AJNE, 1}, {x86.AJPC, 0}}, // next == b.Succs[1]
+ {{Jump: x86.AJNE, Index: 1}, {Jump: x86.AJPS, Index: 1}}, // next == b.Succs[0]
+ {{Jump: x86.AJNE, Index: 1}, {Jump: x86.AJPC, Index: 0}}, // next == b.Succs[1]
}
var nefJumps = [2][2]gc.FloatingEQNEJump{
- {{x86.AJNE, 0}, {x86.AJPC, 1}}, // next == b.Succs[0]
- {{x86.AJNE, 0}, {x86.AJPS, 0}}, // next == b.Succs[1]
+ {{Jump: x86.AJNE, Index: 0}, {Jump: x86.AJPC, Index: 1}}, // next == b.Succs[0]
+ {{Jump: x86.AJNE, Index: 0}, {Jump: x86.AJPS, Index: 0}}, // next == b.Succs[1]
}
func ssaGenBlock(s *gc.SSAGenState, b, next *ssa.Block) {
@@ -982,7 +981,7 @@
if b.Succs[0] != next {
p := gc.Prog(obj.AJMP)
p.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{p, b.Succs[0]})
+ s.Branches = append(s.Branches, gc.Branch{P: p, B: b.Succs[0]})
}
case ssa.BlockDefer:
// defer returns in rax:
@@ -995,11 +994,11 @@
p.To.Reg = x86.REG_AX
p = gc.Prog(x86.AJNE)
p.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{p, b.Succs[1]})
+ s.Branches = append(s.Branches, gc.Branch{P: p, B: b.Succs[1]})
if b.Succs[0] != next {
p := gc.Prog(obj.AJMP)
p.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{p, b.Succs[0]})
+ s.Branches = append(s.Branches, gc.Branch{P: p, B: b.Succs[0]})
}
case ssa.BlockExit:
gc.Prog(obj.AUNDEF) // tell plive.go that we never reach here
@@ -1030,18 +1029,18 @@
p = gc.Prog(jmp.invasm)
likely *= -1
p.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{p, b.Succs[1]})
+ s.Branches = append(s.Branches, gc.Branch{P: p, B: b.Succs[1]})
case b.Succs[1]:
p = gc.Prog(jmp.asm)
p.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{p, b.Succs[0]})
+ s.Branches = append(s.Branches, gc.Branch{P: p, B: b.Succs[0]})
default:
p = gc.Prog(jmp.asm)
p.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{p, b.Succs[0]})
+ s.Branches = append(s.Branches, gc.Branch{P: p, B: b.Succs[0]})
q := gc.Prog(obj.AJMP)
q.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{q, b.Succs[1]})
+ s.Branches = append(s.Branches, gc.Branch{P: q, B: b.Succs[1]})
}
// liblink reorders the instruction stream as it sees fit.
diff --git a/src/cmd/compile/internal/arm/ssa.go b/src/cmd/compile/internal/arm/ssa.go
index ca10f1c..665f8ff 100644
--- a/src/cmd/compile/internal/arm/ssa.go
+++ b/src/cmd/compile/internal/arm/ssa.go
@@ -139,16 +139,16 @@
if b.Succs[0] != next {
p := gc.Prog(obj.AJMP)
p.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{p, b.Succs[0]})
+ s.Branches = append(s.Branches, gc.Branch{P: p, B: b.Succs[0]})
}
case ssa.BlockRet:
gc.Prog(obj.ARET)
case ssa.BlockARMLT:
p := gc.Prog(arm.ABLT)
p.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{p, b.Succs[0]})
+ s.Branches = append(s.Branches, gc.Branch{P: p, B: b.Succs[0]})
p = gc.Prog(obj.AJMP)
p.To.Type = obj.TYPE_BRANCH
- s.Branches = append(s.Branches, gc.Branch{p, b.Succs[1]})
+ s.Branches = append(s.Branches, gc.Branch{P: p, B: b.Succs[1]})
}
}
diff --git a/src/cmd/compile/internal/gc/plive.go b/src/cmd/compile/internal/gc/plive.go
index a15a467..87f4a11 100644
--- a/src/cmd/compile/internal/gc/plive.go
+++ b/src/cmd/compile/internal/gc/plive.go
@@ -1693,7 +1693,7 @@
ls := Linksym(sym)
ls.Name = fmt.Sprintf("gclocals·%x", md5.Sum(ls.P))
ls.Dupok = true
- sv := obj.SymVer{ls.Name, 0}
+ sv := obj.SymVer{Name: ls.Name, Version: 0}
ls2, ok := Ctxt.Hash[sv]
if ok {
sym.Lsym = ls2
diff --git a/src/cmd/compile/internal/gc/ssa.go b/src/cmd/compile/internal/gc/ssa.go
index 61cb16d..9d03f23 100644
--- a/src/cmd/compile/internal/gc/ssa.go
+++ b/src/cmd/compile/internal/gc/ssa.go
@@ -1424,7 +1424,7 @@
s.stmtList(n.Ninit)
switch n.Op {
case OCFUNC:
- aux := s.lookupSymbol(n, &ssa.ExternSymbol{n.Type, n.Left.Sym})
+ aux := s.lookupSymbol(n, &ssa.ExternSymbol{Typ: n.Type, Sym: n.Left.Sym})
return s.entryNewValue1A(ssa.OpAddr, n.Type, aux, s.sb)
case OPARAM:
addr := s.addr(n, false)
@@ -1433,7 +1433,7 @@
if n.Class == PFUNC {
// "value" of a function is the address of the function's closure
sym := funcsym(n.Sym)
- aux := &ssa.ExternSymbol{n.Type, sym}
+ aux := &ssa.ExternSymbol{Typ: n.Type, Sym: sym}
return s.entryNewValue1A(ssa.OpAddr, Ptrto(n.Type), aux, s.sb)
}
if s.canSSA(n) {
@@ -2163,7 +2163,7 @@
// Call growslice
s.startBlock(grow)
- taddr := s.newValue1A(ssa.OpAddr, Types[TUINTPTR], &ssa.ExternSymbol{Types[TUINTPTR], typenamesym(n.Type.Elem())}, s.sb)
+ taddr := s.newValue1A(ssa.OpAddr, Types[TUINTPTR], &ssa.ExternSymbol{Typ: Types[TUINTPTR], Sym: typenamesym(n.Type.Elem())}, s.sb)
r := s.rtcall(growslice, true, []*Type{pt, Types[TINT], Types[TINT]}, taddr, p, l, c, nl)
@@ -2691,7 +2691,7 @@
switch n.Class {
case PEXTERN:
// global variable
- aux := s.lookupSymbol(n, &ssa.ExternSymbol{n.Type, n.Sym})
+ aux := s.lookupSymbol(n, &ssa.ExternSymbol{Typ: n.Type, Sym: n.Sym})
v := s.entryNewValue1A(ssa.OpAddr, t, aux, s.sb)
// TODO: Make OpAddr use AuxInt as well as Aux.
if n.Xoffset != 0 {
@@ -3024,7 +3024,7 @@
bElse := s.f.NewBlock(ssa.BlockPlain)
bEnd := s.f.NewBlock(ssa.BlockPlain)
- aux := &ssa.ExternSymbol{Types[TBOOL], syslook("writeBarrier").Sym}
+ aux := &ssa.ExternSymbol{Typ: Types[TBOOL], Sym: syslook("writeBarrier").Sym}
flagaddr := s.newValue1A(ssa.OpAddr, Ptrto(Types[TUINT32]), aux, s.sb)
// TODO: select the .enabled field. It is currently first, so not needed for now.
// Load word, test byte, avoiding partial register write from load byte.
@@ -3038,7 +3038,7 @@
b.AddEdgeTo(bElse)
s.startBlock(bThen)
- taddr := s.newValue1A(ssa.OpAddr, Types[TUINTPTR], &ssa.ExternSymbol{Types[TUINTPTR], typenamesym(t)}, s.sb)
+ taddr := s.newValue1A(ssa.OpAddr, Types[TUINTPTR], &ssa.ExternSymbol{Typ: Types[TUINTPTR], Sym: typenamesym(t)}, s.sb)
s.rtcall(typedmemmove, true, nil, taddr, left, right)
s.endBlock().AddEdgeTo(bEnd)
@@ -3075,7 +3075,7 @@
bElse := s.f.NewBlock(ssa.BlockPlain)
bEnd := s.f.NewBlock(ssa.BlockPlain)
- aux := &ssa.ExternSymbol{Types[TBOOL], syslook("writeBarrier").Sym}
+ aux := &ssa.ExternSymbol{Typ: Types[TBOOL], Sym: syslook("writeBarrier").Sym}
flagaddr := s.newValue1A(ssa.OpAddr, Ptrto(Types[TUINT32]), aux, s.sb)
// TODO: select the .enabled field. It is currently first, so not needed for now.
// Load word, test byte, avoiding partial register write from load byte.
@@ -3629,7 +3629,7 @@
if !commaok {
// on failure, panic by calling panicdottype
s.startBlock(bFail)
- taddr := s.newValue1A(ssa.OpAddr, byteptr, &ssa.ExternSymbol{byteptr, typenamesym(n.Left.Type)}, s.sb)
+ taddr := s.newValue1A(ssa.OpAddr, byteptr, &ssa.ExternSymbol{Typ: byteptr, Sym: typenamesym(n.Left.Type)}, s.sb)
s.rtcall(panicdottype, false, nil, typ, target, taddr)
// on success, return idata field
@@ -3995,7 +3995,7 @@
if f.StaticData != nil {
for _, n := range f.StaticData.([]*Node) {
if !gen_as_init(n, false) {
- Fatalf("non-static data marked as static: %v\n\n", n, f)
+ Fatalf("non-static data marked as static: %v\n\n", n)
}
}
}
@@ -4084,7 +4084,7 @@
}
func AddAux2(a *obj.Addr, v *ssa.Value, offset int64) {
if a.Type != obj.TYPE_MEM {
- v.Fatalf("bad AddAux addr %s", a)
+ v.Fatalf("bad AddAux addr %v", a)
}
// add integer offset
a.Offset += offset
@@ -4263,10 +4263,10 @@
// Split this string up into two separate variables.
p := e.namedAuto(n.Sym.Name+".ptr", ptrType)
l := e.namedAuto(n.Sym.Name+".len", lenType)
- return ssa.LocalSlot{p, ptrType, 0}, ssa.LocalSlot{l, lenType, 0}
+ return ssa.LocalSlot{N: p, Type: ptrType, Off: 0}, ssa.LocalSlot{N: l, Type: lenType, Off: 0}
}
// Return the two parts of the larger variable.
- return ssa.LocalSlot{n, ptrType, name.Off}, ssa.LocalSlot{n, lenType, name.Off + int64(Widthptr)}
+ return ssa.LocalSlot{N: n, Type: ptrType, Off: name.Off}, ssa.LocalSlot{N: n, Type: lenType, Off: name.Off + int64(Widthptr)}
}
func (e *ssaExport) SplitInterface(name ssa.LocalSlot) (ssa.LocalSlot, ssa.LocalSlot) {
@@ -4280,10 +4280,10 @@
}
c := e.namedAuto(n.Sym.Name+f, t)
d := e.namedAuto(n.Sym.Name+".data", t)
- return ssa.LocalSlot{c, t, 0}, ssa.LocalSlot{d, t, 0}
+ return ssa.LocalSlot{N: c, Type: t, Off: 0}, ssa.LocalSlot{N: d, Type: t, Off: 0}
}
// Return the two parts of the larger variable.
- return ssa.LocalSlot{n, t, name.Off}, ssa.LocalSlot{n, t, name.Off + int64(Widthptr)}
+ return ssa.LocalSlot{N: n, Type: t, Off: name.Off}, ssa.LocalSlot{N: n, Type: t, Off: name.Off + int64(Widthptr)}
}
func (e *ssaExport) SplitSlice(name ssa.LocalSlot) (ssa.LocalSlot, ssa.LocalSlot, ssa.LocalSlot) {
@@ -4295,12 +4295,12 @@
p := e.namedAuto(n.Sym.Name+".ptr", ptrType)
l := e.namedAuto(n.Sym.Name+".len", lenType)
c := e.namedAuto(n.Sym.Name+".cap", lenType)
- return ssa.LocalSlot{p, ptrType, 0}, ssa.LocalSlot{l, lenType, 0}, ssa.LocalSlot{c, lenType, 0}
+ return ssa.LocalSlot{N: p, Type: ptrType, Off: 0}, ssa.LocalSlot{N: l, Type: lenType, Off: 0}, ssa.LocalSlot{N: c, Type: lenType, Off: 0}
}
// Return the three parts of the larger variable.
- return ssa.LocalSlot{n, ptrType, name.Off},
- ssa.LocalSlot{n, lenType, name.Off + int64(Widthptr)},
- ssa.LocalSlot{n, lenType, name.Off + int64(2*Widthptr)}
+ return ssa.LocalSlot{N: n, Type: ptrType, Off: name.Off},
+ ssa.LocalSlot{N: n, Type: lenType, Off: name.Off + int64(Widthptr)},
+ ssa.LocalSlot{N: n, Type: lenType, Off: name.Off + int64(2*Widthptr)}
}
func (e *ssaExport) SplitComplex(name ssa.LocalSlot) (ssa.LocalSlot, ssa.LocalSlot) {
@@ -4316,10 +4316,10 @@
// Split this complex up into two separate variables.
c := e.namedAuto(n.Sym.Name+".real", t)
d := e.namedAuto(n.Sym.Name+".imag", t)
- return ssa.LocalSlot{c, t, 0}, ssa.LocalSlot{d, t, 0}
+ return ssa.LocalSlot{N: c, Type: t, Off: 0}, ssa.LocalSlot{N: d, Type: t, Off: 0}
}
// Return the two parts of the larger variable.
- return ssa.LocalSlot{n, t, name.Off}, ssa.LocalSlot{n, t, name.Off + s}
+ return ssa.LocalSlot{N: n, Type: t, Off: name.Off}, ssa.LocalSlot{N: n, Type: t, Off: name.Off + s}
}
func (e *ssaExport) SplitStruct(name ssa.LocalSlot, i int) ssa.LocalSlot {
@@ -4331,9 +4331,9 @@
// have no fear, identically-named but distinct Autos are
// ok, albeit maybe confusing for a debugger.
x := e.namedAuto(n.Sym.Name+"."+st.FieldName(i), ft)
- return ssa.LocalSlot{x, ft, 0}
+ return ssa.LocalSlot{N: x, Type: ft, Off: 0}
}
- return ssa.LocalSlot{n, ft, name.Off + st.FieldOff(i)}
+ return ssa.LocalSlot{N: n, Type: ft, Off: name.Off + st.FieldOff(i)}
}
// namedAuto returns a new AUTO variable with the given name and type.
diff --git a/src/cmd/compile/internal/ssa/check.go b/src/cmd/compile/internal/ssa/check.go
index af123d8..457bf63 100644
--- a/src/cmd/compile/internal/ssa/check.go
+++ b/src/cmd/compile/internal/ssa/check.go
@@ -67,7 +67,7 @@
f.Fatalf("ret block %s has successors", b)
}
if b.Control == nil {
- f.Fatalf("ret block %s has nil control %s", b)
+ f.Fatalf("ret block %s has nil control", b)
}
if !b.Control.Type.IsMemory() {
f.Fatalf("ret block %s has non-memory control value %s", b, b.Control.LongString())
@@ -77,7 +77,7 @@
f.Fatalf("retjmp block %s len(Succs)==%d, want 0", b, len(b.Succs))
}
if b.Control == nil {
- f.Fatalf("retjmp block %s has nil control %s", b)
+ f.Fatalf("retjmp block %s has nil control", b)
}
if !b.Control.Type.IsMemory() {
f.Fatalf("retjmp block %s has non-memory control value %s", b, b.Control.LongString())
@@ -141,7 +141,7 @@
}
}
if len(b.Succs) > 2 && b.Likely != BranchUnknown {
- f.Fatalf("likeliness prediction %d for block %s with %d successors: %s", b.Likely, b, len(b.Succs))
+ f.Fatalf("likeliness prediction %d for block %s with %d successors", b.Likely, b, len(b.Succs))
}
for _, v := range b.Values {
diff --git a/src/cmd/compile/internal/ssa/gen/AMD64.rules b/src/cmd/compile/internal/ssa/gen/AMD64.rules
index 31d8f12..db274d7 100644
--- a/src/cmd/compile/internal/ssa/gen/AMD64.rules
+++ b/src/cmd/compile/internal/ssa/gen/AMD64.rules
@@ -1196,9 +1196,7 @@
(ADDQconst [c] (ADDQconst [d] x)) && is32Bit(c+d) -> (ADDQconst [c+d] x)
(ADDLconst [c] (ADDLconst [d] x)) -> (ADDLconst [int64(int32(c+d))] x)
(SUBQconst (MOVQconst [d]) [c]) -> (MOVQconst [d-c])
-(SUBLconst (MOVLconst [d]) [c]) -> (MOVLconst [int64(int32(d-c))])
(SUBQconst (SUBQconst x [d]) [c]) && is32Bit(-c-d) -> (ADDQconst [-c-d] x)
-(SUBLconst (SUBLconst x [d]) [c]) -> (ADDLconst [int64(int32(-c-d))] x)
(SARQconst [c] (MOVQconst [d])) -> (MOVQconst [d>>uint64(c)])
(SARLconst [c] (MOVQconst [d])) -> (MOVQconst [d>>uint64(c)])
(SARWconst [c] (MOVQconst [d])) -> (MOVQconst [d>>uint64(c)])
diff --git a/src/cmd/compile/internal/ssa/gen/rulegen.go b/src/cmd/compile/internal/ssa/gen/rulegen.go
index 5f7d1cf..8dd5032 100644
--- a/src/cmd/compile/internal/ssa/gen/rulegen.go
+++ b/src/cmd/compile/internal/ssa/gen/rulegen.go
@@ -173,17 +173,25 @@
fmt.Fprintf(w, "func rewriteValue%s_%s(v *Value, config *Config) bool {\n", arch.name, opName(op, arch))
fmt.Fprintln(w, "b := v.Block")
fmt.Fprintln(w, "_ = b")
- for _, rule := range oprules[op] {
+ var canFail bool
+ for i, rule := range oprules[op] {
match, cond, result := rule.parse()
fmt.Fprintf(w, "// match: %s\n", match)
fmt.Fprintf(w, "// cond: %s\n", cond)
fmt.Fprintf(w, "// result: %s\n", result)
+ canFail = false
fmt.Fprintf(w, "for {\n")
- genMatch(w, arch, match, rule.loc)
+ if genMatch(w, arch, match, rule.loc) {
+ canFail = true
+ }
if cond != "" {
fmt.Fprintf(w, "if !(%s) {\nbreak\n}\n", cond)
+ canFail = true
+ }
+ if !canFail && i != len(oprules[op])-1 {
+ log.Fatalf("unconditional rule %s is followed by other rules", match)
}
genResult(w, arch, result, rule.loc)
@@ -194,7 +202,9 @@
fmt.Fprintf(w, "}\n")
}
- fmt.Fprintf(w, "return false\n")
+ if canFail {
+ fmt.Fprintf(w, "return false\n")
+ }
fmt.Fprintf(w, "}\n")
}
@@ -321,14 +331,16 @@
}
}
-func genMatch(w io.Writer, arch arch, match string, loc string) {
- genMatch0(w, arch, match, "v", map[string]struct{}{}, true, loc)
+// genMatch returns true if the match can fail.
+func genMatch(w io.Writer, arch arch, match string, loc string) bool {
+ return genMatch0(w, arch, match, "v", map[string]struct{}{}, true, loc)
}
-func genMatch0(w io.Writer, arch arch, match, v string, m map[string]struct{}, top bool, loc string) {
+func genMatch0(w io.Writer, arch arch, match, v string, m map[string]struct{}, top bool, loc string) bool {
if match[0] != '(' || match[len(match)-1] != ')' {
panic("non-compound expr in genMatch0: " + match)
}
+ canFail := false
// split body up into regions. Split by spaces/tabs, except those
// contained in () or {}.
@@ -355,6 +367,7 @@
// check op
if !top {
fmt.Fprintf(w, "if %s.Op != %s {\nbreak\n}\n", v, opName(s[0], arch))
+ canFail = true
}
// check type/aux/args
@@ -366,11 +379,13 @@
if !isVariable(t) {
// code. We must match the results of this code.
fmt.Fprintf(w, "if %s.Type != %s {\nbreak\n}\n", v, t)
+ canFail = true
} else {
// variable
if _, ok := m[t]; ok {
// must match previous variable
fmt.Fprintf(w, "if %s.Type != %s {\nbreak\n}\n", v, t)
+ canFail = true
} else {
m[t] = struct{}{}
fmt.Fprintf(w, "%s := %s.Type\n", t, v)
@@ -387,10 +402,12 @@
if !isVariable(x) {
// code
fmt.Fprintf(w, "if %s.AuxInt != %s {\nbreak\n}\n", v, x)
+ canFail = true
} else {
// variable
if _, ok := m[x]; ok {
fmt.Fprintf(w, "if %s.AuxInt != %s {\nbreak\n}\n", v, x)
+ canFail = true
} else {
m[x] = struct{}{}
fmt.Fprintf(w, "%s := %s.AuxInt\n", x, v)
@@ -407,10 +424,12 @@
if !isVariable(x) {
// code
fmt.Fprintf(w, "if %s.Aux != %s {\nbreak\n}\n", v, x)
+ canFail = true
} else {
// variable
if _, ok := m[x]; ok {
fmt.Fprintf(w, "if %s.Aux != %s {\nbreak\n}\n", v, x)
+ canFail = true
} else {
m[x] = struct{}{}
fmt.Fprintf(w, "%s := %s.Aux\n", x, v)
@@ -426,6 +445,7 @@
// For example, (add x x). Equality is just pointer equality
// on Values (so cse is important to do before lowering).
fmt.Fprintf(w, "if %s != %s.Args[%d] {\nbreak\n}\n", a, v, argnum)
+ canFail = true
} else {
// remember that this variable references the given value
m[a] = struct{}{}
@@ -446,15 +466,19 @@
argname = fmt.Sprintf("%s_%d", v, argnum)
}
fmt.Fprintf(w, "%s := %s.Args[%d]\n", argname, v, argnum)
- genMatch0(w, arch, a, argname, m, false, loc)
+ if genMatch0(w, arch, a, argname, m, false, loc) {
+ canFail = true
+ }
argnum++
}
}
if op.argLength == -1 {
fmt.Fprintf(w, "if len(%s.Args) != %d {\nbreak\n}\n", v, argnum)
+ canFail = true
} else if int(op.argLength) != argnum {
log.Fatalf("%s: op %s should have %d args, has %d", loc, op.name, op.argLength, argnum)
}
+ return canFail
}
func genResult(w io.Writer, arch arch, result string, loc string) {
diff --git a/src/cmd/compile/internal/ssa/rewrite.go b/src/cmd/compile/internal/ssa/rewrite.go
index c127e85..852be5e 100644
--- a/src/cmd/compile/internal/ssa/rewrite.go
+++ b/src/cmd/compile/internal/ssa/rewrite.go
@@ -142,7 +142,6 @@
return x
}
panic(fmt.Sprintf("mergeSym with two non-nil syms %s %s", x, y))
- return nil
}
func canMergeSym(x, y interface{}) bool {
return x == nil || y == nil
diff --git a/src/cmd/compile/internal/ssa/rewriteAMD64.go b/src/cmd/compile/internal/ssa/rewriteAMD64.go
index 99639e5b..60ac235 100644
--- a/src/cmd/compile/internal/ssa/rewriteAMD64.go
+++ b/src/cmd/compile/internal/ssa/rewriteAMD64.go
@@ -1539,7 +1539,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAdd32(v *Value, config *Config) bool {
b := v.Block
@@ -1555,7 +1554,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAdd32F(v *Value, config *Config) bool {
b := v.Block
@@ -1571,7 +1569,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAdd64(v *Value, config *Config) bool {
b := v.Block
@@ -1587,7 +1584,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAdd64F(v *Value, config *Config) bool {
b := v.Block
@@ -1603,7 +1599,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAdd8(v *Value, config *Config) bool {
b := v.Block
@@ -1619,7 +1614,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAddPtr(v *Value, config *Config) bool {
b := v.Block
@@ -1635,7 +1629,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAddr(v *Value, config *Config) bool {
b := v.Block
@@ -1651,7 +1644,6 @@
v.AddArg(base)
return true
}
- return false
}
func rewriteValueAMD64_OpAnd16(v *Value, config *Config) bool {
b := v.Block
@@ -1667,7 +1659,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAnd32(v *Value, config *Config) bool {
b := v.Block
@@ -1683,7 +1674,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAnd64(v *Value, config *Config) bool {
b := v.Block
@@ -1699,7 +1689,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAnd8(v *Value, config *Config) bool {
b := v.Block
@@ -1715,7 +1704,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAndB(v *Value, config *Config) bool {
b := v.Block
@@ -1731,7 +1719,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAvg64u(v *Value, config *Config) bool {
b := v.Block
@@ -1747,7 +1734,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpBswap32(v *Value, config *Config) bool {
b := v.Block
@@ -1761,7 +1747,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpBswap64(v *Value, config *Config) bool {
b := v.Block
@@ -1775,7 +1760,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpAMD64CMOVLEQconst(v *Value, config *Config) bool {
b := v.Block
@@ -2899,7 +2883,6 @@
v.AddArg(mem)
return true
}
- return false
}
func rewriteValueAMD64_OpCom16(v *Value, config *Config) bool {
b := v.Block
@@ -2913,7 +2896,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCom32(v *Value, config *Config) bool {
b := v.Block
@@ -2927,7 +2909,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCom64(v *Value, config *Config) bool {
b := v.Block
@@ -2941,7 +2922,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCom8(v *Value, config *Config) bool {
b := v.Block
@@ -2955,7 +2935,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpConst16(v *Value, config *Config) bool {
b := v.Block
@@ -2969,7 +2948,6 @@
v.AuxInt = val
return true
}
- return false
}
func rewriteValueAMD64_OpConst32(v *Value, config *Config) bool {
b := v.Block
@@ -2983,7 +2961,6 @@
v.AuxInt = val
return true
}
- return false
}
func rewriteValueAMD64_OpConst32F(v *Value, config *Config) bool {
b := v.Block
@@ -2997,7 +2974,6 @@
v.AuxInt = val
return true
}
- return false
}
func rewriteValueAMD64_OpConst64(v *Value, config *Config) bool {
b := v.Block
@@ -3011,7 +2987,6 @@
v.AuxInt = val
return true
}
- return false
}
func rewriteValueAMD64_OpConst64F(v *Value, config *Config) bool {
b := v.Block
@@ -3025,7 +3000,6 @@
v.AuxInt = val
return true
}
- return false
}
func rewriteValueAMD64_OpConst8(v *Value, config *Config) bool {
b := v.Block
@@ -3039,7 +3013,6 @@
v.AuxInt = val
return true
}
- return false
}
func rewriteValueAMD64_OpConstBool(v *Value, config *Config) bool {
b := v.Block
@@ -3053,7 +3026,6 @@
v.AuxInt = b
return true
}
- return false
}
func rewriteValueAMD64_OpConstNil(v *Value, config *Config) bool {
b := v.Block
@@ -3066,7 +3038,6 @@
v.AuxInt = 0
return true
}
- return false
}
func rewriteValueAMD64_OpConvert(v *Value, config *Config) bool {
b := v.Block
@@ -3084,7 +3055,6 @@
v.AddArg(mem)
return true
}
- return false
}
func rewriteValueAMD64_OpCtz16(v *Value, config *Config) bool {
b := v.Block
@@ -3106,7 +3076,6 @@
v.AuxInt = 16
return true
}
- return false
}
func rewriteValueAMD64_OpCtz32(v *Value, config *Config) bool {
b := v.Block
@@ -3128,7 +3097,6 @@
v.AuxInt = 32
return true
}
- return false
}
func rewriteValueAMD64_OpCtz64(v *Value, config *Config) bool {
b := v.Block
@@ -3150,7 +3118,6 @@
v.AuxInt = 64
return true
}
- return false
}
func rewriteValueAMD64_OpCvt32Fto32(v *Value, config *Config) bool {
b := v.Block
@@ -3164,7 +3131,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCvt32Fto64(v *Value, config *Config) bool {
b := v.Block
@@ -3178,7 +3144,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCvt32Fto64F(v *Value, config *Config) bool {
b := v.Block
@@ -3192,7 +3157,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCvt32to32F(v *Value, config *Config) bool {
b := v.Block
@@ -3206,7 +3170,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCvt32to64F(v *Value, config *Config) bool {
b := v.Block
@@ -3220,7 +3183,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCvt64Fto32(v *Value, config *Config) bool {
b := v.Block
@@ -3234,7 +3196,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCvt64Fto32F(v *Value, config *Config) bool {
b := v.Block
@@ -3248,7 +3209,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCvt64Fto64(v *Value, config *Config) bool {
b := v.Block
@@ -3262,7 +3222,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCvt64to32F(v *Value, config *Config) bool {
b := v.Block
@@ -3276,7 +3235,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpCvt64to64F(v *Value, config *Config) bool {
b := v.Block
@@ -3290,7 +3248,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpDeferCall(v *Value, config *Config) bool {
b := v.Block
@@ -3306,7 +3263,6 @@
v.AddArg(mem)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv16(v *Value, config *Config) bool {
b := v.Block
@@ -3322,7 +3278,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv16u(v *Value, config *Config) bool {
b := v.Block
@@ -3338,7 +3293,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv32(v *Value, config *Config) bool {
b := v.Block
@@ -3354,7 +3308,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv32F(v *Value, config *Config) bool {
b := v.Block
@@ -3370,7 +3323,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv32u(v *Value, config *Config) bool {
b := v.Block
@@ -3386,7 +3338,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv64(v *Value, config *Config) bool {
b := v.Block
@@ -3402,7 +3353,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv64F(v *Value, config *Config) bool {
b := v.Block
@@ -3418,7 +3368,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv64u(v *Value, config *Config) bool {
b := v.Block
@@ -3434,7 +3383,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv8(v *Value, config *Config) bool {
b := v.Block
@@ -3454,7 +3402,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpDiv8u(v *Value, config *Config) bool {
b := v.Block
@@ -3474,7 +3421,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpEq16(v *Value, config *Config) bool {
b := v.Block
@@ -3492,7 +3438,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpEq32(v *Value, config *Config) bool {
b := v.Block
@@ -3510,7 +3455,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpEq32F(v *Value, config *Config) bool {
b := v.Block
@@ -3528,7 +3472,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpEq64(v *Value, config *Config) bool {
b := v.Block
@@ -3546,7 +3489,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpEq64F(v *Value, config *Config) bool {
b := v.Block
@@ -3564,7 +3506,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpEq8(v *Value, config *Config) bool {
b := v.Block
@@ -3582,7 +3523,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpEqB(v *Value, config *Config) bool {
b := v.Block
@@ -3600,7 +3540,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpEqPtr(v *Value, config *Config) bool {
b := v.Block
@@ -3618,7 +3557,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq16(v *Value, config *Config) bool {
b := v.Block
@@ -3636,7 +3574,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq16U(v *Value, config *Config) bool {
b := v.Block
@@ -3654,7 +3591,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq32(v *Value, config *Config) bool {
b := v.Block
@@ -3672,7 +3608,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq32F(v *Value, config *Config) bool {
b := v.Block
@@ -3690,7 +3625,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq32U(v *Value, config *Config) bool {
b := v.Block
@@ -3708,7 +3642,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq64(v *Value, config *Config) bool {
b := v.Block
@@ -3726,7 +3659,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq64F(v *Value, config *Config) bool {
b := v.Block
@@ -3744,7 +3676,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq64U(v *Value, config *Config) bool {
b := v.Block
@@ -3762,7 +3693,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq8(v *Value, config *Config) bool {
b := v.Block
@@ -3780,7 +3710,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGeq8U(v *Value, config *Config) bool {
b := v.Block
@@ -3798,7 +3727,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGetClosurePtr(v *Value, config *Config) bool {
b := v.Block
@@ -3810,7 +3738,6 @@
v.reset(OpAMD64LoweredGetClosurePtr)
return true
}
- return false
}
func rewriteValueAMD64_OpGetG(v *Value, config *Config) bool {
b := v.Block
@@ -3824,7 +3751,6 @@
v.AddArg(mem)
return true
}
- return false
}
func rewriteValueAMD64_OpGoCall(v *Value, config *Config) bool {
b := v.Block
@@ -3840,7 +3766,6 @@
v.AddArg(mem)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater16(v *Value, config *Config) bool {
b := v.Block
@@ -3858,7 +3783,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater16U(v *Value, config *Config) bool {
b := v.Block
@@ -3876,7 +3800,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater32(v *Value, config *Config) bool {
b := v.Block
@@ -3894,7 +3817,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater32F(v *Value, config *Config) bool {
b := v.Block
@@ -3912,7 +3834,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater32U(v *Value, config *Config) bool {
b := v.Block
@@ -3930,7 +3851,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater64(v *Value, config *Config) bool {
b := v.Block
@@ -3948,7 +3868,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater64F(v *Value, config *Config) bool {
b := v.Block
@@ -3966,7 +3885,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater64U(v *Value, config *Config) bool {
b := v.Block
@@ -3984,7 +3902,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater8(v *Value, config *Config) bool {
b := v.Block
@@ -4002,7 +3919,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpGreater8U(v *Value, config *Config) bool {
b := v.Block
@@ -4020,7 +3936,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpHmul16(v *Value, config *Config) bool {
b := v.Block
@@ -4036,7 +3951,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpHmul16u(v *Value, config *Config) bool {
b := v.Block
@@ -4052,7 +3966,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpHmul32(v *Value, config *Config) bool {
b := v.Block
@@ -4068,7 +3981,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpHmul32u(v *Value, config *Config) bool {
b := v.Block
@@ -4084,7 +3996,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpHmul64(v *Value, config *Config) bool {
b := v.Block
@@ -4100,7 +4011,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpHmul64u(v *Value, config *Config) bool {
b := v.Block
@@ -4116,7 +4026,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpHmul8(v *Value, config *Config) bool {
b := v.Block
@@ -4132,7 +4041,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpHmul8u(v *Value, config *Config) bool {
b := v.Block
@@ -4148,7 +4056,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpITab(v *Value, config *Config) bool {
b := v.Block
@@ -4186,7 +4093,6 @@
v.AddArg(mem)
return true
}
- return false
}
func rewriteValueAMD64_OpIsInBounds(v *Value, config *Config) bool {
b := v.Block
@@ -4204,7 +4110,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpIsNonNil(v *Value, config *Config) bool {
b := v.Block
@@ -4221,7 +4126,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpIsSliceInBounds(v *Value, config *Config) bool {
b := v.Block
@@ -4239,7 +4143,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpAMD64LEAQ(v *Value, config *Config) bool {
b := v.Block
@@ -4945,7 +4848,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLeq16U(v *Value, config *Config) bool {
b := v.Block
@@ -4963,7 +4865,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLeq32(v *Value, config *Config) bool {
b := v.Block
@@ -4981,7 +4882,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLeq32F(v *Value, config *Config) bool {
b := v.Block
@@ -4999,7 +4899,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLeq32U(v *Value, config *Config) bool {
b := v.Block
@@ -5017,7 +4916,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLeq64(v *Value, config *Config) bool {
b := v.Block
@@ -5035,7 +4933,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLeq64F(v *Value, config *Config) bool {
b := v.Block
@@ -5053,7 +4950,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLeq64U(v *Value, config *Config) bool {
b := v.Block
@@ -5071,7 +4967,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLeq8(v *Value, config *Config) bool {
b := v.Block
@@ -5089,7 +4984,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLeq8U(v *Value, config *Config) bool {
b := v.Block
@@ -5107,7 +5001,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess16(v *Value, config *Config) bool {
b := v.Block
@@ -5125,7 +5018,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess16U(v *Value, config *Config) bool {
b := v.Block
@@ -5143,7 +5035,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess32(v *Value, config *Config) bool {
b := v.Block
@@ -5161,7 +5052,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess32F(v *Value, config *Config) bool {
b := v.Block
@@ -5179,7 +5069,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess32U(v *Value, config *Config) bool {
b := v.Block
@@ -5197,7 +5086,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess64(v *Value, config *Config) bool {
b := v.Block
@@ -5215,7 +5103,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess64F(v *Value, config *Config) bool {
b := v.Block
@@ -5233,7 +5120,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess64U(v *Value, config *Config) bool {
b := v.Block
@@ -5251,7 +5137,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess8(v *Value, config *Config) bool {
b := v.Block
@@ -5269,7 +5154,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLess8U(v *Value, config *Config) bool {
b := v.Block
@@ -5287,7 +5171,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpLoad(v *Value, config *Config) bool {
b := v.Block
@@ -5400,7 +5283,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpLrot32(v *Value, config *Config) bool {
b := v.Block
@@ -5418,7 +5300,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpLrot64(v *Value, config *Config) bool {
b := v.Block
@@ -5436,7 +5317,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpLrot8(v *Value, config *Config) bool {
b := v.Block
@@ -5454,7 +5334,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh16x16(v *Value, config *Config) bool {
b := v.Block
@@ -5479,7 +5358,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh16x32(v *Value, config *Config) bool {
b := v.Block
@@ -5504,7 +5382,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh16x64(v *Value, config *Config) bool {
b := v.Block
@@ -5529,7 +5406,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh16x8(v *Value, config *Config) bool {
b := v.Block
@@ -5554,7 +5430,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh32x16(v *Value, config *Config) bool {
b := v.Block
@@ -5579,7 +5454,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh32x32(v *Value, config *Config) bool {
b := v.Block
@@ -5604,7 +5478,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh32x64(v *Value, config *Config) bool {
b := v.Block
@@ -5629,7 +5502,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh32x8(v *Value, config *Config) bool {
b := v.Block
@@ -5654,7 +5526,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh64x16(v *Value, config *Config) bool {
b := v.Block
@@ -5679,7 +5550,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh64x32(v *Value, config *Config) bool {
b := v.Block
@@ -5704,7 +5574,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh64x64(v *Value, config *Config) bool {
b := v.Block
@@ -5729,7 +5598,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh64x8(v *Value, config *Config) bool {
b := v.Block
@@ -5754,7 +5622,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh8x16(v *Value, config *Config) bool {
b := v.Block
@@ -5779,7 +5646,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh8x32(v *Value, config *Config) bool {
b := v.Block
@@ -5804,7 +5670,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh8x64(v *Value, config *Config) bool {
b := v.Block
@@ -5829,7 +5694,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpLsh8x8(v *Value, config *Config) bool {
b := v.Block
@@ -5854,7 +5718,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpAMD64MOVBQSX(v *Value, config *Config) bool {
b := v.Block
@@ -12020,7 +11883,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMod16u(v *Value, config *Config) bool {
b := v.Block
@@ -12036,7 +11898,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMod32(v *Value, config *Config) bool {
b := v.Block
@@ -12052,7 +11913,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMod32u(v *Value, config *Config) bool {
b := v.Block
@@ -12068,7 +11928,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMod64(v *Value, config *Config) bool {
b := v.Block
@@ -12084,7 +11943,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMod64u(v *Value, config *Config) bool {
b := v.Block
@@ -12100,7 +11958,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMod8(v *Value, config *Config) bool {
b := v.Block
@@ -12120,7 +11977,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpMod8u(v *Value, config *Config) bool {
b := v.Block
@@ -12140,7 +11996,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpMove(v *Value, config *Config) bool {
b := v.Block
@@ -12510,7 +12365,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMul32(v *Value, config *Config) bool {
b := v.Block
@@ -12526,7 +12380,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMul32F(v *Value, config *Config) bool {
b := v.Block
@@ -12542,7 +12395,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMul64(v *Value, config *Config) bool {
b := v.Block
@@ -12558,7 +12410,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMul64F(v *Value, config *Config) bool {
b := v.Block
@@ -12574,7 +12425,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpMul8(v *Value, config *Config) bool {
b := v.Block
@@ -12590,7 +12440,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpAMD64NEGL(v *Value, config *Config) bool {
b := v.Block
@@ -12676,7 +12525,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpNeg32(v *Value, config *Config) bool {
b := v.Block
@@ -12690,7 +12538,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpNeg32F(v *Value, config *Config) bool {
b := v.Block
@@ -12707,7 +12554,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNeg64(v *Value, config *Config) bool {
b := v.Block
@@ -12721,7 +12567,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpNeg64F(v *Value, config *Config) bool {
b := v.Block
@@ -12738,7 +12583,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNeg8(v *Value, config *Config) bool {
b := v.Block
@@ -12752,7 +12596,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpNeq16(v *Value, config *Config) bool {
b := v.Block
@@ -12770,7 +12613,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNeq32(v *Value, config *Config) bool {
b := v.Block
@@ -12788,7 +12630,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNeq32F(v *Value, config *Config) bool {
b := v.Block
@@ -12806,7 +12647,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNeq64(v *Value, config *Config) bool {
b := v.Block
@@ -12824,7 +12664,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNeq64F(v *Value, config *Config) bool {
b := v.Block
@@ -12842,7 +12681,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNeq8(v *Value, config *Config) bool {
b := v.Block
@@ -12860,7 +12698,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNeqB(v *Value, config *Config) bool {
b := v.Block
@@ -12878,7 +12715,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNeqPtr(v *Value, config *Config) bool {
b := v.Block
@@ -12896,7 +12732,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpNilCheck(v *Value, config *Config) bool {
b := v.Block
@@ -12912,7 +12747,6 @@
v.AddArg(mem)
return true
}
- return false
}
func rewriteValueAMD64_OpNot(v *Value, config *Config) bool {
b := v.Block
@@ -12927,7 +12761,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpAMD64ORL(v *Value, config *Config) bool {
b := v.Block
@@ -13908,7 +13741,6 @@
v.AddArg(ptr)
return true
}
- return false
}
func rewriteValueAMD64_OpOr16(v *Value, config *Config) bool {
b := v.Block
@@ -13924,7 +13756,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpOr32(v *Value, config *Config) bool {
b := v.Block
@@ -13940,7 +13771,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpOr64(v *Value, config *Config) bool {
b := v.Block
@@ -13956,7 +13786,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpOr8(v *Value, config *Config) bool {
b := v.Block
@@ -13972,7 +13801,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpOrB(v *Value, config *Config) bool {
b := v.Block
@@ -13988,7 +13816,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh16Ux16(v *Value, config *Config) bool {
b := v.Block
@@ -14013,7 +13840,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh16Ux32(v *Value, config *Config) bool {
b := v.Block
@@ -14038,7 +13864,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh16Ux64(v *Value, config *Config) bool {
b := v.Block
@@ -14063,7 +13888,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh16Ux8(v *Value, config *Config) bool {
b := v.Block
@@ -14088,7 +13912,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh16x16(v *Value, config *Config) bool {
b := v.Block
@@ -14116,7 +13939,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh16x32(v *Value, config *Config) bool {
b := v.Block
@@ -14144,7 +13966,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh16x64(v *Value, config *Config) bool {
b := v.Block
@@ -14172,7 +13993,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh16x8(v *Value, config *Config) bool {
b := v.Block
@@ -14200,7 +14020,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh32Ux16(v *Value, config *Config) bool {
b := v.Block
@@ -14225,7 +14044,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh32Ux32(v *Value, config *Config) bool {
b := v.Block
@@ -14250,7 +14068,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh32Ux64(v *Value, config *Config) bool {
b := v.Block
@@ -14275,7 +14092,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh32Ux8(v *Value, config *Config) bool {
b := v.Block
@@ -14300,7 +14116,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh32x16(v *Value, config *Config) bool {
b := v.Block
@@ -14328,7 +14143,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh32x32(v *Value, config *Config) bool {
b := v.Block
@@ -14356,7 +14170,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh32x64(v *Value, config *Config) bool {
b := v.Block
@@ -14384,7 +14197,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh32x8(v *Value, config *Config) bool {
b := v.Block
@@ -14412,7 +14224,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh64Ux16(v *Value, config *Config) bool {
b := v.Block
@@ -14437,7 +14248,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh64Ux32(v *Value, config *Config) bool {
b := v.Block
@@ -14462,7 +14272,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh64Ux64(v *Value, config *Config) bool {
b := v.Block
@@ -14487,7 +14296,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh64Ux8(v *Value, config *Config) bool {
b := v.Block
@@ -14512,7 +14320,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh64x16(v *Value, config *Config) bool {
b := v.Block
@@ -14540,7 +14347,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh64x32(v *Value, config *Config) bool {
b := v.Block
@@ -14568,7 +14374,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh64x64(v *Value, config *Config) bool {
b := v.Block
@@ -14596,7 +14401,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh64x8(v *Value, config *Config) bool {
b := v.Block
@@ -14624,7 +14428,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh8Ux16(v *Value, config *Config) bool {
b := v.Block
@@ -14649,7 +14452,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh8Ux32(v *Value, config *Config) bool {
b := v.Block
@@ -14674,7 +14476,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh8Ux64(v *Value, config *Config) bool {
b := v.Block
@@ -14699,7 +14500,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh8Ux8(v *Value, config *Config) bool {
b := v.Block
@@ -14724,7 +14524,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh8x16(v *Value, config *Config) bool {
b := v.Block
@@ -14752,7 +14551,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh8x32(v *Value, config *Config) bool {
b := v.Block
@@ -14780,7 +14578,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh8x64(v *Value, config *Config) bool {
b := v.Block
@@ -14808,7 +14605,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpRsh8x8(v *Value, config *Config) bool {
b := v.Block
@@ -14836,7 +14632,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueAMD64_OpAMD64SARB(v *Value, config *Config) bool {
b := v.Block
@@ -16359,37 +16154,6 @@
v.AddArg(x)
return true
}
- // match: (SUBLconst (MOVLconst [d]) [c])
- // cond:
- // result: (MOVLconst [int64(int32(d-c))])
- for {
- v_0 := v.Args[0]
- if v_0.Op != OpAMD64MOVLconst {
- break
- }
- d := v_0.AuxInt
- c := v.AuxInt
- v.reset(OpAMD64MOVLconst)
- v.AuxInt = int64(int32(d - c))
- return true
- }
- // match: (SUBLconst (SUBLconst x [d]) [c])
- // cond:
- // result: (ADDLconst [int64(int32(-c-d))] x)
- for {
- v_0 := v.Args[0]
- if v_0.Op != OpAMD64SUBLconst {
- break
- }
- x := v_0.Args[0]
- d := v_0.AuxInt
- c := v.AuxInt
- v.reset(OpAMD64ADDLconst)
- v.AuxInt = int64(int32(-c - d))
- v.AddArg(x)
- return true
- }
- return false
}
func rewriteValueAMD64_OpAMD64SUBQ(v *Value, config *Config) bool {
b := v.Block
@@ -16523,7 +16287,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpSignExt16to64(v *Value, config *Config) bool {
b := v.Block
@@ -16537,7 +16300,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpSignExt32to64(v *Value, config *Config) bool {
b := v.Block
@@ -16551,7 +16313,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpSignExt8to16(v *Value, config *Config) bool {
b := v.Block
@@ -16565,7 +16326,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpSignExt8to32(v *Value, config *Config) bool {
b := v.Block
@@ -16579,7 +16339,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpSignExt8to64(v *Value, config *Config) bool {
b := v.Block
@@ -16593,7 +16352,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpSqrt(v *Value, config *Config) bool {
b := v.Block
@@ -16607,7 +16365,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpStaticCall(v *Value, config *Config) bool {
b := v.Block
@@ -16625,7 +16382,6 @@
v.AddArg(mem)
return true
}
- return false
}
func rewriteValueAMD64_OpStore(v *Value, config *Config) bool {
b := v.Block
@@ -16748,7 +16504,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpSub32(v *Value, config *Config) bool {
b := v.Block
@@ -16764,7 +16519,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpSub32F(v *Value, config *Config) bool {
b := v.Block
@@ -16780,7 +16534,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpSub64(v *Value, config *Config) bool {
b := v.Block
@@ -16796,7 +16549,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpSub64F(v *Value, config *Config) bool {
b := v.Block
@@ -16812,7 +16564,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpSub8(v *Value, config *Config) bool {
b := v.Block
@@ -16828,7 +16579,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpSubPtr(v *Value, config *Config) bool {
b := v.Block
@@ -16844,7 +16594,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpTrunc16to8(v *Value, config *Config) bool {
b := v.Block
@@ -16859,7 +16608,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpTrunc32to16(v *Value, config *Config) bool {
b := v.Block
@@ -16874,7 +16622,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpTrunc32to8(v *Value, config *Config) bool {
b := v.Block
@@ -16889,7 +16636,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpTrunc64to16(v *Value, config *Config) bool {
b := v.Block
@@ -16904,7 +16650,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpTrunc64to32(v *Value, config *Config) bool {
b := v.Block
@@ -16919,7 +16664,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpTrunc64to8(v *Value, config *Config) bool {
b := v.Block
@@ -16934,7 +16678,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpAMD64XORL(v *Value, config *Config) bool {
b := v.Block
@@ -17115,7 +16858,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpXor32(v *Value, config *Config) bool {
b := v.Block
@@ -17131,7 +16873,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpXor64(v *Value, config *Config) bool {
b := v.Block
@@ -17147,7 +16888,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpXor8(v *Value, config *Config) bool {
b := v.Block
@@ -17163,7 +16903,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueAMD64_OpZero(v *Value, config *Config) bool {
b := v.Block
@@ -17491,7 +17230,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpZeroExt16to64(v *Value, config *Config) bool {
b := v.Block
@@ -17505,7 +17243,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpZeroExt32to64(v *Value, config *Config) bool {
b := v.Block
@@ -17519,7 +17256,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpZeroExt8to16(v *Value, config *Config) bool {
b := v.Block
@@ -17533,7 +17269,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpZeroExt8to32(v *Value, config *Config) bool {
b := v.Block
@@ -17547,7 +17282,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteValueAMD64_OpZeroExt8to64(v *Value, config *Config) bool {
b := v.Block
@@ -17561,7 +17295,6 @@
v.AddArg(x)
return true
}
- return false
}
func rewriteBlockAMD64(b *Block) bool {
switch b.Kind {
diff --git a/src/cmd/compile/internal/ssa/rewriteARM.go b/src/cmd/compile/internal/ssa/rewriteARM.go
index 67eff50..598bb9f 100644
--- a/src/cmd/compile/internal/ssa/rewriteARM.go
+++ b/src/cmd/compile/internal/ssa/rewriteARM.go
@@ -82,7 +82,6 @@
v.AddArg(y)
return true
}
- return false
}
func rewriteValueARM_OpAddr(v *Value, config *Config) bool {
b := v.Block
@@ -98,7 +97,6 @@
v.AddArg(base)
return true
}
- return false
}
func rewriteValueARM_OpConst32(v *Value, config *Config) bool {
b := v.Block
@@ -112,7 +110,6 @@
v.AuxInt = val
return true
}
- return false
}
func rewriteValueARM_OpLess32(v *Value, config *Config) bool {
b := v.Block
@@ -130,7 +127,6 @@
v.AddArg(v0)
return true
}
- return false
}
func rewriteValueARM_OpLoad(v *Value, config *Config) bool {
b := v.Block
@@ -228,7 +224,6 @@
v.AddArg(ptr)
return true
}
- return false
}
func rewriteValueARM_OpStaticCall(v *Value, config *Config) bool {
b := v.Block
@@ -246,7 +241,6 @@
v.AddArg(mem)
return true
}
- return false
}
func rewriteValueARM_OpStore(v *Value, config *Config) bool {
b := v.Block
diff --git a/src/cmd/compile/internal/ssa/rewritegeneric.go b/src/cmd/compile/internal/ssa/rewritegeneric.go
index debbf4c..311d5b8 100644
--- a/src/cmd/compile/internal/ssa/rewritegeneric.go
+++ b/src/cmd/compile/internal/ssa/rewritegeneric.go
@@ -1620,7 +1620,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValuegeneric_OpConstSlice(v *Value, config *Config) bool {
b := v.Block
@@ -2497,7 +2496,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValuegeneric_OpEqPtr(v *Value, config *Config) bool {
b := v.Block
@@ -2552,7 +2550,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValuegeneric_OpGeq16(v *Value, config *Config) bool {
b := v.Block
@@ -5780,7 +5777,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValuegeneric_OpNeqPtr(v *Value, config *Config) bool {
b := v.Block
@@ -5831,7 +5827,6 @@
v.AddArg(v1)
return true
}
- return false
}
func rewriteValuegeneric_OpOffPtr(v *Value, config *Config) bool {
b := v.Block