| // Copyright 2015 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // +build gen |
| |
| // This program generates Go code that applies rewrite rules to a Value. |
| // The generated code implements a function of type func (v *Value) bool |
| // which reports whether if did something. |
| // Ideas stolen from Swift: http://www.hpl.hp.com/techreports/Compaq-DEC/WRL-2000-2.html |
| |
| package main |
| |
| import ( |
| "bufio" |
| "bytes" |
| "flag" |
| "fmt" |
| "go/ast" |
| "go/format" |
| "go/parser" |
| "go/printer" |
| "go/token" |
| "io" |
| "io/ioutil" |
| "log" |
| "os" |
| "path" |
| "regexp" |
| "sort" |
| "strconv" |
| "strings" |
| |
| "golang.org/x/tools/go/ast/astutil" |
| ) |
| |
| // rule syntax: |
| // sexpr [&& extra conditions] -> [@block] sexpr (untyped) |
| // sexpr [&& extra conditions] => [@block] sexpr (typed) |
| // |
| // sexpr are s-expressions (lisp-like parenthesized groupings) |
| // sexpr ::= [variable:](opcode sexpr*) |
| // | variable |
| // | <type> |
| // | [auxint] |
| // | {aux} |
| // |
| // aux ::= variable | {code} |
| // type ::= variable | {code} |
| // variable ::= some token |
| // opcode ::= one of the opcodes from the *Ops.go files |
| |
| // extra conditions is just a chunk of Go that evaluates to a boolean. It may use |
| // variables declared in the matching sexpr. The variable "v" is predefined to be |
| // the value matched by the entire rule. |
| |
| // If multiple rules match, the first one in file order is selected. |
| |
| var ( |
| genLog = flag.Bool("log", false, "generate code that logs; for debugging only") |
| addLine = flag.Bool("line", false, "add line number comment to generated rules; for debugging only") |
| ) |
| |
| type Rule struct { |
| Rule string |
| Loc string // file name & line number |
| } |
| |
| func (r Rule) String() string { |
| return fmt.Sprintf("rule %q at %s", r.Rule, r.Loc) |
| } |
| |
| func normalizeSpaces(s string) string { |
| return strings.Join(strings.Fields(strings.TrimSpace(s)), " ") |
| } |
| |
| // parse returns the matching part of the rule, additional conditions, and the result. |
| // parse also reports whether the generated code should use strongly typed aux and auxint fields. |
| func (r Rule) parse() (match, cond, result string, typed bool) { |
| arrow := "->" |
| if strings.Contains(r.Rule, "=>") { |
| arrow = "=>" |
| typed = true |
| } |
| s := strings.Split(r.Rule, arrow) |
| match = normalizeSpaces(s[0]) |
| result = normalizeSpaces(s[1]) |
| cond = "" |
| if i := strings.Index(match, "&&"); i >= 0 { |
| cond = normalizeSpaces(match[i+2:]) |
| match = normalizeSpaces(match[:i]) |
| } |
| return match, cond, result, typed |
| } |
| |
| func genRules(arch arch) { genRulesSuffix(arch, "") } |
| func genSplitLoadRules(arch arch) { genRulesSuffix(arch, "splitload") } |
| |
| func genRulesSuffix(arch arch, suff string) { |
| // Open input file. |
| text, err := os.Open(arch.name + suff + ".rules") |
| if err != nil { |
| if suff == "" { |
| // All architectures must have a plain rules file. |
| log.Fatalf("can't read rule file: %v", err) |
| } |
| // Some architectures have bonus rules files that others don't share. That's fine. |
| return |
| } |
| |
| // oprules contains a list of rules for each block and opcode |
| blockrules := map[string][]Rule{} |
| oprules := map[string][]Rule{} |
| |
| // read rule file |
| scanner := bufio.NewScanner(text) |
| rule := "" |
| var lineno int |
| var ruleLineno int // line number of "->" or "=>" |
| for scanner.Scan() { |
| lineno++ |
| line := scanner.Text() |
| if i := strings.Index(line, "//"); i >= 0 { |
| // Remove comments. Note that this isn't string safe, so |
| // it will truncate lines with // inside strings. Oh well. |
| line = line[:i] |
| } |
| rule += " " + line |
| rule = strings.TrimSpace(rule) |
| if rule == "" { |
| continue |
| } |
| if !strings.Contains(rule, "->") && !strings.Contains(rule, "=>") { |
| continue |
| } |
| if ruleLineno == 0 { |
| ruleLineno = lineno |
| } |
| if strings.HasSuffix(rule, "->") || strings.HasSuffix(rule, "=>") { |
| continue // continue on the next line |
| } |
| if n := balance(rule); n > 0 { |
| continue // open parentheses remain, continue on the next line |
| } else if n < 0 { |
| break // continuing the line can't help, and it will only make errors worse |
| } |
| |
| loc := fmt.Sprintf("%s%s.rules:%d", arch.name, suff, ruleLineno) |
| for _, rule2 := range expandOr(rule) { |
| r := Rule{Rule: rule2, Loc: loc} |
| if rawop := strings.Split(rule2, " ")[0][1:]; isBlock(rawop, arch) { |
| blockrules[rawop] = append(blockrules[rawop], r) |
| continue |
| } |
| // Do fancier value op matching. |
| match, _, _, _ := r.parse() |
| op, oparch, _, _, _, _ := parseValue(match, arch, loc) |
| opname := fmt.Sprintf("Op%s%s", oparch, op.name) |
| oprules[opname] = append(oprules[opname], r) |
| } |
| rule = "" |
| ruleLineno = 0 |
| } |
| if err := scanner.Err(); err != nil { |
| log.Fatalf("scanner failed: %v\n", err) |
| } |
| if balance(rule) != 0 { |
| log.Fatalf("%s.rules:%d: unbalanced rule: %v\n", arch.name, lineno, rule) |
| } |
| |
| // Order all the ops. |
| var ops []string |
| for op := range oprules { |
| ops = append(ops, op) |
| } |
| sort.Strings(ops) |
| |
| genFile := &File{Arch: arch, Suffix: suff} |
| // Main rewrite routine is a switch on v.Op. |
| fn := &Func{Kind: "Value", ArgLen: -1} |
| |
| sw := &Switch{Expr: exprf("v.Op")} |
| for _, op := range ops { |
| eop, ok := parseEllipsisRules(oprules[op], arch) |
| if ok { |
| if strings.Contains(oprules[op][0].Rule, "=>") && opByName(arch, op).aux != opByName(arch, eop).aux { |
| panic(fmt.Sprintf("can't use ... for ops that have different aux types: %s and %s", op, eop)) |
| } |
| swc := &Case{Expr: exprf("%s", op)} |
| swc.add(stmtf("v.Op = %s", eop)) |
| swc.add(stmtf("return true")) |
| sw.add(swc) |
| continue |
| } |
| |
| swc := &Case{Expr: exprf("%s", op)} |
| swc.add(stmtf("return rewriteValue%s%s_%s(v)", arch.name, suff, op)) |
| sw.add(swc) |
| } |
| fn.add(sw) |
| fn.add(stmtf("return false")) |
| genFile.add(fn) |
| |
| // Generate a routine per op. Note that we don't make one giant routine |
| // because it is too big for some compilers. |
| for _, op := range ops { |
| rules := oprules[op] |
| _, ok := parseEllipsisRules(oprules[op], arch) |
| if ok { |
| continue |
| } |
| |
| // rr is kept between iterations, so that each rule can check |
| // that the previous rule wasn't unconditional. |
| var rr *RuleRewrite |
| fn := &Func{ |
| Kind: "Value", |
| Suffix: fmt.Sprintf("_%s", op), |
| ArgLen: opByName(arch, op).argLength, |
| } |
| fn.add(declf("b", "v.Block")) |
| fn.add(declf("config", "b.Func.Config")) |
| fn.add(declf("fe", "b.Func.fe")) |
| fn.add(declf("typ", "&b.Func.Config.Types")) |
| for _, rule := range rules { |
| if rr != nil && !rr.CanFail { |
| log.Fatalf("unconditional rule %s is followed by other rules", rr.Match) |
| } |
| rr = &RuleRewrite{Loc: rule.Loc} |
| rr.Match, rr.Cond, rr.Result, rr.Typed = rule.parse() |
| pos, _ := genMatch(rr, arch, rr.Match, fn.ArgLen >= 0) |
| if pos == "" { |
| pos = "v.Pos" |
| } |
| if rr.Cond != "" { |
| rr.add(breakf("!(%s)", rr.Cond)) |
| } |
| genResult(rr, arch, rr.Result, pos) |
| if *genLog { |
| rr.add(stmtf("logRule(%q)", rule.Loc)) |
| } |
| fn.add(rr) |
| } |
| if rr.CanFail { |
| fn.add(stmtf("return false")) |
| } |
| genFile.add(fn) |
| } |
| |
| // Generate block rewrite function. There are only a few block types |
| // so we can make this one function with a switch. |
| fn = &Func{Kind: "Block"} |
| fn.add(declf("config", "b.Func.Config")) |
| fn.add(declf("typ", "&b.Func.Config.Types")) |
| |
| sw = &Switch{Expr: exprf("b.Kind")} |
| ops = ops[:0] |
| for op := range blockrules { |
| ops = append(ops, op) |
| } |
| sort.Strings(ops) |
| for _, op := range ops { |
| name, data := getBlockInfo(op, arch) |
| swc := &Case{Expr: exprf("%s", name)} |
| for _, rule := range blockrules[op] { |
| swc.add(genBlockRewrite(rule, arch, data)) |
| } |
| sw.add(swc) |
| } |
| fn.add(sw) |
| fn.add(stmtf("return false")) |
| genFile.add(fn) |
| |
| // Remove unused imports and variables. |
| buf := new(bytes.Buffer) |
| fprint(buf, genFile) |
| fset := token.NewFileSet() |
| file, err := parser.ParseFile(fset, "", buf, parser.ParseComments) |
| if err != nil { |
| filename := fmt.Sprintf("%s_broken.go", arch.name) |
| if err := ioutil.WriteFile(filename, buf.Bytes(), 0644); err != nil { |
| log.Printf("failed to dump broken code to %s: %v", filename, err) |
| } else { |
| log.Printf("dumped broken code to %s", filename) |
| } |
| log.Fatalf("failed to parse generated code for arch %s: %v", arch.name, err) |
| } |
| tfile := fset.File(file.Pos()) |
| |
| // First, use unusedInspector to find the unused declarations by their |
| // start position. |
| u := unusedInspector{unused: make(map[token.Pos]bool)} |
| u.node(file) |
| |
| // Then, delete said nodes via astutil.Apply. |
| pre := func(c *astutil.Cursor) bool { |
| node := c.Node() |
| if node == nil { |
| return true |
| } |
| if u.unused[node.Pos()] { |
| c.Delete() |
| // Unused imports and declarations use exactly |
| // one line. Prevent leaving an empty line. |
| tfile.MergeLine(tfile.Position(node.Pos()).Line) |
| return false |
| } |
| return true |
| } |
| post := func(c *astutil.Cursor) bool { |
| switch node := c.Node().(type) { |
| case *ast.GenDecl: |
| if len(node.Specs) == 0 { |
| // Don't leave a broken or empty GenDecl behind, |
| // such as "import ()". |
| c.Delete() |
| } |
| } |
| return true |
| } |
| file = astutil.Apply(file, pre, post).(*ast.File) |
| |
| // Write the well-formatted source to file |
| f, err := os.Create("../rewrite" + arch.name + suff + ".go") |
| if err != nil { |
| log.Fatalf("can't write output: %v", err) |
| } |
| defer f.Close() |
| // gofmt result; use a buffered writer, as otherwise go/format spends |
| // far too much time in syscalls. |
| bw := bufio.NewWriter(f) |
| if err := format.Node(bw, fset, file); err != nil { |
| log.Fatalf("can't format output: %v", err) |
| } |
| if err := bw.Flush(); err != nil { |
| log.Fatalf("can't write output: %v", err) |
| } |
| if err := f.Close(); err != nil { |
| log.Fatalf("can't write output: %v", err) |
| } |
| } |
| |
| // unusedInspector can be used to detect unused variables and imports in an |
| // ast.Node via its node method. The result is available in the "unused" map. |
| // |
| // note that unusedInspector is lazy and best-effort; it only supports the node |
| // types and patterns used by the rulegen program. |
| type unusedInspector struct { |
| // scope is the current scope, which can never be nil when a declaration |
| // is encountered. That is, the unusedInspector.node entrypoint should |
| // generally be an entire file or block. |
| scope *scope |
| |
| // unused is the resulting set of unused declared names, indexed by the |
| // starting position of the node that declared the name. |
| unused map[token.Pos]bool |
| |
| // defining is the object currently being defined; this is useful so |
| // that if "foo := bar" is unused and removed, we can then detect if |
| // "bar" becomes unused as well. |
| defining *object |
| } |
| |
| // scoped opens a new scope when called, and returns a function which closes |
| // that same scope. When a scope is closed, unused variables are recorded. |
| func (u *unusedInspector) scoped() func() { |
| outer := u.scope |
| u.scope = &scope{outer: outer, objects: map[string]*object{}} |
| return func() { |
| for anyUnused := true; anyUnused; { |
| anyUnused = false |
| for _, obj := range u.scope.objects { |
| if obj.numUses > 0 { |
| continue |
| } |
| u.unused[obj.pos] = true |
| for _, used := range obj.used { |
| if used.numUses--; used.numUses == 0 { |
| anyUnused = true |
| } |
| } |
| // We've decremented numUses for each of the |
| // objects in used. Zero this slice too, to keep |
| // everything consistent. |
| obj.used = nil |
| } |
| } |
| u.scope = outer |
| } |
| } |
| |
| func (u *unusedInspector) exprs(list []ast.Expr) { |
| for _, x := range list { |
| u.node(x) |
| } |
| } |
| |
| func (u *unusedInspector) node(node ast.Node) { |
| switch node := node.(type) { |
| case *ast.File: |
| defer u.scoped()() |
| for _, decl := range node.Decls { |
| u.node(decl) |
| } |
| case *ast.GenDecl: |
| for _, spec := range node.Specs { |
| u.node(spec) |
| } |
| case *ast.ImportSpec: |
| impPath, _ := strconv.Unquote(node.Path.Value) |
| name := path.Base(impPath) |
| u.scope.objects[name] = &object{ |
| name: name, |
| pos: node.Pos(), |
| } |
| case *ast.FuncDecl: |
| u.node(node.Type) |
| if node.Body != nil { |
| u.node(node.Body) |
| } |
| case *ast.FuncType: |
| if node.Params != nil { |
| u.node(node.Params) |
| } |
| if node.Results != nil { |
| u.node(node.Results) |
| } |
| case *ast.FieldList: |
| for _, field := range node.List { |
| u.node(field) |
| } |
| case *ast.Field: |
| u.node(node.Type) |
| |
| // statements |
| |
| case *ast.BlockStmt: |
| defer u.scoped()() |
| for _, stmt := range node.List { |
| u.node(stmt) |
| } |
| case *ast.DeclStmt: |
| u.node(node.Decl) |
| case *ast.IfStmt: |
| if node.Init != nil { |
| u.node(node.Init) |
| } |
| u.node(node.Cond) |
| u.node(node.Body) |
| if node.Else != nil { |
| u.node(node.Else) |
| } |
| case *ast.ForStmt: |
| if node.Init != nil { |
| u.node(node.Init) |
| } |
| if node.Cond != nil { |
| u.node(node.Cond) |
| } |
| if node.Post != nil { |
| u.node(node.Post) |
| } |
| u.node(node.Body) |
| case *ast.SwitchStmt: |
| if node.Init != nil { |
| u.node(node.Init) |
| } |
| if node.Tag != nil { |
| u.node(node.Tag) |
| } |
| u.node(node.Body) |
| case *ast.CaseClause: |
| u.exprs(node.List) |
| defer u.scoped()() |
| for _, stmt := range node.Body { |
| u.node(stmt) |
| } |
| case *ast.BranchStmt: |
| case *ast.ExprStmt: |
| u.node(node.X) |
| case *ast.AssignStmt: |
| if node.Tok != token.DEFINE { |
| u.exprs(node.Rhs) |
| u.exprs(node.Lhs) |
| break |
| } |
| lhs := node.Lhs |
| if len(lhs) == 2 && lhs[1].(*ast.Ident).Name == "_" { |
| lhs = lhs[:1] |
| } |
| if len(lhs) != 1 { |
| panic("no support for := with multiple names") |
| } |
| |
| name := lhs[0].(*ast.Ident) |
| obj := &object{ |
| name: name.Name, |
| pos: name.NamePos, |
| } |
| |
| old := u.defining |
| u.defining = obj |
| u.exprs(node.Rhs) |
| u.defining = old |
| |
| u.scope.objects[name.Name] = obj |
| case *ast.ReturnStmt: |
| u.exprs(node.Results) |
| case *ast.IncDecStmt: |
| u.node(node.X) |
| |
| // expressions |
| |
| case *ast.CallExpr: |
| u.node(node.Fun) |
| u.exprs(node.Args) |
| case *ast.SelectorExpr: |
| u.node(node.X) |
| case *ast.UnaryExpr: |
| u.node(node.X) |
| case *ast.BinaryExpr: |
| u.node(node.X) |
| u.node(node.Y) |
| case *ast.StarExpr: |
| u.node(node.X) |
| case *ast.ParenExpr: |
| u.node(node.X) |
| case *ast.IndexExpr: |
| u.node(node.X) |
| u.node(node.Index) |
| case *ast.TypeAssertExpr: |
| u.node(node.X) |
| u.node(node.Type) |
| case *ast.Ident: |
| if obj := u.scope.Lookup(node.Name); obj != nil { |
| obj.numUses++ |
| if u.defining != nil { |
| u.defining.used = append(u.defining.used, obj) |
| } |
| } |
| case *ast.BasicLit: |
| case *ast.ValueSpec: |
| u.exprs(node.Values) |
| default: |
| panic(fmt.Sprintf("unhandled node: %T", node)) |
| } |
| } |
| |
| // scope keeps track of a certain scope and its declared names, as well as the |
| // outer (parent) scope. |
| type scope struct { |
| outer *scope // can be nil, if this is the top-level scope |
| objects map[string]*object // indexed by each declared name |
| } |
| |
| func (s *scope) Lookup(name string) *object { |
| if obj := s.objects[name]; obj != nil { |
| return obj |
| } |
| if s.outer == nil { |
| return nil |
| } |
| return s.outer.Lookup(name) |
| } |
| |
| // object keeps track of a declared name, such as a variable or import. |
| type object struct { |
| name string |
| pos token.Pos // start position of the node declaring the object |
| |
| numUses int // number of times this object is used |
| used []*object // objects that its declaration makes use of |
| } |
| |
| func fprint(w io.Writer, n Node) { |
| switch n := n.(type) { |
| case *File: |
| file := n |
| seenRewrite := make(map[[3]string]string) |
| fmt.Fprintf(w, "// Code generated from gen/%s%s.rules; DO NOT EDIT.\n", n.Arch.name, n.Suffix) |
| fmt.Fprintf(w, "// generated with: cd gen; go run *.go\n") |
| fmt.Fprintf(w, "\npackage ssa\n") |
| for _, path := range append([]string{ |
| "fmt", |
| "math", |
| "cmd/internal/obj", |
| "cmd/internal/objabi", |
| "cmd/compile/internal/types", |
| }, n.Arch.imports...) { |
| fmt.Fprintf(w, "import %q\n", path) |
| } |
| for _, f := range n.List { |
| f := f.(*Func) |
| fmt.Fprintf(w, "func rewrite%s%s%s%s(", f.Kind, n.Arch.name, n.Suffix, f.Suffix) |
| fmt.Fprintf(w, "%c *%s) bool {\n", strings.ToLower(f.Kind)[0], f.Kind) |
| if f.Kind == "Value" && f.ArgLen > 0 { |
| for i := f.ArgLen - 1; i >= 0; i-- { |
| fmt.Fprintf(w, "v_%d := v.Args[%d]\n", i, i) |
| } |
| } |
| for _, n := range f.List { |
| fprint(w, n) |
| |
| if rr, ok := n.(*RuleRewrite); ok { |
| k := [3]string{ |
| normalizeMatch(rr.Match, file.Arch), |
| normalizeWhitespace(rr.Cond), |
| normalizeWhitespace(rr.Result), |
| } |
| if prev, ok := seenRewrite[k]; ok { |
| log.Fatalf("duplicate rule %s, previously seen at %s\n", rr.Loc, prev) |
| } |
| seenRewrite[k] = rr.Loc |
| } |
| } |
| fmt.Fprintf(w, "}\n") |
| } |
| case *Switch: |
| fmt.Fprintf(w, "switch ") |
| fprint(w, n.Expr) |
| fmt.Fprintf(w, " {\n") |
| for _, n := range n.List { |
| fprint(w, n) |
| } |
| fmt.Fprintf(w, "}\n") |
| case *Case: |
| fmt.Fprintf(w, "case ") |
| fprint(w, n.Expr) |
| fmt.Fprintf(w, ":\n") |
| for _, n := range n.List { |
| fprint(w, n) |
| } |
| case *RuleRewrite: |
| if *addLine { |
| fmt.Fprintf(w, "// %s\n", n.Loc) |
| } |
| fmt.Fprintf(w, "// match: %s\n", n.Match) |
| if n.Cond != "" { |
| fmt.Fprintf(w, "// cond: %s\n", n.Cond) |
| } |
| fmt.Fprintf(w, "// result: %s\n", n.Result) |
| fmt.Fprintf(w, "for %s {\n", n.Check) |
| nCommutative := 0 |
| for _, n := range n.List { |
| if b, ok := n.(*CondBreak); ok { |
| b.InsideCommuteLoop = nCommutative > 0 |
| } |
| fprint(w, n) |
| if loop, ok := n.(StartCommuteLoop); ok { |
| if nCommutative != loop.Depth { |
| panic("mismatch commute loop depth") |
| } |
| nCommutative++ |
| } |
| } |
| fmt.Fprintf(w, "return true\n") |
| for i := 0; i < nCommutative; i++ { |
| fmt.Fprintln(w, "}") |
| } |
| if n.CommuteDepth > 0 && n.CanFail { |
| fmt.Fprint(w, "break\n") |
| } |
| fmt.Fprintf(w, "}\n") |
| case *Declare: |
| fmt.Fprintf(w, "%s := ", n.Name) |
| fprint(w, n.Value) |
| fmt.Fprintln(w) |
| case *CondBreak: |
| fmt.Fprintf(w, "if ") |
| fprint(w, n.Cond) |
| fmt.Fprintf(w, " {\n") |
| if n.InsideCommuteLoop { |
| fmt.Fprintf(w, "continue") |
| } else { |
| fmt.Fprintf(w, "break") |
| } |
| fmt.Fprintf(w, "\n}\n") |
| case ast.Node: |
| printConfig.Fprint(w, emptyFset, n) |
| if _, ok := n.(ast.Stmt); ok { |
| fmt.Fprintln(w) |
| } |
| case StartCommuteLoop: |
| fmt.Fprintf(w, "for _i%[1]d := 0; _i%[1]d <= 1; _i%[1]d, %[2]s_0, %[2]s_1 = _i%[1]d + 1, %[2]s_1, %[2]s_0 {\n", n.Depth, n.V) |
| default: |
| log.Fatalf("cannot print %T", n) |
| } |
| } |
| |
| var printConfig = printer.Config{ |
| Mode: printer.RawFormat, // we use go/format later, so skip work here |
| } |
| |
| var emptyFset = token.NewFileSet() |
| |
| // Node can be a Statement or an ast.Expr. |
| type Node interface{} |
| |
| // Statement can be one of our high-level statement struct types, or an |
| // ast.Stmt under some limited circumstances. |
| type Statement interface{} |
| |
| // BodyBase is shared by all of our statement pseudo-node types which can |
| // contain other statements. |
| type BodyBase struct { |
| List []Statement |
| CanFail bool |
| } |
| |
| func (w *BodyBase) add(node Statement) { |
| var last Statement |
| if len(w.List) > 0 { |
| last = w.List[len(w.List)-1] |
| } |
| if node, ok := node.(*CondBreak); ok { |
| w.CanFail = true |
| if last, ok := last.(*CondBreak); ok { |
| // Add to the previous "if <cond> { break }" via a |
| // logical OR, which will save verbosity. |
| last.Cond = &ast.BinaryExpr{ |
| Op: token.LOR, |
| X: last.Cond, |
| Y: node.Cond, |
| } |
| return |
| } |
| } |
| |
| w.List = append(w.List, node) |
| } |
| |
| // predeclared contains globally known tokens that should not be redefined. |
| var predeclared = map[string]bool{ |
| "nil": true, |
| "false": true, |
| "true": true, |
| } |
| |
| // declared reports if the body contains a Declare with the given name. |
| func (w *BodyBase) declared(name string) bool { |
| if predeclared[name] { |
| // Treat predeclared names as having already been declared. |
| // This lets us use nil to match an aux field or |
| // true and false to match an auxint field. |
| return true |
| } |
| for _, s := range w.List { |
| if decl, ok := s.(*Declare); ok && decl.Name == name { |
| return true |
| } |
| } |
| return false |
| } |
| |
| // These types define some high-level statement struct types, which can be used |
| // as a Statement. This allows us to keep some node structs simpler, and have |
| // higher-level nodes such as an entire rule rewrite. |
| // |
| // Note that ast.Expr is always used as-is; we don't declare our own expression |
| // nodes. |
| type ( |
| File struct { |
| BodyBase // []*Func |
| Arch arch |
| Suffix string |
| } |
| Func struct { |
| BodyBase |
| Kind string // "Value" or "Block" |
| Suffix string |
| ArgLen int32 // if kind == "Value", number of args for this op |
| } |
| Switch struct { |
| BodyBase // []*Case |
| Expr ast.Expr |
| } |
| Case struct { |
| BodyBase |
| Expr ast.Expr |
| } |
| RuleRewrite struct { |
| BodyBase |
| Match, Cond, Result string // top comments |
| Check string // top-level boolean expression |
| |
| Alloc int // for unique var names |
| Loc string // file name & line number of the original rule |
| CommuteDepth int // used to track depth of commute loops |
| Typed bool // aux and auxint fields should be strongly typed |
| } |
| Declare struct { |
| Name string |
| Value ast.Expr |
| } |
| CondBreak struct { |
| Cond ast.Expr |
| InsideCommuteLoop bool |
| } |
| StartCommuteLoop struct { |
| Depth int |
| V string |
| } |
| ) |
| |
| // exprf parses a Go expression generated from fmt.Sprintf, panicking if an |
| // error occurs. |
| func exprf(format string, a ...interface{}) ast.Expr { |
| src := fmt.Sprintf(format, a...) |
| expr, err := parser.ParseExpr(src) |
| if err != nil { |
| log.Fatalf("expr parse error on %q: %v", src, err) |
| } |
| return expr |
| } |
| |
| // stmtf parses a Go statement generated from fmt.Sprintf. This function is only |
| // meant for simple statements that don't have a custom Statement node declared |
| // in this package, such as ast.ReturnStmt or ast.ExprStmt. |
| func stmtf(format string, a ...interface{}) Statement { |
| src := fmt.Sprintf(format, a...) |
| fsrc := "package p\nfunc _() {\n" + src + "\n}\n" |
| file, err := parser.ParseFile(token.NewFileSet(), "", fsrc, 0) |
| if err != nil { |
| log.Fatalf("stmt parse error on %q: %v", src, err) |
| } |
| return file.Decls[0].(*ast.FuncDecl).Body.List[0] |
| } |
| |
| // declf constructs a simple "name := value" declaration, using exprf for its |
| // value. |
| func declf(name, format string, a ...interface{}) *Declare { |
| return &Declare{name, exprf(format, a...)} |
| } |
| |
| // breakf constructs a simple "if cond { break }" statement, using exprf for its |
| // condition. |
| func breakf(format string, a ...interface{}) *CondBreak { |
| return &CondBreak{Cond: exprf(format, a...)} |
| } |
| |
| func genBlockRewrite(rule Rule, arch arch, data blockData) *RuleRewrite { |
| rr := &RuleRewrite{Loc: rule.Loc} |
| rr.Match, rr.Cond, rr.Result, rr.Typed = rule.parse() |
| _, _, auxint, aux, s := extract(rr.Match) // remove parens, then split |
| |
| // check match of control values |
| if len(s) < data.controls { |
| log.Fatalf("incorrect number of arguments in %s, got %v wanted at least %v", rule, len(s), data.controls) |
| } |
| controls := s[:data.controls] |
| pos := make([]string, data.controls) |
| for i, arg := range controls { |
| cname := fmt.Sprintf("b.Controls[%v]", i) |
| if strings.Contains(arg, "(") { |
| vname, expr := splitNameExpr(arg) |
| if vname == "" { |
| vname = fmt.Sprintf("v_%v", i) |
| } |
| rr.add(declf(vname, cname)) |
| p, op := genMatch0(rr, arch, expr, vname, nil, false) // TODO: pass non-nil cnt? |
| if op != "" { |
| check := fmt.Sprintf("%s.Op == %s", cname, op) |
| if rr.Check == "" { |
| rr.Check = check |
| } else { |
| rr.Check += " && " + check |
| } |
| } |
| if p == "" { |
| p = vname + ".Pos" |
| } |
| pos[i] = p |
| } else { |
| rr.add(declf(arg, cname)) |
| pos[i] = arg + ".Pos" |
| } |
| } |
| for _, e := range []struct { |
| name, field, dclType string |
| }{ |
| {auxint, "AuxInt", data.auxIntType()}, |
| {aux, "Aux", data.auxType()}, |
| } { |
| if e.name == "" { |
| continue |
| } |
| if !rr.Typed { |
| if !token.IsIdentifier(e.name) || rr.declared(e.name) { |
| // code or variable |
| rr.add(breakf("b.%s != %s", e.field, e.name)) |
| } else { |
| rr.add(declf(e.name, "b.%s", e.field)) |
| } |
| continue |
| } |
| |
| if e.dclType == "" { |
| log.Fatalf("op %s has no declared type for %s", data.name, e.field) |
| } |
| if !token.IsIdentifier(e.name) || rr.declared(e.name) { |
| rr.add(breakf("%sTo%s(b.%s) != %s", unTitle(e.field), title(e.dclType), e.field, e.name)) |
| } else { |
| rr.add(declf(e.name, "%sTo%s(b.%s)", unTitle(e.field), title(e.dclType), e.field)) |
| } |
| } |
| if rr.Cond != "" { |
| rr.add(breakf("!(%s)", rr.Cond)) |
| } |
| |
| // Rule matches. Generate result. |
| outop, _, auxint, aux, t := extract(rr.Result) // remove parens, then split |
| blockName, outdata := getBlockInfo(outop, arch) |
| if len(t) < outdata.controls { |
| log.Fatalf("incorrect number of output arguments in %s, got %v wanted at least %v", rule, len(s), outdata.controls) |
| } |
| |
| // Check if newsuccs is the same set as succs. |
| succs := s[data.controls:] |
| newsuccs := t[outdata.controls:] |
| m := map[string]bool{} |
| for _, succ := range succs { |
| if m[succ] { |
| log.Fatalf("can't have a repeat successor name %s in %s", succ, rule) |
| } |
| m[succ] = true |
| } |
| for _, succ := range newsuccs { |
| if !m[succ] { |
| log.Fatalf("unknown successor %s in %s", succ, rule) |
| } |
| delete(m, succ) |
| } |
| if len(m) != 0 { |
| log.Fatalf("unmatched successors %v in %s", m, rule) |
| } |
| |
| var genControls [2]string |
| for i, control := range t[:outdata.controls] { |
| // Select a source position for any new control values. |
| // TODO: does it always make sense to use the source position |
| // of the original control values or should we be using the |
| // block's source position in some cases? |
| newpos := "b.Pos" // default to block's source position |
| if i < len(pos) && pos[i] != "" { |
| // Use the previous control value's source position. |
| newpos = pos[i] |
| } |
| |
| // Generate a new control value (or copy an existing value). |
| genControls[i] = genResult0(rr, arch, control, false, false, newpos, nil) |
| } |
| switch outdata.controls { |
| case 0: |
| rr.add(stmtf("b.Reset(%s)", blockName)) |
| case 1: |
| rr.add(stmtf("b.resetWithControl(%s, %s)", blockName, genControls[0])) |
| case 2: |
| rr.add(stmtf("b.resetWithControl2(%s, %s, %s)", blockName, genControls[0], genControls[1])) |
| default: |
| log.Fatalf("too many controls: %d", outdata.controls) |
| } |
| |
| if auxint != "" { |
| if rr.Typed { |
| // Make sure auxint value has the right type. |
| rr.add(stmtf("b.AuxInt = %sToAuxInt(%s)", unTitle(outdata.auxIntType()), auxint)) |
| } else { |
| rr.add(stmtf("b.AuxInt = %s", auxint)) |
| } |
| } |
| if aux != "" { |
| if rr.Typed { |
| // Make sure aux value has the right type. |
| rr.add(stmtf("b.Aux = %sToAux(%s)", unTitle(outdata.auxType()), aux)) |
| } else { |
| rr.add(stmtf("b.Aux = %s", aux)) |
| } |
| } |
| |
| succChanged := false |
| for i := 0; i < len(succs); i++ { |
| if succs[i] != newsuccs[i] { |
| succChanged = true |
| } |
| } |
| if succChanged { |
| if len(succs) != 2 { |
| log.Fatalf("changed successors, len!=2 in %s", rule) |
| } |
| if succs[0] != newsuccs[1] || succs[1] != newsuccs[0] { |
| log.Fatalf("can only handle swapped successors in %s", rule) |
| } |
| rr.add(stmtf("b.swapSuccessors()")) |
| } |
| |
| if *genLog { |
| rr.add(stmtf("logRule(%q)", rule.Loc)) |
| } |
| return rr |
| } |
| |
| // genMatch returns the variable whose source position should be used for the |
| // result (or "" if no opinion), and a boolean that reports whether the match can fail. |
| func genMatch(rr *RuleRewrite, arch arch, match string, pregenTop bool) (pos, checkOp string) { |
| cnt := varCount(rr) |
| return genMatch0(rr, arch, match, "v", cnt, pregenTop) |
| } |
| |
| func genMatch0(rr *RuleRewrite, arch arch, match, v string, cnt map[string]int, pregenTop bool) (pos, checkOp string) { |
| if match[0] != '(' || match[len(match)-1] != ')' { |
| log.Fatalf("%s: non-compound expr in genMatch0: %q", rr.Loc, match) |
| } |
| op, oparch, typ, auxint, aux, args := parseValue(match, arch, rr.Loc) |
| |
| checkOp = fmt.Sprintf("Op%s%s", oparch, op.name) |
| |
| if op.faultOnNilArg0 || op.faultOnNilArg1 { |
| // Prefer the position of an instruction which could fault. |
| pos = v + ".Pos" |
| } |
| |
| for _, e := range []struct { |
| name, field, dclType string |
| }{ |
| {typ, "Type", "*types.Type"}, |
| {auxint, "AuxInt", op.auxIntType()}, |
| {aux, "Aux", op.auxType()}, |
| } { |
| if e.name == "" { |
| continue |
| } |
| if !rr.Typed { |
| if !token.IsIdentifier(e.name) || rr.declared(e.name) { |
| // code or variable |
| rr.add(breakf("%s.%s != %s", v, e.field, e.name)) |
| } else { |
| rr.add(declf(e.name, "%s.%s", v, e.field)) |
| } |
| continue |
| } |
| |
| if e.dclType == "" { |
| log.Fatalf("op %s has no declared type for %s", op.name, e.field) |
| } |
| if !token.IsIdentifier(e.name) || rr.declared(e.name) { |
| switch e.field { |
| case "Aux": |
| rr.add(breakf("auxTo%s(%s.%s) != %s", title(e.dclType), v, e.field, e.name)) |
| case "AuxInt": |
| rr.add(breakf("auxIntTo%s(%s.%s) != %s", title(e.dclType), v, e.field, e.name)) |
| case "Type": |
| rr.add(breakf("%s.%s != %s", v, e.field, e.name)) |
| } |
| } else { |
| switch e.field { |
| case "Aux": |
| rr.add(declf(e.name, "auxTo%s(%s.%s)", title(e.dclType), v, e.field)) |
| case "AuxInt": |
| rr.add(declf(e.name, "auxIntTo%s(%s.%s)", title(e.dclType), v, e.field)) |
| case "Type": |
| rr.add(declf(e.name, "%s.%s", v, e.field)) |
| } |
| } |
| } |
| |
| commutative := op.commutative |
| if commutative { |
| if args[0] == args[1] { |
| // When we have (Add x x), for any x, |
| // even if there are other uses of x besides these two, |
| // and even if x is not a variable, |
| // we can skip the commutative match. |
| commutative = false |
| } |
| if cnt[args[0]] == 1 && cnt[args[1]] == 1 { |
| // When we have (Add x y) with no other uses |
| // of x and y in the matching rule and condition, |
| // then we can skip the commutative match (Add y x). |
| commutative = false |
| } |
| } |
| |
| if !pregenTop { |
| // Access last argument first to minimize bounds checks. |
| for n := len(args) - 1; n > 0; n-- { |
| a := args[n] |
| if a == "_" { |
| continue |
| } |
| if !rr.declared(a) && token.IsIdentifier(a) && !(commutative && len(args) == 2) { |
| rr.add(declf(a, "%s.Args[%d]", v, n)) |
| // delete the last argument so it is not reprocessed |
| args = args[:n] |
| } else { |
| rr.add(stmtf("_ = %s.Args[%d]", v, n)) |
| } |
| break |
| } |
| } |
| if commutative && !pregenTop { |
| for i := 0; i <= 1; i++ { |
| vname := fmt.Sprintf("%s_%d", v, i) |
| rr.add(declf(vname, "%s.Args[%d]", v, i)) |
| } |
| } |
| if commutative { |
| rr.add(StartCommuteLoop{rr.CommuteDepth, v}) |
| rr.CommuteDepth++ |
| } |
| for i, arg := range args { |
| if arg == "_" { |
| continue |
| } |
| var rhs string |
| if (commutative && i < 2) || pregenTop { |
| rhs = fmt.Sprintf("%s_%d", v, i) |
| } else { |
| rhs = fmt.Sprintf("%s.Args[%d]", v, i) |
| } |
| if !strings.Contains(arg, "(") { |
| // leaf variable |
| if rr.declared(arg) { |
| // variable already has a definition. Check whether |
| // the old definition and the new definition match. |
| // For example, (add x x). Equality is just pointer equality |
| // on Values (so cse is important to do before lowering). |
| rr.add(breakf("%s != %s", arg, rhs)) |
| } else { |
| if arg != rhs { |
| rr.add(declf(arg, "%s", rhs)) |
| } |
| } |
| continue |
| } |
| // compound sexpr |
| argname, expr := splitNameExpr(arg) |
| if argname == "" { |
| argname = fmt.Sprintf("%s_%d", v, i) |
| } |
| if argname == "b" { |
| log.Fatalf("don't name args 'b', it is ambiguous with blocks") |
| } |
| |
| if argname != rhs { |
| rr.add(declf(argname, "%s", rhs)) |
| } |
| bexpr := exprf("%s.Op != addLater", argname) |
| rr.add(&CondBreak{Cond: bexpr}) |
| argPos, argCheckOp := genMatch0(rr, arch, expr, argname, cnt, false) |
| bexpr.(*ast.BinaryExpr).Y.(*ast.Ident).Name = argCheckOp |
| |
| if argPos != "" { |
| // Keep the argument in preference to the parent, as the |
| // argument is normally earlier in program flow. |
| // Keep the argument in preference to an earlier argument, |
| // as that prefers the memory argument which is also earlier |
| // in the program flow. |
| pos = argPos |
| } |
| } |
| |
| if op.argLength == -1 { |
| rr.add(breakf("len(%s.Args) != %d", v, len(args))) |
| } |
| return pos, checkOp |
| } |
| |
| func genResult(rr *RuleRewrite, arch arch, result, pos string) { |
| move := result[0] == '@' |
| if move { |
| // parse @block directive |
| s := strings.SplitN(result[1:], " ", 2) |
| rr.add(stmtf("b = %s", s[0])) |
| result = s[1] |
| } |
| cse := make(map[string]string) |
| genResult0(rr, arch, result, true, move, pos, cse) |
| } |
| |
| func genResult0(rr *RuleRewrite, arch arch, result string, top, move bool, pos string, cse map[string]string) string { |
| resname, expr := splitNameExpr(result) |
| result = expr |
| // TODO: when generating a constant result, use f.constVal to avoid |
| // introducing copies just to clean them up again. |
| if result[0] != '(' { |
| // variable |
| if top { |
| // It in not safe in general to move a variable between blocks |
| // (and particularly not a phi node). |
| // Introduce a copy. |
| rr.add(stmtf("v.copyOf(%s)", result)) |
| } |
| return result |
| } |
| |
| w := normalizeWhitespace(result) |
| if prev := cse[w]; prev != "" { |
| return prev |
| } |
| |
| op, oparch, typ, auxint, aux, args := parseValue(result, arch, rr.Loc) |
| |
| // Find the type of the variable. |
| typeOverride := typ != "" |
| if typ == "" && op.typ != "" { |
| typ = typeName(op.typ) |
| } |
| |
| v := "v" |
| if top && !move { |
| rr.add(stmtf("v.reset(Op%s%s)", oparch, op.name)) |
| if typeOverride { |
| rr.add(stmtf("v.Type = %s", typ)) |
| } |
| } else { |
| if typ == "" { |
| log.Fatalf("sub-expression %s (op=Op%s%s) at %s must have a type", result, oparch, op.name, rr.Loc) |
| } |
| if resname == "" { |
| v = fmt.Sprintf("v%d", rr.Alloc) |
| } else { |
| v = resname |
| } |
| rr.Alloc++ |
| rr.add(declf(v, "b.NewValue0(%s, Op%s%s, %s)", pos, oparch, op.name, typ)) |
| if move && top { |
| // Rewrite original into a copy |
| rr.add(stmtf("v.copyOf(%s)", v)) |
| } |
| } |
| |
| if auxint != "" { |
| if rr.Typed { |
| // Make sure auxint value has the right type. |
| rr.add(stmtf("%s.AuxInt = %sToAuxInt(%s)", v, unTitle(op.auxIntType()), auxint)) |
| } else { |
| rr.add(stmtf("%s.AuxInt = %s", v, auxint)) |
| } |
| } |
| if aux != "" { |
| if rr.Typed { |
| // Make sure aux value has the right type. |
| rr.add(stmtf("%s.Aux = %sToAux(%s)", v, unTitle(op.auxType()), aux)) |
| } else { |
| rr.add(stmtf("%s.Aux = %s", v, aux)) |
| } |
| } |
| all := new(strings.Builder) |
| for i, arg := range args { |
| x := genResult0(rr, arch, arg, false, move, pos, cse) |
| if i > 0 { |
| all.WriteString(", ") |
| } |
| all.WriteString(x) |
| } |
| switch len(args) { |
| case 0: |
| case 1: |
| rr.add(stmtf("%s.AddArg(%s)", v, all.String())) |
| default: |
| rr.add(stmtf("%s.AddArg%d(%s)", v, len(args), all.String())) |
| } |
| |
| if cse != nil { |
| cse[w] = v |
| } |
| return v |
| } |
| |
| func split(s string) []string { |
| var r []string |
| |
| outer: |
| for s != "" { |
| d := 0 // depth of ({[< |
| var open, close byte // opening and closing markers ({[< or )}]> |
| nonsp := false // found a non-space char so far |
| for i := 0; i < len(s); i++ { |
| switch { |
| case d == 0 && s[i] == '(': |
| open, close = '(', ')' |
| d++ |
| case d == 0 && s[i] == '<': |
| open, close = '<', '>' |
| d++ |
| case d == 0 && s[i] == '[': |
| open, close = '[', ']' |
| d++ |
| case d == 0 && s[i] == '{': |
| open, close = '{', '}' |
| d++ |
| case d == 0 && (s[i] == ' ' || s[i] == '\t'): |
| if nonsp { |
| r = append(r, strings.TrimSpace(s[:i])) |
| s = s[i:] |
| continue outer |
| } |
| case d > 0 && s[i] == open: |
| d++ |
| case d > 0 && s[i] == close: |
| d-- |
| default: |
| nonsp = true |
| } |
| } |
| if d != 0 { |
| log.Fatalf("imbalanced expression: %q", s) |
| } |
| if nonsp { |
| r = append(r, strings.TrimSpace(s)) |
| } |
| break |
| } |
| return r |
| } |
| |
| // isBlock reports whether this op is a block opcode. |
| func isBlock(name string, arch arch) bool { |
| for _, b := range genericBlocks { |
| if b.name == name { |
| return true |
| } |
| } |
| for _, b := range arch.blocks { |
| if b.name == name { |
| return true |
| } |
| } |
| return false |
| } |
| |
| func extract(val string) (op, typ, auxint, aux string, args []string) { |
| val = val[1 : len(val)-1] // remove () |
| |
| // Split val up into regions. |
| // Split by spaces/tabs, except those contained in (), {}, [], or <>. |
| s := split(val) |
| |
| // Extract restrictions and args. |
| op = s[0] |
| for _, a := range s[1:] { |
| switch a[0] { |
| case '<': |
| typ = a[1 : len(a)-1] // remove <> |
| case '[': |
| auxint = a[1 : len(a)-1] // remove [] |
| case '{': |
| aux = a[1 : len(a)-1] // remove {} |
| default: |
| args = append(args, a) |
| } |
| } |
| return |
| } |
| |
| // parseValue parses a parenthesized value from a rule. |
| // The value can be from the match or the result side. |
| // It returns the op and unparsed strings for typ, auxint, and aux restrictions and for all args. |
| // oparch is the architecture that op is located in, or "" for generic. |
| func parseValue(val string, arch arch, loc string) (op opData, oparch, typ, auxint, aux string, args []string) { |
| // Resolve the op. |
| var s string |
| s, typ, auxint, aux, args = extract(val) |
| |
| // match reports whether x is a good op to select. |
| // If strict is true, rule generation might succeed. |
| // If strict is false, rule generation has failed, |
| // but we're trying to generate a useful error. |
| // Doing strict=true then strict=false allows |
| // precise op matching while retaining good error messages. |
| match := func(x opData, strict bool, archname string) bool { |
| if x.name != s { |
| return false |
| } |
| if x.argLength != -1 && int(x.argLength) != len(args) && (len(args) != 1 || args[0] != "...") { |
| if strict { |
| return false |
| } |
| log.Printf("%s: op %s (%s) should have %d args, has %d", loc, s, archname, x.argLength, len(args)) |
| } |
| return true |
| } |
| |
| for _, x := range genericOps { |
| if match(x, true, "generic") { |
| op = x |
| break |
| } |
| } |
| for _, x := range arch.ops { |
| if arch.name != "generic" && match(x, true, arch.name) { |
| if op.name != "" { |
| log.Fatalf("%s: matches for op %s found in both generic and %s", loc, op.name, arch.name) |
| } |
| op = x |
| oparch = arch.name |
| break |
| } |
| } |
| |
| if op.name == "" { |
| // Failed to find the op. |
| // Run through everything again with strict=false |
| // to generate useful diagnosic messages before failing. |
| for _, x := range genericOps { |
| match(x, false, "generic") |
| } |
| for _, x := range arch.ops { |
| match(x, false, arch.name) |
| } |
| log.Fatalf("%s: unknown op %s", loc, s) |
| } |
| |
| // Sanity check aux, auxint. |
| if auxint != "" && !opHasAuxInt(op) { |
| log.Fatalf("%s: op %s %s can't have auxint", loc, op.name, op.aux) |
| } |
| if aux != "" && !opHasAux(op) { |
| log.Fatalf("%s: op %s %s can't have aux", loc, op.name, op.aux) |
| } |
| return |
| } |
| |
| func opHasAuxInt(op opData) bool { |
| switch op.aux { |
| case "Bool", "Int8", "Int16", "Int32", "Int64", "Int128", "Float32", "Float64", |
| "SymOff", "CallOff", "SymValAndOff", "TypSize", "ARM64BitField", "FlagConstant", "CCop": |
| return true |
| } |
| return false |
| } |
| |
| func opHasAux(op opData) bool { |
| switch op.aux { |
| case "String", "Sym", "SymOff", "Call", "CallOff", "SymValAndOff", "Typ", "TypSize", |
| "S390XCCMask", "S390XRotateParams": |
| return true |
| } |
| return false |
| } |
| |
| // splitNameExpr splits s-expr arg, possibly prefixed by "name:", |
| // into name and the unprefixed expression. |
| // For example, "x:(Foo)" yields "x", "(Foo)", |
| // and "(Foo)" yields "", "(Foo)". |
| func splitNameExpr(arg string) (name, expr string) { |
| colon := strings.Index(arg, ":") |
| if colon < 0 { |
| return "", arg |
| } |
| openparen := strings.Index(arg, "(") |
| if openparen < 0 { |
| log.Fatalf("splitNameExpr(%q): colon but no open parens", arg) |
| } |
| if colon > openparen { |
| // colon is inside the parens, such as in "(Foo x:(Bar))". |
| return "", arg |
| } |
| return arg[:colon], arg[colon+1:] |
| } |
| |
| func getBlockInfo(op string, arch arch) (name string, data blockData) { |
| for _, b := range genericBlocks { |
| if b.name == op { |
| return "Block" + op, b |
| } |
| } |
| for _, b := range arch.blocks { |
| if b.name == op { |
| return "Block" + arch.name + op, b |
| } |
| } |
| log.Fatalf("could not find block data for %s", op) |
| panic("unreachable") |
| } |
| |
| // typeName returns the string to use to generate a type. |
| func typeName(typ string) string { |
| if typ[0] == '(' { |
| ts := strings.Split(typ[1:len(typ)-1], ",") |
| if len(ts) != 2 { |
| log.Fatalf("Tuple expect 2 arguments") |
| } |
| return "types.NewTuple(" + typeName(ts[0]) + ", " + typeName(ts[1]) + ")" |
| } |
| switch typ { |
| case "Flags", "Mem", "Void", "Int128": |
| return "types.Type" + typ |
| default: |
| return "typ." + typ |
| } |
| } |
| |
| // balance returns the number of unclosed '(' characters in s. |
| // If a ')' appears without a corresponding '(', balance returns -1. |
| func balance(s string) int { |
| balance := 0 |
| for _, c := range s { |
| switch c { |
| case '(': |
| balance++ |
| case ')': |
| balance-- |
| if balance < 0 { |
| // don't allow ")(" to return 0 |
| return -1 |
| } |
| } |
| } |
| return balance |
| } |
| |
| // findAllOpcode is a function to find the opcode portion of s-expressions. |
| var findAllOpcode = regexp.MustCompile(`[(](\w+[|])+\w+[)]`).FindAllStringIndex |
| |
| // excludeFromExpansion reports whether the substring s[idx[0]:idx[1]] in a rule |
| // should be disregarded as a candidate for | expansion. |
| // It uses simple syntactic checks to see whether the substring |
| // is inside an AuxInt expression or inside the && conditions. |
| func excludeFromExpansion(s string, idx []int) bool { |
| left := s[:idx[0]] |
| if strings.LastIndexByte(left, '[') > strings.LastIndexByte(left, ']') { |
| // Inside an AuxInt expression. |
| return true |
| } |
| right := s[idx[1]:] |
| if strings.Contains(left, "&&") && (strings.Contains(right, "->") || strings.Contains(right, "=>")) { |
| // Inside && conditions. |
| return true |
| } |
| return false |
| } |
| |
| // expandOr converts a rule into multiple rules by expanding | ops. |
| func expandOr(r string) []string { |
| // Find every occurrence of |-separated things. |
| // They look like MOV(B|W|L|Q|SS|SD)load or MOV(Q|L)loadidx(1|8). |
| // Generate rules selecting one case from each |-form. |
| |
| // Count width of |-forms. They must match. |
| n := 1 |
| for _, idx := range findAllOpcode(r, -1) { |
| if excludeFromExpansion(r, idx) { |
| continue |
| } |
| s := r[idx[0]:idx[1]] |
| c := strings.Count(s, "|") + 1 |
| if c == 1 { |
| continue |
| } |
| if n > 1 && n != c { |
| log.Fatalf("'|' count doesn't match in %s: both %d and %d\n", r, n, c) |
| } |
| n = c |
| } |
| if n == 1 { |
| // No |-form in this rule. |
| return []string{r} |
| } |
| // Build each new rule. |
| res := make([]string, n) |
| for i := 0; i < n; i++ { |
| buf := new(strings.Builder) |
| x := 0 |
| for _, idx := range findAllOpcode(r, -1) { |
| if excludeFromExpansion(r, idx) { |
| continue |
| } |
| buf.WriteString(r[x:idx[0]]) // write bytes we've skipped over so far |
| s := r[idx[0]+1 : idx[1]-1] // remove leading "(" and trailing ")" |
| buf.WriteString(strings.Split(s, "|")[i]) // write the op component for this rule |
| x = idx[1] // note that we've written more bytes |
| } |
| buf.WriteString(r[x:]) |
| res[i] = buf.String() |
| } |
| return res |
| } |
| |
| // varCount returns a map which counts the number of occurrences of |
| // Value variables in the s-expression rr.Match and the Go expression rr.Cond. |
| func varCount(rr *RuleRewrite) map[string]int { |
| cnt := map[string]int{} |
| varCount1(rr.Loc, rr.Match, cnt) |
| if rr.Cond != "" { |
| expr, err := parser.ParseExpr(rr.Cond) |
| if err != nil { |
| log.Fatalf("%s: failed to parse cond %q: %v", rr.Loc, rr.Cond, err) |
| } |
| ast.Inspect(expr, func(n ast.Node) bool { |
| if id, ok := n.(*ast.Ident); ok { |
| cnt[id.Name]++ |
| } |
| return true |
| }) |
| } |
| return cnt |
| } |
| |
| func varCount1(loc, m string, cnt map[string]int) { |
| if m[0] == '<' || m[0] == '[' || m[0] == '{' { |
| return |
| } |
| if token.IsIdentifier(m) { |
| cnt[m]++ |
| return |
| } |
| // Split up input. |
| name, expr := splitNameExpr(m) |
| if name != "" { |
| cnt[name]++ |
| } |
| if expr[0] != '(' || expr[len(expr)-1] != ')' { |
| log.Fatalf("%s: non-compound expr in varCount1: %q", loc, expr) |
| } |
| s := split(expr[1 : len(expr)-1]) |
| for _, arg := range s[1:] { |
| varCount1(loc, arg, cnt) |
| } |
| } |
| |
| // normalizeWhitespace replaces 2+ whitespace sequences with a single space. |
| func normalizeWhitespace(x string) string { |
| x = strings.Join(strings.Fields(x), " ") |
| x = strings.Replace(x, "( ", "(", -1) |
| x = strings.Replace(x, " )", ")", -1) |
| x = strings.Replace(x, "[ ", "[", -1) |
| x = strings.Replace(x, " ]", "]", -1) |
| x = strings.Replace(x, ")->", ") ->", -1) |
| x = strings.Replace(x, ")=>", ") =>", -1) |
| return x |
| } |
| |
| // opIsCommutative reports whether op s is commutative. |
| func opIsCommutative(op string, arch arch) bool { |
| for _, x := range genericOps { |
| if op == x.name { |
| if x.commutative { |
| return true |
| } |
| break |
| } |
| } |
| if arch.name != "generic" { |
| for _, x := range arch.ops { |
| if op == x.name { |
| if x.commutative { |
| return true |
| } |
| break |
| } |
| } |
| } |
| return false |
| } |
| |
| func normalizeMatch(m string, arch arch) string { |
| if token.IsIdentifier(m) { |
| return m |
| } |
| op, typ, auxint, aux, args := extract(m) |
| if opIsCommutative(op, arch) { |
| if args[1] < args[0] { |
| args[0], args[1] = args[1], args[0] |
| } |
| } |
| s := new(strings.Builder) |
| fmt.Fprintf(s, "%s <%s> [%s] {%s}", op, typ, auxint, aux) |
| for _, arg := range args { |
| prefix, expr := splitNameExpr(arg) |
| fmt.Fprint(s, " ", prefix, normalizeMatch(expr, arch)) |
| } |
| return s.String() |
| } |
| |
| func parseEllipsisRules(rules []Rule, arch arch) (newop string, ok bool) { |
| if len(rules) != 1 { |
| for _, r := range rules { |
| if strings.Contains(r.Rule, "...") { |
| log.Fatalf("%s: found ellipsis in rule, but there are other rules with the same op", r.Loc) |
| } |
| } |
| return "", false |
| } |
| rule := rules[0] |
| match, cond, result, _ := rule.parse() |
| if cond != "" || !isEllipsisValue(match) || !isEllipsisValue(result) { |
| if strings.Contains(rule.Rule, "...") { |
| log.Fatalf("%s: found ellipsis in non-ellipsis rule", rule.Loc) |
| } |
| checkEllipsisRuleCandidate(rule, arch) |
| return "", false |
| } |
| op, oparch, _, _, _, _ := parseValue(result, arch, rule.Loc) |
| return fmt.Sprintf("Op%s%s", oparch, op.name), true |
| } |
| |
| // isEllipsisValue reports whether s is of the form (OpX ...). |
| func isEllipsisValue(s string) bool { |
| if len(s) < 2 || s[0] != '(' || s[len(s)-1] != ')' { |
| return false |
| } |
| c := split(s[1 : len(s)-1]) |
| if len(c) != 2 || c[1] != "..." { |
| return false |
| } |
| return true |
| } |
| |
| func checkEllipsisRuleCandidate(rule Rule, arch arch) { |
| match, cond, result, _ := rule.parse() |
| if cond != "" { |
| return |
| } |
| op, _, _, auxint, aux, args := parseValue(match, arch, rule.Loc) |
| var auxint2, aux2 string |
| var args2 []string |
| var usingCopy string |
| var eop opData |
| if result[0] != '(' { |
| // Check for (Foo x) -> x, which can be converted to (Foo ...) -> (Copy ...). |
| args2 = []string{result} |
| usingCopy = " using Copy" |
| } else { |
| eop, _, _, auxint2, aux2, args2 = parseValue(result, arch, rule.Loc) |
| } |
| // Check that all restrictions in match are reproduced exactly in result. |
| if aux != aux2 || auxint != auxint2 || len(args) != len(args2) { |
| return |
| } |
| if strings.Contains(rule.Rule, "=>") && op.aux != eop.aux { |
| return |
| } |
| for i := range args { |
| if args[i] != args2[i] { |
| return |
| } |
| } |
| switch { |
| case opHasAux(op) && aux == "" && aux2 == "": |
| fmt.Printf("%s: rule silently zeros aux, either copy aux or explicitly zero\n", rule.Loc) |
| case opHasAuxInt(op) && auxint == "" && auxint2 == "": |
| fmt.Printf("%s: rule silently zeros auxint, either copy auxint or explicitly zero\n", rule.Loc) |
| default: |
| fmt.Printf("%s: possible ellipsis rule candidate%s: %q\n", rule.Loc, usingCopy, rule.Rule) |
| } |
| } |
| |
| func opByName(arch arch, name string) opData { |
| name = name[2:] |
| for _, x := range genericOps { |
| if name == x.name { |
| return x |
| } |
| } |
| if arch.name != "generic" { |
| name = name[len(arch.name):] |
| for _, x := range arch.ops { |
| if name == x.name { |
| return x |
| } |
| } |
| } |
| log.Fatalf("failed to find op named %s in arch %s", name, arch.name) |
| panic("unreachable") |
| } |
| |
| // auxType returns the Go type that this operation should store in its aux field. |
| func (op opData) auxType() string { |
| switch op.aux { |
| case "String": |
| return "string" |
| case "Sym": |
| // Note: a Sym can be an *obj.LSym, a *gc.Node, or nil. |
| return "Sym" |
| case "SymOff": |
| return "Sym" |
| case "Call": |
| return "Call" |
| case "CallOff": |
| return "Call" |
| case "SymValAndOff": |
| return "Sym" |
| case "Typ": |
| return "*types.Type" |
| case "TypSize": |
| return "*types.Type" |
| case "S390XCCMask": |
| return "s390x.CCMask" |
| case "S390XRotateParams": |
| return "s390x.RotateParams" |
| default: |
| return "invalid" |
| } |
| } |
| |
| // auxIntType returns the Go type that this operation should store in its auxInt field. |
| func (op opData) auxIntType() string { |
| switch op.aux { |
| case "Bool": |
| return "bool" |
| case "Int8": |
| return "int8" |
| case "Int16": |
| return "int16" |
| case "Int32": |
| return "int32" |
| case "Int64": |
| return "int64" |
| case "Int128": |
| return "int128" |
| case "Float32": |
| return "float32" |
| case "Float64": |
| return "float64" |
| case "CallOff": |
| return "int32" |
| case "SymOff": |
| return "int32" |
| case "SymValAndOff": |
| return "ValAndOff" |
| case "TypSize": |
| return "int64" |
| case "CCop": |
| return "Op" |
| case "FlagConstant": |
| return "flagConstant" |
| case "ARM64BitField": |
| return "arm64BitField" |
| default: |
| return "invalid" |
| } |
| } |
| |
| // auxType returns the Go type that this block should store in its aux field. |
| func (b blockData) auxType() string { |
| switch b.aux { |
| case "S390XCCMask", "S390XCCMaskInt8", "S390XCCMaskUint8": |
| return "s390x.CCMask" |
| case "S390XRotateParams": |
| return "s390x.RotateParams" |
| default: |
| return "invalid" |
| } |
| } |
| |
| // auxIntType returns the Go type that this block should store in its auxInt field. |
| func (b blockData) auxIntType() string { |
| switch b.aux { |
| case "S390XCCMaskInt8": |
| return "int8" |
| case "S390XCCMaskUint8": |
| return "uint8" |
| case "Int64": |
| return "int64" |
| default: |
| return "invalid" |
| } |
| } |
| |
| func title(s string) string { |
| if i := strings.Index(s, "."); i >= 0 { |
| switch strings.ToLower(s[:i]) { |
| case "s390x": // keep arch prefix for clarity |
| s = s[:i] + s[i+1:] |
| default: |
| s = s[i+1:] |
| } |
| } |
| return strings.Title(s) |
| } |
| |
| func unTitle(s string) string { |
| if i := strings.Index(s, "."); i >= 0 { |
| switch strings.ToLower(s[:i]) { |
| case "s390x": // keep arch prefix for clarity |
| s = s[:i] + s[i+1:] |
| default: |
| s = s[i+1:] |
| } |
| } |
| return strings.ToLower(s[:1]) + s[1:] |
| } |