cmd/compile: remove support for untyped ssa rules

This change removes support in rulegen for untyped -> ssa rules.

Change-Id: I202018e191fc74f027243351bc8cf96145f2482c
Reviewed-on: https://go-review.googlesource.com/c/go/+/264679
Run-TryBot: Alberto Donizetti <alb.donizetti@gmail.com>
TryBot-Result: Go Bot <gobot@golang.org>
Reviewed-by: Keith Randall <khr@golang.org>
Trust: Alberto Donizetti <alb.donizetti@gmail.com>
This commit is contained in:
Alberto Donizetti 2020-10-24 17:27:52 +02:00
parent c515852732
commit 79a3482d9e
1 changed files with 22 additions and 65 deletions

View File

@ -35,8 +35,7 @@ import (
) )
// rule syntax: // rule syntax:
// sexpr [&& extra conditions] -> [@block] sexpr (untyped) // sexpr [&& extra conditions] => [@block] sexpr
// sexpr [&& extra conditions] => [@block] sexpr (typed)
// //
// sexpr are s-expressions (lisp-like parenthesized groupings) // sexpr are s-expressions (lisp-like parenthesized groupings)
// sexpr ::= [variable:](opcode sexpr*) // sexpr ::= [variable:](opcode sexpr*)
@ -79,14 +78,8 @@ func normalizeSpaces(s string) string {
} }
// parse returns the matching part of the rule, additional conditions, and the result. // parse returns the matching part of the rule, additional conditions, and the result.
// parse also reports whether the generated code should use strongly typed aux and auxint fields. func (r Rule) parse() (match, cond, result string) {
func (r Rule) parse() (match, cond, result string, typed bool) { s := strings.Split(r.Rule, "=>")
arrow := "->"
if strings.Contains(r.Rule, "=>") {
arrow = "=>"
typed = true
}
s := strings.Split(r.Rule, arrow)
match = normalizeSpaces(s[0]) match = normalizeSpaces(s[0])
result = normalizeSpaces(s[1]) result = normalizeSpaces(s[1])
cond = "" cond = ""
@ -94,7 +87,7 @@ func (r Rule) parse() (match, cond, result string, typed bool) {
cond = normalizeSpaces(match[i+2:]) cond = normalizeSpaces(match[i+2:])
match = normalizeSpaces(match[:i]) match = normalizeSpaces(match[:i])
} }
return match, cond, result, typed return match, cond, result
} }
func genRules(arch arch) { genRulesSuffix(arch, "") } func genRules(arch arch) { genRulesSuffix(arch, "") }
@ -120,7 +113,7 @@ func genRulesSuffix(arch arch, suff string) {
scanner := bufio.NewScanner(text) scanner := bufio.NewScanner(text)
rule := "" rule := ""
var lineno int var lineno int
var ruleLineno int // line number of "->" or "=>" var ruleLineno int // line number of "=>"
for scanner.Scan() { for scanner.Scan() {
lineno++ lineno++
line := scanner.Text() line := scanner.Text()
@ -134,13 +127,13 @@ func genRulesSuffix(arch arch, suff string) {
if rule == "" { if rule == "" {
continue continue
} }
if !strings.Contains(rule, "->") && !strings.Contains(rule, "=>") { if !strings.Contains(rule, "=>") {
continue continue
} }
if ruleLineno == 0 { if ruleLineno == 0 {
ruleLineno = lineno ruleLineno = lineno
} }
if strings.HasSuffix(rule, "->") || strings.HasSuffix(rule, "=>") { if strings.HasSuffix(rule, "=>") {
continue // continue on the next line continue // continue on the next line
} }
if n := balance(rule); n > 0 { if n := balance(rule); n > 0 {
@ -157,7 +150,7 @@ func genRulesSuffix(arch arch, suff string) {
continue continue
} }
// Do fancier value op matching. // Do fancier value op matching.
match, _, _, _ := r.parse() match, _, _ := r.parse()
op, oparch, _, _, _, _ := parseValue(match, arch, loc) op, oparch, _, _, _, _ := parseValue(match, arch, loc)
opname := fmt.Sprintf("Op%s%s", oparch, op.name) opname := fmt.Sprintf("Op%s%s", oparch, op.name)
oprules[opname] = append(oprules[opname], r) oprules[opname] = append(oprules[opname], r)
@ -231,7 +224,7 @@ func genRulesSuffix(arch arch, suff string) {
log.Fatalf("unconditional rule %s is followed by other rules", rr.Match) log.Fatalf("unconditional rule %s is followed by other rules", rr.Match)
} }
rr = &RuleRewrite{Loc: rule.Loc} rr = &RuleRewrite{Loc: rule.Loc}
rr.Match, rr.Cond, rr.Result, rr.Typed = rule.parse() rr.Match, rr.Cond, rr.Result = rule.parse()
pos, _ := genMatch(rr, arch, rr.Match, fn.ArgLen >= 0) pos, _ := genMatch(rr, arch, rr.Match, fn.ArgLen >= 0)
if pos == "" { if pos == "" {
pos = "v.Pos" pos = "v.Pos"
@ -790,7 +783,6 @@ type (
Alloc int // for unique var names Alloc int // for unique var names
Loc string // file name & line number of the original rule Loc string // file name & line number of the original rule
CommuteDepth int // used to track depth of commute loops CommuteDepth int // used to track depth of commute loops
Typed bool // aux and auxint fields should be strongly typed
} }
Declare struct { Declare struct {
Name string Name string
@ -844,7 +836,7 @@ func breakf(format string, a ...interface{}) *CondBreak {
func genBlockRewrite(rule Rule, arch arch, data blockData) *RuleRewrite { func genBlockRewrite(rule Rule, arch arch, data blockData) *RuleRewrite {
rr := &RuleRewrite{Loc: rule.Loc} rr := &RuleRewrite{Loc: rule.Loc}
rr.Match, rr.Cond, rr.Result, rr.Typed = rule.parse() rr.Match, rr.Cond, rr.Result = rule.parse()
_, _, auxint, aux, s := extract(rr.Match) // remove parens, then split _, _, auxint, aux, s := extract(rr.Match) // remove parens, then split
// check match of control values // check match of control values
@ -888,15 +880,6 @@ func genBlockRewrite(rule Rule, arch arch, data blockData) *RuleRewrite {
if e.name == "" { if e.name == "" {
continue continue
} }
if !rr.Typed {
if !token.IsIdentifier(e.name) || rr.declared(e.name) {
// code or variable
rr.add(breakf("b.%s != %s", e.field, e.name))
} else {
rr.add(declf(e.name, "b.%s", e.field))
}
continue
}
if e.dclType == "" { if e.dclType == "" {
log.Fatalf("op %s has no declared type for %s", data.name, e.field) log.Fatalf("op %s has no declared type for %s", data.name, e.field)
@ -965,20 +948,12 @@ func genBlockRewrite(rule Rule, arch arch, data blockData) *RuleRewrite {
} }
if auxint != "" { if auxint != "" {
if rr.Typed { // Make sure auxint value has the right type.
// Make sure auxint value has the right type. rr.add(stmtf("b.AuxInt = %sToAuxInt(%s)", unTitle(outdata.auxIntType()), auxint))
rr.add(stmtf("b.AuxInt = %sToAuxInt(%s)", unTitle(outdata.auxIntType()), auxint))
} else {
rr.add(stmtf("b.AuxInt = %s", auxint))
}
} }
if aux != "" { if aux != "" {
if rr.Typed { // Make sure aux value has the right type.
// Make sure aux value has the right type. rr.add(stmtf("b.Aux = %sToAux(%s)", unTitle(outdata.auxType()), aux))
rr.add(stmtf("b.Aux = %sToAux(%s)", unTitle(outdata.auxType()), aux))
} else {
rr.add(stmtf("b.Aux = %s", aux))
}
} }
succChanged := false succChanged := false
@ -1046,15 +1021,6 @@ func genMatch0(rr *RuleRewrite, arch arch, match, v string, cnt map[string]int,
if e.name == "" { if e.name == "" {
continue continue
} }
if !rr.Typed {
if !token.IsIdentifier(e.name) || rr.declared(e.name) {
// code or variable
rr.add(breakf("%s.%s != %s", v, e.field, e.name))
} else {
rr.add(declf(e.name, "%s.%s", v, e.field))
}
continue
}
if e.dclType == "" { if e.dclType == "" {
log.Fatalf("op %s has no declared type for %s", op.name, e.field) log.Fatalf("op %s has no declared type for %s", op.name, e.field)
@ -1244,20 +1210,12 @@ func genResult0(rr *RuleRewrite, arch arch, result string, top, move bool, pos s
} }
if auxint != "" { if auxint != "" {
if rr.Typed { // Make sure auxint value has the right type.
// Make sure auxint value has the right type. rr.add(stmtf("%s.AuxInt = %sToAuxInt(%s)", v, unTitle(op.auxIntType()), auxint))
rr.add(stmtf("%s.AuxInt = %sToAuxInt(%s)", v, unTitle(op.auxIntType()), auxint))
} else {
rr.add(stmtf("%s.AuxInt = %s", v, auxint))
}
} }
if aux != "" { if aux != "" {
if rr.Typed { // Make sure aux value has the right type.
// Make sure aux value has the right type. rr.add(stmtf("%s.Aux = %sToAux(%s)", v, unTitle(op.auxType()), aux))
rr.add(stmtf("%s.Aux = %sToAux(%s)", v, unTitle(op.auxType()), aux))
} else {
rr.add(stmtf("%s.Aux = %s", v, aux))
}
} }
all := new(strings.Builder) all := new(strings.Builder)
for i, arg := range args { for i, arg := range args {
@ -1538,7 +1496,7 @@ func excludeFromExpansion(s string, idx []int) bool {
return true return true
} }
right := s[idx[1]:] right := s[idx[1]:]
if strings.Contains(left, "&&") && (strings.Contains(right, "->") || strings.Contains(right, "=>")) { if strings.Contains(left, "&&") && strings.Contains(right, "=>") {
// Inside && conditions. // Inside && conditions.
return true return true
} }
@ -1640,7 +1598,6 @@ func normalizeWhitespace(x string) string {
x = strings.Replace(x, " )", ")", -1) x = strings.Replace(x, " )", ")", -1)
x = strings.Replace(x, "[ ", "[", -1) x = strings.Replace(x, "[ ", "[", -1)
x = strings.Replace(x, " ]", "]", -1) x = strings.Replace(x, " ]", "]", -1)
x = strings.Replace(x, ")->", ") ->", -1)
x = strings.Replace(x, ")=>", ") =>", -1) x = strings.Replace(x, ")=>", ") =>", -1)
return x return x
} }
@ -1697,7 +1654,7 @@ func parseEllipsisRules(rules []Rule, arch arch) (newop string, ok bool) {
return "", false return "", false
} }
rule := rules[0] rule := rules[0]
match, cond, result, _ := rule.parse() match, cond, result := rule.parse()
if cond != "" || !isEllipsisValue(match) || !isEllipsisValue(result) { if cond != "" || !isEllipsisValue(match) || !isEllipsisValue(result) {
if strings.Contains(rule.Rule, "...") { if strings.Contains(rule.Rule, "...") {
log.Fatalf("%s: found ellipsis in non-ellipsis rule", rule.Loc) log.Fatalf("%s: found ellipsis in non-ellipsis rule", rule.Loc)
@ -1722,7 +1679,7 @@ func isEllipsisValue(s string) bool {
} }
func checkEllipsisRuleCandidate(rule Rule, arch arch) { func checkEllipsisRuleCandidate(rule Rule, arch arch) {
match, cond, result, _ := rule.parse() match, cond, result := rule.parse()
if cond != "" { if cond != "" {
return return
} }
@ -1732,7 +1689,7 @@ func checkEllipsisRuleCandidate(rule Rule, arch arch) {
var usingCopy string var usingCopy string
var eop opData var eop opData
if result[0] != '(' { if result[0] != '(' {
// Check for (Foo x) -> x, which can be converted to (Foo ...) -> (Copy ...). // Check for (Foo x) => x, which can be converted to (Foo ...) => (Copy ...).
args2 = []string{result} args2 = []string{result}
usingCopy = " using Copy" usingCopy = " using Copy"
} else { } else {