| // Copyright 2009 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| package walk |
| |
| import ( |
| "go/constant" |
| "internal/abi" |
| |
| "cmd/compile/internal/base" |
| "cmd/compile/internal/ir" |
| "cmd/compile/internal/reflectdata" |
| "cmd/compile/internal/typecheck" |
| "cmd/compile/internal/types" |
| "cmd/internal/src" |
| ) |
| |
| // walkAssign walks an OAS (AssignExpr) or OASOP (AssignOpExpr) node. |
| func walkAssign(init *ir.Nodes, n ir.Node) ir.Node { |
| init.Append(ir.TakeInit(n)...) |
| |
| var left, right ir.Node |
| switch n.Op() { |
| case ir.OAS: |
| n := n.(*ir.AssignStmt) |
| left, right = n.X, n.Y |
| case ir.OASOP: |
| n := n.(*ir.AssignOpStmt) |
| left, right = n.X, n.Y |
| } |
| |
| // Recognize m[k] = append(m[k], ...) so we can reuse |
| // the mapassign call. |
| var mapAppend *ir.CallExpr |
| if left.Op() == ir.OINDEXMAP && right.Op() == ir.OAPPEND { |
| left := left.(*ir.IndexExpr) |
| mapAppend = right.(*ir.CallExpr) |
| if !ir.SameSafeExpr(left, mapAppend.Args[0]) { |
| base.Fatalf("not same expressions: %v != %v", left, mapAppend.Args[0]) |
| } |
| } |
| |
| left = walkExpr(left, init) |
| left = safeExpr(left, init) |
| if mapAppend != nil { |
| mapAppend.Args[0] = left |
| } |
| |
| if n.Op() == ir.OASOP { |
| // Rewrite x op= y into x = x op y. |
| n = ir.NewAssignStmt(base.Pos, left, typecheck.Expr(ir.NewBinaryExpr(base.Pos, n.(*ir.AssignOpStmt).AsOp, left, right))) |
| } else { |
| n.(*ir.AssignStmt).X = left |
| } |
| as := n.(*ir.AssignStmt) |
| |
| if oaslit(as, init) { |
| return ir.NewBlockStmt(as.Pos(), nil) |
| } |
| |
| if as.Y == nil { |
| // TODO(austin): Check all "implicit zeroing" |
| return as |
| } |
| |
| if !base.Flag.Cfg.Instrumenting && ir.IsZero(as.Y) { |
| return as |
| } |
| |
| switch as.Y.Op() { |
| default: |
| as.Y = walkExpr(as.Y, init) |
| |
| case ir.ORECV: |
| // x = <-c; as.Left is x, as.Right.Left is c. |
| // order.stmt made sure x is addressable. |
| recv := as.Y.(*ir.UnaryExpr) |
| recv.X = walkExpr(recv.X, init) |
| |
| n1 := typecheck.NodAddr(as.X) |
| r := recv.X // the channel |
| return mkcall1(chanfn("chanrecv1", 2, r.Type()), nil, init, r, n1) |
| |
| case ir.OAPPEND: |
| // x = append(...) |
| call := as.Y.(*ir.CallExpr) |
| if call.Type().Elem().NotInHeap() { |
| base.Errorf("%v can't be allocated in Go; it is incomplete (or unallocatable)", call.Type().Elem()) |
| } |
| var r ir.Node |
| switch { |
| case isAppendOfMake(call): |
| // x = append(y, make([]T, y)...) |
| r = extendSlice(call, init) |
| case call.IsDDD: |
| r = appendSlice(call, init) // also works for append(slice, string). |
| default: |
| r = walkAppend(call, init, as) |
| } |
| as.Y = r |
| if r.Op() == ir.OAPPEND { |
| r := r.(*ir.CallExpr) |
| // Left in place for back end. |
| // Do not add a new write barrier. |
| // Set up address of type for back end. |
| r.Fun = reflectdata.AppendElemRType(base.Pos, r) |
| return as |
| } |
| // Otherwise, lowered for race detector. |
| // Treat as ordinary assignment. |
| } |
| |
| if as.X != nil && as.Y != nil { |
| return convas(as, init) |
| } |
| return as |
| } |
| |
| // walkAssignDotType walks an OAS2DOTTYPE node. |
| func walkAssignDotType(n *ir.AssignListStmt, init *ir.Nodes) ir.Node { |
| walkExprListSafe(n.Lhs, init) |
| n.Rhs[0] = walkExpr(n.Rhs[0], init) |
| return n |
| } |
| |
| // walkAssignFunc walks an OAS2FUNC node. |
| func walkAssignFunc(init *ir.Nodes, n *ir.AssignListStmt) ir.Node { |
| init.Append(ir.TakeInit(n)...) |
| |
| r := n.Rhs[0] |
| walkExprListSafe(n.Lhs, init) |
| r = walkExpr(r, init) |
| |
| if ir.IsIntrinsicCall(r.(*ir.CallExpr)) { |
| n.Rhs = []ir.Node{r} |
| return n |
| } |
| init.Append(r) |
| |
| ll := ascompatet(n.Lhs, r.Type()) |
| return ir.NewBlockStmt(src.NoXPos, ll) |
| } |
| |
| // walkAssignList walks an OAS2 node. |
| func walkAssignList(init *ir.Nodes, n *ir.AssignListStmt) ir.Node { |
| init.Append(ir.TakeInit(n)...) |
| return ir.NewBlockStmt(src.NoXPos, ascompatee(ir.OAS, n.Lhs, n.Rhs)) |
| } |
| |
| // walkAssignMapRead walks an OAS2MAPR node. |
| func walkAssignMapRead(init *ir.Nodes, n *ir.AssignListStmt) ir.Node { |
| init.Append(ir.TakeInit(n)...) |
| |
| r := n.Rhs[0].(*ir.IndexExpr) |
| walkExprListSafe(n.Lhs, init) |
| r.X = walkExpr(r.X, init) |
| r.Index = walkExpr(r.Index, init) |
| t := r.X.Type() |
| |
| fast := mapfast(t) |
| key := mapKeyArg(fast, r, r.Index, false) |
| |
| // from: |
| // a,b = m[i] |
| // to: |
| // var,b = mapaccess2*(t, m, i) |
| // a = *var |
| a := n.Lhs[0] |
| |
| var call *ir.CallExpr |
| if w := t.Elem().Size(); w <= abi.ZeroValSize { |
| fn := mapfn(mapaccess2[fast], t, false) |
| call = mkcall1(fn, fn.Type().ResultsTuple(), init, reflectdata.IndexMapRType(base.Pos, r), r.X, key) |
| } else { |
| fn := mapfn("mapaccess2_fat", t, true) |
| z := reflectdata.ZeroAddr(w) |
| call = mkcall1(fn, fn.Type().ResultsTuple(), init, reflectdata.IndexMapRType(base.Pos, r), r.X, key, z) |
| } |
| |
| // mapaccess2* returns a typed bool, but due to spec changes, |
| // the boolean result of i.(T) is now untyped so we make it the |
| // same type as the variable on the lhs. |
| if ok := n.Lhs[1]; !ir.IsBlank(ok) && ok.Type().IsBoolean() { |
| call.Type().Field(1).Type = ok.Type() |
| } |
| n.Rhs = []ir.Node{call} |
| n.SetOp(ir.OAS2FUNC) |
| |
| // don't generate a = *var if a is _ |
| if ir.IsBlank(a) { |
| return walkExpr(typecheck.Stmt(n), init) |
| } |
| |
| var_ := typecheck.TempAt(base.Pos, ir.CurFunc, types.NewPtr(t.Elem())) |
| var_.SetTypecheck(1) |
| var_.MarkNonNil() // mapaccess always returns a non-nil pointer |
| |
| n.Lhs[0] = var_ |
| init.Append(walkExpr(n, init)) |
| |
| as := ir.NewAssignStmt(base.Pos, a, ir.NewStarExpr(base.Pos, var_)) |
| return walkExpr(typecheck.Stmt(as), init) |
| } |
| |
| // walkAssignRecv walks an OAS2RECV node. |
| func walkAssignRecv(init *ir.Nodes, n *ir.AssignListStmt) ir.Node { |
| init.Append(ir.TakeInit(n)...) |
| |
| r := n.Rhs[0].(*ir.UnaryExpr) // recv |
| walkExprListSafe(n.Lhs, init) |
| r.X = walkExpr(r.X, init) |
| var n1 ir.Node |
| if ir.IsBlank(n.Lhs[0]) { |
| n1 = typecheck.NodNil() |
| } else { |
| n1 = typecheck.NodAddr(n.Lhs[0]) |
| } |
| fn := chanfn("chanrecv2", 2, r.X.Type()) |
| ok := n.Lhs[1] |
| call := mkcall1(fn, types.Types[types.TBOOL], init, r.X, n1) |
| return typecheck.Stmt(ir.NewAssignStmt(base.Pos, ok, call)) |
| } |
| |
| // walkReturn walks an ORETURN node. |
| func walkReturn(n *ir.ReturnStmt) ir.Node { |
| fn := ir.CurFunc |
| |
| fn.NumReturns++ |
| if len(n.Results) == 0 { |
| return n |
| } |
| |
| results := fn.Type().Results() |
| dsts := make([]ir.Node, len(results)) |
| for i, v := range results { |
| // TODO(mdempsky): typecheck should have already checked the result variables. |
| dsts[i] = typecheck.AssignExpr(v.Nname.(*ir.Name)) |
| } |
| |
| n.Results = ascompatee(n.Op(), dsts, n.Results) |
| return n |
| } |
| |
| // check assign type list to |
| // an expression list. called in |
| // |
| // expr-list = func() |
| func ascompatet(nl ir.Nodes, nr *types.Type) []ir.Node { |
| if len(nl) != nr.NumFields() { |
| base.Fatalf("ascompatet: assignment count mismatch: %d = %d", len(nl), nr.NumFields()) |
| } |
| |
| var nn ir.Nodes |
| for i, l := range nl { |
| if ir.IsBlank(l) { |
| continue |
| } |
| r := nr.Field(i) |
| |
| // Order should have created autotemps of the appropriate type for |
| // us to store results into. |
| if tmp, ok := l.(*ir.Name); !ok || !tmp.AutoTemp() || !types.Identical(tmp.Type(), r.Type) { |
| base.FatalfAt(l.Pos(), "assigning %v to %+v", r.Type, l) |
| } |
| |
| res := ir.NewResultExpr(base.Pos, nil, types.BADWIDTH) |
| res.Index = int64(i) |
| res.SetType(r.Type) |
| res.SetTypecheck(1) |
| |
| nn.Append(ir.NewAssignStmt(base.Pos, l, res)) |
| } |
| return nn |
| } |
| |
| // check assign expression list to |
| // an expression list. called in |
| // |
| // expr-list = expr-list |
| func ascompatee(op ir.Op, nl, nr []ir.Node) []ir.Node { |
| // cannot happen: should have been rejected during type checking |
| if len(nl) != len(nr) { |
| base.Fatalf("assignment operands mismatch: %+v / %+v", ir.Nodes(nl), ir.Nodes(nr)) |
| } |
| |
| var assigned ir.NameSet |
| var memWrite, deferResultWrite bool |
| |
| // affected reports whether expression n could be affected by |
| // the assignments applied so far. |
| affected := func(n ir.Node) bool { |
| if deferResultWrite { |
| return true |
| } |
| return ir.Any(n, func(n ir.Node) bool { |
| if n.Op() == ir.ONAME && assigned.Has(n.(*ir.Name)) { |
| return true |
| } |
| if memWrite && readsMemory(n) { |
| return true |
| } |
| return false |
| }) |
| } |
| |
| // If a needed expression may be affected by an |
| // earlier assignment, make an early copy of that |
| // expression and use the copy instead. |
| var early ir.Nodes |
| save := func(np *ir.Node) { |
| if n := *np; affected(n) { |
| *np = copyExpr(n, n.Type(), &early) |
| } |
| } |
| |
| var late ir.Nodes |
| for i, lorig := range nl { |
| l, r := lorig, nr[i] |
| |
| // Do not generate 'x = x' during return. See issue 4014. |
| if op == ir.ORETURN && ir.SameSafeExpr(l, r) { |
| continue |
| } |
| |
| // Save subexpressions needed on left side. |
| // Drill through non-dereferences. |
| for { |
| // If an expression has init statements, they must be evaluated |
| // before any of its saved sub-operands (#45706). |
| // TODO(mdempsky): Disallow init statements on lvalues. |
| init := ir.TakeInit(l) |
| walkStmtList(init) |
| early.Append(init...) |
| |
| switch ll := l.(type) { |
| case *ir.IndexExpr: |
| if ll.X.Type().IsArray() { |
| save(&ll.Index) |
| l = ll.X |
| continue |
| } |
| case *ir.ParenExpr: |
| l = ll.X |
| continue |
| case *ir.SelectorExpr: |
| if ll.Op() == ir.ODOT { |
| l = ll.X |
| continue |
| } |
| } |
| break |
| } |
| |
| var name *ir.Name |
| switch l.Op() { |
| default: |
| base.Fatalf("unexpected lvalue %v", l.Op()) |
| case ir.ONAME: |
| name = l.(*ir.Name) |
| case ir.OINDEX, ir.OINDEXMAP: |
| l := l.(*ir.IndexExpr) |
| save(&l.X) |
| save(&l.Index) |
| case ir.ODEREF: |
| l := l.(*ir.StarExpr) |
| save(&l.X) |
| case ir.ODOTPTR: |
| l := l.(*ir.SelectorExpr) |
| save(&l.X) |
| } |
| |
| // Save expression on right side. |
| save(&r) |
| |
| appendWalkStmt(&late, convas(ir.NewAssignStmt(base.Pos, lorig, r), &late)) |
| |
| // Check for reasons why we may need to compute later expressions |
| // before this assignment happens. |
| |
| if name == nil { |
| // Not a direct assignment to a declared variable. |
| // Conservatively assume any memory access might alias. |
| memWrite = true |
| continue |
| } |
| |
| if name.Class == ir.PPARAMOUT && ir.CurFunc.HasDefer() { |
| // Assignments to a result parameter in a function with defers |
| // becomes visible early if evaluation of any later expression |
| // panics (#43835). |
| deferResultWrite = true |
| continue |
| } |
| |
| if ir.IsBlank(name) { |
| // We can ignore assignments to blank or anonymous result parameters. |
| // These can't appear in expressions anyway. |
| continue |
| } |
| |
| if name.Addrtaken() || !name.OnStack() { |
| // Global variable, heap escaped, or just addrtaken. |
| // Conservatively assume any memory access might alias. |
| memWrite = true |
| continue |
| } |
| |
| // Local, non-addrtaken variable. |
| // Assignments can only alias with direct uses of this variable. |
| assigned.Add(name) |
| } |
| |
| early.Append(late.Take()...) |
| return early |
| } |
| |
| // readsMemory reports whether the evaluation n directly reads from |
| // memory that might be written to indirectly. |
| func readsMemory(n ir.Node) bool { |
| switch n.Op() { |
| case ir.ONAME: |
| n := n.(*ir.Name) |
| if n.Class == ir.PFUNC { |
| return false |
| } |
| return n.Addrtaken() || !n.OnStack() |
| |
| case ir.OADD, |
| ir.OAND, |
| ir.OANDAND, |
| ir.OANDNOT, |
| ir.OBITNOT, |
| ir.OCONV, |
| ir.OCONVIFACE, |
| ir.OCONVNOP, |
| ir.ODIV, |
| ir.ODOT, |
| ir.ODOTTYPE, |
| ir.OLITERAL, |
| ir.OLSH, |
| ir.OMOD, |
| ir.OMUL, |
| ir.ONEG, |
| ir.ONIL, |
| ir.OOR, |
| ir.OOROR, |
| ir.OPAREN, |
| ir.OPLUS, |
| ir.ORSH, |
| ir.OSUB, |
| ir.OXOR: |
| return false |
| } |
| |
| // Be conservative. |
| return true |
| } |
| |
| // expand append(l1, l2...) to |
| // |
| // init { |
| // s := l1 |
| // newLen := s.len + l2.len |
| // // Compare as uint so growslice can panic on overflow. |
| // if uint(newLen) <= uint(s.cap) { |
| // s = s[:newLen] |
| // } else { |
| // s = growslice(s.ptr, s.len, s.cap, l2.len, T) |
| // } |
| // memmove(&s[s.len-l2.len], &l2[0], l2.len*sizeof(T)) |
| // } |
| // s |
| // |
| // l2 is allowed to be a string. |
| func appendSlice(n *ir.CallExpr, init *ir.Nodes) ir.Node { |
| walkAppendArgs(n, init) |
| |
| l1 := n.Args[0] |
| l2 := n.Args[1] |
| l2 = cheapExpr(l2, init) |
| n.Args[1] = l2 |
| |
| var nodes ir.Nodes |
| |
| // var s []T |
| s := typecheck.TempAt(base.Pos, ir.CurFunc, l1.Type()) |
| nodes.Append(ir.NewAssignStmt(base.Pos, s, l1)) // s = l1 |
| |
| elemtype := s.Type().Elem() |
| |
| // Decompose slice. |
| oldPtr := ir.NewUnaryExpr(base.Pos, ir.OSPTR, s) |
| oldLen := ir.NewUnaryExpr(base.Pos, ir.OLEN, s) |
| oldCap := ir.NewUnaryExpr(base.Pos, ir.OCAP, s) |
| |
| // Number of elements we are adding |
| num := ir.NewUnaryExpr(base.Pos, ir.OLEN, l2) |
| |
| // newLen := oldLen + num |
| newLen := typecheck.TempAt(base.Pos, ir.CurFunc, types.Types[types.TINT]) |
| nodes.Append(ir.NewAssignStmt(base.Pos, newLen, ir.NewBinaryExpr(base.Pos, ir.OADD, oldLen, num))) |
| |
| // if uint(newLen) <= uint(oldCap) |
| nif := ir.NewIfStmt(base.Pos, nil, nil, nil) |
| nuint := typecheck.Conv(newLen, types.Types[types.TUINT]) |
| scapuint := typecheck.Conv(oldCap, types.Types[types.TUINT]) |
| nif.Cond = ir.NewBinaryExpr(base.Pos, ir.OLE, nuint, scapuint) |
| nif.Likely = true |
| |
| // then { s = s[:newLen] } |
| slice := ir.NewSliceExpr(base.Pos, ir.OSLICE, s, nil, newLen, nil) |
| slice.SetBounded(true) |
| nif.Body = []ir.Node{ir.NewAssignStmt(base.Pos, s, slice)} |
| |
| // else { s = growslice(oldPtr, newLen, oldCap, num, T) } |
| call := walkGrowslice(s, nif.PtrInit(), oldPtr, newLen, oldCap, num) |
| nif.Else = []ir.Node{ir.NewAssignStmt(base.Pos, s, call)} |
| |
| nodes.Append(nif) |
| |
| // Index to start copying into s. |
| // idx = newLen - len(l2) |
| // We use this expression instead of oldLen because it avoids |
| // a spill/restore of oldLen. |
| // Note: this doesn't work optimally currently because |
| // the compiler optimizer undoes this arithmetic. |
| idx := ir.NewBinaryExpr(base.Pos, ir.OSUB, newLen, ir.NewUnaryExpr(base.Pos, ir.OLEN, l2)) |
| |
| var ncopy ir.Node |
| if elemtype.HasPointers() { |
| // copy(s[idx:], l2) |
| slice := ir.NewSliceExpr(base.Pos, ir.OSLICE, s, idx, nil, nil) |
| slice.SetType(s.Type()) |
| slice.SetBounded(true) |
| |
| ir.CurFunc.SetWBPos(n.Pos()) |
| |
| // instantiate typedslicecopy(typ *type, dstPtr *any, dstLen int, srcPtr *any, srcLen int) int |
| fn := typecheck.LookupRuntime("typedslicecopy", l1.Type().Elem(), l2.Type().Elem()) |
| ptr1, len1 := backingArrayPtrLen(cheapExpr(slice, &nodes)) |
| ptr2, len2 := backingArrayPtrLen(l2) |
| ncopy = mkcall1(fn, types.Types[types.TINT], &nodes, reflectdata.AppendElemRType(base.Pos, n), ptr1, len1, ptr2, len2) |
| } else if base.Flag.Cfg.Instrumenting && !base.Flag.CompilingRuntime { |
| // rely on runtime to instrument: |
| // copy(s[idx:], l2) |
| // l2 can be a slice or string. |
| slice := ir.NewSliceExpr(base.Pos, ir.OSLICE, s, idx, nil, nil) |
| slice.SetType(s.Type()) |
| slice.SetBounded(true) |
| |
| ptr1, len1 := backingArrayPtrLen(cheapExpr(slice, &nodes)) |
| ptr2, len2 := backingArrayPtrLen(l2) |
| |
| fn := typecheck.LookupRuntime("slicecopy", ptr1.Type().Elem(), ptr2.Type().Elem()) |
| ncopy = mkcall1(fn, types.Types[types.TINT], &nodes, ptr1, len1, ptr2, len2, ir.NewInt(base.Pos, elemtype.Size())) |
| } else { |
| // memmove(&s[idx], &l2[0], len(l2)*sizeof(T)) |
| ix := ir.NewIndexExpr(base.Pos, s, idx) |
| ix.SetBounded(true) |
| addr := typecheck.NodAddr(ix) |
| |
| sptr := ir.NewUnaryExpr(base.Pos, ir.OSPTR, l2) |
| |
| nwid := cheapExpr(typecheck.Conv(ir.NewUnaryExpr(base.Pos, ir.OLEN, l2), types.Types[types.TUINTPTR]), &nodes) |
| nwid = ir.NewBinaryExpr(base.Pos, ir.OMUL, nwid, ir.NewInt(base.Pos, elemtype.Size())) |
| |
| // instantiate func memmove(to *any, frm *any, length uintptr) |
| fn := typecheck.LookupRuntime("memmove", elemtype, elemtype) |
| ncopy = mkcall1(fn, nil, &nodes, addr, sptr, nwid) |
| } |
| ln := append(nodes, ncopy) |
| |
| typecheck.Stmts(ln) |
| walkStmtList(ln) |
| init.Append(ln...) |
| return s |
| } |
| |
| // isAppendOfMake reports whether n is of the form append(x, make([]T, y)...). |
| // isAppendOfMake assumes n has already been typechecked. |
| func isAppendOfMake(n ir.Node) bool { |
| if base.Flag.N != 0 || base.Flag.Cfg.Instrumenting { |
| return false |
| } |
| |
| if n.Typecheck() == 0 { |
| base.Fatalf("missing typecheck: %+v", n) |
| } |
| |
| if n.Op() != ir.OAPPEND { |
| return false |
| } |
| call := n.(*ir.CallExpr) |
| if !call.IsDDD || len(call.Args) != 2 || call.Args[1].Op() != ir.OMAKESLICE { |
| return false |
| } |
| |
| mk := call.Args[1].(*ir.MakeExpr) |
| if mk.Cap != nil { |
| return false |
| } |
| |
| // y must be either an integer constant or the largest possible positive value |
| // of variable y needs to fit into a uint. |
| |
| // typecheck made sure that constant arguments to make are not negative and fit into an int. |
| |
| // The care of overflow of the len argument to make will be handled by an explicit check of int(len) < 0 during runtime. |
| y := mk.Len |
| if !ir.IsConst(y, constant.Int) && y.Type().Size() > types.Types[types.TUINT].Size() { |
| return false |
| } |
| |
| return true |
| } |
| |
| // extendSlice rewrites append(l1, make([]T, l2)...) to |
| // |
| // init { |
| // if l2 >= 0 { // Empty if block here for more meaningful node.SetLikely(true) |
| // } else { |
| // panicmakeslicelen() |
| // } |
| // s := l1 |
| // n := len(s) + l2 |
| // // Compare n and s as uint so growslice can panic on overflow of len(s) + l2. |
| // // cap is a positive int and n can become negative when len(s) + l2 |
| // // overflows int. Interpreting n when negative as uint makes it larger |
| // // than cap(s). growslice will check the int n arg and panic if n is |
| // // negative. This prevents the overflow from being undetected. |
| // if uint(n) <= uint(cap(s)) { |
| // s = s[:n] |
| // } else { |
| // s = growslice(T, s.ptr, n, s.cap, l2, T) |
| // } |
| // // clear the new portion of the underlying array. |
| // hp := &s[len(s)-l2] |
| // hn := l2 * sizeof(T) |
| // memclr(hp, hn) |
| // } |
| // s |
| // |
| // if T has pointers, the final memclr can go inside the "then" branch, as |
| // growslice will have done the clearing for us. |
| |
| func extendSlice(n *ir.CallExpr, init *ir.Nodes) ir.Node { |
| // isAppendOfMake made sure all possible positive values of l2 fit into a uint. |
| // The case of l2 overflow when converting from e.g. uint to int is handled by an explicit |
| // check of l2 < 0 at runtime which is generated below. |
| l2 := typecheck.Conv(n.Args[1].(*ir.MakeExpr).Len, types.Types[types.TINT]) |
| l2 = typecheck.Expr(l2) |
| n.Args[1] = l2 // walkAppendArgs expects l2 in n.List.Second(). |
| |
| walkAppendArgs(n, init) |
| |
| l1 := n.Args[0] |
| l2 = n.Args[1] // re-read l2, as it may have been updated by walkAppendArgs |
| |
| var nodes []ir.Node |
| |
| // if l2 >= 0 (likely happens), do nothing |
| nifneg := ir.NewIfStmt(base.Pos, ir.NewBinaryExpr(base.Pos, ir.OGE, l2, ir.NewInt(base.Pos, 0)), nil, nil) |
| nifneg.Likely = true |
| |
| // else panicmakeslicelen() |
| nifneg.Else = []ir.Node{mkcall("panicmakeslicelen", nil, init)} |
| nodes = append(nodes, nifneg) |
| |
| // s := l1 |
| s := typecheck.TempAt(base.Pos, ir.CurFunc, l1.Type()) |
| nodes = append(nodes, ir.NewAssignStmt(base.Pos, s, l1)) |
| |
| elemtype := s.Type().Elem() |
| |
| // n := s.len + l2 |
| nn := typecheck.TempAt(base.Pos, ir.CurFunc, types.Types[types.TINT]) |
| nodes = append(nodes, ir.NewAssignStmt(base.Pos, nn, ir.NewBinaryExpr(base.Pos, ir.OADD, ir.NewUnaryExpr(base.Pos, ir.OLEN, s), l2))) |
| |
| // if uint(n) <= uint(s.cap) |
| nuint := typecheck.Conv(nn, types.Types[types.TUINT]) |
| capuint := typecheck.Conv(ir.NewUnaryExpr(base.Pos, ir.OCAP, s), types.Types[types.TUINT]) |
| nif := ir.NewIfStmt(base.Pos, ir.NewBinaryExpr(base.Pos, ir.OLE, nuint, capuint), nil, nil) |
| nif.Likely = true |
| |
| // then { s = s[:n] } |
| nt := ir.NewSliceExpr(base.Pos, ir.OSLICE, s, nil, nn, nil) |
| nt.SetBounded(true) |
| nif.Body = []ir.Node{ir.NewAssignStmt(base.Pos, s, nt)} |
| |
| // else { s = growslice(s.ptr, n, s.cap, l2, T) } |
| nif.Else = []ir.Node{ |
| ir.NewAssignStmt(base.Pos, s, walkGrowslice(s, nif.PtrInit(), |
| ir.NewUnaryExpr(base.Pos, ir.OSPTR, s), |
| nn, |
| ir.NewUnaryExpr(base.Pos, ir.OCAP, s), |
| l2)), |
| } |
| |
| nodes = append(nodes, nif) |
| |
| // hp := &s[s.len - l2] |
| // TODO: &s[s.len] - hn? |
| ix := ir.NewIndexExpr(base.Pos, s, ir.NewBinaryExpr(base.Pos, ir.OSUB, ir.NewUnaryExpr(base.Pos, ir.OLEN, s), l2)) |
| ix.SetBounded(true) |
| hp := typecheck.ConvNop(typecheck.NodAddr(ix), types.Types[types.TUNSAFEPTR]) |
| |
| // hn := l2 * sizeof(elem(s)) |
| hn := typecheck.Conv(ir.NewBinaryExpr(base.Pos, ir.OMUL, l2, ir.NewInt(base.Pos, elemtype.Size())), types.Types[types.TUINTPTR]) |
| |
| clrname := "memclrNoHeapPointers" |
| hasPointers := elemtype.HasPointers() |
| if hasPointers { |
| clrname = "memclrHasPointers" |
| ir.CurFunc.SetWBPos(n.Pos()) |
| } |
| |
| var clr ir.Nodes |
| clrfn := mkcall(clrname, nil, &clr, hp, hn) |
| clr.Append(clrfn) |
| if hasPointers { |
| // growslice will have cleared the new entries, so only |
| // if growslice isn't called do we need to do the zeroing ourselves. |
| nif.Body = append(nif.Body, clr...) |
| } else { |
| nodes = append(nodes, clr...) |
| } |
| |
| typecheck.Stmts(nodes) |
| walkStmtList(nodes) |
| init.Append(nodes...) |
| return s |
| } |