| // Code generated from gen/dec64.rules; DO NOT EDIT. |
| // generated with: cd gen; go run *.go |
| |
| package ssa |
| |
| import "fmt" |
| import "math" |
| import "cmd/internal/obj" |
| import "cmd/internal/objabi" |
| import "cmd/compile/internal/types" |
| |
| var _ = fmt.Println // in case not otherwise used |
| var _ = math.MinInt8 // in case not otherwise used |
| var _ = obj.ANOP // in case not otherwise used |
| var _ = objabi.GOROOT // in case not otherwise used |
| var _ = types.TypeMem // in case not otherwise used |
| |
| func rewriteValuedec64(v *Value) bool { |
| switch v.Op { |
| case OpAdd64: |
| return rewriteValuedec64_OpAdd64_0(v) |
| case OpAnd64: |
| return rewriteValuedec64_OpAnd64_0(v) |
| case OpArg: |
| return rewriteValuedec64_OpArg_0(v) |
| case OpBitLen64: |
| return rewriteValuedec64_OpBitLen64_0(v) |
| case OpBswap64: |
| return rewriteValuedec64_OpBswap64_0(v) |
| case OpCom64: |
| return rewriteValuedec64_OpCom64_0(v) |
| case OpConst64: |
| return rewriteValuedec64_OpConst64_0(v) |
| case OpCtz64: |
| return rewriteValuedec64_OpCtz64_0(v) |
| case OpCtz64NonZero: |
| return rewriteValuedec64_OpCtz64NonZero_0(v) |
| case OpEq64: |
| return rewriteValuedec64_OpEq64_0(v) |
| case OpGeq64: |
| return rewriteValuedec64_OpGeq64_0(v) |
| case OpGeq64U: |
| return rewriteValuedec64_OpGeq64U_0(v) |
| case OpGreater64: |
| return rewriteValuedec64_OpGreater64_0(v) |
| case OpGreater64U: |
| return rewriteValuedec64_OpGreater64U_0(v) |
| case OpInt64Hi: |
| return rewriteValuedec64_OpInt64Hi_0(v) |
| case OpInt64Lo: |
| return rewriteValuedec64_OpInt64Lo_0(v) |
| case OpLeq64: |
| return rewriteValuedec64_OpLeq64_0(v) |
| case OpLeq64U: |
| return rewriteValuedec64_OpLeq64U_0(v) |
| case OpLess64: |
| return rewriteValuedec64_OpLess64_0(v) |
| case OpLess64U: |
| return rewriteValuedec64_OpLess64U_0(v) |
| case OpLoad: |
| return rewriteValuedec64_OpLoad_0(v) |
| case OpLsh16x64: |
| return rewriteValuedec64_OpLsh16x64_0(v) |
| case OpLsh32x64: |
| return rewriteValuedec64_OpLsh32x64_0(v) |
| case OpLsh64x16: |
| return rewriteValuedec64_OpLsh64x16_0(v) |
| case OpLsh64x32: |
| return rewriteValuedec64_OpLsh64x32_0(v) |
| case OpLsh64x64: |
| return rewriteValuedec64_OpLsh64x64_0(v) |
| case OpLsh64x8: |
| return rewriteValuedec64_OpLsh64x8_0(v) |
| case OpLsh8x64: |
| return rewriteValuedec64_OpLsh8x64_0(v) |
| case OpMul64: |
| return rewriteValuedec64_OpMul64_0(v) |
| case OpNeg64: |
| return rewriteValuedec64_OpNeg64_0(v) |
| case OpNeq64: |
| return rewriteValuedec64_OpNeq64_0(v) |
| case OpOr64: |
| return rewriteValuedec64_OpOr64_0(v) |
| case OpRsh16Ux64: |
| return rewriteValuedec64_OpRsh16Ux64_0(v) |
| case OpRsh16x64: |
| return rewriteValuedec64_OpRsh16x64_0(v) |
| case OpRsh32Ux64: |
| return rewriteValuedec64_OpRsh32Ux64_0(v) |
| case OpRsh32x64: |
| return rewriteValuedec64_OpRsh32x64_0(v) |
| case OpRsh64Ux16: |
| return rewriteValuedec64_OpRsh64Ux16_0(v) |
| case OpRsh64Ux32: |
| return rewriteValuedec64_OpRsh64Ux32_0(v) |
| case OpRsh64Ux64: |
| return rewriteValuedec64_OpRsh64Ux64_0(v) |
| case OpRsh64Ux8: |
| return rewriteValuedec64_OpRsh64Ux8_0(v) |
| case OpRsh64x16: |
| return rewriteValuedec64_OpRsh64x16_0(v) |
| case OpRsh64x32: |
| return rewriteValuedec64_OpRsh64x32_0(v) |
| case OpRsh64x64: |
| return rewriteValuedec64_OpRsh64x64_0(v) |
| case OpRsh64x8: |
| return rewriteValuedec64_OpRsh64x8_0(v) |
| case OpRsh8Ux64: |
| return rewriteValuedec64_OpRsh8Ux64_0(v) |
| case OpRsh8x64: |
| return rewriteValuedec64_OpRsh8x64_0(v) |
| case OpSignExt16to64: |
| return rewriteValuedec64_OpSignExt16to64_0(v) |
| case OpSignExt32to64: |
| return rewriteValuedec64_OpSignExt32to64_0(v) |
| case OpSignExt8to64: |
| return rewriteValuedec64_OpSignExt8to64_0(v) |
| case OpStore: |
| return rewriteValuedec64_OpStore_0(v) |
| case OpSub64: |
| return rewriteValuedec64_OpSub64_0(v) |
| case OpTrunc64to16: |
| return rewriteValuedec64_OpTrunc64to16_0(v) |
| case OpTrunc64to32: |
| return rewriteValuedec64_OpTrunc64to32_0(v) |
| case OpTrunc64to8: |
| return rewriteValuedec64_OpTrunc64to8_0(v) |
| case OpXor64: |
| return rewriteValuedec64_OpXor64_0(v) |
| case OpZeroExt16to64: |
| return rewriteValuedec64_OpZeroExt16to64_0(v) |
| case OpZeroExt32to64: |
| return rewriteValuedec64_OpZeroExt32to64_0(v) |
| case OpZeroExt8to64: |
| return rewriteValuedec64_OpZeroExt8to64_0(v) |
| } |
| return false |
| } |
| func rewriteValuedec64_OpAdd64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Add64 x y) |
| // cond: |
| // result: (Int64Make (Add32withcarry <typ.Int32> (Int64Hi x) (Int64Hi y) (Select1 <types.TypeFlags> (Add32carry (Int64Lo x) (Int64Lo y)))) (Select0 <typ.UInt32> (Add32carry (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpAdd32withcarry, typ.Int32) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v3 := b.NewValue0(v.Pos, OpSelect1, types.TypeFlags) |
| v4 := b.NewValue0(v.Pos, OpAdd32carry, types.NewTuple(typ.UInt32, types.TypeFlags)) |
| v5 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v0.AddArg(v3) |
| v.AddArg(v0) |
| v7 := b.NewValue0(v.Pos, OpSelect0, typ.UInt32) |
| v8 := b.NewValue0(v.Pos, OpAdd32carry, types.NewTuple(typ.UInt32, types.TypeFlags)) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(x) |
| v8.AddArg(v9) |
| v10 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v10.AddArg(y) |
| v8.AddArg(v10) |
| v7.AddArg(v8) |
| v.AddArg(v7) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpAnd64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (And64 x y) |
| // cond: |
| // result: (Int64Make (And32 <typ.UInt32> (Int64Hi x) (Int64Hi y)) (And32 <typ.UInt32> (Int64Lo x) (Int64Lo y))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpAnd32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpAnd32, typ.UInt32) |
| v4 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v4.AddArg(x) |
| v3.AddArg(v4) |
| v5 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v5.AddArg(y) |
| v3.AddArg(v5) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpArg_0(v *Value) bool { |
| b := v.Block |
| config := b.Func.Config |
| typ := &b.Func.Config.Types |
| // match: (Arg {n} [off]) |
| // cond: is64BitInt(v.Type) && !config.BigEndian && v.Type.IsSigned() |
| // result: (Int64Make (Arg <typ.Int32> {n} [off+4]) (Arg <typ.UInt32> {n} [off])) |
| for { |
| off := v.AuxInt |
| n := v.Aux |
| if !(is64BitInt(v.Type) && !config.BigEndian && v.Type.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpArg, typ.Int32) |
| v0.AuxInt = off + 4 |
| v0.Aux = n |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpArg, typ.UInt32) |
| v1.AuxInt = off |
| v1.Aux = n |
| v.AddArg(v1) |
| return true |
| } |
| // match: (Arg {n} [off]) |
| // cond: is64BitInt(v.Type) && !config.BigEndian && !v.Type.IsSigned() |
| // result: (Int64Make (Arg <typ.UInt32> {n} [off+4]) (Arg <typ.UInt32> {n} [off])) |
| for { |
| off := v.AuxInt |
| n := v.Aux |
| if !(is64BitInt(v.Type) && !config.BigEndian && !v.Type.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpArg, typ.UInt32) |
| v0.AuxInt = off + 4 |
| v0.Aux = n |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpArg, typ.UInt32) |
| v1.AuxInt = off |
| v1.Aux = n |
| v.AddArg(v1) |
| return true |
| } |
| // match: (Arg {n} [off]) |
| // cond: is64BitInt(v.Type) && config.BigEndian && v.Type.IsSigned() |
| // result: (Int64Make (Arg <typ.Int32> {n} [off]) (Arg <typ.UInt32> {n} [off+4])) |
| for { |
| off := v.AuxInt |
| n := v.Aux |
| if !(is64BitInt(v.Type) && config.BigEndian && v.Type.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpArg, typ.Int32) |
| v0.AuxInt = off |
| v0.Aux = n |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpArg, typ.UInt32) |
| v1.AuxInt = off + 4 |
| v1.Aux = n |
| v.AddArg(v1) |
| return true |
| } |
| // match: (Arg {n} [off]) |
| // cond: is64BitInt(v.Type) && config.BigEndian && !v.Type.IsSigned() |
| // result: (Int64Make (Arg <typ.UInt32> {n} [off]) (Arg <typ.UInt32> {n} [off+4])) |
| for { |
| off := v.AuxInt |
| n := v.Aux |
| if !(is64BitInt(v.Type) && config.BigEndian && !v.Type.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpArg, typ.UInt32) |
| v0.AuxInt = off |
| v0.Aux = n |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpArg, typ.UInt32) |
| v1.AuxInt = off + 4 |
| v1.Aux = n |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpBitLen64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (BitLen64 x) |
| // cond: |
| // result: (Add32 <typ.Int> (BitLen32 <typ.Int> (Int64Hi x)) (BitLen32 <typ.Int> (Or32 <typ.UInt32> (Int64Lo x) (Zeromask (Int64Hi x))))) |
| for { |
| x := v.Args[0] |
| v.reset(OpAdd32) |
| v.Type = typ.Int |
| v0 := b.NewValue0(v.Pos, OpBitLen32, typ.Int) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v.AddArg(v0) |
| v2 := b.NewValue0(v.Pos, OpBitLen32, typ.Int) |
| v3 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v4 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v4.AddArg(x) |
| v3.AddArg(v4) |
| v5 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(x) |
| v5.AddArg(v6) |
| v3.AddArg(v5) |
| v2.AddArg(v3) |
| v.AddArg(v2) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpBswap64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Bswap64 x) |
| // cond: |
| // result: (Int64Make (Bswap32 <typ.UInt32> (Int64Lo x)) (Bswap32 <typ.UInt32> (Int64Hi x))) |
| for { |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpBswap32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v.AddArg(v0) |
| v2 := b.NewValue0(v.Pos, OpBswap32, typ.UInt32) |
| v3 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v3.AddArg(x) |
| v2.AddArg(v3) |
| v.AddArg(v2) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpCom64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Com64 x) |
| // cond: |
| // result: (Int64Make (Com32 <typ.UInt32> (Int64Hi x)) (Com32 <typ.UInt32> (Int64Lo x))) |
| for { |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpCom32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v.AddArg(v0) |
| v2 := b.NewValue0(v.Pos, OpCom32, typ.UInt32) |
| v3 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v3.AddArg(x) |
| v2.AddArg(v3) |
| v.AddArg(v2) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpConst64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Const64 <t> [c]) |
| // cond: t.IsSigned() |
| // result: (Int64Make (Const32 <typ.Int32> [c>>32]) (Const32 <typ.UInt32> [int64(int32(c))])) |
| for { |
| t := v.Type |
| c := v.AuxInt |
| if !(t.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpConst32, typ.Int32) |
| v0.AuxInt = c >> 32 |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v1.AuxInt = int64(int32(c)) |
| v.AddArg(v1) |
| return true |
| } |
| // match: (Const64 <t> [c]) |
| // cond: !t.IsSigned() |
| // result: (Int64Make (Const32 <typ.UInt32> [c>>32]) (Const32 <typ.UInt32> [int64(int32(c))])) |
| for { |
| t := v.Type |
| c := v.AuxInt |
| if !(!t.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v0.AuxInt = c >> 32 |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v1.AuxInt = int64(int32(c)) |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpCtz64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Ctz64 x) |
| // cond: |
| // result: (Add32 <typ.UInt32> (Ctz32 <typ.UInt32> (Int64Lo x)) (And32 <typ.UInt32> (Com32 <typ.UInt32> (Zeromask (Int64Lo x))) (Ctz32 <typ.UInt32> (Int64Hi x)))) |
| for { |
| x := v.Args[0] |
| v.reset(OpAdd32) |
| v.Type = typ.UInt32 |
| v0 := b.NewValue0(v.Pos, OpCtz32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v.AddArg(v0) |
| v2 := b.NewValue0(v.Pos, OpAnd32, typ.UInt32) |
| v3 := b.NewValue0(v.Pos, OpCom32, typ.UInt32) |
| v4 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v5 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v3.AddArg(v4) |
| v2.AddArg(v3) |
| v6 := b.NewValue0(v.Pos, OpCtz32, typ.UInt32) |
| v7 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v7.AddArg(x) |
| v6.AddArg(v7) |
| v2.AddArg(v6) |
| v.AddArg(v2) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpCtz64NonZero_0(v *Value) bool { |
| // match: (Ctz64NonZero x) |
| // cond: |
| // result: (Ctz64 x) |
| for { |
| x := v.Args[0] |
| v.reset(OpCtz64) |
| v.AddArg(x) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpEq64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Eq64 x y) |
| // cond: |
| // result: (AndB (Eq32 (Int64Hi x) (Int64Hi y)) (Eq32 (Int64Lo x) (Int64Lo y))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpAndB) |
| v0 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v4.AddArg(x) |
| v3.AddArg(v4) |
| v5 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v5.AddArg(y) |
| v3.AddArg(v5) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpGeq64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Geq64 x y) |
| // cond: |
| // result: (OrB (Greater32 (Int64Hi x) (Int64Hi y)) (AndB (Eq32 (Int64Hi x) (Int64Hi y)) (Geq32U (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpOrB) |
| v0 := b.NewValue0(v.Pos, OpGreater32, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpAndB, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v5 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v7 := b.NewValue0(v.Pos, OpGeq32U, typ.Bool) |
| v8 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v8.AddArg(x) |
| v7.AddArg(v8) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(y) |
| v7.AddArg(v9) |
| v3.AddArg(v7) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpGeq64U_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Geq64U x y) |
| // cond: |
| // result: (OrB (Greater32U (Int64Hi x) (Int64Hi y)) (AndB (Eq32 (Int64Hi x) (Int64Hi y)) (Geq32U (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpOrB) |
| v0 := b.NewValue0(v.Pos, OpGreater32U, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpAndB, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v5 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v7 := b.NewValue0(v.Pos, OpGeq32U, typ.Bool) |
| v8 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v8.AddArg(x) |
| v7.AddArg(v8) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(y) |
| v7.AddArg(v9) |
| v3.AddArg(v7) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpGreater64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Greater64 x y) |
| // cond: |
| // result: (OrB (Greater32 (Int64Hi x) (Int64Hi y)) (AndB (Eq32 (Int64Hi x) (Int64Hi y)) (Greater32U (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpOrB) |
| v0 := b.NewValue0(v.Pos, OpGreater32, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpAndB, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v5 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v7 := b.NewValue0(v.Pos, OpGreater32U, typ.Bool) |
| v8 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v8.AddArg(x) |
| v7.AddArg(v8) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(y) |
| v7.AddArg(v9) |
| v3.AddArg(v7) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpGreater64U_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Greater64U x y) |
| // cond: |
| // result: (OrB (Greater32U (Int64Hi x) (Int64Hi y)) (AndB (Eq32 (Int64Hi x) (Int64Hi y)) (Greater32U (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpOrB) |
| v0 := b.NewValue0(v.Pos, OpGreater32U, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpAndB, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v5 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v7 := b.NewValue0(v.Pos, OpGreater32U, typ.Bool) |
| v8 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v8.AddArg(x) |
| v7.AddArg(v8) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(y) |
| v7.AddArg(v9) |
| v3.AddArg(v7) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpInt64Hi_0(v *Value) bool { |
| // match: (Int64Hi (Int64Make hi _)) |
| // cond: |
| // result: hi |
| for { |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| _ = v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpCopy) |
| v.Type = hi.Type |
| v.AddArg(hi) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpInt64Lo_0(v *Value) bool { |
| // match: (Int64Lo (Int64Make _ lo)) |
| // cond: |
| // result: lo |
| for { |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| v.reset(OpCopy) |
| v.Type = lo.Type |
| v.AddArg(lo) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpLeq64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Leq64 x y) |
| // cond: |
| // result: (OrB (Less32 (Int64Hi x) (Int64Hi y)) (AndB (Eq32 (Int64Hi x) (Int64Hi y)) (Leq32U (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpOrB) |
| v0 := b.NewValue0(v.Pos, OpLess32, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpAndB, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v5 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v7 := b.NewValue0(v.Pos, OpLeq32U, typ.Bool) |
| v8 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v8.AddArg(x) |
| v7.AddArg(v8) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(y) |
| v7.AddArg(v9) |
| v3.AddArg(v7) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpLeq64U_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Leq64U x y) |
| // cond: |
| // result: (OrB (Less32U (Int64Hi x) (Int64Hi y)) (AndB (Eq32 (Int64Hi x) (Int64Hi y)) (Leq32U (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpOrB) |
| v0 := b.NewValue0(v.Pos, OpLess32U, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpAndB, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v5 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v7 := b.NewValue0(v.Pos, OpLeq32U, typ.Bool) |
| v8 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v8.AddArg(x) |
| v7.AddArg(v8) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(y) |
| v7.AddArg(v9) |
| v3.AddArg(v7) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpLess64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Less64 x y) |
| // cond: |
| // result: (OrB (Less32 (Int64Hi x) (Int64Hi y)) (AndB (Eq32 (Int64Hi x) (Int64Hi y)) (Less32U (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpOrB) |
| v0 := b.NewValue0(v.Pos, OpLess32, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpAndB, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v5 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v7 := b.NewValue0(v.Pos, OpLess32U, typ.Bool) |
| v8 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v8.AddArg(x) |
| v7.AddArg(v8) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(y) |
| v7.AddArg(v9) |
| v3.AddArg(v7) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpLess64U_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Less64U x y) |
| // cond: |
| // result: (OrB (Less32U (Int64Hi x) (Int64Hi y)) (AndB (Eq32 (Int64Hi x) (Int64Hi y)) (Less32U (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpOrB) |
| v0 := b.NewValue0(v.Pos, OpLess32U, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpAndB, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpEq32, typ.Bool) |
| v5 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v7 := b.NewValue0(v.Pos, OpLess32U, typ.Bool) |
| v8 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v8.AddArg(x) |
| v7.AddArg(v8) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(y) |
| v7.AddArg(v9) |
| v3.AddArg(v7) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpLoad_0(v *Value) bool { |
| b := v.Block |
| config := b.Func.Config |
| typ := &b.Func.Config.Types |
| // match: (Load <t> ptr mem) |
| // cond: is64BitInt(t) && !config.BigEndian && t.IsSigned() |
| // result: (Int64Make (Load <typ.Int32> (OffPtr <typ.Int32Ptr> [4] ptr) mem) (Load <typ.UInt32> ptr mem)) |
| for { |
| t := v.Type |
| mem := v.Args[1] |
| ptr := v.Args[0] |
| if !(is64BitInt(t) && !config.BigEndian && t.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpLoad, typ.Int32) |
| v1 := b.NewValue0(v.Pos, OpOffPtr, typ.Int32Ptr) |
| v1.AuxInt = 4 |
| v1.AddArg(ptr) |
| v0.AddArg(v1) |
| v0.AddArg(mem) |
| v.AddArg(v0) |
| v2 := b.NewValue0(v.Pos, OpLoad, typ.UInt32) |
| v2.AddArg(ptr) |
| v2.AddArg(mem) |
| v.AddArg(v2) |
| return true |
| } |
| // match: (Load <t> ptr mem) |
| // cond: is64BitInt(t) && !config.BigEndian && !t.IsSigned() |
| // result: (Int64Make (Load <typ.UInt32> (OffPtr <typ.UInt32Ptr> [4] ptr) mem) (Load <typ.UInt32> ptr mem)) |
| for { |
| t := v.Type |
| mem := v.Args[1] |
| ptr := v.Args[0] |
| if !(is64BitInt(t) && !config.BigEndian && !t.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpLoad, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpOffPtr, typ.UInt32Ptr) |
| v1.AuxInt = 4 |
| v1.AddArg(ptr) |
| v0.AddArg(v1) |
| v0.AddArg(mem) |
| v.AddArg(v0) |
| v2 := b.NewValue0(v.Pos, OpLoad, typ.UInt32) |
| v2.AddArg(ptr) |
| v2.AddArg(mem) |
| v.AddArg(v2) |
| return true |
| } |
| // match: (Load <t> ptr mem) |
| // cond: is64BitInt(t) && config.BigEndian && t.IsSigned() |
| // result: (Int64Make (Load <typ.Int32> ptr mem) (Load <typ.UInt32> (OffPtr <typ.UInt32Ptr> [4] ptr) mem)) |
| for { |
| t := v.Type |
| mem := v.Args[1] |
| ptr := v.Args[0] |
| if !(is64BitInt(t) && config.BigEndian && t.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpLoad, typ.Int32) |
| v0.AddArg(ptr) |
| v0.AddArg(mem) |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpLoad, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpOffPtr, typ.UInt32Ptr) |
| v2.AuxInt = 4 |
| v2.AddArg(ptr) |
| v1.AddArg(v2) |
| v1.AddArg(mem) |
| v.AddArg(v1) |
| return true |
| } |
| // match: (Load <t> ptr mem) |
| // cond: is64BitInt(t) && config.BigEndian && !t.IsSigned() |
| // result: (Int64Make (Load <typ.UInt32> ptr mem) (Load <typ.UInt32> (OffPtr <typ.UInt32Ptr> [4] ptr) mem)) |
| for { |
| t := v.Type |
| mem := v.Args[1] |
| ptr := v.Args[0] |
| if !(is64BitInt(t) && config.BigEndian && !t.IsSigned()) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpLoad, typ.UInt32) |
| v0.AddArg(ptr) |
| v0.AddArg(mem) |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpLoad, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpOffPtr, typ.UInt32Ptr) |
| v2.AuxInt = 4 |
| v2.AddArg(ptr) |
| v1.AddArg(v2) |
| v1.AddArg(mem) |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpLsh16x64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Lsh16x64 _ (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Const32 [0]) |
| for { |
| _ = v.Args[1] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpConst32) |
| v.AuxInt = 0 |
| return true |
| } |
| // match: (Lsh16x64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Lsh16x32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpLsh16x32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Lsh16x64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Lsh16x32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpLsh16x32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpLsh32x64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Lsh32x64 _ (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Const32 [0]) |
| for { |
| _ = v.Args[1] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpConst32) |
| v.AuxInt = 0 |
| return true |
| } |
| // match: (Lsh32x64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Lsh32x32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpLsh32x32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Lsh32x64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Lsh32x32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpLsh32x32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpLsh64x16_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Lsh64x16 (Int64Make hi lo) s) |
| // cond: |
| // result: (Int64Make (Or32 <typ.UInt32> (Or32 <typ.UInt32> (Lsh32x16 <typ.UInt32> hi s) (Rsh32Ux16 <typ.UInt32> lo (Sub16 <typ.UInt16> (Const16 <typ.UInt16> [32]) s))) (Lsh32x16 <typ.UInt32> lo (Sub16 <typ.UInt16> s (Const16 <typ.UInt16> [32])))) (Lsh32x16 <typ.UInt32> lo s)) |
| for { |
| s := v.Args[1] |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpLsh32x16, typ.UInt32) |
| v2.AddArg(hi) |
| v2.AddArg(s) |
| v1.AddArg(v2) |
| v3 := b.NewValue0(v.Pos, OpRsh32Ux16, typ.UInt32) |
| v3.AddArg(lo) |
| v4 := b.NewValue0(v.Pos, OpSub16, typ.UInt16) |
| v5 := b.NewValue0(v.Pos, OpConst16, typ.UInt16) |
| v5.AuxInt = 32 |
| v4.AddArg(v5) |
| v4.AddArg(s) |
| v3.AddArg(v4) |
| v1.AddArg(v3) |
| v0.AddArg(v1) |
| v6 := b.NewValue0(v.Pos, OpLsh32x16, typ.UInt32) |
| v6.AddArg(lo) |
| v7 := b.NewValue0(v.Pos, OpSub16, typ.UInt16) |
| v7.AddArg(s) |
| v8 := b.NewValue0(v.Pos, OpConst16, typ.UInt16) |
| v8.AuxInt = 32 |
| v7.AddArg(v8) |
| v6.AddArg(v7) |
| v0.AddArg(v6) |
| v.AddArg(v0) |
| v9 := b.NewValue0(v.Pos, OpLsh32x16, typ.UInt32) |
| v9.AddArg(lo) |
| v9.AddArg(s) |
| v.AddArg(v9) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpLsh64x32_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Lsh64x32 (Int64Make hi lo) s) |
| // cond: |
| // result: (Int64Make (Or32 <typ.UInt32> (Or32 <typ.UInt32> (Lsh32x32 <typ.UInt32> hi s) (Rsh32Ux32 <typ.UInt32> lo (Sub32 <typ.UInt32> (Const32 <typ.UInt32> [32]) s))) (Lsh32x32 <typ.UInt32> lo (Sub32 <typ.UInt32> s (Const32 <typ.UInt32> [32])))) (Lsh32x32 <typ.UInt32> lo s)) |
| for { |
| s := v.Args[1] |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpLsh32x32, typ.UInt32) |
| v2.AddArg(hi) |
| v2.AddArg(s) |
| v1.AddArg(v2) |
| v3 := b.NewValue0(v.Pos, OpRsh32Ux32, typ.UInt32) |
| v3.AddArg(lo) |
| v4 := b.NewValue0(v.Pos, OpSub32, typ.UInt32) |
| v5 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v5.AuxInt = 32 |
| v4.AddArg(v5) |
| v4.AddArg(s) |
| v3.AddArg(v4) |
| v1.AddArg(v3) |
| v0.AddArg(v1) |
| v6 := b.NewValue0(v.Pos, OpLsh32x32, typ.UInt32) |
| v6.AddArg(lo) |
| v7 := b.NewValue0(v.Pos, OpSub32, typ.UInt32) |
| v7.AddArg(s) |
| v8 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v8.AuxInt = 32 |
| v7.AddArg(v8) |
| v6.AddArg(v7) |
| v0.AddArg(v6) |
| v.AddArg(v0) |
| v9 := b.NewValue0(v.Pos, OpLsh32x32, typ.UInt32) |
| v9.AddArg(lo) |
| v9.AddArg(s) |
| v.AddArg(v9) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpLsh64x64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Lsh64x64 _ (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Const64 [0]) |
| for { |
| _ = v.Args[1] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpConst64) |
| v.AuxInt = 0 |
| return true |
| } |
| // match: (Lsh64x64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Lsh64x32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpLsh64x32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Lsh64x64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Lsh64x32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpLsh64x32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpLsh64x8_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Lsh64x8 (Int64Make hi lo) s) |
| // cond: |
| // result: (Int64Make (Or32 <typ.UInt32> (Or32 <typ.UInt32> (Lsh32x8 <typ.UInt32> hi s) (Rsh32Ux8 <typ.UInt32> lo (Sub8 <typ.UInt8> (Const8 <typ.UInt8> [32]) s))) (Lsh32x8 <typ.UInt32> lo (Sub8 <typ.UInt8> s (Const8 <typ.UInt8> [32])))) (Lsh32x8 <typ.UInt32> lo s)) |
| for { |
| s := v.Args[1] |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpLsh32x8, typ.UInt32) |
| v2.AddArg(hi) |
| v2.AddArg(s) |
| v1.AddArg(v2) |
| v3 := b.NewValue0(v.Pos, OpRsh32Ux8, typ.UInt32) |
| v3.AddArg(lo) |
| v4 := b.NewValue0(v.Pos, OpSub8, typ.UInt8) |
| v5 := b.NewValue0(v.Pos, OpConst8, typ.UInt8) |
| v5.AuxInt = 32 |
| v4.AddArg(v5) |
| v4.AddArg(s) |
| v3.AddArg(v4) |
| v1.AddArg(v3) |
| v0.AddArg(v1) |
| v6 := b.NewValue0(v.Pos, OpLsh32x8, typ.UInt32) |
| v6.AddArg(lo) |
| v7 := b.NewValue0(v.Pos, OpSub8, typ.UInt8) |
| v7.AddArg(s) |
| v8 := b.NewValue0(v.Pos, OpConst8, typ.UInt8) |
| v8.AuxInt = 32 |
| v7.AddArg(v8) |
| v6.AddArg(v7) |
| v0.AddArg(v6) |
| v.AddArg(v0) |
| v9 := b.NewValue0(v.Pos, OpLsh32x8, typ.UInt32) |
| v9.AddArg(lo) |
| v9.AddArg(s) |
| v.AddArg(v9) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpLsh8x64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Lsh8x64 _ (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Const32 [0]) |
| for { |
| _ = v.Args[1] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpConst32) |
| v.AuxInt = 0 |
| return true |
| } |
| // match: (Lsh8x64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Lsh8x32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpLsh8x32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Lsh8x64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Lsh8x32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpLsh8x32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpMul64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Mul64 x y) |
| // cond: |
| // result: (Int64Make (Add32 <typ.UInt32> (Mul32 <typ.UInt32> (Int64Lo x) (Int64Hi y)) (Add32 <typ.UInt32> (Mul32 <typ.UInt32> (Int64Hi x) (Int64Lo y)) (Select0 <typ.UInt32> (Mul32uhilo (Int64Lo x) (Int64Lo y))))) (Select1 <typ.UInt32> (Mul32uhilo (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpAdd32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpMul32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v2.AddArg(x) |
| v1.AddArg(v2) |
| v3 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v3.AddArg(y) |
| v1.AddArg(v3) |
| v0.AddArg(v1) |
| v4 := b.NewValue0(v.Pos, OpAdd32, typ.UInt32) |
| v5 := b.NewValue0(v.Pos, OpMul32, typ.UInt32) |
| v6 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v6.AddArg(x) |
| v5.AddArg(v6) |
| v7 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v7.AddArg(y) |
| v5.AddArg(v7) |
| v4.AddArg(v5) |
| v8 := b.NewValue0(v.Pos, OpSelect0, typ.UInt32) |
| v9 := b.NewValue0(v.Pos, OpMul32uhilo, types.NewTuple(typ.UInt32, typ.UInt32)) |
| v10 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v10.AddArg(x) |
| v9.AddArg(v10) |
| v11 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v11.AddArg(y) |
| v9.AddArg(v11) |
| v8.AddArg(v9) |
| v4.AddArg(v8) |
| v0.AddArg(v4) |
| v.AddArg(v0) |
| v12 := b.NewValue0(v.Pos, OpSelect1, typ.UInt32) |
| v13 := b.NewValue0(v.Pos, OpMul32uhilo, types.NewTuple(typ.UInt32, typ.UInt32)) |
| v14 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v14.AddArg(x) |
| v13.AddArg(v14) |
| v15 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v15.AddArg(y) |
| v13.AddArg(v15) |
| v12.AddArg(v13) |
| v.AddArg(v12) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpNeg64_0(v *Value) bool { |
| b := v.Block |
| // match: (Neg64 <t> x) |
| // cond: |
| // result: (Sub64 (Const64 <t> [0]) x) |
| for { |
| t := v.Type |
| x := v.Args[0] |
| v.reset(OpSub64) |
| v0 := b.NewValue0(v.Pos, OpConst64, t) |
| v0.AuxInt = 0 |
| v.AddArg(v0) |
| v.AddArg(x) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpNeq64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Neq64 x y) |
| // cond: |
| // result: (OrB (Neq32 (Int64Hi x) (Int64Hi y)) (Neq32 (Int64Lo x) (Int64Lo y))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpOrB) |
| v0 := b.NewValue0(v.Pos, OpNeq32, typ.Bool) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpNeq32, typ.Bool) |
| v4 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v4.AddArg(x) |
| v3.AddArg(v4) |
| v5 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v5.AddArg(y) |
| v3.AddArg(v5) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpOr64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Or64 x y) |
| // cond: |
| // result: (Int64Make (Or32 <typ.UInt32> (Int64Hi x) (Int64Hi y)) (Or32 <typ.UInt32> (Int64Lo x) (Int64Lo y))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v4 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v4.AddArg(x) |
| v3.AddArg(v4) |
| v5 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v5.AddArg(y) |
| v3.AddArg(v5) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpRsh16Ux64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh16Ux64 _ (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Const32 [0]) |
| for { |
| _ = v.Args[1] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpConst32) |
| v.AuxInt = 0 |
| return true |
| } |
| // match: (Rsh16Ux64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Rsh16Ux32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpRsh16Ux32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Rsh16Ux64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Rsh16Ux32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpRsh16Ux32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh16x64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh16x64 x (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Signmask (SignExt16to32 x)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpSignmask) |
| v0 := b.NewValue0(v.Pos, OpSignExt16to32, typ.Int32) |
| v0.AddArg(x) |
| v.AddArg(v0) |
| return true |
| } |
| // match: (Rsh16x64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Rsh16x32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpRsh16x32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Rsh16x64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Rsh16x32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpRsh16x32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh32Ux64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh32Ux64 _ (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Const32 [0]) |
| for { |
| _ = v.Args[1] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpConst32) |
| v.AuxInt = 0 |
| return true |
| } |
| // match: (Rsh32Ux64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Rsh32Ux32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpRsh32Ux32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Rsh32Ux64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Rsh32Ux32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpRsh32Ux32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh32x64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh32x64 x (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Signmask x) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpSignmask) |
| v.AddArg(x) |
| return true |
| } |
| // match: (Rsh32x64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Rsh32x32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpRsh32x32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Rsh32x64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Rsh32x32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpRsh32x32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh64Ux16_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh64Ux16 (Int64Make hi lo) s) |
| // cond: |
| // result: (Int64Make (Rsh32Ux16 <typ.UInt32> hi s) (Or32 <typ.UInt32> (Or32 <typ.UInt32> (Rsh32Ux16 <typ.UInt32> lo s) (Lsh32x16 <typ.UInt32> hi (Sub16 <typ.UInt16> (Const16 <typ.UInt16> [32]) s))) (Rsh32Ux16 <typ.UInt32> hi (Sub16 <typ.UInt16> s (Const16 <typ.UInt16> [32]))))) |
| for { |
| s := v.Args[1] |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpRsh32Ux16, typ.UInt32) |
| v0.AddArg(hi) |
| v0.AddArg(s) |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v3 := b.NewValue0(v.Pos, OpRsh32Ux16, typ.UInt32) |
| v3.AddArg(lo) |
| v3.AddArg(s) |
| v2.AddArg(v3) |
| v4 := b.NewValue0(v.Pos, OpLsh32x16, typ.UInt32) |
| v4.AddArg(hi) |
| v5 := b.NewValue0(v.Pos, OpSub16, typ.UInt16) |
| v6 := b.NewValue0(v.Pos, OpConst16, typ.UInt16) |
| v6.AuxInt = 32 |
| v5.AddArg(v6) |
| v5.AddArg(s) |
| v4.AddArg(v5) |
| v2.AddArg(v4) |
| v1.AddArg(v2) |
| v7 := b.NewValue0(v.Pos, OpRsh32Ux16, typ.UInt32) |
| v7.AddArg(hi) |
| v8 := b.NewValue0(v.Pos, OpSub16, typ.UInt16) |
| v8.AddArg(s) |
| v9 := b.NewValue0(v.Pos, OpConst16, typ.UInt16) |
| v9.AuxInt = 32 |
| v8.AddArg(v9) |
| v7.AddArg(v8) |
| v1.AddArg(v7) |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh64Ux32_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh64Ux32 (Int64Make hi lo) s) |
| // cond: |
| // result: (Int64Make (Rsh32Ux32 <typ.UInt32> hi s) (Or32 <typ.UInt32> (Or32 <typ.UInt32> (Rsh32Ux32 <typ.UInt32> lo s) (Lsh32x32 <typ.UInt32> hi (Sub32 <typ.UInt32> (Const32 <typ.UInt32> [32]) s))) (Rsh32Ux32 <typ.UInt32> hi (Sub32 <typ.UInt32> s (Const32 <typ.UInt32> [32]))))) |
| for { |
| s := v.Args[1] |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpRsh32Ux32, typ.UInt32) |
| v0.AddArg(hi) |
| v0.AddArg(s) |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v3 := b.NewValue0(v.Pos, OpRsh32Ux32, typ.UInt32) |
| v3.AddArg(lo) |
| v3.AddArg(s) |
| v2.AddArg(v3) |
| v4 := b.NewValue0(v.Pos, OpLsh32x32, typ.UInt32) |
| v4.AddArg(hi) |
| v5 := b.NewValue0(v.Pos, OpSub32, typ.UInt32) |
| v6 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v6.AuxInt = 32 |
| v5.AddArg(v6) |
| v5.AddArg(s) |
| v4.AddArg(v5) |
| v2.AddArg(v4) |
| v1.AddArg(v2) |
| v7 := b.NewValue0(v.Pos, OpRsh32Ux32, typ.UInt32) |
| v7.AddArg(hi) |
| v8 := b.NewValue0(v.Pos, OpSub32, typ.UInt32) |
| v8.AddArg(s) |
| v9 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v9.AuxInt = 32 |
| v8.AddArg(v9) |
| v7.AddArg(v8) |
| v1.AddArg(v7) |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh64Ux64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh64Ux64 _ (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Const64 [0]) |
| for { |
| _ = v.Args[1] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpConst64) |
| v.AuxInt = 0 |
| return true |
| } |
| // match: (Rsh64Ux64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Rsh64Ux32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpRsh64Ux32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Rsh64Ux64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Rsh64Ux32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpRsh64Ux32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh64Ux8_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh64Ux8 (Int64Make hi lo) s) |
| // cond: |
| // result: (Int64Make (Rsh32Ux8 <typ.UInt32> hi s) (Or32 <typ.UInt32> (Or32 <typ.UInt32> (Rsh32Ux8 <typ.UInt32> lo s) (Lsh32x8 <typ.UInt32> hi (Sub8 <typ.UInt8> (Const8 <typ.UInt8> [32]) s))) (Rsh32Ux8 <typ.UInt32> hi (Sub8 <typ.UInt8> s (Const8 <typ.UInt8> [32]))))) |
| for { |
| s := v.Args[1] |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpRsh32Ux8, typ.UInt32) |
| v0.AddArg(hi) |
| v0.AddArg(s) |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v3 := b.NewValue0(v.Pos, OpRsh32Ux8, typ.UInt32) |
| v3.AddArg(lo) |
| v3.AddArg(s) |
| v2.AddArg(v3) |
| v4 := b.NewValue0(v.Pos, OpLsh32x8, typ.UInt32) |
| v4.AddArg(hi) |
| v5 := b.NewValue0(v.Pos, OpSub8, typ.UInt8) |
| v6 := b.NewValue0(v.Pos, OpConst8, typ.UInt8) |
| v6.AuxInt = 32 |
| v5.AddArg(v6) |
| v5.AddArg(s) |
| v4.AddArg(v5) |
| v2.AddArg(v4) |
| v1.AddArg(v2) |
| v7 := b.NewValue0(v.Pos, OpRsh32Ux8, typ.UInt32) |
| v7.AddArg(hi) |
| v8 := b.NewValue0(v.Pos, OpSub8, typ.UInt8) |
| v8.AddArg(s) |
| v9 := b.NewValue0(v.Pos, OpConst8, typ.UInt8) |
| v9.AuxInt = 32 |
| v8.AddArg(v9) |
| v7.AddArg(v8) |
| v1.AddArg(v7) |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh64x16_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh64x16 (Int64Make hi lo) s) |
| // cond: |
| // result: (Int64Make (Rsh32x16 <typ.UInt32> hi s) (Or32 <typ.UInt32> (Or32 <typ.UInt32> (Rsh32Ux16 <typ.UInt32> lo s) (Lsh32x16 <typ.UInt32> hi (Sub16 <typ.UInt16> (Const16 <typ.UInt16> [32]) s))) (And32 <typ.UInt32> (Rsh32x16 <typ.UInt32> hi (Sub16 <typ.UInt16> s (Const16 <typ.UInt16> [32]))) (Zeromask (ZeroExt16to32 (Rsh16Ux32 <typ.UInt16> s (Const32 <typ.UInt32> [5]))))))) |
| for { |
| s := v.Args[1] |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpRsh32x16, typ.UInt32) |
| v0.AddArg(hi) |
| v0.AddArg(s) |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v3 := b.NewValue0(v.Pos, OpRsh32Ux16, typ.UInt32) |
| v3.AddArg(lo) |
| v3.AddArg(s) |
| v2.AddArg(v3) |
| v4 := b.NewValue0(v.Pos, OpLsh32x16, typ.UInt32) |
| v4.AddArg(hi) |
| v5 := b.NewValue0(v.Pos, OpSub16, typ.UInt16) |
| v6 := b.NewValue0(v.Pos, OpConst16, typ.UInt16) |
| v6.AuxInt = 32 |
| v5.AddArg(v6) |
| v5.AddArg(s) |
| v4.AddArg(v5) |
| v2.AddArg(v4) |
| v1.AddArg(v2) |
| v7 := b.NewValue0(v.Pos, OpAnd32, typ.UInt32) |
| v8 := b.NewValue0(v.Pos, OpRsh32x16, typ.UInt32) |
| v8.AddArg(hi) |
| v9 := b.NewValue0(v.Pos, OpSub16, typ.UInt16) |
| v9.AddArg(s) |
| v10 := b.NewValue0(v.Pos, OpConst16, typ.UInt16) |
| v10.AuxInt = 32 |
| v9.AddArg(v10) |
| v8.AddArg(v9) |
| v7.AddArg(v8) |
| v11 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v12 := b.NewValue0(v.Pos, OpZeroExt16to32, typ.UInt32) |
| v13 := b.NewValue0(v.Pos, OpRsh16Ux32, typ.UInt16) |
| v13.AddArg(s) |
| v14 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v14.AuxInt = 5 |
| v13.AddArg(v14) |
| v12.AddArg(v13) |
| v11.AddArg(v12) |
| v7.AddArg(v11) |
| v1.AddArg(v7) |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh64x32_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh64x32 (Int64Make hi lo) s) |
| // cond: |
| // result: (Int64Make (Rsh32x32 <typ.UInt32> hi s) (Or32 <typ.UInt32> (Or32 <typ.UInt32> (Rsh32Ux32 <typ.UInt32> lo s) (Lsh32x32 <typ.UInt32> hi (Sub32 <typ.UInt32> (Const32 <typ.UInt32> [32]) s))) (And32 <typ.UInt32> (Rsh32x32 <typ.UInt32> hi (Sub32 <typ.UInt32> s (Const32 <typ.UInt32> [32]))) (Zeromask (Rsh32Ux32 <typ.UInt32> s (Const32 <typ.UInt32> [5])))))) |
| for { |
| s := v.Args[1] |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpRsh32x32, typ.UInt32) |
| v0.AddArg(hi) |
| v0.AddArg(s) |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v3 := b.NewValue0(v.Pos, OpRsh32Ux32, typ.UInt32) |
| v3.AddArg(lo) |
| v3.AddArg(s) |
| v2.AddArg(v3) |
| v4 := b.NewValue0(v.Pos, OpLsh32x32, typ.UInt32) |
| v4.AddArg(hi) |
| v5 := b.NewValue0(v.Pos, OpSub32, typ.UInt32) |
| v6 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v6.AuxInt = 32 |
| v5.AddArg(v6) |
| v5.AddArg(s) |
| v4.AddArg(v5) |
| v2.AddArg(v4) |
| v1.AddArg(v2) |
| v7 := b.NewValue0(v.Pos, OpAnd32, typ.UInt32) |
| v8 := b.NewValue0(v.Pos, OpRsh32x32, typ.UInt32) |
| v8.AddArg(hi) |
| v9 := b.NewValue0(v.Pos, OpSub32, typ.UInt32) |
| v9.AddArg(s) |
| v10 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v10.AuxInt = 32 |
| v9.AddArg(v10) |
| v8.AddArg(v9) |
| v7.AddArg(v8) |
| v11 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v12 := b.NewValue0(v.Pos, OpRsh32Ux32, typ.UInt32) |
| v12.AddArg(s) |
| v13 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v13.AuxInt = 5 |
| v12.AddArg(v13) |
| v11.AddArg(v12) |
| v7.AddArg(v11) |
| v1.AddArg(v7) |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh64x64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh64x64 x (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Int64Make (Signmask (Int64Hi x)) (Signmask (Int64Hi x))) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpSignmask, typ.Int32) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v.AddArg(v0) |
| v2 := b.NewValue0(v.Pos, OpSignmask, typ.Int32) |
| v3 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v3.AddArg(x) |
| v2.AddArg(v3) |
| v.AddArg(v2) |
| return true |
| } |
| // match: (Rsh64x64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Rsh64x32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpRsh64x32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Rsh64x64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Rsh64x32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpRsh64x32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh64x8_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh64x8 (Int64Make hi lo) s) |
| // cond: |
| // result: (Int64Make (Rsh32x8 <typ.UInt32> hi s) (Or32 <typ.UInt32> (Or32 <typ.UInt32> (Rsh32Ux8 <typ.UInt32> lo s) (Lsh32x8 <typ.UInt32> hi (Sub8 <typ.UInt8> (Const8 <typ.UInt8> [32]) s))) (And32 <typ.UInt32> (Rsh32x8 <typ.UInt32> hi (Sub8 <typ.UInt8> s (Const8 <typ.UInt8> [32]))) (Zeromask (ZeroExt8to32 (Rsh8Ux32 <typ.UInt8> s (Const32 <typ.UInt32> [5]))))))) |
| for { |
| s := v.Args[1] |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| hi := v_0.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpRsh32x8, typ.UInt32) |
| v0.AddArg(hi) |
| v0.AddArg(s) |
| v.AddArg(v0) |
| v1 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v2 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v3 := b.NewValue0(v.Pos, OpRsh32Ux8, typ.UInt32) |
| v3.AddArg(lo) |
| v3.AddArg(s) |
| v2.AddArg(v3) |
| v4 := b.NewValue0(v.Pos, OpLsh32x8, typ.UInt32) |
| v4.AddArg(hi) |
| v5 := b.NewValue0(v.Pos, OpSub8, typ.UInt8) |
| v6 := b.NewValue0(v.Pos, OpConst8, typ.UInt8) |
| v6.AuxInt = 32 |
| v5.AddArg(v6) |
| v5.AddArg(s) |
| v4.AddArg(v5) |
| v2.AddArg(v4) |
| v1.AddArg(v2) |
| v7 := b.NewValue0(v.Pos, OpAnd32, typ.UInt32) |
| v8 := b.NewValue0(v.Pos, OpRsh32x8, typ.UInt32) |
| v8.AddArg(hi) |
| v9 := b.NewValue0(v.Pos, OpSub8, typ.UInt8) |
| v9.AddArg(s) |
| v10 := b.NewValue0(v.Pos, OpConst8, typ.UInt8) |
| v10.AuxInt = 32 |
| v9.AddArg(v10) |
| v8.AddArg(v9) |
| v7.AddArg(v8) |
| v11 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v12 := b.NewValue0(v.Pos, OpZeroExt8to32, typ.UInt32) |
| v13 := b.NewValue0(v.Pos, OpRsh8Ux32, typ.UInt8) |
| v13.AddArg(s) |
| v14 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v14.AuxInt = 5 |
| v13.AddArg(v14) |
| v12.AddArg(v13) |
| v11.AddArg(v12) |
| v7.AddArg(v11) |
| v1.AddArg(v7) |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh8Ux64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh8Ux64 _ (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Const32 [0]) |
| for { |
| _ = v.Args[1] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpConst32) |
| v.AuxInt = 0 |
| return true |
| } |
| // match: (Rsh8Ux64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Rsh8Ux32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpRsh8Ux32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Rsh8Ux64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Rsh8Ux32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpRsh8Ux32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpRsh8x64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Rsh8x64 x (Int64Make (Const32 [c]) _)) |
| // cond: c != 0 |
| // result: (Signmask (SignExt8to32 x)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| _ = v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| c := v_1_0.AuxInt |
| if !(c != 0) { |
| break |
| } |
| v.reset(OpSignmask) |
| v0 := b.NewValue0(v.Pos, OpSignExt8to32, typ.Int32) |
| v0.AddArg(x) |
| v.AddArg(v0) |
| return true |
| } |
| // match: (Rsh8x64 x (Int64Make (Const32 [0]) lo)) |
| // cond: |
| // result: (Rsh8x32 x lo) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| v_1_0 := v_1.Args[0] |
| if v_1_0.Op != OpConst32 { |
| break |
| } |
| if v_1_0.AuxInt != 0 { |
| break |
| } |
| v.reset(OpRsh8x32) |
| v.AddArg(x) |
| v.AddArg(lo) |
| return true |
| } |
| // match: (Rsh8x64 x (Int64Make hi lo)) |
| // cond: hi.Op != OpConst32 |
| // result: (Rsh8x32 x (Or32 <typ.UInt32> (Zeromask hi) lo)) |
| for { |
| _ = v.Args[1] |
| x := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(hi.Op != OpConst32) { |
| break |
| } |
| v.reset(OpRsh8x32) |
| v.AddArg(x) |
| v0 := b.NewValue0(v.Pos, OpOr32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpZeromask, typ.UInt32) |
| v1.AddArg(hi) |
| v0.AddArg(v1) |
| v0.AddArg(lo) |
| v.AddArg(v0) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpSignExt16to64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (SignExt16to64 x) |
| // cond: |
| // result: (SignExt32to64 (SignExt16to32 x)) |
| for { |
| x := v.Args[0] |
| v.reset(OpSignExt32to64) |
| v0 := b.NewValue0(v.Pos, OpSignExt16to32, typ.Int32) |
| v0.AddArg(x) |
| v.AddArg(v0) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpSignExt32to64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (SignExt32to64 x) |
| // cond: |
| // result: (Int64Make (Signmask x) x) |
| for { |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpSignmask, typ.Int32) |
| v0.AddArg(x) |
| v.AddArg(v0) |
| v.AddArg(x) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpSignExt8to64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (SignExt8to64 x) |
| // cond: |
| // result: (SignExt32to64 (SignExt8to32 x)) |
| for { |
| x := v.Args[0] |
| v.reset(OpSignExt32to64) |
| v0 := b.NewValue0(v.Pos, OpSignExt8to32, typ.Int32) |
| v0.AddArg(x) |
| v.AddArg(v0) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpStore_0(v *Value) bool { |
| b := v.Block |
| config := b.Func.Config |
| // match: (Store {t} dst (Int64Make hi lo) mem) |
| // cond: t.(*types.Type).Size() == 8 && !config.BigEndian |
| // result: (Store {hi.Type} (OffPtr <hi.Type.PtrTo()> [4] dst) hi (Store {lo.Type} dst lo mem)) |
| for { |
| t := v.Aux |
| mem := v.Args[2] |
| dst := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(t.(*types.Type).Size() == 8 && !config.BigEndian) { |
| break |
| } |
| v.reset(OpStore) |
| v.Aux = hi.Type |
| v0 := b.NewValue0(v.Pos, OpOffPtr, hi.Type.PtrTo()) |
| v0.AuxInt = 4 |
| v0.AddArg(dst) |
| v.AddArg(v0) |
| v.AddArg(hi) |
| v1 := b.NewValue0(v.Pos, OpStore, types.TypeMem) |
| v1.Aux = lo.Type |
| v1.AddArg(dst) |
| v1.AddArg(lo) |
| v1.AddArg(mem) |
| v.AddArg(v1) |
| return true |
| } |
| // match: (Store {t} dst (Int64Make hi lo) mem) |
| // cond: t.(*types.Type).Size() == 8 && config.BigEndian |
| // result: (Store {lo.Type} (OffPtr <lo.Type.PtrTo()> [4] dst) lo (Store {hi.Type} dst hi mem)) |
| for { |
| t := v.Aux |
| mem := v.Args[2] |
| dst := v.Args[0] |
| v_1 := v.Args[1] |
| if v_1.Op != OpInt64Make { |
| break |
| } |
| lo := v_1.Args[1] |
| hi := v_1.Args[0] |
| if !(t.(*types.Type).Size() == 8 && config.BigEndian) { |
| break |
| } |
| v.reset(OpStore) |
| v.Aux = lo.Type |
| v0 := b.NewValue0(v.Pos, OpOffPtr, lo.Type.PtrTo()) |
| v0.AuxInt = 4 |
| v0.AddArg(dst) |
| v.AddArg(v0) |
| v.AddArg(lo) |
| v1 := b.NewValue0(v.Pos, OpStore, types.TypeMem) |
| v1.Aux = hi.Type |
| v1.AddArg(dst) |
| v1.AddArg(hi) |
| v1.AddArg(mem) |
| v.AddArg(v1) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpSub64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Sub64 x y) |
| // cond: |
| // result: (Int64Make (Sub32withcarry <typ.Int32> (Int64Hi x) (Int64Hi y) (Select1 <types.TypeFlags> (Sub32carry (Int64Lo x) (Int64Lo y)))) (Select0 <typ.UInt32> (Sub32carry (Int64Lo x) (Int64Lo y)))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpSub32withcarry, typ.Int32) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v3 := b.NewValue0(v.Pos, OpSelect1, types.TypeFlags) |
| v4 := b.NewValue0(v.Pos, OpSub32carry, types.NewTuple(typ.UInt32, types.TypeFlags)) |
| v5 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v5.AddArg(x) |
| v4.AddArg(v5) |
| v6 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v6.AddArg(y) |
| v4.AddArg(v6) |
| v3.AddArg(v4) |
| v0.AddArg(v3) |
| v.AddArg(v0) |
| v7 := b.NewValue0(v.Pos, OpSelect0, typ.UInt32) |
| v8 := b.NewValue0(v.Pos, OpSub32carry, types.NewTuple(typ.UInt32, types.TypeFlags)) |
| v9 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v9.AddArg(x) |
| v8.AddArg(v9) |
| v10 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v10.AddArg(y) |
| v8.AddArg(v10) |
| v7.AddArg(v8) |
| v.AddArg(v7) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpTrunc64to16_0(v *Value) bool { |
| // match: (Trunc64to16 (Int64Make _ lo)) |
| // cond: |
| // result: (Trunc32to16 lo) |
| for { |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| v.reset(OpTrunc32to16) |
| v.AddArg(lo) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpTrunc64to32_0(v *Value) bool { |
| // match: (Trunc64to32 (Int64Make _ lo)) |
| // cond: |
| // result: lo |
| for { |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| v.reset(OpCopy) |
| v.Type = lo.Type |
| v.AddArg(lo) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpTrunc64to8_0(v *Value) bool { |
| // match: (Trunc64to8 (Int64Make _ lo)) |
| // cond: |
| // result: (Trunc32to8 lo) |
| for { |
| v_0 := v.Args[0] |
| if v_0.Op != OpInt64Make { |
| break |
| } |
| lo := v_0.Args[1] |
| v.reset(OpTrunc32to8) |
| v.AddArg(lo) |
| return true |
| } |
| return false |
| } |
| func rewriteValuedec64_OpXor64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (Xor64 x y) |
| // cond: |
| // result: (Int64Make (Xor32 <typ.UInt32> (Int64Hi x) (Int64Hi y)) (Xor32 <typ.UInt32> (Int64Lo x) (Int64Lo y))) |
| for { |
| y := v.Args[1] |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpXor32, typ.UInt32) |
| v1 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v1.AddArg(x) |
| v0.AddArg(v1) |
| v2 := b.NewValue0(v.Pos, OpInt64Hi, typ.UInt32) |
| v2.AddArg(y) |
| v0.AddArg(v2) |
| v.AddArg(v0) |
| v3 := b.NewValue0(v.Pos, OpXor32, typ.UInt32) |
| v4 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v4.AddArg(x) |
| v3.AddArg(v4) |
| v5 := b.NewValue0(v.Pos, OpInt64Lo, typ.UInt32) |
| v5.AddArg(y) |
| v3.AddArg(v5) |
| v.AddArg(v3) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpZeroExt16to64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (ZeroExt16to64 x) |
| // cond: |
| // result: (ZeroExt32to64 (ZeroExt16to32 x)) |
| for { |
| x := v.Args[0] |
| v.reset(OpZeroExt32to64) |
| v0 := b.NewValue0(v.Pos, OpZeroExt16to32, typ.UInt32) |
| v0.AddArg(x) |
| v.AddArg(v0) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpZeroExt32to64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (ZeroExt32to64 x) |
| // cond: |
| // result: (Int64Make (Const32 <typ.UInt32> [0]) x) |
| for { |
| x := v.Args[0] |
| v.reset(OpInt64Make) |
| v0 := b.NewValue0(v.Pos, OpConst32, typ.UInt32) |
| v0.AuxInt = 0 |
| v.AddArg(v0) |
| v.AddArg(x) |
| return true |
| } |
| } |
| func rewriteValuedec64_OpZeroExt8to64_0(v *Value) bool { |
| b := v.Block |
| typ := &b.Func.Config.Types |
| // match: (ZeroExt8to64 x) |
| // cond: |
| // result: (ZeroExt32to64 (ZeroExt8to32 x)) |
| for { |
| x := v.Args[0] |
| v.reset(OpZeroExt32to64) |
| v0 := b.NewValue0(v.Pos, OpZeroExt8to32, typ.UInt32) |
| v0.AddArg(x) |
| v.AddArg(v0) |
| return true |
| } |
| } |
| func rewriteBlockdec64(b *Block) bool { |
| config := b.Func.Config |
| typ := &config.Types |
| _ = typ |
| v := b.Control |
| _ = v |
| switch b.Kind { |
| } |
| return false |
| } |