src/cmd/compile/internal/ssa/gen/AMD64.rules | 8 ++++---- src/cmd/compile/internal/ssa/gen/AMD64Ops.go | 2 +- src/cmd/compile/internal/ssa/rewriteAMD64.go | 32 ++++++++++++++++++++++++-------- test/fixedbugs/issue25322.go | 23 +++++++++++++++++++++++ test/fixedbugs/issue25322.out | 1 + diff --git a/src/cmd/compile/internal/ssa/gen/AMD64.rules b/src/cmd/compile/internal/ssa/gen/AMD64.rules index db7c1a447b247e05b6498adca71a5c060f0c9ba8..a90722df6e240d966d8678defa747afa22e1dc24 100644 --- a/src/cmd/compile/internal/ssa/gen/AMD64.rules +++ b/src/cmd/compile/internal/ssa/gen/AMD64.rules @@ -2658,7 +2658,7 @@ (MOVSSstore [off] {sym} ptr (MOVLi2f val) mem) -> (MOVLstore [off] {sym} ptr val mem) // Load args directly into the register class where it will be used. // We do this by just modifying the type of the Arg. -(MOVQf2i (Arg [off] {sym})) -> @b.Func.Entry (Arg [off] {sym}) -(MOVLf2i (Arg [off] {sym})) -> @b.Func.Entry (Arg [off] {sym}) -(MOVQi2f (Arg [off] {sym})) -> @b.Func.Entry (Arg [off] {sym}) -(MOVLi2f (Arg [off] {sym})) -> @b.Func.Entry (Arg [off] {sym}) +(MOVQf2i (Arg [off] {sym})) && t.Size() == u.Size() -> @b.Func.Entry (Arg [off] {sym}) +(MOVLf2i (Arg [off] {sym})) && t.Size() == u.Size() -> @b.Func.Entry (Arg [off] {sym}) +(MOVQi2f (Arg [off] {sym})) && t.Size() == u.Size() -> @b.Func.Entry (Arg [off] {sym}) +(MOVLi2f (Arg [off] {sym})) && t.Size() == u.Size() -> @b.Func.Entry (Arg [off] {sym}) diff --git a/src/cmd/compile/internal/ssa/gen/AMD64Ops.go b/src/cmd/compile/internal/ssa/gen/AMD64Ops.go index b7c0ce9f086fe6d805d88bfcb444f93108e6cab9..c3075e1020a2bf17ae702fe72f31b0622ebb998f 100644 --- a/src/cmd/compile/internal/ssa/gen/AMD64Ops.go +++ b/src/cmd/compile/internal/ssa/gen/AMD64Ops.go @@ -407,7 +407,7 @@ // TODO: should we have generic versions of these? {name: "MOVQi2f", argLength: 1, reg: gpfp, typ: "Float64"}, // move 64 bits from int to float reg {name: "MOVQf2i", argLength: 1, reg: fpgp, typ: "UInt64"}, // move 64 bits from float to int reg {name: "MOVLi2f", argLength: 1, reg: gpfp, typ: "Float32"}, // move 32 bits from int to float reg - {name: "MOVLf2i", argLength: 1, reg: fpgp, typ: "UInt32"}, // move 32 bits from float to int reg + {name: "MOVLf2i", argLength: 1, reg: fpgp, typ: "UInt32"}, // move 32 bits from float to int reg, zero extend {name: "PXOR", argLength: 2, reg: fp21, asm: "PXOR", commutative: true, resultInArg0: true}, // exclusive or, applied to X regs for float negation. diff --git a/src/cmd/compile/internal/ssa/rewriteAMD64.go b/src/cmd/compile/internal/ssa/rewriteAMD64.go index 0b2b321d2566603215b22e78ca676f0820844737..e309ce6bf629f4032fcc2fe7eebb7d50315d707c 100644 --- a/src/cmd/compile/internal/ssa/rewriteAMD64.go +++ b/src/cmd/compile/internal/ssa/rewriteAMD64.go @@ -7428,8 +7428,8 @@ } func rewriteValueAMD64_OpAMD64MOVLf2i_0(v *Value) bool { b := v.Block _ = b - // match: (MOVLf2i (Arg [off] {sym})) - // cond: + // match: (MOVLf2i (Arg [off] {sym})) + // cond: t.Size() == u.Size() // result: @b.Func.Entry (Arg [off] {sym}) for { t := v.Type @@ -7437,8 +7437,12 @@ v_0 := v.Args[0] if v_0.Op != OpArg { break } + u := v_0.Type off := v_0.AuxInt sym := v_0.Aux + if !(t.Size() == u.Size()) { + break + } b = b.Func.Entry v0 := b.NewValue0(v.Pos, OpArg, t) v.reset(OpCopy) @@ -7452,8 +7456,8 @@ } func rewriteValueAMD64_OpAMD64MOVLi2f_0(v *Value) bool { b := v.Block _ = b - // match: (MOVLi2f (Arg [off] {sym})) - // cond: + // match: (MOVLi2f (Arg [off] {sym})) + // cond: t.Size() == u.Size() // result: @b.Func.Entry (Arg [off] {sym}) for { t := v.Type @@ -7461,8 +7465,12 @@ v_0 := v.Args[0] if v_0.Op != OpArg { break } + u := v_0.Type off := v_0.AuxInt sym := v_0.Aux + if !(t.Size() == u.Size()) { + break + } b = b.Func.Entry v0 := b.NewValue0(v.Pos, OpArg, t) v.reset(OpCopy) @@ -9631,8 +9639,8 @@ } func rewriteValueAMD64_OpAMD64MOVQf2i_0(v *Value) bool { b := v.Block _ = b - // match: (MOVQf2i (Arg [off] {sym})) - // cond: + // match: (MOVQf2i (Arg [off] {sym})) + // cond: t.Size() == u.Size() // result: @b.Func.Entry (Arg [off] {sym}) for { t := v.Type @@ -9640,8 +9648,12 @@ v_0 := v.Args[0] if v_0.Op != OpArg { break } + u := v_0.Type off := v_0.AuxInt sym := v_0.Aux + if !(t.Size() == u.Size()) { + break + } b = b.Func.Entry v0 := b.NewValue0(v.Pos, OpArg, t) v.reset(OpCopy) @@ -9655,8 +9667,8 @@ } func rewriteValueAMD64_OpAMD64MOVQi2f_0(v *Value) bool { b := v.Block _ = b - // match: (MOVQi2f (Arg [off] {sym})) - // cond: + // match: (MOVQi2f (Arg [off] {sym})) + // cond: t.Size() == u.Size() // result: @b.Func.Entry (Arg [off] {sym}) for { t := v.Type @@ -9664,8 +9676,12 @@ v_0 := v.Args[0] if v_0.Op != OpArg { break } + u := v_0.Type off := v_0.AuxInt sym := v_0.Aux + if !(t.Size() == u.Size()) { + break + } b = b.Func.Entry v0 := b.NewValue0(v.Pos, OpArg, t) v.reset(OpCopy) diff --git a/test/fixedbugs/issue25322.go b/test/fixedbugs/issue25322.go new file mode 100644 index 0000000000000000000000000000000000000000..7489bbdfc2219307d331d8737463fbfbc2ae3851 --- /dev/null +++ b/test/fixedbugs/issue25322.go @@ -0,0 +1,23 @@ +// cmpout + +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Missing zero extension when converting a float32 +// to a uint64. + +package main + +import ( + "fmt" + "math" +) + +func Foo(v float32) { + fmt.Printf("%x\n", uint64(math.Float32bits(v))) +} + +func main() { + Foo(2.0) +} diff --git a/test/fixedbugs/issue25322.out b/test/fixedbugs/issue25322.out new file mode 100644 index 0000000000000000000000000000000000000000..52f3f6a74518f6f135a1ee2207cbea4e03ff6e89 --- /dev/null +++ b/test/fixedbugs/issue25322.out @@ -0,0 +1 @@ +40000000