Extend JSBuiltinReducer to cover Math.sqrt as well.

R=bmeurer@chromium.org
TEST=compiler-unittests/JSBuiltinReducerTest.MathSqrt

Review URL: https://codereview.chromium.org/595963002

git-svn-id: https://v8.googlecode.com/svn/branches/bleeding_edge@24177 ce2b1a6d-e550-0410-aec6-3dcde31c8c00
parent 98277cfa
...@@ -323,6 +323,9 @@ void CodeGenerator::AssembleArchInstruction(Instruction* instr) { ...@@ -323,6 +323,9 @@ void CodeGenerator::AssembleArchInstruction(Instruction* instr) {
case kArmVnegF64: case kArmVnegF64:
__ vneg(i.OutputDoubleRegister(), i.InputDoubleRegister(0)); __ vneg(i.OutputDoubleRegister(), i.InputDoubleRegister(0));
break; break;
case kArmVsqrtF64:
__ vsqrt(i.OutputDoubleRegister(), i.InputDoubleRegister(0));
break;
case kArmVcvtF64S32: { case kArmVcvtF64S32: {
SwVfpRegister scratch = kScratchDoubleReg.low(); SwVfpRegister scratch = kScratchDoubleReg.low();
__ vmov(scratch, i.InputRegister(0)); __ vmov(scratch, i.InputRegister(0));
......
...@@ -41,6 +41,7 @@ namespace compiler { ...@@ -41,6 +41,7 @@ namespace compiler {
V(ArmVdivF64) \ V(ArmVdivF64) \
V(ArmVmodF64) \ V(ArmVmodF64) \
V(ArmVnegF64) \ V(ArmVnegF64) \
V(ArmVsqrtF64) \
V(ArmVcvtF64S32) \ V(ArmVcvtF64S32) \
V(ArmVcvtF64U32) \ V(ArmVcvtF64U32) \
V(ArmVcvtS32F64) \ V(ArmVcvtS32F64) \
......
...@@ -90,6 +90,7 @@ class ArmOperandGenerator FINAL : public OperandGenerator { ...@@ -90,6 +90,7 @@ class ArmOperandGenerator FINAL : public OperandGenerator {
case kArmVdivF64: case kArmVdivF64:
case kArmVmodF64: case kArmVmodF64:
case kArmVnegF64: case kArmVnegF64:
case kArmVsqrtF64:
case kArmVcvtF64S32: case kArmVcvtF64S32:
case kArmVcvtF64U32: case kArmVcvtF64U32:
case kArmVcvtS32F64: case kArmVcvtS32F64:
...@@ -768,6 +769,12 @@ void InstructionSelector::VisitFloat64Mod(Node* node) { ...@@ -768,6 +769,12 @@ void InstructionSelector::VisitFloat64Mod(Node* node) {
} }
void InstructionSelector::VisitFloat64Sqrt(Node* node) {
ArmOperandGenerator g(this);
Emit(kArmVsqrtF64, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)));
}
void InstructionSelector::VisitCall(Node* call, BasicBlock* continuation, void InstructionSelector::VisitCall(Node* call, BasicBlock* continuation,
BasicBlock* deoptimization) { BasicBlock* deoptimization) {
ArmOperandGenerator g(this); ArmOperandGenerator g(this);
......
...@@ -373,6 +373,9 @@ void CodeGenerator::AssembleArchInstruction(Instruction* instr) { ...@@ -373,6 +373,9 @@ void CodeGenerator::AssembleArchInstruction(Instruction* instr) {
0, 2); 0, 2);
break; break;
} }
case kArm64Float64Sqrt:
__ Fsqrt(i.OutputDoubleRegister(), i.InputDoubleRegister(0));
break;
case kArm64Float64ToInt32: case kArm64Float64ToInt32:
__ Fcvtzs(i.OutputRegister32(), i.InputDoubleRegister(0)); __ Fcvtzs(i.OutputRegister32(), i.InputDoubleRegister(0));
break; break;
......
...@@ -62,6 +62,7 @@ namespace compiler { ...@@ -62,6 +62,7 @@ namespace compiler {
V(Arm64Float64Mul) \ V(Arm64Float64Mul) \
V(Arm64Float64Div) \ V(Arm64Float64Div) \
V(Arm64Float64Mod) \ V(Arm64Float64Mod) \
V(Arm64Float64Sqrt) \
V(Arm64Float64ToInt32) \ V(Arm64Float64ToInt32) \
V(Arm64Float64ToUint32) \ V(Arm64Float64ToUint32) \
V(Arm64Int32ToFloat64) \ V(Arm64Int32ToFloat64) \
......
...@@ -506,6 +506,13 @@ void InstructionSelector::VisitFloat64Mod(Node* node) { ...@@ -506,6 +506,13 @@ void InstructionSelector::VisitFloat64Mod(Node* node) {
} }
void InstructionSelector::VisitFloat64Sqrt(Node* node) {
Arm64OperandGenerator g(this);
Emit(kArm64Float64Sqrt, g.DefineAsRegister(node),
g.UseRegister(node->InputAt(0)));
}
void InstructionSelector::VisitInt32AddWithOverflow(Node* node, void InstructionSelector::VisitInt32AddWithOverflow(Node* node,
FlagsContinuation* cont) { FlagsContinuation* cont) {
VisitBinop<Int32BinopMatcher>(this, node, kArm64Add32, kArithmeticImm, cont); VisitBinop<Int32BinopMatcher>(this, node, kArm64Add32, kArithmeticImm, cont);
......
...@@ -771,6 +771,7 @@ IS_UNOP_MATCHER(ChangeUint32ToFloat64) ...@@ -771,6 +771,7 @@ IS_UNOP_MATCHER(ChangeUint32ToFloat64)
IS_UNOP_MATCHER(ChangeUint32ToUint64) IS_UNOP_MATCHER(ChangeUint32ToUint64)
IS_UNOP_MATCHER(TruncateFloat64ToInt32) IS_UNOP_MATCHER(TruncateFloat64ToInt32)
IS_UNOP_MATCHER(TruncateInt64ToInt32) IS_UNOP_MATCHER(TruncateInt64ToInt32)
IS_UNOP_MATCHER(Float64Sqrt)
#undef IS_UNOP_MATCHER #undef IS_UNOP_MATCHER
} // namespace compiler } // namespace compiler
......
...@@ -131,6 +131,7 @@ Matcher<Node*> IsChangeUint32ToFloat64(const Matcher<Node*>& input_matcher); ...@@ -131,6 +131,7 @@ Matcher<Node*> IsChangeUint32ToFloat64(const Matcher<Node*>& input_matcher);
Matcher<Node*> IsChangeUint32ToUint64(const Matcher<Node*>& input_matcher); Matcher<Node*> IsChangeUint32ToUint64(const Matcher<Node*>& input_matcher);
Matcher<Node*> IsTruncateFloat64ToInt32(const Matcher<Node*>& input_matcher); Matcher<Node*> IsTruncateFloat64ToInt32(const Matcher<Node*>& input_matcher);
Matcher<Node*> IsTruncateInt64ToInt32(const Matcher<Node*>& input_matcher); Matcher<Node*> IsTruncateInt64ToInt32(const Matcher<Node*>& input_matcher);
Matcher<Node*> IsFloat64Sqrt(const Matcher<Node*>& input_matcher);
} // namespace compiler } // namespace compiler
} // namespace internal } // namespace internal
......
...@@ -285,6 +285,9 @@ void CodeGenerator::AssembleArchInstruction(Instruction* instr) { ...@@ -285,6 +285,9 @@ void CodeGenerator::AssembleArchInstruction(Instruction* instr) {
__ add(esp, Immediate(kDoubleSize)); __ add(esp, Immediate(kDoubleSize));
break; break;
} }
case kSSEFloat64Sqrt:
__ sqrtsd(i.OutputDoubleRegister(), i.InputOperand(0));
break;
case kSSEFloat64ToInt32: case kSSEFloat64ToInt32:
__ cvttsd2si(i.OutputRegister(), i.InputOperand(0)); __ cvttsd2si(i.OutputRegister(), i.InputOperand(0));
break; break;
......
...@@ -34,6 +34,7 @@ namespace compiler { ...@@ -34,6 +34,7 @@ namespace compiler {
V(SSEFloat64Mul) \ V(SSEFloat64Mul) \
V(SSEFloat64Div) \ V(SSEFloat64Div) \
V(SSEFloat64Mod) \ V(SSEFloat64Mod) \
V(SSEFloat64Sqrt) \
V(SSEFloat64ToInt32) \ V(SSEFloat64ToInt32) \
V(SSEFloat64ToUint32) \ V(SSEFloat64ToUint32) \
V(SSEInt32ToFloat64) \ V(SSEInt32ToFloat64) \
......
...@@ -417,6 +417,12 @@ void InstructionSelector::VisitFloat64Mod(Node* node) { ...@@ -417,6 +417,12 @@ void InstructionSelector::VisitFloat64Mod(Node* node) {
} }
void InstructionSelector::VisitFloat64Sqrt(Node* node) {
IA32OperandGenerator g(this);
Emit(kSSEFloat64Sqrt, g.DefineAsRegister(node), g.Use(node->InputAt(0)));
}
void InstructionSelector::VisitInt32AddWithOverflow(Node* node, void InstructionSelector::VisitInt32AddWithOverflow(Node* node,
FlagsContinuation* cont) { FlagsContinuation* cont) {
VisitBinop(this, node, kIA32Add, cont); VisitBinop(this, node, kIA32Add, cont);
......
...@@ -596,6 +596,8 @@ void InstructionSelector::VisitNode(Node* node) { ...@@ -596,6 +596,8 @@ void InstructionSelector::VisitNode(Node* node) {
return MarkAsDouble(node), VisitFloat64Div(node); return MarkAsDouble(node), VisitFloat64Div(node);
case IrOpcode::kFloat64Mod: case IrOpcode::kFloat64Mod:
return MarkAsDouble(node), VisitFloat64Mod(node); return MarkAsDouble(node), VisitFloat64Mod(node);
case IrOpcode::kFloat64Sqrt:
return MarkAsDouble(node), VisitFloat64Sqrt(node);
case IrOpcode::kFloat64Equal: case IrOpcode::kFloat64Equal:
return VisitFloat64Equal(node); return VisitFloat64Equal(node);
case IrOpcode::kFloat64LessThan: case IrOpcode::kFloat64LessThan:
......
...@@ -59,6 +59,26 @@ Type* const kNumberTypes[] = { ...@@ -59,6 +59,26 @@ Type* const kNumberTypes[] = {
} // namespace } // namespace
// -----------------------------------------------------------------------------
// Math.sqrt
TEST_F(JSBuiltinReducerTest, MathSqrt) {
Handle<JSFunction> f(isolate()->context()->math_sqrt_fun());
TRACED_FOREACH(Type*, t0, kNumberTypes) {
Node* p0 = Parameter(t0, 0);
Node* fun = HeapConstant(Unique<HeapObject>::CreateUninitialized(f));
Node* call = graph()->NewNode(javascript()->Call(3, NO_CALL_FUNCTION_FLAGS),
fun, UndefinedConstant(), p0);
Reduction r = Reduce(call);
ASSERT_TRUE(r.Changed());
EXPECT_THAT(r.replacement(), IsFloat64Sqrt(p0));
}
}
// ----------------------------------------------------------------------------- // -----------------------------------------------------------------------------
// Math.max // Math.max
...@@ -71,7 +91,7 @@ TEST_F(JSBuiltinReducerTest, MathMax0) { ...@@ -71,7 +91,7 @@ TEST_F(JSBuiltinReducerTest, MathMax0) {
fun, UndefinedConstant()); fun, UndefinedConstant());
Reduction r = Reduce(call); Reduction r = Reduce(call);
EXPECT_TRUE(r.Changed()); ASSERT_TRUE(r.Changed());
EXPECT_THAT(r.replacement(), IsNumberConstant(-V8_INFINITY)); EXPECT_THAT(r.replacement(), IsNumberConstant(-V8_INFINITY));
} }
...@@ -86,7 +106,7 @@ TEST_F(JSBuiltinReducerTest, MathMax1) { ...@@ -86,7 +106,7 @@ TEST_F(JSBuiltinReducerTest, MathMax1) {
fun, UndefinedConstant(), p0); fun, UndefinedConstant(), p0);
Reduction r = Reduce(call); Reduction r = Reduce(call);
EXPECT_TRUE(r.Changed()); ASSERT_TRUE(r.Changed());
EXPECT_THAT(r.replacement(), p0); EXPECT_THAT(r.replacement(), p0);
} }
} }
...@@ -107,7 +127,7 @@ TEST_F(JSBuiltinReducerTest, MathMax2) { ...@@ -107,7 +127,7 @@ TEST_F(JSBuiltinReducerTest, MathMax2) {
if (t0->Is(Type::Integral32()) && t1->Is(Type::Integral32())) { if (t0->Is(Type::Integral32()) && t1->Is(Type::Integral32())) {
Capture<Node*> branch; Capture<Node*> branch;
EXPECT_TRUE(r.Changed()); ASSERT_TRUE(r.Changed());
EXPECT_THAT( EXPECT_THAT(
r.replacement(), r.replacement(),
IsPhi(kMachNone, p1, p0, IsPhi(kMachNone, p1, p0,
...@@ -116,7 +136,7 @@ TEST_F(JSBuiltinReducerTest, MathMax2) { ...@@ -116,7 +136,7 @@ TEST_F(JSBuiltinReducerTest, MathMax2) {
IsBranch(IsNumberLessThan(p0, p1), IsBranch(IsNumberLessThan(p0, p1),
graph()->start())))))); graph()->start()))))));
} else { } else {
EXPECT_FALSE(r.Changed()); ASSERT_FALSE(r.Changed());
EXPECT_EQ(IrOpcode::kJSCallFunction, call->opcode()); EXPECT_EQ(IrOpcode::kJSCallFunction, call->opcode());
} }
} }
...@@ -142,10 +162,10 @@ TEST_F(JSBuiltinReducerTest, MathImul) { ...@@ -142,10 +162,10 @@ TEST_F(JSBuiltinReducerTest, MathImul) {
Reduction r = Reduce(call); Reduction r = Reduce(call);
if (t0->Is(Type::Integral32()) && t1->Is(Type::Integral32())) { if (t0->Is(Type::Integral32()) && t1->Is(Type::Integral32())) {
EXPECT_TRUE(r.Changed()); ASSERT_TRUE(r.Changed());
EXPECT_THAT(r.replacement(), IsInt32Mul(p0, p1)); EXPECT_THAT(r.replacement(), IsInt32Mul(p0, p1));
} else { } else {
EXPECT_FALSE(r.Changed()); ASSERT_FALSE(r.Changed());
EXPECT_EQ(IrOpcode::kJSCallFunction, call->opcode()); EXPECT_EQ(IrOpcode::kJSCallFunction, call->opcode());
} }
} }
......
...@@ -95,6 +95,18 @@ class JSCallReduction { ...@@ -95,6 +95,18 @@ class JSCallReduction {
}; };
// ECMA-262, section 15.8.2.17.
Reduction JSBuiltinReducer::ReduceMathSqrt(Node* node) {
JSCallReduction r(node);
if (r.InputsMatchOne(Type::Number())) {
// Math.sqrt(a:number) -> Float64Sqrt(a)
Node* value = graph()->NewNode(machine()->Float64Sqrt(), r.left());
return Replace(value);
}
return NoChange();
}
// ECMA-262, section 15.8.2.11. // ECMA-262, section 15.8.2.11.
Reduction JSBuiltinReducer::ReduceMathMax(Node* node) { Reduction JSBuiltinReducer::ReduceMathMax(Node* node) {
JSCallReduction r(node); JSCallReduction r(node);
...@@ -145,6 +157,8 @@ Reduction JSBuiltinReducer::Reduce(Node* node) { ...@@ -145,6 +157,8 @@ Reduction JSBuiltinReducer::Reduce(Node* node) {
// Dispatch according to the BuiltinFunctionId if present. // Dispatch according to the BuiltinFunctionId if present.
if (!r.HasBuiltinFunctionId()) return NoChange(); if (!r.HasBuiltinFunctionId()) return NoChange();
switch (r.GetBuiltinFunctionId()) { switch (r.GetBuiltinFunctionId()) {
case kMathSqrt:
return ReplaceWithPureReduction(node, ReduceMathSqrt(node));
case kMathMax: case kMathMax:
return ReplaceWithPureReduction(node, ReduceMathMax(node)); return ReplaceWithPureReduction(node, ReduceMathMax(node));
case kMathImul: case kMathImul:
......
...@@ -30,6 +30,7 @@ class JSBuiltinReducer FINAL : public Reducer { ...@@ -30,6 +30,7 @@ class JSBuiltinReducer FINAL : public Reducer {
MachineOperatorBuilder* machine() const { return jsgraph_->machine(); } MachineOperatorBuilder* machine() const { return jsgraph_->machine(); }
SimplifiedOperatorBuilder* simplified() { return &simplified_; } SimplifiedOperatorBuilder* simplified() { return &simplified_; }
Reduction ReduceMathSqrt(Node* node);
Reduction ReduceMathMax(Node* node); Reduction ReduceMathMax(Node* node);
Reduction ReduceMathImul(Node* node); Reduction ReduceMathImul(Node* node);
......
...@@ -213,8 +213,8 @@ const PureOperator kPureOperators[] = { ...@@ -213,8 +213,8 @@ const PureOperator kPureOperators[] = {
PURE(TruncateInt64ToInt32, 1, 1), PURE(Float64Add, 2, 1), PURE(TruncateInt64ToInt32, 1, 1), PURE(Float64Add, 2, 1),
PURE(Float64Sub, 2, 1), PURE(Float64Mul, 2, 1), PURE(Float64Sub, 2, 1), PURE(Float64Mul, 2, 1),
PURE(Float64Div, 2, 1), PURE(Float64Mod, 2, 1), PURE(Float64Div, 2, 1), PURE(Float64Mod, 2, 1),
PURE(Float64Equal, 2, 1), PURE(Float64LessThan, 2, 1), PURE(Float64Sqrt, 1, 1), PURE(Float64Equal, 2, 1),
PURE(Float64LessThanOrEqual, 2, 1) PURE(Float64LessThan, 2, 1), PURE(Float64LessThanOrEqual, 2, 1)
#undef PURE #undef PURE
}; };
......
...@@ -112,6 +112,7 @@ struct StaticParameterTraits<LoadRepresentation> { ...@@ -112,6 +112,7 @@ struct StaticParameterTraits<LoadRepresentation> {
V(Float64Mul, Operator::kCommutative, 2, 1) \ V(Float64Mul, Operator::kCommutative, 2, 1) \
V(Float64Div, Operator::kNoProperties, 2, 1) \ V(Float64Div, Operator::kNoProperties, 2, 1) \
V(Float64Mod, Operator::kNoProperties, 2, 1) \ V(Float64Mod, Operator::kNoProperties, 2, 1) \
V(Float64Sqrt, Operator::kNoProperties, 1, 1) \
V(Float64Equal, Operator::kCommutative, 2, 1) \ V(Float64Equal, Operator::kCommutative, 2, 1) \
V(Float64LessThan, Operator::kNoProperties, 2, 1) \ V(Float64LessThan, Operator::kNoProperties, 2, 1) \
V(Float64LessThanOrEqual, Operator::kNoProperties, 2, 1) V(Float64LessThanOrEqual, Operator::kNoProperties, 2, 1)
......
...@@ -129,6 +129,7 @@ class MachineOperatorBuilder FINAL { ...@@ -129,6 +129,7 @@ class MachineOperatorBuilder FINAL {
const Operator* Float64Mul(); const Operator* Float64Mul();
const Operator* Float64Div(); const Operator* Float64Div();
const Operator* Float64Mod(); const Operator* Float64Mod();
const Operator* Float64Sqrt();
// Floating point comparisons complying to IEEE 754. // Floating point comparisons complying to IEEE 754.
const Operator* Float64Equal(); const Operator* Float64Equal();
......
...@@ -218,6 +218,7 @@ ...@@ -218,6 +218,7 @@
V(Float64Mul) \ V(Float64Mul) \
V(Float64Div) \ V(Float64Div) \
V(Float64Mod) \ V(Float64Mod) \
V(Float64Sqrt) \
V(Float64Equal) \ V(Float64Equal) \
V(Float64LessThan) \ V(Float64LessThan) \
V(Float64LessThanOrEqual) V(Float64LessThanOrEqual)
......
...@@ -724,6 +724,8 @@ class RepresentationSelector { ...@@ -724,6 +724,8 @@ class RepresentationSelector {
case IrOpcode::kFloat64Div: case IrOpcode::kFloat64Div:
case IrOpcode::kFloat64Mod: case IrOpcode::kFloat64Mod:
return VisitFloat64Binop(node); return VisitFloat64Binop(node);
case IrOpcode::kFloat64Sqrt:
return VisitUnop(node, kMachFloat64, kMachFloat64);
case IrOpcode::kFloat64Equal: case IrOpcode::kFloat64Equal:
case IrOpcode::kFloat64LessThan: case IrOpcode::kFloat64LessThan:
case IrOpcode::kFloat64LessThanOrEqual: case IrOpcode::kFloat64LessThanOrEqual:
......
...@@ -447,6 +447,15 @@ void CodeGenerator::AssembleArchInstruction(Instruction* instr) { ...@@ -447,6 +447,15 @@ void CodeGenerator::AssembleArchInstruction(Instruction* instr) {
__ addq(rsp, Immediate(kDoubleSize)); __ addq(rsp, Immediate(kDoubleSize));
break; break;
} }
case kSSEFloat64Sqrt: {
RegisterOrOperand input = i.InputRegisterOrOperand(0);
if (input.type == kDoubleRegister) {
__ sqrtsd(i.OutputDoubleRegister(), input.double_reg);
} else {
__ sqrtsd(i.OutputDoubleRegister(), input.operand);
}
break;
}
case kSSEFloat64ToInt32: { case kSSEFloat64ToInt32: {
RegisterOrOperand input = i.InputRegisterOrOperand(0); RegisterOrOperand input = i.InputRegisterOrOperand(0);
if (input.type == kDoubleRegister) { if (input.type == kDoubleRegister) {
......
...@@ -50,6 +50,7 @@ namespace compiler { ...@@ -50,6 +50,7 @@ namespace compiler {
V(SSEFloat64Mul) \ V(SSEFloat64Mul) \
V(SSEFloat64Div) \ V(SSEFloat64Div) \
V(SSEFloat64Mod) \ V(SSEFloat64Mod) \
V(SSEFloat64Sqrt) \
V(SSEFloat64ToInt32) \ V(SSEFloat64ToInt32) \
V(SSEFloat64ToUint32) \ V(SSEFloat64ToUint32) \
V(SSEInt32ToFloat64) \ V(SSEInt32ToFloat64) \
......
...@@ -559,6 +559,12 @@ void InstructionSelector::VisitFloat64Mod(Node* node) { ...@@ -559,6 +559,12 @@ void InstructionSelector::VisitFloat64Mod(Node* node) {
} }
void InstructionSelector::VisitFloat64Sqrt(Node* node) {
X64OperandGenerator g(this);
Emit(kSSEFloat64Sqrt, g.DefineAsRegister(node), g.Use(node->InputAt(0)));
}
void InstructionSelector::VisitInt32AddWithOverflow(Node* node, void InstructionSelector::VisitInt32AddWithOverflow(Node* node,
FlagsContinuation* cont) { FlagsContinuation* cont) {
VisitBinop(this, node, kX64Add32, cont); VisitBinop(this, node, kX64Add32, cont);
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment