// Copyright 2012 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "src/code-stubs.h" #include <sstream> #include "src/ast/ast.h" #include "src/bootstrapper.h" #include "src/code-factory.h" #include "src/code-stub-assembler.h" #include "src/factory.h" #include "src/gdb-jit.h" #include "src/ic/handler-compiler.h" #include "src/ic/ic.h" #include "src/macro-assembler.h" namespace v8 { namespace internal { RUNTIME_FUNCTION(UnexpectedStubMiss) { FATAL("Unexpected deopt of a stub"); return Smi::FromInt(0); } CodeStubDescriptor::CodeStubDescriptor(CodeStub* stub) : isolate_(stub->isolate()), call_descriptor_(stub->GetCallInterfaceDescriptor()), stack_parameter_count_(no_reg), hint_stack_parameter_count_(-1), function_mode_(NOT_JS_FUNCTION_STUB_MODE), deoptimization_handler_(NULL), miss_handler_(), has_miss_handler_(false) { stub->InitializeDescriptor(this); } CodeStubDescriptor::CodeStubDescriptor(Isolate* isolate, uint32_t stub_key) : isolate_(isolate), stack_parameter_count_(no_reg), hint_stack_parameter_count_(-1), function_mode_(NOT_JS_FUNCTION_STUB_MODE), deoptimization_handler_(NULL), miss_handler_(), has_miss_handler_(false) { CodeStub::InitializeDescriptor(isolate, stub_key, this); } void CodeStubDescriptor::Initialize(Address deoptimization_handler, int hint_stack_parameter_count, StubFunctionMode function_mode) { deoptimization_handler_ = deoptimization_handler; hint_stack_parameter_count_ = hint_stack_parameter_count; function_mode_ = function_mode; } void CodeStubDescriptor::Initialize(Register stack_parameter_count, Address deoptimization_handler, int hint_stack_parameter_count, StubFunctionMode function_mode) { Initialize(deoptimization_handler, hint_stack_parameter_count, function_mode); stack_parameter_count_ = stack_parameter_count; } bool CodeStub::FindCodeInCache(Code** code_out) { UnseededNumberDictionary* stubs = isolate()->heap()->code_stubs(); int index = stubs->FindEntry(GetKey()); if (index != UnseededNumberDictionary::kNotFound) { *code_out = Code::cast(stubs->ValueAt(index)); return true; } return false; } void CodeStub::RecordCodeGeneration(Handle<Code> code) { std::ostringstream os; os << *this; PROFILE(isolate(), CodeCreateEvent(CodeEventListener::STUB_TAG, AbstractCode::cast(*code), os.str().c_str())); Counters* counters = isolate()->counters(); counters->total_stubs_code_size()->Increment(code->instruction_size()); #ifdef DEBUG code->VerifyEmbeddedObjects(); #endif } Code::Kind CodeStub::GetCodeKind() const { return Code::STUB; } Code::Flags CodeStub::GetCodeFlags() const { return Code::ComputeFlags(GetCodeKind(), GetExtraICState()); } Handle<Code> CodeStub::GetCodeCopy(const Code::FindAndReplacePattern& pattern) { Handle<Code> ic = GetCode(); ic = isolate()->factory()->CopyCode(ic); ic->FindAndReplace(pattern); RecordCodeGeneration(ic); return ic; } Handle<Code> PlatformCodeStub::GenerateCode() { Factory* factory = isolate()->factory(); // Generate the new code. MacroAssembler masm(isolate(), NULL, 256, CodeObjectRequired::kYes); { // Update the static counter each time a new code stub is generated. isolate()->counters()->code_stubs()->Increment(); // Generate the code for the stub. masm.set_generating_stub(true); // TODO(yangguo): remove this once we can serialize IC stubs. masm.enable_serializer(); NoCurrentFrameScope scope(&masm); Generate(&masm); } // Create the code object. CodeDesc desc; masm.GetCode(&desc); // Copy the generated code into a heap object. Code::Flags flags = Code::ComputeFlags(GetCodeKind(), GetExtraICState()); Handle<Code> new_object = factory->NewCode( desc, flags, masm.CodeObject(), NeedsImmovableCode()); return new_object; } Handle<Code> CodeStub::GetCode() { Heap* heap = isolate()->heap(); Code* code; if (UseSpecialCache() ? FindCodeInSpecialCache(&code) : FindCodeInCache(&code)) { DCHECK(GetCodeKind() == code->kind()); return Handle<Code>(code); } { HandleScope scope(isolate()); Handle<Code> new_object = GenerateCode(); new_object->set_stub_key(GetKey()); FinishCode(new_object); RecordCodeGeneration(new_object); #ifdef ENABLE_DISASSEMBLER if (FLAG_print_code_stubs) { CodeTracer::Scope trace_scope(isolate()->GetCodeTracer()); OFStream os(trace_scope.file()); std::ostringstream name; name << *this; new_object->Disassemble(name.str().c_str(), os); os << "\n"; } #endif if (UseSpecialCache()) { AddToSpecialCache(new_object); } else { // Update the dictionary and the root in Heap. Handle<UnseededNumberDictionary> dict = UnseededNumberDictionary::AtNumberPut( Handle<UnseededNumberDictionary>(heap->code_stubs()), GetKey(), new_object); heap->SetRootCodeStubs(*dict); } code = *new_object; } Activate(code); DCHECK(!NeedsImmovableCode() || heap->lo_space()->Contains(code) || heap->code_space()->FirstPage()->Contains(code->address())); return Handle<Code>(code, isolate()); } const char* CodeStub::MajorName(CodeStub::Major major_key) { switch (major_key) { #define DEF_CASE(name) case name: return #name "Stub"; CODE_STUB_LIST(DEF_CASE) #undef DEF_CASE case NoCache: return "<NoCache>Stub"; case NUMBER_OF_IDS: UNREACHABLE(); return NULL; } return NULL; } void CodeStub::PrintBaseName(std::ostream& os) const { // NOLINT os << MajorName(MajorKey()); } void CodeStub::PrintName(std::ostream& os) const { // NOLINT PrintBaseName(os); PrintState(os); } void CodeStub::Dispatch(Isolate* isolate, uint32_t key, void** value_out, DispatchedCall call) { switch (MajorKeyFromKey(key)) { #define DEF_CASE(NAME) \ case NAME: { \ NAME##Stub stub(key, isolate); \ CodeStub* pstub = &stub; \ call(pstub, value_out); \ break; \ } CODE_STUB_LIST(DEF_CASE) #undef DEF_CASE case NUMBER_OF_IDS: case NoCache: UNREACHABLE(); break; } } static void InitializeDescriptorDispatchedCall(CodeStub* stub, void** value_out) { CodeStubDescriptor* descriptor_out = reinterpret_cast<CodeStubDescriptor*>(value_out); stub->InitializeDescriptor(descriptor_out); descriptor_out->set_call_descriptor(stub->GetCallInterfaceDescriptor()); } void CodeStub::InitializeDescriptor(Isolate* isolate, uint32_t key, CodeStubDescriptor* desc) { void** value_out = reinterpret_cast<void**>(desc); Dispatch(isolate, key, value_out, &InitializeDescriptorDispatchedCall); } void CodeStub::GetCodeDispatchCall(CodeStub* stub, void** value_out) { Handle<Code>* code_out = reinterpret_cast<Handle<Code>*>(value_out); // Code stubs with special cache cannot be recreated from stub key. *code_out = stub->UseSpecialCache() ? Handle<Code>() : stub->GetCode(); } MaybeHandle<Code> CodeStub::GetCode(Isolate* isolate, uint32_t key) { HandleScope scope(isolate); Handle<Code> code; void** value_out = reinterpret_cast<void**>(&code); Dispatch(isolate, key, value_out, &GetCodeDispatchCall); return scope.CloseAndEscape(code); } // static void BinaryOpICStub::GenerateAheadOfTime(Isolate* isolate) { if (FLAG_minimal) return; // Generate the uninitialized versions of the stub. for (int op = Token::BIT_OR; op <= Token::MOD; ++op) { BinaryOpICStub stub(isolate, static_cast<Token::Value>(op)); stub.GetCode(); } // Generate special versions of the stub. BinaryOpICState::GenerateAheadOfTime(isolate, &GenerateAheadOfTime); } void BinaryOpICStub::PrintState(std::ostream& os) const { // NOLINT os << state(); } // static void BinaryOpICStub::GenerateAheadOfTime(Isolate* isolate, const BinaryOpICState& state) { if (FLAG_minimal) return; BinaryOpICStub stub(isolate, state); stub.GetCode(); } // static void BinaryOpICWithAllocationSiteStub::GenerateAheadOfTime(Isolate* isolate) { // Generate special versions of the stub. BinaryOpICState::GenerateAheadOfTime(isolate, &GenerateAheadOfTime); } void BinaryOpICWithAllocationSiteStub::PrintState( std::ostream& os) const { // NOLINT os << state(); } // static void BinaryOpICWithAllocationSiteStub::GenerateAheadOfTime( Isolate* isolate, const BinaryOpICState& state) { if (state.CouldCreateAllocationMementos()) { BinaryOpICWithAllocationSiteStub stub(isolate, state); stub.GetCode(); } } std::ostream& operator<<(std::ostream& os, const StringAddFlags& flags) { switch (flags) { case STRING_ADD_CHECK_NONE: return os << "CheckNone"; case STRING_ADD_CHECK_LEFT: return os << "CheckLeft"; case STRING_ADD_CHECK_RIGHT: return os << "CheckRight"; case STRING_ADD_CHECK_BOTH: return os << "CheckBoth"; case STRING_ADD_CONVERT_LEFT: return os << "ConvertLeft"; case STRING_ADD_CONVERT_RIGHT: return os << "ConvertRight"; case STRING_ADD_CONVERT: break; } UNREACHABLE(); return os; } void StringAddStub::PrintBaseName(std::ostream& os) const { // NOLINT os << "StringAddStub_" << flags() << "_" << pretenure_flag(); } InlineCacheState CompareICStub::GetICState() const { CompareICState::State state = Max(left(), right()); switch (state) { case CompareICState::UNINITIALIZED: return ::v8::internal::UNINITIALIZED; case CompareICState::BOOLEAN: case CompareICState::SMI: case CompareICState::NUMBER: case CompareICState::INTERNALIZED_STRING: case CompareICState::STRING: case CompareICState::UNIQUE_NAME: case CompareICState::RECEIVER: case CompareICState::KNOWN_RECEIVER: return MONOMORPHIC; case CompareICState::GENERIC: return ::v8::internal::GENERIC; } UNREACHABLE(); return ::v8::internal::UNINITIALIZED; } Condition CompareICStub::GetCondition() const { return CompareIC::ComputeCondition(op()); } void CompareICStub::Generate(MacroAssembler* masm) { switch (state()) { case CompareICState::UNINITIALIZED: GenerateMiss(masm); break; case CompareICState::BOOLEAN: GenerateBooleans(masm); break; case CompareICState::SMI: GenerateSmis(masm); break; case CompareICState::NUMBER: GenerateNumbers(masm); break; case CompareICState::STRING: GenerateStrings(masm); break; case CompareICState::INTERNALIZED_STRING: GenerateInternalizedStrings(masm); break; case CompareICState::UNIQUE_NAME: GenerateUniqueNames(masm); break; case CompareICState::RECEIVER: GenerateReceivers(masm); break; case CompareICState::KNOWN_RECEIVER: DCHECK(*known_map_ != NULL); GenerateKnownReceivers(masm); break; case CompareICState::GENERIC: GenerateGeneric(masm); break; } } Handle<Code> TurboFanCodeStub::GenerateCode() { const char* name = CodeStub::MajorName(MajorKey()); Zone zone(isolate()->allocator()); CallInterfaceDescriptor descriptor(GetCallInterfaceDescriptor()); CodeStubAssembler assembler(isolate(), &zone, descriptor, GetCodeFlags(), name); GenerateAssembly(&assembler); return assembler.GenerateCode(); } void LoadICTrampolineTFStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* name = assembler->Parameter(Descriptor::kName); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* context = assembler->Parameter(Descriptor::kContext); Node* vector = assembler->LoadTypeFeedbackVectorForStub(); CodeStubAssembler::LoadICParameters p(context, receiver, name, slot, vector); assembler->LoadIC(&p); } void LoadICTFStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* name = assembler->Parameter(Descriptor::kName); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); CodeStubAssembler::LoadICParameters p(context, receiver, name, slot, vector); assembler->LoadIC(&p); } void LoadGlobalICTrampolineStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* slot = assembler->Parameter(Descriptor::kSlot); Node* context = assembler->Parameter(Descriptor::kContext); Node* vector = assembler->LoadTypeFeedbackVectorForStub(); CodeStubAssembler::LoadICParameters p(context, nullptr, nullptr, slot, vector); assembler->LoadGlobalIC(&p); } void LoadGlobalICStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); CodeStubAssembler::LoadICParameters p(context, nullptr, nullptr, slot, vector); assembler->LoadGlobalIC(&p); } void KeyedLoadICTrampolineTFStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* name = assembler->Parameter(Descriptor::kName); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* context = assembler->Parameter(Descriptor::kContext); Node* vector = assembler->LoadTypeFeedbackVectorForStub(); CodeStubAssembler::LoadICParameters p(context, receiver, name, slot, vector); assembler->KeyedLoadIC(&p); } void KeyedLoadICTFStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* name = assembler->Parameter(Descriptor::kName); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); CodeStubAssembler::LoadICParameters p(context, receiver, name, slot, vector); assembler->KeyedLoadIC(&p); } void StoreTransitionStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* name = assembler->Parameter(Descriptor::kName); Node* value = assembler->Parameter(Descriptor::kValue); Node* map = assembler->Parameter(Descriptor::kMap); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); StoreMode store_mode = this->store_mode(); Node* prepared_value = value; Label miss(assembler); bool needs_miss_case = false; if (store_mode != StoreTransitionStub::StoreMapOnly) { Representation representation = this->representation(); FieldIndex index = this->index(); assembler->Comment( "Prepare value for write: representation: %s, index.is_inobject: %d, " "index.offset: %d", representation.Mnemonic(), index.is_inobject(), index.offset()); prepared_value = assembler->PrepareValueForWrite(prepared_value, representation, &miss); // Only store to tagged field never bails out. needs_miss_case |= !representation.IsTagged(); } switch (store_mode) { case StoreTransitionStub::ExtendStorageAndStoreMapAndValue: assembler->Comment("Extend storage"); assembler->ExtendPropertiesBackingStore(receiver); // Fall through. case StoreTransitionStub::StoreMapAndValue: assembler->Comment("Store value"); // Store the new value into the "extended" object. assembler->StoreNamedField(receiver, index(), representation(), prepared_value, true); // Fall through. case StoreTransitionStub::StoreMapOnly: assembler->Comment("Store map"); // And finally update the map. assembler->StoreObjectField(receiver, JSObject::kMapOffset, map); break; } assembler->Return(value); if (needs_miss_case) { assembler->Bind(&miss); { assembler->Comment("Miss"); assembler->TailCallRuntime(Runtime::kStoreIC_Miss, context, value, slot, vector, receiver, name); } } } void ElementsTransitionAndStoreStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* key = assembler->Parameter(Descriptor::kName); Node* value = assembler->Parameter(Descriptor::kValue); Node* map = assembler->Parameter(Descriptor::kMap); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); assembler->Comment( "ElementsTransitionAndStoreStub: from_kind=%s, to_kind=%s," " is_jsarray=%d, store_mode=%d", ElementsKindToString(from_kind()), ElementsKindToString(to_kind()), is_jsarray(), store_mode()); Label miss(assembler); if (FLAG_trace_elements_transitions) { // Tracing elements transitions is the job of the runtime. assembler->Goto(&miss); } else { assembler->TransitionElementsKind(receiver, map, from_kind(), to_kind(), is_jsarray(), &miss); assembler->EmitElementStore(receiver, key, value, is_jsarray(), to_kind(), store_mode(), &miss); assembler->Return(value); } assembler->Bind(&miss); { assembler->Comment("Miss"); assembler->TailCallRuntime(Runtime::kElementsTransitionAndStoreIC_Miss, context, receiver, key, value, map, slot, vector); } } void AllocateHeapNumberStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* result = assembler->AllocateHeapNumber(); assembler->Return(result); } #define SIMD128_GEN_ASM(TYPE, Type, type, lane_count, lane_type) \ void Allocate##Type##Stub::GenerateAssembly(CodeStubAssembler* assembler) \ const { \ compiler::Node* result = \ assembler->Allocate(Simd128Value::kSize, CodeStubAssembler::kNone); \ compiler::Node* map = assembler->LoadMap(result); \ assembler->StoreNoWriteBarrier( \ MachineRepresentation::kTagged, map, \ assembler->HeapConstant(isolate()->factory()->type##_map())); \ assembler->Return(result); \ } SIMD128_TYPES(SIMD128_GEN_ASM) #undef SIMD128_GEN_ASM void StringLengthStub::GenerateAssembly(CodeStubAssembler* assembler) const { compiler::Node* value = assembler->Parameter(0); compiler::Node* string = assembler->LoadJSValueValue(value); compiler::Node* result = assembler->LoadStringLength(string); assembler->Return(result); } // static compiler::Node* AddStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; // Shared entry for floating point addition. Label do_fadd(assembler); Variable var_fadd_lhs(assembler, MachineRepresentation::kFloat64), var_fadd_rhs(assembler, MachineRepresentation::kFloat64); // We might need to loop several times due to ToPrimitive, ToString and/or // ToNumber conversions. Variable var_lhs(assembler, MachineRepresentation::kTagged), var_rhs(assembler, MachineRepresentation::kTagged), var_result(assembler, MachineRepresentation::kTagged); Variable* loop_vars[2] = {&var_lhs, &var_rhs}; Label loop(assembler, 2, loop_vars), end(assembler), string_add_convert_left(assembler, Label::kDeferred), string_add_convert_right(assembler, Label::kDeferred); var_lhs.Bind(left); var_rhs.Bind(right); assembler->Goto(&loop); assembler->Bind(&loop); { // Load the current {lhs} and {rhs} values. Node* lhs = var_lhs.value(); Node* rhs = var_rhs.value(); // Check if the {lhs} is a Smi or a HeapObject. Label if_lhsissmi(assembler), if_lhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(lhs), &if_lhsissmi, &if_lhsisnotsmi); assembler->Bind(&if_lhsissmi); { // Check if the {rhs} is also a Smi. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Try fast Smi addition first. Node* pair = assembler->SmiAddWithOverflow(lhs, rhs); Node* overflow = assembler->Projection(1, pair); // Check if the Smi additon overflowed. Label if_overflow(assembler), if_notoverflow(assembler); assembler->Branch(overflow, &if_overflow, &if_notoverflow); assembler->Bind(&if_overflow); { var_fadd_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fadd_rhs.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fadd); } assembler->Bind(&if_notoverflow); var_result.Bind(assembler->Projection(0, pair)); assembler->Goto(&end); } assembler->Bind(&if_rhsisnotsmi); { // Load the map of {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if the {rhs} is a HeapNumber. Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler, Label::kDeferred); Node* number_map = assembler->HeapNumberMapConstant(); assembler->Branch(assembler->WordEqual(rhs_map, number_map), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { var_fadd_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fadd_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fadd); } assembler->Bind(&if_rhsisnotnumber); { // Load the instance type of {rhs}. Node* rhs_instance_type = assembler->LoadMapInstanceType(rhs_map); // Check if the {rhs} is a String. Label if_rhsisstring(assembler, Label::kDeferred), if_rhsisnotstring(assembler, Label::kDeferred); assembler->Branch(assembler->Int32LessThan( rhs_instance_type, assembler->Int32Constant(FIRST_NONSTRING_TYPE)), &if_rhsisstring, &if_rhsisnotstring); assembler->Bind(&if_rhsisstring); { var_lhs.Bind(lhs); var_rhs.Bind(rhs); assembler->Goto(&string_add_convert_left); } assembler->Bind(&if_rhsisnotstring); { // Check if {rhs} is a JSReceiver. Label if_rhsisreceiver(assembler, Label::kDeferred), if_rhsisnotreceiver(assembler, Label::kDeferred); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), rhs_instance_type), &if_rhsisreceiver, &if_rhsisnotreceiver); assembler->Bind(&if_rhsisreceiver); { // Convert {rhs} to a primitive first passing no hint. Callable callable = CodeFactory::NonPrimitiveToPrimitive(assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotreceiver); { // Convert {rhs} to a Number first. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } } } } } assembler->Bind(&if_lhsisnotsmi); { // Load the map and instance type of {lhs}. Node* lhs_instance_type = assembler->LoadInstanceType(lhs); // Check if {lhs} is a String. Label if_lhsisstring(assembler), if_lhsisnotstring(assembler); assembler->Branch(assembler->Int32LessThan( lhs_instance_type, assembler->Int32Constant(FIRST_NONSTRING_TYPE)), &if_lhsisstring, &if_lhsisnotstring); assembler->Bind(&if_lhsisstring); { var_lhs.Bind(lhs); var_rhs.Bind(rhs); assembler->Goto(&string_add_convert_right); } assembler->Bind(&if_lhsisnotstring); { // Check if {rhs} is a Smi. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Check if {lhs} is a Number. Label if_lhsisnumber(assembler), if_lhsisnotnumber(assembler, Label::kDeferred); assembler->Branch(assembler->Word32Equal( lhs_instance_type, assembler->Int32Constant(HEAP_NUMBER_TYPE)), &if_lhsisnumber, &if_lhsisnotnumber); assembler->Bind(&if_lhsisnumber); { // The {lhs} is a HeapNumber, the {rhs} is a Smi, just add them. var_fadd_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fadd_rhs.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fadd); } assembler->Bind(&if_lhsisnotnumber); { // The {lhs} is neither a Number nor a String, and the {rhs} is a // Smi. Label if_lhsisreceiver(assembler, Label::kDeferred), if_lhsisnotreceiver(assembler, Label::kDeferred); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), lhs_instance_type), &if_lhsisreceiver, &if_lhsisnotreceiver); assembler->Bind(&if_lhsisreceiver); { // Convert {lhs} to a primitive first passing no hint. Callable callable = CodeFactory::NonPrimitiveToPrimitive(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); assembler->Goto(&loop); } assembler->Bind(&if_lhsisnotreceiver); { // Convert {lhs} to a Number first. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); assembler->Goto(&loop); } } } assembler->Bind(&if_rhsisnotsmi); { // Load the instance type of {rhs}. Node* rhs_instance_type = assembler->LoadInstanceType(rhs); // Check if {rhs} is a String. Label if_rhsisstring(assembler), if_rhsisnotstring(assembler); assembler->Branch(assembler->Int32LessThan( rhs_instance_type, assembler->Int32Constant(FIRST_NONSTRING_TYPE)), &if_rhsisstring, &if_rhsisnotstring); assembler->Bind(&if_rhsisstring); { var_lhs.Bind(lhs); var_rhs.Bind(rhs); assembler->Goto(&string_add_convert_left); } assembler->Bind(&if_rhsisnotstring); { // Check if {lhs} is a HeapNumber. Label if_lhsisnumber(assembler), if_lhsisnotnumber(assembler); assembler->Branch(assembler->Word32Equal( lhs_instance_type, assembler->Int32Constant(HEAP_NUMBER_TYPE)), &if_lhsisnumber, &if_lhsisnotnumber); assembler->Bind(&if_lhsisnumber); { // Check if {rhs} is also a HeapNumber. Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler, Label::kDeferred); assembler->Branch(assembler->Word32Equal( rhs_instance_type, assembler->Int32Constant(HEAP_NUMBER_TYPE)), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { // Perform a floating point addition. var_fadd_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fadd_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fadd); } assembler->Bind(&if_rhsisnotnumber); { // Check if {rhs} is a JSReceiver. Label if_rhsisreceiver(assembler, Label::kDeferred), if_rhsisnotreceiver(assembler, Label::kDeferred); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), rhs_instance_type), &if_rhsisreceiver, &if_rhsisnotreceiver); assembler->Bind(&if_rhsisreceiver); { // Convert {rhs} to a primitive first passing no hint. Callable callable = CodeFactory::NonPrimitiveToPrimitive( assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotreceiver); { // Convert {rhs} to a Number first. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } } } assembler->Bind(&if_lhsisnotnumber); { // Check if {lhs} is a JSReceiver. Label if_lhsisreceiver(assembler, Label::kDeferred), if_lhsisnotreceiver(assembler); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), lhs_instance_type), &if_lhsisreceiver, &if_lhsisnotreceiver); assembler->Bind(&if_lhsisreceiver); { // Convert {lhs} to a primitive first passing no hint. Callable callable = CodeFactory::NonPrimitiveToPrimitive(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); assembler->Goto(&loop); } assembler->Bind(&if_lhsisnotreceiver); { // Check if {rhs} is a JSReceiver. Label if_rhsisreceiver(assembler, Label::kDeferred), if_rhsisnotreceiver(assembler, Label::kDeferred); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), rhs_instance_type), &if_rhsisreceiver, &if_rhsisnotreceiver); assembler->Bind(&if_rhsisreceiver); { // Convert {rhs} to a primitive first passing no hint. Callable callable = CodeFactory::NonPrimitiveToPrimitive( assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotreceiver); { // Convert {lhs} to a Number first. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); assembler->Goto(&loop); } } } } } } } } assembler->Bind(&string_add_convert_left); { // Convert {lhs}, which is a Smi, to a String and concatenate the // resulting string with the String {rhs}. Callable callable = CodeFactory::StringAdd( assembler->isolate(), STRING_ADD_CONVERT_LEFT, NOT_TENURED); var_result.Bind(assembler->CallStub(callable, context, var_lhs.value(), var_rhs.value())); assembler->Goto(&end); } assembler->Bind(&string_add_convert_right); { // Convert {lhs}, which is a Smi, to a String and concatenate the // resulting string with the String {rhs}. Callable callable = CodeFactory::StringAdd( assembler->isolate(), STRING_ADD_CONVERT_RIGHT, NOT_TENURED); var_result.Bind(assembler->CallStub(callable, context, var_lhs.value(), var_rhs.value())); assembler->Goto(&end); } assembler->Bind(&do_fadd); { Node* lhs_value = var_fadd_lhs.value(); Node* rhs_value = var_fadd_rhs.value(); Node* value = assembler->Float64Add(lhs_value, rhs_value); Node* result = assembler->ChangeFloat64ToTagged(value); var_result.Bind(result); assembler->Goto(&end); } assembler->Bind(&end); return var_result.value(); } // static compiler::Node* AddWithFeedbackStub::Generate( CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* slot_id, compiler::Node* type_feedback_vector, compiler::Node* context) { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; // Shared entry for floating point addition. Label do_fadd(assembler), end(assembler), call_add_stub(assembler, Label::kDeferred); Variable var_fadd_lhs(assembler, MachineRepresentation::kFloat64), var_fadd_rhs(assembler, MachineRepresentation::kFloat64), var_type_feedback(assembler, MachineRepresentation::kWord32), var_result(assembler, MachineRepresentation::kTagged); // Check if the {lhs} is a Smi or a HeapObject. Label if_lhsissmi(assembler), if_lhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(lhs), &if_lhsissmi, &if_lhsisnotsmi); assembler->Bind(&if_lhsissmi); { // Check if the {rhs} is also a Smi. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Try fast Smi addition first. Node* pair = assembler->SmiAddWithOverflow(lhs, rhs); Node* overflow = assembler->Projection(1, pair); // Check if the Smi additon overflowed. Label if_overflow(assembler), if_notoverflow(assembler); assembler->Branch(overflow, &if_overflow, &if_notoverflow); assembler->Bind(&if_overflow); { var_fadd_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fadd_rhs.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fadd); } assembler->Bind(&if_notoverflow); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kSignedSmall)); var_result.Bind(assembler->Projection(0, pair)); assembler->Goto(&end); } } assembler->Bind(&if_rhsisnotsmi); { // Load the map of {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if the {rhs} is a HeapNumber. assembler->GotoUnless( assembler->WordEqual(rhs_map, assembler->HeapNumberMapConstant()), &call_add_stub); var_fadd_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fadd_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fadd); } } assembler->Bind(&if_lhsisnotsmi); { // Load the map of {lhs}. Node* lhs_map = assembler->LoadMap(lhs); // Check if {lhs} is a HeapNumber. Label if_lhsisnumber(assembler), if_lhsisnotnumber(assembler); assembler->GotoUnless( assembler->WordEqual(lhs_map, assembler->HeapNumberMapConstant()), &call_add_stub); // Check if the {rhs} is Smi. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { var_fadd_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fadd_rhs.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fadd); } assembler->Bind(&if_rhsisnotsmi); { // Load the map of {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if the {rhs} is a HeapNumber. Node* number_map = assembler->HeapNumberMapConstant(); assembler->GotoUnless(assembler->WordEqual(rhs_map, number_map), &call_add_stub); var_fadd_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fadd_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fadd); } } assembler->Bind(&do_fadd); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kNumber)); Node* value = assembler->Float64Add(var_fadd_lhs.value(), var_fadd_rhs.value()); Node* result = assembler->ChangeFloat64ToTagged(value); var_result.Bind(result); assembler->Goto(&end); } assembler->Bind(&call_add_stub); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kAny)); Callable callable = CodeFactory::Add(assembler->isolate()); var_result.Bind(assembler->CallStub(callable, context, lhs, rhs)); assembler->Goto(&end); } assembler->Bind(&end); assembler->UpdateFeedback(var_type_feedback.value(), type_feedback_vector, slot_id); return var_result.value(); } // static compiler::Node* SubtractStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; // Shared entry for floating point subtraction. Label do_fsub(assembler), end(assembler); Variable var_fsub_lhs(assembler, MachineRepresentation::kFloat64), var_fsub_rhs(assembler, MachineRepresentation::kFloat64); // We might need to loop several times due to ToPrimitive and/or ToNumber // conversions. Variable var_lhs(assembler, MachineRepresentation::kTagged), var_rhs(assembler, MachineRepresentation::kTagged), var_result(assembler, MachineRepresentation::kTagged); Variable* loop_vars[2] = {&var_lhs, &var_rhs}; Label loop(assembler, 2, loop_vars); var_lhs.Bind(left); var_rhs.Bind(right); assembler->Goto(&loop); assembler->Bind(&loop); { // Load the current {lhs} and {rhs} values. Node* lhs = var_lhs.value(); Node* rhs = var_rhs.value(); // Check if the {lhs} is a Smi or a HeapObject. Label if_lhsissmi(assembler), if_lhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(lhs), &if_lhsissmi, &if_lhsisnotsmi); assembler->Bind(&if_lhsissmi); { // Check if the {rhs} is also a Smi. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Try a fast Smi subtraction first. Node* pair = assembler->SmiSubWithOverflow(lhs, rhs); Node* overflow = assembler->Projection(1, pair); // Check if the Smi subtraction overflowed. Label if_overflow(assembler), if_notoverflow(assembler); assembler->Branch(overflow, &if_overflow, &if_notoverflow); assembler->Bind(&if_overflow); { // The result doesn't fit into Smi range. var_fsub_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fsub_rhs.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fsub); } assembler->Bind(&if_notoverflow); var_result.Bind(assembler->Projection(0, pair)); assembler->Goto(&end); } assembler->Bind(&if_rhsisnotsmi); { // Load the map of the {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if {rhs} is a HeapNumber. Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler, Label::kDeferred); Node* number_map = assembler->HeapNumberMapConstant(); assembler->Branch(assembler->WordEqual(rhs_map, number_map), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { // Perform a floating point subtraction. var_fsub_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fsub_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fsub); } assembler->Bind(&if_rhsisnotnumber); { // Convert the {rhs} to a Number first. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } } } assembler->Bind(&if_lhsisnotsmi); { // Load the map of the {lhs}. Node* lhs_map = assembler->LoadMap(lhs); // Check if the {lhs} is a HeapNumber. Label if_lhsisnumber(assembler), if_lhsisnotnumber(assembler, Label::kDeferred); Node* number_map = assembler->HeapNumberMapConstant(); assembler->Branch(assembler->WordEqual(lhs_map, number_map), &if_lhsisnumber, &if_lhsisnotnumber); assembler->Bind(&if_lhsisnumber); { // Check if the {rhs} is a Smi. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Perform a floating point subtraction. var_fsub_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fsub_rhs.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fsub); } assembler->Bind(&if_rhsisnotsmi); { // Load the map of the {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if the {rhs} is a HeapNumber. Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(rhs_map, number_map), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { // Perform a floating point subtraction. var_fsub_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fsub_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fsub); } assembler->Bind(&if_rhsisnotnumber); { // Convert the {rhs} to a Number first. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } } } assembler->Bind(&if_lhsisnotnumber); { // Convert the {lhs} to a Number first. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); assembler->Goto(&loop); } } } assembler->Bind(&do_fsub); { Node* lhs_value = var_fsub_lhs.value(); Node* rhs_value = var_fsub_rhs.value(); Node* value = assembler->Float64Sub(lhs_value, rhs_value); var_result.Bind(assembler->ChangeFloat64ToTagged(value)); assembler->Goto(&end); } assembler->Bind(&end); return var_result.value(); } // static compiler::Node* SubtractWithFeedbackStub::Generate( CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* slot_id, compiler::Node* type_feedback_vector, compiler::Node* context) { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; // Shared entry for floating point subtraction. Label do_fsub(assembler), end(assembler), call_subtract_stub(assembler, Label::kDeferred); Variable var_fsub_lhs(assembler, MachineRepresentation::kFloat64), var_fsub_rhs(assembler, MachineRepresentation::kFloat64), var_type_feedback(assembler, MachineRepresentation::kWord32), var_result(assembler, MachineRepresentation::kTagged); // Check if the {lhs} is a Smi or a HeapObject. Label if_lhsissmi(assembler), if_lhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(lhs), &if_lhsissmi, &if_lhsisnotsmi); assembler->Bind(&if_lhsissmi); { // Check if the {rhs} is also a Smi. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Try a fast Smi subtraction first. Node* pair = assembler->SmiSubWithOverflow(lhs, rhs); Node* overflow = assembler->Projection(1, pair); // Check if the Smi subtraction overflowed. Label if_overflow(assembler), if_notoverflow(assembler); assembler->Branch(overflow, &if_overflow, &if_notoverflow); assembler->Bind(&if_overflow); { // lhs, rhs - smi and result - number. combined - number. // The result doesn't fit into Smi range. var_fsub_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fsub_rhs.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fsub); } assembler->Bind(&if_notoverflow); // lhs, rhs, result smi. combined - smi. var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kSignedSmall)); var_result.Bind(assembler->Projection(0, pair)); assembler->Goto(&end); } assembler->Bind(&if_rhsisnotsmi); { // Load the map of the {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if {rhs} is a HeapNumber. assembler->GotoUnless( assembler->WordEqual(rhs_map, assembler->HeapNumberMapConstant()), &call_subtract_stub); // Perform a floating point subtraction. var_fsub_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fsub_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fsub); } } assembler->Bind(&if_lhsisnotsmi); { // Load the map of the {lhs}. Node* lhs_map = assembler->LoadMap(lhs); // Check if the {lhs} is a HeapNumber. assembler->GotoUnless( assembler->WordEqual(lhs_map, assembler->HeapNumberMapConstant()), &call_subtract_stub); // Check if the {rhs} is a Smi. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Perform a floating point subtraction. var_fsub_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fsub_rhs.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fsub); } assembler->Bind(&if_rhsisnotsmi); { // Load the map of the {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if the {rhs} is a HeapNumber. assembler->GotoUnless( assembler->WordEqual(rhs_map, assembler->HeapNumberMapConstant()), &call_subtract_stub); // Perform a floating point subtraction. var_fsub_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fsub_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fsub); } } assembler->Bind(&do_fsub); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kNumber)); Node* lhs_value = var_fsub_lhs.value(); Node* rhs_value = var_fsub_rhs.value(); Node* value = assembler->Float64Sub(lhs_value, rhs_value); var_result.Bind(assembler->ChangeFloat64ToTagged(value)); assembler->Goto(&end); } assembler->Bind(&call_subtract_stub); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kAny)); Callable callable = CodeFactory::Subtract(assembler->isolate()); var_result.Bind(assembler->CallStub(callable, context, lhs, rhs)); assembler->Goto(&end); } assembler->Bind(&end); assembler->UpdateFeedback(var_type_feedback.value(), type_feedback_vector, slot_id); return var_result.value(); } // static compiler::Node* MultiplyStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { using compiler::Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; // Shared entry point for floating point multiplication. Label do_fmul(assembler), return_result(assembler); Variable var_lhs_float64(assembler, MachineRepresentation::kFloat64), var_rhs_float64(assembler, MachineRepresentation::kFloat64); Node* number_map = assembler->HeapNumberMapConstant(); // We might need to loop one or two times due to ToNumber conversions. Variable var_lhs(assembler, MachineRepresentation::kTagged), var_rhs(assembler, MachineRepresentation::kTagged), var_result(assembler, MachineRepresentation::kTagged); Variable* loop_variables[] = {&var_lhs, &var_rhs}; Label loop(assembler, 2, loop_variables); var_lhs.Bind(left); var_rhs.Bind(right); assembler->Goto(&loop); assembler->Bind(&loop); { Node* lhs = var_lhs.value(); Node* rhs = var_rhs.value(); Label lhs_is_smi(assembler), lhs_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(lhs), &lhs_is_smi, &lhs_is_not_smi); assembler->Bind(&lhs_is_smi); { Label rhs_is_smi(assembler), rhs_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &rhs_is_smi, &rhs_is_not_smi); assembler->Bind(&rhs_is_smi); { // Both {lhs} and {rhs} are Smis. The result is not necessarily a smi, // in case of overflow. var_result.Bind(assembler->SmiMul(lhs, rhs)); assembler->Goto(&return_result); } assembler->Bind(&rhs_is_not_smi); { Node* rhs_map = assembler->LoadMap(rhs); // Check if {rhs} is a HeapNumber. Label rhs_is_number(assembler), rhs_is_not_number(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(rhs_map, number_map), &rhs_is_number, &rhs_is_not_number); assembler->Bind(&rhs_is_number); { // Convert {lhs} to a double and multiply it with the value of {rhs}. var_lhs_float64.Bind(assembler->SmiToFloat64(lhs)); var_rhs_float64.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fmul); } assembler->Bind(&rhs_is_not_number); { // Multiplication is commutative, swap {lhs} with {rhs} and loop. var_lhs.Bind(rhs); var_rhs.Bind(lhs); assembler->Goto(&loop); } } } assembler->Bind(&lhs_is_not_smi); { Node* lhs_map = assembler->LoadMap(lhs); // Check if {lhs} is a HeapNumber. Label lhs_is_number(assembler), lhs_is_not_number(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(lhs_map, number_map), &lhs_is_number, &lhs_is_not_number); assembler->Bind(&lhs_is_number); { // Check if {rhs} is a Smi. Label rhs_is_smi(assembler), rhs_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &rhs_is_smi, &rhs_is_not_smi); assembler->Bind(&rhs_is_smi); { // Convert {rhs} to a double and multiply it with the value of {lhs}. var_lhs_float64.Bind(assembler->LoadHeapNumberValue(lhs)); var_rhs_float64.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fmul); } assembler->Bind(&rhs_is_not_smi); { Node* rhs_map = assembler->LoadMap(rhs); // Check if {rhs} is a HeapNumber. Label rhs_is_number(assembler), rhs_is_not_number(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(rhs_map, number_map), &rhs_is_number, &rhs_is_not_number); assembler->Bind(&rhs_is_number); { // Both {lhs} and {rhs} are HeapNumbers. Load their values and // multiply them. var_lhs_float64.Bind(assembler->LoadHeapNumberValue(lhs)); var_rhs_float64.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fmul); } assembler->Bind(&rhs_is_not_number); { // Multiplication is commutative, swap {lhs} with {rhs} and loop. var_lhs.Bind(rhs); var_rhs.Bind(lhs); assembler->Goto(&loop); } } } assembler->Bind(&lhs_is_not_number); { // Convert {lhs} to a Number and loop. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); assembler->Goto(&loop); } } } assembler->Bind(&do_fmul); { Node* value = assembler->Float64Mul(var_lhs_float64.value(), var_rhs_float64.value()); Node* result = assembler->ChangeFloat64ToTagged(value); var_result.Bind(result); assembler->Goto(&return_result); } assembler->Bind(&return_result); return var_result.value(); } // static compiler::Node* MultiplyWithFeedbackStub::Generate( CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* slot_id, compiler::Node* type_feedback_vector, compiler::Node* context) { using compiler::Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; // Shared entry point for floating point multiplication. Label do_fmul(assembler), end(assembler), call_multiply_stub(assembler, Label::kDeferred); Variable var_lhs_float64(assembler, MachineRepresentation::kFloat64), var_rhs_float64(assembler, MachineRepresentation::kFloat64), var_result(assembler, MachineRepresentation::kTagged), var_type_feedback(assembler, MachineRepresentation::kWord32); Node* number_map = assembler->HeapNumberMapConstant(); Label lhs_is_smi(assembler), lhs_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(lhs), &lhs_is_smi, &lhs_is_not_smi); assembler->Bind(&lhs_is_smi); { Label rhs_is_smi(assembler), rhs_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &rhs_is_smi, &rhs_is_not_smi); assembler->Bind(&rhs_is_smi); { // Both {lhs} and {rhs} are Smis. The result is not necessarily a smi, // in case of overflow. var_result.Bind(assembler->SmiMul(lhs, rhs)); var_type_feedback.Bind(assembler->Select( assembler->WordIsSmi(var_result.value()), assembler->Int32Constant(BinaryOperationFeedback::kSignedSmall), assembler->Int32Constant(BinaryOperationFeedback::kNumber), MachineRepresentation::kWord32)); assembler->Goto(&end); } assembler->Bind(&rhs_is_not_smi); { Node* rhs_map = assembler->LoadMap(rhs); // Check if {rhs} is a HeapNumber. assembler->GotoUnless(assembler->WordEqual(rhs_map, number_map), &call_multiply_stub); // Convert {lhs} to a double and multiply it with the value of {rhs}. var_lhs_float64.Bind(assembler->SmiToFloat64(lhs)); var_rhs_float64.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fmul); } } assembler->Bind(&lhs_is_not_smi); { Node* lhs_map = assembler->LoadMap(lhs); // Check if {lhs} is a HeapNumber. assembler->GotoUnless(assembler->WordEqual(lhs_map, number_map), &call_multiply_stub); // Check if {rhs} is a Smi. Label rhs_is_smi(assembler), rhs_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &rhs_is_smi, &rhs_is_not_smi); assembler->Bind(&rhs_is_smi); { // Convert {rhs} to a double and multiply it with the value of {lhs}. var_lhs_float64.Bind(assembler->LoadHeapNumberValue(lhs)); var_rhs_float64.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fmul); } assembler->Bind(&rhs_is_not_smi); { Node* rhs_map = assembler->LoadMap(rhs); // Check if {rhs} is a HeapNumber. assembler->GotoUnless(assembler->WordEqual(rhs_map, number_map), &call_multiply_stub); // Both {lhs} and {rhs} are HeapNumbers. Load their values and // multiply them. var_lhs_float64.Bind(assembler->LoadHeapNumberValue(lhs)); var_rhs_float64.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fmul); } } assembler->Bind(&do_fmul); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kNumber)); Node* value = assembler->Float64Mul(var_lhs_float64.value(), var_rhs_float64.value()); Node* result = assembler->ChangeFloat64ToTagged(value); var_result.Bind(result); assembler->Goto(&end); } assembler->Bind(&call_multiply_stub); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kAny)); Callable callable = CodeFactory::Multiply(assembler->isolate()); var_result.Bind(assembler->CallStub(callable, context, lhs, rhs)); assembler->Goto(&end); } assembler->Bind(&end); assembler->UpdateFeedback(var_type_feedback.value(), type_feedback_vector, slot_id); return var_result.value(); } // static compiler::Node* DivideStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { using compiler::Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; // Shared entry point for floating point division. Label do_fdiv(assembler), end(assembler); Variable var_dividend_float64(assembler, MachineRepresentation::kFloat64), var_divisor_float64(assembler, MachineRepresentation::kFloat64); Node* number_map = assembler->HeapNumberMapConstant(); // We might need to loop one or two times due to ToNumber conversions. Variable var_dividend(assembler, MachineRepresentation::kTagged), var_divisor(assembler, MachineRepresentation::kTagged), var_result(assembler, MachineRepresentation::kTagged); Variable* loop_variables[] = {&var_dividend, &var_divisor}; Label loop(assembler, 2, loop_variables); var_dividend.Bind(left); var_divisor.Bind(right); assembler->Goto(&loop); assembler->Bind(&loop); { Node* dividend = var_dividend.value(); Node* divisor = var_divisor.value(); Label dividend_is_smi(assembler), dividend_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(dividend), ÷nd_is_smi, ÷nd_is_not_smi); assembler->Bind(÷nd_is_smi); { Label divisor_is_smi(assembler), divisor_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(divisor), &divisor_is_smi, &divisor_is_not_smi); assembler->Bind(&divisor_is_smi); { Label bailout(assembler); // Do floating point division if {divisor} is zero. assembler->GotoIf( assembler->WordEqual(divisor, assembler->IntPtrConstant(0)), &bailout); // Do floating point division {dividend} is zero and {divisor} is // negative. Label dividend_is_zero(assembler), dividend_is_not_zero(assembler); assembler->Branch( assembler->WordEqual(dividend, assembler->IntPtrConstant(0)), ÷nd_is_zero, ÷nd_is_not_zero); assembler->Bind(÷nd_is_zero); { assembler->GotoIf( assembler->IntPtrLessThan(divisor, assembler->IntPtrConstant(0)), &bailout); assembler->Goto(÷nd_is_not_zero); } assembler->Bind(÷nd_is_not_zero); Node* untagged_divisor = assembler->SmiUntag(divisor); Node* untagged_dividend = assembler->SmiUntag(dividend); // Do floating point division if {dividend} is kMinInt (or kMinInt - 1 // if the Smi size is 31) and {divisor} is -1. Label divisor_is_minus_one(assembler), divisor_is_not_minus_one(assembler); assembler->Branch(assembler->Word32Equal(untagged_divisor, assembler->Int32Constant(-1)), &divisor_is_minus_one, &divisor_is_not_minus_one); assembler->Bind(&divisor_is_minus_one); { assembler->GotoIf( assembler->Word32Equal( untagged_dividend, assembler->Int32Constant( kSmiValueSize == 32 ? kMinInt : (kMinInt >> 1))), &bailout); assembler->Goto(&divisor_is_not_minus_one); } assembler->Bind(&divisor_is_not_minus_one); // TODO(epertoso): consider adding a machine instruction that returns // both the result and the remainder. Node* untagged_result = assembler->Int32Div(untagged_dividend, untagged_divisor); Node* truncated = assembler->IntPtrMul(untagged_result, untagged_divisor); // Do floating point division if the remainder is not 0. assembler->GotoIf( assembler->Word32NotEqual(untagged_dividend, truncated), &bailout); var_result.Bind(assembler->SmiTag(untagged_result)); assembler->Goto(&end); // Bailout: convert {dividend} and {divisor} to double and do double // division. assembler->Bind(&bailout); { var_dividend_float64.Bind(assembler->SmiToFloat64(dividend)); var_divisor_float64.Bind(assembler->SmiToFloat64(divisor)); assembler->Goto(&do_fdiv); } } assembler->Bind(&divisor_is_not_smi); { Node* divisor_map = assembler->LoadMap(divisor); // Check if {divisor} is a HeapNumber. Label divisor_is_number(assembler), divisor_is_not_number(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(divisor_map, number_map), &divisor_is_number, &divisor_is_not_number); assembler->Bind(&divisor_is_number); { // Convert {dividend} to a double and divide it with the value of // {divisor}. var_dividend_float64.Bind(assembler->SmiToFloat64(dividend)); var_divisor_float64.Bind(assembler->LoadHeapNumberValue(divisor)); assembler->Goto(&do_fdiv); } assembler->Bind(&divisor_is_not_number); { // Convert {divisor} to a number and loop. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_divisor.Bind(assembler->CallStub(callable, context, divisor)); assembler->Goto(&loop); } } } assembler->Bind(÷nd_is_not_smi); { Node* dividend_map = assembler->LoadMap(dividend); // Check if {dividend} is a HeapNumber. Label dividend_is_number(assembler), dividend_is_not_number(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(dividend_map, number_map), ÷nd_is_number, ÷nd_is_not_number); assembler->Bind(÷nd_is_number); { // Check if {divisor} is a Smi. Label divisor_is_smi(assembler), divisor_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(divisor), &divisor_is_smi, &divisor_is_not_smi); assembler->Bind(&divisor_is_smi); { // Convert {divisor} to a double and use it for a floating point // division. var_dividend_float64.Bind(assembler->LoadHeapNumberValue(dividend)); var_divisor_float64.Bind(assembler->SmiToFloat64(divisor)); assembler->Goto(&do_fdiv); } assembler->Bind(&divisor_is_not_smi); { Node* divisor_map = assembler->LoadMap(divisor); // Check if {divisor} is a HeapNumber. Label divisor_is_number(assembler), divisor_is_not_number(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(divisor_map, number_map), &divisor_is_number, &divisor_is_not_number); assembler->Bind(&divisor_is_number); { // Both {dividend} and {divisor} are HeapNumbers. Load their values // and divide them. var_dividend_float64.Bind(assembler->LoadHeapNumberValue(dividend)); var_divisor_float64.Bind(assembler->LoadHeapNumberValue(divisor)); assembler->Goto(&do_fdiv); } assembler->Bind(&divisor_is_not_number); { // Convert {divisor} to a number and loop. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_divisor.Bind(assembler->CallStub(callable, context, divisor)); assembler->Goto(&loop); } } } assembler->Bind(÷nd_is_not_number); { // Convert {dividend} to a Number and loop. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_dividend.Bind(assembler->CallStub(callable, context, dividend)); assembler->Goto(&loop); } } } assembler->Bind(&do_fdiv); { Node* value = assembler->Float64Div(var_dividend_float64.value(), var_divisor_float64.value()); var_result.Bind(assembler->ChangeFloat64ToTagged(value)); assembler->Goto(&end); } assembler->Bind(&end); return var_result.value(); } // static compiler::Node* DivideWithFeedbackStub::Generate( CodeStubAssembler* assembler, compiler::Node* dividend, compiler::Node* divisor, compiler::Node* slot_id, compiler::Node* type_feedback_vector, compiler::Node* context) { using compiler::Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; // Shared entry point for floating point division. Label do_fdiv(assembler), end(assembler), call_divide_stub(assembler); Variable var_dividend_float64(assembler, MachineRepresentation::kFloat64), var_divisor_float64(assembler, MachineRepresentation::kFloat64), var_result(assembler, MachineRepresentation::kTagged), var_type_feedback(assembler, MachineRepresentation::kWord32); Node* number_map = assembler->HeapNumberMapConstant(); Label dividend_is_smi(assembler), dividend_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(dividend), ÷nd_is_smi, ÷nd_is_not_smi); assembler->Bind(÷nd_is_smi); { Label divisor_is_smi(assembler), divisor_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(divisor), &divisor_is_smi, &divisor_is_not_smi); assembler->Bind(&divisor_is_smi); { Label bailout(assembler); // Do floating point division if {divisor} is zero. assembler->GotoIf( assembler->WordEqual(divisor, assembler->IntPtrConstant(0)), &bailout); // Do floating point division {dividend} is zero and {divisor} is // negative. Label dividend_is_zero(assembler), dividend_is_not_zero(assembler); assembler->Branch( assembler->WordEqual(dividend, assembler->IntPtrConstant(0)), ÷nd_is_zero, ÷nd_is_not_zero); assembler->Bind(÷nd_is_zero); { assembler->GotoIf( assembler->IntPtrLessThan(divisor, assembler->IntPtrConstant(0)), &bailout); assembler->Goto(÷nd_is_not_zero); } assembler->Bind(÷nd_is_not_zero); Node* untagged_divisor = assembler->SmiUntag(divisor); Node* untagged_dividend = assembler->SmiUntag(dividend); // Do floating point division if {dividend} is kMinInt (or kMinInt - 1 // if the Smi size is 31) and {divisor} is -1. Label divisor_is_minus_one(assembler), divisor_is_not_minus_one(assembler); assembler->Branch(assembler->Word32Equal(untagged_divisor, assembler->Int32Constant(-1)), &divisor_is_minus_one, &divisor_is_not_minus_one); assembler->Bind(&divisor_is_minus_one); { assembler->GotoIf( assembler->Word32Equal( untagged_dividend, assembler->Int32Constant(kSmiValueSize == 32 ? kMinInt : (kMinInt >> 1))), &bailout); assembler->Goto(&divisor_is_not_minus_one); } assembler->Bind(&divisor_is_not_minus_one); Node* untagged_result = assembler->Int32Div(untagged_dividend, untagged_divisor); Node* truncated = assembler->IntPtrMul(untagged_result, untagged_divisor); // Do floating point division if the remainder is not 0. assembler->GotoIf(assembler->Word32NotEqual(untagged_dividend, truncated), &bailout); var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kSignedSmall)); var_result.Bind(assembler->SmiTag(untagged_result)); assembler->Goto(&end); // Bailout: convert {dividend} and {divisor} to double and do double // division. assembler->Bind(&bailout); { var_dividend_float64.Bind(assembler->SmiToFloat64(dividend)); var_divisor_float64.Bind(assembler->SmiToFloat64(divisor)); assembler->Goto(&do_fdiv); } } assembler->Bind(&divisor_is_not_smi); { Node* divisor_map = assembler->LoadMap(divisor); // Check if {divisor} is a HeapNumber. assembler->GotoUnless(assembler->WordEqual(divisor_map, number_map), &call_divide_stub); // Convert {dividend} to a double and divide it with the value of // {divisor}. var_dividend_float64.Bind(assembler->SmiToFloat64(dividend)); var_divisor_float64.Bind(assembler->LoadHeapNumberValue(divisor)); assembler->Goto(&do_fdiv); } assembler->Bind(÷nd_is_not_smi); { Node* dividend_map = assembler->LoadMap(dividend); // Check if {dividend} is a HeapNumber. assembler->GotoUnless(assembler->WordEqual(dividend_map, number_map), &call_divide_stub); // Check if {divisor} is a Smi. Label divisor_is_smi(assembler), divisor_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(divisor), &divisor_is_smi, &divisor_is_not_smi); assembler->Bind(&divisor_is_smi); { // Convert {divisor} to a double and use it for a floating point // division. var_dividend_float64.Bind(assembler->LoadHeapNumberValue(dividend)); var_divisor_float64.Bind(assembler->SmiToFloat64(divisor)); assembler->Goto(&do_fdiv); } assembler->Bind(&divisor_is_not_smi); { Node* divisor_map = assembler->LoadMap(divisor); // Check if {divisor} is a HeapNumber. assembler->GotoUnless(assembler->WordEqual(divisor_map, number_map), &call_divide_stub); // Both {dividend} and {divisor} are HeapNumbers. Load their values // and divide them. var_dividend_float64.Bind(assembler->LoadHeapNumberValue(dividend)); var_divisor_float64.Bind(assembler->LoadHeapNumberValue(divisor)); assembler->Goto(&do_fdiv); } } } assembler->Bind(&do_fdiv); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kNumber)); Node* value = assembler->Float64Div(var_dividend_float64.value(), var_divisor_float64.value()); var_result.Bind(assembler->ChangeFloat64ToTagged(value)); assembler->Goto(&end); } assembler->Bind(&call_divide_stub); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kAny)); Callable callable = CodeFactory::Divide(assembler->isolate()); var_result.Bind(assembler->CallStub(callable, context, dividend, divisor)); assembler->Goto(&end); } assembler->Bind(&end); assembler->UpdateFeedback(var_type_feedback.value(), type_feedback_vector, slot_id); return var_result.value(); } // static compiler::Node* ModulusStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { using compiler::Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; Variable var_result(assembler, MachineRepresentation::kTagged); Label return_result(assembler, &var_result); // Shared entry point for floating point modulus. Label do_fmod(assembler); Variable var_dividend_float64(assembler, MachineRepresentation::kFloat64), var_divisor_float64(assembler, MachineRepresentation::kFloat64); Node* number_map = assembler->HeapNumberMapConstant(); // We might need to loop one or two times due to ToNumber conversions. Variable var_dividend(assembler, MachineRepresentation::kTagged), var_divisor(assembler, MachineRepresentation::kTagged); Variable* loop_variables[] = {&var_dividend, &var_divisor}; Label loop(assembler, 2, loop_variables); var_dividend.Bind(left); var_divisor.Bind(right); assembler->Goto(&loop); assembler->Bind(&loop); { Node* dividend = var_dividend.value(); Node* divisor = var_divisor.value(); Label dividend_is_smi(assembler), dividend_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(dividend), ÷nd_is_smi, ÷nd_is_not_smi); assembler->Bind(÷nd_is_smi); { Label dividend_is_not_zero(assembler); Label divisor_is_smi(assembler), divisor_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(divisor), &divisor_is_smi, &divisor_is_not_smi); assembler->Bind(&divisor_is_smi); { // Compute the modulus of two Smis. var_result.Bind(assembler->SmiMod(dividend, divisor)); assembler->Goto(&return_result); } assembler->Bind(&divisor_is_not_smi); { Node* divisor_map = assembler->LoadMap(divisor); // Check if {divisor} is a HeapNumber. Label divisor_is_number(assembler), divisor_is_not_number(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(divisor_map, number_map), &divisor_is_number, &divisor_is_not_number); assembler->Bind(&divisor_is_number); { // Convert {dividend} to a double and compute its modulus with the // value of {dividend}. var_dividend_float64.Bind(assembler->SmiToFloat64(dividend)); var_divisor_float64.Bind(assembler->LoadHeapNumberValue(divisor)); assembler->Goto(&do_fmod); } assembler->Bind(&divisor_is_not_number); { // Convert {divisor} to a number and loop. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_divisor.Bind(assembler->CallStub(callable, context, divisor)); assembler->Goto(&loop); } } } assembler->Bind(÷nd_is_not_smi); { Node* dividend_map = assembler->LoadMap(dividend); // Check if {dividend} is a HeapNumber. Label dividend_is_number(assembler), dividend_is_not_number(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(dividend_map, number_map), ÷nd_is_number, ÷nd_is_not_number); assembler->Bind(÷nd_is_number); { // Check if {divisor} is a Smi. Label divisor_is_smi(assembler), divisor_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(divisor), &divisor_is_smi, &divisor_is_not_smi); assembler->Bind(&divisor_is_smi); { // Convert {divisor} to a double and compute {dividend}'s modulus with // it. var_dividend_float64.Bind(assembler->LoadHeapNumberValue(dividend)); var_divisor_float64.Bind(assembler->SmiToFloat64(divisor)); assembler->Goto(&do_fmod); } assembler->Bind(&divisor_is_not_smi); { Node* divisor_map = assembler->LoadMap(divisor); // Check if {divisor} is a HeapNumber. Label divisor_is_number(assembler), divisor_is_not_number(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(divisor_map, number_map), &divisor_is_number, &divisor_is_not_number); assembler->Bind(&divisor_is_number); { // Both {dividend} and {divisor} are HeapNumbers. Load their values // and compute their modulus. var_dividend_float64.Bind(assembler->LoadHeapNumberValue(dividend)); var_divisor_float64.Bind(assembler->LoadHeapNumberValue(divisor)); assembler->Goto(&do_fmod); } assembler->Bind(&divisor_is_not_number); { // Convert {divisor} to a number and loop. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_divisor.Bind(assembler->CallStub(callable, context, divisor)); assembler->Goto(&loop); } } } assembler->Bind(÷nd_is_not_number); { // Convert {dividend} to a Number and loop. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_dividend.Bind(assembler->CallStub(callable, context, dividend)); assembler->Goto(&loop); } } } assembler->Bind(&do_fmod); { Node* value = assembler->Float64Mod(var_dividend_float64.value(), var_divisor_float64.value()); var_result.Bind(assembler->ChangeFloat64ToTagged(value)); assembler->Goto(&return_result); } assembler->Bind(&return_result); return var_result.value(); } // static compiler::Node* ModulusWithFeedbackStub::Generate( CodeStubAssembler* assembler, compiler::Node* dividend, compiler::Node* divisor, compiler::Node* slot_id, compiler::Node* type_feedback_vector, compiler::Node* context) { using compiler::Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; // Shared entry point for floating point division. Label do_fmod(assembler), end(assembler), call_modulus_stub(assembler); Variable var_dividend_float64(assembler, MachineRepresentation::kFloat64), var_divisor_float64(assembler, MachineRepresentation::kFloat64), var_result(assembler, MachineRepresentation::kTagged), var_type_feedback(assembler, MachineRepresentation::kWord32); Node* number_map = assembler->HeapNumberMapConstant(); Label dividend_is_smi(assembler), dividend_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(dividend), ÷nd_is_smi, ÷nd_is_not_smi); assembler->Bind(÷nd_is_smi); { Label divisor_is_smi(assembler), divisor_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(divisor), &divisor_is_smi, &divisor_is_not_smi); assembler->Bind(&divisor_is_smi); { var_result.Bind(assembler->SmiMod(dividend, divisor)); var_type_feedback.Bind(assembler->Select( assembler->WordIsSmi(var_result.value()), assembler->Int32Constant(BinaryOperationFeedback::kSignedSmall), assembler->Int32Constant(BinaryOperationFeedback::kNumber))); assembler->Goto(&end); } assembler->Bind(&divisor_is_not_smi); { Node* divisor_map = assembler->LoadMap(divisor); // Check if {divisor} is a HeapNumber. assembler->GotoUnless(assembler->WordEqual(divisor_map, number_map), &call_modulus_stub); // Convert {dividend} to a double and divide it with the value of // {divisor}. var_dividend_float64.Bind(assembler->SmiToFloat64(dividend)); var_divisor_float64.Bind(assembler->LoadHeapNumberValue(divisor)); assembler->Goto(&do_fmod); } } assembler->Bind(÷nd_is_not_smi); { Node* dividend_map = assembler->LoadMap(dividend); // Check if {dividend} is a HeapNumber. assembler->GotoUnless(assembler->WordEqual(dividend_map, number_map), &call_modulus_stub); // Check if {divisor} is a Smi. Label divisor_is_smi(assembler), divisor_is_not_smi(assembler); assembler->Branch(assembler->WordIsSmi(divisor), &divisor_is_smi, &divisor_is_not_smi); assembler->Bind(&divisor_is_smi); { // Convert {divisor} to a double and use it for a floating point // division. var_dividend_float64.Bind(assembler->LoadHeapNumberValue(dividend)); var_divisor_float64.Bind(assembler->SmiToFloat64(divisor)); assembler->Goto(&do_fmod); } assembler->Bind(&divisor_is_not_smi); { Node* divisor_map = assembler->LoadMap(divisor); // Check if {divisor} is a HeapNumber. assembler->GotoUnless(assembler->WordEqual(divisor_map, number_map), &call_modulus_stub); // Both {dividend} and {divisor} are HeapNumbers. Load their values // and divide them. var_dividend_float64.Bind(assembler->LoadHeapNumberValue(dividend)); var_divisor_float64.Bind(assembler->LoadHeapNumberValue(divisor)); assembler->Goto(&do_fmod); } } assembler->Bind(&do_fmod); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kNumber)); Node* value = assembler->Float64Mod(var_dividend_float64.value(), var_divisor_float64.value()); var_result.Bind(assembler->ChangeFloat64ToTagged(value)); assembler->Goto(&end); } assembler->Bind(&call_modulus_stub); { var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kAny)); Callable callable = CodeFactory::Modulus(assembler->isolate()); var_result.Bind(assembler->CallStub(callable, context, dividend, divisor)); assembler->Goto(&end); } assembler->Bind(&end); assembler->UpdateFeedback(var_type_feedback.value(), type_feedback_vector, slot_id); return var_result.value(); } // static compiler::Node* ShiftLeftStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { using compiler::Node; Node* lhs_value = assembler->TruncateTaggedToWord32(context, left); Node* rhs_value = assembler->TruncateTaggedToWord32(context, right); Node* shift_count = assembler->Word32And(rhs_value, assembler->Int32Constant(0x1f)); Node* value = assembler->Word32Shl(lhs_value, shift_count); Node* result = assembler->ChangeInt32ToTagged(value); return result; } // static compiler::Node* ShiftRightStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { using compiler::Node; Node* lhs_value = assembler->TruncateTaggedToWord32(context, left); Node* rhs_value = assembler->TruncateTaggedToWord32(context, right); Node* shift_count = assembler->Word32And(rhs_value, assembler->Int32Constant(0x1f)); Node* value = assembler->Word32Sar(lhs_value, shift_count); Node* result = assembler->ChangeInt32ToTagged(value); return result; } // static compiler::Node* ShiftRightLogicalStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { using compiler::Node; Node* lhs_value = assembler->TruncateTaggedToWord32(context, left); Node* rhs_value = assembler->TruncateTaggedToWord32(context, right); Node* shift_count = assembler->Word32And(rhs_value, assembler->Int32Constant(0x1f)); Node* value = assembler->Word32Shr(lhs_value, shift_count); Node* result = assembler->ChangeUint32ToTagged(value); return result; } // static compiler::Node* BitwiseAndStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { using compiler::Node; Node* lhs_value = assembler->TruncateTaggedToWord32(context, left); Node* rhs_value = assembler->TruncateTaggedToWord32(context, right); Node* value = assembler->Word32And(lhs_value, rhs_value); Node* result = assembler->ChangeInt32ToTagged(value); return result; } // static compiler::Node* BitwiseOrStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { using compiler::Node; Node* lhs_value = assembler->TruncateTaggedToWord32(context, left); Node* rhs_value = assembler->TruncateTaggedToWord32(context, right); Node* value = assembler->Word32Or(lhs_value, rhs_value); Node* result = assembler->ChangeInt32ToTagged(value); return result; } // static compiler::Node* BitwiseXorStub::Generate(CodeStubAssembler* assembler, compiler::Node* left, compiler::Node* right, compiler::Node* context) { using compiler::Node; Node* lhs_value = assembler->TruncateTaggedToWord32(context, left); Node* rhs_value = assembler->TruncateTaggedToWord32(context, right); Node* value = assembler->Word32Xor(lhs_value, rhs_value); Node* result = assembler->ChangeInt32ToTagged(value); return result; } // static compiler::Node* IncStub::Generate(CodeStubAssembler* assembler, compiler::Node* value, compiler::Node* context, compiler::Node* type_feedback_vector, compiler::Node* slot_id) { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; // Shared entry for floating point increment. Label do_finc(assembler), end(assembler); Variable var_finc_value(assembler, MachineRepresentation::kFloat64); // We might need to try again due to ToNumber conversion. Variable value_var(assembler, MachineRepresentation::kTagged); Variable result_var(assembler, MachineRepresentation::kTagged); Variable var_type_feedback(assembler, MachineRepresentation::kWord32); Variable* loop_vars[] = {&value_var, &var_type_feedback}; Label start(assembler, 2, loop_vars); value_var.Bind(value); var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kNone)); assembler->Goto(&start); assembler->Bind(&start); { value = value_var.value(); Label if_issmi(assembler), if_isnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(value), &if_issmi, &if_isnotsmi); assembler->Bind(&if_issmi); { // Try fast Smi addition first. Node* one = assembler->SmiConstant(Smi::FromInt(1)); Node* pair = assembler->SmiAddWithOverflow(value, one); Node* overflow = assembler->Projection(1, pair); // Check if the Smi addition overflowed. Label if_overflow(assembler), if_notoverflow(assembler); assembler->Branch(overflow, &if_overflow, &if_notoverflow); assembler->Bind(&if_notoverflow); var_type_feedback.Bind(assembler->Word32Or( var_type_feedback.value(), assembler->Int32Constant(BinaryOperationFeedback::kSignedSmall))); result_var.Bind(assembler->Projection(0, pair)); assembler->Goto(&end); assembler->Bind(&if_overflow); { var_finc_value.Bind(assembler->SmiToFloat64(value)); assembler->Goto(&do_finc); } } assembler->Bind(&if_isnotsmi); { // Check if the value is a HeapNumber. Label if_valueisnumber(assembler), if_valuenotnumber(assembler, Label::kDeferred); Node* value_map = assembler->LoadMap(value); Node* number_map = assembler->HeapNumberMapConstant(); assembler->Branch(assembler->WordEqual(value_map, number_map), &if_valueisnumber, &if_valuenotnumber); assembler->Bind(&if_valueisnumber); { // Load the HeapNumber value. var_finc_value.Bind(assembler->LoadHeapNumberValue(value)); assembler->Goto(&do_finc); } assembler->Bind(&if_valuenotnumber); { // Convert to a Number first and try again. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kAny)); value_var.Bind(assembler->CallStub(callable, context, value)); assembler->Goto(&start); } } } assembler->Bind(&do_finc); { Node* finc_value = var_finc_value.value(); Node* one = assembler->Float64Constant(1.0); Node* finc_result = assembler->Float64Add(finc_value, one); var_type_feedback.Bind(assembler->Word32Or( var_type_feedback.value(), assembler->Int32Constant(BinaryOperationFeedback::kNumber))); result_var.Bind(assembler->ChangeFloat64ToTagged(finc_result)); assembler->Goto(&end); } assembler->Bind(&end); assembler->UpdateFeedback(var_type_feedback.value(), type_feedback_vector, slot_id); return result_var.value(); } // static compiler::Node* DecStub::Generate(CodeStubAssembler* assembler, compiler::Node* value, compiler::Node* context, compiler::Node* type_feedback_vector, compiler::Node* slot_id) { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; // Shared entry for floating point decrement. Label do_fdec(assembler), end(assembler); Variable var_fdec_value(assembler, MachineRepresentation::kFloat64); // We might need to try again due to ToNumber conversion. Variable value_var(assembler, MachineRepresentation::kTagged); Variable result_var(assembler, MachineRepresentation::kTagged); Variable var_type_feedback(assembler, MachineRepresentation::kWord32); Variable* loop_vars[] = {&value_var, &var_type_feedback}; Label start(assembler, 2, loop_vars); var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kNone)); value_var.Bind(value); assembler->Goto(&start); assembler->Bind(&start); { value = value_var.value(); Label if_issmi(assembler), if_isnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(value), &if_issmi, &if_isnotsmi); assembler->Bind(&if_issmi); { // Try fast Smi subtraction first. Node* one = assembler->SmiConstant(Smi::FromInt(1)); Node* pair = assembler->SmiSubWithOverflow(value, one); Node* overflow = assembler->Projection(1, pair); // Check if the Smi subtraction overflowed. Label if_overflow(assembler), if_notoverflow(assembler); assembler->Branch(overflow, &if_overflow, &if_notoverflow); assembler->Bind(&if_notoverflow); var_type_feedback.Bind(assembler->Word32Or( var_type_feedback.value(), assembler->Int32Constant(BinaryOperationFeedback::kSignedSmall))); result_var.Bind(assembler->Projection(0, pair)); assembler->Goto(&end); assembler->Bind(&if_overflow); { var_fdec_value.Bind(assembler->SmiToFloat64(value)); assembler->Goto(&do_fdec); } } assembler->Bind(&if_isnotsmi); { // Check if the value is a HeapNumber. Label if_valueisnumber(assembler), if_valuenotnumber(assembler, Label::kDeferred); Node* value_map = assembler->LoadMap(value); Node* number_map = assembler->HeapNumberMapConstant(); assembler->Branch(assembler->WordEqual(value_map, number_map), &if_valueisnumber, &if_valuenotnumber); assembler->Bind(&if_valueisnumber); { // Load the HeapNumber value. var_fdec_value.Bind(assembler->LoadHeapNumberValue(value)); assembler->Goto(&do_fdec); } assembler->Bind(&if_valuenotnumber); { // Convert to a Number first and try again. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_type_feedback.Bind( assembler->Int32Constant(BinaryOperationFeedback::kAny)); value_var.Bind(assembler->CallStub(callable, context, value)); assembler->Goto(&start); } } } assembler->Bind(&do_fdec); { Node* fdec_value = var_fdec_value.value(); Node* one = assembler->Float64Constant(1.0); Node* fdec_result = assembler->Float64Sub(fdec_value, one); var_type_feedback.Bind(assembler->Word32Or( var_type_feedback.value(), assembler->Int32Constant(BinaryOperationFeedback::kNumber))); result_var.Bind(assembler->ChangeFloat64ToTagged(fdec_result)); assembler->Goto(&end); } assembler->Bind(&end); assembler->UpdateFeedback(var_type_feedback.value(), type_feedback_vector, slot_id); return result_var.value(); } // ES6 section 21.1.3.19 String.prototype.substring ( start, end ) compiler::Node* SubStringStub::Generate(CodeStubAssembler* assembler, compiler::Node* string, compiler::Node* from, compiler::Node* to, compiler::Node* context) { return assembler->SubString(context, string, from, to); } // ES6 section 7.1.13 ToObject (argument) void ToObjectStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef compiler::Node Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; Label if_number(assembler, Label::kDeferred), if_notsmi(assembler), if_jsreceiver(assembler), if_noconstructor(assembler, Label::kDeferred), if_wrapjsvalue(assembler); Node* object = assembler->Parameter(Descriptor::kArgument); Node* context = assembler->Parameter(Descriptor::kContext); Variable constructor_function_index_var(assembler, MachineType::PointerRepresentation()); assembler->Branch(assembler->WordIsSmi(object), &if_number, &if_notsmi); assembler->Bind(&if_notsmi); Node* map = assembler->LoadMap(object); assembler->GotoIf( assembler->WordEqual(map, assembler->HeapNumberMapConstant()), &if_number); Node* instance_type = assembler->LoadMapInstanceType(map); assembler->GotoIf( assembler->Int32GreaterThanOrEqual( instance_type, assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE)), &if_jsreceiver); Node* constructor_function_index = assembler->LoadMapConstructorFunctionIndex(map); assembler->GotoIf(assembler->WordEqual(constructor_function_index, assembler->IntPtrConstant( Map::kNoConstructorFunctionIndex)), &if_noconstructor); constructor_function_index_var.Bind(constructor_function_index); assembler->Goto(&if_wrapjsvalue); assembler->Bind(&if_number); constructor_function_index_var.Bind( assembler->IntPtrConstant(Context::NUMBER_FUNCTION_INDEX)); assembler->Goto(&if_wrapjsvalue); assembler->Bind(&if_wrapjsvalue); Node* native_context = assembler->LoadNativeContext(context); Node* constructor = assembler->LoadFixedArrayElement( native_context, constructor_function_index_var.value(), 0, CodeStubAssembler::INTPTR_PARAMETERS); Node* initial_map = assembler->LoadObjectField( constructor, JSFunction::kPrototypeOrInitialMapOffset); Node* js_value = assembler->Allocate(JSValue::kSize); assembler->StoreMapNoWriteBarrier(js_value, initial_map); assembler->StoreObjectFieldRoot(js_value, JSValue::kPropertiesOffset, Heap::kEmptyFixedArrayRootIndex); assembler->StoreObjectFieldRoot(js_value, JSObject::kElementsOffset, Heap::kEmptyFixedArrayRootIndex); assembler->StoreObjectField(js_value, JSValue::kValueOffset, object); assembler->Return(js_value); assembler->Bind(&if_noconstructor); assembler->TailCallRuntime( Runtime::kThrowUndefinedOrNullToObject, context, assembler->HeapConstant(assembler->factory()->NewStringFromAsciiChecked( "ToObject", TENURED))); assembler->Bind(&if_jsreceiver); assembler->Return(object); } // static // ES6 section 12.5.5 typeof operator compiler::Node* TypeofStub::Generate(CodeStubAssembler* assembler, compiler::Node* value, compiler::Node* context) { typedef compiler::Node Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; Variable result_var(assembler, MachineRepresentation::kTagged); Label return_number(assembler, Label::kDeferred), if_oddball(assembler), return_function(assembler), return_undefined(assembler), return_object(assembler), return_string(assembler), return_result(assembler); assembler->GotoIf(assembler->WordIsSmi(value), &return_number); Node* map = assembler->LoadMap(value); assembler->GotoIf( assembler->WordEqual(map, assembler->HeapNumberMapConstant()), &return_number); Node* instance_type = assembler->LoadMapInstanceType(map); assembler->GotoIf(assembler->Word32Equal( instance_type, assembler->Int32Constant(ODDBALL_TYPE)), &if_oddball); Node* callable_or_undetectable_mask = assembler->Word32And(assembler->LoadMapBitField(map), assembler->Int32Constant(1 << Map::kIsCallable | 1 << Map::kIsUndetectable)); assembler->GotoIf( assembler->Word32Equal(callable_or_undetectable_mask, assembler->Int32Constant(1 << Map::kIsCallable)), &return_function); assembler->GotoUnless(assembler->Word32Equal(callable_or_undetectable_mask, assembler->Int32Constant(0)), &return_undefined); assembler->GotoIf( assembler->Int32GreaterThanOrEqual( instance_type, assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE)), &return_object); assembler->GotoIf( assembler->Int32LessThan(instance_type, assembler->Int32Constant(FIRST_NONSTRING_TYPE)), &return_string); #define SIMD128_BRANCH(TYPE, Type, type, lane_count, lane_type) \ Label return_##type(assembler); \ Node* type##_map = \ assembler->HeapConstant(assembler->factory()->type##_map()); \ assembler->GotoIf(assembler->WordEqual(map, type##_map), &return_##type); SIMD128_TYPES(SIMD128_BRANCH) #undef SIMD128_BRANCH assembler->Assert(assembler->Word32Equal( instance_type, assembler->Int32Constant(SYMBOL_TYPE))); result_var.Bind(assembler->HeapConstant( assembler->isolate()->factory()->symbol_string())); assembler->Goto(&return_result); assembler->Bind(&return_number); { result_var.Bind(assembler->HeapConstant( assembler->isolate()->factory()->number_string())); assembler->Goto(&return_result); } assembler->Bind(&if_oddball); { Node* type = assembler->LoadObjectField(value, Oddball::kTypeOfOffset); result_var.Bind(type); assembler->Goto(&return_result); } assembler->Bind(&return_function); { result_var.Bind(assembler->HeapConstant( assembler->isolate()->factory()->function_string())); assembler->Goto(&return_result); } assembler->Bind(&return_undefined); { result_var.Bind(assembler->HeapConstant( assembler->isolate()->factory()->undefined_string())); assembler->Goto(&return_result); } assembler->Bind(&return_object); { result_var.Bind(assembler->HeapConstant( assembler->isolate()->factory()->object_string())); assembler->Goto(&return_result); } assembler->Bind(&return_string); { result_var.Bind(assembler->HeapConstant( assembler->isolate()->factory()->string_string())); assembler->Goto(&return_result); } #define SIMD128_BIND_RETURN(TYPE, Type, type, lane_count, lane_type) \ assembler->Bind(&return_##type); \ { \ result_var.Bind(assembler->HeapConstant( \ assembler->isolate()->factory()->type##_string())); \ assembler->Goto(&return_result); \ } SIMD128_TYPES(SIMD128_BIND_RETURN) #undef SIMD128_BIND_RETURN assembler->Bind(&return_result); return result_var.value(); } // static compiler::Node* InstanceOfStub::Generate(CodeStubAssembler* assembler, compiler::Node* object, compiler::Node* callable, compiler::Node* context) { typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; Label return_runtime(assembler, Label::kDeferred), end(assembler); Variable result(assembler, MachineRepresentation::kTagged); // Check if no one installed @@hasInstance somewhere. assembler->GotoUnless( assembler->WordEqual( assembler->LoadObjectField( assembler->LoadRoot(Heap::kHasInstanceProtectorRootIndex), PropertyCell::kValueOffset), assembler->SmiConstant(Smi::FromInt(Isolate::kArrayProtectorValid))), &return_runtime); // Check if {callable} is a valid receiver. assembler->GotoIf(assembler->WordIsSmi(callable), &return_runtime); assembler->GotoIf( assembler->Word32Equal( assembler->Word32And( assembler->LoadMapBitField(assembler->LoadMap(callable)), assembler->Int32Constant(1 << Map::kIsCallable)), assembler->Int32Constant(0)), &return_runtime); // Use the inline OrdinaryHasInstance directly. result.Bind(assembler->OrdinaryHasInstance(context, callable, object)); assembler->Goto(&end); // TODO(bmeurer): Use GetPropertyStub here once available. assembler->Bind(&return_runtime); { result.Bind(assembler->CallRuntime(Runtime::kInstanceOf, context, object, callable)); assembler->Goto(&end); } assembler->Bind(&end); return result.value(); } namespace { enum RelationalComparisonMode { kLessThan, kLessThanOrEqual, kGreaterThan, kGreaterThanOrEqual }; compiler::Node* GenerateAbstractRelationalComparison( CodeStubAssembler* assembler, RelationalComparisonMode mode, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; Label return_true(assembler), return_false(assembler), end(assembler); Variable result(assembler, MachineRepresentation::kTagged); // Shared entry for floating point comparison. Label do_fcmp(assembler); Variable var_fcmp_lhs(assembler, MachineRepresentation::kFloat64), var_fcmp_rhs(assembler, MachineRepresentation::kFloat64); // We might need to loop several times due to ToPrimitive and/or ToNumber // conversions. Variable var_lhs(assembler, MachineRepresentation::kTagged), var_rhs(assembler, MachineRepresentation::kTagged); Variable* loop_vars[2] = {&var_lhs, &var_rhs}; Label loop(assembler, 2, loop_vars); var_lhs.Bind(lhs); var_rhs.Bind(rhs); assembler->Goto(&loop); assembler->Bind(&loop); { // Load the current {lhs} and {rhs} values. lhs = var_lhs.value(); rhs = var_rhs.value(); // Check if the {lhs} is a Smi or a HeapObject. Label if_lhsissmi(assembler), if_lhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(lhs), &if_lhsissmi, &if_lhsisnotsmi); assembler->Bind(&if_lhsissmi); { // Check if {rhs} is a Smi or a HeapObject. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Both {lhs} and {rhs} are Smi, so just perform a fast Smi comparison. switch (mode) { case kLessThan: assembler->BranchIfSmiLessThan(lhs, rhs, &return_true, &return_false); break; case kLessThanOrEqual: assembler->BranchIfSmiLessThanOrEqual(lhs, rhs, &return_true, &return_false); break; case kGreaterThan: assembler->BranchIfSmiLessThan(rhs, lhs, &return_true, &return_false); break; case kGreaterThanOrEqual: assembler->BranchIfSmiLessThanOrEqual(rhs, lhs, &return_true, &return_false); break; } } assembler->Bind(&if_rhsisnotsmi); { // Load the map of {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if the {rhs} is a HeapNumber. Node* number_map = assembler->HeapNumberMapConstant(); Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(rhs_map, number_map), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { // Convert the {lhs} and {rhs} to floating point values, and // perform a floating point comparison. var_fcmp_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fcmp_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fcmp); } assembler->Bind(&if_rhsisnotnumber); { // Convert the {rhs} to a Number; we don't need to perform the // dedicated ToPrimitive(rhs, hint Number) operation, as the // ToNumber(rhs) will by itself already invoke ToPrimitive with // a Number hint. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } } } assembler->Bind(&if_lhsisnotsmi); { // Load the HeapNumber map for later comparisons. Node* number_map = assembler->HeapNumberMapConstant(); // Load the map of {lhs}. Node* lhs_map = assembler->LoadMap(lhs); // Check if {rhs} is a Smi or a HeapObject. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Check if the {lhs} is a HeapNumber. Label if_lhsisnumber(assembler), if_lhsisnotnumber(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(lhs_map, number_map), &if_lhsisnumber, &if_lhsisnotnumber); assembler->Bind(&if_lhsisnumber); { // Convert the {lhs} and {rhs} to floating point values, and // perform a floating point comparison. var_fcmp_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fcmp_rhs.Bind(assembler->SmiToFloat64(rhs)); assembler->Goto(&do_fcmp); } assembler->Bind(&if_lhsisnotnumber); { // Convert the {lhs} to a Number; we don't need to perform the // dedicated ToPrimitive(lhs, hint Number) operation, as the // ToNumber(lhs) will by itself already invoke ToPrimitive with // a Number hint. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); assembler->Goto(&loop); } } assembler->Bind(&if_rhsisnotsmi); { // Load the map of {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if {lhs} is a HeapNumber. Label if_lhsisnumber(assembler), if_lhsisnotnumber(assembler); assembler->Branch(assembler->WordEqual(lhs_map, number_map), &if_lhsisnumber, &if_lhsisnotnumber); assembler->Bind(&if_lhsisnumber); { // Check if {rhs} is also a HeapNumber. Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(lhs_map, rhs_map), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { // Convert the {lhs} and {rhs} to floating point values, and // perform a floating point comparison. var_fcmp_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fcmp_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fcmp); } assembler->Bind(&if_rhsisnotnumber); { // Convert the {rhs} to a Number; we don't need to perform // dedicated ToPrimitive(rhs, hint Number) operation, as the // ToNumber(rhs) will by itself already invoke ToPrimitive with // a Number hint. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } } assembler->Bind(&if_lhsisnotnumber); { // Load the instance type of {lhs}. Node* lhs_instance_type = assembler->LoadMapInstanceType(lhs_map); // Check if {lhs} is a String. Label if_lhsisstring(assembler), if_lhsisnotstring(assembler, Label::kDeferred); assembler->Branch(assembler->Int32LessThan( lhs_instance_type, assembler->Int32Constant(FIRST_NONSTRING_TYPE)), &if_lhsisstring, &if_lhsisnotstring); assembler->Bind(&if_lhsisstring); { // Load the instance type of {rhs}. Node* rhs_instance_type = assembler->LoadMapInstanceType(rhs_map); // Check if {rhs} is also a String. Label if_rhsisstring(assembler, Label::kDeferred), if_rhsisnotstring(assembler, Label::kDeferred); assembler->Branch(assembler->Int32LessThan( rhs_instance_type, assembler->Int32Constant( FIRST_NONSTRING_TYPE)), &if_rhsisstring, &if_rhsisnotstring); assembler->Bind(&if_rhsisstring); { // Both {lhs} and {rhs} are strings. switch (mode) { case kLessThan: result.Bind(assembler->CallStub( CodeFactory::StringLessThan(assembler->isolate()), context, lhs, rhs)); assembler->Goto(&end); break; case kLessThanOrEqual: result.Bind(assembler->CallStub( CodeFactory::StringLessThanOrEqual(assembler->isolate()), context, lhs, rhs)); assembler->Goto(&end); break; case kGreaterThan: result.Bind(assembler->CallStub( CodeFactory::StringGreaterThan(assembler->isolate()), context, lhs, rhs)); assembler->Goto(&end); break; case kGreaterThanOrEqual: result.Bind( assembler->CallStub(CodeFactory::StringGreaterThanOrEqual( assembler->isolate()), context, lhs, rhs)); assembler->Goto(&end); break; } } assembler->Bind(&if_rhsisnotstring); { // The {lhs} is a String, while {rhs} is neither a Number nor a // String, so we need to call ToPrimitive(rhs, hint Number) if // {rhs} is a receiver or ToNumber(lhs) and ToNumber(rhs) in the // other cases. STATIC_ASSERT(LAST_JS_RECEIVER_TYPE == LAST_TYPE); Label if_rhsisreceiver(assembler, Label::kDeferred), if_rhsisnotreceiver(assembler, Label::kDeferred); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), rhs_instance_type), &if_rhsisreceiver, &if_rhsisnotreceiver); assembler->Bind(&if_rhsisreceiver); { // Convert {rhs} to a primitive first passing Number hint. Callable callable = CodeFactory::NonPrimitiveToPrimitive( assembler->isolate(), ToPrimitiveHint::kNumber); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotreceiver); { // Convert both {lhs} and {rhs} to Number. Callable callable = CodeFactory::ToNumber(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } } } assembler->Bind(&if_lhsisnotstring); { // The {lhs} is neither a Number nor a String, so we need to call // ToPrimitive(lhs, hint Number) if {lhs} is a receiver or // ToNumber(lhs) and ToNumber(rhs) in the other cases. STATIC_ASSERT(LAST_JS_RECEIVER_TYPE == LAST_TYPE); Label if_lhsisreceiver(assembler, Label::kDeferred), if_lhsisnotreceiver(assembler, Label::kDeferred); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), lhs_instance_type), &if_lhsisreceiver, &if_lhsisnotreceiver); assembler->Bind(&if_lhsisreceiver); { // Convert {lhs} to a primitive first passing Number hint. Callable callable = CodeFactory::NonPrimitiveToPrimitive( assembler->isolate(), ToPrimitiveHint::kNumber); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); assembler->Goto(&loop); } assembler->Bind(&if_lhsisnotreceiver); { // Convert both {lhs} and {rhs} to Number. Callable callable = CodeFactory::ToNumber(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } } } } } } assembler->Bind(&do_fcmp); { // Load the {lhs} and {rhs} floating point values. Node* lhs = var_fcmp_lhs.value(); Node* rhs = var_fcmp_rhs.value(); // Perform a fast floating point comparison. switch (mode) { case kLessThan: assembler->BranchIfFloat64LessThan(lhs, rhs, &return_true, &return_false); break; case kLessThanOrEqual: assembler->BranchIfFloat64LessThanOrEqual(lhs, rhs, &return_true, &return_false); break; case kGreaterThan: assembler->BranchIfFloat64GreaterThan(lhs, rhs, &return_true, &return_false); break; case kGreaterThanOrEqual: assembler->BranchIfFloat64GreaterThanOrEqual(lhs, rhs, &return_true, &return_false); break; } } assembler->Bind(&return_true); { result.Bind(assembler->BooleanConstant(true)); assembler->Goto(&end); } assembler->Bind(&return_false); { result.Bind(assembler->BooleanConstant(false)); assembler->Goto(&end); } assembler->Bind(&end); return result.value(); } enum ResultMode { kDontNegateResult, kNegateResult }; void GenerateEqual_Same(CodeStubAssembler* assembler, compiler::Node* value, CodeStubAssembler::Label* if_equal, CodeStubAssembler::Label* if_notequal) { // In case of abstract or strict equality checks, we need additional checks // for NaN values because they are not considered equal, even if both the // left and the right hand side reference exactly the same value. // TODO(bmeurer): This seems to violate the SIMD.js specification, but it // seems to be what is tested in the current SIMD.js testsuite. typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; // Check if {value} is a Smi or a HeapObject. Label if_valueissmi(assembler), if_valueisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(value), &if_valueissmi, &if_valueisnotsmi); assembler->Bind(&if_valueisnotsmi); { // Load the map of {value}. Node* value_map = assembler->LoadMap(value); // Check if {value} (and therefore {rhs}) is a HeapNumber. Node* number_map = assembler->HeapNumberMapConstant(); Label if_valueisnumber(assembler), if_valueisnotnumber(assembler); assembler->Branch(assembler->WordEqual(value_map, number_map), &if_valueisnumber, &if_valueisnotnumber); assembler->Bind(&if_valueisnumber); { // Convert {value} (and therefore {rhs}) to floating point value. Node* value_value = assembler->LoadHeapNumberValue(value); // Check if the HeapNumber value is a NaN. assembler->BranchIfFloat64IsNaN(value_value, if_notequal, if_equal); } assembler->Bind(&if_valueisnotnumber); assembler->Goto(if_equal); } assembler->Bind(&if_valueissmi); assembler->Goto(if_equal); } void GenerateEqual_Simd128Value_HeapObject( CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* lhs_map, compiler::Node* rhs, compiler::Node* rhs_map, CodeStubAssembler::Label* if_equal, CodeStubAssembler::Label* if_notequal) { assembler->BranchIfSimd128Equal(lhs, lhs_map, rhs, rhs_map, if_equal, if_notequal); } // ES6 section 7.2.12 Abstract Equality Comparison compiler::Node* GenerateEqual(CodeStubAssembler* assembler, ResultMode mode, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { // This is a slightly optimized version of Object::Equals represented as // scheduled TurboFan graph utilizing the CodeStubAssembler. Whenever you // change something functionality wise in here, remember to update the // Object::Equals method as well. typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; Label if_equal(assembler), if_notequal(assembler), do_rhsstringtonumber(assembler, Label::kDeferred), end(assembler); Variable result(assembler, MachineRepresentation::kTagged); // Shared entry for floating point comparison. Label do_fcmp(assembler); Variable var_fcmp_lhs(assembler, MachineRepresentation::kFloat64), var_fcmp_rhs(assembler, MachineRepresentation::kFloat64); // We might need to loop several times due to ToPrimitive and/or ToNumber // conversions. Variable var_lhs(assembler, MachineRepresentation::kTagged), var_rhs(assembler, MachineRepresentation::kTagged); Variable* loop_vars[2] = {&var_lhs, &var_rhs}; Label loop(assembler, 2, loop_vars); var_lhs.Bind(lhs); var_rhs.Bind(rhs); assembler->Goto(&loop); assembler->Bind(&loop); { // Load the current {lhs} and {rhs} values. lhs = var_lhs.value(); rhs = var_rhs.value(); // Check if {lhs} and {rhs} refer to the same object. Label if_same(assembler), if_notsame(assembler); assembler->Branch(assembler->WordEqual(lhs, rhs), &if_same, &if_notsame); assembler->Bind(&if_same); { // The {lhs} and {rhs} reference the exact same value, yet we need special // treatment for HeapNumber, as NaN is not equal to NaN. GenerateEqual_Same(assembler, lhs, &if_equal, &if_notequal); } assembler->Bind(&if_notsame); { // Check if {lhs} is a Smi or a HeapObject. Label if_lhsissmi(assembler), if_lhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(lhs), &if_lhsissmi, &if_lhsisnotsmi); assembler->Bind(&if_lhsissmi); { // Check if {rhs} is a Smi or a HeapObject. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); // We have already checked for {lhs} and {rhs} being the same value, so // if both are Smis when we get here they must not be equal. assembler->Goto(&if_notequal); assembler->Bind(&if_rhsisnotsmi); { // Load the map of {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if {rhs} is a HeapNumber. Node* number_map = assembler->HeapNumberMapConstant(); Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler); assembler->Branch(assembler->WordEqual(rhs_map, number_map), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { // Convert {lhs} and {rhs} to floating point values, and // perform a floating point comparison. var_fcmp_lhs.Bind(assembler->SmiToFloat64(lhs)); var_fcmp_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fcmp); } assembler->Bind(&if_rhsisnotnumber); { // Load the instance type of the {rhs}. Node* rhs_instance_type = assembler->LoadMapInstanceType(rhs_map); // Check if the {rhs} is a String. Label if_rhsisstring(assembler, Label::kDeferred), if_rhsisnotstring(assembler); assembler->Branch(assembler->Int32LessThan( rhs_instance_type, assembler->Int32Constant( FIRST_NONSTRING_TYPE)), &if_rhsisstring, &if_rhsisnotstring); assembler->Bind(&if_rhsisstring); { // The {rhs} is a String and the {lhs} is a Smi; we need // to convert the {rhs} to a Number and compare the output to // the Number on the {lhs}. assembler->Goto(&do_rhsstringtonumber); } assembler->Bind(&if_rhsisnotstring); { // Check if the {rhs} is a Boolean. Node* boolean_map = assembler->BooleanMapConstant(); Label if_rhsisboolean(assembler), if_rhsisnotboolean(assembler); assembler->Branch(assembler->WordEqual(rhs_map, boolean_map), &if_rhsisboolean, &if_rhsisnotboolean); assembler->Bind(&if_rhsisboolean); { // The {rhs} is a Boolean, load its number value. var_rhs.Bind( assembler->LoadObjectField(rhs, Oddball::kToNumberOffset)); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotboolean); { // Check if the {rhs} is a Receiver. STATIC_ASSERT(LAST_JS_RECEIVER_TYPE == LAST_TYPE); Label if_rhsisreceiver(assembler, Label::kDeferred), if_rhsisnotreceiver(assembler); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), rhs_instance_type), &if_rhsisreceiver, &if_rhsisnotreceiver); assembler->Bind(&if_rhsisreceiver); { // Convert {rhs} to a primitive first (passing no hint). Callable callable = CodeFactory::NonPrimitiveToPrimitive( assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotreceiver); assembler->Goto(&if_notequal); } } } } } assembler->Bind(&if_lhsisnotsmi); { // Check if {rhs} is a Smi or a HeapObject. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // The {lhs} is a HeapObject and the {rhs} is a Smi; swapping {lhs} // and {rhs} is not observable and doesn't matter for the result, so // we can just swap them and use the Smi handling above (for {lhs} // being a Smi). var_lhs.Bind(rhs); var_rhs.Bind(lhs); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotsmi); { Label if_lhsisstring(assembler), if_lhsisnumber(assembler), if_lhsissymbol(assembler), if_lhsissimd128value(assembler), if_lhsisoddball(assembler), if_lhsisreceiver(assembler); // Both {lhs} and {rhs} are HeapObjects, load their maps // and their instance types. Node* lhs_map = assembler->LoadMap(lhs); Node* rhs_map = assembler->LoadMap(rhs); // Load the instance types of {lhs} and {rhs}. Node* lhs_instance_type = assembler->LoadMapInstanceType(lhs_map); Node* rhs_instance_type = assembler->LoadMapInstanceType(rhs_map); // Dispatch based on the instance type of {lhs}. size_t const kNumCases = FIRST_NONSTRING_TYPE + 4; Label* case_labels[kNumCases]; int32_t case_values[kNumCases]; for (int32_t i = 0; i < FIRST_NONSTRING_TYPE; ++i) { case_labels[i] = new Label(assembler); case_values[i] = i; } case_labels[FIRST_NONSTRING_TYPE + 0] = &if_lhsisnumber; case_values[FIRST_NONSTRING_TYPE + 0] = HEAP_NUMBER_TYPE; case_labels[FIRST_NONSTRING_TYPE + 1] = &if_lhsissymbol; case_values[FIRST_NONSTRING_TYPE + 1] = SYMBOL_TYPE; case_labels[FIRST_NONSTRING_TYPE + 2] = &if_lhsissimd128value; case_values[FIRST_NONSTRING_TYPE + 2] = SIMD128_VALUE_TYPE; case_labels[FIRST_NONSTRING_TYPE + 3] = &if_lhsisoddball; case_values[FIRST_NONSTRING_TYPE + 3] = ODDBALL_TYPE; assembler->Switch(lhs_instance_type, &if_lhsisreceiver, case_values, case_labels, arraysize(case_values)); for (int32_t i = 0; i < FIRST_NONSTRING_TYPE; ++i) { assembler->Bind(case_labels[i]); assembler->Goto(&if_lhsisstring); delete case_labels[i]; } assembler->Bind(&if_lhsisstring); { // Check if {rhs} is also a String. Label if_rhsisstring(assembler, Label::kDeferred), if_rhsisnotstring(assembler); assembler->Branch(assembler->Int32LessThan( rhs_instance_type, assembler->Int32Constant( FIRST_NONSTRING_TYPE)), &if_rhsisstring, &if_rhsisnotstring); assembler->Bind(&if_rhsisstring); { // Both {lhs} and {rhs} are of type String, just do the // string comparison then. Callable callable = (mode == kDontNegateResult) ? CodeFactory::StringEqual(assembler->isolate()) : CodeFactory::StringNotEqual(assembler->isolate()); result.Bind(assembler->CallStub(callable, context, lhs, rhs)); assembler->Goto(&end); } assembler->Bind(&if_rhsisnotstring); { // The {lhs} is a String and the {rhs} is some other HeapObject. // Swapping {lhs} and {rhs} is not observable and doesn't matter // for the result, so we can just swap them and use the String // handling below (for {rhs} being a String). var_lhs.Bind(rhs); var_rhs.Bind(lhs); assembler->Goto(&loop); } } assembler->Bind(&if_lhsisnumber); { // Check if {rhs} is also a HeapNumber. Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler); assembler->Branch( assembler->Word32Equal(lhs_instance_type, rhs_instance_type), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { // Convert {lhs} and {rhs} to floating point values, and // perform a floating point comparison. var_fcmp_lhs.Bind(assembler->LoadHeapNumberValue(lhs)); var_fcmp_rhs.Bind(assembler->LoadHeapNumberValue(rhs)); assembler->Goto(&do_fcmp); } assembler->Bind(&if_rhsisnotnumber); { // The {lhs} is a Number, the {rhs} is some other HeapObject. Label if_rhsisstring(assembler, Label::kDeferred), if_rhsisnotstring(assembler); assembler->Branch( assembler->Int32LessThan( rhs_instance_type, assembler->Int32Constant(FIRST_NONSTRING_TYPE)), &if_rhsisstring, &if_rhsisnotstring); assembler->Bind(&if_rhsisstring); { // The {rhs} is a String and the {lhs} is a HeapNumber; we need // to convert the {rhs} to a Number and compare the output to // the Number on the {lhs}. assembler->Goto(&do_rhsstringtonumber); } assembler->Bind(&if_rhsisnotstring); { // Check if the {rhs} is a JSReceiver. Label if_rhsisreceiver(assembler), if_rhsisnotreceiver(assembler); STATIC_ASSERT(LAST_TYPE == LAST_JS_RECEIVER_TYPE); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), rhs_instance_type), &if_rhsisreceiver, &if_rhsisnotreceiver); assembler->Bind(&if_rhsisreceiver); { // The {lhs} is a Primitive and the {rhs} is a JSReceiver. // Swapping {lhs} and {rhs} is not observable and doesn't // matter for the result, so we can just swap them and use // the JSReceiver handling below (for {lhs} being a // JSReceiver). var_lhs.Bind(rhs); var_rhs.Bind(lhs); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotreceiver); { // Check if {rhs} is a Boolean. Label if_rhsisboolean(assembler), if_rhsisnotboolean(assembler); Node* boolean_map = assembler->BooleanMapConstant(); assembler->Branch(assembler->WordEqual(rhs_map, boolean_map), &if_rhsisboolean, &if_rhsisnotboolean); assembler->Bind(&if_rhsisboolean); { // The {rhs} is a Boolean, convert it to a Smi first. var_rhs.Bind(assembler->LoadObjectField( rhs, Oddball::kToNumberOffset)); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotboolean); assembler->Goto(&if_notequal); } } } } assembler->Bind(&if_lhsisoddball); { // The {lhs} is an Oddball and {rhs} is some other HeapObject. Label if_lhsisboolean(assembler), if_lhsisnotboolean(assembler); Node* boolean_map = assembler->BooleanMapConstant(); assembler->Branch(assembler->WordEqual(lhs_map, boolean_map), &if_lhsisboolean, &if_lhsisnotboolean); assembler->Bind(&if_lhsisboolean); { // The {lhs} is a Boolean, check if {rhs} is also a Boolean. Label if_rhsisboolean(assembler), if_rhsisnotboolean(assembler); assembler->Branch(assembler->WordEqual(rhs_map, boolean_map), &if_rhsisboolean, &if_rhsisnotboolean); assembler->Bind(&if_rhsisboolean); { // Both {lhs} and {rhs} are distinct Boolean values. assembler->Goto(&if_notequal); } assembler->Bind(&if_rhsisnotboolean); { // Convert the {lhs} to a Number first. var_lhs.Bind( assembler->LoadObjectField(lhs, Oddball::kToNumberOffset)); assembler->Goto(&loop); } } assembler->Bind(&if_lhsisnotboolean); { // The {lhs} is either Null or Undefined; check if the {rhs} is // undetectable (i.e. either also Null or Undefined or some // undetectable JSReceiver). Node* rhs_bitfield = assembler->LoadMapBitField(rhs_map); assembler->BranchIfWord32Equal( assembler->Word32And( rhs_bitfield, assembler->Int32Constant(1 << Map::kIsUndetectable)), assembler->Int32Constant(0), &if_notequal, &if_equal); } } assembler->Bind(&if_lhsissymbol); { // Check if the {rhs} is a JSReceiver. Label if_rhsisreceiver(assembler), if_rhsisnotreceiver(assembler); STATIC_ASSERT(LAST_TYPE == LAST_JS_RECEIVER_TYPE); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), rhs_instance_type), &if_rhsisreceiver, &if_rhsisnotreceiver); assembler->Bind(&if_rhsisreceiver); { // The {lhs} is a Primitive and the {rhs} is a JSReceiver. // Swapping {lhs} and {rhs} is not observable and doesn't // matter for the result, so we can just swap them and use // the JSReceiver handling below (for {lhs} being a JSReceiver). var_lhs.Bind(rhs); var_rhs.Bind(lhs); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotreceiver); { // The {rhs} is not a JSReceiver and also not the same Symbol // as the {lhs}, so this is equality check is considered false. assembler->Goto(&if_notequal); } } assembler->Bind(&if_lhsissimd128value); { // Check if the {rhs} is also a Simd128Value. Label if_rhsissimd128value(assembler), if_rhsisnotsimd128value(assembler); assembler->Branch( assembler->Word32Equal(lhs_instance_type, rhs_instance_type), &if_rhsissimd128value, &if_rhsisnotsimd128value); assembler->Bind(&if_rhsissimd128value); { // Both {lhs} and {rhs} is a Simd128Value. GenerateEqual_Simd128Value_HeapObject(assembler, lhs, lhs_map, rhs, rhs_map, &if_equal, &if_notequal); } assembler->Bind(&if_rhsisnotsimd128value); { // Check if the {rhs} is a JSReceiver. Label if_rhsisreceiver(assembler), if_rhsisnotreceiver(assembler); STATIC_ASSERT(LAST_TYPE == LAST_JS_RECEIVER_TYPE); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), rhs_instance_type), &if_rhsisreceiver, &if_rhsisnotreceiver); assembler->Bind(&if_rhsisreceiver); { // The {lhs} is a Primitive and the {rhs} is a JSReceiver. // Swapping {lhs} and {rhs} is not observable and doesn't // matter for the result, so we can just swap them and use // the JSReceiver handling below (for {lhs} being a JSReceiver). var_lhs.Bind(rhs); var_rhs.Bind(lhs); assembler->Goto(&loop); } assembler->Bind(&if_rhsisnotreceiver); { // The {rhs} is some other Primitive. assembler->Goto(&if_notequal); } } } assembler->Bind(&if_lhsisreceiver); { // Check if the {rhs} is also a JSReceiver. Label if_rhsisreceiver(assembler), if_rhsisnotreceiver(assembler); STATIC_ASSERT(LAST_TYPE == LAST_JS_RECEIVER_TYPE); assembler->Branch( assembler->Int32LessThanOrEqual( assembler->Int32Constant(FIRST_JS_RECEIVER_TYPE), rhs_instance_type), &if_rhsisreceiver, &if_rhsisnotreceiver); assembler->Bind(&if_rhsisreceiver); { // Both {lhs} and {rhs} are different JSReceiver references, so // this cannot be considered equal. assembler->Goto(&if_notequal); } assembler->Bind(&if_rhsisnotreceiver); { // Check if {rhs} is Null or Undefined (an undetectable check // is sufficient here, since we already know that {rhs} is not // a JSReceiver). Label if_rhsisundetectable(assembler), if_rhsisnotundetectable(assembler, Label::kDeferred); Node* rhs_bitfield = assembler->LoadMapBitField(rhs_map); assembler->BranchIfWord32Equal( assembler->Word32And( rhs_bitfield, assembler->Int32Constant(1 << Map::kIsUndetectable)), assembler->Int32Constant(0), &if_rhsisnotundetectable, &if_rhsisundetectable); assembler->Bind(&if_rhsisundetectable); { // Check if {lhs} is an undetectable JSReceiver. Node* lhs_bitfield = assembler->LoadMapBitField(lhs_map); assembler->BranchIfWord32Equal( assembler->Word32And( lhs_bitfield, assembler->Int32Constant(1 << Map::kIsUndetectable)), assembler->Int32Constant(0), &if_notequal, &if_equal); } assembler->Bind(&if_rhsisnotundetectable); { // The {rhs} is some Primitive different from Null and // Undefined, need to convert {lhs} to Primitive first. Callable callable = CodeFactory::NonPrimitiveToPrimitive(assembler->isolate()); var_lhs.Bind(assembler->CallStub(callable, context, lhs)); assembler->Goto(&loop); } } } } } } assembler->Bind(&do_rhsstringtonumber); { Callable callable = CodeFactory::StringToNumber(assembler->isolate()); var_rhs.Bind(assembler->CallStub(callable, context, rhs)); assembler->Goto(&loop); } } assembler->Bind(&do_fcmp); { // Load the {lhs} and {rhs} floating point values. Node* lhs = var_fcmp_lhs.value(); Node* rhs = var_fcmp_rhs.value(); // Perform a fast floating point comparison. assembler->BranchIfFloat64Equal(lhs, rhs, &if_equal, &if_notequal); } assembler->Bind(&if_equal); { result.Bind(assembler->BooleanConstant(mode == kDontNegateResult)); assembler->Goto(&end); } assembler->Bind(&if_notequal); { result.Bind(assembler->BooleanConstant(mode == kNegateResult)); assembler->Goto(&end); } assembler->Bind(&end); return result.value(); } compiler::Node* GenerateStrictEqual(CodeStubAssembler* assembler, ResultMode mode, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { // Here's pseudo-code for the algorithm below in case of kDontNegateResult // mode; for kNegateResult mode we properly negate the result. // // if (lhs == rhs) { // if (lhs->IsHeapNumber()) return HeapNumber::cast(lhs)->value() != NaN; // return true; // } // if (!lhs->IsSmi()) { // if (lhs->IsHeapNumber()) { // if (rhs->IsSmi()) { // return Smi::cast(rhs)->value() == HeapNumber::cast(lhs)->value(); // } else if (rhs->IsHeapNumber()) { // return HeapNumber::cast(rhs)->value() == // HeapNumber::cast(lhs)->value(); // } else { // return false; // } // } else { // if (rhs->IsSmi()) { // return false; // } else { // if (lhs->IsString()) { // if (rhs->IsString()) { // return %StringEqual(lhs, rhs); // } else { // return false; // } // } else if (lhs->IsSimd128()) { // if (rhs->IsSimd128()) { // return %StrictEqual(lhs, rhs); // } // } else { // return false; // } // } // } // } else { // if (rhs->IsSmi()) { // return false; // } else { // if (rhs->IsHeapNumber()) { // return Smi::cast(lhs)->value() == HeapNumber::cast(rhs)->value(); // } else { // return false; // } // } // } typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; typedef compiler::Node Node; Label if_equal(assembler), if_notequal(assembler), end(assembler); Variable result(assembler, MachineRepresentation::kTagged); // Check if {lhs} and {rhs} refer to the same object. Label if_same(assembler), if_notsame(assembler); assembler->Branch(assembler->WordEqual(lhs, rhs), &if_same, &if_notsame); assembler->Bind(&if_same); { // The {lhs} and {rhs} reference the exact same value, yet we need special // treatment for HeapNumber, as NaN is not equal to NaN. GenerateEqual_Same(assembler, lhs, &if_equal, &if_notequal); } assembler->Bind(&if_notsame); { // The {lhs} and {rhs} reference different objects, yet for Smi, HeapNumber, // String and Simd128Value they can still be considered equal. Node* number_map = assembler->HeapNumberMapConstant(); // Check if {lhs} is a Smi or a HeapObject. Label if_lhsissmi(assembler), if_lhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(lhs), &if_lhsissmi, &if_lhsisnotsmi); assembler->Bind(&if_lhsisnotsmi); { // Load the map of {lhs}. Node* lhs_map = assembler->LoadMap(lhs); // Check if {lhs} is a HeapNumber. Label if_lhsisnumber(assembler), if_lhsisnotnumber(assembler); assembler->Branch(assembler->WordEqual(lhs_map, number_map), &if_lhsisnumber, &if_lhsisnotnumber); assembler->Bind(&if_lhsisnumber); { // Check if {rhs} is a Smi or a HeapObject. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); { // Convert {lhs} and {rhs} to floating point values. Node* lhs_value = assembler->LoadHeapNumberValue(lhs); Node* rhs_value = assembler->SmiToFloat64(rhs); // Perform a floating point comparison of {lhs} and {rhs}. assembler->BranchIfFloat64Equal(lhs_value, rhs_value, &if_equal, &if_notequal); } assembler->Bind(&if_rhsisnotsmi); { // Load the map of {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if {rhs} is also a HeapNumber. Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler); assembler->Branch(assembler->WordEqual(rhs_map, number_map), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { // Convert {lhs} and {rhs} to floating point values. Node* lhs_value = assembler->LoadHeapNumberValue(lhs); Node* rhs_value = assembler->LoadHeapNumberValue(rhs); // Perform a floating point comparison of {lhs} and {rhs}. assembler->BranchIfFloat64Equal(lhs_value, rhs_value, &if_equal, &if_notequal); } assembler->Bind(&if_rhsisnotnumber); assembler->Goto(&if_notequal); } } assembler->Bind(&if_lhsisnotnumber); { // Check if {rhs} is a Smi or a HeapObject. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); assembler->Goto(&if_notequal); assembler->Bind(&if_rhsisnotsmi); { // Load the instance type of {lhs}. Node* lhs_instance_type = assembler->LoadMapInstanceType(lhs_map); // Check if {lhs} is a String. Label if_lhsisstring(assembler), if_lhsisnotstring(assembler); assembler->Branch(assembler->Int32LessThan( lhs_instance_type, assembler->Int32Constant(FIRST_NONSTRING_TYPE)), &if_lhsisstring, &if_lhsisnotstring); assembler->Bind(&if_lhsisstring); { // Load the instance type of {rhs}. Node* rhs_instance_type = assembler->LoadInstanceType(rhs); // Check if {rhs} is also a String. Label if_rhsisstring(assembler, Label::kDeferred), if_rhsisnotstring(assembler); assembler->Branch(assembler->Int32LessThan( rhs_instance_type, assembler->Int32Constant( FIRST_NONSTRING_TYPE)), &if_rhsisstring, &if_rhsisnotstring); assembler->Bind(&if_rhsisstring); { Callable callable = (mode == kDontNegateResult) ? CodeFactory::StringEqual(assembler->isolate()) : CodeFactory::StringNotEqual(assembler->isolate()); result.Bind(assembler->CallStub(callable, context, lhs, rhs)); assembler->Goto(&end); } assembler->Bind(&if_rhsisnotstring); assembler->Goto(&if_notequal); } assembler->Bind(&if_lhsisnotstring); { // Check if {lhs} is a Simd128Value. Label if_lhsissimd128value(assembler), if_lhsisnotsimd128value(assembler); assembler->Branch(assembler->Word32Equal( lhs_instance_type, assembler->Int32Constant(SIMD128_VALUE_TYPE)), &if_lhsissimd128value, &if_lhsisnotsimd128value); assembler->Bind(&if_lhsissimd128value); { // Load the map of {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // Check if {rhs} is also a Simd128Value that is equal to {lhs}. GenerateEqual_Simd128Value_HeapObject(assembler, lhs, lhs_map, rhs, rhs_map, &if_equal, &if_notequal); } assembler->Bind(&if_lhsisnotsimd128value); assembler->Goto(&if_notequal); } } } } assembler->Bind(&if_lhsissmi); { // We already know that {lhs} and {rhs} are not reference equal, and {lhs} // is a Smi; so {lhs} and {rhs} can only be strictly equal if {rhs} is a // HeapNumber with an equal floating point value. // Check if {rhs} is a Smi or a HeapObject. Label if_rhsissmi(assembler), if_rhsisnotsmi(assembler); assembler->Branch(assembler->WordIsSmi(rhs), &if_rhsissmi, &if_rhsisnotsmi); assembler->Bind(&if_rhsissmi); assembler->Goto(&if_notequal); assembler->Bind(&if_rhsisnotsmi); { // Load the map of the {rhs}. Node* rhs_map = assembler->LoadMap(rhs); // The {rhs} could be a HeapNumber with the same value as {lhs}. Label if_rhsisnumber(assembler), if_rhsisnotnumber(assembler); assembler->Branch(assembler->WordEqual(rhs_map, number_map), &if_rhsisnumber, &if_rhsisnotnumber); assembler->Bind(&if_rhsisnumber); { // Convert {lhs} and {rhs} to floating point values. Node* lhs_value = assembler->SmiToFloat64(lhs); Node* rhs_value = assembler->LoadHeapNumberValue(rhs); // Perform a floating point comparison of {lhs} and {rhs}. assembler->BranchIfFloat64Equal(lhs_value, rhs_value, &if_equal, &if_notequal); } assembler->Bind(&if_rhsisnotnumber); assembler->Goto(&if_notequal); } } } assembler->Bind(&if_equal); { result.Bind(assembler->BooleanConstant(mode == kDontNegateResult)); assembler->Goto(&end); } assembler->Bind(&if_notequal); { result.Bind(assembler->BooleanConstant(mode == kNegateResult)); assembler->Goto(&end); } assembler->Bind(&end); return result.value(); } void GenerateStringRelationalComparison(CodeStubAssembler* assembler, RelationalComparisonMode mode) { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; Node* lhs = assembler->Parameter(0); Node* rhs = assembler->Parameter(1); Node* context = assembler->Parameter(2); Label if_less(assembler), if_equal(assembler), if_greater(assembler); // Fast check to see if {lhs} and {rhs} refer to the same String object. Label if_same(assembler), if_notsame(assembler); assembler->Branch(assembler->WordEqual(lhs, rhs), &if_same, &if_notsame); assembler->Bind(&if_same); assembler->Goto(&if_equal); assembler->Bind(&if_notsame); { // Load instance types of {lhs} and {rhs}. Node* lhs_instance_type = assembler->LoadInstanceType(lhs); Node* rhs_instance_type = assembler->LoadInstanceType(rhs); // Combine the instance types into a single 16-bit value, so we can check // both of them at once. Node* both_instance_types = assembler->Word32Or( lhs_instance_type, assembler->Word32Shl(rhs_instance_type, assembler->Int32Constant(8))); // Check that both {lhs} and {rhs} are flat one-byte strings. int const kBothSeqOneByteStringMask = kStringEncodingMask | kStringRepresentationMask | ((kStringEncodingMask | kStringRepresentationMask) << 8); int const kBothSeqOneByteStringTag = kOneByteStringTag | kSeqStringTag | ((kOneByteStringTag | kSeqStringTag) << 8); Label if_bothonebyteseqstrings(assembler), if_notbothonebyteseqstrings(assembler); assembler->Branch(assembler->Word32Equal( assembler->Word32And(both_instance_types, assembler->Int32Constant( kBothSeqOneByteStringMask)), assembler->Int32Constant(kBothSeqOneByteStringTag)), &if_bothonebyteseqstrings, &if_notbothonebyteseqstrings); assembler->Bind(&if_bothonebyteseqstrings); { // Load the length of {lhs} and {rhs}. Node* lhs_length = assembler->LoadStringLength(lhs); Node* rhs_length = assembler->LoadStringLength(rhs); // Determine the minimum length. Node* length = assembler->SmiMin(lhs_length, rhs_length); // Compute the effective offset of the first character. Node* begin = assembler->IntPtrConstant(SeqOneByteString::kHeaderSize - kHeapObjectTag); // Compute the first offset after the string from the length. Node* end = assembler->IntPtrAdd(begin, assembler->SmiUntag(length)); // Loop over the {lhs} and {rhs} strings to see if they are equal. Variable var_offset(assembler, MachineType::PointerRepresentation()); Label loop(assembler, &var_offset); var_offset.Bind(begin); assembler->Goto(&loop); assembler->Bind(&loop); { // Check if {offset} equals {end}. Node* offset = var_offset.value(); Label if_done(assembler), if_notdone(assembler); assembler->Branch(assembler->WordEqual(offset, end), &if_done, &if_notdone); assembler->Bind(&if_notdone); { // Load the next characters from {lhs} and {rhs}. Node* lhs_value = assembler->Load(MachineType::Uint8(), lhs, offset); Node* rhs_value = assembler->Load(MachineType::Uint8(), rhs, offset); // Check if the characters match. Label if_valueissame(assembler), if_valueisnotsame(assembler); assembler->Branch(assembler->Word32Equal(lhs_value, rhs_value), &if_valueissame, &if_valueisnotsame); assembler->Bind(&if_valueissame); { // Advance to next character. var_offset.Bind( assembler->IntPtrAdd(offset, assembler->IntPtrConstant(1))); } assembler->Goto(&loop); assembler->Bind(&if_valueisnotsame); assembler->BranchIf(assembler->Uint32LessThan(lhs_value, rhs_value), &if_less, &if_greater); } assembler->Bind(&if_done); { // All characters up to the min length are equal, decide based on // string length. Label if_lengthisequal(assembler), if_lengthisnotequal(assembler); assembler->Branch(assembler->SmiEqual(lhs_length, rhs_length), &if_lengthisequal, &if_lengthisnotequal); assembler->Bind(&if_lengthisequal); assembler->Goto(&if_equal); assembler->Bind(&if_lengthisnotequal); assembler->BranchIfSmiLessThan(lhs_length, rhs_length, &if_less, &if_greater); } } } assembler->Bind(&if_notbothonebyteseqstrings); { // TODO(bmeurer): Add fast case support for flattened cons strings; // also add support for two byte string relational comparisons. switch (mode) { case kLessThan: assembler->TailCallRuntime(Runtime::kStringLessThan, context, lhs, rhs); break; case kLessThanOrEqual: assembler->TailCallRuntime(Runtime::kStringLessThanOrEqual, context, lhs, rhs); break; case kGreaterThan: assembler->TailCallRuntime(Runtime::kStringGreaterThan, context, lhs, rhs); break; case kGreaterThanOrEqual: assembler->TailCallRuntime(Runtime::kStringGreaterThanOrEqual, context, lhs, rhs); break; } } } assembler->Bind(&if_less); switch (mode) { case kLessThan: case kLessThanOrEqual: assembler->Return(assembler->BooleanConstant(true)); break; case kGreaterThan: case kGreaterThanOrEqual: assembler->Return(assembler->BooleanConstant(false)); break; } assembler->Bind(&if_equal); switch (mode) { case kLessThan: case kGreaterThan: assembler->Return(assembler->BooleanConstant(false)); break; case kLessThanOrEqual: case kGreaterThanOrEqual: assembler->Return(assembler->BooleanConstant(true)); break; } assembler->Bind(&if_greater); switch (mode) { case kLessThan: case kLessThanOrEqual: assembler->Return(assembler->BooleanConstant(false)); break; case kGreaterThan: case kGreaterThanOrEqual: assembler->Return(assembler->BooleanConstant(true)); break; } } void GenerateStringEqual(CodeStubAssembler* assembler, ResultMode mode) { // Here's pseudo-code for the algorithm below in case of kDontNegateResult // mode; for kNegateResult mode we properly negate the result. // // if (lhs == rhs) return true; // if (lhs->length() != rhs->length()) return false; // if (lhs->IsInternalizedString() && rhs->IsInternalizedString()) { // return false; // } // if (lhs->IsSeqOneByteString() && rhs->IsSeqOneByteString()) { // for (i = 0; i != lhs->length(); ++i) { // if (lhs[i] != rhs[i]) return false; // } // return true; // } // return %StringEqual(lhs, rhs); typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; Node* lhs = assembler->Parameter(0); Node* rhs = assembler->Parameter(1); Node* context = assembler->Parameter(2); Label if_equal(assembler), if_notequal(assembler); // Fast check to see if {lhs} and {rhs} refer to the same String object. Label if_same(assembler), if_notsame(assembler); assembler->Branch(assembler->WordEqual(lhs, rhs), &if_same, &if_notsame); assembler->Bind(&if_same); assembler->Goto(&if_equal); assembler->Bind(&if_notsame); { // The {lhs} and {rhs} don't refer to the exact same String object. // Load the length of {lhs} and {rhs}. Node* lhs_length = assembler->LoadStringLength(lhs); Node* rhs_length = assembler->LoadStringLength(rhs); // Check if the lengths of {lhs} and {rhs} are equal. Label if_lengthisequal(assembler), if_lengthisnotequal(assembler); assembler->Branch(assembler->WordEqual(lhs_length, rhs_length), &if_lengthisequal, &if_lengthisnotequal); assembler->Bind(&if_lengthisequal); { // Load instance types of {lhs} and {rhs}. Node* lhs_instance_type = assembler->LoadInstanceType(lhs); Node* rhs_instance_type = assembler->LoadInstanceType(rhs); // Combine the instance types into a single 16-bit value, so we can check // both of them at once. Node* both_instance_types = assembler->Word32Or( lhs_instance_type, assembler->Word32Shl(rhs_instance_type, assembler->Int32Constant(8))); // Check if both {lhs} and {rhs} are internalized. int const kBothInternalizedMask = kIsNotInternalizedMask | (kIsNotInternalizedMask << 8); int const kBothInternalizedTag = kInternalizedTag | (kInternalizedTag << 8); Label if_bothinternalized(assembler), if_notbothinternalized(assembler); assembler->Branch(assembler->Word32Equal( assembler->Word32And(both_instance_types, assembler->Int32Constant( kBothInternalizedMask)), assembler->Int32Constant(kBothInternalizedTag)), &if_bothinternalized, &if_notbothinternalized); assembler->Bind(&if_bothinternalized); { // Fast negative check for internalized-to-internalized equality. assembler->Goto(&if_notequal); } assembler->Bind(&if_notbothinternalized); { // Check that both {lhs} and {rhs} are flat one-byte strings. int const kBothSeqOneByteStringMask = kStringEncodingMask | kStringRepresentationMask | ((kStringEncodingMask | kStringRepresentationMask) << 8); int const kBothSeqOneByteStringTag = kOneByteStringTag | kSeqStringTag | ((kOneByteStringTag | kSeqStringTag) << 8); Label if_bothonebyteseqstrings(assembler), if_notbothonebyteseqstrings(assembler); assembler->Branch( assembler->Word32Equal( assembler->Word32And( both_instance_types, assembler->Int32Constant(kBothSeqOneByteStringMask)), assembler->Int32Constant(kBothSeqOneByteStringTag)), &if_bothonebyteseqstrings, &if_notbothonebyteseqstrings); assembler->Bind(&if_bothonebyteseqstrings); { // Compute the effective offset of the first character. Node* begin = assembler->IntPtrConstant( SeqOneByteString::kHeaderSize - kHeapObjectTag); // Compute the first offset after the string from the length. Node* end = assembler->IntPtrAdd(begin, assembler->SmiUntag(lhs_length)); // Loop over the {lhs} and {rhs} strings to see if they are equal. Variable var_offset(assembler, MachineType::PointerRepresentation()); Label loop(assembler, &var_offset); var_offset.Bind(begin); assembler->Goto(&loop); assembler->Bind(&loop); { // Check if {offset} equals {end}. Node* offset = var_offset.value(); Label if_done(assembler), if_notdone(assembler); assembler->Branch(assembler->WordEqual(offset, end), &if_done, &if_notdone); assembler->Bind(&if_notdone); { // Load the next characters from {lhs} and {rhs}. Node* lhs_value = assembler->Load(MachineType::Uint8(), lhs, offset); Node* rhs_value = assembler->Load(MachineType::Uint8(), rhs, offset); // Check if the characters match. Label if_valueissame(assembler), if_valueisnotsame(assembler); assembler->Branch(assembler->Word32Equal(lhs_value, rhs_value), &if_valueissame, &if_valueisnotsame); assembler->Bind(&if_valueissame); { // Advance to next character. var_offset.Bind( assembler->IntPtrAdd(offset, assembler->IntPtrConstant(1))); } assembler->Goto(&loop); assembler->Bind(&if_valueisnotsame); assembler->Goto(&if_notequal); } assembler->Bind(&if_done); assembler->Goto(&if_equal); } } assembler->Bind(&if_notbothonebyteseqstrings); { // TODO(bmeurer): Add fast case support for flattened cons strings; // also add support for two byte string equality checks. Runtime::FunctionId function_id = (mode == kDontNegateResult) ? Runtime::kStringEqual : Runtime::kStringNotEqual; assembler->TailCallRuntime(function_id, context, lhs, rhs); } } } assembler->Bind(&if_lengthisnotequal); { // Mismatch in length of {lhs} and {rhs}, cannot be equal. assembler->Goto(&if_notequal); } } assembler->Bind(&if_equal); assembler->Return(assembler->BooleanConstant(mode == kDontNegateResult)); assembler->Bind(&if_notequal); assembler->Return(assembler->BooleanConstant(mode == kNegateResult)); } } // namespace void LoadApiGetterStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* context = assembler->Parameter(Descriptor::kContext); Node* receiver = assembler->Parameter(Descriptor::kReceiver); // For now we only support receiver_is_holder. DCHECK(receiver_is_holder()); Node* holder = receiver; Node* map = assembler->LoadMap(receiver); Node* descriptors = assembler->LoadMapDescriptors(map); Node* value_index = assembler->IntPtrConstant(DescriptorArray::ToValueIndex(index())); Node* callback = assembler->LoadFixedArrayElement( descriptors, value_index, 0, CodeStubAssembler::INTPTR_PARAMETERS); assembler->TailCallStub(CodeFactory::ApiGetter(isolate()), context, receiver, holder, callback); } void StoreFieldStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; FieldIndex index = this->index(); Representation representation = this->representation(); assembler->Comment("StoreFieldStub: inobject=%d, offset=%d, rep=%s", index.is_inobject(), index.offset(), representation.Mnemonic()); Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* name = assembler->Parameter(Descriptor::kName); Node* value = assembler->Parameter(Descriptor::kValue); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); Label miss(assembler); Node* prepared_value = assembler->PrepareValueForWrite(value, representation, &miss); assembler->StoreNamedField(receiver, index, representation, prepared_value, false); assembler->Return(value); // Only stores to tagged field can't bailout. if (!representation.IsTagged()) { assembler->Bind(&miss); { assembler->Comment("Miss"); assembler->TailCallRuntime(Runtime::kStoreIC_Miss, context, value, slot, vector, receiver, name); } } } void StoreGlobalStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; assembler->Comment( "StoreGlobalStub: cell_type=%d, constant_type=%d, check_global=%d", cell_type(), PropertyCellType::kConstantType == cell_type() ? static_cast<int>(constant_type()) : -1, check_global()); Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* name = assembler->Parameter(Descriptor::kName); Node* value = assembler->Parameter(Descriptor::kValue); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); Label miss(assembler); if (check_global()) { // Check that the map of the global has not changed: use a placeholder map // that will be replaced later with the global object's map. Node* proxy_map = assembler->LoadMap(receiver); Node* global = assembler->LoadObjectField(proxy_map, Map::kPrototypeOffset); Node* map_cell = assembler->HeapConstant(isolate()->factory()->NewWeakCell( StoreGlobalStub::global_map_placeholder(isolate()))); Node* expected_map = assembler->LoadWeakCellValue(map_cell); Node* map = assembler->LoadMap(global); assembler->GotoIf(assembler->WordNotEqual(expected_map, map), &miss); } Node* weak_cell = assembler->HeapConstant(isolate()->factory()->NewWeakCell( StoreGlobalStub::property_cell_placeholder(isolate()))); Node* cell = assembler->LoadWeakCellValue(weak_cell); assembler->GotoIf(assembler->WordIsSmi(cell), &miss); // Load the payload of the global parameter cell. A hole indicates that the // cell has been invalidated and that the store must be handled by the // runtime. Node* cell_contents = assembler->LoadObjectField(cell, PropertyCell::kValueOffset); PropertyCellType cell_type = this->cell_type(); if (cell_type == PropertyCellType::kConstant || cell_type == PropertyCellType::kUndefined) { // This is always valid for all states a cell can be in. assembler->GotoIf(assembler->WordNotEqual(cell_contents, value), &miss); } else { assembler->GotoIf( assembler->WordEqual(cell_contents, assembler->TheHoleConstant()), &miss); // When dealing with constant types, the type may be allowed to change, as // long as optimized code remains valid. bool value_is_smi = false; if (cell_type == PropertyCellType::kConstantType) { switch (constant_type()) { case PropertyCellConstantType::kSmi: assembler->GotoUnless(assembler->WordIsSmi(value), &miss); value_is_smi = true; break; case PropertyCellConstantType::kStableMap: { // It is sufficient here to check that the value and cell contents // have identical maps, no matter if they are stable or not or if they // are the maps that were originally in the cell or not. If optimized // code will deopt when a cell has a unstable map and if it has a // dependency on a stable map, it will deopt if the map destabilizes. assembler->GotoIf(assembler->WordIsSmi(value), &miss); assembler->GotoIf(assembler->WordIsSmi(cell_contents), &miss); Node* expected_map = assembler->LoadMap(cell_contents); Node* map = assembler->LoadMap(value); assembler->GotoIf(assembler->WordNotEqual(expected_map, map), &miss); break; } } } if (value_is_smi) { assembler->StoreObjectFieldNoWriteBarrier( cell, PropertyCell::kValueOffset, value); } else { assembler->StoreObjectField(cell, PropertyCell::kValueOffset, value); } } assembler->Return(value); assembler->Bind(&miss); { assembler->Comment("Miss"); assembler->TailCallRuntime(Runtime::kStoreIC_Miss, context, value, slot, vector, receiver, name); } } void KeyedLoadSloppyArgumentsStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* key = assembler->Parameter(Descriptor::kName); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); Label miss(assembler); Node* result = assembler->LoadKeyedSloppyArguments(receiver, key, &miss); assembler->Return(result); assembler->Bind(&miss); { assembler->Comment("Miss"); assembler->TailCallRuntime(Runtime::kKeyedLoadIC_Miss, context, receiver, key, slot, vector); } } void KeyedStoreSloppyArgumentsStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* key = assembler->Parameter(Descriptor::kName); Node* value = assembler->Parameter(Descriptor::kValue); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); Label miss(assembler); assembler->StoreKeyedSloppyArguments(receiver, key, value, &miss); assembler->Return(value); assembler->Bind(&miss); { assembler->Comment("Miss"); assembler->TailCallRuntime(Runtime::kKeyedStoreIC_Miss, context, value, slot, vector, receiver, key); } } void LoadScriptContextFieldStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; assembler->Comment("LoadScriptContextFieldStub: context_index=%d, slot=%d", context_index(), slot_index()); Node* context = assembler->Parameter(Descriptor::kContext); Node* script_context = assembler->LoadScriptContext(context, context_index()); Node* result = assembler->LoadFixedArrayElement( script_context, assembler->IntPtrConstant(slot_index()), 0, CodeStubAssembler::INTPTR_PARAMETERS); assembler->Return(result); } void StoreScriptContextFieldStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; assembler->Comment("StoreScriptContextFieldStub: context_index=%d, slot=%d", context_index(), slot_index()); Node* value = assembler->Parameter(Descriptor::kValue); Node* context = assembler->Parameter(Descriptor::kContext); Node* script_context = assembler->LoadScriptContext(context, context_index()); assembler->StoreFixedArrayElement( script_context, assembler->IntPtrConstant(slot_index()), value, UPDATE_WRITE_BARRIER, CodeStubAssembler::INTPTR_PARAMETERS); assembler->Return(value); } // static compiler::Node* LessThanStub::Generate(CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { return GenerateAbstractRelationalComparison(assembler, kLessThan, lhs, rhs, context); } // static compiler::Node* LessThanOrEqualStub::Generate(CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { return GenerateAbstractRelationalComparison(assembler, kLessThanOrEqual, lhs, rhs, context); } // static compiler::Node* GreaterThanStub::Generate(CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { return GenerateAbstractRelationalComparison(assembler, kGreaterThan, lhs, rhs, context); } // static compiler::Node* GreaterThanOrEqualStub::Generate(CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { return GenerateAbstractRelationalComparison(assembler, kGreaterThanOrEqual, lhs, rhs, context); } // static compiler::Node* EqualStub::Generate(CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { return GenerateEqual(assembler, kDontNegateResult, lhs, rhs, context); } // static compiler::Node* NotEqualStub::Generate(CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { return GenerateEqual(assembler, kNegateResult, lhs, rhs, context); } // static compiler::Node* StrictEqualStub::Generate(CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { return GenerateStrictEqual(assembler, kDontNegateResult, lhs, rhs, context); } // static compiler::Node* StrictNotEqualStub::Generate(CodeStubAssembler* assembler, compiler::Node* lhs, compiler::Node* rhs, compiler::Node* context) { return GenerateStrictEqual(assembler, kNegateResult, lhs, rhs, context); } void StringEqualStub::GenerateAssembly(CodeStubAssembler* assembler) const { GenerateStringEqual(assembler, kDontNegateResult); } void StringNotEqualStub::GenerateAssembly(CodeStubAssembler* assembler) const { GenerateStringEqual(assembler, kNegateResult); } void StringLessThanStub::GenerateAssembly(CodeStubAssembler* assembler) const { GenerateStringRelationalComparison(assembler, kLessThan); } void StringLessThanOrEqualStub::GenerateAssembly( CodeStubAssembler* assembler) const { GenerateStringRelationalComparison(assembler, kLessThanOrEqual); } void StringGreaterThanStub::GenerateAssembly( CodeStubAssembler* assembler) const { GenerateStringRelationalComparison(assembler, kGreaterThan); } void StringGreaterThanOrEqualStub::GenerateAssembly( CodeStubAssembler* assembler) const { GenerateStringRelationalComparison(assembler, kGreaterThanOrEqual); } void ToLengthStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; Node* context = assembler->Parameter(1); // We might need to loop once for ToNumber conversion. Variable var_len(assembler, MachineRepresentation::kTagged); Label loop(assembler, &var_len); var_len.Bind(assembler->Parameter(0)); assembler->Goto(&loop); assembler->Bind(&loop); { // Shared entry points. Label return_len(assembler), return_two53minus1(assembler, Label::kDeferred), return_zero(assembler, Label::kDeferred); // Load the current {len} value. Node* len = var_len.value(); // Check if {len} is a positive Smi. assembler->GotoIf(assembler->WordIsPositiveSmi(len), &return_len); // Check if {len} is a (negative) Smi. assembler->GotoIf(assembler->WordIsSmi(len), &return_zero); // Check if {len} is a HeapNumber. Label if_lenisheapnumber(assembler), if_lenisnotheapnumber(assembler, Label::kDeferred); assembler->Branch(assembler->WordEqual(assembler->LoadMap(len), assembler->HeapNumberMapConstant()), &if_lenisheapnumber, &if_lenisnotheapnumber); assembler->Bind(&if_lenisheapnumber); { // Load the floating-point value of {len}. Node* len_value = assembler->LoadHeapNumberValue(len); // Check if {len} is not greater than zero. assembler->GotoUnless(assembler->Float64GreaterThan( len_value, assembler->Float64Constant(0.0)), &return_zero); // Check if {len} is greater than or equal to 2^53-1. assembler->GotoIf( assembler->Float64GreaterThanOrEqual( len_value, assembler->Float64Constant(kMaxSafeInteger)), &return_two53minus1); // Round the {len} towards -Infinity. Node* value = assembler->Float64Floor(len_value); Node* result = assembler->ChangeFloat64ToTagged(value); assembler->Return(result); } assembler->Bind(&if_lenisnotheapnumber); { // Need to convert {len} to a Number first. Callable callable = CodeFactory::NonNumberToNumber(assembler->isolate()); var_len.Bind(assembler->CallStub(callable, context, len)); assembler->Goto(&loop); } assembler->Bind(&return_len); assembler->Return(var_len.value()); assembler->Bind(&return_two53minus1); assembler->Return(assembler->NumberConstant(kMaxSafeInteger)); assembler->Bind(&return_zero); assembler->Return(assembler->SmiConstant(Smi::FromInt(0))); } } void ToIntegerStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* input = assembler->Parameter(Descriptor::kArgument); Node* context = assembler->Parameter(Descriptor::kContext); assembler->Return(assembler->ToInteger(context, input)); } void StoreInterceptorStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* name = assembler->Parameter(Descriptor::kName); Node* value = assembler->Parameter(Descriptor::kValue); Node* context = assembler->Parameter(Descriptor::kContext); assembler->TailCallRuntime(Runtime::kStorePropertyWithInterceptor, context, receiver, name, value); } void LoadIndexedInterceptorStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; typedef CodeStubAssembler::Label Label; Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* key = assembler->Parameter(Descriptor::kName); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); Label if_keyispositivesmi(assembler), if_keyisinvalid(assembler); assembler->Branch(assembler->WordIsPositiveSmi(key), &if_keyispositivesmi, &if_keyisinvalid); assembler->Bind(&if_keyispositivesmi); assembler->TailCallRuntime(Runtime::kLoadElementWithInterceptor, context, receiver, key); assembler->Bind(&if_keyisinvalid); assembler->TailCallRuntime(Runtime::kKeyedLoadIC_Miss, context, receiver, key, slot, vector); } // static bool FastCloneShallowObjectStub::IsSupported(ObjectLiteral* expr) { // FastCloneShallowObjectStub doesn't copy elements, and object literals don't // support copy-on-write (COW) elements for now. // TODO(mvstanton): make object literals support COW elements. return expr->fast_elements() && expr->has_shallow_properties() && expr->properties_count() <= kMaximumClonedProperties; } // static int FastCloneShallowObjectStub::PropertiesCount(int literal_length) { // This heuristic of setting empty literals to have // kInitialGlobalObjectUnusedPropertiesCount must remain in-sync with the // runtime. // TODO(verwaest): Unify this with the heuristic in the runtime. return literal_length == 0 ? JSObject::kInitialGlobalObjectUnusedPropertiesCount : literal_length; } // static compiler::Node* FastCloneShallowObjectStub::GenerateFastPath( CodeStubAssembler* assembler, compiler::CodeAssembler::Label* call_runtime, compiler::Node* closure, compiler::Node* literals_index, compiler::Node* properties_count) { typedef compiler::Node Node; typedef compiler::CodeAssembler::Label Label; typedef compiler::CodeAssembler::Variable Variable; Node* literals_array = assembler->LoadObjectField(closure, JSFunction::kLiteralsOffset); Node* allocation_site = assembler->LoadFixedArrayElement( literals_array, literals_index, LiteralsArray::kFirstLiteralIndex * kPointerSize, CodeStubAssembler::SMI_PARAMETERS); Node* undefined = assembler->UndefinedConstant(); assembler->GotoIf(assembler->WordEqual(allocation_site, undefined), call_runtime); // Calculate the object and allocation size based on the properties count. Node* object_size = assembler->IntPtrAdd( assembler->WordShl(properties_count, kPointerSizeLog2), assembler->IntPtrConstant(JSObject::kHeaderSize)); Node* allocation_size = object_size; if (FLAG_allocation_site_pretenuring) { allocation_size = assembler->IntPtrAdd( object_size, assembler->IntPtrConstant(AllocationMemento::kSize)); } Node* boilerplate = assembler->LoadObjectField( allocation_site, AllocationSite::kTransitionInfoOffset); Node* boilerplate_map = assembler->LoadMap(boilerplate); Node* instance_size = assembler->LoadMapInstanceSize(boilerplate_map); Node* size_in_words = assembler->WordShr(object_size, kPointerSizeLog2); assembler->GotoUnless(assembler->Word32Equal(instance_size, size_in_words), call_runtime); Node* copy = assembler->Allocate(allocation_size); // Copy boilerplate elements. Variable offset(assembler, MachineType::PointerRepresentation()); offset.Bind(assembler->IntPtrConstant(-kHeapObjectTag)); Node* end_offset = assembler->IntPtrAdd(object_size, offset.value()); Label loop_body(assembler, &offset), loop_check(assembler, &offset); // We should always have an object size greater than zero. assembler->Goto(&loop_body); assembler->Bind(&loop_body); { // The Allocate above guarantees that the copy lies in new space. This // allows us to skip write barriers. This is necessary since we may also be // copying unboxed doubles. Node* field = assembler->Load(MachineType::IntPtr(), boilerplate, offset.value()); assembler->StoreNoWriteBarrier(MachineType::PointerRepresentation(), copy, offset.value(), field); assembler->Goto(&loop_check); } assembler->Bind(&loop_check); { offset.Bind(assembler->IntPtrAdd(offset.value(), assembler->IntPtrConstant(kPointerSize))); assembler->GotoUnless( assembler->IntPtrGreaterThanOrEqual(offset.value(), end_offset), &loop_body); } if (FLAG_allocation_site_pretenuring) { Node* memento = assembler->InnerAllocate(copy, object_size); assembler->StoreObjectFieldNoWriteBarrier( memento, HeapObject::kMapOffset, assembler->LoadRoot(Heap::kAllocationMementoMapRootIndex)); assembler->StoreObjectFieldNoWriteBarrier( memento, AllocationMemento::kAllocationSiteOffset, allocation_site); Node* memento_create_count = assembler->LoadObjectField( allocation_site, AllocationSite::kPretenureCreateCountOffset); memento_create_count = assembler->SmiAdd( memento_create_count, assembler->SmiConstant(Smi::FromInt(1))); assembler->StoreObjectFieldNoWriteBarrier( allocation_site, AllocationSite::kPretenureCreateCountOffset, memento_create_count); } // TODO(verwaest): Allocate and fill in double boxes. return copy; } void FastCloneShallowObjectStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; Label call_runtime(assembler); Node* closure = assembler->Parameter(0); Node* literals_index = assembler->Parameter(1); Node* properties_count = assembler->IntPtrConstant(PropertiesCount(this->length())); Node* copy = GenerateFastPath(assembler, &call_runtime, closure, literals_index, properties_count); assembler->Return(copy); assembler->Bind(&call_runtime); Node* constant_properties = assembler->Parameter(2); Node* flags = assembler->Parameter(3); Node* context = assembler->Parameter(4); assembler->TailCallRuntime(Runtime::kCreateObjectLiteral, context, closure, literals_index, constant_properties, flags); } template<class StateType> void HydrogenCodeStub::TraceTransition(StateType from, StateType to) { // Note: Although a no-op transition is semantically OK, it is hinting at a // bug somewhere in our state transition machinery. DCHECK(from != to); if (!FLAG_trace_ic) return; OFStream os(stdout); os << "["; PrintBaseName(os); os << ": " << from << "=>" << to << "]" << std::endl; } // TODO(svenpanne) Make this a real infix_ostream_iterator. class SimpleListPrinter { public: explicit SimpleListPrinter(std::ostream& os) : os_(os), first_(true) {} void Add(const char* s) { if (first_) { first_ = false; } else { os_ << ","; } os_ << s; } private: std::ostream& os_; bool first_; }; void CallICStub::PrintState(std::ostream& os) const { // NOLINT os << state(); } void JSEntryStub::FinishCode(Handle<Code> code) { Handle<FixedArray> handler_table = code->GetIsolate()->factory()->NewFixedArray(1, TENURED); handler_table->set(0, Smi::FromInt(handler_offset_)); code->set_handler_table(*handler_table); } void LoadDictionaryElementStub::InitializeDescriptor( CodeStubDescriptor* descriptor) { descriptor->Initialize( FUNCTION_ADDR(Runtime_KeyedLoadIC_MissFromStubFailure)); } void KeyedLoadGenericStub::InitializeDescriptor( CodeStubDescriptor* descriptor) { descriptor->Initialize( Runtime::FunctionForId(Runtime::kKeyedGetProperty)->entry); } void HandlerStub::InitializeDescriptor(CodeStubDescriptor* descriptor) { DCHECK(kind() == Code::LOAD_IC || kind() == Code::KEYED_LOAD_IC); if (kind() == Code::KEYED_LOAD_IC) { descriptor->Initialize( FUNCTION_ADDR(Runtime_KeyedLoadIC_MissFromStubFailure)); } } CallInterfaceDescriptor HandlerStub::GetCallInterfaceDescriptor() const { if (kind() == Code::LOAD_IC || kind() == Code::KEYED_LOAD_IC) { return LoadWithVectorDescriptor(isolate()); } else { DCHECK(kind() == Code::STORE_IC || kind() == Code::KEYED_STORE_IC); return StoreWithVectorDescriptor(isolate()); } } void NumberToStringStub::InitializeDescriptor(CodeStubDescriptor* descriptor) { descriptor->Initialize( Runtime::FunctionForId(Runtime::kNumberToString)->entry); descriptor->SetMissHandler(Runtime::kNumberToString); } void RegExpConstructResultStub::InitializeDescriptor( CodeStubDescriptor* descriptor) { descriptor->Initialize( Runtime::FunctionForId(Runtime::kRegExpConstructResult)->entry); descriptor->SetMissHandler(Runtime::kRegExpConstructResult); } void TransitionElementsKindStub::InitializeDescriptor( CodeStubDescriptor* descriptor) { descriptor->Initialize( Runtime::FunctionForId(Runtime::kTransitionElementsKind)->entry); } void AllocateHeapNumberStub::InitializeDescriptor( CodeStubDescriptor* descriptor) { descriptor->Initialize( Runtime::FunctionForId(Runtime::kAllocateHeapNumber)->entry); } #define SIMD128_INIT_DESC(TYPE, Type, type, lane_count, lane_type) \ void Allocate##Type##Stub::InitializeDescriptor( \ CodeStubDescriptor* descriptor) { \ descriptor->Initialize( \ Runtime::FunctionForId(Runtime::kCreate##Type)->entry); \ } SIMD128_TYPES(SIMD128_INIT_DESC) #undef SIMD128_INIT_DESC void ToBooleanICStub::InitializeDescriptor(CodeStubDescriptor* descriptor) { descriptor->Initialize(FUNCTION_ADDR(Runtime_ToBooleanIC_Miss)); descriptor->SetMissHandler(Runtime::kToBooleanIC_Miss); } void BinaryOpICStub::InitializeDescriptor(CodeStubDescriptor* descriptor) { descriptor->Initialize(FUNCTION_ADDR(Runtime_BinaryOpIC_Miss)); descriptor->SetMissHandler(Runtime::kBinaryOpIC_Miss); } void BinaryOpWithAllocationSiteStub::InitializeDescriptor( CodeStubDescriptor* descriptor) { descriptor->Initialize( FUNCTION_ADDR(Runtime_BinaryOpIC_MissWithAllocationSite)); } void StringAddStub::InitializeDescriptor(CodeStubDescriptor* descriptor) { descriptor->Initialize(Runtime::FunctionForId(Runtime::kStringAdd)->entry); descriptor->SetMissHandler(Runtime::kStringAdd); } namespace { compiler::Node* GenerateHasProperty( CodeStubAssembler* assembler, compiler::Node* object, compiler::Node* key, compiler::Node* context, Runtime::FunctionId fallback_runtime_function_id) { typedef compiler::Node Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; Label call_runtime(assembler, Label::kDeferred), return_true(assembler), return_false(assembler), end(assembler); CodeStubAssembler::LookupInHolder lookup_property_in_holder = [assembler, &return_true](Node* receiver, Node* holder, Node* holder_map, Node* holder_instance_type, Node* unique_name, Label* next_holder, Label* if_bailout) { assembler->TryHasOwnProperty(holder, holder_map, holder_instance_type, unique_name, &return_true, next_holder, if_bailout); }; CodeStubAssembler::LookupInHolder lookup_element_in_holder = [assembler, &return_true](Node* receiver, Node* holder, Node* holder_map, Node* holder_instance_type, Node* index, Label* next_holder, Label* if_bailout) { assembler->TryLookupElement(holder, holder_map, holder_instance_type, index, &return_true, next_holder, if_bailout); }; assembler->TryPrototypeChainLookup(object, key, lookup_property_in_holder, lookup_element_in_holder, &return_false, &call_runtime); Variable result(assembler, MachineRepresentation::kTagged); assembler->Bind(&return_true); { result.Bind(assembler->BooleanConstant(true)); assembler->Goto(&end); } assembler->Bind(&return_false); { result.Bind(assembler->BooleanConstant(false)); assembler->Goto(&end); } assembler->Bind(&call_runtime); { result.Bind(assembler->CallRuntime(fallback_runtime_function_id, context, object, key)); assembler->Goto(&end); } assembler->Bind(&end); return result.value(); } } // namespace // static compiler::Node* HasPropertyStub::Generate(CodeStubAssembler* assembler, compiler::Node* key, compiler::Node* object, compiler::Node* context) { return GenerateHasProperty(assembler, object, key, context, Runtime::kHasProperty); } // static compiler::Node* ForInFilterStub::Generate(CodeStubAssembler* assembler, compiler::Node* key, compiler::Node* object, compiler::Node* context) { typedef compiler::Node Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; Label return_undefined(assembler, Label::kDeferred), return_to_name(assembler), end(assembler); Variable var_result(assembler, MachineRepresentation::kTagged); Node* has_property = GenerateHasProperty(assembler, object, key, context, Runtime::kForInHasProperty); assembler->Branch( assembler->WordEqual(has_property, assembler->BooleanConstant(true)), &return_to_name, &return_undefined); assembler->Bind(&return_to_name); { var_result.Bind(assembler->ToName(context, key)); assembler->Goto(&end); } assembler->Bind(&return_undefined); { var_result.Bind(assembler->UndefinedConstant()); assembler->Goto(&end); } assembler->Bind(&end); return var_result.value(); } void GetPropertyStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef compiler::Node Node; typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; Label call_runtime(assembler, Label::kDeferred), return_undefined(assembler), end(assembler); Node* object = assembler->Parameter(0); Node* key = assembler->Parameter(1); Node* context = assembler->Parameter(2); Variable var_result(assembler, MachineRepresentation::kTagged); CodeStubAssembler::LookupInHolder lookup_property_in_holder = [assembler, context, &var_result, &end]( Node* receiver, Node* holder, Node* holder_map, Node* holder_instance_type, Node* unique_name, Label* next_holder, Label* if_bailout) { Variable var_value(assembler, MachineRepresentation::kTagged); Label if_found(assembler); assembler->TryGetOwnProperty( context, receiver, holder, holder_map, holder_instance_type, unique_name, &if_found, &var_value, next_holder, if_bailout); assembler->Bind(&if_found); { var_result.Bind(var_value.value()); assembler->Goto(&end); } }; CodeStubAssembler::LookupInHolder lookup_element_in_holder = [assembler, context, &var_result, &end]( Node* receiver, Node* holder, Node* holder_map, Node* holder_instance_type, Node* index, Label* next_holder, Label* if_bailout) { // Not supported yet. assembler->Use(next_holder); assembler->Goto(if_bailout); }; assembler->TryPrototypeChainLookup(object, key, lookup_property_in_holder, lookup_element_in_holder, &return_undefined, &call_runtime); assembler->Bind(&return_undefined); { var_result.Bind(assembler->UndefinedConstant()); assembler->Goto(&end); } assembler->Bind(&call_runtime); { var_result.Bind( assembler->CallRuntime(Runtime::kGetProperty, context, object, key)); assembler->Goto(&end); } assembler->Bind(&end); assembler->Return(var_result.value()); } // static compiler::Node* FastNewClosureStub::Generate(CodeStubAssembler* assembler, compiler::Node* shared_info, compiler::Node* context) { typedef compiler::Node Node; typedef compiler::CodeAssembler::Label Label; typedef compiler::CodeAssembler::Variable Variable; Isolate* isolate = assembler->isolate(); Factory* factory = assembler->isolate()->factory(); assembler->IncrementCounter(isolate->counters()->fast_new_closure_total(), 1); // Create a new closure from the given function info in new space Node* result = assembler->Allocate(JSFunction::kSize); // Calculate the index of the map we should install on the function based on // the FunctionKind and LanguageMode of the function. // Note: Must be kept in sync with Context::FunctionMapIndex Node* compiler_hints = assembler->LoadObjectField( shared_info, SharedFunctionInfo::kCompilerHintsOffset, MachineType::Uint32()); Node* is_strict = assembler->Word32And( compiler_hints, assembler->Int32Constant(1 << SharedFunctionInfo::kStrictModeBit)); Label if_normal(assembler), if_generator(assembler), if_async(assembler), if_class_constructor(assembler), if_function_without_prototype(assembler), load_map(assembler); Variable map_index(assembler, MachineType::PointerRepresentation()); Node* is_not_normal = assembler->Word32And( compiler_hints, assembler->Int32Constant(SharedFunctionInfo::kFunctionKindMaskBits)); assembler->GotoUnless(is_not_normal, &if_normal); Node* is_generator = assembler->Word32And( compiler_hints, assembler->Int32Constant(1 << SharedFunctionInfo::kIsGeneratorBit)); assembler->GotoIf(is_generator, &if_generator); Node* is_async = assembler->Word32And( compiler_hints, assembler->Int32Constant(1 << SharedFunctionInfo::kIsAsyncFunctionBit)); assembler->GotoIf(is_async, &if_async); Node* is_class_constructor = assembler->Word32And( compiler_hints, assembler->Int32Constant(SharedFunctionInfo::kClassConstructorBits)); assembler->GotoIf(is_class_constructor, &if_class_constructor); if (FLAG_debug_code) { // Function must be a function without a prototype. assembler->Assert(assembler->Word32And( compiler_hints, assembler->Int32Constant( SharedFunctionInfo::kAccessorFunctionBits | (1 << SharedFunctionInfo::kIsArrowBit) | (1 << SharedFunctionInfo::kIsConciseMethodBit)))); } assembler->Goto(&if_function_without_prototype); assembler->Bind(&if_normal); { map_index.Bind(assembler->Select( is_strict, assembler->IntPtrConstant(Context::STRICT_FUNCTION_MAP_INDEX), assembler->IntPtrConstant(Context::SLOPPY_FUNCTION_MAP_INDEX))); assembler->Goto(&load_map); } assembler->Bind(&if_generator); { map_index.Bind(assembler->Select( is_strict, assembler->IntPtrConstant(Context::STRICT_GENERATOR_FUNCTION_MAP_INDEX), assembler->IntPtrConstant( Context::SLOPPY_GENERATOR_FUNCTION_MAP_INDEX))); assembler->Goto(&load_map); } assembler->Bind(&if_async); { map_index.Bind(assembler->Select( is_strict, assembler->IntPtrConstant(Context::STRICT_ASYNC_FUNCTION_MAP_INDEX), assembler->IntPtrConstant(Context::SLOPPY_ASYNC_FUNCTION_MAP_INDEX))); assembler->Goto(&load_map); } assembler->Bind(&if_class_constructor); { map_index.Bind( assembler->IntPtrConstant(Context::STRICT_FUNCTION_MAP_INDEX)); assembler->Goto(&load_map); } assembler->Bind(&if_function_without_prototype); { map_index.Bind(assembler->IntPtrConstant( Context::STRICT_FUNCTION_WITHOUT_PROTOTYPE_MAP_INDEX)); assembler->Goto(&load_map); } assembler->Bind(&load_map); // Get the function map in the current native context and set that // as the map of the allocated object. Node* native_context = assembler->LoadNativeContext(context); Node* map_slot_value = assembler->LoadFixedArrayElement(native_context, map_index.value(), 0, CodeStubAssembler::INTPTR_PARAMETERS); assembler->StoreMapNoWriteBarrier(result, map_slot_value); // Initialize the rest of the function. Node* empty_fixed_array = assembler->HeapConstant(factory->empty_fixed_array()); Node* empty_literals_array = assembler->HeapConstant(factory->empty_literals_array()); assembler->StoreObjectFieldNoWriteBarrier(result, JSObject::kPropertiesOffset, empty_fixed_array); assembler->StoreObjectFieldNoWriteBarrier(result, JSObject::kElementsOffset, empty_fixed_array); assembler->StoreObjectFieldNoWriteBarrier(result, JSFunction::kLiteralsOffset, empty_literals_array); assembler->StoreObjectFieldNoWriteBarrier( result, JSFunction::kPrototypeOrInitialMapOffset, assembler->TheHoleConstant()); assembler->StoreObjectFieldNoWriteBarrier( result, JSFunction::kSharedFunctionInfoOffset, shared_info); assembler->StoreObjectFieldNoWriteBarrier(result, JSFunction::kContextOffset, context); Handle<Code> lazy_builtin_handle( assembler->isolate()->builtins()->builtin(Builtins::kCompileLazy)); Node* lazy_builtin = assembler->HeapConstant(lazy_builtin_handle); Node* lazy_builtin_entry = assembler->IntPtrAdd( lazy_builtin, assembler->IntPtrConstant(Code::kHeaderSize - kHeapObjectTag)); assembler->StoreObjectFieldNoWriteBarrier( result, JSFunction::kCodeEntryOffset, lazy_builtin_entry); assembler->StoreObjectFieldNoWriteBarrier(result, JSFunction::kNextFunctionLinkOffset, assembler->UndefinedConstant()); return result; } void FastNewClosureStub::GenerateAssembly(CodeStubAssembler* assembler) const { assembler->Return( Generate(assembler, assembler->Parameter(0), assembler->Parameter(1))); } // static compiler::Node* FastNewFunctionContextStub::Generate( CodeStubAssembler* assembler, compiler::Node* function, compiler::Node* slots, compiler::Node* context) { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; typedef CodeStubAssembler::Variable Variable; Node* min_context_slots = assembler->Int32Constant(Context::MIN_CONTEXT_SLOTS); Node* length = assembler->Int32Add(slots, min_context_slots); Node* size = assembler->Int32Add( assembler->Word32Shl(length, assembler->Int32Constant(kPointerSizeLog2)), assembler->Int32Constant(FixedArray::kHeaderSize)); // Create a new closure from the given function info in new space Node* function_context = assembler->Allocate(size); Isolate* isolate = assembler->isolate(); assembler->StoreMapNoWriteBarrier( function_context, assembler->HeapConstant(isolate->factory()->function_context_map())); assembler->StoreObjectFieldNoWriteBarrier(function_context, Context::kLengthOffset, assembler->SmiFromWord32(length)); // Set up the fixed slots. assembler->StoreFixedArrayElement( function_context, assembler->Int32Constant(Context::CLOSURE_INDEX), function, SKIP_WRITE_BARRIER); assembler->StoreFixedArrayElement( function_context, assembler->Int32Constant(Context::PREVIOUS_INDEX), context, SKIP_WRITE_BARRIER); assembler->StoreFixedArrayElement( function_context, assembler->Int32Constant(Context::EXTENSION_INDEX), assembler->TheHoleConstant(), SKIP_WRITE_BARRIER); // Copy the native context from the previous context. Node* native_context = assembler->LoadNativeContext(context); assembler->StoreFixedArrayElement( function_context, assembler->Int32Constant(Context::NATIVE_CONTEXT_INDEX), native_context, SKIP_WRITE_BARRIER); // Initialize the rest of the slots to undefined. Node* undefined = assembler->UndefinedConstant(); Variable var_slot_index(assembler, MachineRepresentation::kWord32); var_slot_index.Bind(min_context_slots); Label loop(assembler, &var_slot_index), after_loop(assembler); assembler->Goto(&loop); assembler->Bind(&loop); { Node* slot_index = var_slot_index.value(); assembler->GotoUnless(assembler->Int32LessThan(slot_index, length), &after_loop); assembler->StoreFixedArrayElement(function_context, slot_index, undefined, SKIP_WRITE_BARRIER); Node* one = assembler->Int32Constant(1); Node* next_index = assembler->Int32Add(slot_index, one); var_slot_index.Bind(next_index); assembler->Goto(&loop); } assembler->Bind(&after_loop); return function_context; } void FastNewFunctionContextStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* function = assembler->Parameter(Descriptor::kFunction); Node* slots = assembler->Parameter(FastNewFunctionContextDescriptor::kSlots); Node* context = assembler->Parameter(Descriptor::kContext); assembler->Return(Generate(assembler, function, slots, context)); } // static compiler::Node* FastCloneRegExpStub::Generate(CodeStubAssembler* assembler, compiler::Node* closure, compiler::Node* literal_index, compiler::Node* pattern, compiler::Node* flags, compiler::Node* context) { typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; typedef compiler::Node Node; Label call_runtime(assembler, Label::kDeferred), end(assembler); Variable result(assembler, MachineRepresentation::kTagged); Node* undefined = assembler->UndefinedConstant(); Node* literals_array = assembler->LoadObjectField(closure, JSFunction::kLiteralsOffset); Node* boilerplate = assembler->LoadFixedArrayElement( literals_array, literal_index, LiteralsArray::kFirstLiteralIndex * kPointerSize, CodeStubAssembler::SMI_PARAMETERS); assembler->GotoIf(assembler->WordEqual(boilerplate, undefined), &call_runtime); { int size = JSRegExp::kSize + JSRegExp::kInObjectFieldCount * kPointerSize; Node* copy = assembler->Allocate(size); for (int offset = 0; offset < size; offset += kPointerSize) { Node* value = assembler->LoadObjectField(boilerplate, offset); assembler->StoreObjectFieldNoWriteBarrier(copy, offset, value); } result.Bind(copy); assembler->Goto(&end); } assembler->Bind(&call_runtime); { result.Bind(assembler->CallRuntime(Runtime::kCreateRegExpLiteral, context, closure, literal_index, pattern, flags)); assembler->Goto(&end); } assembler->Bind(&end); return result.value(); } void FastCloneRegExpStub::GenerateAssembly(CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* closure = assembler->Parameter(Descriptor::kClosure); Node* literal_index = assembler->Parameter(Descriptor::kLiteralIndex); Node* pattern = assembler->Parameter(Descriptor::kPattern); Node* flags = assembler->Parameter(Descriptor::kFlags); Node* context = assembler->Parameter(Descriptor::kContext); assembler->Return( Generate(assembler, closure, literal_index, pattern, flags, context)); } namespace { compiler::Node* NonEmptyShallowClone(CodeStubAssembler* assembler, compiler::Node* boilerplate, compiler::Node* boilerplate_map, compiler::Node* boilerplate_elements, compiler::Node* allocation_site, compiler::Node* capacity, ElementsKind kind) { typedef compiler::Node Node; typedef CodeStubAssembler::ParameterMode ParameterMode; ParameterMode param_mode = CodeStubAssembler::SMI_PARAMETERS; Node* length = assembler->LoadJSArrayLength(boilerplate); if (assembler->Is64()) { capacity = assembler->SmiUntag(capacity); param_mode = CodeStubAssembler::INTEGER_PARAMETERS; } Node *array, *elements; std::tie(array, elements) = assembler->AllocateUninitializedJSArrayWithElements( kind, boilerplate_map, length, allocation_site, capacity, param_mode); assembler->Comment("copy elements header"); for (int offset = 0; offset < FixedArrayBase::kHeaderSize; offset += kPointerSize) { Node* value = assembler->LoadObjectField(boilerplate_elements, offset); assembler->StoreObjectField(elements, offset, value); } if (assembler->Is64()) { length = assembler->SmiUntag(length); } assembler->Comment("copy boilerplate elements"); assembler->CopyFixedArrayElements(kind, boilerplate_elements, elements, length, SKIP_WRITE_BARRIER, param_mode); assembler->IncrementCounter( assembler->isolate()->counters()->inlined_copied_elements(), 1); return array; } } // namespace // static compiler::Node* FastCloneShallowArrayStub::Generate( CodeStubAssembler* assembler, compiler::Node* closure, compiler::Node* literal_index, compiler::Node* context, CodeStubAssembler::Label* call_runtime, AllocationSiteMode allocation_site_mode) { typedef CodeStubAssembler::Label Label; typedef CodeStubAssembler::Variable Variable; typedef compiler::Node Node; Label zero_capacity(assembler), cow_elements(assembler), fast_elements(assembler), return_result(assembler); Variable result(assembler, MachineRepresentation::kTagged); Node* literals_array = assembler->LoadObjectField(closure, JSFunction::kLiteralsOffset); Node* allocation_site = assembler->LoadFixedArrayElement( literals_array, literal_index, LiteralsArray::kFirstLiteralIndex * kPointerSize, CodeStubAssembler::SMI_PARAMETERS); Node* undefined = assembler->UndefinedConstant(); assembler->GotoIf(assembler->WordEqual(allocation_site, undefined), call_runtime); allocation_site = assembler->LoadFixedArrayElement( literals_array, literal_index, LiteralsArray::kFirstLiteralIndex * kPointerSize, CodeStubAssembler::SMI_PARAMETERS); Node* boilerplate = assembler->LoadObjectField( allocation_site, AllocationSite::kTransitionInfoOffset); Node* boilerplate_map = assembler->LoadMap(boilerplate); Node* boilerplate_elements = assembler->LoadElements(boilerplate); Node* capacity = assembler->LoadFixedArrayBaseLength(boilerplate_elements); allocation_site = allocation_site_mode == TRACK_ALLOCATION_SITE ? allocation_site : nullptr; Node* zero = assembler->SmiConstant(Smi::FromInt(0)); assembler->GotoIf(assembler->SmiEqual(capacity, zero), &zero_capacity); Node* elements_map = assembler->LoadMap(boilerplate_elements); assembler->GotoIf( assembler->WordEqual(elements_map, assembler->FixedCowArrayMapConstant()), &cow_elements); assembler->GotoIf( assembler->WordEqual(elements_map, assembler->FixedArrayMapConstant()), &fast_elements); { assembler->Comment("fast double elements path"); if (FLAG_debug_code) { Label correct_elements_map(assembler), abort(assembler, Label::kDeferred); assembler->BranchIf( assembler->WordEqual(elements_map, assembler->FixedDoubleArrayMapConstant()), &correct_elements_map, &abort); assembler->Bind(&abort); { Node* abort_id = assembler->SmiConstant( Smi::FromInt(BailoutReason::kExpectedFixedDoubleArrayMap)); assembler->TailCallRuntime(Runtime::kAbort, context, abort_id); } assembler->Bind(&correct_elements_map); } Node* array = NonEmptyShallowClone(assembler, boilerplate, boilerplate_map, boilerplate_elements, allocation_site, capacity, FAST_DOUBLE_ELEMENTS); result.Bind(array); assembler->Goto(&return_result); } assembler->Bind(&fast_elements); { assembler->Comment("fast elements path"); Node* array = NonEmptyShallowClone(assembler, boilerplate, boilerplate_map, boilerplate_elements, allocation_site, capacity, FAST_ELEMENTS); result.Bind(array); assembler->Goto(&return_result); } Variable length(assembler, MachineRepresentation::kTagged), elements(assembler, MachineRepresentation::kTagged); Label allocate_without_elements(assembler); assembler->Bind(&cow_elements); { assembler->Comment("fixed cow path"); length.Bind(assembler->LoadJSArrayLength(boilerplate)); elements.Bind(boilerplate_elements); assembler->Goto(&allocate_without_elements); } assembler->Bind(&zero_capacity); { assembler->Comment("zero capacity path"); length.Bind(zero); elements.Bind(assembler->LoadRoot(Heap::kEmptyFixedArrayRootIndex)); assembler->Goto(&allocate_without_elements); } assembler->Bind(&allocate_without_elements); { Node* array = assembler->AllocateUninitializedJSArrayWithoutElements( FAST_ELEMENTS, boilerplate_map, length.value(), allocation_site); assembler->StoreObjectField(array, JSObject::kElementsOffset, elements.value()); result.Bind(array); assembler->Goto(&return_result); } assembler->Bind(&return_result); return result.value(); } void FastCloneShallowArrayStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; typedef CodeStubAssembler::Label Label; Node* closure = assembler->Parameter(Descriptor::kClosure); Node* literal_index = assembler->Parameter(Descriptor::kLiteralIndex); Node* constant_elements = assembler->Parameter(Descriptor::kConstantElements); Node* context = assembler->Parameter(Descriptor::kContext); Label call_runtime(assembler, Label::kDeferred); assembler->Return(Generate(assembler, closure, literal_index, context, &call_runtime, allocation_site_mode())); assembler->Bind(&call_runtime); { assembler->Comment("call runtime"); Node* flags = assembler->SmiConstant( Smi::FromInt(ArrayLiteral::kShallowElements | (allocation_site_mode() == TRACK_ALLOCATION_SITE ? 0 : ArrayLiteral::kDisableMementos))); assembler->Return(assembler->CallRuntime(Runtime::kCreateArrayLiteral, context, closure, literal_index, constant_elements, flags)); } } void CreateAllocationSiteStub::GenerateAheadOfTime(Isolate* isolate) { CreateAllocationSiteStub stub(isolate); stub.GetCode(); } void CreateWeakCellStub::GenerateAheadOfTime(Isolate* isolate) { CreateWeakCellStub stub(isolate); stub.GetCode(); } void StoreElementStub::Generate(MacroAssembler* masm) { DCHECK_EQ(DICTIONARY_ELEMENTS, elements_kind()); KeyedStoreIC::GenerateSlow(masm); } void StoreFastElementStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef CodeStubAssembler::Label Label; typedef compiler::Node Node; assembler->Comment( "StoreFastElementStub: js_array=%d, elements_kind=%s, store_mode=%d", is_js_array(), ElementsKindToString(elements_kind()), store_mode()); Node* receiver = assembler->Parameter(Descriptor::kReceiver); Node* key = assembler->Parameter(Descriptor::kName); Node* value = assembler->Parameter(Descriptor::kValue); Node* slot = assembler->Parameter(Descriptor::kSlot); Node* vector = assembler->Parameter(Descriptor::kVector); Node* context = assembler->Parameter(Descriptor::kContext); Label miss(assembler); assembler->EmitElementStore(receiver, key, value, is_js_array(), elements_kind(), store_mode(), &miss); assembler->Return(value); assembler->Bind(&miss); { assembler->Comment("Miss"); assembler->TailCallRuntime(Runtime::kKeyedStoreIC_Miss, context, value, slot, vector, receiver, key); } } // static void StoreFastElementStub::GenerateAheadOfTime(Isolate* isolate) { if (FLAG_minimal) return; StoreFastElementStub(isolate, false, FAST_HOLEY_ELEMENTS, STANDARD_STORE) .GetCode(); StoreFastElementStub(isolate, false, FAST_HOLEY_ELEMENTS, STORE_AND_GROW_NO_TRANSITION).GetCode(); for (int i = FIRST_FAST_ELEMENTS_KIND; i <= LAST_FAST_ELEMENTS_KIND; i++) { ElementsKind kind = static_cast<ElementsKind>(i); StoreFastElementStub(isolate, true, kind, STANDARD_STORE).GetCode(); StoreFastElementStub(isolate, true, kind, STORE_AND_GROW_NO_TRANSITION) .GetCode(); } } void ArrayConstructorStub::PrintName(std::ostream& os) const { // NOLINT os << "ArrayConstructorStub"; switch (argument_count()) { case ANY: os << "_Any"; break; case NONE: os << "_None"; break; case ONE: os << "_One"; break; case MORE_THAN_ONE: os << "_More_Than_One"; break; } return; } bool ToBooleanICStub::UpdateStatus(Handle<Object> object) { Types new_types = types(); Types old_types = new_types; bool to_boolean_value = new_types.UpdateStatus(isolate(), object); TraceTransition(old_types, new_types); set_sub_minor_key(TypesBits::update(sub_minor_key(), new_types.ToIntegral())); return to_boolean_value; } void ToBooleanICStub::PrintState(std::ostream& os) const { // NOLINT os << types(); } std::ostream& operator<<(std::ostream& os, const ToBooleanICStub::Types& s) { os << "("; SimpleListPrinter p(os); if (s.IsEmpty()) p.Add("None"); if (s.Contains(ToBooleanICStub::UNDEFINED)) p.Add("Undefined"); if (s.Contains(ToBooleanICStub::BOOLEAN)) p.Add("Bool"); if (s.Contains(ToBooleanICStub::NULL_TYPE)) p.Add("Null"); if (s.Contains(ToBooleanICStub::SMI)) p.Add("Smi"); if (s.Contains(ToBooleanICStub::SPEC_OBJECT)) p.Add("SpecObject"); if (s.Contains(ToBooleanICStub::STRING)) p.Add("String"); if (s.Contains(ToBooleanICStub::SYMBOL)) p.Add("Symbol"); if (s.Contains(ToBooleanICStub::HEAP_NUMBER)) p.Add("HeapNumber"); if (s.Contains(ToBooleanICStub::SIMD_VALUE)) p.Add("SimdValue"); return os << ")"; } bool ToBooleanICStub::Types::UpdateStatus(Isolate* isolate, Handle<Object> object) { if (object->IsUndefined(isolate)) { Add(UNDEFINED); return false; } else if (object->IsBoolean()) { Add(BOOLEAN); return object->IsTrue(isolate); } else if (object->IsNull(isolate)) { Add(NULL_TYPE); return false; } else if (object->IsSmi()) { Add(SMI); return Smi::cast(*object)->value() != 0; } else if (object->IsJSReceiver()) { Add(SPEC_OBJECT); return !object->IsUndetectable(); } else if (object->IsString()) { DCHECK(!object->IsUndetectable()); Add(STRING); return String::cast(*object)->length() != 0; } else if (object->IsSymbol()) { Add(SYMBOL); return true; } else if (object->IsHeapNumber()) { DCHECK(!object->IsUndetectable()); Add(HEAP_NUMBER); double value = HeapNumber::cast(*object)->value(); return value != 0 && !std::isnan(value); } else if (object->IsSimd128Value()) { Add(SIMD_VALUE); return true; } else { // We should never see an internal object at runtime here! UNREACHABLE(); return true; } } bool ToBooleanICStub::Types::NeedsMap() const { return Contains(ToBooleanICStub::SPEC_OBJECT) || Contains(ToBooleanICStub::STRING) || Contains(ToBooleanICStub::SYMBOL) || Contains(ToBooleanICStub::HEAP_NUMBER) || Contains(ToBooleanICStub::SIMD_VALUE); } void StubFailureTrampolineStub::GenerateAheadOfTime(Isolate* isolate) { StubFailureTrampolineStub stub1(isolate, NOT_JS_FUNCTION_STUB_MODE); StubFailureTrampolineStub stub2(isolate, JS_FUNCTION_STUB_MODE); stub1.GetCode(); stub2.GetCode(); } void ProfileEntryHookStub::EntryHookTrampoline(intptr_t function, intptr_t stack_pointer, Isolate* isolate) { FunctionEntryHook entry_hook = isolate->function_entry_hook(); DCHECK(entry_hook != NULL); entry_hook(function, stack_pointer); } void CreateAllocationSiteStub::GenerateAssembly( CodeStubAssembler* assembler) const { assembler->Return(assembler->CreateAllocationSiteInFeedbackVector( assembler->Parameter(Descriptor::kVector), assembler->Parameter(Descriptor::kSlot))); } void CreateWeakCellStub::GenerateAssembly(CodeStubAssembler* assembler) const { assembler->Return(assembler->CreateWeakCellInFeedbackVector( assembler->Parameter(Descriptor::kVector), assembler->Parameter(Descriptor::kSlot), assembler->Parameter(Descriptor::kValue))); } void ArrayNoArgumentConstructorStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* native_context = assembler->LoadObjectField( assembler->Parameter(Descriptor::kFunction), JSFunction::kContextOffset); bool track_allocation_site = AllocationSite::GetMode(elements_kind()) == TRACK_ALLOCATION_SITE && override_mode() != DISABLE_ALLOCATION_SITES; Node* allocation_site = track_allocation_site ? assembler->Parameter(Descriptor::kAllocationSite) : nullptr; Node* array_map = assembler->LoadJSArrayElementsMap(elements_kind(), native_context); Node* array = assembler->AllocateJSArray( elements_kind(), array_map, assembler->IntPtrConstant(JSArray::kPreallocatedArrayElements), assembler->SmiConstant(Smi::FromInt(0)), allocation_site); assembler->Return(array); } void InternalArrayNoArgumentConstructorStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* array_map = assembler->LoadObjectField(assembler->Parameter(Descriptor::kFunction), JSFunction::kPrototypeOrInitialMapOffset); Node* array = assembler->AllocateJSArray( elements_kind(), array_map, assembler->IntPtrConstant(JSArray::kPreallocatedArrayElements), assembler->SmiConstant(Smi::FromInt(0)), nullptr); assembler->Return(array); } namespace { template <typename Descriptor> void SingleArgumentConstructorCommon(CodeStubAssembler* assembler, ElementsKind elements_kind, compiler::Node* array_map, compiler::Node* allocation_site, AllocationSiteMode mode) { typedef compiler::Node Node; typedef CodeStubAssembler::Label Label; Label ok(assembler); Label smi_size(assembler); Label small_smi_size(assembler); Label call_runtime(assembler, Label::kDeferred); Node* size = assembler->Parameter(Descriptor::kArraySizeSmiParameter); assembler->Branch(assembler->WordIsSmi(size), &smi_size, &call_runtime); assembler->Bind(&smi_size); if (IsFastPackedElementsKind(elements_kind)) { Label abort(assembler, Label::kDeferred); assembler->Branch( assembler->SmiEqual(size, assembler->SmiConstant(Smi::FromInt(0))), &small_smi_size, &abort); assembler->Bind(&abort); Node* reason = assembler->SmiConstant(Smi::FromInt(kAllocatingNonEmptyPackedArray)); Node* context = assembler->Parameter(Descriptor::kContext); assembler->TailCallRuntime(Runtime::kAbort, context, reason); } else { int element_size = IsFastDoubleElementsKind(elements_kind) ? kDoubleSize : kPointerSize; int max_fast_elements = (kMaxRegularHeapObjectSize - FixedArray::kHeaderSize - JSArray::kSize - AllocationMemento::kSize) / element_size; assembler->Branch( assembler->SmiAboveOrEqual( size, assembler->SmiConstant(Smi::FromInt(max_fast_elements))), &call_runtime, &small_smi_size); } assembler->Bind(&small_smi_size); { Node* array = assembler->AllocateJSArray( elements_kind, array_map, size, size, mode == DONT_TRACK_ALLOCATION_SITE ? nullptr : allocation_site, CodeStubAssembler::SMI_PARAMETERS); assembler->Return(array); } assembler->Bind(&call_runtime); { Node* context = assembler->Parameter(Descriptor::kContext); Node* function = assembler->Parameter(Descriptor::kFunction); Node* array_size = assembler->Parameter(Descriptor::kArraySizeSmiParameter); Node* allocation_site = assembler->Parameter(Descriptor::kAllocationSite); assembler->TailCallRuntime(Runtime::kNewArray, context, function, array_size, function, allocation_site); } } } // namespace void ArraySingleArgumentConstructorStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* function = assembler->Parameter(Descriptor::kFunction); Node* native_context = assembler->LoadObjectField(function, JSFunction::kContextOffset); Node* array_map = assembler->LoadJSArrayElementsMap(elements_kind(), native_context); AllocationSiteMode mode = override_mode() == DISABLE_ALLOCATION_SITES ? DONT_TRACK_ALLOCATION_SITE : AllocationSite::GetMode(elements_kind()); Node* allocation_site = assembler->Parameter(Descriptor::kAllocationSite); SingleArgumentConstructorCommon<Descriptor>(assembler, elements_kind(), array_map, allocation_site, mode); } void InternalArraySingleArgumentConstructorStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; Node* function = assembler->Parameter(Descriptor::kFunction); Node* array_map = assembler->LoadObjectField( function, JSFunction::kPrototypeOrInitialMapOffset); SingleArgumentConstructorCommon<Descriptor>( assembler, elements_kind(), array_map, assembler->UndefinedConstant(), DONT_TRACK_ALLOCATION_SITE); } void GrowArrayElementsStub::GenerateAssembly( CodeStubAssembler* assembler) const { typedef compiler::Node Node; CodeStubAssembler::Label runtime(assembler, CodeStubAssembler::Label::kDeferred); Node* object = assembler->Parameter(Descriptor::kObject); Node* key = assembler->Parameter(Descriptor::kKey); Node* context = assembler->Parameter(Descriptor::kContext); ElementsKind kind = elements_kind(); Node* elements = assembler->LoadElements(object); Node* new_elements = assembler->TryGrowElementsCapacity(object, elements, kind, key, &runtime); assembler->Return(new_elements); assembler->Bind(&runtime); // TODO(danno): Make this a tail call when the stub is only used from TurboFan // code. This musn't be a tail call for now, since the caller site in lithium // creates a safepoint. This safepoint musn't have a different number of // arguments on the stack in the case that a GC happens from the slow-case // allocation path (zero, since all the stubs inputs are in registers) and // when the call happens (it would be two in the tail call case due to the // tail call pushing the arguments on the stack for the runtime call). By not // tail-calling, the runtime call case also has zero arguments on the stack // for the stub frame. assembler->Return(assembler->CallRuntime(Runtime::kGrowArrayElements, context, object, key)); } ArrayConstructorStub::ArrayConstructorStub(Isolate* isolate) : PlatformCodeStub(isolate) { minor_key_ = ArgumentCountBits::encode(ANY); } ArrayConstructorStub::ArrayConstructorStub(Isolate* isolate, int argument_count) : PlatformCodeStub(isolate) { if (argument_count == 0) { minor_key_ = ArgumentCountBits::encode(NONE); } else if (argument_count == 1) { minor_key_ = ArgumentCountBits::encode(ONE); } else if (argument_count >= 2) { minor_key_ = ArgumentCountBits::encode(MORE_THAN_ONE); } else { UNREACHABLE(); } } InternalArrayConstructorStub::InternalArrayConstructorStub(Isolate* isolate) : PlatformCodeStub(isolate) {} Representation RepresentationFromMachineType(MachineType type) { if (type == MachineType::Int32()) { return Representation::Integer32(); } if (type == MachineType::TaggedSigned()) { return Representation::Smi(); } if (type == MachineType::Pointer()) { return Representation::External(); } return Representation::Tagged(); } } // namespace internal } // namespace v8