| // Copyright 2012 the V8 project authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #ifndef V8_CRANKSHAFT_MIPS_LITHIUM_CODEGEN_MIPS_H_ |
| #define V8_CRANKSHAFT_MIPS_LITHIUM_CODEGEN_MIPS_H_ |
| |
| #include "src/ast/scopes.h" |
| #include "src/crankshaft/lithium-codegen.h" |
| #include "src/crankshaft/mips/lithium-gap-resolver-mips.h" |
| #include "src/crankshaft/mips/lithium-mips.h" |
| #include "src/deoptimizer.h" |
| #include "src/safepoint-table.h" |
| #include "src/utils.h" |
| |
| namespace v8 { |
| namespace internal { |
| |
| // Forward declarations. |
| class LDeferredCode; |
| class SafepointGenerator; |
| |
| class LCodeGen: public LCodeGenBase { |
| public: |
| LCodeGen(LChunk* chunk, MacroAssembler* assembler, CompilationInfo* info) |
| : LCodeGenBase(chunk, assembler, info), |
| jump_table_(4, info->zone()), |
| scope_(info->scope()), |
| deferred_(8, info->zone()), |
| frame_is_built_(false), |
| safepoints_(info->zone()), |
| resolver_(this), |
| expected_safepoint_kind_(Safepoint::kSimple) { |
| PopulateDeoptimizationLiteralsWithInlinedFunctions(); |
| } |
| |
| |
| int LookupDestination(int block_id) const { |
| return chunk()->LookupDestination(block_id); |
| } |
| |
| bool IsNextEmittedBlock(int block_id) const { |
| return LookupDestination(block_id) == GetNextEmittedBlock(); |
| } |
| |
| bool NeedsEagerFrame() const { |
| return HasAllocatedStackSlots() || info()->is_non_deferred_calling() || |
| !info()->IsStub() || info()->requires_frame(); |
| } |
| bool NeedsDeferredFrame() const { |
| return !NeedsEagerFrame() && info()->is_deferred_calling(); |
| } |
| |
| RAStatus GetRAState() const { |
| return frame_is_built_ ? kRAHasBeenSaved : kRAHasNotBeenSaved; |
| } |
| |
| // Support for converting LOperands to assembler types. |
| // LOperand must be a register. |
| Register ToRegister(LOperand* op) const; |
| |
| // LOperand is loaded into scratch, unless already a register. |
| Register EmitLoadRegister(LOperand* op, Register scratch); |
| |
| // LOperand must be a double register. |
| DoubleRegister ToDoubleRegister(LOperand* op) const; |
| |
| // LOperand is loaded into dbl_scratch, unless already a double register. |
| DoubleRegister EmitLoadDoubleRegister(LOperand* op, |
| FloatRegister flt_scratch, |
| DoubleRegister dbl_scratch); |
| int32_t ToRepresentation(LConstantOperand* op, const Representation& r) const; |
| int32_t ToInteger32(LConstantOperand* op) const; |
| Smi* ToSmi(LConstantOperand* op) const; |
| double ToDouble(LConstantOperand* op) const; |
| Operand ToOperand(LOperand* op); |
| MemOperand ToMemOperand(LOperand* op) const; |
| // Returns a MemOperand pointing to the high word of a DoubleStackSlot. |
| MemOperand ToHighMemOperand(LOperand* op) const; |
| |
| bool IsInteger32(LConstantOperand* op) const; |
| bool IsSmi(LConstantOperand* op) const; |
| Handle<Object> ToHandle(LConstantOperand* op) const; |
| |
| // Try to generate code for the entire chunk, but it may fail if the |
| // chunk contains constructs we cannot handle. Returns true if the |
| // code generation attempt succeeded. |
| bool GenerateCode(); |
| |
| // Finish the code by setting stack height, safepoint, and bailout |
| // information on it. |
| void FinishCode(Handle<Code> code); |
| |
| void DoDeferredNumberTagD(LNumberTagD* instr); |
| |
| enum IntegerSignedness { SIGNED_INT32, UNSIGNED_INT32 }; |
| void DoDeferredNumberTagIU(LInstruction* instr, |
| LOperand* value, |
| LOperand* temp1, |
| LOperand* temp2, |
| IntegerSignedness signedness); |
| |
| void DoDeferredTaggedToI(LTaggedToI* instr); |
| void DoDeferredMathAbsTaggedHeapNumber(LMathAbs* instr); |
| void DoDeferredStackCheck(LStackCheck* instr); |
| void DoDeferredMaybeGrowElements(LMaybeGrowElements* instr); |
| void DoDeferredStringCharCodeAt(LStringCharCodeAt* instr); |
| void DoDeferredStringCharFromCode(LStringCharFromCode* instr); |
| void DoDeferredAllocate(LAllocate* instr); |
| void DoDeferredInstanceMigration(LCheckMaps* instr, Register object); |
| void DoDeferredLoadMutableDouble(LLoadFieldByIndex* instr, |
| Register result, |
| Register object, |
| Register index); |
| |
| // Parallel move support. |
| void DoParallelMove(LParallelMove* move); |
| void DoGap(LGap* instr); |
| |
| MemOperand PrepareKeyedOperand(Register key, |
| Register base, |
| bool key_is_constant, |
| int constant_key, |
| int element_size, |
| int shift_size, |
| int base_offset); |
| |
| // Emit frame translation commands for an environment. |
| void WriteTranslation(LEnvironment* environment, Translation* translation); |
| |
| // Declare methods that deal with the individual node types. |
| #define DECLARE_DO(type) void Do##type(L##type* node); |
| LITHIUM_CONCRETE_INSTRUCTION_LIST(DECLARE_DO) |
| #undef DECLARE_DO |
| |
| private: |
| Scope* scope() const { return scope_; } |
| |
| Register scratch0() { return kLithiumScratchReg; } |
| Register scratch1() { return kLithiumScratchReg2; } |
| DoubleRegister double_scratch0() { return kLithiumScratchDouble; } |
| |
| LInstruction* GetNextInstruction(); |
| |
| void EmitClassOfTest(Label* if_true, |
| Label* if_false, |
| Handle<String> class_name, |
| Register input, |
| Register temporary, |
| Register temporary2); |
| |
| bool HasAllocatedStackSlots() const { |
| return chunk()->HasAllocatedStackSlots(); |
| } |
| int GetStackSlotCount() const { return chunk()->GetSpillSlotCount(); } |
| int GetTotalFrameSlotCount() const { |
| return chunk()->GetTotalFrameSlotCount(); |
| } |
| |
| void AddDeferredCode(LDeferredCode* code) { deferred_.Add(code, zone()); } |
| |
| void SaveCallerDoubles(); |
| void RestoreCallerDoubles(); |
| |
| // Code generation passes. Returns true if code generation should |
| // continue. |
| void GenerateBodyInstructionPre(LInstruction* instr) override; |
| bool GeneratePrologue(); |
| bool GenerateDeferredCode(); |
| bool GenerateJumpTable(); |
| bool GenerateSafepointTable(); |
| |
| // Generates the custom OSR entrypoint and sets the osr_pc_offset. |
| void GenerateOsrPrologue(); |
| |
| enum SafepointMode { |
| RECORD_SIMPLE_SAFEPOINT, |
| RECORD_SAFEPOINT_WITH_REGISTERS_AND_NO_ARGUMENTS |
| }; |
| |
| void CallCode(Handle<Code> code, |
| RelocInfo::Mode mode, |
| LInstruction* instr); |
| |
| void CallCodeGeneric(Handle<Code> code, |
| RelocInfo::Mode mode, |
| LInstruction* instr, |
| SafepointMode safepoint_mode); |
| |
| void CallRuntime(const Runtime::Function* function, |
| int num_arguments, |
| LInstruction* instr, |
| SaveFPRegsMode save_doubles = kDontSaveFPRegs); |
| |
| void CallRuntime(Runtime::FunctionId id, |
| int num_arguments, |
| LInstruction* instr) { |
| const Runtime::Function* function = Runtime::FunctionForId(id); |
| CallRuntime(function, num_arguments, instr); |
| } |
| |
| void CallRuntime(Runtime::FunctionId id, LInstruction* instr) { |
| const Runtime::Function* function = Runtime::FunctionForId(id); |
| CallRuntime(function, function->nargs, instr); |
| } |
| |
| void LoadContextFromDeferred(LOperand* context); |
| void CallRuntimeFromDeferred(Runtime::FunctionId id, |
| int argc, |
| LInstruction* instr, |
| LOperand* context); |
| |
| void PrepareForTailCall(const ParameterCount& actual, Register scratch1, |
| Register scratch2, Register scratch3); |
| |
| // Generate a direct call to a known function. Expects the function |
| // to be in a1. |
| void CallKnownFunction(Handle<JSFunction> function, |
| int formal_parameter_count, int arity, |
| bool is_tail_call, LInstruction* instr); |
| |
| void RecordSafepointWithLazyDeopt(LInstruction* instr, |
| SafepointMode safepoint_mode); |
| |
| void RegisterEnvironmentForDeoptimization(LEnvironment* environment, |
| Safepoint::DeoptMode mode); |
| void DeoptimizeIf(Condition condition, LInstruction* instr, |
| DeoptimizeReason deopt_reason, |
| Deoptimizer::BailoutType bailout_type, |
| Register src1 = zero_reg, |
| const Operand& src2 = Operand(zero_reg)); |
| void DeoptimizeIf(Condition condition, LInstruction* instr, |
| DeoptimizeReason deopt_reason = DeoptimizeReason::kNoReason, |
| Register src1 = zero_reg, |
| const Operand& src2 = Operand(zero_reg)); |
| |
| void AddToTranslation(LEnvironment* environment, |
| Translation* translation, |
| LOperand* op, |
| bool is_tagged, |
| bool is_uint32, |
| int* object_index_pointer, |
| int* dematerialized_index_pointer); |
| |
| Register ToRegister(int index) const; |
| DoubleRegister ToDoubleRegister(int index) const; |
| |
| MemOperand BuildSeqStringOperand(Register string, |
| LOperand* index, |
| String::Encoding encoding); |
| |
| void EmitIntegerMathAbs(LMathAbs* instr); |
| |
| // Support for recording safepoint information. |
| void RecordSafepoint(LPointerMap* pointers, |
| Safepoint::Kind kind, |
| int arguments, |
| Safepoint::DeoptMode mode); |
| void RecordSafepoint(LPointerMap* pointers, Safepoint::DeoptMode mode); |
| void RecordSafepoint(Safepoint::DeoptMode mode); |
| void RecordSafepointWithRegisters(LPointerMap* pointers, |
| int arguments, |
| Safepoint::DeoptMode mode); |
| |
| static Condition TokenToCondition(Token::Value op, bool is_unsigned); |
| void EmitGoto(int block); |
| |
| // EmitBranch expects to be the last instruction of a block. |
| template<class InstrType> |
| void EmitBranch(InstrType instr, |
| Condition condition, |
| Register src1, |
| const Operand& src2); |
| template<class InstrType> |
| void EmitBranchF(InstrType instr, |
| Condition condition, |
| FPURegister src1, |
| FPURegister src2); |
| template <class InstrType> |
| void EmitTrueBranch(InstrType instr, Condition condition, Register src1, |
| const Operand& src2); |
| template <class InstrType> |
| void EmitFalseBranch(InstrType instr, Condition condition, Register src1, |
| const Operand& src2); |
| template<class InstrType> |
| void EmitFalseBranchF(InstrType instr, |
| Condition condition, |
| FPURegister src1, |
| FPURegister src2); |
| void EmitCmpI(LOperand* left, LOperand* right); |
| void EmitNumberUntagD(LNumberUntagD* instr, Register input, |
| DoubleRegister result, NumberUntagDMode mode); |
| |
| // Emits optimized code for typeof x == "y". Modifies input register. |
| // Returns the condition on which a final split to |
| // true and false label should be made, to optimize fallthrough. |
| // Returns two registers in cmp1 and cmp2 that can be used in the |
| // Branch instruction after EmitTypeofIs. |
| Condition EmitTypeofIs(Label* true_label, |
| Label* false_label, |
| Register input, |
| Handle<String> type_name, |
| Register* cmp1, |
| Operand* cmp2); |
| |
| // Emits optimized code for %_IsString(x). Preserves input register. |
| // Returns the condition on which a final split to |
| // true and false label should be made, to optimize fallthrough. |
| Condition EmitIsString(Register input, |
| Register temp1, |
| Label* is_not_string, |
| SmiCheck check_needed); |
| |
| // Emits optimized code to deep-copy the contents of statically known |
| // object graphs (e.g. object literal boilerplate). |
| void EmitDeepCopy(Handle<JSObject> object, |
| Register result, |
| Register source, |
| int* offset, |
| AllocationSiteMode mode); |
| // Emit optimized code for integer division. |
| // Inputs are signed. |
| // All registers are clobbered. |
| // If 'remainder' is no_reg, it is not computed. |
| void EmitSignedIntegerDivisionByConstant(Register result, |
| Register dividend, |
| int32_t divisor, |
| Register remainder, |
| Register scratch, |
| LEnvironment* environment); |
| |
| |
| void EnsureSpaceForLazyDeopt(int space_needed) override; |
| void DoLoadKeyedExternalArray(LLoadKeyed* instr); |
| void DoLoadKeyedFixedDoubleArray(LLoadKeyed* instr); |
| void DoLoadKeyedFixedArray(LLoadKeyed* instr); |
| void DoStoreKeyedExternalArray(LStoreKeyed* instr); |
| void DoStoreKeyedFixedDoubleArray(LStoreKeyed* instr); |
| void DoStoreKeyedFixedArray(LStoreKeyed* instr); |
| |
| template <class T> |
| void EmitVectorLoadICRegisters(T* instr); |
| |
| ZoneList<Deoptimizer::JumpTableEntry> jump_table_; |
| Scope* const scope_; |
| ZoneList<LDeferredCode*> deferred_; |
| bool frame_is_built_; |
| |
| // Builder that keeps track of safepoints in the code. The table |
| // itself is emitted at the end of the generated code. |
| SafepointTableBuilder safepoints_; |
| |
| // Compiler from a set of parallel moves to a sequential list of moves. |
| LGapResolver resolver_; |
| |
| Safepoint::Kind expected_safepoint_kind_; |
| |
| class PushSafepointRegistersScope final BASE_EMBEDDED { |
| public: |
| explicit PushSafepointRegistersScope(LCodeGen* codegen); |
| |
| ~PushSafepointRegistersScope(); |
| |
| private: |
| LCodeGen* codegen_; |
| }; |
| |
| friend class LDeferredCode; |
| friend class LEnvironment; |
| friend class SafepointGenerator; |
| DISALLOW_COPY_AND_ASSIGN(LCodeGen); |
| }; |
| |
| |
| class LDeferredCode : public ZoneObject { |
| public: |
| explicit LDeferredCode(LCodeGen* codegen) |
| : codegen_(codegen), |
| external_exit_(NULL), |
| instruction_index_(codegen->current_instruction_) { |
| codegen->AddDeferredCode(this); |
| } |
| |
| virtual ~LDeferredCode() {} |
| virtual void Generate() = 0; |
| virtual LInstruction* instr() = 0; |
| |
| void SetExit(Label* exit) { external_exit_ = exit; } |
| Label* entry() { return &entry_; } |
| Label* exit() { return external_exit_ != NULL ? external_exit_ : &exit_; } |
| int instruction_index() const { return instruction_index_; } |
| |
| protected: |
| LCodeGen* codegen() const { return codegen_; } |
| MacroAssembler* masm() const { return codegen_->masm(); } |
| |
| private: |
| LCodeGen* codegen_; |
| Label entry_; |
| Label exit_; |
| Label* external_exit_; |
| int instruction_index_; |
| }; |
| |
| } // namespace internal |
| } // namespace v8 |
| |
| #endif // V8_CRANKSHAFT_MIPS_LITHIUM_CODEGEN_MIPS_H_ |