| // Copyright 2013 the V8 project authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "src/compiler/code-generator.h" |
| |
| #include "src/compiler/code-generator-impl.h" |
| #include "src/compiler/linkage.h" |
| #include "src/compiler/pipeline.h" |
| |
| namespace v8 { |
| namespace internal { |
| namespace compiler { |
| |
| CodeGenerator::CodeGenerator(Frame* frame, Linkage* linkage, |
| InstructionSequence* code, CompilationInfo* info) |
| : frame_(frame), |
| linkage_(linkage), |
| code_(code), |
| info_(info), |
| labels_(zone()->NewArray<Label>(code->InstructionBlockCount())), |
| current_block_(BasicBlock::RpoNumber::Invalid()), |
| current_source_position_(SourcePosition::Invalid()), |
| masm_(code->zone()->isolate(), NULL, 0), |
| resolver_(this), |
| safepoints_(code->zone()), |
| deoptimization_states_(code->zone()), |
| deoptimization_literals_(code->zone()), |
| translations_(code->zone()), |
| last_lazy_deopt_pc_(0), |
| ools_(nullptr), |
| osr_pc_offset_(-1) { |
| for (int i = 0; i < code->InstructionBlockCount(); ++i) { |
| new (&labels_[i]) Label; |
| } |
| } |
| |
| |
| Handle<Code> CodeGenerator::GenerateCode() { |
| CompilationInfo* info = this->info(); |
| |
| // Emit a code line info recording start event. |
| PositionsRecorder* recorder = masm()->positions_recorder(); |
| LOG_CODE_EVENT(isolate(), CodeStartLinePosInfoRecordEvent(recorder)); |
| |
| // Place function entry hook if requested to do so. |
| if (linkage()->GetIncomingDescriptor()->IsJSFunctionCall()) { |
| ProfileEntryHookStub::MaybeCallEntryHook(masm()); |
| } |
| |
| // Architecture-specific, linkage-specific prologue. |
| info->set_prologue_offset(masm()->pc_offset()); |
| AssemblePrologue(); |
| |
| // Assemble all non-deferred blocks, followed by deferred ones. |
| for (int deferred = 0; deferred < 2; ++deferred) { |
| for (auto const block : code()->instruction_blocks()) { |
| if (block->IsDeferred() == (deferred == 0)) { |
| continue; |
| } |
| // Align loop headers on 16-byte boundaries. |
| if (block->IsLoopHeader()) masm()->Align(16); |
| // Bind a label for a block. |
| current_block_ = block->rpo_number(); |
| if (FLAG_code_comments) { |
| // TODO(titzer): these code comments are a giant memory leak. |
| Vector<char> buffer = Vector<char>::New(32); |
| SNPrintF(buffer, "-- B%d start --", block->id().ToInt()); |
| masm()->RecordComment(buffer.start()); |
| } |
| masm()->bind(GetLabel(current_block_)); |
| for (int i = block->code_start(); i < block->code_end(); ++i) { |
| AssembleInstruction(code()->InstructionAt(i)); |
| } |
| } |
| } |
| |
| // Assemble all out-of-line code. |
| if (ools_) { |
| masm()->RecordComment("-- Out of line code --"); |
| for (OutOfLineCode* ool = ools_; ool; ool = ool->next()) { |
| masm()->bind(ool->entry()); |
| ool->Generate(); |
| masm()->jmp(ool->exit()); |
| } |
| } |
| |
| FinishCode(masm()); |
| |
| // Ensure there is space for lazy deoptimization in the code. |
| if (!info->IsStub()) { |
| int target_offset = masm()->pc_offset() + Deoptimizer::patch_size(); |
| while (masm()->pc_offset() < target_offset) { |
| masm()->nop(); |
| } |
| } |
| |
| safepoints()->Emit(masm(), frame()->GetSpillSlotCount()); |
| |
| // TODO(titzer): what are the right code flags here? |
| Code::Kind kind = Code::STUB; |
| if (linkage()->GetIncomingDescriptor()->IsJSFunctionCall()) { |
| kind = Code::OPTIMIZED_FUNCTION; |
| } |
| Handle<Code> result = v8::internal::CodeGenerator::MakeCodeEpilogue( |
| masm(), Code::ComputeFlags(kind), info); |
| result->set_is_turbofanned(true); |
| result->set_stack_slots(frame()->GetSpillSlotCount()); |
| result->set_safepoint_table_offset(safepoints()->GetCodeOffset()); |
| |
| PopulateDeoptimizationData(result); |
| |
| // Ensure there is space for lazy deoptimization in the relocation info. |
| if (!info->IsStub()) { |
| Deoptimizer::EnsureRelocSpaceForLazyDeoptimization(result); |
| } |
| |
| // Emit a code line info recording stop event. |
| void* line_info = recorder->DetachJITHandlerData(); |
| LOG_CODE_EVENT(isolate(), CodeEndLinePosInfoRecordEvent(*result, line_info)); |
| |
| return result; |
| } |
| |
| |
| bool CodeGenerator::IsNextInAssemblyOrder(BasicBlock::RpoNumber block) const { |
| return code()->InstructionBlockAt(current_block_)->ao_number().IsNext( |
| code()->InstructionBlockAt(block)->ao_number()); |
| } |
| |
| |
| void CodeGenerator::RecordSafepoint(PointerMap* pointers, Safepoint::Kind kind, |
| int arguments, |
| Safepoint::DeoptMode deopt_mode) { |
| const ZoneList<InstructionOperand*>* operands = |
| pointers->GetNormalizedOperands(); |
| Safepoint safepoint = |
| safepoints()->DefineSafepoint(masm(), kind, arguments, deopt_mode); |
| for (int i = 0; i < operands->length(); i++) { |
| InstructionOperand* pointer = operands->at(i); |
| if (pointer->IsStackSlot()) { |
| safepoint.DefinePointerSlot(pointer->index(), zone()); |
| } else if (pointer->IsRegister() && (kind & Safepoint::kWithRegisters)) { |
| Register reg = Register::FromAllocationIndex(pointer->index()); |
| safepoint.DefinePointerRegister(reg, zone()); |
| } |
| } |
| } |
| |
| |
| void CodeGenerator::AssembleInstruction(Instruction* instr) { |
| if (instr->IsGapMoves()) { |
| // Handle parallel moves associated with the gap instruction. |
| AssembleGap(GapInstruction::cast(instr)); |
| } else if (instr->IsSourcePosition()) { |
| AssembleSourcePosition(SourcePositionInstruction::cast(instr)); |
| } else { |
| // Assemble architecture-specific code for the instruction. |
| AssembleArchInstruction(instr); |
| |
| FlagsMode mode = FlagsModeField::decode(instr->opcode()); |
| FlagsCondition condition = FlagsConditionField::decode(instr->opcode()); |
| if (mode == kFlags_branch) { |
| // Assemble a branch after this instruction. |
| InstructionOperandConverter i(this, instr); |
| BasicBlock::RpoNumber true_rpo = |
| i.InputRpo(static_cast<int>(instr->InputCount()) - 2); |
| BasicBlock::RpoNumber false_rpo = |
| i.InputRpo(static_cast<int>(instr->InputCount()) - 1); |
| |
| if (true_rpo == false_rpo) { |
| // redundant branch. |
| if (!IsNextInAssemblyOrder(true_rpo)) { |
| AssembleArchJump(true_rpo); |
| } |
| return; |
| } |
| if (IsNextInAssemblyOrder(true_rpo)) { |
| // true block is next, can fall through if condition negated. |
| std::swap(true_rpo, false_rpo); |
| condition = NegateFlagsCondition(condition); |
| } |
| BranchInfo branch; |
| branch.condition = condition; |
| branch.true_label = GetLabel(true_rpo); |
| branch.false_label = GetLabel(false_rpo); |
| branch.fallthru = IsNextInAssemblyOrder(false_rpo); |
| // Assemble architecture-specific branch. |
| AssembleArchBranch(instr, &branch); |
| } else if (mode == kFlags_set) { |
| // Assemble a boolean materialization after this instruction. |
| AssembleArchBoolean(instr, condition); |
| } |
| } |
| } |
| |
| |
| void CodeGenerator::AssembleSourcePosition(SourcePositionInstruction* instr) { |
| SourcePosition source_position = instr->source_position(); |
| if (source_position == current_source_position_) return; |
| DCHECK(!source_position.IsInvalid()); |
| if (!source_position.IsUnknown()) { |
| int code_pos = source_position.raw(); |
| masm()->positions_recorder()->RecordPosition(source_position.raw()); |
| masm()->positions_recorder()->WriteRecordedPositions(); |
| if (FLAG_code_comments) { |
| Vector<char> buffer = Vector<char>::New(256); |
| CompilationInfo* info = this->info(); |
| int ln = Script::GetLineNumber(info->script(), code_pos); |
| int cn = Script::GetColumnNumber(info->script(), code_pos); |
| if (info->script()->name()->IsString()) { |
| Handle<String> file(String::cast(info->script()->name())); |
| base::OS::SNPrintF(buffer.start(), buffer.length(), "-- %s:%d:%d --", |
| file->ToCString().get(), ln, cn); |
| } else { |
| base::OS::SNPrintF(buffer.start(), buffer.length(), |
| "-- <unknown>:%d:%d --", ln, cn); |
| } |
| masm()->RecordComment(buffer.start()); |
| } |
| } |
| current_source_position_ = source_position; |
| } |
| |
| |
| void CodeGenerator::AssembleGap(GapInstruction* instr) { |
| for (int i = GapInstruction::FIRST_INNER_POSITION; |
| i <= GapInstruction::LAST_INNER_POSITION; i++) { |
| GapInstruction::InnerPosition inner_pos = |
| static_cast<GapInstruction::InnerPosition>(i); |
| ParallelMove* move = instr->GetParallelMove(inner_pos); |
| if (move != NULL) resolver()->Resolve(move); |
| } |
| } |
| |
| |
| void CodeGenerator::PopulateDeoptimizationData(Handle<Code> code_object) { |
| CompilationInfo* info = this->info(); |
| int deopt_count = static_cast<int>(deoptimization_states_.size()); |
| if (deopt_count == 0 && !info->is_osr()) return; |
| Handle<DeoptimizationInputData> data = |
| DeoptimizationInputData::New(isolate(), deopt_count, TENURED); |
| |
| Handle<ByteArray> translation_array = |
| translations_.CreateByteArray(isolate()->factory()); |
| |
| data->SetTranslationByteArray(*translation_array); |
| data->SetInlinedFunctionCount(Smi::FromInt(0)); |
| data->SetOptimizationId(Smi::FromInt(info->optimization_id())); |
| // TODO(jarin) The following code was copied over from Lithium, not sure |
| // whether the scope or the IsOptimizing condition are really needed. |
| if (info->IsOptimizing()) { |
| // Reference to shared function info does not change between phases. |
| AllowDeferredHandleDereference allow_handle_dereference; |
| data->SetSharedFunctionInfo(*info->shared_info()); |
| } else { |
| data->SetSharedFunctionInfo(Smi::FromInt(0)); |
| } |
| |
| Handle<FixedArray> literals = isolate()->factory()->NewFixedArray( |
| static_cast<int>(deoptimization_literals_.size()), TENURED); |
| { |
| AllowDeferredHandleDereference copy_handles; |
| for (unsigned i = 0; i < deoptimization_literals_.size(); i++) { |
| literals->set(i, *deoptimization_literals_[i]); |
| } |
| data->SetLiteralArray(*literals); |
| } |
| |
| if (info->is_osr()) { |
| DCHECK(osr_pc_offset_ >= 0); |
| data->SetOsrAstId(Smi::FromInt(info_->osr_ast_id().ToInt())); |
| data->SetOsrPcOffset(Smi::FromInt(osr_pc_offset_)); |
| } else { |
| BailoutId osr_ast_id = BailoutId::None(); |
| data->SetOsrAstId(Smi::FromInt(osr_ast_id.ToInt())); |
| data->SetOsrPcOffset(Smi::FromInt(-1)); |
| } |
| |
| // Populate deoptimization entries. |
| for (int i = 0; i < deopt_count; i++) { |
| DeoptimizationState* deoptimization_state = deoptimization_states_[i]; |
| data->SetAstId(i, deoptimization_state->bailout_id()); |
| CHECK_NE(NULL, deoptimization_states_[i]); |
| data->SetTranslationIndex( |
| i, Smi::FromInt(deoptimization_states_[i]->translation_id())); |
| data->SetArgumentsStackHeight(i, Smi::FromInt(0)); |
| data->SetPc(i, Smi::FromInt(deoptimization_state->pc_offset())); |
| } |
| |
| code_object->set_deoptimization_data(*data); |
| } |
| |
| |
| void CodeGenerator::AddSafepointAndDeopt(Instruction* instr) { |
| CallDescriptor::Flags flags(MiscField::decode(instr->opcode())); |
| |
| bool needs_frame_state = (flags & CallDescriptor::kNeedsFrameState); |
| |
| RecordSafepoint( |
| instr->pointer_map(), Safepoint::kSimple, 0, |
| needs_frame_state ? Safepoint::kLazyDeopt : Safepoint::kNoLazyDeopt); |
| |
| if (flags & CallDescriptor::kNeedsNopAfterCall) { |
| AddNopForSmiCodeInlining(); |
| } |
| |
| if (needs_frame_state) { |
| MarkLazyDeoptSite(); |
| // If the frame state is present, it starts at argument 1 |
| // (just after the code address). |
| InstructionOperandConverter converter(this, instr); |
| // Deoptimization info starts at argument 1 |
| size_t frame_state_offset = 1; |
| FrameStateDescriptor* descriptor = |
| GetFrameStateDescriptor(instr, frame_state_offset); |
| int pc_offset = masm()->pc_offset(); |
| int deopt_state_id = BuildTranslation(instr, pc_offset, frame_state_offset, |
| descriptor->state_combine()); |
| // If the pre-call frame state differs from the post-call one, produce the |
| // pre-call frame state, too. |
| // TODO(jarin) We might want to avoid building the pre-call frame state |
| // because it is only used to get locals and arguments (by the debugger and |
| // f.arguments), and those are the same in the pre-call and post-call |
| // states. |
| if (!descriptor->state_combine().IsOutputIgnored()) { |
| deopt_state_id = BuildTranslation(instr, -1, frame_state_offset, |
| OutputFrameStateCombine::Ignore()); |
| } |
| #if DEBUG |
| // Make sure all the values live in stack slots or they are immediates. |
| // (The values should not live in register because registers are clobbered |
| // by calls.) |
| for (size_t i = 0; i < descriptor->GetSize(); i++) { |
| InstructionOperand* op = instr->InputAt(frame_state_offset + 1 + i); |
| CHECK(op->IsStackSlot() || op->IsDoubleStackSlot() || op->IsImmediate()); |
| } |
| #endif |
| safepoints()->RecordLazyDeoptimizationIndex(deopt_state_id); |
| } |
| } |
| |
| |
| int CodeGenerator::DefineDeoptimizationLiteral(Handle<Object> literal) { |
| int result = static_cast<int>(deoptimization_literals_.size()); |
| for (unsigned i = 0; i < deoptimization_literals_.size(); ++i) { |
| if (deoptimization_literals_[i].is_identical_to(literal)) return i; |
| } |
| deoptimization_literals_.push_back(literal); |
| return result; |
| } |
| |
| |
| FrameStateDescriptor* CodeGenerator::GetFrameStateDescriptor( |
| Instruction* instr, size_t frame_state_offset) { |
| InstructionOperandConverter i(this, instr); |
| InstructionSequence::StateId state_id = InstructionSequence::StateId::FromInt( |
| i.InputInt32(static_cast<int>(frame_state_offset))); |
| return code()->GetFrameStateDescriptor(state_id); |
| } |
| |
| struct OperandAndType { |
| OperandAndType(InstructionOperand* operand, MachineType type) |
| : operand_(operand), type_(type) {} |
| |
| InstructionOperand* operand_; |
| MachineType type_; |
| }; |
| |
| static OperandAndType TypedOperandForFrameState( |
| FrameStateDescriptor* descriptor, Instruction* instr, |
| size_t frame_state_offset, size_t index, OutputFrameStateCombine combine) { |
| DCHECK(index < descriptor->GetSize(combine)); |
| switch (combine.kind()) { |
| case OutputFrameStateCombine::kPushOutput: { |
| DCHECK(combine.GetPushCount() <= instr->OutputCount()); |
| size_t size_without_output = |
| descriptor->GetSize(OutputFrameStateCombine::Ignore()); |
| // If the index is past the existing stack items, return the output. |
| if (index >= size_without_output) { |
| return OperandAndType(instr->OutputAt(index - size_without_output), |
| kMachAnyTagged); |
| } |
| break; |
| } |
| case OutputFrameStateCombine::kPokeAt: |
| size_t index_from_top = |
| descriptor->GetSize(combine) - 1 - combine.GetOffsetToPokeAt(); |
| if (index >= index_from_top && |
| index < index_from_top + instr->OutputCount()) { |
| return OperandAndType(instr->OutputAt(index - index_from_top), |
| kMachAnyTagged); |
| } |
| break; |
| } |
| return OperandAndType(instr->InputAt(frame_state_offset + index), |
| descriptor->GetType(index)); |
| } |
| |
| |
| void CodeGenerator::BuildTranslationForFrameStateDescriptor( |
| FrameStateDescriptor* descriptor, Instruction* instr, |
| Translation* translation, size_t frame_state_offset, |
| OutputFrameStateCombine state_combine) { |
| // Outer-most state must be added to translation first. |
| if (descriptor->outer_state() != NULL) { |
| BuildTranslationForFrameStateDescriptor(descriptor->outer_state(), instr, |
| translation, frame_state_offset, |
| OutputFrameStateCombine::Ignore()); |
| } |
| |
| int id = Translation::kSelfLiteralId; |
| if (!descriptor->jsfunction().is_null()) { |
| id = DefineDeoptimizationLiteral( |
| Handle<Object>::cast(descriptor->jsfunction().ToHandleChecked())); |
| } |
| |
| switch (descriptor->type()) { |
| case JS_FRAME: |
| translation->BeginJSFrame( |
| descriptor->bailout_id(), id, |
| static_cast<unsigned int>(descriptor->GetSize(state_combine) - |
| descriptor->parameters_count())); |
| break; |
| case ARGUMENTS_ADAPTOR: |
| translation->BeginArgumentsAdaptorFrame( |
| id, static_cast<unsigned int>(descriptor->parameters_count())); |
| break; |
| } |
| |
| frame_state_offset += descriptor->outer_state()->GetTotalSize(); |
| for (size_t i = 0; i < descriptor->GetSize(state_combine); i++) { |
| OperandAndType op = TypedOperandForFrameState( |
| descriptor, instr, frame_state_offset, i, state_combine); |
| AddTranslationForOperand(translation, instr, op.operand_, op.type_); |
| } |
| } |
| |
| |
| int CodeGenerator::BuildTranslation(Instruction* instr, int pc_offset, |
| size_t frame_state_offset, |
| OutputFrameStateCombine state_combine) { |
| FrameStateDescriptor* descriptor = |
| GetFrameStateDescriptor(instr, frame_state_offset); |
| frame_state_offset++; |
| |
| Translation translation( |
| &translations_, static_cast<int>(descriptor->GetFrameCount()), |
| static_cast<int>(descriptor->GetJSFrameCount()), zone()); |
| BuildTranslationForFrameStateDescriptor(descriptor, instr, &translation, |
| frame_state_offset, state_combine); |
| |
| int deoptimization_id = static_cast<int>(deoptimization_states_.size()); |
| |
| deoptimization_states_.push_back(new (zone()) DeoptimizationState( |
| descriptor->bailout_id(), translation.index(), pc_offset)); |
| |
| return deoptimization_id; |
| } |
| |
| |
| void CodeGenerator::AddTranslationForOperand(Translation* translation, |
| Instruction* instr, |
| InstructionOperand* op, |
| MachineType type) { |
| if (op->IsStackSlot()) { |
| if (type == kMachBool || type == kMachInt32 || type == kMachInt8 || |
| type == kMachInt16) { |
| translation->StoreInt32StackSlot(op->index()); |
| } else if (type == kMachUint32 || type == kMachUint16 || |
| type == kMachUint8) { |
| translation->StoreUint32StackSlot(op->index()); |
| } else if ((type & kRepMask) == kRepTagged) { |
| translation->StoreStackSlot(op->index()); |
| } else { |
| CHECK(false); |
| } |
| } else if (op->IsDoubleStackSlot()) { |
| DCHECK((type & (kRepFloat32 | kRepFloat64)) != 0); |
| translation->StoreDoubleStackSlot(op->index()); |
| } else if (op->IsRegister()) { |
| InstructionOperandConverter converter(this, instr); |
| if (type == kMachBool || type == kMachInt32 || type == kMachInt8 || |
| type == kMachInt16) { |
| translation->StoreInt32Register(converter.ToRegister(op)); |
| } else if (type == kMachUint32 || type == kMachUint16 || |
| type == kMachUint8) { |
| translation->StoreUint32Register(converter.ToRegister(op)); |
| } else if ((type & kRepMask) == kRepTagged) { |
| translation->StoreRegister(converter.ToRegister(op)); |
| } else { |
| CHECK(false); |
| } |
| } else if (op->IsDoubleRegister()) { |
| DCHECK((type & (kRepFloat32 | kRepFloat64)) != 0); |
| InstructionOperandConverter converter(this, instr); |
| translation->StoreDoubleRegister(converter.ToDoubleRegister(op)); |
| } else if (op->IsImmediate()) { |
| InstructionOperandConverter converter(this, instr); |
| Constant constant = converter.ToConstant(op); |
| Handle<Object> constant_object; |
| switch (constant.type()) { |
| case Constant::kInt32: |
| DCHECK(type == kMachInt32 || type == kMachUint32); |
| constant_object = |
| isolate()->factory()->NewNumberFromInt(constant.ToInt32()); |
| break; |
| case Constant::kFloat64: |
| DCHECK(type == kMachFloat64 || type == kMachAnyTagged); |
| constant_object = isolate()->factory()->NewNumber(constant.ToFloat64()); |
| break; |
| case Constant::kHeapObject: |
| DCHECK((type & kRepMask) == kRepTagged); |
| constant_object = constant.ToHeapObject(); |
| break; |
| default: |
| CHECK(false); |
| } |
| int literal_id = DefineDeoptimizationLiteral(constant_object); |
| translation->StoreLiteral(literal_id); |
| } else { |
| CHECK(false); |
| } |
| } |
| |
| |
| void CodeGenerator::MarkLazyDeoptSite() { |
| last_lazy_deopt_pc_ = masm()->pc_offset(); |
| } |
| |
| #if !V8_TURBOFAN_BACKEND |
| |
| void CodeGenerator::AssembleArchInstruction(Instruction* instr) { |
| UNIMPLEMENTED(); |
| } |
| |
| |
| void CodeGenerator::AssembleArchBranch(Instruction* instr, |
| BranchInfo* branch) { |
| UNIMPLEMENTED(); |
| } |
| |
| |
| void CodeGenerator::AssembleArchBoolean(Instruction* instr, |
| FlagsCondition condition) { |
| UNIMPLEMENTED(); |
| } |
| |
| |
| void CodeGenerator::AssembleArchJump(BasicBlock::RpoNumber target) { |
| UNIMPLEMENTED(); |
| } |
| |
| |
| void CodeGenerator::AssembleDeoptimizerCall(int deoptimization_id) { |
| UNIMPLEMENTED(); |
| } |
| |
| |
| void CodeGenerator::AssemblePrologue() { UNIMPLEMENTED(); } |
| |
| |
| void CodeGenerator::AssembleReturn() { UNIMPLEMENTED(); } |
| |
| |
| void CodeGenerator::AssembleMove(InstructionOperand* source, |
| InstructionOperand* destination) { |
| UNIMPLEMENTED(); |
| } |
| |
| |
| void CodeGenerator::AssembleSwap(InstructionOperand* source, |
| InstructionOperand* destination) { |
| UNIMPLEMENTED(); |
| } |
| |
| |
| void CodeGenerator::AddNopForSmiCodeInlining() { UNIMPLEMENTED(); } |
| |
| #endif // !V8_TURBOFAN_BACKEND |
| |
| |
| OutOfLineCode::OutOfLineCode(CodeGenerator* gen) |
| : masm_(gen->masm()), next_(gen->ools_) { |
| gen->ools_ = this; |
| } |
| |
| |
| OutOfLineCode::~OutOfLineCode() {} |
| |
| } // namespace compiler |
| } // namespace internal |
| } // namespace v8 |