| /* |
| * Copyright (C) 2011-2022 Apple Inc. All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY |
| * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
| * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR |
| * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
| * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
| * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
| * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
| * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #include "config.h" |
| #include "DFGJITCompiler.h" |
| |
| #if ENABLE(DFG_JIT) |
| |
| #include "CodeBlock.h" |
| #include "CodeBlockWithJITType.h" |
| #include "DFGFailedFinalizer.h" |
| #include "DFGInlineCacheWrapperInlines.h" |
| #include "DFGJITCode.h" |
| #include "DFGJITFinalizer.h" |
| #include "DFGOSRExit.h" |
| #include "DFGSpeculativeJIT.h" |
| #include "DFGThunks.h" |
| #include "JSCJSValueInlines.h" |
| #include "LinkBuffer.h" |
| #include "ProbeContext.h" |
| #include "ThunkGenerators.h" |
| #include "VM.h" |
| |
| namespace JSC { namespace DFG { |
| |
| JITCompiler::JITCompiler(Graph& dfg) |
| : CCallHelpers(dfg.m_codeBlock) |
| , m_graph(dfg) |
| , m_jitCode(adoptRef(new JITCode(m_graph.m_plan.isUnlinked()))) |
| , m_blockHeads(dfg.numBlocks()) |
| , m_pcToCodeOriginMapBuilder(dfg.m_vm) |
| { |
| if (UNLIKELY(shouldDumpDisassembly() || m_graph.m_vm.m_perBytecodeProfiler)) |
| m_disassembler = makeUnique<Disassembler>(dfg); |
| #if ENABLE(FTL_JIT) |
| m_jitCode->tierUpInLoopHierarchy = WTFMove(m_graph.m_plan.tierUpInLoopHierarchy()); |
| for (BytecodeIndex tierUpBytecode : m_graph.m_plan.tierUpAndOSREnterBytecodes()) |
| m_jitCode->tierUpEntryTriggers.add(tierUpBytecode, JITCode::TriggerReason::DontTrigger); |
| #endif |
| } |
| |
| JITCompiler::~JITCompiler() |
| { |
| } |
| |
| void JITCompiler::linkOSRExits() |
| { |
| ASSERT(m_osrExit.size() == m_exitCompilationInfo.size()); |
| if (UNLIKELY(m_graph.compilation())) { |
| for (unsigned i = 0; i < m_osrExit.size(); ++i) { |
| OSRExitCompilationInfo& info = m_exitCompilationInfo[i]; |
| Vector<Label> labels; |
| |
| auto appendLabel = [&] (Label label) { |
| RELEASE_ASSERT(label.isSet()); |
| labels.append(label); |
| }; |
| |
| if (!info.m_failureJumps.empty()) { |
| for (unsigned j = 0; j < info.m_failureJumps.jumps().size(); ++j) |
| appendLabel(info.m_failureJumps.jumps()[j].label()); |
| } else if (info.m_replacementSource.isSet()) |
| appendLabel(info.m_replacementSource); |
| m_exitSiteLabels.append(labels); |
| } |
| } |
| |
| JumpList dispatchCases; |
| JumpList dispatchCasesWithoutLinkedFailures; |
| for (unsigned i = 0; i < m_osrExit.size(); ++i) { |
| OSRExitCompilationInfo& info = m_exitCompilationInfo[i]; |
| JumpList& failureJumps = info.m_failureJumps; |
| if (!failureJumps.empty()) |
| failureJumps.link(this); |
| else |
| info.m_replacementDestination = label(); |
| |
| jitAssertHasValidCallFrame(); |
| #if USE(JSVALUE64) |
| if (m_graph.m_plan.isUnlinked()) { |
| move(TrustedImm32(i), GPRInfo::numberTagRegister); |
| if (info.m_replacementDestination.isSet()) |
| dispatchCasesWithoutLinkedFailures.append(jump()); |
| else |
| dispatchCases.append(jump()); |
| continue; |
| } |
| #endif |
| UNUSED_VARIABLE(dispatchCases); |
| UNUSED_VARIABLE(dispatchCasesWithoutLinkedFailures); |
| store32(TrustedImm32(i), &vm().osrExitIndex); |
| info.m_patchableJump = patchableJump(); |
| } |
| |
| #if USE(JSVALUE64) |
| if (m_graph.m_plan.isUnlinked()) { |
| // When jumping to OSR exit handler via exception, we do not have proper callFrameRegister and constantsRegister. |
| // We should reload appropriate callFrameRegister from VM::callFrameForCatch to materialize constants buffer register. |
| // FIXME: The following code can be a DFG Thunk. |
| if (!dispatchCasesWithoutLinkedFailures.empty()) { |
| dispatchCasesWithoutLinkedFailures.link(this); |
| loadPtr(vm().addressOfCallFrameForCatch(), GPRInfo::notCellMaskRegister); |
| MacroAssembler::Jump didNotHaveException = branchTestPtr(MacroAssembler::Zero, GPRInfo::notCellMaskRegister); |
| move(GPRInfo::notCellMaskRegister, GPRInfo::constantsRegister); |
| emitGetFromCallFrameHeaderPtr(CallFrameSlot::codeBlock, GPRInfo::constantsRegister, GPRInfo::constantsRegister); |
| loadPtr(Address(GPRInfo::constantsRegister, CodeBlock::offsetOfJITData()), GPRInfo::constantsRegister); |
| didNotHaveException.link(this); |
| } |
| dispatchCases.link(this); |
| store32(GPRInfo::numberTagRegister, &vm().osrExitIndex); |
| loadPtr(Address(GPRInfo::constantsRegister, JITData::offsetOfExits()), GPRInfo::constantsRegister); |
| static_assert(sizeof(JITData::ExitVector::value_type) == 16); |
| ASSERT(!JITData::ExitVector::value_type::offsetOfCodePtr()); |
| lshiftPtr(TrustedImm32(4), GPRInfo::numberTagRegister); |
| addPtr(GPRInfo::numberTagRegister, GPRInfo::constantsRegister); |
| emitMaterializeTagCheckRegisters(); |
| farJump(Address(GPRInfo::constantsRegister, JITData::ExitVector::Storage::offsetOfData()), OSRExitPtrTag); |
| } |
| #endif |
| } |
| |
| void JITCompiler::compileEntry() |
| { |
| // This code currently matches the old JIT. In the function header we need to |
| // save return address and call frame via the prologue and perform a fast stack check. |
| // FIXME: https://bugs.webkit.org/show_bug.cgi?id=56292 |
| // We'll need to convert the remaining cti_ style calls (specifically the stack |
| // check) which will be dependent on stack layout. (We'd need to account for this in |
| // both normal return code and when jumping to an exception handler). |
| emitFunctionPrologue(); |
| jitAssertCodeBlockOnCallFrameWithType(GPRInfo::regT2, JITType::DFGJIT); |
| } |
| |
| void JITCompiler::compileSetupRegistersForEntry() |
| { |
| emitSaveCalleeSaves(); |
| emitMaterializeTagCheckRegisters(); |
| #if USE(JSVALUE64) |
| if (m_graph.m_plan.isUnlinked()) { |
| emitGetFromCallFrameHeaderPtr(CallFrameSlot::codeBlock, GPRInfo::constantsRegister); |
| loadPtr(Address(GPRInfo::constantsRegister, CodeBlock::offsetOfJITData()), GPRInfo::constantsRegister); |
| } |
| #endif |
| } |
| |
| void JITCompiler::compileEntryExecutionFlag() |
| { |
| #if ENABLE(FTL_JIT) |
| if (m_graph.m_plan.canTierUpAndOSREnter()) |
| store8(TrustedImm32(0), &m_jitCode->neverExecutedEntry); |
| #endif // ENABLE(FTL_JIT) |
| } |
| |
| void JITCompiler::link(LinkBuffer& linkBuffer) |
| { |
| // Link the code, populate data in CodeBlock data structures. |
| m_jitCode->common.frameRegisterCount = m_graph.frameRegisterCount(); |
| m_jitCode->common.requiredRegisterCountForExit = m_graph.requiredRegisterCountForExit(); |
| |
| if (!m_graph.m_plan.inlineCallFrames()->isEmpty()) |
| m_jitCode->common.inlineCallFrames = m_graph.m_plan.inlineCallFrames(); |
| if (!m_graph.m_stringSearchTable8.isEmpty()) { |
| FixedVector<std::unique_ptr<BoyerMooreHorspoolTable<uint8_t>>> tables(m_graph.m_stringSearchTable8.size()); |
| unsigned index = 0; |
| for (auto& entry : m_graph.m_stringSearchTable8) |
| tables[index++] = WTFMove(entry.value); |
| m_jitCode->common.m_stringSearchTable8 = WTFMove(tables); |
| } |
| |
| #if USE(JSVALUE32_64) |
| m_jitCode->common.doubleConstants = WTFMove(m_graph.m_doubleConstants); |
| #endif |
| |
| m_graph.registerFrozenValues(); |
| |
| ASSERT(m_jitCode->m_stringSwitchJumpTables.isEmpty()); |
| ASSERT(m_jitCode->m_switchJumpTables.isEmpty()); |
| if (!m_graph.m_stringSwitchJumpTables.isEmpty()) |
| m_jitCode->m_stringSwitchJumpTables = WTFMove(m_graph.m_stringSwitchJumpTables); |
| if (!m_graph.m_switchJumpTables.isEmpty()) |
| m_jitCode->m_switchJumpTables = WTFMove(m_graph.m_switchJumpTables); |
| |
| for (Bag<SwitchData>::iterator iter = m_graph.m_switchData.begin(); !!iter; ++iter) { |
| SwitchData& data = **iter; |
| switch (data.kind) { |
| case SwitchChar: |
| case SwitchImm: { |
| if (!data.didUseJumpTable) { |
| ASSERT(m_jitCode->m_switchJumpTables[data.switchTableIndex].isEmpty()); |
| continue; |
| } |
| |
| const UnlinkedSimpleJumpTable& unlinkedTable = m_graph.unlinkedSwitchJumpTable(data.switchTableIndex); |
| SimpleJumpTable& linkedTable = m_jitCode->m_switchJumpTables[data.switchTableIndex]; |
| linkedTable.m_ctiDefault = linkBuffer.locationOf<JSSwitchPtrTag>(m_blockHeads[data.fallThrough.block->index]); |
| RELEASE_ASSERT(linkedTable.m_ctiOffsets.size() == unlinkedTable.m_branchOffsets.size()); |
| for (unsigned j = linkedTable.m_ctiOffsets.size(); j--;) |
| linkedTable.m_ctiOffsets[j] = linkedTable.m_ctiDefault; |
| for (unsigned j = data.cases.size(); j--;) { |
| SwitchCase& myCase = data.cases[j]; |
| linkedTable.m_ctiOffsets[myCase.value.switchLookupValue(data.kind) - unlinkedTable.m_min] = |
| linkBuffer.locationOf<JSSwitchPtrTag>(m_blockHeads[myCase.target.block->index]); |
| } |
| break; |
| } |
| |
| case SwitchString: { |
| if (!data.didUseJumpTable) { |
| ASSERT(m_jitCode->m_stringSwitchJumpTables[data.switchTableIndex].isEmpty()); |
| continue; |
| } |
| |
| const UnlinkedStringJumpTable& unlinkedTable = m_graph.unlinkedStringSwitchJumpTable(data.switchTableIndex); |
| StringJumpTable& linkedTable = m_jitCode->m_stringSwitchJumpTables[data.switchTableIndex]; |
| auto ctiDefault = linkBuffer.locationOf<JSSwitchPtrTag>(m_blockHeads[data.fallThrough.block->index]); |
| RELEASE_ASSERT(linkedTable.m_ctiOffsets.size() == unlinkedTable.m_offsetTable.size() + 1); |
| for (auto& entry : linkedTable.m_ctiOffsets) |
| entry = ctiDefault; |
| for (unsigned j = data.cases.size(); j--;) { |
| SwitchCase& myCase = data.cases[j]; |
| auto iter = unlinkedTable.m_offsetTable.find(myCase.value.stringImpl()); |
| RELEASE_ASSERT(iter != unlinkedTable.m_offsetTable.end()); |
| linkedTable.m_ctiOffsets[iter->value.m_indexInTable] = linkBuffer.locationOf<JSSwitchPtrTag>(m_blockHeads[myCase.target.block->index]); |
| } |
| break; |
| } |
| |
| case SwitchCell: |
| RELEASE_ASSERT_NOT_REACHED(); |
| break; |
| } |
| } |
| |
| // Link all calls out from the JIT code to their respective functions. |
| for (unsigned i = 0; i < m_calls.size(); ++i) |
| linkBuffer.link(m_calls[i].m_call, m_calls[i].m_function); |
| |
| finalizeInlineCaches(m_getByIds, linkBuffer); |
| finalizeInlineCaches(m_getByIdsWithThis, linkBuffer); |
| finalizeInlineCaches(m_getByVals, linkBuffer); |
| finalizeInlineCaches(m_getByValsWithThis, linkBuffer); |
| finalizeInlineCaches(m_putByIds, linkBuffer); |
| finalizeInlineCaches(m_putByVals, linkBuffer); |
| finalizeInlineCaches(m_delByIds, linkBuffer); |
| finalizeInlineCaches(m_delByVals, linkBuffer); |
| finalizeInlineCaches(m_inByIds, linkBuffer); |
| finalizeInlineCaches(m_inByVals, linkBuffer); |
| finalizeInlineCaches(m_instanceOfs, linkBuffer); |
| finalizeInlineCaches(m_privateBrandAccesses, linkBuffer); |
| |
| if (m_graph.m_plan.isUnlinked()) { |
| m_jitCode->m_unlinkedStubInfos = FixedVector<UnlinkedStructureStubInfo>(m_unlinkedStubInfos.size()); |
| if (m_jitCode->m_unlinkedStubInfos.size()) |
| std::move(m_unlinkedStubInfos.begin(), m_unlinkedStubInfos.end(), m_jitCode->m_unlinkedStubInfos.begin()); |
| ASSERT(m_jitCode->common.m_stubInfos.isEmpty()); |
| } |
| |
| for (auto& record : m_jsCalls) { |
| std::visit([&](auto* info) { |
| info->setCodeLocations( |
| linkBuffer.locationOf<JSInternalPtrTag>(record.slowPathStart), |
| linkBuffer.locationOf<JSInternalPtrTag>(record.doneLocation)); |
| }, record.info); |
| } |
| |
| for (auto& record : m_jsDirectCalls) { |
| auto& info = *record.info; |
| info.setCodeLocations( |
| linkBuffer.locationOf<JSInternalPtrTag>(record.slowPath), |
| CodeLocationLabel<JSInternalPtrTag>()); |
| } |
| |
| if (m_graph.m_plan.isUnlinked()) { |
| m_jitCode->m_unlinkedCallLinkInfos = FixedVector<UnlinkedCallLinkInfo>(m_unlinkedCallLinkInfos.size()); |
| if (m_jitCode->m_unlinkedCallLinkInfos.size()) |
| std::move(m_unlinkedCallLinkInfos.begin(), m_unlinkedCallLinkInfos.end(), m_jitCode->m_unlinkedCallLinkInfos.begin()); |
| ASSERT(m_jitCode->common.m_callLinkInfos.isEmpty()); |
| } |
| |
| |
| if (!m_exceptionChecks.empty()) |
| linkBuffer.link(m_exceptionChecks, CodeLocationLabel(vm().getCTIStub(handleExceptionGenerator).retaggedCode<NoPtrTag>())); |
| if (!m_exceptionChecksWithCallFrameRollback.empty()) |
| linkBuffer.link(m_exceptionChecksWithCallFrameRollback, CodeLocationLabel(vm().getCTIStub(handleExceptionWithCallFrameRollbackGenerator).retaggedCode<NoPtrTag>())); |
| |
| if (!m_graph.m_plan.isUnlinked()) { |
| MacroAssemblerCodeRef<JITThunkPtrTag> osrExitThunk = vm().getCTIStub(osrExitGenerationThunkGenerator); |
| auto target = CodeLocationLabel<JITThunkPtrTag>(osrExitThunk.code()); |
| Vector<JumpReplacement> jumpReplacements; |
| for (unsigned i = 0; i < m_osrExit.size(); ++i) { |
| OSRExitCompilationInfo& info = m_exitCompilationInfo[i]; |
| linkBuffer.link(info.m_patchableJump.m_jump, target); |
| OSRExit& exit = m_osrExit[i]; |
| exit.m_patchableJumpLocation = linkBuffer.locationOf<JSInternalPtrTag>(info.m_patchableJump); |
| if (info.m_replacementSource.isSet()) { |
| jumpReplacements.append(JumpReplacement( |
| linkBuffer.locationOf<JSInternalPtrTag>(info.m_replacementSource), |
| linkBuffer.locationOf<OSRExitPtrTag>(info.m_replacementDestination))); |
| } |
| } |
| m_jitCode->common.m_jumpReplacements = WTFMove(jumpReplacements); |
| } |
| |
| #if ASSERT_ENABLED |
| for (auto& info : m_exitCompilationInfo) { |
| if (info.m_replacementSource.isSet()) |
| ASSERT(!m_graph.m_plan.isUnlinked()); |
| } |
| if (m_graph.m_plan.isUnlinked()) |
| ASSERT(m_jitCode->common.m_jumpReplacements.isEmpty()); |
| #endif |
| |
| if (UNLIKELY(m_graph.compilation())) { |
| ASSERT(m_exitSiteLabels.size() == m_osrExit.size()); |
| for (unsigned i = 0; i < m_exitSiteLabels.size(); ++i) { |
| Vector<Label>& labels = m_exitSiteLabels[i]; |
| Vector<CodePtr<JSInternalPtrTag>> addresses; |
| for (unsigned j = 0; j < labels.size(); ++j) |
| addresses.append(linkBuffer.locationOf<JSInternalPtrTag>(labels[j])); |
| m_graph.compilation()->addOSRExitSite(addresses); |
| } |
| } else |
| ASSERT(!m_exitSiteLabels.size()); |
| |
| m_jitCode->common.compilation = m_graph.compilation(); |
| m_jitCode->m_osrExit = WTFMove(m_osrExit); |
| m_jitCode->m_speculationRecovery = WTFMove(m_speculationRecovery); |
| |
| // Link new DFG exception handlers and remove baseline JIT handlers. |
| m_codeBlock->clearExceptionHandlers(); |
| for (unsigned i = 0; i < m_exceptionHandlerOSRExitCallSites.size(); i++) { |
| OSRExitCompilationInfo& info = m_exceptionHandlerOSRExitCallSites[i].exitInfo; |
| if (info.m_replacementDestination.isSet()) { |
| // If this is is *not* set, it means that we already jumped to the OSR exit in pure generated control flow. |
| // i.e, we explicitly emitted an exceptionCheck that we know will be caught in this machine frame. |
| // If this *is set*, it means we will be landing at this code location from genericUnwind from an |
| // exception thrown in a child call frame. |
| CodeLocationLabel<ExceptionHandlerPtrTag> catchLabel = linkBuffer.locationOf<ExceptionHandlerPtrTag>(info.m_replacementDestination); |
| HandlerInfo newExceptionHandler = m_exceptionHandlerOSRExitCallSites[i].baselineExceptionHandler; |
| CallSiteIndex callSite = m_exceptionHandlerOSRExitCallSites[i].callSiteIndex; |
| newExceptionHandler.start = callSite.bits(); |
| newExceptionHandler.end = callSite.bits() + 1; |
| newExceptionHandler.nativeCode = catchLabel; |
| m_codeBlock->appendExceptionHandler(newExceptionHandler); |
| } |
| } |
| |
| if (m_pcToCodeOriginMapBuilder.didBuildMapping()) |
| m_jitCode->common.m_pcToCodeOriginMap = makeUnique<PCToCodeOriginMap>(WTFMove(m_pcToCodeOriginMapBuilder), linkBuffer); |
| |
| m_jitCode->m_linkerIR = LinkerIR(WTFMove(m_graph.m_constantPool)); |
| } |
| |
| void JITCompiler::disassemble(LinkBuffer& linkBuffer) |
| { |
| if (shouldDumpDisassembly()) { |
| m_disassembler->dump(linkBuffer); |
| linkBuffer.didAlreadyDisassemble(); |
| } |
| |
| if (UNLIKELY(m_graph.m_plan.compilation())) |
| m_disassembler->reportToProfiler(m_graph.m_plan.compilation(), linkBuffer); |
| } |
| |
| #if USE(JSVALUE32_64) |
| void* JITCompiler::addressOfDoubleConstant(Node* node) |
| { |
| double value = node->asNumber(); |
| int64_t valueBits = bitwise_cast<int64_t>(value); |
| return m_graph.m_doubleConstantsMap.ensure(valueBits, [&]{ |
| double* addressInConstantPool = m_graph.m_doubleConstants.add(); |
| *addressInConstantPool = value; |
| return addressInConstantPool; |
| }).iterator->value; |
| } |
| #endif |
| |
| void JITCompiler::noticeCatchEntrypoint(BasicBlock& basicBlock, JITCompiler::Label blockHead, LinkBuffer& linkBuffer, Vector<FlushFormat>&& argumentFormats) |
| { |
| RELEASE_ASSERT(basicBlock.isCatchEntrypoint); |
| RELEASE_ASSERT(basicBlock.intersectionOfCFAHasVisited); // An entrypoint is reachable by definition. |
| m_graph.appendCatchEntrypoint(basicBlock.bytecodeBegin, linkBuffer.locationOf<ExceptionHandlerPtrTag>(blockHead), WTFMove(argumentFormats)); |
| } |
| |
| void JITCompiler::noticeOSREntry(BasicBlock& basicBlock, JITCompiler::Label blockHead, LinkBuffer& linkBuffer) |
| { |
| RELEASE_ASSERT(!basicBlock.isCatchEntrypoint); |
| |
| // OSR entry is not allowed into blocks deemed unreachable by control flow analysis. |
| if (!basicBlock.intersectionOfCFAHasVisited) |
| return; |
| |
| OSREntryData entry; |
| entry.m_bytecodeIndex = basicBlock.bytecodeBegin; |
| entry.m_machineCode = linkBuffer.locationOf<OSREntryPtrTag>(blockHead); |
| |
| FixedOperands<AbstractValue> expectedValues(basicBlock.intersectionOfPastValuesAtHead); |
| Vector<OSREntryReshuffling> reshufflings; |
| |
| // Fix the expected values: in our protocol, a dead variable will have an expected |
| // value of (None, []). But the old JIT may stash some values there. So we really |
| // need (Top, TOP). |
| for (size_t argument = 0; argument < basicBlock.variablesAtHead.numberOfArguments(); ++argument) { |
| Node* node = basicBlock.variablesAtHead.argument(argument); |
| if (!node || !node->shouldGenerate()) |
| expectedValues.argument(argument).makeBytecodeTop(); |
| } |
| for (size_t local = 0; local < basicBlock.variablesAtHead.numberOfLocals(); ++local) { |
| Node* node = basicBlock.variablesAtHead.local(local); |
| if (!node || !node->shouldGenerate()) |
| expectedValues.local(local).makeBytecodeTop(); |
| else { |
| VariableAccessData* variable = node->variableAccessData(); |
| entry.m_machineStackUsed.set(variable->machineLocal().toLocal()); |
| |
| switch (variable->flushFormat()) { |
| case FlushedDouble: |
| entry.m_localsForcedDouble.set(local); |
| break; |
| case FlushedInt52: |
| entry.m_localsForcedAnyInt.set(local); |
| break; |
| default: |
| break; |
| } |
| |
| ASSERT(!variable->operand().isTmp()); |
| if (variable->operand().virtualRegister() != variable->machineLocal()) { |
| reshufflings.append( |
| OSREntryReshuffling( |
| variable->operand().virtualRegister().offset(), variable->machineLocal().offset())); |
| } |
| } |
| } |
| |
| entry.m_expectedValues = WTFMove(expectedValues); |
| entry.m_reshufflings = WTFMove(reshufflings); |
| m_osrEntry.append(WTFMove(entry)); |
| } |
| |
| void JITCompiler::appendExceptionHandlingOSRExit(SpeculativeJIT* speculative, ExitKind kind, unsigned eventStreamIndex, CodeOrigin opCatchOrigin, HandlerInfo* exceptionHandler, CallSiteIndex callSite, MacroAssembler::JumpList jumpsToFail) |
| { |
| OSRExit exit(kind, JSValueRegs(), MethodOfGettingAValueProfile(), speculative, eventStreamIndex); |
| exit.m_codeOrigin = opCatchOrigin; |
| exit.m_exceptionHandlerCallSiteIndex = callSite; |
| OSRExitCompilationInfo& exitInfo = appendExitInfo(jumpsToFail); |
| m_osrExit.append(WTFMove(exit)); |
| m_exceptionHandlerOSRExitCallSites.append(ExceptionHandlingOSRExitInfo { exitInfo, *exceptionHandler, callSite }); |
| } |
| |
| void JITCompiler::setEndOfMainPath(CodeOrigin semanticOrigin) |
| { |
| m_pcToCodeOriginMapBuilder.appendItem(labelIgnoringWatchpoints(), semanticOrigin); |
| if (LIKELY(!m_disassembler)) |
| return; |
| m_disassembler->setEndOfMainPath(labelIgnoringWatchpoints()); |
| } |
| |
| void JITCompiler::setEndOfCode() |
| { |
| m_pcToCodeOriginMapBuilder.appendItem(labelIgnoringWatchpoints(), PCToCodeOriginMapBuilder::defaultCodeOrigin()); |
| if (LIKELY(!m_disassembler)) |
| return; |
| m_disassembler->setEndOfCode(labelIgnoringWatchpoints()); |
| } |
| |
| void JITCompiler::makeCatchOSREntryBuffer() |
| { |
| if (m_graph.m_maxLocalsForCatchOSREntry) { |
| uint32_t numberOfLiveLocals = std::max(*m_graph.m_maxLocalsForCatchOSREntry, 1u); // Make sure we always allocate a non-null catchOSREntryBuffer. |
| m_jitCode->common.catchOSREntryBuffer = vm().scratchBufferForSize(sizeof(JSValue) * numberOfLiveLocals); |
| } |
| } |
| |
| void JITCompiler::loadConstant(LinkerIR::Constant index, GPRReg dest) |
| { |
| #if USE(JSVALUE64) |
| loadPtr(Address(GPRInfo::constantsRegister, JITData::offsetOfData() + sizeof(void*) * index), dest); |
| #else |
| UNUSED_PARAM(index); |
| UNUSED_PARAM(dest); |
| RELEASE_ASSERT_NOT_REACHED(); |
| #endif |
| } |
| |
| void JITCompiler::loadLinkableConstant(LinkableConstant constant, GPRReg dest) |
| { |
| constant.materialize(*this, dest); |
| } |
| |
| void JITCompiler::storeLinkableConstant(LinkableConstant constant, Address dest) |
| { |
| constant.store(*this, dest); |
| } |
| |
| JITCompiler::LinkableConstant::LinkableConstant(JITCompiler& jit, JSCell* cell) |
| { |
| jit.m_graph.m_plan.weakReferences().addLazily(cell); |
| if (jit.m_graph.m_plan.isUnlinked()) { |
| m_index = jit.addToConstantPool(LinkerIR::Type::CellPointer, cell); |
| return; |
| } |
| m_pointer = cell; |
| } |
| |
| JITCompiler::LinkableConstant::LinkableConstant(JITCompiler& jit, void* pointer, NonCellTag) |
| { |
| if (jit.m_graph.m_plan.isUnlinked()) { |
| m_index = jit.addToConstantPool(LinkerIR::Type::NonCellPointer, pointer); |
| return; |
| } |
| m_pointer = pointer; |
| } |
| |
| JITCompiler::LinkableConstant::LinkableConstant(JITCompiler& jit, GlobalObjectTag, CodeOrigin codeOrigin) |
| { |
| if (jit.m_graph.m_plan.isUnlinked()) { |
| m_index = jit.addToConstantPool(LinkerIR::Type::GlobalObject, nullptr); |
| return; |
| } |
| m_pointer = jit.m_graph.globalObjectFor(codeOrigin); |
| } |
| |
| void JITCompiler::LinkableConstant::materialize(CCallHelpers& jit, GPRReg resultGPR) |
| { |
| #if USE(JSVALUE64) |
| if (isUnlinked()) { |
| jit.loadPtr(unlinkedAddress(), resultGPR); |
| return; |
| } |
| #endif |
| jit.move(TrustedImmPtr(m_pointer), resultGPR); |
| } |
| |
| void JITCompiler::LinkableConstant::store(CCallHelpers& jit, CCallHelpers::Address address) |
| { |
| #if USE(JSVALUE64) |
| if (isUnlinked()) { |
| jit.transferPtr(unlinkedAddress(), address); |
| return; |
| } |
| #endif |
| jit.storePtr(TrustedImmPtr(m_pointer), address); |
| } |
| |
| LinkerIR::Constant JITCompiler::addToConstantPool(LinkerIR::Type type, void* payload) |
| { |
| LinkerIR::Value value { payload, type }; |
| auto result = m_graph.m_constantPoolMap.add(value, m_graph.m_constantPoolMap.size()); |
| if (result.isNewEntry) |
| m_graph.m_constantPool.append(value); |
| return result.iterator->value; |
| } |
| |
| std::tuple<CompileTimeStructureStubInfo, JITCompiler::LinkableConstant> JITCompiler::addStructureStubInfo() |
| { |
| if (m_graph.m_plan.isUnlinked()) { |
| void* unlinkedStubInfoIndex = bitwise_cast<void*>(static_cast<uintptr_t>(m_unlinkedStubInfos.size())); |
| UnlinkedStructureStubInfo* stubInfo = &m_unlinkedStubInfos.alloc(); |
| LinkerIR::Constant stubInfoIndex = addToConstantPool(LinkerIR::Type::StructureStubInfo, unlinkedStubInfoIndex); |
| return std::tuple { stubInfo, LinkableConstant(stubInfoIndex) }; |
| } |
| StructureStubInfo* stubInfo = jitCode()->common.m_stubInfos.add(); |
| return std::tuple { stubInfo, LinkableConstant() }; |
| } |
| |
| std::tuple<CompileTimeCallLinkInfo, JITCompiler::LinkableConstant> JITCompiler::addCallLinkInfo(CodeOrigin codeOrigin) |
| { |
| if (m_graph.m_plan.isUnlinked()) { |
| void* unlinkedCallLinkInfoIndex = bitwise_cast<void*>(static_cast<uintptr_t>(m_unlinkedCallLinkInfos.size())); |
| UnlinkedCallLinkInfo* callLinkInfo = &m_unlinkedCallLinkInfos.alloc(); |
| callLinkInfo->codeOrigin = codeOrigin; |
| LinkerIR::Constant callLinkInfoIndex = addToConstantPool(LinkerIR::Type::CallLinkInfo, unlinkedCallLinkInfoIndex); |
| return std::tuple { callLinkInfo, LinkableConstant(callLinkInfoIndex) }; |
| } |
| auto* callLinkInfo = jitCode()->common.m_callLinkInfos.add(codeOrigin, CallLinkInfo::UseDataIC::No); |
| return std::tuple { callLinkInfo, LinkableConstant::nonCellPointer(*this, callLinkInfo) }; |
| } |
| |
| |
| } } // namespace JSC::DFG |
| |
| #endif // ENABLE(DFG_JIT) |