| /* |
| * Copyright (C) 2012-2018 Apple Inc. All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY |
| * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
| * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR |
| * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
| * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
| * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
| * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
| * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #pragma once |
| |
| #include "BaselineJITRegisters.h" |
| #include "CallFrame.h" |
| #include "CallFrameShuffleData.h" |
| #include "CallLinkInfoBase.h" |
| #include "CallMode.h" |
| #include "CodeLocation.h" |
| #include "CodeOrigin.h" |
| #include "CodeSpecializationKind.h" |
| #include "PolymorphicCallStubRoutine.h" |
| #include "WriteBarrier.h" |
| #include <wtf/ScopedLambda.h> |
| #include <wtf/SentinelLinkedList.h> |
| |
| namespace JSC { |
| namespace DFG { |
| struct UnlinkedCallLinkInfo; |
| } |
| |
| class CCallHelpers; |
| class ExecutableBase; |
| class FunctionCodeBlock; |
| class JSFunction; |
| class OptimizingCallLinkInfo; |
| class PolymorphicCallStubRoutine; |
| enum OpcodeID : unsigned; |
| |
| struct CallFrameShuffleData; |
| struct UnlinkedCallLinkInfo; |
| struct BaselineUnlinkedCallLinkInfo; |
| |
| |
| using CompileTimeCallLinkInfo = std::variant<OptimizingCallLinkInfo*, BaselineUnlinkedCallLinkInfo*, DFG::UnlinkedCallLinkInfo*>; |
| |
| class CallLinkInfo : public CallLinkInfoBase { |
| public: |
| friend class LLIntOffsetsExtractor; |
| |
| static constexpr uint8_t maxProfiledArgumentCountIncludingThisForVarargs = UINT8_MAX; |
| |
| enum class Type : uint8_t { |
| Baseline, |
| Optimizing, |
| }; |
| |
| enum class Mode : uint8_t { |
| Init, |
| Monomorphic, |
| Polymorphic, |
| Virtual, |
| }; |
| |
| static constexpr uintptr_t polymorphicCalleeMask = 1; |
| |
| static CallType callTypeFor(OpcodeID opcodeID); |
| |
| static bool isVarargsCallType(CallType callType) |
| { |
| switch (callType) { |
| case CallVarargs: |
| case ConstructVarargs: |
| case TailCallVarargs: |
| return true; |
| |
| default: |
| return false; |
| } |
| } |
| |
| ~CallLinkInfo(); |
| |
| static CodeSpecializationKind specializationKindFor(CallType callType) |
| { |
| return specializationFromIsConstruct(callType == Construct || callType == ConstructVarargs || callType == DirectConstruct); |
| } |
| CodeSpecializationKind specializationKind() const |
| { |
| return specializationKindFor(static_cast<CallType>(m_callType)); |
| } |
| |
| CallMode callMode() const |
| { |
| return callModeFor(static_cast<CallType>(m_callType)); |
| } |
| |
| bool isTailCall() const |
| { |
| return callMode() == CallMode::Tail; |
| } |
| |
| NearCallMode nearCallMode() const |
| { |
| return isTailCall() ? NearCallMode::Tail : NearCallMode::Regular; |
| } |
| |
| bool isVarargs() const |
| { |
| return isVarargsCallType(static_cast<CallType>(m_callType)); |
| } |
| |
| bool isLinked() const { return mode() != Mode::Init && mode() != Mode::Virtual; } |
| void unlinkOrUpgradeImpl(VM&, CodeBlock* oldCodeBlock, CodeBlock* newCodeBlock); |
| |
| #if ENABLE(JIT) |
| protected: |
| static std::tuple<MacroAssembler::JumpList, MacroAssembler::Label> emitFastPathImpl(CallLinkInfo*, CCallHelpers&, UseDataIC, bool isTailCall, ScopedLambda<void()>&& prepareForTailCall) WARN_UNUSED_RETURN; |
| static std::tuple<MacroAssembler::JumpList, MacroAssembler::Label> emitDataICFastPath(CCallHelpers&) WARN_UNUSED_RETURN; |
| static std::tuple<MacroAssembler::JumpList, MacroAssembler::Label> emitTailCallDataICFastPath(CCallHelpers&, ScopedLambda<void()>&& prepareForTailCall) WARN_UNUSED_RETURN; |
| |
| static void emitSlowPathImpl(VM&, CCallHelpers&, UseDataIC, bool isTailCall, MacroAssembler::Label); |
| static void emitDataICSlowPath(VM&, CCallHelpers&, bool isTailCall, MacroAssembler::Label); |
| |
| public: |
| static std::tuple<MacroAssembler::JumpList, MacroAssembler::Label> emitFastPath(CCallHelpers&, CompileTimeCallLinkInfo) WARN_UNUSED_RETURN; |
| static std::tuple<MacroAssembler::JumpList, MacroAssembler::Label> emitTailCallFastPath(CCallHelpers&, CompileTimeCallLinkInfo, ScopedLambda<void()>&& prepareForTailCall) WARN_UNUSED_RETURN; |
| static void emitSlowPath(VM&, CCallHelpers&, CompileTimeCallLinkInfo); |
| static void emitTailCallSlowPath(VM&, CCallHelpers&, CompileTimeCallLinkInfo, MacroAssembler::Label); |
| #endif |
| |
| void revertCallToStub(); |
| |
| bool isDataIC() const { return useDataIC() == UseDataIC::Yes; } |
| UseDataIC useDataIC() const { return static_cast<UseDataIC>(m_useDataIC); } |
| |
| bool allowStubs() const { return m_allowStubs; } |
| |
| void disallowStubs() |
| { |
| m_allowStubs = false; |
| } |
| |
| CodeLocationLabel<JSInternalPtrTag> doneLocation(); |
| |
| void setMonomorphicCallee(VM&, JSCell*, JSObject* callee, CodeBlock*, CodePtr<JSEntryPtrTag>); |
| void clearCallee(); |
| JSObject* callee(); |
| |
| void setLastSeenCallee(VM&, const JSCell* owner, JSObject* callee); |
| JSObject* lastSeenCallee() const; |
| bool haveLastSeenCallee() const; |
| |
| void setExecutableDuringCompilation(ExecutableBase*); |
| ExecutableBase* executable(); |
| |
| #if ENABLE(JIT) |
| void setStub(Ref<PolymorphicCallStubRoutine>&&); |
| #endif |
| void clearStub(); |
| |
| void setVirtualCall(VM&); |
| |
| void revertCall(VM&); |
| |
| PolymorphicCallStubRoutine* stub() const |
| { |
| #if ENABLE(JIT) |
| return m_stub.get(); |
| #else |
| return nullptr; |
| #endif |
| } |
| |
| bool seenOnce() |
| { |
| return m_hasSeenShouldRepatch; |
| } |
| |
| void clearSeen() |
| { |
| m_hasSeenShouldRepatch = false; |
| } |
| |
| void setSeen() |
| { |
| m_hasSeenShouldRepatch = true; |
| } |
| |
| bool hasSeenClosure() |
| { |
| return m_hasSeenClosure; |
| } |
| |
| void setHasSeenClosure() |
| { |
| m_hasSeenClosure = true; |
| } |
| |
| bool clearedByGC() |
| { |
| return m_clearedByGC; |
| } |
| |
| bool clearedByVirtual() |
| { |
| return m_clearedByVirtual; |
| } |
| |
| void setClearedByVirtual() |
| { |
| m_clearedByVirtual = true; |
| } |
| |
| void setCallType(CallType callType) |
| { |
| m_callType = callType; |
| } |
| |
| CallType callType() |
| { |
| return static_cast<CallType>(m_callType); |
| } |
| |
| static ptrdiff_t offsetOfMaxArgumentCountIncludingThisForVarargs() |
| { |
| return OBJECT_OFFSETOF(CallLinkInfo, m_maxArgumentCountIncludingThisForVarargs); |
| } |
| |
| uint32_t maxArgumentCountIncludingThisForVarargs() |
| { |
| return m_maxArgumentCountIncludingThisForVarargs; |
| } |
| |
| void updateMaxArgumentCountIncludingThisForVarargs(unsigned argumentCountIncludingThisForVarargs) |
| { |
| if (m_maxArgumentCountIncludingThisForVarargs < argumentCountIncludingThisForVarargs) |
| m_maxArgumentCountIncludingThisForVarargs = std::min<unsigned>(argumentCountIncludingThisForVarargs, maxProfiledArgumentCountIncludingThisForVarargs); |
| } |
| |
| static ptrdiff_t offsetOfSlowPathCount() |
| { |
| return OBJECT_OFFSETOF(CallLinkInfo, m_slowPathCount); |
| } |
| |
| static ptrdiff_t offsetOfCallee() |
| { |
| return OBJECT_OFFSETOF(CallLinkInfo, m_callee); |
| } |
| |
| static ptrdiff_t offsetOfCodeBlock() |
| { |
| return OBJECT_OFFSETOF(CallLinkInfo, u) + OBJECT_OFFSETOF(UnionType, dataIC.m_codeBlock); |
| } |
| |
| static ptrdiff_t offsetOfMonomorphicCallDestination() |
| { |
| return OBJECT_OFFSETOF(CallLinkInfo, u) + OBJECT_OFFSETOF(UnionType, dataIC.m_monomorphicCallDestination); |
| } |
| |
| #if ENABLE(JIT) |
| static ptrdiff_t offsetOfStub() |
| { |
| return OBJECT_OFFSETOF(CallLinkInfo, m_stub); |
| } |
| #endif |
| |
| uint32_t slowPathCount() |
| { |
| return m_slowPathCount; |
| } |
| |
| CodeOrigin codeOrigin() const; |
| |
| template<typename Functor> |
| void forEachDependentCell(const Functor& functor) const |
| { |
| if (isLinked()) { |
| if (stub()) { |
| #if ENABLE(JIT) |
| stub()->forEachDependentCell(functor); |
| #else |
| RELEASE_ASSERT_NOT_REACHED(); |
| #endif |
| } else |
| functor(m_callee.get()); |
| } |
| if (haveLastSeenCallee()) |
| functor(lastSeenCallee()); |
| } |
| |
| void visitWeak(VM&); |
| |
| Type type() const { return static_cast<Type>(m_type); } |
| |
| Mode mode() const { return static_cast<Mode>(m_mode); } |
| |
| JSCell* owner() const { return m_owner; } |
| |
| JSCell* ownerForSlowPath(CallFrame* calleeFrame); |
| |
| JSGlobalObject* globalObjectForSlowPath(JSCell* owner); |
| |
| std::tuple<CodeBlock*, BytecodeIndex> retrieveCaller(JSCell* owner); |
| |
| protected: |
| CallLinkInfo(Type type, UseDataIC useDataIC, JSCell* owner) |
| : CallLinkInfoBase(CallSiteType::CallLinkInfo) |
| , m_useDataIC(static_cast<unsigned>(useDataIC)) |
| , m_type(static_cast<unsigned>(type)) |
| , m_owner(owner) |
| { |
| ASSERT(type == this->type()); |
| ASSERT(useDataIC == this->useDataIC()); |
| } |
| |
| void reset(VM&); |
| |
| bool m_hasSeenShouldRepatch : 1 { false }; |
| bool m_hasSeenClosure : 1 { false }; |
| bool m_clearedByGC : 1 { false }; |
| bool m_clearedByVirtual : 1 { false }; |
| bool m_allowStubs : 1 { true }; |
| unsigned m_callType : 4 { CallType::None }; // CallType |
| unsigned m_useDataIC : 1; // UseDataIC |
| unsigned m_type : 1; // Type |
| unsigned m_mode : 3 { static_cast<unsigned>(Mode::Init) }; // Mode |
| uint8_t m_maxArgumentCountIncludingThisForVarargs { 0 }; // For varargs: the profiled maximum number of arguments. For direct: the number of stack slots allocated for arguments. |
| uint32_t m_slowPathCount { 0 }; |
| |
| CodeLocationLabel<JSInternalPtrTag> m_doneLocation; |
| union UnionType { |
| UnionType() |
| : dataIC { nullptr, nullptr } |
| { } |
| |
| struct DataIC { |
| CodeBlock* m_codeBlock; // This is weakly held. And cleared whenever m_monomorphicCallDestination is changed. |
| CodePtr<JSEntryPtrTag> m_monomorphicCallDestination; |
| } dataIC; |
| |
| struct { |
| CodeLocationDataLabelPtr<JSInternalPtrTag> m_codeBlockLocation; |
| CodeLocationDataLabelPtr<JSInternalPtrTag> m_calleeLocation; |
| } codeIC; |
| } u; |
| |
| WriteBarrier<JSObject> m_callee; |
| WriteBarrier<JSObject> m_lastSeenCallee; |
| #if ENABLE(JIT) |
| RefPtr<PolymorphicCallStubRoutine> m_stub; |
| #endif |
| JSCell* m_owner { nullptr }; |
| }; |
| |
| class BaselineCallLinkInfo final : public CallLinkInfo { |
| public: |
| BaselineCallLinkInfo() |
| : CallLinkInfo(Type::Baseline, UseDataIC::Yes, nullptr) |
| { |
| } |
| |
| void initialize(VM&, CodeBlock*, CallType, BytecodeIndex); |
| |
| void setCodeLocations(CodeLocationLabel<JSInternalPtrTag> doneLocation) |
| { |
| m_doneLocation = doneLocation; |
| } |
| |
| CodeOrigin codeOrigin() const { return CodeOrigin { m_bytecodeIndex }; } |
| |
| private: |
| BytecodeIndex m_bytecodeIndex { }; |
| }; |
| |
| inline CodeOrigin getCallLinkInfoCodeOrigin(CallLinkInfo& callLinkInfo) |
| { |
| return callLinkInfo.codeOrigin(); |
| } |
| |
| struct UnlinkedCallLinkInfo { |
| CodeLocationLabel<JSInternalPtrTag> doneLocation; |
| |
| void setCodeLocations(CodeLocationLabel<JSInternalPtrTag> doneLocation) |
| { |
| this->doneLocation = doneLocation; |
| } |
| }; |
| |
| struct BaselineUnlinkedCallLinkInfo : public JSC::UnlinkedCallLinkInfo { |
| BytecodeIndex bytecodeIndex; // Currently, only used by baseline, so this can trivially produce a CodeOrigin. |
| |
| #if ENABLE(JIT) |
| void setUpCall(CallLinkInfo::CallType) { } |
| #endif |
| }; |
| |
| #if ENABLE(JIT) |
| |
| class DirectCallLinkInfo final : public CallLinkInfoBase { |
| WTF_MAKE_NONCOPYABLE(DirectCallLinkInfo); |
| public: |
| DirectCallLinkInfo(CodeOrigin codeOrigin, UseDataIC useDataIC, JSCell* owner, ExecutableBase* executable) |
| : CallLinkInfoBase(CallSiteType::DirectCall) |
| , m_useDataIC(useDataIC) |
| , m_codeOrigin(codeOrigin) |
| , m_owner(owner) |
| , m_executable(executable) |
| { } |
| |
| ~DirectCallLinkInfo() |
| { |
| m_target = { }; |
| m_codeBlock = nullptr; |
| } |
| |
| void setCallType(CallType callType) |
| { |
| m_callType = callType; |
| } |
| |
| CallType callType() |
| { |
| return static_cast<CallType>(m_callType); |
| } |
| |
| CallMode callMode() const |
| { |
| return callModeFor(static_cast<CallType>(m_callType)); |
| } |
| |
| bool isTailCall() const |
| { |
| return callMode() == CallMode::Tail; |
| } |
| |
| CodeSpecializationKind specializationKind() const |
| { |
| auto callType = static_cast<CallType>(m_callType); |
| return specializationFromIsConstruct(callType == DirectConstruct); |
| } |
| |
| void setCodeLocations(CodeLocationLabel<JSInternalPtrTag> slowPathStart) |
| { |
| m_slowPathStart = slowPathStart; |
| } |
| |
| static ptrdiff_t offsetOfTarget() { return OBJECT_OFFSETOF(DirectCallLinkInfo, m_target); }; |
| static ptrdiff_t offsetOfCodeBlock() { return OBJECT_OFFSETOF(DirectCallLinkInfo, m_codeBlock); }; |
| |
| JSCell* owner() const { return m_owner; } |
| |
| void unlinkOrUpgradeImpl(VM&, CodeBlock* oldCodeBlock, CodeBlock* newCodeBlock); |
| |
| void visitWeak(VM&); |
| |
| CodeOrigin codeOrigin() const { return m_codeOrigin; } |
| bool isDataIC() const { return m_useDataIC == UseDataIC::Yes; } |
| |
| MacroAssembler::JumpList emitDirectFastPath(CCallHelpers&); |
| MacroAssembler::JumpList emitDirectTailCallFastPath(CCallHelpers&, ScopedLambda<void()>&& prepareForTailCall); |
| void setCallTarget(CodeBlock*, CodeLocationLabel<JSEntryPtrTag>); |
| void setMaxArgumentCountIncludingThis(unsigned); |
| unsigned maxArgumentCountIncludingThis() const { return m_maxArgumentCountIncludingThis; } |
| |
| void reset(); |
| |
| void validateSpeculativeRepatchOnMainThread(VM&); |
| |
| private: |
| CodeLocationLabel<JSInternalPtrTag> slowPathStart() const { return m_slowPathStart; } |
| CodeLocationLabel<JSInternalPtrTag> fastPathStart() const { return m_fastPathStart; } |
| |
| void initialize(); |
| void repatchSpeculatively(); |
| |
| std::tuple<CodeBlock*, CodePtr<JSEntryPtrTag>> retrieveCallInfo(FunctionExecutable*); |
| |
| CallType m_callType : 4; |
| UseDataIC m_useDataIC : 1; |
| unsigned m_maxArgumentCountIncludingThis { 0 }; |
| CodePtr<JSEntryPtrTag> m_target; |
| CodeBlock* m_codeBlock { nullptr }; // This is weakly held. And cleared whenever m_target is changed. |
| CodeOrigin m_codeOrigin { }; |
| CodeLocationLabel<JSInternalPtrTag> m_slowPathStart; |
| CodeLocationLabel<JSInternalPtrTag> m_fastPathStart; |
| CodeLocationDataLabelPtr<JSInternalPtrTag> m_codeBlockLocation; |
| CodeLocationNearCall<JSInternalPtrTag> m_callLocation NO_UNIQUE_ADDRESS; |
| JSCell* m_owner; |
| ExecutableBase* m_executable { nullptr }; // This is weakly held. DFG / FTL CommonData already ensures this. |
| }; |
| |
| class OptimizingCallLinkInfo final : public CallLinkInfo { |
| public: |
| friend class CallLinkInfo; |
| |
| OptimizingCallLinkInfo() |
| : CallLinkInfo(Type::Optimizing, UseDataIC::Yes, nullptr) |
| { |
| } |
| |
| OptimizingCallLinkInfo(CodeOrigin codeOrigin, UseDataIC useDataIC, JSCell* owner) |
| : CallLinkInfo(Type::Optimizing, useDataIC, owner) |
| , m_codeOrigin(codeOrigin) |
| { |
| } |
| |
| void setUpCall(CallType callType) |
| { |
| m_callType = callType; |
| } |
| |
| void setCodeLocations(CodeLocationLabel<JSInternalPtrTag> doneLocation) |
| { |
| m_doneLocation = doneLocation; |
| } |
| |
| void setSlowPathCallDestination(CodePtr<JSEntryPtrTag>); |
| |
| CodeOrigin codeOrigin() const { return m_codeOrigin; } |
| |
| void initializeFromDFGUnlinkedCallLinkInfo(VM&, const DFG::UnlinkedCallLinkInfo&, CodeBlock*); |
| |
| static ptrdiff_t offsetOfSlowPathCallDestination() |
| { |
| return OBJECT_OFFSETOF(OptimizingCallLinkInfo, m_slowPathCallDestination); |
| } |
| |
| private: |
| std::tuple<MacroAssembler::JumpList, MacroAssembler::Label> emitFastPath(CCallHelpers&) WARN_UNUSED_RETURN; |
| std::tuple<MacroAssembler::JumpList, MacroAssembler::Label> emitTailCallFastPath(CCallHelpers&, ScopedLambda<void()>&& prepareForTailCall) WARN_UNUSED_RETURN; |
| void emitSlowPath(VM&, CCallHelpers&); |
| void emitTailCallSlowPath(VM&, CCallHelpers&, MacroAssembler::Label); |
| |
| CodeOrigin m_codeOrigin; |
| CodePtr<JSEntryPtrTag> m_slowPathCallDestination; |
| CodeLocationNearCall<JSInternalPtrTag> m_callLocation NO_UNIQUE_ADDRESS; |
| }; |
| |
| #endif |
| |
| inline CodeOrigin CallLinkInfo::codeOrigin() const |
| { |
| switch (type()) { |
| case Type::Baseline: |
| return static_cast<const BaselineCallLinkInfo*>(this)->codeOrigin(); |
| case Type::Optimizing: |
| #if ENABLE(JIT) |
| return static_cast<const OptimizingCallLinkInfo*>(this)->codeOrigin(); |
| #else |
| return { }; |
| #endif |
| } |
| return { }; |
| } |
| |
| inline JSCell* CallLinkInfo::ownerForSlowPath(CallFrame* calleeFrame) |
| { |
| if (m_owner) |
| return m_owner; |
| |
| // Right now, IC (Getter, Setter, Proxy IC etc.) / WasmToJS sets nullptr intentionally since we would like to share IC / WasmToJS thunk eventually. |
| // However, in that case, each IC's data side will have CallLinkInfo. |
| // At that time, they should have appropriate owner. So this is a hack only for now. |
| // This should always works since IC only performs regular-calls and it never does tail-calls. |
| return calleeFrame->callerFrame()->codeOwnerCell(); |
| } |
| |
| } // namespace JSC |