blob: 29ac411a360edfd39e917c531a403ccc6b0634ec [file] [log] [blame] [edit]
// Copyright 2019 the V8 project authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#ifndef V8_EXECUTION_STACK_GUARD_H_
#define V8_EXECUTION_STACK_GUARD_H_
#include "include/v8-internal.h"
#include "src/base/atomicops.h"
#include "src/common/globals.h"
namespace v8 {
namespace internal {
class ExecutionAccess;
class InterruptsScope;
class Isolate;
class Object;
class RootVisitor;
// StackGuard contains the handling of the limits that are used to limit the
// number of nested invocations of JavaScript and the stack size used in each
// invocation.
class V8_EXPORT_PRIVATE V8_NODISCARD StackGuard final {
public:
StackGuard(const StackGuard&) = delete;
StackGuard& operator=(const StackGuard&) = delete;
explicit StackGuard(Isolate* isolate) : isolate_(isolate) {}
// Pass the address beyond which the stack should not grow. The stack
// is assumed to grow downwards.
// When executing on the simulator, we set the stack limits to the limits of
// the simulator's stack instead of using {limit}.
void SetStackLimit(uintptr_t limit);
// Try to compare and swap the given jslimit without the ExecutionAccess lock.
// Expects potential concurrent writes of the interrupt limit, and of the
// interrupt limit only.
void SetStackLimitForStackSwitching(uintptr_t limit);
#ifdef USE_SIMULATOR
// The simulator uses a separate JS stack. Limits on the JS stack might have
// to be adjusted in order to reflect overflows of the C stack, because we
// cannot rely on the interleaving of frames on the simulator.
void AdjustStackLimitForSimulator();
// Reset the limit to the real limit after the stack overflow, if any.
void ResetStackLimitForSimulator();
#endif
// Threading support.
char* ArchiveStackGuard(char* to);
char* RestoreStackGuard(char* from);
static int ArchiveSpacePerThread() { return sizeof(ThreadLocal); }
void FreeThreadResources();
// Sets up the default stack guard for this thread.
void InitThread(const ExecutionAccess& lock);
// Code locations that check for interrupts might only handle a subset of the
// available interrupts, expressed as an `InterruptLevel`. These levels are
// also associated with side effects that are allowed for the respective
// level. The levels are inclusive, which is specified using the order in the
// enum. For example, a site that handles `kAnyEffect` will also handle the
// preceding levels.
enum class InterruptLevel { kNoGC, kNoHeapWrites, kAnyEffect };
static constexpr int kNumberOfInterruptLevels = 3;
#define INTERRUPT_LIST(V) \
V(TERMINATE_EXECUTION, TerminateExecution, 0, InterruptLevel::kNoGC) \
V(GC_REQUEST, GC, 1, InterruptLevel::kNoHeapWrites) \
V(INSTALL_CODE, InstallCode, 2, InterruptLevel::kAnyEffect) \
V(INSTALL_BASELINE_CODE, InstallBaselineCode, 3, InterruptLevel::kAnyEffect) \
V(API_INTERRUPT, ApiInterrupt, 4, InterruptLevel::kNoHeapWrites) \
V(DEOPT_MARKED_ALLOCATION_SITES, DeoptMarkedAllocationSites, 5, \
InterruptLevel::kNoHeapWrites) \
V(GROW_SHARED_MEMORY, GrowSharedMemory, 6, InterruptLevel::kNoGC) \
V(LOG_WASM_CODE, LogWasmCode, 7, InterruptLevel::kAnyEffect) \
V(WASM_CODE_GC, WasmCodeGC, 8, InterruptLevel::kNoHeapWrites) \
V(INSTALL_MAGLEV_CODE, InstallMaglevCode, 9, InterruptLevel::kAnyEffect) \
V(GLOBAL_SAFEPOINT, GlobalSafepoint, 10, InterruptLevel::kNoHeapWrites) \
V(START_INCREMENTAL_MARKING, StartIncrementalMarking, 11, \
InterruptLevel::kNoHeapWrites)
#define V(NAME, Name, id, interrupt_level) \
inline bool Check##Name() { return CheckInterrupt(NAME); } \
inline void Request##Name() { RequestInterrupt(NAME); } \
inline void Clear##Name() { ClearInterrupt(NAME); }
INTERRUPT_LIST(V)
#undef V
// Flag used to set the interrupt causes.
enum InterruptFlag : uint32_t {
#define V(NAME, Name, id, interrupt_level) NAME = (1 << id),
INTERRUPT_LIST(V)
#undef V
#define V(NAME, Name, id, interrupt_level) NAME |
ALL_INTERRUPTS = INTERRUPT_LIST(V) 0
#undef V
};
static_assert(InterruptFlag::ALL_INTERRUPTS <
std::numeric_limits<uint32_t>::max());
static constexpr InterruptFlag InterruptLevelMask(InterruptLevel level) {
#define V(NAME, Name, id, interrupt_level) \
| (interrupt_level <= level ? NAME : 0)
return static_cast<InterruptFlag>(0 INTERRUPT_LIST(V));
#undef V
}
uintptr_t climit() {
#ifdef USE_SIMULATOR
return thread_local_.climit();
#else
return thread_local_.jslimit();
#endif
}
uintptr_t jslimit() { return thread_local_.jslimit(); }
// This provides an asynchronous read of the stack limits for the current
// thread. There are no locks protecting this, but it is assumed that you
// have the global V8 lock if you are using multiple V8 threads.
uintptr_t real_climit() {
#ifdef USE_SIMULATOR
return thread_local_.real_climit_;
#else
return thread_local_.real_jslimit_;
#endif
}
uintptr_t real_jslimit() { return thread_local_.real_jslimit_; }
Address address_of_jslimit() {
return reinterpret_cast<Address>(&thread_local_.jslimit_);
}
Address address_of_real_jslimit() {
return reinterpret_cast<Address>(&thread_local_.real_jslimit_);
}
Address address_of_interrupt_request(InterruptLevel level) {
return reinterpret_cast<Address>(
&thread_local_.interrupt_requested_[static_cast<int>(level)]);
}
static constexpr int jslimit_offset() {
return offsetof(StackGuard, thread_local_) +
offsetof(ThreadLocal, jslimit_);
}
static constexpr int real_jslimit_offset() {
return offsetof(StackGuard, thread_local_) +
offsetof(ThreadLocal, real_jslimit_);
}
static constexpr int no_heap_write_interrupt_request_offset() {
return offsetof(StackGuard, thread_local_) +
offsetof(ThreadLocal, interrupt_requested_) +
static_cast<size_t>(InterruptLevel::kNoHeapWrites);
}
// If the stack guard is triggered, but it is not an actual
// stack overflow, then handle the interruption accordingly.
// Only interrupts that match the given `InterruptLevel` will be handled,
// leaving other interrupts pending as if this method had not been called.
Tagged<Object> HandleInterrupts(
InterruptLevel level = InterruptLevel::kAnyEffect);
// Special case of {HandleInterrupts}: checks for termination requests only.
// This is guaranteed to never cause GC, so can be used to interrupt
// long-running computations that are not GC-safe.
bool HasTerminationRequest();
static constexpr int kSizeInBytes = 8 * kSystemPointerSize;
static char* Iterate(RootVisitor* v, char* thread_storage) {
return thread_storage + ArchiveSpacePerThread();
}
private:
bool CheckInterrupt(InterruptFlag flag);
void RequestInterrupt(InterruptFlag flag);
void ClearInterrupt(InterruptFlag flag);
int FetchAndClearInterrupts(InterruptLevel level);
void SetStackLimitInternal(const ExecutionAccess& lock, uintptr_t limit,
uintptr_t jslimit);
// You should hold the ExecutionAccess lock when calling this method.
bool has_pending_interrupts(const ExecutionAccess& lock) {
return thread_local_.interrupt_flags_ != 0;
}
// You should hold the ExecutionAccess lock when calling this method.
inline void update_interrupt_requests_and_stack_limits(
const ExecutionAccess& lock);
#if V8_TARGET_ARCH_64_BIT
static const uintptr_t kInterruptLimit = uintptr_t{0xfffffffffffffffe};
static const uintptr_t kIllegalLimit = uintptr_t{0xfffffffffffffff8};
#else
static const uintptr_t kInterruptLimit = 0xfffffffe;
static const uintptr_t kIllegalLimit = 0xfffffff8;
#endif
void PushInterruptsScope(InterruptsScope* scope);
void PopInterruptsScope();
class ThreadLocal final {
public:
ThreadLocal() {}
void Initialize(Isolate* isolate, const ExecutionAccess& lock);
// The stack limit has two values: the one with the real_ prefix is the
// actual stack limit set for the VM. The one without the real_ prefix has
// the same value as the actual stack limit except when there is an
// interruption (e.g. debug break or preemption) in which case it is lowered
// to make stack checks fail. Both the generated code and the runtime system
// check against the one without the real_ prefix.
// For simulator builds, we also use a separate C++ stack limit.
// Actual JavaScript stack limit set for the VM.
uintptr_t real_jslimit_ = kIllegalLimit;
#ifdef USE_SIMULATOR
// Actual C++ stack limit set for the VM.
uintptr_t real_climit_ = kIllegalLimit;
#else
// Padding to match the missing {real_climit_} field, renamed to make it
// explicit that this field is unused in this configuration. But the padding
// field is needed:
// - To keep the isolate's LinearAllocationArea fields from crossing cache
// lines (see Isolate::CheckIsolateLayout).
// - To ensure that jslimit_offset() is the same in mksnapshot and in V8:
// When cross-compiling V8, mksnapshot's host and target may be different
// even if they are the same for V8, which results in a different value for
// USE_SIMULATOR. Without this padding, this causes the builtins to use the
// wrong jslimit_offset() for stack checks.
uintptr_t padding1_;
#endif
// jslimit_ and climit_ can be read without any lock.
// Writing requires the ExecutionAccess lock, or may be updated with a
// strong compare-and-swap (e.g. for stack-switching).
base::AtomicWord jslimit_ = kIllegalLimit;
#ifdef USE_SIMULATOR
base::AtomicWord climit_ = kIllegalLimit;
#else
// See {padding1_}.
uintptr_t padding2_;
#endif
uintptr_t jslimit() {
return base::bit_cast<uintptr_t>(base::Relaxed_Load(&jslimit_));
}
void set_jslimit(uintptr_t limit) {
return base::Relaxed_Store(&jslimit_,
static_cast<base::AtomicWord>(limit));
}
#ifdef USE_SIMULATOR
uintptr_t climit() {
return base::bit_cast<uintptr_t>(base::Relaxed_Load(&climit_));
}
void set_climit(uintptr_t limit) {
return base::Relaxed_Store(&climit_,
static_cast<base::AtomicWord>(limit));
}
#endif
// Interrupt request bytes can be read without any lock.
// Writing requires the ExecutionAccess lock.
base::Atomic8 interrupt_requested_[kNumberOfInterruptLevels] = {
false, false, false};
void set_interrupt_requested(InterruptLevel level, bool requested) {
base::Relaxed_Store(&interrupt_requested_[static_cast<int>(level)],
requested);
}
bool has_interrupt_requested(InterruptLevel level) {
return base::Relaxed_Load(&interrupt_requested_[static_cast<int>(level)]);
}
InterruptsScope* interrupt_scopes_ = nullptr;
uint32_t interrupt_flags_ = 0;
};
// TODO(isolates): Technically this could be calculated directly from a
// pointer to StackGuard.
Isolate* isolate_;
ThreadLocal thread_local_;
friend class Isolate;
friend class StackLimitCheck;
friend class InterruptsScope;
static_assert(std::is_standard_layout_v<ThreadLocal>);
};
static_assert(StackGuard::kSizeInBytes == sizeof(StackGuard));
} // namespace internal
} // namespace v8
#endif // V8_EXECUTION_STACK_GUARD_H_