| // Copyright 2013 the V8 project authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #ifndef V8_V8_PLATFORM_H_ |
| #define V8_V8_PLATFORM_H_ |
| |
| #include <math.h> |
| #include <stddef.h> |
| #include <stdint.h> |
| #include <stdlib.h> // For abort. |
| |
| #include <memory> |
| #include <string> |
| |
| #include "v8-source-location.h" // NOLINT(build/include_directory) |
| #include "v8config.h" // NOLINT(build/include_directory) |
| |
| namespace v8 { |
| |
| class Isolate; |
| |
| // Valid priorities supported by the task scheduling infrastructure. |
| enum class TaskPriority : uint8_t { |
| /** |
| * Best effort tasks are not critical for performance of the application. The |
| * platform implementation should preempt such tasks if higher priority tasks |
| * arrive. |
| */ |
| kBestEffort, |
| /** |
| * User visible tasks are long running background tasks that will |
| * improve performance and memory usage of the application upon completion. |
| * Example: background compilation and garbage collection. |
| */ |
| kUserVisible, |
| /** |
| * User blocking tasks are highest priority tasks that block the execution |
| * thread (e.g. major garbage collection). They must be finished as soon as |
| * possible. |
| */ |
| kUserBlocking, |
| }; |
| |
| /** |
| * A Task represents a unit of work. |
| */ |
| class Task { |
| public: |
| virtual ~Task() = default; |
| |
| virtual void Run() = 0; |
| }; |
| |
| /** |
| * An IdleTask represents a unit of work to be performed in idle time. |
| * The Run method is invoked with an argument that specifies the deadline in |
| * seconds returned by MonotonicallyIncreasingTime(). |
| * The idle task is expected to complete by this deadline. |
| */ |
| class IdleTask { |
| public: |
| virtual ~IdleTask() = default; |
| virtual void Run(double deadline_in_seconds) = 0; |
| }; |
| |
| /** |
| * A TaskRunner allows scheduling of tasks. The TaskRunner may still be used to |
| * post tasks after the isolate gets destructed, but these tasks may not get |
| * executed anymore. All tasks posted to a given TaskRunner will be invoked in |
| * sequence. Tasks can be posted from any thread. |
| */ |
| class TaskRunner { |
| public: |
| /** |
| * Schedules a task to be invoked by this TaskRunner. The TaskRunner |
| * implementation takes ownership of |task|. |
| */ |
| virtual void PostTask(std::unique_ptr<Task> task) = 0; |
| |
| /** |
| * Schedules a task to be invoked by this TaskRunner. The TaskRunner |
| * implementation takes ownership of |task|. The |task| cannot be nested |
| * within other task executions. |
| * |
| * Tasks which shouldn't be interleaved with JS execution must be posted with |
| * |PostNonNestableTask| or |PostNonNestableDelayedTask|. This is because the |
| * embedder may process tasks in a callback which is called during JS |
| * execution. |
| * |
| * In particular, tasks which execute JS must be non-nestable, since JS |
| * execution is not allowed to nest. |
| * |
| * Requires that |TaskRunner::NonNestableTasksEnabled()| is true. |
| */ |
| virtual void PostNonNestableTask(std::unique_ptr<Task> task) {} |
| |
| /** |
| * Schedules a task to be invoked by this TaskRunner. The task is scheduled |
| * after the given number of seconds |delay_in_seconds|. The TaskRunner |
| * implementation takes ownership of |task|. |
| */ |
| virtual void PostDelayedTask(std::unique_ptr<Task> task, |
| double delay_in_seconds) = 0; |
| |
| /** |
| * Schedules a task to be invoked by this TaskRunner. The task is scheduled |
| * after the given number of seconds |delay_in_seconds|. The TaskRunner |
| * implementation takes ownership of |task|. The |task| cannot be nested |
| * within other task executions. |
| * |
| * Tasks which shouldn't be interleaved with JS execution must be posted with |
| * |PostNonNestableTask| or |PostNonNestableDelayedTask|. This is because the |
| * embedder may process tasks in a callback which is called during JS |
| * execution. |
| * |
| * In particular, tasks which execute JS must be non-nestable, since JS |
| * execution is not allowed to nest. |
| * |
| * Requires that |TaskRunner::NonNestableDelayedTasksEnabled()| is true. |
| */ |
| virtual void PostNonNestableDelayedTask(std::unique_ptr<Task> task, |
| double delay_in_seconds) {} |
| |
| /** |
| * Schedules an idle task to be invoked by this TaskRunner. The task is |
| * scheduled when the embedder is idle. Requires that |
| * |TaskRunner::IdleTasksEnabled()| is true. Idle tasks may be reordered |
| * relative to other task types and may be starved for an arbitrarily long |
| * time if no idle time is available. The TaskRunner implementation takes |
| * ownership of |task|. |
| */ |
| virtual void PostIdleTask(std::unique_ptr<IdleTask> task) = 0; |
| |
| /** |
| * Returns true if idle tasks are enabled for this TaskRunner. |
| */ |
| virtual bool IdleTasksEnabled() = 0; |
| |
| /** |
| * Returns true if non-nestable tasks are enabled for this TaskRunner. |
| */ |
| virtual bool NonNestableTasksEnabled() const { return false; } |
| |
| /** |
| * Returns true if non-nestable delayed tasks are enabled for this TaskRunner. |
| */ |
| virtual bool NonNestableDelayedTasksEnabled() const { return false; } |
| |
| TaskRunner() = default; |
| virtual ~TaskRunner() = default; |
| |
| TaskRunner(const TaskRunner&) = delete; |
| TaskRunner& operator=(const TaskRunner&) = delete; |
| }; |
| |
| /** |
| * Delegate that's passed to Job's worker task, providing an entry point to |
| * communicate with the scheduler. |
| */ |
| class JobDelegate { |
| public: |
| /** |
| * Returns true if this thread *must* return from the worker task on the |
| * current thread ASAP. Workers should periodically invoke ShouldYield (or |
| * YieldIfNeeded()) as often as is reasonable. |
| * After this method returned true, ShouldYield must not be called again. |
| */ |
| virtual bool ShouldYield() = 0; |
| |
| /** |
| * Notifies the scheduler that max concurrency was increased, and the number |
| * of worker should be adjusted accordingly. See Platform::PostJob() for more |
| * details. |
| */ |
| virtual void NotifyConcurrencyIncrease() = 0; |
| |
| /** |
| * Returns a task_id unique among threads currently running this job, such |
| * that GetTaskId() < worker count. To achieve this, the same task_id may be |
| * reused by a different thread after a worker_task returns. |
| */ |
| virtual uint8_t GetTaskId() = 0; |
| |
| /** |
| * Returns true if the current task is called from the thread currently |
| * running JobHandle::Join(). |
| */ |
| virtual bool IsJoiningThread() const = 0; |
| }; |
| |
| /** |
| * Handle returned when posting a Job. Provides methods to control execution of |
| * the posted Job. |
| */ |
| class JobHandle { |
| public: |
| virtual ~JobHandle() = default; |
| |
| /** |
| * Notifies the scheduler that max concurrency was increased, and the number |
| * of worker should be adjusted accordingly. See Platform::PostJob() for more |
| * details. |
| */ |
| virtual void NotifyConcurrencyIncrease() = 0; |
| |
| /** |
| * Contributes to the job on this thread. Doesn't return until all tasks have |
| * completed and max concurrency becomes 0. When Join() is called and max |
| * concurrency reaches 0, it should not increase again. This also promotes |
| * this Job's priority to be at least as high as the calling thread's |
| * priority. |
| */ |
| virtual void Join() = 0; |
| |
| /** |
| * Forces all existing workers to yield ASAP. Waits until they have all |
| * returned from the Job's callback before returning. |
| */ |
| virtual void Cancel() = 0; |
| |
| /* |
| * Forces all existing workers to yield ASAP but doesn’t wait for them. |
| * Warning, this is dangerous if the Job's callback is bound to or has access |
| * to state which may be deleted after this call. |
| */ |
| virtual void CancelAndDetach() = 0; |
| |
| /** |
| * Returns true if there's any work pending or any worker running. |
| */ |
| virtual bool IsActive() = 0; |
| |
| /** |
| * Returns true if associated with a Job and other methods may be called. |
| * Returns false after Join() or Cancel() was called. This may return true |
| * even if no workers are running and IsCompleted() returns true |
| */ |
| virtual bool IsValid() = 0; |
| |
| /** |
| * Returns true if job priority can be changed. |
| */ |
| virtual bool UpdatePriorityEnabled() const { return false; } |
| |
| /** |
| * Update this Job's priority. |
| */ |
| virtual void UpdatePriority(TaskPriority new_priority) {} |
| }; |
| |
| /** |
| * A JobTask represents work to run in parallel from Platform::PostJob(). |
| */ |
| class JobTask { |
| public: |
| virtual ~JobTask() = default; |
| |
| virtual void Run(JobDelegate* delegate) = 0; |
| |
| /** |
| * Controls the maximum number of threads calling Run() concurrently, given |
| * the number of threads currently assigned to this job and executing Run(). |
| * Run() is only invoked if the number of threads previously running Run() was |
| * less than the value returned. In general, this should return the latest |
| * number of incomplete work items (smallest unit of work) left to process, |
| * including items that are currently in progress. |worker_count| is the |
| * number of threads currently assigned to this job which some callers may |
| * need to determine their return value. Since GetMaxConcurrency() is a leaf |
| * function, it must not call back any JobHandle methods. |
| */ |
| virtual size_t GetMaxConcurrency(size_t worker_count) const = 0; |
| }; |
| |
| /** |
| * A "blocking call" refers to any call that causes the calling thread to wait |
| * off-CPU. It includes but is not limited to calls that wait on synchronous |
| * file I/O operations: read or write a file from disk, interact with a pipe or |
| * a socket, rename or delete a file, enumerate files in a directory, etc. |
| * Acquiring a low contention lock is not considered a blocking call. |
| */ |
| |
| /** |
| * BlockingType indicates the likelihood that a blocking call will actually |
| * block. |
| */ |
| enum class BlockingType { |
| // The call might block (e.g. file I/O that might hit in memory cache). |
| kMayBlock, |
| // The call will definitely block (e.g. cache already checked and now pinging |
| // server synchronously). |
| kWillBlock |
| }; |
| |
| /** |
| * This class is instantiated with CreateBlockingScope() in every scope where a |
| * blocking call is made and serves as a precise annotation of the scope that |
| * may/will block. May be implemented by an embedder to adjust the thread count. |
| * CPU usage should be minimal within that scope. ScopedBlockingCalls can be |
| * nested. |
| */ |
| class ScopedBlockingCall { |
| public: |
| virtual ~ScopedBlockingCall() = default; |
| }; |
| |
| /** |
| * The interface represents complex arguments to trace events. |
| */ |
| class ConvertableToTraceFormat { |
| public: |
| virtual ~ConvertableToTraceFormat() = default; |
| |
| /** |
| * Append the class info to the provided |out| string. The appended |
| * data must be a valid JSON object. Strings must be properly quoted, and |
| * escaped. There is no processing applied to the content after it is |
| * appended. |
| */ |
| virtual void AppendAsTraceFormat(std::string* out) const = 0; |
| }; |
| |
| /** |
| * V8 Tracing controller. |
| * |
| * Can be implemented by an embedder to record trace events from V8. |
| * |
| * Will become obsolete in Perfetto SDK build (v8_use_perfetto = true). |
| */ |
| class TracingController { |
| public: |
| virtual ~TracingController() = default; |
| |
| // In Perfetto mode, trace events are written using Perfetto's Track Event |
| // API directly without going through the embedder. However, it is still |
| // possible to observe tracing being enabled and disabled. |
| #if !defined(V8_USE_PERFETTO) |
| /** |
| * Called by TRACE_EVENT* macros, don't call this directly. |
| * The name parameter is a category group for example: |
| * TRACE_EVENT0("v8,parse", "V8.Parse") |
| * The pointer returned points to a value with zero or more of the bits |
| * defined in CategoryGroupEnabledFlags. |
| **/ |
| virtual const uint8_t* GetCategoryGroupEnabled(const char* name) { |
| static uint8_t no = 0; |
| return &no; |
| } |
| |
| /** |
| * Adds a trace event to the platform tracing system. These function calls are |
| * usually the result of a TRACE_* macro from trace_event_common.h when |
| * tracing and the category of the particular trace are enabled. It is not |
| * advisable to call these functions on their own; they are really only meant |
| * to be used by the trace macros. The returned handle can be used by |
| * UpdateTraceEventDuration to update the duration of COMPLETE events. |
| */ |
| virtual uint64_t AddTraceEvent( |
| char phase, const uint8_t* category_enabled_flag, const char* name, |
| const char* scope, uint64_t id, uint64_t bind_id, int32_t num_args, |
| const char** arg_names, const uint8_t* arg_types, |
| const uint64_t* arg_values, |
| std::unique_ptr<ConvertableToTraceFormat>* arg_convertables, |
| unsigned int flags) { |
| return 0; |
| } |
| virtual uint64_t AddTraceEventWithTimestamp( |
| char phase, const uint8_t* category_enabled_flag, const char* name, |
| const char* scope, uint64_t id, uint64_t bind_id, int32_t num_args, |
| const char** arg_names, const uint8_t* arg_types, |
| const uint64_t* arg_values, |
| std::unique_ptr<ConvertableToTraceFormat>* arg_convertables, |
| unsigned int flags, int64_t timestamp) { |
| return 0; |
| } |
| |
| /** |
| * Sets the duration field of a COMPLETE trace event. It must be called with |
| * the handle returned from AddTraceEvent(). |
| **/ |
| virtual void UpdateTraceEventDuration(const uint8_t* category_enabled_flag, |
| const char* name, uint64_t handle) {} |
| #endif // !defined(V8_USE_PERFETTO) |
| |
| class TraceStateObserver { |
| public: |
| virtual ~TraceStateObserver() = default; |
| virtual void OnTraceEnabled() = 0; |
| virtual void OnTraceDisabled() = 0; |
| }; |
| |
| /** |
| * Adds tracing state change observer. |
| * Does nothing in Perfetto SDK build (v8_use_perfetto = true). |
| */ |
| virtual void AddTraceStateObserver(TraceStateObserver*) {} |
| |
| /** |
| * Removes tracing state change observer. |
| * Does nothing in Perfetto SDK build (v8_use_perfetto = true). |
| */ |
| virtual void RemoveTraceStateObserver(TraceStateObserver*) {} |
| }; |
| |
| /** |
| * A V8 memory page allocator. |
| * |
| * Can be implemented by an embedder to manage large host OS allocations. |
| */ |
| class PageAllocator { |
| public: |
| virtual ~PageAllocator() = default; |
| |
| /** |
| * Gets the page granularity for AllocatePages and FreePages. Addresses and |
| * lengths for those calls should be multiples of AllocatePageSize(). |
| */ |
| virtual size_t AllocatePageSize() = 0; |
| |
| /** |
| * Gets the page granularity for SetPermissions and ReleasePages. Addresses |
| * and lengths for those calls should be multiples of CommitPageSize(). |
| */ |
| virtual size_t CommitPageSize() = 0; |
| |
| /** |
| * Sets the random seed so that GetRandomMmapAddr() will generate repeatable |
| * sequences of random mmap addresses. |
| */ |
| virtual void SetRandomMmapSeed(int64_t seed) = 0; |
| |
| /** |
| * Returns a randomized address, suitable for memory allocation under ASLR. |
| * The address will be aligned to AllocatePageSize. |
| */ |
| virtual void* GetRandomMmapAddr() = 0; |
| |
| /** |
| * Memory permissions. |
| */ |
| enum Permission { |
| kNoAccess, |
| kRead, |
| kReadWrite, |
| kReadWriteExecute, |
| kReadExecute, |
| // Set this when reserving memory that will later require kReadWriteExecute |
| // permissions. The resulting behavior is platform-specific, currently |
| // this is used to set the MAP_JIT flag on Apple Silicon. |
| // TODO(jkummerow): Remove this when Wasm has a platform-independent |
| // w^x implementation. |
| // TODO(saelo): Remove this once all JIT pages are allocated through the |
| // VirtualAddressSpace API. |
| kNoAccessWillJitLater |
| }; |
| |
| /** |
| * Allocates memory in range with the given alignment and permission. |
| */ |
| virtual void* AllocatePages(void* address, size_t length, size_t alignment, |
| Permission permissions) = 0; |
| |
| /** |
| * Frees memory in a range that was allocated by a call to AllocatePages. |
| */ |
| virtual bool FreePages(void* address, size_t length) = 0; |
| |
| /** |
| * Releases memory in a range that was allocated by a call to AllocatePages. |
| */ |
| virtual bool ReleasePages(void* address, size_t length, |
| size_t new_length) = 0; |
| |
| /** |
| * Sets permissions on pages in an allocated range. |
| */ |
| virtual bool SetPermissions(void* address, size_t length, |
| Permission permissions) = 0; |
| |
| /** |
| * Recommits discarded pages in the given range with given permissions. |
| * Discarded pages must be recommitted with their original permissions |
| * before they are used again. |
| */ |
| virtual bool RecommitPages(void* address, size_t length, |
| Permission permissions) { |
| // TODO(v8:12797): make it pure once it's implemented on Chromium side. |
| return false; |
| } |
| |
| /** |
| * Frees memory in the given [address, address + size) range. address and size |
| * should be operating system page-aligned. The next write to this |
| * memory area brings the memory transparently back. This should be treated as |
| * a hint to the OS that the pages are no longer needed. It does not guarantee |
| * that the pages will be discarded immediately or at all. |
| */ |
| virtual bool DiscardSystemPages(void* address, size_t size) { return true; } |
| |
| /** |
| * Decommits any wired memory pages in the given range, allowing the OS to |
| * reclaim them, and marks the region as inacessible (kNoAccess). The address |
| * range stays reserved and can be accessed again later by changing its |
| * permissions. However, in that case the memory content is guaranteed to be |
| * zero-initialized again. The memory must have been previously allocated by a |
| * call to AllocatePages. Returns true on success, false otherwise. |
| */ |
| virtual bool DecommitPages(void* address, size_t size) = 0; |
| |
| /** |
| * INTERNAL ONLY: This interface has not been stabilised and may change |
| * without notice from one release to another without being deprecated first. |
| */ |
| class SharedMemoryMapping { |
| public: |
| // Implementations are expected to free the shared memory mapping in the |
| // destructor. |
| virtual ~SharedMemoryMapping() = default; |
| virtual void* GetMemory() const = 0; |
| }; |
| |
| /** |
| * INTERNAL ONLY: This interface has not been stabilised and may change |
| * without notice from one release to another without being deprecated first. |
| */ |
| class SharedMemory { |
| public: |
| // Implementations are expected to free the shared memory in the destructor. |
| virtual ~SharedMemory() = default; |
| virtual std::unique_ptr<SharedMemoryMapping> RemapTo( |
| void* new_address) const = 0; |
| virtual void* GetMemory() const = 0; |
| virtual size_t GetSize() const = 0; |
| }; |
| |
| /** |
| * INTERNAL ONLY: This interface has not been stabilised and may change |
| * without notice from one release to another without being deprecated first. |
| * |
| * Reserve pages at a fixed address returning whether the reservation is |
| * possible. The reserved memory is detached from the PageAllocator and so |
| * should not be freed by it. It's intended for use with |
| * SharedMemory::RemapTo, where ~SharedMemoryMapping would free the memory. |
| */ |
| virtual bool ReserveForSharedMemoryMapping(void* address, size_t size) { |
| return false; |
| } |
| |
| /** |
| * INTERNAL ONLY: This interface has not been stabilised and may change |
| * without notice from one release to another without being deprecated first. |
| * |
| * Allocates shared memory pages. Not all PageAllocators need support this and |
| * so this method need not be overridden. |
| * Allocates a new read-only shared memory region of size |length| and copies |
| * the memory at |original_address| into it. |
| */ |
| virtual std::unique_ptr<SharedMemory> AllocateSharedPages( |
| size_t length, const void* original_address) { |
| return {}; |
| } |
| |
| /** |
| * INTERNAL ONLY: This interface has not been stabilised and may change |
| * without notice from one release to another without being deprecated first. |
| * |
| * If not overridden and changed to return true, V8 will not attempt to call |
| * AllocateSharedPages or RemapSharedPages. If overridden, AllocateSharedPages |
| * and RemapSharedPages must also be overridden. |
| */ |
| virtual bool CanAllocateSharedPages() { return false; } |
| }; |
| |
| /** |
| * An allocator that uses per-thread permissions to protect the memory. |
| * |
| * The implementation is platform/hardware specific, e.g. using pkeys on x64. |
| * |
| * INTERNAL ONLY: This interface has not been stabilised and may change |
| * without notice from one release to another without being deprecated first. |
| */ |
| class ThreadIsolatedAllocator { |
| public: |
| virtual ~ThreadIsolatedAllocator() = default; |
| |
| virtual void* Allocate(size_t size) = 0; |
| |
| virtual void Free(void* object) = 0; |
| |
| enum class Type { |
| kPkey, |
| }; |
| |
| virtual Type Type() const = 0; |
| |
| /** |
| * Return the pkey used to implement the thread isolation if Type == kPkey. |
| */ |
| virtual int Pkey() const { return -1; } |
| }; |
| |
| // Opaque type representing a handle to a shared memory region. |
| using PlatformSharedMemoryHandle = intptr_t; |
| static constexpr PlatformSharedMemoryHandle kInvalidSharedMemoryHandle = -1; |
| |
| // Conversion routines from the platform-dependent shared memory identifiers |
| // into the opaque PlatformSharedMemoryHandle type. These use the underlying |
| // types (e.g. unsigned int) instead of the typedef'd ones (e.g. mach_port_t) |
| // to avoid pulling in large OS header files into this header file. Instead, |
| // the users of these routines are expected to include the respecitve OS |
| // headers in addition to this one. |
| #if V8_OS_DARWIN |
| // Convert between a shared memory handle and a mach_port_t referencing a memory |
| // entry object. |
| inline PlatformSharedMemoryHandle SharedMemoryHandleFromMachMemoryEntry( |
| unsigned int port) { |
| return static_cast<PlatformSharedMemoryHandle>(port); |
| } |
| inline unsigned int MachMemoryEntryFromSharedMemoryHandle( |
| PlatformSharedMemoryHandle handle) { |
| return static_cast<unsigned int>(handle); |
| } |
| #elif V8_OS_FUCHSIA |
| // Convert between a shared memory handle and a zx_handle_t to a VMO. |
| inline PlatformSharedMemoryHandle SharedMemoryHandleFromVMO(uint32_t handle) { |
| return static_cast<PlatformSharedMemoryHandle>(handle); |
| } |
| inline uint32_t VMOFromSharedMemoryHandle(PlatformSharedMemoryHandle handle) { |
| return static_cast<uint32_t>(handle); |
| } |
| #elif V8_OS_WIN |
| // Convert between a shared memory handle and a Windows HANDLE to a file mapping |
| // object. |
| inline PlatformSharedMemoryHandle SharedMemoryHandleFromFileMapping( |
| void* handle) { |
| return reinterpret_cast<PlatformSharedMemoryHandle>(handle); |
| } |
| inline void* FileMappingFromSharedMemoryHandle( |
| PlatformSharedMemoryHandle handle) { |
| return reinterpret_cast<void*>(handle); |
| } |
| #else |
| // Convert between a shared memory handle and a file descriptor. |
| inline PlatformSharedMemoryHandle SharedMemoryHandleFromFileDescriptor(int fd) { |
| return static_cast<PlatformSharedMemoryHandle>(fd); |
| } |
| inline int FileDescriptorFromSharedMemoryHandle( |
| PlatformSharedMemoryHandle handle) { |
| return static_cast<int>(handle); |
| } |
| #endif |
| |
| /** |
| * Possible permissions for memory pages. |
| */ |
| enum class PagePermissions { |
| kNoAccess, |
| kRead, |
| kReadWrite, |
| kReadWriteExecute, |
| kReadExecute, |
| }; |
| |
| /** |
| * Class to manage a virtual memory address space. |
| * |
| * This class represents a contiguous region of virtual address space in which |
| * sub-spaces and (private or shared) memory pages can be allocated, freed, and |
| * modified. This interface is meant to eventually replace the PageAllocator |
| * interface, and can be used as an alternative in the meantime. |
| * |
| * This API is not yet stable and may change without notice! |
| */ |
| class VirtualAddressSpace { |
| public: |
| using Address = uintptr_t; |
| |
| VirtualAddressSpace(size_t page_size, size_t allocation_granularity, |
| Address base, size_t size, |
| PagePermissions max_page_permissions) |
| : page_size_(page_size), |
| allocation_granularity_(allocation_granularity), |
| base_(base), |
| size_(size), |
| max_page_permissions_(max_page_permissions) {} |
| |
| virtual ~VirtualAddressSpace() = default; |
| |
| /** |
| * The page size used inside this space. Guaranteed to be a power of two. |
| * Used as granularity for all page-related operations except for allocation, |
| * which use the allocation_granularity(), see below. |
| * |
| * \returns the page size in bytes. |
| */ |
| size_t page_size() const { return page_size_; } |
| |
| /** |
| * The granularity of page allocations and, by extension, of subspace |
| * allocations. This is guaranteed to be a power of two and a multiple of the |
| * page_size(). In practice, this is equal to the page size on most OSes, but |
| * on Windows it is usually 64KB, while the page size is 4KB. |
| * |
| * \returns the allocation granularity in bytes. |
| */ |
| size_t allocation_granularity() const { return allocation_granularity_; } |
| |
| /** |
| * The base address of the address space managed by this instance. |
| * |
| * \returns the base address of this address space. |
| */ |
| Address base() const { return base_; } |
| |
| /** |
| * The size of the address space managed by this instance. |
| * |
| * \returns the size of this address space in bytes. |
| */ |
| size_t size() const { return size_; } |
| |
| /** |
| * The maximum page permissions that pages allocated inside this space can |
| * obtain. |
| * |
| * \returns the maximum page permissions. |
| */ |
| PagePermissions max_page_permissions() const { return max_page_permissions_; } |
| |
| /** |
| * Sets the random seed so that GetRandomPageAddress() will generate |
| * repeatable sequences of random addresses. |
| * |
| * \param The seed for the PRNG. |
| */ |
| virtual void SetRandomSeed(int64_t seed) = 0; |
| |
| /** |
| * Returns a random address inside this address space, suitable for page |
| * allocations hints. |
| * |
| * \returns a random address aligned to allocation_granularity(). |
| */ |
| virtual Address RandomPageAddress() = 0; |
| |
| /** |
| * Allocates private memory pages with the given alignment and permissions. |
| * |
| * \param hint If nonzero, the allocation is attempted to be placed at the |
| * given address first. If that fails, the allocation is attempted to be |
| * placed elsewhere, possibly nearby, but that is not guaranteed. Specifying |
| * zero for the hint always causes this function to choose a random address. |
| * The hint, if specified, must be aligned to the specified alignment. |
| * |
| * \param size The size of the allocation in bytes. Must be a multiple of the |
| * allocation_granularity(). |
| * |
| * \param alignment The alignment of the allocation in bytes. Must be a |
| * multiple of the allocation_granularity() and should be a power of two. |
| * |
| * \param permissions The page permissions of the newly allocated pages. |
| * |
| * \returns the start address of the allocated pages on success, zero on |
| * failure. |
| */ |
| static constexpr Address kNoHint = 0; |
| virtual V8_WARN_UNUSED_RESULT Address |
| AllocatePages(Address hint, size_t size, size_t alignment, |
| PagePermissions permissions) = 0; |
| |
| /** |
| * Frees previously allocated pages. |
| * |
| * This function will terminate the process on failure as this implies a bug |
| * in the client. As such, there is no return value. |
| * |
| * \param address The start address of the pages to free. This address must |
| * have been obtained through a call to AllocatePages. |
| * |
| * \param size The size in bytes of the region to free. This must match the |
| * size passed to AllocatePages when the pages were allocated. |
| */ |
| virtual void FreePages(Address address, size_t size) = 0; |
| |
| /** |
| * Sets permissions of all allocated pages in the given range. |
| * |
| * This operation can fail due to OOM, in which case false is returned. If |
| * the operation fails for a reason other than OOM, this function will |
| * terminate the process as this implies a bug in the client. |
| * |
| * \param address The start address of the range. Must be aligned to |
| * page_size(). |
| * |
| * \param size The size in bytes of the range. Must be a multiple |
| * of page_size(). |
| * |
| * \param permissions The new permissions for the range. |
| * |
| * \returns true on success, false on OOM. |
| */ |
| virtual V8_WARN_UNUSED_RESULT bool SetPagePermissions( |
| Address address, size_t size, PagePermissions permissions) = 0; |
| |
| /** |
| * Creates a guard region at the specified address. |
| * |
| * Guard regions are guaranteed to cause a fault when accessed and generally |
| * do not count towards any memory consumption limits. Further, allocating |
| * guard regions can usually not fail in subspaces if the region does not |
| * overlap with another region, subspace, or page allocation. |
| * |
| * \param address The start address of the guard region. Must be aligned to |
| * the allocation_granularity(). |
| * |
| * \param size The size of the guard region in bytes. Must be a multiple of |
| * the allocation_granularity(). |
| * |
| * \returns true on success, false otherwise. |
| */ |
| virtual V8_WARN_UNUSED_RESULT bool AllocateGuardRegion(Address address, |
| size_t size) = 0; |
| |
| /** |
| * Frees an existing guard region. |
| * |
| * This function will terminate the process on failure as this implies a bug |
| * in the client. As such, there is no return value. |
| * |
| * \param address The start address of the guard region to free. This address |
| * must have previously been used as address parameter in a successful |
| * invocation of AllocateGuardRegion. |
| * |
| * \param size The size in bytes of the guard region to free. This must match |
| * the size passed to AllocateGuardRegion when the region was created. |
| */ |
| virtual void FreeGuardRegion(Address address, size_t size) = 0; |
| |
| /** |
| * Allocates shared memory pages with the given permissions. |
| * |
| * \param hint Placement hint. See AllocatePages. |
| * |
| * \param size The size of the allocation in bytes. Must be a multiple of the |
| * allocation_granularity(). |
| * |
| * \param permissions The page permissions of the newly allocated pages. |
| * |
| * \param handle A platform-specific handle to a shared memory object. See |
| * the SharedMemoryHandleFromX routines above for ways to obtain these. |
| * |
| * \param offset The offset in the shared memory object at which the mapping |
| * should start. Must be a multiple of the allocation_granularity(). |
| * |
| * \returns the start address of the allocated pages on success, zero on |
| * failure. |
| */ |
| virtual V8_WARN_UNUSED_RESULT Address |
| AllocateSharedPages(Address hint, size_t size, PagePermissions permissions, |
| PlatformSharedMemoryHandle handle, uint64_t offset) = 0; |
| |
| /** |
| * Frees previously allocated shared pages. |
| * |
| * This function will terminate the process on failure as this implies a bug |
| * in the client. As such, there is no return value. |
| * |
| * \param address The start address of the pages to free. This address must |
| * have been obtained through a call to AllocateSharedPages. |
| * |
| * \param size The size in bytes of the region to free. This must match the |
| * size passed to AllocateSharedPages when the pages were allocated. |
| */ |
| virtual void FreeSharedPages(Address address, size_t size) = 0; |
| |
| /** |
| * Whether this instance can allocate subspaces or not. |
| * |
| * \returns true if subspaces can be allocated, false if not. |
| */ |
| virtual bool CanAllocateSubspaces() = 0; |
| |
| /* |
| * Allocate a subspace. |
| * |
| * The address space of a subspace stays reserved in the parent space for the |
| * lifetime of the subspace. As such, it is guaranteed that page allocations |
| * on the parent space cannot end up inside a subspace. |
| * |
| * \param hint Hints where the subspace should be allocated. See |
| * AllocatePages() for more details. |
| * |
| * \param size The size in bytes of the subspace. Must be a multiple of the |
| * allocation_granularity(). |
| * |
| * \param alignment The alignment of the subspace in bytes. Must be a multiple |
| * of the allocation_granularity() and should be a power of two. |
| * |
| * \param max_page_permissions The maximum permissions that pages allocated in |
| * the subspace can obtain. |
| * |
| * \returns a new subspace or nullptr on failure. |
| */ |
| virtual std::unique_ptr<VirtualAddressSpace> AllocateSubspace( |
| Address hint, size_t size, size_t alignment, |
| PagePermissions max_page_permissions) = 0; |
| |
| // |
| // TODO(v8) maybe refactor the methods below before stabilizing the API. For |
| // example by combining them into some form of page operation method that |
| // takes a command enum as parameter. |
| // |
| |
| /** |
| * Recommits discarded pages in the given range with given permissions. |
| * Discarded pages must be recommitted with their original permissions |
| * before they are used again. |
| * |
| * \param address The start address of the range. Must be aligned to |
| * page_size(). |
| * |
| * \param size The size in bytes of the range. Must be a multiple |
| * of page_size(). |
| * |
| * \param permissions The permissions for the range that the pages must have. |
| * |
| * \returns true on success, false otherwise. |
| */ |
| virtual V8_WARN_UNUSED_RESULT bool RecommitPages( |
| Address address, size_t size, PagePermissions permissions) = 0; |
| |
| /** |
| * Frees memory in the given [address, address + size) range. address and |
| * size should be aligned to the page_size(). The next write to this memory |
| * area brings the memory transparently back. This should be treated as a |
| * hint to the OS that the pages are no longer needed. It does not guarantee |
| * that the pages will be discarded immediately or at all. |
| * |
| * \returns true on success, false otherwise. Since this method is only a |
| * hint, a successful invocation does not imply that pages have been removed. |
| */ |
| virtual V8_WARN_UNUSED_RESULT bool DiscardSystemPages(Address address, |
| size_t size) { |
| return true; |
| } |
| /** |
| * Decommits any wired memory pages in the given range, allowing the OS to |
| * reclaim them, and marks the region as inacessible (kNoAccess). The address |
| * range stays reserved and can be accessed again later by changing its |
| * permissions. However, in that case the memory content is guaranteed to be |
| * zero-initialized again. The memory must have been previously allocated by a |
| * call to AllocatePages. |
| * |
| * \returns true on success, false otherwise. |
| */ |
| virtual V8_WARN_UNUSED_RESULT bool DecommitPages(Address address, |
| size_t size) = 0; |
| |
| private: |
| const size_t page_size_; |
| const size_t allocation_granularity_; |
| const Address base_; |
| const size_t size_; |
| const PagePermissions max_page_permissions_; |
| }; |
| |
| /** |
| * V8 Allocator used for allocating zone backings. |
| */ |
| class ZoneBackingAllocator { |
| public: |
| using MallocFn = void* (*)(size_t); |
| using FreeFn = void (*)(void*); |
| |
| virtual MallocFn GetMallocFn() const { return ::malloc; } |
| virtual FreeFn GetFreeFn() const { return ::free; } |
| }; |
| |
| /** |
| * Observer used by V8 to notify the embedder about entering/leaving sections |
| * with high throughput of malloc/free operations. |
| */ |
| class HighAllocationThroughputObserver { |
| public: |
| virtual void EnterSection() {} |
| virtual void LeaveSection() {} |
| }; |
| |
| /** |
| * V8 Platform abstraction layer. |
| * |
| * The embedder has to provide an implementation of this interface before |
| * initializing the rest of V8. |
| */ |
| class Platform { |
| public: |
| virtual ~Platform() = default; |
| |
| /** |
| * Allows the embedder to manage memory page allocations. |
| * Returning nullptr will cause V8 to use the default page allocator. |
| */ |
| virtual PageAllocator* GetPageAllocator() = 0; |
| |
| /** |
| * Allows the embedder to provide an allocator that uses per-thread memory |
| * permissions to protect allocations. |
| * Returning nullptr will cause V8 to disable protections that rely on this |
| * feature. |
| */ |
| virtual ThreadIsolatedAllocator* GetThreadIsolatedAllocator() { |
| return nullptr; |
| } |
| |
| /** |
| * Allows the embedder to specify a custom allocator used for zones. |
| */ |
| virtual ZoneBackingAllocator* GetZoneBackingAllocator() { |
| static ZoneBackingAllocator default_allocator; |
| return &default_allocator; |
| } |
| |
| /** |
| * Enables the embedder to respond in cases where V8 can't allocate large |
| * blocks of memory. V8 retries the failed allocation once after calling this |
| * method. On success, execution continues; otherwise V8 exits with a fatal |
| * error. |
| * Embedder overrides of this function must NOT call back into V8. |
| */ |
| virtual void OnCriticalMemoryPressure() {} |
| |
| /** |
| * Gets the max number of worker threads that may be used to execute |
| * concurrent work scheduled for any single TaskPriority by |
| * Call(BlockingTask)OnWorkerThread() or PostJob(). This can be used to |
| * estimate the number of tasks a work package should be split into. A return |
| * value of 0 means that there are no worker threads available. Note that a |
| * value of 0 won't prohibit V8 from posting tasks using |CallOnWorkerThread|. |
| */ |
| virtual int NumberOfWorkerThreads() = 0; |
| |
| /** |
| * Returns a TaskRunner which can be used to post a task on the foreground. |
| * The TaskRunner's NonNestableTasksEnabled() must be true. This function |
| * should only be called from a foreground thread. |
| */ |
| virtual std::shared_ptr<v8::TaskRunner> GetForegroundTaskRunner( |
| Isolate* isolate) = 0; |
| |
| /** |
| * Schedules a task to be invoked on a worker thread. |
| * Embedders should override PostTaskOnWorkerThreadImpl() instead of |
| * CallOnWorkerThread(). |
| * TODO(chromium:1424158): Make non-virtual once embedders are migrated to |
| * PostTaskOnWorkerThreadImpl(). |
| */ |
| virtual void CallOnWorkerThread(std::unique_ptr<Task> task) { |
| PostTaskOnWorkerThreadImpl(TaskPriority::kUserVisible, std::move(task), |
| SourceLocation::Current()); |
| } |
| |
| /** |
| * Schedules a task that blocks the main thread to be invoked with |
| * high-priority on a worker thread. |
| * Embedders should override PostTaskOnWorkerThreadImpl() instead of |
| * CallBlockingTaskOnWorkerThread(). |
| * TODO(chromium:1424158): Make non-virtual once embedders are migrated to |
| * PostTaskOnWorkerThreadImpl(). |
| */ |
| virtual void CallBlockingTaskOnWorkerThread(std::unique_ptr<Task> task) { |
| // Embedders may optionally override this to process these tasks in a high |
| // priority pool. |
| CallOnWorkerThread(std::move(task)); |
| } |
| |
| /** |
| * Schedules a task to be invoked with low-priority on a worker thread. |
| * Embedders should override PostTaskOnWorkerThreadImpl() instead of |
| * CallLowPriorityTaskOnWorkerThread(). |
| * TODO(chromium:1424158): Make non-virtual once embedders are migrated to |
| * PostTaskOnWorkerThreadImpl(). |
| */ |
| virtual void CallLowPriorityTaskOnWorkerThread(std::unique_ptr<Task> task) { |
| // Embedders may optionally override this to process these tasks in a low |
| // priority pool. |
| CallOnWorkerThread(std::move(task)); |
| } |
| |
| /** |
| * Schedules a task to be invoked on a worker thread after |delay_in_seconds| |
| * expires. |
| * Embedders should override PostDelayedTaskOnWorkerThreadImpl() instead of |
| * CallDelayedOnWorkerThread(). |
| * TODO(chromium:1424158): Make non-virtual once embedders are migrated to |
| * PostDelayedTaskOnWorkerThreadImpl(). |
| */ |
| virtual void CallDelayedOnWorkerThread(std::unique_ptr<Task> task, |
| double delay_in_seconds) { |
| PostDelayedTaskOnWorkerThreadImpl(TaskPriority::kUserVisible, |
| std::move(task), delay_in_seconds, |
| SourceLocation::Current()); |
| } |
| |
| /** |
| * Returns true if idle tasks are enabled for the given |isolate|. |
| */ |
| virtual bool IdleTasksEnabled(Isolate* isolate) { return false; } |
| |
| /** |
| * Posts |job_task| to run in parallel. Returns a JobHandle associated with |
| * the Job, which can be joined or canceled. |
| * This avoids degenerate cases: |
| * - Calling CallOnWorkerThread() for each work item, causing significant |
| * overhead. |
| * - Fixed number of CallOnWorkerThread() calls that split the work and might |
| * run for a long time. This is problematic when many components post |
| * "num cores" tasks and all expect to use all the cores. In these cases, |
| * the scheduler lacks context to be fair to multiple same-priority requests |
| * and/or ability to request lower priority work to yield when high priority |
| * work comes in. |
| * A canonical implementation of |job_task| looks like: |
| * class MyJobTask : public JobTask { |
| * public: |
| * MyJobTask(...) : worker_queue_(...) {} |
| * // JobTask: |
| * void Run(JobDelegate* delegate) override { |
| * while (!delegate->ShouldYield()) { |
| * // Smallest unit of work. |
| * auto work_item = worker_queue_.TakeWorkItem(); // Thread safe. |
| * if (!work_item) return; |
| * ProcessWork(work_item); |
| * } |
| * } |
| * |
| * size_t GetMaxConcurrency() const override { |
| * return worker_queue_.GetSize(); // Thread safe. |
| * } |
| * }; |
| * auto handle = PostJob(TaskPriority::kUserVisible, |
| * std::make_unique<MyJobTask>(...)); |
| * handle->Join(); |
| * |
| * PostJob() and methods of the returned JobHandle/JobDelegate, must never be |
| * called while holding a lock that could be acquired by JobTask::Run or |
| * JobTask::GetMaxConcurrency -- that could result in a deadlock. This is |
| * because [1] JobTask::GetMaxConcurrency may be invoked while holding |
| * internal lock (A), hence JobTask::GetMaxConcurrency can only use a lock (B) |
| * if that lock is *never* held while calling back into JobHandle from any |
| * thread (A=>B/B=>A deadlock) and [2] JobTask::Run or |
| * JobTask::GetMaxConcurrency may be invoked synchronously from JobHandle |
| * (B=>JobHandle::foo=>B deadlock). |
| * Embedders should override CreateJobImpl() instead of PostJob(). |
| * TODO(chromium:1424158): Make non-virtual once embedders are migrated to |
| * CreateJobImpl(). |
| */ |
| virtual std::unique_ptr<JobHandle> PostJob( |
| TaskPriority priority, std::unique_ptr<JobTask> job_task) { |
| auto handle = CreateJob(priority, std::move(job_task)); |
| handle->NotifyConcurrencyIncrease(); |
| return handle; |
| } |
| |
| /** |
| * Creates and returns a JobHandle associated with a Job. Unlike PostJob(), |
| * this doesn't immediately schedules |worker_task| to run; the Job is then |
| * scheduled by calling either NotifyConcurrencyIncrease() or Join(). |
| * |
| * A sufficient CreateJob() implementation that uses the default Job provided |
| * in libplatform looks like: |
| * std::unique_ptr<JobHandle> CreateJob( |
| * TaskPriority priority, std::unique_ptr<JobTask> job_task) override { |
| * return v8::platform::NewDefaultJobHandle( |
| * this, priority, std::move(job_task), NumberOfWorkerThreads()); |
| * } |
| * |
| * Embedders should override CreateJobImpl() instead of CreateJob(). |
| * TODO(chromium:1424158): Make non-virtual once embedders are migrated to |
| * CreateJobImpl(). |
| */ |
| virtual std::unique_ptr<JobHandle> CreateJob( |
| TaskPriority priority, std::unique_ptr<JobTask> job_task) { |
| return CreateJobImpl(priority, std::move(job_task), |
| SourceLocation::Current()); |
| } |
| |
| /** |
| * Instantiates a ScopedBlockingCall to annotate a scope that may/will block. |
| */ |
| virtual std::unique_ptr<ScopedBlockingCall> CreateBlockingScope( |
| BlockingType blocking_type) { |
| return nullptr; |
| } |
| |
| /** |
| * Monotonically increasing time in seconds from an arbitrary fixed point in |
| * the past. This function is expected to return at least |
| * millisecond-precision values. For this reason, |
| * it is recommended that the fixed point be no further in the past than |
| * the epoch. |
| **/ |
| virtual double MonotonicallyIncreasingTime() = 0; |
| |
| /** |
| * Current wall-clock time in milliseconds since epoch. Use |
| * CurrentClockTimeMillisHighResolution() when higher precision is |
| * required. |
| */ |
| virtual int64_t CurrentClockTimeMilliseconds() { |
| return floor(CurrentClockTimeMillis()); |
| } |
| |
| /** |
| * This function is deprecated and will be deleted. Use either |
| * CurrentClockTimeMilliseconds() or |
| * CurrentClockTimeMillisecondsHighResolution(). |
| */ |
| virtual double CurrentClockTimeMillis() = 0; |
| |
| /** |
| * Same as CurrentClockTimeMilliseconds(), but with more precision. |
| */ |
| virtual double CurrentClockTimeMillisecondsHighResolution() { |
| return CurrentClockTimeMillis(); |
| } |
| |
| typedef void (*StackTracePrinter)(); |
| |
| /** |
| * Returns a function pointer that print a stack trace of the current stack |
| * on invocation. Disables printing of the stack trace if nullptr. |
| */ |
| virtual StackTracePrinter GetStackTracePrinter() { return nullptr; } |
| |
| /** |
| * Returns an instance of a v8::TracingController. This must be non-nullptr. |
| */ |
| virtual TracingController* GetTracingController() = 0; |
| |
| /** |
| * Tells the embedder to generate and upload a crashdump during an unexpected |
| * but non-critical scenario. |
| */ |
| virtual void DumpWithoutCrashing() {} |
| |
| /** |
| * Allows the embedder to observe sections with high throughput allocation |
| * operations. |
| */ |
| virtual HighAllocationThroughputObserver* |
| GetHighAllocationThroughputObserver() { |
| static HighAllocationThroughputObserver default_observer; |
| return &default_observer; |
| } |
| |
| protected: |
| /** |
| * Default implementation of current wall-clock time in milliseconds |
| * since epoch. Useful for implementing |CurrentClockTimeMillis| if |
| * nothing special needed. |
| */ |
| V8_EXPORT static double SystemClockTimeMillis(); |
| |
| /** |
| * Creates and returns a JobHandle associated with a Job. |
| * TODO(chromium:1424158): Make pure virtual once embedders implement it. |
| */ |
| virtual std::unique_ptr<JobHandle> CreateJobImpl( |
| TaskPriority priority, std::unique_ptr<JobTask> job_task, |
| const SourceLocation& location) { |
| return nullptr; |
| } |
| |
| /** |
| * Schedules a task with |priority| to be invoked on a worker thread. |
| * TODO(chromium:1424158): Make pure virtual once embedders implement it. |
| */ |
| virtual void PostTaskOnWorkerThreadImpl(TaskPriority priority, |
| std::unique_ptr<Task> task, |
| const SourceLocation& location) {} |
| |
| /** |
| * Schedules a task with |priority| to be invoked on a worker thread after |
| * |delay_in_seconds| expires. |
| * TODO(chromium:1424158): Make pure virtual once embedders implement it. |
| */ |
| virtual void PostDelayedTaskOnWorkerThreadImpl( |
| TaskPriority priority, std::unique_ptr<Task> task, |
| double delay_in_seconds, const SourceLocation& location) {} |
| }; |
| |
| } // namespace v8 |
| |
| #endif // V8_V8_PLATFORM_H_ |