| // Copyright 2017 the V8 project authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #ifndef V8_OBJECTS_MAP_INL_H_ |
| #define V8_OBJECTS_MAP_INL_H_ |
| |
| #include "src/heap/heap-write-barrier-inl.h" |
| #include "src/objects/api-callbacks-inl.h" |
| #include "src/objects/cell-inl.h" |
| #include "src/objects/dependent-code.h" |
| #include "src/objects/descriptor-array-inl.h" |
| #include "src/objects/dictionary.h" |
| #include "src/objects/field-type.h" |
| #include "src/objects/instance-type-inl.h" |
| #include "src/objects/js-function-inl.h" |
| #include "src/objects/map-updater.h" |
| #include "src/objects/map.h" |
| #include "src/objects/objects-inl.h" |
| #include "src/objects/property.h" |
| #include "src/objects/prototype-info-inl.h" |
| #include "src/objects/prototype-info.h" |
| #include "src/objects/shared-function-info-inl.h" |
| #include "src/objects/templates-inl.h" |
| #include "src/objects/transitions-inl.h" |
| #include "src/objects/transitions.h" |
| |
| #if V8_ENABLE_WEBASSEMBLY |
| #include "src/wasm/wasm-objects-inl.h" |
| #endif // V8_ENABLE_WEBASSEMBLY |
| |
| // Has to be the last include (doesn't have include guards): |
| #include "src/objects/object-macros.h" |
| |
| namespace v8 { |
| namespace internal { |
| |
| #include "torque-generated/src/objects/map-tq-inl.inc" |
| |
| TQ_OBJECT_CONSTRUCTORS_IMPL(Map) |
| |
| ACCESSORS(Map, instance_descriptors, Tagged<DescriptorArray>, |
| kInstanceDescriptorsOffset) |
| RELAXED_ACCESSORS(Map, instance_descriptors, Tagged<DescriptorArray>, |
| kInstanceDescriptorsOffset) |
| RELEASE_ACQUIRE_ACCESSORS(Map, instance_descriptors, Tagged<DescriptorArray>, |
| kInstanceDescriptorsOffset) |
| |
| // A freshly allocated layout descriptor can be set on an existing map. |
| // We need to use release-store and acquire-load accessor pairs to ensure |
| // that the concurrent marking thread observes initializing stores of the |
| // layout descriptor. |
| WEAK_ACCESSORS(Map, raw_transitions, kTransitionsOrPrototypeInfoOffset) |
| RELEASE_ACQUIRE_WEAK_ACCESSORS(Map, raw_transitions, |
| kTransitionsOrPrototypeInfoOffset) |
| |
| ACCESSORS_CHECKED2(Map, prototype, Tagged<HeapObject>, kPrototypeOffset, true, |
| IsNull(value) || IsJSProxy(value) || IsWasmObject(value) || |
| (IsJSObject(value) && |
| (InWritableSharedSpace(value) || |
| value->map()->is_prototype_map()))) |
| |
| DEF_GETTER(Map, prototype_info, Tagged<Object>) { |
| Tagged<Object> value = |
| TaggedField<Object, kTransitionsOrPrototypeInfoOffset>::load(cage_base, |
| *this); |
| DCHECK(this->is_prototype_map()); |
| return value; |
| } |
| RELEASE_ACQUIRE_ACCESSORS(Map, prototype_info, Tagged<Object>, |
| kTransitionsOrPrototypeInfoOffset) |
| |
| void Map::init_prototype_and_constructor_or_back_pointer(ReadOnlyRoots roots) { |
| Tagged<HeapObject> null = roots.null_value(); |
| TaggedField<HeapObject, |
| kConstructorOrBackPointerOrNativeContextOffset>::store(*this, |
| null); |
| TaggedField<HeapObject, kPrototypeOffset>::store(*this, null); |
| } |
| |
| // |bit_field| fields. |
| // Concurrent access to |has_prototype_slot| and |has_non_instance_prototype| |
| // is explicitly allowlisted here. The former is never modified after the map |
| // is setup but it's being read by concurrent marker when pointer compression |
| // is enabled. The latter bit can be modified on a live objects. |
| BIT_FIELD_ACCESSORS(Map, relaxed_bit_field, has_non_instance_prototype, |
| Map::Bits1::HasNonInstancePrototypeBit) |
| BIT_FIELD_ACCESSORS(Map, relaxed_bit_field, has_prototype_slot, |
| Map::Bits1::HasPrototypeSlotBit) |
| |
| // These are fine to be written as non-atomic since we don't have data races. |
| // However, they have to be read atomically from the background since the |
| // |bit_field| as a whole can mutate when using the above setters. |
| BIT_FIELD_ACCESSORS2(Map, relaxed_bit_field, bit_field, is_callable, |
| Map::Bits1::IsCallableBit) |
| BIT_FIELD_ACCESSORS2(Map, relaxed_bit_field, bit_field, has_named_interceptor, |
| Map::Bits1::HasNamedInterceptorBit) |
| BIT_FIELD_ACCESSORS2(Map, relaxed_bit_field, bit_field, has_indexed_interceptor, |
| Map::Bits1::HasIndexedInterceptorBit) |
| BIT_FIELD_ACCESSORS2(Map, relaxed_bit_field, bit_field, is_undetectable, |
| Map::Bits1::IsUndetectableBit) |
| BIT_FIELD_ACCESSORS2(Map, relaxed_bit_field, bit_field, is_access_check_needed, |
| Map::Bits1::IsAccessCheckNeededBit) |
| BIT_FIELD_ACCESSORS2(Map, relaxed_bit_field, bit_field, is_constructor, |
| Map::Bits1::IsConstructorBit) |
| |
| // |bit_field2| fields. |
| BIT_FIELD_ACCESSORS(Map, bit_field2, new_target_is_base, |
| Map::Bits2::NewTargetIsBaseBit) |
| BIT_FIELD_ACCESSORS(Map, bit_field2, is_immutable_proto, |
| Map::Bits2::IsImmutablePrototypeBit) |
| |
| // |bit_field3| fields. |
| BIT_FIELD_ACCESSORS(Map, relaxed_bit_field3, owns_descriptors, |
| Map::Bits3::OwnsDescriptorsBit) |
| BIT_FIELD_ACCESSORS(Map, release_acquire_bit_field3, is_deprecated, |
| Map::Bits3::IsDeprecatedBit) |
| BIT_FIELD_ACCESSORS(Map, relaxed_bit_field3, is_in_retained_map_list, |
| Map::Bits3::IsInRetainedMapListBit) |
| BIT_FIELD_ACCESSORS(Map, release_acquire_bit_field3, is_prototype_map, |
| Map::Bits3::IsPrototypeMapBit) |
| BIT_FIELD_ACCESSORS(Map, relaxed_bit_field3, is_migration_target, |
| Map::Bits3::IsMigrationTargetBit) |
| BIT_FIELD_ACCESSORS2(Map, relaxed_bit_field3, bit_field3, is_extensible, |
| Map::Bits3::IsExtensibleBit) |
| BIT_FIELD_ACCESSORS(Map, bit_field3, may_have_interesting_properties, |
| Map::Bits3::MayHaveInterestingPropertiesBit) |
| BIT_FIELD_ACCESSORS(Map, relaxed_bit_field3, construction_counter, |
| Map::Bits3::ConstructionCounterBits) |
| |
| DEF_GETTER(Map, GetNamedInterceptor, Tagged<InterceptorInfo>) { |
| DCHECK(has_named_interceptor()); |
| Tagged<FunctionTemplateInfo> info = GetFunctionTemplateInfo(cage_base); |
| return InterceptorInfo::cast(info->GetNamedPropertyHandler(cage_base)); |
| } |
| |
| DEF_GETTER(Map, GetIndexedInterceptor, Tagged<InterceptorInfo>) { |
| DCHECK(has_indexed_interceptor()); |
| Tagged<FunctionTemplateInfo> info = GetFunctionTemplateInfo(cage_base); |
| return InterceptorInfo::cast(info->GetIndexedPropertyHandler(cage_base)); |
| } |
| |
| // static |
| bool Map::IsMostGeneralFieldType(Representation representation, |
| Tagged<FieldType> field_type) { |
| return !representation.IsHeapObject() || IsAny(field_type); |
| } |
| |
| // static |
| bool Map::FieldTypeIsCleared(Representation rep, Tagged<FieldType> type) { |
| return IsNone(type) && rep.IsHeapObject(); |
| } |
| |
| // static |
| bool Map::CanHaveFastTransitionableElementsKind(InstanceType instance_type) { |
| return instance_type == JS_ARRAY_TYPE || |
| instance_type == JS_PRIMITIVE_WRAPPER_TYPE || |
| instance_type == JS_ARGUMENTS_OBJECT_TYPE; |
| } |
| |
| bool Map::CanHaveFastTransitionableElementsKind() const { |
| return CanHaveFastTransitionableElementsKind(instance_type()); |
| } |
| |
| bool Map::IsDetached(Isolate* isolate) const { |
| if (is_prototype_map()) return true; |
| return instance_type() == JS_OBJECT_TYPE && NumberOfOwnDescriptors() > 0 && |
| IsUndefined(GetBackPointer(), isolate); |
| } |
| |
| // static |
| void Map::GeneralizeIfCanHaveTransitionableFastElementsKind( |
| Isolate* isolate, InstanceType instance_type, |
| Representation* representation, Handle<FieldType>* field_type) { |
| if (CanHaveFastTransitionableElementsKind(instance_type)) { |
| // We don't support propagation of field generalization through elements |
| // kind transitions because they are inserted into the transition tree |
| // before field transitions. In order to avoid complexity of handling |
| // such a case we ensure that all maps with transitionable elements kinds |
| // have the most general field representation and type. |
| *field_type = FieldType::Any(isolate); |
| *representation = Representation::Tagged(); |
| } |
| } |
| |
| Handle<Map> Map::Normalize(Isolate* isolate, Handle<Map> fast_map, |
| PropertyNormalizationMode mode, const char* reason) { |
| const bool kUseCache = true; |
| return Normalize(isolate, fast_map, fast_map->elements_kind(), mode, |
| kUseCache, reason); |
| } |
| |
| bool Map::EquivalentToForNormalization(const Tagged<Map> other, |
| PropertyNormalizationMode mode) const { |
| return EquivalentToForNormalization(other, elements_kind(), mode); |
| } |
| |
| bool Map::TooManyFastProperties(StoreOrigin store_origin) const { |
| if (UnusedPropertyFields() != 0) return false; |
| if (is_prototype_map()) return false; |
| if (store_origin == StoreOrigin::kNamed) { |
| int limit = std::max( |
| {v8_flags.max_fast_properties.value(), GetInObjectProperties()}); |
| FieldCounts counts = GetFieldCounts(); |
| // Only count mutable fields so that objects with large numbers of |
| // constant functions do not go to dictionary mode. That would be bad |
| // because such objects have often been used as modules. |
| int external = counts.mutable_count() - GetInObjectProperties(); |
| return external > limit || counts.GetTotal() > kMaxNumberOfDescriptors; |
| } else { |
| int limit = std::max( |
| {v8_flags.fast_properties_soft_limit.value(), GetInObjectProperties()}); |
| int external = |
| NumberOfFields(ConcurrencyMode::kSynchronous) - GetInObjectProperties(); |
| return external > limit; |
| } |
| } |
| |
| Tagged<Name> Map::GetLastDescriptorName(Isolate* isolate) const { |
| return instance_descriptors(isolate)->GetKey(LastAdded()); |
| } |
| |
| PropertyDetails Map::GetLastDescriptorDetails(Isolate* isolate) const { |
| return instance_descriptors(isolate)->GetDetails(LastAdded()); |
| } |
| |
| InternalIndex Map::LastAdded() const { |
| int number_of_own_descriptors = NumberOfOwnDescriptors(); |
| DCHECK_GT(number_of_own_descriptors, 0); |
| return InternalIndex(number_of_own_descriptors - 1); |
| } |
| |
| int Map::NumberOfOwnDescriptors() const { |
| return Bits3::NumberOfOwnDescriptorsBits::decode( |
| release_acquire_bit_field3()); |
| } |
| |
| void Map::SetNumberOfOwnDescriptors(int number) { |
| DCHECK_LE(number, instance_descriptors()->number_of_descriptors()); |
| CHECK_LE(static_cast<unsigned>(number), |
| static_cast<unsigned>(kMaxNumberOfDescriptors)); |
| set_release_acquire_bit_field3( |
| Bits3::NumberOfOwnDescriptorsBits::update(bit_field3(), number)); |
| } |
| |
| InternalIndex::Range Map::IterateOwnDescriptors() const { |
| return InternalIndex::Range(NumberOfOwnDescriptors()); |
| } |
| |
| int Map::EnumLength() const { |
| return Bits3::EnumLengthBits::decode(bit_field3()); |
| } |
| |
| void Map::SetEnumLength(int length) { |
| if (length != kInvalidEnumCacheSentinel) { |
| DCHECK_LE(length, NumberOfOwnDescriptors()); |
| CHECK_LE(static_cast<unsigned>(length), |
| static_cast<unsigned>(kMaxNumberOfDescriptors)); |
| } |
| set_relaxed_bit_field3(Bits3::EnumLengthBits::update(bit_field3(), length)); |
| } |
| |
| Tagged<FixedArrayBase> Map::GetInitialElements() const { |
| Tagged<FixedArrayBase> result; |
| if (has_fast_elements() || has_fast_string_wrapper_elements() || |
| has_any_nonextensible_elements()) { |
| result = GetReadOnlyRoots().empty_fixed_array(); |
| } else if (has_typed_array_or_rab_gsab_typed_array_elements()) { |
| result = GetReadOnlyRoots().empty_byte_array(); |
| } else if (has_dictionary_elements()) { |
| result = GetReadOnlyRoots().empty_slow_element_dictionary(); |
| } else { |
| UNREACHABLE(); |
| } |
| DCHECK(!ObjectInYoungGeneration(result)); |
| return result; |
| } |
| |
| VisitorId Map::visitor_id() const { |
| return static_cast<VisitorId>( |
| RELAXED_READ_BYTE_FIELD(*this, kVisitorIdOffset)); |
| } |
| |
| void Map::set_visitor_id(VisitorId id) { |
| CHECK_LT(static_cast<unsigned>(id), 256); |
| RELAXED_WRITE_BYTE_FIELD(*this, kVisitorIdOffset, static_cast<uint8_t>(id)); |
| } |
| |
| int Map::instance_size_in_words() const { |
| return RELAXED_READ_BYTE_FIELD(*this, kInstanceSizeInWordsOffset); |
| } |
| |
| void Map::set_instance_size_in_words(int value) { |
| RELAXED_WRITE_BYTE_FIELD(*this, kInstanceSizeInWordsOffset, |
| static_cast<uint8_t>(value)); |
| } |
| |
| int Map::instance_size() const { |
| return instance_size_in_words() << kTaggedSizeLog2; |
| } |
| |
| void Map::set_instance_size(int size_in_bytes) { |
| CHECK(IsAligned(size_in_bytes, kTaggedSize)); |
| DCHECK_LE(static_cast<unsigned>(size_in_bytes), JSObject::kMaxInstanceSize); |
| int size_in_words = size_in_bytes >>= kTaggedSizeLog2; |
| CHECK_LE(static_cast<unsigned>(size_in_words), kMaxUInt8); |
| set_instance_size_in_words(size_in_words); |
| } |
| |
| int Map::inobject_properties_start_or_constructor_function_index() const { |
| // TODO(solanes, v8:7790, v8:11353): Make this and the setter non-atomic |
| // when TSAN sees the map's store synchronization. |
| return RELAXED_READ_BYTE_FIELD( |
| *this, kInobjectPropertiesStartOrConstructorFunctionIndexOffset); |
| } |
| |
| void Map::set_inobject_properties_start_or_constructor_function_index( |
| int value) { |
| CHECK_LE(static_cast<unsigned>(value), kMaxUInt8); |
| RELAXED_WRITE_BYTE_FIELD( |
| *this, kInobjectPropertiesStartOrConstructorFunctionIndexOffset, |
| static_cast<uint8_t>(value)); |
| } |
| |
| int Map::GetInObjectPropertiesStartInWords() const { |
| DCHECK(IsJSObjectMap(*this)); |
| return inobject_properties_start_or_constructor_function_index(); |
| } |
| |
| void Map::SetInObjectPropertiesStartInWords(int value) { |
| CHECK(IsJSObjectMap(*this)); |
| set_inobject_properties_start_or_constructor_function_index(value); |
| } |
| |
| bool Map::HasOutOfObjectProperties() const { |
| bool ret = used_or_unused_instance_size_in_words() < JSObject::kFieldsAdded; |
| DCHECK_EQ(ret, GetInObjectProperties() < |
| NumberOfFields(ConcurrencyMode::kSynchronous)); |
| return ret; |
| } |
| |
| int Map::GetInObjectProperties() const { |
| DCHECK(IsJSObjectMap(*this)); |
| return instance_size_in_words() - GetInObjectPropertiesStartInWords(); |
| } |
| |
| int Map::GetConstructorFunctionIndex() const { |
| #if V8_ENABLE_WEBASSEMBLY |
| // We allow WasmNull here so builtins can produce error messages when |
| // called from Wasm, without having to special-case WasmNull at every |
| // caller of such a builtin. |
| DCHECK(IsPrimitiveMap(*this) || instance_type() == WASM_NULL_TYPE); |
| #else |
| DCHECK(IsPrimitiveMap(*this)); |
| #endif |
| return inobject_properties_start_or_constructor_function_index(); |
| } |
| |
| void Map::SetConstructorFunctionIndex(int value) { |
| CHECK(IsPrimitiveMap(*this)); |
| set_inobject_properties_start_or_constructor_function_index(value); |
| } |
| |
| int Map::GetInObjectPropertyOffset(int index) const { |
| return (GetInObjectPropertiesStartInWords() + index) * kTaggedSize; |
| } |
| |
| Handle<Map> Map::AddMissingTransitionsForTesting( |
| Isolate* isolate, Handle<Map> split_map, |
| Handle<DescriptorArray> descriptors) { |
| return AddMissingTransitions(isolate, split_map, descriptors); |
| } |
| |
| InstanceType Map::instance_type() const { |
| // TODO(solanes, v8:7790, v8:11353, v8:11945): Make this and the setter |
| // non-atomic when TSAN sees the map's store synchronization. |
| return static_cast<InstanceType>( |
| RELAXED_READ_UINT16_FIELD(*this, kInstanceTypeOffset)); |
| } |
| |
| void Map::set_instance_type(InstanceType value) { |
| RELAXED_WRITE_UINT16_FIELD(*this, kInstanceTypeOffset, value); |
| } |
| |
| int Map::UnusedPropertyFields() const { |
| int value = used_or_unused_instance_size_in_words(); |
| DCHECK_IMPLIES(!IsJSObjectMap(*this), value == 0); |
| int unused; |
| if (value >= JSObject::kFieldsAdded) { |
| unused = instance_size_in_words() - value; |
| } else { |
| // For out of object properties "used_or_unused_instance_size_in_words" |
| // byte encodes the slack in the property array. |
| unused = value; |
| } |
| return unused; |
| } |
| |
| int Map::UnusedInObjectProperties() const { |
| // Like Map::UnusedPropertyFields(), but returns 0 for out of object |
| // properties. |
| int value = used_or_unused_instance_size_in_words(); |
| DCHECK_IMPLIES(!IsJSObjectMap(*this), value == 0); |
| if (value >= JSObject::kFieldsAdded) { |
| return instance_size_in_words() - value; |
| } |
| return 0; |
| } |
| |
| int Map::used_or_unused_instance_size_in_words() const { |
| return RELAXED_READ_BYTE_FIELD(*this, kUsedOrUnusedInstanceSizeInWordsOffset); |
| } |
| |
| void Map::set_used_or_unused_instance_size_in_words(int value) { |
| CHECK_LE(static_cast<unsigned>(value), 255); |
| RELAXED_WRITE_BYTE_FIELD(*this, kUsedOrUnusedInstanceSizeInWordsOffset, |
| static_cast<uint8_t>(value)); |
| } |
| |
| int Map::UsedInstanceSize() const { |
| int words = used_or_unused_instance_size_in_words(); |
| if (words < JSObject::kFieldsAdded) { |
| // All in-object properties are used and the words is tracking the slack |
| // in the property array. |
| return instance_size(); |
| } |
| return words * kTaggedSize; |
| } |
| |
| void Map::SetInObjectUnusedPropertyFields(int value) { |
| static_assert(JSObject::kFieldsAdded == JSObject::kHeaderSize / kTaggedSize); |
| if (!IsJSObjectMap(*this)) { |
| CHECK_EQ(0, value); |
| set_used_or_unused_instance_size_in_words(0); |
| DCHECK_EQ(0, UnusedPropertyFields()); |
| return; |
| } |
| CHECK_LE(0, value); |
| DCHECK_LE(value, GetInObjectProperties()); |
| int used_inobject_properties = GetInObjectProperties() - value; |
| set_used_or_unused_instance_size_in_words( |
| GetInObjectPropertyOffset(used_inobject_properties) / kTaggedSize); |
| DCHECK_EQ(value, UnusedPropertyFields()); |
| } |
| |
| void Map::SetOutOfObjectUnusedPropertyFields(int value) { |
| static_assert(JSObject::kFieldsAdded == JSObject::kHeaderSize / kTaggedSize); |
| CHECK_LT(static_cast<unsigned>(value), JSObject::kFieldsAdded); |
| // For out of object properties "used_instance_size_in_words" byte encodes |
| // the slack in the property array. |
| set_used_or_unused_instance_size_in_words(value); |
| DCHECK_EQ(value, UnusedPropertyFields()); |
| } |
| |
| void Map::CopyUnusedPropertyFields(Tagged<Map> map) { |
| set_used_or_unused_instance_size_in_words( |
| map->used_or_unused_instance_size_in_words()); |
| DCHECK_EQ(UnusedPropertyFields(), map->UnusedPropertyFields()); |
| } |
| |
| void Map::CopyUnusedPropertyFieldsAdjustedForInstanceSize(Tagged<Map> map) { |
| int value = map->used_or_unused_instance_size_in_words(); |
| if (value >= JSPrimitiveWrapper::kFieldsAdded) { |
| // Unused in-object fields. Adjust the offset from the object’s start |
| // so it matches the distance to the object’s end. |
| value += instance_size_in_words() - map->instance_size_in_words(); |
| } |
| set_used_or_unused_instance_size_in_words(value); |
| DCHECK_EQ(UnusedPropertyFields(), map->UnusedPropertyFields()); |
| } |
| |
| void Map::AccountAddedPropertyField() { |
| // Update used instance size and unused property fields number. |
| static_assert(JSObject::kFieldsAdded == JSObject::kHeaderSize / kTaggedSize); |
| #ifdef DEBUG |
| int new_unused = UnusedPropertyFields() - 1; |
| if (new_unused < 0) new_unused += JSObject::kFieldsAdded; |
| #endif |
| int value = used_or_unused_instance_size_in_words(); |
| if (value >= JSObject::kFieldsAdded) { |
| if (value == instance_size_in_words()) { |
| AccountAddedOutOfObjectPropertyField(0); |
| } else { |
| // The property is added in-object, so simply increment the counter. |
| set_used_or_unused_instance_size_in_words(value + 1); |
| } |
| } else { |
| AccountAddedOutOfObjectPropertyField(value); |
| } |
| DCHECK_EQ(new_unused, UnusedPropertyFields()); |
| } |
| |
| void Map::AccountAddedOutOfObjectPropertyField(int unused_in_property_array) { |
| unused_in_property_array--; |
| if (unused_in_property_array < 0) { |
| unused_in_property_array += JSObject::kFieldsAdded; |
| } |
| CHECK_LT(static_cast<unsigned>(unused_in_property_array), |
| JSObject::kFieldsAdded); |
| set_used_or_unused_instance_size_in_words(unused_in_property_array); |
| DCHECK_EQ(unused_in_property_array, UnusedPropertyFields()); |
| } |
| |
| #if V8_ENABLE_WEBASSEMBLY |
| uint8_t Map::WasmByte1() const { |
| DCHECK(IsWasmObjectMap(*this)); |
| return inobject_properties_start_or_constructor_function_index(); |
| } |
| |
| uint8_t Map::WasmByte2() const { |
| DCHECK(IsWasmObjectMap(*this)); |
| return used_or_unused_instance_size_in_words(); |
| } |
| |
| void Map::SetWasmByte1(uint8_t value) { |
| CHECK(IsWasmObjectMap(*this)); |
| set_inobject_properties_start_or_constructor_function_index(value); |
| } |
| |
| void Map::SetWasmByte2(uint8_t value) { |
| CHECK(IsWasmObjectMap(*this)); |
| set_used_or_unused_instance_size_in_words(value); |
| } |
| #endif // V8_ENABLE_WEBASSEMBLY |
| |
| uint8_t Map::bit_field() const { |
| // TODO(solanes, v8:7790, v8:11353): Make this non-atomic when TSAN sees the |
| // map's store synchronization. |
| return relaxed_bit_field(); |
| } |
| |
| void Map::set_bit_field(uint8_t value) { |
| // TODO(solanes, v8:7790, v8:11353): Make this non-atomic when TSAN sees the |
| // map's store synchronization. |
| set_relaxed_bit_field(value); |
| } |
| |
| uint8_t Map::relaxed_bit_field() const { |
| return RELAXED_READ_BYTE_FIELD(*this, kBitFieldOffset); |
| } |
| |
| void Map::set_relaxed_bit_field(uint8_t value) { |
| RELAXED_WRITE_BYTE_FIELD(*this, kBitFieldOffset, value); |
| } |
| |
| uint8_t Map::bit_field2() const { return ReadField<uint8_t>(kBitField2Offset); } |
| |
| void Map::set_bit_field2(uint8_t value) { |
| WriteField<uint8_t>(kBitField2Offset, value); |
| } |
| |
| uint32_t Map::bit_field3() const { |
| // TODO(solanes, v8:7790, v8:11353): Make this and the setter non-atomic |
| // when TSAN sees the map's store synchronization. |
| return relaxed_bit_field3(); |
| } |
| |
| void Map::set_bit_field3(uint32_t value) { set_relaxed_bit_field3(value); } |
| |
| uint32_t Map::relaxed_bit_field3() const { |
| return RELAXED_READ_UINT32_FIELD(*this, kBitField3Offset); |
| } |
| |
| void Map::set_relaxed_bit_field3(uint32_t value) { |
| RELAXED_WRITE_UINT32_FIELD(*this, kBitField3Offset, value); |
| } |
| |
| uint32_t Map::release_acquire_bit_field3() const { |
| return ACQUIRE_READ_UINT32_FIELD(*this, kBitField3Offset); |
| } |
| |
| void Map::set_release_acquire_bit_field3(uint32_t value) { |
| RELEASE_WRITE_UINT32_FIELD(*this, kBitField3Offset, value); |
| } |
| |
| bool Map::is_abandoned_prototype_map() const { |
| return is_prototype_map() && !owns_descriptors(); |
| } |
| |
| bool Map::should_be_fast_prototype_map() const { |
| DCHECK(is_prototype_map()); |
| if (!has_prototype_info()) return false; |
| return PrototypeInfo::cast(prototype_info())->should_be_fast_map(); |
| } |
| |
| bool Map::has_prototype_info() const { |
| DCHECK(is_prototype_map()); |
| return PrototypeInfo::IsPrototypeInfoFast(prototype_info()); |
| } |
| |
| bool Map::TryGetPrototypeInfo(Tagged<PrototypeInfo>* result) const { |
| DCHECK(is_prototype_map()); |
| Tagged<Object> maybe_proto_info = prototype_info(); |
| if (!PrototypeInfo::IsPrototypeInfoFast(maybe_proto_info)) return false; |
| *result = PrototypeInfo::cast(maybe_proto_info); |
| return true; |
| } |
| |
| void Map::set_elements_kind(ElementsKind elements_kind) { |
| CHECK_LT(static_cast<int>(elements_kind), kElementsKindCount); |
| set_bit_field2( |
| Map::Bits2::ElementsKindBits::update(bit_field2(), elements_kind)); |
| } |
| |
| ElementsKind Map::elements_kind() const { |
| return Map::Bits2::ElementsKindBits::decode(bit_field2()); |
| } |
| |
| bool Map::has_fast_smi_elements() const { |
| return IsSmiElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_fast_object_elements() const { |
| return IsObjectElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_fast_smi_or_object_elements() const { |
| return IsSmiOrObjectElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_fast_double_elements() const { |
| return IsDoubleElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_fast_elements() const { |
| return IsFastElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_fast_packed_elements() const { |
| return IsFastPackedElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_sloppy_arguments_elements() const { |
| return IsSloppyArgumentsElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_fast_sloppy_arguments_elements() const { |
| return elements_kind() == FAST_SLOPPY_ARGUMENTS_ELEMENTS; |
| } |
| |
| bool Map::has_fast_string_wrapper_elements() const { |
| return elements_kind() == FAST_STRING_WRAPPER_ELEMENTS; |
| } |
| |
| bool Map::has_typed_array_or_rab_gsab_typed_array_elements() const { |
| return IsTypedArrayOrRabGsabTypedArrayElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_any_typed_array_or_wasm_array_elements() const { |
| ElementsKind kind = elements_kind(); |
| return IsTypedArrayOrRabGsabTypedArrayElementsKind(kind) || |
| #if V8_ENABLE_WEBASSEMBLY |
| IsWasmArrayElementsKind(kind) || |
| #endif // V8_ENABLE_WEBASSEMBLY |
| false; |
| } |
| |
| bool Map::has_dictionary_elements() const { |
| return IsDictionaryElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_any_nonextensible_elements() const { |
| return IsAnyNonextensibleElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_nonextensible_elements() const { |
| return IsNonextensibleElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_sealed_elements() const { |
| return IsSealedElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_frozen_elements() const { |
| return IsFrozenElementsKind(elements_kind()); |
| } |
| |
| bool Map::has_shared_array_elements() const { |
| return IsSharedArrayElementsKind(elements_kind()); |
| } |
| |
| void Map::set_is_dictionary_map(bool value) { |
| uint32_t new_bit_field3 = |
| Bits3::IsDictionaryMapBit::update(bit_field3(), value); |
| new_bit_field3 = Bits3::IsUnstableBit::update(new_bit_field3, value); |
| set_bit_field3(new_bit_field3); |
| } |
| |
| bool Map::is_dictionary_map() const { |
| return Bits3::IsDictionaryMapBit::decode(relaxed_bit_field3()); |
| } |
| |
| void Map::mark_unstable() { |
| set_release_acquire_bit_field3( |
| Bits3::IsUnstableBit::update(bit_field3(), true)); |
| } |
| |
| bool Map::is_stable() const { |
| return !Bits3::IsUnstableBit::decode(release_acquire_bit_field3()); |
| } |
| |
| bool Map::CanBeDeprecated() const { |
| for (InternalIndex i : IterateOwnDescriptors()) { |
| PropertyDetails details = instance_descriptors(kRelaxedLoad)->GetDetails(i); |
| if (details.representation().MightCauseMapDeprecation()) return true; |
| if (details.kind() == PropertyKind::kData && |
| details.location() == PropertyLocation::kDescriptor) { |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| void Map::NotifyLeafMapLayoutChange(Isolate* isolate) { |
| if (is_stable()) { |
| mark_unstable(); |
| DependentCode::DeoptimizeDependencyGroups( |
| isolate, *this, DependentCode::kPrototypeCheckGroup); |
| } |
| } |
| |
| bool Map::CanTransition() const { |
| // Only JSObject and subtypes have map transitions and back pointers. |
| const InstanceType type = instance_type(); |
| // Shared JS objects have fixed shapes and do not transition. Their maps are |
| // either in shared space or RO space. |
| DCHECK_IMPLIES(InstanceTypeChecker::IsAlwaysSharedSpaceJSObject(type), |
| InAnySharedSpace(*this)); |
| return InstanceTypeChecker::IsJSObject(type) && |
| !InstanceTypeChecker::IsAlwaysSharedSpaceJSObject(type); |
| } |
| |
| bool IsBooleanMap(Tagged<Map> map) { |
| return map == map->GetReadOnlyRoots().boolean_map(); |
| } |
| |
| bool IsNullOrUndefinedMap(Tagged<Map> map) { |
| auto roots = map->GetReadOnlyRoots(); |
| return map == roots.null_map() || map == roots.undefined_map(); |
| } |
| |
| bool IsPrimitiveMap(Tagged<Map> map) { |
| return map->instance_type() <= LAST_PRIMITIVE_HEAP_OBJECT_TYPE; |
| } |
| |
| void Map::UpdateDescriptors(Isolate* isolate, |
| Tagged<DescriptorArray> descriptors, |
| int number_of_own_descriptors) { |
| SetInstanceDescriptors(isolate, descriptors, number_of_own_descriptors); |
| } |
| |
| void Map::InitializeDescriptors(Isolate* isolate, |
| Tagged<DescriptorArray> descriptors) { |
| SetInstanceDescriptors(isolate, descriptors, |
| descriptors->number_of_descriptors()); |
| } |
| |
| void Map::clear_padding() { |
| if (FIELD_SIZE(kOptionalPaddingOffset) == 0) return; |
| DCHECK_EQ(4, FIELD_SIZE(kOptionalPaddingOffset)); |
| memset(reinterpret_cast<void*>(address() + kOptionalPaddingOffset), 0, |
| FIELD_SIZE(kOptionalPaddingOffset)); |
| } |
| |
| void Map::AppendDescriptor(Isolate* isolate, Descriptor* desc) { |
| Tagged<DescriptorArray> descriptors = instance_descriptors(isolate); |
| int number_of_own_descriptors = NumberOfOwnDescriptors(); |
| DCHECK(descriptors->number_of_descriptors() == number_of_own_descriptors); |
| { |
| // The following two operations need to happen before the marking write |
| // barrier. |
| descriptors->Append(desc); |
| SetNumberOfOwnDescriptors(number_of_own_descriptors + 1); |
| #ifndef V8_DISABLE_WRITE_BARRIERS |
| WriteBarrier::Marking(descriptors, number_of_own_descriptors + 1); |
| #endif |
| } |
| // Properly mark the map if the {desc} is an "interesting symbol". |
| if (desc->GetKey()->IsInteresting(isolate)) { |
| set_may_have_interesting_properties(true); |
| } |
| PropertyDetails details = desc->GetDetails(); |
| if (details.location() == PropertyLocation::kField) { |
| DCHECK_GT(UnusedPropertyFields(), 0); |
| AccountAddedPropertyField(); |
| } |
| |
| // This function does not support appending double field descriptors and |
| // it should never try to (otherwise, layout descriptor must be updated too). |
| #ifdef DEBUG |
| DCHECK(details.location() != PropertyLocation::kField || |
| !details.representation().IsDouble()); |
| #endif |
| } |
| |
| // static |
| bool Map::ConcurrentIsHeapObjectWithMap(PtrComprCageBase cage_base, |
| Tagged<Object> object, |
| Tagged<Map> meta_map) { |
| if (!IsHeapObject(object)) return false; |
| Tagged<HeapObject> heap_object = HeapObject::cast(object); |
| return heap_object->map(cage_base) == meta_map; |
| } |
| |
| DEF_GETTER(Map, GetBackPointer, Tagged<HeapObject>) { |
| Tagged<Map> back_pointer; |
| if (TryGetBackPointer(cage_base, &back_pointer)) { |
| return back_pointer; |
| } |
| return GetReadOnlyRoots(cage_base).undefined_value(); |
| } |
| |
| bool Map::TryGetBackPointer(PtrComprCageBase cage_base, |
| Tagged<Map>* back_pointer) const { |
| Tagged<Object> object = constructor_or_back_pointer(cage_base, kRelaxedLoad); |
| // We don't expect maps from another native context in the transition tree, |
| // so just compare object's map against current map's meta map. |
| Tagged<Map> meta_map = map(cage_base); |
| if (ConcurrentIsHeapObjectWithMap(cage_base, object, meta_map)) { |
| DCHECK(IsMap(object)); |
| // Sanity check - only contextful maps can transition. |
| DCHECK(IsNativeContext(meta_map->native_context_or_null())); |
| *back_pointer = Map::cast(object); |
| return true; |
| } |
| // If it was a map that'd mean that there are maps from different native |
| // contexts in the transition tree. |
| DCHECK(!IsMap(object)); |
| return false; |
| } |
| |
| void Map::SetBackPointer(Tagged<HeapObject> value, WriteBarrierMode mode) { |
| CHECK_GE(instance_type(), FIRST_JS_RECEIVER_TYPE); |
| CHECK(IsMap(value)); |
| CHECK(IsUndefined(GetBackPointer())); |
| CHECK_EQ(Map::cast(value)->GetConstructorRaw(), |
| constructor_or_back_pointer()); |
| set_constructor_or_back_pointer(value, mode); |
| } |
| |
| // static |
| Tagged<Map> Map::GetMapFor(ReadOnlyRoots roots, InstanceType type) { |
| RootIndex map_idx = TryGetMapRootIdxFor(type).value(); |
| return Map::unchecked_cast(roots.object_at(map_idx)); |
| } |
| |
| // static |
| Tagged<Map> Map::ElementsTransitionMap(Isolate* isolate, |
| ConcurrencyMode cmode) { |
| return TransitionsAccessor(isolate, *this, IsConcurrent(cmode)) |
| .SearchSpecial(ReadOnlyRoots(isolate).elements_transition_symbol()); |
| } |
| |
| ACCESSORS(Map, dependent_code, Tagged<DependentCode>, kDependentCodeOffset) |
| RELAXED_ACCESSORS(Map, prototype_validity_cell, Tagged<Object>, |
| kPrototypeValidityCellOffset) |
| ACCESSORS_CHECKED2(Map, constructor_or_back_pointer, Tagged<Object>, |
| kConstructorOrBackPointerOrNativeContextOffset, |
| !IsContextMap(*this), IsNull(value) || !IsContextMap(*this)) |
| RELAXED_ACCESSORS_CHECKED2(Map, constructor_or_back_pointer, Tagged<Object>, |
| kConstructorOrBackPointerOrNativeContextOffset, |
| !IsContextMap(*this), |
| IsNull(value) || !IsContextMap(*this)) |
| ACCESSORS_CHECKED(Map, native_context, Tagged<NativeContext>, |
| kConstructorOrBackPointerOrNativeContextOffset, |
| IsContextMap(*this) || IsMapMap(*this)) |
| ACCESSORS_CHECKED(Map, native_context_or_null, Tagged<Object>, |
| kConstructorOrBackPointerOrNativeContextOffset, |
| (IsNull(value) || IsNativeContext(value)) && |
| (IsContextMap(*this) || IsMapMap(*this))) |
| // Unlike native_context_or_null() this getter allows the value to be |
| // equal to Smi::uninitialized_deserialization_value(). |
| DEF_GETTER(Map, raw_native_context_or_null, Tagged<Object>) { |
| Tagged<Object> value = TaggedField< |
| Tagged<Object>, |
| kConstructorOrBackPointerOrNativeContextOffset>::load(cage_base, *this); |
| DCHECK(IsNull(value) || IsNativeContext(value) || |
| value == Smi::uninitialized_deserialization_value()); |
| DCHECK(IsContextMap(*this) || IsMapMap(*this)); |
| return value; |
| } |
| #if V8_ENABLE_WEBASSEMBLY |
| ACCESSORS_CHECKED(Map, wasm_type_info, Tagged<WasmTypeInfo>, |
| kConstructorOrBackPointerOrNativeContextOffset, |
| IsWasmStructMap(*this) || IsWasmArrayMap(*this) || |
| IsWasmFuncRefMap(*this)) |
| #endif // V8_ENABLE_WEBASSEMBLY |
| |
| bool Map::IsPrototypeValidityCellValid() const { |
| Tagged<Object> validity_cell = prototype_validity_cell(kRelaxedLoad); |
| if (IsSmi(validity_cell)) { |
| // Smi validity cells should always be considered valid. |
| DCHECK_EQ(Smi::cast(validity_cell).value(), Map::kPrototypeChainValid); |
| return true; |
| } |
| Tagged<Smi> cell_value = Smi::cast(Cell::cast(validity_cell)->value()); |
| return cell_value == Smi::FromInt(Map::kPrototypeChainValid); |
| } |
| |
| DEF_GETTER(Map, GetConstructorRaw, Tagged<Object>) { |
| Tagged<Object> maybe_constructor = constructor_or_back_pointer(cage_base); |
| // Follow any back pointers. |
| // We don't expect maps from another native context in the transition tree, |
| // so just compare object's map against current map's meta map. |
| Tagged<Map> meta_map = map(cage_base); |
| while ( |
| ConcurrentIsHeapObjectWithMap(cage_base, maybe_constructor, meta_map)) { |
| DCHECK(IsMap(maybe_constructor)); |
| // Sanity check - only contextful maps can transition. |
| DCHECK(IsNativeContext(meta_map->native_context_or_null())); |
| maybe_constructor = |
| Map::cast(maybe_constructor)->constructor_or_back_pointer(cage_base); |
| } |
| // If it was a map that'd mean that there are maps from different native |
| // contexts in the transition tree. |
| DCHECK(!IsMap(maybe_constructor)); |
| return maybe_constructor; |
| } |
| |
| DEF_GETTER(Map, GetNonInstancePrototype, Tagged<Object>) { |
| DCHECK(has_non_instance_prototype()); |
| Tagged<Object> raw_constructor = GetConstructorRaw(cage_base); |
| CHECK(IsTuple2(raw_constructor)); |
| // Get prototype from the {constructor, non-instance_prototype} tuple. |
| Tagged<Tuple2> non_instance_prototype_constructor_tuple = |
| Tuple2::cast(raw_constructor); |
| Tagged<Object> result = non_instance_prototype_constructor_tuple->value2(); |
| DCHECK(!IsJSReceiver(result)); |
| DCHECK(!IsFunctionTemplateInfo(result)); |
| return result; |
| } |
| |
| DEF_GETTER(Map, GetConstructor, Tagged<Object>) { |
| Tagged<Object> maybe_constructor = GetConstructorRaw(cage_base); |
| if (IsTuple2(maybe_constructor)) { |
| // Get constructor from the {constructor, non-instance_prototype} tuple. |
| maybe_constructor = Tuple2::cast(maybe_constructor)->value1(); |
| } |
| return maybe_constructor; |
| } |
| |
| Tagged<Object> Map::TryGetConstructor(PtrComprCageBase cage_base, |
| int max_steps) { |
| Tagged<Object> maybe_constructor = constructor_or_back_pointer(cage_base); |
| // Follow any back pointers. |
| while (IsMap(maybe_constructor, cage_base)) { |
| if (max_steps-- == 0) return Smi::FromInt(0); |
| maybe_constructor = |
| Map::cast(maybe_constructor)->constructor_or_back_pointer(cage_base); |
| } |
| if (IsTuple2(maybe_constructor)) { |
| // Get constructor from the {constructor, non-instance_prototype} tuple. |
| maybe_constructor = Tuple2::cast(maybe_constructor)->value1(); |
| } |
| return maybe_constructor; |
| } |
| |
| DEF_GETTER(Map, GetFunctionTemplateInfo, Tagged<FunctionTemplateInfo>) { |
| Tagged<Object> constructor = GetConstructor(cage_base); |
| if (IsJSFunction(constructor, cage_base)) { |
| Tagged<SharedFunctionInfo> sfi = |
| JSFunction::cast(constructor)->shared(cage_base); |
| DCHECK(sfi->IsApiFunction()); |
| return sfi->api_func_data(); |
| } |
| DCHECK(IsFunctionTemplateInfo(constructor, cage_base)); |
| return FunctionTemplateInfo::cast(constructor); |
| } |
| |
| void Map::SetConstructor(Tagged<Object> constructor, WriteBarrierMode mode) { |
| // Never overwrite a back pointer with a constructor. |
| CHECK(!IsMap(constructor_or_back_pointer())); |
| // Constructor field must contain {constructor, non-instance_prototype} tuple |
| // for maps with non-instance prototype. |
| DCHECK_EQ(has_non_instance_prototype(), IsTuple2(constructor)); |
| set_constructor_or_back_pointer(constructor, mode); |
| } |
| |
| Handle<Map> Map::CopyInitialMap(Isolate* isolate, Handle<Map> map) { |
| return CopyInitialMap(isolate, map, map->instance_size(), |
| map->GetInObjectProperties(), |
| map->UnusedPropertyFields()); |
| } |
| |
| bool Map::IsInobjectSlackTrackingInProgress() const { |
| return construction_counter() != Map::kNoSlackTracking; |
| } |
| |
| void Map::InobjectSlackTrackingStep(Isolate* isolate) { |
| DisallowGarbageCollection no_gc; |
| // Slack tracking should only be performed on an initial map. |
| DCHECK(IsUndefined(GetBackPointer())); |
| if (!this->IsInobjectSlackTrackingInProgress()) return; |
| int counter = construction_counter(); |
| set_construction_counter(counter - 1); |
| if (counter == kSlackTrackingCounterEnd) { |
| MapUpdater::CompleteInobjectSlackTracking(isolate, *this); |
| } |
| } |
| |
| int Map::SlackForArraySize(int old_size, int size_limit) { |
| const int max_slack = size_limit - old_size; |
| CHECK_LE(0, max_slack); |
| if (old_size < 4) { |
| DCHECK_LE(1, max_slack); |
| return 1; |
| } |
| return std::min(max_slack, old_size / 4); |
| } |
| |
| int Map::InstanceSizeFromSlack(int slack) const { |
| return instance_size() - slack * kTaggedSize; |
| } |
| |
| OBJECT_CONSTRUCTORS_IMPL(NormalizedMapCache, WeakFixedArray) |
| CAST_ACCESSOR(NormalizedMapCache) |
| NEVER_READ_ONLY_SPACE_IMPL(NormalizedMapCache) |
| |
| int NormalizedMapCache::GetIndex(Handle<Map> map) { |
| return map->Hash() % NormalizedMapCache::kEntries; |
| } |
| |
| DEF_HEAP_OBJECT_PREDICATE(HeapObject, IsNormalizedMapCache) { |
| if (!IsWeakFixedArray(obj, cage_base)) return false; |
| if (WeakFixedArray::cast(obj)->length() != NormalizedMapCache::kEntries) { |
| return false; |
| } |
| return true; |
| } |
| |
| } // namespace internal |
| } // namespace v8 |
| |
| #include "src/objects/object-macros-undef.h" |
| |
| #endif // V8_OBJECTS_MAP_INL_H_ |