| /* | 
 |  * Copyright (C) 2016-2018 Apple Inc. All rights reserved. | 
 |  * | 
 |  * Redistribution and use in source and binary forms, with or without | 
 |  * modification, are permitted provided that the following conditions | 
 |  * are met: | 
 |  * 1. Redistributions of source code must retain the above copyright | 
 |  *    notice, this list of conditions and the following disclaimer. | 
 |  * 2. Redistributions in binary form must reproduce the above copyright | 
 |  *    notice, this list of conditions and the following disclaimer in the | 
 |  *    documentation and/or other materials provided with the distribution. | 
 |  * | 
 |  * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY | 
 |  * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | 
 |  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR | 
 |  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL APPLE INC. OR | 
 |  * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, | 
 |  * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, | 
 |  * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR | 
 |  * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY | 
 |  * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | 
 |  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | 
 |  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.  | 
 |  */ | 
 |  | 
 | #pragma once | 
 |  | 
 | #include "ExceptionHelpers.h" | 
 | #include "JSCJSValueInlines.h" | 
 | #include "JSObject.h" | 
 |  | 
 | namespace JSC { | 
 |  | 
 | JS_EXPORT_PRIVATE const ClassInfo* getHashMapBucketKeyClassInfo(); | 
 | JS_EXPORT_PRIVATE const ClassInfo* getHashMapBucketKeyValueClassInfo(); | 
 | JS_EXPORT_PRIVATE const ClassInfo* getHashMapImplKeyClassInfo(); | 
 | JS_EXPORT_PRIVATE const ClassInfo* getHashMapImplKeyValueClassInfo(); | 
 |  | 
 | enum class HashTableType { | 
 |     Key, | 
 |     KeyValue | 
 | }; | 
 |  | 
 | struct HashMapBucketDataKey { | 
 |     static const HashTableType Type = HashTableType::Key; | 
 |     WriteBarrier<Unknown> key; | 
 | }; | 
 |  | 
 | struct HashMapBucketDataKeyValue { | 
 |     static const HashTableType Type = HashTableType::KeyValue; | 
 |     WriteBarrier<Unknown> key; | 
 |     WriteBarrier<Unknown> value; | 
 | }; | 
 |  | 
 | template <typename Data> | 
 | class HashMapBucket final : public JSCell { | 
 |     using Base = JSCell; | 
 |  | 
 |     template <typename T = Data> | 
 |     static typename std::enable_if<std::is_same<T, HashMapBucketDataKey>::value, Structure*>::type selectStructure(VM& vm) | 
 |     { | 
 |         return vm.hashMapBucketSetStructure.get(); | 
 |     } | 
 |  | 
 |     template <typename T = Data> | 
 |     static typename std::enable_if<std::is_same<T, HashMapBucketDataKeyValue>::value, Structure*>::type selectStructure(VM& vm) | 
 |     { | 
 |         return vm.hashMapBucketMapStructure.get(); | 
 |     } | 
 |  | 
 | public: | 
 |     static const HashTableType Type = Data::Type; | 
 |     static const ClassInfo s_info; // This is never accessed directly, since that would break linkage on some compilers. | 
 |  | 
 |     template<typename CellType, SubspaceAccess mode> | 
 |     static IsoSubspace* subspaceFor(VM& vm) | 
 |     { | 
 |         if constexpr (Type == HashTableType::Key) | 
 |             return vm.setBucketSpace<mode>(); | 
 |         else | 
 |             return vm.mapBucketSpace<mode>(); | 
 |     } | 
 |  | 
 |     static const ClassInfo* info() | 
 |     { | 
 |         if constexpr (Type == HashTableType::Key) | 
 |             return getHashMapBucketKeyClassInfo(); | 
 |         else | 
 |             return getHashMapBucketKeyValueClassInfo(); | 
 |     } | 
 |  | 
 |     static Structure* createStructure(VM& vm, JSGlobalObject* globalObject, JSValue prototype) | 
 |     { | 
 |         return Structure::create(vm, globalObject, prototype, TypeInfo(CellType, StructureFlags), info()); | 
 |     } | 
 |  | 
 |     static HashMapBucket* create(VM& vm) | 
 |     { | 
 |         HashMapBucket* bucket = new (NotNull, allocateCell<HashMapBucket<Data>>(vm.heap)) HashMapBucket(vm, selectStructure(vm)); | 
 |         bucket->finishCreation(vm); | 
 |         ASSERT(!bucket->next()); | 
 |         ASSERT(!bucket->prev()); | 
 |         return bucket; | 
 |     } | 
 |  | 
 |     static HashMapBucket* createSentinel(VM& vm) | 
 |     { | 
 |         auto* bucket = create(vm); | 
 |         bucket->setKey(vm, jsUndefined()); | 
 |         bucket->setValue(vm, jsUndefined()); | 
 |         ASSERT(!bucket->deleted()); | 
 |         return bucket; | 
 |     } | 
 |  | 
 |     HashMapBucket(VM& vm, Structure* structure) | 
 |         : Base(vm, structure) | 
 |     { | 
 |         ASSERT(deleted()); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE void setNext(VM& vm, HashMapBucket* bucket) | 
 |     { | 
 |         m_next.set(vm, this, bucket); | 
 |     } | 
 |     ALWAYS_INLINE void setPrev(VM& vm, HashMapBucket* bucket) | 
 |     { | 
 |         m_prev.set(vm, this, bucket); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE void setKey(VM& vm, JSValue key) | 
 |     { | 
 |         m_data.key.set(vm, this, key); | 
 |     } | 
 |  | 
 |     template <typename T = Data> | 
 |     ALWAYS_INLINE typename std::enable_if<std::is_same<T, HashMapBucketDataKeyValue>::value>::type setValue(VM& vm, JSValue value) | 
 |     { | 
 |         m_data.value.set(vm, this, value); | 
 |     } | 
 |     template <typename T = Data> | 
 |     ALWAYS_INLINE typename std::enable_if<std::is_same<T, HashMapBucketDataKey>::value>::type setValue(VM&, JSValue) { } | 
 |  | 
 |     ALWAYS_INLINE JSValue key() const { return m_data.key.get(); } | 
 |  | 
 |     template <typename T = Data> | 
 |     ALWAYS_INLINE typename std::enable_if<std::is_same<T, HashMapBucketDataKeyValue>::value, JSValue>::type value() const | 
 |     { | 
 |         return m_data.value.get(); | 
 |     } | 
 |  | 
 |     static void visitChildren(JSCell*, SlotVisitor&); | 
 |  | 
 |     ALWAYS_INLINE HashMapBucket* next() const { return m_next.get(); } | 
 |     ALWAYS_INLINE HashMapBucket* prev() const { return m_prev.get(); } | 
 |  | 
 |     ALWAYS_INLINE bool deleted() const { return !key(); } | 
 |     ALWAYS_INLINE void makeDeleted(VM& vm) | 
 |     { | 
 |         setKey(vm, JSValue()); | 
 |         setValue(vm, JSValue()); | 
 |     } | 
 |  | 
 |     static ptrdiff_t offsetOfKey() | 
 |     { | 
 |         return OBJECT_OFFSETOF(HashMapBucket, m_data) + OBJECT_OFFSETOF(Data, key); | 
 |     } | 
 |  | 
 |     template <typename T = Data> | 
 |     static typename std::enable_if<std::is_same<T, HashMapBucketDataKeyValue>::value, ptrdiff_t>::type offsetOfValue() | 
 |     { | 
 |         return OBJECT_OFFSETOF(HashMapBucket, m_data) + OBJECT_OFFSETOF(Data, value); | 
 |     } | 
 |  | 
 |     static ptrdiff_t offsetOfNext() | 
 |     { | 
 |         return OBJECT_OFFSETOF(HashMapBucket, m_next); | 
 |     } | 
 |  | 
 |     template <typename T = Data> | 
 |     ALWAYS_INLINE static typename std::enable_if<std::is_same<T, HashMapBucketDataKeyValue>::value, JSValue>::type extractValue(const HashMapBucket& bucket) | 
 |     { | 
 |         return bucket.value(); | 
 |     } | 
 |  | 
 |     template <typename T = Data> | 
 |     ALWAYS_INLINE static typename std::enable_if<std::is_same<T, HashMapBucketDataKey>::value, JSValue>::type extractValue(const HashMapBucket&) | 
 |     { | 
 |         return JSValue(); | 
 |     } | 
 |  | 
 | private: | 
 |     WriteBarrier<HashMapBucket> m_next; | 
 |     WriteBarrier<HashMapBucket> m_prev; | 
 |     Data m_data; | 
 | }; | 
 |  | 
 | template <typename BucketType> | 
 | class HashMapBuffer { | 
 | public: | 
 |     HashMapBuffer() = delete; | 
 |  | 
 |     static size_t allocationSize(Checked<size_t> capacity) | 
 |     { | 
 |         return (capacity * sizeof(BucketType*)).unsafeGet(); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE BucketType** buffer() const | 
 |     { | 
 |         return bitwise_cast<BucketType**>(this); | 
 |     } | 
 |  | 
 |     static HashMapBuffer* create(JSGlobalObject* globalObject, VM& vm, JSCell*, uint32_t capacity) | 
 |     { | 
 |         auto scope = DECLARE_THROW_SCOPE(vm); | 
 |         size_t allocationSize = HashMapBuffer::allocationSize(capacity); | 
 |         void* data = vm.jsValueGigacageAuxiliarySpace.allocateNonVirtual(vm, allocationSize, nullptr, AllocationFailureMode::ReturnNull); | 
 |         if (!data) { | 
 |             throwOutOfMemoryError(globalObject, scope); | 
 |             return nullptr; | 
 |         } | 
 |  | 
 |         HashMapBuffer* buffer = static_cast<HashMapBuffer*>(data); | 
 |         buffer->reset(capacity); | 
 |         return buffer; | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE void reset(uint32_t capacity) | 
 |     { | 
 |         memset(this, -1, allocationSize(capacity)); | 
 |     } | 
 | }; | 
 |  | 
 | ALWAYS_INLINE static bool areKeysEqual(JSGlobalObject* globalObject, JSValue a, JSValue b) | 
 | { | 
 |     // We want +0 and -0 to be compared to true here. sameValue() itself doesn't | 
 |     // guarantee that, however, we normalize all keys before comparing and storing | 
 |     // them in the map. The normalization will convert -0.0 and 0.0 to the integer | 
 |     // representation for 0. | 
 |     return sameValue(globalObject, a, b); | 
 | } | 
 |  | 
 | // Note that normalization is inlined in DFG's NormalizeMapKey. | 
 | // Keep in sync with the implementation of DFG and FTL normalization. | 
 | ALWAYS_INLINE JSValue normalizeMapKey(JSValue key) | 
 | { | 
 |     if (!key.isNumber()) | 
 |         return key; | 
 |  | 
 |     if (key.isInt32()) | 
 |         return key; | 
 |  | 
 |     double d = key.asDouble(); | 
 |     if (std::isnan(d)) | 
 |         return jsNaN(); | 
 |  | 
 |     int i = static_cast<int>(d); | 
 |     if (i == d) { | 
 |         // When a key is -0, we convert it to positive zero. | 
 |         // When a key is the double representation for an integer, we convert it to an integer. | 
 |         return jsNumber(i); | 
 |     } | 
 |     // This means key is definitely not negative zero, and it's definitely not a double representation of an integer.  | 
 |     return key; | 
 | } | 
 |  | 
 | static ALWAYS_INLINE uint32_t wangsInt64Hash(uint64_t key) | 
 | { | 
 |     key += ~(key << 32); | 
 |     key ^= (key >> 22); | 
 |     key += ~(key << 13); | 
 |     key ^= (key >> 8); | 
 |     key += (key << 3); | 
 |     key ^= (key >> 15); | 
 |     key += ~(key << 27); | 
 |     key ^= (key >> 31); | 
 |     return static_cast<unsigned>(key); | 
 | } | 
 |  | 
 | ALWAYS_INLINE uint32_t jsMapHash(JSGlobalObject* globalObject, VM& vm, JSValue value) | 
 | { | 
 |     ASSERT_WITH_MESSAGE(normalizeMapKey(value) == value, "We expect normalized values flowing into this function."); | 
 |  | 
 |     if (value.isString()) { | 
 |         auto scope = DECLARE_THROW_SCOPE(vm); | 
 |         const String& wtfString = asString(value)->value(globalObject); | 
 |         RETURN_IF_EXCEPTION(scope, UINT_MAX); | 
 |         return wtfString.impl()->hash(); | 
 |     } | 
 |  | 
 |     return wangsInt64Hash(JSValue::encode(value)); | 
 | } | 
 |  | 
 | ALWAYS_INLINE Optional<uint32_t> concurrentJSMapHash(JSValue key) | 
 | { | 
 |     key = normalizeMapKey(key); | 
 |     if (key.isString()) { | 
 |         JSString* string = asString(key); | 
 |         if (string->length() > 10 * 1024) | 
 |             return WTF::nullopt; | 
 |         const StringImpl* impl = string->tryGetValueImpl(); | 
 |         if (!impl) | 
 |             return WTF::nullopt; | 
 |         return impl->concurrentHash(); | 
 |     } | 
 |  | 
 |     uint64_t rawValue = JSValue::encode(key); | 
 |     return wangsInt64Hash(rawValue); | 
 | } | 
 |  | 
 | ALWAYS_INLINE uint32_t shouldShrink(uint32_t capacity, uint32_t keyCount) | 
 | { | 
 |     return 8 * keyCount <= capacity && capacity > 4; | 
 | } | 
 |  | 
 | ALWAYS_INLINE uint32_t shouldRehashAfterAdd(uint32_t capacity, uint32_t keyCount, uint32_t deleteCount) | 
 | { | 
 |     return 2 * (keyCount + deleteCount) >= capacity; | 
 | } | 
 |  | 
 | ALWAYS_INLINE uint32_t nextCapacity(uint32_t capacity, uint32_t keyCount) | 
 | { | 
 |     if (shouldShrink(capacity, keyCount)) { | 
 |         ASSERT((capacity / 2) >= 4); | 
 |         return capacity / 2; | 
 |     } | 
 |  | 
 |     if (3 * keyCount <= capacity && capacity > 64) { | 
 |         // We stay at the same size if rehashing would cause us to be no more than | 
 |         // 1/3rd full. This comes up for programs like this: | 
 |         // Say the hash table grew to a key count of 64, causing it to grow to a capacity of 256. | 
 |         // Then, the table added 63 items. The load is now 127. Then, 63 items are deleted. | 
 |         // The load is still 127. Then, another item is added. The load is now 128, and we | 
 |         // decide that we need to rehash. The key count is 65, almost exactly what it was | 
 |         // when we grew to a capacity of 256. We don't really need to grow to a capacity | 
 |         // of 512 in this situation. Instead, we choose to rehash at the same size. This | 
 |         // will bring the load down to 65. We rehash into the same size when we determine | 
 |         // that the new load ratio will be under 1/3rd. (We also pick a minumum capacity | 
 |         // at which this rule kicks in because otherwise we will be too sensitive to rehashing | 
 |         // at the same capacity). | 
 |         return capacity; | 
 |     } | 
 |     return (Checked<uint32_t>(capacity) * 2).unsafeGet(); | 
 | } | 
 |  | 
 | template <typename HashMapBucketType> | 
 | class HashMapImpl : public JSNonFinalObject { | 
 |     using Base = JSNonFinalObject; | 
 |     using HashMapBufferType = HashMapBuffer<HashMapBucketType>; | 
 |  | 
 | public: | 
 |     using BucketType = HashMapBucketType; | 
 |  | 
 |     static void visitChildren(JSCell*, SlotVisitor&); | 
 |  | 
 |     static size_t estimatedSize(JSCell*, VM&); | 
 |  | 
 |     HashMapImpl(VM& vm, Structure* structure) | 
 |         : Base(vm, structure) | 
 |         , m_keyCount(0) | 
 |         , m_deleteCount(0) | 
 |         , m_capacity(4) | 
 |     { | 
 |     } | 
 |  | 
 |     HashMapImpl(VM& vm, Structure* structure, uint32_t sizeHint) | 
 |         : Base(vm, structure) | 
 |         , m_keyCount(0) | 
 |         , m_deleteCount(0) | 
 |     { | 
 |         uint32_t capacity = ((Checked<uint32_t>(sizeHint) * 2) + 1).unsafeGet(); | 
 |         capacity = std::max<uint32_t>(WTF::roundUpToPowerOfTwo(capacity), 4U); | 
 |         m_capacity = capacity; | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE HashMapBucketType** buffer() const | 
 |     { | 
 |         return m_buffer->buffer(); | 
 |     } | 
 |  | 
 |     void finishCreation(JSGlobalObject* globalObject, VM& vm) | 
 |     { | 
 |         ASSERT_WITH_MESSAGE(HashMapBucket<HashMapBucketDataKey>::offsetOfKey() == HashMapBucket<HashMapBucketDataKeyValue>::offsetOfKey(), "We assume this to be true in the DFG and FTL JIT."); | 
 |  | 
 |         auto scope = DECLARE_THROW_SCOPE(vm); | 
 |         Base::finishCreation(vm); | 
 |  | 
 |         makeAndSetNewBuffer(globalObject, vm); | 
 |         RETURN_IF_EXCEPTION(scope, void()); | 
 |  | 
 |         setUpHeadAndTail(globalObject, vm); | 
 |     } | 
 |  | 
 |     void finishCreation(JSGlobalObject* globalObject, VM& vm, HashMapImpl* base) | 
 |     { | 
 |         auto scope = DECLARE_THROW_SCOPE(vm); | 
 |         Base::finishCreation(vm); | 
 |  | 
 |         // This size should be the same to the case when you clone the map by calling add() repeatedly. | 
 |         uint32_t capacity = ((Checked<uint32_t>(base->m_keyCount) * 2) + 1).unsafeGet(); | 
 |         RELEASE_ASSERT(capacity <= (1U << 31)); | 
 |         capacity = std::max<uint32_t>(WTF::roundUpToPowerOfTwo(capacity), 4U); | 
 |         m_capacity = capacity; | 
 |         makeAndSetNewBuffer(globalObject, vm); | 
 |         RETURN_IF_EXCEPTION(scope, void()); | 
 |  | 
 |         setUpHeadAndTail(globalObject, vm); | 
 |  | 
 |         HashMapBucketType* bucket = base->m_head.get()->next(); | 
 |         while (bucket) { | 
 |             if (!bucket->deleted()) { | 
 |                 addNormalizedNonExistingForCloning(globalObject, bucket->key(), HashMapBucketType::extractValue(*bucket)); | 
 |                 RETURN_IF_EXCEPTION(scope, void()); | 
 |             } | 
 |             bucket = bucket->next(); | 
 |         } | 
 |         checkConsistency(); | 
 |     } | 
 |  | 
 |     static HashMapBucketType* emptyValue() | 
 |     { | 
 |         return bitwise_cast<HashMapBucketType*>(static_cast<uintptr_t>(-1)); | 
 |     } | 
 |  | 
 |     static ALWAYS_INLINE bool isEmpty(HashMapBucketType* bucket) | 
 |     { | 
 |         return bucket == emptyValue(); | 
 |     } | 
 |  | 
 |     static HashMapBucketType* deletedValue() | 
 |     { | 
 |         return bitwise_cast<HashMapBucketType*>(static_cast<uintptr_t>(-3)); | 
 |     } | 
 |  | 
 |     static ALWAYS_INLINE bool isDeleted(HashMapBucketType* bucket) | 
 |     { | 
 |         return bucket == deletedValue(); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE HashMapBucketType** findBucket(JSGlobalObject* globalObject, JSValue key) | 
 |     { | 
 |         VM& vm = getVM(globalObject); | 
 |         auto scope = DECLARE_THROW_SCOPE(vm); | 
 |         key = normalizeMapKey(key); | 
 |         uint32_t hash = jsMapHash(globalObject, vm, key); | 
 |         RETURN_IF_EXCEPTION(scope, nullptr); | 
 |         return findBucket(globalObject, key, hash); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE HashMapBucketType** findBucket(JSGlobalObject* globalObject, JSValue key, uint32_t hash) | 
 |     { | 
 |         ASSERT_WITH_MESSAGE(normalizeMapKey(key) == key, "We expect normalized values flowing into this function."); | 
 |         return findBucketAlreadyHashedAndNormalized(globalObject, key, hash); | 
 |     } | 
 |  | 
 |     template <typename T = HashMapBucketType> | 
 |     ALWAYS_INLINE typename std::enable_if<std::is_same<T, HashMapBucket<HashMapBucketDataKeyValue>>::value, JSValue>::type get(JSGlobalObject* globalObject, JSValue key) | 
 |     { | 
 |         if (HashMapBucketType** bucket = findBucket(globalObject, key)) | 
 |             return (*bucket)->value(); | 
 |         return jsUndefined(); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE bool has(JSGlobalObject* globalObject, JSValue key) | 
 |     { | 
 |         return !!findBucket(globalObject, key); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE void add(JSGlobalObject* globalObject, JSValue key, JSValue value = JSValue()) | 
 |     { | 
 |         key = normalizeMapKey(key); | 
 |         addNormalizedInternal(globalObject, key, value, [&] (HashMapBucketType* bucket) { | 
 |             return !isDeleted(bucket) && areKeysEqual(globalObject, key, bucket->key()); | 
 |         }); | 
 |         if (shouldRehashAfterAdd()) | 
 |             rehash(globalObject); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE HashMapBucketType* addNormalized(JSGlobalObject* globalObject, JSValue key, JSValue value, uint32_t hash) | 
 |     { | 
 |         ASSERT_WITH_MESSAGE(normalizeMapKey(key) == key, "We expect normalized values flowing into this function."); | 
 |         ASSERT_WITH_MESSAGE(jsMapHash(globalObject, getVM(globalObject), key) == hash, "We expect hash value is what we expect."); | 
 |  | 
 |         auto* bucket = addNormalizedInternal(getVM(globalObject), key, value, hash, [&] (HashMapBucketType* bucket) { | 
 |             return !isDeleted(bucket) && areKeysEqual(globalObject, key, bucket->key()); | 
 |         }); | 
 |         if (shouldRehashAfterAdd()) | 
 |             rehash(globalObject); | 
 |         return bucket; | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE bool remove(JSGlobalObject* globalObject, JSValue key) | 
 |     { | 
 |         HashMapBucketType** bucket = findBucket(globalObject, key); | 
 |         if (!bucket) | 
 |             return false; | 
 |  | 
 |         VM& vm = getVM(globalObject); | 
 |         HashMapBucketType* impl = *bucket; | 
 |         impl->next()->setPrev(vm, impl->prev()); | 
 |         impl->prev()->setNext(vm, impl->next()); | 
 |         impl->makeDeleted(vm); | 
 |  | 
 |         *bucket = deletedValue(); | 
 |  | 
 |         ++m_deleteCount; | 
 |         ASSERT(m_keyCount > 0); | 
 |         --m_keyCount; | 
 |  | 
 |         if (shouldShrink()) | 
 |             rehash(globalObject); | 
 |  | 
 |         return true; | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE uint32_t size() const | 
 |     { | 
 |         return m_keyCount; | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE void clear(JSGlobalObject* globalObject) | 
 |     { | 
 |         VM& vm = getVM(globalObject); | 
 |         m_keyCount = 0; | 
 |         m_deleteCount = 0; | 
 |         HashMapBucketType* head = m_head.get(); | 
 |         HashMapBucketType* bucket = m_head->next(); | 
 |         HashMapBucketType* tail = m_tail.get(); | 
 |         while (bucket != tail) { | 
 |             HashMapBucketType* next = bucket->next(); | 
 |             // We restart each iterator by pointing it to the head of the list. | 
 |             bucket->setNext(vm, head); | 
 |             bucket->makeDeleted(vm); | 
 |             bucket = next; | 
 |         } | 
 |         m_head->setNext(vm, m_tail.get()); | 
 |         m_tail->setPrev(vm, m_head.get()); | 
 |         m_capacity = 4; | 
 |         makeAndSetNewBuffer(globalObject, vm); | 
 |         checkConsistency(); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE size_t bufferSizeInBytes() const | 
 |     { | 
 |         return m_capacity * sizeof(HashMapBucketType*); | 
 |     } | 
 |  | 
 |     static ptrdiff_t offsetOfHead() | 
 |     { | 
 |         return OBJECT_OFFSETOF(HashMapImpl<HashMapBucketType>, m_head); | 
 |     } | 
 |  | 
 |     static ptrdiff_t offsetOfBuffer() | 
 |     { | 
 |         return OBJECT_OFFSETOF(HashMapImpl<HashMapBucketType>, m_buffer); | 
 |     } | 
 |  | 
 |     static ptrdiff_t offsetOfCapacity() | 
 |     { | 
 |         return OBJECT_OFFSETOF(HashMapImpl<HashMapBucketType>, m_capacity); | 
 |     } | 
 |  | 
 |     HashMapBucketType* head() { return m_head.get(); } | 
 |     HashMapBucketType* tail() { return m_tail.get(); } | 
 |  | 
 |     size_t approximateSize() const | 
 |     { | 
 |         size_t size = sizeof(HashMapImpl); | 
 |         size += bufferSizeInBytes(); | 
 |         size += 2 * sizeof(HashMapBucketType); // Head and tail members. | 
 |         size += m_keyCount * sizeof(HashMapBucketType); // Number of members that are on the list. | 
 |         return size; | 
 |     } | 
 |  | 
 | private: | 
 |     ALWAYS_INLINE uint32_t shouldRehashAfterAdd() const | 
 |     { | 
 |         return JSC::shouldRehashAfterAdd(m_capacity, m_keyCount, m_deleteCount); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE uint32_t shouldShrink() const | 
 |     { | 
 |         return JSC::shouldShrink(m_capacity, m_keyCount); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE void setUpHeadAndTail(JSGlobalObject*, VM& vm) | 
 |     { | 
 |         m_head.set(vm, this, HashMapBucketType::create(vm)); | 
 |         m_tail.set(vm, this, HashMapBucketType::create(vm)); | 
 |  | 
 |         m_head->setNext(vm, m_tail.get()); | 
 |         m_tail->setPrev(vm, m_head.get()); | 
 |         ASSERT(m_head->deleted()); | 
 |         ASSERT(m_tail->deleted()); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE void addNormalizedNonExistingForCloning(JSGlobalObject* globalObject, JSValue key, JSValue value = JSValue()) | 
 |     { | 
 |         addNormalizedInternal(globalObject, key, value, [&] (HashMapBucketType*) { | 
 |             return false; | 
 |         }); | 
 |     } | 
 |  | 
 |     template<typename CanUseBucket> | 
 |     ALWAYS_INLINE void addNormalizedInternal(JSGlobalObject* globalObject, JSValue key, JSValue value, const CanUseBucket& canUseBucket) | 
 |     { | 
 |         VM& vm = getVM(globalObject); | 
 |         auto scope = DECLARE_THROW_SCOPE(vm); | 
 |  | 
 |         uint32_t hash = jsMapHash(globalObject, vm, key); | 
 |         RETURN_IF_EXCEPTION(scope, void()); | 
 |         scope.release(); | 
 |         addNormalizedInternal(vm, key, value, hash, canUseBucket); | 
 |     } | 
 |  | 
 |     template<typename CanUseBucket> | 
 |     ALWAYS_INLINE HashMapBucketType* addNormalizedInternal(VM& vm, JSValue key, JSValue value, uint32_t hash, const CanUseBucket& canUseBucket) | 
 |     { | 
 |         ASSERT_WITH_MESSAGE(normalizeMapKey(key) == key, "We expect normalized values flowing into this function."); | 
 |  | 
 |         const uint32_t mask = m_capacity - 1; | 
 |         uint32_t index = hash & mask; | 
 |         HashMapBucketType** buffer = this->buffer(); | 
 |         HashMapBucketType* bucket = buffer[index]; | 
 |         while (!isEmpty(bucket)) { | 
 |             if (canUseBucket(bucket)) { | 
 |                 bucket->setValue(vm, value); | 
 |                 return bucket; | 
 |             } | 
 |             index = (index + 1) & mask; | 
 |             bucket = buffer[index]; | 
 |         } | 
 |  | 
 |         HashMapBucketType* newEntry = m_tail.get(); | 
 |         buffer[index] = newEntry; | 
 |         newEntry->setKey(vm, key); | 
 |         newEntry->setValue(vm, value); | 
 |         ASSERT(!newEntry->deleted()); | 
 |         HashMapBucketType* newTail = HashMapBucketType::create(vm); | 
 |         m_tail.set(vm, this, newTail); | 
 |         newTail->setPrev(vm, newEntry); | 
 |         ASSERT(newTail->deleted()); | 
 |         newEntry->setNext(vm, newTail); | 
 |  | 
 |         ++m_keyCount; | 
 |         return newEntry; | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE HashMapBucketType** findBucketAlreadyHashedAndNormalized(JSGlobalObject* globalObject, JSValue key, uint32_t hash) | 
 |     { | 
 |         const uint32_t mask = m_capacity - 1; | 
 |         uint32_t index = hash & mask; | 
 |         HashMapBucketType** buffer = this->buffer(); | 
 |         HashMapBucketType* bucket = buffer[index]; | 
 |  | 
 |         while (!isEmpty(bucket)) { | 
 |             if (!isDeleted(bucket) && areKeysEqual(globalObject, key, bucket->key())) | 
 |                 return buffer + index; | 
 |             index = (index + 1) & mask; | 
 |             bucket = buffer[index]; | 
 |         } | 
 |         return nullptr; | 
 |     } | 
 |  | 
 |     void rehash(JSGlobalObject* globalObject) | 
 |     { | 
 |         VM& vm = getVM(globalObject); | 
 |         auto scope = DECLARE_THROW_SCOPE(vm); | 
 |  | 
 |         uint32_t oldCapacity = m_capacity; | 
 |         m_capacity = nextCapacity(m_capacity, m_keyCount); | 
 |  | 
 |         if (m_capacity != oldCapacity) { | 
 |             makeAndSetNewBuffer(globalObject, vm); | 
 |             RETURN_IF_EXCEPTION(scope, void()); | 
 |         } else { | 
 |             m_buffer->reset(m_capacity); | 
 |             assertBufferIsEmpty(); | 
 |         } | 
 |  | 
 |         HashMapBucketType* iter = m_head->next(); | 
 |         HashMapBucketType* end = m_tail.get(); | 
 |         const uint32_t mask = m_capacity - 1; | 
 |         RELEASE_ASSERT(!(m_capacity & (m_capacity - 1))); | 
 |         HashMapBucketType** buffer = this->buffer(); | 
 |         while (iter != end) { | 
 |             uint32_t index = jsMapHash(globalObject, vm, iter->key()) & mask; | 
 |             EXCEPTION_ASSERT_WITH_MESSAGE(!scope.exception(), "All keys should already be hashed before, so this should not throw because it won't resolve ropes."); | 
 |             { | 
 |                 HashMapBucketType* bucket = buffer[index]; | 
 |                 while (!isEmpty(bucket)) { | 
 |                     index = (index + 1) & mask; | 
 |                     bucket = buffer[index]; | 
 |                 } | 
 |             } | 
 |             buffer[index] = iter; | 
 |             iter = iter->next(); | 
 |         } | 
 |  | 
 |         m_deleteCount = 0; | 
 |  | 
 |         checkConsistency(); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE void checkConsistency() const | 
 |     { | 
 |         if (ASSERT_ENABLED) { | 
 |             HashMapBucketType* iter = m_head->next(); | 
 |             HashMapBucketType* end = m_tail.get(); | 
 |             uint32_t size = 0; | 
 |             while (iter != end) { | 
 |                 ++size; | 
 |                 iter = iter->next(); | 
 |             } | 
 |             ASSERT(size == m_keyCount); | 
 |         } | 
 |     } | 
 |  | 
 |     void makeAndSetNewBuffer(JSGlobalObject* globalObject, VM& vm) | 
 |     { | 
 |         ASSERT(!(m_capacity & (m_capacity - 1))); | 
 |  | 
 |         HashMapBufferType* buffer = HashMapBufferType::create(globalObject, vm, this, m_capacity); | 
 |         if (UNLIKELY(!buffer)) | 
 |             return; | 
 |  | 
 |         m_buffer.set(vm, this, buffer); | 
 |         assertBufferIsEmpty(); | 
 |     } | 
 |  | 
 |     ALWAYS_INLINE void assertBufferIsEmpty() const | 
 |     { | 
 |         if (ASSERT_ENABLED) { | 
 |             for (unsigned i = 0; i < m_capacity; i++) | 
 |                 ASSERT(isEmpty(buffer()[i])); | 
 |         } | 
 |     } | 
 |  | 
 |     WriteBarrier<HashMapBucketType> m_head; | 
 |     WriteBarrier<HashMapBucketType> m_tail; | 
 |     AuxiliaryBarrier<HashMapBufferType*> m_buffer; | 
 |     uint32_t m_keyCount; | 
 |     uint32_t m_deleteCount; | 
 |     uint32_t m_capacity; | 
 | }; | 
 |  | 
 | } // namespace JSC |