|  | /* | 
|  | * Copyright (C) 2017-2021 Apple Inc. All rights reserved. | 
|  | * | 
|  | * Redistribution and use in source and binary forms, with or without | 
|  | * modification, are permitted provided that the following conditions | 
|  | * are met: | 
|  | * 1. Redistributions of source code must retain the above copyright | 
|  | *    notice, this list of conditions and the following disclaimer. | 
|  | * 2. Redistributions in binary form must reproduce the above copyright | 
|  | *    notice, this list of conditions and the following disclaimer in the | 
|  | *    documentation and/or other materials provided with the distribution. | 
|  | * | 
|  | * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY | 
|  | * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | 
|  | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR | 
|  | * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL APPLE INC. OR | 
|  | * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, | 
|  | * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, | 
|  | * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR | 
|  | * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY | 
|  | * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | 
|  | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | 
|  | * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | 
|  | */ | 
|  |  | 
|  | #pragma once | 
|  |  | 
|  | #include "AtomIndices.h" | 
|  | #include "IsoCellSet.h" | 
|  | #include "MarkedBlockInlines.h" | 
|  |  | 
|  | namespace JSC { | 
|  |  | 
|  | inline bool IsoCellSet::add(HeapCell* cell) | 
|  | { | 
|  | if (cell->isPreciseAllocation()) | 
|  | return !m_lowerTierBits.concurrentTestAndSet(cell->preciseAllocation().lowerTierIndex()); | 
|  | AtomIndices atomIndices(cell); | 
|  | auto& bitsPtrRef = m_bits[atomIndices.blockIndex]; | 
|  | auto* bits = bitsPtrRef.get(); | 
|  | if (UNLIKELY(!bits)) | 
|  | bits = addSlow(atomIndices.blockIndex); | 
|  | return !bits->concurrentTestAndSet(atomIndices.atomNumber); | 
|  | } | 
|  |  | 
|  | inline bool IsoCellSet::remove(HeapCell* cell) | 
|  | { | 
|  | if (cell->isPreciseAllocation()) | 
|  | return !m_lowerTierBits.concurrentTestAndClear(cell->preciseAllocation().lowerTierIndex()); | 
|  | AtomIndices atomIndices(cell); | 
|  | auto& bitsPtrRef = m_bits[atomIndices.blockIndex]; | 
|  | auto* bits = bitsPtrRef.get(); | 
|  | if (!bits) | 
|  | return false; | 
|  | return bits->concurrentTestAndClear(atomIndices.atomNumber); | 
|  | } | 
|  |  | 
|  | inline bool IsoCellSet::contains(HeapCell* cell) const | 
|  | { | 
|  | if (cell->isPreciseAllocation()) | 
|  | return !m_lowerTierBits.get(cell->preciseAllocation().lowerTierIndex()); | 
|  | AtomIndices atomIndices(cell); | 
|  | auto* bits = m_bits[atomIndices.blockIndex].get(); | 
|  | if (bits) | 
|  | return bits->get(atomIndices.atomNumber); | 
|  | return false; | 
|  | } | 
|  |  | 
|  | template<typename Func> | 
|  | void IsoCellSet::forEachMarkedCell(const Func& func) | 
|  | { | 
|  | BlockDirectory& directory = m_subspace.m_directory; | 
|  | (directory.m_bits.markingNotEmpty() & m_blocksWithBits).forEachSetBit( | 
|  | [&] (unsigned blockIndex) { | 
|  | MarkedBlock::Handle* block = directory.m_blocks[blockIndex]; | 
|  |  | 
|  | auto* bits = m_bits[blockIndex].get(); | 
|  | block->forEachMarkedCell( | 
|  | [&] (unsigned atomNumber, HeapCell* cell, HeapCell::Kind kind) -> IterationStatus { | 
|  | if (bits->get(atomNumber)) | 
|  | func(cell, kind); | 
|  | return IterationStatus::Continue; | 
|  | }); | 
|  | }); | 
|  |  | 
|  | CellAttributes attributes = m_subspace.attributes(); | 
|  | m_subspace.forEachPreciseAllocation( | 
|  | [&] (PreciseAllocation* allocation) { | 
|  | if (m_lowerTierBits.get(allocation->lowerTierIndex()) && allocation->isMarked()) | 
|  | func(allocation->cell(), attributes.cellKind); | 
|  | }); | 
|  | } | 
|  |  | 
|  | template<typename Visitor, typename Func> | 
|  | Ref<SharedTask<void(Visitor&)>> IsoCellSet::forEachMarkedCellInParallel(const Func& func) | 
|  | { | 
|  | class Task final : public SharedTask<void(Visitor&)> { | 
|  | public: | 
|  | Task(IsoCellSet& set, const Func& func) | 
|  | : m_set(set) | 
|  | , m_blockSource(set.parallelNotEmptyMarkedBlockSource()) | 
|  | , m_func(func) | 
|  | { | 
|  | } | 
|  |  | 
|  | void run(Visitor& visitor) final | 
|  | { | 
|  | while (MarkedBlock::Handle* handle = m_blockSource->run()) { | 
|  | unsigned blockIndex = handle->index(); | 
|  | auto* bits = m_set.m_bits[blockIndex].get(); | 
|  | handle->forEachMarkedCell( | 
|  | [&] (unsigned atomNumber, HeapCell* cell, HeapCell::Kind kind) -> IterationStatus { | 
|  | if (bits->get(atomNumber)) | 
|  | m_func(visitor, cell, kind); | 
|  | return IterationStatus::Continue; | 
|  | }); | 
|  | } | 
|  |  | 
|  | { | 
|  | Locker locker { m_lock }; | 
|  | if (!m_needToVisitPreciseAllocations) | 
|  | return; | 
|  | m_needToVisitPreciseAllocations = false; | 
|  | } | 
|  |  | 
|  | CellAttributes attributes = m_set.m_subspace.attributes(); | 
|  | m_set.m_subspace.forEachPreciseAllocation( | 
|  | [&] (PreciseAllocation* allocation) { | 
|  | if (m_set.m_lowerTierBits.get(allocation->lowerTierIndex()) && allocation->isMarked()) | 
|  | m_func(visitor, allocation->cell(), attributes.cellKind); | 
|  | }); | 
|  | } | 
|  |  | 
|  | private: | 
|  | IsoCellSet& m_set; | 
|  | Ref<SharedTask<MarkedBlock::Handle*()>> m_blockSource; | 
|  | Func m_func; | 
|  | Lock m_lock; | 
|  | bool m_needToVisitPreciseAllocations { true }; | 
|  | }; | 
|  |  | 
|  | return adoptRef(*new Task(*this, func)); | 
|  | } | 
|  |  | 
|  | template<typename Func> | 
|  | void IsoCellSet::forEachLiveCell(const Func& func) | 
|  | { | 
|  | BlockDirectory& directory = m_subspace.m_directory; | 
|  | m_blocksWithBits.forEachSetBit( | 
|  | [&] (unsigned blockIndex) { | 
|  | MarkedBlock::Handle* block = directory.m_blocks[blockIndex]; | 
|  |  | 
|  | auto* bits = m_bits[blockIndex].get(); | 
|  | block->forEachCell( | 
|  | [&] (unsigned atomNumber, HeapCell* cell, HeapCell::Kind kind) -> IterationStatus { | 
|  | if (bits->get(atomNumber) && block->isLive(cell)) | 
|  | func(cell, kind); | 
|  | return IterationStatus::Continue; | 
|  | }); | 
|  | }); | 
|  |  | 
|  | CellAttributes attributes = m_subspace.attributes(); | 
|  | m_subspace.forEachPreciseAllocation( | 
|  | [&] (PreciseAllocation* allocation) { | 
|  | if (m_lowerTierBits.get(allocation->lowerTierIndex()) && allocation->isLive()) | 
|  | func(allocation->cell(), attributes.cellKind); | 
|  | }); | 
|  | } | 
|  |  | 
|  | inline void IsoCellSet::clearLowerTierCell(unsigned index) | 
|  | { | 
|  | m_lowerTierBits.concurrentTestAndClear(index); | 
|  | } | 
|  |  | 
|  | } // namespace JSC | 
|  |  |