From 1099c60f0e3e0fa1059c716a6418a90527929482 Mon Sep 17 00:00:00 2001 From: Gang Zhao Date: Fri, 20 Dec 2024 12:05:30 -0800 Subject: [PATCH] Remove HeapCellIterator and AlignedHeapSegment::cells() Differential Revision: D66727534 --- include/hermes/VM/AlignedHeapSegment.h | 40 +------------------------- lib/VM/gcs/HadesGC.cpp | 11 +++++-- 2 files changed, 10 insertions(+), 41 deletions(-) diff --git a/include/hermes/VM/AlignedHeapSegment.h b/include/hermes/VM/AlignedHeapSegment.h index 4a7d96b197e..727e3b7a41a 100644 --- a/include/hermes/VM/AlignedHeapSegment.h +++ b/include/hermes/VM/AlignedHeapSegment.h @@ -12,20 +12,17 @@ #include "hermes/Support/OSCompat.h" #include "hermes/VM/AdviseUnused.h" #include "hermes/VM/AllocResult.h" -#include "hermes/VM/AllocSource.h" #include "hermes/VM/CardTableNC.h" -#include "hermes/VM/GCBase.h" -#include "hermes/VM/GCCell.h" #include "hermes/VM/HeapAlign.h" #include "llvh/Support/MathExtras.h" #include -#include namespace hermes { namespace vm { +class GCCell; class StorageProvider; #ifndef HERMESVM_LOG_HEAP_SEGMENT_SIZE @@ -191,30 +188,6 @@ class AlignedHeapSegment { 0, "Guard page must be aligned to likely page size"); - class HeapCellIterator : public llvh::iterator_facade_base< - HeapCellIterator, - std::forward_iterator_tag, - GCCell *> { - public: - HeapCellIterator(GCCell *cell) : cell_(cell) {} - - bool operator==(const HeapCellIterator &R) const { - return cell_ == R.cell_; - } - - GCCell *const &operator*() const { - return cell_; - } - - HeapCellIterator &operator++() { - cell_ = cell_->nextCell(); - return *this; - } - - private: - GCCell *cell_{nullptr}; - }; - /// Returns the index of the segment containing \p lowLim, which is required /// to be the start of its containing segment. (This can allow extra /// efficiency, in cases where the segment start has already been computed.) @@ -333,9 +306,6 @@ class AlignedHeapSegment { /// Returns the address at which the next allocation, if any, will occur. inline char *level() const; - /// Returns an iterator range corresponding to the cells in this segment. - inline llvh::iterator_range cells(); - /// Returns whether \p a and \p b are contained in the same /// AlignedHeapSegment. inline static bool containedInSame(const void *a, const void *b); @@ -425,7 +395,6 @@ class AlignedHeapSegment { AllocResult AlignedHeapSegment::alloc(uint32_t size) { assert(lowLim() != nullptr && "Cannot allocate in a null segment"); - assert(size >= sizeof(GCCell) && "cell must be larger than GCCell"); assert(isSizeHeapAligned(size) && "size must be heap aligned"); char *cellPtr; // Initialized in the if below. @@ -545,13 +514,6 @@ char *AlignedHeapSegment::level() const { return level_; } -llvh::iterator_range -AlignedHeapSegment::cells() { - return { - HeapCellIterator(reinterpret_cast(start())), - HeapCellIterator(reinterpret_cast(level()))}; -} - /* static */ bool AlignedHeapSegment::containedInSame(const void *a, const void *b) { return (reinterpret_cast(a) ^ reinterpret_cast(b)) < diff --git a/lib/VM/gcs/HadesGC.cpp b/lib/VM/gcs/HadesGC.cpp index e9bf33f4b28..9397b13856b 100644 --- a/lib/VM/gcs/HadesGC.cpp +++ b/lib/VM/gcs/HadesGC.cpp @@ -174,7 +174,10 @@ template void HadesGC::forAllObjsInSegment( hermes::vm::AlignedHeapSegment &seg, CallbackFunction callback) { - for (GCCell *cell : seg.cells()) { + for (GCCell *cell = reinterpret_cast(seg.start()), + *end = reinterpret_cast(seg.level()); + cell < end; + cell = cell->nextCell()) { // Skip free-list entries. if (!vmisa(cell)) { callback(cell); @@ -1052,7 +1055,11 @@ bool HadesGC::OldGen::sweepNext(bool backgroundThread) { char *freeRangeStart = nullptr, *freeRangeEnd = nullptr; size_t mergedCells = 0; int32_t segmentSweptBytes = 0; - for (GCCell *cell : segments_[sweepIterator_.segNumber].cells()) { + auto &seg = segments_[sweepIterator_.segNumber]; + for (GCCell *cell = reinterpret_cast(seg.start()), + *end = reinterpret_cast(seg.level()); + cell < end; + cell = cell->nextCell()) { assert(cell->isValid() && "Invalid cell in sweeping"); if (AlignedHeapSegment::getCellMarkBit(cell)) { // Cannot concurrently trim storage. Technically just checking