Copying collection shouldn't require O(live bytes) memory overhead
[WebKit-https.git] / Source / JavaScriptCore / heap / Heap.h
index ed9715d..88dc201 100644 (file)
 #ifndef Heap_h
 #define Heap_h
 
-#include "HandleHeap.h"
+#include "BlockAllocator.h"
+#include "CopyVisitor.h"
+#include "DFGCodeBlocks.h"
+#include "GCThreadSharedData.h"
+#include "HandleSet.h"
 #include "HandleStack.h"
-#include "MarkStack.h"
-#include "NewSpace.h"
-#include <wtf/Forward.h>
+#include "JITStubRoutineSet.h"
+#include "MarkedAllocator.h"
+#include "MarkedBlock.h"
+#include "MarkedBlockSet.h"
+#include "MarkedSpace.h"
+#include "Options.h"
+#include "SlotVisitor.h"
+#include "WeakHandleOwner.h"
+#include "WriteBarrierSupport.h"
 #include <wtf/HashCountedSet.h>
 #include <wtf/HashSet.h>
 
+#define COLLECT_ON_EVERY_ALLOCATION 0
+
 namespace JSC {
 
+    class CopiedSpace;
+    class CodeBlock;
+    class ExecutableBase;
     class GCActivityCallback;
+    class GCAwareJITStubRoutine;
     class GlobalCodeBlock;
+    class Heap;
     class HeapRootVisitor;
+    class IncrementalSweeper;
+    class JITStubRoutine;
     class JSCell;
     class JSGlobalData;
+    class JSStack;
     class JSValue;
     class LiveObjectIterator;
-    class MarkStack;
+    class LLIntOffsetsExtractor;
     class MarkedArgumentBuffer;
-    class RegisterFile;
-    class UString;
     class WeakGCHandlePool;
-    typedef MarkStack SlotVisitor;
+    class SlotVisitor;
 
-    typedef std::pair<JSValue, UString> ValueStringPair;
+    typedef std::pair<JSValue, WTF::String> ValueStringPair;
     typedef HashCountedSet<JSCell*> ProtectCountSet;
     typedef HashCountedSet<const char*> TypeCountSet;
 
     enum OperationInProgress { NoOperation, Allocation, Collection };
 
+    enum HeapType { SmallHeap, LargeHeap };
+
     class Heap {
         WTF_MAKE_NONCOPYABLE(Heap);
     public:
-        static Heap* heap(JSValue); // 0 for immediate values
-        static Heap* heap(JSCell*);
+        friend class JIT;
+        friend class GCThreadSharedData;
+        static Heap* heap(const JSValue); // 0 for immediate values
+        static Heap* heap(const JSCell*);
+
+        // This constant determines how many blocks we iterate between checks of our 
+        // deadline when calling Heap::isPagedOut. Decreasing it will cause us to detect 
+        // overstepping our deadline more quickly, while increasing it will cause 
+        // our scan to run faster. 
+        static const unsigned s_timeCheckResolution = 16;
 
         static bool isMarked(const void*);
         static bool testAndSetMarked(const void*);
-        static bool testAndClearMarked(const void*);
         static void setMarked(const void*);
 
+        static bool isWriteBarrierEnabled();
         static void writeBarrier(const JSCell*, JSValue);
         static void writeBarrier(const JSCell*, JSCell*);
+        static uint8_t* addressOfCardFor(JSCell*);
 
-        Heap(JSGlobalData*);
+        Heap(JSGlobalData*, HeapType);
         ~Heap();
-        void destroy(); // JSGlobalData must call destroy() before ~Heap().
+        JS_EXPORT_PRIVATE void lastChanceToFinalize();
 
         JSGlobalData* globalData() const { return m_globalData; }
-        NewSpace& markedSpace() { return m_newSpace; }
+        MarkedSpace& objectSpace() { return m_objectSpace; }
         MachineThreads& machineThreads() { return m_machineThreads; }
 
-        GCActivityCallback* activityCallback();
-        void setActivityCallback(PassOwnPtr<GCActivityCallback>);
+        JS_EXPORT_PRIVATE GCActivityCallback* activityCallback();
+        JS_EXPORT_PRIVATE void setActivityCallback(GCActivityCallback*);
+        JS_EXPORT_PRIVATE void setGarbageCollectionTimerEnabled(bool);
+
+        JS_EXPORT_PRIVATE IncrementalSweeper* sweeper();
 
         // true if an allocation or collection is in progress
         inline bool isBusy();
-
-        void* allocate(size_t);
-        void* allocate(NewSpace::SizeClass&);
-        void collectAllGarbage();
+        
+        MarkedAllocator& firstAllocatorWithoutDestructors() { return m_objectSpace.firstAllocator(); }
+        MarkedAllocator& allocatorForObjectWithoutDestructor(size_t bytes) { return m_objectSpace.allocatorFor(bytes); }
+        MarkedAllocator& allocatorForObjectWithNormalDestructor(size_t bytes) { return m_objectSpace.normalDestructorAllocatorFor(bytes); }
+        MarkedAllocator& allocatorForObjectWithImmortalStructureDestructor(size_t bytes) { return m_objectSpace.immortalStructureDestructorAllocatorFor(bytes); }
+        CopiedAllocator& storageAllocator() { return m_storageSpace.allocator(); }
+        CheckedBoolean tryAllocateStorage(size_t, void**);
+        CheckedBoolean tryReallocateStorage(void**, size_t, size_t);
+
+        typedef void (*Finalizer)(JSCell*);
+        JS_EXPORT_PRIVATE void addFinalizer(JSCell*, Finalizer);
+        void addCompiledCode(ExecutableBase*);
+
+        void notifyIsSafeToCollect() { m_isSafeToCollect = true; }
+        bool isSafeToCollect() const { return m_isSafeToCollect; }
+
+        JS_EXPORT_PRIVATE void collectAllGarbage();
+        enum SweepToggle { DoNotSweep, DoSweep };
+        bool shouldCollect();
+        void collect(SweepToggle);
 
         void reportExtraMemoryCost(size_t cost);
+        JS_EXPORT_PRIVATE void reportAbandonedObjectGraph();
 
-        void protect(JSValue);
-        bool unprotect(JSValue); // True when the protect count drops to 0.
-
-        bool contains(const void*);
-
-        size_t size();
-        size_t capacity();
-        size_t objectCount();
-        size_t globalObjectCount();
-        size_t protectedObjectCount();
-        size_t protectedGlobalObjectCount();
-        PassOwnPtr<TypeCountSet> protectedObjectTypeCounts();
-        PassOwnPtr<TypeCountSet> objectTypeCounts();
+        JS_EXPORT_PRIVATE void protect(JSValue);
+        JS_EXPORT_PRIVATE bool unprotect(JSValue); // True when the protect count drops to 0.
+        
+        void jettisonDFGCodeBlock(PassOwnPtr<CodeBlock>);
+
+        JS_EXPORT_PRIVATE size_t size();
+        JS_EXPORT_PRIVATE size_t capacity();
+        JS_EXPORT_PRIVATE size_t objectCount();
+        JS_EXPORT_PRIVATE size_t globalObjectCount();
+        JS_EXPORT_PRIVATE size_t protectedObjectCount();
+        JS_EXPORT_PRIVATE size_t protectedGlobalObjectCount();
+        JS_EXPORT_PRIVATE PassOwnPtr<TypeCountSet> protectedObjectTypeCounts();
+        JS_EXPORT_PRIVATE PassOwnPtr<TypeCountSet> objectTypeCounts();
+        void showStatistics();
 
         void pushTempSortVector(Vector<ValueStringPair>*);
         void popTempSortVector(Vector<ValueStringPair>*);
     
-        HashSet<MarkedArgumentBuffer*>& markListSet() { if (!m_markListSet) m_markListSet = new HashSet<MarkedArgumentBuffer*>; return *m_markListSet; }
+        HashSet<MarkedArgumentBuffer*>& markListSet() { if (!m_markListSet) m_markListSet = adoptPtr(new HashSet<MarkedArgumentBuffer*>); return *m_markListSet; }
         
         template<typename Functor> typename Functor::ReturnType forEachProtectedCell(Functor&);
         template<typename Functor> typename Functor::ReturnType forEachProtectedCell();
-        template<typename Functor> typename Functor::ReturnType forEachCell(Functor&);
-        template<typename Functor> typename Functor::ReturnType forEachCell();
-        template<typename Functor> typename Functor::ReturnType forEachBlock(Functor&);
-        template<typename Functor> typename Functor::ReturnType forEachBlock();
-        
-        HandleSlot allocateGlobalHandle() { return m_handleHeap.allocate(); }
-        HandleSlot allocateLocalHandle() { return m_handleStack.push(); }
 
+        HandleSet* handleSet() { return &m_handleSet; }
         HandleStack* handleStack() { return &m_handleStack; }
 
+        void getConservativeRegisterRoots(HashSet<JSCell*>& roots);
+
+        double lastGCLength() { return m_lastGCLength; }
+        void increaseLastGCLength(double amount) { m_lastGCLength += amount; }
+
+        JS_EXPORT_PRIVATE void deleteAllCompiledCode();
+
+        void didAllocate(size_t);
+        void didAbandon(size_t);
+
+        bool isPagedOut(double deadline);
+        void didStartVMShutdown();
+
     private:
-        typedef HashSet<MarkedBlock*>::iterator BlockIterator;
+        friend class CodeBlock;
+        friend class GCAwareJITStubRoutine;
+        friend class JITStubRoutine;
+        friend class LLIntOffsetsExtractor;
+        friend class MarkedSpace;
+        friend class MarkedAllocator;
+        friend class MarkedBlock;
+        friend class CopiedSpace;
+        friend class CopyVisitor;
+        friend class SlotVisitor;
+        friend class IncrementalSweeper;
+        friend class HeapStatistics;
+        template<typename T> friend void* allocateCell(Heap&);
+        template<typename T> friend void* allocateCell(Heap&, size_t);
+
+        void* allocateWithImmortalStructureDestructor(size_t); // For use with special objects whose Structures never die.
+        void* allocateWithNormalDestructor(size_t); // For use with objects that inherit directly or indirectly from JSDestructibleObject.
+        void* allocateWithoutDestructor(size_t); // For use with objects without destructors.
 
         static const size_t minExtraCost = 256;
         static const size_t maxExtraCost = 1024 * 1024;
+        
+        class FinalizerOwner : public WeakHandleOwner {
+            virtual void finalize(Handle<Unknown>, void* context);
+        };
 
-        bool isValidAllocation(size_t);
-        void* allocateSlowCase(size_t);
-        void reportExtraMemoryCostSlowCase(size_t);
-        void resetAllocator();
-
-        MarkedBlock* allocateBlock(size_t cellSize);
-        void freeBlocks(MarkedBlock*);
+        JS_EXPORT_PRIVATE bool isValidAllocation(size_t);
+        JS_EXPORT_PRIVATE void reportExtraMemoryCostSlowCase(size_t);
 
-        void clearMarks();
-        void markRoots();
+        void markRoots(bool fullGC);
         void markProtectedObjects(HeapRootVisitor&);
         void markTempSortVectors(HeapRootVisitor&);
-
-        enum SweepToggle { DoNotSweep, DoSweep };
-        void collect(SweepToggle);
-        void shrink();
-        void sweep();
-
-        RegisterFile& registerFile();
-
+        void copyBackingStores();
+        void harvestWeakReferences();
+        void finalizeUnconditionalFinalizers();
+        void deleteUnmarkedCompiledCode();
+        void zombifyDeadObjects();
+        void markDeadObjects();
+
+        JSStack& stack();
+        BlockAllocator& blockAllocator();
+
+        const HeapType m_heapType;
+        const size_t m_ramSize;
+        const size_t m_minBytesPerCycle;
+        size_t m_sizeAfterLastCollect;
+
+        size_t m_bytesAllocatedLimit;
+        size_t m_bytesAllocated;
+        size_t m_bytesAbandoned;
+        
         OperationInProgress m_operationInProgress;
-        NewSpace m_newSpace;
-        HashSet<MarkedBlock*> m_blocks;
+        BlockAllocator m_blockAllocator;
+        MarkedSpace m_objectSpace;
+        CopiedSpace m_storageSpace;
 
-        size_t m_extraCost;
+#if ENABLE(SIMPLE_HEAP_PROFILING)
+        VTableSpectrum m_destroyedTypeCounts;
+#endif
 
         ProtectCountSet m_protectedValues;
         Vector<Vector<ValueStringPair>* > m_tempSortingVectors;
-        HashSet<MarkedArgumentBuffer*>* m_markListSet;
+        OwnPtr<HashSet<MarkedArgumentBuffer*> > m_markListSet;
 
-        OwnPtr<GCActivityCallback> m_activityCallback;
-        
         MachineThreads m_machineThreads;
-        MarkStack m_markStack;
-        HandleHeap m_handleHeap;
+        
+        GCThreadSharedData m_sharedData;
+        SlotVisitor m_slotVisitor;
+        CopyVisitor m_copyVisitor;
+
+        HandleSet m_handleSet;
         HandleStack m_handleStack;
+        DFGCodeBlocks m_dfgCodeBlocks;
+        JITStubRoutineSet m_jitStubRoutines;
+        FinalizerOwner m_finalizerOwner;
+        
+        bool m_isSafeToCollect;
 
         JSGlobalData* m_globalData;
+        double m_lastGCLength;
+        double m_lastCodeDiscardTime;
+
+        DoublyLinkedList<ExecutableBase> m_compiledCode;
+        
+        GCActivityCallback* m_activityCallback;
+        IncrementalSweeper* m_sweeper;
+        Vector<MarkedBlock*> m_blockSnapshot;
+    };
+
+    struct MarkedBlockSnapshotFunctor : public MarkedBlock::VoidFunctor {
+        MarkedBlockSnapshotFunctor(Vector<MarkedBlock*>& blocks) 
+            : m_index(0) 
+            , m_blocks(blocks)
+        {
+        }
+    
+        void operator()(MarkedBlock* block) { m_blocks[m_index++] = block; }
+    
+        size_t m_index;
+        Vector<MarkedBlock*>& m_blocks;
     };
 
+    inline bool Heap::shouldCollect()
+    {
+        if (Options::gcMaxHeapSize())
+            return m_bytesAllocated > Options::gcMaxHeapSize() && m_isSafeToCollect && m_operationInProgress == NoOperation;
+#if ENABLE(GGC)
+        return m_objectSpace.nurseryWaterMark() >= m_minBytesPerCycle && m_isSafeToCollect && m_operationInProgress == NoOperation;
+#else
+        return m_bytesAllocated > m_bytesAllocatedLimit && m_isSafeToCollect && m_operationInProgress == NoOperation;
+#endif
+    }
+
     bool Heap::isBusy()
     {
         return m_operationInProgress != NoOperation;
     }
 
-    inline Heap* Heap::heap(JSCell* cell)
+    inline Heap* Heap::heap(const JSCell* cell)
     {
         return MarkedBlock::blockFor(cell)->heap();
     }
 
-    inline Heap* Heap::heap(JSValue v)
+    inline Heap* Heap::heap(const JSValue v)
     {
         if (!v.isCell())
             return 0;
@@ -190,35 +315,54 @@ namespace JSC {
         return MarkedBlock::blockFor(cell)->testAndSetMarked(cell);
     }
 
-    inline bool Heap::testAndClearMarked(const void* cell)
+    inline void Heap::setMarked(const void* cell)
     {
-        return MarkedBlock::blockFor(cell)->testAndClearMarked(cell);
+        MarkedBlock::blockFor(cell)->setMarked(cell);
     }
 
-    inline void Heap::setMarked(const void* cell)
+    inline bool Heap::isWriteBarrierEnabled()
     {
-        MarkedBlock::blockFor(cell)->setMarked(cell);
+#if ENABLE(GGC) || ENABLE(WRITE_BARRIER_PROFILING)
+        return true;
+#else
+        return false;
+#endif
     }
 
-    inline void Heap::writeBarrier(const JSCell*, JSValue)
+#if ENABLE(GGC)
+    inline uint8_t* Heap::addressOfCardFor(JSCell* cell)
     {
+        return MarkedBlock::blockFor(cell)->addressOfCardFor(cell);
     }
 
-    inline void Heap::writeBarrier(const JSCell*, JSCell*)
+    inline void Heap::writeBarrier(const JSCell* owner, JSCell*)
     {
+        WriteBarrierCounters::countWriteBarrier();
+        MarkedBlock* block = MarkedBlock::blockFor(owner);
+        if (block->isMarked(owner))
+            block->setDirtyObject(owner);
     }
 
-    inline bool Heap::contains(const void* x)
+    inline void Heap::writeBarrier(const JSCell* owner, JSValue value)
     {
-        if (!MarkedBlock::isAtomAligned(x))
-            return false;
+        if (!value)
+            return;
+        if (!value.isCell())
+            return;
+        writeBarrier(owner, value.asCell());
+    }
+#else
 
-        MarkedBlock* block = MarkedBlock::blockFor(x);
-        if (!block || !m_blocks.contains(block))
-            return false;
-            
-        return true;
+    inline void Heap::writeBarrier(const JSCell*, JSCell*)
+    {
+        WriteBarrierCounters::countWriteBarrier();
+    }
+
+    inline void Heap::writeBarrier(const JSCell*, JSValue)
+    {
+        WriteBarrierCounters::countWriteBarrier();
     }
+#endif
 
     inline void Heap::reportExtraMemoryCost(size_t cost)
     {
@@ -230,8 +374,8 @@ namespace JSC {
     {
         ProtectCountSet::iterator end = m_protectedValues.end();
         for (ProtectCountSet::iterator it = m_protectedValues.begin(); it != end; ++it)
-            functor(it->first);
-        m_handleHeap.forEachStrongHandle(functor, m_protectedValues);
+            functor(it->key);
+        m_handleSet.forEachStrongHandle(functor, m_protectedValues);
 
         return functor.returnValue();
     }
@@ -242,39 +386,37 @@ namespace JSC {
         return forEachProtectedCell(functor);
     }
 
-    template<typename Functor> inline typename Functor::ReturnType Heap::forEachCell(Functor& functor)
+    inline void* Heap::allocateWithNormalDestructor(size_t bytes)
     {
-        BlockIterator end = m_blocks.end();
-        for (BlockIterator it = m_blocks.begin(); it != end; ++it)
-            (*it)->forEachCell(functor);
-        return functor.returnValue();
+        ASSERT(isValidAllocation(bytes));
+        return m_objectSpace.allocateWithNormalDestructor(bytes);
     }
-
-    template<typename Functor> inline typename Functor::ReturnType Heap::forEachCell()
+    
+    inline void* Heap::allocateWithImmortalStructureDestructor(size_t bytes)
     {
-        Functor functor;
-        return forEachCell(functor);
+        ASSERT(isValidAllocation(bytes));
+        return m_objectSpace.allocateWithImmortalStructureDestructor(bytes);
     }
-
-    template<typename Functor> inline typename Functor::ReturnType Heap::forEachBlock(Functor& functor)
+    
+    inline void* Heap::allocateWithoutDestructor(size_t bytes)
     {
-        BlockIterator end = m_blocks.end();
-        for (BlockIterator it = m_blocks.begin(); it != end; ++it)
-            functor(*it);
-        return functor.returnValue();
+        ASSERT(isValidAllocation(bytes));
+        return m_objectSpace.allocateWithoutDestructor(bytes);
     }
-
-    template<typename Functor> inline typename Functor::ReturnType Heap::forEachBlock()
+   
+    inline CheckedBoolean Heap::tryAllocateStorage(size_t bytes, void** outPtr)
     {
-        Functor functor;
-        return forEachBlock(functor);
+        return m_storageSpace.tryAllocate(bytes, outPtr);
+    }
+    
+    inline CheckedBoolean Heap::tryReallocateStorage(void** ptr, size_t oldSize, size_t newSize)
+    {
+        return m_storageSpace.tryReallocate(ptr, oldSize, newSize);
     }
 
-    inline void* Heap::allocate(size_t bytes)
+    inline BlockAllocator& Heap::blockAllocator()
     {
-        ASSERT(isValidAllocation(bytes));
-        NewSpace::SizeClass& sizeClass = m_newSpace.sizeClassFor(bytes);
-        return allocate(sizeClass);
+        return m_blockAllocator;
     }
 
 } // namespace JSC