// Licensed to the .NET Foundation under one or more agreements. // The .NET Foundation licenses this file to you under the MIT license. // See the LICENSE file in the project root for more information. /*++ Module Name: gc.h --*/ #ifndef __GC_H #define __GC_H #include "gcinterface.h" #include "env/gcenv.ee.h" #ifdef FEATURE_STANDALONE_GC #include "gcenv.ee.standalone.inl" #endif // FEATURE_STANDALONE_GC /* * Promotion Function Prototypes */ typedef void enum_func (Object*); // callback functions for heap walkers typedef void object_callback_func(void * pvContext, void * pvDataLoc); /*!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!*/ /* If you modify failure_get_memory and */ /* oom_reason be sure to make the corresponding */ /* changes in toolbox\sos\strike\strike.cpp. */ /*!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!*/ enum failure_get_memory { fgm_no_failure = 0, fgm_reserve_segment = 1, fgm_commit_segment_beg = 2, fgm_commit_eph_segment = 3, fgm_grow_table = 4, fgm_commit_table = 5 }; struct fgm_history { failure_get_memory fgm; size_t size; size_t available_pagefile_mb; BOOL loh_p; void set_fgm (failure_get_memory f, size_t s, BOOL l) { fgm = f; size = s; loh_p = l; } }; enum oom_reason { oom_no_failure = 0, oom_budget = 1, oom_cant_commit = 2, oom_cant_reserve = 3, oom_loh = 4, oom_low_mem = 5, oom_unproductive_full_gc = 6 }; // TODO : it would be easier to make this an ORed value enum gc_reason { reason_alloc_soh = 0, reason_induced = 1, reason_lowmemory = 2, reason_empty = 3, reason_alloc_loh = 4, reason_oos_soh = 5, reason_oos_loh = 6, reason_induced_noforce = 7, // it's an induced GC and doesn't have to be blocking. reason_gcstress = 8, // this turns into reason_induced & gc_mechanisms.stress_induced = true reason_lowmemory_blocking = 9, reason_induced_compacting = 10, reason_lowmemory_host = 11, reason_max }; struct oom_history { oom_reason reason; size_t alloc_size; uint8_t* reserved; uint8_t* allocated; size_t gc_index; failure_get_memory fgm; size_t size; size_t available_pagefile_mb; BOOL loh_p; }; /* forward declerations */ class CObjectHeader; class Object; class IGCHeapInternal; /* misc defines */ #define LARGE_OBJECT_SIZE ((size_t)(85000)) #ifdef GC_CONFIG_DRIVEN #define MAX_GLOBAL_GC_MECHANISMS_COUNT 6 GARY_DECL(size_t, gc_global_mechanisms, MAX_GLOBAL_GC_MECHANISMS_COUNT); #endif //GC_CONFIG_DRIVEN #ifdef DACCESS_COMPILE class DacHeapWalker; #endif #ifdef _DEBUG #define _LOGALLOC #endif #define MP_LOCKS namespace WKS { ::IGCHeapInternal* CreateGCHeap(); class GCHeap; class gc_heap; } #if defined(FEATURE_SVR_GC) namespace SVR { ::IGCHeapInternal* CreateGCHeap(); class GCHeap; class gc_heap; } #endif // defined(FEATURE_SVR_GC) #ifdef STRESS_HEAP #define IN_STRESS_HEAP(x) x #define STRESS_HEAP_ARG(x) ,x #else // STRESS_HEAP #define IN_STRESS_HEAP(x) #define STRESS_HEAP_ARG(x) #endif // STRESS_HEAP //dynamic data interface struct gc_counters { size_t current_size; size_t promoted_size; size_t collection_count; }; enum bgc_state { bgc_not_in_process = 0, bgc_initialized, bgc_reset_ww, bgc_mark_handles, bgc_mark_stack, bgc_revisit_soh, bgc_revisit_loh, bgc_overflow_soh, bgc_overflow_loh, bgc_final_marking, bgc_sweep_soh, bgc_sweep_loh, bgc_plan_phase }; enum changed_seg_state { seg_deleted, seg_added }; void record_changed_seg (uint8_t* start, uint8_t* end, size_t current_gc_index, bgc_state current_bgc_state, changed_seg_state changed_state); #ifdef GC_CONFIG_DRIVEN void record_global_mechanism (int mech_index); #endif //GC_CONFIG_DRIVEN struct alloc_context : gc_alloc_context { #ifdef FEATURE_SVR_GC inline SVR::GCHeap* get_alloc_heap() { return static_cast(gc_reserved_1); } inline void set_alloc_heap(SVR::GCHeap* heap) { gc_reserved_1 = heap; } inline SVR::GCHeap* get_home_heap() { return static_cast(gc_reserved_2); } inline void set_home_heap(SVR::GCHeap* heap) { gc_reserved_2 = heap; } #endif // FEATURE_SVR_GC }; class IGCHeapInternal : public IGCHeap { friend struct ::_DacGlobals; #ifdef DACCESS_COMPILE friend class ClrDataAccess; #endif public: virtual ~IGCHeapInternal() {} private: virtual Object* AllocAlign8Common (void* hp, alloc_context* acontext, size_t size, uint32_t flags) = 0; public: virtual int GetNumberOfHeaps () = 0; virtual int GetHomeHeapNumber () = 0; virtual size_t GetPromotedBytes(int heap_index) = 0; unsigned GetMaxGeneration() { return IGCHeap::maxGeneration; } BOOL IsValidSegmentSize(size_t cbSize) { //Must be aligned on a Mb and greater than 4Mb return (((cbSize & (1024*1024-1)) ==0) && (cbSize >> 22)); } BOOL IsValidGen0MaxSize(size_t cbSize) { return (cbSize >= 64*1024); } BOOL IsLargeObject(MethodTable *mt) { WRAPPER_NO_CONTRACT; return mt->GetBaseSize() >= LARGE_OBJECT_SIZE; } protected: public: #if defined(FEATURE_BASICFREEZE) && defined(VERIFY_HEAP) // Return TRUE if object lives in frozen segment virtual BOOL IsInFrozenSegment (Object * object) = 0; #endif // defined(FEATURE_BASICFREEZE) && defined(VERIFY_HEAP) }; // Go through and touch (read) each page straddled by a memory block. void TouchPages(void * pStart, size_t cb); #ifdef WRITE_BARRIER_CHECK void updateGCShadow(Object** ptr, Object* val); #endif // the method table for the WeakReference class extern MethodTable *pWeakReferenceMT; // The canonical method table for WeakReference extern MethodTable *pWeakReferenceOfTCanonMT; extern void FinalizeWeakReference(Object * obj); // The single GC heap instance, shared with the VM. extern IGCHeapInternal* g_theGCHeap; #ifndef DACCESS_COMPILE inline BOOL IsGCInProgress(bool bConsiderGCStart = FALSE) { WRAPPER_NO_CONTRACT; return g_theGCHeap != nullptr ? g_theGCHeap->IsGCInProgressHelper(bConsiderGCStart) : false; } #endif // DACCESS_COMPILE inline BOOL IsServerHeap() { LIMITED_METHOD_CONTRACT; #ifdef FEATURE_SVR_GC _ASSERTE(IGCHeap::gcHeapType != IGCHeap::GC_HEAP_INVALID); return (IGCHeap::gcHeapType == IGCHeap::GC_HEAP_SVR); #else // FEATURE_SVR_GC return false; #endif // FEATURE_SVR_GC } #endif // __GC_H