diff options
author | Jan Vorlicek <janvorli@microsoft.com> | 2019-01-19 21:02:00 +0100 |
---|---|---|
committer | Jan Kotas <jkotas@microsoft.com> | 2019-01-19 12:02:00 -0800 |
commit | d24a3ff0105ab5b9649b27451263e57e7ee6526b (patch) | |
tree | 1c5c5125863cfc1649ec5becd1b3fdcb53471380 | |
parent | dce086f9bbd047296942f615f20a4f9a1e3a8d26 (diff) | |
download | coreclr-d24a3ff0105ab5b9649b27451263e57e7ee6526b.tar.gz coreclr-d24a3ff0105ab5b9649b27451263e57e7ee6526b.tar.bz2 coreclr-d24a3ff0105ab5b9649b27451263e57e7ee6526b.zip |
Cleanup: Remove MethodDesc::GetLoaderAllocatorForCode (#22067)
This function was recently changed to just return the
MethodDesc::GetLoaderAllocator. This is a cleanup that removes the
function completely and replaces all of its usages.
-rw-r--r-- | src/vm/amd64/cgenamd64.cpp | 2 | ||||
-rw-r--r-- | src/vm/arm/stubs.cpp | 2 | ||||
-rw-r--r-- | src/vm/codeman.cpp | 6 | ||||
-rw-r--r-- | src/vm/comdelegate.cpp | 4 | ||||
-rw-r--r-- | src/vm/dynamicmethod.cpp | 6 | ||||
-rw-r--r-- | src/vm/fptrstubs.cpp | 2 | ||||
-rw-r--r-- | src/vm/gccover.cpp | 2 | ||||
-rw-r--r-- | src/vm/i386/stublinkerx86.cpp | 4 | ||||
-rw-r--r-- | src/vm/jitinterface.cpp | 2 | ||||
-rw-r--r-- | src/vm/method.cpp | 7 | ||||
-rw-r--r-- | src/vm/method.hpp | 7 | ||||
-rw-r--r-- | src/vm/precode.cpp | 2 |
12 files changed, 16 insertions, 30 deletions
diff --git a/src/vm/amd64/cgenamd64.cpp b/src/vm/amd64/cgenamd64.cpp index 79f05d24cd..814f04dd9f 100644 --- a/src/vm/amd64/cgenamd64.cpp +++ b/src/vm/amd64/cgenamd64.cpp @@ -634,7 +634,7 @@ INT32 rel32UsingJumpStub(INT32 UNALIGNED * pRel32, PCODE target, MethodDesc *pMe // If a domain is provided, the MethodDesc mustn't yet be set up to have one, or it must match the MethodDesc's domain, // unless we're in a compilation domain (NGen loads assemblies as domain-bound but compiles them as domain neutral). PRECONDITION(!pLoaderAllocator || !pMethod || pMethod->GetMethodDescChunk()->GetMethodTablePtr()->IsNull() || - pLoaderAllocator == pMethod->GetMethodDescChunk()->GetFirstMethodDesc()->GetLoaderAllocatorForCode() || IsCompilationProcess()); + pLoaderAllocator == pMethod->GetMethodDescChunk()->GetFirstMethodDesc()->GetLoaderAllocator() || IsCompilationProcess()); } CONTRACTL_END; diff --git a/src/vm/arm/stubs.cpp b/src/vm/arm/stubs.cpp index 8b59e4eff6..7e7435c107 100644 --- a/src/vm/arm/stubs.cpp +++ b/src/vm/arm/stubs.cpp @@ -3334,7 +3334,7 @@ Stub * StubLinkerCPU::CreateTailCallCopyArgsThunk(CORINFO_SIG_INFO * pSig, pSl->ThumbEmitJumpRegister(thumbRegLr); } - LoaderHeap* pHeap = pMD->GetLoaderAllocatorForCode()->GetStubHeap(); + LoaderHeap* pHeap = pMD->GetLoaderAllocator()->GetStubHeap(); return pSl->Link(pHeap); } diff --git a/src/vm/codeman.cpp b/src/vm/codeman.cpp index c02d4e8f16..0e5c18937c 100644 --- a/src/vm/codeman.cpp +++ b/src/vm/codeman.cpp @@ -631,7 +631,7 @@ BOOL EEJitManager::CodeHeapIterator::Next() // LoaderAllocator filter if (m_pLoaderAllocator && m_pCurrent) { - LoaderAllocator *pCurrentLoaderAllocator = m_pCurrent->GetLoaderAllocatorForCode(); + LoaderAllocator *pCurrentLoaderAllocator = m_pCurrent->GetLoaderAllocator(); if(pCurrentLoaderAllocator != m_pLoaderAllocator) continue; } @@ -2281,7 +2281,7 @@ void CodeHeapRequestInfo::Init() } CONTRACTL_END; if (m_pAllocator == NULL) - m_pAllocator = m_pMD->GetLoaderAllocatorForCode(); + m_pAllocator = m_pMD->GetLoaderAllocator(); m_isDynamicDomain = (m_pMD != NULL) ? m_pMD->IsLCGMethod() : false; m_isCollectible = m_pAllocator->IsCollectible() ? true : false; m_throwOnOutOfMemoryWithinRange = true; @@ -4986,7 +4986,7 @@ PCODE ExecutionManager::jumpStub(MethodDesc* pMD, PCODE target, if (pLoaderAllocator == NULL) { - pLoaderAllocator = pMD->GetLoaderAllocatorForCode(); + pLoaderAllocator = pMD->GetLoaderAllocator(); } _ASSERTE(pLoaderAllocator != NULL); diff --git a/src/vm/comdelegate.cpp b/src/vm/comdelegate.cpp index 8ceab3e4c8..12edc538dd 100644 --- a/src/vm/comdelegate.cpp +++ b/src/vm/comdelegate.cpp @@ -1018,7 +1018,7 @@ void COMDelegate::BindToMethod(DELEGATEREF *pRefThis, else #ifdef HAS_THISPTR_RETBUF_PRECODE if (pTargetMethod->IsStatic() && pTargetMethod->HasRetBuffArg() && IsRetBuffPassedAsFirstArg()) - pTargetCode = pTargetMethod->GetLoaderAllocatorForCode()->GetFuncPtrStubs()->GetFuncPtrStub(pTargetMethod, PRECODE_THISPTR_RETBUF); + pTargetCode = pTargetMethod->GetLoaderAllocator()->GetFuncPtrStubs()->GetFuncPtrStub(pTargetMethod, PRECODE_THISPTR_RETBUF); else #endif // HAS_THISPTR_RETBUF_PRECODE pTargetCode = pTargetMethod->GetMultiCallableAddrOfCode(); @@ -1882,7 +1882,7 @@ FCIMPL3(void, COMDelegate::DelegateConstruct, Object* refThisUNSAFE, Object* tar } #ifdef HAS_THISPTR_RETBUF_PRECODE else if (pMeth->HasRetBuffArg() && IsRetBuffPassedAsFirstArg()) - method = pMeth->GetLoaderAllocatorForCode()->GetFuncPtrStubs()->GetFuncPtrStub(pMeth, PRECODE_THISPTR_RETBUF); + method = pMeth->GetLoaderAllocator()->GetFuncPtrStubs()->GetFuncPtrStub(pMeth, PRECODE_THISPTR_RETBUF); #endif // HAS_THISPTR_RETBUF_PRECODE gc.refThis->SetTarget(gc.target); diff --git a/src/vm/dynamicmethod.cpp b/src/vm/dynamicmethod.cpp index cbf4f63a39..cbec6ae97d 100644 --- a/src/vm/dynamicmethod.cpp +++ b/src/vm/dynamicmethod.cpp @@ -848,7 +848,7 @@ void DynamicMethodDesc::Destroy() CONTRACTL_END; _ASSERTE(IsDynamicMethod()); - LoaderAllocator *pLoaderAllocator = GetLoaderAllocatorForCode(); + LoaderAllocator *pLoaderAllocator = GetLoaderAllocator(); LOG((LF_BCL, LL_INFO1000, "Level3 - Destroying DynamicMethod {0x%p}\n", this)); if (!m_pSig.IsNull()) @@ -924,8 +924,8 @@ void LCGMethodResolver::RecycleIndCells() } // Insert the linked list to the free list of the VirtualCallStubManager of the current domain. - // We should use GetLoaderAllocatorForCode because that is where the ind cell was allocated. - LoaderAllocator *pLoaderAllocator = GetDynamicMethod()->GetLoaderAllocatorForCode(); + // We should use GetLoaderAllocator because that is where the ind cell was allocated. + LoaderAllocator *pLoaderAllocator = GetDynamicMethod()->GetLoaderAllocator(); VirtualCallStubManager *pMgr = pLoaderAllocator->GetVirtualCallStubManager(); pMgr->InsertIntoRecycledIndCellList_Locked(cellhead, cellcurr); m_UsedIndCellList = NULL; diff --git a/src/vm/fptrstubs.cpp b/src/vm/fptrstubs.cpp index 1e326c4b90..79660eccaf 100644 --- a/src/vm/fptrstubs.cpp +++ b/src/vm/fptrstubs.cpp @@ -123,7 +123,7 @@ PCODE FuncPtrStubs::GetFuncPtrStub(MethodDesc * pMD, PrecodeType type) // { AllocMemTracker amt; - Precode* pNewPrecode = Precode::Allocate(type, pMD, pMD->GetLoaderAllocatorForCode(), &amt); + Precode* pNewPrecode = Precode::Allocate(type, pMD, pMD->GetLoaderAllocator(), &amt); if (target != NULL) { diff --git a/src/vm/gccover.cpp b/src/vm/gccover.cpp index 48dcf8b57e..95e5d2aa81 100644 --- a/src/vm/gccover.cpp +++ b/src/vm/gccover.cpp @@ -82,7 +82,7 @@ void SetupAndSprinkleBreakpoints( { // Allocate room for the GCCoverageInfo and copy of the method instructions size_t memSize = sizeof(GCCoverageInfo) + methodRegionInfo.hotSize + methodRegionInfo.coldSize; - GCCoverageInfo* gcCover = (GCCoverageInfo*)(void*) pMD->GetLoaderAllocatorForCode()->GetHighFrequencyHeap()->AllocAlignedMem(memSize, CODE_SIZE_ALIGN); + GCCoverageInfo* gcCover = (GCCoverageInfo*)(void*) pMD->GetLoaderAllocator()->GetHighFrequencyHeap()->AllocAlignedMem(memSize, CODE_SIZE_ALIGN); memset(gcCover, 0, sizeof(GCCoverageInfo)); diff --git a/src/vm/i386/stublinkerx86.cpp b/src/vm/i386/stublinkerx86.cpp index 7dfa083532..9902419c97 100644 --- a/src/vm/i386/stublinkerx86.cpp +++ b/src/vm/i386/stublinkerx86.cpp @@ -6430,7 +6430,7 @@ Stub * StubLinkerCPU::CreateTailCallCopyArgsThunk(CORINFO_SIG_INFO * pSig, EncodeGCOffsets(pSl, gcLayout); } - LoaderHeap* pHeap = pMD->GetLoaderAllocatorForCode()->GetStubHeap(); + LoaderHeap* pHeap = pMD->GetLoaderAllocator()->GetStubHeap(); return pSl->Link(pHeap); } #endif // DACCESS_COMPILE @@ -6805,7 +6805,7 @@ BOOL ThisPtrRetBufPrecode::SetTargetInterlocked(TADDR target, TADDR expected) _ASSERTE(m_rel32 == REL32_JMP_SELF); // Use pMD == NULL to allocate the jump stub in non-dynamic heap that has the same lifetime as the precode itself - INT32 newRel32 = rel32UsingJumpStub(&m_rel32, target, NULL /* pMD */, ((MethodDesc *)GetMethodDesc())->GetLoaderAllocatorForCode()); + INT32 newRel32 = rel32UsingJumpStub(&m_rel32, target, NULL /* pMD */, ((MethodDesc *)GetMethodDesc())->GetLoaderAllocator()); _ASSERTE(IS_ALIGNED(&m_rel32, sizeof(INT32))); FastInterlockExchange((LONG *)&m_rel32, (LONG)newRel32); diff --git a/src/vm/jitinterface.cpp b/src/vm/jitinterface.cpp index ef214f67e3..d05a040c25 100644 --- a/src/vm/jitinterface.cpp +++ b/src/vm/jitinterface.cpp @@ -5554,7 +5554,7 @@ void CEEInfo::getCallInfo( // We shouldn't be using GetLoaderAllocator here because for LCG, we need to get the // VirtualCallStubManager from where the stub will be used. // For normal methods there is no difference. - LoaderAllocator *pLoaderAllocator = m_pMethodBeingCompiled->GetLoaderAllocatorForCode(); + LoaderAllocator *pLoaderAllocator = m_pMethodBeingCompiled->GetLoaderAllocator(); VirtualCallStubManager *pMgr = pLoaderAllocator->GetVirtualCallStubManager(); PCODE addr = pMgr->GetCallStub(exactType, pTargetMD); diff --git a/src/vm/method.cpp b/src/vm/method.cpp index 967d1560c9..00df9d2d28 100644 --- a/src/vm/method.cpp +++ b/src/vm/method.cpp @@ -224,13 +224,6 @@ BaseDomain *MethodDesc::GetDomain() #ifndef DACCESS_COMPILE //******************************************************************************* -LoaderAllocator * MethodDesc::GetLoaderAllocatorForCode() -{ - return GetLoaderAllocator(); -} - - -//******************************************************************************* LoaderAllocator * MethodDesc::GetDomainSpecificLoaderAllocator() { if (GetLoaderModule()->IsCollectible()) diff --git a/src/vm/method.hpp b/src/vm/method.hpp index ac61d3391e..92a91c8305 100644 --- a/src/vm/method.hpp +++ b/src/vm/method.hpp @@ -524,13 +524,6 @@ public: PTR_LoaderAllocator GetLoaderAllocator(); - // GetLoaderAllocatorForCode returns the allocator with the responsibility for allocation. - // This is called from GetMulticallableAddrOfCode when allocating a small trampoline stub for the method. - // Normally a method in a shared domain will allocate memory for stubs in the shared domain. - // That has to be different for DynamicMethod as they need to be allocated always in the AppDomain - // that created the method. - LoaderAllocator * GetLoaderAllocatorForCode(); - // GetDomainSpecificLoaderAllocator returns the collectable loader allocator for collectable types // and the loader allocator in the current domain for non-collectable types LoaderAllocator * GetDomainSpecificLoaderAllocator(); diff --git a/src/vm/precode.cpp b/src/vm/precode.cpp index 38cf9aa82e..649fd1dd7f 100644 --- a/src/vm/precode.cpp +++ b/src/vm/precode.cpp @@ -479,7 +479,7 @@ void Precode::Reset() WRAPPER_NO_CONTRACT; MethodDesc* pMD = GetMethodDesc(); - Init(GetType(), pMD, pMD->GetLoaderAllocatorForCode()); + Init(GetType(), pMD, pMD->GetLoaderAllocator()); ClrFlushInstructionCache(this, SizeOf()); } |