summaryrefslogtreecommitdiff
path: root/src/gc/gc.cpp
diff options
context:
space:
mode:
authorSuchiman <robinsue@live.de>2019-06-11 20:18:39 +0200
committerJan Kotas <jkotas@microsoft.com>2019-06-11 20:26:23 -0700
commit7d0bffa8e5cea35b498dedb36393e0620377d03e (patch)
tree2f56b915da19e509cf39936906be13f74e09b80a /src/gc/gc.cpp
parenta12705bfc76d6f7d7c9f795acffa92a539662b70 (diff)
downloadcoreclr-7d0bffa8e5cea35b498dedb36393e0620377d03e.tar.gz
coreclr-7d0bffa8e5cea35b498dedb36393e0620377d03e.tar.bz2
coreclr-7d0bffa8e5cea35b498dedb36393e0620377d03e.zip
Port typo fixes from CoreRT
Diffstat (limited to 'src/gc/gc.cpp')
-rw-r--r--src/gc/gc.cpp54
1 files changed, 27 insertions, 27 deletions
diff --git a/src/gc/gc.cpp b/src/gc/gc.cpp
index 35f46d4b90..e59c7a931a 100644
--- a/src/gc/gc.cpp
+++ b/src/gc/gc.cpp
@@ -3099,7 +3099,7 @@ void gc_heap::fire_pevents()
gc_data_global.condemned_generation,
gc_data_global.gen0_reduction_count,
gc_data_global.reason,
- gc_data_global.global_mechanims_p,
+ gc_data_global.global_mechanisms_p,
gc_data_global.pause_mode,
gc_data_global.mem_pressure);
@@ -3242,7 +3242,7 @@ gc_heap::dt_estimate_reclaim_space_p (gc_tuning_point tp, int gen_number)
}
// DTREVIEW: Right now we only estimate gen2 fragmentation.
-// on 64-bit though we should consider gen1 or even gen0 fragmentatioin as
+// on 64-bit though we should consider gen1 or even gen0 fragmentation as
// well
inline BOOL
gc_heap::dt_estimate_high_frag_p (gc_tuning_point tp, int gen_number, uint64_t available_mem)
@@ -3798,7 +3798,7 @@ gc_heap* seg_mapping_table_heap_of_worker (uint8_t* o)
gc_heap* hp = ((o > entry->boundary) ? entry->h1 : entry->h0);
- dprintf (2, ("checking obj %Ix, index is %Id, entry: boundry: %Ix, h0: %Ix, seg0: %Ix, h1: %Ix, seg1: %Ix",
+ dprintf (2, ("checking obj %Ix, index is %Id, entry: boundary: %Ix, h0: %Ix, seg0: %Ix, h1: %Ix, seg1: %Ix",
o, index, (entry->boundary + 1),
(uint8_t*)(entry->h0), (uint8_t*)(entry->seg0),
(uint8_t*)(entry->h1), (uint8_t*)(entry->seg1)));
@@ -3867,7 +3867,7 @@ heap_segment* seg_mapping_table_segment_of (uint8_t* o)
size_t index = (size_t)o >> gc_heap::min_segment_size_shr;
seg_mapping* entry = &seg_mapping_table[index];
- dprintf (2, ("checking obj %Ix, index is %Id, entry: boundry: %Ix, seg0: %Ix, seg1: %Ix",
+ dprintf (2, ("checking obj %Ix, index is %Id, entry: boundary: %Ix, seg0: %Ix, seg1: %Ix",
o, index, (entry->boundary + 1),
(uint8_t*)(entry->seg0), (uint8_t*)(entry->seg1)));
@@ -5573,7 +5573,7 @@ public:
// Supposedly Pinned objects cannot have references but we are seeing some from pinvoke
// frames. Also if it's an artificially pinned plug created by us, it can certainly
// have references.
- // We know these cases will be rare so we can optimize this to be only allocated on decommand.
+ // We know these cases will be rare so we can optimize this to be only allocated on demand.
gap_reloc_pair saved_post_plug_reloc;
// We need to calculate this after we are done with plan phase and before compact
@@ -5845,7 +5845,7 @@ void gc_mechanisms::record (gc_history_global* history)
history->reason = reason;
history->pause_mode = (int)pause_mode;
history->mem_pressure = entry_memory_load;
- history->global_mechanims_p = 0;
+ history->global_mechanisms_p = 0;
// start setting the boolean values.
if (concurrent)
@@ -6374,7 +6374,7 @@ void gc_heap::set_pinned_info (uint8_t* last_pinned_plug, size_t plug_len, gener
size_t gc_heap::deque_pinned_plug ()
{
- dprintf (3, ("dequed: %Id", mark_stack_bos));
+ dprintf (3, ("deque: %Id", mark_stack_bos));
size_t m = mark_stack_bos;
mark_stack_bos++;
return m;
@@ -7265,7 +7265,7 @@ int gc_heap::grow_brick_card_tables (uint8_t* start,
if ((la != saved_g_lowest_address ) || (ha != saved_g_highest_address))
{
{
- //modify the higest address so the span covered
+ //modify the highest address so the span covered
//is twice the previous one.
uint8_t* top = (uint8_t*)0 + Align (GCToOSInterface::GetVirtualMemoryLimit());
// On non-Windows systems, we get only an approximate value that can possibly be
@@ -8520,7 +8520,7 @@ void gc_heap::combine_mark_lists()
assert (end_of_list < &g_mark_list [n_heaps*mark_list_size]);
if (end_of_list > &g_mark_list[0])
_sort (&g_mark_list[0], end_of_list, 0);
- //adjust the mark_list to the begining of the resulting mark list.
+ //adjust the mark_list to the beginning of the resulting mark list.
for (int i = 0; i < n_heaps; i++)
{
g_heaps [i]->mark_list = g_mark_list;
@@ -8531,7 +8531,7 @@ void gc_heap::combine_mark_lists()
else
{
uint8_t** end_of_list = g_mark_list;
- //adjust the mark_list to the begining of the resulting mark list.
+ //adjust the mark_list to the beginning of the resulting mark list.
//put the index beyond the end to turn off mark list processing
for (int i = 0; i < n_heaps; i++)
{
@@ -8556,7 +8556,7 @@ class seg_free_spaces
struct free_space_bucket
{
seg_free_space* free_space;
- ptrdiff_t count_add; // Assigned when we first contruct the array.
+ ptrdiff_t count_add; // Assigned when we first construct the array.
ptrdiff_t count_fit; // How many items left when we are fitting plugs.
};
@@ -8840,7 +8840,7 @@ public:
// BARTOKTODO (4841): this code path is disabled (see can_fit_all_blocks_p) until we take alignment requirements into account
_ASSERTE(requiredAlignment == DATA_ALIGNMENT && false);
#endif // FEATURE_STRUCTALIGN
- // TODO: this is also not large alignment ready. We would need to consider alignment when chosing the
+ // TODO: this is also not large alignment ready. We would need to consider alignment when choosing the
// the bucket.
size_t plug_size_to_fit = plug_size;
@@ -9760,7 +9760,7 @@ void gc_heap::restart_vm()
{
//assert (generation_allocation_pointer (youngest_generation) == 0);
dprintf (3, ("Restarting EE"));
- STRESS_LOG0(LF_GC, LL_INFO10000, "Concurrent GC: Retarting EE\n");
+ STRESS_LOG0(LF_GC, LL_INFO10000, "Concurrent GC: Restarting EE\n");
ee_proceed_event.Set();
}
@@ -10652,7 +10652,7 @@ gc_heap::init_gc_heap (int h_number)
{
#ifndef INTERIOR_POINTERS
//set the brick_table for large objects
- //but default value is clearded
+ //but default value is cleared
//clear_brick_table ((uint8_t*)heap_segment_mem (lseg),
// (uint8_t*)heap_segment_reserved (lseg));
@@ -15203,7 +15203,7 @@ int gc_heap::generation_to_condemn (int n_initial,
}
}
- //figure out which ephemeral generation is too fragramented
+ //figure out which ephemeral generation is too fragmented
temp_gen = n;
for (i = n+1; i < max_generation; i++)
{
@@ -25006,7 +25006,7 @@ void gc_heap::copy_cards_range (uint8_t* dest, uint8_t* src, size_t len, BOOL co
clear_card_for_addresses (dest, dest + len);
}
-// POPO TODO: We should actually just recover the artifically made gaps here..because when we copy
+// POPO TODO: We should actually just recover the artificially made gaps here..because when we copy
// we always copy the earlier plugs first which means we won't need the gap sizes anymore. This way
// we won't need to individually recover each overwritten part of plugs.
inline
@@ -25989,7 +25989,7 @@ BOOL gc_heap::commit_mark_array_bgc_init (uint32_t* mark_array_addr)
// the mark_array flag for these segments will remain the same.
BOOL gc_heap::commit_new_mark_array (uint32_t* new_mark_array_addr)
{
- dprintf (GC_TABLE_LOG, ("commiting existing segs on MA %Ix", new_mark_array_addr));
+ dprintf (GC_TABLE_LOG, ("committing existing segs on MA %Ix", new_mark_array_addr));
generation* gen = generation_of (max_generation);
heap_segment* seg = heap_segment_in_range (generation_start_segment (gen));
while (1)
@@ -27220,7 +27220,7 @@ BOOL gc_heap::prepare_bgc_thread(gc_heap* gh)
gh->bgc_threads_timeout_cs.Enter();
if (!(gh->bgc_thread_running))
{
- dprintf (2, ("GC thread not runnning"));
+ dprintf (2, ("GC thread not running"));
if ((gh->bgc_thread == 0) && create_bgc_thread(gh))
{
success = TRUE;
@@ -28545,7 +28545,7 @@ void gc_heap::count_plug (size_t last_plug_size, uint8_t*& last_plug)
{
deque_pinned_plug();
update_oldest_pinned_plug();
- dprintf (3, ("dequed pin,now oldest pin is %Ix", pinned_plug (oldest_pin())));
+ dprintf (3, ("deque pin,now oldest pin is %Ix", pinned_plug (oldest_pin())));
}
else
{
@@ -34669,7 +34669,7 @@ static int32_t GCStressCurCount = 0;
static int32_t GCStressStartAtJit = -1;
// the maximum number of foreground GCs we'll induce during one BGC
-// (this number does not include "naturally" occuring GCs).
+// (this number does not include "naturally" occurring GCs).
static int32_t GCStressMaxFGCsPerBGC = -1;
// CLRRandom implementation can produce FPU exceptions if
@@ -34830,7 +34830,7 @@ bool GCHeap::StressHeap(gc_alloc_context * context)
if (str)
{
// Chop off the end of the string and form a new object out of it.
- // This will 'free' an object at the begining of the heap, which will
+ // This will 'free' an object at the beginning of the heap, which will
// force data movement. Note that we can only do this so many times.
// before we have to move on to the next string.
unsigned sizeOfNewObj = (unsigned)Align(min_obj_size * 31);
@@ -35552,7 +35552,7 @@ void gc_heap::record_interesting_info_per_heap()
heap_number,
(size_t)settings.gc_index,
settings.condemned_generation,
- // TEMP - I am just doing this for wks GC 'cuase I wanna see the pattern of doing C/S GCs.
+ // TEMP - I am just doing this for wks GC 'cause I wanna see the pattern of doing C/S GCs.
(settings.compaction ? (((compact_reason >= 0) && gc_heap_compact_reason_mandatory_p[compact_reason]) ? "M" : "W") : ""), // compaction
((expand_mechanism >= 0)? "X" : ""), // EX
((expand_mechanism == expand_reuse_normal) ? "X" : ""), // NF
@@ -35994,7 +35994,7 @@ GCHeap::GarbageCollectGeneration (unsigned int gen, gc_reason reason)
size_t GCHeap::GetTotalBytesInUse ()
{
#ifdef MULTIPLE_HEAPS
- //enumarate all the heaps and get their size.
+ //enumerate all the heaps and get their size.
size_t tot_size = 0;
for (int i = 0; i < gc_heap::n_heaps; i++)
{
@@ -36225,10 +36225,10 @@ int GCHeap::GetLOHCompactionMode()
return pGenGCHeap->loh_compaction_mode;
}
-void GCHeap::SetLOHCompactionMode (int newLOHCompactionyMode)
+void GCHeap::SetLOHCompactionMode (int newLOHCompactionMode)
{
#ifdef FEATURE_LOH_COMPACTION
- pGenGCHeap->loh_compaction_mode = (gc_loh_compaction_mode)newLOHCompactionyMode;
+ pGenGCHeap->loh_compaction_mode = (gc_loh_compaction_mode)newLOHCompactionMode;
#endif //FEATURE_LOH_COMPACTION
}
@@ -36341,7 +36341,7 @@ HRESULT GCHeap::GetGcCounters(int gen, gc_counters* counters)
counters->promoted_size = 0;
counters->collection_count = 0;
- //enumarate all the heaps and get their counters.
+ //enumerate all the heaps and get their counters.
for (int i = 0; i < gc_heap::n_heaps; i++)
{
dynamic_data* dd = gc_heap::g_heaps [i]->dynamic_data_of (gen);
@@ -36461,7 +36461,7 @@ Object* GCHeap::GetNextFinalizableObject()
if (O)
return O;
}
- //return the first non crtitical/critical one in the first queue.
+ //return the first non critical/critical one in the first queue.
for (int hn = 0; hn < gc_heap::n_heaps; hn++)
{
gc_heap* hp = gc_heap::g_heaps [hn];