src/share/vm/gc_implementation/g1/heapRegionSets.cpp
Index
Unified diffs
Context diffs
Sdiffs
Patch
New
Old
Previous File
Next File
*** old/src/share/vm/gc_implementation/g1/heapRegionSets.cpp Tue Feb 11 11:36:03 2014
--- new/src/share/vm/gc_implementation/g1/heapRegionSets.cpp Tue Feb 11 11:36:03 2014
*** 22,32 ****
--- 22,32 ----
*
*/
#include "precompiled.hpp"
#include "gc_implementation/g1/heapRegionRemSet.hpp"
- #include "gc_implementation/g1/heapRegionSets.hpp"
// Note on the check_mt_safety() methods below:
//
// Verification of the "master" heap region sets / lists that are
// maintained by G1CollectedHeap is always done during a STW pause and
*** 37,117 ****
--- 37,109 ----
// for the "master" heap region sets / lists, the check_mt_safety()
// method should include the VM thread / STW case.
//////////////////// FreeRegionList ////////////////////
! const char* FreeRegionList::verify_region_extra(HeapRegion* hr) {
if (hr->is_young()) {
! return "the region should not be young";
}
// The superclass will check that the region is empty and
// not humongous.
return HeapRegionLinkedList::verify_region_extra(hr);
! void FreeRegionList::verify_list() {
+ HeapRegion* curr = head();
! HeapRegion* prev1 = NULL;
+ HeapRegion* prev0 = NULL;
+ uint count = 0;
+ size_t capacity = 0;
+ while (curr != NULL) {
+ verify_region(curr);
+
+ count++;
+ guarantee(count < _unrealistically_long_length,
+ hrs_err_msg("[%s] the calculated length: %u seems very long, is there maybe a cycle? curr: "PTR_FORMAT" prev0: "PTR_FORMAT" " "prev1: "PTR_FORMAT" length: %u", name(), count, curr, prev0, prev1, length()));
+
+ capacity += curr->capacity();
+
+ prev1 = prev0;
+ prev0 = curr;
+ curr = curr->next();
+ }
+
+ guarantee(tail() == prev0, err_msg("Expected %s to end with %u but it ended with %u.", name(), tail()->hrs_index(), prev0->hrs_index()));
+
+ guarantee(length() == count, err_msg("%s count mismatch. Expected %u, actual %u.", name(), length(), count));
+ guarantee(total_capacity_bytes() == capacity, err_msg("%s capacity mismatch. Expected " SIZE_FORMAT ", actual " SIZE_FORMAT,
+ name(), total_capacity_bytes(), capacity));
}
//////////////////// MasterFreeRegionList ////////////////////
! const char* MasterFreeRegionList::verify_region_extra(HeapRegion* hr) {
// We should reset the RSet for parallel iteration before we add it
// to the master free list so that it is ready when the region is
// re-allocated.
if (!hr->rem_set()->verify_ready_for_par_iteration()) {
return "the region's RSet should be ready for parallel iteration";
}
return FreeRegionList::verify_region_extra(hr);
}
bool MasterFreeRegionList::check_mt_safety() {
! bool MasterFreeRegionListMtSafeChecker::check() {
// Master Free List MT safety protocol:
// (a) If we're at a safepoint, operations on the master free list
// should be invoked by either the VM thread (which will serialize
// them) or by the GC workers while holding the
// FreeList_lock.
// (b) If we're not at a safepoint, operations on the master free
// list should be invoked while holding the Heap_lock.
if (SafepointSynchronize::is_at_safepoint()) {
guarantee(Thread::current()->is_VM_thread() ||
! FreeList_lock->owned_by_self(), "master free list MT safety protocol at a safepoint");
hrs_ext_msg(this, "master free list MT safety protocol "
"at a safepoint"));
} else {
! guarantee(Heap_lock->owned_by_self(), "master free list MT safety protocol outside a safepoint");
hrs_ext_msg(this, "master free list MT safety protocol "
"outside a safepoint"));
}
! return FreeRegionList::check_mt_safety();
! return true;
}
//////////////////// SecondaryFreeRegionList ////////////////////
! bool SecondaryFreeRegionList::check_mt_safety() {
! bool SecondaryFreeRegionListMtSafeChecker::check() {
// Secondary Free List MT safety protocol:
// Operations on the secondary free list should always be invoked
// while holding the SecondaryFreeList_lock.
! guarantee(SecondaryFreeList_lock->owned_by_self(), "secondary free list MT safety protocol");
hrs_ext_msg(this, "secondary free list MT safety protocol"));
! return FreeRegionList::check_mt_safety();
! return true;
}
//////////////////// OldRegionSet ////////////////////
! const char* OldRegionSet::verify_region_extra(HeapRegion* hr) {
if (hr->is_young()) {
return "the region should not be young";
}
// The superclass will check that the region is not empty and not
// humongous.
return HeapRegionSet::verify_region_extra(hr);
}
//////////////////// MasterOldRegionSet ////////////////////
bool MasterOldRegionSet::check_mt_safety() {
! bool OldRegionSetMtSafeChecker::check() {
// Master Old Set MT safety protocol:
// (a) If we're at a safepoint, operations on the master old set
// should be invoked:
// - by the VM thread (which will serialize them), or
// - by the GC workers while holding the FreeList_lock, if we're
*** 122,175 ****
--- 114,150 ----
// safepoint for a cleanup pause.
// (b) If we're not at a safepoint, operations on the master old set
// should be invoked while holding the Heap_lock.
if (SafepointSynchronize::is_at_safepoint()) {
- guarantee(Thread::current()->is_VM_thread() ||
_phase == HRSPhaseEvacuation && FreeList_lock->owned_by_self() ||
_phase == HRSPhaseCleanup && OldSets_lock->owned_by_self(),
hrs_ext_msg(this, "master old set MT safety protocol "
"at a safepoint"));
+ || FreeList_lock->owned_by_self() || OldSets_lock->owned_by_self(),
+ "master old set MT safety protocol at a safepoint");
} else {
! guarantee(Heap_lock->owned_by_self(), "master old set MT safety protocol outside a safepoint");
hrs_ext_msg(this, "master old set MT safety protocol "
"outside a safepoint"));
}
! return OldRegionSet::check_mt_safety();
! return true;
}
//////////////////// HumongousRegionSet ////////////////////
! const char* HumongousRegionSet::verify_region_extra(HeapRegion* hr) {
if (hr->is_young()) {
return "the region should not be young";
}
// The superclass will check that the region is not empty and
// humongous.
return HeapRegionSet::verify_region_extra(hr);
}
//////////////////// MasterHumongousRegionSet ////////////////////
bool MasterHumongousRegionSet::check_mt_safety() {
// Master Humongous Set MT safety protocol:
! bool HumongousRegionSetMtSafeChecker::check() {
+ // Humongous Set MT safety protocol:
// (a) If we're at a safepoint, operations on the master humongous
// set should be invoked by either the VM thread (which will
// serialize them) or by the GC workers while holding the
// OldSets_lock.
// (b) If we're not at a safepoint, operations on the master
// humongous set should be invoked while holding the Heap_lock.
if (SafepointSynchronize::is_at_safepoint()) {
guarantee(Thread::current()->is_VM_thread() ||
OldSets_lock->owned_by_self(),
hrs_ext_msg(this, "master humongous set MT safety protocol "
"at a safepoint"));
+ "master humongous set MT safety protocol at a safepoint");
} else {
guarantee(Heap_lock->owned_by_self(),
hrs_ext_msg(this, "master humongous set MT safety protocol "
"outside a safepoint"));
+ "master humongous set MT safety protocol outside a safepoint");
}
! return HumongousRegionSet::check_mt_safety();
! return true;
}
src/share/vm/gc_implementation/g1/heapRegionSets.cpp
Index
Unified diffs
Context diffs
Sdiffs
Patch
New
Old
Previous File
Next File