--- old/src/share/vm/gc/g1/g1ConcurrentMark.cpp 2016-03-11 00:03:01.405880643 +0100 +++ new/src/share/vm/gc/g1/g1ConcurrentMark.cpp 2016-03-11 00:03:01.310877808 +0100 @@ -852,7 +852,7 @@ // marking. reset_marking_state(true /* clear_overflow */); - log_info(gc)("Concurrent Mark reset for overflow"); + log_info(gc, marking)("Concurrent Mark reset for overflow"); } } @@ -987,13 +987,12 @@ } }; -void G1ConcurrentMark::scanRootRegions() { +void G1ConcurrentMark::scan_root_regions() { // scan_in_progress() will have been set to true only if there was // at least one root region to scan. So, if it's false, we // should not attempt to do any further work. if (root_regions()->scan_in_progress()) { assert(!has_aborted(), "Aborting before root region scanning is finished not supported."); - GCTraceConcTime(Info, gc) tt("Concurrent Root Region Scan"); _parallel_marking_threads = calc_parallel_marking_threads(); assert(parallel_marking_threads() <= max_parallel_marking_threads(), @@ -1051,7 +1050,7 @@ register_concurrent_phase_end_common(true); } -void G1ConcurrentMark::markFromRoots() { +void G1ConcurrentMark::mark_from_roots() { // we might be tempted to assert that: // assert(asynch == !SafepointSynchronize::is_at_safepoint(), // "inconsistent argument?"); @@ -1114,7 +1113,6 @@ if (has_overflown()) { // Oops. We overflowed. Restart concurrent marking. _restart_for_overflow = true; - log_develop_trace(gc)("Remark led to restart for overflow."); // Verify the heap w.r.t. the previous marking bitmap. if (VerifyDuringGC) { @@ -1760,7 +1758,7 @@ g1h->trace_heap_after_concurrent_cycle(); } -void G1ConcurrentMark::completeCleanup() { +void G1ConcurrentMark::complete_cleanup() { if (has_aborted()) return; G1CollectedHeap* g1h = G1CollectedHeap::heap();