< prev index next >

src/hotspot/share/gc/parallel/asPSOldGen.cpp

Print this page

        

*** 90,111 **** assert(virtual_space()->is_aligned(gen_size_limit()), "not aligned"); assert(gen_size_limit() >= virtual_space()->committed_size(), "bad gen size"); ParallelScavengeHeap* heap = ParallelScavengeHeap::heap(); size_t result = gen_size_limit() - virtual_space()->committed_size(); ! size_t result_aligned = align_down(result, heap->generation_alignment()); return result_aligned; } size_t ASPSOldGen::available_for_contraction() { size_t uncommitted_bytes = virtual_space()->uncommitted_size(); if (uncommitted_bytes != 0) { return uncommitted_bytes; } ParallelScavengeHeap* heap = ParallelScavengeHeap::heap(); ! const size_t gen_alignment = heap->generation_alignment(); PSAdaptiveSizePolicy* policy = heap->size_policy(); const size_t working_size = used_in_bytes() + (size_t) policy->avg_promoted()->padded_average(); const size_t working_aligned = align_up(working_size, gen_alignment); const size_t working_or_min = MAX2(working_aligned, min_gen_size()); --- 90,111 ---- assert(virtual_space()->is_aligned(gen_size_limit()), "not aligned"); assert(gen_size_limit() >= virtual_space()->committed_size(), "bad gen size"); ParallelScavengeHeap* heap = ParallelScavengeHeap::heap(); size_t result = gen_size_limit() - virtual_space()->committed_size(); ! size_t result_aligned = align_down(result, heap->gen_alignment()); return result_aligned; } size_t ASPSOldGen::available_for_contraction() { size_t uncommitted_bytes = virtual_space()->uncommitted_size(); if (uncommitted_bytes != 0) { return uncommitted_bytes; } ParallelScavengeHeap* heap = ParallelScavengeHeap::heap(); ! const size_t gen_alignment = heap->gen_alignment(); PSAdaptiveSizePolicy* policy = heap->size_policy(); const size_t working_size = used_in_bytes() + (size_t) policy->avg_promoted()->padded_average(); const size_t working_aligned = align_up(working_size, gen_alignment); const size_t working_or_min = MAX2(working_aligned, min_gen_size());
< prev index next >