14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #include "precompiled.hpp"
26 #include "gc/g1/g1CollectedHeap.hpp"
27 #include "gc/g1/g1HeapSizingPolicy.hpp"
28 #include "gc/g1/g1Analytics.hpp"
29 #include "logging/log.hpp"
30 #include "runtime/globals.hpp"
31 #include "utilities/debug.hpp"
32 #include "utilities/globalDefinitions.hpp"
33
34 G1HeapSizingPolicy::G1HeapSizingPolicy(const G1CollectedHeap* g1, const G1Analytics* analytics) :
35 _g1(g1),
36 _analytics(analytics),
37 _num_prev_pauses_for_heuristics(analytics->number_of_recorded_pause_times()) {
38 assert(MinOverThresholdForGrowth < _num_prev_pauses_for_heuristics, "Threshold must be less than %u", _num_prev_pauses_for_heuristics);
39 clear_ratio_check_data();
40 }
41
42 void G1HeapSizingPolicy::clear_ratio_check_data() {
43 _ratio_over_threshold_count = 0;
44 _ratio_over_threshold_sum = 0.0;
45 _pauses_since_start = 0;
46 }
47
48 size_t G1HeapSizingPolicy::expansion_amount() {
49 double recent_gc_overhead = _analytics->recent_avg_pause_time_ratio() * 100.0;
50 double last_gc_overhead = _analytics->last_pause_time_ratio() * 100.0;
51 assert(GCTimeRatio > 0,
52 "we should have set it to a default value set_g1_gc_flags() "
53 "if a user set it to 0");
54 const double gc_overhead_percent = 100.0 * (1.0 / (1.0 + GCTimeRatio));
55
56 double threshold = gc_overhead_percent;
57 size_t expand_bytes = 0;
58
59 // If the heap is at less than half its maximum size, scale the threshold down,
60 // to a limit of 1. Thus the smaller the heap is, the more likely it is to expand,
61 // though the scaling code will likely keep the increase small.
62 if (_g1->capacity() <= _g1->max_capacity() / 2) {
63 threshold *= (double)_g1->capacity() / (double)(_g1->max_capacity() / 2);
64 threshold = MAX2(threshold, 1.0);
65 }
66
67 // If the last GC time ratio is over the threshold, increment the count of
68 // times it has been exceeded, and add this ratio to the sum of exceeded
69 // ratios.
70 if (last_gc_overhead > threshold) {
71 _ratio_over_threshold_count++;
72 _ratio_over_threshold_sum += last_gc_overhead;
73 }
74
75 // Check if we've had enough GC time ratio checks that were over the
76 // threshold to trigger an expansion. We'll also expand if we've
77 // reached the end of the history buffer and the average of all entries
78 // is still over the threshold. This indicates a smaller number of GCs were
79 // long enough to make the average exceed the threshold.
80 bool filled_history_buffer = _pauses_since_start == _num_prev_pauses_for_heuristics;
81 if ((_ratio_over_threshold_count == MinOverThresholdForGrowth) ||
82 (filled_history_buffer && (recent_gc_overhead > threshold))) {
83 size_t min_expand_bytes = HeapRegion::GrainBytes;
84 size_t reserved_bytes = _g1->max_capacity();
85 size_t committed_bytes = _g1->capacity();
86 size_t uncommitted_bytes = reserved_bytes - committed_bytes;
87 size_t expand_bytes_via_pct =
88 uncommitted_bytes * G1ExpandByPercentOfAvailable / 100;
89 double scale_factor = 1.0;
90
91 // If the current size is less than 1/4 of the Initial heap size, expand
92 // by half of the delta between the current and Initial sizes. IE, grow
93 // back quickly.
94 //
95 // Otherwise, take the current size, or G1ExpandByPercentOfAvailable % of
96 // the available expansion space, whichever is smaller, as the base
97 // expansion size. Then possibly scale this size according to how much the
98 // threshold has (on average) been exceeded by. If the delta is small
99 // (less than the StartScaleDownAt value), scale the size down linearly, but
100 // not by less than MinScaleDownFactor. If the delta is large (greater than
101 // the StartScaleUpAt value), scale up, but adding no more than MaxScaleUpFactor
102 // times the base size. The scaling will be linear in the range from
103 // StartScaleUpAt to (StartScaleUpAt + ScaleUpRange). In other words,
104 // ScaleUpRange sets the rate of scaling up.
105 if (committed_bytes < InitialHeapSize / 4) {
|
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #include "precompiled.hpp"
26 #include "gc/g1/g1CollectedHeap.hpp"
27 #include "gc/g1/g1HeapSizingPolicy.hpp"
28 #include "gc/g1/g1Analytics.hpp"
29 #include "logging/log.hpp"
30 #include "runtime/globals.hpp"
31 #include "utilities/debug.hpp"
32 #include "utilities/globalDefinitions.hpp"
33
34 G1HeapSizingPolicy::G1HeapSizingPolicy(const G1CollectedHeap* g1h, const G1Analytics* analytics) :
35 _g1h(g1h),
36 _analytics(analytics),
37 _num_prev_pauses_for_heuristics(analytics->number_of_recorded_pause_times()) {
38
39 assert(MinOverThresholdForGrowth < _num_prev_pauses_for_heuristics, "Threshold must be less than %u", _num_prev_pauses_for_heuristics);
40 clear_ratio_check_data();
41 }
42
43 void G1HeapSizingPolicy::clear_ratio_check_data() {
44 _ratio_over_threshold_count = 0;
45 _ratio_over_threshold_sum = 0.0;
46 _pauses_since_start = 0;
47 }
48
49 size_t G1HeapSizingPolicy::expansion_amount() {
50 double recent_gc_overhead = _analytics->recent_avg_pause_time_ratio() * 100.0;
51 double last_gc_overhead = _analytics->last_pause_time_ratio() * 100.0;
52 assert(GCTimeRatio > 0,
53 "we should have set it to a default value set_g1_gc_flags() "
54 "if a user set it to 0");
55 const double gc_overhead_percent = 100.0 * (1.0 / (1.0 + GCTimeRatio));
56
57 double threshold = gc_overhead_percent;
58 size_t expand_bytes = 0;
59
60 // If the heap is at less than half its maximum size, scale the threshold down,
61 // to a limit of 1. Thus the smaller the heap is, the more likely it is to expand,
62 // though the scaling code will likely keep the increase small.
63 if (_g1h->capacity() <= _g1h->max_capacity() / 2) {
64 threshold *= (double)_g1h->capacity() / (double)(_g1h->max_capacity() / 2);
65 threshold = MAX2(threshold, 1.0);
66 }
67
68 // If the last GC time ratio is over the threshold, increment the count of
69 // times it has been exceeded, and add this ratio to the sum of exceeded
70 // ratios.
71 if (last_gc_overhead > threshold) {
72 _ratio_over_threshold_count++;
73 _ratio_over_threshold_sum += last_gc_overhead;
74 }
75
76 // Check if we've had enough GC time ratio checks that were over the
77 // threshold to trigger an expansion. We'll also expand if we've
78 // reached the end of the history buffer and the average of all entries
79 // is still over the threshold. This indicates a smaller number of GCs were
80 // long enough to make the average exceed the threshold.
81 bool filled_history_buffer = _pauses_since_start == _num_prev_pauses_for_heuristics;
82 if ((_ratio_over_threshold_count == MinOverThresholdForGrowth) ||
83 (filled_history_buffer && (recent_gc_overhead > threshold))) {
84 size_t min_expand_bytes = HeapRegion::GrainBytes;
85 size_t reserved_bytes = _g1h->max_capacity();
86 size_t committed_bytes = _g1h->capacity();
87 size_t uncommitted_bytes = reserved_bytes - committed_bytes;
88 size_t expand_bytes_via_pct =
89 uncommitted_bytes * G1ExpandByPercentOfAvailable / 100;
90 double scale_factor = 1.0;
91
92 // If the current size is less than 1/4 of the Initial heap size, expand
93 // by half of the delta between the current and Initial sizes. IE, grow
94 // back quickly.
95 //
96 // Otherwise, take the current size, or G1ExpandByPercentOfAvailable % of
97 // the available expansion space, whichever is smaller, as the base
98 // expansion size. Then possibly scale this size according to how much the
99 // threshold has (on average) been exceeded by. If the delta is small
100 // (less than the StartScaleDownAt value), scale the size down linearly, but
101 // not by less than MinScaleDownFactor. If the delta is large (greater than
102 // the StartScaleUpAt value), scale up, but adding no more than MaxScaleUpFactor
103 // times the base size. The scaling will be linear in the range from
104 // StartScaleUpAt to (StartScaleUpAt + ScaleUpRange). In other words,
105 // ScaleUpRange sets the rate of scaling up.
106 if (committed_bytes < InitialHeapSize / 4) {
|