Print this page
Split |
Close |
Expand all |
Collapse all |
--- old/src/share/vm/gc_implementation/g1/concurrentG1Refine.hpp
+++ new/src/share/vm/gc_implementation/g1/concurrentG1Refine.hpp
1 1 /*
2 - * Copyright (c) 2001, 2010, Oracle and/or its affiliates. All rights reserved.
2 + * Copyright (c) 2001, 2011, Oracle and/or its affiliates. All rights reserved.
3 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 4 *
5 5 * This code is free software; you can redistribute it and/or modify it
6 6 * under the terms of the GNU General Public License version 2 only, as
7 7 * published by the Free Software Foundation.
8 8 *
9 9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 12 * version 2 for more details (a copy is included in the LICENSE file that
13 13 * accompanied this code).
14 14 *
15 15 * You should have received a copy of the GNU General Public License version
16 16 * 2 along with this work; if not, write to the Free Software Foundation,
17 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 18 *
19 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 20 * or visit www.oracle.com if you need additional information or have any
21 21 * questions.
22 22 *
23 23 */
24 24
25 25 #ifndef SHARE_VM_GC_IMPLEMENTATION_G1_CONCURRENTG1REFINE_HPP
26 26 #define SHARE_VM_GC_IMPLEMENTATION_G1_CONCURRENTG1REFINE_HPP
27 27
28 28 #include "memory/allocation.hpp"
29 29 #include "memory/cardTableModRefBS.hpp"
30 30 #include "runtime/thread.hpp"
31 31 #include "utilities/globalDefinitions.hpp"
32 32
33 33 // Forward decl
34 34 class ConcurrentG1RefineThread;
35 35 class G1RemSet;
36 36
37 37 class ConcurrentG1Refine: public CHeapObj {
38 38 ConcurrentG1RefineThread** _threads;
39 39 int _n_threads;
40 40 int _n_worker_threads;
41 41 /*
42 42 * The value of the update buffer queue length falls into one of 3 zones:
43 43 * green, yellow, red. If the value is in [0, green) nothing is
44 44 * done, the buffers are left unprocessed to enable the caching effect of the
45 45 * dirtied cards. In the yellow zone [green, yellow) the concurrent refinement
46 46 * threads are gradually activated. In [yellow, red) all threads are
47 47 * running. If the length becomes red (max queue length) the mutators start
48 48 * processing the buffers.
49 49 *
50 50 * There are some interesting cases (when G1UseAdaptiveConcRefinement
51 51 * is turned off):
52 52 * 1) green = yellow = red = 0. In this case the mutator will process all
53 53 * buffers. Except for those that are created by the deferred updates
54 54 * machinery during a collection.
55 55 * 2) green = 0. Means no caching. Can be a good way to minimize the
56 56 * amount of time spent updating rsets during a collection.
57 57 */
58 58 int _green_zone;
59 59 int _yellow_zone;
60 60 int _red_zone;
61 61
62 62 int _thread_threshold_step;
63 63
64 64 // Reset the threshold step value based of the current zone boundaries.
65 65 void reset_threshold_step();
66 66
67 67 // The cache for card refinement.
68 68 bool _use_cache;
69 69 bool _def_use_cache;
70 70
71 71 size_t _n_periods; // Used as clearing epoch
72 72
73 73 // An evicting cache of the number of times each card
74 74 // is accessed. Reduces, but does not eliminate, the amount
75 75 // of duplicated processing of dirty cards.
76 76
77 77 enum SomePrivateConstants {
78 78 epoch_bits = 32,
79 79 card_num_shift = epoch_bits,
80 80 epoch_mask = AllBits,
81 81 card_num_mask = AllBits,
82 82
83 83 // The initial cache size is approximately this fraction
84 84 // of a maximal cache (i.e. the size needed for all cards
85 85 // in the heap)
86 86 InitialCacheFraction = 512
↓ open down ↓ |
74 lines elided |
↑ open up ↑ |
87 87 };
88 88
89 89 const static julong card_num_mask_in_place =
90 90 (julong) card_num_mask << card_num_shift;
91 91
92 92 typedef struct {
93 93 julong _value; // | card_num | epoch |
94 94 } CardEpochCacheEntry;
95 95
96 96 julong make_epoch_entry(unsigned int card_num, unsigned int epoch) {
97 - assert(0 <= card_num && card_num < _max_n_card_counts, "Bounds");
97 + assert(0 <= card_num && card_num < _max_cards, "Bounds");
98 98 assert(0 <= epoch && epoch <= _n_periods, "must be");
99 99
100 100 return ((julong) card_num << card_num_shift) | epoch;
101 101 }
102 102
103 103 unsigned int extract_epoch(julong v) {
104 104 return (v & epoch_mask);
105 105 }
106 106
107 107 unsigned int extract_card_num(julong v) {
108 108 return (v & card_num_mask_in_place) >> card_num_shift;
109 109 }
110 110
111 111 typedef struct {
↓ open down ↓ |
4 lines elided |
↑ open up ↑ |
112 112 unsigned char _count;
113 113 unsigned char _evict_count;
114 114 } CardCountCacheEntry;
115 115
116 116 CardCountCacheEntry* _card_counts;
117 117 CardEpochCacheEntry* _card_epochs;
118 118
119 119 // The current number of buckets in the card count cache
120 120 unsigned _n_card_counts;
121 121
122 - // The max number of buckets required for the number of
123 - // cards for the entire reserved heap
122 + // The number of cards for the entire reserved heap
123 + unsigned _max_cards;
124 +
125 + // The max number of buckets for the card counts and epochs caches.
126 + // This is the maximum that the counts and epochs will grow to.
127 + // It is specified as a fraction or percentage of _max_cards using
128 + // G1MaxHotCardCountSizePercent (currently 50%).
124 129 unsigned _max_n_card_counts;
125 130
126 131 // Possible sizes of the cache: odd primes that roughly double in size.
127 132 // (See jvmtiTagMap.cpp).
128 - static int _cc_cache_sizes[];
133 + enum {
134 + MAX_CC_CACHE_INDEX = 15 // maximum index into the cache size array.
135 + };
136 +
137 + static int _cc_cache_sizes[MAX_CC_CACHE_INDEX];
129 138
130 139 // The index in _cc_cache_sizes corresponding to the size of
131 140 // _card_counts.
132 141 int _cache_size_index;
133 142
134 143 bool _expand_card_counts;
135 144
136 145 const jbyte* _ct_bot;
137 146
138 147 jbyte** _hot_cache;
139 148 int _hot_cache_size;
↓ open down ↓ |
1 lines elided |
↑ open up ↑ |
140 149 int _n_hot;
141 150 int _hot_cache_idx;
142 151
143 152 int _hot_cache_par_chunk_size;
144 153 volatile int _hot_cache_par_claimed_idx;
145 154
146 155 // Needed to workaround 6817995
147 156 CardTableModRefBS* _ct_bs;
148 157 G1CollectedHeap* _g1h;
149 158
150 - // Expands the array that holds the card counts to the next size up
151 - void expand_card_count_cache();
159 + // Helper routine for expand_card_count_cache().
160 + // The arrays used to hold the card counts and the epochs must have
161 + // a 1:1 correspondence. Hence they are allocated and freed together.
162 + // Returns true if the allocations of both the counts and epochs
163 + // were successful; false otherwise.
164 + bool allocate_card_count_cache(int n, CardCountCacheEntry** counts,
165 + CardEpochCacheEntry** epochs);
166 +
167 + // Expands the arrays that hold the card counts and epochs
168 + // to the cache size at index. Returns true if the expansion/
169 + // allocation was successful; false otherwise.
170 + bool expand_card_count_cache(int index);
152 171
153 172 // hash a given key (index of card_ptr) with the specified size
154 173 static unsigned int hash(size_t key, int size) {
155 174 return (unsigned int) key % size;
156 175 }
157 176
158 177 // hash a given key (index of card_ptr)
159 178 unsigned int hash(size_t key) {
160 179 return hash(key, _n_card_counts);
161 180 }
162 181
163 182 unsigned ptr_2_card_num(jbyte* card_ptr) {
164 183 return (unsigned) (card_ptr - _ct_bot);
165 184 }
166 185
167 186 jbyte* card_num_2_ptr(unsigned card_num) {
168 187 return (jbyte*) (_ct_bot + card_num);
169 188 }
170 189
171 190 // Returns the count of this card after incrementing it.
172 191 jbyte* add_card_count(jbyte* card_ptr, int* count, bool* defer);
173 192
174 193 // Returns true if this card is in a young region
175 194 bool is_young_card(jbyte* card_ptr);
176 195
177 196 public:
178 197 ConcurrentG1Refine();
179 198 ~ConcurrentG1Refine();
180 199
181 200 void init(); // Accomplish some initialization that has to wait.
182 201 void stop();
183 202
184 203 void reinitialize_threads();
185 204
186 205 // Iterate over the conc refine threads
187 206 void threads_do(ThreadClosure *tc);
188 207
189 208 // If this is the first entry for the slot, writes into the cache and
190 209 // returns NULL. If it causes an eviction, returns the evicted pointer.
191 210 // Otherwise, its a cache hit, and returns NULL.
192 211 jbyte* cache_insert(jbyte* card_ptr, bool* defer);
193 212
194 213 // Process the cached entries.
195 214 void clean_up_cache(int worker_i, G1RemSet* g1rs, DirtyCardQueue* into_cset_dcq);
196 215
197 216 // Set up for parallel processing of the cards in the hot cache
198 217 void clear_hot_cache_claimed_index() {
199 218 _hot_cache_par_claimed_idx = 0;
200 219 }
201 220
202 221 // Discard entries in the hot cache.
203 222 void clear_hot_cache() {
204 223 _hot_cache_idx = 0; _n_hot = 0;
205 224 }
206 225
207 226 bool hot_cache_is_empty() { return _n_hot == 0; }
208 227
209 228 bool use_cache() { return _use_cache; }
210 229 void set_use_cache(bool b) {
211 230 if (b) _use_cache = _def_use_cache;
212 231 else _use_cache = false;
213 232 }
214 233
215 234 void clear_and_record_card_counts();
216 235
217 236 static int thread_num();
218 237
219 238 void print_worker_threads_on(outputStream* st) const;
220 239
221 240 void set_green_zone(int x) { _green_zone = x; }
222 241 void set_yellow_zone(int x) { _yellow_zone = x; }
223 242 void set_red_zone(int x) { _red_zone = x; }
224 243
225 244 int green_zone() const { return _green_zone; }
226 245 int yellow_zone() const { return _yellow_zone; }
227 246 int red_zone() const { return _red_zone; }
228 247
229 248 int total_thread_num() const { return _n_threads; }
230 249 int worker_thread_num() const { return _n_worker_threads; }
231 250
232 251 int thread_threshold_step() const { return _thread_threshold_step; }
233 252 };
234 253
235 254 #endif // SHARE_VM_GC_IMPLEMENTATION_G1_CONCURRENTG1REFINE_HPP
↓ open down ↓ |
74 lines elided |
↑ open up ↑ |
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX