153 for (uint i = start_idx; i < start_idx + num_regions; i++) {
154 assert(_commit_map.at(i), "Trying to uncommit storage at region %u that is not committed", i);
155 size_t idx = region_idx_to_page_idx(i);
156 uint old_refcount = _refcounts.get_by_index(idx);
157 assert(old_refcount > 0, "must be");
158 if (old_refcount == 1) {
159 _storage.uncommit(idx, 1);
160 }
161 _refcounts.set_by_index(idx, old_refcount - 1);
162 _commit_map.clear_bit(i);
163 }
164 }
165 };
166
167 void G1RegionToSpaceMapper::fire_on_commit(uint start_idx, size_t num_regions, bool zero_filled) {
168 if (_listener != NULL) {
169 _listener->on_commit(start_idx, num_regions, zero_filled);
170 }
171 }
172
173 G1RegionToSpaceMapper* G1RegionToSpaceMapper::create_mapper(ReservedSpace rs,
174 size_t actual_size,
175 size_t page_size,
176 size_t region_granularity,
177 size_t commit_factor,
178 MemoryType type) {
179
180 if (region_granularity >= (page_size * commit_factor)) {
181 return new G1RegionsLargerThanCommitSizeMapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
182 } else {
183 return new G1RegionsSmallerThanCommitSizeMapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
184 }
185 }
|
153 for (uint i = start_idx; i < start_idx + num_regions; i++) {
154 assert(_commit_map.at(i), "Trying to uncommit storage at region %u that is not committed", i);
155 size_t idx = region_idx_to_page_idx(i);
156 uint old_refcount = _refcounts.get_by_index(idx);
157 assert(old_refcount > 0, "must be");
158 if (old_refcount == 1) {
159 _storage.uncommit(idx, 1);
160 }
161 _refcounts.set_by_index(idx, old_refcount - 1);
162 _commit_map.clear_bit(i);
163 }
164 }
165 };
166
167 void G1RegionToSpaceMapper::fire_on_commit(uint start_idx, size_t num_regions, bool zero_filled) {
168 if (_listener != NULL) {
169 _listener->on_commit(start_idx, num_regions, zero_filled);
170 }
171 }
172
173 void G1RegionToHeteroSpaceMapper::map_nvdimm_space(ReservedSpace rs) {
174 assert(AllocateOldGenAt != NULL, "");
175 int _backing_fd = os::create_file_for_heap(AllocateOldGenAt);
176 if (_backing_fd == -1) {
177 vm_exit_during_initialization(
178 err_msg("Could not create file for Old generation at location %s", AllocateOldGenAt));
179 }
180 // commit this memory in nv-dimm
181 char* ret = os::attempt_reserve_memory_at(rs.size(), rs.base(), _backing_fd);
182 //char* ret = os::replace_existing_mapping_with_file_mapping(rs.base(), rs.size(), _backing_fd);
183 if (ret != rs.base()) {
184 if (ret != NULL) {
185 os::unmap_memory(rs.base(), rs.size());
186 }
187 vm_exit_during_initialization(
188 err_msg("Error in mapping Old Gen to given AllocateOldGenAt = %s", AllocateOldGenAt));
189 }
190 }
191
192 G1RegionToHeteroSpaceMapper::G1RegionToHeteroSpaceMapper(ReservedSpace rs,
193 size_t actual_size,
194 size_t page_size,
195 size_t alloc_granularity,
196 size_t commit_factor,
197 MemoryType type) :
198 G1RegionToSpaceMapper(rs, actual_size, page_size, alloc_granularity, commit_factor, type),
199 _num_committed_dram(0), _num_committed_nvdimm(0) {
200 assert(actual_size == 2 * MaxHeapSize, "For 2-way heterogenuous heap, reserved space is two times MaxHeapSize");
201
202 // Since we need to split the reserved space in half and map second half to file in NV-DIMM, we need to release the reserved memory first
203 // Because on some OSes (e.g. Windows) you cannot do a file mapping on memory reserved with regular mapping.
204 os::release_memory(rs.base(), rs.size());
205
206 /* Toggle HeteroHeap
207 // We map first part of size Xmx to DRAM.
208 ReservedSpace rs_dram = rs.first_part(MaxHeapSize);
209 // Second half of reserved memory is mapped to NV-DIMM.
210 ReservedSpace rs_nvdimm = rs.last_part(MaxHeapSize);*/
211 // We map first part of size Xmx to NVDIMM.
212 ReservedSpace rs_nvdimm = rs.first_part(MaxHeapSize);
213 // Second half of reserved memory is mapped to DRAM.
214 ReservedSpace rs_dram = rs.last_part(MaxHeapSize);
215 assert(rs_dram.size() == rs_nvdimm.size() && rs_nvdimm.size() == MaxHeapSize, "They all should be same");
216
217 // Reserve dram memory
218 char* base = os::attempt_reserve_memory_at(rs_dram.size(), rs_dram.base());
219 if (base != rs_dram.base()) {
220 if (base != NULL) {
221 os::release_memory(base, rs_dram.size());
222 }
223 vm_exit_during_initialization(err_msg("Error in allocating heap"));
224 }
225
226 // We reserve and commit this entire space to NV-DIMM.
227 map_nvdimm_space(rs_nvdimm);
228
229 if (alloc_granularity >= (page_size * commit_factor)) {
230 _dram_mapper = new G1RegionsLargerThanCommitSizeMapper(rs_dram, rs_dram.size(), page_size, alloc_granularity, commit_factor, type);
231 }
232 else {
233 _dram_mapper = new G1RegionsSmallerThanCommitSizeMapper(rs_dram, rs_dram.size(), page_size, alloc_granularity, commit_factor, type);
234 }
235
236 /* Toggle HeteroHeap
237 _start_index_of_nvdimm = (uint)(rs_dram.size() / alloc_granularity);
238 _start_index_of_dram = 0; */
239 _start_index_of_nvdimm = 0;
240 _start_index_of_dram = (uint)(rs_nvdimm.size() / alloc_granularity);
241 }
242
243 void G1RegionToHeteroSpaceMapper::commit_regions(uint start_idx, size_t num_regions, WorkGang* pretouch_gang) {
244 uint end_idx = (start_idx + (uint)num_regions - 1);
245
246 /* Toggle HeteroHeap
247 uint num_nvdimm = end_idx >= _start_index_of_nvdimm ? MIN2((end_idx - _start_index_of_nvdimm + 1), (uint)num_regions) : 0;
248 uint num_dram = (uint)num_regions - num_nvdimm;*/
249 uint num_dram = end_idx >= _start_index_of_dram ? MIN2((end_idx - _start_index_of_dram + 1), (uint)num_regions) : 0;
250 uint num_nvdimm = (uint)num_regions - num_dram;
251
252 if (num_nvdimm > 0) {
253 // We do not need to commit nv-dimm regions, since they are committed in the beginning.
254 _num_committed_nvdimm += num_nvdimm;
255 }
256 if (num_dram > 0) {
257 /* Toggle HeteroHeap
258 _dram_mapper->commit_regions(start_idx, num_dram, pretouch_gang); */
259 _dram_mapper->commit_regions(start_idx > _start_index_of_dram ? (start_idx - _start_index_of_dram) : 0, num_dram, pretouch_gang);
260 _num_committed_dram += num_dram;
261 }
262 }
263
264 void G1RegionToHeteroSpaceMapper::uncommit_regions(uint start_idx, size_t num_regions) {
265 uint end_idx = (start_idx + (uint)num_regions - 1);
266 /* Toggle HeterHeap
267 uint num_nvdimm = end_idx >= _start_index_of_nvdimm ? MIN2((end_idx - _start_index_of_nvdimm + 1), (uint)num_regions) : 0;
268 uint num_dram = (uint)num_regions - num_nvdimm;*/
269 uint num_dram = end_idx >= _start_index_of_dram ? MIN2((end_idx - _start_index_of_dram + 1), (uint)num_regions) : 0;
270 uint num_nvdimm = (uint)num_regions - num_dram;
271
272 if (num_nvdimm > 0) {
273 // We do not uncommit memory for nv-dimm regions.
274 _num_committed_nvdimm -= num_nvdimm;
275 }
276
277 if (num_dram > 0) {
278 _dram_mapper->uncommit_regions(start_idx > _start_index_of_dram ? (start_idx - _start_index_of_dram) : 0, num_dram);
279 _num_committed_dram -= num_dram;
280 }
281 }
282
283 uint G1RegionToHeteroSpaceMapper::num_committed_dram() {
284 return _num_committed_dram;
285 }
286
287 uint G1RegionToHeteroSpaceMapper::num_committed_nvdimm() {
288 return _num_committed_nvdimm;
289 }
290
291 G1RegionToSpaceMapper* G1RegionToSpaceMapper::create_heap_mapper(ReservedSpace rs,
292 size_t actual_size,
293 size_t page_size,
294 size_t region_granularity,
295 size_t commit_factor,
296 MemoryType type) {
297 if (AllocateOldGenAt != NULL) {
298 return new G1RegionToHeteroSpaceMapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
299 } else {
300 return create_mapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
301 }
302 }
303
304 G1RegionToSpaceMapper* G1RegionToSpaceMapper::create_mapper(ReservedSpace rs,
305 size_t actual_size,
306 size_t page_size,
307 size_t region_granularity,
308 size_t commit_factor,
309 MemoryType type) {
310 if (region_granularity >= (page_size * commit_factor)) {
311 return new G1RegionsLargerThanCommitSizeMapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
312 } else {
313 return new G1RegionsSmallerThanCommitSizeMapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
314 }
315 }
|