< prev index next >
src/hotspot/share/gc/g1/g1RegionToSpaceMapper.cpp
Print this page
rev 51649 : version 1
rev 51652 : Added support for eager mixed collection of evacuation failure regions
rev 51878 : Minor changes
rev 52017 : All changes for G1 GC moved from 'combined' repo folder
@@ -168,17 +168,147 @@
if (_listener != NULL) {
_listener->on_commit(start_idx, num_regions, zero_filled);
}
}
-G1RegionToSpaceMapper* G1RegionToSpaceMapper::create_mapper(ReservedSpace rs,
+void G1RegionToHeteroSpaceMapper::map_nvdimm_space(ReservedSpace rs) {
+ assert(AllocateOldGenAt != NULL, "");
+ int _backing_fd = os::create_file_for_heap(AllocateOldGenAt);
+ if (_backing_fd == -1) {
+ vm_exit_during_initialization(
+ err_msg("Could not create file for Old generation at location %s", AllocateOldGenAt));
+ }
+ // commit this memory in nv-dimm
+ char* ret = os::attempt_reserve_memory_at(rs.size(), rs.base(), _backing_fd);
+ //char* ret = os::replace_existing_mapping_with_file_mapping(rs.base(), rs.size(), _backing_fd);
+ if (ret != rs.base()) {
+ if (ret != NULL) {
+ os::unmap_memory(rs.base(), rs.size());
+ }
+ vm_exit_during_initialization(
+ err_msg("Error in mapping Old Gen to given AllocateOldGenAt = %s", AllocateOldGenAt));
+ }
+}
+
+G1RegionToHeteroSpaceMapper::G1RegionToHeteroSpaceMapper(ReservedSpace rs,
+ size_t actual_size,
+ size_t page_size,
+ size_t alloc_granularity,
+ size_t commit_factor,
+ MemoryType type) :
+ G1RegionToSpaceMapper(rs, actual_size, page_size, alloc_granularity, commit_factor, type),
+ _num_committed_dram(0), _num_committed_nvdimm(0) {
+ assert(actual_size == 2 * MaxHeapSize, "For 2-way heterogenuous heap, reserved space is two times MaxHeapSize");
+
+ // Since we need to split the reserved space in half and map second half to file in NV-DIMM, we need to release the reserved memory first
+ // Because on some OSes (e.g. Windows) you cannot do a file mapping on memory reserved with regular mapping.
+ os::release_memory(rs.base(), rs.size());
+
+ /* Toggle HeteroHeap
+ // We map first part of size Xmx to DRAM.
+ ReservedSpace rs_dram = rs.first_part(MaxHeapSize);
+ // Second half of reserved memory is mapped to NV-DIMM.
+ ReservedSpace rs_nvdimm = rs.last_part(MaxHeapSize);*/
+ // We map first part of size Xmx to NVDIMM.
+ ReservedSpace rs_nvdimm = rs.first_part(MaxHeapSize);
+ // Second half of reserved memory is mapped to DRAM.
+ ReservedSpace rs_dram = rs.last_part(MaxHeapSize);
+ assert(rs_dram.size() == rs_nvdimm.size() && rs_nvdimm.size() == MaxHeapSize, "They all should be same");
+
+ // Reserve dram memory
+ char* base = os::attempt_reserve_memory_at(rs_dram.size(), rs_dram.base());
+ if (base != rs_dram.base()) {
+ if (base != NULL) {
+ os::release_memory(base, rs_dram.size());
+ }
+ vm_exit_during_initialization(err_msg("Error in allocating heap"));
+ }
+
+ // We reserve and commit this entire space to NV-DIMM.
+ map_nvdimm_space(rs_nvdimm);
+
+ if (alloc_granularity >= (page_size * commit_factor)) {
+ _dram_mapper = new G1RegionsLargerThanCommitSizeMapper(rs_dram, rs_dram.size(), page_size, alloc_granularity, commit_factor, type);
+ }
+ else {
+ _dram_mapper = new G1RegionsSmallerThanCommitSizeMapper(rs_dram, rs_dram.size(), page_size, alloc_granularity, commit_factor, type);
+ }
+
+/* Toggle HeteroHeap
+ _start_index_of_nvdimm = (uint)(rs_dram.size() / alloc_granularity);
+ _start_index_of_dram = 0; */
+ _start_index_of_nvdimm = 0;
+ _start_index_of_dram = (uint)(rs_nvdimm.size() / alloc_granularity);
+}
+
+void G1RegionToHeteroSpaceMapper::commit_regions(uint start_idx, size_t num_regions, WorkGang* pretouch_gang) {
+ uint end_idx = (start_idx + (uint)num_regions - 1);
+
+ /* Toggle HeteroHeap
+ uint num_nvdimm = end_idx >= _start_index_of_nvdimm ? MIN2((end_idx - _start_index_of_nvdimm + 1), (uint)num_regions) : 0;
+ uint num_dram = (uint)num_regions - num_nvdimm;*/
+ uint num_dram = end_idx >= _start_index_of_dram ? MIN2((end_idx - _start_index_of_dram + 1), (uint)num_regions) : 0;
+ uint num_nvdimm = (uint)num_regions - num_dram;
+
+ if (num_nvdimm > 0) {
+ // We do not need to commit nv-dimm regions, since they are committed in the beginning.
+ _num_committed_nvdimm += num_nvdimm;
+ }
+ if (num_dram > 0) {
+ /* Toggle HeteroHeap
+ _dram_mapper->commit_regions(start_idx, num_dram, pretouch_gang); */
+ _dram_mapper->commit_regions(start_idx > _start_index_of_dram ? (start_idx - _start_index_of_dram) : 0, num_dram, pretouch_gang);
+ _num_committed_dram += num_dram;
+ }
+}
+
+void G1RegionToHeteroSpaceMapper::uncommit_regions(uint start_idx, size_t num_regions) {
+ uint end_idx = (start_idx + (uint)num_regions - 1);
+ /* Toggle HeterHeap
+ uint num_nvdimm = end_idx >= _start_index_of_nvdimm ? MIN2((end_idx - _start_index_of_nvdimm + 1), (uint)num_regions) : 0;
+ uint num_dram = (uint)num_regions - num_nvdimm;*/
+ uint num_dram = end_idx >= _start_index_of_dram ? MIN2((end_idx - _start_index_of_dram + 1), (uint)num_regions) : 0;
+ uint num_nvdimm = (uint)num_regions - num_dram;
+
+ if (num_nvdimm > 0) {
+ // We do not uncommit memory for nv-dimm regions.
+ _num_committed_nvdimm -= num_nvdimm;
+ }
+
+ if (num_dram > 0) {
+ _dram_mapper->uncommit_regions(start_idx > _start_index_of_dram ? (start_idx - _start_index_of_dram) : 0, num_dram);
+ _num_committed_dram -= num_dram;
+ }
+}
+
+uint G1RegionToHeteroSpaceMapper::num_committed_dram() {
+ return _num_committed_dram;
+}
+
+uint G1RegionToHeteroSpaceMapper::num_committed_nvdimm() {
+ return _num_committed_nvdimm;
+}
+
+G1RegionToSpaceMapper* G1RegionToSpaceMapper::create_heap_mapper(ReservedSpace rs,
size_t actual_size,
size_t page_size,
size_t region_granularity,
size_t commit_factor,
MemoryType type) {
+ if (AllocateOldGenAt != NULL) {
+ return new G1RegionToHeteroSpaceMapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
+ } else {
+ return create_mapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
+ }
+}
+G1RegionToSpaceMapper* G1RegionToSpaceMapper::create_mapper(ReservedSpace rs,
+ size_t actual_size,
+ size_t page_size,
+ size_t region_granularity,
+ size_t commit_factor,
+ MemoryType type) {
if (region_granularity >= (page_size * commit_factor)) {
return new G1RegionsLargerThanCommitSizeMapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
} else {
return new G1RegionsSmallerThanCommitSizeMapper(rs, actual_size, page_size, region_granularity, commit_factor, type);
}
< prev index next >