src/share/vm/gc_implementation/parallelScavenge/objectStartArray.cpp
Index Unified diffs Context diffs Sdiffs Wdiffs Patch New Old Previous File Next File hotspot Sdiff src/share/vm/gc_implementation/parallelScavenge

src/share/vm/gc_implementation/parallelScavenge/objectStartArray.cpp

Print this page


   1 /*
   2  * Copyright (c) 2001, 2012, Oracle and/or its affiliates. All rights reserved.
   3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4  *
   5  * This code is free software; you can redistribute it and/or modify it
   6  * under the terms of the GNU General Public License version 2 only, as
   7  * published by the Free Software Foundation.
   8  *
   9  * This code is distributed in the hope that it will be useful, but WITHOUT
  10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12  * version 2 for more details (a copy is included in the LICENSE file that
  13  * accompanied this code).
  14  *
  15  * You should have received a copy of the GNU General Public License version
  16  * 2 along with this work; if not, write to the Free Software Foundation,
  17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18  *
  19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  20  * or visit www.oracle.com if you need additional information or have any
  21  * questions.
  22  *


  34   // We're based on the assumption that we use the same
  35   // size blocks as the card table.
  36   assert((int)block_size == (int)CardTableModRefBS::card_size, "Sanity");
  37   assert((int)block_size <= 512, "block_size must be less than or equal to 512");
  38 
  39   // Calculate how much space must be reserved
  40   _reserved_region = reserved_region;
  41 
  42   size_t bytes_to_reserve = reserved_region.word_size() / block_size_in_words;
  43   assert(bytes_to_reserve > 0, "Sanity");
  44 
  45   bytes_to_reserve =
  46     align_size_up(bytes_to_reserve, os::vm_allocation_granularity());
  47 
  48   // Do not use large-pages for the backing store. The one large page region
  49   // will be used for the heap proper.
  50   ReservedSpace backing_store(bytes_to_reserve);
  51   if (!backing_store.is_reserved()) {
  52     vm_exit_during_initialization("Could not reserve space for ObjectStartArray");
  53   }
  54   MemTracker::record_virtual_memory_type((address)backing_store.base(), mtGC);


  55 
  56   // We do not commit any memory initially
  57   if (!_virtual_space.initialize(backing_store, 0)) {
  58     vm_exit_during_initialization("Could not commit space for ObjectStartArray");
  59   }
  60 
  61   _raw_base = (jbyte*)_virtual_space.low_boundary();
  62 
  63   if (_raw_base == NULL) {
  64     vm_exit_during_initialization("Could not get raw_base address");
  65   }
  66 
  67   MemTracker::record_virtual_memory_type((address)_raw_base, mtGC);
  68 
  69 
  70   _offset_base = _raw_base - (size_t(reserved_region.start()) >> block_shift);
  71 
  72   _covered_region.set_start(reserved_region.start());
  73   _covered_region.set_word_size(0);
  74 
  75   _blocks_region.set_start((HeapWord*)_raw_base);
  76   _blocks_region.set_word_size(0);
  77 }
  78 
  79 void ObjectStartArray::set_covered_region(MemRegion mr) {
  80   assert(_reserved_region.contains(mr), "MemRegion outside of reserved space");
  81   assert(_reserved_region.start() == mr.start(), "Attempt to move covered region");
  82 
  83   HeapWord* low_bound  = mr.start();
  84   HeapWord* high_bound = mr.end();
  85   assert((uintptr_t(low_bound)  & (block_size - 1))  == 0, "heap must start at block boundary");
  86   assert((uintptr_t(high_bound) & (block_size - 1))  == 0, "heap must end at block boundary");
  87 
  88   size_t requested_blocks_size_in_bytes = mr.word_size() / block_size_in_words;


   1 /*
   2  * Copyright (c) 2001, 2013, Oracle and/or its affiliates. All rights reserved.
   3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4  *
   5  * This code is free software; you can redistribute it and/or modify it
   6  * under the terms of the GNU General Public License version 2 only, as
   7  * published by the Free Software Foundation.
   8  *
   9  * This code is distributed in the hope that it will be useful, but WITHOUT
  10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12  * version 2 for more details (a copy is included in the LICENSE file that
  13  * accompanied this code).
  14  *
  15  * You should have received a copy of the GNU General Public License version
  16  * 2 along with this work; if not, write to the Free Software Foundation,
  17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18  *
  19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  20  * or visit www.oracle.com if you need additional information or have any
  21  * questions.
  22  *


  34   // We're based on the assumption that we use the same
  35   // size blocks as the card table.
  36   assert((int)block_size == (int)CardTableModRefBS::card_size, "Sanity");
  37   assert((int)block_size <= 512, "block_size must be less than or equal to 512");
  38 
  39   // Calculate how much space must be reserved
  40   _reserved_region = reserved_region;
  41 
  42   size_t bytes_to_reserve = reserved_region.word_size() / block_size_in_words;
  43   assert(bytes_to_reserve > 0, "Sanity");
  44 
  45   bytes_to_reserve =
  46     align_size_up(bytes_to_reserve, os::vm_allocation_granularity());
  47 
  48   // Do not use large-pages for the backing store. The one large page region
  49   // will be used for the heap proper.
  50   ReservedSpace backing_store(bytes_to_reserve);
  51   if (!backing_store.is_reserved()) {
  52     vm_exit_during_initialization("Could not reserve space for ObjectStartArray");
  53   }
  54 
  55   NMTTrackOp op(NMTTrackOp::TypeOp);
  56   op.execute_op((address)backing_store.base(), 0, mtGC);
  57 
  58   // We do not commit any memory initially
  59   if (!_virtual_space.initialize(backing_store, 0)) {
  60     vm_exit_during_initialization("Could not commit space for ObjectStartArray");
  61   }
  62 
  63   _raw_base = (jbyte*)_virtual_space.low_boundary();
  64 
  65   if (_raw_base == NULL) {
  66     vm_exit_during_initialization("Could not get raw_base address");
  67   }
  68 
  69   op.execute_op((address)_raw_base, 0, mtGC);

  70 
  71   _offset_base = _raw_base - (size_t(reserved_region.start()) >> block_shift);
  72 
  73   _covered_region.set_start(reserved_region.start());
  74   _covered_region.set_word_size(0);
  75 
  76   _blocks_region.set_start((HeapWord*)_raw_base);
  77   _blocks_region.set_word_size(0);
  78 }
  79 
  80 void ObjectStartArray::set_covered_region(MemRegion mr) {
  81   assert(_reserved_region.contains(mr), "MemRegion outside of reserved space");
  82   assert(_reserved_region.start() == mr.start(), "Attempt to move covered region");
  83 
  84   HeapWord* low_bound  = mr.start();
  85   HeapWord* high_bound = mr.end();
  86   assert((uintptr_t(low_bound)  & (block_size - 1))  == 0, "heap must start at block boundary");
  87   assert((uintptr_t(high_bound) & (block_size - 1))  == 0, "heap must end at block boundary");
  88 
  89   size_t requested_blocks_size_in_bytes = mr.word_size() / block_size_in_words;


src/share/vm/gc_implementation/parallelScavenge/objectStartArray.cpp
Index Unified diffs Context diffs Sdiffs Wdiffs Patch New Old Previous File Next File