1 /*
   2  * Copyright (c) 2001, 2005, Oracle and/or its affiliates. All rights reserved.
   3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4  *
   5  * This code is free software; you can redistribute it and/or modify it
   6  * under the terms of the GNU General Public License version 2 only, as
   7  * published by the Free Software Foundation.
   8  *
   9  * This code is distributed in the hope that it will be useful, but WITHOUT
  10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12  * version 2 for more details (a copy is included in the LICENSE file that
  13  * accompanied this code).
  14  *
  15  * You should have received a copy of the GNU General Public License version
  16  * 2 along with this work; if not, write to the Free Software Foundation,
  17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18  *
  19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  20  * or visit www.oracle.com if you need additional information or have any
  21  * questions.
  22  *
  23  */
  24 
  25 # include "incls/_precompiled.incl"
  26 # include "incls/_objectStartArray.cpp.incl"
  27 
  28 void ObjectStartArray::initialize(MemRegion reserved_region) {
  29   // We're based on the assumption that we use the same
  30   // size blocks as the card table.
  31   assert((int)block_size == (int)CardTableModRefBS::card_size, "Sanity");
  32   assert((int)block_size <= 512, "block_size must be less than or equal to 512");
  33 
  34   // Calculate how much space must be reserved
  35   _reserved_region = reserved_region;
  36 
  37   size_t bytes_to_reserve = reserved_region.word_size() / block_size_in_words;
  38   assert(bytes_to_reserve > 0, "Sanity");
  39 
  40   bytes_to_reserve =
  41     align_size_up(bytes_to_reserve, os::vm_allocation_granularity());
  42 
  43   // Do not use large-pages for the backing store. The one large page region
  44   // will be used for the heap proper.
  45   ReservedSpace backing_store(bytes_to_reserve);
  46   if (!backing_store.is_reserved()) {
  47     vm_exit_during_initialization("Could not reserve space for ObjectStartArray");
  48   }
  49 
  50   // We do not commit any memory initially
  51   if (!_virtual_space.initialize(backing_store, 0)) {
  52     vm_exit_during_initialization("Could not commit space for ObjectStartArray");
  53   }
  54 
  55   _raw_base = (jbyte*)_virtual_space.low_boundary();
  56   if (_raw_base == NULL) {
  57     vm_exit_during_initialization("Could not get raw_base address");
  58   }
  59 
  60   _offset_base = _raw_base - (size_t(reserved_region.start()) >> block_shift);
  61 
  62   _covered_region.set_start(reserved_region.start());
  63   _covered_region.set_word_size(0);
  64 
  65   _blocks_region.set_start((HeapWord*)_raw_base);
  66   _blocks_region.set_word_size(0);
  67 }
  68 
  69 void ObjectStartArray::set_covered_region(MemRegion mr) {
  70   assert(_reserved_region.contains(mr), "MemRegion outside of reserved space");
  71   assert(_reserved_region.start() == mr.start(), "Attempt to move covered region");
  72 
  73   HeapWord* low_bound  = mr.start();
  74   HeapWord* high_bound = mr.end();
  75   assert((uintptr_t(low_bound)  & (block_size - 1))  == 0, "heap must start at block boundary");
  76   assert((uintptr_t(high_bound) & (block_size - 1))  == 0, "heap must end at block boundary");
  77 
  78   size_t requested_blocks_size_in_bytes = mr.word_size() / block_size_in_words;
  79 
  80   // Only commit memory in page sized chunks
  81   requested_blocks_size_in_bytes =
  82     align_size_up(requested_blocks_size_in_bytes, os::vm_page_size());
  83 
  84   _covered_region = mr;
  85 
  86   size_t current_blocks_size_in_bytes = _blocks_region.byte_size();
  87 
  88   if (requested_blocks_size_in_bytes > current_blocks_size_in_bytes) {
  89     // Expand
  90     size_t expand_by = requested_blocks_size_in_bytes - current_blocks_size_in_bytes;
  91     if (!_virtual_space.expand_by(expand_by)) {
  92       vm_exit_out_of_memory(expand_by, "object start array expansion");
  93     }
  94     // Clear *only* the newly allocated region
  95     memset(_blocks_region.end(), clean_block, expand_by);
  96   }
  97 
  98   if (requested_blocks_size_in_bytes < current_blocks_size_in_bytes) {
  99     // Shrink
 100     size_t shrink_by = current_blocks_size_in_bytes - requested_blocks_size_in_bytes;
 101     _virtual_space.shrink_by(shrink_by);
 102   }
 103 
 104   _blocks_region.set_word_size(requested_blocks_size_in_bytes / sizeof(HeapWord));
 105 
 106   assert(requested_blocks_size_in_bytes % sizeof(HeapWord) == 0, "Block table not expanded in word sized increment");
 107   assert(requested_blocks_size_in_bytes == _blocks_region.byte_size(), "Sanity");
 108   assert(block_for_addr(low_bound) == &_raw_base[0], "Checking start of map");
 109   assert(block_for_addr(high_bound-1) <= &_raw_base[_blocks_region.byte_size()-1], "Checking end of map");
 110 }
 111 
 112 void ObjectStartArray::reset() {
 113   memset(_blocks_region.start(), clean_block, _blocks_region.byte_size());
 114 }
 115 
 116 
 117 bool ObjectStartArray::object_starts_in_range(HeapWord* start_addr,
 118                                               HeapWord* end_addr) const {
 119   assert(start_addr <= end_addr, "range is wrong");
 120   if (start_addr > end_addr) {
 121     return false;
 122   }
 123 
 124   jbyte* start_block = block_for_addr(start_addr);
 125   jbyte* end_block = block_for_addr(end_addr);
 126 
 127   for (jbyte* block = start_block; block <= end_block; block++) {
 128     if (*block != clean_block) {
 129       return true;
 130     }
 131   }
 132   // No object starts in this slice; verify this using
 133   // more traditional methods:
 134   assert(object_start(end_addr - 1) <= start_addr,
 135          "Oops an object does start in this slice?");
 136   return false;
 137 }