1 /*
   2  * Copyright (c) 1997, 2019, Oracle and/or its affiliates. All rights reserved.
   3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4  *
   5  * This code is free software; you can redistribute it and/or modify it
   6  * under the terms of the GNU General Public License version 2 only, as
   7  * published by the Free Software Foundation.
   8  *
   9  * This code is distributed in the hope that it will be useful, but WITHOUT
  10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12  * version 2 for more details (a copy is included in the LICENSE file that
  13  * accompanied this code).
  14  *
  15  * You should have received a copy of the GNU General Public License version
  16  * 2 along with this work; if not, write to the Free Software Foundation,
  17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18  *
  19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  20  * or visit www.oracle.com if you need additional information or have any
  21  * questions.
  22  *
  23  */
  24 
  25 #include "precompiled.hpp"
  26 #include "code/codeBlob.hpp"
  27 #include "code/codeCache.hpp"
  28 #include "code/stubs.hpp"
  29 #include "memory/allocation.inline.hpp"
  30 #include "oops/oop.inline.hpp"
  31 #include "runtime/mutexLocker.hpp"
  32 #include "utilities/align.hpp"
  33 
  34 
  35 // Implementation of StubQueue
  36 //
  37 // Standard wrap-around queue implementation; the queue dimensions
  38 // are specified by the _queue_begin & _queue_end indices. The queue
  39 // can be in two states (transparent to the outside):
  40 //
  41 // a) contiguous state: all queue entries in one block (or empty)
  42 //
  43 // Queue: |...|XXXXXXX|...............|
  44 //        ^0  ^begin  ^end            ^size = limit
  45 //            |_______|
  46 //            one block
  47 //
  48 // b) non-contiguous state: queue entries in two blocks
  49 //
  50 // Queue: |XXX|.......|XXXXXXX|.......|
  51 //        ^0  ^end    ^begin  ^limit  ^size
  52 //        |___|       |_______|
  53 //         1st block  2nd block
  54 //
  55 // In the non-contiguous state, the wrap-around point is
  56 // indicated via the _buffer_limit index since the last
  57 // queue entry may not fill up the queue completely in
  58 // which case we need to know where the 2nd block's end
  59 // is to do the proper wrap-around. When removing the
  60 // last entry of the 2nd block, _buffer_limit is reset
  61 // to _buffer_size.
  62 //
  63 // CAUTION: DO NOT MESS WITH THIS CODE IF YOU CANNOT PROVE
  64 // ITS CORRECTNESS! THIS CODE IS MORE SUBTLE THAN IT LOOKS!
  65 
  66 
  67 StubQueue::StubQueue(StubInterface* stub_interface, int buffer_size,
  68                      Mutex* lock, const char* name) : _mutex(lock) {
  69   intptr_t size = align_up(buffer_size, 2*BytesPerWord);
  70   BufferBlob* blob = BufferBlob::create(name, size);
  71   if( blob == NULL) {
  72     vm_exit_out_of_memory(size, OOM_MALLOC_ERROR, "CodeCache: no room for %s", name);
  73   }
  74   _stub_interface  = stub_interface;
  75   _buffer_size     = blob->content_size();
  76   _buffer_limit    = blob->content_size();
  77   _stub_buffer     = blob->content_begin();
  78   _queue_begin     = 0;
  79   _queue_end       = 0;
  80   _number_of_stubs = 0;
  81 }
  82 
  83 
  84 StubQueue::~StubQueue() {
  85   // Note: Currently StubQueues are never destroyed so nothing needs to be done here.
  86   //       If we want to implement the destructor, we need to release the BufferBlob
  87   //       allocated in the constructor (i.e., we need to keep it around or look it
  88   //       up via CodeCache::find_blob(...).
  89   Unimplemented();
  90 }
  91 
  92 void StubQueue::deallocate_unused_tail() {
  93   CodeBlob* blob = CodeCache::find_blob((void*)_stub_buffer);
  94   CodeCache::free_unused_tail(blob, used_space());
  95   // Update the limits to the new, trimmed CodeBlob size
  96   _buffer_size = blob->content_size();
  97   _buffer_limit = blob->content_size();
  98 }
  99 
 100 Stub* StubQueue::stub_containing(address pc) const {
 101   if (contains(pc)) {
 102     for (Stub* s = first(); s != NULL; s = next(s)) {
 103       if (stub_contains(s, pc)) return s;
 104     }
 105   }
 106   return NULL;
 107 }
 108 
 109 
 110 Stub* StubQueue::request_committed(int code_size) {
 111   Stub* s = request(code_size);
 112   CodeStrings strings;
 113   if (s != NULL) commit(code_size, strings);
 114   return s;
 115 }
 116 
 117 
 118 Stub* StubQueue::request(int requested_code_size) {
 119   assert(requested_code_size > 0, "requested_code_size must be > 0");
 120   if (_mutex != NULL) _mutex->lock_without_safepoint_check();
 121   Stub* s = current_stub();
 122   int requested_size = align_up(stub_code_size_to_size(requested_code_size), CodeEntryAlignment);
 123   if (requested_size <= available_space()) {
 124     if (is_contiguous()) {
 125       // Queue: |...|XXXXXXX|.............|
 126       //        ^0  ^begin  ^end          ^size = limit
 127       assert(_buffer_limit == _buffer_size, "buffer must be fully usable");
 128       if (_queue_end + requested_size <= _buffer_size) {
 129         // code fits in at the end => nothing to do
 130         CodeStrings strings;
 131         stub_initialize(s, requested_size, strings);
 132         return s;
 133       } else {
 134         // stub doesn't fit in at the queue end
 135         // => reduce buffer limit & wrap around
 136         assert(!is_empty(), "just checkin'");
 137         _buffer_limit = _queue_end;
 138         _queue_end = 0;
 139       }
 140     }
 141   }
 142   if (requested_size <= available_space()) {
 143     assert(!is_contiguous(), "just checkin'");
 144     assert(_buffer_limit <= _buffer_size, "queue invariant broken");
 145     // Queue: |XXX|.......|XXXXXXX|.......|
 146     //        ^0  ^end    ^begin  ^limit  ^size
 147     s = current_stub();
 148     CodeStrings strings;
 149     stub_initialize(s, requested_size, strings);
 150     return s;
 151   }
 152   // Not enough space left
 153   if (_mutex != NULL) _mutex->unlock();
 154   return NULL;
 155 }
 156 
 157 
 158 void StubQueue::commit(int committed_code_size, CodeStrings& strings) {
 159   assert(committed_code_size > 0, "committed_code_size must be > 0");
 160   int committed_size = align_up(stub_code_size_to_size(committed_code_size), CodeEntryAlignment);
 161   Stub* s = current_stub();
 162   assert(committed_size <= stub_size(s), "committed size must not exceed requested size");
 163   stub_initialize(s, committed_size, strings);
 164   _queue_end += committed_size;
 165   _number_of_stubs++;
 166   if (_mutex != NULL) _mutex->unlock();
 167   debug_only(stub_verify(s);)
 168 }
 169 
 170 
 171 void StubQueue::remove_first() {
 172   if (number_of_stubs() == 0) return;
 173   Stub* s = first();
 174   debug_only(stub_verify(s);)
 175   stub_finalize(s);
 176   _queue_begin += stub_size(s);
 177   assert(_queue_begin <= _buffer_limit, "sanity check");
 178   if (_queue_begin == _queue_end) {
 179     // buffer empty
 180     // => reset queue indices
 181     _queue_begin  = 0;
 182     _queue_end    = 0;
 183     _buffer_limit = _buffer_size;
 184   } else if (_queue_begin == _buffer_limit) {
 185     // buffer limit reached
 186     // => reset buffer limit & wrap around
 187     _buffer_limit = _buffer_size;
 188     _queue_begin = 0;
 189   }
 190   _number_of_stubs--;
 191 }
 192 
 193 
 194 void StubQueue::remove_first(int n) {
 195   int i = MIN2(n, number_of_stubs());
 196   while (i-- > 0) remove_first();
 197 }
 198 
 199 
 200 void StubQueue::remove_all(){
 201   debug_only(verify();)
 202   remove_first(number_of_stubs());
 203   assert(number_of_stubs() == 0, "sanity check");
 204 }
 205 
 206 
 207 void StubQueue::verify() {
 208   // verify only if initialized
 209   if (_stub_buffer == NULL) return;
 210   MutexLocker lock(_mutex, Mutex::_no_safepoint_check_flag);
 211   // verify index boundaries
 212   guarantee(0 <= _buffer_size, "buffer size must be positive");
 213   guarantee(0 <= _buffer_limit && _buffer_limit <= _buffer_size , "_buffer_limit out of bounds");
 214   guarantee(0 <= _queue_begin  && _queue_begin  <  _buffer_limit, "_queue_begin out of bounds");
 215   guarantee(0 <= _queue_end    && _queue_end    <= _buffer_limit, "_queue_end   out of bounds");
 216   // verify alignment
 217   guarantee(_buffer_size  % CodeEntryAlignment == 0, "_buffer_size  not aligned");
 218   guarantee(_buffer_limit % CodeEntryAlignment == 0, "_buffer_limit not aligned");
 219   guarantee(_queue_begin  % CodeEntryAlignment == 0, "_queue_begin  not aligned");
 220   guarantee(_queue_end    % CodeEntryAlignment == 0, "_queue_end    not aligned");
 221   // verify buffer limit/size relationship
 222   if (is_contiguous()) {
 223     guarantee(_buffer_limit == _buffer_size, "_buffer_limit must equal _buffer_size");
 224   }
 225   // verify contents
 226   int n = 0;
 227   for (Stub* s = first(); s != NULL; s = next(s)) {
 228     stub_verify(s);
 229     n++;
 230   }
 231   guarantee(n == number_of_stubs(), "number of stubs inconsistent");
 232   guarantee(_queue_begin != _queue_end || n == 0, "buffer indices must be the same");
 233 }
 234 
 235 
 236 void StubQueue::print() {
 237   MutexLocker lock(_mutex, Mutex::_no_safepoint_check_flag);
 238   for (Stub* s = first(); s != NULL; s = next(s)) {
 239     stub_print(s);
 240   }
 241 }