1 /* 2 * Copyright (c) 1997, 2018, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #ifndef SHARE_VM_CODE_CODECACHE_HPP 26 #define SHARE_VM_CODE_CODECACHE_HPP 27 28 #include "code/codeBlob.hpp" 29 #include "code/nmethod.hpp" 30 #include "memory/allocation.hpp" 31 #include "memory/heap.hpp" 32 #include "oops/instanceKlass.hpp" 33 #include "oops/oopsHierarchy.hpp" 34 #include "runtime/mutexLocker.hpp" 35 36 // The CodeCache implements the code cache for various pieces of generated 37 // code, e.g., compiled java methods, runtime stubs, transition frames, etc. 38 // The entries in the CodeCache are all CodeBlob's. 39 40 // -- Implementation -- 41 // The CodeCache consists of one or more CodeHeaps, each of which contains 42 // CodeBlobs of a specific CodeBlobType. Currently heaps for the following 43 // types are available: 44 // - Non-nmethods: Non-nmethods like Buffers, Adapters and Runtime Stubs 45 // - Profiled nmethods: nmethods that are profiled, i.e., those 46 // executed at level 2 or 3 47 // - Non-Profiled nmethods: nmethods that are not profiled, i.e., those 48 // executed at level 1 or 4 and native methods 49 // - All: Used for code of all types if code cache segmentation is disabled. 50 // 51 // In the rare case of the non-nmethod code heap getting full, non-nmethod code 52 // will be stored in the non-profiled code heap as a fallback solution. 53 // 54 // Depending on the availability of compilers and TieredCompilation there 55 // may be fewer heaps. The size of the code heaps depends on the values of 56 // ReservedCodeCacheSize, NonProfiledCodeHeapSize and ProfiledCodeHeapSize 57 // (see CodeCache::heap_available(..) and CodeCache::initialize_heaps(..) 58 // for details). 59 // 60 // Code cache segmentation is controlled by the flag SegmentedCodeCache. 61 // If turned off, all code types are stored in a single code heap. By default 62 // code cache segmentation is turned on if TieredCompilation is enabled and 63 // ReservedCodeCacheSize >= 240 MB. 64 // 65 // All methods of the CodeCache accepting a CodeBlobType only apply to 66 // CodeBlobs of the given type. For example, iteration over the 67 // CodeBlobs of a specific type can be done by using CodeCache::first_blob(..) 68 // and CodeCache::next_blob(..) and providing the corresponding CodeBlobType. 69 // 70 // IMPORTANT: If you add new CodeHeaps to the code cache or change the 71 // existing ones, make sure to adapt the dtrace scripts (jhelper.d) for 72 // Solaris and BSD. 73 74 class OopClosure; 75 class KlassDepChange; 76 77 class CodeCache : AllStatic { 78 friend class VMStructs; 79 friend class JVMCIVMStructs; 80 template <class T, class Filter> friend class CodeBlobIterator; 81 friend class WhiteBox; 82 friend class CodeCacheLoader; 83 private: 84 // CodeHeaps of the cache 85 static GrowableArray<CodeHeap*>* _heaps; 86 static GrowableArray<CodeHeap*>* _compiled_heaps; 87 static GrowableArray<CodeHeap*>* _nmethod_heaps; 88 static GrowableArray<CodeHeap*>* _allocable_heaps; 89 90 static address _low_bound; // Lower bound of CodeHeap addresses 91 static address _high_bound; // Upper bound of CodeHeap addresses 92 static int _number_of_nmethods_with_dependencies; // Total number of nmethods with dependencies 93 static bool _needs_cache_clean; // True if inline caches of the nmethods needs to be flushed 94 static nmethod* _scavenge_root_nmethods; // linked via nm->scavenge_root_link() 95 96 static void mark_scavenge_root_nmethods() PRODUCT_RETURN; 97 static void verify_perm_nmethods(CodeBlobClosure* f_or_null) PRODUCT_RETURN; 98 99 // CodeHeap management 100 static void initialize_heaps(); // Initializes the CodeHeaps 101 // Check the code heap sizes set by the user via command line 102 static void check_heap_sizes(size_t non_nmethod_size, size_t profiled_size, size_t non_profiled_size, size_t cache_size, bool all_set); 103 // Creates a new heap with the given name and size, containing CodeBlobs of the given type 104 static void add_heap(ReservedSpace rs, const char* name, int code_blob_type); 105 static CodeHeap* get_code_heap_containing(void* p); // Returns the CodeHeap containing the given pointer, or NULL 106 static CodeHeap* get_code_heap(const CodeBlob* cb); // Returns the CodeHeap for the given CodeBlob 107 static CodeHeap* get_code_heap(int code_blob_type); // Returns the CodeHeap for the given CodeBlobType 108 // Returns the name of the VM option to set the size of the corresponding CodeHeap 109 static const char* get_code_heap_flag_name(int code_blob_type); 110 static size_t page_size(bool aligned = true); // Returns the page size used by the CodeCache 111 static ReservedCodeSpace reserve_heap_memory(size_t size); // Reserves one continuous chunk of memory for the CodeHeaps 112 113 // Iteration 114 static CodeBlob* first_blob(CodeHeap* heap); // Returns the first CodeBlob on the given CodeHeap 115 static CodeBlob* first_blob(int code_blob_type); // Returns the first CodeBlob of the given type 116 static CodeBlob* next_blob(CodeHeap* heap, CodeBlob* cb); // Returns the next CodeBlob on the given CodeHeap 117 118 static size_t bytes_allocated_in_freelists(); 119 static int allocated_segments(); 120 static size_t freelists_length(); 121 122 static void set_scavenge_root_nmethods(nmethod* nm) { _scavenge_root_nmethods = nm; } 123 static void prune_scavenge_root_nmethods(); 124 static void unlink_scavenge_root_nmethod(nmethod* nm, nmethod* prev); 125 126 // Make private to prevent unsafe calls. Not all CodeBlob*'s are embedded in a CodeHeap. 127 static bool contains(CodeBlob *p) { fatal("don't call me!"); return false; } 128 129 public: 130 // Initialization 131 static void initialize(); 132 133 static int code_heap_compare(CodeHeap* const &lhs, CodeHeap* const &rhs); 134 135 static void add_heap(CodeHeap* heap); 136 static const GrowableArray<CodeHeap*>* heaps() { return _heaps; } 137 static const GrowableArray<CodeHeap*>* compiled_heaps() { return _compiled_heaps; } 138 static const GrowableArray<CodeHeap*>* nmethod_heaps() { return _nmethod_heaps; } 139 140 // Allocation/administration 141 static CodeBlob* allocate(int size, int code_blob_type, int orig_code_blob_type = CodeBlobType::All); // allocates a new CodeBlob 142 static void commit(CodeBlob* cb); // called when the allocated CodeBlob has been filled 143 static int alignment_unit(); // guaranteed alignment of all CodeBlobs 144 static int alignment_offset(); // guaranteed offset of first CodeBlob byte within alignment unit (i.e., allocation header) 145 static void free(CodeBlob* cb); // frees a CodeBlob 146 static void free_unused_tail(CodeBlob* cb, size_t used); // frees the unused tail of a CodeBlob (only used by TemplateInterpreter::initialize()) 147 static bool contains(void *p); // returns whether p is included 148 static bool contains(nmethod* nm); // returns whether nm is included 149 static void blobs_do(void f(CodeBlob* cb)); // iterates over all CodeBlobs 150 static void blobs_do(CodeBlobClosure* f); // iterates over all CodeBlobs 151 static void nmethods_do(void f(nmethod* nm)); // iterates over all nmethods 152 static void metadata_do(void f(Metadata* m)); // iterates over metadata in alive nmethods 153 154 // Lookup 155 static CodeBlob* find_blob(void* start); // Returns the CodeBlob containing the given address 156 static CodeBlob* find_blob_unsafe(void* start); // Same as find_blob but does not fail if looking up a zombie method 157 static nmethod* find_nmethod(void* start); // Returns the nmethod containing the given address 158 static CompiledMethod* find_compiled(void* start); 159 160 static int blob_count(); // Returns the total number of CodeBlobs in the cache 161 static int blob_count(int code_blob_type); 162 static int adapter_count(); // Returns the total number of Adapters in the cache 163 static int adapter_count(int code_blob_type); 164 static int nmethod_count(); // Returns the total number of nmethods in the cache 165 static int nmethod_count(int code_blob_type); 166 167 // GC support 168 static void gc_epilogue(); 169 static void gc_prologue(); 170 static void verify_oops(); 171 // If any oops are not marked this method unloads (i.e., breaks root links 172 // to) any unmarked codeBlobs in the cache. Sets "marked_for_unloading" 173 // to "true" iff some code got unloaded. 174 // "unloading_occurred" controls whether metadata should be cleaned because of class unloading. 175 static void do_unloading(BoolObjectClosure* is_alive, bool unloading_occurred); 176 static void asserted_non_scavengable_nmethods_do(CodeBlobClosure* f = NULL) PRODUCT_RETURN; 177 178 // Apply f to every live code blob in scavengable nmethods. Prune nmethods 179 // from the list of scavengable nmethods if f->fix_relocations() and a nmethod 180 // no longer has scavengable oops. If f->fix_relocations(), then f must copy 181 // objects to their new location immediately to avoid fixing nmethods on the 182 // basis of the old object locations. 183 static void scavenge_root_nmethods_do(CodeBlobToOopClosure* f); 184 185 static nmethod* scavenge_root_nmethods() { return _scavenge_root_nmethods; } 186 // register_scavenge_root_nmethod() conditionally adds the nmethod to the list 187 // if it is not already on the list and has a scavengeable root 188 static void register_scavenge_root_nmethod(nmethod* nm); 189 static void verify_scavenge_root_nmethod(nmethod* nm); 190 static void add_scavenge_root_nmethod(nmethod* nm); 191 static void drop_scavenge_root_nmethod(nmethod* nm); 192 193 // Printing/debugging 194 static void print(); // prints summary 195 static void print_internals(); 196 static void print_memory_overhead(); 197 static void verify(); // verifies the code cache 198 static void print_trace(const char* event, CodeBlob* cb, int size = 0) PRODUCT_RETURN; 199 static void print_summary(outputStream* st, bool detailed = true); // Prints a summary of the code cache usage 200 static void log_state(outputStream* st); 201 static const char* get_code_heap_name(int code_blob_type) { return (heap_available(code_blob_type) ? get_code_heap(code_blob_type)->name() : "Unused"); } 202 static void report_codemem_full(int code_blob_type, bool print); 203 204 // Dcmd (Diagnostic commands) 205 static void print_codelist(outputStream* st); 206 static void print_layout(outputStream* st); 207 208 // The full limits of the codeCache 209 static address low_bound() { return _low_bound; } 210 static address low_bound(int code_blob_type); 211 static address high_bound() { return _high_bound; } 212 static address high_bound(int code_blob_type); 213 214 // Have to use far call instructions to call this pc. 215 static bool is_far_target(address pc); 216 217 // Profiling 218 static size_t capacity(); 219 static size_t unallocated_capacity(int code_blob_type); 220 static size_t unallocated_capacity(); 221 static size_t max_capacity(); 222 223 static double reverse_free_ratio(int code_blob_type); 224 225 static bool needs_cache_clean() { return _needs_cache_clean; } 226 static void set_needs_cache_clean(bool v) { _needs_cache_clean = v; } 227 228 static void clear_inline_caches(); // clear all inline caches 229 static void cleanup_inline_caches(); // clean unloaded/zombie nmethods from inline caches 230 static void do_unloading_nmethod_caches(bool class_unloading_occurred); // clean all nmethod caches for unloading, including inline caches 231 232 // Returns true if an own CodeHeap for the given CodeBlobType is available 233 static bool heap_available(int code_blob_type); 234 235 // Returns the CodeBlobType for the given CompiledMethod 236 static int get_code_blob_type(CompiledMethod* cm) { 237 return get_code_heap(cm)->code_blob_type(); 238 } 239 240 static bool code_blob_type_accepts_compiled(int type) { 241 bool result = type == CodeBlobType::All || type <= CodeBlobType::MethodProfiled; 242 AOT_ONLY( result = result || type == CodeBlobType::AOT; ) 243 return result; 244 } 245 246 static bool code_blob_type_accepts_nmethod(int type) { 247 return type == CodeBlobType::All || type <= CodeBlobType::MethodProfiled; 248 } 249 250 static bool code_blob_type_accepts_allocable(int type) { 251 return type <= CodeBlobType::All; 252 } 253 254 255 // Returns the CodeBlobType for the given compilation level 256 static int get_code_blob_type(int comp_level) { 257 if (comp_level == CompLevel_none || 258 comp_level == CompLevel_simple || 259 comp_level == CompLevel_full_optimization) { 260 // Non profiled methods 261 return CodeBlobType::MethodNonProfiled; 262 } else if (comp_level == CompLevel_limited_profile || 263 comp_level == CompLevel_full_profile) { 264 // Profiled methods 265 return CodeBlobType::MethodProfiled; 266 } 267 ShouldNotReachHere(); 268 return 0; 269 } 270 271 static void verify_clean_inline_caches(); 272 static void verify_icholder_relocations(); 273 274 // Deoptimization 275 private: 276 static int mark_for_deoptimization(KlassDepChange& changes); 277 #ifdef HOTSWAP 278 static int mark_for_evol_deoptimization(InstanceKlass* dependee); 279 #endif // HOTSWAP 280 281 public: 282 static void mark_all_nmethods_for_deoptimization(); 283 static int mark_for_deoptimization(Method* dependee); 284 static void make_marked_nmethods_not_entrant(); 285 286 // Flushing and deoptimization 287 static void flush_dependents_on(InstanceKlass* dependee); 288 #ifdef HOTSWAP 289 // Flushing and deoptimization in case of evolution 290 static void flush_evol_dependents_on(InstanceKlass* dependee); 291 #endif // HOTSWAP 292 // Support for fullspeed debugging 293 static void flush_dependents_on_method(const methodHandle& dependee); 294 295 // tells how many nmethods have dependencies 296 static int number_of_nmethods_with_dependencies(); 297 298 static int get_codemem_full_count(int code_blob_type) { 299 CodeHeap* heap = get_code_heap(code_blob_type); 300 return (heap != NULL) ? heap->full_count() : 0; 301 } 302 303 // CodeHeap State Analytics. 304 // interface methods for CodeHeap printing, called by CompileBroker 305 static void aggregate(outputStream *out, size_t granularity); 306 static void discard(outputStream *out); 307 static void print_usedSpace(outputStream *out); 308 static void print_freeSpace(outputStream *out); 309 static void print_count(outputStream *out); 310 static void print_space(outputStream *out); 311 static void print_age(outputStream *out); 312 static void print_names(outputStream *out); 313 }; 314 315 316 // Iterator to iterate over nmethods in the CodeCache. 317 template <class T, class Filter> class CodeBlobIterator : public StackObj { 318 private: 319 CodeBlob* _code_blob; // Current CodeBlob 320 GrowableArrayIterator<CodeHeap*> _heap; 321 GrowableArrayIterator<CodeHeap*> _end; 322 323 public: 324 CodeBlobIterator(T* nm = NULL) { 325 if (Filter::heaps() == NULL) { 326 return; 327 } 328 _heap = Filter::heaps()->begin(); 329 _end = Filter::heaps()->end(); 330 // If set to NULL, initialized by first call to next() 331 _code_blob = (CodeBlob*)nm; 332 if (nm != NULL) { 333 while(!(*_heap)->contains_blob(_code_blob)) { 334 ++_heap; 335 } 336 assert((*_heap)->contains_blob(_code_blob), "match not found"); 337 } 338 } 339 340 // Advance iterator to next blob 341 bool next() { 342 assert_locked_or_safepoint(CodeCache_lock); 343 344 bool result = next_blob(); 345 while (!result && _heap != _end) { 346 // Advance to next code heap of segmented code cache 347 if (++_heap == _end) { 348 break; 349 } 350 result = next_blob(); 351 } 352 353 return result; 354 } 355 356 // Advance iterator to next alive blob 357 bool next_alive() { 358 bool result = next(); 359 while(result && !_code_blob->is_alive()) { 360 result = next(); 361 } 362 return result; 363 } 364 365 bool end() const { return _code_blob == NULL; } 366 T* method() const { return (T*)_code_blob; } 367 368 private: 369 370 // Advance iterator to the next blob in the current code heap 371 bool next_blob() { 372 if (_heap == _end) { 373 return false; 374 } 375 CodeHeap *heap = *_heap; 376 // Get first method CodeBlob 377 if (_code_blob == NULL) { 378 _code_blob = CodeCache::first_blob(heap); 379 if (_code_blob == NULL) { 380 return false; 381 } else if (Filter::apply(_code_blob)) { 382 return true; 383 } 384 } 385 // Search for next method CodeBlob 386 _code_blob = CodeCache::next_blob(heap, _code_blob); 387 while (_code_blob != NULL && !Filter::apply(_code_blob)) { 388 _code_blob = CodeCache::next_blob(heap, _code_blob); 389 } 390 return _code_blob != NULL; 391 } 392 }; 393 394 395 struct CompiledMethodFilter { 396 static bool apply(CodeBlob* cb) { return cb->is_compiled(); } 397 static const GrowableArray<CodeHeap*>* heaps() { return CodeCache::compiled_heaps(); } 398 }; 399 400 401 struct NMethodFilter { 402 static bool apply(CodeBlob* cb) { return cb->is_nmethod(); } 403 static const GrowableArray<CodeHeap*>* heaps() { return CodeCache::nmethod_heaps(); } 404 }; 405 406 407 typedef CodeBlobIterator<CompiledMethod, CompiledMethodFilter> CompiledMethodIterator; 408 typedef CodeBlobIterator<nmethod, NMethodFilter> NMethodIterator; 409 410 #endif // SHARE_VM_CODE_CODECACHE_HPP