1 /* 2 * Copyright (c) 1998, 2019, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "code/codeBlob.hpp" 27 #include "code/codeCache.hpp" 28 #include "code/relocInfo.hpp" 29 #include "compiler/disassembler.hpp" 30 #include "interpreter/bytecode.hpp" 31 #include "memory/allocation.inline.hpp" 32 #include "memory/heap.hpp" 33 #include "oops/oop.inline.hpp" 34 #include "prims/forte.hpp" 35 #include "runtime/handles.inline.hpp" 36 #include "runtime/interfaceSupport.hpp" 37 #include "runtime/mutexLocker.hpp" 38 #include "runtime/safepoint.hpp" 39 #include "runtime/sharedRuntime.hpp" 40 #include "runtime/vframe.hpp" 41 #include "services/memoryService.hpp" 42 #ifdef TARGET_ARCH_x86 43 # include "nativeInst_x86.hpp" 44 #endif 45 #ifdef TARGET_ARCH_sparc 46 # include "nativeInst_sparc.hpp" 47 #endif 48 #ifdef TARGET_ARCH_zero 49 # include "nativeInst_zero.hpp" 50 #endif 51 #ifdef TARGET_ARCH_arm 52 # include "nativeInst_arm.hpp" 53 #endif 54 #ifdef TARGET_ARCH_ppc 55 # include "nativeInst_ppc.hpp" 56 #endif 57 #ifdef COMPILER1 58 #include "c1/c1_Runtime1.hpp" 59 #endif 60 61 unsigned int CodeBlob::align_code_offset(int offset) { 62 // align the size to CodeEntryAlignment 63 return 64 ((offset + (int)CodeHeap::header_size() + (CodeEntryAlignment-1)) & ~(CodeEntryAlignment-1)) 65 - (int)CodeHeap::header_size(); 66 } 67 68 69 // This must be consistent with the CodeBlob constructor's layout actions. 70 unsigned int CodeBlob::allocation_size(CodeBuffer* cb, int header_size) { 71 unsigned int size = header_size; 72 size += round_to(cb->total_relocation_size(), oopSize); 73 // align the size to CodeEntryAlignment 74 size = align_code_offset(size); 75 size += round_to(cb->total_content_size(), oopSize); 76 size += round_to(cb->total_oop_size(), oopSize); 77 size += round_to(cb->total_metadata_size(), oopSize); 78 return size; 79 } 80 81 82 // Creates a simple CodeBlob. Sets up the size of the different regions. 83 CodeBlob::CodeBlob(const char* name, int header_size, int size, int frame_complete, int locs_size) { 84 assert(size == round_to(size, oopSize), "unaligned size"); 85 assert(locs_size == round_to(locs_size, oopSize), "unaligned size"); 86 assert(header_size == round_to(header_size, oopSize), "unaligned size"); 87 assert(!UseRelocIndex, "no space allocated for reloc index yet"); 88 89 // Note: If UseRelocIndex is enabled, there needs to be (at least) one 90 // extra word for the relocation information, containing the reloc 91 // index table length. Unfortunately, the reloc index table imple- 92 // mentation is not easily understandable and thus it is not clear 93 // what exactly the format is supposed to be. For now, we just turn 94 // off the use of this table (gri 7/6/2000). 95 96 _name = name; 97 _size = size; 98 _frame_complete_offset = frame_complete; 99 _header_size = header_size; 100 _relocation_size = locs_size; 101 _content_offset = align_code_offset(header_size + _relocation_size); 102 _code_offset = _content_offset; 103 _data_offset = size; 104 _frame_size = 0; 105 set_oop_maps(NULL); 106 } 107 108 109 // Creates a CodeBlob from a CodeBuffer. Sets up the size of the different regions, 110 // and copy code and relocation info. 111 CodeBlob::CodeBlob( 112 const char* name, 113 CodeBuffer* cb, 114 int header_size, 115 int size, 116 int frame_complete, 117 int frame_size, 118 OopMapSet* oop_maps 119 ) { 120 assert(size == round_to(size, oopSize), "unaligned size"); 121 assert(header_size == round_to(header_size, oopSize), "unaligned size"); 122 123 _name = name; 124 _size = size; 125 _frame_complete_offset = frame_complete; 126 _header_size = header_size; 127 _relocation_size = round_to(cb->total_relocation_size(), oopSize); 128 _content_offset = align_code_offset(header_size + _relocation_size); 129 _code_offset = _content_offset + cb->total_offset_of(cb->insts()); 130 _data_offset = _content_offset + round_to(cb->total_content_size(), oopSize); 131 assert(_data_offset <= size, "codeBlob is too small"); 132 133 cb->copy_code_and_locs_to(this); 134 set_oop_maps(oop_maps); 135 _frame_size = frame_size; 136 #ifdef COMPILER1 137 // probably wrong for tiered 138 assert(_frame_size >= -1, "must use frame size or -1 for runtime stubs"); 139 #endif // COMPILER1 140 } 141 142 143 void CodeBlob::set_oop_maps(OopMapSet* p) { 144 // Danger Will Robinson! This method allocates a big 145 // chunk of memory, its your job to free it. 146 if (p != NULL) { 147 // We need to allocate a chunk big enough to hold the OopMapSet and all of its OopMaps 148 _oop_maps = (OopMapSet* )NEW_C_HEAP_ARRAY(unsigned char, p->heap_size(), mtCode); 149 p->copy_to((address)_oop_maps); 150 } else { 151 _oop_maps = NULL; 152 } 153 } 154 155 156 void CodeBlob::trace_new_stub(CodeBlob* stub, const char* name1, const char* name2) { 157 // Do not hold the CodeCache lock during name formatting. 158 assert(!CodeCache_lock->owned_by_self(), "release CodeCache before registering the stub"); 159 160 if (stub != NULL) { 161 char stub_id[256]; 162 assert(strlen(name1) + strlen(name2) < sizeof(stub_id), ""); 163 jio_snprintf(stub_id, sizeof(stub_id), "%s%s", name1, name2); 164 if (PrintStubCode) { 165 ttyLocker ttyl; 166 tty->print_cr("Decoding %s " INTPTR_FORMAT, stub_id, (intptr_t) stub); 167 Disassembler::decode(stub->code_begin(), stub->code_end()); 168 tty->cr(); 169 } 170 Forte::register_stub(stub_id, stub->code_begin(), stub->code_end()); 171 172 if (JvmtiExport::should_post_dynamic_code_generated()) { 173 const char* stub_name = name2; 174 if (name2[0] == '\0') stub_name = name1; 175 JvmtiExport::post_dynamic_code_generated(stub_name, stub->code_begin(), stub->code_end()); 176 } 177 } 178 179 // Track memory usage statistic after releasing CodeCache_lock 180 MemoryService::track_code_cache_memory_usage(); 181 } 182 183 184 void CodeBlob::flush() { 185 if (_oop_maps) { 186 FREE_C_HEAP_ARRAY(unsigned char, _oop_maps, mtCode); 187 _oop_maps = NULL; 188 } 189 _strings.free(); 190 } 191 192 193 OopMap* CodeBlob::oop_map_for_return_address(address return_address) { 194 assert(oop_maps() != NULL, "nope"); 195 return oop_maps()->find_map_at_offset((intptr_t) return_address - (intptr_t) code_begin()); 196 } 197 198 199 //---------------------------------------------------------------------------------------------------- 200 // Implementation of BufferBlob 201 202 203 BufferBlob::BufferBlob(const char* name, int size) 204 : CodeBlob(name, sizeof(BufferBlob), size, CodeOffsets::frame_never_safe, /*locs_size:*/ 0) 205 {} 206 207 BufferBlob* BufferBlob::create(const char* name, int buffer_size) { 208 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 209 210 BufferBlob* blob = NULL; 211 unsigned int size = sizeof(BufferBlob); 212 // align the size to CodeEntryAlignment 213 size = align_code_offset(size); 214 size += round_to(buffer_size, oopSize); 215 assert(name != NULL, "must provide a name"); 216 { 217 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 218 blob = new (size) BufferBlob(name, size); 219 } 220 // Track memory usage statistic after releasing CodeCache_lock 221 MemoryService::track_code_cache_memory_usage(); 222 223 return blob; 224 } 225 226 227 BufferBlob::BufferBlob(const char* name, int size, CodeBuffer* cb) 228 : CodeBlob(name, cb, sizeof(BufferBlob), size, CodeOffsets::frame_never_safe, 0, NULL) 229 {} 230 231 BufferBlob* BufferBlob::create(const char* name, CodeBuffer* cb) { 232 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 233 234 BufferBlob* blob = NULL; 235 unsigned int size = allocation_size(cb, sizeof(BufferBlob)); 236 assert(name != NULL, "must provide a name"); 237 { 238 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 239 blob = new (size) BufferBlob(name, size, cb); 240 } 241 // Track memory usage statistic after releasing CodeCache_lock 242 MemoryService::track_code_cache_memory_usage(); 243 244 return blob; 245 } 246 247 248 void* BufferBlob::operator new(size_t s, unsigned size, bool is_critical) throw() { 249 void* p = CodeCache::allocate(size, is_critical); 250 return p; 251 } 252 253 254 void BufferBlob::free( BufferBlob *blob ) { 255 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 256 blob->flush(); 257 { 258 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 259 CodeCache::free((CodeBlob*)blob); 260 } 261 // Track memory usage statistic after releasing CodeCache_lock 262 MemoryService::track_code_cache_memory_usage(); 263 } 264 265 266 //---------------------------------------------------------------------------------------------------- 267 // Implementation of AdapterBlob 268 269 AdapterBlob::AdapterBlob(int size, CodeBuffer* cb) : 270 BufferBlob("I2C/C2I adapters", size, cb) { 271 CodeCache::commit(this); 272 } 273 274 AdapterBlob* AdapterBlob::create(CodeBuffer* cb) { 275 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 276 277 AdapterBlob* blob = NULL; 278 unsigned int size = allocation_size(cb, sizeof(AdapterBlob)); 279 { 280 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 281 // The parameter 'true' indicates a critical memory allocation. 282 // This means that CodeCacheMinimumFreeSpace is used, if necessary 283 const bool is_critical = true; 284 blob = new (size, is_critical) AdapterBlob(size, cb); 285 } 286 // Track memory usage statistic after releasing CodeCache_lock 287 MemoryService::track_code_cache_memory_usage(); 288 289 return blob; 290 } 291 292 VtableBlob::VtableBlob(const char* name, int size) : 293 BufferBlob(name, size) { 294 } 295 296 VtableBlob* VtableBlob::create(const char* name, int buffer_size) { 297 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 298 299 VtableBlob* blob = NULL; 300 unsigned int size = sizeof(VtableBlob); 301 // align the size to CodeEntryAlignment 302 size = align_code_offset(size); 303 size += round_to(buffer_size, oopSize); 304 assert(name != NULL, "must provide a name"); 305 { 306 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 307 blob = new (size) VtableBlob(name, size); 308 } 309 // Track memory usage statistic after releasing CodeCache_lock 310 MemoryService::track_code_cache_memory_usage(); 311 312 return blob; 313 } 314 315 //---------------------------------------------------------------------------------------------------- 316 // Implementation of MethodHandlesAdapterBlob 317 318 MethodHandlesAdapterBlob* MethodHandlesAdapterBlob::create(int buffer_size) { 319 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 320 321 MethodHandlesAdapterBlob* blob = NULL; 322 unsigned int size = sizeof(MethodHandlesAdapterBlob); 323 // align the size to CodeEntryAlignment 324 size = align_code_offset(size); 325 size += round_to(buffer_size, oopSize); 326 { 327 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 328 // The parameter 'true' indicates a critical memory allocation. 329 // This means that CodeCacheMinimumFreeSpace is used, if necessary 330 const bool is_critical = true; 331 blob = new (size, is_critical) MethodHandlesAdapterBlob(size); 332 } 333 // Track memory usage statistic after releasing CodeCache_lock 334 MemoryService::track_code_cache_memory_usage(); 335 336 return blob; 337 } 338 339 340 //---------------------------------------------------------------------------------------------------- 341 // Implementation of RuntimeStub 342 343 RuntimeStub::RuntimeStub( 344 const char* name, 345 CodeBuffer* cb, 346 int size, 347 int frame_complete, 348 int frame_size, 349 OopMapSet* oop_maps, 350 bool caller_must_gc_arguments 351 ) 352 : CodeBlob(name, cb, sizeof(RuntimeStub), size, frame_complete, frame_size, oop_maps) 353 { 354 _caller_must_gc_arguments = caller_must_gc_arguments; 355 } 356 357 358 RuntimeStub* RuntimeStub::new_runtime_stub(const char* stub_name, 359 CodeBuffer* cb, 360 int frame_complete, 361 int frame_size, 362 OopMapSet* oop_maps, 363 bool caller_must_gc_arguments) 364 { 365 RuntimeStub* stub = NULL; 366 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 367 { 368 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 369 unsigned int size = allocation_size(cb, sizeof(RuntimeStub)); 370 stub = new (size) RuntimeStub(stub_name, cb, size, frame_complete, frame_size, oop_maps, caller_must_gc_arguments); 371 } 372 373 trace_new_stub(stub, "RuntimeStub - ", stub_name); 374 375 return stub; 376 } 377 378 379 void* RuntimeStub::operator new(size_t s, unsigned size) throw() { 380 void* p = CodeCache::allocate(size, true); 381 if (!p) fatal("Initial size of CodeCache is too small"); 382 return p; 383 } 384 385 // operator new shared by all singletons: 386 void* SingletonBlob::operator new(size_t s, unsigned size) throw() { 387 void* p = CodeCache::allocate(size, true); 388 if (!p) fatal("Initial size of CodeCache is too small"); 389 return p; 390 } 391 392 393 //---------------------------------------------------------------------------------------------------- 394 // Implementation of DeoptimizationBlob 395 396 DeoptimizationBlob::DeoptimizationBlob( 397 CodeBuffer* cb, 398 int size, 399 OopMapSet* oop_maps, 400 int unpack_offset, 401 int unpack_with_exception_offset, 402 int unpack_with_reexecution_offset, 403 int frame_size 404 ) 405 : SingletonBlob("DeoptimizationBlob", cb, sizeof(DeoptimizationBlob), size, frame_size, oop_maps) 406 { 407 _unpack_offset = unpack_offset; 408 _unpack_with_exception = unpack_with_exception_offset; 409 _unpack_with_reexecution = unpack_with_reexecution_offset; 410 #ifdef COMPILER1 411 _unpack_with_exception_in_tls = -1; 412 #endif 413 } 414 415 416 DeoptimizationBlob* DeoptimizationBlob::create( 417 CodeBuffer* cb, 418 OopMapSet* oop_maps, 419 int unpack_offset, 420 int unpack_with_exception_offset, 421 int unpack_with_reexecution_offset, 422 int frame_size) 423 { 424 DeoptimizationBlob* blob = NULL; 425 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 426 { 427 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 428 unsigned int size = allocation_size(cb, sizeof(DeoptimizationBlob)); 429 blob = new (size) DeoptimizationBlob(cb, 430 size, 431 oop_maps, 432 unpack_offset, 433 unpack_with_exception_offset, 434 unpack_with_reexecution_offset, 435 frame_size); 436 } 437 438 trace_new_stub(blob, "DeoptimizationBlob"); 439 440 return blob; 441 } 442 443 444 //---------------------------------------------------------------------------------------------------- 445 // Implementation of UncommonTrapBlob 446 447 #ifdef COMPILER2 448 UncommonTrapBlob::UncommonTrapBlob( 449 CodeBuffer* cb, 450 int size, 451 OopMapSet* oop_maps, 452 int frame_size 453 ) 454 : SingletonBlob("UncommonTrapBlob", cb, sizeof(UncommonTrapBlob), size, frame_size, oop_maps) 455 {} 456 457 458 UncommonTrapBlob* UncommonTrapBlob::create( 459 CodeBuffer* cb, 460 OopMapSet* oop_maps, 461 int frame_size) 462 { 463 UncommonTrapBlob* blob = NULL; 464 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 465 { 466 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 467 unsigned int size = allocation_size(cb, sizeof(UncommonTrapBlob)); 468 blob = new (size) UncommonTrapBlob(cb, size, oop_maps, frame_size); 469 } 470 471 trace_new_stub(blob, "UncommonTrapBlob"); 472 473 return blob; 474 } 475 476 477 #endif // COMPILER2 478 479 480 //---------------------------------------------------------------------------------------------------- 481 // Implementation of ExceptionBlob 482 483 #ifdef COMPILER2 484 ExceptionBlob::ExceptionBlob( 485 CodeBuffer* cb, 486 int size, 487 OopMapSet* oop_maps, 488 int frame_size 489 ) 490 : SingletonBlob("ExceptionBlob", cb, sizeof(ExceptionBlob), size, frame_size, oop_maps) 491 {} 492 493 494 ExceptionBlob* ExceptionBlob::create( 495 CodeBuffer* cb, 496 OopMapSet* oop_maps, 497 int frame_size) 498 { 499 ExceptionBlob* blob = NULL; 500 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 501 { 502 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 503 unsigned int size = allocation_size(cb, sizeof(ExceptionBlob)); 504 blob = new (size) ExceptionBlob(cb, size, oop_maps, frame_size); 505 } 506 507 trace_new_stub(blob, "ExceptionBlob"); 508 509 return blob; 510 } 511 512 513 #endif // COMPILER2 514 515 516 //---------------------------------------------------------------------------------------------------- 517 // Implementation of SafepointBlob 518 519 SafepointBlob::SafepointBlob( 520 CodeBuffer* cb, 521 int size, 522 OopMapSet* oop_maps, 523 int frame_size 524 ) 525 : SingletonBlob("SafepointBlob", cb, sizeof(SafepointBlob), size, frame_size, oop_maps) 526 {} 527 528 529 SafepointBlob* SafepointBlob::create( 530 CodeBuffer* cb, 531 OopMapSet* oop_maps, 532 int frame_size) 533 { 534 SafepointBlob* blob = NULL; 535 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock 536 { 537 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 538 unsigned int size = allocation_size(cb, sizeof(SafepointBlob)); 539 blob = new (size) SafepointBlob(cb, size, oop_maps, frame_size); 540 } 541 542 trace_new_stub(blob, "SafepointBlob"); 543 544 return blob; 545 } 546 547 548 //---------------------------------------------------------------------------------------------------- 549 // Verification and printing 550 551 void CodeBlob::verify() { 552 ShouldNotReachHere(); 553 } 554 555 void CodeBlob::print_on(outputStream* st) const { 556 st->print_cr("[CodeBlob (" INTPTR_FORMAT ")]", p2i(this)); 557 st->print_cr("Framesize: %d", _frame_size); 558 } 559 560 void CodeBlob::print_value_on(outputStream* st) const { 561 st->print_cr("[CodeBlob]"); 562 } 563 564 void BufferBlob::verify() { 565 // unimplemented 566 } 567 568 void BufferBlob::print_on(outputStream* st) const { 569 CodeBlob::print_on(st); 570 print_value_on(st); 571 } 572 573 void BufferBlob::print_value_on(outputStream* st) const { 574 st->print_cr("BufferBlob (" INTPTR_FORMAT ") used for %s", p2i(this), name()); 575 } 576 577 void RuntimeStub::verify() { 578 // unimplemented 579 } 580 581 void RuntimeStub::print_on(outputStream* st) const { 582 ttyLocker ttyl; 583 CodeBlob::print_on(st); 584 st->print("Runtime Stub (" INTPTR_FORMAT "): ", p2i(this)); 585 st->print_cr("%s", name()); 586 Disassembler::decode((CodeBlob*)this, st); 587 } 588 589 void RuntimeStub::print_value_on(outputStream* st) const { 590 st->print("RuntimeStub (" INTPTR_FORMAT "): ", p2i(this)); st->print("%s", name()); 591 } 592 593 void SingletonBlob::verify() { 594 // unimplemented 595 } 596 597 void SingletonBlob::print_on(outputStream* st) const { 598 ttyLocker ttyl; 599 CodeBlob::print_on(st); 600 st->print_cr("%s", name()); 601 Disassembler::decode((CodeBlob*)this, st); 602 } 603 604 void SingletonBlob::print_value_on(outputStream* st) const { 605 st->print_cr("%s", name()); 606 } 607 608 void DeoptimizationBlob::print_value_on(outputStream* st) const { 609 st->print_cr("Deoptimization (frame not available)"); 610 }