1 /* 2 * Copyright (c) 2000, 2017, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "classfile/systemDictionary.hpp" 27 #include "compiler/compilerOracle.hpp" 28 #include "interpreter/bytecode.hpp" 29 #include "interpreter/bytecodeStream.hpp" 30 #include "interpreter/linkResolver.hpp" 31 #include "memory/heapInspection.hpp" 32 #include "memory/resourceArea.hpp" 33 #include "oops/methodData.hpp" 34 #include "prims/jvmtiRedefineClasses.hpp" 35 #include "runtime/arguments.hpp" 36 #include "runtime/compilationPolicy.hpp" 37 #include "runtime/deoptimization.hpp" 38 #include "runtime/handles.inline.hpp" 39 #include "runtime/orderAccess.inline.hpp" 40 #include "utilities/copy.hpp" 41 42 // ================================================================== 43 // DataLayout 44 // 45 // Overlay for generic profiling data. 46 47 // Some types of data layouts need a length field. 48 bool DataLayout::needs_array_len(u1 tag) { 49 return (tag == multi_branch_data_tag) || (tag == arg_info_data_tag) || (tag == parameters_type_data_tag); 50 } 51 52 // Perform generic initialization of the data. More specific 53 // initialization occurs in overrides of ProfileData::post_initialize. 54 void DataLayout::initialize(u1 tag, u2 bci, int cell_count) { 55 _header._bits = (intptr_t)0; 56 _header._struct._tag = tag; 57 _header._struct._bci = bci; 58 for (int i = 0; i < cell_count; i++) { 59 set_cell_at(i, (intptr_t)0); 60 } 61 if (needs_array_len(tag)) { 62 set_cell_at(ArrayData::array_len_off_set, cell_count - 1); // -1 for header. 63 } 64 if (tag == call_type_data_tag) { 65 CallTypeData::initialize(this, cell_count); 66 } else if (tag == virtual_call_type_data_tag) { 67 VirtualCallTypeData::initialize(this, cell_count); 68 } 69 } 70 71 void DataLayout::clean_weak_klass_links(BoolObjectClosure* cl) { 72 ResourceMark m; 73 data_in()->clean_weak_klass_links(cl); 74 } 75 76 77 // ================================================================== 78 // ProfileData 79 // 80 // A ProfileData object is created to refer to a section of profiling 81 // data in a structured way. 82 83 // Constructor for invalid ProfileData. 84 ProfileData::ProfileData() { 85 _data = NULL; 86 } 87 88 char* ProfileData::print_data_on_helper(const MethodData* md) const { 89 DataLayout* dp = md->extra_data_base(); 90 DataLayout* end = md->args_data_limit(); 91 stringStream ss; 92 for (;; dp = MethodData::next_extra(dp)) { 93 assert(dp < end, "moved past end of extra data"); 94 switch(dp->tag()) { 95 case DataLayout::speculative_trap_data_tag: 96 if (dp->bci() == bci()) { 97 SpeculativeTrapData* data = new SpeculativeTrapData(dp); 98 int trap = data->trap_state(); 99 char buf[100]; 100 ss.print("trap/"); 101 data->method()->print_short_name(&ss); 102 ss.print("(%s) ", Deoptimization::format_trap_state(buf, sizeof(buf), trap)); 103 } 104 break; 105 case DataLayout::bit_data_tag: 106 break; 107 case DataLayout::no_tag: 108 case DataLayout::arg_info_data_tag: 109 return ss.as_string(); 110 break; 111 default: 112 fatal("unexpected tag %d", dp->tag()); 113 } 114 } 115 return NULL; 116 } 117 118 void ProfileData::print_data_on(outputStream* st, const MethodData* md) const { 119 print_data_on(st, print_data_on_helper(md)); 120 } 121 122 void ProfileData::print_shared(outputStream* st, const char* name, const char* extra) const { 123 st->print("bci: %d", bci()); 124 st->fill_to(tab_width_one); 125 st->print("%s", name); 126 tab(st); 127 int trap = trap_state(); 128 if (trap != 0) { 129 char buf[100]; 130 st->print("trap(%s) ", Deoptimization::format_trap_state(buf, sizeof(buf), trap)); 131 } 132 if (extra != NULL) { 133 st->print("%s", extra); 134 } 135 int flags = data()->flags(); 136 if (flags != 0) { 137 st->print("flags(%d) ", flags); 138 } 139 } 140 141 void ProfileData::tab(outputStream* st, bool first) const { 142 st->fill_to(first ? tab_width_one : tab_width_two); 143 } 144 145 // ================================================================== 146 // BitData 147 // 148 // A BitData corresponds to a one-bit flag. This is used to indicate 149 // whether a checkcast bytecode has seen a null value. 150 151 152 void BitData::print_data_on(outputStream* st, const char* extra) const { 153 print_shared(st, "BitData", extra); 154 st->cr(); 155 } 156 157 // ================================================================== 158 // CounterData 159 // 160 // A CounterData corresponds to a simple counter. 161 162 void CounterData::print_data_on(outputStream* st, const char* extra) const { 163 print_shared(st, "CounterData", extra); 164 st->print_cr("count(%u)", count()); 165 } 166 167 // ================================================================== 168 // JumpData 169 // 170 // A JumpData is used to access profiling information for a direct 171 // branch. It is a counter, used for counting the number of branches, 172 // plus a data displacement, used for realigning the data pointer to 173 // the corresponding target bci. 174 175 void JumpData::post_initialize(BytecodeStream* stream, MethodData* mdo) { 176 assert(stream->bci() == bci(), "wrong pos"); 177 int target; 178 Bytecodes::Code c = stream->code(); 179 if (c == Bytecodes::_goto_w || c == Bytecodes::_jsr_w) { 180 target = stream->dest_w(); 181 } else { 182 target = stream->dest(); 183 } 184 int my_di = mdo->dp_to_di(dp()); 185 int target_di = mdo->bci_to_di(target); 186 int offset = target_di - my_di; 187 set_displacement(offset); 188 } 189 190 void JumpData::print_data_on(outputStream* st, const char* extra) const { 191 print_shared(st, "JumpData", extra); 192 st->print_cr("taken(%u) displacement(%d)", taken(), displacement()); 193 } 194 195 int TypeStackSlotEntries::compute_cell_count(Symbol* signature, bool include_receiver, int max) { 196 // Parameter profiling include the receiver 197 int args_count = include_receiver ? 1 : 0; 198 ResourceMark rm; 199 SignatureStream ss(signature); 200 args_count += ss.reference_parameter_count(); 201 args_count = MIN2(args_count, max); 202 return args_count * per_arg_cell_count; 203 } 204 205 int TypeEntriesAtCall::compute_cell_count(BytecodeStream* stream) { 206 assert(Bytecodes::is_invoke(stream->code()), "should be invoke"); 207 assert(TypeStackSlotEntries::per_arg_count() > ReturnTypeEntry::static_cell_count(), "code to test for arguments/results broken"); 208 const methodHandle m = stream->method(); 209 int bci = stream->bci(); 210 Bytecode_invoke inv(m, bci); 211 int args_cell = 0; 212 if (MethodData::profile_arguments_for_invoke(m, bci)) { 213 args_cell = TypeStackSlotEntries::compute_cell_count(inv.signature(), false, TypeProfileArgsLimit); 214 } 215 int ret_cell = 0; 216 if (MethodData::profile_return_for_invoke(m, bci) && (inv.result_type() == T_OBJECT || inv.result_type() == T_ARRAY)) { 217 ret_cell = ReturnTypeEntry::static_cell_count(); 218 } 219 int header_cell = 0; 220 if (args_cell + ret_cell > 0) { 221 header_cell = header_cell_count(); 222 } 223 224 return header_cell + args_cell + ret_cell; 225 } 226 227 class ArgumentOffsetComputer : public SignatureInfo { 228 private: 229 int _max; 230 GrowableArray<int> _offsets; 231 232 void set(int size, BasicType type) { _size += size; } 233 void do_object(int begin, int end) { 234 if (_offsets.length() < _max) { 235 _offsets.push(_size); 236 } 237 SignatureInfo::do_object(begin, end); 238 } 239 void do_array (int begin, int end) { 240 if (_offsets.length() < _max) { 241 _offsets.push(_size); 242 } 243 SignatureInfo::do_array(begin, end); 244 } 245 246 public: 247 ArgumentOffsetComputer(Symbol* signature, int max) 248 : SignatureInfo(signature), _max(max), _offsets(Thread::current(), max) { 249 } 250 251 int total() { lazy_iterate_parameters(); return _size; } 252 253 int off_at(int i) const { return _offsets.at(i); } 254 }; 255 256 void TypeStackSlotEntries::post_initialize(Symbol* signature, bool has_receiver, bool include_receiver) { 257 ResourceMark rm; 258 int start = 0; 259 // Parameter profiling include the receiver 260 if (include_receiver && has_receiver) { 261 set_stack_slot(0, 0); 262 set_type(0, type_none()); 263 start += 1; 264 } 265 ArgumentOffsetComputer aos(signature, _number_of_entries-start); 266 aos.total(); 267 for (int i = start; i < _number_of_entries; i++) { 268 set_stack_slot(i, aos.off_at(i-start) + (has_receiver ? 1 : 0)); 269 set_type(i, type_none()); 270 } 271 } 272 273 void CallTypeData::post_initialize(BytecodeStream* stream, MethodData* mdo) { 274 assert(Bytecodes::is_invoke(stream->code()), "should be invoke"); 275 Bytecode_invoke inv(stream->method(), stream->bci()); 276 277 SignatureStream ss(inv.signature()); 278 if (has_arguments()) { 279 #ifdef ASSERT 280 ResourceMark rm; 281 int count = MIN2(ss.reference_parameter_count(), (int)TypeProfileArgsLimit); 282 assert(count > 0, "room for args type but none found?"); 283 check_number_of_arguments(count); 284 #endif 285 _args.post_initialize(inv.signature(), inv.has_receiver(), false); 286 } 287 288 if (has_return()) { 289 assert(inv.result_type() == T_OBJECT || inv.result_type() == T_ARRAY, "room for a ret type but doesn't return obj?"); 290 _ret.post_initialize(); 291 } 292 } 293 294 void VirtualCallTypeData::post_initialize(BytecodeStream* stream, MethodData* mdo) { 295 assert(Bytecodes::is_invoke(stream->code()), "should be invoke"); 296 Bytecode_invoke inv(stream->method(), stream->bci()); 297 298 if (has_arguments()) { 299 #ifdef ASSERT 300 ResourceMark rm; 301 SignatureStream ss(inv.signature()); 302 int count = MIN2(ss.reference_parameter_count(), (int)TypeProfileArgsLimit); 303 assert(count > 0, "room for args type but none found?"); 304 check_number_of_arguments(count); 305 #endif 306 _args.post_initialize(inv.signature(), inv.has_receiver(), false); 307 } 308 309 if (has_return()) { 310 assert(inv.result_type() == T_OBJECT || inv.result_type() == T_ARRAY, "room for a ret type but doesn't return obj?"); 311 _ret.post_initialize(); 312 } 313 } 314 315 bool TypeEntries::is_loader_alive(BoolObjectClosure* is_alive_cl, intptr_t p) { 316 Klass* k = (Klass*)klass_part(p); 317 return k != NULL && k->is_loader_alive(is_alive_cl); 318 } 319 320 void TypeStackSlotEntries::clean_weak_klass_links(BoolObjectClosure* is_alive_cl) { 321 for (int i = 0; i < _number_of_entries; i++) { 322 intptr_t p = type(i); 323 if (!is_loader_alive(is_alive_cl, p)) { 324 set_type(i, with_status((Klass*)NULL, p)); 325 } 326 } 327 } 328 329 void ReturnTypeEntry::clean_weak_klass_links(BoolObjectClosure* is_alive_cl) { 330 intptr_t p = type(); 331 if (!is_loader_alive(is_alive_cl, p)) { 332 set_type(with_status((Klass*)NULL, p)); 333 } 334 } 335 336 bool TypeEntriesAtCall::return_profiling_enabled() { 337 return MethodData::profile_return(); 338 } 339 340 bool TypeEntriesAtCall::arguments_profiling_enabled() { 341 return MethodData::profile_arguments(); 342 } 343 344 void TypeEntries::print_klass(outputStream* st, intptr_t k) { 345 if (is_type_none(k)) { 346 st->print("none"); 347 } else if (is_type_unknown(k)) { 348 st->print("unknown"); 349 } else { 350 valid_klass(k)->print_value_on(st); 351 } 352 if (was_null_seen(k)) { 353 st->print(" (null seen)"); 354 } 355 } 356 357 void TypeStackSlotEntries::print_data_on(outputStream* st) const { 358 for (int i = 0; i < _number_of_entries; i++) { 359 _pd->tab(st); 360 st->print("%d: stack(%u) ", i, stack_slot(i)); 361 print_klass(st, type(i)); 362 st->cr(); 363 } 364 } 365 366 void ReturnTypeEntry::print_data_on(outputStream* st) const { 367 _pd->tab(st); 368 print_klass(st, type()); 369 st->cr(); 370 } 371 372 void CallTypeData::print_data_on(outputStream* st, const char* extra) const { 373 CounterData::print_data_on(st, extra); 374 if (has_arguments()) { 375 tab(st, true); 376 st->print("argument types"); 377 _args.print_data_on(st); 378 } 379 if (has_return()) { 380 tab(st, true); 381 st->print("return type"); 382 _ret.print_data_on(st); 383 } 384 } 385 386 void VirtualCallTypeData::print_data_on(outputStream* st, const char* extra) const { 387 VirtualCallData::print_data_on(st, extra); 388 if (has_arguments()) { 389 tab(st, true); 390 st->print("argument types"); 391 _args.print_data_on(st); 392 } 393 if (has_return()) { 394 tab(st, true); 395 st->print("return type"); 396 _ret.print_data_on(st); 397 } 398 } 399 400 // ================================================================== 401 // ReceiverTypeData 402 // 403 // A ReceiverTypeData is used to access profiling information about a 404 // dynamic type check. It consists of a counter which counts the total times 405 // that the check is reached, and a series of (Klass*, count) pairs 406 // which are used to store a type profile for the receiver of the check. 407 408 void ReceiverTypeData::clean_weak_klass_links(BoolObjectClosure* is_alive_cl) { 409 for (uint row = 0; row < row_limit(); row++) { 410 Klass* p = receiver(row); 411 if (p != NULL && !p->is_loader_alive(is_alive_cl)) { 412 clear_row(row); 413 } 414 } 415 } 416 417 #if INCLUDE_JVMCI 418 void VirtualCallData::clean_weak_klass_links(BoolObjectClosure* is_alive_cl) { 419 ReceiverTypeData::clean_weak_klass_links(is_alive_cl); 420 for (uint row = 0; row < method_row_limit(); row++) { 421 Method* p = method(row); 422 if (p != NULL && !p->method_holder()->is_loader_alive(is_alive_cl)) { 423 clear_method_row(row); 424 } 425 } 426 } 427 428 void VirtualCallData::clean_weak_method_links() { 429 ReceiverTypeData::clean_weak_method_links(); 430 for (uint row = 0; row < method_row_limit(); row++) { 431 Method* p = method(row); 432 if (p != NULL && !p->on_stack()) { 433 clear_method_row(row); 434 } 435 } 436 } 437 #endif // INCLUDE_JVMCI 438 439 void ReceiverTypeData::print_receiver_data_on(outputStream* st) const { 440 uint row; 441 int entries = 0; 442 for (row = 0; row < row_limit(); row++) { 443 if (receiver(row) != NULL) entries++; 444 } 445 #if INCLUDE_JVMCI 446 st->print_cr("count(%u) nonprofiled_count(%u) entries(%u)", count(), nonprofiled_count(), entries); 447 #else 448 st->print_cr("count(%u) entries(%u)", count(), entries); 449 #endif 450 int total = count(); 451 for (row = 0; row < row_limit(); row++) { 452 if (receiver(row) != NULL) { 453 total += receiver_count(row); 454 } 455 } 456 for (row = 0; row < row_limit(); row++) { 457 if (receiver(row) != NULL) { 458 tab(st); 459 receiver(row)->print_value_on(st); 460 st->print_cr("(%u %4.2f)", receiver_count(row), (float) receiver_count(row) / (float) total); 461 } 462 } 463 } 464 void ReceiverTypeData::print_data_on(outputStream* st, const char* extra) const { 465 print_shared(st, "ReceiverTypeData", extra); 466 print_receiver_data_on(st); 467 } 468 469 #if INCLUDE_JVMCI 470 void VirtualCallData::print_method_data_on(outputStream* st) const { 471 uint row; 472 int entries = 0; 473 for (row = 0; row < method_row_limit(); row++) { 474 if (method(row) != NULL) entries++; 475 } 476 tab(st); 477 st->print_cr("method_entries(%u)", entries); 478 int total = count(); 479 for (row = 0; row < method_row_limit(); row++) { 480 if (method(row) != NULL) { 481 total += method_count(row); 482 } 483 } 484 for (row = 0; row < method_row_limit(); row++) { 485 if (method(row) != NULL) { 486 tab(st); 487 method(row)->print_value_on(st); 488 st->print_cr("(%u %4.2f)", method_count(row), (float) method_count(row) / (float) total); 489 } 490 } 491 } 492 #endif // INCLUDE_JVMCI 493 494 void VirtualCallData::print_data_on(outputStream* st, const char* extra) const { 495 print_shared(st, "VirtualCallData", extra); 496 print_receiver_data_on(st); 497 print_method_data_on(st); 498 } 499 500 // ================================================================== 501 // RetData 502 // 503 // A RetData is used to access profiling information for a ret bytecode. 504 // It is composed of a count of the number of times that the ret has 505 // been executed, followed by a series of triples of the form 506 // (bci, count, di) which count the number of times that some bci was the 507 // target of the ret and cache a corresponding displacement. 508 509 void RetData::post_initialize(BytecodeStream* stream, MethodData* mdo) { 510 for (uint row = 0; row < row_limit(); row++) { 511 set_bci_displacement(row, -1); 512 set_bci(row, no_bci); 513 } 514 // release so other threads see a consistent state. bci is used as 515 // a valid flag for bci_displacement. 516 OrderAccess::release(); 517 } 518 519 // This routine needs to atomically update the RetData structure, so the 520 // caller needs to hold the RetData_lock before it gets here. Since taking 521 // the lock can block (and allow GC) and since RetData is a ProfileData is a 522 // wrapper around a derived oop, taking the lock in _this_ method will 523 // basically cause the 'this' pointer's _data field to contain junk after the 524 // lock. We require the caller to take the lock before making the ProfileData 525 // structure. Currently the only caller is InterpreterRuntime::update_mdp_for_ret 526 address RetData::fixup_ret(int return_bci, MethodData* h_mdo) { 527 // First find the mdp which corresponds to the return bci. 528 address mdp = h_mdo->bci_to_dp(return_bci); 529 530 // Now check to see if any of the cache slots are open. 531 for (uint row = 0; row < row_limit(); row++) { 532 if (bci(row) == no_bci) { 533 set_bci_displacement(row, mdp - dp()); 534 set_bci_count(row, DataLayout::counter_increment); 535 // Barrier to ensure displacement is written before the bci; allows 536 // the interpreter to read displacement without fear of race condition. 537 release_set_bci(row, return_bci); 538 break; 539 } 540 } 541 return mdp; 542 } 543 544 #ifdef CC_INTERP 545 DataLayout* RetData::advance(MethodData *md, int bci) { 546 return (DataLayout*) md->bci_to_dp(bci); 547 } 548 #endif // CC_INTERP 549 550 void RetData::print_data_on(outputStream* st, const char* extra) const { 551 print_shared(st, "RetData", extra); 552 uint row; 553 int entries = 0; 554 for (row = 0; row < row_limit(); row++) { 555 if (bci(row) != no_bci) entries++; 556 } 557 st->print_cr("count(%u) entries(%u)", count(), entries); 558 for (row = 0; row < row_limit(); row++) { 559 if (bci(row) != no_bci) { 560 tab(st); 561 st->print_cr("bci(%d: count(%u) displacement(%d))", 562 bci(row), bci_count(row), bci_displacement(row)); 563 } 564 } 565 } 566 567 // ================================================================== 568 // BranchData 569 // 570 // A BranchData is used to access profiling data for a two-way branch. 571 // It consists of taken and not_taken counts as well as a data displacement 572 // for the taken case. 573 574 void BranchData::post_initialize(BytecodeStream* stream, MethodData* mdo) { 575 assert(stream->bci() == bci(), "wrong pos"); 576 int target = stream->dest(); 577 int my_di = mdo->dp_to_di(dp()); 578 int target_di = mdo->bci_to_di(target); 579 int offset = target_di - my_di; 580 set_displacement(offset); 581 } 582 583 void BranchData::print_data_on(outputStream* st, const char* extra) const { 584 print_shared(st, "BranchData", extra); 585 st->print_cr("taken(%u) displacement(%d)", 586 taken(), displacement()); 587 tab(st); 588 st->print_cr("not taken(%u)", not_taken()); 589 } 590 591 // ================================================================== 592 // MultiBranchData 593 // 594 // A MultiBranchData is used to access profiling information for 595 // a multi-way branch (*switch bytecodes). It consists of a series 596 // of (count, displacement) pairs, which count the number of times each 597 // case was taken and specify the data displacment for each branch target. 598 599 int MultiBranchData::compute_cell_count(BytecodeStream* stream) { 600 int cell_count = 0; 601 if (stream->code() == Bytecodes::_tableswitch) { 602 Bytecode_tableswitch sw(stream->method()(), stream->bcp()); 603 cell_count = 1 + per_case_cell_count * (1 + sw.length()); // 1 for default 604 } else { 605 Bytecode_lookupswitch sw(stream->method()(), stream->bcp()); 606 cell_count = 1 + per_case_cell_count * (sw.number_of_pairs() + 1); // 1 for default 607 } 608 return cell_count; 609 } 610 611 void MultiBranchData::post_initialize(BytecodeStream* stream, 612 MethodData* mdo) { 613 assert(stream->bci() == bci(), "wrong pos"); 614 int target; 615 int my_di; 616 int target_di; 617 int offset; 618 if (stream->code() == Bytecodes::_tableswitch) { 619 Bytecode_tableswitch sw(stream->method()(), stream->bcp()); 620 int len = sw.length(); 621 assert(array_len() == per_case_cell_count * (len + 1), "wrong len"); 622 for (int count = 0; count < len; count++) { 623 target = sw.dest_offset_at(count) + bci(); 624 my_di = mdo->dp_to_di(dp()); 625 target_di = mdo->bci_to_di(target); 626 offset = target_di - my_di; 627 set_displacement_at(count, offset); 628 } 629 target = sw.default_offset() + bci(); 630 my_di = mdo->dp_to_di(dp()); 631 target_di = mdo->bci_to_di(target); 632 offset = target_di - my_di; 633 set_default_displacement(offset); 634 635 } else { 636 Bytecode_lookupswitch sw(stream->method()(), stream->bcp()); 637 int npairs = sw.number_of_pairs(); 638 assert(array_len() == per_case_cell_count * (npairs + 1), "wrong len"); 639 for (int count = 0; count < npairs; count++) { 640 LookupswitchPair pair = sw.pair_at(count); 641 target = pair.offset() + bci(); 642 my_di = mdo->dp_to_di(dp()); 643 target_di = mdo->bci_to_di(target); 644 offset = target_di - my_di; 645 set_displacement_at(count, offset); 646 } 647 target = sw.default_offset() + bci(); 648 my_di = mdo->dp_to_di(dp()); 649 target_di = mdo->bci_to_di(target); 650 offset = target_di - my_di; 651 set_default_displacement(offset); 652 } 653 } 654 655 void MultiBranchData::print_data_on(outputStream* st, const char* extra) const { 656 print_shared(st, "MultiBranchData", extra); 657 st->print_cr("default_count(%u) displacement(%d)", 658 default_count(), default_displacement()); 659 int cases = number_of_cases(); 660 for (int i = 0; i < cases; i++) { 661 tab(st); 662 st->print_cr("count(%u) displacement(%d)", 663 count_at(i), displacement_at(i)); 664 } 665 } 666 667 void ArgInfoData::print_data_on(outputStream* st, const char* extra) const { 668 print_shared(st, "ArgInfoData", extra); 669 int nargs = number_of_args(); 670 for (int i = 0; i < nargs; i++) { 671 st->print(" 0x%x", arg_modified(i)); 672 } 673 st->cr(); 674 } 675 676 int ParametersTypeData::compute_cell_count(Method* m) { 677 if (!MethodData::profile_parameters_for_method(m)) { 678 return 0; 679 } 680 int max = TypeProfileParmsLimit == -1 ? INT_MAX : TypeProfileParmsLimit; 681 int obj_args = TypeStackSlotEntries::compute_cell_count(m->signature(), !m->is_static(), max); 682 if (obj_args > 0) { 683 return obj_args + 1; // 1 cell for array len 684 } 685 return 0; 686 } 687 688 void ParametersTypeData::post_initialize(BytecodeStream* stream, MethodData* mdo) { 689 _parameters.post_initialize(mdo->method()->signature(), !mdo->method()->is_static(), true); 690 } 691 692 bool ParametersTypeData::profiling_enabled() { 693 return MethodData::profile_parameters(); 694 } 695 696 void ParametersTypeData::print_data_on(outputStream* st, const char* extra) const { 697 st->print("parameter types"); // FIXME extra ignored? 698 _parameters.print_data_on(st); 699 } 700 701 void SpeculativeTrapData::print_data_on(outputStream* st, const char* extra) const { 702 print_shared(st, "SpeculativeTrapData", extra); 703 tab(st); 704 method()->print_short_name(st); 705 st->cr(); 706 } 707 708 // ================================================================== 709 // MethodData* 710 // 711 // A MethodData* holds information which has been collected about 712 // a method. 713 714 MethodData* MethodData::allocate(ClassLoaderData* loader_data, const methodHandle& method, TRAPS) { 715 int size = MethodData::compute_allocation_size_in_words(method); 716 717 return new (loader_data, size, false, MetaspaceObj::MethodDataType, THREAD) 718 MethodData(method(), size, THREAD); 719 } 720 721 int MethodData::bytecode_cell_count(Bytecodes::Code code) { 722 if (is_client_compilation_mode_vm()) { 723 return no_profile_data; 724 } 725 switch (code) { 726 case Bytecodes::_checkcast: 727 case Bytecodes::_instanceof: 728 case Bytecodes::_aastore: 729 if (TypeProfileCasts) { 730 return ReceiverTypeData::static_cell_count(); 731 } else { 732 return BitData::static_cell_count(); 733 } 734 case Bytecodes::_invokespecial: 735 case Bytecodes::_invokestatic: 736 if (MethodData::profile_arguments() || MethodData::profile_return()) { 737 return variable_cell_count; 738 } else { 739 return CounterData::static_cell_count(); 740 } 741 case Bytecodes::_goto: 742 case Bytecodes::_goto_w: 743 case Bytecodes::_jsr: 744 case Bytecodes::_jsr_w: 745 return JumpData::static_cell_count(); 746 case Bytecodes::_invokevirtual: 747 case Bytecodes::_invokeinterface: 748 if (MethodData::profile_arguments() || MethodData::profile_return()) { 749 return variable_cell_count; 750 } else { 751 return VirtualCallData::static_cell_count(); 752 } 753 case Bytecodes::_invokedynamic: 754 if (MethodData::profile_arguments() || MethodData::profile_return()) { 755 return variable_cell_count; 756 } else { 757 return CounterData::static_cell_count(); 758 } 759 case Bytecodes::_ret: 760 return RetData::static_cell_count(); 761 case Bytecodes::_ifeq: 762 case Bytecodes::_ifne: 763 case Bytecodes::_iflt: 764 case Bytecodes::_ifge: 765 case Bytecodes::_ifgt: 766 case Bytecodes::_ifle: 767 case Bytecodes::_if_icmpeq: 768 case Bytecodes::_if_icmpne: 769 case Bytecodes::_if_icmplt: 770 case Bytecodes::_if_icmpge: 771 case Bytecodes::_if_icmpgt: 772 case Bytecodes::_if_icmple: 773 case Bytecodes::_if_acmpeq: 774 case Bytecodes::_if_acmpne: 775 case Bytecodes::_ifnull: 776 case Bytecodes::_ifnonnull: 777 return BranchData::static_cell_count(); 778 case Bytecodes::_lookupswitch: 779 case Bytecodes::_tableswitch: 780 return variable_cell_count; 781 default: 782 return no_profile_data; 783 } 784 } 785 786 // Compute the size of the profiling information corresponding to 787 // the current bytecode. 788 int MethodData::compute_data_size(BytecodeStream* stream) { 789 int cell_count = bytecode_cell_count(stream->code()); 790 if (cell_count == no_profile_data) { 791 return 0; 792 } 793 if (cell_count == variable_cell_count) { 794 switch (stream->code()) { 795 case Bytecodes::_lookupswitch: 796 case Bytecodes::_tableswitch: 797 cell_count = MultiBranchData::compute_cell_count(stream); 798 break; 799 case Bytecodes::_invokespecial: 800 case Bytecodes::_invokestatic: 801 case Bytecodes::_invokedynamic: 802 assert(MethodData::profile_arguments() || MethodData::profile_return(), "should be collecting args profile"); 803 if (profile_arguments_for_invoke(stream->method(), stream->bci()) || 804 profile_return_for_invoke(stream->method(), stream->bci())) { 805 cell_count = CallTypeData::compute_cell_count(stream); 806 } else { 807 cell_count = CounterData::static_cell_count(); 808 } 809 break; 810 case Bytecodes::_invokevirtual: 811 case Bytecodes::_invokeinterface: { 812 assert(MethodData::profile_arguments() || MethodData::profile_return(), "should be collecting args profile"); 813 if (profile_arguments_for_invoke(stream->method(), stream->bci()) || 814 profile_return_for_invoke(stream->method(), stream->bci())) { 815 cell_count = VirtualCallTypeData::compute_cell_count(stream); 816 } else { 817 cell_count = VirtualCallData::static_cell_count(); 818 } 819 break; 820 } 821 default: 822 fatal("unexpected bytecode for var length profile data"); 823 } 824 } 825 // Note: cell_count might be zero, meaning that there is just 826 // a DataLayout header, with no extra cells. 827 assert(cell_count >= 0, "sanity"); 828 return DataLayout::compute_size_in_bytes(cell_count); 829 } 830 831 bool MethodData::is_speculative_trap_bytecode(Bytecodes::Code code) { 832 // Bytecodes for which we may use speculation 833 switch (code) { 834 case Bytecodes::_checkcast: 835 case Bytecodes::_instanceof: 836 case Bytecodes::_aastore: 837 case Bytecodes::_invokevirtual: 838 case Bytecodes::_invokeinterface: 839 case Bytecodes::_if_acmpeq: 840 case Bytecodes::_if_acmpne: 841 case Bytecodes::_ifnull: 842 case Bytecodes::_ifnonnull: 843 case Bytecodes::_invokestatic: 844 #ifdef COMPILER2 845 if (is_server_compilation_mode_vm()) { 846 return UseTypeSpeculation; 847 } 848 #endif 849 default: 850 return false; 851 } 852 return false; 853 } 854 855 int MethodData::compute_extra_data_count(int data_size, int empty_bc_count, bool needs_speculative_traps) { 856 #if INCLUDE_JVMCI 857 if (ProfileTraps) { 858 // Assume that up to 30% of the possibly trapping BCIs with no MDP will need to allocate one. 859 int extra_data_count = MIN2(empty_bc_count, MAX2(4, (empty_bc_count * 30) / 100)); 860 861 // Make sure we have a minimum number of extra data slots to 862 // allocate SpeculativeTrapData entries. We would want to have one 863 // entry per compilation that inlines this method and for which 864 // some type speculation assumption fails. So the room we need for 865 // the SpeculativeTrapData entries doesn't directly depend on the 866 // size of the method. Because it's hard to estimate, we reserve 867 // space for an arbitrary number of entries. 868 int spec_data_count = (needs_speculative_traps ? SpecTrapLimitExtraEntries : 0) * 869 (SpeculativeTrapData::static_cell_count() + DataLayout::header_size_in_cells()); 870 871 return MAX2(extra_data_count, spec_data_count); 872 } else { 873 return 0; 874 } 875 #else // INCLUDE_JVMCI 876 if (ProfileTraps) { 877 // Assume that up to 3% of BCIs with no MDP will need to allocate one. 878 int extra_data_count = (uint)(empty_bc_count * 3) / 128 + 1; 879 // If the method is large, let the extra BCIs grow numerous (to ~1%). 880 int one_percent_of_data 881 = (uint)data_size / (DataLayout::header_size_in_bytes()*128); 882 if (extra_data_count < one_percent_of_data) 883 extra_data_count = one_percent_of_data; 884 if (extra_data_count > empty_bc_count) 885 extra_data_count = empty_bc_count; // no need for more 886 887 // Make sure we have a minimum number of extra data slots to 888 // allocate SpeculativeTrapData entries. We would want to have one 889 // entry per compilation that inlines this method and for which 890 // some type speculation assumption fails. So the room we need for 891 // the SpeculativeTrapData entries doesn't directly depend on the 892 // size of the method. Because it's hard to estimate, we reserve 893 // space for an arbitrary number of entries. 894 int spec_data_count = (needs_speculative_traps ? SpecTrapLimitExtraEntries : 0) * 895 (SpeculativeTrapData::static_cell_count() + DataLayout::header_size_in_cells()); 896 897 return MAX2(extra_data_count, spec_data_count); 898 } else { 899 return 0; 900 } 901 #endif // INCLUDE_JVMCI 902 } 903 904 // Compute the size of the MethodData* necessary to store 905 // profiling information about a given method. Size is in bytes. 906 int MethodData::compute_allocation_size_in_bytes(const methodHandle& method) { 907 int data_size = 0; 908 BytecodeStream stream(method); 909 Bytecodes::Code c; 910 int empty_bc_count = 0; // number of bytecodes lacking data 911 bool needs_speculative_traps = false; 912 while ((c = stream.next()) >= 0) { 913 int size_in_bytes = compute_data_size(&stream); 914 data_size += size_in_bytes; 915 if (size_in_bytes == 0 JVMCI_ONLY(&& Bytecodes::can_trap(c))) empty_bc_count += 1; 916 needs_speculative_traps = needs_speculative_traps || is_speculative_trap_bytecode(c); 917 } 918 int object_size = in_bytes(data_offset()) + data_size; 919 920 // Add some extra DataLayout cells (at least one) to track stray traps. 921 int extra_data_count = compute_extra_data_count(data_size, empty_bc_count, needs_speculative_traps); 922 object_size += extra_data_count * DataLayout::compute_size_in_bytes(0); 923 924 // Add a cell to record information about modified arguments. 925 int arg_size = method->size_of_parameters(); 926 object_size += DataLayout::compute_size_in_bytes(arg_size+1); 927 928 // Reserve room for an area of the MDO dedicated to profiling of 929 // parameters 930 int args_cell = ParametersTypeData::compute_cell_count(method()); 931 if (args_cell > 0) { 932 object_size += DataLayout::compute_size_in_bytes(args_cell); 933 } 934 return object_size; 935 } 936 937 // Compute the size of the MethodData* necessary to store 938 // profiling information about a given method. Size is in words 939 int MethodData::compute_allocation_size_in_words(const methodHandle& method) { 940 int byte_size = compute_allocation_size_in_bytes(method); 941 int word_size = align_size_up(byte_size, BytesPerWord) / BytesPerWord; 942 return align_metadata_size(word_size); 943 } 944 945 // Initialize an individual data segment. Returns the size of 946 // the segment in bytes. 947 int MethodData::initialize_data(BytecodeStream* stream, 948 int data_index) { 949 if (is_client_compilation_mode_vm()) { 950 return 0; 951 } 952 int cell_count = -1; 953 int tag = DataLayout::no_tag; 954 DataLayout* data_layout = data_layout_at(data_index); 955 Bytecodes::Code c = stream->code(); 956 switch (c) { 957 case Bytecodes::_checkcast: 958 case Bytecodes::_instanceof: 959 case Bytecodes::_aastore: 960 if (TypeProfileCasts) { 961 cell_count = ReceiverTypeData::static_cell_count(); 962 tag = DataLayout::receiver_type_data_tag; 963 } else { 964 cell_count = BitData::static_cell_count(); 965 tag = DataLayout::bit_data_tag; 966 } 967 break; 968 case Bytecodes::_invokespecial: 969 case Bytecodes::_invokestatic: { 970 int counter_data_cell_count = CounterData::static_cell_count(); 971 if (profile_arguments_for_invoke(stream->method(), stream->bci()) || 972 profile_return_for_invoke(stream->method(), stream->bci())) { 973 cell_count = CallTypeData::compute_cell_count(stream); 974 } else { 975 cell_count = counter_data_cell_count; 976 } 977 if (cell_count > counter_data_cell_count) { 978 tag = DataLayout::call_type_data_tag; 979 } else { 980 tag = DataLayout::counter_data_tag; 981 } 982 break; 983 } 984 case Bytecodes::_goto: 985 case Bytecodes::_goto_w: 986 case Bytecodes::_jsr: 987 case Bytecodes::_jsr_w: 988 cell_count = JumpData::static_cell_count(); 989 tag = DataLayout::jump_data_tag; 990 break; 991 case Bytecodes::_invokevirtual: 992 case Bytecodes::_invokeinterface: { 993 int virtual_call_data_cell_count = VirtualCallData::static_cell_count(); 994 if (profile_arguments_for_invoke(stream->method(), stream->bci()) || 995 profile_return_for_invoke(stream->method(), stream->bci())) { 996 cell_count = VirtualCallTypeData::compute_cell_count(stream); 997 } else { 998 cell_count = virtual_call_data_cell_count; 999 } 1000 if (cell_count > virtual_call_data_cell_count) { 1001 tag = DataLayout::virtual_call_type_data_tag; 1002 } else { 1003 tag = DataLayout::virtual_call_data_tag; 1004 } 1005 break; 1006 } 1007 case Bytecodes::_invokedynamic: { 1008 // %%% should make a type profile for any invokedynamic that takes a ref argument 1009 int counter_data_cell_count = CounterData::static_cell_count(); 1010 if (profile_arguments_for_invoke(stream->method(), stream->bci()) || 1011 profile_return_for_invoke(stream->method(), stream->bci())) { 1012 cell_count = CallTypeData::compute_cell_count(stream); 1013 } else { 1014 cell_count = counter_data_cell_count; 1015 } 1016 if (cell_count > counter_data_cell_count) { 1017 tag = DataLayout::call_type_data_tag; 1018 } else { 1019 tag = DataLayout::counter_data_tag; 1020 } 1021 break; 1022 } 1023 case Bytecodes::_ret: 1024 cell_count = RetData::static_cell_count(); 1025 tag = DataLayout::ret_data_tag; 1026 break; 1027 case Bytecodes::_ifeq: 1028 case Bytecodes::_ifne: 1029 case Bytecodes::_iflt: 1030 case Bytecodes::_ifge: 1031 case Bytecodes::_ifgt: 1032 case Bytecodes::_ifle: 1033 case Bytecodes::_if_icmpeq: 1034 case Bytecodes::_if_icmpne: 1035 case Bytecodes::_if_icmplt: 1036 case Bytecodes::_if_icmpge: 1037 case Bytecodes::_if_icmpgt: 1038 case Bytecodes::_if_icmple: 1039 case Bytecodes::_if_acmpeq: 1040 case Bytecodes::_if_acmpne: 1041 case Bytecodes::_ifnull: 1042 case Bytecodes::_ifnonnull: 1043 cell_count = BranchData::static_cell_count(); 1044 tag = DataLayout::branch_data_tag; 1045 break; 1046 case Bytecodes::_lookupswitch: 1047 case Bytecodes::_tableswitch: 1048 cell_count = MultiBranchData::compute_cell_count(stream); 1049 tag = DataLayout::multi_branch_data_tag; 1050 break; 1051 default: 1052 break; 1053 } 1054 assert(tag == DataLayout::multi_branch_data_tag || 1055 ((MethodData::profile_arguments() || MethodData::profile_return()) && 1056 (tag == DataLayout::call_type_data_tag || 1057 tag == DataLayout::counter_data_tag || 1058 tag == DataLayout::virtual_call_type_data_tag || 1059 tag == DataLayout::virtual_call_data_tag)) || 1060 cell_count == bytecode_cell_count(c), "cell counts must agree"); 1061 if (cell_count >= 0) { 1062 assert(tag != DataLayout::no_tag, "bad tag"); 1063 assert(bytecode_has_profile(c), "agree w/ BHP"); 1064 data_layout->initialize(tag, stream->bci(), cell_count); 1065 return DataLayout::compute_size_in_bytes(cell_count); 1066 } else { 1067 assert(!bytecode_has_profile(c), "agree w/ !BHP"); 1068 return 0; 1069 } 1070 } 1071 1072 // Get the data at an arbitrary (sort of) data index. 1073 ProfileData* MethodData::data_at(int data_index) const { 1074 if (out_of_bounds(data_index)) { 1075 return NULL; 1076 } 1077 DataLayout* data_layout = data_layout_at(data_index); 1078 return data_layout->data_in(); 1079 } 1080 1081 ProfileData* DataLayout::data_in() { 1082 switch (tag()) { 1083 case DataLayout::no_tag: 1084 default: 1085 ShouldNotReachHere(); 1086 return NULL; 1087 case DataLayout::bit_data_tag: 1088 return new BitData(this); 1089 case DataLayout::counter_data_tag: 1090 return new CounterData(this); 1091 case DataLayout::jump_data_tag: 1092 return new JumpData(this); 1093 case DataLayout::receiver_type_data_tag: 1094 return new ReceiverTypeData(this); 1095 case DataLayout::virtual_call_data_tag: 1096 return new VirtualCallData(this); 1097 case DataLayout::ret_data_tag: 1098 return new RetData(this); 1099 case DataLayout::branch_data_tag: 1100 return new BranchData(this); 1101 case DataLayout::multi_branch_data_tag: 1102 return new MultiBranchData(this); 1103 case DataLayout::arg_info_data_tag: 1104 return new ArgInfoData(this); 1105 case DataLayout::call_type_data_tag: 1106 return new CallTypeData(this); 1107 case DataLayout::virtual_call_type_data_tag: 1108 return new VirtualCallTypeData(this); 1109 case DataLayout::parameters_type_data_tag: 1110 return new ParametersTypeData(this); 1111 case DataLayout::speculative_trap_data_tag: 1112 return new SpeculativeTrapData(this); 1113 } 1114 } 1115 1116 // Iteration over data. 1117 ProfileData* MethodData::next_data(ProfileData* current) const { 1118 int current_index = dp_to_di(current->dp()); 1119 int next_index = current_index + current->size_in_bytes(); 1120 ProfileData* next = data_at(next_index); 1121 return next; 1122 } 1123 1124 // Give each of the data entries a chance to perform specific 1125 // data initialization. 1126 void MethodData::post_initialize(BytecodeStream* stream) { 1127 ResourceMark rm; 1128 ProfileData* data; 1129 for (data = first_data(); is_valid(data); data = next_data(data)) { 1130 stream->set_start(data->bci()); 1131 stream->next(); 1132 data->post_initialize(stream, this); 1133 } 1134 if (_parameters_type_data_di != no_parameters) { 1135 parameters_type_data()->post_initialize(NULL, this); 1136 } 1137 } 1138 1139 // Initialize the MethodData* corresponding to a given method. 1140 MethodData::MethodData(const methodHandle& method, int size, TRAPS) 1141 : _extra_data_lock(Monitor::leaf, "MDO extra data lock"), 1142 _parameters_type_data_di(parameters_uninitialized) { 1143 // Set the method back-pointer. 1144 _method = method(); 1145 initialize(); 1146 } 1147 1148 void MethodData::initialize() { 1149 NoSafepointVerifier no_safepoint; // init function atomic wrt GC 1150 ResourceMark rm; 1151 1152 init(); 1153 set_creation_mileage(mileage_of(method())); 1154 1155 // Go through the bytecodes and allocate and initialize the 1156 // corresponding data cells. 1157 int data_size = 0; 1158 int empty_bc_count = 0; // number of bytecodes lacking data 1159 _data[0] = 0; // apparently not set below. 1160 BytecodeStream stream(method()); 1161 Bytecodes::Code c; 1162 bool needs_speculative_traps = false; 1163 while ((c = stream.next()) >= 0) { 1164 int size_in_bytes = initialize_data(&stream, data_size); 1165 data_size += size_in_bytes; 1166 if (size_in_bytes == 0 JVMCI_ONLY(&& Bytecodes::can_trap(c))) empty_bc_count += 1; 1167 needs_speculative_traps = needs_speculative_traps || is_speculative_trap_bytecode(c); 1168 } 1169 _data_size = data_size; 1170 int object_size = in_bytes(data_offset()) + data_size; 1171 1172 // Add some extra DataLayout cells (at least one) to track stray traps. 1173 int extra_data_count = compute_extra_data_count(data_size, empty_bc_count, needs_speculative_traps); 1174 int extra_size = extra_data_count * DataLayout::compute_size_in_bytes(0); 1175 1176 // Let's zero the space for the extra data 1177 Copy::zero_to_bytes(((address)_data) + data_size, extra_size); 1178 1179 // Add a cell to record information about modified arguments. 1180 // Set up _args_modified array after traps cells so that 1181 // the code for traps cells works. 1182 DataLayout *dp = data_layout_at(data_size + extra_size); 1183 1184 int arg_size = method()->size_of_parameters(); 1185 dp->initialize(DataLayout::arg_info_data_tag, 0, arg_size+1); 1186 1187 int arg_data_size = DataLayout::compute_size_in_bytes(arg_size+1); 1188 object_size += extra_size + arg_data_size; 1189 1190 int parms_cell = ParametersTypeData::compute_cell_count(method()); 1191 // If we are profiling parameters, we reserver an area near the end 1192 // of the MDO after the slots for bytecodes (because there's no bci 1193 // for method entry so they don't fit with the framework for the 1194 // profiling of bytecodes). We store the offset within the MDO of 1195 // this area (or -1 if no parameter is profiled) 1196 if (parms_cell > 0) { 1197 object_size += DataLayout::compute_size_in_bytes(parms_cell); 1198 _parameters_type_data_di = data_size + extra_size + arg_data_size; 1199 DataLayout *dp = data_layout_at(data_size + extra_size + arg_data_size); 1200 dp->initialize(DataLayout::parameters_type_data_tag, 0, parms_cell); 1201 } else { 1202 _parameters_type_data_di = no_parameters; 1203 } 1204 1205 // Set an initial hint. Don't use set_hint_di() because 1206 // first_di() may be out of bounds if data_size is 0. 1207 // In that situation, _hint_di is never used, but at 1208 // least well-defined. 1209 _hint_di = first_di(); 1210 1211 post_initialize(&stream); 1212 1213 assert(object_size == compute_allocation_size_in_bytes(methodHandle(_method)), "MethodData: computed size != initialized size"); 1214 set_size(object_size); 1215 } 1216 1217 void MethodData::init() { 1218 _invocation_counter.init(); 1219 _backedge_counter.init(); 1220 _invocation_counter_start = 0; 1221 _backedge_counter_start = 0; 1222 1223 // Set per-method invoke- and backedge mask. 1224 double scale = 1.0; 1225 CompilerOracle::has_option_value(_method, "CompileThresholdScaling", scale); 1226 _invoke_mask = right_n_bits(Arguments::scaled_freq_log(Tier0InvokeNotifyFreqLog, scale)) << InvocationCounter::count_shift; 1227 _backedge_mask = right_n_bits(Arguments::scaled_freq_log(Tier0BackedgeNotifyFreqLog, scale)) << InvocationCounter::count_shift; 1228 1229 _tenure_traps = 0; 1230 _num_loops = 0; 1231 _num_blocks = 0; 1232 _would_profile = unknown; 1233 1234 #if INCLUDE_JVMCI 1235 _jvmci_ir_size = 0; 1236 #endif 1237 1238 #if INCLUDE_RTM_OPT 1239 _rtm_state = NoRTM; // No RTM lock eliding by default 1240 if (UseRTMLocking && 1241 !CompilerOracle::has_option_string(_method, "NoRTMLockEliding")) { 1242 if (CompilerOracle::has_option_string(_method, "UseRTMLockEliding") || !UseRTMDeopt) { 1243 // Generate RTM lock eliding code without abort ratio calculation code. 1244 _rtm_state = UseRTM; 1245 } else if (UseRTMDeopt) { 1246 // Generate RTM lock eliding code and include abort ratio calculation 1247 // code if UseRTMDeopt is on. 1248 _rtm_state = ProfileRTM; 1249 } 1250 } 1251 #endif 1252 1253 // Initialize flags and trap history. 1254 _nof_decompiles = 0; 1255 _nof_overflow_recompiles = 0; 1256 _nof_overflow_traps = 0; 1257 clear_escape_info(); 1258 assert(sizeof(_trap_hist) % sizeof(HeapWord) == 0, "align"); 1259 Copy::zero_to_words((HeapWord*) &_trap_hist, 1260 sizeof(_trap_hist) / sizeof(HeapWord)); 1261 } 1262 1263 // Get a measure of how much mileage the method has on it. 1264 int MethodData::mileage_of(Method* method) { 1265 int mileage = 0; 1266 if (TieredCompilation) { 1267 mileage = MAX2(method->invocation_count(), method->backedge_count()); 1268 } else { 1269 int iic = method->interpreter_invocation_count(); 1270 if (mileage < iic) mileage = iic; 1271 MethodCounters* mcs = method->method_counters(); 1272 if (mcs != NULL) { 1273 InvocationCounter* ic = mcs->invocation_counter(); 1274 InvocationCounter* bc = mcs->backedge_counter(); 1275 int icval = ic->count(); 1276 if (ic->carry()) icval += CompileThreshold; 1277 if (mileage < icval) mileage = icval; 1278 int bcval = bc->count(); 1279 if (bc->carry()) bcval += CompileThreshold; 1280 if (mileage < bcval) mileage = bcval; 1281 } 1282 } 1283 return mileage; 1284 } 1285 1286 bool MethodData::is_mature() const { 1287 return CompilationPolicy::policy()->is_mature(_method); 1288 } 1289 1290 // Translate a bci to its corresponding data index (di). 1291 address MethodData::bci_to_dp(int bci) { 1292 ResourceMark rm; 1293 ProfileData* data = data_before(bci); 1294 ProfileData* prev = NULL; 1295 for ( ; is_valid(data); data = next_data(data)) { 1296 if (data->bci() >= bci) { 1297 if (data->bci() == bci) set_hint_di(dp_to_di(data->dp())); 1298 else if (prev != NULL) set_hint_di(dp_to_di(prev->dp())); 1299 return data->dp(); 1300 } 1301 prev = data; 1302 } 1303 return (address)limit_data_position(); 1304 } 1305 1306 // Translate a bci to its corresponding data, or NULL. 1307 ProfileData* MethodData::bci_to_data(int bci) { 1308 ProfileData* data = data_before(bci); 1309 for ( ; is_valid(data); data = next_data(data)) { 1310 if (data->bci() == bci) { 1311 set_hint_di(dp_to_di(data->dp())); 1312 return data; 1313 } else if (data->bci() > bci) { 1314 break; 1315 } 1316 } 1317 return bci_to_extra_data(bci, NULL, false); 1318 } 1319 1320 DataLayout* MethodData::next_extra(DataLayout* dp) { 1321 int nb_cells = 0; 1322 switch(dp->tag()) { 1323 case DataLayout::bit_data_tag: 1324 case DataLayout::no_tag: 1325 nb_cells = BitData::static_cell_count(); 1326 break; 1327 case DataLayout::speculative_trap_data_tag: 1328 nb_cells = SpeculativeTrapData::static_cell_count(); 1329 break; 1330 default: 1331 fatal("unexpected tag %d", dp->tag()); 1332 } 1333 return (DataLayout*)((address)dp + DataLayout::compute_size_in_bytes(nb_cells)); 1334 } 1335 1336 ProfileData* MethodData::bci_to_extra_data_helper(int bci, Method* m, DataLayout*& dp, bool concurrent) { 1337 DataLayout* end = args_data_limit(); 1338 1339 for (;; dp = next_extra(dp)) { 1340 assert(dp < end, "moved past end of extra data"); 1341 // No need for "OrderAccess::load_acquire" ops, 1342 // since the data structure is monotonic. 1343 switch(dp->tag()) { 1344 case DataLayout::no_tag: 1345 return NULL; 1346 case DataLayout::arg_info_data_tag: 1347 dp = end; 1348 return NULL; // ArgInfoData is at the end of extra data section. 1349 case DataLayout::bit_data_tag: 1350 if (m == NULL && dp->bci() == bci) { 1351 return new BitData(dp); 1352 } 1353 break; 1354 case DataLayout::speculative_trap_data_tag: 1355 if (m != NULL) { 1356 SpeculativeTrapData* data = new SpeculativeTrapData(dp); 1357 // data->method() may be null in case of a concurrent 1358 // allocation. Maybe it's for the same method. Try to use that 1359 // entry in that case. 1360 if (dp->bci() == bci) { 1361 if (data->method() == NULL) { 1362 assert(concurrent, "impossible because no concurrent allocation"); 1363 return NULL; 1364 } else if (data->method() == m) { 1365 return data; 1366 } 1367 } 1368 } 1369 break; 1370 default: 1371 fatal("unexpected tag %d", dp->tag()); 1372 } 1373 } 1374 return NULL; 1375 } 1376 1377 1378 // Translate a bci to its corresponding extra data, or NULL. 1379 ProfileData* MethodData::bci_to_extra_data(int bci, Method* m, bool create_if_missing) { 1380 // This code assumes an entry for a SpeculativeTrapData is 2 cells 1381 assert(2*DataLayout::compute_size_in_bytes(BitData::static_cell_count()) == 1382 DataLayout::compute_size_in_bytes(SpeculativeTrapData::static_cell_count()), 1383 "code needs to be adjusted"); 1384 1385 // Do not create one of these if method has been redefined. 1386 if (m != NULL && m->is_old()) { 1387 return NULL; 1388 } 1389 1390 DataLayout* dp = extra_data_base(); 1391 DataLayout* end = args_data_limit(); 1392 1393 // Allocation in the extra data space has to be atomic because not 1394 // all entries have the same size and non atomic concurrent 1395 // allocation would result in a corrupted extra data space. 1396 ProfileData* result = bci_to_extra_data_helper(bci, m, dp, true); 1397 if (result != NULL) { 1398 return result; 1399 } 1400 1401 if (create_if_missing && dp < end) { 1402 MutexLocker ml(&_extra_data_lock); 1403 // Check again now that we have the lock. Another thread may 1404 // have added extra data entries. 1405 ProfileData* result = bci_to_extra_data_helper(bci, m, dp, false); 1406 if (result != NULL || dp >= end) { 1407 return result; 1408 } 1409 1410 assert(dp->tag() == DataLayout::no_tag || (dp->tag() == DataLayout::speculative_trap_data_tag && m != NULL), "should be free"); 1411 assert(next_extra(dp)->tag() == DataLayout::no_tag || next_extra(dp)->tag() == DataLayout::arg_info_data_tag, "should be free or arg info"); 1412 u1 tag = m == NULL ? DataLayout::bit_data_tag : DataLayout::speculative_trap_data_tag; 1413 // SpeculativeTrapData is 2 slots. Make sure we have room. 1414 if (m != NULL && next_extra(dp)->tag() != DataLayout::no_tag) { 1415 return NULL; 1416 } 1417 DataLayout temp; 1418 temp.initialize(tag, bci, 0); 1419 1420 dp->set_header(temp.header()); 1421 assert(dp->tag() == tag, "sane"); 1422 assert(dp->bci() == bci, "no concurrent allocation"); 1423 if (tag == DataLayout::bit_data_tag) { 1424 return new BitData(dp); 1425 } else { 1426 SpeculativeTrapData* data = new SpeculativeTrapData(dp); 1427 data->set_method(m); 1428 return data; 1429 } 1430 } 1431 return NULL; 1432 } 1433 1434 ArgInfoData *MethodData::arg_info() { 1435 DataLayout* dp = extra_data_base(); 1436 DataLayout* end = args_data_limit(); 1437 for (; dp < end; dp = next_extra(dp)) { 1438 if (dp->tag() == DataLayout::arg_info_data_tag) 1439 return new ArgInfoData(dp); 1440 } 1441 return NULL; 1442 } 1443 1444 // Printing 1445 1446 void MethodData::print_on(outputStream* st) const { 1447 assert(is_methodData(), "should be method data"); 1448 st->print("method data for "); 1449 method()->print_value_on(st); 1450 st->cr(); 1451 print_data_on(st); 1452 } 1453 1454 void MethodData::print_value_on(outputStream* st) const { 1455 assert(is_methodData(), "should be method data"); 1456 st->print("method data for "); 1457 method()->print_value_on(st); 1458 } 1459 1460 void MethodData::print_data_on(outputStream* st) const { 1461 ResourceMark rm; 1462 ProfileData* data = first_data(); 1463 if (_parameters_type_data_di != no_parameters) { 1464 parameters_type_data()->print_data_on(st); 1465 } 1466 for ( ; is_valid(data); data = next_data(data)) { 1467 st->print("%d", dp_to_di(data->dp())); 1468 st->fill_to(6); 1469 data->print_data_on(st, this); 1470 } 1471 st->print_cr("--- Extra data:"); 1472 DataLayout* dp = extra_data_base(); 1473 DataLayout* end = args_data_limit(); 1474 for (;; dp = next_extra(dp)) { 1475 assert(dp < end, "moved past end of extra data"); 1476 // No need for "OrderAccess::load_acquire" ops, 1477 // since the data structure is monotonic. 1478 switch(dp->tag()) { 1479 case DataLayout::no_tag: 1480 continue; 1481 case DataLayout::bit_data_tag: 1482 data = new BitData(dp); 1483 break; 1484 case DataLayout::speculative_trap_data_tag: 1485 data = new SpeculativeTrapData(dp); 1486 break; 1487 case DataLayout::arg_info_data_tag: 1488 data = new ArgInfoData(dp); 1489 dp = end; // ArgInfoData is at the end of extra data section. 1490 break; 1491 default: 1492 fatal("unexpected tag %d", dp->tag()); 1493 } 1494 st->print("%d", dp_to_di(data->dp())); 1495 st->fill_to(6); 1496 data->print_data_on(st); 1497 if (dp >= end) return; 1498 } 1499 } 1500 1501 #if INCLUDE_SERVICES 1502 // Size Statistics 1503 void MethodData::collect_statistics(KlassSizeStats *sz) const { 1504 int n = sz->count(this); 1505 sz->_method_data_bytes += n; 1506 sz->_method_all_bytes += n; 1507 sz->_rw_bytes += n; 1508 } 1509 #endif // INCLUDE_SERVICES 1510 1511 // Verification 1512 1513 void MethodData::verify_on(outputStream* st) { 1514 guarantee(is_methodData(), "object must be method data"); 1515 // guarantee(m->is_perm(), "should be in permspace"); 1516 this->verify_data_on(st); 1517 } 1518 1519 void MethodData::verify_data_on(outputStream* st) { 1520 NEEDS_CLEANUP; 1521 // not yet implemented. 1522 } 1523 1524 bool MethodData::profile_jsr292(const methodHandle& m, int bci) { 1525 if (m->is_compiled_lambda_form()) { 1526 return true; 1527 } 1528 1529 Bytecode_invoke inv(m , bci); 1530 return inv.is_invokedynamic() || inv.is_invokehandle(); 1531 } 1532 1533 bool MethodData::profile_unsafe(const methodHandle& m, int bci) { 1534 Bytecode_invoke inv(m , bci); 1535 if (inv.is_invokevirtual() && inv.klass() == vmSymbols::jdk_internal_misc_Unsafe()) { 1536 ResourceMark rm; 1537 char* name = inv.name()->as_C_string(); 1538 if (!strncmp(name, "get", 3) || !strncmp(name, "put", 3)) { 1539 return true; 1540 } 1541 } 1542 return false; 1543 } 1544 1545 int MethodData::profile_arguments_flag() { 1546 return TypeProfileLevel % 10; 1547 } 1548 1549 bool MethodData::profile_arguments() { 1550 return profile_arguments_flag() > no_type_profile && profile_arguments_flag() <= type_profile_all; 1551 } 1552 1553 bool MethodData::profile_arguments_jsr292_only() { 1554 return profile_arguments_flag() == type_profile_jsr292; 1555 } 1556 1557 bool MethodData::profile_all_arguments() { 1558 return profile_arguments_flag() == type_profile_all; 1559 } 1560 1561 bool MethodData::profile_arguments_for_invoke(const methodHandle& m, int bci) { 1562 if (!profile_arguments()) { 1563 return false; 1564 } 1565 1566 if (profile_all_arguments()) { 1567 return true; 1568 } 1569 1570 if (profile_unsafe(m, bci)) { 1571 return true; 1572 } 1573 1574 assert(profile_arguments_jsr292_only(), "inconsistent"); 1575 return profile_jsr292(m, bci); 1576 } 1577 1578 int MethodData::profile_return_flag() { 1579 return (TypeProfileLevel % 100) / 10; 1580 } 1581 1582 bool MethodData::profile_return() { 1583 return profile_return_flag() > no_type_profile && profile_return_flag() <= type_profile_all; 1584 } 1585 1586 bool MethodData::profile_return_jsr292_only() { 1587 return profile_return_flag() == type_profile_jsr292; 1588 } 1589 1590 bool MethodData::profile_all_return() { 1591 return profile_return_flag() == type_profile_all; 1592 } 1593 1594 bool MethodData::profile_return_for_invoke(const methodHandle& m, int bci) { 1595 if (!profile_return()) { 1596 return false; 1597 } 1598 1599 if (profile_all_return()) { 1600 return true; 1601 } 1602 1603 assert(profile_return_jsr292_only(), "inconsistent"); 1604 return profile_jsr292(m, bci); 1605 } 1606 1607 int MethodData::profile_parameters_flag() { 1608 return TypeProfileLevel / 100; 1609 } 1610 1611 bool MethodData::profile_parameters() { 1612 return profile_parameters_flag() > no_type_profile && profile_parameters_flag() <= type_profile_all; 1613 } 1614 1615 bool MethodData::profile_parameters_jsr292_only() { 1616 return profile_parameters_flag() == type_profile_jsr292; 1617 } 1618 1619 bool MethodData::profile_all_parameters() { 1620 return profile_parameters_flag() == type_profile_all; 1621 } 1622 1623 bool MethodData::profile_parameters_for_method(const methodHandle& m) { 1624 if (!profile_parameters()) { 1625 return false; 1626 } 1627 1628 if (profile_all_parameters()) { 1629 return true; 1630 } 1631 1632 assert(profile_parameters_jsr292_only(), "inconsistent"); 1633 return m->is_compiled_lambda_form(); 1634 } 1635 1636 void MethodData::clean_extra_data_helper(DataLayout* dp, int shift, bool reset) { 1637 if (shift == 0) { 1638 return; 1639 } 1640 if (!reset) { 1641 // Move all cells of trap entry at dp left by "shift" cells 1642 intptr_t* start = (intptr_t*)dp; 1643 intptr_t* end = (intptr_t*)next_extra(dp); 1644 for (intptr_t* ptr = start; ptr < end; ptr++) { 1645 *(ptr-shift) = *ptr; 1646 } 1647 } else { 1648 // Reset "shift" cells stopping at dp 1649 intptr_t* start = ((intptr_t*)dp) - shift; 1650 intptr_t* end = (intptr_t*)dp; 1651 for (intptr_t* ptr = start; ptr < end; ptr++) { 1652 *ptr = 0; 1653 } 1654 } 1655 } 1656 1657 class CleanExtraDataClosure : public StackObj { 1658 public: 1659 virtual bool is_live(Method* m) = 0; 1660 }; 1661 1662 // Check for entries that reference an unloaded method 1663 class CleanExtraDataKlassClosure : public CleanExtraDataClosure { 1664 private: 1665 BoolObjectClosure* _is_alive; 1666 public: 1667 CleanExtraDataKlassClosure(BoolObjectClosure* is_alive) : _is_alive(is_alive) {} 1668 bool is_live(Method* m) { 1669 return m->method_holder()->is_loader_alive(_is_alive); 1670 } 1671 }; 1672 1673 // Check for entries that reference a redefined method 1674 class CleanExtraDataMethodClosure : public CleanExtraDataClosure { 1675 public: 1676 CleanExtraDataMethodClosure() {} 1677 bool is_live(Method* m) { return !m->is_old(); } 1678 }; 1679 1680 1681 // Remove SpeculativeTrapData entries that reference an unloaded or 1682 // redefined method 1683 void MethodData::clean_extra_data(CleanExtraDataClosure* cl) { 1684 DataLayout* dp = extra_data_base(); 1685 DataLayout* end = args_data_limit(); 1686 1687 int shift = 0; 1688 for (; dp < end; dp = next_extra(dp)) { 1689 switch(dp->tag()) { 1690 case DataLayout::speculative_trap_data_tag: { 1691 SpeculativeTrapData* data = new SpeculativeTrapData(dp); 1692 Method* m = data->method(); 1693 assert(m != NULL, "should have a method"); 1694 if (!cl->is_live(m)) { 1695 // "shift" accumulates the number of cells for dead 1696 // SpeculativeTrapData entries that have been seen so 1697 // far. Following entries must be shifted left by that many 1698 // cells to remove the dead SpeculativeTrapData entries. 1699 shift += (int)((intptr_t*)next_extra(dp) - (intptr_t*)dp); 1700 } else { 1701 // Shift this entry left if it follows dead 1702 // SpeculativeTrapData entries 1703 clean_extra_data_helper(dp, shift); 1704 } 1705 break; 1706 } 1707 case DataLayout::bit_data_tag: 1708 // Shift this entry left if it follows dead SpeculativeTrapData 1709 // entries 1710 clean_extra_data_helper(dp, shift); 1711 continue; 1712 case DataLayout::no_tag: 1713 case DataLayout::arg_info_data_tag: 1714 // We are at end of the live trap entries. The previous "shift" 1715 // cells contain entries that are either dead or were shifted 1716 // left. They need to be reset to no_tag 1717 clean_extra_data_helper(dp, shift, true); 1718 return; 1719 default: 1720 fatal("unexpected tag %d", dp->tag()); 1721 } 1722 } 1723 } 1724 1725 // Verify there's no unloaded or redefined method referenced by a 1726 // SpeculativeTrapData entry 1727 void MethodData::verify_extra_data_clean(CleanExtraDataClosure* cl) { 1728 #ifdef ASSERT 1729 DataLayout* dp = extra_data_base(); 1730 DataLayout* end = args_data_limit(); 1731 1732 for (; dp < end; dp = next_extra(dp)) { 1733 switch(dp->tag()) { 1734 case DataLayout::speculative_trap_data_tag: { 1735 SpeculativeTrapData* data = new SpeculativeTrapData(dp); 1736 Method* m = data->method(); 1737 assert(m != NULL && cl->is_live(m), "Method should exist"); 1738 break; 1739 } 1740 case DataLayout::bit_data_tag: 1741 continue; 1742 case DataLayout::no_tag: 1743 case DataLayout::arg_info_data_tag: 1744 return; 1745 default: 1746 fatal("unexpected tag %d", dp->tag()); 1747 } 1748 } 1749 #endif 1750 } 1751 1752 void MethodData::clean_method_data(BoolObjectClosure* is_alive) { 1753 ResourceMark rm; 1754 for (ProfileData* data = first_data(); 1755 is_valid(data); 1756 data = next_data(data)) { 1757 data->clean_weak_klass_links(is_alive); 1758 } 1759 ParametersTypeData* parameters = parameters_type_data(); 1760 if (parameters != NULL) { 1761 parameters->clean_weak_klass_links(is_alive); 1762 } 1763 1764 CleanExtraDataKlassClosure cl(is_alive); 1765 clean_extra_data(&cl); 1766 verify_extra_data_clean(&cl); 1767 } 1768 1769 void MethodData::clean_weak_method_links() { 1770 ResourceMark rm; 1771 for (ProfileData* data = first_data(); 1772 is_valid(data); 1773 data = next_data(data)) { 1774 data->clean_weak_method_links(); 1775 } 1776 1777 CleanExtraDataMethodClosure cl; 1778 clean_extra_data(&cl); 1779 verify_extra_data_clean(&cl); 1780 } 1781 1782 #ifdef ASSERT 1783 void MethodData::verify_clean_weak_method_links() { 1784 ResourceMark rm; 1785 for (ProfileData* data = first_data(); 1786 is_valid(data); 1787 data = next_data(data)) { 1788 data->verify_clean_weak_method_links(); 1789 } 1790 1791 CleanExtraDataMethodClosure cl; 1792 verify_extra_data_clean(&cl); 1793 } 1794 #endif // ASSERT