1 /* 2 * Copyright (c) 1997, 2016, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "interpreter/interpreter.hpp" 27 #include "memory/resourceArea.hpp" 28 #include "oops/markOop.hpp" 29 #include "oops/method.hpp" 30 #include "oops/oop.inline.hpp" 31 #include "prims/methodHandles.hpp" 32 #include "runtime/frame.inline.hpp" 33 #include "runtime/handles.inline.hpp" 34 #include "runtime/javaCalls.hpp" 35 #include "runtime/monitorChunk.hpp" 36 #include "runtime/os.inline.hpp" 37 #include "runtime/signature.hpp" 38 #include "runtime/stubCodeGenerator.hpp" 39 #include "runtime/stubRoutines.hpp" 40 #include "vmreg_x86.inline.hpp" 41 #ifdef COMPILER1 42 #include "c1/c1_Runtime1.hpp" 43 #include "runtime/vframeArray.hpp" 44 #endif 45 46 #ifdef ASSERT 47 void RegisterMap::check_location_valid() { 48 } 49 #endif 50 51 // Profiling/safepoint support 52 53 bool frame::safe_for_sender(JavaThread *thread) { 54 address sp = (address)_sp; 55 address fp = (address)_fp; 56 address unextended_sp = (address)_unextended_sp; 57 58 // consider stack guards when trying to determine "safe" stack pointers 59 static size_t stack_guard_size = os::uses_stack_guard_pages() ? 60 JavaThread::stack_red_zone_size() + JavaThread::stack_yellow_zone_size() : 0; 61 size_t usable_stack_size = thread->stack_size() - stack_guard_size; 62 63 // sp must be within the usable part of the stack (not in guards) 64 bool sp_safe = (sp < thread->stack_base()) && 65 (sp >= thread->stack_base() - usable_stack_size); 66 67 68 if (!sp_safe) { 69 return false; 70 } 71 72 // unextended sp must be within the stack and above or equal sp 73 bool unextended_sp_safe = (unextended_sp < thread->stack_base()) && 74 (unextended_sp >= sp); 75 76 if (!unextended_sp_safe) { 77 return false; 78 } 79 80 // an fp must be within the stack and above (but not equal) sp 81 // second evaluation on fp+ is added to handle situation where fp is -1 82 bool fp_safe = (fp < thread->stack_base() && (fp > sp) && (((fp + (return_addr_offset * sizeof(void*))) < thread->stack_base()))); 83 84 // We know sp/unextended_sp are safe only fp is questionable here 85 86 // If the current frame is known to the code cache then we can attempt to 87 // to construct the sender and do some validation of it. This goes a long way 88 // toward eliminating issues when we get in frame construction code 89 90 if (_cb != NULL ) { 91 92 // First check if frame is complete and tester is reliable 93 // Unfortunately we can only check frame complete for runtime stubs and nmethod 94 // other generic buffer blobs are more problematic so we just assume they are 95 // ok. adapter blobs never have a frame complete and are never ok. 96 97 if (!_cb->is_frame_complete_at(_pc)) { 98 if (_cb->is_compiled() || _cb->is_adapter_blob() || _cb->is_runtime_stub()) { 99 return false; 100 } 101 } 102 103 // Could just be some random pointer within the codeBlob 104 if (!_cb->code_contains(_pc)) { 105 return false; 106 } 107 108 // Entry frame checks 109 if (is_entry_frame()) { 110 // an entry frame must have a valid fp. 111 112 if (!fp_safe) return false; 113 114 return entry_frame_is_safe(thread); 115 } 116 117 intptr_t* sender_sp = NULL; 118 intptr_t* sender_unextended_sp = NULL; 119 address sender_pc = NULL; 120 intptr_t* saved_fp = NULL; 121 122 if (is_interpreted_frame()) { 123 // fp must be safe 124 if (!fp_safe) { 125 return false; 126 } 127 128 sender_pc = (address) this->fp()[return_addr_offset]; 129 // for interpreted frames, the value below is the sender "raw" sp, 130 // which can be different from the sender unextended sp (the sp seen 131 // by the sender) because of current frame local variables 132 sender_sp = (intptr_t*) addr_at(sender_sp_offset); 133 sender_unextended_sp = (intptr_t*) this->fp()[interpreter_frame_sender_sp_offset]; 134 saved_fp = (intptr_t*) this->fp()[link_offset]; 135 136 } else { 137 // must be some sort of compiled/runtime frame 138 // fp does not have to be safe (although it could be check for c1?) 139 140 // check for a valid frame_size, otherwise we are unlikely to get a valid sender_pc 141 if (_cb->frame_size() <= 0) { 142 return false; 143 } 144 145 sender_sp = _unextended_sp + _cb->frame_size(); 146 sender_unextended_sp = sender_sp; 147 // On Intel the return_address is always the word on the stack 148 sender_pc = (address) *(sender_sp-1); 149 // Note: frame::sender_sp_offset is only valid for compiled frame 150 saved_fp = (intptr_t*) *(sender_sp - frame::sender_sp_offset); 151 } 152 153 154 // If the potential sender is the interpreter then we can do some more checking 155 if (Interpreter::contains(sender_pc)) { 156 157 // ebp is always saved in a recognizable place in any code we generate. However 158 // only if the sender is interpreted/call_stub (c1 too?) are we certain that the saved ebp 159 // is really a frame pointer. 160 161 bool saved_fp_safe = ((address)saved_fp < thread->stack_base()) && (saved_fp > sender_sp); 162 163 if (!saved_fp_safe) { 164 return false; 165 } 166 167 // construct the potential sender 168 169 frame sender(sender_sp, sender_unextended_sp, saved_fp, sender_pc); 170 171 return sender.is_interpreted_frame_valid(thread); 172 173 } 174 175 // We must always be able to find a recognizable pc 176 CodeBlob* sender_blob = CodeCache::find_blob_unsafe(sender_pc); 177 if (sender_pc == NULL || sender_blob == NULL) { 178 return false; 179 } 180 181 // Could be a zombie method 182 if (sender_blob->is_zombie() || sender_blob->is_unloaded()) { 183 return false; 184 } 185 186 // Could just be some random pointer within the codeBlob 187 if (!sender_blob->code_contains(sender_pc)) { 188 return false; 189 } 190 191 // We should never be able to see an adapter if the current frame is something from code cache 192 if (sender_blob->is_adapter_blob()) { 193 return false; 194 } 195 196 // Could be the call_stub 197 if (StubRoutines::returns_to_call_stub(sender_pc)) { 198 bool saved_fp_safe = ((address)saved_fp < thread->stack_base()) && (saved_fp > sender_sp); 199 200 if (!saved_fp_safe) { 201 return false; 202 } 203 204 // construct the potential sender 205 frame sender(sender_sp, sender_unextended_sp, saved_fp, sender_pc); 206 return sender.entry_frame_is_safe(thread); 207 } 208 209 CompiledMethod* nm = sender_blob->as_compiled_method_or_null(); 210 if (nm != NULL) { 211 if (nm->is_deopt_mh_entry(sender_pc) || nm->is_deopt_entry(sender_pc) || 212 nm->method()->is_method_handle_intrinsic()) { 213 return false; 214 } 215 } 216 217 // If the frame size is 0 something (or less) is bad because every nmethod has a non-zero frame size 218 // because the return address counts against the callee's frame. 219 220 if (sender_blob->frame_size() <= 0) { 221 assert(!sender_blob->is_compiled(), "should count return address at least"); 222 return false; 223 } 224 225 // We should never be able to see anything here except an nmethod. If something in the 226 // code cache (current frame) is called by an entity within the code cache that entity 227 // should not be anything but the call stub (already covered), the interpreter (already covered) 228 // or an nmethod. 229 230 if (!sender_blob->is_compiled()) { 231 return false; 232 } 233 234 // Could put some more validation for the potential non-interpreted sender 235 // frame we'd create by calling sender if I could think of any. Wait for next crash in forte... 236 237 // One idea is seeing if the sender_pc we have is one that we'd expect to call to current cb 238 239 // We've validated the potential sender that would be created 240 return true; 241 } 242 243 // Must be native-compiled frame. Since sender will try and use fp to find 244 // linkages it must be safe 245 246 if (!fp_safe) { 247 return false; 248 } 249 250 // Will the pc we fetch be non-zero (which we'll find at the oldest frame) 251 252 if ( (address) this->fp()[return_addr_offset] == NULL) return false; 253 254 255 // could try and do some more potential verification of native frame if we could think of some... 256 257 return true; 258 259 } 260 261 262 void frame::patch_pc(Thread* thread, address pc) { 263 address* pc_addr = &(((address*) sp())[-1]); 264 if (TracePcPatching) { 265 tty->print_cr("patch_pc at address " INTPTR_FORMAT " [" INTPTR_FORMAT " -> " INTPTR_FORMAT "]", 266 p2i(pc_addr), p2i(*pc_addr), p2i(pc)); 267 } 268 // Either the return address is the original one or we are going to 269 // patch in the same address that's already there. 270 assert(_pc == *pc_addr || pc == *pc_addr, "must be"); 271 *pc_addr = pc; 272 _cb = CodeCache::find_blob(pc); 273 address original_pc = CompiledMethod::get_deopt_original_pc(this); 274 if (original_pc != NULL) { 275 assert(original_pc == _pc, "expected original PC to be stored before patching"); 276 _deopt_state = is_deoptimized; 277 // leave _pc as is 278 } else { 279 _deopt_state = not_deoptimized; 280 _pc = pc; 281 } 282 } 283 284 bool frame::is_interpreted_frame() const { 285 return Interpreter::contains(pc()); 286 } 287 288 int frame::frame_size(RegisterMap* map) const { 289 frame sender = this->sender(map); 290 return sender.sp() - sp(); 291 } 292 293 intptr_t* frame::entry_frame_argument_at(int offset) const { 294 // convert offset to index to deal with tsi 295 int index = (Interpreter::expr_offset_in_bytes(offset)/wordSize); 296 // Entry frame's arguments are always in relation to unextended_sp() 297 return &unextended_sp()[index]; 298 } 299 300 // sender_sp 301 302 intptr_t* frame::interpreter_frame_sender_sp() const { 303 assert(is_interpreted_frame(), "interpreted frame expected"); 304 return (intptr_t*) at(interpreter_frame_sender_sp_offset); 305 } 306 307 void frame::set_interpreter_frame_sender_sp(intptr_t* sender_sp) { 308 assert(is_interpreted_frame(), "interpreted frame expected"); 309 ptr_at_put(interpreter_frame_sender_sp_offset, (intptr_t) sender_sp); 310 } 311 312 313 // monitor elements 314 315 BasicObjectLock* frame::interpreter_frame_monitor_begin() const { 316 return (BasicObjectLock*) addr_at(interpreter_frame_monitor_block_bottom_offset); 317 } 318 319 BasicObjectLock* frame::interpreter_frame_monitor_end() const { 320 BasicObjectLock* result = (BasicObjectLock*) *addr_at(interpreter_frame_monitor_block_top_offset); 321 // make sure the pointer points inside the frame 322 assert(sp() <= (intptr_t*) result, "monitor end should be above the stack pointer"); 323 assert((intptr_t*) result < fp(), "monitor end should be strictly below the frame pointer"); 324 return result; 325 } 326 327 void frame::interpreter_frame_set_monitor_end(BasicObjectLock* value) { 328 *((BasicObjectLock**)addr_at(interpreter_frame_monitor_block_top_offset)) = value; 329 } 330 331 // Used by template based interpreter deoptimization 332 void frame::interpreter_frame_set_last_sp(intptr_t* sp) { 333 *((intptr_t**)addr_at(interpreter_frame_last_sp_offset)) = sp; 334 } 335 336 frame frame::sender_for_entry_frame(RegisterMap* map) const { 337 assert(map != NULL, "map must be set"); 338 // Java frame called from C; skip all C frames and return top C 339 // frame of that chunk as the sender 340 JavaFrameAnchor* jfa = entry_frame_call_wrapper()->anchor(); 341 assert(!entry_frame_is_first(), "next Java fp must be non zero"); 342 assert(jfa->last_Java_sp() > sp(), "must be above this frame on stack"); 343 map->clear(); 344 assert(map->include_argument_oops(), "should be set by clear"); 345 if (jfa->last_Java_pc() != NULL ) { 346 frame fr(jfa->last_Java_sp(), jfa->last_Java_fp(), jfa->last_Java_pc()); 347 return fr; 348 } 349 frame fr(jfa->last_Java_sp(), jfa->last_Java_fp()); 350 return fr; 351 } 352 353 //------------------------------------------------------------------------------ 354 // frame::verify_deopt_original_pc 355 // 356 // Verifies the calculated original PC of a deoptimization PC for the 357 // given unextended SP. 358 #ifdef ASSERT 359 void frame::verify_deopt_original_pc(CompiledMethod* nm, intptr_t* unextended_sp) { 360 frame fr; 361 362 // This is ugly but it's better than to change {get,set}_original_pc 363 // to take an SP value as argument. And it's only a debugging 364 // method anyway. 365 fr._unextended_sp = unextended_sp; 366 367 address original_pc = nm->get_original_pc(&fr); 368 assert(nm->insts_contains(original_pc), "original PC must be in CompiledMethod"); 369 } 370 #endif 371 372 //------------------------------------------------------------------------------ 373 // frame::adjust_unextended_sp 374 void frame::adjust_unextended_sp() { 375 // On x86, sites calling method handle intrinsics and lambda forms are treated 376 // as any other call site. Therefore, no special action is needed when we are 377 // returning to any of these call sites. 378 379 if (_cb != NULL) { 380 CompiledMethod* sender_cm = _cb->as_compiled_method_or_null(); 381 if (sender_cm != NULL) { 382 // If the sender PC is a deoptimization point, get the original PC. 383 if (sender_cm->is_deopt_entry(_pc) || 384 sender_cm->is_deopt_mh_entry(_pc)) { 385 DEBUG_ONLY(verify_deopt_original_pc(sender_cm, _unextended_sp)); 386 } 387 } 388 } 389 } 390 391 //------------------------------------------------------------------------------ 392 // frame::update_map_with_saved_link 393 void frame::update_map_with_saved_link(RegisterMap* map, intptr_t** link_addr) { 394 // The interpreter and compiler(s) always save EBP/RBP in a known 395 // location on entry. We must record where that location is 396 // so this if EBP/RBP was live on callout from c2 we can find 397 // the saved copy no matter what it called. 398 399 // Since the interpreter always saves EBP/RBP if we record where it is then 400 // we don't have to always save EBP/RBP on entry and exit to c2 compiled 401 // code, on entry will be enough. 402 map->set_location(rbp->as_VMReg(), (address) link_addr); 403 #ifdef AMD64 404 // this is weird "H" ought to be at a higher address however the 405 // oopMaps seems to have the "H" regs at the same address and the 406 // vanilla register. 407 // XXXX make this go away 408 if (true) { 409 map->set_location(rbp->as_VMReg()->next(), (address) link_addr); 410 } 411 #endif // AMD64 412 } 413 414 415 //------------------------------------------------------------------------------ 416 // frame::sender_for_interpreter_frame 417 frame frame::sender_for_interpreter_frame(RegisterMap* map) const { 418 // SP is the raw SP from the sender after adapter or interpreter 419 // extension. 420 intptr_t* sender_sp = this->sender_sp(); 421 422 // This is the sp before any possible extension (adapter/locals). 423 intptr_t* unextended_sp = interpreter_frame_sender_sp(); 424 425 #if defined(COMPILER2) || INCLUDE_JVMCI 426 if (map->update_map()) { 427 update_map_with_saved_link(map, (intptr_t**) addr_at(link_offset)); 428 } 429 #endif // COMPILER2 || INCLUDE_JVMCI 430 431 return frame(sender_sp, unextended_sp, link(), sender_pc()); 432 } 433 434 435 //------------------------------------------------------------------------------ 436 // frame::sender_for_compiled_frame 437 frame frame::sender_for_compiled_frame(RegisterMap* map) const { 438 assert(map != NULL, "map must be set"); 439 440 // frame owned by optimizing compiler 441 assert(_cb->frame_size() >= 0, "must have non-zero frame size"); 442 intptr_t* sender_sp = unextended_sp() + _cb->frame_size(); 443 intptr_t* unextended_sp = sender_sp; 444 445 // On Intel the return_address is always the word on the stack 446 address sender_pc = (address) *(sender_sp-1); 447 448 // This is the saved value of EBP which may or may not really be an FP. 449 // It is only an FP if the sender is an interpreter frame (or C1?). 450 intptr_t** saved_fp_addr = (intptr_t**) (sender_sp - frame::sender_sp_offset); 451 452 if (map->update_map()) { 453 // Tell GC to use argument oopmaps for some runtime stubs that need it. 454 // For C1, the runtime stub might not have oop maps, so set this flag 455 // outside of update_register_map. 456 map->set_include_argument_oops(_cb->caller_must_gc_arguments(map->thread())); 457 if (_cb->oop_maps() != NULL) { 458 OopMapSet::update_register_map(this, map); 459 } 460 461 // Since the prolog does the save and restore of EBP there is no oopmap 462 // for it so we must fill in its location as if there was an oopmap entry 463 // since if our caller was compiled code there could be live jvm state in it. 464 update_map_with_saved_link(map, saved_fp_addr); 465 } 466 467 assert(sender_sp != sp(), "must have changed"); 468 return frame(sender_sp, unextended_sp, *saved_fp_addr, sender_pc); 469 } 470 471 472 //------------------------------------------------------------------------------ 473 // frame::sender 474 frame frame::sender(RegisterMap* map) const { 475 // Default is we done have to follow them. The sender_for_xxx will 476 // update it accordingly 477 map->set_include_argument_oops(false); 478 479 if (is_entry_frame()) return sender_for_entry_frame(map); 480 if (is_interpreted_frame()) return sender_for_interpreter_frame(map); 481 assert(_cb == CodeCache::find_blob(pc()),"Must be the same"); 482 483 if (_cb != NULL) { 484 return sender_for_compiled_frame(map); 485 } 486 // Must be native-compiled frame, i.e. the marshaling code for native 487 // methods that exists in the core system. 488 return frame(sender_sp(), link(), sender_pc()); 489 } 490 491 bool frame::is_interpreted_frame_valid(JavaThread* thread) const { 492 assert(is_interpreted_frame(), "Not an interpreted frame"); 493 // These are reasonable sanity checks 494 if (fp() == 0 || (intptr_t(fp()) & (wordSize-1)) != 0) { 495 return false; 496 } 497 if (sp() == 0 || (intptr_t(sp()) & (wordSize-1)) != 0) { 498 return false; 499 } 500 if (fp() + interpreter_frame_initial_sp_offset < sp()) { 501 return false; 502 } 503 // These are hacks to keep us out of trouble. 504 // The problem with these is that they mask other problems 505 if (fp() <= sp()) { // this attempts to deal with unsigned comparison above 506 return false; 507 } 508 509 // do some validation of frame elements 510 // first the method 511 512 Method* m = *interpreter_frame_method_addr(); 513 514 // validate the method we'd find in this potential sender 515 if (!m->is_valid_method()) return false; 516 517 // stack frames shouldn't be much larger than max_stack elements 518 // this test requires the use the unextended_sp which is the sp as seen by 519 // the current frame, and not sp which is the "raw" pc which could point 520 // further because of local variables of the callee method inserted after 521 // method arguments 522 if (fp() - unextended_sp() > 1024 + m->max_stack()*Interpreter::stackElementSize) { 523 return false; 524 } 525 526 // validate bci/bcp 527 528 address bcp = interpreter_frame_bcp(); 529 if (m->validate_bci_from_bcp(bcp) < 0) { 530 return false; 531 } 532 533 // validate ConstantPoolCache* 534 ConstantPoolCache* cp = *interpreter_frame_cache_addr(); 535 if (cp == NULL || !cp->is_metaspace_object()) return false; 536 537 // validate locals 538 539 address locals = (address) *interpreter_frame_locals_addr(); 540 541 if (locals > thread->stack_base() || locals < (address) fp()) return false; 542 543 // We'd have to be pretty unlucky to be mislead at this point 544 return true; 545 } 546 547 BasicType frame::interpreter_frame_result(oop* oop_result, jvalue* value_result) { 548 assert(is_interpreted_frame(), "interpreted frame expected"); 549 Method* method = interpreter_frame_method(); 550 BasicType type = method->result_type(); 551 552 intptr_t* tos_addr; 553 if (method->is_native()) { 554 // Prior to calling into the runtime to report the method_exit the possible 555 // return value is pushed to the native stack. If the result is a jfloat/jdouble 556 // then ST0 is saved before EAX/EDX. See the note in generate_native_result 557 tos_addr = (intptr_t*)sp(); 558 if (type == T_FLOAT || type == T_DOUBLE) { 559 // QQQ seems like this code is equivalent on the two platforms 560 #ifdef AMD64 561 // This is times two because we do a push(ltos) after pushing XMM0 562 // and that takes two interpreter stack slots. 563 tos_addr += 2 * Interpreter::stackElementWords; 564 #else 565 tos_addr += 2; 566 #endif // AMD64 567 } 568 } else { 569 tos_addr = (intptr_t*)interpreter_frame_tos_address(); 570 } 571 572 switch (type) { 573 case T_OBJECT : 574 case T_ARRAY : { 575 oop obj; 576 if (method->is_native()) { 577 obj = cast_to_oop(at(interpreter_frame_oop_temp_offset)); 578 } else { 579 oop* obj_p = (oop*)tos_addr; 580 obj = (obj_p == NULL) ? (oop)NULL : *obj_p; 581 } 582 assert(obj == NULL || Universe::heap()->is_in(obj), "sanity check"); 583 *oop_result = obj; 584 break; 585 } 586 case T_BOOLEAN : value_result->z = *(jboolean*)tos_addr; break; 587 case T_BYTE : value_result->b = *(jbyte*)tos_addr; break; 588 case T_CHAR : value_result->c = *(jchar*)tos_addr; break; 589 case T_SHORT : value_result->s = *(jshort*)tos_addr; break; 590 case T_INT : value_result->i = *(jint*)tos_addr; break; 591 case T_LONG : value_result->j = *(jlong*)tos_addr; break; 592 case T_FLOAT : { 593 #ifdef AMD64 594 value_result->f = *(jfloat*)tos_addr; 595 #else 596 if (method->is_native()) { 597 jdouble d = *(jdouble*)tos_addr; // Result was in ST0 so need to convert to jfloat 598 value_result->f = (jfloat)d; 599 } else { 600 value_result->f = *(jfloat*)tos_addr; 601 } 602 #endif // AMD64 603 break; 604 } 605 case T_DOUBLE : value_result->d = *(jdouble*)tos_addr; break; 606 case T_VOID : /* Nothing to do */ break; 607 default : ShouldNotReachHere(); 608 } 609 610 return type; 611 } 612 613 614 intptr_t* frame::interpreter_frame_tos_at(jint offset) const { 615 int index = (Interpreter::expr_offset_in_bytes(offset)/wordSize); 616 return &interpreter_frame_tos_address()[index]; 617 } 618 619 #ifndef PRODUCT 620 621 #define DESCRIBE_FP_OFFSET(name) \ 622 values.describe(frame_no, fp() + frame::name##_offset, #name) 623 624 void frame::describe_pd(FrameValues& values, int frame_no) { 625 if (is_interpreted_frame()) { 626 DESCRIBE_FP_OFFSET(interpreter_frame_sender_sp); 627 DESCRIBE_FP_OFFSET(interpreter_frame_last_sp); 628 DESCRIBE_FP_OFFSET(interpreter_frame_method); 629 DESCRIBE_FP_OFFSET(interpreter_frame_mirror); 630 DESCRIBE_FP_OFFSET(interpreter_frame_mdp); 631 DESCRIBE_FP_OFFSET(interpreter_frame_cache); 632 DESCRIBE_FP_OFFSET(interpreter_frame_locals); 633 DESCRIBE_FP_OFFSET(interpreter_frame_bcp); 634 DESCRIBE_FP_OFFSET(interpreter_frame_initial_sp); 635 #ifdef AMD64 636 } else if (is_entry_frame()) { 637 // This could be more descriptive if we use the enum in 638 // stubGenerator to map to real names but it's most important to 639 // claim these frame slots so the error checking works. 640 for (int i = 0; i < entry_frame_after_call_words; i++) { 641 values.describe(frame_no, fp() - i, err_msg("call_stub word fp - %d", i)); 642 } 643 #endif // AMD64 644 } 645 } 646 #endif // !PRODUCT 647 648 intptr_t *frame::initial_deoptimization_info() { 649 // used to reset the saved FP 650 return fp(); 651 } 652 653 intptr_t* frame::real_fp() const { 654 if (_cb != NULL) { 655 // use the frame size if valid 656 int size = _cb->frame_size(); 657 if (size > 0) { 658 return unextended_sp() + size; 659 } 660 } 661 // else rely on fp() 662 assert(! is_compiled_frame(), "unknown compiled frame size"); 663 return fp(); 664 } 665 666 #ifndef PRODUCT 667 // This is a generic constructor which is only used by pns() in debug.cpp. 668 frame::frame(void* sp, void* fp, void* pc) { 669 init((intptr_t*)sp, (intptr_t*)fp, (address)pc); 670 } 671 #endif