1 /* 2 * Copyright (c) 2003, 2011, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "classfile/systemDictionary.hpp" 27 #include "interpreter/interpreter.hpp" 28 #include "jvmtifiles/jvmtiEnv.hpp" 29 #include "memory/resourceArea.hpp" 30 #include "oops/instanceKlass.hpp" 31 #include "prims/jvmtiAgentThread.hpp" 32 #include "prims/jvmtiEventController.inline.hpp" 33 #include "prims/jvmtiImpl.hpp" 34 #include "prims/jvmtiRedefineClasses.hpp" 35 #include "runtime/deoptimization.hpp" 36 #include "runtime/handles.hpp" 37 #include "runtime/handles.inline.hpp" 38 #include "runtime/interfaceSupport.hpp" 39 #include "runtime/javaCalls.hpp" 40 #include "runtime/signature.hpp" 41 #include "runtime/vframe.hpp" 42 #include "runtime/vframe_hp.hpp" 43 #include "runtime/vm_operations.hpp" 44 #include "utilities/exceptions.hpp" 45 #ifdef TARGET_OS_FAMILY_linux 46 # include "thread_linux.inline.hpp" 47 #endif 48 #ifdef TARGET_OS_FAMILY_solaris 49 # include "thread_solaris.inline.hpp" 50 #endif 51 #ifdef TARGET_OS_FAMILY_windows 52 # include "thread_windows.inline.hpp" 53 #endif 54 55 // 56 // class JvmtiAgentThread 57 // 58 // JavaThread used to wrap a thread started by an agent 59 // using the JVMTI method RunAgentThread. 60 // 61 62 JvmtiAgentThread::JvmtiAgentThread(JvmtiEnv* env, jvmtiStartFunction start_fn, const void *start_arg) 63 : JavaThread(start_function_wrapper) { 64 _env = env; 65 _start_fn = start_fn; 66 _start_arg = start_arg; 67 } 68 69 void 70 JvmtiAgentThread::start_function_wrapper(JavaThread *thread, TRAPS) { 71 // It is expected that any Agent threads will be created as 72 // Java Threads. If this is the case, notification of the creation 73 // of the thread is given in JavaThread::thread_main(). 74 assert(thread->is_Java_thread(), "debugger thread should be a Java Thread"); 75 assert(thread == JavaThread::current(), "sanity check"); 76 77 JvmtiAgentThread *dthread = (JvmtiAgentThread *)thread; 78 dthread->call_start_function(); 79 } 80 81 void 82 JvmtiAgentThread::call_start_function() { 83 ThreadToNativeFromVM transition(this); 84 _start_fn(_env->jvmti_external(), jni_environment(), (void*)_start_arg); 85 } 86 87 88 // 89 // class GrowableCache - private methods 90 // 91 92 void GrowableCache::recache() { 93 int len = _elements->length(); 94 95 FREE_C_HEAP_ARRAY(address, _cache); 96 _cache = NEW_C_HEAP_ARRAY(address,len+1); 97 98 for (int i=0; i<len; i++) { 99 _cache[i] = _elements->at(i)->getCacheValue(); 100 // 101 // The cache entry has gone bad. Without a valid frame pointer 102 // value, the entry is useless so we simply delete it in product 103 // mode. The call to remove() will rebuild the cache again 104 // without the bad entry. 105 // 106 if (_cache[i] == NULL) { 107 assert(false, "cannot recache NULL elements"); 108 remove(i); 109 return; 110 } 111 } 112 _cache[len] = NULL; 113 114 _listener_fun(_this_obj,_cache); 115 } 116 117 bool GrowableCache::equals(void* v, GrowableElement *e2) { 118 GrowableElement *e1 = (GrowableElement *) v; 119 assert(e1 != NULL, "e1 != NULL"); 120 assert(e2 != NULL, "e2 != NULL"); 121 122 return e1->equals(e2); 123 } 124 125 // 126 // class GrowableCache - public methods 127 // 128 129 GrowableCache::GrowableCache() { 130 _this_obj = NULL; 131 _listener_fun = NULL; 132 _elements = NULL; 133 _cache = NULL; 134 } 135 136 GrowableCache::~GrowableCache() { 137 clear(); 138 delete _elements; 139 FREE_C_HEAP_ARRAY(address, _cache); 140 } 141 142 void GrowableCache::initialize(void *this_obj, void listener_fun(void *, address*) ) { 143 _this_obj = this_obj; 144 _listener_fun = listener_fun; 145 _elements = new (ResourceObj::C_HEAP) GrowableArray<GrowableElement*>(5,true); 146 recache(); 147 } 148 149 // number of elements in the collection 150 int GrowableCache::length() { 151 return _elements->length(); 152 } 153 154 // get the value of the index element in the collection 155 GrowableElement* GrowableCache::at(int index) { 156 GrowableElement *e = (GrowableElement *) _elements->at(index); 157 assert(e != NULL, "e != NULL"); 158 return e; 159 } 160 161 int GrowableCache::find(GrowableElement* e) { 162 return _elements->find(e, GrowableCache::equals); 163 } 164 165 // append a copy of the element to the end of the collection 166 void GrowableCache::append(GrowableElement* e) { 167 GrowableElement *new_e = e->clone(); 168 _elements->append(new_e); 169 recache(); 170 } 171 172 // insert a copy of the element using lessthan() 173 void GrowableCache::insert(GrowableElement* e) { 174 GrowableElement *new_e = e->clone(); 175 _elements->append(new_e); 176 177 int n = length()-2; 178 for (int i=n; i>=0; i--) { 179 GrowableElement *e1 = _elements->at(i); 180 GrowableElement *e2 = _elements->at(i+1); 181 if (e2->lessThan(e1)) { 182 _elements->at_put(i+1, e1); 183 _elements->at_put(i, e2); 184 } 185 } 186 187 recache(); 188 } 189 190 // remove the element at index 191 void GrowableCache::remove (int index) { 192 GrowableElement *e = _elements->at(index); 193 assert(e != NULL, "e != NULL"); 194 _elements->remove(e); 195 delete e; 196 recache(); 197 } 198 199 // clear out all elements, release all heap space and 200 // let our listener know that things have changed. 201 void GrowableCache::clear() { 202 int len = _elements->length(); 203 for (int i=0; i<len; i++) { 204 delete _elements->at(i); 205 } 206 _elements->clear(); 207 recache(); 208 } 209 210 void GrowableCache::oops_do(OopClosure* f) { 211 int len = _elements->length(); 212 for (int i=0; i<len; i++) { 213 GrowableElement *e = _elements->at(i); 214 e->oops_do(f); 215 } 216 } 217 218 void GrowableCache::gc_epilogue() { 219 int len = _elements->length(); 220 for (int i=0; i<len; i++) { 221 _cache[i] = _elements->at(i)->getCacheValue(); 222 } 223 } 224 225 // 226 // class JvmtiBreakpoint 227 // 228 229 JvmtiBreakpoint::JvmtiBreakpoint() { 230 _method = NULL; 231 _bci = 0; 232 #ifdef CHECK_UNHANDLED_OOPS 233 // This one is always allocated with new, but check it just in case. 234 Thread *thread = Thread::current(); 235 if (thread->is_in_stack((address)&_method)) { 236 thread->allow_unhandled_oop((oop*)&_method); 237 } 238 #endif // CHECK_UNHANDLED_OOPS 239 } 240 241 JvmtiBreakpoint::JvmtiBreakpoint(methodOop m_method, jlocation location) { 242 _method = m_method; 243 assert(_method != NULL, "_method != NULL"); 244 _bci = (int) location; 245 #ifdef CHECK_UNHANDLED_OOPS 246 // Could be allocated with new and wouldn't be on the unhandled oop list. 247 Thread *thread = Thread::current(); 248 if (thread->is_in_stack((address)&_method)) { 249 thread->allow_unhandled_oop(&_method); 250 } 251 #endif // CHECK_UNHANDLED_OOPS 252 253 assert(_bci >= 0, "_bci >= 0"); 254 } 255 256 void JvmtiBreakpoint::copy(JvmtiBreakpoint& bp) { 257 _method = bp._method; 258 _bci = bp._bci; 259 } 260 261 bool JvmtiBreakpoint::lessThan(JvmtiBreakpoint& bp) { 262 Unimplemented(); 263 return false; 264 } 265 266 bool JvmtiBreakpoint::equals(JvmtiBreakpoint& bp) { 267 return _method == bp._method 268 && _bci == bp._bci; 269 } 270 271 bool JvmtiBreakpoint::is_valid() { 272 return _method != NULL && 273 _bci >= 0; 274 } 275 276 address JvmtiBreakpoint::getBcp() { 277 return _method->bcp_from(_bci); 278 } 279 280 void JvmtiBreakpoint::each_method_version_do(method_action meth_act) { 281 ((methodOopDesc*)_method->*meth_act)(_bci); 282 283 // add/remove breakpoint to/from versions of the method that 284 // are EMCP. Directly or transitively obsolete methods are 285 // not saved in the PreviousVersionInfo. 286 Thread *thread = Thread::current(); 287 instanceKlassHandle ikh = instanceKlassHandle(thread, _method->method_holder()); 288 symbolOop m_name = _method->name(); 289 symbolOop m_signature = _method->signature(); 290 291 { 292 ResourceMark rm(thread); 293 // PreviousVersionInfo objects returned via PreviousVersionWalker 294 // contain a GrowableArray of handles. We have to clean up the 295 // GrowableArray _after_ the PreviousVersionWalker destructor 296 // has destroyed the handles. 297 { 298 // search previous versions if they exist 299 PreviousVersionWalker pvw((instanceKlass *)ikh()->klass_part()); 300 for (PreviousVersionInfo * pv_info = pvw.next_previous_version(); 301 pv_info != NULL; pv_info = pvw.next_previous_version()) { 302 GrowableArray<methodHandle>* methods = 303 pv_info->prev_EMCP_method_handles(); 304 305 if (methods == NULL) { 306 // We have run into a PreviousVersion generation where 307 // all methods were made obsolete during that generation's 308 // RedefineClasses() operation. At the time of that 309 // operation, all EMCP methods were flushed so we don't 310 // have to go back any further. 311 // 312 // A NULL methods array is different than an empty methods 313 // array. We cannot infer any optimizations about older 314 // generations from an empty methods array for the current 315 // generation. 316 break; 317 } 318 319 for (int i = methods->length() - 1; i >= 0; i--) { 320 methodHandle method = methods->at(i); 321 if (method->name() == m_name && method->signature() == m_signature) { 322 RC_TRACE(0x00000800, ("%sing breakpoint in %s(%s)", 323 meth_act == &methodOopDesc::set_breakpoint ? "sett" : "clear", 324 method->name()->as_C_string(), 325 method->signature()->as_C_string())); 326 assert(!method->is_obsolete(), "only EMCP methods here"); 327 328 ((methodOopDesc*)method()->*meth_act)(_bci); 329 break; 330 } 331 } 332 } 333 } // pvw is cleaned up 334 } // rm is cleaned up 335 } 336 337 void JvmtiBreakpoint::set() { 338 each_method_version_do(&methodOopDesc::set_breakpoint); 339 } 340 341 void JvmtiBreakpoint::clear() { 342 each_method_version_do(&methodOopDesc::clear_breakpoint); 343 } 344 345 void JvmtiBreakpoint::print() { 346 #ifndef PRODUCT 347 const char *class_name = (_method == NULL) ? "NULL" : _method->klass_name()->as_C_string(); 348 const char *method_name = (_method == NULL) ? "NULL" : _method->name()->as_C_string(); 349 350 tty->print("Breakpoint(%s,%s,%d,%p)",class_name, method_name, _bci, getBcp()); 351 #endif 352 } 353 354 355 // 356 // class VM_ChangeBreakpoints 357 // 358 // Modify the Breakpoints data structure at a safepoint 359 // 360 361 void VM_ChangeBreakpoints::doit() { 362 switch (_operation) { 363 case SET_BREAKPOINT: 364 _breakpoints->set_at_safepoint(*_bp); 365 break; 366 case CLEAR_BREAKPOINT: 367 _breakpoints->clear_at_safepoint(*_bp); 368 break; 369 case CLEAR_ALL_BREAKPOINT: 370 _breakpoints->clearall_at_safepoint(); 371 break; 372 default: 373 assert(false, "Unknown operation"); 374 } 375 } 376 377 void VM_ChangeBreakpoints::oops_do(OopClosure* f) { 378 // This operation keeps breakpoints alive 379 if (_breakpoints != NULL) { 380 _breakpoints->oops_do(f); 381 } 382 if (_bp != NULL) { 383 _bp->oops_do(f); 384 } 385 } 386 387 // 388 // class JvmtiBreakpoints 389 // 390 // a JVMTI internal collection of JvmtiBreakpoint 391 // 392 393 JvmtiBreakpoints::JvmtiBreakpoints(void listener_fun(void *,address *)) { 394 _bps.initialize(this,listener_fun); 395 } 396 397 JvmtiBreakpoints:: ~JvmtiBreakpoints() {} 398 399 void JvmtiBreakpoints::oops_do(OopClosure* f) { 400 _bps.oops_do(f); 401 } 402 403 void JvmtiBreakpoints::gc_epilogue() { 404 _bps.gc_epilogue(); 405 } 406 407 void JvmtiBreakpoints::print() { 408 #ifndef PRODUCT 409 ResourceMark rm; 410 411 int n = _bps.length(); 412 for (int i=0; i<n; i++) { 413 JvmtiBreakpoint& bp = _bps.at(i); 414 tty->print("%d: ", i); 415 bp.print(); 416 tty->print_cr(""); 417 } 418 #endif 419 } 420 421 422 void JvmtiBreakpoints::set_at_safepoint(JvmtiBreakpoint& bp) { 423 assert(SafepointSynchronize::is_at_safepoint(), "must be at safepoint"); 424 425 int i = _bps.find(bp); 426 if (i == -1) { 427 _bps.append(bp); 428 bp.set(); 429 } 430 } 431 432 void JvmtiBreakpoints::clear_at_safepoint(JvmtiBreakpoint& bp) { 433 assert(SafepointSynchronize::is_at_safepoint(), "must be at safepoint"); 434 435 int i = _bps.find(bp); 436 if (i != -1) { 437 _bps.remove(i); 438 bp.clear(); 439 } 440 } 441 442 void JvmtiBreakpoints::clearall_at_safepoint() { 443 assert(SafepointSynchronize::is_at_safepoint(), "must be at safepoint"); 444 445 int len = _bps.length(); 446 for (int i=0; i<len; i++) { 447 _bps.at(i).clear(); 448 } 449 _bps.clear(); 450 } 451 452 int JvmtiBreakpoints::length() { return _bps.length(); } 453 454 int JvmtiBreakpoints::set(JvmtiBreakpoint& bp) { 455 if ( _bps.find(bp) != -1) { 456 return JVMTI_ERROR_DUPLICATE; 457 } 458 VM_ChangeBreakpoints set_breakpoint(this,VM_ChangeBreakpoints::SET_BREAKPOINT, &bp); 459 VMThread::execute(&set_breakpoint); 460 return JVMTI_ERROR_NONE; 461 } 462 463 int JvmtiBreakpoints::clear(JvmtiBreakpoint& bp) { 464 if ( _bps.find(bp) == -1) { 465 return JVMTI_ERROR_NOT_FOUND; 466 } 467 468 VM_ChangeBreakpoints clear_breakpoint(this,VM_ChangeBreakpoints::CLEAR_BREAKPOINT, &bp); 469 VMThread::execute(&clear_breakpoint); 470 return JVMTI_ERROR_NONE; 471 } 472 473 void JvmtiBreakpoints::clearall_in_class_at_safepoint(klassOop klass) { 474 bool changed = true; 475 // We are going to run thru the list of bkpts 476 // and delete some. This deletion probably alters 477 // the list in some implementation defined way such 478 // that when we delete entry i, the next entry might 479 // no longer be at i+1. To be safe, each time we delete 480 // an entry, we'll just start again from the beginning. 481 // We'll stop when we make a pass thru the whole list without 482 // deleting anything. 483 while (changed) { 484 int len = _bps.length(); 485 changed = false; 486 for (int i = 0; i < len; i++) { 487 JvmtiBreakpoint& bp = _bps.at(i); 488 if (bp.method()->method_holder() == klass) { 489 bp.clear(); 490 _bps.remove(i); 491 // This changed 'i' so we have to start over. 492 changed = true; 493 break; 494 } 495 } 496 } 497 } 498 499 void JvmtiBreakpoints::clearall() { 500 VM_ChangeBreakpoints clearall_breakpoint(this,VM_ChangeBreakpoints::CLEAR_ALL_BREAKPOINT); 501 VMThread::execute(&clearall_breakpoint); 502 } 503 504 // 505 // class JvmtiCurrentBreakpoints 506 // 507 508 JvmtiBreakpoints *JvmtiCurrentBreakpoints::_jvmti_breakpoints = NULL; 509 address * JvmtiCurrentBreakpoints::_breakpoint_list = NULL; 510 511 512 JvmtiBreakpoints& JvmtiCurrentBreakpoints::get_jvmti_breakpoints() { 513 if (_jvmti_breakpoints != NULL) return (*_jvmti_breakpoints); 514 _jvmti_breakpoints = new JvmtiBreakpoints(listener_fun); 515 assert(_jvmti_breakpoints != NULL, "_jvmti_breakpoints != NULL"); 516 return (*_jvmti_breakpoints); 517 } 518 519 void JvmtiCurrentBreakpoints::listener_fun(void *this_obj, address *cache) { 520 JvmtiBreakpoints *this_jvmti = (JvmtiBreakpoints *) this_obj; 521 assert(this_jvmti != NULL, "this_jvmti != NULL"); 522 523 debug_only(int n = this_jvmti->length();); 524 assert(cache[n] == NULL, "cache must be NULL terminated"); 525 526 set_breakpoint_list(cache); 527 } 528 529 530 void JvmtiCurrentBreakpoints::oops_do(OopClosure* f) { 531 if (_jvmti_breakpoints != NULL) { 532 _jvmti_breakpoints->oops_do(f); 533 } 534 } 535 536 void JvmtiCurrentBreakpoints::gc_epilogue() { 537 if (_jvmti_breakpoints != NULL) { 538 _jvmti_breakpoints->gc_epilogue(); 539 } 540 } 541 542 /////////////////////////////////////////////////////////////// 543 // 544 // class VM_GetOrSetLocal 545 // 546 547 // Constructor for non-object getter 548 VM_GetOrSetLocal::VM_GetOrSetLocal(JavaThread* thread, jint depth, int index, BasicType type) 549 : _thread(thread) 550 , _calling_thread(NULL) 551 , _depth(depth) 552 , _index(index) 553 , _type(type) 554 , _set(false) 555 , _jvf(NULL) 556 , _result(JVMTI_ERROR_NONE) 557 { 558 } 559 560 // Constructor for object or non-object setter 561 VM_GetOrSetLocal::VM_GetOrSetLocal(JavaThread* thread, jint depth, int index, BasicType type, jvalue value) 562 : _thread(thread) 563 , _calling_thread(NULL) 564 , _depth(depth) 565 , _index(index) 566 , _type(type) 567 , _value(value) 568 , _set(true) 569 , _jvf(NULL) 570 , _result(JVMTI_ERROR_NONE) 571 { 572 } 573 574 // Constructor for object getter 575 VM_GetOrSetLocal::VM_GetOrSetLocal(JavaThread* thread, JavaThread* calling_thread, jint depth, int index) 576 : _thread(thread) 577 , _calling_thread(calling_thread) 578 , _depth(depth) 579 , _index(index) 580 , _type(T_OBJECT) 581 , _set(false) 582 , _jvf(NULL) 583 , _result(JVMTI_ERROR_NONE) 584 { 585 } 586 587 vframe *VM_GetOrSetLocal::get_vframe() { 588 if (!_thread->has_last_Java_frame()) { 589 return NULL; 590 } 591 RegisterMap reg_map(_thread); 592 vframe *vf = _thread->last_java_vframe(®_map); 593 int d = 0; 594 while ((vf != NULL) && (d < _depth)) { 595 vf = vf->java_sender(); 596 d++; 597 } 598 return vf; 599 } 600 601 javaVFrame *VM_GetOrSetLocal::get_java_vframe() { 602 vframe* vf = get_vframe(); 603 if (vf == NULL) { 604 _result = JVMTI_ERROR_NO_MORE_FRAMES; 605 return NULL; 606 } 607 javaVFrame *jvf = (javaVFrame*)vf; 608 609 if (!vf->is_java_frame()) { 610 _result = JVMTI_ERROR_OPAQUE_FRAME; 611 return NULL; 612 } 613 return jvf; 614 } 615 616 // Check that the klass is assignable to a type with the given signature. 617 // Another solution could be to use the function Klass::is_subtype_of(type). 618 // But the type class can be forced to load/initialize eagerly in such a case. 619 // This may cause unexpected consequences like CFLH or class-init JVMTI events. 620 // It is better to avoid such a behavior. 621 bool VM_GetOrSetLocal::is_assignable(const char* ty_sign, Klass* klass, Thread* thread) { 622 assert(ty_sign != NULL, "type signature must not be NULL"); 623 assert(thread != NULL, "thread must not be NULL"); 624 assert(klass != NULL, "klass must not be NULL"); 625 626 int len = (int) strlen(ty_sign); 627 if (ty_sign[0] == 'L' && ty_sign[len-1] == ';') { // Need pure class/interface name 628 ty_sign++; 629 len -= 2; 630 } 631 symbolHandle ty_sym = oopFactory::new_symbol_handle(ty_sign, len, thread); 632 if (klass->name() == ty_sym()) { 633 return true; 634 } 635 // Compare primary supers 636 int super_depth = klass->super_depth(); 637 int idx; 638 for (idx = 0; idx < super_depth; idx++) { 639 if (Klass::cast(klass->primary_super_of_depth(idx))->name() == ty_sym()) { 640 return true; 641 } 642 } 643 // Compare secondary supers 644 objArrayOop sec_supers = klass->secondary_supers(); 645 for (idx = 0; idx < sec_supers->length(); idx++) { 646 if (Klass::cast((klassOop) sec_supers->obj_at(idx))->name() == ty_sym()) { 647 return true; 648 } 649 } 650 return false; 651 } 652 653 // Checks error conditions: 654 // JVMTI_ERROR_INVALID_SLOT 655 // JVMTI_ERROR_TYPE_MISMATCH 656 // Returns: 'true' - everything is Ok, 'false' - error code 657 658 bool VM_GetOrSetLocal::check_slot_type(javaVFrame* jvf) { 659 methodOop method_oop = jvf->method(); 660 if (!method_oop->has_localvariable_table()) { 661 // Just to check index boundaries 662 jint extra_slot = (_type == T_LONG || _type == T_DOUBLE) ? 1 : 0; 663 if (_index < 0 || _index + extra_slot >= method_oop->max_locals()) { 664 _result = JVMTI_ERROR_INVALID_SLOT; 665 return false; 666 } 667 return true; 668 } 669 670 jint num_entries = method_oop->localvariable_table_length(); 671 if (num_entries == 0) { 672 _result = JVMTI_ERROR_INVALID_SLOT; 673 return false; // There are no slots 674 } 675 int signature_idx = -1; 676 int vf_bci = jvf->bci(); 677 LocalVariableTableElement* table = method_oop->localvariable_table_start(); 678 for (int i = 0; i < num_entries; i++) { 679 int start_bci = table[i].start_bci; 680 int end_bci = start_bci + table[i].length; 681 682 // Here we assume that locations of LVT entries 683 // with the same slot number cannot be overlapped 684 if (_index == (jint) table[i].slot && start_bci <= vf_bci && vf_bci <= end_bci) { 685 signature_idx = (int) table[i].descriptor_cp_index; 686 break; 687 } 688 } 689 if (signature_idx == -1) { 690 _result = JVMTI_ERROR_INVALID_SLOT; 691 return false; // Incorrect slot index 692 } 693 symbolOop sign_sym = method_oop->constants()->symbol_at(signature_idx); 694 const char* signature = (const char *) sign_sym->as_utf8(); 695 BasicType slot_type = char2type(signature[0]); 696 697 switch (slot_type) { 698 case T_BYTE: 699 case T_SHORT: 700 case T_CHAR: 701 case T_BOOLEAN: 702 slot_type = T_INT; 703 break; 704 case T_ARRAY: 705 slot_type = T_OBJECT; 706 break; 707 }; 708 if (_type != slot_type) { 709 _result = JVMTI_ERROR_TYPE_MISMATCH; 710 return false; 711 } 712 713 jobject jobj = _value.l; 714 if (_set && slot_type == T_OBJECT && jobj != NULL) { // NULL reference is allowed 715 // Check that the jobject class matches the return type signature. 716 JavaThread* cur_thread = JavaThread::current(); 717 HandleMark hm(cur_thread); 718 719 Handle obj = Handle(cur_thread, JNIHandles::resolve_external_guard(jobj)); 720 NULL_CHECK(obj, (_result = JVMTI_ERROR_INVALID_OBJECT, false)); 721 KlassHandle ob_kh = KlassHandle(cur_thread, obj->klass()); 722 NULL_CHECK(ob_kh, (_result = JVMTI_ERROR_INVALID_OBJECT, false)); 723 724 if (!is_assignable(signature, Klass::cast(ob_kh()), cur_thread)) { 725 _result = JVMTI_ERROR_TYPE_MISMATCH; 726 return false; 727 } 728 } 729 return true; 730 } 731 732 static bool can_be_deoptimized(vframe* vf) { 733 return (vf->is_compiled_frame() && vf->fr().can_be_deoptimized()); 734 } 735 736 bool VM_GetOrSetLocal::doit_prologue() { 737 _jvf = get_java_vframe(); 738 NULL_CHECK(_jvf, false); 739 740 if (_jvf->method()->is_native()) { 741 if (getting_receiver() && !_jvf->method()->is_static()) { 742 return true; 743 } else { 744 _result = JVMTI_ERROR_OPAQUE_FRAME; 745 return false; 746 } 747 } 748 749 if (!check_slot_type(_jvf)) { 750 return false; 751 } 752 return true; 753 } 754 755 void VM_GetOrSetLocal::doit() { 756 if (_set) { 757 // Force deoptimization of frame if compiled because it's 758 // possible the compiler emitted some locals as constant values, 759 // meaning they are not mutable. 760 if (can_be_deoptimized(_jvf)) { 761 762 // Schedule deoptimization so that eventually the local 763 // update will be written to an interpreter frame. 764 Deoptimization::deoptimize_frame(_jvf->thread(), _jvf->fr().id()); 765 766 // Now store a new value for the local which will be applied 767 // once deoptimization occurs. Note however that while this 768 // write is deferred until deoptimization actually happens 769 // can vframe created after this point will have its locals 770 // reflecting this update so as far as anyone can see the 771 // write has already taken place. 772 773 // If we are updating an oop then get the oop from the handle 774 // since the handle will be long gone by the time the deopt 775 // happens. The oop stored in the deferred local will be 776 // gc'd on its own. 777 if (_type == T_OBJECT) { 778 _value.l = (jobject) (JNIHandles::resolve_external_guard(_value.l)); 779 } 780 // Re-read the vframe so we can see that it is deoptimized 781 // [ Only need because of assert in update_local() ] 782 _jvf = get_java_vframe(); 783 ((compiledVFrame*)_jvf)->update_local(_type, _index, _value); 784 return; 785 } 786 StackValueCollection *locals = _jvf->locals(); 787 HandleMark hm; 788 789 switch (_type) { 790 case T_INT: locals->set_int_at (_index, _value.i); break; 791 case T_LONG: locals->set_long_at (_index, _value.j); break; 792 case T_FLOAT: locals->set_float_at (_index, _value.f); break; 793 case T_DOUBLE: locals->set_double_at(_index, _value.d); break; 794 case T_OBJECT: { 795 Handle ob_h(JNIHandles::resolve_external_guard(_value.l)); 796 locals->set_obj_at (_index, ob_h); 797 break; 798 } 799 default: ShouldNotReachHere(); 800 } 801 _jvf->set_locals(locals); 802 } else { 803 if (_jvf->method()->is_native() && _jvf->is_compiled_frame()) { 804 assert(getting_receiver(), "Can only get here when getting receiver"); 805 oop receiver = _jvf->fr().get_native_receiver(); 806 _value.l = JNIHandles::make_local(_calling_thread, receiver); 807 } else { 808 StackValueCollection *locals = _jvf->locals(); 809 810 if (locals->at(_index)->type() == T_CONFLICT) { 811 memset(&_value, 0, sizeof(_value)); 812 _value.l = NULL; 813 return; 814 } 815 816 switch (_type) { 817 case T_INT: _value.i = locals->int_at (_index); break; 818 case T_LONG: _value.j = locals->long_at (_index); break; 819 case T_FLOAT: _value.f = locals->float_at (_index); break; 820 case T_DOUBLE: _value.d = locals->double_at(_index); break; 821 case T_OBJECT: { 822 // Wrap the oop to be returned in a local JNI handle since 823 // oops_do() no longer applies after doit() is finished. 824 oop obj = locals->obj_at(_index)(); 825 _value.l = JNIHandles::make_local(_calling_thread, obj); 826 break; 827 } 828 default: ShouldNotReachHere(); 829 } 830 } 831 } 832 } 833 834 835 bool VM_GetOrSetLocal::allow_nested_vm_operations() const { 836 return true; // May need to deoptimize 837 } 838 839 840 VM_GetReceiver::VM_GetReceiver( 841 JavaThread* thread, JavaThread* caller_thread, jint depth) 842 : VM_GetOrSetLocal(thread, caller_thread, depth, 0) {} 843 844 ///////////////////////////////////////////////////////////////////////////////////////// 845 846 // 847 // class JvmtiSuspendControl - see comments in jvmtiImpl.hpp 848 // 849 850 bool JvmtiSuspendControl::suspend(JavaThread *java_thread) { 851 // external suspend should have caught suspending a thread twice 852 853 // Immediate suspension required for JPDA back-end so JVMTI agent threads do 854 // not deadlock due to later suspension on transitions while holding 855 // raw monitors. Passing true causes the immediate suspension. 856 // java_suspend() will catch threads in the process of exiting 857 // and will ignore them. 858 java_thread->java_suspend(); 859 860 // It would be nice to have the following assertion in all the time, 861 // but it is possible for a racing resume request to have resumed 862 // this thread right after we suspended it. Temporarily enable this 863 // assertion if you are chasing a different kind of bug. 864 // 865 // assert(java_lang_Thread::thread(java_thread->threadObj()) == NULL || 866 // java_thread->is_being_ext_suspended(), "thread is not suspended"); 867 868 if (java_lang_Thread::thread(java_thread->threadObj()) == NULL) { 869 // check again because we can get delayed in java_suspend(): 870 // the thread is in process of exiting. 871 return false; 872 } 873 874 return true; 875 } 876 877 bool JvmtiSuspendControl::resume(JavaThread *java_thread) { 878 // external suspend should have caught resuming a thread twice 879 assert(java_thread->is_being_ext_suspended(), "thread should be suspended"); 880 881 // resume thread 882 { 883 // must always grab Threads_lock, see JVM_SuspendThread 884 MutexLocker ml(Threads_lock); 885 java_thread->java_resume(); 886 } 887 888 return true; 889 } 890 891 892 void JvmtiSuspendControl::print() { 893 #ifndef PRODUCT 894 MutexLocker mu(Threads_lock); 895 ResourceMark rm; 896 897 tty->print("Suspended Threads: ["); 898 for (JavaThread *thread = Threads::first(); thread != NULL; thread = thread->next()) { 899 #if JVMTI_TRACE 900 const char *name = JvmtiTrace::safe_get_thread_name(thread); 901 #else 902 const char *name = ""; 903 #endif /*JVMTI_TRACE */ 904 tty->print("%s(%c ", name, thread->is_being_ext_suspended() ? 'S' : '_'); 905 if (!thread->has_last_Java_frame()) { 906 tty->print("no stack"); 907 } 908 tty->print(") "); 909 } 910 tty->print_cr("]"); 911 #endif 912 } 913 914 #ifndef KERNEL 915 916 void JvmtiDeferredEvent::post() { 917 switch(_type) { 918 case TYPE_COMPILED_METHOD_LOAD: 919 JvmtiExport::post_compiled_method_load(compiled_method_loaded()); 920 break; 921 default: 922 ShouldNotReachHere(); 923 } 924 } 925 926 JvmtiDeferredEventQueue::QueueNode* JvmtiDeferredEventQueue::_free_list = NULL; 927 JvmtiDeferredEventQueue::QueueNode* JvmtiDeferredEventQueue::_queue = NULL; 928 929 bool JvmtiDeferredEventQueue::has_events() { 930 return _queue != NULL; 931 } 932 933 void JvmtiDeferredEventQueue::enqueue(const JvmtiDeferredEvent& event) { 934 assert(Service_lock->owned_by_self(), "Must own Service_lock"); 935 936 QueueNode* node = _free_list; 937 if (node == NULL) { 938 node = new QueueNode(); 939 } else { 940 _free_list = node->next(); 941 } 942 943 node->set_event(event); 944 node->set_next(_queue); 945 946 _queue = node; 947 948 Service_lock->notify_all(); 949 } 950 951 JvmtiDeferredEvent JvmtiDeferredEventQueue::dequeue() { 952 assert(Service_lock->owned_by_self(), "Must own Service_lock"); 953 assert(_queue != NULL, "Nothing to dequeue"); 954 955 if (_queue == NULL) { 956 // Just in case this happens in product; it shouldn't be let's not crash 957 return JvmtiDeferredEvent::null_event(); 958 } 959 960 // Events need to be processed in FIFO order, so take the last event. 961 QueueNode* node = _queue->next(); 962 if (node == NULL) { 963 // Only one item in the queue 964 node = _queue; 965 _queue = NULL; 966 } else { 967 QueueNode* prev = _queue; 968 while (node->next() != NULL) { 969 prev = node; 970 node = node->next(); 971 } 972 prev->set_next(NULL); 973 } 974 975 JvmtiDeferredEvent event = node->event(); 976 977 // Stick the node on the freelist for reuse. 978 node->set_next(_free_list); 979 _free_list = node; 980 981 return event; 982 } 983 984 #endif // ndef KERNEL