1 /* 2 * Copyright (c) 1997, 2018, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "classfile/javaClasses.inline.hpp" 27 #include "code/codeCache.hpp" 28 #include "code/debugInfoRec.hpp" 29 #include "code/nmethod.hpp" 30 #include "code/pcDesc.hpp" 31 #include "code/scopeDesc.hpp" 32 #include "interpreter/interpreter.hpp" 33 #include "interpreter/oopMapCache.hpp" 34 #include "oops/instanceKlass.hpp" 35 #include "oops/oop.inline.hpp" 36 #include "runtime/basicLock.hpp" 37 #include "runtime/frame.inline.hpp" 38 #include "runtime/handles.inline.hpp" 39 #include "runtime/monitorChunk.hpp" 40 #include "runtime/signature.hpp" 41 #include "runtime/stubRoutines.hpp" 42 #include "runtime/vframeArray.hpp" 43 #include "runtime/vframe_hp.hpp" 44 #ifdef COMPILER2 45 #include "opto/matcher.hpp" 46 #endif 47 48 49 // ------------- compiledVFrame -------------- 50 51 StackValueCollection* compiledVFrame::locals() const { 52 // Natives has no scope 53 if (scope() == NULL) return new StackValueCollection(0); 54 GrowableArray<ScopeValue*>* scv_list = scope()->locals(); 55 if (scv_list == NULL) return new StackValueCollection(0); 56 57 // scv_list is the list of ScopeValues describing the JVM stack state. 58 // There is one scv_list entry for every JVM stack state in use. 59 int length = scv_list->length(); 60 StackValueCollection* result = new StackValueCollection(length); 61 for (int i = 0; i < length; i++) { 62 result->add(create_stack_value(scv_list->at(i))); 63 } 64 65 // Replace the original values with any stores that have been 66 // performed through compiledVFrame::update_locals. 67 GrowableArray<jvmtiDeferredLocalVariableSet*>* list = thread()->deferred_locals(); 68 if (list != NULL ) { 69 // In real life this never happens or is typically a single element search 70 for (int i = 0; i < list->length(); i++) { 71 if (list->at(i)->matches(this)) { 72 list->at(i)->update_locals(result); 73 break; 74 } 75 } 76 } 77 78 return result; 79 } 80 81 82 void compiledVFrame::set_locals(StackValueCollection* values) const { 83 84 fatal("Should use update_local for each local update"); 85 } 86 87 void compiledVFrame::update_local(BasicType type, int index, jvalue value) { 88 assert(index >= 0 && index < method()->max_locals(), "out of bounds"); 89 update_deferred_value(type, index, value); 90 } 91 92 void compiledVFrame::update_stack(BasicType type, int index, jvalue value) { 93 assert(index >= 0 && index < method()->max_stack(), "out of bounds"); 94 update_deferred_value(type, index + method()->max_locals(), value); 95 } 96 97 void compiledVFrame::update_monitor(int index, MonitorInfo* val) { 98 assert(index >= 0, "out of bounds"); 99 jvalue value; 100 value.l = (jobject) val->owner(); 101 update_deferred_value(T_OBJECT, index + method()->max_locals() + method()->max_stack(), value); 102 } 103 104 void compiledVFrame::update_deferred_value(BasicType type, int index, jvalue value) { 105 assert(fr().is_deoptimized_frame(), "frame must be scheduled for deoptimization"); 106 GrowableArray<jvmtiDeferredLocalVariableSet*>* deferred = thread()->deferred_locals(); 107 jvmtiDeferredLocalVariableSet* locals = NULL; 108 if (deferred != NULL ) { 109 // See if this vframe has already had locals with deferred writes 110 for (int f = 0; f < deferred->length(); f++ ) { 111 if (deferred->at(f)->matches(this)) { 112 locals = deferred->at(f); 113 break; 114 } 115 } 116 // No matching vframe must push a new vframe 117 } else { 118 // No deferred updates pending for this thread. 119 // allocate in C heap 120 thread()->allocate_deferred_updates(); 121 deferred = thread()->deferred_locals(); 122 } 123 if (locals == NULL) { 124 locals = new jvmtiDeferredLocalVariableSet(method(), bci(), fr().id(), vframe_id()); 125 deferred->push(locals); 126 assert(locals->id() == fr().id(), "Huh? Must match"); 127 } 128 locals->set_value_at(index, type, value); 129 } 130 131 StackValueCollection* compiledVFrame::expressions() const { 132 // Natives has no scope 133 if (scope() == NULL) return new StackValueCollection(0); 134 GrowableArray<ScopeValue*>* scv_list = scope()->expressions(); 135 if (scv_list == NULL) return new StackValueCollection(0); 136 137 // scv_list is the list of ScopeValues describing the JVM stack state. 138 // There is one scv_list entry for every JVM stack state in use. 139 int length = scv_list->length(); 140 StackValueCollection* result = new StackValueCollection(length); 141 for (int i = 0; i < length; i++) { 142 result->add(create_stack_value(scv_list->at(i))); 143 } 144 145 // Replace the original values with any stores that have been 146 // performed through compiledVFrame::update_stack. 147 GrowableArray<jvmtiDeferredLocalVariableSet*>* list = thread()->deferred_locals(); 148 if (list != NULL ) { 149 // In real life this never happens or is typically a single element search 150 for (int i = 0; i < list->length(); i++) { 151 if (list->at(i)->matches(this)) { 152 list->at(i)->update_stack(result); 153 break; 154 } 155 } 156 } 157 158 return result; 159 } 160 161 162 // The implementation of the following two methods was factorized into the 163 // class StackValue because it is also used from within deoptimization.cpp for 164 // rematerialization and relocking of non-escaping objects. 165 166 StackValue *compiledVFrame::create_stack_value(ScopeValue *sv) const { 167 return StackValue::create_stack_value(&_fr, register_map(), sv); 168 } 169 170 BasicLock* compiledVFrame::resolve_monitor_lock(Location location) const { 171 return StackValue::resolve_monitor_lock(&_fr, location); 172 } 173 174 175 GrowableArray<MonitorInfo*>* compiledVFrame::monitors() const { 176 // Natives has no scope 177 if (scope() == NULL) { 178 CompiledMethod* nm = code(); 179 Method* method = nm->method(); 180 assert(method->is_native() || nm->is_aot(), "Expect a native method or precompiled method"); 181 if (!method->is_synchronized()) { 182 return new GrowableArray<MonitorInfo*>(0); 183 } 184 // This monitor is really only needed for UseBiasedLocking, but 185 // return it in all cases for now as it might be useful for stack 186 // traces and tools as well 187 GrowableArray<MonitorInfo*> *monitors = new GrowableArray<MonitorInfo*>(1); 188 // Casting away const 189 frame& fr = (frame&) _fr; 190 MonitorInfo* info = new MonitorInfo( 191 fr.get_native_receiver(), fr.get_native_monitor(), false, false); 192 monitors->push(info); 193 return monitors; 194 } 195 GrowableArray<MonitorValue*>* monitors = scope()->monitors(); 196 if (monitors == NULL) { 197 return new GrowableArray<MonitorInfo*>(0); 198 } 199 GrowableArray<MonitorInfo*>* result = new GrowableArray<MonitorInfo*>(monitors->length()); 200 for (int index = 0; index < monitors->length(); index++) { 201 MonitorValue* mv = monitors->at(index); 202 ScopeValue* ov = mv->owner(); 203 StackValue *owner_sv = create_stack_value(ov); // it is an oop 204 if (ov->is_object() && owner_sv->obj_is_scalar_replaced()) { // The owner object was scalar replaced 205 assert(mv->eliminated(), "monitor should be eliminated for scalar replaced object"); 206 // Put klass for scalar replaced object. 207 ScopeValue* kv = ((ObjectValue *)ov)->klass(); 208 assert(kv->is_constant_oop(), "klass should be oop constant for scalar replaced object"); 209 Handle k(Thread::current(), ((ConstantOopReadValue*)kv)->value()()); 210 assert(java_lang_Class::is_instance(k()), "must be"); 211 result->push(new MonitorInfo(k(), resolve_monitor_lock(mv->basic_lock()), 212 mv->eliminated(), true)); 213 } else { 214 result->push(new MonitorInfo(owner_sv->get_obj()(), resolve_monitor_lock(mv->basic_lock()), 215 mv->eliminated(), false)); 216 } 217 } 218 219 // Replace the original values with any stores that have been 220 // performed through compiledVFrame::update_monitors. 221 GrowableArray<jvmtiDeferredLocalVariableSet*>* list = thread()->deferred_locals(); 222 if (list != NULL ) { 223 // In real life this never happens or is typically a single element search 224 for (int i = 0; i < list->length(); i++) { 225 if (list->at(i)->matches(this)) { 226 list->at(i)->update_monitors(result); 227 break; 228 } 229 } 230 } 231 232 return result; 233 } 234 235 236 compiledVFrame::compiledVFrame(const frame* fr, const RegisterMap* reg_map, JavaThread* thread, CompiledMethod* nm) 237 : javaVFrame(fr, reg_map, thread) { 238 _scope = NULL; 239 _vframe_id = 0; 240 // Compiled method (native stub or Java code) 241 // native wrappers have no scope data, it is implied 242 if (!nm->is_compiled() || !nm->as_compiled_method()->is_native_method()) { 243 _scope = nm->scope_desc_at(_fr.pc()); 244 } 245 } 246 247 compiledVFrame::compiledVFrame(const frame* fr, const RegisterMap* reg_map, JavaThread* thread, ScopeDesc* scope, int vframe_id) 248 : javaVFrame(fr, reg_map, thread) { 249 _scope = scope; 250 _vframe_id = vframe_id; 251 guarantee(_scope != NULL, "scope must be present"); 252 } 253 254 compiledVFrame* compiledVFrame::at_scope(int decode_offset, int vframe_id) { 255 if (scope()->decode_offset() != decode_offset) { 256 ScopeDesc* scope = this->scope()->at_offset(decode_offset); 257 return new compiledVFrame(frame_pointer(), register_map(), thread(), scope, vframe_id); 258 } 259 assert(_vframe_id == vframe_id, "wrong frame id"); 260 return this; 261 } 262 263 bool compiledVFrame::is_top() const { 264 // FIX IT: Remove this when new native stubs are in place 265 if (scope() == NULL) return true; 266 return scope()->is_top(); 267 } 268 269 270 CompiledMethod* compiledVFrame::code() const { 271 return CodeCache::find_compiled(_fr.pc()); 272 } 273 274 275 Method* compiledVFrame::method() const { 276 if (scope() == NULL) { 277 // native nmethods have no scope the method is implied 278 nmethod* nm = code()->as_nmethod(); 279 assert(nm->is_native_method(), "must be native"); 280 return nm->method(); 281 } 282 return scope()->method(); 283 } 284 285 286 int compiledVFrame::bci() const { 287 int raw = raw_bci(); 288 return raw == SynchronizationEntryBCI ? 0 : raw; 289 } 290 291 292 int compiledVFrame::raw_bci() const { 293 if (scope() == NULL) { 294 // native nmethods have no scope the method/bci is implied 295 nmethod* nm = code()->as_nmethod(); 296 assert(nm->is_native_method(), "must be native"); 297 return 0; 298 } 299 return scope()->bci(); 300 } 301 302 bool compiledVFrame::should_reexecute() const { 303 if (scope() == NULL) { 304 // native nmethods have no scope the method/bci is implied 305 nmethod* nm = code()->as_nmethod(); 306 assert(nm->is_native_method(), "must be native"); 307 return false; 308 } 309 return scope()->should_reexecute(); 310 } 311 312 bool compiledVFrame::not_global_escape_in_scope() const { 313 if (scope() == NULL) { 314 // native nmethod, all objs escape 315 nmethod* nm = code()->as_nmethod(); 316 assert(nm->is_native_method(), "must be native"); 317 return false; 318 } 319 return (scope()->objects() != NULL) || scope()->not_global_escape_in_scope(); 320 } 321 322 bool compiledVFrame::arg_escape() const { 323 if (scope() == NULL) { 324 // native nmethod, all objs escape 325 nmethod* nm = code()->as_nmethod(); 326 assert(nm->is_native_method(), "must be native"); 327 return false; 328 } 329 return scope()->arg_escape(); 330 } 331 332 vframe* compiledVFrame::sender() const { 333 const frame f = fr(); 334 if (scope() == NULL) { 335 // native nmethods have no scope the method/bci is implied 336 nmethod* nm = code()->as_nmethod(); 337 assert(nm->is_native_method(), "must be native"); 338 return vframe::sender(); 339 } else { 340 return scope()->is_top() 341 ? vframe::sender() 342 : new compiledVFrame(&f, register_map(), thread(), scope()->sender(), vframe_id() + 1); 343 } 344 } 345 346 jvmtiDeferredLocalVariableSet::jvmtiDeferredLocalVariableSet(Method* method, int bci, intptr_t* id, int vframe_id) { 347 _method = method; 348 _bci = bci; 349 _id = id; 350 _vframe_id = vframe_id; 351 // Alway will need at least one, must be on C heap 352 _locals = new(ResourceObj::C_HEAP, mtCompiler) GrowableArray<jvmtiDeferredLocalVariable*> (1, true); 353 _objects_are_deoptimized = false; 354 } 355 356 jvmtiDeferredLocalVariableSet::~jvmtiDeferredLocalVariableSet() { 357 for (int i = 0; i < _locals->length(); i++ ) { 358 delete _locals->at(i); 359 } 360 // Free growableArray and c heap for elements 361 delete _locals; 362 } 363 364 bool jvmtiDeferredLocalVariableSet::matches(const vframe* vf) { 365 if (!vf->is_compiled_frame()) return false; 366 compiledVFrame* cvf = (compiledVFrame*)vf; 367 if (cvf->fr().id() == id() && cvf->vframe_id() == vframe_id()) { 368 assert(cvf->method() == method() && cvf->bci() == bci(), "must agree"); 369 return true; 370 } 371 return false; 372 } 373 374 void jvmtiDeferredLocalVariableSet::set_value_at(int idx, BasicType type, jvalue val) { 375 for (int i = 0; i < _locals->length(); i++) { 376 if (_locals->at(i)->index() == idx) { 377 assert(_locals->at(i)->type() == type, "Wrong type"); 378 _locals->at(i)->set_value(val); 379 return; 380 } 381 } 382 _locals->push(new jvmtiDeferredLocalVariable(idx, type, val)); 383 } 384 385 void jvmtiDeferredLocalVariableSet::update_value(StackValueCollection* locals, BasicType type, int index, jvalue value) { 386 switch (type) { 387 case T_BOOLEAN: 388 locals->set_int_at(index, value.z); 389 break; 390 case T_CHAR: 391 locals->set_int_at(index, value.c); 392 break; 393 case T_FLOAT: 394 locals->set_float_at(index, value.f); 395 break; 396 case T_DOUBLE: 397 locals->set_double_at(index, value.d); 398 break; 399 case T_BYTE: 400 locals->set_int_at(index, value.b); 401 break; 402 case T_SHORT: 403 locals->set_int_at(index, value.s); 404 break; 405 case T_INT: 406 locals->set_int_at(index, value.i); 407 break; 408 case T_LONG: 409 locals->set_long_at(index, value.j); 410 break; 411 case T_OBJECT: 412 { 413 Handle obj(Thread::current(), (oop)value.l); 414 locals->set_obj_at(index, obj); 415 } 416 break; 417 default: 418 ShouldNotReachHere(); 419 } 420 } 421 422 void jvmtiDeferredLocalVariableSet::update_locals(StackValueCollection* locals) { 423 for (int l = 0; l < _locals->length(); l ++) { 424 jvmtiDeferredLocalVariable* val = _locals->at(l); 425 if (val->index() >= 0 && val->index() < method()->max_locals()) { 426 update_value(locals, val->type(), val->index(), val->value()); 427 } 428 } 429 } 430 431 432 void jvmtiDeferredLocalVariableSet::update_stack(StackValueCollection* expressions) { 433 for (int l = 0; l < _locals->length(); l ++) { 434 jvmtiDeferredLocalVariable* val = _locals->at(l); 435 if (val->index() >= method()->max_locals() && val->index() < method()->max_locals() + method()->max_stack()) { 436 update_value(expressions, val->type(), val->index() - method()->max_locals(), val->value()); 437 } 438 } 439 } 440 441 442 void jvmtiDeferredLocalVariableSet::update_monitors(GrowableArray<MonitorInfo*>* monitors) { 443 for (int l = 0; l < _locals->length(); l ++) { 444 jvmtiDeferredLocalVariable* val = _locals->at(l); 445 if (val->index() >= method()->max_locals() + method()->max_stack()) { 446 int lock_index = val->index() - (method()->max_locals() + method()->max_stack()); 447 MonitorInfo* info = monitors->at(lock_index); 448 MonitorInfo* new_info = new MonitorInfo((oopDesc*)val->value().l, info->lock(), info->eliminated(), false); 449 monitors->at_put(lock_index, new_info); 450 } 451 } 452 } 453 454 455 void jvmtiDeferredLocalVariableSet::oops_do(OopClosure* f) { 456 // The Method* is on the stack so a live activation keeps it alive 457 // either by mirror in interpreter or code in compiled code. 458 for (int i = 0; i < _locals->length(); i++) { 459 if (_locals->at(i)->type() == T_OBJECT) { 460 f->do_oop(_locals->at(i)->oop_addr()); 461 } 462 } 463 } 464 465 jvmtiDeferredLocalVariable::jvmtiDeferredLocalVariable(int index, BasicType type, jvalue value) { 466 _index = index; 467 _type = type; 468 _value = value; 469 } 470 471 472 #ifndef PRODUCT 473 void compiledVFrame::verify() const { 474 Unimplemented(); 475 } 476 #endif // PRODUCT