1 /* 2 * Copyright (c) 2012, 2015, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 27 #include <new> 28 29 #include "classfile/classLoaderData.hpp" 30 #include "classfile/stringTable.hpp" 31 #include "code/codeCache.hpp" 32 #include "jvmtifiles/jvmtiEnv.hpp" 33 #include "memory/metadataFactory.hpp" 34 #include "memory/universe.hpp" 35 #include "oops/oop.inline.hpp" 36 #include "prims/wbtestmethods/parserTests.hpp" 37 #include "prims/whitebox.hpp" 38 #include "runtime/arguments.hpp" 39 #include "runtime/compilationPolicy.hpp" 40 #include "runtime/deoptimization.hpp" 41 #include "runtime/interfaceSupport.hpp" 42 #include "runtime/os.hpp" 43 #include "runtime/sweeper.hpp" 44 #include "runtime/javaCalls.hpp" 45 #include "runtime/thread.hpp" 46 #include "runtime/vm_version.hpp" 47 #include "utilities/array.hpp" 48 #include "utilities/debug.hpp" 49 #include "utilities/exceptions.hpp" 50 #include "utilities/macros.hpp" 51 #if INCLUDE_ALL_GCS 52 #include "gc_implementation/parallelScavenge/parallelScavengeHeap.inline.hpp" 53 #include "gc_implementation/g1/concurrentMark.hpp" 54 #include "gc_implementation/g1/concurrentMarkThread.hpp" 55 #include "gc_implementation/g1/g1CollectedHeap.inline.hpp" 56 #include "gc_implementation/g1/heapRegionRemSet.hpp" 57 #endif // INCLUDE_ALL_GCS 58 #if INCLUDE_NMT 59 #include "services/mallocSiteTable.hpp" 60 #include "services/memTracker.hpp" 61 #include "utilities/nativeCallStack.hpp" 62 #endif // INCLUDE_NMT 63 64 65 PRAGMA_FORMAT_MUTE_WARNINGS_FOR_GCC 66 67 #define SIZE_T_MAX_VALUE ((size_t) -1) 68 69 bool WhiteBox::_used = false; 70 volatile bool WhiteBox::compilation_locked = false; 71 72 class VM_WhiteBoxOperation : public VM_Operation { 73 public: 74 VM_WhiteBoxOperation() { } 75 VMOp_Type type() const { return VMOp_WhiteBoxOperation; } 76 bool allow_nested_vm_operations() const { return true; } 77 }; 78 79 80 WB_ENTRY(jlong, WB_GetObjectAddress(JNIEnv* env, jobject o, jobject obj)) 81 return (jlong)(void*)JNIHandles::resolve(obj); 82 WB_END 83 84 WB_ENTRY(jint, WB_GetHeapOopSize(JNIEnv* env, jobject o)) 85 return heapOopSize; 86 WB_END 87 88 WB_ENTRY(jint, WB_GetVMPageSize(JNIEnv* env, jobject o)) 89 return os::vm_page_size(); 90 WB_END 91 92 WB_ENTRY(jlong, WB_GetVMLargePageSize(JNIEnv* env, jobject o)) 93 return os::large_page_size(); 94 WB_END 95 96 class WBIsKlassAliveClosure : public KlassClosure { 97 Symbol* _name; 98 bool _found; 99 public: 100 WBIsKlassAliveClosure(Symbol* name) : _name(name), _found(false) {} 101 102 void do_klass(Klass* k) { 103 if (_found) return; 104 Symbol* ksym = k->name(); 105 if (ksym->fast_compare(_name) == 0) { 106 _found = true; 107 } 108 } 109 110 bool found() const { 111 return _found; 112 } 113 }; 114 115 WB_ENTRY(jboolean, WB_IsClassAlive(JNIEnv* env, jobject target, jstring name)) 116 Handle h_name = JNIHandles::resolve(name); 117 if (h_name.is_null()) return false; 118 Symbol* sym = java_lang_String::as_symbol(h_name, CHECK_false); 119 TempNewSymbol tsym(sym); // Make sure to decrement reference count on sym on return 120 121 WBIsKlassAliveClosure closure(sym); 122 ClassLoaderDataGraph::classes_do(&closure); 123 124 return closure.found(); 125 WB_END 126 127 WB_ENTRY(void, WB_AddToBootstrapClassLoaderSearch(JNIEnv* env, jobject o, jstring segment)) { 128 #if INCLUDE_JVMTI 129 ResourceMark rm; 130 const char* seg = java_lang_String::as_utf8_string(JNIHandles::resolve_non_null(segment)); 131 JvmtiEnv* jvmti_env = JvmtiEnv::create_a_jvmti(JVMTI_VERSION); 132 jvmtiError err = jvmti_env->AddToBootstrapClassLoaderSearch(seg); 133 assert(err == JVMTI_ERROR_NONE, "must not fail"); 134 #endif 135 } 136 WB_END 137 138 WB_ENTRY(void, WB_AddToSystemClassLoaderSearch(JNIEnv* env, jobject o, jstring segment)) { 139 #if INCLUDE_JVMTI 140 ResourceMark rm; 141 const char* seg = java_lang_String::as_utf8_string(JNIHandles::resolve_non_null(segment)); 142 JvmtiEnv* jvmti_env = JvmtiEnv::create_a_jvmti(JVMTI_VERSION); 143 jvmtiError err = jvmti_env->AddToSystemClassLoaderSearch(seg); 144 assert(err == JVMTI_ERROR_NONE, "must not fail"); 145 #endif 146 } 147 WB_END 148 149 150 WB_ENTRY(jlong, WB_GetCompressedOopsMaxHeapSize(JNIEnv* env, jobject o)) { 151 return (jlong)Arguments::max_heap_for_compressed_oops(); 152 } 153 WB_END 154 155 WB_ENTRY(void, WB_PrintHeapSizes(JNIEnv* env, jobject o)) { 156 CollectorPolicy * p = Universe::heap()->collector_policy(); 157 gclog_or_tty->print_cr("Minimum heap "SIZE_FORMAT" Initial heap " 158 SIZE_FORMAT" Maximum heap "SIZE_FORMAT" Space alignment "SIZE_FORMAT" Heap alignment "SIZE_FORMAT, 159 p->min_heap_byte_size(), p->initial_heap_byte_size(), p->max_heap_byte_size(), 160 p->space_alignment(), p->heap_alignment()); 161 } 162 WB_END 163 164 #ifndef PRODUCT 165 // Forward declaration 166 void TestReservedSpace_test(); 167 void TestReserveMemorySpecial_test(); 168 void TestVirtualSpace_test(); 169 void TestMetaspaceAux_test(); 170 #endif 171 172 WB_ENTRY(void, WB_RunMemoryUnitTests(JNIEnv* env, jobject o)) 173 #ifndef PRODUCT 174 TestReservedSpace_test(); 175 TestReserveMemorySpecial_test(); 176 TestVirtualSpace_test(); 177 TestMetaspaceAux_test(); 178 #endif 179 WB_END 180 181 WB_ENTRY(void, WB_ReadFromNoaccessArea(JNIEnv* env, jobject o)) 182 size_t granularity = os::vm_allocation_granularity(); 183 ReservedHeapSpace rhs(100 * granularity, granularity, false); 184 VirtualSpace vs; 185 vs.initialize(rhs, 50 * granularity); 186 187 // Check if constraints are complied 188 if (!( UseCompressedOops && rhs.base() != NULL && 189 Universe::narrow_oop_base() != NULL && 190 Universe::narrow_oop_use_implicit_null_checks() )) { 191 tty->print_cr("WB_ReadFromNoaccessArea method is useless:\n " 192 "\tUseCompressedOops is %d\n" 193 "\trhs.base() is "PTR_FORMAT"\n" 194 "\tUniverse::narrow_oop_base() is "PTR_FORMAT"\n" 195 "\tUniverse::narrow_oop_use_implicit_null_checks() is %d", 196 UseCompressedOops, 197 rhs.base(), 198 Universe::narrow_oop_base(), 199 Universe::narrow_oop_use_implicit_null_checks()); 200 return; 201 } 202 tty->print_cr("Reading from no access area... "); 203 tty->print_cr("*(vs.low_boundary() - rhs.noaccess_prefix() / 2 ) = %c", 204 *(vs.low_boundary() - rhs.noaccess_prefix() / 2 )); 205 WB_END 206 207 static jint wb_stress_virtual_space_resize(size_t reserved_space_size, 208 size_t magnitude, size_t iterations) { 209 size_t granularity = os::vm_allocation_granularity(); 210 ReservedHeapSpace rhs(reserved_space_size * granularity, granularity, false); 211 VirtualSpace vs; 212 if (!vs.initialize(rhs, 0)) { 213 tty->print_cr("Failed to initialize VirtualSpace. Can't proceed."); 214 return 3; 215 } 216 217 long seed = os::random(); 218 tty->print_cr("Random seed is %ld", seed); 219 os::init_random(seed); 220 221 for (size_t i = 0; i < iterations; i++) { 222 223 // Whether we will shrink or grow 224 bool shrink = os::random() % 2L == 0; 225 226 // Get random delta to resize virtual space 227 size_t delta = (size_t)os::random() % magnitude; 228 229 // If we are about to shrink virtual space below zero, then expand instead 230 if (shrink && vs.committed_size() < delta) { 231 shrink = false; 232 } 233 234 // Resizing by delta 235 if (shrink) { 236 vs.shrink_by(delta); 237 } else { 238 // If expanding fails expand_by will silently return false 239 vs.expand_by(delta, true); 240 } 241 } 242 return 0; 243 } 244 245 WB_ENTRY(jint, WB_StressVirtualSpaceResize(JNIEnv* env, jobject o, 246 jlong reserved_space_size, jlong magnitude, jlong iterations)) 247 tty->print_cr("reservedSpaceSize="JLONG_FORMAT", magnitude="JLONG_FORMAT", " 248 "iterations="JLONG_FORMAT"\n", reserved_space_size, magnitude, 249 iterations); 250 if (reserved_space_size < 0 || magnitude < 0 || iterations < 0) { 251 tty->print_cr("One of variables printed above is negative. Can't proceed.\n"); 252 return 1; 253 } 254 255 // sizeof(size_t) depends on whether OS is 32bit or 64bit. sizeof(jlong) is 256 // always 8 byte. That's why we should avoid overflow in case of 32bit platform. 257 if (sizeof(size_t) < sizeof(jlong)) { 258 jlong size_t_max_value = (jlong) SIZE_T_MAX_VALUE; 259 if (reserved_space_size > size_t_max_value || magnitude > size_t_max_value 260 || iterations > size_t_max_value) { 261 tty->print_cr("One of variables printed above overflows size_t. Can't proceed.\n"); 262 return 2; 263 } 264 } 265 266 return wb_stress_virtual_space_resize((size_t) reserved_space_size, 267 (size_t) magnitude, (size_t) iterations); 268 WB_END 269 270 WB_ENTRY(jboolean, WB_isObjectInOldGen(JNIEnv* env, jobject o, jobject obj)) 271 oop p = JNIHandles::resolve(obj); 272 #if INCLUDE_ALL_GCS 273 if (UseG1GC) { 274 G1CollectedHeap* g1 = G1CollectedHeap::heap(); 275 const HeapRegion* hr = g1->heap_region_containing(p); 276 if (hr == NULL) { 277 return false; 278 } 279 return !(hr->is_young()); 280 } else if (UseParallelGC) { 281 ParallelScavengeHeap* psh = ParallelScavengeHeap::heap(); 282 return !psh->is_in_young(p); 283 } 284 #endif // INCLUDE_ALL_GCS 285 GenCollectedHeap* gch = GenCollectedHeap::heap(); 286 return !gch->is_in_young(p); 287 WB_END 288 289 WB_ENTRY(jlong, WB_GetObjectSize(JNIEnv* env, jobject o, jobject obj)) 290 oop p = JNIHandles::resolve(obj); 291 return p->size() * HeapWordSize; 292 WB_END 293 294 #if INCLUDE_ALL_GCS 295 WB_ENTRY(jboolean, WB_G1IsHumongous(JNIEnv* env, jobject o, jobject obj)) 296 G1CollectedHeap* g1 = G1CollectedHeap::heap(); 297 oop result = JNIHandles::resolve(obj); 298 const HeapRegion* hr = g1->heap_region_containing(result); 299 return hr->is_humongous(); 300 WB_END 301 302 WB_ENTRY(jlong, WB_G1NumFreeRegions(JNIEnv* env, jobject o)) 303 G1CollectedHeap* g1 = G1CollectedHeap::heap(); 304 size_t nr = g1->num_free_regions(); 305 return (jlong)nr; 306 WB_END 307 308 WB_ENTRY(jboolean, WB_G1InConcurrentMark(JNIEnv* env, jobject o)) 309 G1CollectedHeap* g1 = G1CollectedHeap::heap(); 310 return g1->concurrent_mark()->cmThread()->during_cycle(); 311 WB_END 312 313 WB_ENTRY(jboolean, WB_G1StartMarkCycle(JNIEnv* env, jobject o)) 314 G1CollectedHeap* g1h = G1CollectedHeap::heap(); 315 if (!g1h->concurrent_mark()->cmThread()->during_cycle()) { 316 g1h->collect(GCCause::_wb_conc_mark); 317 return true; 318 } 319 return false; 320 WB_END 321 322 WB_ENTRY(jint, WB_G1RegionSize(JNIEnv* env, jobject o)) 323 return (jint)HeapRegion::GrainBytes; 324 WB_END 325 #endif // INCLUDE_ALL_GCS 326 327 #if INCLUDE_NMT 328 // Alloc memory using the test memory type so that we can use that to see if 329 // NMT picks it up correctly 330 WB_ENTRY(jlong, WB_NMTMalloc(JNIEnv* env, jobject o, jlong size)) 331 jlong addr = 0; 332 addr = (jlong)(uintptr_t)os::malloc(size, mtTest); 333 return addr; 334 WB_END 335 336 // Alloc memory with pseudo call stack. The test can create psudo malloc 337 // allocation site to stress the malloc tracking. 338 WB_ENTRY(jlong, WB_NMTMallocWithPseudoStack(JNIEnv* env, jobject o, jlong size, jint pseudo_stack)) 339 address pc = (address)(size_t)pseudo_stack; 340 NativeCallStack stack(&pc, 1); 341 return (jlong)(uintptr_t)os::malloc(size, mtTest, stack); 342 WB_END 343 344 // Free the memory allocated by NMTAllocTest 345 WB_ENTRY(void, WB_NMTFree(JNIEnv* env, jobject o, jlong mem)) 346 os::free((void*)(uintptr_t)mem); 347 WB_END 348 349 WB_ENTRY(jlong, WB_NMTReserveMemory(JNIEnv* env, jobject o, jlong size)) 350 jlong addr = 0; 351 352 addr = (jlong)(uintptr_t)os::reserve_memory(size); 353 MemTracker::record_virtual_memory_type((address)addr, mtTest); 354 355 return addr; 356 WB_END 357 358 WB_ENTRY(void, WB_NMTCommitMemory(JNIEnv* env, jobject o, jlong addr, jlong size)) 359 os::commit_memory((char *)(uintptr_t)addr, size, !ExecMem); 360 MemTracker::record_virtual_memory_type((address)(uintptr_t)addr, mtTest); 361 WB_END 362 363 WB_ENTRY(void, WB_NMTUncommitMemory(JNIEnv* env, jobject o, jlong addr, jlong size)) 364 os::uncommit_memory((char *)(uintptr_t)addr, size); 365 WB_END 366 367 WB_ENTRY(void, WB_NMTReleaseMemory(JNIEnv* env, jobject o, jlong addr, jlong size)) 368 os::release_memory((char *)(uintptr_t)addr, size); 369 WB_END 370 371 WB_ENTRY(jboolean, WB_NMTIsDetailSupported(JNIEnv* env)) 372 return MemTracker::tracking_level() == NMT_detail; 373 WB_END 374 375 WB_ENTRY(jboolean, WB_NMTChangeTrackingLevel(JNIEnv* env)) 376 // Test that we can downgrade NMT levels but not upgrade them. 377 if (MemTracker::tracking_level() == NMT_off) { 378 MemTracker::transition_to(NMT_off); 379 return MemTracker::tracking_level() == NMT_off; 380 } else { 381 assert(MemTracker::tracking_level() == NMT_detail, "Should start out as detail tracking"); 382 MemTracker::transition_to(NMT_summary); 383 assert(MemTracker::tracking_level() == NMT_summary, "Should be summary now"); 384 385 // Can't go to detail once NMT is set to summary. 386 MemTracker::transition_to(NMT_detail); 387 assert(MemTracker::tracking_level() == NMT_summary, "Should still be summary now"); 388 389 // Shutdown sets tracking level to minimal. 390 MemTracker::shutdown(); 391 assert(MemTracker::tracking_level() == NMT_minimal, "Should be minimal now"); 392 393 // Once the tracking level is minimal, we cannot increase to summary. 394 // The code ignores this request instead of asserting because if the malloc site 395 // table overflows in another thread, it tries to change the code to summary. 396 MemTracker::transition_to(NMT_summary); 397 assert(MemTracker::tracking_level() == NMT_minimal, "Should still be minimal now"); 398 399 // Really can never go up to detail, verify that the code would never do this. 400 MemTracker::transition_to(NMT_detail); 401 assert(MemTracker::tracking_level() == NMT_minimal, "Should still be minimal now"); 402 return MemTracker::tracking_level() == NMT_minimal; 403 } 404 WB_END 405 406 WB_ENTRY(jint, WB_NMTGetHashSize(JNIEnv* env, jobject o)) 407 int hash_size = MallocSiteTable::hash_buckets(); 408 assert(hash_size > 0, "NMT hash_size should be > 0"); 409 return (jint)hash_size; 410 WB_END 411 #endif // INCLUDE_NMT 412 413 static jmethodID reflected_method_to_jmid(JavaThread* thread, JNIEnv* env, jobject method) { 414 assert(method != NULL, "method should not be null"); 415 ThreadToNativeFromVM ttn(thread); 416 return env->FromReflectedMethod(method); 417 } 418 419 // Deoptimizes all compiled frames and makes nmethods not entrant if it's requested 420 class VM_WhiteBoxDeoptimizeFrames : public VM_WhiteBoxOperation { 421 private: 422 int _result; 423 const bool _make_not_entrant; 424 public: 425 VM_WhiteBoxDeoptimizeFrames(bool make_not_entrant) : 426 _result(0), _make_not_entrant(make_not_entrant) { } 427 int result() const { return _result; } 428 429 void doit() { 430 for (JavaThread* t = Threads::first(); t != NULL; t = t->next()) { 431 if (t->has_last_Java_frame()) { 432 for (StackFrameStream fst(t, UseBiasedLocking); !fst.is_done(); fst.next()) { 433 frame* f = fst.current(); 434 if (f->can_be_deoptimized() && !f->is_deoptimized_frame()) { 435 RegisterMap* reg_map = fst.register_map(); 436 Deoptimization::deoptimize(t, *f, reg_map); 437 if (_make_not_entrant) { 438 nmethod* nm = CodeCache::find_nmethod(f->pc()); 439 assert(nm != NULL, "sanity check"); 440 nm->make_not_entrant(); 441 } 442 ++_result; 443 } 444 } 445 } 446 } 447 } 448 }; 449 450 WB_ENTRY(jint, WB_DeoptimizeFrames(JNIEnv* env, jobject o, jboolean make_not_entrant)) 451 VM_WhiteBoxDeoptimizeFrames op(make_not_entrant == JNI_TRUE); 452 VMThread::execute(&op); 453 return op.result(); 454 WB_END 455 456 WB_ENTRY(void, WB_DeoptimizeAll(JNIEnv* env, jobject o)) 457 MutexLockerEx mu(Compile_lock); 458 CodeCache::mark_all_nmethods_for_deoptimization(); 459 VM_Deoptimize op; 460 VMThread::execute(&op); 461 WB_END 462 463 WB_ENTRY(jint, WB_DeoptimizeMethod(JNIEnv* env, jobject o, jobject method, jboolean is_osr)) 464 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 465 int result = 0; 466 CHECK_JNI_EXCEPTION_(env, result); 467 MutexLockerEx mu(Compile_lock); 468 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 469 if (is_osr) { 470 result += mh->mark_osr_nmethods(); 471 } else if (mh->code() != NULL) { 472 mh->code()->mark_for_deoptimization(); 473 ++result; 474 } 475 result += CodeCache::mark_for_deoptimization(mh()); 476 if (result > 0) { 477 VM_Deoptimize op; 478 VMThread::execute(&op); 479 } 480 return result; 481 WB_END 482 483 WB_ENTRY(jboolean, WB_IsMethodCompiled(JNIEnv* env, jobject o, jobject method, jboolean is_osr)) 484 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 485 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 486 MutexLockerEx mu(Compile_lock); 487 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 488 nmethod* code = is_osr ? mh->lookup_osr_nmethod_for(InvocationEntryBci, CompLevel_none, false) : mh->code(); 489 if (code == NULL) { 490 return JNI_FALSE; 491 } 492 return (code->is_alive() && !code->is_marked_for_deoptimization()); 493 WB_END 494 495 WB_ENTRY(jboolean, WB_IsMethodCompilable(JNIEnv* env, jobject o, jobject method, jint comp_level, jboolean is_osr)) 496 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 497 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 498 MutexLockerEx mu(Compile_lock); 499 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 500 if (is_osr) { 501 return CompilationPolicy::can_be_osr_compiled(mh, comp_level); 502 } else { 503 return CompilationPolicy::can_be_compiled(mh, comp_level); 504 } 505 WB_END 506 507 WB_ENTRY(jboolean, WB_IsMethodQueuedForCompilation(JNIEnv* env, jobject o, jobject method)) 508 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 509 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 510 MutexLockerEx mu(Compile_lock); 511 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 512 return mh->queued_for_compilation(); 513 WB_END 514 515 WB_ENTRY(jint, WB_GetMethodCompilationLevel(JNIEnv* env, jobject o, jobject method, jboolean is_osr)) 516 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 517 CHECK_JNI_EXCEPTION_(env, CompLevel_none); 518 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 519 nmethod* code = is_osr ? mh->lookup_osr_nmethod_for(InvocationEntryBci, CompLevel_none, false) : mh->code(); 520 return (code != NULL ? code->comp_level() : CompLevel_none); 521 WB_END 522 523 WB_ENTRY(void, WB_MakeMethodNotCompilable(JNIEnv* env, jobject o, jobject method, jint comp_level, jboolean is_osr)) 524 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 525 CHECK_JNI_EXCEPTION(env); 526 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 527 if (is_osr) { 528 mh->set_not_osr_compilable(comp_level, true /* report */, "WhiteBox"); 529 } else { 530 mh->set_not_compilable(comp_level, true /* report */, "WhiteBox"); 531 } 532 WB_END 533 534 WB_ENTRY(jint, WB_GetMethodEntryBci(JNIEnv* env, jobject o, jobject method)) 535 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 536 CHECK_JNI_EXCEPTION_(env, InvocationEntryBci); 537 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 538 nmethod* code = mh->lookup_osr_nmethod_for(InvocationEntryBci, CompLevel_none, false); 539 return (code != NULL && code->is_osr_method() ? code->osr_entry_bci() : InvocationEntryBci); 540 WB_END 541 542 WB_ENTRY(jboolean, WB_TestSetDontInlineMethod(JNIEnv* env, jobject o, jobject method, jboolean value)) 543 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 544 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 545 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 546 bool result = mh->dont_inline(); 547 mh->set_dont_inline(value == JNI_TRUE); 548 return result; 549 WB_END 550 551 WB_ENTRY(jint, WB_GetCompileQueueSize(JNIEnv* env, jobject o, jint comp_level)) 552 if (comp_level == CompLevel_any) { 553 return CompileBroker::queue_size(CompLevel_full_optimization) /* C2 */ + 554 CompileBroker::queue_size(CompLevel_full_profile) /* C1 */; 555 } else { 556 return CompileBroker::queue_size(comp_level); 557 } 558 WB_END 559 560 WB_ENTRY(jboolean, WB_TestSetForceInlineMethod(JNIEnv* env, jobject o, jobject method, jboolean value)) 561 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 562 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 563 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 564 bool result = mh->force_inline(); 565 mh->set_force_inline(value == JNI_TRUE); 566 return result; 567 WB_END 568 569 WB_ENTRY(jboolean, WB_EnqueueMethodForCompilation(JNIEnv* env, jobject o, jobject method, jint comp_level, jint bci)) 570 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 571 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 572 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 573 nmethod* nm = CompileBroker::compile_method(mh, bci, comp_level, mh, mh->invocation_count(), "WhiteBox", THREAD); 574 MutexLockerEx mu(Compile_lock); 575 return (mh->queued_for_compilation() || nm != NULL); 576 WB_END 577 578 class AlwaysFalseClosure : public BoolObjectClosure { 579 public: 580 bool do_object_b(oop p) { return false; } 581 }; 582 583 static AlwaysFalseClosure always_false; 584 585 WB_ENTRY(void, WB_ClearMethodState(JNIEnv* env, jobject o, jobject method)) 586 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 587 CHECK_JNI_EXCEPTION(env); 588 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 589 MutexLockerEx mu(Compile_lock); 590 MethodData* mdo = mh->method_data(); 591 MethodCounters* mcs = mh->method_counters(); 592 593 if (mdo != NULL) { 594 mdo->init(); 595 ResourceMark rm; 596 int arg_count = mdo->method()->size_of_parameters(); 597 for (int i = 0; i < arg_count; i++) { 598 mdo->set_arg_modified(i, 0); 599 } 600 MutexLockerEx mu(mdo->extra_data_lock()); 601 mdo->clean_method_data(&always_false); 602 } 603 604 mh->clear_not_c1_compilable(); 605 mh->clear_not_c2_compilable(); 606 mh->clear_not_c2_osr_compilable(); 607 NOT_PRODUCT(mh->set_compiled_invocation_count(0)); 608 if (mcs != NULL) { 609 mcs->backedge_counter()->init(); 610 mcs->invocation_counter()->init(); 611 mcs->set_interpreter_invocation_count(0); 612 mcs->set_interpreter_throwout_count(0); 613 614 #ifdef TIERED 615 mcs->set_rate(0.0F); 616 mh->set_prev_event_count(0); 617 mh->set_prev_time(0); 618 #endif 619 } 620 WB_END 621 622 template <typename T> 623 static bool GetVMFlag(JavaThread* thread, JNIEnv* env, jstring name, T* value, bool (*TAt)(const char*, T*, bool, bool)) { 624 if (name == NULL) { 625 return false; 626 } 627 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 628 const char* flag_name = env->GetStringUTFChars(name, NULL); 629 bool result = (*TAt)(flag_name, value, true, true); 630 env->ReleaseStringUTFChars(name, flag_name); 631 return result; 632 } 633 634 template <typename T> 635 static bool SetVMFlag(JavaThread* thread, JNIEnv* env, jstring name, T* value, bool (*TAtPut)(const char*, T*, Flag::Flags)) { 636 if (name == NULL) { 637 return false; 638 } 639 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 640 const char* flag_name = env->GetStringUTFChars(name, NULL); 641 bool result = (*TAtPut)(flag_name, value, Flag::INTERNAL); 642 env->ReleaseStringUTFChars(name, flag_name); 643 return result; 644 } 645 646 template <typename T> 647 static jobject box(JavaThread* thread, JNIEnv* env, Symbol* name, Symbol* sig, T value) { 648 ResourceMark rm(thread); 649 jclass clazz = env->FindClass(name->as_C_string()); 650 CHECK_JNI_EXCEPTION_(env, NULL); 651 jmethodID methodID = env->GetStaticMethodID(clazz, 652 vmSymbols::valueOf_name()->as_C_string(), 653 sig->as_C_string()); 654 CHECK_JNI_EXCEPTION_(env, NULL); 655 jobject result = env->CallStaticObjectMethod(clazz, methodID, value); 656 CHECK_JNI_EXCEPTION_(env, NULL); 657 return result; 658 } 659 660 static jobject booleanBox(JavaThread* thread, JNIEnv* env, jboolean value) { 661 return box(thread, env, vmSymbols::java_lang_Boolean(), vmSymbols::Boolean_valueOf_signature(), value); 662 } 663 static jobject integerBox(JavaThread* thread, JNIEnv* env, jint value) { 664 return box(thread, env, vmSymbols::java_lang_Integer(), vmSymbols::Integer_valueOf_signature(), value); 665 } 666 static jobject longBox(JavaThread* thread, JNIEnv* env, jlong value) { 667 return box(thread, env, vmSymbols::java_lang_Long(), vmSymbols::Long_valueOf_signature(), value); 668 } 669 /* static jobject floatBox(JavaThread* thread, JNIEnv* env, jfloat value) { 670 return box(thread, env, vmSymbols::java_lang_Float(), vmSymbols::Float_valueOf_signature(), value); 671 }*/ 672 static jobject doubleBox(JavaThread* thread, JNIEnv* env, jdouble value) { 673 return box(thread, env, vmSymbols::java_lang_Double(), vmSymbols::Double_valueOf_signature(), value); 674 } 675 676 static Flag* getVMFlag(JavaThread* thread, JNIEnv* env, jstring name) { 677 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 678 const char* flag_name = env->GetStringUTFChars(name, NULL); 679 Flag* result = Flag::find_flag(flag_name, strlen(flag_name), true, true); 680 env->ReleaseStringUTFChars(name, flag_name); 681 return result; 682 } 683 684 WB_ENTRY(jboolean, WB_IsConstantVMFlag(JNIEnv* env, jobject o, jstring name)) 685 Flag* flag = getVMFlag(thread, env, name); 686 return (flag != NULL) && flag->is_constant_in_binary(); 687 WB_END 688 689 WB_ENTRY(jboolean, WB_IsLockedVMFlag(JNIEnv* env, jobject o, jstring name)) 690 Flag* flag = getVMFlag(thread, env, name); 691 return (flag != NULL) && !(flag->is_unlocked() || flag->is_unlocker()); 692 WB_END 693 694 WB_ENTRY(jobject, WB_GetBooleanVMFlag(JNIEnv* env, jobject o, jstring name)) 695 bool result; 696 if (GetVMFlag <bool> (thread, env, name, &result, &CommandLineFlags::boolAt)) { 697 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 698 return booleanBox(thread, env, result); 699 } 700 return NULL; 701 WB_END 702 703 WB_ENTRY(jobject, WB_GetIntxVMFlag(JNIEnv* env, jobject o, jstring name)) 704 intx result; 705 if (GetVMFlag <intx> (thread, env, name, &result, &CommandLineFlags::intxAt)) { 706 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 707 return longBox(thread, env, result); 708 } 709 return NULL; 710 WB_END 711 712 WB_ENTRY(jobject, WB_GetUintxVMFlag(JNIEnv* env, jobject o, jstring name)) 713 uintx result; 714 if (GetVMFlag <uintx> (thread, env, name, &result, &CommandLineFlags::uintxAt)) { 715 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 716 return longBox(thread, env, result); 717 } 718 return NULL; 719 WB_END 720 721 WB_ENTRY(jobject, WB_GetUint64VMFlag(JNIEnv* env, jobject o, jstring name)) 722 uint64_t result; 723 if (GetVMFlag <uint64_t> (thread, env, name, &result, &CommandLineFlags::uint64_tAt)) { 724 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 725 return longBox(thread, env, result); 726 } 727 return NULL; 728 WB_END 729 730 WB_ENTRY(jobject, WB_GetSizeTVMFlag(JNIEnv* env, jobject o, jstring name)) 731 uintx result; 732 if (GetVMFlag <size_t> (thread, env, name, &result, &CommandLineFlags::size_tAt)) { 733 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 734 return longBox(thread, env, result); 735 } 736 return NULL; 737 WB_END 738 739 WB_ENTRY(jobject, WB_GetDoubleVMFlag(JNIEnv* env, jobject o, jstring name)) 740 double result; 741 if (GetVMFlag <double> (thread, env, name, &result, &CommandLineFlags::doubleAt)) { 742 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 743 return doubleBox(thread, env, result); 744 } 745 return NULL; 746 WB_END 747 748 WB_ENTRY(jstring, WB_GetStringVMFlag(JNIEnv* env, jobject o, jstring name)) 749 ccstr ccstrResult; 750 if (GetVMFlag <ccstr> (thread, env, name, &ccstrResult, &CommandLineFlags::ccstrAt)) { 751 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 752 jstring result = env->NewStringUTF(ccstrResult); 753 CHECK_JNI_EXCEPTION_(env, NULL); 754 return result; 755 } 756 return NULL; 757 WB_END 758 759 WB_ENTRY(void, WB_SetBooleanVMFlag(JNIEnv* env, jobject o, jstring name, jboolean value)) 760 bool result = value == JNI_TRUE ? true : false; 761 SetVMFlag <bool> (thread, env, name, &result, &CommandLineFlags::boolAtPut); 762 WB_END 763 764 WB_ENTRY(void, WB_SetIntxVMFlag(JNIEnv* env, jobject o, jstring name, jlong value)) 765 intx result = value; 766 SetVMFlag <intx> (thread, env, name, &result, &CommandLineFlags::intxAtPut); 767 WB_END 768 769 WB_ENTRY(void, WB_SetUintxVMFlag(JNIEnv* env, jobject o, jstring name, jlong value)) 770 uintx result = value; 771 SetVMFlag <uintx> (thread, env, name, &result, &CommandLineFlags::uintxAtPut); 772 WB_END 773 774 WB_ENTRY(void, WB_SetUint64VMFlag(JNIEnv* env, jobject o, jstring name, jlong value)) 775 uint64_t result = value; 776 SetVMFlag <uint64_t> (thread, env, name, &result, &CommandLineFlags::uint64_tAtPut); 777 WB_END 778 779 WB_ENTRY(void, WB_SetSizeTVMFlag(JNIEnv* env, jobject o, jstring name, jlong value)) 780 size_t result = value; 781 SetVMFlag <size_t> (thread, env, name, &result, &CommandLineFlags::size_tAtPut); 782 WB_END 783 784 WB_ENTRY(void, WB_SetDoubleVMFlag(JNIEnv* env, jobject o, jstring name, jdouble value)) 785 double result = value; 786 SetVMFlag <double> (thread, env, name, &result, &CommandLineFlags::doubleAtPut); 787 WB_END 788 789 WB_ENTRY(void, WB_SetStringVMFlag(JNIEnv* env, jobject o, jstring name, jstring value)) 790 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 791 const char* ccstrValue = (value == NULL) ? NULL : env->GetStringUTFChars(value, NULL); 792 ccstr ccstrResult = ccstrValue; 793 bool needFree; 794 { 795 ThreadInVMfromNative ttvfn(thread); // back to VM 796 needFree = SetVMFlag <ccstr> (thread, env, name, &ccstrResult, &CommandLineFlags::ccstrAtPut); 797 } 798 if (value != NULL) { 799 env->ReleaseStringUTFChars(value, ccstrValue); 800 } 801 if (needFree) { 802 FREE_C_HEAP_ARRAY(char, ccstrResult); 803 } 804 WB_END 805 806 WB_ENTRY(void, WB_LockCompilation(JNIEnv* env, jobject o, jlong timeout)) 807 WhiteBox::compilation_locked = true; 808 WB_END 809 810 WB_ENTRY(void, WB_UnlockCompilation(JNIEnv* env, jobject o)) 811 MonitorLockerEx mo(Compilation_lock, Mutex::_no_safepoint_check_flag); 812 WhiteBox::compilation_locked = false; 813 mo.notify_all(); 814 WB_END 815 816 void WhiteBox::sweeper_thread_entry(JavaThread* thread, TRAPS) { 817 guarantee(WhiteBoxAPI, "internal testing API :: WhiteBox has to be enabled"); 818 { 819 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 820 NMethodSweeper::_should_sweep = true; 821 } 822 NMethodSweeper::possibly_sweep(); 823 } 824 825 JavaThread* WhiteBox::create_sweeper_thread(TRAPS) { 826 // create sweeper thread w/ custom entry -- one iteration instead of loop 827 CodeCacheSweeperThread* sweeper_thread = new CodeCacheSweeperThread(); 828 sweeper_thread->set_entry_point(&WhiteBox::sweeper_thread_entry); 829 830 // create j.l.Thread object and associate it w/ sweeper thread 831 { 832 // inherit deamon property from current thread 833 bool is_daemon = java_lang_Thread::is_daemon(JavaThread::current()->threadObj()); 834 835 HandleMark hm(THREAD); 836 Handle thread_group(THREAD, Universe::system_thread_group()); 837 const char* name = "WB Sweeper thread"; 838 sweeper_thread->allocate_threadObj(thread_group, name, is_daemon, THREAD); 839 } 840 841 { 842 MutexLocker mu(Threads_lock, THREAD); 843 Threads::add(sweeper_thread); 844 } 845 return sweeper_thread; 846 } 847 848 WB_ENTRY(jobject, WB_ForceNMethodSweep(JNIEnv* env, jobject o)) 849 JavaThread* sweeper_thread = WhiteBox::create_sweeper_thread(Thread::current()); 850 if (sweeper_thread == NULL) { 851 return NULL; 852 } 853 jobject result = JNIHandles::make_local(env, sweeper_thread->threadObj()); 854 Thread::start(sweeper_thread); 855 return result; 856 WB_END 857 858 WB_ENTRY(jboolean, WB_IsInStringTable(JNIEnv* env, jobject o, jstring javaString)) 859 ResourceMark rm(THREAD); 860 int len; 861 jchar* name = java_lang_String::as_unicode_string(JNIHandles::resolve(javaString), len, CHECK_false); 862 return (StringTable::lookup(name, len) != NULL); 863 WB_END 864 865 WB_ENTRY(void, WB_FullGC(JNIEnv* env, jobject o)) 866 Universe::heap()->collector_policy()->set_should_clear_all_soft_refs(true); 867 Universe::heap()->collect(GCCause::_last_ditch_collection); 868 #if INCLUDE_ALL_GCS 869 if (UseG1GC) { 870 // Needs to be cleared explicitly for G1 871 Universe::heap()->collector_policy()->set_should_clear_all_soft_refs(false); 872 } 873 #endif // INCLUDE_ALL_GCS 874 WB_END 875 876 WB_ENTRY(void, WB_YoungGC(JNIEnv* env, jobject o)) 877 Universe::heap()->collect(GCCause::_wb_young_gc); 878 WB_END 879 880 WB_ENTRY(void, WB_ReadReservedMemory(JNIEnv* env, jobject o)) 881 // static+volatile in order to force the read to happen 882 // (not be eliminated by the compiler) 883 static char c; 884 static volatile char* p; 885 886 p = os::reserve_memory(os::vm_allocation_granularity(), NULL, 0); 887 if (p == NULL) { 888 THROW_MSG(vmSymbols::java_lang_OutOfMemoryError(), "Failed to reserve memory"); 889 } 890 891 c = *p; 892 WB_END 893 894 WB_ENTRY(jstring, WB_GetCPUFeatures(JNIEnv* env, jobject o)) 895 const char* cpu_features = VM_Version::cpu_features(); 896 ThreadToNativeFromVM ttn(thread); 897 jstring features_string = env->NewStringUTF(cpu_features); 898 899 CHECK_JNI_EXCEPTION_(env, NULL); 900 901 return features_string; 902 WB_END 903 904 int WhiteBox::get_blob_type(const CodeBlob* code) { 905 guarantee(WhiteBoxAPI, "internal testing API :: WhiteBox has to be enabled"); 906 return CodeCache::get_code_heap(code)->code_blob_type(); 907 } 908 909 CodeHeap* WhiteBox::get_code_heap(int blob_type) { 910 guarantee(WhiteBoxAPI, "internal testing API :: WhiteBox has to be enabled"); 911 return CodeCache::get_code_heap(blob_type); 912 } 913 914 struct CodeBlobStub { 915 CodeBlobStub(const CodeBlob* blob) : 916 name(os::strdup(blob->name())), 917 size(blob->size()), 918 blob_type(WhiteBox::get_blob_type(blob)) { } 919 ~CodeBlobStub() { os::free((void*) name); } 920 const char* const name; 921 const int size; 922 const int blob_type; 923 }; 924 925 static jobjectArray codeBlob2objectArray(JavaThread* thread, JNIEnv* env, CodeBlobStub* cb) { 926 jclass clazz = env->FindClass(vmSymbols::java_lang_Object()->as_C_string()); 927 CHECK_JNI_EXCEPTION_(env, NULL); 928 jobjectArray result = env->NewObjectArray(3, clazz, NULL); 929 930 jstring name = env->NewStringUTF(cb->name); 931 CHECK_JNI_EXCEPTION_(env, NULL); 932 env->SetObjectArrayElement(result, 0, name); 933 934 jobject obj = integerBox(thread, env, cb->size); 935 CHECK_JNI_EXCEPTION_(env, NULL); 936 env->SetObjectArrayElement(result, 1, obj); 937 938 obj = integerBox(thread, env, cb->blob_type); 939 CHECK_JNI_EXCEPTION_(env, NULL); 940 env->SetObjectArrayElement(result, 2, obj); 941 942 return result; 943 } 944 945 WB_ENTRY(jobjectArray, WB_GetNMethod(JNIEnv* env, jobject o, jobject method, jboolean is_osr)) 946 ResourceMark rm(THREAD); 947 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 948 CHECK_JNI_EXCEPTION_(env, NULL); 949 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 950 nmethod* code = is_osr ? mh->lookup_osr_nmethod_for(InvocationEntryBci, CompLevel_none, false) : mh->code(); 951 jobjectArray result = NULL; 952 if (code == NULL) { 953 return result; 954 } 955 int insts_size = code->insts_size(); 956 957 ThreadToNativeFromVM ttn(thread); 958 jclass clazz = env->FindClass(vmSymbols::java_lang_Object()->as_C_string()); 959 CHECK_JNI_EXCEPTION_(env, NULL); 960 result = env->NewObjectArray(4, clazz, NULL); 961 if (result == NULL) { 962 return result; 963 } 964 965 CodeBlobStub stub(code); 966 jobjectArray codeBlob = codeBlob2objectArray(thread, env, &stub); 967 env->SetObjectArrayElement(result, 0, codeBlob); 968 969 jobject level = integerBox(thread, env, code->comp_level()); 970 CHECK_JNI_EXCEPTION_(env, NULL); 971 env->SetObjectArrayElement(result, 1, level); 972 973 jbyteArray insts = env->NewByteArray(insts_size); 974 CHECK_JNI_EXCEPTION_(env, NULL); 975 env->SetByteArrayRegion(insts, 0, insts_size, (jbyte*) code->insts_begin()); 976 env->SetObjectArrayElement(result, 2, insts); 977 978 jobject id = integerBox(thread, env, code->compile_id()); 979 CHECK_JNI_EXCEPTION_(env, NULL); 980 env->SetObjectArrayElement(result, 3, id); 981 982 return result; 983 WB_END 984 985 CodeBlob* WhiteBox::allocate_code_blob(int size, int blob_type) { 986 guarantee(WhiteBoxAPI, "internal testing API :: WhiteBox has to be enabled"); 987 BufferBlob* blob; 988 int full_size = CodeBlob::align_code_offset(sizeof(BufferBlob)); 989 if (full_size < size) { 990 full_size += round_to(size - full_size, oopSize); 991 } 992 { 993 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 994 blob = (BufferBlob*) CodeCache::allocate(full_size, blob_type); 995 ::new (blob) BufferBlob("WB::DummyBlob", full_size); 996 } 997 // Track memory usage statistic after releasing CodeCache_lock 998 MemoryService::track_code_cache_memory_usage(); 999 return blob; 1000 } 1001 1002 WB_ENTRY(jlong, WB_AllocateCodeBlob(JNIEnv* env, jobject o, jint size, jint blob_type)) 1003 return (jlong) WhiteBox::allocate_code_blob(size, blob_type); 1004 WB_END 1005 1006 WB_ENTRY(void, WB_FreeCodeBlob(JNIEnv* env, jobject o, jlong addr)) 1007 BufferBlob::free((BufferBlob*) addr); 1008 WB_END 1009 1010 WB_ENTRY(jobjectArray, WB_GetCodeHeapEntries(JNIEnv* env, jobject o, jint blob_type)) 1011 ResourceMark rm; 1012 GrowableArray<CodeBlobStub*> blobs; 1013 { 1014 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 1015 CodeHeap* heap = WhiteBox::get_code_heap(blob_type); 1016 if (heap == NULL) { 1017 return NULL; 1018 } 1019 for (CodeBlob* cb = (CodeBlob*) heap->first(); 1020 cb != NULL; cb = (CodeBlob*) heap->next(cb)) { 1021 CodeBlobStub* stub = NEW_RESOURCE_OBJ(CodeBlobStub); 1022 new (stub) CodeBlobStub(cb); 1023 blobs.append(stub); 1024 } 1025 } 1026 if (blobs.length() == 0) { 1027 return NULL; 1028 } 1029 ThreadToNativeFromVM ttn(thread); 1030 jobjectArray result = NULL; 1031 jclass clazz = env->FindClass(vmSymbols::java_lang_Object()->as_C_string()); 1032 CHECK_JNI_EXCEPTION_(env, NULL); 1033 result = env->NewObjectArray(blobs.length(), clazz, NULL); 1034 if (result == NULL) { 1035 return result; 1036 } 1037 int i = 0; 1038 for (GrowableArrayIterator<CodeBlobStub*> it = blobs.begin(); 1039 it != blobs.end(); ++it) { 1040 jobjectArray obj = codeBlob2objectArray(thread, env, *it); 1041 env->SetObjectArrayElement(result, i, obj); 1042 ++i; 1043 } 1044 return result; 1045 WB_END 1046 1047 WB_ENTRY(jint, WB_GetCompilationActivityMode(JNIEnv* env, jobject o)) 1048 return CompileBroker::get_compilation_activity_mode(); 1049 WB_END 1050 1051 WB_ENTRY(jobjectArray, WB_GetCodeBlob(JNIEnv* env, jobject o, jlong addr)) 1052 ThreadToNativeFromVM ttn(thread); 1053 CodeBlobStub stub((CodeBlob*) addr); 1054 return codeBlob2objectArray(thread, env, &stub); 1055 WB_END 1056 1057 WB_ENTRY(jlong, WB_GetThreadStackSize(JNIEnv* env, jobject o)) 1058 return (jlong) Thread::current()->stack_size(); 1059 WB_END 1060 1061 WB_ENTRY(jlong, WB_GetThreadRemainingStackSize(JNIEnv* env, jobject o)) 1062 JavaThread* t = JavaThread::current(); 1063 return (jlong) t->stack_available(os::current_stack_pointer()) - (jlong) StackShadowPages * os::vm_page_size(); 1064 WB_END 1065 1066 int WhiteBox::array_bytes_to_length(size_t bytes) { 1067 return Array<u1>::bytes_to_length(bytes); 1068 } 1069 1070 WB_ENTRY(jlong, WB_AllocateMetaspace(JNIEnv* env, jobject wb, jobject class_loader, jlong size)) 1071 if (size < 0) { 1072 THROW_MSG_0(vmSymbols::java_lang_IllegalArgumentException(), 1073 err_msg("WB_AllocateMetaspace: size is negative: " JLONG_FORMAT, size)); 1074 } 1075 1076 oop class_loader_oop = JNIHandles::resolve(class_loader); 1077 ClassLoaderData* cld = class_loader_oop != NULL 1078 ? java_lang_ClassLoader::loader_data(class_loader_oop) 1079 : ClassLoaderData::the_null_class_loader_data(); 1080 1081 void* metadata = MetadataFactory::new_writeable_array<u1>(cld, WhiteBox::array_bytes_to_length((size_t)size), thread); 1082 1083 return (jlong)(uintptr_t)metadata; 1084 WB_END 1085 1086 WB_ENTRY(void, WB_FreeMetaspace(JNIEnv* env, jobject wb, jobject class_loader, jlong addr, jlong size)) 1087 oop class_loader_oop = JNIHandles::resolve(class_loader); 1088 ClassLoaderData* cld = class_loader_oop != NULL 1089 ? java_lang_ClassLoader::loader_data(class_loader_oop) 1090 : ClassLoaderData::the_null_class_loader_data(); 1091 1092 MetadataFactory::free_array(cld, (Array<u1>*)(uintptr_t)addr); 1093 WB_END 1094 1095 WB_ENTRY(jlong, WB_IncMetaspaceCapacityUntilGC(JNIEnv* env, jobject wb, jlong inc)) 1096 if (inc < 0) { 1097 THROW_MSG_0(vmSymbols::java_lang_IllegalArgumentException(), 1098 err_msg("WB_IncMetaspaceCapacityUntilGC: inc is negative: " JLONG_FORMAT, inc)); 1099 } 1100 1101 jlong max_size_t = (jlong) ((size_t) -1); 1102 if (inc > max_size_t) { 1103 THROW_MSG_0(vmSymbols::java_lang_IllegalArgumentException(), 1104 err_msg("WB_IncMetaspaceCapacityUntilGC: inc does not fit in size_t: " JLONG_FORMAT, inc)); 1105 } 1106 1107 size_t new_cap_until_GC = 0; 1108 size_t aligned_inc = align_size_down((size_t) inc, Metaspace::commit_alignment()); 1109 bool success = MetaspaceGC::inc_capacity_until_GC(aligned_inc, &new_cap_until_GC); 1110 if (!success) { 1111 THROW_MSG_0(vmSymbols::java_lang_IllegalStateException(), 1112 "WB_IncMetaspaceCapacityUntilGC: could not increase capacity until GC " 1113 "due to contention with another thread"); 1114 } 1115 return (jlong) new_cap_until_GC; 1116 WB_END 1117 1118 WB_ENTRY(jlong, WB_MetaspaceCapacityUntilGC(JNIEnv* env, jobject wb)) 1119 return (jlong) MetaspaceGC::capacity_until_GC(); 1120 WB_END 1121 1122 WB_ENTRY(void, WB_AssertMatchingSafepointCalls(JNIEnv* env, jobject o, jboolean mutexSafepointValue, jboolean attemptedNoSafepointValue)) 1123 Monitor::SafepointCheckRequired sfpt_check_required = mutexSafepointValue ? 1124 Monitor::_safepoint_check_always : 1125 Monitor::_safepoint_check_never; 1126 MutexLockerEx ml(new Mutex(Mutex::leaf, "SFPT_Test_lock", true, sfpt_check_required), 1127 attemptedNoSafepointValue == JNI_TRUE); 1128 WB_END 1129 1130 WB_ENTRY(jboolean, WB_IsMonitorInflated(JNIEnv* env, jobject wb, jobject obj)) 1131 oop obj_oop = JNIHandles::resolve(obj); 1132 return (jboolean) obj_oop->mark()->has_monitor(); 1133 WB_END 1134 1135 WB_ENTRY(void, WB_ForceSafepoint(JNIEnv* env, jobject wb)) 1136 VM_ForceSafepoint force_safepoint_op; 1137 VMThread::execute(&force_safepoint_op); 1138 WB_END 1139 1140 //Some convenience methods to deal with objects from java 1141 int WhiteBox::offset_for_field(const char* field_name, oop object, 1142 Symbol* signature_symbol) { 1143 assert(field_name != NULL && strlen(field_name) > 0, "Field name not valid"); 1144 Thread* THREAD = Thread::current(); 1145 1146 //Get the class of our object 1147 Klass* arg_klass = object->klass(); 1148 //Turn it into an instance-klass 1149 InstanceKlass* ik = InstanceKlass::cast(arg_klass); 1150 1151 //Create symbols to look for in the class 1152 TempNewSymbol name_symbol = SymbolTable::lookup(field_name, (int) strlen(field_name), 1153 THREAD); 1154 1155 //To be filled in with an offset of the field we're looking for 1156 fieldDescriptor fd; 1157 1158 Klass* res = ik->find_field(name_symbol, signature_symbol, &fd); 1159 if (res == NULL) { 1160 tty->print_cr("Invalid layout of %s at %s", ik->external_name(), 1161 name_symbol->as_C_string()); 1162 fatal("Invalid layout of preloaded class"); 1163 } 1164 1165 //fetch the field at the offset we've found 1166 int dest_offset = fd.offset(); 1167 1168 return dest_offset; 1169 } 1170 1171 1172 const char* WhiteBox::lookup_jstring(const char* field_name, oop object) { 1173 int offset = offset_for_field(field_name, object, 1174 vmSymbols::string_signature()); 1175 oop string = object->obj_field(offset); 1176 if (string == NULL) { 1177 return NULL; 1178 } 1179 const char* ret = java_lang_String::as_utf8_string(string); 1180 return ret; 1181 } 1182 1183 bool WhiteBox::lookup_bool(const char* field_name, oop object) { 1184 int offset = 1185 offset_for_field(field_name, object, vmSymbols::bool_signature()); 1186 bool ret = (object->bool_field(offset) == JNI_TRUE); 1187 return ret; 1188 } 1189 1190 void WhiteBox::register_methods(JNIEnv* env, jclass wbclass, JavaThread* thread, JNINativeMethod* method_array, int method_count) { 1191 ResourceMark rm; 1192 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 1193 1194 // one by one registration natives for exception catching 1195 jclass no_such_method_error_klass = env->FindClass(vmSymbols::java_lang_NoSuchMethodError()->as_C_string()); 1196 CHECK_JNI_EXCEPTION(env); 1197 for (int i = 0, n = method_count; i < n; ++i) { 1198 // Skip dummy entries 1199 if (method_array[i].fnPtr == NULL) continue; 1200 if (env->RegisterNatives(wbclass, &method_array[i], 1) != 0) { 1201 jthrowable throwable_obj = env->ExceptionOccurred(); 1202 if (throwable_obj != NULL) { 1203 env->ExceptionClear(); 1204 if (env->IsInstanceOf(throwable_obj, no_such_method_error_klass)) { 1205 // NoSuchMethodError is thrown when a method can't be found or a method is not native. 1206 // Ignoring the exception since it is not preventing use of other WhiteBox methods. 1207 tty->print_cr("Warning: 'NoSuchMethodError' on register of sun.hotspot.WhiteBox::%s%s", 1208 method_array[i].name, method_array[i].signature); 1209 } 1210 } else { 1211 // Registration failed unexpectedly. 1212 tty->print_cr("Warning: unexpected error on register of sun.hotspot.WhiteBox::%s%s. All methods will be unregistered", 1213 method_array[i].name, method_array[i].signature); 1214 env->UnregisterNatives(wbclass); 1215 break; 1216 } 1217 } 1218 } 1219 } 1220 1221 #define CC (char*) 1222 1223 static JNINativeMethod methods[] = { 1224 {CC"getObjectAddress", CC"(Ljava/lang/Object;)J", (void*)&WB_GetObjectAddress }, 1225 {CC"getObjectSize", CC"(Ljava/lang/Object;)J", (void*)&WB_GetObjectSize }, 1226 {CC"isObjectInOldGen", CC"(Ljava/lang/Object;)Z", (void*)&WB_isObjectInOldGen }, 1227 {CC"getHeapOopSize", CC"()I", (void*)&WB_GetHeapOopSize }, 1228 {CC"getVMPageSize", CC"()I", (void*)&WB_GetVMPageSize }, 1229 {CC"getVMLargePageSize", CC"()J", (void*)&WB_GetVMLargePageSize}, 1230 {CC"isClassAlive0", CC"(Ljava/lang/String;)Z", (void*)&WB_IsClassAlive }, 1231 {CC"parseCommandLine", 1232 CC"(Ljava/lang/String;C[Lsun/hotspot/parser/DiagnosticCommand;)[Ljava/lang/Object;", 1233 (void*) &WB_ParseCommandLine 1234 }, 1235 {CC"addToBootstrapClassLoaderSearch", CC"(Ljava/lang/String;)V", 1236 (void*)&WB_AddToBootstrapClassLoaderSearch}, 1237 {CC"addToSystemClassLoaderSearch", CC"(Ljava/lang/String;)V", 1238 (void*)&WB_AddToSystemClassLoaderSearch}, 1239 {CC"getCompressedOopsMaxHeapSize", CC"()J", 1240 (void*)&WB_GetCompressedOopsMaxHeapSize}, 1241 {CC"printHeapSizes", CC"()V", (void*)&WB_PrintHeapSizes }, 1242 {CC"runMemoryUnitTests", CC"()V", (void*)&WB_RunMemoryUnitTests}, 1243 {CC"readFromNoaccessArea",CC"()V", (void*)&WB_ReadFromNoaccessArea}, 1244 {CC"stressVirtualSpaceResize",CC"(JJJ)I", (void*)&WB_StressVirtualSpaceResize}, 1245 #if INCLUDE_ALL_GCS 1246 {CC"g1InConcurrentMark", CC"()Z", (void*)&WB_G1InConcurrentMark}, 1247 {CC"g1IsHumongous", CC"(Ljava/lang/Object;)Z", (void*)&WB_G1IsHumongous }, 1248 {CC"g1NumFreeRegions", CC"()J", (void*)&WB_G1NumFreeRegions }, 1249 {CC"g1RegionSize", CC"()I", (void*)&WB_G1RegionSize }, 1250 {CC"g1StartConcMarkCycle", CC"()Z", (void*)&WB_G1StartMarkCycle }, 1251 #endif // INCLUDE_ALL_GCS 1252 #if INCLUDE_NMT 1253 {CC"NMTMalloc", CC"(J)J", (void*)&WB_NMTMalloc }, 1254 {CC"NMTMallocWithPseudoStack", CC"(JI)J", (void*)&WB_NMTMallocWithPseudoStack}, 1255 {CC"NMTFree", CC"(J)V", (void*)&WB_NMTFree }, 1256 {CC"NMTReserveMemory", CC"(J)J", (void*)&WB_NMTReserveMemory }, 1257 {CC"NMTCommitMemory", CC"(JJ)V", (void*)&WB_NMTCommitMemory }, 1258 {CC"NMTUncommitMemory", CC"(JJ)V", (void*)&WB_NMTUncommitMemory }, 1259 {CC"NMTReleaseMemory", CC"(JJ)V", (void*)&WB_NMTReleaseMemory }, 1260 {CC"NMTIsDetailSupported",CC"()Z", (void*)&WB_NMTIsDetailSupported}, 1261 {CC"NMTChangeTrackingLevel", CC"()Z", (void*)&WB_NMTChangeTrackingLevel}, 1262 {CC"NMTGetHashSize", CC"()I", (void*)&WB_NMTGetHashSize }, 1263 #endif // INCLUDE_NMT 1264 {CC"deoptimizeFrames", CC"(Z)I", (void*)&WB_DeoptimizeFrames }, 1265 {CC"deoptimizeAll", CC"()V", (void*)&WB_DeoptimizeAll }, 1266 {CC"deoptimizeMethod", CC"(Ljava/lang/reflect/Executable;Z)I", 1267 (void*)&WB_DeoptimizeMethod }, 1268 {CC"isMethodCompiled", CC"(Ljava/lang/reflect/Executable;Z)Z", 1269 (void*)&WB_IsMethodCompiled }, 1270 {CC"isMethodCompilable", CC"(Ljava/lang/reflect/Executable;IZ)Z", 1271 (void*)&WB_IsMethodCompilable}, 1272 {CC"isMethodQueuedForCompilation", 1273 CC"(Ljava/lang/reflect/Executable;)Z", (void*)&WB_IsMethodQueuedForCompilation}, 1274 {CC"makeMethodNotCompilable", 1275 CC"(Ljava/lang/reflect/Executable;IZ)V", (void*)&WB_MakeMethodNotCompilable}, 1276 {CC"testSetDontInlineMethod", 1277 CC"(Ljava/lang/reflect/Executable;Z)Z", (void*)&WB_TestSetDontInlineMethod}, 1278 {CC"getMethodCompilationLevel", 1279 CC"(Ljava/lang/reflect/Executable;Z)I", (void*)&WB_GetMethodCompilationLevel}, 1280 {CC"getMethodEntryBci", 1281 CC"(Ljava/lang/reflect/Executable;)I", (void*)&WB_GetMethodEntryBci}, 1282 {CC"getCompileQueueSize", 1283 CC"(I)I", (void*)&WB_GetCompileQueueSize}, 1284 {CC"testSetForceInlineMethod", 1285 CC"(Ljava/lang/reflect/Executable;Z)Z", (void*)&WB_TestSetForceInlineMethod}, 1286 {CC"enqueueMethodForCompilation", 1287 CC"(Ljava/lang/reflect/Executable;II)Z", (void*)&WB_EnqueueMethodForCompilation}, 1288 {CC"clearMethodState", 1289 CC"(Ljava/lang/reflect/Executable;)V", (void*)&WB_ClearMethodState}, 1290 {CC"lockCompilation", CC"()V", (void*)&WB_LockCompilation}, 1291 {CC"unlockCompilation", CC"()V", (void*)&WB_UnlockCompilation}, 1292 {CC"isConstantVMFlag", CC"(Ljava/lang/String;)Z", (void*)&WB_IsConstantVMFlag}, 1293 {CC"isLockedVMFlag", CC"(Ljava/lang/String;)Z", (void*)&WB_IsLockedVMFlag}, 1294 {CC"setBooleanVMFlag", CC"(Ljava/lang/String;Z)V",(void*)&WB_SetBooleanVMFlag}, 1295 {CC"setIntxVMFlag", CC"(Ljava/lang/String;J)V",(void*)&WB_SetIntxVMFlag}, 1296 {CC"setUintxVMFlag", CC"(Ljava/lang/String;J)V",(void*)&WB_SetUintxVMFlag}, 1297 {CC"setUint64VMFlag", CC"(Ljava/lang/String;J)V",(void*)&WB_SetUint64VMFlag}, 1298 {CC"setSizeTVMFlag", CC"(Ljava/lang/String;J)V",(void*)&WB_SetSizeTVMFlag}, 1299 {CC"setDoubleVMFlag", CC"(Ljava/lang/String;D)V",(void*)&WB_SetDoubleVMFlag}, 1300 {CC"setStringVMFlag", CC"(Ljava/lang/String;Ljava/lang/String;)V", 1301 (void*)&WB_SetStringVMFlag}, 1302 {CC"getBooleanVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Boolean;", 1303 (void*)&WB_GetBooleanVMFlag}, 1304 {CC"getIntxVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Long;", 1305 (void*)&WB_GetIntxVMFlag}, 1306 {CC"getUintxVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Long;", 1307 (void*)&WB_GetUintxVMFlag}, 1308 {CC"getUint64VMFlag", CC"(Ljava/lang/String;)Ljava/lang/Long;", 1309 (void*)&WB_GetUint64VMFlag}, 1310 {CC"getSizeTVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Long;", 1311 (void*)&WB_GetSizeTVMFlag}, 1312 {CC"getDoubleVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Double;", 1313 (void*)&WB_GetDoubleVMFlag}, 1314 {CC"getStringVMFlag", CC"(Ljava/lang/String;)Ljava/lang/String;", 1315 (void*)&WB_GetStringVMFlag}, 1316 {CC"isInStringTable", CC"(Ljava/lang/String;)Z", (void*)&WB_IsInStringTable }, 1317 {CC"fullGC", CC"()V", (void*)&WB_FullGC }, 1318 {CC"youngGC", CC"()V", (void*)&WB_YoungGC }, 1319 {CC"readReservedMemory", CC"()V", (void*)&WB_ReadReservedMemory }, 1320 {CC"allocateMetaspace", 1321 CC"(Ljava/lang/ClassLoader;J)J", (void*)&WB_AllocateMetaspace }, 1322 {CC"freeMetaspace", 1323 CC"(Ljava/lang/ClassLoader;JJ)V", (void*)&WB_FreeMetaspace }, 1324 {CC"incMetaspaceCapacityUntilGC", CC"(J)J", (void*)&WB_IncMetaspaceCapacityUntilGC }, 1325 {CC"metaspaceCapacityUntilGC", CC"()J", (void*)&WB_MetaspaceCapacityUntilGC }, 1326 {CC"getCPUFeatures", CC"()Ljava/lang/String;", (void*)&WB_GetCPUFeatures }, 1327 {CC"getNMethod", CC"(Ljava/lang/reflect/Executable;Z)[Ljava/lang/Object;", 1328 (void*)&WB_GetNMethod }, 1329 {CC"forceNMethodSweep0", CC"()Ljava/lang/Thread;", (void*)&WB_ForceNMethodSweep }, 1330 {CC"allocateCodeBlob", CC"(II)J", (void*)&WB_AllocateCodeBlob }, 1331 {CC"freeCodeBlob", CC"(J)V", (void*)&WB_FreeCodeBlob }, 1332 {CC"getCodeHeapEntries", CC"(I)[Ljava/lang/Object;",(void*)&WB_GetCodeHeapEntries }, 1333 {CC"getCompilationActivityMode", 1334 CC"()I", (void*)&WB_GetCompilationActivityMode}, 1335 {CC"getCodeBlob", CC"(J)[Ljava/lang/Object;",(void*)&WB_GetCodeBlob }, 1336 {CC"getThreadStackSize", CC"()J", (void*)&WB_GetThreadStackSize }, 1337 {CC"getThreadRemainingStackSize", CC"()J", (void*)&WB_GetThreadRemainingStackSize }, 1338 {CC"assertMatchingSafepointCalls", CC"(ZZ)V", (void*)&WB_AssertMatchingSafepointCalls }, 1339 {CC"isMonitorInflated", CC"(Ljava/lang/Object;)Z", (void*)&WB_IsMonitorInflated }, 1340 {CC"forceSafepoint", CC"()V", (void*)&WB_ForceSafepoint }, 1341 }; 1342 1343 #undef CC 1344 1345 JVM_ENTRY(void, JVM_RegisterWhiteBoxMethods(JNIEnv* env, jclass wbclass)) 1346 { 1347 if (WhiteBoxAPI) { 1348 // Make sure that wbclass is loaded by the null classloader 1349 instanceKlassHandle ikh = instanceKlassHandle(JNIHandles::resolve(wbclass)->klass()); 1350 Handle loader(ikh->class_loader()); 1351 if (loader.is_null()) { 1352 WhiteBox::register_methods(env, wbclass, thread, methods, sizeof(methods) / sizeof(methods[0])); 1353 WhiteBox::register_extended(env, wbclass, thread); 1354 WhiteBox::set_used(); 1355 } 1356 } 1357 } 1358 JVM_END