1 /* 2 * Copyright (c) 2012, 2014, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 27 #include <new> 28 29 #include "code/codeCache.hpp" 30 #include "memory/metadataFactory.hpp" 31 #include "memory/universe.hpp" 32 #include "oops/oop.inline.hpp" 33 34 #include "classfile/stringTable.hpp" 35 #include "classfile/classLoaderData.hpp" 36 37 #include "prims/whitebox.hpp" 38 #include "prims/wbtestmethods/parserTests.hpp" 39 40 #include "runtime/thread.hpp" 41 #include "runtime/arguments.hpp" 42 #include "runtime/deoptimization.hpp" 43 #include "runtime/interfaceSupport.hpp" 44 #include "runtime/os.hpp" 45 #include "runtime/vm_version.hpp" 46 #include "runtime/sweeper.hpp" 47 48 #include "utilities/array.hpp" 49 #include "utilities/debug.hpp" 50 #include "utilities/macros.hpp" 51 #include "utilities/exceptions.hpp" 52 53 #if INCLUDE_ALL_GCS 54 #include "gc_implementation/parallelScavenge/parallelScavengeHeap.inline.hpp" 55 #include "gc_implementation/g1/concurrentMark.hpp" 56 #include "gc_implementation/g1/g1CollectedHeap.inline.hpp" 57 #include "gc_implementation/g1/heapRegionRemSet.hpp" 58 #endif // INCLUDE_ALL_GCS 59 60 #if INCLUDE_NMT 61 #include "services/mallocSiteTable.hpp" 62 #include "services/memTracker.hpp" 63 #include "utilities/nativeCallStack.hpp" 64 #endif // INCLUDE_NMT 65 66 #include "compiler/compileBroker.hpp" 67 #include "jvmtifiles/jvmtiEnv.hpp" 68 #include "runtime/compilationPolicy.hpp" 69 70 PRAGMA_FORMAT_MUTE_WARNINGS_FOR_GCC 71 72 #define SIZE_T_MAX_VALUE ((size_t) -1) 73 74 bool WhiteBox::_used = false; 75 volatile bool WhiteBox::compilation_locked = false; 76 77 WB_ENTRY(jlong, WB_GetObjectAddress(JNIEnv* env, jobject o, jobject obj)) 78 return (jlong)(void*)JNIHandles::resolve(obj); 79 WB_END 80 81 WB_ENTRY(jint, WB_GetHeapOopSize(JNIEnv* env, jobject o)) 82 return heapOopSize; 83 WB_END 84 85 WB_ENTRY(jint, WB_GetVMPageSize(JNIEnv* env, jobject o)) 86 return os::vm_page_size(); 87 WB_END 88 89 class WBIsKlassAliveClosure : public KlassClosure { 90 Symbol* _name; 91 bool _found; 92 public: 93 WBIsKlassAliveClosure(Symbol* name) : _name(name), _found(false) {} 94 95 void do_klass(Klass* k) { 96 if (_found) return; 97 Symbol* ksym = k->name(); 98 if (ksym->fast_compare(_name) == 0) { 99 _found = true; 100 } 101 } 102 103 bool found() const { 104 return _found; 105 } 106 }; 107 108 WB_ENTRY(jboolean, WB_IsClassAlive(JNIEnv* env, jobject target, jstring name)) 109 Handle h_name = JNIHandles::resolve(name); 110 if (h_name.is_null()) return false; 111 Symbol* sym = java_lang_String::as_symbol(h_name, CHECK_false); 112 TempNewSymbol tsym(sym); // Make sure to decrement reference count on sym on return 113 114 WBIsKlassAliveClosure closure(sym); 115 ClassLoaderDataGraph::classes_do(&closure); 116 117 return closure.found(); 118 WB_END 119 120 WB_ENTRY(void, WB_AddToBootstrapClassLoaderSearch(JNIEnv* env, jobject o, jstring segment)) { 121 #if INCLUDE_JVMTI 122 ResourceMark rm; 123 const char* seg = java_lang_String::as_utf8_string(JNIHandles::resolve_non_null(segment)); 124 JvmtiEnv* jvmti_env = JvmtiEnv::create_a_jvmti(JVMTI_VERSION); 125 jvmtiError err = jvmti_env->AddToBootstrapClassLoaderSearch(seg); 126 assert(err == JVMTI_ERROR_NONE, "must not fail"); 127 #endif 128 } 129 WB_END 130 131 WB_ENTRY(void, WB_AddToSystemClassLoaderSearch(JNIEnv* env, jobject o, jstring segment)) { 132 #if INCLUDE_JVMTI 133 ResourceMark rm; 134 const char* seg = java_lang_String::as_utf8_string(JNIHandles::resolve_non_null(segment)); 135 JvmtiEnv* jvmti_env = JvmtiEnv::create_a_jvmti(JVMTI_VERSION); 136 jvmtiError err = jvmti_env->AddToSystemClassLoaderSearch(seg); 137 assert(err == JVMTI_ERROR_NONE, "must not fail"); 138 #endif 139 } 140 WB_END 141 142 143 WB_ENTRY(jlong, WB_GetCompressedOopsMaxHeapSize(JNIEnv* env, jobject o)) { 144 return (jlong)Arguments::max_heap_for_compressed_oops(); 145 } 146 WB_END 147 148 WB_ENTRY(void, WB_PrintHeapSizes(JNIEnv* env, jobject o)) { 149 CollectorPolicy * p = Universe::heap()->collector_policy(); 150 gclog_or_tty->print_cr("Minimum heap "SIZE_FORMAT" Initial heap " 151 SIZE_FORMAT" Maximum heap "SIZE_FORMAT" Space alignment "SIZE_FORMAT" Heap alignment "SIZE_FORMAT, 152 p->min_heap_byte_size(), p->initial_heap_byte_size(), p->max_heap_byte_size(), 153 p->space_alignment(), p->heap_alignment()); 154 } 155 WB_END 156 157 #ifndef PRODUCT 158 // Forward declaration 159 void TestReservedSpace_test(); 160 void TestReserveMemorySpecial_test(); 161 void TestVirtualSpace_test(); 162 void TestMetaspaceAux_test(); 163 #endif 164 165 WB_ENTRY(void, WB_RunMemoryUnitTests(JNIEnv* env, jobject o)) 166 #ifndef PRODUCT 167 TestReservedSpace_test(); 168 TestReserveMemorySpecial_test(); 169 TestVirtualSpace_test(); 170 TestMetaspaceAux_test(); 171 #endif 172 WB_END 173 174 WB_ENTRY(void, WB_ReadFromNoaccessArea(JNIEnv* env, jobject o)) 175 size_t granularity = os::vm_allocation_granularity(); 176 ReservedHeapSpace rhs(100 * granularity, granularity, false, NULL); 177 VirtualSpace vs; 178 vs.initialize(rhs, 50 * granularity); 179 180 //Check if constraints are complied 181 if (!( UseCompressedOops && rhs.base() != NULL && 182 Universe::narrow_oop_base() != NULL && 183 Universe::narrow_oop_use_implicit_null_checks() )) { 184 tty->print_cr("WB_ReadFromNoaccessArea method is useless:\n " 185 "\tUseCompressedOops is %d\n" 186 "\trhs.base() is "PTR_FORMAT"\n" 187 "\tUniverse::narrow_oop_base() is "PTR_FORMAT"\n" 188 "\tUniverse::narrow_oop_use_implicit_null_checks() is %d", 189 UseCompressedOops, 190 rhs.base(), 191 Universe::narrow_oop_base(), 192 Universe::narrow_oop_use_implicit_null_checks()); 193 return; 194 } 195 tty->print_cr("Reading from no access area... "); 196 tty->print_cr("*(vs.low_boundary() - rhs.noaccess_prefix() / 2 ) = %c", 197 *(vs.low_boundary() - rhs.noaccess_prefix() / 2 )); 198 WB_END 199 200 static jint wb_stress_virtual_space_resize(size_t reserved_space_size, 201 size_t magnitude, size_t iterations) { 202 size_t granularity = os::vm_allocation_granularity(); 203 ReservedHeapSpace rhs(reserved_space_size * granularity, granularity, false, NULL); 204 VirtualSpace vs; 205 if (!vs.initialize(rhs, 0)) { 206 tty->print_cr("Failed to initialize VirtualSpace. Can't proceed."); 207 return 3; 208 } 209 210 long seed = os::random(); 211 tty->print_cr("Random seed is %ld", seed); 212 os::init_random(seed); 213 214 for (size_t i = 0; i < iterations; i++) { 215 216 // Whether we will shrink or grow 217 bool shrink = os::random() % 2L == 0; 218 219 // Get random delta to resize virtual space 220 size_t delta = (size_t)os::random() % magnitude; 221 222 // If we are about to shrink virtual space below zero, then expand instead 223 if (shrink && vs.committed_size() < delta) { 224 shrink = false; 225 } 226 227 // Resizing by delta 228 if (shrink) { 229 vs.shrink_by(delta); 230 } else { 231 // If expanding fails expand_by will silently return false 232 vs.expand_by(delta, true); 233 } 234 } 235 return 0; 236 } 237 238 WB_ENTRY(jint, WB_StressVirtualSpaceResize(JNIEnv* env, jobject o, 239 jlong reserved_space_size, jlong magnitude, jlong iterations)) 240 tty->print_cr("reservedSpaceSize="JLONG_FORMAT", magnitude="JLONG_FORMAT", " 241 "iterations="JLONG_FORMAT"\n", reserved_space_size, magnitude, 242 iterations); 243 if (reserved_space_size < 0 || magnitude < 0 || iterations < 0) { 244 tty->print_cr("One of variables printed above is negative. Can't proceed.\n"); 245 return 1; 246 } 247 248 // sizeof(size_t) depends on whether OS is 32bit or 64bit. sizeof(jlong) is 249 // always 8 byte. That's why we should avoid overflow in case of 32bit platform. 250 if (sizeof(size_t) < sizeof(jlong)) { 251 jlong size_t_max_value = (jlong) SIZE_T_MAX_VALUE; 252 if (reserved_space_size > size_t_max_value || magnitude > size_t_max_value 253 || iterations > size_t_max_value) { 254 tty->print_cr("One of variables printed above overflows size_t. Can't proceed.\n"); 255 return 2; 256 } 257 } 258 259 return wb_stress_virtual_space_resize((size_t) reserved_space_size, 260 (size_t) magnitude, (size_t) iterations); 261 WB_END 262 263 WB_ENTRY(jboolean, WB_isObjectInOldGen(JNIEnv* env, jobject o, jobject obj)) 264 oop p = JNIHandles::resolve(obj); 265 #if INCLUDE_ALL_GCS 266 if (UseG1GC) { 267 G1CollectedHeap* g1 = G1CollectedHeap::heap(); 268 const HeapRegion* hr = g1->heap_region_containing(p); 269 if (hr == NULL) { 270 return false; 271 } 272 return !(hr->is_young()); 273 } else if (UseParallelGC) { 274 ParallelScavengeHeap* psh = ParallelScavengeHeap::heap(); 275 return !psh->is_in_young(p); 276 } 277 #endif // INCLUDE_ALL_GCS 278 GenCollectedHeap* gch = GenCollectedHeap::heap(); 279 return !gch->is_in_young(p); 280 WB_END 281 282 WB_ENTRY(jlong, WB_GetObjectSize(JNIEnv* env, jobject o, jobject obj)) 283 oop p = JNIHandles::resolve(obj); 284 return p->size() * HeapWordSize; 285 WB_END 286 287 #if INCLUDE_ALL_GCS 288 WB_ENTRY(jboolean, WB_G1IsHumongous(JNIEnv* env, jobject o, jobject obj)) 289 G1CollectedHeap* g1 = G1CollectedHeap::heap(); 290 oop result = JNIHandles::resolve(obj); 291 const HeapRegion* hr = g1->heap_region_containing(result); 292 return hr->is_humongous(); 293 WB_END 294 295 WB_ENTRY(jlong, WB_G1NumFreeRegions(JNIEnv* env, jobject o)) 296 G1CollectedHeap* g1 = G1CollectedHeap::heap(); 297 size_t nr = g1->num_free_regions(); 298 return (jlong)nr; 299 WB_END 300 301 WB_ENTRY(jboolean, WB_G1InConcurrentMark(JNIEnv* env, jobject o)) 302 G1CollectedHeap* g1 = G1CollectedHeap::heap(); 303 ConcurrentMark* cm = g1->concurrent_mark(); 304 return cm->concurrent_marking_in_progress(); 305 WB_END 306 307 WB_ENTRY(jint, WB_G1RegionSize(JNIEnv* env, jobject o)) 308 return (jint)HeapRegion::GrainBytes; 309 WB_END 310 #endif // INCLUDE_ALL_GCS 311 312 #if INCLUDE_NMT 313 // Alloc memory using the test memory type so that we can use that to see if 314 // NMT picks it up correctly 315 WB_ENTRY(jlong, WB_NMTMalloc(JNIEnv* env, jobject o, jlong size)) 316 jlong addr = 0; 317 addr = (jlong)(uintptr_t)os::malloc(size, mtTest); 318 return addr; 319 WB_END 320 321 // Alloc memory with pseudo call stack. The test can create psudo malloc 322 // allocation site to stress the malloc tracking. 323 WB_ENTRY(jlong, WB_NMTMallocWithPseudoStack(JNIEnv* env, jobject o, jlong size, jint pseudo_stack)) 324 address pc = (address)(size_t)pseudo_stack; 325 NativeCallStack stack(&pc, 1); 326 return (jlong)(uintptr_t)os::malloc(size, mtTest, stack); 327 WB_END 328 329 // Free the memory allocated by NMTAllocTest 330 WB_ENTRY(void, WB_NMTFree(JNIEnv* env, jobject o, jlong mem)) 331 os::free((void*)(uintptr_t)mem); 332 WB_END 333 334 WB_ENTRY(jlong, WB_NMTReserveMemory(JNIEnv* env, jobject o, jlong size)) 335 jlong addr = 0; 336 337 addr = (jlong)(uintptr_t)os::reserve_memory(size); 338 MemTracker::record_virtual_memory_type((address)addr, mtTest); 339 340 return addr; 341 WB_END 342 343 WB_ENTRY(void, WB_NMTCommitMemory(JNIEnv* env, jobject o, jlong addr, jlong size)) 344 os::commit_memory((char *)(uintptr_t)addr, size, !ExecMem); 345 MemTracker::record_virtual_memory_type((address)(uintptr_t)addr, mtTest); 346 WB_END 347 348 WB_ENTRY(void, WB_NMTUncommitMemory(JNIEnv* env, jobject o, jlong addr, jlong size)) 349 os::uncommit_memory((char *)(uintptr_t)addr, size); 350 WB_END 351 352 WB_ENTRY(void, WB_NMTReleaseMemory(JNIEnv* env, jobject o, jlong addr, jlong size)) 353 os::release_memory((char *)(uintptr_t)addr, size); 354 WB_END 355 356 WB_ENTRY(jboolean, WB_NMTIsDetailSupported(JNIEnv* env)) 357 return MemTracker::tracking_level() == NMT_detail; 358 WB_END 359 360 WB_ENTRY(jboolean, WB_NMTChangeTrackingLevel(JNIEnv* env)) 361 // Test that we can downgrade NMT levels but not upgrade them. 362 if (MemTracker::tracking_level() == NMT_off) { 363 MemTracker::transition_to(NMT_off); 364 return MemTracker::tracking_level() == NMT_off; 365 } else { 366 assert(MemTracker::tracking_level() == NMT_detail, "Should start out as detail tracking"); 367 MemTracker::transition_to(NMT_summary); 368 assert(MemTracker::tracking_level() == NMT_summary, "Should be summary now"); 369 370 // Can't go to detail once NMT is set to summary. 371 MemTracker::transition_to(NMT_detail); 372 assert(MemTracker::tracking_level() == NMT_summary, "Should still be summary now"); 373 374 // Shutdown sets tracking level to minimal. 375 MemTracker::shutdown(); 376 assert(MemTracker::tracking_level() == NMT_minimal, "Should be minimal now"); 377 378 // Once the tracking level is minimal, we cannot increase to summary. 379 // The code ignores this request instead of asserting because if the malloc site 380 // table overflows in another thread, it tries to change the code to summary. 381 MemTracker::transition_to(NMT_summary); 382 assert(MemTracker::tracking_level() == NMT_minimal, "Should still be minimal now"); 383 384 // Really can never go up to detail, verify that the code would never do this. 385 MemTracker::transition_to(NMT_detail); 386 assert(MemTracker::tracking_level() == NMT_minimal, "Should still be minimal now"); 387 return MemTracker::tracking_level() == NMT_minimal; 388 } 389 WB_END 390 391 WB_ENTRY(jint, WB_NMTGetHashSize(JNIEnv* env, jobject o)) 392 int hash_size = MallocSiteTable::hash_buckets(); 393 assert(hash_size > 0, "NMT hash_size should be > 0"); 394 return (jint)hash_size; 395 WB_END 396 #endif // INCLUDE_NMT 397 398 static jmethodID reflected_method_to_jmid(JavaThread* thread, JNIEnv* env, jobject method) { 399 assert(method != NULL, "method should not be null"); 400 ThreadToNativeFromVM ttn(thread); 401 return env->FromReflectedMethod(method); 402 } 403 404 WB_ENTRY(void, WB_DeoptimizeAll(JNIEnv* env, jobject o)) 405 MutexLockerEx mu(Compile_lock); 406 CodeCache::mark_all_nmethods_for_deoptimization(); 407 VM_Deoptimize op; 408 VMThread::execute(&op); 409 WB_END 410 411 WB_ENTRY(jint, WB_DeoptimizeMethod(JNIEnv* env, jobject o, jobject method, jboolean is_osr)) 412 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 413 int result = 0; 414 CHECK_JNI_EXCEPTION_(env, result); 415 MutexLockerEx mu(Compile_lock); 416 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 417 if (is_osr) { 418 result += mh->mark_osr_nmethods(); 419 } else if (mh->code() != NULL) { 420 mh->code()->mark_for_deoptimization(); 421 ++result; 422 } 423 result += CodeCache::mark_for_deoptimization(mh()); 424 if (result > 0) { 425 VM_Deoptimize op; 426 VMThread::execute(&op); 427 } 428 return result; 429 WB_END 430 431 WB_ENTRY(jboolean, WB_IsMethodCompiled(JNIEnv* env, jobject o, jobject method, jboolean is_osr)) 432 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 433 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 434 MutexLockerEx mu(Compile_lock); 435 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 436 nmethod* code = is_osr ? mh->lookup_osr_nmethod_for(InvocationEntryBci, CompLevel_none, false) : mh->code(); 437 if (code == NULL) { 438 return JNI_FALSE; 439 } 440 return (code->is_alive() && !code->is_marked_for_deoptimization()); 441 WB_END 442 443 WB_ENTRY(jboolean, WB_IsMethodCompilable(JNIEnv* env, jobject o, jobject method, jint comp_level, jboolean is_osr)) 444 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 445 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 446 MutexLockerEx mu(Compile_lock); 447 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 448 if (is_osr) { 449 return CompilationPolicy::can_be_osr_compiled(mh, comp_level); 450 } else { 451 return CompilationPolicy::can_be_compiled(mh, comp_level); 452 } 453 WB_END 454 455 WB_ENTRY(jboolean, WB_IsMethodQueuedForCompilation(JNIEnv* env, jobject o, jobject method)) 456 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 457 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 458 MutexLockerEx mu(Compile_lock); 459 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 460 return mh->queued_for_compilation(); 461 WB_END 462 463 WB_ENTRY(jint, WB_GetMethodCompilationLevel(JNIEnv* env, jobject o, jobject method, jboolean is_osr)) 464 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 465 CHECK_JNI_EXCEPTION_(env, CompLevel_none); 466 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 467 nmethod* code = is_osr ? mh->lookup_osr_nmethod_for(InvocationEntryBci, CompLevel_none, false) : mh->code(); 468 return (code != NULL ? code->comp_level() : CompLevel_none); 469 WB_END 470 471 WB_ENTRY(void, WB_MakeMethodNotCompilable(JNIEnv* env, jobject o, jobject method, jint comp_level, jboolean is_osr)) 472 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 473 CHECK_JNI_EXCEPTION(env); 474 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 475 if (is_osr) { 476 mh->set_not_osr_compilable(comp_level, true /* report */, "WhiteBox"); 477 } else { 478 mh->set_not_compilable(comp_level, true /* report */, "WhiteBox"); 479 } 480 WB_END 481 482 WB_ENTRY(jint, WB_GetMethodEntryBci(JNIEnv* env, jobject o, jobject method)) 483 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 484 CHECK_JNI_EXCEPTION_(env, InvocationEntryBci); 485 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 486 nmethod* code = mh->lookup_osr_nmethod_for(InvocationEntryBci, CompLevel_none, false); 487 return (code != NULL && code->is_osr_method() ? code->osr_entry_bci() : InvocationEntryBci); 488 WB_END 489 490 WB_ENTRY(jboolean, WB_TestSetDontInlineMethod(JNIEnv* env, jobject o, jobject method, jboolean value)) 491 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 492 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 493 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 494 bool result = mh->dont_inline(); 495 mh->set_dont_inline(value == JNI_TRUE); 496 return result; 497 WB_END 498 499 WB_ENTRY(jint, WB_GetCompileQueueSize(JNIEnv* env, jobject o, jint comp_level)) 500 if (comp_level == CompLevel_any) { 501 return CompileBroker::queue_size(CompLevel_full_optimization) /* C2 */ + 502 CompileBroker::queue_size(CompLevel_full_profile) /* C1 */; 503 } else { 504 return CompileBroker::queue_size(comp_level); 505 } 506 WB_END 507 508 WB_ENTRY(jboolean, WB_TestSetForceInlineMethod(JNIEnv* env, jobject o, jobject method, jboolean value)) 509 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 510 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 511 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 512 bool result = mh->force_inline(); 513 mh->set_force_inline(value == JNI_TRUE); 514 return result; 515 WB_END 516 517 WB_ENTRY(jboolean, WB_EnqueueMethodForCompilation(JNIEnv* env, jobject o, jobject method, jint comp_level, jint bci)) 518 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 519 CHECK_JNI_EXCEPTION_(env, JNI_FALSE); 520 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 521 nmethod* nm = CompileBroker::compile_method(mh, bci, comp_level, mh, mh->invocation_count(), "WhiteBox", THREAD); 522 MutexLockerEx mu(Compile_lock); 523 return (mh->queued_for_compilation() || nm != NULL); 524 WB_END 525 526 class VM_WhiteBoxOperation : public VM_Operation { 527 public: 528 VM_WhiteBoxOperation() { } 529 VMOp_Type type() const { return VMOp_WhiteBoxOperation; } 530 bool allow_nested_vm_operations() const { return true; } 531 }; 532 533 class AlwaysFalseClosure : public BoolObjectClosure { 534 public: 535 bool do_object_b(oop p) { return false; } 536 }; 537 538 static AlwaysFalseClosure always_false; 539 540 WB_ENTRY(void, WB_ClearMethodState(JNIEnv* env, jobject o, jobject method)) 541 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 542 CHECK_JNI_EXCEPTION(env); 543 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 544 MutexLockerEx mu(Compile_lock); 545 MethodData* mdo = mh->method_data(); 546 MethodCounters* mcs = mh->method_counters(); 547 548 if (mdo != NULL) { 549 mdo->init(); 550 ResourceMark rm; 551 int arg_count = mdo->method()->size_of_parameters(); 552 for (int i = 0; i < arg_count; i++) { 553 mdo->set_arg_modified(i, 0); 554 } 555 MutexLockerEx mu(mdo->extra_data_lock()); 556 mdo->clean_method_data(&always_false); 557 } 558 559 mh->clear_not_c1_compilable(); 560 mh->clear_not_c2_compilable(); 561 mh->clear_not_c2_osr_compilable(); 562 NOT_PRODUCT(mh->set_compiled_invocation_count(0)); 563 if (mcs != NULL) { 564 mcs->backedge_counter()->init(); 565 mcs->invocation_counter()->init(); 566 mcs->set_interpreter_invocation_count(0); 567 mcs->set_interpreter_throwout_count(0); 568 569 #ifdef TIERED 570 mcs->set_rate(0.0F); 571 mh->set_prev_event_count(0); 572 mh->set_prev_time(0); 573 #endif 574 } 575 WB_END 576 577 template <typename T> 578 static bool GetVMFlag(JavaThread* thread, JNIEnv* env, jstring name, T* value, bool (*TAt)(const char*, T*, bool, bool)) { 579 if (name == NULL) { 580 return false; 581 } 582 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 583 const char* flag_name = env->GetStringUTFChars(name, NULL); 584 bool result = (*TAt)(flag_name, value, true, true); 585 env->ReleaseStringUTFChars(name, flag_name); 586 return result; 587 } 588 589 template <typename T> 590 static bool SetVMFlag(JavaThread* thread, JNIEnv* env, jstring name, T* value, bool (*TAtPut)(const char*, T*, Flag::Flags)) { 591 if (name == NULL) { 592 return false; 593 } 594 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 595 const char* flag_name = env->GetStringUTFChars(name, NULL); 596 bool result = (*TAtPut)(flag_name, value, Flag::INTERNAL); 597 env->ReleaseStringUTFChars(name, flag_name); 598 return result; 599 } 600 601 template <typename T> 602 static jobject box(JavaThread* thread, JNIEnv* env, Symbol* name, Symbol* sig, T value) { 603 ResourceMark rm(thread); 604 jclass clazz = env->FindClass(name->as_C_string()); 605 CHECK_JNI_EXCEPTION_(env, NULL); 606 jmethodID methodID = env->GetStaticMethodID(clazz, 607 vmSymbols::valueOf_name()->as_C_string(), 608 sig->as_C_string()); 609 CHECK_JNI_EXCEPTION_(env, NULL); 610 jobject result = env->CallStaticObjectMethod(clazz, methodID, value); 611 CHECK_JNI_EXCEPTION_(env, NULL); 612 return result; 613 } 614 615 static jobject booleanBox(JavaThread* thread, JNIEnv* env, jboolean value) { 616 return box(thread, env, vmSymbols::java_lang_Boolean(), vmSymbols::Boolean_valueOf_signature(), value); 617 } 618 static jobject integerBox(JavaThread* thread, JNIEnv* env, jint value) { 619 return box(thread, env, vmSymbols::java_lang_Integer(), vmSymbols::Integer_valueOf_signature(), value); 620 } 621 static jobject longBox(JavaThread* thread, JNIEnv* env, jlong value) { 622 return box(thread, env, vmSymbols::java_lang_Long(), vmSymbols::Long_valueOf_signature(), value); 623 } 624 /* static jobject floatBox(JavaThread* thread, JNIEnv* env, jfloat value) { 625 return box(thread, env, vmSymbols::java_lang_Float(), vmSymbols::Float_valueOf_signature(), value); 626 }*/ 627 static jobject doubleBox(JavaThread* thread, JNIEnv* env, jdouble value) { 628 return box(thread, env, vmSymbols::java_lang_Double(), vmSymbols::Double_valueOf_signature(), value); 629 } 630 631 static Flag* getVMFlag(JavaThread* thread, JNIEnv* env, jstring name) { 632 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 633 const char* flag_name = env->GetStringUTFChars(name, NULL); 634 Flag* result = Flag::find_flag(flag_name, strlen(flag_name), true, true); 635 env->ReleaseStringUTFChars(name, flag_name); 636 return result; 637 } 638 639 WB_ENTRY(jboolean, WB_IsConstantVMFlag(JNIEnv* env, jobject o, jstring name)) 640 Flag* flag = getVMFlag(thread, env, name); 641 return (flag != NULL) && flag->is_constant_in_binary(); 642 WB_END 643 644 WB_ENTRY(jboolean, WB_IsLockedVMFlag(JNIEnv* env, jobject o, jstring name)) 645 Flag* flag = getVMFlag(thread, env, name); 646 return (flag != NULL) && !(flag->is_unlocked() || flag->is_unlocker()); 647 WB_END 648 649 WB_ENTRY(jobject, WB_GetBooleanVMFlag(JNIEnv* env, jobject o, jstring name)) 650 bool result; 651 if (GetVMFlag <bool> (thread, env, name, &result, &CommandLineFlags::boolAt)) { 652 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 653 return booleanBox(thread, env, result); 654 } 655 return NULL; 656 WB_END 657 658 WB_ENTRY(jobject, WB_GetIntxVMFlag(JNIEnv* env, jobject o, jstring name)) 659 intx result; 660 if (GetVMFlag <intx> (thread, env, name, &result, &CommandLineFlags::intxAt)) { 661 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 662 return longBox(thread, env, result); 663 } 664 return NULL; 665 WB_END 666 667 WB_ENTRY(jobject, WB_GetUintxVMFlag(JNIEnv* env, jobject o, jstring name)) 668 uintx result; 669 if (GetVMFlag <uintx> (thread, env, name, &result, &CommandLineFlags::uintxAt)) { 670 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 671 return longBox(thread, env, result); 672 } 673 return NULL; 674 WB_END 675 676 WB_ENTRY(jobject, WB_GetUint64VMFlag(JNIEnv* env, jobject o, jstring name)) 677 uint64_t result; 678 if (GetVMFlag <uint64_t> (thread, env, name, &result, &CommandLineFlags::uint64_tAt)) { 679 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 680 return longBox(thread, env, result); 681 } 682 return NULL; 683 WB_END 684 685 WB_ENTRY(jobject, WB_GetSizeTVMFlag(JNIEnv* env, jobject o, jstring name)) 686 uintx result; 687 if (GetVMFlag <size_t> (thread, env, name, &result, &CommandLineFlags::size_tAt)) { 688 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 689 return longBox(thread, env, result); 690 } 691 return NULL; 692 WB_END 693 694 WB_ENTRY(jobject, WB_GetDoubleVMFlag(JNIEnv* env, jobject o, jstring name)) 695 double result; 696 if (GetVMFlag <double> (thread, env, name, &result, &CommandLineFlags::doubleAt)) { 697 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 698 return doubleBox(thread, env, result); 699 } 700 return NULL; 701 WB_END 702 703 WB_ENTRY(jstring, WB_GetStringVMFlag(JNIEnv* env, jobject o, jstring name)) 704 ccstr ccstrResult; 705 if (GetVMFlag <ccstr> (thread, env, name, &ccstrResult, &CommandLineFlags::ccstrAt)) { 706 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 707 jstring result = env->NewStringUTF(ccstrResult); 708 CHECK_JNI_EXCEPTION_(env, NULL); 709 return result; 710 } 711 return NULL; 712 WB_END 713 714 WB_ENTRY(void, WB_SetBooleanVMFlag(JNIEnv* env, jobject o, jstring name, jboolean value)) 715 bool result = value == JNI_TRUE ? true : false; 716 SetVMFlag <bool> (thread, env, name, &result, &CommandLineFlags::boolAtPut); 717 WB_END 718 719 WB_ENTRY(void, WB_SetIntxVMFlag(JNIEnv* env, jobject o, jstring name, jlong value)) 720 intx result = value; 721 SetVMFlag <intx> (thread, env, name, &result, &CommandLineFlags::intxAtPut); 722 WB_END 723 724 WB_ENTRY(void, WB_SetUintxVMFlag(JNIEnv* env, jobject o, jstring name, jlong value)) 725 uintx result = value; 726 SetVMFlag <uintx> (thread, env, name, &result, &CommandLineFlags::uintxAtPut); 727 WB_END 728 729 WB_ENTRY(void, WB_SetUint64VMFlag(JNIEnv* env, jobject o, jstring name, jlong value)) 730 uint64_t result = value; 731 SetVMFlag <uint64_t> (thread, env, name, &result, &CommandLineFlags::uint64_tAtPut); 732 WB_END 733 734 WB_ENTRY(void, WB_SetSizeTVMFlag(JNIEnv* env, jobject o, jstring name, jlong value)) 735 size_t result = value; 736 SetVMFlag <size_t> (thread, env, name, &result, &CommandLineFlags::size_tAtPut); 737 WB_END 738 739 WB_ENTRY(void, WB_SetDoubleVMFlag(JNIEnv* env, jobject o, jstring name, jdouble value)) 740 double result = value; 741 SetVMFlag <double> (thread, env, name, &result, &CommandLineFlags::doubleAtPut); 742 WB_END 743 744 WB_ENTRY(void, WB_SetStringVMFlag(JNIEnv* env, jobject o, jstring name, jstring value)) 745 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 746 const char* ccstrValue = (value == NULL) ? NULL : env->GetStringUTFChars(value, NULL); 747 ccstr ccstrResult = ccstrValue; 748 bool needFree; 749 { 750 ThreadInVMfromNative ttvfn(thread); // back to VM 751 needFree = SetVMFlag <ccstr> (thread, env, name, &ccstrResult, &CommandLineFlags::ccstrAtPut); 752 } 753 if (value != NULL) { 754 env->ReleaseStringUTFChars(value, ccstrValue); 755 } 756 if (needFree) { 757 FREE_C_HEAP_ARRAY(char, ccstrResult); 758 } 759 WB_END 760 761 762 WB_ENTRY(void, WB_LockCompilation(JNIEnv* env, jobject o, jlong timeout)) 763 WhiteBox::compilation_locked = true; 764 WB_END 765 766 WB_ENTRY(void, WB_UnlockCompilation(JNIEnv* env, jobject o)) 767 MonitorLockerEx mo(Compilation_lock, Mutex::_no_safepoint_check_flag); 768 WhiteBox::compilation_locked = false; 769 mo.notify_all(); 770 WB_END 771 772 void WhiteBox::force_sweep() { 773 guarantee(WhiteBoxAPI, "internal testing API :: WhiteBox has to enabled"); 774 { 775 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 776 NMethodSweeper::_should_sweep = true; 777 } 778 NMethodSweeper::possibly_sweep(); 779 } 780 781 WB_ENTRY(void, WB_ForceNMethodSweep(JNIEnv* env, jobject o)) 782 WhiteBox::force_sweep(); 783 WB_END 784 785 WB_ENTRY(jboolean, WB_IsInStringTable(JNIEnv* env, jobject o, jstring javaString)) 786 ResourceMark rm(THREAD); 787 int len; 788 jchar* name = java_lang_String::as_unicode_string(JNIHandles::resolve(javaString), len, CHECK_false); 789 return (StringTable::lookup(name, len) != NULL); 790 WB_END 791 792 WB_ENTRY(void, WB_FullGC(JNIEnv* env, jobject o)) 793 Universe::heap()->collector_policy()->set_should_clear_all_soft_refs(true); 794 Universe::heap()->collect(GCCause::_last_ditch_collection); 795 #if INCLUDE_ALL_GCS 796 if (UseG1GC) { 797 // Needs to be cleared explicitly for G1 798 Universe::heap()->collector_policy()->set_should_clear_all_soft_refs(false); 799 } 800 #endif // INCLUDE_ALL_GCS 801 WB_END 802 803 WB_ENTRY(void, WB_YoungGC(JNIEnv* env, jobject o)) 804 Universe::heap()->collect(GCCause::_wb_young_gc); 805 WB_END 806 807 WB_ENTRY(void, WB_ReadReservedMemory(JNIEnv* env, jobject o)) 808 // static+volatile in order to force the read to happen 809 // (not be eliminated by the compiler) 810 static char c; 811 static volatile char* p; 812 813 p = os::reserve_memory(os::vm_allocation_granularity(), NULL, 0); 814 if (p == NULL) { 815 THROW_MSG(vmSymbols::java_lang_OutOfMemoryError(), "Failed to reserve memory"); 816 } 817 818 c = *p; 819 WB_END 820 821 WB_ENTRY(jstring, WB_GetCPUFeatures(JNIEnv* env, jobject o)) 822 const char* cpu_features = VM_Version::cpu_features(); 823 ThreadToNativeFromVM ttn(thread); 824 jstring features_string = env->NewStringUTF(cpu_features); 825 826 CHECK_JNI_EXCEPTION_(env, NULL); 827 828 return features_string; 829 WB_END 830 831 int WhiteBox::get_blob_type(const CodeBlob* code) { 832 guarantee(WhiteBoxAPI, "internal testing API :: WhiteBox has to enabled"); 833 return CodeCache::get_code_heap(code)->code_blob_type(); 834 } 835 836 CodeHeap* WhiteBox::get_code_heap(int blob_type) { 837 guarantee(WhiteBoxAPI, "internal testing API :: WhiteBox has to enabled"); 838 return CodeCache::get_code_heap(blob_type); 839 } 840 841 struct CodeBlobStub { 842 CodeBlobStub(const CodeBlob* blob) : 843 name(os::strdup(blob->name())), 844 size(blob->size()), 845 blob_type(WhiteBox::get_blob_type(blob)) { } 846 ~CodeBlobStub() { os::free((void*) name); } 847 const char* const name; 848 const int size; 849 const int blob_type; 850 }; 851 852 static jobjectArray codeBlob2objectArray(JavaThread* thread, JNIEnv* env, CodeBlobStub* cb) { 853 jclass clazz = env->FindClass(vmSymbols::java_lang_Object()->as_C_string()); 854 CHECK_JNI_EXCEPTION_(env, NULL); 855 jobjectArray result = env->NewObjectArray(3, clazz, NULL); 856 857 jstring name = env->NewStringUTF(cb->name); 858 CHECK_JNI_EXCEPTION_(env, NULL); 859 env->SetObjectArrayElement(result, 0, name); 860 861 jobject obj = integerBox(thread, env, cb->size); 862 CHECK_JNI_EXCEPTION_(env, NULL); 863 env->SetObjectArrayElement(result, 1, obj); 864 865 obj = integerBox(thread, env, cb->blob_type); 866 CHECK_JNI_EXCEPTION_(env, NULL); 867 env->SetObjectArrayElement(result, 2, obj); 868 869 return result; 870 } 871 872 WB_ENTRY(jobjectArray, WB_GetNMethod(JNIEnv* env, jobject o, jobject method, jboolean is_osr)) 873 ResourceMark rm(THREAD); 874 jmethodID jmid = reflected_method_to_jmid(thread, env, method); 875 CHECK_JNI_EXCEPTION_(env, NULL); 876 methodHandle mh(THREAD, Method::checked_resolve_jmethod_id(jmid)); 877 nmethod* code = is_osr ? mh->lookup_osr_nmethod_for(InvocationEntryBci, CompLevel_none, false) : mh->code(); 878 jobjectArray result = NULL; 879 if (code == NULL) { 880 return result; 881 } 882 int insts_size = code->insts_size(); 883 884 ThreadToNativeFromVM ttn(thread); 885 jclass clazz = env->FindClass(vmSymbols::java_lang_Object()->as_C_string()); 886 CHECK_JNI_EXCEPTION_(env, NULL); 887 result = env->NewObjectArray(4, clazz, NULL); 888 if (result == NULL) { 889 return result; 890 } 891 892 CodeBlobStub stub(code); 893 jobjectArray codeBlob = codeBlob2objectArray(thread, env, &stub); 894 env->SetObjectArrayElement(result, 0, codeBlob); 895 896 jobject level = integerBox(thread, env, code->comp_level()); 897 CHECK_JNI_EXCEPTION_(env, NULL); 898 env->SetObjectArrayElement(result, 1, level); 899 900 jbyteArray insts = env->NewByteArray(insts_size); 901 CHECK_JNI_EXCEPTION_(env, NULL); 902 env->SetByteArrayRegion(insts, 0, insts_size, (jbyte*) code->insts_begin()); 903 env->SetObjectArrayElement(result, 2, insts); 904 905 jobject id = integerBox(thread, env, code->compile_id()); 906 CHECK_JNI_EXCEPTION_(env, NULL); 907 env->SetObjectArrayElement(result, 3, id); 908 909 return result; 910 WB_END 911 912 CodeBlob* WhiteBox::allocate_code_blob(int size, int blob_type) { 913 guarantee(WhiteBoxAPI, "internal testing API :: WhiteBox has to enabled"); 914 BufferBlob* blob; 915 int full_size = CodeBlob::align_code_offset(sizeof(BufferBlob)); 916 if (full_size < size) { 917 full_size += round_to(size - full_size, oopSize); 918 } 919 { 920 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 921 blob = (BufferBlob*) CodeCache::allocate(full_size, blob_type); 922 } 923 // Track memory usage statistic after releasing CodeCache_lock 924 MemoryService::track_code_cache_memory_usage(); 925 ::new (blob) BufferBlob("WB::DummyBlob", full_size); 926 return blob; 927 } 928 929 WB_ENTRY(jlong, WB_AllocateCodeBlob(JNIEnv* env, jobject o, jint size, jint blob_type)) 930 return (jlong) WhiteBox::allocate_code_blob(size, blob_type); 931 WB_END 932 933 WB_ENTRY(void, WB_FreeCodeBlob(JNIEnv* env, jobject o, jlong addr)) 934 BufferBlob::free((BufferBlob*) addr); 935 WB_END 936 937 WB_ENTRY(jobjectArray, WB_GetCodeHeapEntries(JNIEnv* env, jobject o, jint blob_type)) 938 ResourceMark rm; 939 GrowableArray<CodeBlobStub*> blobs; 940 { 941 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 942 CodeHeap* heap = WhiteBox::get_code_heap(blob_type); 943 if (heap == NULL) { 944 return NULL; 945 } 946 for (CodeBlob* cb = (CodeBlob*) heap->first(); 947 cb != NULL; cb = (CodeBlob*) heap->next(cb)) { 948 CodeBlobStub* stub = NEW_RESOURCE_OBJ(CodeBlobStub); 949 new (stub) CodeBlobStub(cb); 950 blobs.append(stub); 951 } 952 } 953 if (blobs.length() == 0) { 954 return NULL; 955 } 956 ThreadToNativeFromVM ttn(thread); 957 jobjectArray result = NULL; 958 jclass clazz = env->FindClass(vmSymbols::java_lang_Object()->as_C_string()); 959 CHECK_JNI_EXCEPTION_(env, NULL); 960 result = env->NewObjectArray(blobs.length(), clazz, NULL); 961 if (result == NULL) { 962 return result; 963 } 964 int i = 0; 965 for (GrowableArrayIterator<CodeBlobStub*> it = blobs.begin(); 966 it != blobs.end(); ++it) { 967 jobjectArray obj = codeBlob2objectArray(thread, env, *it); 968 env->SetObjectArrayElement(result, i, obj); 969 ++i; 970 } 971 return result; 972 WB_END 973 974 WB_ENTRY(jlong, WB_GetThreadStackSize(JNIEnv* env, jobject o)) 975 return (jlong) Thread::current()->stack_size(); 976 WB_END 977 978 WB_ENTRY(jlong, WB_GetThreadRemainingStackSize(JNIEnv* env, jobject o)) 979 JavaThread* t = JavaThread::current(); 980 return (jlong) t->stack_available(os::current_stack_pointer()) - (jlong) StackShadowPages * os::vm_page_size(); 981 WB_END 982 983 int WhiteBox::array_bytes_to_length(size_t bytes) { 984 return Array<u1>::bytes_to_length(bytes); 985 } 986 987 WB_ENTRY(jlong, WB_AllocateMetaspace(JNIEnv* env, jobject wb, jobject class_loader, jlong size)) 988 if (size < 0) { 989 THROW_MSG_0(vmSymbols::java_lang_IllegalArgumentException(), 990 err_msg("WB_AllocateMetaspace: size is negative: " JLONG_FORMAT, size)); 991 } 992 993 oop class_loader_oop = JNIHandles::resolve(class_loader); 994 ClassLoaderData* cld = class_loader_oop != NULL 995 ? java_lang_ClassLoader::loader_data(class_loader_oop) 996 : ClassLoaderData::the_null_class_loader_data(); 997 998 void* metadata = MetadataFactory::new_writeable_array<u1>(cld, WhiteBox::array_bytes_to_length((size_t)size), thread); 999 1000 return (jlong)(uintptr_t)metadata; 1001 WB_END 1002 1003 WB_ENTRY(void, WB_FreeMetaspace(JNIEnv* env, jobject wb, jobject class_loader, jlong addr, jlong size)) 1004 oop class_loader_oop = JNIHandles::resolve(class_loader); 1005 ClassLoaderData* cld = class_loader_oop != NULL 1006 ? java_lang_ClassLoader::loader_data(class_loader_oop) 1007 : ClassLoaderData::the_null_class_loader_data(); 1008 1009 MetadataFactory::free_array(cld, (Array<u1>*)(uintptr_t)addr); 1010 WB_END 1011 1012 WB_ENTRY(jlong, WB_IncMetaspaceCapacityUntilGC(JNIEnv* env, jobject wb, jlong inc)) 1013 if (inc < 0) { 1014 THROW_MSG_0(vmSymbols::java_lang_IllegalArgumentException(), 1015 err_msg("WB_IncMetaspaceCapacityUntilGC: inc is negative: " JLONG_FORMAT, inc)); 1016 } 1017 1018 jlong max_size_t = (jlong) ((size_t) -1); 1019 if (inc > max_size_t) { 1020 THROW_MSG_0(vmSymbols::java_lang_IllegalArgumentException(), 1021 err_msg("WB_IncMetaspaceCapacityUntilGC: inc does not fit in size_t: " JLONG_FORMAT, inc)); 1022 } 1023 1024 size_t new_cap_until_GC = 0; 1025 size_t aligned_inc = align_size_down((size_t) inc, Metaspace::commit_alignment()); 1026 bool success = MetaspaceGC::inc_capacity_until_GC(aligned_inc, &new_cap_until_GC); 1027 if (!success) { 1028 THROW_MSG_0(vmSymbols::java_lang_IllegalStateException(), 1029 "WB_IncMetaspaceCapacityUntilGC: could not increase capacity until GC " 1030 "due to contention with another thread"); 1031 } 1032 return (jlong) new_cap_until_GC; 1033 WB_END 1034 1035 WB_ENTRY(jlong, WB_MetaspaceCapacityUntilGC(JNIEnv* env, jobject wb)) 1036 return (jlong) MetaspaceGC::capacity_until_GC(); 1037 WB_END 1038 1039 //Some convenience methods to deal with objects from java 1040 int WhiteBox::offset_for_field(const char* field_name, oop object, 1041 Symbol* signature_symbol) { 1042 assert(field_name != NULL && strlen(field_name) > 0, "Field name not valid"); 1043 Thread* THREAD = Thread::current(); 1044 1045 //Get the class of our object 1046 Klass* arg_klass = object->klass(); 1047 //Turn it into an instance-klass 1048 InstanceKlass* ik = InstanceKlass::cast(arg_klass); 1049 1050 //Create symbols to look for in the class 1051 TempNewSymbol name_symbol = SymbolTable::lookup(field_name, (int) strlen(field_name), 1052 THREAD); 1053 1054 //To be filled in with an offset of the field we're looking for 1055 fieldDescriptor fd; 1056 1057 Klass* res = ik->find_field(name_symbol, signature_symbol, &fd); 1058 if (res == NULL) { 1059 tty->print_cr("Invalid layout of %s at %s", ik->external_name(), 1060 name_symbol->as_C_string()); 1061 fatal("Invalid layout of preloaded class"); 1062 } 1063 1064 //fetch the field at the offset we've found 1065 int dest_offset = fd.offset(); 1066 1067 return dest_offset; 1068 } 1069 1070 1071 const char* WhiteBox::lookup_jstring(const char* field_name, oop object) { 1072 int offset = offset_for_field(field_name, object, 1073 vmSymbols::string_signature()); 1074 oop string = object->obj_field(offset); 1075 if (string == NULL) { 1076 return NULL; 1077 } 1078 const char* ret = java_lang_String::as_utf8_string(string); 1079 return ret; 1080 } 1081 1082 bool WhiteBox::lookup_bool(const char* field_name, oop object) { 1083 int offset = 1084 offset_for_field(field_name, object, vmSymbols::bool_signature()); 1085 bool ret = (object->bool_field(offset) == JNI_TRUE); 1086 return ret; 1087 } 1088 1089 void WhiteBox::register_methods(JNIEnv* env, jclass wbclass, JavaThread* thread, JNINativeMethod* method_array, int method_count) { 1090 ResourceMark rm; 1091 ThreadToNativeFromVM ttnfv(thread); // can't be in VM when we call JNI 1092 1093 // one by one registration natives for exception catching 1094 jclass no_such_method_error_klass = env->FindClass(vmSymbols::java_lang_NoSuchMethodError()->as_C_string()); 1095 CHECK_JNI_EXCEPTION(env); 1096 for (int i = 0, n = method_count; i < n; ++i) { 1097 // Skip dummy entries 1098 if (method_array[i].fnPtr == NULL) continue; 1099 if (env->RegisterNatives(wbclass, &method_array[i], 1) != 0) { 1100 jthrowable throwable_obj = env->ExceptionOccurred(); 1101 if (throwable_obj != NULL) { 1102 env->ExceptionClear(); 1103 if (env->IsInstanceOf(throwable_obj, no_such_method_error_klass)) { 1104 // NoSuchMethodError is thrown when a method can't be found or a method is not native. 1105 // Ignoring the exception since it is not preventing use of other WhiteBox methods. 1106 tty->print_cr("Warning: 'NoSuchMethodError' on register of sun.hotspot.WhiteBox::%s%s", 1107 method_array[i].name, method_array[i].signature); 1108 } 1109 } else { 1110 // Registration failed unexpectedly. 1111 tty->print_cr("Warning: unexpected error on register of sun.hotspot.WhiteBox::%s%s. All methods will be unregistered", 1112 method_array[i].name, method_array[i].signature); 1113 env->UnregisterNatives(wbclass); 1114 break; 1115 } 1116 } 1117 } 1118 } 1119 1120 #define CC (char*) 1121 1122 static JNINativeMethod methods[] = { 1123 {CC"getObjectAddress", CC"(Ljava/lang/Object;)J", (void*)&WB_GetObjectAddress }, 1124 {CC"getObjectSize", CC"(Ljava/lang/Object;)J", (void*)&WB_GetObjectSize }, 1125 {CC"isObjectInOldGen", CC"(Ljava/lang/Object;)Z", (void*)&WB_isObjectInOldGen }, 1126 {CC"getHeapOopSize", CC"()I", (void*)&WB_GetHeapOopSize }, 1127 {CC"getVMPageSize", CC"()I", (void*)&WB_GetVMPageSize }, 1128 {CC"isClassAlive0", CC"(Ljava/lang/String;)Z", (void*)&WB_IsClassAlive }, 1129 {CC"parseCommandLine", 1130 CC"(Ljava/lang/String;C[Lsun/hotspot/parser/DiagnosticCommand;)[Ljava/lang/Object;", 1131 (void*) &WB_ParseCommandLine 1132 }, 1133 {CC"addToBootstrapClassLoaderSearch", CC"(Ljava/lang/String;)V", 1134 (void*)&WB_AddToBootstrapClassLoaderSearch}, 1135 {CC"addToSystemClassLoaderSearch", CC"(Ljava/lang/String;)V", 1136 (void*)&WB_AddToSystemClassLoaderSearch}, 1137 {CC"getCompressedOopsMaxHeapSize", CC"()J", 1138 (void*)&WB_GetCompressedOopsMaxHeapSize}, 1139 {CC"printHeapSizes", CC"()V", (void*)&WB_PrintHeapSizes }, 1140 {CC"runMemoryUnitTests", CC"()V", (void*)&WB_RunMemoryUnitTests}, 1141 {CC"readFromNoaccessArea",CC"()V", (void*)&WB_ReadFromNoaccessArea}, 1142 {CC"stressVirtualSpaceResize",CC"(JJJ)I", (void*)&WB_StressVirtualSpaceResize}, 1143 #if INCLUDE_ALL_GCS 1144 {CC"g1InConcurrentMark", CC"()Z", (void*)&WB_G1InConcurrentMark}, 1145 {CC"g1IsHumongous", CC"(Ljava/lang/Object;)Z", (void*)&WB_G1IsHumongous }, 1146 {CC"g1NumFreeRegions", CC"()J", (void*)&WB_G1NumFreeRegions }, 1147 {CC"g1RegionSize", CC"()I", (void*)&WB_G1RegionSize }, 1148 #endif // INCLUDE_ALL_GCS 1149 #if INCLUDE_NMT 1150 {CC"NMTMalloc", CC"(J)J", (void*)&WB_NMTMalloc }, 1151 {CC"NMTMallocWithPseudoStack", CC"(JI)J", (void*)&WB_NMTMallocWithPseudoStack}, 1152 {CC"NMTFree", CC"(J)V", (void*)&WB_NMTFree }, 1153 {CC"NMTReserveMemory", CC"(J)J", (void*)&WB_NMTReserveMemory }, 1154 {CC"NMTCommitMemory", CC"(JJ)V", (void*)&WB_NMTCommitMemory }, 1155 {CC"NMTUncommitMemory", CC"(JJ)V", (void*)&WB_NMTUncommitMemory }, 1156 {CC"NMTReleaseMemory", CC"(JJ)V", (void*)&WB_NMTReleaseMemory }, 1157 {CC"NMTIsDetailSupported",CC"()Z", (void*)&WB_NMTIsDetailSupported}, 1158 {CC"NMTChangeTrackingLevel", CC"()Z", (void*)&WB_NMTChangeTrackingLevel}, 1159 {CC"NMTGetHashSize", CC"()I", (void*)&WB_NMTGetHashSize }, 1160 #endif // INCLUDE_NMT 1161 {CC"deoptimizeAll", CC"()V", (void*)&WB_DeoptimizeAll }, 1162 {CC"deoptimizeMethod", CC"(Ljava/lang/reflect/Executable;Z)I", 1163 (void*)&WB_DeoptimizeMethod }, 1164 {CC"isMethodCompiled", CC"(Ljava/lang/reflect/Executable;Z)Z", 1165 (void*)&WB_IsMethodCompiled }, 1166 {CC"isMethodCompilable", CC"(Ljava/lang/reflect/Executable;IZ)Z", 1167 (void*)&WB_IsMethodCompilable}, 1168 {CC"isMethodQueuedForCompilation", 1169 CC"(Ljava/lang/reflect/Executable;)Z", (void*)&WB_IsMethodQueuedForCompilation}, 1170 {CC"makeMethodNotCompilable", 1171 CC"(Ljava/lang/reflect/Executable;IZ)V", (void*)&WB_MakeMethodNotCompilable}, 1172 {CC"testSetDontInlineMethod", 1173 CC"(Ljava/lang/reflect/Executable;Z)Z", (void*)&WB_TestSetDontInlineMethod}, 1174 {CC"getMethodCompilationLevel", 1175 CC"(Ljava/lang/reflect/Executable;Z)I", (void*)&WB_GetMethodCompilationLevel}, 1176 {CC"getMethodEntryBci", 1177 CC"(Ljava/lang/reflect/Executable;)I", (void*)&WB_GetMethodEntryBci}, 1178 {CC"getCompileQueueSize", 1179 CC"(I)I", (void*)&WB_GetCompileQueueSize}, 1180 {CC"testSetForceInlineMethod", 1181 CC"(Ljava/lang/reflect/Executable;Z)Z", (void*)&WB_TestSetForceInlineMethod}, 1182 {CC"enqueueMethodForCompilation", 1183 CC"(Ljava/lang/reflect/Executable;II)Z", (void*)&WB_EnqueueMethodForCompilation}, 1184 {CC"clearMethodState", 1185 CC"(Ljava/lang/reflect/Executable;)V", (void*)&WB_ClearMethodState}, 1186 {CC"lockCompilation", CC"()V", (void*)&WB_LockCompilation}, 1187 {CC"unlockCompilation", CC"()V", (void*)&WB_UnlockCompilation}, 1188 {CC"isConstantVMFlag", CC"(Ljava/lang/String;)Z", (void*)&WB_IsConstantVMFlag}, 1189 {CC"isLockedVMFlag", CC"(Ljava/lang/String;)Z", (void*)&WB_IsLockedVMFlag}, 1190 {CC"setBooleanVMFlag", CC"(Ljava/lang/String;Z)V",(void*)&WB_SetBooleanVMFlag}, 1191 {CC"setIntxVMFlag", CC"(Ljava/lang/String;J)V",(void*)&WB_SetIntxVMFlag}, 1192 {CC"setUintxVMFlag", CC"(Ljava/lang/String;J)V",(void*)&WB_SetUintxVMFlag}, 1193 {CC"setUint64VMFlag", CC"(Ljava/lang/String;J)V",(void*)&WB_SetUint64VMFlag}, 1194 {CC"setSizeTVMFlag", CC"(Ljava/lang/String;J)V",(void*)&WB_SetSizeTVMFlag}, 1195 {CC"setDoubleVMFlag", CC"(Ljava/lang/String;D)V",(void*)&WB_SetDoubleVMFlag}, 1196 {CC"setStringVMFlag", CC"(Ljava/lang/String;Ljava/lang/String;)V", 1197 (void*)&WB_SetStringVMFlag}, 1198 {CC"getBooleanVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Boolean;", 1199 (void*)&WB_GetBooleanVMFlag}, 1200 {CC"getIntxVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Long;", 1201 (void*)&WB_GetIntxVMFlag}, 1202 {CC"getUintxVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Long;", 1203 (void*)&WB_GetUintxVMFlag}, 1204 {CC"getUint64VMFlag", CC"(Ljava/lang/String;)Ljava/lang/Long;", 1205 (void*)&WB_GetUint64VMFlag}, 1206 {CC"getSizeTVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Long;", 1207 (void*)&WB_GetSizeTVMFlag}, 1208 {CC"getDoubleVMFlag", CC"(Ljava/lang/String;)Ljava/lang/Double;", 1209 (void*)&WB_GetDoubleVMFlag}, 1210 {CC"getStringVMFlag", CC"(Ljava/lang/String;)Ljava/lang/String;", 1211 (void*)&WB_GetStringVMFlag}, 1212 {CC"isInStringTable", CC"(Ljava/lang/String;)Z", (void*)&WB_IsInStringTable }, 1213 {CC"fullGC", CC"()V", (void*)&WB_FullGC }, 1214 {CC"youngGC", CC"()V", (void*)&WB_YoungGC }, 1215 {CC"readReservedMemory", CC"()V", (void*)&WB_ReadReservedMemory }, 1216 {CC"allocateMetaspace", 1217 CC"(Ljava/lang/ClassLoader;J)J", (void*)&WB_AllocateMetaspace }, 1218 {CC"freeMetaspace", 1219 CC"(Ljava/lang/ClassLoader;JJ)V", (void*)&WB_FreeMetaspace }, 1220 {CC"incMetaspaceCapacityUntilGC", CC"(J)J", (void*)&WB_IncMetaspaceCapacityUntilGC }, 1221 {CC"metaspaceCapacityUntilGC", CC"()J", (void*)&WB_MetaspaceCapacityUntilGC }, 1222 {CC"getCPUFeatures", CC"()Ljava/lang/String;", (void*)&WB_GetCPUFeatures }, 1223 {CC"getNMethod", CC"(Ljava/lang/reflect/Executable;Z)[Ljava/lang/Object;", 1224 (void*)&WB_GetNMethod }, 1225 {CC"forceNMethodSweep", CC"()V", (void*)&WB_ForceNMethodSweep }, 1226 {CC"allocateCodeBlob", CC"(II)J", (void*)&WB_AllocateCodeBlob }, 1227 {CC"freeCodeBlob", CC"(J)V", (void*)&WB_FreeCodeBlob }, 1228 {CC"getCodeHeapEntries", CC"(I)[Ljava/lang/Object;",(void*)&WB_GetCodeHeapEntries }, 1229 {CC"getThreadStackSize", CC"()J", (void*)&WB_GetThreadStackSize }, 1230 {CC"getThreadRemainingStackSize", CC"()J", (void*)&WB_GetThreadRemainingStackSize }, 1231 }; 1232 1233 #undef CC 1234 1235 JVM_ENTRY(void, JVM_RegisterWhiteBoxMethods(JNIEnv* env, jclass wbclass)) 1236 { 1237 if (WhiteBoxAPI) { 1238 // Make sure that wbclass is loaded by the null classloader 1239 instanceKlassHandle ikh = instanceKlassHandle(JNIHandles::resolve(wbclass)->klass()); 1240 Handle loader(ikh->class_loader()); 1241 if (loader.is_null()) { 1242 WhiteBox::register_methods(env, wbclass, thread, methods, sizeof(methods) / sizeof(methods[0])); 1243 WhiteBox::register_extended(env, wbclass, thread); 1244 WhiteBox::set_used(); 1245 } 1246 } 1247 } 1248 JVM_END