1 /* 2 * Copyright (c) 1997, 2018, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #ifndef SHARE_VM_RUNTIME_VMOPERATIONS_HPP 26 #define SHARE_VM_RUNTIME_VMOPERATIONS_HPP 27 28 #include "classfile/javaClasses.hpp" 29 #include "memory/allocation.hpp" 30 #include "oops/oop.hpp" 31 #include "runtime/thread.hpp" 32 #include "runtime/threadSMR.hpp" 33 #include "code/codeCache.hpp" 34 35 // The following classes are used for operations 36 // initiated by a Java thread but that must 37 // take place in the VMThread. 38 39 #define VM_OP_ENUM(type) VMOp_##type, 40 41 // Note: When new VM_XXX comes up, add 'XXX' to the template table. 42 #define VM_OPS_DO(template) \ 43 template(Dummy) \ 44 template(ThreadStop) \ 45 template(ThreadDump) \ 46 template(PrintThreads) \ 47 template(FindDeadlocks) \ 48 template(ClearICs) \ 49 template(ForceSafepoint) \ 50 template(ForceAsyncSafepoint) \ 51 template(Deoptimize) \ 52 template(DeoptimizeFrame) \ 53 template(DeoptimizeAll) \ 54 template(ZombieAll) \ 55 template(UnlinkSymbols) \ 56 template(Verify) \ 57 template(PrintJNI) \ 58 template(HeapDumper) \ 59 template(DeoptimizeTheWorld) \ 60 template(CollectForMetadataAllocation) \ 61 template(GC_HeapInspection) \ 62 template(GenCollectFull) \ 63 template(GenCollectFullConcurrent) \ 64 template(GenCollectForAllocation) \ 65 template(ParallelGCFailedAllocation) \ 66 template(ParallelGCSystemGC) \ 67 template(CGC_Operation) \ 68 template(CMS_Initial_Mark) \ 69 template(CMS_Final_Remark) \ 70 template(G1CollectForAllocation) \ 71 template(G1CollectFull) \ 72 template(ZOperation) \ 73 template(HandshakeOneThread) \ 74 template(HandshakeAllThreads) \ 75 template(HandshakeFallback) \ 76 template(EnableBiasedLocking) \ 77 template(RevokeBias) \ 78 template(BulkRevokeBias) \ 79 template(PopulateDumpSharedSpace) \ 80 template(JNIFunctionTableCopier) \ 81 template(RedefineClasses) \ 82 template(UpdateForPopTopFrame) \ 83 template(SetFramePop) \ 84 template(GetOwnedMonitorInfo) \ 85 template(GetObjectMonitorUsage) \ 86 template(GetCurrentContendedMonitor) \ 87 template(GetStackTrace) \ 88 template(GetMultipleStackTraces) \ 89 template(GetAllStackTraces) \ 90 template(GetThreadListStackTraces) \ 91 template(GetFrameCount) \ 92 template(GetFrameLocation) \ 93 template(ChangeBreakpoints) \ 94 template(GetOrSetLocal) \ 95 template(GetCurrentLocation) \ 96 template(EnterInterpOnlyMode) \ 97 template(ChangeSingleStep) \ 98 template(HeapWalkOperation) \ 99 template(HeapIterateOperation) \ 100 template(ReportJavaOutOfMemory) \ 101 template(JFRCheckpoint) \ 102 template(ShenandoahFullGC) \ 103 template(ShenandoahInitMark) \ 104 template(ShenandoahFinalMarkStartEvac) \ 105 template(ShenandoahInitUpdateRefs) \ 106 template(ShenandoahFinalUpdateRefs) \ 107 template(ShenandoahDegeneratedGC) \ 108 template(Exit) \ 109 template(LinuxDllLoad) \ 110 template(RotateGCLog) \ 111 template(WhiteBoxOperation) \ 112 template(ClassLoaderStatsOperation) \ 113 template(ClassLoaderHierarchyOperation) \ 114 template(DumpHashtable) \ 115 template(DumpTouchedMethods) \ 116 template(MarkActiveNMethods) \ 117 template(PrintCompileQueue) \ 118 template(PrintClassHierarchy) \ 119 template(ThreadSuspend) \ 120 template(CTWThreshold) \ 121 template(ThreadsSuspendJVMTI) \ 122 template(ICBufferFull) \ 123 template(ScavengeMonitors) \ 124 template(PrintMetadata) \ 125 template(GTestExecuteAtSafepoint) \ 126 template(JFROldObject) \ 127 128 class VM_Operation: public CHeapObj<mtInternal> { 129 public: 130 enum Mode { 131 _safepoint, // blocking, safepoint, vm_op C-heap allocated 132 _no_safepoint, // blocking, no safepoint, vm_op C-Heap allocated 133 _concurrent, // non-blocking, no safepoint, vm_op C-Heap allocated 134 _async_safepoint // non-blocking, safepoint, vm_op C-Heap allocated 135 }; 136 137 enum VMOp_Type { 138 VM_OPS_DO(VM_OP_ENUM) 139 VMOp_Terminating 140 }; 141 142 private: 143 Thread* _calling_thread; 144 ThreadPriority _priority; 145 long _timestamp; 146 VM_Operation* _next; 147 VM_Operation* _prev; 148 149 // The VM operation name array 150 static const char* _names[]; 151 152 public: 153 VM_Operation() { _calling_thread = NULL; _next = NULL; _prev = NULL; } 154 virtual ~VM_Operation() {} 155 156 // VM operation support (used by VM thread) 157 Thread* calling_thread() const { return _calling_thread; } 158 ThreadPriority priority() { return _priority; } 159 void set_calling_thread(Thread* thread, ThreadPriority priority); 160 161 long timestamp() const { return _timestamp; } 162 void set_timestamp(long timestamp) { _timestamp = timestamp; } 163 164 // Called by VM thread - does in turn invoke doit(). Do not override this 165 void evaluate(); 166 167 // evaluate() is called by the VMThread and in turn calls doit(). 168 // If the thread invoking VMThread::execute((VM_Operation*) is a JavaThread, 169 // doit_prologue() is called in that thread before transferring control to 170 // the VMThread. 171 // If doit_prologue() returns true the VM operation will proceed, and 172 // doit_epilogue() will be called by the JavaThread once the VM operation 173 // completes. If doit_prologue() returns false the VM operation is cancelled. 174 virtual void doit() = 0; 175 virtual bool doit_prologue() { return true; }; 176 virtual void doit_epilogue() {}; // Note: Not called if mode is: _concurrent 177 178 // Type test 179 virtual bool is_methodCompiler() const { return false; } 180 181 // Linking 182 VM_Operation *next() const { return _next; } 183 VM_Operation *prev() const { return _prev; } 184 void set_next(VM_Operation *next) { _next = next; } 185 void set_prev(VM_Operation *prev) { _prev = prev; } 186 187 // Configuration. Override these appropriately in subclasses. 188 virtual VMOp_Type type() const = 0; 189 virtual Mode evaluation_mode() const { return _safepoint; } 190 virtual bool allow_nested_vm_operations() const { return false; } 191 virtual bool is_cheap_allocated() const { return false; } 192 virtual void oops_do(OopClosure* f) { /* do nothing */ }; 193 194 // CAUTION: <don't hang yourself with following rope> 195 // If you override these methods, make sure that the evaluation 196 // of these methods is race-free and non-blocking, since these 197 // methods may be evaluated either by the mutators or by the 198 // vm thread, either concurrently with mutators or with the mutators 199 // stopped. In other words, taking locks is verboten, and if there 200 // are any races in evaluating the conditions, they'd better be benign. 201 virtual bool evaluate_at_safepoint() const { 202 return evaluation_mode() == _safepoint || 203 evaluation_mode() == _async_safepoint; 204 } 205 virtual bool evaluate_concurrently() const { 206 return evaluation_mode() == _concurrent || 207 evaluation_mode() == _async_safepoint; 208 } 209 210 static const char* mode_to_string(Mode mode); 211 212 // Debugging 213 virtual void print_on_error(outputStream* st) const; 214 const char* name() const { return _names[type()]; } 215 static const char* name(int type) { 216 assert(type >= 0 && type < VMOp_Terminating, "invalid VM operation type"); 217 return _names[type]; 218 } 219 #ifndef PRODUCT 220 void print_on(outputStream* st) const { print_on_error(st); } 221 #endif 222 }; 223 224 class VM_ThreadStop: public VM_Operation { 225 private: 226 oop _thread; // The Thread that the Throwable is thrown against 227 oop _throwable; // The Throwable thrown at the target Thread 228 public: 229 // All oops are passed as JNI handles, since there is no guarantee that a GC might happen before the 230 // VM operation is executed. 231 VM_ThreadStop(oop thread, oop throwable) { 232 _thread = thread; 233 _throwable = throwable; 234 } 235 VMOp_Type type() const { return VMOp_ThreadStop; } 236 oop target_thread() const { return _thread; } 237 oop throwable() const { return _throwable;} 238 void doit(); 239 // We deoptimize if top-most frame is compiled - this might require a C2I adapter to be generated 240 bool allow_nested_vm_operations() const { return true; } 241 Mode evaluation_mode() const { return _async_safepoint; } 242 bool is_cheap_allocated() const { return true; } 243 244 // GC support 245 void oops_do(OopClosure* f) { 246 f->do_oop(&_thread); f->do_oop(&_throwable); 247 } 248 }; 249 250 class VM_ClearICs: public VM_Operation { 251 private: 252 bool _preserve_static_stubs; 253 public: 254 VM_ClearICs(bool preserve_static_stubs) { _preserve_static_stubs = preserve_static_stubs; } 255 void doit(); 256 VMOp_Type type() const { return VMOp_ClearICs; } 257 }; 258 259 // empty vm op, evaluated just to force a safepoint 260 class VM_ForceSafepoint: public VM_Operation { 261 public: 262 void doit() {} 263 VMOp_Type type() const { return VMOp_ForceSafepoint; } 264 }; 265 266 // empty vm op, when forcing a safepoint to suspend a thread 267 class VM_ThreadSuspend: public VM_ForceSafepoint { 268 public: 269 VMOp_Type type() const { return VMOp_ThreadSuspend; } 270 }; 271 272 // empty vm op, when forcing a safepoint due to ctw threshold is reached for the sweeper 273 class VM_CTWThreshold: public VM_ForceSafepoint { 274 public: 275 VMOp_Type type() const { return VMOp_CTWThreshold; } 276 }; 277 278 // empty vm op, when forcing a safepoint to suspend threads from jvmti 279 class VM_ThreadsSuspendJVMTI: public VM_ForceSafepoint { 280 public: 281 VMOp_Type type() const { return VMOp_ThreadsSuspendJVMTI; } 282 }; 283 284 // empty vm op, when forcing a safepoint due to inline cache buffers being full 285 class VM_ICBufferFull: public VM_ForceSafepoint { 286 public: 287 VMOp_Type type() const { return VMOp_ICBufferFull; } 288 }; 289 290 // empty asynchronous vm op, when forcing a safepoint to scavenge monitors 291 class VM_ScavengeMonitors: public VM_ForceSafepoint { 292 public: 293 VMOp_Type type() const { return VMOp_ScavengeMonitors; } 294 Mode evaluation_mode() const { return _async_safepoint; } 295 bool is_cheap_allocated() const { return true; } 296 }; 297 298 // Base class for invoking parts of a gtest in a safepoint. 299 // Derived classes provide the doit method. 300 // Typically also need to transition the gtest thread from native to VM. 301 class VM_GTestExecuteAtSafepoint: public VM_Operation { 302 public: 303 VMOp_Type type() const { return VMOp_GTestExecuteAtSafepoint; } 304 305 protected: 306 VM_GTestExecuteAtSafepoint() {} 307 }; 308 309 class VM_Deoptimize: public VM_Operation { 310 public: 311 VM_Deoptimize() {} 312 VMOp_Type type() const { return VMOp_Deoptimize; } 313 void doit(); 314 bool allow_nested_vm_operations() const { return true; } 315 }; 316 317 class VM_MarkActiveNMethods: public VM_Operation { 318 public: 319 VM_MarkActiveNMethods() {} 320 VMOp_Type type() const { return VMOp_MarkActiveNMethods; } 321 void doit(); 322 bool allow_nested_vm_operations() const { return true; } 323 }; 324 325 // Deopt helper that can deoptimize frames in threads other than the 326 // current thread. Only used through Deoptimization::deoptimize_frame. 327 class VM_DeoptimizeFrame: public VM_Operation { 328 friend class Deoptimization; 329 330 private: 331 JavaThread* _thread; 332 intptr_t* _id; 333 int _reason; 334 VM_DeoptimizeFrame(JavaThread* thread, intptr_t* id, int reason); 335 336 public: 337 VMOp_Type type() const { return VMOp_DeoptimizeFrame; } 338 void doit(); 339 bool allow_nested_vm_operations() const { return true; } 340 }; 341 342 #ifndef PRODUCT 343 class VM_DeoptimizeAll: public VM_Operation { 344 private: 345 Klass* _dependee; 346 public: 347 VM_DeoptimizeAll() {} 348 VMOp_Type type() const { return VMOp_DeoptimizeAll; } 349 void doit(); 350 bool allow_nested_vm_operations() const { return true; } 351 }; 352 353 354 class VM_ZombieAll: public VM_Operation { 355 public: 356 VM_ZombieAll() {} 357 VMOp_Type type() const { return VMOp_ZombieAll; } 358 void doit(); 359 bool allow_nested_vm_operations() const { return true; } 360 }; 361 #endif // PRODUCT 362 363 class VM_UnlinkSymbols: public VM_Operation { 364 public: 365 VM_UnlinkSymbols() {} 366 VMOp_Type type() const { return VMOp_UnlinkSymbols; } 367 void doit(); 368 bool allow_nested_vm_operations() const { return true; } 369 }; 370 371 class VM_Verify: public VM_Operation { 372 public: 373 VMOp_Type type() const { return VMOp_Verify; } 374 void doit(); 375 }; 376 377 378 class VM_PrintThreads: public VM_Operation { 379 private: 380 outputStream* _out; 381 bool _print_concurrent_locks; 382 bool _print_extended_info; 383 public: 384 VM_PrintThreads() 385 : _out(tty), _print_concurrent_locks(PrintConcurrentLocks), _print_extended_info(false) 386 {} 387 VM_PrintThreads(outputStream* out, bool print_concurrent_locks, bool print_extended_info) 388 : _out(out), _print_concurrent_locks(print_concurrent_locks), _print_extended_info(print_extended_info) 389 {} 390 VMOp_Type type() const { 391 return VMOp_PrintThreads; 392 } 393 void doit(); 394 bool doit_prologue(); 395 void doit_epilogue(); 396 }; 397 398 class VM_PrintJNI: public VM_Operation { 399 private: 400 outputStream* _out; 401 public: 402 VM_PrintJNI() { _out = tty; } 403 VM_PrintJNI(outputStream* out) { _out = out; } 404 VMOp_Type type() const { return VMOp_PrintJNI; } 405 void doit(); 406 }; 407 408 class VM_PrintMetadata : public VM_Operation { 409 private: 410 outputStream* const _out; 411 const size_t _scale; 412 const int _flags; 413 414 public: 415 VM_PrintMetadata(outputStream* out, size_t scale, int flags) 416 : _out(out), _scale(scale), _flags(flags) 417 {}; 418 419 VMOp_Type type() const { return VMOp_PrintMetadata; } 420 void doit(); 421 }; 422 423 class DeadlockCycle; 424 class VM_FindDeadlocks: public VM_Operation { 425 private: 426 bool _concurrent_locks; 427 DeadlockCycle* _deadlocks; 428 outputStream* _out; 429 ThreadsListSetter _setter; // Helper to set hazard ptr in the originating thread 430 // which protects the JavaThreads in _deadlocks. 431 432 public: 433 VM_FindDeadlocks(bool concurrent_locks) : _concurrent_locks(concurrent_locks), _out(NULL), _deadlocks(NULL), _setter() {}; 434 VM_FindDeadlocks(outputStream* st) : _concurrent_locks(true), _out(st), _deadlocks(NULL) {}; 435 ~VM_FindDeadlocks(); 436 437 DeadlockCycle* result() { return _deadlocks; }; 438 VMOp_Type type() const { return VMOp_FindDeadlocks; } 439 void doit(); 440 }; 441 442 class ThreadDumpResult; 443 class ThreadSnapshot; 444 class ThreadConcurrentLocks; 445 446 class VM_ThreadDump : public VM_Operation { 447 private: 448 ThreadDumpResult* _result; 449 int _num_threads; 450 GrowableArray<instanceHandle>* _threads; 451 int _max_depth; 452 bool _with_locked_monitors; 453 bool _with_locked_synchronizers; 454 455 ThreadSnapshot* snapshot_thread(JavaThread* java_thread, ThreadConcurrentLocks* tcl); 456 457 public: 458 VM_ThreadDump(ThreadDumpResult* result, 459 int max_depth, // -1 indicates entire stack 460 bool with_locked_monitors, 461 bool with_locked_synchronizers); 462 463 VM_ThreadDump(ThreadDumpResult* result, 464 GrowableArray<instanceHandle>* threads, 465 int num_threads, // -1 indicates entire stack 466 int max_depth, 467 bool with_locked_monitors, 468 bool with_locked_synchronizers); 469 470 VMOp_Type type() const { return VMOp_ThreadDump; } 471 void doit(); 472 bool doit_prologue(); 473 void doit_epilogue(); 474 }; 475 476 477 class VM_Exit: public VM_Operation { 478 private: 479 int _exit_code; 480 static volatile bool _vm_exited; 481 static Thread * volatile _shutdown_thread; 482 static void wait_if_vm_exited(); 483 public: 484 VM_Exit(int exit_code) { 485 _exit_code = exit_code; 486 } 487 static int wait_for_threads_in_native_to_block(); 488 static int set_vm_exited(); 489 static bool vm_exited() { return _vm_exited; } 490 static Thread * shutdown_thread() { return _shutdown_thread; } 491 static void block_if_vm_exited() { 492 if (_vm_exited) { 493 wait_if_vm_exited(); 494 } 495 } 496 VMOp_Type type() const { return VMOp_Exit; } 497 void doit(); 498 }; 499 500 class VM_PrintCompileQueue: public VM_Operation { 501 private: 502 outputStream* _out; 503 504 public: 505 VM_PrintCompileQueue(outputStream* st) : _out(st) {} 506 VMOp_Type type() const { return VMOp_PrintCompileQueue; } 507 Mode evaluation_mode() const { return _safepoint; } 508 void doit(); 509 }; 510 511 #if INCLUDE_SERVICES 512 class VM_PrintClassHierarchy: public VM_Operation { 513 private: 514 outputStream* _out; 515 bool _print_interfaces; 516 bool _print_subclasses; 517 char* _classname; 518 519 public: 520 VM_PrintClassHierarchy(outputStream* st, bool print_interfaces, bool print_subclasses, char* classname) : 521 _out(st), _print_interfaces(print_interfaces), _print_subclasses(print_subclasses), 522 _classname(classname) {} 523 VMOp_Type type() const { return VMOp_PrintClassHierarchy; } 524 void doit(); 525 }; 526 #endif // INCLUDE_SERVICES 527 528 #endif // SHARE_VM_RUNTIME_VMOPERATIONS_HPP