1 /* 2 * Copyright (c) 1997, 2013, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "code/codeCache.hpp" 27 #include "code/compiledIC.hpp" 28 #include "code/icBuffer.hpp" 29 #include "code/nmethod.hpp" 30 #include "compiler/compileBroker.hpp" 31 #include "memory/resourceArea.hpp" 32 #include "oops/method.hpp" 33 #include "runtime/atomic.hpp" 34 #include "runtime/compilationPolicy.hpp" 35 #include "runtime/mutexLocker.hpp" 36 #include "runtime/os.hpp" 37 #include "runtime/sweeper.hpp" 38 #include "runtime/vm_operations.hpp" 39 #include "trace/tracing.hpp" 40 #include "utilities/events.hpp" 41 #include "utilities/xmlstream.hpp" 42 43 #ifdef ASSERT 44 45 #define SWEEP(nm) record_sweep(nm, __LINE__) 46 // Sweeper logging code 47 class SweeperRecord { 48 public: 49 int traversal; 50 int invocation; 51 int compile_id; 52 long traversal_mark; 53 int state; 54 const char* kind; 55 address vep; 56 address uep; 57 int line; 58 59 void print() { 60 tty->print_cr("traversal = %d invocation = %d compile_id = %d %s uep = " PTR_FORMAT " vep = " 61 PTR_FORMAT " state = %d traversal_mark %d line = %d", 62 traversal, 63 invocation, 64 compile_id, 65 kind == NULL ? "" : kind, 66 uep, 67 vep, 68 state, 69 traversal_mark, 70 line); 71 } 72 }; 73 74 static int _sweep_index = 0; 75 static SweeperRecord* _records = NULL; 76 77 void NMethodSweeper::report_events(int id, address entry) { 78 if (_records != NULL) { 79 for (int i = _sweep_index; i < SweeperLogEntries; i++) { 80 if (_records[i].uep == entry || 81 _records[i].vep == entry || 82 _records[i].compile_id == id) { 83 _records[i].print(); 84 } 85 } 86 for (int i = 0; i < _sweep_index; i++) { 87 if (_records[i].uep == entry || 88 _records[i].vep == entry || 89 _records[i].compile_id == id) { 90 _records[i].print(); 91 } 92 } 93 } 94 } 95 96 void NMethodSweeper::report_events() { 97 if (_records != NULL) { 98 for (int i = _sweep_index; i < SweeperLogEntries; i++) { 99 // skip empty records 100 if (_records[i].vep == NULL) continue; 101 _records[i].print(); 102 } 103 for (int i = 0; i < _sweep_index; i++) { 104 // skip empty records 105 if (_records[i].vep == NULL) continue; 106 _records[i].print(); 107 } 108 } 109 } 110 111 void NMethodSweeper::record_sweep(nmethod* nm, int line) { 112 if (_records != NULL) { 113 _records[_sweep_index].traversal = _traversals; 114 _records[_sweep_index].traversal_mark = nm->_stack_traversal_mark; 115 _records[_sweep_index].invocation = _invocations; 116 _records[_sweep_index].compile_id = nm->compile_id(); 117 _records[_sweep_index].kind = nm->compile_kind(); 118 _records[_sweep_index].state = nm->_state; 119 _records[_sweep_index].vep = nm->verified_entry_point(); 120 _records[_sweep_index].uep = nm->entry_point(); 121 _records[_sweep_index].line = line; 122 123 _sweep_index = (_sweep_index + 1) % SweeperLogEntries; 124 } 125 } 126 #else 127 #define SWEEP(nm) 128 #endif 129 130 131 long NMethodSweeper::_traversals = 0; // No. of stack traversals performed 132 nmethod* NMethodSweeper::_current = NULL; // Current nmethod 133 int NMethodSweeper::_seen = 0 ; // No. of nmethods we have currently processed in current pass of CodeCache 134 int NMethodSweeper::_flushed_count = 0; // Nof. nmethods flushed in current sweep 135 size_t NMethodSweeper::_flushed_size = 0; // Total size of nmethods flushed in current sweep 136 int NMethodSweeper::_zombified_count = 0; // Nof. nmethods made zombie in current sweep 137 int NMethodSweeper::_marked_count = 0; // Nof. nmethods marked for reclaim in current sweep 138 139 volatile int NMethodSweeper::_invocations = 0; // No. of invocations left until we are completed with this pass 140 volatile int NMethodSweeper::_sweep_started = 0; // Whether a sweep is in progress. 141 142 jint NMethodSweeper::_locked_seen = 0; 143 jint NMethodSweeper::_not_entrant_seen_on_stack = 0; 144 bool NMethodSweeper::_resweep = false; 145 jint NMethodSweeper::_flush_token = 0; 146 jlong NMethodSweeper::_last_full_flush_time = 0; 147 int NMethodSweeper::_highest_marked = 0; 148 int NMethodSweeper::_dead_compile_ids = 0; 149 long NMethodSweeper::_last_flush_traversal_id = 0; 150 151 int NMethodSweeper::_number_of_flushes = 0; // Total of full traversals caused by full cache 152 int NMethodSweeper::_total_nof_methods_reclaimed = 0; 153 size_t NMethodSweeper::_total_flushed_size = 0; 154 jlong NMethodSweeper::_total_time_sweeping = 0; 155 jlong NMethodSweeper::_total_time_this_sweep = 0; 156 jlong NMethodSweeper::_peak_sweep_time = 0; 157 jlong NMethodSweeper::_peak_sweep_fraction_time = 0; 158 jlong NMethodSweeper::_total_disconnect_time = 0; 159 jlong NMethodSweeper::_peak_disconnect_time = 0; 160 161 class MarkActivationClosure: public CodeBlobClosure { 162 public: 163 virtual void do_code_blob(CodeBlob* cb) { 164 // If we see an activation belonging to a non_entrant nmethod, we mark it. 165 if (cb->is_nmethod() && ((nmethod*)cb)->is_not_entrant()) { 166 ((nmethod*)cb)->mark_as_seen_on_stack(); 167 } 168 } 169 }; 170 static MarkActivationClosure mark_activation_closure; 171 172 bool NMethodSweeper::sweep_in_progress() { 173 return (_current != NULL); 174 } 175 176 void NMethodSweeper::scan_stacks() { 177 assert(SafepointSynchronize::is_at_safepoint(), "must be executed at a safepoint"); 178 if (!MethodFlushing) return; 179 180 // No need to synchronize access, since this is always executed at a 181 // safepoint. 182 183 // Make sure CompiledIC_lock in unlocked, since we might update some 184 // inline caches. If it is, we just bail-out and try later. 185 if (CompiledIC_lock->is_locked() || Patching_lock->is_locked()) return; 186 187 // Check for restart 188 assert(CodeCache::find_blob_unsafe(_current) == _current, "Sweeper nmethod cached state invalid"); 189 if (!sweep_in_progress() && _resweep) { 190 _seen = 0; 191 _invocations = NmethodSweepFraction; 192 _current = CodeCache::first_nmethod(); 193 _traversals += 1; 194 _total_time_this_sweep = 0; 195 196 if (PrintMethodFlushing) { 197 tty->print_cr("### Sweep: stack traversal %d", _traversals); 198 } 199 Threads::nmethods_do(&mark_activation_closure); 200 201 // reset the flags since we started a scan from the beginning. 202 _resweep = false; 203 _locked_seen = 0; 204 _not_entrant_seen_on_stack = 0; 205 } 206 207 if (UseCodeCacheFlushing) { 208 // only allow new flushes after the interval is complete. 209 jlong now = os::javaTimeMillis(); 210 jlong max_interval = (jlong)MinCodeCacheFlushingInterval * (jlong)1000; 211 jlong curr_interval = now - _last_full_flush_time; 212 if (curr_interval > max_interval) { 213 _flush_token = 0; 214 } 215 216 if (!CodeCache::needs_flushing() && !CompileBroker::should_compile_new_jobs()) { 217 CompileBroker::set_should_compile_new_jobs(CompileBroker::run_compilation); 218 log_sweep("restart_compiler"); 219 } 220 } 221 } 222 223 void NMethodSweeper::possibly_sweep() { 224 assert(JavaThread::current()->thread_state() == _thread_in_vm, "must run in vm mode"); 225 if (!MethodFlushing || !sweep_in_progress()) return; 226 227 if (_invocations > 0) { 228 // Only one thread at a time will sweep 229 jint old = Atomic::cmpxchg( 1, &_sweep_started, 0 ); 230 if (old != 0) { 231 return; 232 } 233 #ifdef ASSERT 234 if (LogSweeper && _records == NULL) { 235 // Create the ring buffer for the logging code 236 _records = NEW_C_HEAP_ARRAY(SweeperRecord, SweeperLogEntries, mtGC); 237 memset(_records, 0, sizeof(SweeperRecord) * SweeperLogEntries); 238 } 239 #endif 240 if (_invocations > 0) { 241 sweep_code_cache(); 242 _invocations--; 243 } 244 _sweep_started = 0; 245 } 246 } 247 248 void NMethodSweeper::sweep_code_cache() { 249 250 jlong sweep_start_counter = os::elapsed_counter(); 251 252 _flushed_count = 0; 253 _flushed_size = 0; 254 _zombified_count = 0; 255 _marked_count = 0; 256 257 if (PrintMethodFlushing && Verbose) { 258 tty->print_cr("### Sweep at %d out of %d. Invocations left: %d", _seen, CodeCache::nof_nmethods(), _invocations); 259 } 260 261 if (!CompileBroker::should_compile_new_jobs()) { 262 // If we have turned off compilations we might as well do full sweeps 263 // in order to reach the clean state faster. Otherwise the sleeping compiler 264 // threads will slow down sweeping. After a few iterations the cache 265 // will be clean and sweeping stops (_resweep will not be set) 266 _invocations = 1; 267 } 268 269 // We want to visit all nmethods after NmethodSweepFraction 270 // invocations so divide the remaining number of nmethods by the 271 // remaining number of invocations. This is only an estimate since 272 // the number of nmethods changes during the sweep so the final 273 // stage must iterate until it there are no more nmethods. 274 int todo = (CodeCache::nof_nmethods() - _seen) / _invocations; 275 int swept_count = 0; 276 277 assert(!SafepointSynchronize::is_at_safepoint(), "should not be in safepoint when we get here"); 278 assert(!CodeCache_lock->owned_by_self(), "just checking"); 279 280 { 281 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 282 283 // The last invocation iterates until there are no more nmethods 284 for (int i = 0; (i < todo || _invocations == 1) && _current != NULL; i++) { 285 swept_count++; 286 if (SafepointSynchronize::is_synchronizing()) { // Safepoint request 287 if (PrintMethodFlushing && Verbose) { 288 tty->print_cr("### Sweep at %d out of %d, invocation: %d, yielding to safepoint", _seen, CodeCache::nof_nmethods(), _invocations); 289 } 290 MutexUnlockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 291 292 assert(Thread::current()->is_Java_thread(), "should be java thread"); 293 JavaThread* thread = (JavaThread*)Thread::current(); 294 ThreadBlockInVM tbivm(thread); 295 thread->java_suspend_self(); 296 } 297 // Since we will give up the CodeCache_lock, always skip ahead 298 // to the next nmethod. Other blobs can be deleted by other 299 // threads but nmethods are only reclaimed by the sweeper. 300 nmethod* next = CodeCache::next_nmethod(_current); 301 302 // Now ready to process nmethod and give up CodeCache_lock 303 { 304 MutexUnlockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 305 process_nmethod(_current); 306 } 307 _seen++; 308 _current = next; 309 } 310 } 311 312 assert(_invocations > 1 || _current == NULL, "must have scanned the whole cache"); 313 314 if (!sweep_in_progress() && !_resweep && (_locked_seen || _not_entrant_seen_on_stack)) { 315 // we've completed a scan without making progress but there were 316 // nmethods we were unable to process either because they were 317 // locked or were still on stack. We don't have to aggresively 318 // clean them up so just stop scanning. We could scan once more 319 // but that complicates the control logic and it's unlikely to 320 // matter much. 321 if (PrintMethodFlushing) { 322 tty->print_cr("### Couldn't make progress on some nmethods so stopping sweep"); 323 } 324 } 325 326 jlong sweep_end_counter = os::elapsed_counter(); 327 jlong sweep_time = sweep_end_counter - sweep_start_counter; 328 _total_time_sweeping += sweep_time; 329 _total_time_this_sweep += sweep_time; 330 _peak_sweep_fraction_time = MAX2(sweep_time, _peak_sweep_fraction_time); 331 _total_flushed_size += _flushed_size; 332 _total_nof_methods_reclaimed += _flushed_count; 333 334 EventSweepCodeCache event(UNTIMED); 335 if (event.should_commit()) { 336 event.set_starttime(sweep_start_counter); 337 event.set_endtime(sweep_end_counter); 338 event.set_sweepIndex(_traversals); 339 event.set_sweepFractionIndex(NmethodSweepFraction - _invocations + 1); 340 event.set_sweptCount(swept_count); 341 event.set_flushedCount(_flushed_count); 342 event.set_markedCount(_marked_count); 343 event.set_zombifiedCount(_zombified_count); 344 event.commit(); 345 } 346 347 #ifdef ASSERT 348 if(PrintMethodFlushing) { 349 tty->print_cr("### sweeper: sweep time(%d): " INT64_FORMAT, _invocations, (jlong)sweep_time); 350 } 351 #endif 352 353 if (_invocations == 1) { 354 _peak_sweep_time = MAX2(_peak_sweep_time, _total_time_this_sweep); 355 log_sweep("finished"); 356 } 357 358 // Sweeper is the only case where memory is released, 359 // check here if it is time to restart the compiler. 360 if (UseCodeCacheFlushing && !CompileBroker::should_compile_new_jobs() && !CodeCache::needs_flushing()) { 361 CompileBroker::set_should_compile_new_jobs(CompileBroker::run_compilation); 362 log_sweep("restart_compiler"); 363 } 364 } 365 366 class NMethodMarker: public StackObj { 367 private: 368 CompilerThread* _thread; 369 public: 370 NMethodMarker(nmethod* nm) { 371 _thread = CompilerThread::current(); 372 if (!nm->is_zombie() && !nm->is_unloaded()) { 373 // Only expose live nmethods for scanning 374 _thread->set_scanned_nmethod(nm); 375 } 376 } 377 ~NMethodMarker() { 378 _thread->set_scanned_nmethod(NULL); 379 } 380 }; 381 382 void NMethodSweeper::release_nmethod(nmethod *nm) { 383 // Clean up any CompiledICHolders 384 { 385 ResourceMark rm; 386 MutexLocker ml_patch(CompiledIC_lock); 387 RelocIterator iter(nm); 388 while (iter.next()) { 389 if (iter.type() == relocInfo::virtual_call_type) { 390 CompiledIC::cleanup_call_site(iter.virtual_call_reloc()); 391 } 392 } 393 } 394 395 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); 396 nm->flush(); 397 } 398 399 void NMethodSweeper::process_nmethod(nmethod *nm) { 400 assert(!CodeCache_lock->owned_by_self(), "just checking"); 401 402 // Make sure this nmethod doesn't get unloaded during the scan, 403 // since the locks acquired below might safepoint. 404 NMethodMarker nmm(nm); 405 406 SWEEP(nm); 407 408 // Skip methods that are currently referenced by the VM 409 if (nm->is_locked_by_vm()) { 410 // But still remember to clean-up inline caches for alive nmethods 411 if (nm->is_alive()) { 412 // Clean-up all inline caches that points to zombie/non-reentrant methods 413 MutexLocker cl(CompiledIC_lock); 414 nm->cleanup_inline_caches(); 415 SWEEP(nm); 416 } else { 417 _locked_seen++; 418 SWEEP(nm); 419 } 420 return; 421 } 422 423 if (nm->is_zombie()) { 424 // If it is first time, we see nmethod then we mark it. Otherwise, 425 // we reclame it. When we have seen a zombie method twice, we know that 426 // there are no inline caches that refer to it. 427 if (nm->is_marked_for_reclamation()) { 428 assert(!nm->is_locked_by_vm(), "must not flush locked nmethods"); 429 if (PrintMethodFlushing && Verbose) { 430 tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (marked for reclamation) being flushed", nm->compile_id(), nm); 431 } 432 _flushed_size += nm->size(); 433 release_nmethod(nm); 434 _flushed_count++; 435 } else { 436 if (PrintMethodFlushing && Verbose) { 437 tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (zombie) being marked for reclamation", nm->compile_id(), nm); 438 } 439 nm->mark_for_reclamation(); 440 _resweep = true; 441 _marked_count++; 442 SWEEP(nm); 443 } 444 } else if (nm->is_not_entrant()) { 445 // If there is no current activations of this method on the 446 // stack we can safely convert it to a zombie method 447 if (nm->can_not_entrant_be_converted()) { 448 if (PrintMethodFlushing && Verbose) { 449 tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (not entrant) being made zombie", nm->compile_id(), nm); 450 } 451 nm->make_zombie(); 452 _resweep = true; 453 _zombified_count++; 454 SWEEP(nm); 455 } else { 456 // Still alive, clean up its inline caches 457 MutexLocker cl(CompiledIC_lock); 458 nm->cleanup_inline_caches(); 459 // we coudn't transition this nmethod so don't immediately 460 // request a rescan. If this method stays on the stack for a 461 // long time we don't want to keep rescanning the code cache. 462 _not_entrant_seen_on_stack++; 463 SWEEP(nm); 464 } 465 } else if (nm->is_unloaded()) { 466 // Unloaded code, just make it a zombie 467 if (PrintMethodFlushing && Verbose) 468 tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (unloaded) being made zombie", nm->compile_id(), nm); 469 470 if (nm->is_osr_method()) { 471 SWEEP(nm); 472 // No inline caches will ever point to osr methods, so we can just remove it 473 _flushed_size += nm->size(); 474 release_nmethod(nm); 475 _flushed_count++; 476 } else { 477 nm->make_zombie(); 478 _resweep = true; 479 _zombified_count++; 480 SWEEP(nm); 481 } 482 } else { 483 assert(nm->is_alive(), "should be alive"); 484 485 if (UseCodeCacheFlushing) { 486 if (nm->is_speculatively_disconnected() && !nm->is_locked_by_vm() && !nm->is_osr_method() && 487 (_traversals > _last_flush_traversal_id + 2) && (nm->compile_id() < _highest_marked)) { 488 // This method has not been called since the forced cleanup happened 489 nm->make_not_entrant(); 490 } 491 } 492 493 // Clean-up all inline caches that points to zombie/non-reentrant methods 494 MutexLocker cl(CompiledIC_lock); 495 nm->cleanup_inline_caches(); 496 SWEEP(nm); 497 } 498 } 499 500 // Code cache unloading: when compilers notice the code cache is getting full, 501 // they will call a vm op that comes here. This code attempts to speculatively 502 // unload the oldest half of the nmethods (based on the compile job id) by 503 // saving the old code in a list in the CodeCache. Then 504 // execution resumes. If a method so marked is not called by the second sweeper 505 // stack traversal after the current one, the nmethod will be marked non-entrant and 506 // got rid of by normal sweeping. If the method is called, the Method*'s 507 // _code field is restored and the Method*/nmethod 508 // go back to their normal state. 509 void NMethodSweeper::handle_full_code_cache(bool is_full) { 510 511 if (is_full) { 512 // Since code cache is full, immediately stop new compiles 513 if (CompileBroker::set_should_compile_new_jobs(CompileBroker::stop_compilation)) { 514 log_sweep("disable_compiler"); 515 } 516 } 517 518 // Make sure only one thread can flush 519 // The token is reset after CodeCacheMinimumFlushInterval in scan stacks, 520 // no need to check the timeout here. 521 jint old = Atomic::cmpxchg( 1, &_flush_token, 0 ); 522 if (old != 0) { 523 return; 524 } 525 526 VM_HandleFullCodeCache op(is_full); 527 VMThread::execute(&op); 528 529 // resweep again as soon as possible 530 _resweep = true; 531 } 532 533 void NMethodSweeper::speculative_disconnect_nmethods(bool is_full) { 534 // If there was a race in detecting full code cache, only run 535 // one vm op for it or keep the compiler shut off 536 537 jlong disconnect_start_counter = os::elapsed_counter(); 538 539 // Traverse the code cache trying to dump the oldest nmethods 540 int curr_max_comp_id = CompileBroker::get_compilation_id(); 541 int flush_target = ((curr_max_comp_id - _dead_compile_ids) / CodeCacheFlushingFraction) + _dead_compile_ids; 542 543 log_sweep("start_cleaning"); 544 545 nmethod* nm = CodeCache::alive_nmethod(CodeCache::first()); 546 jint disconnected = 0; 547 jint made_not_entrant = 0; 548 jint nmethod_count = 0; 549 550 while ((nm != NULL)){ 551 int curr_comp_id = nm->compile_id(); 552 553 // OSR methods cannot be flushed like this. Also, don't flush native methods 554 // since they are part of the JDK in most cases 555 if (!nm->is_osr_method() && !nm->is_locked_by_vm() && !nm->is_native_method()) { 556 557 // only count methods that can be speculatively disconnected 558 nmethod_count++; 559 560 if (nm->is_in_use() && (curr_comp_id < flush_target)) { 561 if ((nm->method()->code() == nm)) { 562 // This method has not been previously considered for 563 // unloading or it was restored already 564 CodeCache::speculatively_disconnect(nm); 565 disconnected++; 566 } else if (nm->is_speculatively_disconnected()) { 567 // This method was previously considered for preemptive unloading and was not called since then 568 CompilationPolicy::policy()->delay_compilation(nm->method()); 569 nm->make_not_entrant(); 570 made_not_entrant++; 571 } 572 573 if (curr_comp_id > _highest_marked) { 574 _highest_marked = curr_comp_id; 575 } 576 } 577 } 578 nm = CodeCache::alive_nmethod(CodeCache::next(nm)); 579 } 580 581 // remember how many compile_ids wheren't seen last flush. 582 _dead_compile_ids = curr_max_comp_id - nmethod_count; 583 584 log_sweep("stop_cleaning", 585 "disconnected='" UINT32_FORMAT "' made_not_entrant='" UINT32_FORMAT "'", 586 disconnected, made_not_entrant); 587 588 // Shut off compiler. Sweeper will start over with a new stack scan and 589 // traversal cycle and turn it back on if it clears enough space. 590 if (is_full) { 591 _last_full_flush_time = os::javaTimeMillis(); 592 } 593 594 jlong disconnect_end_counter = os::elapsed_counter(); 595 jlong disconnect_time = disconnect_end_counter - disconnect_start_counter; 596 _total_disconnect_time += disconnect_time; 597 _peak_disconnect_time = MAX2(disconnect_time, _peak_disconnect_time); 598 599 EventCleanCodeCache event(UNTIMED); 600 if (event.should_commit()) { 601 event.set_starttime(disconnect_start_counter); 602 event.set_endtime(disconnect_end_counter); 603 event.set_disconnectedCount(disconnected); 604 event.set_madeNonEntrantCount(made_not_entrant); 605 event.commit(); 606 } 607 _number_of_flushes++; 608 609 // After two more traversals the sweeper will get rid of unrestored nmethods 610 _last_flush_traversal_id = _traversals; 611 _resweep = true; 612 #ifdef ASSERT 613 614 if(PrintMethodFlushing && Verbose) { 615 tty->print_cr("### sweeper: unload time: " INT64_FORMAT, (jlong)disconnect_time); 616 } 617 #endif 618 } 619 620 621 // Print out some state information about the current sweep and the 622 // state of the code cache if it's requested. 623 void NMethodSweeper::log_sweep(const char* msg, const char* format, ...) { 624 if (PrintMethodFlushing) { 625 stringStream s; 626 // Dump code cache state into a buffer before locking the tty, 627 // because log_state() will use locks causing lock conflicts. 628 CodeCache::log_state(&s); 629 630 ttyLocker ttyl; 631 tty->print("### sweeper: %s ", msg); 632 if (format != NULL) { 633 va_list ap; 634 va_start(ap, format); 635 tty->vprint(format, ap); 636 va_end(ap); 637 } 638 tty->print_cr(s.as_string()); 639 } 640 641 if (LogCompilation && (xtty != NULL)) { 642 stringStream s; 643 // Dump code cache state into a buffer before locking the tty, 644 // because log_state() will use locks causing lock conflicts. 645 CodeCache::log_state(&s); 646 647 ttyLocker ttyl; 648 xtty->begin_elem("sweeper state='%s' traversals='" INTX_FORMAT "' ", msg, (intx)traversal_count()); 649 if (format != NULL) { 650 va_list ap; 651 va_start(ap, format); 652 xtty->vprint(format, ap); 653 va_end(ap); 654 } 655 xtty->print(s.as_string()); 656 xtty->stamp(); 657 xtty->end_elem(); 658 } 659 } 660 661 void NMethodSweeper::print() { 662 ttyLocker ttyl; 663 tty->print_cr("Method flushing statistics:"); 664 tty->print_cr(" Total sweep time: %1.3lfms", (double)_total_time_sweeping/1000); 665 tty->print_cr(" Peak sweep time: %1.3lfms", (double)_peak_sweep_time/1000); 666 tty->print_cr(" Peak sweep fraction time: %1.3lfms", (double)_peak_sweep_fraction_time/1000); 667 tty->print_cr(" Total number of stack traversals: %d", _traversals); 668 tty->print_cr(" Total number of flushed methods: %d", _total_nof_methods_reclaimed); 669 tty->print_cr(" Total size of flushed methods: " SIZE_FORMAT "Kb", _total_flushed_size/K); 670 }