1 /* 2 * Copyright (c) 2003, 2013, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "classfile/systemDictionary.hpp" 27 #include "classfile/vmSymbols.hpp" 28 #include "gc_implementation/shared/mutableSpace.hpp" 29 #include "memory/collectorPolicy.hpp" 30 #include "memory/defNewGeneration.hpp" 31 #include "memory/genCollectedHeap.hpp" 32 #include "memory/generation.hpp" 33 #include "memory/generationSpec.hpp" 34 #include "memory/heap.hpp" 35 #include "memory/memRegion.hpp" 36 #include "memory/tenuredGeneration.hpp" 37 #include "oops/oop.inline.hpp" 38 #include "runtime/globals.hpp" 39 #include "runtime/javaCalls.hpp" 40 #include "services/classLoadingService.hpp" 41 #include "services/lowMemoryDetector.hpp" 42 #include "services/management.hpp" 43 #include "services/memoryManager.hpp" 44 #include "services/memoryPool.hpp" 45 #include "services/memoryService.hpp" 46 #include "utilities/growableArray.hpp" 47 #include "utilities/macros.hpp" 48 #if INCLUDE_ALL_GCS 49 #include "gc_implementation/concurrentMarkSweep/concurrentMarkSweepGeneration.hpp" 50 #include "gc_implementation/g1/g1CollectedHeap.inline.hpp" 51 #include "gc_implementation/parNew/parNewGeneration.hpp" 52 #include "gc_implementation/parallelScavenge/parallelScavengeHeap.hpp" 53 #include "gc_implementation/parallelScavenge/psOldGen.hpp" 54 #include "gc_implementation/parallelScavenge/psYoungGen.hpp" 55 #include "services/g1MemoryPool.hpp" 56 #include "services/psMemoryPool.hpp" 57 #endif // INCLUDE_ALL_GCS 58 59 GrowableArray<MemoryPool*>* MemoryService::_pools_list = 60 new (ResourceObj::C_HEAP, mtInternal) GrowableArray<MemoryPool*>(init_pools_list_size, true); 61 GrowableArray<MemoryManager*>* MemoryService::_managers_list = 62 new (ResourceObj::C_HEAP, mtInternal) GrowableArray<MemoryManager*>(init_managers_list_size, true); 63 64 GCMemoryManager* MemoryService::_minor_gc_manager = NULL; 65 GCMemoryManager* MemoryService::_major_gc_manager = NULL; 66 MemoryPool* MemoryService::_code_heap_pool = NULL; 67 MemoryPool* MemoryService::_metaspace_pool = NULL; 68 MemoryPool* MemoryService::_compressed_class_pool = NULL; 69 70 class GcThreadCountClosure: public ThreadClosure { 71 private: 72 int _count; 73 public: 74 GcThreadCountClosure() : _count(0) {}; 75 void do_thread(Thread* thread); 76 int count() { return _count; } 77 }; 78 79 void GcThreadCountClosure::do_thread(Thread* thread) { 80 _count++; 81 } 82 83 void MemoryService::set_universe_heap(CollectedHeap* heap) { 84 CollectedHeap::Name kind = heap->kind(); 85 switch (kind) { 86 case CollectedHeap::GenCollectedHeap : { 87 add_gen_collected_heap_info(GenCollectedHeap::heap()); 88 break; 89 } 90 #if INCLUDE_ALL_GCS 91 case CollectedHeap::ParallelScavengeHeap : { 92 add_parallel_scavenge_heap_info(ParallelScavengeHeap::heap()); 93 break; 94 } 95 case CollectedHeap::G1CollectedHeap : { 96 add_g1_heap_info(G1CollectedHeap::heap()); 97 break; 98 } 99 #endif // INCLUDE_ALL_GCS 100 default: { 101 guarantee(false, "Unrecognized kind of heap"); 102 } 103 } 104 105 // set the GC thread count 106 GcThreadCountClosure gctcc; 107 heap->gc_threads_do(&gctcc); 108 int count = gctcc.count(); 109 if (count > 0) { 110 _minor_gc_manager->set_num_gc_threads(count); 111 _major_gc_manager->set_num_gc_threads(count); 112 } 113 114 // All memory pools and memory managers are initialized. 115 // 116 _minor_gc_manager->initialize_gc_stat_info(); 117 _major_gc_manager->initialize_gc_stat_info(); 118 } 119 120 // Add memory pools for GenCollectedHeap 121 // This function currently only supports two generations collected heap. 122 // The collector for GenCollectedHeap will have two memory managers. 123 void MemoryService::add_gen_collected_heap_info(GenCollectedHeap* heap) { 124 CollectorPolicy* policy = heap->collector_policy(); 125 126 assert(policy->is_two_generation_policy(), "Only support two generations"); 127 guarantee(heap->n_gens() == 2, "Only support two-generation heap"); 128 129 TwoGenerationCollectorPolicy* two_gen_policy = policy->as_two_generation_policy(); 130 if (two_gen_policy != NULL) { 131 GenerationSpec** specs = two_gen_policy->generations(); 132 Generation::Name kind = specs[0]->name(); 133 switch (kind) { 134 case Generation::DefNew: 135 _minor_gc_manager = MemoryManager::get_copy_memory_manager(); 136 break; 137 #if INCLUDE_ALL_GCS 138 case Generation::ParNew: 139 case Generation::ASParNew: 140 _minor_gc_manager = MemoryManager::get_parnew_memory_manager(); 141 break; 142 #endif // INCLUDE_ALL_GCS 143 default: 144 guarantee(false, "Unrecognized generation spec"); 145 break; 146 } 147 if (policy->is_mark_sweep_policy()) { 148 _major_gc_manager = MemoryManager::get_msc_memory_manager(); 149 #if INCLUDE_ALL_GCS 150 } else if (policy->is_concurrent_mark_sweep_policy()) { 151 _major_gc_manager = MemoryManager::get_cms_memory_manager(); 152 #endif // INCLUDE_ALL_GCS 153 } else { 154 guarantee(false, "Unknown two-gen policy"); 155 } 156 } else { 157 guarantee(false, "Non two-gen policy"); 158 } 159 _managers_list->append(_minor_gc_manager); 160 _managers_list->append(_major_gc_manager); 161 162 add_generation_memory_pool(heap->get_gen(minor), _major_gc_manager, _minor_gc_manager); 163 add_generation_memory_pool(heap->get_gen(major), _major_gc_manager); 164 } 165 166 #if INCLUDE_ALL_GCS 167 // Add memory pools for ParallelScavengeHeap 168 // This function currently only supports two generations collected heap. 169 // The collector for ParallelScavengeHeap will have two memory managers. 170 void MemoryService::add_parallel_scavenge_heap_info(ParallelScavengeHeap* heap) { 171 // Two managers to keep statistics about _minor_gc_manager and _major_gc_manager GC. 172 _minor_gc_manager = MemoryManager::get_psScavenge_memory_manager(); 173 _major_gc_manager = MemoryManager::get_psMarkSweep_memory_manager(); 174 _managers_list->append(_minor_gc_manager); 175 _managers_list->append(_major_gc_manager); 176 177 add_psYoung_memory_pool(heap->young_gen(), _major_gc_manager, _minor_gc_manager); 178 add_psOld_memory_pool(heap->old_gen(), _major_gc_manager); 179 } 180 181 void MemoryService::add_g1_heap_info(G1CollectedHeap* g1h) { 182 assert(UseG1GC, "sanity"); 183 184 _minor_gc_manager = MemoryManager::get_g1YoungGen_memory_manager(); 185 _major_gc_manager = MemoryManager::get_g1OldGen_memory_manager(); 186 _managers_list->append(_minor_gc_manager); 187 _managers_list->append(_major_gc_manager); 188 189 add_g1YoungGen_memory_pool(g1h, _major_gc_manager, _minor_gc_manager); 190 add_g1OldGen_memory_pool(g1h, _major_gc_manager); 191 } 192 #endif // INCLUDE_ALL_GCS 193 194 MemoryPool* MemoryService::add_gen(Generation* gen, 195 const char* name, 196 bool is_heap, 197 bool support_usage_threshold) { 198 199 MemoryPool::PoolType type = (is_heap ? MemoryPool::Heap : MemoryPool::NonHeap); 200 GenerationPool* pool = new GenerationPool(gen, name, type, support_usage_threshold); 201 _pools_list->append(pool); 202 return (MemoryPool*) pool; 203 } 204 205 MemoryPool* MemoryService::add_space(ContiguousSpace* space, 206 const char* name, 207 bool is_heap, 208 size_t max_size, 209 bool support_usage_threshold) { 210 MemoryPool::PoolType type = (is_heap ? MemoryPool::Heap : MemoryPool::NonHeap); 211 ContiguousSpacePool* pool = new ContiguousSpacePool(space, name, type, max_size, support_usage_threshold); 212 213 _pools_list->append(pool); 214 return (MemoryPool*) pool; 215 } 216 217 MemoryPool* MemoryService::add_survivor_spaces(DefNewGeneration* gen, 218 const char* name, 219 bool is_heap, 220 size_t max_size, 221 bool support_usage_threshold) { 222 MemoryPool::PoolType type = (is_heap ? MemoryPool::Heap : MemoryPool::NonHeap); 223 SurvivorContiguousSpacePool* pool = new SurvivorContiguousSpacePool(gen, name, type, max_size, support_usage_threshold); 224 225 _pools_list->append(pool); 226 return (MemoryPool*) pool; 227 } 228 229 #if INCLUDE_ALL_GCS 230 MemoryPool* MemoryService::add_cms_space(CompactibleFreeListSpace* space, 231 const char* name, 232 bool is_heap, 233 size_t max_size, 234 bool support_usage_threshold) { 235 MemoryPool::PoolType type = (is_heap ? MemoryPool::Heap : MemoryPool::NonHeap); 236 CompactibleFreeListSpacePool* pool = new CompactibleFreeListSpacePool(space, name, type, max_size, support_usage_threshold); 237 _pools_list->append(pool); 238 return (MemoryPool*) pool; 239 } 240 #endif // INCLUDE_ALL_GCS 241 242 // Add memory pool(s) for one generation 243 void MemoryService::add_generation_memory_pool(Generation* gen, 244 MemoryManager* major_mgr, 245 MemoryManager* minor_mgr) { 246 guarantee(gen != NULL, "No generation for memory pool"); 247 Generation::Name kind = gen->kind(); 248 int index = _pools_list->length(); 249 250 switch (kind) { 251 case Generation::DefNew: { 252 assert(major_mgr != NULL && minor_mgr != NULL, "Should have two managers"); 253 DefNewGeneration* young_gen = (DefNewGeneration*) gen; 254 // Add a memory pool for each space and young gen doesn't 255 // support low memory detection as it is expected to get filled up. 256 MemoryPool* eden = add_space(young_gen->eden(), 257 "Eden Space", 258 true, /* is_heap */ 259 young_gen->max_eden_size(), 260 false /* support_usage_threshold */); 261 MemoryPool* survivor = add_survivor_spaces(young_gen, 262 "Survivor Space", 263 true, /* is_heap */ 264 young_gen->max_survivor_size(), 265 false /* support_usage_threshold */); 266 break; 267 } 268 269 #if INCLUDE_ALL_GCS 270 case Generation::ParNew: 271 case Generation::ASParNew: 272 { 273 assert(major_mgr != NULL && minor_mgr != NULL, "Should have two managers"); 274 // Add a memory pool for each space and young gen doesn't 275 // support low memory detection as it is expected to get filled up. 276 ParNewGeneration* parnew_gen = (ParNewGeneration*) gen; 277 MemoryPool* eden = add_space(parnew_gen->eden(), 278 "Par Eden Space", 279 true /* is_heap */, 280 parnew_gen->max_eden_size(), 281 false /* support_usage_threshold */); 282 MemoryPool* survivor = add_survivor_spaces(parnew_gen, 283 "Par Survivor Space", 284 true, /* is_heap */ 285 parnew_gen->max_survivor_size(), 286 false /* support_usage_threshold */); 287 288 break; 289 } 290 #endif // INCLUDE_ALL_GCS 291 292 case Generation::MarkSweepCompact: { 293 assert(major_mgr != NULL && minor_mgr == NULL, "Should have only one manager"); 294 add_gen(gen, 295 "Tenured Gen", 296 true, /* is_heap */ 297 true /* support_usage_threshold */); 298 break; 299 } 300 301 #if INCLUDE_ALL_GCS 302 case Generation::ConcurrentMarkSweep: 303 case Generation::ASConcurrentMarkSweep: 304 { 305 assert(major_mgr != NULL && minor_mgr == NULL, "Should have only one manager"); 306 ConcurrentMarkSweepGeneration* cms = (ConcurrentMarkSweepGeneration*) gen; 307 MemoryPool* pool = add_cms_space(cms->cmsSpace(), 308 "CMS Old Gen", 309 true, /* is_heap */ 310 cms->reserved().byte_size(), 311 true /* support_usage_threshold */); 312 break; 313 } 314 #endif // INCLUDE_ALL_GCS 315 316 default: 317 assert(false, "should not reach here"); 318 // no memory pool added for others 319 break; 320 } 321 322 assert(major_mgr != NULL, "Should have at least one manager"); 323 // Link managers and the memory pools together 324 for (int i = index; i < _pools_list->length(); i++) { 325 MemoryPool* pool = _pools_list->at(i); 326 major_mgr->add_pool(pool); 327 if (minor_mgr != NULL) { 328 minor_mgr->add_pool(pool); 329 } 330 } 331 } 332 333 334 #if INCLUDE_ALL_GCS 335 void MemoryService::add_psYoung_memory_pool(PSYoungGen* gen, MemoryManager* major_mgr, MemoryManager* minor_mgr) { 336 assert(major_mgr != NULL && minor_mgr != NULL, "Should have two managers"); 337 338 // Add a memory pool for each space and young gen doesn't 339 // support low memory detection as it is expected to get filled up. 340 EdenMutableSpacePool* eden = new EdenMutableSpacePool(gen, 341 gen->eden_space(), 342 "PS Eden Space", 343 MemoryPool::Heap, 344 false /* support_usage_threshold */); 345 346 SurvivorMutableSpacePool* survivor = new SurvivorMutableSpacePool(gen, 347 "PS Survivor Space", 348 MemoryPool::Heap, 349 false /* support_usage_threshold */); 350 351 major_mgr->add_pool(eden); 352 major_mgr->add_pool(survivor); 353 minor_mgr->add_pool(eden); 354 minor_mgr->add_pool(survivor); 355 _pools_list->append(eden); 356 _pools_list->append(survivor); 357 } 358 359 void MemoryService::add_psOld_memory_pool(PSOldGen* gen, MemoryManager* mgr) { 360 PSGenerationPool* old_gen = new PSGenerationPool(gen, 361 "PS Old Gen", 362 MemoryPool::Heap, 363 true /* support_usage_threshold */); 364 mgr->add_pool(old_gen); 365 _pools_list->append(old_gen); 366 } 367 368 void MemoryService::add_g1YoungGen_memory_pool(G1CollectedHeap* g1h, 369 MemoryManager* major_mgr, 370 MemoryManager* minor_mgr) { 371 assert(major_mgr != NULL && minor_mgr != NULL, "should have two managers"); 372 373 G1EdenPool* eden = new G1EdenPool(g1h); 374 G1SurvivorPool* survivor = new G1SurvivorPool(g1h); 375 376 major_mgr->add_pool(eden); 377 major_mgr->add_pool(survivor); 378 minor_mgr->add_pool(eden); 379 minor_mgr->add_pool(survivor); 380 _pools_list->append(eden); 381 _pools_list->append(survivor); 382 } 383 384 void MemoryService::add_g1OldGen_memory_pool(G1CollectedHeap* g1h, 385 MemoryManager* mgr) { 386 assert(mgr != NULL, "should have one manager"); 387 388 G1OldGenPool* old_gen = new G1OldGenPool(g1h); 389 mgr->add_pool(old_gen); 390 _pools_list->append(old_gen); 391 } 392 #endif // INCLUDE_ALL_GCS 393 394 void MemoryService::add_code_heap_memory_pool(CodeHeap* heap) { 395 _code_heap_pool = new CodeHeapPool(heap, 396 "Code Cache", 397 true /* support_usage_threshold */); 398 MemoryManager* mgr = MemoryManager::get_code_cache_memory_manager(); 399 mgr->add_pool(_code_heap_pool); 400 401 _pools_list->append(_code_heap_pool); 402 _managers_list->append(mgr); 403 } 404 405 void MemoryService::add_metaspace_memory_pools() { 406 MemoryManager* mgr = MemoryManager::get_metaspace_memory_manager(); 407 408 _metaspace_pool = new MetaspacePool(); 409 mgr->add_pool(_metaspace_pool); 410 _pools_list->append(_metaspace_pool); 411 412 if (UseCompressedClassPointers) { 413 _compressed_class_pool = new CompressedKlassSpacePool(); 414 mgr->add_pool(_compressed_class_pool); 415 _pools_list->append(_compressed_class_pool); 416 } 417 418 _managers_list->append(mgr); 419 } 420 421 MemoryManager* MemoryService::get_memory_manager(instanceHandle mh) { 422 for (int i = 0; i < _managers_list->length(); i++) { 423 MemoryManager* mgr = _managers_list->at(i); 424 if (mgr->is_manager(mh)) { 425 return mgr; 426 } 427 } 428 return NULL; 429 } 430 431 MemoryPool* MemoryService::get_memory_pool(instanceHandle ph) { 432 for (int i = 0; i < _pools_list->length(); i++) { 433 MemoryPool* pool = _pools_list->at(i); 434 if (pool->is_pool(ph)) { 435 return pool; 436 } 437 } 438 return NULL; 439 } 440 441 void MemoryService::track_memory_usage() { 442 // Track the peak memory usage 443 for (int i = 0; i < _pools_list->length(); i++) { 444 MemoryPool* pool = _pools_list->at(i); 445 pool->record_peak_memory_usage(); 446 } 447 448 // Detect low memory 449 LowMemoryDetector::detect_low_memory(); 450 } 451 452 void MemoryService::track_memory_pool_usage(MemoryPool* pool) { 453 // Track the peak memory usage 454 pool->record_peak_memory_usage(); 455 456 // Detect low memory 457 if (LowMemoryDetector::is_enabled(pool)) { 458 LowMemoryDetector::detect_low_memory(pool); 459 } 460 } 461 462 void MemoryService::gc_begin(bool fullGC, bool recordGCBeginTime, 463 bool recordAccumulatedGCTime, 464 bool recordPreGCUsage, bool recordPeakUsage) { 465 466 GCMemoryManager* mgr; 467 if (fullGC) { 468 mgr = _major_gc_manager; 469 } else { 470 mgr = _minor_gc_manager; 471 } 472 assert(mgr->is_gc_memory_manager(), "Sanity check"); 473 mgr->gc_begin(recordGCBeginTime, recordPreGCUsage, recordAccumulatedGCTime); 474 475 // Track the peak memory usage when GC begins 476 if (recordPeakUsage) { 477 for (int i = 0; i < _pools_list->length(); i++) { 478 MemoryPool* pool = _pools_list->at(i); 479 pool->record_peak_memory_usage(); 480 } 481 } 482 } 483 484 void MemoryService::gc_end(bool fullGC, bool recordPostGCUsage, 485 bool recordAccumulatedGCTime, 486 bool recordGCEndTime, bool countCollection, 487 GCCause::Cause cause) { 488 489 GCMemoryManager* mgr; 490 if (fullGC) { 491 mgr = (GCMemoryManager*) _major_gc_manager; 492 } else { 493 mgr = (GCMemoryManager*) _minor_gc_manager; 494 } 495 assert(mgr->is_gc_memory_manager(), "Sanity check"); 496 497 // register the GC end statistics and memory usage 498 mgr->gc_end(recordPostGCUsage, recordAccumulatedGCTime, recordGCEndTime, 499 countCollection, cause); 500 } 501 502 void MemoryService::oops_do(OopClosure* f) { 503 int i; 504 505 for (i = 0; i < _pools_list->length(); i++) { 506 MemoryPool* pool = _pools_list->at(i); 507 pool->oops_do(f); 508 } 509 for (i = 0; i < _managers_list->length(); i++) { 510 MemoryManager* mgr = _managers_list->at(i); 511 mgr->oops_do(f); 512 } 513 } 514 515 bool MemoryService::set_verbose(bool verbose) { 516 MutexLocker m(Management_lock); 517 // verbose will be set to the previous value 518 bool succeed = CommandLineFlags::boolAtPut((char*)"PrintGC", &verbose, Flag::MANAGEMENT); 519 assert(succeed, "Setting PrintGC flag fails"); 520 ClassLoadingService::reset_trace_class_unloading(); 521 522 return verbose; 523 } 524 525 Handle MemoryService::create_MemoryUsage_obj(MemoryUsage usage, TRAPS) { 526 Klass* k = Management::java_lang_management_MemoryUsage_klass(CHECK_NH); 527 instanceKlassHandle ik(THREAD, k); 528 529 instanceHandle obj = ik->allocate_instance_handle(CHECK_NH); 530 531 JavaValue result(T_VOID); 532 JavaCallArguments args(10); 533 args.push_oop(obj); // receiver 534 args.push_long(usage.init_size_as_jlong()); // Argument 1 535 args.push_long(usage.used_as_jlong()); // Argument 2 536 args.push_long(usage.committed_as_jlong()); // Argument 3 537 args.push_long(usage.max_size_as_jlong()); // Argument 4 538 539 JavaCalls::call_special(&result, 540 ik, 541 vmSymbols::object_initializer_name(), 542 vmSymbols::long_long_long_long_void_signature(), 543 &args, 544 CHECK_NH); 545 return obj; 546 } 547 // 548 // GC manager type depends on the type of Generation. Depending on the space 549 // availablity and vm options the gc uses major gc manager or minor gc 550 // manager or both. The type of gc manager depends on the generation kind. 551 // For DefNew, ParNew and ASParNew generation doing scavenge gc uses minor 552 // gc manager (so _fullGC is set to false ) and for other generation kinds 553 // doing mark-sweep-compact uses major gc manager (so _fullGC is set 554 // to true). 555 TraceMemoryManagerStats::TraceMemoryManagerStats(Generation::Name kind, GCCause::Cause cause) { 556 switch (kind) { 557 case Generation::DefNew: 558 #if INCLUDE_ALL_GCS 559 case Generation::ParNew: 560 case Generation::ASParNew: 561 #endif // INCLUDE_ALL_GCS 562 _fullGC=false; 563 break; 564 case Generation::MarkSweepCompact: 565 #if INCLUDE_ALL_GCS 566 case Generation::ConcurrentMarkSweep: 567 case Generation::ASConcurrentMarkSweep: 568 #endif // INCLUDE_ALL_GCS 569 _fullGC=true; 570 break; 571 default: 572 assert(false, "Unrecognized gc generation kind."); 573 } 574 // this has to be called in a stop the world pause and represent 575 // an entire gc pause, start to finish: 576 initialize(_fullGC, cause,true, true, true, true, true, true, true); 577 } 578 TraceMemoryManagerStats::TraceMemoryManagerStats(bool fullGC, 579 GCCause::Cause cause, 580 bool recordGCBeginTime, 581 bool recordPreGCUsage, 582 bool recordPeakUsage, 583 bool recordPostGCUsage, 584 bool recordAccumulatedGCTime, 585 bool recordGCEndTime, 586 bool countCollection) { 587 initialize(fullGC, cause, recordGCBeginTime, recordPreGCUsage, recordPeakUsage, 588 recordPostGCUsage, recordAccumulatedGCTime, recordGCEndTime, 589 countCollection); 590 } 591 592 // for a subclass to create then initialize an instance before invoking 593 // the MemoryService 594 void TraceMemoryManagerStats::initialize(bool fullGC, 595 GCCause::Cause cause, 596 bool recordGCBeginTime, 597 bool recordPreGCUsage, 598 bool recordPeakUsage, 599 bool recordPostGCUsage, 600 bool recordAccumulatedGCTime, 601 bool recordGCEndTime, 602 bool countCollection) { 603 _fullGC = fullGC; 604 _recordGCBeginTime = recordGCBeginTime; 605 _recordPreGCUsage = recordPreGCUsage; 606 _recordPeakUsage = recordPeakUsage; 607 _recordPostGCUsage = recordPostGCUsage; 608 _recordAccumulatedGCTime = recordAccumulatedGCTime; 609 _recordGCEndTime = recordGCEndTime; 610 _countCollection = countCollection; 611 _cause = cause; 612 613 MemoryService::gc_begin(_fullGC, _recordGCBeginTime, _recordAccumulatedGCTime, 614 _recordPreGCUsage, _recordPeakUsage); 615 } 616 617 TraceMemoryManagerStats::~TraceMemoryManagerStats() { 618 MemoryService::gc_end(_fullGC, _recordPostGCUsage, _recordAccumulatedGCTime, 619 _recordGCEndTime, _countCollection, _cause); 620 }