1 /* 2 * Copyright (c) 2012, 2017, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "gc/shared/copyFailedInfo.hpp" 27 #include "gc/shared/gcHeapSummary.hpp" 28 #include "gc/shared/gcTimer.hpp" 29 #include "gc/shared/gcTrace.hpp" 30 #include "gc/shared/gcWhen.hpp" 31 #include "jfr/jfrEvents.hpp" 32 #include "runtime/os.hpp" 33 #include "utilities/macros.hpp" 34 #if INCLUDE_G1GC 35 #include "gc/g1/g1EvacuationInfo.hpp" 36 #include "gc/g1/g1YCTypes.hpp" 37 #endif 38 39 // All GC dependencies against the trace framework is contained within this file. 40 41 typedef uintptr_t TraceAddress; 42 43 void GCTracer::send_garbage_collection_event() const { 44 EventGarbageCollection event(UNTIMED); 45 if (event.should_commit()) { 46 event.set_gcId(GCId::current()); 47 event.set_name(_shared_gc_info.name()); 48 event.set_cause((u2) _shared_gc_info.cause()); 49 event.set_sumOfPauses(_shared_gc_info.sum_of_pauses()); 50 event.set_longestPause(_shared_gc_info.longest_pause()); 51 event.set_starttime(_shared_gc_info.start_timestamp()); 52 event.set_endtime(_shared_gc_info.end_timestamp()); 53 event.commit(); 54 } 55 } 56 57 void GCTracer::send_reference_stats_event(ReferenceType type, size_t count) const { 58 EventGCReferenceStatistics e; 59 if (e.should_commit()) { 60 e.set_gcId(GCId::current()); 61 e.set_type((u1)type); 62 e.set_count(count); 63 e.commit(); 64 } 65 } 66 67 void GCTracer::send_metaspace_chunk_free_list_summary(GCWhen::Type when, metaspace::MetadataType mdtype, 68 const MetaspaceChunkFreeListSummary& summary) const { 69 EventMetaspaceChunkFreeListSummary e; 70 if (e.should_commit()) { 71 e.set_gcId(GCId::current()); 72 e.set_when(when); 73 e.set_metadataType(mdtype); 74 75 e.set_specializedChunks(summary.num_specialized_chunks()); 76 e.set_specializedChunksTotalSize(summary.specialized_chunks_size_in_bytes()); 77 78 e.set_smallChunks(summary.num_small_chunks()); 79 e.set_smallChunksTotalSize(summary.small_chunks_size_in_bytes()); 80 81 e.set_mediumChunks(summary.num_medium_chunks()); 82 e.set_mediumChunksTotalSize(summary.medium_chunks_size_in_bytes()); 83 84 e.set_humongousChunks(summary.num_humongous_chunks()); 85 e.set_humongousChunksTotalSize(summary.humongous_chunks_size_in_bytes()); 86 87 e.commit(); 88 } 89 } 90 91 void ParallelOldTracer::send_parallel_old_event() const { 92 EventParallelOldGarbageCollection e(UNTIMED); 93 if (e.should_commit()) { 94 e.set_gcId(GCId::current()); 95 e.set_densePrefix((TraceAddress)_parallel_old_gc_info.dense_prefix()); 96 e.set_starttime(_shared_gc_info.start_timestamp()); 97 e.set_endtime(_shared_gc_info.end_timestamp()); 98 e.commit(); 99 } 100 } 101 102 void YoungGCTracer::send_young_gc_event() const { 103 EventYoungGarbageCollection e(UNTIMED); 104 if (e.should_commit()) { 105 e.set_gcId(GCId::current()); 106 e.set_tenuringThreshold(_tenuring_threshold); 107 e.set_starttime(_shared_gc_info.start_timestamp()); 108 e.set_endtime(_shared_gc_info.end_timestamp()); 109 e.commit(); 110 } 111 } 112 113 bool YoungGCTracer::should_send_promotion_in_new_plab_event() const { 114 return EventPromoteObjectInNewPLAB::is_enabled(); 115 } 116 117 bool YoungGCTracer::should_send_promotion_outside_plab_event() const { 118 return EventPromoteObjectOutsidePLAB::is_enabled(); 119 } 120 121 void YoungGCTracer::send_promotion_in_new_plab_event(Klass* klass, size_t obj_size, 122 uint age, bool tenured, 123 size_t plab_size) const { 124 125 EventPromoteObjectInNewPLAB event; 126 if (event.should_commit()) { 127 event.set_gcId(GCId::current()); 128 event.set_objectClass(klass); 129 event.set_objectSize(obj_size); 130 event.set_tenured(tenured); 131 event.set_tenuringAge(age); 132 event.set_plabSize(plab_size); 133 event.commit(); 134 } 135 } 136 137 void YoungGCTracer::send_promotion_outside_plab_event(Klass* klass, size_t obj_size, 138 uint age, bool tenured) const { 139 140 EventPromoteObjectOutsidePLAB event; 141 if (event.should_commit()) { 142 event.set_gcId(GCId::current()); 143 event.set_objectClass(klass); 144 event.set_objectSize(obj_size); 145 event.set_tenured(tenured); 146 event.set_tenuringAge(age); 147 event.commit(); 148 } 149 } 150 151 void OldGCTracer::send_old_gc_event() const { 152 EventOldGarbageCollection e(UNTIMED); 153 if (e.should_commit()) { 154 e.set_gcId(GCId::current()); 155 e.set_starttime(_shared_gc_info.start_timestamp()); 156 e.set_endtime(_shared_gc_info.end_timestamp()); 157 e.commit(); 158 } 159 } 160 161 static JfrStructCopyFailed to_struct(const CopyFailedInfo& cf_info) { 162 JfrStructCopyFailed failed_info; 163 failed_info.set_objectCount(cf_info.failed_count()); 164 failed_info.set_firstSize(cf_info.first_size()); 165 failed_info.set_smallestSize(cf_info.smallest_size()); 166 failed_info.set_totalSize(cf_info.total_size()); 167 return failed_info; 168 } 169 170 void YoungGCTracer::send_promotion_failed_event(const PromotionFailedInfo& pf_info) const { 171 EventPromotionFailed e; 172 if (e.should_commit()) { 173 e.set_gcId(GCId::current()); 174 e.set_promotionFailed(to_struct(pf_info)); 175 e.set_thread(pf_info.thread_trace_id()); 176 e.commit(); 177 } 178 } 179 180 // Common to CMS and G1 181 void OldGCTracer::send_concurrent_mode_failure_event() { 182 EventConcurrentModeFailure e; 183 if (e.should_commit()) { 184 e.set_gcId(GCId::current()); 185 e.commit(); 186 } 187 } 188 189 #if INCLUDE_G1GC 190 void G1NewTracer::send_g1_young_gc_event() { 191 EventG1GarbageCollection e(UNTIMED); 192 if (e.should_commit()) { 193 e.set_gcId(GCId::current()); 194 e.set_type(_g1_young_gc_info.type()); 195 e.set_starttime(_shared_gc_info.start_timestamp()); 196 e.set_endtime(_shared_gc_info.end_timestamp()); 197 e.commit(); 198 } 199 } 200 201 void G1MMUTracer::send_g1_mmu_event(double time_slice_ms, double gc_time_ms, double max_time_ms) { 202 EventG1MMU e; 203 if (e.should_commit()) { 204 e.set_gcId(GCId::current()); 205 e.set_timeSlice(time_slice_ms); 206 e.set_gcTime(gc_time_ms); 207 e.set_pauseTarget(max_time_ms); 208 e.commit(); 209 } 210 } 211 212 void G1NewTracer::send_evacuation_info_event(G1EvacuationInfo* info) { 213 EventEvacuationInformation e; 214 if (e.should_commit()) { 215 e.set_gcId(GCId::current()); 216 e.set_cSetRegions(info->collectionset_regions()); 217 e.set_cSetUsedBefore(info->collectionset_used_before()); 218 e.set_cSetUsedAfter(info->collectionset_used_after()); 219 e.set_allocationRegions(info->allocation_regions()); 220 e.set_allocationRegionsUsedBefore(info->alloc_regions_used_before()); 221 e.set_allocationRegionsUsedAfter(info->alloc_regions_used_before() + info->bytes_copied()); 222 e.set_bytesCopied(info->bytes_copied()); 223 e.set_regionsFreed(info->regions_freed()); 224 e.commit(); 225 } 226 } 227 228 void G1NewTracer::send_evacuation_failed_event(const EvacuationFailedInfo& ef_info) const { 229 EventEvacuationFailed e; 230 if (e.should_commit()) { 231 e.set_gcId(GCId::current()); 232 e.set_evacuationFailed(to_struct(ef_info)); 233 e.commit(); 234 } 235 } 236 237 static JfrStructG1EvacuationStatistics 238 create_g1_evacstats(unsigned gcid, const G1EvacSummary& summary) { 239 JfrStructG1EvacuationStatistics s; 240 s.set_gcId(gcid); 241 s.set_allocated(summary.allocated() * HeapWordSize); 242 s.set_wasted(summary.wasted() * HeapWordSize); 243 s.set_used(summary.used() * HeapWordSize); 244 s.set_undoWaste(summary.undo_wasted() * HeapWordSize); 245 s.set_regionEndWaste(summary.region_end_waste() * HeapWordSize); 246 s.set_regionsRefilled(summary.regions_filled()); 247 s.set_directAllocated(summary.direct_allocated() * HeapWordSize); 248 s.set_failureUsed(summary.failure_used() * HeapWordSize); 249 s.set_failureWaste(summary.failure_waste() * HeapWordSize); 250 return s; 251 } 252 253 void G1NewTracer::send_young_evacuation_statistics(const G1EvacSummary& summary) const { 254 EventG1EvacuationYoungStatistics surv_evt; 255 if (surv_evt.should_commit()) { 256 surv_evt.set_statistics(create_g1_evacstats(GCId::current(), summary)); 257 surv_evt.commit(); 258 } 259 } 260 261 void G1NewTracer::send_old_evacuation_statistics(const G1EvacSummary& summary) const { 262 EventG1EvacuationOldStatistics old_evt; 263 if (old_evt.should_commit()) { 264 old_evt.set_statistics(create_g1_evacstats(GCId::current(), summary)); 265 old_evt.commit(); 266 } 267 } 268 269 void G1NewTracer::send_basic_ihop_statistics(size_t threshold, 270 size_t target_occupancy, 271 size_t current_occupancy, 272 size_t last_allocation_size, 273 double last_allocation_duration, 274 double last_marking_length) { 275 EventG1BasicIHOP evt; 276 if (evt.should_commit()) { 277 evt.set_gcId(GCId::current()); 278 evt.set_threshold(threshold); 279 evt.set_targetOccupancy(target_occupancy); 280 evt.set_thresholdPercentage(target_occupancy > 0 ? ((double)threshold / target_occupancy) : 0.0); 281 evt.set_currentOccupancy(current_occupancy); 282 evt.set_recentMutatorAllocationSize(last_allocation_size); 283 evt.set_recentMutatorDuration(last_allocation_duration * MILLIUNITS); 284 evt.set_recentAllocationRate(last_allocation_duration != 0.0 ? last_allocation_size / last_allocation_duration : 0.0); 285 evt.set_lastMarkingDuration(last_marking_length * MILLIUNITS); 286 evt.commit(); 287 } 288 } 289 290 void G1NewTracer::send_adaptive_ihop_statistics(size_t threshold, 291 size_t internal_target_occupancy, 292 size_t current_occupancy, 293 size_t additional_buffer_size, 294 double predicted_allocation_rate, 295 double predicted_marking_length, 296 bool prediction_active) { 297 EventG1AdaptiveIHOP evt; 298 if (evt.should_commit()) { 299 evt.set_gcId(GCId::current()); 300 evt.set_threshold(threshold); 301 evt.set_thresholdPercentage(internal_target_occupancy > 0 ? ((double)threshold / internal_target_occupancy) : 0.0); 302 evt.set_ihopTargetOccupancy(internal_target_occupancy); 303 evt.set_currentOccupancy(current_occupancy); 304 evt.set_additionalBufferSize(additional_buffer_size); 305 evt.set_predictedAllocationRate(predicted_allocation_rate); 306 evt.set_predictedMarkingDuration(predicted_marking_length * MILLIUNITS); 307 evt.set_predictionActive(prediction_active); 308 evt.commit(); 309 } 310 } 311 312 #endif // INCLUDE_G1GC 313 314 static JfrStructVirtualSpace to_struct(const VirtualSpaceSummary& summary) { 315 JfrStructVirtualSpace space; 316 space.set_start((TraceAddress)summary.start()); 317 space.set_committedEnd((TraceAddress)summary.committed_end()); 318 space.set_committedSize(summary.committed_size()); 319 space.set_reservedEnd((TraceAddress)summary.reserved_end()); 320 space.set_reservedSize(summary.reserved_size()); 321 return space; 322 } 323 324 static JfrStructObjectSpace to_struct(const SpaceSummary& summary) { 325 JfrStructObjectSpace space; 326 space.set_start((TraceAddress)summary.start()); 327 space.set_end((TraceAddress)summary.end()); 328 space.set_used(summary.used()); 329 space.set_size(summary.size()); 330 return space; 331 } 332 333 class GCHeapSummaryEventSender : public GCHeapSummaryVisitor { 334 GCWhen::Type _when; 335 public: 336 GCHeapSummaryEventSender(GCWhen::Type when) : _when(when) {} 337 338 void visit(const GCHeapSummary* heap_summary) const { 339 const VirtualSpaceSummary& heap_space = heap_summary->heap(); 340 341 EventGCHeapSummary e; 342 if (e.should_commit()) { 343 e.set_gcId(GCId::current()); 344 e.set_when((u1)_when); 345 e.set_heapSpace(to_struct(heap_space)); 346 e.set_heapUsed(heap_summary->used()); 347 e.commit(); 348 } 349 } 350 351 void visit(const G1HeapSummary* g1_heap_summary) const { 352 visit((GCHeapSummary*)g1_heap_summary); 353 354 EventG1HeapSummary e; 355 if (e.should_commit()) { 356 e.set_gcId(GCId::current()); 357 e.set_when((u1)_when); 358 e.set_edenUsedSize(g1_heap_summary->edenUsed()); 359 e.set_edenTotalSize(g1_heap_summary->edenCapacity()); 360 e.set_survivorUsedSize(g1_heap_summary->survivorUsed()); 361 e.set_numberOfRegions(g1_heap_summary->numberOfRegions()); 362 e.commit(); 363 } 364 } 365 366 void visit(const PSHeapSummary* ps_heap_summary) const { 367 visit((GCHeapSummary*)ps_heap_summary); 368 369 const VirtualSpaceSummary& old_summary = ps_heap_summary->old(); 370 const SpaceSummary& old_space = ps_heap_summary->old_space(); 371 const VirtualSpaceSummary& young_summary = ps_heap_summary->young(); 372 const SpaceSummary& eden_space = ps_heap_summary->eden(); 373 const SpaceSummary& from_space = ps_heap_summary->from(); 374 const SpaceSummary& to_space = ps_heap_summary->to(); 375 376 EventPSHeapSummary e; 377 if (e.should_commit()) { 378 e.set_gcId(GCId::current()); 379 e.set_when((u1)_when); 380 381 e.set_oldSpace(to_struct(ps_heap_summary->old())); 382 e.set_oldObjectSpace(to_struct(ps_heap_summary->old_space())); 383 e.set_youngSpace(to_struct(ps_heap_summary->young())); 384 e.set_edenSpace(to_struct(ps_heap_summary->eden())); 385 e.set_fromSpace(to_struct(ps_heap_summary->from())); 386 e.set_toSpace(to_struct(ps_heap_summary->to())); 387 e.commit(); 388 } 389 } 390 }; 391 392 void GCTracer::send_gc_heap_summary_event(GCWhen::Type when, const GCHeapSummary& heap_summary) const { 393 GCHeapSummaryEventSender visitor(when); 394 heap_summary.accept(&visitor); 395 } 396 397 static JfrStructMetaspaceSizes to_struct(const MetaspaceSizes& sizes) { 398 JfrStructMetaspaceSizes meta_sizes; 399 400 meta_sizes.set_committed(sizes.committed()); 401 meta_sizes.set_used(sizes.used()); 402 meta_sizes.set_reserved(sizes.reserved()); 403 404 return meta_sizes; 405 } 406 407 void GCTracer::send_meta_space_summary_event(GCWhen::Type when, const MetaspaceSummary& meta_space_summary) const { 408 EventMetaspaceSummary e; 409 if (e.should_commit()) { 410 e.set_gcId(GCId::current()); 411 e.set_when((u1) when); 412 e.set_gcThreshold(meta_space_summary.capacity_until_GC()); 413 e.set_metaspace(to_struct(meta_space_summary.meta_space())); 414 e.set_dataSpace(to_struct(meta_space_summary.data_space())); 415 e.set_classSpace(to_struct(meta_space_summary.class_space())); 416 e.commit(); 417 } 418 } 419 420 class PhaseSender : public PhaseVisitor { 421 void visit_pause(GCPhase* phase) { 422 assert(phase->level() < PhasesStack::PHASE_LEVELS, "Need more event types for PausePhase"); 423 424 switch (phase->level()) { 425 case 0: send_phase<EventGCPhasePause>(phase); break; 426 case 1: send_phase<EventGCPhasePauseLevel1>(phase); break; 427 case 2: send_phase<EventGCPhasePauseLevel2>(phase); break; 428 case 3: send_phase<EventGCPhasePauseLevel3>(phase); break; 429 case 4: send_phase<EventGCPhasePauseLevel4>(phase); break; 430 default: /* Ignore sending this phase */ break; 431 } 432 } 433 434 void visit_concurrent(GCPhase* phase) { 435 assert(phase->level() < 1, "There is only one level for ConcurrentPhase"); 436 437 switch (phase->level()) { 438 case 0: send_phase<EventGCPhaseConcurrent>(phase); break; 439 default: /* Ignore sending this phase */ break; 440 } 441 } 442 443 public: 444 template<typename T> 445 void send_phase(GCPhase* phase) { 446 T event(UNTIMED); 447 if (event.should_commit()) { 448 event.set_gcId(GCId::current()); 449 event.set_name(phase->name()); 450 event.set_starttime(phase->start()); 451 event.set_endtime(phase->end()); 452 event.commit(); 453 } 454 } 455 456 void visit(GCPhase* phase) { 457 if (phase->type() == GCPhase::PausePhaseType) { 458 visit_pause(phase); 459 } else { 460 assert(phase->type() == GCPhase::ConcurrentPhaseType, "Should be ConcurrentPhaseType"); 461 visit_concurrent(phase); 462 } 463 } 464 }; 465 466 void GCTracer::send_phase_events(TimePartitions* time_partitions) const { 467 PhaseSender phase_reporter; 468 469 TimePartitionPhasesIterator iter(time_partitions); 470 while (iter.has_next()) { 471 GCPhase* phase = iter.next(); 472 phase->accept(&phase_reporter); 473 } 474 }