1 /* 2 * Copyright (c) 2012, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 #include "precompiled.hpp" 25 #include "memory/allocation.hpp" 26 #include "services/memBaseline.hpp" 27 #include "services/memTracker.hpp" 28 29 MemType2Name MemBaseline::MemType2NameMap[NUMBER_OF_MEMORY_TYPE] = { 30 {mtJavaHeap, "Java Heap"}, 31 {mtClass, "Class"}, 32 {mtThreadStack,"Thread Stack"}, 33 {mtThread, "Thread"}, 34 {mtCode, "Code"}, 35 {mtGC, "GC"}, 36 {mtCompiler, "Compiler"}, 37 {mtInternal, "Internal"}, 38 {mtOther, "Other"}, 39 {mtSymbol, "Symbol"}, 40 {mtNMT, "Memory Tracking"}, 41 {mtChunk, "Pooled Free Chunks"}, 42 {mtClassShared,"Shared spaces for classes"}, 43 {mtNone, "Unknown"} // It can happen when type tagging records are lagging 44 // behind 45 }; 46 47 MemBaseline::MemBaseline() { 48 _baselined = false; 49 50 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { 51 _malloc_data[index].set_type(MemType2NameMap[index]._flag); 52 _vm_data[index].set_type(MemType2NameMap[index]._flag); 53 _arena_data[index].set_type(MemType2NameMap[index]._flag); 54 } 55 56 _malloc_cs = NULL; 57 _vm_cs = NULL; 58 _vm_map = NULL; 59 60 _number_of_classes = 0; 61 _number_of_threads = 0; 62 } 63 64 65 void MemBaseline::clear() { 66 if (_malloc_cs != NULL) { 67 delete _malloc_cs; 68 _malloc_cs = NULL; 69 } 70 71 if (_vm_cs != NULL) { 72 delete _vm_cs; 73 _vm_cs = NULL; 74 } 75 76 if (_vm_map != NULL) { 77 delete _vm_map; 78 _vm_map = NULL; 79 } 80 81 reset(); 82 } 83 84 85 void MemBaseline::reset() { 86 _baselined = false; 87 _total_vm_reserved = 0; 88 _total_vm_committed = 0; 89 _total_malloced = 0; 90 _number_of_classes = 0; 91 92 if (_malloc_cs != NULL) _malloc_cs->clear(); 93 if (_vm_cs != NULL) _vm_cs->clear(); 94 if (_vm_map != NULL) _vm_map->clear(); 95 96 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { 97 _malloc_data[index].clear(); 98 _vm_data[index].clear(); 99 _arena_data[index].clear(); 100 } 101 } 102 103 MemBaseline::~MemBaseline() { 104 clear(); 105 } 106 107 // baseline malloc'd memory records, generate overall summary and summaries by 108 // memory types 109 bool MemBaseline::baseline_malloc_summary(const MemPointerArray* malloc_records) { 110 MemPointerArrayIteratorImpl malloc_itr((MemPointerArray*)malloc_records); 111 MemPointerRecord* malloc_ptr = (MemPointerRecord*)malloc_itr.current(); 112 size_t used_arena_size = 0; 113 int index; 114 while (malloc_ptr != NULL) { 115 index = flag2index(FLAGS_TO_MEMORY_TYPE(malloc_ptr->flags())); 116 size_t size = malloc_ptr->size(); 117 if (malloc_ptr->is_arena_memory_record()) { 118 // We do have anonymous arenas, they are either used as value objects, 119 // which are embedded inside other objects, or used as stack objects. 120 _arena_data[index].inc(size); 121 used_arena_size += size; 122 } else { 123 _total_malloced += size; 124 _malloc_data[index].inc(size); 125 if (malloc_ptr->is_arena_record()) { 126 // see if arena memory record present 127 MemPointerRecord* next_malloc_ptr = (MemPointerRecordEx*)malloc_itr.peek_next(); 128 if (next_malloc_ptr->is_arena_memory_record()) { 129 assert(next_malloc_ptr->is_memory_record_of_arena(malloc_ptr), 130 "Arena records do not match"); 131 size = next_malloc_ptr->size(); 132 _arena_data[index].inc(size); 133 used_arena_size += size; 134 malloc_itr.next(); 135 } 136 } 137 } 138 malloc_ptr = (MemPointerRecordEx*)malloc_itr.next(); 139 } 140 141 // substract used arena size to get size of arena chunk in free list 142 index = flag2index(mtChunk); 143 _malloc_data[index].reduce(used_arena_size); 144 // we really don't know how many chunks in free list, so just set to 145 // 0 146 _malloc_data[index].overwrite_counter(0); 147 148 return true; 149 } 150 151 // baseline mmap'd memory records, generate overall summary and summaries by 152 // memory types 153 bool MemBaseline::baseline_vm_summary(const MemPointerArray* vm_records) { 154 MemPointerArrayIteratorImpl vm_itr((MemPointerArray*)vm_records); 155 VMMemRegion* vm_ptr = (VMMemRegion*)vm_itr.current(); 156 int index; 157 while (vm_ptr != NULL) { 158 if (vm_ptr->is_reserved_region()) { 159 index = flag2index(FLAGS_TO_MEMORY_TYPE(vm_ptr->flags())); 160 // we use the number of thread stack to count threads 161 if (IS_MEMORY_TYPE(vm_ptr->flags(), mtThreadStack)) { 162 _number_of_threads ++; 163 } 164 _total_vm_reserved += vm_ptr->size(); 165 _vm_data[index].inc(vm_ptr->size(), 0); 166 } else { 167 _total_vm_committed += vm_ptr->size(); 168 _vm_data[index].inc(0, vm_ptr->size()); 169 } 170 vm_ptr = (VMMemRegion*)vm_itr.next(); 171 } 172 return true; 173 } 174 175 // baseline malloc'd memory by callsites, but only the callsites with memory allocation 176 // over 1KB are stored. 177 bool MemBaseline::baseline_malloc_details(const MemPointerArray* malloc_records) { 178 assert(MemTracker::track_callsite(), "detail tracking is off"); 179 180 MemPointerArrayIteratorImpl malloc_itr(const_cast<MemPointerArray*>(malloc_records)); 181 MemPointerRecordEx* malloc_ptr = (MemPointerRecordEx*)malloc_itr.current(); 182 MallocCallsitePointer malloc_callsite; 183 184 // initailize malloc callsite array 185 if (_malloc_cs == NULL) { 186 _malloc_cs = new (std::nothrow) MemPointerArrayImpl<MallocCallsitePointer>(64); 187 // out of native memory 188 if (_malloc_cs == NULL || _malloc_cs->out_of_memory()) { 189 return false; 190 } 191 } else { 192 _malloc_cs->clear(); 193 } 194 195 MemPointerArray* malloc_data = const_cast<MemPointerArray*>(malloc_records); 196 197 // sort into callsite pc order. Details are aggregated by callsites 198 malloc_data->sort((FN_SORT)malloc_sort_by_pc); 199 bool ret = true; 200 201 // baseline memory that is totaled over 1 KB 202 while (malloc_ptr != NULL) { 203 if (!MemPointerRecord::is_arena_memory_record(malloc_ptr->flags())) { 204 // skip thread stacks 205 if (!IS_MEMORY_TYPE(malloc_ptr->flags(), mtThreadStack)) { 206 if (malloc_callsite.addr() != malloc_ptr->pc()) { 207 if ((malloc_callsite.amount()/K) > 0) { 208 if (!_malloc_cs->append(&malloc_callsite)) { 209 ret = false; 210 break; 211 } 212 } 213 malloc_callsite = MallocCallsitePointer(malloc_ptr->pc()); 214 } 215 malloc_callsite.inc(malloc_ptr->size()); 216 } 217 } 218 malloc_ptr = (MemPointerRecordEx*)malloc_itr.next(); 219 } 220 221 // restore to address order. Snapshot malloc data is maintained in memory 222 // address order. 223 malloc_data->sort((FN_SORT)malloc_sort_by_addr); 224 225 if (!ret) { 226 return false; 227 } 228 // deal with last record 229 if (malloc_callsite.addr() != 0 && (malloc_callsite.amount()/K) > 0) { 230 if (!_malloc_cs->append(&malloc_callsite)) { 231 return false; 232 } 233 } 234 return true; 235 } 236 237 // baseline mmap'd memory by callsites 238 bool MemBaseline::baseline_vm_details(const MemPointerArray* vm_records) { 239 assert(MemTracker::track_callsite(), "detail tracking is off"); 240 241 VMCallsitePointer vm_callsite; 242 VMCallsitePointer* cur_callsite = NULL; 243 MemPointerArrayIteratorImpl vm_itr((MemPointerArray*)vm_records); 244 VMMemRegionEx* vm_ptr = (VMMemRegionEx*)vm_itr.current(); 245 246 // initialize virtual memory map array 247 if (_vm_map == NULL) { 248 _vm_map = new (std::nothrow) MemPointerArrayImpl<VMMemRegionEx>(vm_records->length()); 249 if (_vm_map == NULL || _vm_map->out_of_memory()) { 250 return false; 251 } 252 } else { 253 _vm_map->clear(); 254 } 255 256 // initialize virtual memory callsite array 257 if (_vm_cs == NULL) { 258 _vm_cs = new (std::nothrow) MemPointerArrayImpl<VMCallsitePointer>(64); 259 if (_vm_cs == NULL || _vm_cs->out_of_memory()) { 260 return false; 261 } 262 } else { 263 _vm_cs->clear(); 264 } 265 266 // consolidate virtual memory data 267 VMMemRegionEx* reserved_rec = NULL; 268 VMMemRegionEx* committed_rec = NULL; 269 270 // vm_ptr is coming in increasing base address order 271 while (vm_ptr != NULL) { 272 if (vm_ptr->is_reserved_region()) { 273 // consolidate reserved memory regions for virtual memory map. 274 // The criteria for consolidation is: 275 // 1. two adjacent reserved memory regions 276 // 2. belong to the same memory type 277 // 3. reserved from the same callsite 278 if (reserved_rec == NULL || 279 reserved_rec->base() + reserved_rec->size() != vm_ptr->addr() || 280 FLAGS_TO_MEMORY_TYPE(reserved_rec->flags()) != FLAGS_TO_MEMORY_TYPE(vm_ptr->flags()) || 281 reserved_rec->pc() != vm_ptr->pc()) { 282 if (!_vm_map->append(vm_ptr)) { 283 return false; 284 } 285 // inserted reserved region, we need the pointer to the element in virtual 286 // memory map array. 287 reserved_rec = (VMMemRegionEx*)_vm_map->at(_vm_map->length() - 1); 288 } else { 289 reserved_rec->expand_region(vm_ptr->addr(), vm_ptr->size()); 290 } 291 292 if (cur_callsite != NULL && !_vm_cs->append(cur_callsite)) { 293 return false; 294 } 295 vm_callsite = VMCallsitePointer(vm_ptr->pc()); 296 cur_callsite = &vm_callsite; 297 vm_callsite.inc(vm_ptr->size(), 0); 298 } else { 299 // consolidate committed memory regions for virtual memory map 300 // The criterial is: 301 // 1. two adjacent committed memory regions 302 // 2. committed from the same callsite 303 if (committed_rec == NULL || 304 committed_rec->base() + committed_rec->size() != vm_ptr->addr() || 305 committed_rec->pc() != vm_ptr->pc()) { 306 if (!_vm_map->append(vm_ptr)) { 307 return false; 308 } 309 committed_rec = (VMMemRegionEx*)_vm_map->at(_vm_map->length() - 1); 310 } else { 311 committed_rec->expand_region(vm_ptr->addr(), vm_ptr->size()); 312 } 313 vm_callsite.inc(0, vm_ptr->size()); 314 } 315 vm_ptr = (VMMemRegionEx*)vm_itr.next(); 316 } 317 // deal with last record 318 if (cur_callsite != NULL && !_vm_cs->append(cur_callsite)) { 319 return false; 320 } 321 322 // sort it into callsite pc order. Details are aggregated by callsites 323 _vm_cs->sort((FN_SORT)bl_vm_sort_by_pc); 324 325 // walk the array to consolidate record by pc 326 MemPointerArrayIteratorImpl itr(_vm_cs); 327 VMCallsitePointer* callsite_rec = (VMCallsitePointer*)itr.current(); 328 VMCallsitePointer* next_rec = (VMCallsitePointer*)itr.next(); 329 while (next_rec != NULL) { 330 assert(callsite_rec != NULL, "Sanity check"); 331 if (next_rec->addr() == callsite_rec->addr()) { 332 callsite_rec->inc(next_rec->reserved_amount(), next_rec->committed_amount()); 333 itr.remove(); 334 next_rec = (VMCallsitePointer*)itr.current(); 335 } else { 336 callsite_rec = next_rec; 337 next_rec = (VMCallsitePointer*)itr.next(); 338 } 339 } 340 341 return true; 342 } 343 344 // baseline a snapshot. If summary_only = false, memory usages aggregated by 345 // callsites are also baselined. 346 bool MemBaseline::baseline(MemSnapshot& snapshot, bool summary_only) { 347 MutexLockerEx snapshot_locker(snapshot._lock, true); 348 reset(); 349 _baselined = baseline_malloc_summary(snapshot._alloc_ptrs) && 350 baseline_vm_summary(snapshot._vm_ptrs); 351 _number_of_classes = snapshot.number_of_classes(); 352 353 if (!summary_only && MemTracker::track_callsite() && _baselined) { 354 _baselined = baseline_malloc_details(snapshot._alloc_ptrs) && 355 baseline_vm_details(snapshot._vm_ptrs); 356 } 357 return _baselined; 358 } 359 360 361 int MemBaseline::flag2index(MEMFLAGS flag) const { 362 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { 363 if (MemType2NameMap[index]._flag == flag) { 364 return index; 365 } 366 } 367 assert(false, "no type"); 368 return -1; 369 } 370 371 const char* MemBaseline::type2name(MEMFLAGS type) { 372 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { 373 if (MemType2NameMap[index]._flag == type) { 374 return MemType2NameMap[index]._name; 375 } 376 } 377 assert(false, err_msg("bad type %x", type)); 378 return NULL; 379 } 380 381 382 MemBaseline& MemBaseline::operator=(const MemBaseline& other) { 383 _total_malloced = other._total_malloced; 384 _total_vm_reserved = other._total_vm_reserved; 385 _total_vm_committed = other._total_vm_committed; 386 387 _baselined = other._baselined; 388 _number_of_classes = other._number_of_classes; 389 390 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { 391 _malloc_data[index] = other._malloc_data[index]; 392 _vm_data[index] = other._vm_data[index]; 393 _arena_data[index] = other._arena_data[index]; 394 } 395 396 if (MemTracker::track_callsite()) { 397 assert(_malloc_cs != NULL && _vm_cs != NULL, "out of memory"); 398 assert(other._malloc_cs != NULL && other._vm_cs != NULL, 399 "not properly baselined"); 400 _malloc_cs->clear(); 401 _vm_cs->clear(); 402 int index; 403 for (index = 0; index < other._malloc_cs->length(); index ++) { 404 _malloc_cs->append(other._malloc_cs->at(index)); 405 } 406 407 for (index = 0; index < other._vm_cs->length(); index ++) { 408 _vm_cs->append(other._vm_cs->at(index)); 409 } 410 } 411 return *this; 412 } 413 414 /* compare functions for sorting */ 415 416 // sort snapshot malloc'd records in callsite pc order 417 int MemBaseline::malloc_sort_by_pc(const void* p1, const void* p2) { 418 assert(MemTracker::track_callsite(),"Just check"); 419 const MemPointerRecordEx* mp1 = (const MemPointerRecordEx*)p1; 420 const MemPointerRecordEx* mp2 = (const MemPointerRecordEx*)p2; 421 return UNSIGNED_COMPARE(mp1->pc(), mp2->pc()); 422 } 423 424 // sort baselined malloc'd records in size order 425 int MemBaseline::bl_malloc_sort_by_size(const void* p1, const void* p2) { 426 assert(MemTracker::is_on(), "Just check"); 427 const MallocCallsitePointer* mp1 = (const MallocCallsitePointer*)p1; 428 const MallocCallsitePointer* mp2 = (const MallocCallsitePointer*)p2; 429 return UNSIGNED_COMPARE(mp2->amount(), mp1->amount()); 430 } 431 432 // sort baselined malloc'd records in callsite pc order 433 int MemBaseline::bl_malloc_sort_by_pc(const void* p1, const void* p2) { 434 assert(MemTracker::is_on(), "Just check"); 435 const MallocCallsitePointer* mp1 = (const MallocCallsitePointer*)p1; 436 const MallocCallsitePointer* mp2 = (const MallocCallsitePointer*)p2; 437 return UNSIGNED_COMPARE(mp1->addr(), mp2->addr()); 438 } 439 440 441 // sort baselined mmap'd records in size (reserved size) order 442 int MemBaseline::bl_vm_sort_by_size(const void* p1, const void* p2) { 443 assert(MemTracker::is_on(), "Just check"); 444 const VMCallsitePointer* mp1 = (const VMCallsitePointer*)p1; 445 const VMCallsitePointer* mp2 = (const VMCallsitePointer*)p2; 446 return UNSIGNED_COMPARE(mp2->reserved_amount(), mp1->reserved_amount()); 447 } 448 449 // sort baselined mmap'd records in callsite pc order 450 int MemBaseline::bl_vm_sort_by_pc(const void* p1, const void* p2) { 451 assert(MemTracker::is_on(), "Just check"); 452 const VMCallsitePointer* mp1 = (const VMCallsitePointer*)p1; 453 const VMCallsitePointer* mp2 = (const VMCallsitePointer*)p2; 454 return UNSIGNED_COMPARE(mp1->addr(), mp2->addr()); 455 } 456 457 458 // sort snapshot malloc'd records in memory block address order 459 int MemBaseline::malloc_sort_by_addr(const void* p1, const void* p2) { 460 assert(MemTracker::is_on(), "Just check"); 461 const MemPointerRecord* mp1 = (const MemPointerRecord*)p1; 462 const MemPointerRecord* mp2 = (const MemPointerRecord*)p2; 463 int delta = UNSIGNED_COMPARE(mp1->addr(), mp2->addr()); 464 assert(delta != 0, "dup pointer"); 465 return delta; 466 } 467