1 /* 2 * Copyright (c) 2013, 2014, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #ifndef SHARE_VM_SERVICES_VIRTUAL_MEMORY_TRACKER_HPP 26 #define SHARE_VM_SERVICES_VIRTUAL_MEMORY_TRACKER_HPP 27 28 #if INCLUDE_NMT 29 30 #include "memory/allocation.hpp" 31 #include "services/allocationSite.hpp" 32 #include "services/nmtCommon.hpp" 33 #include "utilities/linkedlist.hpp" 34 #include "utilities/nativeCallStack.hpp" 35 #include "utilities/ostream.hpp" 36 37 38 /* 39 * Virtual memory counter 40 */ 41 class VirtualMemory VALUE_OBJ_CLASS_SPEC { 42 private: 43 size_t _reserved; 44 size_t _committed; 45 46 public: 47 VirtualMemory() : _reserved(0), _committed(0) { } 48 49 inline void reserve_memory(size_t sz) { _reserved += sz; } 50 inline void commit_memory (size_t sz) { 51 _committed += sz; 52 assert(_committed <= _reserved, "Sanity check"); 53 } 54 55 inline void release_memory (size_t sz) { 56 assert(_reserved >= sz, "Negative amount"); 57 _reserved -= sz; 58 } 59 60 inline void uncommit_memory(size_t sz) { 61 assert(_committed >= sz, "Negative amount"); 62 _committed -= sz; 63 } 64 65 void reset() { 66 _reserved = 0; 67 _committed = 0; 68 } 69 70 inline size_t reserved() const { return _reserved; } 71 inline size_t committed() const { return _committed; } 72 }; 73 74 // Virtual memory allocation site, keeps track where the virtual memory is reserved. 75 class VirtualMemoryAllocationSite : public AllocationSite<VirtualMemory> { 76 public: 77 VirtualMemoryAllocationSite(const NativeCallStack& stack) : 78 AllocationSite<VirtualMemory>(stack) { } 79 80 inline void reserve_memory(size_t sz) { data()->reserve_memory(sz); } 81 inline void commit_memory (size_t sz) { data()->commit_memory(sz); } 82 inline void uncommit_memory(size_t sz) { data()->uncommit_memory(sz); } 83 inline void release_memory(size_t sz) { data()->release_memory(sz); } 84 inline size_t reserved() const { return peek()->reserved(); } 85 inline size_t committed() const { return peek()->committed(); } 86 }; 87 88 class VirtualMemorySummary; 89 90 // This class represents a snapshot of virtual memory at a given time. 91 // The latest snapshot is saved in a static area. 92 class VirtualMemorySnapshot : public ResourceObj { 93 friend class VirtualMemorySummary; 94 95 private: 96 VirtualMemory _virtual_memory[mt_number_of_types]; 97 98 public: 99 inline VirtualMemory* by_type(MEMFLAGS flag) { 100 int index = NMTUtil::flag_to_index(flag); 101 return &_virtual_memory[index]; 102 } 103 104 inline VirtualMemory* by_index(int index) { 105 assert(index >= 0, "Index out of bound"); 106 assert(index < mt_number_of_types, "Index out of bound"); 107 return &_virtual_memory[index]; 108 } 109 110 inline size_t total_reserved() const { 111 size_t amount = 0; 112 for (int index = 0; index < mt_number_of_types; index ++) { 113 amount += _virtual_memory[index].reserved(); 114 } 115 return amount; 116 } 117 118 inline size_t total_committed() const { 119 size_t amount = 0; 120 for (int index = 0; index < mt_number_of_types; index ++) { 121 amount += _virtual_memory[index].committed(); 122 } 123 return amount; 124 } 125 126 inline void reset() { 127 for (int index = 0; index < mt_number_of_types; index ++) { 128 _virtual_memory[index].reset(); 129 } 130 } 131 132 void copy_to(VirtualMemorySnapshot* s) { 133 for (int index = 0; index < mt_number_of_types; index ++) { 134 s->_virtual_memory[index] = _virtual_memory[index]; 135 } 136 } 137 }; 138 139 class VirtualMemorySummary : AllStatic { 140 public: 141 static void initialize(); 142 143 static inline void record_reserved_memory(size_t size, MEMFLAGS flag) { 144 as_snapshot()->by_type(flag)->reserve_memory(size); 145 } 146 147 static inline void record_committed_memory(size_t size, MEMFLAGS flag) { 148 as_snapshot()->by_type(flag)->commit_memory(size); 149 } 150 151 static inline void record_uncommitted_memory(size_t size, MEMFLAGS flag) { 152 as_snapshot()->by_type(flag)->uncommit_memory(size); 153 } 154 155 static inline void record_released_memory(size_t size, MEMFLAGS flag) { 156 as_snapshot()->by_type(flag)->release_memory(size); 157 } 158 159 // Move virtual memory from one memory type to another. 160 // Virtual memory can be reserved before it is associated with a memory type, and tagged 161 // as 'unknown'. Once the memory is tagged, the virtual memory will be moved from 'unknown' 162 // type to specified memory type. 163 static inline void move_reserved_memory(MEMFLAGS from, MEMFLAGS to, size_t size) { 164 as_snapshot()->by_type(from)->release_memory(size); 165 as_snapshot()->by_type(to)->reserve_memory(size); 166 } 167 168 static inline void move_committed_memory(MEMFLAGS from, MEMFLAGS to, size_t size) { 169 as_snapshot()->by_type(from)->uncommit_memory(size); 170 as_snapshot()->by_type(to)->commit_memory(size); 171 } 172 173 static inline void snapshot(VirtualMemorySnapshot* s) { 174 as_snapshot()->copy_to(s); 175 } 176 177 static inline void reset() { 178 as_snapshot()->reset(); 179 } 180 181 static VirtualMemorySnapshot* as_snapshot() { 182 return (VirtualMemorySnapshot*)_snapshot; 183 } 184 185 private: 186 static size_t _snapshot[CALC_OBJ_SIZE_IN_TYPE(VirtualMemorySnapshot, size_t)]; 187 }; 188 189 190 191 /* 192 * A virtual memory region 193 */ 194 class VirtualMemoryRegion VALUE_OBJ_CLASS_SPEC { 195 private: 196 address _base_address; 197 size_t _size; 198 199 public: 200 VirtualMemoryRegion(address addr, size_t size) : 201 _base_address(addr), _size(size) { 202 assert(addr != NULL, "Invalid address"); 203 assert(size > 0, "Invalid size"); 204 } 205 206 inline address base() const { return _base_address; } 207 inline address end() const { return base() + size(); } 208 inline size_t size() const { return _size; } 209 210 inline bool is_empty() const { return size() == 0; } 211 212 inline bool contain_address(address addr) const { 213 return (addr >= base() && addr < end()); 214 } 215 216 217 inline bool contain_region(address addr, size_t size) const { 218 return contain_address(addr) && contain_address(addr + size - 1); 219 } 220 221 inline bool same_region(address addr, size_t sz) const { 222 return (addr == base() && sz == size()); 223 } 224 225 226 inline bool overlap_region(address addr, size_t sz) const { 227 VirtualMemoryRegion rgn(addr, sz); 228 return contain_address(addr) || 229 contain_address(addr + sz - 1) || 230 rgn.contain_address(base()) || 231 rgn.contain_address(end() - 1); 232 } 233 234 inline bool adjacent_to(address addr, size_t sz) const { 235 return (addr == end() || (addr + sz) == base()); 236 } 237 238 void exclude_region(address addr, size_t sz) { 239 assert(contain_region(addr, sz), "Not containment"); 240 assert(addr == base() || addr + sz == end(), "Can not exclude from middle"); 241 size_t new_size = size() - sz; 242 243 if (addr == base()) { 244 set_base(addr + sz); 245 } 246 set_size(new_size); 247 } 248 249 void expand_region(address addr, size_t sz) { 250 assert(adjacent_to(addr, sz), "Not adjacent regions"); 251 if (base() == addr + sz) { 252 set_base(addr); 253 } 254 set_size(size() + sz); 255 } 256 257 protected: 258 void set_base(address base) { 259 assert(base != NULL, "Sanity check"); 260 _base_address = base; 261 } 262 263 void set_size(size_t size) { 264 assert(size > 0, "Sanity check"); 265 _size = size; 266 } 267 }; 268 269 270 class CommittedMemoryRegion : public VirtualMemoryRegion { 271 private: 272 NativeCallStack _stack; 273 274 public: 275 CommittedMemoryRegion(address addr, size_t size, const NativeCallStack& stack) : 276 VirtualMemoryRegion(addr, size), _stack(stack) { } 277 278 inline int compare(const CommittedMemoryRegion& rgn) const { 279 if (overlap_region(rgn.base(), rgn.size()) || 280 adjacent_to (rgn.base(), rgn.size())) { 281 return 0; 282 } else { 283 if (base() == rgn.base()) { 284 return 0; 285 } else if (base() > rgn.base()) { 286 return 1; 287 } else { 288 return -1; 289 } 290 } 291 } 292 293 inline bool equals(const CommittedMemoryRegion& rgn) const { 294 return compare(rgn) == 0; 295 } 296 297 inline void set_call_stack(const NativeCallStack& stack) { _stack = stack; } 298 inline const NativeCallStack* call_stack() const { return &_stack; } 299 }; 300 301 302 typedef LinkedListIterator<CommittedMemoryRegion> CommittedRegionIterator; 303 304 int compare_committed_region(const CommittedMemoryRegion&, const CommittedMemoryRegion&); 305 class ReservedMemoryRegion : public VirtualMemoryRegion { 306 private: 307 SortedLinkedList<CommittedMemoryRegion, compare_committed_region> 308 _committed_regions; 309 310 NativeCallStack _stack; 311 MEMFLAGS _flag; 312 313 bool _all_committed; 314 315 public: 316 ReservedMemoryRegion(address base, size_t size, const NativeCallStack& stack, 317 MEMFLAGS flag = mtNone) : 318 VirtualMemoryRegion(base, size), _stack(stack), _flag(flag), 319 _all_committed(false) { } 320 321 322 ReservedMemoryRegion(address base, size_t size) : 323 VirtualMemoryRegion(base, size), _stack(emptyStack), _flag(mtNone), 324 _all_committed(false) { } 325 326 // Copy constructor 327 ReservedMemoryRegion(const ReservedMemoryRegion& rr) : 328 VirtualMemoryRegion(rr.base(), rr.size()) { 329 *this = rr; 330 } 331 332 inline void set_call_stack(const NativeCallStack& stack) { _stack = stack; } 333 inline const NativeCallStack* call_stack() const { return &_stack; } 334 335 void set_flag(MEMFLAGS flag); 336 inline MEMFLAGS flag() const { return _flag; } 337 338 inline int compare(const ReservedMemoryRegion& rgn) const { 339 if (overlap_region(rgn.base(), rgn.size())) { 340 return 0; 341 } else { 342 if (base() == rgn.base()) { 343 return 0; 344 } else if (base() > rgn.base()) { 345 return 1; 346 } else { 347 return -1; 348 } 349 } 350 } 351 352 inline bool equals(const ReservedMemoryRegion& rgn) const { 353 return compare(rgn) == 0; 354 } 355 356 bool add_committed_region(address addr, size_t size, const NativeCallStack& stack); 357 bool remove_uncommitted_region(address addr, size_t size); 358 359 size_t committed_size() const; 360 361 // move committed regions that higher than specified address to 362 // the new region 363 void move_committed_regions(address addr, ReservedMemoryRegion& rgn); 364 365 inline bool all_committed() const { return _all_committed; } 366 void set_all_committed(bool b); 367 368 CommittedRegionIterator iterate_committed_regions() const { 369 return CommittedRegionIterator(_committed_regions.head()); 370 } 371 372 ReservedMemoryRegion& operator= (const ReservedMemoryRegion& other) { 373 set_base(other.base()); 374 set_size(other.size()); 375 376 _stack = *other.call_stack(); 377 _flag = other.flag(); 378 _all_committed = other.all_committed(); 379 if (other.all_committed()) { 380 set_all_committed(true); 381 } else { 382 CommittedRegionIterator itr = other.iterate_committed_regions(); 383 const CommittedMemoryRegion* rgn = itr.next(); 384 while (rgn != NULL) { 385 _committed_regions.add(*rgn); 386 rgn = itr.next(); 387 } 388 } 389 return *this; 390 } 391 392 private: 393 // The committed region contains the uncommitted region, subtract the uncommitted 394 // region from this committed region 395 bool remove_uncommitted_region(LinkedListNode<CommittedMemoryRegion>* node, 396 address addr, size_t sz); 397 398 bool add_committed_region(const CommittedMemoryRegion& rgn) { 399 assert(rgn.base() != NULL, "Invalid base address"); 400 assert(size() > 0, "Invalid size"); 401 return _committed_regions.add(rgn) != NULL; 402 } 403 }; 404 405 int compare_reserved_region_base(const ReservedMemoryRegion& r1, const ReservedMemoryRegion& r2); 406 407 class VirtualMemoryWalker : public StackObj { 408 public: 409 virtual bool do_allocation_site(const ReservedMemoryRegion* rgn) { return false; } 410 }; 411 412 // Main class called from MemTracker to track virtual memory allocations, commits and releases. 413 class VirtualMemoryTracker : AllStatic { 414 public: 415 static bool initialize(NMT_TrackingLevel level); 416 417 // Late phase initialization 418 static bool late_initialize(NMT_TrackingLevel level); 419 420 static bool add_reserved_region (address base_addr, size_t size, const NativeCallStack& stack, 421 MEMFLAGS flag = mtNone, bool all_committed = false); 422 423 static bool add_committed_region (address base_addr, size_t size, const NativeCallStack& stack); 424 static bool remove_uncommitted_region (address base_addr, size_t size); 425 static bool remove_released_region (address base_addr, size_t size); 426 static void set_reserved_region_type (address addr, MEMFLAGS flag); 427 428 // Walk virtual memory data structure for creating baseline, etc. 429 static bool walk_virtual_memory(VirtualMemoryWalker* walker); 430 431 static bool transition(NMT_TrackingLevel from, NMT_TrackingLevel to); 432 433 private: 434 static SortedLinkedList<ReservedMemoryRegion, compare_reserved_region_base>* _reserved_regions; 435 }; 436 437 438 #endif // INCLUDE_NMT 439 440 #endif // SHARE_VM_SERVICES_VIRTUAL_MEMORY_TRACKER_HPP