1 /* 2 * Copyright (c) 2017, Google and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #ifndef SHARE_VM_RUNTIME_HEAPMONITORING_HPP 26 #define SHARE_VM_RUNTIME_HEAPMONITORING_HPP 27 28 #include "gc/shared/referenceProcessor.hpp" 29 #include "runtime/sharedRuntime.hpp" 30 31 // Support class for sampling heap allocations across the VM. 32 class HeapMonitoring : AllStatic { 33 private: 34 // Cheap random number generator 35 static uint64_t _rnd; 36 static bool _initialized; 37 static jint _monitoring_rate; 38 static bool _enabled; 39 40 // Statics for the fast log 41 static const int FastLogNumBits = 10; 42 static const int FastLogMask = (1 << FastLogNumBits) - 1; 43 static double _log_table[1<<FastLogNumBits]; // Constant 44 static AlwaysTrueClosure _always_true; 45 46 static void pick_next_sample(JavaThread *t); 47 48 // Returns the next prng value. 49 // pRNG is: aX+b mod c with a = 0x5DEECE66D, b = 0xB, c = 1<<48 50 // This is the lrand64 generator. 51 static inline uint64_t next_random(uint64_t rnd) { 52 const uint64_t PrngMult = 0x5DEECE66DLL; 53 const uint64_t prng_add = 0xB; 54 const uint64_t prng_mod_power = 48; 55 const uint64_t prng_mod_mask = 56 ~((~static_cast<uint64_t>(0)) << prng_mod_power); 57 return (PrngMult * rnd + prng_add) & prng_mod_mask; 58 } 59 60 static inline double fast_log2(const double & d) { 61 assert(d>0, "bad value passed to assert"); 62 uint64_t x = 0; 63 memcpy(&x, &d, sizeof(uint64_t)); 64 const uint32_t x_high = x >> 32; 65 const uint32_t y = x_high >> (20 - FastLogNumBits) & FastLogMask; 66 const int32_t exponent = ((x_high >> 20) & 0x7FF) - 1023; 67 return exponent + _log_table[y]; 68 } 69 70 public: 71 static void pick_next_sample(size_t *ptr); 72 73 static void get_live_traces(jvmtiStackTraces* stack_traces); 74 static void get_sampling_statistics(jvmtiHeapSamplingStats* stats); 75 static void get_garbage_traces(jvmtiStackTraces* stack_traces); 76 static void get_frequent_garbage_traces(jvmtiStackTraces* stack_traces); 77 static void release_traces(jvmtiStackTraces *trace_info); 78 static void initialize_profiling(jint monitoring_rate, jint max_storage); 79 static void stop_profiling(); 80 static bool initialized(); 81 static bool *initialized_address(); 82 83 // Called when o is allocated, called by interpreter and C1. 84 static void object_alloc_unsized(oopDesc* o); 85 static void object_alloc(oopDesc* o, intx byte_size); 86 87 // Called when o is allocated from C2 directly, 88 // we know the thread, and we have done the sampling. 89 static void object_alloc_do_sample(Thread *t, oopDesc *o, intx size_in_bytes); 90 91 // Called to clean up oops that have been saved by our sampling function, 92 // but which no longer have other references in the heap. 93 static size_t weak_oops_do(BoolObjectClosure* is_alive, 94 OopClosure *f); 95 static size_t weak_oops_do(OopClosure* oop_closure) { 96 return weak_oops_do(&_always_true, oop_closure); 97 } 98 99 static bool enabled() { 100 return _enabled; 101 } 102 }; 103 104 #endif // SHARE_VM_RUNTIME_HEAPMONITORING_HPP