/* * Copyright (c) 2012, 2018, Oracle and/or its affiliates. All rights reserved. * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. * * This code is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License version 2 only, as * published by the Free Software Foundation. * * This code is distributed in the hope that it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License * version 2 for more details (a copy is included in the LICENSE file that * accompanied this code). * * You should have received a copy of the GNU General Public License version * 2 along with this work; if not, write to the Free Software Foundation, * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. * * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA * or visit www.oracle.com if you need additional information or have any * questions. * */ // A ClassLoaderData identifies the full set of class types that a class // loader's name resolution strategy produces for a given configuration of the // class loader. // Class types in the ClassLoaderData may be defined by from class file binaries // provided by the class loader, or from other class loader it interacts with // according to its name resolution strategy. // // Class loaders that implement a deterministic name resolution strategy // (including with respect to their delegation behavior), such as the boot, the // platform, and the system loaders of the JDK's built-in class loader // hierarchy, always produce the same linkset for a given configuration. // // ClassLoaderData carries information related to a linkset (e.g., // metaspace holding its klass definitions). // The System Dictionary and related data structures (e.g., placeholder table, // loader constraints table) as well as the runtime representation of classes // only reference ClassLoaderData. // // Instances of java.lang.ClassLoader holds a pointer to a ClassLoaderData that // that represent the loader's "linking domain" in the JVM. // // The bootstrap loader (represented by NULL) also has a ClassLoaderData, // the singleton class the_null_class_loader_data(). #include "precompiled.hpp" #include "classfile/classLoaderData.hpp" #include "classfile/classLoaderData.inline.hpp" #include "classfile/dictionary.hpp" #include "classfile/javaClasses.hpp" #include "classfile/metadataOnStackMark.hpp" #include "classfile/moduleEntry.hpp" #include "classfile/packageEntry.hpp" #include "classfile/symbolTable.hpp" #include "classfile/systemDictionary.hpp" #include "logging/log.hpp" #include "logging/logStream.hpp" #include "memory/allocation.inline.hpp" #include "memory/metadataFactory.hpp" #include "memory/metaspaceShared.hpp" #include "memory/resourceArea.hpp" #include "memory/universe.hpp" #include "oops/access.inline.hpp" #include "oops/oop.inline.hpp" #include "oops/oopHandle.inline.hpp" #include "oops/weakHandle.inline.hpp" #include "runtime/atomic.hpp" #include "runtime/handles.inline.hpp" #include "runtime/mutex.hpp" #include "runtime/orderAccess.hpp" #include "runtime/safepoint.hpp" #include "runtime/safepointVerifiers.hpp" #include "utilities/growableArray.hpp" #include "utilities/macros.hpp" #include "utilities/ostream.hpp" volatile size_t ClassLoaderDataGraph::_num_array_classes = 0; volatile size_t ClassLoaderDataGraph::_num_instance_classes = 0; ClassLoaderData * ClassLoaderData::_the_null_class_loader_data = NULL; void ClassLoaderData::init_null_class_loader_data() { assert(_the_null_class_loader_data == NULL, "cannot initialize twice"); assert(ClassLoaderDataGraph::_head == NULL, "cannot initialize twice"); _the_null_class_loader_data = new ClassLoaderData(Handle(), false); ClassLoaderDataGraph::_head = _the_null_class_loader_data; assert(_the_null_class_loader_data->is_the_null_class_loader_data(), "Must be"); LogTarget(Trace, class, loader, data) lt; if (lt.is_enabled()) { ResourceMark rm; LogStream ls(lt); ls.print("create "); _the_null_class_loader_data->print_value_on(&ls); ls.cr(); } } // Obtain and set the class loader's name within the ClassLoaderData so // it will be available for error messages, logging, JFR, etc. The name // and klass are available after the class_loader oop is no longer alive, // during unloading. void ClassLoaderData::initialize_name(Handle class_loader) { Thread* THREAD = Thread::current(); ResourceMark rm(THREAD); // Obtain the class loader's name. If the class loader's name was not // explicitly set during construction, the CLD's _name field will be null. oop cl_name = java_lang_ClassLoader::name(class_loader()); if (cl_name != NULL) { const char* cl_instance_name = java_lang_String::as_utf8_string(cl_name); if (cl_instance_name != NULL && cl_instance_name[0] != '\0') { // Can't throw InternalError and SymbolTable doesn't throw OOM anymore. _name = SymbolTable::new_symbol(cl_instance_name, CATCH); } } // Obtain the class loader's name and identity hash. If the class loader's // name was not explicitly set during construction, the class loader's name and id // will be set to the qualified class name of the class loader along with its // identity hash. // If for some reason the ClassLoader's constructor has not been run, instead of // leaving the _name_and_id field null, fall back to the external qualified class // name. Thus CLD's _name_and_id field should never have a null value. oop cl_name_and_id = java_lang_ClassLoader::nameAndId(class_loader()); const char* cl_instance_name_and_id = (cl_name_and_id == NULL) ? _class_loader_klass->external_name() : java_lang_String::as_utf8_string(cl_name_and_id); assert(cl_instance_name_and_id != NULL && cl_instance_name_and_id[0] != '\0', "class loader has no name and id"); // Can't throw InternalError and SymbolTable doesn't throw OOM anymore. _name_and_id = SymbolTable::new_symbol(cl_instance_name_and_id, CATCH); } ClassLoaderData::ClassLoaderData(Handle h_class_loader, bool is_unsafe_anonymous) : _metaspace(NULL), _metaspace_lock(new Mutex(Monitor::leaf+1, "Metaspace allocation lock", true, Monitor::_safepoint_check_never)), _unloading(false), _is_unsafe_anonymous(is_unsafe_anonymous), _modified_oops(true), _accumulated_modified_oops(false), // An unsafe anonymous class loader data doesn't have anything to keep // it from being unloaded during parsing of the unsafe anonymous class. // The null-class-loader should always be kept alive. _keep_alive((is_unsafe_anonymous || h_class_loader.is_null()) ? 1 : 0), _claim(0), _handles(), _klasses(NULL), _packages(NULL), _modules(NULL), _unnamed_module(NULL), _dictionary(NULL), _jmethod_ids(NULL), _deallocate_list(NULL), _next(NULL), _class_loader_klass(NULL), _name(NULL), _name_and_id(NULL) { if (!h_class_loader.is_null()) { _class_loader = _handles.add(h_class_loader()); _class_loader_klass = h_class_loader->klass(); } if (!is_unsafe_anonymous) { // The holder is initialized later for unsafe anonymous classes, and before calling anything // that call class_loader(). initialize_holder(h_class_loader); // A ClassLoaderData created solely for an unsafe anonymous class should never have a // ModuleEntryTable or PackageEntryTable created for it. The defining package // and module for an unsafe anonymous class will be found in its host class. _packages = new PackageEntryTable(PackageEntryTable::_packagetable_entry_size); if (h_class_loader.is_null()) { // Create unnamed module for boot loader _unnamed_module = ModuleEntry::create_boot_unnamed_module(this); } else { // Create unnamed module for all other loaders _unnamed_module = ModuleEntry::create_unnamed_module(this); } _dictionary = create_dictionary(); } NOT_PRODUCT(_dependency_count = 0); // number of class loader dependencies JFR_ONLY(INIT_ID(this);) } ClassLoaderData::ChunkedHandleList::~ChunkedHandleList() { Chunk* c = _head; while (c != NULL) { Chunk* next = c->_next; delete c; c = next; } } oop* ClassLoaderData::ChunkedHandleList::add(oop o) { if (_head == NULL || _head->_size == Chunk::CAPACITY) { Chunk* next = new Chunk(_head); OrderAccess::release_store(&_head, next); } oop* handle = &_head->_data[_head->_size]; *handle = o; OrderAccess::release_store(&_head->_size, _head->_size + 1); return handle; } int ClassLoaderData::ChunkedHandleList::count() const { int count = 0; Chunk* chunk = _head; while (chunk != NULL) { count += chunk->_size; chunk = chunk->_next; } return count; } inline void ClassLoaderData::ChunkedHandleList::oops_do_chunk(OopClosure* f, Chunk* c, const juint size) { for (juint i = 0; i < size; i++) { if (c->_data[i] != NULL) { f->do_oop(&c->_data[i]); } } } void ClassLoaderData::ChunkedHandleList::oops_do(OopClosure* f) { Chunk* head = OrderAccess::load_acquire(&_head); if (head != NULL) { // Must be careful when reading size of head oops_do_chunk(f, head, OrderAccess::load_acquire(&head->_size)); for (Chunk* c = head->_next; c != NULL; c = c->_next) { oops_do_chunk(f, c, c->_size); } } } class VerifyContainsOopClosure : public OopClosure { oop _target; bool _found; public: VerifyContainsOopClosure(oop target) : _target(target), _found(false) {} void do_oop(oop* p) { if (p != NULL && oopDesc::equals(RawAccess<>::oop_load(p), _target)) { _found = true; } } void do_oop(narrowOop* p) { // The ChunkedHandleList should not contain any narrowOop ShouldNotReachHere(); } bool found() const { return _found; } }; bool ClassLoaderData::ChunkedHandleList::contains(oop p) { VerifyContainsOopClosure cl(p); oops_do(&cl); return cl.found(); } #ifndef PRODUCT bool ClassLoaderData::ChunkedHandleList::owner_of(oop* oop_handle) { Chunk* chunk = _head; while (chunk != NULL) { if (&(chunk->_data[0]) <= oop_handle && oop_handle < &(chunk->_data[chunk->_size])) { return true; } chunk = chunk->_next; } return false; } #endif // PRODUCT bool ClassLoaderData::try_claim(int claim) { for (;;) { int old_claim = Atomic::load(&_claim); if ((old_claim & claim) == claim) { return false; } int new_claim = old_claim | claim; if (Atomic::cmpxchg(new_claim, &_claim, old_claim) == old_claim) { return true; } } } // Unsafe anonymous classes have their own ClassLoaderData that is marked to keep alive // while the class is being parsed, and if the class appears on the module fixup list. // Due to the uniqueness that no other class shares the unsafe anonymous class' name or // ClassLoaderData, no other non-GC thread has knowledge of the unsafe anonymous class while // it is being defined, therefore _keep_alive is not volatile or atomic. void ClassLoaderData::inc_keep_alive() { if (is_unsafe_anonymous()) { assert(_keep_alive >= 0, "Invalid keep alive increment count"); _keep_alive++; } } void ClassLoaderData::dec_keep_alive() { if (is_unsafe_anonymous()) { assert(_keep_alive > 0, "Invalid keep alive decrement count"); _keep_alive--; } } void ClassLoaderData::oops_do(OopClosure* f, int claim_value, bool clear_mod_oops) { if (claim_value != ClassLoaderData::_claim_none && !try_claim(claim_value)) { return; } // Only clear modified_oops after the ClassLoaderData is claimed. if (clear_mod_oops) { clear_modified_oops(); } _handles.oops_do(f); } void ClassLoaderData::classes_do(KlassClosure* klass_closure) { // Lock-free access requires load_acquire for (Klass* k = OrderAccess::load_acquire(&_klasses); k != NULL; k = k->next_link()) { klass_closure->do_klass(k); assert(k != k->next_link(), "no loops!"); } } void ClassLoaderData::classes_do(void f(Klass * const)) { // Lock-free access requires load_acquire for (Klass* k = OrderAccess::load_acquire(&_klasses); k != NULL; k = k->next_link()) { f(k); assert(k != k->next_link(), "no loops!"); } } void ClassLoaderData::methods_do(void f(Method*)) { // Lock-free access requires load_acquire for (Klass* k = OrderAccess::load_acquire(&_klasses); k != NULL; k = k->next_link()) { if (k->is_instance_klass() && InstanceKlass::cast(k)->is_loaded()) { InstanceKlass::cast(k)->methods_do(f); } } } void ClassLoaderData::loaded_classes_do(KlassClosure* klass_closure) { // Lock-free access requires load_acquire for (Klass* k = OrderAccess::load_acquire(&_klasses); k != NULL; k = k->next_link()) { // Do not filter ArrayKlass oops here... if (k->is_array_klass() || (k->is_instance_klass() && InstanceKlass::cast(k)->is_loaded())) { klass_closure->do_klass(k); } } } void ClassLoaderData::classes_do(void f(InstanceKlass*)) { // Lock-free access requires load_acquire for (Klass* k = OrderAccess::load_acquire(&_klasses); k != NULL; k = k->next_link()) { if (k->is_instance_klass()) { f(InstanceKlass::cast(k)); } assert(k != k->next_link(), "no loops!"); } } void ClassLoaderData::modules_do(void f(ModuleEntry*)) { assert_locked_or_safepoint(Module_lock); if (_unnamed_module != NULL) { f(_unnamed_module); } if (_modules != NULL) { for (int i = 0; i < _modules->table_size(); i++) { for (ModuleEntry* entry = _modules->bucket(i); entry != NULL; entry = entry->next()) { f(entry); } } } } void ClassLoaderData::packages_do(void f(PackageEntry*)) { assert_locked_or_safepoint(Module_lock); if (_packages != NULL) { for (int i = 0; i < _packages->table_size(); i++) { for (PackageEntry* entry = _packages->bucket(i); entry != NULL; entry = entry->next()) { f(entry); } } } } void ClassLoaderData::record_dependency(const Klass* k) { assert(k != NULL, "invariant"); ClassLoaderData * const from_cld = this; ClassLoaderData * const to_cld = k->class_loader_data(); // Do not need to record dependency if the dependency is to a class whose // class loader data is never freed. (i.e. the dependency's class loader // is one of the three builtin class loaders and the dependency is not // unsafe anonymous.) if (to_cld->is_permanent_class_loader_data()) { return; } oop to; if (to_cld->is_unsafe_anonymous()) { // Just return if an unsafe anonymous class is attempting to record a dependency // to itself. (Note that every unsafe anonymous class has its own unique class // loader data.) if (to_cld == from_cld) { return; } // Unsafe anonymous class dependencies are through the mirror. to = k->java_mirror(); } else { to = to_cld->class_loader(); oop from = from_cld->class_loader(); // Just return if this dependency is to a class with the same or a parent // class_loader. if (oopDesc::equals(from, to) || java_lang_ClassLoader::isAncestor(from, to)) { return; // this class loader is in the parent list, no need to add it. } } // It's a dependency we won't find through GC, add it. if (!_handles.contains(to)) { NOT_PRODUCT(Atomic::inc(&_dependency_count)); LogTarget(Trace, class, loader, data) lt; if (lt.is_enabled()) { ResourceMark rm; LogStream ls(lt); ls.print("adding dependency from "); print_value_on(&ls); ls.print(" to "); to_cld->print_value_on(&ls); ls.cr(); } Handle dependency(Thread::current(), to); add_handle(dependency); // Added a potentially young gen oop to the ClassLoaderData record_modified_oops(); } } void ClassLoaderDataGraph::clear_claimed_marks() { for (ClassLoaderData* cld = _head; cld != NULL; cld = cld->next()) { cld->clear_claim(); } } void ClassLoaderData::add_class(Klass* k, bool publicize /* true */) { { MutexLockerEx ml(metaspace_lock(), Mutex::_no_safepoint_check_flag); Klass* old_value = _klasses; k->set_next_link(old_value); // Link the new item into the list, making sure the linked class is stable // since the list can be walked without a lock OrderAccess::release_store(&_klasses, k); if (k->is_array_klass()) { ClassLoaderDataGraph::inc_array_classes(1); } else { ClassLoaderDataGraph::inc_instance_classes(1); } } if (publicize) { LogTarget(Trace, class, loader, data) lt; if (lt.is_enabled()) { ResourceMark rm; LogStream ls(lt); ls.print("Adding k: " PTR_FORMAT " %s to ", p2i(k), k->external_name()); print_value_on(&ls); ls.cr(); } } } // Class iterator used by the compiler. It gets some number of classes at // a safepoint to decay invocation counters on the methods. class ClassLoaderDataGraphKlassIteratorStatic { ClassLoaderData* _current_loader_data; Klass* _current_class_entry; public: ClassLoaderDataGraphKlassIteratorStatic() : _current_loader_data(NULL), _current_class_entry(NULL) {} InstanceKlass* try_get_next_class() { assert(SafepointSynchronize::is_at_safepoint(), "only called at safepoint"); size_t max_classes = ClassLoaderDataGraph::num_instance_classes(); assert(max_classes > 0, "should not be called with no instance classes"); for (size_t i = 0; i < max_classes; ) { if (_current_class_entry != NULL) { Klass* k = _current_class_entry; _current_class_entry = _current_class_entry->next_link(); if (k->is_instance_klass()) { InstanceKlass* ik = InstanceKlass::cast(k); i++; // count all instance classes found // Not yet loaded classes are counted in max_classes // but only return loaded classes. if (ik->is_loaded()) { return ik; } } } else { // Go to next CLD if (_current_loader_data != NULL) { _current_loader_data = _current_loader_data->next(); } // Start at the beginning if (_current_loader_data == NULL) { _current_loader_data = ClassLoaderDataGraph::_head; } _current_class_entry = _current_loader_data->klasses(); } } // Should never be reached unless all instance classes have failed or are not fully loaded. // Caller handles NULL. return NULL; } // If the current class for the static iterator is a class being unloaded or // deallocated, adjust the current class. void adjust_saved_class(ClassLoaderData* cld) { if (_current_loader_data == cld) { _current_loader_data = cld->next(); if (_current_loader_data != NULL) { _current_class_entry = _current_loader_data->klasses(); } // else try_get_next_class will start at the head } } void adjust_saved_class(Klass* klass) { if (_current_class_entry == klass) { _current_class_entry = klass->next_link(); } } }; static ClassLoaderDataGraphKlassIteratorStatic static_klass_iterator; InstanceKlass* ClassLoaderDataGraph::try_get_next_class() { assert(SafepointSynchronize::is_at_safepoint(), "only called at safepoint"); return static_klass_iterator.try_get_next_class(); } void ClassLoaderData::initialize_holder(Handle loader_or_mirror) { if (loader_or_mirror() != NULL) { assert(_holder.is_null(), "never replace holders"); _holder = WeakHandle::create(loader_or_mirror); } } // Remove a klass from the _klasses list for scratch_class during redefinition // or parsed class in the case of an error. void ClassLoaderData::remove_class(Klass* scratch_class) { assert(SafepointSynchronize::is_at_safepoint(), "only called at safepoint"); // Adjust global class iterator. static_klass_iterator.adjust_saved_class(scratch_class); Klass* prev = NULL; for (Klass* k = _klasses; k != NULL; k = k->next_link()) { if (k == scratch_class) { if (prev == NULL) { _klasses = k->next_link(); } else { Klass* next = k->next_link(); prev->set_next_link(next); } if (k->is_array_klass()) { ClassLoaderDataGraph::dec_array_classes(1); } else { ClassLoaderDataGraph::dec_instance_classes(1); } return; } prev = k; assert(k != k->next_link(), "no loops!"); } ShouldNotReachHere(); // should have found this class!! } void ClassLoaderData::unload() { _unloading = true; LogTarget(Trace, class, loader, data) lt; if (lt.is_enabled()) { ResourceMark rm; LogStream ls(lt); ls.print("unload"); print_value_on(&ls); ls.cr(); } // Some items on the _deallocate_list need to free their C heap structures // if they are not already on the _klasses list. free_deallocate_list_C_heap_structures(); // Tell serviceability tools these classes are unloading // after erroneous classes are released. classes_do(InstanceKlass::notify_unload_class); // Clean up global class iterator for compiler static_klass_iterator.adjust_saved_class(this); } ModuleEntryTable* ClassLoaderData::modules() { // Lazily create the module entry table at first request. // Lock-free access requires load_acquire. ModuleEntryTable* modules = OrderAccess::load_acquire(&_modules); if (modules == NULL) { MutexLocker m1(Module_lock); // Check if _modules got allocated while we were waiting for this lock. if ((modules = _modules) == NULL) { modules = new ModuleEntryTable(ModuleEntryTable::_moduletable_entry_size); { MutexLockerEx m1(metaspace_lock(), Mutex::_no_safepoint_check_flag); // Ensure _modules is stable, since it is examined without a lock OrderAccess::release_store(&_modules, modules); } } } return modules; } const int _boot_loader_dictionary_size = 1009; const int _default_loader_dictionary_size = 107; Dictionary* ClassLoaderData::create_dictionary() { assert(!is_unsafe_anonymous(), "unsafe anonymous class loader data do not have a dictionary"); int size; bool resizable = false; if (_the_null_class_loader_data == NULL) { size = _boot_loader_dictionary_size; resizable = true; } else if (class_loader()->is_a(SystemDictionary::reflect_DelegatingClassLoader_klass())) { size = 1; // there's only one class in relection class loader and no initiated classes } else if (is_system_class_loader_data()) { size = _boot_loader_dictionary_size; resizable = true; } else { size = _default_loader_dictionary_size; resizable = true; } if (!DynamicallyResizeSystemDictionaries || DumpSharedSpaces) { resizable = false; } return new Dictionary(this, size, resizable); } // Tell the GC to keep this klass alive while iterating ClassLoaderDataGraph oop ClassLoaderData::holder_phantom() const { // A klass that was previously considered dead can be looked up in the // CLD/SD, and its _java_mirror or _class_loader can be stored in a root // or a reachable object making it alive again. The SATB part of G1 needs // to get notified about this potential resurrection, otherwise the marking // might not find the object. if (!_holder.is_null()) { // NULL class_loader return _holder.resolve(); } else { return NULL; } } // Let the GC read the holder without keeping it alive. oop ClassLoaderData::holder_no_keepalive() const { if (!_holder.is_null()) { // NULL class_loader return _holder.peek(); } else { return NULL; } } // Unloading support bool ClassLoaderData::is_alive() const { bool alive = keep_alive() // null class loader and incomplete unsafe anonymous klasses. || (_holder.peek() != NULL); // and not cleaned by the GC weak handle processing. return alive; } class ReleaseKlassClosure: public KlassClosure { private: size_t _instance_class_released; size_t _array_class_released; public: ReleaseKlassClosure() : _instance_class_released(0), _array_class_released(0) { } size_t instance_class_released() const { return _instance_class_released; } size_t array_class_released() const { return _array_class_released; } void do_klass(Klass* k) { if (k->is_array_klass()) { _array_class_released ++; } else { assert(k->is_instance_klass(), "Must be"); _instance_class_released ++; InstanceKlass::release_C_heap_structures(InstanceKlass::cast(k)); } } }; ClassLoaderData::~ClassLoaderData() { // Release C heap structures for all the classes. ReleaseKlassClosure cl; classes_do(&cl); ClassLoaderDataGraph::dec_array_classes(cl.array_class_released()); ClassLoaderDataGraph::dec_instance_classes(cl.instance_class_released()); // Release the WeakHandle _holder.release(); // Release C heap allocated hashtable for all the packages. if (_packages != NULL) { // Destroy the table itself delete _packages; _packages = NULL; } // Release C heap allocated hashtable for all the modules. if (_modules != NULL) { // Destroy the table itself delete _modules; _modules = NULL; } // Release C heap allocated hashtable for the dictionary if (_dictionary != NULL) { // Destroy the table itself delete _dictionary; _dictionary = NULL; } if (_unnamed_module != NULL) { _unnamed_module->delete_unnamed_module(); _unnamed_module = NULL; } // release the metaspace ClassLoaderMetaspace *m = _metaspace; if (m != NULL) { _metaspace = NULL; delete m; } // Clear all the JNI handles for methods // These aren't deallocated and are going to look like a leak, but that's // needed because we can't really get rid of jmethodIDs because we don't // know when native code is going to stop using them. The spec says that // they're "invalid" but existing programs likely rely on their being // NULL after class unloading. if (_jmethod_ids != NULL) { Method::clear_jmethod_ids(this); } // Delete lock delete _metaspace_lock; // Delete free list if (_deallocate_list != NULL) { delete _deallocate_list; } } // Returns true if this class loader data is for the app class loader // or a user defined system class loader. (Note that the class loader // data may be unsafe anonymous.) bool ClassLoaderData::is_system_class_loader_data() const { return SystemDictionary::is_system_class_loader(class_loader()); } // Returns true if this class loader data is for the platform class loader. // (Note that the class loader data may be unsafe anonymous.) bool ClassLoaderData::is_platform_class_loader_data() const { return SystemDictionary::is_platform_class_loader(class_loader()); } // Returns true if the class loader for this class loader data is one of // the 3 builtin (boot application/system or platform) class loaders, // including a user-defined system class loader. Note that if the class // loader data is for an unsafe anonymous class then it may get freed by a GC // even if its class loader is one of these loaders. bool ClassLoaderData::is_builtin_class_loader_data() const { return (is_boot_class_loader_data() || SystemDictionary::is_system_class_loader(class_loader()) || SystemDictionary::is_platform_class_loader(class_loader())); } // Returns true if this class loader data is a class loader data // that is not ever freed by a GC. It must be the CLD for one of the builtin // class loaders and not the CLD for an unsafe anonymous class. bool ClassLoaderData::is_permanent_class_loader_data() const { return is_builtin_class_loader_data() && !is_unsafe_anonymous(); } ClassLoaderMetaspace* ClassLoaderData::metaspace_non_null() { // If the metaspace has not been allocated, create a new one. Might want // to create smaller arena for Reflection class loaders also. // The reason for the delayed allocation is because some class loaders are // simply for delegating with no metadata of their own. // Lock-free access requires load_acquire. ClassLoaderMetaspace* metaspace = OrderAccess::load_acquire(&_metaspace); if (metaspace == NULL) { MutexLockerEx ml(_metaspace_lock, Mutex::_no_safepoint_check_flag); // Check if _metaspace got allocated while we were waiting for this lock. if ((metaspace = _metaspace) == NULL) { if (this == the_null_class_loader_data()) { assert (class_loader() == NULL, "Must be"); metaspace = new ClassLoaderMetaspace(_metaspace_lock, Metaspace::BootMetaspaceType); } else if (is_unsafe_anonymous()) { metaspace = new ClassLoaderMetaspace(_metaspace_lock, Metaspace::UnsafeAnonymousMetaspaceType); } else if (class_loader()->is_a(SystemDictionary::reflect_DelegatingClassLoader_klass())) { metaspace = new ClassLoaderMetaspace(_metaspace_lock, Metaspace::ReflectionMetaspaceType); } else { metaspace = new ClassLoaderMetaspace(_metaspace_lock, Metaspace::StandardMetaspaceType); } // Ensure _metaspace is stable, since it is examined without a lock OrderAccess::release_store(&_metaspace, metaspace); } } return metaspace; } OopHandle ClassLoaderData::add_handle(Handle h) { MutexLockerEx ml(metaspace_lock(), Mutex::_no_safepoint_check_flag); record_modified_oops(); return OopHandle(_handles.add(h())); } void ClassLoaderData::remove_handle(OopHandle h) { assert(!is_unloading(), "Do not remove a handle for a CLD that is unloading"); oop* ptr = h.ptr_raw(); if (ptr != NULL) { assert(_handles.owner_of(ptr), "Got unexpected handle " PTR_FORMAT, p2i(ptr)); NativeAccess<>::oop_store(ptr, oop(NULL)); } } void ClassLoaderData::init_handle_locked(OopHandle& dest, Handle h) { MutexLockerEx ml(metaspace_lock(), Mutex::_no_safepoint_check_flag); if (dest.resolve() != NULL) { return; } else { dest = _handles.add(h()); } } // Add this metadata pointer to be freed when it's safe. This is only during // a safepoint which checks if handles point to this metadata field. void ClassLoaderData::add_to_deallocate_list(Metadata* m) { // Metadata in shared region isn't deleted. if (!m->is_shared()) { MutexLockerEx ml(metaspace_lock(), Mutex::_no_safepoint_check_flag); if (_deallocate_list == NULL) { _deallocate_list = new (ResourceObj::C_HEAP, mtClass) GrowableArray(100, true); } _deallocate_list->append_if_missing(m); log_debug(class, loader, data)("deallocate added for %s", m->print_value_string()); ClassLoaderDataGraph::set_should_clean_deallocate_lists(); } } // Deallocate free metadata on the free list. How useful the PermGen was! void ClassLoaderData::free_deallocate_list() { // Don't need lock, at safepoint assert(SafepointSynchronize::is_at_safepoint(), "only called at safepoint"); assert(!is_unloading(), "only called for ClassLoaderData that are not unloading"); if (_deallocate_list == NULL) { return; } // Go backwards because this removes entries that are freed. for (int i = _deallocate_list->length() - 1; i >= 0; i--) { Metadata* m = _deallocate_list->at(i); if (!m->on_stack()) { _deallocate_list->remove_at(i); // There are only three types of metadata that we deallocate directly. // Cast them so they can be used by the template function. if (m->is_method()) { MetadataFactory::free_metadata(this, (Method*)m); } else if (m->is_constantPool()) { MetadataFactory::free_metadata(this, (ConstantPool*)m); } else if (m->is_klass()) { MetadataFactory::free_metadata(this, (InstanceKlass*)m); } else { ShouldNotReachHere(); } } else { // Metadata is alive. // If scratch_class is on stack then it shouldn't be on this list! assert(!m->is_klass() || !((InstanceKlass*)m)->is_scratch_class(), "scratch classes on this list should be dead"); // Also should assert that other metadata on the list was found in handles. // Some cleaning remains. ClassLoaderDataGraph::set_should_clean_deallocate_lists(); } } } void ClassLoaderDataGraph::clean_deallocate_lists(bool walk_previous_versions) { assert(SafepointSynchronize::is_at_safepoint(), "must only be called at safepoint"); uint loaders_processed = 0; for (ClassLoaderData* cld = _head; cld != NULL; cld = cld->next()) { // is_alive check will be necessary for concurrent class unloading. if (cld->is_alive()) { // clean metaspace if (walk_previous_versions) { cld->classes_do(InstanceKlass::purge_previous_versions); } cld->free_deallocate_list(); loaders_processed++; } } log_debug(class, loader, data)("clean_deallocate_lists: loaders processed %u %s", loaders_processed, walk_previous_versions ? "walk_previous_versions" : ""); } void ClassLoaderDataGraph::walk_metadata_and_clean_metaspaces() { assert(SafepointSynchronize::is_at_safepoint(), "must only be called at safepoint"); _should_clean_deallocate_lists = false; // assume everything gets cleaned // Mark metadata seen on the stack so we can delete unreferenced entries. // Walk all metadata, including the expensive code cache walk, only for class redefinition. // The MetadataOnStackMark walk during redefinition saves previous versions if it finds old methods // on the stack or in the code cache, so we only have to repeat the full walk if // they were found at that time. // TODO: have redefinition clean old methods out of the code cache. They still exist in some places. bool walk_all_metadata = InstanceKlass::has_previous_versions_and_reset(); MetadataOnStackMark md_on_stack(walk_all_metadata); clean_deallocate_lists(walk_all_metadata); } // This is distinct from free_deallocate_list. For class loader data that are // unloading, this frees the C heap memory for items on the list, and unlinks // scratch or error classes so that unloading events aren't triggered for these // classes. The metadata is removed with the unloading metaspace. // There isn't C heap memory allocated for methods, so nothing is done for them. void ClassLoaderData::free_deallocate_list_C_heap_structures() { // Don't need lock, at safepoint assert(SafepointSynchronize::is_at_safepoint(), "only called at safepoint"); assert(is_unloading(), "only called for ClassLoaderData that are unloading"); if (_deallocate_list == NULL) { return; } // Go backwards because this removes entries that are freed. for (int i = _deallocate_list->length() - 1; i >= 0; i--) { Metadata* m = _deallocate_list->at(i); _deallocate_list->remove_at(i); if (m->is_constantPool()) { ((ConstantPool*)m)->release_C_heap_structures(); } else if (m->is_klass()) { InstanceKlass* ik = (InstanceKlass*)m; // also releases ik->constants() C heap memory InstanceKlass::release_C_heap_structures(ik); // Remove the class so unloading events aren't triggered for // this class (scratch or error class) in do_unloading(). remove_class(ik); } } } // These CLDs are to contain unsafe anonymous classes used for JSR292 ClassLoaderData* ClassLoaderData::unsafe_anonymous_class_loader_data(Handle loader) { // Add a new class loader data to the graph. return ClassLoaderDataGraph::add(loader, true); } // Caller needs ResourceMark // If the class loader's _name has not been explicitly set, the class loader's // qualified class name is returned. const char* ClassLoaderData::loader_name() const { if (_class_loader_klass == NULL) { return BOOTSTRAP_LOADER_NAME; } else if (_name != NULL) { return _name->as_C_string(); } else { return _class_loader_klass->external_name(); } } // Caller needs ResourceMark // Format of the _name_and_id is as follows: // If the defining loader has a name explicitly set then '' @ // If the defining loader has no name then @ // If built-in loader, then omit '@' as there is only one instance. const char* ClassLoaderData::loader_name_and_id() const { if (_class_loader_klass == NULL) { return "'" BOOTSTRAP_LOADER_NAME "'"; } else if (_name_and_id != NULL) { return _name_and_id->as_C_string(); } else { // May be called in a race before _name_and_id is initialized. return _class_loader_klass->external_name(); } } void ClassLoaderData::print_value_on(outputStream* out) const { if (!is_unloading() && class_loader() != NULL) { out->print("loader data: " INTPTR_FORMAT " for instance ", p2i(this)); class_loader()->print_value_on(out); // includes loader_name_and_id() and address of class loader instance } else { // loader data: 0xsomeaddr of 'bootstrap' out->print("loader data: " INTPTR_FORMAT " of %s", p2i(this), loader_name_and_id()); } if (is_unsafe_anonymous()) { out->print(" unsafe anonymous"); } } #ifndef PRODUCT void ClassLoaderData::print_on(outputStream* out) const { out->print("ClassLoaderData CLD: " PTR_FORMAT ", loader: " PTR_FORMAT ", loader_klass: %s {", p2i(this), p2i(_class_loader.ptr_raw()), loader_name_and_id()); if (is_unsafe_anonymous()) out->print(" unsafe anonymous"); if (claimed()) out->print(" claimed"); if (is_unloading()) out->print(" unloading"); out->print(" metaspace: " INTPTR_FORMAT, p2i(metaspace_or_null())); if (_jmethod_ids != NULL) { Method::print_jmethod_ids(this, out); } out->print(" handles count %d", _handles.count()); out->print(" dependencies %d", _dependency_count); out->print_cr("}"); } #endif // PRODUCT void ClassLoaderData::verify() { assert_locked_or_safepoint(_metaspace_lock); oop cl = class_loader(); guarantee(this == class_loader_data(cl) || is_unsafe_anonymous(), "Must be the same"); guarantee(cl != NULL || this == ClassLoaderData::the_null_class_loader_data() || is_unsafe_anonymous(), "must be"); // Verify the integrity of the allocated space. if (metaspace_or_null() != NULL) { metaspace_or_null()->verify(); } for (Klass* k = _klasses; k != NULL; k = k->next_link()) { guarantee(k->class_loader_data() == this, "Must be the same"); k->verify(); assert(k != k->next_link(), "no loops!"); } } bool ClassLoaderData::contains_klass(Klass* klass) { // Lock-free access requires load_acquire for (Klass* k = OrderAccess::load_acquire(&_klasses); k != NULL; k = k->next_link()) { if (k == klass) return true; } return false; } // GC root of class loader data created. ClassLoaderData* ClassLoaderDataGraph::_head = NULL; ClassLoaderData* ClassLoaderDataGraph::_unloading = NULL; ClassLoaderData* ClassLoaderDataGraph::_saved_unloading = NULL; ClassLoaderData* ClassLoaderDataGraph::_saved_head = NULL; bool ClassLoaderDataGraph::_should_purge = false; bool ClassLoaderDataGraph::_should_clean_deallocate_lists = false; bool ClassLoaderDataGraph::_safepoint_cleanup_needed = false; bool ClassLoaderDataGraph::_metaspace_oom = false; // Add a new class loader data node to the list. Assign the newly created // ClassLoaderData into the java/lang/ClassLoader object as a hidden field ClassLoaderData* ClassLoaderDataGraph::add_to_graph(Handle loader, bool is_unsafe_anonymous) { ClassLoaderData* cld; { NoSafepointVerifier no_safepoints; // we mustn't GC until we've installed the // ClassLoaderData in the loader since the CLD // contains oops in _handles that must be walked. // GC will find the CLD through the loader after this. cld = new ClassLoaderData(loader, is_unsafe_anonymous); if (!is_unsafe_anonymous) { // First, Atomically set it ClassLoaderData* old = java_lang_ClassLoader::cmpxchg_loader_data(cld, loader(), NULL); if (old != NULL) { delete cld; // Returns the data. return old; } } } MutexLocker ml(ClassLoaderDataGraph_lock); // We won the race, and therefore the task of adding the data to the list of // class loader data cld->set_next(_head); _head = cld; LogTarget(Trace, class, loader, data) lt; if (lt.is_enabled()) { ResourceMark rm; LogStream ls(lt); ls.print("create "); cld->print_value_on(&ls); ls.cr(); } return cld; } ClassLoaderData* ClassLoaderDataGraph::add(Handle loader, bool is_unsafe_anonymous) { ClassLoaderData* loader_data = add_to_graph(loader, is_unsafe_anonymous); // Initialize _name and _name_and_id after the loader data is added to the // CLDG because adding the Symbol for _name and _name_and_id might safepoint. if (loader.not_null()) { loader_data->initialize_name(loader); } return loader_data; } void ClassLoaderDataGraph::cld_do(CLDClosure* cl) { assert_locked_or_safepoint_weak(ClassLoaderDataGraph_lock); for (ClassLoaderData* cld = _head; cld != NULL; cld = cld->_next) { cl->do_cld(cld); } } void ClassLoaderDataGraph::cld_unloading_do(CLDClosure* cl) { assert_locked_or_safepoint_weak(ClassLoaderDataGraph_lock); // Only walk the head until any clds not purged from prior unloading // (CMS doesn't purge right away). for (ClassLoaderData* cld = _unloading; cld != _saved_unloading; cld = cld->next()) { assert(cld->is_unloading(), "invariant"); cl->do_cld(cld); } } void ClassLoaderDataGraph::roots_cld_do(CLDClosure* strong, CLDClosure* weak) { assert_locked_or_safepoint_weak(ClassLoaderDataGraph_lock); for (ClassLoaderData* cld = _head; cld != NULL; cld = cld->_next) { CLDClosure* closure = cld->keep_alive() ? strong : weak; if (closure != NULL) { closure->do_cld(cld); } } } void ClassLoaderDataGraph::always_strong_cld_do(CLDClosure* cl) { assert_locked_or_safepoint_weak(ClassLoaderDataGraph_lock); if (ClassUnloading) { roots_cld_do(cl, NULL); } else { cld_do(cl); } } // Closure for locking and iterating through classes. LockedClassesDo::LockedClassesDo(classes_do_func_t f) : _function(f) { ClassLoaderDataGraph_lock->lock(); } LockedClassesDo::LockedClassesDo() : _function(NULL) { // callers provide their own do_klass ClassLoaderDataGraph_lock->lock(); } LockedClassesDo::~LockedClassesDo() { ClassLoaderDataGraph_lock->unlock(); } // Iterating over the CLDG needs to be locked because // unloading can remove entries concurrently soon. class ClassLoaderDataGraphIterator : public StackObj { ClassLoaderData* _next; HandleMark _hm; // clean up handles when this is done. Handle _holder; Thread* _thread; void hold_next() { if (_next != NULL) { _holder = Handle(_thread, _next->holder_phantom()); } } public: ClassLoaderDataGraphIterator() : _next(ClassLoaderDataGraph::_head) { _thread = Thread::current(); assert_locked_or_safepoint(ClassLoaderDataGraph_lock); hold_next(); } bool repeat() const { return _next != NULL; } ClassLoaderData* get_next() { ClassLoaderData* next = _next; if (_next != NULL) { _next = _next->next(); hold_next(); } return next; } }; // These functions assume that the caller has locked the ClassLoaderDataGraph_lock // if they are not calling the function from a safepoint. void ClassLoaderDataGraph::classes_do(KlassClosure* klass_closure) { ClassLoaderDataGraphIterator iter; while (iter.repeat()) { ClassLoaderData* cld = iter.get_next(); cld->classes_do(klass_closure); } } void ClassLoaderDataGraph::classes_do(void f(Klass* const)) { ClassLoaderDataGraphIterator iter; while (iter.repeat()) { ClassLoaderData* cld = iter.get_next(); cld->classes_do(f); } } void ClassLoaderDataGraph::methods_do(void f(Method*)) { ClassLoaderDataGraphIterator iter; while (iter.repeat()) { ClassLoaderData* cld = iter.get_next(); cld->methods_do(f); } } void ClassLoaderDataGraph::modules_do(void f(ModuleEntry*)) { assert_locked_or_safepoint(Module_lock); ClassLoaderDataGraphIterator iter; while (iter.repeat()) { ClassLoaderData* cld = iter.get_next(); cld->modules_do(f); } } void ClassLoaderDataGraph::modules_unloading_do(void f(ModuleEntry*)) { assert_locked_or_safepoint(ClassLoaderDataGraph_lock); // Only walk the head until any clds not purged from prior unloading // (CMS doesn't purge right away). for (ClassLoaderData* cld = _unloading; cld != _saved_unloading; cld = cld->next()) { assert(cld->is_unloading(), "invariant"); cld->modules_do(f); } } void ClassLoaderDataGraph::packages_do(void f(PackageEntry*)) { assert_locked_or_safepoint(Module_lock); ClassLoaderDataGraphIterator iter; while (iter.repeat()) { ClassLoaderData* cld = iter.get_next(); cld->packages_do(f); } } void ClassLoaderDataGraph::packages_unloading_do(void f(PackageEntry*)) { assert_locked_or_safepoint(ClassLoaderDataGraph_lock); // Only walk the head until any clds not purged from prior unloading // (CMS doesn't purge right away). for (ClassLoaderData* cld = _unloading; cld != _saved_unloading; cld = cld->next()) { assert(cld->is_unloading(), "invariant"); cld->packages_do(f); } } void ClassLoaderDataGraph::loaded_classes_do(KlassClosure* klass_closure) { ClassLoaderDataGraphIterator iter; while (iter.repeat()) { ClassLoaderData* cld = iter.get_next(); cld->loaded_classes_do(klass_closure); } } // This case can block but cannot do unloading (called from CDS) void ClassLoaderDataGraph::unlocked_loaded_classes_do(KlassClosure* klass_closure) { for (ClassLoaderData* cld = _head; cld != NULL; cld = cld->next()) { cld->loaded_classes_do(klass_closure); } } void ClassLoaderDataGraph::classes_unloading_do(void f(Klass* const)) { assert_locked_or_safepoint(ClassLoaderDataGraph_lock); // Only walk the head until any clds not purged from prior unloading // (CMS doesn't purge right away). for (ClassLoaderData* cld = _unloading; cld != _saved_unloading; cld = cld->next()) { assert(cld->is_unloading(), "invariant"); cld->classes_do(f); } } #define FOR_ALL_DICTIONARY(X) ClassLoaderDataGraphIterator iter; \ ClassLoaderData* X; \ while ((X = iter.get_next()) != NULL) \ if (X->dictionary() != NULL) // Walk classes in the loaded class dictionaries in various forms. // Only walks the classes defined in this class loader. void ClassLoaderDataGraph::dictionary_classes_do(void f(InstanceKlass*)) { FOR_ALL_DICTIONARY(cld) { cld->dictionary()->classes_do(f); } } // Only walks the classes defined in this class loader. void ClassLoaderDataGraph::dictionary_classes_do(void f(InstanceKlass*, TRAPS), TRAPS) { FOR_ALL_DICTIONARY(cld) { cld->dictionary()->classes_do(f, CHECK); } } void ClassLoaderDataGraph::verify_dictionary() { FOR_ALL_DICTIONARY(cld) { cld->dictionary()->verify(); } } void ClassLoaderDataGraph::print_dictionary(outputStream* st) { FOR_ALL_DICTIONARY(cld) { st->print("Dictionary for "); cld->print_value_on(st); st->cr(); cld->dictionary()->print_on(st); st->cr(); } } void ClassLoaderDataGraph::print_dictionary_statistics(outputStream* st) { FOR_ALL_DICTIONARY(cld) { ResourceMark rm; stringStream tempst; tempst.print("System Dictionary for %s class loader", cld->loader_name_and_id()); cld->dictionary()->print_table_statistics(st, tempst.as_string()); } } GrowableArray* ClassLoaderDataGraph::new_clds() { assert_locked_or_safepoint(ClassLoaderDataGraph_lock); assert(_head == NULL || _saved_head != NULL, "remember_new_clds(true) not called?"); GrowableArray* array = new GrowableArray(); // The CLDs in [_head, _saved_head] were all added during last call to remember_new_clds(true); ClassLoaderData* curr = _head; while (curr != _saved_head) { if (!curr->claimed()) { array->push(curr); LogTarget(Debug, class, loader, data) lt; if (lt.is_enabled()) { LogStream ls(lt); ls.print("found new CLD: "); curr->print_value_on(&ls); ls.cr(); } } curr = curr->_next; } return array; } #ifndef PRODUCT bool ClassLoaderDataGraph::contains_loader_data(ClassLoaderData* loader_data) { assert_locked_or_safepoint(ClassLoaderDataGraph_lock); for (ClassLoaderData* data = _head; data != NULL; data = data->next()) { if (loader_data == data) { return true; } } return false; } #endif // PRODUCT // Move class loader data from main list to the unloaded list for unloading // and deallocation later. bool ClassLoaderDataGraph::do_unloading(bool do_cleaning) { assert_locked_or_safepoint(ClassLoaderDataGraph_lock); // Indicate whether safepoint cleanup is needed. _safepoint_cleanup_needed |= do_cleaning; ClassLoaderData* data = _head; ClassLoaderData* prev = NULL; bool seen_dead_loader = false; uint loaders_processed = 0; uint loaders_removed = 0; // Save previous _unloading pointer for CMS which may add to unloading list before // purging and we don't want to rewalk the previously unloaded class loader data. _saved_unloading = _unloading; data = _head; while (data != NULL) { if (data->is_alive()) { prev = data; data = data->next(); loaders_processed++; continue; } seen_dead_loader = true; loaders_removed++; ClassLoaderData* dead = data; dead->unload(); data = data->next(); // Remove from loader list. // This class loader data will no longer be found // in the ClassLoaderDataGraph. if (prev != NULL) { prev->set_next(data); } else { assert(dead == _head, "sanity check"); _head = data; } dead->set_next(_unloading); _unloading = dead; } log_debug(class, loader, data)("do_unloading: loaders processed %u, loaders removed %u", loaders_processed, loaders_removed); return seen_dead_loader; } // There's at least one dead class loader. Purge refererences of healthy module // reads lists and package export lists to modules belonging to dead loaders. void ClassLoaderDataGraph::clean_module_and_package_info() { assert_locked_or_safepoint(ClassLoaderDataGraph_lock); ClassLoaderData* data = _head; while (data != NULL) { // Remove entries in the dictionary of live class loader that have // initiated loading classes in a dead class loader. if (data->dictionary() != NULL) { data->dictionary()->do_unloading(); } // Walk a ModuleEntry's reads, and a PackageEntry's exports // lists to determine if there are modules on those lists that are now // dead and should be removed. A module's life cycle is equivalent // to its defining class loader's life cycle. Since a module is // considered dead if its class loader is dead, these walks must // occur after each class loader's aliveness is determined. if (data->packages() != NULL) { data->packages()->purge_all_package_exports(); } if (data->modules_defined()) { data->modules()->purge_all_module_reads(); } data = data->next(); } } void ClassLoaderDataGraph::purge() { assert(SafepointSynchronize::is_at_safepoint(), "must be at safepoint!"); ClassLoaderData* list = _unloading; _unloading = NULL; ClassLoaderData* next = list; bool classes_unloaded = false; while (next != NULL) { ClassLoaderData* purge_me = next; next = purge_me->next(); delete purge_me; classes_unloaded = true; } if (classes_unloaded) { Metaspace::purge(); set_metaspace_oom(false); } } int ClassLoaderDataGraph::resize_if_needed() { assert(SafepointSynchronize::is_at_safepoint(), "must be at safepoint!"); int resized = 0; if (Dictionary::does_any_dictionary_needs_resizing()) { FOR_ALL_DICTIONARY(cld) { if (cld->dictionary()->resize_if_needed()) { resized++; } } } return resized; } ClassLoaderDataGraphKlassIteratorAtomic::ClassLoaderDataGraphKlassIteratorAtomic() : _next_klass(NULL) { assert(SafepointSynchronize::is_at_safepoint(), "must be at safepoint!"); ClassLoaderData* cld = ClassLoaderDataGraph::_head; Klass* klass = NULL; // Find the first klass in the CLDG. while (cld != NULL) { assert_locked_or_safepoint(cld->metaspace_lock()); klass = cld->_klasses; if (klass != NULL) { _next_klass = klass; return; } cld = cld->next(); } } Klass* ClassLoaderDataGraphKlassIteratorAtomic::next_klass_in_cldg(Klass* klass) { Klass* next = klass->next_link(); if (next != NULL) { return next; } // No more klasses in the current CLD. Time to find a new CLD. ClassLoaderData* cld = klass->class_loader_data(); assert_locked_or_safepoint(cld->metaspace_lock()); while (next == NULL) { cld = cld->next(); if (cld == NULL) { break; } next = cld->_klasses; } return next; } Klass* ClassLoaderDataGraphKlassIteratorAtomic::next_klass() { Klass* head = _next_klass; while (head != NULL) { Klass* next = next_klass_in_cldg(head); Klass* old_head = Atomic::cmpxchg(next, &_next_klass, head); if (old_head == head) { return head; // Won the CAS. } head = old_head; } // Nothing more for the iterator to hand out. assert(head == NULL, "head is " PTR_FORMAT ", expected not null:", p2i(head)); return NULL; } ClassLoaderDataGraphMetaspaceIterator::ClassLoaderDataGraphMetaspaceIterator() { assert(SafepointSynchronize::is_at_safepoint(), "must be at safepoint!"); _data = ClassLoaderDataGraph::_head; } ClassLoaderDataGraphMetaspaceIterator::~ClassLoaderDataGraphMetaspaceIterator() {} #ifndef PRODUCT // callable from debugger extern "C" int print_loader_data_graph() { ResourceMark rm; ClassLoaderDataGraph::print_on(tty); return 0; } void ClassLoaderDataGraph::verify() { ClassLoaderDataGraphIterator iter; while (iter.repeat()) { ClassLoaderData* cld = iter.get_next(); cld->verify(); } } void ClassLoaderDataGraph::print_on(outputStream * const out) { ClassLoaderDataGraphIterator iter; while (iter.repeat()) { ClassLoaderData* cld = iter.get_next(); cld->print_on(out); } } #endif // PRODUCT