1 /* 2 * Copyright (c) 1997, 2017, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "classfile/dictionary.hpp" 27 #include "classfile/javaClasses.hpp" 28 #include "classfile/systemDictionary.hpp" 29 #include "classfile/vmSymbols.hpp" 30 #include "gc/shared/collectedHeap.inline.hpp" 31 #include "logging/log.hpp" 32 #include "memory/heapInspection.hpp" 33 #include "memory/metadataFactory.hpp" 34 #include "memory/metaspaceClosure.hpp" 35 #include "memory/metaspaceShared.hpp" 36 #include "memory/oopFactory.hpp" 37 #include "memory/resourceArea.hpp" 38 #include "oops/instanceKlass.hpp" 39 #include "oops/klass.inline.hpp" 40 #include "oops/oop.inline.hpp" 41 #include "runtime/atomic.hpp" 42 #include "runtime/orderAccess.inline.hpp" 43 #include "trace/traceMacros.hpp" 44 #include "utilities/macros.hpp" 45 #include "utilities/stack.inline.hpp" 46 47 void Klass::set_java_mirror(Handle m) { 48 assert(!m.is_null(), "New mirror should never be null."); 49 assert(_java_mirror.resolve() == NULL, "should only be used to initialize mirror"); 50 _java_mirror = class_loader_data()->add_handle(m); 51 } 52 53 oop Klass::java_mirror() const { 54 return _java_mirror.resolve(); 55 } 56 57 bool Klass::is_cloneable() const { 58 return _access_flags.is_cloneable_fast() || 59 is_subtype_of(SystemDictionary::Cloneable_klass()); 60 } 61 62 void Klass::set_is_cloneable() { 63 if (name() != vmSymbols::java_lang_invoke_MemberName()) { 64 _access_flags.set_is_cloneable_fast(); 65 } else { 66 assert(is_final(), "no subclasses allowed"); 67 // MemberName cloning should not be intrinsified and always happen in JVM_Clone. 68 } 69 } 70 71 void Klass::set_name(Symbol* n) { 72 _name = n; 73 if (_name != NULL) _name->increment_refcount(); 74 } 75 76 bool Klass::is_subclass_of(const Klass* k) const { 77 // Run up the super chain and check 78 if (this == k) return true; 79 80 Klass* t = const_cast<Klass*>(this)->super(); 81 82 while (t != NULL) { 83 if (t == k) return true; 84 t = t->super(); 85 } 86 return false; 87 } 88 89 bool Klass::search_secondary_supers(Klass* k) const { 90 // Put some extra logic here out-of-line, before the search proper. 91 // This cuts down the size of the inline method. 92 93 // This is necessary, since I am never in my own secondary_super list. 94 if (this == k) 95 return true; 96 // Scan the array-of-objects for a match 97 int cnt = secondary_supers()->length(); 98 for (int i = 0; i < cnt; i++) { 99 if (secondary_supers()->at(i) == k) { 100 ((Klass*)this)->set_secondary_super_cache(k); 101 return true; 102 } 103 } 104 return false; 105 } 106 107 // Return self, except for abstract classes with exactly 1 108 // implementor. Then return the 1 concrete implementation. 109 Klass *Klass::up_cast_abstract() { 110 Klass *r = this; 111 while( r->is_abstract() ) { // Receiver is abstract? 112 Klass *s = r->subklass(); // Check for exactly 1 subklass 113 if( !s || s->next_sibling() ) // Oops; wrong count; give up 114 return this; // Return 'this' as a no-progress flag 115 r = s; // Loop till find concrete class 116 } 117 return r; // Return the 1 concrete class 118 } 119 120 // Find LCA in class hierarchy 121 Klass *Klass::LCA( Klass *k2 ) { 122 Klass *k1 = this; 123 while( 1 ) { 124 if( k1->is_subtype_of(k2) ) return k2; 125 if( k2->is_subtype_of(k1) ) return k1; 126 k1 = k1->super(); 127 k2 = k2->super(); 128 } 129 } 130 131 132 void Klass::check_valid_for_instantiation(bool throwError, TRAPS) { 133 ResourceMark rm(THREAD); 134 THROW_MSG(throwError ? vmSymbols::java_lang_InstantiationError() 135 : vmSymbols::java_lang_InstantiationException(), external_name()); 136 } 137 138 139 void Klass::copy_array(arrayOop s, int src_pos, arrayOop d, int dst_pos, int length, TRAPS) { 140 THROW(vmSymbols::java_lang_ArrayStoreException()); 141 } 142 143 144 void Klass::initialize(TRAPS) { 145 ShouldNotReachHere(); 146 } 147 148 bool Klass::compute_is_subtype_of(Klass* k) { 149 assert(k->is_klass(), "argument must be a class"); 150 return is_subclass_of(k); 151 } 152 153 Klass* Klass::find_field(Symbol* name, Symbol* sig, fieldDescriptor* fd) const { 154 #ifdef ASSERT 155 tty->print_cr("Error: find_field called on a klass oop." 156 " Likely error: reflection method does not correctly" 157 " wrap return value in a mirror object."); 158 #endif 159 ShouldNotReachHere(); 160 return NULL; 161 } 162 163 Method* Klass::uncached_lookup_method(const Symbol* name, const Symbol* signature, OverpassLookupMode overpass_mode) const { 164 #ifdef ASSERT 165 tty->print_cr("Error: uncached_lookup_method called on a klass oop." 166 " Likely error: reflection method does not correctly" 167 " wrap return value in a mirror object."); 168 #endif 169 ShouldNotReachHere(); 170 return NULL; 171 } 172 173 #if INCLUDE_VBC 174 void* Klass::operator new(size_t size, ClassLoaderData* loader_data, size_t word_size, bool is_value, TRAPS) throw() { 175 // Pad size in case need adjust to even/odd klass ptr 176 uintptr_t addr = (uintptr_t) Metaspace::allocate(loader_data, word_size + (1 << LogKlassAlignment), MetaspaceObj::ClassType, THREAD); 177 // values are odd, otherwise make even (and vice versa) 178 if (is_value ^ ((addr & KlassPtrEvenOddMask) >> LogKlassAlignmentInBytes)) { 179 addr += (1 << LogKlassAlignmentInBytes); 180 } 181 assert(is_aligned(addr, (1 << LogKlassAlignmentInBytes)), "Klass base alignment incorrect"); 182 assert(is_value || (addr & KlassPtrEvenOddMask) == 0, "Klass even alignment incorrect"); 183 assert((!is_value) || (addr & KlassPtrEvenOddMask), "Klass odd alignment incorrect"); 184 return (void*) addr; 185 #else 186 void* Klass::operator new(size_t size, ClassLoaderData* loader_data, size_t word_size, TRAPS) throw() { 187 return Metaspace::allocate(loader_data, word_size, MetaspaceObj::ClassType, THREAD); 188 #endif 189 } 190 191 192 // "Normal" instantiation is preceeded by a MetaspaceObj allocation 193 // which zeros out memory - calloc equivalent. 194 // The constructor is also used from CppVtableCloner, 195 // which doesn't zero out the memory before calling the constructor. 196 // Need to set the _java_mirror field explicitly to not hit an assert that the field 197 // should be NULL before setting it. 198 Klass::Klass() : _prototype_header(markOopDesc::prototype()), 199 _shared_class_path_index(-1), 200 _java_mirror(NULL) { 201 202 _primary_supers[0] = this; 203 set_super_check_offset(in_bytes(primary_supers_offset())); 204 } 205 206 jint Klass::array_layout_helper(BasicType etype) { 207 assert(etype >= T_BOOLEAN && etype <= T_OBJECT, "valid etype"); 208 // Note that T_ARRAY is not allowed here. 209 int hsize = arrayOopDesc::base_offset_in_bytes(etype); 210 int esize = type2aelembytes(etype); 211 bool isobj = (etype == T_OBJECT); 212 int tag = isobj ? _lh_array_tag_obj_value : _lh_array_tag_type_value; 213 int lh = array_layout_helper(tag, hsize, etype, exact_log2(esize)); 214 215 assert(lh < (int)_lh_neutral_value, "must look like an array layout"); 216 assert(layout_helper_is_array(lh), "correct kind"); 217 assert(layout_helper_is_objArray(lh) == isobj, "correct kind"); 218 assert(layout_helper_is_typeArray(lh) == !isobj, "correct kind"); 219 assert(layout_helper_header_size(lh) == hsize, "correct decode"); 220 assert(layout_helper_element_type(lh) == etype, "correct decode"); 221 assert(1 << layout_helper_log2_element_size(lh) == esize, "correct decode"); 222 223 return lh; 224 } 225 226 bool Klass::can_be_primary_super_slow() const { 227 if (super() == NULL) 228 return true; 229 else if (super()->super_depth() >= primary_super_limit()-1) 230 return false; 231 else 232 return true; 233 } 234 235 void Klass::initialize_supers(Klass* k, TRAPS) { 236 if (FastSuperclassLimit == 0) { 237 // None of the other machinery matters. 238 set_super(k); 239 return; 240 } 241 if (k == NULL) { 242 set_super(NULL); 243 _primary_supers[0] = this; 244 assert(super_depth() == 0, "Object must already be initialized properly"); 245 } else if (k != super() || k == SystemDictionary::Object_klass()) { 246 assert(super() == NULL || super() == SystemDictionary::Object_klass(), 247 "initialize this only once to a non-trivial value"); 248 set_super(k); 249 Klass* sup = k; 250 int sup_depth = sup->super_depth(); 251 juint my_depth = MIN2(sup_depth + 1, (int)primary_super_limit()); 252 if (!can_be_primary_super_slow()) 253 my_depth = primary_super_limit(); 254 for (juint i = 0; i < my_depth; i++) { 255 _primary_supers[i] = sup->_primary_supers[i]; 256 } 257 Klass* *super_check_cell; 258 if (my_depth < primary_super_limit()) { 259 _primary_supers[my_depth] = this; 260 super_check_cell = &_primary_supers[my_depth]; 261 } else { 262 // Overflow of the primary_supers array forces me to be secondary. 263 super_check_cell = &_secondary_super_cache; 264 } 265 set_super_check_offset((address)super_check_cell - (address) this); 266 267 #ifdef ASSERT 268 { 269 juint j = super_depth(); 270 assert(j == my_depth, "computed accessor gets right answer"); 271 Klass* t = this; 272 while (!t->can_be_primary_super()) { 273 t = t->super(); 274 j = t->super_depth(); 275 } 276 for (juint j1 = j+1; j1 < primary_super_limit(); j1++) { 277 assert(primary_super_of_depth(j1) == NULL, "super list padding"); 278 } 279 while (t != NULL) { 280 assert(primary_super_of_depth(j) == t, "super list initialization"); 281 t = t->super(); 282 --j; 283 } 284 assert(j == (juint)-1, "correct depth count"); 285 } 286 #endif 287 } 288 289 if (secondary_supers() == NULL) { 290 291 // Now compute the list of secondary supertypes. 292 // Secondaries can occasionally be on the super chain, 293 // if the inline "_primary_supers" array overflows. 294 int extras = 0; 295 Klass* p; 296 for (p = super(); !(p == NULL || p->can_be_primary_super()); p = p->super()) { 297 ++extras; 298 } 299 300 ResourceMark rm(THREAD); // need to reclaim GrowableArrays allocated below 301 302 // Compute the "real" non-extra secondaries. 303 GrowableArray<Klass*>* secondaries = compute_secondary_supers(extras); 304 if (secondaries == NULL) { 305 // secondary_supers set by compute_secondary_supers 306 return; 307 } 308 309 GrowableArray<Klass*>* primaries = new GrowableArray<Klass*>(extras); 310 311 for (p = super(); !(p == NULL || p->can_be_primary_super()); p = p->super()) { 312 int i; // Scan for overflow primaries being duplicates of 2nd'arys 313 314 // This happens frequently for very deeply nested arrays: the 315 // primary superclass chain overflows into the secondary. The 316 // secondary list contains the element_klass's secondaries with 317 // an extra array dimension added. If the element_klass's 318 // secondary list already contains some primary overflows, they 319 // (with the extra level of array-ness) will collide with the 320 // normal primary superclass overflows. 321 for( i = 0; i < secondaries->length(); i++ ) { 322 if( secondaries->at(i) == p ) 323 break; 324 } 325 if( i < secondaries->length() ) 326 continue; // It's a dup, don't put it in 327 primaries->push(p); 328 } 329 // Combine the two arrays into a metadata object to pack the array. 330 // The primaries are added in the reverse order, then the secondaries. 331 int new_length = primaries->length() + secondaries->length(); 332 Array<Klass*>* s2 = MetadataFactory::new_array<Klass*>( 333 class_loader_data(), new_length, CHECK); 334 int fill_p = primaries->length(); 335 for (int j = 0; j < fill_p; j++) { 336 s2->at_put(j, primaries->pop()); // add primaries in reverse order. 337 } 338 for( int j = 0; j < secondaries->length(); j++ ) { 339 s2->at_put(j+fill_p, secondaries->at(j)); // add secondaries on the end. 340 } 341 342 #ifdef ASSERT 343 // We must not copy any NULL placeholders left over from bootstrap. 344 for (int j = 0; j < s2->length(); j++) { 345 assert(s2->at(j) != NULL, "correct bootstrapping order"); 346 } 347 #endif 348 349 set_secondary_supers(s2); 350 } 351 } 352 353 GrowableArray<Klass*>* Klass::compute_secondary_supers(int num_extra_slots) { 354 assert(num_extra_slots == 0, "override for complex klasses"); 355 set_secondary_supers(Universe::the_empty_klass_array()); 356 return NULL; 357 } 358 359 360 InstanceKlass* Klass::superklass() const { 361 assert(super() == NULL || super()->is_instance_klass(), "must be instance klass"); 362 return _super == NULL ? NULL : InstanceKlass::cast(_super); 363 } 364 365 void Klass::set_subklass(Klass* s) { 366 assert(s != this, "sanity check"); 367 _subklass = s; 368 } 369 370 void Klass::set_next_sibling(Klass* s) { 371 assert(s != this, "sanity check"); 372 _next_sibling = s; 373 } 374 375 void Klass::append_to_sibling_list() { 376 debug_only(verify();) 377 // add ourselves to superklass' subklass list 378 InstanceKlass* super = superklass(); 379 if (super == NULL) return; // special case: class Object 380 assert((!super->is_interface() // interfaces cannot be supers 381 && (super->superklass() == NULL || !is_interface())), 382 "an interface can only be a subklass of Object"); 383 Klass* prev_first_subklass = super->subklass(); 384 if (prev_first_subklass != NULL) { 385 // set our sibling to be the superklass' previous first subklass 386 set_next_sibling(prev_first_subklass); 387 } 388 // make ourselves the superklass' first subklass 389 super->set_subklass(this); 390 debug_only(verify();) 391 } 392 393 bool Klass::is_loader_alive(BoolObjectClosure* is_alive) { 394 #ifdef ASSERT 395 // The class is alive iff the class loader is alive. 396 oop loader = class_loader(); 397 bool loader_alive = (loader == NULL) || is_alive->do_object_b(loader); 398 #endif // ASSERT 399 400 // The class is alive if it's mirror is alive (which should be marked if the 401 // loader is alive) unless it's an anoymous class. 402 bool mirror_alive = is_alive->do_object_b(java_mirror()); 403 assert(!mirror_alive || loader_alive, "loader must be alive if the mirror is" 404 " but not the other way around with anonymous classes"); 405 return mirror_alive; 406 } 407 408 void Klass::clean_weak_klass_links(BoolObjectClosure* is_alive, bool clean_alive_klasses) { 409 if (!ClassUnloading) { 410 return; 411 } 412 413 Klass* root = SystemDictionary::Object_klass(); 414 Stack<Klass*, mtGC> stack; 415 416 stack.push(root); 417 while (!stack.is_empty()) { 418 Klass* current = stack.pop(); 419 420 assert(current->is_loader_alive(is_alive), "just checking, this should be live"); 421 422 // Find and set the first alive subklass 423 Klass* sub = current->subklass(); 424 while (sub != NULL && !sub->is_loader_alive(is_alive)) { 425 #ifndef PRODUCT 426 if (log_is_enabled(Trace, class, unload)) { 427 ResourceMark rm; 428 log_trace(class, unload)("unlinking class (subclass): %s", sub->external_name()); 429 } 430 #endif 431 sub = sub->next_sibling(); 432 } 433 current->set_subklass(sub); 434 if (sub != NULL) { 435 stack.push(sub); 436 } 437 438 // Find and set the first alive sibling 439 Klass* sibling = current->next_sibling(); 440 while (sibling != NULL && !sibling->is_loader_alive(is_alive)) { 441 if (log_is_enabled(Trace, class, unload)) { 442 ResourceMark rm; 443 log_trace(class, unload)("[Unlinking class (sibling) %s]", sibling->external_name()); 444 } 445 sibling = sibling->next_sibling(); 446 } 447 current->set_next_sibling(sibling); 448 if (sibling != NULL) { 449 stack.push(sibling); 450 } 451 452 // Clean the implementors list and method data. 453 if (clean_alive_klasses && current->is_instance_klass()) { 454 InstanceKlass* ik = InstanceKlass::cast(current); 455 ik->clean_weak_instanceklass_links(is_alive); 456 457 // JVMTI RedefineClasses creates previous versions that are not in 458 // the class hierarchy, so process them here. 459 while ((ik = ik->previous_versions()) != NULL) { 460 ik->clean_weak_instanceklass_links(is_alive); 461 } 462 } 463 } 464 } 465 466 void Klass::metaspace_pointers_do(MetaspaceClosure* it) { 467 if (log_is_enabled(Trace, cds)) { 468 ResourceMark rm; 469 log_trace(cds)("Iter(Klass): %p (%s)", this, external_name()); 470 } 471 472 it->push(&_name); 473 it->push(&_secondary_super_cache); 474 it->push(&_secondary_supers); 475 for (int i = 0; i < _primary_super_limit; i++) { 476 it->push(&_primary_supers[i]); 477 } 478 it->push(&_super); 479 it->push(&_subklass); 480 it->push(&_next_sibling); 481 it->push(&_next_link); 482 483 vtableEntry* vt = start_of_vtable(); 484 for (int i=0; i<vtable_length(); i++) { 485 it->push(vt[i].method_addr()); 486 } 487 } 488 489 void Klass::remove_unshareable_info() { 490 assert (DumpSharedSpaces, "only called for DumpSharedSpaces"); 491 TRACE_REMOVE_ID(this); 492 if (log_is_enabled(Trace, cds, unshareable)) { 493 ResourceMark rm; 494 log_trace(cds, unshareable)("remove: %s", external_name()); 495 } 496 497 set_subklass(NULL); 498 set_next_sibling(NULL); 499 set_next_link(NULL); 500 501 // Null out class_loader_data because we don't share that yet. 502 set_class_loader_data(NULL); 503 set_is_shared(); 504 } 505 506 void Klass::remove_java_mirror() { 507 assert (DumpSharedSpaces, "only called for DumpSharedSpaces"); 508 if (log_is_enabled(Trace, cds, unshareable)) { 509 ResourceMark rm; 510 log_trace(cds, unshareable)("remove java_mirror: %s", external_name()); 511 } 512 // Just null out the mirror. The class_loader_data() no longer exists. 513 _java_mirror = NULL; 514 } 515 516 void Klass::restore_unshareable_info(ClassLoaderData* loader_data, Handle protection_domain, TRAPS) { 517 assert(is_klass(), "ensure C++ vtable is restored"); 518 assert(is_shared(), "must be set"); 519 TRACE_RESTORE_ID(this); 520 if (log_is_enabled(Trace, cds, unshareable)) { 521 ResourceMark rm; 522 log_trace(cds, unshareable)("restore: %s", external_name()); 523 } 524 525 // If an exception happened during CDS restore, some of these fields may already be 526 // set. We leave the class on the CLD list, even if incomplete so that we don't 527 // modify the CLD list outside a safepoint. 528 if (class_loader_data() == NULL) { 529 // Restore class_loader_data to the null class loader data 530 set_class_loader_data(loader_data); 531 532 // Add to null class loader list first before creating the mirror 533 // (same order as class file parsing) 534 loader_data->add_class(this); 535 } 536 537 // Recreate the class mirror. 538 // Only recreate it if not present. A previous attempt to restore may have 539 // gotten an OOM later but keep the mirror if it was created. 540 if (java_mirror() == NULL) { 541 Handle loader(THREAD, loader_data->class_loader()); 542 ModuleEntry* module_entry = NULL; 543 Klass* k = this; 544 if (k->is_objArray_klass()) { 545 k = ObjArrayKlass::cast(k)->bottom_klass(); 546 } 547 // Obtain klass' module. 548 if (k->is_instance_klass()) { 549 InstanceKlass* ik = (InstanceKlass*) k; 550 module_entry = ik->module(); 551 } else { 552 module_entry = ModuleEntryTable::javabase_moduleEntry(); 553 } 554 // Obtain java.lang.Module, if available 555 Handle module_handle(THREAD, ((module_entry != NULL) ? module_entry->module() : (oop)NULL)); 556 java_lang_Class::create_mirror(this, loader, module_handle, protection_domain, CHECK); 557 } 558 } 559 560 Klass* Klass::array_klass_or_null(int rank) { 561 EXCEPTION_MARK; 562 // No exception can be thrown by array_klass_impl when called with or_null == true. 563 // (In anycase, the execption mark will fail if it do so) 564 return array_klass_impl(true, rank, THREAD); 565 } 566 567 568 Klass* Klass::array_klass_or_null() { 569 EXCEPTION_MARK; 570 // No exception can be thrown by array_klass_impl when called with or_null == true. 571 // (In anycase, the execption mark will fail if it do so) 572 return array_klass_impl(true, THREAD); 573 } 574 575 576 Klass* Klass::array_klass_impl(bool or_null, int rank, TRAPS) { 577 fatal("array_klass should be dispatched to InstanceKlass, ObjArrayKlass or TypeArrayKlass"); 578 return NULL; 579 } 580 581 582 Klass* Klass::array_klass_impl(bool or_null, TRAPS) { 583 fatal("array_klass should be dispatched to InstanceKlass, ObjArrayKlass or TypeArrayKlass"); 584 return NULL; 585 } 586 587 oop Klass::class_loader() const { return class_loader_data()->class_loader(); } 588 589 // In product mode, this function doesn't have virtual function calls so 590 // there might be some performance advantage to handling InstanceKlass here. 591 const char* Klass::external_name() const { 592 if (is_instance_klass()) { 593 const InstanceKlass* ik = static_cast<const InstanceKlass*>(this); 594 if (ik->is_anonymous()) { 595 intptr_t hash = 0; 596 if (ik->java_mirror() != NULL) { 597 // java_mirror might not be created yet, return 0 as hash. 598 hash = ik->java_mirror()->identity_hash(); 599 } 600 char hash_buf[40]; 601 sprintf(hash_buf, "/" UINTX_FORMAT, (uintx)hash); 602 size_t hash_len = strlen(hash_buf); 603 604 size_t result_len = name()->utf8_length(); 605 char* result = NEW_RESOURCE_ARRAY(char, result_len + hash_len + 1); 606 name()->as_klass_external_name(result, (int) result_len + 1); 607 assert(strlen(result) == result_len, ""); 608 strcpy(result + result_len, hash_buf); 609 assert(strlen(result) == result_len + hash_len, ""); 610 return result; 611 } 612 } 613 if (name() == NULL) return "<unknown>"; 614 return name()->as_klass_external_name(); 615 } 616 617 618 const char* Klass::signature_name() const { 619 if (name() == NULL) return "<unknown>"; 620 return name()->as_C_string(); 621 } 622 623 // Unless overridden, modifier_flags is 0. 624 jint Klass::compute_modifier_flags(TRAPS) const { 625 return 0; 626 } 627 628 int Klass::atomic_incr_biased_lock_revocation_count() { 629 return (int) Atomic::add(1, &_biased_lock_revocation_count); 630 } 631 632 // Unless overridden, jvmti_class_status has no flags set. 633 jint Klass::jvmti_class_status() const { 634 return 0; 635 } 636 637 638 // Printing 639 640 void Klass::print_on(outputStream* st) const { 641 ResourceMark rm; 642 // print title 643 st->print("%s", internal_name()); 644 print_address_on(st); 645 st->cr(); 646 } 647 648 void Klass::oop_print_on(oop obj, outputStream* st) { 649 ResourceMark rm; 650 // print title 651 st->print_cr("%s ", internal_name()); 652 obj->print_address_on(st); 653 654 if (WizardMode) { 655 // print header 656 obj->mark()->print_on(st); 657 } 658 659 // print class 660 st->print(" - klass: "); 661 obj->klass()->print_value_on(st); 662 st->cr(); 663 } 664 665 void Klass::oop_print_value_on(oop obj, outputStream* st) { 666 // print title 667 ResourceMark rm; // Cannot print in debug mode without this 668 st->print("%s", internal_name()); 669 obj->print_address_on(st); 670 } 671 672 #if INCLUDE_SERVICES 673 // Size Statistics 674 void Klass::collect_statistics(KlassSizeStats *sz) const { 675 sz->_klass_bytes = sz->count(this); 676 sz->_mirror_bytes = sz->count(java_mirror()); 677 sz->_secondary_supers_bytes = sz->count_array(secondary_supers()); 678 679 sz->_ro_bytes += sz->_secondary_supers_bytes; 680 sz->_rw_bytes += sz->_klass_bytes + sz->_mirror_bytes; 681 } 682 #endif // INCLUDE_SERVICES 683 684 // Verification 685 686 void Klass::verify_on(outputStream* st) { 687 688 // This can be expensive, but it is worth checking that this klass is actually 689 // in the CLD graph but not in production. 690 assert(Metaspace::contains((address)this), "Should be"); 691 692 guarantee(this->is_klass(),"should be klass"); 693 694 if (super() != NULL) { 695 guarantee(super()->is_klass(), "should be klass"); 696 } 697 if (secondary_super_cache() != NULL) { 698 Klass* ko = secondary_super_cache(); 699 guarantee(ko->is_klass(), "should be klass"); 700 } 701 for ( uint i = 0; i < primary_super_limit(); i++ ) { 702 Klass* ko = _primary_supers[i]; 703 if (ko != NULL) { 704 guarantee(ko->is_klass(), "should be klass"); 705 } 706 } 707 708 if (java_mirror() != NULL) { 709 guarantee(oopDesc::is_oop(java_mirror()), "should be instance"); 710 } 711 } 712 713 void Klass::oop_verify_on(oop obj, outputStream* st) { 714 guarantee(oopDesc::is_oop(obj), "should be oop"); 715 guarantee(obj->klass()->is_klass(), "klass field is not a klass"); 716 } 717 718 klassVtable Klass::vtable() const { 719 return klassVtable(const_cast<Klass*>(this), start_of_vtable(), vtable_length() / vtableEntry::size()); 720 } 721 722 vtableEntry* Klass::start_of_vtable() const { 723 return (vtableEntry*) ((address)this + in_bytes(vtable_start_offset())); 724 } 725 726 Method* Klass::method_at_vtable(int index) { 727 #ifndef PRODUCT 728 assert(index >= 0, "valid vtable index"); 729 if (DebugVtables) { 730 verify_vtable_index(index); 731 } 732 #endif 733 return start_of_vtable()[index].method(); 734 } 735 736 ByteSize Klass::vtable_start_offset() { 737 return in_ByteSize(InstanceKlass::header_size() * wordSize); 738 } 739 740 #ifndef PRODUCT 741 742 bool Klass::verify_vtable_index(int i) { 743 int limit = vtable_length()/vtableEntry::size(); 744 assert(i >= 0 && i < limit, "index %d out of bounds %d", i, limit); 745 return true; 746 } 747 748 bool Klass::verify_itable_index(int i) { 749 assert(is_instance_klass(), ""); 750 int method_count = klassItable::method_count_for_interface(this); 751 assert(i >= 0 && i < method_count, "index out of bounds"); 752 return true; 753 } 754 755 #endif