1 /* 2 * Copyright (c) 1998, 2015, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "code/codeBlob.hpp" 27 #include "code/codeCache.hpp" 28 #include "code/nmethod.hpp" 29 #include "code/scopeDesc.hpp" 30 #include "compiler/oopMap.hpp" 31 #include "gc/shared/collectedHeap.hpp" 32 #include "memory/allocation.inline.hpp" 33 #include "memory/resourceArea.hpp" 34 #include "runtime/frame.inline.hpp" 35 #include "runtime/signature.hpp" 36 #ifdef COMPILER1 37 #include "c1/c1_Defs.hpp" 38 #endif 39 #ifdef COMPILER2 40 #include "opto/optoreg.hpp" 41 #endif 42 #ifdef SPARC 43 #include "vmreg_sparc.inline.hpp" 44 #endif 45 46 // OopMapStream 47 48 OopMapStream::OopMapStream(OopMap* oop_map, int oop_types_mask) { 49 _stream = new CompressedReadStream(oop_map->write_stream()->buffer()); 50 _mask = oop_types_mask; 51 _size = oop_map->omv_count(); 52 _position = 0; 53 _valid_omv = false; 54 } 55 56 OopMapStream::OopMapStream(const ImmutableOopMap* oop_map, int oop_types_mask) { 57 _stream = new CompressedReadStream(oop_map->data_addr()); 58 _mask = oop_types_mask; 59 _size = oop_map->count(); 60 _position = 0; 61 _valid_omv = false; 62 } 63 64 void OopMapStream::find_next() { 65 while(_position++ < _size) { 66 _omv.read_from(_stream); 67 if(((int)_omv.type() & _mask) > 0) { 68 _valid_omv = true; 69 return; 70 } 71 } 72 _valid_omv = false; 73 } 74 75 76 // OopMap 77 78 // frame_size units are stack-slots (4 bytes) NOT intptr_t; we can name odd 79 // slots to hold 4-byte values like ints and floats in the LP64 build. 80 OopMap::OopMap(int frame_size, int arg_count) { 81 // OopMaps are usually quite so small, so pick a small initial size 82 set_write_stream(new CompressedWriteStream(32)); 83 set_omv_count(0); 84 85 #ifdef ASSERT 86 _locs_length = VMRegImpl::stack2reg(0)->value() + frame_size + arg_count; 87 _locs_used = NEW_RESOURCE_ARRAY(OopMapValue::oop_types, _locs_length); 88 for(int i = 0; i < _locs_length; i++) _locs_used[i] = OopMapValue::unused_value; 89 #endif 90 } 91 92 93 OopMap::OopMap(OopMap::DeepCopyToken, OopMap* source) { 94 // This constructor does a deep copy 95 // of the source OopMap. 96 set_write_stream(new CompressedWriteStream(source->omv_count() * 2)); 97 set_omv_count(0); 98 set_offset(source->offset()); 99 100 #ifdef ASSERT 101 _locs_length = source->_locs_length; 102 _locs_used = NEW_RESOURCE_ARRAY(OopMapValue::oop_types, _locs_length); 103 for(int i = 0; i < _locs_length; i++) _locs_used[i] = OopMapValue::unused_value; 104 #endif 105 106 // We need to copy the entries too. 107 for (OopMapStream oms(source); !oms.is_done(); oms.next()) { 108 OopMapValue omv = oms.current(); 109 omv.write_on(write_stream()); 110 increment_count(); 111 } 112 } 113 114 115 OopMap* OopMap::deep_copy() { 116 return new OopMap(_deep_copy_token, this); 117 } 118 119 void OopMap::copy_data_to(address addr) const { 120 memcpy(addr, write_stream()->buffer(), write_stream()->position()); 121 } 122 123 int OopMap::heap_size() const { 124 int size = sizeof(OopMap); 125 int align = sizeof(void *) - 1; 126 size += write_stream()->position(); 127 // Align to a reasonable ending point 128 size = ((size+align) & ~align); 129 return size; 130 } 131 132 // frame_size units are stack-slots (4 bytes) NOT intptr_t; we can name odd 133 // slots to hold 4-byte values like ints and floats in the LP64 build. 134 void OopMap::set_xxx(VMReg reg, OopMapValue::oop_types x, VMReg optional) { 135 136 assert(reg->value() < _locs_length, "too big reg value for stack size"); 137 assert( _locs_used[reg->value()] == OopMapValue::unused_value, "cannot insert twice" ); 138 debug_only( _locs_used[reg->value()] = x; ) 139 140 OopMapValue o(reg, x); 141 142 if(x == OopMapValue::callee_saved_value) { 143 // This can never be a stack location, so we don't need to transform it. 144 assert(optional->is_reg(), "Trying to callee save a stack location"); 145 o.set_content_reg(optional); 146 } else if(x == OopMapValue::derived_oop_value) { 147 o.set_content_reg(optional); 148 } 149 150 o.write_on(write_stream()); 151 increment_count(); 152 } 153 154 155 void OopMap::set_oop(VMReg reg) { 156 set_xxx(reg, OopMapValue::oop_value, VMRegImpl::Bad()); 157 } 158 159 160 void OopMap::set_value(VMReg reg) { 161 // At this time, we don't need value entries in our OopMap. 162 } 163 164 165 void OopMap::set_narrowoop(VMReg reg) { 166 set_xxx(reg, OopMapValue::narrowoop_value, VMRegImpl::Bad()); 167 } 168 169 170 void OopMap::set_callee_saved(VMReg reg, VMReg caller_machine_register ) { 171 set_xxx(reg, OopMapValue::callee_saved_value, caller_machine_register); 172 } 173 174 175 void OopMap::set_derived_oop(VMReg reg, VMReg derived_from_local_register ) { 176 if( reg == derived_from_local_register ) { 177 // Actually an oop, derived shares storage with base, 178 set_oop(reg); 179 } else { 180 set_xxx(reg, OopMapValue::derived_oop_value, derived_from_local_register); 181 } 182 } 183 184 // OopMapSet 185 186 OopMapSet::OopMapSet() { 187 set_om_size(MinOopMapAllocation); 188 set_om_count(0); 189 OopMap** temp = NEW_RESOURCE_ARRAY(OopMap*, om_size()); 190 set_om_data(temp); 191 } 192 193 194 void OopMapSet::grow_om_data() { 195 int new_size = om_size() * 2; 196 OopMap** new_data = NEW_RESOURCE_ARRAY(OopMap*, new_size); 197 memcpy(new_data,om_data(),om_size() * sizeof(OopMap*)); 198 set_om_size(new_size); 199 set_om_data(new_data); 200 } 201 202 void OopMapSet::add_gc_map(int pc_offset, OopMap *map ) { 203 assert(om_size() != -1,"Cannot grow a fixed OopMapSet"); 204 205 if(om_count() >= om_size()) { 206 grow_om_data(); 207 } 208 map->set_offset(pc_offset); 209 210 #ifdef ASSERT 211 if(om_count() > 0) { 212 OopMap* last = at(om_count()-1); 213 if (last->offset() == map->offset() ) { 214 fatal("OopMap inserted twice"); 215 } 216 if(last->offset() > map->offset()) { 217 tty->print_cr( "WARNING, maps not sorted: pc[%d]=%d, pc[%d]=%d", 218 om_count(),last->offset(),om_count()+1,map->offset()); 219 } 220 } 221 #endif // ASSERT 222 223 set(om_count(),map); 224 increment_count(); 225 } 226 227 228 int OopMapSet::heap_size() const { 229 // The space we use 230 int size = sizeof(OopMap); 231 int align = sizeof(void *) - 1; 232 size = ((size+align) & ~align); 233 size += om_count() * sizeof(OopMap*); 234 235 // Now add in the space needed for the indivdiual OopMaps 236 for(int i=0; i < om_count(); i++) { 237 size += at(i)->heap_size(); 238 } 239 // We don't need to align this, it will be naturally pointer aligned 240 return size; 241 } 242 243 244 OopMap* OopMapSet::singular_oop_map() { 245 guarantee(om_count() == 1, "Make sure we only have a single gc point"); 246 return at(0); 247 } 248 249 250 OopMap* OopMapSet::find_map_at_offset(int pc_offset) const { 251 int i, len = om_count(); 252 assert( len > 0, "must have pointer maps" ); 253 254 // Scan through oopmaps. Stop when current offset is either equal or greater 255 // than the one we are looking for. 256 for( i = 0; i < len; i++) { 257 if( at(i)->offset() >= pc_offset ) 258 break; 259 } 260 261 assert( i < len, "oopmap not found" ); 262 263 OopMap* m = at(i); 264 assert( m->offset() == pc_offset, "oopmap not found" ); 265 return m; 266 } 267 268 class DoNothingClosure: public OopClosure { 269 public: 270 void do_oop(oop* p) {} 271 void do_oop(narrowOop* p) {} 272 }; 273 static DoNothingClosure do_nothing; 274 275 static void add_derived_oop(oop* base, oop* derived) { 276 #if !defined(TIERED) && !defined(INCLUDE_JVMCI) 277 COMPILER1_PRESENT(ShouldNotReachHere();) 278 #endif // !defined(TIERED) && !defined(INCLUDE_JVMCI) 279 #if defined(COMPILER2) || INCLUDE_JVMCI 280 DerivedPointerTable::add(derived, base); 281 #endif // COMPILER2 || INCLUDE_JVMCI 282 } 283 284 285 #ifndef PRODUCT 286 static void trace_codeblob_maps(const frame *fr, const RegisterMap *reg_map) { 287 // Print oopmap and regmap 288 tty->print_cr("------ "); 289 CodeBlob* cb = fr->cb(); 290 const ImmutableOopMapSet* maps = cb->oop_maps(); 291 const ImmutableOopMap* map = cb->oop_map_for_return_address(fr->pc()); 292 map->print(); 293 if( cb->is_nmethod() ) { 294 nmethod* nm = (nmethod*)cb; 295 // native wrappers have no scope data, it is implied 296 if (nm->is_native_method()) { 297 tty->print("bci: 0 (native)"); 298 } else { 299 ScopeDesc* scope = nm->scope_desc_at(fr->pc()); 300 tty->print("bci: %d ",scope->bci()); 301 } 302 } 303 tty->cr(); 304 fr->print_on(tty); 305 tty->print(" "); 306 cb->print_value_on(tty); tty->cr(); 307 reg_map->print(); 308 tty->print_cr("------ "); 309 310 } 311 #endif // PRODUCT 312 313 void OopMapSet::oops_do(const frame *fr, const RegisterMap* reg_map, OopClosure* f) { 314 // add derived oops to a table 315 all_do(fr, reg_map, f, add_derived_oop, &do_nothing); 316 } 317 318 319 void OopMapSet::all_do(const frame *fr, const RegisterMap *reg_map, 320 OopClosure* oop_fn, void derived_oop_fn(oop*, oop*), 321 OopClosure* value_fn) { 322 CodeBlob* cb = fr->cb(); 323 assert(cb != NULL, "no codeblob"); 324 325 NOT_PRODUCT(if (TraceCodeBlobStacks) trace_codeblob_maps(fr, reg_map);) 326 327 const ImmutableOopMapSet* maps = cb->oop_maps(); 328 const ImmutableOopMap* map = cb->oop_map_for_return_address(fr->pc()); 329 assert(map != NULL, "no ptr map found"); 330 331 // handle derived pointers first (otherwise base pointer may be 332 // changed before derived pointer offset has been collected) 333 OopMapValue omv; 334 { 335 OopMapStream oms(map,OopMapValue::derived_oop_value); 336 if (!oms.is_done()) { 337 #ifndef TIERED 338 COMPILER1_PRESENT(ShouldNotReachHere();) 339 #if INCLUDE_JVMCI 340 if (UseJVMCICompiler) { 341 ShouldNotReachHere(); 342 } 343 #endif 344 #endif // !TIERED 345 // Protect the operation on the derived pointers. This 346 // protects the addition of derived pointers to the shared 347 // derived pointer table in DerivedPointerTable::add(). 348 MutexLockerEx x(DerivedPointerTableGC_lock, Mutex::_no_safepoint_check_flag); 349 do { 350 omv = oms.current(); 351 oop* loc = fr->oopmapreg_to_location(omv.reg(),reg_map); 352 guarantee(loc != NULL, "missing saved register"); 353 oop *derived_loc = loc; 354 oop *base_loc = fr->oopmapreg_to_location(omv.content_reg(), reg_map); 355 // Ignore NULL oops and decoded NULL narrow oops which 356 // equal to Universe::narrow_oop_base when a narrow oop 357 // implicit null check is used in compiled code. 358 // The narrow_oop_base could be NULL or be the address 359 // of the page below heap depending on compressed oops mode. 360 if (base_loc != NULL && *base_loc != (oop)NULL && !Universe::is_narrow_oop_base(*base_loc)) { 361 derived_oop_fn(base_loc, derived_loc); 362 } 363 oms.next(); 364 } while (!oms.is_done()); 365 } 366 } 367 368 // We want coop and oop oop_types 369 int mask = OopMapValue::oop_value | OopMapValue::narrowoop_value; 370 { 371 for (OopMapStream oms(map,mask); !oms.is_done(); oms.next()) { 372 omv = oms.current(); 373 oop* loc = fr->oopmapreg_to_location(omv.reg(),reg_map); 374 // It should be an error if no location can be found for a 375 // register mentioned as contained an oop of some kind. Maybe 376 // this was allowed previously because value_value items might 377 // be missing? 378 guarantee(loc != NULL, "missing saved register"); 379 if ( omv.type() == OopMapValue::oop_value ) { 380 oop val = *loc; 381 if (val == (oop)NULL || Universe::is_narrow_oop_base(val)) { 382 // Ignore NULL oops and decoded NULL narrow oops which 383 // equal to Universe::narrow_oop_base when a narrow oop 384 // implicit null check is used in compiled code. 385 // The narrow_oop_base could be NULL or be the address 386 // of the page below heap depending on compressed oops mode. 387 continue; 388 } 389 #ifdef ASSERT 390 if ((((uintptr_t)loc & (sizeof(*loc)-1)) != 0) || 391 !Universe::heap()->is_in_or_null(*loc)) { 392 tty->print_cr("# Found non oop pointer. Dumping state at failure"); 393 // try to dump out some helpful debugging information 394 trace_codeblob_maps(fr, reg_map); 395 omv.print(); 396 tty->print_cr("register r"); 397 omv.reg()->print(); 398 tty->print_cr("loc = %p *loc = %p\n", loc, (address)*loc); 399 // do the real assert. 400 assert(Universe::heap()->is_in_or_null(*loc), "found non oop pointer"); 401 } 402 #endif // ASSERT 403 oop_fn->do_oop(loc); 404 } else if ( omv.type() == OopMapValue::narrowoop_value ) { 405 narrowOop *nl = (narrowOop*)loc; 406 #ifndef VM_LITTLE_ENDIAN 407 VMReg vmReg = omv.reg(); 408 // Don't do this on SPARC float registers as they can be individually addressed 409 if (!vmReg->is_stack() SPARC_ONLY(&& !vmReg->is_FloatRegister())) { 410 // compressed oops in registers only take up 4 bytes of an 411 // 8 byte register but they are in the wrong part of the 412 // word so adjust loc to point at the right place. 413 nl = (narrowOop*)((address)nl + 4); 414 } 415 #endif 416 oop_fn->do_oop(nl); 417 } 418 } 419 } 420 } 421 422 423 // Update callee-saved register info for the following frame 424 void OopMapSet::update_register_map(const frame *fr, RegisterMap *reg_map) { 425 ResourceMark rm; 426 CodeBlob* cb = fr->cb(); 427 assert(cb != NULL, "no codeblob"); 428 429 // Any reg might be saved by a safepoint handler (see generate_handler_blob). 430 assert( reg_map->_update_for_id == NULL || fr->is_older(reg_map->_update_for_id), 431 "already updated this map; do not 'update' it twice!" ); 432 debug_only(reg_map->_update_for_id = fr->id()); 433 434 // Check if caller must update oop argument 435 assert((reg_map->include_argument_oops() || 436 !cb->caller_must_gc_arguments(reg_map->thread())), 437 "include_argument_oops should already be set"); 438 439 // Scan through oopmap and find location of all callee-saved registers 440 // (we do not do update in place, since info could be overwritten) 441 442 address pc = fr->pc(); 443 const ImmutableOopMap* map = cb->oop_map_for_return_address(pc); 444 assert(map != NULL, "no ptr map found"); 445 DEBUG_ONLY(int nof_callee = 0;) 446 447 for (OopMapStream oms(map, OopMapValue::callee_saved_value); !oms.is_done(); oms.next()) { 448 OopMapValue omv = oms.current(); 449 VMReg reg = omv.content_reg(); 450 oop* loc = fr->oopmapreg_to_location(omv.reg(), reg_map); 451 reg_map->set_location(reg, (address) loc); 452 DEBUG_ONLY(nof_callee++;) 453 } 454 455 // Check that runtime stubs save all callee-saved registers 456 #ifdef COMPILER2 457 assert(cb->is_compiled_by_c1() || cb->is_compiled_by_jvmci() || !cb->is_runtime_stub() || 458 (nof_callee >= SAVED_ON_ENTRY_REG_COUNT || nof_callee >= C_SAVED_ON_ENTRY_REG_COUNT), 459 "must save all"); 460 #endif // COMPILER2 461 } 462 463 //============================================================================= 464 // Non-Product code 465 466 #ifndef PRODUCT 467 468 bool ImmutableOopMap::has_derived_pointer() const { 469 #if !defined(TIERED) && !defined(INCLUDE_JVMCI) 470 COMPILER1_PRESENT(return false); 471 #endif // !TIERED 472 #if defined(COMPILER2) || INCLUDE_JVMCI 473 OopMapStream oms(this,OopMapValue::derived_oop_value); 474 return oms.is_done(); 475 #else 476 return false; 477 #endif // COMPILER2 || INCLUDE_JVMCI 478 } 479 480 #endif //PRODUCT 481 482 // Printing code is present in product build for -XX:+PrintAssembly. 483 484 static 485 void print_register_type(OopMapValue::oop_types x, VMReg optional, 486 outputStream* st) { 487 switch( x ) { 488 case OopMapValue::oop_value: 489 st->print("Oop"); 490 break; 491 case OopMapValue::narrowoop_value: 492 st->print("NarrowOop"); 493 break; 494 case OopMapValue::callee_saved_value: 495 st->print("Callers_"); 496 optional->print_on(st); 497 break; 498 case OopMapValue::derived_oop_value: 499 st->print("Derived_oop_"); 500 optional->print_on(st); 501 break; 502 default: 503 ShouldNotReachHere(); 504 } 505 } 506 507 void OopMapValue::print_on(outputStream* st) const { 508 reg()->print_on(st); 509 st->print("="); 510 print_register_type(type(),content_reg(),st); 511 st->print(" "); 512 } 513 514 void ImmutableOopMap::print_on(outputStream* st) const { 515 OopMapValue omv; 516 st->print("ImmutableOopMap{"); 517 for(OopMapStream oms(this); !oms.is_done(); oms.next()) { 518 omv = oms.current(); 519 omv.print_on(st); 520 } 521 st->print("}"); 522 } 523 524 void OopMap::print_on(outputStream* st) const { 525 OopMapValue omv; 526 st->print("OopMap{"); 527 for(OopMapStream oms((OopMap*)this); !oms.is_done(); oms.next()) { 528 omv = oms.current(); 529 omv.print_on(st); 530 } 531 st->print("off=%d}", (int) offset()); 532 } 533 534 void ImmutableOopMapSet::print_on(outputStream* st) const { 535 const ImmutableOopMap* last = NULL; 536 for (int i = 0; i < _count; ++i) { 537 const ImmutableOopMapPair* pair = pair_at(i); 538 const ImmutableOopMap* map = pair->get_from(this); 539 if (map != last) { 540 st->cr(); 541 map->print_on(st); 542 st->print("pc offsets: "); 543 } 544 last = map; 545 st->print("%d ", pair->pc_offset()); 546 } 547 } 548 549 void OopMapSet::print_on(outputStream* st) const { 550 int i, len = om_count(); 551 552 st->print_cr("OopMapSet contains %d OopMaps\n",len); 553 554 for( i = 0; i < len; i++) { 555 OopMap* m = at(i); 556 st->print_cr("#%d ",i); 557 m->print_on(st); 558 st->cr(); 559 } 560 } 561 562 bool OopMap::equals(const OopMap* other) const { 563 if (other->_omv_count != _omv_count) { 564 return false; 565 } 566 if (other->write_stream()->position() != write_stream()->position()) { 567 return false; 568 } 569 if (memcmp(other->write_stream()->buffer(), write_stream()->buffer(), write_stream()->position()) != 0) { 570 return false; 571 } 572 return true; 573 } 574 575 const ImmutableOopMap* ImmutableOopMapSet::find_map_at_offset(int pc_offset) const { 576 ImmutableOopMapPair* pairs = get_pairs(); 577 ImmutableOopMapPair* last = NULL; 578 579 for (int i = 0; i < _count; ++i) { 580 if (pairs[i].pc_offset() >= pc_offset) { 581 last = &pairs[i]; 582 break; 583 } 584 } 585 586 assert(last->pc_offset() == pc_offset, "oopmap not found"); 587 return last->get_from(this); 588 } 589 590 const ImmutableOopMap* ImmutableOopMapPair::get_from(const ImmutableOopMapSet* set) const { 591 return set->oopmap_at_offset(_oopmap_offset); 592 } 593 594 ImmutableOopMap::ImmutableOopMap(const OopMap* oopmap) : _count(oopmap->count()) { 595 address addr = data_addr(); 596 oopmap->copy_data_to(addr); 597 } 598 599 #ifdef ASSERT 600 int ImmutableOopMap::nr_of_bytes() const { 601 OopMapStream oms(this); 602 603 while (!oms.is_done()) { 604 oms.next(); 605 } 606 return sizeof(ImmutableOopMap) + oms.stream_position(); 607 } 608 #endif 609 610 ImmutableOopMapBuilder::ImmutableOopMapBuilder(const OopMapSet* set) : _set(set), _new_set(NULL), _empty(NULL), _last(NULL), _empty_offset(-1), _last_offset(-1), _offset(0), _required(-1) { 611 _mapping = NEW_RESOURCE_ARRAY(Mapping, _set->size()); 612 } 613 614 int ImmutableOopMapBuilder::size_for(const OopMap* map) const { 615 return align_up((int)sizeof(ImmutableOopMap) + map->data_size(), 8); 616 } 617 618 int ImmutableOopMapBuilder::heap_size() { 619 int base = sizeof(ImmutableOopMapSet); 620 base = align_up(base, 8); 621 622 // all of ours pc / offset pairs 623 int pairs = _set->size() * sizeof(ImmutableOopMapPair); 624 pairs = align_up(pairs, 8); 625 626 for (int i = 0; i < _set->size(); ++i) { 627 int size = 0; 628 OopMap* map = _set->at(i); 629 630 if (is_empty(map)) { 631 /* only keep a single empty map in the set */ 632 if (has_empty()) { 633 _mapping[i].set(Mapping::OOPMAP_EMPTY, _empty_offset, 0, map, _empty); 634 } else { 635 _empty_offset = _offset; 636 _empty = map; 637 size = size_for(map); 638 _mapping[i].set(Mapping::OOPMAP_NEW, _offset, size, map); 639 } 640 } else if (is_last_duplicate(map)) { 641 /* if this entry is identical to the previous one, just point it there */ 642 _mapping[i].set(Mapping::OOPMAP_DUPLICATE, _last_offset, 0, map, _last); 643 } else { 644 /* not empty, not an identical copy of the previous entry */ 645 size = size_for(map); 646 _mapping[i].set(Mapping::OOPMAP_NEW, _offset, size, map); 647 _last_offset = _offset; 648 _last = map; 649 } 650 651 assert(_mapping[i]._map == map, "check"); 652 _offset += size; 653 } 654 655 int total = base + pairs + _offset; 656 DEBUG_ONLY(total += 8); 657 _required = total; 658 return total; 659 } 660 661 void ImmutableOopMapBuilder::fill_pair(ImmutableOopMapPair* pair, const OopMap* map, int offset, const ImmutableOopMapSet* set) { 662 assert(offset < set->nr_of_bytes(), "check"); 663 new ((address) pair) ImmutableOopMapPair(map->offset(), offset); 664 } 665 666 int ImmutableOopMapBuilder::fill_map(ImmutableOopMapPair* pair, const OopMap* map, int offset, const ImmutableOopMapSet* set) { 667 fill_pair(pair, map, offset, set); 668 address addr = (address) pair->get_from(_new_set); // location of the ImmutableOopMap 669 670 new (addr) ImmutableOopMap(map); 671 return size_for(map); 672 } 673 674 void ImmutableOopMapBuilder::fill(ImmutableOopMapSet* set, int sz) { 675 ImmutableOopMapPair* pairs = set->get_pairs(); 676 677 for (int i = 0; i < set->count(); ++i) { 678 const OopMap* map = _mapping[i]._map; 679 ImmutableOopMapPair* pair = NULL; 680 int size = 0; 681 682 if (_mapping[i]._kind == Mapping::OOPMAP_NEW) { 683 size = fill_map(&pairs[i], map, _mapping[i]._offset, set); 684 } else if (_mapping[i]._kind == Mapping::OOPMAP_DUPLICATE || _mapping[i]._kind == Mapping::OOPMAP_EMPTY) { 685 fill_pair(&pairs[i], map, _mapping[i]._offset, set); 686 } 687 688 const ImmutableOopMap* nv = set->find_map_at_offset(map->offset()); 689 assert(memcmp(map->data(), nv->data_addr(), map->data_size()) == 0, "check identity"); 690 } 691 } 692 693 #ifdef ASSERT 694 void ImmutableOopMapBuilder::verify(address buffer, int size, const ImmutableOopMapSet* set) { 695 for (int i = 0; i < 8; ++i) { 696 assert(buffer[size - 8 + i] == (unsigned char) 0xff, "overwritten memory check"); 697 } 698 699 for (int i = 0; i < set->count(); ++i) { 700 const ImmutableOopMapPair* pair = set->pair_at(i); 701 assert(pair->oopmap_offset() < set->nr_of_bytes(), "check size"); 702 const ImmutableOopMap* map = pair->get_from(set); 703 int nr_of_bytes = map->nr_of_bytes(); 704 assert(pair->oopmap_offset() + nr_of_bytes <= set->nr_of_bytes(), "check size + size"); 705 } 706 } 707 #endif 708 709 ImmutableOopMapSet* ImmutableOopMapBuilder::generate_into(address buffer) { 710 DEBUG_ONLY(memset(&buffer[_required-8], 0xff, 8)); 711 712 _new_set = new (buffer) ImmutableOopMapSet(_set, _required); 713 fill(_new_set, _required); 714 715 DEBUG_ONLY(verify(buffer, _required, _new_set)); 716 717 return _new_set; 718 } 719 720 ImmutableOopMapSet* ImmutableOopMapBuilder::build() { 721 _required = heap_size(); 722 723 // We need to allocate a chunk big enough to hold the ImmutableOopMapSet and all of its ImmutableOopMaps 724 address buffer = (address) NEW_C_HEAP_ARRAY(unsigned char, _required, mtCode); 725 return generate_into(buffer); 726 } 727 728 ImmutableOopMapSet* ImmutableOopMapSet::build_from(const OopMapSet* oopmap_set) { 729 ResourceMark mark; 730 ImmutableOopMapBuilder builder(oopmap_set); 731 return builder.build(); 732 } 733 734 735 //------------------------------DerivedPointerTable--------------------------- 736 737 #if defined(COMPILER2) || INCLUDE_JVMCI 738 739 class DerivedPointerEntry : public CHeapObj<mtCompiler> { 740 private: 741 oop* _location; // Location of derived pointer (also pointing to the base) 742 intptr_t _offset; // Offset from base pointer 743 public: 744 DerivedPointerEntry(oop* location, intptr_t offset) { _location = location; _offset = offset; } 745 oop* location() { return _location; } 746 intptr_t offset() { return _offset; } 747 }; 748 749 750 GrowableArray<DerivedPointerEntry*>* DerivedPointerTable::_list = NULL; 751 bool DerivedPointerTable::_active = false; 752 753 754 void DerivedPointerTable::clear() { 755 // The first time, we create the list. Otherwise it should be 756 // empty. If not, then we have probably forgotton to call 757 // update_pointers after last GC/Scavenge. 758 assert (!_active, "should not be active"); 759 assert(_list == NULL || _list->length() == 0, "table not empty"); 760 if (_list == NULL) { 761 _list = new (ResourceObj::C_HEAP, mtCompiler) GrowableArray<DerivedPointerEntry*>(10, true); // Allocated on C heap 762 } 763 _active = true; 764 } 765 766 767 // Returns value of location as an int 768 intptr_t value_of_loc(oop *pointer) { return cast_from_oop<intptr_t>((*pointer)); } 769 770 771 void DerivedPointerTable::add(oop *derived_loc, oop *base_loc) { 772 assert(Universe::heap()->is_in_or_null(*base_loc), "not an oop"); 773 assert(derived_loc != base_loc, "Base and derived in same location"); 774 if (_active) { 775 assert(*derived_loc != (oop)base_loc, "location already added"); 776 assert(_list != NULL, "list must exist"); 777 intptr_t offset = value_of_loc(derived_loc) - value_of_loc(base_loc); 778 // This assert is invalid because derived pointers can be 779 // arbitrarily far away from their base. 780 // assert(offset >= -1000000, "wrong derived pointer info"); 781 782 if (TraceDerivedPointers) { 783 tty->print_cr( 784 "Add derived pointer@" INTPTR_FORMAT 785 " - Derived: " INTPTR_FORMAT 786 " Base: " INTPTR_FORMAT " (@" INTPTR_FORMAT ") (Offset: " INTX_FORMAT ")", 787 p2i(derived_loc), p2i((address)*derived_loc), p2i((address)*base_loc), p2i(base_loc), offset 788 ); 789 } 790 // Set derived oop location to point to base. 791 *derived_loc = (oop)base_loc; 792 assert_lock_strong(DerivedPointerTableGC_lock); 793 DerivedPointerEntry *entry = new DerivedPointerEntry(derived_loc, offset); 794 _list->append(entry); 795 } 796 } 797 798 799 void DerivedPointerTable::update_pointers() { 800 assert(_list != NULL, "list must exist"); 801 for(int i = 0; i < _list->length(); i++) { 802 DerivedPointerEntry* entry = _list->at(i); 803 oop* derived_loc = entry->location(); 804 intptr_t offset = entry->offset(); 805 // The derived oop was setup to point to location of base 806 oop base = **(oop**)derived_loc; 807 assert(Universe::heap()->is_in_or_null(base), "must be an oop"); 808 809 *derived_loc = (oop)(((address)base) + offset); 810 assert(value_of_loc(derived_loc) - value_of_loc(&base) == offset, "sanity check"); 811 812 if (TraceDerivedPointers) { 813 tty->print_cr("Updating derived pointer@" INTPTR_FORMAT 814 " - Derived: " INTPTR_FORMAT " Base: " INTPTR_FORMAT " (Offset: " INTX_FORMAT ")", 815 p2i(derived_loc), p2i((address)*derived_loc), p2i((address)base), offset); 816 } 817 818 // Delete entry 819 delete entry; 820 _list->at_put(i, NULL); 821 } 822 // Clear list, so it is ready for next traversal (this is an invariant) 823 if (TraceDerivedPointers && !_list->is_empty()) { 824 tty->print_cr("--------------------------"); 825 } 826 _list->clear(); 827 _active = false; 828 } 829 830 #endif // COMPILER2 || INCLUDE_JVMCI