< prev index next >

src/hotspot/cpu/arm/gc/shared/cardTableBarrierSetAssembler_arm.cpp

Print this page
rev 49939 : 8201786: Modularize interpreter GC barriers: leftovers for ARM32
Reviewed-by: enevill, eosterlund

*** 70,74 **** --- 70,178 ---- __ strb(zero, Address(addr, 1, post_indexed)); __ subs(count, count, 1); __ b(L_cardtable_loop, ge); __ BIND(L_done); } + + void CardTableBarrierSetAssembler::oop_store_at(MacroAssembler* masm, DecoratorSet decorators, BasicType type, + Address obj, Register new_val, Register tmp1, Register tmp2, Register tmp3, bool is_null) { + bool on_array = (decorators & IN_HEAP_ARRAY) != 0; + bool on_anonymous = (decorators & ON_UNKNOWN_OOP_REF) != 0; + bool precise = on_array || on_anonymous; + + if (is_null) { + BarrierSetAssembler::store_at(masm, decorators, type, obj, new_val, tmp1, tmp2, tmp3, true); + } else { + assert (!precise || (obj.index() == noreg && obj.disp() == 0), + "store check address should be calculated beforehand"); + + store_check_part1(masm, tmp1); + BarrierSetAssembler::store_at(masm, decorators, type, obj, new_val, tmp1, tmp2, tmp3, false); + new_val = noreg; + store_check_part2(masm, obj.base(), tmp1, tmp2); + } + } + + // The 1st part of the store check. + // Sets card_table_base register. + void CardTableBarrierSetAssembler::store_check_part1(MacroAssembler* masm, Register card_table_base) { + // Check barrier set type (should be card table) and element size + BarrierSet* bs = BarrierSet::barrier_set(); + assert(bs->kind() == BarrierSet::CardTableBarrierSet, + "Wrong barrier set kind"); + + CardTableBarrierSet* ctbs = barrier_set_cast<CardTableBarrierSet>(bs); + CardTable* ct = ctbs->card_table(); + assert(sizeof(*ct->byte_map_base()) == sizeof(jbyte), "Adjust store check code"); + + // Load card table base address. + + /* Performance note. + + There is an alternative way of loading card table base address + from thread descriptor, which may look more efficient: + + ldr(card_table_base, Address(Rthread, JavaThread::card_table_base_offset())); + + However, performance measurements of micro benchmarks and specJVM98 + showed that loading of card table base from thread descriptor is + 7-18% slower compared to loading of literal embedded into the code. + Possible cause is a cache miss (card table base address resides in a + rarely accessed area of thread descriptor). + */ + // TODO-AARCH64 Investigate if mov_slow is faster than ldr from Rthread on AArch64 + __ mov_address(card_table_base, (address)ct->byte_map_base(), symbolic_Relocation::card_table_reference); + } + + // The 2nd part of the store check. + void CardTableBarrierSetAssembler::store_check_part2(MacroAssembler* masm, Register obj, Register card_table_base, Register tmp) { + assert_different_registers(obj, card_table_base, tmp); + + assert(CardTable::dirty_card_val() == 0, "Dirty card value must be 0 due to optimizations."); + #ifdef AARCH64 + add(card_table_base, card_table_base, AsmOperand(obj, lsr, CardTable::card_shift)); + Address card_table_addr(card_table_base); + #else + Address card_table_addr(card_table_base, obj, lsr, CardTable::card_shift); + #endif + + if (UseCondCardMark) { + #if INCLUDE_ALL_GCS + if (UseConcMarkSweepGC) { + __ membar(MacroAssembler::Membar_mask_bits(MacroAssembler::StoreLoad), noreg); + } + #endif + Label already_dirty; + + __ ldrb(tmp, card_table_addr); + __ cbz(tmp, already_dirty); + + set_card(masm, card_table_base, card_table_addr, tmp); + __ bind(already_dirty); + + } else { + #if INCLUDE_ALL_GCS + if (UseConcMarkSweepGC && CMSPrecleaningEnabled) { + __ membar(MacroAssembler::Membar_mask_bits(MacroAssembler::StoreStore), noreg); + } + #endif + set_card(masm, card_table_base, card_table_addr, tmp); + } + } + + void CardTableBarrierSetAssembler::set_card(MacroAssembler* masm, Register card_table_base, Address card_table_addr, Register tmp) { + #ifdef AARCH64 + strb(ZR, card_table_addr); + #else + CardTableBarrierSet* ctbs = barrier_set_cast<CardTableBarrierSet>(BarrierSet::barrier_set()); + CardTable* ct = ctbs->card_table(); + if ((((uintptr_t)ct->byte_map_base() & 0xff) == 0)) { + // Card table is aligned so the lowest byte of the table address base is zero. + // This works only if the code is not saved for later use, possibly + // in a context where the base would no longer be aligned. + __ strb(card_table_base, card_table_addr); + } else { + __ mov(tmp, 0); + __ strb(tmp, card_table_addr); + } + #endif // AARCH64 + }
< prev index next >