/* * Copyright (c) 2016, 2018, Oracle and/or its affiliates. All rights reserved. * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. * * This code is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License version 2 only, as * published by the Free Software Foundation. * * This code is distributed in the hope that it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License * version 2 for more details (a copy is included in the LICENSE file that * accompanied this code). * * You should have received a copy of the GNU General Public License version * 2 along with this work; if not, write to the Free Software Foundation, * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. * * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA * or visit www.oracle.com if you need additional information or have any * questions. * */ #ifndef SHARE_VM_GC_G1_G1BARRIERSET_INLINE_HPP #define SHARE_VM_GC_G1_G1BARRIERSET_INLINE_HPP #include "gc/g1/g1BarrierSet.hpp" #include "gc/g1/g1CardTable.hpp" #include "gc/shared/accessBarrierSupport.inline.hpp" #include "oops/access.inline.hpp" #include "oops/compressedOops.inline.hpp" #include "oops/oop.hpp" template inline void G1BarrierSet::write_ref_field_pre(T* field) { if (HasDecorator::value || HasDecorator::value) { return; } T heap_oop = RawAccess::oop_load(field); if (!CompressedOops::is_null(heap_oop)) { enqueue(CompressedOops::decode_not_null(heap_oop)); } } template inline void G1BarrierSet::write_ref_field_post(T* field, oop new_val) { volatile jbyte* byte = _card_table->byte_for(field); if (*byte != G1CardTable::g1_young_card_val()) { // Take a slow path for cards in old write_ref_field_post_slow(byte); } } inline void G1BarrierSet::enqueue_if_weak(DecoratorSet decorators, oop value) { assert((decorators & ON_UNKNOWN_OOP_REF) == 0, "Reference strength must be known"); // Loading from a weak or phantom reference needs enqueueing, as // the object may not have been reachable (part of the snapshot) // when marking started. const bool on_strong_oop_ref = (decorators & ON_STRONG_OOP_REF) != 0; const bool peek = (decorators & AS_NO_KEEPALIVE) != 0; const bool needs_enqueue = (!peek && !on_strong_oop_ref); if (needs_enqueue && value != NULL) { enqueue(value); } } template template inline oop G1BarrierSet::AccessBarrier:: oop_load_not_in_heap(T* addr) { oop value = ModRef::oop_load_not_in_heap(addr); enqueue_if_weak(decorators, value); return value; } template template inline oop G1BarrierSet::AccessBarrier:: oop_load_in_heap(T* addr) { oop value = ModRef::oop_load_in_heap(addr); enqueue_if_weak(decorators, value); return value; } template inline oop G1BarrierSet::AccessBarrier:: oop_load_in_heap_at(oop base, ptrdiff_t offset) { oop value = ModRef::oop_load_in_heap_at(base, offset); enqueue_if_weak(AccessBarrierSupport::resolve_possibly_unknown_oop_ref_strength(base, offset), value); return value; } template template inline void G1BarrierSet::AccessBarrier:: oop_store_not_in_heap(T* addr, oop new_value) { // Apply SATB barriers for all non-heap references, to allow // concurrent scanning of such references. G1BarrierSet *bs = barrier_set_cast(BarrierSet::barrier_set()); bs->write_ref_field_pre(addr); Raw::oop_store(addr, new_value); } #endif // SHARE_VM_GC_G1_G1BARRIERSET_INLINE_HPP