< prev index next >
src/hotspot/share/runtime/objectMonitor.inline.hpp
Print this page
rev 54415 : 8222295: more baseline cleanups from Async Monitor Deflation project
rev 54416 : Checkpoint latest preliminary review patches for full OpenJDK review; merge with 8222295.patch.
rev 54417 : imported patch dcubed.monitor_deflate_conc.v2.01
*** 54,86 ****
void* owner = _owner;
return owner != DEFLATER_MARKER ? owner : NULL;
}
inline void ObjectMonitor::clear() {
assert(_contentions == 0, "must be 0: contentions=%d", _contentions);
assert(_owner == NULL, "must be NULL: owner=" INTPTR_FORMAT, p2i(_owner));
clear_using_JT();
}
inline void ObjectMonitor::clear_using_JT() {
! // When clearing using a JavaThread, we leave _owner == DEFLATER_MARKER
! // and _contentions < 0 to force any racing threads to retry. Unlike other
! // *_using_JT() functions, we cannot assert AsyncDeflateIdleMonitors
! // or Thread::current()->is_Java_thread() because clear() calls this
! // function for the rest of its checks.
!
! assert(_header != NULL, "must be non-NULL");
assert(_waiters == 0, "must be 0: waiters=%d", _waiters);
assert(_recursions == 0, "must be 0: recursions=" INTPTR_FORMAT, _recursions);
assert(_object != NULL, "must be non-NULL");
// Do not assert _ref_count == 0 here because a racing thread could
// increment _ref_count, observe _owner == DEFLATER_MARKER and then
// decrement _ref_count.
set_allocation_state(Free);
- _header = NULL;
_object = NULL;
// Do not clear _ref_count here because _ref_count is for indicating
// that the ObjectMonitor* is in use which is orthogonal to whether
// the ObjectMonitor itself is in use for a locking operation.
}
--- 54,98 ----
void* owner = _owner;
return owner != DEFLATER_MARKER ? owner : NULL;
}
inline void ObjectMonitor::clear() {
+ assert(_header != NULL, "must be non-NULL");
assert(_contentions == 0, "must be 0: contentions=%d", _contentions);
assert(_owner == NULL, "must be NULL: owner=" INTPTR_FORMAT, p2i(_owner));
+ _header = NULL;
+
clear_using_JT();
}
inline void ObjectMonitor::clear_using_JT() {
! // Unlike other *_using_JT() functions, we cannot assert
! // AsyncDeflateIdleMonitors or Thread::current()->is_Java_thread()
! // because clear() calls this function for the rest of its checks.
!
! if (AsyncDeflateIdleMonitors) {
! // Async deflation protocol uses the _header, _contentions and _owner
! // fields. While the ObjectMonitor being deflated is on the global free
! // list, we leave those three fields alone; _owner == DEFLATER_MARKER
! // and _contentions < 0 will force any racing threads to retry. The
! // _header field is used by install_displaced_markword_in_object()
! // in the last part of the deflation protocol so we cannot check
! // its values here.
! guarantee(_owner == NULL || _owner == DEFLATER_MARKER,
! "must be NULL or DEFLATER_MARKER: owner=" INTPTR_FORMAT,
! p2i(_owner));
! guarantee(_contentions <= 0, "must be <= 0: contentions=%d", _contentions);
! }
assert(_waiters == 0, "must be 0: waiters=%d", _waiters);
assert(_recursions == 0, "must be 0: recursions=" INTPTR_FORMAT, _recursions);
assert(_object != NULL, "must be non-NULL");
// Do not assert _ref_count == 0 here because a racing thread could
// increment _ref_count, observe _owner == DEFLATER_MARKER and then
// decrement _ref_count.
set_allocation_state(Free);
_object = NULL;
// Do not clear _ref_count here because _ref_count is for indicating
// that the ObjectMonitor* is in use which is orthogonal to whether
// the ObjectMonitor itself is in use for a locking operation.
}
*** 146,179 ****
inline bool ObjectMonitor::is_new() const {
return _allocation_state == New;
}
inline void ObjectMonitor::dec_ref_count() {
! // The decrement needs to be MO_ACQ_REL. At the moment, the Atomic::dec
! // backend on PPC does not yet conform to these requirements. Therefore
! // the decrement is simulated with an Atomic::sub(1, &addr). Without
! // this MO_ACQ_REL Atomic::dec simulation, AsyncDeflateIdleMonitors is
! // not safe.
! Atomic::sub((jint)1, &_ref_count);
guarantee(_ref_count >= 0, "sanity check: ref_count=%d", _ref_count);
}
inline void ObjectMonitor::inc_ref_count() {
! // The increment needs to be MO_SEQ_CST. At the moment, the Atomic::inc
! // backend on PPC does not yet conform to these requirements. Therefore
! // the increment is simulated with a load phi; cas phi + 1; loop.
! // Without this MO_SEQ_CST Atomic::inc simulation, AsyncDeflateIdleMonitors
! // is not safe.
! for (;;) {
! jint sample = OrderAccess::load_acquire(&_ref_count);
! guarantee(sample >= 0, "sanity check: sample=%d", (int)sample);
! if (Atomic::cmpxchg(sample + 1, &_ref_count, sample) == sample) {
! // Incremented _ref_count without interference.
! return;
! }
! // Implied else: Saw interference so loop and try again.
! }
}
inline jint ObjectMonitor::ref_count() const {
return OrderAccess::load_acquire(&_ref_count);
}
--- 158,179 ----
inline bool ObjectMonitor::is_new() const {
return _allocation_state == New;
}
inline void ObjectMonitor::dec_ref_count() {
! // The decrement only needs to be MO_ACQ_REL since the reference
! // counter is volatile.
! Atomic::dec(&_ref_count);
guarantee(_ref_count >= 0, "sanity check: ref_count=%d", _ref_count);
}
inline void ObjectMonitor::inc_ref_count() {
! // The increment needs to be MO_SEQ_CST so that the reference
! // counter update is seen as soon as possible in a race with the
! // async deflation protocol.
! Atomic::inc(&_ref_count);
! guarantee(_ref_count > 0, "sanity check: ref_count=%d", _ref_count);
}
inline jint ObjectMonitor::ref_count() const {
return OrderAccess::load_acquire(&_ref_count);
}
< prev index next >