< prev index next >
src/hotspot/share/runtime/objectMonitor.inline.hpp
Print this page
rev 54110 : Checkpoint latest preliminary review patches for full OpenJDK review.
*** 51,74 ****
inline jint ObjectMonitor::waiters() const {
return _waiters;
}
inline void* ObjectMonitor::owner() const {
! return _owner;
}
inline void ObjectMonitor::clear() {
- assert(_header != NULL, "Fatal logic error in ObjectMonitor header!");
assert(_count == 0, "Fatal logic error in ObjectMonitor count!");
assert(_waiters == 0, "Fatal logic error in ObjectMonitor waiters!");
assert(_recursions == 0, "Fatal logic error in ObjectMonitor recursions!");
assert(_object != NULL, "Fatal logic error in ObjectMonitor object!");
! assert(_owner == NULL, "Fatal logic error in ObjectMonitor owner!");
_header = NULL;
_object = NULL;
}
inline void* ObjectMonitor::object() const {
return _object;
}
--- 51,94 ----
inline jint ObjectMonitor::waiters() const {
return _waiters;
}
+ // Returns NULL if DEFLATER_MARKER is observed.
inline void* ObjectMonitor::owner() const {
! void* owner = _owner;
! return owner != DEFLATER_MARKER ? owner : NULL;
}
inline void ObjectMonitor::clear() {
assert(_count == 0, "Fatal logic error in ObjectMonitor count!");
+ assert(_owner == NULL, "Fatal logic error in ObjectMonitor owner!");
+
+ clear_using_JT();
+ }
+
+ inline void ObjectMonitor::clear_using_JT() {
+ // When clearing using a JavaThread, we leave _owner == DEFLATER_MARKER
+ // and _count < 0 to force any racing threads to retry. Unlike other
+ // *_using_JT() functions, we cannot assert AsyncDeflateIdleMonitors
+ // or Thread::current()->is_Java_thread() because clear() calls this
+ // function for the rest of its checks.
+
+ assert(_header != NULL, "Fatal logic error in ObjectMonitor header!");
assert(_waiters == 0, "Fatal logic error in ObjectMonitor waiters!");
assert(_recursions == 0, "Fatal logic error in ObjectMonitor recursions!");
assert(_object != NULL, "Fatal logic error in ObjectMonitor object!");
! // Do not assert _ref_count == 0 here because a racing thread could
! // increment _ref_count, observe _owner == DEFLATER_MARKER and then
! // decrement _ref_count.
+ set_allocation_state(Free);
_header = NULL;
_object = NULL;
+ // Do not clear _ref_count here because _ref_count is for indicating
+ // that the ObjectMonitor* is in use which is orthogonal to whether
+ // the ObjectMonitor itself is in use for a locking operation.
}
inline void* ObjectMonitor::object() const {
return _object;
}
*** 105,110 ****
--- 125,185 ----
inline void ObjectMonitor::set_owner(void* owner) {
_owner = owner;
_recursions = 0;
}
+ inline void ObjectMonitor::set_allocation_state(ObjectMonitor::AllocationState s) {
+ _allocation_state = s;
+ }
+
+ inline ObjectMonitor::AllocationState ObjectMonitor::allocation_state() const {
+ return _allocation_state;
+ }
+
+ inline bool ObjectMonitor::is_free() const {
+ return _allocation_state == Free;
+ }
+
+ inline bool ObjectMonitor::is_active() const {
+ return !is_free();
+ }
+
+ inline bool ObjectMonitor::is_old() const {
+ return _allocation_state == Old;
+ }
+
+ inline bool ObjectMonitor::is_new() const {
+ return _allocation_state == New;
+ }
+
+ inline void ObjectMonitor::dec_ref_count() {
+ // The decrement needs to be MO_ACQ_REL. At the moment, the Atomic::dec
+ // backend on PPC does not yet conform to these requirements. Therefore
+ // the decrement is simulated with an Atomic::sub(1, &addr). Without
+ // this MO_ACQ_REL Atomic::dec simulation, AsyncDeflateIdleMonitors is
+ // not safe.
+ Atomic::sub((jint)1, &_ref_count);
+ guarantee(_ref_count >= 0, "sanity check: ref_count=%d", _ref_count);
+ }
+
+ inline void ObjectMonitor::inc_ref_count() {
+ // The increment needs to be MO_SEQ_CST. At the moment, the Atomic::inc
+ // backend on PPC does not yet conform to these requirements. Therefore
+ // the increment is simulated with a load phi; cas phi + 1; loop.
+ // Without this MO_SEQ_CST Atomic::inc simulation, AsyncDeflateIdleMonitors
+ // is not safe.
+ for (;;) {
+ jint sample = OrderAccess::load_acquire(&_ref_count);
+ guarantee(sample >= 0, "sanity check: sample=%d", (int)sample);
+ if (Atomic::cmpxchg(sample + 1, &_ref_count, sample) == sample) {
+ // Incremented _ref_count without interference.
+ return;
+ }
+ // Implied else: Saw interference so loop and try again.
+ }
+ }
+
+ inline jint ObjectMonitor::ref_count() const {
+ return OrderAccess::load_acquire(&_ref_count);
+ }
+
#endif // SHARE_RUNTIME_OBJECTMONITOR_INLINE_HPP
< prev index next >