< prev index next >

src/hotspot/share/runtime/objectMonitor.inline.hpp

Print this page
rev 56634 : imported patch 8230876.patch
rev 56635 : v2.00 -> v2.05 (CR5/v2.05/8-for-jdk13) patches combined into one; merge with 8229212.patch; merge with jdk-14+11; merge with 8230184.patch; merge with 8230876.patch; merge with jdk-14+15; merge with jdk-14+18.
rev 56639 : loosen a couple more counter checks due to races observed in testing; simplify om_release() extraction of mid since list head or cur_mid_in_use is marked; simplify deflate_monitor_list() extraction of mid since there are no parallel deleters due to the safepoint; simplify deflate_monitor_list_using_JT() extraction of mid since list head or cur_mid_in_use is marked; prepend_block_to_lists() - simplify based on David H's comments; does not need load_acquire() or release_store() because of the cmpxchg(); prepend_to_common() - simplify to use mark_next_loop() for m and use mark_list_head() and release_store() for the non-empty list case; add more debugging for "Non-balanced monitor enter/exit" failure mode; fix race in inflate() in the "CASE: neutral" code path; install_displaced_markword_in_object() does not need to clear the header field since that is handled when the ObjectMonitor is moved from the global free list; LSuccess should clear boxReg to set ICC.ZF=1 to avoid depending on existing boxReg contents; update fast_unlock() to detect when object no longer refers to the same ObjectMonitor and take fast path exit instead; clarify fast_lock() code where we detect when object no longer refers to the same ObjectMonitor; add/update comments for movptr() calls where we move a literal into an Address; remove set_owner(); refactor setting of owner field into set_owner_from(2 versions), set_owner_from_BasicLock(), and try_set_owner_from(); the new functions include monitorinflation+owner logging; extract debug code from v2.06 and v2.07 and move to v2.07.debug; change 'jccb' -> 'jcc' and 'jmpb' -> 'jmp' as needed; checkpoint initial version of MacroAssembler::inc_om_ref_count(); update LP64 MacroAssembler::fast_lock() and fast_unlock() to use inc_om_ref_count(); fast_lock() return flag setting logic can use 'testptr(tmpReg, tmpReg)' instead of 'cmpptr(tmpReg, 0)' since that's more efficient; fast_unlock() LSuccess return flag setting logic can use 'testl (boxReg, 0)' instead of 'xorptr(boxReg, boxReg)' since that's more efficient; cleanup "fast-path" vs "fast path" and "slow-path" vs "slow path"; update MacroAssembler::rtm_inflated_locking() to use inc_om_ref_count(); update MacroAssembler::fast_lock() to preserve the flags before decrementing ref_count and restore the flags afterwards; this is more clean than depending on the contents of rax/tmpReg; coleenp CR - refactor async monitor deflation work from ServiceThread::service_thread_entry() to ObjectSynchronizer::deflate_idle_monitors_using_JT(); rehn,eosterlund CR - add support for HandshakeAfterDeflateIdleMonitors for platforms that don't have ObjectMonitor ref_count support implemented in C2 fast_lock() and fast_unlock().

@@ -23,10 +23,11 @@
  */
 
 #ifndef SHARE_RUNTIME_OBJECTMONITOR_INLINE_HPP
 #define SHARE_RUNTIME_OBJECTMONITOR_INLINE_HPP
 
+#include "logging/log.hpp"
 #include "runtime/atomic.hpp"
 
 inline intptr_t ObjectMonitor::is_entered(TRAPS) const {
   if (THREAD == _owner || THREAD->is_lock_owned((address) _owner)) {
     return 1;

@@ -49,23 +50,57 @@
 
 inline jint ObjectMonitor::waiters() const {
   return _waiters;
 }
 
+// Returns NULL if DEFLATER_MARKER is observed.
 inline void* ObjectMonitor::owner() const {
-  return _owner;
+  void* owner = _owner;
+  return owner != DEFLATER_MARKER ? owner : NULL;
+}
+
+// Returns true if owner field == DEFLATER_MARKER and false otherwise.
+// This accessor is called when we really need to know if the owner
+// field == DEFLATER_MARKER and any non-NULL value won't do the trick.
+inline bool ObjectMonitor::owner_is_DEFLATER_MARKER() {
+  return OrderAccess::load_acquire(&_owner) == DEFLATER_MARKER;
 }
 
 inline void ObjectMonitor::clear() {
   assert(Atomic::load(&_header).value() != 0, "must be non-zero");
+  assert(_owner == NULL, "must be NULL: owner=" INTPTR_FORMAT, p2i(_owner));
+  assert(ref_count() == 0, "must be 0: ref_count=%d", ref_count());
+
+  Atomic::store(markWord::zero(), &_header);
+
+  clear_using_JT();
+}
+
+inline void ObjectMonitor::clear_using_JT() {
+  // Unlike other *_using_JT() functions, we cannot assert
+  // AsyncDeflateIdleMonitors or Thread::current()->is_Java_thread()
+  // because clear() calls this function for the rest of its checks.
+
+  if (AsyncDeflateIdleMonitors) {
+    // Async deflation protocol uses the header, owner and ref_count
+    // fields. While the ObjectMonitor being deflated is on the global free
+    // list, we leave those three fields alone; owner == DEFLATER_MARKER
+    // and ref_count < 0 will force any racing threads to retry. The
+    // header field is used by install_displaced_markword_in_object()
+    // in the last part of the deflation protocol so we cannot check
+    // its value here.
+    guarantee(_owner == NULL || _owner == DEFLATER_MARKER,
+              "must be NULL or DEFLATER_MARKER: owner=" INTPTR_FORMAT,
+              p2i(_owner));
+    guarantee(ref_count() <= 0, "must be <= 0: ref_count=%d", ref_count());
+  }
   assert(_contentions == 0, "must be 0: contentions=%d", _contentions);
   assert(_waiters == 0, "must be 0: waiters=%d", _waiters);
-  assert(_recursions == 0, "must be 0: recursions=" INTPTR_FORMAT, _recursions);
+  assert(_recursions == 0, "must be 0: recursions=" INTX_FORMAT, _recursions);
   assert(_object != NULL, "must be non-NULL");
-  assert(_owner == NULL, "must be NULL: owner=" INTPTR_FORMAT, p2i(_owner));
 
-  Atomic::store(markWord::zero(), &_header);
+  set_allocation_state(Free);
   _object = NULL;
 }
 
 inline void* ObjectMonitor::object() const {
   return _object;

@@ -82,10 +117,105 @@
 // return number of threads contending for this monitor
 inline jint ObjectMonitor::contentions() const {
   return _contentions;
 }
 
-inline void ObjectMonitor::set_owner(void* owner) {
-  _owner = owner;
+// Set _owner field to new_value; current value must match old_value.
+inline void ObjectMonitor::set_owner_from(void* new_value, void* old_value) {
+  void* prev = Atomic::cmpxchg(new_value, &_owner, old_value);
+  ADIM_guarantee(prev == old_value, "unexpected prev owner=" INTPTR_FORMAT
+                 ", expected=" INTPTR_FORMAT, p2i(prev), p2i(old_value));
+  log_trace(monitorinflation, owner)("mid=" INTPTR_FORMAT ", prev="
+                                     INTPTR_FORMAT ", new=" INTPTR_FORMAT,
+                                     p2i(this), p2i(prev), p2i(new_value));
+}
+
+// Set _owner field to new_value; current value must match old_value1 or old_value2.
+inline void ObjectMonitor::set_owner_from(void* new_value, void* old_value1, void* old_value2) {
+  void* prev = Atomic::cmpxchg(new_value, &_owner, old_value1);
+  if (prev != old_value1) {
+    prev = Atomic::cmpxchg(new_value, &_owner, old_value2);
+  }
+  ADIM_guarantee(prev == old_value1 || prev == old_value2,
+                 "unexpected prev owner=" INTPTR_FORMAT ", expected1="
+                 INTPTR_FORMAT " or expected2=" INTPTR_FORMAT, p2i(prev),
+                 p2i(old_value1), p2i(old_value2));
+  log_trace(monitorinflation, owner)("mid=" INTPTR_FORMAT ", prev="
+                                     INTPTR_FORMAT ", new=" INTPTR_FORMAT,
+                                     p2i(this), p2i(prev), p2i(new_value));
+}
+
+// Set _owner field to self; current value must match basic_lock_p.
+inline void ObjectMonitor::set_owner_from_BasicLock(Thread* self, void* basic_lock_p) {
+  assert(self->is_lock_owned((address)basic_lock_p), "self=" INTPTR_FORMAT
+         " must own basic_lock_p=" INTPTR_FORMAT, p2i(self), p2i(basic_lock_p));
+  void* prev = _owner;
+  ADIM_guarantee(prev == basic_lock_p, "unexpected prev owner=" INTPTR_FORMAT
+                 ", expected=" INTPTR_FORMAT, p2i(prev), p2i(basic_lock_p));
+  // Non-null owner field to non-null owner field is safe without
+  // cmpxchg() as long as all readers can tolerate either flavor.
+  _owner = self;
+  log_trace(monitorinflation, owner)("mid=" INTPTR_FORMAT ", prev="
+                                     INTPTR_FORMAT ", new=" INTPTR_FORMAT,
+                                     p2i(this), p2i(prev), p2i(self));
+}
+
+// Try to set _owner field to new_value if the current value matches
+// old_value. Otherwise, does not change the _owner field.
+inline void* ObjectMonitor::try_set_owner_from(void* new_value, void* old_value) {
+  void* prev = Atomic::cmpxchg(new_value, &_owner, old_value);
+  if (prev == old_value) {
+    log_trace(monitorinflation, owner)("mid=" INTPTR_FORMAT ", prev="
+                                       INTPTR_FORMAT ", new=" INTPTR_FORMAT,
+                                       p2i(this), p2i(prev), p2i(new_value));
+  }
+  return prev;
+}
+
+inline void ObjectMonitor::set_allocation_state(ObjectMonitor::AllocationState s) {
+  _allocation_state = s;
+}
+
+inline ObjectMonitor::AllocationState ObjectMonitor::allocation_state() const {
+  return _allocation_state;
+}
+
+inline bool ObjectMonitor::is_free() const {
+  return _allocation_state == Free;
+}
+
+inline bool ObjectMonitor::is_active() const {
+  return !is_free();
+}
+
+inline bool ObjectMonitor::is_old() const {
+  return _allocation_state == Old;
+}
+
+inline bool ObjectMonitor::is_new() const {
+  return _allocation_state == New;
+}
+
+inline void ObjectMonitor::dec_ref_count() {
+  // The decrement only needs to be MO_ACQ_REL since the reference
+  // counter is volatile.
+  Atomic::dec(&_ref_count);
+  // Can be negative as part of async deflation protocol.
+  ADIM_guarantee(AsyncDeflateIdleMonitors || ref_count() >= 0,
+                 "sanity check: ref_count=%d", ref_count());
+}
+
+inline void ObjectMonitor::inc_ref_count() {
+  // The increment needs to be MO_SEQ_CST so that the reference
+  // counter update is seen as soon as possible in a race with the
+  // async deflation protocol.
+  Atomic::inc(&_ref_count);
+  // Can be negative as part of async deflation protocol.
+  ADIM_guarantee(AsyncDeflateIdleMonitors || ref_count() > 0,
+                 "sanity check: ref_count=%d", ref_count());
+}
+
+inline jint ObjectMonitor::ref_count() const {
+  return OrderAccess::load_acquire(&_ref_count);
 }
 
 #endif // SHARE_RUNTIME_OBJECTMONITOR_INLINE_HPP
< prev index next >