< prev index next >
src/hotspot/share/code/nmethod.cpp
Print this page
rev 56098 : imported patch 8226705-8221734-baseline
*** 48,57 ****
--- 48,58 ----
#include "oops/method.inline.hpp"
#include "oops/methodData.hpp"
#include "oops/oop.inline.hpp"
#include "prims/jvmtiImpl.hpp"
#include "runtime/atomic.hpp"
+ #include "runtime/deoptimization.hpp"
#include "runtime/flags/flagSetting.hpp"
#include "runtime/frame.inline.hpp"
#include "runtime/handles.inline.hpp"
#include "runtime/jniHandles.inline.hpp"
#include "runtime/orderAccess.hpp"
*** 1191,1205 ****
// If _method is already NULL the Method* is about to be unloaded,
// so we don't have to break the cycle. Note that it is possible to
// have the Method* live here, in case we unload the nmethod because
// it is pointing to some oop (other than the Method*) being unloaded.
if (_method != NULL) {
! // OSR methods point to the Method*, but the Method* does not
! // point back!
! if (_method->code() == this) {
! _method->clear_code(); // Break a cycle
! }
}
// Make the class unloaded - i.e., change state and notify sweeper
assert(SafepointSynchronize::is_at_safepoint() || Thread::current()->is_ConcurrentGC_thread(),
"must be at safepoint");
--- 1192,1202 ----
// If _method is already NULL the Method* is about to be unloaded,
// so we don't have to break the cycle. Note that it is possible to
// have the Method* live here, in case we unload the nmethod because
// it is pointing to some oop (other than the Method*) being unloaded.
if (_method != NULL) {
! _method->unlink_code(this);
}
// Make the class unloaded - i.e., change state and notify sweeper
assert(SafepointSynchronize::is_at_safepoint() || Thread::current()->is_ConcurrentGC_thread(),
"must be at safepoint");
*** 1279,1298 ****
if (PrintCompilation && _state != unloaded) {
print_on(tty, state_msg);
}
}
! void nmethod::unlink_from_method(bool acquire_lock) {
! // We need to check if both the _code and _from_compiled_code_entry_point
! // refer to this nmethod because there is a race in setting these two fields
! // in Method* as seen in bugid 4947125.
! // If the vep() points to the zombie nmethod, the memory for the nmethod
! // could be flushed and the compiler and vtable stubs could still call
! // through it.
! if (method() != NULL && (method()->code() == this ||
! method()->from_compiled_entry() == verified_entry_point())) {
! method()->clear_code(acquire_lock);
}
}
/**
* Common functionality for both make_not_entrant and make_zombie
--- 1276,1288 ----
if (PrintCompilation && _state != unloaded) {
print_on(tty, state_msg);
}
}
! void nmethod::unlink_from_method() {
! if (method() != NULL) {
! method()->unlink_code();
}
}
/**
* Common functionality for both make_not_entrant and make_zombie
*** 1315,1342 ****
// This can be called while the system is already at a safepoint which is ok
NoSafepointVerifier nsv;
// during patching, depending on the nmethod state we must notify the GC that
// code has been unloaded, unregistering it. We cannot do this right while
! // holding the Patching_lock because we need to use the CodeCache_lock. This
// would be prone to deadlocks.
// This flag is used to remember whether we need to later lock and unregister.
bool nmethod_needs_unregister = false;
- {
// invalidate osr nmethod before acquiring the patching lock since
// they both acquire leaf locks and we don't want a deadlock.
// This logic is equivalent to the logic below for patching the
// verified entry point of regular methods. We check that the
// nmethod is in use to ensure that it is invalidated only once.
if (is_osr_method() && is_in_use()) {
// this effectively makes the osr nmethod not entrant
invalidate_osr_method();
}
// Enter critical section. Does not block for safepoint.
! MutexLocker pl(Patching_lock, Mutex::_no_safepoint_check_flag);
if (Atomic::load(&_state) >= state) {
// another thread already performed this transition so nothing
// to do, but return false to indicate this.
return false;
--- 1305,1332 ----
// This can be called while the system is already at a safepoint which is ok
NoSafepointVerifier nsv;
// during patching, depending on the nmethod state we must notify the GC that
// code has been unloaded, unregistering it. We cannot do this right while
! // holding the CompiledMethod_lock because we need to use the CodeCache_lock. This
// would be prone to deadlocks.
// This flag is used to remember whether we need to later lock and unregister.
bool nmethod_needs_unregister = false;
// invalidate osr nmethod before acquiring the patching lock since
// they both acquire leaf locks and we don't want a deadlock.
// This logic is equivalent to the logic below for patching the
// verified entry point of regular methods. We check that the
// nmethod is in use to ensure that it is invalidated only once.
if (is_osr_method() && is_in_use()) {
// this effectively makes the osr nmethod not entrant
invalidate_osr_method();
}
+ {
// Enter critical section. Does not block for safepoint.
! MutexLocker pl(CompiledMethod_lock, Mutex::_no_safepoint_check_flag);
if (Atomic::load(&_state) >= state) {
// another thread already performed this transition so nothing
// to do, but return false to indicate this.
return false;
*** 1387,1398 ****
// Log the transition once
log_state_change();
// Remove nmethod from method.
! unlink_from_method(false /* already owns Patching_lock */);
! } // leave critical region under Patching_lock
#if INCLUDE_JVMCI
// Invalidate can't occur while holding the Patching lock
JVMCINMethodData* nmethod_data = jvmci_nmethod_data();
if (nmethod_data != NULL) {
--- 1377,1389 ----
// Log the transition once
log_state_change();
// Remove nmethod from method.
! unlink_from_method();
!
! } // leave critical region under CompiledMethod_lock
#if INCLUDE_JVMCI
// Invalidate can't occur while holding the Patching lock
JVMCINMethodData* nmethod_data = jvmci_nmethod_data();
if (nmethod_data != NULL) {
< prev index next >