--- old/src/hotspot/os_cpu/bsd_zero/atomic_bsd_zero.hpp 2019-11-21 11:54:05.491858649 +0100 +++ new/src/hotspot/os_cpu/bsd_zero/atomic_bsd_zero.hpp 2019-11-21 11:54:05.263854792 +0100 @@ -197,15 +197,15 @@ template<> template -inline T Atomic::PlatformXchg<4>::operator()(T exchange_value, - T volatile* dest, +inline T Atomic::PlatformXchg<4>::operator()(T volatile* dest, + T exchange_value, atomic_memory_order order) const { STATIC_ASSERT(4 == sizeof(T)); #ifdef ARM - return xchg_using_helper(arm_lock_test_and_set, exchange_value, dest); + return xchg_using_helper(arm_lock_test_and_set, dest, exchange_value); #else #ifdef M68K - return xchg_using_helper(m68k_lock_test_and_set, exchange_value, dest); + return xchg_using_helper(m68k_lock_test_and_set, dest, exchange_value); #else // __sync_lock_test_and_set is a bizarrely named atomic exchange // operation. Note that some platforms only support this with the @@ -224,8 +224,8 @@ template<> template -inline T Atomic::PlatformXchg<8>::operator()(T exchange_value, - T volatile* dest, +inline T Atomic::PlatformXchg<8>::operator()(T volatile* dest, + T exchange_value, atomic_memory_order order) const { STATIC_ASSERT(8 == sizeof(T)); T result = __sync_lock_test_and_set (dest, exchange_value);