< prev index next >

src/os_cpu/linux_x86/vm/atomic_linux_x86.hpp

Print this page
rev 13487 : imported patch add_linux_x86

*** 38,54 **** inline void Atomic::store (jint store_value, volatile jint* dest) { *dest = store_value; } inline void Atomic::store_ptr(intptr_t store_value, volatile intptr_t* dest) { *dest = store_value; } inline void Atomic::store_ptr(void* store_value, volatile void* dest) { *(void* volatile *)dest = store_value; } ! inline jint Atomic::add (jint add_value, volatile jint* dest) { ! jint addend = add_value; __asm__ volatile ( "lock xaddl %0,(%2)" ! : "=r" (addend) ! : "0" (addend), "r" (dest) : "cc", "memory"); ! return addend + add_value; } inline void Atomic::inc (volatile jint* dest) { __asm__ volatile ( "lock addl $1,(%0)" : : "r" (dest) : "cc", "memory"); --- 38,66 ---- inline void Atomic::store (jint store_value, volatile jint* dest) { *dest = store_value; } inline void Atomic::store_ptr(intptr_t store_value, volatile intptr_t* dest) { *dest = store_value; } inline void Atomic::store_ptr(void* store_value, volatile void* dest) { *(void* volatile *)dest = store_value; } ! template<size_t byte_size> ! struct Atomic::PlatformAdd ! : Atomic::FetchAndAdd<Atomic::PlatformAdd<byte_size> > ! { ! template<typename I, typename D> ! D fetch_and_add(I add_value, D volatile* dest) const; ! }; ! ! template<> ! template<typename I, typename D> ! inline D Atomic::PlatformAdd<4>::fetch_and_add(I add_value, D volatile* dest) const { ! STATIC_ASSERT(4 == sizeof(I)); ! STATIC_ASSERT(4 == sizeof(D)); ! D old_value; __asm__ volatile ( "lock xaddl %0,(%2)" ! : "=r" (old_value) ! : "0" (add_value), "r" (dest) : "cc", "memory"); ! return old_value; } inline void Atomic::inc (volatile jint* dest) { __asm__ volatile ( "lock addl $1,(%0)" : : "r" (dest) : "cc", "memory");
*** 109,129 **** #ifdef AMD64 inline void Atomic::store (jlong store_value, jlong* dest) { *dest = store_value; } inline void Atomic::store (jlong store_value, volatile jlong* dest) { *dest = store_value; } ! inline intptr_t Atomic::add_ptr(intptr_t add_value, volatile intptr_t* dest) { ! intptr_t addend = add_value; __asm__ __volatile__ ("lock xaddq %0,(%2)" ! : "=r" (addend) ! : "0" (addend), "r" (dest) : "cc", "memory"); ! return addend + add_value; ! } ! ! inline void* Atomic::add_ptr(intptr_t add_value, volatile void* dest) { ! return (void*)add_ptr(add_value, (volatile intptr_t*)dest); } inline void Atomic::inc_ptr(volatile intptr_t* dest) { __asm__ __volatile__ ("lock addq $1,(%0)" : --- 121,141 ---- #ifdef AMD64 inline void Atomic::store (jlong store_value, jlong* dest) { *dest = store_value; } inline void Atomic::store (jlong store_value, volatile jlong* dest) { *dest = store_value; } ! template<> ! template<typename I, typename D> ! inline D Atomic::PlatformAdd<8>::fetch_and_add(I add_value, D volatile* dest) const { ! STATIC_ASSERT(8 == sizeof(I)); ! STATIC_ASSERT(8 == sizeof(D)); ! D old_value; __asm__ __volatile__ ("lock xaddq %0,(%2)" ! : "=r" (old_value) ! : "0" (add_value), "r" (dest) : "cc", "memory"); ! return old_value; } inline void Atomic::inc_ptr(volatile intptr_t* dest) { __asm__ __volatile__ ("lock addq $1,(%0)" :
*** 162,180 **** inline jlong Atomic::load(const volatile jlong* src) { return *src; } #else // !AMD64 - inline intptr_t Atomic::add_ptr(intptr_t add_value, volatile intptr_t* dest) { - return (intptr_t)Atomic::add((jint)add_value, (volatile jint*)dest); - } - - inline void* Atomic::add_ptr(intptr_t add_value, volatile void* dest) { - return (void*)Atomic::add((jint)add_value, (volatile jint*)dest); - } - - inline void Atomic::inc_ptr(volatile intptr_t* dest) { inc((volatile jint*)dest); } inline void Atomic::dec_ptr(volatile intptr_t* dest) { --- 174,183 ----
< prev index next >