< prev index next >

src/os_cpu/linux_aarch64/vm/atomic_linux_aarch64.inline.hpp

Print this page
rev 10437 : 8151775: aarch64: add support for 8.1 LSE atomic operations
Reviewed-by: aph

*** 34,43 **** --- 34,56 ---- #define FULL_MEM_BARRIER __sync_synchronize() #define READ_MEM_BARRIER __atomic_thread_fence(__ATOMIC_ACQUIRE); #define WRITE_MEM_BARRIER __atomic_thread_fence(__ATOMIC_RELEASE); + // CASALW w2, w0, [x1] + #define CASALW ".word 0b10001000111000101111110000100000;" + // CASAL x2, x0, [x1] + #define CASAL ".word 0b11001000111000101111110000100000;" + // LDADDALW w0, w2, [x1] + #define LDADDALW ".word 0b10111000111000000000000000100010;" + // LDADDAL w0, w2, [x1] + #define LDADDAL ".word 0b11111000111000000000000000100010;" + // SWPW w0, w2, [x1] + #define SWPW ".word 0b10111000001000001000000000100010;" + // SWP x0, x2, [x1] + #define SWP ".word 0b11111000001000001000000000100010;" + inline void Atomic::store (jbyte store_value, jbyte* dest) { *dest = store_value; } inline void Atomic::store (jshort store_value, jshort* dest) { *dest = store_value; } inline void Atomic::store (jint store_value, jint* dest) { *dest = store_value; } inline void Atomic::store_ptr(intptr_t store_value, intptr_t* dest) { *dest = store_value; } inline void Atomic::store_ptr(void* store_value, void* dest) { *(void**)dest = store_value; }
*** 49,58 **** --- 62,82 ---- inline void Atomic::store_ptr(void* store_value, volatile void* dest) { *(void* volatile *)dest = store_value; } inline jint Atomic::add(jint add_value, volatile jint* dest) { + if (UseLSE) { + register jint r_add_value asm("w0") = add_value; + register volatile jint *r_dest asm("x1") = dest; + register jint r_result asm("w2"); + __asm volatile(LDADDALW + : [_result]"=r"(r_result) + : [_add_value]"r"(r_add_value), + [_dest]"r"(r_dest) + : "memory"); + return r_result+add_value; + } return __sync_add_and_fetch(dest, add_value); } inline void Atomic::inc(volatile jint* dest) {
*** 74,83 **** --- 98,119 ---- add_ptr(-1, dest); } inline jint Atomic::xchg (jint exchange_value, volatile jint* dest) { + if (UseLSE) { + register jint r_exchange_value asm("w0") = exchange_value; + register volatile jint *r_dest asm("x1") = dest; + register jint r_result asm("w2"); + __asm volatile(SWPW + : [_result]"=r"(r_result) + : [_exchange_value]"r"(r_exchange_value), + [_dest]"r"(r_dest) + : "memory"); + FULL_MEM_BARRIER; + return r_result; + } jint res = __sync_lock_test_and_set (dest, exchange_value); FULL_MEM_BARRIER; return res; }
*** 85,105 **** { return (void *) xchg_ptr((intptr_t) exchange_value, (volatile intptr_t*) dest); } - inline jint Atomic::cmpxchg (jint exchange_value, volatile jint* dest, jint compare_value) { return __sync_val_compare_and_swap(dest, compare_value, exchange_value); } inline void Atomic::store (jlong store_value, jlong* dest) { *dest = store_value; } inline void Atomic::store (jlong store_value, volatile jlong* dest) { *dest = store_value; } inline intptr_t Atomic::add_ptr(intptr_t add_value, volatile intptr_t* dest) { return __sync_add_and_fetch(dest, add_value); } inline void* Atomic::add_ptr(intptr_t add_value, volatile void* dest) { --- 121,162 ---- { return (void *) xchg_ptr((intptr_t) exchange_value, (volatile intptr_t*) dest); } inline jint Atomic::cmpxchg (jint exchange_value, volatile jint* dest, jint compare_value) { + if (UseLSE) { + register jint r_exchange_value asm("w0") = exchange_value; + register volatile jint *r_dest asm("x1") = dest; + register jint r_compare_value asm("w2") = compare_value; + __asm volatile(CASALW + : [_compare_value]"+r"(r_compare_value) + : [_exchange_value]"r"(r_exchange_value), + [_dest]"r"(r_dest) + : "memory"); + return r_compare_value; + } return __sync_val_compare_and_swap(dest, compare_value, exchange_value); } inline void Atomic::store (jlong store_value, jlong* dest) { *dest = store_value; } inline void Atomic::store (jlong store_value, volatile jlong* dest) { *dest = store_value; } inline intptr_t Atomic::add_ptr(intptr_t add_value, volatile intptr_t* dest) { + if (UseLSE) { + register intptr_t r_add_value asm("x0") = add_value; + register volatile intptr_t *r_dest asm("x1") = dest; + register intptr_t r_result asm("x2"); + __asm volatile(LDADDAL + : [_result]"=r"(r_result) + : [_add_value]"r"(r_add_value), + [_dest]"r"(r_dest) + : "memory"); + return r_result+add_value; + } return __sync_add_and_fetch(dest, add_value); } inline void* Atomic::add_ptr(intptr_t add_value, volatile void* dest) {
*** 116,137 **** --- 173,228 ---- add_ptr(-1, dest); } inline intptr_t Atomic::xchg_ptr(intptr_t exchange_value, volatile intptr_t* dest) { + if (UseLSE) { + register intptr_t r_exchange_value asm("x0") = exchange_value; + register volatile intptr_t *r_dest asm("x1") = dest; + register intptr_t r_result asm("x2"); + __asm volatile(SWP + : [_result]"=r"(r_result) + : [_exchange_value]"r"(r_exchange_value), + [_dest]"r"(r_dest) + : "memory"); + FULL_MEM_BARRIER; + return r_result; + } intptr_t res = __sync_lock_test_and_set (dest, exchange_value); FULL_MEM_BARRIER; return res; } inline jlong Atomic::cmpxchg (jlong exchange_value, volatile jlong* dest, jlong compare_value) { + if (UseLSE) { + register jlong r_exchange_value asm("x0") = exchange_value; + register volatile jlong *r_dest asm("x1") = dest; + register jlong r_compare_value asm("x2") = compare_value; + __asm volatile(CASAL + : [_compare_value]"+r"(r_compare_value) + : [_exchange_value]"r"(r_exchange_value), + [_dest]"r"(r_dest) + : "memory"); + return r_compare_value; + } return __sync_val_compare_and_swap(dest, compare_value, exchange_value); } inline intptr_t Atomic::cmpxchg_ptr(intptr_t exchange_value, volatile intptr_t* dest, intptr_t compare_value) { + if (UseLSE) { + register intptr_t r_exchange_value asm("x0") = exchange_value; + register volatile intptr_t *r_dest asm("x1") = dest; + register intptr_t r_compare_value asm("x2") = compare_value; + __asm volatile(CASAL + : [_compare_value]"+r"(r_compare_value) + : [_exchange_value]"r"(r_exchange_value), + [_dest]"r"(r_dest) + : "memory"); + return r_compare_value; + } return __sync_val_compare_and_swap(dest, compare_value, exchange_value); } inline void* Atomic::cmpxchg_ptr(void* exchange_value, volatile void* dest, void* compare_value) {
< prev index next >