1 /*
   2  * Copyright (c) 1999, 2017, Oracle and/or its affiliates. All rights reserved.
   3  * Copyright (c) 2014, Red Hat Inc. All rights reserved.
   4  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   5  *
   6  * This code is free software; you can redistribute it and/or modify it
   7  * under the terms of the GNU General Public License version 2 only, as
   8  * published by the Free Software Foundation.
   9  *
  10  * This code is distributed in the hope that it will be useful, but WITHOUT
  11  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  12  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  13  * version 2 for more details (a copy is included in the LICENSE file that
  14  * accompanied this code).
  15  *
  16  * You should have received a copy of the GNU General Public License version
  17  * 2 along with this work; if not, write to the Free Software Foundation,
  18  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  19  *
  20  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  21  * or visit www.oracle.com if you need additional information or have any
  22  * questions.
  23  *
  24  */
  25 
  26 #ifndef OS_CPU_LINUX_AARCH64_VM_ATOMIC_LINUX_AARCH64_HPP
  27 #define OS_CPU_LINUX_AARCH64_VM_ATOMIC_LINUX_AARCH64_HPP
  28 
  29 #include "vm_version_aarch64.hpp"
  30 
  31 // Implementation of class atomic
  32 
  33 #define FULL_MEM_BARRIER  __sync_synchronize()
  34 #define READ_MEM_BARRIER  __atomic_thread_fence(__ATOMIC_ACQUIRE);
  35 #define WRITE_MEM_BARRIER __atomic_thread_fence(__ATOMIC_RELEASE);
  36 
  37 
  38 template <>
  39 inline int32_t Atomic::specialized_add<int32_t>(int32_t add_value, volatile int32_t* dest) {
  40  return __sync_add_and_fetch(dest, add_value);
  41 }
  42 
  43 
  44 template <>
  45 inline int64_t Atomic::specialized_add<int64_t>(int64_t add_value, volatile int64_t* dest) {
  46  return __sync_add_and_fetch(dest, add_value);
  47 }
  48 
  49 
  50 template <>
  51 inline int32_t Atomic::specialized_xchg<int32_t>(int32_t exchange_value, volatile int32_t* dest) {
  52   int32_t res = __sync_lock_test_and_set (dest, exchange_value);
  53   FULL_MEM_BARRIER;
  54   return res;
  55 }
  56 
  57 template <>
  58 inline int64_t Atomic::specialized_xchg<int64_t>(int64_t exchange_value, volatile int64_t* dest) {
  59   int64_t res = __sync_lock_test_and_set (dest, exchange_value);
  60   FULL_MEM_BARRIER;
  61   return res;
  62 }
  63 
  64 template <>
  65 inline Atomic::CanonicalPointer Atomic::specialized_xchg<Atomic::CanonicalPointer>(Atomic::CanonicalPointer exchange_value, volatile Atomic::CanonicalPointer* dest) {
  66   Atomic::CanonicalPointer res = __sync_lock_test_and_set (dest, exchange_value);
  67   FULL_MEM_BARRIER;
  68   return res;
  69 }
  70 
  71 template <typename T> T generic_cmpxchg(T exchange_value, volatile T* dest,
  72                                         T compare_value, cmpxchg_memory_order order)
  73 {
  74   if (order == memory_order_relaxed) {
  75     T value = compare_value;
  76     __atomic_compare_exchange(dest, &value, &exchange_value, /*weak*/false,
  77                               __ATOMIC_RELAXED, __ATOMIC_RELAXED);
  78     return value;
  79   } else {
  80     return __sync_val_compare_and_swap(dest, compare_value, exchange_value);
  81   }
  82 }
  83 
  84 #define VM_HAS_SPECIALIZED_CMPXCHG_BYTE
  85 template <>
  86 inline int8_t Atomic::specialized_cmpxchg<int8_t>(int8_t exchange_value, volatile int8_t* dest, int8_t compare_value, cmpxchg_memory_order order) {
  87   return generic_cmpxchg(exchange_value, dest, compare_value, order);
  88 }
  89 
  90 template <>
  91 inline int32_t Atomic::specialized_cmpxchg<int32_t>(int32_t exchange_value, volatile int32_t* dest, int32_t compare_value, cmpxchg_memory_order order) {
  92   return generic_cmpxchg(exchange_value, dest, compare_value, order);
  93 }
  94 
  95 
  96 template <>
  97 inline int64_t Atomic::specialized_cmpxchg<int64_t>(int64_t exchange_value, volatile int64_t* dest, int64_t compare_value, cmpxchg_memory_order order) {
  98   return generic_cmpxchg(exchange_value, dest, compare_value, order);
  99 }
 100 
 101 
 102 #endif // OS_CPU_LINUX_AARCH64_VM_ATOMIC_LINUX_AARCH64_HPP