1 /* 2 * Copyright (c) 1999, 2017, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #ifndef OS_CPU_BSD_X86_VM_ATOMIC_BSD_X86_HPP 26 #define OS_CPU_BSD_X86_VM_ATOMIC_BSD_X86_HPP 27 28 // Implementation of class atomic 29 30 inline void Atomic::store (jbyte store_value, jbyte* dest) { *dest = store_value; } 31 inline void Atomic::store (jshort store_value, jshort* dest) { *dest = store_value; } 32 inline void Atomic::store (jint store_value, jint* dest) { *dest = store_value; } 33 inline void Atomic::store_ptr(intptr_t store_value, intptr_t* dest) { *dest = store_value; } 34 inline void Atomic::store_ptr(void* store_value, void* dest) { *(void**)dest = store_value; } 35 36 inline void Atomic::store (jbyte store_value, volatile jbyte* dest) { *dest = store_value; } 37 inline void Atomic::store (jshort store_value, volatile jshort* dest) { *dest = store_value; } 38 inline void Atomic::store (jint store_value, volatile jint* dest) { *dest = store_value; } 39 inline void Atomic::store_ptr(intptr_t store_value, volatile intptr_t* dest) { *dest = store_value; } 40 inline void Atomic::store_ptr(void* store_value, volatile void* dest) { *(void* volatile *)dest = store_value; } 41 42 43 template<size_t byte_size> 44 struct Atomic::PlatformAdd 45 : Atomic::FetchAndAdd<Atomic::PlatformAdd<byte_size> > 46 { 47 template<typename I, typename D> 48 D fetch_and_add(I add_value, D volatile* dest) const; 49 }; 50 51 template<> 52 template<typename I, typename D> 53 inline D Atomic::PlatformAdd<4>::fetch_and_add(I add_value, D volatile* dest) const { 54 STATIC_ASSERT(4 == sizeof(I)); 55 STATIC_ASSERT(4 == sizeof(D)); 56 D old_value; 57 __asm__ volatile ( "lock xaddl %0,(%2)" 58 : "=r" (old_value) 59 : "0" (add_value), "r" (dest) 60 : "cc", "memory"); 61 return old_value; 62 } 63 64 inline void Atomic::inc (volatile jint* dest) { 65 __asm__ volatile ( "lock addl $1,(%0)" : 66 : "r" (dest) : "cc", "memory"); 67 } 68 69 inline void Atomic::inc_ptr(volatile void* dest) { 70 inc_ptr((volatile intptr_t*)dest); 71 } 72 73 inline void Atomic::dec (volatile jint* dest) { 74 __asm__ volatile ( "lock subl $1,(%0)" : 75 : "r" (dest) : "cc", "memory"); 76 } 77 78 inline void Atomic::dec_ptr(volatile void* dest) { 79 dec_ptr((volatile intptr_t*)dest); 80 } 81 82 inline jint Atomic::xchg (jint exchange_value, volatile jint* dest) { 83 __asm__ volatile ( "xchgl (%2),%0" 84 : "=r" (exchange_value) 85 : "0" (exchange_value), "r" (dest) 86 : "memory"); 87 return exchange_value; 88 } 89 90 inline void* Atomic::xchg_ptr(void* exchange_value, volatile void* dest) { 91 return (void*)xchg_ptr((intptr_t)exchange_value, (volatile intptr_t*)dest); 92 } 93 94 template<> 95 template<typename T> 96 inline T Atomic::PlatformCmpxchg<1>::operator()(T exchange_value, 97 T volatile* dest, 98 T compare_value, 99 cmpxchg_memory_order /* order */) const { 100 STATIC_ASSERT(1 == sizeof(T)); 101 __asm__ volatile ( "lock cmpxchgb %1,(%3)" 102 : "=a" (exchange_value) 103 : "q" (exchange_value), "a" (compare_value), "r" (dest) 104 : "cc", "memory"); 105 return exchange_value; 106 } 107 108 template<> 109 template<typename T> 110 inline T Atomic::PlatformCmpxchg<4>::operator()(T exchange_value, 111 T volatile* dest, 112 T compare_value, 113 cmpxchg_memory_order /* order */) const { 114 STATIC_ASSERT(4 == sizeof(T)); 115 __asm__ volatile ( "lock cmpxchgl %1,(%3)" 116 : "=a" (exchange_value) 117 : "r" (exchange_value), "a" (compare_value), "r" (dest) 118 : "cc", "memory"); 119 return exchange_value; 120 } 121 122 #ifdef AMD64 123 inline void Atomic::store (jlong store_value, jlong* dest) { *dest = store_value; } 124 inline void Atomic::store (jlong store_value, volatile jlong* dest) { *dest = store_value; } 125 126 template<> 127 template<typename I, typename D> 128 inline D Atomic::PlatformAdd<8>::fetch_and_add(I add_value, D volatile* dest) const { 129 STATIC_ASSERT(8 == sizeof(I)); 130 STATIC_ASSERT(8 == sizeof(D)); 131 D old_value; 132 __asm__ __volatile__ ( "lock xaddq %0,(%2)" 133 : "=r" (old_value) 134 : "0" (add_value), "r" (dest) 135 : "cc", "memory"); 136 return old_value; 137 } 138 139 inline void Atomic::inc_ptr(volatile intptr_t* dest) { 140 __asm__ __volatile__ ( "lock addq $1,(%0)" 141 : 142 : "r" (dest) 143 : "cc", "memory"); 144 } 145 146 inline void Atomic::dec_ptr(volatile intptr_t* dest) { 147 __asm__ __volatile__ ( "lock subq $1,(%0)" 148 : 149 : "r" (dest) 150 : "cc", "memory"); 151 } 152 153 inline intptr_t Atomic::xchg_ptr(intptr_t exchange_value, volatile intptr_t* dest) { 154 __asm__ __volatile__ ("xchgq (%2),%0" 155 : "=r" (exchange_value) 156 : "0" (exchange_value), "r" (dest) 157 : "memory"); 158 return exchange_value; 159 } 160 161 template<> 162 template<typename T> 163 inline T Atomic::PlatformCmpxchg<8>::operator()(T exchange_value, 164 T volatile* dest, 165 T compare_value, 166 cmpxchg_memory_order /* order */) const { 167 STATIC_ASSERT(8 == sizeof(T)); 168 __asm__ __volatile__ ( "lock cmpxchgq %1,(%3)" 169 : "=a" (exchange_value) 170 : "r" (exchange_value), "a" (compare_value), "r" (dest) 171 : "cc", "memory"); 172 return exchange_value; 173 } 174 175 inline jlong Atomic::load(const volatile jlong* src) { return *src; } 176 177 #else // !AMD64 178 179 inline void Atomic::inc_ptr(volatile intptr_t* dest) { 180 inc((volatile jint*)dest); 181 } 182 183 inline void Atomic::dec_ptr(volatile intptr_t* dest) { 184 dec((volatile jint*)dest); 185 } 186 187 inline intptr_t Atomic::xchg_ptr(intptr_t exchange_value, volatile intptr_t* dest) { 188 return (intptr_t)xchg((jint)exchange_value, (volatile jint*)dest); 189 } 190 191 extern "C" { 192 // defined in bsd_x86.s 193 jlong _Atomic_cmpxchg_long(jlong, volatile jlong*, jlong, bool); 194 void _Atomic_move_long(const volatile jlong* src, volatile jlong* dst); 195 } 196 197 template<> 198 template<typename T> 199 inline T Atomic::PlatformCmpxchg<8>::operator()(T exchange_value, 200 T volatile* dest, 201 T compare_value, 202 cmpxchg_memory_order order) const { 203 STATIC_ASSERT(8 == sizeof(T)); 204 return cmpxchg_using_helper<jlong>(_Atomic_cmpxchg_long, exchange_value, dest, compare_value); 205 } 206 207 inline jlong Atomic::load(const volatile jlong* src) { 208 volatile jlong dest; 209 _Atomic_move_long(src, &dest); 210 return dest; 211 } 212 213 inline void Atomic::store(jlong store_value, jlong* dest) { 214 _Atomic_move_long((volatile jlong*)&store_value, (volatile jlong*)dest); 215 } 216 217 inline void Atomic::store(jlong store_value, volatile jlong* dest) { 218 _Atomic_move_long((volatile jlong*)&store_value, dest); 219 } 220 221 #endif // AMD64 222 223 #endif // OS_CPU_BSD_X86_VM_ATOMIC_BSD_X86_HPP