1 /* 2 * Copyright (c) 2001, 2014, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "runtime/atomic.inline.hpp" 27 28 #ifndef X86 29 jbyte Atomic::cmpxchg(jbyte exchange_value, volatile jbyte* dest, jbyte compare_value) { 30 assert(sizeof(jbyte) == 1, "assumption."); 31 uintptr_t dest_addr = (uintptr_t)dest; 32 uintptr_t offset = dest_addr % sizeof(jint); 33 volatile jint* dest_int = (volatile jint*)(dest_addr - offset); 34 jint cur = *dest_int; 35 jbyte* cur_as_bytes = (jbyte*)(&cur); 36 jint new_val = cur; 37 jbyte* new_val_as_bytes = (jbyte*)(&new_val); 38 new_val_as_bytes[offset] = exchange_value; 39 while (cur_as_bytes[offset] == compare_value) { 40 jint res = cmpxchg(new_val, dest_int, cur); 41 if (res == cur) break; 42 cur = res; 43 new_val = cur; 44 new_val_as_bytes[offset] = exchange_value; 45 } 46 return cur_as_bytes[offset]; 47 } 48 #endif 49 50 unsigned Atomic::xchg(unsigned int exchange_value, volatile unsigned int* dest) { 51 assert(sizeof(unsigned int) == sizeof(jint), "more work to do"); 52 return (unsigned int)Atomic::xchg((jint)exchange_value, (volatile jint*)dest); 53 } 54 55 unsigned Atomic::cmpxchg(unsigned int exchange_value, 56 volatile unsigned int* dest, unsigned int compare_value) { 57 assert(sizeof(unsigned int) == sizeof(jint), "more work to do"); 58 return (unsigned int)Atomic::cmpxchg((jint)exchange_value, (volatile jint*)dest, 59 (jint)compare_value); 60 } 61 62 jlong Atomic::add(jlong add_value, volatile jlong* dest) { 63 jlong old = load(dest); 64 jlong new_value = old + add_value; 65 while (old != cmpxchg(new_value, dest, old)) { 66 old = load(dest); 67 new_value = old + add_value; 68 } 69 return old; 70 } 71 72 void Atomic::inc(volatile short* dest) { 73 // Most platforms do not support atomic increment on a 2-byte value. However, 74 // if the value occupies the most significant 16 bits of an aligned 32-bit 75 // word, then we can do this with an atomic add of 0x10000 to the 32-bit word. 76 // 77 // The least significant parts of this 32-bit word will never be affected, even 78 // in case of overflow/underflow. 79 // 80 // Use the ATOMIC_SHORT_PAIR macro to get the desired alignment. 81 #ifdef VM_LITTLE_ENDIAN 82 assert((intx(dest) & 0x03) == 0x02, "wrong alignment"); 83 (void)Atomic::add(0x10000, (volatile int*)(dest-1)); 84 #else 85 assert((intx(dest) & 0x03) == 0x00, "wrong alignment"); 86 (void)Atomic::add(0x10000, (volatile int*)(dest)); 87 #endif 88 } 89 90 void Atomic::dec(volatile short* dest) { 91 #ifdef VM_LITTLE_ENDIAN 92 assert((intx(dest) & 0x03) == 0x02, "wrong alignment"); 93 (void)Atomic::add(-0x10000, (volatile int*)(dest-1)); 94 #else 95 assert((intx(dest) & 0x03) == 0x00, "wrong alignment"); 96 (void)Atomic::add(-0x10000, (volatile int*)(dest)); 97 #endif 98 } 99