1 /*
2 * Copyright (c) 2001, 2013, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
57 if (res == cur) break;
58 cur = res;
59 new_val = cur;
60 new_val_as_bytes[offset] = exchange_value;
61 }
62 return cur_as_bytes[offset];
63 }
64
65 unsigned Atomic::xchg(unsigned int exchange_value, volatile unsigned int* dest) {
66 assert(sizeof(unsigned int) == sizeof(jint), "more work to do");
67 return (unsigned int)Atomic::xchg((jint)exchange_value, (volatile jint*)dest);
68 }
69
70 unsigned Atomic::cmpxchg(unsigned int exchange_value,
71 volatile unsigned int* dest, unsigned int compare_value) {
72 assert(sizeof(unsigned int) == sizeof(jint), "more work to do");
73 return (unsigned int)Atomic::cmpxchg((jint)exchange_value, (volatile jint*)dest,
74 (jint)compare_value);
75 }
76
77 jlong Atomic::add(jlong add_value, volatile jlong* dest) {
78 jlong old = load(dest);
79 jlong new_value = old + add_value;
80 while (old != cmpxchg(new_value, dest, old)) {
81 old = load(dest);
82 new_value = old + add_value;
83 }
84 return old;
85 }
86
87 void Atomic::inc(volatile short* dest) {
88 // Most platforms do not support atomic increment on a 2-byte value. However,
89 // if the value occupies the most significant 16 bits of an aligned 32-bit
90 // word, then we can do this with an atomic add of 0x10000 to the 32-bit word.
91 //
92 // The least significant parts of this 32-bit word will never be affected, even
93 // in case of overflow/underflow.
94 //
95 // Use the ATOMIC_SHORT_PAIR macro to get the desired alignment.
96 #ifdef VM_LITTLE_ENDIAN
97 assert((intx(dest) & 0x03) == 0x02, "wrong alignment");
98 (void)Atomic::add(0x10000, (volatile int*)(dest-1));
99 #else
100 assert((intx(dest) & 0x03) == 0x00, "wrong alignment");
101 (void)Atomic::add(0x10000, (volatile int*)(dest));
102 #endif
103 }
104
105 void Atomic::dec(volatile short* dest) {
106 #ifdef VM_LITTLE_ENDIAN
|
1 /*
2 * Copyright (c) 2001, 2019, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
57 if (res == cur) break;
58 cur = res;
59 new_val = cur;
60 new_val_as_bytes[offset] = exchange_value;
61 }
62 return cur_as_bytes[offset];
63 }
64
65 unsigned Atomic::xchg(unsigned int exchange_value, volatile unsigned int* dest) {
66 assert(sizeof(unsigned int) == sizeof(jint), "more work to do");
67 return (unsigned int)Atomic::xchg((jint)exchange_value, (volatile jint*)dest);
68 }
69
70 unsigned Atomic::cmpxchg(unsigned int exchange_value,
71 volatile unsigned int* dest, unsigned int compare_value) {
72 assert(sizeof(unsigned int) == sizeof(jint), "more work to do");
73 return (unsigned int)Atomic::cmpxchg((jint)exchange_value, (volatile jint*)dest,
74 (jint)compare_value);
75 }
76
77 julong Atomic::cmpxchg(julong exchange_value,
78 volatile julong* dest, julong compare_value) {
79 return (julong)Atomic::cmpxchg((jlong)exchange_value, (volatile jlong*)dest,
80 (jlong)compare_value);
81 }
82
83 julong Atomic::load(volatile julong* src) {
84 return (julong)load((volatile jlong*)src);
85 }
86
87 jlong Atomic::add(jlong add_value, volatile jlong* dest) {
88 jlong old = load(dest);
89 jlong new_value = old + add_value;
90 while (old != cmpxchg(new_value, dest, old)) {
91 old = load(dest);
92 new_value = old + add_value;
93 }
94 return old;
95 }
96
97 julong Atomic::add(julong add_value, volatile julong* dest) {
98 julong old = load(dest);
99 julong new_value = old + add_value;
100 while (old != cmpxchg(new_value, dest, old)) {
101 old = load(dest);
102 new_value = old + add_value;
103 }
104 return old;
105 }
106
107 void Atomic::inc(volatile short* dest) {
108 // Most platforms do not support atomic increment on a 2-byte value. However,
109 // if the value occupies the most significant 16 bits of an aligned 32-bit
110 // word, then we can do this with an atomic add of 0x10000 to the 32-bit word.
111 //
112 // The least significant parts of this 32-bit word will never be affected, even
113 // in case of overflow/underflow.
114 //
115 // Use the ATOMIC_SHORT_PAIR macro to get the desired alignment.
116 #ifdef VM_LITTLE_ENDIAN
117 assert((intx(dest) & 0x03) == 0x02, "wrong alignment");
118 (void)Atomic::add(0x10000, (volatile int*)(dest-1));
119 #else
120 assert((intx(dest) & 0x03) == 0x00, "wrong alignment");
121 (void)Atomic::add(0x10000, (volatile int*)(dest));
122 #endif
123 }
124
125 void Atomic::dec(volatile short* dest) {
126 #ifdef VM_LITTLE_ENDIAN
|