1 /* 2 * Copyright (c) 2018, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "gc/shared/barrierSetAssembler.hpp" 27 #include "gc/shared/collectedHeap.hpp" 28 #include "interpreter/interp_masm.hpp" 29 #include "runtime/jniHandles.hpp" 30 #include "runtime/thread.hpp" 31 32 #define __ masm-> 33 34 void BarrierSetAssembler::load_at(MacroAssembler* masm, DecoratorSet decorators, BasicType type, 35 Register dst, Address src, Register tmp1, Register tmp_thread) { 36 bool in_heap = (decorators & IN_HEAP) != 0; 37 bool in_native = (decorators & IN_NATIVE) != 0; 38 bool oop_not_null = (decorators & OOP_NOT_NULL) != 0; 39 bool atomic = (decorators & MO_RELAXED) != 0; 40 41 switch (type) { 42 case T_OBJECT: 43 case T_ARRAY: { 44 if (in_heap) { 45 #ifdef _LP64 46 if (UseCompressedOops) { 47 __ movl(dst, src); 48 if (oop_not_null) { 49 __ decode_heap_oop_not_null(dst); 50 } else { 51 __ decode_heap_oop(dst); 52 } 53 } else 54 #endif 55 { 56 __ movptr(dst, src); 57 } 58 } else { 59 assert(in_native, "why else?"); 60 __ movptr(dst, src); 61 } 62 break; 63 } 64 case T_BOOLEAN: __ load_unsigned_byte(dst, src); break; 65 case T_BYTE: __ load_signed_byte(dst, src); break; 66 case T_CHAR: __ load_unsigned_short(dst, src); break; 67 case T_SHORT: __ load_signed_short(dst, src); break; 68 case T_INT: __ movl (dst, src); break; 69 case T_ADDRESS: __ movptr(dst, src); break; 70 case T_FLOAT: 71 assert(dst == noreg, "only to ftos"); 72 __ load_float(src); 73 break; 74 case T_DOUBLE: 75 assert(dst == noreg, "only to dtos"); 76 __ load_double(src); 77 break; 78 case T_LONG: 79 assert(dst == noreg, "only to ltos"); 80 #ifdef _LP64 81 __ movq(rax, src); 82 #else 83 if (atomic) { 84 __ fild_d(src); // Must load atomically 85 __ subptr(rsp,2*wordSize); // Make space for store 86 __ fistp_d(Address(rsp,0)); 87 __ pop(rax); 88 __ pop(rdx); 89 } else { 90 __ movl(rax, src); 91 __ movl(rdx, src.plus_disp(wordSize)); 92 } 93 #endif 94 break; 95 default: Unimplemented(); 96 } 97 } 98 99 void BarrierSetAssembler::store_at(MacroAssembler* masm, DecoratorSet decorators, BasicType type, 100 Address dst, Register val, Register tmp1, Register tmp2) { 101 bool in_heap = (decorators & IN_HEAP) != 0; 102 bool in_native = (decorators & IN_NATIVE) != 0; 103 bool oop_not_null = (decorators & OOP_NOT_NULL) != 0; 104 bool atomic = (decorators & MO_RELAXED) != 0; 105 106 switch (type) { 107 case T_OBJECT: 108 case T_ARRAY: { 109 if (in_heap) { 110 if (val == noreg) { 111 assert(!oop_not_null, "inconsistent access"); 112 #ifdef _LP64 113 if (UseCompressedOops) { 114 __ movl(dst, (int32_t)NULL_WORD); 115 } else { 116 __ movslq(dst, (int32_t)NULL_WORD); 117 } 118 #else 119 __ movl(dst, (int32_t)NULL_WORD); 120 #endif 121 } else { 122 #ifdef _LP64 123 if (UseCompressedOops) { 124 assert(!dst.uses(val), "not enough registers"); 125 if (oop_not_null) { 126 __ encode_heap_oop_not_null(val); 127 } else { 128 __ encode_heap_oop(val); 129 } 130 __ movl(dst, val); 131 } else 132 #endif 133 { 134 __ movptr(dst, val); 135 } 136 } 137 } else { 138 assert(in_native, "why else?"); 139 assert(val != noreg, "not supported"); 140 __ movptr(dst, val); 141 } 142 break; 143 } 144 case T_BOOLEAN: 145 __ andl(val, 0x1); // boolean is true if LSB is 1 146 __ movb(dst, val); 147 break; 148 case T_BYTE: 149 __ movb(dst, val); 150 break; 151 case T_SHORT: 152 __ movw(dst, val); 153 break; 154 case T_CHAR: 155 __ movw(dst, val); 156 break; 157 case T_INT: 158 __ movl(dst, val); 159 break; 160 case T_LONG: 161 assert(val == noreg, "only tos"); 162 #ifdef _LP64 163 __ movq(dst, rax); 164 #else 165 if (atomic) { 166 __ push(rdx); 167 __ push(rax); // Must update atomically with FIST 168 __ fild_d(Address(rsp,0)); // So load into FPU register 169 __ fistp_d(dst); // and put into memory atomically 170 __ addptr(rsp, 2*wordSize); 171 } else { 172 __ movptr(dst, rax); 173 __ movptr(dst.plus_disp(wordSize), rdx); 174 } 175 #endif 176 break; 177 case T_FLOAT: 178 assert(val == noreg, "only tos"); 179 __ store_float(dst); 180 break; 181 case T_DOUBLE: 182 assert(val == noreg, "only tos"); 183 __ store_double(dst); 184 break; 185 case T_ADDRESS: 186 __ movptr(dst, val); 187 break; 188 default: Unimplemented(); 189 } 190 } 191 192 #ifndef _LP64 193 void BarrierSetAssembler::obj_equals(MacroAssembler* masm, 194 Address obj1, jobject obj2) { 195 __ cmpoop_raw(obj1, obj2); 196 } 197 198 void BarrierSetAssembler::obj_equals(MacroAssembler* masm, 199 Register obj1, jobject obj2) { 200 __ cmpoop_raw(obj1, obj2); 201 } 202 #endif 203 void BarrierSetAssembler::obj_equals(MacroAssembler* masm, 204 Register obj1, Address obj2) { 205 __ cmpptr(obj1, obj2); 206 } 207 208 void BarrierSetAssembler::obj_equals(MacroAssembler* masm, 209 Register obj1, Register obj2) { 210 __ cmpptr(obj1, obj2); 211 } 212 213 void BarrierSetAssembler::try_resolve_jobject_in_native(MacroAssembler* masm, Register jni_env, 214 Register obj, Register tmp, Label& slowpath) { 215 __ clear_jweak_tag(obj); 216 __ movptr(obj, Address(obj, 0)); 217 } 218 219 void BarrierSetAssembler::tlab_allocate(MacroAssembler* masm, 220 Register thread, Register obj, 221 Register var_size_in_bytes, 222 int con_size_in_bytes, 223 Register t1, 224 Register t2, 225 Label& slow_case) { 226 assert_different_registers(obj, t1, t2); 227 assert_different_registers(obj, var_size_in_bytes, t1); 228 Register end = t2; 229 if (!thread->is_valid()) { 230 #ifdef _LP64 231 thread = r15_thread; 232 #else 233 assert(t1->is_valid(), "need temp reg"); 234 thread = t1; 235 __ get_thread(thread); 236 #endif 237 } 238 239 __ verify_tlab(); 240 241 __ movptr(obj, Address(thread, JavaThread::tlab_top_offset())); 242 if (var_size_in_bytes == noreg) { 243 __ lea(end, Address(obj, con_size_in_bytes)); 244 } else { 245 __ lea(end, Address(obj, var_size_in_bytes, Address::times_1)); 246 } 247 __ cmpptr(end, Address(thread, JavaThread::tlab_end_offset())); 248 __ jcc(Assembler::above, slow_case); 249 250 // update the tlab top pointer 251 __ movptr(Address(thread, JavaThread::tlab_top_offset()), end); 252 253 // recover var_size_in_bytes if necessary 254 if (var_size_in_bytes == end) { 255 __ subptr(var_size_in_bytes, obj); 256 } 257 __ verify_tlab(); 258 } 259 260 // Defines obj, preserves var_size_in_bytes 261 void BarrierSetAssembler::eden_allocate(MacroAssembler* masm, 262 Register thread, Register obj, 263 Register var_size_in_bytes, 264 int con_size_in_bytes, 265 Register t1, 266 Label& slow_case) { 267 assert(obj == rax, "obj must be in rax, for cmpxchg"); 268 assert_different_registers(obj, var_size_in_bytes, t1); 269 if (!Universe::heap()->supports_inline_contig_alloc()) { 270 __ jmp(slow_case); 271 } else { 272 Register end = t1; 273 Label retry; 274 __ bind(retry); 275 ExternalAddress heap_top((address) Universe::heap()->top_addr()); 276 __ movptr(obj, heap_top); 277 if (var_size_in_bytes == noreg) { 278 __ lea(end, Address(obj, con_size_in_bytes)); 279 } else { 280 __ lea(end, Address(obj, var_size_in_bytes, Address::times_1)); 281 } 282 // if end < obj then we wrapped around => object too long => slow case 283 __ cmpptr(end, obj); 284 __ jcc(Assembler::below, slow_case); 285 __ cmpptr(end, ExternalAddress((address) Universe::heap()->end_addr())); 286 __ jcc(Assembler::above, slow_case); 287 // Compare obj with the top addr, and if still equal, store the new top addr in 288 // end at the address of the top addr pointer. Sets ZF if was equal, and clears 289 // it otherwise. Use lock prefix for atomicity on MPs. 290 __ locked_cmpxchgptr(end, heap_top); 291 __ jcc(Assembler::notEqual, retry); 292 incr_allocated_bytes(masm, thread, var_size_in_bytes, con_size_in_bytes, thread->is_valid() ? noreg : t1); 293 } 294 } 295 296 void BarrierSetAssembler::incr_allocated_bytes(MacroAssembler* masm, Register thread, 297 Register var_size_in_bytes, 298 int con_size_in_bytes, 299 Register t1) { 300 if (!thread->is_valid()) { 301 #ifdef _LP64 302 thread = r15_thread; 303 #else 304 assert(t1->is_valid(), "need temp reg"); 305 thread = t1; 306 __ get_thread(thread); 307 #endif 308 } 309 310 #ifdef _LP64 311 if (var_size_in_bytes->is_valid()) { 312 __ addq(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())), var_size_in_bytes); 313 } else { 314 __ addq(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())), con_size_in_bytes); 315 } 316 #else 317 if (var_size_in_bytes->is_valid()) { 318 __ addl(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())), var_size_in_bytes); 319 } else { 320 __ addl(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())), con_size_in_bytes); 321 } 322 __ adcl(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())+4), 0); 323 #endif 324 }