90
91 pre_val = gen->new_register(T_OBJECT);
92
93 if (!addr_opr->is_address()) {
94 assert(addr_opr->is_register(), "must be");
95 addr_opr = LIR_OprFact::address(new LIR_Address(addr_opr, T_OBJECT));
96 }
97 slow = new ShenandoahPreBarrierStub(addr_opr, pre_val, pre_val_patch_code, info ? new CodeEmitInfo(info) : NULL);
98 } else {
99 assert(addr_opr == LIR_OprFact::illegalOpr, "sanity");
100 assert(pre_val->is_register(), "must be");
101 assert(pre_val->type() == T_OBJECT, "must be an object");
102
103 slow = new ShenandoahPreBarrierStub(pre_val);
104 }
105
106 __ branch(lir_cond_notEqual, T_INT, slow);
107 __ branch_destination(slow->continuation());
108 }
109
110 LIR_Opr ShenandoahBarrierSetC1::load_reference_barrier(LIRGenerator* gen, LIR_Opr obj, LIR_Opr addr) {
111 if (ShenandoahLoadRefBarrier) {
112 return load_reference_barrier_impl(gen, obj, addr);
113 } else {
114 return obj;
115 }
116 }
117
118 LIR_Opr ShenandoahBarrierSetC1::load_reference_barrier_impl(LIRGenerator* gen, LIR_Opr obj, LIR_Opr addr) {
119 assert(ShenandoahLoadRefBarrier, "Should be enabled");
120
121 obj = ensure_in_register(gen, obj, T_OBJECT);
122 assert(obj->is_register(), "must be a register at this point");
123 addr = ensure_in_register(gen, addr, T_ADDRESS);
124 assert(addr->is_register(), "must be a register at this point");
125 LIR_Opr result = gen->result_register_for(obj->value_type());
126 __ move(obj, result);
127 LIR_Opr tmp1 = gen->new_register(T_ADDRESS);
128 LIR_Opr tmp2 = gen->new_register(T_ADDRESS);
129
130 LIR_Opr thrd = gen->getThreadPointer();
131 LIR_Address* active_flag_addr =
132 new LIR_Address(thrd,
133 in_bytes(ShenandoahThreadLocalData::gc_state_offset()),
134 T_BYTE);
135 // Read and check the gc-state-flag.
136 LIR_Opr flag_val = gen->new_register(T_INT);
137 __ load(active_flag_addr, flag_val);
138 LIR_Opr mask = LIR_OprFact::intConst(ShenandoahHeap::HAS_FORWARDED |
139 ShenandoahHeap::EVACUATION |
140 ShenandoahHeap::TRAVERSAL);
141 LIR_Opr mask_reg = gen->new_register(T_INT);
142 __ move(mask, mask_reg);
143
144 if (TwoOperandLIRForm) {
145 __ logical_and(flag_val, mask_reg, flag_val);
146 } else {
147 LIR_Opr masked_flag = gen->new_register(T_INT);
148 __ logical_and(flag_val, mask_reg, masked_flag);
149 flag_val = masked_flag;
150 }
151 __ cmp(lir_cond_notEqual, flag_val, LIR_OprFact::intConst(0));
152
153 CodeStub* slow = new ShenandoahLoadReferenceBarrierStub(obj, addr, result, tmp1, tmp2);
154 __ branch(lir_cond_notEqual, T_INT, slow);
155 __ branch_destination(slow->continuation());
156
157 return result;
158 }
159
160 LIR_Opr ShenandoahBarrierSetC1::ensure_in_register(LIRGenerator* gen, LIR_Opr obj, BasicType type) {
161 if (!obj->is_register()) {
162 LIR_Opr obj_reg;
163 if (obj->is_constant()) {
164 obj_reg = gen->new_register(type);
165 __ move(obj, obj_reg);
166 } else {
167 obj_reg = gen->new_pointer_register();
168 __ leal(obj, obj_reg);
169 }
170 obj = obj_reg;
171 }
172 return obj;
173 }
194 // We must resolve in register when patching. This is to avoid
195 // having a patch area in the load barrier stub, since the call
196 // into the runtime to patch will not have the proper oop map.
197 const bool patch_before_barrier = access.is_oop() && (access.decorators() & C1_NEEDS_PATCHING) != 0;
198 return BarrierSetC1::resolve_address(access, resolve_in_register || patch_before_barrier);
199 }
200
201 void ShenandoahBarrierSetC1::load_at_resolved(LIRAccess& access, LIR_Opr result) {
202 // 1: non-reference load, no additional barrier is needed
203 if (!access.is_oop()) {
204 BarrierSetC1::load_at_resolved(access, result);
205 return;
206 }
207
208 LIRGenerator* gen = access.gen();
209 DecoratorSet decorators = access.decorators();
210 BasicType type = access.type();
211
212 // 2: load a reference from src location and apply LRB if ShenandoahLoadRefBarrier is set
213 if (ShenandoahBarrierSet::need_load_reference_barrier(decorators, type)) {
214 if (ShenandoahBarrierSet::use_load_reference_barrier_native(decorators, type)) {
215 BarrierSetC1::load_at_resolved(access, result);
216 LIR_OprList* args = new LIR_OprList();
217 LIR_Opr addr = access.resolved_addr();
218 addr = ensure_in_register(gen, addr, T_ADDRESS);
219 args->append(result);
220 args->append(addr);
221 BasicTypeList signature;
222 signature.append(T_OBJECT);
223 signature.append(T_ADDRESS);
224 LIR_Opr call_result = gen->call_runtime(&signature, args,
225 CAST_FROM_FN_PTR(address, ShenandoahRuntime::load_reference_barrier_native),
226 objectType, NULL);
227 __ move(call_result, result);
228 } else {
229 LIR_Opr tmp = gen->new_register(T_OBJECT);
230 BarrierSetC1::load_at_resolved(access, tmp);
231 tmp = load_reference_barrier(gen, tmp, access.resolved_addr());
232 __ move(tmp, result);
233 }
234 } else {
235 BarrierSetC1::load_at_resolved(access, result);
236 }
237
238 // 3: apply keep-alive barrier if ShenandoahKeepAliveBarrier is set
239 if (ShenandoahKeepAliveBarrier) {
240 bool is_weak = (decorators & ON_WEAK_OOP_REF) != 0;
241 bool is_phantom = (decorators & ON_PHANTOM_OOP_REF) != 0;
242 bool is_anonymous = (decorators & ON_UNKNOWN_OOP_REF) != 0;
243 bool is_traversal_mode = ShenandoahHeap::heap()->is_traversal_mode();
244 bool keep_alive = (decorators & AS_NO_KEEPALIVE) == 0 || is_traversal_mode;
245
246 if ((is_weak || is_phantom || is_anonymous) && keep_alive) {
247 // Register the value in the referent field with the pre-barrier
248 LabelObj *Lcont_anonymous;
249 if (is_anonymous) {
250 Lcont_anonymous = new LabelObj();
251 generate_referent_check(access, Lcont_anonymous);
252 }
253 pre_barrier(gen, access.access_emit_info(), decorators, LIR_OprFact::illegalOpr /* addr_opr */,
254 result /* pre_val */);
255 if (is_anonymous) {
256 __ branch_destination(Lcont_anonymous->label());
257 }
258 }
259 }
260 }
261
262 class C1ShenandoahPreBarrierCodeGenClosure : public StubAssemblerCodeGenClosure {
263 virtual OopMapSet* generate_code(StubAssembler* sasm) {
264 ShenandoahBarrierSetAssembler* bs = (ShenandoahBarrierSetAssembler*)BarrierSet::barrier_set()->barrier_set_assembler();
265 bs->generate_c1_pre_barrier_runtime_stub(sasm);
266 return NULL;
267 }
268 };
269
270 class C1ShenandoahLoadReferenceBarrierCodeGenClosure : public StubAssemblerCodeGenClosure {
271 virtual OopMapSet* generate_code(StubAssembler* sasm) {
272 ShenandoahBarrierSetAssembler* bs = (ShenandoahBarrierSetAssembler*)BarrierSet::barrier_set()->barrier_set_assembler();
273 bs->generate_c1_load_reference_barrier_runtime_stub(sasm);
274 return NULL;
275 }
276 };
277
278 void ShenandoahBarrierSetC1::generate_c1_runtime_stubs(BufferBlob* buffer_blob) {
279 C1ShenandoahPreBarrierCodeGenClosure pre_code_gen_cl;
280 _pre_barrier_c1_runtime_code_blob = Runtime1::generate_blob(buffer_blob, -1,
281 "shenandoah_pre_barrier_slow",
282 false, &pre_code_gen_cl);
283 if (ShenandoahLoadRefBarrier) {
284 C1ShenandoahLoadReferenceBarrierCodeGenClosure lrb_code_gen_cl;
285 _load_reference_barrier_rt_code_blob = Runtime1::generate_blob(buffer_blob, -1,
286 "shenandoah_load_reference_barrier_slow",
287 false, &lrb_code_gen_cl);
288 }
289 }
290
291 const char* ShenandoahBarrierSetC1::rtcall_name_for_address(address entry) {
292 if (entry == CAST_FROM_FN_PTR(address, ShenandoahRuntime::load_reference_barrier_native)) {
293 return "ShenandoahRuntime::load_reference_barrier_native";
294 }
295 return NULL;
296 }
|
90
91 pre_val = gen->new_register(T_OBJECT);
92
93 if (!addr_opr->is_address()) {
94 assert(addr_opr->is_register(), "must be");
95 addr_opr = LIR_OprFact::address(new LIR_Address(addr_opr, T_OBJECT));
96 }
97 slow = new ShenandoahPreBarrierStub(addr_opr, pre_val, pre_val_patch_code, info ? new CodeEmitInfo(info) : NULL);
98 } else {
99 assert(addr_opr == LIR_OprFact::illegalOpr, "sanity");
100 assert(pre_val->is_register(), "must be");
101 assert(pre_val->type() == T_OBJECT, "must be an object");
102
103 slow = new ShenandoahPreBarrierStub(pre_val);
104 }
105
106 __ branch(lir_cond_notEqual, T_INT, slow);
107 __ branch_destination(slow->continuation());
108 }
109
110 LIR_Opr ShenandoahBarrierSetC1::load_reference_barrier(LIRGenerator* gen, LIR_Opr obj, LIR_Opr addr, bool native) {
111 if (ShenandoahLoadRefBarrier) {
112 return load_reference_barrier_impl(gen, obj, addr, native);
113 } else {
114 return obj;
115 }
116 }
117
118 LIR_Opr ShenandoahBarrierSetC1::load_reference_barrier_impl(LIRGenerator* gen, LIR_Opr obj, LIR_Opr addr, bool native) {
119 assert(ShenandoahLoadRefBarrier, "Should be enabled");
120
121 obj = ensure_in_register(gen, obj, T_OBJECT);
122 assert(obj->is_register(), "must be a register at this point");
123 addr = ensure_in_register(gen, addr, T_ADDRESS);
124 assert(addr->is_register(), "must be a register at this point");
125 LIR_Opr result = gen->result_register_for(obj->value_type());
126 __ move(obj, result);
127 LIR_Opr tmp1 = gen->new_register(T_ADDRESS);
128 LIR_Opr tmp2 = gen->new_register(T_ADDRESS);
129
130 LIR_Opr thrd = gen->getThreadPointer();
131 LIR_Address* active_flag_addr =
132 new LIR_Address(thrd,
133 in_bytes(ShenandoahThreadLocalData::gc_state_offset()),
134 T_BYTE);
135 // Read and check the gc-state-flag.
136 LIR_Opr flag_val = gen->new_register(T_INT);
137 __ load(active_flag_addr, flag_val);
138 LIR_Opr mask = LIR_OprFact::intConst(ShenandoahHeap::HAS_FORWARDED |
139 ShenandoahHeap::EVACUATION |
140 ShenandoahHeap::TRAVERSAL);
141 LIR_Opr mask_reg = gen->new_register(T_INT);
142 __ move(mask, mask_reg);
143
144 if (TwoOperandLIRForm) {
145 __ logical_and(flag_val, mask_reg, flag_val);
146 } else {
147 LIR_Opr masked_flag = gen->new_register(T_INT);
148 __ logical_and(flag_val, mask_reg, masked_flag);
149 flag_val = masked_flag;
150 }
151 __ cmp(lir_cond_notEqual, flag_val, LIR_OprFact::intConst(0));
152
153 CodeStub* slow = new ShenandoahLoadReferenceBarrierStub(obj, addr, result, tmp1, tmp2, native);
154 __ branch(lir_cond_notEqual, T_INT, slow);
155 __ branch_destination(slow->continuation());
156
157 return result;
158 }
159
160 LIR_Opr ShenandoahBarrierSetC1::ensure_in_register(LIRGenerator* gen, LIR_Opr obj, BasicType type) {
161 if (!obj->is_register()) {
162 LIR_Opr obj_reg;
163 if (obj->is_constant()) {
164 obj_reg = gen->new_register(type);
165 __ move(obj, obj_reg);
166 } else {
167 obj_reg = gen->new_pointer_register();
168 __ leal(obj, obj_reg);
169 }
170 obj = obj_reg;
171 }
172 return obj;
173 }
194 // We must resolve in register when patching. This is to avoid
195 // having a patch area in the load barrier stub, since the call
196 // into the runtime to patch will not have the proper oop map.
197 const bool patch_before_barrier = access.is_oop() && (access.decorators() & C1_NEEDS_PATCHING) != 0;
198 return BarrierSetC1::resolve_address(access, resolve_in_register || patch_before_barrier);
199 }
200
201 void ShenandoahBarrierSetC1::load_at_resolved(LIRAccess& access, LIR_Opr result) {
202 // 1: non-reference load, no additional barrier is needed
203 if (!access.is_oop()) {
204 BarrierSetC1::load_at_resolved(access, result);
205 return;
206 }
207
208 LIRGenerator* gen = access.gen();
209 DecoratorSet decorators = access.decorators();
210 BasicType type = access.type();
211
212 // 2: load a reference from src location and apply LRB if ShenandoahLoadRefBarrier is set
213 if (ShenandoahBarrierSet::need_load_reference_barrier(decorators, type)) {
214 LIR_Opr tmp = gen->new_register(T_OBJECT);
215 BarrierSetC1::load_at_resolved(access, tmp);
216 bool native = ShenandoahBarrierSet::use_load_reference_barrier_native(decorators, type);
217 tmp = load_reference_barrier(gen, tmp, access.resolved_addr(), native);
218 __ move(tmp, result);
219 } else {
220 BarrierSetC1::load_at_resolved(access, result);
221 }
222
223 // 3: apply keep-alive barrier if ShenandoahKeepAliveBarrier is set
224 if (ShenandoahKeepAliveBarrier) {
225 bool is_weak = (decorators & ON_WEAK_OOP_REF) != 0;
226 bool is_phantom = (decorators & ON_PHANTOM_OOP_REF) != 0;
227 bool is_anonymous = (decorators & ON_UNKNOWN_OOP_REF) != 0;
228 bool is_traversal_mode = ShenandoahHeap::heap()->is_traversal_mode();
229 bool keep_alive = (decorators & AS_NO_KEEPALIVE) == 0 || is_traversal_mode;
230
231 if ((is_weak || is_phantom || is_anonymous) && keep_alive) {
232 // Register the value in the referent field with the pre-barrier
233 LabelObj *Lcont_anonymous;
234 if (is_anonymous) {
235 Lcont_anonymous = new LabelObj();
236 generate_referent_check(access, Lcont_anonymous);
237 }
238 pre_barrier(gen, access.access_emit_info(), decorators, LIR_OprFact::illegalOpr /* addr_opr */,
239 result /* pre_val */);
240 if (is_anonymous) {
241 __ branch_destination(Lcont_anonymous->label());
242 }
243 }
244 }
245 }
246
247 class C1ShenandoahPreBarrierCodeGenClosure : public StubAssemblerCodeGenClosure {
248 virtual OopMapSet* generate_code(StubAssembler* sasm) {
249 ShenandoahBarrierSetAssembler* bs = (ShenandoahBarrierSetAssembler*)BarrierSet::barrier_set()->barrier_set_assembler();
250 bs->generate_c1_pre_barrier_runtime_stub(sasm);
251 return NULL;
252 }
253 };
254
255 class C1ShenandoahLoadReferenceBarrierCodeGenClosure : public StubAssemblerCodeGenClosure {
256 private:
257 bool _native;
258
259 public:
260 C1ShenandoahLoadReferenceBarrierCodeGenClosure(bool native) : _native(native) {}
261
262 virtual OopMapSet* generate_code(StubAssembler* sasm) {
263 ShenandoahBarrierSetAssembler* bs = (ShenandoahBarrierSetAssembler*)BarrierSet::barrier_set()->barrier_set_assembler();
264 bs->generate_c1_load_reference_barrier_runtime_stub(sasm, _native);
265 return NULL;
266 }
267 };
268
269 void ShenandoahBarrierSetC1::generate_c1_runtime_stubs(BufferBlob* buffer_blob) {
270 C1ShenandoahPreBarrierCodeGenClosure pre_code_gen_cl;
271 _pre_barrier_c1_runtime_code_blob = Runtime1::generate_blob(buffer_blob, -1,
272 "shenandoah_pre_barrier_slow",
273 false, &pre_code_gen_cl);
274 if (ShenandoahLoadRefBarrier) {
275 C1ShenandoahLoadReferenceBarrierCodeGenClosure lrb_code_gen_cl(false);
276 _load_reference_barrier_rt_code_blob = Runtime1::generate_blob(buffer_blob, -1,
277 "shenandoah_load_reference_barrier_slow",
278 false, &lrb_code_gen_cl);
279
280 C1ShenandoahLoadReferenceBarrierCodeGenClosure lrb_native_code_gen_cl(true);
281 _load_reference_barrier_native_rt_code_blob = Runtime1::generate_blob(buffer_blob, -1,
282 "shenandoah_load_reference_barrier_native_slow",
283 false, &lrb_native_code_gen_cl);
284 }
285 }
|