140 void emit_exception_entries(ExceptionInfoList* info_list); 141 int emit_deopt_handler(); 142 143 void emit_code(BlockList* hir); 144 void emit_block(BlockBegin* block); 145 void emit_lir_list(LIR_List* list); 146 147 // any last minute peephole optimizations are performed here. In 148 // particular sparc uses this for delay slot filling. 149 void peephole(LIR_List* list); 150 151 void emit_string_compare(LIR_Opr left, LIR_Opr right, LIR_Opr dst, CodeEmitInfo* info); 152 153 void return_op(LIR_Opr result); 154 155 // returns offset of poll instruction 156 int safepoint_poll(LIR_Opr result, CodeEmitInfo* info); 157 158 void const2reg (LIR_Opr src, LIR_Opr dest, LIR_PatchCode patch_code, CodeEmitInfo* info); 159 void const2stack(LIR_Opr src, LIR_Opr dest); 160 void const2mem (LIR_Opr src, LIR_Opr dest, BasicType type, CodeEmitInfo* info); 161 void reg2stack (LIR_Opr src, LIR_Opr dest, BasicType type, bool pop_fpu_stack); 162 void reg2reg (LIR_Opr src, LIR_Opr dest); 163 void reg2mem (LIR_Opr src, LIR_Opr dest, BasicType type, LIR_PatchCode patch_code, CodeEmitInfo* info, bool pop_fpu_stack, bool unaligned); 164 void stack2reg (LIR_Opr src, LIR_Opr dest, BasicType type); 165 void stack2stack(LIR_Opr src, LIR_Opr dest, BasicType type); 166 void mem2reg (LIR_Opr src, LIR_Opr dest, BasicType type, 167 LIR_PatchCode patch_code = lir_patch_none, 168 CodeEmitInfo* info = NULL, bool unaligned = false); 169 170 void prefetchr (LIR_Opr src); 171 void prefetchw (LIR_Opr src); 172 173 void shift_op(LIR_Code code, LIR_Opr left, LIR_Opr count, LIR_Opr dest, LIR_Opr tmp); 174 void shift_op(LIR_Code code, LIR_Opr left, jint count, LIR_Opr dest); 175 176 void move_regs(Register from_reg, Register to_reg); 177 void swap_reg(Register a, Register b); 178 179 void emit_op0(LIR_Op0* op); 180 void emit_op1(LIR_Op1* op); 181 void emit_op2(LIR_Op2* op); 182 void emit_op3(LIR_Op3* op); 183 void emit_opBranch(LIR_OpBranch* op); 184 void emit_opLabel(LIR_OpLabel* op); 185 void emit_arraycopy(LIR_OpArrayCopy* op); 186 void emit_opConvert(LIR_OpConvert* op); 187 void emit_alloc_obj(LIR_OpAllocObj* op); 188 void emit_alloc_array(LIR_OpAllocArray* op); 189 void emit_opTypeCheck(LIR_OpTypeCheck* op); 190 void emit_typecheck_helper(LIR_OpTypeCheck *op, Label* success, Label* failure, Label* obj_is_null); 191 void emit_compare_and_swap(LIR_OpCompareAndSwap* op); 192 void emit_lock(LIR_OpLock* op); 193 void emit_call(LIR_OpJavaCall* op); 194 void emit_rtcall(LIR_OpRTCall* op); 195 void emit_profile_call(LIR_OpProfileCall* op); 196 void emit_delay(LIR_OpDelay* op); 197 198 void arith_op(LIR_Code code, LIR_Opr left, LIR_Opr right, LIR_Opr dest, CodeEmitInfo* info, bool pop_fpu_stack); 199 void arithmetic_idiv(LIR_Code code, LIR_Opr left, LIR_Opr right, LIR_Opr temp, LIR_Opr result, CodeEmitInfo* info); 200 void intrinsic_op(LIR_Code code, LIR_Opr value, LIR_Opr unused, LIR_Opr dest, LIR_Op* op); 201 202 void logic_op(LIR_Code code, LIR_Opr left, LIR_Opr right, LIR_Opr dest); 203 204 void roundfp_op(LIR_Opr src, LIR_Opr tmp, LIR_Opr dest, bool pop_fpu_stack); 205 void move_op(LIR_Opr src, LIR_Opr result, BasicType type, 206 LIR_PatchCode patch_code, CodeEmitInfo* info, bool pop_fpu_stack, bool unaligned); 207 void volatile_move_op(LIR_Opr src, LIR_Opr result, BasicType type, CodeEmitInfo* info); 208 void comp_mem_op(LIR_Opr src, LIR_Opr result, BasicType type, CodeEmitInfo* info); // info set for null exceptions 209 void comp_fl2i(LIR_Code code, LIR_Opr left, LIR_Opr right, LIR_Opr result, LIR_Op2* op); 210 void cmove(LIR_Condition code, LIR_Opr left, LIR_Opr right, LIR_Opr result); 211 212 void call( LIR_OpJavaCall* op, relocInfo::relocType rtype); 213 void ic_call( LIR_OpJavaCall* op); 214 void vtable_call( LIR_OpJavaCall* op); 215 216 void osr_entry(); 217 218 void build_frame(); 219 220 void throw_op(LIR_Opr exceptionPC, LIR_Opr exceptionOop, CodeEmitInfo* info); 221 void unwind_op(LIR_Opr exceptionOop); 222 void monitor_address(int monitor_ix, LIR_Opr dst); 223 224 void align_backward_branch_target(); 225 void align_call(LIR_Code code); 226 | 140 void emit_exception_entries(ExceptionInfoList* info_list); 141 int emit_deopt_handler(); 142 143 void emit_code(BlockList* hir); 144 void emit_block(BlockBegin* block); 145 void emit_lir_list(LIR_List* list); 146 147 // any last minute peephole optimizations are performed here. In 148 // particular sparc uses this for delay slot filling. 149 void peephole(LIR_List* list); 150 151 void emit_string_compare(LIR_Opr left, LIR_Opr right, LIR_Opr dst, CodeEmitInfo* info); 152 153 void return_op(LIR_Opr result); 154 155 // returns offset of poll instruction 156 int safepoint_poll(LIR_Opr result, CodeEmitInfo* info); 157 158 void const2reg (LIR_Opr src, LIR_Opr dest, LIR_PatchCode patch_code, CodeEmitInfo* info); 159 void const2stack(LIR_Opr src, LIR_Opr dest); 160 void const2mem (LIR_Opr src, LIR_Opr dest, BasicType type, CodeEmitInfo* info, bool wide); 161 void reg2stack (LIR_Opr src, LIR_Opr dest, BasicType type, bool pop_fpu_stack); 162 void reg2reg (LIR_Opr src, LIR_Opr dest); 163 void reg2mem (LIR_Opr src, LIR_Opr dest, BasicType type, 164 LIR_PatchCode patch_code, CodeEmitInfo* info, 165 bool pop_fpu_stack, bool unaligned, bool wide); 166 void stack2reg (LIR_Opr src, LIR_Opr dest, BasicType type); 167 void stack2stack(LIR_Opr src, LIR_Opr dest, BasicType type); 168 void mem2reg (LIR_Opr src, LIR_Opr dest, BasicType type, 169 LIR_PatchCode patch_code, 170 CodeEmitInfo* info, bool unaligned, bool wide); 171 172 void prefetchr (LIR_Opr src); 173 void prefetchw (LIR_Opr src); 174 175 void shift_op(LIR_Code code, LIR_Opr left, LIR_Opr count, LIR_Opr dest, LIR_Opr tmp); 176 void shift_op(LIR_Code code, LIR_Opr left, jint count, LIR_Opr dest); 177 178 void move_regs(Register from_reg, Register to_reg); 179 void swap_reg(Register a, Register b); 180 181 void emit_op0(LIR_Op0* op); 182 void emit_op1(LIR_Op1* op); 183 void emit_op2(LIR_Op2* op); 184 void emit_op3(LIR_Op3* op); 185 void emit_opBranch(LIR_OpBranch* op); 186 void emit_opLabel(LIR_OpLabel* op); 187 void emit_arraycopy(LIR_OpArrayCopy* op); 188 void emit_opConvert(LIR_OpConvert* op); 189 void emit_alloc_obj(LIR_OpAllocObj* op); 190 void emit_alloc_array(LIR_OpAllocArray* op); 191 void emit_opTypeCheck(LIR_OpTypeCheck* op); 192 void emit_typecheck_helper(LIR_OpTypeCheck *op, Label* success, Label* failure, Label* obj_is_null); 193 void emit_compare_and_swap(LIR_OpCompareAndSwap* op); 194 void emit_lock(LIR_OpLock* op); 195 void emit_call(LIR_OpJavaCall* op); 196 void emit_rtcall(LIR_OpRTCall* op); 197 void emit_profile_call(LIR_OpProfileCall* op); 198 void emit_delay(LIR_OpDelay* op); 199 200 void arith_op(LIR_Code code, LIR_Opr left, LIR_Opr right, LIR_Opr dest, CodeEmitInfo* info, bool pop_fpu_stack); 201 void arithmetic_idiv(LIR_Code code, LIR_Opr left, LIR_Opr right, LIR_Opr temp, LIR_Opr result, CodeEmitInfo* info); 202 void intrinsic_op(LIR_Code code, LIR_Opr value, LIR_Opr unused, LIR_Opr dest, LIR_Op* op); 203 204 void logic_op(LIR_Code code, LIR_Opr left, LIR_Opr right, LIR_Opr dest); 205 206 void roundfp_op(LIR_Opr src, LIR_Opr tmp, LIR_Opr dest, bool pop_fpu_stack); 207 void move_op(LIR_Opr src, LIR_Opr result, BasicType type, 208 LIR_PatchCode patch_code, CodeEmitInfo* info, bool pop_fpu_stack, bool unaligned, bool wide); 209 void volatile_move_op(LIR_Opr src, LIR_Opr result, BasicType type, CodeEmitInfo* info); 210 void comp_mem_op(LIR_Opr src, LIR_Opr result, BasicType type, CodeEmitInfo* info); // info set for null exceptions 211 void comp_fl2i(LIR_Code code, LIR_Opr left, LIR_Opr right, LIR_Opr result, LIR_Op2* op); 212 void cmove(LIR_Condition code, LIR_Opr left, LIR_Opr right, LIR_Opr result); 213 214 void call( LIR_OpJavaCall* op, relocInfo::relocType rtype); 215 void ic_call( LIR_OpJavaCall* op); 216 void vtable_call( LIR_OpJavaCall* op); 217 218 void osr_entry(); 219 220 void build_frame(); 221 222 void throw_op(LIR_Opr exceptionPC, LIR_Opr exceptionOop, CodeEmitInfo* info); 223 void unwind_op(LIR_Opr exceptionOop); 224 void monitor_address(int monitor_ix, LIR_Opr dst); 225 226 void align_backward_branch_target(); 227 void align_call(LIR_Code code); 228 |