--- /dev/null 2016-08-24 15:41:39.598575000 -0400 +++ new/hotspot/src/cpu/arm/vm/c1_FrameMap_arm.cpp 2016-12-02 11:18:13.185249126 -0500 @@ -0,0 +1,230 @@ +/* + * Copyright (c) 2008, 2015, Oracle and/or its affiliates. All rights reserved. + * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. + * + * This code is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 only, as + * published by the Free Software Foundation. + * + * This code is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License + * version 2 for more details (a copy is included in the LICENSE file that + * accompanied this code). + * + * You should have received a copy of the GNU General Public License version + * 2 along with this work; if not, write to the Free Software Foundation, + * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. + * + * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA + * or visit www.oracle.com if you need additional information or have any + * questions. + * + */ + +#include "precompiled.hpp" +#include "c1/c1_FrameMap.hpp" +#include "c1/c1_LIR.hpp" +#include "runtime/sharedRuntime.hpp" +#include "vmreg_arm.inline.hpp" + +LIR_Opr FrameMap::R0_opr; +LIR_Opr FrameMap::R1_opr; +LIR_Opr FrameMap::R2_opr; +LIR_Opr FrameMap::R3_opr; +LIR_Opr FrameMap::R4_opr; +LIR_Opr FrameMap::R5_opr; + +LIR_Opr FrameMap::R0_oop_opr; +LIR_Opr FrameMap::R1_oop_opr; +LIR_Opr FrameMap::R2_oop_opr; +LIR_Opr FrameMap::R3_oop_opr; +LIR_Opr FrameMap::R4_oop_opr; +LIR_Opr FrameMap::R5_oop_opr; + +LIR_Opr FrameMap::R0_metadata_opr; +LIR_Opr FrameMap::R1_metadata_opr; +LIR_Opr FrameMap::R2_metadata_opr; +LIR_Opr FrameMap::R3_metadata_opr; +LIR_Opr FrameMap::R4_metadata_opr; +LIR_Opr FrameMap::R5_metadata_opr; + +#ifdef AARCH64 +LIR_Opr FrameMap::ZR_opr; +#endif // AARCH64 + +LIR_Opr FrameMap::LR_opr; +LIR_Opr FrameMap::LR_oop_opr; +LIR_Opr FrameMap::LR_ptr_opr; +LIR_Opr FrameMap::FP_opr; +LIR_Opr FrameMap::SP_opr; +LIR_Opr FrameMap::Rthread_opr; + +LIR_Opr FrameMap::Int_result_opr; +LIR_Opr FrameMap::Long_result_opr; +LIR_Opr FrameMap::Object_result_opr; +LIR_Opr FrameMap::Float_result_opr; +LIR_Opr FrameMap::Double_result_opr; + +LIR_Opr FrameMap::Exception_oop_opr; +LIR_Opr FrameMap::Exception_pc_opr; + +LIR_Opr FrameMap::_caller_save_cpu_regs[] = { 0 }; +LIR_Opr FrameMap::_caller_save_fpu_regs[]; // same as initialize to zero + +LIR_Opr FrameMap::map_to_opr(BasicType type, VMRegPair* reg, bool) { + LIR_Opr opr = LIR_OprFact::illegalOpr; + VMReg r_1 = reg->first(); + VMReg r_2 = reg->second(); + if (r_1->is_stack()) { + int st_off = (r_1->reg2stack() + SharedRuntime::out_preserve_stack_slots()) * VMRegImpl::stack_slot_size; + opr = LIR_OprFact::address(new LIR_Address(SP_opr, st_off, type)); + } else if (r_1->is_Register()) { + Register reg = r_1->as_Register(); + if (r_2->is_Register() && (type == T_LONG || type == T_DOUBLE)) { +#ifdef AARCH64 + assert(r_1->next() == r_2, "should be the same"); + opr = as_long_opr(reg); +#else + opr = as_long_opr(reg, r_2->as_Register()); +#endif + } else if (type == T_OBJECT || type == T_ARRAY) { + opr = as_oop_opr(reg); + } else if (type == T_METADATA) { + opr = as_metadata_opr(reg); + } else { + // PreferInterpreterNativeStubs should ensure we never need to + // handle a long opr passed as R3+stack_slot + assert(! r_2->is_stack(), "missing support for ALIGN_WIDE_ARGUMENTS==0"); + opr = as_opr(reg); + } + } else if (r_1->is_FloatRegister()) { + FloatRegister reg = r_1->as_FloatRegister(); + opr = type == T_FLOAT ? as_float_opr(reg) : as_double_opr(reg); + } else { + ShouldNotReachHere(); + } + return opr; +} + + +void FrameMap::initialize() { + if (_init_done) return; + + int i; + int rnum = 0; + + // Registers used for allocation +#ifdef AARCH64 + assert(Rthread == R28 && Rheap_base == R27 && Rtemp == R16, "change the code here"); + for (i = 0; i < 16; i++) { + map_register(rnum++, as_Register(i)); + } + for (i = 17; i < 28; i++) { + map_register(rnum++, as_Register(i)); + } +#else + assert(Rthread == R10 && Rtemp == R12, "change the code here"); + for (i = 0; i < 10; i++) { + map_register(rnum++, as_Register(i)); + } +#endif // AARCH64 + assert(rnum == pd_nof_cpu_regs_reg_alloc, "should be"); + + // Registers not used for allocation + map_register(rnum++, LR); // LR register should be listed first, see c1_LinearScan_arm.hpp::is_processed_reg_num. + assert(rnum == pd_nof_cpu_regs_processed_in_linearscan, "should be"); + + map_register(rnum++, Rtemp); + map_register(rnum++, Rthread); + map_register(rnum++, FP); // ARM32: R7 or R11 + map_register(rnum++, SP); +#ifdef AARCH64 + map_register(rnum++, ZR); +#else + map_register(rnum++, PC); +#endif + assert(rnum == pd_nof_cpu_regs_frame_map, "should be"); + + _init_done = true; + + R0_opr = as_opr(R0); R0_oop_opr = as_oop_opr(R0); R0_metadata_opr = as_metadata_opr(R0); + R1_opr = as_opr(R1); R1_oop_opr = as_oop_opr(R1); R1_metadata_opr = as_metadata_opr(R1); + R2_opr = as_opr(R2); R2_oop_opr = as_oop_opr(R2); R2_metadata_opr = as_metadata_opr(R2); + R3_opr = as_opr(R3); R3_oop_opr = as_oop_opr(R3); R3_metadata_opr = as_metadata_opr(R3); + R4_opr = as_opr(R4); R4_oop_opr = as_oop_opr(R4); R4_metadata_opr = as_metadata_opr(R4); + R5_opr = as_opr(R5); R5_oop_opr = as_oop_opr(R5); R5_metadata_opr = as_metadata_opr(R5); + +#ifdef AARCH64 + ZR_opr = as_opr(ZR); +#endif // AARCH64 + + LR_opr = as_opr(LR); + LR_oop_opr = as_oop_opr(LR); + LR_ptr_opr = as_pointer_opr(LR); + FP_opr = as_pointer_opr(FP); + SP_opr = as_pointer_opr(SP); + Rthread_opr = as_pointer_opr(Rthread); + + // LIR operands for result + Int_result_opr = R0_opr; + Object_result_opr = R0_oop_opr; +#ifdef AARCH64 + Long_result_opr = as_long_opr(R0); + Float_result_opr = as_float_opr(S0); + Double_result_opr = as_double_opr(D0); +#else + Long_result_opr = as_long_opr(R0, R1); +#ifdef __ABI_HARD__ + Float_result_opr = as_float_opr(S0); + Double_result_opr = as_double_opr(D0); +#else + Float_result_opr = LIR_OprFact::single_softfp(0); + Double_result_opr = LIR_OprFact::double_softfp(0, 1); +#endif // __ABI_HARD__ +#endif // AARCH64 + + Exception_oop_opr = as_oop_opr(Rexception_obj); + Exception_pc_opr = as_opr(Rexception_pc); + + for (i = 0; i < nof_caller_save_cpu_regs(); i++) { + _caller_save_cpu_regs[i] = LIR_OprFact::single_cpu(i); + } + for (i = 0; i < nof_caller_save_fpu_regs; i++) { + _caller_save_fpu_regs[i] = LIR_OprFact::single_fpu(i); + } +} + + +Address FrameMap::make_new_address(ByteSize sp_offset) const { + return Address(SP, sp_offset); +} + +LIR_Opr FrameMap::stack_pointer() { + return FrameMap::SP_opr; +} + +LIR_Opr FrameMap::method_handle_invoke_SP_save_opr() { + assert(Rmh_SP_save == FP, "Fix register used for saving SP for MethodHandle calls"); + return FP_opr; +} + +bool FrameMap::validate_frame() { + int max_offset = in_bytes(framesize_in_bytes()); + int java_index = 0; + for (int i = 0; i < _incoming_arguments->length(); i++) { + LIR_Opr opr = _incoming_arguments->at(i); + if (opr->is_stack()) { + int arg_offset = _argument_locations->at(java_index); + if (arg_offset > max_offset) { + max_offset = arg_offset; + } + } + java_index += type2size[opr->type()]; + } + return max_offset < AARCH64_ONLY(16384) NOT_AARCH64(4096); // TODO-AARCH64 check that LIRAssembler does not generate load/store of byte and half-word with SP as address base +} + +VMReg FrameMap::fpu_regname(int n) { + return as_FloatRegister(n)->as_VMReg(); +}