1 /* 2 * Copyright (c) 1997, 2015, Oracle and/or its affiliates. All rights reserved. 3 * Copyright (c) 2014, Red Hat Inc. All rights reserved. 4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 5 * 6 * This code is free software; you can redistribute it and/or modify it 7 * under the terms of the GNU General Public License version 2 only, as 8 * published by the Free Software Foundation. 9 * 10 * This code is distributed in the hope that it will be useful, but WITHOUT 11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 12 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 13 * version 2 for more details (a copy is included in the LICENSE file that 14 * accompanied this code). 15 * 16 * You should have received a copy of the GNU General Public License version 17 * 2 along with this work; if not, write to the Free Software Foundation, 18 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 19 * 20 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 21 * or visit www.oracle.com if you need additional information or have any 22 * questions. 23 * 24 */ 25 26 #ifndef CPU_AARCH64_VM_FRAME_AARCH64_INLINE_HPP 27 #define CPU_AARCH64_VM_FRAME_AARCH64_INLINE_HPP 28 29 #include "code/codeCache.hpp" 30 #include "code/vmreg.inline.hpp" 31 32 // Inline functions for AArch64 frames: 33 34 // Constructors: 35 36 inline frame::frame() { 37 _pc = NULL; 38 _sp = NULL; 39 _unextended_sp = NULL; 40 _fp = NULL; 41 _cb = NULL; 42 _deopt_state = unknown; 43 } 44 45 static int spin; 46 47 inline void frame::init(intptr_t* sp, intptr_t* fp, address pc) { 48 intptr_t a = intptr_t(sp); 49 intptr_t b = intptr_t(fp); 50 _sp = sp; 51 _unextended_sp = sp; 52 _fp = fp; 53 _pc = pc; 54 assert(pc != NULL, "no pc?"); 55 _cb = CodeCache::find_blob(pc); 56 adjust_unextended_sp(); 57 58 address original_pc = nmethod::get_deopt_original_pc(this); 59 if (original_pc != NULL) { 60 _pc = original_pc; 61 _deopt_state = is_deoptimized; 62 } else { 63 _deopt_state = not_deoptimized; 64 } 65 } 66 67 inline frame::frame(intptr_t* sp, intptr_t* fp, address pc) { 68 init(sp, fp, pc); 69 } 70 71 inline frame::frame(intptr_t* sp, intptr_t* unextended_sp, intptr_t* fp, address pc) { 72 intptr_t a = intptr_t(sp); 73 intptr_t b = intptr_t(fp); 74 _sp = sp; 75 _unextended_sp = unextended_sp; 76 _fp = fp; 77 _pc = pc; 78 assert(pc != NULL, "no pc?"); 79 _cb = CodeCache::find_blob(pc); 80 adjust_unextended_sp(); 81 82 address original_pc = nmethod::get_deopt_original_pc(this); 83 if (original_pc != NULL) { 84 _pc = original_pc; 85 assert(((nmethod*)_cb)->insts_contains(_pc), "original PC must be in nmethod"); 86 _deopt_state = is_deoptimized; 87 } else { 88 _deopt_state = not_deoptimized; 89 } 90 } 91 92 inline frame::frame(intptr_t* sp, intptr_t* fp) { 93 intptr_t a = intptr_t(sp); 94 intptr_t b = intptr_t(fp); 95 _sp = sp; 96 _unextended_sp = sp; 97 _fp = fp; 98 _pc = (address)(sp[-1]); 99 100 // Here's a sticky one. This constructor can be called via AsyncGetCallTrace 101 // when last_Java_sp is non-null but the pc fetched is junk. If we are truly 102 // unlucky the junk value could be to a zombied method and we'll die on the 103 // find_blob call. This is also why we can have no asserts on the validity 104 // of the pc we find here. AsyncGetCallTrace -> pd_get_top_frame_for_signal_handler 105 // -> pd_last_frame should use a specialized version of pd_last_frame which could 106 // call a specilaized frame constructor instead of this one. 107 // Then we could use the assert below. However this assert is of somewhat dubious 108 // value. 109 // assert(_pc != NULL, "no pc?"); 110 111 _cb = CodeCache::find_blob(_pc); 112 adjust_unextended_sp(); 113 114 address original_pc = nmethod::get_deopt_original_pc(this); 115 if (original_pc != NULL) { 116 _pc = original_pc; 117 _deopt_state = is_deoptimized; 118 } else { 119 _deopt_state = not_deoptimized; 120 } 121 } 122 123 // Accessors 124 125 inline bool frame::equal(frame other) const { 126 bool ret = sp() == other.sp() 127 && unextended_sp() == other.unextended_sp() 128 && fp() == other.fp() 129 && pc() == other.pc(); 130 assert(!ret || ret && cb() == other.cb() && _deopt_state == other._deopt_state, "inconsistent construction"); 131 return ret; 132 } 133 134 // Return unique id for this frame. The id must have a value where we can distinguish 135 // identity and younger/older relationship. NULL represents an invalid (incomparable) 136 // frame. 137 inline intptr_t* frame::id(void) const { return unextended_sp(); } 138 139 // Relationals on frames based 140 // Return true if the frame is younger (more recent activation) than the frame represented by id 141 inline bool frame::is_younger(intptr_t* id) const { assert(this->id() != NULL && id != NULL, "NULL frame id"); 142 return this->id() < id ; } 143 144 // Return true if the frame is older (less recent activation) than the frame represented by id 145 inline bool frame::is_older(intptr_t* id) const { assert(this->id() != NULL && id != NULL, "NULL frame id"); 146 return this->id() > id ; } 147 148 149 150 inline intptr_t* frame::link() const { return (intptr_t*) *(intptr_t **)addr_at(link_offset); } 151 152 153 inline intptr_t* frame::unextended_sp() const { return _unextended_sp; } 154 155 // Return address: 156 157 inline address* frame::sender_pc_addr() const { return (address*) addr_at( return_addr_offset); } 158 inline address frame::sender_pc() const { return *sender_pc_addr(); } 159 160 #ifdef CC_INTERP 161 162 inline interpreterState frame::get_interpreterState() const { 163 return ((interpreterState)addr_at( -((int)sizeof(BytecodeInterpreter))/wordSize )); 164 } 165 166 inline intptr_t* frame::sender_sp() const { 167 // Hmm this seems awfully expensive QQQ, is this really called with interpreted frames? 168 if (is_interpreted_frame()) { 169 assert(false, "should never happen"); 170 return get_interpreterState()->sender_sp(); 171 } else { 172 return addr_at(sender_sp_offset); 173 } 174 } 175 176 inline intptr_t** frame::interpreter_frame_locals_addr() const { 177 assert(is_interpreted_frame(), "must be interpreted"); 178 return &(get_interpreterState()->_locals); 179 } 180 181 inline intptr_t* frame::interpreter_frame_bcx_addr() const { 182 assert(is_interpreted_frame(), "must be interpreted"); 183 return (intptr_t*) &(get_interpreterState()->_bcp); 184 } 185 186 187 // Constant pool cache 188 189 inline constantPoolCacheOop* frame::interpreter_frame_cache_addr() const { 190 assert(is_interpreted_frame(), "must be interpreted"); 191 return &(get_interpreterState()->_constants); 192 } 193 194 // Method 195 196 inline methodOop* frame::interpreter_frame_method_addr() const { 197 assert(is_interpreted_frame(), "must be interpreted"); 198 return &(get_interpreterState()->_method); 199 } 200 201 inline intptr_t* frame::interpreter_frame_mdx_addr() const { 202 assert(is_interpreted_frame(), "must be interpreted"); 203 return (intptr_t*) &(get_interpreterState()->_mdx); 204 } 205 206 // top of expression stack 207 inline intptr_t* frame::interpreter_frame_tos_address() const { 208 assert(is_interpreted_frame(), "wrong frame type"); 209 return get_interpreterState()->_stack + 1; 210 } 211 212 #else /* asm interpreter */ 213 inline intptr_t* frame::sender_sp() const { return addr_at( sender_sp_offset); } 214 215 inline intptr_t** frame::interpreter_frame_locals_addr() const { 216 return (intptr_t**)addr_at(interpreter_frame_locals_offset); 217 } 218 219 inline intptr_t* frame::interpreter_frame_last_sp() const { 220 return *(intptr_t**)addr_at(interpreter_frame_last_sp_offset); 221 } 222 223 inline intptr_t* frame::interpreter_frame_bcp_addr() const { 224 return (intptr_t*)addr_at(interpreter_frame_bcp_offset); 225 } 226 227 inline intptr_t* frame::interpreter_frame_mdp_addr() const { 228 return (intptr_t*)addr_at(interpreter_frame_mdp_offset); 229 } 230 231 232 // Constant pool cache 233 234 inline ConstantPoolCache** frame::interpreter_frame_cache_addr() const { 235 return (ConstantPoolCache**)addr_at(interpreter_frame_cache_offset); 236 } 237 238 // Method 239 240 inline Method** frame::interpreter_frame_method_addr() const { 241 return (Method**)addr_at(interpreter_frame_method_offset); 242 } 243 244 // top of expression stack 245 inline intptr_t* frame::interpreter_frame_tos_address() const { 246 intptr_t* last_sp = interpreter_frame_last_sp(); 247 if (last_sp == NULL) { 248 return sp(); 249 } else { 250 // sp() may have been extended or shrunk by an adapter. At least 251 // check that we don't fall behind the legal region. 252 // For top deoptimized frame last_sp == interpreter_frame_monitor_end. 253 assert(last_sp <= (intptr_t*) interpreter_frame_monitor_end(), "bad tos"); 254 return last_sp; 255 } 256 } 257 258 inline oop* frame::interpreter_frame_temp_oop_addr() const { 259 return (oop *)(fp() + interpreter_frame_oop_temp_offset); 260 } 261 262 #endif /* CC_INTERP */ 263 264 inline int frame::pd_oop_map_offset_adjustment() const { 265 return 0; 266 } 267 268 inline int frame::interpreter_frame_monitor_size() { 269 return BasicObjectLock::size(); 270 } 271 272 273 // expression stack 274 // (the max_stack arguments are used by the GC; see class FrameClosure) 275 276 inline intptr_t* frame::interpreter_frame_expression_stack() const { 277 intptr_t* monitor_end = (intptr_t*) interpreter_frame_monitor_end(); 278 return monitor_end-1; 279 } 280 281 282 inline jint frame::interpreter_frame_expression_stack_direction() { return -1; } 283 284 285 // Entry frames 286 287 inline JavaCallWrapper** frame::entry_frame_call_wrapper_addr() const { 288 return (JavaCallWrapper**)addr_at(entry_frame_call_wrapper_offset); 289 } 290 291 292 // Compiled frames 293 294 inline int frame::local_offset_for_compiler(int local_index, int nof_args, int max_nof_locals, int max_nof_monitors) { 295 return (nof_args - local_index + (local_index < nof_args ? 1: -1)); 296 } 297 298 inline int frame::monitor_offset_for_compiler(int local_index, int nof_args, int max_nof_locals, int max_nof_monitors) { 299 return local_offset_for_compiler(local_index, nof_args, max_nof_locals, max_nof_monitors); 300 } 301 302 inline int frame::min_local_offset_for_compiler(int nof_args, int max_nof_locals, int max_nof_monitors) { 303 return (nof_args - (max_nof_locals + max_nof_monitors*2) - 1); 304 } 305 306 inline bool frame::volatile_across_calls(Register reg) { 307 return true; 308 } 309 310 311 312 inline oop frame::saved_oop_result(RegisterMap* map) const { 313 oop* result_adr = (oop *)map->location(r0->as_VMReg()); 314 guarantee(result_adr != NULL, "bad register save location"); 315 316 return (*result_adr); 317 } 318 319 inline void frame::set_saved_oop_result(RegisterMap* map, oop obj) { 320 oop* result_adr = (oop *)map->location(r0->as_VMReg()); 321 guarantee(result_adr != NULL, "bad register save location"); 322 323 *result_adr = obj; 324 } 325 326 #endif // CPU_AARCH64_VM_FRAME_AARCH64_INLINE_HPP