1 /*
   2  * Copyright 1997-2008 Sun Microsystems, Inc.  All Rights Reserved.
   3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4  *
   5  * This code is free software; you can redistribute it and/or modify it
   6  * under the terms of the GNU General Public License version 2 only, as
   7  * published by the Free Software Foundation.
   8  *
   9  * This code is distributed in the hope that it will be useful, but WITHOUT
  10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12  * version 2 for more details (a copy is included in the LICENSE file that
  13  * accompanied this code).
  14  *
  15  * You should have received a copy of the GNU General Public License version
  16  * 2 along with this work; if not, write to the Free Software Foundation,
  17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18  *
  19  * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
  20  * CA 95054 USA or visit www.sun.com if you need additional information or
  21  * have any questions.
  22  *
  23  */
  24 
  25 // Portions of code courtesy of Clifford Click
  26 
  27 // Optimization - Graph Style
  28 
  29 class Chaitin;
  30 class NamedCounter;
  31 class MultiNode;
  32 class  SafePointNode;
  33 class   CallNode;
  34 class     CallJavaNode;
  35 class       CallStaticJavaNode;
  36 class       CallDynamicJavaNode;
  37 class     CallRuntimeNode;
  38 class       CallLeafNode;
  39 class         CallLeafNoFPNode;
  40 class     AllocateNode;
  41 class       AllocateArrayNode;
  42 class     LockNode;
  43 class     UnlockNode;
  44 class JVMState;
  45 class OopMap;
  46 class State;
  47 class StartNode;
  48 class MachCallNode;
  49 class FastLockNode;
  50 
  51 //------------------------------StartNode--------------------------------------
  52 // The method start node
  53 class StartNode : public MultiNode {
  54   virtual uint cmp( const Node &n ) const;
  55   virtual uint size_of() const; // Size is bigger
  56 public:
  57   const TypeTuple *_domain;
  58   StartNode( Node *root, const TypeTuple *domain ) : MultiNode(2), _domain(domain) {
  59     init_class_id(Class_Start);
  60     init_flags(Flag_is_block_start);
  61     init_req(0,this);
  62     init_req(1,root);
  63   }
  64   virtual int Opcode() const;
  65   virtual bool pinned() const { return true; };
  66   virtual const Type *bottom_type() const;
  67   virtual const TypePtr *adr_type() const { return TypePtr::BOTTOM; }
  68   virtual const Type *Value( PhaseTransform *phase ) const;
  69   virtual Node *Ideal(PhaseGVN *phase, bool can_reshape);
  70   virtual void  calling_convention( BasicType* sig_bt, VMRegPair *parm_reg, uint length ) const;
  71   virtual const RegMask &in_RegMask(uint) const;
  72   virtual Node *match( const ProjNode *proj, const Matcher *m );
  73   virtual uint ideal_reg() const { return 0; }
  74 #ifndef PRODUCT
  75   virtual void  dump_spec(outputStream *st) const;
  76 #endif
  77 };
  78 
  79 //------------------------------StartOSRNode-----------------------------------
  80 // The method start node for on stack replacement code
  81 class StartOSRNode : public StartNode {
  82 public:
  83   StartOSRNode( Node *root, const TypeTuple *domain ) : StartNode(root, domain) {}
  84   virtual int   Opcode() const;
  85   static  const TypeTuple *osr_domain();
  86 };
  87 
  88 
  89 //------------------------------ParmNode---------------------------------------
  90 // Incoming parameters
  91 class ParmNode : public ProjNode {
  92   static const char * const names[TypeFunc::Parms+1];
  93 public:
  94   ParmNode( StartNode *src, uint con ) : ProjNode(src,con) {
  95     init_class_id(Class_Parm);
  96   }
  97   virtual int Opcode() const;
  98   virtual bool  is_CFG() const { return (_con == TypeFunc::Control); }
  99   virtual uint ideal_reg() const;
 100 #ifndef PRODUCT
 101   virtual void dump_spec(outputStream *st) const;
 102 #endif
 103 };
 104 
 105 
 106 //------------------------------ReturnNode-------------------------------------
 107 // Return from subroutine node
 108 class ReturnNode : public Node {
 109 public:
 110   ReturnNode( uint edges, Node *cntrl, Node *i_o, Node *memory, Node *retadr, Node *frameptr );
 111   virtual int Opcode() const;
 112   virtual bool  is_CFG() const { return true; }
 113   virtual uint hash() const { return NO_HASH; }  // CFG nodes do not hash
 114   virtual bool depends_only_on_test() const { return false; }
 115   virtual Node *Ideal(PhaseGVN *phase, bool can_reshape);
 116   virtual const Type *Value( PhaseTransform *phase ) const;
 117   virtual uint ideal_reg() const { return NotAMachineReg; }
 118   virtual uint match_edge(uint idx) const;
 119 #ifndef PRODUCT
 120   virtual void dump_req() const;
 121 #endif
 122 };
 123 
 124 
 125 //------------------------------RethrowNode------------------------------------
 126 // Rethrow of exception at call site.  Ends a procedure before rethrowing;
 127 // ends the current basic block like a ReturnNode.  Restores registers and
 128 // unwinds stack.  Rethrow happens in the caller's method.
 129 class RethrowNode : public Node {
 130  public:
 131   RethrowNode( Node *cntrl, Node *i_o, Node *memory, Node *frameptr, Node *ret_adr, Node *exception );
 132   virtual int Opcode() const;
 133   virtual bool  is_CFG() const { return true; }
 134   virtual uint hash() const { return NO_HASH; }  // CFG nodes do not hash
 135   virtual bool depends_only_on_test() const { return false; }
 136   virtual Node *Ideal(PhaseGVN *phase, bool can_reshape);
 137   virtual const Type *Value( PhaseTransform *phase ) const;
 138   virtual uint match_edge(uint idx) const;
 139   virtual uint ideal_reg() const { return NotAMachineReg; }
 140 #ifndef PRODUCT
 141   virtual void dump_req() const;
 142 #endif
 143 };
 144 
 145 
 146 //------------------------------TailCallNode-----------------------------------
 147 // Pop stack frame and jump indirect
 148 class TailCallNode : public ReturnNode {
 149 public:
 150   TailCallNode( Node *cntrl, Node *i_o, Node *memory, Node *frameptr, Node *retadr, Node *target, Node *moop )
 151     : ReturnNode( TypeFunc::Parms+2, cntrl, i_o, memory, frameptr, retadr ) {
 152     init_req(TypeFunc::Parms, target);
 153     init_req(TypeFunc::Parms+1, moop);
 154   }
 155 
 156   virtual int Opcode() const;
 157   virtual uint match_edge(uint idx) const;
 158 };
 159 
 160 //------------------------------TailJumpNode-----------------------------------
 161 // Pop stack frame and jump indirect
 162 class TailJumpNode : public ReturnNode {
 163 public:
 164   TailJumpNode( Node *cntrl, Node *i_o, Node *memory, Node *frameptr, Node *target, Node *ex_oop)
 165     : ReturnNode(TypeFunc::Parms+2, cntrl, i_o, memory, frameptr, Compile::current()->top()) {
 166     init_req(TypeFunc::Parms, target);
 167     init_req(TypeFunc::Parms+1, ex_oop);
 168   }
 169 
 170   virtual int Opcode() const;
 171   virtual uint match_edge(uint idx) const;
 172 };
 173 
 174 //-------------------------------JVMState-------------------------------------
 175 // A linked list of JVMState nodes captures the whole interpreter state,
 176 // plus GC roots, for all active calls at some call site in this compilation
 177 // unit.  (If there is no inlining, then the list has exactly one link.)
 178 // This provides a way to map the optimized program back into the interpreter,
 179 // or to let the GC mark the stack.
 180 class JVMState : public ResourceObj {
 181 private:
 182   JVMState*         _caller;    // List pointer for forming scope chains
 183   uint              _depth;     // One mroe than caller depth, or one.
 184   uint              _locoff;    // Offset to locals in input edge mapping
 185   uint              _stkoff;    // Offset to stack in input edge mapping
 186   uint              _monoff;    // Offset to monitors in input edge mapping
 187   uint              _scloff;    // Offset to fields of scalar objs in input edge mapping
 188   uint              _endoff;    // Offset to end of input edge mapping
 189   uint              _sp;        // Jave Expression Stack Pointer for this state
 190   int               _bci;       // Byte Code Index of this JVM point
 191   bool              _reexecute; // Whether this bytecode need to be re-executed
 192   ciMethod*         _method;    // Method Pointer
 193   SafePointNode*    _map;       // Map node associated with this scope
 194 public:
 195   friend class Compile;
 196 
 197   // Because JVMState objects live over the entire lifetime of the
 198   // Compile object, they are allocated into the comp_arena, which
 199   // does not get resource marked or reset during the compile process
 200   void *operator new( size_t x, Compile* C ) { return C->comp_arena()->Amalloc(x); }
 201   void operator delete( void * ) { } // fast deallocation
 202 
 203   // Create a new JVMState, ready for abstract interpretation.
 204   JVMState(ciMethod* method, JVMState* caller);
 205   JVMState(int stack_size);  // root state; has a null method
 206 
 207   // Access functions for the JVM
 208   uint              locoff() const { return _locoff; }
 209   uint              stkoff() const { return _stkoff; }
 210   uint              argoff() const { return _stkoff + _sp; }
 211   uint              monoff() const { return _monoff; }
 212   uint              scloff() const { return _scloff; }
 213   uint              endoff() const { return _endoff; }
 214   uint              oopoff() const { return debug_end(); }
 215 
 216   int            loc_size() const { return _stkoff - _locoff; }
 217   int            stk_size() const { return _monoff - _stkoff; }
 218   int            mon_size() const { return _scloff - _monoff; }
 219   int            scl_size() const { return _endoff - _scloff; }
 220 
 221   bool        is_loc(uint i) const { return i >= _locoff && i < _stkoff; }
 222   bool        is_stk(uint i) const { return i >= _stkoff && i < _monoff; }
 223   bool        is_mon(uint i) const { return i >= _monoff && i < _scloff; }
 224   bool        is_scl(uint i) const { return i >= _scloff && i < _endoff; }
 225 
 226   uint                      sp()     const { return _sp; }
 227   int                      bci()    const { return _bci; }
 228   bool        should_reexecute() const { return _reexecute; }
 229   bool              has_method() const { return _method != NULL; }
 230   ciMethod*             method() const { assert(has_method(), ""); return _method; }
 231   JVMState*             caller() const { return _caller; }
 232   SafePointNode*           map()    const { return _map; }
 233   uint                   depth()  const { return _depth; }
 234   uint             debug_start()  const; // returns locoff of root caller
 235   uint               debug_end()    const; // returns endoff of self
 236   uint              debug_size()   const {
 237     return loc_size() + sp() + mon_size() + scl_size();
 238   }
 239   uint        debug_depth()  const; // returns sum of debug_size values at all depths
 240 
 241   // Returns the JVM state at the desired depth (1 == root).
 242   JVMState* of_depth(int d) const;
 243 
 244   // Tells if two JVM states have the same call chain (depth, methods, & bcis).
 245   bool same_calls_as(const JVMState* that) const;
 246 
 247   // Monitors (monitors are stored as (boxNode, objNode) pairs
 248   enum { logMonitorEdges = 1 };
 249   int  nof_monitors()              const { return mon_size() >> logMonitorEdges; }
 250   int  monitor_depth()             const { return nof_monitors() + (caller() ? caller()->monitor_depth() : 0); }
 251   int  monitor_box_offset(int idx) const { return monoff() + (idx << logMonitorEdges) + 0; }
 252   int  monitor_obj_offset(int idx) const { return monoff() + (idx << logMonitorEdges) + 1; }
 253   bool is_monitor_box(uint off)    const {
 254     assert(is_mon(off), "should be called only for monitor edge");
 255     return (0 == bitfield(off - monoff(), 0, logMonitorEdges));
 256   }
 257   bool is_monitor_use(uint off)    const { return (is_mon(off)
 258                                                    && is_monitor_box(off))
 259                                              || (caller() && caller()->is_monitor_use(off)); }
 260 
 261   // Initialization functions for the JVM
 262   void              set_locoff(uint off) { _locoff = off; }
 263   void              set_stkoff(uint off) { _stkoff = off; }
 264   void              set_monoff(uint off) { _monoff = off; }
 265   void              set_scloff(uint off) { _scloff = off; }
 266   void              set_endoff(uint off) { _endoff = off; }
 267   void              set_offsets(uint off) {
 268     _locoff = _stkoff = _monoff = _scloff = _endoff = off;
 269   }
 270   void              set_map(SafePointNode *map) { _map = map; }
 271   void              set_sp(uint sp) { _sp = sp; }
 272   void              set_bci(int bci) { _bci = bci; }
 273   void              set_reexecute(bool reexec) {_reexecute = reexec;}
 274 
 275   // Miscellaneous utility functions
 276   JVMState* clone_deep(Compile* C) const;    // recursively clones caller chain
 277   JVMState* clone_shallow(Compile* C) const; // retains uncloned caller
 278 
 279 #ifndef PRODUCT
 280   void      format(PhaseRegAlloc *regalloc, const Node *n, outputStream* st) const;
 281   void      dump_spec(outputStream *st) const;
 282   void      dump_on(outputStream* st) const;
 283   void      dump() const {
 284     dump_on(tty);
 285   }
 286 #endif
 287 };
 288 
 289 //------------------------------SafePointNode----------------------------------
 290 // A SafePointNode is a subclass of a MultiNode for convenience (and
 291 // potential code sharing) only - conceptually it is independent of
 292 // the Node semantics.
 293 class SafePointNode : public MultiNode {
 294   virtual uint           cmp( const Node &n ) const;
 295   virtual uint           size_of() const;       // Size is bigger
 296 
 297 public:
 298   SafePointNode(uint edges, JVMState* jvms,
 299                 // A plain safepoint advertises no memory effects (NULL):
 300                 const TypePtr* adr_type = NULL)
 301     : MultiNode( edges ),
 302       _jvms(jvms),
 303       _oop_map(NULL),
 304       _adr_type(adr_type)
 305   {
 306     init_class_id(Class_SafePoint);
 307   }
 308 
 309   OopMap*         _oop_map;   // Array of OopMap info (8-bit char) for GC
 310   JVMState* const _jvms;      // Pointer to list of JVM State objects
 311   const TypePtr*  _adr_type;  // What type of memory does this node produce?
 312 
 313   // Many calls take *all* of memory as input,
 314   // but some produce a limited subset of that memory as output.
 315   // The adr_type reports the call's behavior as a store, not a load.
 316 
 317   virtual JVMState* jvms() const { return _jvms; }
 318   void set_jvms(JVMState* s) {
 319     *(JVMState**)&_jvms = s;  // override const attribute in the accessor
 320   }
 321   OopMap *oop_map() const { return _oop_map; }
 322   void set_oop_map(OopMap *om) { _oop_map = om; }
 323 
 324   // Functionality from old debug nodes which has changed
 325   Node *local(JVMState* jvms, uint idx) const {
 326     assert(verify_jvms(jvms), "jvms must match");
 327     return in(jvms->locoff() + idx);
 328   }
 329   Node *stack(JVMState* jvms, uint idx) const {
 330     assert(verify_jvms(jvms), "jvms must match");
 331     return in(jvms->stkoff() + idx);
 332   }
 333   Node *argument(JVMState* jvms, uint idx) const {
 334     assert(verify_jvms(jvms), "jvms must match");
 335     return in(jvms->argoff() + idx);
 336   }
 337   Node *monitor_box(JVMState* jvms, uint idx) const {
 338     assert(verify_jvms(jvms), "jvms must match");
 339     return in(jvms->monitor_box_offset(idx));
 340   }
 341   Node *monitor_obj(JVMState* jvms, uint idx) const {
 342     assert(verify_jvms(jvms), "jvms must match");
 343     return in(jvms->monitor_obj_offset(idx));
 344   }
 345 
 346   void  set_local(JVMState* jvms, uint idx, Node *c);
 347 
 348   void  set_stack(JVMState* jvms, uint idx, Node *c) {
 349     assert(verify_jvms(jvms), "jvms must match");
 350     set_req(jvms->stkoff() + idx, c);
 351   }
 352   void  set_argument(JVMState* jvms, uint idx, Node *c) {
 353     assert(verify_jvms(jvms), "jvms must match");
 354     set_req(jvms->argoff() + idx, c);
 355   }
 356   void ensure_stack(JVMState* jvms, uint stk_size) {
 357     assert(verify_jvms(jvms), "jvms must match");
 358     int grow_by = (int)stk_size - (int)jvms->stk_size();
 359     if (grow_by > 0)  grow_stack(jvms, grow_by);
 360   }
 361   void grow_stack(JVMState* jvms, uint grow_by);
 362   // Handle monitor stack
 363   void push_monitor( const FastLockNode *lock );
 364   void pop_monitor ();
 365   Node *peek_monitor_box() const;
 366   Node *peek_monitor_obj() const;
 367 
 368   // Access functions for the JVM
 369   Node *control  () const { return in(TypeFunc::Control  ); }
 370   Node *i_o      () const { return in(TypeFunc::I_O      ); }
 371   Node *memory   () const { return in(TypeFunc::Memory   ); }
 372   Node *returnadr() const { return in(TypeFunc::ReturnAdr); }
 373   Node *frameptr () const { return in(TypeFunc::FramePtr ); }
 374 
 375   void set_control  ( Node *c ) { set_req(TypeFunc::Control,c); }
 376   void set_i_o      ( Node *c ) { set_req(TypeFunc::I_O    ,c); }
 377   void set_memory   ( Node *c ) { set_req(TypeFunc::Memory ,c); }
 378 
 379   MergeMemNode* merged_memory() const {
 380     return in(TypeFunc::Memory)->as_MergeMem();
 381   }
 382 
 383   // The parser marks useless maps as dead when it's done with them:
 384   bool is_killed() { return in(TypeFunc::Control) == NULL; }
 385 
 386   // Exception states bubbling out of subgraphs such as inlined calls
 387   // are recorded here.  (There might be more than one, hence the "next".)
 388   // This feature is used only for safepoints which serve as "maps"
 389   // for JVM states during parsing, intrinsic expansion, etc.
 390   SafePointNode*         next_exception() const;
 391   void               set_next_exception(SafePointNode* n);
 392   bool                   has_exceptions() const { return next_exception() != NULL; }
 393 
 394   // Standard Node stuff
 395   virtual int            Opcode() const;
 396   virtual bool           pinned() const { return true; }
 397   virtual const Type    *Value( PhaseTransform *phase ) const;
 398   virtual const Type    *bottom_type() const { return Type::CONTROL; }
 399   virtual const TypePtr *adr_type() const { return _adr_type; }
 400   virtual Node          *Ideal(PhaseGVN *phase, bool can_reshape);
 401   virtual Node          *Identity( PhaseTransform *phase );
 402   virtual uint           ideal_reg() const { return 0; }
 403   virtual const RegMask &in_RegMask(uint) const;
 404   virtual const RegMask &out_RegMask() const;
 405   virtual uint           match_edge(uint idx) const;
 406 
 407   static  bool           needs_polling_address_input();
 408 
 409 #ifndef PRODUCT
 410   virtual void              dump_spec(outputStream *st) const;
 411 #endif
 412 };
 413 
 414 //------------------------------SafePointScalarObjectNode----------------------
 415 // A SafePointScalarObjectNode represents the state of a scalarized object
 416 // at a safepoint.
 417 
 418 class SafePointScalarObjectNode: public TypeNode {
 419   uint _first_index; // First input edge index of a SafePoint node where
 420                      // states of the scalarized object fields are collected.
 421   uint _n_fields;    // Number of non-static fields of the scalarized object.
 422   DEBUG_ONLY(AllocateNode* _alloc;)
 423 public:
 424   SafePointScalarObjectNode(const TypeOopPtr* tp,
 425 #ifdef ASSERT
 426                             AllocateNode* alloc,
 427 #endif
 428                             uint first_index, uint n_fields);
 429   virtual int Opcode() const;
 430   virtual uint           ideal_reg() const;
 431   virtual const RegMask &in_RegMask(uint) const;
 432   virtual const RegMask &out_RegMask() const;
 433   virtual uint           match_edge(uint idx) const;
 434 
 435   uint first_index() const { return _first_index; }
 436   uint n_fields()    const { return _n_fields; }
 437   DEBUG_ONLY(AllocateNode* alloc() const { return _alloc; })
 438 
 439   // SafePointScalarObject should be always pinned to the control edge
 440   // of the SafePoint node for which it was generated.
 441   virtual bool pinned() const; // { return true; }
 442 
 443   // SafePointScalarObject depends on the SafePoint node
 444   // for which it was generated.
 445   virtual bool depends_only_on_test() const; // { return false; }
 446 
 447   virtual uint size_of() const { return sizeof(*this); }
 448 
 449   // Assumes that "this" is an argument to a safepoint node "s", and that
 450   // "new_call" is being created to correspond to "s".  But the difference
 451   // between the start index of the jvmstates of "new_call" and "s" is
 452   // "jvms_adj".  Produce and return a SafePointScalarObjectNode that
 453   // corresponds appropriately to "this" in "new_call".  Assumes that
 454   // "sosn_map" is a map, specific to the translation of "s" to "new_call",
 455   // mapping old SafePointScalarObjectNodes to new, to avoid multiple copies.
 456   SafePointScalarObjectNode* clone(int jvms_adj, Dict* sosn_map) const;
 457 
 458 #ifndef PRODUCT
 459   virtual void              dump_spec(outputStream *st) const;
 460 #endif
 461 };
 462 
 463 //------------------------------CallNode---------------------------------------
 464 // Call nodes now subsume the function of debug nodes at callsites, so they
 465 // contain the functionality of a full scope chain of debug nodes.
 466 class CallNode : public SafePointNode {
 467 public:
 468   const TypeFunc *_tf;        // Function type
 469   address      _entry_point;  // Address of method being called
 470   float        _cnt;          // Estimate of number of times called
 471 
 472   CallNode(const TypeFunc* tf, address addr, const TypePtr* adr_type)
 473     : SafePointNode(tf->domain()->cnt(), NULL, adr_type),
 474       _tf(tf),
 475       _entry_point(addr),
 476       _cnt(COUNT_UNKNOWN)
 477   {
 478     init_class_id(Class_Call);
 479     init_flags(Flag_is_Call);
 480   }
 481 
 482   const TypeFunc* tf()        const { return _tf; }
 483   const address entry_point() const { return _entry_point; }
 484   const float   cnt()         const { return _cnt; }
 485 
 486   void set_tf(const TypeFunc* tf) { _tf = tf; }
 487   void set_entry_point(address p) { _entry_point = p; }
 488   void set_cnt(float c)           { _cnt = c; }
 489 
 490   virtual const Type *bottom_type() const;
 491   virtual const Type *Value( PhaseTransform *phase ) const;
 492   virtual Node *Identity( PhaseTransform *phase ) { return this; }
 493   virtual uint        cmp( const Node &n ) const;
 494   virtual uint        size_of() const = 0;
 495   virtual void        calling_convention( BasicType* sig_bt, VMRegPair *parm_regs, uint argcnt ) const;
 496   virtual Node       *match( const ProjNode *proj, const Matcher *m );
 497   virtual uint        ideal_reg() const { return NotAMachineReg; }
 498   // Are we guaranteed that this node is a safepoint?  Not true for leaf calls and
 499   // for some macro nodes whose expansion does not have a safepoint on the fast path.
 500   virtual bool        guaranteed_safepoint()  { return true; }
 501   // For macro nodes, the JVMState gets modified during expansion, so when cloning
 502   // the node the JVMState must be cloned.
 503   virtual void        clone_jvms() { }   // default is not to clone
 504 
 505   // Returns true if the call may modify n
 506   virtual bool        may_modify(const TypePtr *addr_t, PhaseTransform *phase);
 507   // Does this node have a use of n other than in debug information?
 508   bool                has_non_debug_use(Node *n);
 509   // Returns the unique CheckCastPP of a call
 510   // or result projection is there are several CheckCastPP
 511   // or returns NULL if there is no one.
 512   Node *result_cast();
 513 
 514   virtual uint match_edge(uint idx) const;
 515 
 516 #ifndef PRODUCT
 517   virtual void        dump_req()  const;
 518   virtual void        dump_spec(outputStream *st) const;
 519 #endif
 520 };
 521 
 522 //------------------------------CallJavaNode-----------------------------------
 523 // Make a static or dynamic subroutine call node using Java calling
 524 // convention.  (The "Java" calling convention is the compiler's calling
 525 // convention, as opposed to the interpreter's or that of native C.)
 526 class CallJavaNode : public CallNode {
 527 protected:
 528   virtual uint cmp( const Node &n ) const;
 529   virtual uint size_of() const; // Size is bigger
 530 
 531   bool    _optimized_virtual;
 532   ciMethod* _method;            // Method being direct called
 533 public:
 534   const int       _bci;         // Byte Code Index of call byte code
 535   CallJavaNode(const TypeFunc* tf , address addr, ciMethod* method, int bci)
 536     : CallNode(tf, addr, TypePtr::BOTTOM),
 537       _method(method), _bci(bci), _optimized_virtual(false)
 538   {
 539     init_class_id(Class_CallJava);
 540   }
 541 
 542   virtual int   Opcode() const;
 543   ciMethod* method() const                { return _method; }
 544   void  set_method(ciMethod *m)           { _method = m; }
 545   void  set_optimized_virtual(bool f)     { _optimized_virtual = f; }
 546   bool  is_optimized_virtual() const      { return _optimized_virtual; }
 547 
 548 #ifndef PRODUCT
 549   virtual void  dump_spec(outputStream *st) const;
 550 #endif
 551 };
 552 
 553 //------------------------------CallStaticJavaNode-----------------------------
 554 // Make a direct subroutine call using Java calling convention (for static
 555 // calls and optimized virtual calls, plus calls to wrappers for run-time
 556 // routines); generates static stub.
 557 class CallStaticJavaNode : public CallJavaNode {
 558   virtual uint cmp( const Node &n ) const;
 559   virtual uint size_of() const; // Size is bigger
 560 public:
 561   CallStaticJavaNode(const TypeFunc* tf, address addr, ciMethod* method, int bci)
 562     : CallJavaNode(tf, addr, method, bci), _name(NULL) {
 563     init_class_id(Class_CallStaticJava);
 564   }
 565   CallStaticJavaNode(const TypeFunc* tf, address addr, const char* name, int bci,
 566                      const TypePtr* adr_type)
 567     : CallJavaNode(tf, addr, NULL, bci), _name(name) {
 568     init_class_id(Class_CallStaticJava);
 569     // This node calls a runtime stub, which often has narrow memory effects.
 570     _adr_type = adr_type;
 571   }
 572   const char *_name;            // Runtime wrapper name
 573 
 574   // If this is an uncommon trap, return the request code, else zero.
 575   int uncommon_trap_request() const;
 576   static int extract_uncommon_trap_request(const Node* call);
 577 
 578   virtual int         Opcode() const;
 579 #ifndef PRODUCT
 580   virtual void        dump_spec(outputStream *st) const;
 581 #endif
 582 };
 583 
 584 //------------------------------CallDynamicJavaNode----------------------------
 585 // Make a dispatched call using Java calling convention.
 586 class CallDynamicJavaNode : public CallJavaNode {
 587   virtual uint cmp( const Node &n ) const;
 588   virtual uint size_of() const; // Size is bigger
 589 public:
 590   CallDynamicJavaNode( const TypeFunc *tf , address addr, ciMethod* method, int vtable_index, int bci ) : CallJavaNode(tf,addr,method,bci), _vtable_index(vtable_index) {
 591     init_class_id(Class_CallDynamicJava);
 592   }
 593 
 594   int _vtable_index;
 595   virtual int   Opcode() const;
 596 #ifndef PRODUCT
 597   virtual void  dump_spec(outputStream *st) const;
 598 #endif
 599 };
 600 
 601 //------------------------------CallRuntimeNode--------------------------------
 602 // Make a direct subroutine call node into compiled C++ code.
 603 class CallRuntimeNode : public CallNode {
 604   virtual uint cmp( const Node &n ) const;
 605   virtual uint size_of() const; // Size is bigger
 606 public:
 607   CallRuntimeNode(const TypeFunc* tf, address addr, const char* name,
 608                   const TypePtr* adr_type)
 609     : CallNode(tf, addr, adr_type),
 610       _name(name)
 611   {
 612     init_class_id(Class_CallRuntime);
 613   }
 614 
 615   const char *_name;            // Printable name, if _method is NULL
 616   virtual int   Opcode() const;
 617   virtual void  calling_convention( BasicType* sig_bt, VMRegPair *parm_regs, uint argcnt ) const;
 618 
 619 #ifndef PRODUCT
 620   virtual void  dump_spec(outputStream *st) const;
 621 #endif
 622 };
 623 
 624 //------------------------------CallLeafNode-----------------------------------
 625 // Make a direct subroutine call node into compiled C++ code, without
 626 // safepoints
 627 class CallLeafNode : public CallRuntimeNode {
 628 public:
 629   CallLeafNode(const TypeFunc* tf, address addr, const char* name,
 630                const TypePtr* adr_type)
 631     : CallRuntimeNode(tf, addr, name, adr_type)
 632   {
 633     init_class_id(Class_CallLeaf);
 634   }
 635   virtual int   Opcode() const;
 636   virtual bool        guaranteed_safepoint()  { return false; }
 637 #ifndef PRODUCT
 638   virtual void  dump_spec(outputStream *st) const;
 639 #endif
 640 };
 641 
 642 //------------------------------CallLeafNoFPNode-------------------------------
 643 // CallLeafNode, not using floating point or using it in the same manner as
 644 // the generated code
 645 class CallLeafNoFPNode : public CallLeafNode {
 646 public:
 647   CallLeafNoFPNode(const TypeFunc* tf, address addr, const char* name,
 648                    const TypePtr* adr_type)
 649     : CallLeafNode(tf, addr, name, adr_type)
 650   {
 651   }
 652   virtual int   Opcode() const;
 653 };
 654 
 655 
 656 //------------------------------Allocate---------------------------------------
 657 // High-level memory allocation
 658 //
 659 //  AllocateNode and AllocateArrayNode are subclasses of CallNode because they will
 660 //  get expanded into a code sequence containing a call.  Unlike other CallNodes,
 661 //  they have 2 memory projections and 2 i_o projections (which are distinguished by
 662 //  the _is_io_use flag in the projection.)  This is needed when expanding the node in
 663 //  order to differentiate the uses of the projection on the normal control path from
 664 //  those on the exception return path.
 665 //
 666 class AllocateNode : public CallNode {
 667 public:
 668   enum {
 669     // Output:
 670     RawAddress  = TypeFunc::Parms,    // the newly-allocated raw address
 671     // Inputs:
 672     AllocSize   = TypeFunc::Parms,    // size (in bytes) of the new object
 673     KlassNode,                        // type (maybe dynamic) of the obj.
 674     InitialTest,                      // slow-path test (may be constant)
 675     ALength,                          // array length (or TOP if none)
 676     ParmLimit
 677   };
 678 
 679   static const TypeFunc* alloc_type() {
 680     const Type** fields = TypeTuple::fields(ParmLimit - TypeFunc::Parms);
 681     fields[AllocSize]   = TypeInt::POS;
 682     fields[KlassNode]   = TypeInstPtr::NOTNULL;
 683     fields[InitialTest] = TypeInt::BOOL;
 684     fields[ALength]     = TypeInt::INT;  // length (can be a bad length)
 685 
 686     const TypeTuple *domain = TypeTuple::make(ParmLimit, fields);
 687 
 688     // create result type (range)
 689     fields = TypeTuple::fields(1);
 690     fields[TypeFunc::Parms+0] = TypeRawPtr::NOTNULL; // Returned oop
 691 
 692     const TypeTuple *range = TypeTuple::make(TypeFunc::Parms+1, fields);
 693 
 694     return TypeFunc::make(domain, range);
 695   }
 696 
 697   bool _is_scalar_replaceable;  // Result of Escape Analysis
 698 
 699   virtual uint size_of() const; // Size is bigger
 700   AllocateNode(Compile* C, const TypeFunc *atype, Node *ctrl, Node *mem, Node *abio,
 701                Node *size, Node *klass_node, Node *initial_test);
 702   // Expansion modifies the JVMState, so we need to clone it
 703   virtual void  clone_jvms() {
 704     set_jvms(jvms()->clone_deep(Compile::current()));
 705   }
 706   virtual int Opcode() const;
 707   virtual uint ideal_reg() const { return Op_RegP; }
 708   virtual bool        guaranteed_safepoint()  { return false; }
 709 
 710   // allocations do not modify their arguments
 711   virtual bool        may_modify(const TypePtr *addr_t, PhaseTransform *phase) { return false;}
 712 
 713   // Pattern-match a possible usage of AllocateNode.
 714   // Return null if no allocation is recognized.
 715   // The operand is the pointer produced by the (possible) allocation.
 716   // It must be a projection of the Allocate or its subsequent CastPP.
 717   // (Note:  This function is defined in file graphKit.cpp, near
 718   // GraphKit::new_instance/new_array, whose output it recognizes.)
 719   // The 'ptr' may not have an offset unless the 'offset' argument is given.
 720   static AllocateNode* Ideal_allocation(Node* ptr, PhaseTransform* phase);
 721 
 722   // Fancy version which uses AddPNode::Ideal_base_and_offset to strip
 723   // an offset, which is reported back to the caller.
 724   // (Note:  AllocateNode::Ideal_allocation is defined in graphKit.cpp.)
 725   static AllocateNode* Ideal_allocation(Node* ptr, PhaseTransform* phase,
 726                                         intptr_t& offset);
 727 
 728   // Dig the klass operand out of a (possible) allocation site.
 729   static Node* Ideal_klass(Node* ptr, PhaseTransform* phase) {
 730     AllocateNode* allo = Ideal_allocation(ptr, phase);
 731     return (allo == NULL) ? NULL : allo->in(KlassNode);
 732   }
 733 
 734   // Conservatively small estimate of offset of first non-header byte.
 735   int minimum_header_size() {
 736     return is_AllocateArray() ? arrayOopDesc::base_offset_in_bytes(T_BYTE) :
 737                                 instanceOopDesc::base_offset_in_bytes();
 738   }
 739 
 740   // Return the corresponding initialization barrier (or null if none).
 741   // Walks out edges to find it...
 742   // (Note: Both InitializeNode::allocation and AllocateNode::initialization
 743   // are defined in graphKit.cpp, which sets up the bidirectional relation.)
 744   InitializeNode* initialization();
 745 
 746   // Convenience for initialization->maybe_set_complete(phase)
 747   bool maybe_set_complete(PhaseGVN* phase);
 748 };
 749 
 750 //------------------------------AllocateArray---------------------------------
 751 //
 752 // High-level array allocation
 753 //
 754 class AllocateArrayNode : public AllocateNode {
 755 public:
 756   AllocateArrayNode(Compile* C, const TypeFunc *atype, Node *ctrl, Node *mem, Node *abio,
 757                     Node* size, Node* klass_node, Node* initial_test,
 758                     Node* count_val
 759                     )
 760     : AllocateNode(C, atype, ctrl, mem, abio, size, klass_node,
 761                    initial_test)
 762   {
 763     init_class_id(Class_AllocateArray);
 764     set_req(AllocateNode::ALength,        count_val);
 765   }
 766   virtual int Opcode() const;
 767   virtual uint size_of() const; // Size is bigger
 768   virtual Node *Ideal(PhaseGVN *phase, bool can_reshape);
 769 
 770   // Dig the length operand out of a array allocation site.
 771   Node* Ideal_length() {
 772     return in(AllocateNode::ALength);
 773   }
 774 
 775   // Dig the length operand out of a array allocation site and narrow the
 776   // type with a CastII, if necesssary
 777   Node* make_ideal_length(const TypeOopPtr* ary_type, PhaseTransform *phase, bool can_create = true);
 778 
 779   // Pattern-match a possible usage of AllocateArrayNode.
 780   // Return null if no allocation is recognized.
 781   static AllocateArrayNode* Ideal_array_allocation(Node* ptr, PhaseTransform* phase) {
 782     AllocateNode* allo = Ideal_allocation(ptr, phase);
 783     return (allo == NULL || !allo->is_AllocateArray())
 784            ? NULL : allo->as_AllocateArray();
 785   }
 786 };
 787 
 788 //------------------------------AbstractLockNode-----------------------------------
 789 class AbstractLockNode: public CallNode {
 790 private:
 791   bool _eliminate;    // indicates this lock can be safely eliminated
 792   bool _coarsened;    // indicates this lock was coarsened
 793 #ifndef PRODUCT
 794   NamedCounter* _counter;
 795 #endif
 796 
 797 protected:
 798   // helper functions for lock elimination
 799   //
 800 
 801   bool find_matching_unlock(const Node* ctrl, LockNode* lock,
 802                             GrowableArray<AbstractLockNode*> &lock_ops);
 803   bool find_lock_and_unlock_through_if(Node* node, LockNode* lock,
 804                                        GrowableArray<AbstractLockNode*> &lock_ops);
 805   bool find_unlocks_for_region(const RegionNode* region, LockNode* lock,
 806                                GrowableArray<AbstractLockNode*> &lock_ops);
 807   LockNode *find_matching_lock(UnlockNode* unlock);
 808 
 809 
 810 public:
 811   AbstractLockNode(const TypeFunc *tf)
 812     : CallNode(tf, NULL, TypeRawPtr::BOTTOM),
 813       _coarsened(false),
 814       _eliminate(false)
 815   {
 816 #ifndef PRODUCT
 817     _counter = NULL;
 818 #endif
 819   }
 820   virtual int Opcode() const = 0;
 821   Node *   obj_node() const       {return in(TypeFunc::Parms + 0); }
 822   Node *   box_node() const       {return in(TypeFunc::Parms + 1); }
 823   Node *   fastlock_node() const  {return in(TypeFunc::Parms + 2); }
 824   const Type *sub(const Type *t1, const Type *t2) const { return TypeInt::CC;}
 825 
 826   virtual uint size_of() const { return sizeof(*this); }
 827 
 828   bool is_eliminated()         {return _eliminate; }
 829   // mark node as eliminated and update the counter if there is one
 830   void set_eliminated();
 831 
 832   bool is_coarsened()  { return _coarsened; }
 833   void set_coarsened() { _coarsened = true; }
 834 
 835   // locking does not modify its arguments
 836   virtual bool        may_modify(const TypePtr *addr_t, PhaseTransform *phase){ return false;}
 837 
 838 #ifndef PRODUCT
 839   void create_lock_counter(JVMState* s);
 840   NamedCounter* counter() const { return _counter; }
 841 #endif
 842 };
 843 
 844 //------------------------------Lock---------------------------------------
 845 // High-level lock operation
 846 //
 847 // This is a subclass of CallNode because it is a macro node which gets expanded
 848 // into a code sequence containing a call.  This node takes 3 "parameters":
 849 //    0  -  object to lock
 850 //    1 -   a BoxLockNode
 851 //    2 -   a FastLockNode
 852 //
 853 class LockNode : public AbstractLockNode {
 854 public:
 855 
 856   static const TypeFunc *lock_type() {
 857     // create input type (domain)
 858     const Type **fields = TypeTuple::fields(3);
 859     fields[TypeFunc::Parms+0] = TypeInstPtr::NOTNULL;  // Object to be Locked
 860     fields[TypeFunc::Parms+1] = TypeRawPtr::BOTTOM;    // Address of stack location for lock
 861     fields[TypeFunc::Parms+2] = TypeInt::BOOL;         // FastLock
 862     const TypeTuple *domain = TypeTuple::make(TypeFunc::Parms+3,fields);
 863 
 864     // create result type (range)
 865     fields = TypeTuple::fields(0);
 866 
 867     const TypeTuple *range = TypeTuple::make(TypeFunc::Parms+0,fields);
 868 
 869     return TypeFunc::make(domain,range);
 870   }
 871 
 872   virtual int Opcode() const;
 873   virtual uint size_of() const; // Size is bigger
 874   LockNode(Compile* C, const TypeFunc *tf) : AbstractLockNode( tf ) {
 875     init_class_id(Class_Lock);
 876     init_flags(Flag_is_macro);
 877     C->add_macro_node(this);
 878   }
 879   virtual bool        guaranteed_safepoint()  { return false; }
 880 
 881   virtual Node *Ideal(PhaseGVN *phase, bool can_reshape);
 882   // Expansion modifies the JVMState, so we need to clone it
 883   virtual void  clone_jvms() {
 884     set_jvms(jvms()->clone_deep(Compile::current()));
 885   }
 886 };
 887 
 888 //------------------------------Unlock---------------------------------------
 889 // High-level unlock operation
 890 class UnlockNode : public AbstractLockNode {
 891 public:
 892   virtual int Opcode() const;
 893   virtual uint size_of() const; // Size is bigger
 894   UnlockNode(Compile* C, const TypeFunc *tf) : AbstractLockNode( tf ) {
 895     init_class_id(Class_Unlock);
 896     init_flags(Flag_is_macro);
 897     C->add_macro_node(this);
 898   }
 899   virtual Node *Ideal(PhaseGVN *phase, bool can_reshape);
 900   // unlock is never a safepoint
 901   virtual bool        guaranteed_safepoint()  { return false; }
 902 };