1 /* 2 * Copyright (c) 2007, 2015, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 */ 23 24 #ifndef SHARE_VM_OPTO_SUPERWORD_HPP 25 #define SHARE_VM_OPTO_SUPERWORD_HPP 26 27 #include "opto/loopnode.hpp" 28 #include "opto/node.hpp" 29 #include "opto/phaseX.hpp" 30 #include "opto/vectornode.hpp" 31 #include "utilities/growableArray.hpp" 32 33 // 34 // S U P E R W O R D T R A N S F O R M 35 // 36 // SuperWords are short, fixed length vectors. 37 // 38 // Algorithm from: 39 // 40 // Exploiting SuperWord Level Parallelism with 41 // Multimedia Instruction Sets 42 // by 43 // Samuel Larsen and Saman Amarasinghe 44 // MIT Laboratory for Computer Science 45 // date 46 // May 2000 47 // published in 48 // ACM SIGPLAN Notices 49 // Proceedings of ACM PLDI '00, Volume 35 Issue 5 50 // 51 // Definition 3.1 A Pack is an n-tuple, <s1, ...,sn>, where 52 // s1,...,sn are independent isomorphic statements in a basic 53 // block. 54 // 55 // Definition 3.2 A PackSet is a set of Packs. 56 // 57 // Definition 3.3 A Pair is a Pack of size two, where the 58 // first statement is considered the left element, and the 59 // second statement is considered the right element. 60 61 class SWPointer; 62 class OrderedPair; 63 64 // ========================= Dependence Graph ===================== 65 66 class DepMem; 67 68 //------------------------------DepEdge--------------------------- 69 // An edge in the dependence graph. The edges incident to a dependence 70 // node are threaded through _next_in for incoming edges and _next_out 71 // for outgoing edges. 72 class DepEdge : public ResourceObj { 73 protected: 74 DepMem* _pred; 75 DepMem* _succ; 76 DepEdge* _next_in; // list of in edges, null terminated 77 DepEdge* _next_out; // list of out edges, null terminated 78 79 public: 80 DepEdge(DepMem* pred, DepMem* succ, DepEdge* next_in, DepEdge* next_out) : 81 _pred(pred), _succ(succ), _next_in(next_in), _next_out(next_out) {} 82 83 DepEdge* next_in() { return _next_in; } 84 DepEdge* next_out() { return _next_out; } 85 DepMem* pred() { return _pred; } 86 DepMem* succ() { return _succ; } 87 88 void print(); 89 }; 90 91 //------------------------------DepMem--------------------------- 92 // A node in the dependence graph. _in_head starts the threaded list of 93 // incoming edges, and _out_head starts the list of outgoing edges. 94 class DepMem : public ResourceObj { 95 protected: 96 Node* _node; // Corresponding ideal node 97 DepEdge* _in_head; // Head of list of in edges, null terminated 98 DepEdge* _out_head; // Head of list of out edges, null terminated 99 100 public: 101 DepMem(Node* node) : _node(node), _in_head(NULL), _out_head(NULL) {} 102 103 Node* node() { return _node; } 104 DepEdge* in_head() { return _in_head; } 105 DepEdge* out_head() { return _out_head; } 106 void set_in_head(DepEdge* hd) { _in_head = hd; } 107 void set_out_head(DepEdge* hd) { _out_head = hd; } 108 109 int in_cnt(); // Incoming edge count 110 int out_cnt(); // Outgoing edge count 111 112 void print(); 113 }; 114 115 //------------------------------DepGraph--------------------------- 116 class DepGraph VALUE_OBJ_CLASS_SPEC { 117 protected: 118 Arena* _arena; 119 GrowableArray<DepMem*> _map; 120 DepMem* _root; 121 DepMem* _tail; 122 123 public: 124 DepGraph(Arena* a) : _arena(a), _map(a, 8, 0, NULL) { 125 _root = new (_arena) DepMem(NULL); 126 _tail = new (_arena) DepMem(NULL); 127 } 128 129 DepMem* root() { return _root; } 130 DepMem* tail() { return _tail; } 131 132 // Return dependence node corresponding to an ideal node 133 DepMem* dep(Node* node) { return _map.at(node->_idx); } 134 135 // Make a new dependence graph node for an ideal node. 136 DepMem* make_node(Node* node); 137 138 // Make a new dependence graph edge dprec->dsucc 139 DepEdge* make_edge(DepMem* dpred, DepMem* dsucc); 140 141 DepEdge* make_edge(Node* pred, Node* succ) { return make_edge(dep(pred), dep(succ)); } 142 DepEdge* make_edge(DepMem* pred, Node* succ) { return make_edge(pred, dep(succ)); } 143 DepEdge* make_edge(Node* pred, DepMem* succ) { return make_edge(dep(pred), succ); } 144 145 void init() { _map.clear(); } // initialize 146 147 void print(Node* n) { dep(n)->print(); } 148 void print(DepMem* d) { d->print(); } 149 }; 150 151 //------------------------------DepPreds--------------------------- 152 // Iterator over predecessors in the dependence graph and 153 // non-memory-graph inputs of ideal nodes. 154 class DepPreds : public StackObj { 155 private: 156 Node* _n; 157 int _next_idx, _end_idx; 158 DepEdge* _dep_next; 159 Node* _current; 160 bool _done; 161 162 public: 163 DepPreds(Node* n, DepGraph& dg); 164 Node* current() { return _current; } 165 bool done() { return _done; } 166 void next(); 167 }; 168 169 //------------------------------DepSuccs--------------------------- 170 // Iterator over successors in the dependence graph and 171 // non-memory-graph outputs of ideal nodes. 172 class DepSuccs : public StackObj { 173 private: 174 Node* _n; 175 int _next_idx, _end_idx; 176 DepEdge* _dep_next; 177 Node* _current; 178 bool _done; 179 180 public: 181 DepSuccs(Node* n, DepGraph& dg); 182 Node* current() { return _current; } 183 bool done() { return _done; } 184 void next(); 185 }; 186 187 188 // ========================= SuperWord ===================== 189 190 // -----------------------------SWNodeInfo--------------------------------- 191 // Per node info needed by SuperWord 192 class SWNodeInfo VALUE_OBJ_CLASS_SPEC { 193 public: 194 int _alignment; // memory alignment for a node 195 int _depth; // Max expression (DAG) depth from block start 196 const Type* _velt_type; // vector element type 197 Node_List* _my_pack; // pack containing this node 198 199 SWNodeInfo() : _alignment(-1), _depth(0), _velt_type(NULL), _my_pack(NULL) {} 200 static const SWNodeInfo initial; 201 }; 202 203 // JVMCI: OrderedPair is moved up to deal with compilation issues on Windows 204 //------------------------------OrderedPair--------------------------- 205 // Ordered pair of Node*. 206 class OrderedPair VALUE_OBJ_CLASS_SPEC { 207 protected: 208 Node* _p1; 209 Node* _p2; 210 public: 211 OrderedPair() : _p1(NULL), _p2(NULL) {} 212 OrderedPair(Node* p1, Node* p2) { 213 if (p1->_idx < p2->_idx) { 214 _p1 = p1; _p2 = p2; 215 } else { 216 _p1 = p2; _p2 = p1; 217 } 218 } 219 220 bool operator==(const OrderedPair &rhs) { 221 return _p1 == rhs._p1 && _p2 == rhs._p2; 222 } 223 void print() { tty->print(" (%d, %d)", _p1->_idx, _p2->_idx); } 224 225 static const OrderedPair initial; 226 }; 227 228 // -----------------------------SuperWord--------------------------------- 229 // Transforms scalar operations into packed (superword) operations. 230 class SuperWord : public ResourceObj { 231 friend class SWPointer; 232 private: 233 PhaseIdealLoop* _phase; 234 Arena* _arena; 235 PhaseIterGVN &_igvn; 236 237 enum consts { top_align = -1, bottom_align = -666 }; 238 239 GrowableArray<Node_List*> _packset; // Packs for the current block 240 241 GrowableArray<int> _bb_idx; // Map from Node _idx to index within block 242 243 GrowableArray<Node*> _block; // Nodes in current block 244 GrowableArray<Node*> _data_entry; // Nodes with all inputs from outside 245 GrowableArray<Node*> _mem_slice_head; // Memory slice head nodes 246 GrowableArray<Node*> _mem_slice_tail; // Memory slice tail nodes 247 GrowableArray<Node*> _iteration_first; // nodes in the generation that has deps from phi 248 GrowableArray<Node*> _iteration_last; // nodes in the generation that has deps to phi 249 GrowableArray<SWNodeInfo> _node_info; // Info needed per node 250 CloneMap& _clone_map; // map of nodes created in cloning 251 252 MemNode* _align_to_ref; // Memory reference that pre-loop will align to 253 254 GrowableArray<OrderedPair> _disjoint_ptrs; // runtime disambiguated pointer pairs 255 256 DepGraph _dg; // Dependence graph 257 258 // Scratch pads 259 VectorSet _visited; // Visited set 260 VectorSet _post_visited; // Post-visited set 261 Node_Stack _n_idx_list; // List of (node,index) pairs 262 GrowableArray<Node*> _nlist; // List of nodes 263 GrowableArray<Node*> _stk; // Stack of nodes 264 265 public: 266 SuperWord(PhaseIdealLoop* phase); 267 268 void transform_loop(IdealLoopTree* lpt, bool do_optimization); 269 270 void unrolling_analysis(int &local_loop_unroll_factor); 271 272 // Accessors for SWPointer 273 PhaseIdealLoop* phase() { return _phase; } 274 IdealLoopTree* lpt() { return _lpt; } 275 PhiNode* iv() { return _iv; } 276 277 bool early_return() { return _early_return; } 278 279 #ifndef PRODUCT 280 bool is_debug() { return _vector_loop_debug > 0; } 281 bool is_trace_alignment() { return (_vector_loop_debug & 2) > 0; } 282 bool is_trace_mem_slice() { return (_vector_loop_debug & 4) > 0; } 283 bool is_trace_loop() { return (_vector_loop_debug & 8) > 0; } 284 bool is_trace_adjacent() { return (_vector_loop_debug & 16) > 0; } 285 #endif 286 bool do_vector_loop() { return _do_vector_loop; } 287 private: 288 IdealLoopTree* _lpt; // Current loop tree node 289 LoopNode* _lp; // Current LoopNode 290 Node* _bb; // Current basic block 291 PhiNode* _iv; // Induction var 292 bool _race_possible; // In cases where SDMU is true 293 bool _early_return; // True if we do not initialize 294 bool _do_vector_loop; // whether to do vectorization/simd style 295 int _num_work_vecs; // Number of non memory vector operations 296 int _num_reductions; // Number of reduction expressions applied 297 int _ii_first; // generation with direct deps from mem phi 298 int _ii_last; // generation with direct deps to mem phi 299 GrowableArray<int> _ii_order; 300 #ifndef PRODUCT 301 uintx _vector_loop_debug; // provide more printing in debug mode 302 uintx _CountedLoopReserveKit_debug; // for debugging CountedLoopReserveKit 303 #endif 304 305 // Accessors 306 Arena* arena() { return _arena; } 307 308 Node* bb() { return _bb; } 309 void set_bb(Node* bb) { _bb = bb; } 310 311 void set_lpt(IdealLoopTree* lpt) { _lpt = lpt; } 312 313 LoopNode* lp() { return _lp; } 314 void set_lp(LoopNode* lp) { _lp = lp; 315 _iv = lp->as_CountedLoop()->phi()->as_Phi(); } 316 int iv_stride() { return lp()->as_CountedLoop()->stride_con(); } 317 318 int vector_width(Node* n) { 319 BasicType bt = velt_basic_type(n); 320 return MIN2(ABS(iv_stride()), Matcher::max_vector_size(bt)); 321 } 322 int vector_width_in_bytes(Node* n) { 323 BasicType bt = velt_basic_type(n); 324 return vector_width(n)*type2aelembytes(bt); 325 } 326 MemNode* align_to_ref() { return _align_to_ref; } 327 void set_align_to_ref(MemNode* m) { _align_to_ref = m; } 328 329 Node* ctrl(Node* n) const { return _phase->has_ctrl(n) ? _phase->get_ctrl(n) : n; } 330 331 // block accessors 332 bool in_bb(Node* n) { return n != NULL && n->outcnt() > 0 && ctrl(n) == _bb; } 333 int bb_idx(Node* n) { assert(in_bb(n), "must be"); return _bb_idx.at(n->_idx); } 334 void set_bb_idx(Node* n, int i) { _bb_idx.at_put_grow(n->_idx, i); } 335 336 // visited set accessors 337 void visited_clear() { _visited.Clear(); } 338 void visited_set(Node* n) { return _visited.set(bb_idx(n)); } 339 int visited_test(Node* n) { return _visited.test(bb_idx(n)); } 340 int visited_test_set(Node* n) { return _visited.test_set(bb_idx(n)); } 341 void post_visited_clear() { _post_visited.Clear(); } 342 void post_visited_set(Node* n) { return _post_visited.set(bb_idx(n)); } 343 int post_visited_test(Node* n) { return _post_visited.test(bb_idx(n)); } 344 345 // Ensure node_info contains element "i" 346 void grow_node_info(int i) { if (i >= _node_info.length()) _node_info.at_put_grow(i, SWNodeInfo::initial); } 347 348 // memory alignment for a node 349 int alignment(Node* n) { return _node_info.adr_at(bb_idx(n))->_alignment; } 350 void set_alignment(Node* n, int a) { int i = bb_idx(n); grow_node_info(i); _node_info.adr_at(i)->_alignment = a; } 351 352 // Max expression (DAG) depth from beginning of the block for each node 353 int depth(Node* n) { return _node_info.adr_at(bb_idx(n))->_depth; } 354 void set_depth(Node* n, int d) { int i = bb_idx(n); grow_node_info(i); _node_info.adr_at(i)->_depth = d; } 355 356 // vector element type 357 const Type* velt_type(Node* n) { return _node_info.adr_at(bb_idx(n))->_velt_type; } 358 BasicType velt_basic_type(Node* n) { return velt_type(n)->array_element_basic_type(); } 359 void set_velt_type(Node* n, const Type* t) { int i = bb_idx(n); grow_node_info(i); _node_info.adr_at(i)->_velt_type = t; } 360 bool same_velt_type(Node* n1, Node* n2); 361 362 // my_pack 363 Node_List* my_pack(Node* n) { return !in_bb(n) ? NULL : _node_info.adr_at(bb_idx(n))->_my_pack; } 364 void set_my_pack(Node* n, Node_List* p) { int i = bb_idx(n); grow_node_info(i); _node_info.adr_at(i)->_my_pack = p; } 365 366 // CloneMap utilities 367 bool same_origin_idx(Node* a, Node* b) const; 368 bool same_generation(Node* a, Node* b) const; 369 370 // methods 371 372 // Extract the superword level parallelism 373 void SLP_extract(); 374 // Find the adjacent memory references and create pack pairs for them. 375 void find_adjacent_refs(); 376 // Tracing support 377 #ifndef PRODUCT 378 void find_adjacent_refs_trace_1(Node* best_align_to_mem_ref, int best_iv_adjustment); 379 void print_loop(bool whole); 380 #endif 381 // Find a memory reference to align the loop induction variable to. 382 MemNode* find_align_to_ref(Node_List &memops); 383 // Calculate loop's iv adjustment for this memory ops. 384 int get_iv_adjustment(MemNode* mem); 385 // Can the preloop align the reference to position zero in the vector? 386 bool ref_is_alignable(SWPointer& p); 387 // rebuild the graph so all loads in different iterations of cloned loop become dependant on phi node (in _do_vector_loop only) 388 bool hoist_loads_in_graph(); 389 // Test whether MemNode::Memory dependency to the same load but in the first iteration of this loop is coming from memory phi 390 // Return false if failed 391 Node* find_phi_for_mem_dep(LoadNode* ld); 392 // Return same node but from the first generation. Return 0, if not found 393 Node* first_node(Node* nd); 394 // Return same node as this but from the last generation. Return 0, if not found 395 Node* last_node(Node* n); 396 // Mark nodes belonging to first and last generation 397 // returns first generation index or -1 if vectorization/simd is impossible 398 int mark_generations(); 399 // swapping inputs of commutative instruction (Add or Mul) 400 bool fix_commutative_inputs(Node* gold, Node* fix); 401 // make packs forcefully (in _do_vector_loop only) 402 bool pack_parallel(); 403 // Construct dependency graph. 404 void dependence_graph(); 405 // Return a memory slice (node list) in predecessor order starting at "start" 406 void mem_slice_preds(Node* start, Node* stop, GrowableArray<Node*> &preds); 407 // Can s1 and s2 be in a pack with s1 immediately preceding s2 and s1 aligned at "align" 408 bool stmts_can_pack(Node* s1, Node* s2, int align); 409 // Does s exist in a pack at position pos? 410 bool exists_at(Node* s, uint pos); 411 // Is s1 immediately before s2 in memory? 412 bool are_adjacent_refs(Node* s1, Node* s2); 413 // Are s1 and s2 similar? 414 bool isomorphic(Node* s1, Node* s2); 415 // Is there no data path from s1 to s2 or s2 to s1? 416 bool independent(Node* s1, Node* s2); 417 // Is there a data path between s1 and s2 and both are reductions? 418 bool reduction(Node* s1, Node* s2); 419 // Helper for independent 420 bool independent_path(Node* shallow, Node* deep, uint dp=0); 421 void set_alignment(Node* s1, Node* s2, int align); 422 int data_size(Node* s); 423 // Extend packset by following use->def and def->use links from pack members. 424 void extend_packlist(); 425 // Extend the packset by visiting operand definitions of nodes in pack p 426 bool follow_use_defs(Node_List* p); 427 // Extend the packset by visiting uses of nodes in pack p 428 bool follow_def_uses(Node_List* p); 429 // For extended packsets, ordinally arrange uses packset by major component 430 void order_def_uses(Node_List* p); 431 // Estimate the savings from executing s1 and s2 as a pack 432 int est_savings(Node* s1, Node* s2); 433 int adjacent_profit(Node* s1, Node* s2); 434 int pack_cost(int ct); 435 int unpack_cost(int ct); 436 // Combine packs A and B with A.last == B.first into A.first..,A.last,B.second,..B.last 437 void combine_packs(); 438 // Construct the map from nodes to packs. 439 void construct_my_pack_map(); 440 // Remove packs that are not implemented or not profitable. 441 void filter_packs(); 442 // Adjust the memory graph for the packed operations 443 void schedule(); 444 // Remove "current" from its current position in the memory graph and insert 445 // it after the appropriate insert points (lip or uip); 446 void remove_and_insert(MemNode *current, MemNode *prev, MemNode *lip, Node *uip, Unique_Node_List &schd_before); 447 // Within a store pack, schedule stores together by moving out the sandwiched memory ops according 448 // to dependence info; and within a load pack, move loads down to the last executed load. 449 void co_locate_pack(Node_List* p); 450 // Convert packs into vector node operations 451 void output(); 452 // Create a vector operand for the nodes in pack p for operand: in(opd_idx) 453 Node* vector_opd(Node_List* p, int opd_idx); 454 // Can code be generated for pack p? 455 bool implemented(Node_List* p); 456 // For pack p, are all operands and all uses (with in the block) vector? 457 bool profitable(Node_List* p); 458 // If a use of pack p is not a vector use, then replace the use with an extract operation. 459 void insert_extracts(Node_List* p); 460 // Is use->in(u_idx) a vector use? 461 bool is_vector_use(Node* use, int u_idx); 462 // Construct reverse postorder list of block members 463 bool construct_bb(); 464 // Initialize per node info 465 void initialize_bb(); 466 // Insert n into block after pos 467 void bb_insert_after(Node* n, int pos); 468 // Compute max depth for expressions from beginning of block 469 void compute_max_depth(); 470 // Compute necessary vector element type for expressions 471 void compute_vector_element_type(); 472 // Are s1 and s2 in a pack pair and ordered as s1,s2? 473 bool in_packset(Node* s1, Node* s2); 474 // Is s in pack p? 475 Node_List* in_pack(Node* s, Node_List* p); 476 // Remove the pack at position pos in the packset 477 void remove_pack_at(int pos); 478 // Return the node executed first in pack p. 479 Node* executed_first(Node_List* p); 480 // Return the node executed last in pack p. 481 Node* executed_last(Node_List* p); 482 static LoadNode::ControlDependency control_dependency(Node_List* p); 483 // Alignment within a vector memory reference 484 int memory_alignment(MemNode* s, int iv_adjust); 485 // (Start, end] half-open range defining which operands are vector 486 void vector_opd_range(Node* n, uint* start, uint* end); 487 // Smallest type containing range of values 488 const Type* container_type(Node* n); 489 // Adjust pre-loop limit so that in main loop, a load/store reference 490 // to align_to_ref will be a position zero in the vector. 491 void align_initial_loop_index(MemNode* align_to_ref); 492 // Find pre loop end from main loop. Returns null if none. 493 CountedLoopEndNode* get_pre_loop_end(CountedLoopNode *cl); 494 // Is the use of d1 in u1 at the same operand position as d2 in u2? 495 bool opnd_positions_match(Node* d1, Node* u1, Node* d2, Node* u2); 496 void init(); 497 // clean up some basic structures - used if the ideal graph was rebuilt 498 void restart(); 499 500 // print methods 501 void print_packset(); 502 void print_pack(Node_List* p); 503 void print_bb(); 504 void print_stmt(Node* s); 505 char* blank(uint depth); 506 507 void packset_sort(int n); 508 }; 509 510 511 512 //------------------------------SWPointer--------------------------- 513 // Information about an address for dependence checking and vector alignment 514 class SWPointer VALUE_OBJ_CLASS_SPEC { 515 protected: 516 MemNode* _mem; // My memory reference node 517 SuperWord* _slp; // SuperWord class 518 519 Node* _base; // NULL if unsafe nonheap reference 520 Node* _adr; // address pointer 521 jint _scale; // multiplier for iv (in bytes), 0 if no loop iv 522 jint _offset; // constant offset (in bytes) 523 Node* _invar; // invariant offset (in bytes), NULL if none 524 bool _negate_invar; // if true then use: (0 - _invar) 525 Node_Stack* _nstack; // stack used to record a swpointer trace of variants 526 bool _analyze_only; // Used in loop unrolling only for swpointer trace 527 uint _stack_idx; // Used in loop unrolling only for swpointer trace 528 529 PhaseIdealLoop* phase() { return _slp->phase(); } 530 IdealLoopTree* lpt() { return _slp->lpt(); } 531 PhiNode* iv() { return _slp->iv(); } // Induction var 532 533 bool invariant(Node* n); 534 535 // Match: k*iv + offset 536 bool scaled_iv_plus_offset(Node* n); 537 // Match: k*iv where k is a constant that's not zero 538 bool scaled_iv(Node* n); 539 // Match: offset is (k [+/- invariant]) 540 bool offset_plus_k(Node* n, bool negate = false); 541 542 public: 543 enum CMP { 544 Less = 1, 545 Greater = 2, 546 Equal = 4, 547 NotEqual = (Less | Greater), 548 NotComparable = (Less | Greater | Equal) 549 }; 550 551 SWPointer(MemNode* mem, SuperWord* slp, Node_Stack *nstack, bool analyze_only); 552 // Following is used to create a temporary object during 553 // the pattern match of an address expression. 554 SWPointer(SWPointer* p); 555 556 bool valid() { return _adr != NULL; } 557 bool has_iv() { return _scale != 0; } 558 559 Node* base() { return _base; } 560 Node* adr() { return _adr; } 561 MemNode* mem() { return _mem; } 562 int scale_in_bytes() { return _scale; } 563 Node* invar() { return _invar; } 564 bool negate_invar() { return _negate_invar; } 565 int offset_in_bytes() { return _offset; } 566 int memory_size() { return _mem->memory_size(); } 567 Node_Stack* node_stack() { return _nstack; } 568 569 // Comparable? 570 int cmp(SWPointer& q) { 571 if (valid() && q.valid() && 572 (_adr == q._adr || _base == _adr && q._base == q._adr) && 573 _scale == q._scale && 574 _invar == q._invar && 575 _negate_invar == q._negate_invar) { 576 bool overlap = q._offset < _offset + memory_size() && 577 _offset < q._offset + q.memory_size(); 578 return overlap ? Equal : (_offset < q._offset ? Less : Greater); 579 } else { 580 return NotComparable; 581 } 582 } 583 584 bool not_equal(SWPointer& q) { return not_equal(cmp(q)); } 585 bool equal(SWPointer& q) { return equal(cmp(q)); } 586 bool comparable(SWPointer& q) { return comparable(cmp(q)); } 587 static bool not_equal(int cmp) { return cmp <= NotEqual; } 588 static bool equal(int cmp) { return cmp == Equal; } 589 static bool comparable(int cmp) { return cmp < NotComparable; } 590 591 void print(); 592 593 #ifndef PRODUCT 594 class Tracer { 595 friend class SuperWord; 596 friend class SWPointer; 597 SuperWord* _slp; 598 static int _depth; 599 int _depth_save; 600 void print_depth(); 601 int depth() const { return _depth; } 602 void set_depth(int d) { _depth = d; } 603 void inc_depth() { _depth++;} 604 void dec_depth() { if (_depth > 0) _depth--;} 605 void store_depth() {_depth_save = _depth;} 606 void restore_depth() {_depth = _depth_save;} 607 608 class Depth { 609 friend class Tracer; 610 friend class SWPointer; 611 friend class SuperWord; 612 Depth() { ++_depth; } 613 Depth(int x) { _depth = 0; } 614 ~Depth() { if (_depth > 0) --_depth;} 615 }; 616 Tracer (SuperWord* slp) : _slp(slp) {} 617 618 // tracing functions 619 void ctor_1(Node* mem); 620 void ctor_2(Node* adr); 621 void ctor_3(Node* adr, int i); 622 void ctor_4(Node* adr, int i); 623 void ctor_5(Node* adr, Node* base, int i); 624 void ctor_6(Node* mem); 625 626 void invariant_1(Node *n, Node *n_c); 627 628 void scaled_iv_plus_offset_1(Node* n); 629 void scaled_iv_plus_offset_2(Node* n); 630 void scaled_iv_plus_offset_3(Node* n); 631 void scaled_iv_plus_offset_4(Node* n); 632 void scaled_iv_plus_offset_5(Node* n); 633 void scaled_iv_plus_offset_6(Node* n); 634 void scaled_iv_plus_offset_7(Node* n); 635 void scaled_iv_plus_offset_8(Node* n); 636 637 void scaled_iv_1(Node* n); 638 void scaled_iv_2(Node* n, int scale); 639 void scaled_iv_3(Node* n, int scale); 640 void scaled_iv_4(Node* n, int scale); 641 void scaled_iv_5(Node* n, int scale); 642 void scaled_iv_6(Node* n, int scale); 643 void scaled_iv_7(Node* n); 644 void scaled_iv_8(Node* n, SWPointer* tmp); 645 void scaled_iv_9(Node* n, int _scale, int _offset, int mult); 646 void scaled_iv_10(Node* n); 647 648 void offset_plus_k_1(Node* n); 649 void offset_plus_k_2(Node* n, int _offset); 650 void offset_plus_k_3(Node* n, int _offset); 651 void offset_plus_k_4(Node* n); 652 void offset_plus_k_5(Node* n, Node* _invar); 653 void offset_plus_k_6(Node* n, Node* _invar, bool _negate_invar, int _offset); 654 void offset_plus_k_7(Node* n, Node* _invar, bool _negate_invar, int _offset); 655 void offset_plus_k_8(Node* n, Node* _invar, bool _negate_invar, int _offset); 656 void offset_plus_k_9(Node* n, Node* _invar, bool _negate_invar, int _offset); 657 void offset_plus_k_10(Node* n, Node* _invar, bool _negate_invar, int _offset); 658 void offset_plus_k_11(Node* n); 659 660 } _tracer;//TRacer; 661 #endif 662 }; 663 664 #endif // SHARE_VM_OPTO_SUPERWORD_HPP