1 /* 2 * Copyright (c) 2011, 2019, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 */ 23 24 25 package org.graalvm.compiler.lir.amd64; 26 27 import static jdk.vm.ci.code.ValueUtil.asRegister; 28 import static jdk.vm.ci.code.ValueUtil.isRegister; 29 import static org.graalvm.compiler.lir.LIRInstruction.OperandFlag.CONST; 30 import static org.graalvm.compiler.lir.LIRInstruction.OperandFlag.HINT; 31 import static org.graalvm.compiler.lir.LIRInstruction.OperandFlag.ILLEGAL; 32 import static org.graalvm.compiler.lir.LIRInstruction.OperandFlag.REG; 33 import static org.graalvm.compiler.lir.LIRInstruction.OperandFlag.STACK; 34 35 import org.graalvm.compiler.asm.Label; 36 import org.graalvm.compiler.asm.amd64.AMD64Address; 37 import org.graalvm.compiler.asm.amd64.AMD64Address.Scale; 38 import org.graalvm.compiler.asm.amd64.AMD64Assembler.ConditionFlag; 39 import org.graalvm.compiler.asm.amd64.AMD64MacroAssembler; 40 import org.graalvm.compiler.code.CompilationResult.JumpTable; 41 import org.graalvm.compiler.core.common.NumUtil; 42 import org.graalvm.compiler.core.common.calc.Condition; 43 import org.graalvm.compiler.debug.GraalError; 44 import org.graalvm.compiler.lir.LIRInstructionClass; 45 import org.graalvm.compiler.lir.LabelRef; 46 import org.graalvm.compiler.lir.Opcode; 47 import org.graalvm.compiler.lir.StandardOp; 48 import org.graalvm.compiler.lir.StandardOp.BlockEndOp; 49 import org.graalvm.compiler.lir.SwitchStrategy; 50 import org.graalvm.compiler.lir.SwitchStrategy.BaseSwitchClosure; 51 import org.graalvm.compiler.lir.Variable; 52 import org.graalvm.compiler.lir.asm.CompilationResultBuilder; 53 54 import jdk.vm.ci.amd64.AMD64; 55 import jdk.vm.ci.amd64.AMD64.CPUFeature; 56 import jdk.vm.ci.amd64.AMD64Kind; 57 import jdk.vm.ci.code.Register; 58 import jdk.vm.ci.meta.AllocatableValue; 59 import jdk.vm.ci.meta.Constant; 60 import jdk.vm.ci.meta.JavaConstant; 61 import jdk.vm.ci.meta.Value; 62 63 public class AMD64ControlFlow { 64 65 public static final class ReturnOp extends AMD64BlockEndOp implements BlockEndOp { 66 public static final LIRInstructionClass<ReturnOp> TYPE = LIRInstructionClass.create(ReturnOp.class); 67 @Use({REG, ILLEGAL}) protected Value x; 68 69 public ReturnOp(Value x) { 70 super(TYPE); 71 this.x = x; 72 } 73 74 @Override 75 public void emitCode(CompilationResultBuilder crb, AMD64MacroAssembler masm) { 76 crb.frameContext.leave(crb); 77 /* 78 * We potentially return to the interpreter, and that's an AVX-SSE transition. The only 79 * live value at this point should be the return value in either rax, or in xmm0 with 80 * the upper half of the register unused, so we don't destroy any value here. 81 */ 82 if (masm.supports(CPUFeature.AVX)) { 83 masm.vzeroupper(); 84 } 85 masm.ret(0); 86 } 87 } 88 89 public static class BranchOp extends AMD64BlockEndOp implements StandardOp.BranchOp { 90 public static final LIRInstructionClass<BranchOp> TYPE = LIRInstructionClass.create(BranchOp.class); 91 protected final ConditionFlag condition; 92 protected final LabelRef trueDestination; 93 protected final LabelRef falseDestination; 94 95 private final double trueDestinationProbability; 96 97 public BranchOp(Condition condition, LabelRef trueDestination, LabelRef falseDestination, double trueDestinationProbability) { 98 this(intCond(condition), trueDestination, falseDestination, trueDestinationProbability); 99 } 100 101 public BranchOp(ConditionFlag condition, LabelRef trueDestination, LabelRef falseDestination, double trueDestinationProbability) { 102 this(TYPE, condition, trueDestination, falseDestination, trueDestinationProbability); 103 } 104 105 protected BranchOp(LIRInstructionClass<? extends BranchOp> c, ConditionFlag condition, LabelRef trueDestination, LabelRef falseDestination, double trueDestinationProbability) { 106 super(c); 107 this.condition = condition; 108 this.trueDestination = trueDestination; 109 this.falseDestination = falseDestination; 110 this.trueDestinationProbability = trueDestinationProbability; 111 } 112 113 @Override 114 public void emitCode(CompilationResultBuilder crb, AMD64MacroAssembler masm) { 115 boolean isNegated = false; 116 int jccPos = masm.position(); 117 /* 118 * The strategy for emitting jumps is: If either trueDestination or falseDestination is 119 * the successor block, assume the block scheduler did the correct thing and jcc to the 120 * other. Otherwise, we need a jcc followed by a jmp. Use the branch probability to make 121 * sure it is more likely to branch on the jcc (= less likely to execute both the jcc 122 * and the jmp instead of just the jcc). In the case of loops, that means the jcc is the 123 * back-edge. 124 */ 125 if (crb.isSuccessorEdge(trueDestination)) { 126 jcc(masm, true, falseDestination); 127 isNegated = true; 128 } else if (crb.isSuccessorEdge(falseDestination)) { 129 jcc(masm, false, trueDestination); 130 } else if (trueDestinationProbability < 0.5) { 131 jcc(masm, true, falseDestination); 132 masm.jmp(trueDestination.label()); 133 isNegated = true; 134 } else { 135 jcc(masm, false, trueDestination); 136 masm.jmp(falseDestination.label()); 137 } 138 crb.recordBranch(jccPos, isNegated); 139 } 140 141 protected void jcc(AMD64MacroAssembler masm, boolean negate, LabelRef target) { 142 masm.jcc(negate ? condition.negate() : condition, target.label()); 143 } 144 } 145 146 public static final class FloatBranchOp extends BranchOp { 147 public static final LIRInstructionClass<FloatBranchOp> TYPE = LIRInstructionClass.create(FloatBranchOp.class); 148 protected boolean unorderedIsTrue; 149 150 public FloatBranchOp(Condition condition, boolean unorderedIsTrue, LabelRef trueDestination, LabelRef falseDestination, double trueDestinationProbability) { 151 super(TYPE, floatCond(condition), trueDestination, falseDestination, trueDestinationProbability); 152 this.unorderedIsTrue = unorderedIsTrue; 153 } 154 155 @Override 156 protected void jcc(AMD64MacroAssembler masm, boolean negate, LabelRef target) { 157 floatJcc(masm, negate ? condition.negate() : condition, negate ? !unorderedIsTrue : unorderedIsTrue, target.label()); 158 } 159 } 160 161 public static class StrategySwitchOp extends AMD64BlockEndOp { 162 public static final LIRInstructionClass<StrategySwitchOp> TYPE = LIRInstructionClass.create(StrategySwitchOp.class); 163 protected final Constant[] keyConstants; 164 private final LabelRef[] keyTargets; 165 private LabelRef defaultTarget; 166 @Alive({REG}) protected Value key; 167 @Temp({REG, ILLEGAL}) protected Value scratch; 168 protected final SwitchStrategy strategy; 169 170 public StrategySwitchOp(SwitchStrategy strategy, LabelRef[] keyTargets, LabelRef defaultTarget, Value key, Value scratch) { 171 this(TYPE, strategy, keyTargets, defaultTarget, key, scratch); 172 } 173 174 protected StrategySwitchOp(LIRInstructionClass<? extends StrategySwitchOp> c, SwitchStrategy strategy, LabelRef[] keyTargets, LabelRef defaultTarget, Value key, Value scratch) { 175 super(c); 176 this.strategy = strategy; 177 this.keyConstants = strategy.getKeyConstants(); 178 this.keyTargets = keyTargets; 179 this.defaultTarget = defaultTarget; 180 this.key = key; 181 this.scratch = scratch; 182 assert keyConstants.length == keyTargets.length; 183 assert keyConstants.length == strategy.keyProbabilities.length; 184 } 185 186 @Override 187 public void emitCode(final CompilationResultBuilder crb, final AMD64MacroAssembler masm) { 188 strategy.run(new SwitchClosure(asRegister(key), crb, masm)); 189 } 190 191 public class SwitchClosure extends BaseSwitchClosure { 192 193 protected final Register keyRegister; 194 protected final CompilationResultBuilder crb; 195 protected final AMD64MacroAssembler masm; 196 197 protected SwitchClosure(Register keyRegister, CompilationResultBuilder crb, AMD64MacroAssembler masm) { 198 super(crb, masm, keyTargets, defaultTarget); 199 this.keyRegister = keyRegister; 200 this.crb = crb; 201 this.masm = masm; 202 } 203 204 protected void emitComparison(Constant c) { 205 JavaConstant jc = (JavaConstant) c; 206 switch (jc.getJavaKind()) { 207 case Int: 208 long lc = jc.asLong(); 209 assert NumUtil.isInt(lc); 210 masm.cmpl(keyRegister, (int) lc); 211 break; 212 case Long: 213 masm.cmpq(keyRegister, (AMD64Address) crb.asLongConstRef(jc)); 214 break; 215 case Object: 216 AMD64Move.const2reg(crb, masm, asRegister(scratch), jc); 217 masm.cmpptr(keyRegister, asRegister(scratch)); 218 break; 219 default: 220 throw new GraalError("switch only supported for int, long and object"); 221 } 222 } 223 224 @Override 225 protected void conditionalJump(int index, Condition condition, Label target) { 226 emitComparison(keyConstants[index]); 227 masm.jcc(intCond(condition), target); 228 } 229 } 230 } 231 232 public static final class TableSwitchOp extends AMD64BlockEndOp { 233 public static final LIRInstructionClass<TableSwitchOp> TYPE = LIRInstructionClass.create(TableSwitchOp.class); 234 private final int lowKey; 235 private final LabelRef defaultTarget; 236 private final LabelRef[] targets; 237 @Use protected Value index; 238 @Temp({REG, HINT}) protected Value idxScratch; 239 @Temp protected Value scratch; 240 241 public TableSwitchOp(final int lowKey, final LabelRef defaultTarget, final LabelRef[] targets, Value index, Variable scratch, Variable idxScratch) { 242 super(TYPE); 243 this.lowKey = lowKey; 244 this.defaultTarget = defaultTarget; 245 this.targets = targets; 246 this.index = index; 247 this.scratch = scratch; 248 this.idxScratch = idxScratch; 249 } 250 251 @Override 252 public void emitCode(CompilationResultBuilder crb, AMD64MacroAssembler masm) { 253 Register indexReg = asRegister(index, AMD64Kind.DWORD); 254 Register idxScratchReg = asRegister(idxScratch, AMD64Kind.DWORD); 255 Register scratchReg = asRegister(scratch, AMD64Kind.QWORD); 256 257 if (!indexReg.equals(idxScratchReg)) { 258 masm.movl(idxScratchReg, indexReg); 259 } 260 261 // Compare index against jump table bounds 262 int highKey = lowKey + targets.length - 1; 263 if (lowKey != 0) { 264 // subtract the low value from the switch value 265 masm.subl(idxScratchReg, lowKey); 266 masm.cmpl(idxScratchReg, highKey - lowKey); 267 } else { 268 masm.cmpl(idxScratchReg, highKey); 269 } 270 271 // Jump to default target if index is not within the jump table 272 if (defaultTarget != null) { 273 masm.jcc(ConditionFlag.Above, defaultTarget.label()); 274 } 275 276 // Set scratch to address of jump table 277 masm.leaq(scratchReg, new AMD64Address(AMD64.rip, 0)); 278 final int afterLea = masm.position(); 279 280 // Load jump table entry into scratch and jump to it 281 masm.movslq(idxScratchReg, new AMD64Address(scratchReg, idxScratchReg, Scale.Times4, 0)); 282 masm.addq(scratchReg, idxScratchReg); 283 masm.jmp(scratchReg); 284 285 // Inserting padding so that jump table address is 4-byte aligned 286 masm.align(4); 287 288 // Patch LEA instruction above now that we know the position of the jump table 289 // this is ugly but there is no better way to do this given the assembler API 290 final int jumpTablePos = masm.position(); 291 final int leaDisplacementPosition = afterLea - 4; 292 masm.emitInt(jumpTablePos - afterLea, leaDisplacementPosition); 293 294 // Emit jump table entries 295 for (LabelRef target : targets) { 296 Label label = target.label(); 297 int offsetToJumpTableBase = masm.position() - jumpTablePos; 298 if (label.isBound()) { 299 int imm32 = label.position() - jumpTablePos; 300 masm.emitInt(imm32); 301 } else { 302 label.addPatchAt(masm.position()); 303 304 masm.emitByte(0); // pseudo-opcode for jump table entry 305 masm.emitShort(offsetToJumpTableBase); 306 masm.emitByte(0); // padding to make jump table entry 4 bytes wide 307 } 308 } 309 310 JumpTable jt = new JumpTable(jumpTablePos, lowKey, highKey, 4); 311 crb.compilationResult.addAnnotation(jt); 312 } 313 } 314 315 public static final class HashTableSwitchOp extends AMD64BlockEndOp { 316 public static final LIRInstructionClass<HashTableSwitchOp> TYPE = LIRInstructionClass.create(HashTableSwitchOp.class); 317 private final JavaConstant[] keys; 318 private final LabelRef defaultTarget; 319 private final LabelRef[] targets; 320 @Alive protected Value value; 321 @Alive protected Value hash; 322 @Temp({REG}) protected Value entryScratch; 323 @Temp({REG}) protected Value scratch; 324 325 public HashTableSwitchOp(final JavaConstant[] keys, final LabelRef defaultTarget, LabelRef[] targets, Value value, Value hash, Variable scratch, Variable entryScratch) { 326 super(TYPE); 327 this.keys = keys; 328 this.defaultTarget = defaultTarget; 329 this.targets = targets; 330 this.value = value; 331 this.hash = hash; 332 this.scratch = scratch; 333 this.entryScratch = entryScratch; 334 } 335 336 @Override 337 public void emitCode(CompilationResultBuilder crb, AMD64MacroAssembler masm) { 338 Register valueReg = asRegister(value, AMD64Kind.DWORD); 339 Register indexReg = asRegister(hash, AMD64Kind.DWORD); 340 Register scratchReg = asRegister(scratch, AMD64Kind.QWORD); 341 Register entryScratchReg = asRegister(entryScratch, AMD64Kind.QWORD); 342 343 // Set scratch to address of jump table 344 masm.leaq(scratchReg, new AMD64Address(AMD64.rip, 0)); 345 final int afterLea = masm.position(); 346 347 // When the default target is set, the jump table contains entries with two DWORDS: 348 // the original key before hashing and the label jump address 349 if (defaultTarget != null) { 350 351 // Move the table entry (two DWORDs) into a QWORD 352 masm.movq(entryScratchReg, new AMD64Address(scratchReg, indexReg, Scale.Times8, 0)); 353 354 // Jump to the default target if the first DWORD (original key) doesn't match the 355 // current key. Accounts for hash collisions with unknown keys 356 masm.cmpl(entryScratchReg, valueReg); 357 masm.jcc(ConditionFlag.NotEqual, defaultTarget.label()); 358 359 // Shift to the second DWORD 360 masm.sarq(entryScratchReg, 32); 361 } else { 362 363 // The jump table has a single DWORD with the label address if there's no 364 // default target 365 masm.movslq(entryScratchReg, new AMD64Address(scratchReg, indexReg, Scale.Times4, 0)); 366 } 367 masm.addq(scratchReg, entryScratchReg); 368 masm.jmp(scratchReg); 369 370 // Inserting padding so that jump the table address is aligned 371 if (defaultTarget != null) { 372 masm.align(8); 373 } else { 374 masm.align(4); 375 } 376 377 // Patch LEA instruction above now that we know the position of the jump table 378 // this is ugly but there is no better way to do this given the assembler API 379 final int jumpTablePos = masm.position(); 380 final int leaDisplacementPosition = afterLea - 4; 381 masm.emitInt(jumpTablePos - afterLea, leaDisplacementPosition); 382 383 // Emit jump table entries 384 for (int i = 0; i < targets.length; i++) { 385 386 Label label = targets[i].label(); 387 388 if (defaultTarget != null) { 389 masm.emitInt(keys[i].asInt()); 390 } 391 if (label.isBound()) { 392 int imm32 = label.position() - jumpTablePos; 393 masm.emitInt(imm32); 394 } else { 395 int offsetToJumpTableBase = masm.position() - jumpTablePos; 396 label.addPatchAt(masm.position()); 397 masm.emitByte(0); // pseudo-opcode for jump table entry 398 masm.emitShort(offsetToJumpTableBase); 399 masm.emitByte(0); // padding to make jump table entry 4 bytes wide 400 } 401 } 402 403 JumpTable jt = new JumpTable(jumpTablePos, keys[0].asInt(), keys[keys.length - 1].asInt(), 4); 404 crb.compilationResult.addAnnotation(jt); 405 } 406 } 407 408 @Opcode("SETcc") 409 public static final class CondSetOp extends AMD64LIRInstruction { 410 public static final LIRInstructionClass<CondSetOp> TYPE = LIRInstructionClass.create(CondSetOp.class); 411 @Def({REG, HINT}) protected Value result; 412 private final ConditionFlag condition; 413 414 public CondSetOp(Variable result, Condition condition) { 415 super(TYPE); 416 this.result = result; 417 this.condition = intCond(condition); 418 } 419 420 @Override 421 public void emitCode(CompilationResultBuilder crb, AMD64MacroAssembler masm) { 422 setcc(masm, result, condition); 423 } 424 } 425 426 @Opcode("SETcc") 427 public static final class FloatCondSetOp extends AMD64LIRInstruction { 428 public static final LIRInstructionClass<FloatCondSetOp> TYPE = LIRInstructionClass.create(FloatCondSetOp.class); 429 @Def({REG, HINT}) protected Value result; 430 private final ConditionFlag condition; 431 432 public FloatCondSetOp(Variable result, Condition condition) { 433 super(TYPE); 434 this.result = result; 435 this.condition = floatCond(condition); 436 } 437 438 @Override 439 public void emitCode(CompilationResultBuilder crb, AMD64MacroAssembler masm) { 440 setcc(masm, result, condition); 441 } 442 } 443 444 @Opcode("CMOVE") 445 public static final class CondMoveOp extends AMD64LIRInstruction { 446 public static final LIRInstructionClass<CondMoveOp> TYPE = LIRInstructionClass.create(CondMoveOp.class); 447 @Def({REG, HINT}) protected Value result; 448 @Alive({REG}) protected Value trueValue; 449 @Use({REG, STACK, CONST}) protected Value falseValue; 450 private final ConditionFlag condition; 451 452 public CondMoveOp(Variable result, Condition condition, AllocatableValue trueValue, Value falseValue) { 453 super(TYPE); 454 this.result = result; 455 this.condition = intCond(condition); 456 this.trueValue = trueValue; 457 this.falseValue = falseValue; 458 } 459 460 @Override 461 public void emitCode(CompilationResultBuilder crb, AMD64MacroAssembler masm) { 462 cmove(crb, masm, result, false, condition, false, trueValue, falseValue); 463 } 464 } 465 466 @Opcode("CMOVE") 467 public static final class FloatCondMoveOp extends AMD64LIRInstruction { 468 public static final LIRInstructionClass<FloatCondMoveOp> TYPE = LIRInstructionClass.create(FloatCondMoveOp.class); 469 @Def({REG}) protected Value result; 470 @Alive({REG}) protected Value trueValue; 471 @Alive({REG}) protected Value falseValue; 472 private final ConditionFlag condition; 473 private final boolean unorderedIsTrue; 474 475 public FloatCondMoveOp(Variable result, Condition condition, boolean unorderedIsTrue, Variable trueValue, Variable falseValue) { 476 super(TYPE); 477 this.result = result; 478 this.condition = floatCond(condition); 479 this.unorderedIsTrue = unorderedIsTrue; 480 this.trueValue = trueValue; 481 this.falseValue = falseValue; 482 } 483 484 @Override 485 public void emitCode(CompilationResultBuilder crb, AMD64MacroAssembler masm) { 486 cmove(crb, masm, result, true, condition, unorderedIsTrue, trueValue, falseValue); 487 } 488 } 489 490 private static void floatJcc(AMD64MacroAssembler masm, ConditionFlag condition, boolean unorderedIsTrue, Label label) { 491 Label endLabel = new Label(); 492 if (unorderedIsTrue && !trueOnUnordered(condition)) { 493 masm.jcc(ConditionFlag.Parity, label); 494 } else if (!unorderedIsTrue && trueOnUnordered(condition)) { 495 masm.jccb(ConditionFlag.Parity, endLabel); 496 } 497 masm.jcc(condition, label); 498 masm.bind(endLabel); 499 } 500 501 private static void cmove(CompilationResultBuilder crb, AMD64MacroAssembler masm, Value result, boolean isFloat, ConditionFlag condition, boolean unorderedIsTrue, Value trueValue, 502 Value falseValue) { 503 // check that we don't overwrite an input operand before it is used. 504 assert !result.equals(trueValue); 505 506 AMD64Move.move(crb, masm, result, falseValue); 507 cmove(crb, masm, result, condition, trueValue); 508 509 if (isFloat) { 510 if (unorderedIsTrue && !trueOnUnordered(condition)) { 511 cmove(crb, masm, result, ConditionFlag.Parity, trueValue); 512 } else if (!unorderedIsTrue && trueOnUnordered(condition)) { 513 cmove(crb, masm, result, ConditionFlag.Parity, falseValue); 514 } 515 } 516 } 517 518 private static void cmove(CompilationResultBuilder crb, AMD64MacroAssembler masm, Value result, ConditionFlag cond, Value other) { 519 if (isRegister(other)) { 520 assert !asRegister(other).equals(asRegister(result)) : "other already overwritten by previous move"; 521 switch ((AMD64Kind) other.getPlatformKind()) { 522 case BYTE: 523 case WORD: 524 case DWORD: 525 masm.cmovl(cond, asRegister(result), asRegister(other)); 526 break; 527 case QWORD: 528 masm.cmovq(cond, asRegister(result), asRegister(other)); 529 break; 530 default: 531 throw GraalError.shouldNotReachHere(); 532 } 533 } else { 534 AMD64Address addr = (AMD64Address) crb.asAddress(other); 535 switch ((AMD64Kind) other.getPlatformKind()) { 536 case BYTE: 537 case WORD: 538 case DWORD: 539 masm.cmovl(cond, asRegister(result), addr); 540 break; 541 case QWORD: 542 masm.cmovq(cond, asRegister(result), addr); 543 break; 544 default: 545 throw GraalError.shouldNotReachHere(); 546 } 547 } 548 } 549 550 private static void setcc(AMD64MacroAssembler masm, Value result, ConditionFlag cond) { 551 switch ((AMD64Kind) result.getPlatformKind()) { 552 case BYTE: 553 case WORD: 554 case DWORD: 555 masm.setl(cond, asRegister(result)); 556 break; 557 case QWORD: 558 masm.setq(cond, asRegister(result)); 559 break; 560 default: 561 throw GraalError.shouldNotReachHere(); 562 } 563 } 564 565 private static ConditionFlag intCond(Condition cond) { 566 switch (cond) { 567 case EQ: 568 return ConditionFlag.Equal; 569 case NE: 570 return ConditionFlag.NotEqual; 571 case LT: 572 return ConditionFlag.Less; 573 case LE: 574 return ConditionFlag.LessEqual; 575 case GE: 576 return ConditionFlag.GreaterEqual; 577 case GT: 578 return ConditionFlag.Greater; 579 case BE: 580 return ConditionFlag.BelowEqual; 581 case AE: 582 return ConditionFlag.AboveEqual; 583 case AT: 584 return ConditionFlag.Above; 585 case BT: 586 return ConditionFlag.Below; 587 default: 588 throw GraalError.shouldNotReachHere(); 589 } 590 } 591 592 private static ConditionFlag floatCond(Condition cond) { 593 switch (cond) { 594 case EQ: 595 return ConditionFlag.Equal; 596 case NE: 597 return ConditionFlag.NotEqual; 598 case LT: 599 return ConditionFlag.Below; 600 case LE: 601 return ConditionFlag.BelowEqual; 602 case GE: 603 return ConditionFlag.AboveEqual; 604 case GT: 605 return ConditionFlag.Above; 606 default: 607 throw GraalError.shouldNotReachHere(); 608 } 609 } 610 611 public static boolean trueOnUnordered(Condition condition) { 612 return trueOnUnordered(floatCond(condition)); 613 } 614 615 private static boolean trueOnUnordered(ConditionFlag condition) { 616 switch (condition) { 617 case AboveEqual: 618 case NotEqual: 619 case Above: 620 case Less: 621 case Overflow: 622 return false; 623 case Equal: 624 case BelowEqual: 625 case Below: 626 case GreaterEqual: 627 case NoOverflow: 628 return true; 629 default: 630 throw GraalError.shouldNotReachHere(); 631 } 632 } 633 }