--- /dev/null Thu Jan 01 00:00:00 1970 +0000
+++ b/hotspot/src/jdk.internal.vm.compiler/share/classes/org.graalvm.compiler.core.amd64/src/org/graalvm/compiler/core/amd64/AMD64LIRGenerator.java Thu Feb 16 15:46:09 2017 -0800
@@ -0,0 +1,463 @@
+/*
+ * Copyright (c) 2009, 2016, Oracle and/or its affiliates. All rights reserved.
+ * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
+ *
+ * This code is free software; you can redistribute it and/or modify it
+ * under the terms of the GNU General Public License version 2 only, as
+ * published by the Free Software Foundation.
+ *
+ * This code is distributed in the hope that it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+ * version 2 for more details (a copy is included in the LICENSE file that
+ * accompanied this code).
+ *
+ * You should have received a copy of the GNU General Public License version
+ * 2 along with this work; if not, write to the Free Software Foundation,
+ * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
+ * or visit www.oracle.com if you need additional information or have any
+ * questions.
+ */
+
+package org.graalvm.compiler.core.amd64;
+
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64BinaryArithmetic.CMP;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize.DWORD;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize.PD;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize.PS;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize.QWORD;
+import static org.graalvm.compiler.core.common.GraalOptions.GeneratePIC;
+import static org.graalvm.compiler.lir.LIRValueUtil.asConstantValue;
+import static org.graalvm.compiler.lir.LIRValueUtil.asJavaConstant;
+import static org.graalvm.compiler.lir.LIRValueUtil.isJavaConstant;
+import static jdk.vm.ci.code.ValueUtil.isAllocatableValue;
+
+import org.graalvm.compiler.asm.NumUtil;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64MIOp;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64RMOp;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.ConditionFlag;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.SSEOp;
+import org.graalvm.compiler.core.common.LIRKind;
+import org.graalvm.compiler.core.common.calc.Condition;
+import org.graalvm.compiler.core.common.spi.ForeignCallLinkage;
+import org.graalvm.compiler.core.common.spi.LIRKindTool;
+import org.graalvm.compiler.debug.GraalError;
+import org.graalvm.compiler.lir.ConstantValue;
+import org.graalvm.compiler.lir.LIRFrameState;
+import org.graalvm.compiler.lir.LIRInstruction;
+import org.graalvm.compiler.lir.LIRValueUtil;
+import org.graalvm.compiler.lir.LabelRef;
+import org.graalvm.compiler.lir.StandardOp.JumpOp;
+import org.graalvm.compiler.lir.StandardOp.SaveRegistersOp;
+import org.graalvm.compiler.lir.SwitchStrategy;
+import org.graalvm.compiler.lir.Variable;
+import org.graalvm.compiler.lir.amd64.AMD64AddressValue;
+import org.graalvm.compiler.lir.amd64.AMD64ArithmeticLIRGeneratorTool;
+import org.graalvm.compiler.lir.amd64.AMD64ArrayEqualsOp;
+import org.graalvm.compiler.lir.amd64.AMD64BinaryConsumer;
+import org.graalvm.compiler.lir.amd64.AMD64ByteSwapOp;
+import org.graalvm.compiler.lir.amd64.AMD64Call;
+import org.graalvm.compiler.lir.amd64.AMD64ControlFlow.BranchOp;
+import org.graalvm.compiler.lir.amd64.AMD64ControlFlow.CondMoveOp;
+import org.graalvm.compiler.lir.amd64.AMD64ControlFlow.FloatBranchOp;
+import org.graalvm.compiler.lir.amd64.AMD64ControlFlow.FloatCondMoveOp;
+import org.graalvm.compiler.lir.amd64.AMD64ControlFlow.ReturnOp;
+import org.graalvm.compiler.lir.amd64.AMD64ControlFlow.StrategySwitchOp;
+import org.graalvm.compiler.lir.amd64.AMD64ControlFlow.TableSwitchOp;
+import org.graalvm.compiler.lir.amd64.AMD64Move;
+import org.graalvm.compiler.lir.amd64.AMD64Move.CompareAndSwapOp;
+import org.graalvm.compiler.lir.amd64.AMD64Move.MembarOp;
+import org.graalvm.compiler.lir.amd64.AMD64Move.StackLeaOp;
+import org.graalvm.compiler.lir.amd64.AMD64PauseOp;
+import org.graalvm.compiler.lir.amd64.AMD64ZapRegistersOp;
+import org.graalvm.compiler.lir.amd64.AMD64ZapStackOp;
+import org.graalvm.compiler.lir.gen.LIRGenerationResult;
+import org.graalvm.compiler.lir.gen.LIRGenerator;
+import org.graalvm.compiler.phases.util.Providers;
+
+import jdk.vm.ci.amd64.AMD64;
+import jdk.vm.ci.amd64.AMD64Kind;
+import jdk.vm.ci.code.CallingConvention;
+import jdk.vm.ci.code.Register;
+import jdk.vm.ci.code.RegisterValue;
+import jdk.vm.ci.code.StackSlot;
+import jdk.vm.ci.meta.AllocatableValue;
+import jdk.vm.ci.meta.JavaConstant;
+import jdk.vm.ci.meta.JavaKind;
+import jdk.vm.ci.meta.PlatformKind;
+import jdk.vm.ci.meta.VMConstant;
+import jdk.vm.ci.meta.Value;
+import jdk.vm.ci.meta.ValueKind;
+
+/**
+ * This class implements the AMD64 specific portion of the LIR generator.
+ */
+public abstract class AMD64LIRGenerator extends LIRGenerator {
+
+ public AMD64LIRGenerator(LIRKindTool lirKindTool, AMD64ArithmeticLIRGenerator arithmeticLIRGen, MoveFactory moveFactory, Providers providers, LIRGenerationResult lirGenRes) {
+ super(lirKindTool, arithmeticLIRGen, moveFactory, providers, lirGenRes);
+ }
+
+ /**
+ * Checks whether the supplied constant can be used without loading it into a register for store
+ * operations, i.e., on the right hand side of a memory access.
+ *
+ * @param c The constant to check.
+ * @return True if the constant can be used directly, false if the constant needs to be in a
+ * register.
+ */
+ protected static final boolean canStoreConstant(JavaConstant c) {
+ // there is no immediate move of 64-bit constants on Intel
+ switch (c.getJavaKind()) {
+ case Long: {
+ long l = c.asLong();
+ return (int) l == l;
+ }
+ case Double:
+ return false;
+ case Object:
+ return c.isNull();
+ default:
+ return true;
+ }
+ }
+
+ @Override
+ protected JavaConstant zapValueForKind(PlatformKind kind) {
+ long dead = 0xDEADDEADDEADDEADL;
+ switch ((AMD64Kind) kind) {
+ case BYTE:
+ return JavaConstant.forByte((byte) dead);
+ case WORD:
+ return JavaConstant.forShort((short) dead);
+ case DWORD:
+ return JavaConstant.forInt((int) dead);
+ case QWORD:
+ return JavaConstant.forLong(dead);
+ case SINGLE:
+ return JavaConstant.forFloat(Float.intBitsToFloat((int) dead));
+ default:
+ // we don't support vector types, so just zap with double for all of them
+ return JavaConstant.forDouble(Double.longBitsToDouble(dead));
+ }
+ }
+
+ public AMD64AddressValue asAddressValue(Value address) {
+ if (address instanceof AMD64AddressValue) {
+ return (AMD64AddressValue) address;
+ } else {
+ if (address instanceof JavaConstant) {
+ long displacement = ((JavaConstant) address).asLong();
+ if (NumUtil.isInt(displacement)) {
+ return new AMD64AddressValue(address.getValueKind(), Value.ILLEGAL, (int) displacement);
+ }
+ }
+ return new AMD64AddressValue(address.getValueKind(), asAllocatable(address), 0);
+ }
+ }
+
+ @Override
+ public Variable emitAddress(AllocatableValue stackslot) {
+ Variable result = newVariable(LIRKind.value(target().arch.getWordKind()));
+ append(new StackLeaOp(result, stackslot));
+ return result;
+ }
+
+ /**
+ * The AMD64 backend only uses DWORD and QWORD values in registers because of a performance
+ * penalty when accessing WORD or BYTE registers. This function converts small integer kinds to
+ * DWORD.
+ */
+ @Override
+ public <K extends ValueKind<K>> K toRegisterKind(K kind) {
+ switch ((AMD64Kind) kind.getPlatformKind()) {
+ case BYTE:
+ case WORD:
+ return kind.changeType(AMD64Kind.DWORD);
+ default:
+ return kind;
+ }
+ }
+
+ @Override
+ public Variable emitCompareAndSwap(Value address, Value expectedValue, Value newValue, Value trueValue, Value falseValue) {
+ ValueKind<?> kind = newValue.getValueKind();
+ assert kind.equals(expectedValue.getValueKind());
+ AMD64Kind memKind = (AMD64Kind) kind.getPlatformKind();
+
+ AMD64AddressValue addressValue = asAddressValue(address);
+ RegisterValue raxRes = AMD64.rax.asValue(kind);
+ emitMove(raxRes, expectedValue);
+ append(new CompareAndSwapOp(memKind, raxRes, addressValue, raxRes, asAllocatable(newValue)));
+
+ assert trueValue.getValueKind().equals(falseValue.getValueKind());
+ Variable result = newVariable(trueValue.getValueKind());
+ append(new CondMoveOp(result, Condition.EQ, asAllocatable(trueValue), falseValue));
+ return result;
+ }
+
+ @Override
+ public Value emitAtomicReadAndAdd(Value address, Value delta) {
+ ValueKind<?> kind = delta.getValueKind();
+ Variable result = newVariable(kind);
+ AMD64AddressValue addressValue = asAddressValue(address);
+ append(new AMD64Move.AtomicReadAndAddOp((AMD64Kind) kind.getPlatformKind(), result, addressValue, asAllocatable(delta)));
+ return result;
+ }
+
+ @Override
+ public Value emitAtomicReadAndWrite(Value address, Value newValue) {
+ ValueKind<?> kind = newValue.getValueKind();
+ Variable result = newVariable(kind);
+ AMD64AddressValue addressValue = asAddressValue(address);
+ append(new AMD64Move.AtomicReadAndWriteOp((AMD64Kind) kind.getPlatformKind(), result, addressValue, asAllocatable(newValue)));
+ return result;
+ }
+
+ @Override
+ public void emitNullCheck(Value address, LIRFrameState state) {
+ append(new AMD64Move.NullCheckOp(asAddressValue(address), state));
+ }
+
+ @Override
+ public void emitJump(LabelRef label) {
+ assert label != null;
+ append(new JumpOp(label));
+ }
+
+ @Override
+ public void emitCompareBranch(PlatformKind cmpKind, Value left, Value right, Condition cond, boolean unorderedIsTrue, LabelRef trueLabel, LabelRef falseLabel, double trueLabelProbability) {
+ boolean mirrored = emitCompare(cmpKind, left, right);
+ Condition finalCondition = mirrored ? cond.mirror() : cond;
+ if (cmpKind == AMD64Kind.SINGLE || cmpKind == AMD64Kind.DOUBLE) {
+ append(new FloatBranchOp(finalCondition, unorderedIsTrue, trueLabel, falseLabel, trueLabelProbability));
+ } else {
+ append(new BranchOp(finalCondition, trueLabel, falseLabel, trueLabelProbability));
+ }
+ }
+
+ public void emitCompareBranchMemory(AMD64Kind cmpKind, Value left, AMD64AddressValue right, LIRFrameState state, Condition cond, boolean unorderedIsTrue, LabelRef trueLabel, LabelRef falseLabel,
+ double trueLabelProbability) {
+ boolean mirrored = emitCompareMemory(cmpKind, left, right, state);
+ Condition finalCondition = mirrored ? cond.mirror() : cond;
+ if (cmpKind.isXMM()) {
+ append(new FloatBranchOp(finalCondition, unorderedIsTrue, trueLabel, falseLabel, trueLabelProbability));
+ } else {
+ append(new BranchOp(finalCondition, trueLabel, falseLabel, trueLabelProbability));
+ }
+ }
+
+ @Override
+ public void emitOverflowCheckBranch(LabelRef overflow, LabelRef noOverflow, LIRKind cmpLIRKind, double overflowProbability) {
+ append(new BranchOp(ConditionFlag.Overflow, overflow, noOverflow, overflowProbability));
+ }
+
+ @Override
+ public void emitIntegerTestBranch(Value left, Value right, LabelRef trueDestination, LabelRef falseDestination, double trueDestinationProbability) {
+ emitIntegerTest(left, right);
+ append(new BranchOp(Condition.EQ, trueDestination, falseDestination, trueDestinationProbability));
+ }
+
+ @Override
+ public Variable emitConditionalMove(PlatformKind cmpKind, Value left, Value right, Condition cond, boolean unorderedIsTrue, Value trueValue, Value falseValue) {
+ boolean mirrored = emitCompare(cmpKind, left, right);
+ Condition finalCondition = mirrored ? cond.mirror() : cond;
+
+ Variable result = newVariable(trueValue.getValueKind());
+ if (cmpKind == AMD64Kind.SINGLE || cmpKind == AMD64Kind.DOUBLE) {
+ append(new FloatCondMoveOp(result, finalCondition, unorderedIsTrue, load(trueValue), load(falseValue)));
+ } else {
+ append(new CondMoveOp(result, finalCondition, load(trueValue), loadNonConst(falseValue)));
+ }
+ return result;
+ }
+
+ @Override
+ public Variable emitIntegerTestMove(Value left, Value right, Value trueValue, Value falseValue) {
+ emitIntegerTest(left, right);
+ Variable result = newVariable(trueValue.getValueKind());
+ append(new CondMoveOp(result, Condition.EQ, load(trueValue), loadNonConst(falseValue)));
+ return result;
+ }
+
+ private void emitIntegerTest(Value a, Value b) {
+ assert ((AMD64Kind) a.getPlatformKind()).isInteger();
+ OperandSize size = a.getPlatformKind() == AMD64Kind.QWORD ? QWORD : DWORD;
+ if (isJavaConstant(b) && NumUtil.is32bit(asJavaConstant(b).asLong())) {
+ append(new AMD64BinaryConsumer.ConstOp(AMD64MIOp.TEST, size, asAllocatable(a), (int) asJavaConstant(b).asLong()));
+ } else if (isJavaConstant(a) && NumUtil.is32bit(asJavaConstant(a).asLong())) {
+ append(new AMD64BinaryConsumer.ConstOp(AMD64MIOp.TEST, size, asAllocatable(b), (int) asJavaConstant(a).asLong()));
+ } else if (isAllocatableValue(b)) {
+ append(new AMD64BinaryConsumer.Op(AMD64RMOp.TEST, size, asAllocatable(b), asAllocatable(a)));
+ } else {
+ append(new AMD64BinaryConsumer.Op(AMD64RMOp.TEST, size, asAllocatable(a), asAllocatable(b)));
+ }
+ }
+
+ /**
+ * This method emits the compare against memory instruction, and may reorder the operands. It
+ * returns true if it did so.
+ *
+ * @param b the right operand of the comparison
+ * @return true if the left and right operands were switched, false otherwise
+ */
+ private boolean emitCompareMemory(AMD64Kind cmpKind, Value a, AMD64AddressValue b, LIRFrameState state) {
+ OperandSize size;
+ switch (cmpKind) {
+ case BYTE:
+ size = OperandSize.BYTE;
+ break;
+ case WORD:
+ size = OperandSize.WORD;
+ break;
+ case DWORD:
+ size = OperandSize.DWORD;
+ break;
+ case QWORD:
+ size = OperandSize.QWORD;
+ break;
+ case SINGLE:
+ append(new AMD64BinaryConsumer.MemoryRMOp(SSEOp.UCOMIS, PS, asAllocatable(a), b, state));
+ return false;
+ case DOUBLE:
+ append(new AMD64BinaryConsumer.MemoryRMOp(SSEOp.UCOMIS, PD, asAllocatable(a), b, state));
+ return false;
+ default:
+ throw GraalError.shouldNotReachHere("unexpected kind: " + cmpKind);
+ }
+
+ if (isJavaConstant(a)) {
+ return emitCompareMemoryConOp(size, asConstantValue(a), b, state);
+ } else {
+ return emitCompareRegMemoryOp(size, asAllocatable(a), b, state);
+ }
+ }
+
+ protected boolean emitCompareMemoryConOp(OperandSize size, ConstantValue a, AMD64AddressValue b, LIRFrameState state) {
+ if (JavaConstant.isNull(a.getConstant())) {
+ append(new AMD64BinaryConsumer.MemoryConstOp(CMP, size, b, 0, state));
+ return true;
+ } else if (a.getConstant() instanceof VMConstant && size == DWORD) {
+ VMConstant vc = (VMConstant) a.getConstant();
+ append(new AMD64BinaryConsumer.MemoryVMConstOp(CMP.getMIOpcode(size, false), b, vc, state));
+ return true;
+ } else {
+ long value = a.getJavaConstant().asLong();
+ if (NumUtil.is32bit(value)) {
+ append(new AMD64BinaryConsumer.MemoryConstOp(CMP, size, b, (int) value, state));
+ return true;
+ } else {
+ return emitCompareRegMemoryOp(size, asAllocatable(a), b, state);
+ }
+ }
+ }
+
+ private boolean emitCompareRegMemoryOp(OperandSize size, AllocatableValue a, AMD64AddressValue b, LIRFrameState state) {
+ AMD64RMOp op = CMP.getRMOpcode(size);
+ append(new AMD64BinaryConsumer.MemoryRMOp(op, size, a, b, state));
+ return false;
+ }
+
+ /**
+ * This method emits the compare instruction, and may reorder the operands. It returns true if
+ * it did so.
+ *
+ * @param a the left operand of the comparison
+ * @param b the right operand of the comparison
+ * @return true if the left and right operands were switched, false otherwise
+ */
+ private boolean emitCompare(PlatformKind cmpKind, Value a, Value b) {
+ Variable left;
+ Value right;
+ boolean mirrored;
+ if (LIRValueUtil.isVariable(b)) {
+ left = load(b);
+ right = loadNonConst(a);
+ mirrored = true;
+ } else {
+ left = load(a);
+ right = loadNonConst(b);
+ mirrored = false;
+ }
+ ((AMD64ArithmeticLIRGeneratorTool) arithmeticLIRGen).emitCompareOp((AMD64Kind) cmpKind, left, right);
+ return mirrored;
+ }
+
+ @Override
+ public void emitMembar(int barriers) {
+ int necessaryBarriers = target().arch.requiredBarriers(barriers);
+ if (target().isMP && necessaryBarriers != 0) {
+ append(new MembarOp(necessaryBarriers));
+ }
+ }
+
+ public abstract void emitCCall(long address, CallingConvention nativeCallingConvention, Value[] args, int numberOfFloatingPointArguments);
+
+ @Override
+ protected void emitForeignCallOp(ForeignCallLinkage linkage, Value result, Value[] arguments, Value[] temps, LIRFrameState info) {
+ long maxOffset = linkage.getMaxCallTargetOffset();
+ if (maxOffset != (int) maxOffset && !GeneratePIC.getValue()) {
+ append(new AMD64Call.DirectFarForeignCallOp(linkage, result, arguments, temps, info));
+ } else {
+ append(new AMD64Call.DirectNearForeignCallOp(linkage, result, arguments, temps, info));
+ }
+ }
+
+ @Override
+ public Variable emitByteSwap(Value input) {
+ Variable result = newVariable(LIRKind.combine(input));
+ append(new AMD64ByteSwapOp(result, input));
+ return result;
+ }
+
+ @Override
+ public Variable emitArrayEquals(JavaKind kind, Value array1, Value array2, Value length) {
+ Variable result = newVariable(LIRKind.value(AMD64Kind.DWORD));
+ append(new AMD64ArrayEqualsOp(this, kind, result, array1, array2, asAllocatable(length)));
+ return result;
+ }
+
+ @Override
+ public void emitReturn(JavaKind kind, Value input) {
+ AllocatableValue operand = Value.ILLEGAL;
+ if (input != null) {
+ operand = resultOperandFor(kind, input.getValueKind());
+ emitMove(operand, input);
+ }
+ append(new ReturnOp(operand));
+ }
+
+ protected StrategySwitchOp createStrategySwitchOp(SwitchStrategy strategy, LabelRef[] keyTargets, LabelRef defaultTarget, Variable key, AllocatableValue temp) {
+ return new StrategySwitchOp(strategy, keyTargets, defaultTarget, key, temp);
+ }
+
+ @Override
+ public void emitStrategySwitch(SwitchStrategy strategy, Variable key, LabelRef[] keyTargets, LabelRef defaultTarget) {
+ // a temp is needed for loading object constants
+ boolean needsTemp = !LIRKind.isValue(key);
+ append(createStrategySwitchOp(strategy, keyTargets, defaultTarget, key, needsTemp ? newVariable(key.getValueKind()) : Value.ILLEGAL));
+ }
+
+ @Override
+ protected void emitTableSwitch(int lowKey, LabelRef defaultTarget, LabelRef[] targets, Value key) {
+ append(new TableSwitchOp(lowKey, defaultTarget, targets, key, newVariable(LIRKind.value(target().arch.getWordKind())), newVariable(key.getValueKind())));
+ }
+
+ @Override
+ public void emitPause() {
+ append(new AMD64PauseOp());
+ }
+
+ @Override
+ public SaveRegistersOp createZapRegisters(Register[] zappedRegisters, JavaConstant[] zapValues) {
+ return new AMD64ZapRegistersOp(zappedRegisters, zapValues);
+ }
+
+ @Override
+ public LIRInstruction createZapArgumentSpace(StackSlot[] zappedStack, JavaConstant[] zapValues) {
+ return new AMD64ZapStackOp(zappedStack, zapValues);
+ }
+}