--- /dev/null Thu Jan 01 00:00:00 1970 +0000
+++ b/src/jdk.internal.vm.compiler/share/classes/org.graalvm.compiler.core.amd64/src/org/graalvm/compiler/core/amd64/AMD64NodeMatchRules.java Tue Sep 12 19:03:39 2017 +0200
@@ -0,0 +1,583 @@
+/*
+ * Copyright (c) 2009, 2016, Oracle and/or its affiliates. All rights reserved.
+ * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
+ *
+ * This code is free software; you can redistribute it and/or modify it
+ * under the terms of the GNU General Public License version 2 only, as
+ * published by the Free Software Foundation.
+ *
+ * This code is distributed in the hope that it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+ * version 2 for more details (a copy is included in the LICENSE file that
+ * accompanied this code).
+ *
+ * You should have received a copy of the GNU General Public License version
+ * 2 along with this work; if not, write to the Free Software Foundation,
+ * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
+ * or visit www.oracle.com if you need additional information or have any
+ * questions.
+ */
+
+package org.graalvm.compiler.core.amd64;
+
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64BinaryArithmetic.ADD;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64BinaryArithmetic.AND;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64BinaryArithmetic.OR;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64BinaryArithmetic.SUB;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64BinaryArithmetic.XOR;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64RMOp.MOVSX;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64RMOp.MOVSXB;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64RMOp.MOVSXD;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize.DWORD;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize.QWORD;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize.SD;
+import static org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize.SS;
+
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64MIOp;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64RMOp;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.AMD64RRMOp;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.AVXOp;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.OperandSize;
+import org.graalvm.compiler.asm.amd64.AMD64Assembler.SSEOp;
+import org.graalvm.compiler.core.common.LIRKind;
+import org.graalvm.compiler.core.common.NumUtil;
+import org.graalvm.compiler.core.common.calc.Condition;
+import org.graalvm.compiler.core.gen.NodeLIRBuilder;
+import org.graalvm.compiler.core.gen.NodeMatchRules;
+import org.graalvm.compiler.core.match.ComplexMatchResult;
+import org.graalvm.compiler.core.match.MatchRule;
+import org.graalvm.compiler.debug.GraalError;
+import org.graalvm.compiler.lir.LIRFrameState;
+import org.graalvm.compiler.lir.LIRValueUtil;
+import org.graalvm.compiler.lir.LabelRef;
+import org.graalvm.compiler.lir.amd64.AMD64AddressValue;
+import org.graalvm.compiler.lir.amd64.AMD64BinaryConsumer;
+import org.graalvm.compiler.lir.amd64.AMD64ControlFlow.BranchOp;
+import org.graalvm.compiler.lir.gen.LIRGeneratorTool;
+import org.graalvm.compiler.nodes.ConstantNode;
+import org.graalvm.compiler.nodes.DeoptimizingNode;
+import org.graalvm.compiler.nodes.IfNode;
+import org.graalvm.compiler.nodes.ValueNode;
+import org.graalvm.compiler.nodes.calc.CompareNode;
+import org.graalvm.compiler.nodes.calc.FloatConvertNode;
+import org.graalvm.compiler.nodes.calc.LeftShiftNode;
+import org.graalvm.compiler.nodes.calc.NarrowNode;
+import org.graalvm.compiler.nodes.calc.ReinterpretNode;
+import org.graalvm.compiler.nodes.calc.SignExtendNode;
+import org.graalvm.compiler.nodes.calc.UnsignedRightShiftNode;
+import org.graalvm.compiler.nodes.calc.ZeroExtendNode;
+import org.graalvm.compiler.nodes.java.LogicCompareAndSwapNode;
+import org.graalvm.compiler.nodes.java.ValueCompareAndSwapNode;
+import org.graalvm.compiler.nodes.memory.Access;
+import org.graalvm.compiler.nodes.memory.LIRLowerableAccess;
+import org.graalvm.compiler.nodes.memory.WriteNode;
+import org.graalvm.compiler.nodes.util.GraphUtil;
+
+import jdk.vm.ci.amd64.AMD64;
+import jdk.vm.ci.amd64.AMD64.CPUFeature;
+import jdk.vm.ci.amd64.AMD64Kind;
+import jdk.vm.ci.code.TargetDescription;
+import jdk.vm.ci.meta.AllocatableValue;
+import jdk.vm.ci.meta.JavaConstant;
+import jdk.vm.ci.meta.PlatformKind;
+import jdk.vm.ci.meta.Value;
+import jdk.vm.ci.meta.ValueKind;
+
+public class AMD64NodeMatchRules extends NodeMatchRules {
+
+ public AMD64NodeMatchRules(LIRGeneratorTool gen) {
+ super(gen);
+ }
+
+ protected LIRFrameState getState(Access access) {
+ if (access instanceof DeoptimizingNode) {
+ return state((DeoptimizingNode) access);
+ }
+ return null;
+ }
+
+ protected AMD64Kind getMemoryKind(LIRLowerableAccess access) {
+ return (AMD64Kind) getLirKind(access).getPlatformKind();
+ }
+
+ protected LIRKind getLirKind(LIRLowerableAccess access) {
+ return gen.getLIRKind(access.getAccessStamp());
+ }
+
+ protected OperandSize getMemorySize(LIRLowerableAccess access) {
+ switch (getMemoryKind(access)) {
+ case BYTE:
+ return OperandSize.BYTE;
+ case WORD:
+ return OperandSize.WORD;
+ case DWORD:
+ return OperandSize.DWORD;
+ case QWORD:
+ return OperandSize.QWORD;
+ case SINGLE:
+ return OperandSize.SS;
+ case DOUBLE:
+ return OperandSize.SD;
+ default:
+ throw GraalError.shouldNotReachHere("unsupported memory access type " + getMemoryKind(access));
+ }
+ }
+
+ protected ComplexMatchResult emitCompareBranchMemory(IfNode ifNode, CompareNode compare, ValueNode value, LIRLowerableAccess access) {
+ Condition cond = compare.condition();
+ AMD64Kind kind = getMemoryKind(access);
+ boolean matchedAsConstant = false; // For assertion checking
+
+ if (value.isConstant()) {
+ JavaConstant constant = value.asJavaConstant();
+ if (constant != null) {
+ if (kind == AMD64Kind.QWORD && !constant.getJavaKind().isObject() && !NumUtil.isInt(constant.asLong())) {
+ // Only imm32 as long
+ return null;
+ }
+ // A QWORD that can be encoded as int can be embedded as a constant
+ matchedAsConstant = kind == AMD64Kind.QWORD && !constant.getJavaKind().isObject() && NumUtil.isInt(constant.asLong());
+ }
+ if (kind == AMD64Kind.DWORD) {
+ // Any DWORD value should be embeddable as a constant
+ matchedAsConstant = true;
+ }
+ if (kind.isXMM()) {
+ ifNode.getDebug().log("Skipping constant compares for float kinds");
+ return null;
+ }
+ }
+ boolean matchedAsConstantFinal = matchedAsConstant;
+
+ /*
+ * emitCompareBranchMemory expects the memory on the right, so mirror the condition if
+ * that's not true. It might be mirrored again the actual compare is emitted but that's ok.
+ */
+ Condition finalCondition = GraphUtil.unproxify(compare.getX()) == access ? cond.mirror() : cond;
+ return new ComplexMatchResult() {
+ @Override
+ public Value evaluate(NodeLIRBuilder builder) {
+ LabelRef trueLabel = getLIRBlock(ifNode.trueSuccessor());
+ LabelRef falseLabel = getLIRBlock(ifNode.falseSuccessor());
+ boolean unorderedIsTrue = compare.unorderedIsTrue();
+ double trueLabelProbability = ifNode.probability(ifNode.trueSuccessor());
+ Value other = operand(value);
+ /*
+ * Check that patterns which were matched as a constant actually end up seeing a
+ * constant in the LIR.
+ */
+ assert !matchedAsConstantFinal || !LIRValueUtil.isVariable(other) : "expected constant value " + value;
+ AMD64AddressValue address = (AMD64AddressValue) operand(access.getAddress());
+ getLIRGeneratorTool().emitCompareBranchMemory(kind, other, address, getState(access), finalCondition, unorderedIsTrue, trueLabel, falseLabel, trueLabelProbability);
+ return null;
+ }
+ };
+ }
+
+ private ComplexMatchResult emitIntegerTestBranchMemory(IfNode x, ValueNode value, LIRLowerableAccess access) {
+ LabelRef trueLabel = getLIRBlock(x.trueSuccessor());
+ LabelRef falseLabel = getLIRBlock(x.falseSuccessor());
+ double trueLabelProbability = x.probability(x.trueSuccessor());
+ AMD64Kind kind = getMemoryKind(access);
+ OperandSize size = kind == AMD64Kind.QWORD ? QWORD : DWORD;
+ if (value.isConstant()) {
+ JavaConstant constant = value.asJavaConstant();
+ if (constant != null && kind == AMD64Kind.QWORD && !NumUtil.isInt(constant.asLong())) {
+ // Only imm32 as long
+ return null;
+ }
+ return builder -> {
+ AMD64AddressValue address = (AMD64AddressValue) operand(access.getAddress());
+ gen.append(new AMD64BinaryConsumer.MemoryConstOp(AMD64MIOp.TEST, size, address, (int) constant.asLong(), getState(access)));
+ gen.append(new BranchOp(Condition.EQ, trueLabel, falseLabel, trueLabelProbability));
+ return null;
+ };
+ } else {
+ return builder -> {
+ AMD64AddressValue address = (AMD64AddressValue) operand(access.getAddress());
+ gen.append(new AMD64BinaryConsumer.MemoryRMOp(AMD64RMOp.TEST, size, gen.asAllocatable(operand(value)), address, getState(access)));
+ gen.append(new BranchOp(Condition.EQ, trueLabel, falseLabel, trueLabelProbability));
+ return null;
+ };
+ }
+ }
+
+ protected ComplexMatchResult emitConvertMemoryOp(PlatformKind kind, AMD64RMOp op, OperandSize size, Access access, ValueKind<?> addressKind) {
+ return builder -> {
+ AMD64AddressValue address = (AMD64AddressValue) operand(access.getAddress());
+ LIRFrameState state = getState(access);
+ if (addressKind != null) {
+ address = address.withKind(addressKind);
+ }
+ return getArithmeticLIRGenerator().emitConvertMemoryOp(kind, op, size, address, state);
+ };
+ }
+
+ protected ComplexMatchResult emitConvertMemoryOp(PlatformKind kind, AMD64RMOp op, OperandSize size, Access access) {
+ return emitConvertMemoryOp(kind, op, size, access, null);
+ }
+
+ private ComplexMatchResult emitSignExtendMemory(Access access, int fromBits, int toBits, ValueKind<?> addressKind) {
+ assert fromBits <= toBits && toBits <= 64;
+ AMD64Kind kind = null;
+ AMD64RMOp op;
+ OperandSize size;
+ if (fromBits == toBits) {
+ return null;
+ } else if (toBits > 32) {
+ kind = AMD64Kind.QWORD;
+ size = OperandSize.QWORD;
+ // sign extend to 64 bits
+ switch (fromBits) {
+ case 8:
+ op = MOVSXB;
+ break;
+ case 16:
+ op = MOVSX;
+ break;
+ case 32:
+ op = MOVSXD;
+ break;
+ default:
+ throw GraalError.unimplemented("unsupported sign extension (" + fromBits + " bit -> " + toBits + " bit)");
+ }
+ } else {
+ kind = AMD64Kind.DWORD;
+ size = OperandSize.DWORD;
+ // sign extend to 32 bits (smaller values are internally represented as 32 bit values)
+ switch (fromBits) {
+ case 8:
+ op = MOVSXB;
+ break;
+ case 16:
+ op = MOVSX;
+ break;
+ case 32:
+ return null;
+ default:
+ throw GraalError.unimplemented("unsupported sign extension (" + fromBits + " bit -> " + toBits + " bit)");
+ }
+ }
+ if (kind != null && op != null) {
+ return emitConvertMemoryOp(kind, op, size, access, addressKind);
+ }
+ return null;
+ }
+
+ private Value emitReinterpretMemory(LIRKind to, Access access) {
+ AMD64AddressValue address = (AMD64AddressValue) operand(access.getAddress());
+ LIRFrameState state = getState(access);
+ return getArithmeticLIRGenerator().emitLoad(to, address, state);
+ }
+
+ @MatchRule("(If (IntegerTest Read=access value))")
+ @MatchRule("(If (IntegerTest FloatingRead=access value))")
+ public ComplexMatchResult integerTestBranchMemory(IfNode root, LIRLowerableAccess access, ValueNode value) {
+ return emitIntegerTestBranchMemory(root, value, access);
+ }
+
+ @MatchRule("(If (IntegerEquals=compare value Read=access))")
+ @MatchRule("(If (IntegerLessThan=compare value Read=access))")
+ @MatchRule("(If (IntegerBelow=compare value Read=access))")
+ @MatchRule("(If (IntegerEquals=compare value FloatingRead=access))")
+ @MatchRule("(If (IntegerLessThan=compare value FloatingRead=access))")
+ @MatchRule("(If (IntegerBelow=compare value FloatingRead=access))")
+ @MatchRule("(If (FloatEquals=compare value Read=access))")
+ @MatchRule("(If (FloatEquals=compare value FloatingRead=access))")
+ @MatchRule("(If (FloatLessThan=compare value Read=access))")
+ @MatchRule("(If (FloatLessThan=compare value FloatingRead=access))")
+ @MatchRule("(If (PointerEquals=compare value Read=access))")
+ @MatchRule("(If (PointerEquals=compare value FloatingRead=access))")
+ @MatchRule("(If (ObjectEquals=compare value Read=access))")
+ @MatchRule("(If (ObjectEquals=compare value FloatingRead=access))")
+ public ComplexMatchResult ifCompareMemory(IfNode root, CompareNode compare, ValueNode value, LIRLowerableAccess access) {
+ return emitCompareBranchMemory(root, compare, value, access);
+ }
+
+ @MatchRule("(If (ObjectEquals=compare value ValueCompareAndSwap=cas))")
+ @MatchRule("(If (PointerEquals=compare value ValueCompareAndSwap=cas))")
+ @MatchRule("(If (FloatEquals=compare value ValueCompareAndSwap=cas))")
+ @MatchRule("(If (IntegerEquals=compare value ValueCompareAndSwap=cas))")
+ public ComplexMatchResult ifCompareValueCas(IfNode root, CompareNode compare, ValueNode value, ValueCompareAndSwapNode cas) {
+ assert compare.condition() == Condition.EQ;
+ if (value == cas.getExpectedValue() && cas.usages().count() == 1) {
+ return builder -> {
+ LIRKind kind = getLirKind(cas);
+ LabelRef trueLabel = getLIRBlock(root.trueSuccessor());
+ LabelRef falseLabel = getLIRBlock(root.falseSuccessor());
+ double trueLabelProbability = root.probability(root.trueSuccessor());
+ Value expectedValue = operand(cas.getExpectedValue());
+ Value newValue = operand(cas.getNewValue());
+ AMD64AddressValue address = (AMD64AddressValue) operand(cas.getAddress());
+ getLIRGeneratorTool().emitCompareAndSwapBranch(kind, address, expectedValue, newValue, Condition.EQ, trueLabel, falseLabel, trueLabelProbability);
+ return null;
+ };
+ }
+ return null;
+ }
+
+ @MatchRule("(If (ObjectEquals=compare value LogicCompareAndSwap=cas))")
+ @MatchRule("(If (PointerEquals=compare value LogicCompareAndSwap=cas))")
+ @MatchRule("(If (FloatEquals=compare value LogicCompareAndSwap=cas))")
+ @MatchRule("(If (IntegerEquals=compare value LogicCompareAndSwap=cas))")
+ public ComplexMatchResult ifCompareLogicCas(IfNode root, CompareNode compare, ValueNode value, LogicCompareAndSwapNode cas) {
+ JavaConstant constant = value.asJavaConstant();
+ assert compare.condition() == Condition.EQ;
+ if (constant != null && cas.usages().count() == 1) {
+ long constantValue = constant.asLong();
+ boolean successIsTrue;
+ if (constantValue == 0) {
+ successIsTrue = false;
+ } else if (constantValue == 1) {
+ successIsTrue = true;
+ } else {
+ return null;
+ }
+ return builder -> {
+ LIRKind kind = getLirKind(cas);
+ LabelRef trueLabel = getLIRBlock(root.trueSuccessor());
+ LabelRef falseLabel = getLIRBlock(root.falseSuccessor());
+ double trueLabelProbability = root.probability(root.trueSuccessor());
+ Value expectedValue = operand(cas.getExpectedValue());
+ Value newValue = operand(cas.getNewValue());
+ AMD64AddressValue address = (AMD64AddressValue) operand(cas.getAddress());
+ Condition condition = successIsTrue ? Condition.EQ : Condition.NE;
+ getLIRGeneratorTool().emitCompareAndSwapBranch(kind, address, expectedValue, newValue, condition, trueLabel, falseLabel, trueLabelProbability);
+ return null;
+ };
+ }
+ return null;
+ }
+
+ @MatchRule("(If (ObjectEquals=compare value FloatingRead=access))")
+ public ComplexMatchResult ifLogicCas(IfNode root, CompareNode compare, ValueNode value, LIRLowerableAccess access) {
+ return emitCompareBranchMemory(root, compare, value, access);
+ }
+
+ @MatchRule("(Or (LeftShift=lshift value Constant) (UnsignedRightShift=rshift value Constant))")
+ public ComplexMatchResult rotateLeftConstant(LeftShiftNode lshift, UnsignedRightShiftNode rshift) {
+ if ((lshift.getShiftAmountMask() & (lshift.getY().asJavaConstant().asInt() + rshift.getY().asJavaConstant().asInt())) == 0) {
+ return builder -> getArithmeticLIRGenerator().emitRol(operand(lshift.getX()), operand(lshift.getY()));
+ }
+ return null;
+ }
+
+ @MatchRule("(Or (LeftShift value (Sub Constant=delta shiftAmount)) (UnsignedRightShift value shiftAmount))")
+ public ComplexMatchResult rotateRightVariable(ValueNode value, ConstantNode delta, ValueNode shiftAmount) {
+ if (delta.asJavaConstant().asLong() == 0 || delta.asJavaConstant().asLong() == 32) {
+ return builder -> getArithmeticLIRGenerator().emitRor(operand(value), operand(shiftAmount));
+ }
+ return null;
+ }
+
+ @MatchRule("(Or (LeftShift value shiftAmount) (UnsignedRightShift value (Sub Constant=delta shiftAmount)))")
+ public ComplexMatchResult rotateLeftVariable(ValueNode value, ValueNode shiftAmount, ConstantNode delta) {
+ if (delta.asJavaConstant().asLong() == 0 || delta.asJavaConstant().asLong() == 32) {
+ return builder -> getArithmeticLIRGenerator().emitRol(operand(value), operand(shiftAmount));
+ }
+ return null;
+ }
+
+ private ComplexMatchResult binaryRead(AMD64RMOp op, OperandSize size, ValueNode value, LIRLowerableAccess access) {
+ return builder -> getArithmeticLIRGenerator().emitBinaryMemory(op, size, getLIRGeneratorTool().asAllocatable(operand(value)), (AMD64AddressValue) operand(access.getAddress()),
+ getState(access));
+ }
+
+ private ComplexMatchResult binaryRead(AMD64RRMOp op, OperandSize size, ValueNode value, LIRLowerableAccess access) {
+ return builder -> getArithmeticLIRGenerator().emitBinaryMemory(op, size, getLIRGeneratorTool().asAllocatable(operand(value)), (AMD64AddressValue) operand(access.getAddress()),
+ getState(access));
+ }
+
+ @MatchRule("(Add value Read=access)")
+ @MatchRule("(Add value FloatingRead=access)")
+ public ComplexMatchResult addMemory(ValueNode value, LIRLowerableAccess access) {
+ OperandSize size = getMemorySize(access);
+ if (size.isXmmType()) {
+ TargetDescription target = getLIRGeneratorTool().target();
+ boolean isAvx = ((AMD64) target.arch).getFeatures().contains(CPUFeature.AVX);
+ if (isAvx) {
+ return binaryRead(AVXOp.ADD, size, value, access);
+ } else {
+ return binaryRead(SSEOp.ADD, size, value, access);
+ }
+ } else {
+ return binaryRead(ADD.getRMOpcode(size), size, value, access);
+ }
+ }
+
+ @MatchRule("(Sub value Read=access)")
+ @MatchRule("(Sub value FloatingRead=access)")
+ public ComplexMatchResult subMemory(ValueNode value, LIRLowerableAccess access) {
+ OperandSize size = getMemorySize(access);
+ if (size.isXmmType()) {
+ TargetDescription target = getLIRGeneratorTool().target();
+ boolean isAvx = ((AMD64) target.arch).getFeatures().contains(CPUFeature.AVX);
+ if (isAvx) {
+ return binaryRead(AVXOp.SUB, size, value, access);
+ } else {
+ return binaryRead(SSEOp.SUB, size, value, access);
+ }
+ } else {
+ return binaryRead(SUB.getRMOpcode(size), size, value, access);
+ }
+ }
+
+ @MatchRule("(Mul value Read=access)")
+ @MatchRule("(Mul value FloatingRead=access)")
+ public ComplexMatchResult mulMemory(ValueNode value, LIRLowerableAccess access) {
+ OperandSize size = getMemorySize(access);
+ if (size.isXmmType()) {
+ TargetDescription target = getLIRGeneratorTool().target();
+ boolean isAvx = ((AMD64) target.arch).getFeatures().contains(CPUFeature.AVX);
+ if (isAvx) {
+ return binaryRead(AVXOp.MUL, size, value, access);
+ } else {
+ return binaryRead(SSEOp.MUL, size, value, access);
+ }
+ } else {
+ return binaryRead(AMD64RMOp.IMUL, size, value, access);
+ }
+ }
+
+ @MatchRule("(And value Read=access)")
+ @MatchRule("(And value FloatingRead=access)")
+ public ComplexMatchResult andMemory(ValueNode value, LIRLowerableAccess access) {
+ OperandSize size = getMemorySize(access);
+ if (size.isXmmType()) {
+ return null;
+ } else {
+ return binaryRead(AND.getRMOpcode(size), size, value, access);
+ }
+ }
+
+ @MatchRule("(Or value Read=access)")
+ @MatchRule("(Or value FloatingRead=access)")
+ public ComplexMatchResult orMemory(ValueNode value, LIRLowerableAccess access) {
+ OperandSize size = getMemorySize(access);
+ if (size.isXmmType()) {
+ return null;
+ } else {
+ return binaryRead(OR.getRMOpcode(size), size, value, access);
+ }
+ }
+
+ @MatchRule("(Xor value Read=access)")
+ @MatchRule("(Xor value FloatingRead=access)")
+ public ComplexMatchResult xorMemory(ValueNode value, LIRLowerableAccess access) {
+ OperandSize size = getMemorySize(access);
+ if (size.isXmmType()) {
+ return null;
+ } else {
+ return binaryRead(XOR.getRMOpcode(size), size, value, access);
+ }
+ }
+
+ @MatchRule("(Write object Narrow=narrow)")
+ public ComplexMatchResult writeNarrow(WriteNode root, NarrowNode narrow) {
+ return builder -> {
+ LIRKind writeKind = getLIRGeneratorTool().getLIRKind(root.value().stamp());
+ getArithmeticLIRGenerator().emitStore(writeKind, operand(root.getAddress()), operand(narrow.getValue()), state(root));
+ return null;
+ };
+ }
+
+ @MatchRule("(SignExtend Read=access)")
+ @MatchRule("(SignExtend FloatingRead=access)")
+ public ComplexMatchResult signExtend(SignExtendNode root, LIRLowerableAccess access) {
+ return emitSignExtendMemory(access, root.getInputBits(), root.getResultBits(), null);
+ }
+
+ @MatchRule("(ZeroExtend Read=access)")
+ @MatchRule("(ZeroExtend FloatingRead=access)")
+ public ComplexMatchResult zeroExtend(ZeroExtendNode root, LIRLowerableAccess access) {
+ AMD64Kind memoryKind = getMemoryKind(access);
+ return builder -> getArithmeticLIRGenerator().emitZeroExtendMemory(memoryKind, root.getResultBits(), (AMD64AddressValue) operand(access.getAddress()), getState(access));
+ }
+
+ @MatchRule("(Narrow Read=access)")
+ @MatchRule("(Narrow FloatingRead=access)")
+ public ComplexMatchResult narrowRead(NarrowNode root, LIRLowerableAccess access) {
+ return new ComplexMatchResult() {
+ @Override
+ public Value evaluate(NodeLIRBuilder builder) {
+ AMD64AddressValue address = (AMD64AddressValue) operand(access.getAddress());
+ LIRKind addressKind = LIRKind.combineDerived(getLIRGeneratorTool().getLIRKind(root.asNode().stamp()),
+ address.getBase(), address.getIndex());
+ AMD64AddressValue newAddress = address.withKind(addressKind);
+ LIRKind readKind = getLIRGeneratorTool().getLIRKind(root.stamp());
+ return getArithmeticLIRGenerator().emitZeroExtendMemory((AMD64Kind) readKind.getPlatformKind(),
+ root.getResultBits(), newAddress, getState(access));
+ }
+ };
+ }
+
+ @MatchRule("(SignExtend (Narrow=narrow Read=access))")
+ @MatchRule("(SignExtend (Narrow=narrow FloatingRead=access))")
+ public ComplexMatchResult signExtendNarrowRead(SignExtendNode root, NarrowNode narrow, LIRLowerableAccess access) {
+ LIRKind kind = getLIRGeneratorTool().getLIRKind(narrow.stamp());
+ return emitSignExtendMemory(access, narrow.getResultBits(), root.getResultBits(), kind);
+ }
+
+ @MatchRule("(FloatConvert Read=access)")
+ @MatchRule("(FloatConvert FloatingRead=access)")
+ public ComplexMatchResult floatConvert(FloatConvertNode root, LIRLowerableAccess access) {
+ switch (root.getFloatConvert()) {
+ case D2F:
+ return emitConvertMemoryOp(AMD64Kind.SINGLE, SSEOp.CVTSD2SS, SD, access);
+ case D2I:
+ return emitConvertMemoryOp(AMD64Kind.DWORD, SSEOp.CVTTSD2SI, DWORD, access);
+ case D2L:
+ return emitConvertMemoryOp(AMD64Kind.QWORD, SSEOp.CVTTSD2SI, QWORD, access);
+ case F2D:
+ return emitConvertMemoryOp(AMD64Kind.DOUBLE, SSEOp.CVTSS2SD, SS, access);
+ case F2I:
+ return emitConvertMemoryOp(AMD64Kind.DWORD, SSEOp.CVTTSS2SI, DWORD, access);
+ case F2L:
+ return emitConvertMemoryOp(AMD64Kind.QWORD, SSEOp.CVTTSS2SI, QWORD, access);
+ case I2D:
+ return emitConvertMemoryOp(AMD64Kind.DOUBLE, SSEOp.CVTSI2SD, DWORD, access);
+ case I2F:
+ return emitConvertMemoryOp(AMD64Kind.SINGLE, SSEOp.CVTSI2SS, DWORD, access);
+ case L2D:
+ return emitConvertMemoryOp(AMD64Kind.DOUBLE, SSEOp.CVTSI2SD, QWORD, access);
+ case L2F:
+ return emitConvertMemoryOp(AMD64Kind.SINGLE, SSEOp.CVTSI2SS, QWORD, access);
+ default:
+ throw GraalError.shouldNotReachHere();
+ }
+ }
+
+ @MatchRule("(Reinterpret Read=access)")
+ @MatchRule("(Reinterpret FloatingRead=access)")
+ public ComplexMatchResult reinterpret(ReinterpretNode root, LIRLowerableAccess access) {
+ return builder -> {
+ LIRKind kind = getLIRGeneratorTool().getLIRKind(root.stamp());
+ return emitReinterpretMemory(kind, access);
+ };
+
+ }
+
+ @MatchRule("(Write object Reinterpret=reinterpret)")
+ public ComplexMatchResult writeReinterpret(WriteNode root, ReinterpretNode reinterpret) {
+ return builder -> {
+ LIRKind kind = getLIRGeneratorTool().getLIRKind(reinterpret.getValue().stamp());
+ AllocatableValue value = getLIRGeneratorTool().asAllocatable(operand(reinterpret.getValue()));
+
+ AMD64AddressValue address = (AMD64AddressValue) operand(root.getAddress());
+ getArithmeticLIRGenerator().emitStore((AMD64Kind) kind.getPlatformKind(), address, value, getState(root));
+ return null;
+ };
+ }
+
+ @Override
+ public AMD64LIRGenerator getLIRGeneratorTool() {
+ return (AMD64LIRGenerator) gen;
+ }
+
+ protected AMD64ArithmeticLIRGenerator getArithmeticLIRGenerator() {
+ return (AMD64ArithmeticLIRGenerator) getLIRGeneratorTool().getArithmetic();
+ }
+}