diff -r 4ebc2e2fb97c -r 71c04702a3d5 src/hotspot/share/c1/c1_Instruction.hpp --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/src/hotspot/share/c1/c1_Instruction.hpp Tue Sep 12 19:03:39 2017 +0200 @@ -0,0 +1,2634 @@ +/* + * Copyright (c) 1999, 2016, Oracle and/or its affiliates. All rights reserved. + * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. + * + * This code is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 only, as + * published by the Free Software Foundation. + * + * This code is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License + * version 2 for more details (a copy is included in the LICENSE file that + * accompanied this code). + * + * You should have received a copy of the GNU General Public License version + * 2 along with this work; if not, write to the Free Software Foundation, + * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. + * + * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA + * or visit www.oracle.com if you need additional information or have any + * questions. + * + */ + +#ifndef SHARE_VM_C1_C1_INSTRUCTION_HPP +#define SHARE_VM_C1_C1_INSTRUCTION_HPP + +#include "c1/c1_Compilation.hpp" +#include "c1/c1_LIR.hpp" +#include "c1/c1_ValueType.hpp" +#include "ci/ciField.hpp" + +// Predefined classes +class ciField; +class ValueStack; +class InstructionPrinter; +class IRScope; +class LIR_OprDesc; +typedef LIR_OprDesc* LIR_Opr; + + +// Instruction class hierarchy +// +// All leaf classes in the class hierarchy are concrete classes +// (i.e., are instantiated). All other classes are abstract and +// serve factoring. + +class Instruction; +class Phi; +class Local; +class Constant; +class AccessField; +class LoadField; +class StoreField; +class AccessArray; +class ArrayLength; +class AccessIndexed; +class LoadIndexed; +class StoreIndexed; +class NegateOp; +class Op2; +class ArithmeticOp; +class ShiftOp; +class LogicOp; +class CompareOp; +class IfOp; +class Convert; +class NullCheck; +class TypeCast; +class OsrEntry; +class ExceptionObject; +class StateSplit; +class Invoke; +class NewInstance; +class NewArray; +class NewTypeArray; +class NewObjectArray; +class NewMultiArray; +class TypeCheck; +class CheckCast; +class InstanceOf; +class AccessMonitor; +class MonitorEnter; +class MonitorExit; +class Intrinsic; +class BlockBegin; +class BlockEnd; +class Goto; +class If; +class IfInstanceOf; +class Switch; +class TableSwitch; +class LookupSwitch; +class Return; +class Throw; +class Base; +class RoundFP; +class UnsafeOp; +class UnsafeRawOp; +class UnsafeGetRaw; +class UnsafePutRaw; +class UnsafeObjectOp; +class UnsafeGetObject; +class UnsafePutObject; +class UnsafeGetAndSetObject; +class ProfileCall; +class ProfileReturnType; +class ProfileInvoke; +class RuntimeCall; +class MemBar; +class RangeCheckPredicate; +#ifdef ASSERT +class Assert; +#endif + +// A Value is a reference to the instruction creating the value +typedef Instruction* Value; +typedef GrowableArray Values; +typedef GrowableArray ValueStackStack; + +// BlockClosure is the base class for block traversal/iteration. + +class BlockClosure: public CompilationResourceObj { + public: + virtual void block_do(BlockBegin* block) = 0; +}; + + +// A simple closure class for visiting the values of an Instruction +class ValueVisitor: public StackObj { + public: + virtual void visit(Value* v) = 0; +}; + + +// Some array and list classes +typedef GrowableArray BlockBeginArray; + +class BlockList: public GrowableArray { + public: + BlockList(): GrowableArray() {} + BlockList(const int size): GrowableArray(size) {} + BlockList(const int size, BlockBegin* init): GrowableArray(size, size, init) {} + + void iterate_forward(BlockClosure* closure); + void iterate_backward(BlockClosure* closure); + void blocks_do(void f(BlockBegin*)); + void values_do(ValueVisitor* f); + void print(bool cfg_only = false, bool live_only = false) PRODUCT_RETURN; +}; + + +// InstructionVisitors provide type-based dispatch for instructions. +// For each concrete Instruction class X, a virtual function do_X is +// provided. Functionality that needs to be implemented for all classes +// (e.g., printing, code generation) is factored out into a specialised +// visitor instead of added to the Instruction classes itself. + +class InstructionVisitor: public StackObj { + public: + virtual void do_Phi (Phi* x) = 0; + virtual void do_Local (Local* x) = 0; + virtual void do_Constant (Constant* x) = 0; + virtual void do_LoadField (LoadField* x) = 0; + virtual void do_StoreField (StoreField* x) = 0; + virtual void do_ArrayLength (ArrayLength* x) = 0; + virtual void do_LoadIndexed (LoadIndexed* x) = 0; + virtual void do_StoreIndexed (StoreIndexed* x) = 0; + virtual void do_NegateOp (NegateOp* x) = 0; + virtual void do_ArithmeticOp (ArithmeticOp* x) = 0; + virtual void do_ShiftOp (ShiftOp* x) = 0; + virtual void do_LogicOp (LogicOp* x) = 0; + virtual void do_CompareOp (CompareOp* x) = 0; + virtual void do_IfOp (IfOp* x) = 0; + virtual void do_Convert (Convert* x) = 0; + virtual void do_NullCheck (NullCheck* x) = 0; + virtual void do_TypeCast (TypeCast* x) = 0; + virtual void do_Invoke (Invoke* x) = 0; + virtual void do_NewInstance (NewInstance* x) = 0; + virtual void do_NewTypeArray (NewTypeArray* x) = 0; + virtual void do_NewObjectArray (NewObjectArray* x) = 0; + virtual void do_NewMultiArray (NewMultiArray* x) = 0; + virtual void do_CheckCast (CheckCast* x) = 0; + virtual void do_InstanceOf (InstanceOf* x) = 0; + virtual void do_MonitorEnter (MonitorEnter* x) = 0; + virtual void do_MonitorExit (MonitorExit* x) = 0; + virtual void do_Intrinsic (Intrinsic* x) = 0; + virtual void do_BlockBegin (BlockBegin* x) = 0; + virtual void do_Goto (Goto* x) = 0; + virtual void do_If (If* x) = 0; + virtual void do_IfInstanceOf (IfInstanceOf* x) = 0; + virtual void do_TableSwitch (TableSwitch* x) = 0; + virtual void do_LookupSwitch (LookupSwitch* x) = 0; + virtual void do_Return (Return* x) = 0; + virtual void do_Throw (Throw* x) = 0; + virtual void do_Base (Base* x) = 0; + virtual void do_OsrEntry (OsrEntry* x) = 0; + virtual void do_ExceptionObject(ExceptionObject* x) = 0; + virtual void do_RoundFP (RoundFP* x) = 0; + virtual void do_UnsafeGetRaw (UnsafeGetRaw* x) = 0; + virtual void do_UnsafePutRaw (UnsafePutRaw* x) = 0; + virtual void do_UnsafeGetObject(UnsafeGetObject* x) = 0; + virtual void do_UnsafePutObject(UnsafePutObject* x) = 0; + virtual void do_UnsafeGetAndSetObject(UnsafeGetAndSetObject* x) = 0; + virtual void do_ProfileCall (ProfileCall* x) = 0; + virtual void do_ProfileReturnType (ProfileReturnType* x) = 0; + virtual void do_ProfileInvoke (ProfileInvoke* x) = 0; + virtual void do_RuntimeCall (RuntimeCall* x) = 0; + virtual void do_MemBar (MemBar* x) = 0; + virtual void do_RangeCheckPredicate(RangeCheckPredicate* x) = 0; +#ifdef ASSERT + virtual void do_Assert (Assert* x) = 0; +#endif +}; + + +// Hashing support +// +// Note: This hash functions affect the performance +// of ValueMap - make changes carefully! + +#define HASH1(x1 ) ((intx)(x1)) +#define HASH2(x1, x2 ) ((HASH1(x1 ) << 7) ^ HASH1(x2)) +#define HASH3(x1, x2, x3 ) ((HASH2(x1, x2 ) << 7) ^ HASH1(x3)) +#define HASH4(x1, x2, x3, x4) ((HASH3(x1, x2, x3) << 7) ^ HASH1(x4)) + + +// The following macros are used to implement instruction-specific hashing. +// By default, each instruction implements hash() and is_equal(Value), used +// for value numbering/common subexpression elimination. The default imple- +// mentation disables value numbering. Each instruction which can be value- +// numbered, should define corresponding hash() and is_equal(Value) functions +// via the macros below. The f arguments specify all the values/op codes, etc. +// that need to be identical for two instructions to be identical. +// +// Note: The default implementation of hash() returns 0 in order to indicate +// that the instruction should not be considered for value numbering. +// The currently used hash functions do not guarantee that never a 0 +// is produced. While this is still correct, it may be a performance +// bug (no value numbering for that node). However, this situation is +// so unlikely, that we are not going to handle it specially. + +#define HASHING1(class_name, enabled, f1) \ + virtual intx hash() const { \ + return (enabled) ? HASH2(name(), f1) : 0; \ + } \ + virtual bool is_equal(Value v) const { \ + if (!(enabled) ) return false; \ + class_name* _v = v->as_##class_name(); \ + if (_v == NULL ) return false; \ + if (f1 != _v->f1) return false; \ + return true; \ + } \ + + +#define HASHING2(class_name, enabled, f1, f2) \ + virtual intx hash() const { \ + return (enabled) ? HASH3(name(), f1, f2) : 0; \ + } \ + virtual bool is_equal(Value v) const { \ + if (!(enabled) ) return false; \ + class_name* _v = v->as_##class_name(); \ + if (_v == NULL ) return false; \ + if (f1 != _v->f1) return false; \ + if (f2 != _v->f2) return false; \ + return true; \ + } \ + + +#define HASHING3(class_name, enabled, f1, f2, f3) \ + virtual intx hash() const { \ + return (enabled) ? HASH4(name(), f1, f2, f3) : 0; \ + } \ + virtual bool is_equal(Value v) const { \ + if (!(enabled) ) return false; \ + class_name* _v = v->as_##class_name(); \ + if (_v == NULL ) return false; \ + if (f1 != _v->f1) return false; \ + if (f2 != _v->f2) return false; \ + if (f3 != _v->f3) return false; \ + return true; \ + } \ + + +// The mother of all instructions... + +class Instruction: public CompilationResourceObj { + private: + int _id; // the unique instruction id +#ifndef PRODUCT + int _printable_bci; // the bci of the instruction for printing +#endif + int _use_count; // the number of instructions refering to this value (w/o prev/next); only roots can have use count = 0 or > 1 + int _pin_state; // set of PinReason describing the reason for pinning + ValueType* _type; // the instruction value type + Instruction* _next; // the next instruction if any (NULL for BlockEnd instructions) + Instruction* _subst; // the substitution instruction if any + LIR_Opr _operand; // LIR specific information + unsigned int _flags; // Flag bits + + ValueStack* _state_before; // Copy of state with input operands still on stack (or NULL) + ValueStack* _exception_state; // Copy of state for exception handling + XHandlers* _exception_handlers; // Flat list of exception handlers covering this instruction + + friend class UseCountComputer; + friend class BlockBegin; + + void update_exception_state(ValueStack* state); + + protected: + BlockBegin* _block; // Block that contains this instruction + + void set_type(ValueType* type) { + assert(type != NULL, "type must exist"); + _type = type; + } + + // Helper class to keep track of which arguments need a null check + class ArgsNonNullState { + private: + int _nonnull_state; // mask identifying which args are nonnull + public: + ArgsNonNullState() + : _nonnull_state(AllBits) {} + + // Does argument number i needs a null check? + bool arg_needs_null_check(int i) const { + // No data is kept for arguments starting at position 33 so + // conservatively assume that they need a null check. + if (i >= 0 && i < (int)sizeof(_nonnull_state) * BitsPerByte) { + return is_set_nth_bit(_nonnull_state, i); + } + return true; + } + + // Set whether argument number i needs a null check or not + void set_arg_needs_null_check(int i, bool check) { + if (i >= 0 && i < (int)sizeof(_nonnull_state) * BitsPerByte) { + if (check) { + _nonnull_state |= nth_bit(i); + } else { + _nonnull_state &= ~(nth_bit(i)); + } + } + } + }; + + public: + void* operator new(size_t size) throw() { + Compilation* c = Compilation::current(); + void* res = c->arena()->Amalloc(size); + ((Instruction*)res)->_id = c->get_next_id(); + return res; + } + + static const int no_bci = -99; + + enum InstructionFlag { + NeedsNullCheckFlag = 0, + CanTrapFlag, + DirectCompareFlag, + IsEliminatedFlag, + IsSafepointFlag, + IsStaticFlag, + IsStrictfpFlag, + NeedsStoreCheckFlag, + NeedsWriteBarrierFlag, + PreservesStateFlag, + TargetIsFinalFlag, + TargetIsLoadedFlag, + TargetIsStrictfpFlag, + UnorderedIsTrueFlag, + NeedsPatchingFlag, + ThrowIncompatibleClassChangeErrorFlag, + InvokeSpecialReceiverCheckFlag, + ProfileMDOFlag, + IsLinkedInBlockFlag, + NeedsRangeCheckFlag, + InWorkListFlag, + DeoptimizeOnException, + InstructionLastFlag + }; + + public: + bool check_flag(InstructionFlag id) const { return (_flags & (1 << id)) != 0; } + void set_flag(InstructionFlag id, bool f) { _flags = f ? (_flags | (1 << id)) : (_flags & ~(1 << id)); }; + + // 'globally' used condition values + enum Condition { + eql, neq, lss, leq, gtr, geq, aeq, beq + }; + + // Instructions may be pinned for many reasons and under certain conditions + // with enough knowledge it's possible to safely unpin them. + enum PinReason { + PinUnknown = 1 << 0 + , PinExplicitNullCheck = 1 << 3 + , PinStackForStateSplit= 1 << 12 + , PinStateSplitConstructor= 1 << 13 + , PinGlobalValueNumbering= 1 << 14 + }; + + static Condition mirror(Condition cond); + static Condition negate(Condition cond); + + // initialization + static int number_of_instructions() { + return Compilation::current()->number_of_instructions(); + } + + // creation + Instruction(ValueType* type, ValueStack* state_before = NULL, bool type_is_constant = false) + : _use_count(0) +#ifndef PRODUCT + , _printable_bci(-99) +#endif + , _pin_state(0) + , _type(type) + , _next(NULL) + , _block(NULL) + , _subst(NULL) + , _flags(0) + , _operand(LIR_OprFact::illegalOpr) + , _state_before(state_before) + , _exception_handlers(NULL) + { + check_state(state_before); + assert(type != NULL && (!type->is_constant() || type_is_constant), "type must exist"); + update_exception_state(_state_before); + } + + // accessors + int id() const { return _id; } +#ifndef PRODUCT + bool has_printable_bci() const { return _printable_bci != -99; } + int printable_bci() const { assert(has_printable_bci(), "_printable_bci should have been set"); return _printable_bci; } + void set_printable_bci(int bci) { _printable_bci = bci; } +#endif + int dominator_depth(); + int use_count() const { return _use_count; } + int pin_state() const { return _pin_state; } + bool is_pinned() const { return _pin_state != 0 || PinAllInstructions; } + ValueType* type() const { return _type; } + BlockBegin *block() const { return _block; } + Instruction* prev(); // use carefully, expensive operation + Instruction* next() const { return _next; } + bool has_subst() const { return _subst != NULL; } + Instruction* subst() { return _subst == NULL ? this : _subst->subst(); } + LIR_Opr operand() const { return _operand; } + + void set_needs_null_check(bool f) { set_flag(NeedsNullCheckFlag, f); } + bool needs_null_check() const { return check_flag(NeedsNullCheckFlag); } + bool is_linked() const { return check_flag(IsLinkedInBlockFlag); } + bool can_be_linked() { return as_Local() == NULL && as_Phi() == NULL; } + + bool has_uses() const { return use_count() > 0; } + ValueStack* state_before() const { return _state_before; } + ValueStack* exception_state() const { return _exception_state; } + virtual bool needs_exception_state() const { return true; } + XHandlers* exception_handlers() const { return _exception_handlers; } + + // manipulation + void pin(PinReason reason) { _pin_state |= reason; } + void pin() { _pin_state |= PinUnknown; } + // DANGEROUS: only used by EliminateStores + void unpin(PinReason reason) { assert((reason & PinUnknown) == 0, "can't unpin unknown state"); _pin_state &= ~reason; } + + Instruction* set_next(Instruction* next) { + assert(next->has_printable_bci(), "_printable_bci should have been set"); + assert(next != NULL, "must not be NULL"); + assert(as_BlockEnd() == NULL, "BlockEnd instructions must have no next"); + assert(next->can_be_linked(), "shouldn't link these instructions into list"); + + BlockBegin *block = this->block(); + next->_block = block; + + next->set_flag(Instruction::IsLinkedInBlockFlag, true); + _next = next; + return next; + } + + Instruction* set_next(Instruction* next, int bci) { +#ifndef PRODUCT + next->set_printable_bci(bci); +#endif + return set_next(next); + } + + // when blocks are merged + void fixup_block_pointers() { + Instruction *cur = next()->next(); // next()'s block is set in set_next + while (cur && cur->_block != block()) { + cur->_block = block(); + cur = cur->next(); + } + } + + Instruction *insert_after(Instruction *i) { + Instruction* n = _next; + set_next(i); + i->set_next(n); + return _next; + } + + Instruction *insert_after_same_bci(Instruction *i) { +#ifndef PRODUCT + i->set_printable_bci(printable_bci()); +#endif + return insert_after(i); + } + + void set_subst(Instruction* subst) { + assert(subst == NULL || + type()->base() == subst->type()->base() || + subst->type()->base() == illegalType, "type can't change"); + _subst = subst; + } + void set_exception_handlers(XHandlers *xhandlers) { _exception_handlers = xhandlers; } + void set_exception_state(ValueStack* s) { check_state(s); _exception_state = s; } + void set_state_before(ValueStack* s) { check_state(s); _state_before = s; } + + // machine-specifics + void set_operand(LIR_Opr operand) { assert(operand != LIR_OprFact::illegalOpr, "operand must exist"); _operand = operand; } + void clear_operand() { _operand = LIR_OprFact::illegalOpr; } + + // generic + virtual Instruction* as_Instruction() { return this; } // to satisfy HASHING1 macro + virtual Phi* as_Phi() { return NULL; } + virtual Local* as_Local() { return NULL; } + virtual Constant* as_Constant() { return NULL; } + virtual AccessField* as_AccessField() { return NULL; } + virtual LoadField* as_LoadField() { return NULL; } + virtual StoreField* as_StoreField() { return NULL; } + virtual AccessArray* as_AccessArray() { return NULL; } + virtual ArrayLength* as_ArrayLength() { return NULL; } + virtual AccessIndexed* as_AccessIndexed() { return NULL; } + virtual LoadIndexed* as_LoadIndexed() { return NULL; } + virtual StoreIndexed* as_StoreIndexed() { return NULL; } + virtual NegateOp* as_NegateOp() { return NULL; } + virtual Op2* as_Op2() { return NULL; } + virtual ArithmeticOp* as_ArithmeticOp() { return NULL; } + virtual ShiftOp* as_ShiftOp() { return NULL; } + virtual LogicOp* as_LogicOp() { return NULL; } + virtual CompareOp* as_CompareOp() { return NULL; } + virtual IfOp* as_IfOp() { return NULL; } + virtual Convert* as_Convert() { return NULL; } + virtual NullCheck* as_NullCheck() { return NULL; } + virtual OsrEntry* as_OsrEntry() { return NULL; } + virtual StateSplit* as_StateSplit() { return NULL; } + virtual Invoke* as_Invoke() { return NULL; } + virtual NewInstance* as_NewInstance() { return NULL; } + virtual NewArray* as_NewArray() { return NULL; } + virtual NewTypeArray* as_NewTypeArray() { return NULL; } + virtual NewObjectArray* as_NewObjectArray() { return NULL; } + virtual NewMultiArray* as_NewMultiArray() { return NULL; } + virtual TypeCheck* as_TypeCheck() { return NULL; } + virtual CheckCast* as_CheckCast() { return NULL; } + virtual InstanceOf* as_InstanceOf() { return NULL; } + virtual TypeCast* as_TypeCast() { return NULL; } + virtual AccessMonitor* as_AccessMonitor() { return NULL; } + virtual MonitorEnter* as_MonitorEnter() { return NULL; } + virtual MonitorExit* as_MonitorExit() { return NULL; } + virtual Intrinsic* as_Intrinsic() { return NULL; } + virtual BlockBegin* as_BlockBegin() { return NULL; } + virtual BlockEnd* as_BlockEnd() { return NULL; } + virtual Goto* as_Goto() { return NULL; } + virtual If* as_If() { return NULL; } + virtual IfInstanceOf* as_IfInstanceOf() { return NULL; } + virtual TableSwitch* as_TableSwitch() { return NULL; } + virtual LookupSwitch* as_LookupSwitch() { return NULL; } + virtual Return* as_Return() { return NULL; } + virtual Throw* as_Throw() { return NULL; } + virtual Base* as_Base() { return NULL; } + virtual RoundFP* as_RoundFP() { return NULL; } + virtual ExceptionObject* as_ExceptionObject() { return NULL; } + virtual UnsafeOp* as_UnsafeOp() { return NULL; } + virtual ProfileInvoke* as_ProfileInvoke() { return NULL; } + virtual RangeCheckPredicate* as_RangeCheckPredicate() { return NULL; } + +#ifdef ASSERT + virtual Assert* as_Assert() { return NULL; } +#endif + + virtual void visit(InstructionVisitor* v) = 0; + + virtual bool can_trap() const { return false; } + + virtual void input_values_do(ValueVisitor* f) = 0; + virtual void state_values_do(ValueVisitor* f); + virtual void other_values_do(ValueVisitor* f) { /* usually no other - override on demand */ } + void values_do(ValueVisitor* f) { input_values_do(f); state_values_do(f); other_values_do(f); } + + virtual ciType* exact_type() const; + virtual ciType* declared_type() const { return NULL; } + + // hashing + virtual const char* name() const = 0; + HASHING1(Instruction, false, id()) // hashing disabled by default + + // debugging + static void check_state(ValueStack* state) PRODUCT_RETURN; + void print() PRODUCT_RETURN; + void print_line() PRODUCT_RETURN; + void print(InstructionPrinter& ip) PRODUCT_RETURN; +}; + + +// The following macros are used to define base (i.e., non-leaf) +// and leaf instruction classes. They define class-name related +// generic functionality in one place. + +#define BASE(class_name, super_class_name) \ + class class_name: public super_class_name { \ + public: \ + virtual class_name* as_##class_name() { return this; } \ + + +#define LEAF(class_name, super_class_name) \ + BASE(class_name, super_class_name) \ + public: \ + virtual const char* name() const { return #class_name; } \ + virtual void visit(InstructionVisitor* v) { v->do_##class_name(this); } \ + + +// Debugging support + + +#ifdef ASSERT +class AssertValues: public ValueVisitor { + void visit(Value* x) { assert((*x) != NULL, "value must exist"); } +}; + #define ASSERT_VALUES { AssertValues assert_value; values_do(&assert_value); } +#else + #define ASSERT_VALUES +#endif // ASSERT + + +// A Phi is a phi function in the sense of SSA form. It stands for +// the value of a local variable at the beginning of a join block. +// A Phi consists of n operands, one for every incoming branch. + +LEAF(Phi, Instruction) + private: + int _pf_flags; // the flags of the phi function + int _index; // to value on operand stack (index < 0) or to local + public: + // creation + Phi(ValueType* type, BlockBegin* b, int index) + : Instruction(type->base()) + , _pf_flags(0) + , _index(index) + { + _block = b; + NOT_PRODUCT(set_printable_bci(Value(b)->printable_bci())); + if (type->is_illegal()) { + make_illegal(); + } + } + + // flags + enum Flag { + no_flag = 0, + visited = 1 << 0, + cannot_simplify = 1 << 1 + }; + + // accessors + bool is_local() const { return _index >= 0; } + bool is_on_stack() const { return !is_local(); } + int local_index() const { assert(is_local(), ""); return _index; } + int stack_index() const { assert(is_on_stack(), ""); return -(_index+1); } + + Value operand_at(int i) const; + int operand_count() const; + + void set(Flag f) { _pf_flags |= f; } + void clear(Flag f) { _pf_flags &= ~f; } + bool is_set(Flag f) const { return (_pf_flags & f) != 0; } + + // Invalidates phis corresponding to merges of locals of two different types + // (these should never be referenced, otherwise the bytecodes are illegal) + void make_illegal() { + set(cannot_simplify); + set_type(illegalType); + } + + bool is_illegal() const { + return type()->is_illegal(); + } + + // generic + virtual void input_values_do(ValueVisitor* f) { + } +}; + + +// A local is a placeholder for an incoming argument to a function call. +LEAF(Local, Instruction) + private: + int _java_index; // the local index within the method to which the local belongs + bool _is_receiver; // if local variable holds the receiver: "this" for non-static methods + ciType* _declared_type; + public: + // creation + Local(ciType* declared, ValueType* type, int index, bool receiver) + : Instruction(type) + , _java_index(index) + , _declared_type(declared) + , _is_receiver(receiver) + { + NOT_PRODUCT(set_printable_bci(-1)); + } + + // accessors + int java_index() const { return _java_index; } + bool is_receiver() const { return _is_receiver; } + + virtual ciType* declared_type() const { return _declared_type; } + + // generic + virtual void input_values_do(ValueVisitor* f) { /* no values */ } +}; + + +LEAF(Constant, Instruction) + public: + // creation + Constant(ValueType* type): + Instruction(type, NULL, /*type_is_constant*/ true) + { + assert(type->is_constant(), "must be a constant"); + } + + Constant(ValueType* type, ValueStack* state_before): + Instruction(type, state_before, /*type_is_constant*/ true) + { + assert(state_before != NULL, "only used for constants which need patching"); + assert(type->is_constant(), "must be a constant"); + // since it's patching it needs to be pinned + pin(); + } + + // generic + virtual bool can_trap() const { return state_before() != NULL; } + virtual void input_values_do(ValueVisitor* f) { /* no values */ } + + virtual intx hash() const; + virtual bool is_equal(Value v) const; + + virtual ciType* exact_type() const; + + enum CompareResult { not_comparable = -1, cond_false, cond_true }; + + virtual CompareResult compare(Instruction::Condition condition, Value right) const; + BlockBegin* compare(Instruction::Condition cond, Value right, + BlockBegin* true_sux, BlockBegin* false_sux) const { + switch (compare(cond, right)) { + case not_comparable: + return NULL; + case cond_false: + return false_sux; + case cond_true: + return true_sux; + default: + ShouldNotReachHere(); + return NULL; + } + } +}; + + +BASE(AccessField, Instruction) + private: + Value _obj; + int _offset; + ciField* _field; + NullCheck* _explicit_null_check; // For explicit null check elimination + + public: + // creation + AccessField(Value obj, int offset, ciField* field, bool is_static, + ValueStack* state_before, bool needs_patching) + : Instruction(as_ValueType(field->type()->basic_type()), state_before) + , _obj(obj) + , _offset(offset) + , _field(field) + , _explicit_null_check(NULL) + { + set_needs_null_check(!is_static); + set_flag(IsStaticFlag, is_static); + set_flag(NeedsPatchingFlag, needs_patching); + ASSERT_VALUES + // pin of all instructions with memory access + pin(); + } + + // accessors + Value obj() const { return _obj; } + int offset() const { return _offset; } + ciField* field() const { return _field; } + BasicType field_type() const { return _field->type()->basic_type(); } + bool is_static() const { return check_flag(IsStaticFlag); } + NullCheck* explicit_null_check() const { return _explicit_null_check; } + bool needs_patching() const { return check_flag(NeedsPatchingFlag); } + + // Unresolved getstatic and putstatic can cause initialization. + // Technically it occurs at the Constant that materializes the base + // of the static fields but it's simpler to model it here. + bool is_init_point() const { return is_static() && (needs_patching() || !_field->holder()->is_initialized()); } + + // manipulation + + // Under certain circumstances, if a previous NullCheck instruction + // proved the target object non-null, we can eliminate the explicit + // null check and do an implicit one, simply specifying the debug + // information from the NullCheck. This field should only be consulted + // if needs_null_check() is true. + void set_explicit_null_check(NullCheck* check) { _explicit_null_check = check; } + + // generic + virtual bool can_trap() const { return needs_null_check() || needs_patching(); } + virtual void input_values_do(ValueVisitor* f) { f->visit(&_obj); } +}; + + +LEAF(LoadField, AccessField) + public: + // creation + LoadField(Value obj, int offset, ciField* field, bool is_static, + ValueStack* state_before, bool needs_patching) + : AccessField(obj, offset, field, is_static, state_before, needs_patching) + {} + + ciType* declared_type() const; + + // generic + HASHING2(LoadField, !needs_patching() && !field()->is_volatile(), obj()->subst(), offset()) // cannot be eliminated if needs patching or if volatile +}; + + +LEAF(StoreField, AccessField) + private: + Value _value; + + public: + // creation + StoreField(Value obj, int offset, ciField* field, Value value, bool is_static, + ValueStack* state_before, bool needs_patching) + : AccessField(obj, offset, field, is_static, state_before, needs_patching) + , _value(value) + { + set_flag(NeedsWriteBarrierFlag, as_ValueType(field_type())->is_object()); + ASSERT_VALUES + pin(); + } + + // accessors + Value value() const { return _value; } + bool needs_write_barrier() const { return check_flag(NeedsWriteBarrierFlag); } + + // generic + virtual void input_values_do(ValueVisitor* f) { AccessField::input_values_do(f); f->visit(&_value); } +}; + + +BASE(AccessArray, Instruction) + private: + Value _array; + + public: + // creation + AccessArray(ValueType* type, Value array, ValueStack* state_before) + : Instruction(type, state_before) + , _array(array) + { + set_needs_null_check(true); + ASSERT_VALUES + pin(); // instruction with side effect (null exception or range check throwing) + } + + Value array() const { return _array; } + + // generic + virtual bool can_trap() const { return needs_null_check(); } + virtual void input_values_do(ValueVisitor* f) { f->visit(&_array); } +}; + + +LEAF(ArrayLength, AccessArray) + private: + NullCheck* _explicit_null_check; // For explicit null check elimination + + public: + // creation + ArrayLength(Value array, ValueStack* state_before) + : AccessArray(intType, array, state_before) + , _explicit_null_check(NULL) {} + + // accessors + NullCheck* explicit_null_check() const { return _explicit_null_check; } + + // setters + // See LoadField::set_explicit_null_check for documentation + void set_explicit_null_check(NullCheck* check) { _explicit_null_check = check; } + + // generic + HASHING1(ArrayLength, true, array()->subst()) +}; + + +BASE(AccessIndexed, AccessArray) + private: + Value _index; + Value _length; + BasicType _elt_type; + bool _mismatched; + + public: + // creation + AccessIndexed(Value array, Value index, Value length, BasicType elt_type, ValueStack* state_before, bool mismatched) + : AccessArray(as_ValueType(elt_type), array, state_before) + , _index(index) + , _length(length) + , _elt_type(elt_type) + , _mismatched(mismatched) + { + set_flag(Instruction::NeedsRangeCheckFlag, true); + ASSERT_VALUES + } + + // accessors + Value index() const { return _index; } + Value length() const { return _length; } + BasicType elt_type() const { return _elt_type; } + bool mismatched() const { return _mismatched; } + + void clear_length() { _length = NULL; } + // perform elimination of range checks involving constants + bool compute_needs_range_check(); + + // generic + virtual void input_values_do(ValueVisitor* f) { AccessArray::input_values_do(f); f->visit(&_index); if (_length != NULL) f->visit(&_length); } +}; + + +LEAF(LoadIndexed, AccessIndexed) + private: + NullCheck* _explicit_null_check; // For explicit null check elimination + + public: + // creation + LoadIndexed(Value array, Value index, Value length, BasicType elt_type, ValueStack* state_before, bool mismatched = false) + : AccessIndexed(array, index, length, elt_type, state_before, mismatched) + , _explicit_null_check(NULL) {} + + // accessors + NullCheck* explicit_null_check() const { return _explicit_null_check; } + + // setters + // See LoadField::set_explicit_null_check for documentation + void set_explicit_null_check(NullCheck* check) { _explicit_null_check = check; } + + ciType* exact_type() const; + ciType* declared_type() const; + + // generic + HASHING2(LoadIndexed, true, array()->subst(), index()->subst()) +}; + + +LEAF(StoreIndexed, AccessIndexed) + private: + Value _value; + + ciMethod* _profiled_method; + int _profiled_bci; + bool _check_boolean; + + public: + // creation + StoreIndexed(Value array, Value index, Value length, BasicType elt_type, Value value, ValueStack* state_before, + bool check_boolean, bool mismatched = false) + : AccessIndexed(array, index, length, elt_type, state_before, mismatched) + , _value(value), _profiled_method(NULL), _profiled_bci(0), _check_boolean(check_boolean) + { + set_flag(NeedsWriteBarrierFlag, (as_ValueType(elt_type)->is_object())); + set_flag(NeedsStoreCheckFlag, (as_ValueType(elt_type)->is_object())); + ASSERT_VALUES + pin(); + } + + // accessors + Value value() const { return _value; } + bool needs_write_barrier() const { return check_flag(NeedsWriteBarrierFlag); } + bool needs_store_check() const { return check_flag(NeedsStoreCheckFlag); } + bool check_boolean() const { return _check_boolean; } + // Helpers for MethodData* profiling + void set_should_profile(bool value) { set_flag(ProfileMDOFlag, value); } + void set_profiled_method(ciMethod* method) { _profiled_method = method; } + void set_profiled_bci(int bci) { _profiled_bci = bci; } + bool should_profile() const { return check_flag(ProfileMDOFlag); } + ciMethod* profiled_method() const { return _profiled_method; } + int profiled_bci() const { return _profiled_bci; } + // generic + virtual void input_values_do(ValueVisitor* f) { AccessIndexed::input_values_do(f); f->visit(&_value); } +}; + + +LEAF(NegateOp, Instruction) + private: + Value _x; + + public: + // creation + NegateOp(Value x) : Instruction(x->type()->base()), _x(x) { + ASSERT_VALUES + } + + // accessors + Value x() const { return _x; } + + // generic + virtual void input_values_do(ValueVisitor* f) { f->visit(&_x); } +}; + + +BASE(Op2, Instruction) + private: + Bytecodes::Code _op; + Value _x; + Value _y; + + public: + // creation + Op2(ValueType* type, Bytecodes::Code op, Value x, Value y, ValueStack* state_before = NULL) + : Instruction(type, state_before) + , _op(op) + , _x(x) + , _y(y) + { + ASSERT_VALUES + } + + // accessors + Bytecodes::Code op() const { return _op; } + Value x() const { return _x; } + Value y() const { return _y; } + + // manipulators + void swap_operands() { + assert(is_commutative(), "operation must be commutative"); + Value t = _x; _x = _y; _y = t; + } + + // generic + virtual bool is_commutative() const { return false; } + virtual void input_values_do(ValueVisitor* f) { f->visit(&_x); f->visit(&_y); } +}; + + +LEAF(ArithmeticOp, Op2) + public: + // creation + ArithmeticOp(Bytecodes::Code op, Value x, Value y, bool is_strictfp, ValueStack* state_before) + : Op2(x->type()->meet(y->type()), op, x, y, state_before) + { + set_flag(IsStrictfpFlag, is_strictfp); + if (can_trap()) pin(); + } + + // accessors + bool is_strictfp() const { return check_flag(IsStrictfpFlag); } + + // generic + virtual bool is_commutative() const; + virtual bool can_trap() const; + HASHING3(Op2, true, op(), x()->subst(), y()->subst()) +}; + + +LEAF(ShiftOp, Op2) + public: + // creation + ShiftOp(Bytecodes::Code op, Value x, Value s) : Op2(x->type()->base(), op, x, s) {} + + // generic + HASHING3(Op2, true, op(), x()->subst(), y()->subst()) +}; + + +LEAF(LogicOp, Op2) + public: + // creation + LogicOp(Bytecodes::Code op, Value x, Value y) : Op2(x->type()->meet(y->type()), op, x, y) {} + + // generic + virtual bool is_commutative() const; + HASHING3(Op2, true, op(), x()->subst(), y()->subst()) +}; + + +LEAF(CompareOp, Op2) + public: + // creation + CompareOp(Bytecodes::Code op, Value x, Value y, ValueStack* state_before) + : Op2(intType, op, x, y, state_before) + {} + + // generic + HASHING3(Op2, true, op(), x()->subst(), y()->subst()) +}; + + +LEAF(IfOp, Op2) + private: + Value _tval; + Value _fval; + + public: + // creation + IfOp(Value x, Condition cond, Value y, Value tval, Value fval) + : Op2(tval->type()->meet(fval->type()), (Bytecodes::Code)cond, x, y) + , _tval(tval) + , _fval(fval) + { + ASSERT_VALUES + assert(tval->type()->tag() == fval->type()->tag(), "types must match"); + } + + // accessors + virtual bool is_commutative() const; + Bytecodes::Code op() const { ShouldNotCallThis(); return Bytecodes::_illegal; } + Condition cond() const { return (Condition)Op2::op(); } + Value tval() const { return _tval; } + Value fval() const { return _fval; } + + // generic + virtual void input_values_do(ValueVisitor* f) { Op2::input_values_do(f); f->visit(&_tval); f->visit(&_fval); } +}; + + +LEAF(Convert, Instruction) + private: + Bytecodes::Code _op; + Value _value; + + public: + // creation + Convert(Bytecodes::Code op, Value value, ValueType* to_type) : Instruction(to_type), _op(op), _value(value) { + ASSERT_VALUES + } + + // accessors + Bytecodes::Code op() const { return _op; } + Value value() const { return _value; } + + // generic + virtual void input_values_do(ValueVisitor* f) { f->visit(&_value); } + HASHING2(Convert, true, op(), value()->subst()) +}; + + +LEAF(NullCheck, Instruction) + private: + Value _obj; + + public: + // creation + NullCheck(Value obj, ValueStack* state_before) + : Instruction(obj->type()->base(), state_before) + , _obj(obj) + { + ASSERT_VALUES + set_can_trap(true); + assert(_obj->type()->is_object(), "null check must be applied to objects only"); + pin(Instruction::PinExplicitNullCheck); + } + + // accessors + Value obj() const { return _obj; } + + // setters + void set_can_trap(bool can_trap) { set_flag(CanTrapFlag, can_trap); } + + // generic + virtual bool can_trap() const { return check_flag(CanTrapFlag); /* null-check elimination sets to false */ } + virtual void input_values_do(ValueVisitor* f) { f->visit(&_obj); } + HASHING1(NullCheck, true, obj()->subst()) +}; + + +// This node is supposed to cast the type of another node to a more precise +// declared type. +LEAF(TypeCast, Instruction) + private: + ciType* _declared_type; + Value _obj; + + public: + // The type of this node is the same type as the object type (and it might be constant). + TypeCast(ciType* type, Value obj, ValueStack* state_before) + : Instruction(obj->type(), state_before, obj->type()->is_constant()), + _declared_type(type), + _obj(obj) {} + + // accessors + ciType* declared_type() const { return _declared_type; } + Value obj() const { return _obj; } + + // generic + virtual void input_values_do(ValueVisitor* f) { f->visit(&_obj); } +}; + + +BASE(StateSplit, Instruction) + private: + ValueStack* _state; + + protected: + static void substitute(BlockList& list, BlockBegin* old_block, BlockBegin* new_block); + + public: + // creation + StateSplit(ValueType* type, ValueStack* state_before = NULL) + : Instruction(type, state_before) + , _state(NULL) + { + pin(PinStateSplitConstructor); + } + + // accessors + ValueStack* state() const { return _state; } + IRScope* scope() const; // the state's scope + + // manipulation + void set_state(ValueStack* state) { assert(_state == NULL, "overwriting existing state"); check_state(state); _state = state; } + + // generic + virtual void input_values_do(ValueVisitor* f) { /* no values */ } + virtual void state_values_do(ValueVisitor* f); +}; + + +LEAF(Invoke, StateSplit) + private: + Bytecodes::Code _code; + Value _recv; + Values* _args; + BasicTypeList* _signature; + int _vtable_index; + ciMethod* _target; + + public: + // creation + Invoke(Bytecodes::Code code, ValueType* result_type, Value recv, Values* args, + int vtable_index, ciMethod* target, ValueStack* state_before); + + // accessors + Bytecodes::Code code() const { return _code; } + Value receiver() const { return _recv; } + bool has_receiver() const { return receiver() != NULL; } + int number_of_arguments() const { return _args->length(); } + Value argument_at(int i) const { return _args->at(i); } + int vtable_index() const { return _vtable_index; } + BasicTypeList* signature() const { return _signature; } + ciMethod* target() const { return _target; } + + ciType* declared_type() const; + + // Returns false if target is not loaded + bool target_is_final() const { return check_flag(TargetIsFinalFlag); } + bool target_is_loaded() const { return check_flag(TargetIsLoadedFlag); } + // Returns false if target is not loaded + bool target_is_strictfp() const { return check_flag(TargetIsStrictfpFlag); } + + // JSR 292 support + bool is_invokedynamic() const { return code() == Bytecodes::_invokedynamic; } + bool is_method_handle_intrinsic() const { return target()->is_method_handle_intrinsic(); } + + virtual bool needs_exception_state() const { return false; } + + // generic + virtual bool can_trap() const { return true; } + virtual void input_values_do(ValueVisitor* f) { + StateSplit::input_values_do(f); + if (has_receiver()) f->visit(&_recv); + for (int i = 0; i < _args->length(); i++) f->visit(_args->adr_at(i)); + } + virtual void state_values_do(ValueVisitor *f); +}; + + +LEAF(NewInstance, StateSplit) + private: + ciInstanceKlass* _klass; + bool _is_unresolved; + + public: + // creation + NewInstance(ciInstanceKlass* klass, ValueStack* state_before, bool is_unresolved) + : StateSplit(instanceType, state_before) + , _klass(klass), _is_unresolved(is_unresolved) + {} + + // accessors + ciInstanceKlass* klass() const { return _klass; } + bool is_unresolved() const { return _is_unresolved; } + + virtual bool needs_exception_state() const { return false; } + + // generic + virtual bool can_trap() const { return true; } + ciType* exact_type() const; + ciType* declared_type() const; +}; + + +BASE(NewArray, StateSplit) + private: + Value _length; + + public: + // creation + NewArray(Value length, ValueStack* state_before) + : StateSplit(objectType, state_before) + , _length(length) + { + // Do not ASSERT_VALUES since length is NULL for NewMultiArray + } + + // accessors + Value length() const { return _length; } + + virtual bool needs_exception_state() const { return false; } + + ciType* exact_type() const { return NULL; } + ciType* declared_type() const; + + // generic + virtual bool can_trap() const { return true; } + virtual void input_values_do(ValueVisitor* f) { StateSplit::input_values_do(f); f->visit(&_length); } +}; + + +LEAF(NewTypeArray, NewArray) + private: + BasicType _elt_type; + + public: + // creation + NewTypeArray(Value length, BasicType elt_type, ValueStack* state_before) + : NewArray(length, state_before) + , _elt_type(elt_type) + {} + + // accessors + BasicType elt_type() const { return _elt_type; } + ciType* exact_type() const; +}; + + +LEAF(NewObjectArray, NewArray) + private: + ciKlass* _klass; + + public: + // creation + NewObjectArray(ciKlass* klass, Value length, ValueStack* state_before) : NewArray(length, state_before), _klass(klass) {} + + // accessors + ciKlass* klass() const { return _klass; } + ciType* exact_type() const; +}; + + +LEAF(NewMultiArray, NewArray) + private: + ciKlass* _klass; + Values* _dims; + + public: + // creation + NewMultiArray(ciKlass* klass, Values* dims, ValueStack* state_before) : NewArray(NULL, state_before), _klass(klass), _dims(dims) { + ASSERT_VALUES + } + + // accessors + ciKlass* klass() const { return _klass; } + Values* dims() const { return _dims; } + int rank() const { return dims()->length(); } + + // generic + virtual void input_values_do(ValueVisitor* f) { + // NOTE: we do not call NewArray::input_values_do since "length" + // is meaningless for a multi-dimensional array; passing the + // zeroth element down to NewArray as its length is a bad idea + // since there will be a copy in the "dims" array which doesn't + // get updated, and the value must not be traversed twice. Was bug + // - kbr 4/10/2001 + StateSplit::input_values_do(f); + for (int i = 0; i < _dims->length(); i++) f->visit(_dims->adr_at(i)); + } +}; + + +BASE(TypeCheck, StateSplit) + private: + ciKlass* _klass; + Value _obj; + + ciMethod* _profiled_method; + int _profiled_bci; + + public: + // creation + TypeCheck(ciKlass* klass, Value obj, ValueType* type, ValueStack* state_before) + : StateSplit(type, state_before), _klass(klass), _obj(obj), + _profiled_method(NULL), _profiled_bci(0) { + ASSERT_VALUES + set_direct_compare(false); + } + + // accessors + ciKlass* klass() const { return _klass; } + Value obj() const { return _obj; } + bool is_loaded() const { return klass() != NULL; } + bool direct_compare() const { return check_flag(DirectCompareFlag); } + + // manipulation + void set_direct_compare(bool flag) { set_flag(DirectCompareFlag, flag); } + + // generic + virtual bool can_trap() const { return true; } + virtual void input_values_do(ValueVisitor* f) { StateSplit::input_values_do(f); f->visit(&_obj); } + + // Helpers for MethodData* profiling + void set_should_profile(bool value) { set_flag(ProfileMDOFlag, value); } + void set_profiled_method(ciMethod* method) { _profiled_method = method; } + void set_profiled_bci(int bci) { _profiled_bci = bci; } + bool should_profile() const { return check_flag(ProfileMDOFlag); } + ciMethod* profiled_method() const { return _profiled_method; } + int profiled_bci() const { return _profiled_bci; } +}; + + +LEAF(CheckCast, TypeCheck) + public: + // creation + CheckCast(ciKlass* klass, Value obj, ValueStack* state_before) + : TypeCheck(klass, obj, objectType, state_before) {} + + void set_incompatible_class_change_check() { + set_flag(ThrowIncompatibleClassChangeErrorFlag, true); + } + bool is_incompatible_class_change_check() const { + return check_flag(ThrowIncompatibleClassChangeErrorFlag); + } + void set_invokespecial_receiver_check() { + set_flag(InvokeSpecialReceiverCheckFlag, true); + } + bool is_invokespecial_receiver_check() const { + return check_flag(InvokeSpecialReceiverCheckFlag); + } + + virtual bool needs_exception_state() const { + return !is_invokespecial_receiver_check(); + } + + ciType* declared_type() const; +}; + + +LEAF(InstanceOf, TypeCheck) + public: + // creation + InstanceOf(ciKlass* klass, Value obj, ValueStack* state_before) : TypeCheck(klass, obj, intType, state_before) {} + + virtual bool needs_exception_state() const { return false; } +}; + + +BASE(AccessMonitor, StateSplit) + private: + Value _obj; + int _monitor_no; + + public: + // creation + AccessMonitor(Value obj, int monitor_no, ValueStack* state_before = NULL) + : StateSplit(illegalType, state_before) + , _obj(obj) + , _monitor_no(monitor_no) + { + set_needs_null_check(true); + ASSERT_VALUES + } + + // accessors + Value obj() const { return _obj; } + int monitor_no() const { return _monitor_no; } + + // generic + virtual void input_values_do(ValueVisitor* f) { StateSplit::input_values_do(f); f->visit(&_obj); } +}; + + +LEAF(MonitorEnter, AccessMonitor) + public: + // creation + MonitorEnter(Value obj, int monitor_no, ValueStack* state_before) + : AccessMonitor(obj, monitor_no, state_before) + { + ASSERT_VALUES + } + + // generic + virtual bool can_trap() const { return true; } +}; + + +LEAF(MonitorExit, AccessMonitor) + public: + // creation + MonitorExit(Value obj, int monitor_no) + : AccessMonitor(obj, monitor_no, NULL) + { + ASSERT_VALUES + } +}; + + +LEAF(Intrinsic, StateSplit) + private: + vmIntrinsics::ID _id; + Values* _args; + Value _recv; + ArgsNonNullState _nonnull_state; + + public: + // preserves_state can be set to true for Intrinsics + // which are guaranteed to preserve register state across any slow + // cases; setting it to true does not mean that the Intrinsic can + // not trap, only that if we continue execution in the same basic + // block after the Intrinsic, all of the registers are intact. This + // allows load elimination and common expression elimination to be + // performed across the Intrinsic. The default value is false. + Intrinsic(ValueType* type, + vmIntrinsics::ID id, + Values* args, + bool has_receiver, + ValueStack* state_before, + bool preserves_state, + bool cantrap = true) + : StateSplit(type, state_before) + , _id(id) + , _args(args) + , _recv(NULL) + { + assert(args != NULL, "args must exist"); + ASSERT_VALUES + set_flag(PreservesStateFlag, preserves_state); + set_flag(CanTrapFlag, cantrap); + if (has_receiver) { + _recv = argument_at(0); + } + set_needs_null_check(has_receiver); + + // some intrinsics can't trap, so don't force them to be pinned + if (!can_trap() && !vmIntrinsics::should_be_pinned(_id)) { + unpin(PinStateSplitConstructor); + } + } + + // accessors + vmIntrinsics::ID id() const { return _id; } + int number_of_arguments() const { return _args->length(); } + Value argument_at(int i) const { return _args->at(i); } + + bool has_receiver() const { return (_recv != NULL); } + Value receiver() const { assert(has_receiver(), "must have receiver"); return _recv; } + bool preserves_state() const { return check_flag(PreservesStateFlag); } + + bool arg_needs_null_check(int i) const { + return _nonnull_state.arg_needs_null_check(i); + } + + void set_arg_needs_null_check(int i, bool check) { + _nonnull_state.set_arg_needs_null_check(i, check); + } + + // generic + virtual bool can_trap() const { return check_flag(CanTrapFlag); } + virtual void input_values_do(ValueVisitor* f) { + StateSplit::input_values_do(f); + for (int i = 0; i < _args->length(); i++) f->visit(_args->adr_at(i)); + } +}; + + +class LIR_List; + +LEAF(BlockBegin, StateSplit) + private: + int _block_id; // the unique block id + int _bci; // start-bci of block + int _depth_first_number; // number of this block in a depth-first ordering + int _linear_scan_number; // number of this block in linear-scan ordering + int _dominator_depth; + int _loop_depth; // the loop nesting level of this block + int _loop_index; // number of the innermost loop of this block + int _flags; // the flags associated with this block + + // fields used by BlockListBuilder + int _total_preds; // number of predecessors found by BlockListBuilder + ResourceBitMap _stores_to_locals; // bit is set when a local variable is stored in the block + + // SSA specific fields: (factor out later) + BlockList _successors; // the successors of this block + BlockList _predecessors; // the predecessors of this block + BlockList _dominates; // list of blocks that are dominated by this block + BlockBegin* _dominator; // the dominator of this block + // SSA specific ends + BlockEnd* _end; // the last instruction of this block + BlockList _exception_handlers; // the exception handlers potentially invoked by this block + ValueStackStack* _exception_states; // only for xhandler entries: states of all instructions that have an edge to this xhandler + int _exception_handler_pco; // if this block is the start of an exception handler, + // this records the PC offset in the assembly code of the + // first instruction in this block + Label _label; // the label associated with this block + LIR_List* _lir; // the low level intermediate representation for this block + + ResourceBitMap _live_in; // set of live LIR_Opr registers at entry to this block + ResourceBitMap _live_out; // set of live LIR_Opr registers at exit from this block + ResourceBitMap _live_gen; // set of registers used before any redefinition in this block + ResourceBitMap _live_kill; // set of registers defined in this block + + ResourceBitMap _fpu_register_usage; + intArray* _fpu_stack_state; // For x86 FPU code generation with UseLinearScan + int _first_lir_instruction_id; // ID of first LIR instruction in this block + int _last_lir_instruction_id; // ID of last LIR instruction in this block + + void iterate_preorder (boolArray& mark, BlockClosure* closure); + void iterate_postorder(boolArray& mark, BlockClosure* closure); + + friend class SuxAndWeightAdjuster; + + public: + void* operator new(size_t size) throw() { + Compilation* c = Compilation::current(); + void* res = c->arena()->Amalloc(size); + ((BlockBegin*)res)->_id = c->get_next_id(); + ((BlockBegin*)res)->_block_id = c->get_next_block_id(); + return res; + } + + // initialization/counting + static int number_of_blocks() { + return Compilation::current()->number_of_blocks(); + } + + // creation + BlockBegin(int bci) + : StateSplit(illegalType) + , _bci(bci) + , _depth_first_number(-1) + , _linear_scan_number(-1) + , _loop_depth(0) + , _flags(0) + , _dominator_depth(-1) + , _dominator(NULL) + , _end(NULL) + , _predecessors(2) + , _successors(2) + , _dominates(2) + , _exception_handlers(1) + , _exception_states(NULL) + , _exception_handler_pco(-1) + , _lir(NULL) + , _loop_index(-1) + , _live_in() + , _live_out() + , _live_gen() + , _live_kill() + , _fpu_register_usage() + , _fpu_stack_state(NULL) + , _first_lir_instruction_id(-1) + , _last_lir_instruction_id(-1) + , _total_preds(0) + , _stores_to_locals() + { + _block = this; +#ifndef PRODUCT + set_printable_bci(bci); +#endif + } + + // accessors + int block_id() const { return _block_id; } + int bci() const { return _bci; } + BlockList* successors() { return &_successors; } + BlockList* dominates() { return &_dominates; } + BlockBegin* dominator() const { return _dominator; } + int loop_depth() const { return _loop_depth; } + int dominator_depth() const { return _dominator_depth; } + int depth_first_number() const { return _depth_first_number; } + int linear_scan_number() const { return _linear_scan_number; } + BlockEnd* end() const { return _end; } + Label* label() { return &_label; } + LIR_List* lir() const { return _lir; } + int exception_handler_pco() const { return _exception_handler_pco; } + ResourceBitMap& live_in() { return _live_in; } + ResourceBitMap& live_out() { return _live_out; } + ResourceBitMap& live_gen() { return _live_gen; } + ResourceBitMap& live_kill() { return _live_kill; } + ResourceBitMap& fpu_register_usage() { return _fpu_register_usage; } + intArray* fpu_stack_state() const { return _fpu_stack_state; } + int first_lir_instruction_id() const { return _first_lir_instruction_id; } + int last_lir_instruction_id() const { return _last_lir_instruction_id; } + int total_preds() const { return _total_preds; } + BitMap& stores_to_locals() { return _stores_to_locals; } + + // manipulation + void set_dominator(BlockBegin* dom) { _dominator = dom; } + void set_loop_depth(int d) { _loop_depth = d; } + void set_dominator_depth(int d) { _dominator_depth = d; } + void set_depth_first_number(int dfn) { _depth_first_number = dfn; } + void set_linear_scan_number(int lsn) { _linear_scan_number = lsn; } + void set_end(BlockEnd* end); + void clear_end(); + void disconnect_from_graph(); + static void disconnect_edge(BlockBegin* from, BlockBegin* to); + BlockBegin* insert_block_between(BlockBegin* sux); + void substitute_sux(BlockBegin* old_sux, BlockBegin* new_sux); + void set_lir(LIR_List* lir) { _lir = lir; } + void set_exception_handler_pco(int pco) { _exception_handler_pco = pco; } + void set_live_in (const ResourceBitMap& map) { _live_in = map; } + void set_live_out (const ResourceBitMap& map) { _live_out = map; } + void set_live_gen (const ResourceBitMap& map) { _live_gen = map; } + void set_live_kill(const ResourceBitMap& map) { _live_kill = map; } + void set_fpu_register_usage(const ResourceBitMap& map) { _fpu_register_usage = map; } + void set_fpu_stack_state(intArray* state) { _fpu_stack_state = state; } + void set_first_lir_instruction_id(int id) { _first_lir_instruction_id = id; } + void set_last_lir_instruction_id(int id) { _last_lir_instruction_id = id; } + void increment_total_preds(int n = 1) { _total_preds += n; } + void init_stores_to_locals(int locals_count) { _stores_to_locals.initialize(locals_count); } + + // generic + virtual void state_values_do(ValueVisitor* f); + + // successors and predecessors + int number_of_sux() const; + BlockBegin* sux_at(int i) const; + void add_successor(BlockBegin* sux); + void remove_successor(BlockBegin* pred); + bool is_successor(BlockBegin* sux) const { return _successors.contains(sux); } + + void add_predecessor(BlockBegin* pred); + void remove_predecessor(BlockBegin* pred); + bool is_predecessor(BlockBegin* pred) const { return _predecessors.contains(pred); } + int number_of_preds() const { return _predecessors.length(); } + BlockBegin* pred_at(int i) const { return _predecessors.at(i); } + + // exception handlers potentially invoked by this block + void add_exception_handler(BlockBegin* b); + bool is_exception_handler(BlockBegin* b) const { return _exception_handlers.contains(b); } + int number_of_exception_handlers() const { return _exception_handlers.length(); } + BlockBegin* exception_handler_at(int i) const { return _exception_handlers.at(i); } + + // states of the instructions that have an edge to this exception handler + int number_of_exception_states() { assert(is_set(exception_entry_flag), "only for xhandlers"); return _exception_states == NULL ? 0 : _exception_states->length(); } + ValueStack* exception_state_at(int idx) const { assert(is_set(exception_entry_flag), "only for xhandlers"); return _exception_states->at(idx); } + int add_exception_state(ValueStack* state); + + // flags + enum Flag { + no_flag = 0, + std_entry_flag = 1 << 0, + osr_entry_flag = 1 << 1, + exception_entry_flag = 1 << 2, + subroutine_entry_flag = 1 << 3, + backward_branch_target_flag = 1 << 4, + is_on_work_list_flag = 1 << 5, + was_visited_flag = 1 << 6, + parser_loop_header_flag = 1 << 7, // set by parser to identify blocks where phi functions can not be created on demand + critical_edge_split_flag = 1 << 8, // set for all blocks that are introduced when critical edges are split + linear_scan_loop_header_flag = 1 << 9, // set during loop-detection for LinearScan + linear_scan_loop_end_flag = 1 << 10, // set during loop-detection for LinearScan + donot_eliminate_range_checks = 1 << 11 // Should be try to eliminate range checks in this block + }; + + void set(Flag f) { _flags |= f; } + void clear(Flag f) { _flags &= ~f; } + bool is_set(Flag f) const { return (_flags & f) != 0; } + bool is_entry_block() const { + const int entry_mask = std_entry_flag | osr_entry_flag | exception_entry_flag; + return (_flags & entry_mask) != 0; + } + + // iteration + void iterate_preorder (BlockClosure* closure); + void iterate_postorder (BlockClosure* closure); + + void block_values_do(ValueVisitor* f); + + // loops + void set_loop_index(int ix) { _loop_index = ix; } + int loop_index() const { return _loop_index; } + + // merging + bool try_merge(ValueStack* state); // try to merge states at block begin + void merge(ValueStack* state) { bool b = try_merge(state); assert(b, "merge failed"); } + + // debugging + void print_block() PRODUCT_RETURN; + void print_block(InstructionPrinter& ip, bool live_only = false) PRODUCT_RETURN; +}; + + +BASE(BlockEnd, StateSplit) + private: + BlockList* _sux; + + protected: + BlockList* sux() const { return _sux; } + + void set_sux(BlockList* sux) { +#ifdef ASSERT + assert(sux != NULL, "sux must exist"); + for (int i = sux->length() - 1; i >= 0; i--) assert(sux->at(i) != NULL, "sux must exist"); +#endif + _sux = sux; + } + + public: + // creation + BlockEnd(ValueType* type, ValueStack* state_before, bool is_safepoint) + : StateSplit(type, state_before) + , _sux(NULL) + { + set_flag(IsSafepointFlag, is_safepoint); + } + + // accessors + bool is_safepoint() const { return check_flag(IsSafepointFlag); } + // For compatibility with old code, for new code use block() + BlockBegin* begin() const { return _block; } + + // manipulation + void set_begin(BlockBegin* begin); + + // successors + int number_of_sux() const { return _sux != NULL ? _sux->length() : 0; } + BlockBegin* sux_at(int i) const { return _sux->at(i); } + BlockBegin* default_sux() const { return sux_at(number_of_sux() - 1); } + BlockBegin** addr_sux_at(int i) const { return _sux->adr_at(i); } + int sux_index(BlockBegin* sux) const { return _sux->find(sux); } + void substitute_sux(BlockBegin* old_sux, BlockBegin* new_sux); +}; + + +LEAF(Goto, BlockEnd) + public: + enum Direction { + none, // Just a regular goto + taken, not_taken // Goto produced from If + }; + private: + ciMethod* _profiled_method; + int _profiled_bci; + Direction _direction; + public: + // creation + Goto(BlockBegin* sux, ValueStack* state_before, bool is_safepoint = false) + : BlockEnd(illegalType, state_before, is_safepoint) + , _direction(none) + , _profiled_method(NULL) + , _profiled_bci(0) { + BlockList* s = new BlockList(1); + s->append(sux); + set_sux(s); + } + + Goto(BlockBegin* sux, bool is_safepoint) : BlockEnd(illegalType, NULL, is_safepoint) + , _direction(none) + , _profiled_method(NULL) + , _profiled_bci(0) { + BlockList* s = new BlockList(1); + s->append(sux); + set_sux(s); + } + + bool should_profile() const { return check_flag(ProfileMDOFlag); } + ciMethod* profiled_method() const { return _profiled_method; } // set only for profiled branches + int profiled_bci() const { return _profiled_bci; } + Direction direction() const { return _direction; } + + void set_should_profile(bool value) { set_flag(ProfileMDOFlag, value); } + void set_profiled_method(ciMethod* method) { _profiled_method = method; } + void set_profiled_bci(int bci) { _profiled_bci = bci; } + void set_direction(Direction d) { _direction = d; } +}; + +#ifdef ASSERT +LEAF(Assert, Instruction) + private: + Value _x; + Condition _cond; + Value _y; + char *_message; + + public: + // creation + // unordered_is_true is valid for float/double compares only + Assert(Value x, Condition cond, bool unordered_is_true, Value y); + + // accessors + Value x() const { return _x; } + Condition cond() const { return _cond; } + bool unordered_is_true() const { return check_flag(UnorderedIsTrueFlag); } + Value y() const { return _y; } + const char *message() const { return _message; } + + // generic + virtual void input_values_do(ValueVisitor* f) { f->visit(&_x); f->visit(&_y); } +}; +#endif + +LEAF(RangeCheckPredicate, StateSplit) + private: + Value _x; + Condition _cond; + Value _y; + + void check_state(); + + public: + // creation + // unordered_is_true is valid for float/double compares only + RangeCheckPredicate(Value x, Condition cond, bool unordered_is_true, Value y, ValueStack* state) : StateSplit(illegalType) + , _x(x) + , _cond(cond) + , _y(y) + { + ASSERT_VALUES + set_flag(UnorderedIsTrueFlag, unordered_is_true); + assert(x->type()->tag() == y->type()->tag(), "types must match"); + this->set_state(state); + check_state(); + } + + // Always deoptimize + RangeCheckPredicate(ValueStack* state) : StateSplit(illegalType) + { + this->set_state(state); + _x = _y = NULL; + check_state(); + } + + // accessors + Value x() const { return _x; } + Condition cond() const { return _cond; } + bool unordered_is_true() const { return check_flag(UnorderedIsTrueFlag); } + Value y() const { return _y; } + + void always_fail() { _x = _y = NULL; } + + // generic + virtual void input_values_do(ValueVisitor* f) { StateSplit::input_values_do(f); f->visit(&_x); f->visit(&_y); } + HASHING3(RangeCheckPredicate, true, x()->subst(), y()->subst(), cond()) +}; + +LEAF(If, BlockEnd) + private: + Value _x; + Condition _cond; + Value _y; + ciMethod* _profiled_method; + int _profiled_bci; // Canonicalizer may alter bci of If node + bool _swapped; // Is the order reversed with respect to the original If in the + // bytecode stream? + public: + // creation + // unordered_is_true is valid for float/double compares only + If(Value x, Condition cond, bool unordered_is_true, Value y, BlockBegin* tsux, BlockBegin* fsux, ValueStack* state_before, bool is_safepoint) + : BlockEnd(illegalType, state_before, is_safepoint) + , _x(x) + , _cond(cond) + , _y(y) + , _profiled_method(NULL) + , _profiled_bci(0) + , _swapped(false) + { + ASSERT_VALUES + set_flag(UnorderedIsTrueFlag, unordered_is_true); + assert(x->type()->tag() == y->type()->tag(), "types must match"); + BlockList* s = new BlockList(2); + s->append(tsux); + s->append(fsux); + set_sux(s); + } + + // accessors + Value x() const { return _x; } + Condition cond() const { return _cond; } + bool unordered_is_true() const { return check_flag(UnorderedIsTrueFlag); } + Value y() const { return _y; } + BlockBegin* sux_for(bool is_true) const { return sux_at(is_true ? 0 : 1); } + BlockBegin* tsux() const { return sux_for(true); } + BlockBegin* fsux() const { return sux_for(false); } + BlockBegin* usux() const { return sux_for(unordered_is_true()); } + bool should_profile() const { return check_flag(ProfileMDOFlag); } + ciMethod* profiled_method() const { return _profiled_method; } // set only for profiled branches + int profiled_bci() const { return _profiled_bci; } // set for profiled branches and tiered + bool is_swapped() const { return _swapped; } + + // manipulation + void swap_operands() { + Value t = _x; _x = _y; _y = t; + _cond = mirror(_cond); + } + + void swap_sux() { + assert(number_of_sux() == 2, "wrong number of successors"); + BlockList* s = sux(); + BlockBegin* t = s->at(0); s->at_put(0, s->at(1)); s->at_put(1, t); + _cond = negate(_cond); + set_flag(UnorderedIsTrueFlag, !check_flag(UnorderedIsTrueFlag)); + } + + void set_should_profile(bool value) { set_flag(ProfileMDOFlag, value); } + void set_profiled_method(ciMethod* method) { _profiled_method = method; } + void set_profiled_bci(int bci) { _profiled_bci = bci; } + void set_swapped(bool value) { _swapped = value; } + // generic + virtual void input_values_do(ValueVisitor* f) { BlockEnd::input_values_do(f); f->visit(&_x); f->visit(&_y); } +}; + + +LEAF(IfInstanceOf, BlockEnd) + private: + ciKlass* _klass; + Value _obj; + bool _test_is_instance; // jump if instance + int _instanceof_bci; + + public: + IfInstanceOf(ciKlass* klass, Value obj, bool test_is_instance, int instanceof_bci, BlockBegin* tsux, BlockBegin* fsux) + : BlockEnd(illegalType, NULL, false) // temporary set to false + , _klass(klass) + , _obj(obj) + , _test_is_instance(test_is_instance) + , _instanceof_bci(instanceof_bci) + { + ASSERT_VALUES + assert(instanceof_bci >= 0, "illegal bci"); + BlockList* s = new BlockList(2); + s->append(tsux); + s->append(fsux); + set_sux(s); + } + + // accessors + // + // Note 1: If test_is_instance() is true, IfInstanceOf tests if obj *is* an + // instance of klass; otherwise it tests if it is *not* and instance + // of klass. + // + // Note 2: IfInstanceOf instructions are created by combining an InstanceOf + // and an If instruction. The IfInstanceOf bci() corresponds to the + // bci that the If would have had; the (this->) instanceof_bci() is + // the bci of the original InstanceOf instruction. + ciKlass* klass() const { return _klass; } + Value obj() const { return _obj; } + int instanceof_bci() const { return _instanceof_bci; } + bool test_is_instance() const { return _test_is_instance; } + BlockBegin* sux_for(bool is_true) const { return sux_at(is_true ? 0 : 1); } + BlockBegin* tsux() const { return sux_for(true); } + BlockBegin* fsux() const { return sux_for(false); } + + // manipulation + void swap_sux() { + assert(number_of_sux() == 2, "wrong number of successors"); + BlockList* s = sux(); + BlockBegin* t = s->at(0); s->at_put(0, s->at(1)); s->at_put(1, t); + _test_is_instance = !_test_is_instance; + } + + // generic + virtual void input_values_do(ValueVisitor* f) { BlockEnd::input_values_do(f); f->visit(&_obj); } +}; + + +BASE(Switch, BlockEnd) + private: + Value _tag; + + public: + // creation + Switch(Value tag, BlockList* sux, ValueStack* state_before, bool is_safepoint) + : BlockEnd(illegalType, state_before, is_safepoint) + , _tag(tag) { + ASSERT_VALUES + set_sux(sux); + } + + // accessors + Value tag() const { return _tag; } + int length() const { return number_of_sux() - 1; } + + virtual bool needs_exception_state() const { return false; } + + // generic + virtual void input_values_do(ValueVisitor* f) { BlockEnd::input_values_do(f); f->visit(&_tag); } +}; + + +LEAF(TableSwitch, Switch) + private: + int _lo_key; + + public: + // creation + TableSwitch(Value tag, BlockList* sux, int lo_key, ValueStack* state_before, bool is_safepoint) + : Switch(tag, sux, state_before, is_safepoint) + , _lo_key(lo_key) {} + + // accessors + int lo_key() const { return _lo_key; } + int hi_key() const { return _lo_key + length() - 1; } +}; + + +LEAF(LookupSwitch, Switch) + private: + intArray* _keys; + + public: + // creation + LookupSwitch(Value tag, BlockList* sux, intArray* keys, ValueStack* state_before, bool is_safepoint) + : Switch(tag, sux, state_before, is_safepoint) + , _keys(keys) { + assert(keys != NULL, "keys must exist"); + assert(keys->length() == length(), "sux & keys have incompatible lengths"); + } + + // accessors + int key_at(int i) const { return _keys->at(i); } +}; + + +LEAF(Return, BlockEnd) + private: + Value _result; + + public: + // creation + Return(Value result) : + BlockEnd(result == NULL ? voidType : result->type()->base(), NULL, true), + _result(result) {} + + // accessors + Value result() const { return _result; } + bool has_result() const { return result() != NULL; } + + // generic + virtual void input_values_do(ValueVisitor* f) { + BlockEnd::input_values_do(f); + if (has_result()) f->visit(&_result); + } +}; + + +LEAF(Throw, BlockEnd) + private: + Value _exception; + + public: + // creation + Throw(Value exception, ValueStack* state_before) : BlockEnd(illegalType, state_before, true), _exception(exception) { + ASSERT_VALUES + } + + // accessors + Value exception() const { return _exception; } + + // generic + virtual bool can_trap() const { return true; } + virtual void input_values_do(ValueVisitor* f) { BlockEnd::input_values_do(f); f->visit(&_exception); } +}; + + +LEAF(Base, BlockEnd) + public: + // creation + Base(BlockBegin* std_entry, BlockBegin* osr_entry) : BlockEnd(illegalType, NULL, false) { + assert(std_entry->is_set(BlockBegin::std_entry_flag), "std entry must be flagged"); + assert(osr_entry == NULL || osr_entry->is_set(BlockBegin::osr_entry_flag), "osr entry must be flagged"); + BlockList* s = new BlockList(2); + if (osr_entry != NULL) s->append(osr_entry); + s->append(std_entry); // must be default sux! + set_sux(s); + } + + // accessors + BlockBegin* std_entry() const { return default_sux(); } + BlockBegin* osr_entry() const { return number_of_sux() < 2 ? NULL : sux_at(0); } +}; + + +LEAF(OsrEntry, Instruction) + public: + // creation +#ifdef _LP64 + OsrEntry() : Instruction(longType) { pin(); } +#else + OsrEntry() : Instruction(intType) { pin(); } +#endif + + // generic + virtual void input_values_do(ValueVisitor* f) { } +}; + + +// Models the incoming exception at a catch site +LEAF(ExceptionObject, Instruction) + public: + // creation + ExceptionObject() : Instruction(objectType) { + pin(); + } + + // generic + virtual void input_values_do(ValueVisitor* f) { } +}; + + +// Models needed rounding for floating-point values on Intel. +// Currently only used to represent rounding of double-precision +// values stored into local variables, but could be used to model +// intermediate rounding of single-precision values as well. +LEAF(RoundFP, Instruction) + private: + Value _input; // floating-point value to be rounded + + public: + RoundFP(Value input) + : Instruction(input->type()) // Note: should not be used for constants + , _input(input) + { + ASSERT_VALUES + } + + // accessors + Value input() const { return _input; } + + // generic + virtual void input_values_do(ValueVisitor* f) { f->visit(&_input); } +}; + + +BASE(UnsafeOp, Instruction) + private: + BasicType _basic_type; // ValueType can not express byte-sized integers + + protected: + // creation + UnsafeOp(BasicType basic_type, bool is_put) + : Instruction(is_put ? voidType : as_ValueType(basic_type)) + , _basic_type(basic_type) + { + //Note: Unsafe ops are not not guaranteed to throw NPE. + // Convservatively, Unsafe operations must be pinned though we could be + // looser about this if we wanted to.. + pin(); + } + + public: + // accessors + BasicType basic_type() { return _basic_type; } + + // generic + virtual void input_values_do(ValueVisitor* f) { } +}; + + +BASE(UnsafeRawOp, UnsafeOp) + private: + Value _base; // Base address (a Java long) + Value _index; // Index if computed by optimizer; initialized to NULL + int _log2_scale; // Scale factor: 0, 1, 2, or 3. + // Indicates log2 of number of bytes (1, 2, 4, or 8) + // to scale index by. + + protected: + UnsafeRawOp(BasicType basic_type, Value addr, bool is_put) + : UnsafeOp(basic_type, is_put) + , _base(addr) + , _index(NULL) + , _log2_scale(0) + { + // Can not use ASSERT_VALUES because index may be NULL + assert(addr != NULL && addr->type()->is_long(), "just checking"); + } + + UnsafeRawOp(BasicType basic_type, Value base, Value index, int log2_scale, bool is_put) + : UnsafeOp(basic_type, is_put) + , _base(base) + , _index(index) + , _log2_scale(log2_scale) + { + } + + public: + // accessors + Value base() { return _base; } + Value index() { return _index; } + bool has_index() { return (_index != NULL); } + int log2_scale() { return _log2_scale; } + + // setters + void set_base (Value base) { _base = base; } + void set_index(Value index) { _index = index; } + void set_log2_scale(int log2_scale) { _log2_scale = log2_scale; } + + // generic + virtual void input_values_do(ValueVisitor* f) { UnsafeOp::input_values_do(f); + f->visit(&_base); + if (has_index()) f->visit(&_index); } +}; + + +LEAF(UnsafeGetRaw, UnsafeRawOp) + private: + bool _may_be_unaligned, _is_wide; // For OSREntry + + public: + UnsafeGetRaw(BasicType basic_type, Value addr, bool may_be_unaligned, bool is_wide = false) + : UnsafeRawOp(basic_type, addr, false) { + _may_be_unaligned = may_be_unaligned; + _is_wide = is_wide; + } + + UnsafeGetRaw(BasicType basic_type, Value base, Value index, int log2_scale, bool may_be_unaligned, bool is_wide = false) + : UnsafeRawOp(basic_type, base, index, log2_scale, false) { + _may_be_unaligned = may_be_unaligned; + _is_wide = is_wide; + } + + bool may_be_unaligned() { return _may_be_unaligned; } + bool is_wide() { return _is_wide; } +}; + + +LEAF(UnsafePutRaw, UnsafeRawOp) + private: + Value _value; // Value to be stored + + public: + UnsafePutRaw(BasicType basic_type, Value addr, Value value) + : UnsafeRawOp(basic_type, addr, true) + , _value(value) + { + assert(value != NULL, "just checking"); + ASSERT_VALUES + } + + UnsafePutRaw(BasicType basic_type, Value base, Value index, int log2_scale, Value value) + : UnsafeRawOp(basic_type, base, index, log2_scale, true) + , _value(value) + { + assert(value != NULL, "just checking"); + ASSERT_VALUES + } + + // accessors + Value value() { return _value; } + + // generic + virtual void input_values_do(ValueVisitor* f) { UnsafeRawOp::input_values_do(f); + f->visit(&_value); } +}; + + +BASE(UnsafeObjectOp, UnsafeOp) + private: + Value _object; // Object to be fetched from or mutated + Value _offset; // Offset within object + bool _is_volatile; // true if volatile - dl/JSR166 + public: + UnsafeObjectOp(BasicType basic_type, Value object, Value offset, bool is_put, bool is_volatile) + : UnsafeOp(basic_type, is_put), _object(object), _offset(offset), _is_volatile(is_volatile) + { + } + + // accessors + Value object() { return _object; } + Value offset() { return _offset; } + bool is_volatile() { return _is_volatile; } + // generic + virtual void input_values_do(ValueVisitor* f) { UnsafeOp::input_values_do(f); + f->visit(&_object); + f->visit(&_offset); } +}; + + +LEAF(UnsafeGetObject, UnsafeObjectOp) + public: + UnsafeGetObject(BasicType basic_type, Value object, Value offset, bool is_volatile) + : UnsafeObjectOp(basic_type, object, offset, false, is_volatile) + { + ASSERT_VALUES + } +}; + + +LEAF(UnsafePutObject, UnsafeObjectOp) + private: + Value _value; // Value to be stored + public: + UnsafePutObject(BasicType basic_type, Value object, Value offset, Value value, bool is_volatile) + : UnsafeObjectOp(basic_type, object, offset, true, is_volatile) + , _value(value) + { + ASSERT_VALUES + } + + // accessors + Value value() { return _value; } + + // generic + virtual void input_values_do(ValueVisitor* f) { UnsafeObjectOp::input_values_do(f); + f->visit(&_value); } +}; + +LEAF(UnsafeGetAndSetObject, UnsafeObjectOp) + private: + Value _value; // Value to be stored + bool _is_add; + public: + UnsafeGetAndSetObject(BasicType basic_type, Value object, Value offset, Value value, bool is_add) + : UnsafeObjectOp(basic_type, object, offset, false, false) + , _value(value) + , _is_add(is_add) + { + ASSERT_VALUES + } + + // accessors + bool is_add() const { return _is_add; } + Value value() { return _value; } + + // generic + virtual void input_values_do(ValueVisitor* f) { UnsafeObjectOp::input_values_do(f); + f->visit(&_value); } +}; + +LEAF(ProfileCall, Instruction) + private: + ciMethod* _method; + int _bci_of_invoke; + ciMethod* _callee; // the method that is called at the given bci + Value _recv; + ciKlass* _known_holder; + Values* _obj_args; // arguments for type profiling + ArgsNonNullState _nonnull_state; // Do we know whether some arguments are never null? + bool _inlined; // Are we profiling a call that is inlined + + public: + ProfileCall(ciMethod* method, int bci, ciMethod* callee, Value recv, ciKlass* known_holder, Values* obj_args, bool inlined) + : Instruction(voidType) + , _method(method) + , _bci_of_invoke(bci) + , _callee(callee) + , _recv(recv) + , _known_holder(known_holder) + , _obj_args(obj_args) + , _inlined(inlined) + { + // The ProfileCall has side-effects and must occur precisely where located + pin(); + } + + ciMethod* method() const { return _method; } + int bci_of_invoke() const { return _bci_of_invoke; } + ciMethod* callee() const { return _callee; } + Value recv() const { return _recv; } + ciKlass* known_holder() const { return _known_holder; } + int nb_profiled_args() const { return _obj_args == NULL ? 0 : _obj_args->length(); } + Value profiled_arg_at(int i) const { return _obj_args->at(i); } + bool arg_needs_null_check(int i) const { + return _nonnull_state.arg_needs_null_check(i); + } + bool inlined() const { return _inlined; } + + void set_arg_needs_null_check(int i, bool check) { + _nonnull_state.set_arg_needs_null_check(i, check); + } + + virtual void input_values_do(ValueVisitor* f) { + if (_recv != NULL) { + f->visit(&_recv); + } + for (int i = 0; i < nb_profiled_args(); i++) { + f->visit(_obj_args->adr_at(i)); + } + } +}; + +LEAF(ProfileReturnType, Instruction) + private: + ciMethod* _method; + ciMethod* _callee; + int _bci_of_invoke; + Value _ret; + + public: + ProfileReturnType(ciMethod* method, int bci, ciMethod* callee, Value ret) + : Instruction(voidType) + , _method(method) + , _callee(callee) + , _bci_of_invoke(bci) + , _ret(ret) + { + set_needs_null_check(true); + // The ProfileType has side-effects and must occur precisely where located + pin(); + } + + ciMethod* method() const { return _method; } + ciMethod* callee() const { return _callee; } + int bci_of_invoke() const { return _bci_of_invoke; } + Value ret() const { return _ret; } + + virtual void input_values_do(ValueVisitor* f) { + if (_ret != NULL) { + f->visit(&_ret); + } + } +}; + +// Call some C runtime function that doesn't safepoint, +// optionally passing the current thread as the first argument. +LEAF(RuntimeCall, Instruction) + private: + const char* _entry_name; + address _entry; + Values* _args; + bool _pass_thread; // Pass the JavaThread* as an implicit first argument + + public: + RuntimeCall(ValueType* type, const char* entry_name, address entry, Values* args, bool pass_thread = true) + : Instruction(type) + , _entry(entry) + , _args(args) + , _entry_name(entry_name) + , _pass_thread(pass_thread) { + ASSERT_VALUES + pin(); + } + + const char* entry_name() const { return _entry_name; } + address entry() const { return _entry; } + int number_of_arguments() const { return _args->length(); } + Value argument_at(int i) const { return _args->at(i); } + bool pass_thread() const { return _pass_thread; } + + virtual void input_values_do(ValueVisitor* f) { + for (int i = 0; i < _args->length(); i++) f->visit(_args->adr_at(i)); + } +}; + +// Use to trip invocation counter of an inlined method + +LEAF(ProfileInvoke, Instruction) + private: + ciMethod* _inlinee; + ValueStack* _state; + + public: + ProfileInvoke(ciMethod* inlinee, ValueStack* state) + : Instruction(voidType) + , _inlinee(inlinee) + , _state(state) + { + // The ProfileInvoke has side-effects and must occur precisely where located QQQ??? + pin(); + } + + ciMethod* inlinee() { return _inlinee; } + ValueStack* state() { return _state; } + virtual void input_values_do(ValueVisitor*) {} + virtual void state_values_do(ValueVisitor*); +}; + +LEAF(MemBar, Instruction) + private: + LIR_Code _code; + + public: + MemBar(LIR_Code code) + : Instruction(voidType) + , _code(code) + { + pin(); + } + + LIR_Code code() { return _code; } + + virtual void input_values_do(ValueVisitor*) {} +}; + +class BlockPair: public CompilationResourceObj { + private: + BlockBegin* _from; + BlockBegin* _to; + public: + BlockPair(BlockBegin* from, BlockBegin* to): _from(from), _to(to) {} + BlockBegin* from() const { return _from; } + BlockBegin* to() const { return _to; } + bool is_same(BlockBegin* from, BlockBegin* to) const { return _from == from && _to == to; } + bool is_same(BlockPair* p) const { return _from == p->from() && _to == p->to(); } + void set_to(BlockBegin* b) { _to = b; } + void set_from(BlockBegin* b) { _from = b; } +}; + +typedef GrowableArray BlockPairList; + +inline int BlockBegin::number_of_sux() const { assert(_end == NULL || _end->number_of_sux() == _successors.length(), "mismatch"); return _successors.length(); } +inline BlockBegin* BlockBegin::sux_at(int i) const { assert(_end == NULL || _end->sux_at(i) == _successors.at(i), "mismatch"); return _successors.at(i); } +inline void BlockBegin::add_successor(BlockBegin* sux) { assert(_end == NULL, "Would create mismatch with successors of BlockEnd"); _successors.append(sux); } + +#undef ASSERT_VALUES + +#endif // SHARE_VM_C1_C1_INSTRUCTION_HPP