dfb1ed8a19
Reviewed-by: jrose, roland
2624 lines
90 KiB
C++
2624 lines
90 KiB
C++
/*
|
|
* Copyright (c) 1999, 2015, Oracle and/or its affiliates. All rights reserved.
|
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
*
|
|
* This code is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 only, as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
* version 2 for more details (a copy is included in the LICENSE file that
|
|
* accompanied this code).
|
|
*
|
|
* You should have received a copy of the GNU General Public License version
|
|
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*
|
|
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
* or visit www.oracle.com if you need additional information or have any
|
|
* questions.
|
|
*
|
|
*/
|
|
|
|
#ifndef SHARE_VM_C1_C1_INSTRUCTION_HPP
|
|
#define SHARE_VM_C1_C1_INSTRUCTION_HPP
|
|
|
|
#include "c1/c1_Compilation.hpp"
|
|
#include "c1/c1_LIR.hpp"
|
|
#include "c1/c1_ValueType.hpp"
|
|
#include "ci/ciField.hpp"
|
|
|
|
// Predefined classes
|
|
class ciField;
|
|
class ValueStack;
|
|
class InstructionPrinter;
|
|
class IRScope;
|
|
class LIR_OprDesc;
|
|
typedef LIR_OprDesc* LIR_Opr;
|
|
|
|
|
|
// Instruction class hierarchy
|
|
//
|
|
// All leaf classes in the class hierarchy are concrete classes
|
|
// (i.e., are instantiated). All other classes are abstract and
|
|
// serve factoring.
|
|
|
|
class Instruction;
|
|
class Phi;
|
|
class Local;
|
|
class Constant;
|
|
class AccessField;
|
|
class LoadField;
|
|
class StoreField;
|
|
class AccessArray;
|
|
class ArrayLength;
|
|
class AccessIndexed;
|
|
class LoadIndexed;
|
|
class StoreIndexed;
|
|
class NegateOp;
|
|
class Op2;
|
|
class ArithmeticOp;
|
|
class ShiftOp;
|
|
class LogicOp;
|
|
class CompareOp;
|
|
class IfOp;
|
|
class Convert;
|
|
class NullCheck;
|
|
class TypeCast;
|
|
class OsrEntry;
|
|
class ExceptionObject;
|
|
class StateSplit;
|
|
class Invoke;
|
|
class NewInstance;
|
|
class NewArray;
|
|
class NewTypeArray;
|
|
class NewObjectArray;
|
|
class NewMultiArray;
|
|
class TypeCheck;
|
|
class CheckCast;
|
|
class InstanceOf;
|
|
class AccessMonitor;
|
|
class MonitorEnter;
|
|
class MonitorExit;
|
|
class Intrinsic;
|
|
class BlockBegin;
|
|
class BlockEnd;
|
|
class Goto;
|
|
class If;
|
|
class IfInstanceOf;
|
|
class Switch;
|
|
class TableSwitch;
|
|
class LookupSwitch;
|
|
class Return;
|
|
class Throw;
|
|
class Base;
|
|
class RoundFP;
|
|
class UnsafeOp;
|
|
class UnsafeRawOp;
|
|
class UnsafeGetRaw;
|
|
class UnsafePutRaw;
|
|
class UnsafeObjectOp;
|
|
class UnsafeGetObject;
|
|
class UnsafePutObject;
|
|
class UnsafeGetAndSetObject;
|
|
class ProfileCall;
|
|
class ProfileReturnType;
|
|
class ProfileInvoke;
|
|
class RuntimeCall;
|
|
class MemBar;
|
|
class RangeCheckPredicate;
|
|
#ifdef ASSERT
|
|
class Assert;
|
|
#endif
|
|
|
|
// A Value is a reference to the instruction creating the value
|
|
typedef Instruction* Value;
|
|
define_array(ValueArray, Value)
|
|
define_stack(Values, ValueArray)
|
|
|
|
define_array(ValueStackArray, ValueStack*)
|
|
define_stack(ValueStackStack, ValueStackArray)
|
|
|
|
// BlockClosure is the base class for block traversal/iteration.
|
|
|
|
class BlockClosure: public CompilationResourceObj {
|
|
public:
|
|
virtual void block_do(BlockBegin* block) = 0;
|
|
};
|
|
|
|
|
|
// A simple closure class for visiting the values of an Instruction
|
|
class ValueVisitor: public StackObj {
|
|
public:
|
|
virtual void visit(Value* v) = 0;
|
|
};
|
|
|
|
|
|
// Some array and list classes
|
|
define_array(BlockBeginArray, BlockBegin*)
|
|
define_stack(_BlockList, BlockBeginArray)
|
|
|
|
class BlockList: public _BlockList {
|
|
public:
|
|
BlockList(): _BlockList() {}
|
|
BlockList(const int size): _BlockList(size) {}
|
|
BlockList(const int size, BlockBegin* init): _BlockList(size, init) {}
|
|
|
|
void iterate_forward(BlockClosure* closure);
|
|
void iterate_backward(BlockClosure* closure);
|
|
void blocks_do(void f(BlockBegin*));
|
|
void values_do(ValueVisitor* f);
|
|
void print(bool cfg_only = false, bool live_only = false) PRODUCT_RETURN;
|
|
};
|
|
|
|
|
|
// InstructionVisitors provide type-based dispatch for instructions.
|
|
// For each concrete Instruction class X, a virtual function do_X is
|
|
// provided. Functionality that needs to be implemented for all classes
|
|
// (e.g., printing, code generation) is factored out into a specialised
|
|
// visitor instead of added to the Instruction classes itself.
|
|
|
|
class InstructionVisitor: public StackObj {
|
|
public:
|
|
virtual void do_Phi (Phi* x) = 0;
|
|
virtual void do_Local (Local* x) = 0;
|
|
virtual void do_Constant (Constant* x) = 0;
|
|
virtual void do_LoadField (LoadField* x) = 0;
|
|
virtual void do_StoreField (StoreField* x) = 0;
|
|
virtual void do_ArrayLength (ArrayLength* x) = 0;
|
|
virtual void do_LoadIndexed (LoadIndexed* x) = 0;
|
|
virtual void do_StoreIndexed (StoreIndexed* x) = 0;
|
|
virtual void do_NegateOp (NegateOp* x) = 0;
|
|
virtual void do_ArithmeticOp (ArithmeticOp* x) = 0;
|
|
virtual void do_ShiftOp (ShiftOp* x) = 0;
|
|
virtual void do_LogicOp (LogicOp* x) = 0;
|
|
virtual void do_CompareOp (CompareOp* x) = 0;
|
|
virtual void do_IfOp (IfOp* x) = 0;
|
|
virtual void do_Convert (Convert* x) = 0;
|
|
virtual void do_NullCheck (NullCheck* x) = 0;
|
|
virtual void do_TypeCast (TypeCast* x) = 0;
|
|
virtual void do_Invoke (Invoke* x) = 0;
|
|
virtual void do_NewInstance (NewInstance* x) = 0;
|
|
virtual void do_NewTypeArray (NewTypeArray* x) = 0;
|
|
virtual void do_NewObjectArray (NewObjectArray* x) = 0;
|
|
virtual void do_NewMultiArray (NewMultiArray* x) = 0;
|
|
virtual void do_CheckCast (CheckCast* x) = 0;
|
|
virtual void do_InstanceOf (InstanceOf* x) = 0;
|
|
virtual void do_MonitorEnter (MonitorEnter* x) = 0;
|
|
virtual void do_MonitorExit (MonitorExit* x) = 0;
|
|
virtual void do_Intrinsic (Intrinsic* x) = 0;
|
|
virtual void do_BlockBegin (BlockBegin* x) = 0;
|
|
virtual void do_Goto (Goto* x) = 0;
|
|
virtual void do_If (If* x) = 0;
|
|
virtual void do_IfInstanceOf (IfInstanceOf* x) = 0;
|
|
virtual void do_TableSwitch (TableSwitch* x) = 0;
|
|
virtual void do_LookupSwitch (LookupSwitch* x) = 0;
|
|
virtual void do_Return (Return* x) = 0;
|
|
virtual void do_Throw (Throw* x) = 0;
|
|
virtual void do_Base (Base* x) = 0;
|
|
virtual void do_OsrEntry (OsrEntry* x) = 0;
|
|
virtual void do_ExceptionObject(ExceptionObject* x) = 0;
|
|
virtual void do_RoundFP (RoundFP* x) = 0;
|
|
virtual void do_UnsafeGetRaw (UnsafeGetRaw* x) = 0;
|
|
virtual void do_UnsafePutRaw (UnsafePutRaw* x) = 0;
|
|
virtual void do_UnsafeGetObject(UnsafeGetObject* x) = 0;
|
|
virtual void do_UnsafePutObject(UnsafePutObject* x) = 0;
|
|
virtual void do_UnsafeGetAndSetObject(UnsafeGetAndSetObject* x) = 0;
|
|
virtual void do_ProfileCall (ProfileCall* x) = 0;
|
|
virtual void do_ProfileReturnType (ProfileReturnType* x) = 0;
|
|
virtual void do_ProfileInvoke (ProfileInvoke* x) = 0;
|
|
virtual void do_RuntimeCall (RuntimeCall* x) = 0;
|
|
virtual void do_MemBar (MemBar* x) = 0;
|
|
virtual void do_RangeCheckPredicate(RangeCheckPredicate* x) = 0;
|
|
#ifdef ASSERT
|
|
virtual void do_Assert (Assert* x) = 0;
|
|
#endif
|
|
};
|
|
|
|
|
|
// Hashing support
|
|
//
|
|
// Note: This hash functions affect the performance
|
|
// of ValueMap - make changes carefully!
|
|
|
|
#define HASH1(x1 ) ((intx)(x1))
|
|
#define HASH2(x1, x2 ) ((HASH1(x1 ) << 7) ^ HASH1(x2))
|
|
#define HASH3(x1, x2, x3 ) ((HASH2(x1, x2 ) << 7) ^ HASH1(x3))
|
|
#define HASH4(x1, x2, x3, x4) ((HASH3(x1, x2, x3) << 7) ^ HASH1(x4))
|
|
|
|
|
|
// The following macros are used to implement instruction-specific hashing.
|
|
// By default, each instruction implements hash() and is_equal(Value), used
|
|
// for value numbering/common subexpression elimination. The default imple-
|
|
// mentation disables value numbering. Each instruction which can be value-
|
|
// numbered, should define corresponding hash() and is_equal(Value) functions
|
|
// via the macros below. The f arguments specify all the values/op codes, etc.
|
|
// that need to be identical for two instructions to be identical.
|
|
//
|
|
// Note: The default implementation of hash() returns 0 in order to indicate
|
|
// that the instruction should not be considered for value numbering.
|
|
// The currently used hash functions do not guarantee that never a 0
|
|
// is produced. While this is still correct, it may be a performance
|
|
// bug (no value numbering for that node). However, this situation is
|
|
// so unlikely, that we are not going to handle it specially.
|
|
|
|
#define HASHING1(class_name, enabled, f1) \
|
|
virtual intx hash() const { \
|
|
return (enabled) ? HASH2(name(), f1) : 0; \
|
|
} \
|
|
virtual bool is_equal(Value v) const { \
|
|
if (!(enabled) ) return false; \
|
|
class_name* _v = v->as_##class_name(); \
|
|
if (_v == NULL ) return false; \
|
|
if (f1 != _v->f1) return false; \
|
|
return true; \
|
|
} \
|
|
|
|
|
|
#define HASHING2(class_name, enabled, f1, f2) \
|
|
virtual intx hash() const { \
|
|
return (enabled) ? HASH3(name(), f1, f2) : 0; \
|
|
} \
|
|
virtual bool is_equal(Value v) const { \
|
|
if (!(enabled) ) return false; \
|
|
class_name* _v = v->as_##class_name(); \
|
|
if (_v == NULL ) return false; \
|
|
if (f1 != _v->f1) return false; \
|
|
if (f2 != _v->f2) return false; \
|
|
return true; \
|
|
} \
|
|
|
|
|
|
#define HASHING3(class_name, enabled, f1, f2, f3) \
|
|
virtual intx hash() const { \
|
|
return (enabled) ? HASH4(name(), f1, f2, f3) : 0; \
|
|
} \
|
|
virtual bool is_equal(Value v) const { \
|
|
if (!(enabled) ) return false; \
|
|
class_name* _v = v->as_##class_name(); \
|
|
if (_v == NULL ) return false; \
|
|
if (f1 != _v->f1) return false; \
|
|
if (f2 != _v->f2) return false; \
|
|
if (f3 != _v->f3) return false; \
|
|
return true; \
|
|
} \
|
|
|
|
|
|
// The mother of all instructions...
|
|
|
|
class Instruction: public CompilationResourceObj {
|
|
private:
|
|
int _id; // the unique instruction id
|
|
#ifndef PRODUCT
|
|
int _printable_bci; // the bci of the instruction for printing
|
|
#endif
|
|
int _use_count; // the number of instructions refering to this value (w/o prev/next); only roots can have use count = 0 or > 1
|
|
int _pin_state; // set of PinReason describing the reason for pinning
|
|
ValueType* _type; // the instruction value type
|
|
Instruction* _next; // the next instruction if any (NULL for BlockEnd instructions)
|
|
Instruction* _subst; // the substitution instruction if any
|
|
LIR_Opr _operand; // LIR specific information
|
|
unsigned int _flags; // Flag bits
|
|
|
|
ValueStack* _state_before; // Copy of state with input operands still on stack (or NULL)
|
|
ValueStack* _exception_state; // Copy of state for exception handling
|
|
XHandlers* _exception_handlers; // Flat list of exception handlers covering this instruction
|
|
|
|
friend class UseCountComputer;
|
|
friend class BlockBegin;
|
|
|
|
void update_exception_state(ValueStack* state);
|
|
|
|
protected:
|
|
BlockBegin* _block; // Block that contains this instruction
|
|
|
|
void set_type(ValueType* type) {
|
|
assert(type != NULL, "type must exist");
|
|
_type = type;
|
|
}
|
|
|
|
// Helper class to keep track of which arguments need a null check
|
|
class ArgsNonNullState {
|
|
private:
|
|
int _nonnull_state; // mask identifying which args are nonnull
|
|
public:
|
|
ArgsNonNullState()
|
|
: _nonnull_state(AllBits) {}
|
|
|
|
// Does argument number i needs a null check?
|
|
bool arg_needs_null_check(int i) const {
|
|
// No data is kept for arguments starting at position 33 so
|
|
// conservatively assume that they need a null check.
|
|
if (i >= 0 && i < (int)sizeof(_nonnull_state) * BitsPerByte) {
|
|
return is_set_nth_bit(_nonnull_state, i);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
// Set whether argument number i needs a null check or not
|
|
void set_arg_needs_null_check(int i, bool check) {
|
|
if (i >= 0 && i < (int)sizeof(_nonnull_state) * BitsPerByte) {
|
|
if (check) {
|
|
_nonnull_state |= nth_bit(i);
|
|
} else {
|
|
_nonnull_state &= ~(nth_bit(i));
|
|
}
|
|
}
|
|
}
|
|
};
|
|
|
|
public:
|
|
void* operator new(size_t size) throw() {
|
|
Compilation* c = Compilation::current();
|
|
void* res = c->arena()->Amalloc(size);
|
|
((Instruction*)res)->_id = c->get_next_id();
|
|
return res;
|
|
}
|
|
|
|
static const int no_bci = -99;
|
|
|
|
enum InstructionFlag {
|
|
NeedsNullCheckFlag = 0,
|
|
CanTrapFlag,
|
|
DirectCompareFlag,
|
|
IsEliminatedFlag,
|
|
IsSafepointFlag,
|
|
IsStaticFlag,
|
|
IsStrictfpFlag,
|
|
NeedsStoreCheckFlag,
|
|
NeedsWriteBarrierFlag,
|
|
PreservesStateFlag,
|
|
TargetIsFinalFlag,
|
|
TargetIsLoadedFlag,
|
|
TargetIsStrictfpFlag,
|
|
UnorderedIsTrueFlag,
|
|
NeedsPatchingFlag,
|
|
ThrowIncompatibleClassChangeErrorFlag,
|
|
ProfileMDOFlag,
|
|
IsLinkedInBlockFlag,
|
|
NeedsRangeCheckFlag,
|
|
InWorkListFlag,
|
|
DeoptimizeOnException,
|
|
InstructionLastFlag
|
|
};
|
|
|
|
public:
|
|
bool check_flag(InstructionFlag id) const { return (_flags & (1 << id)) != 0; }
|
|
void set_flag(InstructionFlag id, bool f) { _flags = f ? (_flags | (1 << id)) : (_flags & ~(1 << id)); };
|
|
|
|
// 'globally' used condition values
|
|
enum Condition {
|
|
eql, neq, lss, leq, gtr, geq, aeq, beq
|
|
};
|
|
|
|
// Instructions may be pinned for many reasons and under certain conditions
|
|
// with enough knowledge it's possible to safely unpin them.
|
|
enum PinReason {
|
|
PinUnknown = 1 << 0
|
|
, PinExplicitNullCheck = 1 << 3
|
|
, PinStackForStateSplit= 1 << 12
|
|
, PinStateSplitConstructor= 1 << 13
|
|
, PinGlobalValueNumbering= 1 << 14
|
|
};
|
|
|
|
static Condition mirror(Condition cond);
|
|
static Condition negate(Condition cond);
|
|
|
|
// initialization
|
|
static int number_of_instructions() {
|
|
return Compilation::current()->number_of_instructions();
|
|
}
|
|
|
|
// creation
|
|
Instruction(ValueType* type, ValueStack* state_before = NULL, bool type_is_constant = false)
|
|
: _use_count(0)
|
|
#ifndef PRODUCT
|
|
, _printable_bci(-99)
|
|
#endif
|
|
, _pin_state(0)
|
|
, _type(type)
|
|
, _next(NULL)
|
|
, _block(NULL)
|
|
, _subst(NULL)
|
|
, _flags(0)
|
|
, _operand(LIR_OprFact::illegalOpr)
|
|
, _state_before(state_before)
|
|
, _exception_handlers(NULL)
|
|
{
|
|
check_state(state_before);
|
|
assert(type != NULL && (!type->is_constant() || type_is_constant), "type must exist");
|
|
update_exception_state(_state_before);
|
|
}
|
|
|
|
// accessors
|
|
int id() const { return _id; }
|
|
#ifndef PRODUCT
|
|
bool has_printable_bci() const { return _printable_bci != -99; }
|
|
int printable_bci() const { assert(has_printable_bci(), "_printable_bci should have been set"); return _printable_bci; }
|
|
void set_printable_bci(int bci) { _printable_bci = bci; }
|
|
#endif
|
|
int dominator_depth();
|
|
int use_count() const { return _use_count; }
|
|
int pin_state() const { return _pin_state; }
|
|
bool is_pinned() const { return _pin_state != 0 || PinAllInstructions; }
|
|
ValueType* type() const { return _type; }
|
|
BlockBegin *block() const { return _block; }
|
|
Instruction* prev(); // use carefully, expensive operation
|
|
Instruction* next() const { return _next; }
|
|
bool has_subst() const { return _subst != NULL; }
|
|
Instruction* subst() { return _subst == NULL ? this : _subst->subst(); }
|
|
LIR_Opr operand() const { return _operand; }
|
|
|
|
void set_needs_null_check(bool f) { set_flag(NeedsNullCheckFlag, f); }
|
|
bool needs_null_check() const { return check_flag(NeedsNullCheckFlag); }
|
|
bool is_linked() const { return check_flag(IsLinkedInBlockFlag); }
|
|
bool can_be_linked() { return as_Local() == NULL && as_Phi() == NULL; }
|
|
|
|
bool has_uses() const { return use_count() > 0; }
|
|
ValueStack* state_before() const { return _state_before; }
|
|
ValueStack* exception_state() const { return _exception_state; }
|
|
virtual bool needs_exception_state() const { return true; }
|
|
XHandlers* exception_handlers() const { return _exception_handlers; }
|
|
|
|
// manipulation
|
|
void pin(PinReason reason) { _pin_state |= reason; }
|
|
void pin() { _pin_state |= PinUnknown; }
|
|
// DANGEROUS: only used by EliminateStores
|
|
void unpin(PinReason reason) { assert((reason & PinUnknown) == 0, "can't unpin unknown state"); _pin_state &= ~reason; }
|
|
|
|
Instruction* set_next(Instruction* next) {
|
|
assert(next->has_printable_bci(), "_printable_bci should have been set");
|
|
assert(next != NULL, "must not be NULL");
|
|
assert(as_BlockEnd() == NULL, "BlockEnd instructions must have no next");
|
|
assert(next->can_be_linked(), "shouldn't link these instructions into list");
|
|
|
|
BlockBegin *block = this->block();
|
|
next->_block = block;
|
|
|
|
next->set_flag(Instruction::IsLinkedInBlockFlag, true);
|
|
_next = next;
|
|
return next;
|
|
}
|
|
|
|
Instruction* set_next(Instruction* next, int bci) {
|
|
#ifndef PRODUCT
|
|
next->set_printable_bci(bci);
|
|
#endif
|
|
return set_next(next);
|
|
}
|
|
|
|
// when blocks are merged
|
|
void fixup_block_pointers() {
|
|
Instruction *cur = next()->next(); // next()'s block is set in set_next
|
|
while (cur && cur->_block != block()) {
|
|
cur->_block = block();
|
|
cur = cur->next();
|
|
}
|
|
}
|
|
|
|
Instruction *insert_after(Instruction *i) {
|
|
Instruction* n = _next;
|
|
set_next(i);
|
|
i->set_next(n);
|
|
return _next;
|
|
}
|
|
|
|
Instruction *insert_after_same_bci(Instruction *i) {
|
|
#ifndef PRODUCT
|
|
i->set_printable_bci(printable_bci());
|
|
#endif
|
|
return insert_after(i);
|
|
}
|
|
|
|
void set_subst(Instruction* subst) {
|
|
assert(subst == NULL ||
|
|
type()->base() == subst->type()->base() ||
|
|
subst->type()->base() == illegalType, "type can't change");
|
|
_subst = subst;
|
|
}
|
|
void set_exception_handlers(XHandlers *xhandlers) { _exception_handlers = xhandlers; }
|
|
void set_exception_state(ValueStack* s) { check_state(s); _exception_state = s; }
|
|
void set_state_before(ValueStack* s) { check_state(s); _state_before = s; }
|
|
|
|
// machine-specifics
|
|
void set_operand(LIR_Opr operand) { assert(operand != LIR_OprFact::illegalOpr, "operand must exist"); _operand = operand; }
|
|
void clear_operand() { _operand = LIR_OprFact::illegalOpr; }
|
|
|
|
// generic
|
|
virtual Instruction* as_Instruction() { return this; } // to satisfy HASHING1 macro
|
|
virtual Phi* as_Phi() { return NULL; }
|
|
virtual Local* as_Local() { return NULL; }
|
|
virtual Constant* as_Constant() { return NULL; }
|
|
virtual AccessField* as_AccessField() { return NULL; }
|
|
virtual LoadField* as_LoadField() { return NULL; }
|
|
virtual StoreField* as_StoreField() { return NULL; }
|
|
virtual AccessArray* as_AccessArray() { return NULL; }
|
|
virtual ArrayLength* as_ArrayLength() { return NULL; }
|
|
virtual AccessIndexed* as_AccessIndexed() { return NULL; }
|
|
virtual LoadIndexed* as_LoadIndexed() { return NULL; }
|
|
virtual StoreIndexed* as_StoreIndexed() { return NULL; }
|
|
virtual NegateOp* as_NegateOp() { return NULL; }
|
|
virtual Op2* as_Op2() { return NULL; }
|
|
virtual ArithmeticOp* as_ArithmeticOp() { return NULL; }
|
|
virtual ShiftOp* as_ShiftOp() { return NULL; }
|
|
virtual LogicOp* as_LogicOp() { return NULL; }
|
|
virtual CompareOp* as_CompareOp() { return NULL; }
|
|
virtual IfOp* as_IfOp() { return NULL; }
|
|
virtual Convert* as_Convert() { return NULL; }
|
|
virtual NullCheck* as_NullCheck() { return NULL; }
|
|
virtual OsrEntry* as_OsrEntry() { return NULL; }
|
|
virtual StateSplit* as_StateSplit() { return NULL; }
|
|
virtual Invoke* as_Invoke() { return NULL; }
|
|
virtual NewInstance* as_NewInstance() { return NULL; }
|
|
virtual NewArray* as_NewArray() { return NULL; }
|
|
virtual NewTypeArray* as_NewTypeArray() { return NULL; }
|
|
virtual NewObjectArray* as_NewObjectArray() { return NULL; }
|
|
virtual NewMultiArray* as_NewMultiArray() { return NULL; }
|
|
virtual TypeCheck* as_TypeCheck() { return NULL; }
|
|
virtual CheckCast* as_CheckCast() { return NULL; }
|
|
virtual InstanceOf* as_InstanceOf() { return NULL; }
|
|
virtual TypeCast* as_TypeCast() { return NULL; }
|
|
virtual AccessMonitor* as_AccessMonitor() { return NULL; }
|
|
virtual MonitorEnter* as_MonitorEnter() { return NULL; }
|
|
virtual MonitorExit* as_MonitorExit() { return NULL; }
|
|
virtual Intrinsic* as_Intrinsic() { return NULL; }
|
|
virtual BlockBegin* as_BlockBegin() { return NULL; }
|
|
virtual BlockEnd* as_BlockEnd() { return NULL; }
|
|
virtual Goto* as_Goto() { return NULL; }
|
|
virtual If* as_If() { return NULL; }
|
|
virtual IfInstanceOf* as_IfInstanceOf() { return NULL; }
|
|
virtual TableSwitch* as_TableSwitch() { return NULL; }
|
|
virtual LookupSwitch* as_LookupSwitch() { return NULL; }
|
|
virtual Return* as_Return() { return NULL; }
|
|
virtual Throw* as_Throw() { return NULL; }
|
|
virtual Base* as_Base() { return NULL; }
|
|
virtual RoundFP* as_RoundFP() { return NULL; }
|
|
virtual ExceptionObject* as_ExceptionObject() { return NULL; }
|
|
virtual UnsafeOp* as_UnsafeOp() { return NULL; }
|
|
virtual ProfileInvoke* as_ProfileInvoke() { return NULL; }
|
|
virtual RangeCheckPredicate* as_RangeCheckPredicate() { return NULL; }
|
|
|
|
#ifdef ASSERT
|
|
virtual Assert* as_Assert() { return NULL; }
|
|
#endif
|
|
|
|
virtual void visit(InstructionVisitor* v) = 0;
|
|
|
|
virtual bool can_trap() const { return false; }
|
|
|
|
virtual void input_values_do(ValueVisitor* f) = 0;
|
|
virtual void state_values_do(ValueVisitor* f);
|
|
virtual void other_values_do(ValueVisitor* f) { /* usually no other - override on demand */ }
|
|
void values_do(ValueVisitor* f) { input_values_do(f); state_values_do(f); other_values_do(f); }
|
|
|
|
virtual ciType* exact_type() const;
|
|
virtual ciType* declared_type() const { return NULL; }
|
|
|
|
// hashing
|
|
virtual const char* name() const = 0;
|
|
HASHING1(Instruction, false, id()) // hashing disabled by default
|
|
|
|
// debugging
|
|
static void check_state(ValueStack* state) PRODUCT_RETURN;
|
|
void print() PRODUCT_RETURN;
|
|
void print_line() PRODUCT_RETURN;
|
|
void print(InstructionPrinter& ip) PRODUCT_RETURN;
|
|
};
|
|
|
|
|
|
// The following macros are used to define base (i.e., non-leaf)
|
|
// and leaf instruction classes. They define class-name related
|
|
// generic functionality in one place.
|
|
|
|
#define BASE(class_name, super_class_name) \
|
|
class class_name: public super_class_name { \
|
|
public: \
|
|
virtual class_name* as_##class_name() { return this; } \
|
|
|
|
|
|
#define LEAF(class_name, super_class_name) \
|
|
BASE(class_name, super_class_name) \
|
|
public: \
|
|
virtual const char* name() const { return #class_name; } \
|
|
virtual void visit(InstructionVisitor* v) { v->do_##class_name(this); } \
|
|
|
|
|
|
// Debugging support
|
|
|
|
|
|
#ifdef ASSERT
|
|
class AssertValues: public ValueVisitor {
|
|
void visit(Value* x) { assert((*x) != NULL, "value must exist"); }
|
|
};
|
|
#define ASSERT_VALUES { AssertValues assert_value; values_do(&assert_value); }
|
|
#else
|
|
#define ASSERT_VALUES
|
|
#endif // ASSERT
|
|
|
|
|
|
// A Phi is a phi function in the sense of SSA form. It stands for
|
|
// the value of a local variable at the beginning of a join block.
|
|
// A Phi consists of n operands, one for every incoming branch.
|
|
|
|
LEAF(Phi, Instruction)
|
|
private:
|
|
int _pf_flags; // the flags of the phi function
|
|
int _index; // to value on operand stack (index < 0) or to local
|
|
public:
|
|
// creation
|
|
Phi(ValueType* type, BlockBegin* b, int index)
|
|
: Instruction(type->base())
|
|
, _pf_flags(0)
|
|
, _index(index)
|
|
{
|
|
_block = b;
|
|
NOT_PRODUCT(set_printable_bci(Value(b)->printable_bci()));
|
|
if (type->is_illegal()) {
|
|
make_illegal();
|
|
}
|
|
}
|
|
|
|
// flags
|
|
enum Flag {
|
|
no_flag = 0,
|
|
visited = 1 << 0,
|
|
cannot_simplify = 1 << 1
|
|
};
|
|
|
|
// accessors
|
|
bool is_local() const { return _index >= 0; }
|
|
bool is_on_stack() const { return !is_local(); }
|
|
int local_index() const { assert(is_local(), ""); return _index; }
|
|
int stack_index() const { assert(is_on_stack(), ""); return -(_index+1); }
|
|
|
|
Value operand_at(int i) const;
|
|
int operand_count() const;
|
|
|
|
void set(Flag f) { _pf_flags |= f; }
|
|
void clear(Flag f) { _pf_flags &= ~f; }
|
|
bool is_set(Flag f) const { return (_pf_flags & f) != 0; }
|
|
|
|
// Invalidates phis corresponding to merges of locals of two different types
|
|
// (these should never be referenced, otherwise the bytecodes are illegal)
|
|
void make_illegal() {
|
|
set(cannot_simplify);
|
|
set_type(illegalType);
|
|
}
|
|
|
|
bool is_illegal() const {
|
|
return type()->is_illegal();
|
|
}
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) {
|
|
}
|
|
};
|
|
|
|
|
|
// A local is a placeholder for an incoming argument to a function call.
|
|
LEAF(Local, Instruction)
|
|
private:
|
|
int _java_index; // the local index within the method to which the local belongs
|
|
bool _is_receiver; // if local variable holds the receiver: "this" for non-static methods
|
|
ciType* _declared_type;
|
|
public:
|
|
// creation
|
|
Local(ciType* declared, ValueType* type, int index, bool receiver)
|
|
: Instruction(type)
|
|
, _java_index(index)
|
|
, _declared_type(declared)
|
|
, _is_receiver(receiver)
|
|
{
|
|
NOT_PRODUCT(set_printable_bci(-1));
|
|
}
|
|
|
|
// accessors
|
|
int java_index() const { return _java_index; }
|
|
bool is_receiver() const { return _is_receiver; }
|
|
|
|
virtual ciType* declared_type() const { return _declared_type; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { /* no values */ }
|
|
};
|
|
|
|
|
|
LEAF(Constant, Instruction)
|
|
public:
|
|
// creation
|
|
Constant(ValueType* type):
|
|
Instruction(type, NULL, /*type_is_constant*/ true)
|
|
{
|
|
assert(type->is_constant(), "must be a constant");
|
|
}
|
|
|
|
Constant(ValueType* type, ValueStack* state_before):
|
|
Instruction(type, state_before, /*type_is_constant*/ true)
|
|
{
|
|
assert(state_before != NULL, "only used for constants which need patching");
|
|
assert(type->is_constant(), "must be a constant");
|
|
// since it's patching it needs to be pinned
|
|
pin();
|
|
}
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return state_before() != NULL; }
|
|
virtual void input_values_do(ValueVisitor* f) { /* no values */ }
|
|
|
|
virtual intx hash() const;
|
|
virtual bool is_equal(Value v) const;
|
|
|
|
virtual ciType* exact_type() const;
|
|
|
|
enum CompareResult { not_comparable = -1, cond_false, cond_true };
|
|
|
|
virtual CompareResult compare(Instruction::Condition condition, Value right) const;
|
|
BlockBegin* compare(Instruction::Condition cond, Value right,
|
|
BlockBegin* true_sux, BlockBegin* false_sux) const {
|
|
switch (compare(cond, right)) {
|
|
case not_comparable:
|
|
return NULL;
|
|
case cond_false:
|
|
return false_sux;
|
|
case cond_true:
|
|
return true_sux;
|
|
default:
|
|
ShouldNotReachHere();
|
|
return NULL;
|
|
}
|
|
}
|
|
};
|
|
|
|
|
|
BASE(AccessField, Instruction)
|
|
private:
|
|
Value _obj;
|
|
int _offset;
|
|
ciField* _field;
|
|
NullCheck* _explicit_null_check; // For explicit null check elimination
|
|
|
|
public:
|
|
// creation
|
|
AccessField(Value obj, int offset, ciField* field, bool is_static,
|
|
ValueStack* state_before, bool needs_patching)
|
|
: Instruction(as_ValueType(field->type()->basic_type()), state_before)
|
|
, _obj(obj)
|
|
, _offset(offset)
|
|
, _field(field)
|
|
, _explicit_null_check(NULL)
|
|
{
|
|
set_needs_null_check(!is_static);
|
|
set_flag(IsStaticFlag, is_static);
|
|
set_flag(NeedsPatchingFlag, needs_patching);
|
|
ASSERT_VALUES
|
|
// pin of all instructions with memory access
|
|
pin();
|
|
}
|
|
|
|
// accessors
|
|
Value obj() const { return _obj; }
|
|
int offset() const { return _offset; }
|
|
ciField* field() const { return _field; }
|
|
BasicType field_type() const { return _field->type()->basic_type(); }
|
|
bool is_static() const { return check_flag(IsStaticFlag); }
|
|
NullCheck* explicit_null_check() const { return _explicit_null_check; }
|
|
bool needs_patching() const { return check_flag(NeedsPatchingFlag); }
|
|
|
|
// Unresolved getstatic and putstatic can cause initialization.
|
|
// Technically it occurs at the Constant that materializes the base
|
|
// of the static fields but it's simpler to model it here.
|
|
bool is_init_point() const { return is_static() && (needs_patching() || !_field->holder()->is_initialized()); }
|
|
|
|
// manipulation
|
|
|
|
// Under certain circumstances, if a previous NullCheck instruction
|
|
// proved the target object non-null, we can eliminate the explicit
|
|
// null check and do an implicit one, simply specifying the debug
|
|
// information from the NullCheck. This field should only be consulted
|
|
// if needs_null_check() is true.
|
|
void set_explicit_null_check(NullCheck* check) { _explicit_null_check = check; }
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return needs_null_check() || needs_patching(); }
|
|
virtual void input_values_do(ValueVisitor* f) { f->visit(&_obj); }
|
|
};
|
|
|
|
|
|
LEAF(LoadField, AccessField)
|
|
public:
|
|
// creation
|
|
LoadField(Value obj, int offset, ciField* field, bool is_static,
|
|
ValueStack* state_before, bool needs_patching)
|
|
: AccessField(obj, offset, field, is_static, state_before, needs_patching)
|
|
{}
|
|
|
|
ciType* declared_type() const;
|
|
|
|
// generic
|
|
HASHING2(LoadField, !needs_patching() && !field()->is_volatile(), obj()->subst(), offset()) // cannot be eliminated if needs patching or if volatile
|
|
};
|
|
|
|
|
|
LEAF(StoreField, AccessField)
|
|
private:
|
|
Value _value;
|
|
|
|
public:
|
|
// creation
|
|
StoreField(Value obj, int offset, ciField* field, Value value, bool is_static,
|
|
ValueStack* state_before, bool needs_patching)
|
|
: AccessField(obj, offset, field, is_static, state_before, needs_patching)
|
|
, _value(value)
|
|
{
|
|
set_flag(NeedsWriteBarrierFlag, as_ValueType(field_type())->is_object());
|
|
ASSERT_VALUES
|
|
pin();
|
|
}
|
|
|
|
// accessors
|
|
Value value() const { return _value; }
|
|
bool needs_write_barrier() const { return check_flag(NeedsWriteBarrierFlag); }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { AccessField::input_values_do(f); f->visit(&_value); }
|
|
};
|
|
|
|
|
|
BASE(AccessArray, Instruction)
|
|
private:
|
|
Value _array;
|
|
|
|
public:
|
|
// creation
|
|
AccessArray(ValueType* type, Value array, ValueStack* state_before)
|
|
: Instruction(type, state_before)
|
|
, _array(array)
|
|
{
|
|
set_needs_null_check(true);
|
|
ASSERT_VALUES
|
|
pin(); // instruction with side effect (null exception or range check throwing)
|
|
}
|
|
|
|
Value array() const { return _array; }
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return needs_null_check(); }
|
|
virtual void input_values_do(ValueVisitor* f) { f->visit(&_array); }
|
|
};
|
|
|
|
|
|
LEAF(ArrayLength, AccessArray)
|
|
private:
|
|
NullCheck* _explicit_null_check; // For explicit null check elimination
|
|
|
|
public:
|
|
// creation
|
|
ArrayLength(Value array, ValueStack* state_before)
|
|
: AccessArray(intType, array, state_before)
|
|
, _explicit_null_check(NULL) {}
|
|
|
|
// accessors
|
|
NullCheck* explicit_null_check() const { return _explicit_null_check; }
|
|
|
|
// setters
|
|
// See LoadField::set_explicit_null_check for documentation
|
|
void set_explicit_null_check(NullCheck* check) { _explicit_null_check = check; }
|
|
|
|
// generic
|
|
HASHING1(ArrayLength, true, array()->subst())
|
|
};
|
|
|
|
|
|
BASE(AccessIndexed, AccessArray)
|
|
private:
|
|
Value _index;
|
|
Value _length;
|
|
BasicType _elt_type;
|
|
|
|
public:
|
|
// creation
|
|
AccessIndexed(Value array, Value index, Value length, BasicType elt_type, ValueStack* state_before)
|
|
: AccessArray(as_ValueType(elt_type), array, state_before)
|
|
, _index(index)
|
|
, _length(length)
|
|
, _elt_type(elt_type)
|
|
{
|
|
set_flag(Instruction::NeedsRangeCheckFlag, true);
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
Value index() const { return _index; }
|
|
Value length() const { return _length; }
|
|
BasicType elt_type() const { return _elt_type; }
|
|
|
|
void clear_length() { _length = NULL; }
|
|
// perform elimination of range checks involving constants
|
|
bool compute_needs_range_check();
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { AccessArray::input_values_do(f); f->visit(&_index); if (_length != NULL) f->visit(&_length); }
|
|
};
|
|
|
|
|
|
LEAF(LoadIndexed, AccessIndexed)
|
|
private:
|
|
NullCheck* _explicit_null_check; // For explicit null check elimination
|
|
|
|
public:
|
|
// creation
|
|
LoadIndexed(Value array, Value index, Value length, BasicType elt_type, ValueStack* state_before)
|
|
: AccessIndexed(array, index, length, elt_type, state_before)
|
|
, _explicit_null_check(NULL) {}
|
|
|
|
// accessors
|
|
NullCheck* explicit_null_check() const { return _explicit_null_check; }
|
|
|
|
// setters
|
|
// See LoadField::set_explicit_null_check for documentation
|
|
void set_explicit_null_check(NullCheck* check) { _explicit_null_check = check; }
|
|
|
|
ciType* exact_type() const;
|
|
ciType* declared_type() const;
|
|
|
|
// generic
|
|
HASHING2(LoadIndexed, true, array()->subst(), index()->subst())
|
|
};
|
|
|
|
|
|
LEAF(StoreIndexed, AccessIndexed)
|
|
private:
|
|
Value _value;
|
|
|
|
ciMethod* _profiled_method;
|
|
int _profiled_bci;
|
|
public:
|
|
// creation
|
|
StoreIndexed(Value array, Value index, Value length, BasicType elt_type, Value value, ValueStack* state_before)
|
|
: AccessIndexed(array, index, length, elt_type, state_before)
|
|
, _value(value), _profiled_method(NULL), _profiled_bci(0)
|
|
{
|
|
set_flag(NeedsWriteBarrierFlag, (as_ValueType(elt_type)->is_object()));
|
|
set_flag(NeedsStoreCheckFlag, (as_ValueType(elt_type)->is_object()));
|
|
ASSERT_VALUES
|
|
pin();
|
|
}
|
|
|
|
// accessors
|
|
Value value() const { return _value; }
|
|
bool needs_write_barrier() const { return check_flag(NeedsWriteBarrierFlag); }
|
|
bool needs_store_check() const { return check_flag(NeedsStoreCheckFlag); }
|
|
// Helpers for MethodData* profiling
|
|
void set_should_profile(bool value) { set_flag(ProfileMDOFlag, value); }
|
|
void set_profiled_method(ciMethod* method) { _profiled_method = method; }
|
|
void set_profiled_bci(int bci) { _profiled_bci = bci; }
|
|
bool should_profile() const { return check_flag(ProfileMDOFlag); }
|
|
ciMethod* profiled_method() const { return _profiled_method; }
|
|
int profiled_bci() const { return _profiled_bci; }
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { AccessIndexed::input_values_do(f); f->visit(&_value); }
|
|
};
|
|
|
|
|
|
LEAF(NegateOp, Instruction)
|
|
private:
|
|
Value _x;
|
|
|
|
public:
|
|
// creation
|
|
NegateOp(Value x) : Instruction(x->type()->base()), _x(x) {
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
Value x() const { return _x; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { f->visit(&_x); }
|
|
};
|
|
|
|
|
|
BASE(Op2, Instruction)
|
|
private:
|
|
Bytecodes::Code _op;
|
|
Value _x;
|
|
Value _y;
|
|
|
|
public:
|
|
// creation
|
|
Op2(ValueType* type, Bytecodes::Code op, Value x, Value y, ValueStack* state_before = NULL)
|
|
: Instruction(type, state_before)
|
|
, _op(op)
|
|
, _x(x)
|
|
, _y(y)
|
|
{
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
Bytecodes::Code op() const { return _op; }
|
|
Value x() const { return _x; }
|
|
Value y() const { return _y; }
|
|
|
|
// manipulators
|
|
void swap_operands() {
|
|
assert(is_commutative(), "operation must be commutative");
|
|
Value t = _x; _x = _y; _y = t;
|
|
}
|
|
|
|
// generic
|
|
virtual bool is_commutative() const { return false; }
|
|
virtual void input_values_do(ValueVisitor* f) { f->visit(&_x); f->visit(&_y); }
|
|
};
|
|
|
|
|
|
LEAF(ArithmeticOp, Op2)
|
|
public:
|
|
// creation
|
|
ArithmeticOp(Bytecodes::Code op, Value x, Value y, bool is_strictfp, ValueStack* state_before)
|
|
: Op2(x->type()->meet(y->type()), op, x, y, state_before)
|
|
{
|
|
set_flag(IsStrictfpFlag, is_strictfp);
|
|
if (can_trap()) pin();
|
|
}
|
|
|
|
// accessors
|
|
bool is_strictfp() const { return check_flag(IsStrictfpFlag); }
|
|
|
|
// generic
|
|
virtual bool is_commutative() const;
|
|
virtual bool can_trap() const;
|
|
HASHING3(Op2, true, op(), x()->subst(), y()->subst())
|
|
};
|
|
|
|
|
|
LEAF(ShiftOp, Op2)
|
|
public:
|
|
// creation
|
|
ShiftOp(Bytecodes::Code op, Value x, Value s) : Op2(x->type()->base(), op, x, s) {}
|
|
|
|
// generic
|
|
HASHING3(Op2, true, op(), x()->subst(), y()->subst())
|
|
};
|
|
|
|
|
|
LEAF(LogicOp, Op2)
|
|
public:
|
|
// creation
|
|
LogicOp(Bytecodes::Code op, Value x, Value y) : Op2(x->type()->meet(y->type()), op, x, y) {}
|
|
|
|
// generic
|
|
virtual bool is_commutative() const;
|
|
HASHING3(Op2, true, op(), x()->subst(), y()->subst())
|
|
};
|
|
|
|
|
|
LEAF(CompareOp, Op2)
|
|
public:
|
|
// creation
|
|
CompareOp(Bytecodes::Code op, Value x, Value y, ValueStack* state_before)
|
|
: Op2(intType, op, x, y, state_before)
|
|
{}
|
|
|
|
// generic
|
|
HASHING3(Op2, true, op(), x()->subst(), y()->subst())
|
|
};
|
|
|
|
|
|
LEAF(IfOp, Op2)
|
|
private:
|
|
Value _tval;
|
|
Value _fval;
|
|
|
|
public:
|
|
// creation
|
|
IfOp(Value x, Condition cond, Value y, Value tval, Value fval)
|
|
: Op2(tval->type()->meet(fval->type()), (Bytecodes::Code)cond, x, y)
|
|
, _tval(tval)
|
|
, _fval(fval)
|
|
{
|
|
ASSERT_VALUES
|
|
assert(tval->type()->tag() == fval->type()->tag(), "types must match");
|
|
}
|
|
|
|
// accessors
|
|
virtual bool is_commutative() const;
|
|
Bytecodes::Code op() const { ShouldNotCallThis(); return Bytecodes::_illegal; }
|
|
Condition cond() const { return (Condition)Op2::op(); }
|
|
Value tval() const { return _tval; }
|
|
Value fval() const { return _fval; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { Op2::input_values_do(f); f->visit(&_tval); f->visit(&_fval); }
|
|
};
|
|
|
|
|
|
LEAF(Convert, Instruction)
|
|
private:
|
|
Bytecodes::Code _op;
|
|
Value _value;
|
|
|
|
public:
|
|
// creation
|
|
Convert(Bytecodes::Code op, Value value, ValueType* to_type) : Instruction(to_type), _op(op), _value(value) {
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
Bytecodes::Code op() const { return _op; }
|
|
Value value() const { return _value; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { f->visit(&_value); }
|
|
HASHING2(Convert, true, op(), value()->subst())
|
|
};
|
|
|
|
|
|
LEAF(NullCheck, Instruction)
|
|
private:
|
|
Value _obj;
|
|
|
|
public:
|
|
// creation
|
|
NullCheck(Value obj, ValueStack* state_before)
|
|
: Instruction(obj->type()->base(), state_before)
|
|
, _obj(obj)
|
|
{
|
|
ASSERT_VALUES
|
|
set_can_trap(true);
|
|
assert(_obj->type()->is_object(), "null check must be applied to objects only");
|
|
pin(Instruction::PinExplicitNullCheck);
|
|
}
|
|
|
|
// accessors
|
|
Value obj() const { return _obj; }
|
|
|
|
// setters
|
|
void set_can_trap(bool can_trap) { set_flag(CanTrapFlag, can_trap); }
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return check_flag(CanTrapFlag); /* null-check elimination sets to false */ }
|
|
virtual void input_values_do(ValueVisitor* f) { f->visit(&_obj); }
|
|
HASHING1(NullCheck, true, obj()->subst())
|
|
};
|
|
|
|
|
|
// This node is supposed to cast the type of another node to a more precise
|
|
// declared type.
|
|
LEAF(TypeCast, Instruction)
|
|
private:
|
|
ciType* _declared_type;
|
|
Value _obj;
|
|
|
|
public:
|
|
// The type of this node is the same type as the object type (and it might be constant).
|
|
TypeCast(ciType* type, Value obj, ValueStack* state_before)
|
|
: Instruction(obj->type(), state_before, obj->type()->is_constant()),
|
|
_declared_type(type),
|
|
_obj(obj) {}
|
|
|
|
// accessors
|
|
ciType* declared_type() const { return _declared_type; }
|
|
Value obj() const { return _obj; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { f->visit(&_obj); }
|
|
};
|
|
|
|
|
|
BASE(StateSplit, Instruction)
|
|
private:
|
|
ValueStack* _state;
|
|
|
|
protected:
|
|
static void substitute(BlockList& list, BlockBegin* old_block, BlockBegin* new_block);
|
|
|
|
public:
|
|
// creation
|
|
StateSplit(ValueType* type, ValueStack* state_before = NULL)
|
|
: Instruction(type, state_before)
|
|
, _state(NULL)
|
|
{
|
|
pin(PinStateSplitConstructor);
|
|
}
|
|
|
|
// accessors
|
|
ValueStack* state() const { return _state; }
|
|
IRScope* scope() const; // the state's scope
|
|
|
|
// manipulation
|
|
void set_state(ValueStack* state) { assert(_state == NULL, "overwriting existing state"); check_state(state); _state = state; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { /* no values */ }
|
|
virtual void state_values_do(ValueVisitor* f);
|
|
};
|
|
|
|
|
|
LEAF(Invoke, StateSplit)
|
|
private:
|
|
Bytecodes::Code _code;
|
|
Value _recv;
|
|
Values* _args;
|
|
BasicTypeList* _signature;
|
|
int _vtable_index;
|
|
ciMethod* _target;
|
|
|
|
public:
|
|
// creation
|
|
Invoke(Bytecodes::Code code, ValueType* result_type, Value recv, Values* args,
|
|
int vtable_index, ciMethod* target, ValueStack* state_before);
|
|
|
|
// accessors
|
|
Bytecodes::Code code() const { return _code; }
|
|
Value receiver() const { return _recv; }
|
|
bool has_receiver() const { return receiver() != NULL; }
|
|
int number_of_arguments() const { return _args->length(); }
|
|
Value argument_at(int i) const { return _args->at(i); }
|
|
int vtable_index() const { return _vtable_index; }
|
|
BasicTypeList* signature() const { return _signature; }
|
|
ciMethod* target() const { return _target; }
|
|
|
|
ciType* declared_type() const;
|
|
|
|
// Returns false if target is not loaded
|
|
bool target_is_final() const { return check_flag(TargetIsFinalFlag); }
|
|
bool target_is_loaded() const { return check_flag(TargetIsLoadedFlag); }
|
|
// Returns false if target is not loaded
|
|
bool target_is_strictfp() const { return check_flag(TargetIsStrictfpFlag); }
|
|
|
|
// JSR 292 support
|
|
bool is_invokedynamic() const { return code() == Bytecodes::_invokedynamic; }
|
|
bool is_method_handle_intrinsic() const { return target()->is_method_handle_intrinsic(); }
|
|
|
|
virtual bool needs_exception_state() const { return false; }
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return true; }
|
|
virtual void input_values_do(ValueVisitor* f) {
|
|
StateSplit::input_values_do(f);
|
|
if (has_receiver()) f->visit(&_recv);
|
|
for (int i = 0; i < _args->length(); i++) f->visit(_args->adr_at(i));
|
|
}
|
|
virtual void state_values_do(ValueVisitor *f);
|
|
};
|
|
|
|
|
|
LEAF(NewInstance, StateSplit)
|
|
private:
|
|
ciInstanceKlass* _klass;
|
|
bool _is_unresolved;
|
|
|
|
public:
|
|
// creation
|
|
NewInstance(ciInstanceKlass* klass, ValueStack* state_before, bool is_unresolved)
|
|
: StateSplit(instanceType, state_before)
|
|
, _klass(klass), _is_unresolved(is_unresolved)
|
|
{}
|
|
|
|
// accessors
|
|
ciInstanceKlass* klass() const { return _klass; }
|
|
bool is_unresolved() const { return _is_unresolved; }
|
|
|
|
virtual bool needs_exception_state() const { return false; }
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return true; }
|
|
ciType* exact_type() const;
|
|
ciType* declared_type() const;
|
|
};
|
|
|
|
|
|
BASE(NewArray, StateSplit)
|
|
private:
|
|
Value _length;
|
|
|
|
public:
|
|
// creation
|
|
NewArray(Value length, ValueStack* state_before)
|
|
: StateSplit(objectType, state_before)
|
|
, _length(length)
|
|
{
|
|
// Do not ASSERT_VALUES since length is NULL for NewMultiArray
|
|
}
|
|
|
|
// accessors
|
|
Value length() const { return _length; }
|
|
|
|
virtual bool needs_exception_state() const { return false; }
|
|
|
|
ciType* exact_type() const { return NULL; }
|
|
ciType* declared_type() const;
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return true; }
|
|
virtual void input_values_do(ValueVisitor* f) { StateSplit::input_values_do(f); f->visit(&_length); }
|
|
};
|
|
|
|
|
|
LEAF(NewTypeArray, NewArray)
|
|
private:
|
|
BasicType _elt_type;
|
|
|
|
public:
|
|
// creation
|
|
NewTypeArray(Value length, BasicType elt_type, ValueStack* state_before)
|
|
: NewArray(length, state_before)
|
|
, _elt_type(elt_type)
|
|
{}
|
|
|
|
// accessors
|
|
BasicType elt_type() const { return _elt_type; }
|
|
ciType* exact_type() const;
|
|
};
|
|
|
|
|
|
LEAF(NewObjectArray, NewArray)
|
|
private:
|
|
ciKlass* _klass;
|
|
|
|
public:
|
|
// creation
|
|
NewObjectArray(ciKlass* klass, Value length, ValueStack* state_before) : NewArray(length, state_before), _klass(klass) {}
|
|
|
|
// accessors
|
|
ciKlass* klass() const { return _klass; }
|
|
ciType* exact_type() const;
|
|
};
|
|
|
|
|
|
LEAF(NewMultiArray, NewArray)
|
|
private:
|
|
ciKlass* _klass;
|
|
Values* _dims;
|
|
|
|
public:
|
|
// creation
|
|
NewMultiArray(ciKlass* klass, Values* dims, ValueStack* state_before) : NewArray(NULL, state_before), _klass(klass), _dims(dims) {
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
ciKlass* klass() const { return _klass; }
|
|
Values* dims() const { return _dims; }
|
|
int rank() const { return dims()->length(); }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) {
|
|
// NOTE: we do not call NewArray::input_values_do since "length"
|
|
// is meaningless for a multi-dimensional array; passing the
|
|
// zeroth element down to NewArray as its length is a bad idea
|
|
// since there will be a copy in the "dims" array which doesn't
|
|
// get updated, and the value must not be traversed twice. Was bug
|
|
// - kbr 4/10/2001
|
|
StateSplit::input_values_do(f);
|
|
for (int i = 0; i < _dims->length(); i++) f->visit(_dims->adr_at(i));
|
|
}
|
|
};
|
|
|
|
|
|
BASE(TypeCheck, StateSplit)
|
|
private:
|
|
ciKlass* _klass;
|
|
Value _obj;
|
|
|
|
ciMethod* _profiled_method;
|
|
int _profiled_bci;
|
|
|
|
public:
|
|
// creation
|
|
TypeCheck(ciKlass* klass, Value obj, ValueType* type, ValueStack* state_before)
|
|
: StateSplit(type, state_before), _klass(klass), _obj(obj),
|
|
_profiled_method(NULL), _profiled_bci(0) {
|
|
ASSERT_VALUES
|
|
set_direct_compare(false);
|
|
}
|
|
|
|
// accessors
|
|
ciKlass* klass() const { return _klass; }
|
|
Value obj() const { return _obj; }
|
|
bool is_loaded() const { return klass() != NULL; }
|
|
bool direct_compare() const { return check_flag(DirectCompareFlag); }
|
|
|
|
// manipulation
|
|
void set_direct_compare(bool flag) { set_flag(DirectCompareFlag, flag); }
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return true; }
|
|
virtual void input_values_do(ValueVisitor* f) { StateSplit::input_values_do(f); f->visit(&_obj); }
|
|
|
|
// Helpers for MethodData* profiling
|
|
void set_should_profile(bool value) { set_flag(ProfileMDOFlag, value); }
|
|
void set_profiled_method(ciMethod* method) { _profiled_method = method; }
|
|
void set_profiled_bci(int bci) { _profiled_bci = bci; }
|
|
bool should_profile() const { return check_flag(ProfileMDOFlag); }
|
|
ciMethod* profiled_method() const { return _profiled_method; }
|
|
int profiled_bci() const { return _profiled_bci; }
|
|
};
|
|
|
|
|
|
LEAF(CheckCast, TypeCheck)
|
|
public:
|
|
// creation
|
|
CheckCast(ciKlass* klass, Value obj, ValueStack* state_before)
|
|
: TypeCheck(klass, obj, objectType, state_before) {}
|
|
|
|
void set_incompatible_class_change_check() {
|
|
set_flag(ThrowIncompatibleClassChangeErrorFlag, true);
|
|
}
|
|
bool is_incompatible_class_change_check() const {
|
|
return check_flag(ThrowIncompatibleClassChangeErrorFlag);
|
|
}
|
|
|
|
ciType* declared_type() const;
|
|
};
|
|
|
|
|
|
LEAF(InstanceOf, TypeCheck)
|
|
public:
|
|
// creation
|
|
InstanceOf(ciKlass* klass, Value obj, ValueStack* state_before) : TypeCheck(klass, obj, intType, state_before) {}
|
|
|
|
virtual bool needs_exception_state() const { return false; }
|
|
};
|
|
|
|
|
|
BASE(AccessMonitor, StateSplit)
|
|
private:
|
|
Value _obj;
|
|
int _monitor_no;
|
|
|
|
public:
|
|
// creation
|
|
AccessMonitor(Value obj, int monitor_no, ValueStack* state_before = NULL)
|
|
: StateSplit(illegalType, state_before)
|
|
, _obj(obj)
|
|
, _monitor_no(monitor_no)
|
|
{
|
|
set_needs_null_check(true);
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
Value obj() const { return _obj; }
|
|
int monitor_no() const { return _monitor_no; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { StateSplit::input_values_do(f); f->visit(&_obj); }
|
|
};
|
|
|
|
|
|
LEAF(MonitorEnter, AccessMonitor)
|
|
public:
|
|
// creation
|
|
MonitorEnter(Value obj, int monitor_no, ValueStack* state_before)
|
|
: AccessMonitor(obj, monitor_no, state_before)
|
|
{
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return true; }
|
|
};
|
|
|
|
|
|
LEAF(MonitorExit, AccessMonitor)
|
|
public:
|
|
// creation
|
|
MonitorExit(Value obj, int monitor_no)
|
|
: AccessMonitor(obj, monitor_no, NULL)
|
|
{
|
|
ASSERT_VALUES
|
|
}
|
|
};
|
|
|
|
|
|
LEAF(Intrinsic, StateSplit)
|
|
private:
|
|
vmIntrinsics::ID _id;
|
|
Values* _args;
|
|
Value _recv;
|
|
ArgsNonNullState _nonnull_state;
|
|
|
|
public:
|
|
// preserves_state can be set to true for Intrinsics
|
|
// which are guaranteed to preserve register state across any slow
|
|
// cases; setting it to true does not mean that the Intrinsic can
|
|
// not trap, only that if we continue execution in the same basic
|
|
// block after the Intrinsic, all of the registers are intact. This
|
|
// allows load elimination and common expression elimination to be
|
|
// performed across the Intrinsic. The default value is false.
|
|
Intrinsic(ValueType* type,
|
|
vmIntrinsics::ID id,
|
|
Values* args,
|
|
bool has_receiver,
|
|
ValueStack* state_before,
|
|
bool preserves_state,
|
|
bool cantrap = true)
|
|
: StateSplit(type, state_before)
|
|
, _id(id)
|
|
, _args(args)
|
|
, _recv(NULL)
|
|
{
|
|
assert(args != NULL, "args must exist");
|
|
ASSERT_VALUES
|
|
set_flag(PreservesStateFlag, preserves_state);
|
|
set_flag(CanTrapFlag, cantrap);
|
|
if (has_receiver) {
|
|
_recv = argument_at(0);
|
|
}
|
|
set_needs_null_check(has_receiver);
|
|
|
|
// some intrinsics can't trap, so don't force them to be pinned
|
|
if (!can_trap()) {
|
|
unpin(PinStateSplitConstructor);
|
|
}
|
|
}
|
|
|
|
// accessors
|
|
vmIntrinsics::ID id() const { return _id; }
|
|
int number_of_arguments() const { return _args->length(); }
|
|
Value argument_at(int i) const { return _args->at(i); }
|
|
|
|
bool has_receiver() const { return (_recv != NULL); }
|
|
Value receiver() const { assert(has_receiver(), "must have receiver"); return _recv; }
|
|
bool preserves_state() const { return check_flag(PreservesStateFlag); }
|
|
|
|
bool arg_needs_null_check(int i) const {
|
|
return _nonnull_state.arg_needs_null_check(i);
|
|
}
|
|
|
|
void set_arg_needs_null_check(int i, bool check) {
|
|
_nonnull_state.set_arg_needs_null_check(i, check);
|
|
}
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return check_flag(CanTrapFlag); }
|
|
virtual void input_values_do(ValueVisitor* f) {
|
|
StateSplit::input_values_do(f);
|
|
for (int i = 0; i < _args->length(); i++) f->visit(_args->adr_at(i));
|
|
}
|
|
};
|
|
|
|
|
|
class LIR_List;
|
|
|
|
LEAF(BlockBegin, StateSplit)
|
|
private:
|
|
int _block_id; // the unique block id
|
|
int _bci; // start-bci of block
|
|
int _depth_first_number; // number of this block in a depth-first ordering
|
|
int _linear_scan_number; // number of this block in linear-scan ordering
|
|
int _dominator_depth;
|
|
int _loop_depth; // the loop nesting level of this block
|
|
int _loop_index; // number of the innermost loop of this block
|
|
int _flags; // the flags associated with this block
|
|
|
|
// fields used by BlockListBuilder
|
|
int _total_preds; // number of predecessors found by BlockListBuilder
|
|
BitMap _stores_to_locals; // bit is set when a local variable is stored in the block
|
|
|
|
// SSA specific fields: (factor out later)
|
|
BlockList _successors; // the successors of this block
|
|
BlockList _predecessors; // the predecessors of this block
|
|
BlockList _dominates; // list of blocks that are dominated by this block
|
|
BlockBegin* _dominator; // the dominator of this block
|
|
// SSA specific ends
|
|
BlockEnd* _end; // the last instruction of this block
|
|
BlockList _exception_handlers; // the exception handlers potentially invoked by this block
|
|
ValueStackStack* _exception_states; // only for xhandler entries: states of all instructions that have an edge to this xhandler
|
|
int _exception_handler_pco; // if this block is the start of an exception handler,
|
|
// this records the PC offset in the assembly code of the
|
|
// first instruction in this block
|
|
Label _label; // the label associated with this block
|
|
LIR_List* _lir; // the low level intermediate representation for this block
|
|
|
|
BitMap _live_in; // set of live LIR_Opr registers at entry to this block
|
|
BitMap _live_out; // set of live LIR_Opr registers at exit from this block
|
|
BitMap _live_gen; // set of registers used before any redefinition in this block
|
|
BitMap _live_kill; // set of registers defined in this block
|
|
|
|
BitMap _fpu_register_usage;
|
|
intArray* _fpu_stack_state; // For x86 FPU code generation with UseLinearScan
|
|
int _first_lir_instruction_id; // ID of first LIR instruction in this block
|
|
int _last_lir_instruction_id; // ID of last LIR instruction in this block
|
|
|
|
void iterate_preorder (boolArray& mark, BlockClosure* closure);
|
|
void iterate_postorder(boolArray& mark, BlockClosure* closure);
|
|
|
|
friend class SuxAndWeightAdjuster;
|
|
|
|
public:
|
|
void* operator new(size_t size) throw() {
|
|
Compilation* c = Compilation::current();
|
|
void* res = c->arena()->Amalloc(size);
|
|
((BlockBegin*)res)->_id = c->get_next_id();
|
|
((BlockBegin*)res)->_block_id = c->get_next_block_id();
|
|
return res;
|
|
}
|
|
|
|
// initialization/counting
|
|
static int number_of_blocks() {
|
|
return Compilation::current()->number_of_blocks();
|
|
}
|
|
|
|
// creation
|
|
BlockBegin(int bci)
|
|
: StateSplit(illegalType)
|
|
, _bci(bci)
|
|
, _depth_first_number(-1)
|
|
, _linear_scan_number(-1)
|
|
, _loop_depth(0)
|
|
, _flags(0)
|
|
, _dominator_depth(-1)
|
|
, _dominator(NULL)
|
|
, _end(NULL)
|
|
, _predecessors(2)
|
|
, _successors(2)
|
|
, _dominates(2)
|
|
, _exception_handlers(1)
|
|
, _exception_states(NULL)
|
|
, _exception_handler_pco(-1)
|
|
, _lir(NULL)
|
|
, _loop_index(-1)
|
|
, _live_in()
|
|
, _live_out()
|
|
, _live_gen()
|
|
, _live_kill()
|
|
, _fpu_register_usage()
|
|
, _fpu_stack_state(NULL)
|
|
, _first_lir_instruction_id(-1)
|
|
, _last_lir_instruction_id(-1)
|
|
, _total_preds(0)
|
|
, _stores_to_locals()
|
|
{
|
|
_block = this;
|
|
#ifndef PRODUCT
|
|
set_printable_bci(bci);
|
|
#endif
|
|
}
|
|
|
|
// accessors
|
|
int block_id() const { return _block_id; }
|
|
int bci() const { return _bci; }
|
|
BlockList* successors() { return &_successors; }
|
|
BlockList* dominates() { return &_dominates; }
|
|
BlockBegin* dominator() const { return _dominator; }
|
|
int loop_depth() const { return _loop_depth; }
|
|
int dominator_depth() const { return _dominator_depth; }
|
|
int depth_first_number() const { return _depth_first_number; }
|
|
int linear_scan_number() const { return _linear_scan_number; }
|
|
BlockEnd* end() const { return _end; }
|
|
Label* label() { return &_label; }
|
|
LIR_List* lir() const { return _lir; }
|
|
int exception_handler_pco() const { return _exception_handler_pco; }
|
|
BitMap& live_in() { return _live_in; }
|
|
BitMap& live_out() { return _live_out; }
|
|
BitMap& live_gen() { return _live_gen; }
|
|
BitMap& live_kill() { return _live_kill; }
|
|
BitMap& fpu_register_usage() { return _fpu_register_usage; }
|
|
intArray* fpu_stack_state() const { return _fpu_stack_state; }
|
|
int first_lir_instruction_id() const { return _first_lir_instruction_id; }
|
|
int last_lir_instruction_id() const { return _last_lir_instruction_id; }
|
|
int total_preds() const { return _total_preds; }
|
|
BitMap& stores_to_locals() { return _stores_to_locals; }
|
|
|
|
// manipulation
|
|
void set_dominator(BlockBegin* dom) { _dominator = dom; }
|
|
void set_loop_depth(int d) { _loop_depth = d; }
|
|
void set_dominator_depth(int d) { _dominator_depth = d; }
|
|
void set_depth_first_number(int dfn) { _depth_first_number = dfn; }
|
|
void set_linear_scan_number(int lsn) { _linear_scan_number = lsn; }
|
|
void set_end(BlockEnd* end);
|
|
void clear_end();
|
|
void disconnect_from_graph();
|
|
static void disconnect_edge(BlockBegin* from, BlockBegin* to);
|
|
BlockBegin* insert_block_between(BlockBegin* sux);
|
|
void substitute_sux(BlockBegin* old_sux, BlockBegin* new_sux);
|
|
void set_lir(LIR_List* lir) { _lir = lir; }
|
|
void set_exception_handler_pco(int pco) { _exception_handler_pco = pco; }
|
|
void set_live_in (BitMap map) { _live_in = map; }
|
|
void set_live_out (BitMap map) { _live_out = map; }
|
|
void set_live_gen (BitMap map) { _live_gen = map; }
|
|
void set_live_kill (BitMap map) { _live_kill = map; }
|
|
void set_fpu_register_usage(BitMap map) { _fpu_register_usage = map; }
|
|
void set_fpu_stack_state(intArray* state) { _fpu_stack_state = state; }
|
|
void set_first_lir_instruction_id(int id) { _first_lir_instruction_id = id; }
|
|
void set_last_lir_instruction_id(int id) { _last_lir_instruction_id = id; }
|
|
void increment_total_preds(int n = 1) { _total_preds += n; }
|
|
void init_stores_to_locals(int locals_count) { _stores_to_locals = BitMap(locals_count); _stores_to_locals.clear(); }
|
|
|
|
// generic
|
|
virtual void state_values_do(ValueVisitor* f);
|
|
|
|
// successors and predecessors
|
|
int number_of_sux() const;
|
|
BlockBegin* sux_at(int i) const;
|
|
void add_successor(BlockBegin* sux);
|
|
void remove_successor(BlockBegin* pred);
|
|
bool is_successor(BlockBegin* sux) const { return _successors.contains(sux); }
|
|
|
|
void add_predecessor(BlockBegin* pred);
|
|
void remove_predecessor(BlockBegin* pred);
|
|
bool is_predecessor(BlockBegin* pred) const { return _predecessors.contains(pred); }
|
|
int number_of_preds() const { return _predecessors.length(); }
|
|
BlockBegin* pred_at(int i) const { return _predecessors[i]; }
|
|
|
|
// exception handlers potentially invoked by this block
|
|
void add_exception_handler(BlockBegin* b);
|
|
bool is_exception_handler(BlockBegin* b) const { return _exception_handlers.contains(b); }
|
|
int number_of_exception_handlers() const { return _exception_handlers.length(); }
|
|
BlockBegin* exception_handler_at(int i) const { return _exception_handlers.at(i); }
|
|
|
|
// states of the instructions that have an edge to this exception handler
|
|
int number_of_exception_states() { assert(is_set(exception_entry_flag), "only for xhandlers"); return _exception_states == NULL ? 0 : _exception_states->length(); }
|
|
ValueStack* exception_state_at(int idx) const { assert(is_set(exception_entry_flag), "only for xhandlers"); return _exception_states->at(idx); }
|
|
int add_exception_state(ValueStack* state);
|
|
|
|
// flags
|
|
enum Flag {
|
|
no_flag = 0,
|
|
std_entry_flag = 1 << 0,
|
|
osr_entry_flag = 1 << 1,
|
|
exception_entry_flag = 1 << 2,
|
|
subroutine_entry_flag = 1 << 3,
|
|
backward_branch_target_flag = 1 << 4,
|
|
is_on_work_list_flag = 1 << 5,
|
|
was_visited_flag = 1 << 6,
|
|
parser_loop_header_flag = 1 << 7, // set by parser to identify blocks where phi functions can not be created on demand
|
|
critical_edge_split_flag = 1 << 8, // set for all blocks that are introduced when critical edges are split
|
|
linear_scan_loop_header_flag = 1 << 9, // set during loop-detection for LinearScan
|
|
linear_scan_loop_end_flag = 1 << 10, // set during loop-detection for LinearScan
|
|
donot_eliminate_range_checks = 1 << 11 // Should be try to eliminate range checks in this block
|
|
};
|
|
|
|
void set(Flag f) { _flags |= f; }
|
|
void clear(Flag f) { _flags &= ~f; }
|
|
bool is_set(Flag f) const { return (_flags & f) != 0; }
|
|
bool is_entry_block() const {
|
|
const int entry_mask = std_entry_flag | osr_entry_flag | exception_entry_flag;
|
|
return (_flags & entry_mask) != 0;
|
|
}
|
|
|
|
// iteration
|
|
void iterate_preorder (BlockClosure* closure);
|
|
void iterate_postorder (BlockClosure* closure);
|
|
|
|
void block_values_do(ValueVisitor* f);
|
|
|
|
// loops
|
|
void set_loop_index(int ix) { _loop_index = ix; }
|
|
int loop_index() const { return _loop_index; }
|
|
|
|
// merging
|
|
bool try_merge(ValueStack* state); // try to merge states at block begin
|
|
void merge(ValueStack* state) { bool b = try_merge(state); assert(b, "merge failed"); }
|
|
|
|
// debugging
|
|
void print_block() PRODUCT_RETURN;
|
|
void print_block(InstructionPrinter& ip, bool live_only = false) PRODUCT_RETURN;
|
|
};
|
|
|
|
|
|
BASE(BlockEnd, StateSplit)
|
|
private:
|
|
BlockList* _sux;
|
|
|
|
protected:
|
|
BlockList* sux() const { return _sux; }
|
|
|
|
void set_sux(BlockList* sux) {
|
|
#ifdef ASSERT
|
|
assert(sux != NULL, "sux must exist");
|
|
for (int i = sux->length() - 1; i >= 0; i--) assert(sux->at(i) != NULL, "sux must exist");
|
|
#endif
|
|
_sux = sux;
|
|
}
|
|
|
|
public:
|
|
// creation
|
|
BlockEnd(ValueType* type, ValueStack* state_before, bool is_safepoint)
|
|
: StateSplit(type, state_before)
|
|
, _sux(NULL)
|
|
{
|
|
set_flag(IsSafepointFlag, is_safepoint);
|
|
}
|
|
|
|
// accessors
|
|
bool is_safepoint() const { return check_flag(IsSafepointFlag); }
|
|
// For compatibility with old code, for new code use block()
|
|
BlockBegin* begin() const { return _block; }
|
|
|
|
// manipulation
|
|
void set_begin(BlockBegin* begin);
|
|
|
|
// successors
|
|
int number_of_sux() const { return _sux != NULL ? _sux->length() : 0; }
|
|
BlockBegin* sux_at(int i) const { return _sux->at(i); }
|
|
BlockBegin* default_sux() const { return sux_at(number_of_sux() - 1); }
|
|
BlockBegin** addr_sux_at(int i) const { return _sux->adr_at(i); }
|
|
int sux_index(BlockBegin* sux) const { return _sux->find(sux); }
|
|
void substitute_sux(BlockBegin* old_sux, BlockBegin* new_sux);
|
|
};
|
|
|
|
|
|
LEAF(Goto, BlockEnd)
|
|
public:
|
|
enum Direction {
|
|
none, // Just a regular goto
|
|
taken, not_taken // Goto produced from If
|
|
};
|
|
private:
|
|
ciMethod* _profiled_method;
|
|
int _profiled_bci;
|
|
Direction _direction;
|
|
public:
|
|
// creation
|
|
Goto(BlockBegin* sux, ValueStack* state_before, bool is_safepoint = false)
|
|
: BlockEnd(illegalType, state_before, is_safepoint)
|
|
, _direction(none)
|
|
, _profiled_method(NULL)
|
|
, _profiled_bci(0) {
|
|
BlockList* s = new BlockList(1);
|
|
s->append(sux);
|
|
set_sux(s);
|
|
}
|
|
|
|
Goto(BlockBegin* sux, bool is_safepoint) : BlockEnd(illegalType, NULL, is_safepoint)
|
|
, _direction(none)
|
|
, _profiled_method(NULL)
|
|
, _profiled_bci(0) {
|
|
BlockList* s = new BlockList(1);
|
|
s->append(sux);
|
|
set_sux(s);
|
|
}
|
|
|
|
bool should_profile() const { return check_flag(ProfileMDOFlag); }
|
|
ciMethod* profiled_method() const { return _profiled_method; } // set only for profiled branches
|
|
int profiled_bci() const { return _profiled_bci; }
|
|
Direction direction() const { return _direction; }
|
|
|
|
void set_should_profile(bool value) { set_flag(ProfileMDOFlag, value); }
|
|
void set_profiled_method(ciMethod* method) { _profiled_method = method; }
|
|
void set_profiled_bci(int bci) { _profiled_bci = bci; }
|
|
void set_direction(Direction d) { _direction = d; }
|
|
};
|
|
|
|
#ifdef ASSERT
|
|
LEAF(Assert, Instruction)
|
|
private:
|
|
Value _x;
|
|
Condition _cond;
|
|
Value _y;
|
|
char *_message;
|
|
|
|
public:
|
|
// creation
|
|
// unordered_is_true is valid for float/double compares only
|
|
Assert(Value x, Condition cond, bool unordered_is_true, Value y);
|
|
|
|
// accessors
|
|
Value x() const { return _x; }
|
|
Condition cond() const { return _cond; }
|
|
bool unordered_is_true() const { return check_flag(UnorderedIsTrueFlag); }
|
|
Value y() const { return _y; }
|
|
const char *message() const { return _message; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { f->visit(&_x); f->visit(&_y); }
|
|
};
|
|
#endif
|
|
|
|
LEAF(RangeCheckPredicate, StateSplit)
|
|
private:
|
|
Value _x;
|
|
Condition _cond;
|
|
Value _y;
|
|
|
|
void check_state();
|
|
|
|
public:
|
|
// creation
|
|
// unordered_is_true is valid for float/double compares only
|
|
RangeCheckPredicate(Value x, Condition cond, bool unordered_is_true, Value y, ValueStack* state) : StateSplit(illegalType)
|
|
, _x(x)
|
|
, _cond(cond)
|
|
, _y(y)
|
|
{
|
|
ASSERT_VALUES
|
|
set_flag(UnorderedIsTrueFlag, unordered_is_true);
|
|
assert(x->type()->tag() == y->type()->tag(), "types must match");
|
|
this->set_state(state);
|
|
check_state();
|
|
}
|
|
|
|
// Always deoptimize
|
|
RangeCheckPredicate(ValueStack* state) : StateSplit(illegalType)
|
|
{
|
|
this->set_state(state);
|
|
_x = _y = NULL;
|
|
check_state();
|
|
}
|
|
|
|
// accessors
|
|
Value x() const { return _x; }
|
|
Condition cond() const { return _cond; }
|
|
bool unordered_is_true() const { return check_flag(UnorderedIsTrueFlag); }
|
|
Value y() const { return _y; }
|
|
|
|
void always_fail() { _x = _y = NULL; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { StateSplit::input_values_do(f); f->visit(&_x); f->visit(&_y); }
|
|
HASHING3(RangeCheckPredicate, true, x()->subst(), y()->subst(), cond())
|
|
};
|
|
|
|
LEAF(If, BlockEnd)
|
|
private:
|
|
Value _x;
|
|
Condition _cond;
|
|
Value _y;
|
|
ciMethod* _profiled_method;
|
|
int _profiled_bci; // Canonicalizer may alter bci of If node
|
|
bool _swapped; // Is the order reversed with respect to the original If in the
|
|
// bytecode stream?
|
|
public:
|
|
// creation
|
|
// unordered_is_true is valid for float/double compares only
|
|
If(Value x, Condition cond, bool unordered_is_true, Value y, BlockBegin* tsux, BlockBegin* fsux, ValueStack* state_before, bool is_safepoint)
|
|
: BlockEnd(illegalType, state_before, is_safepoint)
|
|
, _x(x)
|
|
, _cond(cond)
|
|
, _y(y)
|
|
, _profiled_method(NULL)
|
|
, _profiled_bci(0)
|
|
, _swapped(false)
|
|
{
|
|
ASSERT_VALUES
|
|
set_flag(UnorderedIsTrueFlag, unordered_is_true);
|
|
assert(x->type()->tag() == y->type()->tag(), "types must match");
|
|
BlockList* s = new BlockList(2);
|
|
s->append(tsux);
|
|
s->append(fsux);
|
|
set_sux(s);
|
|
}
|
|
|
|
// accessors
|
|
Value x() const { return _x; }
|
|
Condition cond() const { return _cond; }
|
|
bool unordered_is_true() const { return check_flag(UnorderedIsTrueFlag); }
|
|
Value y() const { return _y; }
|
|
BlockBegin* sux_for(bool is_true) const { return sux_at(is_true ? 0 : 1); }
|
|
BlockBegin* tsux() const { return sux_for(true); }
|
|
BlockBegin* fsux() const { return sux_for(false); }
|
|
BlockBegin* usux() const { return sux_for(unordered_is_true()); }
|
|
bool should_profile() const { return check_flag(ProfileMDOFlag); }
|
|
ciMethod* profiled_method() const { return _profiled_method; } // set only for profiled branches
|
|
int profiled_bci() const { return _profiled_bci; } // set for profiled branches and tiered
|
|
bool is_swapped() const { return _swapped; }
|
|
|
|
// manipulation
|
|
void swap_operands() {
|
|
Value t = _x; _x = _y; _y = t;
|
|
_cond = mirror(_cond);
|
|
}
|
|
|
|
void swap_sux() {
|
|
assert(number_of_sux() == 2, "wrong number of successors");
|
|
BlockList* s = sux();
|
|
BlockBegin* t = s->at(0); s->at_put(0, s->at(1)); s->at_put(1, t);
|
|
_cond = negate(_cond);
|
|
set_flag(UnorderedIsTrueFlag, !check_flag(UnorderedIsTrueFlag));
|
|
}
|
|
|
|
void set_should_profile(bool value) { set_flag(ProfileMDOFlag, value); }
|
|
void set_profiled_method(ciMethod* method) { _profiled_method = method; }
|
|
void set_profiled_bci(int bci) { _profiled_bci = bci; }
|
|
void set_swapped(bool value) { _swapped = value; }
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { BlockEnd::input_values_do(f); f->visit(&_x); f->visit(&_y); }
|
|
};
|
|
|
|
|
|
LEAF(IfInstanceOf, BlockEnd)
|
|
private:
|
|
ciKlass* _klass;
|
|
Value _obj;
|
|
bool _test_is_instance; // jump if instance
|
|
int _instanceof_bci;
|
|
|
|
public:
|
|
IfInstanceOf(ciKlass* klass, Value obj, bool test_is_instance, int instanceof_bci, BlockBegin* tsux, BlockBegin* fsux)
|
|
: BlockEnd(illegalType, NULL, false) // temporary set to false
|
|
, _klass(klass)
|
|
, _obj(obj)
|
|
, _test_is_instance(test_is_instance)
|
|
, _instanceof_bci(instanceof_bci)
|
|
{
|
|
ASSERT_VALUES
|
|
assert(instanceof_bci >= 0, "illegal bci");
|
|
BlockList* s = new BlockList(2);
|
|
s->append(tsux);
|
|
s->append(fsux);
|
|
set_sux(s);
|
|
}
|
|
|
|
// accessors
|
|
//
|
|
// Note 1: If test_is_instance() is true, IfInstanceOf tests if obj *is* an
|
|
// instance of klass; otherwise it tests if it is *not* and instance
|
|
// of klass.
|
|
//
|
|
// Note 2: IfInstanceOf instructions are created by combining an InstanceOf
|
|
// and an If instruction. The IfInstanceOf bci() corresponds to the
|
|
// bci that the If would have had; the (this->) instanceof_bci() is
|
|
// the bci of the original InstanceOf instruction.
|
|
ciKlass* klass() const { return _klass; }
|
|
Value obj() const { return _obj; }
|
|
int instanceof_bci() const { return _instanceof_bci; }
|
|
bool test_is_instance() const { return _test_is_instance; }
|
|
BlockBegin* sux_for(bool is_true) const { return sux_at(is_true ? 0 : 1); }
|
|
BlockBegin* tsux() const { return sux_for(true); }
|
|
BlockBegin* fsux() const { return sux_for(false); }
|
|
|
|
// manipulation
|
|
void swap_sux() {
|
|
assert(number_of_sux() == 2, "wrong number of successors");
|
|
BlockList* s = sux();
|
|
BlockBegin* t = s->at(0); s->at_put(0, s->at(1)); s->at_put(1, t);
|
|
_test_is_instance = !_test_is_instance;
|
|
}
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { BlockEnd::input_values_do(f); f->visit(&_obj); }
|
|
};
|
|
|
|
|
|
BASE(Switch, BlockEnd)
|
|
private:
|
|
Value _tag;
|
|
|
|
public:
|
|
// creation
|
|
Switch(Value tag, BlockList* sux, ValueStack* state_before, bool is_safepoint)
|
|
: BlockEnd(illegalType, state_before, is_safepoint)
|
|
, _tag(tag) {
|
|
ASSERT_VALUES
|
|
set_sux(sux);
|
|
}
|
|
|
|
// accessors
|
|
Value tag() const { return _tag; }
|
|
int length() const { return number_of_sux() - 1; }
|
|
|
|
virtual bool needs_exception_state() const { return false; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { BlockEnd::input_values_do(f); f->visit(&_tag); }
|
|
};
|
|
|
|
|
|
LEAF(TableSwitch, Switch)
|
|
private:
|
|
int _lo_key;
|
|
|
|
public:
|
|
// creation
|
|
TableSwitch(Value tag, BlockList* sux, int lo_key, ValueStack* state_before, bool is_safepoint)
|
|
: Switch(tag, sux, state_before, is_safepoint)
|
|
, _lo_key(lo_key) {}
|
|
|
|
// accessors
|
|
int lo_key() const { return _lo_key; }
|
|
int hi_key() const { return _lo_key + length() - 1; }
|
|
};
|
|
|
|
|
|
LEAF(LookupSwitch, Switch)
|
|
private:
|
|
intArray* _keys;
|
|
|
|
public:
|
|
// creation
|
|
LookupSwitch(Value tag, BlockList* sux, intArray* keys, ValueStack* state_before, bool is_safepoint)
|
|
: Switch(tag, sux, state_before, is_safepoint)
|
|
, _keys(keys) {
|
|
assert(keys != NULL, "keys must exist");
|
|
assert(keys->length() == length(), "sux & keys have incompatible lengths");
|
|
}
|
|
|
|
// accessors
|
|
int key_at(int i) const { return _keys->at(i); }
|
|
};
|
|
|
|
|
|
LEAF(Return, BlockEnd)
|
|
private:
|
|
Value _result;
|
|
|
|
public:
|
|
// creation
|
|
Return(Value result) :
|
|
BlockEnd(result == NULL ? voidType : result->type()->base(), NULL, true),
|
|
_result(result) {}
|
|
|
|
// accessors
|
|
Value result() const { return _result; }
|
|
bool has_result() const { return result() != NULL; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) {
|
|
BlockEnd::input_values_do(f);
|
|
if (has_result()) f->visit(&_result);
|
|
}
|
|
};
|
|
|
|
|
|
LEAF(Throw, BlockEnd)
|
|
private:
|
|
Value _exception;
|
|
|
|
public:
|
|
// creation
|
|
Throw(Value exception, ValueStack* state_before) : BlockEnd(illegalType, state_before, true), _exception(exception) {
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
Value exception() const { return _exception; }
|
|
|
|
// generic
|
|
virtual bool can_trap() const { return true; }
|
|
virtual void input_values_do(ValueVisitor* f) { BlockEnd::input_values_do(f); f->visit(&_exception); }
|
|
};
|
|
|
|
|
|
LEAF(Base, BlockEnd)
|
|
public:
|
|
// creation
|
|
Base(BlockBegin* std_entry, BlockBegin* osr_entry) : BlockEnd(illegalType, NULL, false) {
|
|
assert(std_entry->is_set(BlockBegin::std_entry_flag), "std entry must be flagged");
|
|
assert(osr_entry == NULL || osr_entry->is_set(BlockBegin::osr_entry_flag), "osr entry must be flagged");
|
|
BlockList* s = new BlockList(2);
|
|
if (osr_entry != NULL) s->append(osr_entry);
|
|
s->append(std_entry); // must be default sux!
|
|
set_sux(s);
|
|
}
|
|
|
|
// accessors
|
|
BlockBegin* std_entry() const { return default_sux(); }
|
|
BlockBegin* osr_entry() const { return number_of_sux() < 2 ? NULL : sux_at(0); }
|
|
};
|
|
|
|
|
|
LEAF(OsrEntry, Instruction)
|
|
public:
|
|
// creation
|
|
#ifdef _LP64
|
|
OsrEntry() : Instruction(longType) { pin(); }
|
|
#else
|
|
OsrEntry() : Instruction(intType) { pin(); }
|
|
#endif
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { }
|
|
};
|
|
|
|
|
|
// Models the incoming exception at a catch site
|
|
LEAF(ExceptionObject, Instruction)
|
|
public:
|
|
// creation
|
|
ExceptionObject() : Instruction(objectType) {
|
|
pin();
|
|
}
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { }
|
|
};
|
|
|
|
|
|
// Models needed rounding for floating-point values on Intel.
|
|
// Currently only used to represent rounding of double-precision
|
|
// values stored into local variables, but could be used to model
|
|
// intermediate rounding of single-precision values as well.
|
|
LEAF(RoundFP, Instruction)
|
|
private:
|
|
Value _input; // floating-point value to be rounded
|
|
|
|
public:
|
|
RoundFP(Value input)
|
|
: Instruction(input->type()) // Note: should not be used for constants
|
|
, _input(input)
|
|
{
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
Value input() const { return _input; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { f->visit(&_input); }
|
|
};
|
|
|
|
|
|
BASE(UnsafeOp, Instruction)
|
|
private:
|
|
BasicType _basic_type; // ValueType can not express byte-sized integers
|
|
|
|
protected:
|
|
// creation
|
|
UnsafeOp(BasicType basic_type, bool is_put)
|
|
: Instruction(is_put ? voidType : as_ValueType(basic_type))
|
|
, _basic_type(basic_type)
|
|
{
|
|
//Note: Unsafe ops are not not guaranteed to throw NPE.
|
|
// Convservatively, Unsafe operations must be pinned though we could be
|
|
// looser about this if we wanted to..
|
|
pin();
|
|
}
|
|
|
|
public:
|
|
// accessors
|
|
BasicType basic_type() { return _basic_type; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { }
|
|
};
|
|
|
|
|
|
BASE(UnsafeRawOp, UnsafeOp)
|
|
private:
|
|
Value _base; // Base address (a Java long)
|
|
Value _index; // Index if computed by optimizer; initialized to NULL
|
|
int _log2_scale; // Scale factor: 0, 1, 2, or 3.
|
|
// Indicates log2 of number of bytes (1, 2, 4, or 8)
|
|
// to scale index by.
|
|
|
|
protected:
|
|
UnsafeRawOp(BasicType basic_type, Value addr, bool is_put)
|
|
: UnsafeOp(basic_type, is_put)
|
|
, _base(addr)
|
|
, _index(NULL)
|
|
, _log2_scale(0)
|
|
{
|
|
// Can not use ASSERT_VALUES because index may be NULL
|
|
assert(addr != NULL && addr->type()->is_long(), "just checking");
|
|
}
|
|
|
|
UnsafeRawOp(BasicType basic_type, Value base, Value index, int log2_scale, bool is_put)
|
|
: UnsafeOp(basic_type, is_put)
|
|
, _base(base)
|
|
, _index(index)
|
|
, _log2_scale(log2_scale)
|
|
{
|
|
}
|
|
|
|
public:
|
|
// accessors
|
|
Value base() { return _base; }
|
|
Value index() { return _index; }
|
|
bool has_index() { return (_index != NULL); }
|
|
int log2_scale() { return _log2_scale; }
|
|
|
|
// setters
|
|
void set_base (Value base) { _base = base; }
|
|
void set_index(Value index) { _index = index; }
|
|
void set_log2_scale(int log2_scale) { _log2_scale = log2_scale; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { UnsafeOp::input_values_do(f);
|
|
f->visit(&_base);
|
|
if (has_index()) f->visit(&_index); }
|
|
};
|
|
|
|
|
|
LEAF(UnsafeGetRaw, UnsafeRawOp)
|
|
private:
|
|
bool _may_be_unaligned, _is_wide; // For OSREntry
|
|
|
|
public:
|
|
UnsafeGetRaw(BasicType basic_type, Value addr, bool may_be_unaligned, bool is_wide = false)
|
|
: UnsafeRawOp(basic_type, addr, false) {
|
|
_may_be_unaligned = may_be_unaligned;
|
|
_is_wide = is_wide;
|
|
}
|
|
|
|
UnsafeGetRaw(BasicType basic_type, Value base, Value index, int log2_scale, bool may_be_unaligned, bool is_wide = false)
|
|
: UnsafeRawOp(basic_type, base, index, log2_scale, false) {
|
|
_may_be_unaligned = may_be_unaligned;
|
|
_is_wide = is_wide;
|
|
}
|
|
|
|
bool may_be_unaligned() { return _may_be_unaligned; }
|
|
bool is_wide() { return _is_wide; }
|
|
};
|
|
|
|
|
|
LEAF(UnsafePutRaw, UnsafeRawOp)
|
|
private:
|
|
Value _value; // Value to be stored
|
|
|
|
public:
|
|
UnsafePutRaw(BasicType basic_type, Value addr, Value value)
|
|
: UnsafeRawOp(basic_type, addr, true)
|
|
, _value(value)
|
|
{
|
|
assert(value != NULL, "just checking");
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
UnsafePutRaw(BasicType basic_type, Value base, Value index, int log2_scale, Value value)
|
|
: UnsafeRawOp(basic_type, base, index, log2_scale, true)
|
|
, _value(value)
|
|
{
|
|
assert(value != NULL, "just checking");
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
Value value() { return _value; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { UnsafeRawOp::input_values_do(f);
|
|
f->visit(&_value); }
|
|
};
|
|
|
|
|
|
BASE(UnsafeObjectOp, UnsafeOp)
|
|
private:
|
|
Value _object; // Object to be fetched from or mutated
|
|
Value _offset; // Offset within object
|
|
bool _is_volatile; // true if volatile - dl/JSR166
|
|
public:
|
|
UnsafeObjectOp(BasicType basic_type, Value object, Value offset, bool is_put, bool is_volatile)
|
|
: UnsafeOp(basic_type, is_put), _object(object), _offset(offset), _is_volatile(is_volatile)
|
|
{
|
|
}
|
|
|
|
// accessors
|
|
Value object() { return _object; }
|
|
Value offset() { return _offset; }
|
|
bool is_volatile() { return _is_volatile; }
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { UnsafeOp::input_values_do(f);
|
|
f->visit(&_object);
|
|
f->visit(&_offset); }
|
|
};
|
|
|
|
|
|
LEAF(UnsafeGetObject, UnsafeObjectOp)
|
|
public:
|
|
UnsafeGetObject(BasicType basic_type, Value object, Value offset, bool is_volatile)
|
|
: UnsafeObjectOp(basic_type, object, offset, false, is_volatile)
|
|
{
|
|
ASSERT_VALUES
|
|
}
|
|
};
|
|
|
|
|
|
LEAF(UnsafePutObject, UnsafeObjectOp)
|
|
private:
|
|
Value _value; // Value to be stored
|
|
public:
|
|
UnsafePutObject(BasicType basic_type, Value object, Value offset, Value value, bool is_volatile)
|
|
: UnsafeObjectOp(basic_type, object, offset, true, is_volatile)
|
|
, _value(value)
|
|
{
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
Value value() { return _value; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { UnsafeObjectOp::input_values_do(f);
|
|
f->visit(&_value); }
|
|
};
|
|
|
|
LEAF(UnsafeGetAndSetObject, UnsafeObjectOp)
|
|
private:
|
|
Value _value; // Value to be stored
|
|
bool _is_add;
|
|
public:
|
|
UnsafeGetAndSetObject(BasicType basic_type, Value object, Value offset, Value value, bool is_add)
|
|
: UnsafeObjectOp(basic_type, object, offset, false, false)
|
|
, _value(value)
|
|
, _is_add(is_add)
|
|
{
|
|
ASSERT_VALUES
|
|
}
|
|
|
|
// accessors
|
|
bool is_add() const { return _is_add; }
|
|
Value value() { return _value; }
|
|
|
|
// generic
|
|
virtual void input_values_do(ValueVisitor* f) { UnsafeObjectOp::input_values_do(f);
|
|
f->visit(&_value); }
|
|
};
|
|
|
|
LEAF(ProfileCall, Instruction)
|
|
private:
|
|
ciMethod* _method;
|
|
int _bci_of_invoke;
|
|
ciMethod* _callee; // the method that is called at the given bci
|
|
Value _recv;
|
|
ciKlass* _known_holder;
|
|
Values* _obj_args; // arguments for type profiling
|
|
ArgsNonNullState _nonnull_state; // Do we know whether some arguments are never null?
|
|
bool _inlined; // Are we profiling a call that is inlined
|
|
|
|
public:
|
|
ProfileCall(ciMethod* method, int bci, ciMethod* callee, Value recv, ciKlass* known_holder, Values* obj_args, bool inlined)
|
|
: Instruction(voidType)
|
|
, _method(method)
|
|
, _bci_of_invoke(bci)
|
|
, _callee(callee)
|
|
, _recv(recv)
|
|
, _known_holder(known_holder)
|
|
, _obj_args(obj_args)
|
|
, _inlined(inlined)
|
|
{
|
|
// The ProfileCall has side-effects and must occur precisely where located
|
|
pin();
|
|
}
|
|
|
|
ciMethod* method() const { return _method; }
|
|
int bci_of_invoke() const { return _bci_of_invoke; }
|
|
ciMethod* callee() const { return _callee; }
|
|
Value recv() const { return _recv; }
|
|
ciKlass* known_holder() const { return _known_holder; }
|
|
int nb_profiled_args() const { return _obj_args == NULL ? 0 : _obj_args->length(); }
|
|
Value profiled_arg_at(int i) const { return _obj_args->at(i); }
|
|
bool arg_needs_null_check(int i) const {
|
|
return _nonnull_state.arg_needs_null_check(i);
|
|
}
|
|
bool inlined() const { return _inlined; }
|
|
|
|
void set_arg_needs_null_check(int i, bool check) {
|
|
_nonnull_state.set_arg_needs_null_check(i, check);
|
|
}
|
|
|
|
virtual void input_values_do(ValueVisitor* f) {
|
|
if (_recv != NULL) {
|
|
f->visit(&_recv);
|
|
}
|
|
for (int i = 0; i < nb_profiled_args(); i++) {
|
|
f->visit(_obj_args->adr_at(i));
|
|
}
|
|
}
|
|
};
|
|
|
|
LEAF(ProfileReturnType, Instruction)
|
|
private:
|
|
ciMethod* _method;
|
|
ciMethod* _callee;
|
|
int _bci_of_invoke;
|
|
Value _ret;
|
|
|
|
public:
|
|
ProfileReturnType(ciMethod* method, int bci, ciMethod* callee, Value ret)
|
|
: Instruction(voidType)
|
|
, _method(method)
|
|
, _callee(callee)
|
|
, _bci_of_invoke(bci)
|
|
, _ret(ret)
|
|
{
|
|
set_needs_null_check(true);
|
|
// The ProfileType has side-effects and must occur precisely where located
|
|
pin();
|
|
}
|
|
|
|
ciMethod* method() const { return _method; }
|
|
ciMethod* callee() const { return _callee; }
|
|
int bci_of_invoke() const { return _bci_of_invoke; }
|
|
Value ret() const { return _ret; }
|
|
|
|
virtual void input_values_do(ValueVisitor* f) {
|
|
if (_ret != NULL) {
|
|
f->visit(&_ret);
|
|
}
|
|
}
|
|
};
|
|
|
|
// Call some C runtime function that doesn't safepoint,
|
|
// optionally passing the current thread as the first argument.
|
|
LEAF(RuntimeCall, Instruction)
|
|
private:
|
|
const char* _entry_name;
|
|
address _entry;
|
|
Values* _args;
|
|
bool _pass_thread; // Pass the JavaThread* as an implicit first argument
|
|
|
|
public:
|
|
RuntimeCall(ValueType* type, const char* entry_name, address entry, Values* args, bool pass_thread = true)
|
|
: Instruction(type)
|
|
, _entry(entry)
|
|
, _args(args)
|
|
, _entry_name(entry_name)
|
|
, _pass_thread(pass_thread) {
|
|
ASSERT_VALUES
|
|
pin();
|
|
}
|
|
|
|
const char* entry_name() const { return _entry_name; }
|
|
address entry() const { return _entry; }
|
|
int number_of_arguments() const { return _args->length(); }
|
|
Value argument_at(int i) const { return _args->at(i); }
|
|
bool pass_thread() const { return _pass_thread; }
|
|
|
|
virtual void input_values_do(ValueVisitor* f) {
|
|
for (int i = 0; i < _args->length(); i++) f->visit(_args->adr_at(i));
|
|
}
|
|
};
|
|
|
|
// Use to trip invocation counter of an inlined method
|
|
|
|
LEAF(ProfileInvoke, Instruction)
|
|
private:
|
|
ciMethod* _inlinee;
|
|
ValueStack* _state;
|
|
|
|
public:
|
|
ProfileInvoke(ciMethod* inlinee, ValueStack* state)
|
|
: Instruction(voidType)
|
|
, _inlinee(inlinee)
|
|
, _state(state)
|
|
{
|
|
// The ProfileInvoke has side-effects and must occur precisely where located QQQ???
|
|
pin();
|
|
}
|
|
|
|
ciMethod* inlinee() { return _inlinee; }
|
|
ValueStack* state() { return _state; }
|
|
virtual void input_values_do(ValueVisitor*) {}
|
|
virtual void state_values_do(ValueVisitor*);
|
|
};
|
|
|
|
LEAF(MemBar, Instruction)
|
|
private:
|
|
LIR_Code _code;
|
|
|
|
public:
|
|
MemBar(LIR_Code code)
|
|
: Instruction(voidType)
|
|
, _code(code)
|
|
{
|
|
pin();
|
|
}
|
|
|
|
LIR_Code code() { return _code; }
|
|
|
|
virtual void input_values_do(ValueVisitor*) {}
|
|
};
|
|
|
|
class BlockPair: public CompilationResourceObj {
|
|
private:
|
|
BlockBegin* _from;
|
|
BlockBegin* _to;
|
|
public:
|
|
BlockPair(BlockBegin* from, BlockBegin* to): _from(from), _to(to) {}
|
|
BlockBegin* from() const { return _from; }
|
|
BlockBegin* to() const { return _to; }
|
|
bool is_same(BlockBegin* from, BlockBegin* to) const { return _from == from && _to == to; }
|
|
bool is_same(BlockPair* p) const { return _from == p->from() && _to == p->to(); }
|
|
void set_to(BlockBegin* b) { _to = b; }
|
|
void set_from(BlockBegin* b) { _from = b; }
|
|
};
|
|
|
|
|
|
define_array(BlockPairArray, BlockPair*)
|
|
define_stack(BlockPairList, BlockPairArray)
|
|
|
|
|
|
inline int BlockBegin::number_of_sux() const { assert(_end == NULL || _end->number_of_sux() == _successors.length(), "mismatch"); return _successors.length(); }
|
|
inline BlockBegin* BlockBegin::sux_at(int i) const { assert(_end == NULL || _end->sux_at(i) == _successors.at(i), "mismatch"); return _successors.at(i); }
|
|
inline void BlockBegin::add_successor(BlockBegin* sux) { assert(_end == NULL, "Would create mismatch with successors of BlockEnd"); _successors.append(sux); }
|
|
|
|
#undef ASSERT_VALUES
|
|
|
|
#endif // SHARE_VM_C1_C1_INSTRUCTION_HPP
|