8181211: C2: Use profiling data to optimize on/off heap unsafe accesses

Reviewed-by: kvn
This commit is contained in:
Roland Westrelin 2017-06-09 10:51:52 +02:00
parent 0d3624a309
commit 698444f968
16 changed files with 233 additions and 90 deletions

View File

@ -1540,13 +1540,7 @@ void GraphBuilder::method_return(Value x, bool ignore_return) {
ciMethod* caller = state()->scope()->method();
ciMethodData* md = caller->method_data_or_null();
ciProfileData* data = md->bci_to_data(invoke_bci);
if (data != NULL && (data->is_CallTypeData() || data->is_VirtualCallTypeData())) {
bool has_return = data->is_CallTypeData() ? ((ciCallTypeData*)data)->has_return() : ((ciVirtualCallTypeData*)data)->has_return();
// May not be true in case of an inlined call through a method handle intrinsic.
if (has_return) {
profile_return_type(x, method(), caller, invoke_bci);
}
}
profile_return_type(x, method(), caller, invoke_bci);
}
}
Goto* goto_callee = new Goto(continuation(), false);
@ -4366,7 +4360,10 @@ void GraphBuilder::profile_return_type(Value ret, ciMethod* callee, ciMethod* m,
ciMethodData* md = m->method_data_or_null();
ciProfileData* data = md->bci_to_data(invoke_bci);
if (data != NULL && (data->is_CallTypeData() || data->is_VirtualCallTypeData())) {
append(new ProfileReturnType(m , invoke_bci, callee, ret));
bool has_return = data->is_CallTypeData() ? ((ciCallTypeData*)data)->has_return() : ((ciVirtualCallTypeData*)data)->has_return();
if (has_return) {
append(new ProfileReturnType(m , invoke_bci, callee, ret));
}
}
}

View File

@ -594,11 +594,11 @@ void ciMethod::assert_call_type_ok(int bci) {
* @param [in]bci bci of the call
* @param [in]i argument number
* @param [out]type profiled type of argument, NULL if none
* @param [out]maybe_null true if null was seen for argument
* @param [out]ptr_kind whether always null, never null or maybe null
* @return true if profiling exists
*
*/
bool ciMethod::argument_profiled_type(int bci, int i, ciKlass*& type, bool& maybe_null) {
bool ciMethod::argument_profiled_type(int bci, int i, ciKlass*& type, ProfilePtrKind& ptr_kind) {
if (MethodData::profile_parameters() && method_data() != NULL && method_data()->is_mature()) {
ciProfileData* data = method_data()->bci_to_data(bci);
if (data != NULL) {
@ -609,7 +609,7 @@ bool ciMethod::argument_profiled_type(int bci, int i, ciKlass*& type, bool& mayb
return false;
}
type = call->valid_argument_type(i);
maybe_null = call->argument_maybe_null(i);
ptr_kind = call->argument_ptr_kind(i);
return true;
} else if (data->is_CallTypeData()) {
assert_call_type_ok(bci);
@ -618,7 +618,7 @@ bool ciMethod::argument_profiled_type(int bci, int i, ciKlass*& type, bool& mayb
return false;
}
type = call->valid_argument_type(i);
maybe_null = call->argument_maybe_null(i);
ptr_kind = call->argument_ptr_kind(i);
return true;
}
}
@ -632,25 +632,29 @@ bool ciMethod::argument_profiled_type(int bci, int i, ciKlass*& type, bool& mayb
*
* @param [in]bci bci of the call
* @param [out]type profiled type of argument, NULL if none
* @param [out]maybe_null true if null was seen for argument
* @param [out]ptr_kind whether always null, never null or maybe null
* @return true if profiling exists
*
*/
bool ciMethod::return_profiled_type(int bci, ciKlass*& type, bool& maybe_null) {
bool ciMethod::return_profiled_type(int bci, ciKlass*& type, ProfilePtrKind& ptr_kind) {
if (MethodData::profile_return() && method_data() != NULL && method_data()->is_mature()) {
ciProfileData* data = method_data()->bci_to_data(bci);
if (data != NULL) {
if (data->is_VirtualCallTypeData()) {
assert_virtual_call_type_ok(bci);
ciVirtualCallTypeData* call = (ciVirtualCallTypeData*)data->as_VirtualCallTypeData();
type = call->valid_return_type();
maybe_null = call->return_maybe_null();
return true;
if (call->has_return()) {
type = call->valid_return_type();
ptr_kind = call->return_ptr_kind();
return true;
}
} else if (data->is_CallTypeData()) {
assert_call_type_ok(bci);
ciCallTypeData* call = (ciCallTypeData*)data->as_CallTypeData();
type = call->valid_return_type();
maybe_null = call->return_maybe_null();
if (call->has_return()) {
type = call->valid_return_type();
ptr_kind = call->return_ptr_kind();
}
return true;
}
}
@ -663,16 +667,16 @@ bool ciMethod::return_profiled_type(int bci, ciKlass*& type, bool& maybe_null) {
*
* @param [in]i parameter number
* @param [out]type profiled type of parameter, NULL if none
* @param [out]maybe_null true if null was seen for parameter
* @param [out]ptr_kind whether always null, never null or maybe null
* @return true if profiling exists
*
*/
bool ciMethod::parameter_profiled_type(int i, ciKlass*& type, bool& maybe_null) {
bool ciMethod::parameter_profiled_type(int i, ciKlass*& type, ProfilePtrKind& ptr_kind) {
if (MethodData::profile_parameters() && method_data() != NULL && method_data()->is_mature()) {
ciParametersTypeData* parameters = method_data()->parameters_type_data();
if (parameters != NULL && i < parameters->number_of_parameters()) {
type = parameters->valid_parameter_type(i);
maybe_null = parameters->parameter_maybe_null(i);
ptr_kind = parameters->parameter_ptr_kind(i);
return true;
}
}

View File

@ -40,6 +40,14 @@ class Arena;
class BCEscapeAnalyzer;
class InlineTree;
// Whether profiling found an oop to be always, never or sometimes
// null
enum ProfilePtrKind {
ProfileAlwaysNull,
ProfileNeverNull,
ProfileMaybeNull
};
// ciMethod
//
// This class represents a Method* in the HotSpot virtual
@ -248,9 +256,9 @@ class ciMethod : public ciMetadata {
int interpreter_call_site_count(int bci);
// Does type profiling provide any useful information at this point?
bool argument_profiled_type(int bci, int i, ciKlass*& type, bool& maybe_null);
bool parameter_profiled_type(int i, ciKlass*& type, bool& maybe_null);
bool return_profiled_type(int bci, ciKlass*& type, bool& maybe_null);
bool argument_profiled_type(int bci, int i, ciKlass*& type, ProfilePtrKind& ptr_kind);
bool parameter_profiled_type(int i, ciKlass*& type, ProfilePtrKind& ptr_kind);
bool return_profiled_type(int bci, ciKlass*& type, ProfilePtrKind& ptr_kind);
ciField* get_field_at_bci( int bci, bool &will_link);
ciMethod* get_method_at_bci(int bci, bool &will_link, ciSignature* *declared_signature);

View File

@ -88,6 +88,17 @@ public:
}
}
static ProfilePtrKind ptr_kind(intptr_t v) {
bool maybe_null = TypeEntries::was_null_seen(v);
if (!maybe_null) {
return ProfileNeverNull;
} else if (TypeEntries::is_type_none(v)) {
return ProfileAlwaysNull;
} else {
return ProfileMaybeNull;
}
}
static intptr_t with_status(ciKlass* k, intptr_t in) {
return TypeEntries::with_status((intptr_t)k, in);
}
@ -105,8 +116,8 @@ public:
return valid_ciklass(type(i));
}
bool maybe_null(int i) const {
return was_null_seen(type(i));
ProfilePtrKind ptr_kind(int i) const {
return ciTypeEntries::ptr_kind(type(i));
}
#ifndef PRODUCT
@ -122,8 +133,8 @@ public:
return valid_ciklass(type());
}
bool maybe_null() const {
return was_null_seen(type());
ProfilePtrKind ptr_kind() const {
return ciTypeEntries::ptr_kind(type());
}
#ifndef PRODUCT
@ -167,12 +178,12 @@ public:
return ret()->valid_type();
}
bool argument_maybe_null(int i) const {
return args()->maybe_null(i);
ProfilePtrKind argument_ptr_kind(int i) const {
return args()->ptr_kind(i);
}
bool return_maybe_null() const {
return ret()->maybe_null();
ProfilePtrKind return_ptr_kind() const {
return ret()->ptr_kind();
}
#ifndef PRODUCT
@ -281,12 +292,12 @@ public:
return ret()->valid_type();
}
bool argument_maybe_null(int i) const {
return args()->maybe_null(i);
ProfilePtrKind argument_ptr_kind(int i) const {
return args()->ptr_kind(i);
}
bool return_maybe_null() const {
return ret()->maybe_null();
ProfilePtrKind return_ptr_kind() const {
return ret()->ptr_kind();
}
#ifndef PRODUCT
@ -334,8 +345,8 @@ public:
return parameters()->valid_type(i);
}
bool parameter_maybe_null(int i) const {
return parameters()->maybe_null(i);
ProfilePtrKind parameter_ptr_kind(int i) const {
return parameters()->ptr_kind(i);
}
#ifndef PRODUCT

View File

@ -205,13 +205,15 @@ int TypeStackSlotEntries::compute_cell_count(Symbol* signature, bool include_rec
int TypeEntriesAtCall::compute_cell_count(BytecodeStream* stream) {
assert(Bytecodes::is_invoke(stream->code()), "should be invoke");
assert(TypeStackSlotEntries::per_arg_count() > ReturnTypeEntry::static_cell_count(), "code to test for arguments/results broken");
Bytecode_invoke inv(stream->method(), stream->bci());
const methodHandle m = stream->method();
int bci = stream->bci();
Bytecode_invoke inv(m, bci);
int args_cell = 0;
if (arguments_profiling_enabled()) {
if (MethodData::profile_arguments_for_invoke(m, bci)) {
args_cell = TypeStackSlotEntries::compute_cell_count(inv.signature(), false, TypeProfileArgsLimit);
}
int ret_cell = 0;
if (return_profiling_enabled() && (inv.result_type() == T_OBJECT || inv.result_type() == T_ARRAY)) {
if (MethodData::profile_return_for_invoke(m, bci) && (inv.result_type() == T_OBJECT || inv.result_type() == T_ARRAY)) {
ret_cell = ReturnTypeEntry::static_cell_count();
}
int header_cell = 0;
@ -1525,6 +1527,18 @@ bool MethodData::profile_jsr292(const methodHandle& m, int bci) {
return inv.is_invokedynamic() || inv.is_invokehandle();
}
bool MethodData::profile_unsafe(const methodHandle& m, int bci) {
Bytecode_invoke inv(m , bci);
if (inv.is_invokevirtual() && inv.klass() == vmSymbols::jdk_internal_misc_Unsafe()) {
ResourceMark rm;
char* name = inv.name()->as_C_string();
if (!strncmp(name, "get", 3) || !strncmp(name, "put", 3)) {
return true;
}
}
return false;
}
int MethodData::profile_arguments_flag() {
return TypeProfileLevel % 10;
}
@ -1550,6 +1564,10 @@ bool MethodData::profile_arguments_for_invoke(const methodHandle& m, int bci) {
return true;
}
if (profile_unsafe(m, bci)) {
return true;
}
assert(profile_arguments_jsr292_only(), "inconsistent");
return profile_jsr292(m, bci);
}

View File

@ -2151,6 +2151,7 @@ class MethodData : public Metadata {
CC_INTERP_ONLY(friend class BytecodeInterpreter;)
private:
friend class ProfileData;
friend class TypeEntriesAtCall;
// Back pointer to the Method*
Method* _method;
@ -2173,7 +2174,7 @@ public:
// Whole-method sticky bits and flags
enum {
_trap_hist_limit = 22 JVMCI_ONLY(+5), // decoupled from Deoptimization::Reason_LIMIT
_trap_hist_limit = 23 JVMCI_ONLY(+5), // decoupled from Deoptimization::Reason_LIMIT
_trap_hist_mask = max_jubyte,
_extra_data_count = 4 // extra DataLayout headers, for trap history
}; // Public flag values
@ -2302,6 +2303,7 @@ private:
};
static bool profile_jsr292(const methodHandle& m, int bci);
static bool profile_unsafe(const methodHandle& m, int bci);
static int profile_arguments_flag();
static bool profile_all_arguments();
static bool profile_arguments_for_invoke(const methodHandle& m, int bci);

View File

@ -569,7 +569,7 @@ void Parse::do_call() {
// save across call, for a subsequent cast_not_null.
Node* receiver = has_receiver ? argument(0) : NULL;
// The extra CheckCastPP for speculative types mess with PhaseStringOpts
// The extra CheckCastPPs for speculative types mess with PhaseStringOpts
if (receiver != NULL && !call_does_dispatch && !cg->is_string_late_inline()) {
// Feed profiling data for a single receiver to the type system so
// it can propagate it as a speculative type

View File

@ -1294,7 +1294,7 @@ Node* GraphKit::null_check_common(Node* value, BasicType type,
float ok_prob = PROB_MAX; // a priori estimate: nulls never happen
Deoptimization::DeoptReason reason;
if (assert_null) {
reason = Deoptimization::Reason_null_assert;
reason = Deoptimization::reason_null_assert(speculative);
} else if (type == T_OBJECT) {
reason = Deoptimization::reason_null_check(speculative);
} else {
@ -2133,7 +2133,7 @@ void GraphKit::round_double_arguments(ciMethod* dest_method) {
*
* @return node with improved type
*/
Node* GraphKit::record_profile_for_speculation(Node* n, ciKlass* exact_kls, bool maybe_null) {
Node* GraphKit::record_profile_for_speculation(Node* n, ciKlass* exact_kls, ProfilePtrKind ptr_kind) {
const Type* current_type = _gvn.type(n);
assert(UseTypeSpeculation, "type speculation must be on");
@ -2145,19 +2145,24 @@ Node* GraphKit::record_profile_for_speculation(Node* n, ciKlass* exact_kls, bool
const TypeOopPtr* xtype = tklass->as_instance_type();
assert(xtype->klass_is_exact(), "Should be exact");
// Any reason to believe n is not null (from this profiling or a previous one)?
const TypePtr* ptr = (maybe_null && current_type->speculative_maybe_null()) ? TypePtr::BOTTOM : TypePtr::NOTNULL;
assert(ptr_kind != ProfileAlwaysNull, "impossible here");
const TypePtr* ptr = (ptr_kind == ProfileMaybeNull && current_type->speculative_maybe_null()) ? TypePtr::BOTTOM : TypePtr::NOTNULL;
// record the new speculative type's depth
speculative = xtype->cast_to_ptr_type(ptr->ptr())->is_ptr();
speculative = speculative->with_inline_depth(jvms()->depth());
} else if (current_type->would_improve_ptr(maybe_null)) {
} else if (current_type->would_improve_ptr(ptr_kind)) {
// Profiling report that null was never seen so we can change the
// speculative type to non null ptr.
assert(!maybe_null, "nothing to improve");
if (speculative == NULL) {
speculative = TypePtr::NOTNULL;
if (ptr_kind == ProfileAlwaysNull) {
speculative = TypePtr::NULL_PTR;
} else {
assert(ptr_kind == ProfileNeverNull, "nothing else is an improvement");
const TypePtr* ptr = TypePtr::NOTNULL;
speculative = speculative->cast_to_ptr_type(ptr->ptr())->is_ptr();
if (speculative != NULL) {
speculative = speculative->cast_to_ptr_type(ptr->ptr())->is_ptr();
} else {
speculative = ptr;
}
}
}
@ -2191,14 +2196,30 @@ Node* GraphKit::record_profiled_receiver_for_speculation(Node* n) {
return n;
}
ciKlass* exact_kls = profile_has_unique_klass();
bool maybe_null = true;
if (java_bc() == Bytecodes::_checkcast ||
java_bc() == Bytecodes::_instanceof ||
java_bc() == Bytecodes::_aastore) {
ProfilePtrKind ptr_kind = ProfileMaybeNull;
if ((java_bc() == Bytecodes::_checkcast ||
java_bc() == Bytecodes::_instanceof ||
java_bc() == Bytecodes::_aastore) &&
method()->method_data()->is_mature()) {
ciProfileData* data = method()->method_data()->bci_to_data(bci());
maybe_null = data == NULL ? true : data->as_BitData()->null_seen();
if (data != NULL) {
if (!data->as_BitData()->null_seen()) {
ptr_kind = ProfileNeverNull;
} else {
assert(data->is_ReceiverTypeData(), "bad profile data type");
ciReceiverTypeData* call = (ciReceiverTypeData*)data->as_ReceiverTypeData();
uint i = 0;
for (; i < call->row_limit(); i++) {
ciKlass* receiver = call->receiver(i);
if (receiver != NULL) {
break;
}
}
ptr_kind = (i == call->row_limit()) ? ProfileAlwaysNull : ProfileMaybeNull;
}
}
}
return record_profile_for_speculation(n, exact_kls, maybe_null);
return record_profile_for_speculation(n, exact_kls, ptr_kind);
}
/**
@ -2218,10 +2239,10 @@ void GraphKit::record_profiled_arguments_for_speculation(ciMethod* dest_method,
for (int j = skip, i = 0; j < nargs && i < TypeProfileArgsLimit; j++) {
const Type *targ = tf->domain()->field_at(j + TypeFunc::Parms);
if (targ->basic_type() == T_OBJECT || targ->basic_type() == T_ARRAY) {
bool maybe_null = true;
ProfilePtrKind ptr_kind = ProfileMaybeNull;
ciKlass* better_type = NULL;
if (method()->argument_profiled_type(bci(), i, better_type, maybe_null)) {
record_profile_for_speculation(argument(j), better_type, maybe_null);
if (method()->argument_profiled_type(bci(), i, better_type, ptr_kind)) {
record_profile_for_speculation(argument(j), better_type, ptr_kind);
}
i++;
}
@ -2238,10 +2259,10 @@ void GraphKit::record_profiled_parameters_for_speculation() {
}
for (int i = 0, j = 0; i < method()->arg_size() ; i++) {
if (_gvn.type(local(i))->isa_oopptr()) {
bool maybe_null = true;
ProfilePtrKind ptr_kind = ProfileMaybeNull;
ciKlass* better_type = NULL;
if (method()->parameter_profiled_type(j, better_type, maybe_null)) {
record_profile_for_speculation(local(i), better_type, maybe_null);
if (method()->parameter_profiled_type(j, better_type, ptr_kind)) {
record_profile_for_speculation(local(i), better_type, ptr_kind);
}
j++;
}
@ -2256,13 +2277,13 @@ void GraphKit::record_profiled_return_for_speculation() {
if (!UseTypeSpeculation) {
return;
}
bool maybe_null = true;
ProfilePtrKind ptr_kind = ProfileMaybeNull;
ciKlass* better_type = NULL;
if (method()->return_profiled_type(bci(), better_type, maybe_null)) {
if (method()->return_profiled_type(bci(), better_type, ptr_kind)) {
// If profiling reports a single type for the return value,
// feed it to the type system so it can propagate it as a
// speculative type
record_profile_for_speculation(stack(sp()-1), better_type, maybe_null);
record_profile_for_speculation(stack(sp()-1), better_type, ptr_kind);
}
}
@ -2938,12 +2959,7 @@ Node* GraphKit::gen_instanceof(Node* obj, Node* superklass, bool safe_for_replac
}
}
if (known_statically && UseTypeSpeculation) {
// If we know the type check always succeeds then we don't use the
// profiling data at this bytecode. Don't lose it, feed it to the
// type system as a speculative type.
not_null_obj = record_profiled_receiver_for_speculation(not_null_obj);
} else {
if (!known_statically) {
const TypeOopPtr* obj_type = _gvn.type(obj)->is_oopptr();
// We may not have profiling here or it may not help us. If we
// have a speculative type use it to perform an exact cast.
@ -2977,6 +2993,15 @@ Node* GraphKit::gen_instanceof(Node* obj, Node* superklass, bool safe_for_replac
// Return final merged results
set_control( _gvn.transform(region) );
record_for_igvn(region);
// If we know the type check always succeeds then we don't use the
// profiling data at this bytecode. Don't lose it, feed it to the
// type system as a speculative type.
if (safe_for_replace) {
Node* casted_obj = record_profiled_receiver_for_speculation(obj);
replace_in_map(obj, casted_obj);
}
return _gvn.transform(phi);
}
@ -3117,7 +3142,8 @@ Node* GraphKit::gen_checkcast(Node *obj, Node* superklass,
// Return final merged results
set_control( _gvn.transform(region) );
record_for_igvn(region);
return res;
return record_profiled_receiver_for_speculation(res);
}
//------------------------------next_monitor-----------------------------------

View File

@ -365,7 +365,7 @@ class GraphKit : public Phase {
// Throw an uncommon trap if a given value is __not__ null.
// Return the value cast to null, and be clever about dominating checks.
Node* null_assert(Node* value, BasicType type = T_OBJECT) {
return null_check_common(value, type, true);
return null_check_common(value, type, true, NULL, _gvn.type(value)->speculative_always_null());
}
// Check if value is null and abort if it is
@ -397,7 +397,7 @@ class GraphKit : public Phase {
}
// record type from profiling with the type system
Node* record_profile_for_speculation(Node* n, ciKlass* exact_kls, bool maybe_null);
Node* record_profile_for_speculation(Node* n, ciKlass* exact_kls, ProfilePtrKind ptr_kind);
void record_profiled_arguments_for_speculation(ciMethod* dest_method, Bytecodes::Code bc);
void record_profiled_parameters_for_speculation();
void record_profiled_return_for_speculation();

View File

@ -240,7 +240,7 @@ class LibraryCallKit : public GraphKit {
Node* generate_min_max(vmIntrinsics::ID id, Node* x, Node* y);
// This returns Type::AnyPtr, RawPtr, or OopPtr.
int classify_unsafe_addr(Node* &base, Node* &offset, BasicType type);
Node* make_unsafe_address(Node*& base, Node* offset, BasicType type = T_ILLEGAL);
Node* make_unsafe_address(Node*& base, Node* offset, BasicType type = T_ILLEGAL, bool can_cast = false);
// Helper for inline_unsafe_access.
// Generates the guards that check whether the result of
// Unsafe.getObject should be recorded in an SATB log buffer.
@ -2111,13 +2111,33 @@ LibraryCallKit::classify_unsafe_addr(Node* &base, Node* &offset, BasicType type)
}
}
inline Node* LibraryCallKit::make_unsafe_address(Node*& base, Node* offset, BasicType type) {
inline Node* LibraryCallKit::make_unsafe_address(Node*& base, Node* offset, BasicType type, bool can_cast) {
Node* uncasted_base = base;
int kind = classify_unsafe_addr(uncasted_base, offset, type);
if (kind == Type::RawPtr) {
return basic_plus_adr(top(), uncasted_base, offset);
} else if (kind == Type::AnyPtr) {
assert(base == uncasted_base, "unexpected base change");
if (can_cast) {
if (!_gvn.type(base)->speculative_maybe_null() &&
!too_many_traps(Deoptimization::Reason_speculate_null_check)) {
// According to profiling, this access is always on
// heap. Casting the base to not null and thus avoiding membars
// around the access should allow better optimizations
Node* null_ctl = top();
base = null_check_oop(base, &null_ctl, true, true, true);
assert(null_ctl->is_top(), "no null control here");
return basic_plus_adr(base, offset);
} else if (_gvn.type(base)->speculative_always_null() &&
!too_many_traps(Deoptimization::Reason_speculate_null_assert)) {
// According to profiling, this access is always off
// heap.
base = null_assert(base);
Node* raw_base = _gvn.transform(new CastX2PNode(offset));
offset = MakeConX(0);
return basic_plus_adr(top(), raw_base, offset);
}
}
// We don't know if it's an on heap or off heap access. Fall back
// to raw memory access.
Node* raw = _gvn.transform(new CheckCastPPNode(control(), base, TypeRawPtr::BOTTOM));
@ -2359,7 +2379,8 @@ bool LibraryCallKit::inline_unsafe_access(bool is_store, const BasicType type, c
"fieldOffset must be byte-scaled");
// 32-bit machines ignore the high half!
offset = ConvL2X(offset);
adr = make_unsafe_address(base, offset, type);
adr = make_unsafe_address(base, offset, type, kind == Relaxed);
if (_gvn.type(base)->isa_ptr() != TypePtr::NULL_PTR) {
heap_base_oop = base;
} else if (type == T_OBJECT) {
@ -2417,7 +2438,7 @@ bool LibraryCallKit::inline_unsafe_access(bool is_store, const BasicType type, c
bool need_mem_bar = false;
switch (kind) {
case Relaxed:
need_mem_bar = mismatched && !adr_type->isa_aryptr();
need_mem_bar = (mismatched && !adr_type->isa_aryptr()) || can_access_non_heap;
break;
case Opaque:
// Opaque uses CPUOrder membars for protection against code movement.
@ -2521,7 +2542,22 @@ bool LibraryCallKit::inline_unsafe_access(bool is_store, const BasicType type, c
if (p == NULL) {
// To be valid, unsafe loads may depend on other conditions than
// the one that guards them: pin the Load node
p = make_load(control(), adr, value_type, type, adr_type, mo, LoadNode::Pinned, requires_atomic_access, unaligned, mismatched);
LoadNode::ControlDependency dep = LoadNode::Pinned;
Node* ctrl = control();
if (adr_type->isa_instptr()) {
assert(adr_type->meet(TypePtr::NULL_PTR) != adr_type->remove_speculative(), "should be not null");
intptr_t offset = Type::OffsetBot;
AddPNode::Ideal_base_and_offset(adr, &_gvn, offset);
if (offset >= 0) {
int s = Klass::layout_helper_size_in_bytes(adr_type->isa_instptr()->klass()->layout_helper());
if (offset < s) {
// Guaranteed to be a valid access, no need to pin it
dep = LoadNode::DependsOnlyOnTest;
ctrl = NULL;
}
}
}
p = make_load(ctrl, adr, value_type, type, adr_type, mo, dep, requires_atomic_access, unaligned, mismatched);
// load value
switch (type) {
case T_BOOLEAN:
@ -2770,7 +2806,7 @@ bool LibraryCallKit::inline_unsafe_load_store(const BasicType type, const LoadSt
assert(Unsafe_field_offset_to_byte_offset(11) == 11, "fieldOffset must be byte-scaled");
// 32-bit machines ignore the high half of long offsets
offset = ConvL2X(offset);
Node* adr = make_unsafe_address(base, offset, type);
Node* adr = make_unsafe_address(base, offset, type, false);
const TypePtr *adr_type = _gvn.type(adr)->isa_ptr();
Compile::AliasType* alias_type = C->alias_type(adr_type);

View File

@ -2282,6 +2282,11 @@ void Parse::do_one_bytecode() {
b = null_check_oop(b, &null_ctl, true, true, true);
assert(null_ctl->is_top(), "no null control here");
dec_sp(1);
} else if (_gvn.type(b)->speculative_always_null() &&
!too_many_traps(Deoptimization::Reason_speculate_null_assert)) {
inc_sp(1);
b = null_assert(b);
dec_sp(1);
}
c = _gvn.transform( new CmpPNode(b, a) );
do_ifnull(btest, c);

View File

@ -2531,7 +2531,8 @@ const Type* TypePtr::cleanup_speculative() const {
const TypeOopPtr* spec_oopptr = speculative()->isa_oopptr();
// If the speculative may be null and is an inexact klass then it
// doesn't help
if (speculative()->maybe_null() && (spec_oopptr == NULL || !spec_oopptr->klass_is_exact())) {
if (speculative() != TypePtr::NULL_PTR && speculative()->maybe_null() &&
(spec_oopptr == NULL || !spec_oopptr->klass_is_exact())) {
return no_spec;
}
return this;
@ -2660,6 +2661,14 @@ bool TypePtr::speculative_maybe_null() const {
return true;
}
bool TypePtr::speculative_always_null() const {
if (_speculative != NULL) {
const TypePtr* speculative = _speculative->join(this)->is_ptr();
return speculative == TypePtr::NULL_PTR;
}
return false;
}
/**
* Same as TypePtr::speculative_type() but return the klass only if
* the speculative tells us is not null
@ -2684,6 +2693,9 @@ bool TypePtr::would_improve_type(ciKlass* exact_kls, int inline_depth) const {
if (exact_kls == NULL) {
return false;
}
if (speculative() == TypePtr::NULL_PTR) {
return false;
}
// no speculative type or non exact speculative type?
if (speculative_type() == NULL) {
return true;
@ -2703,16 +2715,16 @@ bool TypePtr::would_improve_type(ciKlass* exact_kls, int inline_depth) const {
* Check whether new profiling would improve ptr (= tells us it is non
* null)
*
* @param maybe_null true if profiling tells the ptr may be null
* @param ptr_kind always null or not null?
*
* @return true if ptr profile is valuable
*/
bool TypePtr::would_improve_ptr(bool maybe_null) const {
bool TypePtr::would_improve_ptr(ProfilePtrKind ptr_kind) const {
// profiling doesn't tell us anything useful
if (maybe_null) {
if (ptr_kind != ProfileAlwaysNull && ptr_kind != ProfileNeverNull) {
return false;
}
// We already know this is not be null
// We already know this is not null
if (!this->maybe_null()) {
return false;
}
@ -2720,6 +2732,17 @@ bool TypePtr::would_improve_ptr(bool maybe_null) const {
if (!speculative_maybe_null()) {
return false;
}
// We already know this is always null
if (this == TypePtr::NULL_PTR) {
return false;
}
// We already know the speculative type is always null
if (speculative_always_null()) {
return false;
}
if (ptr_kind == ProfileAlwaysNull && speculative() != NULL && speculative()->isa_oopptr()) {
return false;
}
return true;
}

View File

@ -445,10 +445,11 @@ public:
virtual ciKlass* speculative_type() const { return NULL; }
virtual ciKlass* speculative_type_not_null() const { return NULL; }
virtual bool speculative_maybe_null() const { return true; }
virtual bool speculative_always_null() const { return true; }
virtual const Type* remove_speculative() const { return this; }
virtual const Type* cleanup_speculative() const { return this; }
virtual bool would_improve_type(ciKlass* exact_kls, int inline_depth) const { return exact_kls != NULL; }
virtual bool would_improve_ptr(bool maybe_null) const { return !maybe_null; }
virtual bool would_improve_ptr(ProfilePtrKind ptr_kind) const { return ptr_kind == ProfileAlwaysNull || ptr_kind == ProfileNeverNull; }
const Type* maybe_remove_speculative(bool include_speculative) const;
virtual bool maybe_null() const { return true; }
@ -885,10 +886,11 @@ public:
virtual ciKlass* speculative_type() const;
virtual ciKlass* speculative_type_not_null() const;
virtual bool speculative_maybe_null() const;
virtual bool speculative_always_null() const;
virtual const Type* remove_speculative() const;
virtual const Type* cleanup_speculative() const;
virtual bool would_improve_type(ciKlass* exact_kls, int inline_depth) const;
virtual bool would_improve_ptr(bool maybe_null) const;
virtual bool would_improve_ptr(ProfilePtrKind maybe_null) const;
virtual const TypePtr* with_inline_depth(int depth) const;
virtual bool maybe_null() const { return meet_ptr(Null) == ptr(); }

View File

@ -2179,6 +2179,7 @@ const char* Deoptimization::_trap_reason_name[] = {
"loop_limit_check",
"speculate_class_check",
"speculate_null_check",
"speculate_null_assert",
"rtm_state_change",
"unstable_if",
"unstable_fused_if",

View File

@ -74,6 +74,7 @@ class Deoptimization : AllStatic {
Reason_loop_limit_check, // compiler generated loop limits check failed
Reason_speculate_class_check, // saw unexpected object class from type speculation
Reason_speculate_null_check, // saw unexpected null from type speculation
Reason_speculate_null_assert, // saw unexpected null from type speculation
Reason_rtm_state_change, // rtm state change detected
Reason_unstable_if, // a branch predicted always false was taken
Reason_unstable_fused_if, // fused two ifs that had each one untaken branch. One is now taken.
@ -365,6 +366,8 @@ JVMCI_ONLY(public:)
return Reason_class_check;
else if (reason == Reason_speculate_null_check)
return Reason_null_check;
else if (reason == Reason_speculate_null_assert)
return Reason_null_assert;
else if (reason == Reason_unstable_if)
return Reason_intrinsic;
else if (reason == Reason_unstable_fused_if)
@ -374,7 +377,9 @@ JVMCI_ONLY(public:)
}
static bool reason_is_speculate(int reason) {
if (reason == Reason_speculate_class_check || reason == Reason_speculate_null_check) {
if (reason == Reason_speculate_class_check ||
reason == Reason_speculate_null_check ||
reason == Reason_speculate_null_assert) {
return true;
}
return false;
@ -388,6 +393,10 @@ JVMCI_ONLY(public:)
return speculative ? Deoptimization::Reason_speculate_class_check : Deoptimization::Reason_class_check;
}
static DeoptReason reason_null_assert(bool speculative) {
return speculative ? Deoptimization::Reason_speculate_null_assert : Deoptimization::Reason_null_assert;
}
static uint per_method_trap_limit(int reason) {
return reason_is_speculate(reason) ? (uint)PerMethodSpecTrapLimit : (uint)PerMethodTrapLimit;
}

View File

@ -2656,6 +2656,7 @@ typedef RehashableHashtable<Symbol*, mtSymbol> RehashableSymbolHashtable;
declare_constant(Deoptimization::Reason_loop_limit_check) \
declare_constant(Deoptimization::Reason_speculate_class_check) \
declare_constant(Deoptimization::Reason_speculate_null_check) \
declare_constant(Deoptimization::Reason_speculate_null_assert) \
declare_constant(Deoptimization::Reason_rtm_state_change) \
declare_constant(Deoptimization::Reason_unstable_if) \
declare_constant(Deoptimization::Reason_unstable_fused_if) \