154a1a02bf
Use script and some manual fixup to fix directores names in include guards. Reviewed-by: lfoltan, eosterlund, kbarrett
132 lines
5.0 KiB
C++
132 lines
5.0 KiB
C++
/*
|
|
* Copyright (c) 1999, 2019, Oracle and/or its affiliates. All rights reserved.
|
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
*
|
|
* This code is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 only, as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
* version 2 for more details (a copy is included in the LICENSE file that
|
|
* accompanied this code).
|
|
*
|
|
* You should have received a copy of the GNU General Public License version
|
|
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*
|
|
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
* or visit www.oracle.com if you need additional information or have any
|
|
* questions.
|
|
*
|
|
*/
|
|
|
|
#ifndef SHARE_OPTO_LOCKNODE_HPP
|
|
#define SHARE_OPTO_LOCKNODE_HPP
|
|
|
|
#include "opto/node.hpp"
|
|
#include "opto/opcodes.hpp"
|
|
#include "opto/subnode.hpp"
|
|
|
|
class BiasedLockingCounters;
|
|
class RTMLockingCounters;
|
|
|
|
//------------------------------BoxLockNode------------------------------------
|
|
class BoxLockNode : public Node {
|
|
const int _slot; // stack slot
|
|
RegMask _inmask; // OptoReg corresponding to stack slot
|
|
bool _is_eliminated; // Associated locks were safely eliminated
|
|
|
|
public:
|
|
BoxLockNode( int lock );
|
|
virtual int Opcode() const;
|
|
virtual void emit(CodeBuffer &cbuf, PhaseRegAlloc *ra_) const;
|
|
virtual uint size(PhaseRegAlloc *ra_) const;
|
|
virtual const RegMask &in_RegMask(uint) const;
|
|
virtual const RegMask &out_RegMask() const;
|
|
virtual uint size_of() const;
|
|
virtual uint hash() const;
|
|
virtual uint cmp( const Node &n ) const;
|
|
virtual const class Type *bottom_type() const { return TypeRawPtr::BOTTOM; }
|
|
virtual uint ideal_reg() const { return Op_RegP; }
|
|
|
|
static OptoReg::Name reg(Node* box_node);
|
|
static BoxLockNode* box_node(Node* box_node);
|
|
static bool same_slot(Node* box1, Node* box2) {
|
|
return box1->as_BoxLock()->_slot == box2->as_BoxLock()->_slot;
|
|
}
|
|
int stack_slot() const { return _slot; }
|
|
|
|
bool is_eliminated() const { return _is_eliminated; }
|
|
// mark lock as eliminated.
|
|
void set_eliminated() { _is_eliminated = true; }
|
|
|
|
// Is BoxLock node used for one simple lock region?
|
|
bool is_simple_lock_region(LockNode** unique_lock, Node* obj);
|
|
|
|
#ifndef PRODUCT
|
|
virtual void format( PhaseRegAlloc *, outputStream *st ) const;
|
|
virtual void dump_spec(outputStream *st) const { st->print(" Lock %d",_slot); }
|
|
#endif
|
|
};
|
|
|
|
//------------------------------FastLockNode-----------------------------------
|
|
class FastLockNode: public CmpNode {
|
|
private:
|
|
BiasedLockingCounters* _counters;
|
|
RTMLockingCounters* _rtm_counters; // RTM lock counters for inflated locks
|
|
RTMLockingCounters* _stack_rtm_counters; // RTM lock counters for stack locks
|
|
|
|
public:
|
|
FastLockNode(Node *ctrl, Node *oop, Node *box) : CmpNode(oop,box) {
|
|
init_req(0,ctrl);
|
|
init_class_id(Class_FastLock);
|
|
_counters = NULL;
|
|
_rtm_counters = NULL;
|
|
_stack_rtm_counters = NULL;
|
|
}
|
|
Node* obj_node() const { return in(1); }
|
|
Node* box_node() const { return in(2); }
|
|
void set_box_node(Node* box) { set_req(2, box); }
|
|
|
|
// FastLock and FastUnlockNode do not hash, we need one for each correspoding
|
|
// LockNode/UnLockNode to avoid creating Phi's.
|
|
virtual uint hash() const ; // { return NO_HASH; }
|
|
virtual uint size_of() const;
|
|
virtual uint cmp( const Node &n ) const ; // Always fail, except on self
|
|
virtual int Opcode() const;
|
|
virtual const Type* Value(PhaseGVN* phase) const { return TypeInt::CC; }
|
|
const Type *sub(const Type *t1, const Type *t2) const { return TypeInt::CC;}
|
|
|
|
void create_lock_counter(JVMState* s);
|
|
void create_rtm_lock_counter(JVMState* state);
|
|
BiasedLockingCounters* counters() const { return _counters; }
|
|
RTMLockingCounters* rtm_counters() const { return _rtm_counters; }
|
|
RTMLockingCounters* stack_rtm_counters() const { return _stack_rtm_counters; }
|
|
};
|
|
|
|
|
|
//------------------------------FastUnlockNode---------------------------------
|
|
class FastUnlockNode: public CmpNode {
|
|
public:
|
|
FastUnlockNode(Node *ctrl, Node *oop, Node *box) : CmpNode(oop,box) {
|
|
init_req(0,ctrl);
|
|
init_class_id(Class_FastUnlock);
|
|
}
|
|
Node* obj_node() const { return in(1); }
|
|
Node* box_node() const { return in(2); }
|
|
|
|
|
|
// FastLock and FastUnlockNode do not hash, we need one for each correspoding
|
|
// LockNode/UnLockNode to avoid creating Phi's.
|
|
virtual uint hash() const ; // { return NO_HASH; }
|
|
virtual uint cmp( const Node &n ) const ; // Always fail, except on self
|
|
virtual int Opcode() const;
|
|
virtual const Type* Value(PhaseGVN* phase) const { return TypeInt::CC; }
|
|
const Type *sub(const Type *t1, const Type *t2) const { return TypeInt::CC;}
|
|
|
|
};
|
|
|
|
#endif // SHARE_OPTO_LOCKNODE_HPP
|