8204179: [BACKOUT] OopStorage should use GlobalCounter

Backout JDK-8202945

Reviewed-by: eosterlund, pliden
This commit is contained in:
Kim Barrett 2018-05-31 15:37:18 -04:00
parent e8aba519f7
commit dd3d24341b
2 changed files with 63 additions and 5 deletions

View File

@ -40,7 +40,6 @@
#include "utilities/align.hpp"
#include "utilities/count_trailing_zeros.hpp"
#include "utilities/debug.hpp"
#include "utilities/globalCounter.inline.hpp"
#include "utilities/globalDefinitions.hpp"
#include "utilities/macros.hpp"
#include "utilities/ostream.hpp"
@ -502,6 +501,48 @@ bool OopStorage::expand_active_array() {
return true;
}
OopStorage::ProtectActive::ProtectActive() : _enter(0), _exit() {}
// Begin read-side critical section.
uint OopStorage::ProtectActive::read_enter() {
return Atomic::add(2u, &_enter);
}
// End read-side critical section.
void OopStorage::ProtectActive::read_exit(uint enter_value) {
Atomic::add(2u, &_exit[enter_value & 1]);
}
// Wait until all readers that entered the critical section before
// synchronization have exited that critical section.
void OopStorage::ProtectActive::write_synchronize() {
SpinYield spinner;
// Determine old and new exit counters, based on bit0 of the
// on-entry _enter counter.
uint value = OrderAccess::load_acquire(&_enter);
volatile uint* new_ptr = &_exit[(value + 1) & 1];
// Atomically change the in-use exit counter to the new counter, by
// adding 1 to the _enter counter (flipping bit0 between 0 and 1)
// and initializing the new exit counter to that enter value. Note:
// The new exit counter is not being used by read operations until
// this change succeeds.
uint old;
do {
old = value;
*new_ptr = ++value;
value = Atomic::cmpxchg(value, &_enter, old);
} while (old != value);
// Readers that entered the critical section before we changed the
// selected exit counter will use the old exit counter. Readers
// entering after the change will use the new exit counter. Wait
// for all the critical sections started before the change to
// complete, e.g. for the value of old_ptr to catch up with old.
volatile uint* old_ptr = &_exit[old & 1];
while (old != OrderAccess::load_acquire(old_ptr)) {
spinner.wait();
}
}
// Make new_array the _active_array. Increments new_array's refcount
// to account for the new reference. The assignment is atomic wrto
// obtain_active_array; once this function returns, it is safe for the
@ -513,9 +554,9 @@ void OopStorage::replace_active_array(ActiveArray* new_array) {
// Install new_array, ensuring its initialization is complete first.
OrderAccess::release_store(&_active_array, new_array);
// Wait for any readers that could read the old array from _active_array.
GlobalCounter::write_synchronize();
// All obtain_active_array critical sections that could see the old array
// have completed, having incremented the refcount of the old array. The
_protect_active.write_synchronize();
// All obtain critical sections that could see the old array have
// completed, having incremented the refcount of the old array. The
// caller can now safely relinquish the old array.
}
@ -525,9 +566,10 @@ void OopStorage::replace_active_array(ActiveArray* new_array) {
// _active_array. The caller must relinquish the array when done
// using it.
OopStorage::ActiveArray* OopStorage::obtain_active_array() const {
GlobalCounter::CriticalSection cs(Thread::current());
uint enter_value = _protect_active.read_enter();
ActiveArray* result = OrderAccess::load_acquire(&_active_array);
result->increment_refcount();
_protect_active.read_exit(enter_value);
return result;
}

View File

@ -204,6 +204,19 @@ NOT_AIX( private: )
void unlink(const Block& block);
};
// RCU-inspired protection of access to _active_array.
class ProtectActive {
volatile uint _enter;
volatile uint _exit[2];
public:
ProtectActive();
uint read_enter();
void read_exit(uint enter_value);
void write_synchronize();
};
private:
const char* _name;
ActiveArray* _active_array;
@ -216,6 +229,9 @@ private:
// Volatile for racy unlocked accesses.
volatile size_t _allocation_count;
// Protection for _active_array.
mutable ProtectActive _protect_active;
// mutable because this gets set even for const iteration.
mutable bool _concurrent_iteration_active;