2007-12-01 00:00:00 +00:00
|
|
|
/*
|
2012-10-09 10:09:34 -07:00
|
|
|
* Copyright (c) 2000, 2012, Oracle and/or its affiliates. All rights reserved.
|
2007-12-01 00:00:00 +00:00
|
|
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
|
|
*
|
|
|
|
* This code is free software; you can redistribute it and/or modify it
|
|
|
|
* under the terms of the GNU General Public License version 2 only, as
|
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*
|
|
|
|
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
|
|
* version 2 for more details (a copy is included in the LICENSE file that
|
|
|
|
* accompanied this code).
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License version
|
|
|
|
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
|
|
*
|
2010-05-27 19:08:38 -07:00
|
|
|
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
|
|
* or visit www.oracle.com if you need additional information or have any
|
|
|
|
* questions.
|
2007-12-01 00:00:00 +00:00
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
2010-11-23 13:22:55 -08:00
|
|
|
#ifndef SHARE_VM_MEMORY_SPACE_INLINE_HPP
|
|
|
|
#define SHARE_VM_MEMORY_SPACE_INLINE_HPP
|
|
|
|
|
|
|
|
#include "gc_interface/collectedHeap.hpp"
|
|
|
|
#include "memory/space.hpp"
|
|
|
|
#include "memory/universe.hpp"
|
2014-05-08 15:37:17 +02:00
|
|
|
#include "runtime/prefetch.inline.hpp"
|
2010-11-23 13:22:55 -08:00
|
|
|
#include "runtime/safepoint.hpp"
|
|
|
|
|
2008-06-05 15:57:56 -07:00
|
|
|
inline HeapWord* Space::block_start(const void* p) {
|
|
|
|
return block_start_const(p);
|
|
|
|
}
|
|
|
|
|
2014-05-08 15:37:17 +02:00
|
|
|
#define SCAN_AND_FORWARD(cp,scan_limit,block_is_obj,block_size) { \
|
|
|
|
/* Compute the new addresses for the live objects and store it in the mark \
|
|
|
|
* Used by universe::mark_sweep_phase2() \
|
|
|
|
*/ \
|
|
|
|
HeapWord* compact_top; /* This is where we are currently compacting to. */ \
|
|
|
|
\
|
|
|
|
/* We're sure to be here before any objects are compacted into this \
|
|
|
|
* space, so this is a good time to initialize this: \
|
|
|
|
*/ \
|
|
|
|
set_compaction_top(bottom()); \
|
|
|
|
\
|
|
|
|
if (cp->space == NULL) { \
|
|
|
|
assert(cp->gen != NULL, "need a generation"); \
|
|
|
|
assert(cp->threshold == NULL, "just checking"); \
|
|
|
|
assert(cp->gen->first_compaction_space() == this, "just checking"); \
|
|
|
|
cp->space = cp->gen->first_compaction_space(); \
|
|
|
|
compact_top = cp->space->bottom(); \
|
|
|
|
cp->space->set_compaction_top(compact_top); \
|
|
|
|
cp->threshold = cp->space->initialize_threshold(); \
|
|
|
|
} else { \
|
|
|
|
compact_top = cp->space->compaction_top(); \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* We allow some amount of garbage towards the bottom of the space, so \
|
|
|
|
* we don't start compacting before there is a significant gain to be made.\
|
|
|
|
* Occasionally, we want to ensure a full compaction, which is determined \
|
|
|
|
* by the MarkSweepAlwaysCompactCount parameter. \
|
|
|
|
*/ \
|
|
|
|
uint invocations = MarkSweep::total_invocations(); \
|
|
|
|
bool skip_dead = ((invocations % MarkSweepAlwaysCompactCount) != 0); \
|
|
|
|
\
|
|
|
|
size_t allowed_deadspace = 0; \
|
|
|
|
if (skip_dead) { \
|
|
|
|
const size_t ratio = allowed_dead_ratio(); \
|
|
|
|
allowed_deadspace = (capacity() * ratio / 100) / HeapWordSize; \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
HeapWord* q = bottom(); \
|
|
|
|
HeapWord* t = scan_limit(); \
|
|
|
|
\
|
|
|
|
HeapWord* end_of_live= q; /* One byte beyond the last byte of the last \
|
|
|
|
live object. */ \
|
|
|
|
HeapWord* first_dead = end();/* The first dead object. */ \
|
|
|
|
LiveRange* liveRange = NULL; /* The current live range, recorded in the \
|
|
|
|
first header of preceding free area. */ \
|
|
|
|
_first_dead = first_dead; \
|
|
|
|
\
|
|
|
|
const intx interval = PrefetchScanIntervalInBytes; \
|
|
|
|
\
|
|
|
|
while (q < t) { \
|
|
|
|
assert(!block_is_obj(q) || \
|
|
|
|
oop(q)->mark()->is_marked() || oop(q)->mark()->is_unlocked() || \
|
|
|
|
oop(q)->mark()->has_bias_pattern(), \
|
|
|
|
"these are the only valid states during a mark sweep"); \
|
|
|
|
if (block_is_obj(q) && oop(q)->is_gc_marked()) { \
|
|
|
|
/* prefetch beyond q */ \
|
|
|
|
Prefetch::write(q, interval); \
|
|
|
|
size_t size = block_size(q); \
|
|
|
|
compact_top = cp->space->forward(oop(q), size, cp, compact_top); \
|
|
|
|
q += size; \
|
|
|
|
end_of_live = q; \
|
|
|
|
} else { \
|
|
|
|
/* run over all the contiguous dead objects */ \
|
|
|
|
HeapWord* end = q; \
|
|
|
|
do { \
|
|
|
|
/* prefetch beyond end */ \
|
|
|
|
Prefetch::write(end, interval); \
|
|
|
|
end += block_size(end); \
|
|
|
|
} while (end < t && (!block_is_obj(end) || !oop(end)->is_gc_marked()));\
|
|
|
|
\
|
|
|
|
/* see if we might want to pretend this object is alive so that \
|
|
|
|
* we don't have to compact quite as often. \
|
|
|
|
*/ \
|
|
|
|
if (allowed_deadspace > 0 && q == compact_top) { \
|
|
|
|
size_t sz = pointer_delta(end, q); \
|
|
|
|
if (insert_deadspace(allowed_deadspace, q, sz)) { \
|
|
|
|
compact_top = cp->space->forward(oop(q), sz, cp, compact_top); \
|
|
|
|
q = end; \
|
|
|
|
end_of_live = end; \
|
|
|
|
continue; \
|
|
|
|
} \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* otherwise, it really is a free region. */ \
|
|
|
|
\
|
|
|
|
/* for the previous LiveRange, record the end of the live objects. */ \
|
|
|
|
if (liveRange) { \
|
|
|
|
liveRange->set_end(q); \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* record the current LiveRange object. \
|
|
|
|
* liveRange->start() is overlaid on the mark word. \
|
|
|
|
*/ \
|
|
|
|
liveRange = (LiveRange*)q; \
|
|
|
|
liveRange->set_start(end); \
|
|
|
|
liveRange->set_end(end); \
|
|
|
|
\
|
|
|
|
/* see if this is the first dead region. */ \
|
|
|
|
if (q < first_dead) { \
|
|
|
|
first_dead = q; \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* move on to the next object */ \
|
|
|
|
q = end; \
|
|
|
|
} \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
assert(q == t, "just checking"); \
|
|
|
|
if (liveRange != NULL) { \
|
|
|
|
liveRange->set_end(q); \
|
|
|
|
} \
|
|
|
|
_end_of_live = end_of_live; \
|
|
|
|
if (end_of_live < first_dead) { \
|
|
|
|
first_dead = end_of_live; \
|
|
|
|
} \
|
|
|
|
_first_dead = first_dead; \
|
|
|
|
\
|
|
|
|
/* save the compaction_top of the compaction space. */ \
|
|
|
|
cp->space->set_compaction_top(compact_top); \
|
|
|
|
}
|
|
|
|
|
|
|
|
#define SCAN_AND_ADJUST_POINTERS(adjust_obj_size) { \
|
|
|
|
/* adjust all the interior pointers to point at the new locations of objects \
|
|
|
|
* Used by MarkSweep::mark_sweep_phase3() */ \
|
|
|
|
\
|
|
|
|
HeapWord* q = bottom(); \
|
|
|
|
HeapWord* t = _end_of_live; /* Established by "prepare_for_compaction". */ \
|
|
|
|
\
|
|
|
|
assert(_first_dead <= _end_of_live, "Stands to reason, no?"); \
|
|
|
|
\
|
|
|
|
if (q < t && _first_dead > q && \
|
|
|
|
!oop(q)->is_gc_marked()) { \
|
|
|
|
/* we have a chunk of the space which hasn't moved and we've \
|
|
|
|
* reinitialized the mark word during the previous pass, so we can't \
|
|
|
|
* use is_gc_marked for the traversal. */ \
|
|
|
|
HeapWord* end = _first_dead; \
|
|
|
|
\
|
|
|
|
while (q < end) { \
|
|
|
|
/* I originally tried to conjoin "block_start(q) == q" to the \
|
|
|
|
* assertion below, but that doesn't work, because you can't \
|
|
|
|
* accurately traverse previous objects to get to the current one \
|
|
|
|
* after their pointers have been \
|
|
|
|
* updated, until the actual compaction is done. dld, 4/00 */ \
|
|
|
|
assert(block_is_obj(q), \
|
|
|
|
"should be at block boundaries, and should be looking at objs"); \
|
|
|
|
\
|
|
|
|
/* point all the oops to the new location */ \
|
|
|
|
size_t size = oop(q)->adjust_pointers(); \
|
|
|
|
size = adjust_obj_size(size); \
|
|
|
|
\
|
|
|
|
q += size; \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
if (_first_dead == t) { \
|
|
|
|
q = t; \
|
|
|
|
} else { \
|
|
|
|
/* $$$ This is funky. Using this to read the previously written \
|
|
|
|
* LiveRange. See also use below. */ \
|
|
|
|
q = (HeapWord*)oop(_first_dead)->mark()->decode_pointer(); \
|
|
|
|
} \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
const intx interval = PrefetchScanIntervalInBytes; \
|
|
|
|
\
|
|
|
|
debug_only(HeapWord* prev_q = NULL); \
|
|
|
|
while (q < t) { \
|
|
|
|
/* prefetch beyond q */ \
|
|
|
|
Prefetch::write(q, interval); \
|
|
|
|
if (oop(q)->is_gc_marked()) { \
|
|
|
|
/* q is alive */ \
|
|
|
|
/* point all the oops to the new location */ \
|
|
|
|
size_t size = oop(q)->adjust_pointers(); \
|
|
|
|
size = adjust_obj_size(size); \
|
|
|
|
debug_only(prev_q = q); \
|
|
|
|
q += size; \
|
|
|
|
} else { \
|
|
|
|
/* q is not a live object, so its mark should point at the next \
|
|
|
|
* live object */ \
|
|
|
|
debug_only(prev_q = q); \
|
|
|
|
q = (HeapWord*) oop(q)->mark()->decode_pointer(); \
|
|
|
|
assert(q > prev_q, "we should be moving forward through memory"); \
|
|
|
|
} \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
assert(q == t, "just checking"); \
|
|
|
|
}
|
|
|
|
|
|
|
|
#define SCAN_AND_COMPACT(obj_size) { \
|
|
|
|
/* Copy all live objects to their new location \
|
|
|
|
* Used by MarkSweep::mark_sweep_phase4() */ \
|
|
|
|
\
|
|
|
|
HeapWord* q = bottom(); \
|
|
|
|
HeapWord* const t = _end_of_live; \
|
|
|
|
debug_only(HeapWord* prev_q = NULL); \
|
|
|
|
\
|
|
|
|
if (q < t && _first_dead > q && \
|
|
|
|
!oop(q)->is_gc_marked()) { \
|
|
|
|
debug_only( \
|
|
|
|
/* we have a chunk of the space which hasn't moved and we've reinitialized \
|
|
|
|
* the mark word during the previous pass, so we can't use is_gc_marked for \
|
|
|
|
* the traversal. */ \
|
|
|
|
HeapWord* const end = _first_dead; \
|
|
|
|
\
|
|
|
|
while (q < end) { \
|
|
|
|
size_t size = obj_size(q); \
|
|
|
|
assert(!oop(q)->is_gc_marked(), \
|
|
|
|
"should be unmarked (special dense prefix handling)"); \
|
|
|
|
debug_only(prev_q = q); \
|
|
|
|
q += size; \
|
|
|
|
} \
|
|
|
|
) /* debug_only */ \
|
|
|
|
\
|
|
|
|
if (_first_dead == t) { \
|
|
|
|
q = t; \
|
|
|
|
} else { \
|
|
|
|
/* $$$ Funky */ \
|
|
|
|
q = (HeapWord*) oop(_first_dead)->mark()->decode_pointer(); \
|
|
|
|
} \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
const intx scan_interval = PrefetchScanIntervalInBytes; \
|
|
|
|
const intx copy_interval = PrefetchCopyIntervalInBytes; \
|
|
|
|
while (q < t) { \
|
|
|
|
if (!oop(q)->is_gc_marked()) { \
|
|
|
|
/* mark is pointer to next marked oop */ \
|
|
|
|
debug_only(prev_q = q); \
|
|
|
|
q = (HeapWord*) oop(q)->mark()->decode_pointer(); \
|
|
|
|
assert(q > prev_q, "we should be moving forward through memory"); \
|
|
|
|
} else { \
|
|
|
|
/* prefetch beyond q */ \
|
|
|
|
Prefetch::read(q, scan_interval); \
|
|
|
|
\
|
|
|
|
/* size and destination */ \
|
|
|
|
size_t size = obj_size(q); \
|
|
|
|
HeapWord* compaction_top = (HeapWord*)oop(q)->forwardee(); \
|
|
|
|
\
|
|
|
|
/* prefetch beyond compaction_top */ \
|
|
|
|
Prefetch::write(compaction_top, copy_interval); \
|
|
|
|
\
|
|
|
|
/* copy object and reinit its mark */ \
|
|
|
|
assert(q != compaction_top, "everything in this pass should be moving"); \
|
|
|
|
Copy::aligned_conjoint_words(q, compaction_top, size); \
|
|
|
|
oop(compaction_top)->init_mark(); \
|
|
|
|
assert(oop(compaction_top)->klass() != NULL, "should have a class"); \
|
|
|
|
\
|
|
|
|
debug_only(prev_q = q); \
|
|
|
|
q += size; \
|
|
|
|
} \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* Let's remember if we were empty before we did the compaction. */ \
|
|
|
|
bool was_empty = used_region().is_empty(); \
|
|
|
|
/* Reset space after compaction is complete */ \
|
|
|
|
reset_after_compaction(); \
|
|
|
|
/* We do this clear, below, since it has overloaded meanings for some */ \
|
|
|
|
/* space subtypes. For example, OffsetTableContigSpace's that were */ \
|
|
|
|
/* compacted into will have had their offset table thresholds updated */ \
|
|
|
|
/* continuously, but those that weren't need to have their thresholds */ \
|
|
|
|
/* re-initialized. Also mangles unused area for debugging. */ \
|
|
|
|
if (used_region().is_empty()) { \
|
|
|
|
if (!was_empty) clear(SpaceDecorator::Mangle); \
|
|
|
|
} else { \
|
|
|
|
if (ZapUnusedHeapArea) mangle_unused_area(); \
|
|
|
|
} \
|
|
|
|
}
|
|
|
|
|
2007-12-01 00:00:00 +00:00
|
|
|
inline HeapWord* OffsetTableContigSpace::allocate(size_t size) {
|
|
|
|
HeapWord* res = ContiguousSpace::allocate(size);
|
|
|
|
if (res != NULL) {
|
|
|
|
_offsets.alloc_block(res, size);
|
|
|
|
}
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Because of the requirement of keeping "_offsets" up to date with the
|
|
|
|
// allocations, we sequentialize these with a lock. Therefore, best if
|
|
|
|
// this is used for larger LAB allocations only.
|
|
|
|
inline HeapWord* OffsetTableContigSpace::par_allocate(size_t size) {
|
|
|
|
MutexLocker x(&_par_alloc_lock);
|
|
|
|
// This ought to be just "allocate", because of the lock above, but that
|
|
|
|
// ContiguousSpace::allocate asserts that either the allocating thread
|
|
|
|
// holds the heap lock or it is the VM thread and we're at a safepoint.
|
|
|
|
// The best I (dld) could figure was to put a field in ContiguousSpace
|
|
|
|
// meaning "locking at safepoint taken care of", and set/reset that
|
|
|
|
// here. But this will do for now, especially in light of the comment
|
|
|
|
// above. Perhaps in the future some lock-free manner of keeping the
|
|
|
|
// coordination.
|
|
|
|
HeapWord* res = ContiguousSpace::par_allocate(size);
|
|
|
|
if (res != NULL) {
|
|
|
|
_offsets.alloc_block(res, size);
|
|
|
|
}
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
2008-06-05 15:57:56 -07:00
|
|
|
inline HeapWord*
|
|
|
|
OffsetTableContigSpace::block_start_const(const void* p) const {
|
2007-12-01 00:00:00 +00:00
|
|
|
return _offsets.block_start(p);
|
|
|
|
}
|
|
|
|
|
2010-11-23 13:22:55 -08:00
|
|
|
#endif // SHARE_VM_MEMORY_SPACE_INLINE_HPP
|