8329494: Serial: Merge GenMarkSweep into MarkSweep
Reviewed-by: ihse, ayang, tschatzl
This commit is contained in:
parent
802018306f
commit
41966885b9
@ -1,5 +1,5 @@
|
|||||||
#
|
#
|
||||||
# Copyright (c) 2013, 2023, Oracle and/or its affiliates. All rights reserved.
|
# Copyright (c) 2013, 2024, Oracle and/or its affiliates. All rights reserved.
|
||||||
# DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
# DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
||||||
#
|
#
|
||||||
# This code is free software; you can redistribute it and/or modify it
|
# This code is free software; you can redistribute it and/or modify it
|
||||||
@ -214,7 +214,6 @@ ifeq ($(call check-jvm-feature, opt-size), true)
|
|||||||
frame_x86.cpp \
|
frame_x86.cpp \
|
||||||
genCollectedHeap.cpp \
|
genCollectedHeap.cpp \
|
||||||
generation.cpp \
|
generation.cpp \
|
||||||
genMarkSweep.cpp \
|
|
||||||
growableArray.cpp \
|
growableArray.cpp \
|
||||||
handles.cpp \
|
handles.cpp \
|
||||||
hashtable.cpp \
|
hashtable.cpp \
|
||||||
|
@ -1,544 +0,0 @@
|
|||||||
/*
|
|
||||||
* Copyright (c) 2001, 2023, Oracle and/or its affiliates. All rights reserved.
|
|
||||||
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
||||||
*
|
|
||||||
* This code is free software; you can redistribute it and/or modify it
|
|
||||||
* under the terms of the GNU General Public License version 2 only, as
|
|
||||||
* published by the Free Software Foundation.
|
|
||||||
*
|
|
||||||
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
||||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
||||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
||||||
* version 2 for more details (a copy is included in the LICENSE file that
|
|
||||||
* accompanied this code).
|
|
||||||
*
|
|
||||||
* You should have received a copy of the GNU General Public License version
|
|
||||||
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
||||||
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
||||||
*
|
|
||||||
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
||||||
* or visit www.oracle.com if you need additional information or have any
|
|
||||||
* questions.
|
|
||||||
*
|
|
||||||
*/
|
|
||||||
|
|
||||||
#include "precompiled.hpp"
|
|
||||||
#include "classfile/classLoaderDataGraph.hpp"
|
|
||||||
#include "classfile/javaClasses.hpp"
|
|
||||||
#include "classfile/stringTable.hpp"
|
|
||||||
#include "classfile/symbolTable.hpp"
|
|
||||||
#include "classfile/systemDictionary.hpp"
|
|
||||||
#include "classfile/vmSymbols.hpp"
|
|
||||||
#include "code/codeCache.hpp"
|
|
||||||
#include "compiler/oopMap.hpp"
|
|
||||||
#include "gc/serial/cardTableRS.hpp"
|
|
||||||
#include "gc/serial/defNewGeneration.hpp"
|
|
||||||
#include "gc/serial/generation.hpp"
|
|
||||||
#include "gc/serial/genMarkSweep.hpp"
|
|
||||||
#include "gc/serial/markSweep.inline.hpp"
|
|
||||||
#include "gc/serial/serialGcRefProcProxyTask.hpp"
|
|
||||||
#include "gc/serial/serialHeap.hpp"
|
|
||||||
#include "gc/shared/classUnloadingContext.hpp"
|
|
||||||
#include "gc/shared/collectedHeap.inline.hpp"
|
|
||||||
#include "gc/shared/gcHeapSummary.hpp"
|
|
||||||
#include "gc/shared/gcTimer.hpp"
|
|
||||||
#include "gc/shared/gcTrace.hpp"
|
|
||||||
#include "gc/shared/gcTraceTime.inline.hpp"
|
|
||||||
#include "gc/shared/modRefBarrierSet.hpp"
|
|
||||||
#include "gc/shared/preservedMarks.inline.hpp"
|
|
||||||
#include "gc/shared/referencePolicy.hpp"
|
|
||||||
#include "gc/shared/referenceProcessorPhaseTimes.hpp"
|
|
||||||
#include "gc/shared/space.inline.hpp"
|
|
||||||
#include "gc/shared/strongRootsScope.hpp"
|
|
||||||
#include "gc/shared/weakProcessor.hpp"
|
|
||||||
#include "memory/universe.hpp"
|
|
||||||
#include "oops/instanceRefKlass.hpp"
|
|
||||||
#include "oops/oop.inline.hpp"
|
|
||||||
#include "prims/jvmtiExport.hpp"
|
|
||||||
#include "runtime/handles.inline.hpp"
|
|
||||||
#include "runtime/javaThread.hpp"
|
|
||||||
#include "runtime/prefetch.inline.hpp"
|
|
||||||
#include "runtime/synchronizer.hpp"
|
|
||||||
#include "runtime/vmThread.hpp"
|
|
||||||
#include "utilities/copy.hpp"
|
|
||||||
#include "utilities/events.hpp"
|
|
||||||
#include "utilities/stack.inline.hpp"
|
|
||||||
#if INCLUDE_JVMCI
|
|
||||||
#include "jvmci/jvmci.hpp"
|
|
||||||
#endif
|
|
||||||
|
|
||||||
class DeadSpacer : StackObj {
|
|
||||||
size_t _allowed_deadspace_words;
|
|
||||||
bool _active;
|
|
||||||
ContiguousSpace* _space;
|
|
||||||
|
|
||||||
public:
|
|
||||||
DeadSpacer(ContiguousSpace* space) : _allowed_deadspace_words(0), _space(space) {
|
|
||||||
size_t ratio = _space->allowed_dead_ratio();
|
|
||||||
_active = ratio > 0;
|
|
||||||
|
|
||||||
if (_active) {
|
|
||||||
// We allow some amount of garbage towards the bottom of the space, so
|
|
||||||
// we don't start compacting before there is a significant gain to be made.
|
|
||||||
// Occasionally, we want to ensure a full compaction, which is determined
|
|
||||||
// by the MarkSweepAlwaysCompactCount parameter.
|
|
||||||
if ((MarkSweep::total_invocations() % MarkSweepAlwaysCompactCount) != 0) {
|
|
||||||
_allowed_deadspace_words = (space->capacity() * ratio / 100) / HeapWordSize;
|
|
||||||
} else {
|
|
||||||
_active = false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
bool insert_deadspace(HeapWord* dead_start, HeapWord* dead_end) {
|
|
||||||
if (!_active) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
size_t dead_length = pointer_delta(dead_end, dead_start);
|
|
||||||
if (_allowed_deadspace_words >= dead_length) {
|
|
||||||
_allowed_deadspace_words -= dead_length;
|
|
||||||
CollectedHeap::fill_with_object(dead_start, dead_length);
|
|
||||||
oop obj = cast_to_oop(dead_start);
|
|
||||||
// obj->set_mark(obj->mark().set_marked());
|
|
||||||
|
|
||||||
assert(dead_length == obj->size(), "bad filler object size");
|
|
||||||
log_develop_trace(gc, compaction)("Inserting object to dead space: " PTR_FORMAT ", " PTR_FORMAT ", " SIZE_FORMAT "b",
|
|
||||||
p2i(dead_start), p2i(dead_end), dead_length * HeapWordSize);
|
|
||||||
|
|
||||||
return true;
|
|
||||||
} else {
|
|
||||||
_active = false;
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
// Implement the "compaction" part of the mark-compact GC algorithm.
|
|
||||||
class Compacter {
|
|
||||||
// There are four spaces in total, but only the first three can be used after
|
|
||||||
// compact. IOW, old and eden/from must be enough for all live objs
|
|
||||||
static constexpr uint max_num_spaces = 4;
|
|
||||||
|
|
||||||
struct CompactionSpace {
|
|
||||||
ContiguousSpace* _space;
|
|
||||||
// Will be the new top after compaction is complete.
|
|
||||||
HeapWord* _compaction_top;
|
|
||||||
// The first dead word in this contiguous space. It's an optimization to
|
|
||||||
// skip large chunk of live objects at the beginning.
|
|
||||||
HeapWord* _first_dead;
|
|
||||||
|
|
||||||
void init(ContiguousSpace* space) {
|
|
||||||
_space = space;
|
|
||||||
_compaction_top = space->bottom();
|
|
||||||
_first_dead = nullptr;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
CompactionSpace _spaces[max_num_spaces];
|
|
||||||
// The num of spaces to be compacted, i.e. containing live objs.
|
|
||||||
uint _num_spaces;
|
|
||||||
|
|
||||||
uint _index;
|
|
||||||
|
|
||||||
HeapWord* get_compaction_top(uint index) const {
|
|
||||||
return _spaces[index]._compaction_top;
|
|
||||||
}
|
|
||||||
|
|
||||||
HeapWord* get_first_dead(uint index) const {
|
|
||||||
return _spaces[index]._first_dead;
|
|
||||||
}
|
|
||||||
|
|
||||||
ContiguousSpace* get_space(uint index) const {
|
|
||||||
return _spaces[index]._space;
|
|
||||||
}
|
|
||||||
|
|
||||||
void record_first_dead(uint index, HeapWord* first_dead) {
|
|
||||||
assert(_spaces[index]._first_dead == nullptr, "should write only once");
|
|
||||||
_spaces[index]._first_dead = first_dead;
|
|
||||||
}
|
|
||||||
|
|
||||||
HeapWord* alloc(size_t words) {
|
|
||||||
while (true) {
|
|
||||||
if (words <= pointer_delta(_spaces[_index]._space->end(),
|
|
||||||
_spaces[_index]._compaction_top)) {
|
|
||||||
HeapWord* result = _spaces[_index]._compaction_top;
|
|
||||||
_spaces[_index]._compaction_top += words;
|
|
||||||
if (_index == 0) {
|
|
||||||
// old-gen requires BOT update
|
|
||||||
static_cast<TenuredSpace*>(_spaces[0]._space)->update_for_block(result, result + words);
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
|
|
||||||
// out-of-memory in this space
|
|
||||||
_index++;
|
|
||||||
assert(_index < max_num_spaces - 1, "the last space should not be used");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static void prefetch_read_scan(void* p) {
|
|
||||||
if (PrefetchScanIntervalInBytes >= 0) {
|
|
||||||
Prefetch::read(p, PrefetchScanIntervalInBytes);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static void prefetch_write_scan(void* p) {
|
|
||||||
if (PrefetchScanIntervalInBytes >= 0) {
|
|
||||||
Prefetch::write(p, PrefetchScanIntervalInBytes);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static void prefetch_write_copy(void* p) {
|
|
||||||
if (PrefetchCopyIntervalInBytes >= 0) {
|
|
||||||
Prefetch::write(p, PrefetchCopyIntervalInBytes);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static void forward_obj(oop obj, HeapWord* new_addr) {
|
|
||||||
prefetch_write_scan(obj);
|
|
||||||
if (cast_from_oop<HeapWord*>(obj) != new_addr) {
|
|
||||||
obj->forward_to(cast_to_oop(new_addr));
|
|
||||||
} else {
|
|
||||||
assert(obj->is_gc_marked(), "inv");
|
|
||||||
// This obj will stay in-place. Fix the markword.
|
|
||||||
obj->init_mark();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static HeapWord* find_next_live_addr(HeapWord* start, HeapWord* end) {
|
|
||||||
for (HeapWord* i_addr = start; i_addr < end; /* empty */) {
|
|
||||||
prefetch_read_scan(i_addr);
|
|
||||||
oop obj = cast_to_oop(i_addr);
|
|
||||||
if (obj->is_gc_marked()) {
|
|
||||||
return i_addr;
|
|
||||||
}
|
|
||||||
i_addr += obj->size();
|
|
||||||
}
|
|
||||||
return end;
|
|
||||||
};
|
|
||||||
|
|
||||||
static size_t relocate(HeapWord* addr) {
|
|
||||||
// Prefetch source and destination
|
|
||||||
prefetch_read_scan(addr);
|
|
||||||
|
|
||||||
oop obj = cast_to_oop(addr);
|
|
||||||
oop new_obj = obj->forwardee();
|
|
||||||
HeapWord* new_addr = cast_from_oop<HeapWord*>(new_obj);
|
|
||||||
assert(addr != new_addr, "inv");
|
|
||||||
prefetch_write_copy(new_addr);
|
|
||||||
|
|
||||||
size_t obj_size = obj->size();
|
|
||||||
Copy::aligned_conjoint_words(addr, new_addr, obj_size);
|
|
||||||
new_obj->init_mark();
|
|
||||||
|
|
||||||
return obj_size;
|
|
||||||
}
|
|
||||||
|
|
||||||
public:
|
|
||||||
explicit Compacter(SerialHeap* heap) {
|
|
||||||
// In this order so that heap is compacted towards old-gen.
|
|
||||||
_spaces[0].init(heap->old_gen()->space());
|
|
||||||
_spaces[1].init(heap->young_gen()->eden());
|
|
||||||
_spaces[2].init(heap->young_gen()->from());
|
|
||||||
|
|
||||||
bool is_promotion_failed = (heap->young_gen()->from()->next_compaction_space() != nullptr);
|
|
||||||
if (is_promotion_failed) {
|
|
||||||
_spaces[3].init(heap->young_gen()->to());
|
|
||||||
_num_spaces = 4;
|
|
||||||
} else {
|
|
||||||
_num_spaces = 3;
|
|
||||||
}
|
|
||||||
_index = 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
void phase2_calculate_new_addr() {
|
|
||||||
for (uint i = 0; i < _num_spaces; ++i) {
|
|
||||||
ContiguousSpace* space = get_space(i);
|
|
||||||
HeapWord* cur_addr = space->bottom();
|
|
||||||
HeapWord* top = space->top();
|
|
||||||
|
|
||||||
bool record_first_dead_done = false;
|
|
||||||
|
|
||||||
DeadSpacer dead_spacer(space);
|
|
||||||
|
|
||||||
while (cur_addr < top) {
|
|
||||||
oop obj = cast_to_oop(cur_addr);
|
|
||||||
size_t obj_size = obj->size();
|
|
||||||
if (obj->is_gc_marked()) {
|
|
||||||
HeapWord* new_addr = alloc(obj_size);
|
|
||||||
forward_obj(obj, new_addr);
|
|
||||||
cur_addr += obj_size;
|
|
||||||
} else {
|
|
||||||
// Skipping the current known-unmarked obj
|
|
||||||
HeapWord* next_live_addr = find_next_live_addr(cur_addr + obj_size, top);
|
|
||||||
if (dead_spacer.insert_deadspace(cur_addr, next_live_addr)) {
|
|
||||||
// Register space for the filler obj
|
|
||||||
alloc(pointer_delta(next_live_addr, cur_addr));
|
|
||||||
} else {
|
|
||||||
if (!record_first_dead_done) {
|
|
||||||
record_first_dead(i, cur_addr);
|
|
||||||
record_first_dead_done = true;
|
|
||||||
}
|
|
||||||
*(HeapWord**)cur_addr = next_live_addr;
|
|
||||||
}
|
|
||||||
cur_addr = next_live_addr;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!record_first_dead_done) {
|
|
||||||
record_first_dead(i, top);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void phase3_adjust_pointers() {
|
|
||||||
for (uint i = 0; i < _num_spaces; ++i) {
|
|
||||||
ContiguousSpace* space = get_space(i);
|
|
||||||
HeapWord* cur_addr = space->bottom();
|
|
||||||
HeapWord* const top = space->top();
|
|
||||||
HeapWord* const first_dead = get_first_dead(i);
|
|
||||||
|
|
||||||
while (cur_addr < top) {
|
|
||||||
prefetch_write_scan(cur_addr);
|
|
||||||
if (cur_addr < first_dead || cast_to_oop(cur_addr)->is_gc_marked()) {
|
|
||||||
size_t size = MarkSweep::adjust_pointers(cast_to_oop(cur_addr));
|
|
||||||
cur_addr += size;
|
|
||||||
} else {
|
|
||||||
assert(*(HeapWord**)cur_addr > cur_addr, "forward progress");
|
|
||||||
cur_addr = *(HeapWord**)cur_addr;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void phase4_compact() {
|
|
||||||
for (uint i = 0; i < _num_spaces; ++i) {
|
|
||||||
ContiguousSpace* space = get_space(i);
|
|
||||||
HeapWord* cur_addr = space->bottom();
|
|
||||||
HeapWord* top = space->top();
|
|
||||||
|
|
||||||
// Check if the first obj inside this space is forwarded.
|
|
||||||
if (!cast_to_oop(cur_addr)->is_forwarded()) {
|
|
||||||
// Jump over consecutive (in-place) live-objs-chunk
|
|
||||||
cur_addr = get_first_dead(i);
|
|
||||||
}
|
|
||||||
|
|
||||||
while (cur_addr < top) {
|
|
||||||
if (!cast_to_oop(cur_addr)->is_forwarded()) {
|
|
||||||
cur_addr = *(HeapWord**) cur_addr;
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
cur_addr += relocate(cur_addr);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Reset top and unused memory
|
|
||||||
space->set_top(get_compaction_top(i));
|
|
||||||
if (ZapUnusedHeapArea) {
|
|
||||||
space->mangle_unused_area();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
void GenMarkSweep::phase1_mark(bool clear_all_softrefs) {
|
|
||||||
// Recursively traverse all live objects and mark them
|
|
||||||
GCTraceTime(Info, gc, phases) tm("Phase 1: Mark live objects", _gc_timer);
|
|
||||||
|
|
||||||
SerialHeap* gch = SerialHeap::heap();
|
|
||||||
|
|
||||||
ClassLoaderDataGraph::verify_claimed_marks_cleared(ClassLoaderData::_claim_stw_fullgc_mark);
|
|
||||||
|
|
||||||
ref_processor()->start_discovery(clear_all_softrefs);
|
|
||||||
|
|
||||||
{
|
|
||||||
StrongRootsScope srs(0);
|
|
||||||
|
|
||||||
CLDClosure* weak_cld_closure = ClassUnloading ? nullptr : &follow_cld_closure;
|
|
||||||
MarkingCodeBlobClosure mark_code_closure(&follow_root_closure, !CodeBlobToOopClosure::FixRelocations, true);
|
|
||||||
gch->process_roots(SerialHeap::SO_None,
|
|
||||||
&follow_root_closure,
|
|
||||||
&follow_cld_closure,
|
|
||||||
weak_cld_closure,
|
|
||||||
&mark_code_closure);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Process reference objects found during marking
|
|
||||||
{
|
|
||||||
GCTraceTime(Debug, gc, phases) tm_m("Reference Processing", gc_timer());
|
|
||||||
|
|
||||||
ReferenceProcessorPhaseTimes pt(_gc_timer, ref_processor()->max_num_queues());
|
|
||||||
SerialGCRefProcProxyTask task(is_alive, keep_alive, follow_stack_closure);
|
|
||||||
const ReferenceProcessorStats& stats = ref_processor()->process_discovered_references(task, pt);
|
|
||||||
pt.print_all_references();
|
|
||||||
gc_tracer()->report_gc_reference_stats(stats);
|
|
||||||
}
|
|
||||||
|
|
||||||
// This is the point where the entire marking should have completed.
|
|
||||||
assert(_marking_stack.is_empty(), "Marking should have completed");
|
|
||||||
|
|
||||||
{
|
|
||||||
GCTraceTime(Debug, gc, phases) tm_m("Weak Processing", gc_timer());
|
|
||||||
WeakProcessor::weak_oops_do(&is_alive, &do_nothing_cl);
|
|
||||||
}
|
|
||||||
|
|
||||||
{
|
|
||||||
GCTraceTime(Debug, gc, phases) tm_m("Class Unloading", gc_timer());
|
|
||||||
|
|
||||||
ClassUnloadingContext* ctx = ClassUnloadingContext::context();
|
|
||||||
|
|
||||||
bool unloading_occurred;
|
|
||||||
{
|
|
||||||
CodeCache::UnlinkingScope scope(&is_alive);
|
|
||||||
|
|
||||||
// Unload classes and purge the SystemDictionary.
|
|
||||||
unloading_occurred = SystemDictionary::do_unloading(gc_timer());
|
|
||||||
|
|
||||||
// Unload nmethods.
|
|
||||||
CodeCache::do_unloading(unloading_occurred);
|
|
||||||
}
|
|
||||||
|
|
||||||
{
|
|
||||||
GCTraceTime(Debug, gc, phases) t("Purge Unlinked NMethods", gc_timer());
|
|
||||||
// Release unloaded nmethod's memory.
|
|
||||||
ctx->purge_nmethods();
|
|
||||||
}
|
|
||||||
{
|
|
||||||
GCTraceTime(Debug, gc, phases) ur("Unregister NMethods", gc_timer());
|
|
||||||
gch->prune_unlinked_nmethods();
|
|
||||||
}
|
|
||||||
{
|
|
||||||
GCTraceTime(Debug, gc, phases) t("Free Code Blobs", gc_timer());
|
|
||||||
ctx->free_code_blobs();
|
|
||||||
}
|
|
||||||
|
|
||||||
// Prune dead klasses from subklass/sibling/implementor lists.
|
|
||||||
Klass::clean_weak_klass_links(unloading_occurred);
|
|
||||||
|
|
||||||
// Clean JVMCI metadata handles.
|
|
||||||
JVMCI_ONLY(JVMCI::do_unloading(unloading_occurred));
|
|
||||||
}
|
|
||||||
|
|
||||||
{
|
|
||||||
GCTraceTime(Debug, gc, phases) tm_m("Report Object Count", gc_timer());
|
|
||||||
gc_tracer()->report_object_count_after_gc(&is_alive, nullptr);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void GenMarkSweep::invoke_at_safepoint(bool clear_all_softrefs) {
|
|
||||||
assert(SafepointSynchronize::is_at_safepoint(), "must be at a safepoint");
|
|
||||||
|
|
||||||
SerialHeap* gch = SerialHeap::heap();
|
|
||||||
#ifdef ASSERT
|
|
||||||
if (gch->soft_ref_policy()->should_clear_all_soft_refs()) {
|
|
||||||
assert(clear_all_softrefs, "Policy should have been checked earlier");
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
gch->trace_heap_before_gc(_gc_tracer);
|
|
||||||
|
|
||||||
// Increment the invocation count
|
|
||||||
_total_invocations++;
|
|
||||||
|
|
||||||
// Capture used regions for old-gen to reestablish old-to-young invariant
|
|
||||||
// after full-gc.
|
|
||||||
gch->old_gen()->save_used_region();
|
|
||||||
|
|
||||||
allocate_stacks();
|
|
||||||
|
|
||||||
phase1_mark(clear_all_softrefs);
|
|
||||||
|
|
||||||
Compacter compacter{gch};
|
|
||||||
|
|
||||||
{
|
|
||||||
// Now all live objects are marked, compute the new object addresses.
|
|
||||||
GCTraceTime(Info, gc, phases) tm("Phase 2: Compute new object addresses", _gc_timer);
|
|
||||||
|
|
||||||
compacter.phase2_calculate_new_addr();
|
|
||||||
}
|
|
||||||
|
|
||||||
// Don't add any more derived pointers during phase3
|
|
||||||
#if COMPILER2_OR_JVMCI
|
|
||||||
assert(DerivedPointerTable::is_active(), "Sanity");
|
|
||||||
DerivedPointerTable::set_active(false);
|
|
||||||
#endif
|
|
||||||
|
|
||||||
{
|
|
||||||
// Adjust the pointers to reflect the new locations
|
|
||||||
GCTraceTime(Info, gc, phases) tm("Phase 3: Adjust pointers", gc_timer());
|
|
||||||
|
|
||||||
ClassLoaderDataGraph::verify_claimed_marks_cleared(ClassLoaderData::_claim_stw_fullgc_adjust);
|
|
||||||
|
|
||||||
CodeBlobToOopClosure code_closure(&adjust_pointer_closure, CodeBlobToOopClosure::FixRelocations);
|
|
||||||
gch->process_roots(SerialHeap::SO_AllCodeCache,
|
|
||||||
&adjust_pointer_closure,
|
|
||||||
&adjust_cld_closure,
|
|
||||||
&adjust_cld_closure,
|
|
||||||
&code_closure);
|
|
||||||
|
|
||||||
WeakProcessor::oops_do(&adjust_pointer_closure);
|
|
||||||
|
|
||||||
adjust_marks();
|
|
||||||
compacter.phase3_adjust_pointers();
|
|
||||||
}
|
|
||||||
|
|
||||||
{
|
|
||||||
// All pointers are now adjusted, move objects accordingly
|
|
||||||
GCTraceTime(Info, gc, phases) tm("Phase 4: Move objects", _gc_timer);
|
|
||||||
|
|
||||||
compacter.phase4_compact();
|
|
||||||
}
|
|
||||||
|
|
||||||
restore_marks();
|
|
||||||
|
|
||||||
// Set saved marks for allocation profiler (and other things? -- dld)
|
|
||||||
// (Should this be in general part?)
|
|
||||||
gch->save_marks();
|
|
||||||
|
|
||||||
deallocate_stacks();
|
|
||||||
|
|
||||||
MarkSweep::_string_dedup_requests->flush();
|
|
||||||
|
|
||||||
bool is_young_gen_empty = (gch->young_gen()->used() == 0);
|
|
||||||
gch->rem_set()->maintain_old_to_young_invariant(gch->old_gen(), is_young_gen_empty);
|
|
||||||
|
|
||||||
gch->prune_scavengable_nmethods();
|
|
||||||
|
|
||||||
// Update heap occupancy information which is used as
|
|
||||||
// input to soft ref clearing policy at the next gc.
|
|
||||||
Universe::heap()->update_capacity_and_used_at_gc();
|
|
||||||
|
|
||||||
// Signal that we have completed a visit to all live objects.
|
|
||||||
Universe::heap()->record_whole_heap_examined_timestamp();
|
|
||||||
|
|
||||||
gch->trace_heap_after_gc(_gc_tracer);
|
|
||||||
}
|
|
||||||
|
|
||||||
void GenMarkSweep::allocate_stacks() {
|
|
||||||
void* scratch = nullptr;
|
|
||||||
size_t num_words;
|
|
||||||
DefNewGeneration* young_gen = (DefNewGeneration*)SerialHeap::heap()->young_gen();
|
|
||||||
young_gen->contribute_scratch(scratch, num_words);
|
|
||||||
|
|
||||||
if (scratch != nullptr) {
|
|
||||||
_preserved_count_max = num_words * HeapWordSize / sizeof(PreservedMark);
|
|
||||||
} else {
|
|
||||||
_preserved_count_max = 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
_preserved_marks = (PreservedMark*)scratch;
|
|
||||||
_preserved_count = 0;
|
|
||||||
|
|
||||||
_preserved_overflow_stack_set.init(1);
|
|
||||||
}
|
|
||||||
|
|
||||||
void GenMarkSweep::deallocate_stacks() {
|
|
||||||
if (_preserved_count_max != 0) {
|
|
||||||
DefNewGeneration* young_gen = (DefNewGeneration*)SerialHeap::heap()->young_gen();
|
|
||||||
young_gen->reset_scratch();
|
|
||||||
}
|
|
||||||
|
|
||||||
_preserved_overflow_stack_set.reclaim();
|
|
||||||
_marking_stack.clear();
|
|
||||||
_objarray_stack.clear(true);
|
|
||||||
}
|
|
@ -1,43 +0,0 @@
|
|||||||
/*
|
|
||||||
* Copyright (c) 2001, 2019, Oracle and/or its affiliates. All rights reserved.
|
|
||||||
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
||||||
*
|
|
||||||
* This code is free software; you can redistribute it and/or modify it
|
|
||||||
* under the terms of the GNU General Public License version 2 only, as
|
|
||||||
* published by the Free Software Foundation.
|
|
||||||
*
|
|
||||||
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
||||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
||||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
||||||
* version 2 for more details (a copy is included in the LICENSE file that
|
|
||||||
* accompanied this code).
|
|
||||||
*
|
|
||||||
* You should have received a copy of the GNU General Public License version
|
|
||||||
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
||||||
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
||||||
*
|
|
||||||
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
||||||
* or visit www.oracle.com if you need additional information or have any
|
|
||||||
* questions.
|
|
||||||
*
|
|
||||||
*/
|
|
||||||
|
|
||||||
#ifndef SHARE_GC_SERIAL_GENMARKSWEEP_HPP
|
|
||||||
#define SHARE_GC_SERIAL_GENMARKSWEEP_HPP
|
|
||||||
|
|
||||||
#include "gc/serial/markSweep.hpp"
|
|
||||||
|
|
||||||
class GenMarkSweep : public MarkSweep {
|
|
||||||
public:
|
|
||||||
static void invoke_at_safepoint(bool clear_all_softrefs);
|
|
||||||
|
|
||||||
private:
|
|
||||||
// Mark live objects
|
|
||||||
static void phase1_mark(bool clear_all_softrefs);
|
|
||||||
|
|
||||||
// Temporary data structures for traversal and storing/restoring marks
|
|
||||||
static void allocate_stacks();
|
|
||||||
static void deallocate_stacks();
|
|
||||||
};
|
|
||||||
|
|
||||||
#endif // SHARE_GC_SERIAL_GENMARKSWEEP_HPP
|
|
@ -1,5 +1,5 @@
|
|||||||
/*
|
/*
|
||||||
* Copyright (c) 1997, 2023, Oracle and/or its affiliates. All rights reserved.
|
* Copyright (c) 1997, 2024, Oracle and/or its affiliates. All rights reserved.
|
||||||
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
||||||
*
|
*
|
||||||
* This code is free software; you can redistribute it and/or modify it
|
* This code is free software; you can redistribute it and/or modify it
|
||||||
@ -23,21 +23,49 @@
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
#include "precompiled.hpp"
|
#include "precompiled.hpp"
|
||||||
|
#include "classfile/classLoaderDataGraph.hpp"
|
||||||
|
#include "classfile/javaClasses.hpp"
|
||||||
|
#include "classfile/stringTable.hpp"
|
||||||
|
#include "classfile/symbolTable.hpp"
|
||||||
|
#include "classfile/systemDictionary.hpp"
|
||||||
|
#include "classfile/vmSymbols.hpp"
|
||||||
|
#include "code/codeCache.hpp"
|
||||||
#include "compiler/compileBroker.hpp"
|
#include "compiler/compileBroker.hpp"
|
||||||
|
#include "compiler/oopMap.hpp"
|
||||||
|
#include "gc/serial/cardTableRS.hpp"
|
||||||
|
#include "gc/serial/defNewGeneration.hpp"
|
||||||
#include "gc/serial/markSweep.inline.hpp"
|
#include "gc/serial/markSweep.inline.hpp"
|
||||||
|
#include "gc/serial/serialGcRefProcProxyTask.hpp"
|
||||||
|
#include "gc/serial/serialHeap.hpp"
|
||||||
|
#include "gc/shared/classUnloadingContext.hpp"
|
||||||
#include "gc/shared/collectedHeap.inline.hpp"
|
#include "gc/shared/collectedHeap.inline.hpp"
|
||||||
|
#include "gc/shared/gcHeapSummary.hpp"
|
||||||
#include "gc/shared/gcTimer.hpp"
|
#include "gc/shared/gcTimer.hpp"
|
||||||
#include "gc/shared/gcTrace.hpp"
|
#include "gc/shared/gcTrace.hpp"
|
||||||
|
#include "gc/shared/gcTraceTime.inline.hpp"
|
||||||
#include "gc/shared/gc_globals.hpp"
|
#include "gc/shared/gc_globals.hpp"
|
||||||
|
#include "gc/shared/modRefBarrierSet.hpp"
|
||||||
|
#include "gc/shared/referencePolicy.hpp"
|
||||||
|
#include "gc/shared/referenceProcessorPhaseTimes.hpp"
|
||||||
|
#include "gc/shared/space.inline.hpp"
|
||||||
|
#include "gc/shared/strongRootsScope.hpp"
|
||||||
|
#include "gc/shared/weakProcessor.hpp"
|
||||||
#include "memory/iterator.inline.hpp"
|
#include "memory/iterator.inline.hpp"
|
||||||
#include "memory/universe.hpp"
|
#include "memory/universe.hpp"
|
||||||
#include "oops/access.inline.hpp"
|
#include "oops/access.inline.hpp"
|
||||||
#include "oops/compressedOops.inline.hpp"
|
#include "oops/compressedOops.inline.hpp"
|
||||||
|
#include "oops/instanceRefKlass.hpp"
|
||||||
#include "oops/methodData.hpp"
|
#include "oops/methodData.hpp"
|
||||||
#include "oops/objArrayKlass.inline.hpp"
|
#include "oops/objArrayKlass.inline.hpp"
|
||||||
#include "oops/oop.inline.hpp"
|
#include "oops/oop.inline.hpp"
|
||||||
#include "oops/typeArrayOop.inline.hpp"
|
#include "oops/typeArrayOop.inline.hpp"
|
||||||
|
#include "runtime/prefetch.inline.hpp"
|
||||||
|
#include "utilities/copy.hpp"
|
||||||
|
#include "utilities/events.hpp"
|
||||||
#include "utilities/stack.inline.hpp"
|
#include "utilities/stack.inline.hpp"
|
||||||
|
#if INCLUDE_JVMCI
|
||||||
|
#include "jvmci/jvmci.hpp"
|
||||||
|
#endif
|
||||||
|
|
||||||
uint MarkSweep::_total_invocations = 0;
|
uint MarkSweep::_total_invocations = 0;
|
||||||
|
|
||||||
@ -62,6 +90,280 @@ MarkAndPushClosure MarkSweep::mark_and_push_closure(ClassLoaderData::_claim_stw_
|
|||||||
CLDToOopClosure MarkSweep::follow_cld_closure(&mark_and_push_closure, ClassLoaderData::_claim_stw_fullgc_mark);
|
CLDToOopClosure MarkSweep::follow_cld_closure(&mark_and_push_closure, ClassLoaderData::_claim_stw_fullgc_mark);
|
||||||
CLDToOopClosure MarkSweep::adjust_cld_closure(&adjust_pointer_closure, ClassLoaderData::_claim_stw_fullgc_adjust);
|
CLDToOopClosure MarkSweep::adjust_cld_closure(&adjust_pointer_closure, ClassLoaderData::_claim_stw_fullgc_adjust);
|
||||||
|
|
||||||
|
class DeadSpacer : StackObj {
|
||||||
|
size_t _allowed_deadspace_words;
|
||||||
|
bool _active;
|
||||||
|
ContiguousSpace* _space;
|
||||||
|
|
||||||
|
public:
|
||||||
|
DeadSpacer(ContiguousSpace* space) : _allowed_deadspace_words(0), _space(space) {
|
||||||
|
size_t ratio = _space->allowed_dead_ratio();
|
||||||
|
_active = ratio > 0;
|
||||||
|
|
||||||
|
if (_active) {
|
||||||
|
// We allow some amount of garbage towards the bottom of the space, so
|
||||||
|
// we don't start compacting before there is a significant gain to be made.
|
||||||
|
// Occasionally, we want to ensure a full compaction, which is determined
|
||||||
|
// by the MarkSweepAlwaysCompactCount parameter.
|
||||||
|
if ((MarkSweep::total_invocations() % MarkSweepAlwaysCompactCount) != 0) {
|
||||||
|
_allowed_deadspace_words = (space->capacity() * ratio / 100) / HeapWordSize;
|
||||||
|
} else {
|
||||||
|
_active = false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
bool insert_deadspace(HeapWord* dead_start, HeapWord* dead_end) {
|
||||||
|
if (!_active) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t dead_length = pointer_delta(dead_end, dead_start);
|
||||||
|
if (_allowed_deadspace_words >= dead_length) {
|
||||||
|
_allowed_deadspace_words -= dead_length;
|
||||||
|
CollectedHeap::fill_with_object(dead_start, dead_length);
|
||||||
|
oop obj = cast_to_oop(dead_start);
|
||||||
|
// obj->set_mark(obj->mark().set_marked());
|
||||||
|
|
||||||
|
assert(dead_length == obj->size(), "bad filler object size");
|
||||||
|
log_develop_trace(gc, compaction)("Inserting object to dead space: " PTR_FORMAT ", " PTR_FORMAT ", " SIZE_FORMAT "b",
|
||||||
|
p2i(dead_start), p2i(dead_end), dead_length * HeapWordSize);
|
||||||
|
|
||||||
|
return true;
|
||||||
|
} else {
|
||||||
|
_active = false;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
// Implement the "compaction" part of the mark-compact GC algorithm.
|
||||||
|
class Compacter {
|
||||||
|
// There are four spaces in total, but only the first three can be used after
|
||||||
|
// compact. IOW, old and eden/from must be enough for all live objs
|
||||||
|
static constexpr uint max_num_spaces = 4;
|
||||||
|
|
||||||
|
struct CompactionSpace {
|
||||||
|
ContiguousSpace* _space;
|
||||||
|
// Will be the new top after compaction is complete.
|
||||||
|
HeapWord* _compaction_top;
|
||||||
|
// The first dead word in this contiguous space. It's an optimization to
|
||||||
|
// skip large chunk of live objects at the beginning.
|
||||||
|
HeapWord* _first_dead;
|
||||||
|
|
||||||
|
void init(ContiguousSpace* space) {
|
||||||
|
_space = space;
|
||||||
|
_compaction_top = space->bottom();
|
||||||
|
_first_dead = nullptr;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
CompactionSpace _spaces[max_num_spaces];
|
||||||
|
// The num of spaces to be compacted, i.e. containing live objs.
|
||||||
|
uint _num_spaces;
|
||||||
|
|
||||||
|
uint _index;
|
||||||
|
|
||||||
|
HeapWord* get_compaction_top(uint index) const {
|
||||||
|
return _spaces[index]._compaction_top;
|
||||||
|
}
|
||||||
|
|
||||||
|
HeapWord* get_first_dead(uint index) const {
|
||||||
|
return _spaces[index]._first_dead;
|
||||||
|
}
|
||||||
|
|
||||||
|
ContiguousSpace* get_space(uint index) const {
|
||||||
|
return _spaces[index]._space;
|
||||||
|
}
|
||||||
|
|
||||||
|
void record_first_dead(uint index, HeapWord* first_dead) {
|
||||||
|
assert(_spaces[index]._first_dead == nullptr, "should write only once");
|
||||||
|
_spaces[index]._first_dead = first_dead;
|
||||||
|
}
|
||||||
|
|
||||||
|
HeapWord* alloc(size_t words) {
|
||||||
|
while (true) {
|
||||||
|
if (words <= pointer_delta(_spaces[_index]._space->end(),
|
||||||
|
_spaces[_index]._compaction_top)) {
|
||||||
|
HeapWord* result = _spaces[_index]._compaction_top;
|
||||||
|
_spaces[_index]._compaction_top += words;
|
||||||
|
if (_index == 0) {
|
||||||
|
// old-gen requires BOT update
|
||||||
|
static_cast<TenuredSpace*>(_spaces[0]._space)->update_for_block(result, result + words);
|
||||||
|
}
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
// out-of-memory in this space
|
||||||
|
_index++;
|
||||||
|
assert(_index < max_num_spaces - 1, "the last space should not be used");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static void prefetch_read_scan(void* p) {
|
||||||
|
if (PrefetchScanIntervalInBytes >= 0) {
|
||||||
|
Prefetch::read(p, PrefetchScanIntervalInBytes);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static void prefetch_write_scan(void* p) {
|
||||||
|
if (PrefetchScanIntervalInBytes >= 0) {
|
||||||
|
Prefetch::write(p, PrefetchScanIntervalInBytes);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static void prefetch_write_copy(void* p) {
|
||||||
|
if (PrefetchCopyIntervalInBytes >= 0) {
|
||||||
|
Prefetch::write(p, PrefetchCopyIntervalInBytes);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static void forward_obj(oop obj, HeapWord* new_addr) {
|
||||||
|
prefetch_write_scan(obj);
|
||||||
|
if (cast_from_oop<HeapWord*>(obj) != new_addr) {
|
||||||
|
obj->forward_to(cast_to_oop(new_addr));
|
||||||
|
} else {
|
||||||
|
assert(obj->is_gc_marked(), "inv");
|
||||||
|
// This obj will stay in-place. Fix the markword.
|
||||||
|
obj->init_mark();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static HeapWord* find_next_live_addr(HeapWord* start, HeapWord* end) {
|
||||||
|
for (HeapWord* i_addr = start; i_addr < end; /* empty */) {
|
||||||
|
prefetch_read_scan(i_addr);
|
||||||
|
oop obj = cast_to_oop(i_addr);
|
||||||
|
if (obj->is_gc_marked()) {
|
||||||
|
return i_addr;
|
||||||
|
}
|
||||||
|
i_addr += obj->size();
|
||||||
|
}
|
||||||
|
return end;
|
||||||
|
};
|
||||||
|
|
||||||
|
static size_t relocate(HeapWord* addr) {
|
||||||
|
// Prefetch source and destination
|
||||||
|
prefetch_read_scan(addr);
|
||||||
|
|
||||||
|
oop obj = cast_to_oop(addr);
|
||||||
|
oop new_obj = obj->forwardee();
|
||||||
|
HeapWord* new_addr = cast_from_oop<HeapWord*>(new_obj);
|
||||||
|
assert(addr != new_addr, "inv");
|
||||||
|
prefetch_write_copy(new_addr);
|
||||||
|
|
||||||
|
size_t obj_size = obj->size();
|
||||||
|
Copy::aligned_conjoint_words(addr, new_addr, obj_size);
|
||||||
|
new_obj->init_mark();
|
||||||
|
|
||||||
|
return obj_size;
|
||||||
|
}
|
||||||
|
|
||||||
|
public:
|
||||||
|
explicit Compacter(SerialHeap* heap) {
|
||||||
|
// In this order so that heap is compacted towards old-gen.
|
||||||
|
_spaces[0].init(heap->old_gen()->space());
|
||||||
|
_spaces[1].init(heap->young_gen()->eden());
|
||||||
|
_spaces[2].init(heap->young_gen()->from());
|
||||||
|
|
||||||
|
bool is_promotion_failed = (heap->young_gen()->from()->next_compaction_space() != nullptr);
|
||||||
|
if (is_promotion_failed) {
|
||||||
|
_spaces[3].init(heap->young_gen()->to());
|
||||||
|
_num_spaces = 4;
|
||||||
|
} else {
|
||||||
|
_num_spaces = 3;
|
||||||
|
}
|
||||||
|
_index = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
void phase2_calculate_new_addr() {
|
||||||
|
for (uint i = 0; i < _num_spaces; ++i) {
|
||||||
|
ContiguousSpace* space = get_space(i);
|
||||||
|
HeapWord* cur_addr = space->bottom();
|
||||||
|
HeapWord* top = space->top();
|
||||||
|
|
||||||
|
bool record_first_dead_done = false;
|
||||||
|
|
||||||
|
DeadSpacer dead_spacer(space);
|
||||||
|
|
||||||
|
while (cur_addr < top) {
|
||||||
|
oop obj = cast_to_oop(cur_addr);
|
||||||
|
size_t obj_size = obj->size();
|
||||||
|
if (obj->is_gc_marked()) {
|
||||||
|
HeapWord* new_addr = alloc(obj_size);
|
||||||
|
forward_obj(obj, new_addr);
|
||||||
|
cur_addr += obj_size;
|
||||||
|
} else {
|
||||||
|
// Skipping the current known-unmarked obj
|
||||||
|
HeapWord* next_live_addr = find_next_live_addr(cur_addr + obj_size, top);
|
||||||
|
if (dead_spacer.insert_deadspace(cur_addr, next_live_addr)) {
|
||||||
|
// Register space for the filler obj
|
||||||
|
alloc(pointer_delta(next_live_addr, cur_addr));
|
||||||
|
} else {
|
||||||
|
if (!record_first_dead_done) {
|
||||||
|
record_first_dead(i, cur_addr);
|
||||||
|
record_first_dead_done = true;
|
||||||
|
}
|
||||||
|
*(HeapWord**)cur_addr = next_live_addr;
|
||||||
|
}
|
||||||
|
cur_addr = next_live_addr;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!record_first_dead_done) {
|
||||||
|
record_first_dead(i, top);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void phase3_adjust_pointers() {
|
||||||
|
for (uint i = 0; i < _num_spaces; ++i) {
|
||||||
|
ContiguousSpace* space = get_space(i);
|
||||||
|
HeapWord* cur_addr = space->bottom();
|
||||||
|
HeapWord* const top = space->top();
|
||||||
|
HeapWord* const first_dead = get_first_dead(i);
|
||||||
|
|
||||||
|
while (cur_addr < top) {
|
||||||
|
prefetch_write_scan(cur_addr);
|
||||||
|
if (cur_addr < first_dead || cast_to_oop(cur_addr)->is_gc_marked()) {
|
||||||
|
size_t size = MarkSweep::adjust_pointers(cast_to_oop(cur_addr));
|
||||||
|
cur_addr += size;
|
||||||
|
} else {
|
||||||
|
assert(*(HeapWord**)cur_addr > cur_addr, "forward progress");
|
||||||
|
cur_addr = *(HeapWord**)cur_addr;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void phase4_compact() {
|
||||||
|
for (uint i = 0; i < _num_spaces; ++i) {
|
||||||
|
ContiguousSpace* space = get_space(i);
|
||||||
|
HeapWord* cur_addr = space->bottom();
|
||||||
|
HeapWord* top = space->top();
|
||||||
|
|
||||||
|
// Check if the first obj inside this space is forwarded.
|
||||||
|
if (!cast_to_oop(cur_addr)->is_forwarded()) {
|
||||||
|
// Jump over consecutive (in-place) live-objs-chunk
|
||||||
|
cur_addr = get_first_dead(i);
|
||||||
|
}
|
||||||
|
|
||||||
|
while (cur_addr < top) {
|
||||||
|
if (!cast_to_oop(cur_addr)->is_forwarded()) {
|
||||||
|
cur_addr = *(HeapWord**) cur_addr;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
cur_addr += relocate(cur_addr);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Reset top and unused memory
|
||||||
|
space->set_top(get_compaction_top(i));
|
||||||
|
if (ZapUnusedHeapArea) {
|
||||||
|
space->mangle_unused_area();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
template <class T> void MarkSweep::KeepAliveClosure::do_oop_work(T* p) {
|
template <class T> void MarkSweep::KeepAliveClosure::do_oop_work(T* p) {
|
||||||
mark_and_push(p);
|
mark_and_push(p);
|
||||||
}
|
}
|
||||||
@ -157,6 +459,119 @@ void MarkSweep::preserve_mark(oop obj, markWord mark) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void MarkSweep::phase1_mark(bool clear_all_softrefs) {
|
||||||
|
// Recursively traverse all live objects and mark them
|
||||||
|
GCTraceTime(Info, gc, phases) tm("Phase 1: Mark live objects", _gc_timer);
|
||||||
|
|
||||||
|
SerialHeap* gch = SerialHeap::heap();
|
||||||
|
|
||||||
|
ClassLoaderDataGraph::verify_claimed_marks_cleared(ClassLoaderData::_claim_stw_fullgc_mark);
|
||||||
|
|
||||||
|
ref_processor()->start_discovery(clear_all_softrefs);
|
||||||
|
|
||||||
|
{
|
||||||
|
StrongRootsScope srs(0);
|
||||||
|
|
||||||
|
CLDClosure* weak_cld_closure = ClassUnloading ? nullptr : &follow_cld_closure;
|
||||||
|
MarkingCodeBlobClosure mark_code_closure(&follow_root_closure, !CodeBlobToOopClosure::FixRelocations, true);
|
||||||
|
gch->process_roots(SerialHeap::SO_None,
|
||||||
|
&follow_root_closure,
|
||||||
|
&follow_cld_closure,
|
||||||
|
weak_cld_closure,
|
||||||
|
&mark_code_closure);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process reference objects found during marking
|
||||||
|
{
|
||||||
|
GCTraceTime(Debug, gc, phases) tm_m("Reference Processing", gc_timer());
|
||||||
|
|
||||||
|
ReferenceProcessorPhaseTimes pt(_gc_timer, ref_processor()->max_num_queues());
|
||||||
|
SerialGCRefProcProxyTask task(is_alive, keep_alive, follow_stack_closure);
|
||||||
|
const ReferenceProcessorStats& stats = ref_processor()->process_discovered_references(task, pt);
|
||||||
|
pt.print_all_references();
|
||||||
|
gc_tracer()->report_gc_reference_stats(stats);
|
||||||
|
}
|
||||||
|
|
||||||
|
// This is the point where the entire marking should have completed.
|
||||||
|
assert(_marking_stack.is_empty(), "Marking should have completed");
|
||||||
|
|
||||||
|
{
|
||||||
|
GCTraceTime(Debug, gc, phases) tm_m("Weak Processing", gc_timer());
|
||||||
|
WeakProcessor::weak_oops_do(&is_alive, &do_nothing_cl);
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
GCTraceTime(Debug, gc, phases) tm_m("Class Unloading", gc_timer());
|
||||||
|
|
||||||
|
ClassUnloadingContext* ctx = ClassUnloadingContext::context();
|
||||||
|
|
||||||
|
bool unloading_occurred;
|
||||||
|
{
|
||||||
|
CodeCache::UnlinkingScope scope(&is_alive);
|
||||||
|
|
||||||
|
// Unload classes and purge the SystemDictionary.
|
||||||
|
unloading_occurred = SystemDictionary::do_unloading(gc_timer());
|
||||||
|
|
||||||
|
// Unload nmethods.
|
||||||
|
CodeCache::do_unloading(unloading_occurred);
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
GCTraceTime(Debug, gc, phases) t("Purge Unlinked NMethods", gc_timer());
|
||||||
|
// Release unloaded nmethod's memory.
|
||||||
|
ctx->purge_nmethods();
|
||||||
|
}
|
||||||
|
{
|
||||||
|
GCTraceTime(Debug, gc, phases) ur("Unregister NMethods", gc_timer());
|
||||||
|
gch->prune_unlinked_nmethods();
|
||||||
|
}
|
||||||
|
{
|
||||||
|
GCTraceTime(Debug, gc, phases) t("Free Code Blobs", gc_timer());
|
||||||
|
ctx->free_code_blobs();
|
||||||
|
}
|
||||||
|
|
||||||
|
// Prune dead klasses from subklass/sibling/implementor lists.
|
||||||
|
Klass::clean_weak_klass_links(unloading_occurred);
|
||||||
|
|
||||||
|
// Clean JVMCI metadata handles.
|
||||||
|
JVMCI_ONLY(JVMCI::do_unloading(unloading_occurred));
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
GCTraceTime(Debug, gc, phases) tm_m("Report Object Count", gc_timer());
|
||||||
|
gc_tracer()->report_object_count_after_gc(&is_alive, nullptr);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void MarkSweep::allocate_stacks() {
|
||||||
|
void* scratch = nullptr;
|
||||||
|
size_t num_words;
|
||||||
|
DefNewGeneration* young_gen = (DefNewGeneration*)SerialHeap::heap()->young_gen();
|
||||||
|
young_gen->contribute_scratch(scratch, num_words);
|
||||||
|
|
||||||
|
if (scratch != nullptr) {
|
||||||
|
_preserved_count_max = num_words * HeapWordSize / sizeof(PreservedMark);
|
||||||
|
} else {
|
||||||
|
_preserved_count_max = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
_preserved_marks = (PreservedMark*)scratch;
|
||||||
|
_preserved_count = 0;
|
||||||
|
|
||||||
|
_preserved_overflow_stack_set.init(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
void MarkSweep::deallocate_stacks() {
|
||||||
|
if (_preserved_count_max != 0) {
|
||||||
|
DefNewGeneration* young_gen = (DefNewGeneration*)SerialHeap::heap()->young_gen();
|
||||||
|
young_gen->reset_scratch();
|
||||||
|
}
|
||||||
|
|
||||||
|
_preserved_overflow_stack_set.reclaim();
|
||||||
|
_marking_stack.clear();
|
||||||
|
_objarray_stack.clear(true);
|
||||||
|
}
|
||||||
|
|
||||||
void MarkSweep::mark_object(oop obj) {
|
void MarkSweep::mark_object(oop obj) {
|
||||||
if (StringDedup::is_enabled() &&
|
if (StringDedup::is_enabled() &&
|
||||||
java_lang_String::is_instance(obj) &&
|
java_lang_String::is_instance(obj) &&
|
||||||
@ -235,3 +650,92 @@ void MarkSweep::initialize() {
|
|||||||
MarkSweep::_ref_processor = new ReferenceProcessor(&_always_true_closure);
|
MarkSweep::_ref_processor = new ReferenceProcessor(&_always_true_closure);
|
||||||
mark_and_push_closure.set_ref_discoverer(_ref_processor);
|
mark_and_push_closure.set_ref_discoverer(_ref_processor);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void MarkSweep::invoke_at_safepoint(bool clear_all_softrefs) {
|
||||||
|
assert(SafepointSynchronize::is_at_safepoint(), "must be at a safepoint");
|
||||||
|
|
||||||
|
SerialHeap* gch = SerialHeap::heap();
|
||||||
|
#ifdef ASSERT
|
||||||
|
if (gch->soft_ref_policy()->should_clear_all_soft_refs()) {
|
||||||
|
assert(clear_all_softrefs, "Policy should have been checked earlier");
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
gch->trace_heap_before_gc(_gc_tracer);
|
||||||
|
|
||||||
|
// Increment the invocation count
|
||||||
|
_total_invocations++;
|
||||||
|
|
||||||
|
// Capture used regions for old-gen to reestablish old-to-young invariant
|
||||||
|
// after full-gc.
|
||||||
|
gch->old_gen()->save_used_region();
|
||||||
|
|
||||||
|
allocate_stacks();
|
||||||
|
|
||||||
|
phase1_mark(clear_all_softrefs);
|
||||||
|
|
||||||
|
Compacter compacter{gch};
|
||||||
|
|
||||||
|
{
|
||||||
|
// Now all live objects are marked, compute the new object addresses.
|
||||||
|
GCTraceTime(Info, gc, phases) tm("Phase 2: Compute new object addresses", _gc_timer);
|
||||||
|
|
||||||
|
compacter.phase2_calculate_new_addr();
|
||||||
|
}
|
||||||
|
|
||||||
|
// Don't add any more derived pointers during phase3
|
||||||
|
#if COMPILER2_OR_JVMCI
|
||||||
|
assert(DerivedPointerTable::is_active(), "Sanity");
|
||||||
|
DerivedPointerTable::set_active(false);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
{
|
||||||
|
// Adjust the pointers to reflect the new locations
|
||||||
|
GCTraceTime(Info, gc, phases) tm("Phase 3: Adjust pointers", gc_timer());
|
||||||
|
|
||||||
|
ClassLoaderDataGraph::verify_claimed_marks_cleared(ClassLoaderData::_claim_stw_fullgc_adjust);
|
||||||
|
|
||||||
|
CodeBlobToOopClosure code_closure(&adjust_pointer_closure, CodeBlobToOopClosure::FixRelocations);
|
||||||
|
gch->process_roots(SerialHeap::SO_AllCodeCache,
|
||||||
|
&adjust_pointer_closure,
|
||||||
|
&adjust_cld_closure,
|
||||||
|
&adjust_cld_closure,
|
||||||
|
&code_closure);
|
||||||
|
|
||||||
|
WeakProcessor::oops_do(&adjust_pointer_closure);
|
||||||
|
|
||||||
|
adjust_marks();
|
||||||
|
compacter.phase3_adjust_pointers();
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
// All pointers are now adjusted, move objects accordingly
|
||||||
|
GCTraceTime(Info, gc, phases) tm("Phase 4: Move objects", _gc_timer);
|
||||||
|
|
||||||
|
compacter.phase4_compact();
|
||||||
|
}
|
||||||
|
|
||||||
|
restore_marks();
|
||||||
|
|
||||||
|
// Set saved marks for allocation profiler (and other things? -- dld)
|
||||||
|
// (Should this be in general part?)
|
||||||
|
gch->save_marks();
|
||||||
|
|
||||||
|
deallocate_stacks();
|
||||||
|
|
||||||
|
MarkSweep::_string_dedup_requests->flush();
|
||||||
|
|
||||||
|
bool is_young_gen_empty = (gch->young_gen()->used() == 0);
|
||||||
|
gch->rem_set()->maintain_old_to_young_invariant(gch->old_gen(), is_young_gen_empty);
|
||||||
|
|
||||||
|
gch->prune_scavengable_nmethods();
|
||||||
|
|
||||||
|
// Update heap occupancy information which is used as
|
||||||
|
// input to soft ref clearing policy at the next gc.
|
||||||
|
Universe::heap()->update_capacity_and_used_at_gc();
|
||||||
|
|
||||||
|
// Signal that we have completed a visit to all live objects.
|
||||||
|
Universe::heap()->record_whole_heap_examined_timestamp();
|
||||||
|
|
||||||
|
gch->trace_heap_after_gc(_gc_tracer);
|
||||||
|
}
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
/*
|
/*
|
||||||
* Copyright (c) 1997, 2022, Oracle and/or its affiliates. All rights reserved.
|
* Copyright (c) 1997, 2024, Oracle and/or its affiliates. All rights reserved.
|
||||||
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
||||||
*
|
*
|
||||||
* This code is free software; you can redistribute it and/or modify it
|
* This code is free software; you can redistribute it and/or modify it
|
||||||
@ -128,6 +128,8 @@ class MarkSweep : AllStatic {
|
|||||||
static AdjustPointerClosure adjust_pointer_closure;
|
static AdjustPointerClosure adjust_pointer_closure;
|
||||||
static CLDToOopClosure adjust_cld_closure;
|
static CLDToOopClosure adjust_cld_closure;
|
||||||
|
|
||||||
|
static void invoke_at_safepoint(bool clear_all_softrefs);
|
||||||
|
|
||||||
// Accessors
|
// Accessors
|
||||||
static uint total_invocations() { return _total_invocations; }
|
static uint total_invocations() { return _total_invocations; }
|
||||||
|
|
||||||
@ -152,6 +154,13 @@ class MarkSweep : AllStatic {
|
|||||||
template <class T> static void mark_and_push(T* p);
|
template <class T> static void mark_and_push(T* p);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
|
// Mark live objects
|
||||||
|
static void phase1_mark(bool clear_all_softrefs);
|
||||||
|
|
||||||
|
// Temporary data structures for traversal and storing/restoring marks
|
||||||
|
static void allocate_stacks();
|
||||||
|
static void deallocate_stacks();
|
||||||
|
|
||||||
// Call backs for marking
|
// Call backs for marking
|
||||||
static void mark_object(oop obj);
|
static void mark_object(oop obj);
|
||||||
// Mark pointer and follow contents. Empty marking stack afterwards.
|
// Mark pointer and follow contents. Empty marking stack afterwards.
|
||||||
|
@ -31,7 +31,6 @@
|
|||||||
#include "compiler/oopMap.hpp"
|
#include "compiler/oopMap.hpp"
|
||||||
#include "gc/serial/cardTableRS.hpp"
|
#include "gc/serial/cardTableRS.hpp"
|
||||||
#include "gc/serial/defNewGeneration.inline.hpp"
|
#include "gc/serial/defNewGeneration.inline.hpp"
|
||||||
#include "gc/serial/genMarkSweep.hpp"
|
|
||||||
#include "gc/serial/markSweep.hpp"
|
#include "gc/serial/markSweep.hpp"
|
||||||
#include "gc/serial/serialHeap.hpp"
|
#include "gc/serial/serialHeap.hpp"
|
||||||
#include "gc/serial/serialMemoryPools.hpp"
|
#include "gc/serial/serialMemoryPools.hpp"
|
||||||
@ -561,7 +560,7 @@ void SerialHeap::do_collection(bool full,
|
|||||||
|
|
||||||
if (do_full_collection) {
|
if (do_full_collection) {
|
||||||
GCIdMark gc_id_mark;
|
GCIdMark gc_id_mark;
|
||||||
GCTraceCPUTime tcpu(GenMarkSweep::gc_tracer());
|
GCTraceCPUTime tcpu(MarkSweep::gc_tracer());
|
||||||
GCTraceTime(Info, gc) t("Pause Full", nullptr, gc_cause(), true);
|
GCTraceTime(Info, gc) t("Pause Full", nullptr, gc_cause(), true);
|
||||||
|
|
||||||
print_heap_before_gc();
|
print_heap_before_gc();
|
||||||
|
@ -66,7 +66,7 @@ class SerialHeap : public CollectedHeap {
|
|||||||
friend class Generation;
|
friend class Generation;
|
||||||
friend class DefNewGeneration;
|
friend class DefNewGeneration;
|
||||||
friend class TenuredGeneration;
|
friend class TenuredGeneration;
|
||||||
friend class GenMarkSweep;
|
friend class MarkSweep;
|
||||||
friend class VM_GenCollectForAllocation;
|
friend class VM_GenCollectForAllocation;
|
||||||
friend class VM_GenCollectFull;
|
friend class VM_GenCollectFull;
|
||||||
friend class VM_GC_HeapInspection;
|
friend class VM_GC_HeapInspection;
|
||||||
|
@ -24,7 +24,7 @@
|
|||||||
|
|
||||||
#include "precompiled.hpp"
|
#include "precompiled.hpp"
|
||||||
#include "gc/serial/cardTableRS.hpp"
|
#include "gc/serial/cardTableRS.hpp"
|
||||||
#include "gc/serial/genMarkSweep.hpp"
|
#include "gc/serial/markSweep.hpp"
|
||||||
#include "gc/serial/serialBlockOffsetTable.inline.hpp"
|
#include "gc/serial/serialBlockOffsetTable.inline.hpp"
|
||||||
#include "gc/serial/serialHeap.hpp"
|
#include "gc/serial/serialHeap.hpp"
|
||||||
#include "gc/serial/tenuredGeneration.inline.hpp"
|
#include "gc/serial/tenuredGeneration.inline.hpp"
|
||||||
@ -444,15 +444,15 @@ void TenuredGeneration::collect(bool full,
|
|||||||
bool is_tlab) {
|
bool is_tlab) {
|
||||||
SerialHeap* gch = SerialHeap::heap();
|
SerialHeap* gch = SerialHeap::heap();
|
||||||
|
|
||||||
STWGCTimer* gc_timer = GenMarkSweep::gc_timer();
|
STWGCTimer* gc_timer = MarkSweep::gc_timer();
|
||||||
gc_timer->register_gc_start();
|
gc_timer->register_gc_start();
|
||||||
|
|
||||||
SerialOldTracer* gc_tracer = GenMarkSweep::gc_tracer();
|
SerialOldTracer* gc_tracer = MarkSweep::gc_tracer();
|
||||||
gc_tracer->report_gc_start(gch->gc_cause(), gc_timer->gc_start());
|
gc_tracer->report_gc_start(gch->gc_cause(), gc_timer->gc_start());
|
||||||
|
|
||||||
gch->pre_full_gc_dump(gc_timer);
|
gch->pre_full_gc_dump(gc_timer);
|
||||||
|
|
||||||
GenMarkSweep::invoke_at_safepoint(clear_all_soft_refs);
|
MarkSweep::invoke_at_safepoint(clear_all_soft_refs);
|
||||||
|
|
||||||
gch->post_full_gc_dump(gc_timer);
|
gch->post_full_gc_dump(gc_timer);
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user