mirror of
https://github.com/openjdk/jdk.git
synced 2025-09-17 01:24:33 +02:00
8079792: GC directory structure cleanup
Reviewed-by: brutisso, stefank, david
This commit is contained in:
parent
042902811c
commit
4dc240f785
521 changed files with 2481 additions and 2573 deletions
335
hotspot/src/share/vm/gc/parallel/psCompactionManager.cpp
Normal file
335
hotspot/src/share/vm/gc/parallel/psCompactionManager.cpp
Normal file
|
@ -0,0 +1,335 @@
|
|||
/*
|
||||
* Copyright (c) 2005, 2015, Oracle and/or its affiliates. All rights reserved.
|
||||
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
||||
*
|
||||
* This code is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 only, as
|
||||
* published by the Free Software Foundation.
|
||||
*
|
||||
* This code is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
||||
* version 2 for more details (a copy is included in the LICENSE file that
|
||||
* accompanied this code).
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License version
|
||||
* 2 along with this work; if not, write to the Free Software Foundation,
|
||||
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
||||
*
|
||||
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
||||
* or visit www.oracle.com if you need additional information or have any
|
||||
* questions.
|
||||
*
|
||||
*/
|
||||
|
||||
#include "precompiled.hpp"
|
||||
#include "classfile/systemDictionary.hpp"
|
||||
#include "gc/parallel/gcTaskManager.hpp"
|
||||
#include "gc/parallel/objectStartArray.hpp"
|
||||
#include "gc/parallel/parMarkBitMap.hpp"
|
||||
#include "gc/parallel/parallelScavengeHeap.hpp"
|
||||
#include "gc/parallel/psCompactionManager.inline.hpp"
|
||||
#include "gc/parallel/psOldGen.hpp"
|
||||
#include "gc/parallel/psParallelCompact.inline.hpp"
|
||||
#include "gc/shared/taskqueue.inline.hpp"
|
||||
#include "memory/iterator.inline.hpp"
|
||||
#include "oops/instanceKlass.inline.hpp"
|
||||
#include "oops/instanceMirrorKlass.inline.hpp"
|
||||
#include "oops/objArrayKlass.inline.hpp"
|
||||
#include "oops/oop.inline.hpp"
|
||||
#include "runtime/atomic.inline.hpp"
|
||||
|
||||
PSOldGen* ParCompactionManager::_old_gen = NULL;
|
||||
ParCompactionManager** ParCompactionManager::_manager_array = NULL;
|
||||
|
||||
RegionTaskQueue** ParCompactionManager::_region_list = NULL;
|
||||
|
||||
OopTaskQueueSet* ParCompactionManager::_stack_array = NULL;
|
||||
ParCompactionManager::ObjArrayTaskQueueSet*
|
||||
ParCompactionManager::_objarray_queues = NULL;
|
||||
ObjectStartArray* ParCompactionManager::_start_array = NULL;
|
||||
ParMarkBitMap* ParCompactionManager::_mark_bitmap = NULL;
|
||||
RegionTaskQueueSet* ParCompactionManager::_region_array = NULL;
|
||||
|
||||
uint* ParCompactionManager::_recycled_stack_index = NULL;
|
||||
int ParCompactionManager::_recycled_top = -1;
|
||||
int ParCompactionManager::_recycled_bottom = -1;
|
||||
|
||||
ParCompactionManager::ParCompactionManager() :
|
||||
_action(CopyAndUpdate),
|
||||
_region_stack(NULL),
|
||||
_region_stack_index((uint)max_uintx) {
|
||||
|
||||
ParallelScavengeHeap* heap = ParallelScavengeHeap::heap();
|
||||
|
||||
_old_gen = heap->old_gen();
|
||||
_start_array = old_gen()->start_array();
|
||||
|
||||
marking_stack()->initialize();
|
||||
_objarray_stack.initialize();
|
||||
}
|
||||
|
||||
ParCompactionManager::~ParCompactionManager() {
|
||||
delete _recycled_stack_index;
|
||||
}
|
||||
|
||||
void ParCompactionManager::initialize(ParMarkBitMap* mbm) {
|
||||
assert(PSParallelCompact::gc_task_manager() != NULL,
|
||||
"Needed for initialization");
|
||||
|
||||
_mark_bitmap = mbm;
|
||||
|
||||
uint parallel_gc_threads = PSParallelCompact::gc_task_manager()->workers();
|
||||
|
||||
assert(_manager_array == NULL, "Attempt to initialize twice");
|
||||
_manager_array = NEW_C_HEAP_ARRAY(ParCompactionManager*, parallel_gc_threads+1, mtGC);
|
||||
guarantee(_manager_array != NULL, "Could not allocate manager_array");
|
||||
|
||||
_region_list = NEW_C_HEAP_ARRAY(RegionTaskQueue*,
|
||||
parallel_gc_threads+1, mtGC);
|
||||
guarantee(_region_list != NULL, "Could not initialize promotion manager");
|
||||
|
||||
_recycled_stack_index = NEW_C_HEAP_ARRAY(uint, parallel_gc_threads, mtGC);
|
||||
|
||||
// parallel_gc-threads + 1 to be consistent with the number of
|
||||
// compaction managers.
|
||||
for(uint i=0; i<parallel_gc_threads + 1; i++) {
|
||||
_region_list[i] = new RegionTaskQueue();
|
||||
region_list(i)->initialize();
|
||||
}
|
||||
|
||||
_stack_array = new OopTaskQueueSet(parallel_gc_threads);
|
||||
guarantee(_stack_array != NULL, "Could not allocate stack_array");
|
||||
_objarray_queues = new ObjArrayTaskQueueSet(parallel_gc_threads);
|
||||
guarantee(_objarray_queues != NULL, "Could not allocate objarray_queues");
|
||||
_region_array = new RegionTaskQueueSet(parallel_gc_threads);
|
||||
guarantee(_region_array != NULL, "Could not allocate region_array");
|
||||
|
||||
// Create and register the ParCompactionManager(s) for the worker threads.
|
||||
for(uint i=0; i<parallel_gc_threads; i++) {
|
||||
_manager_array[i] = new ParCompactionManager();
|
||||
guarantee(_manager_array[i] != NULL, "Could not create ParCompactionManager");
|
||||
stack_array()->register_queue(i, _manager_array[i]->marking_stack());
|
||||
_objarray_queues->register_queue(i, &_manager_array[i]->_objarray_stack);
|
||||
region_array()->register_queue(i, region_list(i));
|
||||
}
|
||||
|
||||
// The VMThread gets its own ParCompactionManager, which is not available
|
||||
// for work stealing.
|
||||
_manager_array[parallel_gc_threads] = new ParCompactionManager();
|
||||
guarantee(_manager_array[parallel_gc_threads] != NULL,
|
||||
"Could not create ParCompactionManager");
|
||||
assert(PSParallelCompact::gc_task_manager()->workers() != 0,
|
||||
"Not initialized?");
|
||||
}
|
||||
|
||||
int ParCompactionManager::pop_recycled_stack_index() {
|
||||
assert(_recycled_bottom <= _recycled_top, "list is empty");
|
||||
// Get the next available index
|
||||
if (_recycled_bottom < _recycled_top) {
|
||||
uint cur, next, last;
|
||||
do {
|
||||
cur = _recycled_bottom;
|
||||
next = cur + 1;
|
||||
last = Atomic::cmpxchg(next, &_recycled_bottom, cur);
|
||||
} while (cur != last);
|
||||
return _recycled_stack_index[next];
|
||||
} else {
|
||||
return -1;
|
||||
}
|
||||
}
|
||||
|
||||
void ParCompactionManager::push_recycled_stack_index(uint v) {
|
||||
// Get the next available index
|
||||
int cur = Atomic::add(1, &_recycled_top);
|
||||
_recycled_stack_index[cur] = v;
|
||||
assert(_recycled_bottom <= _recycled_top, "list top and bottom are wrong");
|
||||
}
|
||||
|
||||
bool ParCompactionManager::should_update() {
|
||||
assert(action() != NotValid, "Action is not set");
|
||||
return (action() == ParCompactionManager::Update) ||
|
||||
(action() == ParCompactionManager::CopyAndUpdate) ||
|
||||
(action() == ParCompactionManager::UpdateAndCopy);
|
||||
}
|
||||
|
||||
bool ParCompactionManager::should_copy() {
|
||||
assert(action() != NotValid, "Action is not set");
|
||||
return (action() == ParCompactionManager::Copy) ||
|
||||
(action() == ParCompactionManager::CopyAndUpdate) ||
|
||||
(action() == ParCompactionManager::UpdateAndCopy);
|
||||
}
|
||||
|
||||
void ParCompactionManager::region_list_push(uint list_index,
|
||||
size_t region_index) {
|
||||
region_list(list_index)->push(region_index);
|
||||
}
|
||||
|
||||
void ParCompactionManager::verify_region_list_empty(uint list_index) {
|
||||
assert(region_list(list_index)->is_empty(), "Not empty");
|
||||
}
|
||||
|
||||
ParCompactionManager*
|
||||
ParCompactionManager::gc_thread_compaction_manager(int index) {
|
||||
assert(index >= 0 && index < (int)ParallelGCThreads, "index out of range");
|
||||
assert(_manager_array != NULL, "Sanity");
|
||||
return _manager_array[index];
|
||||
}
|
||||
|
||||
void InstanceKlass::oop_pc_follow_contents(oop obj, ParCompactionManager* cm) {
|
||||
assert(obj != NULL, "can't follow the content of NULL object");
|
||||
|
||||
cm->follow_klass(this);
|
||||
// Only mark the header and let the scan of the meta-data mark
|
||||
// everything else.
|
||||
|
||||
ParCompactionManager::MarkAndPushClosure cl(cm);
|
||||
InstanceKlass::oop_oop_iterate_oop_maps<true>(obj, &cl);
|
||||
}
|
||||
|
||||
void InstanceMirrorKlass::oop_pc_follow_contents(oop obj, ParCompactionManager* cm) {
|
||||
InstanceKlass::oop_pc_follow_contents(obj, cm);
|
||||
|
||||
// Follow the klass field in the mirror.
|
||||
Klass* klass = java_lang_Class::as_Klass(obj);
|
||||
if (klass != NULL) {
|
||||
// An anonymous class doesn't have its own class loader, so the call
|
||||
// to follow_klass will mark and push its java mirror instead of the
|
||||
// class loader. When handling the java mirror for an anonymous class
|
||||
// we need to make sure its class loader data is claimed, this is done
|
||||
// by calling follow_class_loader explicitly. For non-anonymous classes
|
||||
// the call to follow_class_loader is made when the class loader itself
|
||||
// is handled.
|
||||
if (klass->oop_is_instance() && InstanceKlass::cast(klass)->is_anonymous()) {
|
||||
cm->follow_class_loader(klass->class_loader_data());
|
||||
} else {
|
||||
cm->follow_klass(klass);
|
||||
}
|
||||
} else {
|
||||
// If klass is NULL then this a mirror for a primitive type.
|
||||
// We don't have to follow them, since they are handled as strong
|
||||
// roots in Universe::oops_do.
|
||||
assert(java_lang_Class::is_primitive(obj), "Sanity check");
|
||||
}
|
||||
|
||||
ParCompactionManager::MarkAndPushClosure cl(cm);
|
||||
oop_oop_iterate_statics<true>(obj, &cl);
|
||||
}
|
||||
|
||||
void InstanceClassLoaderKlass::oop_pc_follow_contents(oop obj, ParCompactionManager* cm) {
|
||||
InstanceKlass::oop_pc_follow_contents(obj, cm);
|
||||
|
||||
ClassLoaderData * const loader_data = java_lang_ClassLoader::loader_data(obj);
|
||||
if (loader_data != NULL) {
|
||||
cm->follow_class_loader(loader_data);
|
||||
}
|
||||
}
|
||||
|
||||
template <class T>
|
||||
static void oop_pc_follow_contents_specialized(InstanceRefKlass* klass, oop obj, ParCompactionManager* cm) {
|
||||
T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
|
||||
T heap_oop = oopDesc::load_heap_oop(referent_addr);
|
||||
debug_only(
|
||||
if(TraceReferenceGC && PrintGCDetails) {
|
||||
gclog_or_tty->print_cr("InstanceRefKlass::oop_pc_follow_contents " PTR_FORMAT, p2i(obj));
|
||||
}
|
||||
)
|
||||
if (!oopDesc::is_null(heap_oop)) {
|
||||
oop referent = oopDesc::decode_heap_oop_not_null(heap_oop);
|
||||
if (PSParallelCompact::mark_bitmap()->is_unmarked(referent) &&
|
||||
PSParallelCompact::ref_processor()->discover_reference(obj, klass->reference_type())) {
|
||||
// reference already enqueued, referent will be traversed later
|
||||
klass->InstanceKlass::oop_pc_follow_contents(obj, cm);
|
||||
debug_only(
|
||||
if(TraceReferenceGC && PrintGCDetails) {
|
||||
gclog_or_tty->print_cr(" Non NULL enqueued " PTR_FORMAT, p2i(obj));
|
||||
}
|
||||
)
|
||||
return;
|
||||
} else {
|
||||
// treat referent as normal oop
|
||||
debug_only(
|
||||
if(TraceReferenceGC && PrintGCDetails) {
|
||||
gclog_or_tty->print_cr(" Non NULL normal " PTR_FORMAT, p2i(obj));
|
||||
}
|
||||
)
|
||||
cm->mark_and_push(referent_addr);
|
||||
}
|
||||
}
|
||||
T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
|
||||
// Treat discovered as normal oop, if ref is not "active",
|
||||
// i.e. if next is non-NULL.
|
||||
T next_oop = oopDesc::load_heap_oop(next_addr);
|
||||
if (!oopDesc::is_null(next_oop)) { // i.e. ref is not "active"
|
||||
T* discovered_addr = (T*)java_lang_ref_Reference::discovered_addr(obj);
|
||||
debug_only(
|
||||
if(TraceReferenceGC && PrintGCDetails) {
|
||||
gclog_or_tty->print_cr(" Process discovered as normal "
|
||||
PTR_FORMAT, p2i(discovered_addr));
|
||||
}
|
||||
)
|
||||
cm->mark_and_push(discovered_addr);
|
||||
}
|
||||
cm->mark_and_push(next_addr);
|
||||
klass->InstanceKlass::oop_pc_follow_contents(obj, cm);
|
||||
}
|
||||
|
||||
|
||||
void InstanceRefKlass::oop_pc_follow_contents(oop obj, ParCompactionManager* cm) {
|
||||
if (UseCompressedOops) {
|
||||
oop_pc_follow_contents_specialized<narrowOop>(this, obj, cm);
|
||||
} else {
|
||||
oop_pc_follow_contents_specialized<oop>(this, obj, cm);
|
||||
}
|
||||
}
|
||||
|
||||
void ObjArrayKlass::oop_pc_follow_contents(oop obj, ParCompactionManager* cm) {
|
||||
cm->follow_klass(this);
|
||||
|
||||
if (UseCompressedOops) {
|
||||
oop_pc_follow_contents_specialized<narrowOop>(objArrayOop(obj), 0, cm);
|
||||
} else {
|
||||
oop_pc_follow_contents_specialized<oop>(objArrayOop(obj), 0, cm);
|
||||
}
|
||||
}
|
||||
|
||||
void TypeArrayKlass::oop_pc_follow_contents(oop obj, ParCompactionManager* cm) {
|
||||
assert(obj->is_typeArray(),"must be a type array");
|
||||
// Performance tweak: We skip iterating over the klass pointer since we
|
||||
// know that Universe::TypeArrayKlass never moves.
|
||||
}
|
||||
|
||||
void ParCompactionManager::follow_marking_stacks() {
|
||||
do {
|
||||
// Drain the overflow stack first, to allow stealing from the marking stack.
|
||||
oop obj;
|
||||
while (marking_stack()->pop_overflow(obj)) {
|
||||
follow_contents(obj);
|
||||
}
|
||||
while (marking_stack()->pop_local(obj)) {
|
||||
follow_contents(obj);
|
||||
}
|
||||
|
||||
// Process ObjArrays one at a time to avoid marking stack bloat.
|
||||
ObjArrayTask task;
|
||||
if (_objarray_stack.pop_overflow(task) || _objarray_stack.pop_local(task)) {
|
||||
follow_contents((objArrayOop)task.obj(), task.index());
|
||||
}
|
||||
} while (!marking_stacks_empty());
|
||||
|
||||
assert(marking_stacks_empty(), "Sanity");
|
||||
}
|
||||
|
||||
void ParCompactionManager::drain_region_stacks() {
|
||||
do {
|
||||
// Drain overflow stack first so other threads can steal.
|
||||
size_t region_index;
|
||||
while (region_stack()->pop_overflow(region_index)) {
|
||||
PSParallelCompact::fill_and_update_region(this, region_index);
|
||||
}
|
||||
|
||||
while (region_stack()->pop_local(region_index)) {
|
||||
PSParallelCompact::fill_and_update_region(this, region_index);
|
||||
}
|
||||
} while (!region_stack()->is_empty());
|
||||
}
|
Loading…
Add table
Add a link
Reference in a new issue