mirror of
https://github.com/openjdk/jdk.git
synced 2025-09-15 16:44:36 +02:00

Assume pending_list uses discovered field and remove resulting dead code. Reviewed-by: brutisso, pliden
174 lines
6.7 KiB
C++
174 lines
6.7 KiB
C++
/*
|
|
* Copyright (c) 2015, Oracle and/or its affiliates. All rights reserved.
|
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
*
|
|
* This code is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 only, as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
* version 2 for more details (a copy is included in the LICENSE file that
|
|
* accompanied this code).
|
|
*
|
|
* You should have received a copy of the GNU General Public License version
|
|
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*
|
|
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
* or visit www.oracle.com if you need additional information or have any
|
|
* questions.
|
|
*
|
|
*/
|
|
|
|
#ifndef SHARE_VM_OOPS_INSTANCEREFKLASS_INLINE_HPP
|
|
#define SHARE_VM_OOPS_INSTANCEREFKLASS_INLINE_HPP
|
|
|
|
#include "classfile/javaClasses.hpp"
|
|
#include "memory/referenceProcessor.hpp"
|
|
#include "oops/instanceRefKlass.hpp"
|
|
#include "oops/instanceKlass.inline.hpp"
|
|
#include "oops/oop.inline.hpp"
|
|
#include "utilities/debug.hpp"
|
|
#include "utilities/globalDefinitions.hpp"
|
|
#include "utilities/macros.hpp"
|
|
|
|
template <bool nv, typename T, class OopClosureType, class Contains>
|
|
void InstanceRefKlass::oop_oop_iterate_ref_processing_specialized(oop obj, OopClosureType* closure, Contains& contains) {
|
|
T* disc_addr = (T*)java_lang_ref_Reference::discovered_addr(obj);
|
|
if (closure->apply_to_weak_ref_discovered_field()) {
|
|
Devirtualizer<nv>::do_oop(closure, disc_addr);
|
|
}
|
|
|
|
T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
|
|
T heap_oop = oopDesc::load_heap_oop(referent_addr);
|
|
ReferenceProcessor* rp = closure->_ref_processor;
|
|
if (!oopDesc::is_null(heap_oop)) {
|
|
oop referent = oopDesc::decode_heap_oop_not_null(heap_oop);
|
|
if (!referent->is_gc_marked() && (rp != NULL) &&
|
|
rp->discover_reference(obj, reference_type())) {
|
|
return;
|
|
} else if (contains(referent_addr)) {
|
|
// treat referent as normal oop
|
|
Devirtualizer<nv>::do_oop(closure, referent_addr);
|
|
}
|
|
}
|
|
T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
|
|
T next_oop = oopDesc::load_heap_oop(next_addr);
|
|
// Treat discovered as normal oop, if ref is not "active" (next non-NULL)
|
|
if (!oopDesc::is_null(next_oop) && contains(disc_addr)) {
|
|
// i.e. ref is not "active"
|
|
debug_only(
|
|
if(TraceReferenceGC && PrintGCDetails) {
|
|
gclog_or_tty->print_cr(" Process discovered as normal "
|
|
PTR_FORMAT, p2i(disc_addr));
|
|
}
|
|
)
|
|
Devirtualizer<nv>::do_oop(closure, disc_addr);
|
|
}
|
|
// treat next as normal oop
|
|
if (contains(next_addr)) {
|
|
Devirtualizer<nv>::do_oop(closure, next_addr);
|
|
}
|
|
}
|
|
|
|
class AlwaysContains {
|
|
public:
|
|
template <typename T> bool operator()(T* p) const { return true; }
|
|
};
|
|
|
|
template <bool nv, class OopClosureType>
|
|
void InstanceRefKlass::oop_oop_iterate_ref_processing(oop obj, OopClosureType* closure) {
|
|
AlwaysContains always_contains;
|
|
if (UseCompressedOops) {
|
|
oop_oop_iterate_ref_processing_specialized<nv, narrowOop>(obj, closure, always_contains);
|
|
} else {
|
|
oop_oop_iterate_ref_processing_specialized<nv, oop>(obj, closure, always_contains);
|
|
}
|
|
}
|
|
|
|
class MrContains {
|
|
const MemRegion _mr;
|
|
public:
|
|
MrContains(MemRegion mr) : _mr(mr) {}
|
|
template <typename T> bool operator()(T* p) const { return _mr.contains(p); }
|
|
};
|
|
|
|
template <bool nv, class OopClosureType>
|
|
void InstanceRefKlass::oop_oop_iterate_ref_processing_bounded(oop obj, OopClosureType* closure, MemRegion mr) {
|
|
const MrContains contains(mr);
|
|
if (UseCompressedOops) {
|
|
oop_oop_iterate_ref_processing_specialized<nv, narrowOop>(obj, closure, contains);
|
|
} else {
|
|
oop_oop_iterate_ref_processing_specialized<nv, oop>(obj, closure, contains);
|
|
}
|
|
}
|
|
|
|
template <bool nv, class OopClosureType>
|
|
int InstanceRefKlass::oop_oop_iterate(oop obj, OopClosureType* closure) {
|
|
// Get size before changing pointers
|
|
int size = InstanceKlass::oop_oop_iterate<nv>(obj, closure);
|
|
|
|
oop_oop_iterate_ref_processing<nv>(obj, closure);
|
|
|
|
return size;
|
|
}
|
|
|
|
#if INCLUDE_ALL_GCS
|
|
template <bool nv, class OopClosureType>
|
|
int InstanceRefKlass::
|
|
oop_oop_iterate_reverse(oop obj, OopClosureType* closure) {
|
|
// Get size before changing pointers
|
|
int size = InstanceKlass::oop_oop_iterate_reverse<nv>(obj, closure);
|
|
|
|
oop_oop_iterate_ref_processing<nv>(obj, closure);
|
|
|
|
return size;
|
|
}
|
|
#endif // INCLUDE_ALL_GCS
|
|
|
|
|
|
template <bool nv, class OopClosureType>
|
|
int InstanceRefKlass::oop_oop_iterate_bounded(oop obj, OopClosureType* closure, MemRegion mr) {
|
|
// Get size before changing pointers
|
|
int size = InstanceKlass::oop_oop_iterate_bounded<nv>(obj, closure, mr);
|
|
|
|
oop_oop_iterate_ref_processing_bounded<nv>(obj, closure, mr);
|
|
|
|
return size;
|
|
}
|
|
|
|
// Macro to define InstanceRefKlass::oop_oop_iterate for virtual/nonvirtual for
|
|
// all closures. Macros calling macros above for each oop size.
|
|
|
|
#define InstanceRefKlass_OOP_OOP_ITERATE_DEFN(OopClosureType, nv_suffix) \
|
|
\
|
|
int InstanceRefKlass::oop_oop_iterate##nv_suffix(oop obj, OopClosureType* closure) { \
|
|
return oop_oop_iterate<nvs_to_bool(nv_suffix)>(obj, closure); \
|
|
}
|
|
|
|
#if INCLUDE_ALL_GCS
|
|
#define InstanceRefKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN(OopClosureType, nv_suffix) \
|
|
\
|
|
int InstanceRefKlass::oop_oop_iterate_backwards##nv_suffix(oop obj, OopClosureType* closure) { \
|
|
return oop_oop_iterate_reverse<nvs_to_bool(nv_suffix)>(obj, closure); \
|
|
}
|
|
#else
|
|
#define InstanceRefKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN(OopClosureType, nv_suffix)
|
|
#endif
|
|
|
|
|
|
#define InstanceRefKlass_OOP_OOP_ITERATE_DEFN_m(OopClosureType, nv_suffix) \
|
|
\
|
|
int InstanceRefKlass::oop_oop_iterate##nv_suffix##_m(oop obj, OopClosureType* closure, MemRegion mr) { \
|
|
return oop_oop_iterate_bounded<nvs_to_bool(nv_suffix)>(obj, closure, mr); \
|
|
}
|
|
|
|
#define ALL_INSTANCE_REF_KLASS_OOP_OOP_ITERATE_DEFN(OopClosureType, nv_suffix) \
|
|
InstanceRefKlass_OOP_OOP_ITERATE_DEFN( OopClosureType, nv_suffix) \
|
|
InstanceRefKlass_OOP_OOP_ITERATE_DEFN_m( OopClosureType, nv_suffix) \
|
|
InstanceRefKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN(OopClosureType, nv_suffix)
|
|
|
|
|
|
#endif // SHARE_VM_OOPS_INSTANCEREFKLASS_INLINE_HPP
|