mirror of
https://github.com/openjdk/jdk.git
synced 2025-09-21 03:24:38 +02:00
Merge
This commit is contained in:
commit
428d02d5d7
4 changed files with 16 additions and 16 deletions
|
@ -81,20 +81,24 @@ void G1MMUTrackerQueue::add_pause(double start, double end, bool gc_thread) {
|
||||||
|
|
||||||
remove_expired_entries(end);
|
remove_expired_entries(end);
|
||||||
if (_no_entries == QueueLength) {
|
if (_no_entries == QueueLength) {
|
||||||
// OK, right now when we fill up we bomb out
|
// OK, we've filled up the queue. There are a few ways
|
||||||
// there are a few ways of dealing with this "gracefully"
|
// of dealing with this "gracefully"
|
||||||
// increase the array size (:-)
|
// increase the array size (:-)
|
||||||
// remove the oldest entry (this might allow more GC time for
|
// remove the oldest entry (this might allow more GC time for
|
||||||
// the time slice than what's allowed)
|
// the time slice than what's allowed) - this is what we
|
||||||
|
// currently do
|
||||||
// consolidate the two entries with the minimum gap between them
|
// consolidate the two entries with the minimum gap between them
|
||||||
// (this might allow less GC time than what's allowed)
|
// (this might allow less GC time than what's allowed)
|
||||||
guarantee(NOT_PRODUCT(ScavengeALot ||) G1UseFixedWindowMMUTracker,
|
|
||||||
"array full, currently we can't recover unless +G1UseFixedWindowMMUTracker");
|
|
||||||
// In the case where ScavengeALot is true, such overflow is not
|
// In the case where ScavengeALot is true, such overflow is not
|
||||||
// uncommon; in such cases, we can, without much loss of precision
|
// uncommon; in such cases, we can, without much loss of precision
|
||||||
// or performance (we are GC'ing most of the time anyway!),
|
// or performance (we are GC'ing most of the time anyway!),
|
||||||
// simply overwrite the oldest entry in the tracker: this
|
// simply overwrite the oldest entry in the tracker.
|
||||||
// is also the behaviour when G1UseFixedWindowMMUTracker is enabled.
|
|
||||||
|
if (G1PolicyVerbose > 1) {
|
||||||
|
warning("MMU Tracker Queue overflow. Replacing earliest entry.");
|
||||||
|
}
|
||||||
|
|
||||||
_head_index = trim_index(_head_index + 1);
|
_head_index = trim_index(_head_index + 1);
|
||||||
assert(_head_index == _tail_index, "Because we have a full circular buffer");
|
assert(_head_index == _tail_index, "Because we have a full circular buffer");
|
||||||
_tail_index = trim_index(_tail_index + 1);
|
_tail_index = trim_index(_tail_index + 1);
|
||||||
|
|
|
@ -254,9 +254,6 @@
|
||||||
"If non-0 is the size of the G1 survivor space, " \
|
"If non-0 is the size of the G1 survivor space, " \
|
||||||
"otherwise SurvivorRatio is used to determine the size") \
|
"otherwise SurvivorRatio is used to determine the size") \
|
||||||
\
|
\
|
||||||
product(bool, G1UseFixedWindowMMUTracker, false, \
|
|
||||||
"If the MMU tracker's memory is full, forget the oldest entry") \
|
|
||||||
\
|
|
||||||
product(uintx, G1HeapRegionSize, 0, \
|
product(uintx, G1HeapRegionSize, 0, \
|
||||||
"Size of the G1 regions.") \
|
"Size of the G1 regions.") \
|
||||||
\
|
\
|
||||||
|
|
|
@ -3487,7 +3487,6 @@ void GraphKit::g1_write_barrier_post(Node* oop_store,
|
||||||
|
|
||||||
Node* tls = __ thread(); // ThreadLocalStorage
|
Node* tls = __ thread(); // ThreadLocalStorage
|
||||||
|
|
||||||
Node* no_ctrl = NULL;
|
|
||||||
Node* no_base = __ top();
|
Node* no_base = __ top();
|
||||||
float likely = PROB_LIKELY(0.999);
|
float likely = PROB_LIKELY(0.999);
|
||||||
float unlikely = PROB_UNLIKELY(0.999);
|
float unlikely = PROB_UNLIKELY(0.999);
|
||||||
|
@ -3511,10 +3510,10 @@ void GraphKit::g1_write_barrier_post(Node* oop_store,
|
||||||
Node* index_adr = __ AddP(no_base, tls, __ ConX(index_offset));
|
Node* index_adr = __ AddP(no_base, tls, __ ConX(index_offset));
|
||||||
|
|
||||||
// Now some values
|
// Now some values
|
||||||
|
// Use ctrl to avoid hoisting these values past a safepoint, which could
|
||||||
Node* index = __ load(no_ctrl, index_adr, TypeInt::INT, T_INT, Compile::AliasIdxRaw);
|
// potentially reset these fields in the JavaThread.
|
||||||
Node* buffer = __ load(no_ctrl, buffer_adr, TypeRawPtr::NOTNULL, T_ADDRESS, Compile::AliasIdxRaw);
|
Node* index = __ load(__ ctrl(), index_adr, TypeInt::INT, T_INT, Compile::AliasIdxRaw);
|
||||||
|
Node* buffer = __ load(__ ctrl(), buffer_adr, TypeRawPtr::NOTNULL, T_ADDRESS, Compile::AliasIdxRaw);
|
||||||
|
|
||||||
// Convert the store obj pointer to an int prior to doing math on it
|
// Convert the store obj pointer to an int prior to doing math on it
|
||||||
// Must use ctrl to prevent "integerized oop" existing across safepoint
|
// Must use ctrl to prevent "integerized oop" existing across safepoint
|
||||||
|
|
|
@ -65,7 +65,7 @@ do
|
||||||
done
|
done
|
||||||
rm -f $$
|
rm -f $$
|
||||||
|
|
||||||
i=$(expr $i + 1)
|
i=`expr $i + 1`
|
||||||
done
|
done
|
||||||
|
|
||||||
exit $rc
|
exit $rc
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue