RFR (M): 8027959: Investigate early reclamation of large objects in G1
Mikael Gerdin
mikael.gerdin at oracle.com
Thu Jul 17 14:52:22 UTC 2014
Hi Thomas,
On Tuesday 15 July 2014 11.10.53 Thomas Schatzl wrote:
> Hi all,
>
> could I have reviews for the following change that allows G1 to
> eagerly/early reclaim humongous objects on every GC?
>
> Problem:
>
> In G1 large objects are always allocated in the old generation,
> currently requiring a complete heap liveness analysis (full gc, marking)
> to reclaim them.
>
> This is far from ideal for many transaction based enterprise
> applications that create large objects that are only live until a
> (typically short-lived) transaction has been completed (e.g. in a
> ResultSet of a JDBC query that generates a large result,
> byteoutputstreams, etc).
> This results in the heap filling up relatively quickly, typically
> leading to unnecessary marking cycles just to reclaim them.
>
> The solution implemented here is to directly target these types of
> objects by using remembered set and reachability information from any GC
> to make (conservatively) sure that we can reclaim the space.
>
> You can quickly determine this if there are no references from the roots
> or young gen to that object, and if there are no remembered set entries
> to that object. This is sufficient because:
> - g1 root processing walks over all roots and young gen always which
> are sources for potential references.
> - the remembered set contains potential locations that reference this
> object. These are all locations, as humongous objects are always
> allocated into their own regions (so there can be no intra-region
> references).
>
> These are all potential reference locations during GC pause because GC
> pause makes sure that the remembered set is current at pause time.
>
> We can also reclaim if the region is considered live by the marking if
> it has been allocated during marking (and the other conditions hold). At
> reclaim time, if something referenced that object, there either must
> have been a remembered set entry or a reference from the roots/or young
> gen if it is actually live so nobody can install a reference from it any
> more.
> (If there has once been a reference from another old region, it must
> have had a remembered set entry).
> When marking continues after GC, it will simply notice that the region
> has been freed, and skip over it during continue.
>
> After putting the humongous region into the collection set, liveness
> detection occurs by intercepting the slow path for allocation of space
> for that humongous object. As it is humongous, we always end up there.
>
> The change includes some minor optimizations:
> - during registering the humongous regions for inclusion into the
> collection set, we already check whether that humongous object is
> actually one we can potentially remove. E.g. has no remembered set. This
> makes it a "humongous candidate" (note there is no actual state for
> this, just a name for these regions)
> - after finding out that the region is live once, remove that humongous
> region from the collection set so that further references to it do not
> cause use to go into the slow path. This is to avoid going into the slow
> path too often if that object is referenced a lot. (Most likely, if that
> object had many references it would not be a "humongous candidate" btw)
> - if there were no candidates at the start of the GC, then do not
> bother trying to reclaim later.
>
> In total I found no particular slowdown when enabling this feature by
> default. I.e. if there are no humongous candidate objects, there will be
> no change to the current code path at all because none will be added to
> the collection set.
>
> The feature can be disabled completely by disabling
> G1ReclaimDeadHumongousObjectsAtYoungGC.
>
> There is a new log line "Humongous Reclaim" measuring reclaim time, and
> if with G1LogLevel=finest is set it prints some statistics about total,
> candidate and reclaimed humongous objects on the heap.
>
> The CR contains a graph showing large improvements on average humongous
> object reclamation delay. In total we have seen some benchmarks
> reclaiming GBs of heap space over time using this functionality (instead
> of waiting for the marking/full GC). This improves throughput
> significantly as there is more space available for the young gen on
> average now.
>
> Also it might avoid users to manually increase heap region sizes just to
> avoid humongous object troubles.
>
> CR:
> https://bugs.openjdk.java.net/browse/JDK-8027959
>
> Webrev:
> http://cr.openjdk.java.net/~tschatzl/8027959/webrev/
g1CollectedHeap.hpp:
Most of the new functions you add for humongous regions operate on region
indices, except for
+ bool humongous_region_is_always_live(HeapRegion* region);
Is there any reason for it not operating on a region index as well? I think
that would make the API nice and consistent.
g1CollectedHeap.cpp:
This comment is pretty confusing to me, can you try to rephrase it?
3798 if (is_candidate) {
3799 // Do not even try to reclaim a humongous object that we already
know will
3800 // not be treated as live later. A young collection will not
decrease the
3801 // amount of remembered set entries for that region.
forwardee can only be null here for humongous objects, correct?
Would it make sense to add an else-clause asserting that to keep some of the
old assert functionality?
4674 if (forwardee != NULL) {
4675 oopDesc::encode_store_heap_oop(p, forwardee);
4676 if (do_mark_object != G1MarkNone && forwardee != obj) {
4677 // If the object is self-forwarded we don't need to explicitly
4678 // mark it, the evacuation failure protocol will do so.
4679 mark_forwarded_object(obj, forwardee);
4680 }
4681
4682 if (barrier == G1BarrierKlass) {
4683 do_klass_barrier(p, forwardee);
4684 }
4685 needs_marking = false;
4686 }
g1OopClosures.inline.hpp
Are you sure that the change to is_in_cset_or_humongous is correct here?
I think that when FilterIntoCSClosure is used for card refinement we are not
actually interested in references to humongous objects since they don't move.
So this could cause us to needlessly add cards to the into_cset_dcq.
However in the case of scanning remembered sets it's the new check is probably
exactly what you are looking for.
43 template <class T>
44 inline void FilterIntoCSClosure::do_oop_nv(T* p) {
45 T heap_oop = oopDesc::load_heap_oop(p);
46 if (!oopDesc::is_null(heap_oop) &&
47 _g1-
>is_in_cset_or_humongous(oopDesc::decode_heap_oop_not_null(heap_oop))) {
48 _oc->do_oop(p);
49 }
50 }
Perhaps this should be left for a further cleanup, one approach could be to
fold the logic using is_in_cset_or_humongous into HeapRegion_DCTOC since
scanCard seems to be the only place where that closure is used.
Previsously the else-clause here would cause a remembered set entry of the
region containing "obj" to be added, to remember the reference "p"
Now that you're removing humongous objects from this consideration, can we
miss remembered set entries for humongous regions?
65 inline void G1ParScanClosure::do_oop_nv(T* p) {
66 T heap_oop = oopDesc::load_heap_oop(p);
67
68 if (!oopDesc::is_null(heap_oop)) {
69 oop obj = oopDesc::decode_heap_oop_not_null(heap_oop);
70 if (_g1->is_in_cset_or_humongous(obj)) {
71 // We're not going to even bother checking whether the object is
72 // already forwarded or not, as this usually causes an immediate
73 // stall. We'll try to prefetch the object (for write, given that
74 // we might need to install the forwarding reference) and we'll
75 // get back to it when pop it from the queue
76 Prefetch::write(obj->mark_addr(), 0);
77 Prefetch::read(obj->mark_addr(), (HeapWordSize*2));
78
79 // slightly paranoid test; I'm trying to catch potential
80 // problems before we go into push_on_queue to know where the
81 // problem is coming from
82 assert((obj == oopDesc::load_decode_heap_oop(p)) ||
83 (obj->is_forwarded() &&
84 obj->forwardee() == oopDesc::load_decode_heap_oop(p)),
85 "p should still be pointing to obj or to its forwardee");
86
87 _par_scan_state->push_on_queue(p);
88 } else {
89 _par_scan_state->update_rs(_from, p, _worker_id);
90 }
I also somewhat agree with Bengt's opinion about explicitly checking for
humongous objects.
/Mikael
>
> Testing:
> jprt, aurora adhoc, various internal benchmarks
>
> Thanks,
> Thomas
More information about the hotspot-gc-dev
mailing list