comparison src/share/vm/gc_implementation/g1/g1Allocator.cpp @ 20404:227a9e5e4b4a

8057536: Refactor G1 to allow context specific allocations Summary: Splitting out a g1 allocator class to simply specialized allocators which can associate each allocation with a given context. Reviewed-by: mgerdin, brutisso
author sjohanss
date Fri, 05 Sep 2014 09:49:19 +0200
parents
children
comparison
equal deleted inserted replaced
20403:8ec8971f511a 20404:227a9e5e4b4a
1 /*
2 * Copyright (c) 2014, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #include "precompiled.hpp"
26 #include "gc_implementation/g1/g1Allocator.hpp"
27 #include "gc_implementation/g1/g1CollectedHeap.hpp"
28 #include "gc_implementation/g1/g1CollectorPolicy.hpp"
29 #include "gc_implementation/g1/heapRegion.inline.hpp"
30 #include "gc_implementation/g1/heapRegionSet.inline.hpp"
31
32 void G1DefaultAllocator::init_mutator_alloc_region() {
33 assert(_mutator_alloc_region.get() == NULL, "pre-condition");
34 _mutator_alloc_region.init();
35 }
36
37 void G1DefaultAllocator::release_mutator_alloc_region() {
38 _mutator_alloc_region.release();
39 assert(_mutator_alloc_region.get() == NULL, "post-condition");
40 }
41
42 void G1Allocator::reuse_retained_old_region(EvacuationInfo& evacuation_info,
43 OldGCAllocRegion* old,
44 HeapRegion** retained_old) {
45 HeapRegion* retained_region = *retained_old;
46 *retained_old = NULL;
47
48 // We will discard the current GC alloc region if:
49 // a) it's in the collection set (it can happen!),
50 // b) it's already full (no point in using it),
51 // c) it's empty (this means that it was emptied during
52 // a cleanup and it should be on the free list now), or
53 // d) it's humongous (this means that it was emptied
54 // during a cleanup and was added to the free list, but
55 // has been subsequently used to allocate a humongous
56 // object that may be less than the region size).
57 if (retained_region != NULL &&
58 !retained_region->in_collection_set() &&
59 !(retained_region->top() == retained_region->end()) &&
60 !retained_region->is_empty() &&
61 !retained_region->isHumongous()) {
62 retained_region->record_top_and_timestamp();
63 // The retained region was added to the old region set when it was
64 // retired. We have to remove it now, since we don't allow regions
65 // we allocate to in the region sets. We'll re-add it later, when
66 // it's retired again.
67 _g1h->_old_set.remove(retained_region);
68 bool during_im = _g1h->g1_policy()->during_initial_mark_pause();
69 retained_region->note_start_of_copying(during_im);
70 old->set(retained_region);
71 _g1h->_hr_printer.reuse(retained_region);
72 evacuation_info.set_alloc_regions_used_before(retained_region->used());
73 }
74 }
75
76 void G1DefaultAllocator::init_gc_alloc_regions(EvacuationInfo& evacuation_info) {
77 assert_at_safepoint(true /* should_be_vm_thread */);
78
79 _survivor_gc_alloc_region.init();
80 _old_gc_alloc_region.init();
81 reuse_retained_old_region(evacuation_info,
82 &_old_gc_alloc_region,
83 &_retained_old_gc_alloc_region);
84 }
85
86 void G1DefaultAllocator::release_gc_alloc_regions(uint no_of_gc_workers, EvacuationInfo& evacuation_info) {
87 AllocationContext_t context = AllocationContext::current();
88 evacuation_info.set_allocation_regions(survivor_gc_alloc_region(context)->count() +
89 old_gc_alloc_region(context)->count());
90 survivor_gc_alloc_region(context)->release();
91 // If we have an old GC alloc region to release, we'll save it in
92 // _retained_old_gc_alloc_region. If we don't
93 // _retained_old_gc_alloc_region will become NULL. This is what we
94 // want either way so no reason to check explicitly for either
95 // condition.
96 _retained_old_gc_alloc_region = old_gc_alloc_region(context)->release();
97
98 if (ResizePLAB) {
99 _g1h->_survivor_plab_stats.adjust_desired_plab_sz(no_of_gc_workers);
100 _g1h->_old_plab_stats.adjust_desired_plab_sz(no_of_gc_workers);
101 }
102 }
103
104 void G1DefaultAllocator::abandon_gc_alloc_regions() {
105 assert(survivor_gc_alloc_region(AllocationContext::current())->get() == NULL, "pre-condition");
106 assert(old_gc_alloc_region(AllocationContext::current())->get() == NULL, "pre-condition");
107 _retained_old_gc_alloc_region = NULL;
108 }
109
110 G1ParGCAllocBuffer::G1ParGCAllocBuffer(size_t gclab_word_size) :
111 ParGCAllocBuffer(gclab_word_size), _retired(true) { }
112
113 HeapWord* G1ParGCAllocator::allocate_slow(GCAllocPurpose purpose, size_t word_sz, AllocationContext_t context) {
114 HeapWord* obj = NULL;
115 size_t gclab_word_size = _g1h->desired_plab_sz(purpose);
116 if (word_sz * 100 < gclab_word_size * ParallelGCBufferWastePct) {
117 G1ParGCAllocBuffer* alloc_buf = alloc_buffer(purpose, context);
118 add_to_alloc_buffer_waste(alloc_buf->words_remaining());
119 alloc_buf->retire(false /* end_of_gc */, false /* retain */);
120
121 HeapWord* buf = _g1h->par_allocate_during_gc(purpose, gclab_word_size, context);
122 if (buf == NULL) {
123 return NULL; // Let caller handle allocation failure.
124 }
125 // Otherwise.
126 alloc_buf->set_word_size(gclab_word_size);
127 alloc_buf->set_buf(buf);
128
129 obj = alloc_buf->allocate(word_sz);
130 assert(obj != NULL, "buffer was definitely big enough...");
131 } else {
132 obj = _g1h->par_allocate_during_gc(purpose, word_sz, context);
133 }
134 return obj;
135 }
136
137 G1DefaultParGCAllocator::G1DefaultParGCAllocator(G1CollectedHeap* g1h) :
138 G1ParGCAllocator(g1h),
139 _surviving_alloc_buffer(g1h->desired_plab_sz(GCAllocForSurvived)),
140 _tenured_alloc_buffer(g1h->desired_plab_sz(GCAllocForTenured)) {
141
142 _alloc_buffers[GCAllocForSurvived] = &_surviving_alloc_buffer;
143 _alloc_buffers[GCAllocForTenured] = &_tenured_alloc_buffer;
144
145 }
146
147 void G1DefaultParGCAllocator::retire_alloc_buffers() {
148 for (int ap = 0; ap < GCAllocPurposeCount; ++ap) {
149 size_t waste = _alloc_buffers[ap]->words_remaining();
150 add_to_alloc_buffer_waste(waste);
151 _alloc_buffers[ap]->flush_stats_and_retire(_g1h->stats_for_purpose((GCAllocPurpose)ap),
152 true /* end_of_gc */,
153 false /* retain */);
154 }
155 }