Mercurial > hg > graal-compiler
annotate src/share/vm/gc_implementation/g1/heapRegion.hpp @ 4097:dc467e8b2c5e
7112743: G1: Reduce overhead of marking closure during evacuation pauses
Summary: Parallelize the serial code that was used to mark objects reachable from survivor objects in the collection set. Some minor improvments in the timers used to track the freeing of the collection set along with some tweaks to PrintGCDetails.
Reviewed-by: tonyp, brutisso
author | johnc |
---|---|
date | Thu, 17 Nov 2011 12:40:15 -0800 |
parents | 6071e0581859 |
children | 023652e49ac0 |
rev | line source |
---|---|
342 | 1 /* |
2133
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
2 * Copyright (c) 2001, 2011, Oracle and/or its affiliates. All rights reserved. |
342 | 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 * | |
5 * This code is free software; you can redistribute it and/or modify it | |
6 * under the terms of the GNU General Public License version 2 only, as | |
7 * published by the Free Software Foundation. | |
8 * | |
9 * This code is distributed in the hope that it will be useful, but WITHOUT | |
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License | |
12 * version 2 for more details (a copy is included in the LICENSE file that | |
13 * accompanied this code). | |
14 * | |
15 * You should have received a copy of the GNU General Public License version | |
16 * 2 along with this work; if not, write to the Free Software Foundation, | |
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | |
18 * | |
1552
c18cbe5936b8
6941466: Oracle rebranding changes for Hotspot repositories
trims
parents:
1394
diff
changeset
|
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
c18cbe5936b8
6941466: Oracle rebranding changes for Hotspot repositories
trims
parents:
1394
diff
changeset
|
20 * or visit www.oracle.com if you need additional information or have any |
c18cbe5936b8
6941466: Oracle rebranding changes for Hotspot repositories
trims
parents:
1394
diff
changeset
|
21 * questions. |
342 | 22 * |
23 */ | |
24 | |
1972 | 25 #ifndef SHARE_VM_GC_IMPLEMENTATION_G1_HEAPREGION_HPP |
26 #define SHARE_VM_GC_IMPLEMENTATION_G1_HEAPREGION_HPP | |
27 | |
28 #include "gc_implementation/g1/g1BlockOffsetTable.inline.hpp" | |
29 #include "gc_implementation/g1/g1_specialized_oop_closures.hpp" | |
30 #include "gc_implementation/g1/survRateGroup.hpp" | |
31 #include "gc_implementation/shared/ageTable.hpp" | |
32 #include "gc_implementation/shared/spaceDecorator.hpp" | |
33 #include "memory/space.inline.hpp" | |
34 #include "memory/watermark.hpp" | |
35 | |
342 | 36 #ifndef SERIALGC |
37 | |
38 // A HeapRegion is the smallest piece of a G1CollectedHeap that | |
39 // can be collected independently. | |
40 | |
41 // NOTE: Although a HeapRegion is a Space, its | |
42 // Space::initDirtyCardClosure method must not be called. | |
43 // The problem is that the existence of this method breaks | |
44 // the independence of barrier sets from remembered sets. | |
45 // The solution is to remove this method from the definition | |
46 // of a Space. | |
47 | |
48 class CompactibleSpace; | |
49 class ContiguousSpace; | |
50 class HeapRegionRemSet; | |
51 class HeapRegionRemSetIterator; | |
52 class HeapRegion; | |
2152 | 53 class HeapRegionSetBase; |
54 | |
3766 | 55 #define HR_FORMAT SIZE_FORMAT":(%s)["PTR_FORMAT","PTR_FORMAT","PTR_FORMAT"]" |
56 #define HR_FORMAT_PARAMS(_hr_) \ | |
57 (_hr_)->hrs_index(), \ | |
58 (_hr_)->is_survivor() ? "S" : (_hr_)->is_young() ? "E" : "-", \ | |
59 (_hr_)->bottom(), (_hr_)->top(), (_hr_)->end() | |
342 | 60 |
61 // A dirty card to oop closure for heap regions. It | |
62 // knows how to get the G1 heap and how to use the bitmap | |
63 // in the concurrent marker used by G1 to filter remembered | |
64 // sets. | |
65 | |
66 class HeapRegionDCTOC : public ContiguousSpaceDCTOC { | |
67 public: | |
68 // Specification of possible DirtyCardToOopClosure filtering. | |
69 enum FilterKind { | |
70 NoFilterKind, | |
71 IntoCSFilterKind, | |
72 OutOfRegionFilterKind | |
73 }; | |
74 | |
75 protected: | |
76 HeapRegion* _hr; | |
77 FilterKind _fk; | |
78 G1CollectedHeap* _g1; | |
79 | |
80 void walk_mem_region_with_cl(MemRegion mr, | |
81 HeapWord* bottom, HeapWord* top, | |
82 OopClosure* cl); | |
83 | |
84 // We don't specialize this for FilteringClosure; filtering is handled by | |
85 // the "FilterKind" mechanism. But we provide this to avoid a compiler | |
86 // warning. | |
87 void walk_mem_region_with_cl(MemRegion mr, | |
88 HeapWord* bottom, HeapWord* top, | |
89 FilteringClosure* cl) { | |
90 HeapRegionDCTOC::walk_mem_region_with_cl(mr, bottom, top, | |
91 (OopClosure*)cl); | |
92 } | |
93 | |
94 // Get the actual top of the area on which the closure will | |
95 // operate, given where the top is assumed to be (the end of the | |
96 // memory region passed to do_MemRegion) and where the object | |
97 // at the top is assumed to start. For example, an object may | |
98 // start at the top but actually extend past the assumed top, | |
99 // in which case the top becomes the end of the object. | |
100 HeapWord* get_actual_top(HeapWord* top, HeapWord* top_obj) { | |
101 return ContiguousSpaceDCTOC::get_actual_top(top, top_obj); | |
102 } | |
103 | |
104 // Walk the given memory region from bottom to (actual) top | |
105 // looking for objects and applying the oop closure (_cl) to | |
106 // them. The base implementation of this treats the area as | |
107 // blocks, where a block may or may not be an object. Sub- | |
108 // classes should override this to provide more accurate | |
109 // or possibly more efficient walking. | |
110 void walk_mem_region(MemRegion mr, HeapWord* bottom, HeapWord* top) { | |
111 Filtering_DCTOC::walk_mem_region(mr, bottom, top); | |
112 } | |
113 | |
114 public: | |
115 HeapRegionDCTOC(G1CollectedHeap* g1, | |
116 HeapRegion* hr, OopClosure* cl, | |
117 CardTableModRefBS::PrecisionStyle precision, | |
118 FilterKind fk); | |
119 }; | |
120 | |
121 // The complicating factor is that BlockOffsetTable diverged | |
122 // significantly, and we need functionality that is only in the G1 version. | |
123 // So I copied that code, which led to an alternate G1 version of | |
124 // OffsetTableContigSpace. If the two versions of BlockOffsetTable could | |
125 // be reconciled, then G1OffsetTableContigSpace could go away. | |
126 | |
127 // The idea behind time stamps is the following. Doing a save_marks on | |
128 // all regions at every GC pause is time consuming (if I remember | |
129 // well, 10ms or so). So, we would like to do that only for regions | |
130 // that are GC alloc regions. To achieve this, we use time | |
131 // stamps. For every evacuation pause, G1CollectedHeap generates a | |
132 // unique time stamp (essentially a counter that gets | |
133 // incremented). Every time we want to call save_marks on a region, | |
134 // we set the saved_mark_word to top and also copy the current GC | |
135 // time stamp to the time stamp field of the space. Reading the | |
136 // saved_mark_word involves checking the time stamp of the | |
137 // region. If it is the same as the current GC time stamp, then we | |
138 // can safely read the saved_mark_word field, as it is valid. If the | |
139 // time stamp of the region is not the same as the current GC time | |
140 // stamp, then we instead read top, as the saved_mark_word field is | |
141 // invalid. Time stamps (on the regions and also on the | |
142 // G1CollectedHeap) are reset at every cleanup (we iterate over | |
143 // the regions anyway) and at the end of a Full GC. The current scheme | |
144 // that uses sequential unsigned ints will fail only if we have 4b | |
145 // evacuation pauses between two cleanups, which is _highly_ unlikely. | |
146 | |
147 class G1OffsetTableContigSpace: public ContiguousSpace { | |
148 friend class VMStructs; | |
149 protected: | |
150 G1BlockOffsetArrayContigSpace _offsets; | |
151 Mutex _par_alloc_lock; | |
152 volatile unsigned _gc_time_stamp; | |
2433
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
153 // When we need to retire an allocation region, while other threads |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
154 // are also concurrently trying to allocate into it, we typically |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
155 // allocate a dummy object at the end of the region to ensure that |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
156 // no more allocations can take place in it. However, sometimes we |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
157 // want to know where the end of the last "real" object we allocated |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
158 // into the region was and this is what this keeps track. |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
159 HeapWord* _pre_dummy_top; |
342 | 160 |
161 public: | |
162 // Constructor. If "is_zeroed" is true, the MemRegion "mr" may be | |
163 // assumed to contain zeros. | |
164 G1OffsetTableContigSpace(G1BlockOffsetSharedArray* sharedOffsetArray, | |
165 MemRegion mr, bool is_zeroed = false); | |
166 | |
167 void set_bottom(HeapWord* value); | |
168 void set_end(HeapWord* value); | |
169 | |
170 virtual HeapWord* saved_mark_word() const; | |
171 virtual void set_saved_mark(); | |
172 void reset_gc_time_stamp() { _gc_time_stamp = 0; } | |
173 | |
2433
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
174 // See the comment above in the declaration of _pre_dummy_top for an |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
175 // explanation of what it is. |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
176 void set_pre_dummy_top(HeapWord* pre_dummy_top) { |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
177 assert(is_in(pre_dummy_top) && pre_dummy_top <= top(), "pre-condition"); |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
178 _pre_dummy_top = pre_dummy_top; |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
179 } |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
180 HeapWord* pre_dummy_top() { |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
181 return (_pre_dummy_top == NULL) ? top() : _pre_dummy_top; |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
182 } |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
183 void reset_pre_dummy_top() { _pre_dummy_top = NULL; } |
abdfc822206f
7023069: G1: Introduce symmetric locking in the slow allocation path
tonyp
parents:
2361
diff
changeset
|
184 |
356 | 185 virtual void initialize(MemRegion mr, bool clear_space, bool mangle_space); |
186 virtual void clear(bool mangle_space); | |
342 | 187 |
188 HeapWord* block_start(const void* p); | |
189 HeapWord* block_start_const(const void* p) const; | |
190 | |
191 // Add offset table update. | |
192 virtual HeapWord* allocate(size_t word_size); | |
193 HeapWord* par_allocate(size_t word_size); | |
194 | |
195 // MarkSweep support phase3 | |
196 virtual HeapWord* initialize_threshold(); | |
197 virtual HeapWord* cross_threshold(HeapWord* start, HeapWord* end); | |
198 | |
199 virtual void print() const; | |
2133
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
200 |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
201 void reset_bot() { |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
202 _offsets.zero_bottom_entry(); |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
203 _offsets.initialize_threshold(); |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
204 } |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
205 |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
206 void update_bot_for_object(HeapWord* start, size_t word_size) { |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
207 _offsets.alloc_block(start, word_size); |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
208 } |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
209 |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
210 void print_bot_on(outputStream* out) { |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
211 _offsets.print_on(out); |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
212 } |
342 | 213 }; |
214 | |
215 class HeapRegion: public G1OffsetTableContigSpace { | |
216 friend class VMStructs; | |
217 private: | |
218 | |
355 | 219 enum HumongousType { |
220 NotHumongous = 0, | |
221 StartsHumongous, | |
222 ContinuesHumongous | |
223 }; | |
224 | |
342 | 225 // Requires that the region "mr" be dense with objects, and begin and end |
226 // with an object. | |
227 void oops_in_mr_iterate(MemRegion mr, OopClosure* cl); | |
228 | |
229 // The remembered set for this region. | |
230 // (Might want to make this "inline" later, to avoid some alloc failure | |
231 // issues.) | |
232 HeapRegionRemSet* _rem_set; | |
233 | |
234 G1BlockOffsetArrayContigSpace* offsets() { return &_offsets; } | |
235 | |
236 protected: | |
3766 | 237 // The index of this region in the heap region sequence. |
238 size_t _hrs_index; | |
342 | 239 |
355 | 240 HumongousType _humongous_type; |
342 | 241 // For a humongous region, region in which it starts. |
242 HeapRegion* _humongous_start_region; | |
243 // For the start region of a humongous sequence, it's original end(). | |
244 HeapWord* _orig_end; | |
245 | |
246 // True iff the region is in current collection_set. | |
247 bool _in_collection_set; | |
248 | |
249 // True iff an attempt to evacuate an object in the region failed. | |
250 bool _evacuation_failed; | |
251 | |
252 // A heap region may be a member one of a number of special subsets, each | |
253 // represented as linked lists through the field below. Currently, these | |
254 // sets include: | |
255 // The collection set. | |
256 // The set of allocation regions used in a collection pause. | |
257 // Spaces that may contain gray objects. | |
258 HeapRegion* _next_in_special_set; | |
259 | |
260 // next region in the young "generation" region set | |
261 HeapRegion* _next_young_region; | |
262 | |
796
29e7d79232b9
6819065: G1: eliminate high serial card table clearing time
apetrusenko
parents:
677
diff
changeset
|
263 // Next region whose cards need cleaning |
29e7d79232b9
6819065: G1: eliminate high serial card table clearing time
apetrusenko
parents:
677
diff
changeset
|
264 HeapRegion* _next_dirty_cards_region; |
29e7d79232b9
6819065: G1: eliminate high serial card table clearing time
apetrusenko
parents:
677
diff
changeset
|
265 |
2152 | 266 // Fields used by the HeapRegionSetBase class and subclasses. |
267 HeapRegion* _next; | |
268 #ifdef ASSERT | |
269 HeapRegionSetBase* _containing_set; | |
270 #endif // ASSERT | |
271 bool _pending_removal; | |
272 | |
342 | 273 // For parallel heapRegion traversal. |
274 jint _claimed; | |
275 | |
276 // We use concurrent marking to determine the amount of live data | |
277 // in each heap region. | |
278 size_t _prev_marked_bytes; // Bytes known to be live via last completed marking. | |
279 size_t _next_marked_bytes; // Bytes known to be live via in-progress marking. | |
280 | |
281 // See "sort_index" method. -1 means is not in the array. | |
282 int _sort_index; | |
283 | |
284 // <PREDICTION> | |
285 double _gc_efficiency; | |
286 // </PREDICTION> | |
287 | |
288 enum YoungType { | |
289 NotYoung, // a region is not young | |
290 Young, // a region is young | |
3766 | 291 Survivor // a region is young and it contains survivors |
342 | 292 }; |
293 | |
1666
5cbac8938c4c
6956639: G1: assert(cached_ptr != card_ptr) failed: shouldn't be, concurrentG1Refine.cpp:307
johnc
parents:
1552
diff
changeset
|
294 volatile YoungType _young_type; |
342 | 295 int _young_index_in_cset; |
296 SurvRateGroup* _surv_rate_group; | |
297 int _age_index; | |
298 | |
299 // The start of the unmarked area. The unmarked area extends from this | |
300 // word until the top and/or end of the region, and is the part | |
301 // of the region for which no marking was done, i.e. objects may | |
302 // have been allocated in this part since the last mark phase. | |
303 // "prev" is the top at the start of the last completed marking. | |
304 // "next" is the top at the start of the in-progress marking (if any.) | |
305 HeapWord* _prev_top_at_mark_start; | |
306 HeapWord* _next_top_at_mark_start; | |
307 // If a collection pause is in progress, this is the top at the start | |
308 // of that pause. | |
309 | |
310 // We've counted the marked bytes of objects below here. | |
311 HeapWord* _top_at_conc_mark_count; | |
312 | |
313 void init_top_at_mark_start() { | |
314 assert(_prev_marked_bytes == 0 && | |
315 _next_marked_bytes == 0, | |
316 "Must be called after zero_marked_bytes."); | |
317 HeapWord* bot = bottom(); | |
318 _prev_top_at_mark_start = bot; | |
319 _next_top_at_mark_start = bot; | |
320 _top_at_conc_mark_count = bot; | |
321 } | |
322 | |
323 void set_young_type(YoungType new_type) { | |
324 //assert(_young_type != new_type, "setting the same type" ); | |
325 // TODO: add more assertions here | |
326 _young_type = new_type; | |
327 } | |
328 | |
1394
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
329 // Cached attributes used in the collection set policy information |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
330 |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
331 // The RSet length that was added to the total value |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
332 // for the collection set. |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
333 size_t _recorded_rs_length; |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
334 |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
335 // The predicted elapsed time that was added to total value |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
336 // for the collection set. |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
337 double _predicted_elapsed_time_ms; |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
338 |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
339 // The predicted number of bytes to copy that was added to |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
340 // the total value for the collection set. |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
341 size_t _predicted_bytes_to_copy; |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
342 |
342 | 343 public: |
344 // If "is_zeroed" is "true", the region "mr" can be assumed to contain zeros. | |
3766 | 345 HeapRegion(size_t hrs_index, |
346 G1BlockOffsetSharedArray* sharedOffsetArray, | |
342 | 347 MemRegion mr, bool is_zeroed); |
348 | |
3986
65a8ff39a6da
7095194: G1: HeapRegion::GrainBytes, GrainWords, and CardsPerRegion should be size_t
johnc
parents:
3980
diff
changeset
|
349 static int LogOfHRGrainBytes; |
65a8ff39a6da
7095194: G1: HeapRegion::GrainBytes, GrainWords, and CardsPerRegion should be size_t
johnc
parents:
3980
diff
changeset
|
350 static int LogOfHRGrainWords; |
65a8ff39a6da
7095194: G1: HeapRegion::GrainBytes, GrainWords, and CardsPerRegion should be size_t
johnc
parents:
3980
diff
changeset
|
351 |
65a8ff39a6da
7095194: G1: HeapRegion::GrainBytes, GrainWords, and CardsPerRegion should be size_t
johnc
parents:
3980
diff
changeset
|
352 static size_t GrainBytes; |
65a8ff39a6da
7095194: G1: HeapRegion::GrainBytes, GrainWords, and CardsPerRegion should be size_t
johnc
parents:
3980
diff
changeset
|
353 static size_t GrainWords; |
65a8ff39a6da
7095194: G1: HeapRegion::GrainBytes, GrainWords, and CardsPerRegion should be size_t
johnc
parents:
3980
diff
changeset
|
354 static size_t CardsPerRegion; |
942
2c79770d1f6e
6819085: G1: use larger and/or user settable region size
tonyp
parents:
811
diff
changeset
|
355 |
3980
8229bd737950
7075646: G1: fix inconsistencies in the monitoring data
tonyp
parents:
3979
diff
changeset
|
356 static size_t align_up_to_region_byte_size(size_t sz) { |
8229bd737950
7075646: G1: fix inconsistencies in the monitoring data
tonyp
parents:
3979
diff
changeset
|
357 return (sz + (size_t) GrainBytes - 1) & |
8229bd737950
7075646: G1: fix inconsistencies in the monitoring data
tonyp
parents:
3979
diff
changeset
|
358 ~((1 << (size_t) LogOfHRGrainBytes) - 1); |
8229bd737950
7075646: G1: fix inconsistencies in the monitoring data
tonyp
parents:
3979
diff
changeset
|
359 } |
8229bd737950
7075646: G1: fix inconsistencies in the monitoring data
tonyp
parents:
3979
diff
changeset
|
360 |
942
2c79770d1f6e
6819085: G1: use larger and/or user settable region size
tonyp
parents:
811
diff
changeset
|
361 // It sets up the heap region size (GrainBytes / GrainWords), as |
2c79770d1f6e
6819085: G1: use larger and/or user settable region size
tonyp
parents:
811
diff
changeset
|
362 // well as other related fields that are based on the heap region |
2c79770d1f6e
6819085: G1: use larger and/or user settable region size
tonyp
parents:
811
diff
changeset
|
363 // size (LogOfHRGrainBytes / LogOfHRGrainWords / |
2c79770d1f6e
6819085: G1: use larger and/or user settable region size
tonyp
parents:
811
diff
changeset
|
364 // CardsPerRegion). All those fields are considered constant |
2c79770d1f6e
6819085: G1: use larger and/or user settable region size
tonyp
parents:
811
diff
changeset
|
365 // throughout the JVM's execution, therefore they should only be set |
2c79770d1f6e
6819085: G1: use larger and/or user settable region size
tonyp
parents:
811
diff
changeset
|
366 // up once during initialization time. |
2c79770d1f6e
6819085: G1: use larger and/or user settable region size
tonyp
parents:
811
diff
changeset
|
367 static void setup_heap_region_size(uintx min_heap_size); |
342 | 368 |
355 | 369 enum ClaimValues { |
4097
dc467e8b2c5e
7112743: G1: Reduce overhead of marking closure during evacuation pauses
johnc
parents:
4093
diff
changeset
|
370 InitialClaimValue = 0, |
dc467e8b2c5e
7112743: G1: Reduce overhead of marking closure during evacuation pauses
johnc
parents:
4093
diff
changeset
|
371 FinalCountClaimValue = 1, |
dc467e8b2c5e
7112743: G1: Reduce overhead of marking closure during evacuation pauses
johnc
parents:
4093
diff
changeset
|
372 NoteEndClaimValue = 2, |
dc467e8b2c5e
7112743: G1: Reduce overhead of marking closure during evacuation pauses
johnc
parents:
4093
diff
changeset
|
373 ScrubRemSetClaimValue = 3, |
dc467e8b2c5e
7112743: G1: Reduce overhead of marking closure during evacuation pauses
johnc
parents:
4093
diff
changeset
|
374 ParVerifyClaimValue = 4, |
dc467e8b2c5e
7112743: G1: Reduce overhead of marking closure during evacuation pauses
johnc
parents:
4093
diff
changeset
|
375 RebuildRSClaimValue = 5, |
dc467e8b2c5e
7112743: G1: Reduce overhead of marking closure during evacuation pauses
johnc
parents:
4093
diff
changeset
|
376 CompleteMarkCSetClaimValue = 6 |
355 | 377 }; |
378 | |
2134
b158bed62ef5
6994297: G1: do first-level slow-path allocations with a CAS
tonyp
parents:
2133
diff
changeset
|
379 inline HeapWord* par_allocate_no_bot_updates(size_t word_size) { |
b158bed62ef5
6994297: G1: do first-level slow-path allocations with a CAS
tonyp
parents:
2133
diff
changeset
|
380 assert(is_young(), "we can only skip BOT updates on young regions"); |
b158bed62ef5
6994297: G1: do first-level slow-path allocations with a CAS
tonyp
parents:
2133
diff
changeset
|
381 return ContiguousSpace::par_allocate(word_size); |
b158bed62ef5
6994297: G1: do first-level slow-path allocations with a CAS
tonyp
parents:
2133
diff
changeset
|
382 } |
b158bed62ef5
6994297: G1: do first-level slow-path allocations with a CAS
tonyp
parents:
2133
diff
changeset
|
383 inline HeapWord* allocate_no_bot_updates(size_t word_size) { |
b158bed62ef5
6994297: G1: do first-level slow-path allocations with a CAS
tonyp
parents:
2133
diff
changeset
|
384 assert(is_young(), "we can only skip BOT updates on young regions"); |
b158bed62ef5
6994297: G1: do first-level slow-path allocations with a CAS
tonyp
parents:
2133
diff
changeset
|
385 return ContiguousSpace::allocate(word_size); |
b158bed62ef5
6994297: G1: do first-level slow-path allocations with a CAS
tonyp
parents:
2133
diff
changeset
|
386 } |
b158bed62ef5
6994297: G1: do first-level slow-path allocations with a CAS
tonyp
parents:
2133
diff
changeset
|
387 |
342 | 388 // If this region is a member of a HeapRegionSeq, the index in that |
389 // sequence, otherwise -1. | |
3766 | 390 size_t hrs_index() const { return _hrs_index; } |
342 | 391 |
392 // The number of bytes marked live in the region in the last marking phase. | |
393 size_t marked_bytes() { return _prev_marked_bytes; } | |
2435
371bbc844bf1
7027766: G1: introduce flag to dump the liveness information per region at the end of marking
tonyp
parents:
2433
diff
changeset
|
394 size_t live_bytes() { |
371bbc844bf1
7027766: G1: introduce flag to dump the liveness information per region at the end of marking
tonyp
parents:
2433
diff
changeset
|
395 return (top() - prev_top_at_mark_start()) * HeapWordSize + marked_bytes(); |
371bbc844bf1
7027766: G1: introduce flag to dump the liveness information per region at the end of marking
tonyp
parents:
2433
diff
changeset
|
396 } |
371bbc844bf1
7027766: G1: introduce flag to dump the liveness information per region at the end of marking
tonyp
parents:
2433
diff
changeset
|
397 |
342 | 398 // The number of bytes counted in the next marking. |
399 size_t next_marked_bytes() { return _next_marked_bytes; } | |
400 // The number of bytes live wrt the next marking. | |
401 size_t next_live_bytes() { | |
2435
371bbc844bf1
7027766: G1: introduce flag to dump the liveness information per region at the end of marking
tonyp
parents:
2433
diff
changeset
|
402 return |
371bbc844bf1
7027766: G1: introduce flag to dump the liveness information per region at the end of marking
tonyp
parents:
2433
diff
changeset
|
403 (top() - next_top_at_mark_start()) * HeapWordSize + next_marked_bytes(); |
342 | 404 } |
405 | |
406 // A lower bound on the amount of garbage bytes in the region. | |
407 size_t garbage_bytes() { | |
408 size_t used_at_mark_start_bytes = | |
409 (prev_top_at_mark_start() - bottom()) * HeapWordSize; | |
410 assert(used_at_mark_start_bytes >= marked_bytes(), | |
411 "Can't mark more than we have."); | |
412 return used_at_mark_start_bytes - marked_bytes(); | |
413 } | |
414 | |
415 // An upper bound on the number of live bytes in the region. | |
416 size_t max_live_bytes() { return used() - garbage_bytes(); } | |
417 | |
418 void add_to_marked_bytes(size_t incr_bytes) { | |
419 _next_marked_bytes = _next_marked_bytes + incr_bytes; | |
4093
6071e0581859
7111795: G1: Various cleanups identified during walk through of changes for 6484965
johnc
parents:
4023
diff
changeset
|
420 assert(_next_marked_bytes <= used(), "invariant" ); |
342 | 421 } |
422 | |
423 void zero_marked_bytes() { | |
424 _prev_marked_bytes = _next_marked_bytes = 0; | |
425 } | |
426 | |
355 | 427 bool isHumongous() const { return _humongous_type != NotHumongous; } |
428 bool startsHumongous() const { return _humongous_type == StartsHumongous; } | |
429 bool continuesHumongous() const { return _humongous_type == ContinuesHumongous; } | |
342 | 430 // For a humongous region, region in which it starts. |
431 HeapRegion* humongous_start_region() const { | |
432 return _humongous_start_region; | |
433 } | |
434 | |
4020
5e5d4821bf07
7097516: G1: assert(0<= from_card && from_card<HeapRegion::CardsPerRegion) failed: Must be in range.
brutisso
parents:
3986
diff
changeset
|
435 // Same as Space::is_in_reserved, but will use the original size of the region. |
5e5d4821bf07
7097516: G1: assert(0<= from_card && from_card<HeapRegion::CardsPerRegion) failed: Must be in range.
brutisso
parents:
3986
diff
changeset
|
436 // The original size is different only for start humongous regions. They get |
5e5d4821bf07
7097516: G1: assert(0<= from_card && from_card<HeapRegion::CardsPerRegion) failed: Must be in range.
brutisso
parents:
3986
diff
changeset
|
437 // their _end set up to be the end of the last continues region of the |
5e5d4821bf07
7097516: G1: assert(0<= from_card && from_card<HeapRegion::CardsPerRegion) failed: Must be in range.
brutisso
parents:
3986
diff
changeset
|
438 // corresponding humongous object. |
5e5d4821bf07
7097516: G1: assert(0<= from_card && from_card<HeapRegion::CardsPerRegion) failed: Must be in range.
brutisso
parents:
3986
diff
changeset
|
439 bool is_in_reserved_raw(const void* p) const { |
5e5d4821bf07
7097516: G1: assert(0<= from_card && from_card<HeapRegion::CardsPerRegion) failed: Must be in range.
brutisso
parents:
3986
diff
changeset
|
440 return _bottom <= p && p < _orig_end; |
5e5d4821bf07
7097516: G1: assert(0<= from_card && from_card<HeapRegion::CardsPerRegion) failed: Must be in range.
brutisso
parents:
3986
diff
changeset
|
441 } |
5e5d4821bf07
7097516: G1: assert(0<= from_card && from_card<HeapRegion::CardsPerRegion) failed: Must be in range.
brutisso
parents:
3986
diff
changeset
|
442 |
2133
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
443 // Makes the current region be a "starts humongous" region, i.e., |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
444 // the first region in a series of one or more contiguous regions |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
445 // that will contain a single "humongous" object. The two parameters |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
446 // are as follows: |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
447 // |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
448 // new_top : The new value of the top field of this region which |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
449 // points to the end of the humongous object that's being |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
450 // allocated. If there is more than one region in the series, top |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
451 // will lie beyond this region's original end field and on the last |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
452 // region in the series. |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
453 // |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
454 // new_end : The new value of the end field of this region which |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
455 // points to the end of the last region in the series. If there is |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
456 // one region in the series (namely: this one) end will be the same |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
457 // as the original end of this region. |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
458 // |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
459 // Updating top and end as described above makes this region look as |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
460 // if it spans the entire space taken up by all the regions in the |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
461 // series and an single allocation moved its top to new_top. This |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
462 // ensures that the space (capacity / allocated) taken up by all |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
463 // humongous regions can be calculated by just looking at the |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
464 // "starts humongous" regions and by ignoring the "continues |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
465 // humongous" regions. |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
466 void set_startsHumongous(HeapWord* new_top, HeapWord* new_end); |
342 | 467 |
2133
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
468 // Makes the current region be a "continues humongous' |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
469 // region. first_hr is the "start humongous" region of the series |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
470 // which this region will be part of. |
2250ee17e258
7007068: G1: refine the BOT during evac failure handling
tonyp
parents:
1972
diff
changeset
|
471 void set_continuesHumongous(HeapRegion* first_hr); |
342 | 472 |
2152 | 473 // Unsets the humongous-related fields on the region. |
474 void set_notHumongous(); | |
475 | |
342 | 476 // If the region has a remembered set, return a pointer to it. |
477 HeapRegionRemSet* rem_set() const { | |
478 return _rem_set; | |
479 } | |
480 | |
481 // True iff the region is in current collection_set. | |
482 bool in_collection_set() const { | |
483 return _in_collection_set; | |
484 } | |
485 void set_in_collection_set(bool b) { | |
486 _in_collection_set = b; | |
487 } | |
488 HeapRegion* next_in_collection_set() { | |
489 assert(in_collection_set(), "should only invoke on member of CS."); | |
490 assert(_next_in_special_set == NULL || | |
491 _next_in_special_set->in_collection_set(), | |
492 "Malformed CS."); | |
493 return _next_in_special_set; | |
494 } | |
495 void set_next_in_collection_set(HeapRegion* r) { | |
496 assert(in_collection_set(), "should only invoke on member of CS."); | |
497 assert(r == NULL || r->in_collection_set(), "Malformed CS."); | |
498 _next_in_special_set = r; | |
499 } | |
500 | |
2152 | 501 // Methods used by the HeapRegionSetBase class and subclasses. |
342 | 502 |
2152 | 503 // Getter and setter for the next field used to link regions into |
504 // linked lists. | |
505 HeapRegion* next() { return _next; } | |
506 | |
507 void set_next(HeapRegion* next) { _next = next; } | |
342 | 508 |
2152 | 509 // Every region added to a set is tagged with a reference to that |
510 // set. This is used for doing consistency checking to make sure that | |
511 // the contents of a set are as they should be and it's only | |
512 // available in non-product builds. | |
513 #ifdef ASSERT | |
514 void set_containing_set(HeapRegionSetBase* containing_set) { | |
515 assert((containing_set == NULL && _containing_set != NULL) || | |
516 (containing_set != NULL && _containing_set == NULL), | |
517 err_msg("containing_set: "PTR_FORMAT" " | |
518 "_containing_set: "PTR_FORMAT, | |
519 containing_set, _containing_set)); | |
520 | |
521 _containing_set = containing_set; | |
2361 | 522 } |
342 | 523 |
2152 | 524 HeapRegionSetBase* containing_set() { return _containing_set; } |
525 #else // ASSERT | |
526 void set_containing_set(HeapRegionSetBase* containing_set) { } | |
342 | 527 |
2361 | 528 // containing_set() is only used in asserts so there's no reason |
2152 | 529 // to provide a dummy version of it. |
530 #endif // ASSERT | |
342 | 531 |
2152 | 532 // If we want to remove regions from a list in bulk we can simply tag |
533 // them with the pending_removal tag and call the | |
534 // remove_all_pending() method on the list. | |
342 | 535 |
2152 | 536 bool pending_removal() { return _pending_removal; } |
537 | |
538 void set_pending_removal(bool pending_removal) { | |
2361 | 539 if (pending_removal) { |
540 assert(!_pending_removal && containing_set() != NULL, | |
541 "can only set pending removal to true if it's false and " | |
542 "the region belongs to a region set"); | |
543 } else { | |
544 assert( _pending_removal && containing_set() == NULL, | |
545 "can only set pending removal to false if it's true and " | |
546 "the region does not belong to a region set"); | |
547 } | |
2152 | 548 |
549 _pending_removal = pending_removal; | |
342 | 550 } |
551 | |
552 HeapRegion* get_next_young_region() { return _next_young_region; } | |
553 void set_next_young_region(HeapRegion* hr) { | |
554 _next_young_region = hr; | |
555 } | |
556 | |
796
29e7d79232b9
6819065: G1: eliminate high serial card table clearing time
apetrusenko
parents:
677
diff
changeset
|
557 HeapRegion* get_next_dirty_cards_region() const { return _next_dirty_cards_region; } |
29e7d79232b9
6819065: G1: eliminate high serial card table clearing time
apetrusenko
parents:
677
diff
changeset
|
558 HeapRegion** next_dirty_cards_region_addr() { return &_next_dirty_cards_region; } |
29e7d79232b9
6819065: G1: eliminate high serial card table clearing time
apetrusenko
parents:
677
diff
changeset
|
559 void set_next_dirty_cards_region(HeapRegion* hr) { _next_dirty_cards_region = hr; } |
29e7d79232b9
6819065: G1: eliminate high serial card table clearing time
apetrusenko
parents:
677
diff
changeset
|
560 bool is_on_dirty_cards_region_list() const { return get_next_dirty_cards_region() != NULL; } |
29e7d79232b9
6819065: G1: eliminate high serial card table clearing time
apetrusenko
parents:
677
diff
changeset
|
561 |
3766 | 562 HeapWord* orig_end() { return _orig_end; } |
563 | |
342 | 564 // Allows logical separation between objects allocated before and after. |
565 void save_marks(); | |
566 | |
567 // Reset HR stuff to default values. | |
568 void hr_clear(bool par, bool clear_space); | |
3317
063382f9b575
7035144: G1: nightly failure: Non-dirty cards in region that should be dirty (failures still exist...)
tonyp
parents:
2435
diff
changeset
|
569 void par_clear(); |
342 | 570 |
356 | 571 void initialize(MemRegion mr, bool clear_space, bool mangle_space); |
342 | 572 |
573 // Get the start of the unmarked area in this region. | |
574 HeapWord* prev_top_at_mark_start() const { return _prev_top_at_mark_start; } | |
575 HeapWord* next_top_at_mark_start() const { return _next_top_at_mark_start; } | |
576 | |
577 // Apply "cl->do_oop" to (the addresses of) all reference fields in objects | |
578 // allocated in the current region before the last call to "save_mark". | |
579 void oop_before_save_marks_iterate(OopClosure* cl); | |
580 | |
581 // Note the start or end of marking. This tells the heap region | |
582 // that the collector is about to start or has finished (concurrently) | |
583 // marking the heap. | |
584 | |
585 // Note the start of a marking phase. Record the | |
586 // start of the unmarked area of the region here. | |
587 void note_start_of_marking(bool during_initial_mark) { | |
588 init_top_at_conc_mark_count(); | |
589 _next_marked_bytes = 0; | |
590 if (during_initial_mark && is_young() && !is_survivor()) | |
591 _next_top_at_mark_start = bottom(); | |
592 else | |
593 _next_top_at_mark_start = top(); | |
594 } | |
595 | |
596 // Note the end of a marking phase. Install the start of | |
597 // the unmarked area that was captured at start of marking. | |
598 void note_end_of_marking() { | |
599 _prev_top_at_mark_start = _next_top_at_mark_start; | |
600 _prev_marked_bytes = _next_marked_bytes; | |
601 _next_marked_bytes = 0; | |
602 | |
603 guarantee(_prev_marked_bytes <= | |
604 (size_t) (prev_top_at_mark_start() - bottom()) * HeapWordSize, | |
605 "invariant"); | |
606 } | |
607 | |
608 // After an evacuation, we need to update _next_top_at_mark_start | |
609 // to be the current top. Note this is only valid if we have only | |
610 // ever evacuated into this region. If we evacuate, allocate, and | |
611 // then evacuate we are in deep doodoo. | |
612 void note_end_of_copying() { | |
1021
1f19207eefc2
6847956: G1: crash in oopDesc*G1ParCopyHelper::copy_to_survivor_space(oopDesc*)
tonyp
parents:
1020
diff
changeset
|
613 assert(top() >= _next_top_at_mark_start, "Increase only"); |
1f19207eefc2
6847956: G1: crash in oopDesc*G1ParCopyHelper::copy_to_survivor_space(oopDesc*)
tonyp
parents:
1020
diff
changeset
|
614 _next_top_at_mark_start = top(); |
342 | 615 } |
616 | |
617 // Returns "false" iff no object in the region was allocated when the | |
618 // last mark phase ended. | |
619 bool is_marked() { return _prev_top_at_mark_start != bottom(); } | |
620 | |
621 // If "is_marked()" is true, then this is the index of the region in | |
622 // an array constructed at the end of marking of the regions in a | |
623 // "desirability" order. | |
624 int sort_index() { | |
625 return _sort_index; | |
626 } | |
627 void set_sort_index(int i) { | |
628 _sort_index = i; | |
629 } | |
630 | |
631 void init_top_at_conc_mark_count() { | |
632 _top_at_conc_mark_count = bottom(); | |
633 } | |
634 | |
635 void set_top_at_conc_mark_count(HeapWord *cur) { | |
636 assert(bottom() <= cur && cur <= end(), "Sanity."); | |
637 _top_at_conc_mark_count = cur; | |
638 } | |
639 | |
640 HeapWord* top_at_conc_mark_count() { | |
641 return _top_at_conc_mark_count; | |
642 } | |
643 | |
644 void reset_during_compaction() { | |
645 guarantee( isHumongous() && startsHumongous(), | |
646 "should only be called for humongous regions"); | |
647 | |
648 zero_marked_bytes(); | |
649 init_top_at_mark_start(); | |
650 } | |
651 | |
652 // <PREDICTION> | |
653 void calc_gc_efficiency(void); | |
654 double gc_efficiency() { return _gc_efficiency;} | |
655 // </PREDICTION> | |
656 | |
657 bool is_young() const { return _young_type != NotYoung; } | |
658 bool is_survivor() const { return _young_type == Survivor; } | |
659 | |
660 int young_index_in_cset() const { return _young_index_in_cset; } | |
661 void set_young_index_in_cset(int index) { | |
662 assert( (index == -1) || is_young(), "pre-condition" ); | |
663 _young_index_in_cset = index; | |
664 } | |
665 | |
666 int age_in_surv_rate_group() { | |
667 assert( _surv_rate_group != NULL, "pre-condition" ); | |
668 assert( _age_index > -1, "pre-condition" ); | |
669 return _surv_rate_group->age_in_group(_age_index); | |
670 } | |
671 | |
672 void record_surv_words_in_group(size_t words_survived) { | |
673 assert( _surv_rate_group != NULL, "pre-condition" ); | |
674 assert( _age_index > -1, "pre-condition" ); | |
675 int age_in_group = age_in_surv_rate_group(); | |
676 _surv_rate_group->record_surviving_words(age_in_group, words_survived); | |
677 } | |
678 | |
679 int age_in_surv_rate_group_cond() { | |
680 if (_surv_rate_group != NULL) | |
681 return age_in_surv_rate_group(); | |
682 else | |
683 return -1; | |
684 } | |
685 | |
686 SurvRateGroup* surv_rate_group() { | |
687 return _surv_rate_group; | |
688 } | |
689 | |
690 void install_surv_rate_group(SurvRateGroup* surv_rate_group) { | |
691 assert( surv_rate_group != NULL, "pre-condition" ); | |
692 assert( _surv_rate_group == NULL, "pre-condition" ); | |
693 assert( is_young(), "pre-condition" ); | |
694 | |
695 _surv_rate_group = surv_rate_group; | |
696 _age_index = surv_rate_group->next_age_index(); | |
697 } | |
698 | |
699 void uninstall_surv_rate_group() { | |
700 if (_surv_rate_group != NULL) { | |
701 assert( _age_index > -1, "pre-condition" ); | |
702 assert( is_young(), "pre-condition" ); | |
703 | |
704 _surv_rate_group = NULL; | |
705 _age_index = -1; | |
706 } else { | |
707 assert( _age_index == -1, "pre-condition" ); | |
708 } | |
709 } | |
710 | |
711 void set_young() { set_young_type(Young); } | |
712 | |
713 void set_survivor() { set_young_type(Survivor); } | |
714 | |
715 void set_not_young() { set_young_type(NotYoung); } | |
716 | |
717 // Determine if an object has been allocated since the last | |
718 // mark performed by the collector. This returns true iff the object | |
719 // is within the unmarked area of the region. | |
720 bool obj_allocated_since_prev_marking(oop obj) const { | |
721 return (HeapWord *) obj >= prev_top_at_mark_start(); | |
722 } | |
723 bool obj_allocated_since_next_marking(oop obj) const { | |
724 return (HeapWord *) obj >= next_top_at_mark_start(); | |
725 } | |
726 | |
727 // For parallel heapRegion traversal. | |
728 bool claimHeapRegion(int claimValue); | |
729 jint claim_value() { return _claimed; } | |
730 // Use this carefully: only when you're sure no one is claiming... | |
731 void set_claim_value(int claimValue) { _claimed = claimValue; } | |
732 | |
733 // Returns the "evacuation_failed" property of the region. | |
734 bool evacuation_failed() { return _evacuation_failed; } | |
735 | |
736 // Sets the "evacuation_failed" property of the region. | |
737 void set_evacuation_failed(bool b) { | |
738 _evacuation_failed = b; | |
739 | |
740 if (b) { | |
741 init_top_at_conc_mark_count(); | |
742 _next_marked_bytes = 0; | |
743 } | |
744 } | |
745 | |
746 // Requires that "mr" be entirely within the region. | |
747 // Apply "cl->do_object" to all objects that intersect with "mr". | |
748 // If the iteration encounters an unparseable portion of the region, | |
749 // or if "cl->abort()" is true after a closure application, | |
750 // terminate the iteration and return the address of the start of the | |
751 // subregion that isn't done. (The two can be distinguished by querying | |
752 // "cl->abort()".) Return of "NULL" indicates that the iteration | |
753 // completed. | |
754 HeapWord* | |
755 object_iterate_mem_careful(MemRegion mr, ObjectClosure* cl); | |
756 | |
3317
063382f9b575
7035144: G1: nightly failure: Non-dirty cards in region that should be dirty (failures still exist...)
tonyp
parents:
2435
diff
changeset
|
757 // filter_young: if true and the region is a young region then we |
063382f9b575
7035144: G1: nightly failure: Non-dirty cards in region that should be dirty (failures still exist...)
tonyp
parents:
2435
diff
changeset
|
758 // skip the iteration. |
063382f9b575
7035144: G1: nightly failure: Non-dirty cards in region that should be dirty (failures still exist...)
tonyp
parents:
2435
diff
changeset
|
759 // card_ptr: if not NULL, and we decide that the card is not young |
063382f9b575
7035144: G1: nightly failure: Non-dirty cards in region that should be dirty (failures still exist...)
tonyp
parents:
2435
diff
changeset
|
760 // and we iterate over it, we'll clean the card before we start the |
063382f9b575
7035144: G1: nightly failure: Non-dirty cards in region that should be dirty (failures still exist...)
tonyp
parents:
2435
diff
changeset
|
761 // iteration. |
342 | 762 HeapWord* |
763 oops_on_card_seq_iterate_careful(MemRegion mr, | |
1666
5cbac8938c4c
6956639: G1: assert(cached_ptr != card_ptr) failed: shouldn't be, concurrentG1Refine.cpp:307
johnc
parents:
1552
diff
changeset
|
764 FilterOutOfRegionClosure* cl, |
3317
063382f9b575
7035144: G1: nightly failure: Non-dirty cards in region that should be dirty (failures still exist...)
tonyp
parents:
2435
diff
changeset
|
765 bool filter_young, |
063382f9b575
7035144: G1: nightly failure: Non-dirty cards in region that should be dirty (failures still exist...)
tonyp
parents:
2435
diff
changeset
|
766 jbyte* card_ptr); |
342 | 767 |
768 // A version of block start that is guaranteed to find *some* block | |
769 // boundary at or before "p", but does not object iteration, and may | |
770 // therefore be used safely when the heap is unparseable. | |
771 HeapWord* block_start_careful(const void* p) const { | |
772 return _offsets.block_start_careful(p); | |
773 } | |
774 | |
775 // Requires that "addr" is within the region. Returns the start of the | |
776 // first ("careful") block that starts at or after "addr", or else the | |
777 // "end" of the region if there is no such block. | |
778 HeapWord* next_block_start_careful(HeapWord* addr); | |
779 | |
1394
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
780 size_t recorded_rs_length() const { return _recorded_rs_length; } |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
781 double predicted_elapsed_time_ms() const { return _predicted_elapsed_time_ms; } |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
782 size_t predicted_bytes_to_copy() const { return _predicted_bytes_to_copy; } |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
783 |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
784 void set_recorded_rs_length(size_t rs_length) { |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
785 _recorded_rs_length = rs_length; |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
786 } |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
787 |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
788 void set_predicted_elapsed_time_ms(double ms) { |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
789 _predicted_elapsed_time_ms = ms; |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
790 } |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
791 |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
792 void set_predicted_bytes_to_copy(size_t bytes) { |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
793 _predicted_bytes_to_copy = bytes; |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
794 } |
1316cec51b4d
6819061: G1: eliminate serial Other times that are proportional to the collection set length
johnc
parents:
1021
diff
changeset
|
795 |
342 | 796 #define HeapRegion_OOP_SINCE_SAVE_MARKS_DECL(OopClosureType, nv_suffix) \ |
797 virtual void oop_since_save_marks_iterate##nv_suffix(OopClosureType* cl); | |
798 SPECIALIZED_SINCE_SAVE_MARKS_CLOSURES(HeapRegion_OOP_SINCE_SAVE_MARKS_DECL) | |
799 | |
800 CompactibleSpace* next_compaction_space() const; | |
801 | |
802 virtual void reset_after_compaction(); | |
803 | |
804 void print() const; | |
805 void print_on(outputStream* st) const; | |
806 | |
3772
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
807 // vo == UsePrevMarking -> use "prev" marking information, |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
808 // vo == UseNextMarking -> use "next" marking information |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
809 // vo == UseMarkWord -> use the mark word in the object header |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
810 // |
811 | 811 // NOTE: Only the "prev" marking information is guaranteed to be |
812 // consistent most of the time, so most calls to this should use | |
3772
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
813 // vo == UsePrevMarking. |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
814 // Currently, there is only one case where this is called with |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
815 // vo == UseNextMarking, which is to verify the "next" marking |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
816 // information at the end of remark. |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
817 // Currently there is only one place where this is called with |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
818 // vo == UseMarkWord, which is to verify the marking during a |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
819 // full GC. |
6747fd0512e0
7004681: G1: Extend marking verification to Full GCs
johnc
parents:
3766
diff
changeset
|
820 void verify(bool allow_dirty, VerifyOption vo, bool *failures) const; |
811 | 821 |
822 // Override; it uses the "prev" marking information | |
342 | 823 virtual void verify(bool allow_dirty) const; |
824 }; | |
825 | |
826 // HeapRegionClosure is used for iterating over regions. | |
827 // Terminates the iteration when the "doHeapRegion" method returns "true". | |
828 class HeapRegionClosure : public StackObj { | |
829 friend class HeapRegionSeq; | |
830 friend class G1CollectedHeap; | |
831 | |
832 bool _complete; | |
833 void incomplete() { _complete = false; } | |
834 | |
835 public: | |
836 HeapRegionClosure(): _complete(true) {} | |
837 | |
838 // Typically called on each region until it returns true. | |
839 virtual bool doHeapRegion(HeapRegion* r) = 0; | |
840 | |
841 // True after iteration if the closure was applied to all heap regions | |
842 // and returned "false" in all cases. | |
843 bool complete() { return _complete; } | |
844 }; | |
845 | |
846 #endif // SERIALGC | |
1972 | 847 |
848 #endif // SHARE_VM_GC_IMPLEMENTATION_G1_HEAPREGION_HPP |