comparison src/share/vm/gc_implementation/concurrentMarkSweep/adaptiveFreeList.hpp @ 6885:685df3c6f84b

7045397: NPG: Add freelists to class loader arenas. Reviewed-by: coleenp, stefank, jprovino, ohair
author jmasa
date Tue, 18 Sep 2012 23:35:42 -0700
parents
children 7c5a1b62f53d
comparison
equal deleted inserted replaced
6877:d0e7716b179e 6885:685df3c6f84b
1 /*
2 * Copyright (c) 2001, 2010, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #ifndef SHARE_VM_MEMORY_ADAPTIVEFREELIST_HPP
26 #define SHARE_VM_MEMORY_ADAPTIVEFREELIST_HPP
27
28 #include "memory/freeList.hpp"
29 #include "gc_implementation/shared/allocationStats.hpp"
30
31 class CompactibleFreeListSpace;
32
33 // A class for maintaining a free list of Chunk's. The FreeList
34 // maintains a the structure of the list (head, tail, etc.) plus
35 // statistics for allocations from the list. The links between items
36 // are not part of FreeList. The statistics are
37 // used to make decisions about coalescing Chunk's when they
38 // are swept during collection.
39 //
40 // See the corresponding .cpp file for a description of the specifics
41 // for that implementation.
42
43 class Mutex;
44
45 template <class Chunk>
46 class AdaptiveFreeList : public FreeList<Chunk> {
47 friend class CompactibleFreeListSpace;
48 friend class VMStructs;
49 // friend class PrintTreeCensusClosure<Chunk, FreeList_t>;
50
51 size_t _hint; // next larger size list with a positive surplus
52
53 AllocationStats _allocation_stats; // allocation-related statistics
54
55 public:
56
57 AdaptiveFreeList();
58 AdaptiveFreeList(Chunk* fc);
59
60 using FreeList<Chunk>::assert_proper_lock_protection;
61 #ifdef ASSERT
62 using FreeList<Chunk>::protecting_lock;
63 #endif
64 using FreeList<Chunk>::count;
65 using FreeList<Chunk>::size;
66 using FreeList<Chunk>::verify_chunk_in_free_list;
67 using FreeList<Chunk>::getFirstNChunksFromList;
68 using FreeList<Chunk>::print_on;
69 void return_chunk_at_head(Chunk* fc, bool record_return);
70 void return_chunk_at_head(Chunk* fc);
71 void return_chunk_at_tail(Chunk* fc, bool record_return);
72 void return_chunk_at_tail(Chunk* fc);
73 using FreeList<Chunk>::return_chunk_at_tail;
74 using FreeList<Chunk>::remove_chunk;
75 using FreeList<Chunk>::prepend;
76 using FreeList<Chunk>::print_labels_on;
77 using FreeList<Chunk>::get_chunk_at_head;
78
79 // Initialize.
80 void initialize();
81
82 // Reset the head, tail, hint, and count of a free list.
83 void reset(size_t hint);
84
85 void assert_proper_lock_protection_work() const PRODUCT_RETURN;
86
87 void print_on(outputStream* st, const char* c = NULL) const;
88
89 size_t hint() const {
90 return _hint;
91 }
92 void set_hint(size_t v) {
93 assert_proper_lock_protection();
94 assert(v == 0 || size() < v, "Bad hint");
95 _hint = v;
96 }
97
98 size_t get_better_size();
99
100 // Accessors for statistics
101 void init_statistics(bool split_birth = false);
102
103 AllocationStats* allocation_stats() {
104 assert_proper_lock_protection();
105 return &_allocation_stats;
106 }
107
108 ssize_t desired() const {
109 return _allocation_stats.desired();
110 }
111 void set_desired(ssize_t v) {
112 assert_proper_lock_protection();
113 _allocation_stats.set_desired(v);
114 }
115 void compute_desired(float inter_sweep_current,
116 float inter_sweep_estimate,
117 float intra_sweep_estimate) {
118 assert_proper_lock_protection();
119 _allocation_stats.compute_desired(count(),
120 inter_sweep_current,
121 inter_sweep_estimate,
122 intra_sweep_estimate);
123 }
124 ssize_t coal_desired() const {
125 return _allocation_stats.coal_desired();
126 }
127 void set_coal_desired(ssize_t v) {
128 assert_proper_lock_protection();
129 _allocation_stats.set_coal_desired(v);
130 }
131
132 ssize_t surplus() const {
133 return _allocation_stats.surplus();
134 }
135 void set_surplus(ssize_t v) {
136 assert_proper_lock_protection();
137 _allocation_stats.set_surplus(v);
138 }
139 void increment_surplus() {
140 assert_proper_lock_protection();
141 _allocation_stats.increment_surplus();
142 }
143 void decrement_surplus() {
144 assert_proper_lock_protection();
145 _allocation_stats.decrement_surplus();
146 }
147
148 ssize_t bfr_surp() const {
149 return _allocation_stats.bfr_surp();
150 }
151 void set_bfr_surp(ssize_t v) {
152 assert_proper_lock_protection();
153 _allocation_stats.set_bfr_surp(v);
154 }
155 ssize_t prev_sweep() const {
156 return _allocation_stats.prev_sweep();
157 }
158 void set_prev_sweep(ssize_t v) {
159 assert_proper_lock_protection();
160 _allocation_stats.set_prev_sweep(v);
161 }
162 ssize_t before_sweep() const {
163 return _allocation_stats.before_sweep();
164 }
165 void set_before_sweep(ssize_t v) {
166 assert_proper_lock_protection();
167 _allocation_stats.set_before_sweep(v);
168 }
169
170 ssize_t coal_births() const {
171 return _allocation_stats.coal_births();
172 }
173 void set_coal_births(ssize_t v) {
174 assert_proper_lock_protection();
175 _allocation_stats.set_coal_births(v);
176 }
177 void increment_coal_births() {
178 assert_proper_lock_protection();
179 _allocation_stats.increment_coal_births();
180 }
181
182 ssize_t coal_deaths() const {
183 return _allocation_stats.coal_deaths();
184 }
185 void set_coal_deaths(ssize_t v) {
186 assert_proper_lock_protection();
187 _allocation_stats.set_coal_deaths(v);
188 }
189 void increment_coal_deaths() {
190 assert_proper_lock_protection();
191 _allocation_stats.increment_coal_deaths();
192 }
193
194 ssize_t split_births() const {
195 return _allocation_stats.split_births();
196 }
197 void set_split_births(ssize_t v) {
198 assert_proper_lock_protection();
199 _allocation_stats.set_split_births(v);
200 }
201 void increment_split_births() {
202 assert_proper_lock_protection();
203 _allocation_stats.increment_split_births();
204 }
205
206 ssize_t split_deaths() const {
207 return _allocation_stats.split_deaths();
208 }
209 void set_split_deaths(ssize_t v) {
210 assert_proper_lock_protection();
211 _allocation_stats.set_split_deaths(v);
212 }
213 void increment_split_deaths() {
214 assert_proper_lock_protection();
215 _allocation_stats.increment_split_deaths();
216 }
217
218 #ifndef PRODUCT
219 // For debugging. The "_returned_bytes" in all the lists are summed
220 // and compared with the total number of bytes swept during a
221 // collection.
222 size_t returned_bytes() const { return _allocation_stats.returned_bytes(); }
223 void set_returned_bytes(size_t v) { _allocation_stats.set_returned_bytes(v); }
224 void increment_returned_bytes_by(size_t v) {
225 _allocation_stats.set_returned_bytes(_allocation_stats.returned_bytes() + v);
226 }
227 // Stats verification
228 void verify_stats() const;
229 #endif // NOT PRODUCT
230 };
231
232 #endif // SHARE_VM_MEMORY_ADAPTIVEFREELIST_HPP