Mercurial > hg > truffle
annotate src/share/vm/gc_interface/collectedHeap.inline.hpp @ 3543:344264424174
Enable debugging of compiler code from Java IDEs. Moved binary to "graal" subdirectory. New flag for starting Graal is just "-graal".
author | Thomas Wuerthinger <thomas@wuerthinger.net> |
---|---|
date | Sat, 13 Aug 2011 18:08:28 +0200 |
parents | b1a2afa37ec4 |
children | c9ca3f51cf41 |
rev | line source |
---|---|
0 | 1 /* |
2100
b1a2afa37ec4
7003271: Hotspot should track cumulative Java heap bytes allocated on a per-thread basis
phh
parents:
1972
diff
changeset
|
2 * Copyright (c) 2001, 2011, Oracle and/or its affiliates. All rights reserved. |
0 | 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 * | |
5 * This code is free software; you can redistribute it and/or modify it | |
6 * under the terms of the GNU General Public License version 2 only, as | |
7 * published by the Free Software Foundation. | |
8 * | |
9 * This code is distributed in the hope that it will be useful, but WITHOUT | |
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License | |
12 * version 2 for more details (a copy is included in the LICENSE file that | |
13 * accompanied this code). | |
14 * | |
15 * You should have received a copy of the GNU General Public License version | |
16 * 2 along with this work; if not, write to the Free Software Foundation, | |
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | |
18 * | |
1552
c18cbe5936b8
6941466: Oracle rebranding changes for Hotspot repositories
trims
parents:
579
diff
changeset
|
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
c18cbe5936b8
6941466: Oracle rebranding changes for Hotspot repositories
trims
parents:
579
diff
changeset
|
20 * or visit www.oracle.com if you need additional information or have any |
c18cbe5936b8
6941466: Oracle rebranding changes for Hotspot repositories
trims
parents:
579
diff
changeset
|
21 * questions. |
0 | 22 * |
23 */ | |
24 | |
1972 | 25 #ifndef SHARE_VM_GC_INTERFACE_COLLECTEDHEAP_INLINE_HPP |
26 #define SHARE_VM_GC_INTERFACE_COLLECTEDHEAP_INLINE_HPP | |
27 | |
28 #include "gc_interface/collectedHeap.hpp" | |
29 #include "memory/threadLocalAllocBuffer.inline.hpp" | |
30 #include "memory/universe.hpp" | |
31 #include "oops/arrayOop.hpp" | |
32 #include "prims/jvmtiExport.hpp" | |
33 #include "runtime/sharedRuntime.hpp" | |
34 #include "runtime/thread.hpp" | |
35 #include "services/lowMemoryDetector.hpp" | |
36 #include "utilities/copy.hpp" | |
37 #ifdef TARGET_OS_FAMILY_linux | |
38 # include "thread_linux.inline.hpp" | |
39 #endif | |
40 #ifdef TARGET_OS_FAMILY_solaris | |
41 # include "thread_solaris.inline.hpp" | |
42 #endif | |
43 #ifdef TARGET_OS_FAMILY_windows | |
44 # include "thread_windows.inline.hpp" | |
45 #endif | |
46 | |
0 | 47 // Inline allocation implementations. |
48 | |
49 void CollectedHeap::post_allocation_setup_common(KlassHandle klass, | |
50 HeapWord* obj, | |
51 size_t size) { | |
52 post_allocation_setup_no_klass_install(klass, obj, size); | |
53 post_allocation_install_obj_klass(klass, oop(obj), (int) size); | |
54 } | |
55 | |
56 void CollectedHeap::post_allocation_setup_no_klass_install(KlassHandle klass, | |
57 HeapWord* objPtr, | |
58 size_t size) { | |
59 oop obj = (oop)objPtr; | |
60 | |
61 assert(obj != NULL, "NULL object pointer"); | |
62 if (UseBiasedLocking && (klass() != NULL)) { | |
63 obj->set_mark(klass->prototype_header()); | |
64 } else { | |
65 // May be bootstrapping | |
66 obj->set_mark(markOopDesc::prototype()); | |
67 } | |
68 } | |
69 | |
70 void CollectedHeap::post_allocation_install_obj_klass(KlassHandle klass, | |
71 oop obj, | |
72 int size) { | |
73 // These asserts are kind of complicated because of klassKlass | |
74 // and the beginning of the world. | |
75 assert(klass() != NULL || !Universe::is_fully_initialized(), "NULL klass"); | |
76 assert(klass() == NULL || klass()->is_klass(), "not a klass"); | |
77 assert(klass() == NULL || klass()->klass_part() != NULL, "not a klass"); | |
78 assert(obj != NULL, "NULL object pointer"); | |
79 obj->set_klass(klass()); | |
80 assert(!Universe::is_fully_initialized() || obj->blueprint() != NULL, | |
81 "missing blueprint"); | |
113
ba764ed4b6f2
6420645: Create a vm that uses compressed oops for up to 32gb heapsizes
coleenp
parents:
0
diff
changeset
|
82 } |
0 | 83 |
113
ba764ed4b6f2
6420645: Create a vm that uses compressed oops for up to 32gb heapsizes
coleenp
parents:
0
diff
changeset
|
84 // Support for jvmti and dtrace |
ba764ed4b6f2
6420645: Create a vm that uses compressed oops for up to 32gb heapsizes
coleenp
parents:
0
diff
changeset
|
85 inline void post_allocation_notify(KlassHandle klass, oop obj) { |
481
7d7a7c599c17
6578152: fill_region_with_object has usability and safety issues
jcoomes
parents:
356
diff
changeset
|
86 // support low memory notifications (no-op if not enabled) |
7d7a7c599c17
6578152: fill_region_with_object has usability and safety issues
jcoomes
parents:
356
diff
changeset
|
87 LowMemoryDetector::detect_low_memory_for_collected_pools(); |
7d7a7c599c17
6578152: fill_region_with_object has usability and safety issues
jcoomes
parents:
356
diff
changeset
|
88 |
0 | 89 // support for JVMTI VMObjectAlloc event (no-op if not enabled) |
90 JvmtiExport::vm_object_alloc_event_collector(obj); | |
91 | |
92 if (DTraceAllocProbes) { | |
93 // support for Dtrace object alloc event (no-op most of the time) | |
94 if (klass() != NULL && klass()->klass_part()->name() != NULL) { | |
95 SharedRuntime::dtrace_object_alloc(obj); | |
96 } | |
97 } | |
98 } | |
99 | |
100 void CollectedHeap::post_allocation_setup_obj(KlassHandle klass, | |
101 HeapWord* obj, | |
102 size_t size) { | |
103 post_allocation_setup_common(klass, obj, size); | |
104 assert(Universe::is_bootstrapping() || | |
105 !((oop)obj)->blueprint()->oop_is_array(), "must not be an array"); | |
113
ba764ed4b6f2
6420645: Create a vm that uses compressed oops for up to 32gb heapsizes
coleenp
parents:
0
diff
changeset
|
106 // notify jvmti and dtrace |
ba764ed4b6f2
6420645: Create a vm that uses compressed oops for up to 32gb heapsizes
coleenp
parents:
0
diff
changeset
|
107 post_allocation_notify(klass, (oop)obj); |
0 | 108 } |
109 | |
110 void CollectedHeap::post_allocation_setup_array(KlassHandle klass, | |
111 HeapWord* obj, | |
112 size_t size, | |
113 int length) { | |
167
feeb96a45707
6696264: assert("narrow oop can never be zero") for GCBasher & ParNewGC
coleenp
parents:
113
diff
changeset
|
114 // Set array length before setting the _klass field |
feeb96a45707
6696264: assert("narrow oop can never be zero") for GCBasher & ParNewGC
coleenp
parents:
113
diff
changeset
|
115 // in post_allocation_setup_common() because the klass field |
feeb96a45707
6696264: assert("narrow oop can never be zero") for GCBasher & ParNewGC
coleenp
parents:
113
diff
changeset
|
116 // indicates that the object is parsable by concurrent GC. |
0 | 117 assert(length >= 0, "length should be non-negative"); |
167
feeb96a45707
6696264: assert("narrow oop can never be zero") for GCBasher & ParNewGC
coleenp
parents:
113
diff
changeset
|
118 ((arrayOop)obj)->set_length(length); |
113
ba764ed4b6f2
6420645: Create a vm that uses compressed oops for up to 32gb heapsizes
coleenp
parents:
0
diff
changeset
|
119 post_allocation_setup_common(klass, obj, size); |
0 | 120 assert(((oop)obj)->blueprint()->oop_is_array(), "must be an array"); |
113
ba764ed4b6f2
6420645: Create a vm that uses compressed oops for up to 32gb heapsizes
coleenp
parents:
0
diff
changeset
|
121 // notify jvmti and dtrace (must be after length is set for dtrace) |
ba764ed4b6f2
6420645: Create a vm that uses compressed oops for up to 32gb heapsizes
coleenp
parents:
0
diff
changeset
|
122 post_allocation_notify(klass, (oop)obj); |
0 | 123 } |
124 | |
125 HeapWord* CollectedHeap::common_mem_allocate_noinit(size_t size, bool is_noref, TRAPS) { | |
126 | |
127 // Clear unhandled oops for memory allocation. Memory allocation might | |
128 // not take out a lock if from tlab, so clear here. | |
129 CHECK_UNHANDLED_OOPS_ONLY(THREAD->clear_unhandled_oops();) | |
130 | |
131 if (HAS_PENDING_EXCEPTION) { | |
132 NOT_PRODUCT(guarantee(false, "Should not allocate with exception pending")); | |
133 return NULL; // caller does a CHECK_0 too | |
134 } | |
135 | |
136 // We may want to update this, is_noref objects might not be allocated in TLABs. | |
137 HeapWord* result = NULL; | |
138 if (UseTLAB) { | |
139 result = CollectedHeap::allocate_from_tlab(THREAD, size); | |
140 if (result != NULL) { | |
141 assert(!HAS_PENDING_EXCEPTION, | |
142 "Unexpected exception, will result in uninitialized storage"); | |
143 return result; | |
144 } | |
145 } | |
342
37f87013dfd8
6711316: Open source the Garbage-First garbage collector
ysr
parents:
113
diff
changeset
|
146 bool gc_overhead_limit_was_exceeded = false; |
0 | 147 result = Universe::heap()->mem_allocate(size, |
148 is_noref, | |
149 false, | |
150 &gc_overhead_limit_was_exceeded); | |
151 if (result != NULL) { | |
152 NOT_PRODUCT(Universe::heap()-> | |
153 check_for_non_bad_heap_word_value(result, size)); | |
154 assert(!HAS_PENDING_EXCEPTION, | |
155 "Unexpected exception, will result in uninitialized storage"); | |
2100
b1a2afa37ec4
7003271: Hotspot should track cumulative Java heap bytes allocated on a per-thread basis
phh
parents:
1972
diff
changeset
|
156 THREAD->incr_allocated_bytes(size * HeapWordSize); |
0 | 157 return result; |
158 } | |
159 | |
160 | |
161 if (!gc_overhead_limit_was_exceeded) { | |
162 // -XX:+HeapDumpOnOutOfMemoryError and -XX:OnOutOfMemoryError support | |
163 report_java_out_of_memory("Java heap space"); | |
164 | |
165 if (JvmtiExport::should_post_resource_exhausted()) { | |
166 JvmtiExport::post_resource_exhausted( | |
167 JVMTI_RESOURCE_EXHAUSTED_OOM_ERROR | JVMTI_RESOURCE_EXHAUSTED_JAVA_HEAP, | |
168 "Java heap space"); | |
169 } | |
170 | |
171 THROW_OOP_0(Universe::out_of_memory_error_java_heap()); | |
172 } else { | |
173 // -XX:+HeapDumpOnOutOfMemoryError and -XX:OnOutOfMemoryError support | |
174 report_java_out_of_memory("GC overhead limit exceeded"); | |
175 | |
176 if (JvmtiExport::should_post_resource_exhausted()) { | |
177 JvmtiExport::post_resource_exhausted( | |
178 JVMTI_RESOURCE_EXHAUSTED_OOM_ERROR | JVMTI_RESOURCE_EXHAUSTED_JAVA_HEAP, | |
179 "GC overhead limit exceeded"); | |
180 } | |
181 | |
182 THROW_OOP_0(Universe::out_of_memory_error_gc_overhead_limit()); | |
183 } | |
184 } | |
185 | |
186 HeapWord* CollectedHeap::common_mem_allocate_init(size_t size, bool is_noref, TRAPS) { | |
187 HeapWord* obj = common_mem_allocate_noinit(size, is_noref, CHECK_NULL); | |
188 init_obj(obj, size); | |
189 return obj; | |
190 } | |
191 | |
192 // Need to investigate, do we really want to throw OOM exception here? | |
193 HeapWord* CollectedHeap::common_permanent_mem_allocate_noinit(size_t size, TRAPS) { | |
194 if (HAS_PENDING_EXCEPTION) { | |
195 NOT_PRODUCT(guarantee(false, "Should not allocate with exception pending")); | |
196 return NULL; // caller does a CHECK_NULL too | |
197 } | |
198 | |
199 #ifdef ASSERT | |
200 if (CIFireOOMAt > 0 && THREAD->is_Compiler_thread() && | |
201 ++_fire_out_of_memory_count >= CIFireOOMAt) { | |
202 // For testing of OOM handling in the CI throw an OOM and see how | |
203 // it does. Historically improper handling of these has resulted | |
204 // in crashes which we really don't want to have in the CI. | |
205 THROW_OOP_0(Universe::out_of_memory_error_perm_gen()); | |
206 } | |
207 #endif | |
208 | |
209 HeapWord* result = Universe::heap()->permanent_mem_allocate(size); | |
210 if (result != NULL) { | |
211 NOT_PRODUCT(Universe::heap()-> | |
212 check_for_non_bad_heap_word_value(result, size)); | |
213 assert(!HAS_PENDING_EXCEPTION, | |
214 "Unexpected exception, will result in uninitialized storage"); | |
215 return result; | |
216 } | |
217 // -XX:+HeapDumpOnOutOfMemoryError and -XX:OnOutOfMemoryError support | |
218 report_java_out_of_memory("PermGen space"); | |
219 | |
220 if (JvmtiExport::should_post_resource_exhausted()) { | |
221 JvmtiExport::post_resource_exhausted( | |
222 JVMTI_RESOURCE_EXHAUSTED_OOM_ERROR, | |
223 "PermGen space"); | |
224 } | |
225 | |
226 THROW_OOP_0(Universe::out_of_memory_error_perm_gen()); | |
227 } | |
228 | |
229 HeapWord* CollectedHeap::common_permanent_mem_allocate_init(size_t size, TRAPS) { | |
230 HeapWord* obj = common_permanent_mem_allocate_noinit(size, CHECK_NULL); | |
231 init_obj(obj, size); | |
232 return obj; | |
233 } | |
234 | |
235 HeapWord* CollectedHeap::allocate_from_tlab(Thread* thread, size_t size) { | |
236 assert(UseTLAB, "should use UseTLAB"); | |
237 | |
238 HeapWord* obj = thread->tlab().allocate(size); | |
239 if (obj != NULL) { | |
240 return obj; | |
241 } | |
242 // Otherwise... | |
243 return allocate_from_tlab_slow(thread, size); | |
244 } | |
245 | |
246 void CollectedHeap::init_obj(HeapWord* obj, size_t size) { | |
247 assert(obj != NULL, "cannot initialize NULL object"); | |
248 const size_t hs = oopDesc::header_size(); | |
249 assert(size >= hs, "unexpected object size"); | |
167
feeb96a45707
6696264: assert("narrow oop can never be zero") for GCBasher & ParNewGC
coleenp
parents:
113
diff
changeset
|
250 ((oop)obj)->set_klass_gap(0); |
0 | 251 Copy::fill_to_aligned_words(obj + hs, size - hs); |
252 } | |
253 | |
254 oop CollectedHeap::obj_allocate(KlassHandle klass, int size, TRAPS) { | |
255 debug_only(check_for_valid_allocation_state()); | |
256 assert(!Universe::heap()->is_gc_active(), "Allocation during gc not allowed"); | |
257 assert(size >= 0, "int won't convert to size_t"); | |
258 HeapWord* obj = common_mem_allocate_init(size, false, CHECK_NULL); | |
259 post_allocation_setup_obj(klass, obj, size); | |
260 NOT_PRODUCT(Universe::heap()->check_for_bad_heap_word_value(obj, size)); | |
261 return (oop)obj; | |
262 } | |
263 | |
264 oop CollectedHeap::array_allocate(KlassHandle klass, | |
265 int size, | |
266 int length, | |
267 TRAPS) { | |
268 debug_only(check_for_valid_allocation_state()); | |
269 assert(!Universe::heap()->is_gc_active(), "Allocation during gc not allowed"); | |
270 assert(size >= 0, "int won't convert to size_t"); | |
271 HeapWord* obj = common_mem_allocate_init(size, false, CHECK_NULL); | |
272 post_allocation_setup_array(klass, obj, size, length); | |
273 NOT_PRODUCT(Universe::heap()->check_for_bad_heap_word_value(obj, size)); | |
274 return (oop)obj; | |
275 } | |
276 | |
277 oop CollectedHeap::large_typearray_allocate(KlassHandle klass, | |
278 int size, | |
279 int length, | |
280 TRAPS) { | |
281 debug_only(check_for_valid_allocation_state()); | |
282 assert(!Universe::heap()->is_gc_active(), "Allocation during gc not allowed"); | |
283 assert(size >= 0, "int won't convert to size_t"); | |
284 HeapWord* obj = common_mem_allocate_init(size, true, CHECK_NULL); | |
285 post_allocation_setup_array(klass, obj, size, length); | |
286 NOT_PRODUCT(Universe::heap()->check_for_bad_heap_word_value(obj, size)); | |
287 return (oop)obj; | |
288 } | |
289 | |
290 oop CollectedHeap::permanent_obj_allocate(KlassHandle klass, int size, TRAPS) { | |
291 oop obj = permanent_obj_allocate_no_klass_install(klass, size, CHECK_NULL); | |
292 post_allocation_install_obj_klass(klass, obj, size); | |
293 NOT_PRODUCT(Universe::heap()->check_for_bad_heap_word_value((HeapWord*) obj, | |
294 size)); | |
295 return obj; | |
296 } | |
297 | |
298 oop CollectedHeap::permanent_obj_allocate_no_klass_install(KlassHandle klass, | |
299 int size, | |
300 TRAPS) { | |
301 debug_only(check_for_valid_allocation_state()); | |
302 assert(!Universe::heap()->is_gc_active(), "Allocation during gc not allowed"); | |
303 assert(size >= 0, "int won't convert to size_t"); | |
304 HeapWord* obj = common_permanent_mem_allocate_init(size, CHECK_NULL); | |
305 post_allocation_setup_no_klass_install(klass, obj, size); | |
306 NOT_PRODUCT(Universe::heap()->check_for_bad_heap_word_value(obj, size)); | |
307 return (oop)obj; | |
308 } | |
309 | |
310 oop CollectedHeap::permanent_array_allocate(KlassHandle klass, | |
311 int size, | |
312 int length, | |
313 TRAPS) { | |
314 debug_only(check_for_valid_allocation_state()); | |
315 assert(!Universe::heap()->is_gc_active(), "Allocation during gc not allowed"); | |
316 assert(size >= 0, "int won't convert to size_t"); | |
317 HeapWord* obj = common_permanent_mem_allocate_init(size, CHECK_NULL); | |
318 post_allocation_setup_array(klass, obj, size, length); | |
319 NOT_PRODUCT(Universe::heap()->check_for_bad_heap_word_value(obj, size)); | |
320 return (oop)obj; | |
321 } | |
322 | |
323 // Returns "TRUE" if "p" is a method oop in the | |
324 // current heap with high probability. NOTE: The main | |
325 // current consumers of this interface are Forte:: | |
326 // and ThreadProfiler::. In these cases, the | |
327 // interpreter frame from which "p" came, may be | |
328 // under construction when sampled asynchronously, so | |
329 // the clients want to check that it represents a | |
330 // valid method before using it. Nonetheless since | |
331 // the clients do not typically lock out GC, the | |
332 // predicate is_valid_method() is not stable, so | |
333 // it is possible that by the time "p" is used, it | |
334 // is no longer valid. | |
335 inline bool CollectedHeap::is_valid_method(oop p) const { | |
336 return | |
337 p != NULL && | |
338 | |
339 // Check whether it is aligned at a HeapWord boundary. | |
340 Space::is_aligned(p) && | |
341 | |
342 // Check whether "method" is in the allocated part of the | |
343 // permanent generation -- this needs to be checked before | |
344 // p->klass() below to avoid a SEGV (but see below | |
345 // for a potential window of vulnerability). | |
346 is_permanent((void*)p) && | |
347 | |
348 // See if GC is active; however, there is still an | |
349 // apparently unavoidable window after this call | |
350 // and before the client of this interface uses "p". | |
351 // If the client chooses not to lock out GC, then | |
352 // it's a risk the client must accept. | |
353 !is_gc_active() && | |
354 | |
355 // Check that p is a methodOop. | |
356 p->klass() == Universe::methodKlassObj(); | |
357 } | |
358 | |
359 | |
360 #ifndef PRODUCT | |
361 | |
362 inline bool | |
363 CollectedHeap::promotion_should_fail(volatile size_t* count) { | |
364 // Access to count is not atomic; the value does not have to be exact. | |
365 if (PromotionFailureALot) { | |
366 const size_t gc_num = total_collections(); | |
367 const size_t elapsed_gcs = gc_num - _promotion_failure_alot_gc_number; | |
368 if (elapsed_gcs >= PromotionFailureALotInterval) { | |
369 // Test for unsigned arithmetic wrap-around. | |
370 if (++*count >= PromotionFailureALotCount) { | |
371 *count = 0; | |
372 return true; | |
373 } | |
374 } | |
375 } | |
376 return false; | |
377 } | |
378 | |
379 inline bool CollectedHeap::promotion_should_fail() { | |
380 return promotion_should_fail(&_promotion_failure_alot_count); | |
381 } | |
382 | |
383 inline void CollectedHeap::reset_promotion_should_fail(volatile size_t* count) { | |
384 if (PromotionFailureALot) { | |
385 _promotion_failure_alot_gc_number = total_collections(); | |
386 *count = 0; | |
387 } | |
388 } | |
389 | |
390 inline void CollectedHeap::reset_promotion_should_fail() { | |
391 reset_promotion_should_fail(&_promotion_failure_alot_count); | |
392 } | |
393 #endif // #ifndef PRODUCT | |
1972 | 394 |
395 #endif // SHARE_VM_GC_INTERFACE_COLLECTEDHEAP_INLINE_HPP |