Mercurial > hg > graal-jvmci-8
comparison src/share/vm/oops/instanceMirrorKlass.cpp @ 2376:c7f3d0b4570f
7017732: move static fields into Class to prepare for perm gen removal
Reviewed-by: kvn, coleenp, twisti, stefank
author | never |
---|---|
date | Fri, 18 Mar 2011 16:00:34 -0700 |
parents | |
children | e5928e7dab26 |
comparison
equal
deleted
inserted
replaced
2375:d673ef06fe96 | 2376:c7f3d0b4570f |
---|---|
1 /* | |
2 * Copyright (c) 2011, Oracle and/or its affiliates. All rights reserved. | |
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. | |
4 * | |
5 * This code is free software; you can redistribute it and/or modify it | |
6 * under the terms of the GNU General Public License version 2 only, as | |
7 * published by the Free Software Foundation. | |
8 * | |
9 * This code is distributed in the hope that it will be useful, but WITHOUT | |
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License | |
12 * version 2 for more details (a copy is included in the LICENSE file that | |
13 * accompanied this code). | |
14 * | |
15 * You should have received a copy of the GNU General Public License version | |
16 * 2 along with this work; if not, write to the Free Software Foundation, | |
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | |
18 * | |
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA | |
20 * or visit www.oracle.com if you need additional information or have any | |
21 * questions. | |
22 * | |
23 */ | |
24 | |
25 #include "precompiled.hpp" | |
26 #include "classfile/javaClasses.hpp" | |
27 #include "classfile/systemDictionary.hpp" | |
28 #include "gc_implementation/shared/markSweep.inline.hpp" | |
29 #include "gc_interface/collectedHeap.inline.hpp" | |
30 #include "memory/genOopClosures.inline.hpp" | |
31 #include "memory/oopFactory.hpp" | |
32 #include "memory/permGen.hpp" | |
33 #include "oops/instanceKlass.hpp" | |
34 #include "oops/instanceMirrorKlass.hpp" | |
35 #include "oops/instanceOop.hpp" | |
36 #include "oops/oop.inline.hpp" | |
37 #include "oops/symbol.hpp" | |
38 #include "runtime/handles.inline.hpp" | |
39 #ifndef SERIALGC | |
40 #include "gc_implementation/g1/g1CollectedHeap.inline.hpp" | |
41 #include "gc_implementation/g1/g1OopClosures.inline.hpp" | |
42 #include "gc_implementation/g1/g1RemSet.inline.hpp" | |
43 #include "gc_implementation/g1/heapRegionSeq.inline.hpp" | |
44 #include "gc_implementation/parNew/parOopClosures.inline.hpp" | |
45 #include "gc_implementation/parallelScavenge/psPromotionManager.inline.hpp" | |
46 #include "gc_implementation/parallelScavenge/psScavenge.inline.hpp" | |
47 #include "oops/oop.pcgc.inline.hpp" | |
48 #endif | |
49 | |
50 int instanceMirrorKlass::_offset_of_static_fields = 0; | |
51 | |
52 #ifdef ASSERT | |
53 template <class T> void assert_is_in(T *p) { | |
54 T heap_oop = oopDesc::load_heap_oop(p); | |
55 if (!oopDesc::is_null(heap_oop)) { | |
56 oop o = oopDesc::decode_heap_oop_not_null(heap_oop); | |
57 assert(Universe::heap()->is_in(o), "should be in heap"); | |
58 } | |
59 } | |
60 template <class T> void assert_is_in_closed_subset(T *p) { | |
61 T heap_oop = oopDesc::load_heap_oop(p); | |
62 if (!oopDesc::is_null(heap_oop)) { | |
63 oop o = oopDesc::decode_heap_oop_not_null(heap_oop); | |
64 assert(Universe::heap()->is_in_closed_subset(o), "should be in closed"); | |
65 } | |
66 } | |
67 template <class T> void assert_is_in_reserved(T *p) { | |
68 T heap_oop = oopDesc::load_heap_oop(p); | |
69 if (!oopDesc::is_null(heap_oop)) { | |
70 oop o = oopDesc::decode_heap_oop_not_null(heap_oop); | |
71 assert(Universe::heap()->is_in_reserved(o), "should be in reserved"); | |
72 } | |
73 } | |
74 template <class T> void assert_nothing(T *p) {} | |
75 | |
76 #else | |
77 template <class T> void assert_is_in(T *p) {} | |
78 template <class T> void assert_is_in_closed_subset(T *p) {} | |
79 template <class T> void assert_is_in_reserved(T *p) {} | |
80 template <class T> void assert_nothing(T *p) {} | |
81 #endif // ASSERT | |
82 | |
83 #define InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE( \ | |
84 T, start_p, count, do_oop, \ | |
85 assert_fn) \ | |
86 { \ | |
87 T* p = (T*)(start_p); \ | |
88 T* const end = p + (count); \ | |
89 while (p < end) { \ | |
90 (assert_fn)(p); \ | |
91 do_oop; \ | |
92 ++p; \ | |
93 } \ | |
94 } | |
95 | |
96 #define InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE( \ | |
97 T, start_p, count, low, high, \ | |
98 do_oop, assert_fn) \ | |
99 { \ | |
100 T* const l = (T*)(low); \ | |
101 T* const h = (T*)(high); \ | |
102 assert(mask_bits((intptr_t)l, sizeof(T)-1) == 0 && \ | |
103 mask_bits((intptr_t)h, sizeof(T)-1) == 0, \ | |
104 "bounded region must be properly aligned"); \ | |
105 T* p = (T*)(start_p); \ | |
106 T* end = p + (count); \ | |
107 if (p < l) p = l; \ | |
108 if (end > h) end = h; \ | |
109 while (p < end) { \ | |
110 (assert_fn)(p); \ | |
111 do_oop; \ | |
112 ++p; \ | |
113 } \ | |
114 } | |
115 | |
116 | |
117 #define InstanceMirrorKlass_OOP_ITERATE(start_p, count, \ | |
118 do_oop, assert_fn) \ | |
119 { \ | |
120 if (UseCompressedOops) { \ | |
121 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE(narrowOop, \ | |
122 start_p, count, \ | |
123 do_oop, assert_fn) \ | |
124 } else { \ | |
125 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE(oop, \ | |
126 start_p, count, \ | |
127 do_oop, assert_fn) \ | |
128 } \ | |
129 } | |
130 | |
131 // The following macros call specialized macros, passing either oop or | |
132 // narrowOop as the specialization type. These test the UseCompressedOops | |
133 // flag. | |
134 #define InstanceMirrorKlass_BOUNDED_OOP_ITERATE(start_p, count, low, high, \ | |
135 do_oop, assert_fn) \ | |
136 { \ | |
137 if (UseCompressedOops) { \ | |
138 InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE(narrowOop, \ | |
139 start_p, count, \ | |
140 low, high, \ | |
141 do_oop, assert_fn) \ | |
142 } else { \ | |
143 InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE(oop, \ | |
144 start_p, count, \ | |
145 low, high, \ | |
146 do_oop, assert_fn) \ | |
147 } \ | |
148 } | |
149 | |
150 | |
151 void instanceMirrorKlass::oop_follow_contents(oop obj) { | |
152 instanceKlass::oop_follow_contents(obj); | |
153 InstanceMirrorKlass_OOP_ITERATE( \ | |
154 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \ | |
155 MarkSweep::mark_and_push(p), \ | |
156 assert_is_in_closed_subset) | |
157 } | |
158 | |
159 #ifndef SERIALGC | |
160 void instanceMirrorKlass::oop_follow_contents(ParCompactionManager* cm, | |
161 oop obj) { | |
162 instanceKlass::oop_follow_contents(cm, obj); | |
163 InstanceMirrorKlass_OOP_ITERATE( \ | |
164 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \ | |
165 PSParallelCompact::mark_and_push(cm, p), \ | |
166 assert_is_in) | |
167 } | |
168 #endif // SERIALGC | |
169 | |
170 int instanceMirrorKlass::oop_adjust_pointers(oop obj) { | |
171 int size = oop_size(obj); | |
172 instanceKlass::oop_adjust_pointers(obj); | |
173 InstanceMirrorKlass_OOP_ITERATE( \ | |
174 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \ | |
175 MarkSweep::adjust_pointer(p), \ | |
176 assert_nothing) | |
177 return size; | |
178 } | |
179 | |
180 #define InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(T, nv_suffix) \ | |
181 InstanceMirrorKlass_OOP_ITERATE( \ | |
182 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \ | |
183 (closure)->do_oop##nv_suffix(p), \ | |
184 assert_is_in_closed_subset) \ | |
185 return oop_size(obj); \ | |
186 | |
187 #define InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(T, nv_suffix, mr) \ | |
188 InstanceMirrorKlass_BOUNDED_OOP_ITERATE( \ | |
189 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \ | |
190 mr.start(), mr.end(), \ | |
191 (closure)->do_oop##nv_suffix(p), \ | |
192 assert_is_in_closed_subset) \ | |
193 return oop_size(obj); \ | |
194 | |
195 | |
196 // Macro to define instanceMirrorKlass::oop_oop_iterate for virtual/nonvirtual for | |
197 // all closures. Macros calling macros above for each oop size. | |
198 | |
199 #define InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN(OopClosureType, nv_suffix) \ | |
200 \ | |
201 int instanceMirrorKlass:: \ | |
202 oop_oop_iterate##nv_suffix(oop obj, OopClosureType* closure) { \ | |
203 /* Get size before changing pointers */ \ | |
204 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk); \ | |
205 \ | |
206 instanceKlass::oop_oop_iterate##nv_suffix(obj, closure); \ | |
207 \ | |
208 if (UseCompressedOops) { \ | |
209 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(narrowOop, nv_suffix); \ | |
210 } else { \ | |
211 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(oop, nv_suffix); \ | |
212 } \ | |
213 } | |
214 | |
215 #ifndef SERIALGC | |
216 #define InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN(OopClosureType, nv_suffix) \ | |
217 \ | |
218 int instanceMirrorKlass:: \ | |
219 oop_oop_iterate_backwards##nv_suffix(oop obj, OopClosureType* closure) { \ | |
220 /* Get size before changing pointers */ \ | |
221 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk); \ | |
222 \ | |
223 instanceKlass::oop_oop_iterate_backwards##nv_suffix(obj, closure); \ | |
224 \ | |
225 if (UseCompressedOops) { \ | |
226 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(narrowOop, nv_suffix); \ | |
227 } else { \ | |
228 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(oop, nv_suffix); \ | |
229 } \ | |
230 } | |
231 #endif // !SERIALGC | |
232 | |
233 | |
234 #define InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m(OopClosureType, nv_suffix) \ | |
235 \ | |
236 int instanceMirrorKlass:: \ | |
237 oop_oop_iterate##nv_suffix##_m(oop obj, \ | |
238 OopClosureType* closure, \ | |
239 MemRegion mr) { \ | |
240 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk); \ | |
241 \ | |
242 instanceKlass::oop_oop_iterate##nv_suffix##_m(obj, closure, mr); \ | |
243 if (UseCompressedOops) { \ | |
244 InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(narrowOop, nv_suffix, mr); \ | |
245 } else { \ | |
246 InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(oop, nv_suffix, mr); \ | |
247 } \ | |
248 } | |
249 | |
250 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN) | |
251 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN) | |
252 #ifndef SERIALGC | |
253 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN) | |
254 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN) | |
255 #endif // SERIALGC | |
256 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m) | |
257 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m) | |
258 | |
259 #ifndef SERIALGC | |
260 void instanceMirrorKlass::oop_push_contents(PSPromotionManager* pm, oop obj) { | |
261 instanceKlass::oop_push_contents(pm, obj); | |
262 InstanceMirrorKlass_OOP_ITERATE( \ | |
263 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),\ | |
264 if (PSScavenge::should_scavenge(p)) { \ | |
265 pm->claim_or_forward_depth(p); \ | |
266 }, \ | |
267 assert_nothing ) | |
268 } | |
269 | |
270 int instanceMirrorKlass::oop_update_pointers(ParCompactionManager* cm, oop obj) { | |
271 instanceKlass::oop_update_pointers(cm, obj); | |
272 InstanceMirrorKlass_OOP_ITERATE( \ | |
273 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),\ | |
274 PSParallelCompact::adjust_pointer(p), \ | |
275 assert_nothing) | |
276 return oop_size(obj); | |
277 } | |
278 #endif // SERIALGC | |
279 | |
280 int instanceMirrorKlass::instance_size(KlassHandle k) { | |
281 if (k() != NULL && k->oop_is_instance()) { | |
282 return align_object_size(size_helper() + instanceKlass::cast(k())->static_field_size()); | |
283 } | |
284 return size_helper(); | |
285 } | |
286 | |
287 instanceOop instanceMirrorKlass::allocate_instance(KlassHandle k, TRAPS) { | |
288 // Query before forming handle. | |
289 int size = instance_size(k); | |
290 KlassHandle h_k(THREAD, as_klassOop()); | |
291 instanceOop i; | |
292 | |
293 if (JavaObjectsInPerm) { | |
294 i = (instanceOop) CollectedHeap::permanent_obj_allocate(h_k, size, CHECK_NULL); | |
295 } else { | |
296 assert(ScavengeRootsInCode > 0, "must be"); | |
297 i = (instanceOop) CollectedHeap::obj_allocate(h_k, size, CHECK_NULL); | |
298 } | |
299 | |
300 return i; | |
301 } | |
302 | |
303 int instanceMirrorKlass::oop_size(oop obj) const { | |
304 return java_lang_Class::oop_size(obj); | |
305 } | |
306 | |
307 int instanceMirrorKlass::compute_static_oop_field_count(oop obj) { | |
308 klassOop k = java_lang_Class::as_klassOop(obj); | |
309 if (k != NULL && k->klass_part()->oop_is_instance()) { | |
310 return instanceKlass::cast(k)->static_oop_field_count(); | |
311 } | |
312 return 0; | |
313 } |