comparison src/share/vm/oops/instanceMirrorKlass.cpp @ 2376:c7f3d0b4570f

7017732: move static fields into Class to prepare for perm gen removal Reviewed-by: kvn, coleenp, twisti, stefank
author never
date Fri, 18 Mar 2011 16:00:34 -0700
parents
children e5928e7dab26
comparison
equal deleted inserted replaced
2375:d673ef06fe96 2376:c7f3d0b4570f
1 /*
2 * Copyright (c) 2011, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #include "precompiled.hpp"
26 #include "classfile/javaClasses.hpp"
27 #include "classfile/systemDictionary.hpp"
28 #include "gc_implementation/shared/markSweep.inline.hpp"
29 #include "gc_interface/collectedHeap.inline.hpp"
30 #include "memory/genOopClosures.inline.hpp"
31 #include "memory/oopFactory.hpp"
32 #include "memory/permGen.hpp"
33 #include "oops/instanceKlass.hpp"
34 #include "oops/instanceMirrorKlass.hpp"
35 #include "oops/instanceOop.hpp"
36 #include "oops/oop.inline.hpp"
37 #include "oops/symbol.hpp"
38 #include "runtime/handles.inline.hpp"
39 #ifndef SERIALGC
40 #include "gc_implementation/g1/g1CollectedHeap.inline.hpp"
41 #include "gc_implementation/g1/g1OopClosures.inline.hpp"
42 #include "gc_implementation/g1/g1RemSet.inline.hpp"
43 #include "gc_implementation/g1/heapRegionSeq.inline.hpp"
44 #include "gc_implementation/parNew/parOopClosures.inline.hpp"
45 #include "gc_implementation/parallelScavenge/psPromotionManager.inline.hpp"
46 #include "gc_implementation/parallelScavenge/psScavenge.inline.hpp"
47 #include "oops/oop.pcgc.inline.hpp"
48 #endif
49
50 int instanceMirrorKlass::_offset_of_static_fields = 0;
51
52 #ifdef ASSERT
53 template <class T> void assert_is_in(T *p) {
54 T heap_oop = oopDesc::load_heap_oop(p);
55 if (!oopDesc::is_null(heap_oop)) {
56 oop o = oopDesc::decode_heap_oop_not_null(heap_oop);
57 assert(Universe::heap()->is_in(o), "should be in heap");
58 }
59 }
60 template <class T> void assert_is_in_closed_subset(T *p) {
61 T heap_oop = oopDesc::load_heap_oop(p);
62 if (!oopDesc::is_null(heap_oop)) {
63 oop o = oopDesc::decode_heap_oop_not_null(heap_oop);
64 assert(Universe::heap()->is_in_closed_subset(o), "should be in closed");
65 }
66 }
67 template <class T> void assert_is_in_reserved(T *p) {
68 T heap_oop = oopDesc::load_heap_oop(p);
69 if (!oopDesc::is_null(heap_oop)) {
70 oop o = oopDesc::decode_heap_oop_not_null(heap_oop);
71 assert(Universe::heap()->is_in_reserved(o), "should be in reserved");
72 }
73 }
74 template <class T> void assert_nothing(T *p) {}
75
76 #else
77 template <class T> void assert_is_in(T *p) {}
78 template <class T> void assert_is_in_closed_subset(T *p) {}
79 template <class T> void assert_is_in_reserved(T *p) {}
80 template <class T> void assert_nothing(T *p) {}
81 #endif // ASSERT
82
83 #define InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE( \
84 T, start_p, count, do_oop, \
85 assert_fn) \
86 { \
87 T* p = (T*)(start_p); \
88 T* const end = p + (count); \
89 while (p < end) { \
90 (assert_fn)(p); \
91 do_oop; \
92 ++p; \
93 } \
94 }
95
96 #define InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE( \
97 T, start_p, count, low, high, \
98 do_oop, assert_fn) \
99 { \
100 T* const l = (T*)(low); \
101 T* const h = (T*)(high); \
102 assert(mask_bits((intptr_t)l, sizeof(T)-1) == 0 && \
103 mask_bits((intptr_t)h, sizeof(T)-1) == 0, \
104 "bounded region must be properly aligned"); \
105 T* p = (T*)(start_p); \
106 T* end = p + (count); \
107 if (p < l) p = l; \
108 if (end > h) end = h; \
109 while (p < end) { \
110 (assert_fn)(p); \
111 do_oop; \
112 ++p; \
113 } \
114 }
115
116
117 #define InstanceMirrorKlass_OOP_ITERATE(start_p, count, \
118 do_oop, assert_fn) \
119 { \
120 if (UseCompressedOops) { \
121 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE(narrowOop, \
122 start_p, count, \
123 do_oop, assert_fn) \
124 } else { \
125 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE(oop, \
126 start_p, count, \
127 do_oop, assert_fn) \
128 } \
129 }
130
131 // The following macros call specialized macros, passing either oop or
132 // narrowOop as the specialization type. These test the UseCompressedOops
133 // flag.
134 #define InstanceMirrorKlass_BOUNDED_OOP_ITERATE(start_p, count, low, high, \
135 do_oop, assert_fn) \
136 { \
137 if (UseCompressedOops) { \
138 InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE(narrowOop, \
139 start_p, count, \
140 low, high, \
141 do_oop, assert_fn) \
142 } else { \
143 InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE(oop, \
144 start_p, count, \
145 low, high, \
146 do_oop, assert_fn) \
147 } \
148 }
149
150
151 void instanceMirrorKlass::oop_follow_contents(oop obj) {
152 instanceKlass::oop_follow_contents(obj);
153 InstanceMirrorKlass_OOP_ITERATE( \
154 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
155 MarkSweep::mark_and_push(p), \
156 assert_is_in_closed_subset)
157 }
158
159 #ifndef SERIALGC
160 void instanceMirrorKlass::oop_follow_contents(ParCompactionManager* cm,
161 oop obj) {
162 instanceKlass::oop_follow_contents(cm, obj);
163 InstanceMirrorKlass_OOP_ITERATE( \
164 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
165 PSParallelCompact::mark_and_push(cm, p), \
166 assert_is_in)
167 }
168 #endif // SERIALGC
169
170 int instanceMirrorKlass::oop_adjust_pointers(oop obj) {
171 int size = oop_size(obj);
172 instanceKlass::oop_adjust_pointers(obj);
173 InstanceMirrorKlass_OOP_ITERATE( \
174 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
175 MarkSweep::adjust_pointer(p), \
176 assert_nothing)
177 return size;
178 }
179
180 #define InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(T, nv_suffix) \
181 InstanceMirrorKlass_OOP_ITERATE( \
182 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
183 (closure)->do_oop##nv_suffix(p), \
184 assert_is_in_closed_subset) \
185 return oop_size(obj); \
186
187 #define InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(T, nv_suffix, mr) \
188 InstanceMirrorKlass_BOUNDED_OOP_ITERATE( \
189 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
190 mr.start(), mr.end(), \
191 (closure)->do_oop##nv_suffix(p), \
192 assert_is_in_closed_subset) \
193 return oop_size(obj); \
194
195
196 // Macro to define instanceMirrorKlass::oop_oop_iterate for virtual/nonvirtual for
197 // all closures. Macros calling macros above for each oop size.
198
199 #define InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN(OopClosureType, nv_suffix) \
200 \
201 int instanceMirrorKlass:: \
202 oop_oop_iterate##nv_suffix(oop obj, OopClosureType* closure) { \
203 /* Get size before changing pointers */ \
204 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk); \
205 \
206 instanceKlass::oop_oop_iterate##nv_suffix(obj, closure); \
207 \
208 if (UseCompressedOops) { \
209 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(narrowOop, nv_suffix); \
210 } else { \
211 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(oop, nv_suffix); \
212 } \
213 }
214
215 #ifndef SERIALGC
216 #define InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN(OopClosureType, nv_suffix) \
217 \
218 int instanceMirrorKlass:: \
219 oop_oop_iterate_backwards##nv_suffix(oop obj, OopClosureType* closure) { \
220 /* Get size before changing pointers */ \
221 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk); \
222 \
223 instanceKlass::oop_oop_iterate_backwards##nv_suffix(obj, closure); \
224 \
225 if (UseCompressedOops) { \
226 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(narrowOop, nv_suffix); \
227 } else { \
228 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(oop, nv_suffix); \
229 } \
230 }
231 #endif // !SERIALGC
232
233
234 #define InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m(OopClosureType, nv_suffix) \
235 \
236 int instanceMirrorKlass:: \
237 oop_oop_iterate##nv_suffix##_m(oop obj, \
238 OopClosureType* closure, \
239 MemRegion mr) { \
240 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk); \
241 \
242 instanceKlass::oop_oop_iterate##nv_suffix##_m(obj, closure, mr); \
243 if (UseCompressedOops) { \
244 InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(narrowOop, nv_suffix, mr); \
245 } else { \
246 InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(oop, nv_suffix, mr); \
247 } \
248 }
249
250 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN)
251 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN)
252 #ifndef SERIALGC
253 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN)
254 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN)
255 #endif // SERIALGC
256 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m)
257 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m)
258
259 #ifndef SERIALGC
260 void instanceMirrorKlass::oop_push_contents(PSPromotionManager* pm, oop obj) {
261 instanceKlass::oop_push_contents(pm, obj);
262 InstanceMirrorKlass_OOP_ITERATE( \
263 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),\
264 if (PSScavenge::should_scavenge(p)) { \
265 pm->claim_or_forward_depth(p); \
266 }, \
267 assert_nothing )
268 }
269
270 int instanceMirrorKlass::oop_update_pointers(ParCompactionManager* cm, oop obj) {
271 instanceKlass::oop_update_pointers(cm, obj);
272 InstanceMirrorKlass_OOP_ITERATE( \
273 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),\
274 PSParallelCompact::adjust_pointer(p), \
275 assert_nothing)
276 return oop_size(obj);
277 }
278 #endif // SERIALGC
279
280 int instanceMirrorKlass::instance_size(KlassHandle k) {
281 if (k() != NULL && k->oop_is_instance()) {
282 return align_object_size(size_helper() + instanceKlass::cast(k())->static_field_size());
283 }
284 return size_helper();
285 }
286
287 instanceOop instanceMirrorKlass::allocate_instance(KlassHandle k, TRAPS) {
288 // Query before forming handle.
289 int size = instance_size(k);
290 KlassHandle h_k(THREAD, as_klassOop());
291 instanceOop i;
292
293 if (JavaObjectsInPerm) {
294 i = (instanceOop) CollectedHeap::permanent_obj_allocate(h_k, size, CHECK_NULL);
295 } else {
296 assert(ScavengeRootsInCode > 0, "must be");
297 i = (instanceOop) CollectedHeap::obj_allocate(h_k, size, CHECK_NULL);
298 }
299
300 return i;
301 }
302
303 int instanceMirrorKlass::oop_size(oop obj) const {
304 return java_lang_Class::oop_size(obj);
305 }
306
307 int instanceMirrorKlass::compute_static_oop_field_count(oop obj) {
308 klassOop k = java_lang_Class::as_klassOop(obj);
309 if (k != NULL && k->klass_part()->oop_is_instance()) {
310 return instanceKlass::cast(k)->static_oop_field_count();
311 }
312 return 0;
313 }