Mercurial > hg > truffle
comparison src/share/vm/code/compiledIC.cpp @ 6725:da91efe96a93
6964458: Reimplement class meta-data storage to use native memory
Summary: Remove PermGen, allocate meta-data in metaspace linked to class loaders, rewrite GC walking, rewrite and rename metadata to be C++ classes
Reviewed-by: jmasa, stefank, never, coleenp, kvn, brutisso, mgerdin, dholmes, jrose, twisti, roland
Contributed-by: jmasa <jon.masamitsu@oracle.com>, stefank <stefan.karlsson@oracle.com>, mgerdin <mikael.gerdin@oracle.com>, never <tom.rodriguez@oracle.com>
author | coleenp |
---|---|
date | Sat, 01 Sep 2012 13:25:18 -0400 |
parents | aa3d708d67c4 |
children | 18fb7da42534 |
comparison
equal
deleted
inserted
replaced
6724:36d1d483d5d6 | 6725:da91efe96a93 |
---|---|
29 #include "code/icBuffer.hpp" | 29 #include "code/icBuffer.hpp" |
30 #include "code/nmethod.hpp" | 30 #include "code/nmethod.hpp" |
31 #include "code/vtableStubs.hpp" | 31 #include "code/vtableStubs.hpp" |
32 #include "interpreter/interpreter.hpp" | 32 #include "interpreter/interpreter.hpp" |
33 #include "interpreter/linkResolver.hpp" | 33 #include "interpreter/linkResolver.hpp" |
34 #include "memory/metadataFactory.hpp" | |
34 #include "memory/oopFactory.hpp" | 35 #include "memory/oopFactory.hpp" |
35 #include "oops/methodOop.hpp" | 36 #include "oops/method.hpp" |
36 #include "oops/oop.inline.hpp" | 37 #include "oops/oop.inline.hpp" |
37 #include "oops/symbol.hpp" | 38 #include "oops/symbol.hpp" |
38 #include "runtime/icache.hpp" | 39 #include "runtime/icache.hpp" |
39 #include "runtime/sharedRuntime.hpp" | 40 #include "runtime/sharedRuntime.hpp" |
40 #include "runtime/stubRoutines.hpp" | 41 #include "runtime/stubRoutines.hpp" |
42 | 43 |
43 | 44 |
44 // Every time a compiled IC is changed or its type is being accessed, | 45 // Every time a compiled IC is changed or its type is being accessed, |
45 // either the CompiledIC_lock must be set or we must be at a safe point. | 46 // either the CompiledIC_lock must be set or we must be at a safe point. |
46 | 47 |
48 | |
49 // Release the CompiledICHolder* associated with this call site is there is one. | |
50 void CompiledIC::cleanup_call_site(virtual_call_Relocation* call_site) { | |
51 // This call site might have become stale so inspect it carefully. | |
52 NativeCall* call = nativeCall_at(call_site->addr()); | |
53 if (is_icholder_entry(call->destination())) { | |
54 NativeMovConstReg* value = nativeMovConstReg_at(call_site->cached_value()); | |
55 InlineCacheBuffer::queue_for_release((CompiledICHolder*)value->data()); | |
56 } | |
57 } | |
58 | |
59 | |
60 bool CompiledIC::is_icholder_call_site(virtual_call_Relocation* call_site) { | |
61 // This call site might have become stale so inspect it carefully. | |
62 NativeCall* call = nativeCall_at(call_site->addr()); | |
63 return is_icholder_entry(call->destination()); | |
64 } | |
65 | |
66 | |
47 //----------------------------------------------------------------------------- | 67 //----------------------------------------------------------------------------- |
48 // Low-level access to an inline cache. Private, since they might not be | 68 // Low-level access to an inline cache. Private, since they might not be |
49 // MT-safe to use. | 69 // MT-safe to use. |
50 | 70 |
51 void CompiledIC::set_cached_oop(oop cache) { | 71 void* CompiledIC::cached_value() const { |
52 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | 72 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); |
53 assert (!is_optimized(), "an optimized virtual call does not have a cached oop"); | 73 assert (!is_optimized(), "an optimized virtual call does not have a cached metadata"); |
54 assert (cache == NULL || cache != badOop, "invalid oop"); | 74 |
75 if (!is_in_transition_state()) { | |
76 void* data = (void*)_value->data(); | |
77 // If we let the metadata value here be initialized to zero... | |
78 assert(data != NULL || Universe::non_oop_word() == NULL, | |
79 "no raw nulls in CompiledIC metadatas, because of patching races"); | |
80 return (data == (void*)Universe::non_oop_word()) ? NULL : data; | |
81 } else { | |
82 return InlineCacheBuffer::cached_value_for((CompiledIC *)this); | |
83 } | |
84 } | |
85 | |
86 | |
87 void CompiledIC::internal_set_ic_destination(address entry_point, bool is_icstub, void* cache, bool is_icholder) { | |
88 assert(entry_point != NULL, "must set legal entry point"); | |
89 assert(CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | |
90 assert (!is_optimized() || cache == NULL, "an optimized virtual call does not have a cached metadata"); | |
91 assert (cache == NULL || cache != (Metadata*)badOopVal, "invalid metadata"); | |
92 | |
93 assert(!is_icholder || is_icholder_entry(entry_point), "must be"); | |
94 | |
95 // Don't use ic_destination for this test since that forwards | |
96 // through ICBuffer instead of returning the actual current state of | |
97 // the CompiledIC. | |
98 if (is_icholder_entry(_ic_call->destination())) { | |
99 // When patching for the ICStub case the cached value isn't | |
100 // overwritten until the ICStub copied into the CompiledIC during | |
101 // the next safepoint. Make sure that the CompiledICHolder* is | |
102 // marked for release at this point since it won't be identifiable | |
103 // once the entry point is overwritten. | |
104 InlineCacheBuffer::queue_for_release((CompiledICHolder*)_value->data()); | |
105 } | |
55 | 106 |
56 if (TraceCompiledIC) { | 107 if (TraceCompiledIC) { |
57 tty->print(" "); | 108 tty->print(" "); |
58 print_compiled_ic(); | 109 print_compiled_ic(); |
59 tty->print_cr(" changing oop to " INTPTR_FORMAT, (address)cache); | 110 tty->print(" changing destination to " INTPTR_FORMAT, entry_point); |
60 } | 111 if (!is_optimized()) { |
61 | 112 tty->print(" changing cached %s to " INTPTR_FORMAT, is_icholder ? "icholder" : "metadata", (address)cache); |
62 if (cache == NULL) cache = (oop)Universe::non_oop_word(); | 113 } |
63 | 114 if (is_icstub) { |
64 *_oop_addr = cache; | 115 tty->print(" (icstub)"); |
65 // fix up the relocations | 116 } |
66 RelocIterator iter = _oops; | 117 tty->cr(); |
67 while (iter.next()) { | 118 } |
68 if (iter.type() == relocInfo::oop_type) { | 119 |
69 oop_Relocation* r = iter.oop_reloc(); | 120 { |
70 if (r->oop_addr() == _oop_addr) | |
71 r->fix_oop_relocation(); | |
72 } | |
73 } | |
74 return; | |
75 } | |
76 | |
77 | |
78 oop CompiledIC::cached_oop() const { | |
79 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | |
80 assert (!is_optimized(), "an optimized virtual call does not have a cached oop"); | |
81 | |
82 if (!is_in_transition_state()) { | |
83 oop data = *_oop_addr; | |
84 // If we let the oop value here be initialized to zero... | |
85 assert(data != NULL || Universe::non_oop_word() == NULL, | |
86 "no raw nulls in CompiledIC oops, because of patching races"); | |
87 return (data == (oop)Universe::non_oop_word()) ? (oop)NULL : data; | |
88 } else { | |
89 return InlineCacheBuffer::cached_oop_for((CompiledIC *)this); | |
90 } | |
91 } | |
92 | |
93 | |
94 void CompiledIC::set_ic_destination(address entry_point) { | |
95 assert(entry_point != NULL, "must set legal entry point"); | |
96 assert(CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | |
97 if (TraceCompiledIC) { | |
98 tty->print(" "); | |
99 print_compiled_ic(); | |
100 tty->print_cr(" changing destination to " INTPTR_FORMAT, entry_point); | |
101 } | |
102 MutexLockerEx pl(Patching_lock, Mutex::_no_safepoint_check_flag); | 121 MutexLockerEx pl(Patching_lock, Mutex::_no_safepoint_check_flag); |
103 #ifdef ASSERT | 122 #ifdef ASSERT |
104 CodeBlob* cb = CodeCache::find_blob_unsafe(_ic_call); | 123 CodeBlob* cb = CodeCache::find_blob_unsafe(_ic_call); |
105 assert(cb != NULL && cb->is_nmethod(), "must be nmethod"); | 124 assert(cb != NULL && cb->is_nmethod(), "must be nmethod"); |
106 #endif | 125 #endif |
107 _ic_call->set_destination_mt_safe(entry_point); | 126 _ic_call->set_destination_mt_safe(entry_point); |
108 } | 127 } |
128 | |
129 if (is_optimized() || is_icstub) { | |
130 // Optimized call sites don't have a cache value and ICStub call | |
131 // sites only change the entry point. Changing the value in that | |
132 // case could lead to MT safety issues. | |
133 assert(cache == NULL, "must be null"); | |
134 return; | |
135 } | |
136 | |
137 if (cache == NULL) cache = (void*)Universe::non_oop_word(); | |
138 | |
139 _value->set_data((intptr_t)cache); | |
140 } | |
141 | |
142 | |
143 void CompiledIC::set_ic_destination(ICStub* stub) { | |
144 internal_set_ic_destination(stub->code_begin(), true, NULL, false); | |
145 } | |
146 | |
109 | 147 |
110 | 148 |
111 address CompiledIC::ic_destination() const { | 149 address CompiledIC::ic_destination() const { |
112 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | 150 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); |
113 if (!is_in_transition_state()) { | 151 if (!is_in_transition_state()) { |
122 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | 160 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); |
123 return InlineCacheBuffer::contains(_ic_call->destination()); | 161 return InlineCacheBuffer::contains(_ic_call->destination()); |
124 } | 162 } |
125 | 163 |
126 | 164 |
165 bool CompiledIC::is_icholder_call() const { | |
166 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | |
167 return !_is_optimized && is_icholder_entry(ic_destination()); | |
168 } | |
169 | |
127 // Returns native address of 'call' instruction in inline-cache. Used by | 170 // Returns native address of 'call' instruction in inline-cache. Used by |
128 // the InlineCacheBuffer when it needs to find the stub. | 171 // the InlineCacheBuffer when it needs to find the stub. |
129 address CompiledIC::stub_address() const { | 172 address CompiledIC::stub_address() const { |
130 assert(is_in_transition_state(), "should only be called when we are in a transition state"); | 173 assert(is_in_transition_state(), "should only be called when we are in a transition state"); |
131 return _ic_call->destination(); | 174 return _ic_call->destination(); |
138 | 181 |
139 void CompiledIC::set_to_megamorphic(CallInfo* call_info, Bytecodes::Code bytecode, TRAPS) { | 182 void CompiledIC::set_to_megamorphic(CallInfo* call_info, Bytecodes::Code bytecode, TRAPS) { |
140 methodHandle method = call_info->selected_method(); | 183 methodHandle method = call_info->selected_method(); |
141 bool is_invoke_interface = (bytecode == Bytecodes::_invokeinterface && !call_info->has_vtable_index()); | 184 bool is_invoke_interface = (bytecode == Bytecodes::_invokeinterface && !call_info->has_vtable_index()); |
142 assert(CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | 185 assert(CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); |
143 assert(method->is_oop(), "cannot be NULL and must be oop"); | |
144 assert(!is_optimized(), "cannot set an optimized virtual call to megamorphic"); | 186 assert(!is_optimized(), "cannot set an optimized virtual call to megamorphic"); |
145 assert(is_call_to_compiled() || is_call_to_interpreted(), "going directly to megamorphic?"); | 187 assert(is_call_to_compiled() || is_call_to_interpreted(), "going directly to megamorphic?"); |
146 | 188 |
147 address entry; | 189 address entry; |
148 if (is_invoke_interface) { | 190 if (is_invoke_interface) { |
149 int index = klassItable::compute_itable_index(call_info->resolved_method()()); | 191 int index = klassItable::compute_itable_index(call_info->resolved_method()()); |
150 entry = VtableStubs::create_stub(false, index, method()); | 192 entry = VtableStubs::create_stub(false, index, method()); |
151 assert(entry != NULL, "entry not computed"); | 193 assert(entry != NULL, "entry not computed"); |
152 klassOop k = call_info->resolved_method()->method_holder(); | 194 Klass* k = call_info->resolved_method()->method_holder(); |
153 assert(Klass::cast(k)->is_interface(), "sanity check"); | 195 assert(Klass::cast(k)->is_interface(), "sanity check"); |
154 InlineCacheBuffer::create_transition_stub(this, k, entry); | 196 InlineCacheBuffer::create_transition_stub(this, k, entry); |
155 } else { | 197 } else { |
156 // Can be different than method->vtable_index(), due to package-private etc. | 198 // Can be different than method->vtable_index(), due to package-private etc. |
157 int vtable_index = call_info->vtable_index(); | 199 int vtable_index = call_info->vtable_index(); |
178 // true if destination is megamorphic stub | 220 // true if destination is megamorphic stub |
179 bool CompiledIC::is_megamorphic() const { | 221 bool CompiledIC::is_megamorphic() const { |
180 assert(CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | 222 assert(CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); |
181 assert(!is_optimized(), "an optimized call cannot be megamorphic"); | 223 assert(!is_optimized(), "an optimized call cannot be megamorphic"); |
182 | 224 |
183 // Cannot rely on cached_oop. It is either an interface or a method. | 225 // Cannot rely on cached_value. It is either an interface or a method. |
184 return VtableStubs::is_entry_point(ic_destination()); | 226 return VtableStubs::is_entry_point(ic_destination()); |
185 } | 227 } |
186 | 228 |
187 bool CompiledIC::is_call_to_compiled() const { | 229 bool CompiledIC::is_call_to_compiled() const { |
188 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | 230 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); |
190 // Use unsafe, since an inline cache might point to a zombie method. However, the zombie | 232 // Use unsafe, since an inline cache might point to a zombie method. However, the zombie |
191 // method is guaranteed to still exist, since we only remove methods after all inline caches | 233 // method is guaranteed to still exist, since we only remove methods after all inline caches |
192 // has been cleaned up | 234 // has been cleaned up |
193 CodeBlob* cb = CodeCache::find_blob_unsafe(ic_destination()); | 235 CodeBlob* cb = CodeCache::find_blob_unsafe(ic_destination()); |
194 bool is_monomorphic = (cb != NULL && cb->is_nmethod()); | 236 bool is_monomorphic = (cb != NULL && cb->is_nmethod()); |
195 // Check that the cached_oop is a klass for non-optimized monomorphic calls | 237 // Check that the cached_value is a klass for non-optimized monomorphic calls |
196 // This assertion is invalid for compiler1: a call that does not look optimized (no static stub) can be used | 238 // This assertion is invalid for compiler1: a call that does not look optimized (no static stub) can be used |
197 // for calling directly to vep without using the inline cache (i.e., cached_oop == NULL) | 239 // for calling directly to vep without using the inline cache (i.e., cached_value == NULL) |
198 #ifdef ASSERT | 240 #ifdef ASSERT |
199 #ifdef TIERED | |
200 CodeBlob* caller = CodeCache::find_blob_unsafe(instruction_address()); | 241 CodeBlob* caller = CodeCache::find_blob_unsafe(instruction_address()); |
201 bool is_c1_method = caller->is_compiled_by_c1(); | 242 bool is_c1_method = caller->is_compiled_by_c1(); |
202 #else | |
203 #ifdef COMPILER1 | |
204 bool is_c1_method = true; | |
205 #else | |
206 bool is_c1_method = false; | |
207 #endif // COMPILER1 | |
208 #endif // TIERED | |
209 assert( is_c1_method || | 243 assert( is_c1_method || |
210 !is_monomorphic || | 244 !is_monomorphic || |
211 is_optimized() || | 245 is_optimized() || |
212 (cached_oop() != NULL && cached_oop()->is_klass()), "sanity check"); | 246 (cached_metadata() != NULL && cached_metadata()->is_klass()), "sanity check"); |
213 #endif // ASSERT | 247 #endif // ASSERT |
214 return is_monomorphic; | 248 return is_monomorphic; |
215 } | 249 } |
216 | 250 |
217 | 251 |
224 // must use unsafe because the destination can be a zombie (and we're cleaning) | 258 // must use unsafe because the destination can be a zombie (and we're cleaning) |
225 // and the print_compiled_ic code wants to know if site (in the non-zombie) | 259 // and the print_compiled_ic code wants to know if site (in the non-zombie) |
226 // is to the interpreter. | 260 // is to the interpreter. |
227 CodeBlob* cb = CodeCache::find_blob_unsafe(ic_destination()); | 261 CodeBlob* cb = CodeCache::find_blob_unsafe(ic_destination()); |
228 is_call_to_interpreted = (cb != NULL && cb->is_adapter_blob()); | 262 is_call_to_interpreted = (cb != NULL && cb->is_adapter_blob()); |
229 assert(!is_call_to_interpreted || (cached_oop() != NULL && cached_oop()->is_compiledICHolder()), "sanity check"); | 263 assert(!is_call_to_interpreted || (is_icholder_call() && cached_icholder() != NULL), "sanity check"); |
230 } else { | 264 } else { |
231 // Check if we are calling into our own codeblob (i.e., to a stub) | 265 // Check if we are calling into our own codeblob (i.e., to a stub) |
232 CodeBlob* cb = CodeCache::find_blob(_ic_call->instruction_address()); | 266 CodeBlob* cb = CodeCache::find_blob(_ic_call->instruction_address()); |
233 address dest = ic_destination(); | 267 address dest = ic_destination(); |
234 #ifdef ASSERT | 268 #ifdef ASSERT |
255 entry = SharedRuntime::get_resolve_opt_virtual_call_stub(); | 289 entry = SharedRuntime::get_resolve_opt_virtual_call_stub(); |
256 } else { | 290 } else { |
257 entry = SharedRuntime::get_resolve_virtual_call_stub(); | 291 entry = SharedRuntime::get_resolve_virtual_call_stub(); |
258 } | 292 } |
259 | 293 |
260 // A zombie transition will always be safe, since the oop has already been set to NULL, so | 294 // A zombie transition will always be safe, since the metadata has already been set to NULL, so |
261 // we only need to patch the destination | 295 // we only need to patch the destination |
262 bool safe_transition = is_optimized() || SafepointSynchronize::is_at_safepoint(); | 296 bool safe_transition = is_optimized() || SafepointSynchronize::is_at_safepoint(); |
263 | 297 |
264 if (safe_transition) { | 298 if (safe_transition) { |
265 if (!is_optimized()) set_cached_oop(NULL); | |
266 // Kill any leftover stub we might have too | 299 // Kill any leftover stub we might have too |
267 if (is_in_transition_state()) { | 300 if (is_in_transition_state()) { |
268 ICStub* old_stub = ICStub_from_destination_address(stub_address()); | 301 ICStub* old_stub = ICStub_from_destination_address(stub_address()); |
269 old_stub->clear(); | 302 old_stub->clear(); |
270 } | 303 } |
304 if (is_optimized()) { | |
271 set_ic_destination(entry); | 305 set_ic_destination(entry); |
306 } else { | |
307 set_ic_destination_and_value(entry, (void*)NULL); | |
308 } | |
272 } else { | 309 } else { |
273 // Unsafe transition - create stub. | 310 // Unsafe transition - create stub. |
274 InlineCacheBuffer::create_transition_stub(this, NULL, entry); | 311 InlineCacheBuffer::create_transition_stub(this, NULL, entry); |
275 } | 312 } |
276 // We can't check this anymore. With lazy deopt we could have already | 313 // We can't check this anymore. With lazy deopt we could have already |
287 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | 324 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); |
288 bool is_clean = false; | 325 bool is_clean = false; |
289 address dest = ic_destination(); | 326 address dest = ic_destination(); |
290 is_clean = dest == SharedRuntime::get_resolve_opt_virtual_call_stub() || | 327 is_clean = dest == SharedRuntime::get_resolve_opt_virtual_call_stub() || |
291 dest == SharedRuntime::get_resolve_virtual_call_stub(); | 328 dest == SharedRuntime::get_resolve_virtual_call_stub(); |
292 assert(!is_clean || is_optimized() || cached_oop() == NULL, "sanity check"); | 329 assert(!is_clean || is_optimized() || cached_value() == NULL, "sanity check"); |
293 return is_clean; | 330 return is_clean; |
294 } | 331 } |
295 | 332 |
296 | 333 |
297 void CompiledIC::set_to_monomorphic(const CompiledICInfo& info) { | 334 void CompiledIC::set_to_monomorphic(CompiledICInfo& info) { |
298 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); | 335 assert (CompiledIC_lock->is_locked() || SafepointSynchronize::is_at_safepoint(), ""); |
299 // Updating a cache to the wrong entry can cause bugs that are very hard | 336 // Updating a cache to the wrong entry can cause bugs that are very hard |
300 // to track down - if cache entry gets invalid - we just clean it. In | 337 // to track down - if cache entry gets invalid - we just clean it. In |
301 // this way it is always the same code path that is responsible for | 338 // this way it is always the same code path that is responsible for |
302 // updating and resolving an inline cache | 339 // updating and resolving an inline cache |
307 // | 344 // |
308 // In both of these cases the only thing being modifed is the jump/call target and these | 345 // In both of these cases the only thing being modifed is the jump/call target and these |
309 // transitions are mt_safe | 346 // transitions are mt_safe |
310 | 347 |
311 Thread *thread = Thread::current(); | 348 Thread *thread = Thread::current(); |
312 if (info._to_interpreter) { | 349 if (info.to_interpreter()) { |
313 // Call to interpreter | 350 // Call to interpreter |
314 if (info.is_optimized() && is_optimized()) { | 351 if (info.is_optimized() && is_optimized()) { |
315 assert(is_clean(), "unsafe IC path"); | 352 assert(is_clean(), "unsafe IC path"); |
316 MutexLockerEx pl(Patching_lock, Mutex::_no_safepoint_check_flag); | 353 MutexLockerEx pl(Patching_lock, Mutex::_no_safepoint_check_flag); |
317 // the call analysis (callee structure) specifies that the call is optimized | 354 // the call analysis (callee structure) specifies that the call is optimized |
318 // (either because of CHA or the static target is final) | 355 // (either because of CHA or the static target is final) |
319 // At code generation time, this call has been emitted as static call | 356 // At code generation time, this call has been emitted as static call |
320 // Call via stub | 357 // Call via stub |
321 assert(info.cached_oop().not_null() && info.cached_oop()->is_method(), "sanity check"); | 358 assert(info.cached_metadata() != NULL && info.cached_metadata()->is_method(), "sanity check"); |
322 CompiledStaticCall* csc = compiledStaticCall_at(instruction_address()); | 359 CompiledStaticCall* csc = compiledStaticCall_at(instruction_address()); |
323 methodHandle method (thread, (methodOop)info.cached_oop()()); | 360 methodHandle method (thread, (Method*)info.cached_metadata()); |
324 csc->set_to_interpreted(method, info.entry()); | 361 csc->set_to_interpreted(method, info.entry()); |
325 if (TraceICs) { | 362 if (TraceICs) { |
326 ResourceMark rm(thread); | 363 ResourceMark rm(thread); |
327 tty->print_cr ("IC@" INTPTR_FORMAT ": monomorphic to interpreter: %s", | 364 tty->print_cr ("IC@" INTPTR_FORMAT ": monomorphic to interpreter: %s", |
328 instruction_address(), | 365 instruction_address(), |
329 method->print_value_string()); | 366 method->print_value_string()); |
330 } | 367 } |
331 } else { | 368 } else { |
332 // Call via method-klass-holder | 369 // Call via method-klass-holder |
333 assert(info.cached_oop().not_null(), "must be set"); | 370 InlineCacheBuffer::create_transition_stub(this, info.claim_cached_icholder(), info.entry()); |
334 InlineCacheBuffer::create_transition_stub(this, info.cached_oop()(), info.entry()); | |
335 | |
336 if (TraceICs) { | 371 if (TraceICs) { |
337 ResourceMark rm(thread); | 372 ResourceMark rm(thread); |
338 tty->print_cr ("IC@" INTPTR_FORMAT ": monomorphic to interpreter via mkh", instruction_address()); | 373 tty->print_cr ("IC@" INTPTR_FORMAT ": monomorphic to interpreter via icholder ", instruction_address()); |
339 } | 374 } |
340 } | 375 } |
341 } else { | 376 } else { |
342 // Call to compiled code | 377 // Call to compiled code |
343 bool static_bound = info.is_optimized() || (info.cached_oop().is_null()); | 378 bool static_bound = info.is_optimized() || (info.cached_metadata() == NULL); |
344 #ifdef ASSERT | 379 #ifdef ASSERT |
345 CodeBlob* cb = CodeCache::find_blob_unsafe(info.entry()); | 380 CodeBlob* cb = CodeCache::find_blob_unsafe(info.entry()); |
346 assert (cb->is_nmethod(), "must be compiled!"); | 381 assert (cb->is_nmethod(), "must be compiled!"); |
347 #endif /* ASSERT */ | 382 #endif /* ASSERT */ |
348 | 383 |
350 // non-verified entry point | 385 // non-verified entry point |
351 bool safe = SafepointSynchronize::is_at_safepoint() || | 386 bool safe = SafepointSynchronize::is_at_safepoint() || |
352 (!is_in_transition_state() && (info.is_optimized() || static_bound || is_clean())); | 387 (!is_in_transition_state() && (info.is_optimized() || static_bound || is_clean())); |
353 | 388 |
354 if (!safe) { | 389 if (!safe) { |
355 InlineCacheBuffer::create_transition_stub(this, info.cached_oop()(), info.entry()); | 390 InlineCacheBuffer::create_transition_stub(this, info.cached_metadata(), info.entry()); |
356 } else { | 391 } else { |
392 if (is_optimized()) { | |
357 set_ic_destination(info.entry()); | 393 set_ic_destination(info.entry()); |
358 if (!is_optimized()) set_cached_oop(info.cached_oop()()); | 394 } else { |
395 set_ic_destination_and_value(info.entry(), info.cached_metadata()); | |
396 } | |
359 } | 397 } |
360 | 398 |
361 if (TraceICs) { | 399 if (TraceICs) { |
362 ResourceMark rm(thread); | 400 ResourceMark rm(thread); |
363 assert(info.cached_oop() == NULL || info.cached_oop()()->is_klass(), "must be"); | 401 assert(info.cached_metadata() == NULL || info.cached_metadata()->is_klass(), "must be"); |
364 tty->print_cr ("IC@" INTPTR_FORMAT ": monomorphic to compiled (rcvr klass) %s: %s", | 402 tty->print_cr ("IC@" INTPTR_FORMAT ": monomorphic to compiled (rcvr klass) %s: %s", |
365 instruction_address(), | 403 instruction_address(), |
366 ((klassOop)info.cached_oop()())->print_value_string(), | 404 ((Klass*)info.cached_metadata())->print_value_string(), |
367 (safe) ? "" : "via stub"); | 405 (safe) ? "" : "via stub"); |
368 } | 406 } |
369 } | 407 } |
370 // We can't check this anymore. With lazy deopt we could have already | 408 // We can't check this anymore. With lazy deopt we could have already |
371 // cleaned this IC entry before we even return. This is possible if | 409 // cleaned this IC entry before we even return. This is possible if |
384 KlassHandle receiver_klass, | 422 KlassHandle receiver_klass, |
385 bool is_optimized, | 423 bool is_optimized, |
386 bool static_bound, | 424 bool static_bound, |
387 CompiledICInfo& info, | 425 CompiledICInfo& info, |
388 TRAPS) { | 426 TRAPS) { |
389 info._is_optimized = is_optimized; | |
390 | |
391 nmethod* method_code = method->code(); | 427 nmethod* method_code = method->code(); |
392 address entry = NULL; | 428 address entry = NULL; |
393 if (method_code != NULL) { | 429 if (method_code != NULL) { |
394 // Call to compiled code | 430 // Call to compiled code |
395 if (static_bound || is_optimized) { | 431 if (static_bound || is_optimized) { |
398 entry = method_code->entry_point(); | 434 entry = method_code->entry_point(); |
399 } | 435 } |
400 } | 436 } |
401 if (entry != NULL) { | 437 if (entry != NULL) { |
402 // Call to compiled code | 438 // Call to compiled code |
403 info._entry = entry; | 439 info.set_compiled_entry(entry, (static_bound || is_optimized) ? NULL : receiver_klass(), is_optimized); |
404 if (static_bound || is_optimized) { | |
405 info._cached_oop = Handle(THREAD, (oop)NULL); | |
406 } else { | |
407 info._cached_oop = receiver_klass; | |
408 } | |
409 info._to_interpreter = false; | |
410 } else { | 440 } else { |
411 // Note: the following problem exists with Compiler1: | 441 // Note: the following problem exists with Compiler1: |
412 // - at compile time we may or may not know if the destination is final | 442 // - at compile time we may or may not know if the destination is final |
413 // - if we know that the destination is final, we will emit an optimized | 443 // - if we know that the destination is final, we will emit an optimized |
414 // virtual call (no inline cache), and need a methodOop to make a call | 444 // virtual call (no inline cache), and need a Method* to make a call |
415 // to the interpreter | 445 // to the interpreter |
416 // - if we do not know if the destination is final, we emit a standard | 446 // - if we do not know if the destination is final, we emit a standard |
417 // virtual call, and use CompiledICHolder to call interpreted code | 447 // virtual call, and use CompiledICHolder to call interpreted code |
418 // (no static call stub has been generated) | 448 // (no static call stub has been generated) |
419 // However in that case we will now notice it is static_bound | 449 // However in that case we will now notice it is static_bound |
420 // and convert the call into what looks to be an optimized | 450 // and convert the call into what looks to be an optimized |
421 // virtual call. This causes problems in verifying the IC because | 451 // virtual call. This causes problems in verifying the IC because |
422 // it look vanilla but is optimized. Code in is_call_to_interpreted | 452 // it look vanilla but is optimized. Code in is_call_to_interpreted |
423 // is aware of this and weakens its asserts. | 453 // is aware of this and weakens its asserts. |
424 | 454 |
425 info._to_interpreter = true; | |
426 // static_bound should imply is_optimized -- otherwise we have a | 455 // static_bound should imply is_optimized -- otherwise we have a |
427 // performance bug (statically-bindable method is called via | 456 // performance bug (statically-bindable method is called via |
428 // dynamically-dispatched call note: the reverse implication isn't | 457 // dynamically-dispatched call note: the reverse implication isn't |
429 // necessarily true -- the call may have been optimized based on compiler | 458 // necessarily true -- the call may have been optimized based on compiler |
430 // analysis (static_bound is only based on "final" etc.) | 459 // analysis (static_bound is only based on "final" etc.) |
441 assert(!static_bound || is_optimized, "static_bound should imply is_optimized"); | 470 assert(!static_bound || is_optimized, "static_bound should imply is_optimized"); |
442 #endif // TIERED | 471 #endif // TIERED |
443 #endif // COMPILER2 | 472 #endif // COMPILER2 |
444 if (is_optimized) { | 473 if (is_optimized) { |
445 // Use stub entry | 474 // Use stub entry |
446 info._entry = method()->get_c2i_entry(); | 475 info.set_interpreter_entry(method()->get_c2i_entry(), method()); |
447 info._cached_oop = method; | |
448 } else { | 476 } else { |
449 // Use mkh entry | 477 // Use icholder entry |
450 oop holder = oopFactory::new_compiledICHolder(method, receiver_klass, CHECK); | 478 CompiledICHolder* holder = new CompiledICHolder(method(), receiver_klass()); |
451 info._cached_oop = Handle(THREAD, holder); | 479 info.set_icholder_entry(method()->get_c2i_unverified_entry(), holder); |
452 info._entry = method()->get_c2i_unverified_entry(); | 480 } |
453 } | 481 } |
454 } | 482 assert(info.is_optimized() == is_optimized, "must agree"); |
455 } | 483 } |
456 | 484 |
457 | 485 |
458 inline static RelocIterator parse_ic(nmethod* nm, address ic_call, oop* &_oop_addr, bool *is_optimized) { | 486 bool CompiledIC::is_icholder_entry(address entry) { |
459 address first_oop = NULL; | 487 CodeBlob* cb = CodeCache::find_blob_unsafe(entry); |
460 // Mergers please note: Sun SC5.x CC insists on an lvalue for a reference parameter. | 488 return (cb != NULL && cb->is_adapter_blob()); |
461 nmethod* tmp_nm = nm; | 489 } |
462 return virtual_call_Relocation::parse_ic(tmp_nm, ic_call, first_oop, _oop_addr, is_optimized); | 490 |
463 } | 491 |
464 | 492 CompiledIC::CompiledIC(nmethod* nm, NativeCall* call) |
465 CompiledIC::CompiledIC(NativeCall* ic_call) | 493 : _ic_call(call) |
466 : _ic_call(ic_call), | |
467 _oops(parse_ic(NULL, ic_call->instruction_address(), _oop_addr, &_is_optimized)) | |
468 { | 494 { |
469 } | 495 address ic_call = call->instruction_address(); |
470 | 496 |
471 | 497 assert(ic_call != NULL, "ic_call address must be set"); |
472 CompiledIC::CompiledIC(Relocation* ic_reloc) | 498 assert(nm != NULL, "must pass nmethod"); |
473 : _ic_call(nativeCall_at(ic_reloc->addr())), | 499 assert(nm->contains(ic_call), "must be in nmethod"); |
474 _oops(parse_ic(ic_reloc->code(), ic_reloc->addr(), _oop_addr, &_is_optimized)) | 500 |
475 { | 501 // search for the ic_call at the given address |
476 assert(ic_reloc->type() == relocInfo::virtual_call_type || | 502 RelocIterator iter(nm, ic_call, ic_call+1); |
477 ic_reloc->type() == relocInfo::opt_virtual_call_type, "wrong reloc. info"); | 503 bool ret = iter.next(); |
504 assert(ret == true, "relocInfo must exist at this address"); | |
505 assert(iter.addr() == ic_call, "must find ic_call"); | |
506 if (iter.type() == relocInfo::virtual_call_type) { | |
507 virtual_call_Relocation* r = iter.virtual_call_reloc(); | |
508 _is_optimized = false; | |
509 _value = nativeMovConstReg_at(r->cached_value()); | |
510 } else { | |
511 assert(iter.type() == relocInfo::opt_virtual_call_type, "must be a virtual call"); | |
512 _is_optimized = true; | |
513 _value = NULL; | |
514 } | |
478 } | 515 } |
479 | 516 |
480 | 517 |
481 // ---------------------------------------------------------------------------- | 518 // ---------------------------------------------------------------------------- |
482 | 519 |
637 tty->cr(); | 674 tty->cr(); |
638 } | 675 } |
639 | 676 |
640 | 677 |
641 void CompiledIC::print_compiled_ic() { | 678 void CompiledIC::print_compiled_ic() { |
642 tty->print("Inline cache at " INTPTR_FORMAT ", calling %s " INTPTR_FORMAT, | 679 tty->print("Inline cache at " INTPTR_FORMAT ", calling %s " INTPTR_FORMAT " cached_value " INTPTR_FORMAT, |
643 instruction_address(), is_call_to_interpreted() ? "interpreted " : "", ic_destination()); | 680 instruction_address(), is_call_to_interpreted() ? "interpreted " : "", ic_destination(), is_optimized() ? NULL : cached_value()); |
644 } | 681 } |
645 | 682 |
646 | 683 |
647 void CompiledStaticCall::print() { | 684 void CompiledStaticCall::print() { |
648 tty->print("static call at " INTPTR_FORMAT " -> ", instruction_address()); | 685 tty->print("static call at " INTPTR_FORMAT " -> ", instruction_address()); |