Mercurial > hg > graal-compiler
comparison src/share/vm/oops/method.hpp @ 6725:da91efe96a93
6964458: Reimplement class meta-data storage to use native memory
Summary: Remove PermGen, allocate meta-data in metaspace linked to class loaders, rewrite GC walking, rewrite and rename metadata to be C++ classes
Reviewed-by: jmasa, stefank, never, coleenp, kvn, brutisso, mgerdin, dholmes, jrose, twisti, roland
Contributed-by: jmasa <jon.masamitsu@oracle.com>, stefank <stefan.karlsson@oracle.com>, mgerdin <mikael.gerdin@oracle.com>, never <tom.rodriguez@oracle.com>
author | coleenp |
---|---|
date | Sat, 01 Sep 2012 13:25:18 -0400 |
parents | src/share/vm/oops/methodOop.hpp@93c71eb28866 |
children | fa6e618671d7 |
comparison
equal
deleted
inserted
replaced
6724:36d1d483d5d6 | 6725:da91efe96a93 |
---|---|
1 /* | |
2 * Copyright (c) 1997, 2012, Oracle and/or its affiliates. All rights reserved. | |
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. | |
4 * | |
5 * This code is free software; you can redistribute it and/or modify it | |
6 * under the terms of the GNU General Public License version 2 only, as | |
7 * published by the Free Software Foundation. | |
8 * | |
9 * This code is distributed in the hope that it will be useful, but WITHOUT | |
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License | |
12 * version 2 for more details (a copy is included in the LICENSE file that | |
13 * accompanied this code). | |
14 * | |
15 * You should have received a copy of the GNU General Public License version | |
16 * 2 along with this work; if not, write to the Free Software Foundation, | |
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | |
18 * | |
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA | |
20 * or visit www.oracle.com if you need additional information or have any | |
21 * questions. | |
22 * | |
23 */ | |
24 | |
25 #ifndef SHARE_VM_OOPS_METHODOOP_HPP | |
26 #define SHARE_VM_OOPS_METHODOOP_HPP | |
27 | |
28 #include "classfile/vmSymbols.hpp" | |
29 #include "code/compressedStream.hpp" | |
30 #include "compiler/oopMap.hpp" | |
31 #include "interpreter/invocationCounter.hpp" | |
32 #include "oops/annotations.hpp" | |
33 #include "oops/constMethod.hpp" | |
34 #include "oops/constantPool.hpp" | |
35 #include "oops/instanceKlass.hpp" | |
36 #include "oops/oop.hpp" | |
37 #include "oops/typeArrayOop.hpp" | |
38 #include "utilities/accessFlags.hpp" | |
39 #include "utilities/growableArray.hpp" | |
40 | |
41 // A Method* represents a Java method. | |
42 // | |
43 // Memory layout (each line represents a word). Note that most applications load thousands of methods, | |
44 // so keeping the size of this structure small has a big impact on footprint. | |
45 // | |
46 // We put all oops and method_size first for better gc cache locality. | |
47 // | |
48 // The actual bytecodes are inlined after the end of the Method struct. | |
49 // | |
50 // There are bits in the access_flags telling whether inlined tables are present. | |
51 // Note that accessing the line number and local variable tables is not performance critical at all. | |
52 // Accessing the checked exceptions table is used by reflection, so we put that last to make access | |
53 // to it fast. | |
54 // | |
55 // The line number table is compressed and inlined following the byte codes. It is found as the first | |
56 // byte following the byte codes. The checked exceptions table and the local variable table are inlined | |
57 // after the line number table, and indexed from the end of the method. We do not compress the checked | |
58 // exceptions table since the average length is less than 2, and do not bother to compress the local | |
59 // variable table either since it is mostly absent. | |
60 // | |
61 // Note that native_function and signature_handler has to be at fixed offsets (required by the interpreter) | |
62 // | |
63 // |------------------------------------------------------| | |
64 // | header | | |
65 // | klass | | |
66 // |------------------------------------------------------| | |
67 // | ConstMethod* (oop) | | |
68 // |------------------------------------------------------| | |
69 // | methodData (oop) | | |
70 // | interp_invocation_count | | |
71 // |------------------------------------------------------| | |
72 // | access_flags | | |
73 // | vtable_index | | |
74 // |------------------------------------------------------| | |
75 // | result_index (C++ interpreter only) | | |
76 // |------------------------------------------------------| | |
77 // | method_size | max_stack | | |
78 // | max_locals | size_of_parameters | | |
79 // |------------------------------------------------------| | |
80 // |intrinsic_id| flags | throwout_count | | |
81 // |------------------------------------------------------| | |
82 // | num_breakpoints | (unused) | | |
83 // |------------------------------------------------------| | |
84 // | invocation_counter | | |
85 // | backedge_counter | | |
86 // |------------------------------------------------------| | |
87 // | prev_time (tiered only, 64 bit wide) | | |
88 // | | | |
89 // |------------------------------------------------------| | |
90 // | rate (tiered) | | |
91 // |------------------------------------------------------| | |
92 // | code (pointer) | | |
93 // | i2i (pointer) | | |
94 // | adapter (pointer) | | |
95 // | from_compiled_entry (pointer) | | |
96 // | from_interpreted_entry (pointer) | | |
97 // |------------------------------------------------------| | |
98 // | native_function (present only if native) | | |
99 // | signature_handler (present only if native) | | |
100 // |------------------------------------------------------| | |
101 | |
102 | |
103 class CheckedExceptionElement; | |
104 class LocalVariableTableElement; | |
105 class AdapterHandlerEntry; | |
106 class MethodData; | |
107 | |
108 class Method : public Metadata { | |
109 friend class VMStructs; | |
110 private: | |
111 ConstMethod* _constMethod; // Method read-only data. | |
112 MethodData* _method_data; | |
113 int _interpreter_invocation_count; // Count of times invoked (reused as prev_event_count in tiered) | |
114 AccessFlags _access_flags; // Access flags | |
115 int _vtable_index; // vtable index of this method (see VtableIndexFlag) | |
116 // note: can have vtables with >2**16 elements (because of inheritance) | |
117 #ifdef CC_INTERP | |
118 int _result_index; // C++ interpreter needs for converting results to/from stack | |
119 #endif | |
120 u2 _method_size; // size of this object | |
121 u2 _max_stack; // Maximum number of entries on the expression stack | |
122 u2 _max_locals; // Number of local variables used by this method | |
123 u2 _size_of_parameters; // size of the parameter block (receiver + arguments) in words | |
124 u1 _intrinsic_id; // vmSymbols::intrinsic_id (0 == _none) | |
125 u1 _jfr_towrite : 1, // Flags | |
126 _force_inline : 1, | |
127 _hidden : 1, | |
128 _dont_inline : 1, | |
129 : 4; | |
130 u2 _interpreter_throwout_count; // Count of times method was exited via exception while interpreting | |
131 u2 _number_of_breakpoints; // fullspeed debugging support | |
132 InvocationCounter _invocation_counter; // Incremented before each activation of the method - used to trigger frequency-based optimizations | |
133 InvocationCounter _backedge_counter; // Incremented before each backedge taken - used to trigger frequencey-based optimizations | |
134 | |
135 #ifdef TIERED | |
136 jlong _prev_time; // Previous time the rate was acquired | |
137 float _rate; // Events (invocation and backedge counter increments) per millisecond | |
138 #endif | |
139 | |
140 #ifndef PRODUCT | |
141 int _compiled_invocation_count; // Number of nmethod invocations so far (for perf. debugging) | |
142 #endif | |
143 // Entry point for calling both from and to the interpreter. | |
144 address _i2i_entry; // All-args-on-stack calling convention | |
145 // Adapter blob (i2c/c2i) for this Method*. Set once when method is linked. | |
146 AdapterHandlerEntry* _adapter; | |
147 // Entry point for calling from compiled code, to compiled code if it exists | |
148 // or else the interpreter. | |
149 volatile address _from_compiled_entry; // Cache of: _code ? _code->entry_point() : _adapter->c2i_entry() | |
150 // The entry point for calling both from and to compiled code is | |
151 // "_code->entry_point()". Because of tiered compilation and de-opt, this | |
152 // field can come and go. It can transition from NULL to not-null at any | |
153 // time (whenever a compile completes). It can transition from not-null to | |
154 // NULL only at safepoints (because of a de-opt). | |
155 nmethod* volatile _code; // Points to the corresponding piece of native code | |
156 volatile address _from_interpreted_entry; // Cache of _code ? _adapter->i2c_entry() : _i2i_entry | |
157 | |
158 // Constructor | |
159 Method(ConstMethod* xconst, AccessFlags access_flags, int size); | |
160 public: | |
161 static Method* allocate(ClassLoaderData* loader_data, | |
162 int byte_code_size, | |
163 AccessFlags access_flags, | |
164 int compressed_line_number_size, | |
165 int localvariable_table_length, | |
166 int exception_table_length, | |
167 int checked_exceptions_length, | |
168 TRAPS); | |
169 | |
170 Method() { assert(DumpSharedSpaces || UseSharedSpaces, "only for CDS"); } | |
171 bool is_method() const volatile { return true; } | |
172 | |
173 // accessors for instance variables | |
174 ConstMethod* constMethod() const { return _constMethod; } | |
175 void set_constMethod(ConstMethod* xconst) { _constMethod = xconst; } | |
176 | |
177 | |
178 static address make_adapters(methodHandle mh, TRAPS); | |
179 volatile address from_compiled_entry() const { return (address)OrderAccess::load_ptr_acquire(&_from_compiled_entry); } | |
180 volatile address from_interpreted_entry() const{ return (address)OrderAccess::load_ptr_acquire(&_from_interpreted_entry); } | |
181 | |
182 // access flag | |
183 AccessFlags access_flags() const { return _access_flags; } | |
184 void set_access_flags(AccessFlags flags) { _access_flags = flags; } | |
185 | |
186 // name | |
187 Symbol* name() const { return constants()->symbol_at(name_index()); } | |
188 int name_index() const { return constMethod()->name_index(); } | |
189 void set_name_index(int index) { constMethod()->set_name_index(index); } | |
190 | |
191 // signature | |
192 Symbol* signature() const { return constants()->symbol_at(signature_index()); } | |
193 int signature_index() const { return constMethod()->signature_index(); } | |
194 void set_signature_index(int index) { constMethod()->set_signature_index(index); } | |
195 | |
196 // generics support | |
197 Symbol* generic_signature() const { int idx = generic_signature_index(); return ((idx != 0) ? constants()->symbol_at(idx) : (Symbol*)NULL); } | |
198 int generic_signature_index() const { return constMethod()->generic_signature_index(); } | |
199 void set_generic_signature_index(int index) { constMethod()->set_generic_signature_index(index); } | |
200 | |
201 // annotations support | |
202 AnnotationArray* annotations() const { | |
203 InstanceKlass* ik = InstanceKlass::cast(method_holder()); | |
204 if (ik->annotations() == NULL) { | |
205 return NULL; | |
206 } | |
207 return ik->annotations()->get_method_annotations_of(method_idnum()); | |
208 } | |
209 AnnotationArray* parameter_annotations() const { | |
210 InstanceKlass* ik = InstanceKlass::cast(method_holder()); | |
211 if (ik->annotations() == NULL) { | |
212 return NULL; | |
213 } | |
214 return ik->annotations()->get_method_parameter_annotations_of(method_idnum()); | |
215 } | |
216 AnnotationArray* annotation_default() const { | |
217 InstanceKlass* ik = InstanceKlass::cast(method_holder()); | |
218 if (ik->annotations() == NULL) { | |
219 return NULL; | |
220 } | |
221 return ik->annotations()->get_method_default_annotations_of(method_idnum()); | |
222 } | |
223 | |
224 #ifdef CC_INTERP | |
225 void set_result_index(BasicType type); | |
226 int result_index() { return _result_index; } | |
227 #endif | |
228 | |
229 // Helper routine: get klass name + "." + method name + signature as | |
230 // C string, for the purpose of providing more useful NoSuchMethodErrors | |
231 // and fatal error handling. The string is allocated in resource | |
232 // area if a buffer is not provided by the caller. | |
233 char* name_and_sig_as_C_string() const; | |
234 char* name_and_sig_as_C_string(char* buf, int size) const; | |
235 | |
236 // Static routine in the situations we don't have a Method* | |
237 static char* name_and_sig_as_C_string(Klass* klass, Symbol* method_name, Symbol* signature); | |
238 static char* name_and_sig_as_C_string(Klass* klass, Symbol* method_name, Symbol* signature, char* buf, int size); | |
239 | |
240 Bytecodes::Code java_code_at(int bci) const { | |
241 return Bytecodes::java_code_at(this, bcp_from(bci)); | |
242 } | |
243 Bytecodes::Code code_at(int bci) const { | |
244 return Bytecodes::code_at(this, bcp_from(bci)); | |
245 } | |
246 | |
247 // JVMTI breakpoints | |
248 Bytecodes::Code orig_bytecode_at(int bci) const; | |
249 void set_orig_bytecode_at(int bci, Bytecodes::Code code); | |
250 void set_breakpoint(int bci); | |
251 void clear_breakpoint(int bci); | |
252 void clear_all_breakpoints(); | |
253 // Tracking number of breakpoints, for fullspeed debugging. | |
254 // Only mutated by VM thread. | |
255 u2 number_of_breakpoints() const { return _number_of_breakpoints; } | |
256 void incr_number_of_breakpoints() { ++_number_of_breakpoints; } | |
257 void decr_number_of_breakpoints() { --_number_of_breakpoints; } | |
258 // Initialization only | |
259 void clear_number_of_breakpoints() { _number_of_breakpoints = 0; } | |
260 | |
261 // index into InstanceKlass methods() array | |
262 // note: also used by jfr | |
263 u2 method_idnum() const { return constMethod()->method_idnum(); } | |
264 void set_method_idnum(u2 idnum) { constMethod()->set_method_idnum(idnum); } | |
265 | |
266 // code size | |
267 int code_size() const { return constMethod()->code_size(); } | |
268 | |
269 // method size | |
270 int method_size() const { return _method_size; } | |
271 void set_method_size(int size) { | |
272 assert(0 <= size && size < (1 << 16), "invalid method size"); | |
273 _method_size = size; | |
274 } | |
275 | |
276 // constant pool for Klass* holding this method | |
277 ConstantPool* constants() const { return constMethod()->constants(); } | |
278 void set_constants(ConstantPool* c) { constMethod()->set_constants(c); } | |
279 | |
280 // max stack | |
281 // return original max stack size for method verification | |
282 int verifier_max_stack() const { return _max_stack; } | |
283 int max_stack() const { return _max_stack + extra_stack_entries(); } | |
284 void set_max_stack(int size) { _max_stack = size; } | |
285 | |
286 // max locals | |
287 int max_locals() const { return _max_locals; } | |
288 void set_max_locals(int size) { _max_locals = size; } | |
289 | |
290 int highest_comp_level() const; | |
291 void set_highest_comp_level(int level); | |
292 int highest_osr_comp_level() const; | |
293 void set_highest_osr_comp_level(int level); | |
294 | |
295 // Count of times method was exited via exception while interpreting | |
296 void interpreter_throwout_increment() { | |
297 if (_interpreter_throwout_count < 65534) { | |
298 _interpreter_throwout_count++; | |
299 } | |
300 } | |
301 | |
302 int interpreter_throwout_count() const { return _interpreter_throwout_count; } | |
303 void set_interpreter_throwout_count(int count) { _interpreter_throwout_count = count; } | |
304 | |
305 // size of parameters | |
306 int size_of_parameters() const { return _size_of_parameters; } | |
307 | |
308 bool has_stackmap_table() const { | |
309 return constMethod()->has_stackmap_table(); | |
310 } | |
311 | |
312 Array<u1>* stackmap_data() const { | |
313 return constMethod()->stackmap_data(); | |
314 } | |
315 | |
316 void set_stackmap_data(Array<u1>* sd) { | |
317 constMethod()->set_stackmap_data(sd); | |
318 } | |
319 | |
320 // exception handler table | |
321 bool has_exception_handler() const | |
322 { return constMethod()->has_exception_handler(); } | |
323 int exception_table_length() const | |
324 { return constMethod()->exception_table_length(); } | |
325 ExceptionTableElement* exception_table_start() const | |
326 { return constMethod()->exception_table_start(); } | |
327 | |
328 // Finds the first entry point bci of an exception handler for an | |
329 // exception of klass ex_klass thrown at throw_bci. A value of NULL | |
330 // for ex_klass indicates that the exception klass is not known; in | |
331 // this case it matches any constraint class. Returns -1 if the | |
332 // exception cannot be handled in this method. The handler | |
333 // constraint classes are loaded if necessary. Note that this may | |
334 // throw an exception if loading of the constraint classes causes | |
335 // an IllegalAccessError (bugid 4307310) or an OutOfMemoryError. | |
336 // If an exception is thrown, returns the bci of the | |
337 // exception handler which caused the exception to be thrown, which | |
338 // is needed for proper retries. See, for example, | |
339 // InterpreterRuntime::exception_handler_for_exception. | |
340 int fast_exception_handler_bci_for(KlassHandle ex_klass, int throw_bci, TRAPS); | |
341 | |
342 // method data access | |
343 MethodData* method_data() const { | |
344 return _method_data; | |
345 } | |
346 void set_method_data(MethodData* data) { | |
347 _method_data = data; | |
348 } | |
349 | |
350 // invocation counter | |
351 InvocationCounter* invocation_counter() { return &_invocation_counter; } | |
352 InvocationCounter* backedge_counter() { return &_backedge_counter; } | |
353 | |
354 #ifdef TIERED | |
355 // We are reusing interpreter_invocation_count as a holder for the previous event count! | |
356 // We can do that since interpreter_invocation_count is not used in tiered. | |
357 int prev_event_count() const { return _interpreter_invocation_count; } | |
358 void set_prev_event_count(int count) { _interpreter_invocation_count = count; } | |
359 jlong prev_time() const { return _prev_time; } | |
360 void set_prev_time(jlong time) { _prev_time = time; } | |
361 float rate() const { return _rate; } | |
362 void set_rate(float rate) { _rate = rate; } | |
363 #endif | |
364 | |
365 int invocation_count(); | |
366 int backedge_count(); | |
367 | |
368 bool was_executed_more_than(int n); | |
369 bool was_never_executed() { return !was_executed_more_than(0); } | |
370 | |
371 static void build_interpreter_method_data(methodHandle method, TRAPS); | |
372 | |
373 int interpreter_invocation_count() { | |
374 if (TieredCompilation) return invocation_count(); | |
375 else return _interpreter_invocation_count; | |
376 } | |
377 void set_interpreter_invocation_count(int count) { _interpreter_invocation_count = count; } | |
378 int increment_interpreter_invocation_count() { | |
379 if (TieredCompilation) ShouldNotReachHere(); | |
380 return ++_interpreter_invocation_count; | |
381 } | |
382 | |
383 #ifndef PRODUCT | |
384 int compiled_invocation_count() const { return _compiled_invocation_count; } | |
385 void set_compiled_invocation_count(int count) { _compiled_invocation_count = count; } | |
386 #endif // not PRODUCT | |
387 | |
388 // Clear (non-shared space) pointers which could not be relevant | |
389 // if this (shared) method were mapped into another JVM. | |
390 void remove_unshareable_info(); | |
391 | |
392 // nmethod/verified compiler entry | |
393 address verified_code_entry(); | |
394 bool check_code() const; // Not inline to avoid circular ref | |
395 nmethod* volatile code() const { assert( check_code(), "" ); return (nmethod *)OrderAccess::load_ptr_acquire(&_code); } | |
396 void clear_code(); // Clear out any compiled code | |
397 static void set_code(methodHandle mh, nmethod* code); | |
398 void set_adapter_entry(AdapterHandlerEntry* adapter) { _adapter = adapter; } | |
399 address get_i2c_entry(); | |
400 address get_c2i_entry(); | |
401 address get_c2i_unverified_entry(); | |
402 AdapterHandlerEntry* adapter() { return _adapter; } | |
403 // setup entry points | |
404 void link_method(methodHandle method, TRAPS); | |
405 // clear entry points. Used by sharing code | |
406 void unlink_method(); | |
407 | |
408 // vtable index | |
409 enum VtableIndexFlag { | |
410 // Valid vtable indexes are non-negative (>= 0). | |
411 // These few negative values are used as sentinels. | |
412 highest_unused_vtable_index_value = -5, | |
413 invalid_vtable_index = -4, // distinct from any valid vtable index | |
414 garbage_vtable_index = -3, // not yet linked; no vtable layout yet | |
415 nonvirtual_vtable_index = -2 // there is no need for vtable dispatch | |
416 // 6330203 Note: Do not use -1, which was overloaded with many meanings. | |
417 }; | |
418 DEBUG_ONLY(bool valid_vtable_index() const { return _vtable_index >= nonvirtual_vtable_index; }) | |
419 int vtable_index() const { assert(valid_vtable_index(), ""); | |
420 return _vtable_index; } | |
421 void set_vtable_index(int index) { _vtable_index = index; } | |
422 | |
423 // interpreter entry | |
424 address interpreter_entry() const { return _i2i_entry; } | |
425 // Only used when first initialize so we can set _i2i_entry and _from_interpreted_entry | |
426 void set_interpreter_entry(address entry) { _i2i_entry = entry; _from_interpreted_entry = entry; } | |
427 int interpreter_kind(void) { | |
428 return constMethod()->interpreter_kind(); | |
429 } | |
430 void set_interpreter_kind(); | |
431 void set_interpreter_kind(int kind) { | |
432 constMethod()->set_interpreter_kind(kind); | |
433 } | |
434 | |
435 // native function (used for native methods only) | |
436 enum { | |
437 native_bind_event_is_interesting = true | |
438 }; | |
439 address native_function() const { return *(native_function_addr()); } | |
440 address critical_native_function(); | |
441 | |
442 // Must specify a real function (not NULL). | |
443 // Use clear_native_function() to unregister. | |
444 void set_native_function(address function, bool post_event_flag); | |
445 bool has_native_function() const; | |
446 void clear_native_function(); | |
447 | |
448 // signature handler (used for native methods only) | |
449 address signature_handler() const { return *(signature_handler_addr()); } | |
450 void set_signature_handler(address handler); | |
451 | |
452 // Interpreter oopmap support | |
453 void mask_for(int bci, InterpreterOopMap* mask); | |
454 | |
455 #ifndef PRODUCT | |
456 // operations on invocation counter | |
457 void print_invocation_count(); | |
458 #endif | |
459 | |
460 // byte codes | |
461 void set_code(address code) { return constMethod()->set_code(code); } | |
462 address code_base() const { return constMethod()->code_base(); } | |
463 bool contains(address bcp) const { return constMethod()->contains(bcp); } | |
464 | |
465 // prints byte codes | |
466 void print_codes() const { print_codes_on(tty); } | |
467 void print_codes_on(outputStream* st) const PRODUCT_RETURN; | |
468 void print_codes_on(int from, int to, outputStream* st) const PRODUCT_RETURN; | |
469 | |
470 // checked exceptions | |
471 int checked_exceptions_length() const | |
472 { return constMethod()->checked_exceptions_length(); } | |
473 CheckedExceptionElement* checked_exceptions_start() const | |
474 { return constMethod()->checked_exceptions_start(); } | |
475 | |
476 // localvariable table | |
477 bool has_localvariable_table() const | |
478 { return constMethod()->has_localvariable_table(); } | |
479 int localvariable_table_length() const | |
480 { return constMethod()->localvariable_table_length(); } | |
481 LocalVariableTableElement* localvariable_table_start() const | |
482 { return constMethod()->localvariable_table_start(); } | |
483 | |
484 bool has_linenumber_table() const | |
485 { return constMethod()->has_linenumber_table(); } | |
486 u_char* compressed_linenumber_table() const | |
487 { return constMethod()->compressed_linenumber_table(); } | |
488 | |
489 // method holder (the Klass* holding this method) | |
490 Klass* method_holder() const { return constants()->pool_holder(); } | |
491 | |
492 void compute_size_of_parameters(Thread *thread); // word size of parameters (receiver if any + arguments) | |
493 Symbol* klass_name() const; // returns the name of the method holder | |
494 BasicType result_type() const; // type of the method result | |
495 int result_type_index() const; // type index of the method result | |
496 bool is_returning_oop() const { BasicType r = result_type(); return (r == T_OBJECT || r == T_ARRAY); } | |
497 bool is_returning_fp() const { BasicType r = result_type(); return (r == T_FLOAT || r == T_DOUBLE); } | |
498 | |
499 // Checked exceptions thrown by this method (resolved to mirrors) | |
500 objArrayHandle resolved_checked_exceptions(TRAPS) { return resolved_checked_exceptions_impl(this, THREAD); } | |
501 | |
502 // Access flags | |
503 bool is_public() const { return access_flags().is_public(); } | |
504 bool is_private() const { return access_flags().is_private(); } | |
505 bool is_protected() const { return access_flags().is_protected(); } | |
506 bool is_package_private() const { return !is_public() && !is_private() && !is_protected(); } | |
507 bool is_static() const { return access_flags().is_static(); } | |
508 bool is_final() const { return access_flags().is_final(); } | |
509 bool is_synchronized() const { return access_flags().is_synchronized();} | |
510 bool is_native() const { return access_flags().is_native(); } | |
511 bool is_abstract() const { return access_flags().is_abstract(); } | |
512 bool is_strict() const { return access_flags().is_strict(); } | |
513 bool is_synthetic() const { return access_flags().is_synthetic(); } | |
514 | |
515 // returns true if contains only return operation | |
516 bool is_empty_method() const; | |
517 | |
518 // returns true if this is a vanilla constructor | |
519 bool is_vanilla_constructor() const; | |
520 | |
521 // checks method and its method holder | |
522 bool is_final_method() const; | |
523 bool is_strict_method() const; | |
524 | |
525 // true if method needs no dynamic dispatch (final and/or no vtable entry) | |
526 bool can_be_statically_bound() const; | |
527 | |
528 // returns true if the method has any backward branches. | |
529 bool has_loops() { | |
530 return access_flags().loops_flag_init() ? access_flags().has_loops() : compute_has_loops_flag(); | |
531 }; | |
532 | |
533 bool compute_has_loops_flag(); | |
534 | |
535 bool has_jsrs() { | |
536 return access_flags().has_jsrs(); | |
537 }; | |
538 void set_has_jsrs() { | |
539 _access_flags.set_has_jsrs(); | |
540 } | |
541 | |
542 // returns true if the method has any monitors. | |
543 bool has_monitors() const { return is_synchronized() || access_flags().has_monitor_bytecodes(); } | |
544 bool has_monitor_bytecodes() const { return access_flags().has_monitor_bytecodes(); } | |
545 | |
546 void set_has_monitor_bytecodes() { _access_flags.set_has_monitor_bytecodes(); } | |
547 | |
548 // monitor matching. This returns a conservative estimate of whether the monitorenter/monitorexit bytecodes | |
549 // propererly nest in the method. It might return false, even though they actually nest properly, since the info. | |
550 // has not been computed yet. | |
551 bool guaranteed_monitor_matching() const { return access_flags().is_monitor_matching(); } | |
552 void set_guaranteed_monitor_matching() { _access_flags.set_monitor_matching(); } | |
553 | |
554 // returns true if the method is an accessor function (setter/getter). | |
555 bool is_accessor() const; | |
556 | |
557 // returns true if the method is an initializer (<init> or <clinit>). | |
558 bool is_initializer() const; | |
559 | |
560 // returns true if the method is static OR if the classfile version < 51 | |
561 bool has_valid_initializer_flags() const; | |
562 | |
563 // returns true if the method name is <clinit> and the method has | |
564 // valid static initializer flags. | |
565 bool is_static_initializer() const; | |
566 | |
567 // compiled code support | |
568 // NOTE: code() is inherently racy as deopt can be clearing code | |
569 // simultaneously. Use with caution. | |
570 bool has_compiled_code() const { return code() != NULL; } | |
571 | |
572 // sizing | |
573 static int header_size() { return sizeof(Method)/HeapWordSize; } | |
574 static int size(bool is_native); | |
575 int size() const { return method_size(); } | |
576 | |
577 // interpreter support | |
578 static ByteSize const_offset() { return byte_offset_of(Method, _constMethod ); } | |
579 static ByteSize access_flags_offset() { return byte_offset_of(Method, _access_flags ); } | |
580 #ifdef CC_INTERP | |
581 static ByteSize result_index_offset() { return byte_offset_of(Method, _result_index ); } | |
582 #endif /* CC_INTERP */ | |
583 static ByteSize size_of_locals_offset() { return byte_offset_of(Method, _max_locals ); } | |
584 static ByteSize size_of_parameters_offset() { return byte_offset_of(Method, _size_of_parameters); } | |
585 static ByteSize from_compiled_offset() { return byte_offset_of(Method, _from_compiled_entry); } | |
586 static ByteSize code_offset() { return byte_offset_of(Method, _code); } | |
587 static ByteSize invocation_counter_offset() { return byte_offset_of(Method, _invocation_counter); } | |
588 static ByteSize backedge_counter_offset() { return byte_offset_of(Method, _backedge_counter); } | |
589 static ByteSize method_data_offset() { | |
590 return byte_offset_of(Method, _method_data); | |
591 } | |
592 static ByteSize interpreter_invocation_counter_offset() { return byte_offset_of(Method, _interpreter_invocation_count); } | |
593 #ifndef PRODUCT | |
594 static ByteSize compiled_invocation_counter_offset() { return byte_offset_of(Method, _compiled_invocation_count); } | |
595 #endif // not PRODUCT | |
596 static ByteSize native_function_offset() { return in_ByteSize(sizeof(Method)); } | |
597 static ByteSize from_interpreted_offset() { return byte_offset_of(Method, _from_interpreted_entry ); } | |
598 static ByteSize interpreter_entry_offset() { return byte_offset_of(Method, _i2i_entry ); } | |
599 static ByteSize signature_handler_offset() { return in_ByteSize(sizeof(Method) + wordSize); } | |
600 static ByteSize max_stack_offset() { return byte_offset_of(Method, _max_stack ); } | |
601 | |
602 // for code generation | |
603 static int method_data_offset_in_bytes() { return offset_of(Method, _method_data); } | |
604 static int interpreter_invocation_counter_offset_in_bytes() | |
605 { return offset_of(Method, _interpreter_invocation_count); } | |
606 static int intrinsic_id_offset_in_bytes() { return offset_of(Method, _intrinsic_id); } | |
607 static int intrinsic_id_size_in_bytes() { return sizeof(u1); } | |
608 | |
609 // Static methods that are used to implement member methods where an exposed this pointer | |
610 // is needed due to possible GCs | |
611 static objArrayHandle resolved_checked_exceptions_impl(Method* this_oop, TRAPS); | |
612 | |
613 // Returns the byte code index from the byte code pointer | |
614 int bci_from(address bcp) const; | |
615 address bcp_from(int bci) const; | |
616 int validate_bci_from_bcx(intptr_t bcx) const; | |
617 | |
618 // Returns the line number for a bci if debugging information for the method is prowided, | |
619 // -1 is returned otherwise. | |
620 int line_number_from_bci(int bci) const; | |
621 | |
622 // Reflection support | |
623 bool is_overridden_in(Klass* k) const; | |
624 | |
625 // JSR 292 support | |
626 bool is_method_handle_intrinsic() const; // MethodHandles::is_signature_polymorphic_intrinsic(intrinsic_id) | |
627 bool is_compiled_lambda_form() const; // intrinsic_id() == vmIntrinsics::_compiledLambdaForm | |
628 bool has_member_arg() const; // intrinsic_id() == vmIntrinsics::_linkToSpecial, etc. | |
629 static methodHandle make_method_handle_intrinsic(vmIntrinsics::ID iid, // _invokeBasic, _linkToVirtual | |
630 Symbol* signature, //anything at all | |
631 TRAPS); | |
632 static Klass* check_non_bcp_klass(Klass* klass); | |
633 // these operate only on invoke methods: | |
634 // presize interpreter frames for extra interpreter stack entries, if needed | |
635 // method handles want to be able to push a few extra values (e.g., a bound receiver), and | |
636 // invokedynamic sometimes needs to push a bootstrap method, call site, and arglist, | |
637 // all without checking for a stack overflow | |
638 static int extra_stack_entries() { return EnableInvokeDynamic ? 2 : 0; } | |
639 static int extra_stack_words(); // = extra_stack_entries() * Interpreter::stackElementSize() | |
640 | |
641 // RedefineClasses() support: | |
642 bool is_old() const { return access_flags().is_old(); } | |
643 void set_is_old() { _access_flags.set_is_old(); } | |
644 bool is_obsolete() const { return access_flags().is_obsolete(); } | |
645 void set_is_obsolete() { _access_flags.set_is_obsolete(); } | |
646 bool on_stack() const { return access_flags().on_stack(); } | |
647 void set_on_stack(const bool value); | |
648 | |
649 // see the definition in Method*.cpp for the gory details | |
650 bool should_not_be_cached() const; | |
651 | |
652 // JVMTI Native method prefixing support: | |
653 bool is_prefixed_native() const { return access_flags().is_prefixed_native(); } | |
654 void set_is_prefixed_native() { _access_flags.set_is_prefixed_native(); } | |
655 | |
656 // Rewriting support | |
657 static methodHandle clone_with_new_data(methodHandle m, u_char* new_code, int new_code_length, | |
658 u_char* new_compressed_linenumber_table, int new_compressed_linenumber_size, TRAPS); | |
659 | |
660 // jmethodID handling | |
661 // Because the useful life-span of a jmethodID cannot be determined, | |
662 // once created they are never reclaimed. The methods to which they refer, | |
663 // however, can be GC'ed away if the class is unloaded or if the method is | |
664 // made obsolete or deleted -- in these cases, the jmethodID | |
665 // refers to NULL (as is the case for any weak reference). | |
666 static jmethodID make_jmethod_id(ClassLoaderData* loader_data, Method* mh); | |
667 static void destroy_jmethod_id(ClassLoaderData* loader_data, jmethodID mid); | |
668 | |
669 // Use resolve_jmethod_id() in situations where the caller is expected | |
670 // to provide a valid jmethodID; the only sanity checks are in asserts; | |
671 // result guaranteed not to be NULL. | |
672 inline static Method* resolve_jmethod_id(jmethodID mid) { | |
673 assert(mid != NULL, "JNI method id should not be null"); | |
674 return *((Method**)mid); | |
675 } | |
676 | |
677 // Use checked_resolve_jmethod_id() in situations where the caller | |
678 // should provide a valid jmethodID, but might not. NULL is returned | |
679 // when the jmethodID does not refer to a valid method. | |
680 static Method* checked_resolve_jmethod_id(jmethodID mid); | |
681 | |
682 static void change_method_associated_with_jmethod_id(jmethodID old_jmid_ptr, Method* new_method); | |
683 static bool is_method_id(jmethodID mid); | |
684 | |
685 // Clear methods | |
686 static void clear_jmethod_ids(ClassLoaderData* loader_data); | |
687 static void print_jmethod_ids(ClassLoaderData* loader_data, outputStream* out) PRODUCT_RETURN; | |
688 | |
689 // Get this method's jmethodID -- allocate if it doesn't exist | |
690 jmethodID jmethod_id() { methodHandle this_h(this); | |
691 return InstanceKlass::get_jmethod_id(InstanceKlass::cast(method_holder()), this_h); } | |
692 | |
693 // Lookup the jmethodID for this method. Return NULL if not found. | |
694 // NOTE that this function can be called from a signal handler | |
695 // (see AsyncGetCallTrace support for Forte Analyzer) and this | |
696 // needs to be async-safe. No allocation should be done and | |
697 // so handles are not used to avoid deadlock. | |
698 jmethodID find_jmethod_id_or_null() { return InstanceKlass::cast(method_holder())->jmethod_id_or_null(this); } | |
699 | |
700 // JNI static invoke cached itable index accessors | |
701 int cached_itable_index() { return InstanceKlass::cast(method_holder())->cached_itable_index(method_idnum()); } | |
702 void set_cached_itable_index(int index) { InstanceKlass::cast(method_holder())->set_cached_itable_index(method_idnum(), index); } | |
703 | |
704 // Support for inlining of intrinsic methods | |
705 vmIntrinsics::ID intrinsic_id() const { return (vmIntrinsics::ID) _intrinsic_id; } | |
706 void set_intrinsic_id(vmIntrinsics::ID id) { _intrinsic_id = (u1) id; } | |
707 | |
708 // Helper routines for intrinsic_id() and vmIntrinsics::method(). | |
709 void init_intrinsic_id(); // updates from _none if a match | |
710 static vmSymbols::SID klass_id_for_intrinsics(Klass* holder); | |
711 | |
712 bool jfr_towrite() { return _jfr_towrite; } | |
713 void set_jfr_towrite(bool towrite) { _jfr_towrite = towrite; } | |
714 | |
715 bool force_inline() { return _force_inline; } | |
716 void set_force_inline(bool x) { _force_inline = x; } | |
717 bool dont_inline() { return _dont_inline; } | |
718 void set_dont_inline(bool x) { _dont_inline = x; } | |
719 bool is_hidden() { return _hidden; } | |
720 void set_hidden(bool x) { _hidden = x; } | |
721 | |
722 // On-stack replacement support | |
723 bool has_osr_nmethod(int level, bool match_level) { | |
724 return InstanceKlass::cast(method_holder())->lookup_osr_nmethod(this, InvocationEntryBci, level, match_level) != NULL; | |
725 } | |
726 | |
727 nmethod* lookup_osr_nmethod_for(int bci, int level, bool match_level) { | |
728 return InstanceKlass::cast(method_holder())->lookup_osr_nmethod(this, bci, level, match_level); | |
729 } | |
730 | |
731 // Inline cache support | |
732 void cleanup_inline_caches(); | |
733 | |
734 // Find if klass for method is loaded | |
735 bool is_klass_loaded_by_klass_index(int klass_index) const; | |
736 bool is_klass_loaded(int refinfo_index, bool must_be_resolved = false) const; | |
737 | |
738 // Indicates whether compilation failed earlier for this method, or | |
739 // whether it is not compilable for another reason like having a | |
740 // breakpoint set in it. | |
741 bool is_not_compilable(int comp_level = CompLevel_any) const; | |
742 void set_not_compilable(int comp_level = CompLevel_all, bool report = true); | |
743 void set_not_compilable_quietly(int comp_level = CompLevel_all) { | |
744 set_not_compilable(comp_level, false); | |
745 } | |
746 bool is_not_osr_compilable(int comp_level = CompLevel_any) const { | |
747 return is_not_compilable(comp_level) || access_flags().is_not_osr_compilable(); | |
748 } | |
749 void set_not_osr_compilable() { _access_flags.set_not_osr_compilable(); } | |
750 bool is_not_c1_compilable() const { return access_flags().is_not_c1_compilable(); } | |
751 void set_not_c1_compilable() { _access_flags.set_not_c1_compilable(); } | |
752 bool is_not_c2_compilable() const { return access_flags().is_not_c2_compilable(); } | |
753 void set_not_c2_compilable() { _access_flags.set_not_c2_compilable(); } | |
754 | |
755 // Background compilation support | |
756 bool queued_for_compilation() const { return access_flags().queued_for_compilation(); } | |
757 void set_queued_for_compilation() { _access_flags.set_queued_for_compilation(); } | |
758 void clear_queued_for_compilation() { _access_flags.clear_queued_for_compilation(); } | |
759 | |
760 // Resolve all classes in signature, return 'true' if successful | |
761 static bool load_signature_classes(methodHandle m, TRAPS); | |
762 | |
763 // Return if true if not all classes references in signature, including return type, has been loaded | |
764 static bool has_unloaded_classes_in_signature(methodHandle m, TRAPS); | |
765 | |
766 // Printing | |
767 void print_short_name(outputStream* st = tty) /*PRODUCT_RETURN*/; // prints as klassname::methodname; Exposed so field engineers can debug VM | |
768 void print_name(outputStream* st = tty) PRODUCT_RETURN; // prints as "virtual void foo(int)" | |
769 | |
770 // Helper routine used for method sorting | |
771 static void sort_methods(Array<Method*>* methods, | |
772 Array<AnnotationArray*>* methods_annotations, | |
773 Array<AnnotationArray*>* methods_parameter_annotations, | |
774 Array<AnnotationArray*>* methods_default_annotations, | |
775 bool idempotent = false); | |
776 | |
777 // size of parameters | |
778 void set_size_of_parameters(int size) { _size_of_parameters = size; } | |
779 | |
780 // Deallocation function for redefine classes or if an error occurs | |
781 void deallocate_contents(ClassLoaderData* loader_data); | |
782 | |
783 // Printing | |
784 #ifndef PRODUCT | |
785 void print_on(outputStream* st) const; | |
786 #endif | |
787 void print_value_on(outputStream* st) const; | |
788 | |
789 const char* internal_name() const { return "{method}"; } | |
790 | |
791 // Verify | |
792 void verify() { verify_on(tty); } | |
793 void verify_on(outputStream* st); | |
794 | |
795 private: | |
796 | |
797 // Inlined elements | |
798 address* native_function_addr() const { assert(is_native(), "must be native"); return (address*) (this+1); } | |
799 address* signature_handler_addr() const { return native_function_addr() + 1; } | |
800 }; | |
801 | |
802 | |
803 // Utility class for compressing line number tables | |
804 | |
805 class CompressedLineNumberWriteStream: public CompressedWriteStream { | |
806 private: | |
807 int _bci; | |
808 int _line; | |
809 public: | |
810 // Constructor | |
811 CompressedLineNumberWriteStream(int initial_size) : CompressedWriteStream(initial_size), _bci(0), _line(0) {} | |
812 CompressedLineNumberWriteStream(u_char* buffer, int initial_size) : CompressedWriteStream(buffer, initial_size), _bci(0), _line(0) {} | |
813 | |
814 // Write (bci, line number) pair to stream | |
815 void write_pair_regular(int bci_delta, int line_delta); | |
816 | |
817 inline void write_pair_inline(int bci, int line) { | |
818 int bci_delta = bci - _bci; | |
819 int line_delta = line - _line; | |
820 _bci = bci; | |
821 _line = line; | |
822 // Skip (0,0) deltas - they do not add information and conflict with terminator. | |
823 if (bci_delta == 0 && line_delta == 0) return; | |
824 // Check if bci is 5-bit and line number 3-bit unsigned. | |
825 if (((bci_delta & ~0x1F) == 0) && ((line_delta & ~0x7) == 0)) { | |
826 // Compress into single byte. | |
827 jubyte value = ((jubyte) bci_delta << 3) | (jubyte) line_delta; | |
828 // Check that value doesn't match escape character. | |
829 if (value != 0xFF) { | |
830 write_byte(value); | |
831 return; | |
832 } | |
833 } | |
834 write_pair_regular(bci_delta, line_delta); | |
835 } | |
836 | |
837 // Windows AMD64 + Apr 2005 PSDK with /O2 generates bad code for write_pair. | |
838 // Disabling optimization doesn't work for methods in header files | |
839 // so we force it to call through the non-optimized version in the .cpp. | |
840 // It's gross, but it's the only way we can ensure that all callers are | |
841 // fixed. _MSC_VER is defined by the windows compiler | |
842 #if defined(_M_AMD64) && _MSC_VER >= 1400 | |
843 void write_pair(int bci, int line); | |
844 #else | |
845 void write_pair(int bci, int line) { write_pair_inline(bci, line); } | |
846 #endif | |
847 | |
848 // Write end-of-stream marker | |
849 void write_terminator() { write_byte(0); } | |
850 }; | |
851 | |
852 | |
853 // Utility class for decompressing line number tables | |
854 | |
855 class CompressedLineNumberReadStream: public CompressedReadStream { | |
856 private: | |
857 int _bci; | |
858 int _line; | |
859 public: | |
860 // Constructor | |
861 CompressedLineNumberReadStream(u_char* buffer); | |
862 // Read (bci, line number) pair from stream. Returns false at end-of-stream. | |
863 bool read_pair(); | |
864 // Accessing bci and line number (after calling read_pair) | |
865 int bci() const { return _bci; } | |
866 int line() const { return _line; } | |
867 }; | |
868 | |
869 | |
870 /// Fast Breakpoints. | |
871 | |
872 // If this structure gets more complicated (because bpts get numerous), | |
873 // move it into its own header. | |
874 | |
875 // There is presently no provision for concurrent access | |
876 // to breakpoint lists, which is only OK for JVMTI because | |
877 // breakpoints are written only at safepoints, and are read | |
878 // concurrently only outside of safepoints. | |
879 | |
880 class BreakpointInfo : public CHeapObj<mtClass> { | |
881 friend class VMStructs; | |
882 private: | |
883 Bytecodes::Code _orig_bytecode; | |
884 int _bci; | |
885 u2 _name_index; // of method | |
886 u2 _signature_index; // of method | |
887 BreakpointInfo* _next; // simple storage allocation | |
888 | |
889 public: | |
890 BreakpointInfo(Method* m, int bci); | |
891 | |
892 // accessors | |
893 Bytecodes::Code orig_bytecode() { return _orig_bytecode; } | |
894 void set_orig_bytecode(Bytecodes::Code code) { _orig_bytecode = code; } | |
895 int bci() { return _bci; } | |
896 | |
897 BreakpointInfo* next() const { return _next; } | |
898 void set_next(BreakpointInfo* n) { _next = n; } | |
899 | |
900 // helps for searchers | |
901 bool match(const Method* m, int bci) { | |
902 return bci == _bci && match(m); | |
903 } | |
904 | |
905 bool match(const Method* m) { | |
906 return _name_index == m->name_index() && | |
907 _signature_index == m->signature_index(); | |
908 } | |
909 | |
910 void set(Method* method); | |
911 void clear(Method* method); | |
912 }; | |
913 | |
914 // Utility class for access exception handlers | |
915 class ExceptionTable : public StackObj { | |
916 private: | |
917 ExceptionTableElement* _table; | |
918 u2 _length; | |
919 | |
920 public: | |
921 ExceptionTable(Method* m) { | |
922 if (m->has_exception_handler()) { | |
923 _table = m->exception_table_start(); | |
924 _length = m->exception_table_length(); | |
925 } else { | |
926 _table = NULL; | |
927 _length = 0; | |
928 } | |
929 } | |
930 | |
931 int length() const { | |
932 return _length; | |
933 } | |
934 | |
935 u2 start_pc(int idx) const { | |
936 assert(idx < _length, "out of bounds"); | |
937 return _table[idx].start_pc; | |
938 } | |
939 | |
940 void set_start_pc(int idx, u2 value) { | |
941 assert(idx < _length, "out of bounds"); | |
942 _table[idx].start_pc = value; | |
943 } | |
944 | |
945 u2 end_pc(int idx) const { | |
946 assert(idx < _length, "out of bounds"); | |
947 return _table[idx].end_pc; | |
948 } | |
949 | |
950 void set_end_pc(int idx, u2 value) { | |
951 assert(idx < _length, "out of bounds"); | |
952 _table[idx].end_pc = value; | |
953 } | |
954 | |
955 u2 handler_pc(int idx) const { | |
956 assert(idx < _length, "out of bounds"); | |
957 return _table[idx].handler_pc; | |
958 } | |
959 | |
960 void set_handler_pc(int idx, u2 value) { | |
961 assert(idx < _length, "out of bounds"); | |
962 _table[idx].handler_pc = value; | |
963 } | |
964 | |
965 u2 catch_type_index(int idx) const { | |
966 assert(idx < _length, "out of bounds"); | |
967 return _table[idx].catch_type_index; | |
968 } | |
969 | |
970 void set_catch_type_index(int idx, u2 value) { | |
971 assert(idx < _length, "out of bounds"); | |
972 _table[idx].catch_type_index = value; | |
973 } | |
974 }; | |
975 | |
976 #endif // SHARE_VM_OOPS_METHODOOP_HPP |