Mercurial > hg > graal-jvmci-8
annotate src/share/vm/runtime/fprofiler.hpp @ 615:c6c601a0f2d6
6797870: Add -XX:+{HeapDump,PrintClassHistogram}{Before,After}FullGC
Summary: Call newly created CollectedHeap::dump_{pre,post}_full_gc before and after every stop-world full collection cycle on GenCollectedHeap and ParallelScavengeHeap. (Support for G1CollectedHeap forthcoming under CR 6810861.) Small modifications to existing heap dumping and class histogram implementation, especially to allow multiple on-the-fly histos/dumps by the VM thread during a single safepoint.
Reviewed-by: jmasa, alanb, mchung
author | ysr |
---|---|
date | Mon, 02 Mar 2009 16:37:04 -0800 |
parents | d1605aabd0a1 |
children | c18cbe5936b8 |
rev | line source |
---|---|
0 | 1 /* |
196 | 2 * Copyright 1997-2008 Sun Microsystems, Inc. All Rights Reserved. |
0 | 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 * | |
5 * This code is free software; you can redistribute it and/or modify it | |
6 * under the terms of the GNU General Public License version 2 only, as | |
7 * published by the Free Software Foundation. | |
8 * | |
9 * This code is distributed in the hope that it will be useful, but WITHOUT | |
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License | |
12 * version 2 for more details (a copy is included in the LICENSE file that | |
13 * accompanied this code). | |
14 * | |
15 * You should have received a copy of the GNU General Public License version | |
16 * 2 along with this work; if not, write to the Free Software Foundation, | |
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | |
18 * | |
19 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, | |
20 * CA 95054 USA or visit www.sun.com if you need additional information or | |
21 * have any questions. | |
22 * | |
23 */ | |
24 | |
25 // a simple flat profiler for Java | |
26 | |
27 | |
28 // Forward declaration of classes defined in this header file | |
29 class ThreadProfiler; | |
30 class ThreadProfilerMark; | |
31 class FlatProfiler; | |
32 class IntervalData; | |
33 | |
34 // Declarations of classes defined only in the implementation. | |
35 class ProfilerNode; | |
36 class FlatProfilerTask; | |
37 | |
38 enum TickPosition { | |
39 tp_code, | |
40 tp_native | |
41 }; | |
42 | |
43 // One of these guys is constructed as we enter interesting regions | |
44 // and destructed as we exit the region. While we are in the region | |
45 // ticks are allotted to the region. | |
46 class ThreadProfilerMark: public StackObj { | |
47 public: | |
48 // For now, the only thread-specific region is the class loader. | |
49 enum Region { noRegion, classLoaderRegion, extraRegion, maxRegion }; | |
50 | |
51 ThreadProfilerMark(Region) KERNEL_RETURN; | |
52 ~ThreadProfilerMark() KERNEL_RETURN; | |
53 | |
54 private: | |
55 ThreadProfiler* _pp; | |
56 Region _r; | |
57 }; | |
58 | |
59 #ifndef FPROF_KERNEL | |
60 | |
61 class IntervalData VALUE_OBJ_CLASS_SPEC { | |
62 // Just to keep these things all together | |
63 private: | |
64 int _interpreted; | |
65 int _compiled; | |
66 int _native; | |
67 int _compiling; | |
68 public: | |
69 int interpreted() { | |
70 return _interpreted; | |
71 } | |
72 int compiled() { | |
73 return _compiled; | |
74 } | |
75 int native() { | |
76 return _native; | |
77 } | |
78 int compiling() { | |
79 return _compiling; | |
80 } | |
81 int total() { | |
82 return (interpreted() + compiled() + native() + compiling()); | |
83 } | |
84 void inc_interpreted() { | |
85 _interpreted += 1; | |
86 } | |
87 void inc_compiled() { | |
88 _compiled += 1; | |
89 } | |
90 void inc_native() { | |
91 _native += 1; | |
92 } | |
93 void inc_compiling() { | |
94 _compiling += 1; | |
95 } | |
96 void reset() { | |
97 _interpreted = 0; | |
98 _compiled = 0; | |
99 _native = 0; | |
100 _compiling = 0; | |
101 } | |
102 static void print_header(outputStream* st); | |
103 void print_data(outputStream* st); | |
104 }; | |
105 #endif // FPROF_KERNEL | |
106 | |
107 class ThreadProfiler: public CHeapObj { | |
108 public: | |
109 ThreadProfiler() KERNEL_RETURN; | |
110 ~ThreadProfiler() KERNEL_RETURN; | |
111 | |
112 // Resets the profiler | |
113 void reset() KERNEL_RETURN; | |
114 | |
115 // Activates the profiler for a certain thread | |
116 void engage() KERNEL_RETURN; | |
117 | |
118 // Deactivates the profiler | |
119 void disengage() KERNEL_RETURN; | |
120 | |
121 // Prints the collected profiling information | |
122 void print(const char* thread_name) KERNEL_RETURN; | |
123 | |
124 // Garbage Collection Support | |
125 void oops_do(OopClosure* f) KERNEL_RETURN; | |
126 | |
127 #ifndef FPROF_KERNEL | |
128 private: | |
129 // for recording ticks. | |
130 friend class ProfilerNode; | |
131 char* area_bottom; // preallocated area for pnodes | |
132 char* area_top; | |
133 char* area_limit; | |
134 static int table_size; | |
135 ProfilerNode** table; | |
136 | |
137 private: | |
107
93b6525e3b82
6603919: Stackwalking crash on x86 -server with Sun Studio's collect -j on
sgoldman
parents:
0
diff
changeset
|
138 void record_interpreted_tick(JavaThread* thread, frame fr, TickPosition where, int* ticks); |
0 | 139 void record_compiled_tick (JavaThread* thread, frame fr, TickPosition where); |
140 void interpreted_update(methodOop method, TickPosition where); | |
141 void compiled_update (methodOop method, TickPosition where); | |
142 void stub_update (methodOop method, const char* name, TickPosition where); | |
143 void adapter_update (TickPosition where); | |
144 | |
145 void runtime_stub_update(const CodeBlob* stub, const char* name, TickPosition where); | |
146 void unknown_compiled_update (const CodeBlob* cb, TickPosition where); | |
147 | |
148 void vm_update (TickPosition where); | |
149 void vm_update (const char* name, TickPosition where); | |
150 | |
151 void record_tick_for_running_frame(JavaThread* thread, frame fr); | |
152 void record_tick_for_calling_frame(JavaThread* thread, frame fr); | |
153 | |
154 void initialize(); | |
155 | |
156 static int entry(int value); | |
157 | |
158 | |
159 private: | |
160 friend class FlatProfiler; | |
161 void record_tick(JavaThread* thread); | |
162 bool engaged; | |
163 // so we can do percentages for this thread, and quick checks for activity | |
164 int thread_ticks; | |
165 int compiler_ticks; | |
166 int interpreter_ticks; | |
167 | |
168 public: | |
169 void inc_thread_ticks() { thread_ticks += 1; } | |
170 | |
171 private: | |
172 friend class ThreadProfilerMark; | |
173 // counters for thread-specific regions | |
174 bool region_flag[ThreadProfilerMark::maxRegion]; | |
175 int class_loader_ticks; | |
176 int extra_ticks; | |
177 | |
178 private: | |
179 // other thread-specific regions | |
180 int blocked_ticks; | |
181 enum UnknownTickSites { | |
182 ut_null_method, | |
183 ut_vtable_stubs, | |
184 ut_running_frame, | |
185 ut_calling_frame, | |
186 ut_no_pc, | |
187 ut_no_last_Java_frame, | |
188 ut_unknown_thread_state, | |
189 ut_end | |
190 }; | |
191 int unknown_ticks_array[ut_end]; | |
192 int unknown_ticks() { | |
193 int result = 0; | |
194 for (int ut = 0; ut < ut_end; ut += 1) { | |
195 result += unknown_ticks_array[ut]; | |
196 } | |
197 return result; | |
198 } | |
199 | |
200 elapsedTimer timer; | |
201 | |
202 // For interval timing | |
203 private: | |
204 IntervalData _interval_data; | |
205 IntervalData interval_data() { | |
206 return _interval_data; | |
207 } | |
208 IntervalData* interval_data_ref() { | |
209 return &_interval_data; | |
210 } | |
211 #endif // FPROF_KERNEL | |
212 }; | |
213 | |
214 class FlatProfiler: AllStatic { | |
215 public: | |
216 static void reset() KERNEL_RETURN ; | |
217 static void engage(JavaThread* mainThread, bool fullProfile) KERNEL_RETURN ; | |
218 static void disengage() KERNEL_RETURN ; | |
219 static void print(int unused) KERNEL_RETURN ; | |
220 static bool is_active() KERNEL_RETURN_(return false;) ; | |
221 | |
222 // This is NULL if each thread has its own thread profiler, | |
223 // else this is the single thread profiler used by all threads. | |
224 // In particular it makes a difference during garbage collection, | |
225 // where you only want to traverse each thread profiler once. | |
226 static ThreadProfiler* get_thread_profiler() KERNEL_RETURN_(return NULL;); | |
227 | |
228 // Garbage Collection Support | |
229 static void oops_do(OopClosure* f) KERNEL_RETURN ; | |
230 | |
231 // Support for disassembler to inspect the PCRecorder | |
232 | |
233 // Returns the start address for a given pc | |
234 // NULL is returned if the PCRecorder is inactive | |
235 static address bucket_start_for(address pc) KERNEL_RETURN_(return NULL;); | |
236 | |
237 enum { MillisecsPerTick = 10 }; // ms per profiling ticks | |
238 | |
239 // Returns the number of ticks recorded for the bucket | |
240 // pc belongs to. | |
241 static int bucket_count_for(address pc) KERNEL_RETURN_(return 0;); | |
242 | |
243 #ifndef FPROF_KERNEL | |
244 | |
245 private: | |
246 static bool full_profile() { | |
247 return full_profile_flag; | |
248 } | |
249 | |
250 friend class ThreadProfiler; | |
251 // the following group of ticks cover everything that's not attributed to individual Java methods | |
252 static int received_gc_ticks; // ticks during which gc was active | |
253 static int vm_operation_ticks; // total ticks in vm_operations other than GC | |
254 static int threads_lock_ticks; // the number of times we couldn't get the Threads_lock without blocking | |
255 static int blocked_ticks; // ticks when the thread was blocked. | |
256 static int class_loader_ticks; // total ticks in class loader | |
257 static int extra_ticks; // total ticks an extra temporary measuring | |
258 static int compiler_ticks; // total ticks in compilation | |
259 static int interpreter_ticks; // ticks in unknown interpreted method | |
260 static int deopt_ticks; // ticks in deoptimization | |
261 static int unknown_ticks; // ticks that cannot be categorized | |
262 static int received_ticks; // ticks that were received by task | |
263 static int delivered_ticks; // ticks that were delivered by task | |
264 static int non_method_ticks() { | |
265 return | |
266 ( received_gc_ticks | |
267 + vm_operation_ticks | |
268 + deopt_ticks | |
269 + threads_lock_ticks | |
270 + blocked_ticks | |
271 + compiler_ticks | |
272 + interpreter_ticks | |
273 + unknown_ticks ); | |
274 } | |
275 static elapsedTimer timer; | |
276 | |
277 // Counts of each of the byte codes | |
278 static int* bytecode_ticks; | |
279 static int* bytecode_ticks_stub; | |
280 static void print_byte_code_statistics(); | |
281 | |
282 // the ticks below are for continuous profiling (to adjust recompilation, etc.) | |
283 static int all_ticks; // total count of ticks received so far | |
284 static int all_int_ticks; // ticks in interpreter | |
285 static int all_comp_ticks; // ticks in compiled code (+ native) | |
286 static bool full_profile_flag; // collecting full profile? | |
287 | |
288 // to accumulate thread-specific data | |
289 // if we aren't profiling individual threads. | |
290 static ThreadProfiler* thread_profiler; | |
291 static ThreadProfiler* vm_thread_profiler; | |
292 | |
293 static void allocate_table(); | |
294 | |
295 // The task that periodically interrupts things. | |
296 friend class FlatProfilerTask; | |
297 static FlatProfilerTask* task; | |
298 static void record_vm_operation(); | |
299 static void record_vm_tick(); | |
300 static void record_thread_ticks(); | |
301 | |
302 // For interval analysis | |
303 private: | |
304 static int interval_ticks_previous; // delivered_ticks from the last interval | |
305 static void interval_record_thread(ThreadProfiler* tp); // extract ticks from ThreadProfiler. | |
306 static void interval_print(); // print interval data. | |
307 static void interval_reset(); // reset interval data. | |
308 enum {interval_print_size = 10}; | |
309 static IntervalData* interval_data; | |
310 #endif // FPROF_KERNEL | |
311 }; |