comparison src/share/vm/runtime/fprofiler.hpp @ 0:a61af66fc99e jdk7-b24

Initial load
author duke
date Sat, 01 Dec 2007 00:00:00 +0000
parents
children 93b6525e3b82
comparison
equal deleted inserted replaced
-1:000000000000 0:a61af66fc99e
1 /*
2 * Copyright 1997-2007 Sun Microsystems, Inc. All Rights Reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
20 * CA 95054 USA or visit www.sun.com if you need additional information or
21 * have any questions.
22 *
23 */
24
25 // a simple flat profiler for Java
26
27
28 // Forward declaration of classes defined in this header file
29 class ThreadProfiler;
30 class ThreadProfilerMark;
31 class FlatProfiler;
32 class IntervalData;
33
34 // Declarations of classes defined only in the implementation.
35 class ProfilerNode;
36 class FlatProfilerTask;
37
38 enum TickPosition {
39 tp_code,
40 tp_native
41 };
42
43 // One of these guys is constructed as we enter interesting regions
44 // and destructed as we exit the region. While we are in the region
45 // ticks are allotted to the region.
46 class ThreadProfilerMark: public StackObj {
47 public:
48 // For now, the only thread-specific region is the class loader.
49 enum Region { noRegion, classLoaderRegion, extraRegion, maxRegion };
50
51 ThreadProfilerMark(Region) KERNEL_RETURN;
52 ~ThreadProfilerMark() KERNEL_RETURN;
53
54 private:
55 ThreadProfiler* _pp;
56 Region _r;
57 };
58
59 #ifndef FPROF_KERNEL
60
61 class IntervalData VALUE_OBJ_CLASS_SPEC {
62 // Just to keep these things all together
63 private:
64 int _interpreted;
65 int _compiled;
66 int _native;
67 int _compiling;
68 public:
69 int interpreted() {
70 return _interpreted;
71 }
72 int compiled() {
73 return _compiled;
74 }
75 int native() {
76 return _native;
77 }
78 int compiling() {
79 return _compiling;
80 }
81 int total() {
82 return (interpreted() + compiled() + native() + compiling());
83 }
84 void inc_interpreted() {
85 _interpreted += 1;
86 }
87 void inc_compiled() {
88 _compiled += 1;
89 }
90 void inc_native() {
91 _native += 1;
92 }
93 void inc_compiling() {
94 _compiling += 1;
95 }
96 void reset() {
97 _interpreted = 0;
98 _compiled = 0;
99 _native = 0;
100 _compiling = 0;
101 }
102 static void print_header(outputStream* st);
103 void print_data(outputStream* st);
104 };
105 #endif // FPROF_KERNEL
106
107 class ThreadProfiler: public CHeapObj {
108 public:
109 ThreadProfiler() KERNEL_RETURN;
110 ~ThreadProfiler() KERNEL_RETURN;
111
112 // Resets the profiler
113 void reset() KERNEL_RETURN;
114
115 // Activates the profiler for a certain thread
116 void engage() KERNEL_RETURN;
117
118 // Deactivates the profiler
119 void disengage() KERNEL_RETURN;
120
121 // Prints the collected profiling information
122 void print(const char* thread_name) KERNEL_RETURN;
123
124 // Garbage Collection Support
125 void oops_do(OopClosure* f) KERNEL_RETURN;
126
127 #ifndef FPROF_KERNEL
128 private:
129 // for recording ticks.
130 friend class ProfilerNode;
131 char* area_bottom; // preallocated area for pnodes
132 char* area_top;
133 char* area_limit;
134 static int table_size;
135 ProfilerNode** table;
136
137 private:
138 void record_interpreted_tick(frame fr, TickPosition where, int* ticks);
139 void record_compiled_tick (JavaThread* thread, frame fr, TickPosition where);
140 void interpreted_update(methodOop method, TickPosition where);
141 void compiled_update (methodOop method, TickPosition where);
142 void stub_update (methodOop method, const char* name, TickPosition where);
143 void adapter_update (TickPosition where);
144
145 void runtime_stub_update(const CodeBlob* stub, const char* name, TickPosition where);
146 void unknown_compiled_update (const CodeBlob* cb, TickPosition where);
147
148 void vm_update (TickPosition where);
149 void vm_update (const char* name, TickPosition where);
150
151 void record_tick_for_running_frame(JavaThread* thread, frame fr);
152 void record_tick_for_calling_frame(JavaThread* thread, frame fr);
153
154 void initialize();
155
156 static int entry(int value);
157
158
159 private:
160 friend class FlatProfiler;
161 void record_tick(JavaThread* thread);
162 bool engaged;
163 // so we can do percentages for this thread, and quick checks for activity
164 int thread_ticks;
165 int compiler_ticks;
166 int interpreter_ticks;
167
168 public:
169 void inc_thread_ticks() { thread_ticks += 1; }
170
171 private:
172 friend class ThreadProfilerMark;
173 // counters for thread-specific regions
174 bool region_flag[ThreadProfilerMark::maxRegion];
175 int class_loader_ticks;
176 int extra_ticks;
177
178 private:
179 // other thread-specific regions
180 int blocked_ticks;
181 enum UnknownTickSites {
182 ut_null_method,
183 ut_vtable_stubs,
184 ut_running_frame,
185 ut_calling_frame,
186 ut_no_pc,
187 ut_no_last_Java_frame,
188 ut_unknown_thread_state,
189 ut_end
190 };
191 int unknown_ticks_array[ut_end];
192 int unknown_ticks() {
193 int result = 0;
194 for (int ut = 0; ut < ut_end; ut += 1) {
195 result += unknown_ticks_array[ut];
196 }
197 return result;
198 }
199
200 elapsedTimer timer;
201
202 // For interval timing
203 private:
204 IntervalData _interval_data;
205 IntervalData interval_data() {
206 return _interval_data;
207 }
208 IntervalData* interval_data_ref() {
209 return &_interval_data;
210 }
211 #endif // FPROF_KERNEL
212 };
213
214 class FlatProfiler: AllStatic {
215 public:
216 static void reset() KERNEL_RETURN ;
217 static void engage(JavaThread* mainThread, bool fullProfile) KERNEL_RETURN ;
218 static void disengage() KERNEL_RETURN ;
219 static void print(int unused) KERNEL_RETURN ;
220 static bool is_active() KERNEL_RETURN_(return false;) ;
221
222 // This is NULL if each thread has its own thread profiler,
223 // else this is the single thread profiler used by all threads.
224 // In particular it makes a difference during garbage collection,
225 // where you only want to traverse each thread profiler once.
226 static ThreadProfiler* get_thread_profiler() KERNEL_RETURN_(return NULL;);
227
228 // Garbage Collection Support
229 static void oops_do(OopClosure* f) KERNEL_RETURN ;
230
231 // Support for disassembler to inspect the PCRecorder
232
233 // Returns the start address for a given pc
234 // NULL is returned if the PCRecorder is inactive
235 static address bucket_start_for(address pc) KERNEL_RETURN_(return NULL;);
236
237 enum { MillisecsPerTick = 10 }; // ms per profiling ticks
238
239 // Returns the number of ticks recorded for the bucket
240 // pc belongs to.
241 static int bucket_count_for(address pc) KERNEL_RETURN_(return 0;);
242
243 #ifndef FPROF_KERNEL
244
245 private:
246 static bool full_profile() {
247 return full_profile_flag;
248 }
249
250 friend class ThreadProfiler;
251 // the following group of ticks cover everything that's not attributed to individual Java methods
252 static int received_gc_ticks; // ticks during which gc was active
253 static int vm_operation_ticks; // total ticks in vm_operations other than GC
254 static int threads_lock_ticks; // the number of times we couldn't get the Threads_lock without blocking
255 static int blocked_ticks; // ticks when the thread was blocked.
256 static int class_loader_ticks; // total ticks in class loader
257 static int extra_ticks; // total ticks an extra temporary measuring
258 static int compiler_ticks; // total ticks in compilation
259 static int interpreter_ticks; // ticks in unknown interpreted method
260 static int deopt_ticks; // ticks in deoptimization
261 static int unknown_ticks; // ticks that cannot be categorized
262 static int received_ticks; // ticks that were received by task
263 static int delivered_ticks; // ticks that were delivered by task
264 static int non_method_ticks() {
265 return
266 ( received_gc_ticks
267 + vm_operation_ticks
268 + deopt_ticks
269 + threads_lock_ticks
270 + blocked_ticks
271 + compiler_ticks
272 + interpreter_ticks
273 + unknown_ticks );
274 }
275 static elapsedTimer timer;
276
277 // Counts of each of the byte codes
278 static int* bytecode_ticks;
279 static int* bytecode_ticks_stub;
280 static void print_byte_code_statistics();
281
282 // the ticks below are for continuous profiling (to adjust recompilation, etc.)
283 static int all_ticks; // total count of ticks received so far
284 static int all_int_ticks; // ticks in interpreter
285 static int all_comp_ticks; // ticks in compiled code (+ native)
286 static bool full_profile_flag; // collecting full profile?
287
288 // to accumulate thread-specific data
289 // if we aren't profiling individual threads.
290 static ThreadProfiler* thread_profiler;
291 static ThreadProfiler* vm_thread_profiler;
292
293 static void allocate_table();
294
295 // The task that periodically interrupts things.
296 friend class FlatProfilerTask;
297 static FlatProfilerTask* task;
298 static void record_vm_operation();
299 static void record_vm_tick();
300 static void record_thread_ticks();
301
302 // For interval analysis
303 private:
304 static int interval_ticks_previous; // delivered_ticks from the last interval
305 static void interval_record_thread(ThreadProfiler* tp); // extract ticks from ThreadProfiler.
306 static void interval_print(); // print interval data.
307 static void interval_reset(); // reset interval data.
308 enum {interval_print_size = 10};
309 static IntervalData* interval_data;
310 #endif // FPROF_KERNEL
311 };