comparison src/share/vm/services/memBaseline.cpp @ 6197:d2a62e0f25eb

6995781: Native Memory Tracking (Phase 1) 7151532: DCmd for hotspot native memory tracking Summary: Implementation of native memory tracking phase 1, which tracks VM native memory usage, and related DCmd Reviewed-by: acorn, coleenp, fparain
author zgu
date Thu, 28 Jun 2012 17:03:16 -0400
parents
children 716c64bda5ba
comparison
equal deleted inserted replaced
6174:74533f63b116 6197:d2a62e0f25eb
1 /*
2 * Copyright (c) 2012, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24 #include "precompiled.hpp"
25 #include "classfile/systemDictionary.hpp"
26 #include "memory/allocation.hpp"
27 #include "services/memBaseline.hpp"
28 #include "services/memTracker.hpp"
29
30 MemType2Name MemBaseline::MemType2NameMap[NUMBER_OF_MEMORY_TYPE] = {
31 {mtJavaHeap, "Java Heap"},
32 {mtClass, "Class"},
33 {mtThreadStack,"Thread Stack"},
34 {mtThread, "Thread"},
35 {mtCode, "Code"},
36 {mtGC, "GC"},
37 {mtCompiler, "Compiler"},
38 {mtInternal, "Internal"},
39 {mtOther, "Other"},
40 {mtSymbol, "Symbol"},
41 {mtNMT, "Memory Tracking"},
42 {mtChunk, "Pooled Free Chunks"},
43 {mtNone, "Unknown"} // It can happen when type tagging records are lagging
44 // behind
45 };
46
47 MemBaseline::MemBaseline() {
48 _baselined = false;
49
50 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
51 _malloc_data[index].set_type(MemType2NameMap[index]._flag);
52 _vm_data[index].set_type(MemType2NameMap[index]._flag);
53 _arena_data[index].set_type(MemType2NameMap[index]._flag);
54 }
55
56 _malloc_cs = NULL;
57 _vm_cs = NULL;
58
59 _number_of_classes = 0;
60 _number_of_threads = 0;
61 }
62
63
64 void MemBaseline::clear() {
65 if (_malloc_cs != NULL) {
66 delete _malloc_cs;
67 _malloc_cs = NULL;
68 }
69
70 if (_vm_cs != NULL) {
71 delete _vm_cs;
72 _vm_cs = NULL;
73 }
74
75 reset();
76 }
77
78
79 void MemBaseline::reset() {
80 _baselined = false;
81 _total_vm_reserved = 0;
82 _total_vm_committed = 0;
83 _total_malloced = 0;
84 _number_of_classes = 0;
85
86 if (_malloc_cs != NULL) _malloc_cs->clear();
87 if (_vm_cs != NULL) _vm_cs->clear();
88
89 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
90 _malloc_data[index].clear();
91 _vm_data[index].clear();
92 _arena_data[index].clear();
93 }
94 }
95
96 MemBaseline::~MemBaseline() {
97 if (_malloc_cs != NULL) {
98 delete _malloc_cs;
99 }
100
101 if (_vm_cs != NULL) {
102 delete _vm_cs;
103 }
104 }
105
106 // baseline malloc'd memory records, generate overall summary and summaries by
107 // memory types
108 bool MemBaseline::baseline_malloc_summary(const MemPointerArray* malloc_records) {
109 MemPointerArrayIteratorImpl mItr((MemPointerArray*)malloc_records);
110 MemPointerRecord* mptr = (MemPointerRecord*)mItr.current();
111 size_t used_arena_size = 0;
112 int index;
113 while (mptr != NULL) {
114 index = flag2index(FLAGS_TO_MEMORY_TYPE(mptr->flags()));
115 size_t size = mptr->size();
116 _total_malloced += size;
117 _malloc_data[index].inc(size);
118 if (MemPointerRecord::is_arena_record(mptr->flags())) {
119 // see if arena size record present
120 MemPointerRecord* next_p = (MemPointerRecordEx*)mItr.peek_next();
121 if (MemPointerRecord::is_arena_size_record(next_p->flags())) {
122 assert(next_p->is_size_record_of_arena(mptr), "arena records do not match");
123 size = next_p->size();
124 _arena_data[index].inc(size);
125 used_arena_size += size;
126 mItr.next();
127 }
128 }
129 mptr = (MemPointerRecordEx*)mItr.next();
130 }
131
132 // substract used arena size to get size of arena chunk in free list
133 index = flag2index(mtChunk);
134 _malloc_data[index].reduce(used_arena_size);
135 // we really don't know how many chunks in free list, so just set to
136 // 0
137 _malloc_data[index].overwrite_counter(0);
138
139 return true;
140 }
141
142 // baseline mmap'd memory records, generate overall summary and summaries by
143 // memory types
144 bool MemBaseline::baseline_vm_summary(const MemPointerArray* vm_records) {
145 MemPointerArrayIteratorImpl vItr((MemPointerArray*)vm_records);
146 VMMemRegion* vptr = (VMMemRegion*)vItr.current();
147 int index;
148 while (vptr != NULL) {
149 index = flag2index(FLAGS_TO_MEMORY_TYPE(vptr->flags()));
150
151 // we use the number of thread stack to count threads
152 if (IS_MEMORY_TYPE(vptr->flags(), mtThreadStack)) {
153 _number_of_threads ++;
154 }
155 _total_vm_reserved += vptr->reserved_size();
156 _total_vm_committed += vptr->committed_size();
157 _vm_data[index].inc(vptr->reserved_size(), vptr->committed_size());
158 vptr = (VMMemRegion*)vItr.next();
159 }
160 return true;
161 }
162
163 // baseline malloc'd memory by callsites, but only the callsites with memory allocation
164 // over 1KB are stored.
165 bool MemBaseline::baseline_malloc_details(const MemPointerArray* malloc_records) {
166 assert(MemTracker::track_callsite(), "detail tracking is off");
167
168 MemPointerArrayIteratorImpl mItr((MemPointerArray*)malloc_records);
169 MemPointerRecordEx* mptr = (MemPointerRecordEx*)mItr.current();
170 MallocCallsitePointer mp;
171
172 if (_malloc_cs == NULL) {
173 _malloc_cs = new (std::nothrow) MemPointerArrayImpl<MallocCallsitePointer>(64);
174 // out of native memory
175 if (_malloc_cs == NULL) {
176 return false;
177 }
178 } else {
179 _malloc_cs->clear();
180 }
181
182 // baseline memory that is totaled over 1 KB
183 while (mptr != NULL) {
184 if (!MemPointerRecord::is_arena_size_record(mptr->flags())) {
185 // skip thread stacks
186 if (!IS_MEMORY_TYPE(mptr->flags(), mtThreadStack)) {
187 if (mp.addr() != mptr->pc()) {
188 if ((mp.amount()/K) > 0) {
189 if (!_malloc_cs->append(&mp)) {
190 return false;
191 }
192 }
193 mp = MallocCallsitePointer(mptr->pc());
194 }
195 mp.inc(mptr->size());
196 }
197 }
198 mptr = (MemPointerRecordEx*)mItr.next();
199 }
200
201 if (mp.addr() != 0 && (mp.amount()/K) > 0) {
202 if (!_malloc_cs->append(&mp)) {
203 return false;
204 }
205 }
206 return true;
207 }
208
209 // baseline mmap'd memory by callsites
210 bool MemBaseline::baseline_vm_details(const MemPointerArray* vm_records) {
211 assert(MemTracker::track_callsite(), "detail tracking is off");
212
213 VMCallsitePointer vp;
214 MemPointerArrayIteratorImpl vItr((MemPointerArray*)vm_records);
215 VMMemRegionEx* vptr = (VMMemRegionEx*)vItr.current();
216
217 if (_vm_cs == NULL) {
218 _vm_cs = new (std::nothrow) MemPointerArrayImpl<VMCallsitePointer>(64);
219 if (_vm_cs == NULL) {
220 return false;
221 }
222 } else {
223 _vm_cs->clear();
224 }
225
226 while (vptr != NULL) {
227 if (vp.addr() != vptr->pc()) {
228 if (!_vm_cs->append(&vp)) {
229 return false;
230 }
231 vp = VMCallsitePointer(vptr->pc());
232 }
233 vp.inc(vptr->size(), vptr->committed_size());
234 vptr = (VMMemRegionEx*)vItr.next();
235 }
236 if (vp.addr() != 0) {
237 if (!_vm_cs->append(&vp)) {
238 return false;
239 }
240 }
241 return true;
242 }
243
244 // baseline a snapshot. If summary_only = false, memory usages aggregated by
245 // callsites are also baselined.
246 bool MemBaseline::baseline(MemSnapshot& snapshot, bool summary_only) {
247 MutexLockerEx snapshot_locker(snapshot._lock, true);
248 reset();
249 _baselined = baseline_malloc_summary(snapshot._alloc_ptrs) &&
250 baseline_vm_summary(snapshot._vm_ptrs);
251 _number_of_classes = SystemDictionary::number_of_classes();
252
253 if (!summary_only && MemTracker::track_callsite() && _baselined) {
254 ((MemPointerArray*)snapshot._alloc_ptrs)->sort((FN_SORT)malloc_sort_by_pc);
255 ((MemPointerArray*)snapshot._vm_ptrs)->sort((FN_SORT)vm_sort_by_pc);
256 _baselined = baseline_malloc_details(snapshot._alloc_ptrs) &&
257 baseline_vm_details(snapshot._vm_ptrs);
258 ((MemPointerArray*)snapshot._alloc_ptrs)->sort((FN_SORT)malloc_sort_by_addr);
259 ((MemPointerArray*)snapshot._vm_ptrs)->sort((FN_SORT)vm_sort_by_addr);
260 }
261 return _baselined;
262 }
263
264
265 int MemBaseline::flag2index(MEMFLAGS flag) const {
266 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
267 if (MemType2NameMap[index]._flag == flag) {
268 return index;
269 }
270 }
271 assert(false, "no type");
272 return -1;
273 }
274
275 const char* MemBaseline::type2name(MEMFLAGS type) {
276 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
277 if (MemType2NameMap[index]._flag == type) {
278 return MemType2NameMap[index]._name;
279 }
280 }
281 assert(false, "no type");
282 return NULL;
283 }
284
285
286 MemBaseline& MemBaseline::operator=(const MemBaseline& other) {
287 _total_malloced = other._total_malloced;
288 _total_vm_reserved = other._total_vm_reserved;
289 _total_vm_committed = other._total_vm_committed;
290
291 _baselined = other._baselined;
292 _number_of_classes = other._number_of_classes;
293
294 for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
295 _malloc_data[index] = other._malloc_data[index];
296 _vm_data[index] = other._vm_data[index];
297 _arena_data[index] = other._arena_data[index];
298 }
299
300 if (MemTracker::track_callsite()) {
301 assert(_malloc_cs != NULL && _vm_cs != NULL, "out of memory");
302 assert(other._malloc_cs != NULL && other._vm_cs != NULL,
303 "not properly baselined");
304 _malloc_cs->clear();
305 _vm_cs->clear();
306 int index;
307 for (index = 0; index < other._malloc_cs->length(); index ++) {
308 _malloc_cs->append(other._malloc_cs->at(index));
309 }
310
311 for (index = 0; index < other._vm_cs->length(); index ++) {
312 _vm_cs->append(other._vm_cs->at(index));
313 }
314 }
315 return *this;
316 }
317
318 /* compare functions for sorting */
319
320 // sort snapshot malloc'd records in callsite pc order
321 int MemBaseline::malloc_sort_by_pc(const void* p1, const void* p2) {
322 assert(MemTracker::track_callsite(),"Just check");
323 const MemPointerRecordEx* mp1 = (const MemPointerRecordEx*)p1;
324 const MemPointerRecordEx* mp2 = (const MemPointerRecordEx*)p2;
325 return UNSIGNED_COMPARE(mp1->pc(), mp2->pc());
326 }
327
328 // sort baselined malloc'd records in size order
329 int MemBaseline::bl_malloc_sort_by_size(const void* p1, const void* p2) {
330 assert(MemTracker::is_on(), "Just check");
331 const MallocCallsitePointer* mp1 = (const MallocCallsitePointer*)p1;
332 const MallocCallsitePointer* mp2 = (const MallocCallsitePointer*)p2;
333 return UNSIGNED_COMPARE(mp2->amount(), mp1->amount());
334 }
335
336 // sort baselined malloc'd records in callsite pc order
337 int MemBaseline::bl_malloc_sort_by_pc(const void* p1, const void* p2) {
338 assert(MemTracker::is_on(), "Just check");
339 const MallocCallsitePointer* mp1 = (const MallocCallsitePointer*)p1;
340 const MallocCallsitePointer* mp2 = (const MallocCallsitePointer*)p2;
341 return UNSIGNED_COMPARE(mp1->addr(), mp2->addr());
342 }
343
344 // sort snapshot mmap'd records in callsite pc order
345 int MemBaseline::vm_sort_by_pc(const void* p1, const void* p2) {
346 assert(MemTracker::track_callsite(),"Just check");
347 const VMMemRegionEx* mp1 = (const VMMemRegionEx*)p1;
348 const VMMemRegionEx* mp2 = (const VMMemRegionEx*)p2;
349 return UNSIGNED_COMPARE(mp1->pc(), mp2->pc());
350 }
351
352 // sort baselined mmap'd records in size (reserved size) order
353 int MemBaseline::bl_vm_sort_by_size(const void* p1, const void* p2) {
354 assert(MemTracker::is_on(), "Just check");
355 const VMCallsitePointer* mp1 = (const VMCallsitePointer*)p1;
356 const VMCallsitePointer* mp2 = (const VMCallsitePointer*)p2;
357 return UNSIGNED_COMPARE(mp2->reserved_amount(), mp1->reserved_amount());
358 }
359
360 // sort baselined mmap'd records in callsite pc order
361 int MemBaseline::bl_vm_sort_by_pc(const void* p1, const void* p2) {
362 assert(MemTracker::is_on(), "Just check");
363 const VMCallsitePointer* mp1 = (const VMCallsitePointer*)p1;
364 const VMCallsitePointer* mp2 = (const VMCallsitePointer*)p2;
365 return UNSIGNED_COMPARE(mp1->addr(), mp2->addr());
366 }
367
368
369 // sort snapshot malloc'd records in memory block address order
370 int MemBaseline::malloc_sort_by_addr(const void* p1, const void* p2) {
371 assert(MemTracker::is_on(), "Just check");
372 const MemPointerRecord* mp1 = (const MemPointerRecord*)p1;
373 const MemPointerRecord* mp2 = (const MemPointerRecord*)p2;
374 int delta = UNSIGNED_COMPARE(mp1->addr(), mp2->addr());
375 assert(delta != 0, "dup pointer");
376 return delta;
377 }
378
379 // sort snapshot mmap'd records in memory block address order
380 int MemBaseline::vm_sort_by_addr(const void* p1, const void* p2) {
381 assert(MemTracker::is_on(), "Just check");
382 const VMMemRegion* mp1 = (const VMMemRegion*)p1;
383 const VMMemRegion* mp2 = (const VMMemRegion*)p2;
384 int delta = UNSIGNED_COMPARE(mp1->addr(), mp2->addr());
385 assert(delta != 0, "dup pointer");
386 return delta;
387 }