2012-06-28 17:03:16 -04:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2012, Oracle and/or its affiliates. All rights reserved.
|
|
|
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
|
|
*
|
|
|
|
* This code is free software; you can redistribute it and/or modify it
|
|
|
|
* under the terms of the GNU General Public License version 2 only, as
|
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*
|
|
|
|
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
|
|
* version 2 for more details (a copy is included in the LICENSE file that
|
|
|
|
* accompanied this code).
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License version
|
|
|
|
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
|
|
*
|
|
|
|
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
|
|
* or visit www.oracle.com if you need additional information or have any
|
|
|
|
* questions.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
#include "precompiled.hpp"
|
|
|
|
#include "memory/allocation.hpp"
|
|
|
|
#include "services/memBaseline.hpp"
|
|
|
|
#include "services/memTracker.hpp"
|
|
|
|
|
|
|
|
MemType2Name MemBaseline::MemType2NameMap[NUMBER_OF_MEMORY_TYPE] = {
|
|
|
|
{mtJavaHeap, "Java Heap"},
|
|
|
|
{mtClass, "Class"},
|
|
|
|
{mtThreadStack,"Thread Stack"},
|
|
|
|
{mtThread, "Thread"},
|
|
|
|
{mtCode, "Code"},
|
|
|
|
{mtGC, "GC"},
|
|
|
|
{mtCompiler, "Compiler"},
|
|
|
|
{mtInternal, "Internal"},
|
|
|
|
{mtOther, "Other"},
|
|
|
|
{mtSymbol, "Symbol"},
|
|
|
|
{mtNMT, "Memory Tracking"},
|
|
|
|
{mtChunk, "Pooled Free Chunks"},
|
2012-10-19 21:40:07 -04:00
|
|
|
{mtClassShared,"Shared spaces for classes"},
|
2013-02-01 23:48:08 +01:00
|
|
|
{mtTest, "Test"},
|
2012-06-28 17:03:16 -04:00
|
|
|
{mtNone, "Unknown"} // It can happen when type tagging records are lagging
|
|
|
|
// behind
|
|
|
|
};
|
|
|
|
|
|
|
|
MemBaseline::MemBaseline() {
|
|
|
|
_baselined = false;
|
|
|
|
|
|
|
|
for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
|
|
|
|
_malloc_data[index].set_type(MemType2NameMap[index]._flag);
|
|
|
|
_vm_data[index].set_type(MemType2NameMap[index]._flag);
|
|
|
|
_arena_data[index].set_type(MemType2NameMap[index]._flag);
|
|
|
|
}
|
|
|
|
|
|
|
|
_malloc_cs = NULL;
|
|
|
|
_vm_cs = NULL;
|
2012-10-19 21:40:07 -04:00
|
|
|
_vm_map = NULL;
|
2012-06-28 17:03:16 -04:00
|
|
|
|
|
|
|
_number_of_classes = 0;
|
|
|
|
_number_of_threads = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void MemBaseline::clear() {
|
|
|
|
if (_malloc_cs != NULL) {
|
|
|
|
delete _malloc_cs;
|
|
|
|
_malloc_cs = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (_vm_cs != NULL) {
|
|
|
|
delete _vm_cs;
|
|
|
|
_vm_cs = NULL;
|
|
|
|
}
|
|
|
|
|
2012-10-19 21:40:07 -04:00
|
|
|
if (_vm_map != NULL) {
|
|
|
|
delete _vm_map;
|
|
|
|
_vm_map = NULL;
|
|
|
|
}
|
|
|
|
|
2012-06-28 17:03:16 -04:00
|
|
|
reset();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void MemBaseline::reset() {
|
|
|
|
_baselined = false;
|
|
|
|
_total_vm_reserved = 0;
|
|
|
|
_total_vm_committed = 0;
|
|
|
|
_total_malloced = 0;
|
|
|
|
_number_of_classes = 0;
|
|
|
|
|
|
|
|
if (_malloc_cs != NULL) _malloc_cs->clear();
|
|
|
|
if (_vm_cs != NULL) _vm_cs->clear();
|
2012-10-19 21:40:07 -04:00
|
|
|
if (_vm_map != NULL) _vm_map->clear();
|
2012-06-28 17:03:16 -04:00
|
|
|
|
|
|
|
for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
|
|
|
|
_malloc_data[index].clear();
|
|
|
|
_vm_data[index].clear();
|
|
|
|
_arena_data[index].clear();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
MemBaseline::~MemBaseline() {
|
2012-10-19 21:40:07 -04:00
|
|
|
clear();
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
// baseline malloc'd memory records, generate overall summary and summaries by
|
|
|
|
// memory types
|
|
|
|
bool MemBaseline::baseline_malloc_summary(const MemPointerArray* malloc_records) {
|
2012-10-19 21:40:07 -04:00
|
|
|
MemPointerArrayIteratorImpl malloc_itr((MemPointerArray*)malloc_records);
|
|
|
|
MemPointerRecord* malloc_ptr = (MemPointerRecord*)malloc_itr.current();
|
2012-06-28 17:03:16 -04:00
|
|
|
size_t used_arena_size = 0;
|
|
|
|
int index;
|
2012-10-19 21:40:07 -04:00
|
|
|
while (malloc_ptr != NULL) {
|
|
|
|
index = flag2index(FLAGS_TO_MEMORY_TYPE(malloc_ptr->flags()));
|
|
|
|
size_t size = malloc_ptr->size();
|
2012-11-09 19:24:31 -05:00
|
|
|
if (malloc_ptr->is_arena_memory_record()) {
|
|
|
|
// We do have anonymous arenas, they are either used as value objects,
|
|
|
|
// which are embedded inside other objects, or used as stack objects.
|
|
|
|
_arena_data[index].inc(size);
|
|
|
|
used_arena_size += size;
|
|
|
|
} else {
|
|
|
|
_total_malloced += size;
|
|
|
|
_malloc_data[index].inc(size);
|
|
|
|
if (malloc_ptr->is_arena_record()) {
|
|
|
|
// see if arena memory record present
|
|
|
|
MemPointerRecord* next_malloc_ptr = (MemPointerRecordEx*)malloc_itr.peek_next();
|
|
|
|
if (next_malloc_ptr->is_arena_memory_record()) {
|
|
|
|
assert(next_malloc_ptr->is_memory_record_of_arena(malloc_ptr),
|
|
|
|
"Arena records do not match");
|
|
|
|
size = next_malloc_ptr->size();
|
|
|
|
_arena_data[index].inc(size);
|
|
|
|
used_arena_size += size;
|
|
|
|
malloc_itr.next();
|
|
|
|
}
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
malloc_ptr = (MemPointerRecordEx*)malloc_itr.next();
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
// substract used arena size to get size of arena chunk in free list
|
|
|
|
index = flag2index(mtChunk);
|
|
|
|
_malloc_data[index].reduce(used_arena_size);
|
|
|
|
// we really don't know how many chunks in free list, so just set to
|
|
|
|
// 0
|
|
|
|
_malloc_data[index].overwrite_counter(0);
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// baseline mmap'd memory records, generate overall summary and summaries by
|
|
|
|
// memory types
|
|
|
|
bool MemBaseline::baseline_vm_summary(const MemPointerArray* vm_records) {
|
2012-10-19 21:40:07 -04:00
|
|
|
MemPointerArrayIteratorImpl vm_itr((MemPointerArray*)vm_records);
|
|
|
|
VMMemRegion* vm_ptr = (VMMemRegion*)vm_itr.current();
|
2012-06-28 17:03:16 -04:00
|
|
|
int index;
|
2012-10-19 21:40:07 -04:00
|
|
|
while (vm_ptr != NULL) {
|
|
|
|
if (vm_ptr->is_reserved_region()) {
|
|
|
|
index = flag2index(FLAGS_TO_MEMORY_TYPE(vm_ptr->flags()));
|
2012-06-28 17:03:16 -04:00
|
|
|
// we use the number of thread stack to count threads
|
2012-10-19 21:40:07 -04:00
|
|
|
if (IS_MEMORY_TYPE(vm_ptr->flags(), mtThreadStack)) {
|
2012-06-28 17:03:16 -04:00
|
|
|
_number_of_threads ++;
|
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
_total_vm_reserved += vm_ptr->size();
|
|
|
|
_vm_data[index].inc(vm_ptr->size(), 0);
|
|
|
|
} else {
|
|
|
|
_total_vm_committed += vm_ptr->size();
|
|
|
|
_vm_data[index].inc(0, vm_ptr->size());
|
|
|
|
}
|
|
|
|
vm_ptr = (VMMemRegion*)vm_itr.next();
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// baseline malloc'd memory by callsites, but only the callsites with memory allocation
|
|
|
|
// over 1KB are stored.
|
|
|
|
bool MemBaseline::baseline_malloc_details(const MemPointerArray* malloc_records) {
|
|
|
|
assert(MemTracker::track_callsite(), "detail tracking is off");
|
|
|
|
|
2012-10-19 21:40:07 -04:00
|
|
|
MemPointerArrayIteratorImpl malloc_itr(const_cast<MemPointerArray*>(malloc_records));
|
|
|
|
MemPointerRecordEx* malloc_ptr = (MemPointerRecordEx*)malloc_itr.current();
|
|
|
|
MallocCallsitePointer malloc_callsite;
|
2012-06-28 17:03:16 -04:00
|
|
|
|
2012-10-19 21:40:07 -04:00
|
|
|
// initailize malloc callsite array
|
2012-06-28 17:03:16 -04:00
|
|
|
if (_malloc_cs == NULL) {
|
|
|
|
_malloc_cs = new (std::nothrow) MemPointerArrayImpl<MallocCallsitePointer>(64);
|
|
|
|
// out of native memory
|
2012-10-19 21:40:07 -04:00
|
|
|
if (_malloc_cs == NULL || _malloc_cs->out_of_memory()) {
|
2012-06-28 17:03:16 -04:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
_malloc_cs->clear();
|
|
|
|
}
|
|
|
|
|
2012-10-19 21:40:07 -04:00
|
|
|
MemPointerArray* malloc_data = const_cast<MemPointerArray*>(malloc_records);
|
|
|
|
|
|
|
|
// sort into callsite pc order. Details are aggregated by callsites
|
|
|
|
malloc_data->sort((FN_SORT)malloc_sort_by_pc);
|
|
|
|
bool ret = true;
|
|
|
|
|
2012-06-28 17:03:16 -04:00
|
|
|
// baseline memory that is totaled over 1 KB
|
2012-10-19 21:40:07 -04:00
|
|
|
while (malloc_ptr != NULL) {
|
2012-11-09 19:24:31 -05:00
|
|
|
if (!MemPointerRecord::is_arena_memory_record(malloc_ptr->flags())) {
|
2012-06-28 17:03:16 -04:00
|
|
|
// skip thread stacks
|
2012-10-19 21:40:07 -04:00
|
|
|
if (!IS_MEMORY_TYPE(malloc_ptr->flags(), mtThreadStack)) {
|
|
|
|
if (malloc_callsite.addr() != malloc_ptr->pc()) {
|
|
|
|
if ((malloc_callsite.amount()/K) > 0) {
|
|
|
|
if (!_malloc_cs->append(&malloc_callsite)) {
|
|
|
|
ret = false;
|
|
|
|
break;
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
malloc_callsite = MallocCallsitePointer(malloc_ptr->pc());
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
malloc_callsite.inc(malloc_ptr->size());
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
malloc_ptr = (MemPointerRecordEx*)malloc_itr.next();
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
|
|
|
|
2012-10-19 21:40:07 -04:00
|
|
|
// restore to address order. Snapshot malloc data is maintained in memory
|
|
|
|
// address order.
|
|
|
|
malloc_data->sort((FN_SORT)malloc_sort_by_addr);
|
|
|
|
|
|
|
|
if (!ret) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
// deal with last record
|
|
|
|
if (malloc_callsite.addr() != 0 && (malloc_callsite.amount()/K) > 0) {
|
|
|
|
if (!_malloc_cs->append(&malloc_callsite)) {
|
2012-06-28 17:03:16 -04:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// baseline mmap'd memory by callsites
|
|
|
|
bool MemBaseline::baseline_vm_details(const MemPointerArray* vm_records) {
|
|
|
|
assert(MemTracker::track_callsite(), "detail tracking is off");
|
|
|
|
|
2012-10-19 21:40:07 -04:00
|
|
|
VMCallsitePointer vm_callsite;
|
|
|
|
VMCallsitePointer* cur_callsite = NULL;
|
|
|
|
MemPointerArrayIteratorImpl vm_itr((MemPointerArray*)vm_records);
|
|
|
|
VMMemRegionEx* vm_ptr = (VMMemRegionEx*)vm_itr.current();
|
|
|
|
|
|
|
|
// initialize virtual memory map array
|
|
|
|
if (_vm_map == NULL) {
|
|
|
|
_vm_map = new (std::nothrow) MemPointerArrayImpl<VMMemRegionEx>(vm_records->length());
|
|
|
|
if (_vm_map == NULL || _vm_map->out_of_memory()) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
_vm_map->clear();
|
|
|
|
}
|
2012-06-28 17:03:16 -04:00
|
|
|
|
2012-10-19 21:40:07 -04:00
|
|
|
// initialize virtual memory callsite array
|
2012-06-28 17:03:16 -04:00
|
|
|
if (_vm_cs == NULL) {
|
|
|
|
_vm_cs = new (std::nothrow) MemPointerArrayImpl<VMCallsitePointer>(64);
|
2012-10-19 21:40:07 -04:00
|
|
|
if (_vm_cs == NULL || _vm_cs->out_of_memory()) {
|
2012-06-28 17:03:16 -04:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
_vm_cs->clear();
|
|
|
|
}
|
|
|
|
|
2012-10-19 21:40:07 -04:00
|
|
|
// consolidate virtual memory data
|
|
|
|
VMMemRegionEx* reserved_rec = NULL;
|
|
|
|
VMMemRegionEx* committed_rec = NULL;
|
|
|
|
|
|
|
|
// vm_ptr is coming in increasing base address order
|
|
|
|
while (vm_ptr != NULL) {
|
|
|
|
if (vm_ptr->is_reserved_region()) {
|
|
|
|
// consolidate reserved memory regions for virtual memory map.
|
|
|
|
// The criteria for consolidation is:
|
|
|
|
// 1. two adjacent reserved memory regions
|
|
|
|
// 2. belong to the same memory type
|
|
|
|
// 3. reserved from the same callsite
|
|
|
|
if (reserved_rec == NULL ||
|
|
|
|
reserved_rec->base() + reserved_rec->size() != vm_ptr->addr() ||
|
|
|
|
FLAGS_TO_MEMORY_TYPE(reserved_rec->flags()) != FLAGS_TO_MEMORY_TYPE(vm_ptr->flags()) ||
|
|
|
|
reserved_rec->pc() != vm_ptr->pc()) {
|
|
|
|
if (!_vm_map->append(vm_ptr)) {
|
2012-06-28 17:03:16 -04:00
|
|
|
return false;
|
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
// inserted reserved region, we need the pointer to the element in virtual
|
|
|
|
// memory map array.
|
|
|
|
reserved_rec = (VMMemRegionEx*)_vm_map->at(_vm_map->length() - 1);
|
|
|
|
} else {
|
|
|
|
reserved_rec->expand_region(vm_ptr->addr(), vm_ptr->size());
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
|
|
|
|
if (cur_callsite != NULL && !_vm_cs->append(cur_callsite)) {
|
2012-06-28 17:03:16 -04:00
|
|
|
return false;
|
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
vm_callsite = VMCallsitePointer(vm_ptr->pc());
|
|
|
|
cur_callsite = &vm_callsite;
|
|
|
|
vm_callsite.inc(vm_ptr->size(), 0);
|
|
|
|
} else {
|
|
|
|
// consolidate committed memory regions for virtual memory map
|
|
|
|
// The criterial is:
|
|
|
|
// 1. two adjacent committed memory regions
|
|
|
|
// 2. committed from the same callsite
|
|
|
|
if (committed_rec == NULL ||
|
|
|
|
committed_rec->base() + committed_rec->size() != vm_ptr->addr() ||
|
|
|
|
committed_rec->pc() != vm_ptr->pc()) {
|
|
|
|
if (!_vm_map->append(vm_ptr)) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
committed_rec = (VMMemRegionEx*)_vm_map->at(_vm_map->length() - 1);
|
|
|
|
} else {
|
|
|
|
committed_rec->expand_region(vm_ptr->addr(), vm_ptr->size());
|
|
|
|
}
|
|
|
|
vm_callsite.inc(0, vm_ptr->size());
|
|
|
|
}
|
|
|
|
vm_ptr = (VMMemRegionEx*)vm_itr.next();
|
2012-06-28 17:03:16 -04:00
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
// deal with last record
|
|
|
|
if (cur_callsite != NULL && !_vm_cs->append(cur_callsite)) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// sort it into callsite pc order. Details are aggregated by callsites
|
|
|
|
_vm_cs->sort((FN_SORT)bl_vm_sort_by_pc);
|
|
|
|
|
|
|
|
// walk the array to consolidate record by pc
|
|
|
|
MemPointerArrayIteratorImpl itr(_vm_cs);
|
|
|
|
VMCallsitePointer* callsite_rec = (VMCallsitePointer*)itr.current();
|
|
|
|
VMCallsitePointer* next_rec = (VMCallsitePointer*)itr.next();
|
|
|
|
while (next_rec != NULL) {
|
|
|
|
assert(callsite_rec != NULL, "Sanity check");
|
|
|
|
if (next_rec->addr() == callsite_rec->addr()) {
|
|
|
|
callsite_rec->inc(next_rec->reserved_amount(), next_rec->committed_amount());
|
|
|
|
itr.remove();
|
|
|
|
next_rec = (VMCallsitePointer*)itr.current();
|
|
|
|
} else {
|
|
|
|
callsite_rec = next_rec;
|
|
|
|
next_rec = (VMCallsitePointer*)itr.next();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-06-28 17:03:16 -04:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// baseline a snapshot. If summary_only = false, memory usages aggregated by
|
|
|
|
// callsites are also baselined.
|
|
|
|
bool MemBaseline::baseline(MemSnapshot& snapshot, bool summary_only) {
|
|
|
|
MutexLockerEx snapshot_locker(snapshot._lock, true);
|
|
|
|
reset();
|
|
|
|
_baselined = baseline_malloc_summary(snapshot._alloc_ptrs) &&
|
|
|
|
baseline_vm_summary(snapshot._vm_ptrs);
|
2013-01-08 14:04:25 -05:00
|
|
|
_number_of_classes = snapshot.number_of_classes();
|
2012-06-28 17:03:16 -04:00
|
|
|
|
|
|
|
if (!summary_only && MemTracker::track_callsite() && _baselined) {
|
|
|
|
_baselined = baseline_malloc_details(snapshot._alloc_ptrs) &&
|
|
|
|
baseline_vm_details(snapshot._vm_ptrs);
|
|
|
|
}
|
|
|
|
return _baselined;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int MemBaseline::flag2index(MEMFLAGS flag) const {
|
|
|
|
for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
|
|
|
|
if (MemType2NameMap[index]._flag == flag) {
|
|
|
|
return index;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
assert(false, "no type");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
const char* MemBaseline::type2name(MEMFLAGS type) {
|
|
|
|
for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
|
|
|
|
if (MemType2NameMap[index]._flag == type) {
|
|
|
|
return MemType2NameMap[index]._name;
|
|
|
|
}
|
|
|
|
}
|
2012-10-19 21:40:07 -04:00
|
|
|
assert(false, err_msg("bad type %x", type));
|
2012-06-28 17:03:16 -04:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
MemBaseline& MemBaseline::operator=(const MemBaseline& other) {
|
|
|
|
_total_malloced = other._total_malloced;
|
|
|
|
_total_vm_reserved = other._total_vm_reserved;
|
|
|
|
_total_vm_committed = other._total_vm_committed;
|
|
|
|
|
|
|
|
_baselined = other._baselined;
|
|
|
|
_number_of_classes = other._number_of_classes;
|
|
|
|
|
|
|
|
for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) {
|
|
|
|
_malloc_data[index] = other._malloc_data[index];
|
|
|
|
_vm_data[index] = other._vm_data[index];
|
|
|
|
_arena_data[index] = other._arena_data[index];
|
|
|
|
}
|
|
|
|
|
|
|
|
if (MemTracker::track_callsite()) {
|
|
|
|
assert(_malloc_cs != NULL && _vm_cs != NULL, "out of memory");
|
|
|
|
assert(other._malloc_cs != NULL && other._vm_cs != NULL,
|
|
|
|
"not properly baselined");
|
|
|
|
_malloc_cs->clear();
|
|
|
|
_vm_cs->clear();
|
|
|
|
int index;
|
|
|
|
for (index = 0; index < other._malloc_cs->length(); index ++) {
|
|
|
|
_malloc_cs->append(other._malloc_cs->at(index));
|
|
|
|
}
|
|
|
|
|
|
|
|
for (index = 0; index < other._vm_cs->length(); index ++) {
|
|
|
|
_vm_cs->append(other._vm_cs->at(index));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return *this;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* compare functions for sorting */
|
|
|
|
|
|
|
|
// sort snapshot malloc'd records in callsite pc order
|
|
|
|
int MemBaseline::malloc_sort_by_pc(const void* p1, const void* p2) {
|
|
|
|
assert(MemTracker::track_callsite(),"Just check");
|
|
|
|
const MemPointerRecordEx* mp1 = (const MemPointerRecordEx*)p1;
|
|
|
|
const MemPointerRecordEx* mp2 = (const MemPointerRecordEx*)p2;
|
|
|
|
return UNSIGNED_COMPARE(mp1->pc(), mp2->pc());
|
|
|
|
}
|
|
|
|
|
|
|
|
// sort baselined malloc'd records in size order
|
|
|
|
int MemBaseline::bl_malloc_sort_by_size(const void* p1, const void* p2) {
|
|
|
|
assert(MemTracker::is_on(), "Just check");
|
|
|
|
const MallocCallsitePointer* mp1 = (const MallocCallsitePointer*)p1;
|
|
|
|
const MallocCallsitePointer* mp2 = (const MallocCallsitePointer*)p2;
|
|
|
|
return UNSIGNED_COMPARE(mp2->amount(), mp1->amount());
|
|
|
|
}
|
|
|
|
|
|
|
|
// sort baselined malloc'd records in callsite pc order
|
|
|
|
int MemBaseline::bl_malloc_sort_by_pc(const void* p1, const void* p2) {
|
|
|
|
assert(MemTracker::is_on(), "Just check");
|
|
|
|
const MallocCallsitePointer* mp1 = (const MallocCallsitePointer*)p1;
|
|
|
|
const MallocCallsitePointer* mp2 = (const MallocCallsitePointer*)p2;
|
|
|
|
return UNSIGNED_COMPARE(mp1->addr(), mp2->addr());
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// sort baselined mmap'd records in size (reserved size) order
|
|
|
|
int MemBaseline::bl_vm_sort_by_size(const void* p1, const void* p2) {
|
|
|
|
assert(MemTracker::is_on(), "Just check");
|
|
|
|
const VMCallsitePointer* mp1 = (const VMCallsitePointer*)p1;
|
|
|
|
const VMCallsitePointer* mp2 = (const VMCallsitePointer*)p2;
|
|
|
|
return UNSIGNED_COMPARE(mp2->reserved_amount(), mp1->reserved_amount());
|
|
|
|
}
|
|
|
|
|
|
|
|
// sort baselined mmap'd records in callsite pc order
|
|
|
|
int MemBaseline::bl_vm_sort_by_pc(const void* p1, const void* p2) {
|
|
|
|
assert(MemTracker::is_on(), "Just check");
|
|
|
|
const VMCallsitePointer* mp1 = (const VMCallsitePointer*)p1;
|
|
|
|
const VMCallsitePointer* mp2 = (const VMCallsitePointer*)p2;
|
|
|
|
return UNSIGNED_COMPARE(mp1->addr(), mp2->addr());
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// sort snapshot malloc'd records in memory block address order
|
|
|
|
int MemBaseline::malloc_sort_by_addr(const void* p1, const void* p2) {
|
|
|
|
assert(MemTracker::is_on(), "Just check");
|
|
|
|
const MemPointerRecord* mp1 = (const MemPointerRecord*)p1;
|
|
|
|
const MemPointerRecord* mp2 = (const MemPointerRecord*)p2;
|
|
|
|
int delta = UNSIGNED_COMPARE(mp1->addr(), mp2->addr());
|
|
|
|
assert(delta != 0, "dup pointer");
|
|
|
|
return delta;
|
|
|
|
}
|
|
|
|
|