mirror of
https://gitlab.winehq.org/wine/wine-gecko.git
synced 2024-09-13 09:24:08 -07:00
435eec853f
--HG-- extra : rebase_source : d0b233a62c1b50facb27d781f9942628f100180e
2078 lines
55 KiB
C++
2078 lines
55 KiB
C++
/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
|
|
/* vim: set ts=8 sts=2 et sw=2 tw=80: */
|
|
/* This Source Code Form is subject to the terms of the Mozilla Public
|
|
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
|
|
|
|
#include "DMD.h"
|
|
|
|
#include <ctype.h>
|
|
#include <errno.h>
|
|
#include <limits.h>
|
|
#include <stdarg.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
#ifdef XP_WIN
|
|
#if defined(MOZ_OPTIMIZE) && !defined(MOZ_PROFILING)
|
|
#error "Optimized, DMD-enabled builds on Windows must be built with --enable-profiling"
|
|
#endif
|
|
#include <windows.h>
|
|
#include <process.h>
|
|
#else
|
|
#include <unistd.h>
|
|
#endif
|
|
|
|
#ifdef ANDROID
|
|
#include <android/log.h>
|
|
#endif
|
|
|
|
#include "nscore.h"
|
|
#include "nsStackWalk.h"
|
|
|
|
#include "js/HashTable.h"
|
|
#include "js/Vector.h"
|
|
|
|
#include "mozilla/Assertions.h"
|
|
#include "mozilla/HashFunctions.h"
|
|
#include "mozilla/IntegerPrintfMacros.h"
|
|
#include "mozilla/JSONWriter.h"
|
|
#include "mozilla/Likely.h"
|
|
#include "mozilla/MemoryReporting.h"
|
|
|
|
// CodeAddressService is defined entirely in the header, so this does not make
|
|
// DMD depend on XPCOM's object file.
|
|
#include "CodeAddressService.h"
|
|
|
|
// MOZ_REPLACE_ONLY_MEMALIGN saves us from having to define
|
|
// replace_{posix_memalign,aligned_alloc,valloc}. It requires defining
|
|
// PAGE_SIZE. Nb: sysconf() is expensive, but it's only used for (the obsolete
|
|
// and rarely used) valloc.
|
|
#define MOZ_REPLACE_ONLY_MEMALIGN 1
|
|
|
|
#ifdef XP_WIN
|
|
#define PAGE_SIZE GetPageSize()
|
|
static long GetPageSize()
|
|
{
|
|
SYSTEM_INFO si;
|
|
GetSystemInfo(&si);
|
|
return si.dwPageSize;
|
|
}
|
|
#else
|
|
#define PAGE_SIZE sysconf(_SC_PAGESIZE)
|
|
#endif
|
|
#include "replace_malloc.h"
|
|
#undef MOZ_REPLACE_ONLY_MEMALIGN
|
|
#undef PAGE_SIZE
|
|
|
|
namespace mozilla {
|
|
namespace dmd {
|
|
|
|
//---------------------------------------------------------------------------
|
|
// Utilities
|
|
//---------------------------------------------------------------------------
|
|
|
|
#ifndef DISALLOW_COPY_AND_ASSIGN
|
|
#define DISALLOW_COPY_AND_ASSIGN(T) \
|
|
T(const T&); \
|
|
void operator=(const T&)
|
|
#endif
|
|
|
|
static const malloc_table_t* gMallocTable = nullptr;
|
|
|
|
// This enables/disables DMD.
|
|
static bool gIsDMDRunning = false;
|
|
|
|
// This provides infallible allocations (they abort on OOM). We use it for all
|
|
// of DMD's own allocations, which fall into the following three cases.
|
|
// - Direct allocations (the easy case).
|
|
// - Indirect allocations in js::{Vector,HashSet,HashMap} -- this class serves
|
|
// as their AllocPolicy.
|
|
// - Other indirect allocations (e.g. NS_StackWalk) -- see the comments on
|
|
// Thread::mBlockIntercepts and in replace_malloc for how these work.
|
|
//
|
|
class InfallibleAllocPolicy
|
|
{
|
|
static void ExitOnFailure(const void* aP);
|
|
|
|
public:
|
|
static void* malloc_(size_t aSize)
|
|
{
|
|
void* p = gMallocTable->malloc(aSize);
|
|
ExitOnFailure(p);
|
|
return p;
|
|
}
|
|
|
|
template <typename T>
|
|
static T* pod_malloc(size_t aNumElems)
|
|
{
|
|
if (aNumElems & mozilla::tl::MulOverflowMask<sizeof(T)>::value)
|
|
return nullptr;
|
|
void* p = gMallocTable->malloc(aNumElems * sizeof(T));
|
|
ExitOnFailure(p);
|
|
return (T*)p;
|
|
}
|
|
|
|
static void* calloc_(size_t aSize)
|
|
{
|
|
void* p = gMallocTable->calloc(1, aSize);
|
|
ExitOnFailure(p);
|
|
return p;
|
|
}
|
|
|
|
template <typename T>
|
|
static T* pod_calloc(size_t aNumElems)
|
|
{
|
|
void* p = gMallocTable->calloc(aNumElems, sizeof(T));
|
|
ExitOnFailure(p);
|
|
return (T*)p;
|
|
}
|
|
|
|
// This realloc_ is the one we use for direct reallocs within DMD.
|
|
static void* realloc_(void* aPtr, size_t aNewSize)
|
|
{
|
|
void* p = gMallocTable->realloc(aPtr, aNewSize);
|
|
ExitOnFailure(p);
|
|
return p;
|
|
}
|
|
|
|
// This realloc_ is required for this to be a JS container AllocPolicy.
|
|
template <typename T>
|
|
static T* pod_realloc(T* aPtr, size_t aOldSize, size_t aNewSize)
|
|
{
|
|
if (aNewSize & mozilla::tl::MulOverflowMask<sizeof(T)>::value)
|
|
return nullptr;
|
|
return (T*)InfallibleAllocPolicy::realloc_((void *)aPtr, aNewSize * sizeof(T));
|
|
}
|
|
|
|
static void* memalign_(size_t aAlignment, size_t aSize)
|
|
{
|
|
void* p = gMallocTable->memalign(aAlignment, aSize);
|
|
ExitOnFailure(p);
|
|
return p;
|
|
}
|
|
|
|
static void free_(void* aPtr) { gMallocTable->free(aPtr); }
|
|
|
|
static char* strdup_(const char* aStr)
|
|
{
|
|
char* s = (char*) InfallibleAllocPolicy::malloc_(strlen(aStr) + 1);
|
|
strcpy(s, aStr);
|
|
return s;
|
|
}
|
|
|
|
template <class T>
|
|
static T* new_()
|
|
{
|
|
void* mem = malloc_(sizeof(T));
|
|
ExitOnFailure(mem);
|
|
return new (mem) T;
|
|
}
|
|
|
|
template <class T, typename P1>
|
|
static T* new_(P1 p1)
|
|
{
|
|
void* mem = malloc_(sizeof(T));
|
|
ExitOnFailure(mem);
|
|
return new (mem) T(p1);
|
|
}
|
|
|
|
template <class T>
|
|
static void delete_(T *p)
|
|
{
|
|
if (p) {
|
|
p->~T();
|
|
InfallibleAllocPolicy::free_(p);
|
|
}
|
|
}
|
|
|
|
static void reportAllocOverflow() { ExitOnFailure(nullptr); }
|
|
};
|
|
|
|
// This is only needed because of the |const void*| vs |void*| arg mismatch.
|
|
static size_t
|
|
MallocSizeOf(const void* aPtr)
|
|
{
|
|
return gMallocTable->malloc_usable_size(const_cast<void*>(aPtr));
|
|
}
|
|
|
|
MOZ_EXPORT void
|
|
StatusMsg(const char* aFmt, ...)
|
|
{
|
|
va_list ap;
|
|
va_start(ap, aFmt);
|
|
#ifdef ANDROID
|
|
#ifdef MOZ_B2G_LOADER
|
|
// Don't call __android_log_vprint() during initialization, or the magic file
|
|
// descriptors will be occupied by android logcat.
|
|
if (gIsDMDRunning)
|
|
#endif
|
|
__android_log_vprint(ANDROID_LOG_INFO, "DMD", aFmt, ap);
|
|
#else
|
|
// The +64 is easily enough for the "DMD[<pid>] " prefix and the NUL.
|
|
char* fmt = (char*) InfallibleAllocPolicy::malloc_(strlen(aFmt) + 64);
|
|
sprintf(fmt, "DMD[%d] %s", getpid(), aFmt);
|
|
vfprintf(stderr, fmt, ap);
|
|
InfallibleAllocPolicy::free_(fmt);
|
|
#endif
|
|
va_end(ap);
|
|
}
|
|
|
|
/* static */ void
|
|
InfallibleAllocPolicy::ExitOnFailure(const void* aP)
|
|
{
|
|
if (!aP) {
|
|
MOZ_CRASH("DMD out of memory; aborting");
|
|
}
|
|
}
|
|
|
|
class FpWriteFunc : public JSONWriteFunc
|
|
{
|
|
public:
|
|
explicit FpWriteFunc(FILE* aFp) : mFp(aFp) {}
|
|
~FpWriteFunc() { fclose(mFp); }
|
|
|
|
void Write(const char* aStr) { fputs(aStr, mFp); }
|
|
|
|
private:
|
|
FILE* mFp;
|
|
};
|
|
|
|
static double
|
|
Percent(size_t part, size_t whole)
|
|
{
|
|
return (whole == 0) ? 0 : 100 * (double)part / whole;
|
|
}
|
|
|
|
// Commifies the number.
|
|
static char*
|
|
Show(size_t n, char* buf, size_t buflen)
|
|
{
|
|
int nc = 0, i = 0, lasti = buflen - 2;
|
|
buf[lasti + 1] = '\0';
|
|
if (n == 0) {
|
|
buf[lasti - i] = '0';
|
|
i++;
|
|
} else {
|
|
while (n > 0) {
|
|
if (((i - nc) % 3) == 0 && i != 0) {
|
|
buf[lasti - i] = ',';
|
|
i++;
|
|
nc++;
|
|
}
|
|
buf[lasti - i] = static_cast<char>((n % 10) + '0');
|
|
i++;
|
|
n /= 10;
|
|
}
|
|
}
|
|
int firstCharIndex = lasti - i + 1;
|
|
|
|
MOZ_ASSERT(firstCharIndex >= 0);
|
|
return &buf[firstCharIndex];
|
|
}
|
|
|
|
//---------------------------------------------------------------------------
|
|
// Options (Part 1)
|
|
//---------------------------------------------------------------------------
|
|
|
|
class Options
|
|
{
|
|
template <typename T>
|
|
struct NumOption
|
|
{
|
|
const T mDefault;
|
|
const T mMax;
|
|
T mActual;
|
|
NumOption(T aDefault, T aMax)
|
|
: mDefault(aDefault), mMax(aMax), mActual(aDefault)
|
|
{}
|
|
};
|
|
|
|
enum Mode {
|
|
Normal, // run normally
|
|
Test // do some basic correctness tests
|
|
};
|
|
|
|
char* mDMDEnvVar; // a saved copy, for later printing
|
|
|
|
NumOption<size_t> mSampleBelowSize;
|
|
NumOption<uint32_t> mMaxFrames;
|
|
bool mShowDumpStats;
|
|
Mode mMode;
|
|
|
|
void BadArg(const char* aArg);
|
|
static const char* ValueIfMatch(const char* aArg, const char* aOptionName);
|
|
static bool GetLong(const char* aArg, const char* aOptionName,
|
|
long aMin, long aMax, long* aValue);
|
|
static bool GetBool(const char* aArg, const char* aOptionName, bool* aValue);
|
|
|
|
public:
|
|
explicit Options(const char* aDMDEnvVar);
|
|
|
|
const char* DMDEnvVar() const { return mDMDEnvVar; }
|
|
|
|
size_t SampleBelowSize() const { return mSampleBelowSize.mActual; }
|
|
size_t MaxFrames() const { return mMaxFrames.mActual; }
|
|
size_t ShowDumpStats() const { return mShowDumpStats; }
|
|
|
|
void SetSampleBelowSize(size_t aN) { mSampleBelowSize.mActual = aN; }
|
|
|
|
bool IsTestMode() const { return mMode == Test; }
|
|
};
|
|
|
|
static Options *gOptions;
|
|
|
|
//---------------------------------------------------------------------------
|
|
// The global lock
|
|
//---------------------------------------------------------------------------
|
|
|
|
// MutexBase implements the platform-specific parts of a mutex.
|
|
|
|
#ifdef XP_WIN
|
|
|
|
class MutexBase
|
|
{
|
|
CRITICAL_SECTION mCS;
|
|
|
|
DISALLOW_COPY_AND_ASSIGN(MutexBase);
|
|
|
|
public:
|
|
MutexBase()
|
|
{
|
|
InitializeCriticalSection(&mCS);
|
|
}
|
|
|
|
~MutexBase()
|
|
{
|
|
DeleteCriticalSection(&mCS);
|
|
}
|
|
|
|
void Lock()
|
|
{
|
|
EnterCriticalSection(&mCS);
|
|
}
|
|
|
|
void Unlock()
|
|
{
|
|
LeaveCriticalSection(&mCS);
|
|
}
|
|
};
|
|
|
|
#else
|
|
|
|
#include <pthread.h>
|
|
#include <sys/types.h>
|
|
|
|
class MutexBase
|
|
{
|
|
pthread_mutex_t mMutex;
|
|
|
|
DISALLOW_COPY_AND_ASSIGN(MutexBase);
|
|
|
|
public:
|
|
MutexBase()
|
|
{
|
|
pthread_mutex_init(&mMutex, nullptr);
|
|
}
|
|
|
|
void Lock()
|
|
{
|
|
pthread_mutex_lock(&mMutex);
|
|
}
|
|
|
|
void Unlock()
|
|
{
|
|
pthread_mutex_unlock(&mMutex);
|
|
}
|
|
};
|
|
|
|
#endif
|
|
|
|
class Mutex : private MutexBase
|
|
{
|
|
bool mIsLocked;
|
|
|
|
DISALLOW_COPY_AND_ASSIGN(Mutex);
|
|
|
|
public:
|
|
Mutex()
|
|
: mIsLocked(false)
|
|
{}
|
|
|
|
void Lock()
|
|
{
|
|
MutexBase::Lock();
|
|
MOZ_ASSERT(!mIsLocked);
|
|
mIsLocked = true;
|
|
}
|
|
|
|
void Unlock()
|
|
{
|
|
MOZ_ASSERT(mIsLocked);
|
|
mIsLocked = false;
|
|
MutexBase::Unlock();
|
|
}
|
|
|
|
bool IsLocked()
|
|
{
|
|
return mIsLocked;
|
|
}
|
|
};
|
|
|
|
// This lock must be held while manipulating global state, such as
|
|
// gStackTraceTable, gBlockTable, etc.
|
|
static Mutex* gStateLock = nullptr;
|
|
|
|
class AutoLockState
|
|
{
|
|
DISALLOW_COPY_AND_ASSIGN(AutoLockState);
|
|
|
|
public:
|
|
AutoLockState()
|
|
{
|
|
gStateLock->Lock();
|
|
}
|
|
~AutoLockState()
|
|
{
|
|
gStateLock->Unlock();
|
|
}
|
|
};
|
|
|
|
class AutoUnlockState
|
|
{
|
|
DISALLOW_COPY_AND_ASSIGN(AutoUnlockState);
|
|
|
|
public:
|
|
AutoUnlockState()
|
|
{
|
|
gStateLock->Unlock();
|
|
}
|
|
~AutoUnlockState()
|
|
{
|
|
gStateLock->Lock();
|
|
}
|
|
};
|
|
|
|
//---------------------------------------------------------------------------
|
|
// Thread-local storage and blocking of intercepts
|
|
//---------------------------------------------------------------------------
|
|
|
|
#ifdef XP_WIN
|
|
|
|
#define DMD_TLS_INDEX_TYPE DWORD
|
|
#define DMD_CREATE_TLS_INDEX(i_) do { \
|
|
(i_) = TlsAlloc(); \
|
|
} while (0)
|
|
#define DMD_DESTROY_TLS_INDEX(i_) TlsFree((i_))
|
|
#define DMD_GET_TLS_DATA(i_) TlsGetValue((i_))
|
|
#define DMD_SET_TLS_DATA(i_, v_) TlsSetValue((i_), (v_))
|
|
|
|
#else
|
|
|
|
#include <pthread.h>
|
|
|
|
#define DMD_TLS_INDEX_TYPE pthread_key_t
|
|
#define DMD_CREATE_TLS_INDEX(i_) pthread_key_create(&(i_), nullptr)
|
|
#define DMD_DESTROY_TLS_INDEX(i_) pthread_key_delete((i_))
|
|
#define DMD_GET_TLS_DATA(i_) pthread_getspecific((i_))
|
|
#define DMD_SET_TLS_DATA(i_, v_) pthread_setspecific((i_), (v_))
|
|
|
|
#endif
|
|
|
|
static DMD_TLS_INDEX_TYPE gTlsIndex;
|
|
|
|
class Thread
|
|
{
|
|
// Required for allocation via InfallibleAllocPolicy::new_.
|
|
friend class InfallibleAllocPolicy;
|
|
|
|
// When true, this blocks intercepts, which allows malloc interception
|
|
// functions to themselves call malloc. (Nb: for direct calls to malloc we
|
|
// can just use InfallibleAllocPolicy::{malloc_,new_}, but we sometimes
|
|
// indirectly call vanilla malloc via functions like NS_StackWalk.)
|
|
bool mBlockIntercepts;
|
|
|
|
Thread()
|
|
: mBlockIntercepts(false)
|
|
{}
|
|
|
|
DISALLOW_COPY_AND_ASSIGN(Thread);
|
|
|
|
public:
|
|
static Thread* Fetch();
|
|
|
|
bool BlockIntercepts()
|
|
{
|
|
MOZ_ASSERT(!mBlockIntercepts);
|
|
return mBlockIntercepts = true;
|
|
}
|
|
|
|
bool UnblockIntercepts()
|
|
{
|
|
MOZ_ASSERT(mBlockIntercepts);
|
|
return mBlockIntercepts = false;
|
|
}
|
|
|
|
bool InterceptsAreBlocked() const
|
|
{
|
|
return mBlockIntercepts;
|
|
}
|
|
};
|
|
|
|
/* static */ Thread*
|
|
Thread::Fetch()
|
|
{
|
|
Thread* t = static_cast<Thread*>(DMD_GET_TLS_DATA(gTlsIndex));
|
|
|
|
if (MOZ_UNLIKELY(!t)) {
|
|
// This memory is never freed, even if the thread dies. It's a leak, but
|
|
// only a tiny one.
|
|
t = InfallibleAllocPolicy::new_<Thread>();
|
|
DMD_SET_TLS_DATA(gTlsIndex, t);
|
|
}
|
|
|
|
return t;
|
|
}
|
|
|
|
// An object of this class must be created (on the stack) before running any
|
|
// code that might allocate.
|
|
class AutoBlockIntercepts
|
|
{
|
|
Thread* const mT;
|
|
|
|
DISALLOW_COPY_AND_ASSIGN(AutoBlockIntercepts);
|
|
|
|
public:
|
|
explicit AutoBlockIntercepts(Thread* aT)
|
|
: mT(aT)
|
|
{
|
|
mT->BlockIntercepts();
|
|
}
|
|
~AutoBlockIntercepts()
|
|
{
|
|
MOZ_ASSERT(mT->InterceptsAreBlocked());
|
|
mT->UnblockIntercepts();
|
|
}
|
|
};
|
|
|
|
//---------------------------------------------------------------------------
|
|
// Location service
|
|
//---------------------------------------------------------------------------
|
|
|
|
class StringTable
|
|
{
|
|
public:
|
|
StringTable()
|
|
{
|
|
(void)mSet.init(64);
|
|
}
|
|
|
|
const char*
|
|
Intern(const char* aString)
|
|
{
|
|
StringHashSet::AddPtr p = mSet.lookupForAdd(aString);
|
|
if (p) {
|
|
return *p;
|
|
}
|
|
|
|
const char* newString = InfallibleAllocPolicy::strdup_(aString);
|
|
(void)mSet.add(p, newString);
|
|
return newString;
|
|
}
|
|
|
|
size_t
|
|
SizeOfExcludingThis(mozilla::MallocSizeOf aMallocSizeOf) const
|
|
{
|
|
size_t n = 0;
|
|
n += mSet.sizeOfExcludingThis(aMallocSizeOf);
|
|
for (StringHashSet::Range r = mSet.all();
|
|
!r.empty();
|
|
r.popFront()) {
|
|
n += aMallocSizeOf(r.front());
|
|
}
|
|
return n;
|
|
}
|
|
|
|
private:
|
|
struct StringHasher
|
|
{
|
|
typedef const char* Lookup;
|
|
|
|
static uint32_t hash(const char* const& aS)
|
|
{
|
|
return HashString(aS);
|
|
}
|
|
|
|
static bool match(const char* const& aA, const char* const& aB)
|
|
{
|
|
return strcmp(aA, aB) == 0;
|
|
}
|
|
};
|
|
|
|
typedef js::HashSet<const char*, StringHasher, InfallibleAllocPolicy> StringHashSet;
|
|
|
|
StringHashSet mSet;
|
|
};
|
|
|
|
class StringAlloc
|
|
{
|
|
public:
|
|
static char*
|
|
copy(const char* aString)
|
|
{
|
|
return InfallibleAllocPolicy::strdup_(aString);
|
|
}
|
|
static void
|
|
free(char* aString)
|
|
{
|
|
InfallibleAllocPolicy::free_(aString);
|
|
}
|
|
};
|
|
|
|
struct DescribeCodeAddressLock
|
|
{
|
|
static void Unlock() { gStateLock->Unlock(); }
|
|
static void Lock() { gStateLock->Lock(); }
|
|
static bool IsLocked() { return gStateLock->IsLocked(); }
|
|
};
|
|
|
|
typedef CodeAddressService<StringTable, StringAlloc, DescribeCodeAddressLock>
|
|
CodeAddressService;
|
|
|
|
//---------------------------------------------------------------------------
|
|
// Stack traces
|
|
//---------------------------------------------------------------------------
|
|
|
|
class StackTrace
|
|
{
|
|
public:
|
|
static const uint32_t MaxFrames = 24;
|
|
|
|
private:
|
|
uint32_t mLength; // The number of PCs.
|
|
const void* mPcs[MaxFrames]; // The PCs themselves. If --max-frames is less
|
|
// than 24, this array is bigger than
|
|
// necessary, but that case is unusual.
|
|
|
|
public:
|
|
StackTrace() : mLength(0) {}
|
|
|
|
uint32_t Length() const { return mLength; }
|
|
const void* Pc(uint32_t i) const
|
|
{
|
|
MOZ_ASSERT(i < mLength);
|
|
return mPcs[i];
|
|
}
|
|
|
|
uint32_t Size() const { return mLength * sizeof(mPcs[0]); }
|
|
|
|
// The stack trace returned by this function is interned in gStackTraceTable,
|
|
// and so is immortal and unmovable.
|
|
static const StackTrace* Get(Thread* aT);
|
|
|
|
// Hash policy.
|
|
|
|
typedef StackTrace* Lookup;
|
|
|
|
static uint32_t hash(const StackTrace* const& aSt)
|
|
{
|
|
return mozilla::HashBytes(aSt->mPcs, aSt->Size());
|
|
}
|
|
|
|
static bool match(const StackTrace* const& aA,
|
|
const StackTrace* const& aB)
|
|
{
|
|
return aA->mLength == aB->mLength &&
|
|
memcmp(aA->mPcs, aB->mPcs, aA->Size()) == 0;
|
|
}
|
|
|
|
private:
|
|
static void StackWalkCallback(uint32_t aFrameNumber, void* aPc, void* aSp,
|
|
void* aClosure)
|
|
{
|
|
StackTrace* st = (StackTrace*) aClosure;
|
|
MOZ_ASSERT(st->mLength < MaxFrames);
|
|
st->mPcs[st->mLength] = aPc;
|
|
st->mLength++;
|
|
MOZ_ASSERT(st->mLength == aFrameNumber);
|
|
}
|
|
};
|
|
|
|
typedef js::HashSet<StackTrace*, StackTrace, InfallibleAllocPolicy>
|
|
StackTraceTable;
|
|
static StackTraceTable* gStackTraceTable = nullptr;
|
|
|
|
typedef js::HashSet<const StackTrace*, js::DefaultHasher<const StackTrace*>,
|
|
InfallibleAllocPolicy>
|
|
StackTraceSet;
|
|
|
|
typedef js::HashSet<const void*, js::DefaultHasher<const void*>,
|
|
InfallibleAllocPolicy>
|
|
PointerSet;
|
|
typedef js::HashMap<const void*, uint32_t, js::DefaultHasher<const void*>,
|
|
InfallibleAllocPolicy>
|
|
PointerIdMap;
|
|
|
|
// We won't GC the stack trace table until it this many elements.
|
|
static uint32_t gGCStackTraceTableWhenSizeExceeds = 4 * 1024;
|
|
|
|
/* static */ const StackTrace*
|
|
StackTrace::Get(Thread* aT)
|
|
{
|
|
MOZ_ASSERT(gStateLock->IsLocked());
|
|
MOZ_ASSERT(aT->InterceptsAreBlocked());
|
|
|
|
// On Windows, NS_StackWalk can acquire a lock from the shared library
|
|
// loader. Another thread might call malloc while holding that lock (when
|
|
// loading a shared library). So we can't be in gStateLock during the call
|
|
// to NS_StackWalk. For details, see
|
|
// https://bugzilla.mozilla.org/show_bug.cgi?id=374829#c8
|
|
// On Linux, something similar can happen; see bug 824340.
|
|
// So let's just release it on all platforms.
|
|
nsresult rv;
|
|
StackTrace tmp;
|
|
{
|
|
AutoUnlockState unlock;
|
|
uint32_t skipFrames = 2;
|
|
rv = NS_StackWalk(StackWalkCallback, skipFrames,
|
|
gOptions->MaxFrames(), &tmp, 0, nullptr);
|
|
}
|
|
|
|
if (rv == NS_OK) {
|
|
// Handle the common case first. All is ok. Nothing to do.
|
|
} else if (rv == NS_ERROR_NOT_IMPLEMENTED || rv == NS_ERROR_FAILURE) {
|
|
tmp.mLength = 0;
|
|
} else if (rv == NS_ERROR_UNEXPECTED) {
|
|
// XXX: This |rv| only happens on Mac, and it indicates that we're handling
|
|
// a call to malloc that happened inside a mutex-handling function. Any
|
|
// attempt to create a semaphore (which can happen in printf) could
|
|
// deadlock.
|
|
//
|
|
// However, the most complex thing DMD does after Get() returns is to put
|
|
// something in a hash table, which might call
|
|
// InfallibleAllocPolicy::malloc_. I'm not yet sure if this needs special
|
|
// handling, hence the forced abort. Sorry. If you hit this, please file
|
|
// a bug and CC nnethercote.
|
|
MOZ_CRASH("unexpected case in StackTrace::Get()");
|
|
} else {
|
|
MOZ_CRASH("impossible case in StackTrace::Get()");
|
|
}
|
|
|
|
StackTraceTable::AddPtr p = gStackTraceTable->lookupForAdd(&tmp);
|
|
if (!p) {
|
|
StackTrace* stnew = InfallibleAllocPolicy::new_<StackTrace>(tmp);
|
|
(void)gStackTraceTable->add(p, stnew);
|
|
}
|
|
return *p;
|
|
}
|
|
|
|
//---------------------------------------------------------------------------
|
|
// Heap blocks
|
|
//---------------------------------------------------------------------------
|
|
|
|
// This class combines a 2-byte-aligned pointer (i.e. one whose bottom bit
|
|
// is zero) with a 1-bit tag.
|
|
//
|
|
// |T| is the pointer type, e.g. |int*|, not the pointed-to type. This makes
|
|
// is easier to have const pointers, e.g. |TaggedPtr<const int*>|.
|
|
template <typename T>
|
|
class TaggedPtr
|
|
{
|
|
union
|
|
{
|
|
T mPtr;
|
|
uintptr_t mUint;
|
|
};
|
|
|
|
static const uintptr_t kTagMask = uintptr_t(0x1);
|
|
static const uintptr_t kPtrMask = ~kTagMask;
|
|
|
|
static bool IsTwoByteAligned(T aPtr)
|
|
{
|
|
return (uintptr_t(aPtr) & kTagMask) == 0;
|
|
}
|
|
|
|
public:
|
|
TaggedPtr()
|
|
: mPtr(nullptr)
|
|
{}
|
|
|
|
TaggedPtr(T aPtr, bool aBool)
|
|
: mPtr(aPtr)
|
|
{
|
|
MOZ_ASSERT(IsTwoByteAligned(aPtr));
|
|
uintptr_t tag = uintptr_t(aBool);
|
|
MOZ_ASSERT(tag <= kTagMask);
|
|
mUint |= (tag & kTagMask);
|
|
}
|
|
|
|
void Set(T aPtr, bool aBool)
|
|
{
|
|
MOZ_ASSERT(IsTwoByteAligned(aPtr));
|
|
mPtr = aPtr;
|
|
uintptr_t tag = uintptr_t(aBool);
|
|
MOZ_ASSERT(tag <= kTagMask);
|
|
mUint |= (tag & kTagMask);
|
|
}
|
|
|
|
T Ptr() const { return reinterpret_cast<T>(mUint & kPtrMask); }
|
|
|
|
bool Tag() const { return bool(mUint & kTagMask); }
|
|
};
|
|
|
|
// A live heap block.
|
|
class Block
|
|
{
|
|
const void* mPtr;
|
|
const size_t mReqSize; // size requested
|
|
|
|
// Ptr: |mAllocStackTrace| - stack trace where this block was allocated.
|
|
// Tag bit 0: |mSampled| - was this block sampled? (if so, slop == 0).
|
|
TaggedPtr<const StackTrace* const>
|
|
mAllocStackTrace_mSampled;
|
|
|
|
// This array has two elements because we record at most two reports of a
|
|
// block.
|
|
// - Ptr: |mReportStackTrace| - stack trace where this block was reported.
|
|
// nullptr if not reported.
|
|
// - Tag bit 0: |mReportedOnAlloc| - was the block reported immediately on
|
|
// allocation? If so, DMD must not clear the report at the end of
|
|
// AnalyzeReports(). Only relevant if |mReportStackTrace| is non-nullptr.
|
|
//
|
|
// |mPtr| is used as the key in BlockTable, so it's ok for this member
|
|
// to be |mutable|.
|
|
mutable TaggedPtr<const StackTrace*> mReportStackTrace_mReportedOnAlloc[2];
|
|
|
|
public:
|
|
Block(const void* aPtr, size_t aReqSize, const StackTrace* aAllocStackTrace,
|
|
bool aSampled)
|
|
: mPtr(aPtr),
|
|
mReqSize(aReqSize),
|
|
mAllocStackTrace_mSampled(aAllocStackTrace, aSampled),
|
|
mReportStackTrace_mReportedOnAlloc() // all fields get zeroed
|
|
{
|
|
MOZ_ASSERT(aAllocStackTrace);
|
|
}
|
|
|
|
const void* Address() const { return mPtr; }
|
|
|
|
size_t ReqSize() const { return mReqSize; }
|
|
|
|
// Sampled blocks always have zero slop.
|
|
size_t SlopSize() const
|
|
{
|
|
return IsSampled() ? 0 : MallocSizeOf(mPtr) - mReqSize;
|
|
}
|
|
|
|
size_t UsableSize() const
|
|
{
|
|
return IsSampled() ? mReqSize : MallocSizeOf(mPtr);
|
|
}
|
|
|
|
bool IsSampled() const
|
|
{
|
|
return mAllocStackTrace_mSampled.Tag();
|
|
}
|
|
|
|
const StackTrace* AllocStackTrace() const
|
|
{
|
|
return mAllocStackTrace_mSampled.Ptr();
|
|
}
|
|
|
|
const StackTrace* ReportStackTrace1() const
|
|
{
|
|
return mReportStackTrace_mReportedOnAlloc[0].Ptr();
|
|
}
|
|
|
|
const StackTrace* ReportStackTrace2() const
|
|
{
|
|
return mReportStackTrace_mReportedOnAlloc[1].Ptr();
|
|
}
|
|
|
|
bool ReportedOnAlloc1() const
|
|
{
|
|
return mReportStackTrace_mReportedOnAlloc[0].Tag();
|
|
}
|
|
|
|
bool ReportedOnAlloc2() const
|
|
{
|
|
return mReportStackTrace_mReportedOnAlloc[1].Tag();
|
|
}
|
|
|
|
void AddStackTracesToTable(StackTraceSet& aStackTraces) const
|
|
{
|
|
aStackTraces.put(AllocStackTrace()); // never null
|
|
const StackTrace* st;
|
|
if ((st = ReportStackTrace1())) { // may be null
|
|
aStackTraces.put(st);
|
|
}
|
|
if ((st = ReportStackTrace2())) { // may be null
|
|
aStackTraces.put(st);
|
|
}
|
|
}
|
|
|
|
uint32_t NumReports() const
|
|
{
|
|
if (ReportStackTrace2()) {
|
|
MOZ_ASSERT(ReportStackTrace1());
|
|
return 2;
|
|
}
|
|
if (ReportStackTrace1()) {
|
|
return 1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
// This is |const| thanks to the |mutable| fields above.
|
|
void Report(Thread* aT, bool aReportedOnAlloc) const
|
|
{
|
|
// We don't bother recording reports after the 2nd one.
|
|
uint32_t numReports = NumReports();
|
|
if (numReports < 2) {
|
|
mReportStackTrace_mReportedOnAlloc[numReports].Set(StackTrace::Get(aT),
|
|
aReportedOnAlloc);
|
|
}
|
|
}
|
|
|
|
void UnreportIfNotReportedOnAlloc() const
|
|
{
|
|
if (!ReportedOnAlloc1() && !ReportedOnAlloc2()) {
|
|
mReportStackTrace_mReportedOnAlloc[0].Set(nullptr, 0);
|
|
mReportStackTrace_mReportedOnAlloc[1].Set(nullptr, 0);
|
|
|
|
} else if (!ReportedOnAlloc1() && ReportedOnAlloc2()) {
|
|
// Shift the 2nd report down to the 1st one.
|
|
mReportStackTrace_mReportedOnAlloc[0] =
|
|
mReportStackTrace_mReportedOnAlloc[1];
|
|
mReportStackTrace_mReportedOnAlloc[1].Set(nullptr, 0);
|
|
|
|
} else if (ReportedOnAlloc1() && !ReportedOnAlloc2()) {
|
|
mReportStackTrace_mReportedOnAlloc[1].Set(nullptr, 0);
|
|
}
|
|
}
|
|
|
|
// Hash policy.
|
|
|
|
typedef const void* Lookup;
|
|
|
|
static uint32_t hash(const void* const& aPtr)
|
|
{
|
|
return mozilla::HashGeneric(aPtr);
|
|
}
|
|
|
|
static bool match(const Block& aB, const void* const& aPtr)
|
|
{
|
|
return aB.mPtr == aPtr;
|
|
}
|
|
};
|
|
|
|
typedef js::HashSet<Block, Block, InfallibleAllocPolicy> BlockTable;
|
|
static BlockTable* gBlockTable = nullptr;
|
|
|
|
// Add a pointer to each live stack trace into the given StackTraceSet. (A
|
|
// stack trace is live if it's used by one of the live blocks.)
|
|
static void
|
|
GatherUsedStackTraces(StackTraceSet& aStackTraces)
|
|
{
|
|
MOZ_ASSERT(gStateLock->IsLocked());
|
|
MOZ_ASSERT(Thread::Fetch()->InterceptsAreBlocked());
|
|
|
|
aStackTraces.finish();
|
|
aStackTraces.init(512);
|
|
|
|
for (BlockTable::Range r = gBlockTable->all(); !r.empty(); r.popFront()) {
|
|
const Block& b = r.front();
|
|
b.AddStackTracesToTable(aStackTraces);
|
|
}
|
|
}
|
|
|
|
// Delete stack traces that we aren't using, and compact our hashtable.
|
|
static void
|
|
GCStackTraces()
|
|
{
|
|
MOZ_ASSERT(gStateLock->IsLocked());
|
|
MOZ_ASSERT(Thread::Fetch()->InterceptsAreBlocked());
|
|
|
|
StackTraceSet usedStackTraces;
|
|
GatherUsedStackTraces(usedStackTraces);
|
|
|
|
// Delete all unused stack traces from gStackTraceTable. The Enum destructor
|
|
// will automatically rehash and compact the table.
|
|
for (StackTraceTable::Enum e(*gStackTraceTable); !e.empty(); e.popFront()) {
|
|
StackTrace* const& st = e.front();
|
|
if (!usedStackTraces.has(st)) {
|
|
e.removeFront();
|
|
InfallibleAllocPolicy::delete_(st);
|
|
}
|
|
}
|
|
|
|
// Schedule a GC when we have twice as many stack traces as we had right after
|
|
// this GC finished.
|
|
gGCStackTraceTableWhenSizeExceeds = 2 * gStackTraceTable->count();
|
|
}
|
|
|
|
//---------------------------------------------------------------------------
|
|
// malloc/free callbacks
|
|
//---------------------------------------------------------------------------
|
|
|
|
static size_t gSmallBlockActualSizeCounter = 0;
|
|
|
|
static void
|
|
AllocCallback(void* aPtr, size_t aReqSize, Thread* aT)
|
|
{
|
|
MOZ_ASSERT(gIsDMDRunning);
|
|
|
|
if (!aPtr) {
|
|
return;
|
|
}
|
|
|
|
AutoLockState lock;
|
|
AutoBlockIntercepts block(aT);
|
|
|
|
size_t actualSize = gMallocTable->malloc_usable_size(aPtr);
|
|
size_t sampleBelowSize = gOptions->SampleBelowSize();
|
|
|
|
if (actualSize < sampleBelowSize) {
|
|
// If this allocation is smaller than the sample-below size, increment the
|
|
// cumulative counter. Then, if that counter now exceeds the sample size,
|
|
// blame this allocation for |sampleBelowSize| bytes. This precludes the
|
|
// measurement of slop.
|
|
gSmallBlockActualSizeCounter += actualSize;
|
|
if (gSmallBlockActualSizeCounter >= sampleBelowSize) {
|
|
gSmallBlockActualSizeCounter -= sampleBelowSize;
|
|
|
|
Block b(aPtr, sampleBelowSize, StackTrace::Get(aT), /* sampled */ true);
|
|
(void)gBlockTable->putNew(aPtr, b);
|
|
}
|
|
} else {
|
|
// If this block size is larger than the sample size, record it exactly.
|
|
Block b(aPtr, aReqSize, StackTrace::Get(aT), /* sampled */ false);
|
|
(void)gBlockTable->putNew(aPtr, b);
|
|
}
|
|
}
|
|
|
|
static void
|
|
FreeCallback(void* aPtr, Thread* aT)
|
|
{
|
|
MOZ_ASSERT(gIsDMDRunning);
|
|
|
|
if (!aPtr) {
|
|
return;
|
|
}
|
|
|
|
AutoLockState lock;
|
|
AutoBlockIntercepts block(aT);
|
|
|
|
gBlockTable->remove(aPtr);
|
|
|
|
if (gStackTraceTable->count() > gGCStackTraceTableWhenSizeExceeds) {
|
|
GCStackTraces();
|
|
}
|
|
}
|
|
|
|
//---------------------------------------------------------------------------
|
|
// malloc/free interception
|
|
//---------------------------------------------------------------------------
|
|
|
|
static void Init(const malloc_table_t* aMallocTable);
|
|
|
|
} // namespace dmd
|
|
} // namespace mozilla
|
|
|
|
void
|
|
replace_init(const malloc_table_t* aMallocTable)
|
|
{
|
|
mozilla::dmd::Init(aMallocTable);
|
|
}
|
|
|
|
void*
|
|
replace_malloc(size_t aSize)
|
|
{
|
|
using namespace mozilla::dmd;
|
|
|
|
if (!gIsDMDRunning) {
|
|
// DMD hasn't started up, either because it wasn't enabled by the user, or
|
|
// we're still in Init() and something has indirectly called malloc. Do a
|
|
// vanilla malloc. (In the latter case, if it fails we'll crash. But
|
|
// OOM is highly unlikely so early on.)
|
|
return gMallocTable->malloc(aSize);
|
|
}
|
|
|
|
Thread* t = Thread::Fetch();
|
|
if (t->InterceptsAreBlocked()) {
|
|
// Intercepts are blocked, which means this must be a call to malloc
|
|
// triggered indirectly by DMD (e.g. via NS_StackWalk). Be infallible.
|
|
return InfallibleAllocPolicy::malloc_(aSize);
|
|
}
|
|
|
|
// This must be a call to malloc from outside DMD. Intercept it.
|
|
void* ptr = gMallocTable->malloc(aSize);
|
|
AllocCallback(ptr, aSize, t);
|
|
return ptr;
|
|
}
|
|
|
|
void*
|
|
replace_calloc(size_t aCount, size_t aSize)
|
|
{
|
|
using namespace mozilla::dmd;
|
|
|
|
if (!gIsDMDRunning) {
|
|
return gMallocTable->calloc(aCount, aSize);
|
|
}
|
|
|
|
Thread* t = Thread::Fetch();
|
|
if (t->InterceptsAreBlocked()) {
|
|
return InfallibleAllocPolicy::calloc_(aCount * aSize);
|
|
}
|
|
|
|
void* ptr = gMallocTable->calloc(aCount, aSize);
|
|
AllocCallback(ptr, aCount * aSize, t);
|
|
return ptr;
|
|
}
|
|
|
|
void*
|
|
replace_realloc(void* aOldPtr, size_t aSize)
|
|
{
|
|
using namespace mozilla::dmd;
|
|
|
|
if (!gIsDMDRunning) {
|
|
return gMallocTable->realloc(aOldPtr, aSize);
|
|
}
|
|
|
|
Thread* t = Thread::Fetch();
|
|
if (t->InterceptsAreBlocked()) {
|
|
return InfallibleAllocPolicy::realloc_(aOldPtr, aSize);
|
|
}
|
|
|
|
// If |aOldPtr| is nullptr, the call is equivalent to |malloc(aSize)|.
|
|
if (!aOldPtr) {
|
|
return replace_malloc(aSize);
|
|
}
|
|
|
|
// Be very careful here! Must remove the block from the table before doing
|
|
// the realloc to avoid races, just like in replace_free().
|
|
// Nb: This does an unnecessary hashtable remove+add if the block doesn't
|
|
// move, but doing better isn't worth the effort.
|
|
FreeCallback(aOldPtr, t);
|
|
void* ptr = gMallocTable->realloc(aOldPtr, aSize);
|
|
if (ptr) {
|
|
AllocCallback(ptr, aSize, t);
|
|
} else {
|
|
// If realloc fails, we re-insert the old pointer. It will look like it
|
|
// was allocated for the first time here, which is untrue, and the slop
|
|
// bytes will be zero, which may be untrue. But this case is rare and
|
|
// doing better isn't worth the effort.
|
|
AllocCallback(aOldPtr, gMallocTable->malloc_usable_size(aOldPtr), t);
|
|
}
|
|
return ptr;
|
|
}
|
|
|
|
void*
|
|
replace_memalign(size_t aAlignment, size_t aSize)
|
|
{
|
|
using namespace mozilla::dmd;
|
|
|
|
if (!gIsDMDRunning) {
|
|
return gMallocTable->memalign(aAlignment, aSize);
|
|
}
|
|
|
|
Thread* t = Thread::Fetch();
|
|
if (t->InterceptsAreBlocked()) {
|
|
return InfallibleAllocPolicy::memalign_(aAlignment, aSize);
|
|
}
|
|
|
|
void* ptr = gMallocTable->memalign(aAlignment, aSize);
|
|
AllocCallback(ptr, aSize, t);
|
|
return ptr;
|
|
}
|
|
|
|
void
|
|
replace_free(void* aPtr)
|
|
{
|
|
using namespace mozilla::dmd;
|
|
|
|
if (!gIsDMDRunning) {
|
|
gMallocTable->free(aPtr);
|
|
return;
|
|
}
|
|
|
|
Thread* t = Thread::Fetch();
|
|
if (t->InterceptsAreBlocked()) {
|
|
return InfallibleAllocPolicy::free_(aPtr);
|
|
}
|
|
|
|
// Do the actual free after updating the table. Otherwise, another thread
|
|
// could call malloc and get the freed block and update the table, and then
|
|
// our update here would remove the newly-malloc'd block.
|
|
FreeCallback(aPtr, t);
|
|
gMallocTable->free(aPtr);
|
|
}
|
|
|
|
namespace mozilla {
|
|
namespace dmd {
|
|
|
|
//---------------------------------------------------------------------------
|
|
// Options (Part 2)
|
|
//---------------------------------------------------------------------------
|
|
|
|
// Given an |aOptionName| like "foo", succeed if |aArg| has the form "foo=blah"
|
|
// (where "blah" is non-empty) and return the pointer to "blah". |aArg| can
|
|
// have leading space chars (but not other whitespace).
|
|
const char*
|
|
Options::ValueIfMatch(const char* aArg, const char* aOptionName)
|
|
{
|
|
MOZ_ASSERT(!isspace(*aArg)); // any leading whitespace should not remain
|
|
size_t optionLen = strlen(aOptionName);
|
|
if (strncmp(aArg, aOptionName, optionLen) == 0 && aArg[optionLen] == '=' &&
|
|
aArg[optionLen + 1]) {
|
|
return aArg + optionLen + 1;
|
|
}
|
|
return nullptr;
|
|
}
|
|
|
|
// Extracts a |long| value for an option from an argument. It must be within
|
|
// the range |aMin..aMax| (inclusive).
|
|
bool
|
|
Options::GetLong(const char* aArg, const char* aOptionName,
|
|
long aMin, long aMax, long* aValue)
|
|
{
|
|
if (const char* optionValue = ValueIfMatch(aArg, aOptionName)) {
|
|
char* endPtr;
|
|
*aValue = strtol(optionValue, &endPtr, /* base */ 10);
|
|
if (!*endPtr && aMin <= *aValue && *aValue <= aMax &&
|
|
*aValue != LONG_MIN && *aValue != LONG_MAX) {
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
// Extracts a |bool| value for an option -- encoded as "yes" or "no" -- from an
|
|
// argument.
|
|
bool
|
|
Options::GetBool(const char* aArg, const char* aOptionName, bool* aValue)
|
|
{
|
|
if (const char* optionValue = ValueIfMatch(aArg, aOptionName)) {
|
|
if (strcmp(optionValue, "yes") == 0) {
|
|
*aValue = true;
|
|
return true;
|
|
}
|
|
if (strcmp(optionValue, "no") == 0) {
|
|
*aValue = false;
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
// The sample-below default is a prime number close to 4096.
|
|
// - Why that size? Because it's *much* faster but only moderately less precise
|
|
// than a size of 1.
|
|
// - Why prime? Because it makes our sampling more random. If we used a size
|
|
// of 4096, for example, then our alloc counter would only take on even
|
|
// values, because jemalloc always rounds up requests sizes. In contrast, a
|
|
// prime size will explore all possible values of the alloc counter.
|
|
//
|
|
Options::Options(const char* aDMDEnvVar)
|
|
: mDMDEnvVar(InfallibleAllocPolicy::strdup_(aDMDEnvVar)),
|
|
mSampleBelowSize(4093, 100 * 100 * 1000),
|
|
mMaxFrames(StackTrace::MaxFrames, StackTrace::MaxFrames),
|
|
mShowDumpStats(false),
|
|
mMode(Normal)
|
|
{
|
|
char* e = mDMDEnvVar;
|
|
if (strcmp(e, "1") != 0) {
|
|
bool isEnd = false;
|
|
while (!isEnd) {
|
|
// Consume leading whitespace.
|
|
while (isspace(*e)) {
|
|
e++;
|
|
}
|
|
|
|
// Save the start of the arg.
|
|
const char* arg = e;
|
|
|
|
// Find the first char after the arg, and temporarily change it to '\0'
|
|
// to isolate the arg.
|
|
while (!isspace(*e) && *e != '\0') {
|
|
e++;
|
|
}
|
|
char replacedChar = *e;
|
|
isEnd = replacedChar == '\0';
|
|
*e = '\0';
|
|
|
|
// Handle arg
|
|
long myLong;
|
|
bool myBool;
|
|
if (GetLong(arg, "--sample-below", 1, mSampleBelowSize.mMax, &myLong)) {
|
|
mSampleBelowSize.mActual = myLong;
|
|
|
|
} else if (GetLong(arg, "--max-frames", 1, mMaxFrames.mMax, &myLong)) {
|
|
mMaxFrames.mActual = myLong;
|
|
|
|
} else if (GetBool(arg, "--show-dump-stats", &myBool)) {
|
|
mShowDumpStats = myBool;
|
|
|
|
} else if (strcmp(arg, "--mode=normal") == 0) {
|
|
mMode = Options::Normal;
|
|
} else if (strcmp(arg, "--mode=test") == 0) {
|
|
mMode = Options::Test;
|
|
|
|
} else if (strcmp(arg, "") == 0) {
|
|
// This can only happen if there is trailing whitespace. Ignore.
|
|
MOZ_ASSERT(isEnd);
|
|
|
|
} else {
|
|
BadArg(arg);
|
|
}
|
|
|
|
// Undo the temporary isolation.
|
|
*e = replacedChar;
|
|
}
|
|
}
|
|
}
|
|
|
|
void
|
|
Options::BadArg(const char* aArg)
|
|
{
|
|
StatusMsg("\n");
|
|
StatusMsg("Bad entry in the $DMD environment variable: '%s'.\n", aArg);
|
|
StatusMsg("\n");
|
|
StatusMsg("Valid values of $DMD are:\n");
|
|
StatusMsg("- undefined or \"\" or \"0\", which disables DMD, or\n");
|
|
StatusMsg("- \"1\", which enables it with the default options, or\n");
|
|
StatusMsg("- a whitespace-separated list of |--option=val| entries, which\n");
|
|
StatusMsg(" enables it with non-default options.\n");
|
|
StatusMsg("\n");
|
|
StatusMsg("The following options are allowed; defaults are shown in [].\n");
|
|
StatusMsg(" --sample-below=<1..%d> Sample blocks smaller than this [%d]\n",
|
|
int(mSampleBelowSize.mMax),
|
|
int(mSampleBelowSize.mDefault));
|
|
StatusMsg(" (prime numbers are recommended)\n");
|
|
StatusMsg(" --max-frames=<1..%d> Max. depth of stack traces [%d]\n",
|
|
int(mMaxFrames.mMax),
|
|
int(mMaxFrames.mDefault));
|
|
StatusMsg(" --show-dump-stats=<yes|no> Show stats about dumps? [no]\n");
|
|
StatusMsg(" --mode=<normal|test> Mode of operation [normal]\n");
|
|
StatusMsg("\n");
|
|
exit(1);
|
|
}
|
|
|
|
//---------------------------------------------------------------------------
|
|
// DMD start-up
|
|
//---------------------------------------------------------------------------
|
|
|
|
#ifdef XP_MACOSX
|
|
static void
|
|
NopStackWalkCallback(uint32_t aFrameNumber, void* aPc, void* aSp,
|
|
void* aClosure)
|
|
{
|
|
}
|
|
#endif
|
|
|
|
// Note that fopen() can allocate.
|
|
static FILE*
|
|
OpenOutputFile(const char* aFilename)
|
|
{
|
|
FILE* fp = fopen(aFilename, "w");
|
|
if (!fp) {
|
|
StatusMsg("can't create %s file: %s\n", aFilename, strerror(errno));
|
|
exit(1);
|
|
}
|
|
return fp;
|
|
}
|
|
|
|
static void RunTestMode(UniquePtr<FpWriteFunc> aF1, UniquePtr<FpWriteFunc> aF2,
|
|
UniquePtr<FpWriteFunc> aF3, UniquePtr<FpWriteFunc> aF4);
|
|
|
|
// WARNING: this function runs *very* early -- before all static initializers
|
|
// have run. For this reason, non-scalar globals such as gStateLock and
|
|
// gStackTraceTable are allocated dynamically (so we can guarantee their
|
|
// construction in this function) rather than statically.
|
|
static void
|
|
Init(const malloc_table_t* aMallocTable)
|
|
{
|
|
MOZ_ASSERT(!gIsDMDRunning);
|
|
|
|
gMallocTable = aMallocTable;
|
|
|
|
// DMD is controlled by the |DMD| environment variable.
|
|
// - If it's unset or empty or "0", DMD doesn't run.
|
|
// - Otherwise, the contents dictate DMD's behaviour.
|
|
|
|
char* e = getenv("DMD");
|
|
StatusMsg("$DMD = '%s'\n", e);
|
|
|
|
if (!e || strcmp(e, "") == 0 || strcmp(e, "0") == 0) {
|
|
StatusMsg("DMD is not enabled\n");
|
|
return;
|
|
}
|
|
|
|
// Parse $DMD env var.
|
|
gOptions = InfallibleAllocPolicy::new_<Options>(e);
|
|
|
|
StatusMsg("DMD is enabled\n");
|
|
|
|
#ifdef XP_MACOSX
|
|
// On Mac OS X we need to call StackWalkInitCriticalAddress() very early
|
|
// (prior to the creation of any mutexes, apparently) otherwise we can get
|
|
// hangs when getting stack traces (bug 821577). But
|
|
// StackWalkInitCriticalAddress() isn't exported from xpcom/, so instead we
|
|
// just call NS_StackWalk, because that calls StackWalkInitCriticalAddress().
|
|
// See the comment above StackWalkInitCriticalAddress() for more details.
|
|
(void)NS_StackWalk(NopStackWalkCallback, /* skipFrames */ 0,
|
|
/* maxFrames */ 1, nullptr, 0, nullptr);
|
|
#endif
|
|
|
|
gStateLock = InfallibleAllocPolicy::new_<Mutex>();
|
|
|
|
gSmallBlockActualSizeCounter = 0;
|
|
|
|
DMD_CREATE_TLS_INDEX(gTlsIndex);
|
|
|
|
{
|
|
AutoLockState lock;
|
|
|
|
gStackTraceTable = InfallibleAllocPolicy::new_<StackTraceTable>();
|
|
gStackTraceTable->init(8192);
|
|
|
|
gBlockTable = InfallibleAllocPolicy::new_<BlockTable>();
|
|
gBlockTable->init(8192);
|
|
}
|
|
|
|
if (gOptions->IsTestMode()) {
|
|
// Do all necessary allocations before setting gIsDMDRunning so those
|
|
// allocations don't show up in our results. Once gIsDMDRunning is set we
|
|
// are intercepting malloc et al. in earnest.
|
|
auto f1 = MakeUnique<FpWriteFunc>(OpenOutputFile("full1.json"));
|
|
auto f2 = MakeUnique<FpWriteFunc>(OpenOutputFile("full2.json"));
|
|
auto f3 = MakeUnique<FpWriteFunc>(OpenOutputFile("full3.json"));
|
|
auto f4 = MakeUnique<FpWriteFunc>(OpenOutputFile("full4.json"));
|
|
gIsDMDRunning = true;
|
|
|
|
StatusMsg("running test mode...\n");
|
|
RunTestMode(Move(f1), Move(f2), Move(f3), Move(f4));
|
|
StatusMsg("finished test mode\n");
|
|
exit(0);
|
|
}
|
|
|
|
gIsDMDRunning = true;
|
|
}
|
|
|
|
//---------------------------------------------------------------------------
|
|
// DMD reporting and unreporting
|
|
//---------------------------------------------------------------------------
|
|
|
|
static void
|
|
ReportHelper(const void* aPtr, bool aReportedOnAlloc)
|
|
{
|
|
if (!gIsDMDRunning || !aPtr) {
|
|
return;
|
|
}
|
|
|
|
Thread* t = Thread::Fetch();
|
|
|
|
AutoBlockIntercepts block(t);
|
|
AutoLockState lock;
|
|
|
|
if (BlockTable::Ptr p = gBlockTable->lookup(aPtr)) {
|
|
p->Report(t, aReportedOnAlloc);
|
|
} else {
|
|
// We have no record of the block. Do nothing. Either:
|
|
// - We're sampling and we skipped this block. This is likely.
|
|
// - It's a bogus pointer. This is unlikely because Report() is almost
|
|
// always called in conjunction with a malloc_size_of-style function.
|
|
}
|
|
}
|
|
|
|
MOZ_EXPORT void
|
|
Report(const void* aPtr)
|
|
{
|
|
ReportHelper(aPtr, /* onAlloc */ false);
|
|
}
|
|
|
|
MOZ_EXPORT void
|
|
ReportOnAlloc(const void* aPtr)
|
|
{
|
|
ReportHelper(aPtr, /* onAlloc */ true);
|
|
}
|
|
|
|
//---------------------------------------------------------------------------
|
|
// DMD output
|
|
//---------------------------------------------------------------------------
|
|
|
|
// The version number of the output format. Increment this if you make
|
|
// backwards-incompatible changes to the format.
|
|
//
|
|
// Version history:
|
|
// - 1: The original format (bug 1044709).
|
|
//
|
|
static const int kOutputVersionNumber = 1;
|
|
|
|
// Note that, unlike most SizeOf* functions, this function does not take a
|
|
// |mozilla::MallocSizeOf| argument. That's because those arguments are
|
|
// primarily to aid DMD track heap blocks... but DMD deliberately doesn't track
|
|
// heap blocks it allocated for itself!
|
|
//
|
|
// SizeOfInternal should be called while you're holding the state lock and
|
|
// while intercepts are blocked; SizeOf acquires the lock and blocks
|
|
// intercepts.
|
|
|
|
static void
|
|
SizeOfInternal(Sizes* aSizes)
|
|
{
|
|
MOZ_ASSERT(gStateLock->IsLocked());
|
|
MOZ_ASSERT(Thread::Fetch()->InterceptsAreBlocked());
|
|
|
|
aSizes->Clear();
|
|
|
|
if (!gIsDMDRunning) {
|
|
return;
|
|
}
|
|
|
|
StackTraceSet usedStackTraces;
|
|
GatherUsedStackTraces(usedStackTraces);
|
|
|
|
for (StackTraceTable::Range r = gStackTraceTable->all();
|
|
!r.empty();
|
|
r.popFront()) {
|
|
StackTrace* const& st = r.front();
|
|
|
|
if (usedStackTraces.has(st)) {
|
|
aSizes->mStackTracesUsed += MallocSizeOf(st);
|
|
} else {
|
|
aSizes->mStackTracesUnused += MallocSizeOf(st);
|
|
}
|
|
}
|
|
|
|
aSizes->mStackTraceTable =
|
|
gStackTraceTable->sizeOfIncludingThis(MallocSizeOf);
|
|
|
|
aSizes->mBlockTable = gBlockTable->sizeOfIncludingThis(MallocSizeOf);
|
|
}
|
|
|
|
MOZ_EXPORT void
|
|
SizeOf(Sizes* aSizes)
|
|
{
|
|
aSizes->Clear();
|
|
|
|
if (!gIsDMDRunning) {
|
|
return;
|
|
}
|
|
|
|
AutoBlockIntercepts block(Thread::Fetch());
|
|
AutoLockState lock;
|
|
SizeOfInternal(aSizes);
|
|
}
|
|
|
|
MOZ_EXPORT void
|
|
ClearReports()
|
|
{
|
|
if (!gIsDMDRunning) {
|
|
return;
|
|
}
|
|
|
|
AutoLockState lock;
|
|
|
|
// Unreport all blocks that were marked reported by a memory reporter. This
|
|
// excludes those that were reported on allocation, because they need to keep
|
|
// their reported marking.
|
|
for (BlockTable::Range r = gBlockTable->all(); !r.empty(); r.popFront()) {
|
|
r.front().UnreportIfNotReportedOnAlloc();
|
|
}
|
|
}
|
|
|
|
MOZ_EXPORT bool
|
|
IsRunning()
|
|
{
|
|
return gIsDMDRunning;
|
|
}
|
|
|
|
// This function converts an integer to base-32. |aBuf| must have space for at
|
|
// least eight chars, which is the space needed to hold 'Dffffff' (including
|
|
// the terminating null char), which is the base-32 representation of
|
|
// 0xffffffff.
|
|
//
|
|
// We use base-32 values for indexing into the traceTable and the frameTable,
|
|
// for the following reasons.
|
|
//
|
|
// - Base-32 gives more compact indices than base-16.
|
|
//
|
|
// - 32 is a power-of-two, which makes the necessary div/mod calculations fast.
|
|
//
|
|
// - We can (and do) choose non-numeric digits for base-32. When
|
|
// inspecting/debugging the JSON output, non-numeric indices are easier to
|
|
// search for than numeric indices.
|
|
//
|
|
char* Base32(uint32_t aN, char* aBuf, size_t aBufLen)
|
|
{
|
|
static const char digits[] = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdef";
|
|
|
|
char* b = aBuf + aBufLen - 1;
|
|
*b = '\0';
|
|
do {
|
|
b--;
|
|
if (b == aBuf) {
|
|
MOZ_CRASH("Base32 buffer too small");
|
|
}
|
|
*b = digits[aN % 32];
|
|
aN /= 32;
|
|
} while (aN);
|
|
|
|
return b;
|
|
}
|
|
|
|
// Converts a pointer to a unique ID. Reuses the existing ID for the pointer if
|
|
// it's been seen before.
|
|
static const char* Id(PointerIdMap& aIdMap, uint32_t& aNextId,
|
|
const void* aPtr, char* aBuf, size_t aBufLen)
|
|
{
|
|
uint32_t id;
|
|
PointerIdMap::AddPtr p = aIdMap.lookupForAdd(aPtr);
|
|
if (!p) {
|
|
id = aNextId++;
|
|
(void)aIdMap.add(p, aPtr, id);
|
|
} else {
|
|
id = p->value();
|
|
}
|
|
return Base32(id, aBuf, aBufLen);
|
|
}
|
|
|
|
static void
|
|
AnalyzeReportsImpl(JSONWriter& aWriter)
|
|
{
|
|
if (!gIsDMDRunning) {
|
|
return;
|
|
}
|
|
|
|
AutoBlockIntercepts block(Thread::Fetch());
|
|
AutoLockState lock;
|
|
|
|
// Allocate this on the heap instead of the stack because it's fairly large.
|
|
auto locService = InfallibleAllocPolicy::new_<CodeAddressService>();
|
|
|
|
StackTraceSet usedStackTraces;
|
|
usedStackTraces.init(512);
|
|
|
|
PointerSet usedPcs;
|
|
usedPcs.init(512);
|
|
|
|
PointerIdMap idMap;
|
|
idMap.init(512);
|
|
|
|
static int analysisCount = 1;
|
|
StatusMsg("Dump %d {\n", analysisCount++);
|
|
|
|
aWriter.Start();
|
|
{
|
|
#define ID(p) Id(idMap, id, p, idBuf, idBufLen)
|
|
|
|
aWriter.IntProperty("version", kOutputVersionNumber);
|
|
|
|
aWriter.StartObjectProperty("invocation");
|
|
{
|
|
aWriter.StringProperty("dmdEnvVar", gOptions->DMDEnvVar());
|
|
aWriter.IntProperty("sampleBelowSize", gOptions->SampleBelowSize());
|
|
}
|
|
aWriter.EndObject();
|
|
|
|
StatusMsg(" Constructing the heap block list...\n");
|
|
|
|
static const size_t idBufLen = 16;
|
|
char idBuf[idBufLen];
|
|
uint32_t id = 0;
|
|
|
|
aWriter.StartArrayProperty("blockList");
|
|
{
|
|
for (BlockTable::Range r = gBlockTable->all(); !r.empty(); r.popFront()) {
|
|
const Block& b = r.front();
|
|
b.AddStackTracesToTable(usedStackTraces);
|
|
|
|
aWriter.StartObjectElement(aWriter.SingleLineStyle);
|
|
{
|
|
if (!b.IsSampled()) {
|
|
aWriter.IntProperty("req", b.ReqSize());
|
|
if (b.SlopSize() > 0) {
|
|
aWriter.IntProperty("slop", b.SlopSize());
|
|
}
|
|
}
|
|
aWriter.StringProperty("alloc", ID(b.AllocStackTrace()));
|
|
if (b.NumReports() > 0) {
|
|
aWriter.StartArrayProperty("reps");
|
|
{
|
|
if (b.ReportStackTrace1()) {
|
|
aWriter.StringElement(ID(b.ReportStackTrace1()));
|
|
}
|
|
if (b.ReportStackTrace2()) {
|
|
aWriter.StringElement(ID(b.ReportStackTrace2()));
|
|
}
|
|
}
|
|
aWriter.EndArray();
|
|
}
|
|
}
|
|
aWriter.EndObject();
|
|
}
|
|
}
|
|
aWriter.EndArray();
|
|
|
|
StatusMsg(" Constructing the stack trace table...\n");
|
|
|
|
aWriter.StartObjectProperty("traceTable");
|
|
{
|
|
for (StackTraceSet::Enum e(usedStackTraces); !e.empty(); e.popFront()) {
|
|
const StackTrace* const st = e.front();
|
|
aWriter.StartArrayProperty(ID(st), aWriter.SingleLineStyle);
|
|
{
|
|
for (uint32_t i = 0; i < st->Length(); i++) {
|
|
const void* pc = st->Pc(i);
|
|
aWriter.StringElement(ID(pc));
|
|
usedPcs.put(pc);
|
|
}
|
|
}
|
|
aWriter.EndArray();
|
|
}
|
|
}
|
|
aWriter.EndObject();
|
|
|
|
StatusMsg(" Constructing the stack frame table...\n");
|
|
|
|
aWriter.StartObjectProperty("frameTable");
|
|
{
|
|
static const size_t locBufLen = 1024;
|
|
char locBuf[locBufLen];
|
|
|
|
for (PointerSet::Enum e(usedPcs); !e.empty(); e.popFront()) {
|
|
const void* const pc = e.front();
|
|
|
|
// Use 0 for the frame number. See the JSON format description comment
|
|
// in DMD.h to understand why.
|
|
locService->GetLocation(0, pc, locBuf, locBufLen);
|
|
aWriter.StringProperty(ID(pc), locBuf);
|
|
}
|
|
}
|
|
aWriter.EndObject();
|
|
|
|
#undef ID
|
|
}
|
|
aWriter.End();
|
|
|
|
if (gOptions->ShowDumpStats()) {
|
|
Sizes sizes;
|
|
SizeOfInternal(&sizes);
|
|
|
|
static const size_t kBufLen = 64;
|
|
char buf1[kBufLen];
|
|
char buf2[kBufLen];
|
|
char buf3[kBufLen];
|
|
|
|
StatusMsg(" Execution measurements {\n");
|
|
|
|
StatusMsg(" Data structures that persist after Dump() ends {\n");
|
|
|
|
StatusMsg(" Used stack traces: %10s bytes\n",
|
|
Show(sizes.mStackTracesUsed, buf1, kBufLen));
|
|
|
|
StatusMsg(" Unused stack traces: %10s bytes\n",
|
|
Show(sizes.mStackTracesUnused, buf1, kBufLen));
|
|
|
|
StatusMsg(" Stack trace table: %10s bytes (%s entries, %s used)\n",
|
|
Show(sizes.mStackTraceTable, buf1, kBufLen),
|
|
Show(gStackTraceTable->capacity(), buf2, kBufLen),
|
|
Show(gStackTraceTable->count(), buf3, kBufLen));
|
|
|
|
StatusMsg(" Block table: %10s bytes (%s entries, %s used)\n",
|
|
Show(sizes.mBlockTable, buf1, kBufLen),
|
|
Show(gBlockTable->capacity(), buf2, kBufLen),
|
|
Show(gBlockTable->count(), buf3, kBufLen));
|
|
|
|
StatusMsg(" }\n");
|
|
StatusMsg(" Data structures that are destroyed after Dump() ends {\n");
|
|
|
|
StatusMsg(" Location service: %10s bytes\n",
|
|
Show(locService->SizeOfIncludingThis(MallocSizeOf), buf1, kBufLen));
|
|
StatusMsg(" Used stack traces set: %10s bytes\n",
|
|
Show(usedStackTraces.sizeOfExcludingThis(MallocSizeOf), buf1, kBufLen));
|
|
StatusMsg(" Used PCs set: %10s bytes\n",
|
|
Show(usedPcs.sizeOfExcludingThis(MallocSizeOf), buf1, kBufLen));
|
|
StatusMsg(" Pointer ID map: %10s bytes\n",
|
|
Show(idMap.sizeOfExcludingThis(MallocSizeOf), buf1, kBufLen));
|
|
|
|
StatusMsg(" }\n");
|
|
StatusMsg(" Counts {\n");
|
|
|
|
size_t hits = locService->NumCacheHits();
|
|
size_t misses = locService->NumCacheMisses();
|
|
size_t requests = hits + misses;
|
|
StatusMsg(" Location service: %10s requests\n",
|
|
Show(requests, buf1, kBufLen));
|
|
|
|
size_t count = locService->CacheCount();
|
|
size_t capacity = locService->CacheCapacity();
|
|
StatusMsg(" Location service cache: "
|
|
"%4.1f%% hit rate, %.1f%% occupancy at end\n",
|
|
Percent(hits, requests), Percent(count, capacity));
|
|
|
|
StatusMsg(" }\n");
|
|
StatusMsg(" }\n");
|
|
}
|
|
|
|
InfallibleAllocPolicy::delete_(locService);
|
|
|
|
StatusMsg("}\n");
|
|
}
|
|
|
|
MOZ_EXPORT void
|
|
AnalyzeReports(JSONWriter& aWriter)
|
|
{
|
|
AnalyzeReportsImpl(aWriter);
|
|
ClearReports();
|
|
}
|
|
|
|
//---------------------------------------------------------------------------
|
|
// Testing
|
|
//---------------------------------------------------------------------------
|
|
|
|
// This function checks that heap blocks that have the same stack trace but
|
|
// different (or no) reporters get aggregated separately.
|
|
void foo()
|
|
{
|
|
char* a[6];
|
|
for (int i = 0; i < 6; i++) {
|
|
a[i] = (char*) malloc(128 - 16*i);
|
|
}
|
|
|
|
for (int i = 0; i <= 1; i++)
|
|
Report(a[i]); // reported
|
|
Report(a[2]); // reported
|
|
Report(a[3]); // reported
|
|
// a[4], a[5] unreported
|
|
}
|
|
|
|
// This stops otherwise-unused variables from being optimized away.
|
|
static void
|
|
UseItOrLoseIt(void* a)
|
|
{
|
|
char buf[64];
|
|
sprintf(buf, "%p\n", a);
|
|
fwrite(buf, 1, strlen(buf) + 1, stderr);
|
|
}
|
|
|
|
// The output from this should be tested with check_test_output.py. It's been
|
|
// tested on Linux64, and probably will give different results on other
|
|
// platforms.
|
|
static void
|
|
RunTestMode(UniquePtr<FpWriteFunc> aF1, UniquePtr<FpWriteFunc> aF2,
|
|
UniquePtr<FpWriteFunc> aF3, UniquePtr<FpWriteFunc> aF4)
|
|
{
|
|
// The first part of this test requires sampling to be disabled.
|
|
gOptions->SetSampleBelowSize(1);
|
|
|
|
//---------
|
|
|
|
// AnalyzeReports 1. Zero for everything.
|
|
JSONWriter writer1(Move(aF1));
|
|
AnalyzeReports(writer1);
|
|
|
|
//---------
|
|
|
|
// AnalyzeReports 2: 1 freed, 9 out of 10 unreported.
|
|
// AnalyzeReports 3: still present and unreported.
|
|
int i;
|
|
char* a;
|
|
for (i = 0; i < 10; i++) {
|
|
a = (char*) malloc(100);
|
|
UseItOrLoseIt(a);
|
|
}
|
|
free(a);
|
|
|
|
// Min-sized block.
|
|
// AnalyzeReports 2: reported.
|
|
// AnalyzeReports 3: thrice-reported.
|
|
char* a2 = (char*) malloc(0);
|
|
Report(a2);
|
|
|
|
// Operator new[].
|
|
// AnalyzeReports 2: reported.
|
|
// AnalyzeReports 3: reportedness carries over, due to ReportOnAlloc.
|
|
char* b = new char[10];
|
|
ReportOnAlloc(b);
|
|
|
|
// ReportOnAlloc, then freed.
|
|
// AnalyzeReports 2: freed, irrelevant.
|
|
// AnalyzeReports 3: freed, irrelevant.
|
|
char* b2 = new char;
|
|
ReportOnAlloc(b2);
|
|
free(b2);
|
|
|
|
// AnalyzeReports 2: reported 4 times.
|
|
// AnalyzeReports 3: freed, irrelevant.
|
|
char* c = (char*) calloc(10, 3);
|
|
Report(c);
|
|
for (int i = 0; i < 3; i++) {
|
|
Report(c);
|
|
}
|
|
|
|
// AnalyzeReports 2: ignored.
|
|
// AnalyzeReports 3: irrelevant.
|
|
Report((void*)(intptr_t)i);
|
|
|
|
// jemalloc rounds this up to 8192.
|
|
// AnalyzeReports 2: reported.
|
|
// AnalyzeReports 3: freed.
|
|
char* e = (char*) malloc(4096);
|
|
e = (char*) realloc(e, 4097);
|
|
Report(e);
|
|
|
|
// First realloc is like malloc; second realloc is shrinking.
|
|
// AnalyzeReports 2: reported.
|
|
// AnalyzeReports 3: re-reported.
|
|
char* e2 = (char*) realloc(nullptr, 1024);
|
|
e2 = (char*) realloc(e2, 512);
|
|
Report(e2);
|
|
|
|
// First realloc is like malloc; second realloc creates a min-sized block.
|
|
// XXX: on Windows, second realloc frees the block.
|
|
// AnalyzeReports 2: reported.
|
|
// AnalyzeReports 3: freed, irrelevant.
|
|
char* e3 = (char*) realloc(nullptr, 1023);
|
|
//e3 = (char*) realloc(e3, 0);
|
|
MOZ_ASSERT(e3);
|
|
Report(e3);
|
|
|
|
// AnalyzeReports 2: freed, irrelevant.
|
|
// AnalyzeReports 3: freed, irrelevant.
|
|
char* f = (char*) malloc(64);
|
|
free(f);
|
|
|
|
// AnalyzeReports 2: ignored.
|
|
// AnalyzeReports 3: irrelevant.
|
|
Report((void*)(intptr_t)0x0);
|
|
|
|
// AnalyzeReports 2: mixture of reported and unreported.
|
|
// AnalyzeReports 3: all unreported.
|
|
foo();
|
|
foo();
|
|
|
|
// AnalyzeReports 2: twice-reported.
|
|
// AnalyzeReports 3: twice-reported.
|
|
char* g1 = (char*) malloc(77);
|
|
ReportOnAlloc(g1);
|
|
ReportOnAlloc(g1);
|
|
|
|
// AnalyzeReports 2: twice-reported.
|
|
// AnalyzeReports 3: once-reported.
|
|
char* g2 = (char*) malloc(78);
|
|
Report(g2);
|
|
ReportOnAlloc(g2);
|
|
|
|
// AnalyzeReports 2: twice-reported.
|
|
// AnalyzeReports 3: once-reported.
|
|
char* g3 = (char*) malloc(79);
|
|
ReportOnAlloc(g3);
|
|
Report(g3);
|
|
|
|
// All the odd-ball ones.
|
|
// AnalyzeReports 2: all unreported.
|
|
// AnalyzeReports 3: all freed, irrelevant.
|
|
// XXX: no memalign on Mac
|
|
//void* x = memalign(64, 65); // rounds up to 128
|
|
//UseItOrLoseIt(x);
|
|
// XXX: posix_memalign doesn't work on B2G
|
|
//void* y;
|
|
//posix_memalign(&y, 128, 129); // rounds up to 256
|
|
//UseItOrLoseIt(y);
|
|
// XXX: valloc doesn't work on Windows.
|
|
//void* z = valloc(1); // rounds up to 4096
|
|
//UseItOrLoseIt(z);
|
|
//aligned_alloc(64, 256); // XXX: C11 only
|
|
|
|
// AnalyzeReports 2.
|
|
JSONWriter writer2(Move(aF2));
|
|
AnalyzeReports(writer2);
|
|
|
|
//---------
|
|
|
|
Report(a2);
|
|
Report(a2);
|
|
free(c);
|
|
free(e);
|
|
Report(e2);
|
|
free(e3);
|
|
//free(x);
|
|
//free(y);
|
|
//free(z);
|
|
|
|
// AnalyzeReports 3.
|
|
JSONWriter writer3(Move(aF3));
|
|
AnalyzeReports(writer3);
|
|
|
|
//---------
|
|
|
|
// Clear all knowledge of existing blocks to give us a clean slate.
|
|
gBlockTable->clear();
|
|
|
|
gOptions->SetSampleBelowSize(128);
|
|
|
|
char* s;
|
|
|
|
// This equals the sample size, and so is reported exactly. It should be
|
|
// listed before records of the same size that are sampled.
|
|
s = (char*) malloc(128);
|
|
UseItOrLoseIt(s);
|
|
|
|
// This exceeds the sample size, and so is reported exactly.
|
|
s = (char*) malloc(144);
|
|
UseItOrLoseIt(s);
|
|
|
|
// These together constitute exactly one sample.
|
|
for (int i = 0; i < 16; i++) {
|
|
s = (char*) malloc(8);
|
|
UseItOrLoseIt(s);
|
|
}
|
|
MOZ_ASSERT(gSmallBlockActualSizeCounter == 0);
|
|
|
|
// These fall 8 bytes short of a full sample.
|
|
for (int i = 0; i < 15; i++) {
|
|
s = (char*) malloc(8);
|
|
UseItOrLoseIt(s);
|
|
}
|
|
MOZ_ASSERT(gSmallBlockActualSizeCounter == 120);
|
|
|
|
// This exceeds the sample size, and so is recorded exactly.
|
|
s = (char*) malloc(256);
|
|
UseItOrLoseIt(s);
|
|
MOZ_ASSERT(gSmallBlockActualSizeCounter == 120);
|
|
|
|
// This gets more than to a full sample from the |i < 15| loop above.
|
|
s = (char*) malloc(96);
|
|
UseItOrLoseIt(s);
|
|
MOZ_ASSERT(gSmallBlockActualSizeCounter == 88);
|
|
|
|
// This gets to another full sample.
|
|
for (int i = 0; i < 5; i++) {
|
|
s = (char*) malloc(8);
|
|
UseItOrLoseIt(s);
|
|
}
|
|
MOZ_ASSERT(gSmallBlockActualSizeCounter == 0);
|
|
|
|
// This allocates 16, 32, ..., 128 bytes, which results in a heap block
|
|
// record that contains a mix of sample and non-sampled blocks, and so should
|
|
// be printed with '~' signs.
|
|
for (int i = 1; i <= 8; i++) {
|
|
s = (char*) malloc(i * 16);
|
|
UseItOrLoseIt(s);
|
|
}
|
|
MOZ_ASSERT(gSmallBlockActualSizeCounter == 64);
|
|
|
|
// At the end we're 64 bytes into the current sample so we report ~1,424
|
|
// bytes of allocation overall, which is 64 less than the real value 1,488.
|
|
|
|
// AnalyzeReports 4.
|
|
JSONWriter writer4(Move(aF4));
|
|
AnalyzeReports(writer4);
|
|
}
|
|
|
|
} // namespace dmd
|
|
} // namespace mozilla
|