/*
* Copyright (C) 2008 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/*
* Garbage-collecting memory allocator.
*/
#include "Dalvik.h"
#include "alloc/HeapTable.h"
#include "alloc/Heap.h"
#include "alloc/HeapInternal.h"
#include "alloc/DdmHeap.h"
#include "alloc/HeapSource.h"
#include "alloc/MarkSweep.h"
#include "utils/threads.h" // need Android thread priorities
#define kInvalidPriority 10000
#include <cutils/sched_policy.h>
#include <sys/time.h>
#include <sys/resource.h>
#include <limits.h>
#include <errno.h>
#define kNonCollectableRefDefault 16
#define kFinalizableRefDefault 128
static const char* GcReasonStr[] = {
[GC_FOR_MALLOC] = "GC_FOR_MALLOC",
[GC_EXPLICIT] = "GC_EXPLICIT",
[GC_EXTERNAL_ALLOC] = "GC_EXTERNAL_ALLOC",
[GC_HPROF_DUMP_HEAP] = "GC_HPROF_DUMP_HEAP"
};
/*
* Initialize the GC heap.
*
* Returns true if successful, false otherwise.
*/
bool dvmHeapStartup()
{
GcHeap *gcHeap;
#if defined(WITH_ALLOC_LIMITS)
gDvm.checkAllocLimits = false;
gDvm.allocationLimit = -1;
#endif
gcHeap = dvmHeapSourceStartup(gDvm.heapSizeStart, gDvm.heapSizeMax);
if (gcHeap == NULL) {
return false;
}
gcHeap->heapWorkerCurrentObject = NULL;
gcHeap->heapWorkerCurrentMethod = NULL;
gcHeap->heapWorkerInterpStartTime = 0LL;
gcHeap->softReferenceCollectionState = SR_COLLECT_NONE;
gcHeap->softReferenceHeapSizeThreshold = gDvm.heapSizeStart;
gcHeap->ddmHpifWhen = 0;
gcHeap->ddmHpsgWhen = 0;
gcHeap->ddmHpsgWhat = 0;
gcHeap->ddmNhsgWhen = 0;
gcHeap->ddmNhsgWhat = 0;
#if WITH_HPROF
gcHeap->hprofDumpOnGc = false;
gcHeap->hprofContext = NULL;
#endif
/* This needs to be set before we call dvmHeapInitHeapRefTable().
*/
gDvm.gcHeap = gcHeap;
/* Set up the table we'll use for ALLOC_NO_GC.
*/
if (!dvmHeapInitHeapRefTable(&gcHeap->nonCollectableRefs,
kNonCollectableRefDefault))
{
LOGE_HEAP("Can't allocate GC_NO_ALLOC table\n");
goto fail;
}
/* Set up the lists and lock we'll use for finalizable
* and reference objects.
*/
dvmInitMutex(&gDvm.heapWorkerListLock);
gcHeap->finalizableRefs = NULL;
gcHeap->pendingFinalizationRefs = NULL;
gcHeap->referenceOperations = NULL;
/* Initialize the HeapWorker locks and other state
* that the GC uses.
*/
dvmInitializeHeapWorkerState();
return true;
fail:
gDvm.gcHeap = NULL;
dvmHeapSourceShutdown(gcHeap);
return false;
}
bool dvmHeapStartupAfterZygote()
{
/* Update our idea of the last GC start time so that we
* don't use the last time that Zygote happened to GC.
*/
gDvm.gcHeap->gcStartTime = dvmGetRelativeTimeUsec();
return dvmHeapSourceStartupAfterZygote();
}
void dvmHeapShutdown()
{
//TODO: make sure we're locked
if (gDvm.gcHeap != NULL) {
GcHeap *gcHeap;
gcHeap = gDvm.gcHeap;
gDvm.gcHeap = NULL;
/* Tables are allocated on the native heap;
* they need to be cleaned up explicitly.
* The process may stick around, so we don't
* want to leak any native memory.
*/
dvmHeapFreeHeapRefTable(&gcHeap->nonCollectableRefs);
dvmHeapFreeLargeTable(gcHeap->finalizableRefs);
gcHeap->finalizableRefs = NULL;
dvmHeapFreeLargeTable(gcHeap->pendingFinalizationRefs);
gcHeap->pendingFinalizationRefs = NULL;
dvmHeapFreeLargeTable(gcHeap->referenceOperations);
gcHeap->referenceOperations = NULL;
/* Destroy the heap. Any outstanding pointers
* will point to unmapped memory (unless/until
* someone else maps it). This frees gcHeap
* as a side-effect.
*/
dvmHeapSourceShutdown(gcHeap);
}
}
/*
* We've been asked to allocate something we can't, e.g. an array so
* large that (length * elementWidth) is larger than 2^31.
*
* _The Java Programming Language_, 4th edition, says, "you can be sure
* that all SoftReferences to softly reachable objects will be cleared
* before an OutOfMemoryError is thrown."
*
* It's unclear whether that holds for all situations where an OOM can
* be thrown, or just in the context of an allocation that fails due
* to lack of heap space. For simplicity we just throw the exception.
*
* (OOM due to actually running out of space is handled elsewhere.)
*/
void dvmThrowBadAllocException(const char* msg)
{
dvmThrowException("Ljava/lang/OutOfMemoryError;", msg);
}
/*
* Grab the lock, but put ourselves into THREAD_VMWAIT if it looks like
* we're going to have to wait on the mutex.
*/
bool dvmLockHeap()
{
if (pthread_mutex_trylock(&gDvm.gcHeapLock) != 0) {
Thread *self;
ThreadStatus oldStatus;
int cc;
self = dvmThreadSelf();
if (self != NULL) {
oldStatus = dvmChangeStatus(self, THREAD_VMWAIT);
} else {
LOGI("ODD: waiting on heap lock, no self\n");
oldStatus = -1; // shut up gcc
}
cc = pthread_mutex_lock(&gDvm.gcHeapLock);
assert(cc == 0);
if (self != NULL) {
dvmChangeStatus(self, oldStatus);
}
}
return true;
}
void dvmUnlockHeap()
{
dvmUnlockMutex(&gDvm.gcHeapLock);
}
/* Pop an object from the list of pending finalizations and
* reference clears/enqueues, and return the object.
* The caller must call dvmReleaseTrackedAlloc()
* on the object when finished.
*
* Typically only called by the heap worker thread.
*/
Object *dvmGetNextHeapWorkerObject(HeapWorkerOperation *op)
{
Object *obj;
LargeHeapRefTable *table;
GcHeap *gcHeap = gDvm.gcHeap;
assert(op != NULL);
obj = NULL;
dvmLockMutex(&gDvm.heapWorkerListLock);
/* We must handle reference operations before finalizations.
* If:
* a) Someone subclasses WeakReference and overrides clear()
* b) A reference of this type is the last reference to
* a finalizable object
* then we need to guarantee that the overridden clear() is called
* on the reference before finalize() is called on the referent.
* Both of these operations will always be scheduled at the same
* time, so handling reference operations first will guarantee
* the required order.
*/
obj = dvmHeapGetNextObjectFromLargeTable(&gcHeap->referenceOperations);
if (obj != NULL) {
uintptr_t workBits;
workBits = (uintptr_t)obj & WORKER_ENQUEUE;
assert(workBits != 0);
obj = (Object *)((uintptr_t)obj & ~WORKER_ENQUEUE);
*op = workBits;
} else {
obj = dvmHeapGetNextObjectFromLargeTable(
&gcHeap->pendingFinalizationRefs);
if (obj != NULL) {
*op = WORKER_FINALIZE;
}
}
if (obj != NULL) {
/* Don't let the GC collect the object until the
* worker thread is done with it.
*
* This call is safe; it uses thread-local storage
* and doesn't acquire any locks.
*/
dvmAddTrackedAlloc(obj, NULL);
}
dvmUnlockMutex(&gDvm.heapWorkerListLock);
return obj;
}
/* Used for a heap size change hysteresis to avoid collecting
* SoftReferences when the heap only grows by a small amount.
*/
#define SOFT_REFERENCE_GROWTH_SLACK (128 * 1024)
/* Whenever the effective heap size may have changed,
* this function must be called.
*/
void dvmHeapSizeChanged()
{
GcHeap *gcHeap = gDvm.gcHeap;
size_t currentHeapSize;
currentHeapSize = dvmHeapSourceGetIdealFootprint();
/* See if the heap size has changed enough that we should care
* about it.
*/
if (currentHeapSize <= gcHeap->softReferenceHeapSizeThreshold -
4 * SOFT_REFERENCE_GROWTH_SLACK)
{
/* The heap has shrunk enough that we'll use this as a new
* threshold. Since we're doing better on space, there's
* no need to collect any SoftReferences.
*
* This is 4x the growth hysteresis because we don't want
* to snap down so easily after a shrink. If we just cleared
* up a bunch of SoftReferences, we don't want to disallow
* any new ones from being created.
* TODO: determine if the 4x is important, needed, or even good
*/
gcHeap->softReferenceHeapSizeThreshold = currentHeapSize;
gcHeap->softReferenceCollectionState = SR_COLLECT_NONE;
} else if (currentHeapSize >= gcHeap->softReferenceHeapSizeThreshold +
SOFT_REFERENCE_GROWTH_SLACK)
{
/* The heap has grown enough to warrant collecting SoftReferences.
*/
gcHeap->softReferenceHeapSizeThreshold = currentHeapSize;
gcHeap->softReferenceCollectionState = SR_COLLECT_SOME;
}
}
/* Do a full garbage collection, which may grow the
* heap as a side-effect if the live set is large.
*/
static void gcForMalloc(bool collectSoftReferences)
{
#ifdef WITH_PROFILER
if (gDvm.allocProf.enabled) {
Thread* self = dvmThreadSelf();
gDvm.allocProf.gcCount++;
if (self != NULL) {
self->allocProf.gcCount++;
}
}
#endif
/* This may adjust the soft limit as a side-effect.
*/
LOGD_HEAP("dvmMalloc initiating GC%s\n",
collectSoftReferences ? "(collect SoftReferences)" : "");
dvmCollectGarbageInternal(collectSoftReferences, GC_FOR_MALLOC);
}
/* Try as hard as possible to allocate some memory.
*/
static DvmHeapChunk *tryMalloc(size_t size)
{
DvmHeapChunk *hc;
/* Don't try too hard if there's no way the allocation is
* going to succeed. We have to collect SoftReferences before
* throwing an OOME, though.
*/
if (size >= gDvm.heapSizeMax) {
LOGW_HEAP("dvmMalloc(%zu/0x%08zx): "
"someone's allocating a huge buffer\n", size, size);
hc = NULL;
goto collect_soft_refs;
}
//TODO: figure out better heuristics
// There will be a lot of churn if someone allocates a bunch of
// big objects in a row, and we hit the frag case each time.
// A full GC for each.
// Maybe we grow the heap in bigger leaps
// Maybe we skip the GC if the size is large and we did one recently
// (number of allocations ago) (watch for thread effects)
// DeflateTest allocs a bunch of ~128k buffers w/in 0-5 allocs of each other
// (or, at least, there are only 0-5 objects swept each time)
hc = dvmHeapSourceAlloc(size + sizeof(DvmHeapChunk));
if (hc != NULL) {
return hc;
}
/* The allocation failed. Free up some space by doing
* a full garbage collection. This may grow the heap
* if the live set is sufficiently large.
*/
gcForMalloc(false);
hc = dvmHeapSourceAlloc(size + sizeof(DvmHeapChunk));
if (hc != NULL) {
return hc;
}
/* Even that didn't work; this is an exceptional state.
* Try harder, growing the heap if necessary.
*/
hc = dvmHeapSourceAllocAndGrow(size + sizeof(DvmHeapChunk));
dvmHeapSizeChanged();
if (hc != NULL) {
size_t newHeapSize;
newHeapSize = dvmHeapSourceGetIdealFootprint();
//TODO: may want to grow a little bit more so that the amount of free
// space is equal to the old free space + the utilization slop for
// the new allocation.
LOGI_HEAP("Grow heap (frag case) to "
"%zu.%03zuMB for %zu-byte allocation\n",
FRACTIONAL_MB(newHeapSize), size);
return hc;
}
/* Most allocations should have succeeded by now, so the heap
* is really full, really fragmented, or the requested size is
* really big. Do another GC, collecting SoftReferences this
* time. The VM spec requires that all SoftReferences have
* been collected and cleared before throwing an OOME.
*/
//TODO: wait for the finalizers from the previous GC to finish
collect_soft_refs:
LOGI_HEAP("Forcing collection of SoftReferences for %zu-byte allocation\n",
size);
gcForMalloc(true);
hc = dvmHeapSourceAllocAndGrow(size + sizeof(DvmHeapChunk));
dvmHeapSizeChanged();
if (hc != NULL) {
return hc;
}
//TODO: maybe wait for finalizers and try one last time
LOGE_HEAP("Out of memory on a %zd-byte allocation.\n", size);
//TODO: tell the HeapSource to dump its state
dvmDumpThread(dvmThreadSelf(), false);
return NULL;
}
/* Throw an OutOfMemoryError if there's a thread to attach it to.
* Avoid recursing.
*
* The caller must not be holding the heap lock, or else the allocations
* in dvmThrowException() will deadlock.
*/
static void throwOOME()
{
Thread *self;
if ((self = dvmThreadSelf()) != NULL) {
/* If the current (failing) dvmMalloc() happened as part of thread
* creation/attachment before the thread became part of the root set,
* we can't rely on the thread-local trackedAlloc table, so
* we can't keep track of a real allocated OOME object. But, since
* the thread is in the process of being created, it won't have
* a useful stack anyway, so we may as well make things easier
* by throwing the (stackless) pre-built OOME.
*/
if (dvmIsOnThreadList(self) && !self->throwingOOME) {
/* Let ourselves know that we tried to throw an OOM
* error in the normal way in case we run out of
* memory trying to allocate it inside dvmThrowException().
*/
self->throwingOOME = true;
/* Don't include a description string;
* one fewer allocation.
*/
dvmThrowException("Ljava/lang/OutOfMemoryError;", NULL);
} else {
/*
* This thread has already tried to throw an OutOfMemoryError,
* which probably means that we're running out of memory
* while recursively trying to throw.
*
* To avoid any more allocation attempts, "throw" a pre-built
* OutOfMemoryError object (which won't have a useful stack trace).
*
* Note that since this call can't possibly allocate anything,
* we don't care about the state of self->throwingOOME
* (which will usually already be set).
*/
dvmSetException(self, gDvm.outOfMemoryObj);
}
/* We're done with the possible recursion.
*/
self->throwingOOME = false;
}
}
/*
* Allocate storage on the GC heap. We guarantee 8-byte alignment.
*
* The new storage is zeroed out.
*
* Note that, in rare cases, this could get called while a GC is in
* progress. If a non-VM thread tries to attach itself through JNI,
* it will need to allocate some objects. If this becomes annoying to
* deal with, we can block it at the source, but holding the allocation
* mutex should be enough.
*
* In rare circumstances (JNI AttachCurrentThread) we can be called
* from a non-VM thread.
*
* We implement ALLOC_NO_GC by maintaining an internal list of objects
* that should not be collected. This requires no actual flag storage in
* the object itself, which is good, but makes flag queries expensive.
*
* Use ALLOC_DONT_TRACK when we either don't want to track an allocation
* (because it's being done for the interpreter "new" operation and will
* be part of the root set immediately) or we can't (because this allocation
* is for a brand new thread).
*
* Returns NULL and throws an exception on failure.
*
* TODO: don't do a GC if the debugger thinks all threads are suspended
*/
void* dvmMalloc(size_t size, int flags)
{
GcHeap *gcHeap = gDvm.gcHeap;
DvmHeapChunk *hc;
void *ptr;
bool triedGc, triedGrowing;
#if 0
/* handy for spotting large allocations */
if (size >= 100000) {
LOGI("dvmMalloc(%d):\n", size);
dvmDumpThread(dvmThreadSelf(), false);
}
#endif
#if defined(WITH_ALLOC_LIMITS)
/*
* See if they've exceeded the allocation limit for this thread.
*
* A limit value of -1 means "no limit".
*
* This is enabled at compile time because it requires us to do a
* TLS lookup for the Thread pointer. This has enough of a performance
* impact that we don't want to do it if we don't have to. (Now that
* we're using gDvm.checkAllocLimits we may want to reconsider this,
* but it's probably still best to just compile the check out of
* production code -- one less thing to hit on every allocation.)
*/
if (gDvm.checkAllocLimits) {
Thread* self = dvmThreadSelf();
if (self != NULL) {
int count = self->allocLimit;
if (count > 0) {
self->allocLimit--;
} else if (count == 0) {
/* fail! */
assert(!gDvm.initializing);
self->allocLimit = -1;
dvmThrowException("Ldalvik/system/AllocationLimitError;",
"thread allocation limit exceeded");
return NULL;
}
}
}
if (gDvm.allocationLimit >= 0) {
assert(!gDvm.initializing);
gDvm.allocationLimit = -1;
dvmThrowException("Ldalvik/system/AllocationLimitError;",
"global allocation limit exceeded");
return NULL;
}
#endif
dvmLockHeap();
/* Try as hard as possible to allocate some memory.
*/
hc = tryMalloc(size);
if (hc != NULL) {
alloc_succeeded:
/* We've got the memory.
*/
if ((flags & ALLOC_FINALIZABLE) != 0) {
/* This object is an instance of a class that
* overrides finalize(). Add it to the finalizable list.
*
* Note that until DVM_OBJECT_INIT() is called on this
* object, its clazz will be NULL. Since the object is
* in this table, it will be scanned as part of the root
* set. scanObject() explicitly deals with the NULL clazz.
*/
if (!dvmHeapAddRefToLargeTable(&gcHeap->finalizableRefs,
(Object *)hc->data))
{
LOGE_HEAP("dvmMalloc(): no room for any more "
"finalizable objects\n");
dvmAbort();
}
}
ptr = hc->data;
/* The caller may not want us to collect this object.
* If not, throw it in the nonCollectableRefs table, which
* will be added to the root set when we GC.
*
* Note that until DVM_OBJECT_INIT() is called on this
* object, its clazz will be NULL. Since the object is
* in this table, it will be scanned as part of the root
* set. scanObject() explicitly deals with the NULL clazz.
*/
if ((flags & ALLOC_NO_GC) != 0) {
if (!dvmHeapAddToHeapRefTable(&gcHeap->nonCollectableRefs, ptr)) {
LOGE_HEAP("dvmMalloc(): no room for any more "
"ALLOC_NO_GC objects: %zd\n",
dvmHeapNumHeapRefTableEntries(
&gcHeap->nonCollectableRefs));
dvmAbort();
}
}
#ifdef WITH_PROFILER
if (gDvm.allocProf.enabled) {
Thread* self = dvmThreadSelf();
gDvm.allocProf.allocCount++;
gDvm.allocProf.allocSize += size;
if (self != NULL) {
self->allocProf.allocCount++;
self->allocProf.allocSize += size;
}
}
#endif
} else {
/* The allocation failed.
*/
ptr = NULL;
#ifdef WITH_PROFILER
if (gDvm.allocProf.enabled) {
Thread* self = dvmThreadSelf();
gDvm.allocProf.failedAllocCount++;
gDvm.allocProf.failedAllocSize += size;
if (self != NULL) {
self->allocProf.failedAllocCount++;
self->allocProf.failedAllocSize += size;
}
}
#endif
}
dvmUnlockHeap();
if (ptr != NULL) {
/*
* If this block is immediately GCable, and they haven't asked us not
* to track it, add it to the internal tracking list.
*
* If there's no "self" yet, we can't track it. Calls made before
* the Thread exists should use ALLOC_NO_GC.
*/
if ((flags & (ALLOC_DONT_TRACK | ALLOC_NO_GC)) == 0) {
dvmAddTrackedAlloc(ptr, NULL);
}
} else {
/*
* The allocation failed; throw an OutOfMemoryError.
*/
throwOOME();
}
return ptr;
}
/*
* Returns true iff <obj> points to a valid allocated object.
*/
bool dvmIsValidObject(const Object* obj)
{
const DvmHeapChunk *hc;
/* Don't bother if it's NULL or not 8-byte aligned.
*/
hc = ptr2chunk(obj);
if (obj != NULL && ((uintptr_t)hc & (8-1)) == 0) {
/* Even if the heap isn't locked, this shouldn't return
* any false negatives. The only mutation that could
* be happening is allocation, which means that another
* thread could be in the middle of a read-modify-write
* to add a new bit for a new object. However, that
* RMW will have completed by the time any other thread
* could possibly see the new pointer, so there is no
* danger of dvmIsValidObject() being called on a valid
* pointer whose bit isn't set.
*
* Freeing will only happen during the sweep phase, which
* only happens while the heap is locked.
*/
return dvmHeapSourceContains(hc);
}
return false;
}
/*
* Clear flags that were passed into dvmMalloc() et al.
* e.g., ALLOC_NO_GC, ALLOC_DONT_TRACK.
*/
void dvmClearAllocFlags(Object *obj, int mask)
{
if ((mask & ALLOC_NO_GC) != 0) {
dvmLockHeap();
if (dvmIsValidObject(obj)) {
if (!dvmHeapRemoveFromHeapRefTable(&gDvm.gcHeap->nonCollectableRefs,
obj))
{
LOGE_HEAP("dvmMalloc(): failed to remove ALLOC_NO_GC bit from "
"object 0x%08x\n", (uintptr_t)obj);
dvmAbort();
}
//TODO: shrink if the table is very empty
}
dvmUnlockHeap();
}
if ((mask & ALLOC_DONT_TRACK) != 0) {
dvmReleaseTrackedAlloc(obj, NULL);
}
}
size_t dvmObjectSizeInHeap(const Object *obj)
{
return dvmHeapSourceChunkSize(ptr2chunk(obj)) - sizeof(DvmHeapChunk);
}
/*
* Initiate garbage collection.
*
* NOTES:
* - If we don't hold gDvm.threadListLock, it's possible for a thread to
* be added to the thread list while we work. The thread should NOT
* start executing, so this is only interesting when we start chasing
* thread stacks. (Before we do so, grab the lock.)
*
* We are not allowed to GC when the debugger has suspended the VM, which
* is awkward because debugger requests can cause allocations. The easiest
* way to enforce this is to refuse to GC on an allocation made by the
* JDWP thread -- we have to expand the heap or fail.
*/
void dvmCollectGarbageInternal(bool collectSoftReferences, enum GcReason reason)
{
GcHeap *gcHeap = gDvm.gcHeap;
Object *softReferences;
Object *weakReferences;
Object *phantomReferences;
u8 now;
s8 timeSinceLastGc;
s8 gcElapsedTime;
int numFreed;
size_t sizeFreed;
#if DVM_TRACK_HEAP_MARKING
/* Since weak and soft references are always cleared,
* they don't require any marking.
* (Soft are lumped into strong when they aren't cleared.)
*/
size_t strongMarkCount = 0;
size_t strongMarkSize = 0;
size_t finalizeMarkCount = 0;
size_t finalizeMarkSize = 0;
size_t phantomMarkCount = 0;
size_t phantomMarkSize = 0;
#endif
/* The heap lock must be held.
*/
if (gcHeap->gcRunning) {
LOGW_HEAP("Attempted recursive GC\n");
return;
}
gcHeap->gcRunning = true;
now = dvmGetRelativeTimeUsec();
if (gcHeap->gcStartTime != 0) {
timeSinceLastGc = (now - gcHeap->gcStartTime) / 1000;
} else {
timeSinceLastGc = 0;
}
gcHeap->gcStartTime = now;
LOGV_HEAP("%s starting -- suspending threads\n", GcReasonStr[reason]);
dvmSuspendAllThreads(SUSPEND_FOR_GC);
/* Get the priority (the "nice" value) of the current thread. The
* getpriority() call can legitimately return -1, so we have to
* explicitly test errno.
*/
errno = 0;
int oldThreadPriority = kInvalidPriority;
int priorityResult = getpriority(PRIO_PROCESS, 0);
if (errno != 0) {
LOGI_HEAP("getpriority(self) failed: %s\n", strerror(errno));
} else if (priorityResult > ANDROID_PRIORITY_NORMAL) {
/* Current value is numerically greater than "normal", which
* in backward UNIX terms means lower priority.
*/
if (priorityResult >= ANDROID_PRIORITY_BACKGROUND) {
set_sched_policy(dvmGetSysThreadId(), SP_FOREGROUND);
}
if (setpriority(PRIO_PROCESS, 0, ANDROID_PRIORITY_NORMAL) != 0) {
LOGI_HEAP("Unable to elevate priority from %d to %d\n",
priorityResult, ANDROID_PRIORITY_NORMAL);
} else {
/* priority elevated; save value so we can restore it later */
LOGD_HEAP("Elevating priority from %d to %d\n",
priorityResult, ANDROID_PRIORITY_NORMAL);
oldThreadPriority = priorityResult;
}
}
/* Wait for the HeapWorker thread to block.
* (It may also already be suspended in interp code,
* in which case it's not holding heapWorkerLock.)
*/
dvmLockMutex(&gDvm.heapWorkerLock);
/* Make sure that the HeapWorker thread hasn't become
* wedged inside interp code. If it has, this call will
* print a message and abort the VM.
*/
dvmAssertHeapWorkerThreadRunning();
/* Lock the pendingFinalizationRefs list.
*
* Acquire the lock after suspending so the finalizer
* thread can't block in the RUNNING state while
* we try to suspend.
*/
dvmLockMutex(&gDvm.heapWorkerListLock);
#ifdef WITH_PROFILER
dvmMethodTraceGCBegin();
#endif
#if WITH_HPROF
/* Set DUMP_HEAP_ON_DDMS_UPDATE to 1 to enable heap dumps
* whenever DDMS requests a heap update (HPIF chunk).
* The output files will appear in /data/misc, which must
* already exist.
* You must define "WITH_HPROF := true" in your buildspec.mk
* and recompile libdvm for this to work.
*
* To enable stack traces for each allocation, define
* "WITH_HPROF_STACK := true" in buildspec.mk. This option slows down
* allocations and also requires 8 additional bytes per object on the
* GC heap.
*/
#define DUMP_HEAP_ON_DDMS_UPDATE 0
#if DUMP_HEAP_ON_DDMS_UPDATE
gcHeap->hprofDumpOnGc |= (gcHeap->ddmHpifWhen != 0);
#endif
if (gcHeap->hprofDumpOnGc) {
char nameBuf[128];
gcHeap->hprofResult = -1;
if (gcHeap->hprofFileName == NULL) {
/* no filename was provided; invent one */
sprintf(nameBuf, "/data/misc/heap-dump-tm%d-pid%d.hprof",
(int) time(NULL), (int) getpid());
gcHeap->hprofFileName = nameBuf;
}
gcHeap->hprofContext = hprofStartup(gcHeap->hprofFileName,
gcHeap->hprofDirectToDdms);
if (gcHeap->hprofContext != NULL) {
hprofStartHeapDump(gcHeap->hprofContext);
}
gcHeap->hprofDumpOnGc = false;
gcHeap->hprofFileName = NULL;
}
#endif
if (timeSinceLastGc < 10000) {
LOGD_HEAP("GC! (%dms since last GC)\n",
(int)timeSinceLastGc);
} else {
LOGD_HEAP("GC! (%d sec since last GC)\n",
(int)(timeSinceLastGc / 1000));
}
#if DVM_TRACK_HEAP_MARKING
gcHeap->markCount = 0;
gcHeap->markSize = 0;
#endif
/* Set up the marking context.
*/
if (!dvmHeapBeginMarkStep()) {
LOGE_HEAP("dvmHeapBeginMarkStep failed; aborting\n");
dvmAbort();
}
/* Mark the set of objects that are strongly reachable from the roots.
*/
LOGD_HEAP("Marking...");
dvmHeapMarkRootSet();
/* dvmHeapScanMarkedObjects() will build the lists of known
* instances of the Reference classes.
*/
gcHeap->softReferences = NULL;
gcHeap->weakReferences = NULL;
gcHeap->phantomReferences = NULL;
/* Make sure that we don't hard-mark the referents of Reference
* objects by default.
*/
gcHeap->markAllReferents = false;
/* Don't mark SoftReferences if our caller wants us to collect them.
* This has to be set before calling dvmHeapScanMarkedObjects().
*/
if (collectSoftReferences) {
gcHeap->softReferenceCollectionState = SR_COLLECT_ALL;
}
/* Recursively mark any objects that marked objects point to strongly.
* If we're not collecting soft references, soft-reachable
* objects will also be marked.
*/
LOGD_HEAP("Recursing...");
dvmHeapScanMarkedObjects();
#if DVM_TRACK_HEAP_MARKING
strongMarkCount = gcHeap->markCount;
strongMarkSize = gcHeap->markSize;
gcHeap->markCount = 0;
gcHeap->markSize = 0;
#endif
/* Latch these so that the other calls to dvmHeapScanMarkedObjects() don't
* mess with them.
*/
softReferences = gcHeap->softReferences;
weakReferences = gcHeap->weakReferences;
phantomReferences = gcHeap->phantomReferences;
/* All strongly-reachable objects have now been marked.
*/
if (gcHeap->softReferenceCollectionState != SR_COLLECT_NONE) {
LOGD_HEAP("Handling soft references...");
dvmHeapHandleReferences(softReferences, REF_SOFT);
// markCount always zero
/* Now that we've tried collecting SoftReferences,
* fall back to not collecting them. If the heap
* grows, we will start collecting again.
*/
gcHeap->softReferenceCollectionState = SR_COLLECT_NONE;
} // else dvmHeapScanMarkedObjects() already marked the soft-reachable set
LOGD_HEAP("Handling weak references...");
dvmHeapHandleReferences(weakReferences, REF_WEAK);
// markCount always zero
/* Once all weak-reachable objects have been taken
* care of, any remaining unmarked objects can be finalized.
*/
LOGD_HEAP("Finding finalizations...");
dvmHeapScheduleFinalizations();
#if DVM_TRACK_HEAP_MARKING
finalizeMarkCount = gcHeap->markCount;
finalizeMarkSize = gcHeap->markSize;
gcHeap->markCount = 0;
gcHeap->markSize = 0;
#endif
/* Any remaining objects that are not pending finalization
* could be phantom-reachable. This will mark any phantom-reachable
* objects, as well as enqueue their references.
*/
LOGD_HEAP("Handling phantom references...");
dvmHeapHandleReferences(phantomReferences, REF_PHANTOM);
#if DVM_TRACK_HEAP_MARKING
phantomMarkCount = gcHeap->markCount;
phantomMarkSize = gcHeap->markSize;
gcHeap->markCount = 0;
gcHeap->markSize = 0;
#endif
//TODO: take care of JNI weak global references
#if DVM_TRACK_HEAP_MARKING
LOGI_HEAP("Marked objects: %dB strong, %dB final, %dB phantom\n",
strongMarkSize, finalizeMarkSize, phantomMarkSize);
#endif
#ifdef WITH_DEADLOCK_PREDICTION
dvmDumpMonitorInfo("before sweep");
#endif
LOGD_HEAP("Sweeping...");
dvmHeapSweepUnmarkedObjects(&numFreed, &sizeFreed);
#ifdef WITH_DEADLOCK_PREDICTION
dvmDumpMonitorInfo("after sweep");
#endif
LOGD_HEAP("Cleaning up...");
dvmHeapFinishMarkStep();
LOGD_HEAP("Done.");
/* Now's a good time to adjust the heap size, since
* we know what our utilization is.
*
* This doesn't actually resize any memory;
* it just lets the heap grow more when necessary.
*/
dvmHeapSourceGrowForUtilization();
dvmHeapSizeChanged();
#if WITH_HPROF
if (gcHeap->hprofContext != NULL) {
hprofFinishHeapDump(gcHeap->hprofContext);
//TODO: write a HEAP_SUMMARY record
if (hprofShutdown(gcHeap->hprofContext))
gcHeap->hprofResult = 0; /* indicate success */
gcHeap->hprofContext = NULL;
}
#endif
/* Now that we've freed up the GC heap, return any large
* free chunks back to the system. They'll get paged back
* in the next time they're used. Don't do it immediately,
* though; if the process is still allocating a bunch of
* memory, we'll be taking a ton of page faults that we don't
* necessarily need to.
*
* Cancel any old scheduled trims, and schedule a new one.
*/
dvmScheduleHeapSourceTrim(5); // in seconds
#ifdef WITH_PROFILER
dvmMethodTraceGCEnd();
#endif
LOGV_HEAP("GC finished -- resuming threads\n");
gcHeap->gcRunning = false;
dvmUnlockMutex(&gDvm.heapWorkerListLock);
dvmUnlockMutex(&gDvm.heapWorkerLock);
#if defined(WITH_JIT)
extern void dvmCompilerPerformSafePointChecks(void);
/*
* Patching a chaining cell is very cheap as it only updates 4 words. It's
* the overhead of stopping all threads and synchronizing the I/D cache
* that makes it expensive.
*
* Therefore we batch those work orders in a queue and go through them
* when threads are suspended for GC.
*/
dvmCompilerPerformSafePointChecks();
#endif
dvmResumeAllThreads(SUSPEND_FOR_GC);
if (oldThreadPriority != kInvalidPriority) {
if (setpriority(PRIO_PROCESS, 0, oldThreadPriority) != 0) {
LOGW_HEAP("Unable to reset priority to %d: %s\n",
oldThreadPriority, strerror(errno));
} else {
LOGD_HEAP("Reset priority to %d\n", oldThreadPriority);
}
if (oldThreadPriority >= ANDROID_PRIORITY_BACKGROUND) {
set_sched_policy(dvmGetSysThreadId(), SP_BACKGROUND);
}
}
gcElapsedTime = (dvmGetRelativeTimeUsec() - gcHeap->gcStartTime) / 1000;
LOGD("%s freed %d objects / %zd bytes in %dms\n",
GcReasonStr[reason], numFreed, sizeFreed, (int)gcElapsedTime);
dvmLogGcStats(numFreed, sizeFreed, gcElapsedTime);
if (gcHeap->ddmHpifWhen != 0) {
LOGD_HEAP("Sending VM heap info to DDM\n");
dvmDdmSendHeapInfo(gcHeap->ddmHpifWhen, false);
}
if (gcHeap->ddmHpsgWhen != 0) {
LOGD_HEAP("Dumping VM heap to DDM\n");
dvmDdmSendHeapSegments(false, false);
}
if (gcHeap->ddmNhsgWhen != 0) {
LOGD_HEAP("Dumping native heap to DDM\n");
dvmDdmSendHeapSegments(false, true);
}
}
#if WITH_HPROF
/*
* Perform garbage collection, writing heap information to the specified file.
*
* If "fileName" is NULL, a suitable name will be generated automatically.
*
* Returns 0 on success, or an error code on failure.
*/
int hprofDumpHeap(const char* fileName, bool directToDdms)
{
int result;
dvmLockMutex(&gDvm.gcHeapLock);
gDvm.gcHeap->hprofDumpOnGc = true;
gDvm.gcHeap->hprofFileName = fileName;
gDvm.gcHeap->hprofDirectToDdms = directToDdms;
dvmCollectGarbageInternal(false, GC_HPROF_DUMP_HEAP);
result = gDvm.gcHeap->hprofResult;
dvmUnlockMutex(&gDvm.gcHeapLock);
return result;
}
void dvmHeapSetHprofGcScanState(hprof_heap_tag_t state, u4 threadSerialNumber)
{
if (gDvm.gcHeap->hprofContext != NULL) {
hprofSetGcScanState(gDvm.gcHeap->hprofContext, state,
threadSerialNumber);
}
}
#endif