/* * Copyright (C) 2008 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #define ATRACE_TAG ATRACE_TAG_DALVIK /* * Garbage-collecting memory allocator. */ #include "Dalvik.h" #include "alloc/HeapBitmap.h" #include "alloc/Verify.h" #include "alloc/Heap.h" #include "alloc/HeapInternal.h" #include "alloc/DdmHeap.h" #include "alloc/HeapSource.h" #include "alloc/MarkSweep.h" #include "os/os.h" #include <sys/mman.h> #include <sys/resource.h> #include <sys/time.h> #include <limits.h> #include <errno.h> #include <cutils/trace.h> static const GcSpec kGcForMallocSpec = { true, /* isPartial */ false, /* isConcurrent */ true, /* doPreserve */ "GC_FOR_ALLOC" }; const GcSpec *GC_FOR_MALLOC = &kGcForMallocSpec; static const GcSpec kGcConcurrentSpec = { true, /* isPartial */ true, /* isConcurrent */ true, /* doPreserve */ "GC_CONCURRENT" }; const GcSpec *GC_CONCURRENT = &kGcConcurrentSpec; static const GcSpec kGcExplicitSpec = { false, /* isPartial */ true, /* isConcurrent */ true, /* doPreserve */ "GC_EXPLICIT" }; const GcSpec *GC_EXPLICIT = &kGcExplicitSpec; static const GcSpec kGcBeforeOomSpec = { false, /* isPartial */ false, /* isConcurrent */ false, /* doPreserve */ "GC_BEFORE_OOM" }; const GcSpec *GC_BEFORE_OOM = &kGcBeforeOomSpec; /* * Initialize the GC heap. * * Returns true if successful, false otherwise. */ bool dvmHeapStartup() { GcHeap *gcHeap; if (gDvm.heapGrowthLimit == 0) { gDvm.heapGrowthLimit = gDvm.heapMaximumSize; } gcHeap = dvmHeapSourceStartup(gDvm.heapStartingSize, gDvm.heapMaximumSize, gDvm.heapGrowthLimit); if (gcHeap == NULL) { return false; } gcHeap->ddmHpifWhen = 0; gcHeap->ddmHpsgWhen = 0; gcHeap->ddmHpsgWhat = 0; gcHeap->ddmNhsgWhen = 0; gcHeap->ddmNhsgWhat = 0; gDvm.gcHeap = gcHeap; /* Set up the lists we'll use for cleared reference objects. */ gcHeap->clearedReferences = NULL; if (!dvmCardTableStartup(gDvm.heapMaximumSize, gDvm.heapGrowthLimit)) { LOGE_HEAP("card table startup failed."); return false; } return true; } bool dvmHeapStartupAfterZygote() { return dvmHeapSourceStartupAfterZygote(); } void dvmHeapShutdown() { //TODO: make sure we're locked if (gDvm.gcHeap != NULL) { dvmCardTableShutdown(); /* Destroy the heap. Any outstanding pointers will point to * unmapped memory (unless/until someone else maps it). This * frees gDvm.gcHeap as a side-effect. */ dvmHeapSourceShutdown(&gDvm.gcHeap); } } /* * Shutdown any threads internal to the heap. */ void dvmHeapThreadShutdown() { dvmHeapSourceThreadShutdown(); } /* * Grab the lock, but put ourselves into THREAD_VMWAIT if it looks like * we're going to have to wait on the mutex. */ bool dvmLockHeap() { if (dvmTryLockMutex(&gDvm.gcHeapLock) != 0) { Thread *self; ThreadStatus oldStatus; self = dvmThreadSelf(); oldStatus = dvmChangeStatus(self, THREAD_VMWAIT); dvmLockMutex(&gDvm.gcHeapLock); dvmChangeStatus(self, oldStatus); } return true; } void dvmUnlockHeap() { dvmUnlockMutex(&gDvm.gcHeapLock); } /* Do a full garbage collection, which may grow the * heap as a side-effect if the live set is large. */ static void gcForMalloc(bool clearSoftReferences) { if (gDvm.allocProf.enabled) { Thread* self = dvmThreadSelf(); gDvm.allocProf.gcCount++; if (self != NULL) { self->allocProf.gcCount++; } } /* This may adjust the soft limit as a side-effect. */ const GcSpec *spec = clearSoftReferences ? GC_BEFORE_OOM : GC_FOR_MALLOC; dvmCollectGarbageInternal(spec); } /* Try as hard as possible to allocate some memory. */ static void *tryMalloc(size_t size) { void *ptr; //TODO: figure out better heuristics // There will be a lot of churn if someone allocates a bunch of // big objects in a row, and we hit the frag case each time. // A full GC for each. // Maybe we grow the heap in bigger leaps // Maybe we skip the GC if the size is large and we did one recently // (number of allocations ago) (watch for thread effects) // DeflateTest allocs a bunch of ~128k buffers w/in 0-5 allocs of each other // (or, at least, there are only 0-5 objects swept each time) ptr = dvmHeapSourceAlloc(size); if (ptr != NULL) { return ptr; } /* * The allocation failed. If the GC is running, block until it * completes and retry. */ if (gDvm.gcHeap->gcRunning) { /* * The GC is concurrently tracing the heap. Release the heap * lock, wait for the GC to complete, and retrying allocating. */ dvmWaitForConcurrentGcToComplete(); } else { /* * Try a foreground GC since a concurrent GC is not currently running. */ gcForMalloc(false); } ptr = dvmHeapSourceAlloc(size); if (ptr != NULL) { return ptr; } /* Even that didn't work; this is an exceptional state. * Try harder, growing the heap if necessary. */ ptr = dvmHeapSourceAllocAndGrow(size); if (ptr != NULL) { size_t newHeapSize; newHeapSize = dvmHeapSourceGetIdealFootprint(); //TODO: may want to grow a little bit more so that the amount of free // space is equal to the old free space + the utilization slop for // the new allocation. LOGI_HEAP("Grow heap (frag case) to " "%zu.%03zuMB for %zu-byte allocation", FRACTIONAL_MB(newHeapSize), size); return ptr; } /* Most allocations should have succeeded by now, so the heap * is really full, really fragmented, or the requested size is * really big. Do another GC, collecting SoftReferences this * time. The VM spec requires that all SoftReferences have * been collected and cleared before throwing an OOME. */ //TODO: wait for the finalizers from the previous GC to finish LOGI_HEAP("Forcing collection of SoftReferences for %zu-byte allocation", size); gcForMalloc(true); ptr = dvmHeapSourceAllocAndGrow(size); if (ptr != NULL) { return ptr; } //TODO: maybe wait for finalizers and try one last time LOGE_HEAP("Out of memory on a %zd-byte allocation.", size); //TODO: tell the HeapSource to dump its state dvmDumpThread(dvmThreadSelf(), false); return NULL; } /* Throw an OutOfMemoryError if there's a thread to attach it to. * Avoid recursing. * * The caller must not be holding the heap lock, or else the allocations * in dvmThrowException() will deadlock. */ static void throwOOME() { Thread *self; if ((self = dvmThreadSelf()) != NULL) { /* If the current (failing) dvmMalloc() happened as part of thread * creation/attachment before the thread became part of the root set, * we can't rely on the thread-local trackedAlloc table, so * we can't keep track of a real allocated OOME object. But, since * the thread is in the process of being created, it won't have * a useful stack anyway, so we may as well make things easier * by throwing the (stackless) pre-built OOME. */ if (dvmIsOnThreadList(self) && !self->throwingOOME) { /* Let ourselves know that we tried to throw an OOM * error in the normal way in case we run out of * memory trying to allocate it inside dvmThrowException(). */ self->throwingOOME = true; /* Don't include a description string; * one fewer allocation. */ dvmThrowOutOfMemoryError(NULL); } else { /* * This thread has already tried to throw an OutOfMemoryError, * which probably means that we're running out of memory * while recursively trying to throw. * * To avoid any more allocation attempts, "throw" a pre-built * OutOfMemoryError object (which won't have a useful stack trace). * * Note that since this call can't possibly allocate anything, * we don't care about the state of self->throwingOOME * (which will usually already be set). */ dvmSetException(self, gDvm.outOfMemoryObj); } /* We're done with the possible recursion. */ self->throwingOOME = false; } } /* * Allocate storage on the GC heap. We guarantee 8-byte alignment. * * The new storage is zeroed out. * * Note that, in rare cases, this could get called while a GC is in * progress. If a non-VM thread tries to attach itself through JNI, * it will need to allocate some objects. If this becomes annoying to * deal with, we can block it at the source, but holding the allocation * mutex should be enough. * * In rare circumstances (JNI AttachCurrentThread) we can be called * from a non-VM thread. * * Use ALLOC_DONT_TRACK when we either don't want to track an allocation * (because it's being done for the interpreter "new" operation and will * be part of the root set immediately) or we can't (because this allocation * is for a brand new thread). * * Returns NULL and throws an exception on failure. * * TODO: don't do a GC if the debugger thinks all threads are suspended */ void* dvmMalloc(size_t size, int flags) { void *ptr; dvmLockHeap(); /* Try as hard as possible to allocate some memory. */ ptr = tryMalloc(size); if (ptr != NULL) { /* We've got the memory. */ if (gDvm.allocProf.enabled) { Thread* self = dvmThreadSelf(); gDvm.allocProf.allocCount++; gDvm.allocProf.allocSize += size; if (self != NULL) { self->allocProf.allocCount++; self->allocProf.allocSize += size; } } } else { /* The allocation failed. */ if (gDvm.allocProf.enabled) { Thread* self = dvmThreadSelf(); gDvm.allocProf.failedAllocCount++; gDvm.allocProf.failedAllocSize += size; if (self != NULL) { self->allocProf.failedAllocCount++; self->allocProf.failedAllocSize += size; } } } dvmUnlockHeap(); if (ptr != NULL) { /* * If caller hasn't asked us not to track it, add it to the * internal tracking list. */ if ((flags & ALLOC_DONT_TRACK) == 0) { dvmAddTrackedAlloc((Object*)ptr, NULL); } } else { /* * The allocation failed; throw an OutOfMemoryError. */ throwOOME(); } return ptr; } /* * Returns true iff <obj> points to a valid allocated object. */ bool dvmIsValidObject(const Object* obj) { /* Don't bother if it's NULL or not 8-byte aligned. */ if (obj != NULL && ((uintptr_t)obj & (8-1)) == 0) { /* Even if the heap isn't locked, this shouldn't return * any false negatives. The only mutation that could * be happening is allocation, which means that another * thread could be in the middle of a read-modify-write * to add a new bit for a new object. However, that * RMW will have completed by the time any other thread * could possibly see the new pointer, so there is no * danger of dvmIsValidObject() being called on a valid * pointer whose bit isn't set. * * Freeing will only happen during the sweep phase, which * only happens while the heap is locked. */ return dvmHeapSourceContains(obj); } return false; } size_t dvmObjectSizeInHeap(const Object *obj) { return dvmHeapSourceChunkSize(obj); } static void verifyRootsAndHeap() { dvmVerifyRoots(); dvmVerifyBitmap(dvmHeapSourceGetLiveBits()); } /* * Initiate garbage collection. * * NOTES: * - If we don't hold gDvm.threadListLock, it's possible for a thread to * be added to the thread list while we work. The thread should NOT * start executing, so this is only interesting when we start chasing * thread stacks. (Before we do so, grab the lock.) * * We are not allowed to GC when the debugger has suspended the VM, which * is awkward because debugger requests can cause allocations. The easiest * way to enforce this is to refuse to GC on an allocation made by the * JDWP thread -- we have to expand the heap or fail. */ void dvmCollectGarbageInternal(const GcSpec* spec) { GcHeap *gcHeap = gDvm.gcHeap; u4 gcEnd = 0; u4 rootStart = 0 , rootEnd = 0; u4 dirtyStart = 0, dirtyEnd = 0; size_t numObjectsFreed, numBytesFreed; size_t currAllocated, currFootprint; size_t percentFree; int oldThreadPriority = INT_MAX; /* The heap lock must be held. */ if (gcHeap->gcRunning) { LOGW_HEAP("Attempted recursive GC"); return; } // Trace the beginning of the top-level GC. if (spec == GC_FOR_MALLOC) { ATRACE_BEGIN("GC (alloc)"); } else if (spec == GC_CONCURRENT) { ATRACE_BEGIN("GC (concurrent)"); } else if (spec == GC_EXPLICIT) { ATRACE_BEGIN("GC (explicit)"); } else if (spec == GC_BEFORE_OOM) { ATRACE_BEGIN("GC (before OOM)"); } else { ATRACE_BEGIN("GC (unknown)"); } gcHeap->gcRunning = true; rootStart = dvmGetRelativeTimeMsec(); ATRACE_BEGIN("GC: Threads Suspended"); // Suspend A dvmSuspendAllThreads(SUSPEND_FOR_GC); /* * If we are not marking concurrently raise the priority of the * thread performing the garbage collection. */ if (!spec->isConcurrent) { oldThreadPriority = os_raiseThreadPriority(); } if (gDvm.preVerify) { LOGV_HEAP("Verifying roots and heap before GC"); verifyRootsAndHeap(); } dvmMethodTraceGCBegin(); /* Set up the marking context. */ if (!dvmHeapBeginMarkStep(spec->isPartial)) { ATRACE_END(); // Suspend A ATRACE_END(); // Top-level GC LOGE_HEAP("dvmHeapBeginMarkStep failed; aborting"); dvmAbort(); } /* Mark the set of objects that are strongly reachable from the roots. */ LOGD_HEAP("Marking..."); dvmHeapMarkRootSet(); /* dvmHeapScanMarkedObjects() will build the lists of known * instances of the Reference classes. */ assert(gcHeap->softReferences == NULL); assert(gcHeap->weakReferences == NULL); assert(gcHeap->finalizerReferences == NULL); assert(gcHeap->phantomReferences == NULL); assert(gcHeap->clearedReferences == NULL); if (spec->isConcurrent) { /* * Resume threads while tracing from the roots. We unlock the * heap to allow mutator threads to allocate from free space. */ dvmClearCardTable(); dvmUnlockHeap(); dvmResumeAllThreads(SUSPEND_FOR_GC); ATRACE_END(); // Suspend A rootEnd = dvmGetRelativeTimeMsec(); } /* Recursively mark any objects that marked objects point to strongly. * If we're not collecting soft references, soft-reachable * objects will also be marked. */ LOGD_HEAP("Recursing..."); dvmHeapScanMarkedObjects(); if (spec->isConcurrent) { /* * Re-acquire the heap lock and perform the final thread * suspension. */ dirtyStart = dvmGetRelativeTimeMsec(); dvmLockHeap(); ATRACE_BEGIN("GC: Threads Suspended"); // Suspend B dvmSuspendAllThreads(SUSPEND_FOR_GC); /* * As no barrier intercepts root updates, we conservatively * assume all roots may be gray and re-mark them. */ dvmHeapReMarkRootSet(); /* * With the exception of reference objects and weak interned * strings, all gray objects should now be on dirty cards. */ if (gDvm.verifyCardTable) { dvmVerifyCardTable(); } /* * Recursively mark gray objects pointed to by the roots or by * heap objects dirtied during the concurrent mark. */ dvmHeapReScanMarkedObjects(); } /* * All strongly-reachable objects have now been marked. Process * weakly-reachable objects discovered while tracing. */ dvmHeapProcessReferences(&gcHeap->softReferences, spec->doPreserve == false, &gcHeap->weakReferences, &gcHeap->finalizerReferences, &gcHeap->phantomReferences); #if defined(WITH_JIT) /* * Patching a chaining cell is very cheap as it only updates 4 words. It's * the overhead of stopping all threads and synchronizing the I/D cache * that makes it expensive. * * Therefore we batch those work orders in a queue and go through them * when threads are suspended for GC. */ dvmCompilerPerformSafePointChecks(); #endif LOGD_HEAP("Sweeping..."); dvmHeapSweepSystemWeaks(); /* * Live objects have a bit set in the mark bitmap, swap the mark * and live bitmaps. The sweep can proceed concurrently viewing * the new live bitmap as the old mark bitmap, and vice versa. */ dvmHeapSourceSwapBitmaps(); if (gDvm.postVerify) { LOGV_HEAP("Verifying roots and heap after GC"); verifyRootsAndHeap(); } if (spec->isConcurrent) { dvmUnlockHeap(); dvmResumeAllThreads(SUSPEND_FOR_GC); ATRACE_END(); // Suspend B dirtyEnd = dvmGetRelativeTimeMsec(); } dvmHeapSweepUnmarkedObjects(spec->isPartial, spec->isConcurrent, &numObjectsFreed, &numBytesFreed); LOGD_HEAP("Cleaning up..."); dvmHeapFinishMarkStep(); if (spec->isConcurrent) { dvmLockHeap(); } LOGD_HEAP("Done."); /* Now's a good time to adjust the heap size, since * we know what our utilization is. * * This doesn't actually resize any memory; * it just lets the heap grow more when necessary. */ dvmHeapSourceGrowForUtilization(); currAllocated = dvmHeapSourceGetValue(HS_BYTES_ALLOCATED, NULL, 0); currFootprint = dvmHeapSourceGetValue(HS_FOOTPRINT, NULL, 0); dvmMethodTraceGCEnd(); LOGV_HEAP("GC finished"); gcHeap->gcRunning = false; LOGV_HEAP("Resuming threads"); if (spec->isConcurrent) { /* * Wake-up any threads that blocked after a failed allocation * request. */ dvmBroadcastCond(&gDvm.gcHeapCond); } if (!spec->isConcurrent) { dvmResumeAllThreads(SUSPEND_FOR_GC); ATRACE_END(); // Suspend A dirtyEnd = dvmGetRelativeTimeMsec(); /* * Restore the original thread scheduling priority if it was * changed at the start of the current garbage collection. */ if (oldThreadPriority != INT_MAX) { os_lowerThreadPriority(oldThreadPriority); } } /* * Move queue of pending references back into Java. */ dvmEnqueueClearedReferences(&gDvm.gcHeap->clearedReferences); gcEnd = dvmGetRelativeTimeMsec(); percentFree = 100 - (size_t)(100.0f * (float)currAllocated / currFootprint); if (!spec->isConcurrent) { u4 markSweepTime = dirtyEnd - rootStart; u4 gcTime = gcEnd - rootStart; bool isSmall = numBytesFreed > 0 && numBytesFreed < 1024; ALOGD("%s freed %s%zdK, %d%% free %zdK/%zdK, paused %ums, total %ums", spec->reason, isSmall ? "<" : "", numBytesFreed ? MAX(numBytesFreed / 1024, 1) : 0, percentFree, currAllocated / 1024, currFootprint / 1024, markSweepTime, gcTime); } else { u4 rootTime = rootEnd - rootStart; u4 dirtyTime = dirtyEnd - dirtyStart; u4 gcTime = gcEnd - rootStart; bool isSmall = numBytesFreed > 0 && numBytesFreed < 1024; ALOGD("%s freed %s%zdK, %d%% free %zdK/%zdK, paused %ums+%ums, total %ums", spec->reason, isSmall ? "<" : "", numBytesFreed ? MAX(numBytesFreed / 1024, 1) : 0, percentFree, currAllocated / 1024, currFootprint / 1024, rootTime, dirtyTime, gcTime); } if (gcHeap->ddmHpifWhen != 0) { LOGD_HEAP("Sending VM heap info to DDM"); dvmDdmSendHeapInfo(gcHeap->ddmHpifWhen, false); } if (gcHeap->ddmHpsgWhen != 0) { LOGD_HEAP("Dumping VM heap to DDM"); dvmDdmSendHeapSegments(false, false); } if (gcHeap->ddmNhsgWhen != 0) { LOGD_HEAP("Dumping native heap to DDM"); dvmDdmSendHeapSegments(false, true); } ATRACE_END(); // Top-level GC } /* * If the concurrent GC is running, wait for it to finish. The caller * must hold the heap lock. * * Note: the second dvmChangeStatus() could stall if we were in RUNNING * on entry, and some other thread has asked us to suspend. In that * case we will be suspended with the heap lock held, which can lead to * deadlock if the other thread tries to do something with the managed heap. * For example, the debugger might suspend us and then execute a method that * allocates memory. We can avoid this situation by releasing the lock * before self-suspending. (The developer can work around this specific * situation by single-stepping the VM. Alternatively, we could disable * concurrent GC when the debugger is attached, but that might change * behavior more than is desirable.) * * This should not be a problem in production, because any GC-related * activity will grab the lock before issuing a suspend-all. (We may briefly * suspend when the GC thread calls dvmUnlockHeap before dvmResumeAllThreads, * but there's no risk of deadlock.) */ bool dvmWaitForConcurrentGcToComplete() { ATRACE_BEGIN("GC: Wait For Concurrent"); bool waited = gDvm.gcHeap->gcRunning; Thread *self = dvmThreadSelf(); assert(self != NULL); u4 start = dvmGetRelativeTimeMsec(); while (gDvm.gcHeap->gcRunning) { ThreadStatus oldStatus = dvmChangeStatus(self, THREAD_VMWAIT); dvmWaitCond(&gDvm.gcHeapCond, &gDvm.gcHeapLock); dvmChangeStatus(self, oldStatus); } u4 end = dvmGetRelativeTimeMsec(); if (end - start > 0) { ALOGD("WAIT_FOR_CONCURRENT_GC blocked %ums", end - start); } ATRACE_END(); return waited; }