/*
* Copyright (C) 2014 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "scoped_arena_allocator.h"
#include "arena_allocator-inl.h"
#include "memory_tool.h"
namespace art {
static constexpr size_t kMemoryToolRedZoneBytes = 8;
ArenaStack::ArenaStack(ArenaPool* arena_pool)
: DebugStackRefCounter(),
stats_and_pool_(arena_pool),
bottom_arena_(nullptr),
top_arena_(nullptr),
top_ptr_(nullptr),
top_end_(nullptr) {
}
ArenaStack::~ArenaStack() {
DebugStackRefCounter::CheckNoRefs();
stats_and_pool_.pool->FreeArenaChain(bottom_arena_);
}
void ArenaStack::Reset() {
DebugStackRefCounter::CheckNoRefs();
stats_and_pool_.pool->FreeArenaChain(bottom_arena_);
bottom_arena_ = nullptr;
top_arena_ = nullptr;
top_ptr_ = nullptr;
top_end_ = nullptr;
}
MemStats ArenaStack::GetPeakStats() const {
DebugStackRefCounter::CheckNoRefs();
return MemStats("ArenaStack peak", PeakStats(), bottom_arena_);
}
uint8_t* ArenaStack::AllocateFromNextArena(size_t rounded_bytes) {
UpdateBytesAllocated();
size_t allocation_size = std::max(arena_allocator::kArenaDefaultSize, rounded_bytes);
if (UNLIKELY(top_arena_ == nullptr)) {
top_arena_ = bottom_arena_ = stats_and_pool_.pool->AllocArena(allocation_size);
top_arena_->next_ = nullptr;
} else if (top_arena_->next_ != nullptr && top_arena_->next_->Size() >= allocation_size) {
top_arena_ = top_arena_->next_;
} else {
Arena* tail = top_arena_->next_;
top_arena_->next_ = stats_and_pool_.pool->AllocArena(allocation_size);
top_arena_ = top_arena_->next_;
top_arena_->next_ = tail;
}
top_end_ = top_arena_->End();
// top_ptr_ shall be updated by ScopedArenaAllocator.
return top_arena_->Begin();
}
void ArenaStack::UpdatePeakStatsAndRestore(const ArenaAllocatorStats& restore_stats) {
if (PeakStats()->BytesAllocated() < CurrentStats()->BytesAllocated()) {
PeakStats()->Copy(*CurrentStats());
}
CurrentStats()->Copy(restore_stats);
}
void ArenaStack::UpdateBytesAllocated() {
if (top_arena_ != nullptr) {
// Update how many bytes we have allocated into the arena so that the arena pool knows how
// much memory to zero out. Though ScopedArenaAllocator doesn't guarantee the memory is
// zero-initialized, the Arena may be reused by ArenaAllocator which does guarantee this.
size_t allocated = static_cast<size_t>(top_ptr_ - top_arena_->Begin());
if (top_arena_->bytes_allocated_ < allocated) {
top_arena_->bytes_allocated_ = allocated;
}
}
}
void* ArenaStack::AllocWithMemoryTool(size_t bytes, ArenaAllocKind kind) {
// We mark all memory for a newly retrieved arena as inaccessible and then
// mark only the actually allocated memory as defined. That leaves red zones
// and padding between allocations marked as inaccessible.
size_t rounded_bytes = RoundUp(bytes + kMemoryToolRedZoneBytes, 8);
uint8_t* ptr = top_ptr_;
if (UNLIKELY(static_cast<size_t>(top_end_ - ptr) < rounded_bytes)) {
ptr = AllocateFromNextArena(rounded_bytes);
CHECK(ptr != nullptr) << "Failed to allocate memory";
MEMORY_TOOL_MAKE_NOACCESS(ptr, top_end_ - ptr);
}
CurrentStats()->RecordAlloc(bytes, kind);
top_ptr_ = ptr + rounded_bytes;
MEMORY_TOOL_MAKE_UNDEFINED(ptr, bytes);
return ptr;
}
ScopedArenaAllocator::ScopedArenaAllocator(ScopedArenaAllocator&& other) noexcept
: DebugStackReference(std::move(other)),
DebugStackRefCounter(),
ArenaAllocatorStats(other),
arena_stack_(other.arena_stack_),
mark_arena_(other.mark_arena_),
mark_ptr_(other.mark_ptr_),
mark_end_(other.mark_end_) {
other.DebugStackRefCounter::CheckNoRefs();
other.arena_stack_ = nullptr;
}
ScopedArenaAllocator::ScopedArenaAllocator(ArenaStack* arena_stack)
: DebugStackReference(arena_stack),
DebugStackRefCounter(),
ArenaAllocatorStats(*arena_stack->CurrentStats()),
arena_stack_(arena_stack),
mark_arena_(arena_stack->top_arena_),
mark_ptr_(arena_stack->top_ptr_),
mark_end_(arena_stack->top_end_) {
}
ScopedArenaAllocator::~ScopedArenaAllocator() {
if (arena_stack_ != nullptr) {
DoReset();
}
}
void ScopedArenaAllocator::Reset() {
DoReset();
// If this allocator was Create()d, we need to move the arena_stack_->top_ptr_ past *this.
if (mark_ptr_ == reinterpret_cast<uint8_t*>(this)) {
arena_stack_->top_ptr_ = mark_ptr_ + RoundUp(sizeof(ScopedArenaAllocator), 8);
}
}
void ScopedArenaAllocator::DoReset() {
DebugStackReference::CheckTop();
DebugStackRefCounter::CheckNoRefs();
arena_stack_->UpdatePeakStatsAndRestore(*this);
arena_stack_->UpdateBytesAllocated();
if (LIKELY(mark_arena_ != nullptr)) {
arena_stack_->top_arena_ = mark_arena_;
arena_stack_->top_ptr_ = mark_ptr_;
arena_stack_->top_end_ = mark_end_;
} else if (arena_stack_->bottom_arena_ != nullptr) {
mark_arena_ = arena_stack_->top_arena_ = arena_stack_->bottom_arena_;
mark_ptr_ = arena_stack_->top_ptr_ = mark_arena_->Begin();
mark_end_ = arena_stack_->top_end_ = mark_arena_->End();
}
}
} // namespace art