// Copyright 2011 the V8 project authors. All rights reserved.
// Redistribution and use in source and binary forms, with or without
// modification, are permitted provided that the following conditions are
// met:
//
// * Redistributions of source code must retain the above copyright
// notice, this list of conditions and the following disclaimer.
// * Redistributions in binary form must reproduce the above
// copyright notice, this list of conditions and the following
// disclaimer in the documentation and/or other materials provided
// with the distribution.
// * Neither the name of Google Inc. nor the names of its
// contributors may be used to endorse or promote products derived
// from this software without specific prior written permission.
//
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
#ifndef V8_ZONE_INL_H_
#define V8_ZONE_INL_H_
#include "isolate.h"
#include "zone.h"
#include "v8-counters.h"
namespace v8 {
namespace internal {
AssertNoZoneAllocation::AssertNoZoneAllocation()
: prev_(Isolate::Current()->zone_allow_allocation()) {
Isolate::Current()->set_zone_allow_allocation(false);
}
AssertNoZoneAllocation::~AssertNoZoneAllocation() {
Isolate::Current()->set_zone_allow_allocation(prev_);
}
inline void* Zone::New(int size) {
ASSERT(Isolate::Current()->zone_allow_allocation());
ASSERT(ZoneScope::nesting() > 0);
// Round up the requested size to fit the alignment.
size = RoundUp(size, kAlignment);
// Check if the requested size is available without expanding.
Address result = position_;
if ((position_ += size) > limit_) result = NewExpand(size);
// Check that the result has the proper alignment and return it.
ASSERT(IsAddressAligned(result, kAlignment, 0));
allocation_size_ += size;
return reinterpret_cast<void*>(result);
}
template <typename T>
T* Zone::NewArray(int length) {
return static_cast<T*>(New(length * sizeof(T)));
}
bool Zone::excess_allocation() {
return segment_bytes_allocated_ > zone_excess_limit_;
}
void Zone::adjust_segment_bytes_allocated(int delta) {
segment_bytes_allocated_ += delta;
isolate_->counters()->zone_segment_bytes()->Set(segment_bytes_allocated_);
}
template <typename Config>
ZoneSplayTree<Config>::~ZoneSplayTree() {
// Reset the root to avoid unneeded iteration over all tree nodes
// in the destructor. For a zone-allocated tree, nodes will be
// freed by the Zone.
SplayTree<Config, ZoneListAllocationPolicy>::ResetRoot();
}
// TODO(isolates): for performance reasons, this should be replaced with a new
// operator that takes the zone in which the object should be
// allocated.
void* ZoneObject::operator new(size_t size) {
return ZONE->New(static_cast<int>(size));
}
void* ZoneObject::operator new(size_t size, Zone* zone) {
return zone->New(static_cast<int>(size));
}
inline void* ZoneListAllocationPolicy::New(int size) {
return ZONE->New(size);
}
ZoneScope::ZoneScope(ZoneScopeMode mode)
: isolate_(Isolate::Current()),
mode_(mode) {
isolate_->zone()->scope_nesting_++;
}
bool ZoneScope::ShouldDeleteOnExit() {
return isolate_->zone()->scope_nesting_ == 1 && mode_ == DELETE_ON_EXIT;
}
int ZoneScope::nesting() {
return Isolate::Current()->zone()->scope_nesting_;
}
} } // namespace v8::internal
#endif // V8_ZONE_INL_H_