X-Git-Url: https://git.saurik.com/apple/javascriptcore.git/blobdiff_plain/a253471d7f8e4d91bf6ebabab00155c3b387d3d0..93a3786624b2768d89bfa27e46598dc64e2fb70a:/interpreter/JSStack.cpp?ds=inline diff --git a/interpreter/JSStack.cpp b/interpreter/JSStack.cpp new file mode 100644 index 0000000..ec2962a --- /dev/null +++ b/interpreter/JSStack.cpp @@ -0,0 +1,151 @@ +/* + * Copyright (C) 2008 Apple Inc. All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * 3. Neither the name of Apple Computer, Inc. ("Apple") nor the names of + * its contributors may be used to endorse or promote products derived + * from this software without specific prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY + * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY + * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES + * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; + * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND + * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF + * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include "config.h" +#include "JSStack.h" +#include "JSStackInlines.h" + +#include "ConservativeRoots.h" +#include "Interpreter.h" + +namespace JSC { + +static size_t committedBytesCount = 0; + +static Mutex& stackStatisticsMutex() +{ + DEFINE_STATIC_LOCAL(Mutex, staticMutex, ()); + return staticMutex; +} + +JSStack::JSStack(VM& vm, size_t capacity) + : m_end(0) + , m_topCallFrame(vm.topCallFrame) +{ + ASSERT(capacity && isPageAligned(capacity)); + + m_reservation = PageReservation::reserve(roundUpAllocationSize(capacity * sizeof(Register), commitSize), OSAllocator::JSVMStackPages); + m_end = static_cast(m_reservation.base()); + m_commitEnd = static_cast(m_reservation.base()); + + disableErrorStackReserve(); + + m_topCallFrame = 0; +} + +JSStack::~JSStack() +{ + void* base = m_reservation.base(); + m_reservation.decommit(base, reinterpret_cast(m_commitEnd) - reinterpret_cast(base)); + addToCommittedByteCount(-(reinterpret_cast(m_commitEnd) - reinterpret_cast(base))); + m_reservation.deallocate(); +} + +bool JSStack::growSlowCase(Register* newEnd) +{ + // If we have already committed enough memory to satisfy this request, + // just update the end pointer and return. + if (newEnd <= m_commitEnd) { + m_end = newEnd; + return true; + } + + // Compute the chunk size of additional memory to commit, and see if we + // have it is still within our budget. If not, we'll fail to grow and + // return false. + long delta = roundUpAllocationSize(reinterpret_cast(newEnd) - reinterpret_cast(m_commitEnd), commitSize); + if (reinterpret_cast(m_commitEnd) + delta > reinterpret_cast(m_useableEnd)) + return false; + + // Otherwise, the growth is still within our budget. Go ahead and commit + // it and return true. + m_reservation.commit(m_commitEnd, delta); + addToCommittedByteCount(delta); + m_commitEnd = reinterpret_cast_ptr(reinterpret_cast(m_commitEnd) + delta); + m_end = newEnd; + return true; +} + +void JSStack::gatherConservativeRoots(ConservativeRoots& conservativeRoots) +{ + conservativeRoots.add(begin(), getTopOfStack()); +} + +void JSStack::gatherConservativeRoots(ConservativeRoots& conservativeRoots, JITStubRoutineSet& jitStubRoutines, DFGCodeBlocks& dfgCodeBlocks) +{ + conservativeRoots.add(begin(), getTopOfStack(), jitStubRoutines, dfgCodeBlocks); +} + +void JSStack::releaseExcessCapacity() +{ + ptrdiff_t delta = reinterpret_cast(m_commitEnd) - reinterpret_cast(m_reservation.base()); + m_reservation.decommit(m_reservation.base(), delta); + addToCommittedByteCount(-delta); + m_commitEnd = static_cast(m_reservation.base()); +} + +void JSStack::initializeThreading() +{ + stackStatisticsMutex(); +} + +size_t JSStack::committedByteCount() +{ + MutexLocker locker(stackStatisticsMutex()); + return committedBytesCount; +} + +void JSStack::addToCommittedByteCount(long byteCount) +{ + MutexLocker locker(stackStatisticsMutex()); + ASSERT(static_cast(committedBytesCount) + byteCount > -1); + committedBytesCount += byteCount; +} + +void JSStack::enableErrorStackReserve() +{ + m_useableEnd = reservationEnd(); +} + +void JSStack::disableErrorStackReserve() +{ + char* useableEnd = reinterpret_cast(reservationEnd()) - commitSize; + m_useableEnd = reinterpret_cast_ptr(useableEnd); + + // By the time we get here, we are guaranteed to be destructing the last + // Interpreter::ErrorHandlingMode that enabled this reserve in the first + // place. That means the stack space beyond m_useableEnd before we + // enabled the reserve was not previously in use. Hence, it is safe to + // shrink back to that m_useableEnd. + if (m_end > m_useableEnd) { + ASSERT(m_topCallFrame->frameExtent() <= m_useableEnd); + shrink(m_useableEnd); + } +} + +} // namespace JSC