/* * Copyright (C) 2010 Apple Inc. All rights reserved. * Copyright (C) 2011 Nokia Corporation and/or its subsidiary(-ies). All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY APPLE INC. AND ITS CONTRIBUTORS ``AS IS'' * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR ITS CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF * THE POSSIBILITY OF SUCH DAMAGE. */ #include "config.h" #include "OSAllocator.h" #include "PageAllocatorSymbian.h" namespace WTF { // Array to store code chunks used by JIT engine(s) static RPointerArray codeChunksContainer; // The singleton data allocator (non code) static PageAllocatorSymbian dataAllocator; _LIT(KErrorStringInternalConsistency, "OSAllocator:ConsistencyError"); _LIT(KErrorStringChunkCreation, "OSAllocator:ChunkInitError"); _LIT(KErrorStringPageSize, "OSAllocator:WrongPageSize"); // Makes a new code chunk for a JIT engine with everything in committed state static void* allocateCodeChunk(size_t bytes) { RChunk c; TInt error = c.CreateLocalCode(bytes, bytes); __ASSERT_ALWAYS(error == KErrNone, User::Panic(KErrorStringChunkCreation, error)); codeChunksContainer.Append(new SymbianChunk(c.Handle())); return static_cast(c.Base()); } // Frees the _entire_ code chunk in which this address resides. static bool deallocateCodeChunk(void* address) { bool found = false; for (int i = 0; i < codeChunksContainer.Count(); i++) { SymbianChunk* p = codeChunksContainer[i]; if (p && p->contains(address)) { codeChunksContainer.Remove(i); delete p; found = true; } } return found; } // Return the (singleton) object that manages all non-code VM operations static PageAllocatorSymbian* dataAllocatorInstance() { return &dataAllocator; } // Reserve memory and return the base address of the region void* OSAllocator::reserveUncommitted(size_t reservationSize, Usage usage, bool , bool executable) { void* base = 0; if (executable) base = allocateCodeChunk(reservationSize); else base = dataAllocatorInstance()->reserve(reservationSize); return base; } // Inverse operation of reserveUncommitted() void OSAllocator::releaseDecommitted(void* parkedBase, size_t bytes) { if (dataAllocatorInstance()->contains(parkedBase)) dataAllocatorInstance()->release(parkedBase, bytes); // NOOP for code chunks (JIT) because we released them in decommit() } // Commit what was previously reserved via reserveUncommitted() void OSAllocator::commit(void* address, size_t bytes, bool, bool executable) { // For code chunks, we commit (early) in reserveUncommitted(), so NOOP // For data regions, do real work if (!executable) dataAllocatorInstance()->commit(address, bytes); } void OSAllocator::decommit(void* address, size_t bytes) { if (dataAllocatorInstance()->contains(address)) dataAllocatorInstance()->decommit(address, bytes); else deallocateCodeChunk(address); // for code chunk, decommit AND release } void* OSAllocator::reserveAndCommit(size_t bytes, Usage usage, bool writable, bool executable) { void* base = reserveUncommitted(bytes, usage, writable, executable); commit(base, bytes, writable, executable); return base; } // The PageAllocatorSymbian class helps map OSAllocator calls for reserve/commit/decommit // to a single large Symbian chunk. Only works with multiples of page size, and as a corollary // all addresses accepted or returned by it are also page-sized aligned. // Design notes: // - We initialize a chunk up-front with a large reservation size // - The entire reservation reserve is logically divided into pageSized blocks (4K on Symbian) // - The map maintains 1 bit for each of the 4K-sized region in our address space // - OSAllocator::reserveUncommitted() requests lead to 1 or more bits being set in map // to indicate internally reserved state. The VM address corresponding to the first bit is returned. // - OSAllocator::commit() actually calls RChunk.commit() and commits *all or part* of the region // reserved via reserveUncommitted() previously. // - OSAllocator::decommit() calls RChunk.decommit() // - OSAllocator::releaseDecommitted() unparks all the bits in the map, but trusts that a previously // call to decommit() would have returned the memory to the OS PageAllocatorSymbian::PageAllocatorSymbian() { __ASSERT_ALWAYS(m_pageSize == WTF::pageSize(), User::Panic(KErrorStringPageSize, m_pageSize)); RChunk chunk; TInt error = chunk.CreateDisconnectedLocal(0, 0, TInt(largeReservationSize)); __ASSERT_ALWAYS(error == KErrNone, User::Panic(KErrorStringChunkCreation, error)); m_chunk = new SymbianChunk(chunk.Handle()); // takes ownership of chunk } PageAllocatorSymbian::~PageAllocatorSymbian() { delete m_chunk; } // Reserves a region internally in the bitmap void* PageAllocatorSymbian::reserve(size_t bytes) { // Find first available region const size_t nPages = bytes / m_pageSize; const int64_t startIdx = m_map.findRunOfZeros(nPages); // Pseudo OOM if (startIdx < 0) return 0; for (size_t i = startIdx; i < startIdx + nPages ; i++) m_map.set(i); return static_cast( m_chunk->m_base + (TUint)(m_pageSize * startIdx) ); } // Reverses the effects of a reserve() call void PageAllocatorSymbian::release(void* address, size_t bytes) { const size_t startIdx = (static_cast(address) - m_chunk->m_base) / m_pageSize; const size_t nPages = bytes / m_pageSize; for (size_t i = startIdx; i < startIdx + nPages ; i++) m_map.clear(i); } // Actually commit memory from the OS, after a previous call to reserve() bool PageAllocatorSymbian::commit(void* address, size_t bytes) { // sanity check that bits were previously set const size_t idx = (static_cast(address) - m_chunk->m_base) / m_pageSize; const size_t nPages = bytes / m_pageSize; __ASSERT_ALWAYS(m_map.get(idx), User::Panic(KErrorStringInternalConsistency, idx)); __ASSERT_ALWAYS(m_map.get(idx+nPages-1), User::Panic(KErrorStringInternalConsistency, idx+nPages-1)); TInt error = m_chunk->Commit(static_cast(address) - m_chunk->m_base, bytes); return (error == KErrNone); } // Inverse operation of commit(), a release() should follow later bool PageAllocatorSymbian::decommit(void* address, size_t bytes) { TInt error = m_chunk->Decommit(static_cast(address) - m_chunk->m_base, bytes); return (error == KErrNone); } bool PageAllocatorSymbian::contains(const void* address) const { return m_chunk->contains(address); } } // namespace WTF