2 * Copyright (C) 2009 Apple Inc. All rights reserved.
4 * Redistribution and use in source and binary forms, with or without
5 * modification, are permitted provided that the following conditions
7 * 1. Redistributions of source code must retain the above copyright
8 * notice, this list of conditions and the following disclaimer.
9 * 2. Redistributions in binary form must reproduce the above copyright
10 * notice, this list of conditions and the following disclaimer in the
11 * documentation and/or other materials provided with the distribution.
13 * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY
14 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
15 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
16 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR
17 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
18 * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
19 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
20 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
21 * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
22 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
23 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28 #include "ExecutableAllocator.h"
30 #if ENABLE(EXECUTABLE_ALLOCATOR_FIXED)
34 #include "TCSpinLock.h"
37 #include <wtf/AVLTree.h>
38 #include <wtf/VMTags.h>
40 #define MMAP_FLAGS (MAP_PRIVATE | MAP_ANON | MAP_JIT)
46 #define TwoPow(n) (1ull << n)
48 class AllocationTableSizeClass
{
50 AllocationTableSizeClass(size_t size
, size_t blockSize
, unsigned log2BlockSize
)
51 : m_blockSize(blockSize
)
53 ASSERT(blockSize
== TwoPow(log2BlockSize
));
55 // Calculate the number of blocks needed to hold size.
56 size_t blockMask
= blockSize
- 1;
57 m_blockCount
= (size
+ blockMask
) >> log2BlockSize
;
59 // Align to the smallest power of two >= m_blockCount.
61 while (m_blockAlignment
< m_blockCount
)
62 m_blockAlignment
+= m_blockAlignment
;
65 size_t blockSize() const { return m_blockSize
; }
66 size_t blockCount() const { return m_blockCount
; }
67 size_t blockAlignment() const { return m_blockAlignment
; }
71 return m_blockSize
* m_blockCount
;
77 size_t m_blockAlignment
;
80 template<unsigned log2Entries
>
81 class AllocationTableLeaf
{
82 typedef uint64_t BitField
;
85 static const unsigned log2SubregionSize
= 12; // 2^12 == pagesize
86 static const unsigned log2RegionSize
= log2SubregionSize
+ log2Entries
;
88 static const size_t subregionSize
= TwoPow(log2SubregionSize
);
89 static const size_t regionSize
= TwoPow(log2RegionSize
);
90 static const unsigned entries
= TwoPow(log2Entries
);
91 COMPILE_ASSERT(entries
<= (sizeof(BitField
) * 8), AllocationTableLeaf_entries_fit_in_BitField
);
98 ~AllocationTableLeaf()
103 size_t allocate(AllocationTableSizeClass
& sizeClass
)
105 ASSERT(sizeClass
.blockSize() == subregionSize
);
108 size_t alignment
= sizeClass
.blockAlignment();
109 size_t count
= sizeClass
.blockCount();
110 // Use this mask to check for spans of free blocks.
111 BitField mask
= ((1ull << count
) - 1) << (alignment
- count
);
113 // Step in units of alignment size.
114 for (unsigned i
= 0; i
< entries
; i
+= alignment
) {
115 if (!(m_allocated
& mask
)) {
117 return (i
+ (alignment
- count
)) << log2SubregionSize
;
124 void free(size_t location
, AllocationTableSizeClass
& sizeClass
)
126 ASSERT(sizeClass
.blockSize() == subregionSize
);
128 size_t entry
= location
>> log2SubregionSize
;
129 size_t count
= sizeClass
.blockCount();
130 BitField mask
= ((1ull << count
) - 1) << entry
;
132 ASSERT((m_allocated
& mask
) == mask
);
133 m_allocated
&= ~mask
;
143 return !~m_allocated
;
151 static AllocationTableSizeClass
classForSize(size_t size
)
153 return AllocationTableSizeClass(size
, subregionSize
, log2SubregionSize
);
157 void dump(size_t parentOffset
= 0, unsigned indent
= 0)
159 for (unsigned i
= 0; i
< indent
; ++i
)
160 fprintf(stderr
, " ");
161 fprintf(stderr
, "%08x: [%016llx]\n", (int)parentOffset
, m_allocated
);
166 BitField m_allocated
;
170 template<class NextLevel
>
171 class LazyAllocationTable
{
173 static const unsigned log2RegionSize
= NextLevel::log2RegionSize
;
174 static const unsigned entries
= NextLevel::entries
;
176 LazyAllocationTable()
181 ~LazyAllocationTable()
186 size_t allocate(AllocationTableSizeClass
& sizeClass
)
189 m_ptr
= new NextLevel();
190 return m_ptr
->allocate(sizeClass
);
193 void free(size_t location
, AllocationTableSizeClass
& sizeClass
)
196 m_ptr
->free(location
, sizeClass
);
197 if (m_ptr
->isEmpty()) {
210 return m_ptr
&& m_ptr
->isFull();
215 return NextLevel::size();
219 void dump(size_t parentOffset
= 0, unsigned indent
= 0)
222 m_ptr
->dump(parentOffset
, indent
);
226 static AllocationTableSizeClass
classForSize(size_t size
)
228 return NextLevel::classForSize(size
);
235 template<class NextLevel
, unsigned log2Entries
>
236 class AllocationTableDirectory
{
237 typedef uint64_t BitField
;
240 static const unsigned log2SubregionSize
= NextLevel::log2RegionSize
;
241 static const unsigned log2RegionSize
= log2SubregionSize
+ log2Entries
;
243 static const size_t subregionSize
= TwoPow(log2SubregionSize
);
244 static const size_t regionSize
= TwoPow(log2RegionSize
);
245 static const unsigned entries
= TwoPow(log2Entries
);
246 COMPILE_ASSERT(entries
<= (sizeof(BitField
) * 8), AllocationTableDirectory_entries_fit_in_BitField
);
248 AllocationTableDirectory()
250 , m_hasSuballocation(0)
254 ~AllocationTableDirectory()
259 size_t allocate(AllocationTableSizeClass
& sizeClass
)
261 ASSERT(sizeClass
.blockSize() <= subregionSize
);
264 if (sizeClass
.blockSize() < subregionSize
) {
266 for (unsigned i
= 0; i
< entries
; ++i
, bit
+= bit
) {
269 size_t location
= m_suballocations
[i
].allocate(sizeClass
);
270 if (location
!= notFound
) {
271 // If this didn't already have a subregion, it does now!
272 m_hasSuballocation
|= bit
;
273 // Mirror the suballocation's full bit.
274 if (m_suballocations
[i
].isFull())
276 return (i
* subregionSize
) + location
;
282 // A block is allocated if either it is fully allocated or contains suballocations.
283 BitField allocated
= m_full
| m_hasSuballocation
;
285 size_t alignment
= sizeClass
.blockAlignment();
286 size_t count
= sizeClass
.blockCount();
287 // Use this mask to check for spans of free blocks.
288 BitField mask
= ((1ull << count
) - 1) << (alignment
- count
);
290 // Step in units of alignment size.
291 for (unsigned i
= 0; i
< entries
; i
+= alignment
) {
292 if (!(allocated
& mask
)) {
294 return (i
+ (alignment
- count
)) << log2SubregionSize
;
301 void free(size_t location
, AllocationTableSizeClass
& sizeClass
)
303 ASSERT(sizeClass
.blockSize() <= subregionSize
);
305 size_t entry
= location
>> log2SubregionSize
;
307 if (sizeClass
.blockSize() < subregionSize
) {
308 BitField bit
= 1ull << entry
;
309 m_suballocations
[entry
].free(location
& (subregionSize
- 1), sizeClass
);
310 // Check if the suballocation is now empty.
311 if (m_suballocations
[entry
].isEmpty())
312 m_hasSuballocation
&= ~bit
;
313 // No need to check, it clearly isn't full any more!
316 size_t count
= sizeClass
.blockCount();
317 BitField mask
= ((1ull << count
) - 1) << entry
;
318 ASSERT((m_full
& mask
) == mask
);
319 ASSERT(!(m_hasSuballocation
& mask
));
326 return !(m_full
| m_hasSuballocation
);
339 static AllocationTableSizeClass
classForSize(size_t size
)
341 if (size
< subregionSize
) {
342 AllocationTableSizeClass sizeClass
= NextLevel::classForSize(size
);
343 if (sizeClass
.size() < NextLevel::size())
346 return AllocationTableSizeClass(size
, subregionSize
, log2SubregionSize
);
350 void dump(size_t parentOffset
= 0, unsigned indent
= 0)
352 for (unsigned i
= 0; i
< indent
; ++i
)
353 fprintf(stderr
, " ");
354 fprintf(stderr
, "%08x: [", (int)parentOffset
);
355 for (unsigned i
= 0; i
< entries
; ++i
) {
356 BitField bit
= 1ull << i
;
357 char c
= m_hasSuballocation
& bit
358 ? (m_full
& bit
? 'N' : 'n')
359 : (m_full
& bit
? 'F' : '-');
360 fprintf(stderr
, "%c", c
);
362 fprintf(stderr
, "]\n");
364 for (unsigned i
= 0; i
< entries
; ++i
) {
365 BitField bit
= 1ull << i
;
366 size_t offset
= parentOffset
| (subregionSize
* i
);
367 if (m_hasSuballocation
& bit
)
368 m_suballocations
[i
].dump(offset
, indent
+ 1);
374 NextLevel m_suballocations
[entries
];
375 // Subregions exist in one of four states:
376 // (1) empty (both bits clear)
377 // (2) fully allocated as a single allocation (m_full set)
378 // (3) partially allocated through suballocations (m_hasSuballocation set)
379 // (4) fully allocated through suballocations (both bits set)
381 BitField m_hasSuballocation
;
384 typedef AllocationTableLeaf
<6> PageTables256KB
;
385 typedef AllocationTableDirectory
<PageTables256KB
, 6> PageTables16MB
;
386 typedef AllocationTableDirectory
<LazyAllocationTable
<PageTables16MB
>, 1> PageTables32MB
;
387 typedef AllocationTableDirectory
<LazyAllocationTable
<PageTables16MB
>, 6> PageTables1GB
;
390 typedef PageTables16MB FixedVMPoolPageTables
;
391 #elif CPU(X86_64) && !OS(LINUX)
392 typedef PageTables1GB FixedVMPoolPageTables
;
394 typedef PageTables32MB FixedVMPoolPageTables
;
397 class FixedVMPoolAllocator
400 FixedVMPoolAllocator()
402 m_base
= mmap(0, FixedVMPoolPageTables::size(), INITIAL_PROTECTION_FLAGS
, MMAP_FLAGS
, VM_TAG_FOR_EXECUTABLEALLOCATOR_MEMORY
, 0);
404 if (m_base
== MAP_FAILED
) {
405 #if ENABLE(INTERPRETER)
411 // For simplicity, we keep all memory in m_freeList in a 'released' state.
412 // This means that we can simply reuse all memory when allocating, without
413 // worrying about it's previous state, and also makes coalescing m_freeList
414 // simpler since we need not worry about the possibility of coalescing released
415 // chunks with non-released ones.
416 release(m_base
, FixedVMPoolPageTables::size());
420 void* alloc(size_t size
)
423 AllocationTableSizeClass sizeClass
= classForSize(size
);
424 ASSERT(sizeClass
.size());
425 if (sizeClass
.size() >= FixedVMPoolPageTables::size())
428 if (m_pages
.isFull())
430 size_t offset
= m_pages
.allocate(sizeClass
);
431 if (offset
== notFound
)
434 void* result
= offsetToPointer(offset
);
439 void free(void* pointer
, size_t size
)
441 release(pointer
, size
);
444 AllocationTableSizeClass sizeClass
= classForSize(size
);
445 ASSERT(sizeClass
.size());
446 ASSERT(sizeClass
.size() < FixedVMPoolPageTables::size());
448 m_pages
.free(pointerToOffset(pointer
), sizeClass
);
457 // Use madvise as apropriate to prevent freed pages from being spilled,
458 // and to attempt to ensure that used memory is reported correctly.
459 #if HAVE(MADV_FREE_REUSE)
460 void release(void* position
, size_t size
)
462 while (madvise(position
, size
, MADV_FREE_REUSABLE
) == -1 && errno
== EAGAIN
) { }
465 void reuse(void* position
, size_t size
)
467 while (madvise(position
, size
, MADV_FREE_REUSE
) == -1 && errno
== EAGAIN
) { }
469 #elif HAVE(MADV_FREE)
470 void release(void* position
, size_t size
)
472 while (madvise(position
, size
, MADV_FREE
) == -1 && errno
== EAGAIN
) { }
475 void reuse(void*, size_t) {}
476 #elif HAVE(MADV_DONTNEED)
477 void release(void* position
, size_t size
)
479 while (madvise(position
, size
, MADV_DONTNEED
) == -1 && errno
== EAGAIN
) { }
482 void reuse(void*, size_t) {}
484 void release(void*, size_t) {}
485 void reuse(void*, size_t) {}
488 AllocationTableSizeClass
classForSize(size_t size
)
490 return FixedVMPoolPageTables::classForSize(size
);
493 void* offsetToPointer(size_t offset
)
495 return reinterpret_cast<void*>(reinterpret_cast<intptr_t>(m_base
) + offset
);
498 size_t pointerToOffset(void* pointer
)
500 return reinterpret_cast<intptr_t>(pointer
) - reinterpret_cast<intptr_t>(m_base
);
504 FixedVMPoolPageTables m_pages
;
507 void ExecutableAllocator::intializePageSize()
509 ExecutableAllocator::pageSize
= getpagesize();
512 static FixedVMPoolAllocator
* allocator
= 0;
513 static size_t allocatedCount
= 0;
514 static SpinLock spinlock
= SPINLOCK_INITIALIZER
;
516 bool ExecutableAllocator::isValid() const
518 SpinLockHolder
lock_holder(&spinlock
);
520 allocator
= new FixedVMPoolAllocator();
521 return allocator
->isValid();
524 ExecutablePool::Allocation
ExecutablePool::systemAlloc(size_t size
)
526 SpinLockHolder
lock_holder(&spinlock
);
529 allocator
= new FixedVMPoolAllocator();
530 ExecutablePool::Allocation alloc
= {reinterpret_cast<char*>(allocator
->alloc(size
)), size
};
531 allocatedCount
+= size
;
535 void ExecutablePool::systemRelease(const ExecutablePool::Allocation
& allocation
)
537 SpinLockHolder
lock_holder(&spinlock
);
540 allocator
->free(allocation
.pages
, allocation
.size
);
541 allocatedCount
-= allocation
.size
;
544 bool ExecutablePool::underMemoryPressure()
546 // Technically we should take the spin lock here, but we don't
547 // care if we get stale data. This is only really a heuristic
549 return allocatedCount
> (FixedVMPoolPageTables::size() / 2);
554 #endif // HAVE(ASSEMBLER)