]> git.saurik.com Git - apple/javascriptcore.git/blame_incremental - heap/GCThreadSharedData.cpp
JavaScriptCore-7601.1.46.3.tar.gz
[apple/javascriptcore.git] / heap / GCThreadSharedData.cpp
... / ...
CommitLineData
1/*
2 * Copyright (C) 2009, 2011 Apple Inc. All rights reserved.
3 *
4 * Redistribution and use in source and binary forms, with or without
5 * modification, are permitted provided that the following conditions
6 * are met:
7 * 1. Redistributions of source code must retain the above copyright
8 * notice, this list of conditions and the following disclaimer.
9 * 2. Redistributions in binary form must reproduce the above copyright
10 * notice, this list of conditions and the following disclaimer in the
11 * documentation and/or other materials provided with the distribution.
12 *
13 * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY
14 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
15 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
16 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR
17 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
18 * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
19 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
20 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
21 * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
22 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
23 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
24 */
25
26#include "config.h"
27#include "GCThreadSharedData.h"
28
29#include "CopyVisitor.h"
30#include "CopyVisitorInlines.h"
31#include "GCThread.h"
32#include "MarkStack.h"
33#include "JSCInlines.h"
34#include "SlotVisitor.h"
35#include "SlotVisitorInlines.h"
36#include "VM.h"
37
38namespace JSC {
39
40#if ENABLE(PARALLEL_GC)
41void GCThreadSharedData::resetChildren()
42{
43 for (size_t i = 0; i < m_gcThreads.size(); ++i)
44 m_gcThreads[i]->slotVisitor()->reset();
45}
46
47size_t GCThreadSharedData::childVisitCount()
48{
49 unsigned long result = 0;
50 for (unsigned i = 0; i < m_gcThreads.size(); ++i)
51 result += m_gcThreads[i]->slotVisitor()->visitCount();
52 return result;
53}
54
55size_t GCThreadSharedData::childBytesVisited()
56{
57 size_t result = 0;
58 for (unsigned i = 0; i < m_gcThreads.size(); ++i)
59 result += m_gcThreads[i]->slotVisitor()->bytesVisited();
60 return result;
61}
62
63size_t GCThreadSharedData::childBytesCopied()
64{
65 size_t result = 0;
66 for (unsigned i = 0; i < m_gcThreads.size(); ++i)
67 result += m_gcThreads[i]->slotVisitor()->bytesCopied();
68 return result;
69}
70#endif
71
72GCThreadSharedData::GCThreadSharedData(VM* vm)
73 : m_vm(vm)
74 , m_copiedSpace(&vm->heap.m_storageSpace)
75 , m_shouldHashCons(false)
76 , m_sharedMarkStack()
77 , m_numberOfActiveParallelMarkers(0)
78 , m_parallelMarkersShouldExit(false)
79 , m_copyIndex(0)
80 , m_numberOfActiveGCThreads(0)
81 , m_gcThreadsShouldWait(false)
82 , m_currentPhase(NoPhase)
83{
84#if ENABLE(PARALLEL_GC)
85 // Grab the lock so the new GC threads can be properly initialized before they start running.
86 std::unique_lock<std::mutex> lock(m_phaseMutex);
87 for (unsigned i = 1; i < Options::numberOfGCMarkers(); ++i) {
88 m_numberOfActiveGCThreads++;
89 GCThread* newThread = new GCThread(*this, std::make_unique<SlotVisitor>(*this), std::make_unique<CopyVisitor>(*this));
90 ThreadIdentifier threadID = createThread(GCThread::gcThreadStartFunc, newThread, "JavaScriptCore::Marking");
91 newThread->initializeThreadID(threadID);
92 m_gcThreads.append(newThread);
93 }
94
95 // Wait for all the GCThreads to get to the right place.
96 m_activityConditionVariable.wait(lock, [this] { return !m_numberOfActiveGCThreads; });
97#endif
98}
99
100GCThreadSharedData::~GCThreadSharedData()
101{
102#if ENABLE(PARALLEL_GC)
103 // Destroy our marking threads.
104 {
105 std::lock_guard<std::mutex> markingLock(m_markingMutex);
106 std::lock_guard<std::mutex> phaseLock(m_phaseMutex);
107 ASSERT(m_currentPhase == NoPhase);
108 m_parallelMarkersShouldExit = true;
109 m_gcThreadsShouldWait = false;
110 m_currentPhase = Exit;
111 m_phaseConditionVariable.notify_all();
112 }
113 for (unsigned i = 0; i < m_gcThreads.size(); ++i) {
114 waitForThreadCompletion(m_gcThreads[i]->threadID());
115 delete m_gcThreads[i];
116 }
117#endif
118}
119
120void GCThreadSharedData::reset()
121{
122 ASSERT(m_sharedMarkStack.isEmpty());
123
124 m_weakReferenceHarvesters.removeAll();
125
126 if (m_shouldHashCons) {
127 m_vm->resetNewStringsSinceLastHashCons();
128 m_shouldHashCons = false;
129 }
130}
131
132void GCThreadSharedData::startNextPhase(GCPhase phase)
133{
134 std::lock_guard<std::mutex> lock(m_phaseMutex);
135 ASSERT(!m_gcThreadsShouldWait);
136 ASSERT(m_currentPhase == NoPhase);
137 m_gcThreadsShouldWait = true;
138 m_currentPhase = phase;
139 m_phaseConditionVariable.notify_all();
140}
141
142void GCThreadSharedData::endCurrentPhase()
143{
144 ASSERT(m_gcThreadsShouldWait);
145 std::unique_lock<std::mutex> lock(m_phaseMutex);
146 m_currentPhase = NoPhase;
147 m_gcThreadsShouldWait = false;
148 m_phaseConditionVariable.notify_all();
149 m_activityConditionVariable.wait(lock, [this] { return !m_numberOfActiveGCThreads; });
150}
151
152void GCThreadSharedData::didStartMarking()
153{
154 if (m_vm->heap.operationInProgress() == FullCollection) {
155#if ENABLE(PARALLEL_GC)
156 m_opaqueRoots.clear();
157#else
158 ASSERT(m_opaqueRoots.isEmpty());
159#endif
160}
161 std::lock_guard<std::mutex> lock(m_markingMutex);
162 m_parallelMarkersShouldExit = false;
163 startNextPhase(Mark);
164}
165
166void GCThreadSharedData::didFinishMarking()
167{
168 {
169 std::lock_guard<std::mutex> lock(m_markingMutex);
170 m_parallelMarkersShouldExit = true;
171 m_markingConditionVariable.notify_all();
172 }
173
174 ASSERT(m_currentPhase == Mark);
175 endCurrentPhase();
176}
177
178void GCThreadSharedData::didStartCopying()
179{
180 {
181 SpinLockHolder locker(&m_copyLock);
182 if (m_vm->heap.operationInProgress() == EdenCollection) {
183 // Reset the vector to be empty, but don't throw away the backing store.
184 m_blocksToCopy.shrink(0);
185 for (CopiedBlock* block = m_copiedSpace->m_newGen.fromSpace->head(); block; block = block->next())
186 m_blocksToCopy.append(block);
187 } else {
188 ASSERT(m_vm->heap.operationInProgress() == FullCollection);
189 WTF::copyToVector(m_copiedSpace->m_blockSet, m_blocksToCopy);
190 }
191 m_copyIndex = 0;
192 }
193
194 // We do this here so that we avoid a race condition where the main thread can
195 // blow through all of the copying work before the GCThreads fully wake up.
196 // The GCThreads then request a block from the CopiedSpace when the copying phase
197 // has completed, which isn't allowed.
198 for (size_t i = 0; i < m_gcThreads.size(); i++)
199 m_gcThreads[i]->copyVisitor()->startCopying();
200
201 startNextPhase(Copy);
202}
203
204void GCThreadSharedData::didFinishCopying()
205{
206 ASSERT(m_currentPhase == Copy);
207 endCurrentPhase();
208}
209
210} // namespace JSC