1/*
2 * Copyright (C) 2003-2017 Apple Inc. All rights reserved.
3 * Copyright (C) 2007 Eric Seidel <[email protected]>
4 * Copyright (C) 2009 Acision BV. All rights reserved.
5 *
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
10 *
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
15 *
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19 *
20 */
21
22#include "config.h"
23#include "MachineStackMarker.h"
24
25#include "ConservativeRoots.h"
26#include "MachineContext.h"
27#include <setjmp.h>
28#include <stdlib.h>
29#include <wtf/BitVector.h>
30#include <wtf/PageBlock.h>
31#include <wtf/StdLibExtras.h>
32
33namespace JSC {
34
35MachineThreads::MachineThreads()
36 : m_threadGroup(ThreadGroup::create())
37{
38}
39
40SUPPRESS_ASAN
41void MachineThreads::gatherFromCurrentThread(ConservativeRoots& conservativeRoots, JITStubRoutineSet& jitStubRoutines, CodeBlockSet& codeBlocks, CurrentThreadState& currentThreadState)
42{
43 if (currentThreadState.registerState) {
44 void* registersBegin = currentThreadState.registerState;
45 void* registersEnd = reinterpret_cast<void*>(roundUpToMultipleOf<sizeof(void*)>(reinterpret_cast<uintptr_t>(currentThreadState.registerState + 1)));
46 conservativeRoots.add(registersBegin, registersEnd, jitStubRoutines, codeBlocks);
47 }
48
49 conservativeRoots.add(currentThreadState.stackTop, currentThreadState.stackOrigin, jitStubRoutines, codeBlocks);
50}
51
52static inline int osRedZoneAdjustment()
53{
54 int redZoneAdjustment = 0;
55#if !OS(WINDOWS)
56#if CPU(X86_64)
57 // See http://people.freebsd.org/~obrien/amd64-elf-abi.pdf Section 3.2.2.
58 redZoneAdjustment = -128;
59#elif CPU(ARM64)
60 // See https://developer.apple.com/library/ios/documentation/Xcode/Conceptual/iPhoneOSABIReference/Articles/ARM64FunctionCallingConventions.html#//apple_ref/doc/uid/TP40013702-SW7
61 redZoneAdjustment = -128;
62#endif
63#endif // !OS(WINDOWS)
64 return redZoneAdjustment;
65}
66
67static std::pair<void*, size_t> captureStack(Thread& thread, void* stackTop)
68{
69 char* begin = reinterpret_cast_ptr<char*>(thread.stack().origin());
70 char* end = bitwise_cast<char*>(WTF::roundUpToMultipleOf<sizeof(void*)>(reinterpret_cast<uintptr_t>(stackTop)));
71 ASSERT(begin >= end);
72
73 char* endWithRedZone = end + osRedZoneAdjustment();
74 ASSERT(WTF::roundUpToMultipleOf<sizeof(void*)>(reinterpret_cast<uintptr_t>(endWithRedZone)) == reinterpret_cast<uintptr_t>(endWithRedZone));
75
76 if (endWithRedZone < thread.stack().end())
77 endWithRedZone = reinterpret_cast_ptr<char*>(thread.stack().end());
78
79 std::swap(begin, endWithRedZone);
80 return std::make_pair(begin, endWithRedZone - begin);
81}
82
83SUPPRESS_ASAN
84static void copyMemory(void* dst, const void* src, size_t size)
85{
86 size_t dstAsSize = reinterpret_cast<size_t>(dst);
87 size_t srcAsSize = reinterpret_cast<size_t>(src);
88 RELEASE_ASSERT(dstAsSize == WTF::roundUpToMultipleOf<sizeof(CPURegister)>(dstAsSize));
89 RELEASE_ASSERT(srcAsSize == WTF::roundUpToMultipleOf<sizeof(CPURegister)>(srcAsSize));
90 RELEASE_ASSERT(size == WTF::roundUpToMultipleOf<sizeof(CPURegister)>(size));
91
92 CPURegister* dstPtr = reinterpret_cast<CPURegister*>(dst);
93 const CPURegister* srcPtr = reinterpret_cast<const CPURegister*>(src);
94 size /= sizeof(CPURegister);
95 while (size--)
96 *dstPtr++ = *srcPtr++;
97}
98
99
100
101// This function must not call malloc(), free(), or any other function that might
102// acquire a lock. Since 'thread' is suspended, trying to acquire a lock
103// will deadlock if 'thread' holds that lock.
104// This function, specifically the memory copying, was causing problems with Address Sanitizer in
105// apps. Since we cannot blacklist the system memcpy we must use our own naive implementation,
106// copyMemory, for ASan to work on either instrumented or non-instrumented builds. This is not a
107// significant performance loss as tryCopyOtherThreadStack is only called as part of an O(heapsize)
108// operation. As the heap is generally much larger than the stack the performance hit is minimal.
109// See: https://bugs.webkit.org/show_bug.cgi?id=146297
110void MachineThreads::tryCopyOtherThreadStack(Thread& thread, void* buffer, size_t capacity, size_t* size)
111{
112 PlatformRegisters registers;
113 size_t registersSize = thread.getRegisters(registers);
114
115 // This is a workaround for <rdar://problem/27607384>. libdispatch recycles work
116 // queue threads without running pthread exit destructors. This can cause us to scan a
117 // thread during work queue initialization, when the stack pointer is null.
118 if (UNLIKELY(!MachineContext::stackPointer(registers))) {
119 *size = 0;
120 return;
121 }
122
123 std::pair<void*, size_t> stack = captureStack(thread, MachineContext::stackPointer(registers));
124
125 bool canCopy = *size + registersSize + stack.second <= capacity;
126
127 if (canCopy)
128 copyMemory(static_cast<char*>(buffer) + *size, &registers, registersSize);
129 *size += registersSize;
130
131 if (canCopy)
132 copyMemory(static_cast<char*>(buffer) + *size, stack.first, stack.second);
133 *size += stack.second;
134}
135
136bool MachineThreads::tryCopyOtherThreadStacks(const AbstractLocker& locker, void* buffer, size_t capacity, size_t* size, Thread& currentThreadForGC)
137{
138 // Prevent two VMs from suspending each other's threads at the same time,
139 // which can cause deadlock: <rdar://problem/20300842>.
140 static Lock mutex;
141 std::lock_guard<Lock> lock(mutex);
142
143 *size = 0;
144
145 Thread& currentThread = Thread::current();
146 const ListHashSet<Ref<Thread>>& threads = m_threadGroup->threads(locker);
147 BitVector isSuspended(threads.size());
148
149 {
150 unsigned index = 0;
151 for (const Ref<Thread>& thread : threads) {
152 if (thread.ptr() != &currentThread
153 && thread.ptr() != &currentThreadForGC) {
154 auto result = thread->suspend();
155 if (result)
156 isSuspended.set(index);
157 else {
158#if OS(DARWIN)
159 // These threads will be removed from the ThreadGroup. Thus, we do not do anything here except for reporting.
160 ASSERT(result.error() != KERN_SUCCESS);
161 WTFReportError(__FILE__, __LINE__, WTF_PRETTY_FUNCTION,
162 "JavaScript garbage collection encountered an invalid thread (err 0x%x): Thread [%d/%d: %p].",
163 result.error(), index, threads.size(), thread.ptr());
164#endif
165 }
166 }
167 ++index;
168 }
169 }
170
171 {
172 unsigned index = 0;
173 for (auto& thread : threads) {
174 if (isSuspended.get(index))
175 tryCopyOtherThreadStack(thread.get(), buffer, capacity, size);
176 ++index;
177 }
178 }
179
180 {
181 unsigned index = 0;
182 for (auto& thread : threads) {
183 if (isSuspended.get(index))
184 thread->resume();
185 ++index;
186 }
187 }
188
189 return *size <= capacity;
190}
191
192static void growBuffer(size_t size, void** buffer, size_t* capacity)
193{
194 if (*buffer)
195 fastFree(*buffer);
196
197 *capacity = WTF::roundUpToMultipleOf(WTF::pageSize(), size * 2);
198 *buffer = fastMalloc(*capacity);
199}
200
201void MachineThreads::gatherConservativeRoots(ConservativeRoots& conservativeRoots, JITStubRoutineSet& jitStubRoutines, CodeBlockSet& codeBlocks, CurrentThreadState* currentThreadState, Thread* currentThread)
202{
203 if (currentThreadState)
204 gatherFromCurrentThread(conservativeRoots, jitStubRoutines, codeBlocks, *currentThreadState);
205
206 size_t size;
207 size_t capacity = 0;
208 void* buffer = nullptr;
209 auto locker = holdLock(m_threadGroup->getLock());
210 while (!tryCopyOtherThreadStacks(locker, buffer, capacity, &size, *currentThread))
211 growBuffer(size, &buffer, &capacity);
212
213 if (!buffer)
214 return;
215
216 conservativeRoots.add(buffer, static_cast<char*>(buffer) + size, jitStubRoutines, codeBlocks);
217 fastFree(buffer);
218}
219
220NEVER_INLINE int callWithCurrentThreadState(const ScopedLambda<void(CurrentThreadState&)>& lambda)
221{
222 DECLARE_AND_COMPUTE_CURRENT_THREAD_STATE(state);
223 lambda(state);
224 return 42; // Suppress tail call optimization.
225}
226
227} // namespace JSC
228