1 | /* |
2 | * Copyright (C) 2011-2019 Apple Inc. All rights reserved. |
3 | * |
4 | * Redistribution and use in source and binary forms, with or without |
5 | * modification, are permitted provided that the following conditions |
6 | * are met: |
7 | * 1. Redistributions of source code must retain the above copyright |
8 | * notice, this list of conditions and the following disclaimer. |
9 | * 2. Redistributions in binary form must reproduce the above copyright |
10 | * notice, this list of conditions and the following disclaimer in the |
11 | * documentation and/or other materials provided with the distribution. |
12 | * |
13 | * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY |
14 | * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
15 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
16 | * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR |
17 | * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
18 | * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
19 | * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
20 | * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
21 | * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
22 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
23 | * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
24 | */ |
25 | |
26 | #pragma once |
27 | |
28 | #include "BytecodeStructs.h" |
29 | #include "CodeBlock.h" |
30 | #include "CodeSpecializationKind.h" |
31 | #include "DirectArguments.h" |
32 | #include "ExceptionHelpers.h" |
33 | #include "FunctionCodeBlock.h" |
34 | #include "JSImmutableButterfly.h" |
35 | #include "ScopedArguments.h" |
36 | #include "SlowPathReturnType.h" |
37 | #include "StackAlignment.h" |
38 | #include "VMInlines.h" |
39 | #include <wtf/StdLibExtras.h> |
40 | |
41 | namespace JSC { |
42 | |
43 | // The purpose of this namespace is to include slow paths that are shared |
44 | // between the interpreter and baseline JIT. They are written to be agnostic |
45 | // with respect to the slow-path calling convention, but they do rely on the |
46 | // JS code being executed more-or-less directly from bytecode (so the call |
47 | // frame layout is unmodified, making it potentially awkward to use these |
48 | // from any optimizing JIT, like the DFG). |
49 | |
50 | namespace CommonSlowPaths { |
51 | |
52 | ALWAYS_INLINE int (int argumentCountIncludingThis) |
53 | { |
54 | int frameSize = argumentCountIncludingThis + CallFrame::headerSizeInRegisters; |
55 | int alignedFrameSize = WTF::roundUpToMultipleOf(stackAlignmentRegisters(), frameSize); |
56 | return alignedFrameSize - frameSize; |
57 | } |
58 | |
59 | ALWAYS_INLINE int numberOfStackPaddingSlots(CodeBlock* codeBlock, int argumentCountIncludingThis) |
60 | { |
61 | if (argumentCountIncludingThis >= codeBlock->numParameters()) |
62 | return 0; |
63 | int alignedFrameSize = WTF::roundUpToMultipleOf(stackAlignmentRegisters(), argumentCountIncludingThis + CallFrame::headerSizeInRegisters); |
64 | int alignedFrameSizeForParameters = WTF::roundUpToMultipleOf(stackAlignmentRegisters(), codeBlock->numParameters() + CallFrame::headerSizeInRegisters); |
65 | return alignedFrameSizeForParameters - alignedFrameSize; |
66 | } |
67 | |
68 | ALWAYS_INLINE int (CodeBlock* codeBlock, int argumentCountIncludingThis) |
69 | { |
70 | if (argumentCountIncludingThis >= codeBlock->numParameters()) |
71 | return 0; |
72 | return numberOfStackPaddingSlots(codeBlock, argumentCountIncludingThis) + numberOfExtraSlots(argumentCountIncludingThis); |
73 | } |
74 | |
75 | ALWAYS_INLINE CodeBlock* codeBlockFromCallFrameCallee(ExecState* exec, CodeSpecializationKind kind) |
76 | { |
77 | JSFunction* callee = jsCast<JSFunction*>(exec->jsCallee()); |
78 | ASSERT(!callee->isHostFunction()); |
79 | return callee->jsExecutable()->codeBlockFor(kind); |
80 | } |
81 | |
82 | ALWAYS_INLINE int arityCheckFor(ExecState* exec, VM& vm, CodeSpecializationKind kind) |
83 | { |
84 | CodeBlock* newCodeBlock = codeBlockFromCallFrameCallee(exec, kind); |
85 | ASSERT(exec->argumentCountIncludingThis() < static_cast<unsigned>(newCodeBlock->numParameters())); |
86 | int padding = numberOfStackPaddingSlotsWithExtraSlots(newCodeBlock, exec->argumentCountIncludingThis()); |
87 | |
88 | Register* newStack = exec->registers() - WTF::roundUpToMultipleOf(stackAlignmentRegisters(), padding); |
89 | |
90 | if (UNLIKELY(!vm.ensureStackCapacityFor(newStack))) |
91 | return -1; |
92 | return padding; |
93 | } |
94 | |
95 | inline bool opInByVal(ExecState* exec, JSValue baseVal, JSValue propName, ArrayProfile* arrayProfile = nullptr) |
96 | { |
97 | VM& vm = exec->vm(); |
98 | auto scope = DECLARE_THROW_SCOPE(vm); |
99 | if (!baseVal.isObject()) { |
100 | throwException(exec, scope, createInvalidInParameterError(exec, baseVal)); |
101 | return false; |
102 | } |
103 | |
104 | JSObject* baseObj = asObject(baseVal); |
105 | if (arrayProfile) |
106 | arrayProfile->observeStructure(baseObj->structure(vm)); |
107 | |
108 | uint32_t i; |
109 | if (propName.getUInt32(i)) { |
110 | if (arrayProfile) |
111 | arrayProfile->observeIndexedRead(vm, baseObj, i); |
112 | RELEASE_AND_RETURN(scope, baseObj->hasProperty(exec, i)); |
113 | } |
114 | |
115 | auto property = propName.toPropertyKey(exec); |
116 | RETURN_IF_EXCEPTION(scope, false); |
117 | RELEASE_AND_RETURN(scope, baseObj->hasProperty(exec, property)); |
118 | } |
119 | |
120 | inline void tryCachePutToScopeGlobal( |
121 | ExecState* exec, CodeBlock* codeBlock, OpPutToScope& bytecode, JSObject* scope, |
122 | PutPropertySlot& slot, const Identifier& ident) |
123 | { |
124 | // Covers implicit globals. Since they don't exist until they first execute, we didn't know how to cache them at compile time. |
125 | auto& metadata = bytecode.metadata(exec); |
126 | ResolveType resolveType = metadata.m_getPutInfo.resolveType(); |
127 | |
128 | switch (resolveType) { |
129 | case UnresolvedProperty: |
130 | case UnresolvedPropertyWithVarInjectionChecks: { |
131 | if (scope->isGlobalObject()) { |
132 | ResolveType newResolveType = needsVarInjectionChecks(resolveType) ? GlobalPropertyWithVarInjectionChecks : GlobalProperty; |
133 | resolveType = newResolveType; // Allow below caching mechanism to kick in. |
134 | ConcurrentJSLocker locker(codeBlock->m_lock); |
135 | metadata.m_getPutInfo = GetPutInfo(metadata.m_getPutInfo.resolveMode(), newResolveType, metadata.m_getPutInfo.initializationMode()); |
136 | break; |
137 | } |
138 | FALLTHROUGH; |
139 | } |
140 | case GlobalProperty: |
141 | case GlobalPropertyWithVarInjectionChecks: { |
142 | // Global Lexical Binding Epoch is changed. Update op_get_from_scope from GlobalProperty to GlobalLexicalVar. |
143 | if (scope->isGlobalLexicalEnvironment()) { |
144 | JSGlobalLexicalEnvironment* globalLexicalEnvironment = jsCast<JSGlobalLexicalEnvironment*>(scope); |
145 | ResolveType newResolveType = needsVarInjectionChecks(resolveType) ? GlobalLexicalVarWithVarInjectionChecks : GlobalLexicalVar; |
146 | metadata.m_getPutInfo = GetPutInfo(metadata.m_getPutInfo.resolveMode(), newResolveType, metadata.m_getPutInfo.initializationMode()); |
147 | SymbolTableEntry entry = globalLexicalEnvironment->symbolTable()->get(ident.impl()); |
148 | ASSERT(!entry.isNull()); |
149 | ConcurrentJSLocker locker(codeBlock->m_lock); |
150 | metadata.m_watchpointSet = entry.watchpointSet(); |
151 | metadata.m_operand = reinterpret_cast<uintptr_t>(globalLexicalEnvironment->variableAt(entry.scopeOffset()).slot()); |
152 | return; |
153 | } |
154 | break; |
155 | } |
156 | default: |
157 | return; |
158 | } |
159 | |
160 | if (resolveType == GlobalProperty || resolveType == GlobalPropertyWithVarInjectionChecks) { |
161 | VM& vm = exec->vm(); |
162 | JSGlobalObject* globalObject = codeBlock->globalObject(); |
163 | ASSERT(globalObject == scope || globalObject->varInjectionWatchpoint()->hasBeenInvalidated()); |
164 | if (!slot.isCacheablePut() |
165 | || slot.base() != scope |
166 | || scope != globalObject |
167 | || !scope->structure(vm)->propertyAccessesAreCacheable()) |
168 | return; |
169 | |
170 | if (slot.type() == PutPropertySlot::NewProperty) { |
171 | // Don't cache if we've done a transition. We want to detect the first replace so that we |
172 | // can invalidate the watchpoint. |
173 | return; |
174 | } |
175 | |
176 | scope->structure(vm)->didCachePropertyReplacement(vm, slot.cachedOffset()); |
177 | |
178 | ConcurrentJSLocker locker(codeBlock->m_lock); |
179 | metadata.m_structure.set(vm, codeBlock, scope->structure(vm)); |
180 | metadata.m_operand = slot.cachedOffset(); |
181 | } |
182 | } |
183 | |
184 | inline void tryCacheGetFromScopeGlobal( |
185 | ExecState* exec, VM& vm, OpGetFromScope& bytecode, JSObject* scope, PropertySlot& slot, const Identifier& ident) |
186 | { |
187 | auto& metadata = bytecode.metadata(exec); |
188 | ResolveType resolveType = metadata.m_getPutInfo.resolveType(); |
189 | |
190 | switch (resolveType) { |
191 | case UnresolvedProperty: |
192 | case UnresolvedPropertyWithVarInjectionChecks: { |
193 | if (scope->isGlobalObject()) { |
194 | ResolveType newResolveType = needsVarInjectionChecks(resolveType) ? GlobalPropertyWithVarInjectionChecks : GlobalProperty; |
195 | resolveType = newResolveType; // Allow below caching mechanism to kick in. |
196 | ConcurrentJSLocker locker(exec->codeBlock()->m_lock); |
197 | metadata.m_getPutInfo = GetPutInfo(metadata.m_getPutInfo.resolveMode(), newResolveType, metadata.m_getPutInfo.initializationMode()); |
198 | break; |
199 | } |
200 | FALLTHROUGH; |
201 | } |
202 | case GlobalProperty: |
203 | case GlobalPropertyWithVarInjectionChecks: { |
204 | // Global Lexical Binding Epoch is changed. Update op_get_from_scope from GlobalProperty to GlobalLexicalVar. |
205 | if (scope->isGlobalLexicalEnvironment()) { |
206 | JSGlobalLexicalEnvironment* globalLexicalEnvironment = jsCast<JSGlobalLexicalEnvironment*>(scope); |
207 | ResolveType newResolveType = needsVarInjectionChecks(resolveType) ? GlobalLexicalVarWithVarInjectionChecks : GlobalLexicalVar; |
208 | SymbolTableEntry entry = globalLexicalEnvironment->symbolTable()->get(ident.impl()); |
209 | ASSERT(!entry.isNull()); |
210 | ConcurrentJSLocker locker(exec->codeBlock()->m_lock); |
211 | metadata.m_getPutInfo = GetPutInfo(metadata.m_getPutInfo.resolveMode(), newResolveType, metadata.m_getPutInfo.initializationMode()); |
212 | metadata.m_watchpointSet = entry.watchpointSet(); |
213 | metadata.m_operand = reinterpret_cast<uintptr_t>(globalLexicalEnvironment->variableAt(entry.scopeOffset()).slot()); |
214 | return; |
215 | } |
216 | break; |
217 | } |
218 | default: |
219 | return; |
220 | } |
221 | |
222 | // Covers implicit globals. Since they don't exist until they first execute, we didn't know how to cache them at compile time. |
223 | if (resolveType == GlobalProperty || resolveType == GlobalPropertyWithVarInjectionChecks) { |
224 | CodeBlock* codeBlock = exec->codeBlock(); |
225 | JSGlobalObject* globalObject = codeBlock->globalObject(); |
226 | ASSERT(scope == globalObject || globalObject->varInjectionWatchpoint()->hasBeenInvalidated()); |
227 | if (slot.isCacheableValue() && slot.slotBase() == scope && scope == globalObject && scope->structure(vm)->propertyAccessesAreCacheable()) { |
228 | Structure* structure = scope->structure(vm); |
229 | { |
230 | ConcurrentJSLocker locker(codeBlock->m_lock); |
231 | metadata.m_structure.set(vm, codeBlock, structure); |
232 | metadata.m_operand = slot.cachedOffset(); |
233 | } |
234 | structure->startWatchingPropertyForReplacements(vm, slot.cachedOffset()); |
235 | } |
236 | } |
237 | } |
238 | |
239 | inline bool canAccessArgumentIndexQuickly(JSObject& object, uint32_t index) |
240 | { |
241 | switch (object.type()) { |
242 | case DirectArgumentsType: { |
243 | DirectArguments* directArguments = jsCast<DirectArguments*>(&object); |
244 | if (directArguments->isMappedArgumentInDFG(index)) |
245 | return true; |
246 | break; |
247 | } |
248 | case ScopedArgumentsType: { |
249 | ScopedArguments* scopedArguments = jsCast<ScopedArguments*>(&object); |
250 | if (scopedArguments->isMappedArgumentInDFG(index)) |
251 | return true; |
252 | break; |
253 | } |
254 | default: |
255 | break; |
256 | } |
257 | return false; |
258 | } |
259 | |
260 | static ALWAYS_INLINE void putDirectWithReify(VM& vm, ExecState* exec, JSObject* baseObject, PropertyName propertyName, JSValue value, PutPropertySlot& slot, Structure** result = nullptr) |
261 | { |
262 | auto scope = DECLARE_THROW_SCOPE(vm); |
263 | if (baseObject->inherits<JSFunction>(vm)) { |
264 | jsCast<JSFunction*>(baseObject)->reifyLazyPropertyIfNeeded(vm, exec, propertyName); |
265 | RETURN_IF_EXCEPTION(scope, void()); |
266 | } |
267 | if (result) |
268 | *result = baseObject->structure(vm); |
269 | scope.release(); |
270 | baseObject->putDirect(vm, propertyName, value, slot); |
271 | } |
272 | |
273 | static ALWAYS_INLINE void putDirectAccessorWithReify(VM& vm, ExecState* exec, JSObject* baseObject, PropertyName propertyName, GetterSetter* accessor, unsigned attribute) |
274 | { |
275 | auto scope = DECLARE_THROW_SCOPE(vm); |
276 | if (baseObject->inherits<JSFunction>(vm)) { |
277 | jsCast<JSFunction*>(baseObject)->reifyLazyPropertyIfNeeded(vm, exec, propertyName); |
278 | RETURN_IF_EXCEPTION(scope, void()); |
279 | } |
280 | scope.release(); |
281 | baseObject->putDirectAccessor(exec, propertyName, accessor, attribute); |
282 | } |
283 | |
284 | inline JSArray* allocateNewArrayBuffer(VM& vm, Structure* structure, JSImmutableButterfly* immutableButterfly) |
285 | { |
286 | JSGlobalObject* globalObject = structure->globalObject(); |
287 | Structure* originalStructure = globalObject->originalArrayStructureForIndexingType(immutableButterfly->indexingMode()); |
288 | ASSERT(originalStructure->indexingMode() == immutableButterfly->indexingMode()); |
289 | ASSERT(isCopyOnWrite(immutableButterfly->indexingMode())); |
290 | ASSERT(!structure->outOfLineCapacity()); |
291 | |
292 | JSArray* result = JSArray::createWithButterfly(vm, nullptr, originalStructure, immutableButterfly->toButterfly()); |
293 | // FIXME: This works but it's slow. If we cared enough about the perf when having a bad time then we could fix it. |
294 | if (UNLIKELY(originalStructure != structure)) { |
295 | ASSERT(hasSlowPutArrayStorage(structure->indexingMode())); |
296 | ASSERT(globalObject->isHavingABadTime()); |
297 | |
298 | result->switchToSlowPutArrayStorage(vm); |
299 | ASSERT(result->butterfly() != immutableButterfly->toButterfly()); |
300 | ASSERT(!result->butterfly()->arrayStorage()->m_sparseMap.get()); |
301 | ASSERT(result->structureID() == structure->id()); |
302 | } |
303 | |
304 | return result; |
305 | } |
306 | |
307 | } // namespace CommonSlowPaths |
308 | |
309 | class ExecState; |
310 | struct Instruction; |
311 | |
312 | #define SLOW_PATH |
313 | |
314 | #define SLOW_PATH_DECL(name) \ |
315 | extern "C" SlowPathReturnType SLOW_PATH name(ExecState* exec, const Instruction* pc) |
316 | |
317 | #define SLOW_PATH_HIDDEN_DECL(name) \ |
318 | SLOW_PATH_DECL(name) WTF_INTERNAL |
319 | |
320 | SLOW_PATH_HIDDEN_DECL(slow_path_call_arityCheck); |
321 | SLOW_PATH_HIDDEN_DECL(slow_path_construct_arityCheck); |
322 | SLOW_PATH_HIDDEN_DECL(slow_path_create_direct_arguments); |
323 | SLOW_PATH_HIDDEN_DECL(slow_path_create_scoped_arguments); |
324 | SLOW_PATH_HIDDEN_DECL(slow_path_create_cloned_arguments); |
325 | SLOW_PATH_HIDDEN_DECL(slow_path_create_this); |
326 | SLOW_PATH_HIDDEN_DECL(slow_path_enter); |
327 | SLOW_PATH_HIDDEN_DECL(slow_path_get_callee); |
328 | SLOW_PATH_HIDDEN_DECL(slow_path_to_this); |
329 | SLOW_PATH_HIDDEN_DECL(slow_path_throw_tdz_error); |
330 | SLOW_PATH_HIDDEN_DECL(slow_path_check_tdz); |
331 | SLOW_PATH_HIDDEN_DECL(slow_path_throw_strict_mode_readonly_property_write_error); |
332 | SLOW_PATH_HIDDEN_DECL(slow_path_not); |
333 | SLOW_PATH_HIDDEN_DECL(slow_path_eq); |
334 | SLOW_PATH_HIDDEN_DECL(slow_path_neq); |
335 | SLOW_PATH_HIDDEN_DECL(slow_path_stricteq); |
336 | SLOW_PATH_HIDDEN_DECL(slow_path_nstricteq); |
337 | SLOW_PATH_HIDDEN_DECL(slow_path_less); |
338 | SLOW_PATH_HIDDEN_DECL(slow_path_lesseq); |
339 | SLOW_PATH_HIDDEN_DECL(slow_path_greater); |
340 | SLOW_PATH_HIDDEN_DECL(slow_path_greatereq); |
341 | SLOW_PATH_HIDDEN_DECL(slow_path_inc); |
342 | SLOW_PATH_HIDDEN_DECL(slow_path_dec); |
343 | SLOW_PATH_HIDDEN_DECL(slow_path_to_number); |
344 | SLOW_PATH_HIDDEN_DECL(slow_path_to_string); |
345 | SLOW_PATH_HIDDEN_DECL(slow_path_to_object); |
346 | SLOW_PATH_HIDDEN_DECL(slow_path_negate); |
347 | SLOW_PATH_HIDDEN_DECL(slow_path_add); |
348 | SLOW_PATH_HIDDEN_DECL(slow_path_mul); |
349 | SLOW_PATH_HIDDEN_DECL(slow_path_sub); |
350 | SLOW_PATH_HIDDEN_DECL(slow_path_div); |
351 | SLOW_PATH_HIDDEN_DECL(slow_path_mod); |
352 | SLOW_PATH_HIDDEN_DECL(slow_path_pow); |
353 | SLOW_PATH_HIDDEN_DECL(slow_path_lshift); |
354 | SLOW_PATH_HIDDEN_DECL(slow_path_rshift); |
355 | SLOW_PATH_HIDDEN_DECL(slow_path_urshift); |
356 | SLOW_PATH_HIDDEN_DECL(slow_path_unsigned); |
357 | SLOW_PATH_HIDDEN_DECL(slow_path_bitnot); |
358 | SLOW_PATH_HIDDEN_DECL(slow_path_bitand); |
359 | SLOW_PATH_HIDDEN_DECL(slow_path_bitor); |
360 | SLOW_PATH_HIDDEN_DECL(slow_path_bitxor); |
361 | SLOW_PATH_HIDDEN_DECL(slow_path_typeof); |
362 | SLOW_PATH_HIDDEN_DECL(slow_path_is_object); |
363 | SLOW_PATH_HIDDEN_DECL(slow_path_is_object_or_null); |
364 | SLOW_PATH_HIDDEN_DECL(slow_path_is_function); |
365 | SLOW_PATH_HIDDEN_DECL(slow_path_in_by_id); |
366 | SLOW_PATH_HIDDEN_DECL(slow_path_in_by_val); |
367 | SLOW_PATH_HIDDEN_DECL(slow_path_del_by_val); |
368 | SLOW_PATH_HIDDEN_DECL(slow_path_strcat); |
369 | SLOW_PATH_HIDDEN_DECL(slow_path_to_primitive); |
370 | SLOW_PATH_HIDDEN_DECL(slow_path_get_enumerable_length); |
371 | SLOW_PATH_HIDDEN_DECL(slow_path_has_generic_property); |
372 | SLOW_PATH_HIDDEN_DECL(slow_path_has_structure_property); |
373 | SLOW_PATH_HIDDEN_DECL(slow_path_has_indexed_property); |
374 | SLOW_PATH_HIDDEN_DECL(slow_path_get_direct_pname); |
375 | SLOW_PATH_HIDDEN_DECL(slow_path_get_property_enumerator); |
376 | SLOW_PATH_HIDDEN_DECL(slow_path_enumerator_structure_pname); |
377 | SLOW_PATH_HIDDEN_DECL(slow_path_enumerator_generic_pname); |
378 | SLOW_PATH_HIDDEN_DECL(slow_path_to_index_string); |
379 | SLOW_PATH_HIDDEN_DECL(slow_path_profile_type_clear_log); |
380 | SLOW_PATH_HIDDEN_DECL(slow_path_unreachable); |
381 | SLOW_PATH_HIDDEN_DECL(slow_path_create_lexical_environment); |
382 | SLOW_PATH_HIDDEN_DECL(slow_path_push_with_scope); |
383 | SLOW_PATH_HIDDEN_DECL(slow_path_resolve_scope); |
384 | SLOW_PATH_HIDDEN_DECL(slow_path_is_var_scope); |
385 | SLOW_PATH_HIDDEN_DECL(slow_path_resolve_scope_for_hoisting_func_decl_in_eval); |
386 | SLOW_PATH_HIDDEN_DECL(slow_path_create_rest); |
387 | SLOW_PATH_HIDDEN_DECL(slow_path_get_by_id_with_this); |
388 | SLOW_PATH_HIDDEN_DECL(slow_path_get_by_val_with_this); |
389 | SLOW_PATH_HIDDEN_DECL(slow_path_put_by_id_with_this); |
390 | SLOW_PATH_HIDDEN_DECL(slow_path_put_by_val_with_this); |
391 | SLOW_PATH_HIDDEN_DECL(slow_path_define_data_property); |
392 | SLOW_PATH_HIDDEN_DECL(slow_path_define_accessor_property); |
393 | SLOW_PATH_HIDDEN_DECL(slow_path_throw_static_error); |
394 | SLOW_PATH_HIDDEN_DECL(slow_path_new_array_with_spread); |
395 | SLOW_PATH_HIDDEN_DECL(slow_path_new_array_buffer); |
396 | SLOW_PATH_HIDDEN_DECL(slow_path_spread); |
397 | |
398 | using SlowPathFunction = SlowPathReturnType(SLOW_PATH *)(ExecState*, const Instruction*); |
399 | |
400 | } // namespace JSC |
401 | |