1 | /* |
2 | * Copyright (C) 2017-2018 Apple Inc. All rights reserved. |
3 | * |
4 | * Redistribution and use in source and binary forms, with or without |
5 | * modification, are permitted provided that the following conditions |
6 | * are met: |
7 | * 1. Redistributions of source code must retain the above copyright |
8 | * notice, this list of conditions and the following disclaimer. |
9 | * 2. Redistributions in binary form must reproduce the above copyright |
10 | * notice, this list of conditions and the following disclaimer in the |
11 | * documentation and/or other materials provided with the distribution. |
12 | * |
13 | * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY |
14 | * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
15 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
16 | * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR |
17 | * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
18 | * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
19 | * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
20 | * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
21 | * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
22 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
23 | * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
24 | */ |
25 | |
26 | #include "config.h" |
27 | #include "VMInspector.h" |
28 | |
29 | #include "CodeBlock.h" |
30 | #include "CodeBlockSet.h" |
31 | #include "HeapInlines.h" |
32 | #include "HeapIterationScope.h" |
33 | #include "JSCInlines.h" |
34 | #include "MachineContext.h" |
35 | #include "MarkedSpaceInlines.h" |
36 | #include "StackVisitor.h" |
37 | #include <mutex> |
38 | #include <wtf/Expected.h> |
39 | |
40 | #if !OS(WINDOWS) |
41 | #include <unistd.h> |
42 | #endif |
43 | |
44 | namespace JSC { |
45 | |
46 | VMInspector& VMInspector::instance() |
47 | { |
48 | static VMInspector* manager; |
49 | static std::once_flag once; |
50 | std::call_once(once, [] { |
51 | manager = new VMInspector(); |
52 | }); |
53 | return *manager; |
54 | } |
55 | |
56 | void VMInspector::add(VM* vm) |
57 | { |
58 | auto locker = holdLock(m_lock); |
59 | m_list.append(vm); |
60 | } |
61 | |
62 | void VMInspector::remove(VM* vm) |
63 | { |
64 | auto locker = holdLock(m_lock); |
65 | m_list.remove(vm); |
66 | } |
67 | |
68 | auto VMInspector::lock(Seconds timeout) -> Expected<Locker, Error> |
69 | { |
70 | // This function may be called from a signal handler (e.g. via visit()). Hence, |
71 | // it should only use APIs that are safe to call from signal handlers. This is |
72 | // why we use unistd.h's sleep() instead of its alternatives. |
73 | |
74 | // We'll be doing sleep(1) between tries below. Hence, sleepPerRetry is 1. |
75 | unsigned maxRetries = (timeout < Seconds::infinity()) ? timeout.value() : UINT_MAX; |
76 | |
77 | Expected<Locker, Error> locker = Locker::tryLock(m_lock); |
78 | unsigned tryCount = 0; |
79 | while (!locker && tryCount < maxRetries) { |
80 | // We want the version of sleep from unistd.h. Cast to disambiguate. |
81 | #if !OS(WINDOWS) |
82 | (static_cast<unsigned (*)(unsigned)>(sleep))(1); |
83 | #endif |
84 | locker = Locker::tryLock(m_lock); |
85 | } |
86 | |
87 | if (!locker) |
88 | return makeUnexpected(Error::TimedOut); |
89 | return locker; |
90 | } |
91 | |
92 | #if ENABLE(JIT) |
93 | static bool ensureIsSafeToLock(Lock& lock) |
94 | { |
95 | unsigned maxRetries = 2; |
96 | unsigned tryCount = 0; |
97 | while (tryCount <= maxRetries) { |
98 | bool success = lock.tryLock(); |
99 | if (success) { |
100 | lock.unlock(); |
101 | return true; |
102 | } |
103 | tryCount++; |
104 | } |
105 | return false; |
106 | }; |
107 | #endif // ENABLE(JIT) |
108 | |
109 | auto VMInspector::isValidExecutableMemory(const VMInspector::Locker&, void* machinePC) -> Expected<bool, Error> |
110 | { |
111 | #if ENABLE(JIT) |
112 | bool found = false; |
113 | bool hasTimeout = false; |
114 | iterate([&] (VM&) -> FunctorStatus { |
115 | auto& allocator = ExecutableAllocator::singleton(); |
116 | auto& lock = allocator.getLock(); |
117 | |
118 | bool isSafeToLock = ensureIsSafeToLock(lock); |
119 | if (!isSafeToLock) { |
120 | hasTimeout = true; |
121 | return FunctorStatus::Continue; // Skip this VM. |
122 | } |
123 | |
124 | LockHolder executableAllocatorLocker(lock); |
125 | if (allocator.isValidExecutableMemory(executableAllocatorLocker, machinePC)) { |
126 | found = true; |
127 | return FunctorStatus::Done; |
128 | } |
129 | return FunctorStatus::Continue; |
130 | }); |
131 | |
132 | if (!found && hasTimeout) |
133 | return makeUnexpected(Error::TimedOut); |
134 | return found; |
135 | #else |
136 | UNUSED_PARAM(machinePC); |
137 | return false; |
138 | #endif |
139 | } |
140 | |
141 | auto VMInspector::codeBlockForMachinePC(const VMInspector::Locker&, void* machinePC) -> Expected<CodeBlock*, Error> |
142 | { |
143 | #if ENABLE(JIT) |
144 | CodeBlock* codeBlock = nullptr; |
145 | bool hasTimeout = false; |
146 | iterate([&] (VM& vm) { |
147 | if (!vm.currentThreadIsHoldingAPILock()) |
148 | return FunctorStatus::Continue; |
149 | |
150 | // It is safe to call Heap::forEachCodeBlockIgnoringJITPlans here because: |
151 | // 1. CodeBlocks are added to the CodeBlockSet from the main thread before |
152 | // they are handed to the JIT plans. Those codeBlocks will have a null jitCode, |
153 | // but we check for that in our lambda functor. |
154 | // 2. We will acquire the CodeBlockSet lock before iterating. |
155 | // This ensures that a CodeBlock won't be GCed while we're iterating. |
156 | // 3. We do a tryLock on the CodeBlockSet's lock first to ensure that it is |
157 | // safe for the current thread to lock it before calling |
158 | // Heap::forEachCodeBlockIgnoringJITPlans(). Hence, there's no risk of |
159 | // re-entering the lock and deadlocking on it. |
160 | |
161 | auto& codeBlockSetLock = vm.heap.codeBlockSet().getLock(); |
162 | bool isSafeToLock = ensureIsSafeToLock(codeBlockSetLock); |
163 | if (!isSafeToLock) { |
164 | hasTimeout = true; |
165 | return FunctorStatus::Continue; // Skip this VM. |
166 | } |
167 | |
168 | auto locker = holdLock(codeBlockSetLock); |
169 | vm.heap.forEachCodeBlockIgnoringJITPlans(locker, [&] (CodeBlock* cb) { |
170 | JITCode* jitCode = cb->jitCode().get(); |
171 | if (!jitCode) { |
172 | // If the codeBlock is a replacement codeBlock which is in the process of being |
173 | // compiled, its jitCode will be null, and we can disregard it as a match for |
174 | // the machinePC we're searching for. |
175 | return; |
176 | } |
177 | |
178 | if (!JITCode::isJIT(jitCode->jitType())) |
179 | return; |
180 | |
181 | if (jitCode->contains(machinePC)) { |
182 | codeBlock = cb; |
183 | return; |
184 | } |
185 | }); |
186 | if (codeBlock) |
187 | return FunctorStatus::Done; |
188 | return FunctorStatus::Continue; |
189 | }); |
190 | |
191 | if (!codeBlock && hasTimeout) |
192 | return makeUnexpected(Error::TimedOut); |
193 | return codeBlock; |
194 | #else |
195 | UNUSED_PARAM(machinePC); |
196 | return nullptr; |
197 | #endif |
198 | } |
199 | |
200 | bool VMInspector::currentThreadOwnsJSLock(ExecState* exec) |
201 | { |
202 | return exec->vm().currentThreadIsHoldingAPILock(); |
203 | } |
204 | |
205 | static bool ensureCurrentThreadOwnsJSLock(ExecState* exec) |
206 | { |
207 | if (VMInspector::currentThreadOwnsJSLock(exec)) |
208 | return true; |
209 | dataLog("ERROR: current thread does not own the JSLock\n" ); |
210 | return false; |
211 | } |
212 | |
213 | void VMInspector::gc(ExecState* exec) |
214 | { |
215 | VM& vm = exec->vm(); |
216 | if (!ensureCurrentThreadOwnsJSLock(exec)) |
217 | return; |
218 | vm.heap.collectNow(Sync, CollectionScope::Full); |
219 | } |
220 | |
221 | void VMInspector::edenGC(ExecState* exec) |
222 | { |
223 | VM& vm = exec->vm(); |
224 | if (!ensureCurrentThreadOwnsJSLock(exec)) |
225 | return; |
226 | vm.heap.collectSync(CollectionScope::Eden); |
227 | } |
228 | |
229 | bool VMInspector::isInHeap(Heap* heap, void* ptr) |
230 | { |
231 | MarkedBlock* candidate = MarkedBlock::blockFor(ptr); |
232 | if (heap->objectSpace().blocks().set().contains(candidate)) |
233 | return true; |
234 | for (LargeAllocation* allocation : heap->objectSpace().largeAllocations()) { |
235 | if (allocation->contains(ptr)) |
236 | return true; |
237 | } |
238 | return false; |
239 | } |
240 | |
241 | struct CellAddressCheckFunctor : MarkedBlock::CountFunctor { |
242 | CellAddressCheckFunctor(JSCell* candidate) |
243 | : candidate(candidate) |
244 | { |
245 | } |
246 | |
247 | IterationStatus operator()(HeapCell* cell, HeapCell::Kind) const |
248 | { |
249 | if (cell == candidate) { |
250 | found = true; |
251 | return IterationStatus::Done; |
252 | } |
253 | return IterationStatus::Continue; |
254 | } |
255 | |
256 | JSCell* candidate; |
257 | mutable bool found { false }; |
258 | }; |
259 | |
260 | bool VMInspector::isValidCell(Heap* heap, JSCell* candidate) |
261 | { |
262 | HeapIterationScope iterationScope(*heap); |
263 | CellAddressCheckFunctor functor(candidate); |
264 | heap->objectSpace().forEachLiveCell(iterationScope, functor); |
265 | return functor.found; |
266 | } |
267 | |
268 | bool VMInspector::isValidCodeBlock(ExecState* exec, CodeBlock* candidate) |
269 | { |
270 | if (!ensureCurrentThreadOwnsJSLock(exec)) |
271 | return false; |
272 | |
273 | struct CodeBlockValidationFunctor { |
274 | CodeBlockValidationFunctor(CodeBlock* candidate) |
275 | : candidate(candidate) |
276 | { |
277 | } |
278 | |
279 | void operator()(CodeBlock* codeBlock) const |
280 | { |
281 | if (codeBlock == candidate) |
282 | found = true; |
283 | } |
284 | |
285 | CodeBlock* candidate; |
286 | mutable bool found { false }; |
287 | }; |
288 | |
289 | VM& vm = exec->vm(); |
290 | CodeBlockValidationFunctor functor(candidate); |
291 | vm.heap.forEachCodeBlock(functor); |
292 | return functor.found; |
293 | } |
294 | |
295 | CodeBlock* VMInspector::codeBlockForFrame(CallFrame* topCallFrame, unsigned ) |
296 | { |
297 | if (!ensureCurrentThreadOwnsJSLock(topCallFrame)) |
298 | return nullptr; |
299 | |
300 | if (!topCallFrame) |
301 | return nullptr; |
302 | |
303 | struct FetchCodeBlockFunctor { |
304 | public: |
305 | FetchCodeBlockFunctor(unsigned ) |
306 | : targetFrame(targetFrameNumber) |
307 | { |
308 | } |
309 | |
310 | StackVisitor::Status operator()(StackVisitor& visitor) const |
311 | { |
312 | auto currentFrame = nextFrame++; |
313 | if (currentFrame == targetFrame) { |
314 | codeBlock = visitor->codeBlock(); |
315 | return StackVisitor::Done; |
316 | } |
317 | return StackVisitor::Continue; |
318 | } |
319 | |
320 | unsigned targetFrame; |
321 | mutable unsigned nextFrame { 0 }; |
322 | mutable CodeBlock* codeBlock { nullptr }; |
323 | }; |
324 | |
325 | FetchCodeBlockFunctor functor(frameNumber); |
326 | topCallFrame->iterate(functor); |
327 | return functor.codeBlock; |
328 | } |
329 | |
330 | class DumpFrameFunctor { |
331 | public: |
332 | enum Action { |
333 | DumpOne, |
334 | DumpAll |
335 | }; |
336 | |
337 | DumpFrameFunctor(Action action, unsigned framesToSkip) |
338 | : m_action(action) |
339 | , m_framesToSkip(framesToSkip) |
340 | { |
341 | } |
342 | |
343 | StackVisitor::Status operator()(StackVisitor& visitor) const |
344 | { |
345 | m_currentFrame++; |
346 | if (m_currentFrame > m_framesToSkip) { |
347 | visitor->dump(WTF::dataFile(), Indenter(2), [&] (PrintStream& out) { |
348 | out.print("[" , (m_currentFrame - m_framesToSkip - 1), "] " ); |
349 | }); |
350 | } |
351 | if (m_action == DumpOne && m_currentFrame > m_framesToSkip) |
352 | return StackVisitor::Done; |
353 | return StackVisitor::Continue; |
354 | } |
355 | |
356 | private: |
357 | Action m_action; |
358 | unsigned m_framesToSkip; |
359 | mutable unsigned m_currentFrame { 0 }; |
360 | }; |
361 | |
362 | void VMInspector::dumpCallFrame(CallFrame* callFrame, unsigned framesToSkip) |
363 | { |
364 | if (!ensureCurrentThreadOwnsJSLock(callFrame)) |
365 | return; |
366 | DumpFrameFunctor functor(DumpFrameFunctor::DumpOne, framesToSkip); |
367 | callFrame->iterate(functor); |
368 | } |
369 | |
370 | void VMInspector::dumpRegisters(CallFrame* callFrame) |
371 | { |
372 | CodeBlock* codeBlock = callFrame->codeBlock(); |
373 | if (!codeBlock) { |
374 | dataLog("Dumping host frame registers not supported.\n" ); |
375 | return; |
376 | } |
377 | VM& vm = *codeBlock->vm(); |
378 | auto valueAsString = [&] (JSValue v) -> CString { |
379 | if (!v.isCell() || VMInspector::isValidCell(&vm.heap, reinterpret_cast<JSCell*>(JSValue::encode(v)))) |
380 | return toCString(v); |
381 | return "" ; |
382 | }; |
383 | |
384 | dataLogF("Register frame: \n\n" ); |
385 | dataLogF("-----------------------------------------------------------------------------\n" ); |
386 | dataLogF(" use | address | value \n" ); |
387 | dataLogF("-----------------------------------------------------------------------------\n" ); |
388 | |
389 | const Register* it; |
390 | const Register* end; |
391 | |
392 | it = callFrame->registers() + CallFrameSlot::thisArgument + callFrame->argumentCount(); |
393 | end = callFrame->registers() + CallFrameSlot::thisArgument - 1; |
394 | while (it > end) { |
395 | JSValue v = it->jsValue(); |
396 | int registerNumber = it - callFrame->registers(); |
397 | String name = codeBlock->nameForRegister(VirtualRegister(registerNumber)); |
398 | dataLogF("[r% 3d %14s] | %10p | 0x%-16llx %s\n" , registerNumber, name.ascii().data(), it, (long long)JSValue::encode(v), valueAsString(v).data()); |
399 | --it; |
400 | } |
401 | |
402 | dataLogF("-----------------------------------------------------------------------------\n" ); |
403 | dataLogF("[ArgumentCount] | %10p | %lu \n" , it, (unsigned long) callFrame->argumentCount()); |
404 | |
405 | callFrame->iterate([&] (StackVisitor& visitor) { |
406 | if (visitor->callFrame() == callFrame) { |
407 | unsigned line = 0; |
408 | unsigned unusedColumn = 0; |
409 | visitor->computeLineAndColumn(line, unusedColumn); |
410 | dataLogF("[ReturnVPC] | %10p | %d (line %d)\n" , it, visitor->bytecodeOffset(), line); |
411 | return StackVisitor::Done; |
412 | } |
413 | return StackVisitor::Continue; |
414 | }); |
415 | |
416 | --it; |
417 | dataLogF("[Callee] | %10p | 0x%-16llx %s\n" , it, (long long)callFrame->callee().rawPtr(), valueAsString(it->jsValue()).data()); |
418 | --it; |
419 | dataLogF("[CodeBlock] | %10p | 0x%-16llx " , it, (long long)codeBlock); |
420 | dataLogLn(codeBlock); |
421 | --it; |
422 | #if ENABLE(JIT) |
423 | AbstractPC pc = callFrame->abstractReturnPC(callFrame->vm()); |
424 | if (pc.hasJITReturnAddress()) |
425 | dataLogF("[ReturnPC] | %10p | %p \n" , it, pc.jitReturnAddress().value()); |
426 | --it; |
427 | #endif |
428 | dataLogF("[CallerFrame] | %10p | %p \n" , it, callFrame->callerFrame()); |
429 | --it; |
430 | dataLogF("-----------------------------------------------------------------------------\n" ); |
431 | |
432 | size_t numberOfCalleeSaveSlots = codeBlock->calleeSaveSpaceAsVirtualRegisters(); |
433 | const Register* endOfCalleeSaves = it - numberOfCalleeSaveSlots; |
434 | |
435 | end = it - codeBlock->numVars(); |
436 | if (it != end) { |
437 | do { |
438 | JSValue v = it->jsValue(); |
439 | int registerNumber = it - callFrame->registers(); |
440 | String name = (it > endOfCalleeSaves) |
441 | ? "CalleeSaveReg" |
442 | : codeBlock->nameForRegister(VirtualRegister(registerNumber)); |
443 | dataLogF("[r% 3d %14s] | %10p | 0x%-16llx %s\n" , registerNumber, name.ascii().data(), it, (long long)JSValue::encode(v), valueAsString(v).data()); |
444 | --it; |
445 | } while (it != end); |
446 | } |
447 | dataLogF("-----------------------------------------------------------------------------\n" ); |
448 | |
449 | end = it - codeBlock->numCalleeLocals() + codeBlock->numVars(); |
450 | if (it != end) { |
451 | do { |
452 | JSValue v = (*it).jsValue(); |
453 | int registerNumber = it - callFrame->registers(); |
454 | dataLogF("[r% 3d] | %10p | 0x%-16llx %s\n" , registerNumber, it, (long long)JSValue::encode(v), valueAsString(v).data()); |
455 | --it; |
456 | } while (it != end); |
457 | } |
458 | dataLogF("-----------------------------------------------------------------------------\n" ); |
459 | } |
460 | |
461 | void VMInspector::dumpStack(CallFrame* topCallFrame, unsigned framesToSkip) |
462 | { |
463 | if (!ensureCurrentThreadOwnsJSLock(topCallFrame)) |
464 | return; |
465 | if (!topCallFrame) |
466 | return; |
467 | DumpFrameFunctor functor(DumpFrameFunctor::DumpAll, framesToSkip); |
468 | topCallFrame->iterate(functor); |
469 | } |
470 | |
471 | void VMInspector::dumpValue(JSValue value) |
472 | { |
473 | dataLog(value); |
474 | } |
475 | |
476 | void VMInspector::dumpCellMemory(JSCell* cell) |
477 | { |
478 | dumpCellMemoryToStream(cell, WTF::dataFile()); |
479 | } |
480 | |
481 | class IndentationScope { |
482 | public: |
483 | IndentationScope(unsigned& indentation) |
484 | : m_indentation(indentation) |
485 | { |
486 | ++m_indentation; |
487 | } |
488 | |
489 | ~IndentationScope() |
490 | { |
491 | --m_indentation; |
492 | } |
493 | |
494 | private: |
495 | unsigned& m_indentation; |
496 | }; |
497 | |
498 | void VMInspector::dumpCellMemoryToStream(JSCell* cell, PrintStream& out) |
499 | { |
500 | VM& vm = *cell->vm(); |
501 | StructureID structureID = cell->structureID(); |
502 | Structure* structure = cell->structure(vm); |
503 | IndexingType indexingTypeAndMisc = cell->indexingTypeAndMisc(); |
504 | IndexingType indexingType = structure->indexingType(); |
505 | IndexingType indexingMode = structure->indexingMode(); |
506 | JSType type = cell->type(); |
507 | TypeInfo::InlineTypeFlags inlineTypeFlags = cell->inlineTypeFlags(); |
508 | CellState cellState = cell->cellState(); |
509 | size_t cellSize = cell->cellSize(); |
510 | size_t slotCount = cellSize / sizeof(EncodedJSValue); |
511 | |
512 | EncodedJSValue* slots = bitwise_cast<EncodedJSValue*>(cell); |
513 | unsigned indentation = 0; |
514 | |
515 | auto indent = [&] { |
516 | for (unsigned i = 0 ; i < indentation; ++i) |
517 | out.print(" " ); |
518 | }; |
519 | |
520 | #define INDENT indent(), |
521 | |
522 | auto dumpSlot = [&] (EncodedJSValue* slots, unsigned index, const char* label = nullptr) { |
523 | out.print("[" , index, "] " , format("%p : 0x%016" PRIx64, &slots[index], slots[index])); |
524 | if (label) |
525 | out.print(" " , label); |
526 | out.print("\n" ); |
527 | }; |
528 | |
529 | out.printf("<%p, %s>\n" , cell, cell->className(vm)); |
530 | IndentationScope scope(indentation); |
531 | |
532 | INDENT dumpSlot(slots, 0, "header" ); |
533 | { |
534 | IndentationScope scope(indentation); |
535 | INDENT out.println("structureID " , format("%d 0x%" PRIx32, structureID, structureID), " structure " , RawPointer(structure)); |
536 | INDENT out.println("indexingTypeAndMisc " , format("%d 0x%" PRIx8, indexingTypeAndMisc, indexingTypeAndMisc), " " , IndexingTypeDump(indexingMode)); |
537 | INDENT out.println("type " , format("%d 0x%" PRIx8, type, type)); |
538 | INDENT out.println("flags " , format("%d 0x%" PRIx8, inlineTypeFlags, inlineTypeFlags)); |
539 | INDENT out.println("cellState " , format("%d" , cellState)); |
540 | } |
541 | |
542 | unsigned slotIndex = 1; |
543 | if (cell->isObject()) { |
544 | JSObject* obj = static_cast<JSObject*>(const_cast<JSCell*>(cell)); |
545 | Butterfly* butterfly = obj->butterfly(); |
546 | size_t butterflySize = obj->butterflyTotalSize(); |
547 | |
548 | INDENT dumpSlot(slots, slotIndex, "butterfly" ); |
549 | slotIndex++; |
550 | |
551 | if (butterfly) { |
552 | IndentationScope scope(indentation); |
553 | |
554 | bool = structure->hasIndexingHeader(cell); |
555 | bool hasAnyArrayStorage = JSC::hasAnyArrayStorage(indexingType); |
556 | |
557 | size_t preCapacity = obj->butterflyPreCapacity(); |
558 | size_t propertyCapacity = structure->outOfLineCapacity(); |
559 | |
560 | void* base = hasIndexingHeader |
561 | ? butterfly->base(preCapacity, propertyCapacity) |
562 | : butterfly->base(structure); |
563 | |
564 | unsigned publicLength = butterfly->publicLength(); |
565 | unsigned vectorLength = butterfly->vectorLength(); |
566 | size_t butterflyCellSize = MarkedSpace::optimalSizeFor(butterflySize); |
567 | |
568 | size_t endOfIndexedPropertiesIndex = butterflySize / sizeof(EncodedJSValue); |
569 | size_t endOfButterflyIndex = butterflyCellSize / sizeof(EncodedJSValue); |
570 | |
571 | INDENT out.println("base " , RawPointer(base)); |
572 | INDENT out.println("hasIndexingHeader " , (hasIndexingHeader ? "YES" : "NO" ), " hasAnyArrayStorage " , (hasAnyArrayStorage ? "YES" : "NO" )); |
573 | if (hasIndexingHeader) { |
574 | INDENT out.print("publicLength " , publicLength, " vectorLength " , vectorLength); |
575 | if (hasAnyArrayStorage) |
576 | out.print(" indexBias " , butterfly->arrayStorage()->m_indexBias); |
577 | out.print("\n" ); |
578 | } |
579 | INDENT out.println("preCapacity " , preCapacity, " propertyCapacity " , propertyCapacity); |
580 | |
581 | unsigned index = 0; |
582 | EncodedJSValue* slots = reinterpret_cast<EncodedJSValue*>(base); |
583 | |
584 | auto asVoidPtr = [] (void* p) { |
585 | return p; |
586 | }; |
587 | |
588 | auto = [&] (const char* name) { |
589 | out.println("<--- " , name); |
590 | }; |
591 | |
592 | auto dumpSection = [&] (unsigned startIndex, unsigned endIndex, const char* name) -> unsigned { |
593 | for (unsigned index = startIndex; index < endIndex; ++index) { |
594 | if (name && index == startIndex) |
595 | INDENT dumpSectionHeader(name); |
596 | INDENT dumpSlot(slots, index); |
597 | } |
598 | return endIndex; |
599 | }; |
600 | |
601 | { |
602 | IndentationScope scope(indentation); |
603 | |
604 | index = dumpSection(index, preCapacity, "preCapacity" ); |
605 | index = dumpSection(index, preCapacity + propertyCapacity, "propertyCapacity" ); |
606 | |
607 | if (hasIndexingHeader) |
608 | index = dumpSection(index, index + 1, "indexingHeader" ); |
609 | |
610 | INDENT dumpSectionHeader("butterfly" ); |
611 | if (hasAnyArrayStorage) { |
612 | RELEASE_ASSERT(asVoidPtr(butterfly->arrayStorage()) == asVoidPtr(&slots[index])); |
613 | RELEASE_ASSERT(ArrayStorage::vectorOffset() == 2 * sizeof(EncodedJSValue)); |
614 | index = dumpSection(index, index + 2, "arrayStorage" ); |
615 | } |
616 | |
617 | index = dumpSection(index, endOfIndexedPropertiesIndex, "indexedProperties" ); |
618 | index = dumpSection(index, endOfButterflyIndex, "unallocated capacity" ); |
619 | } |
620 | } |
621 | } |
622 | |
623 | for (; slotIndex < slotCount; ++slotIndex) |
624 | INDENT dumpSlot(slots, slotIndex); |
625 | |
626 | #undef INDENT |
627 | } |
628 | |
629 | } // namespace JSC |
630 | |