1 | /* |
2 | * Copyright (C) 2015-2018 Apple Inc. All rights reserved. |
3 | * |
4 | * Redistribution and use in source and binary forms, with or without |
5 | * modification, are permitted provided that the following conditions |
6 | * are met: |
7 | * 1. Redistributions of source code must retain the above copyright |
8 | * notice, this list of conditions and the following disclaimer. |
9 | * 2. Redistributions in binary form must reproduce the above copyright |
10 | * notice, this list of conditions and the following disclaimer in the |
11 | * documentation and/or other materials provided with the distribution. |
12 | * |
13 | * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY |
14 | * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
15 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
16 | * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR |
17 | * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
18 | * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
19 | * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
20 | * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
21 | * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
22 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
23 | * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
24 | */ |
25 | |
26 | #pragma once |
27 | |
28 | #if ENABLE(B3_JIT) |
29 | |
30 | #include "AirArg.h" |
31 | #include "AirBasicBlock.h" |
32 | #include "AirDisassembler.h" |
33 | #include "AirSpecial.h" |
34 | #include "AirStackSlot.h" |
35 | #include "AirTmp.h" |
36 | #include "B3SparseCollection.h" |
37 | #include "CCallHelpers.h" |
38 | #include "RegisterAtOffsetList.h" |
39 | #include "StackAlignment.h" |
40 | #include <wtf/IndexMap.h> |
41 | #include <wtf/WeakRandom.h> |
42 | |
43 | namespace JSC { namespace B3 { |
44 | |
45 | class Procedure; |
46 | |
47 | #if ASSERT_DISABLED |
48 | IGNORE_RETURN_TYPE_WARNINGS_BEGIN |
49 | #endif |
50 | |
51 | namespace Air { |
52 | |
53 | class GenerateAndAllocateRegisters; |
54 | class BlockInsertionSet; |
55 | class CCallSpecial; |
56 | class CFG; |
57 | class Code; |
58 | class Disassembler; |
59 | |
60 | typedef void WasmBoundsCheckGeneratorFunction(CCallHelpers&, GPRReg); |
61 | typedef SharedTask<WasmBoundsCheckGeneratorFunction> WasmBoundsCheckGenerator; |
62 | |
63 | typedef void PrologueGeneratorFunction(CCallHelpers&, Code&); |
64 | typedef SharedTask<PrologueGeneratorFunction> PrologueGenerator; |
65 | |
66 | // This is an IR that is very close to the bare metal. It requires about 40x more bytes than the |
67 | // generated machine code - for example if you're generating 1MB of machine code, you need about |
68 | // 40MB of Air. |
69 | |
70 | class Code { |
71 | WTF_MAKE_NONCOPYABLE(Code); |
72 | WTF_MAKE_FAST_ALLOCATED; |
73 | public: |
74 | ~Code(); |
75 | |
76 | Procedure& proc() { return m_proc; } |
77 | |
78 | const Vector<Reg>& regsInPriorityOrder(Bank bank) const |
79 | { |
80 | switch (bank) { |
81 | case GP: |
82 | return m_gpRegsInPriorityOrder; |
83 | case FP: |
84 | return m_fpRegsInPriorityOrder; |
85 | } |
86 | ASSERT_NOT_REACHED(); |
87 | } |
88 | |
89 | // This is the set of registers that Air is allowed to emit code to mutate. It's derived from |
90 | // regsInPriorityOrder. Any registers not in this set are said to be "pinned". |
91 | const RegisterSet& mutableRegs() const { return m_mutableRegs; } |
92 | |
93 | bool isPinned(Reg reg) const { return !mutableRegs().get(reg); } |
94 | void pinRegister(Reg); |
95 | |
96 | void setOptLevel(unsigned optLevel) { m_optLevel = optLevel; } |
97 | unsigned optLevel() const { return m_optLevel; } |
98 | |
99 | bool needsUsedRegisters() const; |
100 | |
101 | JS_EXPORT_PRIVATE BasicBlock* addBlock(double frequency = 1); |
102 | |
103 | // Note that you can rely on stack slots always getting indices that are larger than the index |
104 | // of any prior stack slot. In fact, all stack slots you create in the future will have an index |
105 | // that is >= stackSlots().size(). |
106 | JS_EXPORT_PRIVATE StackSlot* addStackSlot( |
107 | unsigned byteSize, StackSlotKind, B3::StackSlot* = nullptr); |
108 | StackSlot* addStackSlot(B3::StackSlot*); |
109 | |
110 | JS_EXPORT_PRIVATE Special* addSpecial(std::unique_ptr<Special>); |
111 | |
112 | // This is the special you need to make a C call! |
113 | CCallSpecial* cCallSpecial(); |
114 | |
115 | Tmp newTmp(Bank bank) |
116 | { |
117 | switch (bank) { |
118 | case GP: |
119 | return Tmp::gpTmpForIndex(m_numGPTmps++); |
120 | case FP: |
121 | return Tmp::fpTmpForIndex(m_numFPTmps++); |
122 | } |
123 | ASSERT_NOT_REACHED(); |
124 | } |
125 | |
126 | unsigned numTmps(Bank bank) |
127 | { |
128 | switch (bank) { |
129 | case GP: |
130 | return m_numGPTmps; |
131 | case FP: |
132 | return m_numFPTmps; |
133 | } |
134 | ASSERT_NOT_REACHED(); |
135 | } |
136 | |
137 | template<typename Func> |
138 | void forEachTmp(const Func& func) |
139 | { |
140 | for (unsigned bankIndex = 0; bankIndex < numBanks; ++bankIndex) { |
141 | Bank bank = static_cast<Bank>(bankIndex); |
142 | unsigned numTmps = this->numTmps(bank); |
143 | for (unsigned i = 0; i < numTmps; ++i) |
144 | func(Tmp::tmpForIndex(bank, i)); |
145 | } |
146 | } |
147 | |
148 | unsigned callArgAreaSizeInBytes() const { return m_callArgAreaSize; } |
149 | |
150 | // You can call this before code generation to force a minimum call arg area size. |
151 | void requestCallArgAreaSizeInBytes(unsigned size) |
152 | { |
153 | m_callArgAreaSize = std::max( |
154 | m_callArgAreaSize, |
155 | static_cast<unsigned>(WTF::roundUpToMultipleOf(stackAlignmentBytes(), size))); |
156 | } |
157 | |
158 | unsigned frameSize() const { return m_frameSize; } |
159 | |
160 | // Only phases that do stack allocation are allowed to set this. Currently, only |
161 | // Air::allocateStack() does this. |
162 | void setFrameSize(unsigned frameSize) |
163 | { |
164 | m_frameSize = frameSize; |
165 | } |
166 | |
167 | // Note that this is not the same thing as proc().numEntrypoints(). This value here may be zero |
168 | // until we lower EntrySwitch. |
169 | unsigned numEntrypoints() const { return m_entrypoints.size(); } |
170 | const Vector<FrequentedBlock>& entrypoints() const { return m_entrypoints; } |
171 | const FrequentedBlock& entrypoint(unsigned index) const { return m_entrypoints[index]; } |
172 | bool isEntrypoint(BasicBlock*) const; |
173 | // Note: It is only valid to call this function after LowerEntrySwitch. |
174 | Optional<unsigned> entrypointIndex(BasicBlock*) const; |
175 | |
176 | // Note: We allow this to be called even before we set m_entrypoints just for convenience to users of this API. |
177 | // However, if you call this before setNumEntrypoints, setNumEntrypoints will overwrite this value. |
178 | void setPrologueForEntrypoint(unsigned entrypointIndex, Ref<PrologueGenerator>&& generator) |
179 | { |
180 | m_prologueGenerators[entrypointIndex] = WTFMove(generator); |
181 | } |
182 | const Ref<PrologueGenerator>& prologueGeneratorForEntrypoint(unsigned entrypointIndex) |
183 | { |
184 | return m_prologueGenerators[entrypointIndex]; |
185 | } |
186 | |
187 | void setNumEntrypoints(unsigned); |
188 | |
189 | // This is used by lowerEntrySwitch(). |
190 | template<typename Vector> |
191 | void setEntrypoints(Vector&& vector) |
192 | { |
193 | m_entrypoints = std::forward<Vector>(vector); |
194 | RELEASE_ASSERT(m_entrypoints.size() == m_prologueGenerators.size()); |
195 | } |
196 | |
197 | CCallHelpers::Label entrypointLabel(unsigned index) const |
198 | { |
199 | return m_entrypointLabels[index]; |
200 | } |
201 | |
202 | // This is used by generate(). |
203 | template<typename Vector> |
204 | void setEntrypointLabels(Vector&& vector) |
205 | { |
206 | m_entrypointLabels = std::forward<Vector>(vector); |
207 | RELEASE_ASSERT(m_entrypointLabels.size() == m_prologueGenerators.size()); |
208 | } |
209 | |
210 | void setStackIsAllocated(bool value) |
211 | { |
212 | m_stackIsAllocated = value; |
213 | } |
214 | |
215 | bool stackIsAllocated() const { return m_stackIsAllocated; } |
216 | |
217 | // This sets the callee save registers. |
218 | void setCalleeSaveRegisterAtOffsetList(RegisterAtOffsetList&&, StackSlot*); |
219 | |
220 | // This returns the correctly offset list of callee save registers. |
221 | RegisterAtOffsetList calleeSaveRegisterAtOffsetList() const; |
222 | |
223 | // This just tells you what the callee saves are. |
224 | RegisterSet calleeSaveRegisters() const { return m_calleeSaveRegisters; } |
225 | |
226 | // Recomputes predecessors and deletes unreachable blocks. |
227 | JS_EXPORT_PRIVATE void resetReachability(); |
228 | |
229 | JS_EXPORT_PRIVATE void dump(PrintStream&) const; |
230 | |
231 | unsigned size() const { return m_blocks.size(); } |
232 | BasicBlock* at(unsigned index) const { return m_blocks[index].get(); } |
233 | BasicBlock* operator[](unsigned index) const { return at(index); } |
234 | |
235 | // This is used by phases that optimize the block list. You shouldn't use this unless you really know |
236 | // what you're doing. |
237 | Vector<std::unique_ptr<BasicBlock>>& blockList() { return m_blocks; } |
238 | |
239 | // Finds the smallest index' such that at(index') != null and index' >= index. |
240 | JS_EXPORT_PRIVATE unsigned findFirstBlockIndex(unsigned index) const; |
241 | |
242 | // Finds the smallest index' such that at(index') != null and index' > index. |
243 | unsigned findNextBlockIndex(unsigned index) const; |
244 | |
245 | BasicBlock* findNextBlock(BasicBlock*) const; |
246 | |
247 | class iterator { |
248 | public: |
249 | iterator() |
250 | : m_code(nullptr) |
251 | , m_index(0) |
252 | { |
253 | } |
254 | |
255 | iterator(const Code& code, unsigned index) |
256 | : m_code(&code) |
257 | , m_index(m_code->findFirstBlockIndex(index)) |
258 | { |
259 | } |
260 | |
261 | BasicBlock* operator*() |
262 | { |
263 | return m_code->at(m_index); |
264 | } |
265 | |
266 | iterator& operator++() |
267 | { |
268 | m_index = m_code->findFirstBlockIndex(m_index + 1); |
269 | return *this; |
270 | } |
271 | |
272 | bool operator==(const iterator& other) const |
273 | { |
274 | return m_index == other.m_index; |
275 | } |
276 | |
277 | bool operator!=(const iterator& other) const |
278 | { |
279 | return !(*this == other); |
280 | } |
281 | |
282 | private: |
283 | const Code* m_code; |
284 | unsigned m_index; |
285 | }; |
286 | |
287 | iterator begin() const { return iterator(*this, 0); } |
288 | iterator end() const { return iterator(*this, size()); } |
289 | |
290 | const SparseCollection<StackSlot>& stackSlots() const { return m_stackSlots; } |
291 | SparseCollection<StackSlot>& stackSlots() { return m_stackSlots; } |
292 | |
293 | const SparseCollection<Special>& specials() const { return m_specials; } |
294 | SparseCollection<Special>& specials() { return m_specials; } |
295 | |
296 | template<typename Callback> |
297 | void forAllTmps(const Callback& callback) const |
298 | { |
299 | for (unsigned i = m_numGPTmps; i--;) |
300 | callback(Tmp::gpTmpForIndex(i)); |
301 | for (unsigned i = m_numFPTmps; i--;) |
302 | callback(Tmp::fpTmpForIndex(i)); |
303 | } |
304 | |
305 | void addFastTmp(Tmp); |
306 | bool isFastTmp(Tmp tmp) const { return m_fastTmps.contains(tmp); } |
307 | |
308 | CFG& cfg() const { return *m_cfg; } |
309 | |
310 | void* addDataSection(size_t); |
311 | |
312 | // The name has to be a string literal, since we don't do any memory management for the string. |
313 | void setLastPhaseName(const char* name) |
314 | { |
315 | m_lastPhaseName = name; |
316 | } |
317 | |
318 | const char* lastPhaseName() const { return m_lastPhaseName; } |
319 | |
320 | void setWasmBoundsCheckGenerator(RefPtr<WasmBoundsCheckGenerator> generator) |
321 | { |
322 | m_wasmBoundsCheckGenerator = generator; |
323 | } |
324 | |
325 | RefPtr<WasmBoundsCheckGenerator> wasmBoundsCheckGenerator() const { return m_wasmBoundsCheckGenerator; } |
326 | |
327 | // This is a hash of the code. You can use this if you want to put code into a hashtable, but |
328 | // it's mainly for validating the results from JSAir. |
329 | unsigned jsHash() const; |
330 | |
331 | void setDisassembler(std::unique_ptr<Disassembler>&& disassembler) { m_disassembler = WTFMove(disassembler); } |
332 | Disassembler* disassembler() { return m_disassembler.get(); } |
333 | |
334 | RegisterSet mutableGPRs(); |
335 | RegisterSet mutableFPRs(); |
336 | RegisterSet pinnedRegisters() const { return m_pinnedRegs; } |
337 | |
338 | WeakRandom& weakRandom() { return m_weakRandom; } |
339 | |
340 | void emitDefaultPrologue(CCallHelpers&); |
341 | |
342 | std::unique_ptr<GenerateAndAllocateRegisters> m_generateAndAllocateRegisters; |
343 | |
344 | private: |
345 | friend class ::JSC::B3::Procedure; |
346 | friend class BlockInsertionSet; |
347 | |
348 | Code(Procedure&); |
349 | |
350 | void setRegsInPriorityOrder(Bank, const Vector<Reg>&); |
351 | |
352 | Vector<Reg>& regsInPriorityOrderImpl(Bank bank) |
353 | { |
354 | switch (bank) { |
355 | case GP: |
356 | return m_gpRegsInPriorityOrder; |
357 | case FP: |
358 | return m_fpRegsInPriorityOrder; |
359 | } |
360 | ASSERT_NOT_REACHED(); |
361 | } |
362 | |
363 | WeakRandom m_weakRandom; |
364 | Procedure& m_proc; // Some meta-data, like byproducts, is stored in the Procedure. |
365 | Vector<Reg> m_gpRegsInPriorityOrder; |
366 | Vector<Reg> m_fpRegsInPriorityOrder; |
367 | RegisterSet m_mutableRegs; |
368 | RegisterSet m_pinnedRegs; |
369 | SparseCollection<StackSlot> m_stackSlots; |
370 | Vector<std::unique_ptr<BasicBlock>> m_blocks; |
371 | SparseCollection<Special> m_specials; |
372 | std::unique_ptr<CFG> m_cfg; |
373 | HashSet<Tmp> m_fastTmps; |
374 | CCallSpecial* m_cCallSpecial { nullptr }; |
375 | unsigned m_numGPTmps { 0 }; |
376 | unsigned m_numFPTmps { 0 }; |
377 | unsigned m_frameSize { 0 }; |
378 | unsigned m_callArgAreaSize { 0 }; |
379 | bool m_stackIsAllocated { false }; |
380 | RegisterAtOffsetList m_uncorrectedCalleeSaveRegisterAtOffsetList; |
381 | RegisterSet m_calleeSaveRegisters; |
382 | StackSlot* m_calleeSaveStackSlot { nullptr }; |
383 | Vector<FrequentedBlock> m_entrypoints; // This is empty until after lowerEntrySwitch(). |
384 | Vector<CCallHelpers::Label> m_entrypointLabels; // This is empty until code generation. |
385 | Vector<Ref<PrologueGenerator>, 1> m_prologueGenerators; |
386 | RefPtr<WasmBoundsCheckGenerator> m_wasmBoundsCheckGenerator; |
387 | const char* m_lastPhaseName; |
388 | std::unique_ptr<Disassembler> m_disassembler; |
389 | unsigned m_optLevel { defaultOptLevel() }; |
390 | Ref<PrologueGenerator> m_defaultPrologueGenerator; |
391 | }; |
392 | |
393 | } } } // namespace JSC::B3::Air |
394 | |
395 | #if ASSERT_DISABLED |
396 | IGNORE_RETURN_TYPE_WARNINGS_END |
397 | #endif |
398 | |
399 | #endif // ENABLE(B3_JIT) |
400 | |