Commit c896fe29d6c8ae6cde3917727812ced3f2e536a4
1 parent
56abbcff
TCG code generator
git-svn-id: svn://svn.savannah.nongnu.org/qemu/trunk@3943 c046a42c-6fe2-441c-8c8c-71466251a162
Showing
9 changed files
with
3625 additions
and
0 deletions
Too many changes to show.
To preserve performance only 9 of 13 files are displayed.
tcg/LICENSE
0 โ 100644
tcg/README
0 โ 100644
1 | +Tiny Code Generator - Fabrice Bellard. | ||
2 | + | ||
3 | +1) Introduction | ||
4 | + | ||
5 | +TCG (Tiny Code Generator) began as a generic backend for a C | ||
6 | +compiler. It was simplified to be used in QEMU. It also has its roots | ||
7 | +in the QOP code generator written by Paul Brook. | ||
8 | + | ||
9 | +2) Definitions | ||
10 | + | ||
11 | +The TCG "target" is the architecture for which we generate the | ||
12 | +code. It is of course not the same as the "target" of QEMU which is | ||
13 | +the emulated architecture. As TCG started as a generic C backend used | ||
14 | +for cross compiling, it is assumed that the TCG target is different | ||
15 | +from the host, although it is never the case for QEMU. | ||
16 | + | ||
17 | +A TCG "function" corresponds to a QEMU Translated Block (TB). | ||
18 | + | ||
19 | +A TCG "temporary" is a variable only live in a given | ||
20 | +function. Temporaries are allocated explicitely in each function. | ||
21 | + | ||
22 | +A TCG "global" is a variable which is live in all the functions. They | ||
23 | +are defined before the functions defined. A TCG global can be a memory | ||
24 | +location (e.g. a QEMU CPU register), a fixed host register (e.g. the | ||
25 | +QEMU CPU state pointer) or a memory location which is stored in a | ||
26 | +register outside QEMU TBs (not implemented yet). | ||
27 | + | ||
28 | +A TCG "basic block" corresponds to a list of instructions terminated | ||
29 | +by a branch instruction. | ||
30 | + | ||
31 | +3) Intermediate representation | ||
32 | + | ||
33 | +3.1) Introduction | ||
34 | + | ||
35 | +TCG instructions operate on variables which are temporaries or | ||
36 | +globals. TCG instructions and variables are strongly typed. Two types | ||
37 | +are supported: 32 bit integers and 64 bit integers. Pointers are | ||
38 | +defined as an alias to 32 bit or 64 bit integers depending on the TCG | ||
39 | +target word size. | ||
40 | + | ||
41 | +Each instruction has a fixed number of output variable operands, input | ||
42 | +variable operands and always constant operands. | ||
43 | + | ||
44 | +The notable exception is the call instruction which has a variable | ||
45 | +number of outputs and inputs. | ||
46 | + | ||
47 | +In the textual form, output operands come first, followed by input | ||
48 | +operands, followed by constant operands. The output type is included | ||
49 | +in the instruction name. Constants are prefixed with a '$'. | ||
50 | + | ||
51 | +add_i32 t0, t1, t2 (t0 <- t1 + t2) | ||
52 | + | ||
53 | +sub_i64 t2, t3, $4 (t2 <- t3 - 4) | ||
54 | + | ||
55 | +3.2) Assumptions | ||
56 | + | ||
57 | +* Basic blocks | ||
58 | + | ||
59 | +- Basic blocks end after branches (e.g. brcond_i32 instruction), | ||
60 | + goto_tb and exit_tb instructions. | ||
61 | +- Basic blocks end before legacy dyngen operations. | ||
62 | +- Basic blocks start after the end of a previous basic block, at a | ||
63 | + set_label instruction or after a legacy dyngen operation. | ||
64 | + | ||
65 | +After the end of a basic block, temporaries at destroyed and globals | ||
66 | +are stored at their initial storage (register or memory place | ||
67 | +depending on their declarations). | ||
68 | + | ||
69 | +* Floating point types are not supported yet | ||
70 | + | ||
71 | +* Pointers: depending on the TCG target, pointer size is 32 bit or 64 | ||
72 | + bit. The type TCG_TYPE_PTR is an alias to TCG_TYPE_I32 or | ||
73 | + TCG_TYPE_I64. | ||
74 | + | ||
75 | +* Helpers: | ||
76 | + | ||
77 | +Using the tcg_gen_helper_x_y it is possible to call any function | ||
78 | +taking i32, i64 or pointer types types. Before calling an helper, all | ||
79 | +globals are stored at their canonical location and it is assumed that | ||
80 | +the function can modify them. In the future, function modifiers will | ||
81 | +be allowed to tell that the helper does not read or write some globals. | ||
82 | + | ||
83 | +On some TCG targets (e.g. x86), several calling conventions are | ||
84 | +supported. | ||
85 | + | ||
86 | +* Branches: | ||
87 | + | ||
88 | +Use the instruction 'br' to jump to a label. Use 'jmp' to jump to an | ||
89 | +explicit address. Conditional branches can only jump to labels. | ||
90 | + | ||
91 | +3.3) Code Optimizations | ||
92 | + | ||
93 | +When generating instructions, you can count on at least the following | ||
94 | +optimizations: | ||
95 | + | ||
96 | +- Single instructions are simplified, e.g. | ||
97 | + | ||
98 | + and_i32 t0, t0, $0xffffffff | ||
99 | + | ||
100 | + is suppressed. | ||
101 | + | ||
102 | +- A liveness analysis is done at the basic block level. The | ||
103 | + information is used to suppress moves from a dead temporary to | ||
104 | + another one. It is also used to remove instructions which compute | ||
105 | + dead results. The later is especially useful for condition code | ||
106 | + optimisation in QEMU. | ||
107 | + | ||
108 | + In the following example: | ||
109 | + | ||
110 | + add_i32 t0, t1, t2 | ||
111 | + add_i32 t0, t0, $1 | ||
112 | + mov_i32 t0, $1 | ||
113 | + | ||
114 | + only the last instruction is kept. | ||
115 | + | ||
116 | +- A macro system is supported (may get closer to function inlining | ||
117 | + some day). It is useful if the liveness analysis is likely to prove | ||
118 | + that some results of a computation are indeed not useful. With the | ||
119 | + macro system, the user can provide several alternative | ||
120 | + implementations which are used depending on the used results. It is | ||
121 | + especially useful for condition code optimisation in QEMU. | ||
122 | + | ||
123 | + Here is an example: | ||
124 | + | ||
125 | + macro_2 t0, t1, $1 | ||
126 | + mov_i32 t0, $0x1234 | ||
127 | + | ||
128 | + The macro identified by the ID "$1" normally returns the values t0 | ||
129 | + and t1. Suppose its implementation is: | ||
130 | + | ||
131 | + macro_start | ||
132 | + brcond_i32 t2, $0, $TCG_COND_EQ, $1 | ||
133 | + mov_i32 t0, $2 | ||
134 | + br $2 | ||
135 | + set_label $1 | ||
136 | + mov_i32 t0, $3 | ||
137 | + set_label $2 | ||
138 | + add_i32 t1, t3, t4 | ||
139 | + macro_end | ||
140 | + | ||
141 | + If t0 is not used after the macro, the user can provide a simpler | ||
142 | + implementation: | ||
143 | + | ||
144 | + macro_start | ||
145 | + add_i32 t1, t2, t4 | ||
146 | + macro_end | ||
147 | + | ||
148 | + TCG automatically chooses the right implementation depending on | ||
149 | + which macro outputs are used after it. | ||
150 | + | ||
151 | + Note that if TCG did more expensive optimizations, macros would be | ||
152 | + less useful. In the previous example a macro is useful because the | ||
153 | + liveness analysis is done on each basic block separately. Hence TCG | ||
154 | + cannot remove the code computing 't0' even if it is not used after | ||
155 | + the first macro implementation. | ||
156 | + | ||
157 | +3.4) Instruction Reference | ||
158 | + | ||
159 | +********* Function call | ||
160 | + | ||
161 | +* call <ret> <params> ptr | ||
162 | + | ||
163 | +call function 'ptr' (pointer type) | ||
164 | + | ||
165 | +<ret> optional 32 bit or 64 bit return value | ||
166 | +<params> optional 32 bit or 64 bit parameters | ||
167 | + | ||
168 | +********* Jumps/Labels | ||
169 | + | ||
170 | +* jmp t0 | ||
171 | + | ||
172 | +Absolute jump to address t0 (pointer type). | ||
173 | + | ||
174 | +* set_label $label | ||
175 | + | ||
176 | +Define label 'label' at the current program point. | ||
177 | + | ||
178 | +* br $label | ||
179 | + | ||
180 | +Jump to label. | ||
181 | + | ||
182 | +* brcond_i32/i64 cond, t0, t1, label | ||
183 | + | ||
184 | +Conditional jump if t0 cond t1 is true. cond can be: | ||
185 | + TCG_COND_EQ | ||
186 | + TCG_COND_NE | ||
187 | + TCG_COND_LT /* signed */ | ||
188 | + TCG_COND_GE /* signed */ | ||
189 | + TCG_COND_LE /* signed */ | ||
190 | + TCG_COND_GT /* signed */ | ||
191 | + TCG_COND_LTU /* unsigned */ | ||
192 | + TCG_COND_GEU /* unsigned */ | ||
193 | + TCG_COND_LEU /* unsigned */ | ||
194 | + TCG_COND_GTU /* unsigned */ | ||
195 | + | ||
196 | +********* Arithmetic | ||
197 | + | ||
198 | +* add_i32/i64 t0, t1, t2 | ||
199 | + | ||
200 | +t0=t1+t2 | ||
201 | + | ||
202 | +* sub_i32/i64 t0, t1, t2 | ||
203 | + | ||
204 | +t0=t1-t2 | ||
205 | + | ||
206 | +* mul_i32/i64 t0, t1, t2 | ||
207 | + | ||
208 | +t0=t1*t2 | ||
209 | + | ||
210 | +* div_i32/i64 t0, t1, t2 | ||
211 | + | ||
212 | +t0=t1/t2 (signed). Undefined behavior if division by zero or overflow. | ||
213 | + | ||
214 | +* divu_i32/i64 t0, t1, t2 | ||
215 | + | ||
216 | +t0=t1/t2 (unsigned). Undefined behavior if division by zero. | ||
217 | + | ||
218 | +* rem_i32/i64 t0, t1, t2 | ||
219 | + | ||
220 | +t0=t1%t2 (signed). Undefined behavior if division by zero or overflow. | ||
221 | + | ||
222 | +* remu_i32/i64 t0, t1, t2 | ||
223 | + | ||
224 | +t0=t1%t2 (unsigned). Undefined behavior if division by zero. | ||
225 | + | ||
226 | +* and_i32/i64 t0, t1, t2 | ||
227 | + | ||
228 | +********* Logical | ||
229 | + | ||
230 | +t0=t1&t2 | ||
231 | + | ||
232 | +* or_i32/i64 t0, t1, t2 | ||
233 | + | ||
234 | +t0=t1|t2 | ||
235 | + | ||
236 | +* xor_i32/i64 t0, t1, t2 | ||
237 | + | ||
238 | +t0=t1^t2 | ||
239 | + | ||
240 | +* shl_i32/i64 t0, t1, t2 | ||
241 | + | ||
242 | +********* Shifts | ||
243 | + | ||
244 | +* shl_i32/i64 t0, t1, t2 | ||
245 | + | ||
246 | +t0=t1 << t2. Undefined behavior if t2 < 0 or t2 >= 32 (resp 64) | ||
247 | + | ||
248 | +* shr_i32/i64 t0, t1, t2 | ||
249 | + | ||
250 | +t0=t1 >> t2 (unsigned). Undefined behavior if t2 < 0 or t2 >= 32 (resp 64) | ||
251 | + | ||
252 | +* sar_i32/i64 t0, t1, t2 | ||
253 | + | ||
254 | +t0=t1 >> t2 (signed). Undefined behavior if t2 < 0 or t2 >= 32 (resp 64) | ||
255 | + | ||
256 | +********* Misc | ||
257 | + | ||
258 | +* mov_i32/i64 t0, t1 | ||
259 | + | ||
260 | +t0 = t1 | ||
261 | + | ||
262 | +Move t1 to t0 (both operands must have the same type). | ||
263 | + | ||
264 | +* ext8s_i32/i64 t0, t1 | ||
265 | +ext16s_i32/i64 t0, t1 | ||
266 | +ext32s_i64 t0, t1 | ||
267 | + | ||
268 | +8, 16 or 32 bit sign extension (both operands must have the same type) | ||
269 | + | ||
270 | +* bswap16_i32 t0, t1 | ||
271 | + | ||
272 | +16 bit byte swap on a 32 bit value. The two high order bytes must be set | ||
273 | +to zero. | ||
274 | + | ||
275 | +* bswap_i32 t0, t1 | ||
276 | + | ||
277 | +32 bit byte swap | ||
278 | + | ||
279 | +* bswap_i64 t0, t1 | ||
280 | + | ||
281 | +64 bit byte swap | ||
282 | + | ||
283 | +********* Type conversions | ||
284 | + | ||
285 | +* ext_i32_i64 t0, t1 | ||
286 | +Convert t1 (32 bit) to t0 (64 bit) and does sign extension | ||
287 | + | ||
288 | +* extu_i32_i64 t0, t1 | ||
289 | +Convert t1 (32 bit) to t0 (64 bit) and does zero extension | ||
290 | + | ||
291 | +* trunc_i64_i32 t0, t1 | ||
292 | +Truncate t1 (64 bit) to t0 (32 bit) | ||
293 | + | ||
294 | +********* Load/Store | ||
295 | + | ||
296 | +* ld_i32/i64 t0, t1, offset | ||
297 | +ld8s_i32/i64 t0, t1, offset | ||
298 | +ld8u_i32/i64 t0, t1, offset | ||
299 | +ld16s_i32/i64 t0, t1, offset | ||
300 | +ld16u_i32/i64 t0, t1, offset | ||
301 | +ld32s_i64 t0, t1, offset | ||
302 | +ld32u_i64 t0, t1, offset | ||
303 | + | ||
304 | +t0 = read(t1 + offset) | ||
305 | +Load 8, 16, 32 or 64 bits with or without sign extension from host memory. | ||
306 | +offset must be a constant. | ||
307 | + | ||
308 | +* st_i32/i64 t0, t1, offset | ||
309 | +st8_i32/i64 t0, t1, offset | ||
310 | +st16_i32/i64 t0, t1, offset | ||
311 | +st32_i64 t0, t1, offset | ||
312 | + | ||
313 | +write(t0, t1 + offset) | ||
314 | +Write 8, 16, 32 or 64 bits to host memory. | ||
315 | + | ||
316 | +********* QEMU specific operations | ||
317 | + | ||
318 | +* tb_exit t0 | ||
319 | + | ||
320 | +Exit the current TB and return the value t0 (word type). | ||
321 | + | ||
322 | +* goto_tb index | ||
323 | + | ||
324 | +Exit the current TB and jump to the TB index 'index' (constant) if the | ||
325 | +current TB was linked to this TB. Otherwise execute the next | ||
326 | +instructions. | ||
327 | + | ||
328 | +* qemu_ld_i32/i64 t0, t1, flags | ||
329 | +qemu_ld8u_i32/i64 t0, t1, flags | ||
330 | +qemu_ld8s_i32/i64 t0, t1, flags | ||
331 | +qemu_ld16u_i32/i64 t0, t1, flags | ||
332 | +qemu_ld16s_i32/i64 t0, t1, flags | ||
333 | +qemu_ld32u_i64 t0, t1, flags | ||
334 | +qemu_ld32s_i64 t0, t1, flags | ||
335 | + | ||
336 | +Load data at the QEMU CPU address t1 into t0. t1 has the QEMU CPU | ||
337 | +address type. 'flags' contains the QEMU memory index (selects user or | ||
338 | +kernel access) for example. | ||
339 | + | ||
340 | +* qemu_st_i32/i64 t0, t1, flags | ||
341 | +qemu_st8_i32/i64 t0, t1, flags | ||
342 | +qemu_st16_i32/i64 t0, t1, flags | ||
343 | +qemu_st32_i64 t0, t1, flags | ||
344 | + | ||
345 | +Store the data t0 at the QEMU CPU Address t1. t1 has the QEMU CPU | ||
346 | +address type. 'flags' contains the QEMU memory index (selects user or | ||
347 | +kernel access) for example. | ||
348 | + | ||
349 | +Note 1: Some shortcuts are defined when the last operand is known to be | ||
350 | +a constant (e.g. addi for add, movi for mov). | ||
351 | + | ||
352 | +Note 2: When using TCG, the opcodes must never be generated directly | ||
353 | +as some of them may not be available as "real" opcodes. Always use the | ||
354 | +function tcg_gen_xxx(args). | ||
355 | + | ||
356 | +4) Backend | ||
357 | + | ||
358 | +tcg-target.h contains the target specific definitions. tcg-target.c | ||
359 | +contains the target specific code. | ||
360 | + | ||
361 | +4.1) Assumptions | ||
362 | + | ||
363 | +The target word size (TCG_TARGET_REG_BITS) is expected to be 32 bit or | ||
364 | +64 bit. It is expected that the pointer has the same size as the word. | ||
365 | + | ||
366 | +On a 32 bit target, all 64 bit operations are converted to 32 bits. A | ||
367 | +few specific operations must be implemented to allow it (see add2_i32, | ||
368 | +sub2_i32, brcond2_i32). | ||
369 | + | ||
370 | +Floating point operations are not supported in this version. A | ||
371 | +previous incarnation of the code generator had full support of them, | ||
372 | +but it is better to concentrate on integer operations first. | ||
373 | + | ||
374 | +On a 64 bit target, no assumption is made in TCG about the storage of | ||
375 | +the 32 bit values in 64 bit registers. | ||
376 | + | ||
377 | +4.2) Constraints | ||
378 | + | ||
379 | +GCC like constraints are used to define the constraints of every | ||
380 | +instruction. Memory constraints are not supported in this | ||
381 | +version. Aliases are specified in the input operands as for GCC. | ||
382 | + | ||
383 | +A target can define specific register or constant constraints. If an | ||
384 | +operation uses a constant input constraint which does not allow all | ||
385 | +constants, it must also accept registers in order to have a fallback. | ||
386 | + | ||
387 | +The movi_i32 and movi_i64 operations must accept any constants. | ||
388 | + | ||
389 | +The mov_i32 and mov_i64 operations must accept any registers of the | ||
390 | +same type. | ||
391 | + | ||
392 | +The ld/st instructions must accept signed 32 bit constant offsets. It | ||
393 | +can be implemented by reserving a specific register to compute the | ||
394 | +address if the offset is too big. | ||
395 | + | ||
396 | +The ld/st instructions must accept any destination (ld) or source (st) | ||
397 | +register. | ||
398 | + | ||
399 | +4.3) Function call assumptions | ||
400 | + | ||
401 | +- The only supported types for parameters and return value are: 32 and | ||
402 | + 64 bit integers and pointer. | ||
403 | +- The stack grows downwards. | ||
404 | +- The first N parameters are passed in registers. | ||
405 | +- The next parameters are passed on the stack by storing them as words. | ||
406 | +- Some registers are clobbered during the call. | ||
407 | +- The function can return 0 or 1 value in registers. On a 32 bit | ||
408 | + target, functions must be able to return 2 values in registers for | ||
409 | + 64 bit return type. | ||
410 | + | ||
411 | +5) Migration from dyngen to TCG | ||
412 | + | ||
413 | +TCG is backward compatible with QEMU "dyngen" operations. It means | ||
414 | +that TCG instructions can be freely mixed with dyngen operations. It | ||
415 | +is expected that QEMU targets will be progressively fully converted to | ||
416 | +TCG. Once a target is fully converted to dyngen, it will be possible | ||
417 | +to apply more optimizations because more registers will be free for | ||
418 | +the generated code. | ||
419 | + | ||
420 | +The exception model is the same as the dyngen one. |
tcg/TODO
0 โ 100644
1 | +- test macro system | ||
2 | + | ||
3 | +- test conditional jumps | ||
4 | + | ||
5 | +- test mul, div, ext8s, ext16s, bswap | ||
6 | + | ||
7 | +- generate a global TB prologue and epilogue to save/restore registers | ||
8 | + to/from the CPU state and to reserve a stack frame to optimize | ||
9 | + helper calls. Modify cpu-exec.c so that it does not use global | ||
10 | + register variables (except maybe for 'env'). | ||
11 | + | ||
12 | +- fully convert the x86 target. The minimal amount of work includes: | ||
13 | + - add cc_src, cc_dst and cc_op as globals | ||
14 | + - disable its eflags optimization (the liveness analysis should | ||
15 | + suffice) | ||
16 | + - move complicated operations to helpers (in particular FPU, SSE, MMX). | ||
17 | + | ||
18 | +- optimize the x86 target: | ||
19 | + - move some or all the registers as globals | ||
20 | + - use the TB prologue and epilogue to have QEMU target registers in | ||
21 | + pre assigned host registers. | ||
22 | + | ||
23 | +Ideas: | ||
24 | + | ||
25 | +- Move the slow part of the qemu_ld/st ops after the end of the TB. | ||
26 | + | ||
27 | +- Experiment: change instruction storage to simplify macro handling | ||
28 | + and to handle dynamic allocation and see if the translation speed is | ||
29 | + OK. | ||
30 | + | ||
31 | +- change exception syntax to get closer to QOP system (exception | ||
32 | + parameters given with a specific instruction). |
tcg/i386/tcg-target.c
0 โ 100644
1 | +/* | ||
2 | + * Tiny Code Generator for QEMU | ||
3 | + * | ||
4 | + * Copyright (c) 2008 Fabrice Bellard | ||
5 | + * | ||
6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
7 | + * of this software and associated documentation files (the "Software"), to deal | ||
8 | + * in the Software without restriction, including without limitation the rights | ||
9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
10 | + * copies of the Software, and to permit persons to whom the Software is | ||
11 | + * furnished to do so, subject to the following conditions: | ||
12 | + * | ||
13 | + * The above copyright notice and this permission notice shall be included in | ||
14 | + * all copies or substantial portions of the Software. | ||
15 | + * | ||
16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
22 | + * THE SOFTWARE. | ||
23 | + */ | ||
24 | +const char *tcg_target_reg_names[TCG_TARGET_NB_REGS] = { | ||
25 | + "%eax", | ||
26 | + "%ecx", | ||
27 | + "%edx", | ||
28 | + "%ebx", | ||
29 | + "%esp", | ||
30 | + "%ebp", | ||
31 | + "%esi", | ||
32 | + "%edi", | ||
33 | +}; | ||
34 | + | ||
35 | +int tcg_target_reg_alloc_order[TCG_TARGET_NB_REGS] = { | ||
36 | + TCG_REG_EAX, | ||
37 | + TCG_REG_EDX, | ||
38 | + TCG_REG_ECX, | ||
39 | + TCG_REG_EBX, | ||
40 | + TCG_REG_ESI, | ||
41 | + TCG_REG_EDI, | ||
42 | + TCG_REG_EBP, | ||
43 | + TCG_REG_ESP, | ||
44 | +}; | ||
45 | + | ||
46 | +const int tcg_target_call_iarg_regs[3] = { TCG_REG_EAX, TCG_REG_EDX, TCG_REG_ECX }; | ||
47 | +const int tcg_target_call_oarg_regs[2] = { TCG_REG_EAX, TCG_REG_EDX }; | ||
48 | + | ||
49 | +static void patch_reloc(uint8_t *code_ptr, int type, | ||
50 | + tcg_target_long value) | ||
51 | +{ | ||
52 | + switch(type) { | ||
53 | + case R_386_32: | ||
54 | + *(uint32_t *)code_ptr = value; | ||
55 | + break; | ||
56 | + case R_386_PC32: | ||
57 | + *(uint32_t *)code_ptr = value - (long)code_ptr; | ||
58 | + break; | ||
59 | + default: | ||
60 | + tcg_abort(); | ||
61 | + } | ||
62 | +} | ||
63 | + | ||
64 | +/* maximum number of register used for input function arguments */ | ||
65 | +static inline int tcg_target_get_call_iarg_regs_count(int flags) | ||
66 | +{ | ||
67 | + flags &= TCG_CALL_TYPE_MASK; | ||
68 | + switch(flags) { | ||
69 | + case TCG_CALL_TYPE_STD: | ||
70 | + return 0; | ||
71 | + case TCG_CALL_TYPE_REGPARM_1: | ||
72 | + case TCG_CALL_TYPE_REGPARM_2: | ||
73 | + case TCG_CALL_TYPE_REGPARM: | ||
74 | + return flags - TCG_CALL_TYPE_REGPARM_1 + 1; | ||
75 | + default: | ||
76 | + tcg_abort(); | ||
77 | + } | ||
78 | +} | ||
79 | + | ||
80 | +/* parse target specific constraints */ | ||
81 | +int target_parse_constraint(TCGArgConstraint *ct, const char **pct_str) | ||
82 | +{ | ||
83 | + const char *ct_str; | ||
84 | + | ||
85 | + ct_str = *pct_str; | ||
86 | + switch(ct_str[0]) { | ||
87 | + case 'a': | ||
88 | + ct->ct |= TCG_CT_REG; | ||
89 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EAX); | ||
90 | + break; | ||
91 | + case 'b': | ||
92 | + ct->ct |= TCG_CT_REG; | ||
93 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EBX); | ||
94 | + break; | ||
95 | + case 'c': | ||
96 | + ct->ct |= TCG_CT_REG; | ||
97 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_ECX); | ||
98 | + break; | ||
99 | + case 'd': | ||
100 | + ct->ct |= TCG_CT_REG; | ||
101 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EDX); | ||
102 | + break; | ||
103 | + case 'S': | ||
104 | + ct->ct |= TCG_CT_REG; | ||
105 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_ESI); | ||
106 | + break; | ||
107 | + case 'D': | ||
108 | + ct->ct |= TCG_CT_REG; | ||
109 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EDI); | ||
110 | + break; | ||
111 | + case 'q': | ||
112 | + ct->ct |= TCG_CT_REG; | ||
113 | + tcg_regset_set32(ct->u.regs, 0, 0xf); | ||
114 | + break; | ||
115 | + case 'r': | ||
116 | + ct->ct |= TCG_CT_REG; | ||
117 | + tcg_regset_set32(ct->u.regs, 0, 0xff); | ||
118 | + break; | ||
119 | + | ||
120 | + /* qemu_ld/st address constraint */ | ||
121 | + case 'L': | ||
122 | + ct->ct |= TCG_CT_REG; | ||
123 | + tcg_regset_set32(ct->u.regs, 0, 0xff); | ||
124 | + tcg_regset_reset_reg(ct->u.regs, TCG_REG_EAX); | ||
125 | + tcg_regset_reset_reg(ct->u.regs, TCG_REG_EDX); | ||
126 | + break; | ||
127 | + default: | ||
128 | + return -1; | ||
129 | + } | ||
130 | + ct_str++; | ||
131 | + *pct_str = ct_str; | ||
132 | + return 0; | ||
133 | +} | ||
134 | + | ||
135 | +/* test if a constant matches the constraint */ | ||
136 | +static inline int tcg_target_const_match(tcg_target_long val, | ||
137 | + const TCGArgConstraint *arg_ct) | ||
138 | +{ | ||
139 | + int ct; | ||
140 | + ct = arg_ct->ct; | ||
141 | + if (ct & TCG_CT_CONST) | ||
142 | + return 1; | ||
143 | + else | ||
144 | + return 0; | ||
145 | +} | ||
146 | + | ||
147 | +#define ARITH_ADD 0 | ||
148 | +#define ARITH_OR 1 | ||
149 | +#define ARITH_ADC 2 | ||
150 | +#define ARITH_SBB 3 | ||
151 | +#define ARITH_AND 4 | ||
152 | +#define ARITH_SUB 5 | ||
153 | +#define ARITH_XOR 6 | ||
154 | +#define ARITH_CMP 7 | ||
155 | + | ||
156 | +#define SHIFT_SHL 4 | ||
157 | +#define SHIFT_SHR 5 | ||
158 | +#define SHIFT_SAR 7 | ||
159 | + | ||
160 | +#define JCC_JMP (-1) | ||
161 | +#define JCC_JO 0x0 | ||
162 | +#define JCC_JNO 0x1 | ||
163 | +#define JCC_JB 0x2 | ||
164 | +#define JCC_JAE 0x3 | ||
165 | +#define JCC_JE 0x4 | ||
166 | +#define JCC_JNE 0x5 | ||
167 | +#define JCC_JBE 0x6 | ||
168 | +#define JCC_JA 0x7 | ||
169 | +#define JCC_JS 0x8 | ||
170 | +#define JCC_JNS 0x9 | ||
171 | +#define JCC_JP 0xa | ||
172 | +#define JCC_JNP 0xb | ||
173 | +#define JCC_JL 0xc | ||
174 | +#define JCC_JGE 0xd | ||
175 | +#define JCC_JLE 0xe | ||
176 | +#define JCC_JG 0xf | ||
177 | + | ||
178 | +#define P_EXT 0x100 /* 0x0f opcode prefix */ | ||
179 | + | ||
180 | +static const uint8_t tcg_cond_to_jcc[10] = { | ||
181 | + [TCG_COND_EQ] = JCC_JE, | ||
182 | + [TCG_COND_NE] = JCC_JNE, | ||
183 | + [TCG_COND_LT] = JCC_JL, | ||
184 | + [TCG_COND_GE] = JCC_JGE, | ||
185 | + [TCG_COND_LE] = JCC_JLE, | ||
186 | + [TCG_COND_GT] = JCC_JG, | ||
187 | + [TCG_COND_LTU] = JCC_JB, | ||
188 | + [TCG_COND_GEU] = JCC_JAE, | ||
189 | + [TCG_COND_LEU] = JCC_JBE, | ||
190 | + [TCG_COND_GTU] = JCC_JA, | ||
191 | +}; | ||
192 | + | ||
193 | +static inline void tcg_out_opc(TCGContext *s, int opc) | ||
194 | +{ | ||
195 | + if (opc & P_EXT) | ||
196 | + tcg_out8(s, 0x0f); | ||
197 | + tcg_out8(s, opc); | ||
198 | +} | ||
199 | + | ||
200 | +static inline void tcg_out_modrm(TCGContext *s, int opc, int r, int rm) | ||
201 | +{ | ||
202 | + tcg_out_opc(s, opc); | ||
203 | + tcg_out8(s, 0xc0 | (r << 3) | rm); | ||
204 | +} | ||
205 | + | ||
206 | +/* rm == -1 means no register index */ | ||
207 | +static inline void tcg_out_modrm_offset(TCGContext *s, int opc, int r, int rm, | ||
208 | + int32_t offset) | ||
209 | +{ | ||
210 | + tcg_out_opc(s, opc); | ||
211 | + if (rm == -1) { | ||
212 | + tcg_out8(s, 0x05 | (r << 3)); | ||
213 | + tcg_out32(s, offset); | ||
214 | + } else if (offset == 0 && rm != TCG_REG_EBP) { | ||
215 | + if (rm == TCG_REG_ESP) { | ||
216 | + tcg_out8(s, 0x04 | (r << 3)); | ||
217 | + tcg_out8(s, 0x24); | ||
218 | + } else { | ||
219 | + tcg_out8(s, 0x00 | (r << 3) | rm); | ||
220 | + } | ||
221 | + } else if ((int8_t)offset == offset) { | ||
222 | + if (rm == TCG_REG_ESP) { | ||
223 | + tcg_out8(s, 0x44 | (r << 3)); | ||
224 | + tcg_out8(s, 0x24); | ||
225 | + } else { | ||
226 | + tcg_out8(s, 0x40 | (r << 3) | rm); | ||
227 | + } | ||
228 | + tcg_out8(s, offset); | ||
229 | + } else { | ||
230 | + if (rm == TCG_REG_ESP) { | ||
231 | + tcg_out8(s, 0x84 | (r << 3)); | ||
232 | + tcg_out8(s, 0x24); | ||
233 | + } else { | ||
234 | + tcg_out8(s, 0x80 | (r << 3) | rm); | ||
235 | + } | ||
236 | + tcg_out32(s, offset); | ||
237 | + } | ||
238 | +} | ||
239 | + | ||
240 | +static inline void tcg_out_mov(TCGContext *s, int ret, int arg) | ||
241 | +{ | ||
242 | + if (arg != ret) | ||
243 | + tcg_out_modrm(s, 0x8b, ret, arg); | ||
244 | +} | ||
245 | + | ||
246 | +static inline void tcg_out_movi(TCGContext *s, TCGType type, | ||
247 | + int ret, int32_t arg) | ||
248 | +{ | ||
249 | + if (arg == 0) { | ||
250 | + /* xor r0,r0 */ | ||
251 | + tcg_out_modrm(s, 0x01 | (ARITH_XOR << 3), ret, ret); | ||
252 | + } else { | ||
253 | + tcg_out8(s, 0xb8 + ret); | ||
254 | + tcg_out32(s, arg); | ||
255 | + } | ||
256 | +} | ||
257 | + | ||
258 | +static inline void tcg_out_ld(TCGContext *s, int ret, | ||
259 | + int arg1, int32_t arg2) | ||
260 | +{ | ||
261 | + /* movl */ | ||
262 | + tcg_out_modrm_offset(s, 0x8b, ret, arg1, arg2); | ||
263 | +} | ||
264 | + | ||
265 | +static inline void tcg_out_st(TCGContext *s, int arg, | ||
266 | + int arg1, int32_t arg2) | ||
267 | +{ | ||
268 | + /* movl */ | ||
269 | + tcg_out_modrm_offset(s, 0x89, arg, arg1, arg2); | ||
270 | +} | ||
271 | + | ||
272 | +static inline void tgen_arithi(TCGContext *s, int c, int r0, int32_t val) | ||
273 | +{ | ||
274 | + if (val == (int8_t)val) { | ||
275 | + tcg_out_modrm(s, 0x83, c, r0); | ||
276 | + tcg_out8(s, val); | ||
277 | + } else { | ||
278 | + tcg_out_modrm(s, 0x81, c, r0); | ||
279 | + tcg_out32(s, val); | ||
280 | + } | ||
281 | +} | ||
282 | + | ||
283 | +void tcg_out_addi(TCGContext *s, int reg, tcg_target_long val) | ||
284 | +{ | ||
285 | + if (val != 0) | ||
286 | + tgen_arithi(s, ARITH_ADD, reg, val); | ||
287 | +} | ||
288 | + | ||
289 | +static void tcg_out_jxx(TCGContext *s, int opc, int label_index) | ||
290 | +{ | ||
291 | + int32_t val, val1; | ||
292 | + TCGLabel *l = &s->labels[label_index]; | ||
293 | + | ||
294 | + if (l->has_value) { | ||
295 | + val = l->u.value - (tcg_target_long)s->code_ptr; | ||
296 | + val1 = val - 2; | ||
297 | + if ((int8_t)val1 == val1) { | ||
298 | + if (opc == -1) | ||
299 | + tcg_out8(s, 0xeb); | ||
300 | + else | ||
301 | + tcg_out8(s, 0x70 + opc); | ||
302 | + tcg_out8(s, val1); | ||
303 | + } else { | ||
304 | + if (opc == -1) { | ||
305 | + tcg_out8(s, 0xe9); | ||
306 | + tcg_out32(s, val - 5); | ||
307 | + } else { | ||
308 | + tcg_out8(s, 0x0f); | ||
309 | + tcg_out8(s, 0x80 + opc); | ||
310 | + tcg_out32(s, val - 6); | ||
311 | + } | ||
312 | + } | ||
313 | + } else { | ||
314 | + if (opc == -1) { | ||
315 | + tcg_out8(s, 0xe9); | ||
316 | + } else { | ||
317 | + tcg_out8(s, 0x0f); | ||
318 | + tcg_out8(s, 0x80 + opc); | ||
319 | + } | ||
320 | + tcg_out_reloc(s, s->code_ptr, R_386_PC32, label_index, -4); | ||
321 | + tcg_out32(s, -4); | ||
322 | + } | ||
323 | +} | ||
324 | + | ||
325 | +static void tcg_out_brcond(TCGContext *s, int cond, | ||
326 | + TCGArg arg1, TCGArg arg2, int const_arg2, | ||
327 | + int label_index) | ||
328 | +{ | ||
329 | + int c; | ||
330 | + if (const_arg2) { | ||
331 | + if (arg2 == 0) { | ||
332 | + /* use test */ | ||
333 | + switch(cond) { | ||
334 | + case TCG_COND_EQ: | ||
335 | + c = JCC_JNE; | ||
336 | + break; | ||
337 | + case TCG_COND_NE: | ||
338 | + c = JCC_JNE; | ||
339 | + break; | ||
340 | + case TCG_COND_LT: | ||
341 | + c = JCC_JS; | ||
342 | + break; | ||
343 | + case TCG_COND_GE: | ||
344 | + c = JCC_JNS; | ||
345 | + break; | ||
346 | + default: | ||
347 | + goto do_cmpi; | ||
348 | + } | ||
349 | + /* test r, r */ | ||
350 | + tcg_out_modrm(s, 0x85, arg1, arg1); | ||
351 | + tcg_out_jxx(s, c, label_index); | ||
352 | + } else { | ||
353 | + do_cmpi: | ||
354 | + tgen_arithi(s, ARITH_CMP, arg1, arg2); | ||
355 | + tcg_out_jxx(s, tcg_cond_to_jcc[cond], label_index); | ||
356 | + } | ||
357 | + } else { | ||
358 | + tcg_out_modrm(s, 0x01 | (ARITH_CMP << 3), arg1, arg2); | ||
359 | + tcg_out_jxx(s, tcg_cond_to_jcc[cond], label_index); | ||
360 | + } | ||
361 | +} | ||
362 | + | ||
363 | +/* XXX: we implement it at the target level to avoid having to | ||
364 | + handle cross basic blocks temporaries */ | ||
365 | +static void tcg_out_brcond2(TCGContext *s, | ||
366 | + const TCGArg *args, const int *const_args) | ||
367 | +{ | ||
368 | + int label_next; | ||
369 | + label_next = gen_new_label(); | ||
370 | + switch(args[4]) { | ||
371 | + case TCG_COND_EQ: | ||
372 | + tcg_out_brcond(s, TCG_COND_NE, args[0], args[2], const_args[2], label_next); | ||
373 | + tcg_out_brcond(s, TCG_COND_EQ, args[1], args[3], const_args[3], args[5]); | ||
374 | + break; | ||
375 | + case TCG_COND_NE: | ||
376 | + tcg_out_brcond(s, TCG_COND_NE, args[0], args[2], const_args[2], args[5]); | ||
377 | + tcg_out_brcond(s, TCG_COND_EQ, args[1], args[3], const_args[3], label_next); | ||
378 | + break; | ||
379 | + case TCG_COND_LT: | ||
380 | + tcg_out_brcond(s, TCG_COND_LT, args[1], args[3], const_args[3], args[5]); | ||
381 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
382 | + tcg_out_brcond(s, TCG_COND_LT, args[0], args[2], const_args[2], args[5]); | ||
383 | + break; | ||
384 | + case TCG_COND_LE: | ||
385 | + tcg_out_brcond(s, TCG_COND_LT, args[1], args[3], const_args[3], args[5]); | ||
386 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
387 | + tcg_out_brcond(s, TCG_COND_LE, args[0], args[2], const_args[2], args[5]); | ||
388 | + break; | ||
389 | + case TCG_COND_GT: | ||
390 | + tcg_out_brcond(s, TCG_COND_GT, args[1], args[3], const_args[3], args[5]); | ||
391 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
392 | + tcg_out_brcond(s, TCG_COND_GT, args[0], args[2], const_args[2], args[5]); | ||
393 | + break; | ||
394 | + case TCG_COND_GE: | ||
395 | + tcg_out_brcond(s, TCG_COND_GT, args[1], args[3], const_args[3], args[5]); | ||
396 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
397 | + tcg_out_brcond(s, TCG_COND_GE, args[0], args[2], const_args[2], args[5]); | ||
398 | + break; | ||
399 | + case TCG_COND_LTU: | ||
400 | + tcg_out_brcond(s, TCG_COND_LTU, args[1], args[3], const_args[3], args[5]); | ||
401 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
402 | + tcg_out_brcond(s, TCG_COND_LTU, args[0], args[2], const_args[2], args[5]); | ||
403 | + break; | ||
404 | + case TCG_COND_LEU: | ||
405 | + tcg_out_brcond(s, TCG_COND_LTU, args[1], args[3], const_args[3], args[5]); | ||
406 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
407 | + tcg_out_brcond(s, TCG_COND_LEU, args[0], args[2], const_args[2], args[5]); | ||
408 | + break; | ||
409 | + case TCG_COND_GTU: | ||
410 | + tcg_out_brcond(s, TCG_COND_GTU, args[1], args[3], const_args[3], args[5]); | ||
411 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
412 | + tcg_out_brcond(s, TCG_COND_GTU, args[0], args[2], const_args[2], args[5]); | ||
413 | + break; | ||
414 | + case TCG_COND_GEU: | ||
415 | + tcg_out_brcond(s, TCG_COND_GTU, args[1], args[3], const_args[3], args[5]); | ||
416 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
417 | + tcg_out_brcond(s, TCG_COND_GEU, args[0], args[2], const_args[2], args[5]); | ||
418 | + break; | ||
419 | + default: | ||
420 | + tcg_abort(); | ||
421 | + } | ||
422 | + tcg_out_label(s, label_next, (tcg_target_long)s->code_ptr); | ||
423 | +} | ||
424 | + | ||
425 | +#if defined(CONFIG_SOFTMMU) | ||
426 | +extern void __ldb_mmu(void); | ||
427 | +extern void __ldw_mmu(void); | ||
428 | +extern void __ldl_mmu(void); | ||
429 | +extern void __ldq_mmu(void); | ||
430 | + | ||
431 | +extern void __stb_mmu(void); | ||
432 | +extern void __stw_mmu(void); | ||
433 | +extern void __stl_mmu(void); | ||
434 | +extern void __stq_mmu(void); | ||
435 | + | ||
436 | +static void *qemu_ld_helpers[4] = { | ||
437 | + __ldb_mmu, | ||
438 | + __ldw_mmu, | ||
439 | + __ldl_mmu, | ||
440 | + __ldq_mmu, | ||
441 | +}; | ||
442 | + | ||
443 | +static void *qemu_st_helpers[4] = { | ||
444 | + __stb_mmu, | ||
445 | + __stw_mmu, | ||
446 | + __stl_mmu, | ||
447 | + __stq_mmu, | ||
448 | +}; | ||
449 | +#endif | ||
450 | + | ||
451 | +/* XXX: qemu_ld and qemu_st could be modified to clobber only EDX and | ||
452 | + EAX. It will be useful once fixed registers globals are less | ||
453 | + common. */ | ||
454 | +static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, | ||
455 | + int opc) | ||
456 | +{ | ||
457 | + int addr_reg, data_reg, data_reg2, r0, r1, mem_index, s_bits, bswap; | ||
458 | +#if defined(CONFIG_SOFTMMU) | ||
459 | + uint8_t *label1_ptr, *label2_ptr; | ||
460 | +#endif | ||
461 | +#if TARGET_LONG_BITS == 64 | ||
462 | +#if defined(CONFIG_SOFTMMU) | ||
463 | + uint8_t *label3_ptr; | ||
464 | +#endif | ||
465 | + int addr_reg2; | ||
466 | +#endif | ||
467 | + | ||
468 | + data_reg = *args++; | ||
469 | + if (opc == 3) | ||
470 | + data_reg2 = *args++; | ||
471 | + else | ||
472 | + data_reg2 = 0; | ||
473 | + addr_reg = *args++; | ||
474 | +#if TARGET_LONG_BITS == 64 | ||
475 | + addr_reg2 = *args++; | ||
476 | +#endif | ||
477 | + mem_index = *args; | ||
478 | + s_bits = opc & 3; | ||
479 | + | ||
480 | + r0 = TCG_REG_EAX; | ||
481 | + r1 = TCG_REG_EDX; | ||
482 | + | ||
483 | +#if defined(CONFIG_SOFTMMU) | ||
484 | + tcg_out_mov(s, r1, addr_reg); | ||
485 | + | ||
486 | + tcg_out_mov(s, r0, addr_reg); | ||
487 | + | ||
488 | + tcg_out_modrm(s, 0xc1, 5, r1); /* shr $x, r1 */ | ||
489 | + tcg_out8(s, TARGET_PAGE_BITS - CPU_TLB_ENTRY_BITS); | ||
490 | + | ||
491 | + tcg_out_modrm(s, 0x81, 4, r0); /* andl $x, r0 */ | ||
492 | + tcg_out32(s, TARGET_PAGE_MASK | ((1 << s_bits) - 1)); | ||
493 | + | ||
494 | + tcg_out_modrm(s, 0x81, 4, r1); /* andl $x, r1 */ | ||
495 | + tcg_out32(s, (CPU_TLB_SIZE - 1) << CPU_TLB_ENTRY_BITS); | ||
496 | + | ||
497 | + tcg_out_opc(s, 0x8d); /* lea offset(r1, %ebp), r1 */ | ||
498 | + tcg_out8(s, 0x80 | (r1 << 3) | 0x04); | ||
499 | + tcg_out8(s, (5 << 3) | r1); | ||
500 | + tcg_out32(s, offsetof(CPUState, tlb_table[mem_index][0].addr_read)); | ||
501 | + | ||
502 | + /* cmp 0(r1), r0 */ | ||
503 | + tcg_out_modrm_offset(s, 0x3b, r0, r1, 0); | ||
504 | + | ||
505 | + tcg_out_mov(s, r0, addr_reg); | ||
506 | + | ||
507 | +#if TARGET_LONG_BITS == 32 | ||
508 | + /* je label1 */ | ||
509 | + tcg_out8(s, 0x70 + JCC_JE); | ||
510 | + label1_ptr = s->code_ptr; | ||
511 | + s->code_ptr++; | ||
512 | +#else | ||
513 | + /* jne label3 */ | ||
514 | + tcg_out8(s, 0x70 + JCC_JNE); | ||
515 | + label3_ptr = s->code_ptr; | ||
516 | + s->code_ptr++; | ||
517 | + | ||
518 | + /* cmp 4(r1), addr_reg2 */ | ||
519 | + tcg_out_modrm_offset(s, 0x3b, addr_reg2, r1, 4); | ||
520 | + | ||
521 | + /* je label1 */ | ||
522 | + tcg_out8(s, 0x70 + JCC_JE); | ||
523 | + label1_ptr = s->code_ptr; | ||
524 | + s->code_ptr++; | ||
525 | + | ||
526 | + /* label3: */ | ||
527 | + *label3_ptr = s->code_ptr - label3_ptr - 1; | ||
528 | +#endif | ||
529 | + | ||
530 | + /* XXX: move that code at the end of the TB */ | ||
531 | +#if TARGET_LONG_BITS == 32 | ||
532 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_EDX, mem_index); | ||
533 | +#else | ||
534 | + tcg_out_mov(s, TCG_REG_EDX, addr_reg2); | ||
535 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_ECX, mem_index); | ||
536 | +#endif | ||
537 | + tcg_out8(s, 0xe8); | ||
538 | + tcg_out32(s, (tcg_target_long)qemu_ld_helpers[s_bits] - | ||
539 | + (tcg_target_long)s->code_ptr - 4); | ||
540 | + | ||
541 | + switch(opc) { | ||
542 | + case 0 | 4: | ||
543 | + /* movsbl */ | ||
544 | + tcg_out_modrm(s, 0xbe | P_EXT, data_reg, TCG_REG_EAX); | ||
545 | + break; | ||
546 | + case 1 | 4: | ||
547 | + /* movswl */ | ||
548 | + tcg_out_modrm(s, 0xbf | P_EXT, data_reg, TCG_REG_EAX); | ||
549 | + break; | ||
550 | + case 0: | ||
551 | + case 1: | ||
552 | + case 2: | ||
553 | + default: | ||
554 | + tcg_out_mov(s, data_reg, TCG_REG_EAX); | ||
555 | + break; | ||
556 | + case 3: | ||
557 | + if (data_reg == TCG_REG_EDX) { | ||
558 | + tcg_out_opc(s, 0x90 + TCG_REG_EDX); /* xchg %edx, %eax */ | ||
559 | + tcg_out_mov(s, data_reg2, TCG_REG_EAX); | ||
560 | + } else { | ||
561 | + tcg_out_mov(s, data_reg, TCG_REG_EAX); | ||
562 | + tcg_out_mov(s, data_reg2, TCG_REG_EDX); | ||
563 | + } | ||
564 | + break; | ||
565 | + } | ||
566 | + | ||
567 | + /* jmp label2 */ | ||
568 | + tcg_out8(s, 0xeb); | ||
569 | + label2_ptr = s->code_ptr; | ||
570 | + s->code_ptr++; | ||
571 | + | ||
572 | + /* label1: */ | ||
573 | + *label1_ptr = s->code_ptr - label1_ptr - 1; | ||
574 | + | ||
575 | + /* add x(r1), r0 */ | ||
576 | + tcg_out_modrm_offset(s, 0x03, r0, r1, offsetof(CPUTLBEntry, addend) - | ||
577 | + offsetof(CPUTLBEntry, addr_read)); | ||
578 | +#else | ||
579 | + r0 = addr_reg; | ||
580 | +#endif | ||
581 | + | ||
582 | +#ifdef TARGET_WORDS_BIGENDIAN | ||
583 | + bswap = 1; | ||
584 | +#else | ||
585 | + bswap = 0; | ||
586 | +#endif | ||
587 | + switch(opc) { | ||
588 | + case 0: | ||
589 | + /* movzbl */ | ||
590 | + tcg_out_modrm_offset(s, 0xb6 | P_EXT, data_reg, r0, 0); | ||
591 | + break; | ||
592 | + case 0 | 4: | ||
593 | + /* movsbl */ | ||
594 | + tcg_out_modrm_offset(s, 0xbe | P_EXT, data_reg, r0, 0); | ||
595 | + break; | ||
596 | + case 1: | ||
597 | + /* movzwl */ | ||
598 | + tcg_out_modrm_offset(s, 0xb7 | P_EXT, data_reg, r0, 0); | ||
599 | + if (bswap) { | ||
600 | + /* rolw $8, data_reg */ | ||
601 | + tcg_out8(s, 0x66); | ||
602 | + tcg_out_modrm(s, 0xc1, 0, data_reg); | ||
603 | + tcg_out8(s, 8); | ||
604 | + } | ||
605 | + break; | ||
606 | + case 1 | 4: | ||
607 | + /* movswl */ | ||
608 | + tcg_out_modrm_offset(s, 0xbf | P_EXT, data_reg, r0, 0); | ||
609 | + if (bswap) { | ||
610 | + /* rolw $8, data_reg */ | ||
611 | + tcg_out8(s, 0x66); | ||
612 | + tcg_out_modrm(s, 0xc1, 0, data_reg); | ||
613 | + tcg_out8(s, 8); | ||
614 | + | ||
615 | + /* movswl data_reg, data_reg */ | ||
616 | + tcg_out_modrm(s, 0xbf | P_EXT, data_reg, data_reg); | ||
617 | + } | ||
618 | + break; | ||
619 | + case 2: | ||
620 | + /* movl (r0), data_reg */ | ||
621 | + tcg_out_modrm_offset(s, 0x8b, data_reg, r0, 0); | ||
622 | + if (bswap) { | ||
623 | + /* bswap */ | ||
624 | + tcg_out_opc(s, (0xc8 + data_reg) | P_EXT); | ||
625 | + } | ||
626 | + break; | ||
627 | + case 3: | ||
628 | + /* XXX: could be nicer */ | ||
629 | + if (r0 == data_reg) { | ||
630 | + r1 = TCG_REG_EDX; | ||
631 | + if (r1 == data_reg) | ||
632 | + r1 = TCG_REG_EAX; | ||
633 | + tcg_out_mov(s, r1, r0); | ||
634 | + r0 = r1; | ||
635 | + } | ||
636 | + if (!bswap) { | ||
637 | + tcg_out_modrm_offset(s, 0x8b, data_reg, r0, 0); | ||
638 | + tcg_out_modrm_offset(s, 0x8b, data_reg2, r0, 4); | ||
639 | + } else { | ||
640 | + tcg_out_modrm_offset(s, 0x8b, data_reg, r0, 4); | ||
641 | + tcg_out_opc(s, (0xc8 + data_reg) | P_EXT); | ||
642 | + | ||
643 | + tcg_out_modrm_offset(s, 0x8b, data_reg2, r0, 0); | ||
644 | + /* bswap */ | ||
645 | + tcg_out_opc(s, (0xc8 + data_reg2) | P_EXT); | ||
646 | + } | ||
647 | + break; | ||
648 | + default: | ||
649 | + tcg_abort(); | ||
650 | + } | ||
651 | + | ||
652 | +#if defined(CONFIG_SOFTMMU) | ||
653 | + /* label2: */ | ||
654 | + *label2_ptr = s->code_ptr - label2_ptr - 1; | ||
655 | +#endif | ||
656 | +} | ||
657 | + | ||
658 | + | ||
659 | +static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, | ||
660 | + int opc) | ||
661 | +{ | ||
662 | + int addr_reg, data_reg, data_reg2, r0, r1, mem_index, s_bits, bswap; | ||
663 | +#if defined(CONFIG_SOFTMMU) | ||
664 | + uint8_t *label1_ptr, *label2_ptr; | ||
665 | +#endif | ||
666 | +#if TARGET_LONG_BITS == 64 | ||
667 | +#if defined(CONFIG_SOFTMMU) | ||
668 | + uint8_t *label3_ptr; | ||
669 | +#endif | ||
670 | + int addr_reg2; | ||
671 | +#endif | ||
672 | + | ||
673 | + data_reg = *args++; | ||
674 | + if (opc == 3) | ||
675 | + data_reg2 = *args++; | ||
676 | + else | ||
677 | + data_reg2 = 0; | ||
678 | + addr_reg = *args++; | ||
679 | +#if TARGET_LONG_BITS == 64 | ||
680 | + addr_reg2 = *args++; | ||
681 | +#endif | ||
682 | + mem_index = *args; | ||
683 | + | ||
684 | + s_bits = opc; | ||
685 | + | ||
686 | + r0 = TCG_REG_EAX; | ||
687 | + r1 = TCG_REG_EDX; | ||
688 | + | ||
689 | +#if defined(CONFIG_SOFTMMU) | ||
690 | + tcg_out_mov(s, r1, addr_reg); | ||
691 | + | ||
692 | + tcg_out_mov(s, r0, addr_reg); | ||
693 | + | ||
694 | + tcg_out_modrm(s, 0xc1, 5, r1); /* shr $x, r1 */ | ||
695 | + tcg_out8(s, TARGET_PAGE_BITS - CPU_TLB_ENTRY_BITS); | ||
696 | + | ||
697 | + tcg_out_modrm(s, 0x81, 4, r0); /* andl $x, r0 */ | ||
698 | + tcg_out32(s, TARGET_PAGE_MASK | ((1 << s_bits) - 1)); | ||
699 | + | ||
700 | + tcg_out_modrm(s, 0x81, 4, r1); /* andl $x, r1 */ | ||
701 | + tcg_out32(s, (CPU_TLB_SIZE - 1) << CPU_TLB_ENTRY_BITS); | ||
702 | + | ||
703 | + tcg_out_opc(s, 0x8d); /* lea offset(r1, %ebp), r1 */ | ||
704 | + tcg_out8(s, 0x80 | (r1 << 3) | 0x04); | ||
705 | + tcg_out8(s, (5 << 3) | r1); | ||
706 | + tcg_out32(s, offsetof(CPUState, tlb_table[mem_index][0].addr_write)); | ||
707 | + | ||
708 | + /* cmp 0(r1), r0 */ | ||
709 | + tcg_out_modrm_offset(s, 0x3b, r0, r1, 0); | ||
710 | + | ||
711 | + tcg_out_mov(s, r0, addr_reg); | ||
712 | + | ||
713 | +#if TARGET_LONG_BITS == 32 | ||
714 | + /* je label1 */ | ||
715 | + tcg_out8(s, 0x70 + JCC_JE); | ||
716 | + label1_ptr = s->code_ptr; | ||
717 | + s->code_ptr++; | ||
718 | +#else | ||
719 | + /* jne label3 */ | ||
720 | + tcg_out8(s, 0x70 + JCC_JNE); | ||
721 | + label3_ptr = s->code_ptr; | ||
722 | + s->code_ptr++; | ||
723 | + | ||
724 | + /* cmp 4(r1), addr_reg2 */ | ||
725 | + tcg_out_modrm_offset(s, 0x3b, addr_reg2, r1, 4); | ||
726 | + | ||
727 | + /* je label1 */ | ||
728 | + tcg_out8(s, 0x70 + JCC_JE); | ||
729 | + label1_ptr = s->code_ptr; | ||
730 | + s->code_ptr++; | ||
731 | + | ||
732 | + /* label3: */ | ||
733 | + *label3_ptr = s->code_ptr - label3_ptr - 1; | ||
734 | +#endif | ||
735 | + | ||
736 | + /* XXX: move that code at the end of the TB */ | ||
737 | +#if TARGET_LONG_BITS == 32 | ||
738 | + if (opc == 3) { | ||
739 | + tcg_out_mov(s, TCG_REG_EDX, data_reg); | ||
740 | + tcg_out_mov(s, TCG_REG_ECX, data_reg2); | ||
741 | + tcg_out8(s, 0x6a); /* push Ib */ | ||
742 | + tcg_out8(s, mem_index); | ||
743 | + tcg_out8(s, 0xe8); | ||
744 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | ||
745 | + (tcg_target_long)s->code_ptr - 4); | ||
746 | + tcg_out_addi(s, TCG_REG_ESP, 4); | ||
747 | + } else { | ||
748 | + switch(opc) { | ||
749 | + case 0: | ||
750 | + /* movzbl */ | ||
751 | + tcg_out_modrm(s, 0xb6 | P_EXT, TCG_REG_EDX, data_reg); | ||
752 | + break; | ||
753 | + case 1: | ||
754 | + /* movzwl */ | ||
755 | + tcg_out_modrm(s, 0xb7 | P_EXT, TCG_REG_EDX, data_reg); | ||
756 | + break; | ||
757 | + case 2: | ||
758 | + tcg_out_mov(s, TCG_REG_EDX, data_reg); | ||
759 | + break; | ||
760 | + } | ||
761 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_ECX, mem_index); | ||
762 | + tcg_out8(s, 0xe8); | ||
763 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | ||
764 | + (tcg_target_long)s->code_ptr - 4); | ||
765 | + } | ||
766 | +#else | ||
767 | + if (opc == 3) { | ||
768 | + tcg_out_mov(s, TCG_REG_EDX, addr_reg2); | ||
769 | + tcg_out8(s, 0x6a); /* push Ib */ | ||
770 | + tcg_out8(s, mem_index); | ||
771 | + tcg_out_opc(s, 0x50 + data_reg2); /* push */ | ||
772 | + tcg_out_opc(s, 0x50 + data_reg); /* push */ | ||
773 | + tcg_out8(s, 0xe8); | ||
774 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | ||
775 | + (tcg_target_long)s->code_ptr - 4); | ||
776 | + tcg_out_addi(s, TCG_REG_ESP, 12); | ||
777 | + } else { | ||
778 | + tcg_out_mov(s, TCG_REG_EDX, addr_reg2); | ||
779 | + switch(opc) { | ||
780 | + case 0: | ||
781 | + /* movzbl */ | ||
782 | + tcg_out_modrm(s, 0xb6 | P_EXT, TCG_REG_ECX, data_reg); | ||
783 | + break; | ||
784 | + case 1: | ||
785 | + /* movzwl */ | ||
786 | + tcg_out_modrm(s, 0xb7 | P_EXT, TCG_REG_ECX, data_reg); | ||
787 | + break; | ||
788 | + case 2: | ||
789 | + tcg_out_mov(s, TCG_REG_ECX, data_reg); | ||
790 | + break; | ||
791 | + } | ||
792 | + tcg_out8(s, 0x6a); /* push Ib */ | ||
793 | + tcg_out8(s, mem_index); | ||
794 | + tcg_out8(s, 0xe8); | ||
795 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | ||
796 | + (tcg_target_long)s->code_ptr - 4); | ||
797 | + tcg_out_addi(s, TCG_REG_ESP, 4); | ||
798 | + } | ||
799 | +#endif | ||
800 | + | ||
801 | + /* jmp label2 */ | ||
802 | + tcg_out8(s, 0xeb); | ||
803 | + label2_ptr = s->code_ptr; | ||
804 | + s->code_ptr++; | ||
805 | + | ||
806 | + /* label1: */ | ||
807 | + *label1_ptr = s->code_ptr - label1_ptr - 1; | ||
808 | + | ||
809 | + /* add x(r1), r0 */ | ||
810 | + tcg_out_modrm_offset(s, 0x03, r0, r1, offsetof(CPUTLBEntry, addend) - | ||
811 | + offsetof(CPUTLBEntry, addr_write)); | ||
812 | +#else | ||
813 | + r0 = addr_reg; | ||
814 | +#endif | ||
815 | + | ||
816 | +#ifdef TARGET_WORDS_BIGENDIAN | ||
817 | + bswap = 1; | ||
818 | +#else | ||
819 | + bswap = 0; | ||
820 | +#endif | ||
821 | + switch(opc) { | ||
822 | + case 0: | ||
823 | + /* movb */ | ||
824 | + tcg_out_modrm_offset(s, 0x88, data_reg, r0, 0); | ||
825 | + break; | ||
826 | + case 1: | ||
827 | + if (bswap) { | ||
828 | + tcg_out_mov(s, r1, data_reg); | ||
829 | + tcg_out8(s, 0x66); /* rolw $8, %ecx */ | ||
830 | + tcg_out_modrm(s, 0xc1, 0, r1); | ||
831 | + tcg_out8(s, 8); | ||
832 | + data_reg = r1; | ||
833 | + } | ||
834 | + /* movw */ | ||
835 | + tcg_out8(s, 0x66); | ||
836 | + tcg_out_modrm_offset(s, 0x89, data_reg, r0, 0); | ||
837 | + break; | ||
838 | + case 2: | ||
839 | + if (bswap) { | ||
840 | + tcg_out_mov(s, r1, data_reg); | ||
841 | + /* bswap data_reg */ | ||
842 | + tcg_out_opc(s, (0xc8 + r1) | P_EXT); | ||
843 | + data_reg = r1; | ||
844 | + } | ||
845 | + /* movl */ | ||
846 | + tcg_out_modrm_offset(s, 0x89, data_reg, r0, 0); | ||
847 | + break; | ||
848 | + case 3: | ||
849 | + if (bswap) { | ||
850 | + tcg_out_mov(s, r1, data_reg2); | ||
851 | + /* bswap data_reg */ | ||
852 | + tcg_out_opc(s, (0xc8 + r1) | P_EXT); | ||
853 | + tcg_out_modrm_offset(s, 0x89, r1, r0, 0); | ||
854 | + tcg_out_mov(s, r1, data_reg); | ||
855 | + /* bswap data_reg */ | ||
856 | + tcg_out_opc(s, (0xc8 + r1) | P_EXT); | ||
857 | + tcg_out_modrm_offset(s, 0x89, r1, r0, 4); | ||
858 | + } else { | ||
859 | + tcg_out_modrm_offset(s, 0x89, data_reg, r0, 0); | ||
860 | + tcg_out_modrm_offset(s, 0x89, data_reg2, r0, 4); | ||
861 | + } | ||
862 | + break; | ||
863 | + default: | ||
864 | + tcg_abort(); | ||
865 | + } | ||
866 | + | ||
867 | +#if defined(CONFIG_SOFTMMU) | ||
868 | + /* label2: */ | ||
869 | + *label2_ptr = s->code_ptr - label2_ptr - 1; | ||
870 | +#endif | ||
871 | +} | ||
872 | + | ||
873 | +static inline void tcg_out_op(TCGContext *s, int opc, | ||
874 | + const TCGArg *args, const int *const_args) | ||
875 | +{ | ||
876 | + int c; | ||
877 | + | ||
878 | + switch(opc) { | ||
879 | + case INDEX_op_exit_tb: | ||
880 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_EAX, args[0]); | ||
881 | + tcg_out8(s, 0xc3); /* ret */ | ||
882 | + break; | ||
883 | + case INDEX_op_goto_tb: | ||
884 | + if (s->tb_jmp_offset) { | ||
885 | + /* direct jump method */ | ||
886 | + tcg_out8(s, 0xe9); /* jmp im */ | ||
887 | + s->tb_jmp_offset[args[0]] = s->code_ptr - s->code_buf; | ||
888 | + tcg_out32(s, 0); | ||
889 | + } else { | ||
890 | + /* indirect jump method */ | ||
891 | + /* jmp Ev */ | ||
892 | + tcg_out_modrm_offset(s, 0xff, 4, -1, | ||
893 | + (tcg_target_long)(s->tb_next + args[0])); | ||
894 | + } | ||
895 | + s->tb_next_offset[args[0]] = s->code_ptr - s->code_buf; | ||
896 | + break; | ||
897 | + case INDEX_op_call: | ||
898 | + if (const_args[0]) { | ||
899 | + tcg_out8(s, 0xe8); | ||
900 | + tcg_out32(s, args[0] - (tcg_target_long)s->code_ptr - 4); | ||
901 | + } else { | ||
902 | + tcg_out_modrm(s, 0xff, 2, args[0]); | ||
903 | + } | ||
904 | + break; | ||
905 | + case INDEX_op_jmp: | ||
906 | + if (const_args[0]) { | ||
907 | + tcg_out8(s, 0xe9); | ||
908 | + tcg_out32(s, args[0] - (tcg_target_long)s->code_ptr - 4); | ||
909 | + } else { | ||
910 | + tcg_out_modrm(s, 0xff, 4, args[0]); | ||
911 | + } | ||
912 | + break; | ||
913 | + case INDEX_op_br: | ||
914 | + tcg_out_jxx(s, JCC_JMP, args[0]); | ||
915 | + break; | ||
916 | + case INDEX_op_movi_i32: | ||
917 | + tcg_out_movi(s, TCG_TYPE_I32, args[0], args[1]); | ||
918 | + break; | ||
919 | + case INDEX_op_ld8u_i32: | ||
920 | + /* movzbl */ | ||
921 | + tcg_out_modrm_offset(s, 0xb6 | P_EXT, args[0], args[1], args[2]); | ||
922 | + break; | ||
923 | + case INDEX_op_ld8s_i32: | ||
924 | + /* movsbl */ | ||
925 | + tcg_out_modrm_offset(s, 0xbe | P_EXT, args[0], args[1], args[2]); | ||
926 | + break; | ||
927 | + case INDEX_op_ld16u_i32: | ||
928 | + /* movzwl */ | ||
929 | + tcg_out_modrm_offset(s, 0xb7 | P_EXT, args[0], args[1], args[2]); | ||
930 | + break; | ||
931 | + case INDEX_op_ld16s_i32: | ||
932 | + /* movswl */ | ||
933 | + tcg_out_modrm_offset(s, 0xbf | P_EXT, args[0], args[1], args[2]); | ||
934 | + break; | ||
935 | + case INDEX_op_ld_i32: | ||
936 | + /* movl */ | ||
937 | + tcg_out_modrm_offset(s, 0x8b, args[0], args[1], args[2]); | ||
938 | + break; | ||
939 | + case INDEX_op_st8_i32: | ||
940 | + /* movb */ | ||
941 | + tcg_out_modrm_offset(s, 0x88, args[0], args[1], args[2]); | ||
942 | + break; | ||
943 | + case INDEX_op_st16_i32: | ||
944 | + /* movw */ | ||
945 | + tcg_out8(s, 0x66); | ||
946 | + tcg_out_modrm_offset(s, 0x89, args[0], args[1], args[2]); | ||
947 | + break; | ||
948 | + case INDEX_op_st_i32: | ||
949 | + /* movl */ | ||
950 | + tcg_out_modrm_offset(s, 0x89, args[0], args[1], args[2]); | ||
951 | + break; | ||
952 | + case INDEX_op_sub_i32: | ||
953 | + c = ARITH_SUB; | ||
954 | + goto gen_arith; | ||
955 | + case INDEX_op_and_i32: | ||
956 | + c = ARITH_AND; | ||
957 | + goto gen_arith; | ||
958 | + case INDEX_op_or_i32: | ||
959 | + c = ARITH_OR; | ||
960 | + goto gen_arith; | ||
961 | + case INDEX_op_xor_i32: | ||
962 | + c = ARITH_XOR; | ||
963 | + goto gen_arith; | ||
964 | + case INDEX_op_add_i32: | ||
965 | + c = ARITH_ADD; | ||
966 | + gen_arith: | ||
967 | + if (const_args[2]) { | ||
968 | + tgen_arithi(s, c, args[0], args[2]); | ||
969 | + } else { | ||
970 | + tcg_out_modrm(s, 0x01 | (c << 3), args[2], args[0]); | ||
971 | + } | ||
972 | + break; | ||
973 | + case INDEX_op_mul_i32: | ||
974 | + if (const_args[2]) { | ||
975 | + int32_t val; | ||
976 | + val = args[2]; | ||
977 | + if (val == (int8_t)val) { | ||
978 | + tcg_out_modrm(s, 0x6b, args[0], args[0]); | ||
979 | + tcg_out8(s, val); | ||
980 | + } else { | ||
981 | + tcg_out_modrm(s, 0x69, args[0], args[0]); | ||
982 | + tcg_out32(s, val); | ||
983 | + } | ||
984 | + } else { | ||
985 | + tcg_out_modrm(s, 0xaf | P_EXT, args[0], args[2]); | ||
986 | + } | ||
987 | + break; | ||
988 | + case INDEX_op_mulu2_i32: | ||
989 | + tcg_out_modrm(s, 0xf7, 4, args[3]); | ||
990 | + break; | ||
991 | + case INDEX_op_div2_i32: | ||
992 | + tcg_out_modrm(s, 0xf7, 7, args[4]); | ||
993 | + break; | ||
994 | + case INDEX_op_divu2_i32: | ||
995 | + tcg_out_modrm(s, 0xf7, 6, args[4]); | ||
996 | + break; | ||
997 | + case INDEX_op_shl_i32: | ||
998 | + c = SHIFT_SHL; | ||
999 | + gen_shift32: | ||
1000 | + if (const_args[2]) { | ||
1001 | + if (args[2] == 1) { | ||
1002 | + tcg_out_modrm(s, 0xd1, c, args[0]); | ||
1003 | + } else { | ||
1004 | + tcg_out_modrm(s, 0xc1, c, args[0]); | ||
1005 | + tcg_out8(s, args[2]); | ||
1006 | + } | ||
1007 | + } else { | ||
1008 | + tcg_out_modrm(s, 0xd3, c, args[0]); | ||
1009 | + } | ||
1010 | + break; | ||
1011 | + case INDEX_op_shr_i32: | ||
1012 | + c = SHIFT_SHR; | ||
1013 | + goto gen_shift32; | ||
1014 | + case INDEX_op_sar_i32: | ||
1015 | + c = SHIFT_SAR; | ||
1016 | + goto gen_shift32; | ||
1017 | + | ||
1018 | + case INDEX_op_add2_i32: | ||
1019 | + if (const_args[4]) | ||
1020 | + tgen_arithi(s, ARITH_ADD, args[0], args[4]); | ||
1021 | + else | ||
1022 | + tcg_out_modrm(s, 0x01 | (ARITH_ADD << 3), args[4], args[0]); | ||
1023 | + if (const_args[5]) | ||
1024 | + tgen_arithi(s, ARITH_ADC, args[1], args[5]); | ||
1025 | + else | ||
1026 | + tcg_out_modrm(s, 0x01 | (ARITH_ADC << 3), args[5], args[1]); | ||
1027 | + break; | ||
1028 | + case INDEX_op_sub2_i32: | ||
1029 | + if (const_args[4]) | ||
1030 | + tgen_arithi(s, ARITH_SUB, args[0], args[4]); | ||
1031 | + else | ||
1032 | + tcg_out_modrm(s, 0x01 | (ARITH_SUB << 3), args[4], args[0]); | ||
1033 | + if (const_args[5]) | ||
1034 | + tgen_arithi(s, ARITH_SBB, args[1], args[5]); | ||
1035 | + else | ||
1036 | + tcg_out_modrm(s, 0x01 | (ARITH_SBB << 3), args[5], args[1]); | ||
1037 | + break; | ||
1038 | + case INDEX_op_brcond_i32: | ||
1039 | + tcg_out_brcond(s, args[2], args[0], args[1], const_args[1], args[3]); | ||
1040 | + break; | ||
1041 | + case INDEX_op_brcond2_i32: | ||
1042 | + tcg_out_brcond2(s, args, const_args); | ||
1043 | + break; | ||
1044 | + | ||
1045 | + case INDEX_op_qemu_ld8u: | ||
1046 | + tcg_out_qemu_ld(s, args, 0); | ||
1047 | + break; | ||
1048 | + case INDEX_op_qemu_ld8s: | ||
1049 | + tcg_out_qemu_ld(s, args, 0 | 4); | ||
1050 | + break; | ||
1051 | + case INDEX_op_qemu_ld16u: | ||
1052 | + tcg_out_qemu_ld(s, args, 1); | ||
1053 | + break; | ||
1054 | + case INDEX_op_qemu_ld16s: | ||
1055 | + tcg_out_qemu_ld(s, args, 1 | 4); | ||
1056 | + break; | ||
1057 | + case INDEX_op_qemu_ld32u: | ||
1058 | + tcg_out_qemu_ld(s, args, 2); | ||
1059 | + break; | ||
1060 | + case INDEX_op_qemu_ld64: | ||
1061 | + tcg_out_qemu_ld(s, args, 3); | ||
1062 | + break; | ||
1063 | + | ||
1064 | + case INDEX_op_qemu_st8: | ||
1065 | + tcg_out_qemu_st(s, args, 0); | ||
1066 | + break; | ||
1067 | + case INDEX_op_qemu_st16: | ||
1068 | + tcg_out_qemu_st(s, args, 1); | ||
1069 | + break; | ||
1070 | + case INDEX_op_qemu_st32: | ||
1071 | + tcg_out_qemu_st(s, args, 2); | ||
1072 | + break; | ||
1073 | + case INDEX_op_qemu_st64: | ||
1074 | + tcg_out_qemu_st(s, args, 3); | ||
1075 | + break; | ||
1076 | + | ||
1077 | + default: | ||
1078 | + tcg_abort(); | ||
1079 | + } | ||
1080 | +} | ||
1081 | + | ||
1082 | +static const TCGTargetOpDef x86_op_defs[] = { | ||
1083 | + { INDEX_op_exit_tb, { } }, | ||
1084 | + { INDEX_op_goto_tb, { } }, | ||
1085 | + { INDEX_op_call, { "ri" } }, | ||
1086 | + { INDEX_op_jmp, { "ri" } }, | ||
1087 | + { INDEX_op_br, { } }, | ||
1088 | + { INDEX_op_mov_i32, { "r", "r" } }, | ||
1089 | + { INDEX_op_movi_i32, { "r" } }, | ||
1090 | + { INDEX_op_ld8u_i32, { "r", "r" } }, | ||
1091 | + { INDEX_op_ld8s_i32, { "r", "r" } }, | ||
1092 | + { INDEX_op_ld16u_i32, { "r", "r" } }, | ||
1093 | + { INDEX_op_ld16s_i32, { "r", "r" } }, | ||
1094 | + { INDEX_op_ld_i32, { "r", "r" } }, | ||
1095 | + { INDEX_op_st8_i32, { "q", "r" } }, | ||
1096 | + { INDEX_op_st16_i32, { "r", "r" } }, | ||
1097 | + { INDEX_op_st_i32, { "r", "r" } }, | ||
1098 | + | ||
1099 | + { INDEX_op_add_i32, { "r", "0", "ri" } }, | ||
1100 | + { INDEX_op_sub_i32, { "r", "0", "ri" } }, | ||
1101 | + { INDEX_op_mul_i32, { "r", "0", "ri" } }, | ||
1102 | + { INDEX_op_mulu2_i32, { "a", "d", "a", "r" } }, | ||
1103 | + { INDEX_op_div2_i32, { "a", "d", "0", "1", "r" } }, | ||
1104 | + { INDEX_op_divu2_i32, { "a", "d", "0", "1", "r" } }, | ||
1105 | + { INDEX_op_and_i32, { "r", "0", "ri" } }, | ||
1106 | + { INDEX_op_or_i32, { "r", "0", "ri" } }, | ||
1107 | + { INDEX_op_xor_i32, { "r", "0", "ri" } }, | ||
1108 | + | ||
1109 | + { INDEX_op_shl_i32, { "r", "0", "ci" } }, | ||
1110 | + { INDEX_op_shr_i32, { "r", "0", "ci" } }, | ||
1111 | + { INDEX_op_sar_i32, { "r", "0", "ci" } }, | ||
1112 | + | ||
1113 | + { INDEX_op_brcond_i32, { "r", "ri" } }, | ||
1114 | + | ||
1115 | + { INDEX_op_add2_i32, { "r", "r", "0", "1", "ri", "ri" } }, | ||
1116 | + { INDEX_op_sub2_i32, { "r", "r", "0", "1", "ri", "ri" } }, | ||
1117 | + { INDEX_op_brcond2_i32, { "r", "r", "ri", "ri" } }, | ||
1118 | + | ||
1119 | +#if TARGET_LONG_BITS == 32 | ||
1120 | + { INDEX_op_qemu_ld8u, { "r", "L" } }, | ||
1121 | + { INDEX_op_qemu_ld8s, { "r", "L" } }, | ||
1122 | + { INDEX_op_qemu_ld16u, { "r", "L" } }, | ||
1123 | + { INDEX_op_qemu_ld16s, { "r", "L" } }, | ||
1124 | + { INDEX_op_qemu_ld32u, { "r", "L" } }, | ||
1125 | + { INDEX_op_qemu_ld64, { "r", "r", "L" } }, | ||
1126 | + | ||
1127 | + { INDEX_op_qemu_st8, { "cb", "L" } }, | ||
1128 | + { INDEX_op_qemu_st16, { "L", "L" } }, | ||
1129 | + { INDEX_op_qemu_st32, { "L", "L" } }, | ||
1130 | + { INDEX_op_qemu_st64, { "L", "L", "L" } }, | ||
1131 | +#else | ||
1132 | + { INDEX_op_qemu_ld8u, { "r", "L", "L" } }, | ||
1133 | + { INDEX_op_qemu_ld8s, { "r", "L", "L" } }, | ||
1134 | + { INDEX_op_qemu_ld16u, { "r", "L", "L" } }, | ||
1135 | + { INDEX_op_qemu_ld16s, { "r", "L", "L" } }, | ||
1136 | + { INDEX_op_qemu_ld32u, { "r", "L", "L" } }, | ||
1137 | + { INDEX_op_qemu_ld64, { "r", "r", "L", "L" } }, | ||
1138 | + | ||
1139 | + { INDEX_op_qemu_st8, { "cb", "L", "L" } }, | ||
1140 | + { INDEX_op_qemu_st16, { "L", "L", "L" } }, | ||
1141 | + { INDEX_op_qemu_st32, { "L", "L", "L" } }, | ||
1142 | + { INDEX_op_qemu_st64, { "L", "L", "L", "L" } }, | ||
1143 | +#endif | ||
1144 | + { -1 }, | ||
1145 | +}; | ||
1146 | + | ||
1147 | +void tcg_target_init(TCGContext *s) | ||
1148 | +{ | ||
1149 | + /* fail safe */ | ||
1150 | + if ((1 << CPU_TLB_ENTRY_BITS) != sizeof(CPUTLBEntry)) | ||
1151 | + tcg_abort(); | ||
1152 | + | ||
1153 | + tcg_regset_set32(tcg_target_available_regs[TCG_TYPE_I32], 0, 0xff); | ||
1154 | + tcg_regset_set32(tcg_target_call_clobber_regs, 0, | ||
1155 | + (1 << TCG_REG_EAX) | | ||
1156 | + (1 << TCG_REG_EDX) | | ||
1157 | + (1 << TCG_REG_ECX)); | ||
1158 | + | ||
1159 | + tcg_regset_clear(s->reserved_regs); | ||
1160 | + tcg_regset_set_reg(s->reserved_regs, TCG_REG_ESP); | ||
1161 | + | ||
1162 | + tcg_add_target_add_op_defs(x86_op_defs); | ||
1163 | +} |
tcg/i386/tcg-target.h
0 โ 100644
1 | +/* | ||
2 | + * Tiny Code Generator for QEMU | ||
3 | + * | ||
4 | + * Copyright (c) 2008 Fabrice Bellard | ||
5 | + * | ||
6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
7 | + * of this software and associated documentation files (the "Software"), to deal | ||
8 | + * in the Software without restriction, including without limitation the rights | ||
9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
10 | + * copies of the Software, and to permit persons to whom the Software is | ||
11 | + * furnished to do so, subject to the following conditions: | ||
12 | + * | ||
13 | + * The above copyright notice and this permission notice shall be included in | ||
14 | + * all copies or substantial portions of the Software. | ||
15 | + * | ||
16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
22 | + * THE SOFTWARE. | ||
23 | + */ | ||
24 | +#define TCG_TARGET_I386 1 | ||
25 | + | ||
26 | +#define TCG_TARGET_REG_BITS 32 | ||
27 | +//#define TCG_TARGET_WORDS_BIGENDIAN | ||
28 | + | ||
29 | +#define TCG_TARGET_NB_REGS 8 | ||
30 | + | ||
31 | +enum { | ||
32 | + TCG_REG_EAX = 0, | ||
33 | + TCG_REG_ECX, | ||
34 | + TCG_REG_EDX, | ||
35 | + TCG_REG_EBX, | ||
36 | + TCG_REG_ESP, | ||
37 | + TCG_REG_EBP, | ||
38 | + TCG_REG_ESI, | ||
39 | + TCG_REG_EDI, | ||
40 | +}; | ||
41 | + | ||
42 | +/* used for function call generation */ | ||
43 | +#define TCG_REG_CALL_STACK TCG_REG_ESP | ||
44 | +#define TCG_TARGET_STACK_ALIGN 16 | ||
45 | + | ||
46 | +/* Note: must be synced with dyngen-exec.h */ | ||
47 | +#define TCG_AREG0 TCG_REG_EBP | ||
48 | +#define TCG_AREG1 TCG_REG_EBX | ||
49 | +#define TCG_AREG2 TCG_REG_ESI | ||
50 | +#define TCG_AREG3 TCG_REG_EDI | ||
51 | + | ||
52 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
53 | +{ | ||
54 | +} |
tcg/tcg-dyngen.c
0 โ 100644
1 | +/* | ||
2 | + * Tiny Code Generator for QEMU | ||
3 | + * | ||
4 | + * Copyright (c) 2008 Fabrice Bellard | ||
5 | + * | ||
6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
7 | + * of this software and associated documentation files (the "Software"), to deal | ||
8 | + * in the Software without restriction, including without limitation the rights | ||
9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
10 | + * copies of the Software, and to permit persons to whom the Software is | ||
11 | + * furnished to do so, subject to the following conditions: | ||
12 | + * | ||
13 | + * The above copyright notice and this permission notice shall be included in | ||
14 | + * all copies or substantial portions of the Software. | ||
15 | + * | ||
16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
22 | + * THE SOFTWARE. | ||
23 | + */ | ||
24 | +#include <assert.h> | ||
25 | +#include <stdarg.h> | ||
26 | +#include <stdlib.h> | ||
27 | +#include <stdio.h> | ||
28 | +#include <string.h> | ||
29 | +#include <inttypes.h> | ||
30 | + | ||
31 | +#include "config.h" | ||
32 | +#include "osdep.h" | ||
33 | + | ||
34 | +#include "tcg.h" | ||
35 | + | ||
36 | +int __op_param1, __op_param2, __op_param3; | ||
37 | +#if defined(__sparc__) || defined(__arm__) | ||
38 | + void __op_gen_label1(){} | ||
39 | + void __op_gen_label2(){} | ||
40 | + void __op_gen_label3(){} | ||
41 | +#else | ||
42 | + int __op_gen_label1, __op_gen_label2, __op_gen_label3; | ||
43 | +#endif | ||
44 | +int __op_jmp0, __op_jmp1, __op_jmp2, __op_jmp3; | ||
45 | + | ||
46 | +#if 0 | ||
47 | +#if defined(__s390__) | ||
48 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
49 | +{ | ||
50 | +} | ||
51 | +#elif defined(__ia64__) | ||
52 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
53 | +{ | ||
54 | + while (start < stop) { | ||
55 | + asm volatile ("fc %0" :: "r"(start)); | ||
56 | + start += 32; | ||
57 | + } | ||
58 | + asm volatile (";;sync.i;;srlz.i;;"); | ||
59 | +} | ||
60 | +#elif defined(__powerpc__) | ||
61 | + | ||
62 | +#define MIN_CACHE_LINE_SIZE 8 /* conservative value */ | ||
63 | + | ||
64 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
65 | +{ | ||
66 | + unsigned long p; | ||
67 | + | ||
68 | + start &= ~(MIN_CACHE_LINE_SIZE - 1); | ||
69 | + stop = (stop + MIN_CACHE_LINE_SIZE - 1) & ~(MIN_CACHE_LINE_SIZE - 1); | ||
70 | + | ||
71 | + for (p = start; p < stop; p += MIN_CACHE_LINE_SIZE) { | ||
72 | + asm volatile ("dcbst 0,%0" : : "r"(p) : "memory"); | ||
73 | + } | ||
74 | + asm volatile ("sync" : : : "memory"); | ||
75 | + for (p = start; p < stop; p += MIN_CACHE_LINE_SIZE) { | ||
76 | + asm volatile ("icbi 0,%0" : : "r"(p) : "memory"); | ||
77 | + } | ||
78 | + asm volatile ("sync" : : : "memory"); | ||
79 | + asm volatile ("isync" : : : "memory"); | ||
80 | +} | ||
81 | +#elif defined(__alpha__) | ||
82 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
83 | +{ | ||
84 | + asm ("imb"); | ||
85 | +} | ||
86 | +#elif defined(__sparc__) | ||
87 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
88 | +{ | ||
89 | + unsigned long p; | ||
90 | + | ||
91 | + p = start & ~(8UL - 1UL); | ||
92 | + stop = (stop + (8UL - 1UL)) & ~(8UL - 1UL); | ||
93 | + | ||
94 | + for (; p < stop; p += 8) | ||
95 | + __asm__ __volatile__("flush\t%0" : : "r" (p)); | ||
96 | +} | ||
97 | +#elif defined(__arm__) | ||
98 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
99 | +{ | ||
100 | + register unsigned long _beg __asm ("a1") = start; | ||
101 | + register unsigned long _end __asm ("a2") = stop; | ||
102 | + register unsigned long _flg __asm ("a3") = 0; | ||
103 | + __asm __volatile__ ("swi 0x9f0002" : : "r" (_beg), "r" (_end), "r" (_flg)); | ||
104 | +} | ||
105 | +#elif defined(__mc68000) | ||
106 | + | ||
107 | +# include <asm/cachectl.h> | ||
108 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
109 | +{ | ||
110 | + cacheflush(start,FLUSH_SCOPE_LINE,FLUSH_CACHE_BOTH,stop-start+16); | ||
111 | +} | ||
112 | +#elif defined(__mips__) | ||
113 | + | ||
114 | +#include <sys/cachectl.h> | ||
115 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
116 | +{ | ||
117 | + _flush_cache ((void *)start, stop - start, BCACHE); | ||
118 | +} | ||
119 | +#else | ||
120 | +#error unsupported CPU | ||
121 | +#endif | ||
122 | + | ||
123 | +#ifdef __alpha__ | ||
124 | + | ||
125 | +register int gp asm("$29"); | ||
126 | + | ||
127 | +static inline void immediate_ldah(void *p, int val) { | ||
128 | + uint32_t *dest = p; | ||
129 | + long high = ((val >> 16) + ((val >> 15) & 1)) & 0xffff; | ||
130 | + | ||
131 | + *dest &= ~0xffff; | ||
132 | + *dest |= high; | ||
133 | + *dest |= 31 << 16; | ||
134 | +} | ||
135 | +static inline void immediate_lda(void *dest, int val) { | ||
136 | + *(uint16_t *) dest = val; | ||
137 | +} | ||
138 | +void fix_bsr(void *p, int offset) { | ||
139 | + uint32_t *dest = p; | ||
140 | + *dest &= ~((1 << 21) - 1); | ||
141 | + *dest |= (offset >> 2) & ((1 << 21) - 1); | ||
142 | +} | ||
143 | + | ||
144 | +#endif /* __alpha__ */ | ||
145 | + | ||
146 | +#ifdef __arm__ | ||
147 | + | ||
148 | +#define ARM_LDR_TABLE_SIZE 1024 | ||
149 | + | ||
150 | +typedef struct LDREntry { | ||
151 | + uint8_t *ptr; | ||
152 | + uint32_t *data_ptr; | ||
153 | + unsigned type:2; | ||
154 | +} LDREntry; | ||
155 | + | ||
156 | +static LDREntry arm_ldr_table[1024]; | ||
157 | +static uint32_t arm_data_table[ARM_LDR_TABLE_SIZE]; | ||
158 | + | ||
159 | +extern char exec_loop; | ||
160 | + | ||
161 | +static inline void arm_reloc_pc24(uint32_t *ptr, uint32_t insn, int val) | ||
162 | +{ | ||
163 | + *ptr = (insn & ~0xffffff) | ((insn + ((val - (int)ptr) >> 2)) & 0xffffff); | ||
164 | +} | ||
165 | + | ||
166 | +static uint8_t *arm_flush_ldr(uint8_t *gen_code_ptr, | ||
167 | + LDREntry *ldr_start, LDREntry *ldr_end, | ||
168 | + uint32_t *data_start, uint32_t *data_end, | ||
169 | + int gen_jmp) | ||
170 | +{ | ||
171 | + LDREntry *le; | ||
172 | + uint32_t *ptr; | ||
173 | + int offset, data_size, target; | ||
174 | + uint8_t *data_ptr; | ||
175 | + uint32_t insn; | ||
176 | + uint32_t mask; | ||
177 | + | ||
178 | + data_size = (data_end - data_start) << 2; | ||
179 | + | ||
180 | + if (gen_jmp) { | ||
181 | + /* generate branch to skip the data */ | ||
182 | + if (data_size == 0) | ||
183 | + return gen_code_ptr; | ||
184 | + target = (long)gen_code_ptr + data_size + 4; | ||
185 | + arm_reloc_pc24((uint32_t *)gen_code_ptr, 0xeafffffe, target); | ||
186 | + gen_code_ptr += 4; | ||
187 | + } | ||
188 | + | ||
189 | + /* copy the data */ | ||
190 | + data_ptr = gen_code_ptr; | ||
191 | + memcpy(gen_code_ptr, data_start, data_size); | ||
192 | + gen_code_ptr += data_size; | ||
193 | + | ||
194 | + /* patch the ldr to point to the data */ | ||
195 | + for(le = ldr_start; le < ldr_end; le++) { | ||
196 | + ptr = (uint32_t *)le->ptr; | ||
197 | + offset = ((unsigned long)(le->data_ptr) - (unsigned long)data_start) + | ||
198 | + (unsigned long)data_ptr - | ||
199 | + (unsigned long)ptr - 8; | ||
200 | + if (offset < 0) { | ||
201 | + fprintf(stderr, "Negative constant pool offset\n"); | ||
202 | + tcg_abort(); | ||
203 | + } | ||
204 | + switch (le->type) { | ||
205 | + case 0: /* ldr */ | ||
206 | + mask = ~0x00800fff; | ||
207 | + if (offset >= 4096) { | ||
208 | + fprintf(stderr, "Bad ldr offset\n"); | ||
209 | + tcg_abort(); | ||
210 | + } | ||
211 | + break; | ||
212 | + case 1: /* ldc */ | ||
213 | + mask = ~0x008000ff; | ||
214 | + if (offset >= 1024 ) { | ||
215 | + fprintf(stderr, "Bad ldc offset\n"); | ||
216 | + tcg_abort(); | ||
217 | + } | ||
218 | + break; | ||
219 | + case 2: /* add */ | ||
220 | + mask = ~0xfff; | ||
221 | + if (offset >= 1024 ) { | ||
222 | + fprintf(stderr, "Bad add offset\n"); | ||
223 | + tcg_abort(); | ||
224 | + } | ||
225 | + break; | ||
226 | + default: | ||
227 | + fprintf(stderr, "Bad pc relative fixup\n"); | ||
228 | + tcg_abort(); | ||
229 | + } | ||
230 | + insn = *ptr & mask; | ||
231 | + switch (le->type) { | ||
232 | + case 0: /* ldr */ | ||
233 | + insn |= offset | 0x00800000; | ||
234 | + break; | ||
235 | + case 1: /* ldc */ | ||
236 | + insn |= (offset >> 2) | 0x00800000; | ||
237 | + break; | ||
238 | + case 2: /* add */ | ||
239 | + insn |= (offset >> 2) | 0xf00; | ||
240 | + break; | ||
241 | + } | ||
242 | + *ptr = insn; | ||
243 | + } | ||
244 | + return gen_code_ptr; | ||
245 | +} | ||
246 | + | ||
247 | +#endif /* __arm__ */ | ||
248 | + | ||
249 | +#ifdef __ia64 | ||
250 | + | ||
251 | +/* Patch instruction with "val" where "mask" has 1 bits. */ | ||
252 | +static inline void ia64_patch (uint64_t insn_addr, uint64_t mask, uint64_t val) | ||
253 | +{ | ||
254 | + uint64_t m0, m1, v0, v1, b0, b1, *b = (uint64_t *) (insn_addr & -16); | ||
255 | +# define insn_mask ((1UL << 41) - 1) | ||
256 | + unsigned long shift; | ||
257 | + | ||
258 | + b0 = b[0]; b1 = b[1]; | ||
259 | + shift = 5 + 41 * (insn_addr % 16); /* 5 template, 3 x 41-bit insns */ | ||
260 | + if (shift >= 64) { | ||
261 | + m1 = mask << (shift - 64); | ||
262 | + v1 = val << (shift - 64); | ||
263 | + } else { | ||
264 | + m0 = mask << shift; m1 = mask >> (64 - shift); | ||
265 | + v0 = val << shift; v1 = val >> (64 - shift); | ||
266 | + b[0] = (b0 & ~m0) | (v0 & m0); | ||
267 | + } | ||
268 | + b[1] = (b1 & ~m1) | (v1 & m1); | ||
269 | +} | ||
270 | + | ||
271 | +static inline void ia64_patch_imm60 (uint64_t insn_addr, uint64_t val) | ||
272 | +{ | ||
273 | + ia64_patch(insn_addr, | ||
274 | + 0x011ffffe000UL, | ||
275 | + ( ((val & 0x0800000000000000UL) >> 23) /* bit 59 -> 36 */ | ||
276 | + | ((val & 0x00000000000fffffUL) << 13) /* bit 0 -> 13 */)); | ||
277 | + ia64_patch(insn_addr - 1, 0x1fffffffffcUL, val >> 18); | ||
278 | +} | ||
279 | + | ||
280 | +static inline void ia64_imm64 (void *insn, uint64_t val) | ||
281 | +{ | ||
282 | + /* Ignore the slot number of the relocation; GCC and Intel | ||
283 | + toolchains differed for some time on whether IMM64 relocs are | ||
284 | + against slot 1 (Intel) or slot 2 (GCC). */ | ||
285 | + uint64_t insn_addr = (uint64_t) insn & ~3UL; | ||
286 | + | ||
287 | + ia64_patch(insn_addr + 2, | ||
288 | + 0x01fffefe000UL, | ||
289 | + ( ((val & 0x8000000000000000UL) >> 27) /* bit 63 -> 36 */ | ||
290 | + | ((val & 0x0000000000200000UL) << 0) /* bit 21 -> 21 */ | ||
291 | + | ((val & 0x00000000001f0000UL) << 6) /* bit 16 -> 22 */ | ||
292 | + | ((val & 0x000000000000ff80UL) << 20) /* bit 7 -> 27 */ | ||
293 | + | ((val & 0x000000000000007fUL) << 13) /* bit 0 -> 13 */) | ||
294 | + ); | ||
295 | + ia64_patch(insn_addr + 1, 0x1ffffffffffUL, val >> 22); | ||
296 | +} | ||
297 | + | ||
298 | +static inline void ia64_imm60b (void *insn, uint64_t val) | ||
299 | +{ | ||
300 | + /* Ignore the slot number of the relocation; GCC and Intel | ||
301 | + toolchains differed for some time on whether IMM64 relocs are | ||
302 | + against slot 1 (Intel) or slot 2 (GCC). */ | ||
303 | + uint64_t insn_addr = (uint64_t) insn & ~3UL; | ||
304 | + | ||
305 | + if (val + ((uint64_t) 1 << 59) >= (1UL << 60)) | ||
306 | + fprintf(stderr, "%s: value %ld out of IMM60 range\n", | ||
307 | + __FUNCTION__, (int64_t) val); | ||
308 | + ia64_patch_imm60(insn_addr + 2, val); | ||
309 | +} | ||
310 | + | ||
311 | +static inline void ia64_imm22 (void *insn, uint64_t val) | ||
312 | +{ | ||
313 | + if (val + (1 << 21) >= (1 << 22)) | ||
314 | + fprintf(stderr, "%s: value %li out of IMM22 range\n", | ||
315 | + __FUNCTION__, (int64_t)val); | ||
316 | + ia64_patch((uint64_t) insn, 0x01fffcfe000UL, | ||
317 | + ( ((val & 0x200000UL) << 15) /* bit 21 -> 36 */ | ||
318 | + | ((val & 0x1f0000UL) << 6) /* bit 16 -> 22 */ | ||
319 | + | ((val & 0x00ff80UL) << 20) /* bit 7 -> 27 */ | ||
320 | + | ((val & 0x00007fUL) << 13) /* bit 0 -> 13 */)); | ||
321 | +} | ||
322 | + | ||
323 | +/* Like ia64_imm22(), but also clear bits 20-21. For addl, this has | ||
324 | + the effect of turning "addl rX=imm22,rY" into "addl | ||
325 | + rX=imm22,r0". */ | ||
326 | +static inline void ia64_imm22_r0 (void *insn, uint64_t val) | ||
327 | +{ | ||
328 | + if (val + (1 << 21) >= (1 << 22)) | ||
329 | + fprintf(stderr, "%s: value %li out of IMM22 range\n", | ||
330 | + __FUNCTION__, (int64_t)val); | ||
331 | + ia64_patch((uint64_t) insn, 0x01fffcfe000UL | (0x3UL << 20), | ||
332 | + ( ((val & 0x200000UL) << 15) /* bit 21 -> 36 */ | ||
333 | + | ((val & 0x1f0000UL) << 6) /* bit 16 -> 22 */ | ||
334 | + | ((val & 0x00ff80UL) << 20) /* bit 7 -> 27 */ | ||
335 | + | ((val & 0x00007fUL) << 13) /* bit 0 -> 13 */)); | ||
336 | +} | ||
337 | + | ||
338 | +static inline void ia64_imm21b (void *insn, uint64_t val) | ||
339 | +{ | ||
340 | + if (val + (1 << 20) >= (1 << 21)) | ||
341 | + fprintf(stderr, "%s: value %li out of IMM21b range\n", | ||
342 | + __FUNCTION__, (int64_t)val); | ||
343 | + ia64_patch((uint64_t) insn, 0x11ffffe000UL, | ||
344 | + ( ((val & 0x100000UL) << 16) /* bit 20 -> 36 */ | ||
345 | + | ((val & 0x0fffffUL) << 13) /* bit 0 -> 13 */)); | ||
346 | +} | ||
347 | + | ||
348 | +static inline void ia64_nop_b (void *insn) | ||
349 | +{ | ||
350 | + ia64_patch((uint64_t) insn, (1UL << 41) - 1, 2UL << 37); | ||
351 | +} | ||
352 | + | ||
353 | +static inline void ia64_ldxmov(void *insn, uint64_t val) | ||
354 | +{ | ||
355 | + if (val + (1 << 21) < (1 << 22)) | ||
356 | + ia64_patch((uint64_t) insn, 0x1fff80fe000UL, 8UL << 37); | ||
357 | +} | ||
358 | + | ||
359 | +static inline int ia64_patch_ltoff(void *insn, uint64_t val, | ||
360 | + int relaxable) | ||
361 | +{ | ||
362 | + if (relaxable && (val + (1 << 21) < (1 << 22))) { | ||
363 | + ia64_imm22_r0(insn, val); | ||
364 | + return 0; | ||
365 | + } | ||
366 | + return 1; | ||
367 | +} | ||
368 | + | ||
369 | +struct ia64_fixup { | ||
370 | + struct ia64_fixup *next; | ||
371 | + void *addr; /* address that needs to be patched */ | ||
372 | + long value; | ||
373 | +}; | ||
374 | + | ||
375 | +#define IA64_PLT(insn, plt_index) \ | ||
376 | +do { \ | ||
377 | + struct ia64_fixup *fixup = alloca(sizeof(*fixup)); \ | ||
378 | + fixup->next = plt_fixes; \ | ||
379 | + plt_fixes = fixup; \ | ||
380 | + fixup->addr = (insn); \ | ||
381 | + fixup->value = (plt_index); \ | ||
382 | + plt_offset[(plt_index)] = 1; \ | ||
383 | +} while (0) | ||
384 | + | ||
385 | +#define IA64_LTOFF(insn, val, relaxable) \ | ||
386 | +do { \ | ||
387 | + if (ia64_patch_ltoff(insn, val, relaxable)) { \ | ||
388 | + struct ia64_fixup *fixup = alloca(sizeof(*fixup)); \ | ||
389 | + fixup->next = ltoff_fixes; \ | ||
390 | + ltoff_fixes = fixup; \ | ||
391 | + fixup->addr = (insn); \ | ||
392 | + fixup->value = (val); \ | ||
393 | + } \ | ||
394 | +} while (0) | ||
395 | + | ||
396 | +static inline void ia64_apply_fixes (uint8_t **gen_code_pp, | ||
397 | + struct ia64_fixup *ltoff_fixes, | ||
398 | + uint64_t gp, | ||
399 | + struct ia64_fixup *plt_fixes, | ||
400 | + int num_plts, | ||
401 | + unsigned long *plt_target, | ||
402 | + unsigned int *plt_offset) | ||
403 | +{ | ||
404 | + static const uint8_t plt_bundle[] = { | ||
405 | + 0x04, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, /* nop 0; movl r1=GP */ | ||
406 | + 0x00, 0x00, 0x00, 0x20, 0x00, 0x00, 0x00, 0x60, | ||
407 | + | ||
408 | + 0x05, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, /* nop 0; brl IP */ | ||
409 | + 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0xc0 | ||
410 | + }; | ||
411 | + uint8_t *gen_code_ptr = *gen_code_pp, *plt_start, *got_start; | ||
412 | + uint64_t *vp; | ||
413 | + struct ia64_fixup *fixup; | ||
414 | + unsigned int offset = 0; | ||
415 | + struct fdesc { | ||
416 | + long ip; | ||
417 | + long gp; | ||
418 | + } *fdesc; | ||
419 | + int i; | ||
420 | + | ||
421 | + if (plt_fixes) { | ||
422 | + plt_start = gen_code_ptr; | ||
423 | + | ||
424 | + for (i = 0; i < num_plts; ++i) { | ||
425 | + if (plt_offset[i]) { | ||
426 | + plt_offset[i] = offset; | ||
427 | + offset += sizeof(plt_bundle); | ||
428 | + | ||
429 | + fdesc = (struct fdesc *) plt_target[i]; | ||
430 | + memcpy(gen_code_ptr, plt_bundle, sizeof(plt_bundle)); | ||
431 | + ia64_imm64 (gen_code_ptr + 0x02, fdesc->gp); | ||
432 | + ia64_imm60b(gen_code_ptr + 0x12, | ||
433 | + (fdesc->ip - (long) (gen_code_ptr + 0x10)) >> 4); | ||
434 | + gen_code_ptr += sizeof(plt_bundle); | ||
435 | + } | ||
436 | + } | ||
437 | + | ||
438 | + for (fixup = plt_fixes; fixup; fixup = fixup->next) | ||
439 | + ia64_imm21b(fixup->addr, | ||
440 | + ((long) plt_start + plt_offset[fixup->value] | ||
441 | + - ((long) fixup->addr & ~0xf)) >> 4); | ||
442 | + } | ||
443 | + | ||
444 | + got_start = gen_code_ptr; | ||
445 | + | ||
446 | + /* First, create the GOT: */ | ||
447 | + for (fixup = ltoff_fixes; fixup; fixup = fixup->next) { | ||
448 | + /* first check if we already have this value in the GOT: */ | ||
449 | + for (vp = (uint64_t *) got_start; vp < (uint64_t *) gen_code_ptr; ++vp) | ||
450 | + if (*vp == fixup->value) | ||
451 | + break; | ||
452 | + if (vp == (uint64_t *) gen_code_ptr) { | ||
453 | + /* Nope, we need to put the value in the GOT: */ | ||
454 | + *vp = fixup->value; | ||
455 | + gen_code_ptr += 8; | ||
456 | + } | ||
457 | + ia64_imm22(fixup->addr, (long) vp - gp); | ||
458 | + } | ||
459 | + /* Keep code ptr aligned. */ | ||
460 | + if ((long) gen_code_ptr & 15) | ||
461 | + gen_code_ptr += 8; | ||
462 | + *gen_code_pp = gen_code_ptr; | ||
463 | +} | ||
464 | +#endif | ||
465 | +#endif | ||
466 | + | ||
467 | +const TCGArg *dyngen_op(TCGContext *s, int opc, const TCGArg *opparam_ptr) | ||
468 | +{ | ||
469 | + uint8_t *gen_code_ptr; | ||
470 | + | ||
471 | + gen_code_ptr = s->code_ptr; | ||
472 | + switch(opc) { | ||
473 | + | ||
474 | +/* op.h is dynamically generated by dyngen.c from op.c */ | ||
475 | +#include "op.h" | ||
476 | + | ||
477 | + default: | ||
478 | + tcg_abort(); | ||
479 | + } | ||
480 | + s->code_ptr = gen_code_ptr; | ||
481 | + return opparam_ptr; | ||
482 | +} |
tcg/tcg-op.h
0 โ 100644
1 | +/* | ||
2 | + * Tiny Code Generator for QEMU | ||
3 | + * | ||
4 | + * Copyright (c) 2008 Fabrice Bellard | ||
5 | + * | ||
6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
7 | + * of this software and associated documentation files (the "Software"), to deal | ||
8 | + * in the Software without restriction, including without limitation the rights | ||
9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
10 | + * copies of the Software, and to permit persons to whom the Software is | ||
11 | + * furnished to do so, subject to the following conditions: | ||
12 | + * | ||
13 | + * The above copyright notice and this permission notice shall be included in | ||
14 | + * all copies or substantial portions of the Software. | ||
15 | + * | ||
16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
22 | + * THE SOFTWARE. | ||
23 | + */ | ||
24 | +#include "tcg.h" | ||
25 | + | ||
26 | +/* legacy dyngen operations */ | ||
27 | +#include "gen-op.h" | ||
28 | + | ||
29 | +int gen_new_label(void); | ||
30 | + | ||
31 | +static inline void tcg_gen_op1(int opc, TCGArg arg1) | ||
32 | +{ | ||
33 | + *gen_opc_ptr++ = opc; | ||
34 | + *gen_opparam_ptr++ = arg1; | ||
35 | +} | ||
36 | + | ||
37 | +static inline void tcg_gen_op2(int opc, TCGArg arg1, TCGArg arg2) | ||
38 | +{ | ||
39 | + *gen_opc_ptr++ = opc; | ||
40 | + *gen_opparam_ptr++ = arg1; | ||
41 | + *gen_opparam_ptr++ = arg2; | ||
42 | +} | ||
43 | + | ||
44 | +static inline void tcg_gen_op3(int opc, TCGArg arg1, TCGArg arg2, TCGArg arg3) | ||
45 | +{ | ||
46 | + *gen_opc_ptr++ = opc; | ||
47 | + *gen_opparam_ptr++ = arg1; | ||
48 | + *gen_opparam_ptr++ = arg2; | ||
49 | + *gen_opparam_ptr++ = arg3; | ||
50 | +} | ||
51 | + | ||
52 | +static inline void tcg_gen_op4(int opc, TCGArg arg1, TCGArg arg2, TCGArg arg3, | ||
53 | + TCGArg arg4) | ||
54 | +{ | ||
55 | + *gen_opc_ptr++ = opc; | ||
56 | + *gen_opparam_ptr++ = arg1; | ||
57 | + *gen_opparam_ptr++ = arg2; | ||
58 | + *gen_opparam_ptr++ = arg3; | ||
59 | + *gen_opparam_ptr++ = arg4; | ||
60 | +} | ||
61 | + | ||
62 | +static inline void tcg_gen_op5(int opc, TCGArg arg1, TCGArg arg2, | ||
63 | + TCGArg arg3, TCGArg arg4, | ||
64 | + TCGArg arg5) | ||
65 | +{ | ||
66 | + *gen_opc_ptr++ = opc; | ||
67 | + *gen_opparam_ptr++ = arg1; | ||
68 | + *gen_opparam_ptr++ = arg2; | ||
69 | + *gen_opparam_ptr++ = arg3; | ||
70 | + *gen_opparam_ptr++ = arg4; | ||
71 | + *gen_opparam_ptr++ = arg5; | ||
72 | +} | ||
73 | + | ||
74 | +static inline void tcg_gen_op6(int opc, TCGArg arg1, TCGArg arg2, | ||
75 | + TCGArg arg3, TCGArg arg4, | ||
76 | + TCGArg arg5, TCGArg arg6) | ||
77 | +{ | ||
78 | + *gen_opc_ptr++ = opc; | ||
79 | + *gen_opparam_ptr++ = arg1; | ||
80 | + *gen_opparam_ptr++ = arg2; | ||
81 | + *gen_opparam_ptr++ = arg3; | ||
82 | + *gen_opparam_ptr++ = arg4; | ||
83 | + *gen_opparam_ptr++ = arg5; | ||
84 | + *gen_opparam_ptr++ = arg6; | ||
85 | +} | ||
86 | + | ||
87 | +static inline void gen_set_label(int n) | ||
88 | +{ | ||
89 | + tcg_gen_op1(INDEX_op_set_label, n); | ||
90 | +} | ||
91 | + | ||
92 | +static inline void tcg_gen_mov_i32(int ret, int arg) | ||
93 | +{ | ||
94 | + tcg_gen_op2(INDEX_op_mov_i32, ret, arg); | ||
95 | +} | ||
96 | + | ||
97 | +static inline void tcg_gen_movi_i32(int ret, int32_t arg) | ||
98 | +{ | ||
99 | + tcg_gen_op2(INDEX_op_movi_i32, ret, arg); | ||
100 | +} | ||
101 | + | ||
102 | +/* helper calls */ | ||
103 | +#define TCG_HELPER_CALL_FLAGS 0 | ||
104 | + | ||
105 | +static inline void tcg_gen_helper_0_0(void *func) | ||
106 | +{ | ||
107 | + tcg_gen_call(&tcg_ctx, | ||
108 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | ||
109 | + 0, NULL, 0, NULL); | ||
110 | +} | ||
111 | + | ||
112 | +static inline void tcg_gen_helper_0_1(void *func, TCGArg arg) | ||
113 | +{ | ||
114 | + tcg_gen_call(&tcg_ctx, | ||
115 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | ||
116 | + 0, NULL, 1, &arg); | ||
117 | +} | ||
118 | + | ||
119 | +static inline void tcg_gen_helper_0_2(void *func, TCGArg arg1, TCGArg arg2) | ||
120 | +{ | ||
121 | + TCGArg args[2]; | ||
122 | + args[0] = arg1; | ||
123 | + args[1] = arg2; | ||
124 | + tcg_gen_call(&tcg_ctx, | ||
125 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | ||
126 | + 0, NULL, 2, args); | ||
127 | +} | ||
128 | + | ||
129 | +static inline void tcg_gen_helper_1_2(void *func, TCGArg ret, | ||
130 | + TCGArg arg1, TCGArg arg2) | ||
131 | +{ | ||
132 | + TCGArg args[2]; | ||
133 | + args[0] = arg1; | ||
134 | + args[1] = arg2; | ||
135 | + tcg_gen_call(&tcg_ctx, | ||
136 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | ||
137 | + 1, &ret, 2, args); | ||
138 | +} | ||
139 | + | ||
140 | +/* 32 bit ops */ | ||
141 | + | ||
142 | +static inline void tcg_gen_ld8u_i32(int ret, int arg2, tcg_target_long offset) | ||
143 | +{ | ||
144 | + tcg_gen_op3(INDEX_op_ld8u_i32, ret, arg2, offset); | ||
145 | +} | ||
146 | + | ||
147 | +static inline void tcg_gen_ld8s_i32(int ret, int arg2, tcg_target_long offset) | ||
148 | +{ | ||
149 | + tcg_gen_op3(INDEX_op_ld8s_i32, ret, arg2, offset); | ||
150 | +} | ||
151 | + | ||
152 | +static inline void tcg_gen_ld16u_i32(int ret, int arg2, tcg_target_long offset) | ||
153 | +{ | ||
154 | + tcg_gen_op3(INDEX_op_ld16u_i32, ret, arg2, offset); | ||
155 | +} | ||
156 | + | ||
157 | +static inline void tcg_gen_ld16s_i32(int ret, int arg2, tcg_target_long offset) | ||
158 | +{ | ||
159 | + tcg_gen_op3(INDEX_op_ld16s_i32, ret, arg2, offset); | ||
160 | +} | ||
161 | + | ||
162 | +static inline void tcg_gen_ld_i32(int ret, int arg2, tcg_target_long offset) | ||
163 | +{ | ||
164 | + tcg_gen_op3(INDEX_op_ld_i32, ret, arg2, offset); | ||
165 | +} | ||
166 | + | ||
167 | +static inline void tcg_gen_st8_i32(int arg1, int arg2, tcg_target_long offset) | ||
168 | +{ | ||
169 | + tcg_gen_op3(INDEX_op_st8_i32, arg1, arg2, offset); | ||
170 | +} | ||
171 | + | ||
172 | +static inline void tcg_gen_st16_i32(int arg1, int arg2, tcg_target_long offset) | ||
173 | +{ | ||
174 | + tcg_gen_op3(INDEX_op_st16_i32, arg1, arg2, offset); | ||
175 | +} | ||
176 | + | ||
177 | +static inline void tcg_gen_st_i32(int arg1, int arg2, tcg_target_long offset) | ||
178 | +{ | ||
179 | + tcg_gen_op3(INDEX_op_st_i32, arg1, arg2, offset); | ||
180 | +} | ||
181 | + | ||
182 | +static inline void tcg_gen_add_i32(int ret, int arg1, int arg2) | ||
183 | +{ | ||
184 | + tcg_gen_op3(INDEX_op_add_i32, ret, arg1, arg2); | ||
185 | +} | ||
186 | + | ||
187 | +static inline void tcg_gen_addi_i32(int ret, int arg1, int32_t arg2) | ||
188 | +{ | ||
189 | + tcg_gen_add_i32(ret, arg1, tcg_const_i32(arg2)); | ||
190 | +} | ||
191 | + | ||
192 | +static inline void tcg_gen_sub_i32(int ret, int arg1, int arg2) | ||
193 | +{ | ||
194 | + tcg_gen_op3(INDEX_op_sub_i32, ret, arg1, arg2); | ||
195 | +} | ||
196 | + | ||
197 | +static inline void tcg_gen_subi_i32(int ret, int arg1, int32_t arg2) | ||
198 | +{ | ||
199 | + tcg_gen_sub_i32(ret, arg1, tcg_const_i32(arg2)); | ||
200 | +} | ||
201 | + | ||
202 | +static inline void tcg_gen_and_i32(int ret, int arg1, int arg2) | ||
203 | +{ | ||
204 | + tcg_gen_op3(INDEX_op_and_i32, ret, arg1, arg2); | ||
205 | +} | ||
206 | + | ||
207 | +static inline void tcg_gen_andi_i32(int ret, int arg1, int32_t arg2) | ||
208 | +{ | ||
209 | + /* some cases can be optimized here */ | ||
210 | + if (arg2 == 0) { | ||
211 | + tcg_gen_movi_i32(ret, 0); | ||
212 | + } else if (arg2 == 0xffffffff) { | ||
213 | + tcg_gen_mov_i32(ret, arg1); | ||
214 | + } else { | ||
215 | + tcg_gen_and_i32(ret, arg1, tcg_const_i32(arg2)); | ||
216 | + } | ||
217 | +} | ||
218 | + | ||
219 | +static inline void tcg_gen_or_i32(int ret, int arg1, int arg2) | ||
220 | +{ | ||
221 | + tcg_gen_op3(INDEX_op_or_i32, ret, arg1, arg2); | ||
222 | +} | ||
223 | + | ||
224 | +static inline void tcg_gen_ori_i32(int ret, int arg1, int32_t arg2) | ||
225 | +{ | ||
226 | + /* some cases can be optimized here */ | ||
227 | + if (arg2 == 0xffffffff) { | ||
228 | + tcg_gen_movi_i32(ret, 0); | ||
229 | + } else if (arg2 == 0) { | ||
230 | + tcg_gen_mov_i32(ret, arg1); | ||
231 | + } else { | ||
232 | + tcg_gen_or_i32(ret, arg1, tcg_const_i32(arg2)); | ||
233 | + } | ||
234 | +} | ||
235 | + | ||
236 | +static inline void tcg_gen_xor_i32(int ret, int arg1, int arg2) | ||
237 | +{ | ||
238 | + tcg_gen_op3(INDEX_op_xor_i32, ret, arg1, arg2); | ||
239 | +} | ||
240 | + | ||
241 | +static inline void tcg_gen_xori_i32(int ret, int arg1, int32_t arg2) | ||
242 | +{ | ||
243 | + /* some cases can be optimized here */ | ||
244 | + if (arg2 == 0) { | ||
245 | + tcg_gen_mov_i32(ret, arg1); | ||
246 | + } else { | ||
247 | + tcg_gen_xor_i32(ret, arg1, tcg_const_i32(arg2)); | ||
248 | + } | ||
249 | +} | ||
250 | + | ||
251 | +static inline void tcg_gen_shl_i32(int ret, int arg1, int arg2) | ||
252 | +{ | ||
253 | + tcg_gen_op3(INDEX_op_shl_i32, ret, arg1, arg2); | ||
254 | +} | ||
255 | + | ||
256 | +static inline void tcg_gen_shli_i32(int ret, int arg1, int32_t arg2) | ||
257 | +{ | ||
258 | + tcg_gen_shl_i32(ret, arg1, tcg_const_i32(arg2)); | ||
259 | +} | ||
260 | + | ||
261 | +static inline void tcg_gen_shr_i32(int ret, int arg1, int arg2) | ||
262 | +{ | ||
263 | + tcg_gen_op3(INDEX_op_shr_i32, ret, arg1, arg2); | ||
264 | +} | ||
265 | + | ||
266 | +static inline void tcg_gen_shri_i32(int ret, int arg1, int32_t arg2) | ||
267 | +{ | ||
268 | + tcg_gen_shr_i32(ret, arg1, tcg_const_i32(arg2)); | ||
269 | +} | ||
270 | + | ||
271 | +static inline void tcg_gen_sar_i32(int ret, int arg1, int arg2) | ||
272 | +{ | ||
273 | + tcg_gen_op3(INDEX_op_sar_i32, ret, arg1, arg2); | ||
274 | +} | ||
275 | + | ||
276 | +static inline void tcg_gen_sari_i32(int ret, int arg1, int32_t arg2) | ||
277 | +{ | ||
278 | + tcg_gen_sar_i32(ret, arg1, tcg_const_i32(arg2)); | ||
279 | +} | ||
280 | + | ||
281 | +static inline void tcg_gen_brcond_i32(int cond, TCGArg arg1, TCGArg arg2, | ||
282 | + int label_index) | ||
283 | +{ | ||
284 | + tcg_gen_op4(INDEX_op_brcond_i32, arg1, arg2, cond, label_index); | ||
285 | +} | ||
286 | + | ||
287 | +static inline void tcg_gen_mul_i32(int ret, int arg1, int arg2) | ||
288 | +{ | ||
289 | + tcg_gen_op3(INDEX_op_mul_i32, ret, arg1, arg2); | ||
290 | +} | ||
291 | + | ||
292 | +#ifdef TCG_TARGET_HAS_div_i32 | ||
293 | +static inline void tcg_gen_div_i32(int ret, int arg1, int arg2) | ||
294 | +{ | ||
295 | + tcg_gen_op3(INDEX_op_div_i32, ret, arg1, arg2); | ||
296 | +} | ||
297 | + | ||
298 | +static inline void tcg_gen_rem_i32(int ret, int arg1, int arg2) | ||
299 | +{ | ||
300 | + tcg_gen_op3(INDEX_op_rem_i32, ret, arg1, arg2); | ||
301 | +} | ||
302 | + | ||
303 | +static inline void tcg_gen_divu_i32(int ret, int arg1, int arg2) | ||
304 | +{ | ||
305 | + tcg_gen_op3(INDEX_op_divu_i32, ret, arg1, arg2); | ||
306 | +} | ||
307 | + | ||
308 | +static inline void tcg_gen_remu_i32(int ret, int arg1, int arg2) | ||
309 | +{ | ||
310 | + tcg_gen_op3(INDEX_op_remu_i32, ret, arg1, arg2); | ||
311 | +} | ||
312 | +#else | ||
313 | +static inline void tcg_gen_div_i32(int ret, int arg1, int arg2) | ||
314 | +{ | ||
315 | + int t0; | ||
316 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
317 | + tcg_gen_sari_i32(t0, arg1, 31); | ||
318 | + tcg_gen_op5(INDEX_op_div2_i32, ret, t0, arg1, t0, arg2); | ||
319 | +} | ||
320 | + | ||
321 | +static inline void tcg_gen_rem_i32(int ret, int arg1, int arg2) | ||
322 | +{ | ||
323 | + int t0; | ||
324 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
325 | + tcg_gen_sari_i32(t0, arg1, 31); | ||
326 | + tcg_gen_op5(INDEX_op_div2_i32, t0, ret, arg1, t0, arg2); | ||
327 | +} | ||
328 | + | ||
329 | +static inline void tcg_gen_divu_i32(int ret, int arg1, int arg2) | ||
330 | +{ | ||
331 | + int t0; | ||
332 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
333 | + tcg_gen_movi_i32(t0, 0); | ||
334 | + tcg_gen_op5(INDEX_op_divu2_i32, ret, t0, arg1, t0, arg2); | ||
335 | +} | ||
336 | + | ||
337 | +static inline void tcg_gen_remu_i32(int ret, int arg1, int arg2) | ||
338 | +{ | ||
339 | + int t0; | ||
340 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
341 | + tcg_gen_movi_i32(t0, 0); | ||
342 | + tcg_gen_op5(INDEX_op_divu2_i32, t0, ret, arg1, t0, arg2); | ||
343 | +} | ||
344 | +#endif | ||
345 | + | ||
346 | +#if TCG_TARGET_REG_BITS == 32 | ||
347 | + | ||
348 | +static inline void tcg_gen_mov_i64(int ret, int arg) | ||
349 | +{ | ||
350 | + tcg_gen_mov_i32(ret, arg); | ||
351 | + tcg_gen_mov_i32(ret + 1, arg + 1); | ||
352 | +} | ||
353 | + | ||
354 | +static inline void tcg_gen_movi_i64(int ret, int64_t arg) | ||
355 | +{ | ||
356 | + tcg_gen_movi_i32(ret, arg); | ||
357 | + tcg_gen_movi_i32(ret + 1, arg >> 32); | ||
358 | +} | ||
359 | + | ||
360 | +static inline void tcg_gen_ld8u_i64(int ret, int arg2, tcg_target_long offset) | ||
361 | +{ | ||
362 | + tcg_gen_ld8u_i32(ret, arg2, offset); | ||
363 | + tcg_gen_movi_i32(ret + 1, 0); | ||
364 | +} | ||
365 | + | ||
366 | +static inline void tcg_gen_ld8s_i64(int ret, int arg2, tcg_target_long offset) | ||
367 | +{ | ||
368 | + tcg_gen_ld8s_i32(ret, arg2, offset); | ||
369 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
370 | +} | ||
371 | + | ||
372 | +static inline void tcg_gen_ld16u_i64(int ret, int arg2, tcg_target_long offset) | ||
373 | +{ | ||
374 | + tcg_gen_ld16u_i32(ret, arg2, offset); | ||
375 | + tcg_gen_movi_i32(ret + 1, 0); | ||
376 | +} | ||
377 | + | ||
378 | +static inline void tcg_gen_ld16s_i64(int ret, int arg2, tcg_target_long offset) | ||
379 | +{ | ||
380 | + tcg_gen_ld16s_i32(ret, arg2, offset); | ||
381 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
382 | +} | ||
383 | + | ||
384 | +static inline void tcg_gen_ld32u_i64(int ret, int arg2, tcg_target_long offset) | ||
385 | +{ | ||
386 | + tcg_gen_ld_i32(ret, arg2, offset); | ||
387 | + tcg_gen_movi_i32(ret + 1, 0); | ||
388 | +} | ||
389 | + | ||
390 | +static inline void tcg_gen_ld32s_i64(int ret, int arg2, tcg_target_long offset) | ||
391 | +{ | ||
392 | + tcg_gen_ld_i32(ret, arg2, offset); | ||
393 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
394 | +} | ||
395 | + | ||
396 | +static inline void tcg_gen_ld_i64(int ret, int arg2, tcg_target_long offset) | ||
397 | +{ | ||
398 | + /* since arg2 and ret have different types, they cannot be the | ||
399 | + same temporary */ | ||
400 | +#ifdef TCG_TARGET_WORDS_BIGENDIAN | ||
401 | + tcg_gen_ld_i32(ret + 1, arg2, offset); | ||
402 | + tcg_gen_ld_i32(ret, arg2, offset + 4); | ||
403 | +#else | ||
404 | + tcg_gen_ld_i32(ret, arg2, offset); | ||
405 | + tcg_gen_ld_i32(ret + 1, arg2, offset + 4); | ||
406 | +#endif | ||
407 | +} | ||
408 | + | ||
409 | +static inline void tcg_gen_st8_i64(int arg1, int arg2, tcg_target_long offset) | ||
410 | +{ | ||
411 | + tcg_gen_st8_i32(arg1, arg2, offset); | ||
412 | +} | ||
413 | + | ||
414 | +static inline void tcg_gen_st16_i64(int arg1, int arg2, tcg_target_long offset) | ||
415 | +{ | ||
416 | + tcg_gen_st16_i32(arg1, arg2, offset); | ||
417 | +} | ||
418 | + | ||
419 | +static inline void tcg_gen_st32_i64(int arg1, int arg2, tcg_target_long offset) | ||
420 | +{ | ||
421 | + tcg_gen_st_i32(arg1, arg2, offset); | ||
422 | +} | ||
423 | + | ||
424 | +static inline void tcg_gen_st_i64(int arg1, int arg2, tcg_target_long offset) | ||
425 | +{ | ||
426 | +#ifdef TCG_TARGET_WORDS_BIGENDIAN | ||
427 | + tcg_gen_st_i32(arg1 + 1, arg2, offset); | ||
428 | + tcg_gen_st_i32(arg1, arg2, offset + 4); | ||
429 | +#else | ||
430 | + tcg_gen_st_i32(arg1, arg2, offset); | ||
431 | + tcg_gen_st_i32(arg1 + 1, arg2, offset + 4); | ||
432 | +#endif | ||
433 | +} | ||
434 | + | ||
435 | +static inline void tcg_gen_add_i64(int ret, int arg1, int arg2) | ||
436 | +{ | ||
437 | + tcg_gen_op6(INDEX_op_add2_i32, ret, ret + 1, | ||
438 | + arg1, arg1 + 1, arg2, arg2 + 1); | ||
439 | +} | ||
440 | + | ||
441 | +static inline void tcg_gen_addi_i64(int ret, int arg1, int64_t arg2) | ||
442 | +{ | ||
443 | + tcg_gen_add_i64(ret, arg1, tcg_const_i64(arg2)); | ||
444 | +} | ||
445 | + | ||
446 | +static inline void tcg_gen_sub_i64(int ret, int arg1, int arg2) | ||
447 | +{ | ||
448 | + tcg_gen_op6(INDEX_op_sub2_i32, ret, ret + 1, | ||
449 | + arg1, arg1 + 1, arg2, arg2 + 1); | ||
450 | +} | ||
451 | + | ||
452 | +static inline void tcg_gen_subi_i64(int ret, int arg1, int64_t arg2) | ||
453 | +{ | ||
454 | + tcg_gen_sub_i64(ret, arg1, tcg_const_i64(arg2)); | ||
455 | +} | ||
456 | + | ||
457 | +static inline void tcg_gen_and_i64(int ret, int arg1, int arg2) | ||
458 | +{ | ||
459 | + tcg_gen_and_i32(ret, arg1, arg2); | ||
460 | + tcg_gen_and_i32(ret + 1, arg1 + 1, arg2 + 1); | ||
461 | +} | ||
462 | + | ||
463 | +static inline void tcg_gen_andi_i64(int ret, int arg1, int64_t arg2) | ||
464 | +{ | ||
465 | + tcg_gen_andi_i32(ret, arg1, arg2); | ||
466 | + tcg_gen_andi_i32(ret + 1, arg1 + 1, arg2 >> 32); | ||
467 | +} | ||
468 | + | ||
469 | +static inline void tcg_gen_or_i64(int ret, int arg1, int arg2) | ||
470 | +{ | ||
471 | + tcg_gen_or_i32(ret, arg1, arg2); | ||
472 | + tcg_gen_or_i32(ret + 1, arg1 + 1, arg2 + 1); | ||
473 | +} | ||
474 | + | ||
475 | +static inline void tcg_gen_ori_i64(int ret, int arg1, int64_t arg2) | ||
476 | +{ | ||
477 | + tcg_gen_ori_i32(ret, arg1, arg2); | ||
478 | + tcg_gen_ori_i32(ret + 1, arg1 + 1, arg2 >> 32); | ||
479 | +} | ||
480 | + | ||
481 | +static inline void tcg_gen_xor_i64(int ret, int arg1, int arg2) | ||
482 | +{ | ||
483 | + tcg_gen_xor_i32(ret, arg1, arg2); | ||
484 | + tcg_gen_xor_i32(ret + 1, arg1 + 1, arg2 + 1); | ||
485 | +} | ||
486 | + | ||
487 | +static inline void tcg_gen_xori_i64(int ret, int arg1, int64_t arg2) | ||
488 | +{ | ||
489 | + tcg_gen_xori_i32(ret, arg1, arg2); | ||
490 | + tcg_gen_xori_i32(ret + 1, arg1 + 1, arg2 >> 32); | ||
491 | +} | ||
492 | + | ||
493 | +/* XXX: use generic code when basic block handling is OK or CPU | ||
494 | + specific code (x86) */ | ||
495 | +static inline void tcg_gen_shl_i64(int ret, int arg1, int64_t arg2) | ||
496 | +{ | ||
497 | + tcg_gen_helper_1_2(tcg_helper_shl_i64, ret, arg1, arg2); | ||
498 | +} | ||
499 | + | ||
500 | +static inline void tcg_gen_shli_i64(int ret, int arg1, int64_t arg2) | ||
501 | +{ | ||
502 | + tcg_gen_shifti_i64(ret, arg1, arg2, 0, 0); | ||
503 | +} | ||
504 | + | ||
505 | +static inline void tcg_gen_shr_i64(int ret, int arg1, int64_t arg2) | ||
506 | +{ | ||
507 | + tcg_gen_helper_1_2(tcg_helper_shr_i64, ret, arg1, arg2); | ||
508 | +} | ||
509 | + | ||
510 | +static inline void tcg_gen_shri_i64(int ret, int arg1, int64_t arg2) | ||
511 | +{ | ||
512 | + tcg_gen_shifti_i64(ret, arg1, arg2, 1, 0); | ||
513 | +} | ||
514 | + | ||
515 | +static inline void tcg_gen_sar_i64(int ret, int arg1, int64_t arg2) | ||
516 | +{ | ||
517 | + tcg_gen_helper_1_2(tcg_helper_sar_i64, ret, arg1, arg2); | ||
518 | +} | ||
519 | + | ||
520 | +static inline void tcg_gen_sari_i64(int ret, int arg1, int64_t arg2) | ||
521 | +{ | ||
522 | + tcg_gen_shifti_i64(ret, arg1, arg2, 1, 1); | ||
523 | +} | ||
524 | + | ||
525 | +static inline void tcg_gen_brcond_i64(int cond, TCGArg arg1, TCGArg arg2, | ||
526 | + int label_index) | ||
527 | +{ | ||
528 | + tcg_gen_op6(INDEX_op_brcond2_i32, | ||
529 | + arg1, arg1 + 1, arg2, arg2 + 1, cond, label_index); | ||
530 | +} | ||
531 | + | ||
532 | +static inline void tcg_gen_mul_i64(int ret, int arg1, int arg2) | ||
533 | +{ | ||
534 | + int t0, t1; | ||
535 | + | ||
536 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
537 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
538 | + | ||
539 | + tcg_gen_op4(INDEX_op_mulu2_i32, t0, t0 + 1, arg1, arg2); | ||
540 | + | ||
541 | + tcg_gen_mul_i32(t1, arg1, arg2 + 1); | ||
542 | + tcg_gen_add_i32(t0 + 1, t0 + 1, t1); | ||
543 | + tcg_gen_mul_i32(t1, arg1 + 1, arg2); | ||
544 | + tcg_gen_add_i32(t0 + 1, t0 + 1, t1); | ||
545 | + | ||
546 | + tcg_gen_mov_i64(ret, t0); | ||
547 | +} | ||
548 | + | ||
549 | +static inline void tcg_gen_div_i64(int ret, int arg1, int64_t arg2) | ||
550 | +{ | ||
551 | + tcg_gen_helper_1_2(tcg_helper_div_i64, ret, arg1, arg2); | ||
552 | +} | ||
553 | + | ||
554 | +static inline void tcg_gen_rem_i64(int ret, int arg1, int64_t arg2) | ||
555 | +{ | ||
556 | + tcg_gen_helper_1_2(tcg_helper_rem_i64, ret, arg1, arg2); | ||
557 | +} | ||
558 | + | ||
559 | +static inline void tcg_gen_divu_i64(int ret, int arg1, int64_t arg2) | ||
560 | +{ | ||
561 | + tcg_gen_helper_1_2(tcg_helper_divu_i64, ret, arg1, arg2); | ||
562 | +} | ||
563 | + | ||
564 | +static inline void tcg_gen_remu_i64(int ret, int arg1, int64_t arg2) | ||
565 | +{ | ||
566 | + tcg_gen_helper_1_2(tcg_helper_remu_i64, ret, arg1, arg2); | ||
567 | +} | ||
568 | + | ||
569 | +#else | ||
570 | + | ||
571 | +static inline void tcg_gen_mov_i64(int ret, int arg) | ||
572 | +{ | ||
573 | + tcg_gen_op2(INDEX_op_mov_i64, ret, arg); | ||
574 | +} | ||
575 | + | ||
576 | +static inline void tcg_gen_movi_i64(int ret, int64_t arg) | ||
577 | +{ | ||
578 | + tcg_gen_op2(INDEX_op_movi_i64, ret, arg); | ||
579 | +} | ||
580 | + | ||
581 | +static inline void tcg_gen_ld8u_i64(int ret, int arg2, tcg_target_long offset) | ||
582 | +{ | ||
583 | + tcg_gen_op3(INDEX_op_ld8u_i64, ret, arg2, offset); | ||
584 | +} | ||
585 | + | ||
586 | +static inline void tcg_gen_ld8s_i64(int ret, int arg2, tcg_target_long offset) | ||
587 | +{ | ||
588 | + tcg_gen_op3(INDEX_op_ld8s_i64, ret, arg2, offset); | ||
589 | +} | ||
590 | + | ||
591 | +static inline void tcg_gen_ld16u_i64(int ret, int arg2, tcg_target_long offset) | ||
592 | +{ | ||
593 | + tcg_gen_op3(INDEX_op_ld16u_i64, ret, arg2, offset); | ||
594 | +} | ||
595 | + | ||
596 | +static inline void tcg_gen_ld16s_i64(int ret, int arg2, tcg_target_long offset) | ||
597 | +{ | ||
598 | + tcg_gen_op3(INDEX_op_ld16s_i64, ret, arg2, offset); | ||
599 | +} | ||
600 | + | ||
601 | +static inline void tcg_gen_ld32u_i64(int ret, int arg2, tcg_target_long offset) | ||
602 | +{ | ||
603 | + tcg_gen_op3(INDEX_op_ld32u_i64, ret, arg2, offset); | ||
604 | +} | ||
605 | + | ||
606 | +static inline void tcg_gen_ld32s_i64(int ret, int arg2, tcg_target_long offset) | ||
607 | +{ | ||
608 | + tcg_gen_op3(INDEX_op_ld32s_i64, ret, arg2, offset); | ||
609 | +} | ||
610 | + | ||
611 | +static inline void tcg_gen_ld_i64(int ret, int arg2, tcg_target_long offset) | ||
612 | +{ | ||
613 | + tcg_gen_op3(INDEX_op_ld_i64, ret, arg2, offset); | ||
614 | +} | ||
615 | + | ||
616 | +static inline void tcg_gen_st8_i64(int arg1, int arg2, tcg_target_long offset) | ||
617 | +{ | ||
618 | + tcg_gen_op3(INDEX_op_st8_i64, arg1, arg2, offset); | ||
619 | +} | ||
620 | + | ||
621 | +static inline void tcg_gen_st16_i64(int arg1, int arg2, tcg_target_long offset) | ||
622 | +{ | ||
623 | + tcg_gen_op3(INDEX_op_st16_i64, arg1, arg2, offset); | ||
624 | +} | ||
625 | + | ||
626 | +static inline void tcg_gen_st32_i64(int arg1, int arg2, tcg_target_long offset) | ||
627 | +{ | ||
628 | + tcg_gen_op3(INDEX_op_st32_i64, arg1, arg2, offset); | ||
629 | +} | ||
630 | + | ||
631 | +static inline void tcg_gen_st_i64(int arg1, int arg2, tcg_target_long offset) | ||
632 | +{ | ||
633 | + tcg_gen_op3(INDEX_op_st_i64, arg1, arg2, offset); | ||
634 | +} | ||
635 | + | ||
636 | +static inline void tcg_gen_add_i64(int ret, int arg1, int arg2) | ||
637 | +{ | ||
638 | + tcg_gen_op3(INDEX_op_add_i64, ret, arg1, arg2); | ||
639 | +} | ||
640 | + | ||
641 | +static inline void tcg_gen_addi_i64(int ret, int arg1, int64_t arg2) | ||
642 | +{ | ||
643 | + tcg_gen_add_i64(ret, arg1, tcg_const_i64(arg2)); | ||
644 | +} | ||
645 | + | ||
646 | +static inline void tcg_gen_sub_i64(int ret, int arg1, int arg2) | ||
647 | +{ | ||
648 | + tcg_gen_op3(INDEX_op_sub_i64, ret, arg1, arg2); | ||
649 | +} | ||
650 | + | ||
651 | +static inline void tcg_gen_subi_i64(int ret, int arg1, int64_t arg2) | ||
652 | +{ | ||
653 | + tcg_gen_sub_i64(ret, arg1, tcg_const_i64(arg2)); | ||
654 | +} | ||
655 | + | ||
656 | +static inline void tcg_gen_and_i64(int ret, int arg1, int arg2) | ||
657 | +{ | ||
658 | + tcg_gen_op3(INDEX_op_and_i64, ret, arg1, arg2); | ||
659 | +} | ||
660 | + | ||
661 | +static inline void tcg_gen_andi_i64(int ret, int arg1, int64_t arg2) | ||
662 | +{ | ||
663 | + tcg_gen_and_i64(ret, arg1, tcg_const_i64(arg2)); | ||
664 | +} | ||
665 | + | ||
666 | +static inline void tcg_gen_or_i64(int ret, int arg1, int arg2) | ||
667 | +{ | ||
668 | + tcg_gen_op3(INDEX_op_or_i64, ret, arg1, arg2); | ||
669 | +} | ||
670 | + | ||
671 | +static inline void tcg_gen_ori_i64(int ret, int arg1, int64_t arg2) | ||
672 | +{ | ||
673 | + tcg_gen_or_i64(ret, arg1, tcg_const_i64(arg2)); | ||
674 | +} | ||
675 | + | ||
676 | +static inline void tcg_gen_xor_i64(int ret, int arg1, int arg2) | ||
677 | +{ | ||
678 | + tcg_gen_op3(INDEX_op_xor_i64, ret, arg1, arg2); | ||
679 | +} | ||
680 | + | ||
681 | +static inline void tcg_gen_xori_i64(int ret, int arg1, int64_t arg2) | ||
682 | +{ | ||
683 | + tcg_gen_xor_i64(ret, arg1, tcg_const_i64(arg2)); | ||
684 | +} | ||
685 | + | ||
686 | +static inline void tcg_gen_shl_i64(int ret, int arg1, int arg2) | ||
687 | +{ | ||
688 | + tcg_gen_op3(INDEX_op_shl_i64, ret, arg1, arg2); | ||
689 | +} | ||
690 | + | ||
691 | +static inline void tcg_gen_shli_i64(int ret, int arg1, int64_t arg2) | ||
692 | +{ | ||
693 | + tcg_gen_shl_i64(ret, arg1, tcg_const_i64(arg2)); | ||
694 | +} | ||
695 | + | ||
696 | +static inline void tcg_gen_shr_i64(int ret, int arg1, int arg2) | ||
697 | +{ | ||
698 | + tcg_gen_op3(INDEX_op_shr_i64, ret, arg1, arg2); | ||
699 | +} | ||
700 | + | ||
701 | +static inline void tcg_gen_shri_i64(int ret, int arg1, int64_t arg2) | ||
702 | +{ | ||
703 | + tcg_gen_shr_i64(ret, arg1, tcg_const_i64(arg2)); | ||
704 | +} | ||
705 | + | ||
706 | +static inline void tcg_gen_sar_i64(int ret, int arg1, int arg2) | ||
707 | +{ | ||
708 | + tcg_gen_op3(INDEX_op_sar_i64, ret, arg1, arg2); | ||
709 | +} | ||
710 | + | ||
711 | +static inline void tcg_gen_sari_i64(int ret, int arg1, int64_t arg2) | ||
712 | +{ | ||
713 | + tcg_gen_sar_i64(ret, arg1, tcg_const_i64(arg2)); | ||
714 | +} | ||
715 | + | ||
716 | +static inline void tcg_gen_brcond_i64(int cond, TCGArg arg1, TCGArg arg2, | ||
717 | + int label_index) | ||
718 | +{ | ||
719 | + tcg_gen_op4(INDEX_op_brcond_i64, arg1, arg2, cond, label_index); | ||
720 | +} | ||
721 | + | ||
722 | +static inline void tcg_gen_mul_i64(int ret, int arg1, int arg2) | ||
723 | +{ | ||
724 | + tcg_gen_op3(INDEX_op_mul_i64, ret, arg1, arg2); | ||
725 | +} | ||
726 | + | ||
727 | +#ifdef TCG_TARGET_HAS_div_i64 | ||
728 | +static inline void tcg_gen_div_i64(int ret, int arg1, int arg2) | ||
729 | +{ | ||
730 | + tcg_gen_op3(INDEX_op_div_i64, ret, arg1, arg2); | ||
731 | +} | ||
732 | + | ||
733 | +static inline void tcg_gen_rem_i64(int ret, int arg1, int arg2) | ||
734 | +{ | ||
735 | + tcg_gen_op3(INDEX_op_rem_i64, ret, arg1, arg2); | ||
736 | +} | ||
737 | + | ||
738 | +static inline void tcg_gen_divu_i64(int ret, int arg1, int arg2) | ||
739 | +{ | ||
740 | + tcg_gen_op3(INDEX_op_divu_i64, ret, arg1, arg2); | ||
741 | +} | ||
742 | + | ||
743 | +static inline void tcg_gen_remu_i64(int ret, int arg1, int arg2) | ||
744 | +{ | ||
745 | + tcg_gen_op3(INDEX_op_remu_i64, ret, arg1, arg2); | ||
746 | +} | ||
747 | +#else | ||
748 | +static inline void tcg_gen_div_i64(int ret, int arg1, int arg2) | ||
749 | +{ | ||
750 | + int t0; | ||
751 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
752 | + tcg_gen_sari_i64(t0, arg1, 63); | ||
753 | + tcg_gen_op5(INDEX_op_div2_i64, ret, t0, arg1, t0, arg2); | ||
754 | +} | ||
755 | + | ||
756 | +static inline void tcg_gen_rem_i64(int ret, int arg1, int arg2) | ||
757 | +{ | ||
758 | + int t0; | ||
759 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
760 | + tcg_gen_sari_i64(t0, arg1, 63); | ||
761 | + tcg_gen_op5(INDEX_op_div2_i64, t0, ret, arg1, t0, arg2); | ||
762 | +} | ||
763 | + | ||
764 | +static inline void tcg_gen_divu_i64(int ret, int arg1, int arg2) | ||
765 | +{ | ||
766 | + int t0; | ||
767 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
768 | + tcg_gen_movi_i64(t0, 0); | ||
769 | + tcg_gen_op5(INDEX_op_divu2_i64, ret, t0, arg1, t0, arg2); | ||
770 | +} | ||
771 | + | ||
772 | +static inline void tcg_gen_remu_i64(int ret, int arg1, int arg2) | ||
773 | +{ | ||
774 | + int t0; | ||
775 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
776 | + tcg_gen_movi_i64(t0, 0); | ||
777 | + tcg_gen_op5(INDEX_op_divu2_i64, t0, ret, arg1, t0, arg2); | ||
778 | +} | ||
779 | +#endif | ||
780 | + | ||
781 | +#endif | ||
782 | + | ||
783 | +/***************************************/ | ||
784 | +/* optional operations */ | ||
785 | + | ||
786 | +static inline void tcg_gen_ext8s_i32(int ret, int arg) | ||
787 | +{ | ||
788 | +#ifdef TCG_TARGET_HAS_ext8s_i32 | ||
789 | + tcg_gen_op2(INDEX_op_ext8s_i32, ret, arg); | ||
790 | +#else | ||
791 | + tcg_gen_shli_i32(ret, arg, 24); | ||
792 | + tcg_gen_sari_i32(ret, arg, 24); | ||
793 | +#endif | ||
794 | +} | ||
795 | + | ||
796 | +static inline void tcg_gen_ext16s_i32(int ret, int arg) | ||
797 | +{ | ||
798 | +#ifdef TCG_TARGET_HAS_ext16s_i32 | ||
799 | + tcg_gen_op2(INDEX_op_ext16s_i32, ret, arg); | ||
800 | +#else | ||
801 | + tcg_gen_shli_i32(ret, arg, 16); | ||
802 | + tcg_gen_sari_i32(ret, arg, 16); | ||
803 | +#endif | ||
804 | +} | ||
805 | + | ||
806 | +/* Note: we assume the two high bytes are set to zero */ | ||
807 | +static inline void tcg_gen_bswap16_i32(TCGArg ret, TCGArg arg) | ||
808 | +{ | ||
809 | +#ifdef TCG_TARGET_HAS_bswap16_i32 | ||
810 | + tcg_gen_op2(INDEX_op_bswap16_i32, ret, arg); | ||
811 | +#else | ||
812 | + TCGArg t0, t1; | ||
813 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
814 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
815 | + | ||
816 | + tcg_gen_shri_i32(t0, arg, 8); | ||
817 | + tcg_gen_andi_i32(t1, arg, 0x000000ff); | ||
818 | + tcg_gen_shli_i32(t1, t1, 8); | ||
819 | + tcg_gen_or_i32(ret, t0, t1); | ||
820 | +#endif | ||
821 | +} | ||
822 | + | ||
823 | +static inline void tcg_gen_bswap_i32(TCGArg ret, TCGArg arg) | ||
824 | +{ | ||
825 | +#ifdef TCG_TARGET_HAS_bswap_i32 | ||
826 | + tcg_gen_op2(INDEX_op_bswap_i32, ret, arg); | ||
827 | +#else | ||
828 | + TCGArg t0, t1; | ||
829 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
830 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
831 | + | ||
832 | + tcg_gen_shli_i32(t0, arg, 24); | ||
833 | + | ||
834 | + tcg_gen_andi_i32(t1, arg, 0x0000ff00); | ||
835 | + tcg_gen_shli_i32(t1, t1, 8); | ||
836 | + tcg_gen_or_i32(t0, t0, t1); | ||
837 | + | ||
838 | + tcg_gen_shri_i32(t1, arg, 8); | ||
839 | + tcg_gen_andi_i32(t1, t1, 0x0000ff00); | ||
840 | + tcg_gen_or_i32(t0, t0, t1); | ||
841 | + | ||
842 | + tcg_gen_shri_i32(t1, arg, 24); | ||
843 | + tcg_gen_or_i32(ret, t0, t1); | ||
844 | +#endif | ||
845 | +} | ||
846 | + | ||
847 | +#if TCG_TARGET_REG_BITS == 32 | ||
848 | +static inline void tcg_gen_ext8s_i64(int ret, int arg) | ||
849 | +{ | ||
850 | + tcg_gen_ext8s_i32(ret, arg); | ||
851 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
852 | +} | ||
853 | + | ||
854 | +static inline void tcg_gen_ext16s_i64(int ret, int arg) | ||
855 | +{ | ||
856 | + tcg_gen_ext16s_i32(ret, arg); | ||
857 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
858 | +} | ||
859 | + | ||
860 | +static inline void tcg_gen_ext32s_i64(int ret, int arg) | ||
861 | +{ | ||
862 | + tcg_gen_mov_i32(ret, arg); | ||
863 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
864 | +} | ||
865 | + | ||
866 | +static inline void tcg_gen_trunc_i64_i32(int ret, int arg) | ||
867 | +{ | ||
868 | + tcg_gen_mov_i32(ret, arg); | ||
869 | +} | ||
870 | + | ||
871 | +static inline void tcg_gen_extu_i32_i64(int ret, int arg) | ||
872 | +{ | ||
873 | + tcg_gen_mov_i32(ret, arg); | ||
874 | + tcg_gen_movi_i32(ret + 1, 0); | ||
875 | +} | ||
876 | + | ||
877 | +static inline void tcg_gen_ext_i32_i64(int ret, int arg) | ||
878 | +{ | ||
879 | + tcg_gen_mov_i32(ret, arg); | ||
880 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
881 | +} | ||
882 | + | ||
883 | +static inline void tcg_gen_bswap_i64(int ret, int arg) | ||
884 | +{ | ||
885 | + int t0, t1; | ||
886 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
887 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
888 | + | ||
889 | + tcg_gen_bswap_i32(t0, arg); | ||
890 | + tcg_gen_bswap_i32(t1, arg + 1); | ||
891 | + tcg_gen_mov_i32(ret, t1); | ||
892 | + tcg_gen_mov_i32(ret + 1, t0); | ||
893 | +} | ||
894 | +#else | ||
895 | + | ||
896 | +static inline void tcg_gen_ext8s_i64(int ret, int arg) | ||
897 | +{ | ||
898 | +#ifdef TCG_TARGET_HAS_ext8s_i64 | ||
899 | + tcg_gen_op2(INDEX_op_ext8s_i64, ret, arg); | ||
900 | +#else | ||
901 | + tcg_gen_shli_i64(ret, arg, 56); | ||
902 | + tcg_gen_sari_i64(ret, arg, 56); | ||
903 | +#endif | ||
904 | +} | ||
905 | + | ||
906 | +static inline void tcg_gen_ext16s_i64(int ret, int arg) | ||
907 | +{ | ||
908 | +#ifdef TCG_TARGET_HAS_ext16s_i64 | ||
909 | + tcg_gen_op2(INDEX_op_ext16s_i64, ret, arg); | ||
910 | +#else | ||
911 | + tcg_gen_shli_i64(ret, arg, 48); | ||
912 | + tcg_gen_sari_i64(ret, arg, 48); | ||
913 | +#endif | ||
914 | +} | ||
915 | + | ||
916 | +static inline void tcg_gen_ext32s_i64(int ret, int arg) | ||
917 | +{ | ||
918 | +#ifdef TCG_TARGET_HAS_ext32s_i64 | ||
919 | + tcg_gen_op2(INDEX_op_ext32s_i64, ret, arg); | ||
920 | +#else | ||
921 | + tcg_gen_shli_i64(ret, arg, 32); | ||
922 | + tcg_gen_sari_i64(ret, arg, 32); | ||
923 | +#endif | ||
924 | +} | ||
925 | + | ||
926 | +/* Note: we assume the target supports move between 32 and 64 bit | ||
927 | + registers */ | ||
928 | +static inline void tcg_gen_trunc_i64_i32(int ret, int arg) | ||
929 | +{ | ||
930 | + tcg_gen_mov_i32(ret, arg); | ||
931 | +} | ||
932 | + | ||
933 | +/* Note: we assume the target supports move between 32 and 64 bit | ||
934 | + registers */ | ||
935 | +static inline void tcg_gen_extu_i32_i64(int ret, int arg) | ||
936 | +{ | ||
937 | + tcg_gen_andi_i64(ret, arg, 0xffffffff); | ||
938 | +} | ||
939 | + | ||
940 | +/* Note: we assume the target supports move between 32 and 64 bit | ||
941 | + registers */ | ||
942 | +static inline void tcg_gen_ext_i32_i64(int ret, int arg) | ||
943 | +{ | ||
944 | + tcg_gen_ext32s_i64(ret, arg); | ||
945 | +} | ||
946 | + | ||
947 | +static inline void tcg_gen_bswap_i64(TCGArg ret, TCGArg arg) | ||
948 | +{ | ||
949 | +#ifdef TCG_TARGET_HAS_bswap_i64 | ||
950 | + tcg_gen_op2(INDEX_op_bswap_i64, ret, arg); | ||
951 | +#else | ||
952 | + TCGArg t0, t1; | ||
953 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
954 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
955 | + | ||
956 | + tcg_gen_shli_i64(t0, arg, 56); | ||
957 | + | ||
958 | + tcg_gen_andi_i64(t1, arg, 0x0000ff00); | ||
959 | + tcg_gen_shli_i64(t1, t1, 40); | ||
960 | + tcg_gen_or_i64(t0, t0, t1); | ||
961 | + | ||
962 | + tcg_gen_andi_i64(t1, arg, 0x00ff0000); | ||
963 | + tcg_gen_shli_i64(t1, t1, 24); | ||
964 | + tcg_gen_or_i64(t0, t0, t1); | ||
965 | + | ||
966 | + tcg_gen_andi_i64(t1, arg, 0xff000000); | ||
967 | + tcg_gen_shli_i64(t1, t1, 8); | ||
968 | + tcg_gen_or_i64(t0, t0, t1); | ||
969 | + | ||
970 | + tcg_gen_shri_i64(t1, arg, 8); | ||
971 | + tcg_gen_andi_i64(t1, t1, 0xff000000); | ||
972 | + tcg_gen_or_i64(t0, t0, t1); | ||
973 | + | ||
974 | + tcg_gen_shri_i64(t1, arg, 24); | ||
975 | + tcg_gen_andi_i64(t1, t1, 0x00ff0000); | ||
976 | + tcg_gen_or_i64(t0, t0, t1); | ||
977 | + | ||
978 | + tcg_gen_shri_i64(t1, arg, 40); | ||
979 | + tcg_gen_andi_i64(t1, t1, 0x0000ff00); | ||
980 | + tcg_gen_or_i64(t0, t0, t1); | ||
981 | + | ||
982 | + tcg_gen_shri_i64(t1, arg, 56); | ||
983 | + tcg_gen_or_i64(ret, t0, t1); | ||
984 | +#endif | ||
985 | +} | ||
986 | + | ||
987 | +#endif | ||
988 | + | ||
989 | +/***************************************/ | ||
990 | +static inline void tcg_gen_macro_2(int ret0, int ret1, int macro_id) | ||
991 | +{ | ||
992 | + tcg_gen_op3(INDEX_op_macro_2, ret0, ret1, macro_id); | ||
993 | +} | ||
994 | + | ||
995 | +/***************************************/ | ||
996 | +/* QEMU specific operations. Their type depend on the QEMU CPU | ||
997 | + type. */ | ||
998 | +#ifndef TARGET_LONG_BITS | ||
999 | +#error must include QEMU headers | ||
1000 | +#endif | ||
1001 | + | ||
1002 | +static inline void tcg_gen_exit_tb(tcg_target_long val) | ||
1003 | +{ | ||
1004 | + tcg_gen_op1(INDEX_op_exit_tb, val); | ||
1005 | +} | ||
1006 | + | ||
1007 | +static inline void tcg_gen_goto_tb(int idx) | ||
1008 | +{ | ||
1009 | + tcg_gen_op1(INDEX_op_goto_tb, idx); | ||
1010 | +} | ||
1011 | + | ||
1012 | +#if TCG_TARGET_REG_BITS == 32 | ||
1013 | +static inline void tcg_gen_qemu_ld8u(int ret, int addr, int mem_index) | ||
1014 | +{ | ||
1015 | +#if TARGET_LONG_BITS == 32 | ||
1016 | + tcg_gen_op3(INDEX_op_qemu_ld8u, ret, addr, mem_index); | ||
1017 | +#else | ||
1018 | + tcg_gen_op4(INDEX_op_qemu_ld8u, ret, addr, addr + 1, mem_index); | ||
1019 | + tcg_gen_movi_i32(ret + 1, 0); | ||
1020 | +#endif | ||
1021 | +} | ||
1022 | + | ||
1023 | +static inline void tcg_gen_qemu_ld8s(int ret, int addr, int mem_index) | ||
1024 | +{ | ||
1025 | +#if TARGET_LONG_BITS == 32 | ||
1026 | + tcg_gen_op3(INDEX_op_qemu_ld8s, ret, addr, mem_index); | ||
1027 | +#else | ||
1028 | + tcg_gen_op4(INDEX_op_qemu_ld8s, ret, addr, addr + 1, mem_index); | ||
1029 | + tcg_gen_ext8s_i32(ret + 1, ret); | ||
1030 | +#endif | ||
1031 | +} | ||
1032 | + | ||
1033 | +static inline void tcg_gen_qemu_ld16u(int ret, int addr, int mem_index) | ||
1034 | +{ | ||
1035 | +#if TARGET_LONG_BITS == 32 | ||
1036 | + tcg_gen_op3(INDEX_op_qemu_ld16u, ret, addr, mem_index); | ||
1037 | +#else | ||
1038 | + tcg_gen_op4(INDEX_op_qemu_ld16u, ret, addr, addr + 1, mem_index); | ||
1039 | + tcg_gen_movi_i32(ret + 1, 0); | ||
1040 | +#endif | ||
1041 | +} | ||
1042 | + | ||
1043 | +static inline void tcg_gen_qemu_ld16s(int ret, int addr, int mem_index) | ||
1044 | +{ | ||
1045 | +#if TARGET_LONG_BITS == 32 | ||
1046 | + tcg_gen_op3(INDEX_op_qemu_ld16s, ret, addr, mem_index); | ||
1047 | +#else | ||
1048 | + tcg_gen_op4(INDEX_op_qemu_ld16s, ret, addr, addr + 1, mem_index); | ||
1049 | + tcg_gen_ext16s_i32(ret + 1, ret); | ||
1050 | +#endif | ||
1051 | +} | ||
1052 | + | ||
1053 | +static inline void tcg_gen_qemu_ld32u(int ret, int addr, int mem_index) | ||
1054 | +{ | ||
1055 | +#if TARGET_LONG_BITS == 32 | ||
1056 | + tcg_gen_op3(INDEX_op_qemu_ld32u, ret, addr, mem_index); | ||
1057 | +#else | ||
1058 | + tcg_gen_op4(INDEX_op_qemu_ld32u, ret, addr, addr + 1, mem_index); | ||
1059 | + tcg_gen_movi_i32(ret + 1, 0); | ||
1060 | +#endif | ||
1061 | +} | ||
1062 | + | ||
1063 | +static inline void tcg_gen_qemu_ld32s(int ret, int addr, int mem_index) | ||
1064 | +{ | ||
1065 | +#if TARGET_LONG_BITS == 32 | ||
1066 | + tcg_gen_op3(INDEX_op_qemu_ld32u, ret, addr, mem_index); | ||
1067 | +#else | ||
1068 | + tcg_gen_op4(INDEX_op_qemu_ld32u, ret, addr, addr + 1, mem_index); | ||
1069 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
1070 | +#endif | ||
1071 | +} | ||
1072 | + | ||
1073 | +static inline void tcg_gen_qemu_ld64(int ret, int addr, int mem_index) | ||
1074 | +{ | ||
1075 | +#if TARGET_LONG_BITS == 32 | ||
1076 | + tcg_gen_op4(INDEX_op_qemu_ld64, ret, ret + 1, addr, mem_index); | ||
1077 | +#else | ||
1078 | + tcg_gen_op5(INDEX_op_qemu_ld64, ret, ret + 1, addr, addr + 1, mem_index); | ||
1079 | +#endif | ||
1080 | +} | ||
1081 | + | ||
1082 | +static inline void tcg_gen_qemu_st8(int arg, int addr, int mem_index) | ||
1083 | +{ | ||
1084 | +#if TARGET_LONG_BITS == 32 | ||
1085 | + tcg_gen_op3(INDEX_op_qemu_st8, arg, addr, mem_index); | ||
1086 | +#else | ||
1087 | + tcg_gen_op4(INDEX_op_qemu_st8, arg, addr, addr + 1, mem_index); | ||
1088 | +#endif | ||
1089 | +} | ||
1090 | + | ||
1091 | +static inline void tcg_gen_qemu_st16(int arg, int addr, int mem_index) | ||
1092 | +{ | ||
1093 | +#if TARGET_LONG_BITS == 32 | ||
1094 | + tcg_gen_op3(INDEX_op_qemu_st16, arg, addr, mem_index); | ||
1095 | +#else | ||
1096 | + tcg_gen_op4(INDEX_op_qemu_st16, arg, addr, addr + 1, mem_index); | ||
1097 | +#endif | ||
1098 | +} | ||
1099 | + | ||
1100 | +static inline void tcg_gen_qemu_st32(int arg, int addr, int mem_index) | ||
1101 | +{ | ||
1102 | +#if TARGET_LONG_BITS == 32 | ||
1103 | + tcg_gen_op3(INDEX_op_qemu_st32, arg, addr, mem_index); | ||
1104 | +#else | ||
1105 | + tcg_gen_op4(INDEX_op_qemu_st32, arg, addr, addr + 1, mem_index); | ||
1106 | +#endif | ||
1107 | +} | ||
1108 | + | ||
1109 | +static inline void tcg_gen_qemu_st64(int arg, int addr, int mem_index) | ||
1110 | +{ | ||
1111 | +#if TARGET_LONG_BITS == 32 | ||
1112 | + tcg_gen_op4(INDEX_op_qemu_st64, arg, arg + 1, addr, mem_index); | ||
1113 | +#else | ||
1114 | + tcg_gen_op5(INDEX_op_qemu_st64, arg, arg + 1, addr, addr + 1, mem_index); | ||
1115 | +#endif | ||
1116 | +} | ||
1117 | + | ||
1118 | +#else /* TCG_TARGET_REG_BITS == 32 */ | ||
1119 | + | ||
1120 | +static inline void tcg_gen_qemu_ld8u(int ret, int addr, int mem_index) | ||
1121 | +{ | ||
1122 | + tcg_gen_op3(INDEX_op_qemu_ld8u, ret, addr, mem_index); | ||
1123 | +} | ||
1124 | + | ||
1125 | +static inline void tcg_gen_qemu_ld8s(int ret, int addr, int mem_index) | ||
1126 | +{ | ||
1127 | + tcg_gen_op3(INDEX_op_qemu_ld8s, ret, addr, mem_index); | ||
1128 | +} | ||
1129 | + | ||
1130 | +static inline void tcg_gen_qemu_ld16u(int ret, int addr, int mem_index) | ||
1131 | +{ | ||
1132 | + tcg_gen_op3(INDEX_op_qemu_ld16u, ret, addr, mem_index); | ||
1133 | +} | ||
1134 | + | ||
1135 | +static inline void tcg_gen_qemu_ld16s(int ret, int addr, int mem_index) | ||
1136 | +{ | ||
1137 | + tcg_gen_op3(INDEX_op_qemu_ld16s, ret, addr, mem_index); | ||
1138 | +} | ||
1139 | + | ||
1140 | +static inline void tcg_gen_qemu_ld32u(int ret, int addr, int mem_index) | ||
1141 | +{ | ||
1142 | + tcg_gen_op3(INDEX_op_qemu_ld32u, ret, addr, mem_index); | ||
1143 | +} | ||
1144 | + | ||
1145 | +static inline void tcg_gen_qemu_ld32s(int ret, int addr, int mem_index) | ||
1146 | +{ | ||
1147 | + tcg_gen_op3(INDEX_op_qemu_ld32s, ret, addr, mem_index); | ||
1148 | +} | ||
1149 | + | ||
1150 | +static inline void tcg_gen_qemu_ld64(int ret, int addr, int mem_index) | ||
1151 | +{ | ||
1152 | + tcg_gen_op3(INDEX_op_qemu_ld64, ret, addr, mem_index); | ||
1153 | +} | ||
1154 | + | ||
1155 | +static inline void tcg_gen_qemu_st8(int arg, int addr, int mem_index) | ||
1156 | +{ | ||
1157 | + tcg_gen_op3(INDEX_op_qemu_st8, arg, addr, mem_index); | ||
1158 | +} | ||
1159 | + | ||
1160 | +static inline void tcg_gen_qemu_st16(int arg, int addr, int mem_index) | ||
1161 | +{ | ||
1162 | + tcg_gen_op3(INDEX_op_qemu_st16, arg, addr, mem_index); | ||
1163 | +} | ||
1164 | + | ||
1165 | +static inline void tcg_gen_qemu_st32(int arg, int addr, int mem_index) | ||
1166 | +{ | ||
1167 | + tcg_gen_op3(INDEX_op_qemu_st32, arg, addr, mem_index); | ||
1168 | +} | ||
1169 | + | ||
1170 | +static inline void tcg_gen_qemu_st64(int arg, int addr, int mem_index) | ||
1171 | +{ | ||
1172 | + tcg_gen_op3(INDEX_op_qemu_st64, arg, addr, mem_index); | ||
1173 | +} | ||
1174 | + | ||
1175 | +#endif /* TCG_TARGET_REG_BITS != 32 */ |
tcg/tcg-opc.h
0 โ 100644
1 | +/* | ||
2 | + * Tiny Code Generator for QEMU | ||
3 | + * | ||
4 | + * Copyright (c) 2008 Fabrice Bellard | ||
5 | + * | ||
6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
7 | + * of this software and associated documentation files (the "Software"), to deal | ||
8 | + * in the Software without restriction, including without limitation the rights | ||
9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
10 | + * copies of the Software, and to permit persons to whom the Software is | ||
11 | + * furnished to do so, subject to the following conditions: | ||
12 | + * | ||
13 | + * The above copyright notice and this permission notice shall be included in | ||
14 | + * all copies or substantial portions of the Software. | ||
15 | + * | ||
16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
22 | + * THE SOFTWARE. | ||
23 | + */ | ||
24 | +#include "dyngen-opc.h" | ||
25 | + | ||
26 | +#ifndef DEF2 | ||
27 | +#define DEF2(name, oargs, iargs, cargs, flags) DEF(name, oargs + iargs + cargs, 0) | ||
28 | +#endif | ||
29 | + | ||
30 | +/* predefined ops */ | ||
31 | +DEF2(end, 0, 0, 0, 0) /* must be kept first */ | ||
32 | +DEF2(nop, 0, 0, 0, 0) | ||
33 | +DEF2(nop1, 0, 0, 1, 0) | ||
34 | +DEF2(nop2, 0, 0, 2, 0) | ||
35 | +DEF2(nop3, 0, 0, 3, 0) | ||
36 | +DEF2(nopn, 0, 0, 1, 0) /* variable number of parameters */ | ||
37 | +/* macro handling */ | ||
38 | +DEF2(macro_2, 2, 0, 1, 0) | ||
39 | +DEF2(macro_start, 0, 0, 2, 0) | ||
40 | +DEF2(macro_end, 0, 0, 2, 0) | ||
41 | +DEF2(macro_goto, 0, 0, 3, 0) | ||
42 | + | ||
43 | +DEF2(set_label, 0, 0, 1, 0) | ||
44 | +DEF2(call, 0, 1, 2, 0) /* variable number of parameters */ | ||
45 | +DEF2(jmp, 0, 1, 0, TCG_OPF_BB_END) | ||
46 | +DEF2(br, 0, 0, 1, TCG_OPF_BB_END) | ||
47 | + | ||
48 | +DEF2(mov_i32, 1, 1, 0, 0) | ||
49 | +DEF2(movi_i32, 1, 0, 1, 0) | ||
50 | +/* load/store */ | ||
51 | +DEF2(ld8u_i32, 1, 1, 1, 0) | ||
52 | +DEF2(ld8s_i32, 1, 1, 1, 0) | ||
53 | +DEF2(ld16u_i32, 1, 1, 1, 0) | ||
54 | +DEF2(ld16s_i32, 1, 1, 1, 0) | ||
55 | +DEF2(ld_i32, 1, 1, 1, 0) | ||
56 | +DEF2(st8_i32, 0, 2, 1, 0) | ||
57 | +DEF2(st16_i32, 0, 2, 1, 0) | ||
58 | +DEF2(st_i32, 0, 2, 1, 0) | ||
59 | +/* arith */ | ||
60 | +DEF2(add_i32, 1, 2, 0, 0) | ||
61 | +DEF2(sub_i32, 1, 2, 0, 0) | ||
62 | +DEF2(mul_i32, 1, 2, 0, 0) | ||
63 | +#ifdef TCG_TARGET_HAS_div_i32 | ||
64 | +DEF2(div_i32, 1, 2, 0, 0) | ||
65 | +DEF2(divu_i32, 1, 2, 0, 0) | ||
66 | +DEF2(rem_i32, 1, 2, 0, 0) | ||
67 | +DEF2(remu_i32, 1, 2, 0, 0) | ||
68 | +#else | ||
69 | +DEF2(div2_i32, 2, 3, 0, 0) | ||
70 | +DEF2(divu2_i32, 2, 3, 0, 0) | ||
71 | +#endif | ||
72 | +DEF2(and_i32, 1, 2, 0, 0) | ||
73 | +DEF2(or_i32, 1, 2, 0, 0) | ||
74 | +DEF2(xor_i32, 1, 2, 0, 0) | ||
75 | +/* shifts */ | ||
76 | +DEF2(shl_i32, 1, 2, 0, 0) | ||
77 | +DEF2(shr_i32, 1, 2, 0, 0) | ||
78 | +DEF2(sar_i32, 1, 2, 0, 0) | ||
79 | + | ||
80 | +DEF2(brcond_i32, 0, 2, 2, TCG_OPF_BB_END) | ||
81 | +#if TCG_TARGET_REG_BITS == 32 | ||
82 | +DEF2(add2_i32, 2, 4, 0, 0) | ||
83 | +DEF2(sub2_i32, 2, 4, 0, 0) | ||
84 | +DEF2(brcond2_i32, 0, 4, 2, TCG_OPF_BB_END) | ||
85 | +DEF2(mulu2_i32, 2, 2, 0, 0) | ||
86 | +#endif | ||
87 | +#ifdef TCG_TARGET_HAS_ext8s_i32 | ||
88 | +DEF2(ext8s_i32, 1, 1, 0, 0) | ||
89 | +#endif | ||
90 | +#ifdef TCG_TARGET_HAS_ext16s_i32 | ||
91 | +DEF2(ext16s_i32, 1, 1, 0, 0) | ||
92 | +#endif | ||
93 | +#ifdef TCG_TARGET_HAS_bswap_i32 | ||
94 | +DEF2(bswap_i32, 1, 1, 0, 0) | ||
95 | +#endif | ||
96 | + | ||
97 | +#if TCG_TARGET_REG_BITS == 64 | ||
98 | +DEF2(mov_i64, 1, 1, 0, 0) | ||
99 | +DEF2(movi_i64, 1, 0, 1, 0) | ||
100 | +/* load/store */ | ||
101 | +DEF2(ld8u_i64, 1, 1, 1, 0) | ||
102 | +DEF2(ld8s_i64, 1, 1, 1, 0) | ||
103 | +DEF2(ld16u_i64, 1, 1, 1, 0) | ||
104 | +DEF2(ld16s_i64, 1, 1, 1, 0) | ||
105 | +DEF2(ld32u_i64, 1, 1, 1, 0) | ||
106 | +DEF2(ld32s_i64, 1, 1, 1, 0) | ||
107 | +DEF2(ld_i64, 1, 1, 1, 0) | ||
108 | +DEF2(st8_i64, 0, 2, 1, 0) | ||
109 | +DEF2(st16_i64, 0, 2, 1, 0) | ||
110 | +DEF2(st32_i64, 0, 2, 1, 0) | ||
111 | +DEF2(st_i64, 0, 2, 1, 0) | ||
112 | +/* arith */ | ||
113 | +DEF2(add_i64, 1, 2, 0, 0) | ||
114 | +DEF2(sub_i64, 1, 2, 0, 0) | ||
115 | +DEF2(mul_i64, 1, 2, 0, 0) | ||
116 | +#ifdef TCG_TARGET_HAS_div_i64 | ||
117 | +DEF2(div_i64, 1, 2, 0, 0) | ||
118 | +DEF2(divu_i64, 1, 2, 0, 0) | ||
119 | +DEF2(rem_i64, 1, 2, 0, 0) | ||
120 | +DEF2(remu_i64, 1, 2, 0, 0) | ||
121 | +#else | ||
122 | +DEF2(div2_i64, 2, 3, 0, 0) | ||
123 | +DEF2(divu2_i64, 2, 3, 0, 0) | ||
124 | +#endif | ||
125 | +DEF2(and_i64, 1, 2, 0, 0) | ||
126 | +DEF2(or_i64, 1, 2, 0, 0) | ||
127 | +DEF2(xor_i64, 1, 2, 0, 0) | ||
128 | +/* shifts */ | ||
129 | +DEF2(shl_i64, 1, 2, 0, 0) | ||
130 | +DEF2(shr_i64, 1, 2, 0, 0) | ||
131 | +DEF2(sar_i64, 1, 2, 0, 0) | ||
132 | + | ||
133 | +DEF2(brcond_i64, 0, 2, 2, TCG_OPF_BB_END) | ||
134 | +#ifdef TCG_TARGET_HAS_ext8s_i64 | ||
135 | +DEF2(ext8s_i64, 1, 1, 0, 0) | ||
136 | +#endif | ||
137 | +#ifdef TCG_TARGET_HAS_ext16s_i64 | ||
138 | +DEF2(ext16s_i64, 1, 1, 0, 0) | ||
139 | +#endif | ||
140 | +#ifdef TCG_TARGET_HAS_ext32s_i64 | ||
141 | +DEF2(ext32s_i64, 1, 1, 0, 0) | ||
142 | +#endif | ||
143 | +#ifdef TCG_TARGET_HAS_bswap_i64 | ||
144 | +DEF2(bswap_i64, 1, 1, 0, 0) | ||
145 | +#endif | ||
146 | +#endif | ||
147 | + | ||
148 | +/* QEMU specific */ | ||
149 | +DEF2(exit_tb, 0, 0, 1, TCG_OPF_BB_END) | ||
150 | +DEF2(goto_tb, 0, 0, 1, TCG_OPF_BB_END) | ||
151 | +/* Note: even if TARGET_LONG_BITS is not defined, the INDEX_op | ||
152 | + constants must be defined */ | ||
153 | +#if TCG_TARGET_REG_BITS == 32 | ||
154 | +#if TARGET_LONG_BITS == 32 | ||
155 | +DEF2(qemu_ld8u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
156 | +#else | ||
157 | +DEF2(qemu_ld8u, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
158 | +#endif | ||
159 | +#if TARGET_LONG_BITS == 32 | ||
160 | +DEF2(qemu_ld8s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
161 | +#else | ||
162 | +DEF2(qemu_ld8s, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
163 | +#endif | ||
164 | +#if TARGET_LONG_BITS == 32 | ||
165 | +DEF2(qemu_ld16u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
166 | +#else | ||
167 | +DEF2(qemu_ld16u, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
168 | +#endif | ||
169 | +#if TARGET_LONG_BITS == 32 | ||
170 | +DEF2(qemu_ld16s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
171 | +#else | ||
172 | +DEF2(qemu_ld16s, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
173 | +#endif | ||
174 | +#if TARGET_LONG_BITS == 32 | ||
175 | +DEF2(qemu_ld32u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
176 | +#else | ||
177 | +DEF2(qemu_ld32u, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
178 | +#endif | ||
179 | +#if TARGET_LONG_BITS == 32 | ||
180 | +DEF2(qemu_ld32s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
181 | +#else | ||
182 | +DEF2(qemu_ld32s, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
183 | +#endif | ||
184 | +#if TARGET_LONG_BITS == 32 | ||
185 | +DEF2(qemu_ld64, 2, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
186 | +#else | ||
187 | +DEF2(qemu_ld64, 2, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
188 | +#endif | ||
189 | + | ||
190 | +#if TARGET_LONG_BITS == 32 | ||
191 | +DEF2(qemu_st8, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
192 | +#else | ||
193 | +DEF2(qemu_st8, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | ||
194 | +#endif | ||
195 | +#if TARGET_LONG_BITS == 32 | ||
196 | +DEF2(qemu_st16, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
197 | +#else | ||
198 | +DEF2(qemu_st16, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | ||
199 | +#endif | ||
200 | +#if TARGET_LONG_BITS == 32 | ||
201 | +DEF2(qemu_st32, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
202 | +#else | ||
203 | +DEF2(qemu_st32, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | ||
204 | +#endif | ||
205 | +#if TARGET_LONG_BITS == 32 | ||
206 | +DEF2(qemu_st64, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | ||
207 | +#else | ||
208 | +DEF2(qemu_st64, 0, 4, 1, TCG_OPF_CALL_CLOBBER) | ||
209 | +#endif | ||
210 | + | ||
211 | +#else /* TCG_TARGET_REG_BITS == 32 */ | ||
212 | + | ||
213 | +DEF2(qemu_ld8u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
214 | +DEF2(qemu_ld8s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
215 | +DEF2(qemu_ld16u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
216 | +DEF2(qemu_ld16s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
217 | +DEF2(qemu_ld32u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
218 | +DEF2(qemu_ld32s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
219 | +DEF2(qemu_ld64, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
220 | + | ||
221 | +DEF2(qemu_st8, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
222 | +DEF2(qemu_st16, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
223 | +DEF2(qemu_st32, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
224 | +DEF2(qemu_st64, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
225 | + | ||
226 | +#endif /* TCG_TARGET_REG_BITS != 32 */ | ||
227 | + | ||
228 | +#undef DEF2 |
tcg/tcg-runtime.c
0 โ 100644
1 | +/* | ||
2 | + * Tiny Code Generator for QEMU | ||
3 | + * | ||
4 | + * Copyright (c) 2008 Fabrice Bellard | ||
5 | + * | ||
6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
7 | + * of this software and associated documentation files (the "Software"), to deal | ||
8 | + * in the Software without restriction, including without limitation the rights | ||
9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
10 | + * copies of the Software, and to permit persons to whom the Software is | ||
11 | + * furnished to do so, subject to the following conditions: | ||
12 | + * | ||
13 | + * The above copyright notice and this permission notice shall be included in | ||
14 | + * all copies or substantial portions of the Software. | ||
15 | + * | ||
16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
22 | + * THE SOFTWARE. | ||
23 | + */ | ||
24 | +#include <stdarg.h> | ||
25 | +#include <stdlib.h> | ||
26 | +#include <stdio.h> | ||
27 | +#include <string.h> | ||
28 | +#include <inttypes.h> | ||
29 | + | ||
30 | +#include "config.h" | ||
31 | +#include "osdep.h" | ||
32 | +#include "tcg.h" | ||
33 | + | ||
34 | +int64_t tcg_helper_shl_i64(int64_t arg1, int64_t arg2) | ||
35 | +{ | ||
36 | + return arg1 << arg2; | ||
37 | +} | ||
38 | + | ||
39 | +int64_t tcg_helper_shr_i64(int64_t arg1, int64_t arg2) | ||
40 | +{ | ||
41 | + return (uint64_t)arg1 >> arg2; | ||
42 | +} | ||
43 | + | ||
44 | +int64_t tcg_helper_sar_i64(int64_t arg1, int64_t arg2) | ||
45 | +{ | ||
46 | + return arg1 >> arg2; | ||
47 | +} | ||
48 | + | ||
49 | +int64_t tcg_helper_div_i64(int64_t arg1, int64_t arg2) | ||
50 | +{ | ||
51 | + return arg1 / arg2; | ||
52 | +} | ||
53 | + | ||
54 | +int64_t tcg_helper_rem_i64(int64_t arg1, int64_t arg2) | ||
55 | +{ | ||
56 | + return arg1 / arg2; | ||
57 | +} | ||
58 | + | ||
59 | +uint64_t tcg_helper_divu_i64(uint64_t arg1, uint64_t arg2) | ||
60 | +{ | ||
61 | + return arg1 / arg2; | ||
62 | +} | ||
63 | + | ||
64 | +uint64_t tcg_helper_remu_i64(uint64_t arg1, uint64_t arg2) | ||
65 | +{ | ||
66 | + return arg1 / arg2; | ||
67 | +} | ||
68 | + |