Commit c896fe29d6c8ae6cde3917727812ced3f2e536a4
1 parent
56abbcff
TCG code generator
git-svn-id: svn://svn.savannah.nongnu.org/qemu/trunk@3943 c046a42c-6fe2-441c-8c8c-71466251a162
Showing
9 changed files
with
3625 additions
and
0 deletions
Too many changes to show.
To preserve performance only 9 of 13 files are displayed.
tcg/LICENSE
0 โ 100644
tcg/README
0 โ 100644
| 1 | +Tiny Code Generator - Fabrice Bellard. | |
| 2 | + | |
| 3 | +1) Introduction | |
| 4 | + | |
| 5 | +TCG (Tiny Code Generator) began as a generic backend for a C | |
| 6 | +compiler. It was simplified to be used in QEMU. It also has its roots | |
| 7 | +in the QOP code generator written by Paul Brook. | |
| 8 | + | |
| 9 | +2) Definitions | |
| 10 | + | |
| 11 | +The TCG "target" is the architecture for which we generate the | |
| 12 | +code. It is of course not the same as the "target" of QEMU which is | |
| 13 | +the emulated architecture. As TCG started as a generic C backend used | |
| 14 | +for cross compiling, it is assumed that the TCG target is different | |
| 15 | +from the host, although it is never the case for QEMU. | |
| 16 | + | |
| 17 | +A TCG "function" corresponds to a QEMU Translated Block (TB). | |
| 18 | + | |
| 19 | +A TCG "temporary" is a variable only live in a given | |
| 20 | +function. Temporaries are allocated explicitely in each function. | |
| 21 | + | |
| 22 | +A TCG "global" is a variable which is live in all the functions. They | |
| 23 | +are defined before the functions defined. A TCG global can be a memory | |
| 24 | +location (e.g. a QEMU CPU register), a fixed host register (e.g. the | |
| 25 | +QEMU CPU state pointer) or a memory location which is stored in a | |
| 26 | +register outside QEMU TBs (not implemented yet). | |
| 27 | + | |
| 28 | +A TCG "basic block" corresponds to a list of instructions terminated | |
| 29 | +by a branch instruction. | |
| 30 | + | |
| 31 | +3) Intermediate representation | |
| 32 | + | |
| 33 | +3.1) Introduction | |
| 34 | + | |
| 35 | +TCG instructions operate on variables which are temporaries or | |
| 36 | +globals. TCG instructions and variables are strongly typed. Two types | |
| 37 | +are supported: 32 bit integers and 64 bit integers. Pointers are | |
| 38 | +defined as an alias to 32 bit or 64 bit integers depending on the TCG | |
| 39 | +target word size. | |
| 40 | + | |
| 41 | +Each instruction has a fixed number of output variable operands, input | |
| 42 | +variable operands and always constant operands. | |
| 43 | + | |
| 44 | +The notable exception is the call instruction which has a variable | |
| 45 | +number of outputs and inputs. | |
| 46 | + | |
| 47 | +In the textual form, output operands come first, followed by input | |
| 48 | +operands, followed by constant operands. The output type is included | |
| 49 | +in the instruction name. Constants are prefixed with a '$'. | |
| 50 | + | |
| 51 | +add_i32 t0, t1, t2 (t0 <- t1 + t2) | |
| 52 | + | |
| 53 | +sub_i64 t2, t3, $4 (t2 <- t3 - 4) | |
| 54 | + | |
| 55 | +3.2) Assumptions | |
| 56 | + | |
| 57 | +* Basic blocks | |
| 58 | + | |
| 59 | +- Basic blocks end after branches (e.g. brcond_i32 instruction), | |
| 60 | + goto_tb and exit_tb instructions. | |
| 61 | +- Basic blocks end before legacy dyngen operations. | |
| 62 | +- Basic blocks start after the end of a previous basic block, at a | |
| 63 | + set_label instruction or after a legacy dyngen operation. | |
| 64 | + | |
| 65 | +After the end of a basic block, temporaries at destroyed and globals | |
| 66 | +are stored at their initial storage (register or memory place | |
| 67 | +depending on their declarations). | |
| 68 | + | |
| 69 | +* Floating point types are not supported yet | |
| 70 | + | |
| 71 | +* Pointers: depending on the TCG target, pointer size is 32 bit or 64 | |
| 72 | + bit. The type TCG_TYPE_PTR is an alias to TCG_TYPE_I32 or | |
| 73 | + TCG_TYPE_I64. | |
| 74 | + | |
| 75 | +* Helpers: | |
| 76 | + | |
| 77 | +Using the tcg_gen_helper_x_y it is possible to call any function | |
| 78 | +taking i32, i64 or pointer types types. Before calling an helper, all | |
| 79 | +globals are stored at their canonical location and it is assumed that | |
| 80 | +the function can modify them. In the future, function modifiers will | |
| 81 | +be allowed to tell that the helper does not read or write some globals. | |
| 82 | + | |
| 83 | +On some TCG targets (e.g. x86), several calling conventions are | |
| 84 | +supported. | |
| 85 | + | |
| 86 | +* Branches: | |
| 87 | + | |
| 88 | +Use the instruction 'br' to jump to a label. Use 'jmp' to jump to an | |
| 89 | +explicit address. Conditional branches can only jump to labels. | |
| 90 | + | |
| 91 | +3.3) Code Optimizations | |
| 92 | + | |
| 93 | +When generating instructions, you can count on at least the following | |
| 94 | +optimizations: | |
| 95 | + | |
| 96 | +- Single instructions are simplified, e.g. | |
| 97 | + | |
| 98 | + and_i32 t0, t0, $0xffffffff | |
| 99 | + | |
| 100 | + is suppressed. | |
| 101 | + | |
| 102 | +- A liveness analysis is done at the basic block level. The | |
| 103 | + information is used to suppress moves from a dead temporary to | |
| 104 | + another one. It is also used to remove instructions which compute | |
| 105 | + dead results. The later is especially useful for condition code | |
| 106 | + optimisation in QEMU. | |
| 107 | + | |
| 108 | + In the following example: | |
| 109 | + | |
| 110 | + add_i32 t0, t1, t2 | |
| 111 | + add_i32 t0, t0, $1 | |
| 112 | + mov_i32 t0, $1 | |
| 113 | + | |
| 114 | + only the last instruction is kept. | |
| 115 | + | |
| 116 | +- A macro system is supported (may get closer to function inlining | |
| 117 | + some day). It is useful if the liveness analysis is likely to prove | |
| 118 | + that some results of a computation are indeed not useful. With the | |
| 119 | + macro system, the user can provide several alternative | |
| 120 | + implementations which are used depending on the used results. It is | |
| 121 | + especially useful for condition code optimisation in QEMU. | |
| 122 | + | |
| 123 | + Here is an example: | |
| 124 | + | |
| 125 | + macro_2 t0, t1, $1 | |
| 126 | + mov_i32 t0, $0x1234 | |
| 127 | + | |
| 128 | + The macro identified by the ID "$1" normally returns the values t0 | |
| 129 | + and t1. Suppose its implementation is: | |
| 130 | + | |
| 131 | + macro_start | |
| 132 | + brcond_i32 t2, $0, $TCG_COND_EQ, $1 | |
| 133 | + mov_i32 t0, $2 | |
| 134 | + br $2 | |
| 135 | + set_label $1 | |
| 136 | + mov_i32 t0, $3 | |
| 137 | + set_label $2 | |
| 138 | + add_i32 t1, t3, t4 | |
| 139 | + macro_end | |
| 140 | + | |
| 141 | + If t0 is not used after the macro, the user can provide a simpler | |
| 142 | + implementation: | |
| 143 | + | |
| 144 | + macro_start | |
| 145 | + add_i32 t1, t2, t4 | |
| 146 | + macro_end | |
| 147 | + | |
| 148 | + TCG automatically chooses the right implementation depending on | |
| 149 | + which macro outputs are used after it. | |
| 150 | + | |
| 151 | + Note that if TCG did more expensive optimizations, macros would be | |
| 152 | + less useful. In the previous example a macro is useful because the | |
| 153 | + liveness analysis is done on each basic block separately. Hence TCG | |
| 154 | + cannot remove the code computing 't0' even if it is not used after | |
| 155 | + the first macro implementation. | |
| 156 | + | |
| 157 | +3.4) Instruction Reference | |
| 158 | + | |
| 159 | +********* Function call | |
| 160 | + | |
| 161 | +* call <ret> <params> ptr | |
| 162 | + | |
| 163 | +call function 'ptr' (pointer type) | |
| 164 | + | |
| 165 | +<ret> optional 32 bit or 64 bit return value | |
| 166 | +<params> optional 32 bit or 64 bit parameters | |
| 167 | + | |
| 168 | +********* Jumps/Labels | |
| 169 | + | |
| 170 | +* jmp t0 | |
| 171 | + | |
| 172 | +Absolute jump to address t0 (pointer type). | |
| 173 | + | |
| 174 | +* set_label $label | |
| 175 | + | |
| 176 | +Define label 'label' at the current program point. | |
| 177 | + | |
| 178 | +* br $label | |
| 179 | + | |
| 180 | +Jump to label. | |
| 181 | + | |
| 182 | +* brcond_i32/i64 cond, t0, t1, label | |
| 183 | + | |
| 184 | +Conditional jump if t0 cond t1 is true. cond can be: | |
| 185 | + TCG_COND_EQ | |
| 186 | + TCG_COND_NE | |
| 187 | + TCG_COND_LT /* signed */ | |
| 188 | + TCG_COND_GE /* signed */ | |
| 189 | + TCG_COND_LE /* signed */ | |
| 190 | + TCG_COND_GT /* signed */ | |
| 191 | + TCG_COND_LTU /* unsigned */ | |
| 192 | + TCG_COND_GEU /* unsigned */ | |
| 193 | + TCG_COND_LEU /* unsigned */ | |
| 194 | + TCG_COND_GTU /* unsigned */ | |
| 195 | + | |
| 196 | +********* Arithmetic | |
| 197 | + | |
| 198 | +* add_i32/i64 t0, t1, t2 | |
| 199 | + | |
| 200 | +t0=t1+t2 | |
| 201 | + | |
| 202 | +* sub_i32/i64 t0, t1, t2 | |
| 203 | + | |
| 204 | +t0=t1-t2 | |
| 205 | + | |
| 206 | +* mul_i32/i64 t0, t1, t2 | |
| 207 | + | |
| 208 | +t0=t1*t2 | |
| 209 | + | |
| 210 | +* div_i32/i64 t0, t1, t2 | |
| 211 | + | |
| 212 | +t0=t1/t2 (signed). Undefined behavior if division by zero or overflow. | |
| 213 | + | |
| 214 | +* divu_i32/i64 t0, t1, t2 | |
| 215 | + | |
| 216 | +t0=t1/t2 (unsigned). Undefined behavior if division by zero. | |
| 217 | + | |
| 218 | +* rem_i32/i64 t0, t1, t2 | |
| 219 | + | |
| 220 | +t0=t1%t2 (signed). Undefined behavior if division by zero or overflow. | |
| 221 | + | |
| 222 | +* remu_i32/i64 t0, t1, t2 | |
| 223 | + | |
| 224 | +t0=t1%t2 (unsigned). Undefined behavior if division by zero. | |
| 225 | + | |
| 226 | +* and_i32/i64 t0, t1, t2 | |
| 227 | + | |
| 228 | +********* Logical | |
| 229 | + | |
| 230 | +t0=t1&t2 | |
| 231 | + | |
| 232 | +* or_i32/i64 t0, t1, t2 | |
| 233 | + | |
| 234 | +t0=t1|t2 | |
| 235 | + | |
| 236 | +* xor_i32/i64 t0, t1, t2 | |
| 237 | + | |
| 238 | +t0=t1^t2 | |
| 239 | + | |
| 240 | +* shl_i32/i64 t0, t1, t2 | |
| 241 | + | |
| 242 | +********* Shifts | |
| 243 | + | |
| 244 | +* shl_i32/i64 t0, t1, t2 | |
| 245 | + | |
| 246 | +t0=t1 << t2. Undefined behavior if t2 < 0 or t2 >= 32 (resp 64) | |
| 247 | + | |
| 248 | +* shr_i32/i64 t0, t1, t2 | |
| 249 | + | |
| 250 | +t0=t1 >> t2 (unsigned). Undefined behavior if t2 < 0 or t2 >= 32 (resp 64) | |
| 251 | + | |
| 252 | +* sar_i32/i64 t0, t1, t2 | |
| 253 | + | |
| 254 | +t0=t1 >> t2 (signed). Undefined behavior if t2 < 0 or t2 >= 32 (resp 64) | |
| 255 | + | |
| 256 | +********* Misc | |
| 257 | + | |
| 258 | +* mov_i32/i64 t0, t1 | |
| 259 | + | |
| 260 | +t0 = t1 | |
| 261 | + | |
| 262 | +Move t1 to t0 (both operands must have the same type). | |
| 263 | + | |
| 264 | +* ext8s_i32/i64 t0, t1 | |
| 265 | +ext16s_i32/i64 t0, t1 | |
| 266 | +ext32s_i64 t0, t1 | |
| 267 | + | |
| 268 | +8, 16 or 32 bit sign extension (both operands must have the same type) | |
| 269 | + | |
| 270 | +* bswap16_i32 t0, t1 | |
| 271 | + | |
| 272 | +16 bit byte swap on a 32 bit value. The two high order bytes must be set | |
| 273 | +to zero. | |
| 274 | + | |
| 275 | +* bswap_i32 t0, t1 | |
| 276 | + | |
| 277 | +32 bit byte swap | |
| 278 | + | |
| 279 | +* bswap_i64 t0, t1 | |
| 280 | + | |
| 281 | +64 bit byte swap | |
| 282 | + | |
| 283 | +********* Type conversions | |
| 284 | + | |
| 285 | +* ext_i32_i64 t0, t1 | |
| 286 | +Convert t1 (32 bit) to t0 (64 bit) and does sign extension | |
| 287 | + | |
| 288 | +* extu_i32_i64 t0, t1 | |
| 289 | +Convert t1 (32 bit) to t0 (64 bit) and does zero extension | |
| 290 | + | |
| 291 | +* trunc_i64_i32 t0, t1 | |
| 292 | +Truncate t1 (64 bit) to t0 (32 bit) | |
| 293 | + | |
| 294 | +********* Load/Store | |
| 295 | + | |
| 296 | +* ld_i32/i64 t0, t1, offset | |
| 297 | +ld8s_i32/i64 t0, t1, offset | |
| 298 | +ld8u_i32/i64 t0, t1, offset | |
| 299 | +ld16s_i32/i64 t0, t1, offset | |
| 300 | +ld16u_i32/i64 t0, t1, offset | |
| 301 | +ld32s_i64 t0, t1, offset | |
| 302 | +ld32u_i64 t0, t1, offset | |
| 303 | + | |
| 304 | +t0 = read(t1 + offset) | |
| 305 | +Load 8, 16, 32 or 64 bits with or without sign extension from host memory. | |
| 306 | +offset must be a constant. | |
| 307 | + | |
| 308 | +* st_i32/i64 t0, t1, offset | |
| 309 | +st8_i32/i64 t0, t1, offset | |
| 310 | +st16_i32/i64 t0, t1, offset | |
| 311 | +st32_i64 t0, t1, offset | |
| 312 | + | |
| 313 | +write(t0, t1 + offset) | |
| 314 | +Write 8, 16, 32 or 64 bits to host memory. | |
| 315 | + | |
| 316 | +********* QEMU specific operations | |
| 317 | + | |
| 318 | +* tb_exit t0 | |
| 319 | + | |
| 320 | +Exit the current TB and return the value t0 (word type). | |
| 321 | + | |
| 322 | +* goto_tb index | |
| 323 | + | |
| 324 | +Exit the current TB and jump to the TB index 'index' (constant) if the | |
| 325 | +current TB was linked to this TB. Otherwise execute the next | |
| 326 | +instructions. | |
| 327 | + | |
| 328 | +* qemu_ld_i32/i64 t0, t1, flags | |
| 329 | +qemu_ld8u_i32/i64 t0, t1, flags | |
| 330 | +qemu_ld8s_i32/i64 t0, t1, flags | |
| 331 | +qemu_ld16u_i32/i64 t0, t1, flags | |
| 332 | +qemu_ld16s_i32/i64 t0, t1, flags | |
| 333 | +qemu_ld32u_i64 t0, t1, flags | |
| 334 | +qemu_ld32s_i64 t0, t1, flags | |
| 335 | + | |
| 336 | +Load data at the QEMU CPU address t1 into t0. t1 has the QEMU CPU | |
| 337 | +address type. 'flags' contains the QEMU memory index (selects user or | |
| 338 | +kernel access) for example. | |
| 339 | + | |
| 340 | +* qemu_st_i32/i64 t0, t1, flags | |
| 341 | +qemu_st8_i32/i64 t0, t1, flags | |
| 342 | +qemu_st16_i32/i64 t0, t1, flags | |
| 343 | +qemu_st32_i64 t0, t1, flags | |
| 344 | + | |
| 345 | +Store the data t0 at the QEMU CPU Address t1. t1 has the QEMU CPU | |
| 346 | +address type. 'flags' contains the QEMU memory index (selects user or | |
| 347 | +kernel access) for example. | |
| 348 | + | |
| 349 | +Note 1: Some shortcuts are defined when the last operand is known to be | |
| 350 | +a constant (e.g. addi for add, movi for mov). | |
| 351 | + | |
| 352 | +Note 2: When using TCG, the opcodes must never be generated directly | |
| 353 | +as some of them may not be available as "real" opcodes. Always use the | |
| 354 | +function tcg_gen_xxx(args). | |
| 355 | + | |
| 356 | +4) Backend | |
| 357 | + | |
| 358 | +tcg-target.h contains the target specific definitions. tcg-target.c | |
| 359 | +contains the target specific code. | |
| 360 | + | |
| 361 | +4.1) Assumptions | |
| 362 | + | |
| 363 | +The target word size (TCG_TARGET_REG_BITS) is expected to be 32 bit or | |
| 364 | +64 bit. It is expected that the pointer has the same size as the word. | |
| 365 | + | |
| 366 | +On a 32 bit target, all 64 bit operations are converted to 32 bits. A | |
| 367 | +few specific operations must be implemented to allow it (see add2_i32, | |
| 368 | +sub2_i32, brcond2_i32). | |
| 369 | + | |
| 370 | +Floating point operations are not supported in this version. A | |
| 371 | +previous incarnation of the code generator had full support of them, | |
| 372 | +but it is better to concentrate on integer operations first. | |
| 373 | + | |
| 374 | +On a 64 bit target, no assumption is made in TCG about the storage of | |
| 375 | +the 32 bit values in 64 bit registers. | |
| 376 | + | |
| 377 | +4.2) Constraints | |
| 378 | + | |
| 379 | +GCC like constraints are used to define the constraints of every | |
| 380 | +instruction. Memory constraints are not supported in this | |
| 381 | +version. Aliases are specified in the input operands as for GCC. | |
| 382 | + | |
| 383 | +A target can define specific register or constant constraints. If an | |
| 384 | +operation uses a constant input constraint which does not allow all | |
| 385 | +constants, it must also accept registers in order to have a fallback. | |
| 386 | + | |
| 387 | +The movi_i32 and movi_i64 operations must accept any constants. | |
| 388 | + | |
| 389 | +The mov_i32 and mov_i64 operations must accept any registers of the | |
| 390 | +same type. | |
| 391 | + | |
| 392 | +The ld/st instructions must accept signed 32 bit constant offsets. It | |
| 393 | +can be implemented by reserving a specific register to compute the | |
| 394 | +address if the offset is too big. | |
| 395 | + | |
| 396 | +The ld/st instructions must accept any destination (ld) or source (st) | |
| 397 | +register. | |
| 398 | + | |
| 399 | +4.3) Function call assumptions | |
| 400 | + | |
| 401 | +- The only supported types for parameters and return value are: 32 and | |
| 402 | + 64 bit integers and pointer. | |
| 403 | +- The stack grows downwards. | |
| 404 | +- The first N parameters are passed in registers. | |
| 405 | +- The next parameters are passed on the stack by storing them as words. | |
| 406 | +- Some registers are clobbered during the call. | |
| 407 | +- The function can return 0 or 1 value in registers. On a 32 bit | |
| 408 | + target, functions must be able to return 2 values in registers for | |
| 409 | + 64 bit return type. | |
| 410 | + | |
| 411 | +5) Migration from dyngen to TCG | |
| 412 | + | |
| 413 | +TCG is backward compatible with QEMU "dyngen" operations. It means | |
| 414 | +that TCG instructions can be freely mixed with dyngen operations. It | |
| 415 | +is expected that QEMU targets will be progressively fully converted to | |
| 416 | +TCG. Once a target is fully converted to dyngen, it will be possible | |
| 417 | +to apply more optimizations because more registers will be free for | |
| 418 | +the generated code. | |
| 419 | + | |
| 420 | +The exception model is the same as the dyngen one. | ... | ... |
tcg/TODO
0 โ 100644
| 1 | +- test macro system | |
| 2 | + | |
| 3 | +- test conditional jumps | |
| 4 | + | |
| 5 | +- test mul, div, ext8s, ext16s, bswap | |
| 6 | + | |
| 7 | +- generate a global TB prologue and epilogue to save/restore registers | |
| 8 | + to/from the CPU state and to reserve a stack frame to optimize | |
| 9 | + helper calls. Modify cpu-exec.c so that it does not use global | |
| 10 | + register variables (except maybe for 'env'). | |
| 11 | + | |
| 12 | +- fully convert the x86 target. The minimal amount of work includes: | |
| 13 | + - add cc_src, cc_dst and cc_op as globals | |
| 14 | + - disable its eflags optimization (the liveness analysis should | |
| 15 | + suffice) | |
| 16 | + - move complicated operations to helpers (in particular FPU, SSE, MMX). | |
| 17 | + | |
| 18 | +- optimize the x86 target: | |
| 19 | + - move some or all the registers as globals | |
| 20 | + - use the TB prologue and epilogue to have QEMU target registers in | |
| 21 | + pre assigned host registers. | |
| 22 | + | |
| 23 | +Ideas: | |
| 24 | + | |
| 25 | +- Move the slow part of the qemu_ld/st ops after the end of the TB. | |
| 26 | + | |
| 27 | +- Experiment: change instruction storage to simplify macro handling | |
| 28 | + and to handle dynamic allocation and see if the translation speed is | |
| 29 | + OK. | |
| 30 | + | |
| 31 | +- change exception syntax to get closer to QOP system (exception | |
| 32 | + parameters given with a specific instruction). | ... | ... |
tcg/i386/tcg-target.c
0 โ 100644
| 1 | +/* | |
| 2 | + * Tiny Code Generator for QEMU | |
| 3 | + * | |
| 4 | + * Copyright (c) 2008 Fabrice Bellard | |
| 5 | + * | |
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | |
| 7 | + * of this software and associated documentation files (the "Software"), to deal | |
| 8 | + * in the Software without restriction, including without limitation the rights | |
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | |
| 10 | + * copies of the Software, and to permit persons to whom the Software is | |
| 11 | + * furnished to do so, subject to the following conditions: | |
| 12 | + * | |
| 13 | + * The above copyright notice and this permission notice shall be included in | |
| 14 | + * all copies or substantial portions of the Software. | |
| 15 | + * | |
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | |
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | |
| 22 | + * THE SOFTWARE. | |
| 23 | + */ | |
| 24 | +const char *tcg_target_reg_names[TCG_TARGET_NB_REGS] = { | |
| 25 | + "%eax", | |
| 26 | + "%ecx", | |
| 27 | + "%edx", | |
| 28 | + "%ebx", | |
| 29 | + "%esp", | |
| 30 | + "%ebp", | |
| 31 | + "%esi", | |
| 32 | + "%edi", | |
| 33 | +}; | |
| 34 | + | |
| 35 | +int tcg_target_reg_alloc_order[TCG_TARGET_NB_REGS] = { | |
| 36 | + TCG_REG_EAX, | |
| 37 | + TCG_REG_EDX, | |
| 38 | + TCG_REG_ECX, | |
| 39 | + TCG_REG_EBX, | |
| 40 | + TCG_REG_ESI, | |
| 41 | + TCG_REG_EDI, | |
| 42 | + TCG_REG_EBP, | |
| 43 | + TCG_REG_ESP, | |
| 44 | +}; | |
| 45 | + | |
| 46 | +const int tcg_target_call_iarg_regs[3] = { TCG_REG_EAX, TCG_REG_EDX, TCG_REG_ECX }; | |
| 47 | +const int tcg_target_call_oarg_regs[2] = { TCG_REG_EAX, TCG_REG_EDX }; | |
| 48 | + | |
| 49 | +static void patch_reloc(uint8_t *code_ptr, int type, | |
| 50 | + tcg_target_long value) | |
| 51 | +{ | |
| 52 | + switch(type) { | |
| 53 | + case R_386_32: | |
| 54 | + *(uint32_t *)code_ptr = value; | |
| 55 | + break; | |
| 56 | + case R_386_PC32: | |
| 57 | + *(uint32_t *)code_ptr = value - (long)code_ptr; | |
| 58 | + break; | |
| 59 | + default: | |
| 60 | + tcg_abort(); | |
| 61 | + } | |
| 62 | +} | |
| 63 | + | |
| 64 | +/* maximum number of register used for input function arguments */ | |
| 65 | +static inline int tcg_target_get_call_iarg_regs_count(int flags) | |
| 66 | +{ | |
| 67 | + flags &= TCG_CALL_TYPE_MASK; | |
| 68 | + switch(flags) { | |
| 69 | + case TCG_CALL_TYPE_STD: | |
| 70 | + return 0; | |
| 71 | + case TCG_CALL_TYPE_REGPARM_1: | |
| 72 | + case TCG_CALL_TYPE_REGPARM_2: | |
| 73 | + case TCG_CALL_TYPE_REGPARM: | |
| 74 | + return flags - TCG_CALL_TYPE_REGPARM_1 + 1; | |
| 75 | + default: | |
| 76 | + tcg_abort(); | |
| 77 | + } | |
| 78 | +} | |
| 79 | + | |
| 80 | +/* parse target specific constraints */ | |
| 81 | +int target_parse_constraint(TCGArgConstraint *ct, const char **pct_str) | |
| 82 | +{ | |
| 83 | + const char *ct_str; | |
| 84 | + | |
| 85 | + ct_str = *pct_str; | |
| 86 | + switch(ct_str[0]) { | |
| 87 | + case 'a': | |
| 88 | + ct->ct |= TCG_CT_REG; | |
| 89 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EAX); | |
| 90 | + break; | |
| 91 | + case 'b': | |
| 92 | + ct->ct |= TCG_CT_REG; | |
| 93 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EBX); | |
| 94 | + break; | |
| 95 | + case 'c': | |
| 96 | + ct->ct |= TCG_CT_REG; | |
| 97 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_ECX); | |
| 98 | + break; | |
| 99 | + case 'd': | |
| 100 | + ct->ct |= TCG_CT_REG; | |
| 101 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EDX); | |
| 102 | + break; | |
| 103 | + case 'S': | |
| 104 | + ct->ct |= TCG_CT_REG; | |
| 105 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_ESI); | |
| 106 | + break; | |
| 107 | + case 'D': | |
| 108 | + ct->ct |= TCG_CT_REG; | |
| 109 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EDI); | |
| 110 | + break; | |
| 111 | + case 'q': | |
| 112 | + ct->ct |= TCG_CT_REG; | |
| 113 | + tcg_regset_set32(ct->u.regs, 0, 0xf); | |
| 114 | + break; | |
| 115 | + case 'r': | |
| 116 | + ct->ct |= TCG_CT_REG; | |
| 117 | + tcg_regset_set32(ct->u.regs, 0, 0xff); | |
| 118 | + break; | |
| 119 | + | |
| 120 | + /* qemu_ld/st address constraint */ | |
| 121 | + case 'L': | |
| 122 | + ct->ct |= TCG_CT_REG; | |
| 123 | + tcg_regset_set32(ct->u.regs, 0, 0xff); | |
| 124 | + tcg_regset_reset_reg(ct->u.regs, TCG_REG_EAX); | |
| 125 | + tcg_regset_reset_reg(ct->u.regs, TCG_REG_EDX); | |
| 126 | + break; | |
| 127 | + default: | |
| 128 | + return -1; | |
| 129 | + } | |
| 130 | + ct_str++; | |
| 131 | + *pct_str = ct_str; | |
| 132 | + return 0; | |
| 133 | +} | |
| 134 | + | |
| 135 | +/* test if a constant matches the constraint */ | |
| 136 | +static inline int tcg_target_const_match(tcg_target_long val, | |
| 137 | + const TCGArgConstraint *arg_ct) | |
| 138 | +{ | |
| 139 | + int ct; | |
| 140 | + ct = arg_ct->ct; | |
| 141 | + if (ct & TCG_CT_CONST) | |
| 142 | + return 1; | |
| 143 | + else | |
| 144 | + return 0; | |
| 145 | +} | |
| 146 | + | |
| 147 | +#define ARITH_ADD 0 | |
| 148 | +#define ARITH_OR 1 | |
| 149 | +#define ARITH_ADC 2 | |
| 150 | +#define ARITH_SBB 3 | |
| 151 | +#define ARITH_AND 4 | |
| 152 | +#define ARITH_SUB 5 | |
| 153 | +#define ARITH_XOR 6 | |
| 154 | +#define ARITH_CMP 7 | |
| 155 | + | |
| 156 | +#define SHIFT_SHL 4 | |
| 157 | +#define SHIFT_SHR 5 | |
| 158 | +#define SHIFT_SAR 7 | |
| 159 | + | |
| 160 | +#define JCC_JMP (-1) | |
| 161 | +#define JCC_JO 0x0 | |
| 162 | +#define JCC_JNO 0x1 | |
| 163 | +#define JCC_JB 0x2 | |
| 164 | +#define JCC_JAE 0x3 | |
| 165 | +#define JCC_JE 0x4 | |
| 166 | +#define JCC_JNE 0x5 | |
| 167 | +#define JCC_JBE 0x6 | |
| 168 | +#define JCC_JA 0x7 | |
| 169 | +#define JCC_JS 0x8 | |
| 170 | +#define JCC_JNS 0x9 | |
| 171 | +#define JCC_JP 0xa | |
| 172 | +#define JCC_JNP 0xb | |
| 173 | +#define JCC_JL 0xc | |
| 174 | +#define JCC_JGE 0xd | |
| 175 | +#define JCC_JLE 0xe | |
| 176 | +#define JCC_JG 0xf | |
| 177 | + | |
| 178 | +#define P_EXT 0x100 /* 0x0f opcode prefix */ | |
| 179 | + | |
| 180 | +static const uint8_t tcg_cond_to_jcc[10] = { | |
| 181 | + [TCG_COND_EQ] = JCC_JE, | |
| 182 | + [TCG_COND_NE] = JCC_JNE, | |
| 183 | + [TCG_COND_LT] = JCC_JL, | |
| 184 | + [TCG_COND_GE] = JCC_JGE, | |
| 185 | + [TCG_COND_LE] = JCC_JLE, | |
| 186 | + [TCG_COND_GT] = JCC_JG, | |
| 187 | + [TCG_COND_LTU] = JCC_JB, | |
| 188 | + [TCG_COND_GEU] = JCC_JAE, | |
| 189 | + [TCG_COND_LEU] = JCC_JBE, | |
| 190 | + [TCG_COND_GTU] = JCC_JA, | |
| 191 | +}; | |
| 192 | + | |
| 193 | +static inline void tcg_out_opc(TCGContext *s, int opc) | |
| 194 | +{ | |
| 195 | + if (opc & P_EXT) | |
| 196 | + tcg_out8(s, 0x0f); | |
| 197 | + tcg_out8(s, opc); | |
| 198 | +} | |
| 199 | + | |
| 200 | +static inline void tcg_out_modrm(TCGContext *s, int opc, int r, int rm) | |
| 201 | +{ | |
| 202 | + tcg_out_opc(s, opc); | |
| 203 | + tcg_out8(s, 0xc0 | (r << 3) | rm); | |
| 204 | +} | |
| 205 | + | |
| 206 | +/* rm == -1 means no register index */ | |
| 207 | +static inline void tcg_out_modrm_offset(TCGContext *s, int opc, int r, int rm, | |
| 208 | + int32_t offset) | |
| 209 | +{ | |
| 210 | + tcg_out_opc(s, opc); | |
| 211 | + if (rm == -1) { | |
| 212 | + tcg_out8(s, 0x05 | (r << 3)); | |
| 213 | + tcg_out32(s, offset); | |
| 214 | + } else if (offset == 0 && rm != TCG_REG_EBP) { | |
| 215 | + if (rm == TCG_REG_ESP) { | |
| 216 | + tcg_out8(s, 0x04 | (r << 3)); | |
| 217 | + tcg_out8(s, 0x24); | |
| 218 | + } else { | |
| 219 | + tcg_out8(s, 0x00 | (r << 3) | rm); | |
| 220 | + } | |
| 221 | + } else if ((int8_t)offset == offset) { | |
| 222 | + if (rm == TCG_REG_ESP) { | |
| 223 | + tcg_out8(s, 0x44 | (r << 3)); | |
| 224 | + tcg_out8(s, 0x24); | |
| 225 | + } else { | |
| 226 | + tcg_out8(s, 0x40 | (r << 3) | rm); | |
| 227 | + } | |
| 228 | + tcg_out8(s, offset); | |
| 229 | + } else { | |
| 230 | + if (rm == TCG_REG_ESP) { | |
| 231 | + tcg_out8(s, 0x84 | (r << 3)); | |
| 232 | + tcg_out8(s, 0x24); | |
| 233 | + } else { | |
| 234 | + tcg_out8(s, 0x80 | (r << 3) | rm); | |
| 235 | + } | |
| 236 | + tcg_out32(s, offset); | |
| 237 | + } | |
| 238 | +} | |
| 239 | + | |
| 240 | +static inline void tcg_out_mov(TCGContext *s, int ret, int arg) | |
| 241 | +{ | |
| 242 | + if (arg != ret) | |
| 243 | + tcg_out_modrm(s, 0x8b, ret, arg); | |
| 244 | +} | |
| 245 | + | |
| 246 | +static inline void tcg_out_movi(TCGContext *s, TCGType type, | |
| 247 | + int ret, int32_t arg) | |
| 248 | +{ | |
| 249 | + if (arg == 0) { | |
| 250 | + /* xor r0,r0 */ | |
| 251 | + tcg_out_modrm(s, 0x01 | (ARITH_XOR << 3), ret, ret); | |
| 252 | + } else { | |
| 253 | + tcg_out8(s, 0xb8 + ret); | |
| 254 | + tcg_out32(s, arg); | |
| 255 | + } | |
| 256 | +} | |
| 257 | + | |
| 258 | +static inline void tcg_out_ld(TCGContext *s, int ret, | |
| 259 | + int arg1, int32_t arg2) | |
| 260 | +{ | |
| 261 | + /* movl */ | |
| 262 | + tcg_out_modrm_offset(s, 0x8b, ret, arg1, arg2); | |
| 263 | +} | |
| 264 | + | |
| 265 | +static inline void tcg_out_st(TCGContext *s, int arg, | |
| 266 | + int arg1, int32_t arg2) | |
| 267 | +{ | |
| 268 | + /* movl */ | |
| 269 | + tcg_out_modrm_offset(s, 0x89, arg, arg1, arg2); | |
| 270 | +} | |
| 271 | + | |
| 272 | +static inline void tgen_arithi(TCGContext *s, int c, int r0, int32_t val) | |
| 273 | +{ | |
| 274 | + if (val == (int8_t)val) { | |
| 275 | + tcg_out_modrm(s, 0x83, c, r0); | |
| 276 | + tcg_out8(s, val); | |
| 277 | + } else { | |
| 278 | + tcg_out_modrm(s, 0x81, c, r0); | |
| 279 | + tcg_out32(s, val); | |
| 280 | + } | |
| 281 | +} | |
| 282 | + | |
| 283 | +void tcg_out_addi(TCGContext *s, int reg, tcg_target_long val) | |
| 284 | +{ | |
| 285 | + if (val != 0) | |
| 286 | + tgen_arithi(s, ARITH_ADD, reg, val); | |
| 287 | +} | |
| 288 | + | |
| 289 | +static void tcg_out_jxx(TCGContext *s, int opc, int label_index) | |
| 290 | +{ | |
| 291 | + int32_t val, val1; | |
| 292 | + TCGLabel *l = &s->labels[label_index]; | |
| 293 | + | |
| 294 | + if (l->has_value) { | |
| 295 | + val = l->u.value - (tcg_target_long)s->code_ptr; | |
| 296 | + val1 = val - 2; | |
| 297 | + if ((int8_t)val1 == val1) { | |
| 298 | + if (opc == -1) | |
| 299 | + tcg_out8(s, 0xeb); | |
| 300 | + else | |
| 301 | + tcg_out8(s, 0x70 + opc); | |
| 302 | + tcg_out8(s, val1); | |
| 303 | + } else { | |
| 304 | + if (opc == -1) { | |
| 305 | + tcg_out8(s, 0xe9); | |
| 306 | + tcg_out32(s, val - 5); | |
| 307 | + } else { | |
| 308 | + tcg_out8(s, 0x0f); | |
| 309 | + tcg_out8(s, 0x80 + opc); | |
| 310 | + tcg_out32(s, val - 6); | |
| 311 | + } | |
| 312 | + } | |
| 313 | + } else { | |
| 314 | + if (opc == -1) { | |
| 315 | + tcg_out8(s, 0xe9); | |
| 316 | + } else { | |
| 317 | + tcg_out8(s, 0x0f); | |
| 318 | + tcg_out8(s, 0x80 + opc); | |
| 319 | + } | |
| 320 | + tcg_out_reloc(s, s->code_ptr, R_386_PC32, label_index, -4); | |
| 321 | + tcg_out32(s, -4); | |
| 322 | + } | |
| 323 | +} | |
| 324 | + | |
| 325 | +static void tcg_out_brcond(TCGContext *s, int cond, | |
| 326 | + TCGArg arg1, TCGArg arg2, int const_arg2, | |
| 327 | + int label_index) | |
| 328 | +{ | |
| 329 | + int c; | |
| 330 | + if (const_arg2) { | |
| 331 | + if (arg2 == 0) { | |
| 332 | + /* use test */ | |
| 333 | + switch(cond) { | |
| 334 | + case TCG_COND_EQ: | |
| 335 | + c = JCC_JNE; | |
| 336 | + break; | |
| 337 | + case TCG_COND_NE: | |
| 338 | + c = JCC_JNE; | |
| 339 | + break; | |
| 340 | + case TCG_COND_LT: | |
| 341 | + c = JCC_JS; | |
| 342 | + break; | |
| 343 | + case TCG_COND_GE: | |
| 344 | + c = JCC_JNS; | |
| 345 | + break; | |
| 346 | + default: | |
| 347 | + goto do_cmpi; | |
| 348 | + } | |
| 349 | + /* test r, r */ | |
| 350 | + tcg_out_modrm(s, 0x85, arg1, arg1); | |
| 351 | + tcg_out_jxx(s, c, label_index); | |
| 352 | + } else { | |
| 353 | + do_cmpi: | |
| 354 | + tgen_arithi(s, ARITH_CMP, arg1, arg2); | |
| 355 | + tcg_out_jxx(s, tcg_cond_to_jcc[cond], label_index); | |
| 356 | + } | |
| 357 | + } else { | |
| 358 | + tcg_out_modrm(s, 0x01 | (ARITH_CMP << 3), arg1, arg2); | |
| 359 | + tcg_out_jxx(s, tcg_cond_to_jcc[cond], label_index); | |
| 360 | + } | |
| 361 | +} | |
| 362 | + | |
| 363 | +/* XXX: we implement it at the target level to avoid having to | |
| 364 | + handle cross basic blocks temporaries */ | |
| 365 | +static void tcg_out_brcond2(TCGContext *s, | |
| 366 | + const TCGArg *args, const int *const_args) | |
| 367 | +{ | |
| 368 | + int label_next; | |
| 369 | + label_next = gen_new_label(); | |
| 370 | + switch(args[4]) { | |
| 371 | + case TCG_COND_EQ: | |
| 372 | + tcg_out_brcond(s, TCG_COND_NE, args[0], args[2], const_args[2], label_next); | |
| 373 | + tcg_out_brcond(s, TCG_COND_EQ, args[1], args[3], const_args[3], args[5]); | |
| 374 | + break; | |
| 375 | + case TCG_COND_NE: | |
| 376 | + tcg_out_brcond(s, TCG_COND_NE, args[0], args[2], const_args[2], args[5]); | |
| 377 | + tcg_out_brcond(s, TCG_COND_EQ, args[1], args[3], const_args[3], label_next); | |
| 378 | + break; | |
| 379 | + case TCG_COND_LT: | |
| 380 | + tcg_out_brcond(s, TCG_COND_LT, args[1], args[3], const_args[3], args[5]); | |
| 381 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | |
| 382 | + tcg_out_brcond(s, TCG_COND_LT, args[0], args[2], const_args[2], args[5]); | |
| 383 | + break; | |
| 384 | + case TCG_COND_LE: | |
| 385 | + tcg_out_brcond(s, TCG_COND_LT, args[1], args[3], const_args[3], args[5]); | |
| 386 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | |
| 387 | + tcg_out_brcond(s, TCG_COND_LE, args[0], args[2], const_args[2], args[5]); | |
| 388 | + break; | |
| 389 | + case TCG_COND_GT: | |
| 390 | + tcg_out_brcond(s, TCG_COND_GT, args[1], args[3], const_args[3], args[5]); | |
| 391 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | |
| 392 | + tcg_out_brcond(s, TCG_COND_GT, args[0], args[2], const_args[2], args[5]); | |
| 393 | + break; | |
| 394 | + case TCG_COND_GE: | |
| 395 | + tcg_out_brcond(s, TCG_COND_GT, args[1], args[3], const_args[3], args[5]); | |
| 396 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | |
| 397 | + tcg_out_brcond(s, TCG_COND_GE, args[0], args[2], const_args[2], args[5]); | |
| 398 | + break; | |
| 399 | + case TCG_COND_LTU: | |
| 400 | + tcg_out_brcond(s, TCG_COND_LTU, args[1], args[3], const_args[3], args[5]); | |
| 401 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | |
| 402 | + tcg_out_brcond(s, TCG_COND_LTU, args[0], args[2], const_args[2], args[5]); | |
| 403 | + break; | |
| 404 | + case TCG_COND_LEU: | |
| 405 | + tcg_out_brcond(s, TCG_COND_LTU, args[1], args[3], const_args[3], args[5]); | |
| 406 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | |
| 407 | + tcg_out_brcond(s, TCG_COND_LEU, args[0], args[2], const_args[2], args[5]); | |
| 408 | + break; | |
| 409 | + case TCG_COND_GTU: | |
| 410 | + tcg_out_brcond(s, TCG_COND_GTU, args[1], args[3], const_args[3], args[5]); | |
| 411 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | |
| 412 | + tcg_out_brcond(s, TCG_COND_GTU, args[0], args[2], const_args[2], args[5]); | |
| 413 | + break; | |
| 414 | + case TCG_COND_GEU: | |
| 415 | + tcg_out_brcond(s, TCG_COND_GTU, args[1], args[3], const_args[3], args[5]); | |
| 416 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | |
| 417 | + tcg_out_brcond(s, TCG_COND_GEU, args[0], args[2], const_args[2], args[5]); | |
| 418 | + break; | |
| 419 | + default: | |
| 420 | + tcg_abort(); | |
| 421 | + } | |
| 422 | + tcg_out_label(s, label_next, (tcg_target_long)s->code_ptr); | |
| 423 | +} | |
| 424 | + | |
| 425 | +#if defined(CONFIG_SOFTMMU) | |
| 426 | +extern void __ldb_mmu(void); | |
| 427 | +extern void __ldw_mmu(void); | |
| 428 | +extern void __ldl_mmu(void); | |
| 429 | +extern void __ldq_mmu(void); | |
| 430 | + | |
| 431 | +extern void __stb_mmu(void); | |
| 432 | +extern void __stw_mmu(void); | |
| 433 | +extern void __stl_mmu(void); | |
| 434 | +extern void __stq_mmu(void); | |
| 435 | + | |
| 436 | +static void *qemu_ld_helpers[4] = { | |
| 437 | + __ldb_mmu, | |
| 438 | + __ldw_mmu, | |
| 439 | + __ldl_mmu, | |
| 440 | + __ldq_mmu, | |
| 441 | +}; | |
| 442 | + | |
| 443 | +static void *qemu_st_helpers[4] = { | |
| 444 | + __stb_mmu, | |
| 445 | + __stw_mmu, | |
| 446 | + __stl_mmu, | |
| 447 | + __stq_mmu, | |
| 448 | +}; | |
| 449 | +#endif | |
| 450 | + | |
| 451 | +/* XXX: qemu_ld and qemu_st could be modified to clobber only EDX and | |
| 452 | + EAX. It will be useful once fixed registers globals are less | |
| 453 | + common. */ | |
| 454 | +static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, | |
| 455 | + int opc) | |
| 456 | +{ | |
| 457 | + int addr_reg, data_reg, data_reg2, r0, r1, mem_index, s_bits, bswap; | |
| 458 | +#if defined(CONFIG_SOFTMMU) | |
| 459 | + uint8_t *label1_ptr, *label2_ptr; | |
| 460 | +#endif | |
| 461 | +#if TARGET_LONG_BITS == 64 | |
| 462 | +#if defined(CONFIG_SOFTMMU) | |
| 463 | + uint8_t *label3_ptr; | |
| 464 | +#endif | |
| 465 | + int addr_reg2; | |
| 466 | +#endif | |
| 467 | + | |
| 468 | + data_reg = *args++; | |
| 469 | + if (opc == 3) | |
| 470 | + data_reg2 = *args++; | |
| 471 | + else | |
| 472 | + data_reg2 = 0; | |
| 473 | + addr_reg = *args++; | |
| 474 | +#if TARGET_LONG_BITS == 64 | |
| 475 | + addr_reg2 = *args++; | |
| 476 | +#endif | |
| 477 | + mem_index = *args; | |
| 478 | + s_bits = opc & 3; | |
| 479 | + | |
| 480 | + r0 = TCG_REG_EAX; | |
| 481 | + r1 = TCG_REG_EDX; | |
| 482 | + | |
| 483 | +#if defined(CONFIG_SOFTMMU) | |
| 484 | + tcg_out_mov(s, r1, addr_reg); | |
| 485 | + | |
| 486 | + tcg_out_mov(s, r0, addr_reg); | |
| 487 | + | |
| 488 | + tcg_out_modrm(s, 0xc1, 5, r1); /* shr $x, r1 */ | |
| 489 | + tcg_out8(s, TARGET_PAGE_BITS - CPU_TLB_ENTRY_BITS); | |
| 490 | + | |
| 491 | + tcg_out_modrm(s, 0x81, 4, r0); /* andl $x, r0 */ | |
| 492 | + tcg_out32(s, TARGET_PAGE_MASK | ((1 << s_bits) - 1)); | |
| 493 | + | |
| 494 | + tcg_out_modrm(s, 0x81, 4, r1); /* andl $x, r1 */ | |
| 495 | + tcg_out32(s, (CPU_TLB_SIZE - 1) << CPU_TLB_ENTRY_BITS); | |
| 496 | + | |
| 497 | + tcg_out_opc(s, 0x8d); /* lea offset(r1, %ebp), r1 */ | |
| 498 | + tcg_out8(s, 0x80 | (r1 << 3) | 0x04); | |
| 499 | + tcg_out8(s, (5 << 3) | r1); | |
| 500 | + tcg_out32(s, offsetof(CPUState, tlb_table[mem_index][0].addr_read)); | |
| 501 | + | |
| 502 | + /* cmp 0(r1), r0 */ | |
| 503 | + tcg_out_modrm_offset(s, 0x3b, r0, r1, 0); | |
| 504 | + | |
| 505 | + tcg_out_mov(s, r0, addr_reg); | |
| 506 | + | |
| 507 | +#if TARGET_LONG_BITS == 32 | |
| 508 | + /* je label1 */ | |
| 509 | + tcg_out8(s, 0x70 + JCC_JE); | |
| 510 | + label1_ptr = s->code_ptr; | |
| 511 | + s->code_ptr++; | |
| 512 | +#else | |
| 513 | + /* jne label3 */ | |
| 514 | + tcg_out8(s, 0x70 + JCC_JNE); | |
| 515 | + label3_ptr = s->code_ptr; | |
| 516 | + s->code_ptr++; | |
| 517 | + | |
| 518 | + /* cmp 4(r1), addr_reg2 */ | |
| 519 | + tcg_out_modrm_offset(s, 0x3b, addr_reg2, r1, 4); | |
| 520 | + | |
| 521 | + /* je label1 */ | |
| 522 | + tcg_out8(s, 0x70 + JCC_JE); | |
| 523 | + label1_ptr = s->code_ptr; | |
| 524 | + s->code_ptr++; | |
| 525 | + | |
| 526 | + /* label3: */ | |
| 527 | + *label3_ptr = s->code_ptr - label3_ptr - 1; | |
| 528 | +#endif | |
| 529 | + | |
| 530 | + /* XXX: move that code at the end of the TB */ | |
| 531 | +#if TARGET_LONG_BITS == 32 | |
| 532 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_EDX, mem_index); | |
| 533 | +#else | |
| 534 | + tcg_out_mov(s, TCG_REG_EDX, addr_reg2); | |
| 535 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_ECX, mem_index); | |
| 536 | +#endif | |
| 537 | + tcg_out8(s, 0xe8); | |
| 538 | + tcg_out32(s, (tcg_target_long)qemu_ld_helpers[s_bits] - | |
| 539 | + (tcg_target_long)s->code_ptr - 4); | |
| 540 | + | |
| 541 | + switch(opc) { | |
| 542 | + case 0 | 4: | |
| 543 | + /* movsbl */ | |
| 544 | + tcg_out_modrm(s, 0xbe | P_EXT, data_reg, TCG_REG_EAX); | |
| 545 | + break; | |
| 546 | + case 1 | 4: | |
| 547 | + /* movswl */ | |
| 548 | + tcg_out_modrm(s, 0xbf | P_EXT, data_reg, TCG_REG_EAX); | |
| 549 | + break; | |
| 550 | + case 0: | |
| 551 | + case 1: | |
| 552 | + case 2: | |
| 553 | + default: | |
| 554 | + tcg_out_mov(s, data_reg, TCG_REG_EAX); | |
| 555 | + break; | |
| 556 | + case 3: | |
| 557 | + if (data_reg == TCG_REG_EDX) { | |
| 558 | + tcg_out_opc(s, 0x90 + TCG_REG_EDX); /* xchg %edx, %eax */ | |
| 559 | + tcg_out_mov(s, data_reg2, TCG_REG_EAX); | |
| 560 | + } else { | |
| 561 | + tcg_out_mov(s, data_reg, TCG_REG_EAX); | |
| 562 | + tcg_out_mov(s, data_reg2, TCG_REG_EDX); | |
| 563 | + } | |
| 564 | + break; | |
| 565 | + } | |
| 566 | + | |
| 567 | + /* jmp label2 */ | |
| 568 | + tcg_out8(s, 0xeb); | |
| 569 | + label2_ptr = s->code_ptr; | |
| 570 | + s->code_ptr++; | |
| 571 | + | |
| 572 | + /* label1: */ | |
| 573 | + *label1_ptr = s->code_ptr - label1_ptr - 1; | |
| 574 | + | |
| 575 | + /* add x(r1), r0 */ | |
| 576 | + tcg_out_modrm_offset(s, 0x03, r0, r1, offsetof(CPUTLBEntry, addend) - | |
| 577 | + offsetof(CPUTLBEntry, addr_read)); | |
| 578 | +#else | |
| 579 | + r0 = addr_reg; | |
| 580 | +#endif | |
| 581 | + | |
| 582 | +#ifdef TARGET_WORDS_BIGENDIAN | |
| 583 | + bswap = 1; | |
| 584 | +#else | |
| 585 | + bswap = 0; | |
| 586 | +#endif | |
| 587 | + switch(opc) { | |
| 588 | + case 0: | |
| 589 | + /* movzbl */ | |
| 590 | + tcg_out_modrm_offset(s, 0xb6 | P_EXT, data_reg, r0, 0); | |
| 591 | + break; | |
| 592 | + case 0 | 4: | |
| 593 | + /* movsbl */ | |
| 594 | + tcg_out_modrm_offset(s, 0xbe | P_EXT, data_reg, r0, 0); | |
| 595 | + break; | |
| 596 | + case 1: | |
| 597 | + /* movzwl */ | |
| 598 | + tcg_out_modrm_offset(s, 0xb7 | P_EXT, data_reg, r0, 0); | |
| 599 | + if (bswap) { | |
| 600 | + /* rolw $8, data_reg */ | |
| 601 | + tcg_out8(s, 0x66); | |
| 602 | + tcg_out_modrm(s, 0xc1, 0, data_reg); | |
| 603 | + tcg_out8(s, 8); | |
| 604 | + } | |
| 605 | + break; | |
| 606 | + case 1 | 4: | |
| 607 | + /* movswl */ | |
| 608 | + tcg_out_modrm_offset(s, 0xbf | P_EXT, data_reg, r0, 0); | |
| 609 | + if (bswap) { | |
| 610 | + /* rolw $8, data_reg */ | |
| 611 | + tcg_out8(s, 0x66); | |
| 612 | + tcg_out_modrm(s, 0xc1, 0, data_reg); | |
| 613 | + tcg_out8(s, 8); | |
| 614 | + | |
| 615 | + /* movswl data_reg, data_reg */ | |
| 616 | + tcg_out_modrm(s, 0xbf | P_EXT, data_reg, data_reg); | |
| 617 | + } | |
| 618 | + break; | |
| 619 | + case 2: | |
| 620 | + /* movl (r0), data_reg */ | |
| 621 | + tcg_out_modrm_offset(s, 0x8b, data_reg, r0, 0); | |
| 622 | + if (bswap) { | |
| 623 | + /* bswap */ | |
| 624 | + tcg_out_opc(s, (0xc8 + data_reg) | P_EXT); | |
| 625 | + } | |
| 626 | + break; | |
| 627 | + case 3: | |
| 628 | + /* XXX: could be nicer */ | |
| 629 | + if (r0 == data_reg) { | |
| 630 | + r1 = TCG_REG_EDX; | |
| 631 | + if (r1 == data_reg) | |
| 632 | + r1 = TCG_REG_EAX; | |
| 633 | + tcg_out_mov(s, r1, r0); | |
| 634 | + r0 = r1; | |
| 635 | + } | |
| 636 | + if (!bswap) { | |
| 637 | + tcg_out_modrm_offset(s, 0x8b, data_reg, r0, 0); | |
| 638 | + tcg_out_modrm_offset(s, 0x8b, data_reg2, r0, 4); | |
| 639 | + } else { | |
| 640 | + tcg_out_modrm_offset(s, 0x8b, data_reg, r0, 4); | |
| 641 | + tcg_out_opc(s, (0xc8 + data_reg) | P_EXT); | |
| 642 | + | |
| 643 | + tcg_out_modrm_offset(s, 0x8b, data_reg2, r0, 0); | |
| 644 | + /* bswap */ | |
| 645 | + tcg_out_opc(s, (0xc8 + data_reg2) | P_EXT); | |
| 646 | + } | |
| 647 | + break; | |
| 648 | + default: | |
| 649 | + tcg_abort(); | |
| 650 | + } | |
| 651 | + | |
| 652 | +#if defined(CONFIG_SOFTMMU) | |
| 653 | + /* label2: */ | |
| 654 | + *label2_ptr = s->code_ptr - label2_ptr - 1; | |
| 655 | +#endif | |
| 656 | +} | |
| 657 | + | |
| 658 | + | |
| 659 | +static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, | |
| 660 | + int opc) | |
| 661 | +{ | |
| 662 | + int addr_reg, data_reg, data_reg2, r0, r1, mem_index, s_bits, bswap; | |
| 663 | +#if defined(CONFIG_SOFTMMU) | |
| 664 | + uint8_t *label1_ptr, *label2_ptr; | |
| 665 | +#endif | |
| 666 | +#if TARGET_LONG_BITS == 64 | |
| 667 | +#if defined(CONFIG_SOFTMMU) | |
| 668 | + uint8_t *label3_ptr; | |
| 669 | +#endif | |
| 670 | + int addr_reg2; | |
| 671 | +#endif | |
| 672 | + | |
| 673 | + data_reg = *args++; | |
| 674 | + if (opc == 3) | |
| 675 | + data_reg2 = *args++; | |
| 676 | + else | |
| 677 | + data_reg2 = 0; | |
| 678 | + addr_reg = *args++; | |
| 679 | +#if TARGET_LONG_BITS == 64 | |
| 680 | + addr_reg2 = *args++; | |
| 681 | +#endif | |
| 682 | + mem_index = *args; | |
| 683 | + | |
| 684 | + s_bits = opc; | |
| 685 | + | |
| 686 | + r0 = TCG_REG_EAX; | |
| 687 | + r1 = TCG_REG_EDX; | |
| 688 | + | |
| 689 | +#if defined(CONFIG_SOFTMMU) | |
| 690 | + tcg_out_mov(s, r1, addr_reg); | |
| 691 | + | |
| 692 | + tcg_out_mov(s, r0, addr_reg); | |
| 693 | + | |
| 694 | + tcg_out_modrm(s, 0xc1, 5, r1); /* shr $x, r1 */ | |
| 695 | + tcg_out8(s, TARGET_PAGE_BITS - CPU_TLB_ENTRY_BITS); | |
| 696 | + | |
| 697 | + tcg_out_modrm(s, 0x81, 4, r0); /* andl $x, r0 */ | |
| 698 | + tcg_out32(s, TARGET_PAGE_MASK | ((1 << s_bits) - 1)); | |
| 699 | + | |
| 700 | + tcg_out_modrm(s, 0x81, 4, r1); /* andl $x, r1 */ | |
| 701 | + tcg_out32(s, (CPU_TLB_SIZE - 1) << CPU_TLB_ENTRY_BITS); | |
| 702 | + | |
| 703 | + tcg_out_opc(s, 0x8d); /* lea offset(r1, %ebp), r1 */ | |
| 704 | + tcg_out8(s, 0x80 | (r1 << 3) | 0x04); | |
| 705 | + tcg_out8(s, (5 << 3) | r1); | |
| 706 | + tcg_out32(s, offsetof(CPUState, tlb_table[mem_index][0].addr_write)); | |
| 707 | + | |
| 708 | + /* cmp 0(r1), r0 */ | |
| 709 | + tcg_out_modrm_offset(s, 0x3b, r0, r1, 0); | |
| 710 | + | |
| 711 | + tcg_out_mov(s, r0, addr_reg); | |
| 712 | + | |
| 713 | +#if TARGET_LONG_BITS == 32 | |
| 714 | + /* je label1 */ | |
| 715 | + tcg_out8(s, 0x70 + JCC_JE); | |
| 716 | + label1_ptr = s->code_ptr; | |
| 717 | + s->code_ptr++; | |
| 718 | +#else | |
| 719 | + /* jne label3 */ | |
| 720 | + tcg_out8(s, 0x70 + JCC_JNE); | |
| 721 | + label3_ptr = s->code_ptr; | |
| 722 | + s->code_ptr++; | |
| 723 | + | |
| 724 | + /* cmp 4(r1), addr_reg2 */ | |
| 725 | + tcg_out_modrm_offset(s, 0x3b, addr_reg2, r1, 4); | |
| 726 | + | |
| 727 | + /* je label1 */ | |
| 728 | + tcg_out8(s, 0x70 + JCC_JE); | |
| 729 | + label1_ptr = s->code_ptr; | |
| 730 | + s->code_ptr++; | |
| 731 | + | |
| 732 | + /* label3: */ | |
| 733 | + *label3_ptr = s->code_ptr - label3_ptr - 1; | |
| 734 | +#endif | |
| 735 | + | |
| 736 | + /* XXX: move that code at the end of the TB */ | |
| 737 | +#if TARGET_LONG_BITS == 32 | |
| 738 | + if (opc == 3) { | |
| 739 | + tcg_out_mov(s, TCG_REG_EDX, data_reg); | |
| 740 | + tcg_out_mov(s, TCG_REG_ECX, data_reg2); | |
| 741 | + tcg_out8(s, 0x6a); /* push Ib */ | |
| 742 | + tcg_out8(s, mem_index); | |
| 743 | + tcg_out8(s, 0xe8); | |
| 744 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | |
| 745 | + (tcg_target_long)s->code_ptr - 4); | |
| 746 | + tcg_out_addi(s, TCG_REG_ESP, 4); | |
| 747 | + } else { | |
| 748 | + switch(opc) { | |
| 749 | + case 0: | |
| 750 | + /* movzbl */ | |
| 751 | + tcg_out_modrm(s, 0xb6 | P_EXT, TCG_REG_EDX, data_reg); | |
| 752 | + break; | |
| 753 | + case 1: | |
| 754 | + /* movzwl */ | |
| 755 | + tcg_out_modrm(s, 0xb7 | P_EXT, TCG_REG_EDX, data_reg); | |
| 756 | + break; | |
| 757 | + case 2: | |
| 758 | + tcg_out_mov(s, TCG_REG_EDX, data_reg); | |
| 759 | + break; | |
| 760 | + } | |
| 761 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_ECX, mem_index); | |
| 762 | + tcg_out8(s, 0xe8); | |
| 763 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | |
| 764 | + (tcg_target_long)s->code_ptr - 4); | |
| 765 | + } | |
| 766 | +#else | |
| 767 | + if (opc == 3) { | |
| 768 | + tcg_out_mov(s, TCG_REG_EDX, addr_reg2); | |
| 769 | + tcg_out8(s, 0x6a); /* push Ib */ | |
| 770 | + tcg_out8(s, mem_index); | |
| 771 | + tcg_out_opc(s, 0x50 + data_reg2); /* push */ | |
| 772 | + tcg_out_opc(s, 0x50 + data_reg); /* push */ | |
| 773 | + tcg_out8(s, 0xe8); | |
| 774 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | |
| 775 | + (tcg_target_long)s->code_ptr - 4); | |
| 776 | + tcg_out_addi(s, TCG_REG_ESP, 12); | |
| 777 | + } else { | |
| 778 | + tcg_out_mov(s, TCG_REG_EDX, addr_reg2); | |
| 779 | + switch(opc) { | |
| 780 | + case 0: | |
| 781 | + /* movzbl */ | |
| 782 | + tcg_out_modrm(s, 0xb6 | P_EXT, TCG_REG_ECX, data_reg); | |
| 783 | + break; | |
| 784 | + case 1: | |
| 785 | + /* movzwl */ | |
| 786 | + tcg_out_modrm(s, 0xb7 | P_EXT, TCG_REG_ECX, data_reg); | |
| 787 | + break; | |
| 788 | + case 2: | |
| 789 | + tcg_out_mov(s, TCG_REG_ECX, data_reg); | |
| 790 | + break; | |
| 791 | + } | |
| 792 | + tcg_out8(s, 0x6a); /* push Ib */ | |
| 793 | + tcg_out8(s, mem_index); | |
| 794 | + tcg_out8(s, 0xe8); | |
| 795 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | |
| 796 | + (tcg_target_long)s->code_ptr - 4); | |
| 797 | + tcg_out_addi(s, TCG_REG_ESP, 4); | |
| 798 | + } | |
| 799 | +#endif | |
| 800 | + | |
| 801 | + /* jmp label2 */ | |
| 802 | + tcg_out8(s, 0xeb); | |
| 803 | + label2_ptr = s->code_ptr; | |
| 804 | + s->code_ptr++; | |
| 805 | + | |
| 806 | + /* label1: */ | |
| 807 | + *label1_ptr = s->code_ptr - label1_ptr - 1; | |
| 808 | + | |
| 809 | + /* add x(r1), r0 */ | |
| 810 | + tcg_out_modrm_offset(s, 0x03, r0, r1, offsetof(CPUTLBEntry, addend) - | |
| 811 | + offsetof(CPUTLBEntry, addr_write)); | |
| 812 | +#else | |
| 813 | + r0 = addr_reg; | |
| 814 | +#endif | |
| 815 | + | |
| 816 | +#ifdef TARGET_WORDS_BIGENDIAN | |
| 817 | + bswap = 1; | |
| 818 | +#else | |
| 819 | + bswap = 0; | |
| 820 | +#endif | |
| 821 | + switch(opc) { | |
| 822 | + case 0: | |
| 823 | + /* movb */ | |
| 824 | + tcg_out_modrm_offset(s, 0x88, data_reg, r0, 0); | |
| 825 | + break; | |
| 826 | + case 1: | |
| 827 | + if (bswap) { | |
| 828 | + tcg_out_mov(s, r1, data_reg); | |
| 829 | + tcg_out8(s, 0x66); /* rolw $8, %ecx */ | |
| 830 | + tcg_out_modrm(s, 0xc1, 0, r1); | |
| 831 | + tcg_out8(s, 8); | |
| 832 | + data_reg = r1; | |
| 833 | + } | |
| 834 | + /* movw */ | |
| 835 | + tcg_out8(s, 0x66); | |
| 836 | + tcg_out_modrm_offset(s, 0x89, data_reg, r0, 0); | |
| 837 | + break; | |
| 838 | + case 2: | |
| 839 | + if (bswap) { | |
| 840 | + tcg_out_mov(s, r1, data_reg); | |
| 841 | + /* bswap data_reg */ | |
| 842 | + tcg_out_opc(s, (0xc8 + r1) | P_EXT); | |
| 843 | + data_reg = r1; | |
| 844 | + } | |
| 845 | + /* movl */ | |
| 846 | + tcg_out_modrm_offset(s, 0x89, data_reg, r0, 0); | |
| 847 | + break; | |
| 848 | + case 3: | |
| 849 | + if (bswap) { | |
| 850 | + tcg_out_mov(s, r1, data_reg2); | |
| 851 | + /* bswap data_reg */ | |
| 852 | + tcg_out_opc(s, (0xc8 + r1) | P_EXT); | |
| 853 | + tcg_out_modrm_offset(s, 0x89, r1, r0, 0); | |
| 854 | + tcg_out_mov(s, r1, data_reg); | |
| 855 | + /* bswap data_reg */ | |
| 856 | + tcg_out_opc(s, (0xc8 + r1) | P_EXT); | |
| 857 | + tcg_out_modrm_offset(s, 0x89, r1, r0, 4); | |
| 858 | + } else { | |
| 859 | + tcg_out_modrm_offset(s, 0x89, data_reg, r0, 0); | |
| 860 | + tcg_out_modrm_offset(s, 0x89, data_reg2, r0, 4); | |
| 861 | + } | |
| 862 | + break; | |
| 863 | + default: | |
| 864 | + tcg_abort(); | |
| 865 | + } | |
| 866 | + | |
| 867 | +#if defined(CONFIG_SOFTMMU) | |
| 868 | + /* label2: */ | |
| 869 | + *label2_ptr = s->code_ptr - label2_ptr - 1; | |
| 870 | +#endif | |
| 871 | +} | |
| 872 | + | |
| 873 | +static inline void tcg_out_op(TCGContext *s, int opc, | |
| 874 | + const TCGArg *args, const int *const_args) | |
| 875 | +{ | |
| 876 | + int c; | |
| 877 | + | |
| 878 | + switch(opc) { | |
| 879 | + case INDEX_op_exit_tb: | |
| 880 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_EAX, args[0]); | |
| 881 | + tcg_out8(s, 0xc3); /* ret */ | |
| 882 | + break; | |
| 883 | + case INDEX_op_goto_tb: | |
| 884 | + if (s->tb_jmp_offset) { | |
| 885 | + /* direct jump method */ | |
| 886 | + tcg_out8(s, 0xe9); /* jmp im */ | |
| 887 | + s->tb_jmp_offset[args[0]] = s->code_ptr - s->code_buf; | |
| 888 | + tcg_out32(s, 0); | |
| 889 | + } else { | |
| 890 | + /* indirect jump method */ | |
| 891 | + /* jmp Ev */ | |
| 892 | + tcg_out_modrm_offset(s, 0xff, 4, -1, | |
| 893 | + (tcg_target_long)(s->tb_next + args[0])); | |
| 894 | + } | |
| 895 | + s->tb_next_offset[args[0]] = s->code_ptr - s->code_buf; | |
| 896 | + break; | |
| 897 | + case INDEX_op_call: | |
| 898 | + if (const_args[0]) { | |
| 899 | + tcg_out8(s, 0xe8); | |
| 900 | + tcg_out32(s, args[0] - (tcg_target_long)s->code_ptr - 4); | |
| 901 | + } else { | |
| 902 | + tcg_out_modrm(s, 0xff, 2, args[0]); | |
| 903 | + } | |
| 904 | + break; | |
| 905 | + case INDEX_op_jmp: | |
| 906 | + if (const_args[0]) { | |
| 907 | + tcg_out8(s, 0xe9); | |
| 908 | + tcg_out32(s, args[0] - (tcg_target_long)s->code_ptr - 4); | |
| 909 | + } else { | |
| 910 | + tcg_out_modrm(s, 0xff, 4, args[0]); | |
| 911 | + } | |
| 912 | + break; | |
| 913 | + case INDEX_op_br: | |
| 914 | + tcg_out_jxx(s, JCC_JMP, args[0]); | |
| 915 | + break; | |
| 916 | + case INDEX_op_movi_i32: | |
| 917 | + tcg_out_movi(s, TCG_TYPE_I32, args[0], args[1]); | |
| 918 | + break; | |
| 919 | + case INDEX_op_ld8u_i32: | |
| 920 | + /* movzbl */ | |
| 921 | + tcg_out_modrm_offset(s, 0xb6 | P_EXT, args[0], args[1], args[2]); | |
| 922 | + break; | |
| 923 | + case INDEX_op_ld8s_i32: | |
| 924 | + /* movsbl */ | |
| 925 | + tcg_out_modrm_offset(s, 0xbe | P_EXT, args[0], args[1], args[2]); | |
| 926 | + break; | |
| 927 | + case INDEX_op_ld16u_i32: | |
| 928 | + /* movzwl */ | |
| 929 | + tcg_out_modrm_offset(s, 0xb7 | P_EXT, args[0], args[1], args[2]); | |
| 930 | + break; | |
| 931 | + case INDEX_op_ld16s_i32: | |
| 932 | + /* movswl */ | |
| 933 | + tcg_out_modrm_offset(s, 0xbf | P_EXT, args[0], args[1], args[2]); | |
| 934 | + break; | |
| 935 | + case INDEX_op_ld_i32: | |
| 936 | + /* movl */ | |
| 937 | + tcg_out_modrm_offset(s, 0x8b, args[0], args[1], args[2]); | |
| 938 | + break; | |
| 939 | + case INDEX_op_st8_i32: | |
| 940 | + /* movb */ | |
| 941 | + tcg_out_modrm_offset(s, 0x88, args[0], args[1], args[2]); | |
| 942 | + break; | |
| 943 | + case INDEX_op_st16_i32: | |
| 944 | + /* movw */ | |
| 945 | + tcg_out8(s, 0x66); | |
| 946 | + tcg_out_modrm_offset(s, 0x89, args[0], args[1], args[2]); | |
| 947 | + break; | |
| 948 | + case INDEX_op_st_i32: | |
| 949 | + /* movl */ | |
| 950 | + tcg_out_modrm_offset(s, 0x89, args[0], args[1], args[2]); | |
| 951 | + break; | |
| 952 | + case INDEX_op_sub_i32: | |
| 953 | + c = ARITH_SUB; | |
| 954 | + goto gen_arith; | |
| 955 | + case INDEX_op_and_i32: | |
| 956 | + c = ARITH_AND; | |
| 957 | + goto gen_arith; | |
| 958 | + case INDEX_op_or_i32: | |
| 959 | + c = ARITH_OR; | |
| 960 | + goto gen_arith; | |
| 961 | + case INDEX_op_xor_i32: | |
| 962 | + c = ARITH_XOR; | |
| 963 | + goto gen_arith; | |
| 964 | + case INDEX_op_add_i32: | |
| 965 | + c = ARITH_ADD; | |
| 966 | + gen_arith: | |
| 967 | + if (const_args[2]) { | |
| 968 | + tgen_arithi(s, c, args[0], args[2]); | |
| 969 | + } else { | |
| 970 | + tcg_out_modrm(s, 0x01 | (c << 3), args[2], args[0]); | |
| 971 | + } | |
| 972 | + break; | |
| 973 | + case INDEX_op_mul_i32: | |
| 974 | + if (const_args[2]) { | |
| 975 | + int32_t val; | |
| 976 | + val = args[2]; | |
| 977 | + if (val == (int8_t)val) { | |
| 978 | + tcg_out_modrm(s, 0x6b, args[0], args[0]); | |
| 979 | + tcg_out8(s, val); | |
| 980 | + } else { | |
| 981 | + tcg_out_modrm(s, 0x69, args[0], args[0]); | |
| 982 | + tcg_out32(s, val); | |
| 983 | + } | |
| 984 | + } else { | |
| 985 | + tcg_out_modrm(s, 0xaf | P_EXT, args[0], args[2]); | |
| 986 | + } | |
| 987 | + break; | |
| 988 | + case INDEX_op_mulu2_i32: | |
| 989 | + tcg_out_modrm(s, 0xf7, 4, args[3]); | |
| 990 | + break; | |
| 991 | + case INDEX_op_div2_i32: | |
| 992 | + tcg_out_modrm(s, 0xf7, 7, args[4]); | |
| 993 | + break; | |
| 994 | + case INDEX_op_divu2_i32: | |
| 995 | + tcg_out_modrm(s, 0xf7, 6, args[4]); | |
| 996 | + break; | |
| 997 | + case INDEX_op_shl_i32: | |
| 998 | + c = SHIFT_SHL; | |
| 999 | + gen_shift32: | |
| 1000 | + if (const_args[2]) { | |
| 1001 | + if (args[2] == 1) { | |
| 1002 | + tcg_out_modrm(s, 0xd1, c, args[0]); | |
| 1003 | + } else { | |
| 1004 | + tcg_out_modrm(s, 0xc1, c, args[0]); | |
| 1005 | + tcg_out8(s, args[2]); | |
| 1006 | + } | |
| 1007 | + } else { | |
| 1008 | + tcg_out_modrm(s, 0xd3, c, args[0]); | |
| 1009 | + } | |
| 1010 | + break; | |
| 1011 | + case INDEX_op_shr_i32: | |
| 1012 | + c = SHIFT_SHR; | |
| 1013 | + goto gen_shift32; | |
| 1014 | + case INDEX_op_sar_i32: | |
| 1015 | + c = SHIFT_SAR; | |
| 1016 | + goto gen_shift32; | |
| 1017 | + | |
| 1018 | + case INDEX_op_add2_i32: | |
| 1019 | + if (const_args[4]) | |
| 1020 | + tgen_arithi(s, ARITH_ADD, args[0], args[4]); | |
| 1021 | + else | |
| 1022 | + tcg_out_modrm(s, 0x01 | (ARITH_ADD << 3), args[4], args[0]); | |
| 1023 | + if (const_args[5]) | |
| 1024 | + tgen_arithi(s, ARITH_ADC, args[1], args[5]); | |
| 1025 | + else | |
| 1026 | + tcg_out_modrm(s, 0x01 | (ARITH_ADC << 3), args[5], args[1]); | |
| 1027 | + break; | |
| 1028 | + case INDEX_op_sub2_i32: | |
| 1029 | + if (const_args[4]) | |
| 1030 | + tgen_arithi(s, ARITH_SUB, args[0], args[4]); | |
| 1031 | + else | |
| 1032 | + tcg_out_modrm(s, 0x01 | (ARITH_SUB << 3), args[4], args[0]); | |
| 1033 | + if (const_args[5]) | |
| 1034 | + tgen_arithi(s, ARITH_SBB, args[1], args[5]); | |
| 1035 | + else | |
| 1036 | + tcg_out_modrm(s, 0x01 | (ARITH_SBB << 3), args[5], args[1]); | |
| 1037 | + break; | |
| 1038 | + case INDEX_op_brcond_i32: | |
| 1039 | + tcg_out_brcond(s, args[2], args[0], args[1], const_args[1], args[3]); | |
| 1040 | + break; | |
| 1041 | + case INDEX_op_brcond2_i32: | |
| 1042 | + tcg_out_brcond2(s, args, const_args); | |
| 1043 | + break; | |
| 1044 | + | |
| 1045 | + case INDEX_op_qemu_ld8u: | |
| 1046 | + tcg_out_qemu_ld(s, args, 0); | |
| 1047 | + break; | |
| 1048 | + case INDEX_op_qemu_ld8s: | |
| 1049 | + tcg_out_qemu_ld(s, args, 0 | 4); | |
| 1050 | + break; | |
| 1051 | + case INDEX_op_qemu_ld16u: | |
| 1052 | + tcg_out_qemu_ld(s, args, 1); | |
| 1053 | + break; | |
| 1054 | + case INDEX_op_qemu_ld16s: | |
| 1055 | + tcg_out_qemu_ld(s, args, 1 | 4); | |
| 1056 | + break; | |
| 1057 | + case INDEX_op_qemu_ld32u: | |
| 1058 | + tcg_out_qemu_ld(s, args, 2); | |
| 1059 | + break; | |
| 1060 | + case INDEX_op_qemu_ld64: | |
| 1061 | + tcg_out_qemu_ld(s, args, 3); | |
| 1062 | + break; | |
| 1063 | + | |
| 1064 | + case INDEX_op_qemu_st8: | |
| 1065 | + tcg_out_qemu_st(s, args, 0); | |
| 1066 | + break; | |
| 1067 | + case INDEX_op_qemu_st16: | |
| 1068 | + tcg_out_qemu_st(s, args, 1); | |
| 1069 | + break; | |
| 1070 | + case INDEX_op_qemu_st32: | |
| 1071 | + tcg_out_qemu_st(s, args, 2); | |
| 1072 | + break; | |
| 1073 | + case INDEX_op_qemu_st64: | |
| 1074 | + tcg_out_qemu_st(s, args, 3); | |
| 1075 | + break; | |
| 1076 | + | |
| 1077 | + default: | |
| 1078 | + tcg_abort(); | |
| 1079 | + } | |
| 1080 | +} | |
| 1081 | + | |
| 1082 | +static const TCGTargetOpDef x86_op_defs[] = { | |
| 1083 | + { INDEX_op_exit_tb, { } }, | |
| 1084 | + { INDEX_op_goto_tb, { } }, | |
| 1085 | + { INDEX_op_call, { "ri" } }, | |
| 1086 | + { INDEX_op_jmp, { "ri" } }, | |
| 1087 | + { INDEX_op_br, { } }, | |
| 1088 | + { INDEX_op_mov_i32, { "r", "r" } }, | |
| 1089 | + { INDEX_op_movi_i32, { "r" } }, | |
| 1090 | + { INDEX_op_ld8u_i32, { "r", "r" } }, | |
| 1091 | + { INDEX_op_ld8s_i32, { "r", "r" } }, | |
| 1092 | + { INDEX_op_ld16u_i32, { "r", "r" } }, | |
| 1093 | + { INDEX_op_ld16s_i32, { "r", "r" } }, | |
| 1094 | + { INDEX_op_ld_i32, { "r", "r" } }, | |
| 1095 | + { INDEX_op_st8_i32, { "q", "r" } }, | |
| 1096 | + { INDEX_op_st16_i32, { "r", "r" } }, | |
| 1097 | + { INDEX_op_st_i32, { "r", "r" } }, | |
| 1098 | + | |
| 1099 | + { INDEX_op_add_i32, { "r", "0", "ri" } }, | |
| 1100 | + { INDEX_op_sub_i32, { "r", "0", "ri" } }, | |
| 1101 | + { INDEX_op_mul_i32, { "r", "0", "ri" } }, | |
| 1102 | + { INDEX_op_mulu2_i32, { "a", "d", "a", "r" } }, | |
| 1103 | + { INDEX_op_div2_i32, { "a", "d", "0", "1", "r" } }, | |
| 1104 | + { INDEX_op_divu2_i32, { "a", "d", "0", "1", "r" } }, | |
| 1105 | + { INDEX_op_and_i32, { "r", "0", "ri" } }, | |
| 1106 | + { INDEX_op_or_i32, { "r", "0", "ri" } }, | |
| 1107 | + { INDEX_op_xor_i32, { "r", "0", "ri" } }, | |
| 1108 | + | |
| 1109 | + { INDEX_op_shl_i32, { "r", "0", "ci" } }, | |
| 1110 | + { INDEX_op_shr_i32, { "r", "0", "ci" } }, | |
| 1111 | + { INDEX_op_sar_i32, { "r", "0", "ci" } }, | |
| 1112 | + | |
| 1113 | + { INDEX_op_brcond_i32, { "r", "ri" } }, | |
| 1114 | + | |
| 1115 | + { INDEX_op_add2_i32, { "r", "r", "0", "1", "ri", "ri" } }, | |
| 1116 | + { INDEX_op_sub2_i32, { "r", "r", "0", "1", "ri", "ri" } }, | |
| 1117 | + { INDEX_op_brcond2_i32, { "r", "r", "ri", "ri" } }, | |
| 1118 | + | |
| 1119 | +#if TARGET_LONG_BITS == 32 | |
| 1120 | + { INDEX_op_qemu_ld8u, { "r", "L" } }, | |
| 1121 | + { INDEX_op_qemu_ld8s, { "r", "L" } }, | |
| 1122 | + { INDEX_op_qemu_ld16u, { "r", "L" } }, | |
| 1123 | + { INDEX_op_qemu_ld16s, { "r", "L" } }, | |
| 1124 | + { INDEX_op_qemu_ld32u, { "r", "L" } }, | |
| 1125 | + { INDEX_op_qemu_ld64, { "r", "r", "L" } }, | |
| 1126 | + | |
| 1127 | + { INDEX_op_qemu_st8, { "cb", "L" } }, | |
| 1128 | + { INDEX_op_qemu_st16, { "L", "L" } }, | |
| 1129 | + { INDEX_op_qemu_st32, { "L", "L" } }, | |
| 1130 | + { INDEX_op_qemu_st64, { "L", "L", "L" } }, | |
| 1131 | +#else | |
| 1132 | + { INDEX_op_qemu_ld8u, { "r", "L", "L" } }, | |
| 1133 | + { INDEX_op_qemu_ld8s, { "r", "L", "L" } }, | |
| 1134 | + { INDEX_op_qemu_ld16u, { "r", "L", "L" } }, | |
| 1135 | + { INDEX_op_qemu_ld16s, { "r", "L", "L" } }, | |
| 1136 | + { INDEX_op_qemu_ld32u, { "r", "L", "L" } }, | |
| 1137 | + { INDEX_op_qemu_ld64, { "r", "r", "L", "L" } }, | |
| 1138 | + | |
| 1139 | + { INDEX_op_qemu_st8, { "cb", "L", "L" } }, | |
| 1140 | + { INDEX_op_qemu_st16, { "L", "L", "L" } }, | |
| 1141 | + { INDEX_op_qemu_st32, { "L", "L", "L" } }, | |
| 1142 | + { INDEX_op_qemu_st64, { "L", "L", "L", "L" } }, | |
| 1143 | +#endif | |
| 1144 | + { -1 }, | |
| 1145 | +}; | |
| 1146 | + | |
| 1147 | +void tcg_target_init(TCGContext *s) | |
| 1148 | +{ | |
| 1149 | + /* fail safe */ | |
| 1150 | + if ((1 << CPU_TLB_ENTRY_BITS) != sizeof(CPUTLBEntry)) | |
| 1151 | + tcg_abort(); | |
| 1152 | + | |
| 1153 | + tcg_regset_set32(tcg_target_available_regs[TCG_TYPE_I32], 0, 0xff); | |
| 1154 | + tcg_regset_set32(tcg_target_call_clobber_regs, 0, | |
| 1155 | + (1 << TCG_REG_EAX) | | |
| 1156 | + (1 << TCG_REG_EDX) | | |
| 1157 | + (1 << TCG_REG_ECX)); | |
| 1158 | + | |
| 1159 | + tcg_regset_clear(s->reserved_regs); | |
| 1160 | + tcg_regset_set_reg(s->reserved_regs, TCG_REG_ESP); | |
| 1161 | + | |
| 1162 | + tcg_add_target_add_op_defs(x86_op_defs); | |
| 1163 | +} | ... | ... |
tcg/i386/tcg-target.h
0 โ 100644
| 1 | +/* | |
| 2 | + * Tiny Code Generator for QEMU | |
| 3 | + * | |
| 4 | + * Copyright (c) 2008 Fabrice Bellard | |
| 5 | + * | |
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | |
| 7 | + * of this software and associated documentation files (the "Software"), to deal | |
| 8 | + * in the Software without restriction, including without limitation the rights | |
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | |
| 10 | + * copies of the Software, and to permit persons to whom the Software is | |
| 11 | + * furnished to do so, subject to the following conditions: | |
| 12 | + * | |
| 13 | + * The above copyright notice and this permission notice shall be included in | |
| 14 | + * all copies or substantial portions of the Software. | |
| 15 | + * | |
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | |
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | |
| 22 | + * THE SOFTWARE. | |
| 23 | + */ | |
| 24 | +#define TCG_TARGET_I386 1 | |
| 25 | + | |
| 26 | +#define TCG_TARGET_REG_BITS 32 | |
| 27 | +//#define TCG_TARGET_WORDS_BIGENDIAN | |
| 28 | + | |
| 29 | +#define TCG_TARGET_NB_REGS 8 | |
| 30 | + | |
| 31 | +enum { | |
| 32 | + TCG_REG_EAX = 0, | |
| 33 | + TCG_REG_ECX, | |
| 34 | + TCG_REG_EDX, | |
| 35 | + TCG_REG_EBX, | |
| 36 | + TCG_REG_ESP, | |
| 37 | + TCG_REG_EBP, | |
| 38 | + TCG_REG_ESI, | |
| 39 | + TCG_REG_EDI, | |
| 40 | +}; | |
| 41 | + | |
| 42 | +/* used for function call generation */ | |
| 43 | +#define TCG_REG_CALL_STACK TCG_REG_ESP | |
| 44 | +#define TCG_TARGET_STACK_ALIGN 16 | |
| 45 | + | |
| 46 | +/* Note: must be synced with dyngen-exec.h */ | |
| 47 | +#define TCG_AREG0 TCG_REG_EBP | |
| 48 | +#define TCG_AREG1 TCG_REG_EBX | |
| 49 | +#define TCG_AREG2 TCG_REG_ESI | |
| 50 | +#define TCG_AREG3 TCG_REG_EDI | |
| 51 | + | |
| 52 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | |
| 53 | +{ | |
| 54 | +} | ... | ... |
tcg/tcg-dyngen.c
0 โ 100644
| 1 | +/* | |
| 2 | + * Tiny Code Generator for QEMU | |
| 3 | + * | |
| 4 | + * Copyright (c) 2008 Fabrice Bellard | |
| 5 | + * | |
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | |
| 7 | + * of this software and associated documentation files (the "Software"), to deal | |
| 8 | + * in the Software without restriction, including without limitation the rights | |
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | |
| 10 | + * copies of the Software, and to permit persons to whom the Software is | |
| 11 | + * furnished to do so, subject to the following conditions: | |
| 12 | + * | |
| 13 | + * The above copyright notice and this permission notice shall be included in | |
| 14 | + * all copies or substantial portions of the Software. | |
| 15 | + * | |
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | |
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | |
| 22 | + * THE SOFTWARE. | |
| 23 | + */ | |
| 24 | +#include <assert.h> | |
| 25 | +#include <stdarg.h> | |
| 26 | +#include <stdlib.h> | |
| 27 | +#include <stdio.h> | |
| 28 | +#include <string.h> | |
| 29 | +#include <inttypes.h> | |
| 30 | + | |
| 31 | +#include "config.h" | |
| 32 | +#include "osdep.h" | |
| 33 | + | |
| 34 | +#include "tcg.h" | |
| 35 | + | |
| 36 | +int __op_param1, __op_param2, __op_param3; | |
| 37 | +#if defined(__sparc__) || defined(__arm__) | |
| 38 | + void __op_gen_label1(){} | |
| 39 | + void __op_gen_label2(){} | |
| 40 | + void __op_gen_label3(){} | |
| 41 | +#else | |
| 42 | + int __op_gen_label1, __op_gen_label2, __op_gen_label3; | |
| 43 | +#endif | |
| 44 | +int __op_jmp0, __op_jmp1, __op_jmp2, __op_jmp3; | |
| 45 | + | |
| 46 | +#if 0 | |
| 47 | +#if defined(__s390__) | |
| 48 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | |
| 49 | +{ | |
| 50 | +} | |
| 51 | +#elif defined(__ia64__) | |
| 52 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | |
| 53 | +{ | |
| 54 | + while (start < stop) { | |
| 55 | + asm volatile ("fc %0" :: "r"(start)); | |
| 56 | + start += 32; | |
| 57 | + } | |
| 58 | + asm volatile (";;sync.i;;srlz.i;;"); | |
| 59 | +} | |
| 60 | +#elif defined(__powerpc__) | |
| 61 | + | |
| 62 | +#define MIN_CACHE_LINE_SIZE 8 /* conservative value */ | |
| 63 | + | |
| 64 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | |
| 65 | +{ | |
| 66 | + unsigned long p; | |
| 67 | + | |
| 68 | + start &= ~(MIN_CACHE_LINE_SIZE - 1); | |
| 69 | + stop = (stop + MIN_CACHE_LINE_SIZE - 1) & ~(MIN_CACHE_LINE_SIZE - 1); | |
| 70 | + | |
| 71 | + for (p = start; p < stop; p += MIN_CACHE_LINE_SIZE) { | |
| 72 | + asm volatile ("dcbst 0,%0" : : "r"(p) : "memory"); | |
| 73 | + } | |
| 74 | + asm volatile ("sync" : : : "memory"); | |
| 75 | + for (p = start; p < stop; p += MIN_CACHE_LINE_SIZE) { | |
| 76 | + asm volatile ("icbi 0,%0" : : "r"(p) : "memory"); | |
| 77 | + } | |
| 78 | + asm volatile ("sync" : : : "memory"); | |
| 79 | + asm volatile ("isync" : : : "memory"); | |
| 80 | +} | |
| 81 | +#elif defined(__alpha__) | |
| 82 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | |
| 83 | +{ | |
| 84 | + asm ("imb"); | |
| 85 | +} | |
| 86 | +#elif defined(__sparc__) | |
| 87 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | |
| 88 | +{ | |
| 89 | + unsigned long p; | |
| 90 | + | |
| 91 | + p = start & ~(8UL - 1UL); | |
| 92 | + stop = (stop + (8UL - 1UL)) & ~(8UL - 1UL); | |
| 93 | + | |
| 94 | + for (; p < stop; p += 8) | |
| 95 | + __asm__ __volatile__("flush\t%0" : : "r" (p)); | |
| 96 | +} | |
| 97 | +#elif defined(__arm__) | |
| 98 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | |
| 99 | +{ | |
| 100 | + register unsigned long _beg __asm ("a1") = start; | |
| 101 | + register unsigned long _end __asm ("a2") = stop; | |
| 102 | + register unsigned long _flg __asm ("a3") = 0; | |
| 103 | + __asm __volatile__ ("swi 0x9f0002" : : "r" (_beg), "r" (_end), "r" (_flg)); | |
| 104 | +} | |
| 105 | +#elif defined(__mc68000) | |
| 106 | + | |
| 107 | +# include <asm/cachectl.h> | |
| 108 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | |
| 109 | +{ | |
| 110 | + cacheflush(start,FLUSH_SCOPE_LINE,FLUSH_CACHE_BOTH,stop-start+16); | |
| 111 | +} | |
| 112 | +#elif defined(__mips__) | |
| 113 | + | |
| 114 | +#include <sys/cachectl.h> | |
| 115 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | |
| 116 | +{ | |
| 117 | + _flush_cache ((void *)start, stop - start, BCACHE); | |
| 118 | +} | |
| 119 | +#else | |
| 120 | +#error unsupported CPU | |
| 121 | +#endif | |
| 122 | + | |
| 123 | +#ifdef __alpha__ | |
| 124 | + | |
| 125 | +register int gp asm("$29"); | |
| 126 | + | |
| 127 | +static inline void immediate_ldah(void *p, int val) { | |
| 128 | + uint32_t *dest = p; | |
| 129 | + long high = ((val >> 16) + ((val >> 15) & 1)) & 0xffff; | |
| 130 | + | |
| 131 | + *dest &= ~0xffff; | |
| 132 | + *dest |= high; | |
| 133 | + *dest |= 31 << 16; | |
| 134 | +} | |
| 135 | +static inline void immediate_lda(void *dest, int val) { | |
| 136 | + *(uint16_t *) dest = val; | |
| 137 | +} | |
| 138 | +void fix_bsr(void *p, int offset) { | |
| 139 | + uint32_t *dest = p; | |
| 140 | + *dest &= ~((1 << 21) - 1); | |
| 141 | + *dest |= (offset >> 2) & ((1 << 21) - 1); | |
| 142 | +} | |
| 143 | + | |
| 144 | +#endif /* __alpha__ */ | |
| 145 | + | |
| 146 | +#ifdef __arm__ | |
| 147 | + | |
| 148 | +#define ARM_LDR_TABLE_SIZE 1024 | |
| 149 | + | |
| 150 | +typedef struct LDREntry { | |
| 151 | + uint8_t *ptr; | |
| 152 | + uint32_t *data_ptr; | |
| 153 | + unsigned type:2; | |
| 154 | +} LDREntry; | |
| 155 | + | |
| 156 | +static LDREntry arm_ldr_table[1024]; | |
| 157 | +static uint32_t arm_data_table[ARM_LDR_TABLE_SIZE]; | |
| 158 | + | |
| 159 | +extern char exec_loop; | |
| 160 | + | |
| 161 | +static inline void arm_reloc_pc24(uint32_t *ptr, uint32_t insn, int val) | |
| 162 | +{ | |
| 163 | + *ptr = (insn & ~0xffffff) | ((insn + ((val - (int)ptr) >> 2)) & 0xffffff); | |
| 164 | +} | |
| 165 | + | |
| 166 | +static uint8_t *arm_flush_ldr(uint8_t *gen_code_ptr, | |
| 167 | + LDREntry *ldr_start, LDREntry *ldr_end, | |
| 168 | + uint32_t *data_start, uint32_t *data_end, | |
| 169 | + int gen_jmp) | |
| 170 | +{ | |
| 171 | + LDREntry *le; | |
| 172 | + uint32_t *ptr; | |
| 173 | + int offset, data_size, target; | |
| 174 | + uint8_t *data_ptr; | |
| 175 | + uint32_t insn; | |
| 176 | + uint32_t mask; | |
| 177 | + | |
| 178 | + data_size = (data_end - data_start) << 2; | |
| 179 | + | |
| 180 | + if (gen_jmp) { | |
| 181 | + /* generate branch to skip the data */ | |
| 182 | + if (data_size == 0) | |
| 183 | + return gen_code_ptr; | |
| 184 | + target = (long)gen_code_ptr + data_size + 4; | |
| 185 | + arm_reloc_pc24((uint32_t *)gen_code_ptr, 0xeafffffe, target); | |
| 186 | + gen_code_ptr += 4; | |
| 187 | + } | |
| 188 | + | |
| 189 | + /* copy the data */ | |
| 190 | + data_ptr = gen_code_ptr; | |
| 191 | + memcpy(gen_code_ptr, data_start, data_size); | |
| 192 | + gen_code_ptr += data_size; | |
| 193 | + | |
| 194 | + /* patch the ldr to point to the data */ | |
| 195 | + for(le = ldr_start; le < ldr_end; le++) { | |
| 196 | + ptr = (uint32_t *)le->ptr; | |
| 197 | + offset = ((unsigned long)(le->data_ptr) - (unsigned long)data_start) + | |
| 198 | + (unsigned long)data_ptr - | |
| 199 | + (unsigned long)ptr - 8; | |
| 200 | + if (offset < 0) { | |
| 201 | + fprintf(stderr, "Negative constant pool offset\n"); | |
| 202 | + tcg_abort(); | |
| 203 | + } | |
| 204 | + switch (le->type) { | |
| 205 | + case 0: /* ldr */ | |
| 206 | + mask = ~0x00800fff; | |
| 207 | + if (offset >= 4096) { | |
| 208 | + fprintf(stderr, "Bad ldr offset\n"); | |
| 209 | + tcg_abort(); | |
| 210 | + } | |
| 211 | + break; | |
| 212 | + case 1: /* ldc */ | |
| 213 | + mask = ~0x008000ff; | |
| 214 | + if (offset >= 1024 ) { | |
| 215 | + fprintf(stderr, "Bad ldc offset\n"); | |
| 216 | + tcg_abort(); | |
| 217 | + } | |
| 218 | + break; | |
| 219 | + case 2: /* add */ | |
| 220 | + mask = ~0xfff; | |
| 221 | + if (offset >= 1024 ) { | |
| 222 | + fprintf(stderr, "Bad add offset\n"); | |
| 223 | + tcg_abort(); | |
| 224 | + } | |
| 225 | + break; | |
| 226 | + default: | |
| 227 | + fprintf(stderr, "Bad pc relative fixup\n"); | |
| 228 | + tcg_abort(); | |
| 229 | + } | |
| 230 | + insn = *ptr & mask; | |
| 231 | + switch (le->type) { | |
| 232 | + case 0: /* ldr */ | |
| 233 | + insn |= offset | 0x00800000; | |
| 234 | + break; | |
| 235 | + case 1: /* ldc */ | |
| 236 | + insn |= (offset >> 2) | 0x00800000; | |
| 237 | + break; | |
| 238 | + case 2: /* add */ | |
| 239 | + insn |= (offset >> 2) | 0xf00; | |
| 240 | + break; | |
| 241 | + } | |
| 242 | + *ptr = insn; | |
| 243 | + } | |
| 244 | + return gen_code_ptr; | |
| 245 | +} | |
| 246 | + | |
| 247 | +#endif /* __arm__ */ | |
| 248 | + | |
| 249 | +#ifdef __ia64 | |
| 250 | + | |
| 251 | +/* Patch instruction with "val" where "mask" has 1 bits. */ | |
| 252 | +static inline void ia64_patch (uint64_t insn_addr, uint64_t mask, uint64_t val) | |
| 253 | +{ | |
| 254 | + uint64_t m0, m1, v0, v1, b0, b1, *b = (uint64_t *) (insn_addr & -16); | |
| 255 | +# define insn_mask ((1UL << 41) - 1) | |
| 256 | + unsigned long shift; | |
| 257 | + | |
| 258 | + b0 = b[0]; b1 = b[1]; | |
| 259 | + shift = 5 + 41 * (insn_addr % 16); /* 5 template, 3 x 41-bit insns */ | |
| 260 | + if (shift >= 64) { | |
| 261 | + m1 = mask << (shift - 64); | |
| 262 | + v1 = val << (shift - 64); | |
| 263 | + } else { | |
| 264 | + m0 = mask << shift; m1 = mask >> (64 - shift); | |
| 265 | + v0 = val << shift; v1 = val >> (64 - shift); | |
| 266 | + b[0] = (b0 & ~m0) | (v0 & m0); | |
| 267 | + } | |
| 268 | + b[1] = (b1 & ~m1) | (v1 & m1); | |
| 269 | +} | |
| 270 | + | |
| 271 | +static inline void ia64_patch_imm60 (uint64_t insn_addr, uint64_t val) | |
| 272 | +{ | |
| 273 | + ia64_patch(insn_addr, | |
| 274 | + 0x011ffffe000UL, | |
| 275 | + ( ((val & 0x0800000000000000UL) >> 23) /* bit 59 -> 36 */ | |
| 276 | + | ((val & 0x00000000000fffffUL) << 13) /* bit 0 -> 13 */)); | |
| 277 | + ia64_patch(insn_addr - 1, 0x1fffffffffcUL, val >> 18); | |
| 278 | +} | |
| 279 | + | |
| 280 | +static inline void ia64_imm64 (void *insn, uint64_t val) | |
| 281 | +{ | |
| 282 | + /* Ignore the slot number of the relocation; GCC and Intel | |
| 283 | + toolchains differed for some time on whether IMM64 relocs are | |
| 284 | + against slot 1 (Intel) or slot 2 (GCC). */ | |
| 285 | + uint64_t insn_addr = (uint64_t) insn & ~3UL; | |
| 286 | + | |
| 287 | + ia64_patch(insn_addr + 2, | |
| 288 | + 0x01fffefe000UL, | |
| 289 | + ( ((val & 0x8000000000000000UL) >> 27) /* bit 63 -> 36 */ | |
| 290 | + | ((val & 0x0000000000200000UL) << 0) /* bit 21 -> 21 */ | |
| 291 | + | ((val & 0x00000000001f0000UL) << 6) /* bit 16 -> 22 */ | |
| 292 | + | ((val & 0x000000000000ff80UL) << 20) /* bit 7 -> 27 */ | |
| 293 | + | ((val & 0x000000000000007fUL) << 13) /* bit 0 -> 13 */) | |
| 294 | + ); | |
| 295 | + ia64_patch(insn_addr + 1, 0x1ffffffffffUL, val >> 22); | |
| 296 | +} | |
| 297 | + | |
| 298 | +static inline void ia64_imm60b (void *insn, uint64_t val) | |
| 299 | +{ | |
| 300 | + /* Ignore the slot number of the relocation; GCC and Intel | |
| 301 | + toolchains differed for some time on whether IMM64 relocs are | |
| 302 | + against slot 1 (Intel) or slot 2 (GCC). */ | |
| 303 | + uint64_t insn_addr = (uint64_t) insn & ~3UL; | |
| 304 | + | |
| 305 | + if (val + ((uint64_t) 1 << 59) >= (1UL << 60)) | |
| 306 | + fprintf(stderr, "%s: value %ld out of IMM60 range\n", | |
| 307 | + __FUNCTION__, (int64_t) val); | |
| 308 | + ia64_patch_imm60(insn_addr + 2, val); | |
| 309 | +} | |
| 310 | + | |
| 311 | +static inline void ia64_imm22 (void *insn, uint64_t val) | |
| 312 | +{ | |
| 313 | + if (val + (1 << 21) >= (1 << 22)) | |
| 314 | + fprintf(stderr, "%s: value %li out of IMM22 range\n", | |
| 315 | + __FUNCTION__, (int64_t)val); | |
| 316 | + ia64_patch((uint64_t) insn, 0x01fffcfe000UL, | |
| 317 | + ( ((val & 0x200000UL) << 15) /* bit 21 -> 36 */ | |
| 318 | + | ((val & 0x1f0000UL) << 6) /* bit 16 -> 22 */ | |
| 319 | + | ((val & 0x00ff80UL) << 20) /* bit 7 -> 27 */ | |
| 320 | + | ((val & 0x00007fUL) << 13) /* bit 0 -> 13 */)); | |
| 321 | +} | |
| 322 | + | |
| 323 | +/* Like ia64_imm22(), but also clear bits 20-21. For addl, this has | |
| 324 | + the effect of turning "addl rX=imm22,rY" into "addl | |
| 325 | + rX=imm22,r0". */ | |
| 326 | +static inline void ia64_imm22_r0 (void *insn, uint64_t val) | |
| 327 | +{ | |
| 328 | + if (val + (1 << 21) >= (1 << 22)) | |
| 329 | + fprintf(stderr, "%s: value %li out of IMM22 range\n", | |
| 330 | + __FUNCTION__, (int64_t)val); | |
| 331 | + ia64_patch((uint64_t) insn, 0x01fffcfe000UL | (0x3UL << 20), | |
| 332 | + ( ((val & 0x200000UL) << 15) /* bit 21 -> 36 */ | |
| 333 | + | ((val & 0x1f0000UL) << 6) /* bit 16 -> 22 */ | |
| 334 | + | ((val & 0x00ff80UL) << 20) /* bit 7 -> 27 */ | |
| 335 | + | ((val & 0x00007fUL) << 13) /* bit 0 -> 13 */)); | |
| 336 | +} | |
| 337 | + | |
| 338 | +static inline void ia64_imm21b (void *insn, uint64_t val) | |
| 339 | +{ | |
| 340 | + if (val + (1 << 20) >= (1 << 21)) | |
| 341 | + fprintf(stderr, "%s: value %li out of IMM21b range\n", | |
| 342 | + __FUNCTION__, (int64_t)val); | |
| 343 | + ia64_patch((uint64_t) insn, 0x11ffffe000UL, | |
| 344 | + ( ((val & 0x100000UL) << 16) /* bit 20 -> 36 */ | |
| 345 | + | ((val & 0x0fffffUL) << 13) /* bit 0 -> 13 */)); | |
| 346 | +} | |
| 347 | + | |
| 348 | +static inline void ia64_nop_b (void *insn) | |
| 349 | +{ | |
| 350 | + ia64_patch((uint64_t) insn, (1UL << 41) - 1, 2UL << 37); | |
| 351 | +} | |
| 352 | + | |
| 353 | +static inline void ia64_ldxmov(void *insn, uint64_t val) | |
| 354 | +{ | |
| 355 | + if (val + (1 << 21) < (1 << 22)) | |
| 356 | + ia64_patch((uint64_t) insn, 0x1fff80fe000UL, 8UL << 37); | |
| 357 | +} | |
| 358 | + | |
| 359 | +static inline int ia64_patch_ltoff(void *insn, uint64_t val, | |
| 360 | + int relaxable) | |
| 361 | +{ | |
| 362 | + if (relaxable && (val + (1 << 21) < (1 << 22))) { | |
| 363 | + ia64_imm22_r0(insn, val); | |
| 364 | + return 0; | |
| 365 | + } | |
| 366 | + return 1; | |
| 367 | +} | |
| 368 | + | |
| 369 | +struct ia64_fixup { | |
| 370 | + struct ia64_fixup *next; | |
| 371 | + void *addr; /* address that needs to be patched */ | |
| 372 | + long value; | |
| 373 | +}; | |
| 374 | + | |
| 375 | +#define IA64_PLT(insn, plt_index) \ | |
| 376 | +do { \ | |
| 377 | + struct ia64_fixup *fixup = alloca(sizeof(*fixup)); \ | |
| 378 | + fixup->next = plt_fixes; \ | |
| 379 | + plt_fixes = fixup; \ | |
| 380 | + fixup->addr = (insn); \ | |
| 381 | + fixup->value = (plt_index); \ | |
| 382 | + plt_offset[(plt_index)] = 1; \ | |
| 383 | +} while (0) | |
| 384 | + | |
| 385 | +#define IA64_LTOFF(insn, val, relaxable) \ | |
| 386 | +do { \ | |
| 387 | + if (ia64_patch_ltoff(insn, val, relaxable)) { \ | |
| 388 | + struct ia64_fixup *fixup = alloca(sizeof(*fixup)); \ | |
| 389 | + fixup->next = ltoff_fixes; \ | |
| 390 | + ltoff_fixes = fixup; \ | |
| 391 | + fixup->addr = (insn); \ | |
| 392 | + fixup->value = (val); \ | |
| 393 | + } \ | |
| 394 | +} while (0) | |
| 395 | + | |
| 396 | +static inline void ia64_apply_fixes (uint8_t **gen_code_pp, | |
| 397 | + struct ia64_fixup *ltoff_fixes, | |
| 398 | + uint64_t gp, | |
| 399 | + struct ia64_fixup *plt_fixes, | |
| 400 | + int num_plts, | |
| 401 | + unsigned long *plt_target, | |
| 402 | + unsigned int *plt_offset) | |
| 403 | +{ | |
| 404 | + static const uint8_t plt_bundle[] = { | |
| 405 | + 0x04, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, /* nop 0; movl r1=GP */ | |
| 406 | + 0x00, 0x00, 0x00, 0x20, 0x00, 0x00, 0x00, 0x60, | |
| 407 | + | |
| 408 | + 0x05, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, /* nop 0; brl IP */ | |
| 409 | + 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0xc0 | |
| 410 | + }; | |
| 411 | + uint8_t *gen_code_ptr = *gen_code_pp, *plt_start, *got_start; | |
| 412 | + uint64_t *vp; | |
| 413 | + struct ia64_fixup *fixup; | |
| 414 | + unsigned int offset = 0; | |
| 415 | + struct fdesc { | |
| 416 | + long ip; | |
| 417 | + long gp; | |
| 418 | + } *fdesc; | |
| 419 | + int i; | |
| 420 | + | |
| 421 | + if (plt_fixes) { | |
| 422 | + plt_start = gen_code_ptr; | |
| 423 | + | |
| 424 | + for (i = 0; i < num_plts; ++i) { | |
| 425 | + if (plt_offset[i]) { | |
| 426 | + plt_offset[i] = offset; | |
| 427 | + offset += sizeof(plt_bundle); | |
| 428 | + | |
| 429 | + fdesc = (struct fdesc *) plt_target[i]; | |
| 430 | + memcpy(gen_code_ptr, plt_bundle, sizeof(plt_bundle)); | |
| 431 | + ia64_imm64 (gen_code_ptr + 0x02, fdesc->gp); | |
| 432 | + ia64_imm60b(gen_code_ptr + 0x12, | |
| 433 | + (fdesc->ip - (long) (gen_code_ptr + 0x10)) >> 4); | |
| 434 | + gen_code_ptr += sizeof(plt_bundle); | |
| 435 | + } | |
| 436 | + } | |
| 437 | + | |
| 438 | + for (fixup = plt_fixes; fixup; fixup = fixup->next) | |
| 439 | + ia64_imm21b(fixup->addr, | |
| 440 | + ((long) plt_start + plt_offset[fixup->value] | |
| 441 | + - ((long) fixup->addr & ~0xf)) >> 4); | |
| 442 | + } | |
| 443 | + | |
| 444 | + got_start = gen_code_ptr; | |
| 445 | + | |
| 446 | + /* First, create the GOT: */ | |
| 447 | + for (fixup = ltoff_fixes; fixup; fixup = fixup->next) { | |
| 448 | + /* first check if we already have this value in the GOT: */ | |
| 449 | + for (vp = (uint64_t *) got_start; vp < (uint64_t *) gen_code_ptr; ++vp) | |
| 450 | + if (*vp == fixup->value) | |
| 451 | + break; | |
| 452 | + if (vp == (uint64_t *) gen_code_ptr) { | |
| 453 | + /* Nope, we need to put the value in the GOT: */ | |
| 454 | + *vp = fixup->value; | |
| 455 | + gen_code_ptr += 8; | |
| 456 | + } | |
| 457 | + ia64_imm22(fixup->addr, (long) vp - gp); | |
| 458 | + } | |
| 459 | + /* Keep code ptr aligned. */ | |
| 460 | + if ((long) gen_code_ptr & 15) | |
| 461 | + gen_code_ptr += 8; | |
| 462 | + *gen_code_pp = gen_code_ptr; | |
| 463 | +} | |
| 464 | +#endif | |
| 465 | +#endif | |
| 466 | + | |
| 467 | +const TCGArg *dyngen_op(TCGContext *s, int opc, const TCGArg *opparam_ptr) | |
| 468 | +{ | |
| 469 | + uint8_t *gen_code_ptr; | |
| 470 | + | |
| 471 | + gen_code_ptr = s->code_ptr; | |
| 472 | + switch(opc) { | |
| 473 | + | |
| 474 | +/* op.h is dynamically generated by dyngen.c from op.c */ | |
| 475 | +#include "op.h" | |
| 476 | + | |
| 477 | + default: | |
| 478 | + tcg_abort(); | |
| 479 | + } | |
| 480 | + s->code_ptr = gen_code_ptr; | |
| 481 | + return opparam_ptr; | |
| 482 | +} | ... | ... |
tcg/tcg-op.h
0 โ 100644
| 1 | +/* | |
| 2 | + * Tiny Code Generator for QEMU | |
| 3 | + * | |
| 4 | + * Copyright (c) 2008 Fabrice Bellard | |
| 5 | + * | |
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | |
| 7 | + * of this software and associated documentation files (the "Software"), to deal | |
| 8 | + * in the Software without restriction, including without limitation the rights | |
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | |
| 10 | + * copies of the Software, and to permit persons to whom the Software is | |
| 11 | + * furnished to do so, subject to the following conditions: | |
| 12 | + * | |
| 13 | + * The above copyright notice and this permission notice shall be included in | |
| 14 | + * all copies or substantial portions of the Software. | |
| 15 | + * | |
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | |
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | |
| 22 | + * THE SOFTWARE. | |
| 23 | + */ | |
| 24 | +#include "tcg.h" | |
| 25 | + | |
| 26 | +/* legacy dyngen operations */ | |
| 27 | +#include "gen-op.h" | |
| 28 | + | |
| 29 | +int gen_new_label(void); | |
| 30 | + | |
| 31 | +static inline void tcg_gen_op1(int opc, TCGArg arg1) | |
| 32 | +{ | |
| 33 | + *gen_opc_ptr++ = opc; | |
| 34 | + *gen_opparam_ptr++ = arg1; | |
| 35 | +} | |
| 36 | + | |
| 37 | +static inline void tcg_gen_op2(int opc, TCGArg arg1, TCGArg arg2) | |
| 38 | +{ | |
| 39 | + *gen_opc_ptr++ = opc; | |
| 40 | + *gen_opparam_ptr++ = arg1; | |
| 41 | + *gen_opparam_ptr++ = arg2; | |
| 42 | +} | |
| 43 | + | |
| 44 | +static inline void tcg_gen_op3(int opc, TCGArg arg1, TCGArg arg2, TCGArg arg3) | |
| 45 | +{ | |
| 46 | + *gen_opc_ptr++ = opc; | |
| 47 | + *gen_opparam_ptr++ = arg1; | |
| 48 | + *gen_opparam_ptr++ = arg2; | |
| 49 | + *gen_opparam_ptr++ = arg3; | |
| 50 | +} | |
| 51 | + | |
| 52 | +static inline void tcg_gen_op4(int opc, TCGArg arg1, TCGArg arg2, TCGArg arg3, | |
| 53 | + TCGArg arg4) | |
| 54 | +{ | |
| 55 | + *gen_opc_ptr++ = opc; | |
| 56 | + *gen_opparam_ptr++ = arg1; | |
| 57 | + *gen_opparam_ptr++ = arg2; | |
| 58 | + *gen_opparam_ptr++ = arg3; | |
| 59 | + *gen_opparam_ptr++ = arg4; | |
| 60 | +} | |
| 61 | + | |
| 62 | +static inline void tcg_gen_op5(int opc, TCGArg arg1, TCGArg arg2, | |
| 63 | + TCGArg arg3, TCGArg arg4, | |
| 64 | + TCGArg arg5) | |
| 65 | +{ | |
| 66 | + *gen_opc_ptr++ = opc; | |
| 67 | + *gen_opparam_ptr++ = arg1; | |
| 68 | + *gen_opparam_ptr++ = arg2; | |
| 69 | + *gen_opparam_ptr++ = arg3; | |
| 70 | + *gen_opparam_ptr++ = arg4; | |
| 71 | + *gen_opparam_ptr++ = arg5; | |
| 72 | +} | |
| 73 | + | |
| 74 | +static inline void tcg_gen_op6(int opc, TCGArg arg1, TCGArg arg2, | |
| 75 | + TCGArg arg3, TCGArg arg4, | |
| 76 | + TCGArg arg5, TCGArg arg6) | |
| 77 | +{ | |
| 78 | + *gen_opc_ptr++ = opc; | |
| 79 | + *gen_opparam_ptr++ = arg1; | |
| 80 | + *gen_opparam_ptr++ = arg2; | |
| 81 | + *gen_opparam_ptr++ = arg3; | |
| 82 | + *gen_opparam_ptr++ = arg4; | |
| 83 | + *gen_opparam_ptr++ = arg5; | |
| 84 | + *gen_opparam_ptr++ = arg6; | |
| 85 | +} | |
| 86 | + | |
| 87 | +static inline void gen_set_label(int n) | |
| 88 | +{ | |
| 89 | + tcg_gen_op1(INDEX_op_set_label, n); | |
| 90 | +} | |
| 91 | + | |
| 92 | +static inline void tcg_gen_mov_i32(int ret, int arg) | |
| 93 | +{ | |
| 94 | + tcg_gen_op2(INDEX_op_mov_i32, ret, arg); | |
| 95 | +} | |
| 96 | + | |
| 97 | +static inline void tcg_gen_movi_i32(int ret, int32_t arg) | |
| 98 | +{ | |
| 99 | + tcg_gen_op2(INDEX_op_movi_i32, ret, arg); | |
| 100 | +} | |
| 101 | + | |
| 102 | +/* helper calls */ | |
| 103 | +#define TCG_HELPER_CALL_FLAGS 0 | |
| 104 | + | |
| 105 | +static inline void tcg_gen_helper_0_0(void *func) | |
| 106 | +{ | |
| 107 | + tcg_gen_call(&tcg_ctx, | |
| 108 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | |
| 109 | + 0, NULL, 0, NULL); | |
| 110 | +} | |
| 111 | + | |
| 112 | +static inline void tcg_gen_helper_0_1(void *func, TCGArg arg) | |
| 113 | +{ | |
| 114 | + tcg_gen_call(&tcg_ctx, | |
| 115 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | |
| 116 | + 0, NULL, 1, &arg); | |
| 117 | +} | |
| 118 | + | |
| 119 | +static inline void tcg_gen_helper_0_2(void *func, TCGArg arg1, TCGArg arg2) | |
| 120 | +{ | |
| 121 | + TCGArg args[2]; | |
| 122 | + args[0] = arg1; | |
| 123 | + args[1] = arg2; | |
| 124 | + tcg_gen_call(&tcg_ctx, | |
| 125 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | |
| 126 | + 0, NULL, 2, args); | |
| 127 | +} | |
| 128 | + | |
| 129 | +static inline void tcg_gen_helper_1_2(void *func, TCGArg ret, | |
| 130 | + TCGArg arg1, TCGArg arg2) | |
| 131 | +{ | |
| 132 | + TCGArg args[2]; | |
| 133 | + args[0] = arg1; | |
| 134 | + args[1] = arg2; | |
| 135 | + tcg_gen_call(&tcg_ctx, | |
| 136 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | |
| 137 | + 1, &ret, 2, args); | |
| 138 | +} | |
| 139 | + | |
| 140 | +/* 32 bit ops */ | |
| 141 | + | |
| 142 | +static inline void tcg_gen_ld8u_i32(int ret, int arg2, tcg_target_long offset) | |
| 143 | +{ | |
| 144 | + tcg_gen_op3(INDEX_op_ld8u_i32, ret, arg2, offset); | |
| 145 | +} | |
| 146 | + | |
| 147 | +static inline void tcg_gen_ld8s_i32(int ret, int arg2, tcg_target_long offset) | |
| 148 | +{ | |
| 149 | + tcg_gen_op3(INDEX_op_ld8s_i32, ret, arg2, offset); | |
| 150 | +} | |
| 151 | + | |
| 152 | +static inline void tcg_gen_ld16u_i32(int ret, int arg2, tcg_target_long offset) | |
| 153 | +{ | |
| 154 | + tcg_gen_op3(INDEX_op_ld16u_i32, ret, arg2, offset); | |
| 155 | +} | |
| 156 | + | |
| 157 | +static inline void tcg_gen_ld16s_i32(int ret, int arg2, tcg_target_long offset) | |
| 158 | +{ | |
| 159 | + tcg_gen_op3(INDEX_op_ld16s_i32, ret, arg2, offset); | |
| 160 | +} | |
| 161 | + | |
| 162 | +static inline void tcg_gen_ld_i32(int ret, int arg2, tcg_target_long offset) | |
| 163 | +{ | |
| 164 | + tcg_gen_op3(INDEX_op_ld_i32, ret, arg2, offset); | |
| 165 | +} | |
| 166 | + | |
| 167 | +static inline void tcg_gen_st8_i32(int arg1, int arg2, tcg_target_long offset) | |
| 168 | +{ | |
| 169 | + tcg_gen_op3(INDEX_op_st8_i32, arg1, arg2, offset); | |
| 170 | +} | |
| 171 | + | |
| 172 | +static inline void tcg_gen_st16_i32(int arg1, int arg2, tcg_target_long offset) | |
| 173 | +{ | |
| 174 | + tcg_gen_op3(INDEX_op_st16_i32, arg1, arg2, offset); | |
| 175 | +} | |
| 176 | + | |
| 177 | +static inline void tcg_gen_st_i32(int arg1, int arg2, tcg_target_long offset) | |
| 178 | +{ | |
| 179 | + tcg_gen_op3(INDEX_op_st_i32, arg1, arg2, offset); | |
| 180 | +} | |
| 181 | + | |
| 182 | +static inline void tcg_gen_add_i32(int ret, int arg1, int arg2) | |
| 183 | +{ | |
| 184 | + tcg_gen_op3(INDEX_op_add_i32, ret, arg1, arg2); | |
| 185 | +} | |
| 186 | + | |
| 187 | +static inline void tcg_gen_addi_i32(int ret, int arg1, int32_t arg2) | |
| 188 | +{ | |
| 189 | + tcg_gen_add_i32(ret, arg1, tcg_const_i32(arg2)); | |
| 190 | +} | |
| 191 | + | |
| 192 | +static inline void tcg_gen_sub_i32(int ret, int arg1, int arg2) | |
| 193 | +{ | |
| 194 | + tcg_gen_op3(INDEX_op_sub_i32, ret, arg1, arg2); | |
| 195 | +} | |
| 196 | + | |
| 197 | +static inline void tcg_gen_subi_i32(int ret, int arg1, int32_t arg2) | |
| 198 | +{ | |
| 199 | + tcg_gen_sub_i32(ret, arg1, tcg_const_i32(arg2)); | |
| 200 | +} | |
| 201 | + | |
| 202 | +static inline void tcg_gen_and_i32(int ret, int arg1, int arg2) | |
| 203 | +{ | |
| 204 | + tcg_gen_op3(INDEX_op_and_i32, ret, arg1, arg2); | |
| 205 | +} | |
| 206 | + | |
| 207 | +static inline void tcg_gen_andi_i32(int ret, int arg1, int32_t arg2) | |
| 208 | +{ | |
| 209 | + /* some cases can be optimized here */ | |
| 210 | + if (arg2 == 0) { | |
| 211 | + tcg_gen_movi_i32(ret, 0); | |
| 212 | + } else if (arg2 == 0xffffffff) { | |
| 213 | + tcg_gen_mov_i32(ret, arg1); | |
| 214 | + } else { | |
| 215 | + tcg_gen_and_i32(ret, arg1, tcg_const_i32(arg2)); | |
| 216 | + } | |
| 217 | +} | |
| 218 | + | |
| 219 | +static inline void tcg_gen_or_i32(int ret, int arg1, int arg2) | |
| 220 | +{ | |
| 221 | + tcg_gen_op3(INDEX_op_or_i32, ret, arg1, arg2); | |
| 222 | +} | |
| 223 | + | |
| 224 | +static inline void tcg_gen_ori_i32(int ret, int arg1, int32_t arg2) | |
| 225 | +{ | |
| 226 | + /* some cases can be optimized here */ | |
| 227 | + if (arg2 == 0xffffffff) { | |
| 228 | + tcg_gen_movi_i32(ret, 0); | |
| 229 | + } else if (arg2 == 0) { | |
| 230 | + tcg_gen_mov_i32(ret, arg1); | |
| 231 | + } else { | |
| 232 | + tcg_gen_or_i32(ret, arg1, tcg_const_i32(arg2)); | |
| 233 | + } | |
| 234 | +} | |
| 235 | + | |
| 236 | +static inline void tcg_gen_xor_i32(int ret, int arg1, int arg2) | |
| 237 | +{ | |
| 238 | + tcg_gen_op3(INDEX_op_xor_i32, ret, arg1, arg2); | |
| 239 | +} | |
| 240 | + | |
| 241 | +static inline void tcg_gen_xori_i32(int ret, int arg1, int32_t arg2) | |
| 242 | +{ | |
| 243 | + /* some cases can be optimized here */ | |
| 244 | + if (arg2 == 0) { | |
| 245 | + tcg_gen_mov_i32(ret, arg1); | |
| 246 | + } else { | |
| 247 | + tcg_gen_xor_i32(ret, arg1, tcg_const_i32(arg2)); | |
| 248 | + } | |
| 249 | +} | |
| 250 | + | |
| 251 | +static inline void tcg_gen_shl_i32(int ret, int arg1, int arg2) | |
| 252 | +{ | |
| 253 | + tcg_gen_op3(INDEX_op_shl_i32, ret, arg1, arg2); | |
| 254 | +} | |
| 255 | + | |
| 256 | +static inline void tcg_gen_shli_i32(int ret, int arg1, int32_t arg2) | |
| 257 | +{ | |
| 258 | + tcg_gen_shl_i32(ret, arg1, tcg_const_i32(arg2)); | |
| 259 | +} | |
| 260 | + | |
| 261 | +static inline void tcg_gen_shr_i32(int ret, int arg1, int arg2) | |
| 262 | +{ | |
| 263 | + tcg_gen_op3(INDEX_op_shr_i32, ret, arg1, arg2); | |
| 264 | +} | |
| 265 | + | |
| 266 | +static inline void tcg_gen_shri_i32(int ret, int arg1, int32_t arg2) | |
| 267 | +{ | |
| 268 | + tcg_gen_shr_i32(ret, arg1, tcg_const_i32(arg2)); | |
| 269 | +} | |
| 270 | + | |
| 271 | +static inline void tcg_gen_sar_i32(int ret, int arg1, int arg2) | |
| 272 | +{ | |
| 273 | + tcg_gen_op3(INDEX_op_sar_i32, ret, arg1, arg2); | |
| 274 | +} | |
| 275 | + | |
| 276 | +static inline void tcg_gen_sari_i32(int ret, int arg1, int32_t arg2) | |
| 277 | +{ | |
| 278 | + tcg_gen_sar_i32(ret, arg1, tcg_const_i32(arg2)); | |
| 279 | +} | |
| 280 | + | |
| 281 | +static inline void tcg_gen_brcond_i32(int cond, TCGArg arg1, TCGArg arg2, | |
| 282 | + int label_index) | |
| 283 | +{ | |
| 284 | + tcg_gen_op4(INDEX_op_brcond_i32, arg1, arg2, cond, label_index); | |
| 285 | +} | |
| 286 | + | |
| 287 | +static inline void tcg_gen_mul_i32(int ret, int arg1, int arg2) | |
| 288 | +{ | |
| 289 | + tcg_gen_op3(INDEX_op_mul_i32, ret, arg1, arg2); | |
| 290 | +} | |
| 291 | + | |
| 292 | +#ifdef TCG_TARGET_HAS_div_i32 | |
| 293 | +static inline void tcg_gen_div_i32(int ret, int arg1, int arg2) | |
| 294 | +{ | |
| 295 | + tcg_gen_op3(INDEX_op_div_i32, ret, arg1, arg2); | |
| 296 | +} | |
| 297 | + | |
| 298 | +static inline void tcg_gen_rem_i32(int ret, int arg1, int arg2) | |
| 299 | +{ | |
| 300 | + tcg_gen_op3(INDEX_op_rem_i32, ret, arg1, arg2); | |
| 301 | +} | |
| 302 | + | |
| 303 | +static inline void tcg_gen_divu_i32(int ret, int arg1, int arg2) | |
| 304 | +{ | |
| 305 | + tcg_gen_op3(INDEX_op_divu_i32, ret, arg1, arg2); | |
| 306 | +} | |
| 307 | + | |
| 308 | +static inline void tcg_gen_remu_i32(int ret, int arg1, int arg2) | |
| 309 | +{ | |
| 310 | + tcg_gen_op3(INDEX_op_remu_i32, ret, arg1, arg2); | |
| 311 | +} | |
| 312 | +#else | |
| 313 | +static inline void tcg_gen_div_i32(int ret, int arg1, int arg2) | |
| 314 | +{ | |
| 315 | + int t0; | |
| 316 | + t0 = tcg_temp_new(TCG_TYPE_I32); | |
| 317 | + tcg_gen_sari_i32(t0, arg1, 31); | |
| 318 | + tcg_gen_op5(INDEX_op_div2_i32, ret, t0, arg1, t0, arg2); | |
| 319 | +} | |
| 320 | + | |
| 321 | +static inline void tcg_gen_rem_i32(int ret, int arg1, int arg2) | |
| 322 | +{ | |
| 323 | + int t0; | |
| 324 | + t0 = tcg_temp_new(TCG_TYPE_I32); | |
| 325 | + tcg_gen_sari_i32(t0, arg1, 31); | |
| 326 | + tcg_gen_op5(INDEX_op_div2_i32, t0, ret, arg1, t0, arg2); | |
| 327 | +} | |
| 328 | + | |
| 329 | +static inline void tcg_gen_divu_i32(int ret, int arg1, int arg2) | |
| 330 | +{ | |
| 331 | + int t0; | |
| 332 | + t0 = tcg_temp_new(TCG_TYPE_I32); | |
| 333 | + tcg_gen_movi_i32(t0, 0); | |
| 334 | + tcg_gen_op5(INDEX_op_divu2_i32, ret, t0, arg1, t0, arg2); | |
| 335 | +} | |
| 336 | + | |
| 337 | +static inline void tcg_gen_remu_i32(int ret, int arg1, int arg2) | |
| 338 | +{ | |
| 339 | + int t0; | |
| 340 | + t0 = tcg_temp_new(TCG_TYPE_I32); | |
| 341 | + tcg_gen_movi_i32(t0, 0); | |
| 342 | + tcg_gen_op5(INDEX_op_divu2_i32, t0, ret, arg1, t0, arg2); | |
| 343 | +} | |
| 344 | +#endif | |
| 345 | + | |
| 346 | +#if TCG_TARGET_REG_BITS == 32 | |
| 347 | + | |
| 348 | +static inline void tcg_gen_mov_i64(int ret, int arg) | |
| 349 | +{ | |
| 350 | + tcg_gen_mov_i32(ret, arg); | |
| 351 | + tcg_gen_mov_i32(ret + 1, arg + 1); | |
| 352 | +} | |
| 353 | + | |
| 354 | +static inline void tcg_gen_movi_i64(int ret, int64_t arg) | |
| 355 | +{ | |
| 356 | + tcg_gen_movi_i32(ret, arg); | |
| 357 | + tcg_gen_movi_i32(ret + 1, arg >> 32); | |
| 358 | +} | |
| 359 | + | |
| 360 | +static inline void tcg_gen_ld8u_i64(int ret, int arg2, tcg_target_long offset) | |
| 361 | +{ | |
| 362 | + tcg_gen_ld8u_i32(ret, arg2, offset); | |
| 363 | + tcg_gen_movi_i32(ret + 1, 0); | |
| 364 | +} | |
| 365 | + | |
| 366 | +static inline void tcg_gen_ld8s_i64(int ret, int arg2, tcg_target_long offset) | |
| 367 | +{ | |
| 368 | + tcg_gen_ld8s_i32(ret, arg2, offset); | |
| 369 | + tcg_gen_sari_i32(ret + 1, ret, 31); | |
| 370 | +} | |
| 371 | + | |
| 372 | +static inline void tcg_gen_ld16u_i64(int ret, int arg2, tcg_target_long offset) | |
| 373 | +{ | |
| 374 | + tcg_gen_ld16u_i32(ret, arg2, offset); | |
| 375 | + tcg_gen_movi_i32(ret + 1, 0); | |
| 376 | +} | |
| 377 | + | |
| 378 | +static inline void tcg_gen_ld16s_i64(int ret, int arg2, tcg_target_long offset) | |
| 379 | +{ | |
| 380 | + tcg_gen_ld16s_i32(ret, arg2, offset); | |
| 381 | + tcg_gen_sari_i32(ret + 1, ret, 31); | |
| 382 | +} | |
| 383 | + | |
| 384 | +static inline void tcg_gen_ld32u_i64(int ret, int arg2, tcg_target_long offset) | |
| 385 | +{ | |
| 386 | + tcg_gen_ld_i32(ret, arg2, offset); | |
| 387 | + tcg_gen_movi_i32(ret + 1, 0); | |
| 388 | +} | |
| 389 | + | |
| 390 | +static inline void tcg_gen_ld32s_i64(int ret, int arg2, tcg_target_long offset) | |
| 391 | +{ | |
| 392 | + tcg_gen_ld_i32(ret, arg2, offset); | |
| 393 | + tcg_gen_sari_i32(ret + 1, ret, 31); | |
| 394 | +} | |
| 395 | + | |
| 396 | +static inline void tcg_gen_ld_i64(int ret, int arg2, tcg_target_long offset) | |
| 397 | +{ | |
| 398 | + /* since arg2 and ret have different types, they cannot be the | |
| 399 | + same temporary */ | |
| 400 | +#ifdef TCG_TARGET_WORDS_BIGENDIAN | |
| 401 | + tcg_gen_ld_i32(ret + 1, arg2, offset); | |
| 402 | + tcg_gen_ld_i32(ret, arg2, offset + 4); | |
| 403 | +#else | |
| 404 | + tcg_gen_ld_i32(ret, arg2, offset); | |
| 405 | + tcg_gen_ld_i32(ret + 1, arg2, offset + 4); | |
| 406 | +#endif | |
| 407 | +} | |
| 408 | + | |
| 409 | +static inline void tcg_gen_st8_i64(int arg1, int arg2, tcg_target_long offset) | |
| 410 | +{ | |
| 411 | + tcg_gen_st8_i32(arg1, arg2, offset); | |
| 412 | +} | |
| 413 | + | |
| 414 | +static inline void tcg_gen_st16_i64(int arg1, int arg2, tcg_target_long offset) | |
| 415 | +{ | |
| 416 | + tcg_gen_st16_i32(arg1, arg2, offset); | |
| 417 | +} | |
| 418 | + | |
| 419 | +static inline void tcg_gen_st32_i64(int arg1, int arg2, tcg_target_long offset) | |
| 420 | +{ | |
| 421 | + tcg_gen_st_i32(arg1, arg2, offset); | |
| 422 | +} | |
| 423 | + | |
| 424 | +static inline void tcg_gen_st_i64(int arg1, int arg2, tcg_target_long offset) | |
| 425 | +{ | |
| 426 | +#ifdef TCG_TARGET_WORDS_BIGENDIAN | |
| 427 | + tcg_gen_st_i32(arg1 + 1, arg2, offset); | |
| 428 | + tcg_gen_st_i32(arg1, arg2, offset + 4); | |
| 429 | +#else | |
| 430 | + tcg_gen_st_i32(arg1, arg2, offset); | |
| 431 | + tcg_gen_st_i32(arg1 + 1, arg2, offset + 4); | |
| 432 | +#endif | |
| 433 | +} | |
| 434 | + | |
| 435 | +static inline void tcg_gen_add_i64(int ret, int arg1, int arg2) | |
| 436 | +{ | |
| 437 | + tcg_gen_op6(INDEX_op_add2_i32, ret, ret + 1, | |
| 438 | + arg1, arg1 + 1, arg2, arg2 + 1); | |
| 439 | +} | |
| 440 | + | |
| 441 | +static inline void tcg_gen_addi_i64(int ret, int arg1, int64_t arg2) | |
| 442 | +{ | |
| 443 | + tcg_gen_add_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 444 | +} | |
| 445 | + | |
| 446 | +static inline void tcg_gen_sub_i64(int ret, int arg1, int arg2) | |
| 447 | +{ | |
| 448 | + tcg_gen_op6(INDEX_op_sub2_i32, ret, ret + 1, | |
| 449 | + arg1, arg1 + 1, arg2, arg2 + 1); | |
| 450 | +} | |
| 451 | + | |
| 452 | +static inline void tcg_gen_subi_i64(int ret, int arg1, int64_t arg2) | |
| 453 | +{ | |
| 454 | + tcg_gen_sub_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 455 | +} | |
| 456 | + | |
| 457 | +static inline void tcg_gen_and_i64(int ret, int arg1, int arg2) | |
| 458 | +{ | |
| 459 | + tcg_gen_and_i32(ret, arg1, arg2); | |
| 460 | + tcg_gen_and_i32(ret + 1, arg1 + 1, arg2 + 1); | |
| 461 | +} | |
| 462 | + | |
| 463 | +static inline void tcg_gen_andi_i64(int ret, int arg1, int64_t arg2) | |
| 464 | +{ | |
| 465 | + tcg_gen_andi_i32(ret, arg1, arg2); | |
| 466 | + tcg_gen_andi_i32(ret + 1, arg1 + 1, arg2 >> 32); | |
| 467 | +} | |
| 468 | + | |
| 469 | +static inline void tcg_gen_or_i64(int ret, int arg1, int arg2) | |
| 470 | +{ | |
| 471 | + tcg_gen_or_i32(ret, arg1, arg2); | |
| 472 | + tcg_gen_or_i32(ret + 1, arg1 + 1, arg2 + 1); | |
| 473 | +} | |
| 474 | + | |
| 475 | +static inline void tcg_gen_ori_i64(int ret, int arg1, int64_t arg2) | |
| 476 | +{ | |
| 477 | + tcg_gen_ori_i32(ret, arg1, arg2); | |
| 478 | + tcg_gen_ori_i32(ret + 1, arg1 + 1, arg2 >> 32); | |
| 479 | +} | |
| 480 | + | |
| 481 | +static inline void tcg_gen_xor_i64(int ret, int arg1, int arg2) | |
| 482 | +{ | |
| 483 | + tcg_gen_xor_i32(ret, arg1, arg2); | |
| 484 | + tcg_gen_xor_i32(ret + 1, arg1 + 1, arg2 + 1); | |
| 485 | +} | |
| 486 | + | |
| 487 | +static inline void tcg_gen_xori_i64(int ret, int arg1, int64_t arg2) | |
| 488 | +{ | |
| 489 | + tcg_gen_xori_i32(ret, arg1, arg2); | |
| 490 | + tcg_gen_xori_i32(ret + 1, arg1 + 1, arg2 >> 32); | |
| 491 | +} | |
| 492 | + | |
| 493 | +/* XXX: use generic code when basic block handling is OK or CPU | |
| 494 | + specific code (x86) */ | |
| 495 | +static inline void tcg_gen_shl_i64(int ret, int arg1, int64_t arg2) | |
| 496 | +{ | |
| 497 | + tcg_gen_helper_1_2(tcg_helper_shl_i64, ret, arg1, arg2); | |
| 498 | +} | |
| 499 | + | |
| 500 | +static inline void tcg_gen_shli_i64(int ret, int arg1, int64_t arg2) | |
| 501 | +{ | |
| 502 | + tcg_gen_shifti_i64(ret, arg1, arg2, 0, 0); | |
| 503 | +} | |
| 504 | + | |
| 505 | +static inline void tcg_gen_shr_i64(int ret, int arg1, int64_t arg2) | |
| 506 | +{ | |
| 507 | + tcg_gen_helper_1_2(tcg_helper_shr_i64, ret, arg1, arg2); | |
| 508 | +} | |
| 509 | + | |
| 510 | +static inline void tcg_gen_shri_i64(int ret, int arg1, int64_t arg2) | |
| 511 | +{ | |
| 512 | + tcg_gen_shifti_i64(ret, arg1, arg2, 1, 0); | |
| 513 | +} | |
| 514 | + | |
| 515 | +static inline void tcg_gen_sar_i64(int ret, int arg1, int64_t arg2) | |
| 516 | +{ | |
| 517 | + tcg_gen_helper_1_2(tcg_helper_sar_i64, ret, arg1, arg2); | |
| 518 | +} | |
| 519 | + | |
| 520 | +static inline void tcg_gen_sari_i64(int ret, int arg1, int64_t arg2) | |
| 521 | +{ | |
| 522 | + tcg_gen_shifti_i64(ret, arg1, arg2, 1, 1); | |
| 523 | +} | |
| 524 | + | |
| 525 | +static inline void tcg_gen_brcond_i64(int cond, TCGArg arg1, TCGArg arg2, | |
| 526 | + int label_index) | |
| 527 | +{ | |
| 528 | + tcg_gen_op6(INDEX_op_brcond2_i32, | |
| 529 | + arg1, arg1 + 1, arg2, arg2 + 1, cond, label_index); | |
| 530 | +} | |
| 531 | + | |
| 532 | +static inline void tcg_gen_mul_i64(int ret, int arg1, int arg2) | |
| 533 | +{ | |
| 534 | + int t0, t1; | |
| 535 | + | |
| 536 | + t0 = tcg_temp_new(TCG_TYPE_I64); | |
| 537 | + t1 = tcg_temp_new(TCG_TYPE_I32); | |
| 538 | + | |
| 539 | + tcg_gen_op4(INDEX_op_mulu2_i32, t0, t0 + 1, arg1, arg2); | |
| 540 | + | |
| 541 | + tcg_gen_mul_i32(t1, arg1, arg2 + 1); | |
| 542 | + tcg_gen_add_i32(t0 + 1, t0 + 1, t1); | |
| 543 | + tcg_gen_mul_i32(t1, arg1 + 1, arg2); | |
| 544 | + tcg_gen_add_i32(t0 + 1, t0 + 1, t1); | |
| 545 | + | |
| 546 | + tcg_gen_mov_i64(ret, t0); | |
| 547 | +} | |
| 548 | + | |
| 549 | +static inline void tcg_gen_div_i64(int ret, int arg1, int64_t arg2) | |
| 550 | +{ | |
| 551 | + tcg_gen_helper_1_2(tcg_helper_div_i64, ret, arg1, arg2); | |
| 552 | +} | |
| 553 | + | |
| 554 | +static inline void tcg_gen_rem_i64(int ret, int arg1, int64_t arg2) | |
| 555 | +{ | |
| 556 | + tcg_gen_helper_1_2(tcg_helper_rem_i64, ret, arg1, arg2); | |
| 557 | +} | |
| 558 | + | |
| 559 | +static inline void tcg_gen_divu_i64(int ret, int arg1, int64_t arg2) | |
| 560 | +{ | |
| 561 | + tcg_gen_helper_1_2(tcg_helper_divu_i64, ret, arg1, arg2); | |
| 562 | +} | |
| 563 | + | |
| 564 | +static inline void tcg_gen_remu_i64(int ret, int arg1, int64_t arg2) | |
| 565 | +{ | |
| 566 | + tcg_gen_helper_1_2(tcg_helper_remu_i64, ret, arg1, arg2); | |
| 567 | +} | |
| 568 | + | |
| 569 | +#else | |
| 570 | + | |
| 571 | +static inline void tcg_gen_mov_i64(int ret, int arg) | |
| 572 | +{ | |
| 573 | + tcg_gen_op2(INDEX_op_mov_i64, ret, arg); | |
| 574 | +} | |
| 575 | + | |
| 576 | +static inline void tcg_gen_movi_i64(int ret, int64_t arg) | |
| 577 | +{ | |
| 578 | + tcg_gen_op2(INDEX_op_movi_i64, ret, arg); | |
| 579 | +} | |
| 580 | + | |
| 581 | +static inline void tcg_gen_ld8u_i64(int ret, int arg2, tcg_target_long offset) | |
| 582 | +{ | |
| 583 | + tcg_gen_op3(INDEX_op_ld8u_i64, ret, arg2, offset); | |
| 584 | +} | |
| 585 | + | |
| 586 | +static inline void tcg_gen_ld8s_i64(int ret, int arg2, tcg_target_long offset) | |
| 587 | +{ | |
| 588 | + tcg_gen_op3(INDEX_op_ld8s_i64, ret, arg2, offset); | |
| 589 | +} | |
| 590 | + | |
| 591 | +static inline void tcg_gen_ld16u_i64(int ret, int arg2, tcg_target_long offset) | |
| 592 | +{ | |
| 593 | + tcg_gen_op3(INDEX_op_ld16u_i64, ret, arg2, offset); | |
| 594 | +} | |
| 595 | + | |
| 596 | +static inline void tcg_gen_ld16s_i64(int ret, int arg2, tcg_target_long offset) | |
| 597 | +{ | |
| 598 | + tcg_gen_op3(INDEX_op_ld16s_i64, ret, arg2, offset); | |
| 599 | +} | |
| 600 | + | |
| 601 | +static inline void tcg_gen_ld32u_i64(int ret, int arg2, tcg_target_long offset) | |
| 602 | +{ | |
| 603 | + tcg_gen_op3(INDEX_op_ld32u_i64, ret, arg2, offset); | |
| 604 | +} | |
| 605 | + | |
| 606 | +static inline void tcg_gen_ld32s_i64(int ret, int arg2, tcg_target_long offset) | |
| 607 | +{ | |
| 608 | + tcg_gen_op3(INDEX_op_ld32s_i64, ret, arg2, offset); | |
| 609 | +} | |
| 610 | + | |
| 611 | +static inline void tcg_gen_ld_i64(int ret, int arg2, tcg_target_long offset) | |
| 612 | +{ | |
| 613 | + tcg_gen_op3(INDEX_op_ld_i64, ret, arg2, offset); | |
| 614 | +} | |
| 615 | + | |
| 616 | +static inline void tcg_gen_st8_i64(int arg1, int arg2, tcg_target_long offset) | |
| 617 | +{ | |
| 618 | + tcg_gen_op3(INDEX_op_st8_i64, arg1, arg2, offset); | |
| 619 | +} | |
| 620 | + | |
| 621 | +static inline void tcg_gen_st16_i64(int arg1, int arg2, tcg_target_long offset) | |
| 622 | +{ | |
| 623 | + tcg_gen_op3(INDEX_op_st16_i64, arg1, arg2, offset); | |
| 624 | +} | |
| 625 | + | |
| 626 | +static inline void tcg_gen_st32_i64(int arg1, int arg2, tcg_target_long offset) | |
| 627 | +{ | |
| 628 | + tcg_gen_op3(INDEX_op_st32_i64, arg1, arg2, offset); | |
| 629 | +} | |
| 630 | + | |
| 631 | +static inline void tcg_gen_st_i64(int arg1, int arg2, tcg_target_long offset) | |
| 632 | +{ | |
| 633 | + tcg_gen_op3(INDEX_op_st_i64, arg1, arg2, offset); | |
| 634 | +} | |
| 635 | + | |
| 636 | +static inline void tcg_gen_add_i64(int ret, int arg1, int arg2) | |
| 637 | +{ | |
| 638 | + tcg_gen_op3(INDEX_op_add_i64, ret, arg1, arg2); | |
| 639 | +} | |
| 640 | + | |
| 641 | +static inline void tcg_gen_addi_i64(int ret, int arg1, int64_t arg2) | |
| 642 | +{ | |
| 643 | + tcg_gen_add_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 644 | +} | |
| 645 | + | |
| 646 | +static inline void tcg_gen_sub_i64(int ret, int arg1, int arg2) | |
| 647 | +{ | |
| 648 | + tcg_gen_op3(INDEX_op_sub_i64, ret, arg1, arg2); | |
| 649 | +} | |
| 650 | + | |
| 651 | +static inline void tcg_gen_subi_i64(int ret, int arg1, int64_t arg2) | |
| 652 | +{ | |
| 653 | + tcg_gen_sub_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 654 | +} | |
| 655 | + | |
| 656 | +static inline void tcg_gen_and_i64(int ret, int arg1, int arg2) | |
| 657 | +{ | |
| 658 | + tcg_gen_op3(INDEX_op_and_i64, ret, arg1, arg2); | |
| 659 | +} | |
| 660 | + | |
| 661 | +static inline void tcg_gen_andi_i64(int ret, int arg1, int64_t arg2) | |
| 662 | +{ | |
| 663 | + tcg_gen_and_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 664 | +} | |
| 665 | + | |
| 666 | +static inline void tcg_gen_or_i64(int ret, int arg1, int arg2) | |
| 667 | +{ | |
| 668 | + tcg_gen_op3(INDEX_op_or_i64, ret, arg1, arg2); | |
| 669 | +} | |
| 670 | + | |
| 671 | +static inline void tcg_gen_ori_i64(int ret, int arg1, int64_t arg2) | |
| 672 | +{ | |
| 673 | + tcg_gen_or_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 674 | +} | |
| 675 | + | |
| 676 | +static inline void tcg_gen_xor_i64(int ret, int arg1, int arg2) | |
| 677 | +{ | |
| 678 | + tcg_gen_op3(INDEX_op_xor_i64, ret, arg1, arg2); | |
| 679 | +} | |
| 680 | + | |
| 681 | +static inline void tcg_gen_xori_i64(int ret, int arg1, int64_t arg2) | |
| 682 | +{ | |
| 683 | + tcg_gen_xor_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 684 | +} | |
| 685 | + | |
| 686 | +static inline void tcg_gen_shl_i64(int ret, int arg1, int arg2) | |
| 687 | +{ | |
| 688 | + tcg_gen_op3(INDEX_op_shl_i64, ret, arg1, arg2); | |
| 689 | +} | |
| 690 | + | |
| 691 | +static inline void tcg_gen_shli_i64(int ret, int arg1, int64_t arg2) | |
| 692 | +{ | |
| 693 | + tcg_gen_shl_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 694 | +} | |
| 695 | + | |
| 696 | +static inline void tcg_gen_shr_i64(int ret, int arg1, int arg2) | |
| 697 | +{ | |
| 698 | + tcg_gen_op3(INDEX_op_shr_i64, ret, arg1, arg2); | |
| 699 | +} | |
| 700 | + | |
| 701 | +static inline void tcg_gen_shri_i64(int ret, int arg1, int64_t arg2) | |
| 702 | +{ | |
| 703 | + tcg_gen_shr_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 704 | +} | |
| 705 | + | |
| 706 | +static inline void tcg_gen_sar_i64(int ret, int arg1, int arg2) | |
| 707 | +{ | |
| 708 | + tcg_gen_op3(INDEX_op_sar_i64, ret, arg1, arg2); | |
| 709 | +} | |
| 710 | + | |
| 711 | +static inline void tcg_gen_sari_i64(int ret, int arg1, int64_t arg2) | |
| 712 | +{ | |
| 713 | + tcg_gen_sar_i64(ret, arg1, tcg_const_i64(arg2)); | |
| 714 | +} | |
| 715 | + | |
| 716 | +static inline void tcg_gen_brcond_i64(int cond, TCGArg arg1, TCGArg arg2, | |
| 717 | + int label_index) | |
| 718 | +{ | |
| 719 | + tcg_gen_op4(INDEX_op_brcond_i64, arg1, arg2, cond, label_index); | |
| 720 | +} | |
| 721 | + | |
| 722 | +static inline void tcg_gen_mul_i64(int ret, int arg1, int arg2) | |
| 723 | +{ | |
| 724 | + tcg_gen_op3(INDEX_op_mul_i64, ret, arg1, arg2); | |
| 725 | +} | |
| 726 | + | |
| 727 | +#ifdef TCG_TARGET_HAS_div_i64 | |
| 728 | +static inline void tcg_gen_div_i64(int ret, int arg1, int arg2) | |
| 729 | +{ | |
| 730 | + tcg_gen_op3(INDEX_op_div_i64, ret, arg1, arg2); | |
| 731 | +} | |
| 732 | + | |
| 733 | +static inline void tcg_gen_rem_i64(int ret, int arg1, int arg2) | |
| 734 | +{ | |
| 735 | + tcg_gen_op3(INDEX_op_rem_i64, ret, arg1, arg2); | |
| 736 | +} | |
| 737 | + | |
| 738 | +static inline void tcg_gen_divu_i64(int ret, int arg1, int arg2) | |
| 739 | +{ | |
| 740 | + tcg_gen_op3(INDEX_op_divu_i64, ret, arg1, arg2); | |
| 741 | +} | |
| 742 | + | |
| 743 | +static inline void tcg_gen_remu_i64(int ret, int arg1, int arg2) | |
| 744 | +{ | |
| 745 | + tcg_gen_op3(INDEX_op_remu_i64, ret, arg1, arg2); | |
| 746 | +} | |
| 747 | +#else | |
| 748 | +static inline void tcg_gen_div_i64(int ret, int arg1, int arg2) | |
| 749 | +{ | |
| 750 | + int t0; | |
| 751 | + t0 = tcg_temp_new(TCG_TYPE_I64); | |
| 752 | + tcg_gen_sari_i64(t0, arg1, 63); | |
| 753 | + tcg_gen_op5(INDEX_op_div2_i64, ret, t0, arg1, t0, arg2); | |
| 754 | +} | |
| 755 | + | |
| 756 | +static inline void tcg_gen_rem_i64(int ret, int arg1, int arg2) | |
| 757 | +{ | |
| 758 | + int t0; | |
| 759 | + t0 = tcg_temp_new(TCG_TYPE_I64); | |
| 760 | + tcg_gen_sari_i64(t0, arg1, 63); | |
| 761 | + tcg_gen_op5(INDEX_op_div2_i64, t0, ret, arg1, t0, arg2); | |
| 762 | +} | |
| 763 | + | |
| 764 | +static inline void tcg_gen_divu_i64(int ret, int arg1, int arg2) | |
| 765 | +{ | |
| 766 | + int t0; | |
| 767 | + t0 = tcg_temp_new(TCG_TYPE_I64); | |
| 768 | + tcg_gen_movi_i64(t0, 0); | |
| 769 | + tcg_gen_op5(INDEX_op_divu2_i64, ret, t0, arg1, t0, arg2); | |
| 770 | +} | |
| 771 | + | |
| 772 | +static inline void tcg_gen_remu_i64(int ret, int arg1, int arg2) | |
| 773 | +{ | |
| 774 | + int t0; | |
| 775 | + t0 = tcg_temp_new(TCG_TYPE_I64); | |
| 776 | + tcg_gen_movi_i64(t0, 0); | |
| 777 | + tcg_gen_op5(INDEX_op_divu2_i64, t0, ret, arg1, t0, arg2); | |
| 778 | +} | |
| 779 | +#endif | |
| 780 | + | |
| 781 | +#endif | |
| 782 | + | |
| 783 | +/***************************************/ | |
| 784 | +/* optional operations */ | |
| 785 | + | |
| 786 | +static inline void tcg_gen_ext8s_i32(int ret, int arg) | |
| 787 | +{ | |
| 788 | +#ifdef TCG_TARGET_HAS_ext8s_i32 | |
| 789 | + tcg_gen_op2(INDEX_op_ext8s_i32, ret, arg); | |
| 790 | +#else | |
| 791 | + tcg_gen_shli_i32(ret, arg, 24); | |
| 792 | + tcg_gen_sari_i32(ret, arg, 24); | |
| 793 | +#endif | |
| 794 | +} | |
| 795 | + | |
| 796 | +static inline void tcg_gen_ext16s_i32(int ret, int arg) | |
| 797 | +{ | |
| 798 | +#ifdef TCG_TARGET_HAS_ext16s_i32 | |
| 799 | + tcg_gen_op2(INDEX_op_ext16s_i32, ret, arg); | |
| 800 | +#else | |
| 801 | + tcg_gen_shli_i32(ret, arg, 16); | |
| 802 | + tcg_gen_sari_i32(ret, arg, 16); | |
| 803 | +#endif | |
| 804 | +} | |
| 805 | + | |
| 806 | +/* Note: we assume the two high bytes are set to zero */ | |
| 807 | +static inline void tcg_gen_bswap16_i32(TCGArg ret, TCGArg arg) | |
| 808 | +{ | |
| 809 | +#ifdef TCG_TARGET_HAS_bswap16_i32 | |
| 810 | + tcg_gen_op2(INDEX_op_bswap16_i32, ret, arg); | |
| 811 | +#else | |
| 812 | + TCGArg t0, t1; | |
| 813 | + t0 = tcg_temp_new(TCG_TYPE_I32); | |
| 814 | + t1 = tcg_temp_new(TCG_TYPE_I32); | |
| 815 | + | |
| 816 | + tcg_gen_shri_i32(t0, arg, 8); | |
| 817 | + tcg_gen_andi_i32(t1, arg, 0x000000ff); | |
| 818 | + tcg_gen_shli_i32(t1, t1, 8); | |
| 819 | + tcg_gen_or_i32(ret, t0, t1); | |
| 820 | +#endif | |
| 821 | +} | |
| 822 | + | |
| 823 | +static inline void tcg_gen_bswap_i32(TCGArg ret, TCGArg arg) | |
| 824 | +{ | |
| 825 | +#ifdef TCG_TARGET_HAS_bswap_i32 | |
| 826 | + tcg_gen_op2(INDEX_op_bswap_i32, ret, arg); | |
| 827 | +#else | |
| 828 | + TCGArg t0, t1; | |
| 829 | + t0 = tcg_temp_new(TCG_TYPE_I32); | |
| 830 | + t1 = tcg_temp_new(TCG_TYPE_I32); | |
| 831 | + | |
| 832 | + tcg_gen_shli_i32(t0, arg, 24); | |
| 833 | + | |
| 834 | + tcg_gen_andi_i32(t1, arg, 0x0000ff00); | |
| 835 | + tcg_gen_shli_i32(t1, t1, 8); | |
| 836 | + tcg_gen_or_i32(t0, t0, t1); | |
| 837 | + | |
| 838 | + tcg_gen_shri_i32(t1, arg, 8); | |
| 839 | + tcg_gen_andi_i32(t1, t1, 0x0000ff00); | |
| 840 | + tcg_gen_or_i32(t0, t0, t1); | |
| 841 | + | |
| 842 | + tcg_gen_shri_i32(t1, arg, 24); | |
| 843 | + tcg_gen_or_i32(ret, t0, t1); | |
| 844 | +#endif | |
| 845 | +} | |
| 846 | + | |
| 847 | +#if TCG_TARGET_REG_BITS == 32 | |
| 848 | +static inline void tcg_gen_ext8s_i64(int ret, int arg) | |
| 849 | +{ | |
| 850 | + tcg_gen_ext8s_i32(ret, arg); | |
| 851 | + tcg_gen_sari_i32(ret + 1, ret, 31); | |
| 852 | +} | |
| 853 | + | |
| 854 | +static inline void tcg_gen_ext16s_i64(int ret, int arg) | |
| 855 | +{ | |
| 856 | + tcg_gen_ext16s_i32(ret, arg); | |
| 857 | + tcg_gen_sari_i32(ret + 1, ret, 31); | |
| 858 | +} | |
| 859 | + | |
| 860 | +static inline void tcg_gen_ext32s_i64(int ret, int arg) | |
| 861 | +{ | |
| 862 | + tcg_gen_mov_i32(ret, arg); | |
| 863 | + tcg_gen_sari_i32(ret + 1, ret, 31); | |
| 864 | +} | |
| 865 | + | |
| 866 | +static inline void tcg_gen_trunc_i64_i32(int ret, int arg) | |
| 867 | +{ | |
| 868 | + tcg_gen_mov_i32(ret, arg); | |
| 869 | +} | |
| 870 | + | |
| 871 | +static inline void tcg_gen_extu_i32_i64(int ret, int arg) | |
| 872 | +{ | |
| 873 | + tcg_gen_mov_i32(ret, arg); | |
| 874 | + tcg_gen_movi_i32(ret + 1, 0); | |
| 875 | +} | |
| 876 | + | |
| 877 | +static inline void tcg_gen_ext_i32_i64(int ret, int arg) | |
| 878 | +{ | |
| 879 | + tcg_gen_mov_i32(ret, arg); | |
| 880 | + tcg_gen_sari_i32(ret + 1, ret, 31); | |
| 881 | +} | |
| 882 | + | |
| 883 | +static inline void tcg_gen_bswap_i64(int ret, int arg) | |
| 884 | +{ | |
| 885 | + int t0, t1; | |
| 886 | + t0 = tcg_temp_new(TCG_TYPE_I32); | |
| 887 | + t1 = tcg_temp_new(TCG_TYPE_I32); | |
| 888 | + | |
| 889 | + tcg_gen_bswap_i32(t0, arg); | |
| 890 | + tcg_gen_bswap_i32(t1, arg + 1); | |
| 891 | + tcg_gen_mov_i32(ret, t1); | |
| 892 | + tcg_gen_mov_i32(ret + 1, t0); | |
| 893 | +} | |
| 894 | +#else | |
| 895 | + | |
| 896 | +static inline void tcg_gen_ext8s_i64(int ret, int arg) | |
| 897 | +{ | |
| 898 | +#ifdef TCG_TARGET_HAS_ext8s_i64 | |
| 899 | + tcg_gen_op2(INDEX_op_ext8s_i64, ret, arg); | |
| 900 | +#else | |
| 901 | + tcg_gen_shli_i64(ret, arg, 56); | |
| 902 | + tcg_gen_sari_i64(ret, arg, 56); | |
| 903 | +#endif | |
| 904 | +} | |
| 905 | + | |
| 906 | +static inline void tcg_gen_ext16s_i64(int ret, int arg) | |
| 907 | +{ | |
| 908 | +#ifdef TCG_TARGET_HAS_ext16s_i64 | |
| 909 | + tcg_gen_op2(INDEX_op_ext16s_i64, ret, arg); | |
| 910 | +#else | |
| 911 | + tcg_gen_shli_i64(ret, arg, 48); | |
| 912 | + tcg_gen_sari_i64(ret, arg, 48); | |
| 913 | +#endif | |
| 914 | +} | |
| 915 | + | |
| 916 | +static inline void tcg_gen_ext32s_i64(int ret, int arg) | |
| 917 | +{ | |
| 918 | +#ifdef TCG_TARGET_HAS_ext32s_i64 | |
| 919 | + tcg_gen_op2(INDEX_op_ext32s_i64, ret, arg); | |
| 920 | +#else | |
| 921 | + tcg_gen_shli_i64(ret, arg, 32); | |
| 922 | + tcg_gen_sari_i64(ret, arg, 32); | |
| 923 | +#endif | |
| 924 | +} | |
| 925 | + | |
| 926 | +/* Note: we assume the target supports move between 32 and 64 bit | |
| 927 | + registers */ | |
| 928 | +static inline void tcg_gen_trunc_i64_i32(int ret, int arg) | |
| 929 | +{ | |
| 930 | + tcg_gen_mov_i32(ret, arg); | |
| 931 | +} | |
| 932 | + | |
| 933 | +/* Note: we assume the target supports move between 32 and 64 bit | |
| 934 | + registers */ | |
| 935 | +static inline void tcg_gen_extu_i32_i64(int ret, int arg) | |
| 936 | +{ | |
| 937 | + tcg_gen_andi_i64(ret, arg, 0xffffffff); | |
| 938 | +} | |
| 939 | + | |
| 940 | +/* Note: we assume the target supports move between 32 and 64 bit | |
| 941 | + registers */ | |
| 942 | +static inline void tcg_gen_ext_i32_i64(int ret, int arg) | |
| 943 | +{ | |
| 944 | + tcg_gen_ext32s_i64(ret, arg); | |
| 945 | +} | |
| 946 | + | |
| 947 | +static inline void tcg_gen_bswap_i64(TCGArg ret, TCGArg arg) | |
| 948 | +{ | |
| 949 | +#ifdef TCG_TARGET_HAS_bswap_i64 | |
| 950 | + tcg_gen_op2(INDEX_op_bswap_i64, ret, arg); | |
| 951 | +#else | |
| 952 | + TCGArg t0, t1; | |
| 953 | + t0 = tcg_temp_new(TCG_TYPE_I32); | |
| 954 | + t1 = tcg_temp_new(TCG_TYPE_I32); | |
| 955 | + | |
| 956 | + tcg_gen_shli_i64(t0, arg, 56); | |
| 957 | + | |
| 958 | + tcg_gen_andi_i64(t1, arg, 0x0000ff00); | |
| 959 | + tcg_gen_shli_i64(t1, t1, 40); | |
| 960 | + tcg_gen_or_i64(t0, t0, t1); | |
| 961 | + | |
| 962 | + tcg_gen_andi_i64(t1, arg, 0x00ff0000); | |
| 963 | + tcg_gen_shli_i64(t1, t1, 24); | |
| 964 | + tcg_gen_or_i64(t0, t0, t1); | |
| 965 | + | |
| 966 | + tcg_gen_andi_i64(t1, arg, 0xff000000); | |
| 967 | + tcg_gen_shli_i64(t1, t1, 8); | |
| 968 | + tcg_gen_or_i64(t0, t0, t1); | |
| 969 | + | |
| 970 | + tcg_gen_shri_i64(t1, arg, 8); | |
| 971 | + tcg_gen_andi_i64(t1, t1, 0xff000000); | |
| 972 | + tcg_gen_or_i64(t0, t0, t1); | |
| 973 | + | |
| 974 | + tcg_gen_shri_i64(t1, arg, 24); | |
| 975 | + tcg_gen_andi_i64(t1, t1, 0x00ff0000); | |
| 976 | + tcg_gen_or_i64(t0, t0, t1); | |
| 977 | + | |
| 978 | + tcg_gen_shri_i64(t1, arg, 40); | |
| 979 | + tcg_gen_andi_i64(t1, t1, 0x0000ff00); | |
| 980 | + tcg_gen_or_i64(t0, t0, t1); | |
| 981 | + | |
| 982 | + tcg_gen_shri_i64(t1, arg, 56); | |
| 983 | + tcg_gen_or_i64(ret, t0, t1); | |
| 984 | +#endif | |
| 985 | +} | |
| 986 | + | |
| 987 | +#endif | |
| 988 | + | |
| 989 | +/***************************************/ | |
| 990 | +static inline void tcg_gen_macro_2(int ret0, int ret1, int macro_id) | |
| 991 | +{ | |
| 992 | + tcg_gen_op3(INDEX_op_macro_2, ret0, ret1, macro_id); | |
| 993 | +} | |
| 994 | + | |
| 995 | +/***************************************/ | |
| 996 | +/* QEMU specific operations. Their type depend on the QEMU CPU | |
| 997 | + type. */ | |
| 998 | +#ifndef TARGET_LONG_BITS | |
| 999 | +#error must include QEMU headers | |
| 1000 | +#endif | |
| 1001 | + | |
| 1002 | +static inline void tcg_gen_exit_tb(tcg_target_long val) | |
| 1003 | +{ | |
| 1004 | + tcg_gen_op1(INDEX_op_exit_tb, val); | |
| 1005 | +} | |
| 1006 | + | |
| 1007 | +static inline void tcg_gen_goto_tb(int idx) | |
| 1008 | +{ | |
| 1009 | + tcg_gen_op1(INDEX_op_goto_tb, idx); | |
| 1010 | +} | |
| 1011 | + | |
| 1012 | +#if TCG_TARGET_REG_BITS == 32 | |
| 1013 | +static inline void tcg_gen_qemu_ld8u(int ret, int addr, int mem_index) | |
| 1014 | +{ | |
| 1015 | +#if TARGET_LONG_BITS == 32 | |
| 1016 | + tcg_gen_op3(INDEX_op_qemu_ld8u, ret, addr, mem_index); | |
| 1017 | +#else | |
| 1018 | + tcg_gen_op4(INDEX_op_qemu_ld8u, ret, addr, addr + 1, mem_index); | |
| 1019 | + tcg_gen_movi_i32(ret + 1, 0); | |
| 1020 | +#endif | |
| 1021 | +} | |
| 1022 | + | |
| 1023 | +static inline void tcg_gen_qemu_ld8s(int ret, int addr, int mem_index) | |
| 1024 | +{ | |
| 1025 | +#if TARGET_LONG_BITS == 32 | |
| 1026 | + tcg_gen_op3(INDEX_op_qemu_ld8s, ret, addr, mem_index); | |
| 1027 | +#else | |
| 1028 | + tcg_gen_op4(INDEX_op_qemu_ld8s, ret, addr, addr + 1, mem_index); | |
| 1029 | + tcg_gen_ext8s_i32(ret + 1, ret); | |
| 1030 | +#endif | |
| 1031 | +} | |
| 1032 | + | |
| 1033 | +static inline void tcg_gen_qemu_ld16u(int ret, int addr, int mem_index) | |
| 1034 | +{ | |
| 1035 | +#if TARGET_LONG_BITS == 32 | |
| 1036 | + tcg_gen_op3(INDEX_op_qemu_ld16u, ret, addr, mem_index); | |
| 1037 | +#else | |
| 1038 | + tcg_gen_op4(INDEX_op_qemu_ld16u, ret, addr, addr + 1, mem_index); | |
| 1039 | + tcg_gen_movi_i32(ret + 1, 0); | |
| 1040 | +#endif | |
| 1041 | +} | |
| 1042 | + | |
| 1043 | +static inline void tcg_gen_qemu_ld16s(int ret, int addr, int mem_index) | |
| 1044 | +{ | |
| 1045 | +#if TARGET_LONG_BITS == 32 | |
| 1046 | + tcg_gen_op3(INDEX_op_qemu_ld16s, ret, addr, mem_index); | |
| 1047 | +#else | |
| 1048 | + tcg_gen_op4(INDEX_op_qemu_ld16s, ret, addr, addr + 1, mem_index); | |
| 1049 | + tcg_gen_ext16s_i32(ret + 1, ret); | |
| 1050 | +#endif | |
| 1051 | +} | |
| 1052 | + | |
| 1053 | +static inline void tcg_gen_qemu_ld32u(int ret, int addr, int mem_index) | |
| 1054 | +{ | |
| 1055 | +#if TARGET_LONG_BITS == 32 | |
| 1056 | + tcg_gen_op3(INDEX_op_qemu_ld32u, ret, addr, mem_index); | |
| 1057 | +#else | |
| 1058 | + tcg_gen_op4(INDEX_op_qemu_ld32u, ret, addr, addr + 1, mem_index); | |
| 1059 | + tcg_gen_movi_i32(ret + 1, 0); | |
| 1060 | +#endif | |
| 1061 | +} | |
| 1062 | + | |
| 1063 | +static inline void tcg_gen_qemu_ld32s(int ret, int addr, int mem_index) | |
| 1064 | +{ | |
| 1065 | +#if TARGET_LONG_BITS == 32 | |
| 1066 | + tcg_gen_op3(INDEX_op_qemu_ld32u, ret, addr, mem_index); | |
| 1067 | +#else | |
| 1068 | + tcg_gen_op4(INDEX_op_qemu_ld32u, ret, addr, addr + 1, mem_index); | |
| 1069 | + tcg_gen_sari_i32(ret + 1, ret, 31); | |
| 1070 | +#endif | |
| 1071 | +} | |
| 1072 | + | |
| 1073 | +static inline void tcg_gen_qemu_ld64(int ret, int addr, int mem_index) | |
| 1074 | +{ | |
| 1075 | +#if TARGET_LONG_BITS == 32 | |
| 1076 | + tcg_gen_op4(INDEX_op_qemu_ld64, ret, ret + 1, addr, mem_index); | |
| 1077 | +#else | |
| 1078 | + tcg_gen_op5(INDEX_op_qemu_ld64, ret, ret + 1, addr, addr + 1, mem_index); | |
| 1079 | +#endif | |
| 1080 | +} | |
| 1081 | + | |
| 1082 | +static inline void tcg_gen_qemu_st8(int arg, int addr, int mem_index) | |
| 1083 | +{ | |
| 1084 | +#if TARGET_LONG_BITS == 32 | |
| 1085 | + tcg_gen_op3(INDEX_op_qemu_st8, arg, addr, mem_index); | |
| 1086 | +#else | |
| 1087 | + tcg_gen_op4(INDEX_op_qemu_st8, arg, addr, addr + 1, mem_index); | |
| 1088 | +#endif | |
| 1089 | +} | |
| 1090 | + | |
| 1091 | +static inline void tcg_gen_qemu_st16(int arg, int addr, int mem_index) | |
| 1092 | +{ | |
| 1093 | +#if TARGET_LONG_BITS == 32 | |
| 1094 | + tcg_gen_op3(INDEX_op_qemu_st16, arg, addr, mem_index); | |
| 1095 | +#else | |
| 1096 | + tcg_gen_op4(INDEX_op_qemu_st16, arg, addr, addr + 1, mem_index); | |
| 1097 | +#endif | |
| 1098 | +} | |
| 1099 | + | |
| 1100 | +static inline void tcg_gen_qemu_st32(int arg, int addr, int mem_index) | |
| 1101 | +{ | |
| 1102 | +#if TARGET_LONG_BITS == 32 | |
| 1103 | + tcg_gen_op3(INDEX_op_qemu_st32, arg, addr, mem_index); | |
| 1104 | +#else | |
| 1105 | + tcg_gen_op4(INDEX_op_qemu_st32, arg, addr, addr + 1, mem_index); | |
| 1106 | +#endif | |
| 1107 | +} | |
| 1108 | + | |
| 1109 | +static inline void tcg_gen_qemu_st64(int arg, int addr, int mem_index) | |
| 1110 | +{ | |
| 1111 | +#if TARGET_LONG_BITS == 32 | |
| 1112 | + tcg_gen_op4(INDEX_op_qemu_st64, arg, arg + 1, addr, mem_index); | |
| 1113 | +#else | |
| 1114 | + tcg_gen_op5(INDEX_op_qemu_st64, arg, arg + 1, addr, addr + 1, mem_index); | |
| 1115 | +#endif | |
| 1116 | +} | |
| 1117 | + | |
| 1118 | +#else /* TCG_TARGET_REG_BITS == 32 */ | |
| 1119 | + | |
| 1120 | +static inline void tcg_gen_qemu_ld8u(int ret, int addr, int mem_index) | |
| 1121 | +{ | |
| 1122 | + tcg_gen_op3(INDEX_op_qemu_ld8u, ret, addr, mem_index); | |
| 1123 | +} | |
| 1124 | + | |
| 1125 | +static inline void tcg_gen_qemu_ld8s(int ret, int addr, int mem_index) | |
| 1126 | +{ | |
| 1127 | + tcg_gen_op3(INDEX_op_qemu_ld8s, ret, addr, mem_index); | |
| 1128 | +} | |
| 1129 | + | |
| 1130 | +static inline void tcg_gen_qemu_ld16u(int ret, int addr, int mem_index) | |
| 1131 | +{ | |
| 1132 | + tcg_gen_op3(INDEX_op_qemu_ld16u, ret, addr, mem_index); | |
| 1133 | +} | |
| 1134 | + | |
| 1135 | +static inline void tcg_gen_qemu_ld16s(int ret, int addr, int mem_index) | |
| 1136 | +{ | |
| 1137 | + tcg_gen_op3(INDEX_op_qemu_ld16s, ret, addr, mem_index); | |
| 1138 | +} | |
| 1139 | + | |
| 1140 | +static inline void tcg_gen_qemu_ld32u(int ret, int addr, int mem_index) | |
| 1141 | +{ | |
| 1142 | + tcg_gen_op3(INDEX_op_qemu_ld32u, ret, addr, mem_index); | |
| 1143 | +} | |
| 1144 | + | |
| 1145 | +static inline void tcg_gen_qemu_ld32s(int ret, int addr, int mem_index) | |
| 1146 | +{ | |
| 1147 | + tcg_gen_op3(INDEX_op_qemu_ld32s, ret, addr, mem_index); | |
| 1148 | +} | |
| 1149 | + | |
| 1150 | +static inline void tcg_gen_qemu_ld64(int ret, int addr, int mem_index) | |
| 1151 | +{ | |
| 1152 | + tcg_gen_op3(INDEX_op_qemu_ld64, ret, addr, mem_index); | |
| 1153 | +} | |
| 1154 | + | |
| 1155 | +static inline void tcg_gen_qemu_st8(int arg, int addr, int mem_index) | |
| 1156 | +{ | |
| 1157 | + tcg_gen_op3(INDEX_op_qemu_st8, arg, addr, mem_index); | |
| 1158 | +} | |
| 1159 | + | |
| 1160 | +static inline void tcg_gen_qemu_st16(int arg, int addr, int mem_index) | |
| 1161 | +{ | |
| 1162 | + tcg_gen_op3(INDEX_op_qemu_st16, arg, addr, mem_index); | |
| 1163 | +} | |
| 1164 | + | |
| 1165 | +static inline void tcg_gen_qemu_st32(int arg, int addr, int mem_index) | |
| 1166 | +{ | |
| 1167 | + tcg_gen_op3(INDEX_op_qemu_st32, arg, addr, mem_index); | |
| 1168 | +} | |
| 1169 | + | |
| 1170 | +static inline void tcg_gen_qemu_st64(int arg, int addr, int mem_index) | |
| 1171 | +{ | |
| 1172 | + tcg_gen_op3(INDEX_op_qemu_st64, arg, addr, mem_index); | |
| 1173 | +} | |
| 1174 | + | |
| 1175 | +#endif /* TCG_TARGET_REG_BITS != 32 */ | ... | ... |
tcg/tcg-opc.h
0 โ 100644
| 1 | +/* | |
| 2 | + * Tiny Code Generator for QEMU | |
| 3 | + * | |
| 4 | + * Copyright (c) 2008 Fabrice Bellard | |
| 5 | + * | |
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | |
| 7 | + * of this software and associated documentation files (the "Software"), to deal | |
| 8 | + * in the Software without restriction, including without limitation the rights | |
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | |
| 10 | + * copies of the Software, and to permit persons to whom the Software is | |
| 11 | + * furnished to do so, subject to the following conditions: | |
| 12 | + * | |
| 13 | + * The above copyright notice and this permission notice shall be included in | |
| 14 | + * all copies or substantial portions of the Software. | |
| 15 | + * | |
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | |
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | |
| 22 | + * THE SOFTWARE. | |
| 23 | + */ | |
| 24 | +#include "dyngen-opc.h" | |
| 25 | + | |
| 26 | +#ifndef DEF2 | |
| 27 | +#define DEF2(name, oargs, iargs, cargs, flags) DEF(name, oargs + iargs + cargs, 0) | |
| 28 | +#endif | |
| 29 | + | |
| 30 | +/* predefined ops */ | |
| 31 | +DEF2(end, 0, 0, 0, 0) /* must be kept first */ | |
| 32 | +DEF2(nop, 0, 0, 0, 0) | |
| 33 | +DEF2(nop1, 0, 0, 1, 0) | |
| 34 | +DEF2(nop2, 0, 0, 2, 0) | |
| 35 | +DEF2(nop3, 0, 0, 3, 0) | |
| 36 | +DEF2(nopn, 0, 0, 1, 0) /* variable number of parameters */ | |
| 37 | +/* macro handling */ | |
| 38 | +DEF2(macro_2, 2, 0, 1, 0) | |
| 39 | +DEF2(macro_start, 0, 0, 2, 0) | |
| 40 | +DEF2(macro_end, 0, 0, 2, 0) | |
| 41 | +DEF2(macro_goto, 0, 0, 3, 0) | |
| 42 | + | |
| 43 | +DEF2(set_label, 0, 0, 1, 0) | |
| 44 | +DEF2(call, 0, 1, 2, 0) /* variable number of parameters */ | |
| 45 | +DEF2(jmp, 0, 1, 0, TCG_OPF_BB_END) | |
| 46 | +DEF2(br, 0, 0, 1, TCG_OPF_BB_END) | |
| 47 | + | |
| 48 | +DEF2(mov_i32, 1, 1, 0, 0) | |
| 49 | +DEF2(movi_i32, 1, 0, 1, 0) | |
| 50 | +/* load/store */ | |
| 51 | +DEF2(ld8u_i32, 1, 1, 1, 0) | |
| 52 | +DEF2(ld8s_i32, 1, 1, 1, 0) | |
| 53 | +DEF2(ld16u_i32, 1, 1, 1, 0) | |
| 54 | +DEF2(ld16s_i32, 1, 1, 1, 0) | |
| 55 | +DEF2(ld_i32, 1, 1, 1, 0) | |
| 56 | +DEF2(st8_i32, 0, 2, 1, 0) | |
| 57 | +DEF2(st16_i32, 0, 2, 1, 0) | |
| 58 | +DEF2(st_i32, 0, 2, 1, 0) | |
| 59 | +/* arith */ | |
| 60 | +DEF2(add_i32, 1, 2, 0, 0) | |
| 61 | +DEF2(sub_i32, 1, 2, 0, 0) | |
| 62 | +DEF2(mul_i32, 1, 2, 0, 0) | |
| 63 | +#ifdef TCG_TARGET_HAS_div_i32 | |
| 64 | +DEF2(div_i32, 1, 2, 0, 0) | |
| 65 | +DEF2(divu_i32, 1, 2, 0, 0) | |
| 66 | +DEF2(rem_i32, 1, 2, 0, 0) | |
| 67 | +DEF2(remu_i32, 1, 2, 0, 0) | |
| 68 | +#else | |
| 69 | +DEF2(div2_i32, 2, 3, 0, 0) | |
| 70 | +DEF2(divu2_i32, 2, 3, 0, 0) | |
| 71 | +#endif | |
| 72 | +DEF2(and_i32, 1, 2, 0, 0) | |
| 73 | +DEF2(or_i32, 1, 2, 0, 0) | |
| 74 | +DEF2(xor_i32, 1, 2, 0, 0) | |
| 75 | +/* shifts */ | |
| 76 | +DEF2(shl_i32, 1, 2, 0, 0) | |
| 77 | +DEF2(shr_i32, 1, 2, 0, 0) | |
| 78 | +DEF2(sar_i32, 1, 2, 0, 0) | |
| 79 | + | |
| 80 | +DEF2(brcond_i32, 0, 2, 2, TCG_OPF_BB_END) | |
| 81 | +#if TCG_TARGET_REG_BITS == 32 | |
| 82 | +DEF2(add2_i32, 2, 4, 0, 0) | |
| 83 | +DEF2(sub2_i32, 2, 4, 0, 0) | |
| 84 | +DEF2(brcond2_i32, 0, 4, 2, TCG_OPF_BB_END) | |
| 85 | +DEF2(mulu2_i32, 2, 2, 0, 0) | |
| 86 | +#endif | |
| 87 | +#ifdef TCG_TARGET_HAS_ext8s_i32 | |
| 88 | +DEF2(ext8s_i32, 1, 1, 0, 0) | |
| 89 | +#endif | |
| 90 | +#ifdef TCG_TARGET_HAS_ext16s_i32 | |
| 91 | +DEF2(ext16s_i32, 1, 1, 0, 0) | |
| 92 | +#endif | |
| 93 | +#ifdef TCG_TARGET_HAS_bswap_i32 | |
| 94 | +DEF2(bswap_i32, 1, 1, 0, 0) | |
| 95 | +#endif | |
| 96 | + | |
| 97 | +#if TCG_TARGET_REG_BITS == 64 | |
| 98 | +DEF2(mov_i64, 1, 1, 0, 0) | |
| 99 | +DEF2(movi_i64, 1, 0, 1, 0) | |
| 100 | +/* load/store */ | |
| 101 | +DEF2(ld8u_i64, 1, 1, 1, 0) | |
| 102 | +DEF2(ld8s_i64, 1, 1, 1, 0) | |
| 103 | +DEF2(ld16u_i64, 1, 1, 1, 0) | |
| 104 | +DEF2(ld16s_i64, 1, 1, 1, 0) | |
| 105 | +DEF2(ld32u_i64, 1, 1, 1, 0) | |
| 106 | +DEF2(ld32s_i64, 1, 1, 1, 0) | |
| 107 | +DEF2(ld_i64, 1, 1, 1, 0) | |
| 108 | +DEF2(st8_i64, 0, 2, 1, 0) | |
| 109 | +DEF2(st16_i64, 0, 2, 1, 0) | |
| 110 | +DEF2(st32_i64, 0, 2, 1, 0) | |
| 111 | +DEF2(st_i64, 0, 2, 1, 0) | |
| 112 | +/* arith */ | |
| 113 | +DEF2(add_i64, 1, 2, 0, 0) | |
| 114 | +DEF2(sub_i64, 1, 2, 0, 0) | |
| 115 | +DEF2(mul_i64, 1, 2, 0, 0) | |
| 116 | +#ifdef TCG_TARGET_HAS_div_i64 | |
| 117 | +DEF2(div_i64, 1, 2, 0, 0) | |
| 118 | +DEF2(divu_i64, 1, 2, 0, 0) | |
| 119 | +DEF2(rem_i64, 1, 2, 0, 0) | |
| 120 | +DEF2(remu_i64, 1, 2, 0, 0) | |
| 121 | +#else | |
| 122 | +DEF2(div2_i64, 2, 3, 0, 0) | |
| 123 | +DEF2(divu2_i64, 2, 3, 0, 0) | |
| 124 | +#endif | |
| 125 | +DEF2(and_i64, 1, 2, 0, 0) | |
| 126 | +DEF2(or_i64, 1, 2, 0, 0) | |
| 127 | +DEF2(xor_i64, 1, 2, 0, 0) | |
| 128 | +/* shifts */ | |
| 129 | +DEF2(shl_i64, 1, 2, 0, 0) | |
| 130 | +DEF2(shr_i64, 1, 2, 0, 0) | |
| 131 | +DEF2(sar_i64, 1, 2, 0, 0) | |
| 132 | + | |
| 133 | +DEF2(brcond_i64, 0, 2, 2, TCG_OPF_BB_END) | |
| 134 | +#ifdef TCG_TARGET_HAS_ext8s_i64 | |
| 135 | +DEF2(ext8s_i64, 1, 1, 0, 0) | |
| 136 | +#endif | |
| 137 | +#ifdef TCG_TARGET_HAS_ext16s_i64 | |
| 138 | +DEF2(ext16s_i64, 1, 1, 0, 0) | |
| 139 | +#endif | |
| 140 | +#ifdef TCG_TARGET_HAS_ext32s_i64 | |
| 141 | +DEF2(ext32s_i64, 1, 1, 0, 0) | |
| 142 | +#endif | |
| 143 | +#ifdef TCG_TARGET_HAS_bswap_i64 | |
| 144 | +DEF2(bswap_i64, 1, 1, 0, 0) | |
| 145 | +#endif | |
| 146 | +#endif | |
| 147 | + | |
| 148 | +/* QEMU specific */ | |
| 149 | +DEF2(exit_tb, 0, 0, 1, TCG_OPF_BB_END) | |
| 150 | +DEF2(goto_tb, 0, 0, 1, TCG_OPF_BB_END) | |
| 151 | +/* Note: even if TARGET_LONG_BITS is not defined, the INDEX_op | |
| 152 | + constants must be defined */ | |
| 153 | +#if TCG_TARGET_REG_BITS == 32 | |
| 154 | +#if TARGET_LONG_BITS == 32 | |
| 155 | +DEF2(qemu_ld8u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 156 | +#else | |
| 157 | +DEF2(qemu_ld8u, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 158 | +#endif | |
| 159 | +#if TARGET_LONG_BITS == 32 | |
| 160 | +DEF2(qemu_ld8s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 161 | +#else | |
| 162 | +DEF2(qemu_ld8s, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 163 | +#endif | |
| 164 | +#if TARGET_LONG_BITS == 32 | |
| 165 | +DEF2(qemu_ld16u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 166 | +#else | |
| 167 | +DEF2(qemu_ld16u, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 168 | +#endif | |
| 169 | +#if TARGET_LONG_BITS == 32 | |
| 170 | +DEF2(qemu_ld16s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 171 | +#else | |
| 172 | +DEF2(qemu_ld16s, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 173 | +#endif | |
| 174 | +#if TARGET_LONG_BITS == 32 | |
| 175 | +DEF2(qemu_ld32u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 176 | +#else | |
| 177 | +DEF2(qemu_ld32u, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 178 | +#endif | |
| 179 | +#if TARGET_LONG_BITS == 32 | |
| 180 | +DEF2(qemu_ld32s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 181 | +#else | |
| 182 | +DEF2(qemu_ld32s, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 183 | +#endif | |
| 184 | +#if TARGET_LONG_BITS == 32 | |
| 185 | +DEF2(qemu_ld64, 2, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 186 | +#else | |
| 187 | +DEF2(qemu_ld64, 2, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 188 | +#endif | |
| 189 | + | |
| 190 | +#if TARGET_LONG_BITS == 32 | |
| 191 | +DEF2(qemu_st8, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 192 | +#else | |
| 193 | +DEF2(qemu_st8, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | |
| 194 | +#endif | |
| 195 | +#if TARGET_LONG_BITS == 32 | |
| 196 | +DEF2(qemu_st16, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 197 | +#else | |
| 198 | +DEF2(qemu_st16, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | |
| 199 | +#endif | |
| 200 | +#if TARGET_LONG_BITS == 32 | |
| 201 | +DEF2(qemu_st32, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 202 | +#else | |
| 203 | +DEF2(qemu_st32, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | |
| 204 | +#endif | |
| 205 | +#if TARGET_LONG_BITS == 32 | |
| 206 | +DEF2(qemu_st64, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | |
| 207 | +#else | |
| 208 | +DEF2(qemu_st64, 0, 4, 1, TCG_OPF_CALL_CLOBBER) | |
| 209 | +#endif | |
| 210 | + | |
| 211 | +#else /* TCG_TARGET_REG_BITS == 32 */ | |
| 212 | + | |
| 213 | +DEF2(qemu_ld8u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 214 | +DEF2(qemu_ld8s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 215 | +DEF2(qemu_ld16u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 216 | +DEF2(qemu_ld16s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 217 | +DEF2(qemu_ld32u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 218 | +DEF2(qemu_ld32s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 219 | +DEF2(qemu_ld64, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | |
| 220 | + | |
| 221 | +DEF2(qemu_st8, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 222 | +DEF2(qemu_st16, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 223 | +DEF2(qemu_st32, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 224 | +DEF2(qemu_st64, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | |
| 225 | + | |
| 226 | +#endif /* TCG_TARGET_REG_BITS != 32 */ | |
| 227 | + | |
| 228 | +#undef DEF2 | ... | ... |
tcg/tcg-runtime.c
0 โ 100644
| 1 | +/* | |
| 2 | + * Tiny Code Generator for QEMU | |
| 3 | + * | |
| 4 | + * Copyright (c) 2008 Fabrice Bellard | |
| 5 | + * | |
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | |
| 7 | + * of this software and associated documentation files (the "Software"), to deal | |
| 8 | + * in the Software without restriction, including without limitation the rights | |
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | |
| 10 | + * copies of the Software, and to permit persons to whom the Software is | |
| 11 | + * furnished to do so, subject to the following conditions: | |
| 12 | + * | |
| 13 | + * The above copyright notice and this permission notice shall be included in | |
| 14 | + * all copies or substantial portions of the Software. | |
| 15 | + * | |
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | |
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | |
| 22 | + * THE SOFTWARE. | |
| 23 | + */ | |
| 24 | +#include <stdarg.h> | |
| 25 | +#include <stdlib.h> | |
| 26 | +#include <stdio.h> | |
| 27 | +#include <string.h> | |
| 28 | +#include <inttypes.h> | |
| 29 | + | |
| 30 | +#include "config.h" | |
| 31 | +#include "osdep.h" | |
| 32 | +#include "tcg.h" | |
| 33 | + | |
| 34 | +int64_t tcg_helper_shl_i64(int64_t arg1, int64_t arg2) | |
| 35 | +{ | |
| 36 | + return arg1 << arg2; | |
| 37 | +} | |
| 38 | + | |
| 39 | +int64_t tcg_helper_shr_i64(int64_t arg1, int64_t arg2) | |
| 40 | +{ | |
| 41 | + return (uint64_t)arg1 >> arg2; | |
| 42 | +} | |
| 43 | + | |
| 44 | +int64_t tcg_helper_sar_i64(int64_t arg1, int64_t arg2) | |
| 45 | +{ | |
| 46 | + return arg1 >> arg2; | |
| 47 | +} | |
| 48 | + | |
| 49 | +int64_t tcg_helper_div_i64(int64_t arg1, int64_t arg2) | |
| 50 | +{ | |
| 51 | + return arg1 / arg2; | |
| 52 | +} | |
| 53 | + | |
| 54 | +int64_t tcg_helper_rem_i64(int64_t arg1, int64_t arg2) | |
| 55 | +{ | |
| 56 | + return arg1 / arg2; | |
| 57 | +} | |
| 58 | + | |
| 59 | +uint64_t tcg_helper_divu_i64(uint64_t arg1, uint64_t arg2) | |
| 60 | +{ | |
| 61 | + return arg1 / arg2; | |
| 62 | +} | |
| 63 | + | |
| 64 | +uint64_t tcg_helper_remu_i64(uint64_t arg1, uint64_t arg2) | |
| 65 | +{ | |
| 66 | + return arg1 / arg2; | |
| 67 | +} | |
| 68 | + | ... | ... |