Commit c896fe29d6c8ae6cde3917727812ced3f2e536a4
1 parent
56abbcff
TCG code generator
git-svn-id: svn://svn.savannah.nongnu.org/qemu/trunk@3943 c046a42c-6fe2-441c-8c8c-71466251a162
Showing
9 changed files
with
3625 additions
and
0 deletions
Too many changes to show.
To preserve performance only 9 of 13 files are displayed.
tcg/LICENSE
0 โ 100644
tcg/README
0 โ 100644
| 1 | +Tiny Code Generator - Fabrice Bellard. | ||
| 2 | + | ||
| 3 | +1) Introduction | ||
| 4 | + | ||
| 5 | +TCG (Tiny Code Generator) began as a generic backend for a C | ||
| 6 | +compiler. It was simplified to be used in QEMU. It also has its roots | ||
| 7 | +in the QOP code generator written by Paul Brook. | ||
| 8 | + | ||
| 9 | +2) Definitions | ||
| 10 | + | ||
| 11 | +The TCG "target" is the architecture for which we generate the | ||
| 12 | +code. It is of course not the same as the "target" of QEMU which is | ||
| 13 | +the emulated architecture. As TCG started as a generic C backend used | ||
| 14 | +for cross compiling, it is assumed that the TCG target is different | ||
| 15 | +from the host, although it is never the case for QEMU. | ||
| 16 | + | ||
| 17 | +A TCG "function" corresponds to a QEMU Translated Block (TB). | ||
| 18 | + | ||
| 19 | +A TCG "temporary" is a variable only live in a given | ||
| 20 | +function. Temporaries are allocated explicitely in each function. | ||
| 21 | + | ||
| 22 | +A TCG "global" is a variable which is live in all the functions. They | ||
| 23 | +are defined before the functions defined. A TCG global can be a memory | ||
| 24 | +location (e.g. a QEMU CPU register), a fixed host register (e.g. the | ||
| 25 | +QEMU CPU state pointer) or a memory location which is stored in a | ||
| 26 | +register outside QEMU TBs (not implemented yet). | ||
| 27 | + | ||
| 28 | +A TCG "basic block" corresponds to a list of instructions terminated | ||
| 29 | +by a branch instruction. | ||
| 30 | + | ||
| 31 | +3) Intermediate representation | ||
| 32 | + | ||
| 33 | +3.1) Introduction | ||
| 34 | + | ||
| 35 | +TCG instructions operate on variables which are temporaries or | ||
| 36 | +globals. TCG instructions and variables are strongly typed. Two types | ||
| 37 | +are supported: 32 bit integers and 64 bit integers. Pointers are | ||
| 38 | +defined as an alias to 32 bit or 64 bit integers depending on the TCG | ||
| 39 | +target word size. | ||
| 40 | + | ||
| 41 | +Each instruction has a fixed number of output variable operands, input | ||
| 42 | +variable operands and always constant operands. | ||
| 43 | + | ||
| 44 | +The notable exception is the call instruction which has a variable | ||
| 45 | +number of outputs and inputs. | ||
| 46 | + | ||
| 47 | +In the textual form, output operands come first, followed by input | ||
| 48 | +operands, followed by constant operands. The output type is included | ||
| 49 | +in the instruction name. Constants are prefixed with a '$'. | ||
| 50 | + | ||
| 51 | +add_i32 t0, t1, t2 (t0 <- t1 + t2) | ||
| 52 | + | ||
| 53 | +sub_i64 t2, t3, $4 (t2 <- t3 - 4) | ||
| 54 | + | ||
| 55 | +3.2) Assumptions | ||
| 56 | + | ||
| 57 | +* Basic blocks | ||
| 58 | + | ||
| 59 | +- Basic blocks end after branches (e.g. brcond_i32 instruction), | ||
| 60 | + goto_tb and exit_tb instructions. | ||
| 61 | +- Basic blocks end before legacy dyngen operations. | ||
| 62 | +- Basic blocks start after the end of a previous basic block, at a | ||
| 63 | + set_label instruction or after a legacy dyngen operation. | ||
| 64 | + | ||
| 65 | +After the end of a basic block, temporaries at destroyed and globals | ||
| 66 | +are stored at their initial storage (register or memory place | ||
| 67 | +depending on their declarations). | ||
| 68 | + | ||
| 69 | +* Floating point types are not supported yet | ||
| 70 | + | ||
| 71 | +* Pointers: depending on the TCG target, pointer size is 32 bit or 64 | ||
| 72 | + bit. The type TCG_TYPE_PTR is an alias to TCG_TYPE_I32 or | ||
| 73 | + TCG_TYPE_I64. | ||
| 74 | + | ||
| 75 | +* Helpers: | ||
| 76 | + | ||
| 77 | +Using the tcg_gen_helper_x_y it is possible to call any function | ||
| 78 | +taking i32, i64 or pointer types types. Before calling an helper, all | ||
| 79 | +globals are stored at their canonical location and it is assumed that | ||
| 80 | +the function can modify them. In the future, function modifiers will | ||
| 81 | +be allowed to tell that the helper does not read or write some globals. | ||
| 82 | + | ||
| 83 | +On some TCG targets (e.g. x86), several calling conventions are | ||
| 84 | +supported. | ||
| 85 | + | ||
| 86 | +* Branches: | ||
| 87 | + | ||
| 88 | +Use the instruction 'br' to jump to a label. Use 'jmp' to jump to an | ||
| 89 | +explicit address. Conditional branches can only jump to labels. | ||
| 90 | + | ||
| 91 | +3.3) Code Optimizations | ||
| 92 | + | ||
| 93 | +When generating instructions, you can count on at least the following | ||
| 94 | +optimizations: | ||
| 95 | + | ||
| 96 | +- Single instructions are simplified, e.g. | ||
| 97 | + | ||
| 98 | + and_i32 t0, t0, $0xffffffff | ||
| 99 | + | ||
| 100 | + is suppressed. | ||
| 101 | + | ||
| 102 | +- A liveness analysis is done at the basic block level. The | ||
| 103 | + information is used to suppress moves from a dead temporary to | ||
| 104 | + another one. It is also used to remove instructions which compute | ||
| 105 | + dead results. The later is especially useful for condition code | ||
| 106 | + optimisation in QEMU. | ||
| 107 | + | ||
| 108 | + In the following example: | ||
| 109 | + | ||
| 110 | + add_i32 t0, t1, t2 | ||
| 111 | + add_i32 t0, t0, $1 | ||
| 112 | + mov_i32 t0, $1 | ||
| 113 | + | ||
| 114 | + only the last instruction is kept. | ||
| 115 | + | ||
| 116 | +- A macro system is supported (may get closer to function inlining | ||
| 117 | + some day). It is useful if the liveness analysis is likely to prove | ||
| 118 | + that some results of a computation are indeed not useful. With the | ||
| 119 | + macro system, the user can provide several alternative | ||
| 120 | + implementations which are used depending on the used results. It is | ||
| 121 | + especially useful for condition code optimisation in QEMU. | ||
| 122 | + | ||
| 123 | + Here is an example: | ||
| 124 | + | ||
| 125 | + macro_2 t0, t1, $1 | ||
| 126 | + mov_i32 t0, $0x1234 | ||
| 127 | + | ||
| 128 | + The macro identified by the ID "$1" normally returns the values t0 | ||
| 129 | + and t1. Suppose its implementation is: | ||
| 130 | + | ||
| 131 | + macro_start | ||
| 132 | + brcond_i32 t2, $0, $TCG_COND_EQ, $1 | ||
| 133 | + mov_i32 t0, $2 | ||
| 134 | + br $2 | ||
| 135 | + set_label $1 | ||
| 136 | + mov_i32 t0, $3 | ||
| 137 | + set_label $2 | ||
| 138 | + add_i32 t1, t3, t4 | ||
| 139 | + macro_end | ||
| 140 | + | ||
| 141 | + If t0 is not used after the macro, the user can provide a simpler | ||
| 142 | + implementation: | ||
| 143 | + | ||
| 144 | + macro_start | ||
| 145 | + add_i32 t1, t2, t4 | ||
| 146 | + macro_end | ||
| 147 | + | ||
| 148 | + TCG automatically chooses the right implementation depending on | ||
| 149 | + which macro outputs are used after it. | ||
| 150 | + | ||
| 151 | + Note that if TCG did more expensive optimizations, macros would be | ||
| 152 | + less useful. In the previous example a macro is useful because the | ||
| 153 | + liveness analysis is done on each basic block separately. Hence TCG | ||
| 154 | + cannot remove the code computing 't0' even if it is not used after | ||
| 155 | + the first macro implementation. | ||
| 156 | + | ||
| 157 | +3.4) Instruction Reference | ||
| 158 | + | ||
| 159 | +********* Function call | ||
| 160 | + | ||
| 161 | +* call <ret> <params> ptr | ||
| 162 | + | ||
| 163 | +call function 'ptr' (pointer type) | ||
| 164 | + | ||
| 165 | +<ret> optional 32 bit or 64 bit return value | ||
| 166 | +<params> optional 32 bit or 64 bit parameters | ||
| 167 | + | ||
| 168 | +********* Jumps/Labels | ||
| 169 | + | ||
| 170 | +* jmp t0 | ||
| 171 | + | ||
| 172 | +Absolute jump to address t0 (pointer type). | ||
| 173 | + | ||
| 174 | +* set_label $label | ||
| 175 | + | ||
| 176 | +Define label 'label' at the current program point. | ||
| 177 | + | ||
| 178 | +* br $label | ||
| 179 | + | ||
| 180 | +Jump to label. | ||
| 181 | + | ||
| 182 | +* brcond_i32/i64 cond, t0, t1, label | ||
| 183 | + | ||
| 184 | +Conditional jump if t0 cond t1 is true. cond can be: | ||
| 185 | + TCG_COND_EQ | ||
| 186 | + TCG_COND_NE | ||
| 187 | + TCG_COND_LT /* signed */ | ||
| 188 | + TCG_COND_GE /* signed */ | ||
| 189 | + TCG_COND_LE /* signed */ | ||
| 190 | + TCG_COND_GT /* signed */ | ||
| 191 | + TCG_COND_LTU /* unsigned */ | ||
| 192 | + TCG_COND_GEU /* unsigned */ | ||
| 193 | + TCG_COND_LEU /* unsigned */ | ||
| 194 | + TCG_COND_GTU /* unsigned */ | ||
| 195 | + | ||
| 196 | +********* Arithmetic | ||
| 197 | + | ||
| 198 | +* add_i32/i64 t0, t1, t2 | ||
| 199 | + | ||
| 200 | +t0=t1+t2 | ||
| 201 | + | ||
| 202 | +* sub_i32/i64 t0, t1, t2 | ||
| 203 | + | ||
| 204 | +t0=t1-t2 | ||
| 205 | + | ||
| 206 | +* mul_i32/i64 t0, t1, t2 | ||
| 207 | + | ||
| 208 | +t0=t1*t2 | ||
| 209 | + | ||
| 210 | +* div_i32/i64 t0, t1, t2 | ||
| 211 | + | ||
| 212 | +t0=t1/t2 (signed). Undefined behavior if division by zero or overflow. | ||
| 213 | + | ||
| 214 | +* divu_i32/i64 t0, t1, t2 | ||
| 215 | + | ||
| 216 | +t0=t1/t2 (unsigned). Undefined behavior if division by zero. | ||
| 217 | + | ||
| 218 | +* rem_i32/i64 t0, t1, t2 | ||
| 219 | + | ||
| 220 | +t0=t1%t2 (signed). Undefined behavior if division by zero or overflow. | ||
| 221 | + | ||
| 222 | +* remu_i32/i64 t0, t1, t2 | ||
| 223 | + | ||
| 224 | +t0=t1%t2 (unsigned). Undefined behavior if division by zero. | ||
| 225 | + | ||
| 226 | +* and_i32/i64 t0, t1, t2 | ||
| 227 | + | ||
| 228 | +********* Logical | ||
| 229 | + | ||
| 230 | +t0=t1&t2 | ||
| 231 | + | ||
| 232 | +* or_i32/i64 t0, t1, t2 | ||
| 233 | + | ||
| 234 | +t0=t1|t2 | ||
| 235 | + | ||
| 236 | +* xor_i32/i64 t0, t1, t2 | ||
| 237 | + | ||
| 238 | +t0=t1^t2 | ||
| 239 | + | ||
| 240 | +* shl_i32/i64 t0, t1, t2 | ||
| 241 | + | ||
| 242 | +********* Shifts | ||
| 243 | + | ||
| 244 | +* shl_i32/i64 t0, t1, t2 | ||
| 245 | + | ||
| 246 | +t0=t1 << t2. Undefined behavior if t2 < 0 or t2 >= 32 (resp 64) | ||
| 247 | + | ||
| 248 | +* shr_i32/i64 t0, t1, t2 | ||
| 249 | + | ||
| 250 | +t0=t1 >> t2 (unsigned). Undefined behavior if t2 < 0 or t2 >= 32 (resp 64) | ||
| 251 | + | ||
| 252 | +* sar_i32/i64 t0, t1, t2 | ||
| 253 | + | ||
| 254 | +t0=t1 >> t2 (signed). Undefined behavior if t2 < 0 or t2 >= 32 (resp 64) | ||
| 255 | + | ||
| 256 | +********* Misc | ||
| 257 | + | ||
| 258 | +* mov_i32/i64 t0, t1 | ||
| 259 | + | ||
| 260 | +t0 = t1 | ||
| 261 | + | ||
| 262 | +Move t1 to t0 (both operands must have the same type). | ||
| 263 | + | ||
| 264 | +* ext8s_i32/i64 t0, t1 | ||
| 265 | +ext16s_i32/i64 t0, t1 | ||
| 266 | +ext32s_i64 t0, t1 | ||
| 267 | + | ||
| 268 | +8, 16 or 32 bit sign extension (both operands must have the same type) | ||
| 269 | + | ||
| 270 | +* bswap16_i32 t0, t1 | ||
| 271 | + | ||
| 272 | +16 bit byte swap on a 32 bit value. The two high order bytes must be set | ||
| 273 | +to zero. | ||
| 274 | + | ||
| 275 | +* bswap_i32 t0, t1 | ||
| 276 | + | ||
| 277 | +32 bit byte swap | ||
| 278 | + | ||
| 279 | +* bswap_i64 t0, t1 | ||
| 280 | + | ||
| 281 | +64 bit byte swap | ||
| 282 | + | ||
| 283 | +********* Type conversions | ||
| 284 | + | ||
| 285 | +* ext_i32_i64 t0, t1 | ||
| 286 | +Convert t1 (32 bit) to t0 (64 bit) and does sign extension | ||
| 287 | + | ||
| 288 | +* extu_i32_i64 t0, t1 | ||
| 289 | +Convert t1 (32 bit) to t0 (64 bit) and does zero extension | ||
| 290 | + | ||
| 291 | +* trunc_i64_i32 t0, t1 | ||
| 292 | +Truncate t1 (64 bit) to t0 (32 bit) | ||
| 293 | + | ||
| 294 | +********* Load/Store | ||
| 295 | + | ||
| 296 | +* ld_i32/i64 t0, t1, offset | ||
| 297 | +ld8s_i32/i64 t0, t1, offset | ||
| 298 | +ld8u_i32/i64 t0, t1, offset | ||
| 299 | +ld16s_i32/i64 t0, t1, offset | ||
| 300 | +ld16u_i32/i64 t0, t1, offset | ||
| 301 | +ld32s_i64 t0, t1, offset | ||
| 302 | +ld32u_i64 t0, t1, offset | ||
| 303 | + | ||
| 304 | +t0 = read(t1 + offset) | ||
| 305 | +Load 8, 16, 32 or 64 bits with or without sign extension from host memory. | ||
| 306 | +offset must be a constant. | ||
| 307 | + | ||
| 308 | +* st_i32/i64 t0, t1, offset | ||
| 309 | +st8_i32/i64 t0, t1, offset | ||
| 310 | +st16_i32/i64 t0, t1, offset | ||
| 311 | +st32_i64 t0, t1, offset | ||
| 312 | + | ||
| 313 | +write(t0, t1 + offset) | ||
| 314 | +Write 8, 16, 32 or 64 bits to host memory. | ||
| 315 | + | ||
| 316 | +********* QEMU specific operations | ||
| 317 | + | ||
| 318 | +* tb_exit t0 | ||
| 319 | + | ||
| 320 | +Exit the current TB and return the value t0 (word type). | ||
| 321 | + | ||
| 322 | +* goto_tb index | ||
| 323 | + | ||
| 324 | +Exit the current TB and jump to the TB index 'index' (constant) if the | ||
| 325 | +current TB was linked to this TB. Otherwise execute the next | ||
| 326 | +instructions. | ||
| 327 | + | ||
| 328 | +* qemu_ld_i32/i64 t0, t1, flags | ||
| 329 | +qemu_ld8u_i32/i64 t0, t1, flags | ||
| 330 | +qemu_ld8s_i32/i64 t0, t1, flags | ||
| 331 | +qemu_ld16u_i32/i64 t0, t1, flags | ||
| 332 | +qemu_ld16s_i32/i64 t0, t1, flags | ||
| 333 | +qemu_ld32u_i64 t0, t1, flags | ||
| 334 | +qemu_ld32s_i64 t0, t1, flags | ||
| 335 | + | ||
| 336 | +Load data at the QEMU CPU address t1 into t0. t1 has the QEMU CPU | ||
| 337 | +address type. 'flags' contains the QEMU memory index (selects user or | ||
| 338 | +kernel access) for example. | ||
| 339 | + | ||
| 340 | +* qemu_st_i32/i64 t0, t1, flags | ||
| 341 | +qemu_st8_i32/i64 t0, t1, flags | ||
| 342 | +qemu_st16_i32/i64 t0, t1, flags | ||
| 343 | +qemu_st32_i64 t0, t1, flags | ||
| 344 | + | ||
| 345 | +Store the data t0 at the QEMU CPU Address t1. t1 has the QEMU CPU | ||
| 346 | +address type. 'flags' contains the QEMU memory index (selects user or | ||
| 347 | +kernel access) for example. | ||
| 348 | + | ||
| 349 | +Note 1: Some shortcuts are defined when the last operand is known to be | ||
| 350 | +a constant (e.g. addi for add, movi for mov). | ||
| 351 | + | ||
| 352 | +Note 2: When using TCG, the opcodes must never be generated directly | ||
| 353 | +as some of them may not be available as "real" opcodes. Always use the | ||
| 354 | +function tcg_gen_xxx(args). | ||
| 355 | + | ||
| 356 | +4) Backend | ||
| 357 | + | ||
| 358 | +tcg-target.h contains the target specific definitions. tcg-target.c | ||
| 359 | +contains the target specific code. | ||
| 360 | + | ||
| 361 | +4.1) Assumptions | ||
| 362 | + | ||
| 363 | +The target word size (TCG_TARGET_REG_BITS) is expected to be 32 bit or | ||
| 364 | +64 bit. It is expected that the pointer has the same size as the word. | ||
| 365 | + | ||
| 366 | +On a 32 bit target, all 64 bit operations are converted to 32 bits. A | ||
| 367 | +few specific operations must be implemented to allow it (see add2_i32, | ||
| 368 | +sub2_i32, brcond2_i32). | ||
| 369 | + | ||
| 370 | +Floating point operations are not supported in this version. A | ||
| 371 | +previous incarnation of the code generator had full support of them, | ||
| 372 | +but it is better to concentrate on integer operations first. | ||
| 373 | + | ||
| 374 | +On a 64 bit target, no assumption is made in TCG about the storage of | ||
| 375 | +the 32 bit values in 64 bit registers. | ||
| 376 | + | ||
| 377 | +4.2) Constraints | ||
| 378 | + | ||
| 379 | +GCC like constraints are used to define the constraints of every | ||
| 380 | +instruction. Memory constraints are not supported in this | ||
| 381 | +version. Aliases are specified in the input operands as for GCC. | ||
| 382 | + | ||
| 383 | +A target can define specific register or constant constraints. If an | ||
| 384 | +operation uses a constant input constraint which does not allow all | ||
| 385 | +constants, it must also accept registers in order to have a fallback. | ||
| 386 | + | ||
| 387 | +The movi_i32 and movi_i64 operations must accept any constants. | ||
| 388 | + | ||
| 389 | +The mov_i32 and mov_i64 operations must accept any registers of the | ||
| 390 | +same type. | ||
| 391 | + | ||
| 392 | +The ld/st instructions must accept signed 32 bit constant offsets. It | ||
| 393 | +can be implemented by reserving a specific register to compute the | ||
| 394 | +address if the offset is too big. | ||
| 395 | + | ||
| 396 | +The ld/st instructions must accept any destination (ld) or source (st) | ||
| 397 | +register. | ||
| 398 | + | ||
| 399 | +4.3) Function call assumptions | ||
| 400 | + | ||
| 401 | +- The only supported types for parameters and return value are: 32 and | ||
| 402 | + 64 bit integers and pointer. | ||
| 403 | +- The stack grows downwards. | ||
| 404 | +- The first N parameters are passed in registers. | ||
| 405 | +- The next parameters are passed on the stack by storing them as words. | ||
| 406 | +- Some registers are clobbered during the call. | ||
| 407 | +- The function can return 0 or 1 value in registers. On a 32 bit | ||
| 408 | + target, functions must be able to return 2 values in registers for | ||
| 409 | + 64 bit return type. | ||
| 410 | + | ||
| 411 | +5) Migration from dyngen to TCG | ||
| 412 | + | ||
| 413 | +TCG is backward compatible with QEMU "dyngen" operations. It means | ||
| 414 | +that TCG instructions can be freely mixed with dyngen operations. It | ||
| 415 | +is expected that QEMU targets will be progressively fully converted to | ||
| 416 | +TCG. Once a target is fully converted to dyngen, it will be possible | ||
| 417 | +to apply more optimizations because more registers will be free for | ||
| 418 | +the generated code. | ||
| 419 | + | ||
| 420 | +The exception model is the same as the dyngen one. |
tcg/TODO
0 โ 100644
| 1 | +- test macro system | ||
| 2 | + | ||
| 3 | +- test conditional jumps | ||
| 4 | + | ||
| 5 | +- test mul, div, ext8s, ext16s, bswap | ||
| 6 | + | ||
| 7 | +- generate a global TB prologue and epilogue to save/restore registers | ||
| 8 | + to/from the CPU state and to reserve a stack frame to optimize | ||
| 9 | + helper calls. Modify cpu-exec.c so that it does not use global | ||
| 10 | + register variables (except maybe for 'env'). | ||
| 11 | + | ||
| 12 | +- fully convert the x86 target. The minimal amount of work includes: | ||
| 13 | + - add cc_src, cc_dst and cc_op as globals | ||
| 14 | + - disable its eflags optimization (the liveness analysis should | ||
| 15 | + suffice) | ||
| 16 | + - move complicated operations to helpers (in particular FPU, SSE, MMX). | ||
| 17 | + | ||
| 18 | +- optimize the x86 target: | ||
| 19 | + - move some or all the registers as globals | ||
| 20 | + - use the TB prologue and epilogue to have QEMU target registers in | ||
| 21 | + pre assigned host registers. | ||
| 22 | + | ||
| 23 | +Ideas: | ||
| 24 | + | ||
| 25 | +- Move the slow part of the qemu_ld/st ops after the end of the TB. | ||
| 26 | + | ||
| 27 | +- Experiment: change instruction storage to simplify macro handling | ||
| 28 | + and to handle dynamic allocation and see if the translation speed is | ||
| 29 | + OK. | ||
| 30 | + | ||
| 31 | +- change exception syntax to get closer to QOP system (exception | ||
| 32 | + parameters given with a specific instruction). |
tcg/i386/tcg-target.c
0 โ 100644
| 1 | +/* | ||
| 2 | + * Tiny Code Generator for QEMU | ||
| 3 | + * | ||
| 4 | + * Copyright (c) 2008 Fabrice Bellard | ||
| 5 | + * | ||
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
| 7 | + * of this software and associated documentation files (the "Software"), to deal | ||
| 8 | + * in the Software without restriction, including without limitation the rights | ||
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
| 10 | + * copies of the Software, and to permit persons to whom the Software is | ||
| 11 | + * furnished to do so, subject to the following conditions: | ||
| 12 | + * | ||
| 13 | + * The above copyright notice and this permission notice shall be included in | ||
| 14 | + * all copies or substantial portions of the Software. | ||
| 15 | + * | ||
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
| 22 | + * THE SOFTWARE. | ||
| 23 | + */ | ||
| 24 | +const char *tcg_target_reg_names[TCG_TARGET_NB_REGS] = { | ||
| 25 | + "%eax", | ||
| 26 | + "%ecx", | ||
| 27 | + "%edx", | ||
| 28 | + "%ebx", | ||
| 29 | + "%esp", | ||
| 30 | + "%ebp", | ||
| 31 | + "%esi", | ||
| 32 | + "%edi", | ||
| 33 | +}; | ||
| 34 | + | ||
| 35 | +int tcg_target_reg_alloc_order[TCG_TARGET_NB_REGS] = { | ||
| 36 | + TCG_REG_EAX, | ||
| 37 | + TCG_REG_EDX, | ||
| 38 | + TCG_REG_ECX, | ||
| 39 | + TCG_REG_EBX, | ||
| 40 | + TCG_REG_ESI, | ||
| 41 | + TCG_REG_EDI, | ||
| 42 | + TCG_REG_EBP, | ||
| 43 | + TCG_REG_ESP, | ||
| 44 | +}; | ||
| 45 | + | ||
| 46 | +const int tcg_target_call_iarg_regs[3] = { TCG_REG_EAX, TCG_REG_EDX, TCG_REG_ECX }; | ||
| 47 | +const int tcg_target_call_oarg_regs[2] = { TCG_REG_EAX, TCG_REG_EDX }; | ||
| 48 | + | ||
| 49 | +static void patch_reloc(uint8_t *code_ptr, int type, | ||
| 50 | + tcg_target_long value) | ||
| 51 | +{ | ||
| 52 | + switch(type) { | ||
| 53 | + case R_386_32: | ||
| 54 | + *(uint32_t *)code_ptr = value; | ||
| 55 | + break; | ||
| 56 | + case R_386_PC32: | ||
| 57 | + *(uint32_t *)code_ptr = value - (long)code_ptr; | ||
| 58 | + break; | ||
| 59 | + default: | ||
| 60 | + tcg_abort(); | ||
| 61 | + } | ||
| 62 | +} | ||
| 63 | + | ||
| 64 | +/* maximum number of register used for input function arguments */ | ||
| 65 | +static inline int tcg_target_get_call_iarg_regs_count(int flags) | ||
| 66 | +{ | ||
| 67 | + flags &= TCG_CALL_TYPE_MASK; | ||
| 68 | + switch(flags) { | ||
| 69 | + case TCG_CALL_TYPE_STD: | ||
| 70 | + return 0; | ||
| 71 | + case TCG_CALL_TYPE_REGPARM_1: | ||
| 72 | + case TCG_CALL_TYPE_REGPARM_2: | ||
| 73 | + case TCG_CALL_TYPE_REGPARM: | ||
| 74 | + return flags - TCG_CALL_TYPE_REGPARM_1 + 1; | ||
| 75 | + default: | ||
| 76 | + tcg_abort(); | ||
| 77 | + } | ||
| 78 | +} | ||
| 79 | + | ||
| 80 | +/* parse target specific constraints */ | ||
| 81 | +int target_parse_constraint(TCGArgConstraint *ct, const char **pct_str) | ||
| 82 | +{ | ||
| 83 | + const char *ct_str; | ||
| 84 | + | ||
| 85 | + ct_str = *pct_str; | ||
| 86 | + switch(ct_str[0]) { | ||
| 87 | + case 'a': | ||
| 88 | + ct->ct |= TCG_CT_REG; | ||
| 89 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EAX); | ||
| 90 | + break; | ||
| 91 | + case 'b': | ||
| 92 | + ct->ct |= TCG_CT_REG; | ||
| 93 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EBX); | ||
| 94 | + break; | ||
| 95 | + case 'c': | ||
| 96 | + ct->ct |= TCG_CT_REG; | ||
| 97 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_ECX); | ||
| 98 | + break; | ||
| 99 | + case 'd': | ||
| 100 | + ct->ct |= TCG_CT_REG; | ||
| 101 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EDX); | ||
| 102 | + break; | ||
| 103 | + case 'S': | ||
| 104 | + ct->ct |= TCG_CT_REG; | ||
| 105 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_ESI); | ||
| 106 | + break; | ||
| 107 | + case 'D': | ||
| 108 | + ct->ct |= TCG_CT_REG; | ||
| 109 | + tcg_regset_set_reg(ct->u.regs, TCG_REG_EDI); | ||
| 110 | + break; | ||
| 111 | + case 'q': | ||
| 112 | + ct->ct |= TCG_CT_REG; | ||
| 113 | + tcg_regset_set32(ct->u.regs, 0, 0xf); | ||
| 114 | + break; | ||
| 115 | + case 'r': | ||
| 116 | + ct->ct |= TCG_CT_REG; | ||
| 117 | + tcg_regset_set32(ct->u.regs, 0, 0xff); | ||
| 118 | + break; | ||
| 119 | + | ||
| 120 | + /* qemu_ld/st address constraint */ | ||
| 121 | + case 'L': | ||
| 122 | + ct->ct |= TCG_CT_REG; | ||
| 123 | + tcg_regset_set32(ct->u.regs, 0, 0xff); | ||
| 124 | + tcg_regset_reset_reg(ct->u.regs, TCG_REG_EAX); | ||
| 125 | + tcg_regset_reset_reg(ct->u.regs, TCG_REG_EDX); | ||
| 126 | + break; | ||
| 127 | + default: | ||
| 128 | + return -1; | ||
| 129 | + } | ||
| 130 | + ct_str++; | ||
| 131 | + *pct_str = ct_str; | ||
| 132 | + return 0; | ||
| 133 | +} | ||
| 134 | + | ||
| 135 | +/* test if a constant matches the constraint */ | ||
| 136 | +static inline int tcg_target_const_match(tcg_target_long val, | ||
| 137 | + const TCGArgConstraint *arg_ct) | ||
| 138 | +{ | ||
| 139 | + int ct; | ||
| 140 | + ct = arg_ct->ct; | ||
| 141 | + if (ct & TCG_CT_CONST) | ||
| 142 | + return 1; | ||
| 143 | + else | ||
| 144 | + return 0; | ||
| 145 | +} | ||
| 146 | + | ||
| 147 | +#define ARITH_ADD 0 | ||
| 148 | +#define ARITH_OR 1 | ||
| 149 | +#define ARITH_ADC 2 | ||
| 150 | +#define ARITH_SBB 3 | ||
| 151 | +#define ARITH_AND 4 | ||
| 152 | +#define ARITH_SUB 5 | ||
| 153 | +#define ARITH_XOR 6 | ||
| 154 | +#define ARITH_CMP 7 | ||
| 155 | + | ||
| 156 | +#define SHIFT_SHL 4 | ||
| 157 | +#define SHIFT_SHR 5 | ||
| 158 | +#define SHIFT_SAR 7 | ||
| 159 | + | ||
| 160 | +#define JCC_JMP (-1) | ||
| 161 | +#define JCC_JO 0x0 | ||
| 162 | +#define JCC_JNO 0x1 | ||
| 163 | +#define JCC_JB 0x2 | ||
| 164 | +#define JCC_JAE 0x3 | ||
| 165 | +#define JCC_JE 0x4 | ||
| 166 | +#define JCC_JNE 0x5 | ||
| 167 | +#define JCC_JBE 0x6 | ||
| 168 | +#define JCC_JA 0x7 | ||
| 169 | +#define JCC_JS 0x8 | ||
| 170 | +#define JCC_JNS 0x9 | ||
| 171 | +#define JCC_JP 0xa | ||
| 172 | +#define JCC_JNP 0xb | ||
| 173 | +#define JCC_JL 0xc | ||
| 174 | +#define JCC_JGE 0xd | ||
| 175 | +#define JCC_JLE 0xe | ||
| 176 | +#define JCC_JG 0xf | ||
| 177 | + | ||
| 178 | +#define P_EXT 0x100 /* 0x0f opcode prefix */ | ||
| 179 | + | ||
| 180 | +static const uint8_t tcg_cond_to_jcc[10] = { | ||
| 181 | + [TCG_COND_EQ] = JCC_JE, | ||
| 182 | + [TCG_COND_NE] = JCC_JNE, | ||
| 183 | + [TCG_COND_LT] = JCC_JL, | ||
| 184 | + [TCG_COND_GE] = JCC_JGE, | ||
| 185 | + [TCG_COND_LE] = JCC_JLE, | ||
| 186 | + [TCG_COND_GT] = JCC_JG, | ||
| 187 | + [TCG_COND_LTU] = JCC_JB, | ||
| 188 | + [TCG_COND_GEU] = JCC_JAE, | ||
| 189 | + [TCG_COND_LEU] = JCC_JBE, | ||
| 190 | + [TCG_COND_GTU] = JCC_JA, | ||
| 191 | +}; | ||
| 192 | + | ||
| 193 | +static inline void tcg_out_opc(TCGContext *s, int opc) | ||
| 194 | +{ | ||
| 195 | + if (opc & P_EXT) | ||
| 196 | + tcg_out8(s, 0x0f); | ||
| 197 | + tcg_out8(s, opc); | ||
| 198 | +} | ||
| 199 | + | ||
| 200 | +static inline void tcg_out_modrm(TCGContext *s, int opc, int r, int rm) | ||
| 201 | +{ | ||
| 202 | + tcg_out_opc(s, opc); | ||
| 203 | + tcg_out8(s, 0xc0 | (r << 3) | rm); | ||
| 204 | +} | ||
| 205 | + | ||
| 206 | +/* rm == -1 means no register index */ | ||
| 207 | +static inline void tcg_out_modrm_offset(TCGContext *s, int opc, int r, int rm, | ||
| 208 | + int32_t offset) | ||
| 209 | +{ | ||
| 210 | + tcg_out_opc(s, opc); | ||
| 211 | + if (rm == -1) { | ||
| 212 | + tcg_out8(s, 0x05 | (r << 3)); | ||
| 213 | + tcg_out32(s, offset); | ||
| 214 | + } else if (offset == 0 && rm != TCG_REG_EBP) { | ||
| 215 | + if (rm == TCG_REG_ESP) { | ||
| 216 | + tcg_out8(s, 0x04 | (r << 3)); | ||
| 217 | + tcg_out8(s, 0x24); | ||
| 218 | + } else { | ||
| 219 | + tcg_out8(s, 0x00 | (r << 3) | rm); | ||
| 220 | + } | ||
| 221 | + } else if ((int8_t)offset == offset) { | ||
| 222 | + if (rm == TCG_REG_ESP) { | ||
| 223 | + tcg_out8(s, 0x44 | (r << 3)); | ||
| 224 | + tcg_out8(s, 0x24); | ||
| 225 | + } else { | ||
| 226 | + tcg_out8(s, 0x40 | (r << 3) | rm); | ||
| 227 | + } | ||
| 228 | + tcg_out8(s, offset); | ||
| 229 | + } else { | ||
| 230 | + if (rm == TCG_REG_ESP) { | ||
| 231 | + tcg_out8(s, 0x84 | (r << 3)); | ||
| 232 | + tcg_out8(s, 0x24); | ||
| 233 | + } else { | ||
| 234 | + tcg_out8(s, 0x80 | (r << 3) | rm); | ||
| 235 | + } | ||
| 236 | + tcg_out32(s, offset); | ||
| 237 | + } | ||
| 238 | +} | ||
| 239 | + | ||
| 240 | +static inline void tcg_out_mov(TCGContext *s, int ret, int arg) | ||
| 241 | +{ | ||
| 242 | + if (arg != ret) | ||
| 243 | + tcg_out_modrm(s, 0x8b, ret, arg); | ||
| 244 | +} | ||
| 245 | + | ||
| 246 | +static inline void tcg_out_movi(TCGContext *s, TCGType type, | ||
| 247 | + int ret, int32_t arg) | ||
| 248 | +{ | ||
| 249 | + if (arg == 0) { | ||
| 250 | + /* xor r0,r0 */ | ||
| 251 | + tcg_out_modrm(s, 0x01 | (ARITH_XOR << 3), ret, ret); | ||
| 252 | + } else { | ||
| 253 | + tcg_out8(s, 0xb8 + ret); | ||
| 254 | + tcg_out32(s, arg); | ||
| 255 | + } | ||
| 256 | +} | ||
| 257 | + | ||
| 258 | +static inline void tcg_out_ld(TCGContext *s, int ret, | ||
| 259 | + int arg1, int32_t arg2) | ||
| 260 | +{ | ||
| 261 | + /* movl */ | ||
| 262 | + tcg_out_modrm_offset(s, 0x8b, ret, arg1, arg2); | ||
| 263 | +} | ||
| 264 | + | ||
| 265 | +static inline void tcg_out_st(TCGContext *s, int arg, | ||
| 266 | + int arg1, int32_t arg2) | ||
| 267 | +{ | ||
| 268 | + /* movl */ | ||
| 269 | + tcg_out_modrm_offset(s, 0x89, arg, arg1, arg2); | ||
| 270 | +} | ||
| 271 | + | ||
| 272 | +static inline void tgen_arithi(TCGContext *s, int c, int r0, int32_t val) | ||
| 273 | +{ | ||
| 274 | + if (val == (int8_t)val) { | ||
| 275 | + tcg_out_modrm(s, 0x83, c, r0); | ||
| 276 | + tcg_out8(s, val); | ||
| 277 | + } else { | ||
| 278 | + tcg_out_modrm(s, 0x81, c, r0); | ||
| 279 | + tcg_out32(s, val); | ||
| 280 | + } | ||
| 281 | +} | ||
| 282 | + | ||
| 283 | +void tcg_out_addi(TCGContext *s, int reg, tcg_target_long val) | ||
| 284 | +{ | ||
| 285 | + if (val != 0) | ||
| 286 | + tgen_arithi(s, ARITH_ADD, reg, val); | ||
| 287 | +} | ||
| 288 | + | ||
| 289 | +static void tcg_out_jxx(TCGContext *s, int opc, int label_index) | ||
| 290 | +{ | ||
| 291 | + int32_t val, val1; | ||
| 292 | + TCGLabel *l = &s->labels[label_index]; | ||
| 293 | + | ||
| 294 | + if (l->has_value) { | ||
| 295 | + val = l->u.value - (tcg_target_long)s->code_ptr; | ||
| 296 | + val1 = val - 2; | ||
| 297 | + if ((int8_t)val1 == val1) { | ||
| 298 | + if (opc == -1) | ||
| 299 | + tcg_out8(s, 0xeb); | ||
| 300 | + else | ||
| 301 | + tcg_out8(s, 0x70 + opc); | ||
| 302 | + tcg_out8(s, val1); | ||
| 303 | + } else { | ||
| 304 | + if (opc == -1) { | ||
| 305 | + tcg_out8(s, 0xe9); | ||
| 306 | + tcg_out32(s, val - 5); | ||
| 307 | + } else { | ||
| 308 | + tcg_out8(s, 0x0f); | ||
| 309 | + tcg_out8(s, 0x80 + opc); | ||
| 310 | + tcg_out32(s, val - 6); | ||
| 311 | + } | ||
| 312 | + } | ||
| 313 | + } else { | ||
| 314 | + if (opc == -1) { | ||
| 315 | + tcg_out8(s, 0xe9); | ||
| 316 | + } else { | ||
| 317 | + tcg_out8(s, 0x0f); | ||
| 318 | + tcg_out8(s, 0x80 + opc); | ||
| 319 | + } | ||
| 320 | + tcg_out_reloc(s, s->code_ptr, R_386_PC32, label_index, -4); | ||
| 321 | + tcg_out32(s, -4); | ||
| 322 | + } | ||
| 323 | +} | ||
| 324 | + | ||
| 325 | +static void tcg_out_brcond(TCGContext *s, int cond, | ||
| 326 | + TCGArg arg1, TCGArg arg2, int const_arg2, | ||
| 327 | + int label_index) | ||
| 328 | +{ | ||
| 329 | + int c; | ||
| 330 | + if (const_arg2) { | ||
| 331 | + if (arg2 == 0) { | ||
| 332 | + /* use test */ | ||
| 333 | + switch(cond) { | ||
| 334 | + case TCG_COND_EQ: | ||
| 335 | + c = JCC_JNE; | ||
| 336 | + break; | ||
| 337 | + case TCG_COND_NE: | ||
| 338 | + c = JCC_JNE; | ||
| 339 | + break; | ||
| 340 | + case TCG_COND_LT: | ||
| 341 | + c = JCC_JS; | ||
| 342 | + break; | ||
| 343 | + case TCG_COND_GE: | ||
| 344 | + c = JCC_JNS; | ||
| 345 | + break; | ||
| 346 | + default: | ||
| 347 | + goto do_cmpi; | ||
| 348 | + } | ||
| 349 | + /* test r, r */ | ||
| 350 | + tcg_out_modrm(s, 0x85, arg1, arg1); | ||
| 351 | + tcg_out_jxx(s, c, label_index); | ||
| 352 | + } else { | ||
| 353 | + do_cmpi: | ||
| 354 | + tgen_arithi(s, ARITH_CMP, arg1, arg2); | ||
| 355 | + tcg_out_jxx(s, tcg_cond_to_jcc[cond], label_index); | ||
| 356 | + } | ||
| 357 | + } else { | ||
| 358 | + tcg_out_modrm(s, 0x01 | (ARITH_CMP << 3), arg1, arg2); | ||
| 359 | + tcg_out_jxx(s, tcg_cond_to_jcc[cond], label_index); | ||
| 360 | + } | ||
| 361 | +} | ||
| 362 | + | ||
| 363 | +/* XXX: we implement it at the target level to avoid having to | ||
| 364 | + handle cross basic blocks temporaries */ | ||
| 365 | +static void tcg_out_brcond2(TCGContext *s, | ||
| 366 | + const TCGArg *args, const int *const_args) | ||
| 367 | +{ | ||
| 368 | + int label_next; | ||
| 369 | + label_next = gen_new_label(); | ||
| 370 | + switch(args[4]) { | ||
| 371 | + case TCG_COND_EQ: | ||
| 372 | + tcg_out_brcond(s, TCG_COND_NE, args[0], args[2], const_args[2], label_next); | ||
| 373 | + tcg_out_brcond(s, TCG_COND_EQ, args[1], args[3], const_args[3], args[5]); | ||
| 374 | + break; | ||
| 375 | + case TCG_COND_NE: | ||
| 376 | + tcg_out_brcond(s, TCG_COND_NE, args[0], args[2], const_args[2], args[5]); | ||
| 377 | + tcg_out_brcond(s, TCG_COND_EQ, args[1], args[3], const_args[3], label_next); | ||
| 378 | + break; | ||
| 379 | + case TCG_COND_LT: | ||
| 380 | + tcg_out_brcond(s, TCG_COND_LT, args[1], args[3], const_args[3], args[5]); | ||
| 381 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
| 382 | + tcg_out_brcond(s, TCG_COND_LT, args[0], args[2], const_args[2], args[5]); | ||
| 383 | + break; | ||
| 384 | + case TCG_COND_LE: | ||
| 385 | + tcg_out_brcond(s, TCG_COND_LT, args[1], args[3], const_args[3], args[5]); | ||
| 386 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
| 387 | + tcg_out_brcond(s, TCG_COND_LE, args[0], args[2], const_args[2], args[5]); | ||
| 388 | + break; | ||
| 389 | + case TCG_COND_GT: | ||
| 390 | + tcg_out_brcond(s, TCG_COND_GT, args[1], args[3], const_args[3], args[5]); | ||
| 391 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
| 392 | + tcg_out_brcond(s, TCG_COND_GT, args[0], args[2], const_args[2], args[5]); | ||
| 393 | + break; | ||
| 394 | + case TCG_COND_GE: | ||
| 395 | + tcg_out_brcond(s, TCG_COND_GT, args[1], args[3], const_args[3], args[5]); | ||
| 396 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
| 397 | + tcg_out_brcond(s, TCG_COND_GE, args[0], args[2], const_args[2], args[5]); | ||
| 398 | + break; | ||
| 399 | + case TCG_COND_LTU: | ||
| 400 | + tcg_out_brcond(s, TCG_COND_LTU, args[1], args[3], const_args[3], args[5]); | ||
| 401 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
| 402 | + tcg_out_brcond(s, TCG_COND_LTU, args[0], args[2], const_args[2], args[5]); | ||
| 403 | + break; | ||
| 404 | + case TCG_COND_LEU: | ||
| 405 | + tcg_out_brcond(s, TCG_COND_LTU, args[1], args[3], const_args[3], args[5]); | ||
| 406 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
| 407 | + tcg_out_brcond(s, TCG_COND_LEU, args[0], args[2], const_args[2], args[5]); | ||
| 408 | + break; | ||
| 409 | + case TCG_COND_GTU: | ||
| 410 | + tcg_out_brcond(s, TCG_COND_GTU, args[1], args[3], const_args[3], args[5]); | ||
| 411 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
| 412 | + tcg_out_brcond(s, TCG_COND_GTU, args[0], args[2], const_args[2], args[5]); | ||
| 413 | + break; | ||
| 414 | + case TCG_COND_GEU: | ||
| 415 | + tcg_out_brcond(s, TCG_COND_GTU, args[1], args[3], const_args[3], args[5]); | ||
| 416 | + tcg_out_brcond(s, TCG_COND_NE, args[1], args[3], const_args[3], label_next); | ||
| 417 | + tcg_out_brcond(s, TCG_COND_GEU, args[0], args[2], const_args[2], args[5]); | ||
| 418 | + break; | ||
| 419 | + default: | ||
| 420 | + tcg_abort(); | ||
| 421 | + } | ||
| 422 | + tcg_out_label(s, label_next, (tcg_target_long)s->code_ptr); | ||
| 423 | +} | ||
| 424 | + | ||
| 425 | +#if defined(CONFIG_SOFTMMU) | ||
| 426 | +extern void __ldb_mmu(void); | ||
| 427 | +extern void __ldw_mmu(void); | ||
| 428 | +extern void __ldl_mmu(void); | ||
| 429 | +extern void __ldq_mmu(void); | ||
| 430 | + | ||
| 431 | +extern void __stb_mmu(void); | ||
| 432 | +extern void __stw_mmu(void); | ||
| 433 | +extern void __stl_mmu(void); | ||
| 434 | +extern void __stq_mmu(void); | ||
| 435 | + | ||
| 436 | +static void *qemu_ld_helpers[4] = { | ||
| 437 | + __ldb_mmu, | ||
| 438 | + __ldw_mmu, | ||
| 439 | + __ldl_mmu, | ||
| 440 | + __ldq_mmu, | ||
| 441 | +}; | ||
| 442 | + | ||
| 443 | +static void *qemu_st_helpers[4] = { | ||
| 444 | + __stb_mmu, | ||
| 445 | + __stw_mmu, | ||
| 446 | + __stl_mmu, | ||
| 447 | + __stq_mmu, | ||
| 448 | +}; | ||
| 449 | +#endif | ||
| 450 | + | ||
| 451 | +/* XXX: qemu_ld and qemu_st could be modified to clobber only EDX and | ||
| 452 | + EAX. It will be useful once fixed registers globals are less | ||
| 453 | + common. */ | ||
| 454 | +static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, | ||
| 455 | + int opc) | ||
| 456 | +{ | ||
| 457 | + int addr_reg, data_reg, data_reg2, r0, r1, mem_index, s_bits, bswap; | ||
| 458 | +#if defined(CONFIG_SOFTMMU) | ||
| 459 | + uint8_t *label1_ptr, *label2_ptr; | ||
| 460 | +#endif | ||
| 461 | +#if TARGET_LONG_BITS == 64 | ||
| 462 | +#if defined(CONFIG_SOFTMMU) | ||
| 463 | + uint8_t *label3_ptr; | ||
| 464 | +#endif | ||
| 465 | + int addr_reg2; | ||
| 466 | +#endif | ||
| 467 | + | ||
| 468 | + data_reg = *args++; | ||
| 469 | + if (opc == 3) | ||
| 470 | + data_reg2 = *args++; | ||
| 471 | + else | ||
| 472 | + data_reg2 = 0; | ||
| 473 | + addr_reg = *args++; | ||
| 474 | +#if TARGET_LONG_BITS == 64 | ||
| 475 | + addr_reg2 = *args++; | ||
| 476 | +#endif | ||
| 477 | + mem_index = *args; | ||
| 478 | + s_bits = opc & 3; | ||
| 479 | + | ||
| 480 | + r0 = TCG_REG_EAX; | ||
| 481 | + r1 = TCG_REG_EDX; | ||
| 482 | + | ||
| 483 | +#if defined(CONFIG_SOFTMMU) | ||
| 484 | + tcg_out_mov(s, r1, addr_reg); | ||
| 485 | + | ||
| 486 | + tcg_out_mov(s, r0, addr_reg); | ||
| 487 | + | ||
| 488 | + tcg_out_modrm(s, 0xc1, 5, r1); /* shr $x, r1 */ | ||
| 489 | + tcg_out8(s, TARGET_PAGE_BITS - CPU_TLB_ENTRY_BITS); | ||
| 490 | + | ||
| 491 | + tcg_out_modrm(s, 0x81, 4, r0); /* andl $x, r0 */ | ||
| 492 | + tcg_out32(s, TARGET_PAGE_MASK | ((1 << s_bits) - 1)); | ||
| 493 | + | ||
| 494 | + tcg_out_modrm(s, 0x81, 4, r1); /* andl $x, r1 */ | ||
| 495 | + tcg_out32(s, (CPU_TLB_SIZE - 1) << CPU_TLB_ENTRY_BITS); | ||
| 496 | + | ||
| 497 | + tcg_out_opc(s, 0x8d); /* lea offset(r1, %ebp), r1 */ | ||
| 498 | + tcg_out8(s, 0x80 | (r1 << 3) | 0x04); | ||
| 499 | + tcg_out8(s, (5 << 3) | r1); | ||
| 500 | + tcg_out32(s, offsetof(CPUState, tlb_table[mem_index][0].addr_read)); | ||
| 501 | + | ||
| 502 | + /* cmp 0(r1), r0 */ | ||
| 503 | + tcg_out_modrm_offset(s, 0x3b, r0, r1, 0); | ||
| 504 | + | ||
| 505 | + tcg_out_mov(s, r0, addr_reg); | ||
| 506 | + | ||
| 507 | +#if TARGET_LONG_BITS == 32 | ||
| 508 | + /* je label1 */ | ||
| 509 | + tcg_out8(s, 0x70 + JCC_JE); | ||
| 510 | + label1_ptr = s->code_ptr; | ||
| 511 | + s->code_ptr++; | ||
| 512 | +#else | ||
| 513 | + /* jne label3 */ | ||
| 514 | + tcg_out8(s, 0x70 + JCC_JNE); | ||
| 515 | + label3_ptr = s->code_ptr; | ||
| 516 | + s->code_ptr++; | ||
| 517 | + | ||
| 518 | + /* cmp 4(r1), addr_reg2 */ | ||
| 519 | + tcg_out_modrm_offset(s, 0x3b, addr_reg2, r1, 4); | ||
| 520 | + | ||
| 521 | + /* je label1 */ | ||
| 522 | + tcg_out8(s, 0x70 + JCC_JE); | ||
| 523 | + label1_ptr = s->code_ptr; | ||
| 524 | + s->code_ptr++; | ||
| 525 | + | ||
| 526 | + /* label3: */ | ||
| 527 | + *label3_ptr = s->code_ptr - label3_ptr - 1; | ||
| 528 | +#endif | ||
| 529 | + | ||
| 530 | + /* XXX: move that code at the end of the TB */ | ||
| 531 | +#if TARGET_LONG_BITS == 32 | ||
| 532 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_EDX, mem_index); | ||
| 533 | +#else | ||
| 534 | + tcg_out_mov(s, TCG_REG_EDX, addr_reg2); | ||
| 535 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_ECX, mem_index); | ||
| 536 | +#endif | ||
| 537 | + tcg_out8(s, 0xe8); | ||
| 538 | + tcg_out32(s, (tcg_target_long)qemu_ld_helpers[s_bits] - | ||
| 539 | + (tcg_target_long)s->code_ptr - 4); | ||
| 540 | + | ||
| 541 | + switch(opc) { | ||
| 542 | + case 0 | 4: | ||
| 543 | + /* movsbl */ | ||
| 544 | + tcg_out_modrm(s, 0xbe | P_EXT, data_reg, TCG_REG_EAX); | ||
| 545 | + break; | ||
| 546 | + case 1 | 4: | ||
| 547 | + /* movswl */ | ||
| 548 | + tcg_out_modrm(s, 0xbf | P_EXT, data_reg, TCG_REG_EAX); | ||
| 549 | + break; | ||
| 550 | + case 0: | ||
| 551 | + case 1: | ||
| 552 | + case 2: | ||
| 553 | + default: | ||
| 554 | + tcg_out_mov(s, data_reg, TCG_REG_EAX); | ||
| 555 | + break; | ||
| 556 | + case 3: | ||
| 557 | + if (data_reg == TCG_REG_EDX) { | ||
| 558 | + tcg_out_opc(s, 0x90 + TCG_REG_EDX); /* xchg %edx, %eax */ | ||
| 559 | + tcg_out_mov(s, data_reg2, TCG_REG_EAX); | ||
| 560 | + } else { | ||
| 561 | + tcg_out_mov(s, data_reg, TCG_REG_EAX); | ||
| 562 | + tcg_out_mov(s, data_reg2, TCG_REG_EDX); | ||
| 563 | + } | ||
| 564 | + break; | ||
| 565 | + } | ||
| 566 | + | ||
| 567 | + /* jmp label2 */ | ||
| 568 | + tcg_out8(s, 0xeb); | ||
| 569 | + label2_ptr = s->code_ptr; | ||
| 570 | + s->code_ptr++; | ||
| 571 | + | ||
| 572 | + /* label1: */ | ||
| 573 | + *label1_ptr = s->code_ptr - label1_ptr - 1; | ||
| 574 | + | ||
| 575 | + /* add x(r1), r0 */ | ||
| 576 | + tcg_out_modrm_offset(s, 0x03, r0, r1, offsetof(CPUTLBEntry, addend) - | ||
| 577 | + offsetof(CPUTLBEntry, addr_read)); | ||
| 578 | +#else | ||
| 579 | + r0 = addr_reg; | ||
| 580 | +#endif | ||
| 581 | + | ||
| 582 | +#ifdef TARGET_WORDS_BIGENDIAN | ||
| 583 | + bswap = 1; | ||
| 584 | +#else | ||
| 585 | + bswap = 0; | ||
| 586 | +#endif | ||
| 587 | + switch(opc) { | ||
| 588 | + case 0: | ||
| 589 | + /* movzbl */ | ||
| 590 | + tcg_out_modrm_offset(s, 0xb6 | P_EXT, data_reg, r0, 0); | ||
| 591 | + break; | ||
| 592 | + case 0 | 4: | ||
| 593 | + /* movsbl */ | ||
| 594 | + tcg_out_modrm_offset(s, 0xbe | P_EXT, data_reg, r0, 0); | ||
| 595 | + break; | ||
| 596 | + case 1: | ||
| 597 | + /* movzwl */ | ||
| 598 | + tcg_out_modrm_offset(s, 0xb7 | P_EXT, data_reg, r0, 0); | ||
| 599 | + if (bswap) { | ||
| 600 | + /* rolw $8, data_reg */ | ||
| 601 | + tcg_out8(s, 0x66); | ||
| 602 | + tcg_out_modrm(s, 0xc1, 0, data_reg); | ||
| 603 | + tcg_out8(s, 8); | ||
| 604 | + } | ||
| 605 | + break; | ||
| 606 | + case 1 | 4: | ||
| 607 | + /* movswl */ | ||
| 608 | + tcg_out_modrm_offset(s, 0xbf | P_EXT, data_reg, r0, 0); | ||
| 609 | + if (bswap) { | ||
| 610 | + /* rolw $8, data_reg */ | ||
| 611 | + tcg_out8(s, 0x66); | ||
| 612 | + tcg_out_modrm(s, 0xc1, 0, data_reg); | ||
| 613 | + tcg_out8(s, 8); | ||
| 614 | + | ||
| 615 | + /* movswl data_reg, data_reg */ | ||
| 616 | + tcg_out_modrm(s, 0xbf | P_EXT, data_reg, data_reg); | ||
| 617 | + } | ||
| 618 | + break; | ||
| 619 | + case 2: | ||
| 620 | + /* movl (r0), data_reg */ | ||
| 621 | + tcg_out_modrm_offset(s, 0x8b, data_reg, r0, 0); | ||
| 622 | + if (bswap) { | ||
| 623 | + /* bswap */ | ||
| 624 | + tcg_out_opc(s, (0xc8 + data_reg) | P_EXT); | ||
| 625 | + } | ||
| 626 | + break; | ||
| 627 | + case 3: | ||
| 628 | + /* XXX: could be nicer */ | ||
| 629 | + if (r0 == data_reg) { | ||
| 630 | + r1 = TCG_REG_EDX; | ||
| 631 | + if (r1 == data_reg) | ||
| 632 | + r1 = TCG_REG_EAX; | ||
| 633 | + tcg_out_mov(s, r1, r0); | ||
| 634 | + r0 = r1; | ||
| 635 | + } | ||
| 636 | + if (!bswap) { | ||
| 637 | + tcg_out_modrm_offset(s, 0x8b, data_reg, r0, 0); | ||
| 638 | + tcg_out_modrm_offset(s, 0x8b, data_reg2, r0, 4); | ||
| 639 | + } else { | ||
| 640 | + tcg_out_modrm_offset(s, 0x8b, data_reg, r0, 4); | ||
| 641 | + tcg_out_opc(s, (0xc8 + data_reg) | P_EXT); | ||
| 642 | + | ||
| 643 | + tcg_out_modrm_offset(s, 0x8b, data_reg2, r0, 0); | ||
| 644 | + /* bswap */ | ||
| 645 | + tcg_out_opc(s, (0xc8 + data_reg2) | P_EXT); | ||
| 646 | + } | ||
| 647 | + break; | ||
| 648 | + default: | ||
| 649 | + tcg_abort(); | ||
| 650 | + } | ||
| 651 | + | ||
| 652 | +#if defined(CONFIG_SOFTMMU) | ||
| 653 | + /* label2: */ | ||
| 654 | + *label2_ptr = s->code_ptr - label2_ptr - 1; | ||
| 655 | +#endif | ||
| 656 | +} | ||
| 657 | + | ||
| 658 | + | ||
| 659 | +static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, | ||
| 660 | + int opc) | ||
| 661 | +{ | ||
| 662 | + int addr_reg, data_reg, data_reg2, r0, r1, mem_index, s_bits, bswap; | ||
| 663 | +#if defined(CONFIG_SOFTMMU) | ||
| 664 | + uint8_t *label1_ptr, *label2_ptr; | ||
| 665 | +#endif | ||
| 666 | +#if TARGET_LONG_BITS == 64 | ||
| 667 | +#if defined(CONFIG_SOFTMMU) | ||
| 668 | + uint8_t *label3_ptr; | ||
| 669 | +#endif | ||
| 670 | + int addr_reg2; | ||
| 671 | +#endif | ||
| 672 | + | ||
| 673 | + data_reg = *args++; | ||
| 674 | + if (opc == 3) | ||
| 675 | + data_reg2 = *args++; | ||
| 676 | + else | ||
| 677 | + data_reg2 = 0; | ||
| 678 | + addr_reg = *args++; | ||
| 679 | +#if TARGET_LONG_BITS == 64 | ||
| 680 | + addr_reg2 = *args++; | ||
| 681 | +#endif | ||
| 682 | + mem_index = *args; | ||
| 683 | + | ||
| 684 | + s_bits = opc; | ||
| 685 | + | ||
| 686 | + r0 = TCG_REG_EAX; | ||
| 687 | + r1 = TCG_REG_EDX; | ||
| 688 | + | ||
| 689 | +#if defined(CONFIG_SOFTMMU) | ||
| 690 | + tcg_out_mov(s, r1, addr_reg); | ||
| 691 | + | ||
| 692 | + tcg_out_mov(s, r0, addr_reg); | ||
| 693 | + | ||
| 694 | + tcg_out_modrm(s, 0xc1, 5, r1); /* shr $x, r1 */ | ||
| 695 | + tcg_out8(s, TARGET_PAGE_BITS - CPU_TLB_ENTRY_BITS); | ||
| 696 | + | ||
| 697 | + tcg_out_modrm(s, 0x81, 4, r0); /* andl $x, r0 */ | ||
| 698 | + tcg_out32(s, TARGET_PAGE_MASK | ((1 << s_bits) - 1)); | ||
| 699 | + | ||
| 700 | + tcg_out_modrm(s, 0x81, 4, r1); /* andl $x, r1 */ | ||
| 701 | + tcg_out32(s, (CPU_TLB_SIZE - 1) << CPU_TLB_ENTRY_BITS); | ||
| 702 | + | ||
| 703 | + tcg_out_opc(s, 0x8d); /* lea offset(r1, %ebp), r1 */ | ||
| 704 | + tcg_out8(s, 0x80 | (r1 << 3) | 0x04); | ||
| 705 | + tcg_out8(s, (5 << 3) | r1); | ||
| 706 | + tcg_out32(s, offsetof(CPUState, tlb_table[mem_index][0].addr_write)); | ||
| 707 | + | ||
| 708 | + /* cmp 0(r1), r0 */ | ||
| 709 | + tcg_out_modrm_offset(s, 0x3b, r0, r1, 0); | ||
| 710 | + | ||
| 711 | + tcg_out_mov(s, r0, addr_reg); | ||
| 712 | + | ||
| 713 | +#if TARGET_LONG_BITS == 32 | ||
| 714 | + /* je label1 */ | ||
| 715 | + tcg_out8(s, 0x70 + JCC_JE); | ||
| 716 | + label1_ptr = s->code_ptr; | ||
| 717 | + s->code_ptr++; | ||
| 718 | +#else | ||
| 719 | + /* jne label3 */ | ||
| 720 | + tcg_out8(s, 0x70 + JCC_JNE); | ||
| 721 | + label3_ptr = s->code_ptr; | ||
| 722 | + s->code_ptr++; | ||
| 723 | + | ||
| 724 | + /* cmp 4(r1), addr_reg2 */ | ||
| 725 | + tcg_out_modrm_offset(s, 0x3b, addr_reg2, r1, 4); | ||
| 726 | + | ||
| 727 | + /* je label1 */ | ||
| 728 | + tcg_out8(s, 0x70 + JCC_JE); | ||
| 729 | + label1_ptr = s->code_ptr; | ||
| 730 | + s->code_ptr++; | ||
| 731 | + | ||
| 732 | + /* label3: */ | ||
| 733 | + *label3_ptr = s->code_ptr - label3_ptr - 1; | ||
| 734 | +#endif | ||
| 735 | + | ||
| 736 | + /* XXX: move that code at the end of the TB */ | ||
| 737 | +#if TARGET_LONG_BITS == 32 | ||
| 738 | + if (opc == 3) { | ||
| 739 | + tcg_out_mov(s, TCG_REG_EDX, data_reg); | ||
| 740 | + tcg_out_mov(s, TCG_REG_ECX, data_reg2); | ||
| 741 | + tcg_out8(s, 0x6a); /* push Ib */ | ||
| 742 | + tcg_out8(s, mem_index); | ||
| 743 | + tcg_out8(s, 0xe8); | ||
| 744 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | ||
| 745 | + (tcg_target_long)s->code_ptr - 4); | ||
| 746 | + tcg_out_addi(s, TCG_REG_ESP, 4); | ||
| 747 | + } else { | ||
| 748 | + switch(opc) { | ||
| 749 | + case 0: | ||
| 750 | + /* movzbl */ | ||
| 751 | + tcg_out_modrm(s, 0xb6 | P_EXT, TCG_REG_EDX, data_reg); | ||
| 752 | + break; | ||
| 753 | + case 1: | ||
| 754 | + /* movzwl */ | ||
| 755 | + tcg_out_modrm(s, 0xb7 | P_EXT, TCG_REG_EDX, data_reg); | ||
| 756 | + break; | ||
| 757 | + case 2: | ||
| 758 | + tcg_out_mov(s, TCG_REG_EDX, data_reg); | ||
| 759 | + break; | ||
| 760 | + } | ||
| 761 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_ECX, mem_index); | ||
| 762 | + tcg_out8(s, 0xe8); | ||
| 763 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | ||
| 764 | + (tcg_target_long)s->code_ptr - 4); | ||
| 765 | + } | ||
| 766 | +#else | ||
| 767 | + if (opc == 3) { | ||
| 768 | + tcg_out_mov(s, TCG_REG_EDX, addr_reg2); | ||
| 769 | + tcg_out8(s, 0x6a); /* push Ib */ | ||
| 770 | + tcg_out8(s, mem_index); | ||
| 771 | + tcg_out_opc(s, 0x50 + data_reg2); /* push */ | ||
| 772 | + tcg_out_opc(s, 0x50 + data_reg); /* push */ | ||
| 773 | + tcg_out8(s, 0xe8); | ||
| 774 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | ||
| 775 | + (tcg_target_long)s->code_ptr - 4); | ||
| 776 | + tcg_out_addi(s, TCG_REG_ESP, 12); | ||
| 777 | + } else { | ||
| 778 | + tcg_out_mov(s, TCG_REG_EDX, addr_reg2); | ||
| 779 | + switch(opc) { | ||
| 780 | + case 0: | ||
| 781 | + /* movzbl */ | ||
| 782 | + tcg_out_modrm(s, 0xb6 | P_EXT, TCG_REG_ECX, data_reg); | ||
| 783 | + break; | ||
| 784 | + case 1: | ||
| 785 | + /* movzwl */ | ||
| 786 | + tcg_out_modrm(s, 0xb7 | P_EXT, TCG_REG_ECX, data_reg); | ||
| 787 | + break; | ||
| 788 | + case 2: | ||
| 789 | + tcg_out_mov(s, TCG_REG_ECX, data_reg); | ||
| 790 | + break; | ||
| 791 | + } | ||
| 792 | + tcg_out8(s, 0x6a); /* push Ib */ | ||
| 793 | + tcg_out8(s, mem_index); | ||
| 794 | + tcg_out8(s, 0xe8); | ||
| 795 | + tcg_out32(s, (tcg_target_long)qemu_st_helpers[s_bits] - | ||
| 796 | + (tcg_target_long)s->code_ptr - 4); | ||
| 797 | + tcg_out_addi(s, TCG_REG_ESP, 4); | ||
| 798 | + } | ||
| 799 | +#endif | ||
| 800 | + | ||
| 801 | + /* jmp label2 */ | ||
| 802 | + tcg_out8(s, 0xeb); | ||
| 803 | + label2_ptr = s->code_ptr; | ||
| 804 | + s->code_ptr++; | ||
| 805 | + | ||
| 806 | + /* label1: */ | ||
| 807 | + *label1_ptr = s->code_ptr - label1_ptr - 1; | ||
| 808 | + | ||
| 809 | + /* add x(r1), r0 */ | ||
| 810 | + tcg_out_modrm_offset(s, 0x03, r0, r1, offsetof(CPUTLBEntry, addend) - | ||
| 811 | + offsetof(CPUTLBEntry, addr_write)); | ||
| 812 | +#else | ||
| 813 | + r0 = addr_reg; | ||
| 814 | +#endif | ||
| 815 | + | ||
| 816 | +#ifdef TARGET_WORDS_BIGENDIAN | ||
| 817 | + bswap = 1; | ||
| 818 | +#else | ||
| 819 | + bswap = 0; | ||
| 820 | +#endif | ||
| 821 | + switch(opc) { | ||
| 822 | + case 0: | ||
| 823 | + /* movb */ | ||
| 824 | + tcg_out_modrm_offset(s, 0x88, data_reg, r0, 0); | ||
| 825 | + break; | ||
| 826 | + case 1: | ||
| 827 | + if (bswap) { | ||
| 828 | + tcg_out_mov(s, r1, data_reg); | ||
| 829 | + tcg_out8(s, 0x66); /* rolw $8, %ecx */ | ||
| 830 | + tcg_out_modrm(s, 0xc1, 0, r1); | ||
| 831 | + tcg_out8(s, 8); | ||
| 832 | + data_reg = r1; | ||
| 833 | + } | ||
| 834 | + /* movw */ | ||
| 835 | + tcg_out8(s, 0x66); | ||
| 836 | + tcg_out_modrm_offset(s, 0x89, data_reg, r0, 0); | ||
| 837 | + break; | ||
| 838 | + case 2: | ||
| 839 | + if (bswap) { | ||
| 840 | + tcg_out_mov(s, r1, data_reg); | ||
| 841 | + /* bswap data_reg */ | ||
| 842 | + tcg_out_opc(s, (0xc8 + r1) | P_EXT); | ||
| 843 | + data_reg = r1; | ||
| 844 | + } | ||
| 845 | + /* movl */ | ||
| 846 | + tcg_out_modrm_offset(s, 0x89, data_reg, r0, 0); | ||
| 847 | + break; | ||
| 848 | + case 3: | ||
| 849 | + if (bswap) { | ||
| 850 | + tcg_out_mov(s, r1, data_reg2); | ||
| 851 | + /* bswap data_reg */ | ||
| 852 | + tcg_out_opc(s, (0xc8 + r1) | P_EXT); | ||
| 853 | + tcg_out_modrm_offset(s, 0x89, r1, r0, 0); | ||
| 854 | + tcg_out_mov(s, r1, data_reg); | ||
| 855 | + /* bswap data_reg */ | ||
| 856 | + tcg_out_opc(s, (0xc8 + r1) | P_EXT); | ||
| 857 | + tcg_out_modrm_offset(s, 0x89, r1, r0, 4); | ||
| 858 | + } else { | ||
| 859 | + tcg_out_modrm_offset(s, 0x89, data_reg, r0, 0); | ||
| 860 | + tcg_out_modrm_offset(s, 0x89, data_reg2, r0, 4); | ||
| 861 | + } | ||
| 862 | + break; | ||
| 863 | + default: | ||
| 864 | + tcg_abort(); | ||
| 865 | + } | ||
| 866 | + | ||
| 867 | +#if defined(CONFIG_SOFTMMU) | ||
| 868 | + /* label2: */ | ||
| 869 | + *label2_ptr = s->code_ptr - label2_ptr - 1; | ||
| 870 | +#endif | ||
| 871 | +} | ||
| 872 | + | ||
| 873 | +static inline void tcg_out_op(TCGContext *s, int opc, | ||
| 874 | + const TCGArg *args, const int *const_args) | ||
| 875 | +{ | ||
| 876 | + int c; | ||
| 877 | + | ||
| 878 | + switch(opc) { | ||
| 879 | + case INDEX_op_exit_tb: | ||
| 880 | + tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_EAX, args[0]); | ||
| 881 | + tcg_out8(s, 0xc3); /* ret */ | ||
| 882 | + break; | ||
| 883 | + case INDEX_op_goto_tb: | ||
| 884 | + if (s->tb_jmp_offset) { | ||
| 885 | + /* direct jump method */ | ||
| 886 | + tcg_out8(s, 0xe9); /* jmp im */ | ||
| 887 | + s->tb_jmp_offset[args[0]] = s->code_ptr - s->code_buf; | ||
| 888 | + tcg_out32(s, 0); | ||
| 889 | + } else { | ||
| 890 | + /* indirect jump method */ | ||
| 891 | + /* jmp Ev */ | ||
| 892 | + tcg_out_modrm_offset(s, 0xff, 4, -1, | ||
| 893 | + (tcg_target_long)(s->tb_next + args[0])); | ||
| 894 | + } | ||
| 895 | + s->tb_next_offset[args[0]] = s->code_ptr - s->code_buf; | ||
| 896 | + break; | ||
| 897 | + case INDEX_op_call: | ||
| 898 | + if (const_args[0]) { | ||
| 899 | + tcg_out8(s, 0xe8); | ||
| 900 | + tcg_out32(s, args[0] - (tcg_target_long)s->code_ptr - 4); | ||
| 901 | + } else { | ||
| 902 | + tcg_out_modrm(s, 0xff, 2, args[0]); | ||
| 903 | + } | ||
| 904 | + break; | ||
| 905 | + case INDEX_op_jmp: | ||
| 906 | + if (const_args[0]) { | ||
| 907 | + tcg_out8(s, 0xe9); | ||
| 908 | + tcg_out32(s, args[0] - (tcg_target_long)s->code_ptr - 4); | ||
| 909 | + } else { | ||
| 910 | + tcg_out_modrm(s, 0xff, 4, args[0]); | ||
| 911 | + } | ||
| 912 | + break; | ||
| 913 | + case INDEX_op_br: | ||
| 914 | + tcg_out_jxx(s, JCC_JMP, args[0]); | ||
| 915 | + break; | ||
| 916 | + case INDEX_op_movi_i32: | ||
| 917 | + tcg_out_movi(s, TCG_TYPE_I32, args[0], args[1]); | ||
| 918 | + break; | ||
| 919 | + case INDEX_op_ld8u_i32: | ||
| 920 | + /* movzbl */ | ||
| 921 | + tcg_out_modrm_offset(s, 0xb6 | P_EXT, args[0], args[1], args[2]); | ||
| 922 | + break; | ||
| 923 | + case INDEX_op_ld8s_i32: | ||
| 924 | + /* movsbl */ | ||
| 925 | + tcg_out_modrm_offset(s, 0xbe | P_EXT, args[0], args[1], args[2]); | ||
| 926 | + break; | ||
| 927 | + case INDEX_op_ld16u_i32: | ||
| 928 | + /* movzwl */ | ||
| 929 | + tcg_out_modrm_offset(s, 0xb7 | P_EXT, args[0], args[1], args[2]); | ||
| 930 | + break; | ||
| 931 | + case INDEX_op_ld16s_i32: | ||
| 932 | + /* movswl */ | ||
| 933 | + tcg_out_modrm_offset(s, 0xbf | P_EXT, args[0], args[1], args[2]); | ||
| 934 | + break; | ||
| 935 | + case INDEX_op_ld_i32: | ||
| 936 | + /* movl */ | ||
| 937 | + tcg_out_modrm_offset(s, 0x8b, args[0], args[1], args[2]); | ||
| 938 | + break; | ||
| 939 | + case INDEX_op_st8_i32: | ||
| 940 | + /* movb */ | ||
| 941 | + tcg_out_modrm_offset(s, 0x88, args[0], args[1], args[2]); | ||
| 942 | + break; | ||
| 943 | + case INDEX_op_st16_i32: | ||
| 944 | + /* movw */ | ||
| 945 | + tcg_out8(s, 0x66); | ||
| 946 | + tcg_out_modrm_offset(s, 0x89, args[0], args[1], args[2]); | ||
| 947 | + break; | ||
| 948 | + case INDEX_op_st_i32: | ||
| 949 | + /* movl */ | ||
| 950 | + tcg_out_modrm_offset(s, 0x89, args[0], args[1], args[2]); | ||
| 951 | + break; | ||
| 952 | + case INDEX_op_sub_i32: | ||
| 953 | + c = ARITH_SUB; | ||
| 954 | + goto gen_arith; | ||
| 955 | + case INDEX_op_and_i32: | ||
| 956 | + c = ARITH_AND; | ||
| 957 | + goto gen_arith; | ||
| 958 | + case INDEX_op_or_i32: | ||
| 959 | + c = ARITH_OR; | ||
| 960 | + goto gen_arith; | ||
| 961 | + case INDEX_op_xor_i32: | ||
| 962 | + c = ARITH_XOR; | ||
| 963 | + goto gen_arith; | ||
| 964 | + case INDEX_op_add_i32: | ||
| 965 | + c = ARITH_ADD; | ||
| 966 | + gen_arith: | ||
| 967 | + if (const_args[2]) { | ||
| 968 | + tgen_arithi(s, c, args[0], args[2]); | ||
| 969 | + } else { | ||
| 970 | + tcg_out_modrm(s, 0x01 | (c << 3), args[2], args[0]); | ||
| 971 | + } | ||
| 972 | + break; | ||
| 973 | + case INDEX_op_mul_i32: | ||
| 974 | + if (const_args[2]) { | ||
| 975 | + int32_t val; | ||
| 976 | + val = args[2]; | ||
| 977 | + if (val == (int8_t)val) { | ||
| 978 | + tcg_out_modrm(s, 0x6b, args[0], args[0]); | ||
| 979 | + tcg_out8(s, val); | ||
| 980 | + } else { | ||
| 981 | + tcg_out_modrm(s, 0x69, args[0], args[0]); | ||
| 982 | + tcg_out32(s, val); | ||
| 983 | + } | ||
| 984 | + } else { | ||
| 985 | + tcg_out_modrm(s, 0xaf | P_EXT, args[0], args[2]); | ||
| 986 | + } | ||
| 987 | + break; | ||
| 988 | + case INDEX_op_mulu2_i32: | ||
| 989 | + tcg_out_modrm(s, 0xf7, 4, args[3]); | ||
| 990 | + break; | ||
| 991 | + case INDEX_op_div2_i32: | ||
| 992 | + tcg_out_modrm(s, 0xf7, 7, args[4]); | ||
| 993 | + break; | ||
| 994 | + case INDEX_op_divu2_i32: | ||
| 995 | + tcg_out_modrm(s, 0xf7, 6, args[4]); | ||
| 996 | + break; | ||
| 997 | + case INDEX_op_shl_i32: | ||
| 998 | + c = SHIFT_SHL; | ||
| 999 | + gen_shift32: | ||
| 1000 | + if (const_args[2]) { | ||
| 1001 | + if (args[2] == 1) { | ||
| 1002 | + tcg_out_modrm(s, 0xd1, c, args[0]); | ||
| 1003 | + } else { | ||
| 1004 | + tcg_out_modrm(s, 0xc1, c, args[0]); | ||
| 1005 | + tcg_out8(s, args[2]); | ||
| 1006 | + } | ||
| 1007 | + } else { | ||
| 1008 | + tcg_out_modrm(s, 0xd3, c, args[0]); | ||
| 1009 | + } | ||
| 1010 | + break; | ||
| 1011 | + case INDEX_op_shr_i32: | ||
| 1012 | + c = SHIFT_SHR; | ||
| 1013 | + goto gen_shift32; | ||
| 1014 | + case INDEX_op_sar_i32: | ||
| 1015 | + c = SHIFT_SAR; | ||
| 1016 | + goto gen_shift32; | ||
| 1017 | + | ||
| 1018 | + case INDEX_op_add2_i32: | ||
| 1019 | + if (const_args[4]) | ||
| 1020 | + tgen_arithi(s, ARITH_ADD, args[0], args[4]); | ||
| 1021 | + else | ||
| 1022 | + tcg_out_modrm(s, 0x01 | (ARITH_ADD << 3), args[4], args[0]); | ||
| 1023 | + if (const_args[5]) | ||
| 1024 | + tgen_arithi(s, ARITH_ADC, args[1], args[5]); | ||
| 1025 | + else | ||
| 1026 | + tcg_out_modrm(s, 0x01 | (ARITH_ADC << 3), args[5], args[1]); | ||
| 1027 | + break; | ||
| 1028 | + case INDEX_op_sub2_i32: | ||
| 1029 | + if (const_args[4]) | ||
| 1030 | + tgen_arithi(s, ARITH_SUB, args[0], args[4]); | ||
| 1031 | + else | ||
| 1032 | + tcg_out_modrm(s, 0x01 | (ARITH_SUB << 3), args[4], args[0]); | ||
| 1033 | + if (const_args[5]) | ||
| 1034 | + tgen_arithi(s, ARITH_SBB, args[1], args[5]); | ||
| 1035 | + else | ||
| 1036 | + tcg_out_modrm(s, 0x01 | (ARITH_SBB << 3), args[5], args[1]); | ||
| 1037 | + break; | ||
| 1038 | + case INDEX_op_brcond_i32: | ||
| 1039 | + tcg_out_brcond(s, args[2], args[0], args[1], const_args[1], args[3]); | ||
| 1040 | + break; | ||
| 1041 | + case INDEX_op_brcond2_i32: | ||
| 1042 | + tcg_out_brcond2(s, args, const_args); | ||
| 1043 | + break; | ||
| 1044 | + | ||
| 1045 | + case INDEX_op_qemu_ld8u: | ||
| 1046 | + tcg_out_qemu_ld(s, args, 0); | ||
| 1047 | + break; | ||
| 1048 | + case INDEX_op_qemu_ld8s: | ||
| 1049 | + tcg_out_qemu_ld(s, args, 0 | 4); | ||
| 1050 | + break; | ||
| 1051 | + case INDEX_op_qemu_ld16u: | ||
| 1052 | + tcg_out_qemu_ld(s, args, 1); | ||
| 1053 | + break; | ||
| 1054 | + case INDEX_op_qemu_ld16s: | ||
| 1055 | + tcg_out_qemu_ld(s, args, 1 | 4); | ||
| 1056 | + break; | ||
| 1057 | + case INDEX_op_qemu_ld32u: | ||
| 1058 | + tcg_out_qemu_ld(s, args, 2); | ||
| 1059 | + break; | ||
| 1060 | + case INDEX_op_qemu_ld64: | ||
| 1061 | + tcg_out_qemu_ld(s, args, 3); | ||
| 1062 | + break; | ||
| 1063 | + | ||
| 1064 | + case INDEX_op_qemu_st8: | ||
| 1065 | + tcg_out_qemu_st(s, args, 0); | ||
| 1066 | + break; | ||
| 1067 | + case INDEX_op_qemu_st16: | ||
| 1068 | + tcg_out_qemu_st(s, args, 1); | ||
| 1069 | + break; | ||
| 1070 | + case INDEX_op_qemu_st32: | ||
| 1071 | + tcg_out_qemu_st(s, args, 2); | ||
| 1072 | + break; | ||
| 1073 | + case INDEX_op_qemu_st64: | ||
| 1074 | + tcg_out_qemu_st(s, args, 3); | ||
| 1075 | + break; | ||
| 1076 | + | ||
| 1077 | + default: | ||
| 1078 | + tcg_abort(); | ||
| 1079 | + } | ||
| 1080 | +} | ||
| 1081 | + | ||
| 1082 | +static const TCGTargetOpDef x86_op_defs[] = { | ||
| 1083 | + { INDEX_op_exit_tb, { } }, | ||
| 1084 | + { INDEX_op_goto_tb, { } }, | ||
| 1085 | + { INDEX_op_call, { "ri" } }, | ||
| 1086 | + { INDEX_op_jmp, { "ri" } }, | ||
| 1087 | + { INDEX_op_br, { } }, | ||
| 1088 | + { INDEX_op_mov_i32, { "r", "r" } }, | ||
| 1089 | + { INDEX_op_movi_i32, { "r" } }, | ||
| 1090 | + { INDEX_op_ld8u_i32, { "r", "r" } }, | ||
| 1091 | + { INDEX_op_ld8s_i32, { "r", "r" } }, | ||
| 1092 | + { INDEX_op_ld16u_i32, { "r", "r" } }, | ||
| 1093 | + { INDEX_op_ld16s_i32, { "r", "r" } }, | ||
| 1094 | + { INDEX_op_ld_i32, { "r", "r" } }, | ||
| 1095 | + { INDEX_op_st8_i32, { "q", "r" } }, | ||
| 1096 | + { INDEX_op_st16_i32, { "r", "r" } }, | ||
| 1097 | + { INDEX_op_st_i32, { "r", "r" } }, | ||
| 1098 | + | ||
| 1099 | + { INDEX_op_add_i32, { "r", "0", "ri" } }, | ||
| 1100 | + { INDEX_op_sub_i32, { "r", "0", "ri" } }, | ||
| 1101 | + { INDEX_op_mul_i32, { "r", "0", "ri" } }, | ||
| 1102 | + { INDEX_op_mulu2_i32, { "a", "d", "a", "r" } }, | ||
| 1103 | + { INDEX_op_div2_i32, { "a", "d", "0", "1", "r" } }, | ||
| 1104 | + { INDEX_op_divu2_i32, { "a", "d", "0", "1", "r" } }, | ||
| 1105 | + { INDEX_op_and_i32, { "r", "0", "ri" } }, | ||
| 1106 | + { INDEX_op_or_i32, { "r", "0", "ri" } }, | ||
| 1107 | + { INDEX_op_xor_i32, { "r", "0", "ri" } }, | ||
| 1108 | + | ||
| 1109 | + { INDEX_op_shl_i32, { "r", "0", "ci" } }, | ||
| 1110 | + { INDEX_op_shr_i32, { "r", "0", "ci" } }, | ||
| 1111 | + { INDEX_op_sar_i32, { "r", "0", "ci" } }, | ||
| 1112 | + | ||
| 1113 | + { INDEX_op_brcond_i32, { "r", "ri" } }, | ||
| 1114 | + | ||
| 1115 | + { INDEX_op_add2_i32, { "r", "r", "0", "1", "ri", "ri" } }, | ||
| 1116 | + { INDEX_op_sub2_i32, { "r", "r", "0", "1", "ri", "ri" } }, | ||
| 1117 | + { INDEX_op_brcond2_i32, { "r", "r", "ri", "ri" } }, | ||
| 1118 | + | ||
| 1119 | +#if TARGET_LONG_BITS == 32 | ||
| 1120 | + { INDEX_op_qemu_ld8u, { "r", "L" } }, | ||
| 1121 | + { INDEX_op_qemu_ld8s, { "r", "L" } }, | ||
| 1122 | + { INDEX_op_qemu_ld16u, { "r", "L" } }, | ||
| 1123 | + { INDEX_op_qemu_ld16s, { "r", "L" } }, | ||
| 1124 | + { INDEX_op_qemu_ld32u, { "r", "L" } }, | ||
| 1125 | + { INDEX_op_qemu_ld64, { "r", "r", "L" } }, | ||
| 1126 | + | ||
| 1127 | + { INDEX_op_qemu_st8, { "cb", "L" } }, | ||
| 1128 | + { INDEX_op_qemu_st16, { "L", "L" } }, | ||
| 1129 | + { INDEX_op_qemu_st32, { "L", "L" } }, | ||
| 1130 | + { INDEX_op_qemu_st64, { "L", "L", "L" } }, | ||
| 1131 | +#else | ||
| 1132 | + { INDEX_op_qemu_ld8u, { "r", "L", "L" } }, | ||
| 1133 | + { INDEX_op_qemu_ld8s, { "r", "L", "L" } }, | ||
| 1134 | + { INDEX_op_qemu_ld16u, { "r", "L", "L" } }, | ||
| 1135 | + { INDEX_op_qemu_ld16s, { "r", "L", "L" } }, | ||
| 1136 | + { INDEX_op_qemu_ld32u, { "r", "L", "L" } }, | ||
| 1137 | + { INDEX_op_qemu_ld64, { "r", "r", "L", "L" } }, | ||
| 1138 | + | ||
| 1139 | + { INDEX_op_qemu_st8, { "cb", "L", "L" } }, | ||
| 1140 | + { INDEX_op_qemu_st16, { "L", "L", "L" } }, | ||
| 1141 | + { INDEX_op_qemu_st32, { "L", "L", "L" } }, | ||
| 1142 | + { INDEX_op_qemu_st64, { "L", "L", "L", "L" } }, | ||
| 1143 | +#endif | ||
| 1144 | + { -1 }, | ||
| 1145 | +}; | ||
| 1146 | + | ||
| 1147 | +void tcg_target_init(TCGContext *s) | ||
| 1148 | +{ | ||
| 1149 | + /* fail safe */ | ||
| 1150 | + if ((1 << CPU_TLB_ENTRY_BITS) != sizeof(CPUTLBEntry)) | ||
| 1151 | + tcg_abort(); | ||
| 1152 | + | ||
| 1153 | + tcg_regset_set32(tcg_target_available_regs[TCG_TYPE_I32], 0, 0xff); | ||
| 1154 | + tcg_regset_set32(tcg_target_call_clobber_regs, 0, | ||
| 1155 | + (1 << TCG_REG_EAX) | | ||
| 1156 | + (1 << TCG_REG_EDX) | | ||
| 1157 | + (1 << TCG_REG_ECX)); | ||
| 1158 | + | ||
| 1159 | + tcg_regset_clear(s->reserved_regs); | ||
| 1160 | + tcg_regset_set_reg(s->reserved_regs, TCG_REG_ESP); | ||
| 1161 | + | ||
| 1162 | + tcg_add_target_add_op_defs(x86_op_defs); | ||
| 1163 | +} |
tcg/i386/tcg-target.h
0 โ 100644
| 1 | +/* | ||
| 2 | + * Tiny Code Generator for QEMU | ||
| 3 | + * | ||
| 4 | + * Copyright (c) 2008 Fabrice Bellard | ||
| 5 | + * | ||
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
| 7 | + * of this software and associated documentation files (the "Software"), to deal | ||
| 8 | + * in the Software without restriction, including without limitation the rights | ||
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
| 10 | + * copies of the Software, and to permit persons to whom the Software is | ||
| 11 | + * furnished to do so, subject to the following conditions: | ||
| 12 | + * | ||
| 13 | + * The above copyright notice and this permission notice shall be included in | ||
| 14 | + * all copies or substantial portions of the Software. | ||
| 15 | + * | ||
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
| 22 | + * THE SOFTWARE. | ||
| 23 | + */ | ||
| 24 | +#define TCG_TARGET_I386 1 | ||
| 25 | + | ||
| 26 | +#define TCG_TARGET_REG_BITS 32 | ||
| 27 | +//#define TCG_TARGET_WORDS_BIGENDIAN | ||
| 28 | + | ||
| 29 | +#define TCG_TARGET_NB_REGS 8 | ||
| 30 | + | ||
| 31 | +enum { | ||
| 32 | + TCG_REG_EAX = 0, | ||
| 33 | + TCG_REG_ECX, | ||
| 34 | + TCG_REG_EDX, | ||
| 35 | + TCG_REG_EBX, | ||
| 36 | + TCG_REG_ESP, | ||
| 37 | + TCG_REG_EBP, | ||
| 38 | + TCG_REG_ESI, | ||
| 39 | + TCG_REG_EDI, | ||
| 40 | +}; | ||
| 41 | + | ||
| 42 | +/* used for function call generation */ | ||
| 43 | +#define TCG_REG_CALL_STACK TCG_REG_ESP | ||
| 44 | +#define TCG_TARGET_STACK_ALIGN 16 | ||
| 45 | + | ||
| 46 | +/* Note: must be synced with dyngen-exec.h */ | ||
| 47 | +#define TCG_AREG0 TCG_REG_EBP | ||
| 48 | +#define TCG_AREG1 TCG_REG_EBX | ||
| 49 | +#define TCG_AREG2 TCG_REG_ESI | ||
| 50 | +#define TCG_AREG3 TCG_REG_EDI | ||
| 51 | + | ||
| 52 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
| 53 | +{ | ||
| 54 | +} |
tcg/tcg-dyngen.c
0 โ 100644
| 1 | +/* | ||
| 2 | + * Tiny Code Generator for QEMU | ||
| 3 | + * | ||
| 4 | + * Copyright (c) 2008 Fabrice Bellard | ||
| 5 | + * | ||
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
| 7 | + * of this software and associated documentation files (the "Software"), to deal | ||
| 8 | + * in the Software without restriction, including without limitation the rights | ||
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
| 10 | + * copies of the Software, and to permit persons to whom the Software is | ||
| 11 | + * furnished to do so, subject to the following conditions: | ||
| 12 | + * | ||
| 13 | + * The above copyright notice and this permission notice shall be included in | ||
| 14 | + * all copies or substantial portions of the Software. | ||
| 15 | + * | ||
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
| 22 | + * THE SOFTWARE. | ||
| 23 | + */ | ||
| 24 | +#include <assert.h> | ||
| 25 | +#include <stdarg.h> | ||
| 26 | +#include <stdlib.h> | ||
| 27 | +#include <stdio.h> | ||
| 28 | +#include <string.h> | ||
| 29 | +#include <inttypes.h> | ||
| 30 | + | ||
| 31 | +#include "config.h" | ||
| 32 | +#include "osdep.h" | ||
| 33 | + | ||
| 34 | +#include "tcg.h" | ||
| 35 | + | ||
| 36 | +int __op_param1, __op_param2, __op_param3; | ||
| 37 | +#if defined(__sparc__) || defined(__arm__) | ||
| 38 | + void __op_gen_label1(){} | ||
| 39 | + void __op_gen_label2(){} | ||
| 40 | + void __op_gen_label3(){} | ||
| 41 | +#else | ||
| 42 | + int __op_gen_label1, __op_gen_label2, __op_gen_label3; | ||
| 43 | +#endif | ||
| 44 | +int __op_jmp0, __op_jmp1, __op_jmp2, __op_jmp3; | ||
| 45 | + | ||
| 46 | +#if 0 | ||
| 47 | +#if defined(__s390__) | ||
| 48 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
| 49 | +{ | ||
| 50 | +} | ||
| 51 | +#elif defined(__ia64__) | ||
| 52 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
| 53 | +{ | ||
| 54 | + while (start < stop) { | ||
| 55 | + asm volatile ("fc %0" :: "r"(start)); | ||
| 56 | + start += 32; | ||
| 57 | + } | ||
| 58 | + asm volatile (";;sync.i;;srlz.i;;"); | ||
| 59 | +} | ||
| 60 | +#elif defined(__powerpc__) | ||
| 61 | + | ||
| 62 | +#define MIN_CACHE_LINE_SIZE 8 /* conservative value */ | ||
| 63 | + | ||
| 64 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
| 65 | +{ | ||
| 66 | + unsigned long p; | ||
| 67 | + | ||
| 68 | + start &= ~(MIN_CACHE_LINE_SIZE - 1); | ||
| 69 | + stop = (stop + MIN_CACHE_LINE_SIZE - 1) & ~(MIN_CACHE_LINE_SIZE - 1); | ||
| 70 | + | ||
| 71 | + for (p = start; p < stop; p += MIN_CACHE_LINE_SIZE) { | ||
| 72 | + asm volatile ("dcbst 0,%0" : : "r"(p) : "memory"); | ||
| 73 | + } | ||
| 74 | + asm volatile ("sync" : : : "memory"); | ||
| 75 | + for (p = start; p < stop; p += MIN_CACHE_LINE_SIZE) { | ||
| 76 | + asm volatile ("icbi 0,%0" : : "r"(p) : "memory"); | ||
| 77 | + } | ||
| 78 | + asm volatile ("sync" : : : "memory"); | ||
| 79 | + asm volatile ("isync" : : : "memory"); | ||
| 80 | +} | ||
| 81 | +#elif defined(__alpha__) | ||
| 82 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
| 83 | +{ | ||
| 84 | + asm ("imb"); | ||
| 85 | +} | ||
| 86 | +#elif defined(__sparc__) | ||
| 87 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
| 88 | +{ | ||
| 89 | + unsigned long p; | ||
| 90 | + | ||
| 91 | + p = start & ~(8UL - 1UL); | ||
| 92 | + stop = (stop + (8UL - 1UL)) & ~(8UL - 1UL); | ||
| 93 | + | ||
| 94 | + for (; p < stop; p += 8) | ||
| 95 | + __asm__ __volatile__("flush\t%0" : : "r" (p)); | ||
| 96 | +} | ||
| 97 | +#elif defined(__arm__) | ||
| 98 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
| 99 | +{ | ||
| 100 | + register unsigned long _beg __asm ("a1") = start; | ||
| 101 | + register unsigned long _end __asm ("a2") = stop; | ||
| 102 | + register unsigned long _flg __asm ("a3") = 0; | ||
| 103 | + __asm __volatile__ ("swi 0x9f0002" : : "r" (_beg), "r" (_end), "r" (_flg)); | ||
| 104 | +} | ||
| 105 | +#elif defined(__mc68000) | ||
| 106 | + | ||
| 107 | +# include <asm/cachectl.h> | ||
| 108 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
| 109 | +{ | ||
| 110 | + cacheflush(start,FLUSH_SCOPE_LINE,FLUSH_CACHE_BOTH,stop-start+16); | ||
| 111 | +} | ||
| 112 | +#elif defined(__mips__) | ||
| 113 | + | ||
| 114 | +#include <sys/cachectl.h> | ||
| 115 | +static inline void flush_icache_range(unsigned long start, unsigned long stop) | ||
| 116 | +{ | ||
| 117 | + _flush_cache ((void *)start, stop - start, BCACHE); | ||
| 118 | +} | ||
| 119 | +#else | ||
| 120 | +#error unsupported CPU | ||
| 121 | +#endif | ||
| 122 | + | ||
| 123 | +#ifdef __alpha__ | ||
| 124 | + | ||
| 125 | +register int gp asm("$29"); | ||
| 126 | + | ||
| 127 | +static inline void immediate_ldah(void *p, int val) { | ||
| 128 | + uint32_t *dest = p; | ||
| 129 | + long high = ((val >> 16) + ((val >> 15) & 1)) & 0xffff; | ||
| 130 | + | ||
| 131 | + *dest &= ~0xffff; | ||
| 132 | + *dest |= high; | ||
| 133 | + *dest |= 31 << 16; | ||
| 134 | +} | ||
| 135 | +static inline void immediate_lda(void *dest, int val) { | ||
| 136 | + *(uint16_t *) dest = val; | ||
| 137 | +} | ||
| 138 | +void fix_bsr(void *p, int offset) { | ||
| 139 | + uint32_t *dest = p; | ||
| 140 | + *dest &= ~((1 << 21) - 1); | ||
| 141 | + *dest |= (offset >> 2) & ((1 << 21) - 1); | ||
| 142 | +} | ||
| 143 | + | ||
| 144 | +#endif /* __alpha__ */ | ||
| 145 | + | ||
| 146 | +#ifdef __arm__ | ||
| 147 | + | ||
| 148 | +#define ARM_LDR_TABLE_SIZE 1024 | ||
| 149 | + | ||
| 150 | +typedef struct LDREntry { | ||
| 151 | + uint8_t *ptr; | ||
| 152 | + uint32_t *data_ptr; | ||
| 153 | + unsigned type:2; | ||
| 154 | +} LDREntry; | ||
| 155 | + | ||
| 156 | +static LDREntry arm_ldr_table[1024]; | ||
| 157 | +static uint32_t arm_data_table[ARM_LDR_TABLE_SIZE]; | ||
| 158 | + | ||
| 159 | +extern char exec_loop; | ||
| 160 | + | ||
| 161 | +static inline void arm_reloc_pc24(uint32_t *ptr, uint32_t insn, int val) | ||
| 162 | +{ | ||
| 163 | + *ptr = (insn & ~0xffffff) | ((insn + ((val - (int)ptr) >> 2)) & 0xffffff); | ||
| 164 | +} | ||
| 165 | + | ||
| 166 | +static uint8_t *arm_flush_ldr(uint8_t *gen_code_ptr, | ||
| 167 | + LDREntry *ldr_start, LDREntry *ldr_end, | ||
| 168 | + uint32_t *data_start, uint32_t *data_end, | ||
| 169 | + int gen_jmp) | ||
| 170 | +{ | ||
| 171 | + LDREntry *le; | ||
| 172 | + uint32_t *ptr; | ||
| 173 | + int offset, data_size, target; | ||
| 174 | + uint8_t *data_ptr; | ||
| 175 | + uint32_t insn; | ||
| 176 | + uint32_t mask; | ||
| 177 | + | ||
| 178 | + data_size = (data_end - data_start) << 2; | ||
| 179 | + | ||
| 180 | + if (gen_jmp) { | ||
| 181 | + /* generate branch to skip the data */ | ||
| 182 | + if (data_size == 0) | ||
| 183 | + return gen_code_ptr; | ||
| 184 | + target = (long)gen_code_ptr + data_size + 4; | ||
| 185 | + arm_reloc_pc24((uint32_t *)gen_code_ptr, 0xeafffffe, target); | ||
| 186 | + gen_code_ptr += 4; | ||
| 187 | + } | ||
| 188 | + | ||
| 189 | + /* copy the data */ | ||
| 190 | + data_ptr = gen_code_ptr; | ||
| 191 | + memcpy(gen_code_ptr, data_start, data_size); | ||
| 192 | + gen_code_ptr += data_size; | ||
| 193 | + | ||
| 194 | + /* patch the ldr to point to the data */ | ||
| 195 | + for(le = ldr_start; le < ldr_end; le++) { | ||
| 196 | + ptr = (uint32_t *)le->ptr; | ||
| 197 | + offset = ((unsigned long)(le->data_ptr) - (unsigned long)data_start) + | ||
| 198 | + (unsigned long)data_ptr - | ||
| 199 | + (unsigned long)ptr - 8; | ||
| 200 | + if (offset < 0) { | ||
| 201 | + fprintf(stderr, "Negative constant pool offset\n"); | ||
| 202 | + tcg_abort(); | ||
| 203 | + } | ||
| 204 | + switch (le->type) { | ||
| 205 | + case 0: /* ldr */ | ||
| 206 | + mask = ~0x00800fff; | ||
| 207 | + if (offset >= 4096) { | ||
| 208 | + fprintf(stderr, "Bad ldr offset\n"); | ||
| 209 | + tcg_abort(); | ||
| 210 | + } | ||
| 211 | + break; | ||
| 212 | + case 1: /* ldc */ | ||
| 213 | + mask = ~0x008000ff; | ||
| 214 | + if (offset >= 1024 ) { | ||
| 215 | + fprintf(stderr, "Bad ldc offset\n"); | ||
| 216 | + tcg_abort(); | ||
| 217 | + } | ||
| 218 | + break; | ||
| 219 | + case 2: /* add */ | ||
| 220 | + mask = ~0xfff; | ||
| 221 | + if (offset >= 1024 ) { | ||
| 222 | + fprintf(stderr, "Bad add offset\n"); | ||
| 223 | + tcg_abort(); | ||
| 224 | + } | ||
| 225 | + break; | ||
| 226 | + default: | ||
| 227 | + fprintf(stderr, "Bad pc relative fixup\n"); | ||
| 228 | + tcg_abort(); | ||
| 229 | + } | ||
| 230 | + insn = *ptr & mask; | ||
| 231 | + switch (le->type) { | ||
| 232 | + case 0: /* ldr */ | ||
| 233 | + insn |= offset | 0x00800000; | ||
| 234 | + break; | ||
| 235 | + case 1: /* ldc */ | ||
| 236 | + insn |= (offset >> 2) | 0x00800000; | ||
| 237 | + break; | ||
| 238 | + case 2: /* add */ | ||
| 239 | + insn |= (offset >> 2) | 0xf00; | ||
| 240 | + break; | ||
| 241 | + } | ||
| 242 | + *ptr = insn; | ||
| 243 | + } | ||
| 244 | + return gen_code_ptr; | ||
| 245 | +} | ||
| 246 | + | ||
| 247 | +#endif /* __arm__ */ | ||
| 248 | + | ||
| 249 | +#ifdef __ia64 | ||
| 250 | + | ||
| 251 | +/* Patch instruction with "val" where "mask" has 1 bits. */ | ||
| 252 | +static inline void ia64_patch (uint64_t insn_addr, uint64_t mask, uint64_t val) | ||
| 253 | +{ | ||
| 254 | + uint64_t m0, m1, v0, v1, b0, b1, *b = (uint64_t *) (insn_addr & -16); | ||
| 255 | +# define insn_mask ((1UL << 41) - 1) | ||
| 256 | + unsigned long shift; | ||
| 257 | + | ||
| 258 | + b0 = b[0]; b1 = b[1]; | ||
| 259 | + shift = 5 + 41 * (insn_addr % 16); /* 5 template, 3 x 41-bit insns */ | ||
| 260 | + if (shift >= 64) { | ||
| 261 | + m1 = mask << (shift - 64); | ||
| 262 | + v1 = val << (shift - 64); | ||
| 263 | + } else { | ||
| 264 | + m0 = mask << shift; m1 = mask >> (64 - shift); | ||
| 265 | + v0 = val << shift; v1 = val >> (64 - shift); | ||
| 266 | + b[0] = (b0 & ~m0) | (v0 & m0); | ||
| 267 | + } | ||
| 268 | + b[1] = (b1 & ~m1) | (v1 & m1); | ||
| 269 | +} | ||
| 270 | + | ||
| 271 | +static inline void ia64_patch_imm60 (uint64_t insn_addr, uint64_t val) | ||
| 272 | +{ | ||
| 273 | + ia64_patch(insn_addr, | ||
| 274 | + 0x011ffffe000UL, | ||
| 275 | + ( ((val & 0x0800000000000000UL) >> 23) /* bit 59 -> 36 */ | ||
| 276 | + | ((val & 0x00000000000fffffUL) << 13) /* bit 0 -> 13 */)); | ||
| 277 | + ia64_patch(insn_addr - 1, 0x1fffffffffcUL, val >> 18); | ||
| 278 | +} | ||
| 279 | + | ||
| 280 | +static inline void ia64_imm64 (void *insn, uint64_t val) | ||
| 281 | +{ | ||
| 282 | + /* Ignore the slot number of the relocation; GCC and Intel | ||
| 283 | + toolchains differed for some time on whether IMM64 relocs are | ||
| 284 | + against slot 1 (Intel) or slot 2 (GCC). */ | ||
| 285 | + uint64_t insn_addr = (uint64_t) insn & ~3UL; | ||
| 286 | + | ||
| 287 | + ia64_patch(insn_addr + 2, | ||
| 288 | + 0x01fffefe000UL, | ||
| 289 | + ( ((val & 0x8000000000000000UL) >> 27) /* bit 63 -> 36 */ | ||
| 290 | + | ((val & 0x0000000000200000UL) << 0) /* bit 21 -> 21 */ | ||
| 291 | + | ((val & 0x00000000001f0000UL) << 6) /* bit 16 -> 22 */ | ||
| 292 | + | ((val & 0x000000000000ff80UL) << 20) /* bit 7 -> 27 */ | ||
| 293 | + | ((val & 0x000000000000007fUL) << 13) /* bit 0 -> 13 */) | ||
| 294 | + ); | ||
| 295 | + ia64_patch(insn_addr + 1, 0x1ffffffffffUL, val >> 22); | ||
| 296 | +} | ||
| 297 | + | ||
| 298 | +static inline void ia64_imm60b (void *insn, uint64_t val) | ||
| 299 | +{ | ||
| 300 | + /* Ignore the slot number of the relocation; GCC and Intel | ||
| 301 | + toolchains differed for some time on whether IMM64 relocs are | ||
| 302 | + against slot 1 (Intel) or slot 2 (GCC). */ | ||
| 303 | + uint64_t insn_addr = (uint64_t) insn & ~3UL; | ||
| 304 | + | ||
| 305 | + if (val + ((uint64_t) 1 << 59) >= (1UL << 60)) | ||
| 306 | + fprintf(stderr, "%s: value %ld out of IMM60 range\n", | ||
| 307 | + __FUNCTION__, (int64_t) val); | ||
| 308 | + ia64_patch_imm60(insn_addr + 2, val); | ||
| 309 | +} | ||
| 310 | + | ||
| 311 | +static inline void ia64_imm22 (void *insn, uint64_t val) | ||
| 312 | +{ | ||
| 313 | + if (val + (1 << 21) >= (1 << 22)) | ||
| 314 | + fprintf(stderr, "%s: value %li out of IMM22 range\n", | ||
| 315 | + __FUNCTION__, (int64_t)val); | ||
| 316 | + ia64_patch((uint64_t) insn, 0x01fffcfe000UL, | ||
| 317 | + ( ((val & 0x200000UL) << 15) /* bit 21 -> 36 */ | ||
| 318 | + | ((val & 0x1f0000UL) << 6) /* bit 16 -> 22 */ | ||
| 319 | + | ((val & 0x00ff80UL) << 20) /* bit 7 -> 27 */ | ||
| 320 | + | ((val & 0x00007fUL) << 13) /* bit 0 -> 13 */)); | ||
| 321 | +} | ||
| 322 | + | ||
| 323 | +/* Like ia64_imm22(), but also clear bits 20-21. For addl, this has | ||
| 324 | + the effect of turning "addl rX=imm22,rY" into "addl | ||
| 325 | + rX=imm22,r0". */ | ||
| 326 | +static inline void ia64_imm22_r0 (void *insn, uint64_t val) | ||
| 327 | +{ | ||
| 328 | + if (val + (1 << 21) >= (1 << 22)) | ||
| 329 | + fprintf(stderr, "%s: value %li out of IMM22 range\n", | ||
| 330 | + __FUNCTION__, (int64_t)val); | ||
| 331 | + ia64_patch((uint64_t) insn, 0x01fffcfe000UL | (0x3UL << 20), | ||
| 332 | + ( ((val & 0x200000UL) << 15) /* bit 21 -> 36 */ | ||
| 333 | + | ((val & 0x1f0000UL) << 6) /* bit 16 -> 22 */ | ||
| 334 | + | ((val & 0x00ff80UL) << 20) /* bit 7 -> 27 */ | ||
| 335 | + | ((val & 0x00007fUL) << 13) /* bit 0 -> 13 */)); | ||
| 336 | +} | ||
| 337 | + | ||
| 338 | +static inline void ia64_imm21b (void *insn, uint64_t val) | ||
| 339 | +{ | ||
| 340 | + if (val + (1 << 20) >= (1 << 21)) | ||
| 341 | + fprintf(stderr, "%s: value %li out of IMM21b range\n", | ||
| 342 | + __FUNCTION__, (int64_t)val); | ||
| 343 | + ia64_patch((uint64_t) insn, 0x11ffffe000UL, | ||
| 344 | + ( ((val & 0x100000UL) << 16) /* bit 20 -> 36 */ | ||
| 345 | + | ((val & 0x0fffffUL) << 13) /* bit 0 -> 13 */)); | ||
| 346 | +} | ||
| 347 | + | ||
| 348 | +static inline void ia64_nop_b (void *insn) | ||
| 349 | +{ | ||
| 350 | + ia64_patch((uint64_t) insn, (1UL << 41) - 1, 2UL << 37); | ||
| 351 | +} | ||
| 352 | + | ||
| 353 | +static inline void ia64_ldxmov(void *insn, uint64_t val) | ||
| 354 | +{ | ||
| 355 | + if (val + (1 << 21) < (1 << 22)) | ||
| 356 | + ia64_patch((uint64_t) insn, 0x1fff80fe000UL, 8UL << 37); | ||
| 357 | +} | ||
| 358 | + | ||
| 359 | +static inline int ia64_patch_ltoff(void *insn, uint64_t val, | ||
| 360 | + int relaxable) | ||
| 361 | +{ | ||
| 362 | + if (relaxable && (val + (1 << 21) < (1 << 22))) { | ||
| 363 | + ia64_imm22_r0(insn, val); | ||
| 364 | + return 0; | ||
| 365 | + } | ||
| 366 | + return 1; | ||
| 367 | +} | ||
| 368 | + | ||
| 369 | +struct ia64_fixup { | ||
| 370 | + struct ia64_fixup *next; | ||
| 371 | + void *addr; /* address that needs to be patched */ | ||
| 372 | + long value; | ||
| 373 | +}; | ||
| 374 | + | ||
| 375 | +#define IA64_PLT(insn, plt_index) \ | ||
| 376 | +do { \ | ||
| 377 | + struct ia64_fixup *fixup = alloca(sizeof(*fixup)); \ | ||
| 378 | + fixup->next = plt_fixes; \ | ||
| 379 | + plt_fixes = fixup; \ | ||
| 380 | + fixup->addr = (insn); \ | ||
| 381 | + fixup->value = (plt_index); \ | ||
| 382 | + plt_offset[(plt_index)] = 1; \ | ||
| 383 | +} while (0) | ||
| 384 | + | ||
| 385 | +#define IA64_LTOFF(insn, val, relaxable) \ | ||
| 386 | +do { \ | ||
| 387 | + if (ia64_patch_ltoff(insn, val, relaxable)) { \ | ||
| 388 | + struct ia64_fixup *fixup = alloca(sizeof(*fixup)); \ | ||
| 389 | + fixup->next = ltoff_fixes; \ | ||
| 390 | + ltoff_fixes = fixup; \ | ||
| 391 | + fixup->addr = (insn); \ | ||
| 392 | + fixup->value = (val); \ | ||
| 393 | + } \ | ||
| 394 | +} while (0) | ||
| 395 | + | ||
| 396 | +static inline void ia64_apply_fixes (uint8_t **gen_code_pp, | ||
| 397 | + struct ia64_fixup *ltoff_fixes, | ||
| 398 | + uint64_t gp, | ||
| 399 | + struct ia64_fixup *plt_fixes, | ||
| 400 | + int num_plts, | ||
| 401 | + unsigned long *plt_target, | ||
| 402 | + unsigned int *plt_offset) | ||
| 403 | +{ | ||
| 404 | + static const uint8_t plt_bundle[] = { | ||
| 405 | + 0x04, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, /* nop 0; movl r1=GP */ | ||
| 406 | + 0x00, 0x00, 0x00, 0x20, 0x00, 0x00, 0x00, 0x60, | ||
| 407 | + | ||
| 408 | + 0x05, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, /* nop 0; brl IP */ | ||
| 409 | + 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0xc0 | ||
| 410 | + }; | ||
| 411 | + uint8_t *gen_code_ptr = *gen_code_pp, *plt_start, *got_start; | ||
| 412 | + uint64_t *vp; | ||
| 413 | + struct ia64_fixup *fixup; | ||
| 414 | + unsigned int offset = 0; | ||
| 415 | + struct fdesc { | ||
| 416 | + long ip; | ||
| 417 | + long gp; | ||
| 418 | + } *fdesc; | ||
| 419 | + int i; | ||
| 420 | + | ||
| 421 | + if (plt_fixes) { | ||
| 422 | + plt_start = gen_code_ptr; | ||
| 423 | + | ||
| 424 | + for (i = 0; i < num_plts; ++i) { | ||
| 425 | + if (plt_offset[i]) { | ||
| 426 | + plt_offset[i] = offset; | ||
| 427 | + offset += sizeof(plt_bundle); | ||
| 428 | + | ||
| 429 | + fdesc = (struct fdesc *) plt_target[i]; | ||
| 430 | + memcpy(gen_code_ptr, plt_bundle, sizeof(plt_bundle)); | ||
| 431 | + ia64_imm64 (gen_code_ptr + 0x02, fdesc->gp); | ||
| 432 | + ia64_imm60b(gen_code_ptr + 0x12, | ||
| 433 | + (fdesc->ip - (long) (gen_code_ptr + 0x10)) >> 4); | ||
| 434 | + gen_code_ptr += sizeof(plt_bundle); | ||
| 435 | + } | ||
| 436 | + } | ||
| 437 | + | ||
| 438 | + for (fixup = plt_fixes; fixup; fixup = fixup->next) | ||
| 439 | + ia64_imm21b(fixup->addr, | ||
| 440 | + ((long) plt_start + plt_offset[fixup->value] | ||
| 441 | + - ((long) fixup->addr & ~0xf)) >> 4); | ||
| 442 | + } | ||
| 443 | + | ||
| 444 | + got_start = gen_code_ptr; | ||
| 445 | + | ||
| 446 | + /* First, create the GOT: */ | ||
| 447 | + for (fixup = ltoff_fixes; fixup; fixup = fixup->next) { | ||
| 448 | + /* first check if we already have this value in the GOT: */ | ||
| 449 | + for (vp = (uint64_t *) got_start; vp < (uint64_t *) gen_code_ptr; ++vp) | ||
| 450 | + if (*vp == fixup->value) | ||
| 451 | + break; | ||
| 452 | + if (vp == (uint64_t *) gen_code_ptr) { | ||
| 453 | + /* Nope, we need to put the value in the GOT: */ | ||
| 454 | + *vp = fixup->value; | ||
| 455 | + gen_code_ptr += 8; | ||
| 456 | + } | ||
| 457 | + ia64_imm22(fixup->addr, (long) vp - gp); | ||
| 458 | + } | ||
| 459 | + /* Keep code ptr aligned. */ | ||
| 460 | + if ((long) gen_code_ptr & 15) | ||
| 461 | + gen_code_ptr += 8; | ||
| 462 | + *gen_code_pp = gen_code_ptr; | ||
| 463 | +} | ||
| 464 | +#endif | ||
| 465 | +#endif | ||
| 466 | + | ||
| 467 | +const TCGArg *dyngen_op(TCGContext *s, int opc, const TCGArg *opparam_ptr) | ||
| 468 | +{ | ||
| 469 | + uint8_t *gen_code_ptr; | ||
| 470 | + | ||
| 471 | + gen_code_ptr = s->code_ptr; | ||
| 472 | + switch(opc) { | ||
| 473 | + | ||
| 474 | +/* op.h is dynamically generated by dyngen.c from op.c */ | ||
| 475 | +#include "op.h" | ||
| 476 | + | ||
| 477 | + default: | ||
| 478 | + tcg_abort(); | ||
| 479 | + } | ||
| 480 | + s->code_ptr = gen_code_ptr; | ||
| 481 | + return opparam_ptr; | ||
| 482 | +} |
tcg/tcg-op.h
0 โ 100644
| 1 | +/* | ||
| 2 | + * Tiny Code Generator for QEMU | ||
| 3 | + * | ||
| 4 | + * Copyright (c) 2008 Fabrice Bellard | ||
| 5 | + * | ||
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
| 7 | + * of this software and associated documentation files (the "Software"), to deal | ||
| 8 | + * in the Software without restriction, including without limitation the rights | ||
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
| 10 | + * copies of the Software, and to permit persons to whom the Software is | ||
| 11 | + * furnished to do so, subject to the following conditions: | ||
| 12 | + * | ||
| 13 | + * The above copyright notice and this permission notice shall be included in | ||
| 14 | + * all copies or substantial portions of the Software. | ||
| 15 | + * | ||
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
| 22 | + * THE SOFTWARE. | ||
| 23 | + */ | ||
| 24 | +#include "tcg.h" | ||
| 25 | + | ||
| 26 | +/* legacy dyngen operations */ | ||
| 27 | +#include "gen-op.h" | ||
| 28 | + | ||
| 29 | +int gen_new_label(void); | ||
| 30 | + | ||
| 31 | +static inline void tcg_gen_op1(int opc, TCGArg arg1) | ||
| 32 | +{ | ||
| 33 | + *gen_opc_ptr++ = opc; | ||
| 34 | + *gen_opparam_ptr++ = arg1; | ||
| 35 | +} | ||
| 36 | + | ||
| 37 | +static inline void tcg_gen_op2(int opc, TCGArg arg1, TCGArg arg2) | ||
| 38 | +{ | ||
| 39 | + *gen_opc_ptr++ = opc; | ||
| 40 | + *gen_opparam_ptr++ = arg1; | ||
| 41 | + *gen_opparam_ptr++ = arg2; | ||
| 42 | +} | ||
| 43 | + | ||
| 44 | +static inline void tcg_gen_op3(int opc, TCGArg arg1, TCGArg arg2, TCGArg arg3) | ||
| 45 | +{ | ||
| 46 | + *gen_opc_ptr++ = opc; | ||
| 47 | + *gen_opparam_ptr++ = arg1; | ||
| 48 | + *gen_opparam_ptr++ = arg2; | ||
| 49 | + *gen_opparam_ptr++ = arg3; | ||
| 50 | +} | ||
| 51 | + | ||
| 52 | +static inline void tcg_gen_op4(int opc, TCGArg arg1, TCGArg arg2, TCGArg arg3, | ||
| 53 | + TCGArg arg4) | ||
| 54 | +{ | ||
| 55 | + *gen_opc_ptr++ = opc; | ||
| 56 | + *gen_opparam_ptr++ = arg1; | ||
| 57 | + *gen_opparam_ptr++ = arg2; | ||
| 58 | + *gen_opparam_ptr++ = arg3; | ||
| 59 | + *gen_opparam_ptr++ = arg4; | ||
| 60 | +} | ||
| 61 | + | ||
| 62 | +static inline void tcg_gen_op5(int opc, TCGArg arg1, TCGArg arg2, | ||
| 63 | + TCGArg arg3, TCGArg arg4, | ||
| 64 | + TCGArg arg5) | ||
| 65 | +{ | ||
| 66 | + *gen_opc_ptr++ = opc; | ||
| 67 | + *gen_opparam_ptr++ = arg1; | ||
| 68 | + *gen_opparam_ptr++ = arg2; | ||
| 69 | + *gen_opparam_ptr++ = arg3; | ||
| 70 | + *gen_opparam_ptr++ = arg4; | ||
| 71 | + *gen_opparam_ptr++ = arg5; | ||
| 72 | +} | ||
| 73 | + | ||
| 74 | +static inline void tcg_gen_op6(int opc, TCGArg arg1, TCGArg arg2, | ||
| 75 | + TCGArg arg3, TCGArg arg4, | ||
| 76 | + TCGArg arg5, TCGArg arg6) | ||
| 77 | +{ | ||
| 78 | + *gen_opc_ptr++ = opc; | ||
| 79 | + *gen_opparam_ptr++ = arg1; | ||
| 80 | + *gen_opparam_ptr++ = arg2; | ||
| 81 | + *gen_opparam_ptr++ = arg3; | ||
| 82 | + *gen_opparam_ptr++ = arg4; | ||
| 83 | + *gen_opparam_ptr++ = arg5; | ||
| 84 | + *gen_opparam_ptr++ = arg6; | ||
| 85 | +} | ||
| 86 | + | ||
| 87 | +static inline void gen_set_label(int n) | ||
| 88 | +{ | ||
| 89 | + tcg_gen_op1(INDEX_op_set_label, n); | ||
| 90 | +} | ||
| 91 | + | ||
| 92 | +static inline void tcg_gen_mov_i32(int ret, int arg) | ||
| 93 | +{ | ||
| 94 | + tcg_gen_op2(INDEX_op_mov_i32, ret, arg); | ||
| 95 | +} | ||
| 96 | + | ||
| 97 | +static inline void tcg_gen_movi_i32(int ret, int32_t arg) | ||
| 98 | +{ | ||
| 99 | + tcg_gen_op2(INDEX_op_movi_i32, ret, arg); | ||
| 100 | +} | ||
| 101 | + | ||
| 102 | +/* helper calls */ | ||
| 103 | +#define TCG_HELPER_CALL_FLAGS 0 | ||
| 104 | + | ||
| 105 | +static inline void tcg_gen_helper_0_0(void *func) | ||
| 106 | +{ | ||
| 107 | + tcg_gen_call(&tcg_ctx, | ||
| 108 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | ||
| 109 | + 0, NULL, 0, NULL); | ||
| 110 | +} | ||
| 111 | + | ||
| 112 | +static inline void tcg_gen_helper_0_1(void *func, TCGArg arg) | ||
| 113 | +{ | ||
| 114 | + tcg_gen_call(&tcg_ctx, | ||
| 115 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | ||
| 116 | + 0, NULL, 1, &arg); | ||
| 117 | +} | ||
| 118 | + | ||
| 119 | +static inline void tcg_gen_helper_0_2(void *func, TCGArg arg1, TCGArg arg2) | ||
| 120 | +{ | ||
| 121 | + TCGArg args[2]; | ||
| 122 | + args[0] = arg1; | ||
| 123 | + args[1] = arg2; | ||
| 124 | + tcg_gen_call(&tcg_ctx, | ||
| 125 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | ||
| 126 | + 0, NULL, 2, args); | ||
| 127 | +} | ||
| 128 | + | ||
| 129 | +static inline void tcg_gen_helper_1_2(void *func, TCGArg ret, | ||
| 130 | + TCGArg arg1, TCGArg arg2) | ||
| 131 | +{ | ||
| 132 | + TCGArg args[2]; | ||
| 133 | + args[0] = arg1; | ||
| 134 | + args[1] = arg2; | ||
| 135 | + tcg_gen_call(&tcg_ctx, | ||
| 136 | + tcg_const_ptr((tcg_target_long)func), TCG_HELPER_CALL_FLAGS, | ||
| 137 | + 1, &ret, 2, args); | ||
| 138 | +} | ||
| 139 | + | ||
| 140 | +/* 32 bit ops */ | ||
| 141 | + | ||
| 142 | +static inline void tcg_gen_ld8u_i32(int ret, int arg2, tcg_target_long offset) | ||
| 143 | +{ | ||
| 144 | + tcg_gen_op3(INDEX_op_ld8u_i32, ret, arg2, offset); | ||
| 145 | +} | ||
| 146 | + | ||
| 147 | +static inline void tcg_gen_ld8s_i32(int ret, int arg2, tcg_target_long offset) | ||
| 148 | +{ | ||
| 149 | + tcg_gen_op3(INDEX_op_ld8s_i32, ret, arg2, offset); | ||
| 150 | +} | ||
| 151 | + | ||
| 152 | +static inline void tcg_gen_ld16u_i32(int ret, int arg2, tcg_target_long offset) | ||
| 153 | +{ | ||
| 154 | + tcg_gen_op3(INDEX_op_ld16u_i32, ret, arg2, offset); | ||
| 155 | +} | ||
| 156 | + | ||
| 157 | +static inline void tcg_gen_ld16s_i32(int ret, int arg2, tcg_target_long offset) | ||
| 158 | +{ | ||
| 159 | + tcg_gen_op3(INDEX_op_ld16s_i32, ret, arg2, offset); | ||
| 160 | +} | ||
| 161 | + | ||
| 162 | +static inline void tcg_gen_ld_i32(int ret, int arg2, tcg_target_long offset) | ||
| 163 | +{ | ||
| 164 | + tcg_gen_op3(INDEX_op_ld_i32, ret, arg2, offset); | ||
| 165 | +} | ||
| 166 | + | ||
| 167 | +static inline void tcg_gen_st8_i32(int arg1, int arg2, tcg_target_long offset) | ||
| 168 | +{ | ||
| 169 | + tcg_gen_op3(INDEX_op_st8_i32, arg1, arg2, offset); | ||
| 170 | +} | ||
| 171 | + | ||
| 172 | +static inline void tcg_gen_st16_i32(int arg1, int arg2, tcg_target_long offset) | ||
| 173 | +{ | ||
| 174 | + tcg_gen_op3(INDEX_op_st16_i32, arg1, arg2, offset); | ||
| 175 | +} | ||
| 176 | + | ||
| 177 | +static inline void tcg_gen_st_i32(int arg1, int arg2, tcg_target_long offset) | ||
| 178 | +{ | ||
| 179 | + tcg_gen_op3(INDEX_op_st_i32, arg1, arg2, offset); | ||
| 180 | +} | ||
| 181 | + | ||
| 182 | +static inline void tcg_gen_add_i32(int ret, int arg1, int arg2) | ||
| 183 | +{ | ||
| 184 | + tcg_gen_op3(INDEX_op_add_i32, ret, arg1, arg2); | ||
| 185 | +} | ||
| 186 | + | ||
| 187 | +static inline void tcg_gen_addi_i32(int ret, int arg1, int32_t arg2) | ||
| 188 | +{ | ||
| 189 | + tcg_gen_add_i32(ret, arg1, tcg_const_i32(arg2)); | ||
| 190 | +} | ||
| 191 | + | ||
| 192 | +static inline void tcg_gen_sub_i32(int ret, int arg1, int arg2) | ||
| 193 | +{ | ||
| 194 | + tcg_gen_op3(INDEX_op_sub_i32, ret, arg1, arg2); | ||
| 195 | +} | ||
| 196 | + | ||
| 197 | +static inline void tcg_gen_subi_i32(int ret, int arg1, int32_t arg2) | ||
| 198 | +{ | ||
| 199 | + tcg_gen_sub_i32(ret, arg1, tcg_const_i32(arg2)); | ||
| 200 | +} | ||
| 201 | + | ||
| 202 | +static inline void tcg_gen_and_i32(int ret, int arg1, int arg2) | ||
| 203 | +{ | ||
| 204 | + tcg_gen_op3(INDEX_op_and_i32, ret, arg1, arg2); | ||
| 205 | +} | ||
| 206 | + | ||
| 207 | +static inline void tcg_gen_andi_i32(int ret, int arg1, int32_t arg2) | ||
| 208 | +{ | ||
| 209 | + /* some cases can be optimized here */ | ||
| 210 | + if (arg2 == 0) { | ||
| 211 | + tcg_gen_movi_i32(ret, 0); | ||
| 212 | + } else if (arg2 == 0xffffffff) { | ||
| 213 | + tcg_gen_mov_i32(ret, arg1); | ||
| 214 | + } else { | ||
| 215 | + tcg_gen_and_i32(ret, arg1, tcg_const_i32(arg2)); | ||
| 216 | + } | ||
| 217 | +} | ||
| 218 | + | ||
| 219 | +static inline void tcg_gen_or_i32(int ret, int arg1, int arg2) | ||
| 220 | +{ | ||
| 221 | + tcg_gen_op3(INDEX_op_or_i32, ret, arg1, arg2); | ||
| 222 | +} | ||
| 223 | + | ||
| 224 | +static inline void tcg_gen_ori_i32(int ret, int arg1, int32_t arg2) | ||
| 225 | +{ | ||
| 226 | + /* some cases can be optimized here */ | ||
| 227 | + if (arg2 == 0xffffffff) { | ||
| 228 | + tcg_gen_movi_i32(ret, 0); | ||
| 229 | + } else if (arg2 == 0) { | ||
| 230 | + tcg_gen_mov_i32(ret, arg1); | ||
| 231 | + } else { | ||
| 232 | + tcg_gen_or_i32(ret, arg1, tcg_const_i32(arg2)); | ||
| 233 | + } | ||
| 234 | +} | ||
| 235 | + | ||
| 236 | +static inline void tcg_gen_xor_i32(int ret, int arg1, int arg2) | ||
| 237 | +{ | ||
| 238 | + tcg_gen_op3(INDEX_op_xor_i32, ret, arg1, arg2); | ||
| 239 | +} | ||
| 240 | + | ||
| 241 | +static inline void tcg_gen_xori_i32(int ret, int arg1, int32_t arg2) | ||
| 242 | +{ | ||
| 243 | + /* some cases can be optimized here */ | ||
| 244 | + if (arg2 == 0) { | ||
| 245 | + tcg_gen_mov_i32(ret, arg1); | ||
| 246 | + } else { | ||
| 247 | + tcg_gen_xor_i32(ret, arg1, tcg_const_i32(arg2)); | ||
| 248 | + } | ||
| 249 | +} | ||
| 250 | + | ||
| 251 | +static inline void tcg_gen_shl_i32(int ret, int arg1, int arg2) | ||
| 252 | +{ | ||
| 253 | + tcg_gen_op3(INDEX_op_shl_i32, ret, arg1, arg2); | ||
| 254 | +} | ||
| 255 | + | ||
| 256 | +static inline void tcg_gen_shli_i32(int ret, int arg1, int32_t arg2) | ||
| 257 | +{ | ||
| 258 | + tcg_gen_shl_i32(ret, arg1, tcg_const_i32(arg2)); | ||
| 259 | +} | ||
| 260 | + | ||
| 261 | +static inline void tcg_gen_shr_i32(int ret, int arg1, int arg2) | ||
| 262 | +{ | ||
| 263 | + tcg_gen_op3(INDEX_op_shr_i32, ret, arg1, arg2); | ||
| 264 | +} | ||
| 265 | + | ||
| 266 | +static inline void tcg_gen_shri_i32(int ret, int arg1, int32_t arg2) | ||
| 267 | +{ | ||
| 268 | + tcg_gen_shr_i32(ret, arg1, tcg_const_i32(arg2)); | ||
| 269 | +} | ||
| 270 | + | ||
| 271 | +static inline void tcg_gen_sar_i32(int ret, int arg1, int arg2) | ||
| 272 | +{ | ||
| 273 | + tcg_gen_op3(INDEX_op_sar_i32, ret, arg1, arg2); | ||
| 274 | +} | ||
| 275 | + | ||
| 276 | +static inline void tcg_gen_sari_i32(int ret, int arg1, int32_t arg2) | ||
| 277 | +{ | ||
| 278 | + tcg_gen_sar_i32(ret, arg1, tcg_const_i32(arg2)); | ||
| 279 | +} | ||
| 280 | + | ||
| 281 | +static inline void tcg_gen_brcond_i32(int cond, TCGArg arg1, TCGArg arg2, | ||
| 282 | + int label_index) | ||
| 283 | +{ | ||
| 284 | + tcg_gen_op4(INDEX_op_brcond_i32, arg1, arg2, cond, label_index); | ||
| 285 | +} | ||
| 286 | + | ||
| 287 | +static inline void tcg_gen_mul_i32(int ret, int arg1, int arg2) | ||
| 288 | +{ | ||
| 289 | + tcg_gen_op3(INDEX_op_mul_i32, ret, arg1, arg2); | ||
| 290 | +} | ||
| 291 | + | ||
| 292 | +#ifdef TCG_TARGET_HAS_div_i32 | ||
| 293 | +static inline void tcg_gen_div_i32(int ret, int arg1, int arg2) | ||
| 294 | +{ | ||
| 295 | + tcg_gen_op3(INDEX_op_div_i32, ret, arg1, arg2); | ||
| 296 | +} | ||
| 297 | + | ||
| 298 | +static inline void tcg_gen_rem_i32(int ret, int arg1, int arg2) | ||
| 299 | +{ | ||
| 300 | + tcg_gen_op3(INDEX_op_rem_i32, ret, arg1, arg2); | ||
| 301 | +} | ||
| 302 | + | ||
| 303 | +static inline void tcg_gen_divu_i32(int ret, int arg1, int arg2) | ||
| 304 | +{ | ||
| 305 | + tcg_gen_op3(INDEX_op_divu_i32, ret, arg1, arg2); | ||
| 306 | +} | ||
| 307 | + | ||
| 308 | +static inline void tcg_gen_remu_i32(int ret, int arg1, int arg2) | ||
| 309 | +{ | ||
| 310 | + tcg_gen_op3(INDEX_op_remu_i32, ret, arg1, arg2); | ||
| 311 | +} | ||
| 312 | +#else | ||
| 313 | +static inline void tcg_gen_div_i32(int ret, int arg1, int arg2) | ||
| 314 | +{ | ||
| 315 | + int t0; | ||
| 316 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
| 317 | + tcg_gen_sari_i32(t0, arg1, 31); | ||
| 318 | + tcg_gen_op5(INDEX_op_div2_i32, ret, t0, arg1, t0, arg2); | ||
| 319 | +} | ||
| 320 | + | ||
| 321 | +static inline void tcg_gen_rem_i32(int ret, int arg1, int arg2) | ||
| 322 | +{ | ||
| 323 | + int t0; | ||
| 324 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
| 325 | + tcg_gen_sari_i32(t0, arg1, 31); | ||
| 326 | + tcg_gen_op5(INDEX_op_div2_i32, t0, ret, arg1, t0, arg2); | ||
| 327 | +} | ||
| 328 | + | ||
| 329 | +static inline void tcg_gen_divu_i32(int ret, int arg1, int arg2) | ||
| 330 | +{ | ||
| 331 | + int t0; | ||
| 332 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
| 333 | + tcg_gen_movi_i32(t0, 0); | ||
| 334 | + tcg_gen_op5(INDEX_op_divu2_i32, ret, t0, arg1, t0, arg2); | ||
| 335 | +} | ||
| 336 | + | ||
| 337 | +static inline void tcg_gen_remu_i32(int ret, int arg1, int arg2) | ||
| 338 | +{ | ||
| 339 | + int t0; | ||
| 340 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
| 341 | + tcg_gen_movi_i32(t0, 0); | ||
| 342 | + tcg_gen_op5(INDEX_op_divu2_i32, t0, ret, arg1, t0, arg2); | ||
| 343 | +} | ||
| 344 | +#endif | ||
| 345 | + | ||
| 346 | +#if TCG_TARGET_REG_BITS == 32 | ||
| 347 | + | ||
| 348 | +static inline void tcg_gen_mov_i64(int ret, int arg) | ||
| 349 | +{ | ||
| 350 | + tcg_gen_mov_i32(ret, arg); | ||
| 351 | + tcg_gen_mov_i32(ret + 1, arg + 1); | ||
| 352 | +} | ||
| 353 | + | ||
| 354 | +static inline void tcg_gen_movi_i64(int ret, int64_t arg) | ||
| 355 | +{ | ||
| 356 | + tcg_gen_movi_i32(ret, arg); | ||
| 357 | + tcg_gen_movi_i32(ret + 1, arg >> 32); | ||
| 358 | +} | ||
| 359 | + | ||
| 360 | +static inline void tcg_gen_ld8u_i64(int ret, int arg2, tcg_target_long offset) | ||
| 361 | +{ | ||
| 362 | + tcg_gen_ld8u_i32(ret, arg2, offset); | ||
| 363 | + tcg_gen_movi_i32(ret + 1, 0); | ||
| 364 | +} | ||
| 365 | + | ||
| 366 | +static inline void tcg_gen_ld8s_i64(int ret, int arg2, tcg_target_long offset) | ||
| 367 | +{ | ||
| 368 | + tcg_gen_ld8s_i32(ret, arg2, offset); | ||
| 369 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
| 370 | +} | ||
| 371 | + | ||
| 372 | +static inline void tcg_gen_ld16u_i64(int ret, int arg2, tcg_target_long offset) | ||
| 373 | +{ | ||
| 374 | + tcg_gen_ld16u_i32(ret, arg2, offset); | ||
| 375 | + tcg_gen_movi_i32(ret + 1, 0); | ||
| 376 | +} | ||
| 377 | + | ||
| 378 | +static inline void tcg_gen_ld16s_i64(int ret, int arg2, tcg_target_long offset) | ||
| 379 | +{ | ||
| 380 | + tcg_gen_ld16s_i32(ret, arg2, offset); | ||
| 381 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
| 382 | +} | ||
| 383 | + | ||
| 384 | +static inline void tcg_gen_ld32u_i64(int ret, int arg2, tcg_target_long offset) | ||
| 385 | +{ | ||
| 386 | + tcg_gen_ld_i32(ret, arg2, offset); | ||
| 387 | + tcg_gen_movi_i32(ret + 1, 0); | ||
| 388 | +} | ||
| 389 | + | ||
| 390 | +static inline void tcg_gen_ld32s_i64(int ret, int arg2, tcg_target_long offset) | ||
| 391 | +{ | ||
| 392 | + tcg_gen_ld_i32(ret, arg2, offset); | ||
| 393 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
| 394 | +} | ||
| 395 | + | ||
| 396 | +static inline void tcg_gen_ld_i64(int ret, int arg2, tcg_target_long offset) | ||
| 397 | +{ | ||
| 398 | + /* since arg2 and ret have different types, they cannot be the | ||
| 399 | + same temporary */ | ||
| 400 | +#ifdef TCG_TARGET_WORDS_BIGENDIAN | ||
| 401 | + tcg_gen_ld_i32(ret + 1, arg2, offset); | ||
| 402 | + tcg_gen_ld_i32(ret, arg2, offset + 4); | ||
| 403 | +#else | ||
| 404 | + tcg_gen_ld_i32(ret, arg2, offset); | ||
| 405 | + tcg_gen_ld_i32(ret + 1, arg2, offset + 4); | ||
| 406 | +#endif | ||
| 407 | +} | ||
| 408 | + | ||
| 409 | +static inline void tcg_gen_st8_i64(int arg1, int arg2, tcg_target_long offset) | ||
| 410 | +{ | ||
| 411 | + tcg_gen_st8_i32(arg1, arg2, offset); | ||
| 412 | +} | ||
| 413 | + | ||
| 414 | +static inline void tcg_gen_st16_i64(int arg1, int arg2, tcg_target_long offset) | ||
| 415 | +{ | ||
| 416 | + tcg_gen_st16_i32(arg1, arg2, offset); | ||
| 417 | +} | ||
| 418 | + | ||
| 419 | +static inline void tcg_gen_st32_i64(int arg1, int arg2, tcg_target_long offset) | ||
| 420 | +{ | ||
| 421 | + tcg_gen_st_i32(arg1, arg2, offset); | ||
| 422 | +} | ||
| 423 | + | ||
| 424 | +static inline void tcg_gen_st_i64(int arg1, int arg2, tcg_target_long offset) | ||
| 425 | +{ | ||
| 426 | +#ifdef TCG_TARGET_WORDS_BIGENDIAN | ||
| 427 | + tcg_gen_st_i32(arg1 + 1, arg2, offset); | ||
| 428 | + tcg_gen_st_i32(arg1, arg2, offset + 4); | ||
| 429 | +#else | ||
| 430 | + tcg_gen_st_i32(arg1, arg2, offset); | ||
| 431 | + tcg_gen_st_i32(arg1 + 1, arg2, offset + 4); | ||
| 432 | +#endif | ||
| 433 | +} | ||
| 434 | + | ||
| 435 | +static inline void tcg_gen_add_i64(int ret, int arg1, int arg2) | ||
| 436 | +{ | ||
| 437 | + tcg_gen_op6(INDEX_op_add2_i32, ret, ret + 1, | ||
| 438 | + arg1, arg1 + 1, arg2, arg2 + 1); | ||
| 439 | +} | ||
| 440 | + | ||
| 441 | +static inline void tcg_gen_addi_i64(int ret, int arg1, int64_t arg2) | ||
| 442 | +{ | ||
| 443 | + tcg_gen_add_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 444 | +} | ||
| 445 | + | ||
| 446 | +static inline void tcg_gen_sub_i64(int ret, int arg1, int arg2) | ||
| 447 | +{ | ||
| 448 | + tcg_gen_op6(INDEX_op_sub2_i32, ret, ret + 1, | ||
| 449 | + arg1, arg1 + 1, arg2, arg2 + 1); | ||
| 450 | +} | ||
| 451 | + | ||
| 452 | +static inline void tcg_gen_subi_i64(int ret, int arg1, int64_t arg2) | ||
| 453 | +{ | ||
| 454 | + tcg_gen_sub_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 455 | +} | ||
| 456 | + | ||
| 457 | +static inline void tcg_gen_and_i64(int ret, int arg1, int arg2) | ||
| 458 | +{ | ||
| 459 | + tcg_gen_and_i32(ret, arg1, arg2); | ||
| 460 | + tcg_gen_and_i32(ret + 1, arg1 + 1, arg2 + 1); | ||
| 461 | +} | ||
| 462 | + | ||
| 463 | +static inline void tcg_gen_andi_i64(int ret, int arg1, int64_t arg2) | ||
| 464 | +{ | ||
| 465 | + tcg_gen_andi_i32(ret, arg1, arg2); | ||
| 466 | + tcg_gen_andi_i32(ret + 1, arg1 + 1, arg2 >> 32); | ||
| 467 | +} | ||
| 468 | + | ||
| 469 | +static inline void tcg_gen_or_i64(int ret, int arg1, int arg2) | ||
| 470 | +{ | ||
| 471 | + tcg_gen_or_i32(ret, arg1, arg2); | ||
| 472 | + tcg_gen_or_i32(ret + 1, arg1 + 1, arg2 + 1); | ||
| 473 | +} | ||
| 474 | + | ||
| 475 | +static inline void tcg_gen_ori_i64(int ret, int arg1, int64_t arg2) | ||
| 476 | +{ | ||
| 477 | + tcg_gen_ori_i32(ret, arg1, arg2); | ||
| 478 | + tcg_gen_ori_i32(ret + 1, arg1 + 1, arg2 >> 32); | ||
| 479 | +} | ||
| 480 | + | ||
| 481 | +static inline void tcg_gen_xor_i64(int ret, int arg1, int arg2) | ||
| 482 | +{ | ||
| 483 | + tcg_gen_xor_i32(ret, arg1, arg2); | ||
| 484 | + tcg_gen_xor_i32(ret + 1, arg1 + 1, arg2 + 1); | ||
| 485 | +} | ||
| 486 | + | ||
| 487 | +static inline void tcg_gen_xori_i64(int ret, int arg1, int64_t arg2) | ||
| 488 | +{ | ||
| 489 | + tcg_gen_xori_i32(ret, arg1, arg2); | ||
| 490 | + tcg_gen_xori_i32(ret + 1, arg1 + 1, arg2 >> 32); | ||
| 491 | +} | ||
| 492 | + | ||
| 493 | +/* XXX: use generic code when basic block handling is OK or CPU | ||
| 494 | + specific code (x86) */ | ||
| 495 | +static inline void tcg_gen_shl_i64(int ret, int arg1, int64_t arg2) | ||
| 496 | +{ | ||
| 497 | + tcg_gen_helper_1_2(tcg_helper_shl_i64, ret, arg1, arg2); | ||
| 498 | +} | ||
| 499 | + | ||
| 500 | +static inline void tcg_gen_shli_i64(int ret, int arg1, int64_t arg2) | ||
| 501 | +{ | ||
| 502 | + tcg_gen_shifti_i64(ret, arg1, arg2, 0, 0); | ||
| 503 | +} | ||
| 504 | + | ||
| 505 | +static inline void tcg_gen_shr_i64(int ret, int arg1, int64_t arg2) | ||
| 506 | +{ | ||
| 507 | + tcg_gen_helper_1_2(tcg_helper_shr_i64, ret, arg1, arg2); | ||
| 508 | +} | ||
| 509 | + | ||
| 510 | +static inline void tcg_gen_shri_i64(int ret, int arg1, int64_t arg2) | ||
| 511 | +{ | ||
| 512 | + tcg_gen_shifti_i64(ret, arg1, arg2, 1, 0); | ||
| 513 | +} | ||
| 514 | + | ||
| 515 | +static inline void tcg_gen_sar_i64(int ret, int arg1, int64_t arg2) | ||
| 516 | +{ | ||
| 517 | + tcg_gen_helper_1_2(tcg_helper_sar_i64, ret, arg1, arg2); | ||
| 518 | +} | ||
| 519 | + | ||
| 520 | +static inline void tcg_gen_sari_i64(int ret, int arg1, int64_t arg2) | ||
| 521 | +{ | ||
| 522 | + tcg_gen_shifti_i64(ret, arg1, arg2, 1, 1); | ||
| 523 | +} | ||
| 524 | + | ||
| 525 | +static inline void tcg_gen_brcond_i64(int cond, TCGArg arg1, TCGArg arg2, | ||
| 526 | + int label_index) | ||
| 527 | +{ | ||
| 528 | + tcg_gen_op6(INDEX_op_brcond2_i32, | ||
| 529 | + arg1, arg1 + 1, arg2, arg2 + 1, cond, label_index); | ||
| 530 | +} | ||
| 531 | + | ||
| 532 | +static inline void tcg_gen_mul_i64(int ret, int arg1, int arg2) | ||
| 533 | +{ | ||
| 534 | + int t0, t1; | ||
| 535 | + | ||
| 536 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
| 537 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
| 538 | + | ||
| 539 | + tcg_gen_op4(INDEX_op_mulu2_i32, t0, t0 + 1, arg1, arg2); | ||
| 540 | + | ||
| 541 | + tcg_gen_mul_i32(t1, arg1, arg2 + 1); | ||
| 542 | + tcg_gen_add_i32(t0 + 1, t0 + 1, t1); | ||
| 543 | + tcg_gen_mul_i32(t1, arg1 + 1, arg2); | ||
| 544 | + tcg_gen_add_i32(t0 + 1, t0 + 1, t1); | ||
| 545 | + | ||
| 546 | + tcg_gen_mov_i64(ret, t0); | ||
| 547 | +} | ||
| 548 | + | ||
| 549 | +static inline void tcg_gen_div_i64(int ret, int arg1, int64_t arg2) | ||
| 550 | +{ | ||
| 551 | + tcg_gen_helper_1_2(tcg_helper_div_i64, ret, arg1, arg2); | ||
| 552 | +} | ||
| 553 | + | ||
| 554 | +static inline void tcg_gen_rem_i64(int ret, int arg1, int64_t arg2) | ||
| 555 | +{ | ||
| 556 | + tcg_gen_helper_1_2(tcg_helper_rem_i64, ret, arg1, arg2); | ||
| 557 | +} | ||
| 558 | + | ||
| 559 | +static inline void tcg_gen_divu_i64(int ret, int arg1, int64_t arg2) | ||
| 560 | +{ | ||
| 561 | + tcg_gen_helper_1_2(tcg_helper_divu_i64, ret, arg1, arg2); | ||
| 562 | +} | ||
| 563 | + | ||
| 564 | +static inline void tcg_gen_remu_i64(int ret, int arg1, int64_t arg2) | ||
| 565 | +{ | ||
| 566 | + tcg_gen_helper_1_2(tcg_helper_remu_i64, ret, arg1, arg2); | ||
| 567 | +} | ||
| 568 | + | ||
| 569 | +#else | ||
| 570 | + | ||
| 571 | +static inline void tcg_gen_mov_i64(int ret, int arg) | ||
| 572 | +{ | ||
| 573 | + tcg_gen_op2(INDEX_op_mov_i64, ret, arg); | ||
| 574 | +} | ||
| 575 | + | ||
| 576 | +static inline void tcg_gen_movi_i64(int ret, int64_t arg) | ||
| 577 | +{ | ||
| 578 | + tcg_gen_op2(INDEX_op_movi_i64, ret, arg); | ||
| 579 | +} | ||
| 580 | + | ||
| 581 | +static inline void tcg_gen_ld8u_i64(int ret, int arg2, tcg_target_long offset) | ||
| 582 | +{ | ||
| 583 | + tcg_gen_op3(INDEX_op_ld8u_i64, ret, arg2, offset); | ||
| 584 | +} | ||
| 585 | + | ||
| 586 | +static inline void tcg_gen_ld8s_i64(int ret, int arg2, tcg_target_long offset) | ||
| 587 | +{ | ||
| 588 | + tcg_gen_op3(INDEX_op_ld8s_i64, ret, arg2, offset); | ||
| 589 | +} | ||
| 590 | + | ||
| 591 | +static inline void tcg_gen_ld16u_i64(int ret, int arg2, tcg_target_long offset) | ||
| 592 | +{ | ||
| 593 | + tcg_gen_op3(INDEX_op_ld16u_i64, ret, arg2, offset); | ||
| 594 | +} | ||
| 595 | + | ||
| 596 | +static inline void tcg_gen_ld16s_i64(int ret, int arg2, tcg_target_long offset) | ||
| 597 | +{ | ||
| 598 | + tcg_gen_op3(INDEX_op_ld16s_i64, ret, arg2, offset); | ||
| 599 | +} | ||
| 600 | + | ||
| 601 | +static inline void tcg_gen_ld32u_i64(int ret, int arg2, tcg_target_long offset) | ||
| 602 | +{ | ||
| 603 | + tcg_gen_op3(INDEX_op_ld32u_i64, ret, arg2, offset); | ||
| 604 | +} | ||
| 605 | + | ||
| 606 | +static inline void tcg_gen_ld32s_i64(int ret, int arg2, tcg_target_long offset) | ||
| 607 | +{ | ||
| 608 | + tcg_gen_op3(INDEX_op_ld32s_i64, ret, arg2, offset); | ||
| 609 | +} | ||
| 610 | + | ||
| 611 | +static inline void tcg_gen_ld_i64(int ret, int arg2, tcg_target_long offset) | ||
| 612 | +{ | ||
| 613 | + tcg_gen_op3(INDEX_op_ld_i64, ret, arg2, offset); | ||
| 614 | +} | ||
| 615 | + | ||
| 616 | +static inline void tcg_gen_st8_i64(int arg1, int arg2, tcg_target_long offset) | ||
| 617 | +{ | ||
| 618 | + tcg_gen_op3(INDEX_op_st8_i64, arg1, arg2, offset); | ||
| 619 | +} | ||
| 620 | + | ||
| 621 | +static inline void tcg_gen_st16_i64(int arg1, int arg2, tcg_target_long offset) | ||
| 622 | +{ | ||
| 623 | + tcg_gen_op3(INDEX_op_st16_i64, arg1, arg2, offset); | ||
| 624 | +} | ||
| 625 | + | ||
| 626 | +static inline void tcg_gen_st32_i64(int arg1, int arg2, tcg_target_long offset) | ||
| 627 | +{ | ||
| 628 | + tcg_gen_op3(INDEX_op_st32_i64, arg1, arg2, offset); | ||
| 629 | +} | ||
| 630 | + | ||
| 631 | +static inline void tcg_gen_st_i64(int arg1, int arg2, tcg_target_long offset) | ||
| 632 | +{ | ||
| 633 | + tcg_gen_op3(INDEX_op_st_i64, arg1, arg2, offset); | ||
| 634 | +} | ||
| 635 | + | ||
| 636 | +static inline void tcg_gen_add_i64(int ret, int arg1, int arg2) | ||
| 637 | +{ | ||
| 638 | + tcg_gen_op3(INDEX_op_add_i64, ret, arg1, arg2); | ||
| 639 | +} | ||
| 640 | + | ||
| 641 | +static inline void tcg_gen_addi_i64(int ret, int arg1, int64_t arg2) | ||
| 642 | +{ | ||
| 643 | + tcg_gen_add_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 644 | +} | ||
| 645 | + | ||
| 646 | +static inline void tcg_gen_sub_i64(int ret, int arg1, int arg2) | ||
| 647 | +{ | ||
| 648 | + tcg_gen_op3(INDEX_op_sub_i64, ret, arg1, arg2); | ||
| 649 | +} | ||
| 650 | + | ||
| 651 | +static inline void tcg_gen_subi_i64(int ret, int arg1, int64_t arg2) | ||
| 652 | +{ | ||
| 653 | + tcg_gen_sub_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 654 | +} | ||
| 655 | + | ||
| 656 | +static inline void tcg_gen_and_i64(int ret, int arg1, int arg2) | ||
| 657 | +{ | ||
| 658 | + tcg_gen_op3(INDEX_op_and_i64, ret, arg1, arg2); | ||
| 659 | +} | ||
| 660 | + | ||
| 661 | +static inline void tcg_gen_andi_i64(int ret, int arg1, int64_t arg2) | ||
| 662 | +{ | ||
| 663 | + tcg_gen_and_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 664 | +} | ||
| 665 | + | ||
| 666 | +static inline void tcg_gen_or_i64(int ret, int arg1, int arg2) | ||
| 667 | +{ | ||
| 668 | + tcg_gen_op3(INDEX_op_or_i64, ret, arg1, arg2); | ||
| 669 | +} | ||
| 670 | + | ||
| 671 | +static inline void tcg_gen_ori_i64(int ret, int arg1, int64_t arg2) | ||
| 672 | +{ | ||
| 673 | + tcg_gen_or_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 674 | +} | ||
| 675 | + | ||
| 676 | +static inline void tcg_gen_xor_i64(int ret, int arg1, int arg2) | ||
| 677 | +{ | ||
| 678 | + tcg_gen_op3(INDEX_op_xor_i64, ret, arg1, arg2); | ||
| 679 | +} | ||
| 680 | + | ||
| 681 | +static inline void tcg_gen_xori_i64(int ret, int arg1, int64_t arg2) | ||
| 682 | +{ | ||
| 683 | + tcg_gen_xor_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 684 | +} | ||
| 685 | + | ||
| 686 | +static inline void tcg_gen_shl_i64(int ret, int arg1, int arg2) | ||
| 687 | +{ | ||
| 688 | + tcg_gen_op3(INDEX_op_shl_i64, ret, arg1, arg2); | ||
| 689 | +} | ||
| 690 | + | ||
| 691 | +static inline void tcg_gen_shli_i64(int ret, int arg1, int64_t arg2) | ||
| 692 | +{ | ||
| 693 | + tcg_gen_shl_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 694 | +} | ||
| 695 | + | ||
| 696 | +static inline void tcg_gen_shr_i64(int ret, int arg1, int arg2) | ||
| 697 | +{ | ||
| 698 | + tcg_gen_op3(INDEX_op_shr_i64, ret, arg1, arg2); | ||
| 699 | +} | ||
| 700 | + | ||
| 701 | +static inline void tcg_gen_shri_i64(int ret, int arg1, int64_t arg2) | ||
| 702 | +{ | ||
| 703 | + tcg_gen_shr_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 704 | +} | ||
| 705 | + | ||
| 706 | +static inline void tcg_gen_sar_i64(int ret, int arg1, int arg2) | ||
| 707 | +{ | ||
| 708 | + tcg_gen_op3(INDEX_op_sar_i64, ret, arg1, arg2); | ||
| 709 | +} | ||
| 710 | + | ||
| 711 | +static inline void tcg_gen_sari_i64(int ret, int arg1, int64_t arg2) | ||
| 712 | +{ | ||
| 713 | + tcg_gen_sar_i64(ret, arg1, tcg_const_i64(arg2)); | ||
| 714 | +} | ||
| 715 | + | ||
| 716 | +static inline void tcg_gen_brcond_i64(int cond, TCGArg arg1, TCGArg arg2, | ||
| 717 | + int label_index) | ||
| 718 | +{ | ||
| 719 | + tcg_gen_op4(INDEX_op_brcond_i64, arg1, arg2, cond, label_index); | ||
| 720 | +} | ||
| 721 | + | ||
| 722 | +static inline void tcg_gen_mul_i64(int ret, int arg1, int arg2) | ||
| 723 | +{ | ||
| 724 | + tcg_gen_op3(INDEX_op_mul_i64, ret, arg1, arg2); | ||
| 725 | +} | ||
| 726 | + | ||
| 727 | +#ifdef TCG_TARGET_HAS_div_i64 | ||
| 728 | +static inline void tcg_gen_div_i64(int ret, int arg1, int arg2) | ||
| 729 | +{ | ||
| 730 | + tcg_gen_op3(INDEX_op_div_i64, ret, arg1, arg2); | ||
| 731 | +} | ||
| 732 | + | ||
| 733 | +static inline void tcg_gen_rem_i64(int ret, int arg1, int arg2) | ||
| 734 | +{ | ||
| 735 | + tcg_gen_op3(INDEX_op_rem_i64, ret, arg1, arg2); | ||
| 736 | +} | ||
| 737 | + | ||
| 738 | +static inline void tcg_gen_divu_i64(int ret, int arg1, int arg2) | ||
| 739 | +{ | ||
| 740 | + tcg_gen_op3(INDEX_op_divu_i64, ret, arg1, arg2); | ||
| 741 | +} | ||
| 742 | + | ||
| 743 | +static inline void tcg_gen_remu_i64(int ret, int arg1, int arg2) | ||
| 744 | +{ | ||
| 745 | + tcg_gen_op3(INDEX_op_remu_i64, ret, arg1, arg2); | ||
| 746 | +} | ||
| 747 | +#else | ||
| 748 | +static inline void tcg_gen_div_i64(int ret, int arg1, int arg2) | ||
| 749 | +{ | ||
| 750 | + int t0; | ||
| 751 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
| 752 | + tcg_gen_sari_i64(t0, arg1, 63); | ||
| 753 | + tcg_gen_op5(INDEX_op_div2_i64, ret, t0, arg1, t0, arg2); | ||
| 754 | +} | ||
| 755 | + | ||
| 756 | +static inline void tcg_gen_rem_i64(int ret, int arg1, int arg2) | ||
| 757 | +{ | ||
| 758 | + int t0; | ||
| 759 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
| 760 | + tcg_gen_sari_i64(t0, arg1, 63); | ||
| 761 | + tcg_gen_op5(INDEX_op_div2_i64, t0, ret, arg1, t0, arg2); | ||
| 762 | +} | ||
| 763 | + | ||
| 764 | +static inline void tcg_gen_divu_i64(int ret, int arg1, int arg2) | ||
| 765 | +{ | ||
| 766 | + int t0; | ||
| 767 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
| 768 | + tcg_gen_movi_i64(t0, 0); | ||
| 769 | + tcg_gen_op5(INDEX_op_divu2_i64, ret, t0, arg1, t0, arg2); | ||
| 770 | +} | ||
| 771 | + | ||
| 772 | +static inline void tcg_gen_remu_i64(int ret, int arg1, int arg2) | ||
| 773 | +{ | ||
| 774 | + int t0; | ||
| 775 | + t0 = tcg_temp_new(TCG_TYPE_I64); | ||
| 776 | + tcg_gen_movi_i64(t0, 0); | ||
| 777 | + tcg_gen_op5(INDEX_op_divu2_i64, t0, ret, arg1, t0, arg2); | ||
| 778 | +} | ||
| 779 | +#endif | ||
| 780 | + | ||
| 781 | +#endif | ||
| 782 | + | ||
| 783 | +/***************************************/ | ||
| 784 | +/* optional operations */ | ||
| 785 | + | ||
| 786 | +static inline void tcg_gen_ext8s_i32(int ret, int arg) | ||
| 787 | +{ | ||
| 788 | +#ifdef TCG_TARGET_HAS_ext8s_i32 | ||
| 789 | + tcg_gen_op2(INDEX_op_ext8s_i32, ret, arg); | ||
| 790 | +#else | ||
| 791 | + tcg_gen_shli_i32(ret, arg, 24); | ||
| 792 | + tcg_gen_sari_i32(ret, arg, 24); | ||
| 793 | +#endif | ||
| 794 | +} | ||
| 795 | + | ||
| 796 | +static inline void tcg_gen_ext16s_i32(int ret, int arg) | ||
| 797 | +{ | ||
| 798 | +#ifdef TCG_TARGET_HAS_ext16s_i32 | ||
| 799 | + tcg_gen_op2(INDEX_op_ext16s_i32, ret, arg); | ||
| 800 | +#else | ||
| 801 | + tcg_gen_shli_i32(ret, arg, 16); | ||
| 802 | + tcg_gen_sari_i32(ret, arg, 16); | ||
| 803 | +#endif | ||
| 804 | +} | ||
| 805 | + | ||
| 806 | +/* Note: we assume the two high bytes are set to zero */ | ||
| 807 | +static inline void tcg_gen_bswap16_i32(TCGArg ret, TCGArg arg) | ||
| 808 | +{ | ||
| 809 | +#ifdef TCG_TARGET_HAS_bswap16_i32 | ||
| 810 | + tcg_gen_op2(INDEX_op_bswap16_i32, ret, arg); | ||
| 811 | +#else | ||
| 812 | + TCGArg t0, t1; | ||
| 813 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
| 814 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
| 815 | + | ||
| 816 | + tcg_gen_shri_i32(t0, arg, 8); | ||
| 817 | + tcg_gen_andi_i32(t1, arg, 0x000000ff); | ||
| 818 | + tcg_gen_shli_i32(t1, t1, 8); | ||
| 819 | + tcg_gen_or_i32(ret, t0, t1); | ||
| 820 | +#endif | ||
| 821 | +} | ||
| 822 | + | ||
| 823 | +static inline void tcg_gen_bswap_i32(TCGArg ret, TCGArg arg) | ||
| 824 | +{ | ||
| 825 | +#ifdef TCG_TARGET_HAS_bswap_i32 | ||
| 826 | + tcg_gen_op2(INDEX_op_bswap_i32, ret, arg); | ||
| 827 | +#else | ||
| 828 | + TCGArg t0, t1; | ||
| 829 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
| 830 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
| 831 | + | ||
| 832 | + tcg_gen_shli_i32(t0, arg, 24); | ||
| 833 | + | ||
| 834 | + tcg_gen_andi_i32(t1, arg, 0x0000ff00); | ||
| 835 | + tcg_gen_shli_i32(t1, t1, 8); | ||
| 836 | + tcg_gen_or_i32(t0, t0, t1); | ||
| 837 | + | ||
| 838 | + tcg_gen_shri_i32(t1, arg, 8); | ||
| 839 | + tcg_gen_andi_i32(t1, t1, 0x0000ff00); | ||
| 840 | + tcg_gen_or_i32(t0, t0, t1); | ||
| 841 | + | ||
| 842 | + tcg_gen_shri_i32(t1, arg, 24); | ||
| 843 | + tcg_gen_or_i32(ret, t0, t1); | ||
| 844 | +#endif | ||
| 845 | +} | ||
| 846 | + | ||
| 847 | +#if TCG_TARGET_REG_BITS == 32 | ||
| 848 | +static inline void tcg_gen_ext8s_i64(int ret, int arg) | ||
| 849 | +{ | ||
| 850 | + tcg_gen_ext8s_i32(ret, arg); | ||
| 851 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
| 852 | +} | ||
| 853 | + | ||
| 854 | +static inline void tcg_gen_ext16s_i64(int ret, int arg) | ||
| 855 | +{ | ||
| 856 | + tcg_gen_ext16s_i32(ret, arg); | ||
| 857 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
| 858 | +} | ||
| 859 | + | ||
| 860 | +static inline void tcg_gen_ext32s_i64(int ret, int arg) | ||
| 861 | +{ | ||
| 862 | + tcg_gen_mov_i32(ret, arg); | ||
| 863 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
| 864 | +} | ||
| 865 | + | ||
| 866 | +static inline void tcg_gen_trunc_i64_i32(int ret, int arg) | ||
| 867 | +{ | ||
| 868 | + tcg_gen_mov_i32(ret, arg); | ||
| 869 | +} | ||
| 870 | + | ||
| 871 | +static inline void tcg_gen_extu_i32_i64(int ret, int arg) | ||
| 872 | +{ | ||
| 873 | + tcg_gen_mov_i32(ret, arg); | ||
| 874 | + tcg_gen_movi_i32(ret + 1, 0); | ||
| 875 | +} | ||
| 876 | + | ||
| 877 | +static inline void tcg_gen_ext_i32_i64(int ret, int arg) | ||
| 878 | +{ | ||
| 879 | + tcg_gen_mov_i32(ret, arg); | ||
| 880 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
| 881 | +} | ||
| 882 | + | ||
| 883 | +static inline void tcg_gen_bswap_i64(int ret, int arg) | ||
| 884 | +{ | ||
| 885 | + int t0, t1; | ||
| 886 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
| 887 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
| 888 | + | ||
| 889 | + tcg_gen_bswap_i32(t0, arg); | ||
| 890 | + tcg_gen_bswap_i32(t1, arg + 1); | ||
| 891 | + tcg_gen_mov_i32(ret, t1); | ||
| 892 | + tcg_gen_mov_i32(ret + 1, t0); | ||
| 893 | +} | ||
| 894 | +#else | ||
| 895 | + | ||
| 896 | +static inline void tcg_gen_ext8s_i64(int ret, int arg) | ||
| 897 | +{ | ||
| 898 | +#ifdef TCG_TARGET_HAS_ext8s_i64 | ||
| 899 | + tcg_gen_op2(INDEX_op_ext8s_i64, ret, arg); | ||
| 900 | +#else | ||
| 901 | + tcg_gen_shli_i64(ret, arg, 56); | ||
| 902 | + tcg_gen_sari_i64(ret, arg, 56); | ||
| 903 | +#endif | ||
| 904 | +} | ||
| 905 | + | ||
| 906 | +static inline void tcg_gen_ext16s_i64(int ret, int arg) | ||
| 907 | +{ | ||
| 908 | +#ifdef TCG_TARGET_HAS_ext16s_i64 | ||
| 909 | + tcg_gen_op2(INDEX_op_ext16s_i64, ret, arg); | ||
| 910 | +#else | ||
| 911 | + tcg_gen_shli_i64(ret, arg, 48); | ||
| 912 | + tcg_gen_sari_i64(ret, arg, 48); | ||
| 913 | +#endif | ||
| 914 | +} | ||
| 915 | + | ||
| 916 | +static inline void tcg_gen_ext32s_i64(int ret, int arg) | ||
| 917 | +{ | ||
| 918 | +#ifdef TCG_TARGET_HAS_ext32s_i64 | ||
| 919 | + tcg_gen_op2(INDEX_op_ext32s_i64, ret, arg); | ||
| 920 | +#else | ||
| 921 | + tcg_gen_shli_i64(ret, arg, 32); | ||
| 922 | + tcg_gen_sari_i64(ret, arg, 32); | ||
| 923 | +#endif | ||
| 924 | +} | ||
| 925 | + | ||
| 926 | +/* Note: we assume the target supports move between 32 and 64 bit | ||
| 927 | + registers */ | ||
| 928 | +static inline void tcg_gen_trunc_i64_i32(int ret, int arg) | ||
| 929 | +{ | ||
| 930 | + tcg_gen_mov_i32(ret, arg); | ||
| 931 | +} | ||
| 932 | + | ||
| 933 | +/* Note: we assume the target supports move between 32 and 64 bit | ||
| 934 | + registers */ | ||
| 935 | +static inline void tcg_gen_extu_i32_i64(int ret, int arg) | ||
| 936 | +{ | ||
| 937 | + tcg_gen_andi_i64(ret, arg, 0xffffffff); | ||
| 938 | +} | ||
| 939 | + | ||
| 940 | +/* Note: we assume the target supports move between 32 and 64 bit | ||
| 941 | + registers */ | ||
| 942 | +static inline void tcg_gen_ext_i32_i64(int ret, int arg) | ||
| 943 | +{ | ||
| 944 | + tcg_gen_ext32s_i64(ret, arg); | ||
| 945 | +} | ||
| 946 | + | ||
| 947 | +static inline void tcg_gen_bswap_i64(TCGArg ret, TCGArg arg) | ||
| 948 | +{ | ||
| 949 | +#ifdef TCG_TARGET_HAS_bswap_i64 | ||
| 950 | + tcg_gen_op2(INDEX_op_bswap_i64, ret, arg); | ||
| 951 | +#else | ||
| 952 | + TCGArg t0, t1; | ||
| 953 | + t0 = tcg_temp_new(TCG_TYPE_I32); | ||
| 954 | + t1 = tcg_temp_new(TCG_TYPE_I32); | ||
| 955 | + | ||
| 956 | + tcg_gen_shli_i64(t0, arg, 56); | ||
| 957 | + | ||
| 958 | + tcg_gen_andi_i64(t1, arg, 0x0000ff00); | ||
| 959 | + tcg_gen_shli_i64(t1, t1, 40); | ||
| 960 | + tcg_gen_or_i64(t0, t0, t1); | ||
| 961 | + | ||
| 962 | + tcg_gen_andi_i64(t1, arg, 0x00ff0000); | ||
| 963 | + tcg_gen_shli_i64(t1, t1, 24); | ||
| 964 | + tcg_gen_or_i64(t0, t0, t1); | ||
| 965 | + | ||
| 966 | + tcg_gen_andi_i64(t1, arg, 0xff000000); | ||
| 967 | + tcg_gen_shli_i64(t1, t1, 8); | ||
| 968 | + tcg_gen_or_i64(t0, t0, t1); | ||
| 969 | + | ||
| 970 | + tcg_gen_shri_i64(t1, arg, 8); | ||
| 971 | + tcg_gen_andi_i64(t1, t1, 0xff000000); | ||
| 972 | + tcg_gen_or_i64(t0, t0, t1); | ||
| 973 | + | ||
| 974 | + tcg_gen_shri_i64(t1, arg, 24); | ||
| 975 | + tcg_gen_andi_i64(t1, t1, 0x00ff0000); | ||
| 976 | + tcg_gen_or_i64(t0, t0, t1); | ||
| 977 | + | ||
| 978 | + tcg_gen_shri_i64(t1, arg, 40); | ||
| 979 | + tcg_gen_andi_i64(t1, t1, 0x0000ff00); | ||
| 980 | + tcg_gen_or_i64(t0, t0, t1); | ||
| 981 | + | ||
| 982 | + tcg_gen_shri_i64(t1, arg, 56); | ||
| 983 | + tcg_gen_or_i64(ret, t0, t1); | ||
| 984 | +#endif | ||
| 985 | +} | ||
| 986 | + | ||
| 987 | +#endif | ||
| 988 | + | ||
| 989 | +/***************************************/ | ||
| 990 | +static inline void tcg_gen_macro_2(int ret0, int ret1, int macro_id) | ||
| 991 | +{ | ||
| 992 | + tcg_gen_op3(INDEX_op_macro_2, ret0, ret1, macro_id); | ||
| 993 | +} | ||
| 994 | + | ||
| 995 | +/***************************************/ | ||
| 996 | +/* QEMU specific operations. Their type depend on the QEMU CPU | ||
| 997 | + type. */ | ||
| 998 | +#ifndef TARGET_LONG_BITS | ||
| 999 | +#error must include QEMU headers | ||
| 1000 | +#endif | ||
| 1001 | + | ||
| 1002 | +static inline void tcg_gen_exit_tb(tcg_target_long val) | ||
| 1003 | +{ | ||
| 1004 | + tcg_gen_op1(INDEX_op_exit_tb, val); | ||
| 1005 | +} | ||
| 1006 | + | ||
| 1007 | +static inline void tcg_gen_goto_tb(int idx) | ||
| 1008 | +{ | ||
| 1009 | + tcg_gen_op1(INDEX_op_goto_tb, idx); | ||
| 1010 | +} | ||
| 1011 | + | ||
| 1012 | +#if TCG_TARGET_REG_BITS == 32 | ||
| 1013 | +static inline void tcg_gen_qemu_ld8u(int ret, int addr, int mem_index) | ||
| 1014 | +{ | ||
| 1015 | +#if TARGET_LONG_BITS == 32 | ||
| 1016 | + tcg_gen_op3(INDEX_op_qemu_ld8u, ret, addr, mem_index); | ||
| 1017 | +#else | ||
| 1018 | + tcg_gen_op4(INDEX_op_qemu_ld8u, ret, addr, addr + 1, mem_index); | ||
| 1019 | + tcg_gen_movi_i32(ret + 1, 0); | ||
| 1020 | +#endif | ||
| 1021 | +} | ||
| 1022 | + | ||
| 1023 | +static inline void tcg_gen_qemu_ld8s(int ret, int addr, int mem_index) | ||
| 1024 | +{ | ||
| 1025 | +#if TARGET_LONG_BITS == 32 | ||
| 1026 | + tcg_gen_op3(INDEX_op_qemu_ld8s, ret, addr, mem_index); | ||
| 1027 | +#else | ||
| 1028 | + tcg_gen_op4(INDEX_op_qemu_ld8s, ret, addr, addr + 1, mem_index); | ||
| 1029 | + tcg_gen_ext8s_i32(ret + 1, ret); | ||
| 1030 | +#endif | ||
| 1031 | +} | ||
| 1032 | + | ||
| 1033 | +static inline void tcg_gen_qemu_ld16u(int ret, int addr, int mem_index) | ||
| 1034 | +{ | ||
| 1035 | +#if TARGET_LONG_BITS == 32 | ||
| 1036 | + tcg_gen_op3(INDEX_op_qemu_ld16u, ret, addr, mem_index); | ||
| 1037 | +#else | ||
| 1038 | + tcg_gen_op4(INDEX_op_qemu_ld16u, ret, addr, addr + 1, mem_index); | ||
| 1039 | + tcg_gen_movi_i32(ret + 1, 0); | ||
| 1040 | +#endif | ||
| 1041 | +} | ||
| 1042 | + | ||
| 1043 | +static inline void tcg_gen_qemu_ld16s(int ret, int addr, int mem_index) | ||
| 1044 | +{ | ||
| 1045 | +#if TARGET_LONG_BITS == 32 | ||
| 1046 | + tcg_gen_op3(INDEX_op_qemu_ld16s, ret, addr, mem_index); | ||
| 1047 | +#else | ||
| 1048 | + tcg_gen_op4(INDEX_op_qemu_ld16s, ret, addr, addr + 1, mem_index); | ||
| 1049 | + tcg_gen_ext16s_i32(ret + 1, ret); | ||
| 1050 | +#endif | ||
| 1051 | +} | ||
| 1052 | + | ||
| 1053 | +static inline void tcg_gen_qemu_ld32u(int ret, int addr, int mem_index) | ||
| 1054 | +{ | ||
| 1055 | +#if TARGET_LONG_BITS == 32 | ||
| 1056 | + tcg_gen_op3(INDEX_op_qemu_ld32u, ret, addr, mem_index); | ||
| 1057 | +#else | ||
| 1058 | + tcg_gen_op4(INDEX_op_qemu_ld32u, ret, addr, addr + 1, mem_index); | ||
| 1059 | + tcg_gen_movi_i32(ret + 1, 0); | ||
| 1060 | +#endif | ||
| 1061 | +} | ||
| 1062 | + | ||
| 1063 | +static inline void tcg_gen_qemu_ld32s(int ret, int addr, int mem_index) | ||
| 1064 | +{ | ||
| 1065 | +#if TARGET_LONG_BITS == 32 | ||
| 1066 | + tcg_gen_op3(INDEX_op_qemu_ld32u, ret, addr, mem_index); | ||
| 1067 | +#else | ||
| 1068 | + tcg_gen_op4(INDEX_op_qemu_ld32u, ret, addr, addr + 1, mem_index); | ||
| 1069 | + tcg_gen_sari_i32(ret + 1, ret, 31); | ||
| 1070 | +#endif | ||
| 1071 | +} | ||
| 1072 | + | ||
| 1073 | +static inline void tcg_gen_qemu_ld64(int ret, int addr, int mem_index) | ||
| 1074 | +{ | ||
| 1075 | +#if TARGET_LONG_BITS == 32 | ||
| 1076 | + tcg_gen_op4(INDEX_op_qemu_ld64, ret, ret + 1, addr, mem_index); | ||
| 1077 | +#else | ||
| 1078 | + tcg_gen_op5(INDEX_op_qemu_ld64, ret, ret + 1, addr, addr + 1, mem_index); | ||
| 1079 | +#endif | ||
| 1080 | +} | ||
| 1081 | + | ||
| 1082 | +static inline void tcg_gen_qemu_st8(int arg, int addr, int mem_index) | ||
| 1083 | +{ | ||
| 1084 | +#if TARGET_LONG_BITS == 32 | ||
| 1085 | + tcg_gen_op3(INDEX_op_qemu_st8, arg, addr, mem_index); | ||
| 1086 | +#else | ||
| 1087 | + tcg_gen_op4(INDEX_op_qemu_st8, arg, addr, addr + 1, mem_index); | ||
| 1088 | +#endif | ||
| 1089 | +} | ||
| 1090 | + | ||
| 1091 | +static inline void tcg_gen_qemu_st16(int arg, int addr, int mem_index) | ||
| 1092 | +{ | ||
| 1093 | +#if TARGET_LONG_BITS == 32 | ||
| 1094 | + tcg_gen_op3(INDEX_op_qemu_st16, arg, addr, mem_index); | ||
| 1095 | +#else | ||
| 1096 | + tcg_gen_op4(INDEX_op_qemu_st16, arg, addr, addr + 1, mem_index); | ||
| 1097 | +#endif | ||
| 1098 | +} | ||
| 1099 | + | ||
| 1100 | +static inline void tcg_gen_qemu_st32(int arg, int addr, int mem_index) | ||
| 1101 | +{ | ||
| 1102 | +#if TARGET_LONG_BITS == 32 | ||
| 1103 | + tcg_gen_op3(INDEX_op_qemu_st32, arg, addr, mem_index); | ||
| 1104 | +#else | ||
| 1105 | + tcg_gen_op4(INDEX_op_qemu_st32, arg, addr, addr + 1, mem_index); | ||
| 1106 | +#endif | ||
| 1107 | +} | ||
| 1108 | + | ||
| 1109 | +static inline void tcg_gen_qemu_st64(int arg, int addr, int mem_index) | ||
| 1110 | +{ | ||
| 1111 | +#if TARGET_LONG_BITS == 32 | ||
| 1112 | + tcg_gen_op4(INDEX_op_qemu_st64, arg, arg + 1, addr, mem_index); | ||
| 1113 | +#else | ||
| 1114 | + tcg_gen_op5(INDEX_op_qemu_st64, arg, arg + 1, addr, addr + 1, mem_index); | ||
| 1115 | +#endif | ||
| 1116 | +} | ||
| 1117 | + | ||
| 1118 | +#else /* TCG_TARGET_REG_BITS == 32 */ | ||
| 1119 | + | ||
| 1120 | +static inline void tcg_gen_qemu_ld8u(int ret, int addr, int mem_index) | ||
| 1121 | +{ | ||
| 1122 | + tcg_gen_op3(INDEX_op_qemu_ld8u, ret, addr, mem_index); | ||
| 1123 | +} | ||
| 1124 | + | ||
| 1125 | +static inline void tcg_gen_qemu_ld8s(int ret, int addr, int mem_index) | ||
| 1126 | +{ | ||
| 1127 | + tcg_gen_op3(INDEX_op_qemu_ld8s, ret, addr, mem_index); | ||
| 1128 | +} | ||
| 1129 | + | ||
| 1130 | +static inline void tcg_gen_qemu_ld16u(int ret, int addr, int mem_index) | ||
| 1131 | +{ | ||
| 1132 | + tcg_gen_op3(INDEX_op_qemu_ld16u, ret, addr, mem_index); | ||
| 1133 | +} | ||
| 1134 | + | ||
| 1135 | +static inline void tcg_gen_qemu_ld16s(int ret, int addr, int mem_index) | ||
| 1136 | +{ | ||
| 1137 | + tcg_gen_op3(INDEX_op_qemu_ld16s, ret, addr, mem_index); | ||
| 1138 | +} | ||
| 1139 | + | ||
| 1140 | +static inline void tcg_gen_qemu_ld32u(int ret, int addr, int mem_index) | ||
| 1141 | +{ | ||
| 1142 | + tcg_gen_op3(INDEX_op_qemu_ld32u, ret, addr, mem_index); | ||
| 1143 | +} | ||
| 1144 | + | ||
| 1145 | +static inline void tcg_gen_qemu_ld32s(int ret, int addr, int mem_index) | ||
| 1146 | +{ | ||
| 1147 | + tcg_gen_op3(INDEX_op_qemu_ld32s, ret, addr, mem_index); | ||
| 1148 | +} | ||
| 1149 | + | ||
| 1150 | +static inline void tcg_gen_qemu_ld64(int ret, int addr, int mem_index) | ||
| 1151 | +{ | ||
| 1152 | + tcg_gen_op3(INDEX_op_qemu_ld64, ret, addr, mem_index); | ||
| 1153 | +} | ||
| 1154 | + | ||
| 1155 | +static inline void tcg_gen_qemu_st8(int arg, int addr, int mem_index) | ||
| 1156 | +{ | ||
| 1157 | + tcg_gen_op3(INDEX_op_qemu_st8, arg, addr, mem_index); | ||
| 1158 | +} | ||
| 1159 | + | ||
| 1160 | +static inline void tcg_gen_qemu_st16(int arg, int addr, int mem_index) | ||
| 1161 | +{ | ||
| 1162 | + tcg_gen_op3(INDEX_op_qemu_st16, arg, addr, mem_index); | ||
| 1163 | +} | ||
| 1164 | + | ||
| 1165 | +static inline void tcg_gen_qemu_st32(int arg, int addr, int mem_index) | ||
| 1166 | +{ | ||
| 1167 | + tcg_gen_op3(INDEX_op_qemu_st32, arg, addr, mem_index); | ||
| 1168 | +} | ||
| 1169 | + | ||
| 1170 | +static inline void tcg_gen_qemu_st64(int arg, int addr, int mem_index) | ||
| 1171 | +{ | ||
| 1172 | + tcg_gen_op3(INDEX_op_qemu_st64, arg, addr, mem_index); | ||
| 1173 | +} | ||
| 1174 | + | ||
| 1175 | +#endif /* TCG_TARGET_REG_BITS != 32 */ |
tcg/tcg-opc.h
0 โ 100644
| 1 | +/* | ||
| 2 | + * Tiny Code Generator for QEMU | ||
| 3 | + * | ||
| 4 | + * Copyright (c) 2008 Fabrice Bellard | ||
| 5 | + * | ||
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
| 7 | + * of this software and associated documentation files (the "Software"), to deal | ||
| 8 | + * in the Software without restriction, including without limitation the rights | ||
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
| 10 | + * copies of the Software, and to permit persons to whom the Software is | ||
| 11 | + * furnished to do so, subject to the following conditions: | ||
| 12 | + * | ||
| 13 | + * The above copyright notice and this permission notice shall be included in | ||
| 14 | + * all copies or substantial portions of the Software. | ||
| 15 | + * | ||
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
| 22 | + * THE SOFTWARE. | ||
| 23 | + */ | ||
| 24 | +#include "dyngen-opc.h" | ||
| 25 | + | ||
| 26 | +#ifndef DEF2 | ||
| 27 | +#define DEF2(name, oargs, iargs, cargs, flags) DEF(name, oargs + iargs + cargs, 0) | ||
| 28 | +#endif | ||
| 29 | + | ||
| 30 | +/* predefined ops */ | ||
| 31 | +DEF2(end, 0, 0, 0, 0) /* must be kept first */ | ||
| 32 | +DEF2(nop, 0, 0, 0, 0) | ||
| 33 | +DEF2(nop1, 0, 0, 1, 0) | ||
| 34 | +DEF2(nop2, 0, 0, 2, 0) | ||
| 35 | +DEF2(nop3, 0, 0, 3, 0) | ||
| 36 | +DEF2(nopn, 0, 0, 1, 0) /* variable number of parameters */ | ||
| 37 | +/* macro handling */ | ||
| 38 | +DEF2(macro_2, 2, 0, 1, 0) | ||
| 39 | +DEF2(macro_start, 0, 0, 2, 0) | ||
| 40 | +DEF2(macro_end, 0, 0, 2, 0) | ||
| 41 | +DEF2(macro_goto, 0, 0, 3, 0) | ||
| 42 | + | ||
| 43 | +DEF2(set_label, 0, 0, 1, 0) | ||
| 44 | +DEF2(call, 0, 1, 2, 0) /* variable number of parameters */ | ||
| 45 | +DEF2(jmp, 0, 1, 0, TCG_OPF_BB_END) | ||
| 46 | +DEF2(br, 0, 0, 1, TCG_OPF_BB_END) | ||
| 47 | + | ||
| 48 | +DEF2(mov_i32, 1, 1, 0, 0) | ||
| 49 | +DEF2(movi_i32, 1, 0, 1, 0) | ||
| 50 | +/* load/store */ | ||
| 51 | +DEF2(ld8u_i32, 1, 1, 1, 0) | ||
| 52 | +DEF2(ld8s_i32, 1, 1, 1, 0) | ||
| 53 | +DEF2(ld16u_i32, 1, 1, 1, 0) | ||
| 54 | +DEF2(ld16s_i32, 1, 1, 1, 0) | ||
| 55 | +DEF2(ld_i32, 1, 1, 1, 0) | ||
| 56 | +DEF2(st8_i32, 0, 2, 1, 0) | ||
| 57 | +DEF2(st16_i32, 0, 2, 1, 0) | ||
| 58 | +DEF2(st_i32, 0, 2, 1, 0) | ||
| 59 | +/* arith */ | ||
| 60 | +DEF2(add_i32, 1, 2, 0, 0) | ||
| 61 | +DEF2(sub_i32, 1, 2, 0, 0) | ||
| 62 | +DEF2(mul_i32, 1, 2, 0, 0) | ||
| 63 | +#ifdef TCG_TARGET_HAS_div_i32 | ||
| 64 | +DEF2(div_i32, 1, 2, 0, 0) | ||
| 65 | +DEF2(divu_i32, 1, 2, 0, 0) | ||
| 66 | +DEF2(rem_i32, 1, 2, 0, 0) | ||
| 67 | +DEF2(remu_i32, 1, 2, 0, 0) | ||
| 68 | +#else | ||
| 69 | +DEF2(div2_i32, 2, 3, 0, 0) | ||
| 70 | +DEF2(divu2_i32, 2, 3, 0, 0) | ||
| 71 | +#endif | ||
| 72 | +DEF2(and_i32, 1, 2, 0, 0) | ||
| 73 | +DEF2(or_i32, 1, 2, 0, 0) | ||
| 74 | +DEF2(xor_i32, 1, 2, 0, 0) | ||
| 75 | +/* shifts */ | ||
| 76 | +DEF2(shl_i32, 1, 2, 0, 0) | ||
| 77 | +DEF2(shr_i32, 1, 2, 0, 0) | ||
| 78 | +DEF2(sar_i32, 1, 2, 0, 0) | ||
| 79 | + | ||
| 80 | +DEF2(brcond_i32, 0, 2, 2, TCG_OPF_BB_END) | ||
| 81 | +#if TCG_TARGET_REG_BITS == 32 | ||
| 82 | +DEF2(add2_i32, 2, 4, 0, 0) | ||
| 83 | +DEF2(sub2_i32, 2, 4, 0, 0) | ||
| 84 | +DEF2(brcond2_i32, 0, 4, 2, TCG_OPF_BB_END) | ||
| 85 | +DEF2(mulu2_i32, 2, 2, 0, 0) | ||
| 86 | +#endif | ||
| 87 | +#ifdef TCG_TARGET_HAS_ext8s_i32 | ||
| 88 | +DEF2(ext8s_i32, 1, 1, 0, 0) | ||
| 89 | +#endif | ||
| 90 | +#ifdef TCG_TARGET_HAS_ext16s_i32 | ||
| 91 | +DEF2(ext16s_i32, 1, 1, 0, 0) | ||
| 92 | +#endif | ||
| 93 | +#ifdef TCG_TARGET_HAS_bswap_i32 | ||
| 94 | +DEF2(bswap_i32, 1, 1, 0, 0) | ||
| 95 | +#endif | ||
| 96 | + | ||
| 97 | +#if TCG_TARGET_REG_BITS == 64 | ||
| 98 | +DEF2(mov_i64, 1, 1, 0, 0) | ||
| 99 | +DEF2(movi_i64, 1, 0, 1, 0) | ||
| 100 | +/* load/store */ | ||
| 101 | +DEF2(ld8u_i64, 1, 1, 1, 0) | ||
| 102 | +DEF2(ld8s_i64, 1, 1, 1, 0) | ||
| 103 | +DEF2(ld16u_i64, 1, 1, 1, 0) | ||
| 104 | +DEF2(ld16s_i64, 1, 1, 1, 0) | ||
| 105 | +DEF2(ld32u_i64, 1, 1, 1, 0) | ||
| 106 | +DEF2(ld32s_i64, 1, 1, 1, 0) | ||
| 107 | +DEF2(ld_i64, 1, 1, 1, 0) | ||
| 108 | +DEF2(st8_i64, 0, 2, 1, 0) | ||
| 109 | +DEF2(st16_i64, 0, 2, 1, 0) | ||
| 110 | +DEF2(st32_i64, 0, 2, 1, 0) | ||
| 111 | +DEF2(st_i64, 0, 2, 1, 0) | ||
| 112 | +/* arith */ | ||
| 113 | +DEF2(add_i64, 1, 2, 0, 0) | ||
| 114 | +DEF2(sub_i64, 1, 2, 0, 0) | ||
| 115 | +DEF2(mul_i64, 1, 2, 0, 0) | ||
| 116 | +#ifdef TCG_TARGET_HAS_div_i64 | ||
| 117 | +DEF2(div_i64, 1, 2, 0, 0) | ||
| 118 | +DEF2(divu_i64, 1, 2, 0, 0) | ||
| 119 | +DEF2(rem_i64, 1, 2, 0, 0) | ||
| 120 | +DEF2(remu_i64, 1, 2, 0, 0) | ||
| 121 | +#else | ||
| 122 | +DEF2(div2_i64, 2, 3, 0, 0) | ||
| 123 | +DEF2(divu2_i64, 2, 3, 0, 0) | ||
| 124 | +#endif | ||
| 125 | +DEF2(and_i64, 1, 2, 0, 0) | ||
| 126 | +DEF2(or_i64, 1, 2, 0, 0) | ||
| 127 | +DEF2(xor_i64, 1, 2, 0, 0) | ||
| 128 | +/* shifts */ | ||
| 129 | +DEF2(shl_i64, 1, 2, 0, 0) | ||
| 130 | +DEF2(shr_i64, 1, 2, 0, 0) | ||
| 131 | +DEF2(sar_i64, 1, 2, 0, 0) | ||
| 132 | + | ||
| 133 | +DEF2(brcond_i64, 0, 2, 2, TCG_OPF_BB_END) | ||
| 134 | +#ifdef TCG_TARGET_HAS_ext8s_i64 | ||
| 135 | +DEF2(ext8s_i64, 1, 1, 0, 0) | ||
| 136 | +#endif | ||
| 137 | +#ifdef TCG_TARGET_HAS_ext16s_i64 | ||
| 138 | +DEF2(ext16s_i64, 1, 1, 0, 0) | ||
| 139 | +#endif | ||
| 140 | +#ifdef TCG_TARGET_HAS_ext32s_i64 | ||
| 141 | +DEF2(ext32s_i64, 1, 1, 0, 0) | ||
| 142 | +#endif | ||
| 143 | +#ifdef TCG_TARGET_HAS_bswap_i64 | ||
| 144 | +DEF2(bswap_i64, 1, 1, 0, 0) | ||
| 145 | +#endif | ||
| 146 | +#endif | ||
| 147 | + | ||
| 148 | +/* QEMU specific */ | ||
| 149 | +DEF2(exit_tb, 0, 0, 1, TCG_OPF_BB_END) | ||
| 150 | +DEF2(goto_tb, 0, 0, 1, TCG_OPF_BB_END) | ||
| 151 | +/* Note: even if TARGET_LONG_BITS is not defined, the INDEX_op | ||
| 152 | + constants must be defined */ | ||
| 153 | +#if TCG_TARGET_REG_BITS == 32 | ||
| 154 | +#if TARGET_LONG_BITS == 32 | ||
| 155 | +DEF2(qemu_ld8u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 156 | +#else | ||
| 157 | +DEF2(qemu_ld8u, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 158 | +#endif | ||
| 159 | +#if TARGET_LONG_BITS == 32 | ||
| 160 | +DEF2(qemu_ld8s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 161 | +#else | ||
| 162 | +DEF2(qemu_ld8s, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 163 | +#endif | ||
| 164 | +#if TARGET_LONG_BITS == 32 | ||
| 165 | +DEF2(qemu_ld16u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 166 | +#else | ||
| 167 | +DEF2(qemu_ld16u, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 168 | +#endif | ||
| 169 | +#if TARGET_LONG_BITS == 32 | ||
| 170 | +DEF2(qemu_ld16s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 171 | +#else | ||
| 172 | +DEF2(qemu_ld16s, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 173 | +#endif | ||
| 174 | +#if TARGET_LONG_BITS == 32 | ||
| 175 | +DEF2(qemu_ld32u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 176 | +#else | ||
| 177 | +DEF2(qemu_ld32u, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 178 | +#endif | ||
| 179 | +#if TARGET_LONG_BITS == 32 | ||
| 180 | +DEF2(qemu_ld32s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 181 | +#else | ||
| 182 | +DEF2(qemu_ld32s, 1, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 183 | +#endif | ||
| 184 | +#if TARGET_LONG_BITS == 32 | ||
| 185 | +DEF2(qemu_ld64, 2, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 186 | +#else | ||
| 187 | +DEF2(qemu_ld64, 2, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 188 | +#endif | ||
| 189 | + | ||
| 190 | +#if TARGET_LONG_BITS == 32 | ||
| 191 | +DEF2(qemu_st8, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 192 | +#else | ||
| 193 | +DEF2(qemu_st8, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | ||
| 194 | +#endif | ||
| 195 | +#if TARGET_LONG_BITS == 32 | ||
| 196 | +DEF2(qemu_st16, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 197 | +#else | ||
| 198 | +DEF2(qemu_st16, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | ||
| 199 | +#endif | ||
| 200 | +#if TARGET_LONG_BITS == 32 | ||
| 201 | +DEF2(qemu_st32, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 202 | +#else | ||
| 203 | +DEF2(qemu_st32, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | ||
| 204 | +#endif | ||
| 205 | +#if TARGET_LONG_BITS == 32 | ||
| 206 | +DEF2(qemu_st64, 0, 3, 1, TCG_OPF_CALL_CLOBBER) | ||
| 207 | +#else | ||
| 208 | +DEF2(qemu_st64, 0, 4, 1, TCG_OPF_CALL_CLOBBER) | ||
| 209 | +#endif | ||
| 210 | + | ||
| 211 | +#else /* TCG_TARGET_REG_BITS == 32 */ | ||
| 212 | + | ||
| 213 | +DEF2(qemu_ld8u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 214 | +DEF2(qemu_ld8s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 215 | +DEF2(qemu_ld16u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 216 | +DEF2(qemu_ld16s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 217 | +DEF2(qemu_ld32u, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 218 | +DEF2(qemu_ld32s, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 219 | +DEF2(qemu_ld64, 1, 1, 1, TCG_OPF_CALL_CLOBBER) | ||
| 220 | + | ||
| 221 | +DEF2(qemu_st8, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 222 | +DEF2(qemu_st16, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 223 | +DEF2(qemu_st32, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 224 | +DEF2(qemu_st64, 0, 2, 1, TCG_OPF_CALL_CLOBBER) | ||
| 225 | + | ||
| 226 | +#endif /* TCG_TARGET_REG_BITS != 32 */ | ||
| 227 | + | ||
| 228 | +#undef DEF2 |
tcg/tcg-runtime.c
0 โ 100644
| 1 | +/* | ||
| 2 | + * Tiny Code Generator for QEMU | ||
| 3 | + * | ||
| 4 | + * Copyright (c) 2008 Fabrice Bellard | ||
| 5 | + * | ||
| 6 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
| 7 | + * of this software and associated documentation files (the "Software"), to deal | ||
| 8 | + * in the Software without restriction, including without limitation the rights | ||
| 9 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
| 10 | + * copies of the Software, and to permit persons to whom the Software is | ||
| 11 | + * furnished to do so, subject to the following conditions: | ||
| 12 | + * | ||
| 13 | + * The above copyright notice and this permission notice shall be included in | ||
| 14 | + * all copies or substantial portions of the Software. | ||
| 15 | + * | ||
| 16 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 17 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 18 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
| 19 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
| 20 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
| 21 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
| 22 | + * THE SOFTWARE. | ||
| 23 | + */ | ||
| 24 | +#include <stdarg.h> | ||
| 25 | +#include <stdlib.h> | ||
| 26 | +#include <stdio.h> | ||
| 27 | +#include <string.h> | ||
| 28 | +#include <inttypes.h> | ||
| 29 | + | ||
| 30 | +#include "config.h" | ||
| 31 | +#include "osdep.h" | ||
| 32 | +#include "tcg.h" | ||
| 33 | + | ||
| 34 | +int64_t tcg_helper_shl_i64(int64_t arg1, int64_t arg2) | ||
| 35 | +{ | ||
| 36 | + return arg1 << arg2; | ||
| 37 | +} | ||
| 38 | + | ||
| 39 | +int64_t tcg_helper_shr_i64(int64_t arg1, int64_t arg2) | ||
| 40 | +{ | ||
| 41 | + return (uint64_t)arg1 >> arg2; | ||
| 42 | +} | ||
| 43 | + | ||
| 44 | +int64_t tcg_helper_sar_i64(int64_t arg1, int64_t arg2) | ||
| 45 | +{ | ||
| 46 | + return arg1 >> arg2; | ||
| 47 | +} | ||
| 48 | + | ||
| 49 | +int64_t tcg_helper_div_i64(int64_t arg1, int64_t arg2) | ||
| 50 | +{ | ||
| 51 | + return arg1 / arg2; | ||
| 52 | +} | ||
| 53 | + | ||
| 54 | +int64_t tcg_helper_rem_i64(int64_t arg1, int64_t arg2) | ||
| 55 | +{ | ||
| 56 | + return arg1 / arg2; | ||
| 57 | +} | ||
| 58 | + | ||
| 59 | +uint64_t tcg_helper_divu_i64(uint64_t arg1, uint64_t arg2) | ||
| 60 | +{ | ||
| 61 | + return arg1 / arg2; | ||
| 62 | +} | ||
| 63 | + | ||
| 64 | +uint64_t tcg_helper_remu_i64(uint64_t arg1, uint64_t arg2) | ||
| 65 | +{ | ||
| 66 | + return arg1 / arg2; | ||
| 67 | +} | ||
| 68 | + |