Commit f32fc64851c28e2dd3976d08f93006a3eff68a3d
1 parent
f1c85677
optional support for kernel code virtualization
git-svn-id: svn://svn.savannah.nongnu.org/qemu/trunk@1753 c046a42c-6fe2-441c-8c8c-71466251a162
Showing
4 changed files
with
216 additions
and
37 deletions
cpu-exec.c
| ... | ... | @@ -627,6 +627,9 @@ int cpu_exec(CPUState *env1) |
| 627 | 627 | jump. */ |
| 628 | 628 | { |
| 629 | 629 | if (T0 != 0 && |
| 630 | +#if USE_KQEMU | |
| 631 | + (env->kqemu_enabled != 2) && | |
| 632 | +#endif | |
| 630 | 633 | tb->page_addr[1] == -1 |
| 631 | 634 | #if defined(TARGET_I386) && defined(USE_CODE_COPY) |
| 632 | 635 | && (tb->cflags & CF_CODE_COPY) == |
| ... | ... | @@ -756,6 +759,13 @@ int cpu_exec(CPUState *env1) |
| 756 | 759 | T0 = 0; |
| 757 | 760 | } |
| 758 | 761 | #endif |
| 762 | +#if defined(USE_KQEMU) | |
| 763 | +#define MIN_CYCLE_BEFORE_SWITCH (100 * 1000) | |
| 764 | + if (kqemu_is_ok(env) && | |
| 765 | + (cpu_get_time_fast() - env->last_io_time) >= MIN_CYCLE_BEFORE_SWITCH) { | |
| 766 | + cpu_loop_exit(); | |
| 767 | + } | |
| 768 | +#endif | |
| 759 | 769 | } |
| 760 | 770 | } else { |
| 761 | 771 | env_to_regs(); | ... | ... |
exec-all.h
| ... | ... | @@ -577,21 +577,27 @@ static inline target_ulong get_phys_addr_code(CPUState *env, target_ulong addr) |
| 577 | 577 | |
| 578 | 578 | |
| 579 | 579 | #ifdef USE_KQEMU |
| 580 | +#define KQEMU_MODIFY_PAGE_MASK (0xff & ~(VGA_DIRTY_FLAG | CODE_DIRTY_FLAG)) | |
| 581 | + | |
| 580 | 582 | int kqemu_init(CPUState *env); |
| 581 | 583 | int kqemu_cpu_exec(CPUState *env); |
| 582 | 584 | void kqemu_flush_page(CPUState *env, target_ulong addr); |
| 583 | 585 | void kqemu_flush(CPUState *env, int global); |
| 584 | 586 | void kqemu_set_notdirty(CPUState *env, ram_addr_t ram_addr); |
| 587 | +void kqemu_modify_page(CPUState *env, ram_addr_t ram_addr); | |
| 585 | 588 | void kqemu_cpu_interrupt(CPUState *env); |
| 589 | +void kqemu_record_dump(void); | |
| 586 | 590 | |
| 587 | 591 | static inline int kqemu_is_ok(CPUState *env) |
| 588 | 592 | { |
| 589 | 593 | return(env->kqemu_enabled && |
| 590 | - (env->hflags & HF_CPL_MASK) == 3 && | |
| 591 | - (env->eflags & IOPL_MASK) != IOPL_MASK && | |
| 592 | 594 | (env->cr[0] & CR0_PE_MASK) && |
| 595 | + !(env->hflags & HF_INHIBIT_IRQ_MASK) && | |
| 593 | 596 | (env->eflags & IF_MASK) && |
| 594 | - !(env->eflags & VM_MASK)); | |
| 597 | + !(env->eflags & VM_MASK) && | |
| 598 | + (env->kqemu_enabled == 2 || | |
| 599 | + ((env->hflags & HF_CPL_MASK) == 3 && | |
| 600 | + (env->eflags & IOPL_MASK) != IOPL_MASK))); | |
| 595 | 601 | } |
| 596 | 602 | |
| 597 | 603 | #endif | ... | ... |
exec.c
| ... | ... | @@ -1796,6 +1796,11 @@ static void notdirty_mem_writeb(void *opaque, target_phys_addr_t addr, uint32_t |
| 1796 | 1796 | #endif |
| 1797 | 1797 | } |
| 1798 | 1798 | stb_p((uint8_t *)(long)addr, val); |
| 1799 | +#ifdef USE_KQEMU | |
| 1800 | + if (cpu_single_env->kqemu_enabled && | |
| 1801 | + (dirty_flags & KQEMU_MODIFY_PAGE_MASK) != KQEMU_MODIFY_PAGE_MASK) | |
| 1802 | + kqemu_modify_page(cpu_single_env, ram_addr); | |
| 1803 | +#endif | |
| 1799 | 1804 | dirty_flags |= (0xff & ~CODE_DIRTY_FLAG); |
| 1800 | 1805 | phys_ram_dirty[ram_addr >> TARGET_PAGE_BITS] = dirty_flags; |
| 1801 | 1806 | /* we remove the notdirty callback only if the code has been |
| ... | ... | @@ -1817,6 +1822,11 @@ static void notdirty_mem_writew(void *opaque, target_phys_addr_t addr, uint32_t |
| 1817 | 1822 | #endif |
| 1818 | 1823 | } |
| 1819 | 1824 | stw_p((uint8_t *)(long)addr, val); |
| 1825 | +#ifdef USE_KQEMU | |
| 1826 | + if (cpu_single_env->kqemu_enabled && | |
| 1827 | + (dirty_flags & KQEMU_MODIFY_PAGE_MASK) != KQEMU_MODIFY_PAGE_MASK) | |
| 1828 | + kqemu_modify_page(cpu_single_env, ram_addr); | |
| 1829 | +#endif | |
| 1820 | 1830 | dirty_flags |= (0xff & ~CODE_DIRTY_FLAG); |
| 1821 | 1831 | phys_ram_dirty[ram_addr >> TARGET_PAGE_BITS] = dirty_flags; |
| 1822 | 1832 | /* we remove the notdirty callback only if the code has been |
| ... | ... | @@ -1838,6 +1848,11 @@ static void notdirty_mem_writel(void *opaque, target_phys_addr_t addr, uint32_t |
| 1838 | 1848 | #endif |
| 1839 | 1849 | } |
| 1840 | 1850 | stl_p((uint8_t *)(long)addr, val); |
| 1851 | +#ifdef USE_KQEMU | |
| 1852 | + if (cpu_single_env->kqemu_enabled && | |
| 1853 | + (dirty_flags & KQEMU_MODIFY_PAGE_MASK) != KQEMU_MODIFY_PAGE_MASK) | |
| 1854 | + kqemu_modify_page(cpu_single_env, ram_addr); | |
| 1855 | +#endif | |
| 1841 | 1856 | dirty_flags |= (0xff & ~CODE_DIRTY_FLAG); |
| 1842 | 1857 | phys_ram_dirty[ram_addr >> TARGET_PAGE_BITS] = dirty_flags; |
| 1843 | 1858 | /* we remove the notdirty callback only if the code has been | ... | ... |
kqemu.c
| ... | ... | @@ -54,6 +54,9 @@ |
| 54 | 54 | #define KQEMU_MAX_RAM_PAGES_TO_UPDATE 512 |
| 55 | 55 | #define KQEMU_RAM_PAGES_UPDATE_ALL (KQEMU_MAX_RAM_PAGES_TO_UPDATE + 1) |
| 56 | 56 | #endif |
| 57 | +#ifndef KQEMU_MAX_MODIFIED_RAM_PAGES | |
| 58 | +#define KQEMU_MAX_MODIFIED_RAM_PAGES 512 | |
| 59 | +#endif | |
| 57 | 60 | |
| 58 | 61 | #ifdef _WIN32 |
| 59 | 62 | #define KQEMU_DEVICE "\\\\.\\kqemu" |
| ... | ... | @@ -71,11 +74,18 @@ int kqemu_fd = KQEMU_INVALID_FD; |
| 71 | 74 | #define kqemu_closefd(x) close(x) |
| 72 | 75 | #endif |
| 73 | 76 | |
| 77 | +/* 0 = not allowed | |
| 78 | + 1 = user kqemu | |
| 79 | + 2 = kernel kqemu | |
| 80 | +*/ | |
| 74 | 81 | int kqemu_allowed = 1; |
| 75 | 82 | unsigned long *pages_to_flush; |
| 76 | 83 | unsigned int nb_pages_to_flush; |
| 77 | 84 | unsigned long *ram_pages_to_update; |
| 78 | 85 | unsigned int nb_ram_pages_to_update; |
| 86 | +unsigned long *modified_ram_pages; | |
| 87 | +unsigned int nb_modified_ram_pages; | |
| 88 | +uint8_t *modified_ram_pages_table; | |
| 79 | 89 | extern uint32_t **l1_phys_map; |
| 80 | 90 | |
| 81 | 91 | #define cpuid(index, eax, ebx, ecx, edx) \ |
| ... | ... | @@ -185,6 +195,14 @@ int kqemu_init(CPUState *env) |
| 185 | 195 | if (!ram_pages_to_update) |
| 186 | 196 | goto fail; |
| 187 | 197 | |
| 198 | + modified_ram_pages = qemu_vmalloc(KQEMU_MAX_MODIFIED_RAM_PAGES * | |
| 199 | + sizeof(unsigned long)); | |
| 200 | + if (!modified_ram_pages) | |
| 201 | + goto fail; | |
| 202 | + modified_ram_pages_table = qemu_mallocz(phys_ram_size >> TARGET_PAGE_BITS); | |
| 203 | + if (!modified_ram_pages_table) | |
| 204 | + goto fail; | |
| 205 | + | |
| 188 | 206 | init.ram_base = phys_ram_base; |
| 189 | 207 | init.ram_size = phys_ram_size; |
| 190 | 208 | init.ram_dirty = phys_ram_dirty; |
| ... | ... | @@ -193,6 +211,9 @@ int kqemu_init(CPUState *env) |
| 193 | 211 | #if KQEMU_VERSION >= 0x010200 |
| 194 | 212 | init.ram_pages_to_update = ram_pages_to_update; |
| 195 | 213 | #endif |
| 214 | +#if KQEMU_VERSION >= 0x010300 | |
| 215 | + init.modified_ram_pages = modified_ram_pages; | |
| 216 | +#endif | |
| 196 | 217 | #ifdef _WIN32 |
| 197 | 218 | ret = DeviceIoControl(kqemu_fd, KQEMU_INIT, &init, sizeof(init), |
| 198 | 219 | NULL, 0, &temp, NULL) == TRUE ? 0 : -1; |
| ... | ... | @@ -207,7 +228,7 @@ int kqemu_init(CPUState *env) |
| 207 | 228 | return -1; |
| 208 | 229 | } |
| 209 | 230 | kqemu_update_cpuid(env); |
| 210 | - env->kqemu_enabled = 1; | |
| 231 | + env->kqemu_enabled = kqemu_allowed; | |
| 211 | 232 | nb_pages_to_flush = 0; |
| 212 | 233 | nb_ram_pages_to_update = 0; |
| 213 | 234 | return 0; |
| ... | ... | @@ -215,7 +236,7 @@ int kqemu_init(CPUState *env) |
| 215 | 236 | |
| 216 | 237 | void kqemu_flush_page(CPUState *env, target_ulong addr) |
| 217 | 238 | { |
| 218 | -#ifdef DEBUG | |
| 239 | +#if defined(DEBUG) | |
| 219 | 240 | if (loglevel & CPU_LOG_INT) { |
| 220 | 241 | fprintf(logfile, "kqemu_flush_page: addr=" TARGET_FMT_lx "\n", addr); |
| 221 | 242 | } |
| ... | ... | @@ -252,6 +273,49 @@ void kqemu_set_notdirty(CPUState *env, ram_addr_t ram_addr) |
| 252 | 273 | ram_pages_to_update[nb_ram_pages_to_update++] = ram_addr; |
| 253 | 274 | } |
| 254 | 275 | |
| 276 | +static void kqemu_reset_modified_ram_pages(void) | |
| 277 | +{ | |
| 278 | + int i; | |
| 279 | + unsigned long page_index; | |
| 280 | + | |
| 281 | + for(i = 0; i < nb_modified_ram_pages; i++) { | |
| 282 | + page_index = modified_ram_pages[i] >> TARGET_PAGE_BITS; | |
| 283 | + modified_ram_pages_table[page_index] = 0; | |
| 284 | + } | |
| 285 | + nb_modified_ram_pages = 0; | |
| 286 | +} | |
| 287 | + | |
| 288 | +void kqemu_modify_page(CPUState *env, ram_addr_t ram_addr) | |
| 289 | +{ | |
| 290 | + unsigned long page_index; | |
| 291 | + int ret; | |
| 292 | +#ifdef _WIN32 | |
| 293 | + DWORD temp; | |
| 294 | +#endif | |
| 295 | + | |
| 296 | + page_index = ram_addr >> TARGET_PAGE_BITS; | |
| 297 | + if (!modified_ram_pages_table[page_index]) { | |
| 298 | +#if 0 | |
| 299 | + printf("%d: modify_page=%08lx\n", nb_modified_ram_pages, ram_addr); | |
| 300 | +#endif | |
| 301 | + modified_ram_pages_table[page_index] = 1; | |
| 302 | + modified_ram_pages[nb_modified_ram_pages++] = ram_addr; | |
| 303 | + if (nb_modified_ram_pages >= KQEMU_MAX_MODIFIED_RAM_PAGES) { | |
| 304 | + /* flush */ | |
| 305 | +#ifdef _WIN32 | |
| 306 | + ret = DeviceIoControl(kqemu_fd, KQEMU_MODIFY_RAM_PAGES, | |
| 307 | + &nb_modified_ram_pages, | |
| 308 | + sizeof(nb_modified_ram_pages), | |
| 309 | + NULL, 0, &temp, NULL); | |
| 310 | +#else | |
| 311 | + ret = ioctl(kqemu_fd, KQEMU_MODIFY_RAM_PAGES, | |
| 312 | + &nb_modified_ram_pages); | |
| 313 | +#endif | |
| 314 | + kqemu_reset_modified_ram_pages(); | |
| 315 | + } | |
| 316 | + } | |
| 317 | +} | |
| 318 | + | |
| 255 | 319 | struct fpstate { |
| 256 | 320 | uint16_t fpuc; |
| 257 | 321 | uint16_t dummy1; |
| ... | ... | @@ -442,7 +506,7 @@ static int do_syscall(CPUState *env, |
| 442 | 506 | return 2; |
| 443 | 507 | } |
| 444 | 508 | |
| 445 | -#ifdef PROFILE | |
| 509 | +#ifdef CONFIG_PROFILER | |
| 446 | 510 | |
| 447 | 511 | #define PC_REC_SIZE 1 |
| 448 | 512 | #define PC_REC_HASH_BITS 16 |
| ... | ... | @@ -454,10 +518,10 @@ typedef struct PCRecord { |
| 454 | 518 | struct PCRecord *next; |
| 455 | 519 | } PCRecord; |
| 456 | 520 | |
| 457 | -PCRecord *pc_rec_hash[PC_REC_HASH_SIZE]; | |
| 458 | -int nb_pc_records; | |
| 521 | +static PCRecord *pc_rec_hash[PC_REC_HASH_SIZE]; | |
| 522 | +static int nb_pc_records; | |
| 459 | 523 | |
| 460 | -void kqemu_record_pc(unsigned long pc) | |
| 524 | +static void kqemu_record_pc(unsigned long pc) | |
| 461 | 525 | { |
| 462 | 526 | unsigned long h; |
| 463 | 527 | PCRecord **pr, *r; |
| ... | ... | @@ -484,7 +548,7 @@ void kqemu_record_pc(unsigned long pc) |
| 484 | 548 | nb_pc_records++; |
| 485 | 549 | } |
| 486 | 550 | |
| 487 | -int pc_rec_cmp(const void *p1, const void *p2) | |
| 551 | +static int pc_rec_cmp(const void *p1, const void *p2) | |
| 488 | 552 | { |
| 489 | 553 | PCRecord *r1 = *(PCRecord **)p1; |
| 490 | 554 | PCRecord *r2 = *(PCRecord **)p2; |
| ... | ... | @@ -496,6 +560,21 @@ int pc_rec_cmp(const void *p1, const void *p2) |
| 496 | 560 | return -1; |
| 497 | 561 | } |
| 498 | 562 | |
| 563 | +static void kqemu_record_flush(void) | |
| 564 | +{ | |
| 565 | + PCRecord *r, *r_next; | |
| 566 | + int h; | |
| 567 | + | |
| 568 | + for(h = 0; h < PC_REC_HASH_SIZE; h++) { | |
| 569 | + for(r = pc_rec_hash[h]; r != NULL; r = r_next) { | |
| 570 | + r_next = r->next; | |
| 571 | + free(r); | |
| 572 | + } | |
| 573 | + pc_rec_hash[h] = NULL; | |
| 574 | + } | |
| 575 | + nb_pc_records = 0; | |
| 576 | +} | |
| 577 | + | |
| 499 | 578 | void kqemu_record_dump(void) |
| 500 | 579 | { |
| 501 | 580 | PCRecord **pr, *r; |
| ... | ... | @@ -532,21 +611,26 @@ void kqemu_record_dump(void) |
| 532 | 611 | } |
| 533 | 612 | fclose(f); |
| 534 | 613 | free(pr); |
| 535 | -} | |
| 536 | -#else | |
| 537 | -void kqemu_record_dump(void) | |
| 538 | -{ | |
| 614 | + | |
| 615 | + kqemu_record_flush(); | |
| 539 | 616 | } |
| 540 | 617 | #endif |
| 541 | 618 | |
| 542 | 619 | int kqemu_cpu_exec(CPUState *env) |
| 543 | 620 | { |
| 544 | 621 | struct kqemu_cpu_state kcpu_state, *kenv = &kcpu_state; |
| 545 | - int ret; | |
| 622 | + int ret, cpl, i; | |
| 623 | +#ifdef CONFIG_PROFILER | |
| 624 | + int64_t ti; | |
| 625 | +#endif | |
| 626 | + | |
| 546 | 627 | #ifdef _WIN32 |
| 547 | 628 | DWORD temp; |
| 548 | 629 | #endif |
| 549 | 630 | |
| 631 | +#ifdef CONFIG_PROFILER | |
| 632 | + ti = profile_getclock(); | |
| 633 | +#endif | |
| 550 | 634 | #ifdef DEBUG |
| 551 | 635 | if (loglevel & CPU_LOG_INT) { |
| 552 | 636 | fprintf(logfile, "kqemu: cpu_exec: enter\n"); |
| ... | ... | @@ -569,6 +653,19 @@ int kqemu_cpu_exec(CPUState *env) |
| 569 | 653 | #if KQEMU_VERSION >= 0x010100 |
| 570 | 654 | kenv->efer = env->efer; |
| 571 | 655 | #endif |
| 656 | +#if KQEMU_VERSION >= 0x010300 | |
| 657 | + kenv->tsc_offset = 0; | |
| 658 | + kenv->star = env->star; | |
| 659 | + kenv->sysenter_cs = env->sysenter_cs; | |
| 660 | + kenv->sysenter_esp = env->sysenter_esp; | |
| 661 | + kenv->sysenter_eip = env->sysenter_eip; | |
| 662 | +#ifdef __x86_64__ | |
| 663 | + kenv->lstar = env->lstar; | |
| 664 | + kenv->cstar = env->cstar; | |
| 665 | + kenv->fmask = env->fmask; | |
| 666 | + kenv->kernelgsbase = env->kernelgsbase; | |
| 667 | +#endif | |
| 668 | +#endif | |
| 572 | 669 | if (env->dr[7] & 0xff) { |
| 573 | 670 | kenv->dr7 = env->dr[7]; |
| 574 | 671 | kenv->dr0 = env->dr[0]; |
| ... | ... | @@ -579,21 +676,24 @@ int kqemu_cpu_exec(CPUState *env) |
| 579 | 676 | kenv->dr7 = 0; |
| 580 | 677 | } |
| 581 | 678 | kenv->dr6 = env->dr[6]; |
| 582 | - kenv->cpl = 3; | |
| 679 | + cpl = (env->hflags & HF_CPL_MASK); | |
| 680 | + kenv->cpl = cpl; | |
| 583 | 681 | kenv->nb_pages_to_flush = nb_pages_to_flush; |
| 584 | - nb_pages_to_flush = 0; | |
| 585 | 682 | #if KQEMU_VERSION >= 0x010200 |
| 586 | - kenv->user_only = 1; | |
| 683 | + kenv->user_only = (env->kqemu_enabled == 1); | |
| 587 | 684 | kenv->nb_ram_pages_to_update = nb_ram_pages_to_update; |
| 588 | 685 | #endif |
| 589 | 686 | nb_ram_pages_to_update = 0; |
| 590 | 687 | |
| 591 | - if (!(kenv->cr0 & CR0_TS_MASK)) { | |
| 592 | - if (env->cpuid_features & CPUID_FXSR) | |
| 593 | - restore_native_fp_fxrstor(env); | |
| 594 | - else | |
| 595 | - restore_native_fp_frstor(env); | |
| 596 | - } | |
| 688 | +#if KQEMU_VERSION >= 0x010300 | |
| 689 | + kenv->nb_modified_ram_pages = nb_modified_ram_pages; | |
| 690 | +#endif | |
| 691 | + kqemu_reset_modified_ram_pages(); | |
| 692 | + | |
| 693 | + if (env->cpuid_features & CPUID_FXSR) | |
| 694 | + restore_native_fp_fxrstor(env); | |
| 695 | + else | |
| 696 | + restore_native_fp_frstor(env); | |
| 597 | 697 | |
| 598 | 698 | #ifdef _WIN32 |
| 599 | 699 | if (DeviceIoControl(kqemu_fd, KQEMU_EXEC, |
| ... | ... | @@ -612,30 +712,49 @@ int kqemu_cpu_exec(CPUState *env) |
| 612 | 712 | ret = ioctl(kqemu_fd, KQEMU_EXEC, kenv); |
| 613 | 713 | #endif |
| 614 | 714 | #endif |
| 615 | - if (!(kenv->cr0 & CR0_TS_MASK)) { | |
| 616 | - if (env->cpuid_features & CPUID_FXSR) | |
| 617 | - save_native_fp_fxsave(env); | |
| 618 | - else | |
| 619 | - save_native_fp_fsave(env); | |
| 620 | - } | |
| 715 | + if (env->cpuid_features & CPUID_FXSR) | |
| 716 | + save_native_fp_fxsave(env); | |
| 717 | + else | |
| 718 | + save_native_fp_fsave(env); | |
| 621 | 719 | |
| 622 | 720 | memcpy(env->regs, kenv->regs, sizeof(env->regs)); |
| 623 | 721 | env->eip = kenv->eip; |
| 624 | 722 | env->eflags = kenv->eflags; |
| 625 | 723 | memcpy(env->segs, kenv->segs, sizeof(env->segs)); |
| 724 | + cpu_x86_set_cpl(env, kenv->cpl); | |
| 725 | + memcpy(&env->ldt, &kenv->ldt, sizeof(env->ldt)); | |
| 626 | 726 | #if 0 |
| 627 | 727 | /* no need to restore that */ |
| 628 | - memcpy(env->ldt, kenv->ldt, sizeof(env->ldt)); | |
| 629 | 728 | memcpy(env->tr, kenv->tr, sizeof(env->tr)); |
| 630 | 729 | memcpy(env->gdt, kenv->gdt, sizeof(env->gdt)); |
| 631 | 730 | memcpy(env->idt, kenv->idt, sizeof(env->idt)); |
| 632 | - env->cr[0] = kenv->cr0; | |
| 633 | - env->cr[3] = kenv->cr3; | |
| 634 | - env->cr[4] = kenv->cr4; | |
| 635 | 731 | env->a20_mask = kenv->a20_mask; |
| 636 | 732 | #endif |
| 733 | + env->cr[0] = kenv->cr0; | |
| 734 | + env->cr[4] = kenv->cr4; | |
| 735 | + env->cr[3] = kenv->cr3; | |
| 637 | 736 | env->cr[2] = kenv->cr2; |
| 638 | 737 | env->dr[6] = kenv->dr6; |
| 738 | +#if KQEMU_VERSION >= 0x010300 | |
| 739 | +#ifdef __x86_64__ | |
| 740 | + env->kernelgsbase = kenv->kernelgsbase; | |
| 741 | +#endif | |
| 742 | +#endif | |
| 743 | + | |
| 744 | + /* flush pages as indicated by kqemu */ | |
| 745 | + if (kenv->nb_pages_to_flush >= KQEMU_FLUSH_ALL) { | |
| 746 | + tlb_flush(env, 1); | |
| 747 | + } else { | |
| 748 | + for(i = 0; i < kenv->nb_pages_to_flush; i++) { | |
| 749 | + tlb_flush_page(env, pages_to_flush[i]); | |
| 750 | + } | |
| 751 | + } | |
| 752 | + nb_pages_to_flush = 0; | |
| 753 | + | |
| 754 | +#ifdef CONFIG_PROFILER | |
| 755 | + kqemu_time += profile_getclock() - ti; | |
| 756 | + kqemu_exec_count++; | |
| 757 | +#endif | |
| 639 | 758 | |
| 640 | 759 | #if KQEMU_VERSION >= 0x010200 |
| 641 | 760 | if (kenv->nb_ram_pages_to_update > 0) { |
| ... | ... | @@ -643,6 +762,16 @@ int kqemu_cpu_exec(CPUState *env) |
| 643 | 762 | } |
| 644 | 763 | #endif |
| 645 | 764 | |
| 765 | +#if KQEMU_VERSION >= 0x010300 | |
| 766 | + if (kenv->nb_modified_ram_pages > 0) { | |
| 767 | + for(i = 0; i < kenv->nb_modified_ram_pages; i++) { | |
| 768 | + unsigned long addr; | |
| 769 | + addr = modified_ram_pages[i]; | |
| 770 | + tb_invalidate_phys_page_range(addr, addr + TARGET_PAGE_SIZE, 0); | |
| 771 | + } | |
| 772 | + } | |
| 773 | +#endif | |
| 774 | + | |
| 646 | 775 | /* restore the hidden flags */ |
| 647 | 776 | { |
| 648 | 777 | unsigned int new_hflags; |
| ... | ... | @@ -679,7 +808,14 @@ int kqemu_cpu_exec(CPUState *env) |
| 679 | 808 | ~(HF_CS32_MASK | HF_SS32_MASK | HF_CS64_MASK | HF_ADDSEG_MASK)) | |
| 680 | 809 | new_hflags; |
| 681 | 810 | } |
| 682 | - | |
| 811 | + /* update FPU flags */ | |
| 812 | + env->hflags = (env->hflags & ~(HF_MP_MASK | HF_EM_MASK | HF_TS_MASK)) | | |
| 813 | + ((env->cr[0] << (HF_MP_SHIFT - 1)) & (HF_MP_MASK | HF_EM_MASK | HF_TS_MASK)); | |
| 814 | + if (env->cr[4] & CR4_OSFXSR_MASK) | |
| 815 | + env->hflags |= HF_OSFXSR_MASK; | |
| 816 | + else | |
| 817 | + env->hflags &= ~HF_OSFXSR_MASK; | |
| 818 | + | |
| 683 | 819 | #ifdef DEBUG |
| 684 | 820 | if (loglevel & CPU_LOG_INT) { |
| 685 | 821 | fprintf(logfile, "kqemu: kqemu_cpu_exec: ret=0x%x\n", ret); |
| ... | ... | @@ -694,6 +830,9 @@ int kqemu_cpu_exec(CPUState *env) |
| 694 | 830 | env->error_code = 0; |
| 695 | 831 | env->exception_is_int = 1; |
| 696 | 832 | env->exception_next_eip = kenv->next_eip; |
| 833 | +#ifdef CONFIG_PROFILER | |
| 834 | + kqemu_ret_int_count++; | |
| 835 | +#endif | |
| 697 | 836 | #ifdef DEBUG |
| 698 | 837 | if (loglevel & CPU_LOG_INT) { |
| 699 | 838 | fprintf(logfile, "kqemu: interrupt v=%02x:\n", |
| ... | ... | @@ -707,6 +846,9 @@ int kqemu_cpu_exec(CPUState *env) |
| 707 | 846 | env->error_code = kenv->error_code; |
| 708 | 847 | env->exception_is_int = 0; |
| 709 | 848 | env->exception_next_eip = 0; |
| 849 | +#ifdef CONFIG_PROFILER | |
| 850 | + kqemu_ret_excp_count++; | |
| 851 | +#endif | |
| 710 | 852 | #ifdef DEBUG |
| 711 | 853 | if (loglevel & CPU_LOG_INT) { |
| 712 | 854 | fprintf(logfile, "kqemu: exception v=%02x e=%04x:\n", |
| ... | ... | @@ -716,6 +858,9 @@ int kqemu_cpu_exec(CPUState *env) |
| 716 | 858 | #endif |
| 717 | 859 | return 1; |
| 718 | 860 | } else if (ret == KQEMU_RET_INTR) { |
| 861 | +#ifdef CONFIG_PROFILER | |
| 862 | + kqemu_ret_intr_count++; | |
| 863 | +#endif | |
| 719 | 864 | #ifdef DEBUG |
| 720 | 865 | if (loglevel & CPU_LOG_INT) { |
| 721 | 866 | cpu_dump_state(env, logfile, fprintf, 0); |
| ... | ... | @@ -723,8 +868,11 @@ int kqemu_cpu_exec(CPUState *env) |
| 723 | 868 | #endif |
| 724 | 869 | return 0; |
| 725 | 870 | } else if (ret == KQEMU_RET_SOFTMMU) { |
| 726 | -#ifdef PROFILE | |
| 727 | - kqemu_record_pc(env->eip + env->segs[R_CS].base); | |
| 871 | +#ifdef CONFIG_PROFILER | |
| 872 | + { | |
| 873 | + unsigned long pc = env->eip + env->segs[R_CS].base; | |
| 874 | + kqemu_record_pc(pc); | |
| 875 | + } | |
| 728 | 876 | #endif |
| 729 | 877 | #ifdef DEBUG |
| 730 | 878 | if (loglevel & CPU_LOG_INT) { | ... | ... |