Commit 6fa4cea9e8e904f7aac0c3d4f73a883c9e1e53bd
1 parent
876d4b07
Infrastructure to support more than 2 MMU modes.
Add example for Alpha and PowerPC hypervisor mode. git-svn-id: svn://svn.savannah.nongnu.org/qemu/trunk@2596 c046a42c-6fe2-441c-8c8c-71466251a162
Showing
2 changed files
with
48 additions
and
1 deletions
cpu-defs.h
... | ... | @@ -108,6 +108,15 @@ typedef struct CPUTLBEntry { |
108 | 108 | target_phys_addr_t addend; |
109 | 109 | } CPUTLBEntry; |
110 | 110 | |
111 | +/* Alpha has 4 different running levels */ | |
112 | +#if defined(TARGET_ALPHA) | |
113 | +#define NB_MMU_MODES 4 | |
114 | +#elif defined(TARGET_PPC64H) /* PowerPC 64 with hypervisor mode support */ | |
115 | +#define NB_MMU_MODES 3 | |
116 | +#else | |
117 | +#define NB_MMU_MODES 2 | |
118 | +#endif | |
119 | + | |
111 | 120 | #define CPU_COMMON \ |
112 | 121 | struct TranslationBlock *current_tb; /* currently executing TB */ \ |
113 | 122 | /* soft mmu support */ \ |
... | ... | @@ -119,7 +128,7 @@ typedef struct CPUTLBEntry { |
119 | 128 | target_ulong mem_write_vaddr; /* target virtual addr at which the \ |
120 | 129 | memory was written */ \ |
121 | 130 | /* 0 = kernel, 1 = user */ \ |
122 | - CPUTLBEntry tlb_table[2][CPU_TLB_SIZE]; \ | |
131 | + CPUTLBEntry tlb_table[NB_MMU_MODES][CPU_TLB_SIZE]; \ | |
123 | 132 | struct TranslationBlock *tb_jmp_cache[TB_JMP_CACHE_SIZE]; \ |
124 | 133 | \ |
125 | 134 | /* from this point: preserved by CPU reset */ \ | ... | ... |
exec.c
... | ... | @@ -1300,6 +1300,16 @@ void tlb_flush(CPUState *env, int flush_global) |
1300 | 1300 | env->tlb_table[1][i].addr_read = -1; |
1301 | 1301 | env->tlb_table[1][i].addr_write = -1; |
1302 | 1302 | env->tlb_table[1][i].addr_code = -1; |
1303 | +#if (NB_MMU_MODES >= 3) | |
1304 | + env->tlb_table[2][i].addr_read = -1; | |
1305 | + env->tlb_table[2][i].addr_write = -1; | |
1306 | + env->tlb_table[2][i].addr_code = -1; | |
1307 | +#if (NB_MMU_MODES == 4) | |
1308 | + env->tlb_table[3][i].addr_read = -1; | |
1309 | + env->tlb_table[3][i].addr_write = -1; | |
1310 | + env->tlb_table[3][i].addr_code = -1; | |
1311 | +#endif | |
1312 | +#endif | |
1303 | 1313 | } |
1304 | 1314 | |
1305 | 1315 | memset (env->tb_jmp_cache, 0, TB_JMP_CACHE_SIZE * sizeof (void *)); |
... | ... | @@ -1345,6 +1355,12 @@ void tlb_flush_page(CPUState *env, target_ulong addr) |
1345 | 1355 | i = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); |
1346 | 1356 | tlb_flush_entry(&env->tlb_table[0][i], addr); |
1347 | 1357 | tlb_flush_entry(&env->tlb_table[1][i], addr); |
1358 | +#if (NB_MMU_MODES >= 3) | |
1359 | + tlb_flush_entry(&env->tlb_table[2][i], addr); | |
1360 | +#if (NB_MMU_MODES == 4) | |
1361 | + tlb_flush_entry(&env->tlb_table[3][i], addr); | |
1362 | +#endif | |
1363 | +#endif | |
1348 | 1364 | |
1349 | 1365 | /* Discard jump cache entries for any tb which might potentially |
1350 | 1366 | overlap the flushed page. */ |
... | ... | @@ -1434,6 +1450,14 @@ void cpu_physical_memory_reset_dirty(ram_addr_t start, ram_addr_t end, |
1434 | 1450 | tlb_reset_dirty_range(&env->tlb_table[0][i], start1, length); |
1435 | 1451 | for(i = 0; i < CPU_TLB_SIZE; i++) |
1436 | 1452 | tlb_reset_dirty_range(&env->tlb_table[1][i], start1, length); |
1453 | +#if (NB_MMU_MODES >= 3) | |
1454 | + for(i = 0; i < CPU_TLB_SIZE; i++) | |
1455 | + tlb_reset_dirty_range(&env->tlb_table[2][i], start1, length); | |
1456 | +#if (NB_MMU_MODES == 4) | |
1457 | + for(i = 0; i < CPU_TLB_SIZE; i++) | |
1458 | + tlb_reset_dirty_range(&env->tlb_table[3][i], start1, length); | |
1459 | +#endif | |
1460 | +#endif | |
1437 | 1461 | } |
1438 | 1462 | |
1439 | 1463 | #if !defined(CONFIG_SOFTMMU) |
... | ... | @@ -1486,6 +1510,14 @@ void cpu_tlb_update_dirty(CPUState *env) |
1486 | 1510 | tlb_update_dirty(&env->tlb_table[0][i]); |
1487 | 1511 | for(i = 0; i < CPU_TLB_SIZE; i++) |
1488 | 1512 | tlb_update_dirty(&env->tlb_table[1][i]); |
1513 | +#if (NB_MMU_MODES >= 3) | |
1514 | + for(i = 0; i < CPU_TLB_SIZE; i++) | |
1515 | + tlb_update_dirty(&env->tlb_table[2][i]); | |
1516 | +#if (NB_MMU_MODES == 4) | |
1517 | + for(i = 0; i < CPU_TLB_SIZE; i++) | |
1518 | + tlb_update_dirty(&env->tlb_table[3][i]); | |
1519 | +#endif | |
1520 | +#endif | |
1489 | 1521 | } |
1490 | 1522 | |
1491 | 1523 | static inline void tlb_set_dirty1(CPUTLBEntry *tlb_entry, |
... | ... | @@ -1511,6 +1543,12 @@ static inline void tlb_set_dirty(CPUState *env, |
1511 | 1543 | i = (vaddr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); |
1512 | 1544 | tlb_set_dirty1(&env->tlb_table[0][i], addr); |
1513 | 1545 | tlb_set_dirty1(&env->tlb_table[1][i], addr); |
1546 | +#if (NB_MMU_MODES >= 3) | |
1547 | + tlb_set_dirty1(&env->tlb_table[2][i], addr); | |
1548 | +#if (NB_MMU_MODES == 4) | |
1549 | + tlb_set_dirty1(&env->tlb_table[3][i], addr); | |
1550 | +#endif | |
1551 | +#endif | |
1514 | 1552 | } |
1515 | 1553 | |
1516 | 1554 | /* add a new TLB entry. At most one entry for a given virtual address | ... | ... |