Changeset 2426 in vbox
- Timestamp:
- Apr 30, 2007 12:36:15 PM (18 years ago)
- svn:sync-xref-src-repo-rev:
- 20835
- Location:
- trunk/src/recompiler
- Files:
-
- 8 edited
Legend:
- Unmodified
- Added
- Removed
-
trunk/src/recompiler/target-i386/helper.c
r1953 r2426 18 18 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA 19 19 */ 20 #ifdef VBOX 21 # include <VBox/err.h> 22 #endif 20 23 #include "exec.h" 21 #ifdef VBOX22 #include <VBox/err.h>23 #endif24 24 25 25 //#define DEBUG_PCALL … … 28 28 #define raise_exception_err(a, b)\ 29 29 do {\ 30 fprintf(logfile, "raise_exception line=%d\n", __LINE__);\ 30 if (logfile)\ 31 fprintf(logfile, "raise_exception line=%d\n", __LINE__);\ 31 32 (raise_exception_err)(a, b);\ 32 33 } while (0) … … 224 225 if ((e2 & DESC_C_MASK) && dpl > rpl) 225 226 raise_exception_err(EXCP0A_TSS, selector & 0xfffc); 226 227 227 } else if (seg_reg == R_SS) { 228 228 /* SS must be writable data */ … … 380 380 for(i = 0; i < 6; i++) 381 381 stw_kernel(env->tr.base + (0x48 + i * 4), env->segs[i].selector); 382 383 382 #if defined(VBOX) && defined(DEBUG) 384 383 printf("TSS 32 bits switch\n"); 385 384 printf("Saving CS=%08X\n", env->segs[R_CS].selector); 386 385 #endif 387 388 386 } else { 389 387 /* 16 bit */ … … 567 565 } 568 566 567 #ifdef TARGET_X86_64 568 #define SET_ESP(val, sp_mask)\ 569 do {\ 570 if ((sp_mask) == 0xffff)\ 571 ESP = (ESP & ~0xffff) | ((val) & 0xffff);\ 572 else if ((sp_mask) == 0xffffffffLL)\ 573 ESP = (uint32_t)(val);\ 574 else\ 575 ESP = (val);\ 576 } while (0) 577 #else 578 #define SET_ESP(val, sp_mask) ESP = (ESP & ~(sp_mask)) | ((val) & (sp_mask)) 579 #endif 580 569 581 /* XXX: add a is_user flag to have proper security support */ 570 582 #define PUSHW(ssp, sp, sp_mask, val)\ … … 598 610 SegmentCache *dt; 599 611 target_ulong ptr, ssp; 600 int type, dpl, selector, ss_dpl, cpl , sp_mask;612 int type, dpl, selector, ss_dpl, cpl; 601 613 int has_error_code, new_stack, shift; 602 614 uint32_t e1, e2, offset, ss, esp, ss_e1, ss_e2; 603 uint32_t old_eip ;615 uint32_t old_eip, sp_mask; 604 616 605 617 #ifdef VBOX … … 642 654 switch_tss(intno * 8, e1, e2, SWITCH_TSS_CALL, old_eip); 643 655 if (has_error_code) { 644 int mask, type; 656 int type; 657 uint32_t mask; 645 658 /* push the error code */ 646 659 type = (env->tr.flags >> DESC_TYPE_SHIFT) & 0xf; … … 656 669 else 657 670 stw_kernel(ssp, error_code); 658 ESP = (esp & mask) | (ESP & ~mask);671 SET_ESP(esp, mask); 659 672 } 660 673 return; … … 787 800 ssp, get_seg_limit(ss_e1, ss_e2), ss_e2); 788 801 } 789 ESP = (ESP & ~sp_mask) | (esp &sp_mask);802 SET_ESP(esp, sp_mask); 790 803 791 804 selector = (selector & ~3) | dpl; … … 896 909 env->eflags &= ~IF_MASK; 897 910 } 898 #endif 911 #endif /* VBOX */ 899 912 900 913 #ifdef TARGET_X86_64 … … 1078 1091 cpu_x86_load_seg_cache(env, R_CS, selector & 0xfffc, 1079 1092 0, 0xffffffff, 1080 DESC_G_MASK | DESC_ B_MASK | DESC_P_MASK |1093 DESC_G_MASK | DESC_P_MASK | 1081 1094 DESC_S_MASK | 1082 1095 DESC_CS_MASK | DESC_R_MASK | DESC_A_MASK | DESC_L_MASK); … … 1129 1142 cpu_x86_load_seg_cache(env, R_CS, (selector + 16) | 3, 1130 1143 0, 0xffffffff, 1131 DESC_G_MASK | DESC_ B_MASK | DESC_P_MASK |1144 DESC_G_MASK | DESC_P_MASK | 1132 1145 DESC_S_MASK | (3 << DESC_DPL_SHIFT) | 1133 1146 DESC_CS_MASK | DESC_R_MASK | DESC_A_MASK | … … 1184 1197 void helper_external_event(void) 1185 1198 { 1199 #if defined(__DARWIN__) && defined(VBOX_STRICT) 1200 # if 0 1201 //uintptr_t uFrameAddr = (uintptr_t)__builtin_frame_address(0); - this is broken (uses %ebp) 1202 //AssertMsg(!( (uFrameAddr - sizeof(uintptr_t)) & 7 ), ("uFrameAddr=%#p\n", uFrameAddr)); 1203 # else 1204 uintptr_t uESP; 1205 __asm__ __volatile__("movl %%esp, %0" : "=r" (uESP)); 1206 AssertMsg(!(uESP & 15), ("esp=%#p\n", uESP)); 1207 # endif 1208 #endif 1186 1209 if (env->interrupt_request & CPU_INTERRUPT_EXTERNAL_HARD) 1187 1210 { … … 1278 1301 target_ulong next_eip, int is_hw) 1279 1302 { 1280 #ifdef DEBUG_PCALL 1281 if (loglevel & (CPU_LOG_PCALL | CPU_LOG_INT)) { 1303 if (loglevel & CPU_LOG_INT) { 1282 1304 if ((env->cr[0] & CR0_PE_MASK)) { 1283 1305 static int count; … … 1294 1316 } 1295 1317 fprintf(logfile, "\n"); 1318 cpu_dump_state(env, logfile, fprintf, X86_DUMP_CCOP); 1296 1319 #if 0 1297 cpu_dump_state(env, logfile, fprintf, X86_DUMP_CCOP);1298 1320 { 1299 1321 int i; … … 1310 1332 } 1311 1333 } 1312 #endif1313 1334 if (env->cr[0] & CR0_PE_MASK) { 1314 1335 #if TARGET_X86_64 … … 1328 1349 do_soft_interrupt_vme(intno, error_code, next_eip); 1329 1350 else 1330 #endif 1351 #endif /* VBOX */ 1331 1352 do_interrupt_protected(intno, is_int, error_code, next_eip, is_hw); 1332 1353 } … … 1376 1397 raise_interrupt(exception_index, 0, 0, 0); 1377 1398 } 1399 1400 /* SMM support */ 1401 1402 #if defined(CONFIG_USER_ONLY) 1403 1404 void do_smm_enter(void) 1405 { 1406 } 1407 1408 void helper_rsm(void) 1409 { 1410 } 1411 1412 #else 1413 1414 #ifdef TARGET_X86_64 1415 #define SMM_REVISION_ID 0x00020064 1416 #else 1417 #define SMM_REVISION_ID 0x00020000 1418 #endif 1419 1420 void do_smm_enter(void) 1421 { 1422 #ifdef VBOX 1423 cpu_abort(env, "do_ssm_enter"); 1424 #else /* !VBOX */ 1425 target_ulong sm_state; 1426 SegmentCache *dt; 1427 int i, offset; 1428 1429 if (loglevel & CPU_LOG_INT) { 1430 fprintf(logfile, "SMM: enter\n"); 1431 cpu_dump_state(env, logfile, fprintf, X86_DUMP_CCOP); 1432 } 1433 1434 env->hflags |= HF_SMM_MASK; 1435 cpu_smm_update(env); 1436 1437 sm_state = env->smbase + 0x8000; 1438 1439 #ifdef TARGET_X86_64 1440 for(i = 0; i < 6; i++) { 1441 dt = &env->segs[i]; 1442 offset = 0x7e00 + i * 16; 1443 stw_phys(sm_state + offset, dt->selector); 1444 stw_phys(sm_state + offset + 2, (dt->flags >> 8) & 0xf0ff); 1445 stl_phys(sm_state + offset + 4, dt->limit); 1446 stq_phys(sm_state + offset + 8, dt->base); 1447 } 1448 1449 stq_phys(sm_state + 0x7e68, env->gdt.base); 1450 stl_phys(sm_state + 0x7e64, env->gdt.limit); 1451 1452 stw_phys(sm_state + 0x7e70, env->ldt.selector); 1453 stq_phys(sm_state + 0x7e78, env->ldt.base); 1454 stl_phys(sm_state + 0x7e74, env->ldt.limit); 1455 stw_phys(sm_state + 0x7e72, (env->ldt.flags >> 8) & 0xf0ff); 1456 1457 stq_phys(sm_state + 0x7e88, env->idt.base); 1458 stl_phys(sm_state + 0x7e84, env->idt.limit); 1459 1460 stw_phys(sm_state + 0x7e90, env->tr.selector); 1461 stq_phys(sm_state + 0x7e98, env->tr.base); 1462 stl_phys(sm_state + 0x7e94, env->tr.limit); 1463 stw_phys(sm_state + 0x7e92, (env->tr.flags >> 8) & 0xf0ff); 1464 1465 stq_phys(sm_state + 0x7ed0, env->efer); 1466 1467 stq_phys(sm_state + 0x7ff8, EAX); 1468 stq_phys(sm_state + 0x7ff0, ECX); 1469 stq_phys(sm_state + 0x7fe8, EDX); 1470 stq_phys(sm_state + 0x7fe0, EBX); 1471 stq_phys(sm_state + 0x7fd8, ESP); 1472 stq_phys(sm_state + 0x7fd0, EBP); 1473 stq_phys(sm_state + 0x7fc8, ESI); 1474 stq_phys(sm_state + 0x7fc0, EDI); 1475 for(i = 8; i < 16; i++) 1476 stq_phys(sm_state + 0x7ff8 - i * 8, env->regs[i]); 1477 stq_phys(sm_state + 0x7f78, env->eip); 1478 stl_phys(sm_state + 0x7f70, compute_eflags()); 1479 stl_phys(sm_state + 0x7f68, env->dr[6]); 1480 stl_phys(sm_state + 0x7f60, env->dr[7]); 1481 1482 stl_phys(sm_state + 0x7f48, env->cr[4]); 1483 stl_phys(sm_state + 0x7f50, env->cr[3]); 1484 stl_phys(sm_state + 0x7f58, env->cr[0]); 1485 1486 stl_phys(sm_state + 0x7efc, SMM_REVISION_ID); 1487 stl_phys(sm_state + 0x7f00, env->smbase); 1488 #else 1489 stl_phys(sm_state + 0x7ffc, env->cr[0]); 1490 stl_phys(sm_state + 0x7ff8, env->cr[3]); 1491 stl_phys(sm_state + 0x7ff4, compute_eflags()); 1492 stl_phys(sm_state + 0x7ff0, env->eip); 1493 stl_phys(sm_state + 0x7fec, EDI); 1494 stl_phys(sm_state + 0x7fe8, ESI); 1495 stl_phys(sm_state + 0x7fe4, EBP); 1496 stl_phys(sm_state + 0x7fe0, ESP); 1497 stl_phys(sm_state + 0x7fdc, EBX); 1498 stl_phys(sm_state + 0x7fd8, EDX); 1499 stl_phys(sm_state + 0x7fd4, ECX); 1500 stl_phys(sm_state + 0x7fd0, EAX); 1501 stl_phys(sm_state + 0x7fcc, env->dr[6]); 1502 stl_phys(sm_state + 0x7fc8, env->dr[7]); 1503 1504 stl_phys(sm_state + 0x7fc4, env->tr.selector); 1505 stl_phys(sm_state + 0x7f64, env->tr.base); 1506 stl_phys(sm_state + 0x7f60, env->tr.limit); 1507 stl_phys(sm_state + 0x7f5c, (env->tr.flags >> 8) & 0xf0ff); 1508 1509 stl_phys(sm_state + 0x7fc0, env->ldt.selector); 1510 stl_phys(sm_state + 0x7f80, env->ldt.base); 1511 stl_phys(sm_state + 0x7f7c, env->ldt.limit); 1512 stl_phys(sm_state + 0x7f78, (env->ldt.flags >> 8) & 0xf0ff); 1513 1514 stl_phys(sm_state + 0x7f74, env->gdt.base); 1515 stl_phys(sm_state + 0x7f70, env->gdt.limit); 1516 1517 stl_phys(sm_state + 0x7f58, env->idt.base); 1518 stl_phys(sm_state + 0x7f54, env->idt.limit); 1519 1520 for(i = 0; i < 6; i++) { 1521 dt = &env->segs[i]; 1522 if (i < 3) 1523 offset = 0x7f84 + i * 12; 1524 else 1525 offset = 0x7f2c + (i - 3) * 12; 1526 stl_phys(sm_state + 0x7fa8 + i * 4, dt->selector); 1527 stl_phys(sm_state + offset + 8, dt->base); 1528 stl_phys(sm_state + offset + 4, dt->limit); 1529 stl_phys(sm_state + offset, (dt->flags >> 8) & 0xf0ff); 1530 } 1531 stl_phys(sm_state + 0x7f14, env->cr[4]); 1532 1533 stl_phys(sm_state + 0x7efc, SMM_REVISION_ID); 1534 stl_phys(sm_state + 0x7ef8, env->smbase); 1535 #endif 1536 /* init SMM cpu state */ 1537 1538 #ifdef TARGET_X86_64 1539 env->efer = 0; 1540 env->hflags &= ~HF_LMA_MASK; 1541 #endif 1542 load_eflags(0, ~(CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C | DF_MASK)); 1543 env->eip = 0x00008000; 1544 cpu_x86_load_seg_cache(env, R_CS, (env->smbase >> 4) & 0xffff, env->smbase, 1545 0xffffffff, 0); 1546 cpu_x86_load_seg_cache(env, R_DS, 0, 0, 0xffffffff, 0); 1547 cpu_x86_load_seg_cache(env, R_ES, 0, 0, 0xffffffff, 0); 1548 cpu_x86_load_seg_cache(env, R_SS, 0, 0, 0xffffffff, 0); 1549 cpu_x86_load_seg_cache(env, R_FS, 0, 0, 0xffffffff, 0); 1550 cpu_x86_load_seg_cache(env, R_GS, 0, 0, 0xffffffff, 0); 1551 1552 cpu_x86_update_cr0(env, 1553 env->cr[0] & ~(CR0_PE_MASK | CR0_EM_MASK | CR0_TS_MASK | CR0_PG_MASK)); 1554 cpu_x86_update_cr4(env, 0); 1555 env->dr[7] = 0x00000400; 1556 CC_OP = CC_OP_EFLAGS; 1557 #endif /* VBOX */ 1558 } 1559 1560 void helper_rsm(void) 1561 { 1562 #ifdef VBOX 1563 cpu_abort(env, "helper_rsm"); 1564 #else /* !VBOX */ 1565 target_ulong sm_state; 1566 int i, offset; 1567 uint32_t val; 1568 1569 sm_state = env->smbase + 0x8000; 1570 #ifdef TARGET_X86_64 1571 env->efer = ldq_phys(sm_state + 0x7ed0); 1572 if (env->efer & MSR_EFER_LMA) 1573 env->hflags |= HF_LMA_MASK; 1574 else 1575 env->hflags &= ~HF_LMA_MASK; 1576 1577 for(i = 0; i < 6; i++) { 1578 offset = 0x7e00 + i * 16; 1579 cpu_x86_load_seg_cache(env, i, 1580 lduw_phys(sm_state + offset), 1581 ldq_phys(sm_state + offset + 8), 1582 ldl_phys(sm_state + offset + 4), 1583 (lduw_phys(sm_state + offset + 2) & 0xf0ff) << 8); 1584 } 1585 1586 env->gdt.base = ldq_phys(sm_state + 0x7e68); 1587 env->gdt.limit = ldl_phys(sm_state + 0x7e64); 1588 1589 env->ldt.selector = lduw_phys(sm_state + 0x7e70); 1590 env->ldt.base = ldq_phys(sm_state + 0x7e78); 1591 env->ldt.limit = ldl_phys(sm_state + 0x7e74); 1592 env->ldt.flags = (lduw_phys(sm_state + 0x7e72) & 0xf0ff) << 8; 1593 1594 env->idt.base = ldq_phys(sm_state + 0x7e88); 1595 env->idt.limit = ldl_phys(sm_state + 0x7e84); 1596 1597 env->tr.selector = lduw_phys(sm_state + 0x7e90); 1598 env->tr.base = ldq_phys(sm_state + 0x7e98); 1599 env->tr.limit = ldl_phys(sm_state + 0x7e94); 1600 env->tr.flags = (lduw_phys(sm_state + 0x7e92) & 0xf0ff) << 8; 1601 1602 EAX = ldq_phys(sm_state + 0x7ff8); 1603 ECX = ldq_phys(sm_state + 0x7ff0); 1604 EDX = ldq_phys(sm_state + 0x7fe8); 1605 EBX = ldq_phys(sm_state + 0x7fe0); 1606 ESP = ldq_phys(sm_state + 0x7fd8); 1607 EBP = ldq_phys(sm_state + 0x7fd0); 1608 ESI = ldq_phys(sm_state + 0x7fc8); 1609 EDI = ldq_phys(sm_state + 0x7fc0); 1610 for(i = 8; i < 16; i++) 1611 env->regs[i] = ldq_phys(sm_state + 0x7ff8 - i * 8); 1612 env->eip = ldq_phys(sm_state + 0x7f78); 1613 load_eflags(ldl_phys(sm_state + 0x7f70), 1614 ~(CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C | DF_MASK)); 1615 env->dr[6] = ldl_phys(sm_state + 0x7f68); 1616 env->dr[7] = ldl_phys(sm_state + 0x7f60); 1617 1618 cpu_x86_update_cr4(env, ldl_phys(sm_state + 0x7f48)); 1619 cpu_x86_update_cr3(env, ldl_phys(sm_state + 0x7f50)); 1620 cpu_x86_update_cr0(env, ldl_phys(sm_state + 0x7f58)); 1621 1622 val = ldl_phys(sm_state + 0x7efc); /* revision ID */ 1623 if (val & 0x20000) { 1624 env->smbase = ldl_phys(sm_state + 0x7f00) & ~0x7fff; 1625 } 1626 #else 1627 cpu_x86_update_cr0(env, ldl_phys(sm_state + 0x7ffc)); 1628 cpu_x86_update_cr3(env, ldl_phys(sm_state + 0x7ff8)); 1629 load_eflags(ldl_phys(sm_state + 0x7ff4), 1630 ~(CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C | DF_MASK)); 1631 env->eip = ldl_phys(sm_state + 0x7ff0); 1632 EDI = ldl_phys(sm_state + 0x7fec); 1633 ESI = ldl_phys(sm_state + 0x7fe8); 1634 EBP = ldl_phys(sm_state + 0x7fe4); 1635 ESP = ldl_phys(sm_state + 0x7fe0); 1636 EBX = ldl_phys(sm_state + 0x7fdc); 1637 EDX = ldl_phys(sm_state + 0x7fd8); 1638 ECX = ldl_phys(sm_state + 0x7fd4); 1639 EAX = ldl_phys(sm_state + 0x7fd0); 1640 env->dr[6] = ldl_phys(sm_state + 0x7fcc); 1641 env->dr[7] = ldl_phys(sm_state + 0x7fc8); 1642 1643 env->tr.selector = ldl_phys(sm_state + 0x7fc4) & 0xffff; 1644 env->tr.base = ldl_phys(sm_state + 0x7f64); 1645 env->tr.limit = ldl_phys(sm_state + 0x7f60); 1646 env->tr.flags = (ldl_phys(sm_state + 0x7f5c) & 0xf0ff) << 8; 1647 1648 env->ldt.selector = ldl_phys(sm_state + 0x7fc0) & 0xffff; 1649 env->ldt.base = ldl_phys(sm_state + 0x7f80); 1650 env->ldt.limit = ldl_phys(sm_state + 0x7f7c); 1651 env->ldt.flags = (ldl_phys(sm_state + 0x7f78) & 0xf0ff) << 8; 1652 1653 env->gdt.base = ldl_phys(sm_state + 0x7f74); 1654 env->gdt.limit = ldl_phys(sm_state + 0x7f70); 1655 1656 env->idt.base = ldl_phys(sm_state + 0x7f58); 1657 env->idt.limit = ldl_phys(sm_state + 0x7f54); 1658 1659 for(i = 0; i < 6; i++) { 1660 if (i < 3) 1661 offset = 0x7f84 + i * 12; 1662 else 1663 offset = 0x7f2c + (i - 3) * 12; 1664 cpu_x86_load_seg_cache(env, i, 1665 ldl_phys(sm_state + 0x7fa8 + i * 4) & 0xffff, 1666 ldl_phys(sm_state + offset + 8), 1667 ldl_phys(sm_state + offset + 4), 1668 (ldl_phys(sm_state + offset) & 0xf0ff) << 8); 1669 } 1670 cpu_x86_update_cr4(env, ldl_phys(sm_state + 0x7f14)); 1671 1672 val = ldl_phys(sm_state + 0x7efc); /* revision ID */ 1673 if (val & 0x20000) { 1674 env->smbase = ldl_phys(sm_state + 0x7ef8) & ~0x7fff; 1675 } 1676 #endif 1677 CC_OP = CC_OP_EFLAGS; 1678 env->hflags &= ~HF_SMM_MASK; 1679 cpu_smm_update(env); 1680 1681 if (loglevel & CPU_LOG_INT) { 1682 fprintf(logfile, "SMM: after RSM\n"); 1683 cpu_dump_state(env, logfile, fprintf, X86_DUMP_CCOP); 1684 } 1685 #endif /* !VBOX */ 1686 } 1687 1688 #endif /* !CONFIG_USER_ONLY */ 1689 1378 1690 1379 1691 #ifdef BUGGY_GCC_DIV64 … … 1458 1770 { 1459 1771 #ifndef VBOX 1460 switch((uint32_t)EAX) { 1772 uint32_t index; 1773 index = (uint32_t)EAX; 1774 1775 /* test if maximum index reached */ 1776 if (index & 0x80000000) { 1777 if (index > env->cpuid_xlevel) 1778 index = env->cpuid_level; 1779 } else { 1780 if (index > env->cpuid_level) 1781 index = env->cpuid_level; 1782 } 1783 1784 switch(index) { 1461 1785 case 0: 1462 EAX = 2; /* max EAX index supported */1786 EAX = env->cpuid_level; 1463 1787 EBX = env->cpuid_vendor1; 1464 1788 EDX = env->cpuid_vendor2; … … 1467 1791 case 1: 1468 1792 EAX = env->cpuid_version; 1469 EBX = 0;1793 EBX = 8 << 8; /* CLFLUSH size in quad words, Linux wants it. */ 1470 1794 ECX = env->cpuid_ext_features; 1471 1795 EDX = env->cpuid_features; 1472 1796 break; 1473 1474 default: 1797 case 2: 1475 1798 /* cache info: needed for Pentium Pro compatibility */ 1476 1799 EAX = 0x410601; … … 1479 1802 EDX = 0; 1480 1803 break; 1481 1482 #ifdef TARGET_X86_641483 1804 case 0x80000000: 1484 EAX = 0x80000008;1805 EAX = env->cpuid_xlevel; 1485 1806 EBX = env->cpuid_vendor1; 1486 1807 EDX = env->cpuid_vendor2; … … 1491 1812 EBX = 0; 1492 1813 ECX = 0; 1493 /* long mode + syscall/sysret features */ 1494 EDX = (env->cpuid_features & 0x0183F3FF) | (1 << 29) | (1 << 11); 1814 EDX = env->cpuid_ext2_features; 1815 break; 1816 case 0x80000002: 1817 case 0x80000003: 1818 case 0x80000004: 1819 EAX = env->cpuid_model[(index - 0x80000002) * 4 + 0]; 1820 EBX = env->cpuid_model[(index - 0x80000002) * 4 + 1]; 1821 ECX = env->cpuid_model[(index - 0x80000002) * 4 + 2]; 1822 EDX = env->cpuid_model[(index - 0x80000002) * 4 + 3]; 1823 break; 1824 case 0x80000005: 1825 /* cache info (L1 cache) */ 1826 EAX = 0x01ff01ff; 1827 EBX = 0x01ff01ff; 1828 ECX = 0x40020140; 1829 EDX = 0x40020140; 1830 break; 1831 case 0x80000006: 1832 /* cache info (L2 cache) */ 1833 EAX = 0; 1834 EBX = 0x42004200; 1835 ECX = 0x02008140; 1836 EDX = 0; 1495 1837 break; 1496 1838 case 0x80000008: … … 1501 1843 EDX = 0; 1502 1844 break; 1503 #endif 1845 default: 1846 /* reserved values: zero */ 1847 EAX = 0; 1848 EBX = 0; 1849 ECX = 0; 1850 EDX = 0; 1851 break; 1504 1852 } 1505 1853 #else /* VBOX */ … … 1634 1982 int index, type, entry_limit; 1635 1983 target_ulong ptr; 1984 1636 1985 #ifdef VBOX 1637 1986 Log(("helper_ltr_T0: old tr=%RTsel {.base=%VGv, .limit=%VGv, .flags=%RX32} new=%RTsel\n", … … 1750 2099 if (!(e2 & DESC_CS_MASK) || !(e2 & DESC_C_MASK)) { 1751 2100 /* if not conforming code, test rights */ 1752 if (dpl < cpl || dpl < rpl) 2101 if (dpl < cpl || dpl < rpl) 1753 2102 raise_exception_err(EXCP0D_GPF, selector & 0xfffc); 1754 2103 } … … 1890 2239 } 1891 2240 1892 ESP = (ESP & ~esp_mask) | (esp &esp_mask);2241 SET_ESP(esp, esp_mask); 1893 2242 env->eip = new_eip; 1894 2243 env->segs[R_CS].selector = new_cs; … … 1976 2325 raise_exception_err(EXCP0D_GPF, new_cs & 0xfffc); 1977 2326 /* from this point, not restartable */ 1978 ESP = (ESP & ~sp_mask) | (sp &sp_mask);2327 SET_ESP(sp, sp_mask); 1979 2328 cpu_x86_load_seg_cache(env, R_CS, (new_cs & 0xfffc) | cpl, 1980 2329 get_seg_base(e1, e2), limit, e2); … … 2105 2454 e2); 2106 2455 cpu_x86_set_cpl(env, dpl); 2107 ESP = (ESP & ~sp_mask) | (sp &sp_mask);2456 SET_ESP(sp, sp_mask); 2108 2457 EIP = offset; 2109 2458 } … … 2122 2471 target_ulong ssp; 2123 2472 int eflags_mask; 2124 2125 2473 #ifdef VBOX 2126 2474 bool fVME = false; 2127 2475 2128 2476 remR3TrapClear(env->pVM); 2129 #endif 2477 #endif /* VBOX */ 2130 2478 2131 2479 sp_mask = 0xffff; /* XXXX: use SS segment size ? */ … … 2154 2502 if ( ((new_eflags & IF_MASK) && (env->eflags & VIP_MASK)) 2155 2503 || (new_eflags & TF_MASK)) 2156 {2157 2504 raise_exception(EXCP0D_GPF); 2158 } 2159 } 2160 #endif 2505 } 2506 #endif /* VBOX */ 2161 2507 2162 2508 ESP = (ESP & ~sp_mask) | (sp & sp_mask); … … 2174 2520 if (shift == 0) 2175 2521 eflags_mask &= 0xffff; 2176 2177 2522 load_eflags(new_eflags, eflags_mask); 2178 2523 … … 2185 2530 env->eflags &= ~VIF_MASK; 2186 2531 } 2187 #endif 2532 #endif /* VBOX */ 2188 2533 } 2189 2534 … … 2192 2537 int dpl; 2193 2538 uint32_t e2; 2194 2539 2540 /* XXX: on x86_64, we do not want to nullify FS and GS because 2541 they may still contain a valid base. I would be interested to 2542 know how a real x86_64 CPU behaves */ 2543 if ((seg_reg == R_FS || seg_reg == R_GS) && 2544 (env->segs[seg_reg].selector & 0xfffc) == 0) 2545 return; 2546 2195 2547 e2 = env->segs[seg_reg].flags; 2196 2548 dpl = (e2 >> DESC_DPL_SHIFT) & 3; … … 2217 2569 else 2218 2570 #endif 2219 sp_mask = get_sp_mask(env->segs[R_SS].flags);2571 sp_mask = get_sp_mask(env->segs[R_SS].flags); 2220 2572 sp = ESP; 2221 2573 ssp = env->segs[R_SS].base; … … 2363 2715 #ifdef TARGET_X86_64 2364 2716 /* NULL ss is allowed in long mode if cpl != 3*/ 2717 /* XXX: test CS64 ? */ 2365 2718 if ((env->hflags & HF_LMA_MASK) && rpl != 3) { 2366 2719 cpu_x86_load_seg_cache(env, R_SS, new_ss, … … 2369 2722 DESC_S_MASK | (rpl << DESC_DPL_SHIFT) | 2370 2723 DESC_W_MASK | DESC_A_MASK); 2724 ss_e2 = DESC_B_MASK; /* XXX: should not be needed ? */ 2371 2725 } else 2372 2726 #endif … … 2415 2769 sp += addend; 2416 2770 } 2417 ESP = (ESP & ~sp_mask) | (sp &sp_mask);2771 SET_ESP(sp, sp_mask); 2418 2772 env->eip = new_eip; 2419 2773 if (is_iret) { … … 2426 2780 eflags_mask |= IOPL_MASK; 2427 2781 #endif 2428 2429 2782 iopl = (env->eflags >> IOPL_SHIFT) & 3; 2430 2783 if (cpl <= iopl) … … 2495 2848 helper_ret_protected(shift, 1, 0); 2496 2849 } 2850 #ifdef USE_KQEMU 2851 if (kqemu_is_ok(env)) { 2852 CC_OP = CC_OP_EFLAGS; 2853 env->exception_index = -1; 2854 cpu_loop_exit(); 2855 } 2856 #endif 2497 2857 } 2498 2858 … … 2500 2860 { 2501 2861 helper_ret_protected(shift, 0, addend); 2862 #ifdef USE_KQEMU 2863 if (kqemu_is_ok(env)) { 2864 env->exception_index = -1; 2865 cpu_loop_exit(); 2866 } 2867 #endif 2502 2868 } 2503 2869 … … 2544 2910 ESP = ECX; 2545 2911 EIP = EDX; 2912 #ifdef USE_KQEMU 2913 if (kqemu_is_ok(env)) { 2914 env->exception_index = -1; 2915 cpu_loop_exit(); 2916 } 2917 #endif 2546 2918 } 2547 2919 … … 2575 2947 } 2576 2948 2577 void helper_invlpg( unsigned intaddr)2949 void helper_invlpg(target_ulong addr) 2578 2950 { 2579 2951 cpu_x86_flush_tlb(env, addr); … … 2586 2958 if ((env->cr[4] & CR4_TSD_MASK) && ((env->hflags & HF_CPL_MASK) != 0)) { 2587 2959 raise_exception(EXCP0D_GPF); 2588 } 2960 } 2589 2961 val = cpu_get_tsc(env); 2590 2962 EAX = (uint32_t)(val); … … 2620 2992 cpu_set_apic_base(env, val); 2621 2993 break; 2622 #ifdef TARGET_X86_642623 2994 case MSR_EFER: 2624 #define MSR_EFER_UPDATE_MASK (MSR_EFER_SCE | MSR_EFER_LME | \ 2625 MSR_EFER_NXE | MSR_EFER_FFXSR) 2626 env->efer = (env->efer & ~MSR_EFER_UPDATE_MASK) | 2627 (val & MSR_EFER_UPDATE_MASK); 2995 { 2996 uint64_t update_mask; 2997 update_mask = 0; 2998 if (env->cpuid_ext2_features & CPUID_EXT2_SYSCALL) 2999 update_mask |= MSR_EFER_SCE; 3000 if (env->cpuid_ext2_features & CPUID_EXT2_LM) 3001 update_mask |= MSR_EFER_LME; 3002 if (env->cpuid_ext2_features & CPUID_EXT2_FFXSR) 3003 update_mask |= MSR_EFER_FFXSR; 3004 if (env->cpuid_ext2_features & CPUID_EXT2_NX) 3005 update_mask |= MSR_EFER_NXE; 3006 env->efer = (env->efer & ~update_mask) | 3007 (val & update_mask); 3008 } 2628 3009 break; 2629 3010 case MSR_STAR: 2630 3011 env->star = val; 2631 3012 break; 3013 case MSR_PAT: 3014 env->pat = val; 3015 break; 3016 #ifdef TARGET_X86_64 2632 3017 case MSR_LSTAR: 2633 3018 env->lstar = val; … … 2671 3056 val = cpu_get_apic_base(env); 2672 3057 break; 2673 #ifdef TARGET_X86_642674 3058 case MSR_EFER: 2675 3059 val = env->efer; … … 2678 3062 val = env->star; 2679 3063 break; 3064 case MSR_PAT: 3065 val = env->pat; 3066 break; 3067 #ifdef TARGET_X86_64 2680 3068 case MSR_LSTAR: 2681 3069 val = env->lstar; … … 2927 3315 void helper_fbst_ST0_A0(void) 2928 3316 { 2929 CPU86_LDouble tmp;2930 3317 int v; 2931 3318 target_ulong mem_ref, mem_end; 2932 3319 int64_t val; 2933 3320 2934 tmp = rint(ST0); 2935 val = (int64_t)tmp; 3321 val = floatx_to_int64(ST0, &env->fp_status); 2936 3322 mem_ref = A0; 2937 3323 mem_end = mem_ref + 9; … … 3126 3512 void helper_frndint(void) 3127 3513 { 3128 CPU86_LDouble a; 3129 3130 a = ST0; 3131 #ifdef __arm__ 3132 switch(env->fpuc & RC_MASK) { 3133 default: 3134 case RC_NEAR: 3135 asm("rndd %0, %1" : "=f" (a) : "f"(a)); 3136 break; 3137 case RC_DOWN: 3138 asm("rnddm %0, %1" : "=f" (a) : "f"(a)); 3139 break; 3140 case RC_UP: 3141 asm("rnddp %0, %1" : "=f" (a) : "f"(a)); 3142 break; 3143 case RC_CHOP: 3144 asm("rnddz %0, %1" : "=f" (a) : "f"(a)); 3145 break; 3146 } 3147 #else 3148 a = rint(a); 3149 #endif 3150 ST0 = a; 3514 ST0 = floatx_round_to_int(ST0, &env->fp_status); 3151 3515 } 3152 3516 3153 3517 void helper_fscale(void) 3154 3518 { 3155 CPU86_LDouble fpsrcop, fptemp; 3156 3157 fpsrcop = 2.0; 3158 fptemp = pow(fpsrcop,ST1); 3159 ST0 *= fptemp; 3519 ST0 = ldexp (ST0, (int)(ST1)); 3160 3520 } 3161 3521 … … 3395 3755 3396 3756 if (env->cr[4] & CR4_OSFXSR_MASK) { 3397 /* XXX: finish it , endianness*/3757 /* XXX: finish it */ 3398 3758 env->mxcsr = ldl(ptr + 0x18); 3399 3759 //ldl(ptr + 0x1c); … … 3523 3883 *phigh += v; 3524 3884 #ifdef DEBUG_MULDIV 3525 printf("mul: 0x%016 llx * 0x%016llx = 0x%016llx%016llx\n",3885 printf("mul: 0x%016" PRIx64 " * 0x%016" PRIx64 " = 0x%016" PRIx64 "%016" PRIx64 "\n", 3526 3886 a, b, *phigh, *plow); 3527 3887 #endif … … 3572 3932 } 3573 3933 #if defined(DEBUG_MULDIV) 3574 printf("div: 0x%016 llx%016llx / 0x%016llx: q=0x%016llx r=0x%016llx\n",3934 printf("div: 0x%016" PRIx64 "%016" PRIx64 " / 0x%016" PRIx64 ": q=0x%016" PRIx64 " r=0x%016" PRIx64 "\n", 3575 3935 *phigh, *plow, b, a0, a1); 3576 3936 #endif … … 3666 4026 } 3667 4027 3668 #endif 3669 3670 /* XXX: do it */ 3671 int fpu_isnan(double a) 3672 { 3673 return 0; 3674 } 4028 void helper_bswapq_T0(void) 4029 { 4030 T0 = bswap64(T0); 4031 } 4032 #endif 3675 4033 3676 4034 void helper_hlt(void) 3677 4035 { 3678 4036 env->hflags &= ~HF_INHIBIT_IRQ_MASK; /* needed if sti is just before */ 4037 env->hflags |= HF_HALTED_MASK; 3679 4038 env->exception_index = EXCP_HLT; 3680 4039 cpu_loop_exit(); … … 3683 4042 void helper_monitor(void) 3684 4043 { 3685 if ( ECX != 0)4044 if ((uint32_t)ECX != 0) 3686 4045 raise_exception(EXCP0D_GPF); 3687 4046 /* XXX: store address ? */ … … 3690 4049 void helper_mwait(void) 3691 4050 { 3692 if ( ECX != 0)4051 if ((uint32_t)ECX != 0) 3693 4052 raise_exception(EXCP0D_GPF); 3694 4053 #ifdef VBOX … … 3715 4074 } 3716 4075 3717 /* XXX: find a better solution */ 3718 double helper_sqrt(double a) 3719 { 3720 return sqrt(a); 4076 void update_fp_status(void) 4077 { 4078 int rnd_type; 4079 4080 /* set rounding mode */ 4081 switch(env->fpuc & RC_MASK) { 4082 default: 4083 case RC_NEAR: 4084 rnd_type = float_round_nearest_even; 4085 break; 4086 case RC_DOWN: 4087 rnd_type = float_round_down; 4088 break; 4089 case RC_UP: 4090 rnd_type = float_round_up; 4091 break; 4092 case RC_CHOP: 4093 rnd_type = float_round_to_zero; 4094 break; 4095 } 4096 set_float_rounding_mode(rnd_type, &env->fp_status); 4097 #ifdef FLOATX80 4098 switch((env->fpuc >> 8) & 3) { 4099 case 0: 4100 rnd_type = 32; 4101 break; 4102 case 2: 4103 rnd_type = 64; 4104 break; 4105 case 3: 4106 default: 4107 rnd_type = 80; 4108 break; 4109 } 4110 set_floatx80_rounding_precision(rnd_type, &env->fp_status); 4111 #endif 3721 4112 } 3722 4113 … … 3769 4160 } 3770 4161 if (retaddr) 3771 raise_exception_err(EXCP0E_PAGE, env->error_code); 3772 else { 3773 raise_exception_err_norestore(EXCP0E_PAGE, env->error_code); 3774 } 4162 raise_exception_err(env->exception_index, env->error_code); 4163 else 4164 raise_exception_err_norestore(env->exception_index, env->error_code); 3775 4165 } 3776 4166 env = saved_env; 3777 4167 } 3778 4168 3779 #if defined(VBOX)4169 #ifdef VBOX 3780 4170 3781 4171 /** … … 4029 4419 int csize; 4030 4420 void (*gen_func)(void); 4031 uint8_t * pvCode;4421 uint8_t *tc_ptr; 4032 4422 uint32_t old_eip; 4033 4423 … … 4038 4428 RAWEx_ProfileStart(env, STATS_EMULATE_SINGLE_INSTR); 4039 4429 4040 pvCode = env->pvCodeBuffer; 4041 4042 // Setup temporary translation block 4043 tb_temp.hash_next = 0; 4044 tb_temp.jmp_first = 0; 4045 tb_temp.jmp_next[0] = 0; 4046 tb_temp.jmp_next[1] = 0; 4430 tc_ptr = env->pvCodeBuffer; 4431 4432 /* 4433 * Setup temporary translation block. 4434 */ 4435 /* tb_alloc: */ 4436 tb_temp.pc = env->segs[R_CS].base + env->eip; 4437 tb_temp.cflags = 0; 4438 4439 /* tb_find_slow: */ 4440 tb_temp.tc_ptr = tc_ptr; 4441 tb_temp.cs_base = env->segs[R_CS].base; 4442 tb_temp.flags = env->hflags | (env->eflags & (IOPL_MASK | TF_MASK | VM_MASK)); 4443 4444 /* Initialize the rest with sensible values. */ 4445 tb_temp.size = 0; 4446 tb_temp.phys_hash_next = NULL; 4447 tb_temp.page_next[0] = NULL; 4448 tb_temp.page_next[1] = NULL; 4047 4449 tb_temp.page_addr[0] = 0; 4048 4450 tb_temp.page_addr[1] = 0; 4049 tb_temp.page_next[0] = 0;4050 tb_temp.page_next[1] = 0;4051 tb_temp.hash_next = 0;4052 4053 4451 tb_temp.tb_next_offset[0] = 0xffff; 4054 4452 tb_temp.tb_next_offset[1] = 0xffff; 4055 4056 tb_temp.cs_base = (unsigned long)env->segs[R_CS].base; 4057 tb_temp.pc = tb_temp.cs_base + env->eip; 4058 tb_temp.cflags = 0; 4059 tb_temp.flags = env->hflags; 4060 tb_temp.flags |= (env->eflags & (IOPL_MASK | TF_MASK | VM_MASK)); 4061 tb_temp.tc_ptr = pvCode; 4453 tb_temp.tb_next[0] = 0xffff; 4454 tb_temp.tb_next[1] = 0xffff; 4455 tb_temp.jmp_next[0] = NULL; 4456 tb_temp.jmp_next[1] = NULL; 4457 tb_temp.jmp_first = NULL; 4062 4458 4063 4459 current = env->current_tb; 4064 4460 env->current_tb = NULL; 4065 4461 4066 // Translate only one instruction 4067 ASMAtomicOrS32(&env->state, CPU_EMULATE_SINGLE_INSTR); 4462 /* 4463 * Translate only one instruction. 4464 */ 4465 ASMAtomicOrU32(&env->state, CPU_EMULATE_SINGLE_INSTR); 4068 4466 if (cpu_gen_code(env, &tb_temp, env->cbCodeBuffer, &csize) < 0) 4069 4467 { 4070 4468 AssertFailed(); 4071 4469 RAWEx_ProfileStop(env, STATS_EMULATE_SINGLE_INSTR); 4072 ASMAtomicAnd S32(&env->state, ~CPU_EMULATE_SINGLE_INSTR);4470 ASMAtomicAndU32(&env->state, ~CPU_EMULATE_SINGLE_INSTR); 4073 4471 env = savedenv; 4074 4472 return -1; … … 4079 4477 RAWEx_ProfileStop(env, STATS_EMULATE_SINGLE_INSTR); 4080 4478 AssertFailed(); 4081 ASMAtomicAnd S32(&env->state, ~CPU_EMULATE_SINGLE_INSTR);4479 ASMAtomicAndU32(&env->state, ~CPU_EMULATE_SINGLE_INSTR); 4082 4480 env = savedenv; 4083 4481 return -1; 4084 4482 } 4085 if (tb_temp.tc_ptr != pvCode)4483 if (tb_temp.tc_ptr != tc_ptr) 4086 4484 { 4087 4485 RAWEx_ProfileStop(env, STATS_EMULATE_SINGLE_INSTR); 4088 4486 AssertFailed(); 4089 ASMAtomicAnd S32(&env->state, ~CPU_EMULATE_SINGLE_INSTR);4487 ASMAtomicAndU32(&env->state, ~CPU_EMULATE_SINGLE_INSTR); 4090 4488 env = savedenv; 4091 4489 return -1; 4092 4490 } 4093 4491 #endif 4094 ASMAtomicAndS32(&env->state, ~CPU_EMULATE_SINGLE_INSTR); 4095 4096 tb_link(&tb_temp); 4097 4492 ASMAtomicAndU32(&env->state, ~CPU_EMULATE_SINGLE_INSTR); 4493 4494 /* tb_link_phys: */ 4495 tb_temp.jmp_first = (TranslationBlock *)((intptr_t)&tb_temp | 2); 4496 Assert(tb_temp.jmp_next[0] == NULL); Assert(tb_temp.jmp_next[1] == NULL); 4497 if (tb_temp.tb_next_offset[0] != 0xffff) 4498 tb_set_jmp_target(&tb_temp, 0, (uintptr_t)(tb_temp.tc_ptr + tb_temp.tb_next_offset[0])); 4499 if (tb_temp.tb_next_offset[1] != 0xffff) 4500 tb_set_jmp_target(&tb_temp, 1, (uintptr_t)(tb_temp.tc_ptr + tb_temp.tb_next_offset[1])); 4501 4502 /* 4503 * Execute it using emulation 4504 */ 4098 4505 old_eip = env->eip; 4099 // Execute it using emulation4100 4506 gen_func = (void *)tb_temp.tc_ptr; 4101 4507 env->current_tb = &tb_temp; … … 4119 4525 env->current_tb = current; 4120 4526 4527 Assert(tb_temp.phys_hash_next == NULL); 4528 Assert(tb_temp.page_next[0] == NULL); 4529 Assert(tb_temp.page_next[1] == NULL); 4530 Assert(tb_temp.page_addr[0] == 0); 4531 Assert(tb_temp.page_addr[1] == 0); 4532 /* 4533 Assert(tb_temp.tb_next_offset[0] == 0xffff); 4534 Assert(tb_temp.tb_next_offset[1] == 0xffff); 4535 Assert(tb_temp.tb_next[0] == 0xffff); 4536 Assert(tb_temp.tb_next[1] == 0xffff); 4537 Assert(tb_temp.jmp_next[0] == NULL); 4538 Assert(tb_temp.jmp_next[1] == NULL); 4539 Assert(tb_temp.jmp_first == NULL); */ 4540 4121 4541 RAWEx_ProfileStop(env, STATS_EMULATE_SINGLE_INSTR); 4122 4542 4543 /* 4544 * Execute the next instruction when we encounter instruction fusing. 4545 */ 4123 4546 if (env->hflags & HF_INHIBIT_IRQ_MASK) 4124 4547 { -
trunk/src/recompiler/tests/Makefile
r1 r2426 1 #-include ../config-host.mak2 CC=gcc 1 -include ../config-host.mak 2 3 3 CFLAGS=-Wall -O2 -g 4 #CFLAGS+=-msse2 4 5 LDFLAGS= 5 6 6 7 ifeq ($(ARCH),i386) 7 8 TESTS=linux-test testthread sha1-i386 test-i386 runcom 9 endif 10 ifeq ($(ARCH),x86_64) 11 TESTS=test-x86_64 8 12 endif 9 13 TESTS+=sha1# test_path … … 25 29 ./$@ || { rm $@; exit 1; } 26 30 27 # i386 emulation test (test various opcodes) */31 # i386/x86_64 emulation test (test various opcodes) */ 28 32 test-i386: test-i386.c test-i386-code16.S test-i386-vm86.S \ 29 33 test-i386.h test-i386-shift.h test-i386-muldiv.h 30 $(CC) $(CFLAGS) $(LDFLAGS) -static -o $@ test-i386.c \ 31 test-i386-code16.S test-i386-vm86.S -lm 34 $(CC) $(CFLAGS) $(LDFLAGS) -static -o $@ \ 35 test-i386.c test-i386-code16.S test-i386-vm86.S -lm 36 37 test-x86_64: test-i386.c \ 38 test-i386.h test-i386-shift.h test-i386-muldiv.h 39 $(CC) $(CFLAGS) $(LDFLAGS) -static -o $@ test-i386.c -lm 32 40 33 41 ifeq ($(ARCH),i386) … … 65 73 # NOTE: -fomit-frame-pointer is currently needed : this is a bug in libqemu 66 74 qruncom: qruncom.c ../i386-user/libqemu.a 67 $(CC) $(CFLAGS) -fomit-frame-pointer $(LDFLAGS) -I../target-i386 -I.. -I../i386-user \75 $(CC) $(CFLAGS) -fomit-frame-pointer $(LDFLAGS) -I../target-i386 -I.. -I../i386-user -I../fpu \ 68 76 -o $@ $< -L../i386-user -lqemu -lm 69 77 … … 75 83 arm-linux-gcc -Wall -g -O2 -c -o $@ $< 76 84 85 # MIPS test 86 hello-mips: hello-mips.c 87 mips-linux-gnu-gcc -nostdlib -static -mno-abicalls -fno-PIC -mabi=32 -Wall -Wextra -g -O2 -o $@ $< 88 89 hello-mipsel: hello-mips.c 90 mipsel-linux-gnu-gcc -nostdlib -static -mno-abicalls -fno-PIC -mabi=32 -Wall -Wextra -g -O2 -o $@ $< 91 77 92 # XXX: find a way to compile easily a test for each arch 78 93 test2: 79 @for arch in i386 arm sparc ppc; do \94 @for arch in i386 arm armeb sparc ppc mips mipsel; do \ 80 95 ../$${arch}-user/qemu-$${arch} $${arch}/ls -l linux-test.c ; \ 81 96 done 82 97 83 98 clean: 84 rm -f *~ *.o test-i386.out test-i386.ref qruncom $(TESTS) 99 rm -f *~ *.o test-i386.out test-i386.ref \ 100 test-x86_64.log test-x86_64.ref qruncom $(TESTS) -
trunk/src/recompiler/tests/qruncom.c
r1 r2426 11 11 #include <sys/mman.h> 12 12 #include <signal.h> 13 #include <malloc.h> 13 14 14 15 #include "cpu.h" 15 16 16 17 //#define SIGTEST 17 18 CPUState *cpu_single_env = NULL;19 18 20 19 void cpu_outb(CPUState *env, int addr, int val) … … 87 86 { 88 87 return malloc(size); 88 } 89 90 void *qemu_mallocz(size_t size) 91 { 92 void *ptr; 93 ptr = qemu_malloc(size); 94 if (!ptr) 95 return NULL; 96 memset(ptr, 0, size); 97 return ptr; 98 } 99 100 void *qemu_vmalloc(size_t size) 101 { 102 return memalign(4096, size); 103 } 104 105 void qemu_vfree(void *ptr) 106 { 107 free(ptr); 89 108 } 90 109 … … 207 226 208 227 cpu_x86_load_seg_cache(env, R_CS, seg, 209 ( uint8_t *)(seg << 4), 0xffff, 0);228 (seg << 4), 0xffff, 0); 210 229 cpu_x86_load_seg_cache(env, R_SS, seg, 211 ( uint8_t *)(seg << 4), 0xffff, 0);230 (seg << 4), 0xffff, 0); 212 231 cpu_x86_load_seg_cache(env, R_DS, seg, 213 ( uint8_t *)(seg << 4), 0xffff, 0);232 (seg << 4), 0xffff, 0); 214 233 cpu_x86_load_seg_cache(env, R_ES, seg, 215 ( uint8_t *)(seg << 4), 0xffff, 0);234 (seg << 4), 0xffff, 0); 216 235 cpu_x86_load_seg_cache(env, R_FS, seg, 217 ( uint8_t *)(seg << 4), 0xffff, 0);236 (seg << 4), 0xffff, 0); 218 237 cpu_x86_load_seg_cache(env, R_GS, seg, 219 ( uint8_t *)(seg << 4), 0xffff, 0);238 (seg << 4), 0xffff, 0); 220 239 221 240 /* exception support */ 222 env->idt.base = ( void *)idt_table;241 env->idt.base = (unsigned long)idt_table; 223 242 env->idt.limit = sizeof(idt_table) - 1; 224 243 set_idt(0, 0); … … 266 285 { 267 286 int int_num, ah; 268 int_num = *( env->segs[R_CS].base + env->eip + 1);287 int_num = *(uint8_t *)(env->segs[R_CS].base + env->eip + 1); 269 288 if (int_num != 0x21) 270 289 goto unknown_int; … … 294 313 unknown_int: 295 314 fprintf(stderr, "unsupported int 0x%02x\n", int_num); 296 cpu_dump_state(env, stderr, 0);315 cpu_dump_state(env, stderr, fprintf, 0); 297 316 // exit(1); 298 317 } … … 302 321 default: 303 322 fprintf(stderr, "unhandled cpu_exec return code (0x%x)\n", ret); 304 cpu_dump_state(env, stderr, 0);323 cpu_dump_state(env, stderr, fprintf, 0); 305 324 exit(1); 306 325 } -
trunk/src/recompiler/tests/test-i386-code16.S
r1 r2426 78 78 79 79 code16_end: 80 81 82 /* other 32 bits tests */83 .code3284 85 .globl func_lret3286 func_lret32:87 movl $0x87654321, %eax88 lret89 90 .globl func_iret3291 func_iret32:92 movl $0xabcd4321, %eax93 iret94 95 96 97 -
trunk/src/recompiler/tests/test-i386-muldiv.h
r1 r2426 1 1 2 void glue(glue(test_, OP), b)( int op0, intop1)2 void glue(glue(test_, OP), b)(long op0, long op1) 3 3 { 4 intres, s1, s0, flags;4 long res, s1, s0, flags; 5 5 s0 = op0; 6 6 s1 = op1; … … 11 11 stringify(OP)"b %b2\n\t" 12 12 "pushf\n\t" 13 "pop l%1\n\t"13 "pop %1\n\t" 14 14 : "=a" (res), "=g" (flags) 15 15 : "q" (s1), "0" (res), "1" (flags)); 16 printf("%-10s A= %08x B=%08x R=%08x CC=%04x\n",16 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CC=%04lx\n", 17 17 stringify(OP) "b", s0, s1, res, flags & CC_MASK); 18 18 } 19 19 20 void glue(glue(test_, OP), w)( int op0h, int op0, intop1)20 void glue(glue(test_, OP), w)(long op0h, long op0, long op1) 21 21 { 22 intres, s1, flags, resh;22 long res, s1, flags, resh; 23 23 s1 = op1; 24 24 resh = op0h; … … 29 29 stringify(OP) "w %w3\n\t" 30 30 "pushf\n\t" 31 "pop l%1\n\t"31 "pop %1\n\t" 32 32 : "=a" (res), "=g" (flags), "=d" (resh) 33 33 : "q" (s1), "0" (res), "1" (flags), "2" (resh)); 34 printf("%-10s AH= %08x AL=%08x B=%08x RH=%08x RL=%08x CC=%04x\n",34 printf("%-10s AH=" FMTLX " AL=" FMTLX " B=" FMTLX " RH=" FMTLX " RL=" FMTLX " CC=%04lx\n", 35 35 stringify(OP) "w", op0h, op0, s1, resh, res, flags & CC_MASK); 36 36 } 37 37 38 void glue(glue(test_, OP), l)( int op0h, int op0, intop1)38 void glue(glue(test_, OP), l)(long op0h, long op0, long op1) 39 39 { 40 intres, s1, flags, resh;40 long res, s1, flags, resh; 41 41 s1 = op1; 42 42 resh = op0h; … … 45 45 asm ("push %5\n\t" 46 46 "popf\n\t" 47 stringify(OP) "l % 3\n\t"47 stringify(OP) "l %k3\n\t" 48 48 "pushf\n\t" 49 "pop l%1\n\t"49 "pop %1\n\t" 50 50 : "=a" (res), "=g" (flags), "=d" (resh) 51 51 : "q" (s1), "0" (res), "1" (flags), "2" (resh)); 52 printf("%-10s AH= %08x AL=%08x B=%08x RH=%08x RL=%08x CC=%04x\n",52 printf("%-10s AH=" FMTLX " AL=" FMTLX " B=" FMTLX " RH=" FMTLX " RL=" FMTLX " CC=%04lx\n", 53 53 stringify(OP) "l", op0h, op0, s1, resh, res, flags & CC_MASK); 54 54 } 55 55 56 #if defined(__x86_64__) 57 void glue(glue(test_, OP), q)(long op0h, long op0, long op1) 58 { 59 long res, s1, flags, resh; 60 s1 = op1; 61 resh = op0h; 62 res = op0; 63 flags = 0; 64 asm ("push %5\n\t" 65 "popf\n\t" 66 stringify(OP) "q %3\n\t" 67 "pushf\n\t" 68 "pop %1\n\t" 69 : "=a" (res), "=g" (flags), "=d" (resh) 70 : "q" (s1), "0" (res), "1" (flags), "2" (resh)); 71 printf("%-10s AH=" FMTLX " AL=" FMTLX " B=" FMTLX " RH=" FMTLX " RL=" FMTLX " CC=%04lx\n", 72 stringify(OP) "q", op0h, op0, s1, resh, res, flags & CC_MASK); 73 } 74 #endif 75 56 76 #undef OP -
trunk/src/recompiler/tests/test-i386-shift.h
r1 r2426 1 1 2 2 #define exec_op glue(exec_, OP) 3 #define exec_opq glue(glue(exec_, OP), q) 3 4 #define exec_opl glue(glue(exec_, OP), l) 4 5 #define exec_opw glue(glue(exec_, OP), w) … … 8 9 9 10 #ifdef OP_NOBYTE 10 #define EXECSHIFT(size, r es, s1, s2, flags) \11 #define EXECSHIFT(size, rsize, res, s1, s2, flags) \ 11 12 asm ("push %4\n\t"\ 12 13 "popf\n\t"\ 13 stringify(OP) size " %" size "2, %"size "0\n\t" \14 stringify(OP) size " %" rsize "2, %" rsize "0\n\t" \ 14 15 "pushf\n\t"\ 15 "pop l%1\n\t"\16 "pop %1\n\t"\ 16 17 : "=g" (res), "=g" (flags)\ 17 18 : "r" (s1), "0" (res), "1" (flags)); 18 19 #else 19 #define EXECSHIFT(size, r es, s1, s2, flags) \20 #define EXECSHIFT(size, rsize, res, s1, s2, flags) \ 20 21 asm ("push %4\n\t"\ 21 22 "popf\n\t"\ 22 stringify(OP) size " %%cl, %" size "0\n\t" \23 stringify(OP) size " %%cl, %" rsize "0\n\t" \ 23 24 "pushf\n\t"\ 24 "pop l%1\n\t"\25 "pop %1\n\t"\ 25 26 : "=q" (res), "=g" (flags)\ 26 27 : "c" (s1), "0" (res), "1" (flags)); 27 28 #endif 28 29 29 void exec_opl(int s2, int s0, int s1, int iflags) 30 #if defined(__x86_64__) 31 void exec_opq(long s2, long s0, long s1, long iflags) 30 32 { 31 intres, flags;33 long res, flags; 32 34 res = s0; 33 35 flags = iflags; 34 EXECSHIFT(" ", res, s1, s2, flags);36 EXECSHIFT("q", "", res, s1, s2, flags); 35 37 /* overflow is undefined if count != 1 */ 36 38 if (s1 != 1) 37 39 flags &= ~CC_O; 38 printf("%-10s A=%08x B=%08x R=%08x CCIN=%04x CC=%04x\n", 40 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CCIN=%04lx CC=%04lx\n", 41 stringify(OP) "q", s0, s1, res, iflags, flags & CC_MASK); 42 } 43 #endif 44 45 void exec_opl(long s2, long s0, long s1, long iflags) 46 { 47 long res, flags; 48 res = s0; 49 flags = iflags; 50 EXECSHIFT("l", "k", res, s1, s2, flags); 51 /* overflow is undefined if count != 1 */ 52 if (s1 != 1) 53 flags &= ~CC_O; 54 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CCIN=%04lx CC=%04lx\n", 39 55 stringify(OP) "l", s0, s1, res, iflags, flags & CC_MASK); 40 56 } 41 57 42 void exec_opw( int s2, int s0, int s1, intiflags)58 void exec_opw(long s2, long s0, long s1, long iflags) 43 59 { 44 intres, flags;60 long res, flags; 45 61 res = s0; 46 62 flags = iflags; 47 EXECSHIFT("w", res, s1, s2, flags);63 EXECSHIFT("w", "w", res, s1, s2, flags); 48 64 /* overflow is undefined if count != 1 */ 49 65 if (s1 != 1) 50 66 flags &= ~CC_O; 51 printf("%-10s A= %08x B=%08x R=%08x CCIN=%04x CC=%04x\n",67 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CCIN=%04lx CC=%04lx\n", 52 68 stringify(OP) "w", s0, s1, res, iflags, flags & CC_MASK); 53 69 } 54 70 55 71 #else 56 #define EXECSHIFT(size, r es, s1, s2, flags) \72 #define EXECSHIFT(size, rsize, res, s1, s2, flags) \ 57 73 asm ("push %4\n\t"\ 58 74 "popf\n\t"\ 59 stringify(OP) size " %%cl, %" size "5, %"size "0\n\t" \75 stringify(OP) size " %%cl, %" rsize "5, %" rsize "0\n\t" \ 60 76 "pushf\n\t"\ 61 "pop l%1\n\t"\77 "pop %1\n\t"\ 62 78 : "=g" (res), "=g" (flags)\ 63 79 : "c" (s1), "0" (res), "1" (flags), "r" (s2)); 64 80 65 void exec_opl(int s2, int s0, int s1, int iflags) 81 #if defined(__x86_64__) 82 void exec_opq(long s2, long s0, long s1, long iflags) 66 83 { 67 intres, flags;84 long res, flags; 68 85 res = s0; 69 86 flags = iflags; 70 EXECSHIFT(" ", res, s1, s2, flags);87 EXECSHIFT("q", "", res, s1, s2, flags); 71 88 /* overflow is undefined if count != 1 */ 72 89 if (s1 != 1) 73 90 flags &= ~CC_O; 74 printf("%-10s A=%08x B=%08x C=%08x R=%08x CCIN=%04x CC=%04x\n", 91 printf("%-10s A=" FMTLX " B=" FMTLX " C=" FMTLX " R=" FMTLX " CCIN=%04lx CC=%04lx\n", 92 stringify(OP) "q", s0, s2, s1, res, iflags, flags & CC_MASK); 93 } 94 #endif 95 96 void exec_opl(long s2, long s0, long s1, long iflags) 97 { 98 long res, flags; 99 res = s0; 100 flags = iflags; 101 EXECSHIFT("l", "k", res, s1, s2, flags); 102 /* overflow is undefined if count != 1 */ 103 if (s1 != 1) 104 flags &= ~CC_O; 105 printf("%-10s A=" FMTLX " B=" FMTLX " C=" FMTLX " R=" FMTLX " CCIN=%04lx CC=%04lx\n", 75 106 stringify(OP) "l", s0, s2, s1, res, iflags, flags & CC_MASK); 76 107 } 77 108 78 void exec_opw( int s2, int s0, int s1, intiflags)109 void exec_opw(long s2, long s0, long s1, long iflags) 79 110 { 80 intres, flags;111 long res, flags; 81 112 res = s0; 82 113 flags = iflags; 83 EXECSHIFT("w", res, s1, s2, flags);114 EXECSHIFT("w", "w", res, s1, s2, flags); 84 115 /* overflow is undefined if count != 1 */ 85 116 if (s1 != 1) 86 117 flags &= ~CC_O; 87 printf("%-10s A= %08x B=%08x C=%08x R=%08x CCIN=%04x CC=%04x\n",118 printf("%-10s A=" FMTLX " B=" FMTLX " C=" FMTLX " R=" FMTLX " CCIN=%04lx CC=%04lx\n", 88 119 stringify(OP) "w", s0, s2, s1, res, iflags, flags & CC_MASK); 89 120 } … … 92 123 93 124 #ifndef OP_NOBYTE 94 void exec_opb( int s0, int s1, intiflags)125 void exec_opb(long s0, long s1, long iflags) 95 126 { 96 intres, flags;127 long res, flags; 97 128 res = s0; 98 129 flags = iflags; 99 EXECSHIFT("b", res, s1, 0, flags);130 EXECSHIFT("b", "b", res, s1, 0, flags); 100 131 /* overflow is undefined if count != 1 */ 101 132 if (s1 != 1) 102 133 flags &= ~CC_O; 103 printf("%-10s A= %08x B=%08x R=%08x CCIN=%04x CC=%04x\n",134 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CCIN=%04lx CC=%04lx\n", 104 135 stringify(OP) "b", s0, s1, res, iflags, flags & CC_MASK); 105 136 } 106 137 #endif 107 138 108 void exec_op( int s2, int s0, ints1)139 void exec_op(long s2, long s0, long s1) 109 140 { 141 s2 = i2l(s2); 142 s0 = i2l(s0); 143 #if defined(__x86_64__) 144 exec_opq(s2, s0, s1, 0); 145 #endif 110 146 exec_opl(s2, s0, s1, 0); 111 147 #ifdef OP_SHIFTD … … 119 155 #endif 120 156 #ifdef OP_CC 157 #if defined(__x86_64__) 158 exec_opq(s2, s0, s1, CC_C); 159 #endif 121 160 exec_opl(s2, s0, s1, CC_C); 122 161 exec_opw(s2, s0, s1, CC_C); … … 127 166 void glue(test_, OP)(void) 128 167 { 129 int i; 130 for(i = 0; i < 32; i++) 168 int i, n; 169 #if defined(__x86_64__) 170 n = 64; 171 #else 172 n = 32; 173 #endif 174 for(i = 0; i < n; i++) 131 175 exec_op(0x21ad3d34, 0x12345678, i); 132 for(i = 0; i < 32; i++)133 exec_op(0x813f3421, 0x8234567 8, i);176 for(i = 0; i < n; i++) 177 exec_op(0x813f3421, 0x82345679, i); 134 178 } 135 179 -
trunk/src/recompiler/tests/test-i386.c
r1 r2426 29 29 #include <sys/ucontext.h> 30 30 #include <sys/mman.h> 31 #include <asm/vm86.h> 32 31 32 #if !defined(__x86_64__) 33 #define TEST_VM86 34 #define TEST_SEGS 35 #endif 36 //#define LINUX_VM86_IOPL_FIX 37 //#define TEST_P4_FLAGS 38 #if defined(__x86_64__) 39 #define TEST_SSE 40 #define TEST_CMOV 1 41 #define TEST_FCOMI 1 42 #else 43 //#define TEST_SSE 33 44 #define TEST_CMOV 0 34 45 #define TEST_FCOMI 0 35 //#define LINUX_VM86_IOPL_FIX 36 //#define TEST_P4_FLAGS 46 #endif 47 48 #if defined(__x86_64__) 49 #define FMT64X "%016lx" 50 #define FMTLX "%016lx" 51 #define X86_64_ONLY(x) x 52 #else 53 #define FMT64X "%016" PRIx64 54 #define FMTLX "%08lx" 55 #define X86_64_ONLY(x) 56 #endif 57 58 #ifdef TEST_VM86 59 #include <asm/vm86.h> 60 #endif 37 61 38 62 #define xglue(x, y) x ## y … … 48 72 #define CC_O 0x0800 49 73 50 #define __init_call __attribute__ ((unused,__section__ (".initcall.init"))) 51 52 static void *call_start __init_call = NULL; 74 #define __init_call __attribute__ ((unused,__section__ ("initcall"))) 53 75 54 76 #define CC_MASK (CC_C | CC_P | CC_Z | CC_S | CC_O | CC_A) 77 78 #if defined(__x86_64__) 79 static inline long i2l(long v) 80 { 81 return v | ((v ^ 0xabcd) << 32); 82 } 83 #else 84 static inline long i2l(long v) 85 { 86 return v; 87 } 88 #endif 55 89 56 90 #define OP add … … 157 191 158 192 /* lea test (modrm support) */ 159 #define TEST_LEA (STR)\160 {\ 161 asm("lea l" STR ", %0"\193 #define TEST_LEAQ(STR)\ 194 {\ 195 asm("lea " STR ", %0"\ 162 196 : "=r" (res)\ 163 197 : "a" (eax), "b" (ebx), "c" (ecx), "d" (edx), "S" (esi), "D" (edi));\ 164 printf("lea %s = %08x\n", STR, res);\ 198 printf("lea %s = " FMTLX "\n", STR, res);\ 199 } 200 201 #define TEST_LEA(STR)\ 202 {\ 203 asm("lea " STR ", %0"\ 204 : "=r" (res)\ 205 : "a" (eax), "b" (ebx), "c" (ecx), "d" (edx), "S" (esi), "D" (edi));\ 206 printf("lea %s = " FMTLX "\n", STR, res);\ 165 207 } 166 208 … … 170 212 : "=wq" (res)\ 171 213 : "a" (eax), "b" (ebx), "c" (ecx), "d" (edx), "S" (esi), "D" (edi));\ 172 printf("lea %s = %08 x\n", STR, res);\214 printf("lea %s = %08lx\n", STR, res);\ 173 215 } 174 216 … … 176 218 void test_lea(void) 177 219 { 178 inteax, ebx, ecx, edx, esi, edi, res;179 eax = 0x0001;180 ebx = 0x0002;181 ecx = 0x0004;182 edx = 0x0008;183 esi = 0x0010;184 edi = 0x0020;220 long eax, ebx, ecx, edx, esi, edi, res; 221 eax = i2l(0x0001); 222 ebx = i2l(0x0002); 223 ecx = i2l(0x0004); 224 edx = i2l(0x0008); 225 esi = i2l(0x0010); 226 edi = i2l(0x0020); 185 227 186 228 TEST_LEA("0x4000"); … … 238 280 TEST_LEA("0x4000(%%esi, %%ecx, 8)"); 239 281 282 #if defined(__x86_64__) 283 TEST_LEAQ("0x4000"); 284 TEST_LEAQ("0x4000(%%rip)"); 285 286 TEST_LEAQ("(%%rax)"); 287 TEST_LEAQ("(%%rbx)"); 288 TEST_LEAQ("(%%rcx)"); 289 TEST_LEAQ("(%%rdx)"); 290 TEST_LEAQ("(%%rsi)"); 291 TEST_LEAQ("(%%rdi)"); 292 293 TEST_LEAQ("0x40(%%rax)"); 294 TEST_LEAQ("0x40(%%rbx)"); 295 TEST_LEAQ("0x40(%%rcx)"); 296 TEST_LEAQ("0x40(%%rdx)"); 297 TEST_LEAQ("0x40(%%rsi)"); 298 TEST_LEAQ("0x40(%%rdi)"); 299 300 TEST_LEAQ("0x4000(%%rax)"); 301 TEST_LEAQ("0x4000(%%rbx)"); 302 TEST_LEAQ("0x4000(%%rcx)"); 303 TEST_LEAQ("0x4000(%%rdx)"); 304 TEST_LEAQ("0x4000(%%rsi)"); 305 TEST_LEAQ("0x4000(%%rdi)"); 306 307 TEST_LEAQ("(%%rax, %%rcx)"); 308 TEST_LEAQ("(%%rbx, %%rdx)"); 309 TEST_LEAQ("(%%rcx, %%rcx)"); 310 TEST_LEAQ("(%%rdx, %%rcx)"); 311 TEST_LEAQ("(%%rsi, %%rcx)"); 312 TEST_LEAQ("(%%rdi, %%rcx)"); 313 314 TEST_LEAQ("0x40(%%rax, %%rcx)"); 315 TEST_LEAQ("0x4000(%%rbx, %%rdx)"); 316 317 TEST_LEAQ("(%%rcx, %%rcx, 2)"); 318 TEST_LEAQ("(%%rdx, %%rcx, 4)"); 319 TEST_LEAQ("(%%rsi, %%rcx, 8)"); 320 321 TEST_LEAQ("(,%%rax, 2)"); 322 TEST_LEAQ("(,%%rbx, 4)"); 323 TEST_LEAQ("(,%%rcx, 8)"); 324 325 TEST_LEAQ("0x40(,%%rax, 2)"); 326 TEST_LEAQ("0x40(,%%rbx, 4)"); 327 TEST_LEAQ("0x40(,%%rcx, 8)"); 328 329 330 TEST_LEAQ("-10(%%rcx, %%rcx, 2)"); 331 TEST_LEAQ("-10(%%rdx, %%rcx, 4)"); 332 TEST_LEAQ("-10(%%rsi, %%rcx, 8)"); 333 334 TEST_LEAQ("0x4000(%%rcx, %%rcx, 2)"); 335 TEST_LEAQ("0x4000(%%rdx, %%rcx, 4)"); 336 TEST_LEAQ("0x4000(%%rsi, %%rcx, 8)"); 337 #else 240 338 /* limited 16 bit addressing test */ 241 339 TEST_LEA16("0x4000"); … … 254 352 TEST_LEA16("0x4000(%%bx,%%si)"); 255 353 TEST_LEA16("0x4000(%%bx,%%di)"); 354 #endif 256 355 } 257 356 … … 275 374 printf("%-10s %d\n", "set" JCC, res);\ 276 375 if (TEST_CMOV) {\ 277 asm("movl $0x12345678, %0\n\t"\ 278 "cmpl %2, %1\n\t"\ 279 "cmov" JCC "l %3, %0\n\t"\ 376 long val = i2l(1);\ 377 long res = i2l(0x12345678);\ 378 X86_64_ONLY(\ 379 asm("cmpl %2, %1\n\t"\ 380 "cmov" JCC "q %3, %0\n\t"\ 280 381 : "=r" (res)\ 281 : "r" (v1), "r" (v2), "m" (1));\ 282 printf("%-10s R=0x%08x\n", "cmov" JCC "l", res);\ 283 asm("movl $0x12345678, %0\n\t"\ 284 "cmpl %2, %1\n\t"\ 382 : "r" (v1), "r" (v2), "m" (val), "0" (res));\ 383 printf("%-10s R=" FMTLX "\n", "cmov" JCC "q", res);)\ 384 asm("cmpl %2, %1\n\t"\ 385 "cmov" JCC "l %k3, %k0\n\t"\ 386 : "=r" (res)\ 387 : "r" (v1), "r" (v2), "m" (val), "0" (res));\ 388 printf("%-10s R=" FMTLX "\n", "cmov" JCC "l", res);\ 389 asm("cmpl %2, %1\n\t"\ 285 390 "cmov" JCC "w %w3, %w0\n\t"\ 286 391 : "=r" (res)\ 287 : "r" (v1), "r" (v2), "r" (1) );\288 printf("%-10s R= 0x%08x\n", "cmov" JCC "w", res);\392 : "r" (v1), "r" (v2), "r" (1), "0" (res));\ 393 printf("%-10s R=" FMTLX "\n", "cmov" JCC "w", res);\ 289 394 } \ 290 395 } … … 366 471 #include "test-i386-muldiv.h" 367 472 368 void test_imulw2( int op0, intop1)369 { 370 intres, s1, s0, flags;473 void test_imulw2(long op0, long op1) 474 { 475 long res, s1, s0, flags; 371 476 s0 = op0; 372 477 s1 = op1; 373 478 res = s0; 374 479 flags = 0; 375 asm ("push %4\n\t"480 asm volatile ("push %4\n\t" 376 481 "popf\n\t" 377 482 "imulw %w2, %w0\n\t" 378 483 "pushf\n\t" 379 "pop l%1\n\t"484 "pop %1\n\t" 380 485 : "=q" (res), "=g" (flags) 381 486 : "q" (s1), "0" (res), "1" (flags)); 382 printf("%-10s A= %08x B=%08x R=%08x CC=%04x\n",487 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CC=%04lx\n", 383 488 "imulw", s0, s1, res, flags & CC_MASK); 384 489 } 385 490 386 void test_imull2( int op0, intop1)387 { 388 intres, s1, s0, flags;491 void test_imull2(long op0, long op1) 492 { 493 long res, s1, s0, flags; 389 494 s0 = op0; 390 495 s1 = op1; 391 496 res = s0; 392 497 flags = 0; 393 asm ("push %4\n\t"498 asm volatile ("push %4\n\t" 394 499 "popf\n\t" 395 "imull % 2, %0\n\t"500 "imull %k2, %k0\n\t" 396 501 "pushf\n\t" 397 "pop l%1\n\t"502 "pop %1\n\t" 398 503 : "=q" (res), "=g" (flags) 399 504 : "q" (s1), "0" (res), "1" (flags)); 400 printf("%-10s A= %08x B=%08x R=%08x CC=%04x\n",505 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CC=%04lx\n", 401 506 "imull", s0, s1, res, flags & CC_MASK); 402 507 } 403 508 404 #define TEST_IMUL_IM(size, size1, op0, op1)\ 405 {\ 406 int res, flags;\ 509 #if defined(__x86_64__) 510 void test_imulq2(long op0, long op1) 511 { 512 long res, s1, s0, flags; 513 s0 = op0; 514 s1 = op1; 515 res = s0; 516 flags = 0; 517 asm volatile ("push %4\n\t" 518 "popf\n\t" 519 "imulq %2, %0\n\t" 520 "pushf\n\t" 521 "pop %1\n\t" 522 : "=q" (res), "=g" (flags) 523 : "q" (s1), "0" (res), "1" (flags)); 524 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CC=%04lx\n", 525 "imulq", s0, s1, res, flags & CC_MASK); 526 } 527 #endif 528 529 #define TEST_IMUL_IM(size, rsize, op0, op1)\ 530 {\ 531 long res, flags, s1;\ 407 532 flags = 0;\ 408 533 res = 0;\ 409 asm ("push %3\n\t"\ 534 s1 = op1;\ 535 asm volatile ("push %3\n\t"\ 410 536 "popf\n\t"\ 411 "imul" size " $" #op0 ", %" size1 "2, %" size1"0\n\t" \537 "imul" size " $" #op0 ", %" rsize "2, %" rsize "0\n\t" \ 412 538 "pushf\n\t"\ 413 "pop l%1\n\t"\539 "pop %1\n\t"\ 414 540 : "=r" (res), "=g" (flags)\ 415 : "r" ( op1), "1" (flags), "0" (res));\416 printf("%-10s A= %08x B=%08x R=%08x CC=%04x\n",\417 "imul" size , op0,op1, res, flags & CC_MASK);\541 : "r" (s1), "1" (flags), "0" (res));\ 542 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CC=%04lx\n",\ 543 "imul" size " im", (long)op0, (long)op1, res, flags & CC_MASK);\ 418 544 } 419 545 … … 475 601 TEST_IMUL_IM("w", "w", 0x7fff, 0x1000); 476 602 477 TEST_IMUL_IM("l", " ", 45, 0x1234);478 TEST_IMUL_IM("l", " ", -45, 23);479 TEST_IMUL_IM("l", " ", 0x8000, 0x80000000);480 TEST_IMUL_IM("l", " ", 0x7fff, 0x1000);603 TEST_IMUL_IM("l", "k", 45, 0x1234); 604 TEST_IMUL_IM("l", "k", -45, 23); 605 TEST_IMUL_IM("l", "k", 0x8000, 0x80000000); 606 TEST_IMUL_IM("l", "k", 0x7fff, 0x1000); 481 607 482 608 test_idivb(0x12341678, 0x127e); … … 507 633 test_divl(0, 0x80000000, -1); 508 634 test_divl(0x12343, 0x12345678, 0x81234567); 635 636 #if defined(__x86_64__) 637 test_imulq(0, 0x1234001d1234001d, 45); 638 test_imulq(0, 23, -45); 639 test_imulq(0, 0x8000000000000000, 0x8000000000000000); 640 test_imulq(0, 0x100000000, 0x100000000); 641 642 test_mulq(0, 0x1234001d1234001d, 45); 643 test_mulq(0, 23, -45); 644 test_mulq(0, 0x8000000000000000, 0x8000000000000000); 645 test_mulq(0, 0x100000000, 0x100000000); 646 647 test_imulq2(0x1234001d1234001d, 45); 648 test_imulq2(23, -45); 649 test_imulq2(0x8000000000000000, 0x8000000000000000); 650 test_imulq2(0x100000000, 0x100000000); 651 652 TEST_IMUL_IM("q", "", 45, 0x12341234); 653 TEST_IMUL_IM("q", "", -45, 23); 654 TEST_IMUL_IM("q", "", 0x8000, 0x8000000000000000); 655 TEST_IMUL_IM("q", "", 0x7fff, 0x10000000); 656 657 test_idivq(0, 0x12345678abcdef, 12347); 658 test_idivq(0, -233223, -45); 659 test_idivq(0, 0x8000000000000000, -1); 660 test_idivq(0x12343, 0x12345678, 0x81234567); 661 662 test_divq(0, 0x12345678abcdef, 12347); 663 test_divq(0, -233223, -45); 664 test_divq(0, 0x8000000000000000, -1); 665 test_divq(0x12343, 0x12345678, 0x81234567); 666 #endif 509 667 } 510 668 511 669 #define TEST_BSX(op, size, op0)\ 512 670 {\ 513 intres, val, resz;\671 long res, val, resz;\ 514 672 val = op0;\ 515 asm("xor l%1, %1\n"\516 "mov l$0x12345678, %0\n"\673 asm("xor %1, %1\n"\ 674 "mov $0x12345678, %0\n"\ 517 675 #op " %" size "2, %" size "0 ; setz %b1" \ 518 676 : "=r" (res), "=q" (resz)\ 519 677 : "g" (val));\ 520 printf("%-10s A= %08x R=%08x %d\n", #op, val, res, resz);\678 printf("%-10s A=" FMTLX " R=" FMTLX " %ld\n", #op, val, res, resz);\ 521 679 } 522 680 … … 525 683 TEST_BSX(bsrw, "w", 0); 526 684 TEST_BSX(bsrw, "w", 0x12340128); 527 TEST_BSX(bsrl, "", 0);528 TEST_BSX(bsrl, "", 0x00340128);529 685 TEST_BSX(bsfw, "w", 0); 530 686 TEST_BSX(bsfw, "w", 0x12340128); 531 TEST_BSX(bsfl, "", 0); 532 TEST_BSX(bsfl, "", 0x00340128); 687 TEST_BSX(bsrl, "k", 0); 688 TEST_BSX(bsrl, "k", 0x00340128); 689 TEST_BSX(bsfl, "k", 0); 690 TEST_BSX(bsfl, "k", 0x00340128); 691 #if defined(__x86_64__) 692 TEST_BSX(bsrq, "", 0); 693 TEST_BSX(bsrq, "", 0x003401281234); 694 TEST_BSX(bsfq, "", 0); 695 TEST_BSX(bsfq, "", 0x003401281234); 696 #endif 533 697 } 534 698 535 699 /**********************************************/ 700 701 union float64u { 702 double d; 703 uint64_t l; 704 }; 705 706 union float64u q_nan = { .l = 0xFFF8000000000000 }; 707 union float64u s_nan = { .l = 0xFFF0000000000000 }; 536 708 537 709 void test_fops(double a, double b) … … 556 728 } 557 729 730 void fpu_clear_exceptions(void) 731 { 732 struct __attribute__((packed)) { 733 uint16_t fpuc; 734 uint16_t dummy1; 735 uint16_t fpus; 736 uint16_t dummy2; 737 uint16_t fptag; 738 uint16_t dummy3; 739 uint32_t ignored[4]; 740 long double fpregs[8]; 741 } float_env32; 742 743 asm volatile ("fnstenv %0\n" : : "m" (float_env32)); 744 float_env32.fpus &= ~0x7f; 745 asm volatile ("fldenv %0\n" : : "m" (float_env32)); 746 } 747 748 /* XXX: display exception bits when supported */ 749 #define FPUS_EMASK 0x0000 750 //#define FPUS_EMASK 0x007f 751 558 752 void test_fcmp(double a, double b) 559 753 { 560 printf("(%f<%f)=%d\n", 561 a, b, a < b); 562 printf("(%f<=%f)=%d\n", 563 a, b, a <= b); 564 printf("(%f==%f)=%d\n", 565 a, b, a == b); 566 printf("(%f>%f)=%d\n", 567 a, b, a > b); 568 printf("(%f<=%f)=%d\n", 569 a, b, a >= b); 754 long eflags, fpus; 755 756 fpu_clear_exceptions(); 757 asm("fcom %2\n" 758 "fstsw %%ax\n" 759 : "=a" (fpus) 760 : "t" (a), "u" (b)); 761 printf("fcom(%f %f)=%04lx \n", 762 a, b, fpus & (0x4500 | FPUS_EMASK)); 763 fpu_clear_exceptions(); 764 asm("fucom %2\n" 765 "fstsw %%ax\n" 766 : "=a" (fpus) 767 : "t" (a), "u" (b)); 768 printf("fucom(%f %f)=%04lx\n", 769 a, b, fpus & (0x4500 | FPUS_EMASK)); 570 770 if (TEST_FCOMI) { 571 unsigned int eflags;572 771 /* test f(u)comi instruction */ 573 asm("fcomi %2, %1\n" 772 fpu_clear_exceptions(); 773 asm("fcomi %3, %2\n" 774 "fstsw %%ax\n" 574 775 "pushf\n" 575 776 "pop %0\n" 576 : "=r" (eflags) 777 : "=r" (eflags), "=a" (fpus) 577 778 : "t" (a), "u" (b)); 578 printf("fcomi(%f %f)=%08x\n", a, b, eflags & (CC_Z | CC_P | CC_C)); 579 } 779 printf("fcomi(%f %f)=%04lx %02lx\n", 780 a, b, fpus & FPUS_EMASK, eflags & (CC_Z | CC_P | CC_C)); 781 fpu_clear_exceptions(); 782 asm("fucomi %3, %2\n" 783 "fstsw %%ax\n" 784 "pushf\n" 785 "pop %0\n" 786 : "=r" (eflags), "=a" (fpus) 787 : "t" (a), "u" (b)); 788 printf("fucomi(%f %f)=%04lx %02lx\n", 789 a, b, fpus & FPUS_EMASK, eflags & (CC_Z | CC_P | CC_C)); 790 } 791 fpu_clear_exceptions(); 792 asm volatile("fxam\n" 793 "fstsw %%ax\n" 794 : "=a" (fpus) 795 : "t" (a)); 796 printf("fxam(%f)=%04lx\n", a, fpus & 0x4700); 797 fpu_clear_exceptions(); 580 798 } 581 799 … … 595 813 printf("(float)%f = %f\n", a, fa); 596 814 printf("(long double)%f = %Lf\n", a, la); 597 printf("a= %016Lx\n", *(long long*)&a);598 printf("la= %016Lx %04x\n", *(long long*)&la,815 printf("a=" FMT64X "\n", *(uint64_t *)&a); 816 printf("la=" FMT64X " %04x\n", *(uint64_t *)&la, 599 817 *(unsigned short *)((char *)(&la) + 8)); 600 818 … … 610 828 printf("(short)a = %d\n", wa); 611 829 printf("(int)a = %d\n", ia); 612 printf("(int64_t)a = %Ld\n", lla);830 printf("(int64_t)a = " FMT64X "\n", lla); 613 831 printf("rint(a) = %f\n", ra); 614 832 } … … 647 865 for(i=0;i<5;i++)\ 648 866 asm volatile ("fldl %0" : : "m" (dtab[i]));\ 649 asm (save " %0\n" : : "m" (*(env)));\650 asm (restore " %0\n": : "m" (*(env)));\867 asm volatile (save " %0\n" : : "m" (*(env)));\ 868 asm volatile (restore " %0\n": : "m" (*(env)));\ 651 869 for(i=0;i<5;i++)\ 652 870 asm volatile ("fstpl %0" : "=m" (rtab[i]));\ … … 708 926 : "=t" (res)\ 709 927 : "0" (a), "u" (b), "g" (eflags));\ 710 printf("fcmov%s eflags=0x%04 x-> %f\n", \711 CC, eflags, res);\928 printf("fcmov%s eflags=0x%04lx-> %f\n", \ 929 CC, (long)eflags, res);\ 712 930 } 713 931 … … 715 933 { 716 934 double a, b; 717 inteflags, i;935 long eflags, i; 718 936 719 937 a = 1.0; … … 745 963 test_fcmp(2, 2); 746 964 test_fcmp(2, 3); 965 test_fcmp(2, q_nan.d); 966 test_fcmp(q_nan.d, -1); 967 test_fcmp(-1.0/0.0, -1); 968 test_fcmp(1.0/0.0, -1); 747 969 test_fcvt(0.5); 748 970 test_fcvt(-0.5); … … 751 973 test_fcvt(32768); 752 974 test_fcvt(-1e20); 975 test_fcvt(-1.0/0.0); 976 test_fcvt(1.0/0.0); 977 test_fcvt(q_nan.d); 753 978 test_fconst(); 754 979 test_fbcd(1234567890123456); … … 761 986 762 987 /**********************************************/ 988 #if !defined(__x86_64__) 763 989 764 990 #define TEST_BCD(op, op0, cc_in, cc_mask)\ … … 771 997 #op "\n\t"\ 772 998 "pushf\n\t"\ 773 "pop l%1\n\t"\999 "pop %1\n\t"\ 774 1000 : "=a" (res), "=g" (flags)\ 775 1001 : "0" (res), "1" (flags));\ … … 829 1055 TEST_BCD(aad, 0x12340407, CC_A, (CC_C | CC_P | CC_Z | CC_S | CC_O | CC_A)); 830 1056 } 1057 #endif 831 1058 832 1059 #define TEST_XCHG(op, size, opconst)\ 833 1060 {\ 834 intop0, op1;\835 op0 = 0x12345678;\836 op1 = 0xfbca7654;\1061 long op0, op1;\ 1062 op0 = i2l(0x12345678);\ 1063 op1 = i2l(0xfbca7654);\ 837 1064 asm(#op " %" size "0, %" size "1" \ 838 1065 : "=q" (op0), opconst (op1) \ 839 1066 : "0" (op0), "1" (op1));\ 840 printf("%-10s A= %08x B=%08x\n",\1067 printf("%-10s A=" FMTLX " B=" FMTLX "\n",\ 841 1068 #op, op0, op1);\ 842 1069 } … … 844 1071 #define TEST_CMPXCHG(op, size, opconst, eax)\ 845 1072 {\ 846 int op0, op1;\ 847 op0 = 0x12345678;\ 848 op1 = 0xfbca7654;\ 1073 long op0, op1, op2;\ 1074 op0 = i2l(0x12345678);\ 1075 op1 = i2l(0xfbca7654);\ 1076 op2 = i2l(eax);\ 849 1077 asm(#op " %" size "0, %" size "1" \ 850 1078 : "=q" (op0), opconst (op1) \ 851 : "0" (op0), "1" (op1), "a" ( eax));\852 printf("%-10s EAX= %08x A=%08x C=%08x\n",\853 #op, eax, op0, op1);\1079 : "0" (op0), "1" (op1), "a" (op2));\ 1080 printf("%-10s EAX=" FMTLX " A=" FMTLX " C=" FMTLX "\n",\ 1081 #op, op2, op0, op1);\ 854 1082 } 855 1083 856 1084 void test_xchg(void) 857 1085 { 858 TEST_XCHG(xchgl, "", "=q"); 1086 #if defined(__x86_64__) 1087 TEST_XCHG(xchgq, "", "=q"); 1088 #endif 1089 TEST_XCHG(xchgl, "k", "=q"); 859 1090 TEST_XCHG(xchgw, "w", "=q"); 860 1091 TEST_XCHG(xchgb, "b", "=q"); 861 1092 862 TEST_XCHG(xchgl, "", "=m"); 1093 #if defined(__x86_64__) 1094 TEST_XCHG(xchgq, "", "=m"); 1095 #endif 1096 TEST_XCHG(xchgl, "k", "=m"); 863 1097 TEST_XCHG(xchgw, "w", "=m"); 864 1098 TEST_XCHG(xchgb, "b", "=m"); 865 1099 866 TEST_XCHG(xaddl, "", "=q"); 1100 #if defined(__x86_64__) 1101 TEST_XCHG(xaddq, "", "=q"); 1102 #endif 1103 TEST_XCHG(xaddl, "k", "=q"); 867 1104 TEST_XCHG(xaddw, "w", "=q"); 868 1105 TEST_XCHG(xaddb, "b", "=q"); … … 875 1112 } 876 1113 877 TEST_XCHG(xaddl, "", "=m"); 1114 #if defined(__x86_64__) 1115 TEST_XCHG(xaddq, "", "=m"); 1116 #endif 1117 TEST_XCHG(xaddl, "k", "=m"); 878 1118 TEST_XCHG(xaddw, "w", "=m"); 879 1119 TEST_XCHG(xaddb, "b", "=m"); 880 1120 881 TEST_CMPXCHG(cmpxchgl, "", "=q", 0xfbca7654); 1121 #if defined(__x86_64__) 1122 TEST_CMPXCHG(cmpxchgq, "", "=q", 0xfbca7654); 1123 #endif 1124 TEST_CMPXCHG(cmpxchgl, "k", "=q", 0xfbca7654); 882 1125 TEST_CMPXCHG(cmpxchgw, "w", "=q", 0xfbca7654); 883 1126 TEST_CMPXCHG(cmpxchgb, "b", "=q", 0xfbca7654); 884 1127 885 TEST_CMPXCHG(cmpxchgl, "", "=q", 0xfffefdfc); 1128 #if defined(__x86_64__) 1129 TEST_CMPXCHG(cmpxchgq, "", "=q", 0xfffefdfc); 1130 #endif 1131 TEST_CMPXCHG(cmpxchgl, "k", "=q", 0xfffefdfc); 886 1132 TEST_CMPXCHG(cmpxchgw, "w", "=q", 0xfffefdfc); 887 1133 TEST_CMPXCHG(cmpxchgb, "b", "=q", 0xfffefdfc); 888 1134 889 TEST_CMPXCHG(cmpxchgl, "", "=m", 0xfbca7654); 1135 #if defined(__x86_64__) 1136 TEST_CMPXCHG(cmpxchgq, "", "=m", 0xfbca7654); 1137 #endif 1138 TEST_CMPXCHG(cmpxchgl, "k", "=m", 0xfbca7654); 890 1139 TEST_CMPXCHG(cmpxchgw, "w", "=m", 0xfbca7654); 891 1140 TEST_CMPXCHG(cmpxchgb, "b", "=m", 0xfbca7654); 892 1141 893 TEST_CMPXCHG(cmpxchgl, "", "=m", 0xfffefdfc); 1142 #if defined(__x86_64__) 1143 TEST_CMPXCHG(cmpxchgq, "", "=m", 0xfffefdfc); 1144 #endif 1145 TEST_CMPXCHG(cmpxchgl, "k", "=m", 0xfffefdfc); 894 1146 TEST_CMPXCHG(cmpxchgw, "w", "=m", 0xfffefdfc); 895 1147 TEST_CMPXCHG(cmpxchgb, "b", "=m", 0xfffefdfc); … … 897 1149 { 898 1150 uint64_t op0, op1, op2; 899 inti, eflags;1151 long i, eflags; 900 1152 901 1153 for(i = 0; i < 2; i++) { … … 908 1160 asm("cmpxchg8b %1\n" 909 1161 "pushf\n" 910 "pop l%2\n"1162 "pop %2\n" 911 1163 : "=A" (op0), "=m" (op1), "=g" (eflags) 912 1164 : "0" (op0), "m" (op1), "b" ((int)op2), "c" ((int)(op2 >> 32))); 913 printf("cmpxchg8b: op0= %016llx op1=%016llx CC=%02x\n",1165 printf("cmpxchg8b: op0=" FMT64X " op1=" FMT64X " CC=%02lx\n", 914 1166 op0, op1, eflags & CC_Z); 915 1167 } … … 917 1169 } 918 1170 1171 #ifdef TEST_SEGS 919 1172 /**********************************************/ 920 1173 /* segmentation tests */ … … 930 1183 #endif 931 1184 1185 #define MK_SEL(n) (((n) << 3) | 7) 1186 932 1187 uint8_t seg_data1[4096]; 933 1188 uint8_t seg_data2[4096]; 934 935 #define MK_SEL(n) (((n) << 3) | 7)936 1189 937 1190 #define TEST_LR(op, size, seg, mask)\ … … 1078 1331 printf("func3() = 0x%08x\n", res); 1079 1332 } 1080 1081 extern char func_lret32; 1082 extern char func_iret32; 1333 #endif 1334 1335 #if defined(__x86_64__) 1336 asm(".globl func_lret\n" 1337 "func_lret:\n" 1338 "movl $0x87654641, %eax\n" 1339 "lretq\n"); 1340 #else 1341 asm(".globl func_lret\n" 1342 "func_lret:\n" 1343 "movl $0x87654321, %eax\n" 1344 "lret\n" 1345 1346 ".globl func_iret\n" 1347 "func_iret:\n" 1348 "movl $0xabcd4321, %eax\n" 1349 "iret\n"); 1350 #endif 1351 1352 extern char func_lret; 1353 extern char func_iret; 1083 1354 1084 1355 void test_misc(void) 1085 1356 { 1086 1357 char table[256]; 1087 intres, i;1358 long res, i; 1088 1359 1089 1360 for(i=0;i<256;i++) table[i] = 256 - i; 1090 1361 res = 0x12345678; 1091 1362 asm ("xlat" : "=a" (res) : "b" (table), "0" (res)); 1092 printf("xlat: EAX=%08x\n", res); 1093 1094 asm volatile ("pushl %%cs ; call %1" 1363 printf("xlat: EAX=" FMTLX "\n", res); 1364 1365 #if defined(__x86_64__) 1366 { 1367 static struct __attribute__((packed)) { 1368 uint32_t offset; 1369 uint16_t seg; 1370 } desc; 1371 long cs_sel; 1372 1373 asm volatile ("mov %%cs, %0" : "=r" (cs_sel)); 1374 1375 asm volatile ("push %1\n" 1376 "call func_lret\n" 1377 : "=a" (res) 1378 : "r" (cs_sel) : "memory", "cc"); 1379 printf("func_lret=" FMTLX "\n", res); 1380 1381 /* NOTE: we assume that &func_lret < 4GB */ 1382 desc.offset = (long)&func_lret; 1383 desc.seg = cs_sel; 1384 1385 asm volatile ("xor %%rax, %%rax\n" 1386 "rex64 lcall %1\n" 1387 : "=a" (res) 1388 : "m" (desc) 1389 : "memory", "cc"); 1390 printf("func_lret2=" FMTLX "\n", res); 1391 1392 asm volatile ("push %2\n" 1393 "mov $ 1f, %%rax\n" 1394 "push %%rax\n" 1395 "ljmp %1\n" 1396 "1:\n" 1397 : "=a" (res) 1398 : "m" (desc), "b" (cs_sel) 1399 : "memory", "cc"); 1400 printf("func_lret3=" FMTLX "\n", res); 1401 } 1402 #else 1403 asm volatile ("push %%cs ; call %1" 1095 1404 : "=a" (res) 1096 : "m" (func_lret 32): "memory", "cc");1097 printf("func_lret 32=%x\n", res);1098 1099 asm volatile ("pushf l ; pushl%%cs ; call %1"1405 : "m" (func_lret): "memory", "cc"); 1406 printf("func_lret=" FMTLX "\n", res); 1407 1408 asm volatile ("pushf ; push %%cs ; call %1" 1100 1409 : "=a" (res) 1101 : "m" (func_iret32): "memory", "cc"); 1102 printf("func_iret32=%x\n", res); 1103 1410 : "m" (func_iret): "memory", "cc"); 1411 printf("func_iret=" FMTLX "\n", res); 1412 #endif 1413 1414 #if defined(__x86_64__) 1415 /* specific popl test */ 1416 asm volatile ("push $12345432 ; push $0x9abcdef ; pop (%%rsp) ; pop %0" 1417 : "=g" (res)); 1418 printf("popl esp=" FMTLX "\n", res); 1419 #else 1104 1420 /* specific popl test */ 1105 1421 asm volatile ("pushl $12345432 ; pushl $0x9abcdef ; popl (%%esp) ; popl %0" 1106 1422 : "=g" (res)); 1107 printf("popl esp= %x\n", res);1423 printf("popl esp=" FMTLX "\n", res); 1108 1424 1109 1425 /* specific popw test */ 1110 1426 asm volatile ("pushl $12345432 ; pushl $0x9abcdef ; popw (%%esp) ; addl $2, %%esp ; popl %0" 1111 1427 : "=g" (res)); 1112 printf("popw esp=%x\n", res); 1428 printf("popw esp=" FMTLX "\n", res); 1429 #endif 1113 1430 } 1114 1431 … … 1117 1434 #define TEST_STRING1(OP, size, DF, REP)\ 1118 1435 {\ 1119 intesi, edi, eax, ecx, eflags;\1436 long esi, edi, eax, ecx, eflags;\ 1120 1437 \ 1121 1438 esi = (long)(str_buffer + sizeof(str_buffer) / 2);\ 1122 1439 edi = (long)(str_buffer + sizeof(str_buffer) / 2) + 16;\ 1123 eax = 0x12345678;\1440 eax = i2l(0x12345678);\ 1124 1441 ecx = 17;\ 1125 1442 \ 1126 asm volatile ("push l$0\n\t"\1443 asm volatile ("push $0\n\t"\ 1127 1444 "popf\n\t"\ 1128 1445 DF "\n\t"\ … … 1130 1447 "cld\n\t"\ 1131 1448 "pushf\n\t"\ 1132 "pop l%4\n\t"\1449 "pop %4\n\t"\ 1133 1450 : "=S" (esi), "=D" (edi), "=a" (eax), "=c" (ecx), "=g" (eflags)\ 1134 1451 : "0" (esi), "1" (edi), "2" (eax), "3" (ecx));\ 1135 printf("%-10s ESI= %08x EDI=%08x EAX=%08x ECX=%08xEFL=%04x\n",\1452 printf("%-10s ESI=" FMTLX " EDI=" FMTLX " EAX=" FMTLX " ECX=" FMTLX " EFL=%04x\n",\ 1136 1453 REP #OP size, esi, edi, eax, ecx,\ 1137 eflags & (CC_C | CC_P | CC_Z | CC_S | CC_O | CC_A));\1454 (int)(eflags & (CC_C | CC_P | CC_Z | CC_S | CC_O | CC_A)));\ 1138 1455 } 1139 1456 … … 1142 1459 TEST_STRING1(OP, "w", "", REP);\ 1143 1460 TEST_STRING1(OP, "l", "", REP);\ 1461 X86_64_ONLY(TEST_STRING1(OP, "q", "", REP));\ 1144 1462 TEST_STRING1(OP, "b", "std", REP);\ 1145 1463 TEST_STRING1(OP, "w", "std", REP);\ 1146 TEST_STRING1(OP, "l", "std", REP) 1464 TEST_STRING1(OP, "l", "std", REP);\ 1465 X86_64_ONLY(TEST_STRING1(OP, "q", "std", REP)) 1147 1466 1148 1467 void test_string(void) … … 1168 1487 } 1169 1488 1489 #ifdef TEST_VM86 1170 1490 /* VM86 test */ 1171 1491 … … 1298 1618 munmap(vm86_mem, 0x110000); 1299 1619 } 1620 #endif 1300 1621 1301 1622 /* exception tests */ 1302 #if ndef REG_EAX1623 #if defined(__i386__) && !defined(REG_EAX) 1303 1624 #define REG_EAX EAX 1304 1625 #define REG_EBX EBX … … 1315 1636 #endif 1316 1637 1638 #if defined(__x86_64__) 1639 #define REG_EIP REG_RIP 1640 #endif 1641 1317 1642 jmp_buf jmp_env; 1318 1643 int v1; … … 1329 1654 printf("\n"); 1330 1655 1331 printf("trapno= 0x%02x err=0x%08x",1332 uc->uc_mcontext.gregs[REG_TRAPNO],1333 uc->uc_mcontext.gregs[REG_ERR]);1334 printf(" EIP= 0x%08x",uc->uc_mcontext.gregs[REG_EIP]);1656 printf("trapno=" FMTLX " err=" FMTLX, 1657 (long)uc->uc_mcontext.gregs[REG_TRAPNO], 1658 (long)uc->uc_mcontext.gregs[REG_ERR]); 1659 printf(" EIP=" FMTLX, (long)uc->uc_mcontext.gregs[REG_EIP]); 1335 1660 printf("\n"); 1336 1661 longjmp(jmp_env, 1); … … 1339 1664 void test_exceptions(void) 1340 1665 { 1341 struct modify_ldt_ldt_s ldt;1342 1666 struct sigaction act; 1343 1667 volatile int val; … … 1345 1669 act.sa_sigaction = sig_handler; 1346 1670 sigemptyset(&act.sa_mask); 1347 act.sa_flags = SA_SIGINFO ;1671 act.sa_flags = SA_SIGINFO | SA_NODEFER; 1348 1672 sigaction(SIGFPE, &act, NULL); 1349 1673 sigaction(SIGILL, &act, NULL); … … 1360 1684 } 1361 1685 1686 #if !defined(__x86_64__) 1362 1687 printf("BOUND exception:\n"); 1363 1688 if (setjmp(jmp_env) == 0) { … … 1367 1692 asm volatile ("bound %0, %1" : : "r" (11), "m" (tab[0])); 1368 1693 } 1369 1694 #endif 1695 1696 #ifdef TEST_SEGS 1370 1697 printf("segment exceptions:\n"); 1371 1698 if (setjmp(jmp_env) == 0) { … … 1380 1707 } 1381 1708 1382 ldt.entry_number = 1; 1383 ldt.base_addr = (unsigned long)&seg_data1; 1384 ldt.limit = (sizeof(seg_data1) + 0xfff) >> 12; 1385 ldt.seg_32bit = 1; 1386 ldt.contents = MODIFY_LDT_CONTENTS_DATA; 1387 ldt.read_exec_only = 0; 1388 ldt.limit_in_pages = 1; 1389 ldt.seg_not_present = 1; 1390 ldt.useable = 1; 1391 modify_ldt(1, &ldt, sizeof(ldt)); /* write ldt entry */ 1392 1393 if (setjmp(jmp_env) == 0) { 1394 /* segment not present */ 1395 asm volatile ("movl %0, %%fs" : : "r" (MK_SEL(1))); 1396 } 1709 { 1710 struct modify_ldt_ldt_s ldt; 1711 ldt.entry_number = 1; 1712 ldt.base_addr = (unsigned long)&seg_data1; 1713 ldt.limit = (sizeof(seg_data1) + 0xfff) >> 12; 1714 ldt.seg_32bit = 1; 1715 ldt.contents = MODIFY_LDT_CONTENTS_DATA; 1716 ldt.read_exec_only = 0; 1717 ldt.limit_in_pages = 1; 1718 ldt.seg_not_present = 1; 1719 ldt.useable = 1; 1720 modify_ldt(1, &ldt, sizeof(ldt)); /* write ldt entry */ 1721 1722 if (setjmp(jmp_env) == 0) { 1723 /* segment not present */ 1724 asm volatile ("movl %0, %%fs" : : "r" (MK_SEL(1))); 1725 } 1726 } 1727 #endif 1397 1728 1398 1729 /* test SEGV reporting */ … … 1458 1789 } 1459 1790 1791 #if !defined(__x86_64__) 1460 1792 printf("INTO exception:\n"); 1461 1793 if (setjmp(jmp_env) == 0) { … … 1463 1795 asm volatile ("addl $1, %0 ; into" : : "r" (0x7fffffff)); 1464 1796 } 1797 #endif 1465 1798 1466 1799 printf("OUTB exception:\n"); … … 1501 1834 } 1502 1835 1836 #if !defined(__x86_64__) 1503 1837 /* specific precise single step test */ 1504 1838 void sig_trap_handler(int sig, siginfo_t *info, void *puc) 1505 1839 { 1506 1840 struct ucontext *uc = puc; 1507 printf("EIP= 0x%08x\n",uc->uc_mcontext.gregs[REG_EIP]);1841 printf("EIP=" FMTLX "\n", (long)uc->uc_mcontext.gregs[REG_EIP]); 1508 1842 } 1509 1843 … … 1626 1960 } 1627 1961 } 1962 #endif 1963 1964 long enter_stack[4096]; 1965 1966 #if defined(__x86_64__) 1967 #define RSP "%%rsp" 1968 #define RBP "%%rbp" 1969 #else 1970 #define RSP "%%esp" 1971 #define RBP "%%ebp" 1972 #endif 1973 1974 #define TEST_ENTER(size, stack_type, level)\ 1975 {\ 1976 long esp_save, esp_val, ebp_val, ebp_save, i;\ 1977 stack_type *ptr, *stack_end, *stack_ptr;\ 1978 memset(enter_stack, 0, sizeof(enter_stack));\ 1979 stack_end = stack_ptr = (stack_type *)(enter_stack + 4096);\ 1980 ebp_val = (long)stack_ptr;\ 1981 for(i=1;i<=32;i++)\ 1982 *--stack_ptr = i;\ 1983 esp_val = (long)stack_ptr;\ 1984 asm("mov " RSP ", %[esp_save]\n"\ 1985 "mov " RBP ", %[ebp_save]\n"\ 1986 "mov %[esp_val], " RSP "\n"\ 1987 "mov %[ebp_val], " RBP "\n"\ 1988 "enter" size " $8, $" #level "\n"\ 1989 "mov " RSP ", %[esp_val]\n"\ 1990 "mov " RBP ", %[ebp_val]\n"\ 1991 "mov %[esp_save], " RSP "\n"\ 1992 "mov %[ebp_save], " RBP "\n"\ 1993 : [esp_save] "=r" (esp_save),\ 1994 [ebp_save] "=r" (ebp_save),\ 1995 [esp_val] "=r" (esp_val),\ 1996 [ebp_val] "=r" (ebp_val)\ 1997 : "[esp_val]" (esp_val),\ 1998 "[ebp_val]" (ebp_val));\ 1999 printf("level=%d:\n", level);\ 2000 printf("esp_val=" FMTLX "\n", esp_val - (long)stack_end);\ 2001 printf("ebp_val=" FMTLX "\n", ebp_val - (long)stack_end);\ 2002 for(ptr = (stack_type *)esp_val; ptr < stack_end; ptr++)\ 2003 printf(FMTLX "\n", (long)ptr[0]);\ 2004 } 2005 2006 static void test_enter(void) 2007 { 2008 #if defined(__x86_64__) 2009 TEST_ENTER("q", uint64_t, 0); 2010 TEST_ENTER("q", uint64_t, 1); 2011 TEST_ENTER("q", uint64_t, 2); 2012 TEST_ENTER("q", uint64_t, 31); 2013 #else 2014 TEST_ENTER("l", uint32_t, 0); 2015 TEST_ENTER("l", uint32_t, 1); 2016 TEST_ENTER("l", uint32_t, 2); 2017 TEST_ENTER("l", uint32_t, 31); 2018 #endif 2019 2020 TEST_ENTER("w", uint16_t, 0); 2021 TEST_ENTER("w", uint16_t, 1); 2022 TEST_ENTER("w", uint16_t, 2); 2023 TEST_ENTER("w", uint16_t, 31); 2024 } 2025 2026 #ifdef TEST_SSE 2027 2028 typedef int __m64 __attribute__ ((__mode__ (__V2SI__))); 2029 typedef int __m128 __attribute__ ((__mode__(__V4SF__))); 2030 2031 typedef union { 2032 double d[2]; 2033 float s[4]; 2034 uint32_t l[4]; 2035 uint64_t q[2]; 2036 __m128 dq; 2037 } XMMReg; 2038 2039 static uint64_t __attribute__((aligned(16))) test_values[4][2] = { 2040 { 0x456723c698694873, 0xdc515cff944a58ec }, 2041 { 0x1f297ccd58bad7ab, 0x41f21efba9e3e146 }, 2042 { 0x007c62c2085427f8, 0x231be9e8cde7438d }, 2043 { 0x0f76255a085427f8, 0xc233e9e8c4c9439a }, 2044 }; 2045 2046 #define SSE_OP(op)\ 2047 {\ 2048 asm volatile (#op " %2, %0" : "=x" (r.dq) : "0" (a.dq), "x" (b.dq));\ 2049 printf("%-9s: a=" FMT64X "" FMT64X " b=" FMT64X "" FMT64X " r=" FMT64X "" FMT64X "\n",\ 2050 #op,\ 2051 a.q[1], a.q[0],\ 2052 b.q[1], b.q[0],\ 2053 r.q[1], r.q[0]);\ 2054 } 2055 2056 #define SSE_OP2(op)\ 2057 {\ 2058 int i;\ 2059 for(i=0;i<2;i++) {\ 2060 a.q[0] = test_values[2*i][0];\ 2061 a.q[1] = test_values[2*i][1];\ 2062 b.q[0] = test_values[2*i+1][0];\ 2063 b.q[1] = test_values[2*i+1][1];\ 2064 SSE_OP(op);\ 2065 }\ 2066 } 2067 2068 #define MMX_OP2(op)\ 2069 {\ 2070 int i;\ 2071 for(i=0;i<2;i++) {\ 2072 a.q[0] = test_values[2*i][0];\ 2073 b.q[0] = test_values[2*i+1][0];\ 2074 asm volatile (#op " %2, %0" : "=y" (r.q[0]) : "0" (a.q[0]), "y" (b.q[0]));\ 2075 printf("%-9s: a=" FMT64X " b=" FMT64X " r=" FMT64X "\n",\ 2076 #op,\ 2077 a.q[0],\ 2078 b.q[0],\ 2079 r.q[0]);\ 2080 }\ 2081 SSE_OP2(op);\ 2082 } 2083 2084 #define SHUF_OP(op, ib)\ 2085 {\ 2086 a.q[0] = test_values[0][0];\ 2087 a.q[1] = test_values[0][1];\ 2088 b.q[0] = test_values[1][0];\ 2089 b.q[1] = test_values[1][1];\ 2090 asm volatile (#op " $" #ib ", %2, %0" : "=x" (r.dq) : "0" (a.dq), "x" (b.dq));\ 2091 printf("%-9s: a=" FMT64X "" FMT64X " b=" FMT64X "" FMT64X " ib=%02x r=" FMT64X "" FMT64X "\n",\ 2092 #op,\ 2093 a.q[1], a.q[0],\ 2094 b.q[1], b.q[0],\ 2095 ib,\ 2096 r.q[1], r.q[0]);\ 2097 } 2098 2099 #define PSHUF_OP(op, ib)\ 2100 {\ 2101 int i;\ 2102 for(i=0;i<2;i++) {\ 2103 a.q[0] = test_values[2*i][0];\ 2104 a.q[1] = test_values[2*i][1];\ 2105 asm volatile (#op " $" #ib ", %1, %0" : "=x" (r.dq) : "x" (a.dq));\ 2106 printf("%-9s: a=" FMT64X "" FMT64X " ib=%02x r=" FMT64X "" FMT64X "\n",\ 2107 #op,\ 2108 a.q[1], a.q[0],\ 2109 ib,\ 2110 r.q[1], r.q[0]);\ 2111 }\ 2112 } 2113 2114 #define SHIFT_IM(op, ib)\ 2115 {\ 2116 int i;\ 2117 for(i=0;i<2;i++) {\ 2118 a.q[0] = test_values[2*i][0];\ 2119 a.q[1] = test_values[2*i][1];\ 2120 asm volatile (#op " $" #ib ", %0" : "=x" (r.dq) : "0" (a.dq));\ 2121 printf("%-9s: a=" FMT64X "" FMT64X " ib=%02x r=" FMT64X "" FMT64X "\n",\ 2122 #op,\ 2123 a.q[1], a.q[0],\ 2124 ib,\ 2125 r.q[1], r.q[0]);\ 2126 }\ 2127 } 2128 2129 #define SHIFT_OP(op, ib)\ 2130 {\ 2131 int i;\ 2132 SHIFT_IM(op, ib);\ 2133 for(i=0;i<2;i++) {\ 2134 a.q[0] = test_values[2*i][0];\ 2135 a.q[1] = test_values[2*i][1];\ 2136 b.q[0] = ib;\ 2137 b.q[1] = 0;\ 2138 asm volatile (#op " %2, %0" : "=x" (r.dq) : "0" (a.dq), "x" (b.dq));\ 2139 printf("%-9s: a=" FMT64X "" FMT64X " b=" FMT64X "" FMT64X " r=" FMT64X "" FMT64X "\n",\ 2140 #op,\ 2141 a.q[1], a.q[0],\ 2142 b.q[1], b.q[0],\ 2143 r.q[1], r.q[0]);\ 2144 }\ 2145 } 2146 2147 #define MOVMSK(op)\ 2148 {\ 2149 int i, reg;\ 2150 for(i=0;i<2;i++) {\ 2151 a.q[0] = test_values[2*i][0];\ 2152 a.q[1] = test_values[2*i][1];\ 2153 asm volatile (#op " %1, %0" : "=r" (reg) : "x" (a.dq));\ 2154 printf("%-9s: a=" FMT64X "" FMT64X " r=%08x\n",\ 2155 #op,\ 2156 a.q[1], a.q[0],\ 2157 reg);\ 2158 }\ 2159 } 2160 2161 #define SSE_OPS(a) \ 2162 SSE_OP(a ## ps);\ 2163 SSE_OP(a ## ss); 2164 2165 #define SSE_OPD(a) \ 2166 SSE_OP(a ## pd);\ 2167 SSE_OP(a ## sd); 2168 2169 #define SSE_COMI(op, field)\ 2170 {\ 2171 unsigned int eflags;\ 2172 XMMReg a, b;\ 2173 a.field[0] = a1;\ 2174 b.field[0] = b1;\ 2175 asm volatile (#op " %2, %1\n"\ 2176 "pushf\n"\ 2177 "pop %0\n"\ 2178 : "=m" (eflags)\ 2179 : "x" (a.dq), "x" (b.dq));\ 2180 printf("%-9s: a=%f b=%f cc=%04x\n",\ 2181 #op, a1, b1,\ 2182 eflags & (CC_C | CC_P | CC_Z | CC_S | CC_O | CC_A));\ 2183 } 2184 2185 void test_sse_comi(double a1, double b1) 2186 { 2187 SSE_COMI(ucomiss, s); 2188 SSE_COMI(ucomisd, d); 2189 SSE_COMI(comiss, s); 2190 SSE_COMI(comisd, d); 2191 } 2192 2193 #define CVT_OP_XMM(op)\ 2194 {\ 2195 asm volatile (#op " %1, %0" : "=x" (r.dq) : "x" (a.dq));\ 2196 printf("%-9s: a=" FMT64X "" FMT64X " r=" FMT64X "" FMT64X "\n",\ 2197 #op,\ 2198 a.q[1], a.q[0],\ 2199 r.q[1], r.q[0]);\ 2200 } 2201 2202 /* Force %xmm0 usage to avoid the case where both register index are 0 2203 to test intruction decoding more extensively */ 2204 #define CVT_OP_XMM2MMX(op)\ 2205 {\ 2206 asm volatile (#op " %1, %0" : "=y" (r.q[0]) : "x" (a.dq) \ 2207 : "%xmm0");\ 2208 printf("%-9s: a=" FMT64X "" FMT64X " r=" FMT64X "\n",\ 2209 #op,\ 2210 a.q[1], a.q[0],\ 2211 r.q[0]);\ 2212 } 2213 2214 #define CVT_OP_MMX2XMM(op)\ 2215 {\ 2216 asm volatile (#op " %1, %0" : "=x" (r.dq) : "y" (a.q[0]));\ 2217 printf("%-9s: a=" FMT64X " r=" FMT64X "" FMT64X "\n",\ 2218 #op,\ 2219 a.q[0],\ 2220 r.q[1], r.q[0]);\ 2221 } 2222 2223 #define CVT_OP_REG2XMM(op)\ 2224 {\ 2225 asm volatile (#op " %1, %0" : "=x" (r.dq) : "r" (a.l[0]));\ 2226 printf("%-9s: a=%08x r=" FMT64X "" FMT64X "\n",\ 2227 #op,\ 2228 a.l[0],\ 2229 r.q[1], r.q[0]);\ 2230 } 2231 2232 #define CVT_OP_XMM2REG(op)\ 2233 {\ 2234 asm volatile (#op " %1, %0" : "=r" (r.l[0]) : "x" (a.dq));\ 2235 printf("%-9s: a=" FMT64X "" FMT64X " r=%08x\n",\ 2236 #op,\ 2237 a.q[1], a.q[0],\ 2238 r.l[0]);\ 2239 } 2240 2241 struct fpxstate { 2242 uint16_t fpuc; 2243 uint16_t fpus; 2244 uint16_t fptag; 2245 uint16_t fop; 2246 uint32_t fpuip; 2247 uint16_t cs_sel; 2248 uint16_t dummy0; 2249 uint32_t fpudp; 2250 uint16_t ds_sel; 2251 uint16_t dummy1; 2252 uint32_t mxcsr; 2253 uint32_t mxcsr_mask; 2254 uint8_t fpregs1[8 * 16]; 2255 uint8_t xmm_regs[8 * 16]; 2256 uint8_t dummy2[224]; 2257 }; 2258 2259 static struct fpxstate fpx_state __attribute__((aligned(16))); 2260 static struct fpxstate fpx_state2 __attribute__((aligned(16))); 2261 2262 void test_fxsave(void) 2263 { 2264 struct fpxstate *fp = &fpx_state; 2265 struct fpxstate *fp2 = &fpx_state2; 2266 int i, nb_xmm; 2267 XMMReg a, b; 2268 a.q[0] = test_values[0][0]; 2269 a.q[1] = test_values[0][1]; 2270 b.q[0] = test_values[1][0]; 2271 b.q[1] = test_values[1][1]; 2272 2273 asm("movdqa %2, %%xmm0\n" 2274 "movdqa %3, %%xmm7\n" 2275 #if defined(__x86_64__) 2276 "movdqa %2, %%xmm15\n" 2277 #endif 2278 " fld1\n" 2279 " fldpi\n" 2280 " fldln2\n" 2281 " fxsave %0\n" 2282 " fxrstor %0\n" 2283 " fxsave %1\n" 2284 " fninit\n" 2285 : "=m" (*(uint32_t *)fp2), "=m" (*(uint32_t *)fp) 2286 : "m" (a), "m" (b)); 2287 printf("fpuc=%04x\n", fp->fpuc); 2288 printf("fpus=%04x\n", fp->fpus); 2289 printf("fptag=%04x\n", fp->fptag); 2290 for(i = 0; i < 3; i++) { 2291 printf("ST%d: " FMT64X " %04x\n", 2292 i, 2293 *(uint64_t *)&fp->fpregs1[i * 16], 2294 *(uint16_t *)&fp->fpregs1[i * 16 + 8]); 2295 } 2296 printf("mxcsr=%08x\n", fp->mxcsr & 0x1f80); 2297 #if defined(__x86_64__) 2298 nb_xmm = 16; 2299 #else 2300 nb_xmm = 8; 2301 #endif 2302 for(i = 0; i < nb_xmm; i++) { 2303 printf("xmm%d: " FMT64X "" FMT64X "\n", 2304 i, 2305 *(uint64_t *)&fp->xmm_regs[i * 16], 2306 *(uint64_t *)&fp->xmm_regs[i * 16 + 8]); 2307 } 2308 } 2309 2310 void test_sse(void) 2311 { 2312 XMMReg r, a, b; 2313 int i; 2314 2315 MMX_OP2(punpcklbw); 2316 MMX_OP2(punpcklwd); 2317 MMX_OP2(punpckldq); 2318 MMX_OP2(packsswb); 2319 MMX_OP2(pcmpgtb); 2320 MMX_OP2(pcmpgtw); 2321 MMX_OP2(pcmpgtd); 2322 MMX_OP2(packuswb); 2323 MMX_OP2(punpckhbw); 2324 MMX_OP2(punpckhwd); 2325 MMX_OP2(punpckhdq); 2326 MMX_OP2(packssdw); 2327 MMX_OP2(pcmpeqb); 2328 MMX_OP2(pcmpeqw); 2329 MMX_OP2(pcmpeqd); 2330 2331 MMX_OP2(paddq); 2332 MMX_OP2(pmullw); 2333 MMX_OP2(psubusb); 2334 MMX_OP2(psubusw); 2335 MMX_OP2(pminub); 2336 MMX_OP2(pand); 2337 MMX_OP2(paddusb); 2338 MMX_OP2(paddusw); 2339 MMX_OP2(pmaxub); 2340 MMX_OP2(pandn); 2341 2342 MMX_OP2(pmulhuw); 2343 MMX_OP2(pmulhw); 1628 2344 1629 static void *call_end __init_call = NULL; 2345 MMX_OP2(psubsb); 2346 MMX_OP2(psubsw); 2347 MMX_OP2(pminsw); 2348 MMX_OP2(por); 2349 MMX_OP2(paddsb); 2350 MMX_OP2(paddsw); 2351 MMX_OP2(pmaxsw); 2352 MMX_OP2(pxor); 2353 MMX_OP2(pmuludq); 2354 MMX_OP2(pmaddwd); 2355 MMX_OP2(psadbw); 2356 MMX_OP2(psubb); 2357 MMX_OP2(psubw); 2358 MMX_OP2(psubd); 2359 MMX_OP2(psubq); 2360 MMX_OP2(paddb); 2361 MMX_OP2(paddw); 2362 MMX_OP2(paddd); 2363 2364 MMX_OP2(pavgb); 2365 MMX_OP2(pavgw); 2366 2367 asm volatile ("pinsrw $1, %1, %0" : "=y" (r.q[0]) : "r" (0x12345678)); 2368 printf("%-9s: r=" FMT64X "\n", "pinsrw", r.q[0]); 2369 2370 asm volatile ("pinsrw $5, %1, %0" : "=x" (r.dq) : "r" (0x12345678)); 2371 printf("%-9s: r=" FMT64X "" FMT64X "\n", "pinsrw", r.q[1], r.q[0]); 2372 2373 a.q[0] = test_values[0][0]; 2374 a.q[1] = test_values[0][1]; 2375 asm volatile ("pextrw $1, %1, %0" : "=r" (r.l[0]) : "y" (a.q[0])); 2376 printf("%-9s: r=%08x\n", "pextrw", r.l[0]); 2377 2378 asm volatile ("pextrw $5, %1, %0" : "=r" (r.l[0]) : "x" (a.dq)); 2379 printf("%-9s: r=%08x\n", "pextrw", r.l[0]); 2380 2381 asm volatile ("pmovmskb %1, %0" : "=r" (r.l[0]) : "y" (a.q[0])); 2382 printf("%-9s: r=%08x\n", "pmovmskb", r.l[0]); 2383 2384 asm volatile ("pmovmskb %1, %0" : "=r" (r.l[0]) : "x" (a.dq)); 2385 printf("%-9s: r=%08x\n", "pmovmskb", r.l[0]); 2386 2387 { 2388 r.q[0] = -1; 2389 r.q[1] = -1; 2390 2391 a.q[0] = test_values[0][0]; 2392 a.q[1] = test_values[0][1]; 2393 b.q[0] = test_values[1][0]; 2394 b.q[1] = test_values[1][1]; 2395 asm volatile("maskmovq %1, %0" : 2396 : "y" (a.q[0]), "y" (b.q[0]), "D" (&r) 2397 : "memory"); 2398 printf("%-9s: r=" FMT64X " a=" FMT64X " b=" FMT64X "\n", 2399 "maskmov", 2400 r.q[0], 2401 a.q[0], 2402 b.q[0]); 2403 asm volatile("maskmovdqu %1, %0" : 2404 : "x" (a.dq), "x" (b.dq), "D" (&r) 2405 : "memory"); 2406 printf("%-9s: r=" FMT64X "" FMT64X " a=" FMT64X "" FMT64X " b=" FMT64X "" FMT64X "\n", 2407 "maskmov", 2408 r.q[1], r.q[0], 2409 a.q[1], a.q[0], 2410 b.q[1], b.q[0]); 2411 } 2412 2413 asm volatile ("emms"); 2414 2415 SSE_OP2(punpcklqdq); 2416 SSE_OP2(punpckhqdq); 2417 SSE_OP2(andps); 2418 SSE_OP2(andpd); 2419 SSE_OP2(andnps); 2420 SSE_OP2(andnpd); 2421 SSE_OP2(orps); 2422 SSE_OP2(orpd); 2423 SSE_OP2(xorps); 2424 SSE_OP2(xorpd); 2425 2426 SSE_OP2(unpcklps); 2427 SSE_OP2(unpcklpd); 2428 SSE_OP2(unpckhps); 2429 SSE_OP2(unpckhpd); 2430 2431 SHUF_OP(shufps, 0x78); 2432 SHUF_OP(shufpd, 0x02); 2433 2434 PSHUF_OP(pshufd, 0x78); 2435 PSHUF_OP(pshuflw, 0x78); 2436 PSHUF_OP(pshufhw, 0x78); 2437 2438 SHIFT_OP(psrlw, 7); 2439 SHIFT_OP(psrlw, 16); 2440 SHIFT_OP(psraw, 7); 2441 SHIFT_OP(psraw, 16); 2442 SHIFT_OP(psllw, 7); 2443 SHIFT_OP(psllw, 16); 2444 2445 SHIFT_OP(psrld, 7); 2446 SHIFT_OP(psrld, 32); 2447 SHIFT_OP(psrad, 7); 2448 SHIFT_OP(psrad, 32); 2449 SHIFT_OP(pslld, 7); 2450 SHIFT_OP(pslld, 32); 2451 2452 SHIFT_OP(psrlq, 7); 2453 SHIFT_OP(psrlq, 32); 2454 SHIFT_OP(psllq, 7); 2455 SHIFT_OP(psllq, 32); 2456 2457 SHIFT_IM(psrldq, 16); 2458 SHIFT_IM(psrldq, 7); 2459 SHIFT_IM(pslldq, 16); 2460 SHIFT_IM(pslldq, 7); 2461 2462 MOVMSK(movmskps); 2463 MOVMSK(movmskpd); 2464 2465 /* FPU specific ops */ 2466 2467 { 2468 uint32_t mxcsr; 2469 asm volatile("stmxcsr %0" : "=m" (mxcsr)); 2470 printf("mxcsr=%08x\n", mxcsr & 0x1f80); 2471 asm volatile("ldmxcsr %0" : : "m" (mxcsr)); 2472 } 2473 2474 test_sse_comi(2, -1); 2475 test_sse_comi(2, 2); 2476 test_sse_comi(2, 3); 2477 test_sse_comi(2, q_nan.d); 2478 test_sse_comi(q_nan.d, -1); 2479 2480 for(i = 0; i < 2; i++) { 2481 a.s[0] = 2.7; 2482 a.s[1] = 3.4; 2483 a.s[2] = 4; 2484 a.s[3] = -6.3; 2485 b.s[0] = 45.7; 2486 b.s[1] = 353.4; 2487 b.s[2] = 4; 2488 b.s[3] = 56.3; 2489 if (i == 1) { 2490 a.s[0] = q_nan.d; 2491 b.s[3] = q_nan.d; 2492 } 2493 2494 SSE_OPS(add); 2495 SSE_OPS(mul); 2496 SSE_OPS(sub); 2497 SSE_OPS(min); 2498 SSE_OPS(div); 2499 SSE_OPS(max); 2500 SSE_OPS(sqrt); 2501 SSE_OPS(cmpeq); 2502 SSE_OPS(cmplt); 2503 SSE_OPS(cmple); 2504 SSE_OPS(cmpunord); 2505 SSE_OPS(cmpneq); 2506 SSE_OPS(cmpnlt); 2507 SSE_OPS(cmpnle); 2508 SSE_OPS(cmpord); 2509 2510 2511 a.d[0] = 2.7; 2512 a.d[1] = -3.4; 2513 b.d[0] = 45.7; 2514 b.d[1] = -53.4; 2515 if (i == 1) { 2516 a.d[0] = q_nan.d; 2517 b.d[1] = q_nan.d; 2518 } 2519 SSE_OPD(add); 2520 SSE_OPD(mul); 2521 SSE_OPD(sub); 2522 SSE_OPD(min); 2523 SSE_OPD(div); 2524 SSE_OPD(max); 2525 SSE_OPD(sqrt); 2526 SSE_OPD(cmpeq); 2527 SSE_OPD(cmplt); 2528 SSE_OPD(cmple); 2529 SSE_OPD(cmpunord); 2530 SSE_OPD(cmpneq); 2531 SSE_OPD(cmpnlt); 2532 SSE_OPD(cmpnle); 2533 SSE_OPD(cmpord); 2534 } 2535 2536 /* float to float/int */ 2537 a.s[0] = 2.7; 2538 a.s[1] = 3.4; 2539 a.s[2] = 4; 2540 a.s[3] = -6.3; 2541 CVT_OP_XMM(cvtps2pd); 2542 CVT_OP_XMM(cvtss2sd); 2543 CVT_OP_XMM2MMX(cvtps2pi); 2544 CVT_OP_XMM2MMX(cvttps2pi); 2545 CVT_OP_XMM2REG(cvtss2si); 2546 CVT_OP_XMM2REG(cvttss2si); 2547 CVT_OP_XMM(cvtps2dq); 2548 CVT_OP_XMM(cvttps2dq); 2549 2550 a.d[0] = 2.6; 2551 a.d[1] = -3.4; 2552 CVT_OP_XMM(cvtpd2ps); 2553 CVT_OP_XMM(cvtsd2ss); 2554 CVT_OP_XMM2MMX(cvtpd2pi); 2555 CVT_OP_XMM2MMX(cvttpd2pi); 2556 CVT_OP_XMM2REG(cvtsd2si); 2557 CVT_OP_XMM2REG(cvttsd2si); 2558 CVT_OP_XMM(cvtpd2dq); 2559 CVT_OP_XMM(cvttpd2dq); 2560 2561 /* sse/mmx moves */ 2562 CVT_OP_XMM2MMX(movdq2q); 2563 CVT_OP_MMX2XMM(movq2dq); 2564 2565 /* int to float */ 2566 a.l[0] = -6; 2567 a.l[1] = 2; 2568 a.l[2] = 100; 2569 a.l[3] = -60000; 2570 CVT_OP_MMX2XMM(cvtpi2ps); 2571 CVT_OP_MMX2XMM(cvtpi2pd); 2572 CVT_OP_REG2XMM(cvtsi2ss); 2573 CVT_OP_REG2XMM(cvtsi2sd); 2574 CVT_OP_XMM(cvtdq2ps); 2575 CVT_OP_XMM(cvtdq2pd); 2576 2577 /* XXX: test PNI insns */ 2578 #if 0 2579 SSE_OP2(movshdup); 2580 #endif 2581 asm volatile ("emms"); 2582 } 2583 2584 #endif 2585 2586 extern void *__start_initcall; 2587 extern void *__stop_initcall; 2588 1630 2589 1631 2590 int main(int argc, char **argv) … … 1634 2593 void (*func)(void); 1635 2594 1636 ptr = & call_start + 1;1637 while ( *ptr != NULL) {2595 ptr = &__start_initcall; 2596 while (ptr != &__stop_initcall) { 1638 2597 func = *ptr++; 1639 2598 func(); … … 1643 2602 test_jcc(); 1644 2603 test_floats(); 2604 #if !defined(__x86_64__) 1645 2605 test_bcd(); 2606 #endif 1646 2607 test_xchg(); 1647 2608 test_string(); 1648 2609 test_misc(); 1649 2610 test_lea(); 2611 #ifdef TEST_SEGS 1650 2612 test_segs(); 1651 2613 test_code16(); 2614 #endif 2615 #ifdef TEST_VM86 1652 2616 test_vm86(); 2617 #endif 1653 2618 test_exceptions(); 2619 #if !defined(__x86_64__) 1654 2620 test_self_modifying_code(); 1655 2621 test_single_step(); 2622 #endif 2623 test_enter(); 2624 #ifdef TEST_SSE 2625 test_sse(); 2626 test_fxsave(); 2627 #endif 1656 2628 return 0; 1657 2629 } -
trunk/src/recompiler/tests/test-i386.h
r1 r2426 1 1 2 2 #define exec_op glue(exec_, OP) 3 #define exec_opq glue(glue(exec_, OP), q) 3 4 #define exec_opl glue(glue(exec_, OP), l) 4 5 #define exec_opw glue(glue(exec_, OP), w) 5 6 #define exec_opb glue(glue(exec_, OP), b) 6 7 7 #define EXECOP2(size, r es, s1, flags) \8 #define EXECOP2(size, rsize, res, s1, flags) \ 8 9 asm ("push %4\n\t"\ 9 10 "popf\n\t"\ 10 stringify(OP) size " %" size "2, %"size "0\n\t" \11 stringify(OP) size " %" rsize "2, %" rsize "0\n\t" \ 11 12 "pushf\n\t"\ 12 "pop l%1\n\t"\13 "pop %1\n\t"\ 13 14 : "=q" (res), "=g" (flags)\ 14 : "q" (s1), "0" (res), "1" (flags)); 15 : "q" (s1), "0" (res), "1" (flags)); \ 16 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CCIN=%04lx CC=%04lx\n", \ 17 stringify(OP) size, s0, s1, res, iflags, flags & CC_MASK); 15 18 16 #define EXECOP1(size, r es, flags) \19 #define EXECOP1(size, rsize, res, flags) \ 17 20 asm ("push %3\n\t"\ 18 21 "popf\n\t"\ 19 stringify(OP) size " %" size "0\n\t" \22 stringify(OP) size " %" rsize "0\n\t" \ 20 23 "pushf\n\t"\ 21 "pop l%1\n\t"\24 "pop %1\n\t"\ 22 25 : "=q" (res), "=g" (flags)\ 23 : "0" (res), "1" (flags)); 26 : "0" (res), "1" (flags)); \ 27 printf("%-10s A=" FMTLX " R=" FMTLX " CCIN=%04lx CC=%04lx\n", \ 28 stringify(OP) size, s0, res, iflags, flags & CC_MASK); 24 29 25 30 #ifdef OP1 26 void exec_opl(int s0, int s1, int iflags) 31 #if defined(__x86_64__) 32 void exec_opq(long s0, long s1, long iflags) 27 33 { 28 intres, flags;34 long res, flags; 29 35 res = s0; 30 36 flags = iflags; 31 EXECOP1("", res, flags); 32 printf("%-10s A=%08x R=%08x CCIN=%04x CC=%04x\n", 33 stringify(OP) "l", s0, res, iflags, flags & CC_MASK); 34 } 35 36 void exec_opw(int s0, int s1, int iflags) 37 { 38 int res, flags; 39 res = s0; 40 flags = iflags; 41 EXECOP1("w", res, flags); 42 printf("%-10s A=%08x R=%08x CCIN=%04x CC=%04x\n", 43 stringify(OP) "w", s0, res, iflags, flags & CC_MASK); 44 } 45 46 void exec_opb(int s0, int s1, int iflags) 47 { 48 int res, flags; 49 res = s0; 50 flags = iflags; 51 EXECOP1("b", res, flags); 52 printf("%-10s A=%08x R=%08x CCIN=%04x CC=%04x\n", 53 stringify(OP) "b", s0, res, iflags, flags & CC_MASK); 54 } 55 #else 56 void exec_opl(int s0, int s1, int iflags) 57 { 58 int res, flags; 59 res = s0; 60 flags = iflags; 61 EXECOP2("", res, s1, flags); 62 printf("%-10s A=%08x B=%08x R=%08x CCIN=%04x CC=%04x\n", 63 stringify(OP) "l", s0, s1, res, iflags, flags & CC_MASK); 64 } 65 66 void exec_opw(int s0, int s1, int iflags) 67 { 68 int res, flags; 69 res = s0; 70 flags = iflags; 71 EXECOP2("w", res, s1, flags); 72 printf("%-10s A=%08x B=%08x R=%08x CCIN=%04x CC=%04x\n", 73 stringify(OP) "w", s0, s1, res, iflags, flags & CC_MASK); 74 } 75 76 void exec_opb(int s0, int s1, int iflags) 77 { 78 int res, flags; 79 res = s0; 80 flags = iflags; 81 EXECOP2("b", res, s1, flags); 82 printf("%-10s A=%08x B=%08x R=%08x CCIN=%04x CC=%04x\n", 83 stringify(OP) "b", s0, s1, res, iflags, flags & CC_MASK); 37 EXECOP1("q", "", res, flags); 84 38 } 85 39 #endif 86 40 87 void exec_op (int s0, int s1)41 void exec_opl(long s0, long s1, long iflags) 88 42 { 43 long res, flags; 44 res = s0; 45 flags = iflags; 46 EXECOP1("l", "k", res, flags); 47 } 48 49 void exec_opw(long s0, long s1, long iflags) 50 { 51 long res, flags; 52 res = s0; 53 flags = iflags; 54 EXECOP1("w", "w", res, flags); 55 } 56 57 void exec_opb(long s0, long s1, long iflags) 58 { 59 long res, flags; 60 res = s0; 61 flags = iflags; 62 EXECOP1("b", "b", res, flags); 63 } 64 #else 65 #if defined(__x86_64__) 66 void exec_opq(long s0, long s1, long iflags) 67 { 68 long res, flags; 69 res = s0; 70 flags = iflags; 71 EXECOP2("q", "", res, s1, flags); 72 } 73 #endif 74 75 void exec_opl(long s0, long s1, long iflags) 76 { 77 long res, flags; 78 res = s0; 79 flags = iflags; 80 EXECOP2("l", "k", res, s1, flags); 81 } 82 83 void exec_opw(long s0, long s1, long iflags) 84 { 85 long res, flags; 86 res = s0; 87 flags = iflags; 88 EXECOP2("w", "w", res, s1, flags); 89 } 90 91 void exec_opb(long s0, long s1, long iflags) 92 { 93 long res, flags; 94 res = s0; 95 flags = iflags; 96 EXECOP2("b", "b", res, s1, flags); 97 } 98 #endif 99 100 void exec_op(long s0, long s1) 101 { 102 s0 = i2l(s0); 103 s1 = i2l(s1); 104 #if defined(__x86_64__) 105 exec_opq(s0, s1, 0); 106 #endif 89 107 exec_opl(s0, s1, 0); 90 108 exec_opw(s0, s1, 0); 91 109 exec_opb(s0, s1, 0); 92 110 #ifdef OP_CC 111 #if defined(__x86_64__) 112 exec_opq(s0, s1, CC_C); 113 #endif 93 114 exec_opl(s0, s1, CC_C); 94 115 exec_opw(s0, s1, CC_C);
Note:
See TracChangeset
for help on using the changeset viewer.