VirtualBox

source: vbox/trunk/src/VBox/VMM/VMMR3/EMHM.cpp@ 71800

Last change on this file since 71800 was 70979, checked in by vboxsync, 7 years ago

NEM: Working on the EM loops. bugref:9044

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 18.5 KB
Line 
1/* $Id: EMHM.cpp 70979 2018-02-13 01:38:48Z vboxsync $ */
2/** @file
3 * EM - Execution Monitor / Manager - hardware virtualization
4 */
5
6/*
7 * Copyright (C) 2006-2017 Oracle Corporation
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 */
17
18
19/*********************************************************************************************************************************
20* Header Files *
21*********************************************************************************************************************************/
22#define LOG_GROUP LOG_GROUP_EM
23#include <VBox/vmm/em.h>
24#include <VBox/vmm/vmm.h>
25#include <VBox/vmm/csam.h>
26#include <VBox/vmm/selm.h>
27#include <VBox/vmm/trpm.h>
28#include <VBox/vmm/iem.h>
29#include <VBox/vmm/iom.h>
30#include <VBox/vmm/dbgf.h>
31#include <VBox/vmm/pgm.h>
32#ifdef VBOX_WITH_REM
33# include <VBox/vmm/rem.h>
34#endif
35#include <VBox/vmm/tm.h>
36#include <VBox/vmm/mm.h>
37#include <VBox/vmm/ssm.h>
38#include <VBox/vmm/pdmapi.h>
39#include <VBox/vmm/pdmcritsect.h>
40#include <VBox/vmm/pdmqueue.h>
41#include <VBox/vmm/hm.h>
42#include "EMInternal.h"
43#include <VBox/vmm/vm.h>
44#include <VBox/vmm/gim.h>
45#include <VBox/vmm/cpumdis.h>
46#include <VBox/dis.h>
47#include <VBox/disopcode.h>
48#include <VBox/vmm/dbgf.h>
49#include "VMMTracing.h"
50
51#include <iprt/asm.h>
52
53
54/*********************************************************************************************************************************
55* Defined Constants And Macros *
56*********************************************************************************************************************************/
57#if 0 /* Disabled till after 2.1.0 when we've time to test it. */
58#define EM_NOTIFY_HM
59#endif
60
61
62/*********************************************************************************************************************************
63* Internal Functions *
64*********************************************************************************************************************************/
65DECLINLINE(int) emR3HmExecuteInstruction(PVM pVM, PVMCPU pVCpu, const char *pszPrefix, int rcGC = VINF_SUCCESS);
66static int emR3HmExecuteIOInstruction(PVM pVM, PVMCPU pVCpu);
67static int emR3HmForcedActions(PVM pVM, PVMCPU pVCpu, PCPUMCTX pCtx);
68
69#define EMHANDLERC_WITH_HM
70#define emR3ExecuteInstruction emR3HmExecuteInstruction
71#define emR3ExecuteIOInstruction emR3HmExecuteIOInstruction
72#include "EMHandleRCTmpl.h"
73
74
75/**
76 * Executes instruction in HM mode if we can.
77 *
78 * This is somewhat comparable to REMR3EmulateInstruction.
79 *
80 * @returns VBox strict status code.
81 * @retval VINF_EM_DBG_STEPPED on success.
82 * @retval VERR_EM_CANNOT_EXEC_GUEST if we cannot execute guest instructions in
83 * HM right now.
84 *
85 * @param pVM The cross context VM structure.
86 * @param pVCpu The cross context virtual CPU structure for the calling EMT.
87 * @param fFlags Combinations of EM_ONE_INS_FLAGS_XXX.
88 * @thread EMT.
89 */
90VMMR3_INT_DECL(VBOXSTRICTRC) EMR3HmSingleInstruction(PVM pVM, PVMCPU pVCpu, uint32_t fFlags)
91{
92 PCPUMCTX pCtx = pVCpu->em.s.pCtx;
93 Assert(!(fFlags & ~EM_ONE_INS_FLAGS_MASK));
94
95 if (!HMR3CanExecuteGuest(pVM, pCtx))
96 return VINF_EM_RESCHEDULE;
97
98 uint64_t const uOldRip = pCtx->rip;
99 for (;;)
100 {
101 /*
102 * Service necessary FFs before going into HM.
103 */
104 if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_PRE_RAW_MASK)
105 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_PRE_RAW_MASK))
106 {
107 VBOXSTRICTRC rcStrict = emR3HmForcedActions(pVM, pVCpu, pCtx);
108 if (rcStrict != VINF_SUCCESS)
109 {
110 Log(("EMR3HmSingleInstruction: FFs before -> %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
111 return rcStrict;
112 }
113 }
114
115 /*
116 * Go execute it.
117 */
118 bool fOld = HMSetSingleInstruction(pVM, pVCpu, true);
119 VBOXSTRICTRC rcStrict = VMMR3HmRunGC(pVM, pVCpu);
120 HMSetSingleInstruction(pVM, pVCpu, fOld);
121 LogFlow(("EMR3HmSingleInstruction: %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
122
123 /*
124 * Handle high priority FFs and informational status codes. We don't do
125 * normal FF processing the caller or the next call can deal with them.
126 */
127 VMCPU_FF_CLEAR(pVCpu, VMCPU_FF_RESUME_GUEST_MASK);
128 if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_POST_MASK)
129 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_POST_MASK))
130 {
131 rcStrict = emR3HighPriorityPostForcedActions(pVM, pVCpu, VBOXSTRICTRC_TODO(rcStrict));
132 LogFlow(("EMR3HmSingleInstruction: FFs after -> %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
133 }
134
135 if (rcStrict != VINF_SUCCESS && (rcStrict < VINF_EM_FIRST || rcStrict > VINF_EM_LAST))
136 {
137 rcStrict = emR3HmHandleRC(pVM, pVCpu, pCtx, VBOXSTRICTRC_TODO(rcStrict));
138 Log(("EMR3HmSingleInstruction: emR3HmHandleRC -> %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
139 }
140
141 /*
142 * Done?
143 */
144 if ( (rcStrict != VINF_SUCCESS && rcStrict != VINF_EM_DBG_STEPPED)
145 || !(fFlags & EM_ONE_INS_FLAGS_RIP_CHANGE)
146 || pCtx->rip != uOldRip)
147 {
148 if (rcStrict == VINF_SUCCESS && pCtx->rip != uOldRip)
149 rcStrict = VINF_EM_DBG_STEPPED;
150 Log(("EMR3HmSingleInstruction: returns %Rrc (rip %llx -> %llx)\n", VBOXSTRICTRC_VAL(rcStrict), uOldRip, pCtx->rip));
151 return rcStrict;
152 }
153 }
154}
155
156
157/**
158 * Executes one (or perhaps a few more) instruction(s).
159 *
160 * @returns VBox status code suitable for EM.
161 *
162 * @param pVM The cross context VM structure.
163 * @param pVCpu The cross context virtual CPU structure.
164 * @param rcRC Return code from RC.
165 * @param pszPrefix Disassembly prefix. If not NULL we'll disassemble the
166 * instruction and prefix the log output with this text.
167 */
168#if defined(LOG_ENABLED) || defined(DOXYGEN_RUNNING)
169static int emR3HmExecuteInstructionWorker(PVM pVM, PVMCPU pVCpu, int rcRC, const char *pszPrefix)
170#else
171static int emR3HmExecuteInstructionWorker(PVM pVM, PVMCPU pVCpu, int rcRC)
172#endif
173{
174#ifdef LOG_ENABLED
175 PCPUMCTX pCtx = pVCpu->em.s.pCtx;
176#endif
177 int rc;
178 NOREF(rcRC);
179
180#ifdef LOG_ENABLED
181 /*
182 * Log it.
183 */
184 Log(("EMINS: %04x:%RGv RSP=%RGv\n", pCtx->cs.Sel, (RTGCPTR)pCtx->rip, (RTGCPTR)pCtx->rsp));
185 if (pszPrefix)
186 {
187 DBGFR3_INFO_LOG(pVM, pVCpu, "cpumguest", pszPrefix);
188 DBGFR3_DISAS_INSTR_CUR_LOG(pVCpu, pszPrefix);
189 }
190#endif
191
192 /*
193 * Use IEM and fallback on REM if the functionality is missing.
194 * Once IEM gets mature enough, nothing should ever fall back.
195 */
196 STAM_PROFILE_START(&pVCpu->em.s.StatIEMEmu, a);
197 rc = VBOXSTRICTRC_TODO(IEMExecOne(pVCpu));
198 STAM_PROFILE_STOP(&pVCpu->em.s.StatIEMEmu, a);
199
200 if ( rc == VERR_IEM_ASPECT_NOT_IMPLEMENTED
201 || rc == VERR_IEM_INSTR_NOT_IMPLEMENTED)
202 {
203#ifdef VBOX_WITH_REM
204 STAM_PROFILE_START(&pVCpu->em.s.StatREMEmu, b);
205 EMRemLock(pVM);
206 /* Flush the recompiler TLB if the VCPU has changed. */
207 if (pVM->em.s.idLastRemCpu != pVCpu->idCpu)
208 CPUMSetChangedFlags(pVCpu, CPUM_CHANGED_ALL);
209 pVM->em.s.idLastRemCpu = pVCpu->idCpu;
210
211 rc = REMR3EmulateInstruction(pVM, pVCpu);
212 EMRemUnlock(pVM);
213 STAM_PROFILE_STOP(&pVCpu->em.s.StatREMEmu, b);
214#else /* !VBOX_WITH_REM */
215 NOREF(pVM);
216#endif /* !VBOX_WITH_REM */
217 }
218
219#ifdef EM_NOTIFY_HM
220 if (pVCpu->em.s.enmState == EMSTATE_DEBUG_GUEST_HM)
221 HMR3NotifyEmulated(pVCpu);
222#endif
223 return rc;
224}
225
226
227/**
228 * Executes one (or perhaps a few more) instruction(s).
229 * This is just a wrapper for discarding pszPrefix in non-logging builds.
230 *
231 * @returns VBox status code suitable for EM.
232 * @param pVM The cross context VM structure.
233 * @param pVCpu The cross context virtual CPU structure.
234 * @param pszPrefix Disassembly prefix. If not NULL we'll disassemble the
235 * instruction and prefix the log output with this text.
236 * @param rcGC GC return code
237 */
238DECLINLINE(int) emR3HmExecuteInstruction(PVM pVM, PVMCPU pVCpu, const char *pszPrefix, int rcGC)
239{
240#ifdef LOG_ENABLED
241 return emR3HmExecuteInstructionWorker(pVM, pVCpu, rcGC, pszPrefix);
242#else
243 RT_NOREF_PV(pszPrefix);
244 return emR3HmExecuteInstructionWorker(pVM, pVCpu, rcGC);
245#endif
246}
247
248/**
249 * Executes one (or perhaps a few more) IO instruction(s).
250 *
251 * @returns VBox status code suitable for EM.
252 * @param pVM The cross context VM structure.
253 * @param pVCpu The cross context virtual CPU structure.
254 */
255static int emR3HmExecuteIOInstruction(PVM pVM, PVMCPU pVCpu)
256{
257 PCPUMCTX pCtx = pVCpu->em.s.pCtx;
258
259 STAM_PROFILE_START(&pVCpu->em.s.StatIOEmu, a);
260
261 /*
262 * Try to restart the io instruction that was refused in ring-0.
263 */
264 VBOXSTRICTRC rcStrict = HMR3RestartPendingIOInstr(pVM, pVCpu, pCtx);
265 if (IOM_SUCCESS(rcStrict))
266 {
267 STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatIoRestarted);
268 STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
269 return VBOXSTRICTRC_TODO(rcStrict); /* rip already updated. */
270 }
271 AssertMsgReturn(rcStrict == VERR_NOT_FOUND, ("%Rrc\n", VBOXSTRICTRC_VAL(rcStrict)),
272 RT_SUCCESS_NP(rcStrict) ? VERR_IPE_UNEXPECTED_INFO_STATUS : VBOXSTRICTRC_TODO(rcStrict));
273
274 /*
275 * Hand it over to the interpreter.
276 */
277 rcStrict = IEMExecOne(pVCpu);
278 LogFlow(("emR3HmExecuteIOInstruction: %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
279 STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatIoIem);
280 STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
281 return VBOXSTRICTRC_TODO(rcStrict);
282}
283
284
285/**
286 * Process HM specific forced actions.
287 *
288 * This function is called when any FFs in the VM_FF_HIGH_PRIORITY_PRE_RAW_MASK
289 * or/and VMCPU_FF_HIGH_PRIORITY_PRE_RAW_MASK are pending.
290 *
291 * @returns VBox status code. May return VINF_EM_NO_MEMORY but none of the other
292 * EM statuses.
293 * @param pVM The cross context VM structure.
294 * @param pVCpu The cross context virtual CPU structure.
295 * @param pCtx Pointer to the guest CPU context.
296 */
297static int emR3HmForcedActions(PVM pVM, PVMCPU pVCpu, PCPUMCTX pCtx)
298{
299 /*
300 * Sync page directory.
301 */
302 if (VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_PGM_SYNC_CR3 | VMCPU_FF_PGM_SYNC_CR3_NON_GLOBAL))
303 {
304 Assert(pVCpu->em.s.enmState != EMSTATE_WAIT_SIPI);
305 int rc = PGMSyncCR3(pVCpu, pCtx->cr0, pCtx->cr3, pCtx->cr4, VMCPU_FF_IS_SET(pVCpu, VMCPU_FF_PGM_SYNC_CR3));
306 if (RT_FAILURE(rc))
307 return rc;
308
309#ifdef VBOX_WITH_RAW_MODE
310 Assert(!VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT));
311#endif
312
313 /* Prefetch pages for EIP and ESP. */
314 /** @todo This is rather expensive. Should investigate if it really helps at all. */
315 rc = PGMPrefetchPage(pVCpu, SELMToFlat(pVM, DISSELREG_CS, CPUMCTX2CORE(pCtx), pCtx->rip));
316 if (rc == VINF_SUCCESS)
317 rc = PGMPrefetchPage(pVCpu, SELMToFlat(pVM, DISSELREG_SS, CPUMCTX2CORE(pCtx), pCtx->rsp));
318 if (rc != VINF_SUCCESS)
319 {
320 if (rc != VINF_PGM_SYNC_CR3)
321 {
322 AssertLogRelMsgReturn(RT_FAILURE(rc), ("%Rrc\n", rc), VERR_IPE_UNEXPECTED_INFO_STATUS);
323 return rc;
324 }
325 rc = PGMSyncCR3(pVCpu, pCtx->cr0, pCtx->cr3, pCtx->cr4, VMCPU_FF_IS_SET(pVCpu, VMCPU_FF_PGM_SYNC_CR3));
326 if (RT_FAILURE(rc))
327 return rc;
328 }
329 /** @todo maybe prefetch the supervisor stack page as well */
330#ifdef VBOX_WITH_RAW_MODE
331 Assert(!VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT));
332#endif
333 }
334
335 /*
336 * Allocate handy pages (just in case the above actions have consumed some pages).
337 */
338 if (VM_FF_IS_PENDING_EXCEPT(pVM, VM_FF_PGM_NEED_HANDY_PAGES, VM_FF_PGM_NO_MEMORY))
339 {
340 int rc = PGMR3PhysAllocateHandyPages(pVM);
341 if (RT_FAILURE(rc))
342 return rc;
343 }
344
345 /*
346 * Check whether we're out of memory now.
347 *
348 * This may stem from some of the above actions or operations that has been executed
349 * since we ran FFs. The allocate handy pages must for instance always be followed by
350 * this check.
351 */
352 if (VM_FF_IS_PENDING(pVM, VM_FF_PGM_NO_MEMORY))
353 return VINF_EM_NO_MEMORY;
354
355 return VINF_SUCCESS;
356}
357
358
359/**
360 * Executes hardware accelerated raw code. (Intel VT-x & AMD-V)
361 *
362 * This function contains the raw-mode version of the inner
363 * execution loop (the outer loop being in EMR3ExecuteVM()).
364 *
365 * @returns VBox status code. The most important ones are: VINF_EM_RESCHEDULE, VINF_EM_RESCHEDULE_RAW,
366 * VINF_EM_RESCHEDULE_REM, VINF_EM_SUSPEND, VINF_EM_RESET and VINF_EM_TERMINATE.
367 *
368 * @param pVM The cross context VM structure.
369 * @param pVCpu The cross context virtual CPU structure.
370 * @param pfFFDone Where to store an indicator telling whether or not
371 * FFs were done before returning.
372 */
373int emR3HmExecute(PVM pVM, PVMCPU pVCpu, bool *pfFFDone)
374{
375 int rc = VERR_IPE_UNINITIALIZED_STATUS;
376 PCPUMCTX pCtx = pVCpu->em.s.pCtx;
377
378 LogFlow(("emR3HmExecute%d: (cs:eip=%04x:%RGv)\n", pVCpu->idCpu, pCtx->cs.Sel, (RTGCPTR)pCtx->rip));
379 *pfFFDone = false;
380
381 STAM_COUNTER_INC(&pVCpu->em.s.StatHMExecuteCalled);
382
383#ifdef EM_NOTIFY_HM
384 HMR3NotifyScheduled(pVCpu);
385#endif
386
387 /*
388 * Spin till we get a forced action which returns anything but VINF_SUCCESS.
389 */
390 for (;;)
391 {
392 STAM_PROFILE_ADV_START(&pVCpu->em.s.StatHMEntry, a);
393
394 /* Check if a forced reschedule is pending. */
395 if (HMR3IsRescheduleRequired(pVM, pCtx))
396 {
397 rc = VINF_EM_RESCHEDULE;
398 break;
399 }
400
401 /*
402 * Process high priority pre-execution raw-mode FFs.
403 */
404#ifdef VBOX_WITH_RAW_MODE
405 Assert(!VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_SELM_SYNC_TSS | VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT));
406#endif
407 if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_PRE_RAW_MASK)
408 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_PRE_RAW_MASK))
409 {
410 rc = emR3HmForcedActions(pVM, pVCpu, pCtx);
411 if (rc != VINF_SUCCESS)
412 break;
413 }
414
415#ifdef LOG_ENABLED
416 /*
417 * Log important stuff before entering GC.
418 */
419 if (TRPMHasTrap(pVCpu))
420 Log(("CPU%d: Pending hardware interrupt=0x%x cs:rip=%04X:%RGv\n", pVCpu->idCpu, TRPMGetTrapNo(pVCpu), pCtx->cs.Sel, (RTGCPTR)pCtx->rip));
421
422 uint32_t cpl = CPUMGetGuestCPL(pVCpu);
423 if (pVM->cCpus == 1)
424 {
425 if (pCtx->eflags.Bits.u1VM)
426 Log(("HWV86: %08X IF=%d\n", pCtx->eip, pCtx->eflags.Bits.u1IF));
427 else if (CPUMIsGuestIn64BitCodeEx(pCtx))
428 Log(("HWR%d: %04X:%RGv ESP=%RGv IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pCtx->cs.Sel, (RTGCPTR)pCtx->rip, pCtx->rsp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
429 else
430 Log(("HWR%d: %04X:%08X ESP=%08X IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pCtx->cs.Sel, pCtx->eip, pCtx->esp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
431 }
432 else
433 {
434 if (pCtx->eflags.Bits.u1VM)
435 Log(("HWV86-CPU%d: %08X IF=%d\n", pVCpu->idCpu, pCtx->eip, pCtx->eflags.Bits.u1IF));
436 else if (CPUMIsGuestIn64BitCodeEx(pCtx))
437 Log(("HWR%d-CPU%d: %04X:%RGv ESP=%RGv IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pVCpu->idCpu, pCtx->cs.Sel, (RTGCPTR)pCtx->rip, pCtx->rsp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
438 else
439 Log(("HWR%d-CPU%d: %04X:%08X ESP=%08X IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pVCpu->idCpu, pCtx->cs.Sel, pCtx->eip, pCtx->esp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
440 }
441#endif /* LOG_ENABLED */
442
443 /*
444 * Execute the code.
445 */
446 STAM_PROFILE_ADV_STOP(&pVCpu->em.s.StatHMEntry, a);
447
448 if (RT_LIKELY(emR3IsExecutionAllowed(pVM, pVCpu)))
449 {
450 STAM_PROFILE_START(&pVCpu->em.s.StatHMExec, x);
451 rc = VMMR3HmRunGC(pVM, pVCpu);
452 STAM_PROFILE_STOP(&pVCpu->em.s.StatHMExec, x);
453 }
454 else
455 {
456 /* Give up this time slice; virtual time continues */
457 STAM_REL_PROFILE_ADV_START(&pVCpu->em.s.StatCapped, u);
458 RTThreadSleep(5);
459 STAM_REL_PROFILE_ADV_STOP(&pVCpu->em.s.StatCapped, u);
460 rc = VINF_SUCCESS;
461 }
462
463
464 /*
465 * Deal with high priority post execution FFs before doing anything else.
466 */
467 VMCPU_FF_CLEAR(pVCpu, VMCPU_FF_RESUME_GUEST_MASK);
468 if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_POST_MASK)
469 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_POST_MASK))
470 rc = emR3HighPriorityPostForcedActions(pVM, pVCpu, rc);
471
472 /*
473 * Process the returned status code.
474 */
475 if (rc >= VINF_EM_FIRST && rc <= VINF_EM_LAST)
476 break;
477
478 rc = emR3HmHandleRC(pVM, pVCpu, pCtx, rc);
479 if (rc != VINF_SUCCESS)
480 break;
481
482 /*
483 * Check and execute forced actions.
484 */
485#ifdef VBOX_HIGH_RES_TIMERS_HACK
486 TMTimerPollVoid(pVM, pVCpu);
487#endif
488 if ( VM_FF_IS_PENDING(pVM, VM_FF_ALL_MASK)
489 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_ALL_MASK))
490 {
491 rc = emR3ForcedActions(pVM, pVCpu, rc);
492 VBOXVMM_EM_FF_ALL_RET(pVCpu, rc);
493 if ( rc != VINF_SUCCESS
494 && rc != VINF_EM_RESCHEDULE_HM)
495 {
496 *pfFFDone = true;
497 break;
498 }
499 }
500 }
501
502 /*
503 * Return to outer loop.
504 */
505#if defined(LOG_ENABLED) && defined(DEBUG)
506 RTLogFlush(NULL);
507#endif
508 return rc;
509}
510
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette