VirtualBox

source: vbox/trunk/include/VBox/vmm/hm.h@ 77213

Last change on this file since 77213 was 76993, checked in by vboxsync, 6 years ago

VMM: Nested VMX: bugref:9180 Allowing fetching VM-exit names from ring-0 as well. Various naming cleanups. Added HMDumpHwvirtVmxState() to be able to dump virtual VMCS state from ring-0 as well. Remove unusued HMIsVmxSupported() function.

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 11.7 KB
Line 
1/** @file
2 * HM - Intel/AMD VM Hardware Assisted Virtualization Manager (VMM)
3 */
4
5/*
6 * Copyright (C) 2006-2019 Oracle Corporation
7 *
8 * This file is part of VirtualBox Open Source Edition (OSE), as
9 * available from http://www.virtualbox.org. This file is free software;
10 * you can redistribute it and/or modify it under the terms of the GNU
11 * General Public License (GPL) as published by the Free Software
12 * Foundation, in version 2 as it comes in the "COPYING" file of the
13 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
14 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
15 *
16 * The contents of this file may alternatively be used under the terms
17 * of the Common Development and Distribution License Version 1.0
18 * (CDDL) only, as it comes in the "COPYING.CDDL" file of the
19 * VirtualBox OSE distribution, in which case the provisions of the
20 * CDDL are applicable instead of those of the GPL.
21 *
22 * You may elect to license modified versions of this file under the
23 * terms and conditions of either the GPL or the CDDL or both.
24 */
25
26#ifndef VBOX_INCLUDED_vmm_hm_h
27#define VBOX_INCLUDED_vmm_hm_h
28#ifndef RT_WITHOUT_PRAGMA_ONCE
29# pragma once
30#endif
31
32#include <VBox/vmm/pgm.h>
33#include <VBox/vmm/cpum.h>
34#include <VBox/vmm/vmm.h>
35#include <VBox/vmm/hm_svm.h>
36#include <VBox/vmm/hm_vmx.h>
37#include <VBox/vmm/trpm.h>
38#include <iprt/mp.h>
39
40
41/** @defgroup grp_hm The Hardware Assisted Virtualization Manager API
42 * @ingroup grp_vmm
43 * @{
44 */
45
46RT_C_DECLS_BEGIN
47
48/**
49 * Checks whether HM (VT-x/AMD-V) is being used by this VM.
50 *
51 * @retval true if used.
52 * @retval false if software virtualization (raw-mode) or NEM is used.
53 *
54 * @param a_pVM The cross context VM structure.
55 * @deprecated Please use VM_IS_RAW_MODE_ENABLED, VM_IS_HM_OR_NEM_ENABLED, or
56 * VM_IS_HM_ENABLED instead.
57 * @internal
58 */
59#if defined(VBOX_STRICT) && defined(IN_RING3)
60# define HMIsEnabled(a_pVM) HMIsEnabledNotMacro(a_pVM)
61#else
62# define HMIsEnabled(a_pVM) ((a_pVM)->fHMEnabled)
63#endif
64
65/**
66 * Checks whether raw-mode context is required for HM purposes
67 *
68 * @retval true if required by HM for doing switching the cpu to 64-bit mode.
69 * @retval false if not required by HM.
70 *
71 * @param a_pVM The cross context VM structure.
72 * @internal
73 */
74#if HC_ARCH_BITS == 64
75# define HMIsRawModeCtxNeeded(a_pVM) (false)
76#else
77# define HMIsRawModeCtxNeeded(a_pVM) ((a_pVM)->fHMNeedRawModeCtx)
78#endif
79
80/**
81 * Checks whether we're in the special hardware virtualization context.
82 * @returns true / false.
83 * @param a_pVCpu The caller's cross context virtual CPU structure.
84 * @thread EMT
85 */
86#ifdef IN_RING0
87# define HMIsInHwVirtCtx(a_pVCpu) (VMCPU_GET_STATE(a_pVCpu) == VMCPUSTATE_STARTED_HM)
88#else
89# define HMIsInHwVirtCtx(a_pVCpu) (false)
90#endif
91
92/**
93 * Checks whether we're in the special hardware virtualization context and we
94 * cannot perform long jump without guru meditating and possibly messing up the
95 * host and/or guest state.
96 *
97 * This is after we've turned interrupts off and such.
98 *
99 * @returns true / false.
100 * @param a_pVCpu The caller's cross context virtual CPU structure.
101 * @thread EMT
102 */
103#ifdef IN_RING0
104# define HMIsInHwVirtNoLongJmpCtx(a_pVCpu) (VMCPU_GET_STATE(a_pVCpu) == VMCPUSTATE_STARTED_EXEC)
105#else
106# define HMIsInHwVirtNoLongJmpCtx(a_pVCpu) (false)
107#endif
108
109/**
110 * 64-bit raw-mode (intermediate memory context) operations.
111 *
112 * These are special hypervisor eip values used when running 64-bit guests on
113 * 32-bit hosts. Each operation corresponds to a routine.
114 *
115 * @note Duplicated in the assembly code!
116 */
117typedef enum HM64ON32OP
118{
119 HM64ON32OP_INVALID = 0,
120 HM64ON32OP_VMXRCStartVM64,
121 HM64ON32OP_SVMRCVMRun64,
122 HM64ON32OP_HMRCSaveGuestFPU64,
123 HM64ON32OP_HMRCSaveGuestDebug64,
124 HM64ON32OP_HMRCTestSwitcher64,
125 HM64ON32OP_END,
126 HM64ON32OP_32BIT_HACK = 0x7fffffff
127} HM64ON32OP;
128
129/** @name All-context HM API.
130 * @{ */
131VMMDECL(bool) HMIsEnabledNotMacro(PVM pVM);
132VMMDECL(bool) HMCanExecuteGuest(PVMCPU pVCpu, PCCPUMCTX pCtx);
133VMM_INT_DECL(int) HMInvalidatePage(PVMCPU pVCpu, RTGCPTR GCVirt);
134VMM_INT_DECL(bool) HMHasPendingIrq(PVM pVM);
135VMM_INT_DECL(PX86PDPE) HMGetPaePdpes(PVMCPU pVCpu);
136VMM_INT_DECL(bool) HMSetSingleInstruction(PVM pVM, PVMCPU pVCpu, bool fEnable);
137VMM_INT_DECL(bool) HMIsSvmActive(PVM pVM);
138VMM_INT_DECL(bool) HMIsVmxActive(PVM pVM);
139VMM_INT_DECL(const char *) HMGetVmxDiagDesc(VMXVDIAG enmDiag);
140VMM_INT_DECL(const char *) HMGetVmxAbortDesc(VMXABORT enmAbort);
141VMM_INT_DECL(const char *) HMGetVmxVmcsStateDesc(uint8_t fVmcsState);
142VMM_INT_DECL(const char *) HMGetVmxIdtVectoringInfoTypeDesc(uint8_t uType);
143VMM_INT_DECL(const char *) HMGetVmxExitIntInfoTypeDesc(uint8_t uType);
144VMM_INT_DECL(const char *) HMGetVmxEntryIntInfoTypeDesc(uint8_t uType);
145VMM_INT_DECL(const char *) HMGetVmxExitName(uint32_t uExit);
146VMM_INT_DECL(const char *) HMGetSvmExitName(uint32_t uExit);
147VMM_INT_DECL(void) HMDumpHwvirtVmxState(PVMCPU pVCpu);
148VMM_INT_DECL(void) HMHCChangedPagingMode(PVM pVM, PVMCPU pVCpu, PGMMODE enmShadowMode, PGMMODE enmGuestMode);
149VMM_INT_DECL(void) HMGetVmxMsrsFromHwvirtMsrs(PCSUPHWVIRTMSRS pMsrs, PVMXMSRS pVmxMsrs);
150VMM_INT_DECL(void) HMGetSvmMsrsFromHwvirtMsrs(PCSUPHWVIRTMSRS pMsrs, PSVMMSRS pSvmMsrs);
151/** @} */
152
153/** @name All-context VMX helpers.
154 *
155 * These are hardware-assisted VMX functions (used by IEM/REM/CPUM and HM). Helpers
156 * based purely on the Intel VT-x specification (used by IEM/REM and HM) can be
157 * found in CPUM.
158 * @{ */
159VMM_INT_DECL(bool) HMCanExecuteVmxGuest(PVMCPU pVCpu, PCCPUMCTX pCtx);
160/** @} */
161
162/** @name All-context SVM helpers.
163 *
164 * These are hardware-assisted SVM functions (used by IEM/REM/CPUM and HM). Helpers
165 * based purely on the AMD SVM specification (used by IEM/REM and HM) can be found
166 * in CPUM.
167 * @{ */
168VMM_INT_DECL(TRPMEVENT) HMSvmEventToTrpmEventType(PCSVMEVENT pSvmEvent);
169/** @} */
170
171#ifndef IN_RC
172
173/** @name R0, R3 HM (VMX/SVM agnostic) handlers.
174 * @{ */
175VMM_INT_DECL(int) HMFlushTlb(PVMCPU pVCpu);
176VMM_INT_DECL(int) HMFlushTlbOnAllVCpus(PVM pVM);
177VMM_INT_DECL(int) HMInvalidatePageOnAllVCpus(PVM pVM, RTGCPTR GCVirt);
178VMM_INT_DECL(int) HMInvalidatePhysPage(PVM pVM, RTGCPHYS GCPhys);
179VMM_INT_DECL(bool) HMAreNestedPagingAndFullGuestExecEnabled(PVM pVM);
180VMM_INT_DECL(bool) HMIsLongModeAllowed(PVM pVM);
181VMM_INT_DECL(bool) HMIsNestedPagingActive(PVM pVM);
182VMM_INT_DECL(bool) HMIsMsrBitmapActive(PVM pVM);
183/** @} */
184
185/** @name R0, R3 SVM handlers.
186 * @{ */
187VMM_INT_DECL(bool) HMIsSvmVGifActive(PVM pVM);
188VMM_INT_DECL(uint64_t) HMApplySvmNstGstTscOffset(PVMCPU pVCpu, uint64_t uTicks);
189# ifdef VBOX_WITH_NESTED_HWVIRT_SVM
190VMM_INT_DECL(void) HMNotifySvmNstGstVmexit(PVMCPU pVCpu, PCPUMCTX pCtx);
191# endif
192VMM_INT_DECL(int) HMIsSubjectToSvmErratum170(uint32_t *pu32Family, uint32_t *pu32Model, uint32_t *pu32Stepping);
193VMM_INT_DECL(int) HMHCMaybeMovTprSvmHypercall(PVMCPU pVCpu);
194/** @} */
195
196#else /* Nops in RC: */
197
198/** @name RC HM (VMX/SVM agnostic) handlers.
199 * @{ */
200# define HMFlushTlb(pVCpu) do { } while (0)
201# define HMFlushTlbOnAllVCpus(pVM) do { } while (0)
202# define HMInvalidatePageOnAllVCpus(pVM, GCVirt) do { } while (0)
203# define HMInvalidatePhysPage(pVM, GCVirt) do { } while (0)
204# define HMAreNestedPagingAndFullGuestExecEnabled(pVM) false
205# define HMIsLongModeAllowed(pVM) false
206# define HMIsNestedPagingActive(pVM) false
207# define HMIsMsrBitmapsActive(pVM) false
208/** @} */
209
210/** @name RC SVM handlers.
211 * @{ */
212# define HMIsSvmVGifActive(pVM) false
213# define HMApplySvmNstGstTscOffset(pVCpu, uTicks) (uTicks)
214# define HMNotifySvmNstGstVmexit(pVCpu, pCtx) do { } while (0)
215# define HMIsSubjectToSvmErratum170(puFamily, puModel, puStepping) false
216# define HMHCMaybeMovTprSvmHypercall(pVCpu) do { } while (0)
217/** @} */
218
219#endif
220
221#ifdef IN_RING0
222/** @defgroup grp_hm_r0 The HM ring-0 Context API
223 * @{
224 */
225VMMR0_INT_DECL(int) HMR0Init(void);
226VMMR0_INT_DECL(int) HMR0Term(void);
227VMMR0_INT_DECL(int) HMR0InitVM(PVM pVM);
228VMMR0_INT_DECL(int) HMR0TermVM(PVM pVM);
229VMMR0_INT_DECL(int) HMR0EnableAllCpus(PVM pVM);
230# ifdef VBOX_WITH_RAW_MODE
231VMMR0_INT_DECL(int) HMR0EnterSwitcher(PVM pVM, VMMSWITCHER enmSwitcher, bool *pfVTxDisabled);
232VMMR0_INT_DECL(void) HMR0LeaveSwitcher(PVM pVM, bool fVTxDisabled);
233# endif
234
235VMMR0_INT_DECL(int) HMR0SetupVM(PVM pVM);
236VMMR0_INT_DECL(int) HMR0RunGuestCode(PVM pVM, PVMCPU pVCpu);
237VMMR0_INT_DECL(int) HMR0Enter(PVMCPU pVCpu);
238VMMR0_INT_DECL(int) HMR0LeaveCpu(PVMCPU pVCpu);
239VMMR0_INT_DECL(void) HMR0ThreadCtxCallback(RTTHREADCTXEVENT enmEvent, void *pvUser);
240VMMR0_INT_DECL(void) HMR0NotifyCpumUnloadedGuestFpuState(PVMCPU VCpu);
241VMMR0_INT_DECL(void) HMR0NotifyCpumModifiedHostCr0(PVMCPU VCpu);
242VMMR0_INT_DECL(bool) HMR0SuspendPending(void);
243VMMR0_INT_DECL(int) HMR0InvalidatePage(PVMCPU pVCpu, RTGCPTR GCVirt);
244VMMR0_INT_DECL(int) HMR0ImportStateOnDemand(PVMCPU pVCpu, uint64_t fWhat);
245
246# if HC_ARCH_BITS == 32 && defined(VBOX_WITH_64_BITS_GUESTS)
247VMMR0_INT_DECL(int) HMR0SaveFPUState(PVM pVM, PVMCPU pVCpu, PCPUMCTX pCtx);
248VMMR0_INT_DECL(int) HMR0SaveDebugState(PVM pVM, PVMCPU pVCpu, PCPUMCTX pCtx);
249VMMR0_INT_DECL(int) HMR0TestSwitcher3264(PVM pVM);
250# endif
251
252/** @} */
253#endif /* IN_RING0 */
254
255
256#ifdef IN_RING3
257/** @defgroup grp_hm_r3 The HM ring-3 Context API
258 * @{
259 */
260VMMR3DECL(bool) HMR3IsEnabled(PUVM pUVM);
261VMMR3DECL(bool) HMR3IsNestedPagingActive(PUVM pUVM);
262VMMR3DECL(bool) HMR3IsVirtApicRegsEnabled(PUVM pUVM);
263VMMR3DECL(bool) HMR3IsPostedIntrsEnabled(PUVM pUVM);
264VMMR3DECL(bool) HMR3IsVpidActive(PUVM pUVM);
265VMMR3DECL(bool) HMR3IsUXActive(PUVM pUVM);
266VMMR3DECL(bool) HMR3IsSvmEnabled(PUVM pUVM);
267VMMR3DECL(bool) HMR3IsVmxEnabled(PUVM pUVM);
268
269VMMR3_INT_DECL(bool) HMR3IsEventPending(PVMCPU pVCpu);
270VMMR3_INT_DECL(int) HMR3Init(PVM pVM);
271VMMR3_INT_DECL(int) HMR3InitCompleted(PVM pVM, VMINITCOMPLETED enmWhat);
272VMMR3_INT_DECL(void) HMR3Relocate(PVM pVM);
273VMMR3_INT_DECL(int) HMR3Term(PVM pVM);
274VMMR3_INT_DECL(void) HMR3Reset(PVM pVM);
275VMMR3_INT_DECL(void) HMR3ResetCpu(PVMCPU pVCpu);
276VMMR3_INT_DECL(void) HMR3CheckError(PVM pVM, int iStatusCode);
277VMMR3_INT_DECL(void) HMR3NotifyDebugEventChanged(PVM pVM);
278VMMR3_INT_DECL(void) HMR3NotifyDebugEventChangedPerCpu(PVM pVM, PVMCPU pVCpu);
279VMMR3_INT_DECL(bool) HMR3IsActive(PVMCPU pVCpu);
280VMMR3_INT_DECL(int) HMR3EnablePatching(PVM pVM, RTGCPTR pPatchMem, unsigned cbPatchMem);
281VMMR3_INT_DECL(int) HMR3DisablePatching(PVM pVM, RTGCPTR pPatchMem, unsigned cbPatchMem);
282VMMR3_INT_DECL(int) HMR3PatchTprInstr(PVM pVM, PVMCPU pVCpu);
283VMMR3_INT_DECL(bool) HMR3IsRescheduleRequired(PVM pVM, PCPUMCTX pCtx);
284VMMR3_INT_DECL(bool) HMR3IsVmxPreemptionTimerUsed(PVM pVM);
285/** @} */
286#endif /* IN_RING3 */
287
288/** @} */
289RT_C_DECLS_END
290
291
292#endif /* !VBOX_INCLUDED_vmm_hm_h */
293
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette