VirtualBox

source: vbox/trunk/include/VBox/vmm/cpum.h@ 72891

Last change on this file since 72891 was 72741, checked in by vboxsync, 6 years ago

VMM: Extend HM changed flags. bugref:9193

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 79.1 KB
Line 
1/** @file
2 * CPUM - CPU Monitor(/ Manager).
3 */
4
5/*
6 * Copyright (C) 2006-2017 Oracle Corporation
7 *
8 * This file is part of VirtualBox Open Source Edition (OSE), as
9 * available from http://www.virtualbox.org. This file is free software;
10 * you can redistribute it and/or modify it under the terms of the GNU
11 * General Public License (GPL) as published by the Free Software
12 * Foundation, in version 2 as it comes in the "COPYING" file of the
13 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
14 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
15 *
16 * The contents of this file may alternatively be used under the terms
17 * of the Common Development and Distribution License Version 1.0
18 * (CDDL) only, as it comes in the "COPYING.CDDL" file of the
19 * VirtualBox OSE distribution, in which case the provisions of the
20 * CDDL are applicable instead of those of the GPL.
21 *
22 * You may elect to license modified versions of this file under the
23 * terms and conditions of either the GPL or the CDDL or both.
24 */
25
26#ifndef ___VBox_vmm_cpum_h
27#define ___VBox_vmm_cpum_h
28
29#include <iprt/x86.h>
30#include <VBox/types.h>
31#include <VBox/vmm/cpumctx.h>
32#include <VBox/vmm/stam.h>
33#include <VBox/vmm/vmapi.h>
34
35RT_C_DECLS_BEGIN
36
37/** @defgroup grp_cpum The CPU Monitor / Manager API
38 * @ingroup grp_vmm
39 * @{
40 */
41
42/**
43 * CPUID feature to set or clear.
44 */
45typedef enum CPUMCPUIDFEATURE
46{
47 CPUMCPUIDFEATURE_INVALID = 0,
48 /** The APIC feature bit. (Std+Ext)
49 * Note! There is a per-cpu flag for masking this CPUID feature bit when the
50 * APICBASE.ENABLED bit is zero. So, this feature is only set/cleared
51 * at VM construction time like all the others. This didn't used to be
52 * that way, this is new with 5.1. */
53 CPUMCPUIDFEATURE_APIC,
54 /** The sysenter/sysexit feature bit. (Std) */
55 CPUMCPUIDFEATURE_SEP,
56 /** The SYSCALL/SYSEXIT feature bit (64 bits mode only for Intel CPUs). (Ext) */
57 CPUMCPUIDFEATURE_SYSCALL,
58 /** The PAE feature bit. (Std+Ext) */
59 CPUMCPUIDFEATURE_PAE,
60 /** The NX feature bit. (Ext) */
61 CPUMCPUIDFEATURE_NX,
62 /** The LAHF/SAHF feature bit (64 bits mode only). (Ext) */
63 CPUMCPUIDFEATURE_LAHF,
64 /** The LONG MODE feature bit. (Ext) */
65 CPUMCPUIDFEATURE_LONG_MODE,
66 /** The PAT feature bit. (Std+Ext) */
67 CPUMCPUIDFEATURE_PAT,
68 /** The x2APIC feature bit. (Std) */
69 CPUMCPUIDFEATURE_X2APIC,
70 /** The RDTSCP feature bit. (Ext) */
71 CPUMCPUIDFEATURE_RDTSCP,
72 /** The Hypervisor Present bit. (Std) */
73 CPUMCPUIDFEATURE_HVP,
74 /** The MWait Extensions bits (Std) */
75 CPUMCPUIDFEATURE_MWAIT_EXTS,
76 /** The speculation control feature bits. (StExt) */
77 CPUMCPUIDFEATURE_SPEC_CTRL,
78 /** 32bit hackishness. */
79 CPUMCPUIDFEATURE_32BIT_HACK = 0x7fffffff
80} CPUMCPUIDFEATURE;
81
82/**
83 * CPU Vendor.
84 */
85typedef enum CPUMCPUVENDOR
86{
87 CPUMCPUVENDOR_INVALID = 0,
88 CPUMCPUVENDOR_INTEL,
89 CPUMCPUVENDOR_AMD,
90 CPUMCPUVENDOR_VIA,
91 CPUMCPUVENDOR_CYRIX,
92 CPUMCPUVENDOR_UNKNOWN,
93 /** 32bit hackishness. */
94 CPUMCPUVENDOR_32BIT_HACK = 0x7fffffff
95} CPUMCPUVENDOR;
96
97
98/**
99 * X86 and AMD64 CPU microarchitectures and in processor generations.
100 *
101 * @remarks The separation here is sometimes a little bit too finely grained,
102 * and the differences is more like processor generation than micro
103 * arch. This can be useful, so we'll provide functions for getting at
104 * more coarse grained info.
105 */
106typedef enum CPUMMICROARCH
107{
108 kCpumMicroarch_Invalid = 0,
109
110 kCpumMicroarch_Intel_First,
111
112 kCpumMicroarch_Intel_8086 = kCpumMicroarch_Intel_First,
113 kCpumMicroarch_Intel_80186,
114 kCpumMicroarch_Intel_80286,
115 kCpumMicroarch_Intel_80386,
116 kCpumMicroarch_Intel_80486,
117 kCpumMicroarch_Intel_P5,
118
119 kCpumMicroarch_Intel_P6_Core_Atom_First,
120 kCpumMicroarch_Intel_P6 = kCpumMicroarch_Intel_P6_Core_Atom_First,
121 kCpumMicroarch_Intel_P6_II,
122 kCpumMicroarch_Intel_P6_III,
123
124 kCpumMicroarch_Intel_P6_M_Banias,
125 kCpumMicroarch_Intel_P6_M_Dothan,
126 kCpumMicroarch_Intel_Core_Yonah, /**< Core, also known as Enhanced Pentium M. */
127
128 kCpumMicroarch_Intel_Core2_First,
129 kCpumMicroarch_Intel_Core2_Merom = kCpumMicroarch_Intel_Core2_First, /**< 65nm, Merom/Conroe/Kentsfield/Tigerton */
130 kCpumMicroarch_Intel_Core2_Penryn, /**< 45nm, Penryn/Wolfdale/Yorkfield/Harpertown */
131 kCpumMicroarch_Intel_Core2_End,
132
133 kCpumMicroarch_Intel_Core7_First,
134 kCpumMicroarch_Intel_Core7_Nehalem = kCpumMicroarch_Intel_Core7_First,
135 kCpumMicroarch_Intel_Core7_Westmere,
136 kCpumMicroarch_Intel_Core7_SandyBridge,
137 kCpumMicroarch_Intel_Core7_IvyBridge,
138 kCpumMicroarch_Intel_Core7_Haswell,
139 kCpumMicroarch_Intel_Core7_Broadwell,
140 kCpumMicroarch_Intel_Core7_Skylake,
141 kCpumMicroarch_Intel_Core7_KabyLake,
142 kCpumMicroarch_Intel_Core7_CoffeeLake,
143 kCpumMicroarch_Intel_Core7_CannonLake,
144 kCpumMicroarch_Intel_Core7_IceLake,
145 kCpumMicroarch_Intel_Core7_TigerLake,
146 kCpumMicroarch_Intel_Core7_End,
147
148 kCpumMicroarch_Intel_Atom_First,
149 kCpumMicroarch_Intel_Atom_Bonnell = kCpumMicroarch_Intel_Atom_First,
150 kCpumMicroarch_Intel_Atom_Lincroft, /**< Second generation bonnell (44nm). */
151 kCpumMicroarch_Intel_Atom_Saltwell, /**< 32nm shrink of Bonnell. */
152 kCpumMicroarch_Intel_Atom_Silvermont, /**< 22nm */
153 kCpumMicroarch_Intel_Atom_Airmount, /**< 14nm */
154 kCpumMicroarch_Intel_Atom_Goldmont, /**< 14nm */
155 kCpumMicroarch_Intel_Atom_GoldmontPlus, /**< 14nm */
156 kCpumMicroarch_Intel_Atom_Unknown,
157 kCpumMicroarch_Intel_Atom_End,
158
159
160 kCpumMicroarch_Intel_Phi_First,
161 kCpumMicroarch_Intel_Phi_KnightsFerry = kCpumMicroarch_Intel_Phi_First,
162 kCpumMicroarch_Intel_Phi_KnightsCorner,
163 kCpumMicroarch_Intel_Phi_KnightsLanding,
164 kCpumMicroarch_Intel_Phi_KnightsHill,
165 kCpumMicroarch_Intel_Phi_KnightsMill,
166 kCpumMicroarch_Intel_Phi_End,
167
168 kCpumMicroarch_Intel_P6_Core_Atom_End,
169
170 kCpumMicroarch_Intel_NB_First,
171 kCpumMicroarch_Intel_NB_Willamette = kCpumMicroarch_Intel_NB_First, /**< 180nm */
172 kCpumMicroarch_Intel_NB_Northwood, /**< 130nm */
173 kCpumMicroarch_Intel_NB_Prescott, /**< 90nm */
174 kCpumMicroarch_Intel_NB_Prescott2M, /**< 90nm */
175 kCpumMicroarch_Intel_NB_CedarMill, /**< 65nm */
176 kCpumMicroarch_Intel_NB_Gallatin, /**< 90nm Xeon, Pentium 4 Extreme Edition ("Emergency Edition"). */
177 kCpumMicroarch_Intel_NB_Unknown,
178 kCpumMicroarch_Intel_NB_End,
179
180 kCpumMicroarch_Intel_Unknown,
181 kCpumMicroarch_Intel_End,
182
183 kCpumMicroarch_AMD_First,
184 kCpumMicroarch_AMD_Am286 = kCpumMicroarch_AMD_First,
185 kCpumMicroarch_AMD_Am386,
186 kCpumMicroarch_AMD_Am486,
187 kCpumMicroarch_AMD_Am486Enh, /**< Covers Am5x86 as well. */
188 kCpumMicroarch_AMD_K5,
189 kCpumMicroarch_AMD_K6,
190
191 kCpumMicroarch_AMD_K7_First,
192 kCpumMicroarch_AMD_K7_Palomino = kCpumMicroarch_AMD_K7_First,
193 kCpumMicroarch_AMD_K7_Spitfire,
194 kCpumMicroarch_AMD_K7_Thunderbird,
195 kCpumMicroarch_AMD_K7_Morgan,
196 kCpumMicroarch_AMD_K7_Thoroughbred,
197 kCpumMicroarch_AMD_K7_Barton,
198 kCpumMicroarch_AMD_K7_Unknown,
199 kCpumMicroarch_AMD_K7_End,
200
201 kCpumMicroarch_AMD_K8_First,
202 kCpumMicroarch_AMD_K8_130nm = kCpumMicroarch_AMD_K8_First, /**< 130nm Clawhammer, Sledgehammer, Newcastle, Paris, Odessa, Dublin */
203 kCpumMicroarch_AMD_K8_90nm, /**< 90nm shrink */
204 kCpumMicroarch_AMD_K8_90nm_DualCore, /**< 90nm with two cores. */
205 kCpumMicroarch_AMD_K8_90nm_AMDV, /**< 90nm with AMD-V (usually) and two cores (usually). */
206 kCpumMicroarch_AMD_K8_65nm, /**< 65nm shrink. */
207 kCpumMicroarch_AMD_K8_End,
208
209 kCpumMicroarch_AMD_K10,
210 kCpumMicroarch_AMD_K10_Lion,
211 kCpumMicroarch_AMD_K10_Llano,
212 kCpumMicroarch_AMD_Bobcat,
213 kCpumMicroarch_AMD_Jaguar,
214
215 kCpumMicroarch_AMD_15h_First,
216 kCpumMicroarch_AMD_15h_Bulldozer = kCpumMicroarch_AMD_15h_First,
217 kCpumMicroarch_AMD_15h_Piledriver,
218 kCpumMicroarch_AMD_15h_Steamroller, /**< Yet to be released, might have different family. */
219 kCpumMicroarch_AMD_15h_Excavator, /**< Yet to be released, might have different family. */
220 kCpumMicroarch_AMD_15h_Unknown,
221 kCpumMicroarch_AMD_15h_End,
222
223 kCpumMicroarch_AMD_16h_First,
224 kCpumMicroarch_AMD_16h_End,
225
226 kCpumMicroarch_AMD_Zen_First,
227 kCpumMicroarch_AMD_Zen_Ryzen = kCpumMicroarch_AMD_Zen_First,
228 kCpumMicroarch_AMD_Zen_End,
229
230 kCpumMicroarch_AMD_Unknown,
231 kCpumMicroarch_AMD_End,
232
233 kCpumMicroarch_VIA_First,
234 kCpumMicroarch_Centaur_C6 = kCpumMicroarch_VIA_First,
235 kCpumMicroarch_Centaur_C2,
236 kCpumMicroarch_Centaur_C3,
237 kCpumMicroarch_VIA_C3_M2,
238 kCpumMicroarch_VIA_C3_C5A, /**< 180nm Samuel - Cyrix III, C3, 1GigaPro. */
239 kCpumMicroarch_VIA_C3_C5B, /**< 150nm Samuel 2 - Cyrix III, C3, 1GigaPro, Eden ESP, XP 2000+. */
240 kCpumMicroarch_VIA_C3_C5C, /**< 130nm Ezra - C3, Eden ESP. */
241 kCpumMicroarch_VIA_C3_C5N, /**< 130nm Ezra-T - C3. */
242 kCpumMicroarch_VIA_C3_C5XL, /**< 130nm Nehemiah - C3, Eden ESP, Eden-N. */
243 kCpumMicroarch_VIA_C3_C5P, /**< 130nm Nehemiah+ - C3. */
244 kCpumMicroarch_VIA_C7_C5J, /**< 90nm Esther - C7, C7-D, C7-M, Eden, Eden ULV. */
245 kCpumMicroarch_VIA_Isaiah,
246 kCpumMicroarch_VIA_Unknown,
247 kCpumMicroarch_VIA_End,
248
249 kCpumMicroarch_Cyrix_First,
250 kCpumMicroarch_Cyrix_5x86 = kCpumMicroarch_Cyrix_First,
251 kCpumMicroarch_Cyrix_M1,
252 kCpumMicroarch_Cyrix_MediaGX,
253 kCpumMicroarch_Cyrix_MediaGXm,
254 kCpumMicroarch_Cyrix_M2,
255 kCpumMicroarch_Cyrix_Unknown,
256 kCpumMicroarch_Cyrix_End,
257
258 kCpumMicroarch_NEC_First,
259 kCpumMicroarch_NEC_V20 = kCpumMicroarch_NEC_First,
260 kCpumMicroarch_NEC_V30,
261 kCpumMicroarch_NEC_End,
262
263 kCpumMicroarch_Unknown,
264
265 kCpumMicroarch_32BitHack = 0x7fffffff
266} CPUMMICROARCH;
267
268
269/** Predicate macro for catching netburst CPUs. */
270#define CPUMMICROARCH_IS_INTEL_NETBURST(a_enmMicroarch) \
271 ((a_enmMicroarch) >= kCpumMicroarch_Intel_NB_First && (a_enmMicroarch) <= kCpumMicroarch_Intel_NB_End)
272
273/** Predicate macro for catching Core7 CPUs. */
274#define CPUMMICROARCH_IS_INTEL_CORE7(a_enmMicroarch) \
275 ((a_enmMicroarch) >= kCpumMicroarch_Intel_Core7_First && (a_enmMicroarch) <= kCpumMicroarch_Intel_Core7_End)
276
277/** Predicate macro for catching Core 2 CPUs. */
278#define CPUMMICROARCH_IS_INTEL_CORE2(a_enmMicroarch) \
279 ((a_enmMicroarch) >= kCpumMicroarch_Intel_Core2_First && (a_enmMicroarch) <= kCpumMicroarch_Intel_Core2_End)
280
281/** Predicate macro for catching Atom CPUs, Silvermont and upwards. */
282#define CPUMMICROARCH_IS_INTEL_SILVERMONT_PLUS(a_enmMicroarch) \
283 ((a_enmMicroarch) >= kCpumMicroarch_Intel_Atom_Silvermont && (a_enmMicroarch) <= kCpumMicroarch_Intel_Atom_End)
284
285/** Predicate macro for catching AMD Family OFh CPUs (aka K8). */
286#define CPUMMICROARCH_IS_AMD_FAM_0FH(a_enmMicroarch) \
287 ((a_enmMicroarch) >= kCpumMicroarch_AMD_K8_First && (a_enmMicroarch) <= kCpumMicroarch_AMD_K8_End)
288
289/** Predicate macro for catching AMD Family 10H CPUs (aka K10). */
290#define CPUMMICROARCH_IS_AMD_FAM_10H(a_enmMicroarch) ((a_enmMicroarch) == kCpumMicroarch_AMD_K10)
291
292/** Predicate macro for catching AMD Family 11H CPUs (aka Lion). */
293#define CPUMMICROARCH_IS_AMD_FAM_11H(a_enmMicroarch) ((a_enmMicroarch) == kCpumMicroarch_AMD_K10_Lion)
294
295/** Predicate macro for catching AMD Family 12H CPUs (aka Llano). */
296#define CPUMMICROARCH_IS_AMD_FAM_12H(a_enmMicroarch) ((a_enmMicroarch) == kCpumMicroarch_AMD_K10_Llano)
297
298/** Predicate macro for catching AMD Family 14H CPUs (aka Bobcat). */
299#define CPUMMICROARCH_IS_AMD_FAM_14H(a_enmMicroarch) ((a_enmMicroarch) == kCpumMicroarch_AMD_Bobcat)
300
301/** Predicate macro for catching AMD Family 15H CPUs (bulldozer and it's
302 * decendants). */
303#define CPUMMICROARCH_IS_AMD_FAM_15H(a_enmMicroarch) \
304 ((a_enmMicroarch) >= kCpumMicroarch_AMD_15h_First && (a_enmMicroarch) <= kCpumMicroarch_AMD_15h_End)
305
306/** Predicate macro for catching AMD Family 16H CPUs. */
307#define CPUMMICROARCH_IS_AMD_FAM_16H(a_enmMicroarch) \
308 ((a_enmMicroarch) >= kCpumMicroarch_AMD_16h_First && (a_enmMicroarch) <= kCpumMicroarch_AMD_16h_End)
309
310
311
312/**
313 * CPUID leaf.
314 *
315 * @remarks This structure is used by the patch manager and is therefore
316 * more or less set in stone.
317 */
318typedef struct CPUMCPUIDLEAF
319{
320 /** The leaf number. */
321 uint32_t uLeaf;
322 /** The sub-leaf number. */
323 uint32_t uSubLeaf;
324 /** Sub-leaf mask. This is 0 when sub-leaves aren't used. */
325 uint32_t fSubLeafMask;
326
327 /** The EAX value. */
328 uint32_t uEax;
329 /** The EBX value. */
330 uint32_t uEbx;
331 /** The ECX value. */
332 uint32_t uEcx;
333 /** The EDX value. */
334 uint32_t uEdx;
335
336 /** Flags. */
337 uint32_t fFlags;
338} CPUMCPUIDLEAF;
339#ifndef VBOX_FOR_DTRACE_LIB
340AssertCompileSize(CPUMCPUIDLEAF, 32);
341#endif
342/** Pointer to a CPUID leaf. */
343typedef CPUMCPUIDLEAF *PCPUMCPUIDLEAF;
344/** Pointer to a const CPUID leaf. */
345typedef CPUMCPUIDLEAF const *PCCPUMCPUIDLEAF;
346
347/** @name CPUMCPUIDLEAF::fFlags
348 * @{ */
349/** Indicates working intel leaf 0xb where the lower 8 ECX bits are not modified
350 * and EDX containing the extended APIC ID. */
351#define CPUMCPUIDLEAF_F_INTEL_TOPOLOGY_SUBLEAVES RT_BIT_32(0)
352/** The leaf contains an APIC ID that needs changing to that of the current CPU. */
353#define CPUMCPUIDLEAF_F_CONTAINS_APIC_ID RT_BIT_32(1)
354/** The leaf contains an OSXSAVE which needs individual handling on each CPU. */
355#define CPUMCPUIDLEAF_F_CONTAINS_OSXSAVE RT_BIT_32(2)
356/** The leaf contains an APIC feature bit which is tied to APICBASE.EN. */
357#define CPUMCPUIDLEAF_F_CONTAINS_APIC RT_BIT_32(3)
358/** Mask of the valid flags. */
359#define CPUMCPUIDLEAF_F_VALID_MASK UINT32_C(0xf)
360/** @} */
361
362/**
363 * Method used to deal with unknown CPUID leaves.
364 * @remarks Used in patch code.
365 */
366typedef enum CPUMUNKNOWNCPUID
367{
368 /** Invalid zero value. */
369 CPUMUNKNOWNCPUID_INVALID = 0,
370 /** Use given default values (DefCpuId). */
371 CPUMUNKNOWNCPUID_DEFAULTS,
372 /** Return the last standard leaf.
373 * Intel Sandy Bridge has been observed doing this. */
374 CPUMUNKNOWNCPUID_LAST_STD_LEAF,
375 /** Return the last standard leaf, with ecx observed.
376 * Intel Sandy Bridge has been observed doing this. */
377 CPUMUNKNOWNCPUID_LAST_STD_LEAF_WITH_ECX,
378 /** The register values are passed thru unmodified. */
379 CPUMUNKNOWNCPUID_PASSTHRU,
380 /** End of valid value. */
381 CPUMUNKNOWNCPUID_END,
382 /** Ensure 32-bit type. */
383 CPUMUNKNOWNCPUID_32BIT_HACK = 0x7fffffff
384} CPUMUNKNOWNCPUID;
385/** Pointer to unknown CPUID leaf method. */
386typedef CPUMUNKNOWNCPUID *PCPUMUNKNOWNCPUID;
387
388
389/**
390 * MSR read functions.
391 */
392typedef enum CPUMMSRRDFN
393{
394 /** Invalid zero value. */
395 kCpumMsrRdFn_Invalid = 0,
396 /** Return the CPUMMSRRANGE::uValue. */
397 kCpumMsrRdFn_FixedValue,
398 /** Alias to the MSR range starting at the MSR given by
399 * CPUMMSRRANGE::uValue. Must be used in pair with
400 * kCpumMsrWrFn_MsrAlias. */
401 kCpumMsrRdFn_MsrAlias,
402 /** Write only register, GP all read attempts. */
403 kCpumMsrRdFn_WriteOnly,
404
405 kCpumMsrRdFn_Ia32P5McAddr,
406 kCpumMsrRdFn_Ia32P5McType,
407 kCpumMsrRdFn_Ia32TimestampCounter,
408 kCpumMsrRdFn_Ia32PlatformId, /**< Takes real CPU value for reference. */
409 kCpumMsrRdFn_Ia32ApicBase,
410 kCpumMsrRdFn_Ia32FeatureControl,
411 kCpumMsrRdFn_Ia32BiosSignId, /**< Range value returned. */
412 kCpumMsrRdFn_Ia32SmmMonitorCtl,
413 kCpumMsrRdFn_Ia32PmcN,
414 kCpumMsrRdFn_Ia32MonitorFilterLineSize,
415 kCpumMsrRdFn_Ia32MPerf,
416 kCpumMsrRdFn_Ia32APerf,
417 kCpumMsrRdFn_Ia32MtrrCap, /**< Takes real CPU value for reference. */
418 kCpumMsrRdFn_Ia32MtrrPhysBaseN, /**< Takes register number. */
419 kCpumMsrRdFn_Ia32MtrrPhysMaskN, /**< Takes register number. */
420 kCpumMsrRdFn_Ia32MtrrFixed, /**< Takes CPUMCPU offset. */
421 kCpumMsrRdFn_Ia32MtrrDefType,
422 kCpumMsrRdFn_Ia32Pat,
423 kCpumMsrRdFn_Ia32SysEnterCs,
424 kCpumMsrRdFn_Ia32SysEnterEsp,
425 kCpumMsrRdFn_Ia32SysEnterEip,
426 kCpumMsrRdFn_Ia32McgCap,
427 kCpumMsrRdFn_Ia32McgStatus,
428 kCpumMsrRdFn_Ia32McgCtl,
429 kCpumMsrRdFn_Ia32DebugCtl,
430 kCpumMsrRdFn_Ia32SmrrPhysBase,
431 kCpumMsrRdFn_Ia32SmrrPhysMask,
432 kCpumMsrRdFn_Ia32PlatformDcaCap,
433 kCpumMsrRdFn_Ia32CpuDcaCap,
434 kCpumMsrRdFn_Ia32Dca0Cap,
435 kCpumMsrRdFn_Ia32PerfEvtSelN, /**< Range value indicates the register number. */
436 kCpumMsrRdFn_Ia32PerfStatus, /**< Range value returned. */
437 kCpumMsrRdFn_Ia32PerfCtl, /**< Range value returned. */
438 kCpumMsrRdFn_Ia32FixedCtrN, /**< Takes register number of start of range. */
439 kCpumMsrRdFn_Ia32PerfCapabilities, /**< Takes reference value. */
440 kCpumMsrRdFn_Ia32FixedCtrCtrl,
441 kCpumMsrRdFn_Ia32PerfGlobalStatus, /**< Takes reference value. */
442 kCpumMsrRdFn_Ia32PerfGlobalCtrl,
443 kCpumMsrRdFn_Ia32PerfGlobalOvfCtrl,
444 kCpumMsrRdFn_Ia32PebsEnable,
445 kCpumMsrRdFn_Ia32ClockModulation, /**< Range value returned. */
446 kCpumMsrRdFn_Ia32ThermInterrupt, /**< Range value returned. */
447 kCpumMsrRdFn_Ia32ThermStatus, /**< Range value returned. */
448 kCpumMsrRdFn_Ia32Therm2Ctl, /**< Range value returned. */
449 kCpumMsrRdFn_Ia32MiscEnable, /**< Range value returned. */
450 kCpumMsrRdFn_Ia32McCtlStatusAddrMiscN, /**< Takes bank number. */
451 kCpumMsrRdFn_Ia32McNCtl2, /**< Takes register number of start of range. */
452 kCpumMsrRdFn_Ia32DsArea,
453 kCpumMsrRdFn_Ia32TscDeadline,
454 kCpumMsrRdFn_Ia32X2ApicN,
455 kCpumMsrRdFn_Ia32DebugInterface,
456 kCpumMsrRdFn_Ia32VmxBase, /**< Takes real value as reference. */
457 kCpumMsrRdFn_Ia32VmxPinbasedCtls, /**< Takes real value as reference. */
458 kCpumMsrRdFn_Ia32VmxProcbasedCtls, /**< Takes real value as reference. */
459 kCpumMsrRdFn_Ia32VmxExitCtls, /**< Takes real value as reference. */
460 kCpumMsrRdFn_Ia32VmxEntryCtls, /**< Takes real value as reference. */
461 kCpumMsrRdFn_Ia32VmxMisc, /**< Takes real value as reference. */
462 kCpumMsrRdFn_Ia32VmxCr0Fixed0, /**< Takes real value as reference. */
463 kCpumMsrRdFn_Ia32VmxCr0Fixed1, /**< Takes real value as reference. */
464 kCpumMsrRdFn_Ia32VmxCr4Fixed0, /**< Takes real value as reference. */
465 kCpumMsrRdFn_Ia32VmxCr4Fixed1, /**< Takes real value as reference. */
466 kCpumMsrRdFn_Ia32VmxVmcsEnum, /**< Takes real value as reference. */
467 kCpumMsrRdFn_Ia32VmxProcBasedCtls2, /**< Takes real value as reference. */
468 kCpumMsrRdFn_Ia32VmxEptVpidCap, /**< Takes real value as reference. */
469 kCpumMsrRdFn_Ia32VmxTruePinbasedCtls, /**< Takes real value as reference. */
470 kCpumMsrRdFn_Ia32VmxTrueProcbasedCtls, /**< Takes real value as reference. */
471 kCpumMsrRdFn_Ia32VmxTrueExitCtls, /**< Takes real value as reference. */
472 kCpumMsrRdFn_Ia32VmxTrueEntryCtls, /**< Takes real value as reference. */
473 kCpumMsrRdFn_Ia32VmxVmFunc, /**< Takes real value as reference. */
474 kCpumMsrRdFn_Ia32SpecCtrl,
475 kCpumMsrRdFn_Ia32ArchCapabilities,
476
477 kCpumMsrRdFn_Amd64Efer,
478 kCpumMsrRdFn_Amd64SyscallTarget,
479 kCpumMsrRdFn_Amd64LongSyscallTarget,
480 kCpumMsrRdFn_Amd64CompSyscallTarget,
481 kCpumMsrRdFn_Amd64SyscallFlagMask,
482 kCpumMsrRdFn_Amd64FsBase,
483 kCpumMsrRdFn_Amd64GsBase,
484 kCpumMsrRdFn_Amd64KernelGsBase,
485 kCpumMsrRdFn_Amd64TscAux,
486
487 kCpumMsrRdFn_IntelEblCrPowerOn,
488 kCpumMsrRdFn_IntelI7CoreThreadCount,
489 kCpumMsrRdFn_IntelP4EbcHardPowerOn,
490 kCpumMsrRdFn_IntelP4EbcSoftPowerOn,
491 kCpumMsrRdFn_IntelP4EbcFrequencyId,
492 kCpumMsrRdFn_IntelP6FsbFrequency, /**< Takes real value as reference. */
493 kCpumMsrRdFn_IntelPlatformInfo,
494 kCpumMsrRdFn_IntelFlexRatio, /**< Takes real value as reference. */
495 kCpumMsrRdFn_IntelPkgCStConfigControl,
496 kCpumMsrRdFn_IntelPmgIoCaptureBase,
497 kCpumMsrRdFn_IntelLastBranchFromToN,
498 kCpumMsrRdFn_IntelLastBranchFromN,
499 kCpumMsrRdFn_IntelLastBranchToN,
500 kCpumMsrRdFn_IntelLastBranchTos,
501 kCpumMsrRdFn_IntelBblCrCtl,
502 kCpumMsrRdFn_IntelBblCrCtl3,
503 kCpumMsrRdFn_IntelI7TemperatureTarget, /**< Range value returned. */
504 kCpumMsrRdFn_IntelI7MsrOffCoreResponseN,/**< Takes register number. */
505 kCpumMsrRdFn_IntelI7MiscPwrMgmt,
506 kCpumMsrRdFn_IntelP6CrN,
507 kCpumMsrRdFn_IntelCpuId1FeatureMaskEcdx,
508 kCpumMsrRdFn_IntelCpuId1FeatureMaskEax,
509 kCpumMsrRdFn_IntelCpuId80000001FeatureMaskEcdx,
510 kCpumMsrRdFn_IntelI7SandyAesNiCtl,
511 kCpumMsrRdFn_IntelI7TurboRatioLimit, /**< Returns range value. */
512 kCpumMsrRdFn_IntelI7LbrSelect,
513 kCpumMsrRdFn_IntelI7SandyErrorControl,
514 kCpumMsrRdFn_IntelI7VirtualLegacyWireCap,/**< Returns range value. */
515 kCpumMsrRdFn_IntelI7PowerCtl,
516 kCpumMsrRdFn_IntelI7SandyPebsNumAlt,
517 kCpumMsrRdFn_IntelI7PebsLdLat,
518 kCpumMsrRdFn_IntelI7PkgCnResidencyN, /**< Takes C-state number. */
519 kCpumMsrRdFn_IntelI7CoreCnResidencyN, /**< Takes C-state number. */
520 kCpumMsrRdFn_IntelI7SandyVrCurrentConfig,/**< Takes real value as reference. */
521 kCpumMsrRdFn_IntelI7SandyVrMiscConfig, /**< Takes real value as reference. */
522 kCpumMsrRdFn_IntelI7SandyRaplPowerUnit, /**< Takes real value as reference. */
523 kCpumMsrRdFn_IntelI7SandyPkgCnIrtlN, /**< Takes real value as reference. */
524 kCpumMsrRdFn_IntelI7SandyPkgC2Residency, /**< Takes real value as reference. */
525 kCpumMsrRdFn_IntelI7RaplPkgPowerLimit, /**< Takes real value as reference. */
526 kCpumMsrRdFn_IntelI7RaplPkgEnergyStatus, /**< Takes real value as reference. */
527 kCpumMsrRdFn_IntelI7RaplPkgPerfStatus, /**< Takes real value as reference. */
528 kCpumMsrRdFn_IntelI7RaplPkgPowerInfo, /**< Takes real value as reference. */
529 kCpumMsrRdFn_IntelI7RaplDramPowerLimit, /**< Takes real value as reference. */
530 kCpumMsrRdFn_IntelI7RaplDramEnergyStatus,/**< Takes real value as reference. */
531 kCpumMsrRdFn_IntelI7RaplDramPerfStatus, /**< Takes real value as reference. */
532 kCpumMsrRdFn_IntelI7RaplDramPowerInfo, /**< Takes real value as reference. */
533 kCpumMsrRdFn_IntelI7RaplPp0PowerLimit, /**< Takes real value as reference. */
534 kCpumMsrRdFn_IntelI7RaplPp0EnergyStatus, /**< Takes real value as reference. */
535 kCpumMsrRdFn_IntelI7RaplPp0Policy, /**< Takes real value as reference. */
536 kCpumMsrRdFn_IntelI7RaplPp0PerfStatus, /**< Takes real value as reference. */
537 kCpumMsrRdFn_IntelI7RaplPp1PowerLimit, /**< Takes real value as reference. */
538 kCpumMsrRdFn_IntelI7RaplPp1EnergyStatus, /**< Takes real value as reference. */
539 kCpumMsrRdFn_IntelI7RaplPp1Policy, /**< Takes real value as reference. */
540 kCpumMsrRdFn_IntelI7IvyConfigTdpNominal, /**< Takes real value as reference. */
541 kCpumMsrRdFn_IntelI7IvyConfigTdpLevel1, /**< Takes real value as reference. */
542 kCpumMsrRdFn_IntelI7IvyConfigTdpLevel2, /**< Takes real value as reference. */
543 kCpumMsrRdFn_IntelI7IvyConfigTdpControl,
544 kCpumMsrRdFn_IntelI7IvyTurboActivationRatio,
545 kCpumMsrRdFn_IntelI7UncPerfGlobalCtrl,
546 kCpumMsrRdFn_IntelI7UncPerfGlobalStatus,
547 kCpumMsrRdFn_IntelI7UncPerfGlobalOvfCtrl,
548 kCpumMsrRdFn_IntelI7UncPerfFixedCtrCtrl,
549 kCpumMsrRdFn_IntelI7UncPerfFixedCtr,
550 kCpumMsrRdFn_IntelI7UncCBoxConfig,
551 kCpumMsrRdFn_IntelI7UncArbPerfCtrN,
552 kCpumMsrRdFn_IntelI7UncArbPerfEvtSelN,
553 kCpumMsrRdFn_IntelI7SmiCount,
554 kCpumMsrRdFn_IntelCore2EmttmCrTablesN, /**< Range value returned. */
555 kCpumMsrRdFn_IntelCore2SmmCStMiscInfo,
556 kCpumMsrRdFn_IntelCore1ExtConfig,
557 kCpumMsrRdFn_IntelCore1DtsCalControl,
558 kCpumMsrRdFn_IntelCore2PeciControl,
559 kCpumMsrRdFn_IntelAtSilvCoreC1Recidency,
560
561 kCpumMsrRdFn_P6LastBranchFromIp,
562 kCpumMsrRdFn_P6LastBranchToIp,
563 kCpumMsrRdFn_P6LastIntFromIp,
564 kCpumMsrRdFn_P6LastIntToIp,
565
566 kCpumMsrRdFn_AmdFam15hTscRate,
567 kCpumMsrRdFn_AmdFam15hLwpCfg,
568 kCpumMsrRdFn_AmdFam15hLwpCbAddr,
569 kCpumMsrRdFn_AmdFam10hMc4MiscN,
570 kCpumMsrRdFn_AmdK8PerfCtlN,
571 kCpumMsrRdFn_AmdK8PerfCtrN,
572 kCpumMsrRdFn_AmdK8SysCfg, /**< Range value returned. */
573 kCpumMsrRdFn_AmdK8HwCr,
574 kCpumMsrRdFn_AmdK8IorrBaseN,
575 kCpumMsrRdFn_AmdK8IorrMaskN,
576 kCpumMsrRdFn_AmdK8TopOfMemN,
577 kCpumMsrRdFn_AmdK8NbCfg1,
578 kCpumMsrRdFn_AmdK8McXcptRedir,
579 kCpumMsrRdFn_AmdK8CpuNameN,
580 kCpumMsrRdFn_AmdK8HwThermalCtrl, /**< Range value returned. */
581 kCpumMsrRdFn_AmdK8SwThermalCtrl,
582 kCpumMsrRdFn_AmdK8FidVidControl, /**< Range value returned. */
583 kCpumMsrRdFn_AmdK8FidVidStatus, /**< Range value returned. */
584 kCpumMsrRdFn_AmdK8McCtlMaskN,
585 kCpumMsrRdFn_AmdK8SmiOnIoTrapN,
586 kCpumMsrRdFn_AmdK8SmiOnIoTrapCtlSts,
587 kCpumMsrRdFn_AmdK8IntPendingMessage,
588 kCpumMsrRdFn_AmdK8SmiTriggerIoCycle,
589 kCpumMsrRdFn_AmdFam10hMmioCfgBaseAddr,
590 kCpumMsrRdFn_AmdFam10hTrapCtlMaybe,
591 kCpumMsrRdFn_AmdFam10hPStateCurLimit, /**< Returns range value. */
592 kCpumMsrRdFn_AmdFam10hPStateControl, /**< Returns range value. */
593 kCpumMsrRdFn_AmdFam10hPStateStatus, /**< Returns range value. */
594 kCpumMsrRdFn_AmdFam10hPStateN, /**< Returns range value. This isn't an register index! */
595 kCpumMsrRdFn_AmdFam10hCofVidControl, /**< Returns range value. */
596 kCpumMsrRdFn_AmdFam10hCofVidStatus, /**< Returns range value. */
597 kCpumMsrRdFn_AmdFam10hCStateIoBaseAddr,
598 kCpumMsrRdFn_AmdFam10hCpuWatchdogTimer,
599 kCpumMsrRdFn_AmdK8SmmBase,
600 kCpumMsrRdFn_AmdK8SmmAddr,
601 kCpumMsrRdFn_AmdK8SmmMask,
602 kCpumMsrRdFn_AmdK8VmCr,
603 kCpumMsrRdFn_AmdK8IgnNe,
604 kCpumMsrRdFn_AmdK8SmmCtl,
605 kCpumMsrRdFn_AmdK8VmHSavePa,
606 kCpumMsrRdFn_AmdFam10hVmLockKey,
607 kCpumMsrRdFn_AmdFam10hSmmLockKey,
608 kCpumMsrRdFn_AmdFam10hLocalSmiStatus,
609 kCpumMsrRdFn_AmdFam10hOsVisWrkIdLength,
610 kCpumMsrRdFn_AmdFam10hOsVisWrkStatus,
611 kCpumMsrRdFn_AmdFam16hL2IPerfCtlN,
612 kCpumMsrRdFn_AmdFam16hL2IPerfCtrN,
613 kCpumMsrRdFn_AmdFam15hNorthbridgePerfCtlN,
614 kCpumMsrRdFn_AmdFam15hNorthbridgePerfCtrN,
615 kCpumMsrRdFn_AmdK7MicrocodeCtl, /**< Returns range value. */
616 kCpumMsrRdFn_AmdK7ClusterIdMaybe, /**< Returns range value. */
617 kCpumMsrRdFn_AmdK8CpuIdCtlStd07hEbax,
618 kCpumMsrRdFn_AmdK8CpuIdCtlStd06hEcx,
619 kCpumMsrRdFn_AmdK8CpuIdCtlStd01hEdcx,
620 kCpumMsrRdFn_AmdK8CpuIdCtlExt01hEdcx,
621 kCpumMsrRdFn_AmdK8PatchLevel, /**< Returns range value. */
622 kCpumMsrRdFn_AmdK7DebugStatusMaybe,
623 kCpumMsrRdFn_AmdK7BHTraceBaseMaybe,
624 kCpumMsrRdFn_AmdK7BHTracePtrMaybe,
625 kCpumMsrRdFn_AmdK7BHTraceLimitMaybe,
626 kCpumMsrRdFn_AmdK7HardwareDebugToolCfgMaybe,
627 kCpumMsrRdFn_AmdK7FastFlushCountMaybe,
628 kCpumMsrRdFn_AmdK7NodeId,
629 kCpumMsrRdFn_AmdK7DrXAddrMaskN, /**< Takes register index. */
630 kCpumMsrRdFn_AmdK7Dr0DataMatchMaybe,
631 kCpumMsrRdFn_AmdK7Dr0DataMaskMaybe,
632 kCpumMsrRdFn_AmdK7LoadStoreCfg,
633 kCpumMsrRdFn_AmdK7InstrCacheCfg,
634 kCpumMsrRdFn_AmdK7DataCacheCfg,
635 kCpumMsrRdFn_AmdK7BusUnitCfg,
636 kCpumMsrRdFn_AmdK7DebugCtl2Maybe,
637 kCpumMsrRdFn_AmdFam15hFpuCfg,
638 kCpumMsrRdFn_AmdFam15hDecoderCfg,
639 kCpumMsrRdFn_AmdFam10hBusUnitCfg2,
640 kCpumMsrRdFn_AmdFam15hCombUnitCfg,
641 kCpumMsrRdFn_AmdFam15hCombUnitCfg2,
642 kCpumMsrRdFn_AmdFam15hCombUnitCfg3,
643 kCpumMsrRdFn_AmdFam15hExecUnitCfg,
644 kCpumMsrRdFn_AmdFam15hLoadStoreCfg2,
645 kCpumMsrRdFn_AmdFam10hIbsFetchCtl,
646 kCpumMsrRdFn_AmdFam10hIbsFetchLinAddr,
647 kCpumMsrRdFn_AmdFam10hIbsFetchPhysAddr,
648 kCpumMsrRdFn_AmdFam10hIbsOpExecCtl,
649 kCpumMsrRdFn_AmdFam10hIbsOpRip,
650 kCpumMsrRdFn_AmdFam10hIbsOpData,
651 kCpumMsrRdFn_AmdFam10hIbsOpData2,
652 kCpumMsrRdFn_AmdFam10hIbsOpData3,
653 kCpumMsrRdFn_AmdFam10hIbsDcLinAddr,
654 kCpumMsrRdFn_AmdFam10hIbsDcPhysAddr,
655 kCpumMsrRdFn_AmdFam10hIbsCtl,
656 kCpumMsrRdFn_AmdFam14hIbsBrTarget,
657
658 kCpumMsrRdFn_Gim,
659
660 /** End of valid MSR read function indexes. */
661 kCpumMsrRdFn_End
662} CPUMMSRRDFN;
663
664/**
665 * MSR write functions.
666 */
667typedef enum CPUMMSRWRFN
668{
669 /** Invalid zero value. */
670 kCpumMsrWrFn_Invalid = 0,
671 /** Writes are ignored, the fWrGpMask is observed though. */
672 kCpumMsrWrFn_IgnoreWrite,
673 /** Writes cause GP(0) to be raised, the fWrGpMask should be UINT64_MAX. */
674 kCpumMsrWrFn_ReadOnly,
675 /** Alias to the MSR range starting at the MSR given by
676 * CPUMMSRRANGE::uValue. Must be used in pair with
677 * kCpumMsrRdFn_MsrAlias. */
678 kCpumMsrWrFn_MsrAlias,
679
680 kCpumMsrWrFn_Ia32P5McAddr,
681 kCpumMsrWrFn_Ia32P5McType,
682 kCpumMsrWrFn_Ia32TimestampCounter,
683 kCpumMsrWrFn_Ia32ApicBase,
684 kCpumMsrWrFn_Ia32FeatureControl,
685 kCpumMsrWrFn_Ia32BiosSignId,
686 kCpumMsrWrFn_Ia32BiosUpdateTrigger,
687 kCpumMsrWrFn_Ia32SmmMonitorCtl,
688 kCpumMsrWrFn_Ia32PmcN,
689 kCpumMsrWrFn_Ia32MonitorFilterLineSize,
690 kCpumMsrWrFn_Ia32MPerf,
691 kCpumMsrWrFn_Ia32APerf,
692 kCpumMsrWrFn_Ia32MtrrPhysBaseN, /**< Takes register number. */
693 kCpumMsrWrFn_Ia32MtrrPhysMaskN, /**< Takes register number. */
694 kCpumMsrWrFn_Ia32MtrrFixed, /**< Takes CPUMCPU offset. */
695 kCpumMsrWrFn_Ia32MtrrDefType,
696 kCpumMsrWrFn_Ia32Pat,
697 kCpumMsrWrFn_Ia32SysEnterCs,
698 kCpumMsrWrFn_Ia32SysEnterEsp,
699 kCpumMsrWrFn_Ia32SysEnterEip,
700 kCpumMsrWrFn_Ia32McgStatus,
701 kCpumMsrWrFn_Ia32McgCtl,
702 kCpumMsrWrFn_Ia32DebugCtl,
703 kCpumMsrWrFn_Ia32SmrrPhysBase,
704 kCpumMsrWrFn_Ia32SmrrPhysMask,
705 kCpumMsrWrFn_Ia32PlatformDcaCap,
706 kCpumMsrWrFn_Ia32Dca0Cap,
707 kCpumMsrWrFn_Ia32PerfEvtSelN, /**< Range value indicates the register number. */
708 kCpumMsrWrFn_Ia32PerfStatus,
709 kCpumMsrWrFn_Ia32PerfCtl,
710 kCpumMsrWrFn_Ia32FixedCtrN, /**< Takes register number of start of range. */
711 kCpumMsrWrFn_Ia32PerfCapabilities,
712 kCpumMsrWrFn_Ia32FixedCtrCtrl,
713 kCpumMsrWrFn_Ia32PerfGlobalStatus,
714 kCpumMsrWrFn_Ia32PerfGlobalCtrl,
715 kCpumMsrWrFn_Ia32PerfGlobalOvfCtrl,
716 kCpumMsrWrFn_Ia32PebsEnable,
717 kCpumMsrWrFn_Ia32ClockModulation,
718 kCpumMsrWrFn_Ia32ThermInterrupt,
719 kCpumMsrWrFn_Ia32ThermStatus,
720 kCpumMsrWrFn_Ia32Therm2Ctl,
721 kCpumMsrWrFn_Ia32MiscEnable,
722 kCpumMsrWrFn_Ia32McCtlStatusAddrMiscN, /**< Takes bank number. */
723 kCpumMsrWrFn_Ia32McNCtl2, /**< Takes register number of start of range. */
724 kCpumMsrWrFn_Ia32DsArea,
725 kCpumMsrWrFn_Ia32TscDeadline,
726 kCpumMsrWrFn_Ia32X2ApicN,
727 kCpumMsrWrFn_Ia32DebugInterface,
728 kCpumMsrWrFn_Ia32SpecCtrl,
729 kCpumMsrWrFn_Ia32PredCmd,
730
731 kCpumMsrWrFn_Amd64Efer,
732 kCpumMsrWrFn_Amd64SyscallTarget,
733 kCpumMsrWrFn_Amd64LongSyscallTarget,
734 kCpumMsrWrFn_Amd64CompSyscallTarget,
735 kCpumMsrWrFn_Amd64SyscallFlagMask,
736 kCpumMsrWrFn_Amd64FsBase,
737 kCpumMsrWrFn_Amd64GsBase,
738 kCpumMsrWrFn_Amd64KernelGsBase,
739 kCpumMsrWrFn_Amd64TscAux,
740 kCpumMsrWrFn_IntelEblCrPowerOn,
741 kCpumMsrWrFn_IntelP4EbcHardPowerOn,
742 kCpumMsrWrFn_IntelP4EbcSoftPowerOn,
743 kCpumMsrWrFn_IntelP4EbcFrequencyId,
744 kCpumMsrWrFn_IntelFlexRatio,
745 kCpumMsrWrFn_IntelPkgCStConfigControl,
746 kCpumMsrWrFn_IntelPmgIoCaptureBase,
747 kCpumMsrWrFn_IntelLastBranchFromToN,
748 kCpumMsrWrFn_IntelLastBranchFromN,
749 kCpumMsrWrFn_IntelLastBranchToN,
750 kCpumMsrWrFn_IntelLastBranchTos,
751 kCpumMsrWrFn_IntelBblCrCtl,
752 kCpumMsrWrFn_IntelBblCrCtl3,
753 kCpumMsrWrFn_IntelI7TemperatureTarget,
754 kCpumMsrWrFn_IntelI7MsrOffCoreResponseN, /**< Takes register number. */
755 kCpumMsrWrFn_IntelI7MiscPwrMgmt,
756 kCpumMsrWrFn_IntelP6CrN,
757 kCpumMsrWrFn_IntelCpuId1FeatureMaskEcdx,
758 kCpumMsrWrFn_IntelCpuId1FeatureMaskEax,
759 kCpumMsrWrFn_IntelCpuId80000001FeatureMaskEcdx,
760 kCpumMsrWrFn_IntelI7SandyAesNiCtl,
761 kCpumMsrWrFn_IntelI7TurboRatioLimit,
762 kCpumMsrWrFn_IntelI7LbrSelect,
763 kCpumMsrWrFn_IntelI7SandyErrorControl,
764 kCpumMsrWrFn_IntelI7PowerCtl,
765 kCpumMsrWrFn_IntelI7SandyPebsNumAlt,
766 kCpumMsrWrFn_IntelI7PebsLdLat,
767 kCpumMsrWrFn_IntelI7SandyVrCurrentConfig,
768 kCpumMsrWrFn_IntelI7SandyVrMiscConfig,
769 kCpumMsrWrFn_IntelI7SandyRaplPowerUnit, /**< R/O but found writable bits on a Silvermont CPU here. */
770 kCpumMsrWrFn_IntelI7SandyPkgCnIrtlN,
771 kCpumMsrWrFn_IntelI7SandyPkgC2Residency, /**< R/O but found writable bits on a Silvermont CPU here. */
772 kCpumMsrWrFn_IntelI7RaplPkgPowerLimit,
773 kCpumMsrWrFn_IntelI7RaplDramPowerLimit,
774 kCpumMsrWrFn_IntelI7RaplPp0PowerLimit,
775 kCpumMsrWrFn_IntelI7RaplPp0Policy,
776 kCpumMsrWrFn_IntelI7RaplPp1PowerLimit,
777 kCpumMsrWrFn_IntelI7RaplPp1Policy,
778 kCpumMsrWrFn_IntelI7IvyConfigTdpControl,
779 kCpumMsrWrFn_IntelI7IvyTurboActivationRatio,
780 kCpumMsrWrFn_IntelI7UncPerfGlobalCtrl,
781 kCpumMsrWrFn_IntelI7UncPerfGlobalStatus,
782 kCpumMsrWrFn_IntelI7UncPerfGlobalOvfCtrl,
783 kCpumMsrWrFn_IntelI7UncPerfFixedCtrCtrl,
784 kCpumMsrWrFn_IntelI7UncPerfFixedCtr,
785 kCpumMsrWrFn_IntelI7UncArbPerfCtrN,
786 kCpumMsrWrFn_IntelI7UncArbPerfEvtSelN,
787 kCpumMsrWrFn_IntelCore2EmttmCrTablesN,
788 kCpumMsrWrFn_IntelCore2SmmCStMiscInfo,
789 kCpumMsrWrFn_IntelCore1ExtConfig,
790 kCpumMsrWrFn_IntelCore1DtsCalControl,
791 kCpumMsrWrFn_IntelCore2PeciControl,
792
793 kCpumMsrWrFn_P6LastIntFromIp,
794 kCpumMsrWrFn_P6LastIntToIp,
795
796 kCpumMsrWrFn_AmdFam15hTscRate,
797 kCpumMsrWrFn_AmdFam15hLwpCfg,
798 kCpumMsrWrFn_AmdFam15hLwpCbAddr,
799 kCpumMsrWrFn_AmdFam10hMc4MiscN,
800 kCpumMsrWrFn_AmdK8PerfCtlN,
801 kCpumMsrWrFn_AmdK8PerfCtrN,
802 kCpumMsrWrFn_AmdK8SysCfg,
803 kCpumMsrWrFn_AmdK8HwCr,
804 kCpumMsrWrFn_AmdK8IorrBaseN,
805 kCpumMsrWrFn_AmdK8IorrMaskN,
806 kCpumMsrWrFn_AmdK8TopOfMemN,
807 kCpumMsrWrFn_AmdK8NbCfg1,
808 kCpumMsrWrFn_AmdK8McXcptRedir,
809 kCpumMsrWrFn_AmdK8CpuNameN,
810 kCpumMsrWrFn_AmdK8HwThermalCtrl,
811 kCpumMsrWrFn_AmdK8SwThermalCtrl,
812 kCpumMsrWrFn_AmdK8FidVidControl,
813 kCpumMsrWrFn_AmdK8McCtlMaskN,
814 kCpumMsrWrFn_AmdK8SmiOnIoTrapN,
815 kCpumMsrWrFn_AmdK8SmiOnIoTrapCtlSts,
816 kCpumMsrWrFn_AmdK8IntPendingMessage,
817 kCpumMsrWrFn_AmdK8SmiTriggerIoCycle,
818 kCpumMsrWrFn_AmdFam10hMmioCfgBaseAddr,
819 kCpumMsrWrFn_AmdFam10hTrapCtlMaybe,
820 kCpumMsrWrFn_AmdFam10hPStateControl,
821 kCpumMsrWrFn_AmdFam10hPStateStatus,
822 kCpumMsrWrFn_AmdFam10hPStateN,
823 kCpumMsrWrFn_AmdFam10hCofVidControl,
824 kCpumMsrWrFn_AmdFam10hCofVidStatus,
825 kCpumMsrWrFn_AmdFam10hCStateIoBaseAddr,
826 kCpumMsrWrFn_AmdFam10hCpuWatchdogTimer,
827 kCpumMsrWrFn_AmdK8SmmBase,
828 kCpumMsrWrFn_AmdK8SmmAddr,
829 kCpumMsrWrFn_AmdK8SmmMask,
830 kCpumMsrWrFn_AmdK8VmCr,
831 kCpumMsrWrFn_AmdK8IgnNe,
832 kCpumMsrWrFn_AmdK8SmmCtl,
833 kCpumMsrWrFn_AmdK8VmHSavePa,
834 kCpumMsrWrFn_AmdFam10hVmLockKey,
835 kCpumMsrWrFn_AmdFam10hSmmLockKey,
836 kCpumMsrWrFn_AmdFam10hLocalSmiStatus,
837 kCpumMsrWrFn_AmdFam10hOsVisWrkIdLength,
838 kCpumMsrWrFn_AmdFam10hOsVisWrkStatus,
839 kCpumMsrWrFn_AmdFam16hL2IPerfCtlN,
840 kCpumMsrWrFn_AmdFam16hL2IPerfCtrN,
841 kCpumMsrWrFn_AmdFam15hNorthbridgePerfCtlN,
842 kCpumMsrWrFn_AmdFam15hNorthbridgePerfCtrN,
843 kCpumMsrWrFn_AmdK7MicrocodeCtl,
844 kCpumMsrWrFn_AmdK7ClusterIdMaybe,
845 kCpumMsrWrFn_AmdK8CpuIdCtlStd07hEbax,
846 kCpumMsrWrFn_AmdK8CpuIdCtlStd06hEcx,
847 kCpumMsrWrFn_AmdK8CpuIdCtlStd01hEdcx,
848 kCpumMsrWrFn_AmdK8CpuIdCtlExt01hEdcx,
849 kCpumMsrWrFn_AmdK8PatchLoader,
850 kCpumMsrWrFn_AmdK7DebugStatusMaybe,
851 kCpumMsrWrFn_AmdK7BHTraceBaseMaybe,
852 kCpumMsrWrFn_AmdK7BHTracePtrMaybe,
853 kCpumMsrWrFn_AmdK7BHTraceLimitMaybe,
854 kCpumMsrWrFn_AmdK7HardwareDebugToolCfgMaybe,
855 kCpumMsrWrFn_AmdK7FastFlushCountMaybe,
856 kCpumMsrWrFn_AmdK7NodeId,
857 kCpumMsrWrFn_AmdK7DrXAddrMaskN, /**< Takes register index. */
858 kCpumMsrWrFn_AmdK7Dr0DataMatchMaybe,
859 kCpumMsrWrFn_AmdK7Dr0DataMaskMaybe,
860 kCpumMsrWrFn_AmdK7LoadStoreCfg,
861 kCpumMsrWrFn_AmdK7InstrCacheCfg,
862 kCpumMsrWrFn_AmdK7DataCacheCfg,
863 kCpumMsrWrFn_AmdK7BusUnitCfg,
864 kCpumMsrWrFn_AmdK7DebugCtl2Maybe,
865 kCpumMsrWrFn_AmdFam15hFpuCfg,
866 kCpumMsrWrFn_AmdFam15hDecoderCfg,
867 kCpumMsrWrFn_AmdFam10hBusUnitCfg2,
868 kCpumMsrWrFn_AmdFam15hCombUnitCfg,
869 kCpumMsrWrFn_AmdFam15hCombUnitCfg2,
870 kCpumMsrWrFn_AmdFam15hCombUnitCfg3,
871 kCpumMsrWrFn_AmdFam15hExecUnitCfg,
872 kCpumMsrWrFn_AmdFam15hLoadStoreCfg2,
873 kCpumMsrWrFn_AmdFam10hIbsFetchCtl,
874 kCpumMsrWrFn_AmdFam10hIbsFetchLinAddr,
875 kCpumMsrWrFn_AmdFam10hIbsFetchPhysAddr,
876 kCpumMsrWrFn_AmdFam10hIbsOpExecCtl,
877 kCpumMsrWrFn_AmdFam10hIbsOpRip,
878 kCpumMsrWrFn_AmdFam10hIbsOpData,
879 kCpumMsrWrFn_AmdFam10hIbsOpData2,
880 kCpumMsrWrFn_AmdFam10hIbsOpData3,
881 kCpumMsrWrFn_AmdFam10hIbsDcLinAddr,
882 kCpumMsrWrFn_AmdFam10hIbsDcPhysAddr,
883 kCpumMsrWrFn_AmdFam10hIbsCtl,
884 kCpumMsrWrFn_AmdFam14hIbsBrTarget,
885
886 kCpumMsrWrFn_Gim,
887
888 /** End of valid MSR write function indexes. */
889 kCpumMsrWrFn_End
890} CPUMMSRWRFN;
891
892/**
893 * MSR range.
894 */
895typedef struct CPUMMSRRANGE
896{
897 /** The first MSR. [0] */
898 uint32_t uFirst;
899 /** The last MSR. [4] */
900 uint32_t uLast;
901 /** The read function (CPUMMSRRDFN). [8] */
902 uint16_t enmRdFn;
903 /** The write function (CPUMMSRWRFN). [10] */
904 uint16_t enmWrFn;
905 /** The offset of the 64-bit MSR value relative to the start of CPUMCPU.
906 * UINT16_MAX if not used by the read and write functions. [12] */
907 uint16_t offCpumCpu;
908 /** Reserved for future hacks. [14] */
909 uint16_t fReserved;
910 /** The init/read value. [16]
911 * When enmRdFn is kCpumMsrRdFn_INIT_VALUE, this is the value returned on RDMSR.
912 * offCpumCpu must be UINT16_MAX in that case, otherwise it must be a valid
913 * offset into CPUM. */
914 uint64_t uValue;
915 /** The bits to ignore when writing. [24] */
916 uint64_t fWrIgnMask;
917 /** The bits that will cause a GP(0) when writing. [32]
918 * This is always checked prior to calling the write function. Using
919 * UINT64_MAX effectively marks the MSR as read-only. */
920 uint64_t fWrGpMask;
921 /** The register name, if applicable. [40] */
922 char szName[56];
923
924#ifdef VBOX_WITH_STATISTICS
925 /** The number of reads. */
926 STAMCOUNTER cReads;
927 /** The number of writes. */
928 STAMCOUNTER cWrites;
929 /** The number of times ignored bits were written. */
930 STAMCOUNTER cIgnoredBits;
931 /** The number of GPs generated. */
932 STAMCOUNTER cGps;
933#endif
934} CPUMMSRRANGE;
935#ifndef VBOX_FOR_DTRACE_LIB
936# ifdef VBOX_WITH_STATISTICS
937AssertCompileSize(CPUMMSRRANGE, 128);
938# else
939AssertCompileSize(CPUMMSRRANGE, 96);
940# endif
941#endif
942/** Pointer to an MSR range. */
943typedef CPUMMSRRANGE *PCPUMMSRRANGE;
944/** Pointer to a const MSR range. */
945typedef CPUMMSRRANGE const *PCCPUMMSRRANGE;
946
947
948/**
949 * CPU features and quirks.
950 * This is mostly exploded CPUID info.
951 */
952typedef struct CPUMFEATURES
953{
954 /** The CPU vendor (CPUMCPUVENDOR). */
955 uint8_t enmCpuVendor;
956 /** The CPU family. */
957 uint8_t uFamily;
958 /** The CPU model. */
959 uint8_t uModel;
960 /** The CPU stepping. */
961 uint8_t uStepping;
962 /** The microarchitecture. */
963#ifndef VBOX_FOR_DTRACE_LIB
964 CPUMMICROARCH enmMicroarch;
965#else
966 uint32_t enmMicroarch;
967#endif
968 /** The maximum physical address with of the CPU. */
969 uint8_t cMaxPhysAddrWidth;
970 /** Alignment padding. */
971 uint8_t abPadding[1];
972 /** Max size of the extended state (or FPU state if no XSAVE). */
973 uint16_t cbMaxExtendedState;
974
975 /** Supports MSRs. */
976 uint32_t fMsr : 1;
977 /** Supports the page size extension (4/2 MB pages). */
978 uint32_t fPse : 1;
979 /** Supports 36-bit page size extension (4 MB pages can map memory above
980 * 4GB). */
981 uint32_t fPse36 : 1;
982 /** Supports physical address extension (PAE). */
983 uint32_t fPae : 1;
984 /** Page attribute table (PAT) support (page level cache control). */
985 uint32_t fPat : 1;
986 /** Supports the FXSAVE and FXRSTOR instructions. */
987 uint32_t fFxSaveRstor : 1;
988 /** Supports the XSAVE and XRSTOR instructions. */
989 uint32_t fXSaveRstor : 1;
990 /** The XSAVE/XRSTOR bit in CR4 has been set (only applicable for host!). */
991 uint32_t fOpSysXSaveRstor : 1;
992 /** Supports MMX. */
993 uint32_t fMmx : 1;
994 /** Supports AMD extensions to MMX instructions. */
995 uint32_t fAmdMmxExts : 1;
996 /** Supports SSE. */
997 uint32_t fSse : 1;
998 /** Supports SSE2. */
999 uint32_t fSse2 : 1;
1000 /** Supports SSE3. */
1001 uint32_t fSse3 : 1;
1002 /** Supports SSSE3. */
1003 uint32_t fSsse3 : 1;
1004 /** Supports SSE4.1. */
1005 uint32_t fSse41 : 1;
1006 /** Supports SSE4.2. */
1007 uint32_t fSse42 : 1;
1008 /** Supports AVX. */
1009 uint32_t fAvx : 1;
1010 /** Supports AVX2. */
1011 uint32_t fAvx2 : 1;
1012 /** Supports AVX512 foundation. */
1013 uint32_t fAvx512Foundation : 1;
1014 /** Supports RDTSC. */
1015 uint32_t fTsc : 1;
1016 /** Intel SYSENTER/SYSEXIT support */
1017 uint32_t fSysEnter : 1;
1018 /** First generation APIC. */
1019 uint32_t fApic : 1;
1020 /** Second generation APIC. */
1021 uint32_t fX2Apic : 1;
1022 /** Hypervisor present. */
1023 uint32_t fHypervisorPresent : 1;
1024 /** MWAIT & MONITOR instructions supported. */
1025 uint32_t fMonitorMWait : 1;
1026 /** MWAIT Extensions present. */
1027 uint32_t fMWaitExtensions : 1;
1028 /** Supports CMPXCHG16B in 64-bit mode. */
1029 uint32_t fMovCmpXchg16b : 1;
1030 /** Supports CLFLUSH. */
1031 uint32_t fClFlush : 1;
1032 /** Supports CLFLUSHOPT. */
1033 uint32_t fClFlushOpt : 1;
1034 /** Supports IA32_PRED_CMD.IBPB. */
1035 uint32_t fIbpb : 1;
1036 /** Supports IA32_SPEC_CTRL.IBRS. */
1037 uint32_t fIbrs : 1;
1038 /** Supports IA32_SPEC_CTRL.STIBP. */
1039 uint32_t fStibp : 1;
1040 /** Supports IA32_ARCH_CAP. */
1041 uint32_t fArchCap : 1;
1042 /** Supports PCID. */
1043 uint32_t fPcid : 1;
1044 /** Supports INVPCID. */
1045 uint32_t fInvpcid : 1;
1046 /** Supports read/write FSGSBASE instructions. */
1047 uint32_t fFsGsBase : 1;
1048
1049 /** Supports AMD 3DNow instructions. */
1050 uint32_t f3DNow : 1;
1051 /** Supports the 3DNow/AMD64 prefetch instructions (could be nops). */
1052 uint32_t f3DNowPrefetch : 1;
1053
1054 /** AMD64: Supports long mode. */
1055 uint32_t fLongMode : 1;
1056 /** AMD64: SYSCALL/SYSRET support. */
1057 uint32_t fSysCall : 1;
1058 /** AMD64: No-execute page table bit. */
1059 uint32_t fNoExecute : 1;
1060 /** AMD64: Supports LAHF & SAHF instructions in 64-bit mode. */
1061 uint32_t fLahfSahf : 1;
1062 /** AMD64: Supports RDTSCP. */
1063 uint32_t fRdTscP : 1;
1064 /** AMD64: Supports MOV CR8 in 32-bit code (lock prefix hack). */
1065 uint32_t fMovCr8In32Bit : 1;
1066 /** AMD64: Supports XOP (similar to VEX3/AVX). */
1067 uint32_t fXop : 1;
1068
1069 /** Indicates that FPU instruction and data pointers may leak.
1070 * This generally applies to recent AMD CPUs, where the FPU IP and DP pointer
1071 * is only saved and restored if an exception is pending. */
1072 uint32_t fLeakyFxSR : 1;
1073
1074 /** AMD64: Supports AMD SVM. */
1075 uint32_t fSvm : 1;
1076
1077 /** Support for Intel VMX. */
1078 uint32_t fVmx : 1;
1079
1080 /** Indicates that speculative execution control CPUID bits and
1081 * MSRs are exposed. The details are different for Intel and
1082 * AMD but both have similar functionality. */
1083 uint32_t fSpeculationControl : 1;
1084
1085 /** Alignment padding / reserved for future use. */
1086 uint32_t fPadding : 15;
1087
1088 /** SVM: Supports Nested-paging. */
1089 uint32_t fSvmNestedPaging : 1;
1090 /** SVM: Support LBR (Last Branch Record) virtualization. */
1091 uint32_t fSvmLbrVirt : 1;
1092 /** SVM: Supports SVM lock. */
1093 uint32_t fSvmSvmLock : 1;
1094 /** SVM: Supports Next RIP save. */
1095 uint32_t fSvmNextRipSave : 1;
1096 /** SVM: Supports TSC rate MSR. */
1097 uint32_t fSvmTscRateMsr : 1;
1098 /** SVM: Supports VMCB clean bits. */
1099 uint32_t fSvmVmcbClean : 1;
1100 /** SVM: Supports Flush-by-ASID. */
1101 uint32_t fSvmFlusbByAsid : 1;
1102 /** SVM: Supports decode assist. */
1103 uint32_t fSvmDecodeAssists : 1;
1104 /** SVM: Supports Pause filter. */
1105 uint32_t fSvmPauseFilter : 1;
1106 /** SVM: Supports Pause filter threshold. */
1107 uint32_t fSvmPauseFilterThreshold : 1;
1108 /** SVM: Supports AVIC (Advanced Virtual Interrupt Controller). */
1109 uint32_t fSvmAvic : 1;
1110 /** SVM: Supports Virtualized VMSAVE/VMLOAD. */
1111 uint32_t fSvmVirtVmsaveVmload : 1;
1112 /** SVM: Supports VGIF (Virtual Global Interrupt Flag). */
1113 uint32_t fSvmVGif : 1;
1114 /** SVM: Padding / reserved for future features. */
1115 uint32_t fSvmPadding0 : 19;
1116 /** SVM: Maximum supported ASID. */
1117 uint32_t uSvmMaxAsid;
1118
1119 /** @todo VMX features. */
1120 uint32_t auPadding[1];
1121} CPUMFEATURES;
1122#ifndef VBOX_FOR_DTRACE_LIB
1123AssertCompileSize(CPUMFEATURES, 32);
1124#endif
1125/** Pointer to a CPU feature structure. */
1126typedef CPUMFEATURES *PCPUMFEATURES;
1127/** Pointer to a const CPU feature structure. */
1128typedef CPUMFEATURES const *PCCPUMFEATURES;
1129
1130
1131#ifndef VBOX_FOR_DTRACE_LIB
1132
1133/** @name Guest Register Getters.
1134 * @{ */
1135VMMDECL(void) CPUMGetGuestGDTR(PVMCPU pVCpu, PVBOXGDTR pGDTR);
1136VMMDECL(RTGCPTR) CPUMGetGuestIDTR(PVMCPU pVCpu, uint16_t *pcbLimit);
1137VMMDECL(RTSEL) CPUMGetGuestTR(PVMCPU pVCpu, PCPUMSELREGHID pHidden);
1138VMMDECL(RTSEL) CPUMGetGuestLDTR(PVMCPU pVCpu);
1139VMMDECL(RTSEL) CPUMGetGuestLdtrEx(PVMCPU pVCpu, uint64_t *pGCPtrBase, uint32_t *pcbLimit);
1140VMMDECL(uint64_t) CPUMGetGuestCR0(PVMCPU pVCpu);
1141VMMDECL(uint64_t) CPUMGetGuestCR2(PVMCPU pVCpu);
1142VMMDECL(uint64_t) CPUMGetGuestCR3(PVMCPU pVCpu);
1143VMMDECL(uint64_t) CPUMGetGuestCR4(PVMCPU pVCpu);
1144VMMDECL(uint64_t) CPUMGetGuestCR8(PVMCPU pVCpu);
1145VMMDECL(int) CPUMGetGuestCRx(PVMCPU pVCpu, unsigned iReg, uint64_t *pValue);
1146VMMDECL(uint32_t) CPUMGetGuestEFlags(PVMCPU pVCpu);
1147VMMDECL(uint32_t) CPUMGetGuestEIP(PVMCPU pVCpu);
1148VMMDECL(uint64_t) CPUMGetGuestRIP(PVMCPU pVCpu);
1149VMMDECL(uint32_t) CPUMGetGuestEAX(PVMCPU pVCpu);
1150VMMDECL(uint32_t) CPUMGetGuestEBX(PVMCPU pVCpu);
1151VMMDECL(uint32_t) CPUMGetGuestECX(PVMCPU pVCpu);
1152VMMDECL(uint32_t) CPUMGetGuestEDX(PVMCPU pVCpu);
1153VMMDECL(uint32_t) CPUMGetGuestESI(PVMCPU pVCpu);
1154VMMDECL(uint32_t) CPUMGetGuestEDI(PVMCPU pVCpu);
1155VMMDECL(uint32_t) CPUMGetGuestESP(PVMCPU pVCpu);
1156VMMDECL(uint32_t) CPUMGetGuestEBP(PVMCPU pVCpu);
1157VMMDECL(RTSEL) CPUMGetGuestCS(PVMCPU pVCpu);
1158VMMDECL(RTSEL) CPUMGetGuestDS(PVMCPU pVCpu);
1159VMMDECL(RTSEL) CPUMGetGuestES(PVMCPU pVCpu);
1160VMMDECL(RTSEL) CPUMGetGuestFS(PVMCPU pVCpu);
1161VMMDECL(RTSEL) CPUMGetGuestGS(PVMCPU pVCpu);
1162VMMDECL(RTSEL) CPUMGetGuestSS(PVMCPU pVCpu);
1163VMMDECL(uint64_t) CPUMGetGuestFlatPC(PVMCPU pVCpu);
1164VMMDECL(uint64_t) CPUMGetGuestFlatSP(PVMCPU pVCpu);
1165VMMDECL(uint64_t) CPUMGetGuestDR0(PVMCPU pVCpu);
1166VMMDECL(uint64_t) CPUMGetGuestDR1(PVMCPU pVCpu);
1167VMMDECL(uint64_t) CPUMGetGuestDR2(PVMCPU pVCpu);
1168VMMDECL(uint64_t) CPUMGetGuestDR3(PVMCPU pVCpu);
1169VMMDECL(uint64_t) CPUMGetGuestDR6(PVMCPU pVCpu);
1170VMMDECL(uint64_t) CPUMGetGuestDR7(PVMCPU pVCpu);
1171VMMDECL(int) CPUMGetGuestDRx(PVMCPU pVCpu, uint32_t iReg, uint64_t *pValue);
1172VMMDECL(void) CPUMGetGuestCpuId(PVMCPU pVCpu, uint32_t iLeaf, uint32_t iSubLeaf,
1173 uint32_t *pEax, uint32_t *pEbx, uint32_t *pEcx, uint32_t *pEdx);
1174VMMDECL(uint64_t) CPUMGetGuestEFER(PVMCPU pVCpu);
1175VMM_INT_DECL(uint64_t) CPUMGetGuestIa32MtrrCap(PVMCPU pVCpu);
1176VMM_INT_DECL(uint64_t) CPUMGetGuestIa32FeatureControl(PVMCPU pVCpu);
1177VMMDECL(VBOXSTRICTRC) CPUMQueryGuestMsr(PVMCPU pVCpu, uint32_t idMsr, uint64_t *puValue);
1178VMMDECL(VBOXSTRICTRC) CPUMSetGuestMsr(PVMCPU pVCpu, uint32_t idMsr, uint64_t uValue);
1179VMMDECL(CPUMCPUVENDOR) CPUMGetGuestCpuVendor(PVM pVM);
1180VMMDECL(CPUMCPUVENDOR) CPUMGetHostCpuVendor(PVM pVM);
1181/** @} */
1182
1183/** @name Guest Register Setters.
1184 * @{ */
1185VMMDECL(int) CPUMSetGuestGDTR(PVMCPU pVCpu, uint64_t GCPtrBase, uint16_t cbLimit);
1186VMMDECL(int) CPUMSetGuestIDTR(PVMCPU pVCpu, uint64_t GCPtrBase, uint16_t cbLimit);
1187VMMDECL(int) CPUMSetGuestTR(PVMCPU pVCpu, uint16_t tr);
1188VMMDECL(int) CPUMSetGuestLDTR(PVMCPU pVCpu, uint16_t ldtr);
1189VMMDECL(int) CPUMSetGuestCR0(PVMCPU pVCpu, uint64_t cr0);
1190VMMDECL(int) CPUMSetGuestCR2(PVMCPU pVCpu, uint64_t cr2);
1191VMMDECL(int) CPUMSetGuestCR3(PVMCPU pVCpu, uint64_t cr3);
1192VMMDECL(int) CPUMSetGuestCR4(PVMCPU pVCpu, uint64_t cr4);
1193VMMDECL(int) CPUMSetGuestDR0(PVMCPU pVCpu, uint64_t uDr0);
1194VMMDECL(int) CPUMSetGuestDR1(PVMCPU pVCpu, uint64_t uDr1);
1195VMMDECL(int) CPUMSetGuestDR2(PVMCPU pVCpu, uint64_t uDr2);
1196VMMDECL(int) CPUMSetGuestDR3(PVMCPU pVCpu, uint64_t uDr3);
1197VMMDECL(int) CPUMSetGuestDR6(PVMCPU pVCpu, uint64_t uDr6);
1198VMMDECL(int) CPUMSetGuestDR7(PVMCPU pVCpu, uint64_t uDr7);
1199VMMDECL(int) CPUMSetGuestDRx(PVMCPU pVCpu, uint32_t iReg, uint64_t Value);
1200VMM_INT_DECL(int) CPUMSetGuestXcr0(PVMCPU pVCpu, uint64_t uNewValue);
1201VMMDECL(int) CPUMSetGuestEFlags(PVMCPU pVCpu, uint32_t eflags);
1202VMMDECL(int) CPUMSetGuestEIP(PVMCPU pVCpu, uint32_t eip);
1203VMMDECL(int) CPUMSetGuestEAX(PVMCPU pVCpu, uint32_t eax);
1204VMMDECL(int) CPUMSetGuestEBX(PVMCPU pVCpu, uint32_t ebx);
1205VMMDECL(int) CPUMSetGuestECX(PVMCPU pVCpu, uint32_t ecx);
1206VMMDECL(int) CPUMSetGuestEDX(PVMCPU pVCpu, uint32_t edx);
1207VMMDECL(int) CPUMSetGuestESI(PVMCPU pVCpu, uint32_t esi);
1208VMMDECL(int) CPUMSetGuestEDI(PVMCPU pVCpu, uint32_t edi);
1209VMMDECL(int) CPUMSetGuestESP(PVMCPU pVCpu, uint32_t esp);
1210VMMDECL(int) CPUMSetGuestEBP(PVMCPU pVCpu, uint32_t ebp);
1211VMMDECL(int) CPUMSetGuestCS(PVMCPU pVCpu, uint16_t cs);
1212VMMDECL(int) CPUMSetGuestDS(PVMCPU pVCpu, uint16_t ds);
1213VMMDECL(int) CPUMSetGuestES(PVMCPU pVCpu, uint16_t es);
1214VMMDECL(int) CPUMSetGuestFS(PVMCPU pVCpu, uint16_t fs);
1215VMMDECL(int) CPUMSetGuestGS(PVMCPU pVCpu, uint16_t gs);
1216VMMDECL(int) CPUMSetGuestSS(PVMCPU pVCpu, uint16_t ss);
1217VMMDECL(void) CPUMSetGuestEFER(PVMCPU pVCpu, uint64_t val);
1218VMMR3_INT_DECL(void) CPUMR3SetGuestCpuIdFeature(PVM pVM, CPUMCPUIDFEATURE enmFeature);
1219VMMR3_INT_DECL(void) CPUMR3ClearGuestCpuIdFeature(PVM pVM, CPUMCPUIDFEATURE enmFeature);
1220VMMR3_INT_DECL(bool) CPUMR3GetGuestCpuIdFeature(PVM pVM, CPUMCPUIDFEATURE enmFeature);
1221VMMDECL(bool) CPUMSetGuestCpuIdPerCpuApicFeature(PVMCPU pVCpu, bool fVisible);
1222VMMDECL(void) CPUMSetGuestCtx(PVMCPU pVCpu, const PCPUMCTX pCtx);
1223VMM_INT_DECL(void) CPUMGuestLazyLoadHiddenCsAndSs(PVMCPU pVCpu);
1224VMM_INT_DECL(void) CPUMGuestLazyLoadHiddenSelectorReg(PVMCPU pVCpu, PCPUMSELREG pSReg);
1225VMM_INT_DECL(void) CPUMSetGuestTscAux(PVMCPU pVCpu, uint64_t uValue);
1226VMM_INT_DECL(uint64_t) CPUMGetGuestTscAux(PVMCPU pVCpu);
1227VMM_INT_DECL(void) CPUMSetGuestSpecCtrl(PVMCPU pVCpu, uint64_t uValue);
1228VMM_INT_DECL(uint64_t) CPUMGetGuestSpecCtrl(PVMCPU pVCpu);
1229/** @} */
1230
1231
1232/** @name Misc Guest Predicate Functions.
1233 * @{ */
1234VMMDECL(bool) CPUMIsGuestIn16BitCode(PVMCPU pVCpu);
1235VMMDECL(bool) CPUMIsGuestIn32BitCode(PVMCPU pVCpu);
1236VMMDECL(bool) CPUMIsGuestIn64BitCode(PVMCPU pVCpu);
1237VMMDECL(bool) CPUMIsGuestNXEnabled(PVMCPU pVCpu);
1238VMMDECL(bool) CPUMIsGuestPageSizeExtEnabled(PVMCPU pVCpu);
1239VMMDECL(bool) CPUMIsGuestPagingEnabled(PVMCPU pVCpu);
1240VMMDECL(bool) CPUMIsGuestR0WriteProtEnabled(PVMCPU pVCpu);
1241VMMDECL(bool) CPUMIsGuestInRealMode(PVMCPU pVCpu);
1242VMMDECL(bool) CPUMIsGuestInRealOrV86Mode(PVMCPU pVCpu);
1243VMMDECL(bool) CPUMIsGuestInProtectedMode(PVMCPU pVCpu);
1244VMMDECL(bool) CPUMIsGuestInPagedProtectedMode(PVMCPU pVCpu);
1245VMMDECL(bool) CPUMIsGuestInLongMode(PVMCPU pVCpu);
1246VMMDECL(bool) CPUMIsGuestInPAEMode(PVMCPU pVCpu);
1247VMM_INT_DECL(bool) CPUMIsGuestInRawMode(PVMCPU pVCpu);
1248/** @} */
1249
1250/** @name Nested Hardware-Virtualization Helpers.
1251 * @{ */
1252VMM_INT_DECL(bool) CPUMCanSvmNstGstTakePhysIntr(PVMCPU pVCpu, PCCPUMCTX pCtx);
1253VMM_INT_DECL(bool) CPUMCanSvmNstGstTakeVirtIntr(PVMCPU pVCpu, PCCPUMCTX pCtx);
1254VMM_INT_DECL(uint8_t) CPUMGetSvmNstGstInterrupt(PCCPUMCTX pCtx);
1255VMM_INT_DECL(bool) CPUMGetSvmNstGstVGif(PCCPUMCTX pCtx);
1256VMM_INT_DECL(void) CPUMSvmVmExitRestoreHostState(PVMCPU pVCpu, PCPUMCTX pCtx);
1257VMM_INT_DECL(void) CPUMSvmVmRunSaveHostState(PCPUMCTX pCtx, uint8_t cbInstr);
1258VMM_INT_DECL(uint64_t) CPUMApplyNestedGuestTscOffset(PVMCPU pVCpu, uint64_t uTicks);
1259/** @} */
1260
1261/** @name Externalized State Helpers.
1262 * @{ */
1263/** @def CPUM_ASSERT_NOT_EXTRN
1264 * Macro for asserting that @a a_fNotExtrn are present.
1265 *
1266 * @param a_pVCpu The cross context virtual CPU structure of the calling EMT.
1267 * @param a_fNotExtrn Mask of CPUMCTX_EXTRN_XXX bits to check.
1268 *
1269 * @remarks Requires VMCPU_INCL_CPUM_GST_CTX to be defined.
1270 */
1271#define CPUM_ASSERT_NOT_EXTRN(a_pVCpu, a_fNotExtrn) \
1272 AssertMsg(!((a_pVCpu)->cpum.GstCtx.fExtrn & (a_fNotExtrn)), \
1273 ("%#RX64; a_fNotExtrn=%#RX64\n", (a_pVCpu)->cpum.GstCtx.fExtrn, (a_fNotExtrn)))
1274
1275/** @def CPUM_IMPORT_EXTRN_RET
1276 * Macro for making sure the state specified by @a fExtrnImport is present,
1277 * calling CPUMImportGuestStateOnDemand() to get it if necessary.
1278 *
1279 * Will return if CPUMImportGuestStateOnDemand() fails.
1280 *
1281 * @param a_pVCpu The cross context virtual CPU structure of the calling EMT.
1282 * @param a_fExtrnImport Mask of CPUMCTX_EXTRN_XXX bits to get.
1283 * @thread EMT(a_pVCpu)
1284 *
1285 * @remarks Requires VMCPU_INCL_CPUM_GST_CTX to be defined.
1286 */
1287#define CPUM_IMPORT_EXTRN_RET(a_pVCpu, a_fExtrnImport) \
1288 do { \
1289 if (!((a_pVCpu)->cpum.GstCtx.fExtrn & (a_fExtrnImport))) \
1290 { /* already present, consider this likely */ } \
1291 else \
1292 { \
1293 int rcCpumImport = CPUMImportGuestStateOnDemand(a_pVCpu, a_fExtrnImport); \
1294 AssertRCReturn(rcCpumImport, rcCpumImport); \
1295 } \
1296 } while (0)
1297
1298/** @def CPUM_IMPORT_EXTRN_RCSTRICT
1299 * Macro for making sure the state specified by @a fExtrnImport is present,
1300 * calling CPUMImportGuestStateOnDemand() to get it if necessary.
1301 *
1302 * Will update a_rcStrict if CPUMImportGuestStateOnDemand() fails.
1303 *
1304 * @param a_pVCpu The cross context virtual CPU structure of the calling EMT.
1305 * @param a_fExtrnImport Mask of CPUMCTX_EXTRN_XXX bits to get.
1306 * @param a_rcStrict Strict status code variable to update on failure.
1307 * @thread EMT(a_pVCpu)
1308 *
1309 * @remarks Requires VMCPU_INCL_CPUM_GST_CTX to be defined.
1310 */
1311#define CPUM_IMPORT_EXTRN_RCSTRICT(a_pVCpu, a_fExtrnImport, a_rcStrict) \
1312 do { \
1313 if (!((a_pVCpu)->cpum.GstCtx.fExtrn & (a_fExtrnImport))) \
1314 { /* already present, consider this likely */ } \
1315 else \
1316 { \
1317 int rcCpumImport = CPUMImportGuestStateOnDemand(a_pVCpu, a_fExtrnImport); \
1318 AssertStmt(RT_SUCCESS(rcCpumImport) || RT_FAILURE_NP(a_rcStrict), a_rcStrict = rcCpumImport); \
1319 } \
1320 } while (0)
1321
1322VMM_INT_DECL(int) CPUMImportGuestStateOnDemand(PVMCPU pVCpu, uint64_t fExtrnImport);
1323/** @} */
1324
1325#ifndef IPRT_WITHOUT_NAMED_UNIONS_AND_STRUCTS
1326
1327/**
1328 * Tests if the guest is running in real mode or not.
1329 *
1330 * @returns true if in real mode, otherwise false.
1331 * @param pCtx Current CPU context.
1332 */
1333DECLINLINE(bool) CPUMIsGuestInRealModeEx(PCCPUMCTX pCtx)
1334{
1335 return !(pCtx->cr0 & X86_CR0_PE);
1336}
1337
1338/**
1339 * Tests if the guest is running in real or virtual 8086 mode.
1340 *
1341 * @returns @c true if it is, @c false if not.
1342 * @param pCtx Current CPU context.
1343 */
1344DECLINLINE(bool) CPUMIsGuestInRealOrV86ModeEx(PCPUMCTX pCtx)
1345{
1346 return !(pCtx->cr0 & X86_CR0_PE)
1347 || pCtx->eflags.Bits.u1VM; /* Cannot be set in long mode. Intel spec 2.3.1 "System Flags and Fields in IA-32e Mode". */
1348}
1349
1350/**
1351 * Tests if the guest is running in virtual 8086 mode.
1352 *
1353 * @returns @c true if it is, @c false if not.
1354 * @param pCtx Current CPU context.
1355 */
1356DECLINLINE(bool) CPUMIsGuestInV86ModeEx(PCCPUMCTX pCtx)
1357{
1358 return (pCtx->eflags.Bits.u1VM == 1);
1359}
1360
1361/**
1362 * Tests if the guest is running in paged protected or not.
1363 *
1364 * @returns true if in paged protected mode, otherwise false.
1365 * @param pCtx Current CPU context.
1366 */
1367DECLINLINE(bool) CPUMIsGuestInPagedProtectedModeEx(PCPUMCTX pCtx)
1368{
1369 return (pCtx->cr0 & (X86_CR0_PE | X86_CR0_PG)) == (X86_CR0_PE | X86_CR0_PG);
1370}
1371
1372/**
1373 * Tests if the guest is running in long mode or not.
1374 *
1375 * @returns true if in long mode, otherwise false.
1376 * @param pCtx Current CPU context.
1377 */
1378DECLINLINE(bool) CPUMIsGuestInLongModeEx(PCCPUMCTX pCtx)
1379{
1380 return (pCtx->msrEFER & MSR_K6_EFER_LMA) == MSR_K6_EFER_LMA;
1381}
1382
1383VMM_INT_DECL(bool) CPUMIsGuestIn64BitCodeSlow(PCPUMCTX pCtx);
1384
1385/**
1386 * Tests if the guest is running in 64 bits mode or not.
1387 *
1388 * @returns true if in 64 bits protected mode, otherwise false.
1389 * @param pCtx Current CPU context.
1390 */
1391DECLINLINE(bool) CPUMIsGuestIn64BitCodeEx(PCPUMCTX pCtx)
1392{
1393 if (!(pCtx->msrEFER & MSR_K6_EFER_LMA))
1394 return false;
1395 if (!CPUMSELREG_ARE_HIDDEN_PARTS_VALID(NULL, &pCtx->cs))
1396 return CPUMIsGuestIn64BitCodeSlow(pCtx);
1397 return pCtx->cs.Attr.n.u1Long;
1398}
1399
1400/**
1401 * Tests if the guest has paging enabled or not.
1402 *
1403 * @returns true if paging is enabled, otherwise false.
1404 * @param pCtx Current CPU context.
1405 */
1406DECLINLINE(bool) CPUMIsGuestPagingEnabledEx(PCCPUMCTX pCtx)
1407{
1408 return !!(pCtx->cr0 & X86_CR0_PG);
1409}
1410
1411/**
1412 * Tests if the guest is running in PAE mode or not.
1413 *
1414 * @returns true if in PAE mode, otherwise false.
1415 * @param pCtx Current CPU context.
1416 */
1417DECLINLINE(bool) CPUMIsGuestInPAEModeEx(PCCPUMCTX pCtx)
1418{
1419 /* Intel mentions EFER.LMA and EFER.LME in different parts of their spec. We shall use EFER.LMA rather
1420 than EFER.LME as it reflects if the CPU has entered paging with EFER.LME set. */
1421 return ( (pCtx->cr4 & X86_CR4_PAE)
1422 && CPUMIsGuestPagingEnabledEx(pCtx)
1423 && !(pCtx->msrEFER & MSR_K6_EFER_LMA));
1424}
1425
1426/**
1427 * Tests is if the guest has AMD SVM enabled or not.
1428 *
1429 * @returns true if SMV is enabled, otherwise false.
1430 * @param pCtx Current CPU context.
1431 */
1432DECLINLINE(bool) CPUMIsGuestSvmEnabled(PCCPUMCTX pCtx)
1433{
1434 return RT_BOOL(pCtx->msrEFER & MSR_K6_EFER_SVME);
1435}
1436
1437#ifndef IN_RC
1438
1439/**
1440 * Checks if the nested-guest VMCB has the specified ctrl/instruction intercept
1441 * active.
1442 *
1443 * @returns @c true if in intercept is set, @c false otherwise.
1444 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1445 * @param pCtx Pointer to the context.
1446 * @param fIntercept The SVM control/instruction intercept, see
1447 * SVM_CTRL_INTERCEPT_*.
1448 */
1449DECLINLINE(bool) CPUMIsGuestSvmCtrlInterceptSet(PVMCPU pVCpu, PCPUMCTX pCtx, uint64_t fIntercept)
1450{
1451 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1452 if (!pVmcb)
1453 return false;
1454 if (HMHasGuestSvmVmcbCached(pVCpu))
1455 return HMIsGuestSvmCtrlInterceptSet(pVCpu, fIntercept);
1456 return RT_BOOL(pVmcb->ctrl.u64InterceptCtrl & fIntercept);
1457}
1458
1459/**
1460 * Checks if the nested-guest VMCB has the specified CR read intercept active.
1461 *
1462 * @returns @c true if in intercept is set, @c false otherwise.
1463 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1464 * @param pCtx Pointer to the context.
1465 * @param uCr The CR register number (0 to 15).
1466 */
1467DECLINLINE(bool) CPUMIsGuestSvmReadCRxInterceptSet(PVMCPU pVCpu, PCCPUMCTX pCtx, uint8_t uCr)
1468{
1469 Assert(uCr < 16);
1470 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1471 if (!pVmcb)
1472 return false;
1473 if (HMHasGuestSvmVmcbCached(pVCpu))
1474 return HMIsGuestSvmReadCRxInterceptSet(pVCpu, uCr);
1475 return RT_BOOL(pVmcb->ctrl.u16InterceptRdCRx & (UINT16_C(1) << uCr));
1476}
1477
1478/**
1479 * Checks if the nested-guest VMCB has the specified CR write intercept active.
1480 *
1481 * @returns @c true if in intercept is set, @c false otherwise.
1482 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1483 * @param pCtx Pointer to the context.
1484 * @param uCr The CR register number (0 to 15).
1485 */
1486DECLINLINE(bool) CPUMIsGuestSvmWriteCRxInterceptSet(PVMCPU pVCpu, PCCPUMCTX pCtx, uint8_t uCr)
1487{
1488 Assert(uCr < 16);
1489 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1490 if (!pVmcb)
1491 return false;
1492 if (HMHasGuestSvmVmcbCached(pVCpu))
1493 return HMIsGuestSvmWriteCRxInterceptSet(pVCpu, uCr);
1494 return RT_BOOL(pVmcb->ctrl.u16InterceptWrCRx & (UINT16_C(1) << uCr));
1495}
1496
1497/**
1498 * Checks if the nested-guest VMCB has the specified DR read intercept active.
1499 *
1500 * @returns @c true if in intercept is set, @c false otherwise.
1501 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1502 * @param pCtx Pointer to the context.
1503 * @param uDr The DR register number (0 to 15).
1504 */
1505DECLINLINE(bool) CPUMIsGuestSvmReadDRxInterceptSet(PVMCPU pVCpu, PCCPUMCTX pCtx, uint8_t uDr)
1506{
1507 Assert(uDr < 16);
1508 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1509 if (!pVmcb)
1510 return false;
1511 if (HMHasGuestSvmVmcbCached(pVCpu))
1512 return HMIsGuestSvmReadDRxInterceptSet(pVCpu, uDr);
1513 return RT_BOOL(pVmcb->ctrl.u16InterceptRdDRx & (UINT16_C(1) << uDr));
1514}
1515
1516/**
1517 * Checks if the nested-guest VMCB has the specified DR write intercept active.
1518 *
1519 * @returns @c true if in intercept is set, @c false otherwise.
1520 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1521 * @param pCtx Pointer to the context.
1522 * @param uDr The DR register number (0 to 15).
1523 */
1524DECLINLINE(bool) CPUMIsGuestSvmWriteDRxInterceptSet(PVMCPU pVCpu, PCCPUMCTX pCtx, uint8_t uDr)
1525{
1526 Assert(uDr < 16);
1527 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1528 if (!pVmcb)
1529 return false;
1530 if (HMHasGuestSvmVmcbCached(pVCpu))
1531 return HMIsGuestSvmWriteDRxInterceptSet(pVCpu, uDr);
1532 return RT_BOOL(pVmcb->ctrl.u16InterceptWrDRx & (UINT16_C(1) << uDr));
1533}
1534
1535/**
1536 * Checks if the nested-guest VMCB has the specified exception intercept active.
1537 *
1538 * @returns @c true if in intercept is active, @c false otherwise.
1539 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1540 * @param pCtx Pointer to the context.
1541 * @param uVector The exception / interrupt vector.
1542 */
1543DECLINLINE(bool) CPUMIsGuestSvmXcptInterceptSet(PVMCPU pVCpu, PCCPUMCTX pCtx, uint8_t uVector)
1544{
1545 Assert(uVector < 32);
1546 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1547 if (!pVmcb)
1548 return false;
1549 if (HMHasGuestSvmVmcbCached(pVCpu))
1550 return HMIsGuestSvmXcptInterceptSet(pVCpu, uVector);
1551 return RT_BOOL(pVmcb->ctrl.u32InterceptXcpt & (UINT32_C(1) << uVector));
1552}
1553
1554/**
1555 * Checks if the nested-guest VMCB has virtual-interrupt masking enabled.
1556 *
1557 * @returns @c true if virtual-interrupts are masked, @c false otherwise.
1558 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1559 * @param pCtx Pointer to the context.
1560 *
1561 * @remarks Should only be called when SVM feature is exposed to the guest.
1562 */
1563DECLINLINE(bool) CPUMIsGuestSvmVirtIntrMasking(PVMCPU pVCpu, PCCPUMCTX pCtx)
1564{
1565 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1566 if (!pVmcb)
1567 return false;
1568 if (HMHasGuestSvmVmcbCached(pVCpu))
1569 return HMIsGuestSvmVirtIntrMasking(pVCpu);
1570 return pVmcb->ctrl.IntCtrl.n.u1VIntrMasking;
1571}
1572
1573/**
1574 * Checks if the nested-guest VMCB has nested-paging enabled.
1575 *
1576 * @returns @c true if nested-paging is enabled, @c false otherwise.
1577 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1578 * @param pCtx Pointer to the context.
1579 *
1580 * @remarks Should only be called when SVM feature is exposed to the guest.
1581 */
1582DECLINLINE(bool) CPUMIsGuestSvmNestedPagingEnabled(PVMCPU pVCpu, PCCPUMCTX pCtx)
1583{
1584 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1585 if (!pVmcb)
1586 return false;
1587 if (HMHasGuestSvmVmcbCached(pVCpu))
1588 return HMIsGuestSvmNestedPagingEnabled(pVCpu);
1589 return pVmcb->ctrl.NestedPagingCtrl.n.u1NestedPaging;
1590}
1591
1592/**
1593 * Gets the nested-guest VMCB pause-filter count.
1594 *
1595 * @returns The pause-filter count.
1596 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1597 * @param pCtx Pointer to the context.
1598 *
1599 * @remarks Should only be called when SVM feature is exposed to the guest.
1600 */
1601DECLINLINE(uint16_t) CPUMGetGuestSvmPauseFilterCount(PVMCPU pVCpu, PCCPUMCTX pCtx)
1602{
1603 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1604 if (!pVmcb)
1605 return false;
1606 if (HMHasGuestSvmVmcbCached(pVCpu))
1607 return HMGetGuestSvmPauseFilterCount(pVCpu);
1608 return pVmcb->ctrl.u16PauseFilterCount;
1609}
1610
1611/**
1612 * Updates the NextRIP (NRIP) field in the nested-guest VMCB.
1613 *
1614 * @param pVCpu The cross context virtual CPU structure of the calling EMT.
1615 * @param pCtx Pointer to the context.
1616 * @param cbInstr The length of the current instruction in bytes.
1617 *
1618 * @remarks Should only be called when SVM feature is exposed to the guest.
1619 */
1620DECLINLINE(void) CPUMGuestSvmUpdateNRip(PVMCPU pVCpu, PCCPUMCTX pCtx, uint8_t cbInstr)
1621{
1622 RT_NOREF(pVCpu);
1623 PSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1624 Assert(pVmcb);
1625 pVmcb->ctrl.u64NextRIP = pCtx->rip + cbInstr;
1626}
1627
1628#endif /* !IN_RC */
1629
1630/**
1631 * Checks if we are executing inside an SVM nested hardware-virtualized guest.
1632 *
1633 * @returns @c true if in SVM nested-guest mode, @c false otherwise.
1634 * @param pCtx Pointer to the context.
1635 */
1636DECLINLINE(bool) CPUMIsGuestInSvmNestedHwVirtMode(PCCPUMCTX pCtx)
1637{
1638 /*
1639 * With AMD-V, the VMRUN intercept is a pre-requisite to entering SVM guest-mode.
1640 * See AMD spec. 15.5 "VMRUN instruction" subsection "Canonicalization and Consistency Checks".
1641 */
1642#ifndef IN_RC
1643 PCSVMVMCB pVmcb = pCtx->hwvirt.svm.CTX_SUFF(pVmcb);
1644 return pVmcb && (pVmcb->ctrl.u64InterceptCtrl & SVM_CTRL_INTERCEPT_VMRUN);
1645#else
1646 NOREF(pCtx);
1647 return false;
1648#endif
1649}
1650
1651/**
1652 * Checks if we are executing inside a VMX nested hardware-virtualized guest.
1653 *
1654 * @returns @c true if in VMX nested-guest mode, @c false otherwise.
1655 * @param pCtx Pointer to the context.
1656 */
1657DECLINLINE(bool) CPUMIsGuestInVmxNestedHwVirtMode(PCCPUMCTX pCtx)
1658{
1659 /** @todo Intel. */
1660 NOREF(pCtx);
1661 return false;
1662}
1663
1664/**
1665 * Checks if we are executing inside a nested hardware-virtualized guest.
1666 *
1667 * @returns @c true if in SVM/VMX nested-guest mode, @c false otherwise.
1668 * @param pCtx Pointer to the context.
1669 */
1670DECLINLINE(bool) CPUMIsGuestInNestedHwVirtMode(PCCPUMCTX pCtx)
1671{
1672 return CPUMIsGuestInSvmNestedHwVirtMode(pCtx) || CPUMIsGuestInVmxNestedHwVirtMode(pCtx);
1673}
1674#endif /* IPRT_WITHOUT_NAMED_UNIONS_AND_STRUCTS */
1675
1676/** @} */
1677
1678
1679/** @name Hypervisor Register Getters.
1680 * @{ */
1681VMMDECL(RTSEL) CPUMGetHyperCS(PVMCPU pVCpu);
1682VMMDECL(RTSEL) CPUMGetHyperDS(PVMCPU pVCpu);
1683VMMDECL(RTSEL) CPUMGetHyperES(PVMCPU pVCpu);
1684VMMDECL(RTSEL) CPUMGetHyperFS(PVMCPU pVCpu);
1685VMMDECL(RTSEL) CPUMGetHyperGS(PVMCPU pVCpu);
1686VMMDECL(RTSEL) CPUMGetHyperSS(PVMCPU pVCpu);
1687#if 0 /* these are not correct. */
1688VMMDECL(uint32_t) CPUMGetHyperCR0(PVMCPU pVCpu);
1689VMMDECL(uint32_t) CPUMGetHyperCR2(PVMCPU pVCpu);
1690VMMDECL(uint32_t) CPUMGetHyperCR3(PVMCPU pVCpu);
1691VMMDECL(uint32_t) CPUMGetHyperCR4(PVMCPU pVCpu);
1692#endif
1693/** This register is only saved on fatal traps. */
1694VMMDECL(uint32_t) CPUMGetHyperEAX(PVMCPU pVCpu);
1695VMMDECL(uint32_t) CPUMGetHyperEBX(PVMCPU pVCpu);
1696/** This register is only saved on fatal traps. */
1697VMMDECL(uint32_t) CPUMGetHyperECX(PVMCPU pVCpu);
1698/** This register is only saved on fatal traps. */
1699VMMDECL(uint32_t) CPUMGetHyperEDX(PVMCPU pVCpu);
1700VMMDECL(uint32_t) CPUMGetHyperESI(PVMCPU pVCpu);
1701VMMDECL(uint32_t) CPUMGetHyperEDI(PVMCPU pVCpu);
1702VMMDECL(uint32_t) CPUMGetHyperEBP(PVMCPU pVCpu);
1703VMMDECL(uint32_t) CPUMGetHyperESP(PVMCPU pVCpu);
1704VMMDECL(uint32_t) CPUMGetHyperEFlags(PVMCPU pVCpu);
1705VMMDECL(uint32_t) CPUMGetHyperEIP(PVMCPU pVCpu);
1706VMMDECL(uint64_t) CPUMGetHyperRIP(PVMCPU pVCpu);
1707VMMDECL(uint32_t) CPUMGetHyperIDTR(PVMCPU pVCpu, uint16_t *pcbLimit);
1708VMMDECL(uint32_t) CPUMGetHyperGDTR(PVMCPU pVCpu, uint16_t *pcbLimit);
1709VMMDECL(RTSEL) CPUMGetHyperLDTR(PVMCPU pVCpu);
1710VMMDECL(RTGCUINTREG) CPUMGetHyperDR0(PVMCPU pVCpu);
1711VMMDECL(RTGCUINTREG) CPUMGetHyperDR1(PVMCPU pVCpu);
1712VMMDECL(RTGCUINTREG) CPUMGetHyperDR2(PVMCPU pVCpu);
1713VMMDECL(RTGCUINTREG) CPUMGetHyperDR3(PVMCPU pVCpu);
1714VMMDECL(RTGCUINTREG) CPUMGetHyperDR6(PVMCPU pVCpu);
1715VMMDECL(RTGCUINTREG) CPUMGetHyperDR7(PVMCPU pVCpu);
1716VMMDECL(void) CPUMGetHyperCtx(PVMCPU pVCpu, PCPUMCTX pCtx);
1717VMMDECL(uint32_t) CPUMGetHyperCR3(PVMCPU pVCpu);
1718/** @} */
1719
1720/** @name Hypervisor Register Setters.
1721 * @{ */
1722VMMDECL(void) CPUMSetHyperGDTR(PVMCPU pVCpu, uint32_t addr, uint16_t limit);
1723VMMDECL(void) CPUMSetHyperLDTR(PVMCPU pVCpu, RTSEL SelLDTR);
1724VMMDECL(void) CPUMSetHyperIDTR(PVMCPU pVCpu, uint32_t addr, uint16_t limit);
1725VMMDECL(void) CPUMSetHyperCR3(PVMCPU pVCpu, uint32_t cr3);
1726VMMDECL(void) CPUMSetHyperTR(PVMCPU pVCpu, RTSEL SelTR);
1727VMMDECL(void) CPUMSetHyperCS(PVMCPU pVCpu, RTSEL SelCS);
1728VMMDECL(void) CPUMSetHyperDS(PVMCPU pVCpu, RTSEL SelDS);
1729VMMDECL(void) CPUMSetHyperES(PVMCPU pVCpu, RTSEL SelDS);
1730VMMDECL(void) CPUMSetHyperFS(PVMCPU pVCpu, RTSEL SelDS);
1731VMMDECL(void) CPUMSetHyperGS(PVMCPU pVCpu, RTSEL SelDS);
1732VMMDECL(void) CPUMSetHyperSS(PVMCPU pVCpu, RTSEL SelSS);
1733VMMDECL(void) CPUMSetHyperESP(PVMCPU pVCpu, uint32_t u32ESP);
1734VMMDECL(int) CPUMSetHyperEFlags(PVMCPU pVCpu, uint32_t Efl);
1735VMMDECL(void) CPUMSetHyperEIP(PVMCPU pVCpu, uint32_t u32EIP);
1736VMM_INT_DECL(void) CPUMSetHyperState(PVMCPU pVCpu, uint32_t u32EIP, uint32_t u32ESP, uint32_t u32EAX, uint32_t u32EDX);
1737VMMDECL(void) CPUMSetHyperDR0(PVMCPU pVCpu, RTGCUINTREG uDr0);
1738VMMDECL(void) CPUMSetHyperDR1(PVMCPU pVCpu, RTGCUINTREG uDr1);
1739VMMDECL(void) CPUMSetHyperDR2(PVMCPU pVCpu, RTGCUINTREG uDr2);
1740VMMDECL(void) CPUMSetHyperDR3(PVMCPU pVCpu, RTGCUINTREG uDr3);
1741VMMDECL(void) CPUMSetHyperDR6(PVMCPU pVCpu, RTGCUINTREG uDr6);
1742VMMDECL(void) CPUMSetHyperDR7(PVMCPU pVCpu, RTGCUINTREG uDr7);
1743VMMDECL(void) CPUMSetHyperCtx(PVMCPU pVCpu, const PCPUMCTX pCtx);
1744VMMDECL(int) CPUMRecalcHyperDRx(PVMCPU pVCpu, uint8_t iGstReg, bool fForceHyper);
1745/** @} */
1746
1747VMMDECL(void) CPUMPushHyper(PVMCPU pVCpu, uint32_t u32);
1748VMMDECL(int) CPUMQueryHyperCtxPtr(PVMCPU pVCpu, PCPUMCTX *ppCtx);
1749VMMDECL(PCPUMCTX) CPUMGetHyperCtxPtr(PVMCPU pVCpu);
1750VMMDECL(PCCPUMCTXCORE) CPUMGetHyperCtxCore(PVMCPU pVCpu);
1751VMMDECL(PCPUMCTX) CPUMQueryGuestCtxPtr(PVMCPU pVCpu);
1752VMM_INT_DECL(PCPUMCTXMSRS) CPUMQueryGuestCtxMsrsPtr(PVMCPU pVCpu);
1753VMMDECL(PCCPUMCTXCORE) CPUMGetGuestCtxCore(PVMCPU pVCpu);
1754VMM_INT_DECL(int) CPUMRawEnter(PVMCPU pVCpu);
1755VMM_INT_DECL(int) CPUMRawLeave(PVMCPU pVCpu, int rc);
1756VMMDECL(uint32_t) CPUMRawGetEFlags(PVMCPU pVCpu);
1757VMMDECL(void) CPUMRawSetEFlags(PVMCPU pVCpu, uint32_t fEfl);
1758
1759/** @name Changed flags.
1760 * These flags are used to keep track of which important register that
1761 * have been changed since last they were reset. The only one allowed
1762 * to clear them is REM!
1763 * @{
1764 */
1765#define CPUM_CHANGED_FPU_REM RT_BIT(0)
1766#define CPUM_CHANGED_CR0 RT_BIT(1)
1767#define CPUM_CHANGED_CR4 RT_BIT(2)
1768#define CPUM_CHANGED_GLOBAL_TLB_FLUSH RT_BIT(3)
1769#define CPUM_CHANGED_CR3 RT_BIT(4)
1770#define CPUM_CHANGED_GDTR RT_BIT(5)
1771#define CPUM_CHANGED_IDTR RT_BIT(6)
1772#define CPUM_CHANGED_LDTR RT_BIT(7)
1773#define CPUM_CHANGED_TR RT_BIT(8) /**@< Currently unused. */
1774#define CPUM_CHANGED_SYSENTER_MSR RT_BIT(9)
1775#define CPUM_CHANGED_HIDDEN_SEL_REGS RT_BIT(10) /**@< Currently unused. */
1776#define CPUM_CHANGED_CPUID RT_BIT(11)
1777#define CPUM_CHANGED_ALL ( CPUM_CHANGED_FPU_REM \
1778 | CPUM_CHANGED_CR0 \
1779 | CPUM_CHANGED_CR4 \
1780 | CPUM_CHANGED_GLOBAL_TLB_FLUSH \
1781 | CPUM_CHANGED_CR3 \
1782 | CPUM_CHANGED_GDTR \
1783 | CPUM_CHANGED_IDTR \
1784 | CPUM_CHANGED_LDTR \
1785 | CPUM_CHANGED_TR \
1786 | CPUM_CHANGED_SYSENTER_MSR \
1787 | CPUM_CHANGED_HIDDEN_SEL_REGS \
1788 | CPUM_CHANGED_CPUID )
1789/** @} */
1790
1791VMMDECL(void) CPUMSetChangedFlags(PVMCPU pVCpu, uint32_t fChangedAdd);
1792VMMR3DECL(uint32_t) CPUMR3RemEnter(PVMCPU pVCpu, uint32_t *puCpl);
1793VMMR3DECL(void) CPUMR3RemLeave(PVMCPU pVCpu, bool fNoOutOfSyncSels);
1794VMMDECL(bool) CPUMSupportsXSave(PVM pVM);
1795VMMDECL(bool) CPUMIsHostUsingSysEnter(PVM pVM);
1796VMMDECL(bool) CPUMIsHostUsingSysCall(PVM pVM);
1797VMMDECL(bool) CPUMIsGuestFPUStateActive(PVMCPU pVCpu);
1798VMMDECL(bool) CPUMIsGuestFPUStateLoaded(PVMCPU pVCpu);
1799VMMDECL(bool) CPUMIsHostFPUStateSaved(PVMCPU pVCpu);
1800VMMDECL(bool) CPUMIsGuestDebugStateActive(PVMCPU pVCpu);
1801VMMDECL(bool) CPUMIsGuestDebugStateActivePending(PVMCPU pVCpu);
1802VMMDECL(void) CPUMDeactivateGuestDebugState(PVMCPU pVCpu);
1803VMMDECL(bool) CPUMIsHyperDebugStateActive(PVMCPU pVCpu);
1804VMMDECL(bool) CPUMIsHyperDebugStateActivePending(PVMCPU pVCpu);
1805VMMDECL(uint32_t) CPUMGetGuestCPL(PVMCPU pVCpu);
1806VMMDECL(CPUMMODE) CPUMGetGuestMode(PVMCPU pVCpu);
1807VMMDECL(uint32_t) CPUMGetGuestCodeBits(PVMCPU pVCpu);
1808VMMDECL(DISCPUMODE) CPUMGetGuestDisMode(PVMCPU pVCpu);
1809VMMDECL(uint32_t) CPUMGetGuestMxCsrMask(PVM pVM);
1810VMMDECL(uint64_t) CPUMGetGuestScalableBusFrequency(PVM pVM);
1811VMMDECL(int) CPUMQueryValidatedGuestEfer(PVM pVM, uint64_t uCr0, uint64_t uOldEfer, uint64_t uNewEfer,
1812 uint64_t *puValidEfer);
1813VMMDECL(void) CPUMSetGuestMsrEferNoCheck(PVMCPU pVCpu, uint64_t uOldEfer, uint64_t uValidEfer);
1814VMMDECL(bool) CPUMIsPatMsrValid(uint64_t uValue);
1815
1816/** @name Typical scalable bus frequency values.
1817 * @{ */
1818/** Special internal value indicating that we don't know the frequency.
1819 * @internal */
1820#define CPUM_SBUSFREQ_UNKNOWN UINT64_C(1)
1821#define CPUM_SBUSFREQ_100MHZ UINT64_C(100000000)
1822#define CPUM_SBUSFREQ_133MHZ UINT64_C(133333333)
1823#define CPUM_SBUSFREQ_167MHZ UINT64_C(166666666)
1824#define CPUM_SBUSFREQ_200MHZ UINT64_C(200000000)
1825#define CPUM_SBUSFREQ_267MHZ UINT64_C(266666666)
1826#define CPUM_SBUSFREQ_333MHZ UINT64_C(333333333)
1827#define CPUM_SBUSFREQ_400MHZ UINT64_C(400000000)
1828/** @} */
1829
1830
1831#ifdef IN_RING3
1832/** @defgroup grp_cpum_r3 The CPUM ring-3 API
1833 * @{
1834 */
1835
1836VMMR3DECL(int) CPUMR3Init(PVM pVM);
1837VMMR3DECL(int) CPUMR3InitCompleted(PVM pVM, VMINITCOMPLETED enmWhat);
1838VMMR3DECL(void) CPUMR3LogCpuIds(PVM pVM);
1839VMMR3DECL(void) CPUMR3Relocate(PVM pVM);
1840VMMR3DECL(int) CPUMR3Term(PVM pVM);
1841VMMR3DECL(void) CPUMR3Reset(PVM pVM);
1842VMMR3DECL(void) CPUMR3ResetCpu(PVM pVM, PVMCPU pVCpu);
1843VMMDECL(bool) CPUMR3IsStateRestorePending(PVM pVM);
1844VMMR3DECL(int) CPUMR3SetCR4Feature(PVM pVM, RTHCUINTREG fOr, RTHCUINTREG fAnd);
1845
1846VMMR3DECL(int) CPUMR3CpuIdInsert(PVM pVM, PCPUMCPUIDLEAF pNewLeaf);
1847VMMR3DECL(int) CPUMR3CpuIdGetLeaf(PVM pVM, PCPUMCPUIDLEAF pLeaf, uint32_t uLeaf, uint32_t uSubLeaf);
1848VMMR3DECL(CPUMMICROARCH) CPUMR3CpuIdDetermineMicroarchEx(CPUMCPUVENDOR enmVendor, uint8_t bFamily,
1849 uint8_t bModel, uint8_t bStepping);
1850VMMR3DECL(const char *) CPUMR3MicroarchName(CPUMMICROARCH enmMicroarch);
1851VMMR3DECL(int) CPUMR3CpuIdCollectLeaves(PCPUMCPUIDLEAF *ppaLeaves, uint32_t *pcLeaves);
1852VMMR3DECL(int) CPUMR3CpuIdDetectUnknownLeafMethod(PCPUMUNKNOWNCPUID penmUnknownMethod, PCPUMCPUID pDefUnknown);
1853VMMR3DECL(const char *) CPUMR3CpuIdUnknownLeafMethodName(CPUMUNKNOWNCPUID enmUnknownMethod);
1854VMMR3DECL(CPUMCPUVENDOR) CPUMR3CpuIdDetectVendorEx(uint32_t uEAX, uint32_t uEBX, uint32_t uECX, uint32_t uEDX);
1855VMMR3DECL(const char *) CPUMR3CpuVendorName(CPUMCPUVENDOR enmVendor);
1856VMMR3DECL(uint32_t) CPUMR3DeterminHostMxCsrMask(void);
1857
1858VMMR3DECL(int) CPUMR3MsrRangesInsert(PVM pVM, PCCPUMMSRRANGE pNewRange);
1859
1860# if defined(VBOX_WITH_RAW_MODE) || defined(DOXYGEN_RUNNING)
1861/** @name APIs for the CPUID raw-mode patch (legacy).
1862 * @{ */
1863VMMR3_INT_DECL(RCPTRTYPE(PCCPUMCPUID)) CPUMR3GetGuestCpuIdPatmDefRCPtr(PVM pVM);
1864VMMR3_INT_DECL(uint32_t) CPUMR3GetGuestCpuIdPatmStdMax(PVM pVM);
1865VMMR3_INT_DECL(uint32_t) CPUMR3GetGuestCpuIdPatmExtMax(PVM pVM);
1866VMMR3_INT_DECL(uint32_t) CPUMR3GetGuestCpuIdPatmCentaurMax(PVM pVM);
1867VMMR3_INT_DECL(RCPTRTYPE(PCCPUMCPUID)) CPUMR3GetGuestCpuIdPatmStdRCPtr(PVM pVM);
1868VMMR3_INT_DECL(RCPTRTYPE(PCCPUMCPUID)) CPUMR3GetGuestCpuIdPatmExtRCPtr(PVM pVM);
1869VMMR3_INT_DECL(RCPTRTYPE(PCCPUMCPUID)) CPUMR3GetGuestCpuIdPatmCentaurRCPtr(PVM pVM);
1870/** @} */
1871# endif
1872
1873/** @} */
1874#endif /* IN_RING3 */
1875
1876#ifdef IN_RC
1877/** @defgroup grp_cpum_rc The CPUM Raw-mode Context API
1878 * @{
1879 */
1880
1881/**
1882 * Calls a guest trap/interrupt handler directly
1883 *
1884 * Assumes a trap stack frame has already been setup on the guest's stack!
1885 * This function does not return!
1886 *
1887 * @param pRegFrame Original trap/interrupt context
1888 * @param selCS Code selector of handler
1889 * @param pHandler GC virtual address of handler
1890 * @param eflags Callee's EFLAGS
1891 * @param selSS Stack selector for handler
1892 * @param pEsp Stack address for handler
1893 */
1894DECLASM(void) CPUMGCCallGuestTrapHandler(PCPUMCTXCORE pRegFrame, uint32_t selCS, RTRCPTR pHandler,
1895 uint32_t eflags, uint32_t selSS, RTRCPTR pEsp);
1896
1897/**
1898 * Call guest V86 code directly.
1899 *
1900 * This function does not return!
1901 *
1902 * @param pRegFrame Original trap/interrupt context
1903 */
1904DECLASM(void) CPUMGCCallV86Code(PCPUMCTXCORE pRegFrame);
1905
1906VMMDECL(int) CPUMHandleLazyFPU(PVMCPU pVCpu);
1907VMMDECL(uint32_t) CPUMRCGetGuestCPL(PVMCPU pVCpu, PCPUMCTXCORE pRegFrame);
1908#ifdef VBOX_WITH_RAW_RING1
1909VMMDECL(void) CPUMRCRecheckRawState(PVMCPU pVCpu, PCPUMCTXCORE pCtxCore);
1910#endif
1911VMMRCDECL(void) CPUMRCProcessForceFlag(PVMCPU pVCpu);
1912
1913/** @} */
1914#endif /* IN_RC */
1915
1916#ifdef IN_RING0
1917/** @defgroup grp_cpum_r0 The CPUM ring-0 API
1918 * @{
1919 */
1920VMMR0_INT_DECL(int) CPUMR0ModuleInit(void);
1921VMMR0_INT_DECL(int) CPUMR0ModuleTerm(void);
1922VMMR0_INT_DECL(int) CPUMR0InitVM(PVM pVM);
1923DECLASM(void) CPUMR0RegisterVCpuThread(PVMCPU pVCpu);
1924DECLASM(void) CPUMR0TouchHostFpu(void);
1925VMMR0_INT_DECL(int) CPUMR0Trap07Handler(PVM pVM, PVMCPU pVCpu);
1926VMMR0_INT_DECL(int) CPUMR0LoadGuestFPU(PVM pVM, PVMCPU pVCpu);
1927VMMR0_INT_DECL(bool) CPUMR0FpuStateMaybeSaveGuestAndRestoreHost(PVMCPU pVCpu);
1928VMMR0_INT_DECL(int) CPUMR0SaveHostDebugState(PVM pVM, PVMCPU pVCpu);
1929VMMR0_INT_DECL(bool) CPUMR0DebugStateMaybeSaveGuestAndRestoreHost(PVMCPU pVCpu, bool fDr6);
1930VMMR0_INT_DECL(bool) CPUMR0DebugStateMaybeSaveGuest(PVMCPU pVCpu, bool fDr6);
1931
1932VMMR0_INT_DECL(void) CPUMR0LoadGuestDebugState(PVMCPU pVCpu, bool fDr6);
1933VMMR0_INT_DECL(void) CPUMR0LoadHyperDebugState(PVMCPU pVCpu, bool fDr6);
1934#ifdef VBOX_WITH_VMMR0_DISABLE_LAPIC_NMI
1935VMMR0_INT_DECL(void) CPUMR0SetLApic(PVMCPU pVCpu, uint32_t iHostCpuSet);
1936#endif
1937
1938/** @} */
1939#endif /* IN_RING0 */
1940
1941/** @defgroup grp_cpum_rz The CPUM raw-mode and ring-0 context API
1942 * @{
1943 */
1944VMMRZ_INT_DECL(void) CPUMRZFpuStatePrepareHostCpuForUse(PVMCPU pVCpu);
1945VMMRZ_INT_DECL(void) CPUMRZFpuStateActualizeForRead(PVMCPU pVCpu);
1946VMMRZ_INT_DECL(void) CPUMRZFpuStateActualizeForChange(PVMCPU pVCpu);
1947VMMRZ_INT_DECL(void) CPUMRZFpuStateActualizeSseForRead(PVMCPU pVCpu);
1948VMMRZ_INT_DECL(void) CPUMRZFpuStateActualizeAvxForRead(PVMCPU pVCpu);
1949/** @} */
1950
1951
1952#endif /* !VBOX_FOR_DTRACE_LIB */
1953/** @} */
1954RT_C_DECLS_END
1955
1956
1957#endif
1958
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette