VirtualBox

source: vbox/trunk/src/VBox/VMM/include/TMInternal.h@ 73016

Last change on this file since 73016 was 72522, checked in by vboxsync, 6 years ago

NEM,TM: Work on TSC and NEM/win. bugref:9044 [=>office]

  • Property svn:eol-style set to native
  • Property svn:keywords set to Id Revision
File size: 32.7 KB
Line 
1/* $Id: TMInternal.h 72522 2018-06-12 08:45:27Z vboxsync $ */
2/** @file
3 * TM - Internal header file.
4 */
5
6/*
7 * Copyright (C) 2006-2017 Oracle Corporation
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 */
17
18#ifndef ___TMInternal_h
19#define ___TMInternal_h
20
21#include <VBox/cdefs.h>
22#include <VBox/types.h>
23#include <iprt/time.h>
24#include <iprt/timer.h>
25#include <iprt/assert.h>
26#include <VBox/vmm/stam.h>
27#include <VBox/vmm/pdmcritsect.h>
28
29RT_C_DECLS_BEGIN
30
31
32/** @defgroup grp_tm_int Internal
33 * @ingroup grp_tm
34 * @internal
35 * @{
36 */
37
38/** Frequency of the real clock. */
39#define TMCLOCK_FREQ_REAL UINT32_C(1000)
40/** Frequency of the virtual clock. */
41#define TMCLOCK_FREQ_VIRTUAL UINT32_C(1000000000)
42
43
44/**
45 * Timer type.
46 */
47typedef enum TMTIMERTYPE
48{
49 /** Device timer. */
50 TMTIMERTYPE_DEV = 1,
51 /** USB device timer. */
52 TMTIMERTYPE_USB,
53 /** Driver timer. */
54 TMTIMERTYPE_DRV,
55 /** Internal timer . */
56 TMTIMERTYPE_INTERNAL,
57 /** External timer. */
58 TMTIMERTYPE_EXTERNAL
59} TMTIMERTYPE;
60
61/**
62 * Timer state
63 */
64typedef enum TMTIMERSTATE
65{
66 /** Timer is stopped. */
67 TMTIMERSTATE_STOPPED = 1,
68 /** Timer is active. */
69 TMTIMERSTATE_ACTIVE,
70 /** Timer is expired, getting expire and unlinking. */
71 TMTIMERSTATE_EXPIRED_GET_UNLINK,
72 /** Timer is expired and is being delivered. */
73 TMTIMERSTATE_EXPIRED_DELIVER,
74
75 /** Timer is stopped but still in the active list.
76 * Currently in the ScheduleTimers list. */
77 TMTIMERSTATE_PENDING_STOP,
78 /** Timer is stopped but needs unlinking from the ScheduleTimers list.
79 * Currently in the ScheduleTimers list. */
80 TMTIMERSTATE_PENDING_STOP_SCHEDULE,
81 /** Timer is being modified and will soon be pending scheduling.
82 * Currently in the ScheduleTimers list. */
83 TMTIMERSTATE_PENDING_SCHEDULE_SET_EXPIRE,
84 /** Timer is pending scheduling.
85 * Currently in the ScheduleTimers list. */
86 TMTIMERSTATE_PENDING_SCHEDULE,
87 /** Timer is being modified and will soon be pending rescheduling.
88 * Currently in the ScheduleTimers list and the active list. */
89 TMTIMERSTATE_PENDING_RESCHEDULE_SET_EXPIRE,
90 /** Timer is modified and is now pending rescheduling.
91 * Currently in the ScheduleTimers list and the active list. */
92 TMTIMERSTATE_PENDING_RESCHEDULE,
93 /** Timer is being destroyed. */
94 TMTIMERSTATE_DESTROY,
95 /** Timer is free. */
96 TMTIMERSTATE_FREE
97} TMTIMERSTATE;
98
99/** Predicate that returns true if the give state is pending scheduling or
100 * rescheduling of any kind. Will reference the argument more than once! */
101#define TMTIMERSTATE_IS_PENDING_SCHEDULING(enmState) \
102 ( (enmState) <= TMTIMERSTATE_PENDING_RESCHEDULE \
103 && (enmState) >= TMTIMERSTATE_PENDING_SCHEDULE_SET_EXPIRE)
104
105
106/**
107 * Internal representation of a timer.
108 *
109 * For correct serialization (without the use of semaphores and
110 * other blocking/slow constructs) certain rules applies to updating
111 * this structure:
112 * - For thread other than EMT only u64Expire, enmState and pScheduleNext*
113 * are changeable. Everything else is out of bounds.
114 * - Updating of u64Expire timer can only happen in the TMTIMERSTATE_STOPPED
115 * and TMTIMERSTATE_PENDING_RESCHEDULING_SET_EXPIRE states.
116 * - Timers in the TMTIMERSTATE_EXPIRED state are only accessible from EMT.
117 * - Actual destruction of a timer can only be done at scheduling time.
118 */
119typedef struct TMTIMER
120{
121 /** Expire time. */
122 volatile uint64_t u64Expire;
123 /** Clock to apply to u64Expire. */
124 TMCLOCK enmClock;
125 /** Timer callback type. */
126 TMTIMERTYPE enmType;
127 /** Type specific data. */
128 union
129 {
130 /** TMTIMERTYPE_DEV. */
131 struct
132 {
133 /** Callback. */
134 R3PTRTYPE(PFNTMTIMERDEV) pfnTimer;
135 /** Device instance. */
136 PPDMDEVINSR3 pDevIns;
137 } Dev;
138
139 /** TMTIMERTYPE_DEV. */
140 struct
141 {
142 /** Callback. */
143 R3PTRTYPE(PFNTMTIMERUSB) pfnTimer;
144 /** USB device instance. */
145 PPDMUSBINS pUsbIns;
146 } Usb;
147
148 /** TMTIMERTYPE_DRV. */
149 struct
150 {
151 /** Callback. */
152 R3PTRTYPE(PFNTMTIMERDRV) pfnTimer;
153 /** Device instance. */
154 R3PTRTYPE(PPDMDRVINS) pDrvIns;
155 } Drv;
156
157 /** TMTIMERTYPE_INTERNAL. */
158 struct
159 {
160 /** Callback. */
161 R3PTRTYPE(PFNTMTIMERINT) pfnTimer;
162 } Internal;
163
164 /** TMTIMERTYPE_EXTERNAL. */
165 struct
166 {
167 /** Callback. */
168 R3PTRTYPE(PFNTMTIMEREXT) pfnTimer;
169 } External;
170 } u;
171
172 /** Timer state. */
173 volatile TMTIMERSTATE enmState;
174 /** Timer relative offset to the next timer in the schedule list. */
175 int32_t volatile offScheduleNext;
176
177 /** Timer relative offset to the next timer in the chain. */
178 int32_t offNext;
179 /** Timer relative offset to the previous timer in the chain. */
180 int32_t offPrev;
181
182 /** Pointer to the VM the timer belongs to - R3 Ptr. */
183 PVMR3 pVMR3;
184 /** Pointer to the VM the timer belongs to - R0 Ptr. */
185 PVMR0 pVMR0;
186 /** Pointer to the VM the timer belongs to - RC Ptr. */
187 PVMRC pVMRC;
188 /** The timer frequency hint. This is 0 if not hint was given. */
189 uint32_t volatile uHzHint;
190
191 /** User argument. */
192 RTR3PTR pvUser;
193 /** The critical section associated with the lock. */
194 R3PTRTYPE(PPDMCRITSECT) pCritSect;
195
196 /** Pointer to the next timer in the list of created or free timers. (TM::pTimers or TM::pFree) */
197 PTMTIMERR3 pBigNext;
198 /** Pointer to the previous timer in the list of all created timers. (TM::pTimers) */
199 PTMTIMERR3 pBigPrev;
200 /** Pointer to the timer description. */
201 R3PTRTYPE(const char *) pszDesc;
202#if HC_ARCH_BITS == 32
203 uint32_t padding0; /**< pad structure to multiple of 8 bytes. */
204#endif
205} TMTIMER;
206AssertCompileMemberSize(TMTIMER, enmState, sizeof(uint32_t));
207
208
209/**
210 * Updates a timer state in the correct atomic manner.
211 */
212#if 1
213# define TM_SET_STATE(pTimer, state) \
214 ASMAtomicWriteU32((uint32_t volatile *)&(pTimer)->enmState, state)
215#else
216# define TM_SET_STATE(pTimer, state) \
217 do { \
218 uint32_t uOld1 = (pTimer)->enmState; \
219 Log(("%s: %p: %d -> %d\n", __FUNCTION__, (pTimer), (pTimer)->enmState, state)); \
220 uint32_t uOld2 = ASMAtomicXchgU32((uint32_t volatile *)&(pTimer)->enmState, state); \
221 Assert(uOld1 == uOld2); \
222 } while (0)
223#endif
224
225/**
226 * Tries to updates a timer state in the correct atomic manner.
227 */
228#if 1
229# define TM_TRY_SET_STATE(pTimer, StateNew, StateOld, fRc) \
230 (fRc) = ASMAtomicCmpXchgU32((uint32_t volatile *)&(pTimer)->enmState, StateNew, StateOld)
231#else
232# define TM_TRY_SET_STATE(pTimer, StateNew, StateOld, fRc) \
233 do { (fRc) = ASMAtomicCmpXchgU32((uint32_t volatile *)&(pTimer)->enmState, StateNew, StateOld); \
234 Log(("%s: %p: %d -> %d %RTbool\n", __FUNCTION__, (pTimer), StateOld, StateNew, fRc)); \
235 } while (0)
236#endif
237
238/** Get the previous timer. */
239#define TMTIMER_GET_PREV(pTimer) ((PTMTIMER)((pTimer)->offPrev ? (intptr_t)(pTimer) + (pTimer)->offPrev : 0))
240/** Get the next timer. */
241#define TMTIMER_GET_NEXT(pTimer) ((PTMTIMER)((pTimer)->offNext ? (intptr_t)(pTimer) + (pTimer)->offNext : 0))
242/** Set the previous timer link. */
243#define TMTIMER_SET_PREV(pTimer, pPrev) ((pTimer)->offPrev = (pPrev) ? (intptr_t)(pPrev) - (intptr_t)(pTimer) : 0)
244/** Set the next timer link. */
245#define TMTIMER_SET_NEXT(pTimer, pNext) ((pTimer)->offNext = (pNext) ? (intptr_t)(pNext) - (intptr_t)(pTimer) : 0)
246
247
248/**
249 * A timer queue.
250 *
251 * This is allocated on the hyper heap.
252 */
253typedef struct TMTIMERQUEUE
254{
255 /** The cached expire time for this queue.
256 * Updated by EMT when scheduling the queue or modifying the head timer.
257 * Assigned UINT64_MAX when there is no head timer. */
258 uint64_t u64Expire;
259 /** Doubly linked list of active timers.
260 *
261 * When no scheduling is pending, this list is will be ordered by expire time (ascending).
262 * Access is serialized by only letting the emulation thread (EMT) do changes.
263 *
264 * The offset is relative to the queue structure.
265 */
266 int32_t offActive;
267 /** List of timers pending scheduling of some kind.
268 *
269 * Timer stats allowed in the list are TMTIMERSTATE_PENDING_STOPPING,
270 * TMTIMERSTATE_PENDING_DESTRUCTION, TMTIMERSTATE_PENDING_STOPPING_DESTRUCTION,
271 * TMTIMERSTATE_PENDING_RESCHEDULING and TMTIMERSTATE_PENDING_SCHEDULE.
272 *
273 * The offset is relative to the queue structure.
274 */
275 int32_t volatile offSchedule;
276 /** The clock for this queue. */
277 TMCLOCK enmClock;
278 /** Pad the structure up to 32 bytes. */
279 uint32_t au32Padding[3];
280} TMTIMERQUEUE;
281
282/** Pointer to a timer queue. */
283typedef TMTIMERQUEUE *PTMTIMERQUEUE;
284
285/** Get the head of the active timer list. */
286#define TMTIMER_GET_HEAD(pQueue) ((PTMTIMER)((pQueue)->offActive ? (intptr_t)(pQueue) + (pQueue)->offActive : 0))
287/** Set the head of the active timer list. */
288#define TMTIMER_SET_HEAD(pQueue, pHead) ((pQueue)->offActive = pHead ? (intptr_t)pHead - (intptr_t)(pQueue) : 0)
289
290
291/**
292 * CPU load data set.
293 * Mainly used by tmR3CpuLoadTimer.
294 */
295typedef struct TMCPULOADSTATE
296{
297 /** The percent of the period spent executing guest code. */
298 uint8_t cPctExecuting;
299 /** The percent of the period spent halted. */
300 uint8_t cPctHalted;
301 /** The percent of the period spent on other things. */
302 uint8_t cPctOther;
303 /** Explicit alignment padding */
304 uint8_t au8Alignment[5];
305
306 /** Previous cNsTotal value. */
307 uint64_t cNsPrevTotal;
308 /** Previous cNsExecuting value. */
309 uint64_t cNsPrevExecuting;
310 /** Previous cNsHalted value. */
311 uint64_t cNsPrevHalted;
312} TMCPULOADSTATE;
313AssertCompileSizeAlignment(TMCPULOADSTATE, 8);
314AssertCompileMemberAlignment(TMCPULOADSTATE, cNsPrevTotal, 8);
315/** Pointer to a CPU load data set. */
316typedef TMCPULOADSTATE *PTMCPULOADSTATE;
317
318
319/**
320 * TSC mode.
321 *
322 * The main modes of how TM implements the TSC clock (TMCLOCK_TSC).
323 */
324typedef enum TMTSCMODE
325{
326 /** The guest TSC is an emulated, virtual TSC. */
327 TMTSCMODE_VIRT_TSC_EMULATED = 1,
328 /** The guest TSC is an offset of the real TSC. */
329 TMTSCMODE_REAL_TSC_OFFSET,
330 /** The guest TSC is dynamically derived through emulating or offsetting. */
331 TMTSCMODE_DYNAMIC,
332 /** The native API provides it. */
333 TMTSCMODE_NATIVE_API
334} TMTSCMODE;
335AssertCompileSize(TMTSCMODE, sizeof(uint32_t));
336
337
338/**
339 * Converts a TM pointer into a VM pointer.
340 * @returns Pointer to the VM structure the TM is part of.
341 * @param pTM Pointer to TM instance data.
342 */
343#define TM2VM(pTM) ( (PVM)((char*)pTM - pTM->offVM) )
344
345
346/**
347 * TM VM Instance data.
348 * Changes to this must checked against the padding of the cfgm union in VM!
349 */
350typedef struct TM
351{
352 /** Offset to the VM structure.
353 * See TM2VM(). */
354 RTUINT offVM;
355
356 /** The current TSC mode of the VM.
357 * Config variable: Mode (string). */
358 TMTSCMODE enmTSCMode;
359 /** The original TSC mode of the VM. */
360 TMTSCMODE enmOriginalTSCMode;
361 /** Alignment padding. */
362 uint32_t u32Alignment0;
363 /** Whether the TSC is tied to the execution of code.
364 * Config variable: TSCTiedToExecution (bool) */
365 bool fTSCTiedToExecution;
366 /** Modifier for fTSCTiedToExecution which pauses the TSC while halting if true.
367 * Config variable: TSCNotTiedToHalt (bool) */
368 bool fTSCNotTiedToHalt;
369 /** Whether TM TSC mode switching is allowed at runtime. */
370 bool fTSCModeSwitchAllowed;
371 /** Whether the guest has enabled use of paravirtualized TSC. */
372 bool fParavirtTscEnabled;
373 /** The ID of the virtual CPU that normally runs the timers. */
374 VMCPUID idTimerCpu;
375
376 /** The number of CPU clock ticks per second (TMCLOCK_TSC).
377 * Config variable: TSCTicksPerSecond (64-bit unsigned int)
378 * The config variable implies @c enmTSCMode would be
379 * TMTSCMODE_VIRT_TSC_EMULATED. */
380 uint64_t cTSCTicksPerSecond;
381 /** The TSC difference introduced by pausing the VM. */
382 uint64_t offTSCPause;
383 /** The TSC value when the last TSC was paused. */
384 uint64_t u64LastPausedTSC;
385 /** CPU TSCs ticking indicator (one for each VCPU). */
386 uint32_t volatile cTSCsTicking;
387
388 /** Virtual time ticking enabled indicator (counter for each VCPU). (TMCLOCK_VIRTUAL) */
389 uint32_t volatile cVirtualTicking;
390 /** Virtual time is not running at 100%. */
391 bool fVirtualWarpDrive;
392 /** Virtual timer synchronous time ticking enabled indicator (bool). (TMCLOCK_VIRTUAL_SYNC) */
393 bool volatile fVirtualSyncTicking;
394 /** Virtual timer synchronous time catch-up active. */
395 bool volatile fVirtualSyncCatchUp;
396 /** Alignment padding. */
397 bool afAlignment1[1];
398 /** WarpDrive percentage.
399 * 100% is normal (fVirtualSyncNormal == true). When other than 100% we apply
400 * this percentage to the raw time source for the period it's been valid in,
401 * i.e. since u64VirtualWarpDriveStart. */
402 uint32_t u32VirtualWarpDrivePercentage;
403
404 /** The offset of the virtual clock relative to it's timesource.
405 * Only valid if fVirtualTicking is set. */
406 uint64_t u64VirtualOffset;
407 /** The guest virtual time when fVirtualTicking is cleared. */
408 uint64_t u64Virtual;
409 /** When the Warp drive was started or last adjusted.
410 * Only valid when fVirtualWarpDrive is set. */
411 uint64_t u64VirtualWarpDriveStart;
412 /** The previously returned nano TS.
413 * This handles TSC drift on SMP systems and expired interval.
414 * This is a valid range u64NanoTS to u64NanoTS + 1000000000 (ie. 1sec). */
415 uint64_t volatile u64VirtualRawPrev;
416 /** The ring-3 data structure for the RTTimeNanoTS workers used by tmVirtualGetRawNanoTS. */
417 RTTIMENANOTSDATAR3 VirtualGetRawDataR3;
418 /** The ring-0 data structure for the RTTimeNanoTS workers used by tmVirtualGetRawNanoTS. */
419 RTTIMENANOTSDATAR0 VirtualGetRawDataR0;
420 /** The ring-0 data structure for the RTTimeNanoTS workers used by tmVirtualGetRawNanoTS. */
421 RTTIMENANOTSDATARC VirtualGetRawDataRC;
422 /** Pointer to the ring-3 tmVirtualGetRawNanoTS worker function. */
423 R3PTRTYPE(PFNTIMENANOTSINTERNAL) pfnVirtualGetRawR3;
424 /** Pointer to the ring-0 tmVirtualGetRawNanoTS worker function. */
425 R0PTRTYPE(PFNTIMENANOTSINTERNAL) pfnVirtualGetRawR0;
426 /** Pointer to the raw-mode tmVirtualGetRawNanoTS worker function. */
427 RCPTRTYPE(PFNTIMENANOTSINTERNAL) pfnVirtualGetRawRC;
428 /** Alignment. */
429 RTRCPTR AlignmentRCPtr;
430 /** The guest virtual timer synchronous time when fVirtualSyncTicking is cleared.
431 * When fVirtualSyncTicking is set it holds the last time returned to
432 * the guest (while the lock was held). */
433 uint64_t volatile u64VirtualSync;
434 /** The offset of the timer synchronous virtual clock (TMCLOCK_VIRTUAL_SYNC) relative
435 * to the virtual clock (TMCLOCK_VIRTUAL).
436 * (This is accessed by the timer thread and must be updated atomically.) */
437 uint64_t volatile offVirtualSync;
438 /** The offset into offVirtualSync that's been irrevocably given up by failed catch-up attempts.
439 * Thus the current lag is offVirtualSync - offVirtualSyncGivenUp. */
440 uint64_t offVirtualSyncGivenUp;
441 /** The TMCLOCK_VIRTUAL at the previous TMVirtualGetSync call when catch-up is active. */
442 uint64_t volatile u64VirtualSyncCatchUpPrev;
443 /** The current catch-up percentage. */
444 uint32_t volatile u32VirtualSyncCatchUpPercentage;
445 /** How much slack when processing timers. */
446 uint32_t u32VirtualSyncScheduleSlack;
447 /** When to stop catch-up. */
448 uint64_t u64VirtualSyncCatchUpStopThreshold;
449 /** When to give up catch-up. */
450 uint64_t u64VirtualSyncCatchUpGiveUpThreshold;
451/** @def TM_MAX_CATCHUP_PERIODS
452 * The number of catchup rates. */
453#define TM_MAX_CATCHUP_PERIODS 10
454 /** The aggressiveness of the catch-up relative to how far we've lagged behind.
455 * The idea is to have increasing catch-up percentage as the lag increases. */
456 struct TMCATCHUPPERIOD
457 {
458 uint64_t u64Start; /**< When this period starts. (u64VirtualSyncOffset). */
459 uint32_t u32Percentage; /**< The catch-up percent to apply. */
460 uint32_t u32Alignment; /**< Structure alignment */
461 } aVirtualSyncCatchUpPeriods[TM_MAX_CATCHUP_PERIODS];
462
463 /** The current max timer Hz hint. */
464 uint32_t volatile uMaxHzHint;
465 /** Whether to recalulate the HzHint next time its queried. */
466 bool volatile fHzHintNeedsUpdating;
467 /** Alignment */
468 bool afAlignment2[3];
469 /** @cfgm{/TM/HostHzMax, uint32_t, Hz, 0, UINT32_MAX, 20000}
470 * The max host Hz frequency hint returned by TMCalcHostTimerFrequency. */
471 uint32_t cHostHzMax;
472 /** @cfgm{/TM/HostHzFudgeFactorTimerCpu, uint32_t, Hz, 0, UINT32_MAX, 111}
473 * The number of Hz TMCalcHostTimerFrequency adds for the timer CPU. */
474 uint32_t cPctHostHzFudgeFactorTimerCpu;
475 /** @cfgm{/TM/HostHzFudgeFactorOtherCpu, uint32_t, Hz, 0, UINT32_MAX, 110}
476 * The number of Hz TMCalcHostTimerFrequency adds for the other CPUs. */
477 uint32_t cPctHostHzFudgeFactorOtherCpu;
478 /** @cfgm{/TM/HostHzFudgeFactorCatchUp100, uint32_t, Hz, 0, UINT32_MAX, 300}
479 * The fudge factor (expressed in percent) that catch-up percentages below
480 * 100% is multiplied by. */
481 uint32_t cPctHostHzFudgeFactorCatchUp100;
482 /** @cfgm{/TM/HostHzFudgeFactorCatchUp200, uint32_t, Hz, 0, UINT32_MAX, 250}
483 * The fudge factor (expressed in percent) that catch-up percentages
484 * 100%-199% is multiplied by. */
485 uint32_t cPctHostHzFudgeFactorCatchUp200;
486 /** @cfgm{/TM/HostHzFudgeFactorCatchUp400, uint32_t, Hz, 0, UINT32_MAX, 200}
487 * The fudge factor (expressed in percent) that catch-up percentages
488 * 200%-399% is multiplied by. */
489 uint32_t cPctHostHzFudgeFactorCatchUp400;
490
491 /** The UTC offset in ns.
492 * This is *NOT* for converting UTC to local time. It is for converting real
493 * world UTC time to VM UTC time. This feature is indented for doing date
494 * testing of software and similar.
495 * @todo Implement warpdrive on UTC. */
496 int64_t offUTC;
497 /** The last value TMR3UtcNow returned. */
498 int64_t volatile nsLastUtcNow;
499 /** File to touch on UTC jump. */
500 R3PTRTYPE(char *) pszUtcTouchFileOnJump;
501 /** Just to avoid dealing with 32-bit alignment trouble. */
502 R3PTRTYPE(char *) pszAlignment2b;
503
504 /** Timer queues for the different clock types - R3 Ptr */
505 R3PTRTYPE(PTMTIMERQUEUE) paTimerQueuesR3;
506 /** Timer queues for the different clock types - R0 Ptr */
507 R0PTRTYPE(PTMTIMERQUEUE) paTimerQueuesR0;
508 /** Timer queues for the different clock types - RC Ptr */
509 RCPTRTYPE(PTMTIMERQUEUE) paTimerQueuesRC;
510
511 /** Pointer to our RC mapping of the GIP. */
512 RCPTRTYPE(void *) pvGIPRC;
513 /** Pointer to our R3 mapping of the GIP. */
514 R3PTRTYPE(void *) pvGIPR3;
515
516 /** Pointer to a singly linked list of free timers.
517 * This chain is using the TMTIMER::pBigNext members.
518 * Only accessible from the emulation thread. */
519 PTMTIMERR3 pFree;
520
521 /** Pointer to a doubly linked list of created timers.
522 * This chain is using the TMTIMER::pBigNext and TMTIMER::pBigPrev members.
523 * Only accessible from the emulation thread. */
524 PTMTIMERR3 pCreated;
525
526 /** The schedule timer timer handle (runtime timer).
527 * This timer will do frequent check on pending queue schedules and
528 * raise VM_FF_TIMER to pull EMTs attention to them.
529 */
530 R3PTRTYPE(PRTTIMER) pTimer;
531 /** Interval in milliseconds of the pTimer timer. */
532 uint32_t u32TimerMillies;
533
534 /** Indicates that queues are being run. */
535 bool volatile fRunningQueues;
536 /** Indicates that the virtual sync queue is being run. */
537 bool volatile fRunningVirtualSyncQueue;
538 /** Alignment */
539 bool afAlignment3[2];
540
541 /** Lock serializing access to the timer lists. */
542 PDMCRITSECT TimerCritSect;
543 /** Lock serializing access to the VirtualSync clock and the associated
544 * timer queue. */
545 PDMCRITSECT VirtualSyncLock;
546
547 /** CPU load state for all the virtual CPUs (tmR3CpuLoadTimer). */
548 TMCPULOADSTATE CpuLoad;
549
550 /** TMR3TimerQueuesDo
551 * @{ */
552 STAMPROFILE StatDoQueues;
553 STAMPROFILEADV aStatDoQueues[TMCLOCK_MAX];
554 /** @} */
555 /** tmSchedule
556 * @{ */
557 STAMPROFILE StatScheduleOneRZ;
558 STAMPROFILE StatScheduleOneR3;
559 STAMCOUNTER StatScheduleSetFF;
560 STAMCOUNTER StatPostponedR3;
561 STAMCOUNTER StatPostponedRZ;
562 /** @} */
563 /** Read the time
564 * @{ */
565 STAMCOUNTER StatVirtualGet;
566 STAMCOUNTER StatVirtualGetSetFF;
567 STAMCOUNTER StatVirtualSyncGet;
568 STAMCOUNTER StatVirtualSyncGetAdjLast;
569 STAMCOUNTER StatVirtualSyncGetELoop;
570 STAMCOUNTER StatVirtualSyncGetExpired;
571 STAMCOUNTER StatVirtualSyncGetLockless;
572 STAMCOUNTER StatVirtualSyncGetLocked;
573 STAMCOUNTER StatVirtualSyncGetSetFF;
574 STAMCOUNTER StatVirtualPause;
575 STAMCOUNTER StatVirtualResume;
576 /** @} */
577 /** TMTimerPoll
578 * @{ */
579 STAMCOUNTER StatPoll;
580 STAMCOUNTER StatPollAlreadySet;
581 STAMCOUNTER StatPollELoop;
582 STAMCOUNTER StatPollMiss;
583 STAMCOUNTER StatPollRunning;
584 STAMCOUNTER StatPollSimple;
585 STAMCOUNTER StatPollVirtual;
586 STAMCOUNTER StatPollVirtualSync;
587 /** @} */
588 /** TMTimerSet sans virtual sync timers.
589 * @{ */
590 STAMCOUNTER StatTimerSet;
591 STAMCOUNTER StatTimerSetOpt;
592 STAMPROFILE StatTimerSetRZ;
593 STAMPROFILE StatTimerSetR3;
594 STAMCOUNTER StatTimerSetStStopped;
595 STAMCOUNTER StatTimerSetStExpDeliver;
596 STAMCOUNTER StatTimerSetStActive;
597 STAMCOUNTER StatTimerSetStPendStop;
598 STAMCOUNTER StatTimerSetStPendStopSched;
599 STAMCOUNTER StatTimerSetStPendSched;
600 STAMCOUNTER StatTimerSetStPendResched;
601 STAMCOUNTER StatTimerSetStOther;
602 /** @} */
603 /** TMTimerSet on virtual sync timers.
604 * @{ */
605 STAMCOUNTER StatTimerSetVs;
606 STAMPROFILE StatTimerSetVsRZ;
607 STAMPROFILE StatTimerSetVsR3;
608 STAMCOUNTER StatTimerSetVsStStopped;
609 STAMCOUNTER StatTimerSetVsStExpDeliver;
610 STAMCOUNTER StatTimerSetVsStActive;
611 /** @} */
612 /** TMTimerSetRelative sans virtual sync timers
613 * @{ */
614 STAMCOUNTER StatTimerSetRelative;
615 STAMPROFILE StatTimerSetRelativeRZ;
616 STAMPROFILE StatTimerSetRelativeR3;
617 STAMCOUNTER StatTimerSetRelativeOpt;
618 STAMCOUNTER StatTimerSetRelativeStStopped;
619 STAMCOUNTER StatTimerSetRelativeStExpDeliver;
620 STAMCOUNTER StatTimerSetRelativeStActive;
621 STAMCOUNTER StatTimerSetRelativeStPendStop;
622 STAMCOUNTER StatTimerSetRelativeStPendStopSched;
623 STAMCOUNTER StatTimerSetRelativeStPendSched;
624 STAMCOUNTER StatTimerSetRelativeStPendResched;
625 STAMCOUNTER StatTimerSetRelativeStOther;
626 /** @} */
627 /** TMTimerSetRelative on virtual sync timers.
628 * @{ */
629 STAMCOUNTER StatTimerSetRelativeVs;
630 STAMPROFILE StatTimerSetRelativeVsRZ;
631 STAMPROFILE StatTimerSetRelativeVsR3;
632 STAMCOUNTER StatTimerSetRelativeVsStStopped;
633 STAMCOUNTER StatTimerSetRelativeVsStExpDeliver;
634 STAMCOUNTER StatTimerSetRelativeVsStActive;
635 /** @} */
636 /** TMTimerStop sans virtual sync.
637 * @{ */
638 STAMPROFILE StatTimerStopRZ;
639 STAMPROFILE StatTimerStopR3;
640 /** @} */
641 /** TMTimerStop on virtual sync timers.
642 * @{ */
643 STAMPROFILE StatTimerStopVsRZ;
644 STAMPROFILE StatTimerStopVsR3;
645 /** @} */
646 /** VirtualSync - Running and Catching Up
647 * @{ */
648 STAMCOUNTER StatVirtualSyncRun;
649 STAMCOUNTER StatVirtualSyncRunRestart;
650 STAMPROFILE StatVirtualSyncRunSlack;
651 STAMCOUNTER StatVirtualSyncRunStop;
652 STAMCOUNTER StatVirtualSyncRunStoppedAlready;
653 STAMCOUNTER StatVirtualSyncGiveUp;
654 STAMCOUNTER StatVirtualSyncGiveUpBeforeStarting;
655 STAMPROFILEADV StatVirtualSyncCatchup;
656 STAMCOUNTER aStatVirtualSyncCatchupInitial[TM_MAX_CATCHUP_PERIODS];
657 STAMCOUNTER aStatVirtualSyncCatchupAdjust[TM_MAX_CATCHUP_PERIODS];
658 /** @} */
659 /** TMR3VirtualSyncFF (non dedicated EMT). */
660 STAMPROFILE StatVirtualSyncFF;
661 /** The timer callback. */
662 STAMCOUNTER StatTimerCallbackSetFF;
663 STAMCOUNTER StatTimerCallback;
664
665 /** Calls to TMCpuTickSet. */
666 STAMCOUNTER StatTSCSet;
667
668 /** TSC starts and stops. */
669 STAMCOUNTER StatTSCPause;
670 STAMCOUNTER StatTSCResume;
671
672 /** @name Reasons for refusing TSC offsetting in TMCpuTickCanUseRealTSC.
673 * @{ */
674 STAMCOUNTER StatTSCNotFixed;
675 STAMCOUNTER StatTSCNotTicking;
676 STAMCOUNTER StatTSCCatchupLE010;
677 STAMCOUNTER StatTSCCatchupLE025;
678 STAMCOUNTER StatTSCCatchupLE100;
679 STAMCOUNTER StatTSCCatchupOther;
680 STAMCOUNTER StatTSCWarp;
681 STAMCOUNTER StatTSCUnderflow;
682 STAMCOUNTER StatTSCSyncNotTicking;
683 /** @} */
684} TM;
685/** Pointer to TM VM instance data. */
686typedef TM *PTM;
687
688/**
689 * TM VMCPU Instance data.
690 * Changes to this must checked against the padding of the tm union in VM!
691 */
692typedef struct TMCPU
693{
694 /** Offset to the VMCPU structure.
695 * See TMCPU2VM(). */
696 RTUINT offVMCPU;
697
698 /** CPU timestamp ticking enabled indicator (bool). (RDTSC) */
699 bool fTSCTicking;
700 bool afAlignment0[3]; /**< alignment padding */
701
702 /** The offset between the host tick (TSC/virtual depending on the TSC mode) and
703 * the guest tick. */
704 uint64_t offTSCRawSrc;
705
706 /** The guest TSC when fTicking is cleared. */
707 uint64_t u64TSC;
708
709 /** The last seen TSC by the guest. */
710 uint64_t u64TSCLastSeen;
711
712#ifndef VBOX_WITHOUT_NS_ACCOUNTING
713 /** The nanosecond timestamp of the CPU start or resume.
714 * This is recalculated when the VM is started so that
715 * cNsTotal = RTTimeNanoTS() - u64NsTsStartCpu. */
716 uint64_t u64NsTsStartTotal;
717 /** The nanosecond timestamp of the last start-execute notification. */
718 uint64_t u64NsTsStartExecuting;
719 /** The nanosecond timestamp of the last start-halt notification. */
720 uint64_t u64NsTsStartHalting;
721 /** The cNsXXX generation. */
722 uint32_t volatile uTimesGen;
723 /** Explicit alignment padding. */
724 uint32_t u32Alignment;
725 /** The number of nanoseconds total run time.
726 * @remarks This is updated when cNsExecuting and cNsHalted are updated. */
727 uint64_t cNsTotal;
728 /** The number of nanoseconds spent executing. */
729 uint64_t cNsExecuting;
730 /** The number of nanoseconds being halted. */
731 uint64_t cNsHalted;
732 /** The number of nanoseconds spent on other things.
733 * @remarks This is updated when cNsExecuting and cNsHalted are updated. */
734 uint64_t cNsOther;
735 /** The number of halts. */
736 uint64_t cPeriodsHalted;
737 /** The number of guest execution runs. */
738 uint64_t cPeriodsExecuting;
739# if defined(VBOX_WITH_STATISTICS) || defined(VBOX_WITH_NS_ACCOUNTING_STATS)
740 /** Resettable version of cNsTotal. */
741 STAMCOUNTER StatNsTotal;
742 /** Resettable version of cNsExecuting. */
743 STAMPROFILE StatNsExecuting;
744 /** Long execution intervals. */
745 STAMPROFILE StatNsExecLong;
746 /** Short execution intervals . */
747 STAMPROFILE StatNsExecShort;
748 /** Tiny execution intervals . */
749 STAMPROFILE StatNsExecTiny;
750 /** Resettable version of cNsHalted. */
751 STAMPROFILE StatNsHalted;
752 /** Resettable version of cNsOther. */
753 STAMPROFILE StatNsOther;
754# endif
755
756 /** CPU load state for this virtual CPU (tmR3CpuLoadTimer). */
757 TMCPULOADSTATE CpuLoad;
758#endif
759} TMCPU;
760/** Pointer to TM VMCPU instance data. */
761typedef TMCPU *PTMCPU;
762
763const char *tmTimerState(TMTIMERSTATE enmState);
764void tmTimerQueueSchedule(PVM pVM, PTMTIMERQUEUE pQueue);
765#ifdef VBOX_STRICT
766void tmTimerQueuesSanityChecks(PVM pVM, const char *pszWhere);
767#endif
768
769uint64_t tmR3CpuTickGetRawVirtualNoCheck(PVM pVM);
770int tmCpuTickPause(PVMCPU pVCpu);
771int tmCpuTickPauseLocked(PVM pVM, PVMCPU pVCpu);
772int tmCpuTickResume(PVM pVM, PVMCPU pVCpu);
773int tmCpuTickResumeLocked(PVM pVM, PVMCPU pVCpu);
774
775int tmVirtualPauseLocked(PVM pVM);
776int tmVirtualResumeLocked(PVM pVM);
777DECLCALLBACK(DECLEXPORT(void)) tmVirtualNanoTSBad(PRTTIMENANOTSDATA pData, uint64_t u64NanoTS,
778 uint64_t u64DeltaPrev, uint64_t u64PrevNanoTS);
779DECLCALLBACK(DECLEXPORT(uint64_t)) tmVirtualNanoTSRediscover(PRTTIMENANOTSDATA pData);
780DECLCALLBACK(DECLEXPORT(uint64_t)) tmVirtualNanoTSBadCpuIndex(PRTTIMENANOTSDATA pData, uint16_t idApic,
781 uint16_t iCpuSet, uint16_t iGipCpu);
782
783/**
784 * Try take the timer lock, wait in ring-3 return VERR_SEM_BUSY in R0/RC.
785 *
786 * @retval VINF_SUCCESS on success (always in ring-3).
787 * @retval VERR_SEM_BUSY in RC and R0 if the semaphore is busy.
788 *
789 * @param a_pVM Pointer to the VM.
790 *
791 * @remarks The virtual sync timer queue requires the virtual sync lock.
792 */
793#define TM_LOCK_TIMERS(a_pVM) PDMCritSectEnter(&(a_pVM)->tm.s.TimerCritSect, VERR_SEM_BUSY)
794
795/**
796 * Try take the timer lock, no waiting.
797 *
798 * @retval VINF_SUCCESS on success.
799 * @retval VERR_SEM_BUSY if busy.
800 *
801 * @param a_pVM Pointer to the VM.
802 *
803 * @remarks The virtual sync timer queue requires the virtual sync lock.
804 */
805#define TM_TRY_LOCK_TIMERS(a_pVM) PDMCritSectTryEnter(&(a_pVM)->tm.s.TimerCritSect)
806
807/** Lock the timers (sans the virtual sync queue). */
808#define TM_UNLOCK_TIMERS(a_pVM) do { PDMCritSectLeave(&(a_pVM)->tm.s.TimerCritSect); } while (0)
809
810/** Checks that the caller owns the timer lock. */
811#define TM_ASSERT_TIMER_LOCK_OWNERSHIP(a_pVM) \
812 Assert(PDMCritSectIsOwner(&(a_pVM)->tm.s.TimerCritSect))
813
814/** @} */
815
816RT_C_DECLS_END
817
818#endif
819
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette