VirtualBox

source: vbox/trunk/src/VBox/Storage/VD.cpp@ 44412

Last change on this file since 44412 was 44412, checked in by vboxsync, 12 years ago

Storage: Merge sync/async interface for the cache

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 350.5 KB
Line 
1/* $Id: VD.cpp 44412 2013-01-28 10:39:34Z vboxsync $ */
2/** @file
3 * VBoxHDD - VBox HDD Container implementation.
4 */
5
6/*
7 * Copyright (C) 2006-2013 Oracle Corporation
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 */
17
18/*******************************************************************************
19* Header Files *
20*******************************************************************************/
21#define LOG_GROUP LOG_GROUP_VD
22#include <VBox/vd.h>
23#include <VBox/err.h>
24#include <VBox/sup.h>
25#include <VBox/log.h>
26
27#include <iprt/alloc.h>
28#include <iprt/assert.h>
29#include <iprt/uuid.h>
30#include <iprt/file.h>
31#include <iprt/string.h>
32#include <iprt/asm.h>
33#include <iprt/ldr.h>
34#include <iprt/dir.h>
35#include <iprt/path.h>
36#include <iprt/param.h>
37#include <iprt/memcache.h>
38#include <iprt/sg.h>
39#include <iprt/list.h>
40#include <iprt/avl.h>
41#include <iprt/semaphore.h>
42
43#include <VBox/vd-plugin.h>
44#include <VBox/vd-cache-plugin.h>
45
46/** Disable dynamic backends on non x86 architectures. This feature
47 * requires the SUPR3 library which is not available there.
48 */
49#if !defined(VBOX_HDD_NO_DYNAMIC_BACKENDS) && !defined(RT_ARCH_X86) && !defined(RT_ARCH_AMD64)
50# define VBOX_HDD_NO_DYNAMIC_BACKENDS
51#endif
52
53#define VBOXHDDDISK_SIGNATURE 0x6f0e2a7d
54
55/** Buffer size used for merging images. */
56#define VD_MERGE_BUFFER_SIZE (16 * _1M)
57
58/** Maximum number of segments in one I/O task. */
59#define VD_IO_TASK_SEGMENTS_MAX 64
60
61/** Threshold after not recently used blocks are removed from the list. */
62#define VD_DISCARD_REMOVE_THRESHOLD (10 * _1M) /** @todo: experiment */
63
64/**
65 * VD async I/O interface storage descriptor.
66 */
67typedef struct VDIIOFALLBACKSTORAGE
68{
69 /** File handle. */
70 RTFILE File;
71 /** Completion callback. */
72 PFNVDCOMPLETED pfnCompleted;
73 /** Thread for async access. */
74 RTTHREAD ThreadAsync;
75} VDIIOFALLBACKSTORAGE, *PVDIIOFALLBACKSTORAGE;
76
77/**
78 * Structure containing everything I/O related
79 * for the image and cache descriptors.
80 */
81typedef struct VDIO
82{
83 /** I/O interface to the upper layer. */
84 PVDINTERFACEIO pInterfaceIo;
85
86 /** Per image internal I/O interface. */
87 VDINTERFACEIOINT VDIfIoInt;
88
89 /** Fallback I/O interface, only used if the caller doesn't provide it. */
90 VDINTERFACEIO VDIfIo;
91
92 /** Opaque backend data. */
93 void *pBackendData;
94 /** Disk this image is part of */
95 PVBOXHDD pDisk;
96 /** Flag whether to ignore flush requests. */
97 bool fIgnoreFlush;
98} VDIO, *PVDIO;
99
100/** Forward declaration of an I/O task */
101typedef struct VDIOTASK *PVDIOTASK;
102
103/**
104 * VBox HDD Container image descriptor.
105 */
106typedef struct VDIMAGE
107{
108 /** Link to parent image descriptor, if any. */
109 struct VDIMAGE *pPrev;
110 /** Link to child image descriptor, if any. */
111 struct VDIMAGE *pNext;
112 /** Container base filename. (UTF-8) */
113 char *pszFilename;
114 /** Data managed by the backend which keeps the actual info. */
115 void *pBackendData;
116 /** Cached sanitized image flags. */
117 unsigned uImageFlags;
118 /** Image open flags (only those handled generically in this code and which
119 * the backends will never ever see). */
120 unsigned uOpenFlags;
121
122 /** Function pointers for the various backend methods. */
123 PCVBOXHDDBACKEND Backend;
124 /** Pointer to list of VD interfaces, per-image. */
125 PVDINTERFACE pVDIfsImage;
126 /** I/O related things. */
127 VDIO VDIo;
128} VDIMAGE, *PVDIMAGE;
129
130/**
131 * uModified bit flags.
132 */
133#define VD_IMAGE_MODIFIED_FLAG RT_BIT(0)
134#define VD_IMAGE_MODIFIED_FIRST RT_BIT(1)
135#define VD_IMAGE_MODIFIED_DISABLE_UUID_UPDATE RT_BIT(2)
136
137
138/**
139 * VBox HDD Cache image descriptor.
140 */
141typedef struct VDCACHE
142{
143 /** Cache base filename. (UTF-8) */
144 char *pszFilename;
145 /** Data managed by the backend which keeps the actual info. */
146 void *pBackendData;
147 /** Cached sanitized image flags. */
148 unsigned uImageFlags;
149 /** Image open flags (only those handled generically in this code and which
150 * the backends will never ever see). */
151 unsigned uOpenFlags;
152
153 /** Function pointers for the various backend methods. */
154 PCVDCACHEBACKEND Backend;
155
156 /** Pointer to list of VD interfaces, per-cache. */
157 PVDINTERFACE pVDIfsCache;
158 /** I/O related things. */
159 VDIO VDIo;
160} VDCACHE, *PVDCACHE;
161
162/**
163 * A block waiting for a discard.
164 */
165typedef struct VDDISCARDBLOCK
166{
167 /** AVL core. */
168 AVLRU64NODECORE Core;
169 /** LRU list node. */
170 RTLISTNODE NodeLru;
171 /** Number of bytes to discard. */
172 size_t cbDiscard;
173 /** Bitmap of allocated sectors. */
174 void *pbmAllocated;
175} VDDISCARDBLOCK, *PVDDISCARDBLOCK;
176
177/**
178 * VD discard state.
179 */
180typedef struct VDDISCARDSTATE
181{
182 /** Number of bytes waiting for a discard. */
183 size_t cbDiscarding;
184 /** AVL tree with blocks waiting for a discard.
185 * The uOffset + cbDiscard range is the search key. */
186 PAVLRU64TREE pTreeBlocks;
187 /** LRU list of the least frequently discarded blocks.
188 * If there are to many blocks waiting the least frequently used
189 * will be removed and the range will be set to 0.
190 */
191 RTLISTNODE ListLru;
192} VDDISCARDSTATE, *PVDDISCARDSTATE;
193
194/**
195 * VBox HDD Container main structure, private part.
196 */
197struct VBOXHDD
198{
199 /** Structure signature (VBOXHDDDISK_SIGNATURE). */
200 uint32_t u32Signature;
201
202 /** Image type. */
203 VDTYPE enmType;
204
205 /** Number of opened images. */
206 unsigned cImages;
207
208 /** Base image. */
209 PVDIMAGE pBase;
210
211 /** Last opened image in the chain.
212 * The same as pBase if only one image is used. */
213 PVDIMAGE pLast;
214
215 /** If a merge to one of the parents is running this may be non-NULL
216 * to indicate to what image the writes should be additionally relayed. */
217 PVDIMAGE pImageRelay;
218
219 /** Flags representing the modification state. */
220 unsigned uModified;
221
222 /** Cached size of this disk. */
223 uint64_t cbSize;
224 /** Cached PCHS geometry for this disk. */
225 VDGEOMETRY PCHSGeometry;
226 /** Cached LCHS geometry for this disk. */
227 VDGEOMETRY LCHSGeometry;
228
229 /** Pointer to list of VD interfaces, per-disk. */
230 PVDINTERFACE pVDIfsDisk;
231 /** Pointer to the common interface structure for error reporting. */
232 PVDINTERFACEERROR pInterfaceError;
233 /** Pointer to the optional thread synchronization callbacks. */
234 PVDINTERFACETHREADSYNC pInterfaceThreadSync;
235
236 /** Memory cache for I/O contexts */
237 RTMEMCACHE hMemCacheIoCtx;
238 /** Memory cache for I/O tasks. */
239 RTMEMCACHE hMemCacheIoTask;
240 /** An I/O context is currently using the disk structures
241 * Every I/O context must be placed on one of the lists below. */
242 volatile bool fLocked;
243 /** Head of pending I/O tasks waiting for completion - LIFO order. */
244 volatile PVDIOTASK pIoTasksPendingHead;
245 /** Head of newly queued I/O contexts - LIFO order. */
246 volatile PVDIOCTX pIoCtxHead;
247 /** Head of halted I/O contexts which are given back to generic
248 * disk framework by the backend. - LIFO order. */
249 volatile PVDIOCTX pIoCtxHaltedHead;
250
251 /** Head of blocked I/O contexts, processed only
252 * after pIoCtxLockOwner was freed - LIFO order. */
253 volatile PVDIOCTX pIoCtxBlockedHead;
254 /** I/O context which locked the disk for a growing write or flush request.
255 * Other flush or growing write requests need to wait until
256 * the current one completes. - NIL_VDIOCTX if unlocked. */
257 volatile PVDIOCTX pIoCtxLockOwner;
258
259 /** Pointer to the L2 disk cache if any. */
260 PVDCACHE pCache;
261 /** Pointer to the discard state if any. */
262 PVDDISCARDSTATE pDiscard;
263
264 /** Event semaphore for synchronous I/O. */
265 RTSEMEVENT hEventSemSyncIo;
266 /** Status code of the last synchronous I/O request. */
267 int rcSync;
268};
269
270# define VD_IS_LOCKED(a_pDisk) \
271 do \
272 { \
273 AssertMsg(a_pDisk->fLocked, \
274 ("Lock not held\n"));\
275 } while(0)
276
277/**
278 * VBox parent read descriptor, used internally for compaction.
279 */
280typedef struct VDPARENTSTATEDESC
281{
282 /** Pointer to disk descriptor. */
283 PVBOXHDD pDisk;
284 /** Pointer to image descriptor. */
285 PVDIMAGE pImage;
286} VDPARENTSTATEDESC, *PVDPARENTSTATEDESC;
287
288/**
289 * Transfer direction.
290 */
291typedef enum VDIOCTXTXDIR
292{
293 /** Read */
294 VDIOCTXTXDIR_READ = 0,
295 /** Write */
296 VDIOCTXTXDIR_WRITE,
297 /** Flush */
298 VDIOCTXTXDIR_FLUSH,
299 /** Discard */
300 VDIOCTXTXDIR_DISCARD,
301 /** 32bit hack */
302 VDIOCTXTXDIR_32BIT_HACK = 0x7fffffff
303} VDIOCTXTXDIR, *PVDIOCTXTXDIR;
304
305/** Transfer function */
306typedef DECLCALLBACK(int) FNVDIOCTXTRANSFER (PVDIOCTX pIoCtx);
307/** Pointer to a transfer function. */
308typedef FNVDIOCTXTRANSFER *PFNVDIOCTXTRANSFER;
309
310/**
311 * I/O context
312 */
313typedef struct VDIOCTX
314{
315 /** Pointer to the next I/O context. */
316 struct VDIOCTX * volatile pIoCtxNext;
317 /** Disk this is request is for. */
318 PVBOXHDD pDisk;
319 /** Return code. */
320 int rcReq;
321 /** Various flags for the I/O context. */
322 uint32_t fFlags;
323 /** Number of data transfers currently pending. */
324 volatile uint32_t cDataTransfersPending;
325 /** How many meta data transfers are pending. */
326 volatile uint32_t cMetaTransfersPending;
327 /** Flag whether the request finished */
328 volatile bool fComplete;
329 /** Temporary allocated memory which is freed
330 * when the context completes. */
331 void *pvAllocation;
332 /** Transfer function. */
333 PFNVDIOCTXTRANSFER pfnIoCtxTransfer;
334 /** Next transfer part after the current one completed. */
335 PFNVDIOCTXTRANSFER pfnIoCtxTransferNext;
336 /** Transfer direction */
337 VDIOCTXTXDIR enmTxDir;
338 /** Request type dependent data. */
339 union
340 {
341 /** I/O request (read/write). */
342 struct
343 {
344 /** Number of bytes left until this context completes. */
345 volatile uint32_t cbTransferLeft;
346 /** Current offset */
347 volatile uint64_t uOffset;
348 /** Number of bytes to transfer */
349 volatile size_t cbTransfer;
350 /** Current image in the chain. */
351 PVDIMAGE pImageCur;
352 /** Start image to read from. pImageCur is reset to this
353 * value after it reached the first image in the chain. */
354 PVDIMAGE pImageStart;
355 /** S/G buffer */
356 RTSGBUF SgBuf;
357 } Io;
358 /** Discard requests. */
359 struct
360 {
361 /** Pointer to the range descriptor array. */
362 PCRTRANGE paRanges;
363 /** Number of ranges in the array. */
364 unsigned cRanges;
365 /** Range descriptor index which is processed. */
366 unsigned idxRange;
367 /** Start offset to discard currently. */
368 uint64_t offCur;
369 /** How many bytes left to discard in the current range. */
370 size_t cbDiscardLeft;
371 /** How many bytes to discard in the current block (<= cbDiscardLeft). */
372 size_t cbThisDiscard;
373 /** Discard block handled currently. */
374 PVDDISCARDBLOCK pBlock;
375 } Discard;
376 } Req;
377 /** Parent I/O context if any. Sets the type of the context (root/child) */
378 PVDIOCTX pIoCtxParent;
379 /** Type dependent data (root/child) */
380 union
381 {
382 /** Root data */
383 struct
384 {
385 /** Completion callback */
386 PFNVDASYNCTRANSFERCOMPLETE pfnComplete;
387 /** User argument 1 passed on completion. */
388 void *pvUser1;
389 /** User argument 2 passed on completion. */
390 void *pvUser2;
391 } Root;
392 /** Child data */
393 struct
394 {
395 /** Saved start offset */
396 uint64_t uOffsetSaved;
397 /** Saved transfer size */
398 size_t cbTransferLeftSaved;
399 /** Number of bytes transferred from the parent if this context completes. */
400 size_t cbTransferParent;
401 /** Number of bytes to pre read */
402 size_t cbPreRead;
403 /** Number of bytes to post read. */
404 size_t cbPostRead;
405 /** Number of bytes to write left in the parent. */
406 size_t cbWriteParent;
407 /** Write type dependent data. */
408 union
409 {
410 /** Optimized */
411 struct
412 {
413 /** Bytes to fill to satisfy the block size. Not part of the virtual disk. */
414 size_t cbFill;
415 /** Bytes to copy instead of reading from the parent */
416 size_t cbWriteCopy;
417 /** Bytes to read from the image. */
418 size_t cbReadImage;
419 } Optimized;
420 } Write;
421 } Child;
422 } Type;
423} VDIOCTX;
424
425/** Flag whether the context is blocked. */
426#define VDIOCTX_FLAGS_BLOCKED RT_BIT_32(0)
427/** Flag whether the I/O context is using synchronous I/O. */
428#define VDIOCTX_FLAGS_SYNC RT_BIT_32(1)
429
430/** NIL I/O context pointer value. */
431#define NIL_VDIOCTX ((PVDIOCTX)0)
432
433/**
434 * List node for deferred I/O contexts.
435 */
436typedef struct VDIOCTXDEFERRED
437{
438 /** Node in the list of deferred requests.
439 * A request can be deferred if the image is growing
440 * and the request accesses the same range or if
441 * the backend needs to read or write metadata from the disk
442 * before it can continue. */
443 RTLISTNODE NodeDeferred;
444 /** I/O context this entry points to. */
445 PVDIOCTX pIoCtx;
446} VDIOCTXDEFERRED, *PVDIOCTXDEFERRED;
447
448/**
449 * I/O task.
450 */
451typedef struct VDIOTASK
452{
453 /** Next I/O task waiting in the list. */
454 struct VDIOTASK * volatile pNext;
455 /** Storage this task belongs to. */
456 PVDIOSTORAGE pIoStorage;
457 /** Optional completion callback. */
458 PFNVDXFERCOMPLETED pfnComplete;
459 /** Opaque user data. */
460 void *pvUser;
461 /** Completion status code for the task. */
462 int rcReq;
463 /** Flag whether this is a meta data transfer. */
464 bool fMeta;
465 /** Type dependent data. */
466 union
467 {
468 /** User data transfer. */
469 struct
470 {
471 /** Number of bytes this task transferred. */
472 uint32_t cbTransfer;
473 /** Pointer to the I/O context the task belongs. */
474 PVDIOCTX pIoCtx;
475 } User;
476 /** Meta data transfer. */
477 struct
478 {
479 /** Meta transfer this task is for. */
480 PVDMETAXFER pMetaXfer;
481 } Meta;
482 } Type;
483} VDIOTASK;
484
485/**
486 * Storage handle.
487 */
488typedef struct VDIOSTORAGE
489{
490 /** Image I/O state this storage handle belongs to. */
491 PVDIO pVDIo;
492 /** AVL tree for pending async metadata transfers. */
493 PAVLRFOFFTREE pTreeMetaXfers;
494 /** Storage handle */
495 void *pStorage;
496} VDIOSTORAGE;
497
498/**
499 * Metadata transfer.
500 *
501 * @note This entry can't be freed if either the list is not empty or
502 * the reference counter is not 0.
503 * The assumption is that the backends don't need to read huge amounts of
504 * metadata to complete a transfer so the additional memory overhead should
505 * be relatively small.
506 */
507typedef struct VDMETAXFER
508{
509 /** AVL core for fast search (the file offset is the key) */
510 AVLRFOFFNODECORE Core;
511 /** I/O storage for this transfer. */
512 PVDIOSTORAGE pIoStorage;
513 /** Flags. */
514 uint32_t fFlags;
515 /** List of I/O contexts waiting for this metadata transfer to complete. */
516 RTLISTNODE ListIoCtxWaiting;
517 /** Number of references to this entry. */
518 unsigned cRefs;
519 /** Size of the data stored with this entry. */
520 size_t cbMeta;
521 /** Data stored - variable size. */
522 uint8_t abData[1];
523} VDMETAXFER;
524
525/**
526 * The transfer direction for the metadata.
527 */
528#define VDMETAXFER_TXDIR_MASK 0x3
529#define VDMETAXFER_TXDIR_NONE 0x0
530#define VDMETAXFER_TXDIR_WRITE 0x1
531#define VDMETAXFER_TXDIR_READ 0x2
532#define VDMETAXFER_TXDIR_FLUSH 0x3
533#define VDMETAXFER_TXDIR_GET(flags) ((flags) & VDMETAXFER_TXDIR_MASK)
534#define VDMETAXFER_TXDIR_SET(flags, dir) ((flags) = (flags & ~VDMETAXFER_TXDIR_MASK) | (dir))
535
536extern VBOXHDDBACKEND g_RawBackend;
537extern VBOXHDDBACKEND g_VmdkBackend;
538extern VBOXHDDBACKEND g_VDIBackend;
539extern VBOXHDDBACKEND g_VhdBackend;
540extern VBOXHDDBACKEND g_ParallelsBackend;
541extern VBOXHDDBACKEND g_DmgBackend;
542extern VBOXHDDBACKEND g_ISCSIBackend;
543extern VBOXHDDBACKEND g_QedBackend;
544extern VBOXHDDBACKEND g_QCowBackend;
545extern VBOXHDDBACKEND g_VhdxBackend;
546
547static unsigned g_cBackends = 0;
548static PVBOXHDDBACKEND *g_apBackends = NULL;
549static PVBOXHDDBACKEND aStaticBackends[] =
550{
551 &g_VmdkBackend,
552 &g_VDIBackend,
553 &g_VhdBackend,
554 &g_ParallelsBackend,
555 &g_DmgBackend,
556 &g_QedBackend,
557 &g_QCowBackend,
558 &g_VhdxBackend,
559 &g_RawBackend,
560 &g_ISCSIBackend
561};
562
563/**
564 * Supported backends for the disk cache.
565 */
566extern VDCACHEBACKEND g_VciCacheBackend;
567
568static unsigned g_cCacheBackends = 0;
569static PVDCACHEBACKEND *g_apCacheBackends = NULL;
570static PVDCACHEBACKEND aStaticCacheBackends[] =
571{
572 &g_VciCacheBackend
573};
574
575/** Forward declaration of the async discard helper. */
576static int vdDiscardHelperAsync(PVDIOCTX pIoCtx);
577static void vdDiskProcessBlockedIoCtx(PVBOXHDD pDisk);
578static int vdDiskUnlock(PVBOXHDD pDisk, PVDIOCTX pIoCtxRc);
579
580/**
581 * internal: add several backends.
582 */
583static int vdAddBackends(PVBOXHDDBACKEND *ppBackends, unsigned cBackends)
584{
585 PVBOXHDDBACKEND *pTmp = (PVBOXHDDBACKEND*)RTMemRealloc(g_apBackends,
586 (g_cBackends + cBackends) * sizeof(PVBOXHDDBACKEND));
587 if (RT_UNLIKELY(!pTmp))
588 return VERR_NO_MEMORY;
589 g_apBackends = pTmp;
590 memcpy(&g_apBackends[g_cBackends], ppBackends, cBackends * sizeof(PVBOXHDDBACKEND));
591 g_cBackends += cBackends;
592 return VINF_SUCCESS;
593}
594
595/**
596 * internal: add single backend.
597 */
598DECLINLINE(int) vdAddBackend(PVBOXHDDBACKEND pBackend)
599{
600 return vdAddBackends(&pBackend, 1);
601}
602
603/**
604 * internal: add several cache backends.
605 */
606static int vdAddCacheBackends(PVDCACHEBACKEND *ppBackends, unsigned cBackends)
607{
608 PVDCACHEBACKEND *pTmp = (PVDCACHEBACKEND*)RTMemRealloc(g_apCacheBackends,
609 (g_cCacheBackends + cBackends) * sizeof(PVDCACHEBACKEND));
610 if (RT_UNLIKELY(!pTmp))
611 return VERR_NO_MEMORY;
612 g_apCacheBackends = pTmp;
613 memcpy(&g_apCacheBackends[g_cCacheBackends], ppBackends, cBackends * sizeof(PVDCACHEBACKEND));
614 g_cCacheBackends += cBackends;
615 return VINF_SUCCESS;
616}
617
618/**
619 * internal: add single cache backend.
620 */
621DECLINLINE(int) vdAddCacheBackend(PVDCACHEBACKEND pBackend)
622{
623 return vdAddCacheBackends(&pBackend, 1);
624}
625
626/**
627 * internal: issue error message.
628 */
629static int vdError(PVBOXHDD pDisk, int rc, RT_SRC_POS_DECL,
630 const char *pszFormat, ...)
631{
632 va_list va;
633 va_start(va, pszFormat);
634 if (pDisk->pInterfaceError)
635 pDisk->pInterfaceError->pfnError(pDisk->pInterfaceError->Core.pvUser, rc, RT_SRC_POS_ARGS, pszFormat, va);
636 va_end(va);
637 return rc;
638}
639
640/**
641 * internal: thread synchronization, start read.
642 */
643DECLINLINE(int) vdThreadStartRead(PVBOXHDD pDisk)
644{
645 int rc = VINF_SUCCESS;
646 if (RT_UNLIKELY(pDisk->pInterfaceThreadSync))
647 rc = pDisk->pInterfaceThreadSync->pfnStartRead(pDisk->pInterfaceThreadSync->Core.pvUser);
648 return rc;
649}
650
651/**
652 * internal: thread synchronization, finish read.
653 */
654DECLINLINE(int) vdThreadFinishRead(PVBOXHDD pDisk)
655{
656 int rc = VINF_SUCCESS;
657 if (RT_UNLIKELY(pDisk->pInterfaceThreadSync))
658 rc = pDisk->pInterfaceThreadSync->pfnFinishRead(pDisk->pInterfaceThreadSync->Core.pvUser);
659 return rc;
660}
661
662/**
663 * internal: thread synchronization, start write.
664 */
665DECLINLINE(int) vdThreadStartWrite(PVBOXHDD pDisk)
666{
667 int rc = VINF_SUCCESS;
668 if (RT_UNLIKELY(pDisk->pInterfaceThreadSync))
669 rc = pDisk->pInterfaceThreadSync->pfnStartWrite(pDisk->pInterfaceThreadSync->Core.pvUser);
670 return rc;
671}
672
673/**
674 * internal: thread synchronization, finish write.
675 */
676DECLINLINE(int) vdThreadFinishWrite(PVBOXHDD pDisk)
677{
678 int rc = VINF_SUCCESS;
679 if (RT_UNLIKELY(pDisk->pInterfaceThreadSync))
680 rc = pDisk->pInterfaceThreadSync->pfnFinishWrite(pDisk->pInterfaceThreadSync->Core.pvUser);
681 return rc;
682}
683
684/**
685 * internal: find image format backend.
686 */
687static int vdFindBackend(const char *pszBackend, PCVBOXHDDBACKEND *ppBackend)
688{
689 int rc = VINF_SUCCESS;
690 PCVBOXHDDBACKEND pBackend = NULL;
691
692 if (!g_apBackends)
693 VDInit();
694
695 for (unsigned i = 0; i < g_cBackends; i++)
696 {
697 if (!RTStrICmp(pszBackend, g_apBackends[i]->pszBackendName))
698 {
699 pBackend = g_apBackends[i];
700 break;
701 }
702 }
703 *ppBackend = pBackend;
704 return rc;
705}
706
707/**
708 * internal: find cache format backend.
709 */
710static int vdFindCacheBackend(const char *pszBackend, PCVDCACHEBACKEND *ppBackend)
711{
712 int rc = VINF_SUCCESS;
713 PCVDCACHEBACKEND pBackend = NULL;
714
715 if (!g_apCacheBackends)
716 VDInit();
717
718 for (unsigned i = 0; i < g_cCacheBackends; i++)
719 {
720 if (!RTStrICmp(pszBackend, g_apCacheBackends[i]->pszBackendName))
721 {
722 pBackend = g_apCacheBackends[i];
723 break;
724 }
725 }
726 *ppBackend = pBackend;
727 return rc;
728}
729
730/**
731 * internal: add image structure to the end of images list.
732 */
733static void vdAddImageToList(PVBOXHDD pDisk, PVDIMAGE pImage)
734{
735 pImage->pPrev = NULL;
736 pImage->pNext = NULL;
737
738 if (pDisk->pBase)
739 {
740 Assert(pDisk->cImages > 0);
741 pImage->pPrev = pDisk->pLast;
742 pDisk->pLast->pNext = pImage;
743 pDisk->pLast = pImage;
744 }
745 else
746 {
747 Assert(pDisk->cImages == 0);
748 pDisk->pBase = pImage;
749 pDisk->pLast = pImage;
750 }
751
752 pDisk->cImages++;
753}
754
755/**
756 * internal: remove image structure from the images list.
757 */
758static void vdRemoveImageFromList(PVBOXHDD pDisk, PVDIMAGE pImage)
759{
760 Assert(pDisk->cImages > 0);
761
762 if (pImage->pPrev)
763 pImage->pPrev->pNext = pImage->pNext;
764 else
765 pDisk->pBase = pImage->pNext;
766
767 if (pImage->pNext)
768 pImage->pNext->pPrev = pImage->pPrev;
769 else
770 pDisk->pLast = pImage->pPrev;
771
772 pImage->pPrev = NULL;
773 pImage->pNext = NULL;
774
775 pDisk->cImages--;
776}
777
778/**
779 * internal: find image by index into the images list.
780 */
781static PVDIMAGE vdGetImageByNumber(PVBOXHDD pDisk, unsigned nImage)
782{
783 PVDIMAGE pImage = pDisk->pBase;
784 if (nImage == VD_LAST_IMAGE)
785 return pDisk->pLast;
786 while (pImage && nImage)
787 {
788 pImage = pImage->pNext;
789 nImage--;
790 }
791 return pImage;
792}
793
794/**
795 * Initialize the structure members of a given I/O context.
796 */
797DECLINLINE(void) vdIoCtxInit(PVDIOCTX pIoCtx, PVBOXHDD pDisk, VDIOCTXTXDIR enmTxDir,
798 uint64_t uOffset, size_t cbTransfer, PVDIMAGE pImageStart,
799 PCRTSGBUF pcSgBuf, void *pvAllocation,
800 PFNVDIOCTXTRANSFER pfnIoCtxTransfer, uint32_t fFlags)
801{
802 pIoCtx->pDisk = pDisk;
803 pIoCtx->enmTxDir = enmTxDir;
804 pIoCtx->Req.Io.cbTransferLeft = cbTransfer;
805 pIoCtx->Req.Io.uOffset = uOffset;
806 pIoCtx->Req.Io.cbTransfer = cbTransfer;
807 pIoCtx->Req.Io.pImageStart = pImageStart;
808 pIoCtx->Req.Io.pImageCur = pImageStart;
809 pIoCtx->cDataTransfersPending = 0;
810 pIoCtx->cMetaTransfersPending = 0;
811 pIoCtx->fComplete = false;
812 pIoCtx->fFlags = fFlags;
813 pIoCtx->pvAllocation = pvAllocation;
814 pIoCtx->pfnIoCtxTransfer = pfnIoCtxTransfer;
815 pIoCtx->pfnIoCtxTransferNext = NULL;
816 pIoCtx->rcReq = VINF_SUCCESS;
817
818 /* There is no S/G list for a flush request. */
819 if ( enmTxDir != VDIOCTXTXDIR_FLUSH
820 && enmTxDir != VDIOCTXTXDIR_DISCARD)
821 RTSgBufClone(&pIoCtx->Req.Io.SgBuf, pcSgBuf);
822 else
823 memset(&pIoCtx->Req.Io.SgBuf, 0, sizeof(RTSGBUF));
824}
825
826/**
827 * Internal: Tries to read the desired range from the given cache.
828 *
829 * @returns VBox status code.
830 * @retval VERR_VD_BLOCK_FREE if the block is not in the cache.
831 * pcbRead will be set to the number of bytes not in the cache.
832 * Everything thereafter might be in the cache.
833 * @param pCache The cache to read from.
834 * @param uOffset Offset of the virtual disk to read.
835 * @param cbRead How much to read.
836 * @param pIoCtx The I/O context to read into.
837 * @param pcbRead Where to store the number of bytes actually read.
838 * On success this indicates the number of bytes read from the cache.
839 * If VERR_VD_BLOCK_FREE is returned this gives the number of bytes
840 * which are not in the cache.
841 * In both cases everything beyond this value
842 * might or might not be in the cache.
843 */
844static int vdCacheReadHelper(PVDCACHE pCache, uint64_t uOffset,
845 PVDIOCTX pIoCtx, size_t cbRead, size_t *pcbRead)
846{
847 int rc = VINF_SUCCESS;
848
849 LogFlowFunc(("pCache=%#p uOffset=%llu pIoCtx=%p cbRead=%zu pcbRead=%#p\n",
850 pCache, uOffset, pIoCtx, cbRead, pcbRead));
851
852 AssertPtr(pCache);
853 AssertPtr(pcbRead);
854
855 rc = pCache->Backend->pfnRead(pCache->pBackendData, uOffset, cbRead,
856 pIoCtx, pcbRead);
857
858 LogFlowFunc(("returns rc=%Rrc pcbRead=%zu\n", rc, *pcbRead));
859 return rc;
860}
861
862/**
863 * Internal: Writes data for the given block into the cache.
864 *
865 * @returns VBox status code.
866 * @param pCache The cache to write to.
867 * @param uOffset Offset of the virtual disk to write to the cache.
868 * @param cbWrite How much to write.
869 * @param pIoCtx The I/O context to ẃrite from.
870 * @param pcbWritten How much data could be written, optional.
871 */
872static int vdCacheWriteHelper(PVDCACHE pCache, uint64_t uOffset, size_t cbWrite,
873 PVDIOCTX pIoCtx, size_t *pcbWritten)
874{
875 int rc = VINF_SUCCESS;
876
877 LogFlowFunc(("pCache=%#p uOffset=%llu pIoCtx=%p cbWrite=%zu pcbWritten=%#p\n",
878 pCache, uOffset, pIoCtx, cbWrite, pcbWritten));
879
880 AssertPtr(pCache);
881 AssertPtr(pIoCtx);
882 Assert(cbWrite > 0);
883
884 if (pcbWritten)
885 rc = pCache->Backend->pfnWrite(pCache->pBackendData, uOffset, cbWrite,
886 pIoCtx, pcbWritten);
887 else
888 {
889 size_t cbWritten = 0;
890
891 do
892 {
893 rc = pCache->Backend->pfnWrite(pCache->pBackendData, uOffset, cbWrite,
894 pIoCtx, &cbWritten);
895 uOffset += cbWritten;
896 cbWrite -= cbWritten;
897 } while ( cbWrite
898 && ( RT_SUCCESS(rc)
899 || rc == VERR_VD_ASYNC_IO_IN_PROGRESS));
900 }
901
902 LogFlowFunc(("returns rc=%Rrc pcbWritten=%zu\n",
903 rc, pcbWritten ? *pcbWritten : cbWrite));
904 return rc;
905}
906
907/**
908 * Internal: Reads a given amount of data from the image chain of the disk.
909 **/
910static int vdDiskReadHelper(PVBOXHDD pDisk, PVDIMAGE pImage, PVDIMAGE pImageParentOverride,
911 uint64_t uOffset, void *pvBuf, size_t cbRead, size_t *pcbThisRead)
912{
913 int rc = VINF_SUCCESS;
914 size_t cbThisRead = cbRead;
915 RTSGSEG SegmentBuf;
916 RTSGBUF SgBuf;
917 VDIOCTX IoCtx;
918
919 AssertPtr(pcbThisRead);
920
921 *pcbThisRead = 0;
922
923 SegmentBuf.pvSeg = pvBuf;
924 SegmentBuf.cbSeg = VD_MERGE_BUFFER_SIZE;
925 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
926 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_READ, 0, 0, NULL,
927 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
928
929 /*
930 * Try to read from the given image.
931 * If the block is not allocated read from override chain if present.
932 */
933 rc = pImage->Backend->pfnRead(pImage->pBackendData,
934 uOffset, cbThisRead, &IoCtx,
935 &cbThisRead);
936
937 if (rc == VERR_VD_BLOCK_FREE)
938 {
939 for (PVDIMAGE pCurrImage = pImageParentOverride ? pImageParentOverride : pImage->pPrev;
940 pCurrImage != NULL && rc == VERR_VD_BLOCK_FREE;
941 pCurrImage = pCurrImage->pPrev)
942 {
943 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
944 uOffset, cbThisRead, &IoCtx,
945 &cbThisRead);
946 }
947 }
948
949 if (RT_SUCCESS(rc) || rc == VERR_VD_BLOCK_FREE)
950 *pcbThisRead = cbThisRead;
951
952 return rc;
953}
954
955/**
956 * Extended version of vdReadHelper(), implementing certain optimizations
957 * for image cloning.
958 *
959 * @returns VBox status code.
960 * @param pDisk The disk to read from.
961 * @param pImage The image to start reading from.
962 * @param pImageParentOverride The parent image to read from
963 * if the starting image returns a free block.
964 * If NULL is passed the real parent of the image
965 * in the chain is used.
966 * @param uOffset Offset in the disk to start reading from.
967 * @param pvBuf Where to store the read data.
968 * @param cbRead How much to read.
969 * @param fZeroFreeBlocks Flag whether free blocks should be zeroed.
970 * If false and no image has data for sepcified
971 * range VERR_VD_BLOCK_FREE is returned.
972 * Note that unallocated blocks are still zeroed
973 * if at least one image has valid data for a part
974 * of the range.
975 * @param fUpdateCache Flag whether to update the attached cache if
976 * available.
977 * @param cImagesRead Number of images in the chain to read until
978 * the read is cut off. A value of 0 disables the cut off.
979 */
980static int vdReadHelperEx(PVBOXHDD pDisk, PVDIMAGE pImage, PVDIMAGE pImageParentOverride,
981 uint64_t uOffset, void *pvBuf, size_t cbRead,
982 bool fZeroFreeBlocks, bool fUpdateCache, unsigned cImagesRead)
983{
984 int rc = VINF_SUCCESS;
985 size_t cbThisRead;
986 bool fAllFree = true;
987 size_t cbBufClear = 0;
988
989 /* Loop until all read. */
990 do
991 {
992 /* Search for image with allocated block. Do not attempt to read more
993 * than the previous reads marked as valid. Otherwise this would return
994 * stale data when different block sizes are used for the images. */
995 cbThisRead = cbRead;
996
997 if ( pDisk->pCache
998 && !pImageParentOverride)
999 {
1000#if 0 /** @todo: Will go soon when the sync and async read helper versions are merged. */
1001 rc = vdCacheReadHelper(pDisk->pCache, uOffset, pvBuf,
1002 cbThisRead, &cbThisRead);
1003#endif
1004 if (rc == VERR_VD_BLOCK_FREE)
1005 {
1006 rc = vdDiskReadHelper(pDisk, pImage, NULL, uOffset, pvBuf, cbThisRead,
1007 &cbThisRead);
1008
1009 /* If the read was successful, write the data back into the cache. */
1010 if ( RT_SUCCESS(rc)
1011 && fUpdateCache)
1012 {
1013#if 0 /** @todo: Will go soon when the sync and async read helper versions are merged. */
1014 rc = vdCacheWriteHelper(pDisk->pCache, uOffset, pvBuf,
1015 cbThisRead, NULL);
1016#endif
1017 }
1018 }
1019 }
1020 else
1021 {
1022 RTSGSEG SegmentBuf;
1023 RTSGBUF SgBuf;
1024 VDIOCTX IoCtx;
1025
1026 SegmentBuf.pvSeg = pvBuf;
1027 SegmentBuf.cbSeg = cbThisRead;
1028 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
1029 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_READ, 0, 0, NULL,
1030 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
1031
1032 /*
1033 * Try to read from the given image.
1034 * If the block is not allocated read from override chain if present.
1035 */
1036 rc = pImage->Backend->pfnRead(pImage->pBackendData,
1037 uOffset, cbThisRead, &IoCtx,
1038 &cbThisRead);
1039
1040 if ( rc == VERR_VD_BLOCK_FREE
1041 && cImagesRead != 1)
1042 {
1043 unsigned cImagesToProcess = cImagesRead;
1044
1045 for (PVDIMAGE pCurrImage = pImageParentOverride ? pImageParentOverride : pImage->pPrev;
1046 pCurrImage != NULL && rc == VERR_VD_BLOCK_FREE;
1047 pCurrImage = pCurrImage->pPrev)
1048 {
1049 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
1050 uOffset, cbThisRead,
1051 &IoCtx, &cbThisRead);
1052 if (cImagesToProcess == 1)
1053 break;
1054 else if (cImagesToProcess > 0)
1055 cImagesToProcess--;
1056 }
1057 }
1058 }
1059
1060 /* No image in the chain contains the data for the block. */
1061 if (rc == VERR_VD_BLOCK_FREE)
1062 {
1063 /* Fill the free space with 0 if we are told to do so
1064 * or a previous read returned valid data. */
1065 if (fZeroFreeBlocks || !fAllFree)
1066 memset(pvBuf, '\0', cbThisRead);
1067 else
1068 cbBufClear += cbThisRead;
1069
1070 if (pImage->uOpenFlags & VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS)
1071 rc = VINF_VD_NEW_ZEROED_BLOCK;
1072 else
1073 rc = VINF_SUCCESS;
1074 }
1075 else if (RT_SUCCESS(rc))
1076 {
1077 /* First not free block, fill the space before with 0. */
1078 if (!fZeroFreeBlocks)
1079 {
1080 memset((char *)pvBuf - cbBufClear, '\0', cbBufClear);
1081 cbBufClear = 0;
1082 fAllFree = false;
1083 }
1084 }
1085
1086 cbRead -= cbThisRead;
1087 uOffset += cbThisRead;
1088 pvBuf = (char *)pvBuf + cbThisRead;
1089 } while (cbRead != 0 && RT_SUCCESS(rc));
1090
1091 return (!fZeroFreeBlocks && fAllFree) ? VERR_VD_BLOCK_FREE : rc;
1092}
1093
1094/**
1095 * internal: read the specified amount of data in whatever blocks the backend
1096 * will give us.
1097 */
1098static int vdReadHelper(PVBOXHDD pDisk, PVDIMAGE pImage, uint64_t uOffset,
1099 void *pvBuf, size_t cbRead, bool fUpdateCache)
1100{
1101 return vdReadHelperEx(pDisk, pImage, NULL, uOffset, pvBuf, cbRead,
1102 true /* fZeroFreeBlocks */, fUpdateCache, 0);
1103}
1104
1105/**
1106 * Creates a new empty discard state.
1107 *
1108 * @returns Pointer to the new discard state or NULL if out of memory.
1109 */
1110static PVDDISCARDSTATE vdDiscardStateCreate(void)
1111{
1112 PVDDISCARDSTATE pDiscard = (PVDDISCARDSTATE)RTMemAllocZ(sizeof(VDDISCARDSTATE));
1113
1114 if (pDiscard)
1115 {
1116 RTListInit(&pDiscard->ListLru);
1117 pDiscard->pTreeBlocks = (PAVLRU64TREE)RTMemAllocZ(sizeof(AVLRU64TREE));
1118 if (!pDiscard->pTreeBlocks)
1119 {
1120 RTMemFree(pDiscard);
1121 pDiscard = NULL;
1122 }
1123 }
1124
1125 return pDiscard;
1126}
1127
1128/**
1129 * Removes the least recently used blocks from the waiting list until
1130 * the new value is reached.
1131 *
1132 * @returns VBox status code.
1133 * @param pDisk VD disk container.
1134 * @param pDiscard The discard state.
1135 * @param cbDiscardingNew How many bytes should be waiting on success.
1136 * The number of bytes waiting can be less.
1137 */
1138static int vdDiscardRemoveBlocks(PVBOXHDD pDisk, PVDDISCARDSTATE pDiscard, size_t cbDiscardingNew)
1139{
1140 int rc = VINF_SUCCESS;
1141
1142 LogFlowFunc(("pDisk=%#p pDiscard=%#p cbDiscardingNew=%zu\n",
1143 pDisk, pDiscard, cbDiscardingNew));
1144
1145 while (pDiscard->cbDiscarding > cbDiscardingNew)
1146 {
1147 PVDDISCARDBLOCK pBlock = RTListGetLast(&pDiscard->ListLru, VDDISCARDBLOCK, NodeLru);
1148
1149 Assert(!RTListIsEmpty(&pDiscard->ListLru));
1150
1151 /* Go over the allocation bitmap and mark all discarded sectors as unused. */
1152 uint64_t offStart = pBlock->Core.Key;
1153 uint32_t idxStart = 0;
1154 size_t cbLeft = pBlock->cbDiscard;
1155 bool fAllocated = ASMBitTest(pBlock->pbmAllocated, idxStart);
1156 uint32_t cSectors = pBlock->cbDiscard / 512;
1157
1158 while (cbLeft > 0)
1159 {
1160 int32_t idxEnd;
1161 size_t cbThis = cbLeft;
1162
1163 if (fAllocated)
1164 {
1165 /* Check for the first unallocated bit. */
1166 idxEnd = ASMBitNextClear(pBlock->pbmAllocated, cSectors, idxStart);
1167 if (idxEnd != -1)
1168 {
1169 cbThis = (idxEnd - idxStart) * 512;
1170 fAllocated = false;
1171 }
1172 }
1173 else
1174 {
1175 /* Mark as unused and check for the first set bit. */
1176 idxEnd = ASMBitNextSet(pBlock->pbmAllocated, cSectors, idxStart);
1177 if (idxEnd != -1)
1178 cbThis = (idxEnd - idxStart) * 512;
1179
1180
1181 VDIOCTX IoCtx;
1182 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_DISCARD, 0, 0, NULL,
1183 NULL, NULL, NULL, VDIOCTX_FLAGS_SYNC);
1184 rc = pDisk->pLast->Backend->pfnDiscard(pDisk->pLast->pBackendData,
1185 &IoCtx, offStart, cbThis, NULL,
1186 NULL, &cbThis, NULL,
1187 VD_DISCARD_MARK_UNUSED);
1188 if (RT_FAILURE(rc))
1189 break;
1190
1191 fAllocated = true;
1192 }
1193
1194 idxStart = idxEnd;
1195 offStart += cbThis;
1196 cbLeft -= cbThis;
1197 }
1198
1199 if (RT_FAILURE(rc))
1200 break;
1201
1202 PVDDISCARDBLOCK pBlockRemove = (PVDDISCARDBLOCK)RTAvlrU64RangeRemove(pDiscard->pTreeBlocks, pBlock->Core.Key);
1203 Assert(pBlockRemove == pBlock);
1204 RTListNodeRemove(&pBlock->NodeLru);
1205
1206 pDiscard->cbDiscarding -= pBlock->cbDiscard;
1207 RTMemFree(pBlock->pbmAllocated);
1208 RTMemFree(pBlock);
1209 }
1210
1211 Assert(RT_FAILURE(rc) || pDiscard->cbDiscarding <= cbDiscardingNew);
1212
1213 LogFlowFunc(("returns rc=%Rrc\n", rc));
1214 return rc;
1215}
1216
1217/**
1218 * Destroys the current discard state, writing any waiting blocks to the image.
1219 *
1220 * @returns VBox status code.
1221 * @param pDisk VD disk container.
1222 */
1223static int vdDiscardStateDestroy(PVBOXHDD pDisk)
1224{
1225 int rc = VINF_SUCCESS;
1226
1227 if (pDisk->pDiscard)
1228 {
1229 rc = vdDiscardRemoveBlocks(pDisk, pDisk->pDiscard, 0 /* Remove all blocks. */);
1230 AssertRC(rc);
1231 RTMemFree(pDisk->pDiscard->pTreeBlocks);
1232 RTMemFree(pDisk->pDiscard);
1233 pDisk->pDiscard = NULL;
1234 }
1235
1236 return rc;
1237}
1238
1239/**
1240 * Marks the given range as allocated in the image.
1241 * Required if there are discards in progress and a write to a block which can get discarded
1242 * is written to.
1243 *
1244 * @returns VBox status code.
1245 * @param pDisk VD container data.
1246 * @param uOffset First byte to mark as allocated.
1247 * @param cbRange Number of bytes to mark as allocated.
1248 */
1249static int vdDiscardSetRangeAllocated(PVBOXHDD pDisk, uint64_t uOffset, size_t cbRange)
1250{
1251 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
1252 int rc = VINF_SUCCESS;
1253
1254 if (pDiscard)
1255 {
1256 do
1257 {
1258 size_t cbThisRange = cbRange;
1259 PVDDISCARDBLOCK pBlock = (PVDDISCARDBLOCK)RTAvlrU64RangeGet(pDiscard->pTreeBlocks, uOffset);
1260
1261 if (pBlock)
1262 {
1263 int32_t idxStart, idxEnd;
1264
1265 Assert(!(cbThisRange % 512));
1266 Assert(!((uOffset - pBlock->Core.Key) % 512));
1267
1268 cbThisRange = RT_MIN(cbThisRange, pBlock->Core.KeyLast - uOffset + 1);
1269
1270 idxStart = (uOffset - pBlock->Core.Key) / 512;
1271 idxEnd = idxStart + (cbThisRange / 512);
1272 ASMBitSetRange(pBlock->pbmAllocated, idxStart, idxEnd);
1273 }
1274 else
1275 {
1276 pBlock = (PVDDISCARDBLOCK)RTAvlrU64GetBestFit(pDiscard->pTreeBlocks, uOffset, true);
1277 if (pBlock)
1278 cbThisRange = RT_MIN(cbThisRange, pBlock->Core.Key - uOffset);
1279 }
1280
1281 Assert(cbRange >= cbThisRange);
1282
1283 uOffset += cbThisRange;
1284 cbRange -= cbThisRange;
1285 } while (cbRange != 0);
1286 }
1287
1288 return rc;
1289}
1290
1291DECLINLINE(PVDIOCTX) vdIoCtxAlloc(PVBOXHDD pDisk, VDIOCTXTXDIR enmTxDir,
1292 uint64_t uOffset, size_t cbTransfer,
1293 PVDIMAGE pImageStart,PCRTSGBUF pcSgBuf,
1294 void *pvAllocation, PFNVDIOCTXTRANSFER pfnIoCtxTransfer,
1295 uint32_t fFlags)
1296{
1297 PVDIOCTX pIoCtx = NULL;
1298
1299 pIoCtx = (PVDIOCTX)RTMemCacheAlloc(pDisk->hMemCacheIoCtx);
1300 if (RT_LIKELY(pIoCtx))
1301 {
1302 vdIoCtxInit(pIoCtx, pDisk, enmTxDir, uOffset, cbTransfer, pImageStart,
1303 pcSgBuf, pvAllocation, pfnIoCtxTransfer, fFlags);
1304 }
1305
1306 return pIoCtx;
1307}
1308
1309DECLINLINE(PVDIOCTX) vdIoCtxRootAlloc(PVBOXHDD pDisk, VDIOCTXTXDIR enmTxDir,
1310 uint64_t uOffset, size_t cbTransfer,
1311 PVDIMAGE pImageStart, PCRTSGBUF pcSgBuf,
1312 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
1313 void *pvUser1, void *pvUser2,
1314 void *pvAllocation,
1315 PFNVDIOCTXTRANSFER pfnIoCtxTransfer)
1316{
1317 PVDIOCTX pIoCtx = vdIoCtxAlloc(pDisk, enmTxDir, uOffset, cbTransfer, pImageStart,
1318 pcSgBuf, pvAllocation, pfnIoCtxTransfer, 0);
1319
1320 if (RT_LIKELY(pIoCtx))
1321 {
1322 pIoCtx->pIoCtxParent = NULL;
1323 pIoCtx->Type.Root.pfnComplete = pfnComplete;
1324 pIoCtx->Type.Root.pvUser1 = pvUser1;
1325 pIoCtx->Type.Root.pvUser2 = pvUser2;
1326 }
1327
1328 LogFlow(("Allocated root I/O context %#p\n", pIoCtx));
1329 return pIoCtx;
1330}
1331
1332DECLINLINE(PVDIOCTX) vdIoCtxDiscardAlloc(PVBOXHDD pDisk, PCRTRANGE paRanges,
1333 unsigned cRanges,
1334 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
1335 void *pvUser1, void *pvUser2,
1336 void *pvAllocation,
1337 PFNVDIOCTXTRANSFER pfnIoCtxTransfer,
1338 uint32_t fFlags)
1339{
1340 PVDIOCTX pIoCtx = NULL;
1341
1342 pIoCtx = (PVDIOCTX)RTMemCacheAlloc(pDisk->hMemCacheIoCtx);
1343 if (RT_LIKELY(pIoCtx))
1344 {
1345 pIoCtx->pIoCtxNext = NULL;
1346 pIoCtx->pDisk = pDisk;
1347 pIoCtx->enmTxDir = VDIOCTXTXDIR_DISCARD;
1348 pIoCtx->cDataTransfersPending = 0;
1349 pIoCtx->cMetaTransfersPending = 0;
1350 pIoCtx->fComplete = false;
1351 pIoCtx->fFlags = fFlags;
1352 pIoCtx->pvAllocation = pvAllocation;
1353 pIoCtx->pfnIoCtxTransfer = pfnIoCtxTransfer;
1354 pIoCtx->pfnIoCtxTransferNext = NULL;
1355 pIoCtx->rcReq = VINF_SUCCESS;
1356 pIoCtx->Req.Discard.paRanges = paRanges;
1357 pIoCtx->Req.Discard.cRanges = cRanges;
1358 pIoCtx->Req.Discard.idxRange = 0;
1359 pIoCtx->Req.Discard.cbDiscardLeft = 0;
1360 pIoCtx->Req.Discard.offCur = 0;
1361 pIoCtx->Req.Discard.cbThisDiscard = 0;
1362
1363 pIoCtx->pIoCtxParent = NULL;
1364 pIoCtx->Type.Root.pfnComplete = pfnComplete;
1365 pIoCtx->Type.Root.pvUser1 = pvUser1;
1366 pIoCtx->Type.Root.pvUser2 = pvUser2;
1367 }
1368
1369 LogFlow(("Allocated discard I/O context %#p\n", pIoCtx));
1370 return pIoCtx;
1371}
1372
1373DECLINLINE(PVDIOCTX) vdIoCtxChildAlloc(PVBOXHDD pDisk, VDIOCTXTXDIR enmTxDir,
1374 uint64_t uOffset, size_t cbTransfer,
1375 PVDIMAGE pImageStart, PCRTSGBUF pcSgBuf,
1376 PVDIOCTX pIoCtxParent, size_t cbTransferParent,
1377 size_t cbWriteParent, void *pvAllocation,
1378 PFNVDIOCTXTRANSFER pfnIoCtxTransfer)
1379{
1380 PVDIOCTX pIoCtx = vdIoCtxAlloc(pDisk, enmTxDir, uOffset, cbTransfer, pImageStart,
1381 pcSgBuf, pvAllocation, pfnIoCtxTransfer, 0);
1382
1383 AssertPtr(pIoCtxParent);
1384 Assert(!pIoCtxParent->pIoCtxParent);
1385
1386 if (RT_LIKELY(pIoCtx))
1387 {
1388 pIoCtx->pIoCtxParent = pIoCtxParent;
1389 pIoCtx->Type.Child.uOffsetSaved = uOffset;
1390 pIoCtx->Type.Child.cbTransferLeftSaved = cbTransfer;
1391 pIoCtx->Type.Child.cbTransferParent = cbTransferParent;
1392 pIoCtx->Type.Child.cbWriteParent = cbWriteParent;
1393 }
1394
1395 LogFlow(("Allocated child I/O context %#p\n", pIoCtx));
1396 return pIoCtx;
1397}
1398
1399DECLINLINE(PVDIOTASK) vdIoTaskUserAlloc(PVDIOSTORAGE pIoStorage, PFNVDXFERCOMPLETED pfnComplete, void *pvUser, PVDIOCTX pIoCtx, uint32_t cbTransfer)
1400{
1401 PVDIOTASK pIoTask = NULL;
1402
1403 pIoTask = (PVDIOTASK)RTMemCacheAlloc(pIoStorage->pVDIo->pDisk->hMemCacheIoTask);
1404 if (pIoTask)
1405 {
1406 pIoTask->pIoStorage = pIoStorage;
1407 pIoTask->pfnComplete = pfnComplete;
1408 pIoTask->pvUser = pvUser;
1409 pIoTask->fMeta = false;
1410 pIoTask->Type.User.cbTransfer = cbTransfer;
1411 pIoTask->Type.User.pIoCtx = pIoCtx;
1412 }
1413
1414 return pIoTask;
1415}
1416
1417DECLINLINE(PVDIOTASK) vdIoTaskMetaAlloc(PVDIOSTORAGE pIoStorage, PFNVDXFERCOMPLETED pfnComplete, void *pvUser, PVDMETAXFER pMetaXfer)
1418{
1419 PVDIOTASK pIoTask = NULL;
1420
1421 pIoTask = (PVDIOTASK)RTMemCacheAlloc(pIoStorage->pVDIo->pDisk->hMemCacheIoTask);
1422 if (pIoTask)
1423 {
1424 pIoTask->pIoStorage = pIoStorage;
1425 pIoTask->pfnComplete = pfnComplete;
1426 pIoTask->pvUser = pvUser;
1427 pIoTask->fMeta = true;
1428 pIoTask->Type.Meta.pMetaXfer = pMetaXfer;
1429 }
1430
1431 return pIoTask;
1432}
1433
1434DECLINLINE(void) vdIoCtxFree(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
1435{
1436 LogFlow(("Freeing I/O context %#p\n", pIoCtx));
1437 if (pIoCtx->pvAllocation)
1438 RTMemFree(pIoCtx->pvAllocation);
1439#ifdef DEBUG
1440 memset(pIoCtx, 0xff, sizeof(VDIOCTX));
1441#endif
1442 RTMemCacheFree(pDisk->hMemCacheIoCtx, pIoCtx);
1443}
1444
1445DECLINLINE(void) vdIoTaskFree(PVBOXHDD pDisk, PVDIOTASK pIoTask)
1446{
1447 RTMemCacheFree(pDisk->hMemCacheIoTask, pIoTask);
1448}
1449
1450DECLINLINE(void) vdIoCtxChildReset(PVDIOCTX pIoCtx)
1451{
1452 AssertPtr(pIoCtx->pIoCtxParent);
1453
1454 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
1455 pIoCtx->Req.Io.uOffset = pIoCtx->Type.Child.uOffsetSaved;
1456 pIoCtx->Req.Io.cbTransferLeft = pIoCtx->Type.Child.cbTransferLeftSaved;
1457}
1458
1459DECLINLINE(PVDMETAXFER) vdMetaXferAlloc(PVDIOSTORAGE pIoStorage, uint64_t uOffset, size_t cb)
1460{
1461 PVDMETAXFER pMetaXfer = (PVDMETAXFER)RTMemAlloc(RT_OFFSETOF(VDMETAXFER, abData[cb]));
1462
1463 if (RT_LIKELY(pMetaXfer))
1464 {
1465 pMetaXfer->Core.Key = uOffset;
1466 pMetaXfer->Core.KeyLast = uOffset + cb - 1;
1467 pMetaXfer->fFlags = VDMETAXFER_TXDIR_NONE;
1468 pMetaXfer->cbMeta = cb;
1469 pMetaXfer->pIoStorage = pIoStorage;
1470 pMetaXfer->cRefs = 0;
1471 RTListInit(&pMetaXfer->ListIoCtxWaiting);
1472 }
1473 return pMetaXfer;
1474}
1475
1476DECLINLINE(void) vdIoCtxAddToWaitingList(volatile PVDIOCTX *ppList, PVDIOCTX pIoCtx)
1477{
1478 /* Put it on the waiting list. */
1479 PVDIOCTX pNext = ASMAtomicUoReadPtrT(ppList, PVDIOCTX);
1480 PVDIOCTX pHeadOld;
1481 pIoCtx->pIoCtxNext = pNext;
1482 while (!ASMAtomicCmpXchgExPtr(ppList, pIoCtx, pNext, &pHeadOld))
1483 {
1484 pNext = pHeadOld;
1485 Assert(pNext != pIoCtx);
1486 pIoCtx->pIoCtxNext = pNext;
1487 ASMNopPause();
1488 }
1489}
1490
1491DECLINLINE(void) vdIoCtxDefer(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
1492{
1493 LogFlowFunc(("Deferring write pIoCtx=%#p\n", pIoCtx));
1494
1495 Assert(!pIoCtx->pIoCtxParent && !(pIoCtx->fFlags & VDIOCTX_FLAGS_BLOCKED));
1496 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
1497 vdIoCtxAddToWaitingList(&pDisk->pIoCtxBlockedHead, pIoCtx);
1498}
1499
1500static size_t vdIoCtxCopy(PVDIOCTX pIoCtxDst, PVDIOCTX pIoCtxSrc, size_t cbData)
1501{
1502 return RTSgBufCopy(&pIoCtxDst->Req.Io.SgBuf, &pIoCtxSrc->Req.Io.SgBuf, cbData);
1503}
1504
1505static int vdIoCtxCmp(PVDIOCTX pIoCtx1, PVDIOCTX pIoCtx2, size_t cbData)
1506{
1507 return RTSgBufCmp(&pIoCtx1->Req.Io.SgBuf, &pIoCtx2->Req.Io.SgBuf, cbData);
1508}
1509
1510static size_t vdIoCtxCopyTo(PVDIOCTX pIoCtx, const uint8_t *pbData, size_t cbData)
1511{
1512 return RTSgBufCopyFromBuf(&pIoCtx->Req.Io.SgBuf, pbData, cbData);
1513}
1514
1515static size_t vdIoCtxCopyFrom(PVDIOCTX pIoCtx, uint8_t *pbData, size_t cbData)
1516{
1517 return RTSgBufCopyToBuf(&pIoCtx->Req.Io.SgBuf, pbData, cbData);
1518}
1519
1520static size_t vdIoCtxSet(PVDIOCTX pIoCtx, uint8_t ch, size_t cbData)
1521{
1522 return RTSgBufSet(&pIoCtx->Req.Io.SgBuf, ch, cbData);
1523}
1524
1525/**
1526 * Process the I/O context, core method which assumes that the I/O context
1527 * acquired the lock.
1528 *
1529 * @returns VBox status code.
1530 * @param pIoCtx I/O context to process.
1531 */
1532static int vdIoCtxProcessLocked(PVDIOCTX pIoCtx)
1533{
1534 int rc = VINF_SUCCESS;
1535
1536 VD_IS_LOCKED(pIoCtx->pDisk);
1537
1538 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
1539
1540 if ( !pIoCtx->cMetaTransfersPending
1541 && !pIoCtx->cDataTransfersPending
1542 && !pIoCtx->pfnIoCtxTransfer)
1543 {
1544 rc = VINF_VD_ASYNC_IO_FINISHED;
1545 goto out;
1546 }
1547
1548 /*
1549 * We complete the I/O context in case of an error
1550 * if there is no I/O task pending.
1551 */
1552 if ( RT_FAILURE(pIoCtx->rcReq)
1553 && !pIoCtx->cMetaTransfersPending
1554 && !pIoCtx->cDataTransfersPending)
1555 {
1556 rc = VINF_VD_ASYNC_IO_FINISHED;
1557 goto out;
1558 }
1559
1560 /* Don't change anything if there is a metadata transfer pending or we are blocked. */
1561 if ( pIoCtx->cMetaTransfersPending
1562 || (pIoCtx->fFlags & VDIOCTX_FLAGS_BLOCKED))
1563 {
1564 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1565 goto out;
1566 }
1567
1568 if (pIoCtx->pfnIoCtxTransfer)
1569 {
1570 /* Call the transfer function advancing to the next while there is no error. */
1571 while ( pIoCtx->pfnIoCtxTransfer
1572 && !pIoCtx->cMetaTransfersPending
1573 && RT_SUCCESS(rc))
1574 {
1575 LogFlowFunc(("calling transfer function %#p\n", pIoCtx->pfnIoCtxTransfer));
1576 rc = pIoCtx->pfnIoCtxTransfer(pIoCtx);
1577
1578 /* Advance to the next part of the transfer if the current one succeeded. */
1579 if (RT_SUCCESS(rc))
1580 {
1581 pIoCtx->pfnIoCtxTransfer = pIoCtx->pfnIoCtxTransferNext;
1582 pIoCtx->pfnIoCtxTransferNext = NULL;
1583 }
1584 }
1585 }
1586
1587 if ( RT_SUCCESS(rc)
1588 && !pIoCtx->cMetaTransfersPending
1589 && !pIoCtx->cDataTransfersPending)
1590 rc = VINF_VD_ASYNC_IO_FINISHED;
1591 else if ( RT_SUCCESS(rc)
1592 || rc == VERR_VD_NOT_ENOUGH_METADATA
1593 || rc == VERR_VD_IOCTX_HALT)
1594 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1595 else if (RT_FAILURE(rc) && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
1596 {
1597 ASMAtomicCmpXchgS32(&pIoCtx->rcReq, rc, VINF_SUCCESS);
1598 /*
1599 * The I/O context completed if we have an error and there is no data
1600 * or meta data transfer pending.
1601 */
1602 if ( !pIoCtx->cMetaTransfersPending
1603 && !pIoCtx->cDataTransfersPending)
1604 rc = VINF_VD_ASYNC_IO_FINISHED;
1605 else
1606 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1607 }
1608
1609out:
1610 LogFlowFunc(("pIoCtx=%#p rc=%Rrc cDataTransfersPending=%u cMetaTransfersPending=%u fComplete=%RTbool\n",
1611 pIoCtx, rc, pIoCtx->cDataTransfersPending, pIoCtx->cMetaTransfersPending,
1612 pIoCtx->fComplete));
1613
1614 return rc;
1615}
1616
1617/**
1618 * Processes the list of waiting I/O contexts.
1619 *
1620 * @returns VBox status code.
1621 * @param pDisk The disk structure.
1622 * @param pIoCtxRc An I/O context handle which waits on the list. When processed
1623 * The status code is returned. NULL if there is no I/O context
1624 * to return the status code for.
1625 */
1626static int vdDiskProcessWaitingIoCtx(PVBOXHDD pDisk, PVDIOCTX pIoCtxRc)
1627{
1628 int rc = VINF_SUCCESS;
1629
1630 LogFlowFunc(("pDisk=%#p pIoCtxRc=%#p\n", pDisk, pIoCtxRc));
1631
1632 VD_IS_LOCKED(pDisk);
1633
1634 /* Get the waiting list and process it in FIFO order. */
1635 PVDIOCTX pIoCtxHead = ASMAtomicXchgPtrT(&pDisk->pIoCtxHead, NULL, PVDIOCTX);
1636
1637 /* Reverse it. */
1638 PVDIOCTX pCur = pIoCtxHead;
1639 pIoCtxHead = NULL;
1640 while (pCur)
1641 {
1642 PVDIOCTX pInsert = pCur;
1643 pCur = pCur->pIoCtxNext;
1644 pInsert->pIoCtxNext = pIoCtxHead;
1645 pIoCtxHead = pInsert;
1646 }
1647
1648 /* Process now. */
1649 pCur = pIoCtxHead;
1650 while (pCur)
1651 {
1652 int rcTmp;
1653 PVDIOCTX pTmp = pCur;
1654
1655 pCur = pCur->pIoCtxNext;
1656 pTmp->pIoCtxNext = NULL;
1657
1658 rcTmp = vdIoCtxProcessLocked(pTmp);
1659 if (pTmp == pIoCtxRc)
1660 {
1661 /* The given I/O context was processed, pass the return code to the caller. */
1662 rc = rcTmp;
1663 }
1664 else if ( rcTmp == VINF_VD_ASYNC_IO_FINISHED
1665 && ASMAtomicCmpXchgBool(&pTmp->fComplete, true, false))
1666 {
1667 LogFlowFunc(("Waiting I/O context completed pTmp=%#p\n", pTmp));
1668 vdThreadFinishWrite(pDisk);
1669 pTmp->Type.Root.pfnComplete(pTmp->Type.Root.pvUser1,
1670 pTmp->Type.Root.pvUser2,
1671 pTmp->rcReq);
1672 vdIoCtxFree(pDisk, pTmp);
1673 }
1674 }
1675
1676 LogFlowFunc(("returns rc=%Rrc\n", rc));
1677 return rc;
1678}
1679
1680/**
1681 * Processes the list of blocked I/O contexts.
1682 *
1683 * @returns nothing.
1684 * @param pDisk The disk structure.
1685 */
1686static void vdDiskProcessBlockedIoCtx(PVBOXHDD pDisk)
1687{
1688 LogFlowFunc(("pDisk=%#p\n", pDisk));
1689
1690 VD_IS_LOCKED(pDisk);
1691
1692 /* Get the waiting list and process it in FIFO order. */
1693 PVDIOCTX pIoCtxHead = ASMAtomicXchgPtrT(&pDisk->pIoCtxBlockedHead, NULL, PVDIOCTX);
1694
1695 /* Reverse it. */
1696 PVDIOCTX pCur = pIoCtxHead;
1697 pIoCtxHead = NULL;
1698 while (pCur)
1699 {
1700 PVDIOCTX pInsert = pCur;
1701 pCur = pCur->pIoCtxNext;
1702 pInsert->pIoCtxNext = pIoCtxHead;
1703 pIoCtxHead = pInsert;
1704 }
1705
1706 /* Process now. */
1707 pCur = pIoCtxHead;
1708 while (pCur)
1709 {
1710 int rc;
1711 PVDIOCTX pTmp = pCur;
1712
1713 pCur = pCur->pIoCtxNext;
1714 pTmp->pIoCtxNext = NULL;
1715
1716 Assert(!pTmp->pIoCtxParent);
1717 Assert(pTmp->fFlags & VDIOCTX_FLAGS_BLOCKED);
1718 pTmp->fFlags &= ~VDIOCTX_FLAGS_BLOCKED;
1719
1720 rc = vdIoCtxProcessLocked(pTmp);
1721 if ( rc == VINF_VD_ASYNC_IO_FINISHED
1722 && ASMAtomicCmpXchgBool(&pTmp->fComplete, true, false))
1723 {
1724 LogFlowFunc(("Waiting I/O context completed pTmp=%#p\n", pTmp));
1725 vdThreadFinishWrite(pDisk);
1726 pTmp->Type.Root.pfnComplete(pTmp->Type.Root.pvUser1,
1727 pTmp->Type.Root.pvUser2,
1728 pTmp->rcReq);
1729 vdIoCtxFree(pDisk, pTmp);
1730 }
1731 }
1732
1733 LogFlowFunc(("returns\n"));
1734}
1735
1736/**
1737 * Processes the I/O context trying to lock the criticial section.
1738 * The context is deferred if the critical section is busy.
1739 *
1740 * @returns VBox status code.
1741 * @param pIoCtx The I/O context to process.
1742 */
1743static int vdIoCtxProcessTryLockDefer(PVDIOCTX pIoCtx)
1744{
1745 int rc = VINF_SUCCESS;
1746 PVBOXHDD pDisk = pIoCtx->pDisk;
1747
1748 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
1749
1750 /* Put it on the waiting list first. */
1751 vdIoCtxAddToWaitingList(&pDisk->pIoCtxHead, pIoCtx);
1752
1753 if (ASMAtomicCmpXchgBool(&pDisk->fLocked, true, false))
1754 {
1755 /* Leave it again, the context will be processed just before leaving the lock. */
1756 LogFlowFunc(("Successfully acquired the lock\n"));
1757 rc = vdDiskUnlock(pDisk, pIoCtx);
1758 }
1759 else
1760 {
1761 LogFlowFunc(("Lock is held\n"));
1762 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1763 }
1764
1765 return rc;
1766}
1767
1768DECLINLINE(bool) vdIoCtxIsDiskLockOwner(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
1769{
1770 return pDisk->pIoCtxLockOwner == pIoCtx;
1771}
1772
1773static int vdIoCtxLockDisk(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
1774{
1775 int rc = VINF_SUCCESS;
1776
1777 VD_IS_LOCKED(pDisk);
1778
1779 LogFlowFunc(("pDisk=%#p pIoCtx=%#p\n", pDisk, pIoCtx));
1780
1781 if (!ASMAtomicCmpXchgPtr(&pDisk->pIoCtxLockOwner, pIoCtx, NIL_VDIOCTX))
1782 {
1783 Assert(pDisk->pIoCtxLockOwner != pIoCtx); /* No nesting allowed. */
1784 vdIoCtxDefer(pDisk, pIoCtx);
1785 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1786 }
1787
1788 LogFlowFunc(("returns -> %Rrc\n", rc));
1789 return rc;
1790}
1791
1792static void vdIoCtxUnlockDisk(PVBOXHDD pDisk, PVDIOCTX pIoCtx, bool fProcessBlockedReqs)
1793{
1794 LogFlowFunc(("pDisk=%#p pIoCtx=%#p fProcessBlockedReqs=%RTbool\n",
1795 pDisk, pIoCtx, fProcessBlockedReqs));
1796
1797 VD_IS_LOCKED(pDisk);
1798
1799 LogFlow(("Unlocking disk lock owner is %#p\n", pDisk->pIoCtxLockOwner));
1800 Assert(pDisk->pIoCtxLockOwner == pIoCtx);
1801 ASMAtomicXchgPtrT(&pDisk->pIoCtxLockOwner, NIL_VDIOCTX, PVDIOCTX);
1802
1803 if (fProcessBlockedReqs)
1804 {
1805 /* Process any blocked writes if the current request didn't caused another growing. */
1806 vdDiskProcessBlockedIoCtx(pDisk);
1807 }
1808
1809 LogFlowFunc(("returns\n"));
1810}
1811
1812/**
1813 * internal: read the specified amount of data in whatever blocks the backend
1814 * will give us - async version.
1815 */
1816static int vdReadHelperAsync(PVDIOCTX pIoCtx)
1817{
1818 int rc;
1819 size_t cbToRead = pIoCtx->Req.Io.cbTransfer;
1820 uint64_t uOffset = pIoCtx->Req.Io.uOffset;
1821 PVDIMAGE pCurrImage = pIoCtx->Req.Io.pImageCur;;
1822 size_t cbThisRead;
1823
1824 /* Loop until all reads started or we have a backend which needs to read metadata. */
1825 do
1826 {
1827 /* Search for image with allocated block. Do not attempt to read more
1828 * than the previous reads marked as valid. Otherwise this would return
1829 * stale data when different block sizes are used for the images. */
1830 cbThisRead = cbToRead;
1831
1832 /*
1833 * Try to read from the given image.
1834 * If the block is not allocated read from override chain if present.
1835 */
1836 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
1837 uOffset, cbThisRead,
1838 pIoCtx, &cbThisRead);
1839
1840 if (rc == VERR_VD_BLOCK_FREE)
1841 {
1842 while ( pCurrImage->pPrev != NULL
1843 && rc == VERR_VD_BLOCK_FREE)
1844 {
1845 pCurrImage = pCurrImage->pPrev;
1846 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
1847 uOffset, cbThisRead,
1848 pIoCtx, &cbThisRead);
1849 }
1850 }
1851
1852 /* The task state will be updated on success already, don't do it here!. */
1853 if (rc == VERR_VD_BLOCK_FREE)
1854 {
1855 /* No image in the chain contains the data for the block. */
1856 vdIoCtxSet(pIoCtx, '\0', cbThisRead);
1857 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, cbThisRead);
1858 rc = VINF_SUCCESS;
1859 }
1860 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
1861 rc = VINF_SUCCESS;
1862 else if (rc == VERR_VD_IOCTX_HALT)
1863 {
1864 uOffset += cbThisRead;
1865 cbToRead -= cbThisRead;
1866 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
1867 }
1868
1869 if (RT_FAILURE(rc))
1870 break;
1871
1872 cbToRead -= cbThisRead;
1873 uOffset += cbThisRead;
1874 pCurrImage = pIoCtx->Req.Io.pImageStart; /* Start with the highest image in the chain. */
1875 } while (cbToRead != 0 && RT_SUCCESS(rc));
1876
1877 if ( rc == VERR_VD_NOT_ENOUGH_METADATA
1878 || rc == VERR_VD_IOCTX_HALT)
1879 {
1880 /* Save the current state. */
1881 pIoCtx->Req.Io.uOffset = uOffset;
1882 pIoCtx->Req.Io.cbTransfer = cbToRead;
1883 pIoCtx->Req.Io.pImageCur = pCurrImage ? pCurrImage : pIoCtx->Req.Io.pImageStart;
1884 }
1885
1886 return rc;
1887}
1888
1889/**
1890 * internal: parent image read wrapper for compacting.
1891 */
1892static int vdParentRead(void *pvUser, uint64_t uOffset, void *pvBuf,
1893 size_t cbRead)
1894{
1895 PVDPARENTSTATEDESC pParentState = (PVDPARENTSTATEDESC)pvUser;
1896 return vdReadHelper(pParentState->pDisk, pParentState->pImage, uOffset,
1897 pvBuf, cbRead, false /* fUpdateCache */);
1898}
1899
1900/**
1901 * internal: mark the disk as not modified.
1902 */
1903static void vdResetModifiedFlag(PVBOXHDD pDisk)
1904{
1905 if (pDisk->uModified & VD_IMAGE_MODIFIED_FLAG)
1906 {
1907 /* generate new last-modified uuid */
1908 if (!(pDisk->uModified & VD_IMAGE_MODIFIED_DISABLE_UUID_UPDATE))
1909 {
1910 RTUUID Uuid;
1911
1912 RTUuidCreate(&Uuid);
1913 pDisk->pLast->Backend->pfnSetModificationUuid(pDisk->pLast->pBackendData,
1914 &Uuid);
1915
1916 if (pDisk->pCache)
1917 pDisk->pCache->Backend->pfnSetModificationUuid(pDisk->pCache->pBackendData,
1918 &Uuid);
1919 }
1920
1921 pDisk->uModified &= ~VD_IMAGE_MODIFIED_FLAG;
1922 }
1923}
1924
1925/**
1926 * internal: mark the disk as modified.
1927 */
1928static void vdSetModifiedFlag(PVBOXHDD pDisk)
1929{
1930 pDisk->uModified |= VD_IMAGE_MODIFIED_FLAG;
1931 if (pDisk->uModified & VD_IMAGE_MODIFIED_FIRST)
1932 {
1933 pDisk->uModified &= ~VD_IMAGE_MODIFIED_FIRST;
1934
1935 /* First modify, so create a UUID and ensure it's written to disk. */
1936 vdResetModifiedFlag(pDisk);
1937
1938 if (!(pDisk->uModified & VD_IMAGE_MODIFIED_DISABLE_UUID_UPDATE))
1939 {
1940 VDIOCTX IoCtx;
1941 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_FLUSH, 0, 0, NULL,
1942 NULL, NULL, NULL, VDIOCTX_FLAGS_SYNC);
1943 pDisk->pLast->Backend->pfnFlush(pDisk->pLast->pBackendData, &IoCtx);
1944 }
1945 }
1946}
1947
1948/**
1949 * internal: write a complete block (only used for diff images), taking the
1950 * remaining data from parent images. This implementation does not optimize
1951 * anything (except that it tries to read only that portions from parent
1952 * images that are really needed).
1953 */
1954static int vdWriteHelperStandard(PVBOXHDD pDisk, PVDIMAGE pImage,
1955 PVDIMAGE pImageParentOverride,
1956 uint64_t uOffset, size_t cbWrite,
1957 size_t cbThisWrite, size_t cbPreRead,
1958 size_t cbPostRead, const void *pvBuf,
1959 void *pvTmp)
1960{
1961 int rc = VINF_SUCCESS;
1962
1963 LogFlowFunc(("pDisk=%p pImage=%p pImageParentOverride=%p uOffset=%llu cbWrite=%zu\n",
1964 pDisk, pImage, pImageParentOverride, uOffset, cbWrite));
1965
1966 /* Read the data that goes before the write to fill the block. */
1967 if (cbPreRead)
1968 {
1969 /*
1970 * Updating the cache doesn't make sense here because
1971 * this will be done after the complete block was written.
1972 */
1973 rc = vdReadHelperEx(pDisk, pImage, pImageParentOverride,
1974 uOffset - cbPreRead, pvTmp, cbPreRead,
1975 true /* fZeroFreeBlocks*/,
1976 false /* fUpdateCache */, 0);
1977 if (RT_FAILURE(rc))
1978 return rc;
1979 }
1980
1981 /* Copy the data to the right place in the buffer. */
1982 memcpy((char *)pvTmp + cbPreRead, pvBuf, cbThisWrite);
1983
1984 /* Read the data that goes after the write to fill the block. */
1985 if (cbPostRead)
1986 {
1987 /* If we have data to be written, use that instead of reading
1988 * data from the image. */
1989 size_t cbWriteCopy;
1990 if (cbWrite > cbThisWrite)
1991 cbWriteCopy = RT_MIN(cbWrite - cbThisWrite, cbPostRead);
1992 else
1993 cbWriteCopy = 0;
1994 /* Figure out how much we cannot read from the image, because
1995 * the last block to write might exceed the nominal size of the
1996 * image for technical reasons. */
1997 size_t cbFill;
1998 if (uOffset + cbThisWrite + cbPostRead > pDisk->cbSize)
1999 cbFill = uOffset + cbThisWrite + cbPostRead - pDisk->cbSize;
2000 else
2001 cbFill = 0;
2002 /* The rest must be read from the image. */
2003 size_t cbReadImage = cbPostRead - cbWriteCopy - cbFill;
2004
2005 /* Now assemble the remaining data. */
2006 if (cbWriteCopy)
2007 memcpy((char *)pvTmp + cbPreRead + cbThisWrite,
2008 (char *)pvBuf + cbThisWrite, cbWriteCopy);
2009 if (cbReadImage)
2010 rc = vdReadHelperEx(pDisk, pImage, pImageParentOverride,
2011 uOffset + cbThisWrite + cbWriteCopy,
2012 (char *)pvTmp + cbPreRead + cbThisWrite + cbWriteCopy,
2013 cbReadImage, true /* fZeroFreeBlocks */,
2014 false /* fUpdateCache */, 0);
2015 if (RT_FAILURE(rc))
2016 return rc;
2017 /* Zero out the remainder of this block. Will never be visible, as this
2018 * is beyond the limit of the image. */
2019 if (cbFill)
2020 memset((char *)pvTmp + cbPreRead + cbThisWrite + cbWriteCopy + cbReadImage,
2021 '\0', cbFill);
2022 }
2023
2024 /* Write the full block to the virtual disk. */
2025 RTSGSEG SegmentBuf;
2026 RTSGBUF SgBuf;
2027 VDIOCTX IoCtx;
2028
2029 SegmentBuf.pvSeg = pvTmp;
2030 SegmentBuf.cbSeg = cbPreRead + cbThisWrite + cbPostRead;
2031 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
2032 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_WRITE, 0, 0, NULL,
2033 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
2034 rc = pImage->Backend->pfnWrite(pImage->pBackendData, uOffset - cbPreRead,
2035 cbPreRead + cbThisWrite + cbPostRead,
2036 &IoCtx, NULL, &cbPreRead, &cbPostRead, 0);
2037 Assert(rc != VERR_VD_BLOCK_FREE);
2038 Assert(cbPreRead == 0);
2039 Assert(cbPostRead == 0);
2040
2041 return rc;
2042}
2043
2044/**
2045 * internal: write a complete block (only used for diff images), taking the
2046 * remaining data from parent images. This implementation optimizes out writes
2047 * that do not change the data relative to the state as of the parent images.
2048 * All backends which support differential/growing images support this.
2049 */
2050static int vdWriteHelperOptimized(PVBOXHDD pDisk, PVDIMAGE pImage,
2051 PVDIMAGE pImageParentOverride,
2052 uint64_t uOffset, size_t cbWrite,
2053 size_t cbThisWrite, size_t cbPreRead,
2054 size_t cbPostRead, const void *pvBuf,
2055 void *pvTmp, unsigned cImagesRead)
2056{
2057 size_t cbFill = 0;
2058 size_t cbWriteCopy = 0;
2059 size_t cbReadImage = 0;
2060 int rc;
2061
2062 LogFlowFunc(("pDisk=%p pImage=%p pImageParentOverride=%p uOffset=%llu cbWrite=%zu\n",
2063 pDisk, pImage, pImageParentOverride, uOffset, cbWrite));
2064
2065 if (cbPostRead)
2066 {
2067 /* Figure out how much we cannot read from the image, because
2068 * the last block to write might exceed the nominal size of the
2069 * image for technical reasons. */
2070 if (uOffset + cbThisWrite + cbPostRead > pDisk->cbSize)
2071 cbFill = uOffset + cbThisWrite + cbPostRead - pDisk->cbSize;
2072
2073 /* If we have data to be written, use that instead of reading
2074 * data from the image. */
2075 if (cbWrite > cbThisWrite)
2076 cbWriteCopy = RT_MIN(cbWrite - cbThisWrite, cbPostRead);
2077
2078 /* The rest must be read from the image. */
2079 cbReadImage = cbPostRead - cbWriteCopy - cbFill;
2080 }
2081
2082 /* Read the entire data of the block so that we can compare whether it will
2083 * be modified by the write or not. */
2084 rc = vdReadHelperEx(pDisk, pImage, pImageParentOverride, uOffset - cbPreRead, pvTmp,
2085 cbPreRead + cbThisWrite + cbPostRead - cbFill,
2086 true /* fZeroFreeBlocks */, false /* fUpdateCache */,
2087 cImagesRead);
2088 if (RT_FAILURE(rc))
2089 return rc;
2090
2091 /* Check if the write would modify anything in this block. */
2092 if ( !memcmp((char *)pvTmp + cbPreRead, pvBuf, cbThisWrite)
2093 && (!cbWriteCopy || !memcmp((char *)pvTmp + cbPreRead + cbThisWrite,
2094 (char *)pvBuf + cbThisWrite, cbWriteCopy)))
2095 {
2096 /* Block is completely unchanged, so no need to write anything. */
2097 return VINF_SUCCESS;
2098 }
2099
2100 /* Copy the data to the right place in the buffer. */
2101 memcpy((char *)pvTmp + cbPreRead, pvBuf, cbThisWrite);
2102
2103 /* Handle the data that goes after the write to fill the block. */
2104 if (cbPostRead)
2105 {
2106 /* Now assemble the remaining data. */
2107 if (cbWriteCopy)
2108 memcpy((char *)pvTmp + cbPreRead + cbThisWrite,
2109 (char *)pvBuf + cbThisWrite, cbWriteCopy);
2110 /* Zero out the remainder of this block. Will never be visible, as this
2111 * is beyond the limit of the image. */
2112 if (cbFill)
2113 memset((char *)pvTmp + cbPreRead + cbThisWrite + cbWriteCopy + cbReadImage,
2114 '\0', cbFill);
2115 }
2116
2117 /* Write the full block to the virtual disk. */
2118 RTSGSEG SegmentBuf;
2119 RTSGBUF SgBuf;
2120 VDIOCTX IoCtx;
2121
2122 SegmentBuf.pvSeg = pvTmp;
2123 SegmentBuf.cbSeg = cbPreRead + cbThisWrite + cbPostRead;
2124 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
2125 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_WRITE, 0, 0, NULL,
2126 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
2127 rc = pImage->Backend->pfnWrite(pImage->pBackendData, uOffset - cbPreRead,
2128 cbPreRead + cbThisWrite + cbPostRead,
2129 &IoCtx, NULL, &cbPreRead, &cbPostRead, 0);
2130 Assert(rc != VERR_VD_BLOCK_FREE);
2131 Assert(cbPreRead == 0);
2132 Assert(cbPostRead == 0);
2133
2134 return rc;
2135}
2136
2137/**
2138 * internal: write buffer to the image, taking care of block boundaries and
2139 * write optimizations.
2140 */
2141static int vdWriteHelperEx(PVBOXHDD pDisk, PVDIMAGE pImage,
2142 PVDIMAGE pImageParentOverride, uint64_t uOffset,
2143 const void *pvBuf, size_t cbWrite,
2144 bool fUpdateCache, unsigned cImagesRead)
2145{
2146 int rc;
2147 unsigned fWrite;
2148 size_t cbThisWrite;
2149 size_t cbPreRead, cbPostRead;
2150 uint64_t uOffsetCur = uOffset;
2151 size_t cbWriteCur = cbWrite;
2152 const void *pcvBufCur = pvBuf;
2153 RTSGSEG SegmentBuf;
2154 RTSGBUF SgBuf;
2155 VDIOCTX IoCtx;
2156
2157 /* Loop until all written. */
2158 do
2159 {
2160 /* Try to write the possibly partial block to the last opened image.
2161 * This works when the block is already allocated in this image or
2162 * if it is a full-block write (and allocation isn't suppressed below).
2163 * For image formats which don't support zero blocks, it's beneficial
2164 * to avoid unnecessarily allocating unchanged blocks. This prevents
2165 * unwanted expanding of images. VMDK is an example. */
2166 cbThisWrite = cbWriteCur;
2167 fWrite = (pImage->uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME)
2168 ? 0 : VD_WRITE_NO_ALLOC;
2169
2170 SegmentBuf.pvSeg = (void *)pcvBufCur;
2171 SegmentBuf.cbSeg = cbWrite;
2172 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
2173 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_WRITE, 0, 0, NULL,
2174 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
2175 rc = pImage->Backend->pfnWrite(pImage->pBackendData, uOffsetCur, cbThisWrite,
2176 &IoCtx, &cbThisWrite, &cbPreRead,
2177 &cbPostRead, fWrite);
2178 if (rc == VERR_VD_BLOCK_FREE)
2179 {
2180 void *pvTmp = RTMemTmpAlloc(cbPreRead + cbThisWrite + cbPostRead);
2181 AssertBreakStmt(VALID_PTR(pvTmp), rc = VERR_NO_MEMORY);
2182
2183 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME))
2184 {
2185 /* Optimized write, suppress writing to a so far unallocated
2186 * block if the data is in fact not changed. */
2187 rc = vdWriteHelperOptimized(pDisk, pImage, pImageParentOverride,
2188 uOffsetCur, cbWriteCur,
2189 cbThisWrite, cbPreRead, cbPostRead,
2190 pcvBufCur, pvTmp, cImagesRead);
2191 }
2192 else
2193 {
2194 /* Normal write, not optimized in any way. The block will
2195 * be written no matter what. This will usually (unless the
2196 * backend has some further optimization enabled) cause the
2197 * block to be allocated. */
2198 rc = vdWriteHelperStandard(pDisk, pImage, pImageParentOverride,
2199 uOffsetCur, cbWriteCur,
2200 cbThisWrite, cbPreRead, cbPostRead,
2201 pcvBufCur, pvTmp);
2202 }
2203 RTMemTmpFree(pvTmp);
2204 if (RT_FAILURE(rc))
2205 break;
2206 }
2207
2208 cbWriteCur -= cbThisWrite;
2209 uOffsetCur += cbThisWrite;
2210 pcvBufCur = (char *)pcvBufCur + cbThisWrite;
2211 } while (cbWriteCur != 0 && RT_SUCCESS(rc));
2212
2213#if 0 /** @todo: Soon removed when sync and async version of the write helper are merged. */
2214 /* Update the cache on success */
2215 if ( RT_SUCCESS(rc)
2216 && pDisk->pCache
2217 && fUpdateCache)
2218 rc = vdCacheWriteHelper(pDisk->pCache, uOffset, pvBuf, cbWrite, NULL);
2219
2220 if (RT_SUCCESS(rc))
2221 rc = vdDiscardSetRangeAllocated(pDisk, uOffset, cbWrite);
2222#endif
2223
2224 return rc;
2225}
2226
2227/**
2228 * internal: write buffer to the image, taking care of block boundaries and
2229 * write optimizations.
2230 */
2231static int vdWriteHelper(PVBOXHDD pDisk, PVDIMAGE pImage, uint64_t uOffset,
2232 const void *pvBuf, size_t cbWrite, bool fUpdateCache)
2233{
2234 return vdWriteHelperEx(pDisk, pImage, NULL, uOffset, pvBuf, cbWrite,
2235 fUpdateCache, 0);
2236}
2237
2238/**
2239 * Internal: Copies the content of one disk to another one applying optimizations
2240 * to speed up the copy process if possible.
2241 */
2242static int vdCopyHelper(PVBOXHDD pDiskFrom, PVDIMAGE pImageFrom, PVBOXHDD pDiskTo,
2243 uint64_t cbSize, unsigned cImagesFromRead, unsigned cImagesToRead,
2244 bool fSuppressRedundantIo, PVDINTERFACEPROGRESS pIfProgress,
2245 PVDINTERFACEPROGRESS pDstIfProgress)
2246{
2247 int rc = VINF_SUCCESS;
2248 int rc2;
2249 uint64_t uOffset = 0;
2250 uint64_t cbRemaining = cbSize;
2251 void *pvBuf = NULL;
2252 bool fLockReadFrom = false;
2253 bool fLockWriteTo = false;
2254 bool fBlockwiseCopy = fSuppressRedundantIo || (cImagesFromRead > 0);
2255 unsigned uProgressOld = 0;
2256
2257 LogFlowFunc(("pDiskFrom=%#p pImageFrom=%#p pDiskTo=%#p cbSize=%llu cImagesFromRead=%u cImagesToRead=%u fSuppressRedundantIo=%RTbool pIfProgress=%#p pDstIfProgress=%#p\n",
2258 pDiskFrom, pImageFrom, pDiskTo, cbSize, cImagesFromRead, cImagesToRead, fSuppressRedundantIo, pDstIfProgress, pDstIfProgress));
2259
2260 /* Allocate tmp buffer. */
2261 pvBuf = RTMemTmpAlloc(VD_MERGE_BUFFER_SIZE);
2262 if (!pvBuf)
2263 return rc;
2264
2265 do
2266 {
2267 size_t cbThisRead = RT_MIN(VD_MERGE_BUFFER_SIZE, cbRemaining);
2268
2269 /* Note that we don't attempt to synchronize cross-disk accesses.
2270 * It wouldn't be very difficult to do, just the lock order would
2271 * need to be defined somehow to prevent deadlocks. Postpone such
2272 * magic as there is no use case for this. */
2273
2274 rc2 = vdThreadStartRead(pDiskFrom);
2275 AssertRC(rc2);
2276 fLockReadFrom = true;
2277
2278 if (fBlockwiseCopy)
2279 {
2280 RTSGSEG SegmentBuf;
2281 RTSGBUF SgBuf;
2282 VDIOCTX IoCtx;
2283
2284 SegmentBuf.pvSeg = pvBuf;
2285 SegmentBuf.cbSeg = VD_MERGE_BUFFER_SIZE;
2286 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
2287 vdIoCtxInit(&IoCtx, pDiskFrom, VDIOCTXTXDIR_READ, 0, 0, NULL,
2288 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
2289
2290 /* Read the source data. */
2291 rc = pImageFrom->Backend->pfnRead(pImageFrom->pBackendData,
2292 uOffset, cbThisRead, &IoCtx,
2293 &cbThisRead);
2294
2295 if ( rc == VERR_VD_BLOCK_FREE
2296 && cImagesFromRead != 1)
2297 {
2298 unsigned cImagesToProcess = cImagesFromRead;
2299
2300 for (PVDIMAGE pCurrImage = pImageFrom->pPrev;
2301 pCurrImage != NULL && rc == VERR_VD_BLOCK_FREE;
2302 pCurrImage = pCurrImage->pPrev)
2303 {
2304 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
2305 uOffset, cbThisRead,
2306 &IoCtx, &cbThisRead);
2307 if (cImagesToProcess == 1)
2308 break;
2309 else if (cImagesToProcess > 0)
2310 cImagesToProcess--;
2311 }
2312 }
2313 }
2314 else
2315 rc = vdReadHelper(pDiskFrom, pImageFrom, uOffset, pvBuf, cbThisRead,
2316 false /* fUpdateCache */);
2317
2318 if (RT_FAILURE(rc) && rc != VERR_VD_BLOCK_FREE)
2319 break;
2320
2321 rc2 = vdThreadFinishRead(pDiskFrom);
2322 AssertRC(rc2);
2323 fLockReadFrom = false;
2324
2325 if (rc != VERR_VD_BLOCK_FREE)
2326 {
2327 rc2 = vdThreadStartWrite(pDiskTo);
2328 AssertRC(rc2);
2329 fLockWriteTo = true;
2330
2331 /* Only do collapsed I/O if we are copying the data blockwise. */
2332 rc = vdWriteHelperEx(pDiskTo, pDiskTo->pLast, NULL, uOffset, pvBuf,
2333 cbThisRead, false /* fUpdateCache */,
2334 fBlockwiseCopy ? cImagesToRead : 0);
2335 if (RT_FAILURE(rc))
2336 break;
2337
2338 rc2 = vdThreadFinishWrite(pDiskTo);
2339 AssertRC(rc2);
2340 fLockWriteTo = false;
2341 }
2342 else /* Don't propagate the error to the outside */
2343 rc = VINF_SUCCESS;
2344
2345 uOffset += cbThisRead;
2346 cbRemaining -= cbThisRead;
2347
2348 unsigned uProgressNew = uOffset * 99 / cbSize;
2349 if (uProgressNew != uProgressOld)
2350 {
2351 uProgressOld = uProgressNew;
2352
2353 if (pIfProgress && pIfProgress->pfnProgress)
2354 {
2355 rc = pIfProgress->pfnProgress(pIfProgress->Core.pvUser,
2356 uProgressOld);
2357 if (RT_FAILURE(rc))
2358 break;
2359 }
2360 if (pDstIfProgress && pDstIfProgress->pfnProgress)
2361 {
2362 rc = pDstIfProgress->pfnProgress(pDstIfProgress->Core.pvUser,
2363 uProgressOld);
2364 if (RT_FAILURE(rc))
2365 break;
2366 }
2367 }
2368 } while (uOffset < cbSize);
2369
2370 RTMemFree(pvBuf);
2371
2372 if (fLockReadFrom)
2373 {
2374 rc2 = vdThreadFinishRead(pDiskFrom);
2375 AssertRC(rc2);
2376 }
2377
2378 if (fLockWriteTo)
2379 {
2380 rc2 = vdThreadFinishWrite(pDiskTo);
2381 AssertRC(rc2);
2382 }
2383
2384 LogFlowFunc(("returns rc=%Rrc\n", rc));
2385 return rc;
2386}
2387
2388/**
2389 * Flush helper async version.
2390 */
2391static int vdSetModifiedHelperAsync(PVDIOCTX pIoCtx)
2392{
2393 int rc = VINF_SUCCESS;
2394 PVBOXHDD pDisk = pIoCtx->pDisk;
2395 PVDIMAGE pImage = pIoCtx->Req.Io.pImageCur;
2396
2397 rc = pImage->Backend->pfnFlush(pImage->pBackendData, pIoCtx);
2398 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
2399 rc = VINF_SUCCESS;
2400
2401 return rc;
2402}
2403
2404/**
2405 * internal: mark the disk as modified - async version.
2406 */
2407static int vdSetModifiedFlagAsync(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
2408{
2409 int rc = VINF_SUCCESS;
2410
2411 VD_IS_LOCKED(pDisk);
2412
2413 pDisk->uModified |= VD_IMAGE_MODIFIED_FLAG;
2414 if (pDisk->uModified & VD_IMAGE_MODIFIED_FIRST)
2415 {
2416 rc = vdIoCtxLockDisk(pDisk, pIoCtx);
2417 if (RT_SUCCESS(rc))
2418 {
2419 pDisk->uModified &= ~VD_IMAGE_MODIFIED_FIRST;
2420
2421 /* First modify, so create a UUID and ensure it's written to disk. */
2422 vdResetModifiedFlag(pDisk);
2423
2424 if (!(pDisk->uModified & VD_IMAGE_MODIFIED_DISABLE_UUID_UPDATE))
2425 {
2426 PVDIOCTX pIoCtxFlush = vdIoCtxChildAlloc(pDisk, VDIOCTXTXDIR_FLUSH,
2427 0, 0, pDisk->pLast,
2428 NULL, pIoCtx, 0, 0, NULL,
2429 vdSetModifiedHelperAsync);
2430
2431 if (pIoCtxFlush)
2432 {
2433 rc = vdIoCtxProcessLocked(pIoCtxFlush);
2434 if (rc == VINF_VD_ASYNC_IO_FINISHED)
2435 {
2436 vdIoCtxUnlockDisk(pDisk, pIoCtx, false /* fProcessDeferredReqs */);
2437 vdIoCtxFree(pDisk, pIoCtxFlush);
2438 }
2439 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
2440 {
2441 ASMAtomicIncU32(&pIoCtx->cDataTransfersPending);
2442 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
2443 }
2444 else /* Another error */
2445 vdIoCtxFree(pDisk, pIoCtxFlush);
2446 }
2447 else
2448 rc = VERR_NO_MEMORY;
2449 }
2450 }
2451 }
2452
2453 return rc;
2454}
2455
2456/**
2457 * internal: write a complete block (only used for diff images), taking the
2458 * remaining data from parent images. This implementation does not optimize
2459 * anything (except that it tries to read only that portions from parent
2460 * images that are really needed) - async version.
2461 */
2462static int vdWriteHelperStandardAsync(PVDIOCTX pIoCtx)
2463{
2464 int rc = VINF_SUCCESS;
2465
2466#if 0
2467
2468 /* Read the data that goes before the write to fill the block. */
2469 if (cbPreRead)
2470 {
2471 rc = vdReadHelperAsync(pIoCtxDst);
2472 if (RT_FAILURE(rc))
2473 return rc;
2474 }
2475
2476 /* Copy the data to the right place in the buffer. */
2477 vdIoCtxCopy(pIoCtxDst, pIoCtxSrc, cbThisWrite);
2478
2479 /* Read the data that goes after the write to fill the block. */
2480 if (cbPostRead)
2481 {
2482 /* If we have data to be written, use that instead of reading
2483 * data from the image. */
2484 size_t cbWriteCopy;
2485 if (cbWrite > cbThisWrite)
2486 cbWriteCopy = RT_MIN(cbWrite - cbThisWrite, cbPostRead);
2487 else
2488 cbWriteCopy = 0;
2489 /* Figure out how much we cannot read from the image, because
2490 * the last block to write might exceed the nominal size of the
2491 * image for technical reasons. */
2492 size_t cbFill;
2493 if (uOffset + cbThisWrite + cbPostRead > pDisk->cbSize)
2494 cbFill = uOffset + cbThisWrite + cbPostRead - pDisk->cbSize;
2495 else
2496 cbFill = 0;
2497 /* The rest must be read from the image. */
2498 size_t cbReadImage = cbPostRead - cbWriteCopy - cbFill;
2499
2500 /* Now assemble the remaining data. */
2501 if (cbWriteCopy)
2502 {
2503 vdIoCtxCopy(pIoCtxDst, pIoCtxSrc, cbWriteCopy);
2504 ASMAtomicSubU32(&pIoCtxDst->cbTransferLeft, cbWriteCopy);
2505 }
2506
2507 if (cbReadImage)
2508 rc = vdReadHelperAsync(pDisk, pImage, pImageParentOverride, pIoCtxDst,
2509 uOffset + cbThisWrite + cbWriteCopy,
2510 cbReadImage);
2511 if (RT_FAILURE(rc))
2512 return rc;
2513 /* Zero out the remainder of this block. Will never be visible, as this
2514 * is beyond the limit of the image. */
2515 if (cbFill)
2516 {
2517 vdIoCtxSet(pIoCtxDst, '\0', cbFill);
2518 ASMAtomicSubU32(&pIoCtxDst->cbTransferLeft, cbFill);
2519 }
2520 }
2521
2522 if ( !pIoCtxDst->cbTransferLeft
2523 && !pIoCtxDst->cMetaTransfersPending
2524 && ASMAtomicCmpXchgBool(&pIoCtxDst->fComplete, true, false))
2525 {
2526 /* Write the full block to the virtual disk. */
2527 vdIoCtxChildReset(pIoCtxDst);
2528 rc = pImage->Backend->pfnWrite(pImage->pBackendData,
2529 uOffset - cbPreRead,
2530 cbPreRead + cbThisWrite + cbPostRead,
2531 pIoCtxDst,
2532 NULL, &cbPreRead, &cbPostRead, 0);
2533 Assert(rc != VERR_VD_BLOCK_FREE);
2534 Assert(cbPreRead == 0);
2535 Assert(cbPostRead == 0);
2536 }
2537 else
2538 {
2539 LogFlow(("cbTransferLeft=%u cMetaTransfersPending=%u fComplete=%RTbool\n",
2540 pIoCtxDst->cbTransferLeft, pIoCtxDst->cMetaTransfersPending,
2541 pIoCtxDst->fComplete));
2542 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
2543 }
2544
2545 return rc;
2546#endif
2547 return VERR_NOT_IMPLEMENTED;
2548}
2549
2550static int vdWriteHelperOptimizedCommitAsync(PVDIOCTX pIoCtx)
2551{
2552 int rc = VINF_SUCCESS;
2553 PVDIMAGE pImage = pIoCtx->Req.Io.pImageStart;
2554 size_t cbPreRead = pIoCtx->Type.Child.cbPreRead;
2555 size_t cbPostRead = pIoCtx->Type.Child.cbPostRead;
2556 size_t cbThisWrite = pIoCtx->Type.Child.cbTransferParent;
2557
2558 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2559 rc = pImage->Backend->pfnWrite(pImage->pBackendData,
2560 pIoCtx->Req.Io.uOffset - cbPreRead,
2561 cbPreRead + cbThisWrite + cbPostRead,
2562 pIoCtx, NULL, &cbPreRead, &cbPostRead, 0);
2563 Assert(rc != VERR_VD_BLOCK_FREE);
2564 Assert(rc == VERR_VD_NOT_ENOUGH_METADATA || cbPreRead == 0);
2565 Assert(rc == VERR_VD_NOT_ENOUGH_METADATA || cbPostRead == 0);
2566 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
2567 rc = VINF_SUCCESS;
2568 else if (rc == VERR_VD_IOCTX_HALT)
2569 {
2570 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
2571 rc = VINF_SUCCESS;
2572 }
2573
2574 LogFlowFunc(("returns rc=%Rrc\n", rc));
2575 return rc;
2576}
2577
2578static int vdWriteHelperOptimizedCmpAndWriteAsync(PVDIOCTX pIoCtx)
2579{
2580 int rc = VINF_SUCCESS;
2581 PVDIMAGE pImage = pIoCtx->Req.Io.pImageCur;
2582 size_t cbThisWrite = 0;
2583 size_t cbPreRead = pIoCtx->Type.Child.cbPreRead;
2584 size_t cbPostRead = pIoCtx->Type.Child.cbPostRead;
2585 size_t cbWriteCopy = pIoCtx->Type.Child.Write.Optimized.cbWriteCopy;
2586 size_t cbFill = pIoCtx->Type.Child.Write.Optimized.cbFill;
2587 size_t cbReadImage = pIoCtx->Type.Child.Write.Optimized.cbReadImage;
2588 PVDIOCTX pIoCtxParent = pIoCtx->pIoCtxParent;
2589
2590 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2591
2592 AssertPtr(pIoCtxParent);
2593 Assert(!pIoCtxParent->pIoCtxParent);
2594 Assert(!pIoCtx->Req.Io.cbTransferLeft && !pIoCtx->cMetaTransfersPending);
2595
2596 vdIoCtxChildReset(pIoCtx);
2597 cbThisWrite = pIoCtx->Type.Child.cbTransferParent;
2598 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbPreRead);
2599
2600 /* Check if the write would modify anything in this block. */
2601 if (!RTSgBufCmp(&pIoCtx->Req.Io.SgBuf, &pIoCtxParent->Req.Io.SgBuf, cbThisWrite))
2602 {
2603 RTSGBUF SgBufSrcTmp;
2604
2605 RTSgBufClone(&SgBufSrcTmp, &pIoCtxParent->Req.Io.SgBuf);
2606 RTSgBufAdvance(&SgBufSrcTmp, cbThisWrite);
2607 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbThisWrite);
2608
2609 if (!cbWriteCopy || !RTSgBufCmp(&pIoCtx->Req.Io.SgBuf, &SgBufSrcTmp, cbWriteCopy))
2610 {
2611 /* Block is completely unchanged, so no need to write anything. */
2612 LogFlowFunc(("Block didn't changed\n"));
2613 ASMAtomicWriteU32(&pIoCtx->Req.Io.cbTransferLeft, 0);
2614 RTSgBufAdvance(&pIoCtxParent->Req.Io.SgBuf, cbThisWrite);
2615 return VINF_VD_ASYNC_IO_FINISHED;
2616 }
2617 }
2618
2619 /* Copy the data to the right place in the buffer. */
2620 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
2621 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbPreRead);
2622 vdIoCtxCopy(pIoCtx, pIoCtxParent, cbThisWrite);
2623
2624 /* Handle the data that goes after the write to fill the block. */
2625 if (cbPostRead)
2626 {
2627 /* Now assemble the remaining data. */
2628 if (cbWriteCopy)
2629 {
2630 /*
2631 * The S/G buffer of the parent needs to be cloned because
2632 * it is not allowed to modify the state.
2633 */
2634 RTSGBUF SgBufParentTmp;
2635
2636 RTSgBufClone(&SgBufParentTmp, &pIoCtxParent->Req.Io.SgBuf);
2637 RTSgBufCopy(&pIoCtx->Req.Io.SgBuf, &SgBufParentTmp, cbWriteCopy);
2638 }
2639
2640 /* Zero out the remainder of this block. Will never be visible, as this
2641 * is beyond the limit of the image. */
2642 if (cbFill)
2643 {
2644 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbReadImage);
2645 vdIoCtxSet(pIoCtx, '\0', cbFill);
2646 }
2647 }
2648
2649 /* Write the full block to the virtual disk. */
2650 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
2651 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperOptimizedCommitAsync;
2652
2653 return rc;
2654}
2655
2656static int vdWriteHelperOptimizedPreReadAsync(PVDIOCTX pIoCtx)
2657{
2658 int rc = VINF_SUCCESS;
2659
2660 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2661
2662 if (pIoCtx->Req.Io.cbTransferLeft)
2663 rc = vdReadHelperAsync(pIoCtx);
2664
2665 if ( RT_SUCCESS(rc)
2666 && ( pIoCtx->Req.Io.cbTransferLeft
2667 || pIoCtx->cMetaTransfersPending))
2668 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
2669 else
2670 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperOptimizedCmpAndWriteAsync;
2671
2672 return rc;
2673}
2674
2675/**
2676 * internal: write a complete block (only used for diff images), taking the
2677 * remaining data from parent images. This implementation optimizes out writes
2678 * that do not change the data relative to the state as of the parent images.
2679 * All backends which support differential/growing images support this - async version.
2680 */
2681static int vdWriteHelperOptimizedAsync(PVDIOCTX pIoCtx)
2682{
2683 PVBOXHDD pDisk = pIoCtx->pDisk;
2684 uint64_t uOffset = pIoCtx->Type.Child.uOffsetSaved;
2685 size_t cbThisWrite = pIoCtx->Type.Child.cbTransferParent;
2686 size_t cbPreRead = pIoCtx->Type.Child.cbPreRead;
2687 size_t cbPostRead = pIoCtx->Type.Child.cbPostRead;
2688 size_t cbWrite = pIoCtx->Type.Child.cbWriteParent;
2689 size_t cbFill = 0;
2690 size_t cbWriteCopy = 0;
2691 size_t cbReadImage = 0;
2692
2693 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2694
2695 AssertPtr(pIoCtx->pIoCtxParent);
2696 Assert(!pIoCtx->pIoCtxParent->pIoCtxParent);
2697
2698 if (cbPostRead)
2699 {
2700 /* Figure out how much we cannot read from the image, because
2701 * the last block to write might exceed the nominal size of the
2702 * image for technical reasons. */
2703 if (uOffset + cbThisWrite + cbPostRead > pDisk->cbSize)
2704 cbFill = uOffset + cbThisWrite + cbPostRead - pDisk->cbSize;
2705
2706 /* If we have data to be written, use that instead of reading
2707 * data from the image. */
2708 if (cbWrite > cbThisWrite)
2709 cbWriteCopy = RT_MIN(cbWrite - cbThisWrite, cbPostRead);
2710
2711 /* The rest must be read from the image. */
2712 cbReadImage = cbPostRead - cbWriteCopy - cbFill;
2713 }
2714
2715 pIoCtx->Type.Child.Write.Optimized.cbFill = cbFill;
2716 pIoCtx->Type.Child.Write.Optimized.cbWriteCopy = cbWriteCopy;
2717 pIoCtx->Type.Child.Write.Optimized.cbReadImage = cbReadImage;
2718
2719 /* Read the entire data of the block so that we can compare whether it will
2720 * be modified by the write or not. */
2721 pIoCtx->Req.Io.cbTransferLeft = cbPreRead + cbThisWrite + cbPostRead - cbFill;
2722 pIoCtx->Req.Io.cbTransfer = pIoCtx->Req.Io.cbTransferLeft;
2723 pIoCtx->Req.Io.uOffset -= cbPreRead;
2724
2725 /* Next step */
2726 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperOptimizedPreReadAsync;
2727 return VINF_SUCCESS;
2728}
2729
2730/**
2731 * internal: write buffer to the image, taking care of block boundaries and
2732 * write optimizations - async version.
2733 */
2734static int vdWriteHelperAsync(PVDIOCTX pIoCtx)
2735{
2736 int rc;
2737 size_t cbWrite = pIoCtx->Req.Io.cbTransfer;
2738 uint64_t uOffset = pIoCtx->Req.Io.uOffset;
2739 PVDIMAGE pImage = pIoCtx->Req.Io.pImageCur;
2740 PVBOXHDD pDisk = pIoCtx->pDisk;
2741 unsigned fWrite;
2742 size_t cbThisWrite;
2743 size_t cbPreRead, cbPostRead;
2744
2745 rc = vdSetModifiedFlagAsync(pDisk, pIoCtx);
2746 if (RT_FAILURE(rc)) /* Includes I/O in progress. */
2747 return rc;
2748
2749 rc = vdDiscardSetRangeAllocated(pDisk, uOffset, cbWrite);
2750 if (RT_FAILURE(rc))
2751 return rc;
2752
2753 /* Loop until all written. */
2754 do
2755 {
2756 /* Try to write the possibly partial block to the last opened image.
2757 * This works when the block is already allocated in this image or
2758 * if it is a full-block write (and allocation isn't suppressed below).
2759 * For image formats which don't support zero blocks, it's beneficial
2760 * to avoid unnecessarily allocating unchanged blocks. This prevents
2761 * unwanted expanding of images. VMDK is an example. */
2762 cbThisWrite = cbWrite;
2763 fWrite = (pImage->uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME)
2764 ? 0 : VD_WRITE_NO_ALLOC;
2765 rc = pImage->Backend->pfnWrite(pImage->pBackendData, uOffset,
2766 cbThisWrite, pIoCtx,
2767 &cbThisWrite, &cbPreRead,
2768 &cbPostRead, fWrite);
2769 if (rc == VERR_VD_BLOCK_FREE)
2770 {
2771 /* Lock the disk .*/
2772 rc = vdIoCtxLockDisk(pDisk, pIoCtx);
2773 if (RT_SUCCESS(rc))
2774 {
2775 /*
2776 * Allocate segment and buffer in one go.
2777 * A bit hackish but avoids the need to allocate memory twice.
2778 */
2779 PRTSGBUF pTmp = (PRTSGBUF)RTMemAlloc(cbPreRead + cbThisWrite + cbPostRead + sizeof(RTSGSEG) + sizeof(RTSGBUF));
2780 AssertBreakStmt(VALID_PTR(pTmp), rc = VERR_NO_MEMORY);
2781 PRTSGSEG pSeg = (PRTSGSEG)(pTmp + 1);
2782
2783 pSeg->pvSeg = pSeg + 1;
2784 pSeg->cbSeg = cbPreRead + cbThisWrite + cbPostRead;
2785 RTSgBufInit(pTmp, pSeg, 1);
2786
2787 PVDIOCTX pIoCtxWrite = vdIoCtxChildAlloc(pDisk, VDIOCTXTXDIR_WRITE,
2788 uOffset, pSeg->cbSeg, pImage,
2789 pTmp,
2790 pIoCtx, cbThisWrite,
2791 cbWrite,
2792 pTmp,
2793 (pImage->uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME)
2794 ? vdWriteHelperStandardAsync
2795 : vdWriteHelperOptimizedAsync);
2796 if (!VALID_PTR(pIoCtxWrite))
2797 {
2798 RTMemTmpFree(pTmp);
2799 rc = VERR_NO_MEMORY;
2800 break;
2801 }
2802
2803 LogFlowFunc(("Disk is growing because of pIoCtx=%#p pIoCtxWrite=%#p\n",
2804 pIoCtx, pIoCtxWrite));
2805
2806 pIoCtxWrite->Type.Child.cbPreRead = cbPreRead;
2807 pIoCtxWrite->Type.Child.cbPostRead = cbPostRead;
2808
2809 /* Process the write request */
2810 rc = vdIoCtxProcessLocked(pIoCtxWrite);
2811
2812 if (RT_FAILURE(rc) && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
2813 {
2814 vdIoCtxFree(pDisk, pIoCtxWrite);
2815 break;
2816 }
2817 else if ( rc == VINF_VD_ASYNC_IO_FINISHED
2818 && ASMAtomicCmpXchgBool(&pIoCtxWrite->fComplete, true, false))
2819 {
2820 LogFlow(("Child write request completed\n"));
2821 Assert(pIoCtx->Req.Io.cbTransferLeft >= cbThisWrite);
2822 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, cbThisWrite);
2823 vdIoCtxUnlockDisk(pDisk, pIoCtx, false /* fProcessDeferredReqs*/ );
2824 vdIoCtxFree(pDisk, pIoCtxWrite);
2825
2826 rc = VINF_SUCCESS;
2827 }
2828 else
2829 {
2830 LogFlow(("Child write pending\n"));
2831 ASMAtomicIncU32(&pIoCtx->cDataTransfersPending);
2832 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
2833 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
2834 cbWrite -= cbThisWrite;
2835 uOffset += cbThisWrite;
2836 break;
2837 }
2838 }
2839 else
2840 {
2841 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
2842 break;
2843 }
2844 }
2845
2846 if (rc == VERR_VD_IOCTX_HALT)
2847 {
2848 cbWrite -= cbThisWrite;
2849 uOffset += cbThisWrite;
2850 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
2851 break;
2852 }
2853 else if (rc == VERR_VD_NOT_ENOUGH_METADATA)
2854 break;
2855
2856 cbWrite -= cbThisWrite;
2857 uOffset += cbThisWrite;
2858 } while (cbWrite != 0 && (RT_SUCCESS(rc) || rc == VERR_VD_ASYNC_IO_IN_PROGRESS));
2859
2860 if ( rc == VERR_VD_ASYNC_IO_IN_PROGRESS
2861 || rc == VERR_VD_NOT_ENOUGH_METADATA
2862 || rc == VERR_VD_IOCTX_HALT)
2863 {
2864 /*
2865 * Tell the caller that we don't need to go back here because all
2866 * writes are initiated.
2867 */
2868 if ( !cbWrite
2869 && rc != VERR_VD_IOCTX_HALT)
2870 rc = VINF_SUCCESS;
2871
2872 pIoCtx->Req.Io.uOffset = uOffset;
2873 pIoCtx->Req.Io.cbTransfer = cbWrite;
2874 }
2875
2876 return rc;
2877}
2878
2879/**
2880 * Flush helper async version.
2881 */
2882static int vdFlushHelperAsync(PVDIOCTX pIoCtx)
2883{
2884 int rc = VINF_SUCCESS;
2885 PVBOXHDD pDisk = pIoCtx->pDisk;
2886 PVDIMAGE pImage = pIoCtx->Req.Io.pImageCur;
2887
2888 rc = vdIoCtxLockDisk(pDisk, pIoCtx);
2889 if (RT_SUCCESS(rc))
2890 {
2891 vdResetModifiedFlag(pDisk);
2892 rc = pImage->Backend->pfnFlush(pImage->pBackendData, pIoCtx);
2893 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
2894 rc = VINF_SUCCESS;
2895 else if (rc == VINF_VD_ASYNC_IO_FINISHED)
2896 vdIoCtxUnlockDisk(pDisk, pIoCtx, true /* fProcessDeferredReqs */);
2897 }
2898
2899 return rc;
2900}
2901
2902/**
2903 * Async discard helper - discards a whole block which is recorded in the block
2904 * tree.
2905 *
2906 * @returns VBox status code.
2907 * @param pIoCtx The I/O context to operate on.
2908 */
2909static int vdDiscardWholeBlockAsync(PVDIOCTX pIoCtx)
2910{
2911 int rc = VINF_SUCCESS;
2912 PVBOXHDD pDisk = pIoCtx->pDisk;
2913 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
2914 PVDDISCARDBLOCK pBlock = pIoCtx->Req.Discard.pBlock;
2915 size_t cbPreAllocated, cbPostAllocated, cbActuallyDiscarded;
2916
2917 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2918
2919 AssertPtr(pBlock);
2920
2921 rc = pDisk->pLast->Backend->pfnDiscard(pDisk->pLast->pBackendData, pIoCtx,
2922 pBlock->Core.Key, pBlock->cbDiscard,
2923 &cbPreAllocated, &cbPostAllocated,
2924 &cbActuallyDiscarded, NULL, 0);
2925 Assert(rc != VERR_VD_DISCARD_ALIGNMENT_NOT_MET);
2926 Assert(!cbPreAllocated);
2927 Assert(!cbPostAllocated);
2928 Assert(cbActuallyDiscarded == pBlock->cbDiscard || RT_FAILURE(rc));
2929
2930 /* Remove the block on success. */
2931 if ( RT_SUCCESS(rc)
2932 || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
2933 {
2934 PVDDISCARDBLOCK pBlockRemove = (PVDDISCARDBLOCK)RTAvlrU64RangeRemove(pDiscard->pTreeBlocks, pBlock->Core.Key);
2935 Assert(pBlockRemove == pBlock);
2936
2937 pDiscard->cbDiscarding -= pBlock->cbDiscard;
2938 RTListNodeRemove(&pBlock->NodeLru);
2939 RTMemFree(pBlock->pbmAllocated);
2940 RTMemFree(pBlock);
2941 pIoCtx->Req.Discard.pBlock = NULL;/* Safety precaution. */
2942 pIoCtx->pfnIoCtxTransferNext = vdDiscardHelperAsync; /* Next part. */
2943 rc = VINF_SUCCESS;
2944 }
2945
2946 LogFlowFunc(("returns rc=%Rrc\n", rc));
2947 return rc;
2948}
2949
2950/**
2951 * Removes the least recently used blocks from the waiting list until
2952 * the new value is reached - version for async I/O.
2953 *
2954 * @returns VBox status code.
2955 * @param pDisk VD disk container.
2956 * @param pDiscard The discard state.
2957 * @param cbDiscardingNew How many bytes should be waiting on success.
2958 * The number of bytes waiting can be less.
2959 */
2960static int vdDiscardRemoveBlocksAsync(PVBOXHDD pDisk, PVDIOCTX pIoCtx, size_t cbDiscardingNew)
2961{
2962 int rc = VINF_SUCCESS;
2963 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
2964
2965 LogFlowFunc(("pDisk=%#p pDiscard=%#p cbDiscardingNew=%zu\n",
2966 pDisk, pDiscard, cbDiscardingNew));
2967
2968 while (pDiscard->cbDiscarding > cbDiscardingNew)
2969 {
2970 PVDDISCARDBLOCK pBlock = RTListGetLast(&pDiscard->ListLru, VDDISCARDBLOCK, NodeLru);
2971
2972 Assert(!RTListIsEmpty(&pDiscard->ListLru));
2973
2974 /* Go over the allocation bitmap and mark all discarded sectors as unused. */
2975 uint64_t offStart = pBlock->Core.Key;
2976 uint32_t idxStart = 0;
2977 size_t cbLeft = pBlock->cbDiscard;
2978 bool fAllocated = ASMBitTest(pBlock->pbmAllocated, idxStart);
2979 uint32_t cSectors = pBlock->cbDiscard / 512;
2980
2981 while (cbLeft > 0)
2982 {
2983 int32_t idxEnd;
2984 size_t cbThis = cbLeft;
2985
2986 if (fAllocated)
2987 {
2988 /* Check for the first unallocated bit. */
2989 idxEnd = ASMBitNextClear(pBlock->pbmAllocated, cSectors, idxStart);
2990 if (idxEnd != -1)
2991 {
2992 cbThis = (idxEnd - idxStart) * 512;
2993 fAllocated = false;
2994 }
2995 }
2996 else
2997 {
2998 /* Mark as unused and check for the first set bit. */
2999 idxEnd = ASMBitNextSet(pBlock->pbmAllocated, cSectors, idxStart);
3000 if (idxEnd != -1)
3001 cbThis = (idxEnd - idxStart) * 512;
3002
3003 rc = pDisk->pLast->Backend->pfnDiscard(pDisk->pLast->pBackendData, pIoCtx,
3004 offStart, cbThis, NULL, NULL, &cbThis,
3005 NULL, VD_DISCARD_MARK_UNUSED);
3006 if ( RT_FAILURE(rc)
3007 && rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
3008 break;
3009
3010 fAllocated = true;
3011 }
3012
3013 idxStart = idxEnd;
3014 offStart += cbThis;
3015 cbLeft -= cbThis;
3016 }
3017
3018 if ( RT_FAILURE(rc)
3019 && rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
3020 break;
3021
3022 PVDDISCARDBLOCK pBlockRemove = (PVDDISCARDBLOCK)RTAvlrU64RangeRemove(pDiscard->pTreeBlocks, pBlock->Core.Key);
3023 Assert(pBlockRemove == pBlock);
3024 RTListNodeRemove(&pBlock->NodeLru);
3025
3026 pDiscard->cbDiscarding -= pBlock->cbDiscard;
3027 RTMemFree(pBlock->pbmAllocated);
3028 RTMemFree(pBlock);
3029 }
3030
3031 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
3032 rc = VINF_SUCCESS;
3033
3034 Assert(RT_FAILURE(rc) || pDiscard->cbDiscarding <= cbDiscardingNew);
3035
3036 LogFlowFunc(("returns rc=%Rrc\n", rc));
3037 return rc;
3038}
3039
3040/**
3041 * Async discard helper - discards the current range if there is no matching
3042 * block in the tree.
3043 *
3044 * @returns VBox status code.
3045 * @param pIoCtx The I/O context to operate on.
3046 */
3047static int vdDiscardCurrentRangeAsync(PVDIOCTX pIoCtx)
3048{
3049 PVBOXHDD pDisk = pIoCtx->pDisk;
3050 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
3051 uint64_t offStart = pIoCtx->Req.Discard.offCur;
3052 size_t cbThisDiscard = pIoCtx->Req.Discard.cbThisDiscard;
3053 void *pbmAllocated = NULL;
3054 size_t cbPreAllocated, cbPostAllocated;
3055 int rc = VINF_SUCCESS;
3056
3057 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
3058
3059 /* No block found, try to discard using the backend first. */
3060 rc = pDisk->pLast->Backend->pfnDiscard(pDisk->pLast->pBackendData, pIoCtx,
3061 offStart, cbThisDiscard, &cbPreAllocated,
3062 &cbPostAllocated, &cbThisDiscard,
3063 &pbmAllocated, 0);
3064 if (rc == VERR_VD_DISCARD_ALIGNMENT_NOT_MET)
3065 {
3066 /* Create new discard block. */
3067 PVDDISCARDBLOCK pBlock = (PVDDISCARDBLOCK)RTMemAllocZ(sizeof(VDDISCARDBLOCK));
3068 if (pBlock)
3069 {
3070 pBlock->Core.Key = offStart - cbPreAllocated;
3071 pBlock->Core.KeyLast = offStart + cbThisDiscard + cbPostAllocated - 1;
3072 pBlock->cbDiscard = cbPreAllocated + cbThisDiscard + cbPostAllocated;
3073 pBlock->pbmAllocated = pbmAllocated;
3074 bool fInserted = RTAvlrU64Insert(pDiscard->pTreeBlocks, &pBlock->Core);
3075 Assert(fInserted);
3076
3077 RTListPrepend(&pDiscard->ListLru, &pBlock->NodeLru);
3078 pDiscard->cbDiscarding += pBlock->cbDiscard;
3079
3080 Assert(pIoCtx->Req.Discard.cbDiscardLeft >= cbThisDiscard);
3081 pIoCtx->Req.Discard.cbDiscardLeft -= cbThisDiscard;
3082 pIoCtx->Req.Discard.offCur += cbThisDiscard;
3083 pIoCtx->Req.Discard.cbThisDiscard = cbThisDiscard;
3084
3085 if (pDiscard->cbDiscarding > VD_DISCARD_REMOVE_THRESHOLD)
3086 rc = vdDiscardRemoveBlocksAsync(pDisk, pIoCtx, VD_DISCARD_REMOVE_THRESHOLD);
3087 else
3088 rc = VINF_SUCCESS;
3089
3090 if (RT_SUCCESS(rc))
3091 pIoCtx->pfnIoCtxTransferNext = vdDiscardHelperAsync; /* Next part. */
3092 }
3093 else
3094 {
3095 RTMemFree(pbmAllocated);
3096 rc = VERR_NO_MEMORY;
3097 }
3098 }
3099 else if ( RT_SUCCESS(rc)
3100 || rc == VERR_VD_ASYNC_IO_IN_PROGRESS) /* Save state and andvance to next range. */
3101 {
3102 Assert(pIoCtx->Req.Discard.cbDiscardLeft >= cbThisDiscard);
3103 pIoCtx->Req.Discard.cbDiscardLeft -= cbThisDiscard;
3104 pIoCtx->Req.Discard.offCur += cbThisDiscard;
3105 pIoCtx->Req.Discard.cbThisDiscard = cbThisDiscard;
3106 pIoCtx->pfnIoCtxTransferNext = vdDiscardHelperAsync;
3107 rc = VINF_SUCCESS;
3108 }
3109
3110 LogFlowFunc(("returns rc=%Rrc\n", rc));
3111 return rc;
3112}
3113
3114/**
3115 * Async discard helper - entry point.
3116 *
3117 * @returns VBox status code.
3118 * @param pIoCtx The I/O context to operate on.
3119 */
3120static int vdDiscardHelperAsync(PVDIOCTX pIoCtx)
3121{
3122 int rc = VINF_SUCCESS;
3123 PVBOXHDD pDisk = pIoCtx->pDisk;
3124 PCRTRANGE paRanges = pIoCtx->Req.Discard.paRanges;
3125 unsigned cRanges = pIoCtx->Req.Discard.cRanges;
3126 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
3127
3128 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
3129
3130 /* Check if the I/O context processed all ranges. */
3131 if ( pIoCtx->Req.Discard.idxRange == cRanges
3132 && !pIoCtx->Req.Discard.cbDiscardLeft)
3133 {
3134 LogFlowFunc(("All ranges discarded, completing\n"));
3135 vdIoCtxUnlockDisk(pDisk, pIoCtx, true /* fProcessDeferredReqs*/);
3136 return VINF_SUCCESS;
3137 }
3138
3139 if (pDisk->pIoCtxLockOwner != pIoCtx)
3140 rc = vdIoCtxLockDisk(pDisk, pIoCtx);
3141
3142 if (RT_SUCCESS(rc))
3143 {
3144 uint64_t offStart = pIoCtx->Req.Discard.offCur;
3145 size_t cbDiscardLeft = pIoCtx->Req.Discard.cbDiscardLeft;
3146 size_t cbThisDiscard;
3147
3148 if (RT_UNLIKELY(!pDiscard))
3149 {
3150 pDiscard = vdDiscardStateCreate();
3151 if (!pDiscard)
3152 return VERR_NO_MEMORY;
3153
3154 pDisk->pDiscard = pDiscard;
3155 }
3156
3157 if (!pIoCtx->Req.Discard.cbDiscardLeft)
3158 {
3159 offStart = paRanges[pIoCtx->Req.Discard.idxRange].offStart;
3160 cbDiscardLeft = paRanges[pIoCtx->Req.Discard.idxRange].cbRange;
3161 LogFlowFunc(("New range descriptor loaded (%u) offStart=%llu cbDiscard=%zu\n",
3162 pIoCtx->Req.Discard.idxRange, offStart, cbDiscardLeft));
3163 pIoCtx->Req.Discard.idxRange++;
3164 }
3165
3166 /* Look for a matching block in the AVL tree first. */
3167 PVDDISCARDBLOCK pBlock = (PVDDISCARDBLOCK)RTAvlrU64GetBestFit(pDiscard->pTreeBlocks, offStart, false);
3168 if (!pBlock || pBlock->Core.KeyLast < offStart)
3169 {
3170 PVDDISCARDBLOCK pBlockAbove = (PVDDISCARDBLOCK)RTAvlrU64GetBestFit(pDiscard->pTreeBlocks, offStart, true);
3171
3172 /* Clip range to remain in the current block. */
3173 if (pBlockAbove)
3174 cbThisDiscard = RT_MIN(cbDiscardLeft, pBlockAbove->Core.KeyLast - offStart + 1);
3175 else
3176 cbThisDiscard = cbDiscardLeft;
3177
3178 Assert(!(cbThisDiscard % 512));
3179 pIoCtx->Req.Discard.pBlock = NULL;
3180 pIoCtx->pfnIoCtxTransferNext = vdDiscardCurrentRangeAsync;
3181 }
3182 else
3183 {
3184 /* Range lies partly in the block, update allocation bitmap. */
3185 int32_t idxStart, idxEnd;
3186
3187 cbThisDiscard = RT_MIN(cbDiscardLeft, pBlock->Core.KeyLast - offStart + 1);
3188
3189 AssertPtr(pBlock);
3190
3191 Assert(!(cbThisDiscard % 512));
3192 Assert(!((offStart - pBlock->Core.Key) % 512));
3193
3194 idxStart = (offStart - pBlock->Core.Key) / 512;
3195 idxEnd = idxStart + (cbThisDiscard / 512);
3196
3197 ASMBitClearRange(pBlock->pbmAllocated, idxStart, idxEnd);
3198
3199 cbDiscardLeft -= cbThisDiscard;
3200 offStart += cbThisDiscard;
3201
3202 /* Call the backend to discard the block if it is completely unallocated now. */
3203 if (ASMBitFirstSet((volatile void *)pBlock->pbmAllocated, pBlock->cbDiscard / 512) == -1)
3204 {
3205 pIoCtx->Req.Discard.pBlock = pBlock;
3206 pIoCtx->pfnIoCtxTransferNext = vdDiscardWholeBlockAsync;
3207 rc = VINF_SUCCESS;
3208 }
3209 else
3210 {
3211 RTListNodeRemove(&pBlock->NodeLru);
3212 RTListPrepend(&pDiscard->ListLru, &pBlock->NodeLru);
3213
3214 /* Start with next range. */
3215 pIoCtx->pfnIoCtxTransferNext = vdDiscardHelperAsync;
3216 rc = VINF_SUCCESS;
3217 }
3218 }
3219
3220 /* Save state in the context. */
3221 pIoCtx->Req.Discard.offCur = offStart;
3222 pIoCtx->Req.Discard.cbDiscardLeft = cbDiscardLeft;
3223 pIoCtx->Req.Discard.cbThisDiscard = cbThisDiscard;
3224 }
3225
3226 LogFlowFunc(("returns rc=%Rrc\n", rc));
3227 return rc;
3228}
3229
3230/**
3231 * internal: scans plugin directory and loads the backends have been found.
3232 */
3233static int vdLoadDynamicBackends()
3234{
3235#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
3236 int rc = VINF_SUCCESS;
3237 PRTDIR pPluginDir = NULL;
3238
3239 /* Enumerate plugin backends. */
3240 char szPath[RTPATH_MAX];
3241 rc = RTPathAppPrivateArch(szPath, sizeof(szPath));
3242 if (RT_FAILURE(rc))
3243 return rc;
3244
3245 /* To get all entries with VBoxHDD as prefix. */
3246 char *pszPluginFilter = RTPathJoinA(szPath, VBOX_HDDFORMAT_PLUGIN_PREFIX "*");
3247 if (!pszPluginFilter)
3248 return VERR_NO_STR_MEMORY;
3249
3250 PRTDIRENTRYEX pPluginDirEntry = NULL;
3251 size_t cbPluginDirEntry = sizeof(RTDIRENTRYEX);
3252 /* The plugins are in the same directory as the other shared libs. */
3253 rc = RTDirOpenFiltered(&pPluginDir, pszPluginFilter, RTDIRFILTER_WINNT, 0);
3254 if (RT_FAILURE(rc))
3255 {
3256 /* On Windows the above immediately signals that there are no
3257 * files matching, while on other platforms enumerating the
3258 * files below fails. Either way: no plugins. */
3259 goto out;
3260 }
3261
3262 pPluginDirEntry = (PRTDIRENTRYEX)RTMemAllocZ(sizeof(RTDIRENTRYEX));
3263 if (!pPluginDirEntry)
3264 {
3265 rc = VERR_NO_MEMORY;
3266 goto out;
3267 }
3268
3269 while ((rc = RTDirReadEx(pPluginDir, pPluginDirEntry, &cbPluginDirEntry, RTFSOBJATTRADD_NOTHING, RTPATH_F_ON_LINK)) != VERR_NO_MORE_FILES)
3270 {
3271 RTLDRMOD hPlugin = NIL_RTLDRMOD;
3272 PFNVBOXHDDFORMATLOAD pfnHDDFormatLoad = NULL;
3273 PVBOXHDDBACKEND pBackend = NULL;
3274 char *pszPluginPath = NULL;
3275
3276 if (rc == VERR_BUFFER_OVERFLOW)
3277 {
3278 /* allocate new buffer. */
3279 RTMemFree(pPluginDirEntry);
3280 pPluginDirEntry = (PRTDIRENTRYEX)RTMemAllocZ(cbPluginDirEntry);
3281 if (!pPluginDirEntry)
3282 {
3283 rc = VERR_NO_MEMORY;
3284 break;
3285 }
3286 /* Retry. */
3287 rc = RTDirReadEx(pPluginDir, pPluginDirEntry, &cbPluginDirEntry, RTFSOBJATTRADD_NOTHING, RTPATH_F_ON_LINK);
3288 if (RT_FAILURE(rc))
3289 break;
3290 }
3291 else if (RT_FAILURE(rc))
3292 break;
3293
3294 /* We got the new entry. */
3295 if (!RTFS_IS_FILE(pPluginDirEntry->Info.Attr.fMode))
3296 continue;
3297
3298 /* Prepend the path to the libraries. */
3299 pszPluginPath = RTPathJoinA(szPath, pPluginDirEntry->szName);
3300 if (!pszPluginPath)
3301 {
3302 rc = VERR_NO_STR_MEMORY;
3303 break;
3304 }
3305
3306 rc = SUPR3HardenedLdrLoadPlugIn(pszPluginPath, &hPlugin, NULL);
3307 if (RT_SUCCESS(rc))
3308 {
3309 rc = RTLdrGetSymbol(hPlugin, VBOX_HDDFORMAT_LOAD_NAME, (void**)&pfnHDDFormatLoad);
3310 if (RT_FAILURE(rc) || !pfnHDDFormatLoad)
3311 {
3312 LogFunc(("error resolving the entry point %s in plugin %s, rc=%Rrc, pfnHDDFormat=%#p\n", VBOX_HDDFORMAT_LOAD_NAME, pPluginDirEntry->szName, rc, pfnHDDFormatLoad));
3313 if (RT_SUCCESS(rc))
3314 rc = VERR_SYMBOL_NOT_FOUND;
3315 }
3316
3317 if (RT_SUCCESS(rc))
3318 {
3319 /* Get the function table. */
3320 rc = pfnHDDFormatLoad(&pBackend);
3321 if (RT_SUCCESS(rc) && pBackend->cbSize == sizeof(VBOXHDDBACKEND))
3322 {
3323 pBackend->hPlugin = hPlugin;
3324 vdAddBackend(pBackend);
3325 }
3326 else
3327 LogFunc(("ignored plugin '%s': pBackend->cbSize=%d rc=%Rrc\n", pszPluginPath, pBackend->cbSize, rc));
3328 }
3329 else
3330 LogFunc(("ignored plugin '%s': rc=%Rrc\n", pszPluginPath, rc));
3331
3332 if (RT_FAILURE(rc))
3333 RTLdrClose(hPlugin);
3334 }
3335 RTStrFree(pszPluginPath);
3336 }
3337out:
3338 if (rc == VERR_NO_MORE_FILES)
3339 rc = VINF_SUCCESS;
3340 RTStrFree(pszPluginFilter);
3341 if (pPluginDirEntry)
3342 RTMemFree(pPluginDirEntry);
3343 if (pPluginDir)
3344 RTDirClose(pPluginDir);
3345 return rc;
3346#else
3347 return VINF_SUCCESS;
3348#endif
3349}
3350
3351/**
3352 * internal: scans plugin directory and loads the cache backends have been found.
3353 */
3354static int vdLoadDynamicCacheBackends()
3355{
3356#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
3357 int rc = VINF_SUCCESS;
3358 PRTDIR pPluginDir = NULL;
3359
3360 /* Enumerate plugin backends. */
3361 char szPath[RTPATH_MAX];
3362 rc = RTPathAppPrivateArch(szPath, sizeof(szPath));
3363 if (RT_FAILURE(rc))
3364 return rc;
3365
3366 /* To get all entries with VBoxHDD as prefix. */
3367 char *pszPluginFilter = RTPathJoinA(szPath, VD_CACHEFORMAT_PLUGIN_PREFIX "*");
3368 if (!pszPluginFilter)
3369 {
3370 rc = VERR_NO_STR_MEMORY;
3371 return rc;
3372 }
3373
3374 PRTDIRENTRYEX pPluginDirEntry = NULL;
3375 size_t cbPluginDirEntry = sizeof(RTDIRENTRYEX);
3376 /* The plugins are in the same directory as the other shared libs. */
3377 rc = RTDirOpenFiltered(&pPluginDir, pszPluginFilter, RTDIRFILTER_WINNT, 0);
3378 if (RT_FAILURE(rc))
3379 {
3380 /* On Windows the above immediately signals that there are no
3381 * files matching, while on other platforms enumerating the
3382 * files below fails. Either way: no plugins. */
3383 goto out;
3384 }
3385
3386 pPluginDirEntry = (PRTDIRENTRYEX)RTMemAllocZ(sizeof(RTDIRENTRYEX));
3387 if (!pPluginDirEntry)
3388 {
3389 rc = VERR_NO_MEMORY;
3390 goto out;
3391 }
3392
3393 while ((rc = RTDirReadEx(pPluginDir, pPluginDirEntry, &cbPluginDirEntry, RTFSOBJATTRADD_NOTHING, RTPATH_F_ON_LINK)) != VERR_NO_MORE_FILES)
3394 {
3395 RTLDRMOD hPlugin = NIL_RTLDRMOD;
3396 PFNVDCACHEFORMATLOAD pfnVDCacheLoad = NULL;
3397 PVDCACHEBACKEND pBackend = NULL;
3398 char *pszPluginPath = NULL;
3399
3400 if (rc == VERR_BUFFER_OVERFLOW)
3401 {
3402 /* allocate new buffer. */
3403 RTMemFree(pPluginDirEntry);
3404 pPluginDirEntry = (PRTDIRENTRYEX)RTMemAllocZ(cbPluginDirEntry);
3405 if (!pPluginDirEntry)
3406 {
3407 rc = VERR_NO_MEMORY;
3408 break;
3409 }
3410 /* Retry. */
3411 rc = RTDirReadEx(pPluginDir, pPluginDirEntry, &cbPluginDirEntry, RTFSOBJATTRADD_NOTHING, RTPATH_F_ON_LINK);
3412 if (RT_FAILURE(rc))
3413 break;
3414 }
3415 else if (RT_FAILURE(rc))
3416 break;
3417
3418 /* We got the new entry. */
3419 if (!RTFS_IS_FILE(pPluginDirEntry->Info.Attr.fMode))
3420 continue;
3421
3422 /* Prepend the path to the libraries. */
3423 pszPluginPath = RTPathJoinA(szPath, pPluginDirEntry->szName);
3424 if (!pszPluginPath)
3425 {
3426 rc = VERR_NO_STR_MEMORY;
3427 break;
3428 }
3429
3430 rc = SUPR3HardenedLdrLoadPlugIn(pszPluginPath, &hPlugin, NULL);
3431 if (RT_SUCCESS(rc))
3432 {
3433 rc = RTLdrGetSymbol(hPlugin, VD_CACHEFORMAT_LOAD_NAME, (void**)&pfnVDCacheLoad);
3434 if (RT_FAILURE(rc) || !pfnVDCacheLoad)
3435 {
3436 LogFunc(("error resolving the entry point %s in plugin %s, rc=%Rrc, pfnVDCacheLoad=%#p\n",
3437 VD_CACHEFORMAT_LOAD_NAME, pPluginDirEntry->szName, rc, pfnVDCacheLoad));
3438 if (RT_SUCCESS(rc))
3439 rc = VERR_SYMBOL_NOT_FOUND;
3440 }
3441
3442 if (RT_SUCCESS(rc))
3443 {
3444 /* Get the function table. */
3445 rc = pfnVDCacheLoad(&pBackend);
3446 if (RT_SUCCESS(rc) && pBackend->cbSize == sizeof(VDCACHEBACKEND))
3447 {
3448 pBackend->hPlugin = hPlugin;
3449 vdAddCacheBackend(pBackend);
3450 }
3451 else
3452 LogFunc(("ignored plugin '%s': pBackend->cbSize=%d rc=%Rrc\n", pszPluginPath, pBackend->cbSize, rc));
3453 }
3454 else
3455 LogFunc(("ignored plugin '%s': rc=%Rrc\n", pszPluginPath, rc));
3456
3457 if (RT_FAILURE(rc))
3458 RTLdrClose(hPlugin);
3459 }
3460 RTStrFree(pszPluginPath);
3461 }
3462out:
3463 if (rc == VERR_NO_MORE_FILES)
3464 rc = VINF_SUCCESS;
3465 RTStrFree(pszPluginFilter);
3466 if (pPluginDirEntry)
3467 RTMemFree(pPluginDirEntry);
3468 if (pPluginDir)
3469 RTDirClose(pPluginDir);
3470 return rc;
3471#else
3472 return VINF_SUCCESS;
3473#endif
3474}
3475
3476/**
3477 * VD async I/O interface open callback.
3478 */
3479static int vdIOOpenFallback(void *pvUser, const char *pszLocation,
3480 uint32_t fOpen, PFNVDCOMPLETED pfnCompleted,
3481 void **ppStorage)
3482{
3483 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)RTMemAllocZ(sizeof(VDIIOFALLBACKSTORAGE));
3484
3485 if (!pStorage)
3486 return VERR_NO_MEMORY;
3487
3488 pStorage->pfnCompleted = pfnCompleted;
3489
3490 /* Open the file. */
3491 int rc = RTFileOpen(&pStorage->File, pszLocation, fOpen);
3492 if (RT_SUCCESS(rc))
3493 {
3494 *ppStorage = pStorage;
3495 return VINF_SUCCESS;
3496 }
3497
3498 RTMemFree(pStorage);
3499 return rc;
3500}
3501
3502/**
3503 * VD async I/O interface close callback.
3504 */
3505static int vdIOCloseFallback(void *pvUser, void *pvStorage)
3506{
3507 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
3508
3509 RTFileClose(pStorage->File);
3510 RTMemFree(pStorage);
3511 return VINF_SUCCESS;
3512}
3513
3514static int vdIODeleteFallback(void *pvUser, const char *pcszFilename)
3515{
3516 return RTFileDelete(pcszFilename);
3517}
3518
3519static int vdIOMoveFallback(void *pvUser, const char *pcszSrc, const char *pcszDst, unsigned fMove)
3520{
3521 return RTFileMove(pcszSrc, pcszDst, fMove);
3522}
3523
3524static int vdIOGetFreeSpaceFallback(void *pvUser, const char *pcszFilename, int64_t *pcbFreeSpace)
3525{
3526 return RTFsQuerySizes(pcszFilename, NULL, pcbFreeSpace, NULL, NULL);
3527}
3528
3529static int vdIOGetModificationTimeFallback(void *pvUser, const char *pcszFilename, PRTTIMESPEC pModificationTime)
3530{
3531 RTFSOBJINFO info;
3532 int rc = RTPathQueryInfo(pcszFilename, &info, RTFSOBJATTRADD_NOTHING);
3533 if (RT_SUCCESS(rc))
3534 *pModificationTime = info.ModificationTime;
3535 return rc;
3536}
3537
3538/**
3539 * VD async I/O interface callback for retrieving the file size.
3540 */
3541static int vdIOGetSizeFallback(void *pvUser, void *pvStorage, uint64_t *pcbSize)
3542{
3543 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
3544
3545 return RTFileGetSize(pStorage->File, pcbSize);
3546}
3547
3548/**
3549 * VD async I/O interface callback for setting the file size.
3550 */
3551static int vdIOSetSizeFallback(void *pvUser, void *pvStorage, uint64_t cbSize)
3552{
3553 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
3554
3555 return RTFileSetSize(pStorage->File, cbSize);
3556}
3557
3558/**
3559 * VD async I/O interface callback for a synchronous write to the file.
3560 */
3561static int vdIOWriteSyncFallback(void *pvUser, void *pvStorage, uint64_t uOffset,
3562 const void *pvBuf, size_t cbWrite, size_t *pcbWritten)
3563{
3564 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
3565
3566 return RTFileWriteAt(pStorage->File, uOffset, pvBuf, cbWrite, pcbWritten);
3567}
3568
3569/**
3570 * VD async I/O interface callback for a synchronous read from the file.
3571 */
3572static int vdIOReadSyncFallback(void *pvUser, void *pvStorage, uint64_t uOffset,
3573 void *pvBuf, size_t cbRead, size_t *pcbRead)
3574{
3575 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
3576
3577 return RTFileReadAt(pStorage->File, uOffset, pvBuf, cbRead, pcbRead);
3578}
3579
3580/**
3581 * VD async I/O interface callback for a synchronous flush of the file data.
3582 */
3583static int vdIOFlushSyncFallback(void *pvUser, void *pvStorage)
3584{
3585 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
3586
3587 return RTFileFlush(pStorage->File);
3588}
3589
3590/**
3591 * VD async I/O interface callback for a asynchronous read from the file.
3592 */
3593static int vdIOReadAsyncFallback(void *pvUser, void *pStorage, uint64_t uOffset,
3594 PCRTSGSEG paSegments, size_t cSegments,
3595 size_t cbRead, void *pvCompletion,
3596 void **ppTask)
3597{
3598 return VERR_NOT_IMPLEMENTED;
3599}
3600
3601/**
3602 * VD async I/O interface callback for a asynchronous write to the file.
3603 */
3604static int vdIOWriteAsyncFallback(void *pvUser, void *pStorage, uint64_t uOffset,
3605 PCRTSGSEG paSegments, size_t cSegments,
3606 size_t cbWrite, void *pvCompletion,
3607 void **ppTask)
3608{
3609 return VERR_NOT_IMPLEMENTED;
3610}
3611
3612/**
3613 * VD async I/O interface callback for a asynchronous flush of the file data.
3614 */
3615static int vdIOFlushAsyncFallback(void *pvUser, void *pStorage,
3616 void *pvCompletion, void **ppTask)
3617{
3618 return VERR_NOT_IMPLEMENTED;
3619}
3620
3621/**
3622 * Internal - Continues an I/O context after
3623 * it was halted because of an active transfer.
3624 */
3625static int vdIoCtxContinue(PVDIOCTX pIoCtx, int rcReq)
3626{
3627 PVBOXHDD pDisk = pIoCtx->pDisk;
3628 int rc = VINF_SUCCESS;
3629
3630 VD_IS_LOCKED(pDisk);
3631
3632 if (RT_FAILURE(rcReq))
3633 ASMAtomicCmpXchgS32(&pIoCtx->rcReq, rcReq, VINF_SUCCESS);
3634
3635 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_BLOCKED))
3636 {
3637 /* Continue the transfer */
3638 rc = vdIoCtxProcessLocked(pIoCtx);
3639
3640 if ( rc == VINF_VD_ASYNC_IO_FINISHED
3641 && ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
3642 {
3643 LogFlowFunc(("I/O context completed pIoCtx=%#p\n", pIoCtx));
3644 if (pIoCtx->pIoCtxParent)
3645 {
3646 PVDIOCTX pIoCtxParent = pIoCtx->pIoCtxParent;
3647
3648 Assert(!pIoCtxParent->pIoCtxParent);
3649 if (RT_FAILURE(pIoCtx->rcReq))
3650 ASMAtomicCmpXchgS32(&pIoCtxParent->rcReq, pIoCtx->rcReq, VINF_SUCCESS);
3651
3652 ASMAtomicDecU32(&pIoCtxParent->cDataTransfersPending);
3653
3654 if (pIoCtx->enmTxDir == VDIOCTXTXDIR_WRITE)
3655 {
3656 LogFlowFunc(("I/O context transferred %u bytes for the parent pIoCtxParent=%p\n",
3657 pIoCtx->Type.Child.cbTransferParent, pIoCtxParent));
3658
3659 /* Update the parent state. */
3660 Assert(pIoCtxParent->Req.Io.cbTransferLeft >= pIoCtx->Type.Child.cbTransferParent);
3661 ASMAtomicSubU32(&pIoCtxParent->Req.Io.cbTransferLeft, pIoCtx->Type.Child.cbTransferParent);
3662 }
3663 else
3664 Assert(pIoCtx->enmTxDir == VDIOCTXTXDIR_FLUSH);
3665
3666 /*
3667 * A completed child write means that we finished growing the image.
3668 * We have to process any pending writes now.
3669 */
3670 vdIoCtxUnlockDisk(pDisk, pIoCtxParent, false /* fProcessDeferredReqs */);
3671
3672 /* Unblock the parent */
3673 pIoCtxParent->fFlags &= ~VDIOCTX_FLAGS_BLOCKED;
3674
3675 rc = vdIoCtxProcessLocked(pIoCtxParent);
3676
3677 if ( rc == VINF_VD_ASYNC_IO_FINISHED
3678 && ASMAtomicCmpXchgBool(&pIoCtxParent->fComplete, true, false))
3679 {
3680 LogFlowFunc(("Parent I/O context completed pIoCtxParent=%#p rcReq=%Rrc\n", pIoCtxParent, pIoCtxParent->rcReq));
3681 pIoCtxParent->Type.Root.pfnComplete(pIoCtxParent->Type.Root.pvUser1,
3682 pIoCtxParent->Type.Root.pvUser2,
3683 pIoCtxParent->rcReq);
3684 vdThreadFinishWrite(pDisk);
3685 vdIoCtxFree(pDisk, pIoCtxParent);
3686 vdDiskProcessBlockedIoCtx(pDisk);
3687 }
3688 else if (!vdIoCtxIsDiskLockOwner(pDisk, pIoCtx))
3689 {
3690 /* Process any pending writes if the current request didn't caused another growing. */
3691 vdDiskProcessBlockedIoCtx(pDisk);
3692 }
3693 }
3694 else
3695 {
3696 if (pIoCtx->enmTxDir == VDIOCTXTXDIR_FLUSH)
3697 {
3698 vdIoCtxUnlockDisk(pDisk, pIoCtx, true /* fProcessDerredReqs */);
3699 vdThreadFinishWrite(pDisk);
3700 }
3701 else if ( pIoCtx->enmTxDir == VDIOCTXTXDIR_WRITE
3702 || pIoCtx->enmTxDir == VDIOCTXTXDIR_DISCARD)
3703 vdThreadFinishWrite(pDisk);
3704 else
3705 {
3706 Assert(pIoCtx->enmTxDir == VDIOCTXTXDIR_READ);
3707 vdThreadFinishRead(pDisk);
3708 }
3709
3710 LogFlowFunc(("I/O context completed pIoCtx=%#p rcReq=%Rrc\n", pIoCtx, pIoCtx->rcReq));
3711 pIoCtx->Type.Root.pfnComplete(pIoCtx->Type.Root.pvUser1,
3712 pIoCtx->Type.Root.pvUser2,
3713 pIoCtx->rcReq);
3714 }
3715
3716 vdIoCtxFree(pDisk, pIoCtx);
3717 }
3718 }
3719
3720 return VINF_SUCCESS;
3721}
3722
3723/**
3724 * Internal - Called when user transfer completed.
3725 */
3726static int vdUserXferCompleted(PVDIOSTORAGE pIoStorage, PVDIOCTX pIoCtx,
3727 PFNVDXFERCOMPLETED pfnComplete, void *pvUser,
3728 size_t cbTransfer, int rcReq)
3729{
3730 int rc = VINF_SUCCESS;
3731 bool fIoCtxContinue = true;
3732 PVBOXHDD pDisk = pIoCtx->pDisk;
3733
3734 LogFlowFunc(("pIoStorage=%#p pIoCtx=%#p pfnComplete=%#p pvUser=%#p cbTransfer=%zu rcReq=%Rrc\n",
3735 pIoStorage, pIoCtx, pfnComplete, pvUser, cbTransfer, rcReq));
3736
3737 VD_IS_LOCKED(pDisk);
3738
3739 Assert(pIoCtx->Req.Io.cbTransferLeft >= cbTransfer);
3740 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, cbTransfer);
3741 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
3742
3743 if (pfnComplete)
3744 rc = pfnComplete(pIoStorage->pVDIo->pBackendData, pIoCtx, pvUser, rcReq);
3745
3746 if (RT_SUCCESS(rc))
3747 rc = vdIoCtxContinue(pIoCtx, rcReq);
3748 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
3749 rc = VINF_SUCCESS;
3750
3751 return rc;
3752}
3753
3754/**
3755 * Internal - Called when a meta transfer completed.
3756 */
3757static int vdMetaXferCompleted(PVDIOSTORAGE pIoStorage, PFNVDXFERCOMPLETED pfnComplete, void *pvUser,
3758 PVDMETAXFER pMetaXfer, int rcReq)
3759{
3760 PVBOXHDD pDisk = pIoStorage->pVDIo->pDisk;
3761 RTLISTNODE ListIoCtxWaiting;
3762 bool fFlush;
3763
3764 LogFlowFunc(("pIoStorage=%#p pfnComplete=%#p pvUser=%#p pMetaXfer=%#p rcReq=%Rrc\n",
3765 pIoStorage, pfnComplete, pvUser, pMetaXfer, rcReq));
3766
3767 VD_IS_LOCKED(pDisk);
3768
3769 fFlush = VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_FLUSH;
3770 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_NONE);
3771
3772 if (!fFlush)
3773 {
3774 RTListMove(&ListIoCtxWaiting, &pMetaXfer->ListIoCtxWaiting);
3775
3776 if (RT_FAILURE(rcReq))
3777 {
3778 /* Remove from the AVL tree. */
3779 LogFlow(("Removing meta xfer=%#p\n", pMetaXfer));
3780 bool fRemoved = RTAvlrFileOffsetRemove(pIoStorage->pTreeMetaXfers, pMetaXfer->Core.Key) != NULL;
3781 Assert(fRemoved);
3782 RTMemFree(pMetaXfer);
3783 }
3784 else
3785 {
3786 /* Increase the reference counter to make sure it doesn't go away before the last context is processed. */
3787 pMetaXfer->cRefs++;
3788 }
3789 }
3790 else
3791 RTListMove(&ListIoCtxWaiting, &pMetaXfer->ListIoCtxWaiting);
3792
3793 /* Go through the waiting list and continue the I/O contexts. */
3794 while (!RTListIsEmpty(&ListIoCtxWaiting))
3795 {
3796 int rc = VINF_SUCCESS;
3797 bool fContinue = true;
3798 PVDIOCTXDEFERRED pDeferred = RTListGetFirst(&ListIoCtxWaiting, VDIOCTXDEFERRED, NodeDeferred);
3799 PVDIOCTX pIoCtx = pDeferred->pIoCtx;
3800 RTListNodeRemove(&pDeferred->NodeDeferred);
3801
3802 RTMemFree(pDeferred);
3803 ASMAtomicDecU32(&pIoCtx->cMetaTransfersPending);
3804
3805 if (pfnComplete)
3806 rc = pfnComplete(pIoStorage->pVDIo->pBackendData, pIoCtx, pvUser, rcReq);
3807
3808 LogFlow(("Completion callback for I/O context %#p returned %Rrc\n", pIoCtx, rc));
3809
3810 if (RT_SUCCESS(rc))
3811 {
3812 rc = vdIoCtxContinue(pIoCtx, rcReq);
3813 AssertRC(rc);
3814 }
3815 else
3816 Assert(rc == VERR_VD_ASYNC_IO_IN_PROGRESS);
3817 }
3818
3819 /* Remove if not used anymore. */
3820 if (RT_SUCCESS(rcReq) && !fFlush)
3821 {
3822 pMetaXfer->cRefs--;
3823 if (!pMetaXfer->cRefs && RTListIsEmpty(&pMetaXfer->ListIoCtxWaiting))
3824 {
3825 /* Remove from the AVL tree. */
3826 LogFlow(("Removing meta xfer=%#p\n", pMetaXfer));
3827 bool fRemoved = RTAvlrFileOffsetRemove(pIoStorage->pTreeMetaXfers, pMetaXfer->Core.Key) != NULL;
3828 Assert(fRemoved);
3829 RTMemFree(pMetaXfer);
3830 }
3831 }
3832 else if (fFlush)
3833 RTMemFree(pMetaXfer);
3834
3835 return VINF_SUCCESS;
3836}
3837
3838/**
3839 * Processes a list of waiting I/O tasks. The disk lock must be held by caller.
3840 *
3841 * @returns nothing.
3842 * @param pDisk The disk to process the list for.
3843 */
3844static void vdIoTaskProcessWaitingList(PVBOXHDD pDisk)
3845{
3846 LogFlowFunc(("pDisk=%#p\n", pDisk));
3847
3848 VD_IS_LOCKED(pDisk);
3849
3850 PVDIOTASK pHead = ASMAtomicXchgPtrT(&pDisk->pIoTasksPendingHead, NULL, PVDIOTASK);
3851
3852 /* Reverse order. */
3853 PVDIOTASK pCur = pHead;
3854 pHead = NULL;
3855 while (pCur)
3856 {
3857 PVDIOTASK pInsert = pCur;
3858 pCur = pCur->pNext;
3859 pInsert->pNext = pHead;
3860 pHead = pInsert;
3861 }
3862
3863 while (pHead)
3864 {
3865 PVDIOSTORAGE pIoStorage = pHead->pIoStorage;
3866
3867 if (!pHead->fMeta)
3868 vdUserXferCompleted(pIoStorage, pHead->Type.User.pIoCtx,
3869 pHead->pfnComplete, pHead->pvUser,
3870 pHead->Type.User.cbTransfer, pHead->rcReq);
3871 else
3872 vdMetaXferCompleted(pIoStorage, pHead->pfnComplete, pHead->pvUser,
3873 pHead->Type.Meta.pMetaXfer, pHead->rcReq);
3874
3875 pCur = pHead;
3876 pHead = pHead->pNext;
3877 vdIoTaskFree(pDisk, pCur);
3878 }
3879}
3880
3881/**
3882 * Process any I/O context on the halted list.
3883 *
3884 * @returns nothing.
3885 * @param pDisk The disk.
3886 */
3887static void vdIoCtxProcessHaltedList(PVBOXHDD pDisk)
3888{
3889 LogFlowFunc(("pDisk=%#p\n", pDisk));
3890
3891 VD_IS_LOCKED(pDisk);
3892
3893 /* Get the waiting list and process it in FIFO order. */
3894 PVDIOCTX pIoCtxHead = ASMAtomicXchgPtrT(&pDisk->pIoCtxHaltedHead, NULL, PVDIOCTX);
3895
3896 /* Reverse it. */
3897 PVDIOCTX pCur = pIoCtxHead;
3898 pIoCtxHead = NULL;
3899 while (pCur)
3900 {
3901 PVDIOCTX pInsert = pCur;
3902 pCur = pCur->pIoCtxNext;
3903 pInsert->pIoCtxNext = pIoCtxHead;
3904 pIoCtxHead = pInsert;
3905 }
3906
3907 /* Process now. */
3908 pCur = pIoCtxHead;
3909 while (pCur)
3910 {
3911 int rcTmp;
3912 PVDIOCTX pTmp = pCur;
3913
3914 pCur = pCur->pIoCtxNext;
3915 pTmp->pIoCtxNext = NULL;
3916
3917 /* Continue */
3918 pTmp->fFlags &= ~VDIOCTX_FLAGS_BLOCKED;
3919 vdIoCtxContinue(pTmp, pTmp->rcReq);
3920 }
3921}
3922
3923/**
3924 * Unlock the disk and process pending tasks.
3925 *
3926 * @returns VBox status code.
3927 * @param pDisk The disk to unlock.
3928 */
3929static int vdDiskUnlock(PVBOXHDD pDisk, PVDIOCTX pIoCtxRc)
3930{
3931 int rc = VINF_SUCCESS;
3932
3933 VD_IS_LOCKED(pDisk);
3934
3935 /*
3936 * Process the list of waiting I/O tasks first
3937 * because they might complete I/O contexts.
3938 * Same for the list of halted I/O contexts.
3939 * Afterwards comes the list of new I/O contexts.
3940 */
3941 vdIoTaskProcessWaitingList(pDisk);
3942 vdIoCtxProcessHaltedList(pDisk);
3943 rc = vdDiskProcessWaitingIoCtx(pDisk, pIoCtxRc);
3944 ASMAtomicXchgBool(&pDisk->fLocked, false);
3945
3946 /*
3947 * Need to check for new I/O tasks and waiting I/O contexts now
3948 * again as other threads might added them while we processed
3949 * previous lists.
3950 */
3951 while ( ASMAtomicUoReadPtrT(&pDisk->pIoCtxHead, PVDIOCTX) != NULL
3952 || ASMAtomicUoReadPtrT(&pDisk->pIoTasksPendingHead, PVDIOTASK) != NULL
3953 || ASMAtomicUoReadPtrT(&pDisk->pIoCtxHaltedHead, PVDIOCTX) != NULL)
3954 {
3955 /* Try lock disk again. */
3956 if (ASMAtomicCmpXchgBool(&pDisk->fLocked, true, false))
3957 {
3958 vdIoTaskProcessWaitingList(pDisk);
3959 vdIoCtxProcessHaltedList(pDisk);
3960 vdDiskProcessWaitingIoCtx(pDisk, NULL);
3961 ASMAtomicXchgBool(&pDisk->fLocked, false);
3962 }
3963 else /* Let the other thread everything when he unlocks the disk. */
3964 break;
3965 }
3966
3967 return rc;
3968}
3969
3970/**
3971 * Try to lock the disk to complete pressing of the I/O task.
3972 * The completion is deferred if the disk is locked already.
3973 *
3974 * @returns nothing.
3975 * @param pIoTask The I/O task to complete.
3976 */
3977static void vdXferTryLockDiskDeferIoTask(PVDIOTASK pIoTask)
3978{
3979 PVDIOSTORAGE pIoStorage = pIoTask->pIoStorage;
3980 PVBOXHDD pDisk = pIoStorage->pVDIo->pDisk;
3981
3982 /* Put it on the waiting list. */
3983 PVDIOTASK pNext = ASMAtomicUoReadPtrT(&pDisk->pIoTasksPendingHead, PVDIOTASK);
3984 PVDIOTASK pHeadOld;
3985 pIoTask->pNext = pNext;
3986 while (!ASMAtomicCmpXchgExPtr(&pDisk->pIoTasksPendingHead, pIoTask, pNext, &pHeadOld))
3987 {
3988 pNext = pHeadOld;
3989 Assert(pNext != pIoTask);
3990 pIoTask->pNext = pNext;
3991 ASMNopPause();
3992 }
3993
3994 if (ASMAtomicCmpXchgBool(&pDisk->fLocked, true, false))
3995 {
3996 /* Release disk lock, it will take care of processing all lists. */
3997 vdDiskUnlock(pDisk, NULL);
3998 }
3999}
4000
4001static int vdIOIntReqCompleted(void *pvUser, int rcReq)
4002{
4003 PVDIOTASK pIoTask = (PVDIOTASK)pvUser;
4004
4005 LogFlowFunc(("Task completed pIoTask=%#p\n", pIoTask));
4006
4007 pIoTask->rcReq = rcReq;
4008 vdXferTryLockDiskDeferIoTask(pIoTask);
4009 return VINF_SUCCESS;
4010}
4011
4012/**
4013 * VD I/O interface callback for opening a file.
4014 */
4015static int vdIOIntOpen(void *pvUser, const char *pszLocation,
4016 unsigned uOpenFlags, PPVDIOSTORAGE ppIoStorage)
4017{
4018 int rc = VINF_SUCCESS;
4019 PVDIO pVDIo = (PVDIO)pvUser;
4020 PVDIOSTORAGE pIoStorage = (PVDIOSTORAGE)RTMemAllocZ(sizeof(VDIOSTORAGE));
4021
4022 if (!pIoStorage)
4023 return VERR_NO_MEMORY;
4024
4025 /* Create the AVl tree. */
4026 pIoStorage->pTreeMetaXfers = (PAVLRFOFFTREE)RTMemAllocZ(sizeof(AVLRFOFFTREE));
4027 if (pIoStorage->pTreeMetaXfers)
4028 {
4029 rc = pVDIo->pInterfaceIo->pfnOpen(pVDIo->pInterfaceIo->Core.pvUser,
4030 pszLocation, uOpenFlags,
4031 vdIOIntReqCompleted,
4032 &pIoStorage->pStorage);
4033 if (RT_SUCCESS(rc))
4034 {
4035 pIoStorage->pVDIo = pVDIo;
4036 *ppIoStorage = pIoStorage;
4037 return VINF_SUCCESS;
4038 }
4039
4040 RTMemFree(pIoStorage->pTreeMetaXfers);
4041 }
4042 else
4043 rc = VERR_NO_MEMORY;
4044
4045 RTMemFree(pIoStorage);
4046 return rc;
4047}
4048
4049static int vdIOIntTreeMetaXferDestroy(PAVLRFOFFNODECORE pNode, void *pvUser)
4050{
4051 AssertMsgFailed(("Tree should be empty at this point!\n"));
4052 return VINF_SUCCESS;
4053}
4054
4055static int vdIOIntClose(void *pvUser, PVDIOSTORAGE pIoStorage)
4056{
4057 PVDIO pVDIo = (PVDIO)pvUser;
4058
4059 int rc = pVDIo->pInterfaceIo->pfnClose(pVDIo->pInterfaceIo->Core.pvUser,
4060 pIoStorage->pStorage);
4061 AssertRC(rc);
4062
4063 RTAvlrFileOffsetDestroy(pIoStorage->pTreeMetaXfers, vdIOIntTreeMetaXferDestroy, NULL);
4064 RTMemFree(pIoStorage->pTreeMetaXfers);
4065 RTMemFree(pIoStorage);
4066 return VINF_SUCCESS;
4067}
4068
4069static int vdIOIntDelete(void *pvUser, const char *pcszFilename)
4070{
4071 PVDIO pVDIo = (PVDIO)pvUser;
4072 return pVDIo->pInterfaceIo->pfnDelete(pVDIo->pInterfaceIo->Core.pvUser,
4073 pcszFilename);
4074}
4075
4076static int vdIOIntMove(void *pvUser, const char *pcszSrc, const char *pcszDst,
4077 unsigned fMove)
4078{
4079 PVDIO pVDIo = (PVDIO)pvUser;
4080 return pVDIo->pInterfaceIo->pfnMove(pVDIo->pInterfaceIo->Core.pvUser,
4081 pcszSrc, pcszDst, fMove);
4082}
4083
4084static int vdIOIntGetFreeSpace(void *pvUser, const char *pcszFilename,
4085 int64_t *pcbFreeSpace)
4086{
4087 PVDIO pVDIo = (PVDIO)pvUser;
4088 return pVDIo->pInterfaceIo->pfnGetFreeSpace(pVDIo->pInterfaceIo->Core.pvUser,
4089 pcszFilename, pcbFreeSpace);
4090}
4091
4092static int vdIOIntGetModificationTime(void *pvUser, const char *pcszFilename,
4093 PRTTIMESPEC pModificationTime)
4094{
4095 PVDIO pVDIo = (PVDIO)pvUser;
4096 return pVDIo->pInterfaceIo->pfnGetModificationTime(pVDIo->pInterfaceIo->Core.pvUser,
4097 pcszFilename, pModificationTime);
4098}
4099
4100static int vdIOIntGetSize(void *pvUser, PVDIOSTORAGE pIoStorage,
4101 uint64_t *pcbSize)
4102{
4103 PVDIO pVDIo = (PVDIO)pvUser;
4104 return pVDIo->pInterfaceIo->pfnGetSize(pVDIo->pInterfaceIo->Core.pvUser,
4105 pIoStorage->pStorage, pcbSize);
4106}
4107
4108static int vdIOIntSetSize(void *pvUser, PVDIOSTORAGE pIoStorage,
4109 uint64_t cbSize)
4110{
4111 PVDIO pVDIo = (PVDIO)pvUser;
4112 return pVDIo->pInterfaceIo->pfnSetSize(pVDIo->pInterfaceIo->Core.pvUser,
4113 pIoStorage->pStorage, cbSize);
4114}
4115
4116static int vdIOIntReadUser(void *pvUser, PVDIOSTORAGE pIoStorage, uint64_t uOffset,
4117 PVDIOCTX pIoCtx, size_t cbRead)
4118{
4119 int rc = VINF_SUCCESS;
4120 PVDIO pVDIo = (PVDIO)pvUser;
4121 PVBOXHDD pDisk = pVDIo->pDisk;
4122
4123 LogFlowFunc(("pvUser=%#p pIoStorage=%#p uOffset=%llu pIoCtx=%#p cbRead=%u\n",
4124 pvUser, pIoStorage, uOffset, pIoCtx, cbRead));
4125
4126 /** @todo: Enable check for sync I/O later. */
4127 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4128 VD_IS_LOCKED(pDisk);
4129
4130 Assert(cbRead > 0);
4131
4132 if (pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
4133 {
4134 RTSGSEG Seg;
4135 unsigned cSegments = 1;
4136 size_t cbTaskRead = 0;
4137
4138 /* Synchronous I/O contexts only have one buffer segment. */
4139 AssertMsgReturn(pIoCtx->Req.Io.SgBuf.cSegs == 1,
4140 ("Invalid number of buffer segments for synchronous I/O context"),
4141 VERR_INVALID_PARAMETER);
4142
4143 cbTaskRead = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, &Seg, &cSegments, cbRead);
4144 Assert(cbRead == cbTaskRead);
4145 Assert(cSegments == 1);
4146 rc = pVDIo->pInterfaceIo->pfnReadSync(pVDIo->pInterfaceIo->Core.pvUser,
4147 pIoStorage->pStorage, uOffset,
4148 Seg.pvSeg, cbRead, NULL);
4149 }
4150 else
4151 {
4152 /* Build the S/G array and spawn a new I/O task */
4153 while (cbRead)
4154 {
4155 RTSGSEG aSeg[VD_IO_TASK_SEGMENTS_MAX];
4156 unsigned cSegments = VD_IO_TASK_SEGMENTS_MAX;
4157 size_t cbTaskRead = 0;
4158
4159 cbTaskRead = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, aSeg, &cSegments, cbRead);
4160
4161 Assert(cSegments > 0);
4162 Assert(cbTaskRead > 0);
4163 AssertMsg(cbTaskRead <= cbRead, ("Invalid number of bytes to read\n"));
4164
4165 LogFlow(("Reading %u bytes into %u segments\n", cbTaskRead, cSegments));
4166
4167#ifdef RT_STRICT
4168 for (unsigned i = 0; i < cSegments; i++)
4169 AssertMsg(aSeg[i].pvSeg && !(aSeg[i].cbSeg % 512),
4170 ("Segment %u is invalid\n", i));
4171#endif
4172
4173 PVDIOTASK pIoTask = vdIoTaskUserAlloc(pIoStorage, NULL, NULL, pIoCtx, cbTaskRead);
4174
4175 if (!pIoTask)
4176 return VERR_NO_MEMORY;
4177
4178 ASMAtomicIncU32(&pIoCtx->cDataTransfersPending);
4179
4180 void *pvTask;
4181 rc = pVDIo->pInterfaceIo->pfnReadAsync(pVDIo->pInterfaceIo->Core.pvUser,
4182 pIoStorage->pStorage, uOffset,
4183 aSeg, cSegments, cbTaskRead, pIoTask,
4184 &pvTask);
4185 if (RT_SUCCESS(rc))
4186 {
4187 AssertMsg(cbTaskRead <= pIoCtx->Req.Io.cbTransferLeft, ("Impossible!\n"));
4188 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, cbTaskRead);
4189 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
4190 vdIoTaskFree(pDisk, pIoTask);
4191 }
4192 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
4193 {
4194 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
4195 vdIoTaskFree(pDisk, pIoTask);
4196 break;
4197 }
4198
4199 uOffset += cbTaskRead;
4200 cbRead -= cbTaskRead;
4201 }
4202 }
4203
4204 LogFlowFunc(("returns rc=%Rrc\n", rc));
4205 return rc;
4206}
4207
4208static int vdIOIntWriteUser(void *pvUser, PVDIOSTORAGE pIoStorage, uint64_t uOffset,
4209 PVDIOCTX pIoCtx, size_t cbWrite, PFNVDXFERCOMPLETED pfnComplete,
4210 void *pvCompleteUser)
4211{
4212 int rc = VINF_SUCCESS;
4213 PVDIO pVDIo = (PVDIO)pvUser;
4214 PVBOXHDD pDisk = pVDIo->pDisk;
4215
4216 LogFlowFunc(("pvUser=%#p pIoStorage=%#p uOffset=%llu pIoCtx=%#p cbWrite=%u\n",
4217 pvUser, pIoStorage, uOffset, pIoCtx, cbWrite));
4218
4219 /** @todo: Enable check for sync I/O later. */
4220 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4221 VD_IS_LOCKED(pDisk);
4222
4223 Assert(cbWrite > 0);
4224
4225 if (pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
4226 {
4227 RTSGSEG Seg;
4228 unsigned cSegments = 1;
4229 size_t cbTaskWrite = 0;
4230
4231 /* Synchronous I/O contexts only have one buffer segment. */
4232 AssertMsgReturn(pIoCtx->Req.Io.SgBuf.cSegs == 1,
4233 ("Invalid number of buffer segments for synchronous I/O context"),
4234 VERR_INVALID_PARAMETER);
4235
4236 cbTaskWrite = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, &Seg, &cSegments, cbWrite);
4237 Assert(cbWrite == cbTaskWrite);
4238 Assert(cSegments == 1);
4239 rc = pVDIo->pInterfaceIo->pfnWriteSync(pVDIo->pInterfaceIo->Core.pvUser,
4240 pIoStorage->pStorage, uOffset,
4241 Seg.pvSeg, cbWrite, NULL);
4242 }
4243 else
4244 {
4245 /* Build the S/G array and spawn a new I/O task */
4246 while (cbWrite)
4247 {
4248 RTSGSEG aSeg[VD_IO_TASK_SEGMENTS_MAX];
4249 unsigned cSegments = VD_IO_TASK_SEGMENTS_MAX;
4250 size_t cbTaskWrite = 0;
4251
4252 cbTaskWrite = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, aSeg, &cSegments, cbWrite);
4253
4254 Assert(cSegments > 0);
4255 Assert(cbTaskWrite > 0);
4256 AssertMsg(cbTaskWrite <= cbWrite, ("Invalid number of bytes to write\n"));
4257
4258 LogFlow(("Writing %u bytes from %u segments\n", cbTaskWrite, cSegments));
4259
4260#ifdef DEBUG
4261 for (unsigned i = 0; i < cSegments; i++)
4262 AssertMsg(aSeg[i].pvSeg && !(aSeg[i].cbSeg % 512),
4263 ("Segment %u is invalid\n", i));
4264#endif
4265
4266 PVDIOTASK pIoTask = vdIoTaskUserAlloc(pIoStorage, pfnComplete, pvCompleteUser, pIoCtx, cbTaskWrite);
4267
4268 if (!pIoTask)
4269 return VERR_NO_MEMORY;
4270
4271 ASMAtomicIncU32(&pIoCtx->cDataTransfersPending);
4272
4273 void *pvTask;
4274 rc = pVDIo->pInterfaceIo->pfnWriteAsync(pVDIo->pInterfaceIo->Core.pvUser,
4275 pIoStorage->pStorage,
4276 uOffset, aSeg, cSegments,
4277 cbTaskWrite, pIoTask, &pvTask);
4278 if (RT_SUCCESS(rc))
4279 {
4280 AssertMsg(cbTaskWrite <= pIoCtx->Req.Io.cbTransferLeft, ("Impossible!\n"));
4281 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, cbTaskWrite);
4282 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
4283 vdIoTaskFree(pDisk, pIoTask);
4284 }
4285 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
4286 {
4287 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
4288 vdIoTaskFree(pDisk, pIoTask);
4289 break;
4290 }
4291
4292 uOffset += cbTaskWrite;
4293 cbWrite -= cbTaskWrite;
4294 }
4295 }
4296
4297 LogFlowFunc(("returns rc=%Rrc\n", rc));
4298 return rc;
4299}
4300
4301static int vdIOIntReadMeta(void *pvUser, PVDIOSTORAGE pIoStorage, uint64_t uOffset,
4302 void *pvBuf, size_t cbRead, PVDIOCTX pIoCtx,
4303 PPVDMETAXFER ppMetaXfer, PFNVDXFERCOMPLETED pfnComplete,
4304 void *pvCompleteUser)
4305{
4306 PVDIO pVDIo = (PVDIO)pvUser;
4307 PVBOXHDD pDisk = pVDIo->pDisk;
4308 int rc = VINF_SUCCESS;
4309 RTSGSEG Seg;
4310 PVDIOTASK pIoTask;
4311 PVDMETAXFER pMetaXfer = NULL;
4312 void *pvTask = NULL;
4313
4314 LogFlowFunc(("pvUser=%#p pIoStorage=%#p uOffset=%llu pvBuf=%#p cbRead=%u\n",
4315 pvUser, pIoStorage, uOffset, pvBuf, cbRead));
4316
4317 AssertMsgReturn( pIoCtx
4318 || (!ppMetaXfer && !pfnComplete && !pvCompleteUser),
4319 ("A synchronous metadata read is requested but the parameters are wrong\n"),
4320 VERR_INVALID_POINTER);
4321
4322 /** @todo: Enable check for sync I/O later. */
4323 if ( pIoCtx
4324 && !(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4325 VD_IS_LOCKED(pDisk);
4326
4327 if ( !pIoCtx
4328 || pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
4329 {
4330 /* Handle synchronous metadata I/O. */
4331 /** @todo: Integrate with metadata transfers below. */
4332 rc = pVDIo->pInterfaceIo->pfnReadSync(pVDIo->pInterfaceIo->Core.pvUser,
4333 pIoStorage->pStorage, uOffset,
4334 pvBuf, cbRead, NULL);
4335 if (ppMetaXfer)
4336 *ppMetaXfer = NULL;
4337 }
4338 else
4339 {
4340 pMetaXfer = (PVDMETAXFER)RTAvlrFileOffsetGet(pIoStorage->pTreeMetaXfers, uOffset);
4341 if (!pMetaXfer)
4342 {
4343#ifdef RT_STRICT
4344 pMetaXfer = (PVDMETAXFER)RTAvlrFileOffsetGetBestFit(pIoStorage->pTreeMetaXfers, uOffset, false /* fAbove */);
4345 AssertMsg(!pMetaXfer || (pMetaXfer->Core.Key + (RTFOFF)pMetaXfer->cbMeta <= (RTFOFF)uOffset),
4346 ("Overlapping meta transfers!\n"));
4347#endif
4348
4349 /* Allocate a new meta transfer. */
4350 pMetaXfer = vdMetaXferAlloc(pIoStorage, uOffset, cbRead);
4351 if (!pMetaXfer)
4352 return VERR_NO_MEMORY;
4353
4354 pIoTask = vdIoTaskMetaAlloc(pIoStorage, pfnComplete, pvCompleteUser, pMetaXfer);
4355 if (!pIoTask)
4356 {
4357 RTMemFree(pMetaXfer);
4358 return VERR_NO_MEMORY;
4359 }
4360
4361 Seg.cbSeg = cbRead;
4362 Seg.pvSeg = pMetaXfer->abData;
4363
4364 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_READ);
4365 rc = pVDIo->pInterfaceIo->pfnReadAsync(pVDIo->pInterfaceIo->Core.pvUser,
4366 pIoStorage->pStorage,
4367 uOffset, &Seg, 1,
4368 cbRead, pIoTask, &pvTask);
4369
4370 if (RT_SUCCESS(rc) || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4371 {
4372 bool fInserted = RTAvlrFileOffsetInsert(pIoStorage->pTreeMetaXfers, &pMetaXfer->Core);
4373 Assert(fInserted);
4374 }
4375 else
4376 RTMemFree(pMetaXfer);
4377
4378 if (RT_SUCCESS(rc))
4379 {
4380 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_NONE);
4381 vdIoTaskFree(pDisk, pIoTask);
4382 }
4383 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS && !pfnComplete)
4384 rc = VERR_VD_NOT_ENOUGH_METADATA;
4385 }
4386
4387 Assert(VALID_PTR(pMetaXfer) || RT_FAILURE(rc));
4388
4389 if (RT_SUCCESS(rc) || rc == VERR_VD_NOT_ENOUGH_METADATA || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4390 {
4391 /* If it is pending add the request to the list. */
4392 if (VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_READ)
4393 {
4394 PVDIOCTXDEFERRED pDeferred = (PVDIOCTXDEFERRED)RTMemAllocZ(sizeof(VDIOCTXDEFERRED));
4395 AssertPtr(pDeferred);
4396
4397 RTListInit(&pDeferred->NodeDeferred);
4398 pDeferred->pIoCtx = pIoCtx;
4399
4400 ASMAtomicIncU32(&pIoCtx->cMetaTransfersPending);
4401 RTListAppend(&pMetaXfer->ListIoCtxWaiting, &pDeferred->NodeDeferred);
4402 rc = VERR_VD_NOT_ENOUGH_METADATA;
4403 }
4404 else
4405 {
4406 /* Transfer the data. */
4407 pMetaXfer->cRefs++;
4408 Assert(pMetaXfer->cbMeta >= cbRead);
4409 Assert(pMetaXfer->Core.Key == (RTFOFF)uOffset);
4410 memcpy(pvBuf, pMetaXfer->abData, cbRead);
4411 *ppMetaXfer = pMetaXfer;
4412 }
4413 }
4414 }
4415
4416 LogFlowFunc(("returns rc=%Rrc\n", rc));
4417 return rc;
4418}
4419
4420static int vdIOIntWriteMeta(void *pvUser, PVDIOSTORAGE pIoStorage, uint64_t uOffset,
4421 const void *pvBuf, size_t cbWrite, PVDIOCTX pIoCtx,
4422 PFNVDXFERCOMPLETED pfnComplete, void *pvCompleteUser)
4423{
4424 PVDIO pVDIo = (PVDIO)pvUser;
4425 PVBOXHDD pDisk = pVDIo->pDisk;
4426 int rc = VINF_SUCCESS;
4427 RTSGSEG Seg;
4428 PVDIOTASK pIoTask;
4429 PVDMETAXFER pMetaXfer = NULL;
4430 bool fInTree = false;
4431 void *pvTask = NULL;
4432
4433 LogFlowFunc(("pvUser=%#p pIoStorage=%#p uOffset=%llu pvBuf=%#p cbWrite=%u\n",
4434 pvUser, pIoStorage, uOffset, pvBuf, cbWrite));
4435
4436 AssertMsgReturn( pIoCtx
4437 || (!pfnComplete && !pvCompleteUser),
4438 ("A synchronous metadata write is requested but the parameters are wrong\n"),
4439 VERR_INVALID_POINTER);
4440
4441 /** @todo: Enable check for sync I/O later. */
4442 if ( pIoCtx
4443 && !(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4444 VD_IS_LOCKED(pDisk);
4445
4446 if ( !pIoCtx
4447 || pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
4448 {
4449 /* Handle synchronous metadata I/O. */
4450 /** @todo: Integrate with metadata transfers below. */
4451 rc = pVDIo->pInterfaceIo->pfnWriteSync(pVDIo->pInterfaceIo->Core.pvUser,
4452 pIoStorage->pStorage, uOffset,
4453 pvBuf, cbWrite, NULL);
4454 }
4455 else
4456 {
4457 pMetaXfer = (PVDMETAXFER)RTAvlrFileOffsetGet(pIoStorage->pTreeMetaXfers, uOffset);
4458 if (!pMetaXfer)
4459 {
4460 /* Allocate a new meta transfer. */
4461 pMetaXfer = vdMetaXferAlloc(pIoStorage, uOffset, cbWrite);
4462 if (!pMetaXfer)
4463 return VERR_NO_MEMORY;
4464 }
4465 else
4466 {
4467 Assert(pMetaXfer->cbMeta >= cbWrite);
4468 Assert(pMetaXfer->Core.Key == (RTFOFF)uOffset);
4469 fInTree = true;
4470 }
4471
4472 Assert(VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_NONE);
4473
4474 pIoTask = vdIoTaskMetaAlloc(pIoStorage, pfnComplete, pvCompleteUser, pMetaXfer);
4475 if (!pIoTask)
4476 {
4477 RTMemFree(pMetaXfer);
4478 return VERR_NO_MEMORY;
4479 }
4480
4481 memcpy(pMetaXfer->abData, pvBuf, cbWrite);
4482 Seg.cbSeg = cbWrite;
4483 Seg.pvSeg = pMetaXfer->abData;
4484
4485 ASMAtomicIncU32(&pIoCtx->cMetaTransfersPending);
4486
4487 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_WRITE);
4488 rc = pVDIo->pInterfaceIo->pfnWriteAsync(pVDIo->pInterfaceIo->Core.pvUser,
4489 pIoStorage->pStorage,
4490 uOffset, &Seg, 1, cbWrite, pIoTask,
4491 &pvTask);
4492 if (RT_SUCCESS(rc))
4493 {
4494 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_NONE);
4495 ASMAtomicDecU32(&pIoCtx->cMetaTransfersPending);
4496 vdIoTaskFree(pDisk, pIoTask);
4497 if (fInTree && !pMetaXfer->cRefs)
4498 {
4499 LogFlow(("Removing meta xfer=%#p\n", pMetaXfer));
4500 bool fRemoved = RTAvlrFileOffsetRemove(pIoStorage->pTreeMetaXfers, pMetaXfer->Core.Key) != NULL;
4501 AssertMsg(fRemoved, ("Metadata transfer wasn't removed\n"));
4502 RTMemFree(pMetaXfer);
4503 pMetaXfer = NULL;
4504 }
4505 }
4506 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4507 {
4508 PVDIOCTXDEFERRED pDeferred = (PVDIOCTXDEFERRED)RTMemAllocZ(sizeof(VDIOCTXDEFERRED));
4509 AssertPtr(pDeferred);
4510
4511 RTListInit(&pDeferred->NodeDeferred);
4512 pDeferred->pIoCtx = pIoCtx;
4513
4514 if (!fInTree)
4515 {
4516 bool fInserted = RTAvlrFileOffsetInsert(pIoStorage->pTreeMetaXfers, &pMetaXfer->Core);
4517 Assert(fInserted);
4518 }
4519
4520 RTListAppend(&pMetaXfer->ListIoCtxWaiting, &pDeferred->NodeDeferred);
4521 }
4522 else
4523 {
4524 RTMemFree(pMetaXfer);
4525 pMetaXfer = NULL;
4526 }
4527 }
4528
4529 LogFlowFunc(("returns rc=%Rrc\n", rc));
4530 return rc;
4531}
4532
4533static void vdIOIntMetaXferRelease(void *pvUser, PVDMETAXFER pMetaXfer)
4534{
4535 PVDIO pVDIo = (PVDIO)pvUser;
4536 PVBOXHDD pDisk = pVDIo->pDisk;
4537 PVDIOSTORAGE pIoStorage;
4538
4539 /*
4540 * It is possible that we get called with a NULL metadata xfer handle
4541 * for synchronous I/O. Just exit.
4542 */
4543 if (!pMetaXfer)
4544 return;
4545
4546 pIoStorage = pMetaXfer->pIoStorage;
4547
4548 VD_IS_LOCKED(pDisk);
4549
4550 Assert( VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_NONE
4551 || VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_WRITE);
4552 Assert(pMetaXfer->cRefs > 0);
4553
4554 pMetaXfer->cRefs--;
4555 if ( !pMetaXfer->cRefs
4556 && RTListIsEmpty(&pMetaXfer->ListIoCtxWaiting)
4557 && VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_NONE)
4558 {
4559 /* Free the meta data entry. */
4560 LogFlow(("Removing meta xfer=%#p\n", pMetaXfer));
4561 bool fRemoved = RTAvlrFileOffsetRemove(pIoStorage->pTreeMetaXfers, pMetaXfer->Core.Key) != NULL;
4562 AssertMsg(fRemoved, ("Metadata transfer wasn't removed\n"));
4563
4564 RTMemFree(pMetaXfer);
4565 }
4566}
4567
4568static int vdIOIntFlush(void *pvUser, PVDIOSTORAGE pIoStorage, PVDIOCTX pIoCtx,
4569 PFNVDXFERCOMPLETED pfnComplete, void *pvCompleteUser)
4570{
4571 PVDIO pVDIo = (PVDIO)pvUser;
4572 PVBOXHDD pDisk = pVDIo->pDisk;
4573 int rc = VINF_SUCCESS;
4574 PVDIOTASK pIoTask;
4575 PVDMETAXFER pMetaXfer = NULL;
4576 void *pvTask = NULL;
4577
4578 LogFlowFunc(("pvUser=%#p pIoStorage=%#p pIoCtx=%#p\n",
4579 pvUser, pIoStorage, pIoCtx));
4580
4581 AssertMsgReturn( pIoCtx
4582 || (!pfnComplete && !pvCompleteUser),
4583 ("A synchronous metadata write is requested but the parameters are wrong\n"),
4584 VERR_INVALID_POINTER);
4585
4586 /** @todo: Enable check for sync I/O later. */
4587 if ( pIoCtx
4588 && !(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4589 VD_IS_LOCKED(pDisk);
4590
4591 if (pVDIo->fIgnoreFlush)
4592 return VINF_SUCCESS;
4593
4594 if ( !pIoCtx
4595 || pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
4596 {
4597 /* Handle synchronous flushes. */
4598 /** @todo: Integrate with metadata transfers below. */
4599 rc = pVDIo->pInterfaceIo->pfnFlushSync(pVDIo->pInterfaceIo->Core.pvUser,
4600 pIoStorage->pStorage);
4601 }
4602 else
4603 {
4604 /* Allocate a new meta transfer. */
4605 pMetaXfer = vdMetaXferAlloc(pIoStorage, 0, 0);
4606 if (!pMetaXfer)
4607 return VERR_NO_MEMORY;
4608
4609 pIoTask = vdIoTaskMetaAlloc(pIoStorage, pfnComplete, pvUser, pMetaXfer);
4610 if (!pIoTask)
4611 {
4612 RTMemFree(pMetaXfer);
4613 return VERR_NO_MEMORY;
4614 }
4615
4616 ASMAtomicIncU32(&pIoCtx->cMetaTransfersPending);
4617
4618 PVDIOCTXDEFERRED pDeferred = (PVDIOCTXDEFERRED)RTMemAllocZ(sizeof(VDIOCTXDEFERRED));
4619 AssertPtr(pDeferred);
4620
4621 RTListInit(&pDeferred->NodeDeferred);
4622 pDeferred->pIoCtx = pIoCtx;
4623
4624 RTListAppend(&pMetaXfer->ListIoCtxWaiting, &pDeferred->NodeDeferred);
4625 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_FLUSH);
4626 rc = pVDIo->pInterfaceIo->pfnFlushAsync(pVDIo->pInterfaceIo->Core.pvUser,
4627 pIoStorage->pStorage,
4628 pIoTask, &pvTask);
4629 if (RT_SUCCESS(rc))
4630 {
4631 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_NONE);
4632 ASMAtomicDecU32(&pIoCtx->cMetaTransfersPending);
4633 vdIoTaskFree(pDisk, pIoTask);
4634 RTMemFree(pDeferred);
4635 RTMemFree(pMetaXfer);
4636 }
4637 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
4638 RTMemFree(pMetaXfer);
4639 }
4640
4641 LogFlowFunc(("returns rc=%Rrc\n", rc));
4642 return rc;
4643}
4644
4645static size_t vdIOIntIoCtxCopyTo(void *pvUser, PVDIOCTX pIoCtx,
4646 const void *pvBuf, size_t cbBuf)
4647{
4648 PVDIO pVDIo = (PVDIO)pvUser;
4649 PVBOXHDD pDisk = pVDIo->pDisk;
4650 size_t cbCopied = 0;
4651
4652 /** @todo: Enable check for sync I/O later. */
4653 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4654 VD_IS_LOCKED(pDisk);
4655
4656 cbCopied = vdIoCtxCopyTo(pIoCtx, (uint8_t *)pvBuf, cbBuf);
4657 Assert(cbCopied == cbBuf);
4658
4659 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, cbCopied);
4660
4661 return cbCopied;
4662}
4663
4664static size_t vdIOIntIoCtxCopyFrom(void *pvUser, PVDIOCTX pIoCtx,
4665 void *pvBuf, size_t cbBuf)
4666{
4667 PVDIO pVDIo = (PVDIO)pvUser;
4668 PVBOXHDD pDisk = pVDIo->pDisk;
4669 size_t cbCopied = 0;
4670
4671 /** @todo: Enable check for sync I/O later. */
4672 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4673 VD_IS_LOCKED(pDisk);
4674
4675 cbCopied = vdIoCtxCopyFrom(pIoCtx, (uint8_t *)pvBuf, cbBuf);
4676 Assert(cbCopied == cbBuf);
4677
4678 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, cbCopied);
4679
4680 return cbCopied;
4681}
4682
4683static size_t vdIOIntIoCtxSet(void *pvUser, PVDIOCTX pIoCtx, int ch, size_t cb)
4684{
4685 PVDIO pVDIo = (PVDIO)pvUser;
4686 PVBOXHDD pDisk = pVDIo->pDisk;
4687 size_t cbSet = 0;
4688
4689 /** @todo: Enable check for sync I/O later. */
4690 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4691 VD_IS_LOCKED(pDisk);
4692
4693 cbSet = vdIoCtxSet(pIoCtx, ch, cb);
4694 Assert(cbSet == cb);
4695
4696 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, cbSet);
4697
4698 return cbSet;
4699}
4700
4701static size_t vdIOIntIoCtxSegArrayCreate(void *pvUser, PVDIOCTX pIoCtx,
4702 PRTSGSEG paSeg, unsigned *pcSeg,
4703 size_t cbData)
4704{
4705 PVDIO pVDIo = (PVDIO)pvUser;
4706 PVBOXHDD pDisk = pVDIo->pDisk;
4707 size_t cbCreated = 0;
4708
4709 /** @todo: Enable check for sync I/O later. */
4710 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4711 VD_IS_LOCKED(pDisk);
4712
4713 cbCreated = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, paSeg, pcSeg, cbData);
4714 Assert(!paSeg || cbData == cbCreated);
4715
4716 return cbCreated;
4717}
4718
4719static void vdIOIntIoCtxCompleted(void *pvUser, PVDIOCTX pIoCtx, int rcReq,
4720 size_t cbCompleted)
4721{
4722 PVDIO pVDIo = (PVDIO)pvUser;
4723 PVBOXHDD pDisk = pVDIo->pDisk;
4724
4725 /*
4726 * Grab the disk critical section to avoid races with other threads which
4727 * might still modify the I/O context.
4728 * Example is that iSCSI is doing an asynchronous write but calls us already
4729 * while the other thread is still hanging in vdWriteHelperAsync and couldn't update
4730 * the blocked state yet.
4731 * It can overwrite the state to true before we call vdIoCtxContinue and the
4732 * the request would hang indefinite.
4733 */
4734 pIoCtx->rcReq = rcReq;
4735 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, cbCompleted);
4736
4737 /* Clear the pointer to next transfer function in case we have nothing to transfer anymore.
4738 * @todo: Find a better way to prevent vdIoCtxContinue from calling the read/write helper again. */
4739 if (!pIoCtx->Req.Io.cbTransferLeft)
4740 pIoCtx->pfnIoCtxTransfer = NULL;
4741
4742 vdIoCtxAddToWaitingList(&pDisk->pIoCtxHaltedHead, pIoCtx);
4743 if (ASMAtomicCmpXchgBool(&pDisk->fLocked, true, false))
4744 {
4745 /* Immediately drop the lock again, it will take of processing the list. */
4746 vdDiskUnlock(pDisk, NULL);
4747 }
4748}
4749
4750static DECLCALLBACK(bool) vdIOIntIoCtxIsSynchronous(void *pvUser, PVDIOCTX pIoCtx)
4751{
4752 NOREF(pvUser);
4753 return !!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC);
4754}
4755
4756static DECLCALLBACK(bool) vdIOIntIoCtxIsZero(void *pvUser, PVDIOCTX pIoCtx, size_t cbCheck,
4757 bool fAdvance)
4758{
4759 NOREF(pvUser);
4760
4761 bool fIsZero = RTSgBufIsZero(&pIoCtx->Req.Io.SgBuf, cbCheck);
4762 if (fIsZero && fAdvance)
4763 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbCheck);
4764
4765 return fIsZero;
4766}
4767
4768/**
4769 * VD I/O interface callback for opening a file (limited version for VDGetFormat).
4770 */
4771static int vdIOIntOpenLimited(void *pvUser, const char *pszLocation,
4772 uint32_t fOpen, PPVDIOSTORAGE ppIoStorage)
4773{
4774 int rc = VINF_SUCCESS;
4775 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4776 PVDIOSTORAGE pIoStorage = (PVDIOSTORAGE)RTMemAllocZ(sizeof(VDIOSTORAGE));
4777
4778 if (!pIoStorage)
4779 return VERR_NO_MEMORY;
4780
4781 rc = pInterfaceIo->pfnOpen(NULL, pszLocation, fOpen, NULL, &pIoStorage->pStorage);
4782 if (RT_SUCCESS(rc))
4783 *ppIoStorage = pIoStorage;
4784 else
4785 RTMemFree(pIoStorage);
4786
4787 return rc;
4788}
4789
4790static int vdIOIntCloseLimited(void *pvUser, PVDIOSTORAGE pIoStorage)
4791{
4792 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4793 int rc = pInterfaceIo->pfnClose(NULL, pIoStorage->pStorage);
4794 AssertRC(rc);
4795
4796 RTMemFree(pIoStorage);
4797 return VINF_SUCCESS;
4798}
4799
4800static int vdIOIntDeleteLimited(void *pvUser, const char *pcszFilename)
4801{
4802 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4803 return pInterfaceIo->pfnDelete(NULL, pcszFilename);
4804}
4805
4806static int vdIOIntMoveLimited(void *pvUser, const char *pcszSrc,
4807 const char *pcszDst, unsigned fMove)
4808{
4809 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4810 return pInterfaceIo->pfnMove(NULL, pcszSrc, pcszDst, fMove);
4811}
4812
4813static int vdIOIntGetFreeSpaceLimited(void *pvUser, const char *pcszFilename,
4814 int64_t *pcbFreeSpace)
4815{
4816 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4817 return pInterfaceIo->pfnGetFreeSpace(NULL, pcszFilename, pcbFreeSpace);
4818}
4819
4820static int vdIOIntGetModificationTimeLimited(void *pvUser,
4821 const char *pcszFilename,
4822 PRTTIMESPEC pModificationTime)
4823{
4824 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4825 return pInterfaceIo->pfnGetModificationTime(NULL, pcszFilename, pModificationTime);
4826}
4827
4828static int vdIOIntGetSizeLimited(void *pvUser, PVDIOSTORAGE pIoStorage,
4829 uint64_t *pcbSize)
4830{
4831 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4832 return pInterfaceIo->pfnGetSize(NULL, pIoStorage->pStorage, pcbSize);
4833}
4834
4835static int vdIOIntSetSizeLimited(void *pvUser, PVDIOSTORAGE pIoStorage,
4836 uint64_t cbSize)
4837{
4838 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4839 return pInterfaceIo->pfnSetSize(NULL, pIoStorage->pStorage, cbSize);
4840}
4841
4842static int vdIOIntWriteUserLimited(void *pvUser, PVDIOSTORAGE pStorage,
4843 uint64_t uOffset, PVDIOCTX pIoCtx,
4844 size_t cbWrite,
4845 PFNVDXFERCOMPLETED pfnComplete,
4846 void *pvCompleteUser)
4847{
4848 NOREF(pvUser);
4849 NOREF(pStorage);
4850 NOREF(uOffset);
4851 NOREF(pIoCtx);
4852 NOREF(cbWrite);
4853 NOREF(pfnComplete);
4854 NOREF(pvCompleteUser);
4855 AssertMsgFailedReturn(("This needs to be implemented when called\n"), VERR_NOT_IMPLEMENTED);
4856}
4857
4858static int vdIOIntReadUserLimited(void *pvUser, PVDIOSTORAGE pStorage,
4859 uint64_t uOffset, PVDIOCTX pIoCtx,
4860 size_t cbRead)
4861{
4862 NOREF(pvUser);
4863 NOREF(pStorage);
4864 NOREF(uOffset);
4865 NOREF(pIoCtx);
4866 NOREF(cbRead);
4867 AssertMsgFailedReturn(("This needs to be implemented when called\n"), VERR_NOT_IMPLEMENTED);
4868}
4869
4870static int vdIOIntWriteMetaLimited(void *pvUser, PVDIOSTORAGE pStorage,
4871 uint64_t uOffset, const void *pvBuffer,
4872 size_t cbBuffer, PVDIOCTX pIoCtx,
4873 PFNVDXFERCOMPLETED pfnComplete,
4874 void *pvCompleteUser)
4875{
4876 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4877
4878 AssertMsgReturn(!pIoCtx && !pfnComplete && !pvCompleteUser,
4879 ("Async I/O not implemented for the limited interface"),
4880 VERR_NOT_SUPPORTED);
4881
4882 return pInterfaceIo->pfnWriteSync(NULL, pStorage->pStorage, uOffset, pvBuffer, cbBuffer, NULL);
4883}
4884
4885static int vdIOIntReadMetaLimited(void *pvUser, PVDIOSTORAGE pStorage,
4886 uint64_t uOffset, void *pvBuffer,
4887 size_t cbBuffer, PVDIOCTX pIoCtx,
4888 PPVDMETAXFER ppMetaXfer,
4889 PFNVDXFERCOMPLETED pfnComplete,
4890 void *pvCompleteUser)
4891{
4892 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4893
4894 AssertMsgReturn(!pIoCtx && !ppMetaXfer && !pfnComplete && !pvCompleteUser,
4895 ("Async I/O not implemented for the limited interface"),
4896 VERR_NOT_SUPPORTED);
4897
4898 return pInterfaceIo->pfnReadSync(NULL, pStorage->pStorage, uOffset, pvBuffer, cbBuffer, NULL);
4899}
4900
4901static int vdIOIntMetaXferReleaseLimited(void *pvUser, PVDMETAXFER pMetaXfer)
4902{
4903 /* This is a NOP in this case. */
4904 NOREF(pvUser);
4905 NOREF(pMetaXfer);
4906 return VINF_SUCCESS;
4907}
4908
4909static int vdIOIntFlushLimited(void *pvUser, PVDIOSTORAGE pStorage,
4910 PVDIOCTX pIoCtx,
4911 PFNVDXFERCOMPLETED pfnComplete,
4912 void *pvCompleteUser)
4913{
4914 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
4915
4916 AssertMsgReturn(!pIoCtx && !pfnComplete && !pvCompleteUser,
4917 ("Async I/O not implemented for the limited interface"),
4918 VERR_NOT_SUPPORTED);
4919
4920 return pInterfaceIo->pfnFlushSync(NULL, pStorage->pStorage);
4921}
4922
4923/**
4924 * internal: send output to the log (unconditionally).
4925 */
4926int vdLogMessage(void *pvUser, const char *pszFormat, va_list args)
4927{
4928 NOREF(pvUser);
4929 RTLogPrintfV(pszFormat, args);
4930 return VINF_SUCCESS;
4931}
4932
4933DECLINLINE(int) vdMessageWrapper(PVBOXHDD pDisk, const char *pszFormat, ...)
4934{
4935 va_list va;
4936 va_start(va, pszFormat);
4937 int rc = pDisk->pInterfaceError->pfnMessage(pDisk->pInterfaceError->Core.pvUser,
4938 pszFormat, va);
4939 va_end(va);
4940 return rc;
4941}
4942
4943
4944/**
4945 * internal: adjust PCHS geometry
4946 */
4947static void vdFixupPCHSGeometry(PVDGEOMETRY pPCHS, uint64_t cbSize)
4948{
4949 /* Fix broken PCHS geometry. Can happen for two reasons: either the backend
4950 * mixes up PCHS and LCHS, or the application used to create the source
4951 * image has put garbage in it. Additionally, if the PCHS geometry covers
4952 * more than the image size, set it back to the default. */
4953 if ( pPCHS->cHeads > 16
4954 || pPCHS->cSectors > 63
4955 || pPCHS->cCylinders == 0
4956 || (uint64_t)pPCHS->cHeads * pPCHS->cSectors * pPCHS->cCylinders * 512 > cbSize)
4957 {
4958 Assert(!(RT_MIN(cbSize / 512 / 16 / 63, 16383) - (uint32_t)RT_MIN(cbSize / 512 / 16 / 63, 16383)));
4959 pPCHS->cCylinders = (uint32_t)RT_MIN(cbSize / 512 / 16 / 63, 16383);
4960 pPCHS->cHeads = 16;
4961 pPCHS->cSectors = 63;
4962 }
4963}
4964
4965/**
4966 * internal: adjust PCHS geometry
4967 */
4968static void vdFixupLCHSGeometry(PVDGEOMETRY pLCHS, uint64_t cbSize)
4969{
4970 /* Fix broken LCHS geometry. Can happen for two reasons: either the backend
4971 * mixes up PCHS and LCHS, or the application used to create the source
4972 * image has put garbage in it. The fix in this case is to clear the LCHS
4973 * geometry to trigger autodetection when it is used next. If the geometry
4974 * already says "please autodetect" (cylinders=0) keep it. */
4975 if ( ( pLCHS->cHeads > 255
4976 || pLCHS->cHeads == 0
4977 || pLCHS->cSectors > 63
4978 || pLCHS->cSectors == 0)
4979 && pLCHS->cCylinders != 0)
4980 {
4981 pLCHS->cCylinders = 0;
4982 pLCHS->cHeads = 0;
4983 pLCHS->cSectors = 0;
4984 }
4985 /* Always recompute the number of cylinders stored in the LCHS
4986 * geometry if it isn't set to "autotedetect" at the moment.
4987 * This is very useful if the destination image size is
4988 * larger or smaller than the source image size. Do not modify
4989 * the number of heads and sectors. Windows guests hate it. */
4990 if ( pLCHS->cCylinders != 0
4991 && pLCHS->cHeads != 0 /* paranoia */
4992 && pLCHS->cSectors != 0 /* paranoia */)
4993 {
4994 Assert(!(RT_MIN(cbSize / 512 / pLCHS->cHeads / pLCHS->cSectors, 1024) - (uint32_t)RT_MIN(cbSize / 512 / pLCHS->cHeads / pLCHS->cSectors, 1024)));
4995 pLCHS->cCylinders = (uint32_t)RT_MIN(cbSize / 512 / pLCHS->cHeads / pLCHS->cSectors, 1024);
4996 }
4997}
4998
4999/**
5000 * Sets the I/O callbacks of the given interface to the fallback methods
5001 *
5002 * @returns nothing.
5003 * @param pIfIo The I/O interface to setup.
5004 */
5005static void vdIfIoFallbackCallbacksSetup(PVDINTERFACEIO pIfIo)
5006{
5007 pIfIo->pfnOpen = vdIOOpenFallback;
5008 pIfIo->pfnClose = vdIOCloseFallback;
5009 pIfIo->pfnDelete = vdIODeleteFallback;
5010 pIfIo->pfnMove = vdIOMoveFallback;
5011 pIfIo->pfnGetFreeSpace = vdIOGetFreeSpaceFallback;
5012 pIfIo->pfnGetModificationTime = vdIOGetModificationTimeFallback;
5013 pIfIo->pfnGetSize = vdIOGetSizeFallback;
5014 pIfIo->pfnSetSize = vdIOSetSizeFallback;
5015 pIfIo->pfnReadSync = vdIOReadSyncFallback;
5016 pIfIo->pfnWriteSync = vdIOWriteSyncFallback;
5017 pIfIo->pfnFlushSync = vdIOFlushSyncFallback;
5018 pIfIo->pfnReadAsync = vdIOReadAsyncFallback;
5019 pIfIo->pfnWriteAsync = vdIOWriteAsyncFallback;
5020 pIfIo->pfnFlushAsync = vdIOFlushAsyncFallback;
5021}
5022
5023/**
5024 * Sets the internal I/O callbacks of the given interface.
5025 *
5026 * @returns nothing.
5027 * @param pIfIoInt The internal I/O interface to setup.
5028 */
5029static void vdIfIoIntCallbacksSetup(PVDINTERFACEIOINT pIfIoInt)
5030{
5031 pIfIoInt->pfnOpen = vdIOIntOpen;
5032 pIfIoInt->pfnClose = vdIOIntClose;
5033 pIfIoInt->pfnDelete = vdIOIntDelete;
5034 pIfIoInt->pfnMove = vdIOIntMove;
5035 pIfIoInt->pfnGetFreeSpace = vdIOIntGetFreeSpace;
5036 pIfIoInt->pfnGetModificationTime = vdIOIntGetModificationTime;
5037 pIfIoInt->pfnGetSize = vdIOIntGetSize;
5038 pIfIoInt->pfnSetSize = vdIOIntSetSize;
5039 pIfIoInt->pfnReadUser = vdIOIntReadUser;
5040 pIfIoInt->pfnWriteUser = vdIOIntWriteUser;
5041 pIfIoInt->pfnReadMeta = vdIOIntReadMeta;
5042 pIfIoInt->pfnWriteMeta = vdIOIntWriteMeta;
5043 pIfIoInt->pfnMetaXferRelease = vdIOIntMetaXferRelease;
5044 pIfIoInt->pfnFlush = vdIOIntFlush;
5045 pIfIoInt->pfnIoCtxCopyFrom = vdIOIntIoCtxCopyFrom;
5046 pIfIoInt->pfnIoCtxCopyTo = vdIOIntIoCtxCopyTo;
5047 pIfIoInt->pfnIoCtxSet = vdIOIntIoCtxSet;
5048 pIfIoInt->pfnIoCtxSegArrayCreate = vdIOIntIoCtxSegArrayCreate;
5049 pIfIoInt->pfnIoCtxCompleted = vdIOIntIoCtxCompleted;
5050 pIfIoInt->pfnIoCtxIsSynchronous = vdIOIntIoCtxIsSynchronous;
5051 pIfIoInt->pfnIoCtxIsZero = vdIOIntIoCtxIsZero;
5052}
5053
5054/**
5055 * Internally used completion handler for synchronous I/O contexts.
5056 */
5057static DECLCALLBACK(void) vdIoCtxSyncComplete(void *pvUser1, void *pvUser2, int rcReq)
5058{
5059 PVBOXHDD pDisk = (PVBOXHDD)pvUser1;
5060
5061 pDisk->rcSync = rcReq;
5062 RTSemEventSignal(pDisk->hEventSemSyncIo);
5063}
5064
5065/**
5066 * Initializes HDD backends.
5067 *
5068 * @returns VBox status code.
5069 */
5070VBOXDDU_DECL(int) VDInit(void)
5071{
5072 int rc = vdAddBackends(aStaticBackends, RT_ELEMENTS(aStaticBackends));
5073 if (RT_SUCCESS(rc))
5074 {
5075 rc = vdAddCacheBackends(aStaticCacheBackends, RT_ELEMENTS(aStaticCacheBackends));
5076 if (RT_SUCCESS(rc))
5077 {
5078 rc = vdLoadDynamicBackends();
5079 if (RT_SUCCESS(rc))
5080 rc = vdLoadDynamicCacheBackends();
5081 }
5082 }
5083 LogRel(("VDInit finished\n"));
5084 return rc;
5085}
5086
5087/**
5088 * Destroys loaded HDD backends.
5089 *
5090 * @returns VBox status code.
5091 */
5092VBOXDDU_DECL(int) VDShutdown(void)
5093{
5094 PVBOXHDDBACKEND *pBackends = g_apBackends;
5095 PVDCACHEBACKEND *pCacheBackends = g_apCacheBackends;
5096 unsigned cBackends = g_cBackends;
5097
5098 if (!pBackends)
5099 return VERR_INTERNAL_ERROR;
5100
5101 g_cBackends = 0;
5102 g_apBackends = NULL;
5103
5104#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
5105 for (unsigned i = 0; i < cBackends; i++)
5106 if (pBackends[i]->hPlugin != NIL_RTLDRMOD)
5107 RTLdrClose(pBackends[i]->hPlugin);
5108#endif
5109
5110 /* Clear the supported cache backends. */
5111 cBackends = g_cCacheBackends;
5112 g_cCacheBackends = 0;
5113 g_apCacheBackends = NULL;
5114
5115#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
5116 for (unsigned i = 0; i < cBackends; i++)
5117 if (pCacheBackends[i]->hPlugin != NIL_RTLDRMOD)
5118 RTLdrClose(pCacheBackends[i]->hPlugin);
5119#endif
5120
5121 if (pCacheBackends)
5122 RTMemFree(pCacheBackends);
5123 RTMemFree(pBackends);
5124 return VINF_SUCCESS;
5125}
5126
5127
5128/**
5129 * Lists all HDD backends and their capabilities in a caller-provided buffer.
5130 *
5131 * @returns VBox status code.
5132 * VERR_BUFFER_OVERFLOW if not enough space is passed.
5133 * @param cEntriesAlloc Number of list entries available.
5134 * @param pEntries Pointer to array for the entries.
5135 * @param pcEntriesUsed Number of entries returned.
5136 */
5137VBOXDDU_DECL(int) VDBackendInfo(unsigned cEntriesAlloc, PVDBACKENDINFO pEntries,
5138 unsigned *pcEntriesUsed)
5139{
5140 int rc = VINF_SUCCESS;
5141 PRTDIR pPluginDir = NULL;
5142 unsigned cEntries = 0;
5143
5144 LogFlowFunc(("cEntriesAlloc=%u pEntries=%#p pcEntriesUsed=%#p\n", cEntriesAlloc, pEntries, pcEntriesUsed));
5145 /* Check arguments. */
5146 AssertMsgReturn(cEntriesAlloc,
5147 ("cEntriesAlloc=%u\n", cEntriesAlloc),
5148 VERR_INVALID_PARAMETER);
5149 AssertMsgReturn(VALID_PTR(pEntries),
5150 ("pEntries=%#p\n", pEntries),
5151 VERR_INVALID_PARAMETER);
5152 AssertMsgReturn(VALID_PTR(pcEntriesUsed),
5153 ("pcEntriesUsed=%#p\n", pcEntriesUsed),
5154 VERR_INVALID_PARAMETER);
5155 if (!g_apBackends)
5156 VDInit();
5157
5158 if (cEntriesAlloc < g_cBackends)
5159 {
5160 *pcEntriesUsed = g_cBackends;
5161 return VERR_BUFFER_OVERFLOW;
5162 }
5163
5164 for (unsigned i = 0; i < g_cBackends; i++)
5165 {
5166 pEntries[i].pszBackend = g_apBackends[i]->pszBackendName;
5167 pEntries[i].uBackendCaps = g_apBackends[i]->uBackendCaps;
5168 pEntries[i].paFileExtensions = g_apBackends[i]->paFileExtensions;
5169 pEntries[i].paConfigInfo = g_apBackends[i]->paConfigInfo;
5170 pEntries[i].pfnComposeLocation = g_apBackends[i]->pfnComposeLocation;
5171 pEntries[i].pfnComposeName = g_apBackends[i]->pfnComposeName;
5172 }
5173
5174 LogFlowFunc(("returns %Rrc *pcEntriesUsed=%u\n", rc, cEntries));
5175 *pcEntriesUsed = g_cBackends;
5176 return rc;
5177}
5178
5179/**
5180 * Lists the capabilities of a backend identified by its name.
5181 *
5182 * @returns VBox status code.
5183 * @param pszBackend The backend name.
5184 * @param pEntries Pointer to an entry.
5185 */
5186VBOXDDU_DECL(int) VDBackendInfoOne(const char *pszBackend, PVDBACKENDINFO pEntry)
5187{
5188 LogFlowFunc(("pszBackend=%#p pEntry=%#p\n", pszBackend, pEntry));
5189 /* Check arguments. */
5190 AssertMsgReturn(VALID_PTR(pszBackend),
5191 ("pszBackend=%#p\n", pszBackend),
5192 VERR_INVALID_PARAMETER);
5193 AssertMsgReturn(VALID_PTR(pEntry),
5194 ("pEntry=%#p\n", pEntry),
5195 VERR_INVALID_PARAMETER);
5196 if (!g_apBackends)
5197 VDInit();
5198
5199 /* Go through loaded backends. */
5200 for (unsigned i = 0; i < g_cBackends; i++)
5201 {
5202 if (!RTStrICmp(pszBackend, g_apBackends[i]->pszBackendName))
5203 {
5204 pEntry->pszBackend = g_apBackends[i]->pszBackendName;
5205 pEntry->uBackendCaps = g_apBackends[i]->uBackendCaps;
5206 pEntry->paFileExtensions = g_apBackends[i]->paFileExtensions;
5207 pEntry->paConfigInfo = g_apBackends[i]->paConfigInfo;
5208 return VINF_SUCCESS;
5209 }
5210 }
5211
5212 return VERR_NOT_FOUND;
5213}
5214
5215/**
5216 * Allocates and initializes an empty HDD container.
5217 * No image files are opened.
5218 *
5219 * @returns VBox status code.
5220 * @param pVDIfsDisk Pointer to the per-disk VD interface list.
5221 * @param enmType Type of the image container.
5222 * @param ppDisk Where to store the reference to HDD container.
5223 */
5224VBOXDDU_DECL(int) VDCreate(PVDINTERFACE pVDIfsDisk, VDTYPE enmType, PVBOXHDD *ppDisk)
5225{
5226 int rc = VINF_SUCCESS;
5227 PVBOXHDD pDisk = NULL;
5228
5229 LogFlowFunc(("pVDIfsDisk=%#p\n", pVDIfsDisk));
5230 do
5231 {
5232 /* Check arguments. */
5233 AssertMsgBreakStmt(VALID_PTR(ppDisk),
5234 ("ppDisk=%#p\n", ppDisk),
5235 rc = VERR_INVALID_PARAMETER);
5236
5237 pDisk = (PVBOXHDD)RTMemAllocZ(sizeof(VBOXHDD));
5238 if (pDisk)
5239 {
5240 pDisk->u32Signature = VBOXHDDDISK_SIGNATURE;
5241 pDisk->enmType = enmType;
5242 pDisk->cImages = 0;
5243 pDisk->pBase = NULL;
5244 pDisk->pLast = NULL;
5245 pDisk->cbSize = 0;
5246 pDisk->PCHSGeometry.cCylinders = 0;
5247 pDisk->PCHSGeometry.cHeads = 0;
5248 pDisk->PCHSGeometry.cSectors = 0;
5249 pDisk->LCHSGeometry.cCylinders = 0;
5250 pDisk->LCHSGeometry.cHeads = 0;
5251 pDisk->LCHSGeometry.cSectors = 0;
5252 pDisk->pVDIfsDisk = pVDIfsDisk;
5253 pDisk->pInterfaceError = NULL;
5254 pDisk->pInterfaceThreadSync = NULL;
5255 pDisk->pIoCtxLockOwner = NULL;
5256 pDisk->pIoCtxHead = NULL;
5257 pDisk->fLocked = false;
5258 pDisk->hEventSemSyncIo = NIL_RTSEMEVENT;
5259 pDisk->hMemCacheIoCtx = NIL_RTMEMCACHE;
5260 pDisk->hMemCacheIoTask = NIL_RTMEMCACHE;
5261
5262 rc = RTSemEventCreate(&pDisk->hEventSemSyncIo);
5263 if (RT_FAILURE(rc))
5264 break;
5265
5266 /* Create the I/O ctx cache */
5267 rc = RTMemCacheCreate(&pDisk->hMemCacheIoCtx, sizeof(VDIOCTX), 0, UINT32_MAX,
5268 NULL, NULL, NULL, 0);
5269 if (RT_FAILURE(rc))
5270 break;
5271
5272 /* Create the I/O task cache */
5273 rc = RTMemCacheCreate(&pDisk->hMemCacheIoTask, sizeof(VDIOTASK), 0, UINT32_MAX,
5274 NULL, NULL, NULL, 0);
5275 if (RT_FAILURE(rc))
5276 break;
5277
5278 pDisk->pInterfaceError = VDIfErrorGet(pVDIfsDisk);
5279 pDisk->pInterfaceThreadSync = VDIfThreadSyncGet(pVDIfsDisk);
5280
5281 *ppDisk = pDisk;
5282 }
5283 else
5284 {
5285 rc = VERR_NO_MEMORY;
5286 break;
5287 }
5288 } while (0);
5289
5290 if ( RT_FAILURE(rc)
5291 && pDisk)
5292 {
5293 if (pDisk->hEventSemSyncIo != NIL_RTSEMEVENT)
5294 RTSemEventDestroy(pDisk->hEventSemSyncIo);
5295 if (pDisk->hMemCacheIoCtx != NIL_RTMEMCACHE)
5296 RTMemCacheDestroy(pDisk->hMemCacheIoCtx);
5297 if (pDisk->hMemCacheIoTask != NIL_RTMEMCACHE)
5298 RTMemCacheDestroy(pDisk->hMemCacheIoTask);
5299 }
5300
5301 LogFlowFunc(("returns %Rrc (pDisk=%#p)\n", rc, pDisk));
5302 return rc;
5303}
5304
5305/**
5306 * Destroys HDD container.
5307 * If container has opened image files they will be closed.
5308 *
5309 * @returns VBox status code.
5310 * @param pDisk Pointer to HDD container.
5311 */
5312VBOXDDU_DECL(int) VDDestroy(PVBOXHDD pDisk)
5313{
5314 int rc = VINF_SUCCESS;
5315 LogFlowFunc(("pDisk=%#p\n", pDisk));
5316 do
5317 {
5318 /* sanity check */
5319 AssertPtrBreak(pDisk);
5320 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
5321 Assert(!pDisk->fLocked);
5322
5323 rc = VDCloseAll(pDisk);
5324 RTMemCacheDestroy(pDisk->hMemCacheIoCtx);
5325 RTMemCacheDestroy(pDisk->hMemCacheIoTask);
5326 RTSemEventDestroy(pDisk->hEventSemSyncIo);
5327 RTMemFree(pDisk);
5328 } while (0);
5329 LogFlowFunc(("returns %Rrc\n", rc));
5330 return rc;
5331}
5332
5333/**
5334 * Try to get the backend name which can use this image.
5335 *
5336 * @returns VBox status code.
5337 * VINF_SUCCESS if a plugin was found.
5338 * ppszFormat contains the string which can be used as backend name.
5339 * VERR_NOT_SUPPORTED if no backend was found.
5340 * @param pVDIfsDisk Pointer to the per-disk VD interface list.
5341 * @param pVDIfsImage Pointer to the per-image VD interface list.
5342 * @param pszFilename Name of the image file for which the backend is queried.
5343 * @param ppszFormat Receives pointer of the UTF-8 string which contains the format name.
5344 * The returned pointer must be freed using RTStrFree().
5345 */
5346VBOXDDU_DECL(int) VDGetFormat(PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
5347 const char *pszFilename, char **ppszFormat, VDTYPE *penmType)
5348{
5349 int rc = VERR_NOT_SUPPORTED;
5350 VDINTERFACEIOINT VDIfIoInt;
5351 VDINTERFACEIO VDIfIoFallback;
5352 PVDINTERFACEIO pInterfaceIo;
5353
5354 LogFlowFunc(("pszFilename=\"%s\"\n", pszFilename));
5355 /* Check arguments. */
5356 AssertMsgReturn(VALID_PTR(pszFilename) && *pszFilename,
5357 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
5358 VERR_INVALID_PARAMETER);
5359 AssertMsgReturn(VALID_PTR(ppszFormat),
5360 ("ppszFormat=%#p\n", ppszFormat),
5361 VERR_INVALID_PARAMETER);
5362 AssertMsgReturn(VALID_PTR(penmType),
5363 ("penmType=%#p\n", penmType),
5364 VERR_INVALID_PARAMETER);
5365
5366 if (!g_apBackends)
5367 VDInit();
5368
5369 pInterfaceIo = VDIfIoGet(pVDIfsImage);
5370 if (!pInterfaceIo)
5371 {
5372 /*
5373 * Caller doesn't provide an I/O interface, create our own using the
5374 * native file API.
5375 */
5376 vdIfIoFallbackCallbacksSetup(&VDIfIoFallback);
5377 pInterfaceIo = &VDIfIoFallback;
5378 }
5379
5380 /* Set up the internal I/O interface. */
5381 AssertReturn(!VDIfIoIntGet(pVDIfsImage), VERR_INVALID_PARAMETER);
5382 VDIfIoInt.pfnOpen = vdIOIntOpenLimited;
5383 VDIfIoInt.pfnClose = vdIOIntCloseLimited;
5384 VDIfIoInt.pfnDelete = vdIOIntDeleteLimited;
5385 VDIfIoInt.pfnMove = vdIOIntMoveLimited;
5386 VDIfIoInt.pfnGetFreeSpace = vdIOIntGetFreeSpaceLimited;
5387 VDIfIoInt.pfnGetModificationTime = vdIOIntGetModificationTimeLimited;
5388 VDIfIoInt.pfnGetSize = vdIOIntGetSizeLimited;
5389 VDIfIoInt.pfnSetSize = vdIOIntSetSizeLimited;
5390 VDIfIoInt.pfnReadUser = vdIOIntReadUserLimited;
5391 VDIfIoInt.pfnWriteUser = vdIOIntWriteUserLimited;
5392 VDIfIoInt.pfnReadMeta = vdIOIntReadMetaLimited;
5393 VDIfIoInt.pfnWriteMeta = vdIOIntWriteMetaLimited;
5394 VDIfIoInt.pfnFlush = vdIOIntFlushLimited;
5395 rc = VDInterfaceAdd(&VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
5396 pInterfaceIo, sizeof(VDINTERFACEIOINT), &pVDIfsImage);
5397 AssertRC(rc);
5398
5399 /* Find the backend supporting this file format. */
5400 for (unsigned i = 0; i < g_cBackends; i++)
5401 {
5402 if (g_apBackends[i]->pfnCheckIfValid)
5403 {
5404 rc = g_apBackends[i]->pfnCheckIfValid(pszFilename, pVDIfsDisk,
5405 pVDIfsImage, penmType);
5406 if ( RT_SUCCESS(rc)
5407 /* The correct backend has been found, but there is a small
5408 * incompatibility so that the file cannot be used. Stop here
5409 * and signal success - the actual open will of course fail,
5410 * but that will create a really sensible error message. */
5411 || ( rc != VERR_VD_GEN_INVALID_HEADER
5412 && rc != VERR_VD_VDI_INVALID_HEADER
5413 && rc != VERR_VD_VMDK_INVALID_HEADER
5414 && rc != VERR_VD_ISCSI_INVALID_HEADER
5415 && rc != VERR_VD_VHD_INVALID_HEADER
5416 && rc != VERR_VD_RAW_INVALID_HEADER
5417 && rc != VERR_VD_PARALLELS_INVALID_HEADER
5418 && rc != VERR_VD_DMG_INVALID_HEADER))
5419 {
5420 /* Copy the name into the new string. */
5421 char *pszFormat = RTStrDup(g_apBackends[i]->pszBackendName);
5422 if (!pszFormat)
5423 {
5424 rc = VERR_NO_MEMORY;
5425 break;
5426 }
5427 *ppszFormat = pszFormat;
5428 /* Do not consider the typical file access errors as success,
5429 * which allows the caller to deal with such issues. */
5430 if ( rc != VERR_ACCESS_DENIED
5431 && rc != VERR_PATH_NOT_FOUND
5432 && rc != VERR_FILE_NOT_FOUND)
5433 rc = VINF_SUCCESS;
5434 break;
5435 }
5436 rc = VERR_NOT_SUPPORTED;
5437 }
5438 }
5439
5440 /* Try the cache backends. */
5441 if (rc == VERR_NOT_SUPPORTED)
5442 {
5443 for (unsigned i = 0; i < g_cCacheBackends; i++)
5444 {
5445 if (g_apCacheBackends[i]->pfnProbe)
5446 {
5447 rc = g_apCacheBackends[i]->pfnProbe(pszFilename, pVDIfsDisk,
5448 pVDIfsImage);
5449 if ( RT_SUCCESS(rc)
5450 || (rc != VERR_VD_GEN_INVALID_HEADER))
5451 {
5452 /* Copy the name into the new string. */
5453 char *pszFormat = RTStrDup(g_apBackends[i]->pszBackendName);
5454 if (!pszFormat)
5455 {
5456 rc = VERR_NO_MEMORY;
5457 break;
5458 }
5459 *ppszFormat = pszFormat;
5460 rc = VINF_SUCCESS;
5461 break;
5462 }
5463 rc = VERR_NOT_SUPPORTED;
5464 }
5465 }
5466 }
5467
5468 LogFlowFunc(("returns %Rrc *ppszFormat=\"%s\"\n", rc, *ppszFormat));
5469 return rc;
5470}
5471
5472/**
5473 * Opens an image file.
5474 *
5475 * The first opened image file in HDD container must have a base image type,
5476 * others (next opened images) must be a differencing or undo images.
5477 * Linkage is checked for differencing image to be in consistence with the previously opened image.
5478 * When another differencing image is opened and the last image was opened in read/write access
5479 * mode, then the last image is reopened in read-only with deny write sharing mode. This allows
5480 * other processes to use images in read-only mode too.
5481 *
5482 * Note that the image is opened in read-only mode if a read/write open is not possible.
5483 * Use VDIsReadOnly to check open mode.
5484 *
5485 * @returns VBox status code.
5486 * @param pDisk Pointer to HDD container.
5487 * @param pszBackend Name of the image file backend to use.
5488 * @param pszFilename Name of the image file to open.
5489 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
5490 * @param pVDIfsImage Pointer to the per-image VD interface list.
5491 */
5492VBOXDDU_DECL(int) VDOpen(PVBOXHDD pDisk, const char *pszBackend,
5493 const char *pszFilename, unsigned uOpenFlags,
5494 PVDINTERFACE pVDIfsImage)
5495{
5496 int rc = VINF_SUCCESS;
5497 int rc2;
5498 bool fLockWrite = false;
5499 PVDIMAGE pImage = NULL;
5500
5501 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" uOpenFlags=%#x, pVDIfsImage=%#p\n",
5502 pDisk, pszBackend, pszFilename, uOpenFlags, pVDIfsImage));
5503
5504 do
5505 {
5506 /* sanity check */
5507 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
5508 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
5509
5510 /* Check arguments. */
5511 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
5512 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
5513 rc = VERR_INVALID_PARAMETER);
5514 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
5515 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
5516 rc = VERR_INVALID_PARAMETER);
5517 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
5518 ("uOpenFlags=%#x\n", uOpenFlags),
5519 rc = VERR_INVALID_PARAMETER);
5520 AssertMsgBreakStmt( !(uOpenFlags & VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS)
5521 || (uOpenFlags & VD_OPEN_FLAGS_READONLY),
5522 ("uOpenFlags=%#x\n", uOpenFlags),
5523 rc = VERR_INVALID_PARAMETER);
5524
5525 /*
5526 * Destroy the current discard state first which might still have pending blocks
5527 * for the currently opened image which will be switched to readonly mode.
5528 */
5529 /* Lock disk for writing, as we modify pDisk information below. */
5530 rc2 = vdThreadStartWrite(pDisk);
5531 AssertRC(rc2);
5532 fLockWrite = true;
5533 rc = vdDiscardStateDestroy(pDisk);
5534 if (RT_FAILURE(rc))
5535 break;
5536 rc2 = vdThreadFinishWrite(pDisk);
5537 AssertRC(rc2);
5538 fLockWrite = false;
5539
5540 /* Set up image descriptor. */
5541 pImage = (PVDIMAGE)RTMemAllocZ(sizeof(VDIMAGE));
5542 if (!pImage)
5543 {
5544 rc = VERR_NO_MEMORY;
5545 break;
5546 }
5547 pImage->pszFilename = RTStrDup(pszFilename);
5548 if (!pImage->pszFilename)
5549 {
5550 rc = VERR_NO_MEMORY;
5551 break;
5552 }
5553
5554 pImage->VDIo.pDisk = pDisk;
5555 pImage->pVDIfsImage = pVDIfsImage;
5556
5557 rc = vdFindBackend(pszBackend, &pImage->Backend);
5558 if (RT_FAILURE(rc))
5559 break;
5560 if (!pImage->Backend)
5561 {
5562 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
5563 N_("VD: unknown backend name '%s'"), pszBackend);
5564 break;
5565 }
5566
5567 /*
5568 * Fail if the backend can't do async I/O but the
5569 * flag is set.
5570 */
5571 if ( !(pImage->Backend->uBackendCaps & VD_CAP_ASYNC)
5572 && (uOpenFlags & VD_OPEN_FLAGS_ASYNC_IO))
5573 {
5574 rc = vdError(pDisk, VERR_NOT_SUPPORTED, RT_SRC_POS,
5575 N_("VD: Backend '%s' does not support async I/O"), pszBackend);
5576 break;
5577 }
5578
5579 /*
5580 * Fail if the backend doesn't support the discard operation but the
5581 * flag is set.
5582 */
5583 if ( !(pImage->Backend->uBackendCaps & VD_CAP_DISCARD)
5584 && (uOpenFlags & VD_OPEN_FLAGS_DISCARD))
5585 {
5586 rc = vdError(pDisk, VERR_VD_DISCARD_NOT_SUPPORTED, RT_SRC_POS,
5587 N_("VD: Backend '%s' does not support discard"), pszBackend);
5588 break;
5589 }
5590
5591 /* Set up the I/O interface. */
5592 pImage->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsImage);
5593 if (!pImage->VDIo.pInterfaceIo)
5594 {
5595 vdIfIoFallbackCallbacksSetup(&pImage->VDIo.VDIfIo);
5596 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
5597 pDisk, sizeof(VDINTERFACEIO), &pVDIfsImage);
5598 pImage->VDIo.pInterfaceIo = &pImage->VDIo.VDIfIo;
5599 }
5600
5601 /* Set up the internal I/O interface. */
5602 AssertBreakStmt(!VDIfIoIntGet(pVDIfsImage), rc = VERR_INVALID_PARAMETER);
5603 vdIfIoIntCallbacksSetup(&pImage->VDIo.VDIfIoInt);
5604 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
5605 &pImage->VDIo, sizeof(VDINTERFACEIOINT), &pImage->pVDIfsImage);
5606 AssertRC(rc);
5607
5608 pImage->uOpenFlags = uOpenFlags & (VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_DISCARD | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS);
5609 pImage->VDIo.fIgnoreFlush = (uOpenFlags & VD_OPEN_FLAGS_IGNORE_FLUSH) != 0;
5610 rc = pImage->Backend->pfnOpen(pImage->pszFilename,
5611 uOpenFlags & ~(VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS),
5612 pDisk->pVDIfsDisk,
5613 pImage->pVDIfsImage,
5614 pDisk->enmType,
5615 &pImage->pBackendData);
5616 /*
5617 * If the image is corrupted and there is a repair method try to repair it
5618 * first if it was openend in read-write mode and open again afterwards.
5619 */
5620 if ( RT_UNLIKELY(rc == VERR_VD_IMAGE_CORRUPTED)
5621 && pImage->Backend->pfnRepair)
5622 {
5623 rc = pImage->Backend->pfnRepair(pszFilename, pDisk->pVDIfsDisk, pImage->pVDIfsImage, 0 /* fFlags */);
5624 if (RT_SUCCESS(rc))
5625 rc = pImage->Backend->pfnOpen(pImage->pszFilename,
5626 uOpenFlags & ~(VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS),
5627 pDisk->pVDIfsDisk,
5628 pImage->pVDIfsImage,
5629 pDisk->enmType,
5630 &pImage->pBackendData);
5631 else
5632 {
5633 rc = vdError(pDisk, rc, RT_SRC_POS,
5634 N_("VD: error %Rrc repairing corrupted image file '%s'"), rc, pszFilename);
5635 break;
5636 }
5637 }
5638 else if (RT_UNLIKELY(rc == VERR_VD_IMAGE_CORRUPTED))
5639 {
5640 rc = vdError(pDisk, rc, RT_SRC_POS,
5641 N_("VD: Image file '%s' is corrupted and can't be opened"), pszFilename);
5642 break;
5643 }
5644
5645 /* If the open in read-write mode failed, retry in read-only mode. */
5646 if (RT_FAILURE(rc))
5647 {
5648 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY)
5649 && ( rc == VERR_ACCESS_DENIED
5650 || rc == VERR_PERMISSION_DENIED
5651 || rc == VERR_WRITE_PROTECT
5652 || rc == VERR_SHARING_VIOLATION
5653 || rc == VERR_FILE_LOCK_FAILED))
5654 rc = pImage->Backend->pfnOpen(pImage->pszFilename,
5655 (uOpenFlags & ~(VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS))
5656 | VD_OPEN_FLAGS_READONLY,
5657 pDisk->pVDIfsDisk,
5658 pImage->pVDIfsImage,
5659 pDisk->enmType,
5660 &pImage->pBackendData);
5661 if (RT_FAILURE(rc))
5662 {
5663 rc = vdError(pDisk, rc, RT_SRC_POS,
5664 N_("VD: error %Rrc opening image file '%s'"), rc, pszFilename);
5665 break;
5666 }
5667 }
5668
5669 /* Lock disk for writing, as we modify pDisk information below. */
5670 rc2 = vdThreadStartWrite(pDisk);
5671 AssertRC(rc2);
5672 fLockWrite = true;
5673
5674 pImage->VDIo.pBackendData = pImage->pBackendData;
5675
5676 /* Check image type. As the image itself has only partial knowledge
5677 * whether it's a base image or not, this info is derived here. The
5678 * base image can be fixed or normal, all others must be normal or
5679 * diff images. Some image formats don't distinguish between normal
5680 * and diff images, so this must be corrected here. */
5681 unsigned uImageFlags;
5682 uImageFlags = pImage->Backend->pfnGetImageFlags(pImage->pBackendData);
5683 if (RT_FAILURE(rc))
5684 uImageFlags = VD_IMAGE_FLAGS_NONE;
5685 if ( RT_SUCCESS(rc)
5686 && !(uOpenFlags & VD_OPEN_FLAGS_INFO))
5687 {
5688 if ( pDisk->cImages == 0
5689 && (uImageFlags & VD_IMAGE_FLAGS_DIFF))
5690 {
5691 rc = VERR_VD_INVALID_TYPE;
5692 break;
5693 }
5694 else if (pDisk->cImages != 0)
5695 {
5696 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
5697 {
5698 rc = VERR_VD_INVALID_TYPE;
5699 break;
5700 }
5701 else
5702 uImageFlags |= VD_IMAGE_FLAGS_DIFF;
5703 }
5704 }
5705
5706 /* Ensure we always get correct diff information, even if the backend
5707 * doesn't actually have a stored flag for this. It must not return
5708 * bogus information for the parent UUID if it is not a diff image. */
5709 RTUUID parentUuid;
5710 RTUuidClear(&parentUuid);
5711 rc2 = pImage->Backend->pfnGetParentUuid(pImage->pBackendData, &parentUuid);
5712 if (RT_SUCCESS(rc2) && !RTUuidIsNull(&parentUuid))
5713 uImageFlags |= VD_IMAGE_FLAGS_DIFF;
5714
5715 pImage->uImageFlags = uImageFlags;
5716
5717 /* Force sane optimization settings. It's not worth avoiding writes
5718 * to fixed size images. The overhead would have almost no payback. */
5719 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
5720 pImage->uOpenFlags |= VD_OPEN_FLAGS_HONOR_SAME;
5721
5722 /** @todo optionally check UUIDs */
5723
5724 /* Cache disk information. */
5725 pDisk->cbSize = pImage->Backend->pfnGetSize(pImage->pBackendData);
5726
5727 /* Cache PCHS geometry. */
5728 rc2 = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
5729 &pDisk->PCHSGeometry);
5730 if (RT_FAILURE(rc2))
5731 {
5732 pDisk->PCHSGeometry.cCylinders = 0;
5733 pDisk->PCHSGeometry.cHeads = 0;
5734 pDisk->PCHSGeometry.cSectors = 0;
5735 }
5736 else
5737 {
5738 /* Make sure the PCHS geometry is properly clipped. */
5739 pDisk->PCHSGeometry.cCylinders = RT_MIN(pDisk->PCHSGeometry.cCylinders, 16383);
5740 pDisk->PCHSGeometry.cHeads = RT_MIN(pDisk->PCHSGeometry.cHeads, 16);
5741 pDisk->PCHSGeometry.cSectors = RT_MIN(pDisk->PCHSGeometry.cSectors, 63);
5742 }
5743
5744 /* Cache LCHS geometry. */
5745 rc2 = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
5746 &pDisk->LCHSGeometry);
5747 if (RT_FAILURE(rc2))
5748 {
5749 pDisk->LCHSGeometry.cCylinders = 0;
5750 pDisk->LCHSGeometry.cHeads = 0;
5751 pDisk->LCHSGeometry.cSectors = 0;
5752 }
5753 else
5754 {
5755 /* Make sure the LCHS geometry is properly clipped. */
5756 pDisk->LCHSGeometry.cHeads = RT_MIN(pDisk->LCHSGeometry.cHeads, 255);
5757 pDisk->LCHSGeometry.cSectors = RT_MIN(pDisk->LCHSGeometry.cSectors, 63);
5758 }
5759
5760 if (pDisk->cImages != 0)
5761 {
5762 /* Switch previous image to read-only mode. */
5763 unsigned uOpenFlagsPrevImg;
5764 uOpenFlagsPrevImg = pDisk->pLast->Backend->pfnGetOpenFlags(pDisk->pLast->pBackendData);
5765 if (!(uOpenFlagsPrevImg & VD_OPEN_FLAGS_READONLY))
5766 {
5767 uOpenFlagsPrevImg |= VD_OPEN_FLAGS_READONLY;
5768 rc = pDisk->pLast->Backend->pfnSetOpenFlags(pDisk->pLast->pBackendData, uOpenFlagsPrevImg);
5769 }
5770 }
5771
5772 if (RT_SUCCESS(rc))
5773 {
5774 /* Image successfully opened, make it the last image. */
5775 vdAddImageToList(pDisk, pImage);
5776 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
5777 pDisk->uModified = VD_IMAGE_MODIFIED_FIRST;
5778 }
5779 else
5780 {
5781 /* Error detected, but image opened. Close image. */
5782 rc2 = pImage->Backend->pfnClose(pImage->pBackendData, false);
5783 AssertRC(rc2);
5784 pImage->pBackendData = NULL;
5785 }
5786 } while (0);
5787
5788 if (RT_UNLIKELY(fLockWrite))
5789 {
5790 rc2 = vdThreadFinishWrite(pDisk);
5791 AssertRC(rc2);
5792 }
5793
5794 if (RT_FAILURE(rc))
5795 {
5796 if (pImage)
5797 {
5798 if (pImage->pszFilename)
5799 RTStrFree(pImage->pszFilename);
5800 RTMemFree(pImage);
5801 }
5802 }
5803
5804 LogFlowFunc(("returns %Rrc\n", rc));
5805 return rc;
5806}
5807
5808/**
5809 * Opens a cache image.
5810 *
5811 * @return VBox status code.
5812 * @param pDisk Pointer to the HDD container which should use the cache image.
5813 * @param pszBackend Name of the cache file backend to use (case insensitive).
5814 * @param pszFilename Name of the cache image to open.
5815 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
5816 * @param pVDIfsCache Pointer to the per-cache VD interface list.
5817 */
5818VBOXDDU_DECL(int) VDCacheOpen(PVBOXHDD pDisk, const char *pszBackend,
5819 const char *pszFilename, unsigned uOpenFlags,
5820 PVDINTERFACE pVDIfsCache)
5821{
5822 int rc = VINF_SUCCESS;
5823 int rc2;
5824 bool fLockWrite = false;
5825 PVDCACHE pCache = NULL;
5826
5827 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" uOpenFlags=%#x, pVDIfsCache=%#p\n",
5828 pDisk, pszBackend, pszFilename, uOpenFlags, pVDIfsCache));
5829
5830 do
5831 {
5832 /* sanity check */
5833 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
5834 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
5835
5836 /* Check arguments. */
5837 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
5838 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
5839 rc = VERR_INVALID_PARAMETER);
5840 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
5841 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
5842 rc = VERR_INVALID_PARAMETER);
5843 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
5844 ("uOpenFlags=%#x\n", uOpenFlags),
5845 rc = VERR_INVALID_PARAMETER);
5846
5847 /* Set up image descriptor. */
5848 pCache = (PVDCACHE)RTMemAllocZ(sizeof(VDCACHE));
5849 if (!pCache)
5850 {
5851 rc = VERR_NO_MEMORY;
5852 break;
5853 }
5854 pCache->pszFilename = RTStrDup(pszFilename);
5855 if (!pCache->pszFilename)
5856 {
5857 rc = VERR_NO_MEMORY;
5858 break;
5859 }
5860
5861 pCache->VDIo.pDisk = pDisk;
5862 pCache->pVDIfsCache = pVDIfsCache;
5863
5864 rc = vdFindCacheBackend(pszBackend, &pCache->Backend);
5865 if (RT_FAILURE(rc))
5866 break;
5867 if (!pCache->Backend)
5868 {
5869 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
5870 N_("VD: unknown backend name '%s'"), pszBackend);
5871 break;
5872 }
5873
5874 /* Set up the I/O interface. */
5875 pCache->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsCache);
5876 if (!pCache->VDIo.pInterfaceIo)
5877 {
5878 vdIfIoFallbackCallbacksSetup(&pCache->VDIo.VDIfIo);
5879 rc = VDInterfaceAdd(&pCache->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
5880 pDisk, sizeof(VDINTERFACEIO), &pVDIfsCache);
5881 pCache->VDIo.pInterfaceIo = &pCache->VDIo.VDIfIo;
5882 }
5883
5884 /* Set up the internal I/O interface. */
5885 AssertBreakStmt(!VDIfIoIntGet(pVDIfsCache), rc = VERR_INVALID_PARAMETER);
5886 vdIfIoIntCallbacksSetup(&pCache->VDIo.VDIfIoInt);
5887 rc = VDInterfaceAdd(&pCache->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
5888 &pCache->VDIo, sizeof(VDINTERFACEIOINT), &pCache->pVDIfsCache);
5889 AssertRC(rc);
5890
5891 pCache->uOpenFlags = uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME;
5892 rc = pCache->Backend->pfnOpen(pCache->pszFilename,
5893 uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME,
5894 pDisk->pVDIfsDisk,
5895 pCache->pVDIfsCache,
5896 &pCache->pBackendData);
5897 /* If the open in read-write mode failed, retry in read-only mode. */
5898 if (RT_FAILURE(rc))
5899 {
5900 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY)
5901 && ( rc == VERR_ACCESS_DENIED
5902 || rc == VERR_PERMISSION_DENIED
5903 || rc == VERR_WRITE_PROTECT
5904 || rc == VERR_SHARING_VIOLATION
5905 || rc == VERR_FILE_LOCK_FAILED))
5906 rc = pCache->Backend->pfnOpen(pCache->pszFilename,
5907 (uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME)
5908 | VD_OPEN_FLAGS_READONLY,
5909 pDisk->pVDIfsDisk,
5910 pCache->pVDIfsCache,
5911 &pCache->pBackendData);
5912 if (RT_FAILURE(rc))
5913 {
5914 rc = vdError(pDisk, rc, RT_SRC_POS,
5915 N_("VD: error %Rrc opening image file '%s'"), rc, pszFilename);
5916 break;
5917 }
5918 }
5919
5920 /* Lock disk for writing, as we modify pDisk information below. */
5921 rc2 = vdThreadStartWrite(pDisk);
5922 AssertRC(rc2);
5923 fLockWrite = true;
5924
5925 /*
5926 * Check that the modification UUID of the cache and last image
5927 * match. If not the image was modified in-between without the cache.
5928 * The cache might contain stale data.
5929 */
5930 RTUUID UuidImage, UuidCache;
5931
5932 rc = pCache->Backend->pfnGetModificationUuid(pCache->pBackendData,
5933 &UuidCache);
5934 if (RT_SUCCESS(rc))
5935 {
5936 rc = pDisk->pLast->Backend->pfnGetModificationUuid(pDisk->pLast->pBackendData,
5937 &UuidImage);
5938 if (RT_SUCCESS(rc))
5939 {
5940 if (RTUuidCompare(&UuidImage, &UuidCache))
5941 rc = VERR_VD_CACHE_NOT_UP_TO_DATE;
5942 }
5943 }
5944
5945 /*
5946 * We assume that the user knows what he is doing if one of the images
5947 * doesn't support the modification uuid.
5948 */
5949 if (rc == VERR_NOT_SUPPORTED)
5950 rc = VINF_SUCCESS;
5951
5952 if (RT_SUCCESS(rc))
5953 {
5954 /* Cache successfully opened, make it the current one. */
5955 if (!pDisk->pCache)
5956 pDisk->pCache = pCache;
5957 else
5958 rc = VERR_VD_CACHE_ALREADY_EXISTS;
5959 }
5960
5961 if (RT_FAILURE(rc))
5962 {
5963 /* Error detected, but image opened. Close image. */
5964 rc2 = pCache->Backend->pfnClose(pCache->pBackendData, false);
5965 AssertRC(rc2);
5966 pCache->pBackendData = NULL;
5967 }
5968 } while (0);
5969
5970 if (RT_UNLIKELY(fLockWrite))
5971 {
5972 rc2 = vdThreadFinishWrite(pDisk);
5973 AssertRC(rc2);
5974 }
5975
5976 if (RT_FAILURE(rc))
5977 {
5978 if (pCache)
5979 {
5980 if (pCache->pszFilename)
5981 RTStrFree(pCache->pszFilename);
5982 RTMemFree(pCache);
5983 }
5984 }
5985
5986 LogFlowFunc(("returns %Rrc\n", rc));
5987 return rc;
5988}
5989
5990/**
5991 * Creates and opens a new base image file.
5992 *
5993 * @returns VBox status code.
5994 * @param pDisk Pointer to HDD container.
5995 * @param pszBackend Name of the image file backend to use.
5996 * @param pszFilename Name of the image file to create.
5997 * @param cbSize Image size in bytes.
5998 * @param uImageFlags Flags specifying special image features.
5999 * @param pszComment Pointer to image comment. NULL is ok.
6000 * @param pPCHSGeometry Pointer to physical disk geometry <= (16383,16,63). Not NULL.
6001 * @param pLCHSGeometry Pointer to logical disk geometry <= (x,255,63). Not NULL.
6002 * @param pUuid New UUID of the image. If NULL, a new UUID is created.
6003 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
6004 * @param pVDIfsImage Pointer to the per-image VD interface list.
6005 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
6006 */
6007VBOXDDU_DECL(int) VDCreateBase(PVBOXHDD pDisk, const char *pszBackend,
6008 const char *pszFilename, uint64_t cbSize,
6009 unsigned uImageFlags, const char *pszComment,
6010 PCVDGEOMETRY pPCHSGeometry,
6011 PCVDGEOMETRY pLCHSGeometry,
6012 PCRTUUID pUuid, unsigned uOpenFlags,
6013 PVDINTERFACE pVDIfsImage,
6014 PVDINTERFACE pVDIfsOperation)
6015{
6016 int rc = VINF_SUCCESS;
6017 int rc2;
6018 bool fLockWrite = false, fLockRead = false;
6019 PVDIMAGE pImage = NULL;
6020 RTUUID uuid;
6021
6022 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" cbSize=%llu uImageFlags=%#x pszComment=\"%s\" PCHS=%u/%u/%u LCHS=%u/%u/%u Uuid=%RTuuid uOpenFlags=%#x pVDIfsImage=%#p pVDIfsOperation=%#p\n",
6023 pDisk, pszBackend, pszFilename, cbSize, uImageFlags, pszComment,
6024 pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads,
6025 pPCHSGeometry->cSectors, pLCHSGeometry->cCylinders,
6026 pLCHSGeometry->cHeads, pLCHSGeometry->cSectors, pUuid,
6027 uOpenFlags, pVDIfsImage, pVDIfsOperation));
6028
6029 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
6030
6031 do
6032 {
6033 /* sanity check */
6034 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
6035 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
6036
6037 /* Check arguments. */
6038 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
6039 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
6040 rc = VERR_INVALID_PARAMETER);
6041 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
6042 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
6043 rc = VERR_INVALID_PARAMETER);
6044 AssertMsgBreakStmt(cbSize,
6045 ("cbSize=%llu\n", cbSize),
6046 rc = VERR_INVALID_PARAMETER);
6047 AssertMsgBreakStmt( ((uImageFlags & ~VD_IMAGE_FLAGS_MASK) == 0)
6048 || ((uImageFlags & (VD_IMAGE_FLAGS_FIXED | VD_IMAGE_FLAGS_DIFF)) != VD_IMAGE_FLAGS_FIXED),
6049 ("uImageFlags=%#x\n", uImageFlags),
6050 rc = VERR_INVALID_PARAMETER);
6051 /* The PCHS geometry fields may be 0 to leave it for later. */
6052 AssertMsgBreakStmt( VALID_PTR(pPCHSGeometry)
6053 && pPCHSGeometry->cHeads <= 16
6054 && pPCHSGeometry->cSectors <= 63,
6055 ("pPCHSGeometry=%#p PCHS=%u/%u/%u\n", pPCHSGeometry,
6056 pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads,
6057 pPCHSGeometry->cSectors),
6058 rc = VERR_INVALID_PARAMETER);
6059 /* The LCHS geometry fields may be 0 to leave it to later autodetection. */
6060 AssertMsgBreakStmt( VALID_PTR(pLCHSGeometry)
6061 && pLCHSGeometry->cHeads <= 255
6062 && pLCHSGeometry->cSectors <= 63,
6063 ("pLCHSGeometry=%#p LCHS=%u/%u/%u\n", pLCHSGeometry,
6064 pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads,
6065 pLCHSGeometry->cSectors),
6066 rc = VERR_INVALID_PARAMETER);
6067 /* The UUID may be NULL. */
6068 AssertMsgBreakStmt(pUuid == NULL || VALID_PTR(pUuid),
6069 ("pUuid=%#p UUID=%RTuuid\n", pUuid, pUuid),
6070 rc = VERR_INVALID_PARAMETER);
6071 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
6072 ("uOpenFlags=%#x\n", uOpenFlags),
6073 rc = VERR_INVALID_PARAMETER);
6074
6075 /* Check state. Needs a temporary read lock. Holding the write lock
6076 * all the time would be blocking other activities for too long. */
6077 rc2 = vdThreadStartRead(pDisk);
6078 AssertRC(rc2);
6079 fLockRead = true;
6080 AssertMsgBreakStmt(pDisk->cImages == 0,
6081 ("Create base image cannot be done with other images open\n"),
6082 rc = VERR_VD_INVALID_STATE);
6083 rc2 = vdThreadFinishRead(pDisk);
6084 AssertRC(rc2);
6085 fLockRead = false;
6086
6087 /* Set up image descriptor. */
6088 pImage = (PVDIMAGE)RTMemAllocZ(sizeof(VDIMAGE));
6089 if (!pImage)
6090 {
6091 rc = VERR_NO_MEMORY;
6092 break;
6093 }
6094 pImage->pszFilename = RTStrDup(pszFilename);
6095 if (!pImage->pszFilename)
6096 {
6097 rc = VERR_NO_MEMORY;
6098 break;
6099 }
6100 pImage->VDIo.pDisk = pDisk;
6101 pImage->pVDIfsImage = pVDIfsImage;
6102
6103 /* Set up the I/O interface. */
6104 pImage->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsImage);
6105 if (!pImage->VDIo.pInterfaceIo)
6106 {
6107 vdIfIoFallbackCallbacksSetup(&pImage->VDIo.VDIfIo);
6108 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
6109 pDisk, sizeof(VDINTERFACEIO), &pVDIfsImage);
6110 pImage->VDIo.pInterfaceIo = &pImage->VDIo.VDIfIo;
6111 }
6112
6113 /* Set up the internal I/O interface. */
6114 AssertBreakStmt(!VDIfIoIntGet(pVDIfsImage), rc = VERR_INVALID_PARAMETER);
6115 vdIfIoIntCallbacksSetup(&pImage->VDIo.VDIfIoInt);
6116 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
6117 &pImage->VDIo, sizeof(VDINTERFACEIOINT), &pImage->pVDIfsImage);
6118 AssertRC(rc);
6119
6120 rc = vdFindBackend(pszBackend, &pImage->Backend);
6121 if (RT_FAILURE(rc))
6122 break;
6123 if (!pImage->Backend)
6124 {
6125 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
6126 N_("VD: unknown backend name '%s'"), pszBackend);
6127 break;
6128 }
6129 if (!(pImage->Backend->uBackendCaps & ( VD_CAP_CREATE_FIXED
6130 | VD_CAP_CREATE_DYNAMIC)))
6131 {
6132 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
6133 N_("VD: backend '%s' cannot create base images"), pszBackend);
6134 break;
6135 }
6136
6137 /* Create UUID if the caller didn't specify one. */
6138 if (!pUuid)
6139 {
6140 rc = RTUuidCreate(&uuid);
6141 if (RT_FAILURE(rc))
6142 {
6143 rc = vdError(pDisk, rc, RT_SRC_POS,
6144 N_("VD: cannot generate UUID for image '%s'"),
6145 pszFilename);
6146 break;
6147 }
6148 pUuid = &uuid;
6149 }
6150
6151 pImage->uOpenFlags = uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME;
6152 uImageFlags &= ~VD_IMAGE_FLAGS_DIFF;
6153 pImage->VDIo.fIgnoreFlush = (uOpenFlags & VD_OPEN_FLAGS_IGNORE_FLUSH) != 0;
6154 rc = pImage->Backend->pfnCreate(pImage->pszFilename, cbSize,
6155 uImageFlags, pszComment, pPCHSGeometry,
6156 pLCHSGeometry, pUuid,
6157 uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME,
6158 0, 99,
6159 pDisk->pVDIfsDisk,
6160 pImage->pVDIfsImage,
6161 pVDIfsOperation,
6162 &pImage->pBackendData);
6163
6164 if (RT_SUCCESS(rc))
6165 {
6166 pImage->VDIo.pBackendData = pImage->pBackendData;
6167 pImage->uImageFlags = uImageFlags;
6168
6169 /* Force sane optimization settings. It's not worth avoiding writes
6170 * to fixed size images. The overhead would have almost no payback. */
6171 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
6172 pImage->uOpenFlags |= VD_OPEN_FLAGS_HONOR_SAME;
6173
6174 /* Lock disk for writing, as we modify pDisk information below. */
6175 rc2 = vdThreadStartWrite(pDisk);
6176 AssertRC(rc2);
6177 fLockWrite = true;
6178
6179 /** @todo optionally check UUIDs */
6180
6181 /* Re-check state, as the lock wasn't held and another image
6182 * creation call could have been done by another thread. */
6183 AssertMsgStmt(pDisk->cImages == 0,
6184 ("Create base image cannot be done with other images open\n"),
6185 rc = VERR_VD_INVALID_STATE);
6186 }
6187
6188 if (RT_SUCCESS(rc))
6189 {
6190 /* Cache disk information. */
6191 pDisk->cbSize = pImage->Backend->pfnGetSize(pImage->pBackendData);
6192
6193 /* Cache PCHS geometry. */
6194 rc2 = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
6195 &pDisk->PCHSGeometry);
6196 if (RT_FAILURE(rc2))
6197 {
6198 pDisk->PCHSGeometry.cCylinders = 0;
6199 pDisk->PCHSGeometry.cHeads = 0;
6200 pDisk->PCHSGeometry.cSectors = 0;
6201 }
6202 else
6203 {
6204 /* Make sure the CHS geometry is properly clipped. */
6205 pDisk->PCHSGeometry.cCylinders = RT_MIN(pDisk->PCHSGeometry.cCylinders, 16383);
6206 pDisk->PCHSGeometry.cHeads = RT_MIN(pDisk->PCHSGeometry.cHeads, 16);
6207 pDisk->PCHSGeometry.cSectors = RT_MIN(pDisk->PCHSGeometry.cSectors, 63);
6208 }
6209
6210 /* Cache LCHS geometry. */
6211 rc2 = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
6212 &pDisk->LCHSGeometry);
6213 if (RT_FAILURE(rc2))
6214 {
6215 pDisk->LCHSGeometry.cCylinders = 0;
6216 pDisk->LCHSGeometry.cHeads = 0;
6217 pDisk->LCHSGeometry.cSectors = 0;
6218 }
6219 else
6220 {
6221 /* Make sure the CHS geometry is properly clipped. */
6222 pDisk->LCHSGeometry.cHeads = RT_MIN(pDisk->LCHSGeometry.cHeads, 255);
6223 pDisk->LCHSGeometry.cSectors = RT_MIN(pDisk->LCHSGeometry.cSectors, 63);
6224 }
6225
6226 /* Image successfully opened, make it the last image. */
6227 vdAddImageToList(pDisk, pImage);
6228 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
6229 pDisk->uModified = VD_IMAGE_MODIFIED_FIRST;
6230 }
6231 else
6232 {
6233 /* Error detected, image may or may not be opened. Close and delete
6234 * image if it was opened. */
6235 if (pImage->pBackendData)
6236 {
6237 rc2 = pImage->Backend->pfnClose(pImage->pBackendData, true);
6238 AssertRC(rc2);
6239 pImage->pBackendData = NULL;
6240 }
6241 }
6242 } while (0);
6243
6244 if (RT_UNLIKELY(fLockWrite))
6245 {
6246 rc2 = vdThreadFinishWrite(pDisk);
6247 AssertRC(rc2);
6248 }
6249 else if (RT_UNLIKELY(fLockRead))
6250 {
6251 rc2 = vdThreadFinishRead(pDisk);
6252 AssertRC(rc2);
6253 }
6254
6255 if (RT_FAILURE(rc))
6256 {
6257 if (pImage)
6258 {
6259 if (pImage->pszFilename)
6260 RTStrFree(pImage->pszFilename);
6261 RTMemFree(pImage);
6262 }
6263 }
6264
6265 if (RT_SUCCESS(rc) && pIfProgress && pIfProgress->pfnProgress)
6266 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
6267
6268 LogFlowFunc(("returns %Rrc\n", rc));
6269 return rc;
6270}
6271
6272/**
6273 * Creates and opens a new differencing image file in HDD container.
6274 * See comments for VDOpen function about differencing images.
6275 *
6276 * @returns VBox status code.
6277 * @param pDisk Pointer to HDD container.
6278 * @param pszBackend Name of the image file backend to use.
6279 * @param pszFilename Name of the differencing image file to create.
6280 * @param uImageFlags Flags specifying special image features.
6281 * @param pszComment Pointer to image comment. NULL is ok.
6282 * @param pUuid New UUID of the image. If NULL, a new UUID is created.
6283 * @param pParentUuid New parent UUID of the image. If NULL, the UUID is queried automatically.
6284 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
6285 * @param pVDIfsImage Pointer to the per-image VD interface list.
6286 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
6287 */
6288VBOXDDU_DECL(int) VDCreateDiff(PVBOXHDD pDisk, const char *pszBackend,
6289 const char *pszFilename, unsigned uImageFlags,
6290 const char *pszComment, PCRTUUID pUuid,
6291 PCRTUUID pParentUuid, unsigned uOpenFlags,
6292 PVDINTERFACE pVDIfsImage,
6293 PVDINTERFACE pVDIfsOperation)
6294{
6295 int rc = VINF_SUCCESS;
6296 int rc2;
6297 bool fLockWrite = false, fLockRead = false;
6298 PVDIMAGE pImage = NULL;
6299 RTUUID uuid;
6300
6301 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" uImageFlags=%#x pszComment=\"%s\" Uuid=%RTuuid uOpenFlags=%#x pVDIfsImage=%#p pVDIfsOperation=%#p\n",
6302 pDisk, pszBackend, pszFilename, uImageFlags, pszComment, pUuid, uOpenFlags, pVDIfsImage, pVDIfsOperation));
6303
6304 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
6305
6306 do
6307 {
6308 /* sanity check */
6309 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
6310 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
6311
6312 /* Check arguments. */
6313 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
6314 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
6315 rc = VERR_INVALID_PARAMETER);
6316 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
6317 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
6318 rc = VERR_INVALID_PARAMETER);
6319 AssertMsgBreakStmt((uImageFlags & ~VD_IMAGE_FLAGS_MASK) == 0,
6320 ("uImageFlags=%#x\n", uImageFlags),
6321 rc = VERR_INVALID_PARAMETER);
6322 /* The UUID may be NULL. */
6323 AssertMsgBreakStmt(pUuid == NULL || VALID_PTR(pUuid),
6324 ("pUuid=%#p UUID=%RTuuid\n", pUuid, pUuid),
6325 rc = VERR_INVALID_PARAMETER);
6326 /* The parent UUID may be NULL. */
6327 AssertMsgBreakStmt(pParentUuid == NULL || VALID_PTR(pParentUuid),
6328 ("pParentUuid=%#p ParentUUID=%RTuuid\n", pParentUuid, pParentUuid),
6329 rc = VERR_INVALID_PARAMETER);
6330 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
6331 ("uOpenFlags=%#x\n", uOpenFlags),
6332 rc = VERR_INVALID_PARAMETER);
6333
6334 /* Check state. Needs a temporary read lock. Holding the write lock
6335 * all the time would be blocking other activities for too long. */
6336 rc2 = vdThreadStartRead(pDisk);
6337 AssertRC(rc2);
6338 fLockRead = true;
6339 AssertMsgBreakStmt(pDisk->cImages != 0,
6340 ("Create diff image cannot be done without other images open\n"),
6341 rc = VERR_VD_INVALID_STATE);
6342 rc2 = vdThreadFinishRead(pDisk);
6343 AssertRC(rc2);
6344 fLockRead = false;
6345
6346 /*
6347 * Destroy the current discard state first which might still have pending blocks
6348 * for the currently opened image which will be switched to readonly mode.
6349 */
6350 /* Lock disk for writing, as we modify pDisk information below. */
6351 rc2 = vdThreadStartWrite(pDisk);
6352 AssertRC(rc2);
6353 fLockWrite = true;
6354 rc = vdDiscardStateDestroy(pDisk);
6355 if (RT_FAILURE(rc))
6356 break;
6357 rc2 = vdThreadFinishWrite(pDisk);
6358 AssertRC(rc2);
6359 fLockWrite = false;
6360
6361 /* Set up image descriptor. */
6362 pImage = (PVDIMAGE)RTMemAllocZ(sizeof(VDIMAGE));
6363 if (!pImage)
6364 {
6365 rc = VERR_NO_MEMORY;
6366 break;
6367 }
6368 pImage->pszFilename = RTStrDup(pszFilename);
6369 if (!pImage->pszFilename)
6370 {
6371 rc = VERR_NO_MEMORY;
6372 break;
6373 }
6374
6375 rc = vdFindBackend(pszBackend, &pImage->Backend);
6376 if (RT_FAILURE(rc))
6377 break;
6378 if (!pImage->Backend)
6379 {
6380 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
6381 N_("VD: unknown backend name '%s'"), pszBackend);
6382 break;
6383 }
6384 if ( !(pImage->Backend->uBackendCaps & VD_CAP_DIFF)
6385 || !(pImage->Backend->uBackendCaps & ( VD_CAP_CREATE_FIXED
6386 | VD_CAP_CREATE_DYNAMIC)))
6387 {
6388 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
6389 N_("VD: backend '%s' cannot create diff images"), pszBackend);
6390 break;
6391 }
6392
6393 pImage->VDIo.pDisk = pDisk;
6394 pImage->pVDIfsImage = pVDIfsImage;
6395
6396 /* Set up the I/O interface. */
6397 pImage->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsImage);
6398 if (!pImage->VDIo.pInterfaceIo)
6399 {
6400 vdIfIoFallbackCallbacksSetup(&pImage->VDIo.VDIfIo);
6401 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
6402 pDisk, sizeof(VDINTERFACEIO), &pVDIfsImage);
6403 pImage->VDIo.pInterfaceIo = &pImage->VDIo.VDIfIo;
6404 }
6405
6406 /* Set up the internal I/O interface. */
6407 AssertBreakStmt(!VDIfIoIntGet(pVDIfsImage), rc = VERR_INVALID_PARAMETER);
6408 vdIfIoIntCallbacksSetup(&pImage->VDIo.VDIfIoInt);
6409 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
6410 &pImage->VDIo, sizeof(VDINTERFACEIOINT), &pImage->pVDIfsImage);
6411 AssertRC(rc);
6412
6413 /* Create UUID if the caller didn't specify one. */
6414 if (!pUuid)
6415 {
6416 rc = RTUuidCreate(&uuid);
6417 if (RT_FAILURE(rc))
6418 {
6419 rc = vdError(pDisk, rc, RT_SRC_POS,
6420 N_("VD: cannot generate UUID for image '%s'"),
6421 pszFilename);
6422 break;
6423 }
6424 pUuid = &uuid;
6425 }
6426
6427 pImage->uOpenFlags = uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME;
6428 pImage->VDIo.fIgnoreFlush = (uOpenFlags & VD_OPEN_FLAGS_IGNORE_FLUSH) != 0;
6429 uImageFlags |= VD_IMAGE_FLAGS_DIFF;
6430 rc = pImage->Backend->pfnCreate(pImage->pszFilename, pDisk->cbSize,
6431 uImageFlags | VD_IMAGE_FLAGS_DIFF,
6432 pszComment, &pDisk->PCHSGeometry,
6433 &pDisk->LCHSGeometry, pUuid,
6434 uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME,
6435 0, 99,
6436 pDisk->pVDIfsDisk,
6437 pImage->pVDIfsImage,
6438 pVDIfsOperation,
6439 &pImage->pBackendData);
6440
6441 if (RT_SUCCESS(rc))
6442 {
6443 pImage->VDIo.pBackendData = pImage->pBackendData;
6444 pImage->uImageFlags = uImageFlags;
6445
6446 /* Lock disk for writing, as we modify pDisk information below. */
6447 rc2 = vdThreadStartWrite(pDisk);
6448 AssertRC(rc2);
6449 fLockWrite = true;
6450
6451 /* Switch previous image to read-only mode. */
6452 unsigned uOpenFlagsPrevImg;
6453 uOpenFlagsPrevImg = pDisk->pLast->Backend->pfnGetOpenFlags(pDisk->pLast->pBackendData);
6454 if (!(uOpenFlagsPrevImg & VD_OPEN_FLAGS_READONLY))
6455 {
6456 uOpenFlagsPrevImg |= VD_OPEN_FLAGS_READONLY;
6457 rc = pDisk->pLast->Backend->pfnSetOpenFlags(pDisk->pLast->pBackendData, uOpenFlagsPrevImg);
6458 }
6459
6460 /** @todo optionally check UUIDs */
6461
6462 /* Re-check state, as the lock wasn't held and another image
6463 * creation call could have been done by another thread. */
6464 AssertMsgStmt(pDisk->cImages != 0,
6465 ("Create diff image cannot be done without other images open\n"),
6466 rc = VERR_VD_INVALID_STATE);
6467 }
6468
6469 if (RT_SUCCESS(rc))
6470 {
6471 RTUUID Uuid;
6472 RTTIMESPEC ts;
6473
6474 if (pParentUuid && !RTUuidIsNull(pParentUuid))
6475 {
6476 Uuid = *pParentUuid;
6477 pImage->Backend->pfnSetParentUuid(pImage->pBackendData, &Uuid);
6478 }
6479 else
6480 {
6481 rc2 = pDisk->pLast->Backend->pfnGetUuid(pDisk->pLast->pBackendData,
6482 &Uuid);
6483 if (RT_SUCCESS(rc2))
6484 pImage->Backend->pfnSetParentUuid(pImage->pBackendData, &Uuid);
6485 }
6486 rc2 = pDisk->pLast->Backend->pfnGetModificationUuid(pDisk->pLast->pBackendData,
6487 &Uuid);
6488 if (RT_SUCCESS(rc2))
6489 pImage->Backend->pfnSetParentModificationUuid(pImage->pBackendData,
6490 &Uuid);
6491 if (pDisk->pLast->Backend->pfnGetTimeStamp)
6492 rc2 = pDisk->pLast->Backend->pfnGetTimeStamp(pDisk->pLast->pBackendData,
6493 &ts);
6494 else
6495 rc2 = VERR_NOT_IMPLEMENTED;
6496 if (RT_SUCCESS(rc2) && pImage->Backend->pfnSetParentTimeStamp)
6497 pImage->Backend->pfnSetParentTimeStamp(pImage->pBackendData, &ts);
6498
6499 if (pImage->Backend->pfnSetParentFilename)
6500 rc2 = pImage->Backend->pfnSetParentFilename(pImage->pBackendData, pDisk->pLast->pszFilename);
6501 }
6502
6503 if (RT_SUCCESS(rc))
6504 {
6505 /* Image successfully opened, make it the last image. */
6506 vdAddImageToList(pDisk, pImage);
6507 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
6508 pDisk->uModified = VD_IMAGE_MODIFIED_FIRST;
6509 }
6510 else
6511 {
6512 /* Error detected, but image opened. Close and delete image. */
6513 rc2 = pImage->Backend->pfnClose(pImage->pBackendData, true);
6514 AssertRC(rc2);
6515 pImage->pBackendData = NULL;
6516 }
6517 } while (0);
6518
6519 if (RT_UNLIKELY(fLockWrite))
6520 {
6521 rc2 = vdThreadFinishWrite(pDisk);
6522 AssertRC(rc2);
6523 }
6524 else if (RT_UNLIKELY(fLockRead))
6525 {
6526 rc2 = vdThreadFinishRead(pDisk);
6527 AssertRC(rc2);
6528 }
6529
6530 if (RT_FAILURE(rc))
6531 {
6532 if (pImage)
6533 {
6534 if (pImage->pszFilename)
6535 RTStrFree(pImage->pszFilename);
6536 RTMemFree(pImage);
6537 }
6538 }
6539
6540 if (RT_SUCCESS(rc) && pIfProgress && pIfProgress->pfnProgress)
6541 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
6542
6543 LogFlowFunc(("returns %Rrc\n", rc));
6544 return rc;
6545}
6546
6547
6548/**
6549 * Creates and opens new cache image file in HDD container.
6550 *
6551 * @return VBox status code.
6552 * @param pDisk Name of the cache file backend to use (case insensitive).
6553 * @param pszFilename Name of the differencing cache file to create.
6554 * @param cbSize Maximum size of the cache.
6555 * @param uImageFlags Flags specifying special cache features.
6556 * @param pszComment Pointer to image comment. NULL is ok.
6557 * @param pUuid New UUID of the image. If NULL, a new UUID is created.
6558 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
6559 * @param pVDIfsCache Pointer to the per-cache VD interface list.
6560 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
6561 */
6562VBOXDDU_DECL(int) VDCreateCache(PVBOXHDD pDisk, const char *pszBackend,
6563 const char *pszFilename, uint64_t cbSize,
6564 unsigned uImageFlags, const char *pszComment,
6565 PCRTUUID pUuid, unsigned uOpenFlags,
6566 PVDINTERFACE pVDIfsCache, PVDINTERFACE pVDIfsOperation)
6567{
6568 int rc = VINF_SUCCESS;
6569 int rc2;
6570 bool fLockWrite = false, fLockRead = false;
6571 PVDCACHE pCache = NULL;
6572 RTUUID uuid;
6573
6574 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" cbSize=%llu uImageFlags=%#x pszComment=\"%s\" Uuid=%RTuuid uOpenFlags=%#x pVDIfsImage=%#p pVDIfsOperation=%#p\n",
6575 pDisk, pszBackend, pszFilename, cbSize, uImageFlags, pszComment, pUuid, uOpenFlags, pVDIfsCache, pVDIfsOperation));
6576
6577 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
6578
6579 do
6580 {
6581 /* sanity check */
6582 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
6583 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
6584
6585 /* Check arguments. */
6586 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
6587 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
6588 rc = VERR_INVALID_PARAMETER);
6589 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
6590 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
6591 rc = VERR_INVALID_PARAMETER);
6592 AssertMsgBreakStmt(cbSize,
6593 ("cbSize=%llu\n", cbSize),
6594 rc = VERR_INVALID_PARAMETER);
6595 AssertMsgBreakStmt((uImageFlags & ~VD_IMAGE_FLAGS_MASK) == 0,
6596 ("uImageFlags=%#x\n", uImageFlags),
6597 rc = VERR_INVALID_PARAMETER);
6598 /* The UUID may be NULL. */
6599 AssertMsgBreakStmt(pUuid == NULL || VALID_PTR(pUuid),
6600 ("pUuid=%#p UUID=%RTuuid\n", pUuid, pUuid),
6601 rc = VERR_INVALID_PARAMETER);
6602 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
6603 ("uOpenFlags=%#x\n", uOpenFlags),
6604 rc = VERR_INVALID_PARAMETER);
6605
6606 /* Check state. Needs a temporary read lock. Holding the write lock
6607 * all the time would be blocking other activities for too long. */
6608 rc2 = vdThreadStartRead(pDisk);
6609 AssertRC(rc2);
6610 fLockRead = true;
6611 AssertMsgBreakStmt(!pDisk->pCache,
6612 ("Create cache image cannot be done with a cache already attached\n"),
6613 rc = VERR_VD_CACHE_ALREADY_EXISTS);
6614 rc2 = vdThreadFinishRead(pDisk);
6615 AssertRC(rc2);
6616 fLockRead = false;
6617
6618 /* Set up image descriptor. */
6619 pCache = (PVDCACHE)RTMemAllocZ(sizeof(VDCACHE));
6620 if (!pCache)
6621 {
6622 rc = VERR_NO_MEMORY;
6623 break;
6624 }
6625 pCache->pszFilename = RTStrDup(pszFilename);
6626 if (!pCache->pszFilename)
6627 {
6628 rc = VERR_NO_MEMORY;
6629 break;
6630 }
6631
6632 rc = vdFindCacheBackend(pszBackend, &pCache->Backend);
6633 if (RT_FAILURE(rc))
6634 break;
6635 if (!pCache->Backend)
6636 {
6637 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
6638 N_("VD: unknown backend name '%s'"), pszBackend);
6639 break;
6640 }
6641
6642 pCache->VDIo.pDisk = pDisk;
6643 pCache->pVDIfsCache = pVDIfsCache;
6644
6645 /* Set up the I/O interface. */
6646 pCache->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsCache);
6647 if (!pCache->VDIo.pInterfaceIo)
6648 {
6649 vdIfIoFallbackCallbacksSetup(&pCache->VDIo.VDIfIo);
6650 rc = VDInterfaceAdd(&pCache->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
6651 pDisk, sizeof(VDINTERFACEIO), &pVDIfsCache);
6652 pCache->VDIo.pInterfaceIo = &pCache->VDIo.VDIfIo;
6653 }
6654
6655 /* Set up the internal I/O interface. */
6656 AssertBreakStmt(!VDIfIoIntGet(pVDIfsCache), rc = VERR_INVALID_PARAMETER);
6657 vdIfIoIntCallbacksSetup(&pCache->VDIo.VDIfIoInt);
6658 rc = VDInterfaceAdd(&pCache->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
6659 &pCache->VDIo, sizeof(VDINTERFACEIOINT), &pCache->pVDIfsCache);
6660 AssertRC(rc);
6661
6662 /* Create UUID if the caller didn't specify one. */
6663 if (!pUuid)
6664 {
6665 rc = RTUuidCreate(&uuid);
6666 if (RT_FAILURE(rc))
6667 {
6668 rc = vdError(pDisk, rc, RT_SRC_POS,
6669 N_("VD: cannot generate UUID for image '%s'"),
6670 pszFilename);
6671 break;
6672 }
6673 pUuid = &uuid;
6674 }
6675
6676 pCache->uOpenFlags = uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME;
6677 pCache->VDIo.fIgnoreFlush = (uOpenFlags & VD_OPEN_FLAGS_IGNORE_FLUSH) != 0;
6678 rc = pCache->Backend->pfnCreate(pCache->pszFilename, cbSize,
6679 uImageFlags,
6680 pszComment, pUuid,
6681 uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME,
6682 0, 99,
6683 pDisk->pVDIfsDisk,
6684 pCache->pVDIfsCache,
6685 pVDIfsOperation,
6686 &pCache->pBackendData);
6687
6688 if (RT_SUCCESS(rc))
6689 {
6690 /* Lock disk for writing, as we modify pDisk information below. */
6691 rc2 = vdThreadStartWrite(pDisk);
6692 AssertRC(rc2);
6693 fLockWrite = true;
6694
6695 pCache->VDIo.pBackendData = pCache->pBackendData;
6696
6697 /* Re-check state, as the lock wasn't held and another image
6698 * creation call could have been done by another thread. */
6699 AssertMsgStmt(!pDisk->pCache,
6700 ("Create cache image cannot be done with another cache open\n"),
6701 rc = VERR_VD_CACHE_ALREADY_EXISTS);
6702 }
6703
6704 if ( RT_SUCCESS(rc)
6705 && pDisk->pLast)
6706 {
6707 RTUUID UuidModification;
6708
6709 /* Set same modification Uuid as the last image. */
6710 rc = pDisk->pLast->Backend->pfnGetModificationUuid(pDisk->pLast->pBackendData,
6711 &UuidModification);
6712 if (RT_SUCCESS(rc))
6713 {
6714 rc = pCache->Backend->pfnSetModificationUuid(pCache->pBackendData,
6715 &UuidModification);
6716 }
6717
6718 if (rc == VERR_NOT_SUPPORTED)
6719 rc = VINF_SUCCESS;
6720 }
6721
6722 if (RT_SUCCESS(rc))
6723 {
6724 /* Cache successfully created. */
6725 pDisk->pCache = pCache;
6726 }
6727 else
6728 {
6729 /* Error detected, but image opened. Close and delete image. */
6730 rc2 = pCache->Backend->pfnClose(pCache->pBackendData, true);
6731 AssertRC(rc2);
6732 pCache->pBackendData = NULL;
6733 }
6734 } while (0);
6735
6736 if (RT_UNLIKELY(fLockWrite))
6737 {
6738 rc2 = vdThreadFinishWrite(pDisk);
6739 AssertRC(rc2);
6740 }
6741 else if (RT_UNLIKELY(fLockRead))
6742 {
6743 rc2 = vdThreadFinishRead(pDisk);
6744 AssertRC(rc2);
6745 }
6746
6747 if (RT_FAILURE(rc))
6748 {
6749 if (pCache)
6750 {
6751 if (pCache->pszFilename)
6752 RTStrFree(pCache->pszFilename);
6753 RTMemFree(pCache);
6754 }
6755 }
6756
6757 if (RT_SUCCESS(rc) && pIfProgress && pIfProgress->pfnProgress)
6758 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
6759
6760 LogFlowFunc(("returns %Rrc\n", rc));
6761 return rc;
6762}
6763
6764/**
6765 * Merges two images (not necessarily with direct parent/child relationship).
6766 * As a side effect the source image and potentially the other images which
6767 * are also merged to the destination are deleted from both the disk and the
6768 * images in the HDD container.
6769 *
6770 * @returns VBox status code.
6771 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
6772 * @param pDisk Pointer to HDD container.
6773 * @param nImageFrom Name of the image file to merge from.
6774 * @param nImageTo Name of the image file to merge to.
6775 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
6776 */
6777VBOXDDU_DECL(int) VDMerge(PVBOXHDD pDisk, unsigned nImageFrom,
6778 unsigned nImageTo, PVDINTERFACE pVDIfsOperation)
6779{
6780 int rc = VINF_SUCCESS;
6781 int rc2;
6782 bool fLockWrite = false, fLockRead = false;
6783 void *pvBuf = NULL;
6784
6785 LogFlowFunc(("pDisk=%#p nImageFrom=%u nImageTo=%u pVDIfsOperation=%#p\n",
6786 pDisk, nImageFrom, nImageTo, pVDIfsOperation));
6787
6788 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
6789
6790 do
6791 {
6792 /* sanity check */
6793 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
6794 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
6795
6796 /* For simplicity reasons lock for writing as the image reopen below
6797 * might need it. After all the reopen is usually needed. */
6798 rc2 = vdThreadStartWrite(pDisk);
6799 AssertRC(rc2);
6800 fLockWrite = true;
6801 PVDIMAGE pImageFrom = vdGetImageByNumber(pDisk, nImageFrom);
6802 PVDIMAGE pImageTo = vdGetImageByNumber(pDisk, nImageTo);
6803 if (!pImageFrom || !pImageTo)
6804 {
6805 rc = VERR_VD_IMAGE_NOT_FOUND;
6806 break;
6807 }
6808 AssertBreakStmt(pImageFrom != pImageTo, rc = VERR_INVALID_PARAMETER);
6809
6810 /* Make sure destination image is writable. */
6811 unsigned uOpenFlags = pImageTo->Backend->pfnGetOpenFlags(pImageTo->pBackendData);
6812 if (uOpenFlags & VD_OPEN_FLAGS_READONLY)
6813 {
6814 /*
6815 * Clear skip consistency checks because the image is made writable now and
6816 * skipping consistency checks is only possible for readonly images.
6817 */
6818 uOpenFlags &= ~(VD_OPEN_FLAGS_READONLY | VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS);
6819 rc = pImageTo->Backend->pfnSetOpenFlags(pImageTo->pBackendData,
6820 uOpenFlags);
6821 if (RT_FAILURE(rc))
6822 break;
6823 }
6824
6825 /* Get size of destination image. */
6826 uint64_t cbSize = pImageTo->Backend->pfnGetSize(pImageTo->pBackendData);
6827 rc2 = vdThreadFinishWrite(pDisk);
6828 AssertRC(rc2);
6829 fLockWrite = false;
6830
6831 /* Allocate tmp buffer. */
6832 pvBuf = RTMemTmpAlloc(VD_MERGE_BUFFER_SIZE);
6833 if (!pvBuf)
6834 {
6835 rc = VERR_NO_MEMORY;
6836 break;
6837 }
6838
6839 /* Merging is done directly on the images itself. This potentially
6840 * causes trouble if the disk is full in the middle of operation. */
6841 if (nImageFrom < nImageTo)
6842 {
6843 /* Merge parent state into child. This means writing all not
6844 * allocated blocks in the destination image which are allocated in
6845 * the images to be merged. */
6846 uint64_t uOffset = 0;
6847 uint64_t cbRemaining = cbSize;
6848 do
6849 {
6850 size_t cbThisRead = RT_MIN(VD_MERGE_BUFFER_SIZE, cbRemaining);
6851 RTSGSEG SegmentBuf;
6852 RTSGBUF SgBuf;
6853 VDIOCTX IoCtx;
6854
6855 SegmentBuf.pvSeg = pvBuf;
6856 SegmentBuf.cbSeg = VD_MERGE_BUFFER_SIZE;
6857 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
6858 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_READ, 0, 0, NULL,
6859 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
6860
6861 /* Need to hold the write lock during a read-write operation. */
6862 rc2 = vdThreadStartWrite(pDisk);
6863 AssertRC(rc2);
6864 fLockWrite = true;
6865
6866 rc = pImageTo->Backend->pfnRead(pImageTo->pBackendData,
6867 uOffset, cbThisRead,
6868 &IoCtx, &cbThisRead);
6869 if (rc == VERR_VD_BLOCK_FREE)
6870 {
6871 /* Search for image with allocated block. Do not attempt to
6872 * read more than the previous reads marked as valid.
6873 * Otherwise this would return stale data when different
6874 * block sizes are used for the images. */
6875 for (PVDIMAGE pCurrImage = pImageTo->pPrev;
6876 pCurrImage != NULL && pCurrImage != pImageFrom->pPrev && rc == VERR_VD_BLOCK_FREE;
6877 pCurrImage = pCurrImage->pPrev)
6878 {
6879 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
6880 uOffset, cbThisRead,
6881 &IoCtx, &cbThisRead);
6882 }
6883
6884 if (rc != VERR_VD_BLOCK_FREE)
6885 {
6886 if (RT_FAILURE(rc))
6887 break;
6888 /* Updating the cache is required because this might be a live merge. */
6889 rc = vdWriteHelperEx(pDisk, pImageTo, pImageFrom->pPrev,
6890 uOffset, pvBuf, cbThisRead,
6891 true /* fUpdateCache */, 0);
6892 if (RT_FAILURE(rc))
6893 break;
6894 }
6895 else
6896 rc = VINF_SUCCESS;
6897 }
6898 else if (RT_FAILURE(rc))
6899 break;
6900
6901 rc2 = vdThreadFinishWrite(pDisk);
6902 AssertRC(rc2);
6903 fLockWrite = false;
6904
6905 uOffset += cbThisRead;
6906 cbRemaining -= cbThisRead;
6907
6908 if (pIfProgress && pIfProgress->pfnProgress)
6909 {
6910 /** @todo r=klaus: this can update the progress to the same
6911 * percentage over and over again if the image format makes
6912 * relatively small increments. */
6913 rc = pIfProgress->pfnProgress(pIfProgress->Core.pvUser,
6914 uOffset * 99 / cbSize);
6915 if (RT_FAILURE(rc))
6916 break;
6917 }
6918 } while (uOffset < cbSize);
6919 }
6920 else
6921 {
6922 /*
6923 * We may need to update the parent uuid of the child coming after
6924 * the last image to be merged. We have to reopen it read/write.
6925 *
6926 * This is done before we do the actual merge to prevent an
6927 * inconsistent chain if the mode change fails for some reason.
6928 */
6929 if (pImageFrom->pNext)
6930 {
6931 PVDIMAGE pImageChild = pImageFrom->pNext;
6932
6933 /* Take the write lock. */
6934 rc2 = vdThreadStartWrite(pDisk);
6935 AssertRC(rc2);
6936 fLockWrite = true;
6937
6938 /* We need to open the image in read/write mode. */
6939 uOpenFlags = pImageChild->Backend->pfnGetOpenFlags(pImageChild->pBackendData);
6940
6941 if (uOpenFlags & VD_OPEN_FLAGS_READONLY)
6942 {
6943 uOpenFlags &= ~VD_OPEN_FLAGS_READONLY;
6944 rc = pImageChild->Backend->pfnSetOpenFlags(pImageChild->pBackendData,
6945 uOpenFlags);
6946 if (RT_FAILURE(rc))
6947 break;
6948 }
6949
6950 rc2 = vdThreadFinishWrite(pDisk);
6951 AssertRC(rc2);
6952 fLockWrite = false;
6953 }
6954
6955 /* If the merge is from the last image we have to relay all writes
6956 * to the merge destination as well, so that concurrent writes
6957 * (in case of a live merge) are handled correctly. */
6958 if (!pImageFrom->pNext)
6959 {
6960 /* Take the write lock. */
6961 rc2 = vdThreadStartWrite(pDisk);
6962 AssertRC(rc2);
6963 fLockWrite = true;
6964
6965 pDisk->pImageRelay = pImageTo;
6966
6967 rc2 = vdThreadFinishWrite(pDisk);
6968 AssertRC(rc2);
6969 fLockWrite = false;
6970 }
6971
6972 /* Merge child state into parent. This means writing all blocks
6973 * which are allocated in the image up to the source image to the
6974 * destination image. */
6975 uint64_t uOffset = 0;
6976 uint64_t cbRemaining = cbSize;
6977 do
6978 {
6979 size_t cbThisRead = RT_MIN(VD_MERGE_BUFFER_SIZE, cbRemaining);
6980 RTSGSEG SegmentBuf;
6981 RTSGBUF SgBuf;
6982 VDIOCTX IoCtx;
6983
6984 rc = VERR_VD_BLOCK_FREE;
6985
6986 SegmentBuf.pvSeg = pvBuf;
6987 SegmentBuf.cbSeg = VD_MERGE_BUFFER_SIZE;
6988 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
6989 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_READ, 0, 0, NULL,
6990 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
6991
6992 /* Need to hold the write lock during a read-write operation. */
6993 rc2 = vdThreadStartWrite(pDisk);
6994 AssertRC(rc2);
6995 fLockWrite = true;
6996
6997 /* Search for image with allocated block. Do not attempt to
6998 * read more than the previous reads marked as valid. Otherwise
6999 * this would return stale data when different block sizes are
7000 * used for the images. */
7001 for (PVDIMAGE pCurrImage = pImageFrom;
7002 pCurrImage != NULL && pCurrImage != pImageTo && rc == VERR_VD_BLOCK_FREE;
7003 pCurrImage = pCurrImage->pPrev)
7004 {
7005 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
7006 uOffset, cbThisRead,
7007 &IoCtx, &cbThisRead);
7008 }
7009
7010 if (rc != VERR_VD_BLOCK_FREE)
7011 {
7012 if (RT_FAILURE(rc))
7013 break;
7014 rc = vdWriteHelper(pDisk, pImageTo, uOffset, pvBuf,
7015 cbThisRead, true /* fUpdateCache */);
7016 if (RT_FAILURE(rc))
7017 break;
7018 }
7019 else
7020 rc = VINF_SUCCESS;
7021
7022 rc2 = vdThreadFinishWrite(pDisk);
7023 AssertRC(rc2);
7024 fLockWrite = false;
7025
7026 uOffset += cbThisRead;
7027 cbRemaining -= cbThisRead;
7028
7029 if (pIfProgress && pIfProgress->pfnProgress)
7030 {
7031 /** @todo r=klaus: this can update the progress to the same
7032 * percentage over and over again if the image format makes
7033 * relatively small increments. */
7034 rc = pIfProgress->pfnProgress(pIfProgress->Core.pvUser,
7035 uOffset * 99 / cbSize);
7036 if (RT_FAILURE(rc))
7037 break;
7038 }
7039 } while (uOffset < cbSize);
7040
7041 /* In case we set up a "write proxy" image above we must clear
7042 * this again now to prevent stray writes. Failure or not. */
7043 if (!pImageFrom->pNext)
7044 {
7045 /* Take the write lock. */
7046 rc2 = vdThreadStartWrite(pDisk);
7047 AssertRC(rc2);
7048 fLockWrite = true;
7049
7050 pDisk->pImageRelay = NULL;
7051
7052 rc2 = vdThreadFinishWrite(pDisk);
7053 AssertRC(rc2);
7054 fLockWrite = false;
7055 }
7056 }
7057
7058 /*
7059 * Leave in case of an error to avoid corrupted data in the image chain
7060 * (includes cancelling the operation by the user).
7061 */
7062 if (RT_FAILURE(rc))
7063 break;
7064
7065 /* Need to hold the write lock while finishing the merge. */
7066 rc2 = vdThreadStartWrite(pDisk);
7067 AssertRC(rc2);
7068 fLockWrite = true;
7069
7070 /* Update parent UUID so that image chain is consistent.
7071 * The two attempts work around the problem that some backends
7072 * (e.g. iSCSI) do not support UUIDs, so we exploit the fact that
7073 * so far there can only be one such image in the chain. */
7074 /** @todo needs a better long-term solution, passing the UUID
7075 * knowledge from the caller or some such */
7076 RTUUID Uuid;
7077 PVDIMAGE pImageChild = NULL;
7078 if (nImageFrom < nImageTo)
7079 {
7080 if (pImageFrom->pPrev)
7081 {
7082 /* plan A: ask the parent itself for its UUID */
7083 rc = pImageFrom->pPrev->Backend->pfnGetUuid(pImageFrom->pPrev->pBackendData,
7084 &Uuid);
7085 if (RT_FAILURE(rc))
7086 {
7087 /* plan B: ask the child of the parent for parent UUID */
7088 rc = pImageFrom->Backend->pfnGetParentUuid(pImageFrom->pBackendData,
7089 &Uuid);
7090 }
7091 AssertRC(rc);
7092 }
7093 else
7094 RTUuidClear(&Uuid);
7095 rc = pImageTo->Backend->pfnSetParentUuid(pImageTo->pBackendData,
7096 &Uuid);
7097 AssertRC(rc);
7098 }
7099 else
7100 {
7101 /* Update the parent uuid of the child of the last merged image. */
7102 if (pImageFrom->pNext)
7103 {
7104 /* plan A: ask the parent itself for its UUID */
7105 rc = pImageTo->Backend->pfnGetUuid(pImageTo->pBackendData,
7106 &Uuid);
7107 if (RT_FAILURE(rc))
7108 {
7109 /* plan B: ask the child of the parent for parent UUID */
7110 rc = pImageTo->pNext->Backend->pfnGetParentUuid(pImageTo->pNext->pBackendData,
7111 &Uuid);
7112 }
7113 AssertRC(rc);
7114
7115 rc = pImageFrom->Backend->pfnSetParentUuid(pImageFrom->pNext->pBackendData,
7116 &Uuid);
7117 AssertRC(rc);
7118
7119 pImageChild = pImageFrom->pNext;
7120 }
7121 }
7122
7123 /* Delete the no longer needed images. */
7124 PVDIMAGE pImg = pImageFrom, pTmp;
7125 while (pImg != pImageTo)
7126 {
7127 if (nImageFrom < nImageTo)
7128 pTmp = pImg->pNext;
7129 else
7130 pTmp = pImg->pPrev;
7131 vdRemoveImageFromList(pDisk, pImg);
7132 pImg->Backend->pfnClose(pImg->pBackendData, true);
7133 RTMemFree(pImg->pszFilename);
7134 RTMemFree(pImg);
7135 pImg = pTmp;
7136 }
7137
7138 /* Make sure destination image is back to read only if necessary. */
7139 if (pImageTo != pDisk->pLast)
7140 {
7141 uOpenFlags = pImageTo->Backend->pfnGetOpenFlags(pImageTo->pBackendData);
7142 uOpenFlags |= VD_OPEN_FLAGS_READONLY;
7143 rc = pImageTo->Backend->pfnSetOpenFlags(pImageTo->pBackendData,
7144 uOpenFlags);
7145 if (RT_FAILURE(rc))
7146 break;
7147 }
7148
7149 /*
7150 * Make sure the child is readonly
7151 * for the child -> parent merge direction
7152 * if necessary.
7153 */
7154 if ( nImageFrom > nImageTo
7155 && pImageChild
7156 && pImageChild != pDisk->pLast)
7157 {
7158 uOpenFlags = pImageChild->Backend->pfnGetOpenFlags(pImageChild->pBackendData);
7159 uOpenFlags |= VD_OPEN_FLAGS_READONLY;
7160 rc = pImageChild->Backend->pfnSetOpenFlags(pImageChild->pBackendData,
7161 uOpenFlags);
7162 if (RT_FAILURE(rc))
7163 break;
7164 }
7165 } while (0);
7166
7167 if (RT_UNLIKELY(fLockWrite))
7168 {
7169 rc2 = vdThreadFinishWrite(pDisk);
7170 AssertRC(rc2);
7171 }
7172 else if (RT_UNLIKELY(fLockRead))
7173 {
7174 rc2 = vdThreadFinishRead(pDisk);
7175 AssertRC(rc2);
7176 }
7177
7178 if (pvBuf)
7179 RTMemTmpFree(pvBuf);
7180
7181 if (RT_SUCCESS(rc) && pIfProgress && pIfProgress->pfnProgress)
7182 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
7183
7184 LogFlowFunc(("returns %Rrc\n", rc));
7185 return rc;
7186}
7187
7188/**
7189 * Copies an image from one HDD container to another - extended version.
7190 * The copy is opened in the target HDD container.
7191 * It is possible to convert between different image formats, because the
7192 * backend for the destination may be different from the source.
7193 * If both the source and destination reference the same HDD container,
7194 * then the image is moved (by copying/deleting or renaming) to the new location.
7195 * The source container is unchanged if the move operation fails, otherwise
7196 * the image at the new location is opened in the same way as the old one was.
7197 *
7198 * @note The read/write accesses across disks are not synchronized, just the
7199 * accesses to each disk. Once there is a use case which requires a defined
7200 * read/write behavior in this situation this needs to be extended.
7201 *
7202 * @return VBox status code.
7203 * @return VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
7204 * @param pDiskFrom Pointer to source HDD container.
7205 * @param nImage Image number, counts from 0. 0 is always base image of container.
7206 * @param pDiskTo Pointer to destination HDD container.
7207 * @param pszBackend Name of the image file backend to use (may be NULL to use the same as the source, case insensitive).
7208 * @param pszFilename New name of the image (may be NULL to specify that the
7209 * copy destination is the destination container, or
7210 * if pDiskFrom == pDiskTo, i.e. when moving).
7211 * @param fMoveByRename If true, attempt to perform a move by renaming (if successful the new size is ignored).
7212 * @param cbSize New image size (0 means leave unchanged).
7213 * @param nImageSameFrom todo
7214 * @param nImageSameTo todo
7215 * @param uImageFlags Flags specifying special destination image features.
7216 * @param pDstUuid New UUID of the destination image. If NULL, a new UUID is created.
7217 * This parameter is used if and only if a true copy is created.
7218 * In all rename/move cases or copy to existing image cases the modification UUIDs are copied over.
7219 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
7220 * Only used if the destination image is created.
7221 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
7222 * @param pDstVDIfsImage Pointer to the per-image VD interface list, for the
7223 * destination image.
7224 * @param pDstVDIfsOperation Pointer to the per-operation VD interface list,
7225 * for the destination operation.
7226 */
7227VBOXDDU_DECL(int) VDCopyEx(PVBOXHDD pDiskFrom, unsigned nImage, PVBOXHDD pDiskTo,
7228 const char *pszBackend, const char *pszFilename,
7229 bool fMoveByRename, uint64_t cbSize,
7230 unsigned nImageFromSame, unsigned nImageToSame,
7231 unsigned uImageFlags, PCRTUUID pDstUuid,
7232 unsigned uOpenFlags, PVDINTERFACE pVDIfsOperation,
7233 PVDINTERFACE pDstVDIfsImage,
7234 PVDINTERFACE pDstVDIfsOperation)
7235{
7236 int rc = VINF_SUCCESS;
7237 int rc2;
7238 bool fLockReadFrom = false, fLockWriteFrom = false, fLockWriteTo = false;
7239 PVDIMAGE pImageTo = NULL;
7240
7241 LogFlowFunc(("pDiskFrom=%#p nImage=%u pDiskTo=%#p pszBackend=\"%s\" pszFilename=\"%s\" fMoveByRename=%d cbSize=%llu nImageFromSame=%u nImageToSame=%u uImageFlags=%#x pDstUuid=%#p uOpenFlags=%#x pVDIfsOperation=%#p pDstVDIfsImage=%#p pDstVDIfsOperation=%#p\n",
7242 pDiskFrom, nImage, pDiskTo, pszBackend, pszFilename, fMoveByRename, cbSize, nImageFromSame, nImageToSame, uImageFlags, pDstUuid, uOpenFlags, pVDIfsOperation, pDstVDIfsImage, pDstVDIfsOperation));
7243
7244 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
7245 PVDINTERFACEPROGRESS pDstIfProgress = VDIfProgressGet(pDstVDIfsOperation);
7246
7247 do {
7248 /* Check arguments. */
7249 AssertMsgBreakStmt(VALID_PTR(pDiskFrom), ("pDiskFrom=%#p\n", pDiskFrom),
7250 rc = VERR_INVALID_PARAMETER);
7251 AssertMsg(pDiskFrom->u32Signature == VBOXHDDDISK_SIGNATURE,
7252 ("u32Signature=%08x\n", pDiskFrom->u32Signature));
7253
7254 rc2 = vdThreadStartRead(pDiskFrom);
7255 AssertRC(rc2);
7256 fLockReadFrom = true;
7257 PVDIMAGE pImageFrom = vdGetImageByNumber(pDiskFrom, nImage);
7258 AssertPtrBreakStmt(pImageFrom, rc = VERR_VD_IMAGE_NOT_FOUND);
7259 AssertMsgBreakStmt(VALID_PTR(pDiskTo), ("pDiskTo=%#p\n", pDiskTo),
7260 rc = VERR_INVALID_PARAMETER);
7261 AssertMsg(pDiskTo->u32Signature == VBOXHDDDISK_SIGNATURE,
7262 ("u32Signature=%08x\n", pDiskTo->u32Signature));
7263 AssertMsgBreakStmt( (nImageFromSame < nImage || nImageFromSame == VD_IMAGE_CONTENT_UNKNOWN)
7264 && (nImageToSame < pDiskTo->cImages || nImageToSame == VD_IMAGE_CONTENT_UNKNOWN)
7265 && ( (nImageFromSame == VD_IMAGE_CONTENT_UNKNOWN && nImageToSame == VD_IMAGE_CONTENT_UNKNOWN)
7266 || (nImageFromSame != VD_IMAGE_CONTENT_UNKNOWN && nImageToSame != VD_IMAGE_CONTENT_UNKNOWN)),
7267 ("nImageFromSame=%u nImageToSame=%u\n", nImageFromSame, nImageToSame),
7268 rc = VERR_INVALID_PARAMETER);
7269
7270 /* Move the image. */
7271 if (pDiskFrom == pDiskTo)
7272 {
7273 /* Rename only works when backends are the same, are file based
7274 * and the rename method is implemented. */
7275 if ( fMoveByRename
7276 && !RTStrICmp(pszBackend, pImageFrom->Backend->pszBackendName)
7277 && pImageFrom->Backend->uBackendCaps & VD_CAP_FILE
7278 && pImageFrom->Backend->pfnRename)
7279 {
7280 rc2 = vdThreadFinishRead(pDiskFrom);
7281 AssertRC(rc2);
7282 fLockReadFrom = false;
7283
7284 rc2 = vdThreadStartWrite(pDiskFrom);
7285 AssertRC(rc2);
7286 fLockWriteFrom = true;
7287 rc = pImageFrom->Backend->pfnRename(pImageFrom->pBackendData, pszFilename ? pszFilename : pImageFrom->pszFilename);
7288 break;
7289 }
7290
7291 /** @todo Moving (including shrinking/growing) of the image is
7292 * requested, but the rename attempt failed or it wasn't possible.
7293 * Must now copy image to temp location. */
7294 AssertReleaseMsgFailed(("VDCopy: moving by copy/delete not implemented\n"));
7295 }
7296
7297 /* pszFilename is allowed to be NULL, as this indicates copy to the existing image. */
7298 AssertMsgBreakStmt(pszFilename == NULL || (VALID_PTR(pszFilename) && *pszFilename),
7299 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
7300 rc = VERR_INVALID_PARAMETER);
7301
7302 uint64_t cbSizeFrom;
7303 cbSizeFrom = pImageFrom->Backend->pfnGetSize(pImageFrom->pBackendData);
7304 if (cbSizeFrom == 0)
7305 {
7306 rc = VERR_VD_VALUE_NOT_FOUND;
7307 break;
7308 }
7309
7310 VDGEOMETRY PCHSGeometryFrom = {0, 0, 0};
7311 VDGEOMETRY LCHSGeometryFrom = {0, 0, 0};
7312 pImageFrom->Backend->pfnGetPCHSGeometry(pImageFrom->pBackendData, &PCHSGeometryFrom);
7313 pImageFrom->Backend->pfnGetLCHSGeometry(pImageFrom->pBackendData, &LCHSGeometryFrom);
7314
7315 RTUUID ImageUuid, ImageModificationUuid;
7316 if (pDiskFrom != pDiskTo)
7317 {
7318 if (pDstUuid)
7319 ImageUuid = *pDstUuid;
7320 else
7321 RTUuidCreate(&ImageUuid);
7322 }
7323 else
7324 {
7325 rc = pImageFrom->Backend->pfnGetUuid(pImageFrom->pBackendData, &ImageUuid);
7326 if (RT_FAILURE(rc))
7327 RTUuidCreate(&ImageUuid);
7328 }
7329 rc = pImageFrom->Backend->pfnGetModificationUuid(pImageFrom->pBackendData, &ImageModificationUuid);
7330 if (RT_FAILURE(rc))
7331 RTUuidClear(&ImageModificationUuid);
7332
7333 char szComment[1024];
7334 rc = pImageFrom->Backend->pfnGetComment(pImageFrom->pBackendData, szComment, sizeof(szComment));
7335 if (RT_FAILURE(rc))
7336 szComment[0] = '\0';
7337 else
7338 szComment[sizeof(szComment) - 1] = '\0';
7339
7340 rc2 = vdThreadFinishRead(pDiskFrom);
7341 AssertRC(rc2);
7342 fLockReadFrom = false;
7343
7344 rc2 = vdThreadStartRead(pDiskTo);
7345 AssertRC(rc2);
7346 unsigned cImagesTo = pDiskTo->cImages;
7347 rc2 = vdThreadFinishRead(pDiskTo);
7348 AssertRC(rc2);
7349
7350 if (pszFilename)
7351 {
7352 if (cbSize == 0)
7353 cbSize = cbSizeFrom;
7354
7355 /* Create destination image with the properties of source image. */
7356 /** @todo replace the VDCreateDiff/VDCreateBase calls by direct
7357 * calls to the backend. Unifies the code and reduces the API
7358 * dependencies. Would also make the synchronization explicit. */
7359 if (cImagesTo > 0)
7360 {
7361 rc = VDCreateDiff(pDiskTo, pszBackend, pszFilename,
7362 uImageFlags, szComment, &ImageUuid,
7363 NULL /* pParentUuid */,
7364 uOpenFlags & ~VD_OPEN_FLAGS_READONLY,
7365 pDstVDIfsImage, NULL);
7366
7367 rc2 = vdThreadStartWrite(pDiskTo);
7368 AssertRC(rc2);
7369 fLockWriteTo = true;
7370 } else {
7371 /** @todo hack to force creation of a fixed image for
7372 * the RAW backend, which can't handle anything else. */
7373 if (!RTStrICmp(pszBackend, "RAW"))
7374 uImageFlags |= VD_IMAGE_FLAGS_FIXED;
7375
7376 vdFixupPCHSGeometry(&PCHSGeometryFrom, cbSize);
7377 vdFixupLCHSGeometry(&LCHSGeometryFrom, cbSize);
7378
7379 rc = VDCreateBase(pDiskTo, pszBackend, pszFilename, cbSize,
7380 uImageFlags, szComment,
7381 &PCHSGeometryFrom, &LCHSGeometryFrom,
7382 NULL, uOpenFlags & ~VD_OPEN_FLAGS_READONLY,
7383 pDstVDIfsImage, NULL);
7384
7385 rc2 = vdThreadStartWrite(pDiskTo);
7386 AssertRC(rc2);
7387 fLockWriteTo = true;
7388
7389 if (RT_SUCCESS(rc) && !RTUuidIsNull(&ImageUuid))
7390 pDiskTo->pLast->Backend->pfnSetUuid(pDiskTo->pLast->pBackendData, &ImageUuid);
7391 }
7392 if (RT_FAILURE(rc))
7393 break;
7394
7395 pImageTo = pDiskTo->pLast;
7396 AssertPtrBreakStmt(pImageTo, rc = VERR_VD_IMAGE_NOT_FOUND);
7397
7398 cbSize = RT_MIN(cbSize, cbSizeFrom);
7399 }
7400 else
7401 {
7402 pImageTo = pDiskTo->pLast;
7403 AssertPtrBreakStmt(pImageTo, rc = VERR_VD_IMAGE_NOT_FOUND);
7404
7405 uint64_t cbSizeTo;
7406 cbSizeTo = pImageTo->Backend->pfnGetSize(pImageTo->pBackendData);
7407 if (cbSizeTo == 0)
7408 {
7409 rc = VERR_VD_VALUE_NOT_FOUND;
7410 break;
7411 }
7412
7413 if (cbSize == 0)
7414 cbSize = RT_MIN(cbSizeFrom, cbSizeTo);
7415
7416 vdFixupPCHSGeometry(&PCHSGeometryFrom, cbSize);
7417 vdFixupLCHSGeometry(&LCHSGeometryFrom, cbSize);
7418
7419 /* Update the geometry in the destination image. */
7420 pImageTo->Backend->pfnSetPCHSGeometry(pImageTo->pBackendData, &PCHSGeometryFrom);
7421 pImageTo->Backend->pfnSetLCHSGeometry(pImageTo->pBackendData, &LCHSGeometryFrom);
7422 }
7423
7424 rc2 = vdThreadFinishWrite(pDiskTo);
7425 AssertRC(rc2);
7426 fLockWriteTo = false;
7427
7428 /* Whether we can take the optimized copy path (false) or not.
7429 * Don't optimize if the image existed or if it is a child image. */
7430 bool fSuppressRedundantIo = ( !(pszFilename == NULL || cImagesTo > 0)
7431 || (nImageToSame != VD_IMAGE_CONTENT_UNKNOWN));
7432 unsigned cImagesFromReadBack, cImagesToReadBack;
7433
7434 if (nImageFromSame == VD_IMAGE_CONTENT_UNKNOWN)
7435 cImagesFromReadBack = 0;
7436 else
7437 {
7438 if (nImage == VD_LAST_IMAGE)
7439 cImagesFromReadBack = pDiskFrom->cImages - nImageFromSame - 1;
7440 else
7441 cImagesFromReadBack = nImage - nImageFromSame;
7442 }
7443
7444 if (nImageToSame == VD_IMAGE_CONTENT_UNKNOWN)
7445 cImagesToReadBack = 0;
7446 else
7447 cImagesToReadBack = pDiskTo->cImages - nImageToSame - 1;
7448
7449 /* Copy the data. */
7450 rc = vdCopyHelper(pDiskFrom, pImageFrom, pDiskTo, cbSize,
7451 cImagesFromReadBack, cImagesToReadBack,
7452 fSuppressRedundantIo, pIfProgress, pDstIfProgress);
7453
7454 if (RT_SUCCESS(rc))
7455 {
7456 rc2 = vdThreadStartWrite(pDiskTo);
7457 AssertRC(rc2);
7458 fLockWriteTo = true;
7459
7460 /* Only set modification UUID if it is non-null, since the source
7461 * backend might not provide a valid modification UUID. */
7462 if (!RTUuidIsNull(&ImageModificationUuid))
7463 pImageTo->Backend->pfnSetModificationUuid(pImageTo->pBackendData, &ImageModificationUuid);
7464
7465 /* Set the requested open flags if they differ from the value
7466 * required for creating the image and copying the contents. */
7467 if ( pImageTo && pszFilename
7468 && uOpenFlags != (uOpenFlags & ~VD_OPEN_FLAGS_READONLY))
7469 rc = pImageTo->Backend->pfnSetOpenFlags(pImageTo->pBackendData,
7470 uOpenFlags);
7471 }
7472 } while (0);
7473
7474 if (RT_FAILURE(rc) && pImageTo && pszFilename)
7475 {
7476 /* Take the write lock only if it is not taken. Not worth making the
7477 * above code even more complicated. */
7478 if (RT_UNLIKELY(!fLockWriteTo))
7479 {
7480 rc2 = vdThreadStartWrite(pDiskTo);
7481 AssertRC(rc2);
7482 fLockWriteTo = true;
7483 }
7484 /* Error detected, but new image created. Remove image from list. */
7485 vdRemoveImageFromList(pDiskTo, pImageTo);
7486
7487 /* Close and delete image. */
7488 rc2 = pImageTo->Backend->pfnClose(pImageTo->pBackendData, true);
7489 AssertRC(rc2);
7490 pImageTo->pBackendData = NULL;
7491
7492 /* Free remaining resources. */
7493 if (pImageTo->pszFilename)
7494 RTStrFree(pImageTo->pszFilename);
7495
7496 RTMemFree(pImageTo);
7497 }
7498
7499 if (RT_UNLIKELY(fLockWriteTo))
7500 {
7501 rc2 = vdThreadFinishWrite(pDiskTo);
7502 AssertRC(rc2);
7503 }
7504 if (RT_UNLIKELY(fLockWriteFrom))
7505 {
7506 rc2 = vdThreadFinishWrite(pDiskFrom);
7507 AssertRC(rc2);
7508 }
7509 else if (RT_UNLIKELY(fLockReadFrom))
7510 {
7511 rc2 = vdThreadFinishRead(pDiskFrom);
7512 AssertRC(rc2);
7513 }
7514
7515 if (RT_SUCCESS(rc))
7516 {
7517 if (pIfProgress && pIfProgress->pfnProgress)
7518 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
7519 if (pDstIfProgress && pDstIfProgress->pfnProgress)
7520 pDstIfProgress->pfnProgress(pDstIfProgress->Core.pvUser, 100);
7521 }
7522
7523 LogFlowFunc(("returns %Rrc\n", rc));
7524 return rc;
7525}
7526
7527/**
7528 * Copies an image from one HDD container to another.
7529 * The copy is opened in the target HDD container.
7530 * It is possible to convert between different image formats, because the
7531 * backend for the destination may be different from the source.
7532 * If both the source and destination reference the same HDD container,
7533 * then the image is moved (by copying/deleting or renaming) to the new location.
7534 * The source container is unchanged if the move operation fails, otherwise
7535 * the image at the new location is opened in the same way as the old one was.
7536 *
7537 * @returns VBox status code.
7538 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
7539 * @param pDiskFrom Pointer to source HDD container.
7540 * @param nImage Image number, counts from 0. 0 is always base image of container.
7541 * @param pDiskTo Pointer to destination HDD container.
7542 * @param pszBackend Name of the image file backend to use.
7543 * @param pszFilename New name of the image (may be NULL if pDiskFrom == pDiskTo).
7544 * @param fMoveByRename If true, attempt to perform a move by renaming (if successful the new size is ignored).
7545 * @param cbSize New image size (0 means leave unchanged).
7546 * @param uImageFlags Flags specifying special destination image features.
7547 * @param pDstUuid New UUID of the destination image. If NULL, a new UUID is created.
7548 * This parameter is used if and only if a true copy is created.
7549 * In all rename/move cases the UUIDs are copied over.
7550 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
7551 * Only used if the destination image is created.
7552 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
7553 * @param pDstVDIfsImage Pointer to the per-image VD interface list, for the
7554 * destination image.
7555 * @param pDstVDIfsOperation Pointer to the per-image VD interface list,
7556 * for the destination image.
7557 */
7558VBOXDDU_DECL(int) VDCopy(PVBOXHDD pDiskFrom, unsigned nImage, PVBOXHDD pDiskTo,
7559 const char *pszBackend, const char *pszFilename,
7560 bool fMoveByRename, uint64_t cbSize,
7561 unsigned uImageFlags, PCRTUUID pDstUuid,
7562 unsigned uOpenFlags, PVDINTERFACE pVDIfsOperation,
7563 PVDINTERFACE pDstVDIfsImage,
7564 PVDINTERFACE pDstVDIfsOperation)
7565{
7566 return VDCopyEx(pDiskFrom, nImage, pDiskTo, pszBackend, pszFilename, fMoveByRename,
7567 cbSize, VD_IMAGE_CONTENT_UNKNOWN, VD_IMAGE_CONTENT_UNKNOWN,
7568 uImageFlags, pDstUuid, uOpenFlags, pVDIfsOperation,
7569 pDstVDIfsImage, pDstVDIfsOperation);
7570}
7571
7572/**
7573 * Optimizes the storage consumption of an image. Typically the unused blocks
7574 * have to be wiped with zeroes to achieve a substantial reduced storage use.
7575 * Another optimization done is reordering the image blocks, which can provide
7576 * a significant performance boost, as reads and writes tend to use less random
7577 * file offsets.
7578 *
7579 * @return VBox status code.
7580 * @return VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
7581 * @return VERR_VD_IMAGE_READ_ONLY if image is not writable.
7582 * @return VERR_NOT_SUPPORTED if this kind of image can be compacted, but
7583 * the code for this isn't implemented yet.
7584 * @param pDisk Pointer to HDD container.
7585 * @param nImage Image number, counts from 0. 0 is always base image of container.
7586 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
7587 */
7588VBOXDDU_DECL(int) VDCompact(PVBOXHDD pDisk, unsigned nImage,
7589 PVDINTERFACE pVDIfsOperation)
7590{
7591 int rc = VINF_SUCCESS;
7592 int rc2;
7593 bool fLockRead = false, fLockWrite = false;
7594 void *pvBuf = NULL;
7595 void *pvTmp = NULL;
7596
7597 LogFlowFunc(("pDisk=%#p nImage=%u pVDIfsOperation=%#p\n",
7598 pDisk, nImage, pVDIfsOperation));
7599
7600 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
7601
7602 do {
7603 /* Check arguments. */
7604 AssertMsgBreakStmt(VALID_PTR(pDisk), ("pDisk=%#p\n", pDisk),
7605 rc = VERR_INVALID_PARAMETER);
7606 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE,
7607 ("u32Signature=%08x\n", pDisk->u32Signature));
7608
7609 rc2 = vdThreadStartRead(pDisk);
7610 AssertRC(rc2);
7611 fLockRead = true;
7612
7613 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
7614 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
7615
7616 /* If there is no compact callback for not file based backends then
7617 * the backend doesn't need compaction. No need to make much fuss about
7618 * this. For file based ones signal this as not yet supported. */
7619 if (!pImage->Backend->pfnCompact)
7620 {
7621 if (pImage->Backend->uBackendCaps & VD_CAP_FILE)
7622 rc = VERR_NOT_SUPPORTED;
7623 else
7624 rc = VINF_SUCCESS;
7625 break;
7626 }
7627
7628 /* Insert interface for reading parent state into per-operation list,
7629 * if there is a parent image. */
7630 VDINTERFACEPARENTSTATE VDIfParent;
7631 VDPARENTSTATEDESC ParentUser;
7632 if (pImage->pPrev)
7633 {
7634 VDIfParent.pfnParentRead = vdParentRead;
7635 ParentUser.pDisk = pDisk;
7636 ParentUser.pImage = pImage->pPrev;
7637 rc = VDInterfaceAdd(&VDIfParent.Core, "VDCompact_ParentState", VDINTERFACETYPE_PARENTSTATE,
7638 &ParentUser, sizeof(VDINTERFACEPARENTSTATE), &pVDIfsOperation);
7639 AssertRC(rc);
7640 }
7641
7642 rc2 = vdThreadFinishRead(pDisk);
7643 AssertRC(rc2);
7644 fLockRead = false;
7645
7646 rc2 = vdThreadStartWrite(pDisk);
7647 AssertRC(rc2);
7648 fLockWrite = true;
7649
7650 rc = pImage->Backend->pfnCompact(pImage->pBackendData,
7651 0, 99,
7652 pDisk->pVDIfsDisk,
7653 pImage->pVDIfsImage,
7654 pVDIfsOperation);
7655 } while (0);
7656
7657 if (RT_UNLIKELY(fLockWrite))
7658 {
7659 rc2 = vdThreadFinishWrite(pDisk);
7660 AssertRC(rc2);
7661 }
7662 else if (RT_UNLIKELY(fLockRead))
7663 {
7664 rc2 = vdThreadFinishRead(pDisk);
7665 AssertRC(rc2);
7666 }
7667
7668 if (pvBuf)
7669 RTMemTmpFree(pvBuf);
7670 if (pvTmp)
7671 RTMemTmpFree(pvTmp);
7672
7673 if (RT_SUCCESS(rc))
7674 {
7675 if (pIfProgress && pIfProgress->pfnProgress)
7676 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
7677 }
7678
7679 LogFlowFunc(("returns %Rrc\n", rc));
7680 return rc;
7681}
7682
7683/**
7684 * Resizes the given disk image to the given size.
7685 *
7686 * @return VBox status
7687 * @return VERR_VD_IMAGE_READ_ONLY if image is not writable.
7688 * @return VERR_NOT_SUPPORTED if this kind of image can be compacted, but
7689 *
7690 * @param pDisk Pointer to the HDD container.
7691 * @param cbSize New size of the image.
7692 * @param pPCHSGeometry Pointer to the new physical disk geometry <= (16383,16,63). Not NULL.
7693 * @param pLCHSGeometry Pointer to the new logical disk geometry <= (x,255,63). Not NULL.
7694 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
7695 */
7696VBOXDDU_DECL(int) VDResize(PVBOXHDD pDisk, uint64_t cbSize,
7697 PCVDGEOMETRY pPCHSGeometry,
7698 PCVDGEOMETRY pLCHSGeometry,
7699 PVDINTERFACE pVDIfsOperation)
7700{
7701 /** @todo r=klaus resizing was designed to be part of VDCopy, so having a separate function is not desirable. */
7702 int rc = VINF_SUCCESS;
7703 int rc2;
7704 bool fLockRead = false, fLockWrite = false;
7705
7706 LogFlowFunc(("pDisk=%#p cbSize=%llu pVDIfsOperation=%#p\n",
7707 pDisk, cbSize, pVDIfsOperation));
7708
7709 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
7710
7711 do {
7712 /* Check arguments. */
7713 AssertMsgBreakStmt(VALID_PTR(pDisk), ("pDisk=%#p\n", pDisk),
7714 rc = VERR_INVALID_PARAMETER);
7715 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE,
7716 ("u32Signature=%08x\n", pDisk->u32Signature));
7717
7718 rc2 = vdThreadStartRead(pDisk);
7719 AssertRC(rc2);
7720 fLockRead = true;
7721
7722 /* Must have at least one image in the chain, will resize last. */
7723 AssertMsgBreakStmt(pDisk->cImages >= 1, ("cImages=%u\n", pDisk->cImages),
7724 rc = VERR_NOT_SUPPORTED);
7725
7726 PVDIMAGE pImage = pDisk->pLast;
7727
7728 /* If there is no compact callback for not file based backends then
7729 * the backend doesn't need compaction. No need to make much fuss about
7730 * this. For file based ones signal this as not yet supported. */
7731 if (!pImage->Backend->pfnResize)
7732 {
7733 if (pImage->Backend->uBackendCaps & VD_CAP_FILE)
7734 rc = VERR_NOT_SUPPORTED;
7735 else
7736 rc = VINF_SUCCESS;
7737 break;
7738 }
7739
7740 rc2 = vdThreadFinishRead(pDisk);
7741 AssertRC(rc2);
7742 fLockRead = false;
7743
7744 rc2 = vdThreadStartWrite(pDisk);
7745 AssertRC(rc2);
7746 fLockWrite = true;
7747
7748 VDGEOMETRY PCHSGeometryOld;
7749 VDGEOMETRY LCHSGeometryOld;
7750 PCVDGEOMETRY pPCHSGeometryNew;
7751 PCVDGEOMETRY pLCHSGeometryNew;
7752
7753 if (pPCHSGeometry->cCylinders == 0)
7754 {
7755 /* Auto-detect marker, calculate new value ourself. */
7756 rc = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData, &PCHSGeometryOld);
7757 if (RT_SUCCESS(rc) && (PCHSGeometryOld.cCylinders != 0))
7758 PCHSGeometryOld.cCylinders = RT_MIN(cbSize / 512 / PCHSGeometryOld.cHeads / PCHSGeometryOld.cSectors, 16383);
7759 else if (rc == VERR_VD_GEOMETRY_NOT_SET)
7760 rc = VINF_SUCCESS;
7761
7762 pPCHSGeometryNew = &PCHSGeometryOld;
7763 }
7764 else
7765 pPCHSGeometryNew = pPCHSGeometry;
7766
7767 if (pLCHSGeometry->cCylinders == 0)
7768 {
7769 /* Auto-detect marker, calculate new value ourself. */
7770 rc = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData, &LCHSGeometryOld);
7771 if (RT_SUCCESS(rc) && (LCHSGeometryOld.cCylinders != 0))
7772 LCHSGeometryOld.cCylinders = cbSize / 512 / LCHSGeometryOld.cHeads / LCHSGeometryOld.cSectors;
7773 else if (rc == VERR_VD_GEOMETRY_NOT_SET)
7774 rc = VINF_SUCCESS;
7775
7776 pLCHSGeometryNew = &LCHSGeometryOld;
7777 }
7778 else
7779 pLCHSGeometryNew = pLCHSGeometry;
7780
7781 if (RT_SUCCESS(rc))
7782 rc = pImage->Backend->pfnResize(pImage->pBackendData,
7783 cbSize,
7784 pPCHSGeometryNew,
7785 pLCHSGeometryNew,
7786 0, 99,
7787 pDisk->pVDIfsDisk,
7788 pImage->pVDIfsImage,
7789 pVDIfsOperation);
7790 } while (0);
7791
7792 if (RT_UNLIKELY(fLockWrite))
7793 {
7794 rc2 = vdThreadFinishWrite(pDisk);
7795 AssertRC(rc2);
7796 }
7797 else if (RT_UNLIKELY(fLockRead))
7798 {
7799 rc2 = vdThreadFinishRead(pDisk);
7800 AssertRC(rc2);
7801 }
7802
7803 if (RT_SUCCESS(rc))
7804 {
7805 if (pIfProgress && pIfProgress->pfnProgress)
7806 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
7807
7808 pDisk->cbSize = cbSize;
7809 }
7810
7811 LogFlowFunc(("returns %Rrc\n", rc));
7812 return rc;
7813}
7814
7815/**
7816 * Closes the last opened image file in HDD container.
7817 * If previous image file was opened in read-only mode (the normal case) and
7818 * the last opened image is in read-write mode then the previous image will be
7819 * reopened in read/write mode.
7820 *
7821 * @returns VBox status code.
7822 * @returns VERR_VD_NOT_OPENED if no image is opened in HDD container.
7823 * @param pDisk Pointer to HDD container.
7824 * @param fDelete If true, delete the image from the host disk.
7825 */
7826VBOXDDU_DECL(int) VDClose(PVBOXHDD pDisk, bool fDelete)
7827{
7828 int rc = VINF_SUCCESS;
7829 int rc2;
7830 bool fLockWrite = false;
7831
7832 LogFlowFunc(("pDisk=%#p fDelete=%d\n", pDisk, fDelete));
7833 do
7834 {
7835 /* sanity check */
7836 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
7837 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
7838
7839 /* Not worth splitting this up into a read lock phase and write
7840 * lock phase, as closing an image is a relatively fast operation
7841 * dominated by the part which needs the write lock. */
7842 rc2 = vdThreadStartWrite(pDisk);
7843 AssertRC(rc2);
7844 fLockWrite = true;
7845
7846 PVDIMAGE pImage = pDisk->pLast;
7847 if (!pImage)
7848 {
7849 rc = VERR_VD_NOT_OPENED;
7850 break;
7851 }
7852
7853 /* Destroy the current discard state first which might still have pending blocks. */
7854 rc = vdDiscardStateDestroy(pDisk);
7855 if (RT_FAILURE(rc))
7856 break;
7857
7858 unsigned uOpenFlags = pImage->Backend->pfnGetOpenFlags(pImage->pBackendData);
7859 /* Remove image from list of opened images. */
7860 vdRemoveImageFromList(pDisk, pImage);
7861 /* Close (and optionally delete) image. */
7862 rc = pImage->Backend->pfnClose(pImage->pBackendData, fDelete);
7863 /* Free remaining resources related to the image. */
7864 RTStrFree(pImage->pszFilename);
7865 RTMemFree(pImage);
7866
7867 pImage = pDisk->pLast;
7868 if (!pImage)
7869 break;
7870
7871 /* If disk was previously in read/write mode, make sure it will stay
7872 * like this (if possible) after closing this image. Set the open flags
7873 * accordingly. */
7874 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
7875 {
7876 uOpenFlags = pImage->Backend->pfnGetOpenFlags(pImage->pBackendData);
7877 uOpenFlags &= ~ VD_OPEN_FLAGS_READONLY;
7878 rc = pImage->Backend->pfnSetOpenFlags(pImage->pBackendData, uOpenFlags);
7879 }
7880
7881 /* Cache disk information. */
7882 pDisk->cbSize = pImage->Backend->pfnGetSize(pImage->pBackendData);
7883
7884 /* Cache PCHS geometry. */
7885 rc2 = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
7886 &pDisk->PCHSGeometry);
7887 if (RT_FAILURE(rc2))
7888 {
7889 pDisk->PCHSGeometry.cCylinders = 0;
7890 pDisk->PCHSGeometry.cHeads = 0;
7891 pDisk->PCHSGeometry.cSectors = 0;
7892 }
7893 else
7894 {
7895 /* Make sure the PCHS geometry is properly clipped. */
7896 pDisk->PCHSGeometry.cCylinders = RT_MIN(pDisk->PCHSGeometry.cCylinders, 16383);
7897 pDisk->PCHSGeometry.cHeads = RT_MIN(pDisk->PCHSGeometry.cHeads, 16);
7898 pDisk->PCHSGeometry.cSectors = RT_MIN(pDisk->PCHSGeometry.cSectors, 63);
7899 }
7900
7901 /* Cache LCHS geometry. */
7902 rc2 = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
7903 &pDisk->LCHSGeometry);
7904 if (RT_FAILURE(rc2))
7905 {
7906 pDisk->LCHSGeometry.cCylinders = 0;
7907 pDisk->LCHSGeometry.cHeads = 0;
7908 pDisk->LCHSGeometry.cSectors = 0;
7909 }
7910 else
7911 {
7912 /* Make sure the LCHS geometry is properly clipped. */
7913 pDisk->LCHSGeometry.cHeads = RT_MIN(pDisk->LCHSGeometry.cHeads, 255);
7914 pDisk->LCHSGeometry.cSectors = RT_MIN(pDisk->LCHSGeometry.cSectors, 63);
7915 }
7916 } while (0);
7917
7918 if (RT_UNLIKELY(fLockWrite))
7919 {
7920 rc2 = vdThreadFinishWrite(pDisk);
7921 AssertRC(rc2);
7922 }
7923
7924 LogFlowFunc(("returns %Rrc\n", rc));
7925 return rc;
7926}
7927
7928/**
7929 * Closes the currently opened cache image file in HDD container.
7930 *
7931 * @return VBox status code.
7932 * @return VERR_VD_NOT_OPENED if no cache is opened in HDD container.
7933 * @param pDisk Pointer to HDD container.
7934 * @param fDelete If true, delete the image from the host disk.
7935 */
7936VBOXDDU_DECL(int) VDCacheClose(PVBOXHDD pDisk, bool fDelete)
7937{
7938 int rc = VINF_SUCCESS;
7939 int rc2;
7940 bool fLockWrite = false;
7941 PVDCACHE pCache = NULL;
7942
7943 LogFlowFunc(("pDisk=%#p fDelete=%d\n", pDisk, fDelete));
7944
7945 do
7946 {
7947 /* sanity check */
7948 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
7949 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
7950
7951 rc2 = vdThreadStartWrite(pDisk);
7952 AssertRC(rc2);
7953 fLockWrite = true;
7954
7955 AssertPtrBreakStmt(pDisk->pCache, rc = VERR_VD_CACHE_NOT_FOUND);
7956
7957 pCache = pDisk->pCache;
7958 pDisk->pCache = NULL;
7959
7960 pCache->Backend->pfnClose(pCache->pBackendData, fDelete);
7961 if (pCache->pszFilename)
7962 RTStrFree(pCache->pszFilename);
7963 RTMemFree(pCache);
7964 } while (0);
7965
7966 if (RT_LIKELY(fLockWrite))
7967 {
7968 rc2 = vdThreadFinishWrite(pDisk);
7969 AssertRC(rc2);
7970 }
7971
7972 LogFlowFunc(("returns %Rrc\n", rc));
7973 return rc;
7974}
7975
7976/**
7977 * Closes all opened image files in HDD container.
7978 *
7979 * @returns VBox status code.
7980 * @param pDisk Pointer to HDD container.
7981 */
7982VBOXDDU_DECL(int) VDCloseAll(PVBOXHDD pDisk)
7983{
7984 int rc = VINF_SUCCESS;
7985 int rc2;
7986 bool fLockWrite = false;
7987
7988 LogFlowFunc(("pDisk=%#p\n", pDisk));
7989 do
7990 {
7991 /* sanity check */
7992 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
7993 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
7994
7995 /* Lock the entire operation. */
7996 rc2 = vdThreadStartWrite(pDisk);
7997 AssertRC(rc2);
7998 fLockWrite = true;
7999
8000 PVDCACHE pCache = pDisk->pCache;
8001 if (pCache)
8002 {
8003 rc2 = pCache->Backend->pfnClose(pCache->pBackendData, false);
8004 if (RT_FAILURE(rc2) && RT_SUCCESS(rc))
8005 rc = rc2;
8006
8007 if (pCache->pszFilename)
8008 RTStrFree(pCache->pszFilename);
8009 RTMemFree(pCache);
8010 }
8011
8012 PVDIMAGE pImage = pDisk->pLast;
8013 while (VALID_PTR(pImage))
8014 {
8015 PVDIMAGE pPrev = pImage->pPrev;
8016 /* Remove image from list of opened images. */
8017 vdRemoveImageFromList(pDisk, pImage);
8018 /* Close image. */
8019 rc2 = pImage->Backend->pfnClose(pImage->pBackendData, false);
8020 if (RT_FAILURE(rc2) && RT_SUCCESS(rc))
8021 rc = rc2;
8022 /* Free remaining resources related to the image. */
8023 RTStrFree(pImage->pszFilename);
8024 RTMemFree(pImage);
8025 pImage = pPrev;
8026 }
8027 Assert(!VALID_PTR(pDisk->pLast));
8028 } while (0);
8029
8030 if (RT_UNLIKELY(fLockWrite))
8031 {
8032 rc2 = vdThreadFinishWrite(pDisk);
8033 AssertRC(rc2);
8034 }
8035
8036 LogFlowFunc(("returns %Rrc\n", rc));
8037 return rc;
8038}
8039
8040/**
8041 * Read data from virtual HDD.
8042 *
8043 * @returns VBox status code.
8044 * @returns VERR_VD_NOT_OPENED if no image is opened in HDD container.
8045 * @param pDisk Pointer to HDD container.
8046 * @param uOffset Offset of first reading byte from start of disk.
8047 * @param pvBuf Pointer to buffer for reading data.
8048 * @param cbRead Number of bytes to read.
8049 */
8050VBOXDDU_DECL(int) VDRead(PVBOXHDD pDisk, uint64_t uOffset, void *pvBuf,
8051 size_t cbRead)
8052{
8053 int rc = VINF_SUCCESS;
8054 int rc2;
8055 bool fLockRead = false;
8056
8057 LogFlowFunc(("pDisk=%#p uOffset=%llu pvBuf=%p cbRead=%zu\n",
8058 pDisk, uOffset, pvBuf, cbRead));
8059 do
8060 {
8061 /* sanity check */
8062 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8063 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8064
8065 /* Check arguments. */
8066 AssertMsgBreakStmt(VALID_PTR(pvBuf),
8067 ("pvBuf=%#p\n", pvBuf),
8068 rc = VERR_INVALID_PARAMETER);
8069 AssertMsgBreakStmt(cbRead,
8070 ("cbRead=%zu\n", cbRead),
8071 rc = VERR_INVALID_PARAMETER);
8072
8073 rc2 = vdThreadStartRead(pDisk);
8074 AssertRC(rc2);
8075 fLockRead = true;
8076
8077 AssertMsgBreakStmt(uOffset + cbRead <= pDisk->cbSize,
8078 ("uOffset=%llu cbRead=%zu pDisk->cbSize=%llu\n",
8079 uOffset, cbRead, pDisk->cbSize),
8080 rc = VERR_INVALID_PARAMETER);
8081
8082 PVDIMAGE pImage = pDisk->pLast;
8083 AssertPtrBreakStmt(pImage, rc = VERR_VD_NOT_OPENED);
8084
8085 rc = vdReadHelper(pDisk, pImage, uOffset, pvBuf, cbRead,
8086 true /* fUpdateCache */);
8087 } while (0);
8088
8089 if (RT_UNLIKELY(fLockRead))
8090 {
8091 rc2 = vdThreadFinishRead(pDisk);
8092 AssertRC(rc2);
8093 }
8094
8095 LogFlowFunc(("returns %Rrc\n", rc));
8096 return rc;
8097}
8098
8099/**
8100 * Write data to virtual HDD.
8101 *
8102 * @returns VBox status code.
8103 * @returns VERR_VD_NOT_OPENED if no image is opened in HDD container.
8104 * @param pDisk Pointer to HDD container.
8105 * @param uOffset Offset of the first byte being
8106 * written from start of disk.
8107 * @param pvBuf Pointer to buffer for writing data.
8108 * @param cbWrite Number of bytes to write.
8109 */
8110VBOXDDU_DECL(int) VDWrite(PVBOXHDD pDisk, uint64_t uOffset, const void *pvBuf,
8111 size_t cbWrite)
8112{
8113 int rc = VINF_SUCCESS;
8114 int rc2;
8115 bool fLockWrite = false;
8116
8117 LogFlowFunc(("pDisk=%#p uOffset=%llu pvBuf=%p cbWrite=%zu\n",
8118 pDisk, uOffset, pvBuf, cbWrite));
8119 do
8120 {
8121 /* sanity check */
8122 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8123 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8124
8125 /* Check arguments. */
8126 AssertMsgBreakStmt(VALID_PTR(pvBuf),
8127 ("pvBuf=%#p\n", pvBuf),
8128 rc = VERR_INVALID_PARAMETER);
8129 AssertMsgBreakStmt(cbWrite,
8130 ("cbWrite=%zu\n", cbWrite),
8131 rc = VERR_INVALID_PARAMETER);
8132
8133 rc2 = vdThreadStartWrite(pDisk);
8134 AssertRC(rc2);
8135 fLockWrite = true;
8136
8137 AssertMsgBreakStmt(uOffset + cbWrite <= pDisk->cbSize,
8138 ("uOffset=%llu cbWrite=%zu pDisk->cbSize=%llu\n",
8139 uOffset, cbWrite, pDisk->cbSize),
8140 rc = VERR_INVALID_PARAMETER);
8141
8142 PVDIMAGE pImage = pDisk->pLast;
8143 AssertPtrBreakStmt(pImage, rc = VERR_VD_NOT_OPENED);
8144
8145 vdSetModifiedFlag(pDisk);
8146 rc = vdWriteHelper(pDisk, pImage, uOffset, pvBuf, cbWrite,
8147 true /* fUpdateCache */);
8148 if (RT_FAILURE(rc))
8149 break;
8150
8151 /* If there is a merge (in the direction towards a parent) running
8152 * concurrently then we have to also "relay" the write to this parent,
8153 * as the merge position might be already past the position where
8154 * this write is going. The "context" of the write can come from the
8155 * natural chain, since merging either already did or will take care
8156 * of the "other" content which is might be needed to fill the block
8157 * to a full allocation size. The cache doesn't need to be touched
8158 * as this write is covered by the previous one. */
8159 if (RT_UNLIKELY(pDisk->pImageRelay))
8160 rc = vdWriteHelper(pDisk, pDisk->pImageRelay, uOffset,
8161 pvBuf, cbWrite, false /* fUpdateCache */);
8162 } while (0);
8163
8164 if (RT_UNLIKELY(fLockWrite))
8165 {
8166 rc2 = vdThreadFinishWrite(pDisk);
8167 AssertRC(rc2);
8168 }
8169
8170 LogFlowFunc(("returns %Rrc\n", rc));
8171 return rc;
8172}
8173
8174/**
8175 * Make sure the on disk representation of a virtual HDD is up to date.
8176 *
8177 * @returns VBox status code.
8178 * @returns VERR_VD_NOT_OPENED if no image is opened in HDD container.
8179 * @param pDisk Pointer to HDD container.
8180 */
8181VBOXDDU_DECL(int) VDFlush(PVBOXHDD pDisk)
8182{
8183 int rc = VINF_SUCCESS;
8184 int rc2;
8185 bool fLockWrite = false;
8186
8187 LogFlowFunc(("pDisk=%#p\n", pDisk));
8188 do
8189 {
8190 /* sanity check */
8191 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8192 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8193
8194 rc2 = vdThreadStartWrite(pDisk);
8195 AssertRC(rc2);
8196 fLockWrite = true;
8197
8198 PVDIMAGE pImage = pDisk->pLast;
8199 AssertPtrBreakStmt(pImage, rc = VERR_VD_NOT_OPENED);
8200
8201 vdResetModifiedFlag(pDisk);
8202
8203 VDIOCTX IoCtx;
8204 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_FLUSH, 0, 0, NULL,
8205 NULL, NULL, NULL, VDIOCTX_FLAGS_SYNC);
8206 rc = pImage->Backend->pfnFlush(pImage->pBackendData, &IoCtx);
8207
8208 if ( RT_SUCCESS(rc)
8209 && pDisk->pCache)
8210 rc = pDisk->pCache->Backend->pfnFlush(pDisk->pCache->pBackendData, &IoCtx);
8211 } while (0);
8212
8213 if (RT_UNLIKELY(fLockWrite))
8214 {
8215 rc2 = vdThreadFinishWrite(pDisk);
8216 AssertRC(rc2);
8217 }
8218
8219 LogFlowFunc(("returns %Rrc\n", rc));
8220 return rc;
8221}
8222
8223/**
8224 * Get number of opened images in HDD container.
8225 *
8226 * @returns Number of opened images for HDD container. 0 if no images have been opened.
8227 * @param pDisk Pointer to HDD container.
8228 */
8229VBOXDDU_DECL(unsigned) VDGetCount(PVBOXHDD pDisk)
8230{
8231 unsigned cImages;
8232 int rc2;
8233 bool fLockRead = false;
8234
8235 LogFlowFunc(("pDisk=%#p\n", pDisk));
8236 do
8237 {
8238 /* sanity check */
8239 AssertPtrBreakStmt(pDisk, cImages = 0);
8240 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8241
8242 rc2 = vdThreadStartRead(pDisk);
8243 AssertRC(rc2);
8244 fLockRead = true;
8245
8246 cImages = pDisk->cImages;
8247 } while (0);
8248
8249 if (RT_UNLIKELY(fLockRead))
8250 {
8251 rc2 = vdThreadFinishRead(pDisk);
8252 AssertRC(rc2);
8253 }
8254
8255 LogFlowFunc(("returns %u\n", cImages));
8256 return cImages;
8257}
8258
8259/**
8260 * Get read/write mode of HDD container.
8261 *
8262 * @returns Virtual disk ReadOnly status.
8263 * @returns true if no image is opened in HDD container.
8264 * @param pDisk Pointer to HDD container.
8265 */
8266VBOXDDU_DECL(bool) VDIsReadOnly(PVBOXHDD pDisk)
8267{
8268 bool fReadOnly;
8269 int rc2;
8270 bool fLockRead = false;
8271
8272 LogFlowFunc(("pDisk=%#p\n", pDisk));
8273 do
8274 {
8275 /* sanity check */
8276 AssertPtrBreakStmt(pDisk, fReadOnly = false);
8277 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8278
8279 rc2 = vdThreadStartRead(pDisk);
8280 AssertRC(rc2);
8281 fLockRead = true;
8282
8283 PVDIMAGE pImage = pDisk->pLast;
8284 AssertPtrBreakStmt(pImage, fReadOnly = true);
8285
8286 unsigned uOpenFlags;
8287 uOpenFlags = pDisk->pLast->Backend->pfnGetOpenFlags(pDisk->pLast->pBackendData);
8288 fReadOnly = !!(uOpenFlags & VD_OPEN_FLAGS_READONLY);
8289 } while (0);
8290
8291 if (RT_UNLIKELY(fLockRead))
8292 {
8293 rc2 = vdThreadFinishRead(pDisk);
8294 AssertRC(rc2);
8295 }
8296
8297 LogFlowFunc(("returns %d\n", fReadOnly));
8298 return fReadOnly;
8299}
8300
8301/**
8302 * Get total capacity of an image in HDD container.
8303 *
8304 * @returns Virtual disk size in bytes.
8305 * @returns 0 if no image with specified number was not opened.
8306 * @param pDisk Pointer to HDD container.
8307 * @param nImage Image number, counts from 0. 0 is always base image of container.
8308 */
8309VBOXDDU_DECL(uint64_t) VDGetSize(PVBOXHDD pDisk, unsigned nImage)
8310{
8311 uint64_t cbSize;
8312 int rc2;
8313 bool fLockRead = false;
8314
8315 LogFlowFunc(("pDisk=%#p nImage=%u\n", pDisk, nImage));
8316 do
8317 {
8318 /* sanity check */
8319 AssertPtrBreakStmt(pDisk, cbSize = 0);
8320 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8321
8322 rc2 = vdThreadStartRead(pDisk);
8323 AssertRC(rc2);
8324 fLockRead = true;
8325
8326 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8327 AssertPtrBreakStmt(pImage, cbSize = 0);
8328 cbSize = pImage->Backend->pfnGetSize(pImage->pBackendData);
8329 } while (0);
8330
8331 if (RT_UNLIKELY(fLockRead))
8332 {
8333 rc2 = vdThreadFinishRead(pDisk);
8334 AssertRC(rc2);
8335 }
8336
8337 LogFlowFunc(("returns %llu\n", cbSize));
8338 return cbSize;
8339}
8340
8341/**
8342 * Get total file size of an image in HDD container.
8343 *
8344 * @returns Virtual disk size in bytes.
8345 * @returns 0 if no image is opened in HDD container.
8346 * @param pDisk Pointer to HDD container.
8347 * @param nImage Image number, counts from 0. 0 is always base image of container.
8348 */
8349VBOXDDU_DECL(uint64_t) VDGetFileSize(PVBOXHDD pDisk, unsigned nImage)
8350{
8351 uint64_t cbSize;
8352 int rc2;
8353 bool fLockRead = false;
8354
8355 LogFlowFunc(("pDisk=%#p nImage=%u\n", pDisk, nImage));
8356 do
8357 {
8358 /* sanity check */
8359 AssertPtrBreakStmt(pDisk, cbSize = 0);
8360 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8361
8362 rc2 = vdThreadStartRead(pDisk);
8363 AssertRC(rc2);
8364 fLockRead = true;
8365
8366 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8367 AssertPtrBreakStmt(pImage, cbSize = 0);
8368 cbSize = pImage->Backend->pfnGetFileSize(pImage->pBackendData);
8369 } while (0);
8370
8371 if (RT_UNLIKELY(fLockRead))
8372 {
8373 rc2 = vdThreadFinishRead(pDisk);
8374 AssertRC(rc2);
8375 }
8376
8377 LogFlowFunc(("returns %llu\n", cbSize));
8378 return cbSize;
8379}
8380
8381/**
8382 * Get virtual disk PCHS geometry stored in HDD container.
8383 *
8384 * @returns VBox status code.
8385 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8386 * @returns VERR_VD_GEOMETRY_NOT_SET if no geometry present in the HDD container.
8387 * @param pDisk Pointer to HDD container.
8388 * @param nImage Image number, counts from 0. 0 is always base image of container.
8389 * @param pPCHSGeometry Where to store PCHS geometry. Not NULL.
8390 */
8391VBOXDDU_DECL(int) VDGetPCHSGeometry(PVBOXHDD pDisk, unsigned nImage,
8392 PVDGEOMETRY pPCHSGeometry)
8393{
8394 int rc = VINF_SUCCESS;
8395 int rc2;
8396 bool fLockRead = false;
8397
8398 LogFlowFunc(("pDisk=%#p nImage=%u pPCHSGeometry=%#p\n",
8399 pDisk, nImage, pPCHSGeometry));
8400 do
8401 {
8402 /* sanity check */
8403 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8404 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8405
8406 /* Check arguments. */
8407 AssertMsgBreakStmt(VALID_PTR(pPCHSGeometry),
8408 ("pPCHSGeometry=%#p\n", pPCHSGeometry),
8409 rc = VERR_INVALID_PARAMETER);
8410
8411 rc2 = vdThreadStartRead(pDisk);
8412 AssertRC(rc2);
8413 fLockRead = true;
8414
8415 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8416 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8417
8418 if (pImage == pDisk->pLast)
8419 {
8420 /* Use cached information if possible. */
8421 if (pDisk->PCHSGeometry.cCylinders != 0)
8422 *pPCHSGeometry = pDisk->PCHSGeometry;
8423 else
8424 rc = VERR_VD_GEOMETRY_NOT_SET;
8425 }
8426 else
8427 rc = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
8428 pPCHSGeometry);
8429 } while (0);
8430
8431 if (RT_UNLIKELY(fLockRead))
8432 {
8433 rc2 = vdThreadFinishRead(pDisk);
8434 AssertRC(rc2);
8435 }
8436
8437 LogFlowFunc(("%Rrc (PCHS=%u/%u/%u)\n", rc,
8438 pDisk->PCHSGeometry.cCylinders, pDisk->PCHSGeometry.cHeads,
8439 pDisk->PCHSGeometry.cSectors));
8440 return rc;
8441}
8442
8443/**
8444 * Store virtual disk PCHS geometry in HDD container.
8445 *
8446 * Note that in case of unrecoverable error all images in HDD container will be closed.
8447 *
8448 * @returns VBox status code.
8449 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8450 * @returns VERR_VD_GEOMETRY_NOT_SET if no geometry present in the HDD container.
8451 * @param pDisk Pointer to HDD container.
8452 * @param nImage Image number, counts from 0. 0 is always base image of container.
8453 * @param pPCHSGeometry Where to load PCHS geometry from. Not NULL.
8454 */
8455VBOXDDU_DECL(int) VDSetPCHSGeometry(PVBOXHDD pDisk, unsigned nImage,
8456 PCVDGEOMETRY pPCHSGeometry)
8457{
8458 int rc = VINF_SUCCESS;
8459 int rc2;
8460 bool fLockWrite = false;
8461
8462 LogFlowFunc(("pDisk=%#p nImage=%u pPCHSGeometry=%#p PCHS=%u/%u/%u\n",
8463 pDisk, nImage, pPCHSGeometry, pPCHSGeometry->cCylinders,
8464 pPCHSGeometry->cHeads, pPCHSGeometry->cSectors));
8465 do
8466 {
8467 /* sanity check */
8468 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8469 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8470
8471 /* Check arguments. */
8472 AssertMsgBreakStmt( VALID_PTR(pPCHSGeometry)
8473 && pPCHSGeometry->cHeads <= 16
8474 && pPCHSGeometry->cSectors <= 63,
8475 ("pPCHSGeometry=%#p PCHS=%u/%u/%u\n", pPCHSGeometry,
8476 pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads,
8477 pPCHSGeometry->cSectors),
8478 rc = VERR_INVALID_PARAMETER);
8479
8480 rc2 = vdThreadStartWrite(pDisk);
8481 AssertRC(rc2);
8482 fLockWrite = true;
8483
8484 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8485 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8486
8487 if (pImage == pDisk->pLast)
8488 {
8489 if ( pPCHSGeometry->cCylinders != pDisk->PCHSGeometry.cCylinders
8490 || pPCHSGeometry->cHeads != pDisk->PCHSGeometry.cHeads
8491 || pPCHSGeometry->cSectors != pDisk->PCHSGeometry.cSectors)
8492 {
8493 /* Only update geometry if it is changed. Avoids similar checks
8494 * in every backend. Most of the time the new geometry is set
8495 * to the previous values, so no need to go through the hassle
8496 * of updating an image which could be opened in read-only mode
8497 * right now. */
8498 rc = pImage->Backend->pfnSetPCHSGeometry(pImage->pBackendData,
8499 pPCHSGeometry);
8500
8501 /* Cache new geometry values in any case. */
8502 rc2 = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
8503 &pDisk->PCHSGeometry);
8504 if (RT_FAILURE(rc2))
8505 {
8506 pDisk->PCHSGeometry.cCylinders = 0;
8507 pDisk->PCHSGeometry.cHeads = 0;
8508 pDisk->PCHSGeometry.cSectors = 0;
8509 }
8510 else
8511 {
8512 /* Make sure the CHS geometry is properly clipped. */
8513 pDisk->PCHSGeometry.cHeads = RT_MIN(pDisk->PCHSGeometry.cHeads, 255);
8514 pDisk->PCHSGeometry.cSectors = RT_MIN(pDisk->PCHSGeometry.cSectors, 63);
8515 }
8516 }
8517 }
8518 else
8519 {
8520 VDGEOMETRY PCHS;
8521 rc = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
8522 &PCHS);
8523 if ( RT_FAILURE(rc)
8524 || pPCHSGeometry->cCylinders != PCHS.cCylinders
8525 || pPCHSGeometry->cHeads != PCHS.cHeads
8526 || pPCHSGeometry->cSectors != PCHS.cSectors)
8527 {
8528 /* Only update geometry if it is changed. Avoids similar checks
8529 * in every backend. Most of the time the new geometry is set
8530 * to the previous values, so no need to go through the hassle
8531 * of updating an image which could be opened in read-only mode
8532 * right now. */
8533 rc = pImage->Backend->pfnSetPCHSGeometry(pImage->pBackendData,
8534 pPCHSGeometry);
8535 }
8536 }
8537 } while (0);
8538
8539 if (RT_UNLIKELY(fLockWrite))
8540 {
8541 rc2 = vdThreadFinishWrite(pDisk);
8542 AssertRC(rc2);
8543 }
8544
8545 LogFlowFunc(("returns %Rrc\n", rc));
8546 return rc;
8547}
8548
8549/**
8550 * Get virtual disk LCHS geometry stored in HDD container.
8551 *
8552 * @returns VBox status code.
8553 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8554 * @returns VERR_VD_GEOMETRY_NOT_SET if no geometry present in the HDD container.
8555 * @param pDisk Pointer to HDD container.
8556 * @param nImage Image number, counts from 0. 0 is always base image of container.
8557 * @param pLCHSGeometry Where to store LCHS geometry. Not NULL.
8558 */
8559VBOXDDU_DECL(int) VDGetLCHSGeometry(PVBOXHDD pDisk, unsigned nImage,
8560 PVDGEOMETRY pLCHSGeometry)
8561{
8562 int rc = VINF_SUCCESS;
8563 int rc2;
8564 bool fLockRead = false;
8565
8566 LogFlowFunc(("pDisk=%#p nImage=%u pLCHSGeometry=%#p\n",
8567 pDisk, nImage, pLCHSGeometry));
8568 do
8569 {
8570 /* sanity check */
8571 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8572 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8573
8574 /* Check arguments. */
8575 AssertMsgBreakStmt(VALID_PTR(pLCHSGeometry),
8576 ("pLCHSGeometry=%#p\n", pLCHSGeometry),
8577 rc = VERR_INVALID_PARAMETER);
8578
8579 rc2 = vdThreadStartRead(pDisk);
8580 AssertRC(rc2);
8581 fLockRead = true;
8582
8583 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8584 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8585
8586 if (pImage == pDisk->pLast)
8587 {
8588 /* Use cached information if possible. */
8589 if (pDisk->LCHSGeometry.cCylinders != 0)
8590 *pLCHSGeometry = pDisk->LCHSGeometry;
8591 else
8592 rc = VERR_VD_GEOMETRY_NOT_SET;
8593 }
8594 else
8595 rc = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
8596 pLCHSGeometry);
8597 } while (0);
8598
8599 if (RT_UNLIKELY(fLockRead))
8600 {
8601 rc2 = vdThreadFinishRead(pDisk);
8602 AssertRC(rc2);
8603 }
8604
8605 LogFlowFunc((": %Rrc (LCHS=%u/%u/%u)\n", rc,
8606 pDisk->LCHSGeometry.cCylinders, pDisk->LCHSGeometry.cHeads,
8607 pDisk->LCHSGeometry.cSectors));
8608 return rc;
8609}
8610
8611/**
8612 * Store virtual disk LCHS geometry in HDD container.
8613 *
8614 * Note that in case of unrecoverable error all images in HDD container will be closed.
8615 *
8616 * @returns VBox status code.
8617 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8618 * @returns VERR_VD_GEOMETRY_NOT_SET if no geometry present in the HDD container.
8619 * @param pDisk Pointer to HDD container.
8620 * @param nImage Image number, counts from 0. 0 is always base image of container.
8621 * @param pLCHSGeometry Where to load LCHS geometry from. Not NULL.
8622 */
8623VBOXDDU_DECL(int) VDSetLCHSGeometry(PVBOXHDD pDisk, unsigned nImage,
8624 PCVDGEOMETRY pLCHSGeometry)
8625{
8626 int rc = VINF_SUCCESS;
8627 int rc2;
8628 bool fLockWrite = false;
8629
8630 LogFlowFunc(("pDisk=%#p nImage=%u pLCHSGeometry=%#p LCHS=%u/%u/%u\n",
8631 pDisk, nImage, pLCHSGeometry, pLCHSGeometry->cCylinders,
8632 pLCHSGeometry->cHeads, pLCHSGeometry->cSectors));
8633 do
8634 {
8635 /* sanity check */
8636 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8637 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8638
8639 /* Check arguments. */
8640 AssertMsgBreakStmt( VALID_PTR(pLCHSGeometry)
8641 && pLCHSGeometry->cHeads <= 255
8642 && pLCHSGeometry->cSectors <= 63,
8643 ("pLCHSGeometry=%#p LCHS=%u/%u/%u\n", pLCHSGeometry,
8644 pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads,
8645 pLCHSGeometry->cSectors),
8646 rc = VERR_INVALID_PARAMETER);
8647
8648 rc2 = vdThreadStartWrite(pDisk);
8649 AssertRC(rc2);
8650 fLockWrite = true;
8651
8652 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8653 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8654
8655 if (pImage == pDisk->pLast)
8656 {
8657 if ( pLCHSGeometry->cCylinders != pDisk->LCHSGeometry.cCylinders
8658 || pLCHSGeometry->cHeads != pDisk->LCHSGeometry.cHeads
8659 || pLCHSGeometry->cSectors != pDisk->LCHSGeometry.cSectors)
8660 {
8661 /* Only update geometry if it is changed. Avoids similar checks
8662 * in every backend. Most of the time the new geometry is set
8663 * to the previous values, so no need to go through the hassle
8664 * of updating an image which could be opened in read-only mode
8665 * right now. */
8666 rc = pImage->Backend->pfnSetLCHSGeometry(pImage->pBackendData,
8667 pLCHSGeometry);
8668
8669 /* Cache new geometry values in any case. */
8670 rc2 = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
8671 &pDisk->LCHSGeometry);
8672 if (RT_FAILURE(rc2))
8673 {
8674 pDisk->LCHSGeometry.cCylinders = 0;
8675 pDisk->LCHSGeometry.cHeads = 0;
8676 pDisk->LCHSGeometry.cSectors = 0;
8677 }
8678 else
8679 {
8680 /* Make sure the CHS geometry is properly clipped. */
8681 pDisk->LCHSGeometry.cHeads = RT_MIN(pDisk->LCHSGeometry.cHeads, 255);
8682 pDisk->LCHSGeometry.cSectors = RT_MIN(pDisk->LCHSGeometry.cSectors, 63);
8683 }
8684 }
8685 }
8686 else
8687 {
8688 VDGEOMETRY LCHS;
8689 rc = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
8690 &LCHS);
8691 if ( RT_FAILURE(rc)
8692 || pLCHSGeometry->cCylinders != LCHS.cCylinders
8693 || pLCHSGeometry->cHeads != LCHS.cHeads
8694 || pLCHSGeometry->cSectors != LCHS.cSectors)
8695 {
8696 /* Only update geometry if it is changed. Avoids similar checks
8697 * in every backend. Most of the time the new geometry is set
8698 * to the previous values, so no need to go through the hassle
8699 * of updating an image which could be opened in read-only mode
8700 * right now. */
8701 rc = pImage->Backend->pfnSetLCHSGeometry(pImage->pBackendData,
8702 pLCHSGeometry);
8703 }
8704 }
8705 } while (0);
8706
8707 if (RT_UNLIKELY(fLockWrite))
8708 {
8709 rc2 = vdThreadFinishWrite(pDisk);
8710 AssertRC(rc2);
8711 }
8712
8713 LogFlowFunc(("returns %Rrc\n", rc));
8714 return rc;
8715}
8716
8717/**
8718 * Get version of image in HDD container.
8719 *
8720 * @returns VBox status code.
8721 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8722 * @param pDisk Pointer to HDD container.
8723 * @param nImage Image number, counts from 0. 0 is always base image of container.
8724 * @param puVersion Where to store the image version.
8725 */
8726VBOXDDU_DECL(int) VDGetVersion(PVBOXHDD pDisk, unsigned nImage,
8727 unsigned *puVersion)
8728{
8729 int rc = VINF_SUCCESS;
8730 int rc2;
8731 bool fLockRead = false;
8732
8733 LogFlowFunc(("pDisk=%#p nImage=%u puVersion=%#p\n",
8734 pDisk, nImage, puVersion));
8735 do
8736 {
8737 /* sanity check */
8738 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8739 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8740
8741 /* Check arguments. */
8742 AssertMsgBreakStmt(VALID_PTR(puVersion),
8743 ("puVersion=%#p\n", puVersion),
8744 rc = VERR_INVALID_PARAMETER);
8745
8746 rc2 = vdThreadStartRead(pDisk);
8747 AssertRC(rc2);
8748 fLockRead = true;
8749
8750 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8751 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8752
8753 *puVersion = pImage->Backend->pfnGetVersion(pImage->pBackendData);
8754 } while (0);
8755
8756 if (RT_UNLIKELY(fLockRead))
8757 {
8758 rc2 = vdThreadFinishRead(pDisk);
8759 AssertRC(rc2);
8760 }
8761
8762 LogFlowFunc(("returns %Rrc uVersion=%#x\n", rc, *puVersion));
8763 return rc;
8764}
8765
8766/**
8767 * List the capabilities of image backend in HDD container.
8768 *
8769 * @returns VBox status code.
8770 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8771 * @param pDisk Pointer to the HDD container.
8772 * @param nImage Image number, counts from 0. 0 is always base image of container.
8773 * @param pbackendInfo Where to store the backend information.
8774 */
8775VBOXDDU_DECL(int) VDBackendInfoSingle(PVBOXHDD pDisk, unsigned nImage,
8776 PVDBACKENDINFO pBackendInfo)
8777{
8778 int rc = VINF_SUCCESS;
8779 int rc2;
8780 bool fLockRead = false;
8781
8782 LogFlowFunc(("pDisk=%#p nImage=%u pBackendInfo=%#p\n",
8783 pDisk, nImage, pBackendInfo));
8784 do
8785 {
8786 /* sanity check */
8787 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8788 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8789
8790 /* Check arguments. */
8791 AssertMsgBreakStmt(VALID_PTR(pBackendInfo),
8792 ("pBackendInfo=%#p\n", pBackendInfo),
8793 rc = VERR_INVALID_PARAMETER);
8794
8795 rc2 = vdThreadStartRead(pDisk);
8796 AssertRC(rc2);
8797 fLockRead = true;
8798
8799 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8800 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8801
8802 pBackendInfo->pszBackend = pImage->Backend->pszBackendName;
8803 pBackendInfo->uBackendCaps = pImage->Backend->uBackendCaps;
8804 pBackendInfo->paFileExtensions = pImage->Backend->paFileExtensions;
8805 pBackendInfo->paConfigInfo = pImage->Backend->paConfigInfo;
8806 } while (0);
8807
8808 if (RT_UNLIKELY(fLockRead))
8809 {
8810 rc2 = vdThreadFinishRead(pDisk);
8811 AssertRC(rc2);
8812 }
8813
8814 LogFlowFunc(("returns %Rrc\n", rc));
8815 return rc;
8816}
8817
8818/**
8819 * Get flags of image in HDD container.
8820 *
8821 * @returns VBox status code.
8822 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8823 * @param pDisk Pointer to HDD container.
8824 * @param nImage Image number, counts from 0. 0 is always base image of container.
8825 * @param puImageFlags Where to store the image flags.
8826 */
8827VBOXDDU_DECL(int) VDGetImageFlags(PVBOXHDD pDisk, unsigned nImage,
8828 unsigned *puImageFlags)
8829{
8830 int rc = VINF_SUCCESS;
8831 int rc2;
8832 bool fLockRead = false;
8833
8834 LogFlowFunc(("pDisk=%#p nImage=%u puImageFlags=%#p\n",
8835 pDisk, nImage, puImageFlags));
8836 do
8837 {
8838 /* sanity check */
8839 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8840 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8841
8842 /* Check arguments. */
8843 AssertMsgBreakStmt(VALID_PTR(puImageFlags),
8844 ("puImageFlags=%#p\n", puImageFlags),
8845 rc = VERR_INVALID_PARAMETER);
8846
8847 rc2 = vdThreadStartRead(pDisk);
8848 AssertRC(rc2);
8849 fLockRead = true;
8850
8851 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8852 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8853
8854 *puImageFlags = pImage->uImageFlags;
8855 } while (0);
8856
8857 if (RT_UNLIKELY(fLockRead))
8858 {
8859 rc2 = vdThreadFinishRead(pDisk);
8860 AssertRC(rc2);
8861 }
8862
8863 LogFlowFunc(("returns %Rrc uImageFlags=%#x\n", rc, *puImageFlags));
8864 return rc;
8865}
8866
8867/**
8868 * Get open flags of image in HDD container.
8869 *
8870 * @returns VBox status code.
8871 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8872 * @param pDisk Pointer to HDD container.
8873 * @param nImage Image number, counts from 0. 0 is always base image of container.
8874 * @param puOpenFlags Where to store the image open flags.
8875 */
8876VBOXDDU_DECL(int) VDGetOpenFlags(PVBOXHDD pDisk, unsigned nImage,
8877 unsigned *puOpenFlags)
8878{
8879 int rc = VINF_SUCCESS;
8880 int rc2;
8881 bool fLockRead = false;
8882
8883 LogFlowFunc(("pDisk=%#p nImage=%u puOpenFlags=%#p\n",
8884 pDisk, nImage, puOpenFlags));
8885 do
8886 {
8887 /* sanity check */
8888 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8889 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8890
8891 /* Check arguments. */
8892 AssertMsgBreakStmt(VALID_PTR(puOpenFlags),
8893 ("puOpenFlags=%#p\n", puOpenFlags),
8894 rc = VERR_INVALID_PARAMETER);
8895
8896 rc2 = vdThreadStartRead(pDisk);
8897 AssertRC(rc2);
8898 fLockRead = true;
8899
8900 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8901 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8902
8903 *puOpenFlags = pImage->Backend->pfnGetOpenFlags(pImage->pBackendData);
8904 } while (0);
8905
8906 if (RT_UNLIKELY(fLockRead))
8907 {
8908 rc2 = vdThreadFinishRead(pDisk);
8909 AssertRC(rc2);
8910 }
8911
8912 LogFlowFunc(("returns %Rrc uOpenFlags=%#x\n", rc, *puOpenFlags));
8913 return rc;
8914}
8915
8916/**
8917 * Set open flags of image in HDD container.
8918 * This operation may cause file locking changes and/or files being reopened.
8919 * Note that in case of unrecoverable error all images in HDD container will be closed.
8920 *
8921 * @returns VBox status code.
8922 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8923 * @param pDisk Pointer to HDD container.
8924 * @param nImage Image number, counts from 0. 0 is always base image of container.
8925 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
8926 */
8927VBOXDDU_DECL(int) VDSetOpenFlags(PVBOXHDD pDisk, unsigned nImage,
8928 unsigned uOpenFlags)
8929{
8930 int rc;
8931 int rc2;
8932 bool fLockWrite = false;
8933
8934 LogFlowFunc(("pDisk=%#p uOpenFlags=%#u\n", pDisk, uOpenFlags));
8935 do
8936 {
8937 /* sanity check */
8938 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
8939 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
8940
8941 /* Check arguments. */
8942 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
8943 ("uOpenFlags=%#x\n", uOpenFlags),
8944 rc = VERR_INVALID_PARAMETER);
8945
8946 rc2 = vdThreadStartWrite(pDisk);
8947 AssertRC(rc2);
8948 fLockWrite = true;
8949
8950 /* Destroy any discard state because the image might be changed to readonly mode. */
8951 rc = vdDiscardStateDestroy(pDisk);
8952 if (RT_FAILURE(rc))
8953 break;
8954
8955 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8956 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8957
8958 rc = pImage->Backend->pfnSetOpenFlags(pImage->pBackendData,
8959 uOpenFlags & ~(VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS));
8960 if (RT_SUCCESS(rc))
8961 pImage->uOpenFlags = uOpenFlags & (VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_DISCARD | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS);
8962 } while (0);
8963
8964 if (RT_UNLIKELY(fLockWrite))
8965 {
8966 rc2 = vdThreadFinishWrite(pDisk);
8967 AssertRC(rc2);
8968 }
8969
8970 LogFlowFunc(("returns %Rrc\n", rc));
8971 return rc;
8972}
8973
8974/**
8975 * Get base filename of image in HDD container. Some image formats use
8976 * other filenames as well, so don't use this for anything but informational
8977 * purposes.
8978 *
8979 * @returns VBox status code.
8980 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8981 * @returns VERR_BUFFER_OVERFLOW if pszFilename buffer too small to hold filename.
8982 * @param pDisk Pointer to HDD container.
8983 * @param nImage Image number, counts from 0. 0 is always base image of container.
8984 * @param pszFilename Where to store the image file name.
8985 * @param cbFilename Size of buffer pszFilename points to.
8986 */
8987VBOXDDU_DECL(int) VDGetFilename(PVBOXHDD pDisk, unsigned nImage,
8988 char *pszFilename, unsigned cbFilename)
8989{
8990 int rc;
8991 int rc2;
8992 bool fLockRead = false;
8993
8994 LogFlowFunc(("pDisk=%#p nImage=%u pszFilename=%#p cbFilename=%u\n",
8995 pDisk, nImage, pszFilename, cbFilename));
8996 do
8997 {
8998 /* sanity check */
8999 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9000 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9001
9002 /* Check arguments. */
9003 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
9004 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
9005 rc = VERR_INVALID_PARAMETER);
9006 AssertMsgBreakStmt(cbFilename,
9007 ("cbFilename=%u\n", cbFilename),
9008 rc = VERR_INVALID_PARAMETER);
9009
9010 rc2 = vdThreadStartRead(pDisk);
9011 AssertRC(rc2);
9012 fLockRead = true;
9013
9014 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9015 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9016
9017 size_t cb = strlen(pImage->pszFilename);
9018 if (cb <= cbFilename)
9019 {
9020 strcpy(pszFilename, pImage->pszFilename);
9021 rc = VINF_SUCCESS;
9022 }
9023 else
9024 {
9025 strncpy(pszFilename, pImage->pszFilename, cbFilename - 1);
9026 pszFilename[cbFilename - 1] = '\0';
9027 rc = VERR_BUFFER_OVERFLOW;
9028 }
9029 } while (0);
9030
9031 if (RT_UNLIKELY(fLockRead))
9032 {
9033 rc2 = vdThreadFinishRead(pDisk);
9034 AssertRC(rc2);
9035 }
9036
9037 LogFlowFunc(("returns %Rrc, pszFilename=\"%s\"\n", rc, pszFilename));
9038 return rc;
9039}
9040
9041/**
9042 * Get the comment line of image in HDD container.
9043 *
9044 * @returns VBox status code.
9045 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9046 * @returns VERR_BUFFER_OVERFLOW if pszComment buffer too small to hold comment text.
9047 * @param pDisk Pointer to HDD container.
9048 * @param nImage Image number, counts from 0. 0 is always base image of container.
9049 * @param pszComment Where to store the comment string of image. NULL is ok.
9050 * @param cbComment The size of pszComment buffer. 0 is ok.
9051 */
9052VBOXDDU_DECL(int) VDGetComment(PVBOXHDD pDisk, unsigned nImage,
9053 char *pszComment, unsigned cbComment)
9054{
9055 int rc;
9056 int rc2;
9057 bool fLockRead = false;
9058
9059 LogFlowFunc(("pDisk=%#p nImage=%u pszComment=%#p cbComment=%u\n",
9060 pDisk, nImage, pszComment, cbComment));
9061 do
9062 {
9063 /* sanity check */
9064 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9065 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9066
9067 /* Check arguments. */
9068 AssertMsgBreakStmt(VALID_PTR(pszComment),
9069 ("pszComment=%#p \"%s\"\n", pszComment, pszComment),
9070 rc = VERR_INVALID_PARAMETER);
9071 AssertMsgBreakStmt(cbComment,
9072 ("cbComment=%u\n", cbComment),
9073 rc = VERR_INVALID_PARAMETER);
9074
9075 rc2 = vdThreadStartRead(pDisk);
9076 AssertRC(rc2);
9077 fLockRead = true;
9078
9079 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9080 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9081
9082 rc = pImage->Backend->pfnGetComment(pImage->pBackendData, pszComment,
9083 cbComment);
9084 } while (0);
9085
9086 if (RT_UNLIKELY(fLockRead))
9087 {
9088 rc2 = vdThreadFinishRead(pDisk);
9089 AssertRC(rc2);
9090 }
9091
9092 LogFlowFunc(("returns %Rrc, pszComment=\"%s\"\n", rc, pszComment));
9093 return rc;
9094}
9095
9096/**
9097 * Changes the comment line of image in HDD container.
9098 *
9099 * @returns VBox status code.
9100 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9101 * @param pDisk Pointer to HDD container.
9102 * @param nImage Image number, counts from 0. 0 is always base image of container.
9103 * @param pszComment New comment string (UTF-8). NULL is allowed to reset the comment.
9104 */
9105VBOXDDU_DECL(int) VDSetComment(PVBOXHDD pDisk, unsigned nImage,
9106 const char *pszComment)
9107{
9108 int rc;
9109 int rc2;
9110 bool fLockWrite = false;
9111
9112 LogFlowFunc(("pDisk=%#p nImage=%u pszComment=%#p \"%s\"\n",
9113 pDisk, nImage, pszComment, pszComment));
9114 do
9115 {
9116 /* sanity check */
9117 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9118 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9119
9120 /* Check arguments. */
9121 AssertMsgBreakStmt(VALID_PTR(pszComment) || pszComment == NULL,
9122 ("pszComment=%#p \"%s\"\n", pszComment, pszComment),
9123 rc = VERR_INVALID_PARAMETER);
9124
9125 rc2 = vdThreadStartWrite(pDisk);
9126 AssertRC(rc2);
9127 fLockWrite = true;
9128
9129 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9130 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9131
9132 rc = pImage->Backend->pfnSetComment(pImage->pBackendData, pszComment);
9133 } while (0);
9134
9135 if (RT_UNLIKELY(fLockWrite))
9136 {
9137 rc2 = vdThreadFinishWrite(pDisk);
9138 AssertRC(rc2);
9139 }
9140
9141 LogFlowFunc(("returns %Rrc\n", rc));
9142 return rc;
9143}
9144
9145
9146/**
9147 * Get UUID of image in HDD container.
9148 *
9149 * @returns VBox status code.
9150 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9151 * @param pDisk Pointer to HDD container.
9152 * @param nImage Image number, counts from 0. 0 is always base image of container.
9153 * @param pUuid Where to store the image creation UUID.
9154 */
9155VBOXDDU_DECL(int) VDGetUuid(PVBOXHDD pDisk, unsigned nImage, PRTUUID pUuid)
9156{
9157 int rc;
9158 int rc2;
9159 bool fLockRead = false;
9160
9161 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p\n", pDisk, nImage, pUuid));
9162 do
9163 {
9164 /* sanity check */
9165 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9166 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9167
9168 /* Check arguments. */
9169 AssertMsgBreakStmt(VALID_PTR(pUuid),
9170 ("pUuid=%#p\n", pUuid),
9171 rc = VERR_INVALID_PARAMETER);
9172
9173 rc2 = vdThreadStartRead(pDisk);
9174 AssertRC(rc2);
9175 fLockRead = true;
9176
9177 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9178 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9179
9180 rc = pImage->Backend->pfnGetUuid(pImage->pBackendData, pUuid);
9181 } while (0);
9182
9183 if (RT_UNLIKELY(fLockRead))
9184 {
9185 rc2 = vdThreadFinishRead(pDisk);
9186 AssertRC(rc2);
9187 }
9188
9189 LogFlowFunc(("returns %Rrc, Uuid={%RTuuid}\n", rc, pUuid));
9190 return rc;
9191}
9192
9193/**
9194 * Set the image's UUID. Should not be used by normal applications.
9195 *
9196 * @returns VBox status code.
9197 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9198 * @param pDisk Pointer to HDD container.
9199 * @param nImage Image number, counts from 0. 0 is always base image of container.
9200 * @param pUuid New UUID of the image. If NULL, a new UUID is created.
9201 */
9202VBOXDDU_DECL(int) VDSetUuid(PVBOXHDD pDisk, unsigned nImage, PCRTUUID pUuid)
9203{
9204 int rc;
9205 int rc2;
9206 bool fLockWrite = false;
9207
9208 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p {%RTuuid}\n",
9209 pDisk, nImage, pUuid, pUuid));
9210 do
9211 {
9212 /* sanity check */
9213 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9214 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9215
9216 AssertMsgBreakStmt(VALID_PTR(pUuid) || pUuid == NULL,
9217 ("pUuid=%#p\n", pUuid),
9218 rc = VERR_INVALID_PARAMETER);
9219
9220 rc2 = vdThreadStartWrite(pDisk);
9221 AssertRC(rc2);
9222 fLockWrite = true;
9223
9224 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9225 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9226
9227 RTUUID Uuid;
9228 if (!pUuid)
9229 {
9230 RTUuidCreate(&Uuid);
9231 pUuid = &Uuid;
9232 }
9233 rc = pImage->Backend->pfnSetUuid(pImage->pBackendData, pUuid);
9234 } while (0);
9235
9236 if (RT_UNLIKELY(fLockWrite))
9237 {
9238 rc2 = vdThreadFinishWrite(pDisk);
9239 AssertRC(rc2);
9240 }
9241
9242 LogFlowFunc(("returns %Rrc\n", rc));
9243 return rc;
9244}
9245
9246/**
9247 * Get last modification UUID of image in HDD container.
9248 *
9249 * @returns VBox status code.
9250 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9251 * @param pDisk Pointer to HDD container.
9252 * @param nImage Image number, counts from 0. 0 is always base image of container.
9253 * @param pUuid Where to store the image modification UUID.
9254 */
9255VBOXDDU_DECL(int) VDGetModificationUuid(PVBOXHDD pDisk, unsigned nImage, PRTUUID pUuid)
9256{
9257 int rc = VINF_SUCCESS;
9258 int rc2;
9259 bool fLockRead = false;
9260
9261 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p\n", pDisk, nImage, pUuid));
9262 do
9263 {
9264 /* sanity check */
9265 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9266 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9267
9268 /* Check arguments. */
9269 AssertMsgBreakStmt(VALID_PTR(pUuid),
9270 ("pUuid=%#p\n", pUuid),
9271 rc = VERR_INVALID_PARAMETER);
9272
9273 rc2 = vdThreadStartRead(pDisk);
9274 AssertRC(rc2);
9275 fLockRead = true;
9276
9277 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9278 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9279
9280 rc = pImage->Backend->pfnGetModificationUuid(pImage->pBackendData,
9281 pUuid);
9282 } while (0);
9283
9284 if (RT_UNLIKELY(fLockRead))
9285 {
9286 rc2 = vdThreadFinishRead(pDisk);
9287 AssertRC(rc2);
9288 }
9289
9290 LogFlowFunc(("returns %Rrc, Uuid={%RTuuid}\n", rc, pUuid));
9291 return rc;
9292}
9293
9294/**
9295 * Set the image's last modification UUID. Should not be used by normal applications.
9296 *
9297 * @returns VBox status code.
9298 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9299 * @param pDisk Pointer to HDD container.
9300 * @param nImage Image number, counts from 0. 0 is always base image of container.
9301 * @param pUuid New modification UUID of the image. If NULL, a new UUID is created.
9302 */
9303VBOXDDU_DECL(int) VDSetModificationUuid(PVBOXHDD pDisk, unsigned nImage, PCRTUUID pUuid)
9304{
9305 int rc;
9306 int rc2;
9307 bool fLockWrite = false;
9308
9309 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p {%RTuuid}\n",
9310 pDisk, nImage, pUuid, pUuid));
9311 do
9312 {
9313 /* sanity check */
9314 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9315 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9316
9317 /* Check arguments. */
9318 AssertMsgBreakStmt(VALID_PTR(pUuid) || pUuid == NULL,
9319 ("pUuid=%#p\n", pUuid),
9320 rc = VERR_INVALID_PARAMETER);
9321
9322 rc2 = vdThreadStartWrite(pDisk);
9323 AssertRC(rc2);
9324 fLockWrite = true;
9325
9326 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9327 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9328
9329 RTUUID Uuid;
9330 if (!pUuid)
9331 {
9332 RTUuidCreate(&Uuid);
9333 pUuid = &Uuid;
9334 }
9335 rc = pImage->Backend->pfnSetModificationUuid(pImage->pBackendData,
9336 pUuid);
9337 } while (0);
9338
9339 if (RT_UNLIKELY(fLockWrite))
9340 {
9341 rc2 = vdThreadFinishWrite(pDisk);
9342 AssertRC(rc2);
9343 }
9344
9345 LogFlowFunc(("returns %Rrc\n", rc));
9346 return rc;
9347}
9348
9349/**
9350 * Get parent UUID of image in HDD container.
9351 *
9352 * @returns VBox status code.
9353 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9354 * @param pDisk Pointer to HDD container.
9355 * @param nImage Image number, counts from 0. 0 is always base image of container.
9356 * @param pUuid Where to store the parent image UUID.
9357 */
9358VBOXDDU_DECL(int) VDGetParentUuid(PVBOXHDD pDisk, unsigned nImage,
9359 PRTUUID pUuid)
9360{
9361 int rc = VINF_SUCCESS;
9362 int rc2;
9363 bool fLockRead = false;
9364
9365 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p\n", pDisk, nImage, pUuid));
9366 do
9367 {
9368 /* sanity check */
9369 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9370 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9371
9372 /* Check arguments. */
9373 AssertMsgBreakStmt(VALID_PTR(pUuid),
9374 ("pUuid=%#p\n", pUuid),
9375 rc = VERR_INVALID_PARAMETER);
9376
9377 rc2 = vdThreadStartRead(pDisk);
9378 AssertRC(rc2);
9379 fLockRead = true;
9380
9381 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9382 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9383
9384 rc = pImage->Backend->pfnGetParentUuid(pImage->pBackendData, pUuid);
9385 } while (0);
9386
9387 if (RT_UNLIKELY(fLockRead))
9388 {
9389 rc2 = vdThreadFinishRead(pDisk);
9390 AssertRC(rc2);
9391 }
9392
9393 LogFlowFunc(("returns %Rrc, Uuid={%RTuuid}\n", rc, pUuid));
9394 return rc;
9395}
9396
9397/**
9398 * Set the image's parent UUID. Should not be used by normal applications.
9399 *
9400 * @returns VBox status code.
9401 * @param pDisk Pointer to HDD container.
9402 * @param nImage Image number, counts from 0. 0 is always base image of container.
9403 * @param pUuid New parent UUID of the image. If NULL, a new UUID is created.
9404 */
9405VBOXDDU_DECL(int) VDSetParentUuid(PVBOXHDD pDisk, unsigned nImage,
9406 PCRTUUID pUuid)
9407{
9408 int rc;
9409 int rc2;
9410 bool fLockWrite = false;
9411
9412 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p {%RTuuid}\n",
9413 pDisk, nImage, pUuid, pUuid));
9414 do
9415 {
9416 /* sanity check */
9417 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9418 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9419
9420 /* Check arguments. */
9421 AssertMsgBreakStmt(VALID_PTR(pUuid) || pUuid == NULL,
9422 ("pUuid=%#p\n", pUuid),
9423 rc = VERR_INVALID_PARAMETER);
9424
9425 rc2 = vdThreadStartWrite(pDisk);
9426 AssertRC(rc2);
9427 fLockWrite = true;
9428
9429 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9430 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9431
9432 RTUUID Uuid;
9433 if (!pUuid)
9434 {
9435 RTUuidCreate(&Uuid);
9436 pUuid = &Uuid;
9437 }
9438 rc = pImage->Backend->pfnSetParentUuid(pImage->pBackendData, pUuid);
9439 } while (0);
9440
9441 if (RT_UNLIKELY(fLockWrite))
9442 {
9443 rc2 = vdThreadFinishWrite(pDisk);
9444 AssertRC(rc2);
9445 }
9446
9447 LogFlowFunc(("returns %Rrc\n", rc));
9448 return rc;
9449}
9450
9451
9452/**
9453 * Debug helper - dumps all opened images in HDD container into the log file.
9454 *
9455 * @param pDisk Pointer to HDD container.
9456 */
9457VBOXDDU_DECL(void) VDDumpImages(PVBOXHDD pDisk)
9458{
9459 int rc2;
9460 bool fLockRead = false;
9461
9462 do
9463 {
9464 /* sanity check */
9465 AssertPtrBreak(pDisk);
9466 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9467
9468 if (!pDisk->pInterfaceError || !VALID_PTR(pDisk->pInterfaceError->pfnMessage))
9469 pDisk->pInterfaceError->pfnMessage = vdLogMessage;
9470
9471 rc2 = vdThreadStartRead(pDisk);
9472 AssertRC(rc2);
9473 fLockRead = true;
9474
9475 vdMessageWrapper(pDisk, "--- Dumping VD Disk, Images=%u\n", pDisk->cImages);
9476 for (PVDIMAGE pImage = pDisk->pBase; pImage; pImage = pImage->pNext)
9477 {
9478 vdMessageWrapper(pDisk, "Dumping VD image \"%s\" (Backend=%s)\n",
9479 pImage->pszFilename, pImage->Backend->pszBackendName);
9480 pImage->Backend->pfnDump(pImage->pBackendData);
9481 }
9482 } while (0);
9483
9484 if (RT_UNLIKELY(fLockRead))
9485 {
9486 rc2 = vdThreadFinishRead(pDisk);
9487 AssertRC(rc2);
9488 }
9489}
9490
9491
9492VBOXDDU_DECL(int) VDDiscardRanges(PVBOXHDD pDisk, PCRTRANGE paRanges, unsigned cRanges)
9493{
9494 int rc;
9495 int rc2;
9496 bool fLockWrite = false;
9497
9498 LogFlowFunc(("pDisk=%#p paRanges=%#p cRanges=%u\n",
9499 pDisk, paRanges, cRanges));
9500 do
9501 {
9502 /* sanity check */
9503 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9504 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9505
9506 /* Check arguments. */
9507 AssertMsgBreakStmt(cRanges,
9508 ("cRanges=%u\n", cRanges),
9509 rc = VERR_INVALID_PARAMETER);
9510 AssertMsgBreakStmt(VALID_PTR(paRanges),
9511 ("paRanges=%#p\n", paRanges),
9512 rc = VERR_INVALID_PARAMETER);
9513
9514 rc2 = vdThreadStartWrite(pDisk);
9515 AssertRC(rc2);
9516 fLockWrite = true;
9517
9518 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
9519
9520 AssertMsgBreakStmt(pDisk->pLast->uOpenFlags & VD_OPEN_FLAGS_DISCARD,
9521 ("Discarding not supported\n"),
9522 rc = VERR_NOT_SUPPORTED);
9523
9524 PVDIOCTX pIoCtx = vdIoCtxDiscardAlloc(pDisk, paRanges, cRanges,
9525 vdIoCtxSyncComplete, pDisk, NULL, NULL,
9526 vdDiscardHelperAsync,
9527 VDIOCTX_FLAGS_SYNC);
9528 if (!pIoCtx)
9529 {
9530 rc = VERR_NO_MEMORY;
9531 break;
9532 }
9533
9534 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
9535 if (rc == VINF_VD_ASYNC_IO_FINISHED)
9536 rc = VINF_SUCCESS;
9537
9538 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
9539 {
9540 rc = RTSemEventWait(pDisk->hEventSemSyncIo, RT_INDEFINITE_WAIT);
9541 AssertRC(rc);
9542
9543 rc = pDisk->rcSync;
9544 }
9545 else /* Success or error. */
9546 vdIoCtxFree(pDisk, pIoCtx);
9547 } while (0);
9548
9549 if (RT_UNLIKELY(fLockWrite))
9550 {
9551 rc2 = vdThreadFinishWrite(pDisk);
9552 AssertRC(rc2);
9553 }
9554
9555 LogFlowFunc(("returns %Rrc\n", rc));
9556 return rc;
9557}
9558
9559
9560VBOXDDU_DECL(int) VDAsyncRead(PVBOXHDD pDisk, uint64_t uOffset, size_t cbRead,
9561 PCRTSGBUF pcSgBuf,
9562 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
9563 void *pvUser1, void *pvUser2)
9564{
9565 int rc = VERR_VD_BLOCK_FREE;
9566 int rc2;
9567 bool fLockRead = false;
9568 PVDIOCTX pIoCtx = NULL;
9569
9570 LogFlowFunc(("pDisk=%#p uOffset=%llu pcSgBuf=%#p cbRead=%zu pvUser1=%#p pvUser2=%#p\n",
9571 pDisk, uOffset, pcSgBuf, cbRead, pvUser1, pvUser2));
9572
9573 do
9574 {
9575 /* sanity check */
9576 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9577 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9578
9579 /* Check arguments. */
9580 AssertMsgBreakStmt(cbRead,
9581 ("cbRead=%zu\n", cbRead),
9582 rc = VERR_INVALID_PARAMETER);
9583 AssertMsgBreakStmt(VALID_PTR(pcSgBuf),
9584 ("pcSgBuf=%#p\n", pcSgBuf),
9585 rc = VERR_INVALID_PARAMETER);
9586
9587 rc2 = vdThreadStartRead(pDisk);
9588 AssertRC(rc2);
9589 fLockRead = true;
9590
9591 AssertMsgBreakStmt(uOffset + cbRead <= pDisk->cbSize,
9592 ("uOffset=%llu cbRead=%zu pDisk->cbSize=%llu\n",
9593 uOffset, cbRead, pDisk->cbSize),
9594 rc = VERR_INVALID_PARAMETER);
9595 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
9596
9597 pIoCtx = vdIoCtxRootAlloc(pDisk, VDIOCTXTXDIR_READ, uOffset,
9598 cbRead, pDisk->pLast, pcSgBuf,
9599 pfnComplete, pvUser1, pvUser2,
9600 NULL, vdReadHelperAsync);
9601 if (!pIoCtx)
9602 {
9603 rc = VERR_NO_MEMORY;
9604 break;
9605 }
9606
9607 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
9608 if (rc == VINF_VD_ASYNC_IO_FINISHED)
9609 {
9610 if (ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
9611 vdIoCtxFree(pDisk, pIoCtx);
9612 else
9613 rc = VERR_VD_ASYNC_IO_IN_PROGRESS; /* Let the other handler complete the request. */
9614 }
9615 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS) /* Another error */
9616 vdIoCtxFree(pDisk, pIoCtx);
9617
9618 } while (0);
9619
9620 if (RT_UNLIKELY(fLockRead) && ( rc == VINF_VD_ASYNC_IO_FINISHED
9621 || rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
9622 {
9623 rc2 = vdThreadFinishRead(pDisk);
9624 AssertRC(rc2);
9625 }
9626
9627 LogFlowFunc(("returns %Rrc\n", rc));
9628 return rc;
9629}
9630
9631
9632VBOXDDU_DECL(int) VDAsyncWrite(PVBOXHDD pDisk, uint64_t uOffset, size_t cbWrite,
9633 PCRTSGBUF pcSgBuf,
9634 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
9635 void *pvUser1, void *pvUser2)
9636{
9637 int rc;
9638 int rc2;
9639 bool fLockWrite = false;
9640 PVDIOCTX pIoCtx = NULL;
9641
9642 LogFlowFunc(("pDisk=%#p uOffset=%llu cSgBuf=%#p cbWrite=%zu pvUser1=%#p pvUser2=%#p\n",
9643 pDisk, uOffset, pcSgBuf, cbWrite, pvUser1, pvUser2));
9644 do
9645 {
9646 /* sanity check */
9647 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9648 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9649
9650 /* Check arguments. */
9651 AssertMsgBreakStmt(cbWrite,
9652 ("cbWrite=%zu\n", cbWrite),
9653 rc = VERR_INVALID_PARAMETER);
9654 AssertMsgBreakStmt(VALID_PTR(pcSgBuf),
9655 ("pcSgBuf=%#p\n", pcSgBuf),
9656 rc = VERR_INVALID_PARAMETER);
9657
9658 rc2 = vdThreadStartWrite(pDisk);
9659 AssertRC(rc2);
9660 fLockWrite = true;
9661
9662 AssertMsgBreakStmt(uOffset + cbWrite <= pDisk->cbSize,
9663 ("uOffset=%llu cbWrite=%zu pDisk->cbSize=%llu\n",
9664 uOffset, cbWrite, pDisk->cbSize),
9665 rc = VERR_INVALID_PARAMETER);
9666 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
9667
9668 pIoCtx = vdIoCtxRootAlloc(pDisk, VDIOCTXTXDIR_WRITE, uOffset,
9669 cbWrite, pDisk->pLast, pcSgBuf,
9670 pfnComplete, pvUser1, pvUser2,
9671 NULL, vdWriteHelperAsync);
9672 if (!pIoCtx)
9673 {
9674 rc = VERR_NO_MEMORY;
9675 break;
9676 }
9677
9678 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
9679 if (rc == VINF_VD_ASYNC_IO_FINISHED)
9680 {
9681 if (ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
9682 vdIoCtxFree(pDisk, pIoCtx);
9683 else
9684 rc = VERR_VD_ASYNC_IO_IN_PROGRESS; /* Let the other handler complete the request. */
9685 }
9686 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS) /* Another error */
9687 vdIoCtxFree(pDisk, pIoCtx);
9688 } while (0);
9689
9690 if (RT_UNLIKELY(fLockWrite) && ( rc == VINF_VD_ASYNC_IO_FINISHED
9691 || rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
9692 {
9693 rc2 = vdThreadFinishWrite(pDisk);
9694 AssertRC(rc2);
9695 }
9696
9697 LogFlowFunc(("returns %Rrc\n", rc));
9698 return rc;
9699}
9700
9701
9702VBOXDDU_DECL(int) VDAsyncFlush(PVBOXHDD pDisk, PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
9703 void *pvUser1, void *pvUser2)
9704{
9705 int rc;
9706 int rc2;
9707 bool fLockWrite = false;
9708 PVDIOCTX pIoCtx = NULL;
9709
9710 LogFlowFunc(("pDisk=%#p\n", pDisk));
9711
9712 do
9713 {
9714 /* sanity check */
9715 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9716 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9717
9718 rc2 = vdThreadStartWrite(pDisk);
9719 AssertRC(rc2);
9720 fLockWrite = true;
9721
9722 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
9723
9724 pIoCtx = vdIoCtxRootAlloc(pDisk, VDIOCTXTXDIR_FLUSH, 0,
9725 0, pDisk->pLast, NULL,
9726 pfnComplete, pvUser1, pvUser2,
9727 NULL, vdFlushHelperAsync);
9728 if (!pIoCtx)
9729 {
9730 rc = VERR_NO_MEMORY;
9731 break;
9732 }
9733
9734 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
9735 if (rc == VINF_VD_ASYNC_IO_FINISHED)
9736 {
9737 if (ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
9738 vdIoCtxFree(pDisk, pIoCtx);
9739 else
9740 rc = VERR_VD_ASYNC_IO_IN_PROGRESS; /* Let the other handler complete the request. */
9741 }
9742 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS) /* Another error */
9743 vdIoCtxFree(pDisk, pIoCtx);
9744 } while (0);
9745
9746 if (RT_UNLIKELY(fLockWrite) && ( rc == VINF_VD_ASYNC_IO_FINISHED
9747 || rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
9748 {
9749 rc2 = vdThreadFinishWrite(pDisk);
9750 AssertRC(rc2);
9751 }
9752
9753 LogFlowFunc(("returns %Rrc\n", rc));
9754 return rc;
9755}
9756
9757VBOXDDU_DECL(int) VDAsyncDiscardRanges(PVBOXHDD pDisk, PCRTRANGE paRanges, unsigned cRanges,
9758 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
9759 void *pvUser1, void *pvUser2)
9760{
9761 int rc;
9762 int rc2;
9763 bool fLockWrite = false;
9764 PVDIOCTX pIoCtx = NULL;
9765
9766 LogFlowFunc(("pDisk=%#p\n", pDisk));
9767
9768 do
9769 {
9770 /* sanity check */
9771 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9772 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9773
9774 rc2 = vdThreadStartWrite(pDisk);
9775 AssertRC(rc2);
9776 fLockWrite = true;
9777
9778 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
9779
9780 pIoCtx = vdIoCtxDiscardAlloc(pDisk, paRanges, cRanges,
9781 pfnComplete, pvUser1, pvUser2, NULL,
9782 vdDiscardHelperAsync,
9783 0 /* fFlags */);
9784 if (!pIoCtx)
9785 {
9786 rc = VERR_NO_MEMORY;
9787 break;
9788 }
9789
9790 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
9791 if (rc == VINF_VD_ASYNC_IO_FINISHED)
9792 {
9793 if (ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
9794 vdIoCtxFree(pDisk, pIoCtx);
9795 else
9796 rc = VERR_VD_ASYNC_IO_IN_PROGRESS; /* Let the other handler complete the request. */
9797 }
9798 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS) /* Another error */
9799 vdIoCtxFree(pDisk, pIoCtx);
9800 } while (0);
9801
9802 if (RT_UNLIKELY(fLockWrite) && ( rc == VINF_VD_ASYNC_IO_FINISHED
9803 || rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
9804 {
9805 rc2 = vdThreadFinishWrite(pDisk);
9806 AssertRC(rc2);
9807 }
9808
9809 LogFlowFunc(("returns %Rrc\n", rc));
9810 return rc;
9811}
9812
9813VBOXDDU_DECL(int) VDRepair(PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
9814 const char *pszFilename, const char *pszBackend,
9815 uint32_t fFlags)
9816{
9817 int rc = VERR_NOT_SUPPORTED;
9818 PCVBOXHDDBACKEND pBackend = NULL;
9819 VDINTERFACEIOINT VDIfIoInt;
9820 VDINTERFACEIO VDIfIoFallback;
9821 PVDINTERFACEIO pInterfaceIo;
9822
9823 LogFlowFunc(("pszFilename=\"%s\"\n", pszFilename));
9824 /* Check arguments. */
9825 AssertMsgReturn(VALID_PTR(pszFilename) && *pszFilename,
9826 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
9827 VERR_INVALID_PARAMETER);
9828 AssertMsgReturn(VALID_PTR(pszBackend),
9829 ("pszBackend=%#p\n", pszBackend),
9830 VERR_INVALID_PARAMETER);
9831 AssertMsgReturn((fFlags & ~VD_REPAIR_FLAGS_MASK) == 0,
9832 ("fFlags=%#x\n", fFlags),
9833 VERR_INVALID_PARAMETER);
9834
9835 pInterfaceIo = VDIfIoGet(pVDIfsImage);
9836 if (!pInterfaceIo)
9837 {
9838 /*
9839 * Caller doesn't provide an I/O interface, create our own using the
9840 * native file API.
9841 */
9842 vdIfIoFallbackCallbacksSetup(&VDIfIoFallback);
9843 pInterfaceIo = &VDIfIoFallback;
9844 }
9845
9846 /* Set up the internal I/O interface. */
9847 AssertReturn(!VDIfIoIntGet(pVDIfsImage), VERR_INVALID_PARAMETER);
9848 VDIfIoInt.pfnOpen = vdIOIntOpenLimited;
9849 VDIfIoInt.pfnClose = vdIOIntCloseLimited;
9850 VDIfIoInt.pfnDelete = vdIOIntDeleteLimited;
9851 VDIfIoInt.pfnMove = vdIOIntMoveLimited;
9852 VDIfIoInt.pfnGetFreeSpace = vdIOIntGetFreeSpaceLimited;
9853 VDIfIoInt.pfnGetModificationTime = vdIOIntGetModificationTimeLimited;
9854 VDIfIoInt.pfnGetSize = vdIOIntGetSizeLimited;
9855 VDIfIoInt.pfnSetSize = vdIOIntSetSizeLimited;
9856 VDIfIoInt.pfnReadUser = vdIOIntReadUserLimited;
9857 VDIfIoInt.pfnWriteUser = vdIOIntWriteUserLimited;
9858 VDIfIoInt.pfnReadMeta = vdIOIntReadMetaLimited;
9859 VDIfIoInt.pfnWriteMeta = vdIOIntWriteMetaLimited;
9860 VDIfIoInt.pfnFlush = vdIOIntFlushLimited;
9861 rc = VDInterfaceAdd(&VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
9862 pInterfaceIo, sizeof(VDINTERFACEIOINT), &pVDIfsImage);
9863 AssertRC(rc);
9864
9865 rc = vdFindBackend(pszBackend, &pBackend);
9866 if (RT_SUCCESS(rc))
9867 {
9868 if (pBackend->pfnRepair)
9869 rc = pBackend->pfnRepair(pszFilename, pVDIfsDisk, pVDIfsImage, fFlags);
9870 else
9871 rc = VERR_VD_IMAGE_REPAIR_NOT_SUPPORTED;
9872 }
9873
9874 LogFlowFunc(("returns %Rrc\n", rc));
9875 return rc;
9876}
9877
Note: See TracBrowser for help on using the repository browser.

© 2025 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette