VirtualBox

source: vbox/trunk/src/VBox/Storage/VMDK.cpp@ 49971

Last change on this file since 49971 was 49231, checked in by vboxsync, 11 years ago

VMDK: Sketched the code for using RTZipBlockDecompress in vmdkFileInflateSync.

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 242.0 KB
Line 
1/* $Id: VMDK.cpp 49231 2013-10-22 14:38:45Z vboxsync $ */
2/** @file
3 * VMDK disk image, core code.
4 */
5
6/*
7 * Copyright (C) 2006-2013 Oracle Corporation
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 */
17
18/*******************************************************************************
19* Header Files *
20*******************************************************************************/
21#define LOG_GROUP LOG_GROUP_VD_VMDK
22#include <VBox/vd-plugin.h>
23#include <VBox/err.h>
24
25#include <VBox/log.h>
26#include <iprt/assert.h>
27#include <iprt/alloc.h>
28#include <iprt/uuid.h>
29#include <iprt/path.h>
30#include <iprt/string.h>
31#include <iprt/rand.h>
32#include <iprt/zip.h>
33#include <iprt/asm.h>
34
35/*******************************************************************************
36* Constants And Macros, Structures and Typedefs *
37*******************************************************************************/
38
39/** Maximum encoded string size (including NUL) we allow for VMDK images.
40 * Deliberately not set high to avoid running out of descriptor space. */
41#define VMDK_ENCODED_COMMENT_MAX 1024
42
43/** VMDK descriptor DDB entry for PCHS cylinders. */
44#define VMDK_DDB_GEO_PCHS_CYLINDERS "ddb.geometry.cylinders"
45
46/** VMDK descriptor DDB entry for PCHS heads. */
47#define VMDK_DDB_GEO_PCHS_HEADS "ddb.geometry.heads"
48
49/** VMDK descriptor DDB entry for PCHS sectors. */
50#define VMDK_DDB_GEO_PCHS_SECTORS "ddb.geometry.sectors"
51
52/** VMDK descriptor DDB entry for LCHS cylinders. */
53#define VMDK_DDB_GEO_LCHS_CYLINDERS "ddb.geometry.biosCylinders"
54
55/** VMDK descriptor DDB entry for LCHS heads. */
56#define VMDK_DDB_GEO_LCHS_HEADS "ddb.geometry.biosHeads"
57
58/** VMDK descriptor DDB entry for LCHS sectors. */
59#define VMDK_DDB_GEO_LCHS_SECTORS "ddb.geometry.biosSectors"
60
61/** VMDK descriptor DDB entry for image UUID. */
62#define VMDK_DDB_IMAGE_UUID "ddb.uuid.image"
63
64/** VMDK descriptor DDB entry for image modification UUID. */
65#define VMDK_DDB_MODIFICATION_UUID "ddb.uuid.modification"
66
67/** VMDK descriptor DDB entry for parent image UUID. */
68#define VMDK_DDB_PARENT_UUID "ddb.uuid.parent"
69
70/** VMDK descriptor DDB entry for parent image modification UUID. */
71#define VMDK_DDB_PARENT_MODIFICATION_UUID "ddb.uuid.parentmodification"
72
73/** No compression for streamOptimized files. */
74#define VMDK_COMPRESSION_NONE 0
75
76/** Deflate compression for streamOptimized files. */
77#define VMDK_COMPRESSION_DEFLATE 1
78
79/** Marker that the actual GD value is stored in the footer. */
80#define VMDK_GD_AT_END 0xffffffffffffffffULL
81
82/** Marker for end-of-stream in streamOptimized images. */
83#define VMDK_MARKER_EOS 0
84
85/** Marker for grain table block in streamOptimized images. */
86#define VMDK_MARKER_GT 1
87
88/** Marker for grain directory block in streamOptimized images. */
89#define VMDK_MARKER_GD 2
90
91/** Marker for footer in streamOptimized images. */
92#define VMDK_MARKER_FOOTER 3
93
94/** Marker for unknown purpose in streamOptimized images.
95 * Shows up in very recent images created by vSphere, but only sporadically.
96 * They "forgot" to document that one in the VMDK specification. */
97#define VMDK_MARKER_UNSPECIFIED 4
98
99/** Dummy marker for "don't check the marker value". */
100#define VMDK_MARKER_IGNORE 0xffffffffU
101
102/**
103 * Magic number for hosted images created by VMware Workstation 4, VMware
104 * Workstation 5, VMware Server or VMware Player. Not necessarily sparse.
105 */
106#define VMDK_SPARSE_MAGICNUMBER 0x564d444b /* 'V' 'M' 'D' 'K' */
107
108/**
109 * VMDK hosted binary extent header. The "Sparse" is a total misnomer, as
110 * this header is also used for monolithic flat images.
111 */
112#pragma pack(1)
113typedef struct SparseExtentHeader
114{
115 uint32_t magicNumber;
116 uint32_t version;
117 uint32_t flags;
118 uint64_t capacity;
119 uint64_t grainSize;
120 uint64_t descriptorOffset;
121 uint64_t descriptorSize;
122 uint32_t numGTEsPerGT;
123 uint64_t rgdOffset;
124 uint64_t gdOffset;
125 uint64_t overHead;
126 bool uncleanShutdown;
127 char singleEndLineChar;
128 char nonEndLineChar;
129 char doubleEndLineChar1;
130 char doubleEndLineChar2;
131 uint16_t compressAlgorithm;
132 uint8_t pad[433];
133} SparseExtentHeader;
134#pragma pack()
135
136/** VMDK capacity for a single chunk when 2G splitting is turned on. Should be
137 * divisible by the default grain size (64K) */
138#define VMDK_2G_SPLIT_SIZE (2047 * 1024 * 1024)
139
140/** VMDK streamOptimized file format marker. The type field may or may not
141 * be actually valid, but there's always data to read there. */
142#pragma pack(1)
143typedef struct VMDKMARKER
144{
145 uint64_t uSector;
146 uint32_t cbSize;
147 uint32_t uType;
148} VMDKMARKER, *PVMDKMARKER;
149#pragma pack()
150
151
152#ifdef VBOX_WITH_VMDK_ESX
153
154/** @todo the ESX code is not tested, not used, and lacks error messages. */
155
156/**
157 * Magic number for images created by VMware GSX Server 3 or ESX Server 3.
158 */
159#define VMDK_ESX_SPARSE_MAGICNUMBER 0x44574f43 /* 'C' 'O' 'W' 'D' */
160
161#pragma pack(1)
162typedef struct COWDisk_Header
163{
164 uint32_t magicNumber;
165 uint32_t version;
166 uint32_t flags;
167 uint32_t numSectors;
168 uint32_t grainSize;
169 uint32_t gdOffset;
170 uint32_t numGDEntries;
171 uint32_t freeSector;
172 /* The spec incompletely documents quite a few further fields, but states
173 * that they are unused by the current format. Replace them by padding. */
174 char reserved1[1604];
175 uint32_t savedGeneration;
176 char reserved2[8];
177 uint32_t uncleanShutdown;
178 char padding[396];
179} COWDisk_Header;
180#pragma pack()
181#endif /* VBOX_WITH_VMDK_ESX */
182
183
184/** Convert sector number/size to byte offset/size. */
185#define VMDK_SECTOR2BYTE(u) ((uint64_t)(u) << 9)
186
187/** Convert byte offset/size to sector number/size. */
188#define VMDK_BYTE2SECTOR(u) ((u) >> 9)
189
190/**
191 * VMDK extent type.
192 */
193typedef enum VMDKETYPE
194{
195 /** Hosted sparse extent. */
196 VMDKETYPE_HOSTED_SPARSE = 1,
197 /** Flat extent. */
198 VMDKETYPE_FLAT,
199 /** Zero extent. */
200 VMDKETYPE_ZERO,
201 /** VMFS extent, used by ESX. */
202 VMDKETYPE_VMFS
203#ifdef VBOX_WITH_VMDK_ESX
204 ,
205 /** ESX sparse extent. */
206 VMDKETYPE_ESX_SPARSE
207#endif /* VBOX_WITH_VMDK_ESX */
208} VMDKETYPE, *PVMDKETYPE;
209
210/**
211 * VMDK access type for a extent.
212 */
213typedef enum VMDKACCESS
214{
215 /** No access allowed. */
216 VMDKACCESS_NOACCESS = 0,
217 /** Read-only access. */
218 VMDKACCESS_READONLY,
219 /** Read-write access. */
220 VMDKACCESS_READWRITE
221} VMDKACCESS, *PVMDKACCESS;
222
223/** Forward declaration for PVMDKIMAGE. */
224typedef struct VMDKIMAGE *PVMDKIMAGE;
225
226/**
227 * Extents files entry. Used for opening a particular file only once.
228 */
229typedef struct VMDKFILE
230{
231 /** Pointer to filename. Local copy. */
232 const char *pszFilename;
233 /** File open flags for consistency checking. */
234 unsigned fOpen;
235 /** Handle for sync/async file abstraction.*/
236 PVDIOSTORAGE pStorage;
237 /** Reference counter. */
238 unsigned uReferences;
239 /** Flag whether the file should be deleted on last close. */
240 bool fDelete;
241 /** Pointer to the image we belong to (for debugging purposes). */
242 PVMDKIMAGE pImage;
243 /** Pointer to next file descriptor. */
244 struct VMDKFILE *pNext;
245 /** Pointer to the previous file descriptor. */
246 struct VMDKFILE *pPrev;
247} VMDKFILE, *PVMDKFILE;
248
249/**
250 * VMDK extent data structure.
251 */
252typedef struct VMDKEXTENT
253{
254 /** File handle. */
255 PVMDKFILE pFile;
256 /** Base name of the image extent. */
257 const char *pszBasename;
258 /** Full name of the image extent. */
259 const char *pszFullname;
260 /** Number of sectors in this extent. */
261 uint64_t cSectors;
262 /** Number of sectors per block (grain in VMDK speak). */
263 uint64_t cSectorsPerGrain;
264 /** Starting sector number of descriptor. */
265 uint64_t uDescriptorSector;
266 /** Size of descriptor in sectors. */
267 uint64_t cDescriptorSectors;
268 /** Starting sector number of grain directory. */
269 uint64_t uSectorGD;
270 /** Starting sector number of redundant grain directory. */
271 uint64_t uSectorRGD;
272 /** Total number of metadata sectors. */
273 uint64_t cOverheadSectors;
274 /** Nominal size (i.e. as described by the descriptor) of this extent. */
275 uint64_t cNominalSectors;
276 /** Sector offset (i.e. as described by the descriptor) of this extent. */
277 uint64_t uSectorOffset;
278 /** Number of entries in a grain table. */
279 uint32_t cGTEntries;
280 /** Number of sectors reachable via a grain directory entry. */
281 uint32_t cSectorsPerGDE;
282 /** Number of entries in the grain directory. */
283 uint32_t cGDEntries;
284 /** Pointer to the next free sector. Legacy information. Do not use. */
285 uint32_t uFreeSector;
286 /** Number of this extent in the list of images. */
287 uint32_t uExtent;
288 /** Pointer to the descriptor (NULL if no descriptor in this extent). */
289 char *pDescData;
290 /** Pointer to the grain directory. */
291 uint32_t *pGD;
292 /** Pointer to the redundant grain directory. */
293 uint32_t *pRGD;
294 /** VMDK version of this extent. 1=1.0/1.1 */
295 uint32_t uVersion;
296 /** Type of this extent. */
297 VMDKETYPE enmType;
298 /** Access to this extent. */
299 VMDKACCESS enmAccess;
300 /** Flag whether this extent is marked as unclean. */
301 bool fUncleanShutdown;
302 /** Flag whether the metadata in the extent header needs to be updated. */
303 bool fMetaDirty;
304 /** Flag whether there is a footer in this extent. */
305 bool fFooter;
306 /** Compression type for this extent. */
307 uint16_t uCompression;
308 /** Append position for writing new grain. Only for sparse extents. */
309 uint64_t uAppendPosition;
310 /** Last grain which was accessed. Only for streamOptimized extents. */
311 uint32_t uLastGrainAccess;
312 /** Starting sector corresponding to the grain buffer. */
313 uint32_t uGrainSectorAbs;
314 /** Grain number corresponding to the grain buffer. */
315 uint32_t uGrain;
316 /** Actual size of the compressed data, only valid for reading. */
317 uint32_t cbGrainStreamRead;
318 /** Size of compressed grain buffer for streamOptimized extents. */
319 size_t cbCompGrain;
320 /** Compressed grain buffer for streamOptimized extents, with marker. */
321 void *pvCompGrain;
322 /** Decompressed grain buffer for streamOptimized extents. */
323 void *pvGrain;
324 /** Reference to the image in which this extent is used. Do not use this
325 * on a regular basis to avoid passing pImage references to functions
326 * explicitly. */
327 struct VMDKIMAGE *pImage;
328} VMDKEXTENT, *PVMDKEXTENT;
329
330/**
331 * Grain table cache size. Allocated per image.
332 */
333#define VMDK_GT_CACHE_SIZE 256
334
335/**
336 * Grain table block size. Smaller than an actual grain table block to allow
337 * more grain table blocks to be cached without having to allocate excessive
338 * amounts of memory for the cache.
339 */
340#define VMDK_GT_CACHELINE_SIZE 128
341
342
343/**
344 * Maximum number of lines in a descriptor file. Not worth the effort of
345 * making it variable. Descriptor files are generally very short (~20 lines),
346 * with the exception of sparse files split in 2G chunks, which need for the
347 * maximum size (almost 2T) exactly 1025 lines for the disk database.
348 */
349#define VMDK_DESCRIPTOR_LINES_MAX 1100U
350
351/**
352 * Parsed descriptor information. Allows easy access and update of the
353 * descriptor (whether separate file or not). Free form text files suck.
354 */
355typedef struct VMDKDESCRIPTOR
356{
357 /** Line number of first entry of the disk descriptor. */
358 unsigned uFirstDesc;
359 /** Line number of first entry in the extent description. */
360 unsigned uFirstExtent;
361 /** Line number of first disk database entry. */
362 unsigned uFirstDDB;
363 /** Total number of lines. */
364 unsigned cLines;
365 /** Total amount of memory available for the descriptor. */
366 size_t cbDescAlloc;
367 /** Set if descriptor has been changed and not yet written to disk. */
368 bool fDirty;
369 /** Array of pointers to the data in the descriptor. */
370 char *aLines[VMDK_DESCRIPTOR_LINES_MAX];
371 /** Array of line indices pointing to the next non-comment line. */
372 unsigned aNextLines[VMDK_DESCRIPTOR_LINES_MAX];
373} VMDKDESCRIPTOR, *PVMDKDESCRIPTOR;
374
375
376/**
377 * Cache entry for translating extent/sector to a sector number in that
378 * extent.
379 */
380typedef struct VMDKGTCACHEENTRY
381{
382 /** Extent number for which this entry is valid. */
383 uint32_t uExtent;
384 /** GT data block number. */
385 uint64_t uGTBlock;
386 /** Data part of the cache entry. */
387 uint32_t aGTData[VMDK_GT_CACHELINE_SIZE];
388} VMDKGTCACHEENTRY, *PVMDKGTCACHEENTRY;
389
390/**
391 * Cache data structure for blocks of grain table entries. For now this is a
392 * fixed size direct mapping cache, but this should be adapted to the size of
393 * the sparse image and maybe converted to a set-associative cache. The
394 * implementation below implements a write-through cache with write allocate.
395 */
396typedef struct VMDKGTCACHE
397{
398 /** Cache entries. */
399 VMDKGTCACHEENTRY aGTCache[VMDK_GT_CACHE_SIZE];
400 /** Number of cache entries (currently unused). */
401 unsigned cEntries;
402} VMDKGTCACHE, *PVMDKGTCACHE;
403
404/**
405 * Complete VMDK image data structure. Mainly a collection of extents and a few
406 * extra global data fields.
407 */
408typedef struct VMDKIMAGE
409{
410 /** Image name. */
411 const char *pszFilename;
412 /** Descriptor file if applicable. */
413 PVMDKFILE pFile;
414
415 /** Pointer to the per-disk VD interface list. */
416 PVDINTERFACE pVDIfsDisk;
417 /** Pointer to the per-image VD interface list. */
418 PVDINTERFACE pVDIfsImage;
419
420 /** Error interface. */
421 PVDINTERFACEERROR pIfError;
422 /** I/O interface. */
423 PVDINTERFACEIOINT pIfIo;
424
425
426 /** Pointer to the image extents. */
427 PVMDKEXTENT pExtents;
428 /** Number of image extents. */
429 unsigned cExtents;
430 /** Pointer to the files list, for opening a file referenced multiple
431 * times only once (happens mainly with raw partition access). */
432 PVMDKFILE pFiles;
433
434 /**
435 * Pointer to an array of segment entries for async I/O.
436 * This is an optimization because the task number to submit is not known
437 * and allocating/freeing an array in the read/write functions every time
438 * is too expensive.
439 */
440 PPDMDATASEG paSegments;
441 /** Entries available in the segments array. */
442 unsigned cSegments;
443
444 /** Open flags passed by VBoxHD layer. */
445 unsigned uOpenFlags;
446 /** Image flags defined during creation or determined during open. */
447 unsigned uImageFlags;
448 /** Total size of the image. */
449 uint64_t cbSize;
450 /** Physical geometry of this image. */
451 VDGEOMETRY PCHSGeometry;
452 /** Logical geometry of this image. */
453 VDGEOMETRY LCHSGeometry;
454 /** Image UUID. */
455 RTUUID ImageUuid;
456 /** Image modification UUID. */
457 RTUUID ModificationUuid;
458 /** Parent image UUID. */
459 RTUUID ParentUuid;
460 /** Parent image modification UUID. */
461 RTUUID ParentModificationUuid;
462
463 /** Pointer to grain table cache, if this image contains sparse extents. */
464 PVMDKGTCACHE pGTCache;
465 /** Pointer to the descriptor (NULL if no separate descriptor file). */
466 char *pDescData;
467 /** Allocation size of the descriptor file. */
468 size_t cbDescAlloc;
469 /** Parsed descriptor file content. */
470 VMDKDESCRIPTOR Descriptor;
471} VMDKIMAGE;
472
473
474/** State for the input/output callout of the inflate reader/deflate writer. */
475typedef struct VMDKCOMPRESSIO
476{
477 /* Image this operation relates to. */
478 PVMDKIMAGE pImage;
479 /* Current read position. */
480 ssize_t iOffset;
481 /* Size of the compressed grain buffer (available data). */
482 size_t cbCompGrain;
483 /* Pointer to the compressed grain buffer. */
484 void *pvCompGrain;
485} VMDKCOMPRESSIO;
486
487
488/** Tracks async grain allocation. */
489typedef struct VMDKGRAINALLOCASYNC
490{
491 /** Flag whether the allocation failed. */
492 bool fIoErr;
493 /** Current number of transfers pending.
494 * If reached 0 and there is an error the old state is restored. */
495 unsigned cIoXfersPending;
496 /** Sector number */
497 uint64_t uSector;
498 /** Flag whether the grain table needs to be updated. */
499 bool fGTUpdateNeeded;
500 /** Extent the allocation happens. */
501 PVMDKEXTENT pExtent;
502 /** Position of the new grain, required for the grain table update. */
503 uint64_t uGrainOffset;
504 /** Grain table sector. */
505 uint64_t uGTSector;
506 /** Backup grain table sector. */
507 uint64_t uRGTSector;
508} VMDKGRAINALLOCASYNC, *PVMDKGRAINALLOCASYNC;
509
510/*******************************************************************************
511* Static Variables *
512*******************************************************************************/
513
514/** NULL-terminated array of supported file extensions. */
515static const VDFILEEXTENSION s_aVmdkFileExtensions[] =
516{
517 {"vmdk", VDTYPE_HDD},
518 {NULL, VDTYPE_INVALID}
519};
520
521/*******************************************************************************
522* Internal Functions *
523*******************************************************************************/
524
525static void vmdkFreeStreamBuffers(PVMDKEXTENT pExtent);
526static int vmdkFreeExtentData(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
527 bool fDelete);
528
529static int vmdkCreateExtents(PVMDKIMAGE pImage, unsigned cExtents);
530static int vmdkFlushImage(PVMDKIMAGE pImage, PVDIOCTX pIoCtx);
531static int vmdkSetImageComment(PVMDKIMAGE pImage, const char *pszComment);
532static int vmdkFreeImage(PVMDKIMAGE pImage, bool fDelete);
533
534static int vmdkAllocGrainComplete(void *pBackendData, PVDIOCTX pIoCtx,
535 void *pvUser, int rcReq);
536
537/**
538 * Internal: open a file (using a file descriptor cache to ensure each file
539 * is only opened once - anything else can cause locking problems).
540 */
541static int vmdkFileOpen(PVMDKIMAGE pImage, PVMDKFILE *ppVmdkFile,
542 const char *pszFilename, uint32_t fOpen)
543{
544 int rc = VINF_SUCCESS;
545 PVMDKFILE pVmdkFile;
546
547 for (pVmdkFile = pImage->pFiles;
548 pVmdkFile != NULL;
549 pVmdkFile = pVmdkFile->pNext)
550 {
551 if (!strcmp(pszFilename, pVmdkFile->pszFilename))
552 {
553 Assert(fOpen == pVmdkFile->fOpen);
554 pVmdkFile->uReferences++;
555
556 *ppVmdkFile = pVmdkFile;
557
558 return rc;
559 }
560 }
561
562 /* If we get here, there's no matching entry in the cache. */
563 pVmdkFile = (PVMDKFILE)RTMemAllocZ(sizeof(VMDKFILE));
564 if (!VALID_PTR(pVmdkFile))
565 {
566 *ppVmdkFile = NULL;
567 return VERR_NO_MEMORY;
568 }
569
570 pVmdkFile->pszFilename = RTStrDup(pszFilename);
571 if (!VALID_PTR(pVmdkFile->pszFilename))
572 {
573 RTMemFree(pVmdkFile);
574 *ppVmdkFile = NULL;
575 return VERR_NO_MEMORY;
576 }
577 pVmdkFile->fOpen = fOpen;
578
579 rc = vdIfIoIntFileOpen(pImage->pIfIo, pszFilename, fOpen,
580 &pVmdkFile->pStorage);
581 if (RT_SUCCESS(rc))
582 {
583 pVmdkFile->uReferences = 1;
584 pVmdkFile->pImage = pImage;
585 pVmdkFile->pNext = pImage->pFiles;
586 if (pImage->pFiles)
587 pImage->pFiles->pPrev = pVmdkFile;
588 pImage->pFiles = pVmdkFile;
589 *ppVmdkFile = pVmdkFile;
590 }
591 else
592 {
593 RTStrFree((char *)(void *)pVmdkFile->pszFilename);
594 RTMemFree(pVmdkFile);
595 *ppVmdkFile = NULL;
596 }
597
598 return rc;
599}
600
601/**
602 * Internal: close a file, updating the file descriptor cache.
603 */
604static int vmdkFileClose(PVMDKIMAGE pImage, PVMDKFILE *ppVmdkFile, bool fDelete)
605{
606 int rc = VINF_SUCCESS;
607 PVMDKFILE pVmdkFile = *ppVmdkFile;
608
609 AssertPtr(pVmdkFile);
610
611 pVmdkFile->fDelete |= fDelete;
612 Assert(pVmdkFile->uReferences);
613 pVmdkFile->uReferences--;
614 if (pVmdkFile->uReferences == 0)
615 {
616 PVMDKFILE pPrev;
617 PVMDKFILE pNext;
618
619 /* Unchain the element from the list. */
620 pPrev = pVmdkFile->pPrev;
621 pNext = pVmdkFile->pNext;
622
623 if (pNext)
624 pNext->pPrev = pPrev;
625 if (pPrev)
626 pPrev->pNext = pNext;
627 else
628 pImage->pFiles = pNext;
629
630 rc = vdIfIoIntFileClose(pImage->pIfIo, pVmdkFile->pStorage);
631 if (RT_SUCCESS(rc) && pVmdkFile->fDelete)
632 rc = vdIfIoIntFileDelete(pImage->pIfIo, pVmdkFile->pszFilename);
633 RTStrFree((char *)(void *)pVmdkFile->pszFilename);
634 RTMemFree(pVmdkFile);
635 }
636
637 *ppVmdkFile = NULL;
638 return rc;
639}
640
641/*#define VMDK_USE_BLOCK_DECOMP_API - test and enable */
642#ifndef VMDK_USE_BLOCK_DECOMP_API
643static DECLCALLBACK(int) vmdkFileInflateHelper(void *pvUser, void *pvBuf, size_t cbBuf, size_t *pcbBuf)
644{
645 VMDKCOMPRESSIO *pInflateState = (VMDKCOMPRESSIO *)pvUser;
646 size_t cbInjected = 0;
647
648 Assert(cbBuf);
649 if (pInflateState->iOffset < 0)
650 {
651 *(uint8_t *)pvBuf = RTZIPTYPE_ZLIB;
652 pvBuf = (uint8_t *)pvBuf + 1;
653 cbBuf--;
654 cbInjected = 1;
655 pInflateState->iOffset = RT_OFFSETOF(VMDKMARKER, uType);
656 }
657 if (!cbBuf)
658 {
659 if (pcbBuf)
660 *pcbBuf = cbInjected;
661 return VINF_SUCCESS;
662 }
663 cbBuf = RT_MIN(cbBuf, pInflateState->cbCompGrain - pInflateState->iOffset);
664 memcpy(pvBuf,
665 (uint8_t *)pInflateState->pvCompGrain + pInflateState->iOffset,
666 cbBuf);
667 pInflateState->iOffset += cbBuf;
668 Assert(pcbBuf);
669 *pcbBuf = cbBuf + cbInjected;
670 return VINF_SUCCESS;
671}
672#endif
673
674/**
675 * Internal: read from a file and inflate the compressed data,
676 * distinguishing between async and normal operation
677 */
678DECLINLINE(int) vmdkFileInflateSync(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
679 uint64_t uOffset, void *pvBuf,
680 size_t cbToRead, const void *pcvMarker,
681 uint64_t *puLBA, uint32_t *pcbMarkerData)
682{
683 int rc;
684#ifndef VMDK_USE_BLOCK_DECOMP_API
685 PRTZIPDECOMP pZip = NULL;
686#endif
687 VMDKMARKER *pMarker = (VMDKMARKER *)pExtent->pvCompGrain;
688 size_t cbCompSize, cbActuallyRead;
689
690 if (!pcvMarker)
691 {
692 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
693 uOffset, pMarker, RT_OFFSETOF(VMDKMARKER, uType));
694 if (RT_FAILURE(rc))
695 return rc;
696 }
697 else
698 {
699 memcpy(pMarker, pcvMarker, RT_OFFSETOF(VMDKMARKER, uType));
700 /* pcvMarker endianness has already been partially transformed, fix it */
701 pMarker->uSector = RT_H2LE_U64(pMarker->uSector);
702 pMarker->cbSize = RT_H2LE_U32(pMarker->cbSize);
703 }
704
705 cbCompSize = RT_LE2H_U32(pMarker->cbSize);
706 if (cbCompSize == 0)
707 {
708 AssertMsgFailed(("VMDK: corrupted marker\n"));
709 return VERR_VD_VMDK_INVALID_FORMAT;
710 }
711
712 /* Sanity check - the expansion ratio should be much less than 2. */
713 Assert(cbCompSize < 2 * cbToRead);
714 if (cbCompSize >= 2 * cbToRead)
715 return VERR_VD_VMDK_INVALID_FORMAT;
716
717 /* Compressed grain marker. Data follows immediately. */
718 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
719 uOffset + RT_OFFSETOF(VMDKMARKER, uType),
720 (uint8_t *)pExtent->pvCompGrain
721 + RT_OFFSETOF(VMDKMARKER, uType),
722 RT_ALIGN_Z( cbCompSize
723 + RT_OFFSETOF(VMDKMARKER, uType),
724 512)
725 - RT_OFFSETOF(VMDKMARKER, uType));
726
727 if (puLBA)
728 *puLBA = RT_LE2H_U64(pMarker->uSector);
729 if (pcbMarkerData)
730 *pcbMarkerData = RT_ALIGN( cbCompSize
731 + RT_OFFSETOF(VMDKMARKER, uType),
732 512);
733
734#ifdef VMDK_USE_BLOCK_DECOMP_API
735 rc = RTZipBlockDecompress(RTZIPTYPE_ZLIB, 0 /*fFlags*/,
736 pExtent->pvCompGrain, cbCompSize + RT_OFFSETOF(VMDKMARKER, uType), NULL,
737 pvBuf, cbToRead, &cbActuallyRead);
738#else
739 VMDKCOMPRESSIO InflateState;
740 InflateState.pImage = pImage;
741 InflateState.iOffset = -1;
742 InflateState.cbCompGrain = cbCompSize + RT_OFFSETOF(VMDKMARKER, uType);
743 InflateState.pvCompGrain = pExtent->pvCompGrain;
744
745 rc = RTZipDecompCreate(&pZip, &InflateState, vmdkFileInflateHelper);
746 if (RT_FAILURE(rc))
747 return rc;
748 rc = RTZipDecompress(pZip, pvBuf, cbToRead, &cbActuallyRead);
749 RTZipDecompDestroy(pZip);
750#endif /* !VMDK_USE_BLOCK_DECOMP_API */
751 if (RT_FAILURE(rc))
752 {
753 if (rc == VERR_ZIP_CORRUPTED)
754 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: Compressed image is corrupted '%s'"), pExtent->pszFullname);
755 return rc;
756 }
757 if (cbActuallyRead != cbToRead)
758 rc = VERR_VD_VMDK_INVALID_FORMAT;
759 return rc;
760}
761
762static DECLCALLBACK(int) vmdkFileDeflateHelper(void *pvUser, const void *pvBuf, size_t cbBuf)
763{
764 VMDKCOMPRESSIO *pDeflateState = (VMDKCOMPRESSIO *)pvUser;
765
766 Assert(cbBuf);
767 if (pDeflateState->iOffset < 0)
768 {
769 pvBuf = (const uint8_t *)pvBuf + 1;
770 cbBuf--;
771 pDeflateState->iOffset = RT_OFFSETOF(VMDKMARKER, uType);
772 }
773 if (!cbBuf)
774 return VINF_SUCCESS;
775 if (pDeflateState->iOffset + cbBuf > pDeflateState->cbCompGrain)
776 return VERR_BUFFER_OVERFLOW;
777 memcpy((uint8_t *)pDeflateState->pvCompGrain + pDeflateState->iOffset,
778 pvBuf, cbBuf);
779 pDeflateState->iOffset += cbBuf;
780 return VINF_SUCCESS;
781}
782
783/**
784 * Internal: deflate the uncompressed data and write to a file,
785 * distinguishing between async and normal operation
786 */
787DECLINLINE(int) vmdkFileDeflateSync(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
788 uint64_t uOffset, const void *pvBuf,
789 size_t cbToWrite, uint64_t uLBA,
790 uint32_t *pcbMarkerData)
791{
792 int rc;
793 PRTZIPCOMP pZip = NULL;
794 VMDKCOMPRESSIO DeflateState;
795
796 DeflateState.pImage = pImage;
797 DeflateState.iOffset = -1;
798 DeflateState.cbCompGrain = pExtent->cbCompGrain;
799 DeflateState.pvCompGrain = pExtent->pvCompGrain;
800
801 rc = RTZipCompCreate(&pZip, &DeflateState, vmdkFileDeflateHelper,
802 RTZIPTYPE_ZLIB, RTZIPLEVEL_DEFAULT);
803 if (RT_FAILURE(rc))
804 return rc;
805 rc = RTZipCompress(pZip, pvBuf, cbToWrite);
806 if (RT_SUCCESS(rc))
807 rc = RTZipCompFinish(pZip);
808 RTZipCompDestroy(pZip);
809 if (RT_SUCCESS(rc))
810 {
811 Assert( DeflateState.iOffset > 0
812 && (size_t)DeflateState.iOffset <= DeflateState.cbCompGrain);
813
814 /* pad with zeroes to get to a full sector size */
815 uint32_t uSize = DeflateState.iOffset;
816 if (uSize % 512)
817 {
818 uint32_t uSizeAlign = RT_ALIGN(uSize, 512);
819 memset((uint8_t *)pExtent->pvCompGrain + uSize, '\0',
820 uSizeAlign - uSize);
821 uSize = uSizeAlign;
822 }
823
824 if (pcbMarkerData)
825 *pcbMarkerData = uSize;
826
827 /* Compressed grain marker. Data follows immediately. */
828 VMDKMARKER *pMarker = (VMDKMARKER *)pExtent->pvCompGrain;
829 pMarker->uSector = RT_H2LE_U64(uLBA);
830 pMarker->cbSize = RT_H2LE_U32( DeflateState.iOffset
831 - RT_OFFSETOF(VMDKMARKER, uType));
832 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
833 uOffset, pMarker, uSize);
834 if (RT_FAILURE(rc))
835 return rc;
836 }
837 return rc;
838}
839
840
841/**
842 * Internal: check if all files are closed, prevent leaking resources.
843 */
844static int vmdkFileCheckAllClose(PVMDKIMAGE pImage)
845{
846 int rc = VINF_SUCCESS, rc2;
847 PVMDKFILE pVmdkFile;
848
849 Assert(pImage->pFiles == NULL);
850 for (pVmdkFile = pImage->pFiles;
851 pVmdkFile != NULL;
852 pVmdkFile = pVmdkFile->pNext)
853 {
854 LogRel(("VMDK: leaking reference to file \"%s\"\n",
855 pVmdkFile->pszFilename));
856 pImage->pFiles = pVmdkFile->pNext;
857
858 rc2 = vmdkFileClose(pImage, &pVmdkFile, pVmdkFile->fDelete);
859
860 if (RT_SUCCESS(rc))
861 rc = rc2;
862 }
863 return rc;
864}
865
866/**
867 * Internal: truncate a string (at a UTF8 code point boundary) and encode the
868 * critical non-ASCII characters.
869 */
870static char *vmdkEncodeString(const char *psz)
871{
872 char szEnc[VMDK_ENCODED_COMMENT_MAX + 3];
873 char *pszDst = szEnc;
874
875 AssertPtr(psz);
876
877 for (; *psz; psz = RTStrNextCp(psz))
878 {
879 char *pszDstPrev = pszDst;
880 RTUNICP Cp = RTStrGetCp(psz);
881 if (Cp == '\\')
882 {
883 pszDst = RTStrPutCp(pszDst, Cp);
884 pszDst = RTStrPutCp(pszDst, Cp);
885 }
886 else if (Cp == '\n')
887 {
888 pszDst = RTStrPutCp(pszDst, '\\');
889 pszDst = RTStrPutCp(pszDst, 'n');
890 }
891 else if (Cp == '\r')
892 {
893 pszDst = RTStrPutCp(pszDst, '\\');
894 pszDst = RTStrPutCp(pszDst, 'r');
895 }
896 else
897 pszDst = RTStrPutCp(pszDst, Cp);
898 if (pszDst - szEnc >= VMDK_ENCODED_COMMENT_MAX - 1)
899 {
900 pszDst = pszDstPrev;
901 break;
902 }
903 }
904 *pszDst = '\0';
905 return RTStrDup(szEnc);
906}
907
908/**
909 * Internal: decode a string and store it into the specified string.
910 */
911static int vmdkDecodeString(const char *pszEncoded, char *psz, size_t cb)
912{
913 int rc = VINF_SUCCESS;
914 char szBuf[4];
915
916 if (!cb)
917 return VERR_BUFFER_OVERFLOW;
918
919 AssertPtr(psz);
920
921 for (; *pszEncoded; pszEncoded = RTStrNextCp(pszEncoded))
922 {
923 char *pszDst = szBuf;
924 RTUNICP Cp = RTStrGetCp(pszEncoded);
925 if (Cp == '\\')
926 {
927 pszEncoded = RTStrNextCp(pszEncoded);
928 RTUNICP CpQ = RTStrGetCp(pszEncoded);
929 if (CpQ == 'n')
930 RTStrPutCp(pszDst, '\n');
931 else if (CpQ == 'r')
932 RTStrPutCp(pszDst, '\r');
933 else if (CpQ == '\0')
934 {
935 rc = VERR_VD_VMDK_INVALID_HEADER;
936 break;
937 }
938 else
939 RTStrPutCp(pszDst, CpQ);
940 }
941 else
942 pszDst = RTStrPutCp(pszDst, Cp);
943
944 /* Need to leave space for terminating NUL. */
945 if ((size_t)(pszDst - szBuf) + 1 >= cb)
946 {
947 rc = VERR_BUFFER_OVERFLOW;
948 break;
949 }
950 memcpy(psz, szBuf, pszDst - szBuf);
951 psz += pszDst - szBuf;
952 }
953 *psz = '\0';
954 return rc;
955}
956
957/**
958 * Internal: free all buffers associated with grain directories.
959 */
960static void vmdkFreeGrainDirectory(PVMDKEXTENT pExtent)
961{
962 if (pExtent->pGD)
963 {
964 RTMemFree(pExtent->pGD);
965 pExtent->pGD = NULL;
966 }
967 if (pExtent->pRGD)
968 {
969 RTMemFree(pExtent->pRGD);
970 pExtent->pRGD = NULL;
971 }
972}
973
974/**
975 * Internal: allocate the compressed/uncompressed buffers for streamOptimized
976 * images.
977 */
978static int vmdkAllocStreamBuffers(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
979{
980 int rc = VINF_SUCCESS;
981
982 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
983 {
984 /* streamOptimized extents need a compressed grain buffer, which must
985 * be big enough to hold uncompressible data (which needs ~8 bytes
986 * more than the uncompressed data), the marker and padding. */
987 pExtent->cbCompGrain = RT_ALIGN_Z( VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain)
988 + 8 + sizeof(VMDKMARKER), 512);
989 pExtent->pvCompGrain = RTMemAlloc(pExtent->cbCompGrain);
990 if (!pExtent->pvCompGrain)
991 {
992 rc = VERR_NO_MEMORY;
993 goto out;
994 }
995
996 /* streamOptimized extents need a decompressed grain buffer. */
997 pExtent->pvGrain = RTMemAlloc(VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
998 if (!pExtent->pvGrain)
999 {
1000 rc = VERR_NO_MEMORY;
1001 goto out;
1002 }
1003 }
1004
1005out:
1006 if (RT_FAILURE(rc))
1007 vmdkFreeStreamBuffers(pExtent);
1008 return rc;
1009}
1010
1011/**
1012 * Internal: allocate all buffers associated with grain directories.
1013 */
1014static int vmdkAllocGrainDirectory(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
1015{
1016 int rc = VINF_SUCCESS;
1017 size_t cbGD = pExtent->cGDEntries * sizeof(uint32_t);
1018 uint32_t *pGD = NULL, *pRGD = NULL;
1019
1020 pGD = (uint32_t *)RTMemAllocZ(cbGD);
1021 if (!pGD)
1022 {
1023 rc = VERR_NO_MEMORY;
1024 goto out;
1025 }
1026 pExtent->pGD = pGD;
1027
1028 if (pExtent->uSectorRGD)
1029 {
1030 pRGD = (uint32_t *)RTMemAllocZ(cbGD);
1031 if (!pRGD)
1032 {
1033 rc = VERR_NO_MEMORY;
1034 goto out;
1035 }
1036 pExtent->pRGD = pRGD;
1037 }
1038
1039out:
1040 if (RT_FAILURE(rc))
1041 vmdkFreeGrainDirectory(pExtent);
1042 return rc;
1043}
1044
1045static int vmdkReadGrainDirectory(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
1046{
1047 int rc = VINF_SUCCESS;
1048 size_t i;
1049 uint32_t *pGDTmp, *pRGDTmp;
1050 size_t cbGD = pExtent->cGDEntries * sizeof(uint32_t);
1051
1052 if (pExtent->enmType != VMDKETYPE_HOSTED_SPARSE)
1053 goto out;
1054
1055 if ( pExtent->uSectorGD == VMDK_GD_AT_END
1056 || pExtent->uSectorRGD == VMDK_GD_AT_END)
1057 {
1058 rc = VERR_INTERNAL_ERROR;
1059 goto out;
1060 }
1061
1062 rc = vmdkAllocGrainDirectory(pImage, pExtent);
1063 if (RT_FAILURE(rc))
1064 goto out;
1065
1066 /* The VMDK 1.1 spec seems to talk about compressed grain directories,
1067 * but in reality they are not compressed. */
1068 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
1069 VMDK_SECTOR2BYTE(pExtent->uSectorGD),
1070 pExtent->pGD, cbGD);
1071 AssertRC(rc);
1072 if (RT_FAILURE(rc))
1073 {
1074 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not read grain directory in '%s': %Rrc"), pExtent->pszFullname);
1075 goto out;
1076 }
1077 for (i = 0, pGDTmp = pExtent->pGD; i < pExtent->cGDEntries; i++, pGDTmp++)
1078 *pGDTmp = RT_LE2H_U32(*pGDTmp);
1079
1080 if ( pExtent->uSectorRGD
1081 && !(pImage->uOpenFlags & VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS))
1082 {
1083 /* The VMDK 1.1 spec seems to talk about compressed grain directories,
1084 * but in reality they are not compressed. */
1085 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
1086 VMDK_SECTOR2BYTE(pExtent->uSectorRGD),
1087 pExtent->pRGD, cbGD);
1088 AssertRC(rc);
1089 if (RT_FAILURE(rc))
1090 {
1091 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not read redundant grain directory in '%s'"), pExtent->pszFullname);
1092 goto out;
1093 }
1094 for (i = 0, pRGDTmp = pExtent->pRGD; i < pExtent->cGDEntries; i++, pRGDTmp++)
1095 *pRGDTmp = RT_LE2H_U32(*pRGDTmp);
1096
1097 /* Check grain table and redundant grain table for consistency. */
1098 size_t cbGT = pExtent->cGTEntries * sizeof(uint32_t);
1099 size_t cbGTBuffers = cbGT; /* Start with space for one GT. */
1100 size_t cbGTBuffersMax = _1M;
1101
1102 uint32_t *pTmpGT1 = (uint32_t *)RTMemAlloc(cbGTBuffers);
1103 uint32_t *pTmpGT2 = (uint32_t *)RTMemAlloc(cbGTBuffers);
1104
1105 if ( !pTmpGT1
1106 || !pTmpGT2)
1107 rc = VERR_NO_MEMORY;
1108
1109 i = 0;
1110 pGDTmp = pExtent->pGD;
1111 pRGDTmp = pExtent->pRGD;
1112
1113 /* Loop through all entries. */
1114 while (i < pExtent->cGDEntries)
1115 {
1116 uint32_t uGTStart = *pGDTmp;
1117 uint32_t uRGTStart = *pRGDTmp;
1118 size_t cbGTRead = cbGT;
1119
1120 /* If no grain table is allocated skip the entry. */
1121 if (*pGDTmp == 0 && *pRGDTmp == 0)
1122 {
1123 i++;
1124 continue;
1125 }
1126
1127 if (*pGDTmp == 0 || *pRGDTmp == 0 || *pGDTmp == *pRGDTmp)
1128 {
1129 /* Just one grain directory entry refers to a not yet allocated
1130 * grain table or both grain directory copies refer to the same
1131 * grain table. Not allowed. */
1132 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: inconsistent references to grain directory in '%s'"), pExtent->pszFullname);
1133 break;
1134 }
1135
1136 i++;
1137 pGDTmp++;
1138 pRGDTmp++;
1139
1140 /*
1141 * Read a few tables at once if adjacent to decrease the number
1142 * of I/O requests. Read at maximum 1MB at once.
1143 */
1144 while ( i < pExtent->cGDEntries
1145 && cbGTRead < cbGTBuffersMax)
1146 {
1147 /* If no grain table is allocated skip the entry. */
1148 if (*pGDTmp == 0 && *pRGDTmp == 0)
1149 continue;
1150
1151 if (*pGDTmp == 0 || *pRGDTmp == 0 || *pGDTmp == *pRGDTmp)
1152 {
1153 /* Just one grain directory entry refers to a not yet allocated
1154 * grain table or both grain directory copies refer to the same
1155 * grain table. Not allowed. */
1156 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: inconsistent references to grain directory in '%s'"), pExtent->pszFullname);
1157 break;
1158 }
1159
1160 /* Check that the start offsets are adjacent.*/
1161 if ( VMDK_SECTOR2BYTE(uGTStart) + cbGTRead != VMDK_SECTOR2BYTE(*pGDTmp)
1162 || VMDK_SECTOR2BYTE(uRGTStart) + cbGTRead != VMDK_SECTOR2BYTE(*pRGDTmp))
1163 break;
1164
1165 i++;
1166 pGDTmp++;
1167 pRGDTmp++;
1168 cbGTRead += cbGT;
1169 }
1170
1171 /* Increase buffers if required. */
1172 if ( RT_SUCCESS(rc)
1173 && cbGTBuffers < cbGTRead)
1174 {
1175 uint32_t *pTmp;
1176 pTmp = (uint32_t *)RTMemRealloc(pTmpGT1, cbGTRead);
1177 if (pTmp)
1178 {
1179 pTmpGT1 = pTmp;
1180 pTmp = (uint32_t *)RTMemRealloc(pTmpGT2, cbGTRead);
1181 if (pTmp)
1182 pTmpGT2 = pTmp;
1183 else
1184 rc = VERR_NO_MEMORY;
1185 }
1186 else
1187 rc = VERR_NO_MEMORY;
1188
1189 if (rc == VERR_NO_MEMORY)
1190 {
1191 /* Reset to the old values. */
1192 rc = VINF_SUCCESS;
1193 i -= cbGTRead / cbGT;
1194 cbGTRead = cbGT;
1195
1196 /* Don't try to increase the buffer again in the next run. */
1197 cbGTBuffersMax = cbGTBuffers;
1198 }
1199 }
1200
1201 if (RT_SUCCESS(rc))
1202 {
1203 /* The VMDK 1.1 spec seems to talk about compressed grain tables,
1204 * but in reality they are not compressed. */
1205 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
1206 VMDK_SECTOR2BYTE(uGTStart),
1207 pTmpGT1, cbGTRead);
1208 if (RT_FAILURE(rc))
1209 {
1210 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
1211 N_("VMDK: error reading grain table in '%s'"), pExtent->pszFullname);
1212 break;
1213 }
1214 /* The VMDK 1.1 spec seems to talk about compressed grain tables,
1215 * but in reality they are not compressed. */
1216 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
1217 VMDK_SECTOR2BYTE(uRGTStart),
1218 pTmpGT2, cbGTRead);
1219 if (RT_FAILURE(rc))
1220 {
1221 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
1222 N_("VMDK: error reading backup grain table in '%s'"), pExtent->pszFullname);
1223 break;
1224 }
1225 if (memcmp(pTmpGT1, pTmpGT2, cbGTRead))
1226 {
1227 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
1228 N_("VMDK: inconsistency between grain table and backup grain table in '%s'"), pExtent->pszFullname);
1229 break;
1230 }
1231 }
1232 } /* while (i < pExtent->cGDEntries) */
1233
1234 /** @todo figure out what to do for unclean VMDKs. */
1235 if (pTmpGT1)
1236 RTMemFree(pTmpGT1);
1237 if (pTmpGT2)
1238 RTMemFree(pTmpGT2);
1239 }
1240
1241out:
1242 if (RT_FAILURE(rc))
1243 vmdkFreeGrainDirectory(pExtent);
1244 return rc;
1245}
1246
1247static int vmdkCreateGrainDirectory(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
1248 uint64_t uStartSector, bool fPreAlloc)
1249{
1250 int rc = VINF_SUCCESS;
1251 unsigned i;
1252 size_t cbGD = pExtent->cGDEntries * sizeof(uint32_t);
1253 size_t cbGDRounded = RT_ALIGN_64(cbGD, 512);
1254 size_t cbGTRounded;
1255 uint64_t cbOverhead;
1256
1257 if (fPreAlloc)
1258 {
1259 cbGTRounded = RT_ALIGN_64(pExtent->cGDEntries * pExtent->cGTEntries * sizeof(uint32_t), 512);
1260 cbOverhead = VMDK_SECTOR2BYTE(uStartSector) + cbGDRounded
1261 + cbGTRounded;
1262 }
1263 else
1264 {
1265 /* Use a dummy start sector for layout computation. */
1266 if (uStartSector == VMDK_GD_AT_END)
1267 uStartSector = 1;
1268 cbGTRounded = 0;
1269 cbOverhead = VMDK_SECTOR2BYTE(uStartSector) + cbGDRounded;
1270 }
1271
1272 /* For streamOptimized extents there is only one grain directory,
1273 * and for all others take redundant grain directory into account. */
1274 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
1275 {
1276 cbOverhead = RT_ALIGN_64(cbOverhead,
1277 VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
1278 }
1279 else
1280 {
1281 cbOverhead += cbGDRounded + cbGTRounded;
1282 cbOverhead = RT_ALIGN_64(cbOverhead,
1283 VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
1284 rc = vdIfIoIntFileSetSize(pImage->pIfIo, pExtent->pFile->pStorage, cbOverhead);
1285 }
1286 if (RT_FAILURE(rc))
1287 goto out;
1288 pExtent->uAppendPosition = cbOverhead;
1289 pExtent->cOverheadSectors = VMDK_BYTE2SECTOR(cbOverhead);
1290
1291 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
1292 {
1293 pExtent->uSectorRGD = 0;
1294 pExtent->uSectorGD = uStartSector;
1295 }
1296 else
1297 {
1298 pExtent->uSectorRGD = uStartSector;
1299 pExtent->uSectorGD = uStartSector + VMDK_BYTE2SECTOR(cbGDRounded + cbGTRounded);
1300 }
1301
1302 rc = vmdkAllocStreamBuffers(pImage, pExtent);
1303 if (RT_FAILURE(rc))
1304 goto out;
1305
1306 rc = vmdkAllocGrainDirectory(pImage, pExtent);
1307 if (RT_FAILURE(rc))
1308 goto out;
1309
1310 if (fPreAlloc)
1311 {
1312 uint32_t uGTSectorLE;
1313 uint64_t uOffsetSectors;
1314
1315 if (pExtent->pRGD)
1316 {
1317 uOffsetSectors = pExtent->uSectorRGD + VMDK_BYTE2SECTOR(cbGDRounded);
1318 for (i = 0; i < pExtent->cGDEntries; i++)
1319 {
1320 pExtent->pRGD[i] = uOffsetSectors;
1321 uGTSectorLE = RT_H2LE_U64(uOffsetSectors);
1322 /* Write the redundant grain directory entry to disk. */
1323 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
1324 VMDK_SECTOR2BYTE(pExtent->uSectorRGD) + i * sizeof(uGTSectorLE),
1325 &uGTSectorLE, sizeof(uGTSectorLE));
1326 if (RT_FAILURE(rc))
1327 {
1328 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write new redundant grain directory entry in '%s'"), pExtent->pszFullname);
1329 goto out;
1330 }
1331 uOffsetSectors += VMDK_BYTE2SECTOR(pExtent->cGTEntries * sizeof(uint32_t));
1332 }
1333 }
1334
1335 uOffsetSectors = pExtent->uSectorGD + VMDK_BYTE2SECTOR(cbGDRounded);
1336 for (i = 0; i < pExtent->cGDEntries; i++)
1337 {
1338 pExtent->pGD[i] = uOffsetSectors;
1339 uGTSectorLE = RT_H2LE_U64(uOffsetSectors);
1340 /* Write the grain directory entry to disk. */
1341 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
1342 VMDK_SECTOR2BYTE(pExtent->uSectorGD) + i * sizeof(uGTSectorLE),
1343 &uGTSectorLE, sizeof(uGTSectorLE));
1344 if (RT_FAILURE(rc))
1345 {
1346 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write new grain directory entry in '%s'"), pExtent->pszFullname);
1347 goto out;
1348 }
1349 uOffsetSectors += VMDK_BYTE2SECTOR(pExtent->cGTEntries * sizeof(uint32_t));
1350 }
1351 }
1352
1353out:
1354 if (RT_FAILURE(rc))
1355 vmdkFreeGrainDirectory(pExtent);
1356 return rc;
1357}
1358
1359static int vmdkStringUnquote(PVMDKIMAGE pImage, const char *pszStr,
1360 char **ppszUnquoted, char **ppszNext)
1361{
1362 char *pszQ;
1363 char *pszUnquoted;
1364
1365 /* Skip over whitespace. */
1366 while (*pszStr == ' ' || *pszStr == '\t')
1367 pszStr++;
1368
1369 if (*pszStr != '"')
1370 {
1371 pszQ = (char *)pszStr;
1372 while (*pszQ && *pszQ != ' ' && *pszQ != '\t')
1373 pszQ++;
1374 }
1375 else
1376 {
1377 pszStr++;
1378 pszQ = (char *)strchr(pszStr, '"');
1379 if (pszQ == NULL)
1380 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: incorrectly quoted value in descriptor in '%s'"), pImage->pszFilename);
1381 }
1382
1383 pszUnquoted = (char *)RTMemTmpAlloc(pszQ - pszStr + 1);
1384 if (!pszUnquoted)
1385 return VERR_NO_MEMORY;
1386 memcpy(pszUnquoted, pszStr, pszQ - pszStr);
1387 pszUnquoted[pszQ - pszStr] = '\0';
1388 *ppszUnquoted = pszUnquoted;
1389 if (ppszNext)
1390 *ppszNext = pszQ + 1;
1391 return VINF_SUCCESS;
1392}
1393
1394static int vmdkDescInitStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1395 const char *pszLine)
1396{
1397 char *pEnd = pDescriptor->aLines[pDescriptor->cLines];
1398 ssize_t cbDiff = strlen(pszLine) + 1;
1399
1400 if ( pDescriptor->cLines >= VMDK_DESCRIPTOR_LINES_MAX - 1
1401 && pEnd - pDescriptor->aLines[0] > (ptrdiff_t)pDescriptor->cbDescAlloc - cbDiff)
1402 return vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
1403
1404 memcpy(pEnd, pszLine, cbDiff);
1405 pDescriptor->cLines++;
1406 pDescriptor->aLines[pDescriptor->cLines] = pEnd + cbDiff;
1407 pDescriptor->fDirty = true;
1408
1409 return VINF_SUCCESS;
1410}
1411
1412static bool vmdkDescGetStr(PVMDKDESCRIPTOR pDescriptor, unsigned uStart,
1413 const char *pszKey, const char **ppszValue)
1414{
1415 size_t cbKey = strlen(pszKey);
1416 const char *pszValue;
1417
1418 while (uStart != 0)
1419 {
1420 if (!strncmp(pDescriptor->aLines[uStart], pszKey, cbKey))
1421 {
1422 /* Key matches, check for a '=' (preceded by whitespace). */
1423 pszValue = pDescriptor->aLines[uStart] + cbKey;
1424 while (*pszValue == ' ' || *pszValue == '\t')
1425 pszValue++;
1426 if (*pszValue == '=')
1427 {
1428 *ppszValue = pszValue + 1;
1429 break;
1430 }
1431 }
1432 uStart = pDescriptor->aNextLines[uStart];
1433 }
1434 return !!uStart;
1435}
1436
1437static int vmdkDescSetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1438 unsigned uStart,
1439 const char *pszKey, const char *pszValue)
1440{
1441 char *pszTmp;
1442 size_t cbKey = strlen(pszKey);
1443 unsigned uLast = 0;
1444
1445 while (uStart != 0)
1446 {
1447 if (!strncmp(pDescriptor->aLines[uStart], pszKey, cbKey))
1448 {
1449 /* Key matches, check for a '=' (preceded by whitespace). */
1450 pszTmp = pDescriptor->aLines[uStart] + cbKey;
1451 while (*pszTmp == ' ' || *pszTmp == '\t')
1452 pszTmp++;
1453 if (*pszTmp == '=')
1454 {
1455 pszTmp++;
1456 while (*pszTmp == ' ' || *pszTmp == '\t')
1457 pszTmp++;
1458 break;
1459 }
1460 }
1461 if (!pDescriptor->aNextLines[uStart])
1462 uLast = uStart;
1463 uStart = pDescriptor->aNextLines[uStart];
1464 }
1465 if (uStart)
1466 {
1467 if (pszValue)
1468 {
1469 /* Key already exists, replace existing value. */
1470 size_t cbOldVal = strlen(pszTmp);
1471 size_t cbNewVal = strlen(pszValue);
1472 ssize_t cbDiff = cbNewVal - cbOldVal;
1473 /* Check for buffer overflow. */
1474 if ( pDescriptor->aLines[pDescriptor->cLines]
1475 - pDescriptor->aLines[0] > (ptrdiff_t)pDescriptor->cbDescAlloc - cbDiff)
1476 return vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
1477
1478 memmove(pszTmp + cbNewVal, pszTmp + cbOldVal,
1479 pDescriptor->aLines[pDescriptor->cLines] - pszTmp - cbOldVal);
1480 memcpy(pszTmp, pszValue, cbNewVal + 1);
1481 for (unsigned i = uStart + 1; i <= pDescriptor->cLines; i++)
1482 pDescriptor->aLines[i] += cbDiff;
1483 }
1484 else
1485 {
1486 memmove(pDescriptor->aLines[uStart], pDescriptor->aLines[uStart+1],
1487 pDescriptor->aLines[pDescriptor->cLines] - pDescriptor->aLines[uStart+1] + 1);
1488 for (unsigned i = uStart + 1; i <= pDescriptor->cLines; i++)
1489 {
1490 pDescriptor->aLines[i-1] = pDescriptor->aLines[i];
1491 if (pDescriptor->aNextLines[i])
1492 pDescriptor->aNextLines[i-1] = pDescriptor->aNextLines[i] - 1;
1493 else
1494 pDescriptor->aNextLines[i-1] = 0;
1495 }
1496 pDescriptor->cLines--;
1497 /* Adjust starting line numbers of following descriptor sections. */
1498 if (uStart < pDescriptor->uFirstExtent)
1499 pDescriptor->uFirstExtent--;
1500 if (uStart < pDescriptor->uFirstDDB)
1501 pDescriptor->uFirstDDB--;
1502 }
1503 }
1504 else
1505 {
1506 /* Key doesn't exist, append after the last entry in this category. */
1507 if (!pszValue)
1508 {
1509 /* Key doesn't exist, and it should be removed. Simply a no-op. */
1510 return VINF_SUCCESS;
1511 }
1512 cbKey = strlen(pszKey);
1513 size_t cbValue = strlen(pszValue);
1514 ssize_t cbDiff = cbKey + 1 + cbValue + 1;
1515 /* Check for buffer overflow. */
1516 if ( (pDescriptor->cLines >= VMDK_DESCRIPTOR_LINES_MAX - 1)
1517 || ( pDescriptor->aLines[pDescriptor->cLines]
1518 - pDescriptor->aLines[0] > (ptrdiff_t)pDescriptor->cbDescAlloc - cbDiff))
1519 return vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
1520 for (unsigned i = pDescriptor->cLines + 1; i > uLast + 1; i--)
1521 {
1522 pDescriptor->aLines[i] = pDescriptor->aLines[i - 1];
1523 if (pDescriptor->aNextLines[i - 1])
1524 pDescriptor->aNextLines[i] = pDescriptor->aNextLines[i - 1] + 1;
1525 else
1526 pDescriptor->aNextLines[i] = 0;
1527 }
1528 uStart = uLast + 1;
1529 pDescriptor->aNextLines[uLast] = uStart;
1530 pDescriptor->aNextLines[uStart] = 0;
1531 pDescriptor->cLines++;
1532 pszTmp = pDescriptor->aLines[uStart];
1533 memmove(pszTmp + cbDiff, pszTmp,
1534 pDescriptor->aLines[pDescriptor->cLines] - pszTmp);
1535 memcpy(pDescriptor->aLines[uStart], pszKey, cbKey);
1536 pDescriptor->aLines[uStart][cbKey] = '=';
1537 memcpy(pDescriptor->aLines[uStart] + cbKey + 1, pszValue, cbValue + 1);
1538 for (unsigned i = uStart + 1; i <= pDescriptor->cLines; i++)
1539 pDescriptor->aLines[i] += cbDiff;
1540
1541 /* Adjust starting line numbers of following descriptor sections. */
1542 if (uStart <= pDescriptor->uFirstExtent)
1543 pDescriptor->uFirstExtent++;
1544 if (uStart <= pDescriptor->uFirstDDB)
1545 pDescriptor->uFirstDDB++;
1546 }
1547 pDescriptor->fDirty = true;
1548 return VINF_SUCCESS;
1549}
1550
1551static int vmdkDescBaseGetU32(PVMDKDESCRIPTOR pDescriptor, const char *pszKey,
1552 uint32_t *puValue)
1553{
1554 const char *pszValue;
1555
1556 if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDesc, pszKey,
1557 &pszValue))
1558 return VERR_VD_VMDK_VALUE_NOT_FOUND;
1559 return RTStrToUInt32Ex(pszValue, NULL, 10, puValue);
1560}
1561
1562static int vmdkDescBaseGetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1563 const char *pszKey, const char **ppszValue)
1564{
1565 const char *pszValue;
1566 char *pszValueUnquoted;
1567
1568 if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDesc, pszKey,
1569 &pszValue))
1570 return VERR_VD_VMDK_VALUE_NOT_FOUND;
1571 int rc = vmdkStringUnquote(pImage, pszValue, &pszValueUnquoted, NULL);
1572 if (RT_FAILURE(rc))
1573 return rc;
1574 *ppszValue = pszValueUnquoted;
1575 return rc;
1576}
1577
1578static int vmdkDescBaseSetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1579 const char *pszKey, const char *pszValue)
1580{
1581 char *pszValueQuoted;
1582
1583 RTStrAPrintf(&pszValueQuoted, "\"%s\"", pszValue);
1584 if (!pszValueQuoted)
1585 return VERR_NO_STR_MEMORY;
1586 int rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDesc, pszKey,
1587 pszValueQuoted);
1588 RTStrFree(pszValueQuoted);
1589 return rc;
1590}
1591
1592static void vmdkDescExtRemoveDummy(PVMDKIMAGE pImage,
1593 PVMDKDESCRIPTOR pDescriptor)
1594{
1595 unsigned uEntry = pDescriptor->uFirstExtent;
1596 ssize_t cbDiff;
1597
1598 if (!uEntry)
1599 return;
1600
1601 cbDiff = strlen(pDescriptor->aLines[uEntry]) + 1;
1602 /* Move everything including \0 in the entry marking the end of buffer. */
1603 memmove(pDescriptor->aLines[uEntry], pDescriptor->aLines[uEntry + 1],
1604 pDescriptor->aLines[pDescriptor->cLines] - pDescriptor->aLines[uEntry + 1] + 1);
1605 for (unsigned i = uEntry + 1; i <= pDescriptor->cLines; i++)
1606 {
1607 pDescriptor->aLines[i - 1] = pDescriptor->aLines[i] - cbDiff;
1608 if (pDescriptor->aNextLines[i])
1609 pDescriptor->aNextLines[i - 1] = pDescriptor->aNextLines[i] - 1;
1610 else
1611 pDescriptor->aNextLines[i - 1] = 0;
1612 }
1613 pDescriptor->cLines--;
1614 if (pDescriptor->uFirstDDB)
1615 pDescriptor->uFirstDDB--;
1616
1617 return;
1618}
1619
1620static int vmdkDescExtInsert(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1621 VMDKACCESS enmAccess, uint64_t cNominalSectors,
1622 VMDKETYPE enmType, const char *pszBasename,
1623 uint64_t uSectorOffset)
1624{
1625 static const char *apszAccess[] = { "NOACCESS", "RDONLY", "RW" };
1626 static const char *apszType[] = { "", "SPARSE", "FLAT", "ZERO", "VMFS" };
1627 char *pszTmp;
1628 unsigned uStart = pDescriptor->uFirstExtent, uLast = 0;
1629 char szExt[1024];
1630 ssize_t cbDiff;
1631
1632 Assert((unsigned)enmAccess < RT_ELEMENTS(apszAccess));
1633 Assert((unsigned)enmType < RT_ELEMENTS(apszType));
1634
1635 /* Find last entry in extent description. */
1636 while (uStart)
1637 {
1638 if (!pDescriptor->aNextLines[uStart])
1639 uLast = uStart;
1640 uStart = pDescriptor->aNextLines[uStart];
1641 }
1642
1643 if (enmType == VMDKETYPE_ZERO)
1644 {
1645 RTStrPrintf(szExt, sizeof(szExt), "%s %llu %s ", apszAccess[enmAccess],
1646 cNominalSectors, apszType[enmType]);
1647 }
1648 else if (enmType == VMDKETYPE_FLAT)
1649 {
1650 RTStrPrintf(szExt, sizeof(szExt), "%s %llu %s \"%s\" %llu",
1651 apszAccess[enmAccess], cNominalSectors,
1652 apszType[enmType], pszBasename, uSectorOffset);
1653 }
1654 else
1655 {
1656 RTStrPrintf(szExt, sizeof(szExt), "%s %llu %s \"%s\"",
1657 apszAccess[enmAccess], cNominalSectors,
1658 apszType[enmType], pszBasename);
1659 }
1660 cbDiff = strlen(szExt) + 1;
1661
1662 /* Check for buffer overflow. */
1663 if ( (pDescriptor->cLines >= VMDK_DESCRIPTOR_LINES_MAX - 1)
1664 || ( pDescriptor->aLines[pDescriptor->cLines]
1665 - pDescriptor->aLines[0] > (ptrdiff_t)pDescriptor->cbDescAlloc - cbDiff))
1666 return vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
1667
1668 for (unsigned i = pDescriptor->cLines + 1; i > uLast + 1; i--)
1669 {
1670 pDescriptor->aLines[i] = pDescriptor->aLines[i - 1];
1671 if (pDescriptor->aNextLines[i - 1])
1672 pDescriptor->aNextLines[i] = pDescriptor->aNextLines[i - 1] + 1;
1673 else
1674 pDescriptor->aNextLines[i] = 0;
1675 }
1676 uStart = uLast + 1;
1677 pDescriptor->aNextLines[uLast] = uStart;
1678 pDescriptor->aNextLines[uStart] = 0;
1679 pDescriptor->cLines++;
1680 pszTmp = pDescriptor->aLines[uStart];
1681 memmove(pszTmp + cbDiff, pszTmp,
1682 pDescriptor->aLines[pDescriptor->cLines] - pszTmp);
1683 memcpy(pDescriptor->aLines[uStart], szExt, cbDiff);
1684 for (unsigned i = uStart + 1; i <= pDescriptor->cLines; i++)
1685 pDescriptor->aLines[i] += cbDiff;
1686
1687 /* Adjust starting line numbers of following descriptor sections. */
1688 if (uStart <= pDescriptor->uFirstDDB)
1689 pDescriptor->uFirstDDB++;
1690
1691 pDescriptor->fDirty = true;
1692 return VINF_SUCCESS;
1693}
1694
1695static int vmdkDescDDBGetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1696 const char *pszKey, const char **ppszValue)
1697{
1698 const char *pszValue;
1699 char *pszValueUnquoted;
1700
1701 if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDDB, pszKey,
1702 &pszValue))
1703 return VERR_VD_VMDK_VALUE_NOT_FOUND;
1704 int rc = vmdkStringUnquote(pImage, pszValue, &pszValueUnquoted, NULL);
1705 if (RT_FAILURE(rc))
1706 return rc;
1707 *ppszValue = pszValueUnquoted;
1708 return rc;
1709}
1710
1711static int vmdkDescDDBGetU32(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1712 const char *pszKey, uint32_t *puValue)
1713{
1714 const char *pszValue;
1715 char *pszValueUnquoted;
1716
1717 if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDDB, pszKey,
1718 &pszValue))
1719 return VERR_VD_VMDK_VALUE_NOT_FOUND;
1720 int rc = vmdkStringUnquote(pImage, pszValue, &pszValueUnquoted, NULL);
1721 if (RT_FAILURE(rc))
1722 return rc;
1723 rc = RTStrToUInt32Ex(pszValueUnquoted, NULL, 10, puValue);
1724 RTMemTmpFree(pszValueUnquoted);
1725 return rc;
1726}
1727
1728static int vmdkDescDDBGetUuid(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1729 const char *pszKey, PRTUUID pUuid)
1730{
1731 const char *pszValue;
1732 char *pszValueUnquoted;
1733
1734 if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDDB, pszKey,
1735 &pszValue))
1736 return VERR_VD_VMDK_VALUE_NOT_FOUND;
1737 int rc = vmdkStringUnquote(pImage, pszValue, &pszValueUnquoted, NULL);
1738 if (RT_FAILURE(rc))
1739 return rc;
1740 rc = RTUuidFromStr(pUuid, pszValueUnquoted);
1741 RTMemTmpFree(pszValueUnquoted);
1742 return rc;
1743}
1744
1745static int vmdkDescDDBSetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1746 const char *pszKey, const char *pszVal)
1747{
1748 int rc;
1749 char *pszValQuoted;
1750
1751 if (pszVal)
1752 {
1753 RTStrAPrintf(&pszValQuoted, "\"%s\"", pszVal);
1754 if (!pszValQuoted)
1755 return VERR_NO_STR_MEMORY;
1756 }
1757 else
1758 pszValQuoted = NULL;
1759 rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDDB, pszKey,
1760 pszValQuoted);
1761 if (pszValQuoted)
1762 RTStrFree(pszValQuoted);
1763 return rc;
1764}
1765
1766static int vmdkDescDDBSetUuid(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1767 const char *pszKey, PCRTUUID pUuid)
1768{
1769 char *pszUuid;
1770
1771 RTStrAPrintf(&pszUuid, "\"%RTuuid\"", pUuid);
1772 if (!pszUuid)
1773 return VERR_NO_STR_MEMORY;
1774 int rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDDB, pszKey,
1775 pszUuid);
1776 RTStrFree(pszUuid);
1777 return rc;
1778}
1779
1780static int vmdkDescDDBSetU32(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
1781 const char *pszKey, uint32_t uValue)
1782{
1783 char *pszValue;
1784
1785 RTStrAPrintf(&pszValue, "\"%d\"", uValue);
1786 if (!pszValue)
1787 return VERR_NO_STR_MEMORY;
1788 int rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDDB, pszKey,
1789 pszValue);
1790 RTStrFree(pszValue);
1791 return rc;
1792}
1793
1794static int vmdkPreprocessDescriptor(PVMDKIMAGE pImage, char *pDescData,
1795 size_t cbDescData,
1796 PVMDKDESCRIPTOR pDescriptor)
1797{
1798 int rc = VINF_SUCCESS;
1799 unsigned cLine = 0, uLastNonEmptyLine = 0;
1800 char *pTmp = pDescData;
1801
1802 pDescriptor->cbDescAlloc = cbDescData;
1803 while (*pTmp != '\0')
1804 {
1805 pDescriptor->aLines[cLine++] = pTmp;
1806 if (cLine >= VMDK_DESCRIPTOR_LINES_MAX)
1807 {
1808 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
1809 goto out;
1810 }
1811
1812 while (*pTmp != '\0' && *pTmp != '\n')
1813 {
1814 if (*pTmp == '\r')
1815 {
1816 if (*(pTmp + 1) != '\n')
1817 {
1818 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: unsupported end of line in descriptor in '%s'"), pImage->pszFilename);
1819 goto out;
1820 }
1821 else
1822 {
1823 /* Get rid of CR character. */
1824 *pTmp = '\0';
1825 }
1826 }
1827 pTmp++;
1828 }
1829 /* Get rid of LF character. */
1830 if (*pTmp == '\n')
1831 {
1832 *pTmp = '\0';
1833 pTmp++;
1834 }
1835 }
1836 pDescriptor->cLines = cLine;
1837 /* Pointer right after the end of the used part of the buffer. */
1838 pDescriptor->aLines[cLine] = pTmp;
1839
1840 if ( strcmp(pDescriptor->aLines[0], "# Disk DescriptorFile")
1841 && strcmp(pDescriptor->aLines[0], "# Disk Descriptor File"))
1842 {
1843 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: descriptor does not start as expected in '%s'"), pImage->pszFilename);
1844 goto out;
1845 }
1846
1847 /* Initialize those, because we need to be able to reopen an image. */
1848 pDescriptor->uFirstDesc = 0;
1849 pDescriptor->uFirstExtent = 0;
1850 pDescriptor->uFirstDDB = 0;
1851 for (unsigned i = 0; i < cLine; i++)
1852 {
1853 if (*pDescriptor->aLines[i] != '#' && *pDescriptor->aLines[i] != '\0')
1854 {
1855 if ( !strncmp(pDescriptor->aLines[i], "RW", 2)
1856 || !strncmp(pDescriptor->aLines[i], "RDONLY", 6)
1857 || !strncmp(pDescriptor->aLines[i], "NOACCESS", 8) )
1858 {
1859 /* An extent descriptor. */
1860 if (!pDescriptor->uFirstDesc || pDescriptor->uFirstDDB)
1861 {
1862 /* Incorrect ordering of entries. */
1863 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: incorrect ordering of entries in descriptor in '%s'"), pImage->pszFilename);
1864 goto out;
1865 }
1866 if (!pDescriptor->uFirstExtent)
1867 {
1868 pDescriptor->uFirstExtent = i;
1869 uLastNonEmptyLine = 0;
1870 }
1871 }
1872 else if (!strncmp(pDescriptor->aLines[i], "ddb.", 4))
1873 {
1874 /* A disk database entry. */
1875 if (!pDescriptor->uFirstDesc || !pDescriptor->uFirstExtent)
1876 {
1877 /* Incorrect ordering of entries. */
1878 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: incorrect ordering of entries in descriptor in '%s'"), pImage->pszFilename);
1879 goto out;
1880 }
1881 if (!pDescriptor->uFirstDDB)
1882 {
1883 pDescriptor->uFirstDDB = i;
1884 uLastNonEmptyLine = 0;
1885 }
1886 }
1887 else
1888 {
1889 /* A normal entry. */
1890 if (pDescriptor->uFirstExtent || pDescriptor->uFirstDDB)
1891 {
1892 /* Incorrect ordering of entries. */
1893 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: incorrect ordering of entries in descriptor in '%s'"), pImage->pszFilename);
1894 goto out;
1895 }
1896 if (!pDescriptor->uFirstDesc)
1897 {
1898 pDescriptor->uFirstDesc = i;
1899 uLastNonEmptyLine = 0;
1900 }
1901 }
1902 if (uLastNonEmptyLine)
1903 pDescriptor->aNextLines[uLastNonEmptyLine] = i;
1904 uLastNonEmptyLine = i;
1905 }
1906 }
1907
1908out:
1909 return rc;
1910}
1911
1912static int vmdkDescSetPCHSGeometry(PVMDKIMAGE pImage,
1913 PCVDGEOMETRY pPCHSGeometry)
1914{
1915 int rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
1916 VMDK_DDB_GEO_PCHS_CYLINDERS,
1917 pPCHSGeometry->cCylinders);
1918 if (RT_FAILURE(rc))
1919 return rc;
1920 rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
1921 VMDK_DDB_GEO_PCHS_HEADS,
1922 pPCHSGeometry->cHeads);
1923 if (RT_FAILURE(rc))
1924 return rc;
1925 rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
1926 VMDK_DDB_GEO_PCHS_SECTORS,
1927 pPCHSGeometry->cSectors);
1928 return rc;
1929}
1930
1931static int vmdkDescSetLCHSGeometry(PVMDKIMAGE pImage,
1932 PCVDGEOMETRY pLCHSGeometry)
1933{
1934 int rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
1935 VMDK_DDB_GEO_LCHS_CYLINDERS,
1936 pLCHSGeometry->cCylinders);
1937 if (RT_FAILURE(rc))
1938 return rc;
1939 rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
1940 VMDK_DDB_GEO_LCHS_HEADS,
1941
1942 pLCHSGeometry->cHeads);
1943 if (RT_FAILURE(rc))
1944 return rc;
1945 rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
1946 VMDK_DDB_GEO_LCHS_SECTORS,
1947 pLCHSGeometry->cSectors);
1948 return rc;
1949}
1950
1951static int vmdkCreateDescriptor(PVMDKIMAGE pImage, char *pDescData,
1952 size_t cbDescData, PVMDKDESCRIPTOR pDescriptor)
1953{
1954 int rc;
1955
1956 pDescriptor->uFirstDesc = 0;
1957 pDescriptor->uFirstExtent = 0;
1958 pDescriptor->uFirstDDB = 0;
1959 pDescriptor->cLines = 0;
1960 pDescriptor->cbDescAlloc = cbDescData;
1961 pDescriptor->fDirty = false;
1962 pDescriptor->aLines[pDescriptor->cLines] = pDescData;
1963 memset(pDescriptor->aNextLines, '\0', sizeof(pDescriptor->aNextLines));
1964
1965 rc = vmdkDescInitStr(pImage, pDescriptor, "# Disk DescriptorFile");
1966 if (RT_FAILURE(rc))
1967 goto out;
1968 rc = vmdkDescInitStr(pImage, pDescriptor, "version=1");
1969 if (RT_FAILURE(rc))
1970 goto out;
1971 pDescriptor->uFirstDesc = pDescriptor->cLines - 1;
1972 rc = vmdkDescInitStr(pImage, pDescriptor, "");
1973 if (RT_FAILURE(rc))
1974 goto out;
1975 rc = vmdkDescInitStr(pImage, pDescriptor, "# Extent description");
1976 if (RT_FAILURE(rc))
1977 goto out;
1978 rc = vmdkDescInitStr(pImage, pDescriptor, "NOACCESS 0 ZERO ");
1979 if (RT_FAILURE(rc))
1980 goto out;
1981 pDescriptor->uFirstExtent = pDescriptor->cLines - 1;
1982 rc = vmdkDescInitStr(pImage, pDescriptor, "");
1983 if (RT_FAILURE(rc))
1984 goto out;
1985 /* The trailing space is created by VMware, too. */
1986 rc = vmdkDescInitStr(pImage, pDescriptor, "# The disk Data Base ");
1987 if (RT_FAILURE(rc))
1988 goto out;
1989 rc = vmdkDescInitStr(pImage, pDescriptor, "#DDB");
1990 if (RT_FAILURE(rc))
1991 goto out;
1992 rc = vmdkDescInitStr(pImage, pDescriptor, "");
1993 if (RT_FAILURE(rc))
1994 goto out;
1995 rc = vmdkDescInitStr(pImage, pDescriptor, "ddb.virtualHWVersion = \"4\"");
1996 if (RT_FAILURE(rc))
1997 goto out;
1998 pDescriptor->uFirstDDB = pDescriptor->cLines - 1;
1999
2000 /* Now that the framework is in place, use the normal functions to insert
2001 * the remaining keys. */
2002 char szBuf[9];
2003 RTStrPrintf(szBuf, sizeof(szBuf), "%08x", RTRandU32());
2004 rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDesc,
2005 "CID", szBuf);
2006 if (RT_FAILURE(rc))
2007 goto out;
2008 rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDesc,
2009 "parentCID", "ffffffff");
2010 if (RT_FAILURE(rc))
2011 goto out;
2012
2013 rc = vmdkDescDDBSetStr(pImage, pDescriptor, "ddb.adapterType", "ide");
2014 if (RT_FAILURE(rc))
2015 goto out;
2016
2017out:
2018 return rc;
2019}
2020
2021static int vmdkParseDescriptor(PVMDKIMAGE pImage, char *pDescData,
2022 size_t cbDescData)
2023{
2024 int rc;
2025 unsigned cExtents;
2026 unsigned uLine;
2027 unsigned i;
2028
2029 rc = vmdkPreprocessDescriptor(pImage, pDescData, cbDescData,
2030 &pImage->Descriptor);
2031 if (RT_FAILURE(rc))
2032 return rc;
2033
2034 /* Check version, must be 1. */
2035 uint32_t uVersion;
2036 rc = vmdkDescBaseGetU32(&pImage->Descriptor, "version", &uVersion);
2037 if (RT_FAILURE(rc))
2038 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error finding key 'version' in descriptor in '%s'"), pImage->pszFilename);
2039 if (uVersion != 1)
2040 return vdIfError(pImage->pIfError, VERR_VD_VMDK_UNSUPPORTED_VERSION, RT_SRC_POS, N_("VMDK: unsupported format version in descriptor in '%s'"), pImage->pszFilename);
2041
2042 /* Get image creation type and determine image flags. */
2043 const char *pszCreateType = NULL; /* initialized to make gcc shut up */
2044 rc = vmdkDescBaseGetStr(pImage, &pImage->Descriptor, "createType",
2045 &pszCreateType);
2046 if (RT_FAILURE(rc))
2047 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot get image type from descriptor in '%s'"), pImage->pszFilename);
2048 if ( !strcmp(pszCreateType, "twoGbMaxExtentSparse")
2049 || !strcmp(pszCreateType, "twoGbMaxExtentFlat"))
2050 pImage->uImageFlags |= VD_VMDK_IMAGE_FLAGS_SPLIT_2G;
2051 else if ( !strcmp(pszCreateType, "partitionedDevice")
2052 || !strcmp(pszCreateType, "fullDevice"))
2053 pImage->uImageFlags |= VD_VMDK_IMAGE_FLAGS_RAWDISK;
2054 else if (!strcmp(pszCreateType, "streamOptimized"))
2055 pImage->uImageFlags |= VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED;
2056 else if (!strcmp(pszCreateType, "vmfs"))
2057 pImage->uImageFlags |= VD_IMAGE_FLAGS_FIXED | VD_VMDK_IMAGE_FLAGS_ESX;
2058 RTStrFree((char *)(void *)pszCreateType);
2059
2060 /* Count the number of extent config entries. */
2061 for (uLine = pImage->Descriptor.uFirstExtent, cExtents = 0;
2062 uLine != 0;
2063 uLine = pImage->Descriptor.aNextLines[uLine], cExtents++)
2064 /* nothing */;
2065
2066 if (!pImage->pDescData && cExtents != 1)
2067 {
2068 /* Monolithic image, must have only one extent (already opened). */
2069 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: monolithic image may only have one extent in '%s'"), pImage->pszFilename);
2070 }
2071
2072 if (pImage->pDescData)
2073 {
2074 /* Non-monolithic image, extents need to be allocated. */
2075 rc = vmdkCreateExtents(pImage, cExtents);
2076 if (RT_FAILURE(rc))
2077 return rc;
2078 }
2079
2080 for (i = 0, uLine = pImage->Descriptor.uFirstExtent;
2081 i < cExtents; i++, uLine = pImage->Descriptor.aNextLines[uLine])
2082 {
2083 char *pszLine = pImage->Descriptor.aLines[uLine];
2084
2085 /* Access type of the extent. */
2086 if (!strncmp(pszLine, "RW", 2))
2087 {
2088 pImage->pExtents[i].enmAccess = VMDKACCESS_READWRITE;
2089 pszLine += 2;
2090 }
2091 else if (!strncmp(pszLine, "RDONLY", 6))
2092 {
2093 pImage->pExtents[i].enmAccess = VMDKACCESS_READONLY;
2094 pszLine += 6;
2095 }
2096 else if (!strncmp(pszLine, "NOACCESS", 8))
2097 {
2098 pImage->pExtents[i].enmAccess = VMDKACCESS_NOACCESS;
2099 pszLine += 8;
2100 }
2101 else
2102 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
2103 if (*pszLine++ != ' ')
2104 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
2105
2106 /* Nominal size of the extent. */
2107 rc = RTStrToUInt64Ex(pszLine, &pszLine, 10,
2108 &pImage->pExtents[i].cNominalSectors);
2109 if (RT_FAILURE(rc))
2110 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
2111 if (*pszLine++ != ' ')
2112 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
2113
2114 /* Type of the extent. */
2115#ifdef VBOX_WITH_VMDK_ESX
2116 /** @todo Add the ESX extent types. Not necessary for now because
2117 * the ESX extent types are only used inside an ESX server. They are
2118 * automatically converted if the VMDK is exported. */
2119#endif /* VBOX_WITH_VMDK_ESX */
2120 if (!strncmp(pszLine, "SPARSE", 6))
2121 {
2122 pImage->pExtents[i].enmType = VMDKETYPE_HOSTED_SPARSE;
2123 pszLine += 6;
2124 }
2125 else if (!strncmp(pszLine, "FLAT", 4))
2126 {
2127 pImage->pExtents[i].enmType = VMDKETYPE_FLAT;
2128 pszLine += 4;
2129 }
2130 else if (!strncmp(pszLine, "ZERO", 4))
2131 {
2132 pImage->pExtents[i].enmType = VMDKETYPE_ZERO;
2133 pszLine += 4;
2134 }
2135 else if (!strncmp(pszLine, "VMFS", 4))
2136 {
2137 pImage->pExtents[i].enmType = VMDKETYPE_VMFS;
2138 pszLine += 4;
2139 }
2140 else
2141 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
2142
2143 if (pImage->pExtents[i].enmType == VMDKETYPE_ZERO)
2144 {
2145 /* This one has no basename or offset. */
2146 if (*pszLine == ' ')
2147 pszLine++;
2148 if (*pszLine != '\0')
2149 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
2150 pImage->pExtents[i].pszBasename = NULL;
2151 }
2152 else
2153 {
2154 /* All other extent types have basename and optional offset. */
2155 if (*pszLine++ != ' ')
2156 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
2157
2158 /* Basename of the image. Surrounded by quotes. */
2159 char *pszBasename;
2160 rc = vmdkStringUnquote(pImage, pszLine, &pszBasename, &pszLine);
2161 if (RT_FAILURE(rc))
2162 return rc;
2163 pImage->pExtents[i].pszBasename = pszBasename;
2164 if (*pszLine == ' ')
2165 {
2166 pszLine++;
2167 if (*pszLine != '\0')
2168 {
2169 /* Optional offset in extent specified. */
2170 rc = RTStrToUInt64Ex(pszLine, &pszLine, 10,
2171 &pImage->pExtents[i].uSectorOffset);
2172 if (RT_FAILURE(rc))
2173 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
2174 }
2175 }
2176
2177 if (*pszLine != '\0')
2178 return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
2179 }
2180 }
2181
2182 /* Determine PCHS geometry (autogenerate if necessary). */
2183 rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
2184 VMDK_DDB_GEO_PCHS_CYLINDERS,
2185 &pImage->PCHSGeometry.cCylinders);
2186 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2187 pImage->PCHSGeometry.cCylinders = 0;
2188 else if (RT_FAILURE(rc))
2189 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting PCHS geometry from extent description in '%s'"), pImage->pszFilename);
2190 rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
2191 VMDK_DDB_GEO_PCHS_HEADS,
2192 &pImage->PCHSGeometry.cHeads);
2193 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2194 pImage->PCHSGeometry.cHeads = 0;
2195 else if (RT_FAILURE(rc))
2196 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting PCHS geometry from extent description in '%s'"), pImage->pszFilename);
2197 rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
2198 VMDK_DDB_GEO_PCHS_SECTORS,
2199 &pImage->PCHSGeometry.cSectors);
2200 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2201 pImage->PCHSGeometry.cSectors = 0;
2202 else if (RT_FAILURE(rc))
2203 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting PCHS geometry from extent description in '%s'"), pImage->pszFilename);
2204 if ( pImage->PCHSGeometry.cCylinders == 0
2205 || pImage->PCHSGeometry.cHeads == 0
2206 || pImage->PCHSGeometry.cHeads > 16
2207 || pImage->PCHSGeometry.cSectors == 0
2208 || pImage->PCHSGeometry.cSectors > 63)
2209 {
2210 /* Mark PCHS geometry as not yet valid (can't do the calculation here
2211 * as the total image size isn't known yet). */
2212 pImage->PCHSGeometry.cCylinders = 0;
2213 pImage->PCHSGeometry.cHeads = 16;
2214 pImage->PCHSGeometry.cSectors = 63;
2215 }
2216
2217 /* Determine LCHS geometry (set to 0 if not specified). */
2218 rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
2219 VMDK_DDB_GEO_LCHS_CYLINDERS,
2220 &pImage->LCHSGeometry.cCylinders);
2221 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2222 pImage->LCHSGeometry.cCylinders = 0;
2223 else if (RT_FAILURE(rc))
2224 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting LCHS geometry from extent description in '%s'"), pImage->pszFilename);
2225 rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
2226 VMDK_DDB_GEO_LCHS_HEADS,
2227 &pImage->LCHSGeometry.cHeads);
2228 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2229 pImage->LCHSGeometry.cHeads = 0;
2230 else if (RT_FAILURE(rc))
2231 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting LCHS geometry from extent description in '%s'"), pImage->pszFilename);
2232 rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
2233 VMDK_DDB_GEO_LCHS_SECTORS,
2234 &pImage->LCHSGeometry.cSectors);
2235 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2236 pImage->LCHSGeometry.cSectors = 0;
2237 else if (RT_FAILURE(rc))
2238 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting LCHS geometry from extent description in '%s'"), pImage->pszFilename);
2239 if ( pImage->LCHSGeometry.cCylinders == 0
2240 || pImage->LCHSGeometry.cHeads == 0
2241 || pImage->LCHSGeometry.cSectors == 0)
2242 {
2243 pImage->LCHSGeometry.cCylinders = 0;
2244 pImage->LCHSGeometry.cHeads = 0;
2245 pImage->LCHSGeometry.cSectors = 0;
2246 }
2247
2248 /* Get image UUID. */
2249 rc = vmdkDescDDBGetUuid(pImage, &pImage->Descriptor, VMDK_DDB_IMAGE_UUID,
2250 &pImage->ImageUuid);
2251 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2252 {
2253 /* Image without UUID. Probably created by VMware and not yet used
2254 * by VirtualBox. Can only be added for images opened in read/write
2255 * mode, so don't bother producing a sensible UUID otherwise. */
2256 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2257 RTUuidClear(&pImage->ImageUuid);
2258 else
2259 {
2260 rc = RTUuidCreate(&pImage->ImageUuid);
2261 if (RT_FAILURE(rc))
2262 return rc;
2263 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
2264 VMDK_DDB_IMAGE_UUID, &pImage->ImageUuid);
2265 if (RT_FAILURE(rc))
2266 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing image UUID in descriptor in '%s'"), pImage->pszFilename);
2267 }
2268 }
2269 else if (RT_FAILURE(rc))
2270 return rc;
2271
2272 /* Get image modification UUID. */
2273 rc = vmdkDescDDBGetUuid(pImage, &pImage->Descriptor,
2274 VMDK_DDB_MODIFICATION_UUID,
2275 &pImage->ModificationUuid);
2276 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2277 {
2278 /* Image without UUID. Probably created by VMware and not yet used
2279 * by VirtualBox. Can only be added for images opened in read/write
2280 * mode, so don't bother producing a sensible UUID otherwise. */
2281 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2282 RTUuidClear(&pImage->ModificationUuid);
2283 else
2284 {
2285 rc = RTUuidCreate(&pImage->ModificationUuid);
2286 if (RT_FAILURE(rc))
2287 return rc;
2288 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
2289 VMDK_DDB_MODIFICATION_UUID,
2290 &pImage->ModificationUuid);
2291 if (RT_FAILURE(rc))
2292 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing image modification UUID in descriptor in '%s'"), pImage->pszFilename);
2293 }
2294 }
2295 else if (RT_FAILURE(rc))
2296 return rc;
2297
2298 /* Get UUID of parent image. */
2299 rc = vmdkDescDDBGetUuid(pImage, &pImage->Descriptor, VMDK_DDB_PARENT_UUID,
2300 &pImage->ParentUuid);
2301 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2302 {
2303 /* Image without UUID. Probably created by VMware and not yet used
2304 * by VirtualBox. Can only be added for images opened in read/write
2305 * mode, so don't bother producing a sensible UUID otherwise. */
2306 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2307 RTUuidClear(&pImage->ParentUuid);
2308 else
2309 {
2310 rc = RTUuidClear(&pImage->ParentUuid);
2311 if (RT_FAILURE(rc))
2312 return rc;
2313 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
2314 VMDK_DDB_PARENT_UUID, &pImage->ParentUuid);
2315 if (RT_FAILURE(rc))
2316 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing parent UUID in descriptor in '%s'"), pImage->pszFilename);
2317 }
2318 }
2319 else if (RT_FAILURE(rc))
2320 return rc;
2321
2322 /* Get parent image modification UUID. */
2323 rc = vmdkDescDDBGetUuid(pImage, &pImage->Descriptor,
2324 VMDK_DDB_PARENT_MODIFICATION_UUID,
2325 &pImage->ParentModificationUuid);
2326 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
2327 {
2328 /* Image without UUID. Probably created by VMware and not yet used
2329 * by VirtualBox. Can only be added for images opened in read/write
2330 * mode, so don't bother producing a sensible UUID otherwise. */
2331 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2332 RTUuidClear(&pImage->ParentModificationUuid);
2333 else
2334 {
2335 RTUuidClear(&pImage->ParentModificationUuid);
2336 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
2337 VMDK_DDB_PARENT_MODIFICATION_UUID,
2338 &pImage->ParentModificationUuid);
2339 if (RT_FAILURE(rc))
2340 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing parent modification UUID in descriptor in '%s'"), pImage->pszFilename);
2341 }
2342 }
2343 else if (RT_FAILURE(rc))
2344 return rc;
2345
2346 return VINF_SUCCESS;
2347}
2348
2349/**
2350 * Internal : Prepares the descriptor to write to the image.
2351 */
2352static int vmdkDescriptorPrepare(PVMDKIMAGE pImage, uint64_t cbLimit,
2353 void **ppvData, size_t *pcbData)
2354{
2355 int rc = VINF_SUCCESS;
2356
2357 /*
2358 * Allocate temporary descriptor buffer.
2359 * In case there is no limit allocate a default
2360 * and increase if required.
2361 */
2362 size_t cbDescriptor = cbLimit ? cbLimit : 4 * _1K;
2363 char *pszDescriptor = (char *)RTMemAllocZ(cbDescriptor);
2364 size_t offDescriptor = 0;
2365
2366 if (!pszDescriptor)
2367 return VERR_NO_MEMORY;
2368
2369 for (unsigned i = 0; i < pImage->Descriptor.cLines; i++)
2370 {
2371 const char *psz = pImage->Descriptor.aLines[i];
2372 size_t cb = strlen(psz);
2373
2374 /*
2375 * Increase the descriptor if there is no limit and
2376 * there is not enough room left for this line.
2377 */
2378 if (offDescriptor + cb + 1 > cbDescriptor)
2379 {
2380 if (cbLimit)
2381 {
2382 rc = vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too long in '%s'"), pImage->pszFilename);
2383 break;
2384 }
2385 else
2386 {
2387 char *pszDescriptorNew = NULL;
2388 LogFlow(("Increasing descriptor cache\n"));
2389
2390 pszDescriptorNew = (char *)RTMemRealloc(pszDescriptor, cbDescriptor + cb + 4 * _1K);
2391 if (!pszDescriptorNew)
2392 {
2393 rc = VERR_NO_MEMORY;
2394 break;
2395 }
2396 pszDescriptor = pszDescriptorNew;
2397 cbDescriptor += cb + 4 * _1K;
2398 }
2399 }
2400
2401 if (cb > 0)
2402 {
2403 memcpy(pszDescriptor + offDescriptor, psz, cb);
2404 offDescriptor += cb;
2405 }
2406
2407 memcpy(pszDescriptor + offDescriptor, "\n", 1);
2408 offDescriptor++;
2409 }
2410
2411 if (RT_SUCCESS(rc))
2412 {
2413 *ppvData = pszDescriptor;
2414 *pcbData = offDescriptor;
2415 }
2416 else if (pszDescriptor)
2417 RTMemFree(pszDescriptor);
2418
2419 return rc;
2420}
2421
2422/**
2423 * Internal: write/update the descriptor part of the image.
2424 */
2425static int vmdkWriteDescriptor(PVMDKIMAGE pImage, PVDIOCTX pIoCtx)
2426{
2427 int rc = VINF_SUCCESS;
2428 uint64_t cbLimit;
2429 uint64_t uOffset;
2430 PVMDKFILE pDescFile;
2431 void *pvDescriptor = NULL;
2432 size_t cbDescriptor;
2433
2434 if (pImage->pDescData)
2435 {
2436 /* Separate descriptor file. */
2437 uOffset = 0;
2438 cbLimit = 0;
2439 pDescFile = pImage->pFile;
2440 }
2441 else
2442 {
2443 /* Embedded descriptor file. */
2444 uOffset = VMDK_SECTOR2BYTE(pImage->pExtents[0].uDescriptorSector);
2445 cbLimit = VMDK_SECTOR2BYTE(pImage->pExtents[0].cDescriptorSectors);
2446 pDescFile = pImage->pExtents[0].pFile;
2447 }
2448 /* Bail out if there is no file to write to. */
2449 if (pDescFile == NULL)
2450 return VERR_INVALID_PARAMETER;
2451
2452 rc = vmdkDescriptorPrepare(pImage, cbLimit, &pvDescriptor, &cbDescriptor);
2453 if (RT_SUCCESS(rc))
2454 {
2455 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pDescFile->pStorage,
2456 uOffset, pvDescriptor,
2457 cbLimit ? cbLimit : cbDescriptor,
2458 pIoCtx, NULL, NULL);
2459 if ( RT_FAILURE(rc)
2460 && rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
2461 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error writing descriptor in '%s'"), pImage->pszFilename);
2462 }
2463
2464 if (RT_SUCCESS(rc) && !cbLimit)
2465 {
2466 rc = vdIfIoIntFileSetSize(pImage->pIfIo, pDescFile->pStorage, cbDescriptor);
2467 if (RT_FAILURE(rc))
2468 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error truncating descriptor in '%s'"), pImage->pszFilename);
2469 }
2470
2471 if (RT_SUCCESS(rc))
2472 pImage->Descriptor.fDirty = false;
2473
2474 if (pvDescriptor)
2475 RTMemFree(pvDescriptor);
2476 return rc;
2477
2478}
2479
2480/**
2481 * Internal: validate the consistency check values in a binary header.
2482 */
2483static int vmdkValidateHeader(PVMDKIMAGE pImage, PVMDKEXTENT pExtent, const SparseExtentHeader *pHeader)
2484{
2485 int rc = VINF_SUCCESS;
2486 if (RT_LE2H_U32(pHeader->magicNumber) != VMDK_SPARSE_MAGICNUMBER)
2487 {
2488 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: incorrect magic in sparse extent header in '%s'"), pExtent->pszFullname);
2489 return rc;
2490 }
2491 if (RT_LE2H_U32(pHeader->version) != 1 && RT_LE2H_U32(pHeader->version) != 3)
2492 {
2493 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_UNSUPPORTED_VERSION, RT_SRC_POS, N_("VMDK: incorrect version in sparse extent header in '%s', not a VMDK 1.0/1.1 conforming file"), pExtent->pszFullname);
2494 return rc;
2495 }
2496 if ( (RT_LE2H_U32(pHeader->flags) & 1)
2497 && ( pHeader->singleEndLineChar != '\n'
2498 || pHeader->nonEndLineChar != ' '
2499 || pHeader->doubleEndLineChar1 != '\r'
2500 || pHeader->doubleEndLineChar2 != '\n') )
2501 {
2502 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: corrupted by CR/LF translation in '%s'"), pExtent->pszFullname);
2503 return rc;
2504 }
2505 return rc;
2506}
2507
2508/**
2509 * Internal: read metadata belonging to an extent with binary header, i.e.
2510 * as found in monolithic files.
2511 */
2512static int vmdkReadBinaryMetaExtent(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
2513 bool fMagicAlreadyRead)
2514{
2515 SparseExtentHeader Header;
2516 uint64_t cSectorsPerGDE;
2517 uint64_t cbFile = 0;
2518 int rc;
2519
2520 if (!fMagicAlreadyRead)
2521 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage, 0,
2522 &Header, sizeof(Header));
2523 else
2524 {
2525 Header.magicNumber = RT_H2LE_U32(VMDK_SPARSE_MAGICNUMBER);
2526 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
2527 RT_OFFSETOF(SparseExtentHeader, version),
2528 &Header.version,
2529 sizeof(Header)
2530 - RT_OFFSETOF(SparseExtentHeader, version));
2531 }
2532 AssertRC(rc);
2533 if (RT_FAILURE(rc))
2534 {
2535 vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error reading extent header in '%s'"), pExtent->pszFullname);
2536 rc = VERR_VD_VMDK_INVALID_HEADER;
2537 goto out;
2538 }
2539 rc = vmdkValidateHeader(pImage, pExtent, &Header);
2540 if (RT_FAILURE(rc))
2541 goto out;
2542
2543 if ( (RT_LE2H_U32(Header.flags) & RT_BIT(17))
2544 && RT_LE2H_U64(Header.gdOffset) == VMDK_GD_AT_END)
2545 pExtent->fFooter = true;
2546
2547 if ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2548 || ( pExtent->fFooter
2549 && !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL)))
2550 {
2551 rc = vdIfIoIntFileGetSize(pImage->pIfIo, pExtent->pFile->pStorage, &cbFile);
2552 AssertRC(rc);
2553 if (RT_FAILURE(rc))
2554 {
2555 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot get size of '%s'"), pExtent->pszFullname);
2556 goto out;
2557 }
2558 }
2559
2560 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
2561 pExtent->uAppendPosition = RT_ALIGN_64(cbFile, 512);
2562
2563 if ( pExtent->fFooter
2564 && ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2565 || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL)))
2566 {
2567 /* Read the footer, which comes before the end-of-stream marker. */
2568 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
2569 cbFile - 2*512, &Header,
2570 sizeof(Header));
2571 AssertRC(rc);
2572 if (RT_FAILURE(rc))
2573 {
2574 vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error reading extent footer in '%s'"), pExtent->pszFullname);
2575 rc = VERR_VD_VMDK_INVALID_HEADER;
2576 goto out;
2577 }
2578 rc = vmdkValidateHeader(pImage, pExtent, &Header);
2579 if (RT_FAILURE(rc))
2580 goto out;
2581 /* Prohibit any writes to this extent. */
2582 pExtent->uAppendPosition = 0;
2583 }
2584
2585 pExtent->uVersion = RT_LE2H_U32(Header.version);
2586 pExtent->enmType = VMDKETYPE_HOSTED_SPARSE; /* Just dummy value, changed later. */
2587 pExtent->cSectors = RT_LE2H_U64(Header.capacity);
2588 pExtent->cSectorsPerGrain = RT_LE2H_U64(Header.grainSize);
2589 pExtent->uDescriptorSector = RT_LE2H_U64(Header.descriptorOffset);
2590 pExtent->cDescriptorSectors = RT_LE2H_U64(Header.descriptorSize);
2591 if (pExtent->uDescriptorSector && !pExtent->cDescriptorSectors)
2592 {
2593 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: inconsistent embedded descriptor config in '%s'"), pExtent->pszFullname);
2594 goto out;
2595 }
2596 pExtent->cGTEntries = RT_LE2H_U32(Header.numGTEsPerGT);
2597 if (RT_LE2H_U32(Header.flags) & RT_BIT(1))
2598 {
2599 pExtent->uSectorRGD = RT_LE2H_U64(Header.rgdOffset);
2600 pExtent->uSectorGD = RT_LE2H_U64(Header.gdOffset);
2601 }
2602 else
2603 {
2604 pExtent->uSectorGD = RT_LE2H_U64(Header.gdOffset);
2605 pExtent->uSectorRGD = 0;
2606 }
2607 if ( ( pExtent->uSectorGD == VMDK_GD_AT_END
2608 || pExtent->uSectorRGD == VMDK_GD_AT_END)
2609 && ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2610 || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL)))
2611 {
2612 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: cannot resolve grain directory offset in '%s'"), pExtent->pszFullname);
2613 goto out;
2614 }
2615 pExtent->cOverheadSectors = RT_LE2H_U64(Header.overHead);
2616 pExtent->fUncleanShutdown = !!Header.uncleanShutdown;
2617 pExtent->uCompression = RT_LE2H_U16(Header.compressAlgorithm);
2618 cSectorsPerGDE = pExtent->cGTEntries * pExtent->cSectorsPerGrain;
2619 if (!cSectorsPerGDE || cSectorsPerGDE > UINT32_MAX)
2620 {
2621 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: incorrect grain directory size in '%s'"), pExtent->pszFullname);
2622 goto out;
2623 }
2624 pExtent->cSectorsPerGDE = cSectorsPerGDE;
2625 pExtent->cGDEntries = (pExtent->cSectors + cSectorsPerGDE - 1) / cSectorsPerGDE;
2626
2627 /* Fix up the number of descriptor sectors, as some flat images have
2628 * really just one, and this causes failures when inserting the UUID
2629 * values and other extra information. */
2630 if (pExtent->cDescriptorSectors != 0 && pExtent->cDescriptorSectors < 4)
2631 {
2632 /* Do it the easy way - just fix it for flat images which have no
2633 * other complicated metadata which needs space too. */
2634 if ( pExtent->uDescriptorSector + 4 < pExtent->cOverheadSectors
2635 && pExtent->cGTEntries * pExtent->cGDEntries == 0)
2636 pExtent->cDescriptorSectors = 4;
2637 }
2638
2639out:
2640 if (RT_FAILURE(rc))
2641 vmdkFreeExtentData(pImage, pExtent, false);
2642
2643 return rc;
2644}
2645
2646/**
2647 * Internal: read additional metadata belonging to an extent. For those
2648 * extents which have no additional metadata just verify the information.
2649 */
2650static int vmdkReadMetaExtent(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
2651{
2652 int rc = VINF_SUCCESS;
2653
2654/* disabled the check as there are too many truncated vmdk images out there */
2655#ifdef VBOX_WITH_VMDK_STRICT_SIZE_CHECK
2656 uint64_t cbExtentSize;
2657 /* The image must be a multiple of a sector in size and contain the data
2658 * area (flat images only). If not, it means the image is at least
2659 * truncated, or even seriously garbled. */
2660 rc = vdIfIoIntFileGetSize(pImage->pIfIo, pExtent->pFile->pStorage, &cbExtentSize);
2661 if (RT_FAILURE(rc))
2662 {
2663 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting size in '%s'"), pExtent->pszFullname);
2664 goto out;
2665 }
2666 if ( cbExtentSize != RT_ALIGN_64(cbExtentSize, 512)
2667 && (pExtent->enmType != VMDKETYPE_FLAT || pExtent->cNominalSectors + pExtent->uSectorOffset > VMDK_BYTE2SECTOR(cbExtentSize)))
2668 {
2669 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: file size is not a multiple of 512 in '%s', file is truncated or otherwise garbled"), pExtent->pszFullname);
2670 goto out;
2671 }
2672#endif /* VBOX_WITH_VMDK_STRICT_SIZE_CHECK */
2673 if (pExtent->enmType != VMDKETYPE_HOSTED_SPARSE)
2674 goto out;
2675
2676 /* The spec says that this must be a power of two and greater than 8,
2677 * but probably they meant not less than 8. */
2678 if ( (pExtent->cSectorsPerGrain & (pExtent->cSectorsPerGrain - 1))
2679 || pExtent->cSectorsPerGrain < 8)
2680 {
2681 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: invalid extent grain size %u in '%s'"), pExtent->cSectorsPerGrain, pExtent->pszFullname);
2682 goto out;
2683 }
2684
2685 /* This code requires that a grain table must hold a power of two multiple
2686 * of the number of entries per GT cache entry. */
2687 if ( (pExtent->cGTEntries & (pExtent->cGTEntries - 1))
2688 || pExtent->cGTEntries < VMDK_GT_CACHELINE_SIZE)
2689 {
2690 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: grain table cache size problem in '%s'"), pExtent->pszFullname);
2691 goto out;
2692 }
2693
2694 rc = vmdkAllocStreamBuffers(pImage, pExtent);
2695 if (RT_FAILURE(rc))
2696 goto out;
2697
2698 /* Prohibit any writes to this streamOptimized extent. */
2699 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
2700 pExtent->uAppendPosition = 0;
2701
2702 if ( !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
2703 || !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2704 || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL))
2705 rc = vmdkReadGrainDirectory(pImage, pExtent);
2706 else
2707 {
2708 pExtent->uGrainSectorAbs = pExtent->cOverheadSectors;
2709 pExtent->cbGrainStreamRead = 0;
2710 }
2711
2712out:
2713 if (RT_FAILURE(rc))
2714 vmdkFreeExtentData(pImage, pExtent, false);
2715
2716 return rc;
2717}
2718
2719/**
2720 * Internal: write/update the metadata for a sparse extent.
2721 */
2722static int vmdkWriteMetaSparseExtent(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
2723 uint64_t uOffset, PVDIOCTX pIoCtx)
2724{
2725 SparseExtentHeader Header;
2726
2727 memset(&Header, '\0', sizeof(Header));
2728 Header.magicNumber = RT_H2LE_U32(VMDK_SPARSE_MAGICNUMBER);
2729 Header.version = RT_H2LE_U32(pExtent->uVersion);
2730 Header.flags = RT_H2LE_U32(RT_BIT(0));
2731 if (pExtent->pRGD)
2732 Header.flags |= RT_H2LE_U32(RT_BIT(1));
2733 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
2734 Header.flags |= RT_H2LE_U32(RT_BIT(16) | RT_BIT(17));
2735 Header.capacity = RT_H2LE_U64(pExtent->cSectors);
2736 Header.grainSize = RT_H2LE_U64(pExtent->cSectorsPerGrain);
2737 Header.descriptorOffset = RT_H2LE_U64(pExtent->uDescriptorSector);
2738 Header.descriptorSize = RT_H2LE_U64(pExtent->cDescriptorSectors);
2739 Header.numGTEsPerGT = RT_H2LE_U32(pExtent->cGTEntries);
2740 if (pExtent->fFooter && uOffset == 0)
2741 {
2742 if (pExtent->pRGD)
2743 {
2744 Assert(pExtent->uSectorRGD);
2745 Header.rgdOffset = RT_H2LE_U64(VMDK_GD_AT_END);
2746 Header.gdOffset = RT_H2LE_U64(VMDK_GD_AT_END);
2747 }
2748 else
2749 {
2750 Header.gdOffset = RT_H2LE_U64(VMDK_GD_AT_END);
2751 }
2752 }
2753 else
2754 {
2755 if (pExtent->pRGD)
2756 {
2757 Assert(pExtent->uSectorRGD);
2758 Header.rgdOffset = RT_H2LE_U64(pExtent->uSectorRGD);
2759 Header.gdOffset = RT_H2LE_U64(pExtent->uSectorGD);
2760 }
2761 else
2762 {
2763 Header.gdOffset = RT_H2LE_U64(pExtent->uSectorGD);
2764 }
2765 }
2766 Header.overHead = RT_H2LE_U64(pExtent->cOverheadSectors);
2767 Header.uncleanShutdown = pExtent->fUncleanShutdown;
2768 Header.singleEndLineChar = '\n';
2769 Header.nonEndLineChar = ' ';
2770 Header.doubleEndLineChar1 = '\r';
2771 Header.doubleEndLineChar2 = '\n';
2772 Header.compressAlgorithm = RT_H2LE_U16(pExtent->uCompression);
2773
2774 int rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
2775 uOffset, &Header, sizeof(Header),
2776 pIoCtx, NULL, NULL);
2777 if (RT_FAILURE(rc) && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
2778 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error writing extent header in '%s'"), pExtent->pszFullname);
2779 return rc;
2780}
2781
2782#ifdef VBOX_WITH_VMDK_ESX
2783/**
2784 * Internal: unused code to read the metadata of a sparse ESX extent.
2785 *
2786 * Such extents never leave ESX server, so this isn't ever used.
2787 */
2788static int vmdkReadMetaESXSparseExtent(PVMDKEXTENT pExtent)
2789{
2790 COWDisk_Header Header;
2791 uint64_t cSectorsPerGDE;
2792
2793 int rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage, 0,
2794 &Header, sizeof(Header));
2795 AssertRC(rc);
2796 if (RT_FAILURE(rc))
2797 {
2798 vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error reading ESX sparse extent header in '%s'"), pExtent->pszFullname);
2799 rc = VERR_VD_VMDK_INVALID_HEADER;
2800 goto out;
2801 }
2802 if ( RT_LE2H_U32(Header.magicNumber) != VMDK_ESX_SPARSE_MAGICNUMBER
2803 || RT_LE2H_U32(Header.version) != 1
2804 || RT_LE2H_U32(Header.flags) != 3)
2805 {
2806 rc = VERR_VD_VMDK_INVALID_HEADER;
2807 goto out;
2808 }
2809 pExtent->enmType = VMDKETYPE_ESX_SPARSE;
2810 pExtent->cSectors = RT_LE2H_U32(Header.numSectors);
2811 pExtent->cSectorsPerGrain = RT_LE2H_U32(Header.grainSize);
2812 /* The spec says that this must be between 1 sector and 1MB. This code
2813 * assumes it's a power of two, so check that requirement, too. */
2814 if ( (pExtent->cSectorsPerGrain & (pExtent->cSectorsPerGrain - 1))
2815 || pExtent->cSectorsPerGrain == 0
2816 || pExtent->cSectorsPerGrain > 2048)
2817 {
2818 rc = VERR_VD_VMDK_INVALID_HEADER;
2819 goto out;
2820 }
2821 pExtent->uDescriptorSector = 0;
2822 pExtent->cDescriptorSectors = 0;
2823 pExtent->uSectorGD = RT_LE2H_U32(Header.gdOffset);
2824 pExtent->uSectorRGD = 0;
2825 pExtent->cOverheadSectors = 0;
2826 pExtent->cGTEntries = 4096;
2827 cSectorsPerGDE = pExtent->cGTEntries * pExtent->cSectorsPerGrain;
2828 if (!cSectorsPerGDE || cSectorsPerGDE > UINT32_MAX)
2829 {
2830 rc = VERR_VD_VMDK_INVALID_HEADER;
2831 goto out;
2832 }
2833 pExtent->cSectorsPerGDE = cSectorsPerGDE;
2834 pExtent->cGDEntries = (pExtent->cSectors + cSectorsPerGDE - 1) / cSectorsPerGDE;
2835 if (pExtent->cGDEntries != RT_LE2H_U32(Header.numGDEntries))
2836 {
2837 /* Inconsistency detected. Computed number of GD entries doesn't match
2838 * stored value. Better be safe than sorry. */
2839 rc = VERR_VD_VMDK_INVALID_HEADER;
2840 goto out;
2841 }
2842 pExtent->uFreeSector = RT_LE2H_U32(Header.freeSector);
2843 pExtent->fUncleanShutdown = !!Header.uncleanShutdown;
2844
2845 rc = vmdkReadGrainDirectory(pImage, pExtent);
2846
2847out:
2848 if (RT_FAILURE(rc))
2849 vmdkFreeExtentData(pImage, pExtent, false);
2850
2851 return rc;
2852}
2853#endif /* VBOX_WITH_VMDK_ESX */
2854
2855/**
2856 * Internal: free the buffers used for streamOptimized images.
2857 */
2858static void vmdkFreeStreamBuffers(PVMDKEXTENT pExtent)
2859{
2860 if (pExtent->pvCompGrain)
2861 {
2862 RTMemFree(pExtent->pvCompGrain);
2863 pExtent->pvCompGrain = NULL;
2864 }
2865 if (pExtent->pvGrain)
2866 {
2867 RTMemFree(pExtent->pvGrain);
2868 pExtent->pvGrain = NULL;
2869 }
2870}
2871
2872/**
2873 * Internal: free the memory used by the extent data structure, optionally
2874 * deleting the referenced files.
2875 *
2876 * @returns VBox status code.
2877 * @param pImage Pointer to the image instance data.
2878 * @param pExtent The extent to free.
2879 * @param fDelete Flag whether to delete the backing storage.
2880 */
2881static int vmdkFreeExtentData(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
2882 bool fDelete)
2883{
2884 int rc = VINF_SUCCESS;
2885
2886 vmdkFreeGrainDirectory(pExtent);
2887 if (pExtent->pDescData)
2888 {
2889 RTMemFree(pExtent->pDescData);
2890 pExtent->pDescData = NULL;
2891 }
2892 if (pExtent->pFile != NULL)
2893 {
2894 /* Do not delete raw extents, these have full and base names equal. */
2895 rc =vmdkFileClose(pImage, &pExtent->pFile,
2896 fDelete
2897 && pExtent->pszFullname
2898 && strcmp(pExtent->pszFullname, pExtent->pszBasename));
2899 }
2900 if (pExtent->pszBasename)
2901 {
2902 RTMemTmpFree((void *)pExtent->pszBasename);
2903 pExtent->pszBasename = NULL;
2904 }
2905 if (pExtent->pszFullname)
2906 {
2907 RTStrFree((char *)(void *)pExtent->pszFullname);
2908 pExtent->pszFullname = NULL;
2909 }
2910 vmdkFreeStreamBuffers(pExtent);
2911
2912 return rc;
2913}
2914
2915/**
2916 * Internal: allocate grain table cache if necessary for this image.
2917 */
2918static int vmdkAllocateGrainTableCache(PVMDKIMAGE pImage)
2919{
2920 PVMDKEXTENT pExtent;
2921
2922 /* Allocate grain table cache if any sparse extent is present. */
2923 for (unsigned i = 0; i < pImage->cExtents; i++)
2924 {
2925 pExtent = &pImage->pExtents[i];
2926 if ( pExtent->enmType == VMDKETYPE_HOSTED_SPARSE
2927#ifdef VBOX_WITH_VMDK_ESX
2928 || pExtent->enmType == VMDKETYPE_ESX_SPARSE
2929#endif /* VBOX_WITH_VMDK_ESX */
2930 )
2931 {
2932 /* Allocate grain table cache. */
2933 pImage->pGTCache = (PVMDKGTCACHE)RTMemAllocZ(sizeof(VMDKGTCACHE));
2934 if (!pImage->pGTCache)
2935 return VERR_NO_MEMORY;
2936 for (unsigned j = 0; j < VMDK_GT_CACHE_SIZE; j++)
2937 {
2938 PVMDKGTCACHEENTRY pGCE = &pImage->pGTCache->aGTCache[j];
2939 pGCE->uExtent = UINT32_MAX;
2940 }
2941 pImage->pGTCache->cEntries = VMDK_GT_CACHE_SIZE;
2942 break;
2943 }
2944 }
2945
2946 return VINF_SUCCESS;
2947}
2948
2949/**
2950 * Internal: allocate the given number of extents.
2951 */
2952static int vmdkCreateExtents(PVMDKIMAGE pImage, unsigned cExtents)
2953{
2954 int rc = VINF_SUCCESS;
2955 PVMDKEXTENT pExtents = (PVMDKEXTENT)RTMemAllocZ(cExtents * sizeof(VMDKEXTENT));
2956 if (pExtents)
2957 {
2958 for (unsigned i = 0; i < cExtents; i++)
2959 {
2960 pExtents[i].pFile = NULL;
2961 pExtents[i].pszBasename = NULL;
2962 pExtents[i].pszFullname = NULL;
2963 pExtents[i].pGD = NULL;
2964 pExtents[i].pRGD = NULL;
2965 pExtents[i].pDescData = NULL;
2966 pExtents[i].uVersion = 1;
2967 pExtents[i].uCompression = VMDK_COMPRESSION_NONE;
2968 pExtents[i].uExtent = i;
2969 pExtents[i].pImage = pImage;
2970 }
2971 pImage->pExtents = pExtents;
2972 pImage->cExtents = cExtents;
2973 }
2974 else
2975 rc = VERR_NO_MEMORY;
2976
2977 return rc;
2978}
2979
2980/**
2981 * Internal: Open an image, constructing all necessary data structures.
2982 */
2983static int vmdkOpenImage(PVMDKIMAGE pImage, unsigned uOpenFlags)
2984{
2985 int rc;
2986 uint32_t u32Magic;
2987 PVMDKFILE pFile;
2988 PVMDKEXTENT pExtent;
2989
2990 pImage->uOpenFlags = uOpenFlags;
2991
2992 pImage->pIfError = VDIfErrorGet(pImage->pVDIfsDisk);
2993 pImage->pIfIo = VDIfIoIntGet(pImage->pVDIfsImage);
2994 AssertPtrReturn(pImage->pIfIo, VERR_INVALID_PARAMETER);
2995
2996 /*
2997 * Open the image.
2998 * We don't have to check for asynchronous access because
2999 * we only support raw access and the opened file is a description
3000 * file were no data is stored.
3001 */
3002
3003 rc = vmdkFileOpen(pImage, &pFile, pImage->pszFilename,
3004 VDOpenFlagsToFileOpenFlags(uOpenFlags, false /* fCreate */));
3005 if (RT_FAILURE(rc))
3006 {
3007 /* Do NOT signal an appropriate error here, as the VD layer has the
3008 * choice of retrying the open if it failed. */
3009 goto out;
3010 }
3011 pImage->pFile = pFile;
3012
3013 /* Read magic (if present). */
3014 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pFile->pStorage, 0,
3015 &u32Magic, sizeof(u32Magic));
3016 if (RT_FAILURE(rc))
3017 {
3018 vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error reading the magic number in '%s'"), pImage->pszFilename);
3019 rc = VERR_VD_VMDK_INVALID_HEADER;
3020 goto out;
3021 }
3022
3023 /* Handle the file according to its magic number. */
3024 if (RT_LE2H_U32(u32Magic) == VMDK_SPARSE_MAGICNUMBER)
3025 {
3026 /* It's a hosted single-extent image. */
3027 rc = vmdkCreateExtents(pImage, 1);
3028 if (RT_FAILURE(rc))
3029 goto out;
3030 /* The opened file is passed to the extent. No separate descriptor
3031 * file, so no need to keep anything open for the image. */
3032 pExtent = &pImage->pExtents[0];
3033 pExtent->pFile = pFile;
3034 pImage->pFile = NULL;
3035 pExtent->pszFullname = RTPathAbsDup(pImage->pszFilename);
3036 if (!pExtent->pszFullname)
3037 {
3038 rc = VERR_NO_MEMORY;
3039 goto out;
3040 }
3041 rc = vmdkReadBinaryMetaExtent(pImage, pExtent, true /* fMagicAlreadyRead */);
3042 if (RT_FAILURE(rc))
3043 goto out;
3044
3045 /* As we're dealing with a monolithic image here, there must
3046 * be a descriptor embedded in the image file. */
3047 if (!pExtent->uDescriptorSector || !pExtent->cDescriptorSectors)
3048 {
3049 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: monolithic image without descriptor in '%s'"), pImage->pszFilename);
3050 goto out;
3051 }
3052 /* HACK: extend the descriptor if it is unusually small and it fits in
3053 * the unused space after the image header. Allows opening VMDK files
3054 * with extremely small descriptor in read/write mode. */
3055 if ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
3056 && pExtent->cDescriptorSectors < 3
3057 && (int64_t)pExtent->uSectorGD - pExtent->uDescriptorSector >= 4
3058 && (!pExtent->uSectorRGD || (int64_t)pExtent->uSectorRGD - pExtent->uDescriptorSector >= 4))
3059 {
3060 pExtent->cDescriptorSectors = 4;
3061 pExtent->fMetaDirty = true;
3062 }
3063 /* Read the descriptor from the extent. */
3064 pExtent->pDescData = (char *)RTMemAllocZ(VMDK_SECTOR2BYTE(pExtent->cDescriptorSectors));
3065 if (!pExtent->pDescData)
3066 {
3067 rc = VERR_NO_MEMORY;
3068 goto out;
3069 }
3070 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
3071 VMDK_SECTOR2BYTE(pExtent->uDescriptorSector),
3072 pExtent->pDescData,
3073 VMDK_SECTOR2BYTE(pExtent->cDescriptorSectors));
3074 AssertRC(rc);
3075 if (RT_FAILURE(rc))
3076 {
3077 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: read error for descriptor in '%s'"), pExtent->pszFullname);
3078 goto out;
3079 }
3080
3081 rc = vmdkParseDescriptor(pImage, pExtent->pDescData,
3082 VMDK_SECTOR2BYTE(pExtent->cDescriptorSectors));
3083 if (RT_FAILURE(rc))
3084 goto out;
3085
3086 if ( pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED
3087 && uOpenFlags & VD_OPEN_FLAGS_ASYNC_IO)
3088 {
3089 rc = VERR_NOT_SUPPORTED;
3090 goto out;
3091 }
3092
3093 rc = vmdkReadMetaExtent(pImage, pExtent);
3094 if (RT_FAILURE(rc))
3095 goto out;
3096
3097 /* Mark the extent as unclean if opened in read-write mode. */
3098 if ( !(uOpenFlags & VD_OPEN_FLAGS_READONLY)
3099 && !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
3100 {
3101 pExtent->fUncleanShutdown = true;
3102 pExtent->fMetaDirty = true;
3103 }
3104 }
3105 else
3106 {
3107 /* Allocate at least 10K, and make sure that there is 5K free space
3108 * in case new entries need to be added to the descriptor. Never
3109 * allocate more than 128K, because that's no valid descriptor file
3110 * and will result in the correct "truncated read" error handling. */
3111 uint64_t cbFileSize;
3112 rc = vdIfIoIntFileGetSize(pImage->pIfIo, pFile->pStorage, &cbFileSize);
3113 if (RT_FAILURE(rc))
3114 goto out;
3115
3116 /* If the descriptor file is shorter than 50 bytes it can't be valid. */
3117 if (cbFileSize < 50)
3118 {
3119 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: descriptor in '%s' is too short"), pImage->pszFilename);
3120 goto out;
3121 }
3122
3123 uint64_t cbSize = cbFileSize;
3124 if (cbSize % VMDK_SECTOR2BYTE(10))
3125 cbSize += VMDK_SECTOR2BYTE(20) - cbSize % VMDK_SECTOR2BYTE(10);
3126 else
3127 cbSize += VMDK_SECTOR2BYTE(10);
3128 cbSize = RT_MIN(cbSize, _128K);
3129 pImage->cbDescAlloc = RT_MAX(VMDK_SECTOR2BYTE(20), cbSize);
3130 pImage->pDescData = (char *)RTMemAllocZ(pImage->cbDescAlloc);
3131 if (!pImage->pDescData)
3132 {
3133 rc = VERR_NO_MEMORY;
3134 goto out;
3135 }
3136
3137 /* Don't reread the place where the magic would live in a sparse
3138 * image if it's a descriptor based one. */
3139 memcpy(pImage->pDescData, &u32Magic, sizeof(u32Magic));
3140 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pFile->pStorage, sizeof(u32Magic),
3141 pImage->pDescData + sizeof(u32Magic),
3142 RT_MIN(pImage->cbDescAlloc - sizeof(u32Magic),
3143 cbFileSize - sizeof(u32Magic)));
3144 if (RT_FAILURE(rc))
3145 {
3146 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: read error for descriptor in '%s'"), pImage->pszFilename);
3147 goto out;
3148 }
3149
3150#if 0 /** @todo: Revisit */
3151 cbRead += sizeof(u32Magic);
3152 if (cbRead == pImage->cbDescAlloc)
3153 {
3154 /* Likely the read is truncated. Better fail a bit too early
3155 * (normally the descriptor is much smaller than our buffer). */
3156 rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: cannot read descriptor in '%s'"), pImage->pszFilename);
3157 goto out;
3158 }
3159#endif
3160
3161 rc = vmdkParseDescriptor(pImage, pImage->pDescData,
3162 pImage->cbDescAlloc);
3163 if (RT_FAILURE(rc))
3164 goto out;
3165
3166 /*
3167 * We have to check for the asynchronous open flag. The
3168 * extents are parsed and the type of all are known now.
3169 * Check if every extent is either FLAT or ZERO.
3170 */
3171 if (uOpenFlags & VD_OPEN_FLAGS_ASYNC_IO)
3172 {
3173 unsigned cFlatExtents = 0;
3174
3175 for (unsigned i = 0; i < pImage->cExtents; i++)
3176 {
3177 pExtent = &pImage->pExtents[i];
3178
3179 if (( pExtent->enmType != VMDKETYPE_FLAT
3180 && pExtent->enmType != VMDKETYPE_ZERO
3181 && pExtent->enmType != VMDKETYPE_VMFS)
3182 || ((pImage->pExtents[i].enmType == VMDKETYPE_FLAT) && (cFlatExtents > 0)))
3183 {
3184 /*
3185 * Opened image contains at least one none flat or zero extent.
3186 * Return error but don't set error message as the caller
3187 * has the chance to open in non async I/O mode.
3188 */
3189 rc = VERR_NOT_SUPPORTED;
3190 goto out;
3191 }
3192 if (pExtent->enmType == VMDKETYPE_FLAT)
3193 cFlatExtents++;
3194 }
3195 }
3196
3197 for (unsigned i = 0; i < pImage->cExtents; i++)
3198 {
3199 pExtent = &pImage->pExtents[i];
3200
3201 if (pExtent->pszBasename)
3202 {
3203 /* Hack to figure out whether the specified name in the
3204 * extent descriptor is absolute. Doesn't always work, but
3205 * should be good enough for now. */
3206 char *pszFullname;
3207 /** @todo implement proper path absolute check. */
3208 if (pExtent->pszBasename[0] == RTPATH_SLASH)
3209 {
3210 pszFullname = RTStrDup(pExtent->pszBasename);
3211 if (!pszFullname)
3212 {
3213 rc = VERR_NO_MEMORY;
3214 goto out;
3215 }
3216 }
3217 else
3218 {
3219 char *pszDirname = RTStrDup(pImage->pszFilename);
3220 if (!pszDirname)
3221 {
3222 rc = VERR_NO_MEMORY;
3223 goto out;
3224 }
3225 RTPathStripFilename(pszDirname);
3226 pszFullname = RTPathJoinA(pszDirname, pExtent->pszBasename);
3227 RTStrFree(pszDirname);
3228 if (!pszFullname)
3229 {
3230 rc = VERR_NO_STR_MEMORY;
3231 goto out;
3232 }
3233 }
3234 pExtent->pszFullname = pszFullname;
3235 }
3236 else
3237 pExtent->pszFullname = NULL;
3238
3239 switch (pExtent->enmType)
3240 {
3241 case VMDKETYPE_HOSTED_SPARSE:
3242 rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszFullname,
3243 VDOpenFlagsToFileOpenFlags(uOpenFlags,
3244 false /* fCreate */));
3245 if (RT_FAILURE(rc))
3246 {
3247 /* Do NOT signal an appropriate error here, as the VD
3248 * layer has the choice of retrying the open if it
3249 * failed. */
3250 goto out;
3251 }
3252 rc = vmdkReadBinaryMetaExtent(pImage, pExtent,
3253 false /* fMagicAlreadyRead */);
3254 if (RT_FAILURE(rc))
3255 goto out;
3256 rc = vmdkReadMetaExtent(pImage, pExtent);
3257 if (RT_FAILURE(rc))
3258 goto out;
3259
3260 /* Mark extent as unclean if opened in read-write mode. */
3261 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
3262 {
3263 pExtent->fUncleanShutdown = true;
3264 pExtent->fMetaDirty = true;
3265 }
3266 break;
3267 case VMDKETYPE_VMFS:
3268 case VMDKETYPE_FLAT:
3269 rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszFullname,
3270 VDOpenFlagsToFileOpenFlags(uOpenFlags,
3271 false /* fCreate */));
3272 if (RT_FAILURE(rc))
3273 {
3274 /* Do NOT signal an appropriate error here, as the VD
3275 * layer has the choice of retrying the open if it
3276 * failed. */
3277 goto out;
3278 }
3279 break;
3280 case VMDKETYPE_ZERO:
3281 /* Nothing to do. */
3282 break;
3283 default:
3284 AssertMsgFailed(("unknown vmdk extent type %d\n", pExtent->enmType));
3285 }
3286 }
3287 }
3288
3289 /* Make sure this is not reached accidentally with an error status. */
3290 AssertRC(rc);
3291
3292 /* Determine PCHS geometry if not set. */
3293 if (pImage->PCHSGeometry.cCylinders == 0)
3294 {
3295 uint64_t cCylinders = VMDK_BYTE2SECTOR(pImage->cbSize)
3296 / pImage->PCHSGeometry.cHeads
3297 / pImage->PCHSGeometry.cSectors;
3298 pImage->PCHSGeometry.cCylinders = (unsigned)RT_MIN(cCylinders, 16383);
3299 if ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
3300 && !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
3301 {
3302 rc = vmdkDescSetPCHSGeometry(pImage, &pImage->PCHSGeometry);
3303 AssertRC(rc);
3304 }
3305 }
3306
3307 /* Update the image metadata now in case has changed. */
3308 rc = vmdkFlushImage(pImage, NULL);
3309 if (RT_FAILURE(rc))
3310 goto out;
3311
3312 /* Figure out a few per-image constants from the extents. */
3313 pImage->cbSize = 0;
3314 for (unsigned i = 0; i < pImage->cExtents; i++)
3315 {
3316 pExtent = &pImage->pExtents[i];
3317 if ( pExtent->enmType == VMDKETYPE_HOSTED_SPARSE
3318#ifdef VBOX_WITH_VMDK_ESX
3319 || pExtent->enmType == VMDKETYPE_ESX_SPARSE
3320#endif /* VBOX_WITH_VMDK_ESX */
3321 )
3322 {
3323 /* Here used to be a check whether the nominal size of an extent
3324 * is a multiple of the grain size. The spec says that this is
3325 * always the case, but unfortunately some files out there in the
3326 * wild violate the spec (e.g. ReactOS 0.3.1). */
3327 }
3328 pImage->cbSize += VMDK_SECTOR2BYTE(pExtent->cNominalSectors);
3329 }
3330
3331 for (unsigned i = 0; i < pImage->cExtents; i++)
3332 {
3333 pExtent = &pImage->pExtents[i];
3334 if ( pImage->pExtents[i].enmType == VMDKETYPE_FLAT
3335 || pImage->pExtents[i].enmType == VMDKETYPE_ZERO)
3336 {
3337 pImage->uImageFlags |= VD_IMAGE_FLAGS_FIXED;
3338 break;
3339 }
3340 }
3341
3342 if ( !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
3343 || !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
3344 || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL))
3345 rc = vmdkAllocateGrainTableCache(pImage);
3346
3347out:
3348 if (RT_FAILURE(rc))
3349 vmdkFreeImage(pImage, false);
3350 return rc;
3351}
3352
3353/**
3354 * Internal: create VMDK images for raw disk/partition access.
3355 */
3356static int vmdkCreateRawImage(PVMDKIMAGE pImage, const PVBOXHDDRAW pRaw,
3357 uint64_t cbSize)
3358{
3359 int rc = VINF_SUCCESS;
3360 PVMDKEXTENT pExtent;
3361
3362 if (pRaw->fRawDisk)
3363 {
3364 /* Full raw disk access. This requires setting up a descriptor
3365 * file and open the (flat) raw disk. */
3366 rc = vmdkCreateExtents(pImage, 1);
3367 if (RT_FAILURE(rc))
3368 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new extent list in '%s'"), pImage->pszFilename);
3369 pExtent = &pImage->pExtents[0];
3370 /* Create raw disk descriptor file. */
3371 rc = vmdkFileOpen(pImage, &pImage->pFile, pImage->pszFilename,
3372 VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
3373 true /* fCreate */));
3374 if (RT_FAILURE(rc))
3375 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new file '%s'"), pImage->pszFilename);
3376
3377 /* Set up basename for extent description. Cannot use StrDup. */
3378 size_t cbBasename = strlen(pRaw->pszRawDisk) + 1;
3379 char *pszBasename = (char *)RTMemTmpAlloc(cbBasename);
3380 if (!pszBasename)
3381 return VERR_NO_MEMORY;
3382 memcpy(pszBasename, pRaw->pszRawDisk, cbBasename);
3383 pExtent->pszBasename = pszBasename;
3384 /* For raw disks the full name is identical to the base name. */
3385 pExtent->pszFullname = RTStrDup(pszBasename);
3386 if (!pExtent->pszFullname)
3387 return VERR_NO_MEMORY;
3388 pExtent->enmType = VMDKETYPE_FLAT;
3389 pExtent->cNominalSectors = VMDK_BYTE2SECTOR(cbSize);
3390 pExtent->uSectorOffset = 0;
3391 pExtent->enmAccess = VMDKACCESS_READWRITE;
3392 pExtent->fMetaDirty = false;
3393
3394 /* Open flat image, the raw disk. */
3395 rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszFullname,
3396 VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags & ~VD_OPEN_FLAGS_READONLY,
3397 false /* fCreate */));
3398 if (RT_FAILURE(rc))
3399 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not open raw disk file '%s'"), pExtent->pszFullname);
3400 }
3401 else
3402 {
3403 /* Raw partition access. This requires setting up a descriptor
3404 * file, write the partition information to a flat extent and
3405 * open all the (flat) raw disk partitions. */
3406
3407 /* First pass over the partition data areas to determine how many
3408 * extents we need. One data area can require up to 2 extents, as
3409 * it might be necessary to skip over unpartitioned space. */
3410 unsigned cExtents = 0;
3411 uint64_t uStart = 0;
3412 for (unsigned i = 0; i < pRaw->cPartDescs; i++)
3413 {
3414 PVBOXHDDRAWPARTDESC pPart = &pRaw->pPartDescs[i];
3415 if (uStart > pPart->uStart)
3416 return vdIfError(pImage->pIfError, VERR_INVALID_PARAMETER, RT_SRC_POS, N_("VMDK: incorrect partition data area ordering set up by the caller in '%s'"), pImage->pszFilename);
3417
3418 if (uStart < pPart->uStart)
3419 cExtents++;
3420 uStart = pPart->uStart + pPart->cbData;
3421 cExtents++;
3422 }
3423 /* Another extent for filling up the rest of the image. */
3424 if (uStart != cbSize)
3425 cExtents++;
3426
3427 rc = vmdkCreateExtents(pImage, cExtents);
3428 if (RT_FAILURE(rc))
3429 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new extent list in '%s'"), pImage->pszFilename);
3430
3431 /* Create raw partition descriptor file. */
3432 rc = vmdkFileOpen(pImage, &pImage->pFile, pImage->pszFilename,
3433 VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
3434 true /* fCreate */));
3435 if (RT_FAILURE(rc))
3436 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new file '%s'"), pImage->pszFilename);
3437
3438 /* Create base filename for the partition table extent. */
3439 /** @todo remove fixed buffer without creating memory leaks. */
3440 char pszPartition[1024];
3441 const char *pszBase = RTPathFilename(pImage->pszFilename);
3442 const char *pszSuff = RTPathSuffix(pszBase);
3443 if (pszSuff == NULL)
3444 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: invalid filename '%s'"), pImage->pszFilename);
3445 char *pszBaseBase = RTStrDup(pszBase);
3446 if (!pszBaseBase)
3447 return VERR_NO_MEMORY;
3448 RTPathStripSuffix(pszBaseBase);
3449 RTStrPrintf(pszPartition, sizeof(pszPartition), "%s-pt%s",
3450 pszBaseBase, pszSuff);
3451 RTStrFree(pszBaseBase);
3452
3453 /* Second pass over the partitions, now define all extents. */
3454 uint64_t uPartOffset = 0;
3455 cExtents = 0;
3456 uStart = 0;
3457 for (unsigned i = 0; i < pRaw->cPartDescs; i++)
3458 {
3459 PVBOXHDDRAWPARTDESC pPart = &pRaw->pPartDescs[i];
3460 pExtent = &pImage->pExtents[cExtents++];
3461
3462 if (uStart < pPart->uStart)
3463 {
3464 pExtent->pszBasename = NULL;
3465 pExtent->pszFullname = NULL;
3466 pExtent->enmType = VMDKETYPE_ZERO;
3467 pExtent->cNominalSectors = VMDK_BYTE2SECTOR(pPart->uStart - uStart);
3468 pExtent->uSectorOffset = 0;
3469 pExtent->enmAccess = VMDKACCESS_READWRITE;
3470 pExtent->fMetaDirty = false;
3471 /* go to next extent */
3472 pExtent = &pImage->pExtents[cExtents++];
3473 }
3474 uStart = pPart->uStart + pPart->cbData;
3475
3476 if (pPart->pvPartitionData)
3477 {
3478 /* Set up basename for extent description. Can't use StrDup. */
3479 size_t cbBasename = strlen(pszPartition) + 1;
3480 char *pszBasename = (char *)RTMemTmpAlloc(cbBasename);
3481 if (!pszBasename)
3482 return VERR_NO_MEMORY;
3483 memcpy(pszBasename, pszPartition, cbBasename);
3484 pExtent->pszBasename = pszBasename;
3485
3486 /* Set up full name for partition extent. */
3487 char *pszDirname = RTStrDup(pImage->pszFilename);
3488 if (!pszDirname)
3489 return VERR_NO_STR_MEMORY;
3490 RTPathStripFilename(pszDirname);
3491 char *pszFullname = RTPathJoinA(pszDirname, pExtent->pszBasename);
3492 RTStrFree(pszDirname);
3493 if (!pszDirname)
3494 return VERR_NO_STR_MEMORY;
3495 pExtent->pszFullname = pszFullname;
3496 pExtent->enmType = VMDKETYPE_FLAT;
3497 pExtent->cNominalSectors = VMDK_BYTE2SECTOR(pPart->cbData);
3498 pExtent->uSectorOffset = uPartOffset;
3499 pExtent->enmAccess = VMDKACCESS_READWRITE;
3500 pExtent->fMetaDirty = false;
3501
3502 /* Create partition table flat image. */
3503 rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszFullname,
3504 VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
3505 true /* fCreate */));
3506 if (RT_FAILURE(rc))
3507 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new partition data file '%s'"), pExtent->pszFullname);
3508 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
3509 VMDK_SECTOR2BYTE(uPartOffset),
3510 pPart->pvPartitionData,
3511 pPart->cbData);
3512 if (RT_FAILURE(rc))
3513 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not write partition data to '%s'"), pExtent->pszFullname);
3514 uPartOffset += VMDK_BYTE2SECTOR(pPart->cbData);
3515 }
3516 else
3517 {
3518 if (pPart->pszRawDevice)
3519 {
3520 /* Set up basename for extent descr. Can't use StrDup. */
3521 size_t cbBasename = strlen(pPart->pszRawDevice) + 1;
3522 char *pszBasename = (char *)RTMemTmpAlloc(cbBasename);
3523 if (!pszBasename)
3524 return VERR_NO_MEMORY;
3525 memcpy(pszBasename, pPart->pszRawDevice, cbBasename);
3526 pExtent->pszBasename = pszBasename;
3527 /* For raw disks full name is identical to base name. */
3528 pExtent->pszFullname = RTStrDup(pszBasename);
3529 if (!pExtent->pszFullname)
3530 return VERR_NO_MEMORY;
3531 pExtent->enmType = VMDKETYPE_FLAT;
3532 pExtent->cNominalSectors = VMDK_BYTE2SECTOR(pPart->cbData);
3533 pExtent->uSectorOffset = VMDK_BYTE2SECTOR(pPart->uStartOffset);
3534 pExtent->enmAccess = VMDKACCESS_READWRITE;
3535 pExtent->fMetaDirty = false;
3536
3537 /* Open flat image, the raw partition. */
3538 rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszFullname,
3539 VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags & ~VD_OPEN_FLAGS_READONLY,
3540 false /* fCreate */));
3541 if (RT_FAILURE(rc))
3542 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not open raw partition file '%s'"), pExtent->pszFullname);
3543 }
3544 else
3545 {
3546 pExtent->pszBasename = NULL;
3547 pExtent->pszFullname = NULL;
3548 pExtent->enmType = VMDKETYPE_ZERO;
3549 pExtent->cNominalSectors = VMDK_BYTE2SECTOR(pPart->cbData);
3550 pExtent->uSectorOffset = 0;
3551 pExtent->enmAccess = VMDKACCESS_READWRITE;
3552 pExtent->fMetaDirty = false;
3553 }
3554 }
3555 }
3556 /* Another extent for filling up the rest of the image. */
3557 if (uStart != cbSize)
3558 {
3559 pExtent = &pImage->pExtents[cExtents++];
3560 pExtent->pszBasename = NULL;
3561 pExtent->pszFullname = NULL;
3562 pExtent->enmType = VMDKETYPE_ZERO;
3563 pExtent->cNominalSectors = VMDK_BYTE2SECTOR(cbSize - uStart);
3564 pExtent->uSectorOffset = 0;
3565 pExtent->enmAccess = VMDKACCESS_READWRITE;
3566 pExtent->fMetaDirty = false;
3567 }
3568 }
3569
3570 rc = vmdkDescBaseSetStr(pImage, &pImage->Descriptor, "createType",
3571 pRaw->fRawDisk ?
3572 "fullDevice" : "partitionedDevice");
3573 if (RT_FAILURE(rc))
3574 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set the image type in '%s'"), pImage->pszFilename);
3575 return rc;
3576}
3577
3578/**
3579 * Internal: create a regular (i.e. file-backed) VMDK image.
3580 */
3581static int vmdkCreateRegularImage(PVMDKIMAGE pImage, uint64_t cbSize,
3582 unsigned uImageFlags,
3583 PFNVDPROGRESS pfnProgress, void *pvUser,
3584 unsigned uPercentStart, unsigned uPercentSpan)
3585{
3586 int rc = VINF_SUCCESS;
3587 unsigned cExtents = 1;
3588 uint64_t cbOffset = 0;
3589 uint64_t cbRemaining = cbSize;
3590
3591 if (uImageFlags & VD_VMDK_IMAGE_FLAGS_SPLIT_2G)
3592 {
3593 cExtents = cbSize / VMDK_2G_SPLIT_SIZE;
3594 /* Do proper extent computation: need one smaller extent if the total
3595 * size isn't evenly divisible by the split size. */
3596 if (cbSize % VMDK_2G_SPLIT_SIZE)
3597 cExtents++;
3598 }
3599 rc = vmdkCreateExtents(pImage, cExtents);
3600 if (RT_FAILURE(rc))
3601 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new extent list in '%s'"), pImage->pszFilename);
3602
3603 /* Basename strings needed for constructing the extent names. */
3604 char *pszBasenameSubstr = RTPathFilename(pImage->pszFilename);
3605 AssertPtr(pszBasenameSubstr);
3606 size_t cbBasenameSubstr = strlen(pszBasenameSubstr) + 1;
3607
3608 /* Create separate descriptor file if necessary. */
3609 if (cExtents != 1 || (uImageFlags & VD_IMAGE_FLAGS_FIXED))
3610 {
3611 rc = vmdkFileOpen(pImage, &pImage->pFile, pImage->pszFilename,
3612 VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
3613 true /* fCreate */));
3614 if (RT_FAILURE(rc))
3615 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new sparse descriptor file '%s'"), pImage->pszFilename);
3616 }
3617 else
3618 pImage->pFile = NULL;
3619
3620 /* Set up all extents. */
3621 for (unsigned i = 0; i < cExtents; i++)
3622 {
3623 PVMDKEXTENT pExtent = &pImage->pExtents[i];
3624 uint64_t cbExtent = cbRemaining;
3625
3626 /* Set up fullname/basename for extent description. Cannot use StrDup
3627 * for basename, as it is not guaranteed that the memory can be freed
3628 * with RTMemTmpFree, which must be used as in other code paths
3629 * StrDup is not usable. */
3630 if (cExtents == 1 && !(uImageFlags & VD_IMAGE_FLAGS_FIXED))
3631 {
3632 char *pszBasename = (char *)RTMemTmpAlloc(cbBasenameSubstr);
3633 if (!pszBasename)
3634 return VERR_NO_MEMORY;
3635 memcpy(pszBasename, pszBasenameSubstr, cbBasenameSubstr);
3636 pExtent->pszBasename = pszBasename;
3637 }
3638 else
3639 {
3640 char *pszBasenameSuff = RTPathSuffix(pszBasenameSubstr);
3641 char *pszBasenameBase = RTStrDup(pszBasenameSubstr);
3642 RTPathStripSuffix(pszBasenameBase);
3643 char *pszTmp;
3644 size_t cbTmp;
3645 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
3646 {
3647 if (cExtents == 1)
3648 RTStrAPrintf(&pszTmp, "%s-flat%s", pszBasenameBase,
3649 pszBasenameSuff);
3650 else
3651 RTStrAPrintf(&pszTmp, "%s-f%03d%s", pszBasenameBase,
3652 i+1, pszBasenameSuff);
3653 }
3654 else
3655 RTStrAPrintf(&pszTmp, "%s-s%03d%s", pszBasenameBase, i+1,
3656 pszBasenameSuff);
3657 RTStrFree(pszBasenameBase);
3658 if (!pszTmp)
3659 return VERR_NO_STR_MEMORY;
3660 cbTmp = strlen(pszTmp) + 1;
3661 char *pszBasename = (char *)RTMemTmpAlloc(cbTmp);
3662 if (!pszBasename)
3663 return VERR_NO_MEMORY;
3664 memcpy(pszBasename, pszTmp, cbTmp);
3665 RTStrFree(pszTmp);
3666 pExtent->pszBasename = pszBasename;
3667 if (uImageFlags & VD_VMDK_IMAGE_FLAGS_SPLIT_2G)
3668 cbExtent = RT_MIN(cbRemaining, VMDK_2G_SPLIT_SIZE);
3669 }
3670 char *pszBasedirectory = RTStrDup(pImage->pszFilename);
3671 if (!pszBasedirectory)
3672 return VERR_NO_STR_MEMORY;
3673 RTPathStripFilename(pszBasedirectory);
3674 char *pszFullname = RTPathJoinA(pszBasedirectory, pExtent->pszBasename);
3675 RTStrFree(pszBasedirectory);
3676 if (!pszFullname)
3677 return VERR_NO_STR_MEMORY;
3678 pExtent->pszFullname = pszFullname;
3679
3680 /* Create file for extent. */
3681 rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszFullname,
3682 VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
3683 true /* fCreate */));
3684 if (RT_FAILURE(rc))
3685 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new file '%s'"), pExtent->pszFullname);
3686 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
3687 {
3688 rc = vdIfIoIntFileSetSize(pImage->pIfIo, pExtent->pFile->pStorage, cbExtent);
3689 if (RT_FAILURE(rc))
3690 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set size of new file '%s'"), pExtent->pszFullname);
3691
3692 /* Fill image with zeroes. We do this for every fixed-size image since on some systems
3693 * (for example Windows Vista), it takes ages to write a block near the end of a sparse
3694 * file and the guest could complain about an ATA timeout. */
3695
3696 /** @todo Starting with Linux 2.6.23, there is an fallocate() system call.
3697 * Currently supported file systems are ext4 and ocfs2. */
3698
3699 /* Allocate a temporary zero-filled buffer. Use a bigger block size to optimize writing */
3700 const size_t cbBuf = 128 * _1K;
3701 void *pvBuf = RTMemTmpAllocZ(cbBuf);
3702 if (!pvBuf)
3703 return VERR_NO_MEMORY;
3704
3705 uint64_t uOff = 0;
3706 /* Write data to all image blocks. */
3707 while (uOff < cbExtent)
3708 {
3709 unsigned cbChunk = (unsigned)RT_MIN(cbExtent, cbBuf);
3710
3711 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
3712 uOff, pvBuf, cbChunk);
3713 if (RT_FAILURE(rc))
3714 {
3715 RTMemFree(pvBuf);
3716 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: writing block failed for '%s'"), pImage->pszFilename);
3717 }
3718
3719 uOff += cbChunk;
3720
3721 if (pfnProgress)
3722 {
3723 rc = pfnProgress(pvUser,
3724 uPercentStart + (cbOffset + uOff) * uPercentSpan / cbSize);
3725 if (RT_FAILURE(rc))
3726 {
3727 RTMemFree(pvBuf);
3728 return rc;
3729 }
3730 }
3731 }
3732 RTMemTmpFree(pvBuf);
3733 }
3734
3735 /* Place descriptor file information (where integrated). */
3736 if (cExtents == 1 && !(uImageFlags & VD_IMAGE_FLAGS_FIXED))
3737 {
3738 pExtent->uDescriptorSector = 1;
3739 pExtent->cDescriptorSectors = VMDK_BYTE2SECTOR(pImage->cbDescAlloc);
3740 /* The descriptor is part of the (only) extent. */
3741 pExtent->pDescData = pImage->pDescData;
3742 pImage->pDescData = NULL;
3743 }
3744
3745 if (!(uImageFlags & VD_IMAGE_FLAGS_FIXED))
3746 {
3747 uint64_t cSectorsPerGDE, cSectorsPerGD;
3748 pExtent->enmType = VMDKETYPE_HOSTED_SPARSE;
3749 pExtent->cSectors = VMDK_BYTE2SECTOR(RT_ALIGN_64(cbExtent, _64K));
3750 pExtent->cSectorsPerGrain = VMDK_BYTE2SECTOR(_64K);
3751 pExtent->cGTEntries = 512;
3752 cSectorsPerGDE = pExtent->cGTEntries * pExtent->cSectorsPerGrain;
3753 pExtent->cSectorsPerGDE = cSectorsPerGDE;
3754 pExtent->cGDEntries = (pExtent->cSectors + cSectorsPerGDE - 1) / cSectorsPerGDE;
3755 cSectorsPerGD = (pExtent->cGDEntries + (512 / sizeof(uint32_t) - 1)) / (512 / sizeof(uint32_t));
3756 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
3757 {
3758 /* The spec says version is 1 for all VMDKs, but the vast
3759 * majority of streamOptimized VMDKs actually contain
3760 * version 3 - so go with the majority. Both are accepted. */
3761 pExtent->uVersion = 3;
3762 pExtent->uCompression = VMDK_COMPRESSION_DEFLATE;
3763 }
3764 }
3765 else
3766 {
3767 if (uImageFlags & VD_VMDK_IMAGE_FLAGS_ESX)
3768 pExtent->enmType = VMDKETYPE_VMFS;
3769 else
3770 pExtent->enmType = VMDKETYPE_FLAT;
3771 }
3772
3773 pExtent->enmAccess = VMDKACCESS_READWRITE;
3774 pExtent->fUncleanShutdown = true;
3775 pExtent->cNominalSectors = VMDK_BYTE2SECTOR(cbExtent);
3776 pExtent->uSectorOffset = 0;
3777 pExtent->fMetaDirty = true;
3778
3779 if (!(uImageFlags & VD_IMAGE_FLAGS_FIXED))
3780 {
3781 /* fPreAlloc should never be false because VMware can't use such images. */
3782 rc = vmdkCreateGrainDirectory(pImage, pExtent,
3783 RT_MAX( pExtent->uDescriptorSector
3784 + pExtent->cDescriptorSectors,
3785 1),
3786 true /* fPreAlloc */);
3787 if (RT_FAILURE(rc))
3788 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new grain directory in '%s'"), pExtent->pszFullname);
3789 }
3790
3791 cbOffset += cbExtent;
3792
3793 if (RT_SUCCESS(rc) && pfnProgress)
3794 pfnProgress(pvUser, uPercentStart + cbOffset * uPercentSpan / cbSize);
3795
3796 cbRemaining -= cbExtent;
3797 }
3798
3799 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_ESX)
3800 {
3801 /* VirtualBox doesn't care, but VMWare ESX freaks out if the wrong
3802 * controller type is set in an image. */
3803 rc = vmdkDescDDBSetStr(pImage, &pImage->Descriptor, "ddb.adapterType", "lsilogic");
3804 if (RT_FAILURE(rc))
3805 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set controller type to lsilogic in '%s'"), pImage->pszFilename);
3806 }
3807
3808 const char *pszDescType = NULL;
3809 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
3810 {
3811 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_ESX)
3812 pszDescType = "vmfs";
3813 else
3814 pszDescType = (cExtents == 1)
3815 ? "monolithicFlat" : "twoGbMaxExtentFlat";
3816 }
3817 else
3818 {
3819 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
3820 pszDescType = "streamOptimized";
3821 else
3822 {
3823 pszDescType = (cExtents == 1)
3824 ? "monolithicSparse" : "twoGbMaxExtentSparse";
3825 }
3826 }
3827 rc = vmdkDescBaseSetStr(pImage, &pImage->Descriptor, "createType",
3828 pszDescType);
3829 if (RT_FAILURE(rc))
3830 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set the image type in '%s'"), pImage->pszFilename);
3831 return rc;
3832}
3833
3834/**
3835 * Internal: Create a real stream optimized VMDK using only linear writes.
3836 */
3837static int vmdkCreateStreamImage(PVMDKIMAGE pImage, uint64_t cbSize,
3838 unsigned uImageFlags,
3839 PFNVDPROGRESS pfnProgress, void *pvUser,
3840 unsigned uPercentStart, unsigned uPercentSpan)
3841{
3842 int rc;
3843
3844 rc = vmdkCreateExtents(pImage, 1);
3845 if (RT_FAILURE(rc))
3846 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new extent list in '%s'"), pImage->pszFilename);
3847
3848 /* Basename strings needed for constructing the extent names. */
3849 const char *pszBasenameSubstr = RTPathFilename(pImage->pszFilename);
3850 AssertPtr(pszBasenameSubstr);
3851 size_t cbBasenameSubstr = strlen(pszBasenameSubstr) + 1;
3852
3853 /* No separate descriptor file. */
3854 pImage->pFile = NULL;
3855
3856 /* Set up all extents. */
3857 PVMDKEXTENT pExtent = &pImage->pExtents[0];
3858
3859 /* Set up fullname/basename for extent description. Cannot use StrDup
3860 * for basename, as it is not guaranteed that the memory can be freed
3861 * with RTMemTmpFree, which must be used as in other code paths
3862 * StrDup is not usable. */
3863 char *pszBasename = (char *)RTMemTmpAlloc(cbBasenameSubstr);
3864 if (!pszBasename)
3865 return VERR_NO_MEMORY;
3866 memcpy(pszBasename, pszBasenameSubstr, cbBasenameSubstr);
3867 pExtent->pszBasename = pszBasename;
3868
3869 char *pszBasedirectory = RTStrDup(pImage->pszFilename);
3870 RTPathStripFilename(pszBasedirectory);
3871 char *pszFullname = RTPathJoinA(pszBasedirectory, pExtent->pszBasename);
3872 RTStrFree(pszBasedirectory);
3873 if (!pszFullname)
3874 return VERR_NO_STR_MEMORY;
3875 pExtent->pszFullname = pszFullname;
3876
3877 /* Create file for extent. Make it write only, no reading allowed. */
3878 rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszFullname,
3879 VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
3880 true /* fCreate */)
3881 & ~RTFILE_O_READ);
3882 if (RT_FAILURE(rc))
3883 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new file '%s'"), pExtent->pszFullname);
3884
3885 /* Place descriptor file information. */
3886 pExtent->uDescriptorSector = 1;
3887 pExtent->cDescriptorSectors = VMDK_BYTE2SECTOR(pImage->cbDescAlloc);
3888 /* The descriptor is part of the (only) extent. */
3889 pExtent->pDescData = pImage->pDescData;
3890 pImage->pDescData = NULL;
3891
3892 uint64_t cSectorsPerGDE, cSectorsPerGD;
3893 pExtent->enmType = VMDKETYPE_HOSTED_SPARSE;
3894 pExtent->cSectors = VMDK_BYTE2SECTOR(RT_ALIGN_64(cbSize, _64K));
3895 pExtent->cSectorsPerGrain = VMDK_BYTE2SECTOR(_64K);
3896 pExtent->cGTEntries = 512;
3897 cSectorsPerGDE = pExtent->cGTEntries * pExtent->cSectorsPerGrain;
3898 pExtent->cSectorsPerGDE = cSectorsPerGDE;
3899 pExtent->cGDEntries = (pExtent->cSectors + cSectorsPerGDE - 1) / cSectorsPerGDE;
3900 cSectorsPerGD = (pExtent->cGDEntries + (512 / sizeof(uint32_t) - 1)) / (512 / sizeof(uint32_t));
3901
3902 /* The spec says version is 1 for all VMDKs, but the vast
3903 * majority of streamOptimized VMDKs actually contain
3904 * version 3 - so go with the majority. Both are accepted. */
3905 pExtent->uVersion = 3;
3906 pExtent->uCompression = VMDK_COMPRESSION_DEFLATE;
3907 pExtent->fFooter = true;
3908
3909 pExtent->enmAccess = VMDKACCESS_READONLY;
3910 pExtent->fUncleanShutdown = false;
3911 pExtent->cNominalSectors = VMDK_BYTE2SECTOR(cbSize);
3912 pExtent->uSectorOffset = 0;
3913 pExtent->fMetaDirty = true;
3914
3915 /* Create grain directory, without preallocating it straight away. It will
3916 * be constructed on the fly when writing out the data and written when
3917 * closing the image. The end effect is that the full grain directory is
3918 * allocated, which is a requirement of the VMDK specs. */
3919 rc = vmdkCreateGrainDirectory(pImage, pExtent, VMDK_GD_AT_END,
3920 false /* fPreAlloc */);
3921 if (RT_FAILURE(rc))
3922 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new grain directory in '%s'"), pExtent->pszFullname);
3923
3924 rc = vmdkDescBaseSetStr(pImage, &pImage->Descriptor, "createType",
3925 "streamOptimized");
3926 if (RT_FAILURE(rc))
3927 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set the image type in '%s'"), pImage->pszFilename);
3928
3929 return rc;
3930}
3931
3932/**
3933 * Internal: The actual code for creating any VMDK variant currently in
3934 * existence on hosted environments.
3935 */
3936static int vmdkCreateImage(PVMDKIMAGE pImage, uint64_t cbSize,
3937 unsigned uImageFlags, const char *pszComment,
3938 PCVDGEOMETRY pPCHSGeometry,
3939 PCVDGEOMETRY pLCHSGeometry, PCRTUUID pUuid,
3940 PFNVDPROGRESS pfnProgress, void *pvUser,
3941 unsigned uPercentStart, unsigned uPercentSpan)
3942{
3943 int rc;
3944
3945 pImage->uImageFlags = uImageFlags;
3946
3947 pImage->pIfError = VDIfErrorGet(pImage->pVDIfsDisk);
3948 pImage->pIfIo = VDIfIoIntGet(pImage->pVDIfsImage);
3949 AssertPtrReturn(pImage->pIfIo, VERR_INVALID_PARAMETER);
3950
3951 rc = vmdkCreateDescriptor(pImage, pImage->pDescData, pImage->cbDescAlloc,
3952 &pImage->Descriptor);
3953 if (RT_FAILURE(rc))
3954 {
3955 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new descriptor in '%s'"), pImage->pszFilename);
3956 goto out;
3957 }
3958
3959 if ( (uImageFlags & VD_IMAGE_FLAGS_FIXED)
3960 && (uImageFlags & VD_VMDK_IMAGE_FLAGS_RAWDISK))
3961 {
3962 /* Raw disk image (includes raw partition). */
3963 const PVBOXHDDRAW pRaw = (const PVBOXHDDRAW)pszComment;
3964 /* As the comment is misused, zap it so that no garbage comment
3965 * is set below. */
3966 pszComment = NULL;
3967 rc = vmdkCreateRawImage(pImage, pRaw, cbSize);
3968 }
3969 else
3970 {
3971 if (uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
3972 {
3973 /* Stream optimized sparse image (monolithic). */
3974 rc = vmdkCreateStreamImage(pImage, cbSize, uImageFlags,
3975 pfnProgress, pvUser, uPercentStart,
3976 uPercentSpan * 95 / 100);
3977 }
3978 else
3979 {
3980 /* Regular fixed or sparse image (monolithic or split). */
3981 rc = vmdkCreateRegularImage(pImage, cbSize, uImageFlags,
3982 pfnProgress, pvUser, uPercentStart,
3983 uPercentSpan * 95 / 100);
3984 }
3985 }
3986
3987 if (RT_FAILURE(rc))
3988 goto out;
3989
3990 if (RT_SUCCESS(rc) && pfnProgress)
3991 pfnProgress(pvUser, uPercentStart + uPercentSpan * 98 / 100);
3992
3993 pImage->cbSize = cbSize;
3994
3995 for (unsigned i = 0; i < pImage->cExtents; i++)
3996 {
3997 PVMDKEXTENT pExtent = &pImage->pExtents[i];
3998
3999 rc = vmdkDescExtInsert(pImage, &pImage->Descriptor, pExtent->enmAccess,
4000 pExtent->cNominalSectors, pExtent->enmType,
4001 pExtent->pszBasename, pExtent->uSectorOffset);
4002 if (RT_FAILURE(rc))
4003 {
4004 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not insert the extent list into descriptor in '%s'"), pImage->pszFilename);
4005 goto out;
4006 }
4007 }
4008 vmdkDescExtRemoveDummy(pImage, &pImage->Descriptor);
4009
4010 if ( pPCHSGeometry->cCylinders != 0
4011 && pPCHSGeometry->cHeads != 0
4012 && pPCHSGeometry->cSectors != 0)
4013 {
4014 rc = vmdkDescSetPCHSGeometry(pImage, pPCHSGeometry);
4015 if (RT_FAILURE(rc))
4016 goto out;
4017 }
4018 if ( pLCHSGeometry->cCylinders != 0
4019 && pLCHSGeometry->cHeads != 0
4020 && pLCHSGeometry->cSectors != 0)
4021 {
4022 rc = vmdkDescSetLCHSGeometry(pImage, pLCHSGeometry);
4023 if (RT_FAILURE(rc))
4024 goto out;
4025 }
4026
4027 pImage->LCHSGeometry = *pLCHSGeometry;
4028 pImage->PCHSGeometry = *pPCHSGeometry;
4029
4030 pImage->ImageUuid = *pUuid;
4031 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
4032 VMDK_DDB_IMAGE_UUID, &pImage->ImageUuid);
4033 if (RT_FAILURE(rc))
4034 {
4035 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing image UUID in new descriptor in '%s'"), pImage->pszFilename);
4036 goto out;
4037 }
4038 RTUuidClear(&pImage->ParentUuid);
4039 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
4040 VMDK_DDB_PARENT_UUID, &pImage->ParentUuid);
4041 if (RT_FAILURE(rc))
4042 {
4043 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing parent image UUID in new descriptor in '%s'"), pImage->pszFilename);
4044 goto out;
4045 }
4046 RTUuidClear(&pImage->ModificationUuid);
4047 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
4048 VMDK_DDB_MODIFICATION_UUID,
4049 &pImage->ModificationUuid);
4050 if (RT_FAILURE(rc))
4051 {
4052 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing modification UUID in new descriptor in '%s'"), pImage->pszFilename);
4053 goto out;
4054 }
4055 RTUuidClear(&pImage->ParentModificationUuid);
4056 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
4057 VMDK_DDB_PARENT_MODIFICATION_UUID,
4058 &pImage->ParentModificationUuid);
4059 if (RT_FAILURE(rc))
4060 {
4061 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing parent modification UUID in new descriptor in '%s'"), pImage->pszFilename);
4062 goto out;
4063 }
4064
4065 rc = vmdkAllocateGrainTableCache(pImage);
4066 if (RT_FAILURE(rc))
4067 goto out;
4068
4069 rc = vmdkSetImageComment(pImage, pszComment);
4070 if (RT_FAILURE(rc))
4071 {
4072 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot set image comment in '%s'"), pImage->pszFilename);
4073 goto out;
4074 }
4075
4076 if (RT_SUCCESS(rc) && pfnProgress)
4077 pfnProgress(pvUser, uPercentStart + uPercentSpan * 99 / 100);
4078
4079 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
4080 {
4081 /* streamOptimized is a bit special, we cannot trigger the flush
4082 * until all data has been written. So we write the necessary
4083 * information explicitly. */
4084 pImage->pExtents[0].cDescriptorSectors = VMDK_BYTE2SECTOR(RT_ALIGN_64( pImage->Descriptor.aLines[pImage->Descriptor.cLines]
4085 - pImage->Descriptor.aLines[0], 512));
4086 rc = vmdkWriteMetaSparseExtent(pImage, &pImage->pExtents[0], 0, NULL);
4087 if (RT_FAILURE(rc))
4088 {
4089 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write VMDK header in '%s'"), pImage->pszFilename);
4090 goto out;
4091 }
4092
4093 rc = vmdkWriteDescriptor(pImage, NULL);
4094 if (RT_FAILURE(rc))
4095 {
4096 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write VMDK descriptor in '%s'"), pImage->pszFilename);
4097 goto out;
4098 }
4099 }
4100 else
4101 rc = vmdkFlushImage(pImage, NULL);
4102
4103out:
4104 if (RT_SUCCESS(rc) && pfnProgress)
4105 pfnProgress(pvUser, uPercentStart + uPercentSpan);
4106
4107 if (RT_FAILURE(rc))
4108 vmdkFreeImage(pImage, rc != VERR_ALREADY_EXISTS);
4109 return rc;
4110}
4111
4112/**
4113 * Internal: Update image comment.
4114 */
4115static int vmdkSetImageComment(PVMDKIMAGE pImage, const char *pszComment)
4116{
4117 char *pszCommentEncoded;
4118 if (pszComment)
4119 {
4120 pszCommentEncoded = vmdkEncodeString(pszComment);
4121 if (!pszCommentEncoded)
4122 return VERR_NO_MEMORY;
4123 }
4124 else
4125 pszCommentEncoded = NULL;
4126 int rc = vmdkDescDDBSetStr(pImage, &pImage->Descriptor,
4127 "ddb.comment", pszCommentEncoded);
4128 if (pszComment)
4129 RTStrFree(pszCommentEncoded);
4130 if (RT_FAILURE(rc))
4131 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing image comment in descriptor in '%s'"), pImage->pszFilename);
4132 return VINF_SUCCESS;
4133}
4134
4135/**
4136 * Internal. Clear the grain table buffer for real stream optimized writing.
4137 */
4138static void vmdkStreamClearGT(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
4139{
4140 uint32_t cCacheLines = RT_ALIGN(pExtent->cGTEntries, VMDK_GT_CACHELINE_SIZE) / VMDK_GT_CACHELINE_SIZE;
4141 for (uint32_t i = 0; i < cCacheLines; i++)
4142 memset(&pImage->pGTCache->aGTCache[i].aGTData[0], '\0',
4143 VMDK_GT_CACHELINE_SIZE * sizeof(uint32_t));
4144}
4145
4146/**
4147 * Internal. Flush the grain table buffer for real stream optimized writing.
4148 */
4149static int vmdkStreamFlushGT(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
4150 uint32_t uGDEntry)
4151{
4152 int rc = VINF_SUCCESS;
4153 uint32_t cCacheLines = RT_ALIGN(pExtent->cGTEntries, VMDK_GT_CACHELINE_SIZE) / VMDK_GT_CACHELINE_SIZE;
4154
4155 /* VMware does not write out completely empty grain tables in the case
4156 * of streamOptimized images, which according to my interpretation of
4157 * the VMDK 1.1 spec is bending the rules. Since they do it and we can
4158 * handle it without problems do it the same way and save some bytes. */
4159 bool fAllZero = true;
4160 for (uint32_t i = 0; i < cCacheLines; i++)
4161 {
4162 /* Convert the grain table to little endian in place, as it will not
4163 * be used at all after this function has been called. */
4164 uint32_t *pGTTmp = &pImage->pGTCache->aGTCache[i].aGTData[0];
4165 for (uint32_t j = 0; j < VMDK_GT_CACHELINE_SIZE; j++, pGTTmp++)
4166 if (*pGTTmp)
4167 {
4168 fAllZero = false;
4169 break;
4170 }
4171 if (!fAllZero)
4172 break;
4173 }
4174 if (fAllZero)
4175 return VINF_SUCCESS;
4176
4177 uint64_t uFileOffset = pExtent->uAppendPosition;
4178 if (!uFileOffset)
4179 return VERR_INTERNAL_ERROR;
4180 /* Align to sector, as the previous write could have been any size. */
4181 uFileOffset = RT_ALIGN_64(uFileOffset, 512);
4182
4183 /* Grain table marker. */
4184 uint8_t aMarker[512];
4185 PVMDKMARKER pMarker = (PVMDKMARKER)&aMarker[0];
4186 memset(pMarker, '\0', sizeof(aMarker));
4187 pMarker->uSector = RT_H2LE_U64(VMDK_BYTE2SECTOR((uint64_t)pExtent->cGTEntries * sizeof(uint32_t)));
4188 pMarker->uType = RT_H2LE_U32(VMDK_MARKER_GT);
4189 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage, uFileOffset,
4190 aMarker, sizeof(aMarker));
4191 AssertRC(rc);
4192 uFileOffset += 512;
4193
4194 if (!pExtent->pGD || pExtent->pGD[uGDEntry])
4195 return VERR_INTERNAL_ERROR;
4196
4197 pExtent->pGD[uGDEntry] = VMDK_BYTE2SECTOR(uFileOffset);
4198
4199 for (uint32_t i = 0; i < cCacheLines; i++)
4200 {
4201 /* Convert the grain table to little endian in place, as it will not
4202 * be used at all after this function has been called. */
4203 uint32_t *pGTTmp = &pImage->pGTCache->aGTCache[i].aGTData[0];
4204 for (uint32_t j = 0; j < VMDK_GT_CACHELINE_SIZE; j++, pGTTmp++)
4205 *pGTTmp = RT_H2LE_U32(*pGTTmp);
4206
4207 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage, uFileOffset,
4208 &pImage->pGTCache->aGTCache[i].aGTData[0],
4209 VMDK_GT_CACHELINE_SIZE * sizeof(uint32_t));
4210 uFileOffset += VMDK_GT_CACHELINE_SIZE * sizeof(uint32_t);
4211 if (RT_FAILURE(rc))
4212 break;
4213 }
4214 Assert(!(uFileOffset % 512));
4215 pExtent->uAppendPosition = RT_ALIGN_64(uFileOffset, 512);
4216 return rc;
4217}
4218
4219/**
4220 * Internal. Free all allocated space for representing an image, and optionally
4221 * delete the image from disk.
4222 */
4223static int vmdkFreeImage(PVMDKIMAGE pImage, bool fDelete)
4224{
4225 int rc = VINF_SUCCESS;
4226
4227 /* Freeing a never allocated image (e.g. because the open failed) is
4228 * not signalled as an error. After all nothing bad happens. */
4229 if (pImage)
4230 {
4231 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
4232 {
4233 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
4234 {
4235 /* Check if all extents are clean. */
4236 for (unsigned i = 0; i < pImage->cExtents; i++)
4237 {
4238 Assert(!pImage->pExtents[i].fUncleanShutdown);
4239 }
4240 }
4241 else
4242 {
4243 /* Mark all extents as clean. */
4244 for (unsigned i = 0; i < pImage->cExtents; i++)
4245 {
4246 if ( ( pImage->pExtents[i].enmType == VMDKETYPE_HOSTED_SPARSE
4247#ifdef VBOX_WITH_VMDK_ESX
4248 || pImage->pExtents[i].enmType == VMDKETYPE_ESX_SPARSE
4249#endif /* VBOX_WITH_VMDK_ESX */
4250 )
4251 && pImage->pExtents[i].fUncleanShutdown)
4252 {
4253 pImage->pExtents[i].fUncleanShutdown = false;
4254 pImage->pExtents[i].fMetaDirty = true;
4255 }
4256
4257 /* From now on it's not safe to append any more data. */
4258 pImage->pExtents[i].uAppendPosition = 0;
4259 }
4260 }
4261 }
4262
4263 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
4264 {
4265 /* No need to write any pending data if the file will be deleted
4266 * or if the new file wasn't successfully created. */
4267 if ( !fDelete && pImage->pExtents
4268 && pImage->pExtents[0].cGTEntries
4269 && pImage->pExtents[0].uAppendPosition)
4270 {
4271 PVMDKEXTENT pExtent = &pImage->pExtents[0];
4272 uint32_t uLastGDEntry = pExtent->uLastGrainAccess / pExtent->cGTEntries;
4273 rc = vmdkStreamFlushGT(pImage, pExtent, uLastGDEntry);
4274 AssertRC(rc);
4275 vmdkStreamClearGT(pImage, pExtent);
4276 for (uint32_t i = uLastGDEntry + 1; i < pExtent->cGDEntries; i++)
4277 {
4278 rc = vmdkStreamFlushGT(pImage, pExtent, i);
4279 AssertRC(rc);
4280 }
4281
4282 uint64_t uFileOffset = pExtent->uAppendPosition;
4283 if (!uFileOffset)
4284 return VERR_INTERNAL_ERROR;
4285 uFileOffset = RT_ALIGN_64(uFileOffset, 512);
4286
4287 /* From now on it's not safe to append any more data. */
4288 pExtent->uAppendPosition = 0;
4289
4290 /* Grain directory marker. */
4291 uint8_t aMarker[512];
4292 PVMDKMARKER pMarker = (PVMDKMARKER)&aMarker[0];
4293 memset(pMarker, '\0', sizeof(aMarker));
4294 pMarker->uSector = VMDK_BYTE2SECTOR(RT_ALIGN_64(RT_H2LE_U64((uint64_t)pExtent->cGDEntries * sizeof(uint32_t)), 512));
4295 pMarker->uType = RT_H2LE_U32(VMDK_MARKER_GD);
4296 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage, uFileOffset,
4297 aMarker, sizeof(aMarker));
4298 AssertRC(rc);
4299 uFileOffset += 512;
4300
4301 /* Write grain directory in little endian style. The array will
4302 * not be used after this, so convert in place. */
4303 uint32_t *pGDTmp = pExtent->pGD;
4304 for (uint32_t i = 0; i < pExtent->cGDEntries; i++, pGDTmp++)
4305 *pGDTmp = RT_H2LE_U32(*pGDTmp);
4306 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
4307 uFileOffset, pExtent->pGD,
4308 pExtent->cGDEntries * sizeof(uint32_t));
4309 AssertRC(rc);
4310
4311 pExtent->uSectorGD = VMDK_BYTE2SECTOR(uFileOffset);
4312 pExtent->uSectorRGD = VMDK_BYTE2SECTOR(uFileOffset);
4313 uFileOffset = RT_ALIGN_64( uFileOffset
4314 + pExtent->cGDEntries * sizeof(uint32_t),
4315 512);
4316
4317 /* Footer marker. */
4318 memset(pMarker, '\0', sizeof(aMarker));
4319 pMarker->uSector = VMDK_BYTE2SECTOR(512);
4320 pMarker->uType = RT_H2LE_U32(VMDK_MARKER_FOOTER);
4321 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
4322 uFileOffset, aMarker, sizeof(aMarker));
4323 AssertRC(rc);
4324
4325 uFileOffset += 512;
4326 rc = vmdkWriteMetaSparseExtent(pImage, pExtent, uFileOffset, NULL);
4327 AssertRC(rc);
4328
4329 uFileOffset += 512;
4330 /* End-of-stream marker. */
4331 memset(pMarker, '\0', sizeof(aMarker));
4332 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
4333 uFileOffset, aMarker, sizeof(aMarker));
4334 AssertRC(rc);
4335 }
4336 }
4337 else
4338 vmdkFlushImage(pImage, NULL);
4339
4340 if (pImage->pExtents != NULL)
4341 {
4342 for (unsigned i = 0 ; i < pImage->cExtents; i++)
4343 {
4344 int rc2 = vmdkFreeExtentData(pImage, &pImage->pExtents[i], fDelete);
4345 if (RT_SUCCESS(rc))
4346 rc = rc2; /* Propogate any error when closing the file. */
4347 }
4348 RTMemFree(pImage->pExtents);
4349 pImage->pExtents = NULL;
4350 }
4351 pImage->cExtents = 0;
4352 if (pImage->pFile != NULL)
4353 {
4354 int rc2 = vmdkFileClose(pImage, &pImage->pFile, fDelete);
4355 if (RT_SUCCESS(rc))
4356 rc = rc2; /* Propogate any error when closing the file. */
4357 }
4358 int rc2 = vmdkFileCheckAllClose(pImage);
4359 if (RT_SUCCESS(rc))
4360 rc = rc2; /* Propogate any error when closing the file. */
4361
4362 if (pImage->pGTCache)
4363 {
4364 RTMemFree(pImage->pGTCache);
4365 pImage->pGTCache = NULL;
4366 }
4367 if (pImage->pDescData)
4368 {
4369 RTMemFree(pImage->pDescData);
4370 pImage->pDescData = NULL;
4371 }
4372 }
4373
4374 LogFlowFunc(("returns %Rrc\n", rc));
4375 return rc;
4376}
4377
4378/**
4379 * Internal. Flush image data (and metadata) to disk.
4380 */
4381static int vmdkFlushImage(PVMDKIMAGE pImage, PVDIOCTX pIoCtx)
4382{
4383 PVMDKEXTENT pExtent;
4384 int rc = VINF_SUCCESS;
4385
4386 /* Update descriptor if changed. */
4387 if (pImage->Descriptor.fDirty)
4388 {
4389 rc = vmdkWriteDescriptor(pImage, pIoCtx);
4390 if (RT_FAILURE(rc))
4391 goto out;
4392 }
4393
4394 for (unsigned i = 0; i < pImage->cExtents; i++)
4395 {
4396 pExtent = &pImage->pExtents[i];
4397 if (pExtent->pFile != NULL && pExtent->fMetaDirty)
4398 {
4399 switch (pExtent->enmType)
4400 {
4401 case VMDKETYPE_HOSTED_SPARSE:
4402 if (!pExtent->fFooter)
4403 {
4404 rc = vmdkWriteMetaSparseExtent(pImage, pExtent, 0, pIoCtx);
4405 if (RT_FAILURE(rc))
4406 goto out;
4407 }
4408 else
4409 {
4410 uint64_t uFileOffset = pExtent->uAppendPosition;
4411 /* Simply skip writing anything if the streamOptimized
4412 * image hasn't been just created. */
4413 if (!uFileOffset)
4414 break;
4415 uFileOffset = RT_ALIGN_64(uFileOffset, 512);
4416 rc = vmdkWriteMetaSparseExtent(pImage, pExtent,
4417 uFileOffset, pIoCtx);
4418 if (RT_FAILURE(rc))
4419 goto out;
4420 }
4421 break;
4422#ifdef VBOX_WITH_VMDK_ESX
4423 case VMDKETYPE_ESX_SPARSE:
4424 /** @todo update the header. */
4425 break;
4426#endif /* VBOX_WITH_VMDK_ESX */
4427 case VMDKETYPE_VMFS:
4428 case VMDKETYPE_FLAT:
4429 /* Nothing to do. */
4430 break;
4431 case VMDKETYPE_ZERO:
4432 default:
4433 AssertMsgFailed(("extent with type %d marked as dirty\n",
4434 pExtent->enmType));
4435 break;
4436 }
4437 }
4438 switch (pExtent->enmType)
4439 {
4440 case VMDKETYPE_HOSTED_SPARSE:
4441#ifdef VBOX_WITH_VMDK_ESX
4442 case VMDKETYPE_ESX_SPARSE:
4443#endif /* VBOX_WITH_VMDK_ESX */
4444 case VMDKETYPE_VMFS:
4445 case VMDKETYPE_FLAT:
4446 /** @todo implement proper path absolute check. */
4447 if ( pExtent->pFile != NULL
4448 && !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
4449 && !(pExtent->pszBasename[0] == RTPATH_SLASH))
4450 rc = vdIfIoIntFileFlush(pImage->pIfIo, pExtent->pFile->pStorage, pIoCtx,
4451 NULL, NULL);
4452 break;
4453 case VMDKETYPE_ZERO:
4454 /* No need to do anything for this extent. */
4455 break;
4456 default:
4457 AssertMsgFailed(("unknown extent type %d\n", pExtent->enmType));
4458 break;
4459 }
4460 }
4461
4462out:
4463 return rc;
4464}
4465
4466/**
4467 * Internal. Find extent corresponding to the sector number in the disk.
4468 */
4469static int vmdkFindExtent(PVMDKIMAGE pImage, uint64_t offSector,
4470 PVMDKEXTENT *ppExtent, uint64_t *puSectorInExtent)
4471{
4472 PVMDKEXTENT pExtent = NULL;
4473 int rc = VINF_SUCCESS;
4474
4475 for (unsigned i = 0; i < pImage->cExtents; i++)
4476 {
4477 if (offSector < pImage->pExtents[i].cNominalSectors)
4478 {
4479 pExtent = &pImage->pExtents[i];
4480 *puSectorInExtent = offSector + pImage->pExtents[i].uSectorOffset;
4481 break;
4482 }
4483 offSector -= pImage->pExtents[i].cNominalSectors;
4484 }
4485
4486 if (pExtent)
4487 *ppExtent = pExtent;
4488 else
4489 rc = VERR_IO_SECTOR_NOT_FOUND;
4490
4491 return rc;
4492}
4493
4494/**
4495 * Internal. Hash function for placing the grain table hash entries.
4496 */
4497static uint32_t vmdkGTCacheHash(PVMDKGTCACHE pCache, uint64_t uSector,
4498 unsigned uExtent)
4499{
4500 /** @todo this hash function is quite simple, maybe use a better one which
4501 * scrambles the bits better. */
4502 return (uSector + uExtent) % pCache->cEntries;
4503}
4504
4505/**
4506 * Internal. Get sector number in the extent file from the relative sector
4507 * number in the extent.
4508 */
4509static int vmdkGetSector(PVMDKIMAGE pImage, PVDIOCTX pIoCtx,
4510 PVMDKEXTENT pExtent, uint64_t uSector,
4511 uint64_t *puExtentSector)
4512{
4513 PVMDKGTCACHE pCache = pImage->pGTCache;
4514 uint64_t uGDIndex, uGTSector, uGTBlock;
4515 uint32_t uGTHash, uGTBlockIndex;
4516 PVMDKGTCACHEENTRY pGTCacheEntry;
4517 uint32_t aGTDataTmp[VMDK_GT_CACHELINE_SIZE];
4518 int rc;
4519
4520 /* For newly created and readonly/sequentially opened streamOptimized
4521 * images this must be a no-op, as the grain directory is not there. */
4522 if ( ( pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED
4523 && pExtent->uAppendPosition)
4524 || ( pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED
4525 && pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY
4526 && pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL))
4527 {
4528 *puExtentSector = 0;
4529 return VINF_SUCCESS;
4530 }
4531
4532 uGDIndex = uSector / pExtent->cSectorsPerGDE;
4533 if (uGDIndex >= pExtent->cGDEntries)
4534 return VERR_OUT_OF_RANGE;
4535 uGTSector = pExtent->pGD[uGDIndex];
4536 if (!uGTSector)
4537 {
4538 /* There is no grain table referenced by this grain directory
4539 * entry. So there is absolutely no data in this area. */
4540 *puExtentSector = 0;
4541 return VINF_SUCCESS;
4542 }
4543
4544 uGTBlock = uSector / (pExtent->cSectorsPerGrain * VMDK_GT_CACHELINE_SIZE);
4545 uGTHash = vmdkGTCacheHash(pCache, uGTBlock, pExtent->uExtent);
4546 pGTCacheEntry = &pCache->aGTCache[uGTHash];
4547 if ( pGTCacheEntry->uExtent != pExtent->uExtent
4548 || pGTCacheEntry->uGTBlock != uGTBlock)
4549 {
4550 /* Cache miss, fetch data from disk. */
4551 PVDMETAXFER pMetaXfer;
4552 rc = vdIfIoIntFileReadMeta(pImage->pIfIo, pExtent->pFile->pStorage,
4553 VMDK_SECTOR2BYTE(uGTSector) + (uGTBlock % (pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE)) * sizeof(aGTDataTmp),
4554 aGTDataTmp, sizeof(aGTDataTmp), pIoCtx, &pMetaXfer, NULL, NULL);
4555 if (RT_FAILURE(rc))
4556 return rc;
4557 /* We can release the metadata transfer immediately. */
4558 vdIfIoIntMetaXferRelease(pImage->pIfIo, pMetaXfer);
4559 pGTCacheEntry->uExtent = pExtent->uExtent;
4560 pGTCacheEntry->uGTBlock = uGTBlock;
4561 for (unsigned i = 0; i < VMDK_GT_CACHELINE_SIZE; i++)
4562 pGTCacheEntry->aGTData[i] = RT_LE2H_U32(aGTDataTmp[i]);
4563 }
4564 uGTBlockIndex = (uSector / pExtent->cSectorsPerGrain) % VMDK_GT_CACHELINE_SIZE;
4565 uint32_t uGrainSector = pGTCacheEntry->aGTData[uGTBlockIndex];
4566 if (uGrainSector)
4567 *puExtentSector = uGrainSector + uSector % pExtent->cSectorsPerGrain;
4568 else
4569 *puExtentSector = 0;
4570 return VINF_SUCCESS;
4571}
4572
4573/**
4574 * Internal. Writes the grain and also if necessary the grain tables.
4575 * Uses the grain table cache as a true grain table.
4576 */
4577static int vmdkStreamAllocGrain(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
4578 uint64_t uSector, PVDIOCTX pIoCtx,
4579 uint64_t cbWrite)
4580{
4581 uint32_t uGrain;
4582 uint32_t uGDEntry, uLastGDEntry;
4583 uint32_t cbGrain = 0;
4584 uint32_t uCacheLine, uCacheEntry;
4585 const void *pData;
4586 int rc;
4587
4588 /* Very strict requirements: always write at least one full grain, with
4589 * proper alignment. Everything else would require reading of already
4590 * written data, which we don't support for obvious reasons. The only
4591 * exception is the last grain, and only if the image size specifies
4592 * that only some portion holds data. In any case the write must be
4593 * within the image limits, no "overshoot" allowed. */
4594 if ( cbWrite == 0
4595 || ( cbWrite < VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain)
4596 && pExtent->cNominalSectors - uSector >= pExtent->cSectorsPerGrain)
4597 || uSector % pExtent->cSectorsPerGrain
4598 || uSector + VMDK_BYTE2SECTOR(cbWrite) > pExtent->cNominalSectors)
4599 return VERR_INVALID_PARAMETER;
4600
4601 /* Clip write range to at most the rest of the grain. */
4602 cbWrite = RT_MIN(cbWrite, VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain - uSector % pExtent->cSectorsPerGrain));
4603
4604 /* Do not allow to go back. */
4605 uGrain = uSector / pExtent->cSectorsPerGrain;
4606 uCacheLine = uGrain % pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE;
4607 uCacheEntry = uGrain % VMDK_GT_CACHELINE_SIZE;
4608 uGDEntry = uGrain / pExtent->cGTEntries;
4609 uLastGDEntry = pExtent->uLastGrainAccess / pExtent->cGTEntries;
4610 if (uGrain < pExtent->uLastGrainAccess)
4611 return VERR_VD_VMDK_INVALID_WRITE;
4612
4613 /* Zero byte write optimization. Since we don't tell VBoxHDD that we need
4614 * to allocate something, we also need to detect the situation ourself. */
4615 if ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_HONOR_ZEROES)
4616 && vdIfIoIntIoCtxIsZero(pImage->pIfIo, pIoCtx, cbWrite, true /* fAdvance */))
4617 return VINF_SUCCESS;
4618
4619 if (uGDEntry != uLastGDEntry)
4620 {
4621 rc = vmdkStreamFlushGT(pImage, pExtent, uLastGDEntry);
4622 if (RT_FAILURE(rc))
4623 return rc;
4624 vmdkStreamClearGT(pImage, pExtent);
4625 for (uint32_t i = uLastGDEntry + 1; i < uGDEntry; i++)
4626 {
4627 rc = vmdkStreamFlushGT(pImage, pExtent, i);
4628 if (RT_FAILURE(rc))
4629 return rc;
4630 }
4631 }
4632
4633 uint64_t uFileOffset;
4634 uFileOffset = pExtent->uAppendPosition;
4635 if (!uFileOffset)
4636 return VERR_INTERNAL_ERROR;
4637 /* Align to sector, as the previous write could have been any size. */
4638 uFileOffset = RT_ALIGN_64(uFileOffset, 512);
4639
4640 /* Paranoia check: extent type, grain table buffer presence and
4641 * grain table buffer space. Also grain table entry must be clear. */
4642 if ( pExtent->enmType != VMDKETYPE_HOSTED_SPARSE
4643 || !pImage->pGTCache
4644 || pExtent->cGTEntries > VMDK_GT_CACHE_SIZE * VMDK_GT_CACHELINE_SIZE
4645 || pImage->pGTCache->aGTCache[uCacheLine].aGTData[uCacheEntry])
4646 return VERR_INTERNAL_ERROR;
4647
4648 /* Update grain table entry. */
4649 pImage->pGTCache->aGTCache[uCacheLine].aGTData[uCacheEntry] = VMDK_BYTE2SECTOR(uFileOffset);
4650
4651 if (cbWrite != VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain))
4652 {
4653 vdIfIoIntIoCtxCopyFrom(pImage->pIfIo, pIoCtx, pExtent->pvGrain, cbWrite);
4654 memset((char *)pExtent->pvGrain + cbWrite, '\0',
4655 VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain) - cbWrite);
4656 pData = pExtent->pvGrain;
4657 }
4658 else
4659 {
4660 RTSGSEG Segment;
4661 unsigned cSegments = 1;
4662 size_t cbSeg = 0;
4663
4664 cbSeg = vdIfIoIntIoCtxSegArrayCreate(pImage->pIfIo, pIoCtx, &Segment,
4665 &cSegments, VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
4666 Assert(cbSeg == VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
4667 pData = Segment.pvSeg;
4668 }
4669 rc = vmdkFileDeflateSync(pImage, pExtent, uFileOffset, pData,
4670 VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain),
4671 uSector, &cbGrain);
4672 if (RT_FAILURE(rc))
4673 {
4674 pExtent->uGrainSectorAbs = 0;
4675 AssertRC(rc);
4676 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write compressed data block in '%s'"), pExtent->pszFullname);
4677 }
4678 pExtent->uLastGrainAccess = uGrain;
4679 pExtent->uAppendPosition += cbGrain;
4680
4681 return rc;
4682}
4683
4684/**
4685 * Internal: Updates the grain table during grain allocation.
4686 */
4687static int vmdkAllocGrainGTUpdate(PVMDKIMAGE pImage, PVMDKEXTENT pExtent, PVDIOCTX pIoCtx,
4688 PVMDKGRAINALLOCASYNC pGrainAlloc)
4689{
4690 int rc = VINF_SUCCESS;
4691 PVMDKGTCACHE pCache = pImage->pGTCache;
4692 uint32_t aGTDataTmp[VMDK_GT_CACHELINE_SIZE];
4693 uint32_t uGTHash, uGTBlockIndex;
4694 uint64_t uGTSector, uRGTSector, uGTBlock;
4695 uint64_t uSector = pGrainAlloc->uSector;
4696 PVMDKGTCACHEENTRY pGTCacheEntry;
4697
4698 LogFlowFunc(("pImage=%#p pExtent=%#p pCache=%#p pIoCtx=%#p pGrainAlloc=%#p\n",
4699 pImage, pExtent, pCache, pIoCtx, pGrainAlloc));
4700
4701 uGTSector = pGrainAlloc->uGTSector;
4702 uRGTSector = pGrainAlloc->uRGTSector;
4703 LogFlow(("uGTSector=%llu uRGTSector=%llu\n", uGTSector, uRGTSector));
4704
4705 /* Update the grain table (and the cache). */
4706 uGTBlock = uSector / (pExtent->cSectorsPerGrain * VMDK_GT_CACHELINE_SIZE);
4707 uGTHash = vmdkGTCacheHash(pCache, uGTBlock, pExtent->uExtent);
4708 pGTCacheEntry = &pCache->aGTCache[uGTHash];
4709 if ( pGTCacheEntry->uExtent != pExtent->uExtent
4710 || pGTCacheEntry->uGTBlock != uGTBlock)
4711 {
4712 /* Cache miss, fetch data from disk. */
4713 LogFlow(("Cache miss, fetch data from disk\n"));
4714 PVDMETAXFER pMetaXfer = NULL;
4715 rc = vdIfIoIntFileReadMeta(pImage->pIfIo, pExtent->pFile->pStorage,
4716 VMDK_SECTOR2BYTE(uGTSector) + (uGTBlock % (pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE)) * sizeof(aGTDataTmp),
4717 aGTDataTmp, sizeof(aGTDataTmp), pIoCtx,
4718 &pMetaXfer, vmdkAllocGrainComplete, pGrainAlloc);
4719 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4720 {
4721 pGrainAlloc->cIoXfersPending++;
4722 pGrainAlloc->fGTUpdateNeeded = true;
4723 /* Leave early, we will be called again after the read completed. */
4724 LogFlowFunc(("Metadata read in progress, leaving\n"));
4725 return rc;
4726 }
4727 else if (RT_FAILURE(rc))
4728 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot read allocated grain table entry in '%s'"), pExtent->pszFullname);
4729 vdIfIoIntMetaXferRelease(pImage->pIfIo, pMetaXfer);
4730 pGTCacheEntry->uExtent = pExtent->uExtent;
4731 pGTCacheEntry->uGTBlock = uGTBlock;
4732 for (unsigned i = 0; i < VMDK_GT_CACHELINE_SIZE; i++)
4733 pGTCacheEntry->aGTData[i] = RT_LE2H_U32(aGTDataTmp[i]);
4734 }
4735 else
4736 {
4737 /* Cache hit. Convert grain table block back to disk format, otherwise
4738 * the code below will write garbage for all but the updated entry. */
4739 for (unsigned i = 0; i < VMDK_GT_CACHELINE_SIZE; i++)
4740 aGTDataTmp[i] = RT_H2LE_U32(pGTCacheEntry->aGTData[i]);
4741 }
4742 pGrainAlloc->fGTUpdateNeeded = false;
4743 uGTBlockIndex = (uSector / pExtent->cSectorsPerGrain) % VMDK_GT_CACHELINE_SIZE;
4744 aGTDataTmp[uGTBlockIndex] = RT_H2LE_U32(VMDK_BYTE2SECTOR(pGrainAlloc->uGrainOffset));
4745 pGTCacheEntry->aGTData[uGTBlockIndex] = VMDK_BYTE2SECTOR(pGrainAlloc->uGrainOffset);
4746 /* Update grain table on disk. */
4747 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
4748 VMDK_SECTOR2BYTE(uGTSector) + (uGTBlock % (pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE)) * sizeof(aGTDataTmp),
4749 aGTDataTmp, sizeof(aGTDataTmp), pIoCtx,
4750 vmdkAllocGrainComplete, pGrainAlloc);
4751 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4752 pGrainAlloc->cIoXfersPending++;
4753 else if (RT_FAILURE(rc))
4754 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write updated grain table in '%s'"), pExtent->pszFullname);
4755 if (pExtent->pRGD)
4756 {
4757 /* Update backup grain table on disk. */
4758 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
4759 VMDK_SECTOR2BYTE(uRGTSector) + (uGTBlock % (pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE)) * sizeof(aGTDataTmp),
4760 aGTDataTmp, sizeof(aGTDataTmp), pIoCtx,
4761 vmdkAllocGrainComplete, pGrainAlloc);
4762 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4763 pGrainAlloc->cIoXfersPending++;
4764 else if (RT_FAILURE(rc))
4765 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write updated backup grain table in '%s'"), pExtent->pszFullname);
4766 }
4767#ifdef VBOX_WITH_VMDK_ESX
4768 if (RT_SUCCESS(rc) && pExtent->enmType == VMDKETYPE_ESX_SPARSE)
4769 {
4770 pExtent->uFreeSector = uGTSector + VMDK_BYTE2SECTOR(cbWrite);
4771 pExtent->fMetaDirty = true;
4772 }
4773#endif /* VBOX_WITH_VMDK_ESX */
4774
4775 LogFlowFunc(("leaving rc=%Rrc\n", rc));
4776
4777 return rc;
4778}
4779
4780/**
4781 * Internal - complete the grain allocation by updating disk grain table if required.
4782 */
4783static int vmdkAllocGrainComplete(void *pBackendData, PVDIOCTX pIoCtx, void *pvUser, int rcReq)
4784{
4785 int rc = VINF_SUCCESS;
4786 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
4787 PVMDKGRAINALLOCASYNC pGrainAlloc = (PVMDKGRAINALLOCASYNC)pvUser;
4788 PVMDKEXTENT pExtent = pGrainAlloc->pExtent;
4789
4790 LogFlowFunc(("pBackendData=%#p pIoCtx=%#p pvUser=%#p rcReq=%Rrc\n",
4791 pBackendData, pIoCtx, pvUser, rcReq));
4792
4793 pGrainAlloc->cIoXfersPending--;
4794 if (!pGrainAlloc->cIoXfersPending && pGrainAlloc->fGTUpdateNeeded)
4795 rc = vmdkAllocGrainGTUpdate(pImage, pGrainAlloc->pExtent, pIoCtx, pGrainAlloc);
4796
4797 if (!pGrainAlloc->cIoXfersPending)
4798 {
4799 /* Grain allocation completed. */
4800 RTMemFree(pGrainAlloc);
4801 }
4802
4803 LogFlowFunc(("Leaving rc=%Rrc\n", rc));
4804 return rc;
4805}
4806
4807/**
4808 * Internal. Allocates a new grain table (if necessary).
4809 */
4810static int vmdkAllocGrain(PVMDKIMAGE pImage, PVMDKEXTENT pExtent, PVDIOCTX pIoCtx,
4811 uint64_t uSector, uint64_t cbWrite)
4812{
4813 PVMDKGTCACHE pCache = pImage->pGTCache;
4814 uint64_t uGDIndex, uGTSector, uRGTSector;
4815 uint64_t uFileOffset;
4816 PVMDKGRAINALLOCASYNC pGrainAlloc = NULL;
4817 int rc;
4818
4819 LogFlowFunc(("pCache=%#p pExtent=%#p pIoCtx=%#p uSector=%llu cbWrite=%llu\n",
4820 pCache, pExtent, pIoCtx, uSector, cbWrite));
4821
4822 pGrainAlloc = (PVMDKGRAINALLOCASYNC)RTMemAllocZ(sizeof(VMDKGRAINALLOCASYNC));
4823 if (!pGrainAlloc)
4824 return VERR_NO_MEMORY;
4825
4826 pGrainAlloc->pExtent = pExtent;
4827 pGrainAlloc->uSector = uSector;
4828
4829 uGDIndex = uSector / pExtent->cSectorsPerGDE;
4830 if (uGDIndex >= pExtent->cGDEntries)
4831 {
4832 RTMemFree(pGrainAlloc);
4833 return VERR_OUT_OF_RANGE;
4834 }
4835 uGTSector = pExtent->pGD[uGDIndex];
4836 if (pExtent->pRGD)
4837 uRGTSector = pExtent->pRGD[uGDIndex];
4838 else
4839 uRGTSector = 0; /**< avoid compiler warning */
4840 if (!uGTSector)
4841 {
4842 LogFlow(("Allocating new grain table\n"));
4843
4844 /* There is no grain table referenced by this grain directory
4845 * entry. So there is absolutely no data in this area. Allocate
4846 * a new grain table and put the reference to it in the GDs. */
4847 uFileOffset = pExtent->uAppendPosition;
4848 if (!uFileOffset)
4849 return VERR_INTERNAL_ERROR;
4850 Assert(!(uFileOffset % 512));
4851
4852 uFileOffset = RT_ALIGN_64(uFileOffset, 512);
4853 uGTSector = VMDK_BYTE2SECTOR(uFileOffset);
4854
4855 /* Normally the grain table is preallocated for hosted sparse extents
4856 * that support more than 32 bit sector numbers. So this shouldn't
4857 * ever happen on a valid extent. */
4858 if (uGTSector > UINT32_MAX)
4859 return VERR_VD_VMDK_INVALID_HEADER;
4860
4861 /* Write grain table by writing the required number of grain table
4862 * cache chunks. Allocate memory dynamically here or we flood the
4863 * metadata cache with very small entries. */
4864 size_t cbGTDataTmp = pExtent->cGTEntries * sizeof(uint32_t);
4865 uint32_t *paGTDataTmp = (uint32_t *)RTMemTmpAllocZ(cbGTDataTmp);
4866
4867 if (!paGTDataTmp)
4868 return VERR_NO_MEMORY;
4869
4870 memset(paGTDataTmp, '\0', cbGTDataTmp);
4871 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
4872 VMDK_SECTOR2BYTE(uGTSector),
4873 paGTDataTmp, cbGTDataTmp, pIoCtx,
4874 vmdkAllocGrainComplete, pGrainAlloc);
4875 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4876 pGrainAlloc->cIoXfersPending++;
4877 else if (RT_FAILURE(rc))
4878 {
4879 RTMemTmpFree(paGTDataTmp);
4880 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write grain table allocation in '%s'"), pExtent->pszFullname);
4881 }
4882 pExtent->uAppendPosition = RT_ALIGN_64( pExtent->uAppendPosition
4883 + cbGTDataTmp, 512);
4884
4885 if (pExtent->pRGD)
4886 {
4887 AssertReturn(!uRGTSector, VERR_VD_VMDK_INVALID_HEADER);
4888 uFileOffset = pExtent->uAppendPosition;
4889 if (!uFileOffset)
4890 return VERR_INTERNAL_ERROR;
4891 Assert(!(uFileOffset % 512));
4892 uRGTSector = VMDK_BYTE2SECTOR(uFileOffset);
4893
4894 /* Normally the redundant grain table is preallocated for hosted
4895 * sparse extents that support more than 32 bit sector numbers. So
4896 * this shouldn't ever happen on a valid extent. */
4897 if (uRGTSector > UINT32_MAX)
4898 {
4899 RTMemTmpFree(paGTDataTmp);
4900 return VERR_VD_VMDK_INVALID_HEADER;
4901 }
4902
4903 /* Write grain table by writing the required number of grain table
4904 * cache chunks. Allocate memory dynamically here or we flood the
4905 * metadata cache with very small entries. */
4906 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
4907 VMDK_SECTOR2BYTE(uRGTSector),
4908 paGTDataTmp, cbGTDataTmp, pIoCtx,
4909 vmdkAllocGrainComplete, pGrainAlloc);
4910 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4911 pGrainAlloc->cIoXfersPending++;
4912 else if (RT_FAILURE(rc))
4913 {
4914 RTMemTmpFree(paGTDataTmp);
4915 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write backup grain table allocation in '%s'"), pExtent->pszFullname);
4916 }
4917
4918 pExtent->uAppendPosition = pExtent->uAppendPosition + cbGTDataTmp;
4919 }
4920
4921 RTMemTmpFree(paGTDataTmp);
4922
4923 /* Update the grain directory on disk (doing it before writing the
4924 * grain table will result in a garbled extent if the operation is
4925 * aborted for some reason. Otherwise the worst that can happen is
4926 * some unused sectors in the extent. */
4927 uint32_t uGTSectorLE = RT_H2LE_U64(uGTSector);
4928 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
4929 VMDK_SECTOR2BYTE(pExtent->uSectorGD) + uGDIndex * sizeof(uGTSectorLE),
4930 &uGTSectorLE, sizeof(uGTSectorLE), pIoCtx,
4931 vmdkAllocGrainComplete, pGrainAlloc);
4932 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4933 pGrainAlloc->cIoXfersPending++;
4934 else if (RT_FAILURE(rc))
4935 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write grain directory entry in '%s'"), pExtent->pszFullname);
4936 if (pExtent->pRGD)
4937 {
4938 uint32_t uRGTSectorLE = RT_H2LE_U64(uRGTSector);
4939 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
4940 VMDK_SECTOR2BYTE(pExtent->uSectorRGD) + uGDIndex * sizeof(uGTSectorLE),
4941 &uRGTSectorLE, sizeof(uRGTSectorLE), pIoCtx,
4942 vmdkAllocGrainComplete, pGrainAlloc);
4943 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4944 pGrainAlloc->cIoXfersPending++;
4945 else if (RT_FAILURE(rc))
4946 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write backup grain directory entry in '%s'"), pExtent->pszFullname);
4947 }
4948
4949 /* As the final step update the in-memory copy of the GDs. */
4950 pExtent->pGD[uGDIndex] = uGTSector;
4951 if (pExtent->pRGD)
4952 pExtent->pRGD[uGDIndex] = uRGTSector;
4953 }
4954
4955 LogFlow(("uGTSector=%llu uRGTSector=%llu\n", uGTSector, uRGTSector));
4956 pGrainAlloc->uGTSector = uGTSector;
4957 pGrainAlloc->uRGTSector = uRGTSector;
4958
4959 uFileOffset = pExtent->uAppendPosition;
4960 if (!uFileOffset)
4961 return VERR_INTERNAL_ERROR;
4962 Assert(!(uFileOffset % 512));
4963
4964 pGrainAlloc->uGrainOffset = uFileOffset;
4965
4966 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
4967 {
4968 AssertMsgReturn(vdIfIoIntIoCtxIsSynchronous(pImage->pIfIo, pIoCtx),
4969 ("Accesses to stream optimized images must be synchronous\n"),
4970 VERR_INVALID_STATE);
4971
4972 if (cbWrite != VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain))
4973 return vdIfError(pImage->pIfError, VERR_INTERNAL_ERROR, RT_SRC_POS, N_("VMDK: not enough data for a compressed data block in '%s'"), pExtent->pszFullname);
4974
4975 /* Invalidate cache, just in case some code incorrectly allows mixing
4976 * of reads and writes. Normally shouldn't be needed. */
4977 pExtent->uGrainSectorAbs = 0;
4978
4979 /* Write compressed data block and the markers. */
4980 uint32_t cbGrain = 0;
4981 size_t cbSeg = 0;
4982 RTSGSEG Segment;
4983 unsigned cSegments = 1;
4984
4985 cbSeg = vdIfIoIntIoCtxSegArrayCreate(pImage->pIfIo, pIoCtx, &Segment,
4986 &cSegments, cbWrite);
4987 Assert(cbSeg == cbWrite);
4988
4989 rc = vmdkFileDeflateSync(pImage, pExtent, uFileOffset,
4990 Segment.pvSeg, cbWrite, uSector, &cbGrain);
4991 if (RT_FAILURE(rc))
4992 {
4993 AssertRC(rc);
4994 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write allocated compressed data block in '%s'"), pExtent->pszFullname);
4995 }
4996 pExtent->uLastGrainAccess = uSector / pExtent->cSectorsPerGrain;
4997 pExtent->uAppendPosition += cbGrain;
4998 }
4999 else
5000 {
5001 /* Write the data. Always a full grain, or we're in big trouble. */
5002 rc = vdIfIoIntFileWriteUser(pImage->pIfIo, pExtent->pFile->pStorage,
5003 uFileOffset, pIoCtx, cbWrite,
5004 vmdkAllocGrainComplete, pGrainAlloc);
5005 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
5006 pGrainAlloc->cIoXfersPending++;
5007 else if (RT_FAILURE(rc))
5008 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write allocated data block in '%s'"), pExtent->pszFullname);
5009
5010 pExtent->uAppendPosition += cbWrite;
5011 }
5012
5013 rc = vmdkAllocGrainGTUpdate(pImage, pExtent, pIoCtx, pGrainAlloc);
5014
5015 if (!pGrainAlloc->cIoXfersPending)
5016 {
5017 /* Grain allocation completed. */
5018 RTMemFree(pGrainAlloc);
5019 }
5020
5021 LogFlowFunc(("leaving rc=%Rrc\n", rc));
5022
5023 return rc;
5024}
5025
5026/**
5027 * Internal. Reads the contents by sequentially going over the compressed
5028 * grains (hoping that they are in sequence).
5029 */
5030static int vmdkStreamReadSequential(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
5031 uint64_t uSector, PVDIOCTX pIoCtx,
5032 uint64_t cbRead)
5033{
5034 int rc;
5035
5036 LogFlowFunc(("pImage=%#p pExtent=%#p uSector=%llu pIoCtx=%#p cbRead=%llu\n",
5037 pImage, pExtent, uSector, pIoCtx, cbRead));
5038
5039 AssertMsgReturn(vdIfIoIntIoCtxIsSynchronous(pImage->pIfIo, pIoCtx),
5040 ("Async I/O not supported for sequential stream optimized images\n"),
5041 VERR_INVALID_STATE);
5042
5043 /* Do not allow to go back. */
5044 uint32_t uGrain = uSector / pExtent->cSectorsPerGrain;
5045 if (uGrain < pExtent->uLastGrainAccess)
5046 return VERR_VD_VMDK_INVALID_STATE;
5047 pExtent->uLastGrainAccess = uGrain;
5048
5049 /* After a previous error do not attempt to recover, as it would need
5050 * seeking (in the general case backwards which is forbidden). */
5051 if (!pExtent->uGrainSectorAbs)
5052 return VERR_VD_VMDK_INVALID_STATE;
5053
5054 /* Check if we need to read something from the image or if what we have
5055 * in the buffer is good to fulfill the request. */
5056 if (!pExtent->cbGrainStreamRead || uGrain > pExtent->uGrain)
5057 {
5058 uint32_t uGrainSectorAbs = pExtent->uGrainSectorAbs
5059 + VMDK_BYTE2SECTOR(pExtent->cbGrainStreamRead);
5060
5061 /* Get the marker from the next data block - and skip everything which
5062 * is not a compressed grain. If it's a compressed grain which is for
5063 * the requested sector (or after), read it. */
5064 VMDKMARKER Marker;
5065 do
5066 {
5067 RT_ZERO(Marker);
5068 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
5069 VMDK_SECTOR2BYTE(uGrainSectorAbs),
5070 &Marker, RT_OFFSETOF(VMDKMARKER, uType));
5071 if (RT_FAILURE(rc))
5072 return rc;
5073 Marker.uSector = RT_LE2H_U64(Marker.uSector);
5074 Marker.cbSize = RT_LE2H_U32(Marker.cbSize);
5075
5076 if (Marker.cbSize == 0)
5077 {
5078 /* A marker for something else than a compressed grain. */
5079 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
5080 VMDK_SECTOR2BYTE(uGrainSectorAbs)
5081 + RT_OFFSETOF(VMDKMARKER, uType),
5082 &Marker.uType, sizeof(Marker.uType));
5083 if (RT_FAILURE(rc))
5084 return rc;
5085 Marker.uType = RT_LE2H_U32(Marker.uType);
5086 switch (Marker.uType)
5087 {
5088 case VMDK_MARKER_EOS:
5089 uGrainSectorAbs++;
5090 /* Read (or mostly skip) to the end of file. Uses the
5091 * Marker (LBA sector) as it is unused anyway. This
5092 * makes sure that really everything is read in the
5093 * success case. If this read fails it means the image
5094 * is truncated, but this is harmless so ignore. */
5095 vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
5096 VMDK_SECTOR2BYTE(uGrainSectorAbs)
5097 + 511,
5098 &Marker.uSector, 1);
5099 break;
5100 case VMDK_MARKER_GT:
5101 uGrainSectorAbs += 1 + VMDK_BYTE2SECTOR(pExtent->cGTEntries * sizeof(uint32_t));
5102 break;
5103 case VMDK_MARKER_GD:
5104 uGrainSectorAbs += 1 + VMDK_BYTE2SECTOR(RT_ALIGN(pExtent->cGDEntries * sizeof(uint32_t), 512));
5105 break;
5106 case VMDK_MARKER_FOOTER:
5107 uGrainSectorAbs += 2;
5108 break;
5109 case VMDK_MARKER_UNSPECIFIED:
5110 /* Skip over the contents of the unspecified marker
5111 * type 4 which exists in some vSphere created files. */
5112 /** @todo figure out what the payload means. */
5113 uGrainSectorAbs += 1;
5114 break;
5115 default:
5116 AssertMsgFailed(("VMDK: corrupted marker, type=%#x\n", Marker.uType));
5117 pExtent->uGrainSectorAbs = 0;
5118 return VERR_VD_VMDK_INVALID_STATE;
5119 }
5120 pExtent->cbGrainStreamRead = 0;
5121 }
5122 else
5123 {
5124 /* A compressed grain marker. If it is at/after what we're
5125 * interested in read and decompress data. */
5126 if (uSector > Marker.uSector + pExtent->cSectorsPerGrain)
5127 {
5128 uGrainSectorAbs += VMDK_BYTE2SECTOR(RT_ALIGN(Marker.cbSize + RT_OFFSETOF(VMDKMARKER, uType), 512));
5129 continue;
5130 }
5131 uint64_t uLBA = 0;
5132 uint32_t cbGrainStreamRead = 0;
5133 rc = vmdkFileInflateSync(pImage, pExtent,
5134 VMDK_SECTOR2BYTE(uGrainSectorAbs),
5135 pExtent->pvGrain,
5136 VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain),
5137 &Marker, &uLBA, &cbGrainStreamRead);
5138 if (RT_FAILURE(rc))
5139 {
5140 pExtent->uGrainSectorAbs = 0;
5141 return rc;
5142 }
5143 if ( pExtent->uGrain
5144 && uLBA / pExtent->cSectorsPerGrain <= pExtent->uGrain)
5145 {
5146 pExtent->uGrainSectorAbs = 0;
5147 return VERR_VD_VMDK_INVALID_STATE;
5148 }
5149 pExtent->uGrain = uLBA / pExtent->cSectorsPerGrain;
5150 pExtent->cbGrainStreamRead = cbGrainStreamRead;
5151 break;
5152 }
5153 } while (Marker.uType != VMDK_MARKER_EOS);
5154
5155 pExtent->uGrainSectorAbs = uGrainSectorAbs;
5156
5157 if (!pExtent->cbGrainStreamRead && Marker.uType == VMDK_MARKER_EOS)
5158 {
5159 pExtent->uGrain = UINT32_MAX;
5160 /* Must set a non-zero value for pExtent->cbGrainStreamRead or
5161 * the next read would try to get more data, and we're at EOF. */
5162 pExtent->cbGrainStreamRead = 1;
5163 }
5164 }
5165
5166 if (pExtent->uGrain > uSector / pExtent->cSectorsPerGrain)
5167 {
5168 /* The next data block we have is not for this area, so just return
5169 * that there is no data. */
5170 LogFlowFunc(("returns VERR_VD_BLOCK_FREE\n"));
5171 return VERR_VD_BLOCK_FREE;
5172 }
5173
5174 uint32_t uSectorInGrain = uSector % pExtent->cSectorsPerGrain;
5175 vdIfIoIntIoCtxCopyTo(pImage->pIfIo, pIoCtx,
5176 (uint8_t *)pExtent->pvGrain + VMDK_SECTOR2BYTE(uSectorInGrain),
5177 cbRead);
5178 LogFlowFunc(("returns VINF_SUCCESS\n"));
5179 return VINF_SUCCESS;
5180}
5181
5182/**
5183 * Replaces a fragment of a string with the specified string.
5184 *
5185 * @returns Pointer to the allocated UTF-8 string.
5186 * @param pszWhere UTF-8 string to search in.
5187 * @param pszWhat UTF-8 string to search for.
5188 * @param pszByWhat UTF-8 string to replace the found string with.
5189 */
5190static char *vmdkStrReplace(const char *pszWhere, const char *pszWhat,
5191 const char *pszByWhat)
5192{
5193 AssertPtr(pszWhere);
5194 AssertPtr(pszWhat);
5195 AssertPtr(pszByWhat);
5196 const char *pszFoundStr = strstr(pszWhere, pszWhat);
5197 if (!pszFoundStr)
5198 return NULL;
5199 size_t cFinal = strlen(pszWhere) + 1 + strlen(pszByWhat) - strlen(pszWhat);
5200 char *pszNewStr = (char *)RTMemAlloc(cFinal);
5201 if (pszNewStr)
5202 {
5203 char *pszTmp = pszNewStr;
5204 memcpy(pszTmp, pszWhere, pszFoundStr - pszWhere);
5205 pszTmp += pszFoundStr - pszWhere;
5206 memcpy(pszTmp, pszByWhat, strlen(pszByWhat));
5207 pszTmp += strlen(pszByWhat);
5208 strcpy(pszTmp, pszFoundStr + strlen(pszWhat));
5209 }
5210 return pszNewStr;
5211}
5212
5213
5214/** @copydoc VBOXHDDBACKEND::pfnCheckIfValid */
5215static int vmdkCheckIfValid(const char *pszFilename, PVDINTERFACE pVDIfsDisk,
5216 PVDINTERFACE pVDIfsImage, VDTYPE *penmType)
5217{
5218 LogFlowFunc(("pszFilename=\"%s\" pVDIfsDisk=%#p pVDIfsImage=%#p penmType=%#p\n",
5219 pszFilename, pVDIfsDisk, pVDIfsImage, penmType));
5220 int rc = VINF_SUCCESS;
5221 PVMDKIMAGE pImage;
5222
5223 if ( !pszFilename
5224 || !*pszFilename
5225 || strchr(pszFilename, '"'))
5226 {
5227 rc = VERR_INVALID_PARAMETER;
5228 goto out;
5229 }
5230
5231 pImage = (PVMDKIMAGE)RTMemAllocZ(sizeof(VMDKIMAGE));
5232 if (!pImage)
5233 {
5234 rc = VERR_NO_MEMORY;
5235 goto out;
5236 }
5237 pImage->pszFilename = pszFilename;
5238 pImage->pFile = NULL;
5239 pImage->pExtents = NULL;
5240 pImage->pFiles = NULL;
5241 pImage->pGTCache = NULL;
5242 pImage->pDescData = NULL;
5243 pImage->pVDIfsDisk = pVDIfsDisk;
5244 pImage->pVDIfsImage = pVDIfsImage;
5245 /** @todo speed up this test open (VD_OPEN_FLAGS_INFO) by skipping as
5246 * much as possible in vmdkOpenImage. */
5247 rc = vmdkOpenImage(pImage, VD_OPEN_FLAGS_INFO | VD_OPEN_FLAGS_READONLY);
5248 vmdkFreeImage(pImage, false);
5249 RTMemFree(pImage);
5250
5251 if (RT_SUCCESS(rc))
5252 *penmType = VDTYPE_HDD;
5253
5254out:
5255 LogFlowFunc(("returns %Rrc\n", rc));
5256 return rc;
5257}
5258
5259/** @copydoc VBOXHDDBACKEND::pfnOpen */
5260static int vmdkOpen(const char *pszFilename, unsigned uOpenFlags,
5261 PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
5262 VDTYPE enmType, void **ppBackendData)
5263{
5264 LogFlowFunc(("pszFilename=\"%s\" uOpenFlags=%#x pVDIfsDisk=%#p pVDIfsImage=%#p ppBackendData=%#p\n", pszFilename, uOpenFlags, pVDIfsDisk, pVDIfsImage, ppBackendData));
5265 int rc;
5266 PVMDKIMAGE pImage;
5267
5268 /* Check open flags. All valid flags are supported. */
5269 if (uOpenFlags & ~VD_OPEN_FLAGS_MASK)
5270 {
5271 rc = VERR_INVALID_PARAMETER;
5272 goto out;
5273 }
5274
5275 /* Check remaining arguments. */
5276 if ( !VALID_PTR(pszFilename)
5277 || !*pszFilename
5278 || strchr(pszFilename, '"'))
5279 {
5280 rc = VERR_INVALID_PARAMETER;
5281 goto out;
5282 }
5283
5284 pImage = (PVMDKIMAGE)RTMemAllocZ(sizeof(VMDKIMAGE));
5285 if (!pImage)
5286 {
5287 rc = VERR_NO_MEMORY;
5288 goto out;
5289 }
5290 pImage->pszFilename = pszFilename;
5291 pImage->pFile = NULL;
5292 pImage->pExtents = NULL;
5293 pImage->pFiles = NULL;
5294 pImage->pGTCache = NULL;
5295 pImage->pDescData = NULL;
5296 pImage->pVDIfsDisk = pVDIfsDisk;
5297 pImage->pVDIfsImage = pVDIfsImage;
5298
5299 rc = vmdkOpenImage(pImage, uOpenFlags);
5300 if (RT_SUCCESS(rc))
5301 *ppBackendData = pImage;
5302 else
5303 RTMemFree(pImage);
5304
5305out:
5306 LogFlowFunc(("returns %Rrc (pBackendData=%#p)\n", rc, *ppBackendData));
5307 return rc;
5308}
5309
5310/** @copydoc VBOXHDDBACKEND::pfnCreate */
5311static int vmdkCreate(const char *pszFilename, uint64_t cbSize,
5312 unsigned uImageFlags, const char *pszComment,
5313 PCVDGEOMETRY pPCHSGeometry, PCVDGEOMETRY pLCHSGeometry,
5314 PCRTUUID pUuid, unsigned uOpenFlags,
5315 unsigned uPercentStart, unsigned uPercentSpan,
5316 PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
5317 PVDINTERFACE pVDIfsOperation, void **ppBackendData)
5318{
5319 LogFlowFunc(("pszFilename=\"%s\" cbSize=%llu uImageFlags=%#x pszComment=\"%s\" pPCHSGeometry=%#p pLCHSGeometry=%#p Uuid=%RTuuid uOpenFlags=%#x uPercentStart=%u uPercentSpan=%u pVDIfsDisk=%#p pVDIfsImage=%#p pVDIfsOperation=%#p ppBackendData=%#p\n", pszFilename, cbSize, uImageFlags, pszComment, pPCHSGeometry, pLCHSGeometry, pUuid, uOpenFlags, uPercentStart, uPercentSpan, pVDIfsDisk, pVDIfsImage, pVDIfsOperation, ppBackendData));
5320 int rc;
5321 PVMDKIMAGE pImage;
5322
5323 PFNVDPROGRESS pfnProgress = NULL;
5324 void *pvUser = NULL;
5325 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
5326 if (pIfProgress)
5327 {
5328 pfnProgress = pIfProgress->pfnProgress;
5329 pvUser = pIfProgress->Core.pvUser;
5330 }
5331
5332 /* Check the image flags. */
5333 if ((uImageFlags & ~VD_VMDK_IMAGE_FLAGS_MASK) != 0)
5334 {
5335 rc = VERR_VD_INVALID_TYPE;
5336 goto out;
5337 }
5338
5339 /* Check open flags. All valid flags are supported. */
5340 if (uOpenFlags & ~VD_OPEN_FLAGS_MASK)
5341 {
5342 rc = VERR_INVALID_PARAMETER;
5343 goto out;
5344 }
5345
5346 /* Check size. Maximum 2TB-64K for sparse images, otherwise unlimited. */
5347 if ( !cbSize
5348 || (!(uImageFlags & VD_IMAGE_FLAGS_FIXED) && cbSize >= _1T * 2 - _64K))
5349 {
5350 rc = VERR_VD_INVALID_SIZE;
5351 goto out;
5352 }
5353
5354 /* Check remaining arguments. */
5355 if ( !VALID_PTR(pszFilename)
5356 || !*pszFilename
5357 || strchr(pszFilename, '"')
5358 || !VALID_PTR(pPCHSGeometry)
5359 || !VALID_PTR(pLCHSGeometry)
5360#ifndef VBOX_WITH_VMDK_ESX
5361 || ( uImageFlags & VD_VMDK_IMAGE_FLAGS_ESX
5362 && !(uImageFlags & VD_IMAGE_FLAGS_FIXED))
5363#endif
5364 || ( (uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
5365 && (uImageFlags & ~(VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED | VD_IMAGE_FLAGS_DIFF))))
5366 {
5367 rc = VERR_INVALID_PARAMETER;
5368 goto out;
5369 }
5370
5371 pImage = (PVMDKIMAGE)RTMemAllocZ(sizeof(VMDKIMAGE));
5372 if (!pImage)
5373 {
5374 rc = VERR_NO_MEMORY;
5375 goto out;
5376 }
5377 pImage->pszFilename = pszFilename;
5378 pImage->pFile = NULL;
5379 pImage->pExtents = NULL;
5380 pImage->pFiles = NULL;
5381 pImage->pGTCache = NULL;
5382 pImage->pDescData = NULL;
5383 pImage->pVDIfsDisk = pVDIfsDisk;
5384 pImage->pVDIfsImage = pVDIfsImage;
5385 /* Descriptors for split images can be pretty large, especially if the
5386 * filename is long. So prepare for the worst, and allocate quite some
5387 * memory for the descriptor in this case. */
5388 if (uImageFlags & VD_VMDK_IMAGE_FLAGS_SPLIT_2G)
5389 pImage->cbDescAlloc = VMDK_SECTOR2BYTE(200);
5390 else
5391 pImage->cbDescAlloc = VMDK_SECTOR2BYTE(20);
5392 pImage->pDescData = (char *)RTMemAllocZ(pImage->cbDescAlloc);
5393 if (!pImage->pDescData)
5394 {
5395 RTMemFree(pImage);
5396 rc = VERR_NO_MEMORY;
5397 goto out;
5398 }
5399
5400 rc = vmdkCreateImage(pImage, cbSize, uImageFlags, pszComment,
5401 pPCHSGeometry, pLCHSGeometry, pUuid,
5402 pfnProgress, pvUser, uPercentStart, uPercentSpan);
5403 if (RT_SUCCESS(rc))
5404 {
5405 /* So far the image is opened in read/write mode. Make sure the
5406 * image is opened in read-only mode if the caller requested that. */
5407 if (uOpenFlags & VD_OPEN_FLAGS_READONLY)
5408 {
5409 vmdkFreeImage(pImage, false);
5410 rc = vmdkOpenImage(pImage, uOpenFlags);
5411 if (RT_FAILURE(rc))
5412 goto out;
5413 }
5414 *ppBackendData = pImage;
5415 }
5416 else
5417 {
5418 RTMemFree(pImage->pDescData);
5419 RTMemFree(pImage);
5420 }
5421
5422out:
5423 LogFlowFunc(("returns %Rrc (pBackendData=%#p)\n", rc, *ppBackendData));
5424 return rc;
5425}
5426
5427/** @copydoc VBOXHDDBACKEND::pfnRename */
5428static int vmdkRename(void *pBackendData, const char *pszFilename)
5429{
5430 LogFlowFunc(("pBackendData=%#p pszFilename=%#p\n", pBackendData, pszFilename));
5431
5432 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
5433 int rc = VINF_SUCCESS;
5434 char **apszOldName = NULL;
5435 char **apszNewName = NULL;
5436 char **apszNewLines = NULL;
5437 char *pszOldDescName = NULL;
5438 bool fImageFreed = false;
5439 bool fEmbeddedDesc = false;
5440 unsigned cExtents = 0;
5441 char *pszNewBaseName = NULL;
5442 char *pszOldBaseName = NULL;
5443 char *pszNewFullName = NULL;
5444 char *pszOldFullName = NULL;
5445 const char *pszOldImageName;
5446 unsigned i, line;
5447 VMDKDESCRIPTOR DescriptorCopy;
5448 VMDKEXTENT ExtentCopy;
5449
5450 memset(&DescriptorCopy, 0, sizeof(DescriptorCopy));
5451
5452 /* Check arguments. */
5453 if ( !pImage
5454 || (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_RAWDISK)
5455 || !VALID_PTR(pszFilename)
5456 || !*pszFilename)
5457 {
5458 rc = VERR_INVALID_PARAMETER;
5459 goto out;
5460 }
5461
5462 cExtents = pImage->cExtents;
5463
5464 /*
5465 * Allocate an array to store both old and new names of renamed files
5466 * in case we have to roll back the changes. Arrays are initialized
5467 * with zeros. We actually save stuff when and if we change it.
5468 */
5469 apszOldName = (char **)RTMemTmpAllocZ((cExtents + 1) * sizeof(char*));
5470 apszNewName = (char **)RTMemTmpAllocZ((cExtents + 1) * sizeof(char*));
5471 apszNewLines = (char **)RTMemTmpAllocZ((cExtents) * sizeof(char*));
5472 if (!apszOldName || !apszNewName || !apszNewLines)
5473 {
5474 rc = VERR_NO_MEMORY;
5475 goto out;
5476 }
5477
5478 /* Save the descriptor size and position. */
5479 if (pImage->pDescData)
5480 {
5481 /* Separate descriptor file. */
5482 fEmbeddedDesc = false;
5483 }
5484 else
5485 {
5486 /* Embedded descriptor file. */
5487 ExtentCopy = pImage->pExtents[0];
5488 fEmbeddedDesc = true;
5489 }
5490 /* Save the descriptor content. */
5491 DescriptorCopy.cLines = pImage->Descriptor.cLines;
5492 for (i = 0; i < DescriptorCopy.cLines; i++)
5493 {
5494 DescriptorCopy.aLines[i] = RTStrDup(pImage->Descriptor.aLines[i]);
5495 if (!DescriptorCopy.aLines[i])
5496 {
5497 rc = VERR_NO_MEMORY;
5498 goto out;
5499 }
5500 }
5501
5502 /* Prepare both old and new base names used for string replacement. */
5503 pszNewBaseName = RTStrDup(RTPathFilename(pszFilename));
5504 RTPathStripSuffix(pszNewBaseName);
5505 pszOldBaseName = RTStrDup(RTPathFilename(pImage->pszFilename));
5506 RTPathStripSuffix(pszOldBaseName);
5507 /* Prepare both old and new full names used for string replacement. */
5508 pszNewFullName = RTStrDup(pszFilename);
5509 RTPathStripSuffix(pszNewFullName);
5510 pszOldFullName = RTStrDup(pImage->pszFilename);
5511 RTPathStripSuffix(pszOldFullName);
5512
5513 /* --- Up to this point we have not done any damage yet. --- */
5514
5515 /* Save the old name for easy access to the old descriptor file. */
5516 pszOldDescName = RTStrDup(pImage->pszFilename);
5517 /* Save old image name. */
5518 pszOldImageName = pImage->pszFilename;
5519
5520 /* Update the descriptor with modified extent names. */
5521 for (i = 0, line = pImage->Descriptor.uFirstExtent;
5522 i < cExtents;
5523 i++, line = pImage->Descriptor.aNextLines[line])
5524 {
5525 /* Assume that vmdkStrReplace will fail. */
5526 rc = VERR_NO_MEMORY;
5527 /* Update the descriptor. */
5528 apszNewLines[i] = vmdkStrReplace(pImage->Descriptor.aLines[line],
5529 pszOldBaseName, pszNewBaseName);
5530 if (!apszNewLines[i])
5531 goto rollback;
5532 pImage->Descriptor.aLines[line] = apszNewLines[i];
5533 }
5534 /* Make sure the descriptor gets written back. */
5535 pImage->Descriptor.fDirty = true;
5536 /* Flush the descriptor now, in case it is embedded. */
5537 vmdkFlushImage(pImage, NULL);
5538
5539 /* Close and rename/move extents. */
5540 for (i = 0; i < cExtents; i++)
5541 {
5542 PVMDKEXTENT pExtent = &pImage->pExtents[i];
5543 /* Compose new name for the extent. */
5544 apszNewName[i] = vmdkStrReplace(pExtent->pszFullname,
5545 pszOldFullName, pszNewFullName);
5546 if (!apszNewName[i])
5547 goto rollback;
5548 /* Close the extent file. */
5549 rc = vmdkFileClose(pImage, &pExtent->pFile, false);
5550 if (RT_FAILURE(rc))
5551 goto rollback;
5552
5553 /* Rename the extent file. */
5554 rc = vdIfIoIntFileMove(pImage->pIfIo, pExtent->pszFullname, apszNewName[i], 0);
5555 if (RT_FAILURE(rc))
5556 goto rollback;
5557 /* Remember the old name. */
5558 apszOldName[i] = RTStrDup(pExtent->pszFullname);
5559 }
5560 /* Release all old stuff. */
5561 rc = vmdkFreeImage(pImage, false);
5562 if (RT_FAILURE(rc))
5563 goto rollback;
5564
5565 fImageFreed = true;
5566
5567 /* Last elements of new/old name arrays are intended for
5568 * storing descriptor's names.
5569 */
5570 apszNewName[cExtents] = RTStrDup(pszFilename);
5571 /* Rename the descriptor file if it's separate. */
5572 if (!fEmbeddedDesc)
5573 {
5574 rc = vdIfIoIntFileMove(pImage->pIfIo, pImage->pszFilename, apszNewName[cExtents], 0);
5575 if (RT_FAILURE(rc))
5576 goto rollback;
5577 /* Save old name only if we may need to change it back. */
5578 apszOldName[cExtents] = RTStrDup(pszFilename);
5579 }
5580
5581 /* Update pImage with the new information. */
5582 pImage->pszFilename = pszFilename;
5583
5584 /* Open the new image. */
5585 rc = vmdkOpenImage(pImage, pImage->uOpenFlags);
5586 if (RT_SUCCESS(rc))
5587 goto out;
5588
5589rollback:
5590 /* Roll back all changes in case of failure. */
5591 if (RT_FAILURE(rc))
5592 {
5593 int rrc;
5594 if (!fImageFreed)
5595 {
5596 /*
5597 * Some extents may have been closed, close the rest. We will
5598 * re-open the whole thing later.
5599 */
5600 vmdkFreeImage(pImage, false);
5601 }
5602 /* Rename files back. */
5603 for (i = 0; i <= cExtents; i++)
5604 {
5605 if (apszOldName[i])
5606 {
5607 rrc = vdIfIoIntFileMove(pImage->pIfIo, apszNewName[i], apszOldName[i], 0);
5608 AssertRC(rrc);
5609 }
5610 }
5611 /* Restore the old descriptor. */
5612 PVMDKFILE pFile;
5613 rrc = vmdkFileOpen(pImage, &pFile, pszOldDescName,
5614 VDOpenFlagsToFileOpenFlags(VD_OPEN_FLAGS_NORMAL,
5615 false /* fCreate */));
5616 AssertRC(rrc);
5617 if (fEmbeddedDesc)
5618 {
5619 ExtentCopy.pFile = pFile;
5620 pImage->pExtents = &ExtentCopy;
5621 }
5622 else
5623 {
5624 /* Shouldn't be null for separate descriptor.
5625 * There will be no access to the actual content.
5626 */
5627 pImage->pDescData = pszOldDescName;
5628 pImage->pFile = pFile;
5629 }
5630 pImage->Descriptor = DescriptorCopy;
5631 vmdkWriteDescriptor(pImage, NULL);
5632 vmdkFileClose(pImage, &pFile, false);
5633 /* Get rid of the stuff we implanted. */
5634 pImage->pExtents = NULL;
5635 pImage->pFile = NULL;
5636 pImage->pDescData = NULL;
5637 /* Re-open the image back. */
5638 pImage->pszFilename = pszOldImageName;
5639 rrc = vmdkOpenImage(pImage, pImage->uOpenFlags);
5640 AssertRC(rrc);
5641 }
5642
5643out:
5644 for (i = 0; i < DescriptorCopy.cLines; i++)
5645 if (DescriptorCopy.aLines[i])
5646 RTStrFree(DescriptorCopy.aLines[i]);
5647 if (apszOldName)
5648 {
5649 for (i = 0; i <= cExtents; i++)
5650 if (apszOldName[i])
5651 RTStrFree(apszOldName[i]);
5652 RTMemTmpFree(apszOldName);
5653 }
5654 if (apszNewName)
5655 {
5656 for (i = 0; i <= cExtents; i++)
5657 if (apszNewName[i])
5658 RTStrFree(apszNewName[i]);
5659 RTMemTmpFree(apszNewName);
5660 }
5661 if (apszNewLines)
5662 {
5663 for (i = 0; i < cExtents; i++)
5664 if (apszNewLines[i])
5665 RTStrFree(apszNewLines[i]);
5666 RTMemTmpFree(apszNewLines);
5667 }
5668 if (pszOldDescName)
5669 RTStrFree(pszOldDescName);
5670 if (pszOldBaseName)
5671 RTStrFree(pszOldBaseName);
5672 if (pszNewBaseName)
5673 RTStrFree(pszNewBaseName);
5674 if (pszOldFullName)
5675 RTStrFree(pszOldFullName);
5676 if (pszNewFullName)
5677 RTStrFree(pszNewFullName);
5678 LogFlowFunc(("returns %Rrc\n", rc));
5679 return rc;
5680}
5681
5682/** @copydoc VBOXHDDBACKEND::pfnClose */
5683static int vmdkClose(void *pBackendData, bool fDelete)
5684{
5685 LogFlowFunc(("pBackendData=%#p fDelete=%d\n", pBackendData, fDelete));
5686 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
5687 int rc;
5688
5689 rc = vmdkFreeImage(pImage, fDelete);
5690 RTMemFree(pImage);
5691
5692 LogFlowFunc(("returns %Rrc\n", rc));
5693 return rc;
5694}
5695
5696/** @copydoc VBOXHDDBACKEND::pfnRead */
5697static int vmdkRead(void *pBackendData, uint64_t uOffset, size_t cbToRead,
5698 PVDIOCTX pIoCtx, size_t *pcbActuallyRead)
5699{
5700 LogFlowFunc(("pBackendData=%#p uOffset=%llu pIoCtx=%#p cbToRead=%zu pcbActuallyRead=%#p\n",
5701 pBackendData, uOffset, pIoCtx, cbToRead, pcbActuallyRead));
5702 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
5703 PVMDKEXTENT pExtent;
5704 uint64_t uSectorExtentRel;
5705 uint64_t uSectorExtentAbs;
5706 int rc;
5707
5708 AssertPtr(pImage);
5709 Assert(uOffset % 512 == 0);
5710 Assert(cbToRead % 512 == 0);
5711
5712 if ( uOffset + cbToRead > pImage->cbSize
5713 || cbToRead == 0)
5714 {
5715 rc = VERR_INVALID_PARAMETER;
5716 goto out;
5717 }
5718
5719 rc = vmdkFindExtent(pImage, VMDK_BYTE2SECTOR(uOffset),
5720 &pExtent, &uSectorExtentRel);
5721 if (RT_FAILURE(rc))
5722 goto out;
5723
5724 /* Check access permissions as defined in the extent descriptor. */
5725 if (pExtent->enmAccess == VMDKACCESS_NOACCESS)
5726 {
5727 rc = VERR_VD_VMDK_INVALID_STATE;
5728 goto out;
5729 }
5730
5731 /* Clip read range to remain in this extent. */
5732 cbToRead = RT_MIN(cbToRead, VMDK_SECTOR2BYTE(pExtent->uSectorOffset + pExtent->cNominalSectors - uSectorExtentRel));
5733
5734 /* Handle the read according to the current extent type. */
5735 switch (pExtent->enmType)
5736 {
5737 case VMDKETYPE_HOSTED_SPARSE:
5738#ifdef VBOX_WITH_VMDK_ESX
5739 case VMDKETYPE_ESX_SPARSE:
5740#endif /* VBOX_WITH_VMDK_ESX */
5741 rc = vmdkGetSector(pImage, pIoCtx, pExtent, uSectorExtentRel, &uSectorExtentAbs);
5742 if (RT_FAILURE(rc))
5743 goto out;
5744 /* Clip read range to at most the rest of the grain. */
5745 cbToRead = RT_MIN(cbToRead, VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain - uSectorExtentRel % pExtent->cSectorsPerGrain));
5746 Assert(!(cbToRead % 512));
5747 if (uSectorExtentAbs == 0)
5748 {
5749 if ( !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
5750 || !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
5751 || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL))
5752 rc = VERR_VD_BLOCK_FREE;
5753 else
5754 rc = vmdkStreamReadSequential(pImage, pExtent,
5755 uSectorExtentRel,
5756 pIoCtx, cbToRead);
5757 }
5758 else
5759 {
5760 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
5761 {
5762 AssertMsg(vdIfIoIntIoCtxIsSynchronous(pImage->pIfIo, pIoCtx),
5763 ("Async I/O is not supported for stream optimized VMDK's\n"));
5764
5765 uint32_t uSectorInGrain = uSectorExtentRel % pExtent->cSectorsPerGrain;
5766 uSectorExtentAbs -= uSectorInGrain;
5767 if (pExtent->uGrainSectorAbs != uSectorExtentAbs)
5768 {
5769 uint64_t uLBA = 0; /* gcc maybe uninitialized */
5770 rc = vmdkFileInflateSync(pImage, pExtent,
5771 VMDK_SECTOR2BYTE(uSectorExtentAbs),
5772 pExtent->pvGrain,
5773 VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain),
5774 NULL, &uLBA, NULL);
5775 if (RT_FAILURE(rc))
5776 {
5777 pExtent->uGrainSectorAbs = 0;
5778 AssertRC(rc);
5779 goto out;
5780 }
5781 pExtent->uGrainSectorAbs = uSectorExtentAbs;
5782 pExtent->uGrain = uSectorExtentRel / pExtent->cSectorsPerGrain;
5783 Assert(uLBA == uSectorExtentRel);
5784 }
5785 vdIfIoIntIoCtxCopyTo(pImage->pIfIo, pIoCtx,
5786 (uint8_t *)pExtent->pvGrain
5787 + VMDK_SECTOR2BYTE(uSectorInGrain),
5788 cbToRead);
5789 }
5790 else
5791 rc = vdIfIoIntFileReadUser(pImage->pIfIo, pExtent->pFile->pStorage,
5792 VMDK_SECTOR2BYTE(uSectorExtentAbs),
5793 pIoCtx, cbToRead);
5794 }
5795 break;
5796 case VMDKETYPE_VMFS:
5797 case VMDKETYPE_FLAT:
5798 rc = vdIfIoIntFileReadUser(pImage->pIfIo, pExtent->pFile->pStorage,
5799 VMDK_SECTOR2BYTE(uSectorExtentRel),
5800 pIoCtx, cbToRead);
5801 break;
5802 case VMDKETYPE_ZERO:
5803 size_t cbSet;
5804
5805 cbSet = vdIfIoIntIoCtxSet(pImage->pIfIo, pIoCtx, 0, cbToRead);
5806 Assert(cbSet == cbToRead);
5807
5808 rc = VINF_SUCCESS;
5809 break;
5810 }
5811 if (pcbActuallyRead)
5812 *pcbActuallyRead = cbToRead;
5813
5814out:
5815 LogFlowFunc(("returns %Rrc\n", rc));
5816 return rc;
5817}
5818
5819/** @copydoc VBOXHDDBACKEND::pfnWrite */
5820static int vmdkWrite(void *pBackendData, uint64_t uOffset, size_t cbToWrite,
5821 PVDIOCTX pIoCtx, size_t *pcbWriteProcess, size_t *pcbPreRead,
5822 size_t *pcbPostRead, unsigned fWrite)
5823{
5824 LogFlowFunc(("pBackendData=%#p uOffset=%llu pIoCtx=%#p cbToWrite=%zu pcbWriteProcess=%#p pcbPreRead=%#p pcbPostRead=%#p\n",
5825 pBackendData, uOffset, pIoCtx, cbToWrite, pcbWriteProcess, pcbPreRead, pcbPostRead));
5826 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
5827 PVMDKEXTENT pExtent;
5828 uint64_t uSectorExtentRel;
5829 uint64_t uSectorExtentAbs;
5830 int rc;
5831
5832 AssertPtr(pImage);
5833 Assert(uOffset % 512 == 0);
5834 Assert(cbToWrite % 512 == 0);
5835
5836 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
5837 {
5838 rc = VERR_VD_IMAGE_READ_ONLY;
5839 goto out;
5840 }
5841
5842 if (cbToWrite == 0)
5843 {
5844 rc = VERR_INVALID_PARAMETER;
5845 goto out;
5846 }
5847
5848 /* No size check here, will do that later when the extent is located.
5849 * There are sparse images out there which according to the spec are
5850 * invalid, because the total size is not a multiple of the grain size.
5851 * Also for sparse images which are stitched together in odd ways (not at
5852 * grain boundaries, and with the nominal size not being a multiple of the
5853 * grain size), this would prevent writing to the last grain. */
5854
5855 rc = vmdkFindExtent(pImage, VMDK_BYTE2SECTOR(uOffset),
5856 &pExtent, &uSectorExtentRel);
5857 if (RT_FAILURE(rc))
5858 goto out;
5859
5860 /* Check access permissions as defined in the extent descriptor. */
5861 if ( pExtent->enmAccess != VMDKACCESS_READWRITE
5862 && ( !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
5863 && !pImage->pExtents[0].uAppendPosition
5864 && pExtent->enmAccess != VMDKACCESS_READONLY))
5865 {
5866 rc = VERR_VD_VMDK_INVALID_STATE;
5867 goto out;
5868 }
5869
5870 /* Handle the write according to the current extent type. */
5871 switch (pExtent->enmType)
5872 {
5873 case VMDKETYPE_HOSTED_SPARSE:
5874#ifdef VBOX_WITH_VMDK_ESX
5875 case VMDKETYPE_ESX_SPARSE:
5876#endif /* VBOX_WITH_VMDK_ESX */
5877 rc = vmdkGetSector(pImage, pIoCtx, pExtent, uSectorExtentRel, &uSectorExtentAbs);
5878 if (RT_FAILURE(rc))
5879 goto out;
5880 /* Clip write range to at most the rest of the grain. */
5881 cbToWrite = RT_MIN(cbToWrite, VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain - uSectorExtentRel % pExtent->cSectorsPerGrain));
5882 if ( pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED
5883 && uSectorExtentRel < (uint64_t)pExtent->uLastGrainAccess * pExtent->cSectorsPerGrain)
5884 {
5885 rc = VERR_VD_VMDK_INVALID_WRITE;
5886 goto out;
5887 }
5888 if (uSectorExtentAbs == 0)
5889 {
5890 if (!(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
5891 {
5892 if (cbToWrite == VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain))
5893 {
5894 /* Full block write to a previously unallocated block.
5895 * Check if the caller wants to avoid the automatic alloc. */
5896 if (!(fWrite & VD_WRITE_NO_ALLOC))
5897 {
5898 /* Allocate GT and find out where to store the grain. */
5899 rc = vmdkAllocGrain(pImage, pExtent, pIoCtx,
5900 uSectorExtentRel, cbToWrite);
5901 }
5902 else
5903 rc = VERR_VD_BLOCK_FREE;
5904 *pcbPreRead = 0;
5905 *pcbPostRead = 0;
5906 }
5907 else
5908 {
5909 /* Clip write range to remain in this extent. */
5910 cbToWrite = RT_MIN(cbToWrite, VMDK_SECTOR2BYTE(pExtent->uSectorOffset + pExtent->cNominalSectors - uSectorExtentRel));
5911 *pcbPreRead = VMDK_SECTOR2BYTE(uSectorExtentRel % pExtent->cSectorsPerGrain);
5912 *pcbPostRead = VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain) - cbToWrite - *pcbPreRead;
5913 rc = VERR_VD_BLOCK_FREE;
5914 }
5915 }
5916 else
5917 {
5918 rc = vmdkStreamAllocGrain(pImage, pExtent,
5919 uSectorExtentRel,
5920 pIoCtx, cbToWrite);
5921 }
5922 }
5923 else
5924 {
5925 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
5926 {
5927 /* A partial write to a streamOptimized image is simply
5928 * invalid. It requires rewriting already compressed data
5929 * which is somewhere between expensive and impossible. */
5930 rc = VERR_VD_VMDK_INVALID_STATE;
5931 pExtent->uGrainSectorAbs = 0;
5932 AssertRC(rc);
5933 }
5934 else
5935 {
5936 Assert(!(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED));
5937 rc = vdIfIoIntFileWriteUser(pImage->pIfIo, pExtent->pFile->pStorage,
5938 VMDK_SECTOR2BYTE(uSectorExtentAbs),
5939 pIoCtx, cbToWrite, NULL, NULL);
5940 }
5941 }
5942 break;
5943 case VMDKETYPE_VMFS:
5944 case VMDKETYPE_FLAT:
5945 /* Clip write range to remain in this extent. */
5946 cbToWrite = RT_MIN(cbToWrite, VMDK_SECTOR2BYTE(pExtent->uSectorOffset + pExtent->cNominalSectors - uSectorExtentRel));
5947 rc = vdIfIoIntFileWriteUser(pImage->pIfIo, pExtent->pFile->pStorage,
5948 VMDK_SECTOR2BYTE(uSectorExtentRel),
5949 pIoCtx, cbToWrite, NULL, NULL);
5950 break;
5951 case VMDKETYPE_ZERO:
5952 /* Clip write range to remain in this extent. */
5953 cbToWrite = RT_MIN(cbToWrite, VMDK_SECTOR2BYTE(pExtent->uSectorOffset + pExtent->cNominalSectors - uSectorExtentRel));
5954 break;
5955 }
5956
5957 if (pcbWriteProcess)
5958 *pcbWriteProcess = cbToWrite;
5959
5960out:
5961 LogFlowFunc(("returns %Rrc\n", rc));
5962 return rc;
5963}
5964
5965/** @copydoc VBOXHDDBACKEND::pfnFlush */
5966static int vmdkFlush(void *pBackendData, PVDIOCTX pIoCtx)
5967{
5968 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
5969
5970 return vmdkFlushImage(pImage, pIoCtx);
5971}
5972
5973/** @copydoc VBOXHDDBACKEND::pfnGetVersion */
5974static unsigned vmdkGetVersion(void *pBackendData)
5975{
5976 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
5977 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
5978
5979 AssertPtr(pImage);
5980
5981 if (pImage)
5982 return VMDK_IMAGE_VERSION;
5983 else
5984 return 0;
5985}
5986
5987/** @copydoc VBOXHDDBACKEND::pfnGetSectorSize */
5988static uint32_t vmdkGetSectorSize(void *pBackendData)
5989{
5990 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
5991 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
5992
5993 AssertPtr(pImage);
5994
5995 if (pImage)
5996 return 512;
5997 else
5998 return 0;
5999}
6000
6001/** @copydoc VBOXHDDBACKEND::pfnGetSize */
6002static uint64_t vmdkGetSize(void *pBackendData)
6003{
6004 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
6005 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6006
6007 AssertPtr(pImage);
6008
6009 if (pImage)
6010 return pImage->cbSize;
6011 else
6012 return 0;
6013}
6014
6015/** @copydoc VBOXHDDBACKEND::pfnGetFileSize */
6016static uint64_t vmdkGetFileSize(void *pBackendData)
6017{
6018 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
6019 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6020 uint64_t cb = 0;
6021
6022 AssertPtr(pImage);
6023
6024 if (pImage)
6025 {
6026 uint64_t cbFile;
6027 if (pImage->pFile != NULL)
6028 {
6029 int rc = vdIfIoIntFileGetSize(pImage->pIfIo, pImage->pFile->pStorage, &cbFile);
6030 if (RT_SUCCESS(rc))
6031 cb += cbFile;
6032 }
6033 for (unsigned i = 0; i < pImage->cExtents; i++)
6034 {
6035 if (pImage->pExtents[i].pFile != NULL)
6036 {
6037 int rc = vdIfIoIntFileGetSize(pImage->pIfIo, pImage->pExtents[i].pFile->pStorage, &cbFile);
6038 if (RT_SUCCESS(rc))
6039 cb += cbFile;
6040 }
6041 }
6042 }
6043
6044 LogFlowFunc(("returns %lld\n", cb));
6045 return cb;
6046}
6047
6048/** @copydoc VBOXHDDBACKEND::pfnGetPCHSGeometry */
6049static int vmdkGetPCHSGeometry(void *pBackendData, PVDGEOMETRY pPCHSGeometry)
6050{
6051 LogFlowFunc(("pBackendData=%#p pPCHSGeometry=%#p\n", pBackendData, pPCHSGeometry));
6052 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6053 int rc;
6054
6055 AssertPtr(pImage);
6056
6057 if (pImage)
6058 {
6059 if (pImage->PCHSGeometry.cCylinders)
6060 {
6061 *pPCHSGeometry = pImage->PCHSGeometry;
6062 rc = VINF_SUCCESS;
6063 }
6064 else
6065 rc = VERR_VD_GEOMETRY_NOT_SET;
6066 }
6067 else
6068 rc = VERR_VD_NOT_OPENED;
6069
6070 LogFlowFunc(("returns %Rrc (PCHS=%u/%u/%u)\n", rc, pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads, pPCHSGeometry->cSectors));
6071 return rc;
6072}
6073
6074/** @copydoc VBOXHDDBACKEND::pfnSetPCHSGeometry */
6075static int vmdkSetPCHSGeometry(void *pBackendData, PCVDGEOMETRY pPCHSGeometry)
6076{
6077 LogFlowFunc(("pBackendData=%#p pPCHSGeometry=%#p PCHS=%u/%u/%u\n", pBackendData, pPCHSGeometry, pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads, pPCHSGeometry->cSectors));
6078 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6079 int rc;
6080
6081 AssertPtr(pImage);
6082
6083 if (pImage)
6084 {
6085 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
6086 {
6087 rc = VERR_VD_IMAGE_READ_ONLY;
6088 goto out;
6089 }
6090 if (pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
6091 {
6092 rc = VERR_NOT_SUPPORTED;
6093 goto out;
6094 }
6095 rc = vmdkDescSetPCHSGeometry(pImage, pPCHSGeometry);
6096 if (RT_FAILURE(rc))
6097 goto out;
6098
6099 pImage->PCHSGeometry = *pPCHSGeometry;
6100 rc = VINF_SUCCESS;
6101 }
6102 else
6103 rc = VERR_VD_NOT_OPENED;
6104
6105out:
6106 LogFlowFunc(("returns %Rrc\n", rc));
6107 return rc;
6108}
6109
6110/** @copydoc VBOXHDDBACKEND::pfnGetLCHSGeometry */
6111static int vmdkGetLCHSGeometry(void *pBackendData, PVDGEOMETRY pLCHSGeometry)
6112{
6113 LogFlowFunc(("pBackendData=%#p pLCHSGeometry=%#p\n", pBackendData, pLCHSGeometry));
6114 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6115 int rc;
6116
6117 AssertPtr(pImage);
6118
6119 if (pImage)
6120 {
6121 if (pImage->LCHSGeometry.cCylinders)
6122 {
6123 *pLCHSGeometry = pImage->LCHSGeometry;
6124 rc = VINF_SUCCESS;
6125 }
6126 else
6127 rc = VERR_VD_GEOMETRY_NOT_SET;
6128 }
6129 else
6130 rc = VERR_VD_NOT_OPENED;
6131
6132 LogFlowFunc(("returns %Rrc (LCHS=%u/%u/%u)\n", rc, pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads, pLCHSGeometry->cSectors));
6133 return rc;
6134}
6135
6136/** @copydoc VBOXHDDBACKEND::pfnSetLCHSGeometry */
6137static int vmdkSetLCHSGeometry(void *pBackendData, PCVDGEOMETRY pLCHSGeometry)
6138{
6139 LogFlowFunc(("pBackendData=%#p pLCHSGeometry=%#p LCHS=%u/%u/%u\n", pBackendData, pLCHSGeometry, pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads, pLCHSGeometry->cSectors));
6140 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6141 int rc;
6142
6143 AssertPtr(pImage);
6144
6145 if (pImage)
6146 {
6147 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
6148 {
6149 rc = VERR_VD_IMAGE_READ_ONLY;
6150 goto out;
6151 }
6152 if (pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
6153 {
6154 rc = VERR_NOT_SUPPORTED;
6155 goto out;
6156 }
6157 rc = vmdkDescSetLCHSGeometry(pImage, pLCHSGeometry);
6158 if (RT_FAILURE(rc))
6159 goto out;
6160
6161 pImage->LCHSGeometry = *pLCHSGeometry;
6162 rc = VINF_SUCCESS;
6163 }
6164 else
6165 rc = VERR_VD_NOT_OPENED;
6166
6167out:
6168 LogFlowFunc(("returns %Rrc\n", rc));
6169 return rc;
6170}
6171
6172/** @copydoc VBOXHDDBACKEND::pfnGetImageFlags */
6173static unsigned vmdkGetImageFlags(void *pBackendData)
6174{
6175 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
6176 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6177 unsigned uImageFlags;
6178
6179 AssertPtr(pImage);
6180
6181 if (pImage)
6182 uImageFlags = pImage->uImageFlags;
6183 else
6184 uImageFlags = 0;
6185
6186 LogFlowFunc(("returns %#x\n", uImageFlags));
6187 return uImageFlags;
6188}
6189
6190/** @copydoc VBOXHDDBACKEND::pfnGetOpenFlags */
6191static unsigned vmdkGetOpenFlags(void *pBackendData)
6192{
6193 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
6194 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6195 unsigned uOpenFlags;
6196
6197 AssertPtr(pImage);
6198
6199 if (pImage)
6200 uOpenFlags = pImage->uOpenFlags;
6201 else
6202 uOpenFlags = 0;
6203
6204 LogFlowFunc(("returns %#x\n", uOpenFlags));
6205 return uOpenFlags;
6206}
6207
6208/** @copydoc VBOXHDDBACKEND::pfnSetOpenFlags */
6209static int vmdkSetOpenFlags(void *pBackendData, unsigned uOpenFlags)
6210{
6211 LogFlowFunc(("pBackendData=%#p uOpenFlags=%#x\n", pBackendData, uOpenFlags));
6212 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6213 int rc;
6214
6215 /* Image must be opened and the new flags must be valid. */
6216 if (!pImage || (uOpenFlags & ~( VD_OPEN_FLAGS_READONLY | VD_OPEN_FLAGS_INFO
6217 | VD_OPEN_FLAGS_ASYNC_IO | VD_OPEN_FLAGS_SHAREABLE
6218 | VD_OPEN_FLAGS_SEQUENTIAL | VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS)))
6219 {
6220 rc = VERR_INVALID_PARAMETER;
6221 goto out;
6222 }
6223
6224 /* StreamOptimized images need special treatment: reopen is prohibited. */
6225 if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
6226 {
6227 if (pImage->uOpenFlags == uOpenFlags)
6228 rc = VINF_SUCCESS;
6229 else
6230 rc = VERR_INVALID_PARAMETER;
6231 }
6232 else
6233 {
6234 /* Implement this operation via reopening the image. */
6235 vmdkFreeImage(pImage, false);
6236 rc = vmdkOpenImage(pImage, uOpenFlags);
6237 }
6238
6239out:
6240 LogFlowFunc(("returns %Rrc\n", rc));
6241 return rc;
6242}
6243
6244/** @copydoc VBOXHDDBACKEND::pfnGetComment */
6245static int vmdkGetComment(void *pBackendData, char *pszComment,
6246 size_t cbComment)
6247{
6248 LogFlowFunc(("pBackendData=%#p pszComment=%#p cbComment=%zu\n", pBackendData, pszComment, cbComment));
6249 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6250 int rc;
6251
6252 AssertPtr(pImage);
6253
6254 if (pImage)
6255 {
6256 const char *pszCommentEncoded = NULL;
6257 rc = vmdkDescDDBGetStr(pImage, &pImage->Descriptor,
6258 "ddb.comment", &pszCommentEncoded);
6259 if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
6260 pszCommentEncoded = NULL;
6261 else if (RT_FAILURE(rc))
6262 goto out;
6263
6264 if (pszComment && pszCommentEncoded)
6265 rc = vmdkDecodeString(pszCommentEncoded, pszComment, cbComment);
6266 else
6267 {
6268 if (pszComment)
6269 *pszComment = '\0';
6270 rc = VINF_SUCCESS;
6271 }
6272 if (pszCommentEncoded)
6273 RTStrFree((char *)(void *)pszCommentEncoded);
6274 }
6275 else
6276 rc = VERR_VD_NOT_OPENED;
6277
6278out:
6279 LogFlowFunc(("returns %Rrc comment='%s'\n", rc, pszComment));
6280 return rc;
6281}
6282
6283/** @copydoc VBOXHDDBACKEND::pfnSetComment */
6284static int vmdkSetComment(void *pBackendData, const char *pszComment)
6285{
6286 LogFlowFunc(("pBackendData=%#p pszComment=\"%s\"\n", pBackendData, pszComment));
6287 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6288 int rc;
6289
6290 AssertPtr(pImage);
6291
6292 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
6293 {
6294 rc = VERR_VD_IMAGE_READ_ONLY;
6295 goto out;
6296 }
6297 if (pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
6298 {
6299 rc = VERR_NOT_SUPPORTED;
6300 goto out;
6301 }
6302
6303 if (pImage)
6304 rc = vmdkSetImageComment(pImage, pszComment);
6305 else
6306 rc = VERR_VD_NOT_OPENED;
6307
6308out:
6309 LogFlowFunc(("returns %Rrc\n", rc));
6310 return rc;
6311}
6312
6313/** @copydoc VBOXHDDBACKEND::pfnGetUuid */
6314static int vmdkGetUuid(void *pBackendData, PRTUUID pUuid)
6315{
6316 LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
6317 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6318 int rc;
6319
6320 AssertPtr(pImage);
6321
6322 if (pImage)
6323 {
6324 *pUuid = pImage->ImageUuid;
6325 rc = VINF_SUCCESS;
6326 }
6327 else
6328 rc = VERR_VD_NOT_OPENED;
6329
6330 LogFlowFunc(("returns %Rrc (%RTuuid)\n", rc, pUuid));
6331 return rc;
6332}
6333
6334/** @copydoc VBOXHDDBACKEND::pfnSetUuid */
6335static int vmdkSetUuid(void *pBackendData, PCRTUUID pUuid)
6336{
6337 LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
6338 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6339 int rc;
6340
6341 LogFlowFunc(("%RTuuid\n", pUuid));
6342 AssertPtr(pImage);
6343
6344 if (pImage)
6345 {
6346 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
6347 {
6348 if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
6349 {
6350 pImage->ImageUuid = *pUuid;
6351 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
6352 VMDK_DDB_IMAGE_UUID, pUuid);
6353 if (RT_FAILURE(rc))
6354 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing image UUID in descriptor in '%s'"), pImage->pszFilename);
6355 rc = VINF_SUCCESS;
6356 }
6357 else
6358 rc = VERR_NOT_SUPPORTED;
6359 }
6360 else
6361 rc = VERR_VD_IMAGE_READ_ONLY;
6362 }
6363 else
6364 rc = VERR_VD_NOT_OPENED;
6365
6366 LogFlowFunc(("returns %Rrc\n", rc));
6367 return rc;
6368}
6369
6370/** @copydoc VBOXHDDBACKEND::pfnGetModificationUuid */
6371static int vmdkGetModificationUuid(void *pBackendData, PRTUUID pUuid)
6372{
6373 LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
6374 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6375 int rc;
6376
6377 AssertPtr(pImage);
6378
6379 if (pImage)
6380 {
6381 *pUuid = pImage->ModificationUuid;
6382 rc = VINF_SUCCESS;
6383 }
6384 else
6385 rc = VERR_VD_NOT_OPENED;
6386
6387 LogFlowFunc(("returns %Rrc (%RTuuid)\n", rc, pUuid));
6388 return rc;
6389}
6390
6391/** @copydoc VBOXHDDBACKEND::pfnSetModificationUuid */
6392static int vmdkSetModificationUuid(void *pBackendData, PCRTUUID pUuid)
6393{
6394 LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
6395 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6396 int rc;
6397
6398 AssertPtr(pImage);
6399
6400 if (pImage)
6401 {
6402 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
6403 {
6404 if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
6405 {
6406 /* Only touch the modification uuid if it changed. */
6407 if (RTUuidCompare(&pImage->ModificationUuid, pUuid))
6408 {
6409 pImage->ModificationUuid = *pUuid;
6410 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
6411 VMDK_DDB_MODIFICATION_UUID, pUuid);
6412 if (RT_FAILURE(rc))
6413 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing modification UUID in descriptor in '%s'"), pImage->pszFilename);
6414 }
6415 rc = VINF_SUCCESS;
6416 }
6417 else
6418 rc = VERR_NOT_SUPPORTED;
6419 }
6420 else
6421 rc = VERR_VD_IMAGE_READ_ONLY;
6422 }
6423 else
6424 rc = VERR_VD_NOT_OPENED;
6425
6426 LogFlowFunc(("returns %Rrc\n", rc));
6427 return rc;
6428}
6429
6430/** @copydoc VBOXHDDBACKEND::pfnGetParentUuid */
6431static int vmdkGetParentUuid(void *pBackendData, PRTUUID pUuid)
6432{
6433 LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
6434 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6435 int rc;
6436
6437 AssertPtr(pImage);
6438
6439 if (pImage)
6440 {
6441 *pUuid = pImage->ParentUuid;
6442 rc = VINF_SUCCESS;
6443 }
6444 else
6445 rc = VERR_VD_NOT_OPENED;
6446
6447 LogFlowFunc(("returns %Rrc (%RTuuid)\n", rc, pUuid));
6448 return rc;
6449}
6450
6451/** @copydoc VBOXHDDBACKEND::pfnSetParentUuid */
6452static int vmdkSetParentUuid(void *pBackendData, PCRTUUID pUuid)
6453{
6454 LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
6455 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6456 int rc;
6457
6458 AssertPtr(pImage);
6459
6460 if (pImage)
6461 {
6462 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
6463 {
6464 if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
6465 {
6466 pImage->ParentUuid = *pUuid;
6467 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
6468 VMDK_DDB_PARENT_UUID, pUuid);
6469 if (RT_FAILURE(rc))
6470 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing parent image UUID in descriptor in '%s'"), pImage->pszFilename);
6471 rc = VINF_SUCCESS;
6472 }
6473 else
6474 rc = VERR_NOT_SUPPORTED;
6475 }
6476 else
6477 rc = VERR_VD_IMAGE_READ_ONLY;
6478 }
6479 else
6480 rc = VERR_VD_NOT_OPENED;
6481
6482 LogFlowFunc(("returns %Rrc\n", rc));
6483 return rc;
6484}
6485
6486/** @copydoc VBOXHDDBACKEND::pfnGetParentModificationUuid */
6487static int vmdkGetParentModificationUuid(void *pBackendData, PRTUUID pUuid)
6488{
6489 LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
6490 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6491 int rc;
6492
6493 AssertPtr(pImage);
6494
6495 if (pImage)
6496 {
6497 *pUuid = pImage->ParentModificationUuid;
6498 rc = VINF_SUCCESS;
6499 }
6500 else
6501 rc = VERR_VD_NOT_OPENED;
6502
6503 LogFlowFunc(("returns %Rrc (%RTuuid)\n", rc, pUuid));
6504 return rc;
6505}
6506
6507/** @copydoc VBOXHDDBACKEND::pfnSetParentModificationUuid */
6508static int vmdkSetParentModificationUuid(void *pBackendData, PCRTUUID pUuid)
6509{
6510 LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
6511 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6512 int rc;
6513
6514 AssertPtr(pImage);
6515
6516 if (pImage)
6517 {
6518 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
6519 {
6520 if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
6521 {
6522 pImage->ParentModificationUuid = *pUuid;
6523 rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
6524 VMDK_DDB_PARENT_MODIFICATION_UUID, pUuid);
6525 if (RT_FAILURE(rc))
6526 return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing parent image UUID in descriptor in '%s'"), pImage->pszFilename);
6527 rc = VINF_SUCCESS;
6528 }
6529 else
6530 rc = VERR_NOT_SUPPORTED;
6531 }
6532 else
6533 rc = VERR_VD_IMAGE_READ_ONLY;
6534 }
6535 else
6536 rc = VERR_VD_NOT_OPENED;
6537
6538 LogFlowFunc(("returns %Rrc\n", rc));
6539 return rc;
6540}
6541
6542/** @copydoc VBOXHDDBACKEND::pfnDump */
6543static void vmdkDump(void *pBackendData)
6544{
6545 PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
6546
6547 AssertPtr(pImage);
6548 if (pImage)
6549 {
6550 vdIfErrorMessage(pImage->pIfError, "Header: Geometry PCHS=%u/%u/%u LCHS=%u/%u/%u cbSector=%llu\n",
6551 pImage->PCHSGeometry.cCylinders, pImage->PCHSGeometry.cHeads, pImage->PCHSGeometry.cSectors,
6552 pImage->LCHSGeometry.cCylinders, pImage->LCHSGeometry.cHeads, pImage->LCHSGeometry.cSectors,
6553 VMDK_BYTE2SECTOR(pImage->cbSize));
6554 vdIfErrorMessage(pImage->pIfError, "Header: uuidCreation={%RTuuid}\n", &pImage->ImageUuid);
6555 vdIfErrorMessage(pImage->pIfError, "Header: uuidModification={%RTuuid}\n", &pImage->ModificationUuid);
6556 vdIfErrorMessage(pImage->pIfError, "Header: uuidParent={%RTuuid}\n", &pImage->ParentUuid);
6557 vdIfErrorMessage(pImage->pIfError, "Header: uuidParentModification={%RTuuid}\n", &pImage->ParentModificationUuid);
6558 }
6559}
6560
6561
6562
6563VBOXHDDBACKEND g_VmdkBackend =
6564{
6565 /* pszBackendName */
6566 "VMDK",
6567 /* cbSize */
6568 sizeof(VBOXHDDBACKEND),
6569 /* uBackendCaps */
6570 VD_CAP_UUID | VD_CAP_CREATE_FIXED | VD_CAP_CREATE_DYNAMIC
6571 | VD_CAP_CREATE_SPLIT_2G | VD_CAP_DIFF | VD_CAP_FILE | VD_CAP_ASYNC
6572 | VD_CAP_VFS,
6573 /* paFileExtensions */
6574 s_aVmdkFileExtensions,
6575 /* paConfigInfo */
6576 NULL,
6577 /* hPlugin */
6578 NIL_RTLDRMOD,
6579 /* pfnCheckIfValid */
6580 vmdkCheckIfValid,
6581 /* pfnOpen */
6582 vmdkOpen,
6583 /* pfnCreate */
6584 vmdkCreate,
6585 /* pfnRename */
6586 vmdkRename,
6587 /* pfnClose */
6588 vmdkClose,
6589 /* pfnRead */
6590 vmdkRead,
6591 /* pfnWrite */
6592 vmdkWrite,
6593 /* pfnFlush */
6594 vmdkFlush,
6595 /* pfnDiscard */
6596 NULL,
6597 /* pfnGetVersion */
6598 vmdkGetVersion,
6599 /* pfnGetSectorSize */
6600 vmdkGetSectorSize,
6601 /* pfnGetSize */
6602 vmdkGetSize,
6603 /* pfnGetFileSize */
6604 vmdkGetFileSize,
6605 /* pfnGetPCHSGeometry */
6606 vmdkGetPCHSGeometry,
6607 /* pfnSetPCHSGeometry */
6608 vmdkSetPCHSGeometry,
6609 /* pfnGetLCHSGeometry */
6610 vmdkGetLCHSGeometry,
6611 /* pfnSetLCHSGeometry */
6612 vmdkSetLCHSGeometry,
6613 /* pfnGetImageFlags */
6614 vmdkGetImageFlags,
6615 /* pfnGetOpenFlags */
6616 vmdkGetOpenFlags,
6617 /* pfnSetOpenFlags */
6618 vmdkSetOpenFlags,
6619 /* pfnGetComment */
6620 vmdkGetComment,
6621 /* pfnSetComment */
6622 vmdkSetComment,
6623 /* pfnGetUuid */
6624 vmdkGetUuid,
6625 /* pfnSetUuid */
6626 vmdkSetUuid,
6627 /* pfnGetModificationUuid */
6628 vmdkGetModificationUuid,
6629 /* pfnSetModificationUuid */
6630 vmdkSetModificationUuid,
6631 /* pfnGetParentUuid */
6632 vmdkGetParentUuid,
6633 /* pfnSetParentUuid */
6634 vmdkSetParentUuid,
6635 /* pfnGetParentModificationUuid */
6636 vmdkGetParentModificationUuid,
6637 /* pfnSetParentModificationUuid */
6638 vmdkSetParentModificationUuid,
6639 /* pfnDump */
6640 vmdkDump,
6641 /* pfnGetTimeStamp */
6642 NULL,
6643 /* pfnGetParentTimeStamp */
6644 NULL,
6645 /* pfnSetParentTimeStamp */
6646 NULL,
6647 /* pfnGetParentFilename */
6648 NULL,
6649 /* pfnSetParentFilename */
6650 NULL,
6651 /* pfnComposeLocation */
6652 genericFileComposeLocation,
6653 /* pfnComposeName */
6654 genericFileComposeName,
6655 /* pfnCompact */
6656 NULL,
6657 /* pfnResize */
6658 NULL,
6659 /* pfnRepair */
6660 NULL
6661};
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette