VirtualBox

source: vbox/trunk/src/VBox/Storage/VD.cpp@ 64765

Last change on this file since 64765 was 64765, checked in by vboxsync, 8 years ago

Storage/VD: (bugref:8685) Fixed VBoxSVC crash on removing extpack while no VD plugins are loaded.

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 397.0 KB
Line 
1/* $Id: VD.cpp 64765 2016-11-30 10:28:00Z vboxsync $ */
2/** @file
3 * VBoxHDD - VBox HDD Container implementation.
4 */
5
6/*
7 * Copyright (C) 2006-2016 Oracle Corporation
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 */
17
18
19/*********************************************************************************************************************************
20* Header Files *
21*********************************************************************************************************************************/
22#define LOG_GROUP LOG_GROUP_VD
23#include <VBox/vd.h>
24#include <VBox/err.h>
25#include <VBox/sup.h>
26#include <VBox/log.h>
27
28#include <iprt/alloc.h>
29#include <iprt/assert.h>
30#include <iprt/uuid.h>
31#include <iprt/file.h>
32#include <iprt/string.h>
33#include <iprt/asm.h>
34#include <iprt/ldr.h>
35#include <iprt/dir.h>
36#include <iprt/path.h>
37#include <iprt/param.h>
38#include <iprt/memcache.h>
39#include <iprt/sg.h>
40#include <iprt/list.h>
41#include <iprt/avl.h>
42#include <iprt/semaphore.h>
43
44#include <VBox/vd-plugin.h>
45
46#include "VDBackends.h"
47
48/** Disable dynamic backends on non x86 architectures. This feature
49 * requires the SUPR3 library which is not available there.
50 */
51#if !defined(VBOX_HDD_NO_DYNAMIC_BACKENDS) && !defined(RT_ARCH_X86) && !defined(RT_ARCH_AMD64)
52# define VBOX_HDD_NO_DYNAMIC_BACKENDS
53#endif
54
55#define VBOXHDDDISK_SIGNATURE 0x6f0e2a7d
56
57/** Buffer size used for merging images. */
58#define VD_MERGE_BUFFER_SIZE (16 * _1M)
59
60/** Maximum number of segments in one I/O task. */
61#define VD_IO_TASK_SEGMENTS_MAX 64
62
63/** Threshold after not recently used blocks are removed from the list. */
64#define VD_DISCARD_REMOVE_THRESHOLD (10 * _1M) /** @todo experiment */
65
66/**
67 * VD async I/O interface storage descriptor.
68 */
69typedef struct VDIIOFALLBACKSTORAGE
70{
71 /** File handle. */
72 RTFILE File;
73 /** Completion callback. */
74 PFNVDCOMPLETED pfnCompleted;
75 /** Thread for async access. */
76 RTTHREAD ThreadAsync;
77} VDIIOFALLBACKSTORAGE, *PVDIIOFALLBACKSTORAGE;
78
79/**
80 * Structure containing everything I/O related
81 * for the image and cache descriptors.
82 */
83typedef struct VDIO
84{
85 /** I/O interface to the upper layer. */
86 PVDINTERFACEIO pInterfaceIo;
87
88 /** Per image internal I/O interface. */
89 VDINTERFACEIOINT VDIfIoInt;
90
91 /** Fallback I/O interface, only used if the caller doesn't provide it. */
92 VDINTERFACEIO VDIfIo;
93
94 /** Opaque backend data. */
95 void *pBackendData;
96 /** Disk this image is part of */
97 PVBOXHDD pDisk;
98 /** Flag whether to ignore flush requests. */
99 bool fIgnoreFlush;
100} VDIO, *PVDIO;
101
102/** Forward declaration of an I/O task */
103typedef struct VDIOTASK *PVDIOTASK;
104
105/**
106 * VBox HDD Container image descriptor.
107 */
108typedef struct VDIMAGE
109{
110 /** Link to parent image descriptor, if any. */
111 struct VDIMAGE *pPrev;
112 /** Link to child image descriptor, if any. */
113 struct VDIMAGE *pNext;
114 /** Container base filename. (UTF-8) */
115 char *pszFilename;
116 /** Data managed by the backend which keeps the actual info. */
117 void *pBackendData;
118 /** Cached sanitized image flags. */
119 unsigned uImageFlags;
120 /** Image open flags (only those handled generically in this code and which
121 * the backends will never ever see). */
122 unsigned uOpenFlags;
123
124 /** Function pointers for the various backend methods. */
125 PCVDIMAGEBACKEND Backend;
126 /** Pointer to list of VD interfaces, per-image. */
127 PVDINTERFACE pVDIfsImage;
128 /** I/O related things. */
129 VDIO VDIo;
130} VDIMAGE, *PVDIMAGE;
131
132/**
133 * uModified bit flags.
134 */
135#define VD_IMAGE_MODIFIED_FLAG RT_BIT(0)
136#define VD_IMAGE_MODIFIED_FIRST RT_BIT(1)
137#define VD_IMAGE_MODIFIED_DISABLE_UUID_UPDATE RT_BIT(2)
138
139
140/**
141 * VBox HDD Cache image descriptor.
142 */
143typedef struct VDCACHE
144{
145 /** Cache base filename. (UTF-8) */
146 char *pszFilename;
147 /** Data managed by the backend which keeps the actual info. */
148 void *pBackendData;
149 /** Cached sanitized image flags. */
150 unsigned uImageFlags;
151 /** Image open flags (only those handled generically in this code and which
152 * the backends will never ever see). */
153 unsigned uOpenFlags;
154
155 /** Function pointers for the various backend methods. */
156 PCVDCACHEBACKEND Backend;
157
158 /** Pointer to list of VD interfaces, per-cache. */
159 PVDINTERFACE pVDIfsCache;
160 /** I/O related things. */
161 VDIO VDIo;
162} VDCACHE, *PVDCACHE;
163
164/**
165 * A block waiting for a discard.
166 */
167typedef struct VDDISCARDBLOCK
168{
169 /** AVL core. */
170 AVLRU64NODECORE Core;
171 /** LRU list node. */
172 RTLISTNODE NodeLru;
173 /** Number of bytes to discard. */
174 size_t cbDiscard;
175 /** Bitmap of allocated sectors. */
176 void *pbmAllocated;
177} VDDISCARDBLOCK, *PVDDISCARDBLOCK;
178
179/**
180 * VD discard state.
181 */
182typedef struct VDDISCARDSTATE
183{
184 /** Number of bytes waiting for a discard. */
185 size_t cbDiscarding;
186 /** AVL tree with blocks waiting for a discard.
187 * The uOffset + cbDiscard range is the search key. */
188 PAVLRU64TREE pTreeBlocks;
189 /** LRU list of the least frequently discarded blocks.
190 * If there are to many blocks waiting the least frequently used
191 * will be removed and the range will be set to 0.
192 */
193 RTLISTNODE ListLru;
194} VDDISCARDSTATE, *PVDDISCARDSTATE;
195
196/**
197 * VD filter instance.
198 */
199typedef struct VDFILTER
200{
201 /** List node for the read filter chain. */
202 RTLISTNODE ListNodeChainRead;
203 /** List node for the write filter chain. */
204 RTLISTNODE ListNodeChainWrite;
205 /** Number of references to this filter. */
206 uint32_t cRefs;
207 /** Opaque VD filter backend instance data. */
208 void *pvBackendData;
209 /** Pointer to the filter backend interface. */
210 PCVDFILTERBACKEND pBackend;
211 /** Pointer to list of VD interfaces, per-filter. */
212 PVDINTERFACE pVDIfsFilter;
213 /** I/O related things. */
214 VDIO VDIo;
215} VDFILTER;
216/** Pointer to a VD filter instance. */
217typedef VDFILTER *PVDFILTER;
218
219/**
220 * VBox HDD Container main structure, private part.
221 */
222struct VBOXHDD
223{
224 /** Structure signature (VBOXHDDDISK_SIGNATURE). */
225 uint32_t u32Signature;
226
227 /** Image type. */
228 VDTYPE enmType;
229
230 /** Number of opened images. */
231 unsigned cImages;
232
233 /** Base image. */
234 PVDIMAGE pBase;
235
236 /** Last opened image in the chain.
237 * The same as pBase if only one image is used. */
238 PVDIMAGE pLast;
239
240 /** If a merge to one of the parents is running this may be non-NULL
241 * to indicate to what image the writes should be additionally relayed. */
242 PVDIMAGE pImageRelay;
243
244 /** Flags representing the modification state. */
245 unsigned uModified;
246
247 /** Cached size of this disk. */
248 uint64_t cbSize;
249 /** Cached PCHS geometry for this disk. */
250 VDGEOMETRY PCHSGeometry;
251 /** Cached LCHS geometry for this disk. */
252 VDGEOMETRY LCHSGeometry;
253
254 /** Pointer to list of VD interfaces, per-disk. */
255 PVDINTERFACE pVDIfsDisk;
256 /** Pointer to the common interface structure for error reporting. */
257 PVDINTERFACEERROR pInterfaceError;
258 /** Pointer to the optional thread synchronization callbacks. */
259 PVDINTERFACETHREADSYNC pInterfaceThreadSync;
260
261 /** Memory cache for I/O contexts */
262 RTMEMCACHE hMemCacheIoCtx;
263 /** Memory cache for I/O tasks. */
264 RTMEMCACHE hMemCacheIoTask;
265 /** An I/O context is currently using the disk structures
266 * Every I/O context must be placed on one of the lists below. */
267 volatile bool fLocked;
268 /** Head of pending I/O tasks waiting for completion - LIFO order. */
269 volatile PVDIOTASK pIoTasksPendingHead;
270 /** Head of newly queued I/O contexts - LIFO order. */
271 volatile PVDIOCTX pIoCtxHead;
272 /** Head of halted I/O contexts which are given back to generic
273 * disk framework by the backend. - LIFO order. */
274 volatile PVDIOCTX pIoCtxHaltedHead;
275
276 /** Head of blocked I/O contexts, processed only
277 * after pIoCtxLockOwner was freed - LIFO order. */
278 volatile PVDIOCTX pIoCtxBlockedHead;
279 /** I/O context which locked the disk for a growing write or flush request.
280 * Other flush or growing write requests need to wait until
281 * the current one completes. - NIL_VDIOCTX if unlocked. */
282 volatile PVDIOCTX pIoCtxLockOwner;
283 /** If the disk was locked by a growing write, flush or discard request this
284 * contains the start offset to check for interfering I/O while it is in progress. */
285 uint64_t uOffsetStartLocked;
286 /** If the disk was locked by a growing write, flush or discard request this contains
287 * the first non affected offset to check for interfering I/O while it is in progress. */
288 uint64_t uOffsetEndLocked;
289
290 /** Pointer to the L2 disk cache if any. */
291 PVDCACHE pCache;
292 /** Pointer to the discard state if any. */
293 PVDDISCARDSTATE pDiscard;
294
295 /** Read filter chain - PVDFILTER. */
296 RTLISTANCHOR ListFilterChainRead;
297 /** Write filter chain - PVDFILTER. */
298 RTLISTANCHOR ListFilterChainWrite;
299};
300
301# define VD_IS_LOCKED(a_pDisk) \
302 do \
303 { \
304 NOREF(a_pDisk); \
305 AssertMsg((a_pDisk)->fLocked, \
306 ("Lock not held\n"));\
307 } while(0)
308
309/**
310 * VBox parent read descriptor, used internally for compaction.
311 */
312typedef struct VDPARENTSTATEDESC
313{
314 /** Pointer to disk descriptor. */
315 PVBOXHDD pDisk;
316 /** Pointer to image descriptor. */
317 PVDIMAGE pImage;
318} VDPARENTSTATEDESC, *PVDPARENTSTATEDESC;
319
320/**
321 * Transfer direction.
322 */
323typedef enum VDIOCTXTXDIR
324{
325 /** Read */
326 VDIOCTXTXDIR_READ = 0,
327 /** Write */
328 VDIOCTXTXDIR_WRITE,
329 /** Flush */
330 VDIOCTXTXDIR_FLUSH,
331 /** Discard */
332 VDIOCTXTXDIR_DISCARD,
333 /** 32bit hack */
334 VDIOCTXTXDIR_32BIT_HACK = 0x7fffffff
335} VDIOCTXTXDIR, *PVDIOCTXTXDIR;
336
337/** Transfer function */
338typedef DECLCALLBACK(int) FNVDIOCTXTRANSFER (PVDIOCTX pIoCtx);
339/** Pointer to a transfer function. */
340typedef FNVDIOCTXTRANSFER *PFNVDIOCTXTRANSFER;
341
342/**
343 * I/O context
344 */
345typedef struct VDIOCTX
346{
347 /** Pointer to the next I/O context. */
348 struct VDIOCTX * volatile pIoCtxNext;
349 /** Disk this is request is for. */
350 PVBOXHDD pDisk;
351 /** Return code. */
352 int rcReq;
353 /** Various flags for the I/O context. */
354 uint32_t fFlags;
355 /** Number of data transfers currently pending. */
356 volatile uint32_t cDataTransfersPending;
357 /** How many meta data transfers are pending. */
358 volatile uint32_t cMetaTransfersPending;
359 /** Flag whether the request finished */
360 volatile bool fComplete;
361 /** Temporary allocated memory which is freed
362 * when the context completes. */
363 void *pvAllocation;
364 /** Transfer function. */
365 PFNVDIOCTXTRANSFER pfnIoCtxTransfer;
366 /** Next transfer part after the current one completed. */
367 PFNVDIOCTXTRANSFER pfnIoCtxTransferNext;
368 /** Transfer direction */
369 VDIOCTXTXDIR enmTxDir;
370 /** Request type dependent data. */
371 union
372 {
373 /** I/O request (read/write). */
374 struct
375 {
376 /** Number of bytes left until this context completes. */
377 volatile uint32_t cbTransferLeft;
378 /** Current offset */
379 volatile uint64_t uOffset;
380 /** Number of bytes to transfer */
381 volatile size_t cbTransfer;
382 /** Current image in the chain. */
383 PVDIMAGE pImageCur;
384 /** Start image to read from. pImageCur is reset to this
385 * value after it reached the first image in the chain. */
386 PVDIMAGE pImageStart;
387 /** S/G buffer */
388 RTSGBUF SgBuf;
389 /** Number of bytes to clear in the buffer before the current read. */
390 size_t cbBufClear;
391 /** Number of images to read. */
392 unsigned cImagesRead;
393 /** Override for the parent image to start reading from. */
394 PVDIMAGE pImageParentOverride;
395 /** Original offset of the transfer - required for filtering read requests. */
396 uint64_t uOffsetXferOrig;
397 /** Original size of the transfer - required for fitlering read requests. */
398 size_t cbXferOrig;
399 } Io;
400 /** Discard requests. */
401 struct
402 {
403 /** Pointer to the range descriptor array. */
404 PCRTRANGE paRanges;
405 /** Number of ranges in the array. */
406 unsigned cRanges;
407 /** Range descriptor index which is processed. */
408 unsigned idxRange;
409 /** Start offset to discard currently. */
410 uint64_t offCur;
411 /** How many bytes left to discard in the current range. */
412 size_t cbDiscardLeft;
413 /** How many bytes to discard in the current block (<= cbDiscardLeft). */
414 size_t cbThisDiscard;
415 /** Discard block handled currently. */
416 PVDDISCARDBLOCK pBlock;
417 } Discard;
418 } Req;
419 /** Parent I/O context if any. Sets the type of the context (root/child) */
420 PVDIOCTX pIoCtxParent;
421 /** Type dependent data (root/child) */
422 union
423 {
424 /** Root data */
425 struct
426 {
427 /** Completion callback */
428 PFNVDASYNCTRANSFERCOMPLETE pfnComplete;
429 /** User argument 1 passed on completion. */
430 void *pvUser1;
431 /** User argument 2 passed on completion. */
432 void *pvUser2;
433 } Root;
434 /** Child data */
435 struct
436 {
437 /** Saved start offset */
438 uint64_t uOffsetSaved;
439 /** Saved transfer size */
440 size_t cbTransferLeftSaved;
441 /** Number of bytes transferred from the parent if this context completes. */
442 size_t cbTransferParent;
443 /** Number of bytes to pre read */
444 size_t cbPreRead;
445 /** Number of bytes to post read. */
446 size_t cbPostRead;
447 /** Number of bytes to write left in the parent. */
448 size_t cbWriteParent;
449 /** Write type dependent data. */
450 union
451 {
452 /** Optimized */
453 struct
454 {
455 /** Bytes to fill to satisfy the block size. Not part of the virtual disk. */
456 size_t cbFill;
457 /** Bytes to copy instead of reading from the parent */
458 size_t cbWriteCopy;
459 /** Bytes to read from the image. */
460 size_t cbReadImage;
461 } Optimized;
462 } Write;
463 } Child;
464 } Type;
465} VDIOCTX;
466
467/** Default flags for an I/O context, i.e. unblocked and async. */
468#define VDIOCTX_FLAGS_DEFAULT (0)
469/** Flag whether the context is blocked. */
470#define VDIOCTX_FLAGS_BLOCKED RT_BIT_32(0)
471/** Flag whether the I/O context is using synchronous I/O. */
472#define VDIOCTX_FLAGS_SYNC RT_BIT_32(1)
473/** Flag whether the read should update the cache. */
474#define VDIOCTX_FLAGS_READ_UPDATE_CACHE RT_BIT_32(2)
475/** Flag whether free blocks should be zeroed.
476 * If false and no image has data for sepcified
477 * range VERR_VD_BLOCK_FREE is returned for the I/O context.
478 * Note that unallocated blocks are still zeroed
479 * if at least one image has valid data for a part
480 * of the range.
481 */
482#define VDIOCTX_FLAGS_ZERO_FREE_BLOCKS RT_BIT_32(3)
483/** Don't free the I/O context when complete because
484 * it was alloacted elsewhere (stack, ...). */
485#define VDIOCTX_FLAGS_DONT_FREE RT_BIT_32(4)
486/** Don't set the modified flag for this I/O context when writing. */
487#define VDIOCTX_FLAGS_DONT_SET_MODIFIED_FLAG RT_BIT_32(5)
488/** The write filter was applied already and shouldn't be applied a second time.
489 * Used at the beginning of vdWriteHelperAsync() because it might be called
490 * multiple times.
491 */
492#define VDIOCTX_FLAGS_WRITE_FILTER_APPLIED RT_BIT_32(6)
493
494/** NIL I/O context pointer value. */
495#define NIL_VDIOCTX ((PVDIOCTX)0)
496
497/**
498 * List node for deferred I/O contexts.
499 */
500typedef struct VDIOCTXDEFERRED
501{
502 /** Node in the list of deferred requests.
503 * A request can be deferred if the image is growing
504 * and the request accesses the same range or if
505 * the backend needs to read or write metadata from the disk
506 * before it can continue. */
507 RTLISTNODE NodeDeferred;
508 /** I/O context this entry points to. */
509 PVDIOCTX pIoCtx;
510} VDIOCTXDEFERRED, *PVDIOCTXDEFERRED;
511
512/**
513 * I/O task.
514 */
515typedef struct VDIOTASK
516{
517 /** Next I/O task waiting in the list. */
518 struct VDIOTASK * volatile pNext;
519 /** Storage this task belongs to. */
520 PVDIOSTORAGE pIoStorage;
521 /** Optional completion callback. */
522 PFNVDXFERCOMPLETED pfnComplete;
523 /** Opaque user data. */
524 void *pvUser;
525 /** Completion status code for the task. */
526 int rcReq;
527 /** Flag whether this is a meta data transfer. */
528 bool fMeta;
529 /** Type dependent data. */
530 union
531 {
532 /** User data transfer. */
533 struct
534 {
535 /** Number of bytes this task transferred. */
536 uint32_t cbTransfer;
537 /** Pointer to the I/O context the task belongs. */
538 PVDIOCTX pIoCtx;
539 } User;
540 /** Meta data transfer. */
541 struct
542 {
543 /** Meta transfer this task is for. */
544 PVDMETAXFER pMetaXfer;
545 } Meta;
546 } Type;
547} VDIOTASK;
548
549/**
550 * Storage handle.
551 */
552typedef struct VDIOSTORAGE
553{
554 /** Image I/O state this storage handle belongs to. */
555 PVDIO pVDIo;
556 /** AVL tree for pending async metadata transfers. */
557 PAVLRFOFFTREE pTreeMetaXfers;
558 /** Storage handle */
559 void *pStorage;
560} VDIOSTORAGE;
561
562/**
563 * Metadata transfer.
564 *
565 * @note This entry can't be freed if either the list is not empty or
566 * the reference counter is not 0.
567 * The assumption is that the backends don't need to read huge amounts of
568 * metadata to complete a transfer so the additional memory overhead should
569 * be relatively small.
570 */
571typedef struct VDMETAXFER
572{
573 /** AVL core for fast search (the file offset is the key) */
574 AVLRFOFFNODECORE Core;
575 /** I/O storage for this transfer. */
576 PVDIOSTORAGE pIoStorage;
577 /** Flags. */
578 uint32_t fFlags;
579 /** List of I/O contexts waiting for this metadata transfer to complete. */
580 RTLISTNODE ListIoCtxWaiting;
581 /** Number of references to this entry. */
582 unsigned cRefs;
583 /** Size of the data stored with this entry. */
584 size_t cbMeta;
585 /** Shadow buffer which is used in case a write is still active and other
586 * writes update the shadow buffer. */
587 uint8_t *pbDataShw;
588 /** List of I/O contexts updating the shadow buffer while there is a write
589 * in progress. */
590 RTLISTNODE ListIoCtxShwWrites;
591 /** Data stored - variable size. */
592 uint8_t abData[1];
593} VDMETAXFER;
594
595/**
596 * The transfer direction for the metadata.
597 */
598#define VDMETAXFER_TXDIR_MASK 0x3
599#define VDMETAXFER_TXDIR_NONE 0x0
600#define VDMETAXFER_TXDIR_WRITE 0x1
601#define VDMETAXFER_TXDIR_READ 0x2
602#define VDMETAXFER_TXDIR_FLUSH 0x3
603#define VDMETAXFER_TXDIR_GET(flags) ((flags) & VDMETAXFER_TXDIR_MASK)
604#define VDMETAXFER_TXDIR_SET(flags, dir) ((flags) = (flags & ~VDMETAXFER_TXDIR_MASK) | (dir))
605
606/**
607 * Plugin structure.
608 */
609typedef struct VDPLUGIN
610{
611 /** Pointer to the next plugin structure. */
612 RTLISTNODE NodePlugin;
613 /** Handle of loaded plugin library. */
614 RTLDRMOD hPlugin;
615 /** Filename of the loaded plugin. */
616 char *pszFilename;
617} VDPLUGIN;
618/** Pointer to a plugin structure. */
619typedef VDPLUGIN *PVDPLUGIN;
620
621/** Head of loaded plugin list. */
622static RTLISTANCHOR g_ListPluginsLoaded;
623
624/** Number of image backends supported. */
625static unsigned g_cBackends = 0;
626/** Array of pointers to the image backends. */
627static PCVDIMAGEBACKEND *g_apBackends = NULL;
628/** Array of handles to the corresponding plugin. */
629static RTLDRMOD *g_ahBackendPlugins = NULL;
630/** Builtin image backends. */
631static PCVDIMAGEBACKEND aStaticBackends[] =
632{
633 &g_VmdkBackend,
634 &g_VDIBackend,
635 &g_VhdBackend,
636 &g_ParallelsBackend,
637 &g_DmgBackend,
638 &g_QedBackend,
639 &g_QCowBackend,
640 &g_VhdxBackend,
641 &g_RawBackend,
642 &g_ISCSIBackend
643};
644
645/** Number of supported cache backends. */
646static unsigned g_cCacheBackends = 0;
647/** Array of pointers to the cache backends. */
648static PCVDCACHEBACKEND *g_apCacheBackends = NULL;
649/** Array of handles to the corresponding plugin. */
650static RTLDRMOD *g_ahCacheBackendPlugins = NULL;
651/** Builtin cache backends. */
652static PCVDCACHEBACKEND aStaticCacheBackends[] =
653{
654 &g_VciCacheBackend
655};
656
657/** Number of supported filter backends. */
658static unsigned g_cFilterBackends = 0;
659/** Array of pointers to the filters backends. */
660static PCVDFILTERBACKEND *g_apFilterBackends = NULL;
661#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
662/** Array of handles to the corresponding plugin. */
663static PRTLDRMOD g_pahFilterBackendPlugins = NULL;
664#endif
665
666/** Forward declaration of the async discard helper. */
667static DECLCALLBACK(int) vdDiscardHelperAsync(PVDIOCTX pIoCtx);
668static DECLCALLBACK(int) vdWriteHelperAsync(PVDIOCTX pIoCtx);
669static void vdDiskProcessBlockedIoCtx(PVBOXHDD pDisk);
670static int vdDiskUnlock(PVBOXHDD pDisk, PVDIOCTX pIoCtxRc);
671static DECLCALLBACK(void) vdIoCtxSyncComplete(void *pvUser1, void *pvUser2, int rcReq);
672
673/**
674 * internal: add several backends.
675 */
676static int vdAddBackends(RTLDRMOD hPlugin, PCVDIMAGEBACKEND *ppBackends, unsigned cBackends)
677{
678 PCVDIMAGEBACKEND *pTmp = (PCVDIMAGEBACKEND *)RTMemRealloc(g_apBackends,
679 (g_cBackends + cBackends) * sizeof(PCVDIMAGEBACKEND));
680 if (RT_UNLIKELY(!pTmp))
681 return VERR_NO_MEMORY;
682 g_apBackends = pTmp;
683
684 RTLDRMOD *pTmpPlugins = (RTLDRMOD*)RTMemRealloc(g_ahBackendPlugins,
685 (g_cBackends + cBackends) * sizeof(RTLDRMOD));
686 if (RT_UNLIKELY(!pTmpPlugins))
687 return VERR_NO_MEMORY;
688 g_ahBackendPlugins = pTmpPlugins;
689 memcpy(&g_apBackends[g_cBackends], ppBackends, cBackends * sizeof(PCVDIMAGEBACKEND));
690 for (unsigned i = g_cBackends; i < g_cBackends + cBackends; i++)
691 g_ahBackendPlugins[i] = hPlugin;
692 g_cBackends += cBackends;
693 return VINF_SUCCESS;
694}
695
696#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
697/**
698 * internal: add single backend.
699 */
700DECLINLINE(int) vdAddBackend(RTLDRMOD hPlugin, PCVDIMAGEBACKEND pBackend)
701{
702 return vdAddBackends(hPlugin, &pBackend, 1);
703}
704#endif
705
706/**
707 * internal: add several cache backends.
708 */
709static int vdAddCacheBackends(RTLDRMOD hPlugin, PCVDCACHEBACKEND *ppBackends, unsigned cBackends)
710{
711 PCVDCACHEBACKEND *pTmp = (PCVDCACHEBACKEND*)RTMemRealloc(g_apCacheBackends,
712 (g_cCacheBackends + cBackends) * sizeof(PCVDCACHEBACKEND));
713 if (RT_UNLIKELY(!pTmp))
714 return VERR_NO_MEMORY;
715 g_apCacheBackends = pTmp;
716
717 RTLDRMOD *pTmpPlugins = (RTLDRMOD*)RTMemRealloc(g_ahCacheBackendPlugins,
718 (g_cCacheBackends + cBackends) * sizeof(RTLDRMOD));
719 if (RT_UNLIKELY(!pTmpPlugins))
720 return VERR_NO_MEMORY;
721 g_ahCacheBackendPlugins = pTmpPlugins;
722 memcpy(&g_apCacheBackends[g_cCacheBackends], ppBackends, cBackends * sizeof(PCVDCACHEBACKEND));
723 for (unsigned i = g_cCacheBackends; i < g_cCacheBackends + cBackends; i++)
724 g_ahCacheBackendPlugins[i] = hPlugin;
725 g_cCacheBackends += cBackends;
726 return VINF_SUCCESS;
727}
728
729#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
730
731/**
732 * internal: add single cache backend.
733 */
734DECLINLINE(int) vdAddCacheBackend(RTLDRMOD hPlugin, PCVDCACHEBACKEND pBackend)
735{
736 return vdAddCacheBackends(hPlugin, &pBackend, 1);
737}
738
739
740/**
741 * Add several filter backends.
742 *
743 * @returns VBox status code.
744 * @param hPlugin Plugin handle to add.
745 * @param ppBackends Array of filter backends to add.
746 * @param cBackends Number of backends to add.
747 */
748static int vdAddFilterBackends(RTLDRMOD hPlugin, PCVDFILTERBACKEND *ppBackends, unsigned cBackends)
749{
750 PCVDFILTERBACKEND *pTmp = (PCVDFILTERBACKEND *)RTMemRealloc(g_apFilterBackends,
751 (g_cFilterBackends + cBackends) * sizeof(PCVDFILTERBACKEND));
752 if (RT_UNLIKELY(!pTmp))
753 return VERR_NO_MEMORY;
754 g_apFilterBackends = pTmp;
755
756 PRTLDRMOD pTmpPlugins = (PRTLDRMOD)RTMemRealloc(g_pahFilterBackendPlugins,
757 (g_cFilterBackends + cBackends) * sizeof(RTLDRMOD));
758 if (RT_UNLIKELY(!pTmpPlugins))
759 return VERR_NO_MEMORY;
760
761 g_pahFilterBackendPlugins = pTmpPlugins;
762 memcpy(&g_apFilterBackends[g_cFilterBackends], ppBackends, cBackends * sizeof(PCVDFILTERBACKEND));
763 for (unsigned i = g_cFilterBackends; i < g_cFilterBackends + cBackends; i++)
764 g_pahFilterBackendPlugins[i] = hPlugin;
765 g_cFilterBackends += cBackends;
766 return VINF_SUCCESS;
767}
768
769
770/**
771 * Add a single filter backend to the list of supported filters.
772 *
773 * @returns VBox status code.
774 * @param hPlugin Plugin handle to add.
775 * @param pBackend The backend to add.
776 */
777DECLINLINE(int) vdAddFilterBackend(RTLDRMOD hPlugin, PCVDFILTERBACKEND pBackend)
778{
779 return vdAddFilterBackends(hPlugin, &pBackend, 1);
780}
781
782#endif /* VBOX_HDD_NO_DYNAMIC_BACKENDS*/
783
784/**
785 * internal: issue error message.
786 */
787static int vdError(PVBOXHDD pDisk, int rc, RT_SRC_POS_DECL,
788 const char *pszFormat, ...)
789{
790 va_list va;
791 va_start(va, pszFormat);
792 if (pDisk->pInterfaceError)
793 pDisk->pInterfaceError->pfnError(pDisk->pInterfaceError->Core.pvUser, rc, RT_SRC_POS_ARGS, pszFormat, va);
794 va_end(va);
795 return rc;
796}
797
798/**
799 * internal: thread synchronization, start read.
800 */
801DECLINLINE(int) vdThreadStartRead(PVBOXHDD pDisk)
802{
803 int rc = VINF_SUCCESS;
804 if (RT_UNLIKELY(pDisk->pInterfaceThreadSync))
805 rc = pDisk->pInterfaceThreadSync->pfnStartRead(pDisk->pInterfaceThreadSync->Core.pvUser);
806 return rc;
807}
808
809/**
810 * internal: thread synchronization, finish read.
811 */
812DECLINLINE(int) vdThreadFinishRead(PVBOXHDD pDisk)
813{
814 int rc = VINF_SUCCESS;
815 if (RT_UNLIKELY(pDisk->pInterfaceThreadSync))
816 rc = pDisk->pInterfaceThreadSync->pfnFinishRead(pDisk->pInterfaceThreadSync->Core.pvUser);
817 return rc;
818}
819
820/**
821 * internal: thread synchronization, start write.
822 */
823DECLINLINE(int) vdThreadStartWrite(PVBOXHDD pDisk)
824{
825 int rc = VINF_SUCCESS;
826 if (RT_UNLIKELY(pDisk->pInterfaceThreadSync))
827 rc = pDisk->pInterfaceThreadSync->pfnStartWrite(pDisk->pInterfaceThreadSync->Core.pvUser);
828 return rc;
829}
830
831/**
832 * internal: thread synchronization, finish write.
833 */
834DECLINLINE(int) vdThreadFinishWrite(PVBOXHDD pDisk)
835{
836 int rc = VINF_SUCCESS;
837 if (RT_UNLIKELY(pDisk->pInterfaceThreadSync))
838 rc = pDisk->pInterfaceThreadSync->pfnFinishWrite(pDisk->pInterfaceThreadSync->Core.pvUser);
839 return rc;
840}
841
842/**
843 * internal: find image format backend.
844 */
845static int vdFindBackend(const char *pszBackend, PCVDIMAGEBACKEND *ppBackend)
846{
847 int rc = VINF_SUCCESS;
848 PCVDIMAGEBACKEND pBackend = NULL;
849
850 if (!g_apBackends)
851 VDInit();
852
853 for (unsigned i = 0; i < g_cBackends; i++)
854 {
855 if (!RTStrICmp(pszBackend, g_apBackends[i]->pszBackendName))
856 {
857 pBackend = g_apBackends[i];
858 break;
859 }
860 }
861 *ppBackend = pBackend;
862 return rc;
863}
864
865/**
866 * internal: find cache format backend.
867 */
868static int vdFindCacheBackend(const char *pszBackend, PCVDCACHEBACKEND *ppBackend)
869{
870 int rc = VINF_SUCCESS;
871 PCVDCACHEBACKEND pBackend = NULL;
872
873 if (!g_apCacheBackends)
874 VDInit();
875
876 for (unsigned i = 0; i < g_cCacheBackends; i++)
877 {
878 if (!RTStrICmp(pszBackend, g_apCacheBackends[i]->pszBackendName))
879 {
880 pBackend = g_apCacheBackends[i];
881 break;
882 }
883 }
884 *ppBackend = pBackend;
885 return rc;
886}
887
888/**
889 * internal: find filter backend.
890 */
891static int vdFindFilterBackend(const char *pszFilter, PCVDFILTERBACKEND *ppBackend)
892{
893 int rc = VINF_SUCCESS;
894 PCVDFILTERBACKEND pBackend = NULL;
895
896 for (unsigned i = 0; i < g_cFilterBackends; i++)
897 {
898 if (!RTStrICmp(pszFilter, g_apFilterBackends[i]->pszBackendName))
899 {
900 pBackend = g_apFilterBackends[i];
901 break;
902 }
903 }
904 *ppBackend = pBackend;
905 return rc;
906}
907
908
909/**
910 * internal: add image structure to the end of images list.
911 */
912static void vdAddImageToList(PVBOXHDD pDisk, PVDIMAGE pImage)
913{
914 pImage->pPrev = NULL;
915 pImage->pNext = NULL;
916
917 if (pDisk->pBase)
918 {
919 Assert(pDisk->cImages > 0);
920 pImage->pPrev = pDisk->pLast;
921 pDisk->pLast->pNext = pImage;
922 pDisk->pLast = pImage;
923 }
924 else
925 {
926 Assert(pDisk->cImages == 0);
927 pDisk->pBase = pImage;
928 pDisk->pLast = pImage;
929 }
930
931 pDisk->cImages++;
932}
933
934/**
935 * internal: remove image structure from the images list.
936 */
937static void vdRemoveImageFromList(PVBOXHDD pDisk, PVDIMAGE pImage)
938{
939 Assert(pDisk->cImages > 0);
940
941 if (pImage->pPrev)
942 pImage->pPrev->pNext = pImage->pNext;
943 else
944 pDisk->pBase = pImage->pNext;
945
946 if (pImage->pNext)
947 pImage->pNext->pPrev = pImage->pPrev;
948 else
949 pDisk->pLast = pImage->pPrev;
950
951 pImage->pPrev = NULL;
952 pImage->pNext = NULL;
953
954 pDisk->cImages--;
955}
956
957/**
958 * Release a referene to the filter decrementing the counter and destroying the filter
959 * when the counter reaches zero.
960 *
961 * @returns The new reference count.
962 * @param pFilter The filter to release.
963 */
964static uint32_t vdFilterRelease(PVDFILTER pFilter)
965{
966 uint32_t cRefs = ASMAtomicDecU32(&pFilter->cRefs);
967 if (!cRefs)
968 {
969 pFilter->pBackend->pfnDestroy(pFilter->pvBackendData);
970 RTMemFree(pFilter);
971 }
972
973 return cRefs;
974}
975
976/**
977 * Increments the reference counter of the given filter.
978 *
979 * @return The new reference count.
980 * @param pFilter The filter.
981 */
982static uint32_t vdFilterRetain(PVDFILTER pFilter)
983{
984 return ASMAtomicIncU32(&pFilter->cRefs);
985}
986
987/**
988 * internal: find image by index into the images list.
989 */
990static PVDIMAGE vdGetImageByNumber(PVBOXHDD pDisk, unsigned nImage)
991{
992 PVDIMAGE pImage = pDisk->pBase;
993 if (nImage == VD_LAST_IMAGE)
994 return pDisk->pLast;
995 while (pImage && nImage)
996 {
997 pImage = pImage->pNext;
998 nImage--;
999 }
1000 return pImage;
1001}
1002
1003/**
1004 * Applies the filter chain to the given write request.
1005 *
1006 * @returns VBox status code.
1007 * @param pDisk The HDD container.
1008 * @param uOffset The start offset of the write.
1009 * @param cbWrite Number of bytes to write.
1010 * @param pIoCtx The I/O context associated with the request.
1011 */
1012static int vdFilterChainApplyWrite(PVBOXHDD pDisk, uint64_t uOffset, size_t cbWrite,
1013 PVDIOCTX pIoCtx)
1014{
1015 int rc = VINF_SUCCESS;
1016
1017 VD_IS_LOCKED(pDisk);
1018
1019 PVDFILTER pFilter;
1020 RTListForEach(&pDisk->ListFilterChainWrite, pFilter, VDFILTER, ListNodeChainWrite)
1021 {
1022 rc = pFilter->pBackend->pfnFilterWrite(pFilter->pvBackendData, uOffset, cbWrite, pIoCtx);
1023 if (RT_FAILURE(rc))
1024 break;
1025 /* Reset S/G buffer for the next filter. */
1026 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
1027 }
1028
1029 return rc;
1030}
1031
1032/**
1033 * Applies the filter chain to the given read request.
1034 *
1035 * @returns VBox status code.
1036 * @param pDisk The HDD container.
1037 * @param uOffset The start offset of the read.
1038 * @param cbRead Number of bytes read.
1039 * @param pIoCtx The I/O context associated with the request.
1040 */
1041static int vdFilterChainApplyRead(PVBOXHDD pDisk, uint64_t uOffset, size_t cbRead,
1042 PVDIOCTX pIoCtx)
1043{
1044 int rc = VINF_SUCCESS;
1045
1046 VD_IS_LOCKED(pDisk);
1047
1048 /* Reset buffer before starting. */
1049 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
1050
1051 PVDFILTER pFilter;
1052 RTListForEach(&pDisk->ListFilterChainRead, pFilter, VDFILTER, ListNodeChainRead)
1053 {
1054 rc = pFilter->pBackend->pfnFilterRead(pFilter->pvBackendData, uOffset, cbRead, pIoCtx);
1055 if (RT_FAILURE(rc))
1056 break;
1057 /* Reset S/G buffer for the next filter. */
1058 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
1059 }
1060
1061 return rc;
1062}
1063
1064DECLINLINE(void) vdIoCtxRootComplete(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
1065{
1066 if ( RT_SUCCESS(pIoCtx->rcReq)
1067 && pIoCtx->enmTxDir == VDIOCTXTXDIR_READ)
1068 pIoCtx->rcReq = vdFilterChainApplyRead(pDisk, pIoCtx->Req.Io.uOffsetXferOrig,
1069 pIoCtx->Req.Io.cbXferOrig, pIoCtx);
1070
1071 pIoCtx->Type.Root.pfnComplete(pIoCtx->Type.Root.pvUser1,
1072 pIoCtx->Type.Root.pvUser2,
1073 pIoCtx->rcReq);
1074}
1075
1076/**
1077 * Initialize the structure members of a given I/O context.
1078 */
1079DECLINLINE(void) vdIoCtxInit(PVDIOCTX pIoCtx, PVBOXHDD pDisk, VDIOCTXTXDIR enmTxDir,
1080 uint64_t uOffset, size_t cbTransfer, PVDIMAGE pImageStart,
1081 PCRTSGBUF pcSgBuf, void *pvAllocation,
1082 PFNVDIOCTXTRANSFER pfnIoCtxTransfer, uint32_t fFlags)
1083{
1084 pIoCtx->pDisk = pDisk;
1085 pIoCtx->enmTxDir = enmTxDir;
1086 pIoCtx->Req.Io.cbTransferLeft = (uint32_t)cbTransfer; Assert((uint32_t)cbTransfer == cbTransfer);
1087 pIoCtx->Req.Io.uOffset = uOffset;
1088 pIoCtx->Req.Io.cbTransfer = cbTransfer;
1089 pIoCtx->Req.Io.pImageStart = pImageStart;
1090 pIoCtx->Req.Io.pImageCur = pImageStart;
1091 pIoCtx->Req.Io.cbBufClear = 0;
1092 pIoCtx->Req.Io.pImageParentOverride = NULL;
1093 pIoCtx->Req.Io.uOffsetXferOrig = uOffset;
1094 pIoCtx->Req.Io.cbXferOrig = cbTransfer;
1095 pIoCtx->cDataTransfersPending = 0;
1096 pIoCtx->cMetaTransfersPending = 0;
1097 pIoCtx->fComplete = false;
1098 pIoCtx->fFlags = fFlags;
1099 pIoCtx->pvAllocation = pvAllocation;
1100 pIoCtx->pfnIoCtxTransfer = pfnIoCtxTransfer;
1101 pIoCtx->pfnIoCtxTransferNext = NULL;
1102 pIoCtx->rcReq = VINF_SUCCESS;
1103 pIoCtx->pIoCtxParent = NULL;
1104
1105 /* There is no S/G list for a flush request. */
1106 if ( enmTxDir != VDIOCTXTXDIR_FLUSH
1107 && enmTxDir != VDIOCTXTXDIR_DISCARD)
1108 RTSgBufClone(&pIoCtx->Req.Io.SgBuf, pcSgBuf);
1109 else
1110 memset(&pIoCtx->Req.Io.SgBuf, 0, sizeof(RTSGBUF));
1111}
1112
1113/**
1114 * Internal: Tries to read the desired range from the given cache.
1115 *
1116 * @returns VBox status code.
1117 * @retval VERR_VD_BLOCK_FREE if the block is not in the cache.
1118 * pcbRead will be set to the number of bytes not in the cache.
1119 * Everything thereafter might be in the cache.
1120 * @param pCache The cache to read from.
1121 * @param uOffset Offset of the virtual disk to read.
1122 * @param cbRead How much to read.
1123 * @param pIoCtx The I/O context to read into.
1124 * @param pcbRead Where to store the number of bytes actually read.
1125 * On success this indicates the number of bytes read from the cache.
1126 * If VERR_VD_BLOCK_FREE is returned this gives the number of bytes
1127 * which are not in the cache.
1128 * In both cases everything beyond this value
1129 * might or might not be in the cache.
1130 */
1131static int vdCacheReadHelper(PVDCACHE pCache, uint64_t uOffset,
1132 size_t cbRead, PVDIOCTX pIoCtx, size_t *pcbRead)
1133{
1134 int rc = VINF_SUCCESS;
1135
1136 LogFlowFunc(("pCache=%#p uOffset=%llu pIoCtx=%p cbRead=%zu pcbRead=%#p\n",
1137 pCache, uOffset, pIoCtx, cbRead, pcbRead));
1138
1139 AssertPtr(pCache);
1140 AssertPtr(pcbRead);
1141
1142 rc = pCache->Backend->pfnRead(pCache->pBackendData, uOffset, cbRead,
1143 pIoCtx, pcbRead);
1144
1145 LogFlowFunc(("returns rc=%Rrc pcbRead=%zu\n", rc, *pcbRead));
1146 return rc;
1147}
1148
1149/**
1150 * Internal: Writes data for the given block into the cache.
1151 *
1152 * @returns VBox status code.
1153 * @param pCache The cache to write to.
1154 * @param uOffset Offset of the virtual disk to write to the cache.
1155 * @param cbWrite How much to write.
1156 * @param pIoCtx The I/O context to write from.
1157 * @param pcbWritten How much data could be written, optional.
1158 */
1159static int vdCacheWriteHelper(PVDCACHE pCache, uint64_t uOffset, size_t cbWrite,
1160 PVDIOCTX pIoCtx, size_t *pcbWritten)
1161{
1162 int rc = VINF_SUCCESS;
1163
1164 LogFlowFunc(("pCache=%#p uOffset=%llu pIoCtx=%p cbWrite=%zu pcbWritten=%#p\n",
1165 pCache, uOffset, pIoCtx, cbWrite, pcbWritten));
1166
1167 AssertPtr(pCache);
1168 AssertPtr(pIoCtx);
1169 Assert(cbWrite > 0);
1170
1171 if (pcbWritten)
1172 rc = pCache->Backend->pfnWrite(pCache->pBackendData, uOffset, cbWrite,
1173 pIoCtx, pcbWritten);
1174 else
1175 {
1176 size_t cbWritten = 0;
1177
1178 do
1179 {
1180 rc = pCache->Backend->pfnWrite(pCache->pBackendData, uOffset, cbWrite,
1181 pIoCtx, &cbWritten);
1182 uOffset += cbWritten;
1183 cbWrite -= cbWritten;
1184 } while ( cbWrite
1185 && ( RT_SUCCESS(rc)
1186 || rc == VERR_VD_ASYNC_IO_IN_PROGRESS));
1187 }
1188
1189 LogFlowFunc(("returns rc=%Rrc pcbWritten=%zu\n",
1190 rc, pcbWritten ? *pcbWritten : cbWrite));
1191 return rc;
1192}
1193
1194/**
1195 * Creates a new empty discard state.
1196 *
1197 * @returns Pointer to the new discard state or NULL if out of memory.
1198 */
1199static PVDDISCARDSTATE vdDiscardStateCreate(void)
1200{
1201 PVDDISCARDSTATE pDiscard = (PVDDISCARDSTATE)RTMemAllocZ(sizeof(VDDISCARDSTATE));
1202
1203 if (pDiscard)
1204 {
1205 RTListInit(&pDiscard->ListLru);
1206 pDiscard->pTreeBlocks = (PAVLRU64TREE)RTMemAllocZ(sizeof(AVLRU64TREE));
1207 if (!pDiscard->pTreeBlocks)
1208 {
1209 RTMemFree(pDiscard);
1210 pDiscard = NULL;
1211 }
1212 }
1213
1214 return pDiscard;
1215}
1216
1217/**
1218 * Removes the least recently used blocks from the waiting list until
1219 * the new value is reached.
1220 *
1221 * @returns VBox status code.
1222 * @param pDisk VD disk container.
1223 * @param pDiscard The discard state.
1224 * @param cbDiscardingNew How many bytes should be waiting on success.
1225 * The number of bytes waiting can be less.
1226 */
1227static int vdDiscardRemoveBlocks(PVBOXHDD pDisk, PVDDISCARDSTATE pDiscard, size_t cbDiscardingNew)
1228{
1229 int rc = VINF_SUCCESS;
1230
1231 LogFlowFunc(("pDisk=%#p pDiscard=%#p cbDiscardingNew=%zu\n",
1232 pDisk, pDiscard, cbDiscardingNew));
1233
1234 while (pDiscard->cbDiscarding > cbDiscardingNew)
1235 {
1236 PVDDISCARDBLOCK pBlock = RTListGetLast(&pDiscard->ListLru, VDDISCARDBLOCK, NodeLru);
1237
1238 Assert(!RTListIsEmpty(&pDiscard->ListLru));
1239
1240 /* Go over the allocation bitmap and mark all discarded sectors as unused. */
1241 uint64_t offStart = pBlock->Core.Key;
1242 uint32_t idxStart = 0;
1243 size_t cbLeft = pBlock->cbDiscard;
1244 bool fAllocated = ASMBitTest(pBlock->pbmAllocated, idxStart);
1245 uint32_t cSectors = (uint32_t)(pBlock->cbDiscard / 512);
1246
1247 while (cbLeft > 0)
1248 {
1249 int32_t idxEnd;
1250 size_t cbThis = cbLeft;
1251
1252 if (fAllocated)
1253 {
1254 /* Check for the first unallocated bit. */
1255 idxEnd = ASMBitNextClear(pBlock->pbmAllocated, cSectors, idxStart);
1256 if (idxEnd != -1)
1257 {
1258 cbThis = (idxEnd - idxStart) * 512;
1259 fAllocated = false;
1260 }
1261 }
1262 else
1263 {
1264 /* Mark as unused and check for the first set bit. */
1265 idxEnd = ASMBitNextSet(pBlock->pbmAllocated, cSectors, idxStart);
1266 if (idxEnd != -1)
1267 cbThis = (idxEnd - idxStart) * 512;
1268
1269
1270 VDIOCTX IoCtx;
1271 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_DISCARD, 0, 0, NULL,
1272 NULL, NULL, NULL, VDIOCTX_FLAGS_SYNC);
1273 rc = pDisk->pLast->Backend->pfnDiscard(pDisk->pLast->pBackendData,
1274 &IoCtx, offStart, cbThis, NULL,
1275 NULL, &cbThis, NULL,
1276 VD_DISCARD_MARK_UNUSED);
1277 if (RT_FAILURE(rc))
1278 break;
1279
1280 fAllocated = true;
1281 }
1282
1283 idxStart = idxEnd;
1284 offStart += cbThis;
1285 cbLeft -= cbThis;
1286 }
1287
1288 if (RT_FAILURE(rc))
1289 break;
1290
1291 PVDDISCARDBLOCK pBlockRemove = (PVDDISCARDBLOCK)RTAvlrU64RangeRemove(pDiscard->pTreeBlocks, pBlock->Core.Key);
1292 Assert(pBlockRemove == pBlock); NOREF(pBlockRemove);
1293 RTListNodeRemove(&pBlock->NodeLru);
1294
1295 pDiscard->cbDiscarding -= pBlock->cbDiscard;
1296 RTMemFree(pBlock->pbmAllocated);
1297 RTMemFree(pBlock);
1298 }
1299
1300 Assert(RT_FAILURE(rc) || pDiscard->cbDiscarding <= cbDiscardingNew);
1301
1302 LogFlowFunc(("returns rc=%Rrc\n", rc));
1303 return rc;
1304}
1305
1306/**
1307 * Destroys the current discard state, writing any waiting blocks to the image.
1308 *
1309 * @returns VBox status code.
1310 * @param pDisk VD disk container.
1311 */
1312static int vdDiscardStateDestroy(PVBOXHDD pDisk)
1313{
1314 int rc = VINF_SUCCESS;
1315
1316 if (pDisk->pDiscard)
1317 {
1318 rc = vdDiscardRemoveBlocks(pDisk, pDisk->pDiscard, 0 /* Remove all blocks. */);
1319 AssertRC(rc);
1320 RTMemFree(pDisk->pDiscard->pTreeBlocks);
1321 RTMemFree(pDisk->pDiscard);
1322 pDisk->pDiscard = NULL;
1323 }
1324
1325 return rc;
1326}
1327
1328/**
1329 * Marks the given range as allocated in the image.
1330 * Required if there are discards in progress and a write to a block which can get discarded
1331 * is written to.
1332 *
1333 * @returns VBox status code.
1334 * @param pDisk VD container data.
1335 * @param uOffset First byte to mark as allocated.
1336 * @param cbRange Number of bytes to mark as allocated.
1337 */
1338static int vdDiscardSetRangeAllocated(PVBOXHDD pDisk, uint64_t uOffset, size_t cbRange)
1339{
1340 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
1341 int rc = VINF_SUCCESS;
1342
1343 if (pDiscard)
1344 {
1345 do
1346 {
1347 size_t cbThisRange = cbRange;
1348 PVDDISCARDBLOCK pBlock = (PVDDISCARDBLOCK)RTAvlrU64RangeGet(pDiscard->pTreeBlocks, uOffset);
1349
1350 if (pBlock)
1351 {
1352 int32_t idxStart, idxEnd;
1353
1354 Assert(!(cbThisRange % 512));
1355 Assert(!((uOffset - pBlock->Core.Key) % 512));
1356
1357 cbThisRange = RT_MIN(cbThisRange, pBlock->Core.KeyLast - uOffset + 1);
1358
1359 idxStart = (uOffset - pBlock->Core.Key) / 512;
1360 idxEnd = idxStart + (int32_t)(cbThisRange / 512);
1361 ASMBitSetRange(pBlock->pbmAllocated, idxStart, idxEnd);
1362 }
1363 else
1364 {
1365 pBlock = (PVDDISCARDBLOCK)RTAvlrU64GetBestFit(pDiscard->pTreeBlocks, uOffset, true);
1366 if (pBlock)
1367 cbThisRange = RT_MIN(cbThisRange, pBlock->Core.Key - uOffset);
1368 }
1369
1370 Assert(cbRange >= cbThisRange);
1371
1372 uOffset += cbThisRange;
1373 cbRange -= cbThisRange;
1374 } while (cbRange != 0);
1375 }
1376
1377 return rc;
1378}
1379
1380DECLINLINE(PVDIOCTX) vdIoCtxAlloc(PVBOXHDD pDisk, VDIOCTXTXDIR enmTxDir,
1381 uint64_t uOffset, size_t cbTransfer,
1382 PVDIMAGE pImageStart,PCRTSGBUF pcSgBuf,
1383 void *pvAllocation, PFNVDIOCTXTRANSFER pfnIoCtxTransfer,
1384 uint32_t fFlags)
1385{
1386 PVDIOCTX pIoCtx = NULL;
1387
1388 pIoCtx = (PVDIOCTX)RTMemCacheAlloc(pDisk->hMemCacheIoCtx);
1389 if (RT_LIKELY(pIoCtx))
1390 {
1391 vdIoCtxInit(pIoCtx, pDisk, enmTxDir, uOffset, cbTransfer, pImageStart,
1392 pcSgBuf, pvAllocation, pfnIoCtxTransfer, fFlags);
1393 }
1394
1395 return pIoCtx;
1396}
1397
1398DECLINLINE(PVDIOCTX) vdIoCtxRootAlloc(PVBOXHDD pDisk, VDIOCTXTXDIR enmTxDir,
1399 uint64_t uOffset, size_t cbTransfer,
1400 PVDIMAGE pImageStart, PCRTSGBUF pcSgBuf,
1401 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
1402 void *pvUser1, void *pvUser2,
1403 void *pvAllocation,
1404 PFNVDIOCTXTRANSFER pfnIoCtxTransfer,
1405 uint32_t fFlags)
1406{
1407 PVDIOCTX pIoCtx = vdIoCtxAlloc(pDisk, enmTxDir, uOffset, cbTransfer, pImageStart,
1408 pcSgBuf, pvAllocation, pfnIoCtxTransfer, fFlags);
1409
1410 if (RT_LIKELY(pIoCtx))
1411 {
1412 pIoCtx->pIoCtxParent = NULL;
1413 pIoCtx->Type.Root.pfnComplete = pfnComplete;
1414 pIoCtx->Type.Root.pvUser1 = pvUser1;
1415 pIoCtx->Type.Root.pvUser2 = pvUser2;
1416 }
1417
1418 LogFlow(("Allocated root I/O context %#p\n", pIoCtx));
1419 return pIoCtx;
1420}
1421
1422DECLINLINE(void) vdIoCtxDiscardInit(PVDIOCTX pIoCtx, PVBOXHDD pDisk, PCRTRANGE paRanges,
1423 unsigned cRanges, PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
1424 void *pvUser1, void *pvUser2, void *pvAllocation,
1425 PFNVDIOCTXTRANSFER pfnIoCtxTransfer, uint32_t fFlags)
1426{
1427 pIoCtx->pIoCtxNext = NULL;
1428 pIoCtx->pDisk = pDisk;
1429 pIoCtx->enmTxDir = VDIOCTXTXDIR_DISCARD;
1430 pIoCtx->cDataTransfersPending = 0;
1431 pIoCtx->cMetaTransfersPending = 0;
1432 pIoCtx->fComplete = false;
1433 pIoCtx->fFlags = fFlags;
1434 pIoCtx->pvAllocation = pvAllocation;
1435 pIoCtx->pfnIoCtxTransfer = pfnIoCtxTransfer;
1436 pIoCtx->pfnIoCtxTransferNext = NULL;
1437 pIoCtx->rcReq = VINF_SUCCESS;
1438 pIoCtx->Req.Discard.paRanges = paRanges;
1439 pIoCtx->Req.Discard.cRanges = cRanges;
1440 pIoCtx->Req.Discard.idxRange = 0;
1441 pIoCtx->Req.Discard.cbDiscardLeft = 0;
1442 pIoCtx->Req.Discard.offCur = 0;
1443 pIoCtx->Req.Discard.cbThisDiscard = 0;
1444
1445 pIoCtx->pIoCtxParent = NULL;
1446 pIoCtx->Type.Root.pfnComplete = pfnComplete;
1447 pIoCtx->Type.Root.pvUser1 = pvUser1;
1448 pIoCtx->Type.Root.pvUser2 = pvUser2;
1449}
1450
1451DECLINLINE(PVDIOCTX) vdIoCtxDiscardAlloc(PVBOXHDD pDisk, PCRTRANGE paRanges,
1452 unsigned cRanges,
1453 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
1454 void *pvUser1, void *pvUser2,
1455 void *pvAllocation,
1456 PFNVDIOCTXTRANSFER pfnIoCtxTransfer,
1457 uint32_t fFlags)
1458{
1459 PVDIOCTX pIoCtx = NULL;
1460
1461 pIoCtx = (PVDIOCTX)RTMemCacheAlloc(pDisk->hMemCacheIoCtx);
1462 if (RT_LIKELY(pIoCtx))
1463 {
1464 vdIoCtxDiscardInit(pIoCtx, pDisk, paRanges, cRanges, pfnComplete, pvUser1,
1465 pvUser2, pvAllocation, pfnIoCtxTransfer, fFlags);
1466 }
1467
1468 LogFlow(("Allocated discard I/O context %#p\n", pIoCtx));
1469 return pIoCtx;
1470}
1471
1472DECLINLINE(PVDIOCTX) vdIoCtxChildAlloc(PVBOXHDD pDisk, VDIOCTXTXDIR enmTxDir,
1473 uint64_t uOffset, size_t cbTransfer,
1474 PVDIMAGE pImageStart, PCRTSGBUF pcSgBuf,
1475 PVDIOCTX pIoCtxParent, size_t cbTransferParent,
1476 size_t cbWriteParent, void *pvAllocation,
1477 PFNVDIOCTXTRANSFER pfnIoCtxTransfer)
1478{
1479 PVDIOCTX pIoCtx = vdIoCtxAlloc(pDisk, enmTxDir, uOffset, cbTransfer, pImageStart,
1480 pcSgBuf, pvAllocation, pfnIoCtxTransfer, pIoCtxParent->fFlags & ~VDIOCTX_FLAGS_DONT_FREE);
1481
1482 AssertPtr(pIoCtxParent);
1483 Assert(!pIoCtxParent->pIoCtxParent);
1484
1485 if (RT_LIKELY(pIoCtx))
1486 {
1487 pIoCtx->pIoCtxParent = pIoCtxParent;
1488 pIoCtx->Type.Child.uOffsetSaved = uOffset;
1489 pIoCtx->Type.Child.cbTransferLeftSaved = cbTransfer;
1490 pIoCtx->Type.Child.cbTransferParent = cbTransferParent;
1491 pIoCtx->Type.Child.cbWriteParent = cbWriteParent;
1492 }
1493
1494 LogFlow(("Allocated child I/O context %#p\n", pIoCtx));
1495 return pIoCtx;
1496}
1497
1498DECLINLINE(PVDIOTASK) vdIoTaskUserAlloc(PVDIOSTORAGE pIoStorage, PFNVDXFERCOMPLETED pfnComplete, void *pvUser, PVDIOCTX pIoCtx, uint32_t cbTransfer)
1499{
1500 PVDIOTASK pIoTask = NULL;
1501
1502 pIoTask = (PVDIOTASK)RTMemCacheAlloc(pIoStorage->pVDIo->pDisk->hMemCacheIoTask);
1503 if (pIoTask)
1504 {
1505 pIoTask->pIoStorage = pIoStorage;
1506 pIoTask->pfnComplete = pfnComplete;
1507 pIoTask->pvUser = pvUser;
1508 pIoTask->fMeta = false;
1509 pIoTask->Type.User.cbTransfer = cbTransfer;
1510 pIoTask->Type.User.pIoCtx = pIoCtx;
1511 }
1512
1513 return pIoTask;
1514}
1515
1516DECLINLINE(PVDIOTASK) vdIoTaskMetaAlloc(PVDIOSTORAGE pIoStorage, PFNVDXFERCOMPLETED pfnComplete, void *pvUser, PVDMETAXFER pMetaXfer)
1517{
1518 PVDIOTASK pIoTask = NULL;
1519
1520 pIoTask = (PVDIOTASK)RTMemCacheAlloc(pIoStorage->pVDIo->pDisk->hMemCacheIoTask);
1521 if (pIoTask)
1522 {
1523 pIoTask->pIoStorage = pIoStorage;
1524 pIoTask->pfnComplete = pfnComplete;
1525 pIoTask->pvUser = pvUser;
1526 pIoTask->fMeta = true;
1527 pIoTask->Type.Meta.pMetaXfer = pMetaXfer;
1528 }
1529
1530 return pIoTask;
1531}
1532
1533DECLINLINE(void) vdIoCtxFree(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
1534{
1535 Log(("Freeing I/O context %#p\n", pIoCtx));
1536
1537 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_DONT_FREE))
1538 {
1539 if (pIoCtx->pvAllocation)
1540 RTMemFree(pIoCtx->pvAllocation);
1541#ifdef DEBUG
1542 memset(&pIoCtx->pDisk, 0xff, sizeof(void *));
1543#endif
1544 RTMemCacheFree(pDisk->hMemCacheIoCtx, pIoCtx);
1545 }
1546}
1547
1548DECLINLINE(void) vdIoTaskFree(PVBOXHDD pDisk, PVDIOTASK pIoTask)
1549{
1550#ifdef DEBUG
1551 memset(pIoTask, 0xff, sizeof(VDIOTASK));
1552#endif
1553 RTMemCacheFree(pDisk->hMemCacheIoTask, pIoTask);
1554}
1555
1556DECLINLINE(void) vdIoCtxChildReset(PVDIOCTX pIoCtx)
1557{
1558 AssertPtr(pIoCtx->pIoCtxParent);
1559
1560 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
1561 pIoCtx->Req.Io.uOffset = pIoCtx->Type.Child.uOffsetSaved;
1562 pIoCtx->Req.Io.cbTransferLeft = (uint32_t)pIoCtx->Type.Child.cbTransferLeftSaved;
1563 Assert((uint32_t)pIoCtx->Type.Child.cbTransferLeftSaved == pIoCtx->Type.Child.cbTransferLeftSaved);
1564}
1565
1566DECLINLINE(PVDMETAXFER) vdMetaXferAlloc(PVDIOSTORAGE pIoStorage, uint64_t uOffset, size_t cb)
1567{
1568 PVDMETAXFER pMetaXfer = (PVDMETAXFER)RTMemAlloc(RT_OFFSETOF(VDMETAXFER, abData[cb]));
1569
1570 if (RT_LIKELY(pMetaXfer))
1571 {
1572 pMetaXfer->Core.Key = uOffset;
1573 pMetaXfer->Core.KeyLast = uOffset + cb - 1;
1574 pMetaXfer->fFlags = VDMETAXFER_TXDIR_NONE;
1575 pMetaXfer->cbMeta = cb;
1576 pMetaXfer->pIoStorage = pIoStorage;
1577 pMetaXfer->cRefs = 0;
1578 pMetaXfer->pbDataShw = NULL;
1579 RTListInit(&pMetaXfer->ListIoCtxWaiting);
1580 RTListInit(&pMetaXfer->ListIoCtxShwWrites);
1581 }
1582 return pMetaXfer;
1583}
1584
1585DECLINLINE(void) vdIoCtxAddToWaitingList(volatile PVDIOCTX *ppList, PVDIOCTX pIoCtx)
1586{
1587 /* Put it on the waiting list. */
1588 PVDIOCTX pNext = ASMAtomicUoReadPtrT(ppList, PVDIOCTX);
1589 PVDIOCTX pHeadOld;
1590 pIoCtx->pIoCtxNext = pNext;
1591 while (!ASMAtomicCmpXchgExPtr(ppList, pIoCtx, pNext, &pHeadOld))
1592 {
1593 pNext = pHeadOld;
1594 Assert(pNext != pIoCtx);
1595 pIoCtx->pIoCtxNext = pNext;
1596 ASMNopPause();
1597 }
1598}
1599
1600DECLINLINE(void) vdIoCtxDefer(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
1601{
1602 LogFlowFunc(("Deferring I/O context pIoCtx=%#p\n", pIoCtx));
1603
1604 Assert(!pIoCtx->pIoCtxParent && !(pIoCtx->fFlags & VDIOCTX_FLAGS_BLOCKED));
1605 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
1606 vdIoCtxAddToWaitingList(&pDisk->pIoCtxBlockedHead, pIoCtx);
1607}
1608
1609static size_t vdIoCtxCopy(PVDIOCTX pIoCtxDst, PVDIOCTX pIoCtxSrc, size_t cbData)
1610{
1611 return RTSgBufCopy(&pIoCtxDst->Req.Io.SgBuf, &pIoCtxSrc->Req.Io.SgBuf, cbData);
1612}
1613
1614#if 0 /* unused */
1615static int vdIoCtxCmp(PVDIOCTX pIoCtx1, PVDIOCTX pIoCtx2, size_t cbData)
1616{
1617 return RTSgBufCmp(&pIoCtx1->Req.Io.SgBuf, &pIoCtx2->Req.Io.SgBuf, cbData);
1618}
1619#endif
1620
1621static size_t vdIoCtxCopyTo(PVDIOCTX pIoCtx, const uint8_t *pbData, size_t cbData)
1622{
1623 return RTSgBufCopyFromBuf(&pIoCtx->Req.Io.SgBuf, pbData, cbData);
1624}
1625
1626static size_t vdIoCtxCopyFrom(PVDIOCTX pIoCtx, uint8_t *pbData, size_t cbData)
1627{
1628 return RTSgBufCopyToBuf(&pIoCtx->Req.Io.SgBuf, pbData, cbData);
1629}
1630
1631static size_t vdIoCtxSet(PVDIOCTX pIoCtx, uint8_t ch, size_t cbData)
1632{
1633 return RTSgBufSet(&pIoCtx->Req.Io.SgBuf, ch, cbData);
1634}
1635
1636/**
1637 * Returns whether the given I/O context has completed.
1638 *
1639 * @returns Flag whether the I/O context is complete.
1640 * @param pIoCtx The I/O context to check.
1641 */
1642DECLINLINE(bool) vdIoCtxIsComplete(PVDIOCTX pIoCtx)
1643{
1644 if ( !pIoCtx->cMetaTransfersPending
1645 && !pIoCtx->cDataTransfersPending
1646 && !pIoCtx->pfnIoCtxTransfer)
1647 return true;
1648
1649 /*
1650 * We complete the I/O context in case of an error
1651 * if there is no I/O task pending.
1652 */
1653 if ( RT_FAILURE(pIoCtx->rcReq)
1654 && !pIoCtx->cMetaTransfersPending
1655 && !pIoCtx->cDataTransfersPending)
1656 return true;
1657
1658 return false;
1659}
1660
1661/**
1662 * Returns whether the given I/O context is blocked due to a metadata transfer
1663 * or because the backend blocked it.
1664 *
1665 * @returns Flag whether the I/O context is blocked.
1666 * @param pIoCtx The I/O context to check.
1667 */
1668DECLINLINE(bool) vdIoCtxIsBlocked(PVDIOCTX pIoCtx)
1669{
1670 /* Don't change anything if there is a metadata transfer pending or we are blocked. */
1671 if ( pIoCtx->cMetaTransfersPending
1672 || (pIoCtx->fFlags & VDIOCTX_FLAGS_BLOCKED))
1673 return true;
1674
1675 return false;
1676}
1677
1678/**
1679 * Process the I/O context, core method which assumes that the I/O context
1680 * acquired the lock.
1681 *
1682 * @returns VBox status code.
1683 * @param pIoCtx I/O context to process.
1684 */
1685static int vdIoCtxProcessLocked(PVDIOCTX pIoCtx)
1686{
1687 int rc = VINF_SUCCESS;
1688
1689 VD_IS_LOCKED(pIoCtx->pDisk);
1690
1691 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
1692
1693 if (!vdIoCtxIsComplete(pIoCtx))
1694 {
1695 if (!vdIoCtxIsBlocked(pIoCtx))
1696 {
1697 if (pIoCtx->pfnIoCtxTransfer)
1698 {
1699 /* Call the transfer function advancing to the next while there is no error. */
1700 while ( pIoCtx->pfnIoCtxTransfer
1701 && !pIoCtx->cMetaTransfersPending
1702 && RT_SUCCESS(rc))
1703 {
1704 LogFlowFunc(("calling transfer function %#p\n", pIoCtx->pfnIoCtxTransfer));
1705 rc = pIoCtx->pfnIoCtxTransfer(pIoCtx);
1706
1707 /* Advance to the next part of the transfer if the current one succeeded. */
1708 if (RT_SUCCESS(rc))
1709 {
1710 pIoCtx->pfnIoCtxTransfer = pIoCtx->pfnIoCtxTransferNext;
1711 pIoCtx->pfnIoCtxTransferNext = NULL;
1712 }
1713 }
1714 }
1715
1716 if ( RT_SUCCESS(rc)
1717 && !pIoCtx->cMetaTransfersPending
1718 && !pIoCtx->cDataTransfersPending
1719 && !(pIoCtx->fFlags & VDIOCTX_FLAGS_BLOCKED))
1720 rc = VINF_VD_ASYNC_IO_FINISHED;
1721 else if ( RT_SUCCESS(rc)
1722 || rc == VERR_VD_NOT_ENOUGH_METADATA
1723 || rc == VERR_VD_IOCTX_HALT)
1724 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1725 else if ( RT_FAILURE(rc)
1726 && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
1727 {
1728 ASMAtomicCmpXchgS32(&pIoCtx->rcReq, rc, VINF_SUCCESS);
1729
1730 /*
1731 * The I/O context completed if we have an error and there is no data
1732 * or meta data transfer pending.
1733 */
1734 if ( !pIoCtx->cMetaTransfersPending
1735 && !pIoCtx->cDataTransfersPending)
1736 rc = VINF_VD_ASYNC_IO_FINISHED;
1737 else
1738 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1739 }
1740 }
1741 else
1742 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1743 }
1744 else
1745 rc = VINF_VD_ASYNC_IO_FINISHED;
1746
1747 LogFlowFunc(("pIoCtx=%#p rc=%Rrc cDataTransfersPending=%u cMetaTransfersPending=%u fComplete=%RTbool\n",
1748 pIoCtx, rc, pIoCtx->cDataTransfersPending, pIoCtx->cMetaTransfersPending,
1749 pIoCtx->fComplete));
1750
1751 return rc;
1752}
1753
1754/**
1755 * Processes the list of waiting I/O contexts.
1756 *
1757 * @returns VBox status code, only valid if pIoCtxRc is not NULL, treat as void
1758 * function otherwise.
1759 * @param pDisk The disk structure.
1760 * @param pIoCtxRc An I/O context handle which waits on the list. When processed
1761 * The status code is returned. NULL if there is no I/O context
1762 * to return the status code for.
1763 */
1764static int vdDiskProcessWaitingIoCtx(PVBOXHDD pDisk, PVDIOCTX pIoCtxRc)
1765{
1766 int rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1767
1768 LogFlowFunc(("pDisk=%#p pIoCtxRc=%#p\n", pDisk, pIoCtxRc));
1769
1770 VD_IS_LOCKED(pDisk);
1771
1772 /* Get the waiting list and process it in FIFO order. */
1773 PVDIOCTX pIoCtxHead = ASMAtomicXchgPtrT(&pDisk->pIoCtxHead, NULL, PVDIOCTX);
1774
1775 /* Reverse it. */
1776 PVDIOCTX pCur = pIoCtxHead;
1777 pIoCtxHead = NULL;
1778 while (pCur)
1779 {
1780 PVDIOCTX pInsert = pCur;
1781 pCur = pCur->pIoCtxNext;
1782 pInsert->pIoCtxNext = pIoCtxHead;
1783 pIoCtxHead = pInsert;
1784 }
1785
1786 /* Process now. */
1787 pCur = pIoCtxHead;
1788 while (pCur)
1789 {
1790 int rcTmp;
1791 PVDIOCTX pTmp = pCur;
1792
1793 pCur = pCur->pIoCtxNext;
1794 pTmp->pIoCtxNext = NULL;
1795
1796 /*
1797 * Need to clear the sync flag here if there is a new I/O context
1798 * with it set and the context is not given in pIoCtxRc.
1799 * This happens most likely on a different thread and that one shouldn't
1800 * process the context synchronously.
1801 *
1802 * The thread who issued the context will wait on the event semaphore
1803 * anyway which is signalled when the completion handler is called.
1804 */
1805 if ( pTmp->fFlags & VDIOCTX_FLAGS_SYNC
1806 && pTmp != pIoCtxRc)
1807 pTmp->fFlags &= ~VDIOCTX_FLAGS_SYNC;
1808
1809 rcTmp = vdIoCtxProcessLocked(pTmp);
1810 if (pTmp == pIoCtxRc)
1811 {
1812 if ( rcTmp == VINF_VD_ASYNC_IO_FINISHED
1813 && RT_SUCCESS(pTmp->rcReq)
1814 && pTmp->enmTxDir == VDIOCTXTXDIR_READ)
1815 {
1816 int rc2 = vdFilterChainApplyRead(pDisk, pTmp->Req.Io.uOffsetXferOrig,
1817 pTmp->Req.Io.cbXferOrig, pTmp);
1818 if (RT_FAILURE(rc2))
1819 rcTmp = rc2;
1820 }
1821
1822 /* The given I/O context was processed, pass the return code to the caller. */
1823 if ( rcTmp == VINF_VD_ASYNC_IO_FINISHED
1824 && (pTmp->fFlags & VDIOCTX_FLAGS_SYNC))
1825 rc = pTmp->rcReq;
1826 else
1827 rc = rcTmp;
1828 }
1829 else if ( rcTmp == VINF_VD_ASYNC_IO_FINISHED
1830 && ASMAtomicCmpXchgBool(&pTmp->fComplete, true, false))
1831 {
1832 LogFlowFunc(("Waiting I/O context completed pTmp=%#p\n", pTmp));
1833 vdThreadFinishWrite(pDisk);
1834 vdIoCtxRootComplete(pDisk, pTmp);
1835 vdIoCtxFree(pDisk, pTmp);
1836 }
1837 }
1838
1839 LogFlowFunc(("returns rc=%Rrc\n", rc));
1840 return rc;
1841}
1842
1843/**
1844 * Processes the list of blocked I/O contexts.
1845 *
1846 * @returns nothing.
1847 * @param pDisk The disk structure.
1848 */
1849static void vdDiskProcessBlockedIoCtx(PVBOXHDD pDisk)
1850{
1851 LogFlowFunc(("pDisk=%#p\n", pDisk));
1852
1853 VD_IS_LOCKED(pDisk);
1854
1855 /* Get the waiting list and process it in FIFO order. */
1856 PVDIOCTX pIoCtxHead = ASMAtomicXchgPtrT(&pDisk->pIoCtxBlockedHead, NULL, PVDIOCTX);
1857
1858 /* Reverse it. */
1859 PVDIOCTX pCur = pIoCtxHead;
1860 pIoCtxHead = NULL;
1861 while (pCur)
1862 {
1863 PVDIOCTX pInsert = pCur;
1864 pCur = pCur->pIoCtxNext;
1865 pInsert->pIoCtxNext = pIoCtxHead;
1866 pIoCtxHead = pInsert;
1867 }
1868
1869 /* Process now. */
1870 pCur = pIoCtxHead;
1871 while (pCur)
1872 {
1873 int rc;
1874 PVDIOCTX pTmp = pCur;
1875
1876 pCur = pCur->pIoCtxNext;
1877 pTmp->pIoCtxNext = NULL;
1878
1879 Assert(!pTmp->pIoCtxParent);
1880 Assert(pTmp->fFlags & VDIOCTX_FLAGS_BLOCKED);
1881 pTmp->fFlags &= ~VDIOCTX_FLAGS_BLOCKED;
1882
1883 rc = vdIoCtxProcessLocked(pTmp);
1884 if ( rc == VINF_VD_ASYNC_IO_FINISHED
1885 && ASMAtomicCmpXchgBool(&pTmp->fComplete, true, false))
1886 {
1887 LogFlowFunc(("Waiting I/O context completed pTmp=%#p\n", pTmp));
1888 vdThreadFinishWrite(pDisk);
1889 vdIoCtxRootComplete(pDisk, pTmp);
1890 vdIoCtxFree(pDisk, pTmp);
1891 }
1892 }
1893
1894 LogFlowFunc(("returns\n"));
1895}
1896
1897/**
1898 * Processes the I/O context trying to lock the criticial section.
1899 * The context is deferred if the critical section is busy.
1900 *
1901 * @returns VBox status code.
1902 * @param pIoCtx The I/O context to process.
1903 */
1904static int vdIoCtxProcessTryLockDefer(PVDIOCTX pIoCtx)
1905{
1906 int rc = VINF_SUCCESS;
1907 PVBOXHDD pDisk = pIoCtx->pDisk;
1908
1909 Log(("Defer pIoCtx=%#p\n", pIoCtx));
1910
1911 /* Put it on the waiting list first. */
1912 vdIoCtxAddToWaitingList(&pDisk->pIoCtxHead, pIoCtx);
1913
1914 if (ASMAtomicCmpXchgBool(&pDisk->fLocked, true, false))
1915 {
1916 /* Leave it again, the context will be processed just before leaving the lock. */
1917 LogFlowFunc(("Successfully acquired the lock\n"));
1918 rc = vdDiskUnlock(pDisk, pIoCtx);
1919 }
1920 else
1921 {
1922 LogFlowFunc(("Lock is held\n"));
1923 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1924 }
1925
1926 return rc;
1927}
1928
1929/**
1930 * Process the I/O context in a synchronous manner, waiting
1931 * for it to complete.
1932 *
1933 * @returns VBox status code of the completed request.
1934 * @param pIoCtx The sync I/O context.
1935 * @param hEventComplete Event sempahore to wait on for completion.
1936 */
1937static int vdIoCtxProcessSync(PVDIOCTX pIoCtx, RTSEMEVENT hEventComplete)
1938{
1939 int rc = VINF_SUCCESS;
1940 PVBOXHDD pDisk = pIoCtx->pDisk;
1941
1942 LogFlowFunc(("pIoCtx=%p\n", pIoCtx));
1943
1944 AssertMsg(pIoCtx->fFlags & (VDIOCTX_FLAGS_SYNC | VDIOCTX_FLAGS_DONT_FREE),
1945 ("I/O context is not marked as synchronous\n"));
1946
1947 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
1948 if (rc == VINF_VD_ASYNC_IO_FINISHED)
1949 rc = VINF_SUCCESS;
1950
1951 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
1952 {
1953 rc = RTSemEventWait(hEventComplete, RT_INDEFINITE_WAIT);
1954 AssertRC(rc);
1955 }
1956
1957 rc = pIoCtx->rcReq;
1958 vdIoCtxFree(pDisk, pIoCtx);
1959
1960 return rc;
1961}
1962
1963DECLINLINE(bool) vdIoCtxIsDiskLockOwner(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
1964{
1965 return pDisk->pIoCtxLockOwner == pIoCtx;
1966}
1967
1968static int vdIoCtxLockDisk(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
1969{
1970 int rc = VINF_SUCCESS;
1971
1972 VD_IS_LOCKED(pDisk);
1973
1974 LogFlowFunc(("pDisk=%#p pIoCtx=%#p\n", pDisk, pIoCtx));
1975
1976 if (!ASMAtomicCmpXchgPtr(&pDisk->pIoCtxLockOwner, pIoCtx, NIL_VDIOCTX))
1977 {
1978 Assert(pDisk->pIoCtxLockOwner != pIoCtx); /* No nesting allowed. */
1979 vdIoCtxDefer(pDisk, pIoCtx);
1980 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
1981 }
1982
1983 LogFlowFunc(("returns -> %Rrc\n", rc));
1984 return rc;
1985}
1986
1987static void vdIoCtxUnlockDisk(PVBOXHDD pDisk, PVDIOCTX pIoCtx, bool fProcessBlockedReqs)
1988{
1989 RT_NOREF1(pIoCtx);
1990 LogFlowFunc(("pDisk=%#p pIoCtx=%#p fProcessBlockedReqs=%RTbool\n",
1991 pDisk, pIoCtx, fProcessBlockedReqs));
1992
1993 VD_IS_LOCKED(pDisk);
1994
1995 LogFlow(("Unlocking disk lock owner is %#p\n", pDisk->pIoCtxLockOwner));
1996 Assert(pDisk->pIoCtxLockOwner == pIoCtx);
1997 ASMAtomicXchgPtrT(&pDisk->pIoCtxLockOwner, NIL_VDIOCTX, PVDIOCTX);
1998
1999 if (fProcessBlockedReqs)
2000 {
2001 /* Process any blocked writes if the current request didn't caused another growing. */
2002 vdDiskProcessBlockedIoCtx(pDisk);
2003 }
2004
2005 LogFlowFunc(("returns\n"));
2006}
2007
2008/**
2009 * Internal: Reads a given amount of data from the image chain of the disk.
2010 **/
2011static int vdDiskReadHelper(PVBOXHDD pDisk, PVDIMAGE pImage, PVDIMAGE pImageParentOverride,
2012 uint64_t uOffset, size_t cbRead, PVDIOCTX pIoCtx, size_t *pcbThisRead)
2013{
2014 RT_NOREF1(pDisk);
2015 int rc = VINF_SUCCESS;
2016 size_t cbThisRead = cbRead;
2017
2018 AssertPtr(pcbThisRead);
2019
2020 *pcbThisRead = 0;
2021
2022 /*
2023 * Try to read from the given image.
2024 * If the block is not allocated read from override chain if present.
2025 */
2026 rc = pImage->Backend->pfnRead(pImage->pBackendData,
2027 uOffset, cbThisRead, pIoCtx,
2028 &cbThisRead);
2029
2030 if (rc == VERR_VD_BLOCK_FREE)
2031 {
2032 for (PVDIMAGE pCurrImage = pImageParentOverride ? pImageParentOverride : pImage->pPrev;
2033 pCurrImage != NULL && rc == VERR_VD_BLOCK_FREE;
2034 pCurrImage = pCurrImage->pPrev)
2035 {
2036 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
2037 uOffset, cbThisRead, pIoCtx,
2038 &cbThisRead);
2039 }
2040 }
2041
2042 if (RT_SUCCESS(rc) || rc == VERR_VD_BLOCK_FREE)
2043 *pcbThisRead = cbThisRead;
2044
2045 return rc;
2046}
2047
2048/**
2049 * internal: read the specified amount of data in whatever blocks the backend
2050 * will give us - async version.
2051 */
2052static DECLCALLBACK(int) vdReadHelperAsync(PVDIOCTX pIoCtx)
2053{
2054 int rc;
2055 PVBOXHDD pDisk = pIoCtx->pDisk;
2056 size_t cbToRead = pIoCtx->Req.Io.cbTransfer;
2057 uint64_t uOffset = pIoCtx->Req.Io.uOffset;
2058 PVDIMAGE pCurrImage = pIoCtx->Req.Io.pImageCur;
2059 PVDIMAGE pImageParentOverride = pIoCtx->Req.Io.pImageParentOverride;
2060 unsigned cImagesRead = pIoCtx->Req.Io.cImagesRead;
2061 size_t cbThisRead;
2062
2063 /*
2064 * Check whether there is a full block write in progress which was not allocated.
2065 * Defer I/O if the range interferes but only if it does not belong to the
2066 * write doing the allocation.
2067 */
2068 if ( pDisk->pIoCtxLockOwner != NIL_VDIOCTX
2069 && uOffset >= pDisk->uOffsetStartLocked
2070 && uOffset < pDisk->uOffsetEndLocked
2071 && ( !pIoCtx->pIoCtxParent
2072 || pIoCtx->pIoCtxParent != pDisk->pIoCtxLockOwner))
2073 {
2074 Log(("Interferring read while allocating a new block => deferring read\n"));
2075 vdIoCtxDefer(pDisk, pIoCtx);
2076 return VERR_VD_ASYNC_IO_IN_PROGRESS;
2077 }
2078
2079 /* Loop until all reads started or we have a backend which needs to read metadata. */
2080 do
2081 {
2082 /* Search for image with allocated block. Do not attempt to read more
2083 * than the previous reads marked as valid. Otherwise this would return
2084 * stale data when different block sizes are used for the images. */
2085 cbThisRead = cbToRead;
2086
2087 if ( pDisk->pCache
2088 && !pImageParentOverride)
2089 {
2090 rc = vdCacheReadHelper(pDisk->pCache, uOffset, cbThisRead,
2091 pIoCtx, &cbThisRead);
2092 if (rc == VERR_VD_BLOCK_FREE)
2093 {
2094 rc = vdDiskReadHelper(pDisk, pCurrImage, NULL, uOffset, cbThisRead,
2095 pIoCtx, &cbThisRead);
2096
2097 /* If the read was successful, write the data back into the cache. */
2098 if ( RT_SUCCESS(rc)
2099 && pIoCtx->fFlags & VDIOCTX_FLAGS_READ_UPDATE_CACHE)
2100 {
2101 rc = vdCacheWriteHelper(pDisk->pCache, uOffset, cbThisRead,
2102 pIoCtx, NULL);
2103 }
2104 }
2105 }
2106 else
2107 {
2108 /*
2109 * Try to read from the given image.
2110 * If the block is not allocated read from override chain if present.
2111 */
2112 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
2113 uOffset, cbThisRead, pIoCtx,
2114 &cbThisRead);
2115
2116 if ( rc == VERR_VD_BLOCK_FREE
2117 && cImagesRead != 1)
2118 {
2119 unsigned cImagesToProcess = cImagesRead;
2120
2121 pCurrImage = pImageParentOverride ? pImageParentOverride : pCurrImage->pPrev;
2122 pIoCtx->Req.Io.pImageParentOverride = NULL;
2123
2124 while (pCurrImage && rc == VERR_VD_BLOCK_FREE)
2125 {
2126 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
2127 uOffset, cbThisRead,
2128 pIoCtx, &cbThisRead);
2129 if (cImagesToProcess == 1)
2130 break;
2131 else if (cImagesToProcess > 0)
2132 cImagesToProcess--;
2133
2134 if (rc == VERR_VD_BLOCK_FREE)
2135 pCurrImage = pCurrImage->pPrev;
2136 }
2137 }
2138 }
2139
2140 /* The task state will be updated on success already, don't do it here!. */
2141 if (rc == VERR_VD_BLOCK_FREE)
2142 {
2143 /* No image in the chain contains the data for the block. */
2144 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbThisRead); Assert(cbThisRead == (uint32_t)cbThisRead);
2145
2146 /* Fill the free space with 0 if we are told to do so
2147 * or a previous read returned valid data. */
2148 if (pIoCtx->fFlags & VDIOCTX_FLAGS_ZERO_FREE_BLOCKS)
2149 vdIoCtxSet(pIoCtx, '\0', cbThisRead);
2150 else
2151 pIoCtx->Req.Io.cbBufClear += cbThisRead;
2152
2153 if (pIoCtx->Req.Io.pImageCur->uOpenFlags & VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS)
2154 rc = VINF_VD_NEW_ZEROED_BLOCK;
2155 else
2156 rc = VINF_SUCCESS;
2157 }
2158 else if (rc == VERR_VD_IOCTX_HALT)
2159 {
2160 uOffset += cbThisRead;
2161 cbToRead -= cbThisRead;
2162 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
2163 }
2164 else if ( RT_SUCCESS(rc)
2165 || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
2166 {
2167 /* First not free block, fill the space before with 0. */
2168 if ( pIoCtx->Req.Io.cbBufClear
2169 && !(pIoCtx->fFlags & VDIOCTX_FLAGS_ZERO_FREE_BLOCKS))
2170 {
2171 RTSGBUF SgBuf;
2172 RTSgBufClone(&SgBuf, &pIoCtx->Req.Io.SgBuf);
2173 RTSgBufReset(&SgBuf);
2174 RTSgBufSet(&SgBuf, 0, pIoCtx->Req.Io.cbBufClear);
2175 pIoCtx->Req.Io.cbBufClear = 0;
2176 pIoCtx->fFlags |= VDIOCTX_FLAGS_ZERO_FREE_BLOCKS;
2177 }
2178 rc = VINF_SUCCESS;
2179 }
2180
2181 if (RT_FAILURE(rc))
2182 break;
2183
2184 cbToRead -= cbThisRead;
2185 uOffset += cbThisRead;
2186 pCurrImage = pIoCtx->Req.Io.pImageStart; /* Start with the highest image in the chain. */
2187 } while (cbToRead != 0 && RT_SUCCESS(rc));
2188
2189 if ( rc == VERR_VD_NOT_ENOUGH_METADATA
2190 || rc == VERR_VD_IOCTX_HALT)
2191 {
2192 /* Save the current state. */
2193 pIoCtx->Req.Io.uOffset = uOffset;
2194 pIoCtx->Req.Io.cbTransfer = cbToRead;
2195 pIoCtx->Req.Io.pImageCur = pCurrImage ? pCurrImage : pIoCtx->Req.Io.pImageStart;
2196 }
2197
2198 return (!(pIoCtx->fFlags & VDIOCTX_FLAGS_ZERO_FREE_BLOCKS))
2199 ? VERR_VD_BLOCK_FREE
2200 : rc;
2201}
2202
2203/**
2204 * internal: parent image read wrapper for compacting.
2205 */
2206static DECLCALLBACK(int) vdParentRead(void *pvUser, uint64_t uOffset, void *pvBuf,
2207 size_t cbRead)
2208{
2209 PVDPARENTSTATEDESC pParentState = (PVDPARENTSTATEDESC)pvUser;
2210
2211 /** @todo
2212 * Only used for compaction so far which is not possible to mix with async I/O.
2213 * Needs to be changed if we want to support online compaction of images.
2214 */
2215 bool fLocked = ASMAtomicXchgBool(&pParentState->pDisk->fLocked, true);
2216 AssertMsgReturn(!fLocked,
2217 ("Calling synchronous parent read while another thread holds the disk lock\n"),
2218 VERR_VD_INVALID_STATE);
2219
2220 /* Fake an I/O context. */
2221 RTSGSEG Segment;
2222 RTSGBUF SgBuf;
2223 VDIOCTX IoCtx;
2224
2225 Segment.pvSeg = pvBuf;
2226 Segment.cbSeg = cbRead;
2227 RTSgBufInit(&SgBuf, &Segment, 1);
2228 vdIoCtxInit(&IoCtx, pParentState->pDisk, VDIOCTXTXDIR_READ, uOffset, cbRead, pParentState->pImage,
2229 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC | VDIOCTX_FLAGS_ZERO_FREE_BLOCKS);
2230 int rc = vdReadHelperAsync(&IoCtx);
2231 ASMAtomicXchgBool(&pParentState->pDisk->fLocked, false);
2232 return rc;
2233}
2234
2235/**
2236 * Extended version of vdReadHelper(), implementing certain optimizations
2237 * for image cloning.
2238 *
2239 * @returns VBox status code.
2240 * @param pDisk The disk to read from.
2241 * @param pImage The image to start reading from.
2242 * @param pImageParentOverride The parent image to read from
2243 * if the starting image returns a free block.
2244 * If NULL is passed the real parent of the image
2245 * in the chain is used.
2246 * @param uOffset Offset in the disk to start reading from.
2247 * @param pvBuf Where to store the read data.
2248 * @param cbRead How much to read.
2249 * @param fZeroFreeBlocks Flag whether free blocks should be zeroed.
2250 * If false and no image has data for sepcified
2251 * range VERR_VD_BLOCK_FREE is returned.
2252 * Note that unallocated blocks are still zeroed
2253 * if at least one image has valid data for a part
2254 * of the range.
2255 * @param fUpdateCache Flag whether to update the attached cache if
2256 * available.
2257 * @param cImagesRead Number of images in the chain to read until
2258 * the read is cut off. A value of 0 disables the cut off.
2259 */
2260static int vdReadHelperEx(PVBOXHDD pDisk, PVDIMAGE pImage, PVDIMAGE pImageParentOverride,
2261 uint64_t uOffset, void *pvBuf, size_t cbRead,
2262 bool fZeroFreeBlocks, bool fUpdateCache, unsigned cImagesRead)
2263{
2264 int rc = VINF_SUCCESS;
2265 uint32_t fFlags = VDIOCTX_FLAGS_SYNC | VDIOCTX_FLAGS_DONT_FREE;
2266 RTSGSEG Segment;
2267 RTSGBUF SgBuf;
2268 VDIOCTX IoCtx;
2269 RTSEMEVENT hEventComplete = NIL_RTSEMEVENT;
2270
2271 rc = RTSemEventCreate(&hEventComplete);
2272 if (RT_FAILURE(rc))
2273 return rc;
2274
2275 if (fZeroFreeBlocks)
2276 fFlags |= VDIOCTX_FLAGS_ZERO_FREE_BLOCKS;
2277 if (fUpdateCache)
2278 fFlags |= VDIOCTX_FLAGS_READ_UPDATE_CACHE;
2279
2280 Segment.pvSeg = pvBuf;
2281 Segment.cbSeg = cbRead;
2282 RTSgBufInit(&SgBuf, &Segment, 1);
2283 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_READ, uOffset, cbRead, pImage, &SgBuf,
2284 NULL, vdReadHelperAsync, fFlags);
2285
2286 IoCtx.Req.Io.pImageParentOverride = pImageParentOverride;
2287 IoCtx.Req.Io.cImagesRead = cImagesRead;
2288 IoCtx.Type.Root.pfnComplete = vdIoCtxSyncComplete;
2289 IoCtx.Type.Root.pvUser1 = pDisk;
2290 IoCtx.Type.Root.pvUser2 = hEventComplete;
2291 rc = vdIoCtxProcessSync(&IoCtx, hEventComplete);
2292 RTSemEventDestroy(hEventComplete);
2293 return rc;
2294}
2295
2296/**
2297 * internal: read the specified amount of data in whatever blocks the backend
2298 * will give us.
2299 */
2300static int vdReadHelper(PVBOXHDD pDisk, PVDIMAGE pImage, uint64_t uOffset,
2301 void *pvBuf, size_t cbRead, bool fUpdateCache)
2302{
2303 return vdReadHelperEx(pDisk, pImage, NULL, uOffset, pvBuf, cbRead,
2304 true /* fZeroFreeBlocks */, fUpdateCache, 0);
2305}
2306
2307/**
2308 * internal: mark the disk as not modified.
2309 */
2310static void vdResetModifiedFlag(PVBOXHDD pDisk)
2311{
2312 if (pDisk->uModified & VD_IMAGE_MODIFIED_FLAG)
2313 {
2314 /* generate new last-modified uuid */
2315 if (!(pDisk->uModified & VD_IMAGE_MODIFIED_DISABLE_UUID_UPDATE))
2316 {
2317 RTUUID Uuid;
2318
2319 RTUuidCreate(&Uuid);
2320 pDisk->pLast->Backend->pfnSetModificationUuid(pDisk->pLast->pBackendData,
2321 &Uuid);
2322
2323 if (pDisk->pCache)
2324 pDisk->pCache->Backend->pfnSetModificationUuid(pDisk->pCache->pBackendData,
2325 &Uuid);
2326 }
2327
2328 pDisk->uModified &= ~VD_IMAGE_MODIFIED_FLAG;
2329 }
2330}
2331
2332/**
2333 * internal: mark the disk as modified.
2334 */
2335static void vdSetModifiedFlag(PVBOXHDD pDisk)
2336{
2337 pDisk->uModified |= VD_IMAGE_MODIFIED_FLAG;
2338 if (pDisk->uModified & VD_IMAGE_MODIFIED_FIRST)
2339 {
2340 pDisk->uModified &= ~VD_IMAGE_MODIFIED_FIRST;
2341
2342 /* First modify, so create a UUID and ensure it's written to disk. */
2343 vdResetModifiedFlag(pDisk);
2344
2345 if (!(pDisk->uModified & VD_IMAGE_MODIFIED_DISABLE_UUID_UPDATE))
2346 {
2347 VDIOCTX IoCtx;
2348 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_FLUSH, 0, 0, NULL,
2349 NULL, NULL, NULL, VDIOCTX_FLAGS_SYNC);
2350 pDisk->pLast->Backend->pfnFlush(pDisk->pLast->pBackendData, &IoCtx);
2351 }
2352 }
2353}
2354
2355/**
2356 * internal: write buffer to the image, taking care of block boundaries and
2357 * write optimizations.
2358 */
2359static int vdWriteHelperEx(PVBOXHDD pDisk, PVDIMAGE pImage,
2360 PVDIMAGE pImageParentOverride, uint64_t uOffset,
2361 const void *pvBuf, size_t cbWrite,
2362 uint32_t fFlags, unsigned cImagesRead)
2363{
2364 int rc = VINF_SUCCESS;
2365 RTSGSEG Segment;
2366 RTSGBUF SgBuf;
2367 VDIOCTX IoCtx;
2368 RTSEMEVENT hEventComplete = NIL_RTSEMEVENT;
2369
2370 rc = RTSemEventCreate(&hEventComplete);
2371 if (RT_FAILURE(rc))
2372 return rc;
2373
2374 fFlags |= VDIOCTX_FLAGS_SYNC | VDIOCTX_FLAGS_DONT_FREE;
2375
2376 Segment.pvSeg = (void *)pvBuf;
2377 Segment.cbSeg = cbWrite;
2378 RTSgBufInit(&SgBuf, &Segment, 1);
2379 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_WRITE, uOffset, cbWrite, pImage, &SgBuf,
2380 NULL, vdWriteHelperAsync, fFlags);
2381
2382 IoCtx.Req.Io.pImageParentOverride = pImageParentOverride;
2383 IoCtx.Req.Io.cImagesRead = cImagesRead;
2384 IoCtx.pIoCtxParent = NULL;
2385 IoCtx.Type.Root.pfnComplete = vdIoCtxSyncComplete;
2386 IoCtx.Type.Root.pvUser1 = pDisk;
2387 IoCtx.Type.Root.pvUser2 = hEventComplete;
2388 if (RT_SUCCESS(rc))
2389 rc = vdIoCtxProcessSync(&IoCtx, hEventComplete);
2390
2391 RTSemEventDestroy(hEventComplete);
2392 return rc;
2393}
2394
2395/**
2396 * internal: write buffer to the image, taking care of block boundaries and
2397 * write optimizations.
2398 */
2399static int vdWriteHelper(PVBOXHDD pDisk, PVDIMAGE pImage, uint64_t uOffset,
2400 const void *pvBuf, size_t cbWrite, uint32_t fFlags)
2401{
2402 return vdWriteHelperEx(pDisk, pImage, NULL, uOffset, pvBuf, cbWrite,
2403 fFlags, 0);
2404}
2405
2406/**
2407 * Internal: Copies the content of one disk to another one applying optimizations
2408 * to speed up the copy process if possible.
2409 */
2410static int vdCopyHelper(PVBOXHDD pDiskFrom, PVDIMAGE pImageFrom, PVBOXHDD pDiskTo,
2411 uint64_t cbSize, unsigned cImagesFromRead, unsigned cImagesToRead,
2412 bool fSuppressRedundantIo, PVDINTERFACEPROGRESS pIfProgress,
2413 PVDINTERFACEPROGRESS pDstIfProgress)
2414{
2415 int rc = VINF_SUCCESS;
2416 int rc2;
2417 uint64_t uOffset = 0;
2418 uint64_t cbRemaining = cbSize;
2419 void *pvBuf = NULL;
2420 bool fLockReadFrom = false;
2421 bool fLockWriteTo = false;
2422 bool fBlockwiseCopy = false;
2423 unsigned uProgressOld = 0;
2424
2425 LogFlowFunc(("pDiskFrom=%#p pImageFrom=%#p pDiskTo=%#p cbSize=%llu cImagesFromRead=%u cImagesToRead=%u fSuppressRedundantIo=%RTbool pIfProgress=%#p pDstIfProgress=%#p\n",
2426 pDiskFrom, pImageFrom, pDiskTo, cbSize, cImagesFromRead, cImagesToRead, fSuppressRedundantIo, pDstIfProgress, pDstIfProgress));
2427
2428 if ( (fSuppressRedundantIo || (cImagesFromRead > 0))
2429 && RTListIsEmpty(&pDiskFrom->ListFilterChainRead))
2430 fBlockwiseCopy = true;
2431
2432 /* Allocate tmp buffer. */
2433 pvBuf = RTMemTmpAlloc(VD_MERGE_BUFFER_SIZE);
2434 if (!pvBuf)
2435 return rc;
2436
2437 do
2438 {
2439 size_t cbThisRead = RT_MIN(VD_MERGE_BUFFER_SIZE, cbRemaining);
2440
2441 /* Note that we don't attempt to synchronize cross-disk accesses.
2442 * It wouldn't be very difficult to do, just the lock order would
2443 * need to be defined somehow to prevent deadlocks. Postpone such
2444 * magic as there is no use case for this. */
2445
2446 rc2 = vdThreadStartRead(pDiskFrom);
2447 AssertRC(rc2);
2448 fLockReadFrom = true;
2449
2450 if (fBlockwiseCopy)
2451 {
2452 RTSGSEG SegmentBuf;
2453 RTSGBUF SgBuf;
2454 VDIOCTX IoCtx;
2455
2456 SegmentBuf.pvSeg = pvBuf;
2457 SegmentBuf.cbSeg = VD_MERGE_BUFFER_SIZE;
2458 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
2459 vdIoCtxInit(&IoCtx, pDiskFrom, VDIOCTXTXDIR_READ, 0, 0, NULL,
2460 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
2461
2462 /* Read the source data. */
2463 rc = pImageFrom->Backend->pfnRead(pImageFrom->pBackendData,
2464 uOffset, cbThisRead, &IoCtx,
2465 &cbThisRead);
2466
2467 if ( rc == VERR_VD_BLOCK_FREE
2468 && cImagesFromRead != 1)
2469 {
2470 unsigned cImagesToProcess = cImagesFromRead;
2471
2472 for (PVDIMAGE pCurrImage = pImageFrom->pPrev;
2473 pCurrImage != NULL && rc == VERR_VD_BLOCK_FREE;
2474 pCurrImage = pCurrImage->pPrev)
2475 {
2476 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
2477 uOffset, cbThisRead,
2478 &IoCtx, &cbThisRead);
2479 if (cImagesToProcess == 1)
2480 break;
2481 else if (cImagesToProcess > 0)
2482 cImagesToProcess--;
2483 }
2484 }
2485 }
2486 else
2487 rc = vdReadHelper(pDiskFrom, pImageFrom, uOffset, pvBuf, cbThisRead,
2488 false /* fUpdateCache */);
2489
2490 if (RT_FAILURE(rc) && rc != VERR_VD_BLOCK_FREE)
2491 break;
2492
2493 rc2 = vdThreadFinishRead(pDiskFrom);
2494 AssertRC(rc2);
2495 fLockReadFrom = false;
2496
2497 if (rc != VERR_VD_BLOCK_FREE)
2498 {
2499 rc2 = vdThreadStartWrite(pDiskTo);
2500 AssertRC(rc2);
2501 fLockWriteTo = true;
2502
2503 /* Only do collapsed I/O if we are copying the data blockwise. */
2504 rc = vdWriteHelperEx(pDiskTo, pDiskTo->pLast, NULL, uOffset, pvBuf,
2505 cbThisRead, VDIOCTX_FLAGS_DONT_SET_MODIFIED_FLAG /* fFlags */,
2506 fBlockwiseCopy ? cImagesToRead : 0);
2507 if (RT_FAILURE(rc))
2508 break;
2509
2510 rc2 = vdThreadFinishWrite(pDiskTo);
2511 AssertRC(rc2);
2512 fLockWriteTo = false;
2513 }
2514 else /* Don't propagate the error to the outside */
2515 rc = VINF_SUCCESS;
2516
2517 uOffset += cbThisRead;
2518 cbRemaining -= cbThisRead;
2519
2520 unsigned uProgressNew = uOffset * 99 / cbSize;
2521 if (uProgressNew != uProgressOld)
2522 {
2523 uProgressOld = uProgressNew;
2524
2525 if (pIfProgress && pIfProgress->pfnProgress)
2526 {
2527 rc = pIfProgress->pfnProgress(pIfProgress->Core.pvUser,
2528 uProgressOld);
2529 if (RT_FAILURE(rc))
2530 break;
2531 }
2532 if (pDstIfProgress && pDstIfProgress->pfnProgress)
2533 {
2534 rc = pDstIfProgress->pfnProgress(pDstIfProgress->Core.pvUser,
2535 uProgressOld);
2536 if (RT_FAILURE(rc))
2537 break;
2538 }
2539 }
2540 } while (uOffset < cbSize);
2541
2542 RTMemFree(pvBuf);
2543
2544 if (fLockReadFrom)
2545 {
2546 rc2 = vdThreadFinishRead(pDiskFrom);
2547 AssertRC(rc2);
2548 }
2549
2550 if (fLockWriteTo)
2551 {
2552 rc2 = vdThreadFinishWrite(pDiskTo);
2553 AssertRC(rc2);
2554 }
2555
2556 LogFlowFunc(("returns rc=%Rrc\n", rc));
2557 return rc;
2558}
2559
2560/**
2561 * Flush helper async version.
2562 */
2563static DECLCALLBACK(int) vdSetModifiedHelperAsync(PVDIOCTX pIoCtx)
2564{
2565 int rc = VINF_SUCCESS;
2566 PVDIMAGE pImage = pIoCtx->Req.Io.pImageCur;
2567
2568 rc = pImage->Backend->pfnFlush(pImage->pBackendData, pIoCtx);
2569 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
2570 rc = VINF_SUCCESS;
2571
2572 return rc;
2573}
2574
2575/**
2576 * internal: mark the disk as modified - async version.
2577 */
2578static int vdSetModifiedFlagAsync(PVBOXHDD pDisk, PVDIOCTX pIoCtx)
2579{
2580 int rc = VINF_SUCCESS;
2581
2582 VD_IS_LOCKED(pDisk);
2583
2584 pDisk->uModified |= VD_IMAGE_MODIFIED_FLAG;
2585 if (pDisk->uModified & VD_IMAGE_MODIFIED_FIRST)
2586 {
2587 rc = vdIoCtxLockDisk(pDisk, pIoCtx);
2588 if (RT_SUCCESS(rc))
2589 {
2590 pDisk->uModified &= ~VD_IMAGE_MODIFIED_FIRST;
2591
2592 /* First modify, so create a UUID and ensure it's written to disk. */
2593 vdResetModifiedFlag(pDisk);
2594
2595 if (!(pDisk->uModified & VD_IMAGE_MODIFIED_DISABLE_UUID_UPDATE))
2596 {
2597 PVDIOCTX pIoCtxFlush = vdIoCtxChildAlloc(pDisk, VDIOCTXTXDIR_FLUSH,
2598 0, 0, pDisk->pLast,
2599 NULL, pIoCtx, 0, 0, NULL,
2600 vdSetModifiedHelperAsync);
2601
2602 if (pIoCtxFlush)
2603 {
2604 rc = vdIoCtxProcessLocked(pIoCtxFlush);
2605 if (rc == VINF_VD_ASYNC_IO_FINISHED)
2606 {
2607 vdIoCtxUnlockDisk(pDisk, pIoCtx, false /* fProcessDeferredReqs */);
2608 vdIoCtxFree(pDisk, pIoCtxFlush);
2609 }
2610 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
2611 {
2612 ASMAtomicIncU32(&pIoCtx->cDataTransfersPending);
2613 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
2614 }
2615 else /* Another error */
2616 vdIoCtxFree(pDisk, pIoCtxFlush);
2617 }
2618 else
2619 rc = VERR_NO_MEMORY;
2620 }
2621 }
2622 }
2623
2624 return rc;
2625}
2626
2627static DECLCALLBACK(int) vdWriteHelperCommitAsync(PVDIOCTX pIoCtx)
2628{
2629 int rc = VINF_SUCCESS;
2630 PVDIMAGE pImage = pIoCtx->Req.Io.pImageStart;
2631 size_t cbPreRead = pIoCtx->Type.Child.cbPreRead;
2632 size_t cbPostRead = pIoCtx->Type.Child.cbPostRead;
2633 size_t cbThisWrite = pIoCtx->Type.Child.cbTransferParent;
2634
2635 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2636 rc = pImage->Backend->pfnWrite(pImage->pBackendData,
2637 pIoCtx->Req.Io.uOffset - cbPreRead,
2638 cbPreRead + cbThisWrite + cbPostRead,
2639 pIoCtx, NULL, &cbPreRead, &cbPostRead, 0);
2640 Assert(rc != VERR_VD_BLOCK_FREE);
2641 Assert(rc == VERR_VD_NOT_ENOUGH_METADATA || cbPreRead == 0);
2642 Assert(rc == VERR_VD_NOT_ENOUGH_METADATA || cbPostRead == 0);
2643 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
2644 rc = VINF_SUCCESS;
2645 else if (rc == VERR_VD_IOCTX_HALT)
2646 {
2647 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
2648 rc = VINF_SUCCESS;
2649 }
2650
2651 LogFlowFunc(("returns rc=%Rrc\n", rc));
2652 return rc;
2653}
2654
2655static DECLCALLBACK(int) vdWriteHelperOptimizedCmpAndWriteAsync(PVDIOCTX pIoCtx)
2656{
2657 int rc = VINF_SUCCESS;
2658 size_t cbThisWrite = 0;
2659 size_t cbPreRead = pIoCtx->Type.Child.cbPreRead;
2660 size_t cbPostRead = pIoCtx->Type.Child.cbPostRead;
2661 size_t cbWriteCopy = pIoCtx->Type.Child.Write.Optimized.cbWriteCopy;
2662 size_t cbFill = pIoCtx->Type.Child.Write.Optimized.cbFill;
2663 size_t cbReadImage = pIoCtx->Type.Child.Write.Optimized.cbReadImage;
2664 PVDIOCTX pIoCtxParent = pIoCtx->pIoCtxParent;
2665
2666 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2667
2668 AssertPtr(pIoCtxParent);
2669 Assert(!pIoCtxParent->pIoCtxParent);
2670 Assert(!pIoCtx->Req.Io.cbTransferLeft && !pIoCtx->cMetaTransfersPending);
2671
2672 vdIoCtxChildReset(pIoCtx);
2673 cbThisWrite = pIoCtx->Type.Child.cbTransferParent;
2674 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbPreRead);
2675
2676 /* Check if the write would modify anything in this block. */
2677 if (!RTSgBufCmp(&pIoCtx->Req.Io.SgBuf, &pIoCtxParent->Req.Io.SgBuf, cbThisWrite))
2678 {
2679 RTSGBUF SgBufSrcTmp;
2680
2681 RTSgBufClone(&SgBufSrcTmp, &pIoCtxParent->Req.Io.SgBuf);
2682 RTSgBufAdvance(&SgBufSrcTmp, cbThisWrite);
2683 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbThisWrite);
2684
2685 if (!cbWriteCopy || !RTSgBufCmp(&pIoCtx->Req.Io.SgBuf, &SgBufSrcTmp, cbWriteCopy))
2686 {
2687 /* Block is completely unchanged, so no need to write anything. */
2688 LogFlowFunc(("Block didn't changed\n"));
2689 ASMAtomicWriteU32(&pIoCtx->Req.Io.cbTransferLeft, 0);
2690 RTSgBufAdvance(&pIoCtxParent->Req.Io.SgBuf, cbThisWrite);
2691 return VINF_VD_ASYNC_IO_FINISHED;
2692 }
2693 }
2694
2695 /* Copy the data to the right place in the buffer. */
2696 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
2697 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbPreRead);
2698 vdIoCtxCopy(pIoCtx, pIoCtxParent, cbThisWrite);
2699
2700 /* Handle the data that goes after the write to fill the block. */
2701 if (cbPostRead)
2702 {
2703 /* Now assemble the remaining data. */
2704 if (cbWriteCopy)
2705 {
2706 /*
2707 * The S/G buffer of the parent needs to be cloned because
2708 * it is not allowed to modify the state.
2709 */
2710 RTSGBUF SgBufParentTmp;
2711
2712 RTSgBufClone(&SgBufParentTmp, &pIoCtxParent->Req.Io.SgBuf);
2713 RTSgBufCopy(&pIoCtx->Req.Io.SgBuf, &SgBufParentTmp, cbWriteCopy);
2714 }
2715
2716 /* Zero out the remainder of this block. Will never be visible, as this
2717 * is beyond the limit of the image. */
2718 if (cbFill)
2719 {
2720 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbReadImage);
2721 vdIoCtxSet(pIoCtx, '\0', cbFill);
2722 }
2723 }
2724
2725 /* Write the full block to the virtual disk. */
2726 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
2727 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperCommitAsync;
2728
2729 return rc;
2730}
2731
2732static DECLCALLBACK(int) vdWriteHelperOptimizedPreReadAsync(PVDIOCTX pIoCtx)
2733{
2734 int rc = VINF_SUCCESS;
2735
2736 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2737
2738 pIoCtx->fFlags |= VDIOCTX_FLAGS_ZERO_FREE_BLOCKS;
2739
2740 if ( pIoCtx->Req.Io.cbTransferLeft
2741 && !pIoCtx->cDataTransfersPending)
2742 rc = vdReadHelperAsync(pIoCtx);
2743
2744 if ( ( RT_SUCCESS(rc)
2745 || (rc == VERR_VD_ASYNC_IO_IN_PROGRESS))
2746 && ( pIoCtx->Req.Io.cbTransferLeft
2747 || pIoCtx->cMetaTransfersPending))
2748 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
2749 else
2750 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperOptimizedCmpAndWriteAsync;
2751
2752 return rc;
2753}
2754
2755/**
2756 * internal: write a complete block (only used for diff images), taking the
2757 * remaining data from parent images. This implementation optimizes out writes
2758 * that do not change the data relative to the state as of the parent images.
2759 * All backends which support differential/growing images support this - async version.
2760 */
2761static DECLCALLBACK(int) vdWriteHelperOptimizedAsync(PVDIOCTX pIoCtx)
2762{
2763 PVBOXHDD pDisk = pIoCtx->pDisk;
2764 uint64_t uOffset = pIoCtx->Type.Child.uOffsetSaved;
2765 size_t cbThisWrite = pIoCtx->Type.Child.cbTransferParent;
2766 size_t cbPreRead = pIoCtx->Type.Child.cbPreRead;
2767 size_t cbPostRead = pIoCtx->Type.Child.cbPostRead;
2768 size_t cbWrite = pIoCtx->Type.Child.cbWriteParent;
2769 size_t cbFill = 0;
2770 size_t cbWriteCopy = 0;
2771 size_t cbReadImage = 0;
2772
2773 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2774
2775 AssertPtr(pIoCtx->pIoCtxParent);
2776 Assert(!pIoCtx->pIoCtxParent->pIoCtxParent);
2777
2778 if (cbPostRead)
2779 {
2780 /* Figure out how much we cannot read from the image, because
2781 * the last block to write might exceed the nominal size of the
2782 * image for technical reasons. */
2783 if (uOffset + cbThisWrite + cbPostRead > pDisk->cbSize)
2784 cbFill = uOffset + cbThisWrite + cbPostRead - pDisk->cbSize;
2785
2786 /* If we have data to be written, use that instead of reading
2787 * data from the image. */
2788 if (cbWrite > cbThisWrite)
2789 cbWriteCopy = RT_MIN(cbWrite - cbThisWrite, cbPostRead);
2790
2791 /* The rest must be read from the image. */
2792 cbReadImage = cbPostRead - cbWriteCopy - cbFill;
2793 }
2794
2795 pIoCtx->Type.Child.Write.Optimized.cbFill = cbFill;
2796 pIoCtx->Type.Child.Write.Optimized.cbWriteCopy = cbWriteCopy;
2797 pIoCtx->Type.Child.Write.Optimized.cbReadImage = cbReadImage;
2798
2799 /* Read the entire data of the block so that we can compare whether it will
2800 * be modified by the write or not. */
2801 size_t cbTmp = cbPreRead + cbThisWrite + cbPostRead - cbFill; Assert(cbTmp == (uint32_t)cbTmp);
2802 pIoCtx->Req.Io.cbTransferLeft = (uint32_t)cbTmp;
2803 pIoCtx->Req.Io.cbTransfer = pIoCtx->Req.Io.cbTransferLeft;
2804 pIoCtx->Req.Io.uOffset -= cbPreRead;
2805
2806 /* Next step */
2807 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperOptimizedPreReadAsync;
2808 return VINF_SUCCESS;
2809}
2810
2811static DECLCALLBACK(int) vdWriteHelperStandardReadImageAsync(PVDIOCTX pIoCtx)
2812{
2813 int rc = VINF_SUCCESS;
2814
2815 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2816
2817 pIoCtx->fFlags |= VDIOCTX_FLAGS_ZERO_FREE_BLOCKS;
2818
2819 if ( pIoCtx->Req.Io.cbTransferLeft
2820 && !pIoCtx->cDataTransfersPending)
2821 rc = vdReadHelperAsync(pIoCtx);
2822
2823 if ( RT_SUCCESS(rc)
2824 && ( pIoCtx->Req.Io.cbTransferLeft
2825 || pIoCtx->cMetaTransfersPending))
2826 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
2827 else
2828 {
2829 size_t cbFill = pIoCtx->Type.Child.Write.Optimized.cbFill;
2830
2831 /* Zero out the remainder of this block. Will never be visible, as this
2832 * is beyond the limit of the image. */
2833 if (cbFill)
2834 vdIoCtxSet(pIoCtx, '\0', cbFill);
2835
2836 /* Write the full block to the virtual disk. */
2837 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
2838
2839 vdIoCtxChildReset(pIoCtx);
2840 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperCommitAsync;
2841 }
2842
2843 return rc;
2844}
2845
2846static DECLCALLBACK(int) vdWriteHelperStandardAssemble(PVDIOCTX pIoCtx)
2847{
2848 int rc = VINF_SUCCESS;
2849 size_t cbPostRead = pIoCtx->Type.Child.cbPostRead;
2850 size_t cbThisWrite = pIoCtx->Type.Child.cbTransferParent;
2851 PVDIOCTX pIoCtxParent = pIoCtx->pIoCtxParent;
2852
2853 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2854
2855 vdIoCtxCopy(pIoCtx, pIoCtxParent, cbThisWrite);
2856 if (cbPostRead)
2857 {
2858 size_t cbFill = pIoCtx->Type.Child.Write.Optimized.cbFill;
2859 size_t cbWriteCopy = pIoCtx->Type.Child.Write.Optimized.cbWriteCopy;
2860 size_t cbReadImage = pIoCtx->Type.Child.Write.Optimized.cbReadImage;
2861
2862 /* Now assemble the remaining data. */
2863 if (cbWriteCopy)
2864 {
2865 /*
2866 * The S/G buffer of the parent needs to be cloned because
2867 * it is not allowed to modify the state.
2868 */
2869 RTSGBUF SgBufParentTmp;
2870
2871 RTSgBufClone(&SgBufParentTmp, &pIoCtxParent->Req.Io.SgBuf);
2872 RTSgBufCopy(&pIoCtx->Req.Io.SgBuf, &SgBufParentTmp, cbWriteCopy);
2873 }
2874
2875 if (cbReadImage)
2876 {
2877 /* Read remaining data. */
2878 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperStandardReadImageAsync;
2879
2880 /* Read the data that goes before the write to fill the block. */
2881 pIoCtx->Req.Io.cbTransferLeft = (uint32_t)cbReadImage; Assert(cbReadImage == (uint32_t)cbReadImage);
2882 pIoCtx->Req.Io.cbTransfer = pIoCtx->Req.Io.cbTransferLeft;
2883 pIoCtx->Req.Io.uOffset += cbWriteCopy;
2884 }
2885 else
2886 {
2887 /* Zero out the remainder of this block. Will never be visible, as this
2888 * is beyond the limit of the image. */
2889 if (cbFill)
2890 vdIoCtxSet(pIoCtx, '\0', cbFill);
2891
2892 /* Write the full block to the virtual disk. */
2893 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
2894 vdIoCtxChildReset(pIoCtx);
2895 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperCommitAsync;
2896 }
2897 }
2898 else
2899 {
2900 /* Write the full block to the virtual disk. */
2901 RTSgBufReset(&pIoCtx->Req.Io.SgBuf);
2902 vdIoCtxChildReset(pIoCtx);
2903 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperCommitAsync;
2904 }
2905
2906 return rc;
2907}
2908
2909static DECLCALLBACK(int) vdWriteHelperStandardPreReadAsync(PVDIOCTX pIoCtx)
2910{
2911 int rc = VINF_SUCCESS;
2912
2913 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2914
2915 pIoCtx->fFlags |= VDIOCTX_FLAGS_ZERO_FREE_BLOCKS;
2916
2917 if ( pIoCtx->Req.Io.cbTransferLeft
2918 && !pIoCtx->cDataTransfersPending)
2919 rc = vdReadHelperAsync(pIoCtx);
2920
2921 if ( RT_SUCCESS(rc)
2922 && ( pIoCtx->Req.Io.cbTransferLeft
2923 || pIoCtx->cMetaTransfersPending))
2924 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
2925 else
2926 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperStandardAssemble;
2927
2928 return rc;
2929}
2930
2931static DECLCALLBACK(int) vdWriteHelperStandardAsync(PVDIOCTX pIoCtx)
2932{
2933 PVBOXHDD pDisk = pIoCtx->pDisk;
2934 uint64_t uOffset = pIoCtx->Type.Child.uOffsetSaved;
2935 size_t cbThisWrite = pIoCtx->Type.Child.cbTransferParent;
2936 size_t cbPreRead = pIoCtx->Type.Child.cbPreRead;
2937 size_t cbPostRead = pIoCtx->Type.Child.cbPostRead;
2938 size_t cbWrite = pIoCtx->Type.Child.cbWriteParent;
2939 size_t cbFill = 0;
2940 size_t cbWriteCopy = 0;
2941 size_t cbReadImage = 0;
2942
2943 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
2944
2945 AssertPtr(pIoCtx->pIoCtxParent);
2946 Assert(!pIoCtx->pIoCtxParent->pIoCtxParent);
2947
2948 /* Calculate the amount of data to read that goes after the write to fill the block. */
2949 if (cbPostRead)
2950 {
2951 /* If we have data to be written, use that instead of reading
2952 * data from the image. */
2953 if (cbWrite > cbThisWrite)
2954 cbWriteCopy = RT_MIN(cbWrite - cbThisWrite, cbPostRead);
2955 else
2956 cbWriteCopy = 0;
2957
2958 /* Figure out how much we cannot read from the image, because
2959 * the last block to write might exceed the nominal size of the
2960 * image for technical reasons. */
2961 if (uOffset + cbThisWrite + cbPostRead > pDisk->cbSize)
2962 cbFill = uOffset + cbThisWrite + cbPostRead - pDisk->cbSize;
2963
2964 /* The rest must be read from the image. */
2965 cbReadImage = cbPostRead - cbWriteCopy - cbFill;
2966 }
2967
2968 pIoCtx->Type.Child.Write.Optimized.cbFill = cbFill;
2969 pIoCtx->Type.Child.Write.Optimized.cbWriteCopy = cbWriteCopy;
2970 pIoCtx->Type.Child.Write.Optimized.cbReadImage = cbReadImage;
2971
2972 /* Next step */
2973 if (cbPreRead)
2974 {
2975 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperStandardPreReadAsync;
2976
2977 /* Read the data that goes before the write to fill the block. */
2978 pIoCtx->Req.Io.cbTransferLeft = (uint32_t)cbPreRead; Assert(cbPreRead == (uint32_t)cbPreRead);
2979 pIoCtx->Req.Io.cbTransfer = pIoCtx->Req.Io.cbTransferLeft;
2980 pIoCtx->Req.Io.uOffset -= cbPreRead;
2981 }
2982 else
2983 pIoCtx->pfnIoCtxTransferNext = vdWriteHelperStandardAssemble;
2984
2985 return VINF_SUCCESS;
2986}
2987
2988/**
2989 * internal: write buffer to the image, taking care of block boundaries and
2990 * write optimizations - async version.
2991 */
2992static DECLCALLBACK(int) vdWriteHelperAsync(PVDIOCTX pIoCtx)
2993{
2994 int rc;
2995 size_t cbWrite = pIoCtx->Req.Io.cbTransfer;
2996 uint64_t uOffset = pIoCtx->Req.Io.uOffset;
2997 PVDIMAGE pImage = pIoCtx->Req.Io.pImageCur;
2998 PVBOXHDD pDisk = pIoCtx->pDisk;
2999 unsigned fWrite;
3000 size_t cbThisWrite;
3001 size_t cbPreRead, cbPostRead;
3002
3003 /* Apply write filter chain here if it was not done already. */
3004 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_WRITE_FILTER_APPLIED))
3005 {
3006 rc = vdFilterChainApplyWrite(pDisk, uOffset, cbWrite, pIoCtx);
3007 if (RT_FAILURE(rc))
3008 return rc;
3009 pIoCtx->fFlags |= VDIOCTX_FLAGS_WRITE_FILTER_APPLIED;
3010 }
3011
3012 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_DONT_SET_MODIFIED_FLAG))
3013 {
3014 rc = vdSetModifiedFlagAsync(pDisk, pIoCtx);
3015 if (RT_FAILURE(rc)) /* Includes I/O in progress. */
3016 return rc;
3017 }
3018
3019 rc = vdDiscardSetRangeAllocated(pDisk, uOffset, cbWrite);
3020 if (RT_FAILURE(rc))
3021 return rc;
3022
3023 /* Loop until all written. */
3024 do
3025 {
3026 /* Try to write the possibly partial block to the last opened image.
3027 * This works when the block is already allocated in this image or
3028 * if it is a full-block write (and allocation isn't suppressed below).
3029 * For image formats which don't support zero blocks, it's beneficial
3030 * to avoid unnecessarily allocating unchanged blocks. This prevents
3031 * unwanted expanding of images. VMDK is an example. */
3032 cbThisWrite = cbWrite;
3033
3034 /*
3035 * Check whether there is a full block write in progress which was not allocated.
3036 * Defer I/O if the range interferes.
3037 */
3038 if ( pDisk->pIoCtxLockOwner != NIL_VDIOCTX
3039 && uOffset >= pDisk->uOffsetStartLocked
3040 && uOffset < pDisk->uOffsetEndLocked)
3041 {
3042 Log(("Interferring write while allocating a new block => deferring write\n"));
3043 vdIoCtxDefer(pDisk, pIoCtx);
3044 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
3045 break;
3046 }
3047
3048 fWrite = (pImage->uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME)
3049 ? 0 : VD_WRITE_NO_ALLOC;
3050 rc = pImage->Backend->pfnWrite(pImage->pBackendData, uOffset, cbThisWrite,
3051 pIoCtx, &cbThisWrite, &cbPreRead, &cbPostRead,
3052 fWrite);
3053 if (rc == VERR_VD_BLOCK_FREE)
3054 {
3055 /* Lock the disk .*/
3056 rc = vdIoCtxLockDisk(pDisk, pIoCtx);
3057 if (RT_SUCCESS(rc))
3058 {
3059 /*
3060 * Allocate segment and buffer in one go.
3061 * A bit hackish but avoids the need to allocate memory twice.
3062 */
3063 PRTSGBUF pTmp = (PRTSGBUF)RTMemAlloc(cbPreRead + cbThisWrite + cbPostRead + sizeof(RTSGSEG) + sizeof(RTSGBUF));
3064 AssertBreakStmt(pTmp, rc = VERR_NO_MEMORY);
3065 PRTSGSEG pSeg = (PRTSGSEG)(pTmp + 1);
3066
3067 pSeg->pvSeg = pSeg + 1;
3068 pSeg->cbSeg = cbPreRead + cbThisWrite + cbPostRead;
3069 RTSgBufInit(pTmp, pSeg, 1);
3070
3071 PVDIOCTX pIoCtxWrite = vdIoCtxChildAlloc(pDisk, VDIOCTXTXDIR_WRITE,
3072 uOffset, pSeg->cbSeg, pImage,
3073 pTmp,
3074 pIoCtx, cbThisWrite,
3075 cbWrite,
3076 pTmp,
3077 (pImage->uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME)
3078 ? vdWriteHelperStandardAsync
3079 : vdWriteHelperOptimizedAsync);
3080 if (!VALID_PTR(pIoCtxWrite))
3081 {
3082 RTMemTmpFree(pTmp);
3083 rc = VERR_NO_MEMORY;
3084 break;
3085 }
3086
3087 LogFlowFunc(("Disk is growing because of pIoCtx=%#p pIoCtxWrite=%#p\n",
3088 pIoCtx, pIoCtxWrite));
3089
3090 /* Save the current range for the growing operation to check for intersecting requests later. */
3091 pDisk->uOffsetStartLocked = uOffset - cbPreRead;
3092 pDisk->uOffsetEndLocked = uOffset + cbThisWrite + cbPostRead;
3093
3094 pIoCtxWrite->Type.Child.cbPreRead = cbPreRead;
3095 pIoCtxWrite->Type.Child.cbPostRead = cbPostRead;
3096 pIoCtxWrite->Req.Io.pImageParentOverride = pIoCtx->Req.Io.pImageParentOverride;
3097
3098 /* Process the write request */
3099 rc = vdIoCtxProcessLocked(pIoCtxWrite);
3100
3101 if (RT_FAILURE(rc) && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
3102 {
3103 vdIoCtxUnlockDisk(pDisk, pIoCtx, false /* fProcessDeferredReqs*/ );
3104 vdIoCtxFree(pDisk, pIoCtxWrite);
3105 break;
3106 }
3107 else if ( rc == VINF_VD_ASYNC_IO_FINISHED
3108 && ASMAtomicCmpXchgBool(&pIoCtxWrite->fComplete, true, false))
3109 {
3110 LogFlow(("Child write request completed\n"));
3111 Assert(pIoCtx->Req.Io.cbTransferLeft >= cbThisWrite);
3112 Assert(cbThisWrite == (uint32_t)cbThisWrite);
3113 rc = pIoCtxWrite->rcReq;
3114 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbThisWrite);
3115 vdIoCtxUnlockDisk(pDisk, pIoCtx, false /* fProcessDeferredReqs*/ );
3116 vdIoCtxFree(pDisk, pIoCtxWrite);
3117 }
3118 else
3119 {
3120 LogFlow(("Child write pending\n"));
3121 ASMAtomicIncU32(&pIoCtx->cDataTransfersPending);
3122 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
3123 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
3124 cbWrite -= cbThisWrite;
3125 uOffset += cbThisWrite;
3126 break;
3127 }
3128 }
3129 else
3130 {
3131 rc = VERR_VD_ASYNC_IO_IN_PROGRESS;
3132 break;
3133 }
3134 }
3135
3136 if (rc == VERR_VD_IOCTX_HALT)
3137 {
3138 cbWrite -= cbThisWrite;
3139 uOffset += cbThisWrite;
3140 pIoCtx->fFlags |= VDIOCTX_FLAGS_BLOCKED;
3141 break;
3142 }
3143 else if (rc == VERR_VD_NOT_ENOUGH_METADATA)
3144 break;
3145
3146 cbWrite -= cbThisWrite;
3147 uOffset += cbThisWrite;
3148 } while (cbWrite != 0 && (RT_SUCCESS(rc) || rc == VERR_VD_ASYNC_IO_IN_PROGRESS));
3149
3150 if ( rc == VERR_VD_ASYNC_IO_IN_PROGRESS
3151 || rc == VERR_VD_NOT_ENOUGH_METADATA
3152 || rc == VERR_VD_IOCTX_HALT)
3153 {
3154 /*
3155 * Tell the caller that we don't need to go back here because all
3156 * writes are initiated.
3157 */
3158 if ( !cbWrite
3159 && rc != VERR_VD_IOCTX_HALT)
3160 rc = VINF_SUCCESS;
3161
3162 pIoCtx->Req.Io.uOffset = uOffset;
3163 pIoCtx->Req.Io.cbTransfer = cbWrite;
3164 }
3165
3166 return rc;
3167}
3168
3169/**
3170 * Flush helper async version.
3171 */
3172static DECLCALLBACK(int) vdFlushHelperAsync(PVDIOCTX pIoCtx)
3173{
3174 int rc = VINF_SUCCESS;
3175 PVBOXHDD pDisk = pIoCtx->pDisk;
3176 PVDIMAGE pImage = pIoCtx->Req.Io.pImageCur;
3177
3178 rc = vdIoCtxLockDisk(pDisk, pIoCtx);
3179 if (RT_SUCCESS(rc))
3180 {
3181 /* Mark the whole disk as locked. */
3182 pDisk->uOffsetStartLocked = 0;
3183 pDisk->uOffsetEndLocked = UINT64_C(0xffffffffffffffff);
3184
3185 vdResetModifiedFlag(pDisk);
3186 rc = pImage->Backend->pfnFlush(pImage->pBackendData, pIoCtx);
3187 if ( ( RT_SUCCESS(rc)
3188 || rc == VERR_VD_ASYNC_IO_IN_PROGRESS
3189 || rc == VERR_VD_IOCTX_HALT)
3190 && pDisk->pCache)
3191 {
3192 rc = pDisk->pCache->Backend->pfnFlush(pDisk->pCache->pBackendData, pIoCtx);
3193 if ( RT_SUCCESS(rc)
3194 || ( rc != VERR_VD_ASYNC_IO_IN_PROGRESS
3195 && rc != VERR_VD_IOCTX_HALT))
3196 vdIoCtxUnlockDisk(pDisk, pIoCtx, true /* fProcessBlockedReqs */);
3197 else if (rc != VERR_VD_IOCTX_HALT)
3198 rc = VINF_SUCCESS;
3199 }
3200 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
3201 rc = VINF_SUCCESS;
3202 else if (rc != VERR_VD_IOCTX_HALT)/* Some other error. */
3203 vdIoCtxUnlockDisk(pDisk, pIoCtx, true /* fProcessBlockedReqs */);
3204 }
3205
3206 return rc;
3207}
3208
3209/**
3210 * Async discard helper - discards a whole block which is recorded in the block
3211 * tree.
3212 *
3213 * @returns VBox status code.
3214 * @param pIoCtx The I/O context to operate on.
3215 */
3216static DECLCALLBACK(int) vdDiscardWholeBlockAsync(PVDIOCTX pIoCtx)
3217{
3218 int rc = VINF_SUCCESS;
3219 PVBOXHDD pDisk = pIoCtx->pDisk;
3220 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
3221 PVDDISCARDBLOCK pBlock = pIoCtx->Req.Discard.pBlock;
3222 size_t cbPreAllocated, cbPostAllocated, cbActuallyDiscarded;
3223
3224 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
3225
3226 AssertPtr(pBlock);
3227
3228 rc = pDisk->pLast->Backend->pfnDiscard(pDisk->pLast->pBackendData, pIoCtx,
3229 pBlock->Core.Key, pBlock->cbDiscard,
3230 &cbPreAllocated, &cbPostAllocated,
3231 &cbActuallyDiscarded, NULL, 0);
3232 Assert(rc != VERR_VD_DISCARD_ALIGNMENT_NOT_MET);
3233 Assert(!cbPreAllocated);
3234 Assert(!cbPostAllocated);
3235 Assert(cbActuallyDiscarded == pBlock->cbDiscard || RT_FAILURE(rc));
3236
3237 /* Remove the block on success. */
3238 if ( RT_SUCCESS(rc)
3239 || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
3240 {
3241 PVDDISCARDBLOCK pBlockRemove = (PVDDISCARDBLOCK)RTAvlrU64RangeRemove(pDiscard->pTreeBlocks, pBlock->Core.Key);
3242 Assert(pBlockRemove == pBlock); RT_NOREF1(pBlockRemove);
3243
3244 pDiscard->cbDiscarding -= pBlock->cbDiscard;
3245 RTListNodeRemove(&pBlock->NodeLru);
3246 RTMemFree(pBlock->pbmAllocated);
3247 RTMemFree(pBlock);
3248 pIoCtx->Req.Discard.pBlock = NULL;/* Safety precaution. */
3249 pIoCtx->pfnIoCtxTransferNext = vdDiscardHelperAsync; /* Next part. */
3250 rc = VINF_SUCCESS;
3251 }
3252
3253 LogFlowFunc(("returns rc=%Rrc\n", rc));
3254 return rc;
3255}
3256
3257/**
3258 * Removes the least recently used blocks from the waiting list until
3259 * the new value is reached - version for async I/O.
3260 *
3261 * @returns VBox status code.
3262 * @param pDisk VD disk container.
3263 * @param pIoCtx The I/O context associated with this discard operation.
3264 * @param cbDiscardingNew How many bytes should be waiting on success.
3265 * The number of bytes waiting can be less.
3266 */
3267static int vdDiscardRemoveBlocksAsync(PVBOXHDD pDisk, PVDIOCTX pIoCtx, size_t cbDiscardingNew)
3268{
3269 int rc = VINF_SUCCESS;
3270 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
3271
3272 LogFlowFunc(("pDisk=%#p pDiscard=%#p cbDiscardingNew=%zu\n",
3273 pDisk, pDiscard, cbDiscardingNew));
3274
3275 while (pDiscard->cbDiscarding > cbDiscardingNew)
3276 {
3277 PVDDISCARDBLOCK pBlock = RTListGetLast(&pDiscard->ListLru, VDDISCARDBLOCK, NodeLru);
3278
3279 Assert(!RTListIsEmpty(&pDiscard->ListLru));
3280
3281 /* Go over the allocation bitmap and mark all discarded sectors as unused. */
3282 uint64_t offStart = pBlock->Core.Key;
3283 uint32_t idxStart = 0;
3284 size_t cbLeft = pBlock->cbDiscard;
3285 bool fAllocated = ASMBitTest(pBlock->pbmAllocated, idxStart);
3286 uint32_t cSectors = (uint32_t)(pBlock->cbDiscard / 512);
3287
3288 while (cbLeft > 0)
3289 {
3290 int32_t idxEnd;
3291 size_t cbThis = cbLeft;
3292
3293 if (fAllocated)
3294 {
3295 /* Check for the first unallocated bit. */
3296 idxEnd = ASMBitNextClear(pBlock->pbmAllocated, cSectors, idxStart);
3297 if (idxEnd != -1)
3298 {
3299 cbThis = (idxEnd - idxStart) * 512;
3300 fAllocated = false;
3301 }
3302 }
3303 else
3304 {
3305 /* Mark as unused and check for the first set bit. */
3306 idxEnd = ASMBitNextSet(pBlock->pbmAllocated, cSectors, idxStart);
3307 if (idxEnd != -1)
3308 cbThis = (idxEnd - idxStart) * 512;
3309
3310 rc = pDisk->pLast->Backend->pfnDiscard(pDisk->pLast->pBackendData, pIoCtx,
3311 offStart, cbThis, NULL, NULL, &cbThis,
3312 NULL, VD_DISCARD_MARK_UNUSED);
3313 if ( RT_FAILURE(rc)
3314 && rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
3315 break;
3316
3317 fAllocated = true;
3318 }
3319
3320 idxStart = idxEnd;
3321 offStart += cbThis;
3322 cbLeft -= cbThis;
3323 }
3324
3325 if ( RT_FAILURE(rc)
3326 && rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
3327 break;
3328
3329 PVDDISCARDBLOCK pBlockRemove = (PVDDISCARDBLOCK)RTAvlrU64RangeRemove(pDiscard->pTreeBlocks, pBlock->Core.Key);
3330 Assert(pBlockRemove == pBlock); NOREF(pBlockRemove);
3331 RTListNodeRemove(&pBlock->NodeLru);
3332
3333 pDiscard->cbDiscarding -= pBlock->cbDiscard;
3334 RTMemFree(pBlock->pbmAllocated);
3335 RTMemFree(pBlock);
3336 }
3337
3338 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
3339 rc = VINF_SUCCESS;
3340
3341 Assert(RT_FAILURE(rc) || pDiscard->cbDiscarding <= cbDiscardingNew);
3342
3343 LogFlowFunc(("returns rc=%Rrc\n", rc));
3344 return rc;
3345}
3346
3347/**
3348 * Async discard helper - discards the current range if there is no matching
3349 * block in the tree.
3350 *
3351 * @returns VBox status code.
3352 * @param pIoCtx The I/O context to operate on.
3353 */
3354static DECLCALLBACK(int) vdDiscardCurrentRangeAsync(PVDIOCTX pIoCtx)
3355{
3356 PVBOXHDD pDisk = pIoCtx->pDisk;
3357 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
3358 uint64_t offStart = pIoCtx->Req.Discard.offCur;
3359 size_t cbThisDiscard = pIoCtx->Req.Discard.cbThisDiscard;
3360 void *pbmAllocated = NULL;
3361 size_t cbPreAllocated, cbPostAllocated;
3362 int rc = VINF_SUCCESS;
3363
3364 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
3365
3366 /* No block found, try to discard using the backend first. */
3367 rc = pDisk->pLast->Backend->pfnDiscard(pDisk->pLast->pBackendData, pIoCtx,
3368 offStart, cbThisDiscard, &cbPreAllocated,
3369 &cbPostAllocated, &cbThisDiscard,
3370 &pbmAllocated, 0);
3371 if (rc == VERR_VD_DISCARD_ALIGNMENT_NOT_MET)
3372 {
3373 /* Create new discard block. */
3374 PVDDISCARDBLOCK pBlock = (PVDDISCARDBLOCK)RTMemAllocZ(sizeof(VDDISCARDBLOCK));
3375 if (pBlock)
3376 {
3377 pBlock->Core.Key = offStart - cbPreAllocated;
3378 pBlock->Core.KeyLast = offStart + cbThisDiscard + cbPostAllocated - 1;
3379 pBlock->cbDiscard = cbPreAllocated + cbThisDiscard + cbPostAllocated;
3380 pBlock->pbmAllocated = pbmAllocated;
3381 bool fInserted = RTAvlrU64Insert(pDiscard->pTreeBlocks, &pBlock->Core);
3382 Assert(fInserted); NOREF(fInserted);
3383
3384 RTListPrepend(&pDiscard->ListLru, &pBlock->NodeLru);
3385 pDiscard->cbDiscarding += pBlock->cbDiscard;
3386
3387 Assert(pIoCtx->Req.Discard.cbDiscardLeft >= cbThisDiscard);
3388 pIoCtx->Req.Discard.cbDiscardLeft -= cbThisDiscard;
3389 pIoCtx->Req.Discard.offCur += cbThisDiscard;
3390 pIoCtx->Req.Discard.cbThisDiscard = cbThisDiscard;
3391
3392 if (pDiscard->cbDiscarding > VD_DISCARD_REMOVE_THRESHOLD)
3393 rc = vdDiscardRemoveBlocksAsync(pDisk, pIoCtx, VD_DISCARD_REMOVE_THRESHOLD);
3394 else
3395 rc = VINF_SUCCESS;
3396
3397 if (RT_SUCCESS(rc))
3398 pIoCtx->pfnIoCtxTransferNext = vdDiscardHelperAsync; /* Next part. */
3399 }
3400 else
3401 {
3402 RTMemFree(pbmAllocated);
3403 rc = VERR_NO_MEMORY;
3404 }
3405 }
3406 else if ( RT_SUCCESS(rc)
3407 || rc == VERR_VD_ASYNC_IO_IN_PROGRESS) /* Save state and andvance to next range. */
3408 {
3409 Assert(pIoCtx->Req.Discard.cbDiscardLeft >= cbThisDiscard);
3410 pIoCtx->Req.Discard.cbDiscardLeft -= cbThisDiscard;
3411 pIoCtx->Req.Discard.offCur += cbThisDiscard;
3412 pIoCtx->Req.Discard.cbThisDiscard = cbThisDiscard;
3413 pIoCtx->pfnIoCtxTransferNext = vdDiscardHelperAsync;
3414 rc = VINF_SUCCESS;
3415 }
3416
3417 LogFlowFunc(("returns rc=%Rrc\n", rc));
3418 return rc;
3419}
3420
3421/**
3422 * Async discard helper - entry point.
3423 *
3424 * @returns VBox status code.
3425 * @param pIoCtx The I/O context to operate on.
3426 */
3427static DECLCALLBACK(int) vdDiscardHelperAsync(PVDIOCTX pIoCtx)
3428{
3429 int rc = VINF_SUCCESS;
3430 PVBOXHDD pDisk = pIoCtx->pDisk;
3431 PCRTRANGE paRanges = pIoCtx->Req.Discard.paRanges;
3432 unsigned cRanges = pIoCtx->Req.Discard.cRanges;
3433 PVDDISCARDSTATE pDiscard = pDisk->pDiscard;
3434
3435 LogFlowFunc(("pIoCtx=%#p\n", pIoCtx));
3436
3437 /* Check if the I/O context processed all ranges. */
3438 if ( pIoCtx->Req.Discard.idxRange == cRanges
3439 && !pIoCtx->Req.Discard.cbDiscardLeft)
3440 {
3441 LogFlowFunc(("All ranges discarded, completing\n"));
3442 vdIoCtxUnlockDisk(pDisk, pIoCtx, true /* fProcessDeferredReqs*/);
3443 return VINF_SUCCESS;
3444 }
3445
3446 if (pDisk->pIoCtxLockOwner != pIoCtx)
3447 rc = vdIoCtxLockDisk(pDisk, pIoCtx);
3448
3449 if (RT_SUCCESS(rc))
3450 {
3451 uint64_t offStart = pIoCtx->Req.Discard.offCur;
3452 size_t cbDiscardLeft = pIoCtx->Req.Discard.cbDiscardLeft;
3453 size_t cbThisDiscard;
3454
3455 pDisk->uOffsetStartLocked = offStart;
3456 pDisk->uOffsetEndLocked = offStart + cbDiscardLeft;
3457
3458 if (RT_UNLIKELY(!pDiscard))
3459 {
3460 pDiscard = vdDiscardStateCreate();
3461 if (!pDiscard)
3462 return VERR_NO_MEMORY;
3463
3464 pDisk->pDiscard = pDiscard;
3465 }
3466
3467 if (!pIoCtx->Req.Discard.cbDiscardLeft)
3468 {
3469 offStart = paRanges[pIoCtx->Req.Discard.idxRange].offStart;
3470 cbDiscardLeft = paRanges[pIoCtx->Req.Discard.idxRange].cbRange;
3471 LogFlowFunc(("New range descriptor loaded (%u) offStart=%llu cbDiscard=%zu\n",
3472 pIoCtx->Req.Discard.idxRange, offStart, cbDiscardLeft));
3473 pIoCtx->Req.Discard.idxRange++;
3474 }
3475
3476 /* Look for a matching block in the AVL tree first. */
3477 PVDDISCARDBLOCK pBlock = (PVDDISCARDBLOCK)RTAvlrU64GetBestFit(pDiscard->pTreeBlocks, offStart, false);
3478 if (!pBlock || pBlock->Core.KeyLast < offStart)
3479 {
3480 PVDDISCARDBLOCK pBlockAbove = (PVDDISCARDBLOCK)RTAvlrU64GetBestFit(pDiscard->pTreeBlocks, offStart, true);
3481
3482 /* Clip range to remain in the current block. */
3483 if (pBlockAbove)
3484 cbThisDiscard = RT_MIN(cbDiscardLeft, pBlockAbove->Core.KeyLast - offStart + 1);
3485 else
3486 cbThisDiscard = cbDiscardLeft;
3487
3488 Assert(!(cbThisDiscard % 512));
3489 pIoCtx->Req.Discard.pBlock = NULL;
3490 pIoCtx->pfnIoCtxTransferNext = vdDiscardCurrentRangeAsync;
3491 }
3492 else
3493 {
3494 /* Range lies partly in the block, update allocation bitmap. */
3495 int32_t idxStart, idxEnd;
3496
3497 cbThisDiscard = RT_MIN(cbDiscardLeft, pBlock->Core.KeyLast - offStart + 1);
3498
3499 AssertPtr(pBlock);
3500
3501 Assert(!(cbThisDiscard % 512));
3502 Assert(!((offStart - pBlock->Core.Key) % 512));
3503
3504 idxStart = (offStart - pBlock->Core.Key) / 512;
3505 idxEnd = idxStart + (int32_t)(cbThisDiscard / 512);
3506
3507 ASMBitClearRange(pBlock->pbmAllocated, idxStart, idxEnd);
3508
3509 cbDiscardLeft -= cbThisDiscard;
3510 offStart += cbThisDiscard;
3511
3512 /* Call the backend to discard the block if it is completely unallocated now. */
3513 if (ASMBitFirstSet((volatile void *)pBlock->pbmAllocated, (uint32_t)(pBlock->cbDiscard / 512)) == -1)
3514 {
3515 pIoCtx->Req.Discard.pBlock = pBlock;
3516 pIoCtx->pfnIoCtxTransferNext = vdDiscardWholeBlockAsync;
3517 rc = VINF_SUCCESS;
3518 }
3519 else
3520 {
3521 RTListNodeRemove(&pBlock->NodeLru);
3522 RTListPrepend(&pDiscard->ListLru, &pBlock->NodeLru);
3523
3524 /* Start with next range. */
3525 pIoCtx->pfnIoCtxTransferNext = vdDiscardHelperAsync;
3526 rc = VINF_SUCCESS;
3527 }
3528 }
3529
3530 /* Save state in the context. */
3531 pIoCtx->Req.Discard.offCur = offStart;
3532 pIoCtx->Req.Discard.cbDiscardLeft = cbDiscardLeft;
3533 pIoCtx->Req.Discard.cbThisDiscard = cbThisDiscard;
3534 }
3535
3536 LogFlowFunc(("returns rc=%Rrc\n", rc));
3537 return rc;
3538}
3539
3540#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
3541
3542/**
3543 * @interface_method_impl{VDBACKENDREGISTER,pfnRegisterImage}
3544 */
3545static DECLCALLBACK(int) vdPluginRegisterImage(void *pvUser, PCVDIMAGEBACKEND pBackend)
3546{
3547 int rc = VINF_SUCCESS;
3548
3549 if (VD_VERSION_ARE_COMPATIBLE(VD_IMGBACKEND_VERSION, pBackend->u32Version))
3550 vdAddBackend((RTLDRMOD)pvUser, pBackend);
3551 else
3552 {
3553 LogFunc(("ignored plugin: pBackend->u32Version=%u rc=%Rrc\n", pBackend->u32Version, rc));
3554 rc = VERR_IGNORED;
3555 }
3556
3557 return rc;
3558}
3559
3560/**
3561 * @interface_method_impl{VDBACKENDREGISTER,pfnRegisterCache}
3562 */
3563static DECLCALLBACK(int) vdPluginRegisterCache(void *pvUser, PCVDCACHEBACKEND pBackend)
3564{
3565 int rc = VINF_SUCCESS;
3566
3567 if (VD_VERSION_ARE_COMPATIBLE(VD_CACHEBACKEND_VERSION, pBackend->u32Version))
3568 vdAddCacheBackend((RTLDRMOD)pvUser, pBackend);
3569 else
3570 {
3571 LogFunc(("ignored plugin: pBackend->u32Version=%u rc=%Rrc\n", pBackend->u32Version, rc));
3572 rc = VERR_IGNORED;
3573 }
3574
3575 return rc;
3576}
3577
3578/**
3579 * @interface_method_impl{VDBACKENDREGISTER,pfnRegisterFilter}
3580 */
3581static DECLCALLBACK(int) vdPluginRegisterFilter(void *pvUser, PCVDFILTERBACKEND pBackend)
3582{
3583 int rc = VINF_SUCCESS;
3584
3585 if (VD_VERSION_ARE_COMPATIBLE(VD_FLTBACKEND_VERSION, pBackend->u32Version))
3586 vdAddFilterBackend((RTLDRMOD)pvUser, pBackend);
3587 else
3588 {
3589 LogFunc(("ignored plugin: pBackend->u32Version=%u rc=%Rrc\n", pBackend->u32Version, rc));
3590 rc = VERR_IGNORED;
3591 }
3592
3593 return rc;
3594}
3595
3596/**
3597 * Checks whether the given plugin filename was already loaded.
3598 *
3599 * @returns Pointer to already loaded plugin, NULL if not found.
3600 * @param pszFilename The filename to check.
3601 */
3602static PVDPLUGIN vdPluginFind(const char *pszFilename)
3603{
3604 PVDPLUGIN pIt = NULL;
3605
3606 RTListForEach(&g_ListPluginsLoaded, pIt, VDPLUGIN, NodePlugin)
3607 {
3608 if (!RTStrCmp(pIt->pszFilename, pszFilename))
3609 return pIt;
3610 }
3611
3612 return NULL;
3613}
3614
3615/**
3616 * Adds a plugin to the list of loaded plugins.
3617 *
3618 * @returns VBox status code.
3619 * @param hPlugin Plugin handle to add.
3620 * @param pszFilename The associated filename, used for finding duplicates.
3621 */
3622static int vdAddPlugin(RTLDRMOD hPlugin, const char *pszFilename)
3623{
3624 int rc = VINF_SUCCESS;
3625 PVDPLUGIN pPlugin = (PVDPLUGIN)RTMemAllocZ(sizeof(VDPLUGIN));
3626
3627 if (pPlugin)
3628 {
3629 pPlugin->hPlugin = hPlugin;
3630 pPlugin->pszFilename = RTStrDup(pszFilename);
3631 if (pPlugin->pszFilename)
3632 RTListAppend(&g_ListPluginsLoaded, &pPlugin->NodePlugin);
3633 else
3634 {
3635 RTMemFree(pPlugin);
3636 rc = VERR_NO_MEMORY;
3637 }
3638 }
3639 else
3640 rc = VERR_NO_MEMORY;
3641
3642 return rc;
3643}
3644
3645static int vdRemovePlugin(const char *pszFilename)
3646{
3647 /* Find plugin to be removed from the list. */
3648 PVDPLUGIN pIt = vdPluginFind(pszFilename);
3649 if (!pIt)
3650 return VINF_SUCCESS;
3651
3652 /** @todo r=klaus: need to add a plugin entry point for unregistering the
3653 * backends. Only if this doesn't exist (or fails to work) we should fall
3654 * back to the following uncoordinated backend cleanup. */
3655 for (unsigned i = 0; i < g_cBackends; i++)
3656 {
3657 while (i < g_cBackends && g_ahBackendPlugins[i] == pIt->hPlugin)
3658 {
3659 memcpy(&g_apBackends[i], &g_apBackends[i + 1], (g_cBackends - i - 1) * sizeof(PCVDIMAGEBACKEND));
3660 memcpy(&g_ahBackendPlugins[i], &g_ahBackendPlugins[i + 1], (g_cBackends - i - 1) * sizeof(RTLDRMOD));
3661 /** @todo for now skip reallocating, doesn't save much */
3662 g_cBackends--;
3663 }
3664 }
3665 for (unsigned i = 0; i < g_cCacheBackends; i++)
3666 {
3667 while (i < g_cCacheBackends && g_ahCacheBackendPlugins[i] == pIt->hPlugin)
3668 {
3669 memcpy(&g_apCacheBackends[i], &g_apCacheBackends[i + 1], (g_cCacheBackends - i - 1) * sizeof(PCVDCACHEBACKEND));
3670 memcpy(&g_ahCacheBackendPlugins[i], &g_ahCacheBackendPlugins[i + 1], (g_cCacheBackends - i - 1) * sizeof(RTLDRMOD));
3671 /** @todo for now skip reallocating, doesn't save much */
3672 g_cCacheBackends--;
3673 }
3674 }
3675 for (unsigned i = 0; i < g_cFilterBackends; i++)
3676 {
3677 while (i < g_cFilterBackends && g_pahFilterBackendPlugins[i] == pIt->hPlugin)
3678 {
3679 memcpy(&g_apFilterBackends[i], &g_apFilterBackends[i + 1], (g_cFilterBackends - i - 1) * sizeof(PCVDFILTERBACKEND));
3680 memcpy(&g_pahFilterBackendPlugins[i], &g_pahFilterBackendPlugins[i + 1], (g_cFilterBackends - i - 1) * sizeof(RTLDRMOD));
3681 /** @todo for now skip reallocating, doesn't save much */
3682 g_cFilterBackends--;
3683 }
3684 }
3685
3686 /* Remove the plugin node now, all traces of it are gone. */
3687 RTListNodeRemove(&pIt->NodePlugin);
3688 RTLdrClose(pIt->hPlugin);
3689 RTStrFree(pIt->pszFilename);
3690 RTMemFree(pIt);
3691
3692 return VINF_SUCCESS;
3693}
3694
3695#endif /* !VBOX_HDD_NO_DYNAMIC_BACKENDS */
3696
3697/**
3698 * Worker for VDPluginLoadFromFilename() and vdPluginLoadFromPath().
3699 *
3700 * @returns VBox status code.
3701 * @param pszFilename The plugin filename to load.
3702 */
3703static int vdPluginLoadFromFilename(const char *pszFilename)
3704{
3705#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
3706 /* Plugin loaded? Nothing to do. */
3707 if (vdPluginFind(pszFilename))
3708 return VINF_SUCCESS;
3709
3710 RTLDRMOD hPlugin = NIL_RTLDRMOD;
3711 int rc = SUPR3HardenedLdrLoadPlugIn(pszFilename, &hPlugin, NULL);
3712 if (RT_SUCCESS(rc))
3713 {
3714 VDBACKENDREGISTER BackendRegister;
3715 PFNVDPLUGINLOAD pfnVDPluginLoad = NULL;
3716
3717 BackendRegister.u32Version = VD_BACKENDREG_CB_VERSION;
3718 BackendRegister.pfnRegisterImage = vdPluginRegisterImage;
3719 BackendRegister.pfnRegisterCache = vdPluginRegisterCache;
3720 BackendRegister.pfnRegisterFilter = vdPluginRegisterFilter;
3721
3722 rc = RTLdrGetSymbol(hPlugin, VD_PLUGIN_LOAD_NAME, (void**)&pfnVDPluginLoad);
3723 if (RT_FAILURE(rc) || !pfnVDPluginLoad)
3724 {
3725 LogFunc(("error resolving the entry point %s in plugin %s, rc=%Rrc, pfnVDPluginLoad=%#p\n",
3726 VD_PLUGIN_LOAD_NAME, pszFilename, rc, pfnVDPluginLoad));
3727 if (RT_SUCCESS(rc))
3728 rc = VERR_SYMBOL_NOT_FOUND;
3729 }
3730
3731 if (RT_SUCCESS(rc))
3732 {
3733 /* Get the function table. */
3734 rc = pfnVDPluginLoad(hPlugin, &BackendRegister);
3735 }
3736 else
3737 LogFunc(("ignored plugin '%s': rc=%Rrc\n", pszFilename, rc));
3738
3739 /* Create a plugin entry on success. */
3740 if (RT_SUCCESS(rc))
3741 vdAddPlugin(hPlugin, pszFilename);
3742 else
3743 RTLdrClose(hPlugin);
3744 }
3745
3746 return rc;
3747#else
3748 RT_NOREF1(pszFilename);
3749 return VERR_NOT_IMPLEMENTED;
3750#endif
3751}
3752
3753/**
3754 * Worker for VDPluginLoadFromPath() and vdLoadDynamicBackends().
3755 *
3756 * @returns VBox status code.
3757 * @param pszPath The path to load plugins from.
3758 */
3759static int vdPluginLoadFromPath(const char *pszPath)
3760{
3761#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
3762 /* To get all entries with VBoxHDD as prefix. */
3763 char *pszPluginFilter = RTPathJoinA(pszPath, VD_PLUGIN_PREFIX "*");
3764 if (!pszPluginFilter)
3765 return VERR_NO_STR_MEMORY;
3766
3767 PRTDIRENTRYEX pPluginDirEntry = NULL;
3768 PRTDIR pPluginDir = NULL;
3769 size_t cbPluginDirEntry = sizeof(RTDIRENTRYEX);
3770 int rc = RTDirOpenFiltered(&pPluginDir, pszPluginFilter, RTDIRFILTER_WINNT, 0);
3771 if (RT_SUCCESS(rc))
3772 {
3773 pPluginDirEntry = (PRTDIRENTRYEX)RTMemAllocZ(sizeof(RTDIRENTRYEX));
3774 if (pPluginDirEntry)
3775 {
3776 while ( (rc = RTDirReadEx(pPluginDir, pPluginDirEntry, &cbPluginDirEntry, RTFSOBJATTRADD_NOTHING, RTPATH_F_ON_LINK))
3777 != VERR_NO_MORE_FILES)
3778 {
3779 char *pszPluginPath = NULL;
3780
3781 if (rc == VERR_BUFFER_OVERFLOW)
3782 {
3783 /* allocate new buffer. */
3784 RTMemFree(pPluginDirEntry);
3785 pPluginDirEntry = (PRTDIRENTRYEX)RTMemAllocZ(cbPluginDirEntry);
3786 if (!pPluginDirEntry)
3787 {
3788 rc = VERR_NO_MEMORY;
3789 break;
3790 }
3791 /* Retry. */
3792 rc = RTDirReadEx(pPluginDir, pPluginDirEntry, &cbPluginDirEntry, RTFSOBJATTRADD_NOTHING, RTPATH_F_ON_LINK);
3793 if (RT_FAILURE(rc))
3794 break;
3795 }
3796 else if (RT_FAILURE(rc))
3797 break;
3798
3799 /* We got the new entry. */
3800 if (!RTFS_IS_FILE(pPluginDirEntry->Info.Attr.fMode))
3801 continue;
3802
3803 /* Prepend the path to the libraries. */
3804 pszPluginPath = RTPathJoinA(pszPath, pPluginDirEntry->szName);
3805 if (!pszPluginPath)
3806 {
3807 rc = VERR_NO_STR_MEMORY;
3808 break;
3809 }
3810
3811 rc = vdPluginLoadFromFilename(pszPluginPath);
3812 RTStrFree(pszPluginPath);
3813 }
3814
3815 RTMemFree(pPluginDirEntry);
3816 }
3817 else
3818 rc = VERR_NO_MEMORY;
3819
3820 RTDirClose(pPluginDir);
3821 }
3822 else
3823 {
3824 /* On Windows the above immediately signals that there are no
3825 * files matching, while on other platforms enumerating the
3826 * files below fails. Either way: no plugins. */
3827 }
3828
3829 if (rc == VERR_NO_MORE_FILES)
3830 rc = VINF_SUCCESS;
3831 RTStrFree(pszPluginFilter);
3832 return rc;
3833#else
3834 RT_NOREF1(pszPath);
3835 return VERR_NOT_IMPLEMENTED;
3836#endif
3837}
3838
3839/**
3840 * internal: scans plugin directory and loads found plugins.
3841 */
3842static int vdLoadDynamicBackends(void)
3843{
3844#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
3845 /*
3846 * Enumerate plugin backends from the application directory where the other
3847 * shared libraries are.
3848 */
3849 char szPath[RTPATH_MAX];
3850 int rc = RTPathAppPrivateArch(szPath, sizeof(szPath));
3851 if (RT_FAILURE(rc))
3852 return rc;
3853
3854 return vdPluginLoadFromPath(szPath);
3855#else
3856 return VINF_SUCCESS;
3857#endif
3858}
3859
3860/**
3861 * Worker for VDPluginUnloadFromFilename() and vdPluginUnloadFromPath().
3862 *
3863 * @returns VBox status code.
3864 * @param pszFilename The plugin filename to unload.
3865 */
3866static int vdPluginUnloadFromFilename(const char *pszFilename)
3867{
3868#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
3869 return vdRemovePlugin(pszFilename);
3870#else
3871 RT_NOREF1(pszFilename);
3872 return VERR_NOT_IMPLEMENTED;
3873#endif
3874}
3875
3876/**
3877 * Worker for VDPluginUnloadFromPath().
3878 *
3879 * @returns VBox status code.
3880 * @param pszPath The path to unload plugins from.
3881 */
3882static int vdPluginUnloadFromPath(const char *pszPath)
3883{
3884#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
3885 /* To get all entries with VBoxHDD as prefix. */
3886 char *pszPluginFilter = RTPathJoinA(pszPath, VD_PLUGIN_PREFIX "*");
3887 if (!pszPluginFilter)
3888 return VERR_NO_STR_MEMORY;
3889
3890 PRTDIRENTRYEX pPluginDirEntry = NULL;
3891 PRTDIR pPluginDir = NULL;
3892 size_t cbPluginDirEntry = sizeof(RTDIRENTRYEX);
3893 int rc = RTDirOpenFiltered(&pPluginDir, pszPluginFilter, RTDIRFILTER_WINNT, 0);
3894 if (RT_SUCCESS(rc))
3895 {
3896 pPluginDirEntry = (PRTDIRENTRYEX)RTMemAllocZ(sizeof(RTDIRENTRYEX));
3897 if (pPluginDirEntry)
3898 {
3899 while ((rc = RTDirReadEx(pPluginDir, pPluginDirEntry, &cbPluginDirEntry, RTFSOBJATTRADD_NOTHING, RTPATH_F_ON_LINK)) != VERR_NO_MORE_FILES)
3900 {
3901 char *pszPluginPath = NULL;
3902
3903 if (rc == VERR_BUFFER_OVERFLOW)
3904 {
3905 /* allocate new buffer. */
3906 RTMemFree(pPluginDirEntry);
3907 pPluginDirEntry = (PRTDIRENTRYEX)RTMemAllocZ(cbPluginDirEntry);
3908 if (!pPluginDirEntry)
3909 {
3910 rc = VERR_NO_MEMORY;
3911 break;
3912 }
3913 /* Retry. */
3914 rc = RTDirReadEx(pPluginDir, pPluginDirEntry, &cbPluginDirEntry, RTFSOBJATTRADD_NOTHING, RTPATH_F_ON_LINK);
3915 if (RT_FAILURE(rc))
3916 break;
3917 }
3918 else if (RT_FAILURE(rc))
3919 break;
3920
3921 /* We got the new entry. */
3922 if (!RTFS_IS_FILE(pPluginDirEntry->Info.Attr.fMode))
3923 continue;
3924
3925 /* Prepend the path to the libraries. */
3926 pszPluginPath = RTPathJoinA(pszPath, pPluginDirEntry->szName);
3927 if (!pszPluginPath)
3928 {
3929 rc = VERR_NO_STR_MEMORY;
3930 break;
3931 }
3932
3933 rc = vdPluginUnloadFromFilename(pszPluginPath);
3934 RTStrFree(pszPluginPath);
3935 }
3936
3937 RTMemFree(pPluginDirEntry);
3938 }
3939 else
3940 rc = VERR_NO_MEMORY;
3941
3942 RTDirClose(pPluginDir);
3943 }
3944 else
3945 {
3946 /* On Windows the above immediately signals that there are no
3947 * files matching, while on other platforms enumerating the
3948 * files below fails. Either way: no plugins. */
3949 }
3950
3951 if (rc == VERR_NO_MORE_FILES)
3952 rc = VINF_SUCCESS;
3953 RTStrFree(pszPluginFilter);
3954 return rc;
3955#else
3956 RT_NOREF1(pszPath);
3957 return VERR_NOT_IMPLEMENTED;
3958#endif
3959}
3960
3961/**
3962 * VD async I/O interface open callback.
3963 */
3964static DECLCALLBACK(int) vdIOOpenFallback(void *pvUser, const char *pszLocation,
3965 uint32_t fOpen, PFNVDCOMPLETED pfnCompleted,
3966 void **ppStorage)
3967{
3968 RT_NOREF1(pvUser);
3969 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)RTMemAllocZ(sizeof(VDIIOFALLBACKSTORAGE));
3970
3971 if (!pStorage)
3972 return VERR_NO_MEMORY;
3973
3974 pStorage->pfnCompleted = pfnCompleted;
3975
3976 /* Open the file. */
3977 int rc = RTFileOpen(&pStorage->File, pszLocation, fOpen);
3978 if (RT_SUCCESS(rc))
3979 {
3980 *ppStorage = pStorage;
3981 return VINF_SUCCESS;
3982 }
3983
3984 RTMemFree(pStorage);
3985 return rc;
3986}
3987
3988/**
3989 * VD async I/O interface close callback.
3990 */
3991static DECLCALLBACK(int) vdIOCloseFallback(void *pvUser, void *pvStorage)
3992{
3993 RT_NOREF1(pvUser);
3994 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
3995
3996 RTFileClose(pStorage->File);
3997 RTMemFree(pStorage);
3998 return VINF_SUCCESS;
3999}
4000
4001static DECLCALLBACK(int) vdIODeleteFallback(void *pvUser, const char *pcszFilename)
4002{
4003 RT_NOREF1(pvUser);
4004 return RTFileDelete(pcszFilename);
4005}
4006
4007static DECLCALLBACK(int) vdIOMoveFallback(void *pvUser, const char *pcszSrc, const char *pcszDst, unsigned fMove)
4008{
4009 RT_NOREF1(pvUser);
4010 return RTFileMove(pcszSrc, pcszDst, fMove);
4011}
4012
4013static DECLCALLBACK(int) vdIOGetFreeSpaceFallback(void *pvUser, const char *pcszFilename, int64_t *pcbFreeSpace)
4014{
4015 RT_NOREF1(pvUser);
4016 return RTFsQuerySizes(pcszFilename, NULL, pcbFreeSpace, NULL, NULL);
4017}
4018
4019static DECLCALLBACK(int) vdIOGetModificationTimeFallback(void *pvUser, const char *pcszFilename, PRTTIMESPEC pModificationTime)
4020{
4021 RT_NOREF1(pvUser);
4022 RTFSOBJINFO info;
4023 int rc = RTPathQueryInfo(pcszFilename, &info, RTFSOBJATTRADD_NOTHING);
4024 if (RT_SUCCESS(rc))
4025 *pModificationTime = info.ModificationTime;
4026 return rc;
4027}
4028
4029/**
4030 * VD async I/O interface callback for retrieving the file size.
4031 */
4032static DECLCALLBACK(int) vdIOGetSizeFallback(void *pvUser, void *pvStorage, uint64_t *pcbSize)
4033{
4034 RT_NOREF1(pvUser);
4035 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
4036
4037 return RTFileGetSize(pStorage->File, pcbSize);
4038}
4039
4040/**
4041 * VD async I/O interface callback for setting the file size.
4042 */
4043static DECLCALLBACK(int) vdIOSetSizeFallback(void *pvUser, void *pvStorage, uint64_t cbSize)
4044{
4045 RT_NOREF1(pvUser);
4046 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
4047
4048 return RTFileSetSize(pStorage->File, cbSize);
4049}
4050
4051/**
4052 * VD async I/O interface callback for setting the file allocation size.
4053 */
4054static DECLCALLBACK(int) vdIOSetAllocationSizeFallback(void *pvUser, void *pvStorage, uint64_t cbSize,
4055 uint32_t fFlags)
4056{
4057 RT_NOREF2(pvUser, fFlags);
4058 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
4059
4060 return RTFileSetAllocationSize(pStorage->File, cbSize, RTFILE_ALLOC_SIZE_F_DEFAULT);
4061}
4062
4063/**
4064 * VD async I/O interface callback for a synchronous write to the file.
4065 */
4066static DECLCALLBACK(int) vdIOWriteSyncFallback(void *pvUser, void *pvStorage, uint64_t uOffset,
4067 const void *pvBuf, size_t cbWrite, size_t *pcbWritten)
4068{
4069 RT_NOREF1(pvUser);
4070 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
4071
4072 return RTFileWriteAt(pStorage->File, uOffset, pvBuf, cbWrite, pcbWritten);
4073}
4074
4075/**
4076 * VD async I/O interface callback for a synchronous read from the file.
4077 */
4078static DECLCALLBACK(int) vdIOReadSyncFallback(void *pvUser, void *pvStorage, uint64_t uOffset,
4079 void *pvBuf, size_t cbRead, size_t *pcbRead)
4080{
4081 RT_NOREF1(pvUser);
4082 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
4083
4084 return RTFileReadAt(pStorage->File, uOffset, pvBuf, cbRead, pcbRead);
4085}
4086
4087/**
4088 * VD async I/O interface callback for a synchronous flush of the file data.
4089 */
4090static DECLCALLBACK(int) vdIOFlushSyncFallback(void *pvUser, void *pvStorage)
4091{
4092 RT_NOREF1(pvUser);
4093 PVDIIOFALLBACKSTORAGE pStorage = (PVDIIOFALLBACKSTORAGE)pvStorage;
4094
4095 return RTFileFlush(pStorage->File);
4096}
4097
4098/**
4099 * VD async I/O interface callback for a asynchronous read from the file.
4100 */
4101static DECLCALLBACK(int) vdIOReadAsyncFallback(void *pvUser, void *pStorage, uint64_t uOffset,
4102 PCRTSGSEG paSegments, size_t cSegments,
4103 size_t cbRead, void *pvCompletion,
4104 void **ppTask)
4105{
4106 RT_NOREF8(pvUser, pStorage, uOffset, paSegments, cSegments, cbRead, pvCompletion, ppTask);
4107 return VERR_NOT_IMPLEMENTED;
4108}
4109
4110/**
4111 * VD async I/O interface callback for a asynchronous write to the file.
4112 */
4113static DECLCALLBACK(int) vdIOWriteAsyncFallback(void *pvUser, void *pStorage, uint64_t uOffset,
4114 PCRTSGSEG paSegments, size_t cSegments,
4115 size_t cbWrite, void *pvCompletion,
4116 void **ppTask)
4117{
4118 RT_NOREF8(pvUser, pStorage, uOffset, paSegments, cSegments, cbWrite, pvCompletion, ppTask);
4119 return VERR_NOT_IMPLEMENTED;
4120}
4121
4122/**
4123 * VD async I/O interface callback for a asynchronous flush of the file data.
4124 */
4125static DECLCALLBACK(int) vdIOFlushAsyncFallback(void *pvUser, void *pStorage,
4126 void *pvCompletion, void **ppTask)
4127{
4128 RT_NOREF4(pvUser, pStorage, pvCompletion, ppTask);
4129 return VERR_NOT_IMPLEMENTED;
4130}
4131
4132/**
4133 * Internal - Continues an I/O context after
4134 * it was halted because of an active transfer.
4135 */
4136static int vdIoCtxContinue(PVDIOCTX pIoCtx, int rcReq)
4137{
4138 PVBOXHDD pDisk = pIoCtx->pDisk;
4139 int rc = VINF_SUCCESS;
4140
4141 VD_IS_LOCKED(pDisk);
4142
4143 if (RT_FAILURE(rcReq))
4144 ASMAtomicCmpXchgS32(&pIoCtx->rcReq, rcReq, VINF_SUCCESS);
4145
4146 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_BLOCKED))
4147 {
4148 /* Continue the transfer */
4149 rc = vdIoCtxProcessLocked(pIoCtx);
4150
4151 if ( rc == VINF_VD_ASYNC_IO_FINISHED
4152 && ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
4153 {
4154 LogFlowFunc(("I/O context completed pIoCtx=%#p\n", pIoCtx));
4155 if (pIoCtx->pIoCtxParent)
4156 {
4157 PVDIOCTX pIoCtxParent = pIoCtx->pIoCtxParent;
4158
4159 Assert(!pIoCtxParent->pIoCtxParent);
4160 if (RT_FAILURE(pIoCtx->rcReq))
4161 ASMAtomicCmpXchgS32(&pIoCtxParent->rcReq, pIoCtx->rcReq, VINF_SUCCESS);
4162
4163 ASMAtomicDecU32(&pIoCtxParent->cDataTransfersPending);
4164
4165 if (pIoCtx->enmTxDir == VDIOCTXTXDIR_WRITE)
4166 {
4167 LogFlowFunc(("I/O context transferred %u bytes for the parent pIoCtxParent=%p\n",
4168 pIoCtx->Type.Child.cbTransferParent, pIoCtxParent));
4169
4170 /* Update the parent state. */
4171 Assert(pIoCtxParent->Req.Io.cbTransferLeft >= pIoCtx->Type.Child.cbTransferParent);
4172 ASMAtomicSubU32(&pIoCtxParent->Req.Io.cbTransferLeft, (uint32_t)pIoCtx->Type.Child.cbTransferParent);
4173 }
4174 else
4175 Assert(pIoCtx->enmTxDir == VDIOCTXTXDIR_FLUSH);
4176
4177 /*
4178 * A completed child write means that we finished growing the image.
4179 * We have to process any pending writes now.
4180 */
4181 vdIoCtxUnlockDisk(pDisk, pIoCtxParent, false /* fProcessDeferredReqs */);
4182
4183 /* Unblock the parent */
4184 pIoCtxParent->fFlags &= ~VDIOCTX_FLAGS_BLOCKED;
4185
4186 rc = vdIoCtxProcessLocked(pIoCtxParent);
4187
4188 if ( rc == VINF_VD_ASYNC_IO_FINISHED
4189 && ASMAtomicCmpXchgBool(&pIoCtxParent->fComplete, true, false))
4190 {
4191 LogFlowFunc(("Parent I/O context completed pIoCtxParent=%#p rcReq=%Rrc\n", pIoCtxParent, pIoCtxParent->rcReq));
4192 vdIoCtxRootComplete(pDisk, pIoCtxParent);
4193 vdThreadFinishWrite(pDisk);
4194 vdIoCtxFree(pDisk, pIoCtxParent);
4195 vdDiskProcessBlockedIoCtx(pDisk);
4196 }
4197 else if (!vdIoCtxIsDiskLockOwner(pDisk, pIoCtx))
4198 {
4199 /* Process any pending writes if the current request didn't caused another growing. */
4200 vdDiskProcessBlockedIoCtx(pDisk);
4201 }
4202 }
4203 else
4204 {
4205 if (pIoCtx->enmTxDir == VDIOCTXTXDIR_FLUSH)
4206 {
4207 vdIoCtxUnlockDisk(pDisk, pIoCtx, true /* fProcessDerredReqs */);
4208 vdThreadFinishWrite(pDisk);
4209 }
4210 else if ( pIoCtx->enmTxDir == VDIOCTXTXDIR_WRITE
4211 || pIoCtx->enmTxDir == VDIOCTXTXDIR_DISCARD)
4212 vdThreadFinishWrite(pDisk);
4213 else
4214 {
4215 Assert(pIoCtx->enmTxDir == VDIOCTXTXDIR_READ);
4216 vdThreadFinishRead(pDisk);
4217 }
4218
4219 LogFlowFunc(("I/O context completed pIoCtx=%#p rcReq=%Rrc\n", pIoCtx, pIoCtx->rcReq));
4220 vdIoCtxRootComplete(pDisk, pIoCtx);
4221 }
4222
4223 vdIoCtxFree(pDisk, pIoCtx);
4224 }
4225 }
4226
4227 return VINF_SUCCESS;
4228}
4229
4230/**
4231 * Internal - Called when user transfer completed.
4232 */
4233static int vdUserXferCompleted(PVDIOSTORAGE pIoStorage, PVDIOCTX pIoCtx,
4234 PFNVDXFERCOMPLETED pfnComplete, void *pvUser,
4235 size_t cbTransfer, int rcReq)
4236{
4237 int rc = VINF_SUCCESS;
4238 PVBOXHDD pDisk = pIoCtx->pDisk;
4239
4240 LogFlowFunc(("pIoStorage=%#p pIoCtx=%#p pfnComplete=%#p pvUser=%#p cbTransfer=%zu rcReq=%Rrc\n",
4241 pIoStorage, pIoCtx, pfnComplete, pvUser, cbTransfer, rcReq));
4242
4243 VD_IS_LOCKED(pDisk);
4244
4245 Assert(pIoCtx->Req.Io.cbTransferLeft >= cbTransfer);
4246 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbTransfer); Assert(cbTransfer == (uint32_t)cbTransfer);
4247 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
4248
4249 if (pfnComplete)
4250 rc = pfnComplete(pIoStorage->pVDIo->pBackendData, pIoCtx, pvUser, rcReq);
4251
4252 if (RT_SUCCESS(rc))
4253 rc = vdIoCtxContinue(pIoCtx, rcReq);
4254 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
4255 rc = VINF_SUCCESS;
4256
4257 return rc;
4258}
4259
4260static void vdIoCtxContinueDeferredList(PVDIOSTORAGE pIoStorage, PRTLISTANCHOR pListWaiting,
4261 PFNVDXFERCOMPLETED pfnComplete, void *pvUser, int rcReq)
4262{
4263 LogFlowFunc(("pIoStorage=%#p pListWaiting=%#p pfnComplete=%#p pvUser=%#p rcReq=%Rrc\n",
4264 pIoStorage, pListWaiting, pfnComplete, pvUser, rcReq));
4265
4266 /* Go through the waiting list and continue the I/O contexts. */
4267 while (!RTListIsEmpty(pListWaiting))
4268 {
4269 int rc = VINF_SUCCESS;
4270 PVDIOCTXDEFERRED pDeferred = RTListGetFirst(pListWaiting, VDIOCTXDEFERRED, NodeDeferred);
4271 PVDIOCTX pIoCtx = pDeferred->pIoCtx;
4272 RTListNodeRemove(&pDeferred->NodeDeferred);
4273
4274 RTMemFree(pDeferred);
4275 ASMAtomicDecU32(&pIoCtx->cMetaTransfersPending);
4276
4277 if (pfnComplete)
4278 rc = pfnComplete(pIoStorage->pVDIo->pBackendData, pIoCtx, pvUser, rcReq);
4279
4280 LogFlow(("Completion callback for I/O context %#p returned %Rrc\n", pIoCtx, rc));
4281
4282 if (RT_SUCCESS(rc))
4283 {
4284 rc = vdIoCtxContinue(pIoCtx, rcReq);
4285 AssertRC(rc);
4286 }
4287 else
4288 Assert(rc == VERR_VD_ASYNC_IO_IN_PROGRESS);
4289 }
4290}
4291
4292/**
4293 * Internal - Called when a meta transfer completed.
4294 */
4295static int vdMetaXferCompleted(PVDIOSTORAGE pIoStorage, PFNVDXFERCOMPLETED pfnComplete, void *pvUser,
4296 PVDMETAXFER pMetaXfer, int rcReq)
4297{
4298 PVBOXHDD pDisk = pIoStorage->pVDIo->pDisk;
4299 RTLISTNODE ListIoCtxWaiting;
4300 bool fFlush;
4301
4302 LogFlowFunc(("pIoStorage=%#p pfnComplete=%#p pvUser=%#p pMetaXfer=%#p rcReq=%Rrc\n",
4303 pIoStorage, pfnComplete, pvUser, pMetaXfer, rcReq));
4304
4305 VD_IS_LOCKED(pDisk);
4306
4307 fFlush = VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_FLUSH;
4308
4309 if (!fFlush)
4310 {
4311 RTListMove(&ListIoCtxWaiting, &pMetaXfer->ListIoCtxWaiting);
4312
4313 if (RT_FAILURE(rcReq))
4314 {
4315 /* Remove from the AVL tree. */
4316 LogFlow(("Removing meta xfer=%#p\n", pMetaXfer));
4317 bool fRemoved = RTAvlrFileOffsetRemove(pIoStorage->pTreeMetaXfers, pMetaXfer->Core.Key) != NULL;
4318 Assert(fRemoved); NOREF(fRemoved);
4319 /* If this was a write check if there is a shadow buffer with updated data. */
4320 if (pMetaXfer->pbDataShw)
4321 {
4322 Assert(VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_WRITE);
4323 Assert(!RTListIsEmpty(&pMetaXfer->ListIoCtxShwWrites));
4324 RTListConcatenate(&ListIoCtxWaiting, &pMetaXfer->ListIoCtxShwWrites);
4325 RTMemFree(pMetaXfer->pbDataShw);
4326 pMetaXfer->pbDataShw = NULL;
4327 }
4328 RTMemFree(pMetaXfer);
4329 }
4330 else
4331 {
4332 /* Increase the reference counter to make sure it doesn't go away before the last context is processed. */
4333 pMetaXfer->cRefs++;
4334 }
4335 }
4336 else
4337 RTListMove(&ListIoCtxWaiting, &pMetaXfer->ListIoCtxWaiting);
4338
4339 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_NONE);
4340 vdIoCtxContinueDeferredList(pIoStorage, &ListIoCtxWaiting, pfnComplete, pvUser, rcReq);
4341
4342 /*
4343 * If there is a shadow buffer and the previous write was successful update with the
4344 * new data and trigger a new write.
4345 */
4346 if ( pMetaXfer->pbDataShw
4347 && RT_SUCCESS(rcReq)
4348 && VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_NONE)
4349 {
4350 LogFlowFunc(("pMetaXfer=%#p Updating from shadow buffer and triggering new write\n", pMetaXfer));
4351 memcpy(pMetaXfer->abData, pMetaXfer->pbDataShw, pMetaXfer->cbMeta);
4352 RTMemFree(pMetaXfer->pbDataShw);
4353 pMetaXfer->pbDataShw = NULL;
4354 Assert(!RTListIsEmpty(&pMetaXfer->ListIoCtxShwWrites));
4355
4356 /* Setup a new I/O write. */
4357 PVDIOTASK pIoTask = vdIoTaskMetaAlloc(pIoStorage, pfnComplete, pvUser, pMetaXfer);
4358 if (RT_LIKELY(pIoTask))
4359 {
4360 void *pvTask = NULL;
4361 RTSGSEG Seg;
4362
4363 Seg.cbSeg = pMetaXfer->cbMeta;
4364 Seg.pvSeg = pMetaXfer->abData;
4365
4366 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_WRITE);
4367 rcReq = pIoStorage->pVDIo->pInterfaceIo->pfnWriteAsync(pIoStorage->pVDIo->pInterfaceIo->Core.pvUser,
4368 pIoStorage->pStorage,
4369 pMetaXfer->Core.Key, &Seg, 1,
4370 pMetaXfer->cbMeta, pIoTask,
4371 &pvTask);
4372 if ( RT_SUCCESS(rcReq)
4373 || rcReq != VERR_VD_ASYNC_IO_IN_PROGRESS)
4374 {
4375 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_NONE);
4376 vdIoTaskFree(pDisk, pIoTask);
4377 }
4378 else
4379 RTListMove(&pMetaXfer->ListIoCtxWaiting, &pMetaXfer->ListIoCtxShwWrites);
4380 }
4381 else
4382 rcReq = VERR_NO_MEMORY;
4383
4384 /* Cleanup if there was an error or the request completed already. */
4385 if (rcReq != VERR_VD_ASYNC_IO_IN_PROGRESS)
4386 vdIoCtxContinueDeferredList(pIoStorage, &pMetaXfer->ListIoCtxShwWrites, pfnComplete, pvUser, rcReq);
4387 }
4388
4389 /* Remove if not used anymore. */
4390 if (!fFlush)
4391 {
4392 pMetaXfer->cRefs--;
4393 if (!pMetaXfer->cRefs && RTListIsEmpty(&pMetaXfer->ListIoCtxWaiting))
4394 {
4395 /* Remove from the AVL tree. */
4396 LogFlow(("Removing meta xfer=%#p\n", pMetaXfer));
4397 bool fRemoved = RTAvlrFileOffsetRemove(pIoStorage->pTreeMetaXfers, pMetaXfer->Core.Key) != NULL;
4398 Assert(fRemoved); NOREF(fRemoved);
4399 RTMemFree(pMetaXfer);
4400 }
4401 }
4402 else if (fFlush)
4403 RTMemFree(pMetaXfer);
4404
4405 return VINF_SUCCESS;
4406}
4407
4408/**
4409 * Processes a list of waiting I/O tasks. The disk lock must be held by caller.
4410 *
4411 * @returns nothing.
4412 * @param pDisk The disk to process the list for.
4413 */
4414static void vdIoTaskProcessWaitingList(PVBOXHDD pDisk)
4415{
4416 LogFlowFunc(("pDisk=%#p\n", pDisk));
4417
4418 VD_IS_LOCKED(pDisk);
4419
4420 PVDIOTASK pHead = ASMAtomicXchgPtrT(&pDisk->pIoTasksPendingHead, NULL, PVDIOTASK);
4421
4422 Log(("I/O task list cleared\n"));
4423
4424 /* Reverse order. */
4425 PVDIOTASK pCur = pHead;
4426 pHead = NULL;
4427 while (pCur)
4428 {
4429 PVDIOTASK pInsert = pCur;
4430 pCur = pCur->pNext;
4431 pInsert->pNext = pHead;
4432 pHead = pInsert;
4433 }
4434
4435 while (pHead)
4436 {
4437 PVDIOSTORAGE pIoStorage = pHead->pIoStorage;
4438
4439 if (!pHead->fMeta)
4440 vdUserXferCompleted(pIoStorage, pHead->Type.User.pIoCtx,
4441 pHead->pfnComplete, pHead->pvUser,
4442 pHead->Type.User.cbTransfer, pHead->rcReq);
4443 else
4444 vdMetaXferCompleted(pIoStorage, pHead->pfnComplete, pHead->pvUser,
4445 pHead->Type.Meta.pMetaXfer, pHead->rcReq);
4446
4447 pCur = pHead;
4448 pHead = pHead->pNext;
4449 vdIoTaskFree(pDisk, pCur);
4450 }
4451}
4452
4453/**
4454 * Process any I/O context on the halted list.
4455 *
4456 * @returns nothing.
4457 * @param pDisk The disk.
4458 */
4459static void vdIoCtxProcessHaltedList(PVBOXHDD pDisk)
4460{
4461 LogFlowFunc(("pDisk=%#p\n", pDisk));
4462
4463 VD_IS_LOCKED(pDisk);
4464
4465 /* Get the waiting list and process it in FIFO order. */
4466 PVDIOCTX pIoCtxHead = ASMAtomicXchgPtrT(&pDisk->pIoCtxHaltedHead, NULL, PVDIOCTX);
4467
4468 /* Reverse it. */
4469 PVDIOCTX pCur = pIoCtxHead;
4470 pIoCtxHead = NULL;
4471 while (pCur)
4472 {
4473 PVDIOCTX pInsert = pCur;
4474 pCur = pCur->pIoCtxNext;
4475 pInsert->pIoCtxNext = pIoCtxHead;
4476 pIoCtxHead = pInsert;
4477 }
4478
4479 /* Process now. */
4480 pCur = pIoCtxHead;
4481 while (pCur)
4482 {
4483 PVDIOCTX pTmp = pCur;
4484
4485 pCur = pCur->pIoCtxNext;
4486 pTmp->pIoCtxNext = NULL;
4487
4488 /* Continue */
4489 pTmp->fFlags &= ~VDIOCTX_FLAGS_BLOCKED;
4490 vdIoCtxContinue(pTmp, pTmp->rcReq);
4491 }
4492}
4493
4494/**
4495 * Unlock the disk and process pending tasks.
4496 *
4497 * @returns VBox status code.
4498 * @param pDisk The disk to unlock.
4499 * @param pIoCtxRc The I/O context to get the status code from, optional.
4500 */
4501static int vdDiskUnlock(PVBOXHDD pDisk, PVDIOCTX pIoCtxRc)
4502{
4503 int rc = VINF_SUCCESS;
4504
4505 VD_IS_LOCKED(pDisk);
4506
4507 /*
4508 * Process the list of waiting I/O tasks first
4509 * because they might complete I/O contexts.
4510 * Same for the list of halted I/O contexts.
4511 * Afterwards comes the list of new I/O contexts.
4512 */
4513 vdIoTaskProcessWaitingList(pDisk);
4514 vdIoCtxProcessHaltedList(pDisk);
4515 rc = vdDiskProcessWaitingIoCtx(pDisk, pIoCtxRc);
4516 ASMAtomicXchgBool(&pDisk->fLocked, false);
4517
4518 /*
4519 * Need to check for new I/O tasks and waiting I/O contexts now
4520 * again as other threads might added them while we processed
4521 * previous lists.
4522 */
4523 while ( ASMAtomicUoReadPtrT(&pDisk->pIoCtxHead, PVDIOCTX) != NULL
4524 || ASMAtomicUoReadPtrT(&pDisk->pIoTasksPendingHead, PVDIOTASK) != NULL
4525 || ASMAtomicUoReadPtrT(&pDisk->pIoCtxHaltedHead, PVDIOCTX) != NULL)
4526 {
4527 /* Try lock disk again. */
4528 if (ASMAtomicCmpXchgBool(&pDisk->fLocked, true, false))
4529 {
4530 vdIoTaskProcessWaitingList(pDisk);
4531 vdIoCtxProcessHaltedList(pDisk);
4532 vdDiskProcessWaitingIoCtx(pDisk, NULL);
4533 ASMAtomicXchgBool(&pDisk->fLocked, false);
4534 }
4535 else /* Let the other thread everything when he unlocks the disk. */
4536 break;
4537 }
4538
4539 return rc;
4540}
4541
4542/**
4543 * Try to lock the disk to complete pressing of the I/O task.
4544 * The completion is deferred if the disk is locked already.
4545 *
4546 * @returns nothing.
4547 * @param pIoTask The I/O task to complete.
4548 */
4549static void vdXferTryLockDiskDeferIoTask(PVDIOTASK pIoTask)
4550{
4551 PVDIOSTORAGE pIoStorage = pIoTask->pIoStorage;
4552 PVBOXHDD pDisk = pIoStorage->pVDIo->pDisk;
4553
4554 Log(("Deferring I/O task pIoTask=%p\n", pIoTask));
4555
4556 /* Put it on the waiting list. */
4557 PVDIOTASK pNext = ASMAtomicUoReadPtrT(&pDisk->pIoTasksPendingHead, PVDIOTASK);
4558 PVDIOTASK pHeadOld;
4559 pIoTask->pNext = pNext;
4560 while (!ASMAtomicCmpXchgExPtr(&pDisk->pIoTasksPendingHead, pIoTask, pNext, &pHeadOld))
4561 {
4562 pNext = pHeadOld;
4563 Assert(pNext != pIoTask);
4564 pIoTask->pNext = pNext;
4565 ASMNopPause();
4566 }
4567
4568 if (ASMAtomicCmpXchgBool(&pDisk->fLocked, true, false))
4569 {
4570 /* Release disk lock, it will take care of processing all lists. */
4571 vdDiskUnlock(pDisk, NULL);
4572 }
4573}
4574
4575static DECLCALLBACK(int) vdIOIntReqCompleted(void *pvUser, int rcReq)
4576{
4577 PVDIOTASK pIoTask = (PVDIOTASK)pvUser;
4578
4579 LogFlowFunc(("Task completed pIoTask=%#p\n", pIoTask));
4580
4581 pIoTask->rcReq = rcReq;
4582 vdXferTryLockDiskDeferIoTask(pIoTask);
4583 return VINF_SUCCESS;
4584}
4585
4586/**
4587 * VD I/O interface callback for opening a file.
4588 */
4589static DECLCALLBACK(int) vdIOIntOpen(void *pvUser, const char *pszLocation,
4590 unsigned uOpenFlags, PPVDIOSTORAGE ppIoStorage)
4591{
4592 int rc = VINF_SUCCESS;
4593 PVDIO pVDIo = (PVDIO)pvUser;
4594 PVDIOSTORAGE pIoStorage = (PVDIOSTORAGE)RTMemAllocZ(sizeof(VDIOSTORAGE));
4595
4596 if (!pIoStorage)
4597 return VERR_NO_MEMORY;
4598
4599 /* Create the AVl tree. */
4600 pIoStorage->pTreeMetaXfers = (PAVLRFOFFTREE)RTMemAllocZ(sizeof(AVLRFOFFTREE));
4601 if (pIoStorage->pTreeMetaXfers)
4602 {
4603 rc = pVDIo->pInterfaceIo->pfnOpen(pVDIo->pInterfaceIo->Core.pvUser,
4604 pszLocation, uOpenFlags,
4605 vdIOIntReqCompleted,
4606 &pIoStorage->pStorage);
4607 if (RT_SUCCESS(rc))
4608 {
4609 pIoStorage->pVDIo = pVDIo;
4610 *ppIoStorage = pIoStorage;
4611 return VINF_SUCCESS;
4612 }
4613
4614 RTMemFree(pIoStorage->pTreeMetaXfers);
4615 }
4616 else
4617 rc = VERR_NO_MEMORY;
4618
4619 RTMemFree(pIoStorage);
4620 return rc;
4621}
4622
4623static DECLCALLBACK(int) vdIOIntTreeMetaXferDestroy(PAVLRFOFFNODECORE pNode, void *pvUser)
4624{
4625 RT_NOREF2(pNode, pvUser);
4626 AssertMsgFailed(("Tree should be empty at this point!\n"));
4627 return VINF_SUCCESS;
4628}
4629
4630static DECLCALLBACK(int) vdIOIntClose(void *pvUser, PVDIOSTORAGE pIoStorage)
4631{
4632 int rc = VINF_SUCCESS;
4633 PVDIO pVDIo = (PVDIO)pvUser;
4634
4635 /* We free everything here, even if closing the file failed for some reason. */
4636 rc = pVDIo->pInterfaceIo->pfnClose(pVDIo->pInterfaceIo->Core.pvUser, pIoStorage->pStorage);
4637 RTAvlrFileOffsetDestroy(pIoStorage->pTreeMetaXfers, vdIOIntTreeMetaXferDestroy, NULL);
4638 RTMemFree(pIoStorage->pTreeMetaXfers);
4639 RTMemFree(pIoStorage);
4640 return rc;
4641}
4642
4643static DECLCALLBACK(int) vdIOIntDelete(void *pvUser, const char *pcszFilename)
4644{
4645 PVDIO pVDIo = (PVDIO)pvUser;
4646 return pVDIo->pInterfaceIo->pfnDelete(pVDIo->pInterfaceIo->Core.pvUser,
4647 pcszFilename);
4648}
4649
4650static DECLCALLBACK(int) vdIOIntMove(void *pvUser, const char *pcszSrc, const char *pcszDst,
4651 unsigned fMove)
4652{
4653 PVDIO pVDIo = (PVDIO)pvUser;
4654 return pVDIo->pInterfaceIo->pfnMove(pVDIo->pInterfaceIo->Core.pvUser,
4655 pcszSrc, pcszDst, fMove);
4656}
4657
4658static DECLCALLBACK(int) vdIOIntGetFreeSpace(void *pvUser, const char *pcszFilename,
4659 int64_t *pcbFreeSpace)
4660{
4661 PVDIO pVDIo = (PVDIO)pvUser;
4662 return pVDIo->pInterfaceIo->pfnGetFreeSpace(pVDIo->pInterfaceIo->Core.pvUser,
4663 pcszFilename, pcbFreeSpace);
4664}
4665
4666static DECLCALLBACK(int) vdIOIntGetModificationTime(void *pvUser, const char *pcszFilename,
4667 PRTTIMESPEC pModificationTime)
4668{
4669 PVDIO pVDIo = (PVDIO)pvUser;
4670 return pVDIo->pInterfaceIo->pfnGetModificationTime(pVDIo->pInterfaceIo->Core.pvUser,
4671 pcszFilename, pModificationTime);
4672}
4673
4674static DECLCALLBACK(int) vdIOIntGetSize(void *pvUser, PVDIOSTORAGE pIoStorage,
4675 uint64_t *pcbSize)
4676{
4677 PVDIO pVDIo = (PVDIO)pvUser;
4678 return pVDIo->pInterfaceIo->pfnGetSize(pVDIo->pInterfaceIo->Core.pvUser,
4679 pIoStorage->pStorage, pcbSize);
4680}
4681
4682static DECLCALLBACK(int) vdIOIntSetSize(void *pvUser, PVDIOSTORAGE pIoStorage,
4683 uint64_t cbSize)
4684{
4685 PVDIO pVDIo = (PVDIO)pvUser;
4686 return pVDIo->pInterfaceIo->pfnSetSize(pVDIo->pInterfaceIo->Core.pvUser,
4687 pIoStorage->pStorage, cbSize);
4688}
4689
4690static DECLCALLBACK(int) vdIOIntSetAllocationSize(void *pvUser, PVDIOSTORAGE pIoStorage,
4691 uint64_t cbSize, uint32_t fFlags,
4692 PVDINTERFACEPROGRESS pIfProgress,
4693 unsigned uPercentStart, unsigned uPercentSpan)
4694{
4695 PVDIO pVDIo = (PVDIO)pvUser;
4696 int rc = pVDIo->pInterfaceIo->pfnSetAllocationSize(pVDIo->pInterfaceIo->Core.pvUser,
4697 pIoStorage->pStorage, cbSize, fFlags);
4698 if (rc == VERR_NOT_SUPPORTED)
4699 {
4700 /* Fallback if the underlying medium does not support optimized storage allocation. */
4701 uint64_t cbSizeCur = 0;
4702 rc = pVDIo->pInterfaceIo->pfnGetSize(pVDIo->pInterfaceIo->Core.pvUser,
4703 pIoStorage->pStorage, &cbSizeCur);
4704 if (RT_SUCCESS(rc))
4705 {
4706 if (cbSizeCur < cbSize)
4707 {
4708 const size_t cbBuf = 128 * _1K;
4709 void *pvBuf = RTMemTmpAllocZ(cbBuf);
4710 if (RT_LIKELY(pvBuf))
4711 {
4712 uint64_t cbFill = cbSize - cbSizeCur;
4713 uint64_t uOff = 0;
4714
4715 /* Write data to all blocks. */
4716 while ( uOff < cbFill
4717 && RT_SUCCESS(rc))
4718 {
4719 size_t cbChunk = (size_t)RT_MIN(cbFill - uOff, cbBuf);
4720
4721 rc = pVDIo->pInterfaceIo->pfnWriteSync(pVDIo->pInterfaceIo->Core.pvUser,
4722 pIoStorage->pStorage, cbSizeCur + uOff,
4723 pvBuf, cbChunk, NULL);
4724 if (RT_SUCCESS(rc))
4725 {
4726 uOff += cbChunk;
4727
4728 rc = vdIfProgress(pIfProgress, uPercentStart + uOff * uPercentSpan / cbFill);
4729 }
4730 }
4731
4732 RTMemTmpFree(pvBuf);
4733 }
4734 else
4735 rc = VERR_NO_MEMORY;
4736 }
4737 else if (cbSizeCur > cbSize)
4738 rc = pVDIo->pInterfaceIo->pfnSetSize(pVDIo->pInterfaceIo->Core.pvUser,
4739 pIoStorage->pStorage, cbSize);
4740 }
4741 }
4742
4743 if (RT_SUCCESS(rc))
4744 rc = vdIfProgress(pIfProgress, uPercentStart + uPercentSpan);
4745
4746 return rc;
4747}
4748
4749static DECLCALLBACK(int) vdIOIntReadUser(void *pvUser, PVDIOSTORAGE pIoStorage, uint64_t uOffset,
4750 PVDIOCTX pIoCtx, size_t cbRead)
4751{
4752 int rc = VINF_SUCCESS;
4753 PVDIO pVDIo = (PVDIO)pvUser;
4754 PVBOXHDD pDisk = pVDIo->pDisk;
4755
4756 LogFlowFunc(("pvUser=%#p pIoStorage=%#p uOffset=%llu pIoCtx=%#p cbRead=%u\n",
4757 pvUser, pIoStorage, uOffset, pIoCtx, cbRead));
4758
4759 /** @todo Enable check for sync I/O later. */
4760 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4761 VD_IS_LOCKED(pDisk);
4762
4763 Assert(cbRead > 0);
4764
4765 if (pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
4766 {
4767 RTSGSEG Seg;
4768 unsigned cSegments = 1;
4769 size_t cbTaskRead = 0;
4770
4771 /* Synchronous I/O contexts only have one buffer segment. */
4772 AssertMsgReturn(pIoCtx->Req.Io.SgBuf.cSegs == 1,
4773 ("Invalid number of buffer segments for synchronous I/O context"),
4774 VERR_INVALID_PARAMETER);
4775
4776 cbTaskRead = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, &Seg, &cSegments, cbRead);
4777 Assert(cbRead == cbTaskRead);
4778 Assert(cSegments == 1);
4779 rc = pVDIo->pInterfaceIo->pfnReadSync(pVDIo->pInterfaceIo->Core.pvUser,
4780 pIoStorage->pStorage, uOffset,
4781 Seg.pvSeg, cbRead, NULL);
4782 if (RT_SUCCESS(rc))
4783 {
4784 Assert(cbRead == (uint32_t)cbRead);
4785 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbRead);
4786 }
4787 }
4788 else
4789 {
4790 /* Build the S/G array and spawn a new I/O task */
4791 while (cbRead)
4792 {
4793 RTSGSEG aSeg[VD_IO_TASK_SEGMENTS_MAX];
4794 unsigned cSegments = VD_IO_TASK_SEGMENTS_MAX;
4795 size_t cbTaskRead = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, aSeg, &cSegments, cbRead);
4796
4797 Assert(cSegments > 0);
4798 Assert(cbTaskRead > 0);
4799 AssertMsg(cbTaskRead <= cbRead, ("Invalid number of bytes to read\n"));
4800
4801 LogFlow(("Reading %u bytes into %u segments\n", cbTaskRead, cSegments));
4802
4803#ifdef RT_STRICT
4804 for (unsigned i = 0; i < cSegments; i++)
4805 AssertMsg(aSeg[i].pvSeg && !(aSeg[i].cbSeg % 512),
4806 ("Segment %u is invalid\n", i));
4807#endif
4808
4809 Assert(cbTaskRead == (uint32_t)cbTaskRead);
4810 PVDIOTASK pIoTask = vdIoTaskUserAlloc(pIoStorage, NULL, NULL, pIoCtx, (uint32_t)cbTaskRead);
4811
4812 if (!pIoTask)
4813 return VERR_NO_MEMORY;
4814
4815 ASMAtomicIncU32(&pIoCtx->cDataTransfersPending);
4816
4817 void *pvTask;
4818 Log(("Spawning pIoTask=%p pIoCtx=%p\n", pIoTask, pIoCtx));
4819 rc = pVDIo->pInterfaceIo->pfnReadAsync(pVDIo->pInterfaceIo->Core.pvUser,
4820 pIoStorage->pStorage, uOffset,
4821 aSeg, cSegments, cbTaskRead, pIoTask,
4822 &pvTask);
4823 if (RT_SUCCESS(rc))
4824 {
4825 AssertMsg(cbTaskRead <= pIoCtx->Req.Io.cbTransferLeft, ("Impossible!\n"));
4826 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbTaskRead);
4827 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
4828 vdIoTaskFree(pDisk, pIoTask);
4829 }
4830 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
4831 {
4832 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
4833 vdIoTaskFree(pDisk, pIoTask);
4834 break;
4835 }
4836
4837 uOffset += cbTaskRead;
4838 cbRead -= cbTaskRead;
4839 }
4840 }
4841
4842 LogFlowFunc(("returns rc=%Rrc\n", rc));
4843 return rc;
4844}
4845
4846static DECLCALLBACK(int) vdIOIntWriteUser(void *pvUser, PVDIOSTORAGE pIoStorage, uint64_t uOffset,
4847 PVDIOCTX pIoCtx, size_t cbWrite, PFNVDXFERCOMPLETED pfnComplete,
4848 void *pvCompleteUser)
4849{
4850 int rc = VINF_SUCCESS;
4851 PVDIO pVDIo = (PVDIO)pvUser;
4852 PVBOXHDD pDisk = pVDIo->pDisk;
4853
4854 LogFlowFunc(("pvUser=%#p pIoStorage=%#p uOffset=%llu pIoCtx=%#p cbWrite=%u\n",
4855 pvUser, pIoStorage, uOffset, pIoCtx, cbWrite));
4856
4857 /** @todo Enable check for sync I/O later. */
4858 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4859 VD_IS_LOCKED(pDisk);
4860
4861 Assert(cbWrite > 0);
4862
4863 if (pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
4864 {
4865 RTSGSEG Seg;
4866 unsigned cSegments = 1;
4867 size_t cbTaskWrite = 0;
4868
4869 /* Synchronous I/O contexts only have one buffer segment. */
4870 AssertMsgReturn(pIoCtx->Req.Io.SgBuf.cSegs == 1,
4871 ("Invalid number of buffer segments for synchronous I/O context"),
4872 VERR_INVALID_PARAMETER);
4873
4874 cbTaskWrite = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, &Seg, &cSegments, cbWrite);
4875 Assert(cbWrite == cbTaskWrite);
4876 Assert(cSegments == 1);
4877 rc = pVDIo->pInterfaceIo->pfnWriteSync(pVDIo->pInterfaceIo->Core.pvUser,
4878 pIoStorage->pStorage, uOffset,
4879 Seg.pvSeg, cbWrite, NULL);
4880 if (RT_SUCCESS(rc))
4881 {
4882 Assert(pIoCtx->Req.Io.cbTransferLeft >= cbWrite);
4883 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbWrite);
4884 }
4885 }
4886 else
4887 {
4888 /* Build the S/G array and spawn a new I/O task */
4889 while (cbWrite)
4890 {
4891 RTSGSEG aSeg[VD_IO_TASK_SEGMENTS_MAX];
4892 unsigned cSegments = VD_IO_TASK_SEGMENTS_MAX;
4893 size_t cbTaskWrite = 0;
4894
4895 cbTaskWrite = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, aSeg, &cSegments, cbWrite);
4896
4897 Assert(cSegments > 0);
4898 Assert(cbTaskWrite > 0);
4899 AssertMsg(cbTaskWrite <= cbWrite, ("Invalid number of bytes to write\n"));
4900
4901 LogFlow(("Writing %u bytes from %u segments\n", cbTaskWrite, cSegments));
4902
4903#ifdef DEBUG
4904 for (unsigned i = 0; i < cSegments; i++)
4905 AssertMsg(aSeg[i].pvSeg && !(aSeg[i].cbSeg % 512),
4906 ("Segment %u is invalid\n", i));
4907#endif
4908
4909 Assert(cbTaskWrite == (uint32_t)cbTaskWrite);
4910 PVDIOTASK pIoTask = vdIoTaskUserAlloc(pIoStorage, pfnComplete, pvCompleteUser, pIoCtx, (uint32_t)cbTaskWrite);
4911
4912 if (!pIoTask)
4913 return VERR_NO_MEMORY;
4914
4915 ASMAtomicIncU32(&pIoCtx->cDataTransfersPending);
4916
4917 void *pvTask;
4918 Log(("Spawning pIoTask=%p pIoCtx=%p\n", pIoTask, pIoCtx));
4919 rc = pVDIo->pInterfaceIo->pfnWriteAsync(pVDIo->pInterfaceIo->Core.pvUser,
4920 pIoStorage->pStorage,
4921 uOffset, aSeg, cSegments,
4922 cbTaskWrite, pIoTask, &pvTask);
4923 if (RT_SUCCESS(rc))
4924 {
4925 AssertMsg(cbTaskWrite <= pIoCtx->Req.Io.cbTransferLeft, ("Impossible!\n"));
4926 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbTaskWrite);
4927 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
4928 vdIoTaskFree(pDisk, pIoTask);
4929 }
4930 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
4931 {
4932 ASMAtomicDecU32(&pIoCtx->cDataTransfersPending);
4933 vdIoTaskFree(pDisk, pIoTask);
4934 break;
4935 }
4936
4937 uOffset += cbTaskWrite;
4938 cbWrite -= cbTaskWrite;
4939 }
4940 }
4941
4942 LogFlowFunc(("returns rc=%Rrc\n", rc));
4943 return rc;
4944}
4945
4946static DECLCALLBACK(int) vdIOIntReadMeta(void *pvUser, PVDIOSTORAGE pIoStorage, uint64_t uOffset,
4947 void *pvBuf, size_t cbRead, PVDIOCTX pIoCtx,
4948 PPVDMETAXFER ppMetaXfer, PFNVDXFERCOMPLETED pfnComplete,
4949 void *pvCompleteUser)
4950{
4951 PVDIO pVDIo = (PVDIO)pvUser;
4952 PVBOXHDD pDisk = pVDIo->pDisk;
4953 int rc = VINF_SUCCESS;
4954 RTSGSEG Seg;
4955 PVDIOTASK pIoTask;
4956 PVDMETAXFER pMetaXfer = NULL;
4957 void *pvTask = NULL;
4958
4959 LogFlowFunc(("pvUser=%#p pIoStorage=%#p uOffset=%llu pvBuf=%#p cbRead=%u\n",
4960 pvUser, pIoStorage, uOffset, pvBuf, cbRead));
4961
4962 AssertMsgReturn( pIoCtx
4963 || (!ppMetaXfer && !pfnComplete && !pvCompleteUser),
4964 ("A synchronous metadata read is requested but the parameters are wrong\n"),
4965 VERR_INVALID_POINTER);
4966
4967 /** @todo Enable check for sync I/O later. */
4968 if ( pIoCtx
4969 && !(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
4970 VD_IS_LOCKED(pDisk);
4971
4972 if ( !pIoCtx
4973 || pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
4974 {
4975 /* Handle synchronous metadata I/O. */
4976 /** @todo Integrate with metadata transfers below. */
4977 rc = pVDIo->pInterfaceIo->pfnReadSync(pVDIo->pInterfaceIo->Core.pvUser,
4978 pIoStorage->pStorage, uOffset,
4979 pvBuf, cbRead, NULL);
4980 if (ppMetaXfer)
4981 *ppMetaXfer = NULL;
4982 }
4983 else
4984 {
4985 pMetaXfer = (PVDMETAXFER)RTAvlrFileOffsetGet(pIoStorage->pTreeMetaXfers, uOffset);
4986 if (!pMetaXfer)
4987 {
4988#ifdef RT_STRICT
4989 pMetaXfer = (PVDMETAXFER)RTAvlrFileOffsetGetBestFit(pIoStorage->pTreeMetaXfers, uOffset, false /* fAbove */);
4990 AssertMsg(!pMetaXfer || (pMetaXfer->Core.Key + (RTFOFF)pMetaXfer->cbMeta <= (RTFOFF)uOffset),
4991 ("Overlapping meta transfers!\n"));
4992#endif
4993
4994 /* Allocate a new meta transfer. */
4995 pMetaXfer = vdMetaXferAlloc(pIoStorage, uOffset, cbRead);
4996 if (!pMetaXfer)
4997 return VERR_NO_MEMORY;
4998
4999 pIoTask = vdIoTaskMetaAlloc(pIoStorage, pfnComplete, pvCompleteUser, pMetaXfer);
5000 if (!pIoTask)
5001 {
5002 RTMemFree(pMetaXfer);
5003 return VERR_NO_MEMORY;
5004 }
5005
5006 Seg.cbSeg = cbRead;
5007 Seg.pvSeg = pMetaXfer->abData;
5008
5009 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_READ);
5010 rc = pVDIo->pInterfaceIo->pfnReadAsync(pVDIo->pInterfaceIo->Core.pvUser,
5011 pIoStorage->pStorage,
5012 uOffset, &Seg, 1,
5013 cbRead, pIoTask, &pvTask);
5014
5015 if (RT_SUCCESS(rc) || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
5016 {
5017 bool fInserted = RTAvlrFileOffsetInsert(pIoStorage->pTreeMetaXfers, &pMetaXfer->Core);
5018 Assert(fInserted); NOREF(fInserted);
5019 }
5020 else
5021 RTMemFree(pMetaXfer);
5022
5023 if (RT_SUCCESS(rc))
5024 {
5025 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_NONE);
5026 vdIoTaskFree(pDisk, pIoTask);
5027 }
5028 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS && !pfnComplete)
5029 rc = VERR_VD_NOT_ENOUGH_METADATA;
5030 }
5031
5032 Assert(VALID_PTR(pMetaXfer) || RT_FAILURE(rc));
5033
5034 if (RT_SUCCESS(rc) || rc == VERR_VD_NOT_ENOUGH_METADATA || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
5035 {
5036 /* If it is pending add the request to the list. */
5037 if (VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_READ)
5038 {
5039 PVDIOCTXDEFERRED pDeferred = (PVDIOCTXDEFERRED)RTMemAllocZ(sizeof(VDIOCTXDEFERRED));
5040 AssertPtr(pDeferred);
5041
5042 RTListInit(&pDeferred->NodeDeferred);
5043 pDeferred->pIoCtx = pIoCtx;
5044
5045 ASMAtomicIncU32(&pIoCtx->cMetaTransfersPending);
5046 RTListAppend(&pMetaXfer->ListIoCtxWaiting, &pDeferred->NodeDeferred);
5047 rc = VERR_VD_NOT_ENOUGH_METADATA;
5048 }
5049 else
5050 {
5051 /* Transfer the data. */
5052 pMetaXfer->cRefs++;
5053 Assert(pMetaXfer->cbMeta >= cbRead);
5054 Assert(pMetaXfer->Core.Key == (RTFOFF)uOffset);
5055 if (pMetaXfer->pbDataShw)
5056 memcpy(pvBuf, pMetaXfer->pbDataShw, cbRead);
5057 else
5058 memcpy(pvBuf, pMetaXfer->abData, cbRead);
5059 *ppMetaXfer = pMetaXfer;
5060 }
5061 }
5062 }
5063
5064 LogFlowFunc(("returns rc=%Rrc\n", rc));
5065 return rc;
5066}
5067
5068static DECLCALLBACK(int) vdIOIntWriteMeta(void *pvUser, PVDIOSTORAGE pIoStorage, uint64_t uOffset,
5069 const void *pvBuf, size_t cbWrite, PVDIOCTX pIoCtx,
5070 PFNVDXFERCOMPLETED pfnComplete, void *pvCompleteUser)
5071{
5072 PVDIO pVDIo = (PVDIO)pvUser;
5073 PVBOXHDD pDisk = pVDIo->pDisk;
5074 int rc = VINF_SUCCESS;
5075 RTSGSEG Seg;
5076 PVDIOTASK pIoTask;
5077 PVDMETAXFER pMetaXfer = NULL;
5078 bool fInTree = false;
5079 void *pvTask = NULL;
5080
5081 LogFlowFunc(("pvUser=%#p pIoStorage=%#p uOffset=%llu pvBuf=%#p cbWrite=%u\n",
5082 pvUser, pIoStorage, uOffset, pvBuf, cbWrite));
5083
5084 AssertMsgReturn( pIoCtx
5085 || (!pfnComplete && !pvCompleteUser),
5086 ("A synchronous metadata write is requested but the parameters are wrong\n"),
5087 VERR_INVALID_POINTER);
5088
5089 /** @todo Enable check for sync I/O later. */
5090 if ( pIoCtx
5091 && !(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
5092 VD_IS_LOCKED(pDisk);
5093
5094 if ( !pIoCtx
5095 || pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
5096 {
5097 /* Handle synchronous metadata I/O. */
5098 /** @todo Integrate with metadata transfers below. */
5099 rc = pVDIo->pInterfaceIo->pfnWriteSync(pVDIo->pInterfaceIo->Core.pvUser,
5100 pIoStorage->pStorage, uOffset,
5101 pvBuf, cbWrite, NULL);
5102 }
5103 else
5104 {
5105 pMetaXfer = (PVDMETAXFER)RTAvlrFileOffsetGet(pIoStorage->pTreeMetaXfers, uOffset);
5106 if (!pMetaXfer)
5107 {
5108 /* Allocate a new meta transfer. */
5109 pMetaXfer = vdMetaXferAlloc(pIoStorage, uOffset, cbWrite);
5110 if (!pMetaXfer)
5111 return VERR_NO_MEMORY;
5112 }
5113 else
5114 {
5115 Assert(pMetaXfer->cbMeta >= cbWrite);
5116 Assert(pMetaXfer->Core.Key == (RTFOFF)uOffset);
5117 fInTree = true;
5118 }
5119
5120 if (VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_NONE)
5121 {
5122 pIoTask = vdIoTaskMetaAlloc(pIoStorage, pfnComplete, pvCompleteUser, pMetaXfer);
5123 if (!pIoTask)
5124 {
5125 RTMemFree(pMetaXfer);
5126 return VERR_NO_MEMORY;
5127 }
5128
5129 memcpy(pMetaXfer->abData, pvBuf, cbWrite);
5130 Seg.cbSeg = cbWrite;
5131 Seg.pvSeg = pMetaXfer->abData;
5132
5133 ASMAtomicIncU32(&pIoCtx->cMetaTransfersPending);
5134
5135 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_WRITE);
5136 rc = pVDIo->pInterfaceIo->pfnWriteAsync(pVDIo->pInterfaceIo->Core.pvUser,
5137 pIoStorage->pStorage,
5138 uOffset, &Seg, 1, cbWrite, pIoTask,
5139 &pvTask);
5140 if (RT_SUCCESS(rc))
5141 {
5142 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_NONE);
5143 ASMAtomicDecU32(&pIoCtx->cMetaTransfersPending);
5144 vdIoTaskFree(pDisk, pIoTask);
5145 if (fInTree && !pMetaXfer->cRefs)
5146 {
5147 LogFlow(("Removing meta xfer=%#p\n", pMetaXfer));
5148 bool fRemoved = RTAvlrFileOffsetRemove(pIoStorage->pTreeMetaXfers, pMetaXfer->Core.Key) != NULL;
5149 AssertMsg(fRemoved, ("Metadata transfer wasn't removed\n")); NOREF(fRemoved);
5150 RTMemFree(pMetaXfer);
5151 pMetaXfer = NULL;
5152 }
5153 }
5154 else if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
5155 {
5156 PVDIOCTXDEFERRED pDeferred = (PVDIOCTXDEFERRED)RTMemAllocZ(sizeof(VDIOCTXDEFERRED));
5157 AssertPtr(pDeferred);
5158
5159 RTListInit(&pDeferred->NodeDeferred);
5160 pDeferred->pIoCtx = pIoCtx;
5161
5162 if (!fInTree)
5163 {
5164 bool fInserted = RTAvlrFileOffsetInsert(pIoStorage->pTreeMetaXfers, &pMetaXfer->Core);
5165 Assert(fInserted); NOREF(fInserted);
5166 }
5167
5168 RTListAppend(&pMetaXfer->ListIoCtxWaiting, &pDeferred->NodeDeferred);
5169 }
5170 else
5171 {
5172 RTMemFree(pMetaXfer);
5173 pMetaXfer = NULL;
5174 }
5175 }
5176 else
5177 {
5178 /* I/O is in progress, update shadow buffer and add to waiting list. */
5179 Assert(VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_WRITE);
5180 if (!pMetaXfer->pbDataShw)
5181 {
5182 /* Allocate shadow buffer and set initial state. */
5183 LogFlowFunc(("pMetaXfer=%#p Creating shadow buffer\n", pMetaXfer));
5184 pMetaXfer->pbDataShw = (uint8_t *)RTMemAlloc(pMetaXfer->cbMeta);
5185 if (RT_LIKELY(pMetaXfer->pbDataShw))
5186 memcpy(pMetaXfer->pbDataShw, pMetaXfer->abData, pMetaXfer->cbMeta);
5187 else
5188 rc = VERR_NO_MEMORY;
5189 }
5190
5191 if (RT_SUCCESS(rc))
5192 {
5193 /* Update with written data and append to waiting list. */
5194 PVDIOCTXDEFERRED pDeferred = (PVDIOCTXDEFERRED)RTMemAllocZ(sizeof(VDIOCTXDEFERRED));
5195 if (pDeferred)
5196 {
5197 LogFlowFunc(("pMetaXfer=%#p Updating shadow buffer\n", pMetaXfer));
5198
5199 RTListInit(&pDeferred->NodeDeferred);
5200 pDeferred->pIoCtx = pIoCtx;
5201 ASMAtomicIncU32(&pIoCtx->cMetaTransfersPending);
5202 memcpy(pMetaXfer->pbDataShw, pvBuf, cbWrite);
5203 RTListAppend(&pMetaXfer->ListIoCtxShwWrites, &pDeferred->NodeDeferred);
5204 }
5205 else
5206 {
5207 /*
5208 * Free shadow buffer if there is no one depending on it, i.e.
5209 * we just allocated it.
5210 */
5211 if (RTListIsEmpty(&pMetaXfer->ListIoCtxShwWrites))
5212 {
5213 RTMemFree(pMetaXfer->pbDataShw);
5214 pMetaXfer->pbDataShw = NULL;
5215 }
5216 rc = VERR_NO_MEMORY;
5217 }
5218 }
5219 }
5220 }
5221
5222 LogFlowFunc(("returns rc=%Rrc\n", rc));
5223 return rc;
5224}
5225
5226static DECLCALLBACK(void) vdIOIntMetaXferRelease(void *pvUser, PVDMETAXFER pMetaXfer)
5227{
5228 PVDIO pVDIo = (PVDIO)pvUser;
5229 PVBOXHDD pDisk = pVDIo->pDisk;
5230 PVDIOSTORAGE pIoStorage;
5231
5232 /*
5233 * It is possible that we get called with a NULL metadata xfer handle
5234 * for synchronous I/O. Just exit.
5235 */
5236 if (!pMetaXfer)
5237 return;
5238
5239 pIoStorage = pMetaXfer->pIoStorage;
5240
5241 VD_IS_LOCKED(pDisk);
5242
5243 Assert( VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_NONE
5244 || VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_WRITE);
5245 Assert(pMetaXfer->cRefs > 0);
5246
5247 pMetaXfer->cRefs--;
5248 if ( !pMetaXfer->cRefs
5249 && RTListIsEmpty(&pMetaXfer->ListIoCtxWaiting)
5250 && VDMETAXFER_TXDIR_GET(pMetaXfer->fFlags) == VDMETAXFER_TXDIR_NONE)
5251 {
5252 /* Free the meta data entry. */
5253 LogFlow(("Removing meta xfer=%#p\n", pMetaXfer));
5254 bool fRemoved = RTAvlrFileOffsetRemove(pIoStorage->pTreeMetaXfers, pMetaXfer->Core.Key) != NULL;
5255 AssertMsg(fRemoved, ("Metadata transfer wasn't removed\n")); NOREF(fRemoved);
5256
5257 RTMemFree(pMetaXfer);
5258 }
5259}
5260
5261static DECLCALLBACK(int) vdIOIntFlush(void *pvUser, PVDIOSTORAGE pIoStorage, PVDIOCTX pIoCtx,
5262 PFNVDXFERCOMPLETED pfnComplete, void *pvCompleteUser)
5263{
5264 PVDIO pVDIo = (PVDIO)pvUser;
5265 PVBOXHDD pDisk = pVDIo->pDisk;
5266 int rc = VINF_SUCCESS;
5267 PVDIOTASK pIoTask;
5268 PVDMETAXFER pMetaXfer = NULL;
5269 void *pvTask = NULL;
5270
5271 LogFlowFunc(("pvUser=%#p pIoStorage=%#p pIoCtx=%#p\n",
5272 pvUser, pIoStorage, pIoCtx));
5273
5274 AssertMsgReturn( pIoCtx
5275 || (!pfnComplete && !pvCompleteUser),
5276 ("A synchronous metadata write is requested but the parameters are wrong\n"),
5277 VERR_INVALID_POINTER);
5278
5279 /** @todo Enable check for sync I/O later. */
5280 if ( pIoCtx
5281 && !(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
5282 VD_IS_LOCKED(pDisk);
5283
5284 if (pVDIo->fIgnoreFlush)
5285 return VINF_SUCCESS;
5286
5287 if ( !pIoCtx
5288 || pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC)
5289 {
5290 /* Handle synchronous flushes. */
5291 /** @todo Integrate with metadata transfers below. */
5292 rc = pVDIo->pInterfaceIo->pfnFlushSync(pVDIo->pInterfaceIo->Core.pvUser,
5293 pIoStorage->pStorage);
5294 }
5295 else
5296 {
5297 /* Allocate a new meta transfer. */
5298 pMetaXfer = vdMetaXferAlloc(pIoStorage, 0, 0);
5299 if (!pMetaXfer)
5300 return VERR_NO_MEMORY;
5301
5302 pIoTask = vdIoTaskMetaAlloc(pIoStorage, pfnComplete, pvUser, pMetaXfer);
5303 if (!pIoTask)
5304 {
5305 RTMemFree(pMetaXfer);
5306 return VERR_NO_MEMORY;
5307 }
5308
5309 ASMAtomicIncU32(&pIoCtx->cMetaTransfersPending);
5310
5311 PVDIOCTXDEFERRED pDeferred = (PVDIOCTXDEFERRED)RTMemAllocZ(sizeof(VDIOCTXDEFERRED));
5312 AssertPtr(pDeferred);
5313
5314 RTListInit(&pDeferred->NodeDeferred);
5315 pDeferred->pIoCtx = pIoCtx;
5316
5317 RTListAppend(&pMetaXfer->ListIoCtxWaiting, &pDeferred->NodeDeferred);
5318 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_FLUSH);
5319 rc = pVDIo->pInterfaceIo->pfnFlushAsync(pVDIo->pInterfaceIo->Core.pvUser,
5320 pIoStorage->pStorage,
5321 pIoTask, &pvTask);
5322 if (RT_SUCCESS(rc))
5323 {
5324 VDMETAXFER_TXDIR_SET(pMetaXfer->fFlags, VDMETAXFER_TXDIR_NONE);
5325 ASMAtomicDecU32(&pIoCtx->cMetaTransfersPending);
5326 vdIoTaskFree(pDisk, pIoTask);
5327 RTMemFree(pDeferred);
5328 RTMemFree(pMetaXfer);
5329 }
5330 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
5331 RTMemFree(pMetaXfer);
5332 }
5333
5334 LogFlowFunc(("returns rc=%Rrc\n", rc));
5335 return rc;
5336}
5337
5338static DECLCALLBACK(size_t) vdIOIntIoCtxCopyTo(void *pvUser, PVDIOCTX pIoCtx,
5339 const void *pvBuf, size_t cbBuf)
5340{
5341 PVDIO pVDIo = (PVDIO)pvUser;
5342 PVBOXHDD pDisk = pVDIo->pDisk;
5343 size_t cbCopied = 0;
5344
5345 /** @todo Enable check for sync I/O later. */
5346 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
5347 VD_IS_LOCKED(pDisk);
5348
5349 cbCopied = vdIoCtxCopyTo(pIoCtx, (uint8_t *)pvBuf, cbBuf);
5350 Assert(cbCopied == cbBuf);
5351
5352 /// @todo Assert(pIoCtx->Req.Io.cbTransferLeft >= cbCopied); - triggers with vdCopyHelper/dmgRead.
5353 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbCopied);
5354
5355 return cbCopied;
5356}
5357
5358static DECLCALLBACK(size_t) vdIOIntIoCtxCopyFrom(void *pvUser, PVDIOCTX pIoCtx,
5359 void *pvBuf, size_t cbBuf)
5360{
5361 PVDIO pVDIo = (PVDIO)pvUser;
5362 PVBOXHDD pDisk = pVDIo->pDisk;
5363 size_t cbCopied = 0;
5364
5365 /** @todo Enable check for sync I/O later. */
5366 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
5367 VD_IS_LOCKED(pDisk);
5368
5369 cbCopied = vdIoCtxCopyFrom(pIoCtx, (uint8_t *)pvBuf, cbBuf);
5370 Assert(cbCopied == cbBuf);
5371
5372 /// @todo Assert(pIoCtx->Req.Io.cbTransferLeft > cbCopied); - triggers with vdCopyHelper/dmgRead.
5373 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbCopied);
5374
5375 return cbCopied;
5376}
5377
5378static DECLCALLBACK(size_t) vdIOIntIoCtxSet(void *pvUser, PVDIOCTX pIoCtx, int ch, size_t cb)
5379{
5380 PVDIO pVDIo = (PVDIO)pvUser;
5381 PVBOXHDD pDisk = pVDIo->pDisk;
5382 size_t cbSet = 0;
5383
5384 /** @todo Enable check for sync I/O later. */
5385 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
5386 VD_IS_LOCKED(pDisk);
5387
5388 cbSet = vdIoCtxSet(pIoCtx, ch, cb);
5389 Assert(cbSet == cb);
5390
5391 /// @todo Assert(pIoCtx->Req.Io.cbTransferLeft >= cbSet); - triggers with vdCopyHelper/dmgRead.
5392 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbSet);
5393
5394 return cbSet;
5395}
5396
5397static DECLCALLBACK(size_t) vdIOIntIoCtxSegArrayCreate(void *pvUser, PVDIOCTX pIoCtx,
5398 PRTSGSEG paSeg, unsigned *pcSeg,
5399 size_t cbData)
5400{
5401 PVDIO pVDIo = (PVDIO)pvUser;
5402 PVBOXHDD pDisk = pVDIo->pDisk;
5403 size_t cbCreated = 0;
5404
5405 /** @todo It is possible that this gets called from a filter plugin
5406 * outside of the disk lock. Refine assertion or remove completely. */
5407#if 0
5408 /** @todo Enable check for sync I/O later. */
5409 if (!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC))
5410 VD_IS_LOCKED(pDisk);
5411#else
5412 NOREF(pDisk);
5413#endif
5414
5415 cbCreated = RTSgBufSegArrayCreate(&pIoCtx->Req.Io.SgBuf, paSeg, pcSeg, cbData);
5416 Assert(!paSeg || cbData == cbCreated);
5417
5418 return cbCreated;
5419}
5420
5421static DECLCALLBACK(void) vdIOIntIoCtxCompleted(void *pvUser, PVDIOCTX pIoCtx, int rcReq,
5422 size_t cbCompleted)
5423{
5424 PVDIO pVDIo = (PVDIO)pvUser;
5425 PVBOXHDD pDisk = pVDIo->pDisk;
5426
5427 LogFlowFunc(("pvUser=%#p pIoCtx=%#p rcReq=%Rrc cbCompleted=%zu\n",
5428 pvUser, pIoCtx, rcReq, cbCompleted));
5429
5430 /*
5431 * Grab the disk critical section to avoid races with other threads which
5432 * might still modify the I/O context.
5433 * Example is that iSCSI is doing an asynchronous write but calls us already
5434 * while the other thread is still hanging in vdWriteHelperAsync and couldn't update
5435 * the blocked state yet.
5436 * It can overwrite the state to true before we call vdIoCtxContinue and the
5437 * the request would hang indefinite.
5438 */
5439 ASMAtomicCmpXchgS32(&pIoCtx->rcReq, rcReq, VINF_SUCCESS);
5440 Assert(pIoCtx->Req.Io.cbTransferLeft >= cbCompleted);
5441 ASMAtomicSubU32(&pIoCtx->Req.Io.cbTransferLeft, (uint32_t)cbCompleted);
5442
5443 /* Set next transfer function if the current one finished.
5444 * @todo: Find a better way to prevent vdIoCtxContinue from calling the current helper again. */
5445 if (!pIoCtx->Req.Io.cbTransferLeft)
5446 {
5447 pIoCtx->pfnIoCtxTransfer = pIoCtx->pfnIoCtxTransferNext;
5448 pIoCtx->pfnIoCtxTransferNext = NULL;
5449 }
5450
5451 vdIoCtxAddToWaitingList(&pDisk->pIoCtxHaltedHead, pIoCtx);
5452 if (ASMAtomicCmpXchgBool(&pDisk->fLocked, true, false))
5453 {
5454 /* Immediately drop the lock again, it will take care of processing the list. */
5455 vdDiskUnlock(pDisk, NULL);
5456 }
5457}
5458
5459static DECLCALLBACK(bool) vdIOIntIoCtxIsSynchronous(void *pvUser, PVDIOCTX pIoCtx)
5460{
5461 NOREF(pvUser);
5462 return !!(pIoCtx->fFlags & VDIOCTX_FLAGS_SYNC);
5463}
5464
5465static DECLCALLBACK(bool) vdIOIntIoCtxIsZero(void *pvUser, PVDIOCTX pIoCtx, size_t cbCheck,
5466 bool fAdvance)
5467{
5468 NOREF(pvUser);
5469
5470 bool fIsZero = RTSgBufIsZero(&pIoCtx->Req.Io.SgBuf, cbCheck);
5471 if (fIsZero && fAdvance)
5472 RTSgBufAdvance(&pIoCtx->Req.Io.SgBuf, cbCheck);
5473
5474 return fIsZero;
5475}
5476
5477static DECLCALLBACK(size_t) vdIOIntIoCtxGetDataUnitSize(void *pvUser, PVDIOCTX pIoCtx)
5478{
5479 RT_NOREF1(pIoCtx);
5480 PVDIO pVDIo = (PVDIO)pvUser;
5481 PVBOXHDD pDisk = pVDIo->pDisk;
5482
5483 PVDIMAGE pImage = vdGetImageByNumber(pDisk, VD_LAST_IMAGE);
5484 AssertPtrReturn(pImage, 0);
5485 return pImage->Backend->pfnGetSectorSize(pImage->pBackendData);
5486}
5487
5488/**
5489 * VD I/O interface callback for opening a file (limited version for VDGetFormat).
5490 */
5491static DECLCALLBACK(int) vdIOIntOpenLimited(void *pvUser, const char *pszLocation,
5492 uint32_t fOpen, PPVDIOSTORAGE ppIoStorage)
5493{
5494 int rc = VINF_SUCCESS;
5495 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5496 PVDIOSTORAGE pIoStorage = (PVDIOSTORAGE)RTMemAllocZ(sizeof(VDIOSTORAGE));
5497
5498 if (!pIoStorage)
5499 return VERR_NO_MEMORY;
5500
5501 rc = pInterfaceIo->pfnOpen(NULL, pszLocation, fOpen, NULL, &pIoStorage->pStorage);
5502 if (RT_SUCCESS(rc))
5503 *ppIoStorage = pIoStorage;
5504 else
5505 RTMemFree(pIoStorage);
5506
5507 return rc;
5508}
5509
5510static DECLCALLBACK(int) vdIOIntCloseLimited(void *pvUser, PVDIOSTORAGE pIoStorage)
5511{
5512 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5513 int rc = pInterfaceIo->pfnClose(NULL, pIoStorage->pStorage);
5514
5515 RTMemFree(pIoStorage);
5516 return rc;
5517}
5518
5519static DECLCALLBACK(int) vdIOIntDeleteLimited(void *pvUser, const char *pcszFilename)
5520{
5521 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5522 return pInterfaceIo->pfnDelete(NULL, pcszFilename);
5523}
5524
5525static DECLCALLBACK(int) vdIOIntMoveLimited(void *pvUser, const char *pcszSrc,
5526 const char *pcszDst, unsigned fMove)
5527{
5528 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5529 return pInterfaceIo->pfnMove(NULL, pcszSrc, pcszDst, fMove);
5530}
5531
5532static DECLCALLBACK(int) vdIOIntGetFreeSpaceLimited(void *pvUser, const char *pcszFilename,
5533 int64_t *pcbFreeSpace)
5534{
5535 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5536 return pInterfaceIo->pfnGetFreeSpace(NULL, pcszFilename, pcbFreeSpace);
5537}
5538
5539static DECLCALLBACK(int) vdIOIntGetModificationTimeLimited(void *pvUser,
5540 const char *pcszFilename,
5541 PRTTIMESPEC pModificationTime)
5542{
5543 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5544 return pInterfaceIo->pfnGetModificationTime(NULL, pcszFilename, pModificationTime);
5545}
5546
5547static DECLCALLBACK(int) vdIOIntGetSizeLimited(void *pvUser, PVDIOSTORAGE pIoStorage,
5548 uint64_t *pcbSize)
5549{
5550 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5551 return pInterfaceIo->pfnGetSize(NULL, pIoStorage->pStorage, pcbSize);
5552}
5553
5554static DECLCALLBACK(int) vdIOIntSetSizeLimited(void *pvUser, PVDIOSTORAGE pIoStorage,
5555 uint64_t cbSize)
5556{
5557 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5558 return pInterfaceIo->pfnSetSize(NULL, pIoStorage->pStorage, cbSize);
5559}
5560
5561static DECLCALLBACK(int) vdIOIntWriteUserLimited(void *pvUser, PVDIOSTORAGE pStorage,
5562 uint64_t uOffset, PVDIOCTX pIoCtx,
5563 size_t cbWrite,
5564 PFNVDXFERCOMPLETED pfnComplete,
5565 void *pvCompleteUser)
5566{
5567 NOREF(pvUser);
5568 NOREF(pStorage);
5569 NOREF(uOffset);
5570 NOREF(pIoCtx);
5571 NOREF(cbWrite);
5572 NOREF(pfnComplete);
5573 NOREF(pvCompleteUser);
5574 AssertMsgFailedReturn(("This needs to be implemented when called\n"), VERR_NOT_IMPLEMENTED);
5575}
5576
5577static DECLCALLBACK(int) vdIOIntReadUserLimited(void *pvUser, PVDIOSTORAGE pStorage,
5578 uint64_t uOffset, PVDIOCTX pIoCtx,
5579 size_t cbRead)
5580{
5581 NOREF(pvUser);
5582 NOREF(pStorage);
5583 NOREF(uOffset);
5584 NOREF(pIoCtx);
5585 NOREF(cbRead);
5586 AssertMsgFailedReturn(("This needs to be implemented when called\n"), VERR_NOT_IMPLEMENTED);
5587}
5588
5589static DECLCALLBACK(int) vdIOIntWriteMetaLimited(void *pvUser, PVDIOSTORAGE pStorage,
5590 uint64_t uOffset, const void *pvBuffer,
5591 size_t cbBuffer, PVDIOCTX pIoCtx,
5592 PFNVDXFERCOMPLETED pfnComplete,
5593 void *pvCompleteUser)
5594{
5595 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5596
5597 AssertMsgReturn(!pIoCtx && !pfnComplete && !pvCompleteUser,
5598 ("Async I/O not implemented for the limited interface"),
5599 VERR_NOT_SUPPORTED);
5600
5601 return pInterfaceIo->pfnWriteSync(NULL, pStorage->pStorage, uOffset, pvBuffer, cbBuffer, NULL);
5602}
5603
5604static DECLCALLBACK(int) vdIOIntReadMetaLimited(void *pvUser, PVDIOSTORAGE pStorage,
5605 uint64_t uOffset, void *pvBuffer,
5606 size_t cbBuffer, PVDIOCTX pIoCtx,
5607 PPVDMETAXFER ppMetaXfer,
5608 PFNVDXFERCOMPLETED pfnComplete,
5609 void *pvCompleteUser)
5610{
5611 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5612
5613 AssertMsgReturn(!pIoCtx && !ppMetaXfer && !pfnComplete && !pvCompleteUser,
5614 ("Async I/O not implemented for the limited interface"),
5615 VERR_NOT_SUPPORTED);
5616
5617 return pInterfaceIo->pfnReadSync(NULL, pStorage->pStorage, uOffset, pvBuffer, cbBuffer, NULL);
5618}
5619
5620#if 0 /* unsed */
5621static int vdIOIntMetaXferReleaseLimited(void *pvUser, PVDMETAXFER pMetaXfer)
5622{
5623 /* This is a NOP in this case. */
5624 NOREF(pvUser);
5625 NOREF(pMetaXfer);
5626 return VINF_SUCCESS;
5627}
5628#endif
5629
5630static DECLCALLBACK(int) vdIOIntFlushLimited(void *pvUser, PVDIOSTORAGE pStorage,
5631 PVDIOCTX pIoCtx,
5632 PFNVDXFERCOMPLETED pfnComplete,
5633 void *pvCompleteUser)
5634{
5635 PVDINTERFACEIO pInterfaceIo = (PVDINTERFACEIO)pvUser;
5636
5637 AssertMsgReturn(!pIoCtx && !pfnComplete && !pvCompleteUser,
5638 ("Async I/O not implemented for the limited interface"),
5639 VERR_NOT_SUPPORTED);
5640
5641 return pInterfaceIo->pfnFlushSync(NULL, pStorage->pStorage);
5642}
5643
5644/**
5645 * internal: send output to the log (unconditionally).
5646 */
5647static DECLCALLBACK(int) vdLogMessage(void *pvUser, const char *pszFormat, va_list args)
5648{
5649 NOREF(pvUser);
5650 RTLogPrintfV(pszFormat, args);
5651 return VINF_SUCCESS;
5652}
5653
5654DECLINLINE(int) vdMessageWrapper(PVBOXHDD pDisk, const char *pszFormat, ...)
5655{
5656 va_list va;
5657 va_start(va, pszFormat);
5658 int rc = pDisk->pInterfaceError->pfnMessage(pDisk->pInterfaceError->Core.pvUser,
5659 pszFormat, va);
5660 va_end(va);
5661 return rc;
5662}
5663
5664
5665/**
5666 * internal: adjust PCHS geometry
5667 */
5668static void vdFixupPCHSGeometry(PVDGEOMETRY pPCHS, uint64_t cbSize)
5669{
5670 /* Fix broken PCHS geometry. Can happen for two reasons: either the backend
5671 * mixes up PCHS and LCHS, or the application used to create the source
5672 * image has put garbage in it. Additionally, if the PCHS geometry covers
5673 * more than the image size, set it back to the default. */
5674 if ( pPCHS->cHeads > 16
5675 || pPCHS->cSectors > 63
5676 || pPCHS->cCylinders == 0
5677 || (uint64_t)pPCHS->cHeads * pPCHS->cSectors * pPCHS->cCylinders * 512 > cbSize)
5678 {
5679 Assert(!(RT_MIN(cbSize / 512 / 16 / 63, 16383) - (uint32_t)RT_MIN(cbSize / 512 / 16 / 63, 16383)));
5680 pPCHS->cCylinders = (uint32_t)RT_MIN(cbSize / 512 / 16 / 63, 16383);
5681 pPCHS->cHeads = 16;
5682 pPCHS->cSectors = 63;
5683 }
5684}
5685
5686/**
5687 * internal: adjust PCHS geometry
5688 */
5689static void vdFixupLCHSGeometry(PVDGEOMETRY pLCHS, uint64_t cbSize)
5690{
5691 /* Fix broken LCHS geometry. Can happen for two reasons: either the backend
5692 * mixes up PCHS and LCHS, or the application used to create the source
5693 * image has put garbage in it. The fix in this case is to clear the LCHS
5694 * geometry to trigger autodetection when it is used next. If the geometry
5695 * already says "please autodetect" (cylinders=0) keep it. */
5696 if ( ( pLCHS->cHeads > 255
5697 || pLCHS->cHeads == 0
5698 || pLCHS->cSectors > 63
5699 || pLCHS->cSectors == 0)
5700 && pLCHS->cCylinders != 0)
5701 {
5702 pLCHS->cCylinders = 0;
5703 pLCHS->cHeads = 0;
5704 pLCHS->cSectors = 0;
5705 }
5706 /* Always recompute the number of cylinders stored in the LCHS
5707 * geometry if it isn't set to "autotedetect" at the moment.
5708 * This is very useful if the destination image size is
5709 * larger or smaller than the source image size. Do not modify
5710 * the number of heads and sectors. Windows guests hate it. */
5711 if ( pLCHS->cCylinders != 0
5712 && pLCHS->cHeads != 0 /* paranoia */
5713 && pLCHS->cSectors != 0 /* paranoia */)
5714 {
5715 Assert(!(RT_MIN(cbSize / 512 / pLCHS->cHeads / pLCHS->cSectors, 1024) - (uint32_t)RT_MIN(cbSize / 512 / pLCHS->cHeads / pLCHS->cSectors, 1024)));
5716 pLCHS->cCylinders = (uint32_t)RT_MIN(cbSize / 512 / pLCHS->cHeads / pLCHS->cSectors, 1024);
5717 }
5718}
5719
5720/**
5721 * Sets the I/O callbacks of the given interface to the fallback methods
5722 *
5723 * @returns nothing.
5724 * @param pIfIo The I/O interface to setup.
5725 */
5726static void vdIfIoFallbackCallbacksSetup(PVDINTERFACEIO pIfIo)
5727{
5728 pIfIo->pfnOpen = vdIOOpenFallback;
5729 pIfIo->pfnClose = vdIOCloseFallback;
5730 pIfIo->pfnDelete = vdIODeleteFallback;
5731 pIfIo->pfnMove = vdIOMoveFallback;
5732 pIfIo->pfnGetFreeSpace = vdIOGetFreeSpaceFallback;
5733 pIfIo->pfnGetModificationTime = vdIOGetModificationTimeFallback;
5734 pIfIo->pfnGetSize = vdIOGetSizeFallback;
5735 pIfIo->pfnSetSize = vdIOSetSizeFallback;
5736 pIfIo->pfnSetAllocationSize = vdIOSetAllocationSizeFallback;
5737 pIfIo->pfnReadSync = vdIOReadSyncFallback;
5738 pIfIo->pfnWriteSync = vdIOWriteSyncFallback;
5739 pIfIo->pfnFlushSync = vdIOFlushSyncFallback;
5740 pIfIo->pfnReadAsync = vdIOReadAsyncFallback;
5741 pIfIo->pfnWriteAsync = vdIOWriteAsyncFallback;
5742 pIfIo->pfnFlushAsync = vdIOFlushAsyncFallback;
5743}
5744
5745/**
5746 * Sets the internal I/O callbacks of the given interface.
5747 *
5748 * @returns nothing.
5749 * @param pIfIoInt The internal I/O interface to setup.
5750 */
5751static void vdIfIoIntCallbacksSetup(PVDINTERFACEIOINT pIfIoInt)
5752{
5753 pIfIoInt->pfnOpen = vdIOIntOpen;
5754 pIfIoInt->pfnClose = vdIOIntClose;
5755 pIfIoInt->pfnDelete = vdIOIntDelete;
5756 pIfIoInt->pfnMove = vdIOIntMove;
5757 pIfIoInt->pfnGetFreeSpace = vdIOIntGetFreeSpace;
5758 pIfIoInt->pfnGetModificationTime = vdIOIntGetModificationTime;
5759 pIfIoInt->pfnGetSize = vdIOIntGetSize;
5760 pIfIoInt->pfnSetSize = vdIOIntSetSize;
5761 pIfIoInt->pfnSetAllocationSize = vdIOIntSetAllocationSize;
5762 pIfIoInt->pfnReadUser = vdIOIntReadUser;
5763 pIfIoInt->pfnWriteUser = vdIOIntWriteUser;
5764 pIfIoInt->pfnReadMeta = vdIOIntReadMeta;
5765 pIfIoInt->pfnWriteMeta = vdIOIntWriteMeta;
5766 pIfIoInt->pfnMetaXferRelease = vdIOIntMetaXferRelease;
5767 pIfIoInt->pfnFlush = vdIOIntFlush;
5768 pIfIoInt->pfnIoCtxCopyFrom = vdIOIntIoCtxCopyFrom;
5769 pIfIoInt->pfnIoCtxCopyTo = vdIOIntIoCtxCopyTo;
5770 pIfIoInt->pfnIoCtxSet = vdIOIntIoCtxSet;
5771 pIfIoInt->pfnIoCtxSegArrayCreate = vdIOIntIoCtxSegArrayCreate;
5772 pIfIoInt->pfnIoCtxCompleted = vdIOIntIoCtxCompleted;
5773 pIfIoInt->pfnIoCtxIsSynchronous = vdIOIntIoCtxIsSynchronous;
5774 pIfIoInt->pfnIoCtxIsZero = vdIOIntIoCtxIsZero;
5775 pIfIoInt->pfnIoCtxGetDataUnitSize = vdIOIntIoCtxGetDataUnitSize;
5776}
5777
5778/**
5779 * Internally used completion handler for synchronous I/O contexts.
5780 */
5781static DECLCALLBACK(void) vdIoCtxSyncComplete(void *pvUser1, void *pvUser2, int rcReq)
5782{
5783 RT_NOREF2(pvUser1, rcReq);
5784 RTSEMEVENT hEvent = (RTSEMEVENT)pvUser2;
5785
5786 RTSemEventSignal(hEvent);
5787}
5788
5789/**
5790 * Initializes HDD backends.
5791 *
5792 * @returns VBox status code.
5793 */
5794VBOXDDU_DECL(int) VDInit(void)
5795{
5796 int rc = vdAddBackends(NIL_RTLDRMOD, aStaticBackends, RT_ELEMENTS(aStaticBackends));
5797 if (RT_SUCCESS(rc))
5798 {
5799 rc = vdAddCacheBackends(NIL_RTLDRMOD, aStaticCacheBackends, RT_ELEMENTS(aStaticCacheBackends));
5800 if (RT_SUCCESS(rc))
5801 {
5802 RTListInit(&g_ListPluginsLoaded);
5803 rc = vdLoadDynamicBackends();
5804 }
5805 }
5806 LogRel(("VD: VDInit finished\n"));
5807 return rc;
5808}
5809
5810/**
5811 * Destroys loaded HDD backends.
5812 *
5813 * @returns VBox status code.
5814 */
5815VBOXDDU_DECL(int) VDShutdown(void)
5816{
5817 if (!g_apBackends)
5818 return VERR_INTERNAL_ERROR;
5819
5820 if (g_apCacheBackends)
5821 RTMemFree(g_apCacheBackends);
5822 RTMemFree(g_apBackends);
5823
5824 g_cBackends = 0;
5825 g_apBackends = NULL;
5826
5827 /* Clear the supported cache backends. */
5828 g_cCacheBackends = 0;
5829 g_apCacheBackends = NULL;
5830
5831#ifndef VBOX_HDD_NO_DYNAMIC_BACKENDS
5832 PVDPLUGIN pPlugin, pPluginNext;
5833
5834 RTListForEachSafe(&g_ListPluginsLoaded, pPlugin, pPluginNext, VDPLUGIN, NodePlugin)
5835 {
5836 RTLdrClose(pPlugin->hPlugin);
5837 RTStrFree(pPlugin->pszFilename);
5838 RTListNodeRemove(&pPlugin->NodePlugin);
5839 RTMemFree(pPlugin);
5840 }
5841#endif
5842
5843 return VINF_SUCCESS;
5844}
5845
5846/**
5847 * Loads a single plugin given by filename.
5848 *
5849 * @returns VBox status code.
5850 * @param pszFilename The plugin filename to load.
5851 */
5852VBOXDDU_DECL(int) VDPluginLoadFromFilename(const char *pszFilename)
5853{
5854 if (!g_apBackends)
5855 {
5856 int rc = VDInit();
5857 if (RT_FAILURE(rc))
5858 return rc;
5859 }
5860
5861 return vdPluginLoadFromFilename(pszFilename);
5862}
5863
5864/**
5865 * Load all plugins from a given path.
5866 *
5867 * @returns VBox statuse code.
5868 * @param pszPath The path to load plugins from.
5869 */
5870VBOXDDU_DECL(int) VDPluginLoadFromPath(const char *pszPath)
5871{
5872 if (!g_apBackends)
5873 {
5874 int rc = VDInit();
5875 if (RT_FAILURE(rc))
5876 return rc;
5877 }
5878
5879 return vdPluginLoadFromPath(pszPath);
5880}
5881
5882/**
5883 * Unloads a single plugin given by filename.
5884 *
5885 * @returns VBox status code.
5886 * @param pszFilename The plugin filename to unload.
5887 */
5888VBOXDDU_DECL(int) VDPluginUnloadFromFilename(const char *pszFilename)
5889{
5890 if (!g_apBackends)
5891 {
5892 int rc = VDInit();
5893 if (RT_FAILURE(rc))
5894 return rc;
5895 }
5896
5897 return vdPluginUnloadFromFilename(pszFilename);
5898}
5899
5900/**
5901 * Unload all plugins from a given path.
5902 *
5903 * @returns VBox statuse code.
5904 * @param pszPath The path to unload plugins from.
5905 */
5906VBOXDDU_DECL(int) VDPluginUnloadFromPath(const char *pszPath)
5907{
5908 if (!g_apBackends)
5909 {
5910 int rc = VDInit();
5911 if (RT_FAILURE(rc))
5912 return rc;
5913 }
5914
5915 return vdPluginUnloadFromPath(pszPath);
5916}
5917
5918/**
5919 * Lists all HDD backends and their capabilities in a caller-provided buffer.
5920 *
5921 * @returns VBox status code.
5922 * VERR_BUFFER_OVERFLOW if not enough space is passed.
5923 * @param cEntriesAlloc Number of list entries available.
5924 * @param pEntries Pointer to array for the entries.
5925 * @param pcEntriesUsed Number of entries returned.
5926 */
5927VBOXDDU_DECL(int) VDBackendInfo(unsigned cEntriesAlloc, PVDBACKENDINFO pEntries,
5928 unsigned *pcEntriesUsed)
5929{
5930 int rc = VINF_SUCCESS;
5931
5932 LogFlowFunc(("cEntriesAlloc=%u pEntries=%#p pcEntriesUsed=%#p\n", cEntriesAlloc, pEntries, pcEntriesUsed));
5933 /* Check arguments. */
5934 AssertMsgReturn(cEntriesAlloc,
5935 ("cEntriesAlloc=%u\n", cEntriesAlloc),
5936 VERR_INVALID_PARAMETER);
5937 AssertMsgReturn(VALID_PTR(pEntries),
5938 ("pEntries=%#p\n", pEntries),
5939 VERR_INVALID_PARAMETER);
5940 AssertMsgReturn(VALID_PTR(pcEntriesUsed),
5941 ("pcEntriesUsed=%#p\n", pcEntriesUsed),
5942 VERR_INVALID_PARAMETER);
5943 if (!g_apBackends)
5944 VDInit();
5945
5946 if (cEntriesAlloc < g_cBackends)
5947 {
5948 *pcEntriesUsed = g_cBackends;
5949 return VERR_BUFFER_OVERFLOW;
5950 }
5951
5952 for (unsigned i = 0; i < g_cBackends; i++)
5953 {
5954 pEntries[i].pszBackend = g_apBackends[i]->pszBackendName;
5955 pEntries[i].uBackendCaps = g_apBackends[i]->uBackendCaps;
5956 pEntries[i].paFileExtensions = g_apBackends[i]->paFileExtensions;
5957 pEntries[i].paConfigInfo = g_apBackends[i]->paConfigInfo;
5958 pEntries[i].pfnComposeLocation = g_apBackends[i]->pfnComposeLocation;
5959 pEntries[i].pfnComposeName = g_apBackends[i]->pfnComposeName;
5960 }
5961
5962 LogFlowFunc(("returns %Rrc *pcEntriesUsed=%u\n", rc, g_cBackends));
5963 *pcEntriesUsed = g_cBackends;
5964 return rc;
5965}
5966
5967/**
5968 * Lists the capabilities of a backend identified by its name.
5969 *
5970 * @returns VBox status code.
5971 * @param pszBackend The backend name.
5972 * @param pEntry Pointer to an entry.
5973 */
5974VBOXDDU_DECL(int) VDBackendInfoOne(const char *pszBackend, PVDBACKENDINFO pEntry)
5975{
5976 LogFlowFunc(("pszBackend=%#p pEntry=%#p\n", pszBackend, pEntry));
5977 /* Check arguments. */
5978 AssertMsgReturn(VALID_PTR(pszBackend),
5979 ("pszBackend=%#p\n", pszBackend),
5980 VERR_INVALID_PARAMETER);
5981 AssertMsgReturn(VALID_PTR(pEntry),
5982 ("pEntry=%#p\n", pEntry),
5983 VERR_INVALID_PARAMETER);
5984 if (!g_apBackends)
5985 VDInit();
5986
5987 /* Go through loaded backends. */
5988 for (unsigned i = 0; i < g_cBackends; i++)
5989 {
5990 if (!RTStrICmp(pszBackend, g_apBackends[i]->pszBackendName))
5991 {
5992 pEntry->pszBackend = g_apBackends[i]->pszBackendName;
5993 pEntry->uBackendCaps = g_apBackends[i]->uBackendCaps;
5994 pEntry->paFileExtensions = g_apBackends[i]->paFileExtensions;
5995 pEntry->paConfigInfo = g_apBackends[i]->paConfigInfo;
5996 return VINF_SUCCESS;
5997 }
5998 }
5999
6000 return VERR_NOT_FOUND;
6001}
6002
6003/**
6004 * Lists all filters and their capabilities in a caller-provided buffer.
6005 *
6006 * @return VBox status code.
6007 * VERR_BUFFER_OVERFLOW if not enough space is passed.
6008 * @param cEntriesAlloc Number of list entries available.
6009 * @param pEntries Pointer to array for the entries.
6010 * @param pcEntriesUsed Number of entries returned.
6011 */
6012VBOXDDU_DECL(int) VDFilterInfo(unsigned cEntriesAlloc, PVDFILTERINFO pEntries,
6013 unsigned *pcEntriesUsed)
6014{
6015 int rc = VINF_SUCCESS;
6016
6017 LogFlowFunc(("cEntriesAlloc=%u pEntries=%#p pcEntriesUsed=%#p\n", cEntriesAlloc, pEntries, pcEntriesUsed));
6018 /* Check arguments. */
6019 AssertMsgReturn(cEntriesAlloc,
6020 ("cEntriesAlloc=%u\n", cEntriesAlloc),
6021 VERR_INVALID_PARAMETER);
6022 AssertMsgReturn(VALID_PTR(pEntries),
6023 ("pEntries=%#p\n", pEntries),
6024 VERR_INVALID_PARAMETER);
6025 AssertMsgReturn(VALID_PTR(pcEntriesUsed),
6026 ("pcEntriesUsed=%#p\n", pcEntriesUsed),
6027 VERR_INVALID_PARAMETER);
6028 if (!g_apBackends)
6029 VDInit();
6030
6031 if (cEntriesAlloc < g_cFilterBackends)
6032 {
6033 *pcEntriesUsed = g_cFilterBackends;
6034 return VERR_BUFFER_OVERFLOW;
6035 }
6036
6037 for (unsigned i = 0; i < g_cFilterBackends; i++)
6038 {
6039 pEntries[i].pszFilter = g_apFilterBackends[i]->pszBackendName;
6040 pEntries[i].paConfigInfo = g_apFilterBackends[i]->paConfigInfo;
6041 }
6042
6043 LogFlowFunc(("returns %Rrc *pcEntriesUsed=%u\n", rc, g_cFilterBackends));
6044 *pcEntriesUsed = g_cFilterBackends;
6045 return rc;
6046}
6047
6048/**
6049 * Lists the capabilities of a filter identified by its name.
6050 *
6051 * @return VBox status code.
6052 * @param pszFilter The filter name (case insensitive).
6053 * @param pEntry Pointer to an entry.
6054 */
6055VBOXDDU_DECL(int) VDFilterInfoOne(const char *pszFilter, PVDFILTERINFO pEntry)
6056{
6057 LogFlowFunc(("pszFilter=%#p pEntry=%#p\n", pszFilter, pEntry));
6058 /* Check arguments. */
6059 AssertMsgReturn(VALID_PTR(pszFilter),
6060 ("pszFilter=%#p\n", pszFilter),
6061 VERR_INVALID_PARAMETER);
6062 AssertMsgReturn(VALID_PTR(pEntry),
6063 ("pEntry=%#p\n", pEntry),
6064 VERR_INVALID_PARAMETER);
6065 if (!g_apBackends)
6066 VDInit();
6067
6068 /* Go through loaded backends. */
6069 for (unsigned i = 0; i < g_cFilterBackends; i++)
6070 {
6071 if (!RTStrICmp(pszFilter, g_apFilterBackends[i]->pszBackendName))
6072 {
6073 pEntry->pszFilter = g_apFilterBackends[i]->pszBackendName;
6074 pEntry->paConfigInfo = g_apFilterBackends[i]->paConfigInfo;
6075 return VINF_SUCCESS;
6076 }
6077 }
6078
6079 return VERR_NOT_FOUND;
6080}
6081
6082/**
6083 * Allocates and initializes an empty HDD container.
6084 * No image files are opened.
6085 *
6086 * @returns VBox status code.
6087 * @param pVDIfsDisk Pointer to the per-disk VD interface list.
6088 * @param enmType Type of the image container.
6089 * @param ppDisk Where to store the reference to HDD container.
6090 */
6091VBOXDDU_DECL(int) VDCreate(PVDINTERFACE pVDIfsDisk, VDTYPE enmType, PVBOXHDD *ppDisk)
6092{
6093 int rc = VINF_SUCCESS;
6094 PVBOXHDD pDisk = NULL;
6095
6096 LogFlowFunc(("pVDIfsDisk=%#p\n", pVDIfsDisk));
6097 do
6098 {
6099 /* Check arguments. */
6100 AssertMsgBreakStmt(VALID_PTR(ppDisk),
6101 ("ppDisk=%#p\n", ppDisk),
6102 rc = VERR_INVALID_PARAMETER);
6103
6104 pDisk = (PVBOXHDD)RTMemAllocZ(sizeof(VBOXHDD));
6105 if (pDisk)
6106 {
6107 pDisk->u32Signature = VBOXHDDDISK_SIGNATURE;
6108 pDisk->enmType = enmType;
6109 pDisk->cImages = 0;
6110 pDisk->pBase = NULL;
6111 pDisk->pLast = NULL;
6112 pDisk->cbSize = 0;
6113 pDisk->PCHSGeometry.cCylinders = 0;
6114 pDisk->PCHSGeometry.cHeads = 0;
6115 pDisk->PCHSGeometry.cSectors = 0;
6116 pDisk->LCHSGeometry.cCylinders = 0;
6117 pDisk->LCHSGeometry.cHeads = 0;
6118 pDisk->LCHSGeometry.cSectors = 0;
6119 pDisk->pVDIfsDisk = pVDIfsDisk;
6120 pDisk->pInterfaceError = NULL;
6121 pDisk->pInterfaceThreadSync = NULL;
6122 pDisk->pIoCtxLockOwner = NULL;
6123 pDisk->pIoCtxHead = NULL;
6124 pDisk->fLocked = false;
6125 pDisk->hMemCacheIoCtx = NIL_RTMEMCACHE;
6126 pDisk->hMemCacheIoTask = NIL_RTMEMCACHE;
6127 RTListInit(&pDisk->ListFilterChainWrite);
6128 RTListInit(&pDisk->ListFilterChainRead);
6129
6130 /* Create the I/O ctx cache */
6131 rc = RTMemCacheCreate(&pDisk->hMemCacheIoCtx, sizeof(VDIOCTX), 0, UINT32_MAX,
6132 NULL, NULL, NULL, 0);
6133 if (RT_FAILURE(rc))
6134 break;
6135
6136 /* Create the I/O task cache */
6137 rc = RTMemCacheCreate(&pDisk->hMemCacheIoTask, sizeof(VDIOTASK), 0, UINT32_MAX,
6138 NULL, NULL, NULL, 0);
6139 if (RT_FAILURE(rc))
6140 break;
6141
6142 pDisk->pInterfaceError = VDIfErrorGet(pVDIfsDisk);
6143 pDisk->pInterfaceThreadSync = VDIfThreadSyncGet(pVDIfsDisk);
6144
6145 *ppDisk = pDisk;
6146 }
6147 else
6148 {
6149 rc = VERR_NO_MEMORY;
6150 break;
6151 }
6152 } while (0);
6153
6154 if ( RT_FAILURE(rc)
6155 && pDisk)
6156 {
6157 if (pDisk->hMemCacheIoCtx != NIL_RTMEMCACHE)
6158 RTMemCacheDestroy(pDisk->hMemCacheIoCtx);
6159 if (pDisk->hMemCacheIoTask != NIL_RTMEMCACHE)
6160 RTMemCacheDestroy(pDisk->hMemCacheIoTask);
6161 }
6162
6163 LogFlowFunc(("returns %Rrc (pDisk=%#p)\n", rc, pDisk));
6164 return rc;
6165}
6166
6167/**
6168 * Destroys HDD container.
6169 * If container has opened image files they will be closed.
6170 *
6171 * @returns VBox status code.
6172 * @param pDisk Pointer to HDD container.
6173 */
6174VBOXDDU_DECL(int) VDDestroy(PVBOXHDD pDisk)
6175{
6176 int rc = VINF_SUCCESS;
6177 LogFlowFunc(("pDisk=%#p\n", pDisk));
6178 do
6179 {
6180 /* sanity check */
6181 AssertPtrBreak(pDisk);
6182 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
6183 Assert(!pDisk->fLocked);
6184
6185 rc = VDCloseAll(pDisk);
6186 int rc2 = VDFilterRemoveAll(pDisk);
6187 if (RT_SUCCESS(rc))
6188 rc = rc2;
6189
6190 RTMemCacheDestroy(pDisk->hMemCacheIoCtx);
6191 RTMemCacheDestroy(pDisk->hMemCacheIoTask);
6192 RTMemFree(pDisk);
6193 } while (0);
6194 LogFlowFunc(("returns %Rrc\n", rc));
6195 return rc;
6196}
6197
6198/**
6199 * Try to get the backend name which can use this image.
6200 *
6201 * @returns VBox status code.
6202 * VINF_SUCCESS if a plugin was found.
6203 * ppszFormat contains the string which can be used as backend name.
6204 * VERR_NOT_SUPPORTED if no backend was found.
6205 * @param pVDIfsDisk Pointer to the per-disk VD interface list.
6206 * @param pVDIfsImage Pointer to the per-image VD interface list.
6207 * @param pszFilename Name of the image file for which the backend is queried.
6208 * @param ppszFormat Receives pointer of the UTF-8 string which contains the format name.
6209 * The returned pointer must be freed using RTStrFree().
6210 */
6211VBOXDDU_DECL(int) VDGetFormat(PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
6212 const char *pszFilename, char **ppszFormat, VDTYPE *penmType)
6213{
6214 int rc = VERR_NOT_SUPPORTED;
6215 VDINTERFACEIOINT VDIfIoInt;
6216 VDINTERFACEIO VDIfIoFallback;
6217 PVDINTERFACEIO pInterfaceIo;
6218
6219 LogFlowFunc(("pszFilename=\"%s\"\n", pszFilename));
6220 /* Check arguments. */
6221 AssertMsgReturn(VALID_PTR(pszFilename) && *pszFilename,
6222 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
6223 VERR_INVALID_PARAMETER);
6224 AssertMsgReturn(VALID_PTR(ppszFormat),
6225 ("ppszFormat=%#p\n", ppszFormat),
6226 VERR_INVALID_PARAMETER);
6227 AssertMsgReturn(VALID_PTR(penmType),
6228 ("penmType=%#p\n", penmType),
6229 VERR_INVALID_PARAMETER);
6230
6231 if (!g_apBackends)
6232 VDInit();
6233
6234 pInterfaceIo = VDIfIoGet(pVDIfsImage);
6235 if (!pInterfaceIo)
6236 {
6237 /*
6238 * Caller doesn't provide an I/O interface, create our own using the
6239 * native file API.
6240 */
6241 vdIfIoFallbackCallbacksSetup(&VDIfIoFallback);
6242 pInterfaceIo = &VDIfIoFallback;
6243 }
6244
6245 /* Set up the internal I/O interface. */
6246 AssertReturn(!VDIfIoIntGet(pVDIfsImage), VERR_INVALID_PARAMETER);
6247 VDIfIoInt.pfnOpen = vdIOIntOpenLimited;
6248 VDIfIoInt.pfnClose = vdIOIntCloseLimited;
6249 VDIfIoInt.pfnDelete = vdIOIntDeleteLimited;
6250 VDIfIoInt.pfnMove = vdIOIntMoveLimited;
6251 VDIfIoInt.pfnGetFreeSpace = vdIOIntGetFreeSpaceLimited;
6252 VDIfIoInt.pfnGetModificationTime = vdIOIntGetModificationTimeLimited;
6253 VDIfIoInt.pfnGetSize = vdIOIntGetSizeLimited;
6254 VDIfIoInt.pfnSetSize = vdIOIntSetSizeLimited;
6255 VDIfIoInt.pfnReadUser = vdIOIntReadUserLimited;
6256 VDIfIoInt.pfnWriteUser = vdIOIntWriteUserLimited;
6257 VDIfIoInt.pfnReadMeta = vdIOIntReadMetaLimited;
6258 VDIfIoInt.pfnWriteMeta = vdIOIntWriteMetaLimited;
6259 VDIfIoInt.pfnFlush = vdIOIntFlushLimited;
6260 rc = VDInterfaceAdd(&VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
6261 pInterfaceIo, sizeof(VDINTERFACEIOINT), &pVDIfsImage);
6262 AssertRC(rc);
6263
6264 /* Find the backend supporting this file format. */
6265 for (unsigned i = 0; i < g_cBackends; i++)
6266 {
6267 if (g_apBackends[i]->pfnProbe)
6268 {
6269 rc = g_apBackends[i]->pfnProbe(pszFilename, pVDIfsDisk, pVDIfsImage, penmType);
6270 if ( RT_SUCCESS(rc)
6271 /* The correct backend has been found, but there is a small
6272 * incompatibility so that the file cannot be used. Stop here
6273 * and signal success - the actual open will of course fail,
6274 * but that will create a really sensible error message. */
6275 || ( rc != VERR_VD_GEN_INVALID_HEADER
6276 && rc != VERR_VD_VDI_INVALID_HEADER
6277 && rc != VERR_VD_VMDK_INVALID_HEADER
6278 && rc != VERR_VD_ISCSI_INVALID_HEADER
6279 && rc != VERR_VD_VHD_INVALID_HEADER
6280 && rc != VERR_VD_RAW_INVALID_HEADER
6281 && rc != VERR_VD_RAW_SIZE_MODULO_512
6282 && rc != VERR_VD_RAW_SIZE_MODULO_2048
6283 && rc != VERR_VD_RAW_SIZE_OPTICAL_TOO_SMALL
6284 && rc != VERR_VD_RAW_SIZE_FLOPPY_TOO_BIG
6285 && rc != VERR_VD_PARALLELS_INVALID_HEADER
6286 && rc != VERR_VD_DMG_INVALID_HEADER))
6287 {
6288 /* Copy the name into the new string. */
6289 char *pszFormat = RTStrDup(g_apBackends[i]->pszBackendName);
6290 if (!pszFormat)
6291 {
6292 rc = VERR_NO_MEMORY;
6293 break;
6294 }
6295 *ppszFormat = pszFormat;
6296 /* Do not consider the typical file access errors as success,
6297 * which allows the caller to deal with such issues. */
6298 if ( rc != VERR_ACCESS_DENIED
6299 && rc != VERR_PATH_NOT_FOUND
6300 && rc != VERR_FILE_NOT_FOUND)
6301 rc = VINF_SUCCESS;
6302 break;
6303 }
6304 rc = VERR_NOT_SUPPORTED;
6305 }
6306 }
6307
6308 /* Try the cache backends. */
6309 if (rc == VERR_NOT_SUPPORTED)
6310 {
6311 for (unsigned i = 0; i < g_cCacheBackends; i++)
6312 {
6313 if (g_apCacheBackends[i]->pfnProbe)
6314 {
6315 rc = g_apCacheBackends[i]->pfnProbe(pszFilename, pVDIfsDisk,
6316 pVDIfsImage);
6317 if ( RT_SUCCESS(rc)
6318 || (rc != VERR_VD_GEN_INVALID_HEADER))
6319 {
6320 /* Copy the name into the new string. */
6321 char *pszFormat = RTStrDup(g_apBackends[i]->pszBackendName);
6322 if (!pszFormat)
6323 {
6324 rc = VERR_NO_MEMORY;
6325 break;
6326 }
6327 *ppszFormat = pszFormat;
6328 rc = VINF_SUCCESS;
6329 break;
6330 }
6331 rc = VERR_NOT_SUPPORTED;
6332 }
6333 }
6334 }
6335
6336 LogFlowFunc(("returns %Rrc *ppszFormat=\"%s\"\n", rc, *ppszFormat));
6337 return rc;
6338}
6339
6340/**
6341 * Opens an image file.
6342 *
6343 * The first opened image file in HDD container must have a base image type,
6344 * others (next opened images) must be a differencing or undo images.
6345 * Linkage is checked for differencing image to be in consistence with the previously opened image.
6346 * When another differencing image is opened and the last image was opened in read/write access
6347 * mode, then the last image is reopened in read-only with deny write sharing mode. This allows
6348 * other processes to use images in read-only mode too.
6349 *
6350 * Note that the image is opened in read-only mode if a read/write open is not possible.
6351 * Use VDIsReadOnly to check open mode.
6352 *
6353 * @returns VBox status code.
6354 * @param pDisk Pointer to HDD container.
6355 * @param pszBackend Name of the image file backend to use.
6356 * @param pszFilename Name of the image file to open.
6357 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
6358 * @param pVDIfsImage Pointer to the per-image VD interface list.
6359 */
6360VBOXDDU_DECL(int) VDOpen(PVBOXHDD pDisk, const char *pszBackend,
6361 const char *pszFilename, unsigned uOpenFlags,
6362 PVDINTERFACE pVDIfsImage)
6363{
6364 int rc = VINF_SUCCESS;
6365 int rc2;
6366 bool fLockWrite = false;
6367 PVDIMAGE pImage = NULL;
6368
6369 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" uOpenFlags=%#x, pVDIfsImage=%#p\n",
6370 pDisk, pszBackend, pszFilename, uOpenFlags, pVDIfsImage));
6371
6372 do
6373 {
6374 /* sanity check */
6375 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
6376 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
6377
6378 /* Check arguments. */
6379 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
6380 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
6381 rc = VERR_INVALID_PARAMETER);
6382 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
6383 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
6384 rc = VERR_INVALID_PARAMETER);
6385 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
6386 ("uOpenFlags=%#x\n", uOpenFlags),
6387 rc = VERR_INVALID_PARAMETER);
6388 AssertMsgBreakStmt( !(uOpenFlags & VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS)
6389 || (uOpenFlags & VD_OPEN_FLAGS_READONLY),
6390 ("uOpenFlags=%#x\n", uOpenFlags),
6391 rc = VERR_INVALID_PARAMETER);
6392
6393 /*
6394 * Destroy the current discard state first which might still have pending blocks
6395 * for the currently opened image which will be switched to readonly mode.
6396 */
6397 /* Lock disk for writing, as we modify pDisk information below. */
6398 rc2 = vdThreadStartWrite(pDisk);
6399 AssertRC(rc2);
6400 fLockWrite = true;
6401 rc = vdDiscardStateDestroy(pDisk);
6402 if (RT_FAILURE(rc))
6403 break;
6404 rc2 = vdThreadFinishWrite(pDisk);
6405 AssertRC(rc2);
6406 fLockWrite = false;
6407
6408 /* Set up image descriptor. */
6409 pImage = (PVDIMAGE)RTMemAllocZ(sizeof(VDIMAGE));
6410 if (!pImage)
6411 {
6412 rc = VERR_NO_MEMORY;
6413 break;
6414 }
6415 pImage->pszFilename = RTStrDup(pszFilename);
6416 if (!pImage->pszFilename)
6417 {
6418 rc = VERR_NO_MEMORY;
6419 break;
6420 }
6421
6422 pImage->VDIo.pDisk = pDisk;
6423 pImage->pVDIfsImage = pVDIfsImage;
6424
6425 rc = vdFindBackend(pszBackend, &pImage->Backend);
6426 if (RT_FAILURE(rc))
6427 break;
6428 if (!pImage->Backend)
6429 {
6430 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
6431 N_("VD: unknown backend name '%s'"), pszBackend);
6432 break;
6433 }
6434
6435 /*
6436 * Fail if the backend can't do async I/O but the
6437 * flag is set.
6438 */
6439 if ( !(pImage->Backend->uBackendCaps & VD_CAP_ASYNC)
6440 && (uOpenFlags & VD_OPEN_FLAGS_ASYNC_IO))
6441 {
6442 rc = vdError(pDisk, VERR_NOT_SUPPORTED, RT_SRC_POS,
6443 N_("VD: Backend '%s' does not support async I/O"), pszBackend);
6444 break;
6445 }
6446
6447 /*
6448 * Fail if the backend doesn't support the discard operation but the
6449 * flag is set.
6450 */
6451 if ( !(pImage->Backend->uBackendCaps & VD_CAP_DISCARD)
6452 && (uOpenFlags & VD_OPEN_FLAGS_DISCARD))
6453 {
6454 rc = vdError(pDisk, VERR_VD_DISCARD_NOT_SUPPORTED, RT_SRC_POS,
6455 N_("VD: Backend '%s' does not support discard"), pszBackend);
6456 break;
6457 }
6458
6459 /* Set up the I/O interface. */
6460 pImage->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsImage);
6461 if (!pImage->VDIo.pInterfaceIo)
6462 {
6463 vdIfIoFallbackCallbacksSetup(&pImage->VDIo.VDIfIo);
6464 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
6465 pDisk, sizeof(VDINTERFACEIO), &pVDIfsImage);
6466 pImage->VDIo.pInterfaceIo = &pImage->VDIo.VDIfIo;
6467 }
6468
6469 /* Set up the internal I/O interface. */
6470 AssertBreakStmt(!VDIfIoIntGet(pVDIfsImage), rc = VERR_INVALID_PARAMETER);
6471 vdIfIoIntCallbacksSetup(&pImage->VDIo.VDIfIoInt);
6472 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
6473 &pImage->VDIo, sizeof(VDINTERFACEIOINT), &pImage->pVDIfsImage);
6474 AssertRC(rc);
6475
6476 pImage->uOpenFlags = uOpenFlags & (VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_DISCARD | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS);
6477 pImage->VDIo.fIgnoreFlush = (uOpenFlags & VD_OPEN_FLAGS_IGNORE_FLUSH) != 0;
6478 rc = pImage->Backend->pfnOpen(pImage->pszFilename,
6479 uOpenFlags & ~(VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS),
6480 pDisk->pVDIfsDisk,
6481 pImage->pVDIfsImage,
6482 pDisk->enmType,
6483 &pImage->pBackendData);
6484 /*
6485 * If the image is corrupted and there is a repair method try to repair it
6486 * first if it was openend in read-write mode and open again afterwards.
6487 */
6488 if ( RT_UNLIKELY(rc == VERR_VD_IMAGE_CORRUPTED)
6489 && !(uOpenFlags & VD_OPEN_FLAGS_READONLY)
6490 && pImage->Backend->pfnRepair)
6491 {
6492 rc = pImage->Backend->pfnRepair(pszFilename, pDisk->pVDIfsDisk, pImage->pVDIfsImage, 0 /* fFlags */);
6493 if (RT_SUCCESS(rc))
6494 rc = pImage->Backend->pfnOpen(pImage->pszFilename,
6495 uOpenFlags & ~(VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS),
6496 pDisk->pVDIfsDisk,
6497 pImage->pVDIfsImage,
6498 pDisk->enmType,
6499 &pImage->pBackendData);
6500 else
6501 {
6502 rc = vdError(pDisk, rc, RT_SRC_POS,
6503 N_("VD: error %Rrc repairing corrupted image file '%s'"), rc, pszFilename);
6504 break;
6505 }
6506 }
6507 else if (RT_UNLIKELY(rc == VERR_VD_IMAGE_CORRUPTED))
6508 {
6509 rc = vdError(pDisk, rc, RT_SRC_POS,
6510 N_("VD: Image file '%s' is corrupted and can't be opened"), pszFilename);
6511 break;
6512 }
6513
6514 /* If the open in read-write mode failed, retry in read-only mode. */
6515 if (RT_FAILURE(rc))
6516 {
6517 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY)
6518 && ( rc == VERR_ACCESS_DENIED
6519 || rc == VERR_PERMISSION_DENIED
6520 || rc == VERR_WRITE_PROTECT
6521 || rc == VERR_SHARING_VIOLATION
6522 || rc == VERR_FILE_LOCK_FAILED))
6523 rc = pImage->Backend->pfnOpen(pImage->pszFilename,
6524 (uOpenFlags & ~(VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS))
6525 | VD_OPEN_FLAGS_READONLY,
6526 pDisk->pVDIfsDisk,
6527 pImage->pVDIfsImage,
6528 pDisk->enmType,
6529 &pImage->pBackendData);
6530 if (RT_FAILURE(rc))
6531 {
6532 rc = vdError(pDisk, rc, RT_SRC_POS,
6533 N_("VD: error %Rrc opening image file '%s'"), rc, pszFilename);
6534 break;
6535 }
6536 }
6537
6538 /* Lock disk for writing, as we modify pDisk information below. */
6539 rc2 = vdThreadStartWrite(pDisk);
6540 AssertRC(rc2);
6541 fLockWrite = true;
6542
6543 pImage->VDIo.pBackendData = pImage->pBackendData;
6544
6545 /* Check image type. As the image itself has only partial knowledge
6546 * whether it's a base image or not, this info is derived here. The
6547 * base image can be fixed or normal, all others must be normal or
6548 * diff images. Some image formats don't distinguish between normal
6549 * and diff images, so this must be corrected here. */
6550 unsigned uImageFlags;
6551 uImageFlags = pImage->Backend->pfnGetImageFlags(pImage->pBackendData);
6552 if (RT_FAILURE(rc))
6553 uImageFlags = VD_IMAGE_FLAGS_NONE;
6554 if ( RT_SUCCESS(rc)
6555 && !(uOpenFlags & VD_OPEN_FLAGS_INFO))
6556 {
6557 if ( pDisk->cImages == 0
6558 && (uImageFlags & VD_IMAGE_FLAGS_DIFF))
6559 {
6560 rc = VERR_VD_INVALID_TYPE;
6561 break;
6562 }
6563 else if (pDisk->cImages != 0)
6564 {
6565 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
6566 {
6567 rc = VERR_VD_INVALID_TYPE;
6568 break;
6569 }
6570 else
6571 uImageFlags |= VD_IMAGE_FLAGS_DIFF;
6572 }
6573 }
6574
6575 /* Ensure we always get correct diff information, even if the backend
6576 * doesn't actually have a stored flag for this. It must not return
6577 * bogus information for the parent UUID if it is not a diff image. */
6578 RTUUID parentUuid;
6579 RTUuidClear(&parentUuid);
6580 rc2 = pImage->Backend->pfnGetParentUuid(pImage->pBackendData, &parentUuid);
6581 if (RT_SUCCESS(rc2) && !RTUuidIsNull(&parentUuid))
6582 uImageFlags |= VD_IMAGE_FLAGS_DIFF;
6583
6584 pImage->uImageFlags = uImageFlags;
6585
6586 /* Force sane optimization settings. It's not worth avoiding writes
6587 * to fixed size images. The overhead would have almost no payback. */
6588 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
6589 pImage->uOpenFlags |= VD_OPEN_FLAGS_HONOR_SAME;
6590
6591 /** @todo optionally check UUIDs */
6592
6593 /* Cache disk information. */
6594 pDisk->cbSize = pImage->Backend->pfnGetSize(pImage->pBackendData);
6595
6596 /* Cache PCHS geometry. */
6597 rc2 = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
6598 &pDisk->PCHSGeometry);
6599 if (RT_FAILURE(rc2))
6600 {
6601 pDisk->PCHSGeometry.cCylinders = 0;
6602 pDisk->PCHSGeometry.cHeads = 0;
6603 pDisk->PCHSGeometry.cSectors = 0;
6604 }
6605 else
6606 {
6607 /* Make sure the PCHS geometry is properly clipped. */
6608 pDisk->PCHSGeometry.cCylinders = RT_MIN(pDisk->PCHSGeometry.cCylinders, 16383);
6609 pDisk->PCHSGeometry.cHeads = RT_MIN(pDisk->PCHSGeometry.cHeads, 16);
6610 pDisk->PCHSGeometry.cSectors = RT_MIN(pDisk->PCHSGeometry.cSectors, 63);
6611 }
6612
6613 /* Cache LCHS geometry. */
6614 rc2 = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
6615 &pDisk->LCHSGeometry);
6616 if (RT_FAILURE(rc2))
6617 {
6618 pDisk->LCHSGeometry.cCylinders = 0;
6619 pDisk->LCHSGeometry.cHeads = 0;
6620 pDisk->LCHSGeometry.cSectors = 0;
6621 }
6622 else
6623 {
6624 /* Make sure the LCHS geometry is properly clipped. */
6625 pDisk->LCHSGeometry.cHeads = RT_MIN(pDisk->LCHSGeometry.cHeads, 255);
6626 pDisk->LCHSGeometry.cSectors = RT_MIN(pDisk->LCHSGeometry.cSectors, 63);
6627 }
6628
6629 if (pDisk->cImages != 0)
6630 {
6631 /* Switch previous image to read-only mode. */
6632 unsigned uOpenFlagsPrevImg;
6633 uOpenFlagsPrevImg = pDisk->pLast->Backend->pfnGetOpenFlags(pDisk->pLast->pBackendData);
6634 if (!(uOpenFlagsPrevImg & VD_OPEN_FLAGS_READONLY))
6635 {
6636 uOpenFlagsPrevImg |= VD_OPEN_FLAGS_READONLY;
6637 rc = pDisk->pLast->Backend->pfnSetOpenFlags(pDisk->pLast->pBackendData, uOpenFlagsPrevImg);
6638 }
6639 }
6640
6641 if (RT_SUCCESS(rc))
6642 {
6643 /* Image successfully opened, make it the last image. */
6644 vdAddImageToList(pDisk, pImage);
6645 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
6646 pDisk->uModified = VD_IMAGE_MODIFIED_FIRST;
6647 }
6648 else
6649 {
6650 /* Error detected, but image opened. Close image. */
6651 rc2 = pImage->Backend->pfnClose(pImage->pBackendData, false);
6652 AssertRC(rc2);
6653 pImage->pBackendData = NULL;
6654 }
6655 } while (0);
6656
6657 if (RT_UNLIKELY(fLockWrite))
6658 {
6659 rc2 = vdThreadFinishWrite(pDisk);
6660 AssertRC(rc2);
6661 }
6662
6663 if (RT_FAILURE(rc))
6664 {
6665 if (pImage)
6666 {
6667 if (pImage->pszFilename)
6668 RTStrFree(pImage->pszFilename);
6669 RTMemFree(pImage);
6670 }
6671 }
6672
6673 LogFlowFunc(("returns %Rrc\n", rc));
6674 return rc;
6675}
6676
6677/**
6678 * Opens a cache image.
6679 *
6680 * @return VBox status code.
6681 * @param pDisk Pointer to the HDD container which should use the cache image.
6682 * @param pszBackend Name of the cache file backend to use (case insensitive).
6683 * @param pszFilename Name of the cache image to open.
6684 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
6685 * @param pVDIfsCache Pointer to the per-cache VD interface list.
6686 */
6687VBOXDDU_DECL(int) VDCacheOpen(PVBOXHDD pDisk, const char *pszBackend,
6688 const char *pszFilename, unsigned uOpenFlags,
6689 PVDINTERFACE pVDIfsCache)
6690{
6691 int rc = VINF_SUCCESS;
6692 int rc2;
6693 bool fLockWrite = false;
6694 PVDCACHE pCache = NULL;
6695
6696 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" uOpenFlags=%#x, pVDIfsCache=%#p\n",
6697 pDisk, pszBackend, pszFilename, uOpenFlags, pVDIfsCache));
6698
6699 do
6700 {
6701 /* sanity check */
6702 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
6703 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
6704
6705 /* Check arguments. */
6706 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
6707 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
6708 rc = VERR_INVALID_PARAMETER);
6709 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
6710 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
6711 rc = VERR_INVALID_PARAMETER);
6712 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
6713 ("uOpenFlags=%#x\n", uOpenFlags),
6714 rc = VERR_INVALID_PARAMETER);
6715
6716 /* Set up image descriptor. */
6717 pCache = (PVDCACHE)RTMemAllocZ(sizeof(VDCACHE));
6718 if (!pCache)
6719 {
6720 rc = VERR_NO_MEMORY;
6721 break;
6722 }
6723 pCache->pszFilename = RTStrDup(pszFilename);
6724 if (!pCache->pszFilename)
6725 {
6726 rc = VERR_NO_MEMORY;
6727 break;
6728 }
6729
6730 pCache->VDIo.pDisk = pDisk;
6731 pCache->pVDIfsCache = pVDIfsCache;
6732
6733 rc = vdFindCacheBackend(pszBackend, &pCache->Backend);
6734 if (RT_FAILURE(rc))
6735 break;
6736 if (!pCache->Backend)
6737 {
6738 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
6739 N_("VD: unknown backend name '%s'"), pszBackend);
6740 break;
6741 }
6742
6743 /* Set up the I/O interface. */
6744 pCache->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsCache);
6745 if (!pCache->VDIo.pInterfaceIo)
6746 {
6747 vdIfIoFallbackCallbacksSetup(&pCache->VDIo.VDIfIo);
6748 rc = VDInterfaceAdd(&pCache->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
6749 pDisk, sizeof(VDINTERFACEIO), &pVDIfsCache);
6750 pCache->VDIo.pInterfaceIo = &pCache->VDIo.VDIfIo;
6751 }
6752
6753 /* Set up the internal I/O interface. */
6754 AssertBreakStmt(!VDIfIoIntGet(pVDIfsCache), rc = VERR_INVALID_PARAMETER);
6755 vdIfIoIntCallbacksSetup(&pCache->VDIo.VDIfIoInt);
6756 rc = VDInterfaceAdd(&pCache->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
6757 &pCache->VDIo, sizeof(VDINTERFACEIOINT), &pCache->pVDIfsCache);
6758 AssertRC(rc);
6759
6760 pCache->uOpenFlags = uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME;
6761 rc = pCache->Backend->pfnOpen(pCache->pszFilename,
6762 uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME,
6763 pDisk->pVDIfsDisk,
6764 pCache->pVDIfsCache,
6765 &pCache->pBackendData);
6766 /* If the open in read-write mode failed, retry in read-only mode. */
6767 if (RT_FAILURE(rc))
6768 {
6769 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY)
6770 && ( rc == VERR_ACCESS_DENIED
6771 || rc == VERR_PERMISSION_DENIED
6772 || rc == VERR_WRITE_PROTECT
6773 || rc == VERR_SHARING_VIOLATION
6774 || rc == VERR_FILE_LOCK_FAILED))
6775 rc = pCache->Backend->pfnOpen(pCache->pszFilename,
6776 (uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME)
6777 | VD_OPEN_FLAGS_READONLY,
6778 pDisk->pVDIfsDisk,
6779 pCache->pVDIfsCache,
6780 &pCache->pBackendData);
6781 if (RT_FAILURE(rc))
6782 {
6783 rc = vdError(pDisk, rc, RT_SRC_POS,
6784 N_("VD: error %Rrc opening image file '%s'"), rc, pszFilename);
6785 break;
6786 }
6787 }
6788
6789 /* Lock disk for writing, as we modify pDisk information below. */
6790 rc2 = vdThreadStartWrite(pDisk);
6791 AssertRC(rc2);
6792 fLockWrite = true;
6793
6794 /*
6795 * Check that the modification UUID of the cache and last image
6796 * match. If not the image was modified in-between without the cache.
6797 * The cache might contain stale data.
6798 */
6799 RTUUID UuidImage, UuidCache;
6800
6801 rc = pCache->Backend->pfnGetModificationUuid(pCache->pBackendData,
6802 &UuidCache);
6803 if (RT_SUCCESS(rc))
6804 {
6805 rc = pDisk->pLast->Backend->pfnGetModificationUuid(pDisk->pLast->pBackendData,
6806 &UuidImage);
6807 if (RT_SUCCESS(rc))
6808 {
6809 if (RTUuidCompare(&UuidImage, &UuidCache))
6810 rc = VERR_VD_CACHE_NOT_UP_TO_DATE;
6811 }
6812 }
6813
6814 /*
6815 * We assume that the user knows what he is doing if one of the images
6816 * doesn't support the modification uuid.
6817 */
6818 if (rc == VERR_NOT_SUPPORTED)
6819 rc = VINF_SUCCESS;
6820
6821 if (RT_SUCCESS(rc))
6822 {
6823 /* Cache successfully opened, make it the current one. */
6824 if (!pDisk->pCache)
6825 pDisk->pCache = pCache;
6826 else
6827 rc = VERR_VD_CACHE_ALREADY_EXISTS;
6828 }
6829
6830 if (RT_FAILURE(rc))
6831 {
6832 /* Error detected, but image opened. Close image. */
6833 rc2 = pCache->Backend->pfnClose(pCache->pBackendData, false);
6834 AssertRC(rc2);
6835 pCache->pBackendData = NULL;
6836 }
6837 } while (0);
6838
6839 if (RT_UNLIKELY(fLockWrite))
6840 {
6841 rc2 = vdThreadFinishWrite(pDisk);
6842 AssertRC(rc2);
6843 }
6844
6845 if (RT_FAILURE(rc))
6846 {
6847 if (pCache)
6848 {
6849 if (pCache->pszFilename)
6850 RTStrFree(pCache->pszFilename);
6851 RTMemFree(pCache);
6852 }
6853 }
6854
6855 LogFlowFunc(("returns %Rrc\n", rc));
6856 return rc;
6857}
6858
6859VBOXDDU_DECL(int) VDFilterAdd(PVBOXHDD pDisk, const char *pszFilter, uint32_t fFlags,
6860 PVDINTERFACE pVDIfsFilter)
6861{
6862 int rc = VINF_SUCCESS;
6863 int rc2;
6864 bool fLockWrite = false;
6865 PVDFILTER pFilter = NULL;
6866
6867 LogFlowFunc(("pDisk=%#p pszFilter=\"%s\" pVDIfsFilter=%#p\n",
6868 pDisk, pszFilter, pVDIfsFilter));
6869
6870 do
6871 {
6872 /* sanity check */
6873 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
6874 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
6875
6876 /* Check arguments. */
6877 AssertMsgBreakStmt(VALID_PTR(pszFilter) && *pszFilter,
6878 ("pszFilter=%#p \"%s\"\n", pszFilter, pszFilter),
6879 rc = VERR_INVALID_PARAMETER);
6880
6881 AssertMsgBreakStmt(!(fFlags & ~VD_FILTER_FLAGS_MASK),
6882 ("Invalid flags set (fFlags=%#x)\n", fFlags),
6883 rc = VERR_INVALID_PARAMETER);
6884
6885 /* Set up image descriptor. */
6886 pFilter = (PVDFILTER)RTMemAllocZ(sizeof(VDFILTER));
6887 if (!pFilter)
6888 {
6889 rc = VERR_NO_MEMORY;
6890 break;
6891 }
6892
6893 rc = vdFindFilterBackend(pszFilter, &pFilter->pBackend);
6894 if (RT_FAILURE(rc))
6895 break;
6896 if (!pFilter->pBackend)
6897 {
6898 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
6899 N_("VD: unknown filter backend name '%s'"), pszFilter);
6900 break;
6901 }
6902
6903 pFilter->VDIo.pDisk = pDisk;
6904 pFilter->pVDIfsFilter = pVDIfsFilter;
6905
6906 /* Set up the internal I/O interface. */
6907 AssertBreakStmt(!VDIfIoIntGet(pVDIfsFilter), rc = VERR_INVALID_PARAMETER);
6908 vdIfIoIntCallbacksSetup(&pFilter->VDIo.VDIfIoInt);
6909 rc = VDInterfaceAdd(&pFilter->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
6910 &pFilter->VDIo, sizeof(VDINTERFACEIOINT), &pFilter->pVDIfsFilter);
6911 AssertRC(rc);
6912
6913 rc = pFilter->pBackend->pfnCreate(pDisk->pVDIfsDisk, fFlags & VD_FILTER_FLAGS_INFO,
6914 pFilter->pVDIfsFilter, &pFilter->pvBackendData);
6915 if (RT_FAILURE(rc))
6916 break;
6917
6918 /* Lock disk for writing, as we modify pDisk information below. */
6919 rc2 = vdThreadStartWrite(pDisk);
6920 AssertRC(rc2);
6921 fLockWrite = true;
6922
6923 /* Add filter to chains. */
6924 if (fFlags & VD_FILTER_FLAGS_WRITE)
6925 {
6926 RTListAppend(&pDisk->ListFilterChainWrite, &pFilter->ListNodeChainWrite);
6927 vdFilterRetain(pFilter);
6928 }
6929
6930 if (fFlags & VD_FILTER_FLAGS_READ)
6931 {
6932 RTListAppend(&pDisk->ListFilterChainRead, &pFilter->ListNodeChainRead);
6933 vdFilterRetain(pFilter);
6934 }
6935 } while (0);
6936
6937 if (RT_UNLIKELY(fLockWrite))
6938 {
6939 rc2 = vdThreadFinishWrite(pDisk);
6940 AssertRC(rc2);
6941 }
6942
6943 if (RT_FAILURE(rc))
6944 {
6945 if (pFilter)
6946 RTMemFree(pFilter);
6947 }
6948
6949 LogFlowFunc(("returns %Rrc\n", rc));
6950 return rc;
6951}
6952
6953/**
6954 * Creates and opens a new base image file.
6955 *
6956 * @returns VBox status code.
6957 * @param pDisk Pointer to HDD container.
6958 * @param pszBackend Name of the image file backend to use.
6959 * @param pszFilename Name of the image file to create.
6960 * @param cbSize Image size in bytes.
6961 * @param uImageFlags Flags specifying special image features.
6962 * @param pszComment Pointer to image comment. NULL is ok.
6963 * @param pPCHSGeometry Pointer to physical disk geometry <= (16383,16,63). Not NULL.
6964 * @param pLCHSGeometry Pointer to logical disk geometry <= (x,255,63). Not NULL.
6965 * @param pUuid New UUID of the image. If NULL, a new UUID is created.
6966 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
6967 * @param pVDIfsImage Pointer to the per-image VD interface list.
6968 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
6969 */
6970VBOXDDU_DECL(int) VDCreateBase(PVBOXHDD pDisk, const char *pszBackend,
6971 const char *pszFilename, uint64_t cbSize,
6972 unsigned uImageFlags, const char *pszComment,
6973 PCVDGEOMETRY pPCHSGeometry,
6974 PCVDGEOMETRY pLCHSGeometry,
6975 PCRTUUID pUuid, unsigned uOpenFlags,
6976 PVDINTERFACE pVDIfsImage,
6977 PVDINTERFACE pVDIfsOperation)
6978{
6979 int rc = VINF_SUCCESS;
6980 int rc2;
6981 bool fLockWrite = false, fLockRead = false;
6982 PVDIMAGE pImage = NULL;
6983 RTUUID uuid;
6984
6985 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" cbSize=%llu uImageFlags=%#x pszComment=\"%s\" PCHS=%u/%u/%u LCHS=%u/%u/%u Uuid=%RTuuid uOpenFlags=%#x pVDIfsImage=%#p pVDIfsOperation=%#p\n",
6986 pDisk, pszBackend, pszFilename, cbSize, uImageFlags, pszComment,
6987 pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads,
6988 pPCHSGeometry->cSectors, pLCHSGeometry->cCylinders,
6989 pLCHSGeometry->cHeads, pLCHSGeometry->cSectors, pUuid,
6990 uOpenFlags, pVDIfsImage, pVDIfsOperation));
6991
6992 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
6993
6994 do
6995 {
6996 /* sanity check */
6997 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
6998 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
6999
7000 /* Check arguments. */
7001 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
7002 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
7003 rc = VERR_INVALID_PARAMETER);
7004 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
7005 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
7006 rc = VERR_INVALID_PARAMETER);
7007 AssertMsgBreakStmt(cbSize,
7008 ("cbSize=%llu\n", cbSize),
7009 rc = VERR_INVALID_PARAMETER);
7010 if (cbSize % 512)
7011 {
7012 rc = vdError(pDisk, VERR_VD_INVALID_SIZE, RT_SRC_POS,
7013 N_("VD: The given disk size %llu is not aligned on a sector boundary (512 bytes)"), cbSize);
7014 break;
7015 }
7016 AssertMsgBreakStmt( ((uImageFlags & ~VD_IMAGE_FLAGS_MASK) == 0)
7017 || ((uImageFlags & (VD_IMAGE_FLAGS_FIXED | VD_IMAGE_FLAGS_DIFF)) != VD_IMAGE_FLAGS_FIXED),
7018 ("uImageFlags=%#x\n", uImageFlags),
7019 rc = VERR_INVALID_PARAMETER);
7020 /* The PCHS geometry fields may be 0 to leave it for later. */
7021 AssertMsgBreakStmt( VALID_PTR(pPCHSGeometry)
7022 && pPCHSGeometry->cHeads <= 16
7023 && pPCHSGeometry->cSectors <= 63,
7024 ("pPCHSGeometry=%#p PCHS=%u/%u/%u\n", pPCHSGeometry,
7025 pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads,
7026 pPCHSGeometry->cSectors),
7027 rc = VERR_INVALID_PARAMETER);
7028 /* The LCHS geometry fields may be 0 to leave it to later autodetection. */
7029 AssertMsgBreakStmt( VALID_PTR(pLCHSGeometry)
7030 && pLCHSGeometry->cHeads <= 255
7031 && pLCHSGeometry->cSectors <= 63,
7032 ("pLCHSGeometry=%#p LCHS=%u/%u/%u\n", pLCHSGeometry,
7033 pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads,
7034 pLCHSGeometry->cSectors),
7035 rc = VERR_INVALID_PARAMETER);
7036 /* The UUID may be NULL. */
7037 AssertMsgBreakStmt(pUuid == NULL || VALID_PTR(pUuid),
7038 ("pUuid=%#p UUID=%RTuuid\n", pUuid, pUuid),
7039 rc = VERR_INVALID_PARAMETER);
7040 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
7041 ("uOpenFlags=%#x\n", uOpenFlags),
7042 rc = VERR_INVALID_PARAMETER);
7043
7044 /* Check state. Needs a temporary read lock. Holding the write lock
7045 * all the time would be blocking other activities for too long. */
7046 rc2 = vdThreadStartRead(pDisk);
7047 AssertRC(rc2);
7048 fLockRead = true;
7049 AssertMsgBreakStmt(pDisk->cImages == 0,
7050 ("Create base image cannot be done with other images open\n"),
7051 rc = VERR_VD_INVALID_STATE);
7052 rc2 = vdThreadFinishRead(pDisk);
7053 AssertRC(rc2);
7054 fLockRead = false;
7055
7056 /* Set up image descriptor. */
7057 pImage = (PVDIMAGE)RTMemAllocZ(sizeof(VDIMAGE));
7058 if (!pImage)
7059 {
7060 rc = VERR_NO_MEMORY;
7061 break;
7062 }
7063 pImage->pszFilename = RTStrDup(pszFilename);
7064 if (!pImage->pszFilename)
7065 {
7066 rc = VERR_NO_MEMORY;
7067 break;
7068 }
7069 pImage->VDIo.pDisk = pDisk;
7070 pImage->pVDIfsImage = pVDIfsImage;
7071
7072 /* Set up the I/O interface. */
7073 pImage->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsImage);
7074 if (!pImage->VDIo.pInterfaceIo)
7075 {
7076 vdIfIoFallbackCallbacksSetup(&pImage->VDIo.VDIfIo);
7077 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
7078 pDisk, sizeof(VDINTERFACEIO), &pVDIfsImage);
7079 pImage->VDIo.pInterfaceIo = &pImage->VDIo.VDIfIo;
7080 }
7081
7082 /* Set up the internal I/O interface. */
7083 AssertBreakStmt(!VDIfIoIntGet(pVDIfsImage), rc = VERR_INVALID_PARAMETER);
7084 vdIfIoIntCallbacksSetup(&pImage->VDIo.VDIfIoInt);
7085 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
7086 &pImage->VDIo, sizeof(VDINTERFACEIOINT), &pImage->pVDIfsImage);
7087 AssertRC(rc);
7088
7089 rc = vdFindBackend(pszBackend, &pImage->Backend);
7090 if (RT_FAILURE(rc))
7091 break;
7092 if (!pImage->Backend)
7093 {
7094 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
7095 N_("VD: unknown backend name '%s'"), pszBackend);
7096 break;
7097 }
7098 if (!(pImage->Backend->uBackendCaps & ( VD_CAP_CREATE_FIXED
7099 | VD_CAP_CREATE_DYNAMIC)))
7100 {
7101 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
7102 N_("VD: backend '%s' cannot create base images"), pszBackend);
7103 break;
7104 }
7105 if ( ( (uImageFlags & VD_VMDK_IMAGE_FLAGS_SPLIT_2G)
7106 && !(pImage->Backend->uBackendCaps & VD_CAP_CREATE_SPLIT_2G))
7107 || ( (uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
7108 && RTStrICmp(pszBackend, "VMDK")))
7109 {
7110 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
7111 N_("VD: backend '%s' does not support the selected image variant"), pszBackend);
7112 break;
7113 }
7114
7115 /* Create UUID if the caller didn't specify one. */
7116 if (!pUuid)
7117 {
7118 rc = RTUuidCreate(&uuid);
7119 if (RT_FAILURE(rc))
7120 {
7121 rc = vdError(pDisk, rc, RT_SRC_POS,
7122 N_("VD: cannot generate UUID for image '%s'"),
7123 pszFilename);
7124 break;
7125 }
7126 pUuid = &uuid;
7127 }
7128
7129 pImage->uOpenFlags = uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME;
7130 uImageFlags &= ~VD_IMAGE_FLAGS_DIFF;
7131 pImage->VDIo.fIgnoreFlush = (uOpenFlags & VD_OPEN_FLAGS_IGNORE_FLUSH) != 0;
7132 rc = pImage->Backend->pfnCreate(pImage->pszFilename, cbSize,
7133 uImageFlags, pszComment, pPCHSGeometry,
7134 pLCHSGeometry, pUuid,
7135 uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME,
7136 0, 99,
7137 pDisk->pVDIfsDisk,
7138 pImage->pVDIfsImage,
7139 pVDIfsOperation,
7140 pDisk->enmType,
7141 &pImage->pBackendData);
7142
7143 if (RT_SUCCESS(rc))
7144 {
7145 pImage->VDIo.pBackendData = pImage->pBackendData;
7146 pImage->uImageFlags = uImageFlags;
7147
7148 /* Force sane optimization settings. It's not worth avoiding writes
7149 * to fixed size images. The overhead would have almost no payback. */
7150 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
7151 pImage->uOpenFlags |= VD_OPEN_FLAGS_HONOR_SAME;
7152
7153 /* Lock disk for writing, as we modify pDisk information below. */
7154 rc2 = vdThreadStartWrite(pDisk);
7155 AssertRC(rc2);
7156 fLockWrite = true;
7157
7158 /** @todo optionally check UUIDs */
7159
7160 /* Re-check state, as the lock wasn't held and another image
7161 * creation call could have been done by another thread. */
7162 AssertMsgStmt(pDisk->cImages == 0,
7163 ("Create base image cannot be done with other images open\n"),
7164 rc = VERR_VD_INVALID_STATE);
7165 }
7166
7167 if (RT_SUCCESS(rc))
7168 {
7169 /* Cache disk information. */
7170 pDisk->cbSize = pImage->Backend->pfnGetSize(pImage->pBackendData);
7171
7172 /* Cache PCHS geometry. */
7173 rc2 = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
7174 &pDisk->PCHSGeometry);
7175 if (RT_FAILURE(rc2))
7176 {
7177 pDisk->PCHSGeometry.cCylinders = 0;
7178 pDisk->PCHSGeometry.cHeads = 0;
7179 pDisk->PCHSGeometry.cSectors = 0;
7180 }
7181 else
7182 {
7183 /* Make sure the CHS geometry is properly clipped. */
7184 pDisk->PCHSGeometry.cCylinders = RT_MIN(pDisk->PCHSGeometry.cCylinders, 16383);
7185 pDisk->PCHSGeometry.cHeads = RT_MIN(pDisk->PCHSGeometry.cHeads, 16);
7186 pDisk->PCHSGeometry.cSectors = RT_MIN(pDisk->PCHSGeometry.cSectors, 63);
7187 }
7188
7189 /* Cache LCHS geometry. */
7190 rc2 = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
7191 &pDisk->LCHSGeometry);
7192 if (RT_FAILURE(rc2))
7193 {
7194 pDisk->LCHSGeometry.cCylinders = 0;
7195 pDisk->LCHSGeometry.cHeads = 0;
7196 pDisk->LCHSGeometry.cSectors = 0;
7197 }
7198 else
7199 {
7200 /* Make sure the CHS geometry is properly clipped. */
7201 pDisk->LCHSGeometry.cHeads = RT_MIN(pDisk->LCHSGeometry.cHeads, 255);
7202 pDisk->LCHSGeometry.cSectors = RT_MIN(pDisk->LCHSGeometry.cSectors, 63);
7203 }
7204
7205 /* Image successfully opened, make it the last image. */
7206 vdAddImageToList(pDisk, pImage);
7207 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
7208 pDisk->uModified = VD_IMAGE_MODIFIED_FIRST;
7209 }
7210 else
7211 {
7212 /* Error detected, image may or may not be opened. Close and delete
7213 * image if it was opened. */
7214 if (pImage->pBackendData)
7215 {
7216 rc2 = pImage->Backend->pfnClose(pImage->pBackendData, true);
7217 AssertRC(rc2);
7218 pImage->pBackendData = NULL;
7219 }
7220 }
7221 } while (0);
7222
7223 if (RT_UNLIKELY(fLockWrite))
7224 {
7225 rc2 = vdThreadFinishWrite(pDisk);
7226 AssertRC(rc2);
7227 }
7228 else if (RT_UNLIKELY(fLockRead))
7229 {
7230 rc2 = vdThreadFinishRead(pDisk);
7231 AssertRC(rc2);
7232 }
7233
7234 if (RT_FAILURE(rc))
7235 {
7236 if (pImage)
7237 {
7238 if (pImage->pszFilename)
7239 RTStrFree(pImage->pszFilename);
7240 RTMemFree(pImage);
7241 }
7242 }
7243
7244 if (RT_SUCCESS(rc) && pIfProgress && pIfProgress->pfnProgress)
7245 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
7246
7247 LogFlowFunc(("returns %Rrc\n", rc));
7248 return rc;
7249}
7250
7251/**
7252 * Creates and opens a new differencing image file in HDD container.
7253 * See comments for VDOpen function about differencing images.
7254 *
7255 * @returns VBox status code.
7256 * @param pDisk Pointer to HDD container.
7257 * @param pszBackend Name of the image file backend to use.
7258 * @param pszFilename Name of the differencing image file to create.
7259 * @param uImageFlags Flags specifying special image features.
7260 * @param pszComment Pointer to image comment. NULL is ok.
7261 * @param pUuid New UUID of the image. If NULL, a new UUID is created.
7262 * @param pParentUuid New parent UUID of the image. If NULL, the UUID is queried automatically.
7263 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
7264 * @param pVDIfsImage Pointer to the per-image VD interface list.
7265 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
7266 */
7267VBOXDDU_DECL(int) VDCreateDiff(PVBOXHDD pDisk, const char *pszBackend,
7268 const char *pszFilename, unsigned uImageFlags,
7269 const char *pszComment, PCRTUUID pUuid,
7270 PCRTUUID pParentUuid, unsigned uOpenFlags,
7271 PVDINTERFACE pVDIfsImage,
7272 PVDINTERFACE pVDIfsOperation)
7273{
7274 int rc = VINF_SUCCESS;
7275 int rc2;
7276 bool fLockWrite = false, fLockRead = false;
7277 PVDIMAGE pImage = NULL;
7278 RTUUID uuid;
7279
7280 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" uImageFlags=%#x pszComment=\"%s\" Uuid=%RTuuid uOpenFlags=%#x pVDIfsImage=%#p pVDIfsOperation=%#p\n",
7281 pDisk, pszBackend, pszFilename, uImageFlags, pszComment, pUuid, uOpenFlags, pVDIfsImage, pVDIfsOperation));
7282
7283 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
7284
7285 do
7286 {
7287 /* sanity check */
7288 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
7289 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
7290
7291 /* Check arguments. */
7292 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
7293 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
7294 rc = VERR_INVALID_PARAMETER);
7295 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
7296 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
7297 rc = VERR_INVALID_PARAMETER);
7298 AssertMsgBreakStmt((uImageFlags & ~VD_IMAGE_FLAGS_MASK) == 0,
7299 ("uImageFlags=%#x\n", uImageFlags),
7300 rc = VERR_INVALID_PARAMETER);
7301 /* The UUID may be NULL. */
7302 AssertMsgBreakStmt(pUuid == NULL || VALID_PTR(pUuid),
7303 ("pUuid=%#p UUID=%RTuuid\n", pUuid, pUuid),
7304 rc = VERR_INVALID_PARAMETER);
7305 /* The parent UUID may be NULL. */
7306 AssertMsgBreakStmt(pParentUuid == NULL || VALID_PTR(pParentUuid),
7307 ("pParentUuid=%#p ParentUUID=%RTuuid\n", pParentUuid, pParentUuid),
7308 rc = VERR_INVALID_PARAMETER);
7309 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
7310 ("uOpenFlags=%#x\n", uOpenFlags),
7311 rc = VERR_INVALID_PARAMETER);
7312
7313 /* Check state. Needs a temporary read lock. Holding the write lock
7314 * all the time would be blocking other activities for too long. */
7315 rc2 = vdThreadStartRead(pDisk);
7316 AssertRC(rc2);
7317 fLockRead = true;
7318 AssertMsgBreakStmt(pDisk->cImages != 0,
7319 ("Create diff image cannot be done without other images open\n"),
7320 rc = VERR_VD_INVALID_STATE);
7321 rc2 = vdThreadFinishRead(pDisk);
7322 AssertRC(rc2);
7323 fLockRead = false;
7324
7325 /*
7326 * Destroy the current discard state first which might still have pending blocks
7327 * for the currently opened image which will be switched to readonly mode.
7328 */
7329 /* Lock disk for writing, as we modify pDisk information below. */
7330 rc2 = vdThreadStartWrite(pDisk);
7331 AssertRC(rc2);
7332 fLockWrite = true;
7333 rc = vdDiscardStateDestroy(pDisk);
7334 if (RT_FAILURE(rc))
7335 break;
7336 rc2 = vdThreadFinishWrite(pDisk);
7337 AssertRC(rc2);
7338 fLockWrite = false;
7339
7340 /* Set up image descriptor. */
7341 pImage = (PVDIMAGE)RTMemAllocZ(sizeof(VDIMAGE));
7342 if (!pImage)
7343 {
7344 rc = VERR_NO_MEMORY;
7345 break;
7346 }
7347 pImage->pszFilename = RTStrDup(pszFilename);
7348 if (!pImage->pszFilename)
7349 {
7350 rc = VERR_NO_MEMORY;
7351 break;
7352 }
7353
7354 rc = vdFindBackend(pszBackend, &pImage->Backend);
7355 if (RT_FAILURE(rc))
7356 break;
7357 if (!pImage->Backend)
7358 {
7359 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
7360 N_("VD: unknown backend name '%s'"), pszBackend);
7361 break;
7362 }
7363 if ( !(pImage->Backend->uBackendCaps & VD_CAP_DIFF)
7364 || !(pImage->Backend->uBackendCaps & ( VD_CAP_CREATE_FIXED
7365 | VD_CAP_CREATE_DYNAMIC)))
7366 {
7367 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
7368 N_("VD: backend '%s' cannot create diff images"), pszBackend);
7369 break;
7370 }
7371
7372 pImage->VDIo.pDisk = pDisk;
7373 pImage->pVDIfsImage = pVDIfsImage;
7374
7375 /* Set up the I/O interface. */
7376 pImage->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsImage);
7377 if (!pImage->VDIo.pInterfaceIo)
7378 {
7379 vdIfIoFallbackCallbacksSetup(&pImage->VDIo.VDIfIo);
7380 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
7381 pDisk, sizeof(VDINTERFACEIO), &pVDIfsImage);
7382 pImage->VDIo.pInterfaceIo = &pImage->VDIo.VDIfIo;
7383 }
7384
7385 /* Set up the internal I/O interface. */
7386 AssertBreakStmt(!VDIfIoIntGet(pVDIfsImage), rc = VERR_INVALID_PARAMETER);
7387 vdIfIoIntCallbacksSetup(&pImage->VDIo.VDIfIoInt);
7388 rc = VDInterfaceAdd(&pImage->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
7389 &pImage->VDIo, sizeof(VDINTERFACEIOINT), &pImage->pVDIfsImage);
7390 AssertRC(rc);
7391
7392 /* Create UUID if the caller didn't specify one. */
7393 if (!pUuid)
7394 {
7395 rc = RTUuidCreate(&uuid);
7396 if (RT_FAILURE(rc))
7397 {
7398 rc = vdError(pDisk, rc, RT_SRC_POS,
7399 N_("VD: cannot generate UUID for image '%s'"),
7400 pszFilename);
7401 break;
7402 }
7403 pUuid = &uuid;
7404 }
7405
7406 pImage->uOpenFlags = uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME;
7407 pImage->VDIo.fIgnoreFlush = (uOpenFlags & VD_OPEN_FLAGS_IGNORE_FLUSH) != 0;
7408 uImageFlags |= VD_IMAGE_FLAGS_DIFF;
7409 rc = pImage->Backend->pfnCreate(pImage->pszFilename, pDisk->cbSize,
7410 uImageFlags | VD_IMAGE_FLAGS_DIFF,
7411 pszComment, &pDisk->PCHSGeometry,
7412 &pDisk->LCHSGeometry, pUuid,
7413 uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME,
7414 0, 99,
7415 pDisk->pVDIfsDisk,
7416 pImage->pVDIfsImage,
7417 pVDIfsOperation,
7418 pDisk->enmType,
7419 &pImage->pBackendData);
7420
7421 if (RT_SUCCESS(rc))
7422 {
7423 pImage->VDIo.pBackendData = pImage->pBackendData;
7424 pImage->uImageFlags = uImageFlags;
7425
7426 /* Lock disk for writing, as we modify pDisk information below. */
7427 rc2 = vdThreadStartWrite(pDisk);
7428 AssertRC(rc2);
7429 fLockWrite = true;
7430
7431 /* Switch previous image to read-only mode. */
7432 unsigned uOpenFlagsPrevImg;
7433 uOpenFlagsPrevImg = pDisk->pLast->Backend->pfnGetOpenFlags(pDisk->pLast->pBackendData);
7434 if (!(uOpenFlagsPrevImg & VD_OPEN_FLAGS_READONLY))
7435 {
7436 uOpenFlagsPrevImg |= VD_OPEN_FLAGS_READONLY;
7437 rc = pDisk->pLast->Backend->pfnSetOpenFlags(pDisk->pLast->pBackendData, uOpenFlagsPrevImg);
7438 }
7439
7440 /** @todo optionally check UUIDs */
7441
7442 /* Re-check state, as the lock wasn't held and another image
7443 * creation call could have been done by another thread. */
7444 AssertMsgStmt(pDisk->cImages != 0,
7445 ("Create diff image cannot be done without other images open\n"),
7446 rc = VERR_VD_INVALID_STATE);
7447 }
7448
7449 if (RT_SUCCESS(rc))
7450 {
7451 RTUUID Uuid;
7452 RTTIMESPEC ts;
7453
7454 if (pParentUuid && !RTUuidIsNull(pParentUuid))
7455 {
7456 Uuid = *pParentUuid;
7457 pImage->Backend->pfnSetParentUuid(pImage->pBackendData, &Uuid);
7458 }
7459 else
7460 {
7461 rc2 = pDisk->pLast->Backend->pfnGetUuid(pDisk->pLast->pBackendData,
7462 &Uuid);
7463 if (RT_SUCCESS(rc2))
7464 pImage->Backend->pfnSetParentUuid(pImage->pBackendData, &Uuid);
7465 }
7466 rc2 = pDisk->pLast->Backend->pfnGetModificationUuid(pDisk->pLast->pBackendData,
7467 &Uuid);
7468 if (RT_SUCCESS(rc2))
7469 pImage->Backend->pfnSetParentModificationUuid(pImage->pBackendData,
7470 &Uuid);
7471 if (pDisk->pLast->Backend->pfnGetTimestamp)
7472 rc2 = pDisk->pLast->Backend->pfnGetTimestamp(pDisk->pLast->pBackendData,
7473 &ts);
7474 else
7475 rc2 = VERR_NOT_IMPLEMENTED;
7476 if (RT_SUCCESS(rc2) && pImage->Backend->pfnSetParentTimestamp)
7477 pImage->Backend->pfnSetParentTimestamp(pImage->pBackendData, &ts);
7478
7479 if (pImage->Backend->pfnSetParentFilename)
7480 rc2 = pImage->Backend->pfnSetParentFilename(pImage->pBackendData, pDisk->pLast->pszFilename);
7481 }
7482
7483 if (RT_SUCCESS(rc))
7484 {
7485 /* Image successfully opened, make it the last image. */
7486 vdAddImageToList(pDisk, pImage);
7487 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
7488 pDisk->uModified = VD_IMAGE_MODIFIED_FIRST;
7489 }
7490 else
7491 {
7492 /* Error detected, but image opened. Close and delete image. */
7493 rc2 = pImage->Backend->pfnClose(pImage->pBackendData, true);
7494 AssertRC(rc2);
7495 pImage->pBackendData = NULL;
7496 }
7497 } while (0);
7498
7499 if (RT_UNLIKELY(fLockWrite))
7500 {
7501 rc2 = vdThreadFinishWrite(pDisk);
7502 AssertRC(rc2);
7503 }
7504 else if (RT_UNLIKELY(fLockRead))
7505 {
7506 rc2 = vdThreadFinishRead(pDisk);
7507 AssertRC(rc2);
7508 }
7509
7510 if (RT_FAILURE(rc))
7511 {
7512 if (pImage)
7513 {
7514 if (pImage->pszFilename)
7515 RTStrFree(pImage->pszFilename);
7516 RTMemFree(pImage);
7517 }
7518 }
7519
7520 if (RT_SUCCESS(rc) && pIfProgress && pIfProgress->pfnProgress)
7521 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
7522
7523 LogFlowFunc(("returns %Rrc\n", rc));
7524 return rc;
7525}
7526
7527
7528/**
7529 * Creates and opens new cache image file in HDD container.
7530 *
7531 * @return VBox status code.
7532 * @param pDisk Name of the cache file backend to use (case insensitive).
7533 * @param pszFilename Name of the differencing cache file to create.
7534 * @param cbSize Maximum size of the cache.
7535 * @param uImageFlags Flags specifying special cache features.
7536 * @param pszComment Pointer to image comment. NULL is ok.
7537 * @param pUuid New UUID of the image. If NULL, a new UUID is created.
7538 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
7539 * @param pVDIfsCache Pointer to the per-cache VD interface list.
7540 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
7541 */
7542VBOXDDU_DECL(int) VDCreateCache(PVBOXHDD pDisk, const char *pszBackend,
7543 const char *pszFilename, uint64_t cbSize,
7544 unsigned uImageFlags, const char *pszComment,
7545 PCRTUUID pUuid, unsigned uOpenFlags,
7546 PVDINTERFACE pVDIfsCache, PVDINTERFACE pVDIfsOperation)
7547{
7548 int rc = VINF_SUCCESS;
7549 int rc2;
7550 bool fLockWrite = false, fLockRead = false;
7551 PVDCACHE pCache = NULL;
7552 RTUUID uuid;
7553
7554 LogFlowFunc(("pDisk=%#p pszBackend=\"%s\" pszFilename=\"%s\" cbSize=%llu uImageFlags=%#x pszComment=\"%s\" Uuid=%RTuuid uOpenFlags=%#x pVDIfsImage=%#p pVDIfsOperation=%#p\n",
7555 pDisk, pszBackend, pszFilename, cbSize, uImageFlags, pszComment, pUuid, uOpenFlags, pVDIfsCache, pVDIfsOperation));
7556
7557 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
7558
7559 do
7560 {
7561 /* sanity check */
7562 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
7563 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
7564
7565 /* Check arguments. */
7566 AssertMsgBreakStmt(VALID_PTR(pszBackend) && *pszBackend,
7567 ("pszBackend=%#p \"%s\"\n", pszBackend, pszBackend),
7568 rc = VERR_INVALID_PARAMETER);
7569 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
7570 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
7571 rc = VERR_INVALID_PARAMETER);
7572 AssertMsgBreakStmt(cbSize,
7573 ("cbSize=%llu\n", cbSize),
7574 rc = VERR_INVALID_PARAMETER);
7575 AssertMsgBreakStmt((uImageFlags & ~VD_IMAGE_FLAGS_MASK) == 0,
7576 ("uImageFlags=%#x\n", uImageFlags),
7577 rc = VERR_INVALID_PARAMETER);
7578 /* The UUID may be NULL. */
7579 AssertMsgBreakStmt(pUuid == NULL || VALID_PTR(pUuid),
7580 ("pUuid=%#p UUID=%RTuuid\n", pUuid, pUuid),
7581 rc = VERR_INVALID_PARAMETER);
7582 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
7583 ("uOpenFlags=%#x\n", uOpenFlags),
7584 rc = VERR_INVALID_PARAMETER);
7585
7586 /* Check state. Needs a temporary read lock. Holding the write lock
7587 * all the time would be blocking other activities for too long. */
7588 rc2 = vdThreadStartRead(pDisk);
7589 AssertRC(rc2);
7590 fLockRead = true;
7591 AssertMsgBreakStmt(!pDisk->pCache,
7592 ("Create cache image cannot be done with a cache already attached\n"),
7593 rc = VERR_VD_CACHE_ALREADY_EXISTS);
7594 rc2 = vdThreadFinishRead(pDisk);
7595 AssertRC(rc2);
7596 fLockRead = false;
7597
7598 /* Set up image descriptor. */
7599 pCache = (PVDCACHE)RTMemAllocZ(sizeof(VDCACHE));
7600 if (!pCache)
7601 {
7602 rc = VERR_NO_MEMORY;
7603 break;
7604 }
7605 pCache->pszFilename = RTStrDup(pszFilename);
7606 if (!pCache->pszFilename)
7607 {
7608 rc = VERR_NO_MEMORY;
7609 break;
7610 }
7611
7612 rc = vdFindCacheBackend(pszBackend, &pCache->Backend);
7613 if (RT_FAILURE(rc))
7614 break;
7615 if (!pCache->Backend)
7616 {
7617 rc = vdError(pDisk, VERR_INVALID_PARAMETER, RT_SRC_POS,
7618 N_("VD: unknown backend name '%s'"), pszBackend);
7619 break;
7620 }
7621
7622 pCache->VDIo.pDisk = pDisk;
7623 pCache->pVDIfsCache = pVDIfsCache;
7624
7625 /* Set up the I/O interface. */
7626 pCache->VDIo.pInterfaceIo = VDIfIoGet(pVDIfsCache);
7627 if (!pCache->VDIo.pInterfaceIo)
7628 {
7629 vdIfIoFallbackCallbacksSetup(&pCache->VDIo.VDIfIo);
7630 rc = VDInterfaceAdd(&pCache->VDIo.VDIfIo.Core, "VD_IO", VDINTERFACETYPE_IO,
7631 pDisk, sizeof(VDINTERFACEIO), &pVDIfsCache);
7632 pCache->VDIo.pInterfaceIo = &pCache->VDIo.VDIfIo;
7633 }
7634
7635 /* Set up the internal I/O interface. */
7636 AssertBreakStmt(!VDIfIoIntGet(pVDIfsCache), rc = VERR_INVALID_PARAMETER);
7637 vdIfIoIntCallbacksSetup(&pCache->VDIo.VDIfIoInt);
7638 rc = VDInterfaceAdd(&pCache->VDIo.VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
7639 &pCache->VDIo, sizeof(VDINTERFACEIOINT), &pCache->pVDIfsCache);
7640 AssertRC(rc);
7641
7642 /* Create UUID if the caller didn't specify one. */
7643 if (!pUuid)
7644 {
7645 rc = RTUuidCreate(&uuid);
7646 if (RT_FAILURE(rc))
7647 {
7648 rc = vdError(pDisk, rc, RT_SRC_POS,
7649 N_("VD: cannot generate UUID for image '%s'"),
7650 pszFilename);
7651 break;
7652 }
7653 pUuid = &uuid;
7654 }
7655
7656 pCache->uOpenFlags = uOpenFlags & VD_OPEN_FLAGS_HONOR_SAME;
7657 pCache->VDIo.fIgnoreFlush = (uOpenFlags & VD_OPEN_FLAGS_IGNORE_FLUSH) != 0;
7658 rc = pCache->Backend->pfnCreate(pCache->pszFilename, cbSize,
7659 uImageFlags,
7660 pszComment, pUuid,
7661 uOpenFlags & ~VD_OPEN_FLAGS_HONOR_SAME,
7662 0, 99,
7663 pDisk->pVDIfsDisk,
7664 pCache->pVDIfsCache,
7665 pVDIfsOperation,
7666 &pCache->pBackendData);
7667
7668 if (RT_SUCCESS(rc))
7669 {
7670 /* Lock disk for writing, as we modify pDisk information below. */
7671 rc2 = vdThreadStartWrite(pDisk);
7672 AssertRC(rc2);
7673 fLockWrite = true;
7674
7675 pCache->VDIo.pBackendData = pCache->pBackendData;
7676
7677 /* Re-check state, as the lock wasn't held and another image
7678 * creation call could have been done by another thread. */
7679 AssertMsgStmt(!pDisk->pCache,
7680 ("Create cache image cannot be done with another cache open\n"),
7681 rc = VERR_VD_CACHE_ALREADY_EXISTS);
7682 }
7683
7684 if ( RT_SUCCESS(rc)
7685 && pDisk->pLast)
7686 {
7687 RTUUID UuidModification;
7688
7689 /* Set same modification Uuid as the last image. */
7690 rc = pDisk->pLast->Backend->pfnGetModificationUuid(pDisk->pLast->pBackendData,
7691 &UuidModification);
7692 if (RT_SUCCESS(rc))
7693 {
7694 rc = pCache->Backend->pfnSetModificationUuid(pCache->pBackendData,
7695 &UuidModification);
7696 }
7697
7698 if (rc == VERR_NOT_SUPPORTED)
7699 rc = VINF_SUCCESS;
7700 }
7701
7702 if (RT_SUCCESS(rc))
7703 {
7704 /* Cache successfully created. */
7705 pDisk->pCache = pCache;
7706 }
7707 else
7708 {
7709 /* Error detected, but image opened. Close and delete image. */
7710 rc2 = pCache->Backend->pfnClose(pCache->pBackendData, true);
7711 AssertRC(rc2);
7712 pCache->pBackendData = NULL;
7713 }
7714 } while (0);
7715
7716 if (RT_UNLIKELY(fLockWrite))
7717 {
7718 rc2 = vdThreadFinishWrite(pDisk);
7719 AssertRC(rc2);
7720 }
7721 else if (RT_UNLIKELY(fLockRead))
7722 {
7723 rc2 = vdThreadFinishRead(pDisk);
7724 AssertRC(rc2);
7725 }
7726
7727 if (RT_FAILURE(rc))
7728 {
7729 if (pCache)
7730 {
7731 if (pCache->pszFilename)
7732 RTStrFree(pCache->pszFilename);
7733 RTMemFree(pCache);
7734 }
7735 }
7736
7737 if (RT_SUCCESS(rc) && pIfProgress && pIfProgress->pfnProgress)
7738 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
7739
7740 LogFlowFunc(("returns %Rrc\n", rc));
7741 return rc;
7742}
7743
7744/**
7745 * Merges two images (not necessarily with direct parent/child relationship).
7746 * As a side effect the source image and potentially the other images which
7747 * are also merged to the destination are deleted from both the disk and the
7748 * images in the HDD container.
7749 *
7750 * @returns VBox status code.
7751 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
7752 * @param pDisk Pointer to HDD container.
7753 * @param nImageFrom Name of the image file to merge from.
7754 * @param nImageTo Name of the image file to merge to.
7755 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
7756 */
7757VBOXDDU_DECL(int) VDMerge(PVBOXHDD pDisk, unsigned nImageFrom,
7758 unsigned nImageTo, PVDINTERFACE pVDIfsOperation)
7759{
7760 int rc = VINF_SUCCESS;
7761 int rc2;
7762 bool fLockWrite = false, fLockRead = false;
7763 void *pvBuf = NULL;
7764
7765 LogFlowFunc(("pDisk=%#p nImageFrom=%u nImageTo=%u pVDIfsOperation=%#p\n",
7766 pDisk, nImageFrom, nImageTo, pVDIfsOperation));
7767
7768 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
7769
7770 do
7771 {
7772 /* sanity check */
7773 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
7774 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
7775
7776 /* For simplicity reasons lock for writing as the image reopen below
7777 * might need it. After all the reopen is usually needed. */
7778 rc2 = vdThreadStartWrite(pDisk);
7779 AssertRC(rc2);
7780 fLockWrite = true;
7781 PVDIMAGE pImageFrom = vdGetImageByNumber(pDisk, nImageFrom);
7782 PVDIMAGE pImageTo = vdGetImageByNumber(pDisk, nImageTo);
7783 if (!pImageFrom || !pImageTo)
7784 {
7785 rc = VERR_VD_IMAGE_NOT_FOUND;
7786 break;
7787 }
7788 AssertBreakStmt(pImageFrom != pImageTo, rc = VERR_INVALID_PARAMETER);
7789
7790 /* Make sure destination image is writable. */
7791 unsigned uOpenFlags = pImageTo->Backend->pfnGetOpenFlags(pImageTo->pBackendData);
7792 if (uOpenFlags & VD_OPEN_FLAGS_READONLY)
7793 {
7794 /*
7795 * Clear skip consistency checks because the image is made writable now and
7796 * skipping consistency checks is only possible for readonly images.
7797 */
7798 uOpenFlags &= ~(VD_OPEN_FLAGS_READONLY | VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS);
7799 rc = pImageTo->Backend->pfnSetOpenFlags(pImageTo->pBackendData,
7800 uOpenFlags);
7801 if (RT_FAILURE(rc))
7802 break;
7803 }
7804
7805 /* Get size of destination image. */
7806 uint64_t cbSize = pImageTo->Backend->pfnGetSize(pImageTo->pBackendData);
7807 rc2 = vdThreadFinishWrite(pDisk);
7808 AssertRC(rc2);
7809 fLockWrite = false;
7810
7811 /* Allocate tmp buffer. */
7812 pvBuf = RTMemTmpAlloc(VD_MERGE_BUFFER_SIZE);
7813 if (!pvBuf)
7814 {
7815 rc = VERR_NO_MEMORY;
7816 break;
7817 }
7818
7819 /* Merging is done directly on the images itself. This potentially
7820 * causes trouble if the disk is full in the middle of operation. */
7821 if (nImageFrom < nImageTo)
7822 {
7823 /* Merge parent state into child. This means writing all not
7824 * allocated blocks in the destination image which are allocated in
7825 * the images to be merged. */
7826 uint64_t uOffset = 0;
7827 uint64_t cbRemaining = cbSize;
7828 do
7829 {
7830 size_t cbThisRead = RT_MIN(VD_MERGE_BUFFER_SIZE, cbRemaining);
7831 RTSGSEG SegmentBuf;
7832 RTSGBUF SgBuf;
7833 VDIOCTX IoCtx;
7834
7835 SegmentBuf.pvSeg = pvBuf;
7836 SegmentBuf.cbSeg = VD_MERGE_BUFFER_SIZE;
7837 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
7838 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_READ, 0, 0, NULL,
7839 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
7840
7841 /* Need to hold the write lock during a read-write operation. */
7842 rc2 = vdThreadStartWrite(pDisk);
7843 AssertRC(rc2);
7844 fLockWrite = true;
7845
7846 rc = pImageTo->Backend->pfnRead(pImageTo->pBackendData,
7847 uOffset, cbThisRead,
7848 &IoCtx, &cbThisRead);
7849 if (rc == VERR_VD_BLOCK_FREE)
7850 {
7851 /* Search for image with allocated block. Do not attempt to
7852 * read more than the previous reads marked as valid.
7853 * Otherwise this would return stale data when different
7854 * block sizes are used for the images. */
7855 for (PVDIMAGE pCurrImage = pImageTo->pPrev;
7856 pCurrImage != NULL && pCurrImage != pImageFrom->pPrev && rc == VERR_VD_BLOCK_FREE;
7857 pCurrImage = pCurrImage->pPrev)
7858 {
7859 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
7860 uOffset, cbThisRead,
7861 &IoCtx, &cbThisRead);
7862 }
7863
7864 if (rc != VERR_VD_BLOCK_FREE)
7865 {
7866 if (RT_FAILURE(rc))
7867 break;
7868 /* Updating the cache is required because this might be a live merge. */
7869 rc = vdWriteHelperEx(pDisk, pImageTo, pImageFrom->pPrev,
7870 uOffset, pvBuf, cbThisRead,
7871 VDIOCTX_FLAGS_READ_UPDATE_CACHE, 0);
7872 if (RT_FAILURE(rc))
7873 break;
7874 }
7875 else
7876 rc = VINF_SUCCESS;
7877 }
7878 else if (RT_FAILURE(rc))
7879 break;
7880
7881 rc2 = vdThreadFinishWrite(pDisk);
7882 AssertRC(rc2);
7883 fLockWrite = false;
7884
7885 uOffset += cbThisRead;
7886 cbRemaining -= cbThisRead;
7887
7888 if (pIfProgress && pIfProgress->pfnProgress)
7889 {
7890 /** @todo r=klaus: this can update the progress to the same
7891 * percentage over and over again if the image format makes
7892 * relatively small increments. */
7893 rc = pIfProgress->pfnProgress(pIfProgress->Core.pvUser,
7894 uOffset * 99 / cbSize);
7895 if (RT_FAILURE(rc))
7896 break;
7897 }
7898 } while (uOffset < cbSize);
7899 }
7900 else
7901 {
7902 /*
7903 * We may need to update the parent uuid of the child coming after
7904 * the last image to be merged. We have to reopen it read/write.
7905 *
7906 * This is done before we do the actual merge to prevent an
7907 * inconsistent chain if the mode change fails for some reason.
7908 */
7909 if (pImageFrom->pNext)
7910 {
7911 PVDIMAGE pImageChild = pImageFrom->pNext;
7912
7913 /* Take the write lock. */
7914 rc2 = vdThreadStartWrite(pDisk);
7915 AssertRC(rc2);
7916 fLockWrite = true;
7917
7918 /* We need to open the image in read/write mode. */
7919 uOpenFlags = pImageChild->Backend->pfnGetOpenFlags(pImageChild->pBackendData);
7920
7921 if (uOpenFlags & VD_OPEN_FLAGS_READONLY)
7922 {
7923 uOpenFlags &= ~VD_OPEN_FLAGS_READONLY;
7924 rc = pImageChild->Backend->pfnSetOpenFlags(pImageChild->pBackendData,
7925 uOpenFlags);
7926 if (RT_FAILURE(rc))
7927 break;
7928 }
7929
7930 rc2 = vdThreadFinishWrite(pDisk);
7931 AssertRC(rc2);
7932 fLockWrite = false;
7933 }
7934
7935 /* If the merge is from the last image we have to relay all writes
7936 * to the merge destination as well, so that concurrent writes
7937 * (in case of a live merge) are handled correctly. */
7938 if (!pImageFrom->pNext)
7939 {
7940 /* Take the write lock. */
7941 rc2 = vdThreadStartWrite(pDisk);
7942 AssertRC(rc2);
7943 fLockWrite = true;
7944
7945 pDisk->pImageRelay = pImageTo;
7946
7947 rc2 = vdThreadFinishWrite(pDisk);
7948 AssertRC(rc2);
7949 fLockWrite = false;
7950 }
7951
7952 /* Merge child state into parent. This means writing all blocks
7953 * which are allocated in the image up to the source image to the
7954 * destination image. */
7955 uint64_t uOffset = 0;
7956 uint64_t cbRemaining = cbSize;
7957 do
7958 {
7959 size_t cbThisRead = RT_MIN(VD_MERGE_BUFFER_SIZE, cbRemaining);
7960 RTSGSEG SegmentBuf;
7961 RTSGBUF SgBuf;
7962 VDIOCTX IoCtx;
7963
7964 rc = VERR_VD_BLOCK_FREE;
7965
7966 SegmentBuf.pvSeg = pvBuf;
7967 SegmentBuf.cbSeg = VD_MERGE_BUFFER_SIZE;
7968 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
7969 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_READ, 0, 0, NULL,
7970 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
7971
7972 /* Need to hold the write lock during a read-write operation. */
7973 rc2 = vdThreadStartWrite(pDisk);
7974 AssertRC(rc2);
7975 fLockWrite = true;
7976
7977 /* Search for image with allocated block. Do not attempt to
7978 * read more than the previous reads marked as valid. Otherwise
7979 * this would return stale data when different block sizes are
7980 * used for the images. */
7981 for (PVDIMAGE pCurrImage = pImageFrom;
7982 pCurrImage != NULL && pCurrImage != pImageTo && rc == VERR_VD_BLOCK_FREE;
7983 pCurrImage = pCurrImage->pPrev)
7984 {
7985 rc = pCurrImage->Backend->pfnRead(pCurrImage->pBackendData,
7986 uOffset, cbThisRead,
7987 &IoCtx, &cbThisRead);
7988 }
7989
7990 if (rc != VERR_VD_BLOCK_FREE)
7991 {
7992 if (RT_FAILURE(rc))
7993 break;
7994 rc = vdWriteHelper(pDisk, pImageTo, uOffset, pvBuf,
7995 cbThisRead, VDIOCTX_FLAGS_READ_UPDATE_CACHE);
7996 if (RT_FAILURE(rc))
7997 break;
7998 }
7999 else
8000 rc = VINF_SUCCESS;
8001
8002 rc2 = vdThreadFinishWrite(pDisk);
8003 AssertRC(rc2);
8004 fLockWrite = false;
8005
8006 uOffset += cbThisRead;
8007 cbRemaining -= cbThisRead;
8008
8009 if (pIfProgress && pIfProgress->pfnProgress)
8010 {
8011 /** @todo r=klaus: this can update the progress to the same
8012 * percentage over and over again if the image format makes
8013 * relatively small increments. */
8014 rc = pIfProgress->pfnProgress(pIfProgress->Core.pvUser,
8015 uOffset * 99 / cbSize);
8016 if (RT_FAILURE(rc))
8017 break;
8018 }
8019 } while (uOffset < cbSize);
8020
8021 /* In case we set up a "write proxy" image above we must clear
8022 * this again now to prevent stray writes. Failure or not. */
8023 if (!pImageFrom->pNext)
8024 {
8025 /* Take the write lock. */
8026 rc2 = vdThreadStartWrite(pDisk);
8027 AssertRC(rc2);
8028 fLockWrite = true;
8029
8030 pDisk->pImageRelay = NULL;
8031
8032 rc2 = vdThreadFinishWrite(pDisk);
8033 AssertRC(rc2);
8034 fLockWrite = false;
8035 }
8036 }
8037
8038 /*
8039 * Leave in case of an error to avoid corrupted data in the image chain
8040 * (includes cancelling the operation by the user).
8041 */
8042 if (RT_FAILURE(rc))
8043 break;
8044
8045 /* Need to hold the write lock while finishing the merge. */
8046 rc2 = vdThreadStartWrite(pDisk);
8047 AssertRC(rc2);
8048 fLockWrite = true;
8049
8050 /* Update parent UUID so that image chain is consistent.
8051 * The two attempts work around the problem that some backends
8052 * (e.g. iSCSI) do not support UUIDs, so we exploit the fact that
8053 * so far there can only be one such image in the chain. */
8054 /** @todo needs a better long-term solution, passing the UUID
8055 * knowledge from the caller or some such */
8056 RTUUID Uuid;
8057 PVDIMAGE pImageChild = NULL;
8058 if (nImageFrom < nImageTo)
8059 {
8060 if (pImageFrom->pPrev)
8061 {
8062 /* plan A: ask the parent itself for its UUID */
8063 rc = pImageFrom->pPrev->Backend->pfnGetUuid(pImageFrom->pPrev->pBackendData,
8064 &Uuid);
8065 if (RT_FAILURE(rc))
8066 {
8067 /* plan B: ask the child of the parent for parent UUID */
8068 rc = pImageFrom->Backend->pfnGetParentUuid(pImageFrom->pBackendData,
8069 &Uuid);
8070 }
8071 AssertRC(rc);
8072 }
8073 else
8074 RTUuidClear(&Uuid);
8075 rc = pImageTo->Backend->pfnSetParentUuid(pImageTo->pBackendData,
8076 &Uuid);
8077 AssertRC(rc);
8078 }
8079 else
8080 {
8081 /* Update the parent uuid of the child of the last merged image. */
8082 if (pImageFrom->pNext)
8083 {
8084 /* plan A: ask the parent itself for its UUID */
8085 rc = pImageTo->Backend->pfnGetUuid(pImageTo->pBackendData,
8086 &Uuid);
8087 if (RT_FAILURE(rc))
8088 {
8089 /* plan B: ask the child of the parent for parent UUID */
8090 rc = pImageTo->pNext->Backend->pfnGetParentUuid(pImageTo->pNext->pBackendData,
8091 &Uuid);
8092 }
8093 AssertRC(rc);
8094
8095 rc = pImageFrom->Backend->pfnSetParentUuid(pImageFrom->pNext->pBackendData,
8096 &Uuid);
8097 AssertRC(rc);
8098
8099 pImageChild = pImageFrom->pNext;
8100 }
8101 }
8102
8103 /* Delete the no longer needed images. */
8104 PVDIMAGE pImg = pImageFrom, pTmp;
8105 while (pImg != pImageTo)
8106 {
8107 if (nImageFrom < nImageTo)
8108 pTmp = pImg->pNext;
8109 else
8110 pTmp = pImg->pPrev;
8111 vdRemoveImageFromList(pDisk, pImg);
8112 pImg->Backend->pfnClose(pImg->pBackendData, true);
8113 RTMemFree(pImg->pszFilename);
8114 RTMemFree(pImg);
8115 pImg = pTmp;
8116 }
8117
8118 /* Make sure destination image is back to read only if necessary. */
8119 if (pImageTo != pDisk->pLast)
8120 {
8121 uOpenFlags = pImageTo->Backend->pfnGetOpenFlags(pImageTo->pBackendData);
8122 uOpenFlags |= VD_OPEN_FLAGS_READONLY;
8123 rc = pImageTo->Backend->pfnSetOpenFlags(pImageTo->pBackendData,
8124 uOpenFlags);
8125 if (RT_FAILURE(rc))
8126 break;
8127 }
8128
8129 /*
8130 * Make sure the child is readonly
8131 * for the child -> parent merge direction
8132 * if necessary.
8133 */
8134 if ( nImageFrom > nImageTo
8135 && pImageChild
8136 && pImageChild != pDisk->pLast)
8137 {
8138 uOpenFlags = pImageChild->Backend->pfnGetOpenFlags(pImageChild->pBackendData);
8139 uOpenFlags |= VD_OPEN_FLAGS_READONLY;
8140 rc = pImageChild->Backend->pfnSetOpenFlags(pImageChild->pBackendData,
8141 uOpenFlags);
8142 if (RT_FAILURE(rc))
8143 break;
8144 }
8145 } while (0);
8146
8147 if (RT_UNLIKELY(fLockWrite))
8148 {
8149 rc2 = vdThreadFinishWrite(pDisk);
8150 AssertRC(rc2);
8151 }
8152 else if (RT_UNLIKELY(fLockRead))
8153 {
8154 rc2 = vdThreadFinishRead(pDisk);
8155 AssertRC(rc2);
8156 }
8157
8158 if (pvBuf)
8159 RTMemTmpFree(pvBuf);
8160
8161 if (RT_SUCCESS(rc) && pIfProgress && pIfProgress->pfnProgress)
8162 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
8163
8164 LogFlowFunc(("returns %Rrc\n", rc));
8165 return rc;
8166}
8167
8168/**
8169 * Copies an image from one HDD container to another - extended version.
8170 * The copy is opened in the target HDD container.
8171 * It is possible to convert between different image formats, because the
8172 * backend for the destination may be different from the source.
8173 * If both the source and destination reference the same HDD container,
8174 * then the image is moved (by copying/deleting or renaming) to the new location.
8175 * The source container is unchanged if the move operation fails, otherwise
8176 * the image at the new location is opened in the same way as the old one was.
8177 *
8178 * @note The read/write accesses across disks are not synchronized, just the
8179 * accesses to each disk. Once there is a use case which requires a defined
8180 * read/write behavior in this situation this needs to be extended.
8181 *
8182 * @returns VBox status code.
8183 * @retval VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8184 * @param pDiskFrom Pointer to source HDD container.
8185 * @param nImage Image number, counts from 0. 0 is always base image of container.
8186 * @param pDiskTo Pointer to destination HDD container.
8187 * @param pszBackend Name of the image file backend to use (may be NULL to use the same as the source, case insensitive).
8188 * @param pszFilename New name of the image (may be NULL to specify that the
8189 * copy destination is the destination container, or
8190 * if pDiskFrom == pDiskTo, i.e. when moving).
8191 * @param fMoveByRename If true, attempt to perform a move by renaming (if successful the new size is ignored).
8192 * @param cbSize New image size (0 means leave unchanged).
8193 * @param nImageFromSame todo
8194 * @param nImageToSame todo
8195 * @param uImageFlags Flags specifying special destination image features.
8196 * @param pDstUuid New UUID of the destination image. If NULL, a new UUID is created.
8197 * This parameter is used if and only if a true copy is created.
8198 * In all rename/move cases or copy to existing image cases the modification UUIDs are copied over.
8199 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
8200 * Only used if the destination image is created.
8201 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
8202 * @param pDstVDIfsImage Pointer to the per-image VD interface list, for the
8203 * destination image.
8204 * @param pDstVDIfsOperation Pointer to the per-operation VD interface list,
8205 * for the destination operation.
8206 */
8207VBOXDDU_DECL(int) VDCopyEx(PVBOXHDD pDiskFrom, unsigned nImage, PVBOXHDD pDiskTo,
8208 const char *pszBackend, const char *pszFilename,
8209 bool fMoveByRename, uint64_t cbSize,
8210 unsigned nImageFromSame, unsigned nImageToSame,
8211 unsigned uImageFlags, PCRTUUID pDstUuid,
8212 unsigned uOpenFlags, PVDINTERFACE pVDIfsOperation,
8213 PVDINTERFACE pDstVDIfsImage,
8214 PVDINTERFACE pDstVDIfsOperation)
8215{
8216 int rc = VINF_SUCCESS;
8217 int rc2;
8218 bool fLockReadFrom = false, fLockWriteFrom = false, fLockWriteTo = false;
8219 PVDIMAGE pImageTo = NULL;
8220
8221 LogFlowFunc(("pDiskFrom=%#p nImage=%u pDiskTo=%#p pszBackend=\"%s\" pszFilename=\"%s\" fMoveByRename=%d cbSize=%llu nImageFromSame=%u nImageToSame=%u uImageFlags=%#x pDstUuid=%#p uOpenFlags=%#x pVDIfsOperation=%#p pDstVDIfsImage=%#p pDstVDIfsOperation=%#p\n",
8222 pDiskFrom, nImage, pDiskTo, pszBackend, pszFilename, fMoveByRename, cbSize, nImageFromSame, nImageToSame, uImageFlags, pDstUuid, uOpenFlags, pVDIfsOperation, pDstVDIfsImage, pDstVDIfsOperation));
8223
8224 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
8225 PVDINTERFACEPROGRESS pDstIfProgress = VDIfProgressGet(pDstVDIfsOperation);
8226
8227 do {
8228 /* Check arguments. */
8229 AssertMsgBreakStmt(VALID_PTR(pDiskFrom), ("pDiskFrom=%#p\n", pDiskFrom),
8230 rc = VERR_INVALID_PARAMETER);
8231 AssertMsg(pDiskFrom->u32Signature == VBOXHDDDISK_SIGNATURE,
8232 ("u32Signature=%08x\n", pDiskFrom->u32Signature));
8233
8234 rc2 = vdThreadStartRead(pDiskFrom);
8235 AssertRC(rc2);
8236 fLockReadFrom = true;
8237 PVDIMAGE pImageFrom = vdGetImageByNumber(pDiskFrom, nImage);
8238 AssertPtrBreakStmt(pImageFrom, rc = VERR_VD_IMAGE_NOT_FOUND);
8239 AssertMsgBreakStmt(VALID_PTR(pDiskTo), ("pDiskTo=%#p\n", pDiskTo),
8240 rc = VERR_INVALID_PARAMETER);
8241 AssertMsg(pDiskTo->u32Signature == VBOXHDDDISK_SIGNATURE,
8242 ("u32Signature=%08x\n", pDiskTo->u32Signature));
8243 AssertMsgBreakStmt( (nImageFromSame < nImage || nImageFromSame == VD_IMAGE_CONTENT_UNKNOWN)
8244 && (nImageToSame < pDiskTo->cImages || nImageToSame == VD_IMAGE_CONTENT_UNKNOWN)
8245 && ( (nImageFromSame == VD_IMAGE_CONTENT_UNKNOWN && nImageToSame == VD_IMAGE_CONTENT_UNKNOWN)
8246 || (nImageFromSame != VD_IMAGE_CONTENT_UNKNOWN && nImageToSame != VD_IMAGE_CONTENT_UNKNOWN)),
8247 ("nImageFromSame=%u nImageToSame=%u\n", nImageFromSame, nImageToSame),
8248 rc = VERR_INVALID_PARAMETER);
8249
8250 /* Move the image. */
8251 if (pDiskFrom == pDiskTo)
8252 {
8253 /* Rename only works when backends are the same, are file based
8254 * and the rename method is implemented. */
8255 if ( fMoveByRename
8256 && !RTStrICmp(pszBackend, pImageFrom->Backend->pszBackendName)
8257 && pImageFrom->Backend->uBackendCaps & VD_CAP_FILE
8258 && pImageFrom->Backend->pfnRename)
8259 {
8260 rc2 = vdThreadFinishRead(pDiskFrom);
8261 AssertRC(rc2);
8262 fLockReadFrom = false;
8263
8264 rc2 = vdThreadStartWrite(pDiskFrom);
8265 AssertRC(rc2);
8266 fLockWriteFrom = true;
8267 rc = pImageFrom->Backend->pfnRename(pImageFrom->pBackendData, pszFilename ? pszFilename : pImageFrom->pszFilename);
8268 break;
8269 }
8270
8271 /** @todo Moving (including shrinking/growing) of the image is
8272 * requested, but the rename attempt failed or it wasn't possible.
8273 * Must now copy image to temp location. */
8274 AssertReleaseMsgFailed(("VDCopy: moving by copy/delete not implemented\n"));
8275 }
8276
8277 /* pszFilename is allowed to be NULL, as this indicates copy to the existing image. */
8278 AssertMsgBreakStmt(pszFilename == NULL || (VALID_PTR(pszFilename) && *pszFilename),
8279 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
8280 rc = VERR_INVALID_PARAMETER);
8281
8282 uint64_t cbSizeFrom;
8283 cbSizeFrom = pImageFrom->Backend->pfnGetSize(pImageFrom->pBackendData);
8284 if (cbSizeFrom == 0)
8285 {
8286 rc = VERR_VD_VALUE_NOT_FOUND;
8287 break;
8288 }
8289
8290 VDGEOMETRY PCHSGeometryFrom = {0, 0, 0};
8291 VDGEOMETRY LCHSGeometryFrom = {0, 0, 0};
8292 pImageFrom->Backend->pfnGetPCHSGeometry(pImageFrom->pBackendData, &PCHSGeometryFrom);
8293 pImageFrom->Backend->pfnGetLCHSGeometry(pImageFrom->pBackendData, &LCHSGeometryFrom);
8294
8295 RTUUID ImageUuid, ImageModificationUuid;
8296 if (pDiskFrom != pDiskTo)
8297 {
8298 if (pDstUuid)
8299 ImageUuid = *pDstUuid;
8300 else
8301 RTUuidCreate(&ImageUuid);
8302 }
8303 else
8304 {
8305 rc = pImageFrom->Backend->pfnGetUuid(pImageFrom->pBackendData, &ImageUuid);
8306 if (RT_FAILURE(rc))
8307 RTUuidCreate(&ImageUuid);
8308 }
8309 rc = pImageFrom->Backend->pfnGetModificationUuid(pImageFrom->pBackendData, &ImageModificationUuid);
8310 if (RT_FAILURE(rc))
8311 RTUuidClear(&ImageModificationUuid);
8312
8313 char szComment[1024];
8314 rc = pImageFrom->Backend->pfnGetComment(pImageFrom->pBackendData, szComment, sizeof(szComment));
8315 if (RT_FAILURE(rc))
8316 szComment[0] = '\0';
8317 else
8318 szComment[sizeof(szComment) - 1] = '\0';
8319
8320 rc2 = vdThreadFinishRead(pDiskFrom);
8321 AssertRC(rc2);
8322 fLockReadFrom = false;
8323
8324 rc2 = vdThreadStartRead(pDiskTo);
8325 AssertRC(rc2);
8326 unsigned cImagesTo = pDiskTo->cImages;
8327 rc2 = vdThreadFinishRead(pDiskTo);
8328 AssertRC(rc2);
8329
8330 if (pszFilename)
8331 {
8332 if (cbSize == 0)
8333 cbSize = cbSizeFrom;
8334
8335 /* Create destination image with the properties of source image. */
8336 /** @todo replace the VDCreateDiff/VDCreateBase calls by direct
8337 * calls to the backend. Unifies the code and reduces the API
8338 * dependencies. Would also make the synchronization explicit. */
8339 if (cImagesTo > 0)
8340 {
8341 rc = VDCreateDiff(pDiskTo, pszBackend, pszFilename,
8342 uImageFlags, szComment, &ImageUuid,
8343 NULL /* pParentUuid */,
8344 uOpenFlags & ~VD_OPEN_FLAGS_READONLY,
8345 pDstVDIfsImage, NULL);
8346
8347 rc2 = vdThreadStartWrite(pDiskTo);
8348 AssertRC(rc2);
8349 fLockWriteTo = true;
8350 } else {
8351 /** @todo hack to force creation of a fixed image for
8352 * the RAW backend, which can't handle anything else. */
8353 if (!RTStrICmp(pszBackend, "RAW"))
8354 uImageFlags |= VD_IMAGE_FLAGS_FIXED;
8355
8356 vdFixupPCHSGeometry(&PCHSGeometryFrom, cbSize);
8357 vdFixupLCHSGeometry(&LCHSGeometryFrom, cbSize);
8358
8359 rc = VDCreateBase(pDiskTo, pszBackend, pszFilename, cbSize,
8360 uImageFlags, szComment,
8361 &PCHSGeometryFrom, &LCHSGeometryFrom,
8362 NULL, uOpenFlags & ~VD_OPEN_FLAGS_READONLY,
8363 pDstVDIfsImage, NULL);
8364
8365 rc2 = vdThreadStartWrite(pDiskTo);
8366 AssertRC(rc2);
8367 fLockWriteTo = true;
8368
8369 if (RT_SUCCESS(rc) && !RTUuidIsNull(&ImageUuid))
8370 pDiskTo->pLast->Backend->pfnSetUuid(pDiskTo->pLast->pBackendData, &ImageUuid);
8371 }
8372 if (RT_FAILURE(rc))
8373 break;
8374
8375 pImageTo = pDiskTo->pLast;
8376 AssertPtrBreakStmt(pImageTo, rc = VERR_VD_IMAGE_NOT_FOUND);
8377
8378 cbSize = RT_MIN(cbSize, cbSizeFrom);
8379 }
8380 else
8381 {
8382 pImageTo = pDiskTo->pLast;
8383 AssertPtrBreakStmt(pImageTo, rc = VERR_VD_IMAGE_NOT_FOUND);
8384
8385 uint64_t cbSizeTo;
8386 cbSizeTo = pImageTo->Backend->pfnGetSize(pImageTo->pBackendData);
8387 if (cbSizeTo == 0)
8388 {
8389 rc = VERR_VD_VALUE_NOT_FOUND;
8390 break;
8391 }
8392
8393 if (cbSize == 0)
8394 cbSize = RT_MIN(cbSizeFrom, cbSizeTo);
8395
8396 vdFixupPCHSGeometry(&PCHSGeometryFrom, cbSize);
8397 vdFixupLCHSGeometry(&LCHSGeometryFrom, cbSize);
8398
8399 /* Update the geometry in the destination image. */
8400 pImageTo->Backend->pfnSetPCHSGeometry(pImageTo->pBackendData, &PCHSGeometryFrom);
8401 pImageTo->Backend->pfnSetLCHSGeometry(pImageTo->pBackendData, &LCHSGeometryFrom);
8402 }
8403
8404 rc2 = vdThreadFinishWrite(pDiskTo);
8405 AssertRC(rc2);
8406 fLockWriteTo = false;
8407
8408 /* Whether we can take the optimized copy path (false) or not.
8409 * Don't optimize if the image existed or if it is a child image. */
8410 bool fSuppressRedundantIo = ( !(pszFilename == NULL || cImagesTo > 0)
8411 || (nImageToSame != VD_IMAGE_CONTENT_UNKNOWN));
8412 unsigned cImagesFromReadBack, cImagesToReadBack;
8413
8414 if (nImageFromSame == VD_IMAGE_CONTENT_UNKNOWN)
8415 cImagesFromReadBack = 0;
8416 else
8417 {
8418 if (nImage == VD_LAST_IMAGE)
8419 cImagesFromReadBack = pDiskFrom->cImages - nImageFromSame - 1;
8420 else
8421 cImagesFromReadBack = nImage - nImageFromSame;
8422 }
8423
8424 if (nImageToSame == VD_IMAGE_CONTENT_UNKNOWN)
8425 cImagesToReadBack = 0;
8426 else
8427 cImagesToReadBack = pDiskTo->cImages - nImageToSame - 1;
8428
8429 /* Copy the data. */
8430 rc = vdCopyHelper(pDiskFrom, pImageFrom, pDiskTo, cbSize,
8431 cImagesFromReadBack, cImagesToReadBack,
8432 fSuppressRedundantIo, pIfProgress, pDstIfProgress);
8433
8434 if (RT_SUCCESS(rc))
8435 {
8436 rc2 = vdThreadStartWrite(pDiskTo);
8437 AssertRC(rc2);
8438 fLockWriteTo = true;
8439
8440 /* Only set modification UUID if it is non-null, since the source
8441 * backend might not provide a valid modification UUID. */
8442 if (!RTUuidIsNull(&ImageModificationUuid))
8443 pImageTo->Backend->pfnSetModificationUuid(pImageTo->pBackendData, &ImageModificationUuid);
8444
8445 /* Set the requested open flags if they differ from the value
8446 * required for creating the image and copying the contents. */
8447 if ( pImageTo && pszFilename
8448 && uOpenFlags != (uOpenFlags & ~VD_OPEN_FLAGS_READONLY))
8449 rc = pImageTo->Backend->pfnSetOpenFlags(pImageTo->pBackendData,
8450 uOpenFlags);
8451 }
8452 } while (0);
8453
8454 if (RT_FAILURE(rc) && pImageTo && pszFilename)
8455 {
8456 /* Take the write lock only if it is not taken. Not worth making the
8457 * above code even more complicated. */
8458 if (RT_UNLIKELY(!fLockWriteTo))
8459 {
8460 rc2 = vdThreadStartWrite(pDiskTo);
8461 AssertRC(rc2);
8462 fLockWriteTo = true;
8463 }
8464 /* Error detected, but new image created. Remove image from list. */
8465 vdRemoveImageFromList(pDiskTo, pImageTo);
8466
8467 /* Close and delete image. */
8468 rc2 = pImageTo->Backend->pfnClose(pImageTo->pBackendData, true);
8469 AssertRC(rc2);
8470 pImageTo->pBackendData = NULL;
8471
8472 /* Free remaining resources. */
8473 if (pImageTo->pszFilename)
8474 RTStrFree(pImageTo->pszFilename);
8475
8476 RTMemFree(pImageTo);
8477 }
8478
8479 if (RT_UNLIKELY(fLockWriteTo))
8480 {
8481 rc2 = vdThreadFinishWrite(pDiskTo);
8482 AssertRC(rc2);
8483 }
8484 if (RT_UNLIKELY(fLockWriteFrom))
8485 {
8486 rc2 = vdThreadFinishWrite(pDiskFrom);
8487 AssertRC(rc2);
8488 }
8489 else if (RT_UNLIKELY(fLockReadFrom))
8490 {
8491 rc2 = vdThreadFinishRead(pDiskFrom);
8492 AssertRC(rc2);
8493 }
8494
8495 if (RT_SUCCESS(rc))
8496 {
8497 if (pIfProgress && pIfProgress->pfnProgress)
8498 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
8499 if (pDstIfProgress && pDstIfProgress->pfnProgress)
8500 pDstIfProgress->pfnProgress(pDstIfProgress->Core.pvUser, 100);
8501 }
8502
8503 LogFlowFunc(("returns %Rrc\n", rc));
8504 return rc;
8505}
8506
8507/**
8508 * Copies an image from one HDD container to another.
8509 * The copy is opened in the target HDD container.
8510 * It is possible to convert between different image formats, because the
8511 * backend for the destination may be different from the source.
8512 * If both the source and destination reference the same HDD container,
8513 * then the image is moved (by copying/deleting or renaming) to the new location.
8514 * The source container is unchanged if the move operation fails, otherwise
8515 * the image at the new location is opened in the same way as the old one was.
8516 *
8517 * @returns VBox status code.
8518 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8519 * @param pDiskFrom Pointer to source HDD container.
8520 * @param nImage Image number, counts from 0. 0 is always base image of container.
8521 * @param pDiskTo Pointer to destination HDD container.
8522 * @param pszBackend Name of the image file backend to use.
8523 * @param pszFilename New name of the image (may be NULL if pDiskFrom == pDiskTo).
8524 * @param fMoveByRename If true, attempt to perform a move by renaming (if successful the new size is ignored).
8525 * @param cbSize New image size (0 means leave unchanged).
8526 * @param uImageFlags Flags specifying special destination image features.
8527 * @param pDstUuid New UUID of the destination image. If NULL, a new UUID is created.
8528 * This parameter is used if and only if a true copy is created.
8529 * In all rename/move cases the UUIDs are copied over.
8530 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
8531 * Only used if the destination image is created.
8532 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
8533 * @param pDstVDIfsImage Pointer to the per-image VD interface list, for the
8534 * destination image.
8535 * @param pDstVDIfsOperation Pointer to the per-image VD interface list,
8536 * for the destination image.
8537 */
8538VBOXDDU_DECL(int) VDCopy(PVBOXHDD pDiskFrom, unsigned nImage, PVBOXHDD pDiskTo,
8539 const char *pszBackend, const char *pszFilename,
8540 bool fMoveByRename, uint64_t cbSize,
8541 unsigned uImageFlags, PCRTUUID pDstUuid,
8542 unsigned uOpenFlags, PVDINTERFACE pVDIfsOperation,
8543 PVDINTERFACE pDstVDIfsImage,
8544 PVDINTERFACE pDstVDIfsOperation)
8545{
8546 return VDCopyEx(pDiskFrom, nImage, pDiskTo, pszBackend, pszFilename, fMoveByRename,
8547 cbSize, VD_IMAGE_CONTENT_UNKNOWN, VD_IMAGE_CONTENT_UNKNOWN,
8548 uImageFlags, pDstUuid, uOpenFlags, pVDIfsOperation,
8549 pDstVDIfsImage, pDstVDIfsOperation);
8550}
8551
8552/**
8553 * Optimizes the storage consumption of an image. Typically the unused blocks
8554 * have to be wiped with zeroes to achieve a substantial reduced storage use.
8555 * Another optimization done is reordering the image blocks, which can provide
8556 * a significant performance boost, as reads and writes tend to use less random
8557 * file offsets.
8558 *
8559 * @return VBox status code.
8560 * @return VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
8561 * @return VERR_VD_IMAGE_READ_ONLY if image is not writable.
8562 * @return VERR_NOT_SUPPORTED if this kind of image can be compacted, but
8563 * the code for this isn't implemented yet.
8564 * @param pDisk Pointer to HDD container.
8565 * @param nImage Image number, counts from 0. 0 is always base image of container.
8566 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
8567 */
8568VBOXDDU_DECL(int) VDCompact(PVBOXHDD pDisk, unsigned nImage,
8569 PVDINTERFACE pVDIfsOperation)
8570{
8571 int rc = VINF_SUCCESS;
8572 int rc2;
8573 bool fLockRead = false, fLockWrite = false;
8574 void *pvBuf = NULL;
8575 void *pvTmp = NULL;
8576
8577 LogFlowFunc(("pDisk=%#p nImage=%u pVDIfsOperation=%#p\n",
8578 pDisk, nImage, pVDIfsOperation));
8579
8580 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
8581
8582 do {
8583 /* Check arguments. */
8584 AssertMsgBreakStmt(VALID_PTR(pDisk), ("pDisk=%#p\n", pDisk),
8585 rc = VERR_INVALID_PARAMETER);
8586 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE,
8587 ("u32Signature=%08x\n", pDisk->u32Signature));
8588
8589 rc2 = vdThreadStartRead(pDisk);
8590 AssertRC(rc2);
8591 fLockRead = true;
8592
8593 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
8594 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
8595
8596 /* If there is no compact callback for not file based backends then
8597 * the backend doesn't need compaction. No need to make much fuss about
8598 * this. For file based ones signal this as not yet supported. */
8599 if (!pImage->Backend->pfnCompact)
8600 {
8601 if (pImage->Backend->uBackendCaps & VD_CAP_FILE)
8602 rc = VERR_NOT_SUPPORTED;
8603 else
8604 rc = VINF_SUCCESS;
8605 break;
8606 }
8607
8608 /* Insert interface for reading parent state into per-operation list,
8609 * if there is a parent image. */
8610 VDINTERFACEPARENTSTATE VDIfParent;
8611 VDPARENTSTATEDESC ParentUser;
8612 if (pImage->pPrev)
8613 {
8614 VDIfParent.pfnParentRead = vdParentRead;
8615 ParentUser.pDisk = pDisk;
8616 ParentUser.pImage = pImage->pPrev;
8617 rc = VDInterfaceAdd(&VDIfParent.Core, "VDCompact_ParentState", VDINTERFACETYPE_PARENTSTATE,
8618 &ParentUser, sizeof(VDINTERFACEPARENTSTATE), &pVDIfsOperation);
8619 AssertRC(rc);
8620 }
8621
8622 rc2 = vdThreadFinishRead(pDisk);
8623 AssertRC(rc2);
8624 fLockRead = false;
8625
8626 rc2 = vdThreadStartWrite(pDisk);
8627 AssertRC(rc2);
8628 fLockWrite = true;
8629
8630 rc = pImage->Backend->pfnCompact(pImage->pBackendData,
8631 0, 99,
8632 pDisk->pVDIfsDisk,
8633 pImage->pVDIfsImage,
8634 pVDIfsOperation);
8635 } while (0);
8636
8637 if (RT_UNLIKELY(fLockWrite))
8638 {
8639 rc2 = vdThreadFinishWrite(pDisk);
8640 AssertRC(rc2);
8641 }
8642 else if (RT_UNLIKELY(fLockRead))
8643 {
8644 rc2 = vdThreadFinishRead(pDisk);
8645 AssertRC(rc2);
8646 }
8647
8648 if (pvBuf)
8649 RTMemTmpFree(pvBuf);
8650 if (pvTmp)
8651 RTMemTmpFree(pvTmp);
8652
8653 if (RT_SUCCESS(rc))
8654 {
8655 if (pIfProgress && pIfProgress->pfnProgress)
8656 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
8657 }
8658
8659 LogFlowFunc(("returns %Rrc\n", rc));
8660 return rc;
8661}
8662
8663/**
8664 * Resizes the given disk image to the given size.
8665 *
8666 * @return VBox status
8667 * @return VERR_VD_IMAGE_READ_ONLY if image is not writable.
8668 * @return VERR_NOT_SUPPORTED if this kind of image can be compacted, but
8669 *
8670 * @param pDisk Pointer to the HDD container.
8671 * @param cbSize New size of the image.
8672 * @param pPCHSGeometry Pointer to the new physical disk geometry <= (16383,16,63). Not NULL.
8673 * @param pLCHSGeometry Pointer to the new logical disk geometry <= (x,255,63). Not NULL.
8674 * @param pVDIfsOperation Pointer to the per-operation VD interface list.
8675 */
8676VBOXDDU_DECL(int) VDResize(PVBOXHDD pDisk, uint64_t cbSize,
8677 PCVDGEOMETRY pPCHSGeometry,
8678 PCVDGEOMETRY pLCHSGeometry,
8679 PVDINTERFACE pVDIfsOperation)
8680{
8681 /** @todo r=klaus resizing was designed to be part of VDCopy, so having a separate function is not desirable. */
8682 int rc = VINF_SUCCESS;
8683 int rc2;
8684 bool fLockRead = false, fLockWrite = false;
8685
8686 LogFlowFunc(("pDisk=%#p cbSize=%llu pVDIfsOperation=%#p\n",
8687 pDisk, cbSize, pVDIfsOperation));
8688
8689 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
8690
8691 do {
8692 /* Check arguments. */
8693 AssertMsgBreakStmt(VALID_PTR(pDisk), ("pDisk=%#p\n", pDisk),
8694 rc = VERR_INVALID_PARAMETER);
8695 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE,
8696 ("u32Signature=%08x\n", pDisk->u32Signature));
8697
8698 rc2 = vdThreadStartRead(pDisk);
8699 AssertRC(rc2);
8700 fLockRead = true;
8701
8702 /* Must have at least one image in the chain, will resize last. */
8703 AssertMsgBreakStmt(pDisk->cImages >= 1, ("cImages=%u\n", pDisk->cImages),
8704 rc = VERR_NOT_SUPPORTED);
8705
8706 PVDIMAGE pImage = pDisk->pLast;
8707
8708 /* If there is no compact callback for not file based backends then
8709 * the backend doesn't need compaction. No need to make much fuss about
8710 * this. For file based ones signal this as not yet supported. */
8711 if (!pImage->Backend->pfnResize)
8712 {
8713 if (pImage->Backend->uBackendCaps & VD_CAP_FILE)
8714 rc = VERR_NOT_SUPPORTED;
8715 else
8716 rc = VINF_SUCCESS;
8717 break;
8718 }
8719
8720 rc2 = vdThreadFinishRead(pDisk);
8721 AssertRC(rc2);
8722 fLockRead = false;
8723
8724 rc2 = vdThreadStartWrite(pDisk);
8725 AssertRC(rc2);
8726 fLockWrite = true;
8727
8728 VDGEOMETRY PCHSGeometryOld;
8729 VDGEOMETRY LCHSGeometryOld;
8730 PCVDGEOMETRY pPCHSGeometryNew;
8731 PCVDGEOMETRY pLCHSGeometryNew;
8732
8733 if (pPCHSGeometry->cCylinders == 0)
8734 {
8735 /* Auto-detect marker, calculate new value ourself. */
8736 rc = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData, &PCHSGeometryOld);
8737 if (RT_SUCCESS(rc) && (PCHSGeometryOld.cCylinders != 0))
8738 PCHSGeometryOld.cCylinders = RT_MIN(cbSize / 512 / PCHSGeometryOld.cHeads / PCHSGeometryOld.cSectors, 16383);
8739 else if (rc == VERR_VD_GEOMETRY_NOT_SET)
8740 rc = VINF_SUCCESS;
8741
8742 pPCHSGeometryNew = &PCHSGeometryOld;
8743 }
8744 else
8745 pPCHSGeometryNew = pPCHSGeometry;
8746
8747 if (pLCHSGeometry->cCylinders == 0)
8748 {
8749 /* Auto-detect marker, calculate new value ourself. */
8750 rc = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData, &LCHSGeometryOld);
8751 if (RT_SUCCESS(rc) && (LCHSGeometryOld.cCylinders != 0))
8752 LCHSGeometryOld.cCylinders = cbSize / 512 / LCHSGeometryOld.cHeads / LCHSGeometryOld.cSectors;
8753 else if (rc == VERR_VD_GEOMETRY_NOT_SET)
8754 rc = VINF_SUCCESS;
8755
8756 pLCHSGeometryNew = &LCHSGeometryOld;
8757 }
8758 else
8759 pLCHSGeometryNew = pLCHSGeometry;
8760
8761 if (RT_SUCCESS(rc))
8762 rc = pImage->Backend->pfnResize(pImage->pBackendData,
8763 cbSize,
8764 pPCHSGeometryNew,
8765 pLCHSGeometryNew,
8766 0, 99,
8767 pDisk->pVDIfsDisk,
8768 pImage->pVDIfsImage,
8769 pVDIfsOperation);
8770 } while (0);
8771
8772 if (RT_UNLIKELY(fLockWrite))
8773 {
8774 rc2 = vdThreadFinishWrite(pDisk);
8775 AssertRC(rc2);
8776 }
8777 else if (RT_UNLIKELY(fLockRead))
8778 {
8779 rc2 = vdThreadFinishRead(pDisk);
8780 AssertRC(rc2);
8781 }
8782
8783 if (RT_SUCCESS(rc))
8784 {
8785 if (pIfProgress && pIfProgress->pfnProgress)
8786 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
8787
8788 pDisk->cbSize = cbSize;
8789 }
8790
8791 LogFlowFunc(("returns %Rrc\n", rc));
8792 return rc;
8793}
8794
8795VBOXDDU_DECL(int) VDPrepareWithFilters(PVBOXHDD pDisk, PVDINTERFACE pVDIfsOperation)
8796{
8797 int rc = VINF_SUCCESS;
8798 int rc2;
8799 bool fLockRead = false, fLockWrite = false;
8800
8801 LogFlowFunc(("pDisk=%#p pVDIfsOperation=%#p\n", pDisk, pVDIfsOperation));
8802
8803 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
8804
8805 do {
8806 /* Check arguments. */
8807 AssertMsgBreakStmt(VALID_PTR(pDisk), ("pDisk=%#p\n", pDisk),
8808 rc = VERR_INVALID_PARAMETER);
8809 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE,
8810 ("u32Signature=%08x\n", pDisk->u32Signature));
8811
8812 rc2 = vdThreadStartRead(pDisk);
8813 AssertRC(rc2);
8814 fLockRead = true;
8815
8816 /* Must have at least one image in the chain. */
8817 AssertMsgBreakStmt(pDisk->cImages >= 1, ("cImages=%u\n", pDisk->cImages),
8818 rc = VERR_VD_NOT_OPENED);
8819
8820 unsigned uOpenFlags = pDisk->pLast->Backend->pfnGetOpenFlags(pDisk->pLast->pBackendData);
8821 AssertMsgBreakStmt(!(uOpenFlags & VD_OPEN_FLAGS_READONLY),
8822 ("Last image should be read write"),
8823 rc = VERR_VD_IMAGE_READ_ONLY);
8824
8825 rc2 = vdThreadFinishRead(pDisk);
8826 AssertRC(rc2);
8827 fLockRead = false;
8828
8829 rc2 = vdThreadStartWrite(pDisk);
8830 AssertRC(rc2);
8831 fLockWrite = true;
8832
8833 /*
8834 * Open all images in the chain in read write mode first to avoid running
8835 * into an error in the middle of the process.
8836 */
8837 PVDIMAGE pImage = pDisk->pBase;
8838
8839 while (pImage)
8840 {
8841 uOpenFlags = pImage->Backend->pfnGetOpenFlags(pImage->pBackendData);
8842 if (uOpenFlags & VD_OPEN_FLAGS_READONLY)
8843 {
8844 /*
8845 * Clear skip consistency checks because the image is made writable now and
8846 * skipping consistency checks is only possible for readonly images.
8847 */
8848 uOpenFlags &= ~(VD_OPEN_FLAGS_READONLY | VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS);
8849 rc = pImage->Backend->pfnSetOpenFlags(pImage->pBackendData, uOpenFlags);
8850 if (RT_FAILURE(rc))
8851 break;
8852 }
8853 pImage = pImage->pNext;
8854 }
8855
8856 if (RT_SUCCESS(rc))
8857 {
8858 unsigned cImgCur = 0;
8859 unsigned uPercentStart = 0;
8860 unsigned uPercentSpan = 100 / pDisk->cImages - 1;
8861
8862 /* Allocate tmp buffer. */
8863 void *pvBuf = RTMemTmpAlloc(VD_MERGE_BUFFER_SIZE);
8864 if (!pvBuf)
8865 {
8866 rc = VERR_NO_MEMORY;
8867 break;
8868 }
8869
8870 pImage = pDisk->pBase;
8871 pDisk->fLocked = true;
8872
8873 while ( pImage
8874 && RT_SUCCESS(rc))
8875 {
8876 /* Get size of image. */
8877 uint64_t cbSize = pImage->Backend->pfnGetSize(pImage->pBackendData);
8878 uint64_t cbSizeFile = pImage->Backend->pfnGetFileSize(pImage->pBackendData);
8879 uint64_t cbFileWritten = 0;
8880 uint64_t uOffset = 0;
8881 uint64_t cbRemaining = cbSize;
8882
8883 do
8884 {
8885 size_t cbThisRead = RT_MIN(VD_MERGE_BUFFER_SIZE, cbRemaining);
8886 RTSGSEG SegmentBuf;
8887 RTSGBUF SgBuf;
8888 VDIOCTX IoCtx;
8889
8890 SegmentBuf.pvSeg = pvBuf;
8891 SegmentBuf.cbSeg = VD_MERGE_BUFFER_SIZE;
8892 RTSgBufInit(&SgBuf, &SegmentBuf, 1);
8893 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_READ, 0, 0, NULL,
8894 &SgBuf, NULL, NULL, VDIOCTX_FLAGS_SYNC);
8895
8896 rc = pImage->Backend->pfnRead(pImage->pBackendData, uOffset,
8897 cbThisRead, &IoCtx, &cbThisRead);
8898 if (rc != VERR_VD_BLOCK_FREE)
8899 {
8900 if (RT_FAILURE(rc))
8901 break;
8902
8903 /* Apply filter chains. */
8904 rc = vdFilterChainApplyRead(pDisk, uOffset, cbThisRead, &IoCtx);
8905 if (RT_FAILURE(rc))
8906 break;
8907
8908 rc = vdFilterChainApplyWrite(pDisk, uOffset, cbThisRead, &IoCtx);
8909 if (RT_FAILURE(rc))
8910 break;
8911
8912 RTSgBufReset(&SgBuf);
8913 size_t cbThisWrite = 0;
8914 size_t cbPreRead = 0;
8915 size_t cbPostRead = 0;
8916 rc = pImage->Backend->pfnWrite(pImage->pBackendData, uOffset,
8917 cbThisRead, &IoCtx, &cbThisWrite,
8918 &cbPreRead, &cbPostRead, 0);
8919 if (RT_FAILURE(rc))
8920 break;
8921 Assert(cbThisWrite == cbThisRead);
8922 cbFileWritten += cbThisWrite;
8923 }
8924 else
8925 rc = VINF_SUCCESS;
8926
8927 uOffset += cbThisRead;
8928 cbRemaining -= cbThisRead;
8929
8930 if (pIfProgress && pIfProgress->pfnProgress)
8931 {
8932 rc2 = pIfProgress->pfnProgress(pIfProgress->Core.pvUser,
8933 uPercentStart + cbFileWritten * uPercentSpan / cbSizeFile);
8934 AssertRC(rc2); /* Cancelling this operation without leaving an inconsistent state is not possible. */
8935 }
8936 } while (uOffset < cbSize);
8937
8938 pImage = pImage->pNext;
8939 cImgCur++;
8940 uPercentStart += uPercentSpan;
8941 }
8942
8943 pDisk->fLocked = false;
8944 if (pvBuf)
8945 RTMemTmpFree(pvBuf);
8946 }
8947
8948 /* Change images except last one back to readonly. */
8949 pImage = pDisk->pBase;
8950 while ( pImage != pDisk->pLast
8951 && pImage)
8952 {
8953 uOpenFlags = pImage->Backend->pfnGetOpenFlags(pImage->pBackendData);
8954 uOpenFlags |= VD_OPEN_FLAGS_READONLY;
8955 rc2 = pImage->Backend->pfnSetOpenFlags(pImage->pBackendData, uOpenFlags);
8956 if (RT_FAILURE(rc2))
8957 {
8958 if (RT_SUCCESS(rc))
8959 rc = rc2;
8960 break;
8961 }
8962 pImage = pImage->pNext;
8963 }
8964 } while (0);
8965
8966 if (RT_UNLIKELY(fLockWrite))
8967 {
8968 rc2 = vdThreadFinishWrite(pDisk);
8969 AssertRC(rc2);
8970 }
8971 else if (RT_UNLIKELY(fLockRead))
8972 {
8973 rc2 = vdThreadFinishRead(pDisk);
8974 AssertRC(rc2);
8975 }
8976
8977 if ( RT_SUCCESS(rc)
8978 && pIfProgress
8979 && pIfProgress->pfnProgress)
8980 pIfProgress->pfnProgress(pIfProgress->Core.pvUser, 100);
8981
8982 LogFlowFunc(("returns %Rrc\n", rc));
8983 return rc;
8984}
8985
8986/**
8987 * Closes the last opened image file in HDD container.
8988 * If previous image file was opened in read-only mode (the normal case) and
8989 * the last opened image is in read-write mode then the previous image will be
8990 * reopened in read/write mode.
8991 *
8992 * @returns VBox status code.
8993 * @returns VERR_VD_NOT_OPENED if no image is opened in HDD container.
8994 * @param pDisk Pointer to HDD container.
8995 * @param fDelete If true, delete the image from the host disk.
8996 */
8997VBOXDDU_DECL(int) VDClose(PVBOXHDD pDisk, bool fDelete)
8998{
8999 int rc = VINF_SUCCESS;
9000 int rc2;
9001 bool fLockWrite = false;
9002
9003 LogFlowFunc(("pDisk=%#p fDelete=%d\n", pDisk, fDelete));
9004 do
9005 {
9006 /* sanity check */
9007 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9008 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9009
9010 /* Not worth splitting this up into a read lock phase and write
9011 * lock phase, as closing an image is a relatively fast operation
9012 * dominated by the part which needs the write lock. */
9013 rc2 = vdThreadStartWrite(pDisk);
9014 AssertRC(rc2);
9015 fLockWrite = true;
9016
9017 PVDIMAGE pImage = pDisk->pLast;
9018 if (!pImage)
9019 {
9020 rc = VERR_VD_NOT_OPENED;
9021 break;
9022 }
9023
9024 /* Destroy the current discard state first which might still have pending blocks. */
9025 rc = vdDiscardStateDestroy(pDisk);
9026 if (RT_FAILURE(rc))
9027 break;
9028
9029 unsigned uOpenFlags = pImage->Backend->pfnGetOpenFlags(pImage->pBackendData);
9030 /* Remove image from list of opened images. */
9031 vdRemoveImageFromList(pDisk, pImage);
9032 /* Close (and optionally delete) image. */
9033 rc = pImage->Backend->pfnClose(pImage->pBackendData, fDelete);
9034 /* Free remaining resources related to the image. */
9035 RTStrFree(pImage->pszFilename);
9036 RTMemFree(pImage);
9037
9038 pImage = pDisk->pLast;
9039 if (!pImage)
9040 break;
9041
9042 /* If disk was previously in read/write mode, make sure it will stay
9043 * like this (if possible) after closing this image. Set the open flags
9044 * accordingly. */
9045 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
9046 {
9047 uOpenFlags = pImage->Backend->pfnGetOpenFlags(pImage->pBackendData);
9048 uOpenFlags &= ~ VD_OPEN_FLAGS_READONLY;
9049 rc = pImage->Backend->pfnSetOpenFlags(pImage->pBackendData, uOpenFlags);
9050 }
9051
9052 /* Cache disk information. */
9053 pDisk->cbSize = pImage->Backend->pfnGetSize(pImage->pBackendData);
9054
9055 /* Cache PCHS geometry. */
9056 rc2 = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
9057 &pDisk->PCHSGeometry);
9058 if (RT_FAILURE(rc2))
9059 {
9060 pDisk->PCHSGeometry.cCylinders = 0;
9061 pDisk->PCHSGeometry.cHeads = 0;
9062 pDisk->PCHSGeometry.cSectors = 0;
9063 }
9064 else
9065 {
9066 /* Make sure the PCHS geometry is properly clipped. */
9067 pDisk->PCHSGeometry.cCylinders = RT_MIN(pDisk->PCHSGeometry.cCylinders, 16383);
9068 pDisk->PCHSGeometry.cHeads = RT_MIN(pDisk->PCHSGeometry.cHeads, 16);
9069 pDisk->PCHSGeometry.cSectors = RT_MIN(pDisk->PCHSGeometry.cSectors, 63);
9070 }
9071
9072 /* Cache LCHS geometry. */
9073 rc2 = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
9074 &pDisk->LCHSGeometry);
9075 if (RT_FAILURE(rc2))
9076 {
9077 pDisk->LCHSGeometry.cCylinders = 0;
9078 pDisk->LCHSGeometry.cHeads = 0;
9079 pDisk->LCHSGeometry.cSectors = 0;
9080 }
9081 else
9082 {
9083 /* Make sure the LCHS geometry is properly clipped. */
9084 pDisk->LCHSGeometry.cHeads = RT_MIN(pDisk->LCHSGeometry.cHeads, 255);
9085 pDisk->LCHSGeometry.cSectors = RT_MIN(pDisk->LCHSGeometry.cSectors, 63);
9086 }
9087 } while (0);
9088
9089 if (RT_UNLIKELY(fLockWrite))
9090 {
9091 rc2 = vdThreadFinishWrite(pDisk);
9092 AssertRC(rc2);
9093 }
9094
9095 LogFlowFunc(("returns %Rrc\n", rc));
9096 return rc;
9097}
9098
9099/**
9100 * Closes the currently opened cache image file in HDD container.
9101 *
9102 * @return VBox status code.
9103 * @return VERR_VD_NOT_OPENED if no cache is opened in HDD container.
9104 * @param pDisk Pointer to HDD container.
9105 * @param fDelete If true, delete the image from the host disk.
9106 */
9107VBOXDDU_DECL(int) VDCacheClose(PVBOXHDD pDisk, bool fDelete)
9108{
9109 int rc = VINF_SUCCESS;
9110 int rc2;
9111 bool fLockWrite = false;
9112 PVDCACHE pCache = NULL;
9113
9114 LogFlowFunc(("pDisk=%#p fDelete=%d\n", pDisk, fDelete));
9115
9116 do
9117 {
9118 /* sanity check */
9119 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9120 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9121
9122 rc2 = vdThreadStartWrite(pDisk);
9123 AssertRC(rc2);
9124 fLockWrite = true;
9125
9126 AssertPtrBreakStmt(pDisk->pCache, rc = VERR_VD_CACHE_NOT_FOUND);
9127
9128 pCache = pDisk->pCache;
9129 pDisk->pCache = NULL;
9130
9131 pCache->Backend->pfnClose(pCache->pBackendData, fDelete);
9132 if (pCache->pszFilename)
9133 RTStrFree(pCache->pszFilename);
9134 RTMemFree(pCache);
9135 } while (0);
9136
9137 if (RT_LIKELY(fLockWrite))
9138 {
9139 rc2 = vdThreadFinishWrite(pDisk);
9140 AssertRC(rc2);
9141 }
9142
9143 LogFlowFunc(("returns %Rrc\n", rc));
9144 return rc;
9145}
9146
9147VBOXDDU_DECL(int) VDFilterRemove(PVBOXHDD pDisk, uint32_t fFlags)
9148{
9149 int rc = VINF_SUCCESS;
9150 int rc2;
9151 bool fLockWrite = false;
9152 PVDFILTER pFilter = NULL;
9153
9154 LogFlowFunc(("pDisk=%#p\n", pDisk));
9155
9156 do
9157 {
9158 /* sanity check */
9159 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9160 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9161
9162 AssertMsgBreakStmt(!(fFlags & ~VD_FILTER_FLAGS_MASK),
9163 ("Invalid flags set (fFlags=%#x)\n", fFlags),
9164 rc = VERR_INVALID_PARAMETER);
9165
9166 rc2 = vdThreadStartWrite(pDisk);
9167 AssertRC(rc2);
9168 fLockWrite = true;
9169
9170 if (fFlags & VD_FILTER_FLAGS_WRITE)
9171 {
9172 AssertBreakStmt(!RTListIsEmpty(&pDisk->ListFilterChainWrite), rc = VERR_VD_NOT_OPENED);
9173 pFilter = RTListGetLast(&pDisk->ListFilterChainWrite, VDFILTER, ListNodeChainWrite);
9174 AssertPtr(pFilter);
9175 RTListNodeRemove(&pFilter->ListNodeChainWrite);
9176 vdFilterRelease(pFilter);
9177 }
9178
9179 if (fFlags & VD_FILTER_FLAGS_READ)
9180 {
9181 AssertBreakStmt(!RTListIsEmpty(&pDisk->ListFilterChainRead), rc = VERR_VD_NOT_OPENED);
9182 pFilter = RTListGetLast(&pDisk->ListFilterChainRead, VDFILTER, ListNodeChainRead);
9183 AssertPtr(pFilter);
9184 RTListNodeRemove(&pFilter->ListNodeChainRead);
9185 vdFilterRelease(pFilter);
9186 }
9187 } while (0);
9188
9189 if (RT_LIKELY(fLockWrite))
9190 {
9191 rc2 = vdThreadFinishWrite(pDisk);
9192 AssertRC(rc2);
9193 }
9194
9195 LogFlowFunc(("returns %Rrc\n", rc));
9196 return rc;
9197}
9198
9199/**
9200 * Closes all opened image files in HDD container.
9201 *
9202 * @returns VBox status code.
9203 * @param pDisk Pointer to HDD container.
9204 */
9205VBOXDDU_DECL(int) VDCloseAll(PVBOXHDD pDisk)
9206{
9207 int rc = VINF_SUCCESS;
9208 int rc2;
9209 bool fLockWrite = false;
9210
9211 LogFlowFunc(("pDisk=%#p\n", pDisk));
9212 do
9213 {
9214 /* sanity check */
9215 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9216 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9217
9218 /* Lock the entire operation. */
9219 rc2 = vdThreadStartWrite(pDisk);
9220 AssertRC(rc2);
9221 fLockWrite = true;
9222
9223 PVDCACHE pCache = pDisk->pCache;
9224 if (pCache)
9225 {
9226 rc2 = pCache->Backend->pfnClose(pCache->pBackendData, false);
9227 if (RT_FAILURE(rc2) && RT_SUCCESS(rc))
9228 rc = rc2;
9229
9230 if (pCache->pszFilename)
9231 RTStrFree(pCache->pszFilename);
9232 RTMemFree(pCache);
9233 }
9234
9235 PVDIMAGE pImage = pDisk->pLast;
9236 while (VALID_PTR(pImage))
9237 {
9238 PVDIMAGE pPrev = pImage->pPrev;
9239 /* Remove image from list of opened images. */
9240 vdRemoveImageFromList(pDisk, pImage);
9241 /* Close image. */
9242 rc2 = pImage->Backend->pfnClose(pImage->pBackendData, false);
9243 if (RT_FAILURE(rc2) && RT_SUCCESS(rc))
9244 rc = rc2;
9245 /* Free remaining resources related to the image. */
9246 RTStrFree(pImage->pszFilename);
9247 RTMemFree(pImage);
9248 pImage = pPrev;
9249 }
9250 Assert(!VALID_PTR(pDisk->pLast));
9251 } while (0);
9252
9253 if (RT_UNLIKELY(fLockWrite))
9254 {
9255 rc2 = vdThreadFinishWrite(pDisk);
9256 AssertRC(rc2);
9257 }
9258
9259 LogFlowFunc(("returns %Rrc\n", rc));
9260 return rc;
9261}
9262
9263/**
9264 * Removes all filters of the given HDD container.
9265 *
9266 * @return VBox status code.
9267 * @param pDisk Pointer to HDD container.
9268 */
9269VBOXDDU_DECL(int) VDFilterRemoveAll(PVBOXHDD pDisk)
9270{
9271 int rc = VINF_SUCCESS;
9272 int rc2;
9273 bool fLockWrite = false;
9274
9275 LogFlowFunc(("pDisk=%#p\n", pDisk));
9276 do
9277 {
9278 /* sanity check */
9279 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9280 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9281
9282 /* Lock the entire operation. */
9283 rc2 = vdThreadStartWrite(pDisk);
9284 AssertRC(rc2);
9285 fLockWrite = true;
9286
9287 PVDFILTER pFilter, pFilterNext;
9288 RTListForEachSafe(&pDisk->ListFilterChainWrite, pFilter, pFilterNext, VDFILTER, ListNodeChainWrite)
9289 {
9290 RTListNodeRemove(&pFilter->ListNodeChainWrite);
9291 vdFilterRelease(pFilter);
9292 }
9293
9294 RTListForEachSafe(&pDisk->ListFilterChainRead, pFilter, pFilterNext, VDFILTER, ListNodeChainRead)
9295 {
9296 RTListNodeRemove(&pFilter->ListNodeChainRead);
9297 vdFilterRelease(pFilter);
9298 }
9299 Assert(RTListIsEmpty(&pDisk->ListFilterChainRead));
9300 Assert(RTListIsEmpty(&pDisk->ListFilterChainWrite));
9301 } while (0);
9302
9303 if (RT_UNLIKELY(fLockWrite))
9304 {
9305 rc2 = vdThreadFinishWrite(pDisk);
9306 AssertRC(rc2);
9307 }
9308
9309 LogFlowFunc(("returns %Rrc\n", rc));
9310 return rc;
9311}
9312
9313/**
9314 * Read data from virtual HDD.
9315 *
9316 * @returns VBox status code.
9317 * @retval VERR_VD_NOT_OPENED if no image is opened in HDD container.
9318 * @param pDisk Pointer to HDD container.
9319 * @param uOffset Offset of first reading byte from start of disk.
9320 * @param pvBuf Pointer to buffer for reading data.
9321 * @param cbRead Number of bytes to read.
9322 */
9323VBOXDDU_DECL(int) VDRead(PVBOXHDD pDisk, uint64_t uOffset, void *pvBuf,
9324 size_t cbRead)
9325{
9326 int rc = VINF_SUCCESS;
9327 int rc2;
9328 bool fLockRead = false;
9329
9330 LogFlowFunc(("pDisk=%#p uOffset=%llu pvBuf=%p cbRead=%zu\n",
9331 pDisk, uOffset, pvBuf, cbRead));
9332 do
9333 {
9334 /* sanity check */
9335 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9336 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9337
9338 /* Check arguments. */
9339 AssertMsgBreakStmt(VALID_PTR(pvBuf),
9340 ("pvBuf=%#p\n", pvBuf),
9341 rc = VERR_INVALID_PARAMETER);
9342 AssertMsgBreakStmt(cbRead,
9343 ("cbRead=%zu\n", cbRead),
9344 rc = VERR_INVALID_PARAMETER);
9345
9346 rc2 = vdThreadStartRead(pDisk);
9347 AssertRC(rc2);
9348 fLockRead = true;
9349
9350 PVDIMAGE pImage = pDisk->pLast;
9351 AssertPtrBreakStmt(pImage, rc = VERR_VD_NOT_OPENED);
9352
9353 if (uOffset + cbRead > pDisk->cbSize)
9354 {
9355 /* Floppy images might be smaller than the standard expected by
9356 the floppy controller code. So, we won't fail here. */
9357 AssertMsgBreakStmt(pDisk->enmType == VDTYPE_FLOPPY,
9358 ("uOffset=%llu cbRead=%zu pDisk->cbSize=%llu\n",
9359 uOffset, cbRead, pDisk->cbSize),
9360 rc = VERR_EOF);
9361 memset(pvBuf, 0xf6, cbRead); /* f6h = format.com filler byte */
9362 if (uOffset >= pDisk->cbSize)
9363 break;
9364 cbRead = pDisk->cbSize - uOffset;
9365 }
9366
9367 rc = vdReadHelper(pDisk, pImage, uOffset, pvBuf, cbRead,
9368 true /* fUpdateCache */);
9369 } while (0);
9370
9371 if (RT_UNLIKELY(fLockRead))
9372 {
9373 rc2 = vdThreadFinishRead(pDisk);
9374 AssertRC(rc2);
9375 }
9376
9377 LogFlowFunc(("returns %Rrc\n", rc));
9378 return rc;
9379}
9380
9381/**
9382 * Write data to virtual HDD.
9383 *
9384 * @returns VBox status code.
9385 * @retval VERR_VD_NOT_OPENED if no image is opened in HDD container.
9386 * @param pDisk Pointer to HDD container.
9387 * @param uOffset Offset of the first byte being
9388 * written from start of disk.
9389 * @param pvBuf Pointer to buffer for writing data.
9390 * @param cbWrite Number of bytes to write.
9391 */
9392VBOXDDU_DECL(int) VDWrite(PVBOXHDD pDisk, uint64_t uOffset, const void *pvBuf,
9393 size_t cbWrite)
9394{
9395 int rc = VINF_SUCCESS;
9396 int rc2;
9397 bool fLockWrite = false;
9398
9399 LogFlowFunc(("pDisk=%#p uOffset=%llu pvBuf=%p cbWrite=%zu\n",
9400 pDisk, uOffset, pvBuf, cbWrite));
9401 do
9402 {
9403 /* sanity check */
9404 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9405 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9406
9407 /* Check arguments. */
9408 AssertMsgBreakStmt(VALID_PTR(pvBuf),
9409 ("pvBuf=%#p\n", pvBuf),
9410 rc = VERR_INVALID_PARAMETER);
9411 AssertMsgBreakStmt(cbWrite,
9412 ("cbWrite=%zu\n", cbWrite),
9413 rc = VERR_INVALID_PARAMETER);
9414
9415 rc2 = vdThreadStartWrite(pDisk);
9416 AssertRC(rc2);
9417 fLockWrite = true;
9418
9419 AssertMsgBreakStmt(uOffset + cbWrite <= pDisk->cbSize,
9420 ("uOffset=%llu cbWrite=%zu pDisk->cbSize=%llu\n",
9421 uOffset, cbWrite, pDisk->cbSize),
9422 rc = VERR_INVALID_PARAMETER);
9423
9424 PVDIMAGE pImage = pDisk->pLast;
9425 AssertPtrBreakStmt(pImage, rc = VERR_VD_NOT_OPENED);
9426
9427 vdSetModifiedFlag(pDisk);
9428 rc = vdWriteHelper(pDisk, pImage, uOffset, pvBuf, cbWrite,
9429 VDIOCTX_FLAGS_READ_UPDATE_CACHE);
9430 if (RT_FAILURE(rc))
9431 break;
9432
9433 /* If there is a merge (in the direction towards a parent) running
9434 * concurrently then we have to also "relay" the write to this parent,
9435 * as the merge position might be already past the position where
9436 * this write is going. The "context" of the write can come from the
9437 * natural chain, since merging either already did or will take care
9438 * of the "other" content which is might be needed to fill the block
9439 * to a full allocation size. The cache doesn't need to be touched
9440 * as this write is covered by the previous one. */
9441 if (RT_UNLIKELY(pDisk->pImageRelay))
9442 rc = vdWriteHelper(pDisk, pDisk->pImageRelay, uOffset,
9443 pvBuf, cbWrite, VDIOCTX_FLAGS_DEFAULT);
9444 } while (0);
9445
9446 if (RT_UNLIKELY(fLockWrite))
9447 {
9448 rc2 = vdThreadFinishWrite(pDisk);
9449 AssertRC(rc2);
9450 }
9451
9452 LogFlowFunc(("returns %Rrc\n", rc));
9453 return rc;
9454}
9455
9456/**
9457 * Make sure the on disk representation of a virtual HDD is up to date.
9458 *
9459 * @returns VBox status code.
9460 * @retval VERR_VD_NOT_OPENED if no image is opened in HDD container.
9461 * @param pDisk Pointer to HDD container.
9462 */
9463VBOXDDU_DECL(int) VDFlush(PVBOXHDD pDisk)
9464{
9465 int rc = VINF_SUCCESS;
9466 int rc2;
9467 bool fLockWrite = false;
9468
9469 LogFlowFunc(("pDisk=%#p\n", pDisk));
9470 do
9471 {
9472 /* sanity check */
9473 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9474 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9475
9476 rc2 = vdThreadStartWrite(pDisk);
9477 AssertRC(rc2);
9478 fLockWrite = true;
9479
9480 PVDIMAGE pImage = pDisk->pLast;
9481 AssertPtrBreakStmt(pImage, rc = VERR_VD_NOT_OPENED);
9482
9483 VDIOCTX IoCtx;
9484 RTSEMEVENT hEventComplete = NIL_RTSEMEVENT;
9485
9486 rc = RTSemEventCreate(&hEventComplete);
9487 if (RT_FAILURE(rc))
9488 break;
9489
9490 vdIoCtxInit(&IoCtx, pDisk, VDIOCTXTXDIR_FLUSH, 0, 0, pImage, NULL,
9491 NULL, vdFlushHelperAsync, VDIOCTX_FLAGS_SYNC | VDIOCTX_FLAGS_DONT_FREE);
9492
9493 IoCtx.Type.Root.pfnComplete = vdIoCtxSyncComplete;
9494 IoCtx.Type.Root.pvUser1 = pDisk;
9495 IoCtx.Type.Root.pvUser2 = hEventComplete;
9496 rc = vdIoCtxProcessSync(&IoCtx, hEventComplete);
9497
9498 RTSemEventDestroy(hEventComplete);
9499 } while (0);
9500
9501 if (RT_UNLIKELY(fLockWrite))
9502 {
9503 rc2 = vdThreadFinishWrite(pDisk);
9504 AssertRC(rc2);
9505 }
9506
9507 LogFlowFunc(("returns %Rrc\n", rc));
9508 return rc;
9509}
9510
9511/**
9512 * Get number of opened images in HDD container.
9513 *
9514 * @returns Number of opened images for HDD container. 0 if no images have been opened.
9515 * @param pDisk Pointer to HDD container.
9516 */
9517VBOXDDU_DECL(unsigned) VDGetCount(PVBOXHDD pDisk)
9518{
9519 unsigned cImages;
9520 int rc2;
9521 bool fLockRead = false;
9522
9523 LogFlowFunc(("pDisk=%#p\n", pDisk));
9524 do
9525 {
9526 /* sanity check */
9527 AssertPtrBreakStmt(pDisk, cImages = 0);
9528 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9529
9530 rc2 = vdThreadStartRead(pDisk);
9531 AssertRC(rc2);
9532 fLockRead = true;
9533
9534 cImages = pDisk->cImages;
9535 } while (0);
9536
9537 if (RT_UNLIKELY(fLockRead))
9538 {
9539 rc2 = vdThreadFinishRead(pDisk);
9540 AssertRC(rc2);
9541 }
9542
9543 LogFlowFunc(("returns %u\n", cImages));
9544 return cImages;
9545}
9546
9547/**
9548 * Get read/write mode of HDD container.
9549 *
9550 * @returns Virtual disk ReadOnly status.
9551 * @returns true if no image is opened in HDD container.
9552 * @param pDisk Pointer to HDD container.
9553 */
9554VBOXDDU_DECL(bool) VDIsReadOnly(PVBOXHDD pDisk)
9555{
9556 bool fReadOnly;
9557 int rc2;
9558 bool fLockRead = false;
9559
9560 LogFlowFunc(("pDisk=%#p\n", pDisk));
9561 do
9562 {
9563 /* sanity check */
9564 AssertPtrBreakStmt(pDisk, fReadOnly = false);
9565 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9566
9567 rc2 = vdThreadStartRead(pDisk);
9568 AssertRC(rc2);
9569 fLockRead = true;
9570
9571 PVDIMAGE pImage = pDisk->pLast;
9572 AssertPtrBreakStmt(pImage, fReadOnly = true);
9573
9574 unsigned uOpenFlags;
9575 uOpenFlags = pDisk->pLast->Backend->pfnGetOpenFlags(pDisk->pLast->pBackendData);
9576 fReadOnly = !!(uOpenFlags & VD_OPEN_FLAGS_READONLY);
9577 } while (0);
9578
9579 if (RT_UNLIKELY(fLockRead))
9580 {
9581 rc2 = vdThreadFinishRead(pDisk);
9582 AssertRC(rc2);
9583 }
9584
9585 LogFlowFunc(("returns %d\n", fReadOnly));
9586 return fReadOnly;
9587}
9588
9589/**
9590 * Get sector size of an image in HDD container.
9591 *
9592 * @return Virtual disk sector size in bytes.
9593 * @return 0 if image with specified number was not opened.
9594 * @param pDisk Pointer to HDD container.
9595 * @param nImage Image number, counts from 0. 0 is always base image of container.
9596 */
9597VBOXDDU_DECL(uint32_t) VDGetSectorSize(PVBOXHDD pDisk, unsigned nImage)
9598{
9599 uint64_t cbSector;
9600 int rc2;
9601 bool fLockRead = false;
9602
9603 LogFlowFunc(("pDisk=%#p nImage=%u\n", pDisk, nImage));
9604 do
9605 {
9606 /* sanity check */
9607 AssertPtrBreakStmt(pDisk, cbSector = 0);
9608 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9609
9610 rc2 = vdThreadStartRead(pDisk);
9611 AssertRC(rc2);
9612 fLockRead = true;
9613
9614 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9615 AssertPtrBreakStmt(pImage, cbSector = 0);
9616 cbSector = pImage->Backend->pfnGetSectorSize(pImage->pBackendData);
9617 } while (0);
9618
9619 if (RT_UNLIKELY(fLockRead))
9620 {
9621 rc2 = vdThreadFinishRead(pDisk);
9622 AssertRC(rc2);
9623 }
9624
9625 LogFlowFunc(("returns %u\n", cbSector));
9626 return cbSector;
9627}
9628
9629/**
9630 * Get total capacity of an image in HDD container.
9631 *
9632 * @returns Virtual disk size in bytes.
9633 * @returns 0 if no image with specified number was not opened.
9634 * @param pDisk Pointer to HDD container.
9635 * @param nImage Image number, counts from 0. 0 is always base image of container.
9636 */
9637VBOXDDU_DECL(uint64_t) VDGetSize(PVBOXHDD pDisk, unsigned nImage)
9638{
9639 uint64_t cbSize;
9640 int rc2;
9641 bool fLockRead = false;
9642
9643 LogFlowFunc(("pDisk=%#p nImage=%u\n", pDisk, nImage));
9644 do
9645 {
9646 /* sanity check */
9647 AssertPtrBreakStmt(pDisk, cbSize = 0);
9648 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9649
9650 rc2 = vdThreadStartRead(pDisk);
9651 AssertRC(rc2);
9652 fLockRead = true;
9653
9654 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9655 AssertPtrBreakStmt(pImage, cbSize = 0);
9656 cbSize = pImage->Backend->pfnGetSize(pImage->pBackendData);
9657 } while (0);
9658
9659 if (RT_UNLIKELY(fLockRead))
9660 {
9661 rc2 = vdThreadFinishRead(pDisk);
9662 AssertRC(rc2);
9663 }
9664
9665 LogFlowFunc(("returns %llu\n", cbSize));
9666 return cbSize;
9667}
9668
9669/**
9670 * Get total file size of an image in HDD container.
9671 *
9672 * @returns Virtual disk size in bytes.
9673 * @returns 0 if no image is opened in HDD container.
9674 * @param pDisk Pointer to HDD container.
9675 * @param nImage Image number, counts from 0. 0 is always base image of container.
9676 */
9677VBOXDDU_DECL(uint64_t) VDGetFileSize(PVBOXHDD pDisk, unsigned nImage)
9678{
9679 uint64_t cbSize;
9680 int rc2;
9681 bool fLockRead = false;
9682
9683 LogFlowFunc(("pDisk=%#p nImage=%u\n", pDisk, nImage));
9684 do
9685 {
9686 /* sanity check */
9687 AssertPtrBreakStmt(pDisk, cbSize = 0);
9688 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9689
9690 rc2 = vdThreadStartRead(pDisk);
9691 AssertRC(rc2);
9692 fLockRead = true;
9693
9694 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9695 AssertPtrBreakStmt(pImage, cbSize = 0);
9696 cbSize = pImage->Backend->pfnGetFileSize(pImage->pBackendData);
9697 } while (0);
9698
9699 if (RT_UNLIKELY(fLockRead))
9700 {
9701 rc2 = vdThreadFinishRead(pDisk);
9702 AssertRC(rc2);
9703 }
9704
9705 LogFlowFunc(("returns %llu\n", cbSize));
9706 return cbSize;
9707}
9708
9709/**
9710 * Get virtual disk PCHS geometry stored in HDD container.
9711 *
9712 * @returns VBox status code.
9713 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9714 * @returns VERR_VD_GEOMETRY_NOT_SET if no geometry present in the HDD container.
9715 * @param pDisk Pointer to HDD container.
9716 * @param nImage Image number, counts from 0. 0 is always base image of container.
9717 * @param pPCHSGeometry Where to store PCHS geometry. Not NULL.
9718 */
9719VBOXDDU_DECL(int) VDGetPCHSGeometry(PVBOXHDD pDisk, unsigned nImage,
9720 PVDGEOMETRY pPCHSGeometry)
9721{
9722 int rc = VINF_SUCCESS;
9723 int rc2;
9724 bool fLockRead = false;
9725
9726 LogFlowFunc(("pDisk=%#p nImage=%u pPCHSGeometry=%#p\n",
9727 pDisk, nImage, pPCHSGeometry));
9728 do
9729 {
9730 /* sanity check */
9731 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9732 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9733
9734 /* Check arguments. */
9735 AssertMsgBreakStmt(VALID_PTR(pPCHSGeometry),
9736 ("pPCHSGeometry=%#p\n", pPCHSGeometry),
9737 rc = VERR_INVALID_PARAMETER);
9738
9739 rc2 = vdThreadStartRead(pDisk);
9740 AssertRC(rc2);
9741 fLockRead = true;
9742
9743 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9744 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9745
9746 if (pImage == pDisk->pLast)
9747 {
9748 /* Use cached information if possible. */
9749 if (pDisk->PCHSGeometry.cCylinders != 0)
9750 *pPCHSGeometry = pDisk->PCHSGeometry;
9751 else
9752 rc = VERR_VD_GEOMETRY_NOT_SET;
9753 }
9754 else
9755 rc = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
9756 pPCHSGeometry);
9757 } while (0);
9758
9759 if (RT_UNLIKELY(fLockRead))
9760 {
9761 rc2 = vdThreadFinishRead(pDisk);
9762 AssertRC(rc2);
9763 }
9764
9765 LogFlowFunc(("%Rrc (PCHS=%u/%u/%u)\n", rc,
9766 pDisk->PCHSGeometry.cCylinders, pDisk->PCHSGeometry.cHeads,
9767 pDisk->PCHSGeometry.cSectors));
9768 return rc;
9769}
9770
9771/**
9772 * Store virtual disk PCHS geometry in HDD container.
9773 *
9774 * Note that in case of unrecoverable error all images in HDD container will be closed.
9775 *
9776 * @returns VBox status code.
9777 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9778 * @returns VERR_VD_GEOMETRY_NOT_SET if no geometry present in the HDD container.
9779 * @param pDisk Pointer to HDD container.
9780 * @param nImage Image number, counts from 0. 0 is always base image of container.
9781 * @param pPCHSGeometry Where to load PCHS geometry from. Not NULL.
9782 */
9783VBOXDDU_DECL(int) VDSetPCHSGeometry(PVBOXHDD pDisk, unsigned nImage,
9784 PCVDGEOMETRY pPCHSGeometry)
9785{
9786 int rc = VINF_SUCCESS;
9787 int rc2;
9788 bool fLockWrite = false;
9789
9790 LogFlowFunc(("pDisk=%#p nImage=%u pPCHSGeometry=%#p PCHS=%u/%u/%u\n",
9791 pDisk, nImage, pPCHSGeometry, pPCHSGeometry->cCylinders,
9792 pPCHSGeometry->cHeads, pPCHSGeometry->cSectors));
9793 do
9794 {
9795 /* sanity check */
9796 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9797 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9798
9799 /* Check arguments. */
9800 AssertMsgBreakStmt( VALID_PTR(pPCHSGeometry)
9801 && pPCHSGeometry->cHeads <= 16
9802 && pPCHSGeometry->cSectors <= 63,
9803 ("pPCHSGeometry=%#p PCHS=%u/%u/%u\n", pPCHSGeometry,
9804 pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads,
9805 pPCHSGeometry->cSectors),
9806 rc = VERR_INVALID_PARAMETER);
9807
9808 rc2 = vdThreadStartWrite(pDisk);
9809 AssertRC(rc2);
9810 fLockWrite = true;
9811
9812 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9813 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9814
9815 if (pImage == pDisk->pLast)
9816 {
9817 if ( pPCHSGeometry->cCylinders != pDisk->PCHSGeometry.cCylinders
9818 || pPCHSGeometry->cHeads != pDisk->PCHSGeometry.cHeads
9819 || pPCHSGeometry->cSectors != pDisk->PCHSGeometry.cSectors)
9820 {
9821 /* Only update geometry if it is changed. Avoids similar checks
9822 * in every backend. Most of the time the new geometry is set
9823 * to the previous values, so no need to go through the hassle
9824 * of updating an image which could be opened in read-only mode
9825 * right now. */
9826 rc = pImage->Backend->pfnSetPCHSGeometry(pImage->pBackendData,
9827 pPCHSGeometry);
9828
9829 /* Cache new geometry values in any case. */
9830 rc2 = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
9831 &pDisk->PCHSGeometry);
9832 if (RT_FAILURE(rc2))
9833 {
9834 pDisk->PCHSGeometry.cCylinders = 0;
9835 pDisk->PCHSGeometry.cHeads = 0;
9836 pDisk->PCHSGeometry.cSectors = 0;
9837 }
9838 else
9839 {
9840 /* Make sure the CHS geometry is properly clipped. */
9841 pDisk->PCHSGeometry.cHeads = RT_MIN(pDisk->PCHSGeometry.cHeads, 255);
9842 pDisk->PCHSGeometry.cSectors = RT_MIN(pDisk->PCHSGeometry.cSectors, 63);
9843 }
9844 }
9845 }
9846 else
9847 {
9848 VDGEOMETRY PCHS;
9849 rc = pImage->Backend->pfnGetPCHSGeometry(pImage->pBackendData,
9850 &PCHS);
9851 if ( RT_FAILURE(rc)
9852 || pPCHSGeometry->cCylinders != PCHS.cCylinders
9853 || pPCHSGeometry->cHeads != PCHS.cHeads
9854 || pPCHSGeometry->cSectors != PCHS.cSectors)
9855 {
9856 /* Only update geometry if it is changed. Avoids similar checks
9857 * in every backend. Most of the time the new geometry is set
9858 * to the previous values, so no need to go through the hassle
9859 * of updating an image which could be opened in read-only mode
9860 * right now. */
9861 rc = pImage->Backend->pfnSetPCHSGeometry(pImage->pBackendData,
9862 pPCHSGeometry);
9863 }
9864 }
9865 } while (0);
9866
9867 if (RT_UNLIKELY(fLockWrite))
9868 {
9869 rc2 = vdThreadFinishWrite(pDisk);
9870 AssertRC(rc2);
9871 }
9872
9873 LogFlowFunc(("returns %Rrc\n", rc));
9874 return rc;
9875}
9876
9877/**
9878 * Get virtual disk LCHS geometry stored in HDD container.
9879 *
9880 * @returns VBox status code.
9881 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9882 * @returns VERR_VD_GEOMETRY_NOT_SET if no geometry present in the HDD container.
9883 * @param pDisk Pointer to HDD container.
9884 * @param nImage Image number, counts from 0. 0 is always base image of container.
9885 * @param pLCHSGeometry Where to store LCHS geometry. Not NULL.
9886 */
9887VBOXDDU_DECL(int) VDGetLCHSGeometry(PVBOXHDD pDisk, unsigned nImage,
9888 PVDGEOMETRY pLCHSGeometry)
9889{
9890 int rc = VINF_SUCCESS;
9891 int rc2;
9892 bool fLockRead = false;
9893
9894 LogFlowFunc(("pDisk=%#p nImage=%u pLCHSGeometry=%#p\n",
9895 pDisk, nImage, pLCHSGeometry));
9896 do
9897 {
9898 /* sanity check */
9899 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9900 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9901
9902 /* Check arguments. */
9903 AssertMsgBreakStmt(VALID_PTR(pLCHSGeometry),
9904 ("pLCHSGeometry=%#p\n", pLCHSGeometry),
9905 rc = VERR_INVALID_PARAMETER);
9906
9907 rc2 = vdThreadStartRead(pDisk);
9908 AssertRC(rc2);
9909 fLockRead = true;
9910
9911 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9912 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9913
9914 if (pImage == pDisk->pLast)
9915 {
9916 /* Use cached information if possible. */
9917 if (pDisk->LCHSGeometry.cCylinders != 0)
9918 *pLCHSGeometry = pDisk->LCHSGeometry;
9919 else
9920 rc = VERR_VD_GEOMETRY_NOT_SET;
9921 }
9922 else
9923 rc = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
9924 pLCHSGeometry);
9925 } while (0);
9926
9927 if (RT_UNLIKELY(fLockRead))
9928 {
9929 rc2 = vdThreadFinishRead(pDisk);
9930 AssertRC(rc2);
9931 }
9932
9933 LogFlowFunc((": %Rrc (LCHS=%u/%u/%u)\n", rc,
9934 pDisk->LCHSGeometry.cCylinders, pDisk->LCHSGeometry.cHeads,
9935 pDisk->LCHSGeometry.cSectors));
9936 return rc;
9937}
9938
9939/**
9940 * Store virtual disk LCHS geometry in HDD container.
9941 *
9942 * Note that in case of unrecoverable error all images in HDD container will be closed.
9943 *
9944 * @returns VBox status code.
9945 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
9946 * @returns VERR_VD_GEOMETRY_NOT_SET if no geometry present in the HDD container.
9947 * @param pDisk Pointer to HDD container.
9948 * @param nImage Image number, counts from 0. 0 is always base image of container.
9949 * @param pLCHSGeometry Where to load LCHS geometry from. Not NULL.
9950 */
9951VBOXDDU_DECL(int) VDSetLCHSGeometry(PVBOXHDD pDisk, unsigned nImage,
9952 PCVDGEOMETRY pLCHSGeometry)
9953{
9954 int rc = VINF_SUCCESS;
9955 int rc2;
9956 bool fLockWrite = false;
9957
9958 LogFlowFunc(("pDisk=%#p nImage=%u pLCHSGeometry=%#p LCHS=%u/%u/%u\n",
9959 pDisk, nImage, pLCHSGeometry, pLCHSGeometry->cCylinders,
9960 pLCHSGeometry->cHeads, pLCHSGeometry->cSectors));
9961 do
9962 {
9963 /* sanity check */
9964 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
9965 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
9966
9967 /* Check arguments. */
9968 AssertMsgBreakStmt( VALID_PTR(pLCHSGeometry)
9969 && pLCHSGeometry->cHeads <= 255
9970 && pLCHSGeometry->cSectors <= 63,
9971 ("pLCHSGeometry=%#p LCHS=%u/%u/%u\n", pLCHSGeometry,
9972 pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads,
9973 pLCHSGeometry->cSectors),
9974 rc = VERR_INVALID_PARAMETER);
9975
9976 rc2 = vdThreadStartWrite(pDisk);
9977 AssertRC(rc2);
9978 fLockWrite = true;
9979
9980 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
9981 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
9982
9983 if (pImage == pDisk->pLast)
9984 {
9985 if ( pLCHSGeometry->cCylinders != pDisk->LCHSGeometry.cCylinders
9986 || pLCHSGeometry->cHeads != pDisk->LCHSGeometry.cHeads
9987 || pLCHSGeometry->cSectors != pDisk->LCHSGeometry.cSectors)
9988 {
9989 /* Only update geometry if it is changed. Avoids similar checks
9990 * in every backend. Most of the time the new geometry is set
9991 * to the previous values, so no need to go through the hassle
9992 * of updating an image which could be opened in read-only mode
9993 * right now. */
9994 rc = pImage->Backend->pfnSetLCHSGeometry(pImage->pBackendData,
9995 pLCHSGeometry);
9996
9997 /* Cache new geometry values in any case. */
9998 rc2 = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
9999 &pDisk->LCHSGeometry);
10000 if (RT_FAILURE(rc2))
10001 {
10002 pDisk->LCHSGeometry.cCylinders = 0;
10003 pDisk->LCHSGeometry.cHeads = 0;
10004 pDisk->LCHSGeometry.cSectors = 0;
10005 }
10006 else
10007 {
10008 /* Make sure the CHS geometry is properly clipped. */
10009 pDisk->LCHSGeometry.cHeads = RT_MIN(pDisk->LCHSGeometry.cHeads, 255);
10010 pDisk->LCHSGeometry.cSectors = RT_MIN(pDisk->LCHSGeometry.cSectors, 63);
10011 }
10012 }
10013 }
10014 else
10015 {
10016 VDGEOMETRY LCHS;
10017 rc = pImage->Backend->pfnGetLCHSGeometry(pImage->pBackendData,
10018 &LCHS);
10019 if ( RT_FAILURE(rc)
10020 || pLCHSGeometry->cCylinders != LCHS.cCylinders
10021 || pLCHSGeometry->cHeads != LCHS.cHeads
10022 || pLCHSGeometry->cSectors != LCHS.cSectors)
10023 {
10024 /* Only update geometry if it is changed. Avoids similar checks
10025 * in every backend. Most of the time the new geometry is set
10026 * to the previous values, so no need to go through the hassle
10027 * of updating an image which could be opened in read-only mode
10028 * right now. */
10029 rc = pImage->Backend->pfnSetLCHSGeometry(pImage->pBackendData,
10030 pLCHSGeometry);
10031 }
10032 }
10033 } while (0);
10034
10035 if (RT_UNLIKELY(fLockWrite))
10036 {
10037 rc2 = vdThreadFinishWrite(pDisk);
10038 AssertRC(rc2);
10039 }
10040
10041 LogFlowFunc(("returns %Rrc\n", rc));
10042 return rc;
10043}
10044
10045/**
10046 * Get version of image in HDD container.
10047 *
10048 * @returns VBox status code.
10049 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10050 * @param pDisk Pointer to HDD container.
10051 * @param nImage Image number, counts from 0. 0 is always base image of container.
10052 * @param puVersion Where to store the image version.
10053 */
10054VBOXDDU_DECL(int) VDGetVersion(PVBOXHDD pDisk, unsigned nImage,
10055 unsigned *puVersion)
10056{
10057 int rc = VINF_SUCCESS;
10058 int rc2;
10059 bool fLockRead = false;
10060
10061 LogFlowFunc(("pDisk=%#p nImage=%u puVersion=%#p\n",
10062 pDisk, nImage, puVersion));
10063 do
10064 {
10065 /* sanity check */
10066 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10067 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10068
10069 /* Check arguments. */
10070 AssertMsgBreakStmt(VALID_PTR(puVersion),
10071 ("puVersion=%#p\n", puVersion),
10072 rc = VERR_INVALID_PARAMETER);
10073
10074 rc2 = vdThreadStartRead(pDisk);
10075 AssertRC(rc2);
10076 fLockRead = true;
10077
10078 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10079 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10080
10081 *puVersion = pImage->Backend->pfnGetVersion(pImage->pBackendData);
10082 } while (0);
10083
10084 if (RT_UNLIKELY(fLockRead))
10085 {
10086 rc2 = vdThreadFinishRead(pDisk);
10087 AssertRC(rc2);
10088 }
10089
10090 LogFlowFunc(("returns %Rrc uVersion=%#x\n", rc, *puVersion));
10091 return rc;
10092}
10093
10094/**
10095 * List the capabilities of image backend in HDD container.
10096 *
10097 * @returns VBox status code.
10098 * @retval VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10099 * @param pDisk Pointer to the HDD container.
10100 * @param nImage Image number, counts from 0. 0 is always base image of container.
10101 * @param pBackendInfo Where to store the backend information.
10102 */
10103VBOXDDU_DECL(int) VDBackendInfoSingle(PVBOXHDD pDisk, unsigned nImage,
10104 PVDBACKENDINFO pBackendInfo)
10105{
10106 int rc = VINF_SUCCESS;
10107 int rc2;
10108 bool fLockRead = false;
10109
10110 LogFlowFunc(("pDisk=%#p nImage=%u pBackendInfo=%#p\n",
10111 pDisk, nImage, pBackendInfo));
10112 do
10113 {
10114 /* sanity check */
10115 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10116 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10117
10118 /* Check arguments. */
10119 AssertMsgBreakStmt(VALID_PTR(pBackendInfo),
10120 ("pBackendInfo=%#p\n", pBackendInfo),
10121 rc = VERR_INVALID_PARAMETER);
10122
10123 rc2 = vdThreadStartRead(pDisk);
10124 AssertRC(rc2);
10125 fLockRead = true;
10126
10127 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10128 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10129
10130 pBackendInfo->pszBackend = pImage->Backend->pszBackendName;
10131 pBackendInfo->uBackendCaps = pImage->Backend->uBackendCaps;
10132 pBackendInfo->paFileExtensions = pImage->Backend->paFileExtensions;
10133 pBackendInfo->paConfigInfo = pImage->Backend->paConfigInfo;
10134 } while (0);
10135
10136 if (RT_UNLIKELY(fLockRead))
10137 {
10138 rc2 = vdThreadFinishRead(pDisk);
10139 AssertRC(rc2);
10140 }
10141
10142 LogFlowFunc(("returns %Rrc\n", rc));
10143 return rc;
10144}
10145
10146/**
10147 * Get flags of image in HDD container.
10148 *
10149 * @returns VBox status code.
10150 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10151 * @param pDisk Pointer to HDD container.
10152 * @param nImage Image number, counts from 0. 0 is always base image of container.
10153 * @param puImageFlags Where to store the image flags.
10154 */
10155VBOXDDU_DECL(int) VDGetImageFlags(PVBOXHDD pDisk, unsigned nImage,
10156 unsigned *puImageFlags)
10157{
10158 int rc = VINF_SUCCESS;
10159 int rc2;
10160 bool fLockRead = false;
10161
10162 LogFlowFunc(("pDisk=%#p nImage=%u puImageFlags=%#p\n",
10163 pDisk, nImage, puImageFlags));
10164 do
10165 {
10166 /* sanity check */
10167 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10168 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10169
10170 /* Check arguments. */
10171 AssertMsgBreakStmt(VALID_PTR(puImageFlags),
10172 ("puImageFlags=%#p\n", puImageFlags),
10173 rc = VERR_INVALID_PARAMETER);
10174
10175 rc2 = vdThreadStartRead(pDisk);
10176 AssertRC(rc2);
10177 fLockRead = true;
10178
10179 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10180 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10181
10182 *puImageFlags = pImage->uImageFlags;
10183 } while (0);
10184
10185 if (RT_UNLIKELY(fLockRead))
10186 {
10187 rc2 = vdThreadFinishRead(pDisk);
10188 AssertRC(rc2);
10189 }
10190
10191 LogFlowFunc(("returns %Rrc uImageFlags=%#x\n", rc, *puImageFlags));
10192 return rc;
10193}
10194
10195/**
10196 * Get open flags of image in HDD container.
10197 *
10198 * @returns VBox status code.
10199 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10200 * @param pDisk Pointer to HDD container.
10201 * @param nImage Image number, counts from 0. 0 is always base image of container.
10202 * @param puOpenFlags Where to store the image open flags.
10203 */
10204VBOXDDU_DECL(int) VDGetOpenFlags(PVBOXHDD pDisk, unsigned nImage,
10205 unsigned *puOpenFlags)
10206{
10207 int rc = VINF_SUCCESS;
10208 int rc2;
10209 bool fLockRead = false;
10210
10211 LogFlowFunc(("pDisk=%#p nImage=%u puOpenFlags=%#p\n",
10212 pDisk, nImage, puOpenFlags));
10213 do
10214 {
10215 /* sanity check */
10216 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10217 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10218
10219 /* Check arguments. */
10220 AssertMsgBreakStmt(VALID_PTR(puOpenFlags),
10221 ("puOpenFlags=%#p\n", puOpenFlags),
10222 rc = VERR_INVALID_PARAMETER);
10223
10224 rc2 = vdThreadStartRead(pDisk);
10225 AssertRC(rc2);
10226 fLockRead = true;
10227
10228 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10229 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10230
10231 *puOpenFlags = pImage->Backend->pfnGetOpenFlags(pImage->pBackendData);
10232 } while (0);
10233
10234 if (RT_UNLIKELY(fLockRead))
10235 {
10236 rc2 = vdThreadFinishRead(pDisk);
10237 AssertRC(rc2);
10238 }
10239
10240 LogFlowFunc(("returns %Rrc uOpenFlags=%#x\n", rc, *puOpenFlags));
10241 return rc;
10242}
10243
10244/**
10245 * Set open flags of image in HDD container.
10246 * This operation may cause file locking changes and/or files being reopened.
10247 * Note that in case of unrecoverable error all images in HDD container will be closed.
10248 *
10249 * @returns VBox status code.
10250 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10251 * @param pDisk Pointer to HDD container.
10252 * @param nImage Image number, counts from 0. 0 is always base image of container.
10253 * @param uOpenFlags Image file open mode, see VD_OPEN_FLAGS_* constants.
10254 */
10255VBOXDDU_DECL(int) VDSetOpenFlags(PVBOXHDD pDisk, unsigned nImage,
10256 unsigned uOpenFlags)
10257{
10258 int rc;
10259 int rc2;
10260 bool fLockWrite = false;
10261
10262 LogFlowFunc(("pDisk=%#p uOpenFlags=%#u\n", pDisk, uOpenFlags));
10263 do
10264 {
10265 /* sanity check */
10266 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10267 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10268
10269 /* Check arguments. */
10270 AssertMsgBreakStmt((uOpenFlags & ~VD_OPEN_FLAGS_MASK) == 0,
10271 ("uOpenFlags=%#x\n", uOpenFlags),
10272 rc = VERR_INVALID_PARAMETER);
10273
10274 rc2 = vdThreadStartWrite(pDisk);
10275 AssertRC(rc2);
10276 fLockWrite = true;
10277
10278 /* Destroy any discard state because the image might be changed to readonly mode. */
10279 rc = vdDiscardStateDestroy(pDisk);
10280 if (RT_FAILURE(rc))
10281 break;
10282
10283 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10284 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10285
10286 rc = pImage->Backend->pfnSetOpenFlags(pImage->pBackendData,
10287 uOpenFlags & ~(VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS));
10288 if (RT_SUCCESS(rc))
10289 pImage->uOpenFlags = uOpenFlags & (VD_OPEN_FLAGS_HONOR_SAME | VD_OPEN_FLAGS_DISCARD | VD_OPEN_FLAGS_IGNORE_FLUSH | VD_OPEN_FLAGS_INFORM_ABOUT_ZERO_BLOCKS);
10290 } while (0);
10291
10292 if (RT_UNLIKELY(fLockWrite))
10293 {
10294 rc2 = vdThreadFinishWrite(pDisk);
10295 AssertRC(rc2);
10296 }
10297
10298 LogFlowFunc(("returns %Rrc\n", rc));
10299 return rc;
10300}
10301
10302/**
10303 * Get base filename of image in HDD container. Some image formats use
10304 * other filenames as well, so don't use this for anything but informational
10305 * purposes.
10306 *
10307 * @returns VBox status code.
10308 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10309 * @returns VERR_BUFFER_OVERFLOW if pszFilename buffer too small to hold filename.
10310 * @param pDisk Pointer to HDD container.
10311 * @param nImage Image number, counts from 0. 0 is always base image of container.
10312 * @param pszFilename Where to store the image file name.
10313 * @param cbFilename Size of buffer pszFilename points to.
10314 */
10315VBOXDDU_DECL(int) VDGetFilename(PVBOXHDD pDisk, unsigned nImage,
10316 char *pszFilename, unsigned cbFilename)
10317{
10318 int rc;
10319 int rc2;
10320 bool fLockRead = false;
10321
10322 LogFlowFunc(("pDisk=%#p nImage=%u pszFilename=%#p cbFilename=%u\n",
10323 pDisk, nImage, pszFilename, cbFilename));
10324 do
10325 {
10326 /* sanity check */
10327 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10328 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10329
10330 /* Check arguments. */
10331 AssertMsgBreakStmt(VALID_PTR(pszFilename) && *pszFilename,
10332 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
10333 rc = VERR_INVALID_PARAMETER);
10334 AssertMsgBreakStmt(cbFilename,
10335 ("cbFilename=%u\n", cbFilename),
10336 rc = VERR_INVALID_PARAMETER);
10337
10338 rc2 = vdThreadStartRead(pDisk);
10339 AssertRC(rc2);
10340 fLockRead = true;
10341
10342 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10343 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10344
10345 size_t cb = strlen(pImage->pszFilename);
10346 if (cb <= cbFilename)
10347 {
10348 strcpy(pszFilename, pImage->pszFilename);
10349 rc = VINF_SUCCESS;
10350 }
10351 else
10352 {
10353 strncpy(pszFilename, pImage->pszFilename, cbFilename - 1);
10354 pszFilename[cbFilename - 1] = '\0';
10355 rc = VERR_BUFFER_OVERFLOW;
10356 }
10357 } while (0);
10358
10359 if (RT_UNLIKELY(fLockRead))
10360 {
10361 rc2 = vdThreadFinishRead(pDisk);
10362 AssertRC(rc2);
10363 }
10364
10365 LogFlowFunc(("returns %Rrc, pszFilename=\"%s\"\n", rc, pszFilename));
10366 return rc;
10367}
10368
10369/**
10370 * Get the comment line of image in HDD container.
10371 *
10372 * @returns VBox status code.
10373 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10374 * @returns VERR_BUFFER_OVERFLOW if pszComment buffer too small to hold comment text.
10375 * @param pDisk Pointer to HDD container.
10376 * @param nImage Image number, counts from 0. 0 is always base image of container.
10377 * @param pszComment Where to store the comment string of image. NULL is ok.
10378 * @param cbComment The size of pszComment buffer. 0 is ok.
10379 */
10380VBOXDDU_DECL(int) VDGetComment(PVBOXHDD pDisk, unsigned nImage,
10381 char *pszComment, unsigned cbComment)
10382{
10383 int rc;
10384 int rc2;
10385 bool fLockRead = false;
10386
10387 LogFlowFunc(("pDisk=%#p nImage=%u pszComment=%#p cbComment=%u\n",
10388 pDisk, nImage, pszComment, cbComment));
10389 do
10390 {
10391 /* sanity check */
10392 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10393 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10394
10395 /* Check arguments. */
10396 AssertMsgBreakStmt(VALID_PTR(pszComment),
10397 ("pszComment=%#p \"%s\"\n", pszComment, pszComment),
10398 rc = VERR_INVALID_PARAMETER);
10399 AssertMsgBreakStmt(cbComment,
10400 ("cbComment=%u\n", cbComment),
10401 rc = VERR_INVALID_PARAMETER);
10402
10403 rc2 = vdThreadStartRead(pDisk);
10404 AssertRC(rc2);
10405 fLockRead = true;
10406
10407 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10408 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10409
10410 rc = pImage->Backend->pfnGetComment(pImage->pBackendData, pszComment,
10411 cbComment);
10412 } while (0);
10413
10414 if (RT_UNLIKELY(fLockRead))
10415 {
10416 rc2 = vdThreadFinishRead(pDisk);
10417 AssertRC(rc2);
10418 }
10419
10420 LogFlowFunc(("returns %Rrc, pszComment=\"%s\"\n", rc, pszComment));
10421 return rc;
10422}
10423
10424/**
10425 * Changes the comment line of image in HDD container.
10426 *
10427 * @returns VBox status code.
10428 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10429 * @param pDisk Pointer to HDD container.
10430 * @param nImage Image number, counts from 0. 0 is always base image of container.
10431 * @param pszComment New comment string (UTF-8). NULL is allowed to reset the comment.
10432 */
10433VBOXDDU_DECL(int) VDSetComment(PVBOXHDD pDisk, unsigned nImage,
10434 const char *pszComment)
10435{
10436 int rc;
10437 int rc2;
10438 bool fLockWrite = false;
10439
10440 LogFlowFunc(("pDisk=%#p nImage=%u pszComment=%#p \"%s\"\n",
10441 pDisk, nImage, pszComment, pszComment));
10442 do
10443 {
10444 /* sanity check */
10445 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10446 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10447
10448 /* Check arguments. */
10449 AssertMsgBreakStmt(VALID_PTR(pszComment) || pszComment == NULL,
10450 ("pszComment=%#p \"%s\"\n", pszComment, pszComment),
10451 rc = VERR_INVALID_PARAMETER);
10452
10453 rc2 = vdThreadStartWrite(pDisk);
10454 AssertRC(rc2);
10455 fLockWrite = true;
10456
10457 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10458 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10459
10460 rc = pImage->Backend->pfnSetComment(pImage->pBackendData, pszComment);
10461 } while (0);
10462
10463 if (RT_UNLIKELY(fLockWrite))
10464 {
10465 rc2 = vdThreadFinishWrite(pDisk);
10466 AssertRC(rc2);
10467 }
10468
10469 LogFlowFunc(("returns %Rrc\n", rc));
10470 return rc;
10471}
10472
10473
10474/**
10475 * Get UUID of image in HDD container.
10476 *
10477 * @returns VBox status code.
10478 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10479 * @param pDisk Pointer to HDD container.
10480 * @param nImage Image number, counts from 0. 0 is always base image of container.
10481 * @param pUuid Where to store the image creation UUID.
10482 */
10483VBOXDDU_DECL(int) VDGetUuid(PVBOXHDD pDisk, unsigned nImage, PRTUUID pUuid)
10484{
10485 int rc;
10486 int rc2;
10487 bool fLockRead = false;
10488
10489 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p\n", pDisk, nImage, pUuid));
10490 do
10491 {
10492 /* sanity check */
10493 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10494 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10495
10496 /* Check arguments. */
10497 AssertMsgBreakStmt(VALID_PTR(pUuid),
10498 ("pUuid=%#p\n", pUuid),
10499 rc = VERR_INVALID_PARAMETER);
10500
10501 rc2 = vdThreadStartRead(pDisk);
10502 AssertRC(rc2);
10503 fLockRead = true;
10504
10505 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10506 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10507
10508 rc = pImage->Backend->pfnGetUuid(pImage->pBackendData, pUuid);
10509 } while (0);
10510
10511 if (RT_UNLIKELY(fLockRead))
10512 {
10513 rc2 = vdThreadFinishRead(pDisk);
10514 AssertRC(rc2);
10515 }
10516
10517 LogFlowFunc(("returns %Rrc, Uuid={%RTuuid}\n", rc, pUuid));
10518 return rc;
10519}
10520
10521/**
10522 * Set the image's UUID. Should not be used by normal applications.
10523 *
10524 * @returns VBox status code.
10525 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10526 * @param pDisk Pointer to HDD container.
10527 * @param nImage Image number, counts from 0. 0 is always base image of container.
10528 * @param pUuid New UUID of the image. If NULL, a new UUID is created.
10529 */
10530VBOXDDU_DECL(int) VDSetUuid(PVBOXHDD pDisk, unsigned nImage, PCRTUUID pUuid)
10531{
10532 int rc;
10533 int rc2;
10534 bool fLockWrite = false;
10535
10536 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p {%RTuuid}\n",
10537 pDisk, nImage, pUuid, pUuid));
10538 do
10539 {
10540 /* sanity check */
10541 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10542 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10543
10544 AssertMsgBreakStmt(VALID_PTR(pUuid) || pUuid == NULL,
10545 ("pUuid=%#p\n", pUuid),
10546 rc = VERR_INVALID_PARAMETER);
10547
10548 rc2 = vdThreadStartWrite(pDisk);
10549 AssertRC(rc2);
10550 fLockWrite = true;
10551
10552 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10553 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10554
10555 RTUUID Uuid;
10556 if (!pUuid)
10557 {
10558 RTUuidCreate(&Uuid);
10559 pUuid = &Uuid;
10560 }
10561 rc = pImage->Backend->pfnSetUuid(pImage->pBackendData, pUuid);
10562 } while (0);
10563
10564 if (RT_UNLIKELY(fLockWrite))
10565 {
10566 rc2 = vdThreadFinishWrite(pDisk);
10567 AssertRC(rc2);
10568 }
10569
10570 LogFlowFunc(("returns %Rrc\n", rc));
10571 return rc;
10572}
10573
10574/**
10575 * Get last modification UUID of image in HDD container.
10576 *
10577 * @returns VBox status code.
10578 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10579 * @param pDisk Pointer to HDD container.
10580 * @param nImage Image number, counts from 0. 0 is always base image of container.
10581 * @param pUuid Where to store the image modification UUID.
10582 */
10583VBOXDDU_DECL(int) VDGetModificationUuid(PVBOXHDD pDisk, unsigned nImage, PRTUUID pUuid)
10584{
10585 int rc = VINF_SUCCESS;
10586 int rc2;
10587 bool fLockRead = false;
10588
10589 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p\n", pDisk, nImage, pUuid));
10590 do
10591 {
10592 /* sanity check */
10593 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10594 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10595
10596 /* Check arguments. */
10597 AssertMsgBreakStmt(VALID_PTR(pUuid),
10598 ("pUuid=%#p\n", pUuid),
10599 rc = VERR_INVALID_PARAMETER);
10600
10601 rc2 = vdThreadStartRead(pDisk);
10602 AssertRC(rc2);
10603 fLockRead = true;
10604
10605 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10606 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10607
10608 rc = pImage->Backend->pfnGetModificationUuid(pImage->pBackendData,
10609 pUuid);
10610 } while (0);
10611
10612 if (RT_UNLIKELY(fLockRead))
10613 {
10614 rc2 = vdThreadFinishRead(pDisk);
10615 AssertRC(rc2);
10616 }
10617
10618 LogFlowFunc(("returns %Rrc, Uuid={%RTuuid}\n", rc, pUuid));
10619 return rc;
10620}
10621
10622/**
10623 * Set the image's last modification UUID. Should not be used by normal applications.
10624 *
10625 * @returns VBox status code.
10626 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10627 * @param pDisk Pointer to HDD container.
10628 * @param nImage Image number, counts from 0. 0 is always base image of container.
10629 * @param pUuid New modification UUID of the image. If NULL, a new UUID is created.
10630 */
10631VBOXDDU_DECL(int) VDSetModificationUuid(PVBOXHDD pDisk, unsigned nImage, PCRTUUID pUuid)
10632{
10633 int rc;
10634 int rc2;
10635 bool fLockWrite = false;
10636
10637 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p {%RTuuid}\n",
10638 pDisk, nImage, pUuid, pUuid));
10639 do
10640 {
10641 /* sanity check */
10642 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10643 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10644
10645 /* Check arguments. */
10646 AssertMsgBreakStmt(VALID_PTR(pUuid) || pUuid == NULL,
10647 ("pUuid=%#p\n", pUuid),
10648 rc = VERR_INVALID_PARAMETER);
10649
10650 rc2 = vdThreadStartWrite(pDisk);
10651 AssertRC(rc2);
10652 fLockWrite = true;
10653
10654 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10655 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10656
10657 RTUUID Uuid;
10658 if (!pUuid)
10659 {
10660 RTUuidCreate(&Uuid);
10661 pUuid = &Uuid;
10662 }
10663 rc = pImage->Backend->pfnSetModificationUuid(pImage->pBackendData,
10664 pUuid);
10665 } while (0);
10666
10667 if (RT_UNLIKELY(fLockWrite))
10668 {
10669 rc2 = vdThreadFinishWrite(pDisk);
10670 AssertRC(rc2);
10671 }
10672
10673 LogFlowFunc(("returns %Rrc\n", rc));
10674 return rc;
10675}
10676
10677/**
10678 * Get parent UUID of image in HDD container.
10679 *
10680 * @returns VBox status code.
10681 * @returns VERR_VD_IMAGE_NOT_FOUND if image with specified number was not opened.
10682 * @param pDisk Pointer to HDD container.
10683 * @param nImage Image number, counts from 0. 0 is always base image of container.
10684 * @param pUuid Where to store the parent image UUID.
10685 */
10686VBOXDDU_DECL(int) VDGetParentUuid(PVBOXHDD pDisk, unsigned nImage,
10687 PRTUUID pUuid)
10688{
10689 int rc = VINF_SUCCESS;
10690 int rc2;
10691 bool fLockRead = false;
10692
10693 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p\n", pDisk, nImage, pUuid));
10694 do
10695 {
10696 /* sanity check */
10697 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10698 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10699
10700 /* Check arguments. */
10701 AssertMsgBreakStmt(VALID_PTR(pUuid),
10702 ("pUuid=%#p\n", pUuid),
10703 rc = VERR_INVALID_PARAMETER);
10704
10705 rc2 = vdThreadStartRead(pDisk);
10706 AssertRC(rc2);
10707 fLockRead = true;
10708
10709 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10710 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10711
10712 rc = pImage->Backend->pfnGetParentUuid(pImage->pBackendData, pUuid);
10713 } while (0);
10714
10715 if (RT_UNLIKELY(fLockRead))
10716 {
10717 rc2 = vdThreadFinishRead(pDisk);
10718 AssertRC(rc2);
10719 }
10720
10721 LogFlowFunc(("returns %Rrc, Uuid={%RTuuid}\n", rc, pUuid));
10722 return rc;
10723}
10724
10725/**
10726 * Set the image's parent UUID. Should not be used by normal applications.
10727 *
10728 * @returns VBox status code.
10729 * @param pDisk Pointer to HDD container.
10730 * @param nImage Image number, counts from 0. 0 is always base image of container.
10731 * @param pUuid New parent UUID of the image. If NULL, a new UUID is created.
10732 */
10733VBOXDDU_DECL(int) VDSetParentUuid(PVBOXHDD pDisk, unsigned nImage,
10734 PCRTUUID pUuid)
10735{
10736 int rc;
10737 int rc2;
10738 bool fLockWrite = false;
10739
10740 LogFlowFunc(("pDisk=%#p nImage=%u pUuid=%#p {%RTuuid}\n",
10741 pDisk, nImage, pUuid, pUuid));
10742 do
10743 {
10744 /* sanity check */
10745 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10746 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10747
10748 /* Check arguments. */
10749 AssertMsgBreakStmt(VALID_PTR(pUuid) || pUuid == NULL,
10750 ("pUuid=%#p\n", pUuid),
10751 rc = VERR_INVALID_PARAMETER);
10752
10753 rc2 = vdThreadStartWrite(pDisk);
10754 AssertRC(rc2);
10755 fLockWrite = true;
10756
10757 PVDIMAGE pImage = vdGetImageByNumber(pDisk, nImage);
10758 AssertPtrBreakStmt(pImage, rc = VERR_VD_IMAGE_NOT_FOUND);
10759
10760 RTUUID Uuid;
10761 if (!pUuid)
10762 {
10763 RTUuidCreate(&Uuid);
10764 pUuid = &Uuid;
10765 }
10766 rc = pImage->Backend->pfnSetParentUuid(pImage->pBackendData, pUuid);
10767 } while (0);
10768
10769 if (RT_UNLIKELY(fLockWrite))
10770 {
10771 rc2 = vdThreadFinishWrite(pDisk);
10772 AssertRC(rc2);
10773 }
10774
10775 LogFlowFunc(("returns %Rrc\n", rc));
10776 return rc;
10777}
10778
10779
10780/**
10781 * Debug helper - dumps all opened images in HDD container into the log file.
10782 *
10783 * @param pDisk Pointer to HDD container.
10784 */
10785VBOXDDU_DECL(void) VDDumpImages(PVBOXHDD pDisk)
10786{
10787 int rc2;
10788 bool fLockRead = false;
10789
10790 do
10791 {
10792 /* sanity check */
10793 AssertPtrBreak(pDisk);
10794 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10795
10796 if (!pDisk->pInterfaceError || !VALID_PTR(pDisk->pInterfaceError->pfnMessage))
10797 pDisk->pInterfaceError->pfnMessage = vdLogMessage;
10798
10799 rc2 = vdThreadStartRead(pDisk);
10800 AssertRC(rc2);
10801 fLockRead = true;
10802
10803 vdMessageWrapper(pDisk, "--- Dumping VD Disk, Images=%u\n", pDisk->cImages);
10804 for (PVDIMAGE pImage = pDisk->pBase; pImage; pImage = pImage->pNext)
10805 {
10806 vdMessageWrapper(pDisk, "Dumping VD image \"%s\" (Backend=%s)\n",
10807 pImage->pszFilename, pImage->Backend->pszBackendName);
10808 pImage->Backend->pfnDump(pImage->pBackendData);
10809 }
10810 } while (0);
10811
10812 if (RT_UNLIKELY(fLockRead))
10813 {
10814 rc2 = vdThreadFinishRead(pDisk);
10815 AssertRC(rc2);
10816 }
10817}
10818
10819
10820VBOXDDU_DECL(int) VDDiscardRanges(PVBOXHDD pDisk, PCRTRANGE paRanges, unsigned cRanges)
10821{
10822 int rc;
10823 int rc2;
10824 bool fLockWrite = false;
10825
10826 LogFlowFunc(("pDisk=%#p paRanges=%#p cRanges=%u\n",
10827 pDisk, paRanges, cRanges));
10828 do
10829 {
10830 /* sanity check */
10831 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10832 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10833
10834 /* Check arguments. */
10835 AssertMsgBreakStmt(cRanges,
10836 ("cRanges=%u\n", cRanges),
10837 rc = VERR_INVALID_PARAMETER);
10838 AssertMsgBreakStmt(VALID_PTR(paRanges),
10839 ("paRanges=%#p\n", paRanges),
10840 rc = VERR_INVALID_PARAMETER);
10841
10842 rc2 = vdThreadStartWrite(pDisk);
10843 AssertRC(rc2);
10844 fLockWrite = true;
10845
10846 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
10847
10848 AssertMsgBreakStmt(pDisk->pLast->uOpenFlags & VD_OPEN_FLAGS_DISCARD,
10849 ("Discarding not supported\n"),
10850 rc = VERR_NOT_SUPPORTED);
10851
10852 VDIOCTX IoCtx;
10853 RTSEMEVENT hEventComplete = NIL_RTSEMEVENT;
10854
10855 rc = RTSemEventCreate(&hEventComplete);
10856 if (RT_FAILURE(rc))
10857 break;
10858
10859 vdIoCtxDiscardInit(&IoCtx, pDisk, paRanges, cRanges,
10860 vdIoCtxSyncComplete, pDisk, hEventComplete, NULL,
10861 vdDiscardHelperAsync, VDIOCTX_FLAGS_SYNC | VDIOCTX_FLAGS_DONT_FREE);
10862 rc = vdIoCtxProcessSync(&IoCtx, hEventComplete);
10863
10864 RTSemEventDestroy(hEventComplete);
10865 } while (0);
10866
10867 if (RT_UNLIKELY(fLockWrite))
10868 {
10869 rc2 = vdThreadFinishWrite(pDisk);
10870 AssertRC(rc2);
10871 }
10872
10873 LogFlowFunc(("returns %Rrc\n", rc));
10874 return rc;
10875}
10876
10877
10878VBOXDDU_DECL(int) VDAsyncRead(PVBOXHDD pDisk, uint64_t uOffset, size_t cbRead,
10879 PCRTSGBUF pcSgBuf,
10880 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
10881 void *pvUser1, void *pvUser2)
10882{
10883 int rc = VERR_VD_BLOCK_FREE;
10884 int rc2;
10885 bool fLockRead = false;
10886 PVDIOCTX pIoCtx = NULL;
10887
10888 LogFlowFunc(("pDisk=%#p uOffset=%llu pcSgBuf=%#p cbRead=%zu pvUser1=%#p pvUser2=%#p\n",
10889 pDisk, uOffset, pcSgBuf, cbRead, pvUser1, pvUser2));
10890
10891 do
10892 {
10893 /* sanity check */
10894 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10895 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10896
10897 /* Check arguments. */
10898 AssertMsgBreakStmt(cbRead,
10899 ("cbRead=%zu\n", cbRead),
10900 rc = VERR_INVALID_PARAMETER);
10901 AssertMsgBreakStmt(VALID_PTR(pcSgBuf),
10902 ("pcSgBuf=%#p\n", pcSgBuf),
10903 rc = VERR_INVALID_PARAMETER);
10904
10905 rc2 = vdThreadStartRead(pDisk);
10906 AssertRC(rc2);
10907 fLockRead = true;
10908
10909 AssertMsgBreakStmt(uOffset + cbRead <= pDisk->cbSize,
10910 ("uOffset=%llu cbRead=%zu pDisk->cbSize=%llu\n",
10911 uOffset, cbRead, pDisk->cbSize),
10912 rc = VERR_INVALID_PARAMETER);
10913 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
10914
10915 pIoCtx = vdIoCtxRootAlloc(pDisk, VDIOCTXTXDIR_READ, uOffset,
10916 cbRead, pDisk->pLast, pcSgBuf,
10917 pfnComplete, pvUser1, pvUser2,
10918 NULL, vdReadHelperAsync,
10919 VDIOCTX_FLAGS_ZERO_FREE_BLOCKS);
10920 if (!pIoCtx)
10921 {
10922 rc = VERR_NO_MEMORY;
10923 break;
10924 }
10925
10926 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
10927 if (rc == VINF_VD_ASYNC_IO_FINISHED)
10928 {
10929 if (ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
10930 vdIoCtxFree(pDisk, pIoCtx);
10931 else
10932 rc = VERR_VD_ASYNC_IO_IN_PROGRESS; /* Let the other handler complete the request. */
10933 }
10934 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS) /* Another error */
10935 vdIoCtxFree(pDisk, pIoCtx);
10936
10937 } while (0);
10938
10939 if (RT_UNLIKELY(fLockRead) && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
10940 {
10941 rc2 = vdThreadFinishRead(pDisk);
10942 AssertRC(rc2);
10943 }
10944
10945 LogFlowFunc(("returns %Rrc\n", rc));
10946 return rc;
10947}
10948
10949
10950VBOXDDU_DECL(int) VDAsyncWrite(PVBOXHDD pDisk, uint64_t uOffset, size_t cbWrite,
10951 PCRTSGBUF pcSgBuf,
10952 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
10953 void *pvUser1, void *pvUser2)
10954{
10955 int rc;
10956 int rc2;
10957 bool fLockWrite = false;
10958 PVDIOCTX pIoCtx = NULL;
10959
10960 LogFlowFunc(("pDisk=%#p uOffset=%llu cSgBuf=%#p cbWrite=%zu pvUser1=%#p pvUser2=%#p\n",
10961 pDisk, uOffset, pcSgBuf, cbWrite, pvUser1, pvUser2));
10962 do
10963 {
10964 /* sanity check */
10965 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
10966 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
10967
10968 /* Check arguments. */
10969 AssertMsgBreakStmt(cbWrite,
10970 ("cbWrite=%zu\n", cbWrite),
10971 rc = VERR_INVALID_PARAMETER);
10972 AssertMsgBreakStmt(VALID_PTR(pcSgBuf),
10973 ("pcSgBuf=%#p\n", pcSgBuf),
10974 rc = VERR_INVALID_PARAMETER);
10975
10976 rc2 = vdThreadStartWrite(pDisk);
10977 AssertRC(rc2);
10978 fLockWrite = true;
10979
10980 AssertMsgBreakStmt(uOffset + cbWrite <= pDisk->cbSize,
10981 ("uOffset=%llu cbWrite=%zu pDisk->cbSize=%llu\n",
10982 uOffset, cbWrite, pDisk->cbSize),
10983 rc = VERR_INVALID_PARAMETER);
10984 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
10985
10986 pIoCtx = vdIoCtxRootAlloc(pDisk, VDIOCTXTXDIR_WRITE, uOffset,
10987 cbWrite, pDisk->pLast, pcSgBuf,
10988 pfnComplete, pvUser1, pvUser2,
10989 NULL, vdWriteHelperAsync,
10990 VDIOCTX_FLAGS_DEFAULT);
10991 if (!pIoCtx)
10992 {
10993 rc = VERR_NO_MEMORY;
10994 break;
10995 }
10996
10997 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
10998 if (rc == VINF_VD_ASYNC_IO_FINISHED)
10999 {
11000 if (ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
11001 vdIoCtxFree(pDisk, pIoCtx);
11002 else
11003 rc = VERR_VD_ASYNC_IO_IN_PROGRESS; /* Let the other handler complete the request. */
11004 }
11005 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS) /* Another error */
11006 vdIoCtxFree(pDisk, pIoCtx);
11007 } while (0);
11008
11009 if (RT_UNLIKELY(fLockWrite) && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
11010 {
11011 rc2 = vdThreadFinishWrite(pDisk);
11012 AssertRC(rc2);
11013 }
11014
11015 LogFlowFunc(("returns %Rrc\n", rc));
11016 return rc;
11017}
11018
11019
11020VBOXDDU_DECL(int) VDAsyncFlush(PVBOXHDD pDisk, PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
11021 void *pvUser1, void *pvUser2)
11022{
11023 int rc;
11024 int rc2;
11025 bool fLockWrite = false;
11026 PVDIOCTX pIoCtx = NULL;
11027
11028 LogFlowFunc(("pDisk=%#p\n", pDisk));
11029
11030 do
11031 {
11032 /* sanity check */
11033 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
11034 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
11035
11036 rc2 = vdThreadStartWrite(pDisk);
11037 AssertRC(rc2);
11038 fLockWrite = true;
11039
11040 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
11041
11042 pIoCtx = vdIoCtxRootAlloc(pDisk, VDIOCTXTXDIR_FLUSH, 0,
11043 0, pDisk->pLast, NULL,
11044 pfnComplete, pvUser1, pvUser2,
11045 NULL, vdFlushHelperAsync,
11046 VDIOCTX_FLAGS_DEFAULT);
11047 if (!pIoCtx)
11048 {
11049 rc = VERR_NO_MEMORY;
11050 break;
11051 }
11052
11053 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
11054 if (rc == VINF_VD_ASYNC_IO_FINISHED)
11055 {
11056 if (ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
11057 vdIoCtxFree(pDisk, pIoCtx);
11058 else
11059 rc = VERR_VD_ASYNC_IO_IN_PROGRESS; /* Let the other handler complete the request. */
11060 }
11061 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS) /* Another error */
11062 vdIoCtxFree(pDisk, pIoCtx);
11063 } while (0);
11064
11065 if (RT_UNLIKELY(fLockWrite) && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
11066 {
11067 rc2 = vdThreadFinishWrite(pDisk);
11068 AssertRC(rc2);
11069 }
11070
11071 LogFlowFunc(("returns %Rrc\n", rc));
11072 return rc;
11073}
11074
11075VBOXDDU_DECL(int) VDAsyncDiscardRanges(PVBOXHDD pDisk, PCRTRANGE paRanges, unsigned cRanges,
11076 PFNVDASYNCTRANSFERCOMPLETE pfnComplete,
11077 void *pvUser1, void *pvUser2)
11078{
11079 int rc;
11080 int rc2;
11081 bool fLockWrite = false;
11082 PVDIOCTX pIoCtx = NULL;
11083
11084 LogFlowFunc(("pDisk=%#p\n", pDisk));
11085
11086 do
11087 {
11088 /* sanity check */
11089 AssertPtrBreakStmt(pDisk, rc = VERR_INVALID_PARAMETER);
11090 AssertMsg(pDisk->u32Signature == VBOXHDDDISK_SIGNATURE, ("u32Signature=%08x\n", pDisk->u32Signature));
11091
11092 rc2 = vdThreadStartWrite(pDisk);
11093 AssertRC(rc2);
11094 fLockWrite = true;
11095
11096 AssertPtrBreakStmt(pDisk->pLast, rc = VERR_VD_NOT_OPENED);
11097
11098 pIoCtx = vdIoCtxDiscardAlloc(pDisk, paRanges, cRanges,
11099 pfnComplete, pvUser1, pvUser2, NULL,
11100 vdDiscardHelperAsync,
11101 VDIOCTX_FLAGS_DEFAULT);
11102 if (!pIoCtx)
11103 {
11104 rc = VERR_NO_MEMORY;
11105 break;
11106 }
11107
11108 rc = vdIoCtxProcessTryLockDefer(pIoCtx);
11109 if (rc == VINF_VD_ASYNC_IO_FINISHED)
11110 {
11111 if (ASMAtomicCmpXchgBool(&pIoCtx->fComplete, true, false))
11112 vdIoCtxFree(pDisk, pIoCtx);
11113 else
11114 rc = VERR_VD_ASYNC_IO_IN_PROGRESS; /* Let the other handler complete the request. */
11115 }
11116 else if (rc != VERR_VD_ASYNC_IO_IN_PROGRESS) /* Another error */
11117 vdIoCtxFree(pDisk, pIoCtx);
11118 } while (0);
11119
11120 if (RT_UNLIKELY(fLockWrite) && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
11121 {
11122 rc2 = vdThreadFinishWrite(pDisk);
11123 AssertRC(rc2);
11124 }
11125
11126 LogFlowFunc(("returns %Rrc\n", rc));
11127 return rc;
11128}
11129
11130VBOXDDU_DECL(int) VDRepair(PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
11131 const char *pszFilename, const char *pszBackend,
11132 uint32_t fFlags)
11133{
11134 int rc = VERR_NOT_SUPPORTED;
11135 PCVDIMAGEBACKEND pBackend = NULL;
11136 VDINTERFACEIOINT VDIfIoInt;
11137 VDINTERFACEIO VDIfIoFallback;
11138 PVDINTERFACEIO pInterfaceIo;
11139
11140 LogFlowFunc(("pszFilename=\"%s\"\n", pszFilename));
11141 /* Check arguments. */
11142 AssertMsgReturn(VALID_PTR(pszFilename) && *pszFilename,
11143 ("pszFilename=%#p \"%s\"\n", pszFilename, pszFilename),
11144 VERR_INVALID_PARAMETER);
11145 AssertMsgReturn(VALID_PTR(pszBackend),
11146 ("pszBackend=%#p\n", pszBackend),
11147 VERR_INVALID_PARAMETER);
11148 AssertMsgReturn((fFlags & ~VD_REPAIR_FLAGS_MASK) == 0,
11149 ("fFlags=%#x\n", fFlags),
11150 VERR_INVALID_PARAMETER);
11151
11152 pInterfaceIo = VDIfIoGet(pVDIfsImage);
11153 if (!pInterfaceIo)
11154 {
11155 /*
11156 * Caller doesn't provide an I/O interface, create our own using the
11157 * native file API.
11158 */
11159 vdIfIoFallbackCallbacksSetup(&VDIfIoFallback);
11160 pInterfaceIo = &VDIfIoFallback;
11161 }
11162
11163 /* Set up the internal I/O interface. */
11164 AssertReturn(!VDIfIoIntGet(pVDIfsImage), VERR_INVALID_PARAMETER);
11165 VDIfIoInt.pfnOpen = vdIOIntOpenLimited;
11166 VDIfIoInt.pfnClose = vdIOIntCloseLimited;
11167 VDIfIoInt.pfnDelete = vdIOIntDeleteLimited;
11168 VDIfIoInt.pfnMove = vdIOIntMoveLimited;
11169 VDIfIoInt.pfnGetFreeSpace = vdIOIntGetFreeSpaceLimited;
11170 VDIfIoInt.pfnGetModificationTime = vdIOIntGetModificationTimeLimited;
11171 VDIfIoInt.pfnGetSize = vdIOIntGetSizeLimited;
11172 VDIfIoInt.pfnSetSize = vdIOIntSetSizeLimited;
11173 VDIfIoInt.pfnReadUser = vdIOIntReadUserLimited;
11174 VDIfIoInt.pfnWriteUser = vdIOIntWriteUserLimited;
11175 VDIfIoInt.pfnReadMeta = vdIOIntReadMetaLimited;
11176 VDIfIoInt.pfnWriteMeta = vdIOIntWriteMetaLimited;
11177 VDIfIoInt.pfnFlush = vdIOIntFlushLimited;
11178 rc = VDInterfaceAdd(&VDIfIoInt.Core, "VD_IOINT", VDINTERFACETYPE_IOINT,
11179 pInterfaceIo, sizeof(VDINTERFACEIOINT), &pVDIfsImage);
11180 AssertRC(rc);
11181
11182 rc = vdFindBackend(pszBackend, &pBackend);
11183 if (RT_SUCCESS(rc))
11184 {
11185 if (pBackend->pfnRepair)
11186 rc = pBackend->pfnRepair(pszFilename, pVDIfsDisk, pVDIfsImage, fFlags);
11187 else
11188 rc = VERR_VD_IMAGE_REPAIR_NOT_SUPPORTED;
11189 }
11190
11191 LogFlowFunc(("returns %Rrc\n", rc));
11192 return rc;
11193}
11194
11195
11196/*
11197 * generic plugin functions
11198 */
11199
11200/**
11201 * @interface_method_impl{VDIMAGEBACKEND,pfnComposeLocation}
11202 */
11203DECLCALLBACK(int) genericFileComposeLocation(PVDINTERFACE pConfig, char **pszLocation)
11204{
11205 RT_NOREF1(pConfig);
11206 *pszLocation = NULL;
11207 return VINF_SUCCESS;
11208}
11209
11210/**
11211 * @interface_method_impl{VDIMAGEBACKEND,pfnComposeName}
11212 */
11213DECLCALLBACK(int) genericFileComposeName(PVDINTERFACE pConfig, char **pszName)
11214{
11215 RT_NOREF1(pConfig);
11216 *pszName = NULL;
11217 return VINF_SUCCESS;
11218}
11219
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette