VirtualBox

source: vbox/trunk/src/VBox/Storage/QED.cpp@ 62956

Last change on this file since 62956 was 62873, checked in by vboxsync, 8 years ago

Storage: warnings.

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 90.0 KB
Line 
1/* $Id: QED.cpp 62873 2016-08-02 14:00:15Z vboxsync $ */
2/** @file
3 * QED - QED Disk image.
4 */
5
6/*
7 * Copyright (C) 2011-2016 Oracle Corporation
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 */
17
18
19/*********************************************************************************************************************************
20* Header Files *
21*********************************************************************************************************************************/
22#define LOG_GROUP LOG_GROUP_VD_QED
23#include <VBox/vd-plugin.h>
24#include <VBox/err.h>
25
26#include <VBox/log.h>
27#include <iprt/asm.h>
28#include <iprt/assert.h>
29#include <iprt/string.h>
30#include <iprt/alloc.h>
31#include <iprt/path.h>
32#include <iprt/list.h>
33
34#include "VDBackends.h"
35
36/**
37 * The QED backend implements support for the qemu enhanced disk format (short QED)
38 * The specification for the format is available under http://wiki.qemu.org/Features/QED/Specification
39 *
40 * Missing things to implement:
41 * - compaction
42 * - resizing which requires block relocation (very rare case)
43 */
44
45
46/*********************************************************************************************************************************
47* Structures in a QED image, little endian *
48*********************************************************************************************************************************/
49
50#pragma pack(1)
51typedef struct QedHeader
52{
53 /** Magic value. */
54 uint32_t u32Magic;
55 /** Cluster size in bytes. */
56 uint32_t u32ClusterSize;
57 /** Size of L1 and L2 tables in clusters. */
58 uint32_t u32TableSize;
59 /** size of this header structure in clusters. */
60 uint32_t u32HeaderSize;
61 /** Features used for the image. */
62 uint64_t u64FeatureFlags;
63 /** Compatibility features used for the image. */
64 uint64_t u64CompatFeatureFlags;
65 /** Self resetting feature bits. */
66 uint64_t u64AutoresetFeatureFlags;
67 /** Offset of the L1 table in bytes. */
68 uint64_t u64OffL1Table;
69 /** Logical image size as seen by the guest. */
70 uint64_t u64Size;
71 /** Offset of the backing filename in bytes. */
72 uint32_t u32OffBackingFilename;
73 /** Size of the backing filename. */
74 uint32_t u32BackingFilenameSize;
75} QedHeader;
76#pragma pack()
77/** Pointer to a on disk QED header. */
78typedef QedHeader *PQedHeader;
79
80/** QED magic value. */
81#define QED_MAGIC UINT32_C(0x00444551) /* QED\0 */
82/** Cluster size minimum. */
83#define QED_CLUSTER_SIZE_MIN RT_BIT(12)
84/** Cluster size maximum. */
85#define QED_CLUSTER_SIZE_MAX RT_BIT(26)
86/** L1 and L2 Table size minimum. */
87#define QED_TABLE_SIZE_MIN 1
88/** L1 and L2 Table size maximum. */
89#define QED_TABLE_SIZE_MAX 16
90
91/** QED default cluster size when creating an image. */
92#define QED_CLUSTER_SIZE_DEFAULT (64 * _1K)
93/** The default table size in clusters. */
94#define QED_TABLE_SIZE_DEFAULT 4
95
96/** Feature flags.
97 * @{
98 */
99/** Image uses a backing file to provide data for unallocated clusters. */
100#define QED_FEATURE_BACKING_FILE RT_BIT_64(0)
101/** Image needs checking before use. */
102#define QED_FEATURE_NEED_CHECK RT_BIT_64(1)
103/** Don't probe for format of the backing file, treat as raw image. */
104#define QED_FEATURE_BACKING_FILE_NO_PROBE RT_BIT_64(2)
105/** Mask of valid features. */
106#define QED_FEATURE_MASK (QED_FEATURE_BACKING_FILE | QED_FEATURE_NEED_CHECK | QED_FEATURE_BACKING_FILE_NO_PROBE)
107/** @} */
108
109/** Compatibility feature flags.
110 * @{
111 */
112/** Mask of valid compatibility features. */
113#define QED_COMPAT_FEATURE_MASK (0)
114/** @} */
115
116/** Autoreset feature flags.
117 * @{
118 */
119/** Mask of valid autoreset features. */
120#define QED_AUTORESET_FEATURE_MASK (0)
121/** @} */
122
123
124/*********************************************************************************************************************************
125* Constants And Macros, Structures and Typedefs *
126*********************************************************************************************************************************/
127
128/**
129 * QED L2 cache entry.
130 */
131typedef struct QEDL2CACHEENTRY
132{
133 /** List node for the search list. */
134 RTLISTNODE NodeSearch;
135 /** List node for the LRU list. */
136 RTLISTNODE NodeLru;
137 /** Reference counter. */
138 uint32_t cRefs;
139 /** The offset of the L2 table, used as search key. */
140 uint64_t offL2Tbl;
141 /** Pointer to the cached L2 table. */
142 uint64_t *paL2Tbl;
143} QEDL2CACHEENTRY, *PQEDL2CACHEENTRY;
144
145/** Maximum amount of memory the cache is allowed to use. */
146#define QED_L2_CACHE_MEMORY_MAX (2*_1M)
147
148/**
149 * QED image data structure.
150 */
151typedef struct QEDIMAGE
152{
153 /** Image name. */
154 const char *pszFilename;
155 /** Storage handle. */
156 PVDIOSTORAGE pStorage;
157
158 /** Pointer to the per-disk VD interface list. */
159 PVDINTERFACE pVDIfsDisk;
160 /** Pointer to the per-image VD interface list. */
161 PVDINTERFACE pVDIfsImage;
162 /** Error interface. */
163 PVDINTERFACEERROR pIfError;
164 /** I/O interface. */
165 PVDINTERFACEIOINT pIfIo;
166
167 /** Open flags passed by VBoxHD layer. */
168 unsigned uOpenFlags;
169 /** Image flags defined during creation or determined during open. */
170 unsigned uImageFlags;
171 /** Total size of the image. */
172 uint64_t cbSize;
173 /** Physical geometry of this image. */
174 VDGEOMETRY PCHSGeometry;
175 /** Logical geometry of this image. */
176 VDGEOMETRY LCHSGeometry;
177
178 /** Filename of the backing file if any. */
179 char *pszBackingFilename;
180 /** Offset of the filename in the image. */
181 uint32_t offBackingFilename;
182 /** Size of the backing filename excluding \0. */
183 uint32_t cbBackingFilename;
184
185 /** Size of the image, multiple of clusters. */
186 uint64_t cbImage;
187 /** Cluster size in bytes. */
188 uint32_t cbCluster;
189 /** Number of entries in the L1 and L2 table. */
190 uint32_t cTableEntries;
191 /** Size of an L1 or L2 table rounded to the next cluster size. */
192 uint32_t cbTable;
193 /** Pointer to the L1 table. */
194 uint64_t *paL1Table;
195 /** Offset of the L1 table. */
196 uint64_t offL1Table;
197
198 /** Offset mask for a cluster. */
199 uint64_t fOffsetMask;
200 /** L1 table mask to get the L1 index. */
201 uint64_t fL1Mask;
202 /** Number of bits to shift to get the L1 index. */
203 uint32_t cL1Shift;
204 /** L2 table mask to get the L2 index. */
205 uint64_t fL2Mask;
206 /** Number of bits to shift to get the L2 index. */
207 uint32_t cL2Shift;
208
209 /** Memory occupied by the L2 table cache. */
210 size_t cbL2Cache;
211 /** The sorted L2 entry list used for searching. */
212 RTLISTNODE ListSearch;
213 /** The LRU L2 entry list used for eviction. */
214 RTLISTNODE ListLru;
215
216} QEDIMAGE, *PQEDIMAGE;
217
218/**
219 * State of the async cluster allocation.
220 */
221typedef enum QEDCLUSTERASYNCALLOCSTATE
222{
223 /** Invalid. */
224 QEDCLUSTERASYNCALLOCSTATE_INVALID = 0,
225 /** L2 table allocation. */
226 QEDCLUSTERASYNCALLOCSTATE_L2_ALLOC,
227 /** Link L2 table into L1. */
228 QEDCLUSTERASYNCALLOCSTATE_L2_LINK,
229 /** Allocate user data cluster. */
230 QEDCLUSTERASYNCALLOCSTATE_USER_ALLOC,
231 /** Link user data cluster. */
232 QEDCLUSTERASYNCALLOCSTATE_USER_LINK,
233 /** 32bit blowup. */
234 QEDCLUSTERASYNCALLOCSTATE_32BIT_HACK = 0x7fffffff
235} QEDCLUSTERASYNCALLOCSTATE, *PQEDCLUSTERASYNCALLOCSTATE;
236
237/**
238 * Data needed to track async cluster allocation.
239 */
240typedef struct QEDCLUSTERASYNCALLOC
241{
242 /** The state of the cluster allocation. */
243 QEDCLUSTERASYNCALLOCSTATE enmAllocState;
244 /** Old image size to rollback in case of an error. */
245 uint64_t cbImageOld;
246 /** L1 index to link if any. */
247 uint32_t idxL1;
248 /** L2 index to link, required in any case. */
249 uint32_t idxL2;
250 /** Start offset of the allocated cluster. */
251 uint64_t offClusterNew;
252 /** L2 cache entry if a L2 table is allocated. */
253 PQEDL2CACHEENTRY pL2Entry;
254 /** Number of bytes to write. */
255 size_t cbToWrite;
256} QEDCLUSTERASYNCALLOC, *PQEDCLUSTERASYNCALLOC;
257
258
259/*********************************************************************************************************************************
260* Static Variables *
261*********************************************************************************************************************************/
262
263/** NULL-terminated array of supported file extensions. */
264static const VDFILEEXTENSION s_aQedFileExtensions[] =
265{
266 {"qed", VDTYPE_HDD},
267 {NULL, VDTYPE_INVALID}
268};
269
270
271/*********************************************************************************************************************************
272* Internal Functions *
273*********************************************************************************************************************************/
274
275/**
276 * Converts the image header to the host endianess and performs basic checks.
277 *
278 * @returns Whether the given header is valid or not.
279 * @param pHeader Pointer to the header to convert.
280 */
281static bool qedHdrConvertToHostEndianess(PQedHeader pHeader)
282{
283 pHeader->u32Magic = RT_LE2H_U32(pHeader->u32Magic);
284 pHeader->u32ClusterSize = RT_LE2H_U32(pHeader->u32ClusterSize);
285 pHeader->u32TableSize = RT_LE2H_U32(pHeader->u32TableSize);
286 pHeader->u32HeaderSize = RT_LE2H_U32(pHeader->u32HeaderSize);
287 pHeader->u64FeatureFlags = RT_LE2H_U64(pHeader->u64FeatureFlags);
288 pHeader->u64CompatFeatureFlags = RT_LE2H_U64(pHeader->u64CompatFeatureFlags);
289 pHeader->u64AutoresetFeatureFlags = RT_LE2H_U64(pHeader->u64AutoresetFeatureFlags);
290 pHeader->u64OffL1Table = RT_LE2H_U64(pHeader->u64OffL1Table);
291 pHeader->u64Size = RT_LE2H_U64(pHeader->u64Size);
292 pHeader->u32OffBackingFilename = RT_LE2H_U32(pHeader->u32OffBackingFilename);
293 pHeader->u32BackingFilenameSize = RT_LE2H_U32(pHeader->u32BackingFilenameSize);
294
295 if (RT_UNLIKELY(pHeader->u32Magic != QED_MAGIC))
296 return false;
297 if (RT_UNLIKELY( pHeader->u32ClusterSize < QED_CLUSTER_SIZE_MIN
298 || pHeader->u32ClusterSize > QED_CLUSTER_SIZE_MAX))
299 return false;
300 if (RT_UNLIKELY( pHeader->u32TableSize < QED_TABLE_SIZE_MIN
301 || pHeader->u32TableSize > QED_TABLE_SIZE_MAX))
302 return false;
303 if (RT_UNLIKELY(pHeader->u64Size % 512 != 0))
304 return false;
305
306 return true;
307}
308
309/**
310 * Creates a QED header from the given image state.
311 *
312 * @returns nothing.
313 * @param pImage Image instance data.
314 * @param pHeader Pointer to the header to convert.
315 */
316static void qedHdrConvertFromHostEndianess(PQEDIMAGE pImage, PQedHeader pHeader)
317{
318 pHeader->u32Magic = RT_H2LE_U32(QED_MAGIC);
319 pHeader->u32ClusterSize = RT_H2LE_U32(pImage->cbCluster);
320 pHeader->u32TableSize = RT_H2LE_U32(pImage->cbTable / pImage->cbCluster);
321 pHeader->u32HeaderSize = RT_H2LE_U32(1);
322 pHeader->u64FeatureFlags = RT_H2LE_U64(pImage->pszBackingFilename ? QED_FEATURE_BACKING_FILE : UINT64_C(0));
323 pHeader->u64CompatFeatureFlags = RT_H2LE_U64(UINT64_C(0));
324 pHeader->u64AutoresetFeatureFlags = RT_H2LE_U64(UINT64_C(0));
325 pHeader->u64OffL1Table = RT_H2LE_U64(pImage->offL1Table);
326 pHeader->u64Size = RT_H2LE_U64(pImage->cbSize);
327 pHeader->u32OffBackingFilename = RT_H2LE_U32(pImage->offBackingFilename);
328 pHeader->u32BackingFilenameSize = RT_H2LE_U32(pImage->cbBackingFilename);
329}
330
331/**
332 * Convert table entries from little endian to host endianess.
333 *
334 * @returns nothing.
335 * @param paTbl Pointer to the table.
336 * @param cEntries Number of entries in the table.
337 */
338static void qedTableConvertToHostEndianess(uint64_t *paTbl, uint32_t cEntries)
339{
340 while(cEntries-- > 0)
341 {
342 *paTbl = RT_LE2H_U64(*paTbl);
343 paTbl++;
344 }
345}
346
347#if defined(RT_BIG_ENDIAN)
348/**
349 * Convert table entries from host to little endian format.
350 *
351 * @returns nothing.
352 * @param paTblImg Pointer to the table which will store the little endian table.
353 * @param paTbl The source table to convert.
354 * @param cEntries Number of entries in the table.
355 */
356static void qedTableConvertFromHostEndianess(uint64_t *paTblImg, uint64_t *paTbl,
357 uint32_t cEntries)
358{
359 while(cEntries-- > 0)
360 {
361 *paTblImg = RT_H2LE_U64(*paTbl);
362 paTbl++;
363 paTblImg++;
364 }
365}
366#endif
367
368/**
369 * Creates the L2 table cache.
370 *
371 * @returns VBox status code.
372 * @param pImage The image instance data.
373 */
374static int qedL2TblCacheCreate(PQEDIMAGE pImage)
375{
376 pImage->cbL2Cache = 0;
377 RTListInit(&pImage->ListSearch);
378 RTListInit(&pImage->ListLru);
379
380 return VINF_SUCCESS;
381}
382
383/**
384 * Destroys the L2 table cache.
385 *
386 * @returns nothing.
387 * @param pImage The image instance data.
388 */
389static void qedL2TblCacheDestroy(PQEDIMAGE pImage)
390{
391 PQEDL2CACHEENTRY pL2Entry = NULL;
392 PQEDL2CACHEENTRY pL2Next = NULL;
393
394 RTListForEachSafe(&pImage->ListSearch, pL2Entry, pL2Next, QEDL2CACHEENTRY, NodeSearch)
395 {
396 Assert(!pL2Entry->cRefs);
397
398 RTListNodeRemove(&pL2Entry->NodeSearch);
399 RTMemPageFree(pL2Entry->paL2Tbl, pImage->cbTable);
400 RTMemFree(pL2Entry);
401 }
402
403 pImage->cbL2Cache = 0;
404 RTListInit(&pImage->ListSearch);
405 RTListInit(&pImage->ListLru);
406}
407
408/**
409 * Returns the L2 table matching the given offset or NULL if none could be found.
410 *
411 * @returns Pointer to the L2 table cache entry or NULL.
412 * @param pImage The image instance data.
413 * @param offL2Tbl Offset of the L2 table to search for.
414 */
415static PQEDL2CACHEENTRY qedL2TblCacheRetain(PQEDIMAGE pImage, uint64_t offL2Tbl)
416{
417 PQEDL2CACHEENTRY pL2Entry = NULL;
418
419 RTListForEach(&pImage->ListSearch, pL2Entry, QEDL2CACHEENTRY, NodeSearch)
420 {
421 if (pL2Entry->offL2Tbl == offL2Tbl)
422 break;
423 }
424
425 if (!RTListNodeIsDummy(&pImage->ListSearch, pL2Entry, QEDL2CACHEENTRY, NodeSearch))
426 {
427 /* Update LRU list. */
428 RTListNodeRemove(&pL2Entry->NodeLru);
429 RTListPrepend(&pImage->ListLru, &pL2Entry->NodeLru);
430 pL2Entry->cRefs++;
431 return pL2Entry;
432 }
433 else
434 return NULL;
435}
436
437/**
438 * Releases a L2 table cache entry.
439 *
440 * @returns nothing.
441 * @param pL2Entry The L2 cache entry.
442 */
443static void qedL2TblCacheEntryRelease(PQEDL2CACHEENTRY pL2Entry)
444{
445 Assert(pL2Entry->cRefs > 0);
446 pL2Entry->cRefs--;
447}
448
449/**
450 * Allocates a new L2 table from the cache evicting old entries if required.
451 *
452 * @returns Pointer to the L2 cache entry or NULL.
453 * @param pImage The image instance data.
454 */
455static PQEDL2CACHEENTRY qedL2TblCacheEntryAlloc(PQEDIMAGE pImage)
456{
457 PQEDL2CACHEENTRY pL2Entry = NULL;
458
459 if (pImage->cbL2Cache + pImage->cbTable <= QED_L2_CACHE_MEMORY_MAX)
460 {
461 /* Add a new entry. */
462 pL2Entry = (PQEDL2CACHEENTRY)RTMemAllocZ(sizeof(QEDL2CACHEENTRY));
463 if (pL2Entry)
464 {
465 pL2Entry->paL2Tbl = (uint64_t *)RTMemPageAllocZ(pImage->cbTable);
466 if (RT_UNLIKELY(!pL2Entry->paL2Tbl))
467 {
468 RTMemFree(pL2Entry);
469 pL2Entry = NULL;
470 }
471 else
472 {
473 pL2Entry->cRefs = 1;
474 pImage->cbL2Cache += pImage->cbTable;
475 }
476 }
477 }
478 else
479 {
480 /* Evict the last not in use entry and use it */
481 Assert(!RTListIsEmpty(&pImage->ListLru));
482
483 RTListForEachReverse(&pImage->ListLru, pL2Entry, QEDL2CACHEENTRY, NodeLru)
484 {
485 if (!pL2Entry->cRefs)
486 break;
487 }
488
489 if (!RTListNodeIsDummy(&pImage->ListSearch, pL2Entry, QEDL2CACHEENTRY, NodeSearch))
490 {
491 RTListNodeRemove(&pL2Entry->NodeSearch);
492 RTListNodeRemove(&pL2Entry->NodeLru);
493 pL2Entry->offL2Tbl = 0;
494 pL2Entry->cRefs = 1;
495 }
496 else
497 pL2Entry = NULL;
498 }
499
500 return pL2Entry;
501}
502
503/**
504 * Frees a L2 table cache entry.
505 *
506 * @returns nothing.
507 * @param pImage The image instance data.
508 * @param pL2Entry The L2 cache entry to free.
509 */
510static void qedL2TblCacheEntryFree(PQEDIMAGE pImage, PQEDL2CACHEENTRY pL2Entry)
511{
512 Assert(!pL2Entry->cRefs);
513 RTMemPageFree(pL2Entry->paL2Tbl, pImage->cbTable);
514 RTMemFree(pL2Entry);
515
516 pImage->cbL2Cache -= pImage->cbTable;
517}
518
519/**
520 * Inserts an entry in the L2 table cache.
521 *
522 * @returns nothing.
523 * @param pImage The image instance data.
524 * @param pL2Entry The L2 cache entry to insert.
525 */
526static void qedL2TblCacheEntryInsert(PQEDIMAGE pImage, PQEDL2CACHEENTRY pL2Entry)
527{
528 PQEDL2CACHEENTRY pIt = NULL;
529
530 Assert(pL2Entry->offL2Tbl > 0);
531
532 /* Insert at the top of the LRU list. */
533 RTListPrepend(&pImage->ListLru, &pL2Entry->NodeLru);
534
535 if (RTListIsEmpty(&pImage->ListSearch))
536 {
537 RTListAppend(&pImage->ListSearch, &pL2Entry->NodeSearch);
538 }
539 else
540 {
541 /* Insert into search list. */
542 pIt = RTListGetFirst(&pImage->ListSearch, QEDL2CACHEENTRY, NodeSearch);
543 if (pIt->offL2Tbl > pL2Entry->offL2Tbl)
544 RTListPrepend(&pImage->ListSearch, &pL2Entry->NodeSearch);
545 else
546 {
547 bool fInserted = false;
548
549 RTListForEach(&pImage->ListSearch, pIt, QEDL2CACHEENTRY, NodeSearch)
550 {
551 Assert(pIt->offL2Tbl != pL2Entry->offL2Tbl);
552 if (pIt->offL2Tbl < pL2Entry->offL2Tbl)
553 {
554 RTListNodeInsertAfter(&pIt->NodeSearch, &pL2Entry->NodeSearch);
555 fInserted = true;
556 break;
557 }
558 }
559 Assert(fInserted);
560 }
561 }
562}
563
564#if 0 /* unused */
565/**
566 * Fetches the L2 from the given offset trying the LRU cache first and
567 * reading it from the image after a cache miss.
568 *
569 * @returns VBox status code.
570 * @param pImage Image instance data.
571 * @param offL2Tbl The offset of the L2 table in the image.
572 * @param ppL2Entry Where to store the L2 table on success.
573 */
574static int qedL2TblCacheFetch(PQEDIMAGE pImage, uint64_t offL2Tbl, PQEDL2CACHEENTRY *ppL2Entry)
575{
576 int rc = VINF_SUCCESS;
577
578 LogFlowFunc(("pImage=%#p offL2Tbl=%llu ppL2Entry=%#p\n", pImage, offL2Tbl, ppL2Entry));
579
580 /* Try to fetch the L2 table from the cache first. */
581 PQEDL2CACHEENTRY pL2Entry = qedL2TblCacheRetain(pImage, offL2Tbl);
582 if (!pL2Entry)
583 {
584 LogFlowFunc(("Reading L2 table from image\n"));
585 pL2Entry = qedL2TblCacheEntryAlloc(pImage);
586
587 if (pL2Entry)
588 {
589 /* Read from the image. */
590 pL2Entry->offL2Tbl = offL2Tbl;
591 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pImage->pStorage, offL2Tbl,
592 pL2Entry->paL2Tbl, pImage->cbTable);
593 if (RT_SUCCESS(rc))
594 {
595#if defined(RT_BIG_ENDIAN)
596 qedTableConvertToHostEndianess(pL2Entry->paL2Tbl, pImage->cTableEntries);
597#endif
598 qedL2TblCacheEntryInsert(pImage, pL2Entry);
599 }
600 else
601 {
602 qedL2TblCacheEntryRelease(pL2Entry);
603 qedL2TblCacheEntryFree(pImage, pL2Entry);
604 }
605 }
606 else
607 rc = VERR_NO_MEMORY;
608 }
609
610 if (RT_SUCCESS(rc))
611 *ppL2Entry = pL2Entry;
612
613 LogFlowFunc(("returns rc=%Rrc\n", rc));
614 return rc;
615}
616#endif
617
618/**
619 * Fetches the L2 from the given offset trying the LRU cache first and
620 * reading it from the image after a cache miss - version for async I/O.
621 *
622 * @returns VBox status code.
623 * @param pImage Image instance data.
624 * @param pIoCtx The I/O context.
625 * @param offL2Tbl The offset of the L2 table in the image.
626 * @param ppL2Entry Where to store the L2 table on success.
627 */
628static int qedL2TblCacheFetchAsync(PQEDIMAGE pImage, PVDIOCTX pIoCtx,
629 uint64_t offL2Tbl, PQEDL2CACHEENTRY *ppL2Entry)
630{
631 int rc = VINF_SUCCESS;
632
633 /* Try to fetch the L2 table from the cache first. */
634 PQEDL2CACHEENTRY pL2Entry = qedL2TblCacheRetain(pImage, offL2Tbl);
635 if (!pL2Entry)
636 {
637 pL2Entry = qedL2TblCacheEntryAlloc(pImage);
638
639 if (pL2Entry)
640 {
641 /* Read from the image. */
642 PVDMETAXFER pMetaXfer;
643
644 pL2Entry->offL2Tbl = offL2Tbl;
645 rc = vdIfIoIntFileReadMeta(pImage->pIfIo, pImage->pStorage,
646 offL2Tbl, pL2Entry->paL2Tbl,
647 pImage->cbTable, pIoCtx,
648 &pMetaXfer, NULL, NULL);
649 if (RT_SUCCESS(rc))
650 {
651 vdIfIoIntMetaXferRelease(pImage->pIfIo, pMetaXfer);
652#if defined(RT_BIG_ENDIAN)
653 qedTableConvertToHostEndianess(pL2Entry->paL2Tbl, pImage->cTableEntries);
654#endif
655 qedL2TblCacheEntryInsert(pImage, pL2Entry);
656 }
657 else
658 {
659 qedL2TblCacheEntryRelease(pL2Entry);
660 qedL2TblCacheEntryFree(pImage, pL2Entry);
661 }
662 }
663 else
664 rc = VERR_NO_MEMORY;
665 }
666
667 if (RT_SUCCESS(rc))
668 *ppL2Entry = pL2Entry;
669
670 return rc;
671}
672
673/**
674 * Return power of 2 or 0 if num error.
675 *
676 * @returns The power of 2 or 0 if the given number is not a power of 2.
677 * @param u32 The number.
678 */
679static uint32_t qedGetPowerOfTwo(uint32_t u32)
680{
681 if (u32 == 0)
682 return 0;
683 uint32_t uPower2 = 0;
684 while ((u32 & 1) == 0)
685 {
686 u32 >>= 1;
687 uPower2++;
688 }
689 return u32 == 1 ? uPower2 : 0;
690}
691
692/**
693 * Sets the L1, L2 and offset bitmasks and L1 and L2 bit shift members.
694 *
695 * @returns nothing.
696 * @param pImage The image instance data.
697 */
698static void qedTableMasksInit(PQEDIMAGE pImage)
699{
700 uint32_t cClusterBits, cTableBits;
701
702 cClusterBits = qedGetPowerOfTwo(pImage->cbCluster);
703 cTableBits = qedGetPowerOfTwo(pImage->cTableEntries);
704
705 Assert(cClusterBits + 2 * cTableBits <= 64);
706
707 pImage->fOffsetMask = ((uint64_t)pImage->cbCluster - 1);
708 pImage->fL2Mask = ((uint64_t)pImage->cTableEntries - 1) << cClusterBits;
709 pImage->cL2Shift = cClusterBits;
710 pImage->fL1Mask = ((uint64_t)pImage->cTableEntries - 1) << (cClusterBits + cTableBits);
711 pImage->cL1Shift = cClusterBits + cTableBits;
712}
713
714/**
715 * Converts a given logical offset into the
716 *
717 * @returns nothing.
718 * @param pImage The image instance data.
719 * @param off The logical offset to convert.
720 * @param pidxL1 Where to store the index in the L1 table on success.
721 * @param pidxL2 Where to store the index in the L2 table on success.
722 * @param poffCluster Where to store the offset in the cluster on success.
723 */
724DECLINLINE(void) qedConvertLogicalOffset(PQEDIMAGE pImage, uint64_t off, uint32_t *pidxL1,
725 uint32_t *pidxL2, uint32_t *poffCluster)
726{
727 AssertPtr(pidxL1);
728 AssertPtr(pidxL2);
729 AssertPtr(poffCluster);
730
731 *poffCluster = off & pImage->fOffsetMask;
732 *pidxL1 = (off & pImage->fL1Mask) >> pImage->cL1Shift;
733 *pidxL2 = (off & pImage->fL2Mask) >> pImage->cL2Shift;
734}
735
736/**
737 * Converts Cluster size to a byte size.
738 *
739 * @returns Number of bytes derived from the given number of clusters.
740 * @param pImage The image instance data.
741 * @param cClusters The clusters to convert.
742 */
743DECLINLINE(uint64_t) qedCluster2Byte(PQEDIMAGE pImage, uint64_t cClusters)
744{
745 return cClusters * pImage->cbCluster;
746}
747
748/**
749 * Converts number of bytes to cluster size rounding to the next cluster.
750 *
751 * @returns Number of bytes derived from the given number of clusters.
752 * @param pImage The image instance data.
753 * @param cb Number of bytes to convert.
754 */
755DECLINLINE(uint64_t) qedByte2Cluster(PQEDIMAGE pImage, uint64_t cb)
756{
757 return cb / pImage->cbCluster + (cb % pImage->cbCluster ? 1 : 0);
758}
759
760/**
761 * Allocates a new cluster in the image.
762 *
763 * @returns The start offset of the new cluster in the image.
764 * @param pImage The image instance data.
765 * @param cCLusters Number of clusters to allocate.
766 */
767DECLINLINE(uint64_t) qedClusterAllocate(PQEDIMAGE pImage, uint32_t cClusters)
768{
769 uint64_t offCluster;
770
771 offCluster = pImage->cbImage;
772 pImage->cbImage += cClusters*pImage->cbCluster;
773
774 return offCluster;
775}
776
777/**
778 * Returns the real image offset for a given cluster or an error if the cluster is not
779 * yet allocated.
780 *
781 * @returns VBox status code.
782 * VERR_VD_BLOCK_FREE if the cluster is not yet allocated.
783 * @param pImage The image instance data.
784 * @param pIoCtx The I/O context.
785 * @param idxL1 The L1 index.
786 * @param idxL2 The L2 index.
787 * @param offCluster Offset inside the cluster.
788 * @param poffImage Where to store the image offset on success;
789 */
790static int qedConvertToImageOffset(PQEDIMAGE pImage, PVDIOCTX pIoCtx,
791 uint32_t idxL1, uint32_t idxL2,
792 uint32_t offCluster, uint64_t *poffImage)
793{
794 int rc = VERR_VD_BLOCK_FREE;
795
796 AssertReturn(idxL1 < pImage->cTableEntries, VERR_INVALID_PARAMETER);
797 AssertReturn(idxL2 < pImage->cTableEntries, VERR_INVALID_PARAMETER);
798
799 if (pImage->paL1Table[idxL1])
800 {
801 PQEDL2CACHEENTRY pL2Entry;
802
803 rc = qedL2TblCacheFetchAsync(pImage, pIoCtx, pImage->paL1Table[idxL1],
804 &pL2Entry);
805 if (RT_SUCCESS(rc))
806 {
807 /* Get real file offset. */
808 if (pL2Entry->paL2Tbl[idxL2])
809 *poffImage = pL2Entry->paL2Tbl[idxL2] + offCluster;
810 else
811 rc = VERR_VD_BLOCK_FREE;
812
813 qedL2TblCacheEntryRelease(pL2Entry);
814 }
815 }
816
817 return rc;
818}
819
820
821/**
822 * Internal. Flush image data to disk.
823 */
824static int qedFlushImage(PQEDIMAGE pImage)
825{
826 int rc = VINF_SUCCESS;
827
828 if ( pImage->pStorage
829 && !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
830 {
831 QedHeader Header;
832
833 Assert(!(pImage->cbTable % pImage->cbCluster));
834#if defined(RT_BIG_ENDIAN)
835 uint64_t *paL1TblImg = (uint64_t *)RTMemAllocZ(pImage->cbTable);
836 if (paL1TblImg)
837 {
838 qedTableConvertFromHostEndianess(paL1TblImg, pImage->paL1Table,
839 pImage->cTableEntries);
840 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pImage->pStorage,
841 pImage->offL1Table, paL1TblImg,
842 pImage->cbTable);
843 RTMemFree(paL1TblImg);
844 }
845 else
846 rc = VERR_NO_MEMORY;
847#else
848 /* Write L1 table directly. */
849 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pImage->pStorage, pImage->offL1Table,
850 pImage->paL1Table, pImage->cbTable);
851#endif
852 if (RT_SUCCESS(rc))
853 {
854 /* Write header. */
855 qedHdrConvertFromHostEndianess(pImage, &Header);
856 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pImage->pStorage, 0, &Header,
857 sizeof(Header));
858 if (RT_SUCCESS(rc))
859 rc = vdIfIoIntFileFlushSync(pImage->pIfIo, pImage->pStorage);
860 }
861 }
862
863 return rc;
864}
865
866/**
867 * Flush image data to disk - version for async I/O.
868 *
869 * @returns VBox status code.
870 * @param pImage The image instance data.
871 * @param pIoCtx The I/o context
872 */
873static int qedFlushImageAsync(PQEDIMAGE pImage, PVDIOCTX pIoCtx)
874{
875 int rc = VINF_SUCCESS;
876
877 if ( pImage->pStorage
878 && !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
879 {
880 QedHeader Header;
881
882 Assert(!(pImage->cbTable % pImage->cbCluster));
883#if defined(RT_BIG_ENDIAN)
884 uint64_t *paL1TblImg = (uint64_t *)RTMemAllocZ(pImage->cbTable);
885 if (paL1TblImg)
886 {
887 qedTableConvertFromHostEndianess(paL1TblImg, pImage->paL1Table,
888 pImage->cTableEntries);
889 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pImage->pStorage,
890 pImage->offL1Table, paL1TblImg,
891 pImage->cbTable, pIoCtx, NULL, NULL);
892 RTMemFree(paL1TblImg);
893 }
894 else
895 rc = VERR_NO_MEMORY;
896#else
897 /* Write L1 table directly. */
898 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pImage->pStorage,
899 pImage->offL1Table, pImage->paL1Table,
900 pImage->cbTable, pIoCtx, NULL, NULL);
901#endif
902 if (RT_SUCCESS(rc) || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
903 {
904 /* Write header. */
905 qedHdrConvertFromHostEndianess(pImage, &Header);
906 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pImage->pStorage,
907 0, &Header, sizeof(Header),
908 pIoCtx, NULL, NULL);
909 if (RT_SUCCESS(rc) || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
910 rc = vdIfIoIntFileFlush(pImage->pIfIo, pImage->pStorage,
911 pIoCtx, NULL, NULL);
912 }
913 }
914
915 return rc;
916}
917
918/**
919 * Checks whether the given cluster offset is valid.
920 *
921 * @returns Whether the given cluster offset is valid.
922 * @param offCluster The table offset to check.
923 * @param cbFile The real file size of the image.
924 * @param cbCluster The cluster size in bytes.
925 */
926DECLINLINE(bool) qedIsClusterOffsetValid(uint64_t offCluster, uint64_t cbFile, size_t cbCluster)
927{
928 return (offCluster <= cbFile - cbCluster)
929 && !(offCluster & (cbCluster - 1));
930}
931
932/**
933 * Checks whether the given table offset is valid.
934 *
935 * @returns Whether the given table offset is valid.
936 * @param offTbl The table offset to check.
937 * @param cbFile The real file size of the image.
938 * @param cbTable The table size in bytes.
939 * @param cbCluster The cluster size in bytes.
940 */
941DECLINLINE(bool) qedIsTblOffsetValid(uint64_t offTbl, uint64_t cbFile, size_t cbTable, size_t cbCluster)
942{
943 return (offTbl <= cbFile - cbTable)
944 && !(offTbl & (cbCluster - 1));
945}
946
947/**
948 * Sets the specified range in the cluster bitmap checking whether any of the clusters is already
949 * used before.
950 *
951 * @returns Whether the range was clear and is set now.
952 * @param pvClusterBitmap The cluster bitmap to use.
953 * @param offClusterStart The first cluster to check and set.
954 * @param offClusterEnd The first cluster to not check and set anymore.
955 */
956static bool qedClusterBitmapCheckAndSet(void *pvClusterBitmap, uint32_t offClusterStart, uint32_t offClusterEnd)
957{
958 for (uint32_t offCluster = offClusterStart; offCluster < offClusterEnd; offCluster++)
959 if (ASMBitTest(pvClusterBitmap, offCluster))
960 return false;
961
962 ASMBitSetRange(pvClusterBitmap, offClusterStart, offClusterEnd);
963 return true;
964}
965
966/**
967 * Checks the given image for consistency, usually called when the
968 * QED_FEATURE_NEED_CHECK bit is set.
969 *
970 * @returns VBox status code.
971 * @retval VINF_SUCCESS when the image can be accessed.
972 * @param pImage The image instance data.
973 * @param pHeader The header to use for checking.
974 *
975 * @note It is not required that the image state is fully initialized Only
976 * The I/O interface and storage handle need to be valid.
977 * @note The header must be converted to the host CPU endian format already
978 * and should be validated already.
979 */
980static int qedCheckImage(PQEDIMAGE pImage, PQedHeader pHeader)
981{
982 uint64_t cbFile;
983 uint32_t cbTable;
984 uint32_t cTableEntries;
985 uint64_t *paL1Tbl = NULL;
986 uint64_t *paL2Tbl = NULL;
987 void *pvClusterBitmap = NULL;
988 uint32_t offClusterStart;
989 int rc = VINF_SUCCESS;
990
991 pImage->cbCluster = pHeader->u32ClusterSize;
992 cbTable = pHeader->u32TableSize * pHeader->u32ClusterSize;
993 cTableEntries = cbTable / sizeof(uint64_t);
994
995 do
996 {
997 rc = vdIfIoIntFileGetSize(pImage->pIfIo, pImage->pStorage, &cbFile);
998 if (RT_FAILURE(rc))
999 {
1000 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
1001 N_("Qed: Querying the file size of image '%s' failed"),
1002 pImage->pszFilename);
1003 break;
1004 }
1005
1006 /* Allocate L1 table. */
1007 paL1Tbl = (uint64_t *)RTMemAllocZ(cbTable);
1008 if (!paL1Tbl)
1009 {
1010 rc = vdIfError(pImage->pIfError, VERR_NO_MEMORY, RT_SRC_POS,
1011 N_("Qed: Allocating memory for the L1 table for image '%s' failed"),
1012 pImage->pszFilename);
1013 break;
1014 }
1015
1016 paL2Tbl = (uint64_t *)RTMemAllocZ(cbTable);
1017 if (!paL2Tbl)
1018 {
1019 rc = vdIfError(pImage->pIfError, VERR_NO_MEMORY, RT_SRC_POS,
1020 N_("Qed: Allocating memory for the L2 table for image '%s' failed"),
1021 pImage->pszFilename);
1022 break;
1023 }
1024
1025 pvClusterBitmap = RTMemAllocZ(cbFile / pHeader->u32ClusterSize / 8);
1026 if (!pvClusterBitmap)
1027 {
1028 rc = vdIfError(pImage->pIfError, VERR_NO_MEMORY, RT_SRC_POS,
1029 N_("Qed: Allocating memory for the cluster bitmap for image '%s' failed"),
1030 pImage->pszFilename);
1031 break;
1032 }
1033
1034 /* Validate L1 table offset. */
1035 if (!qedIsTblOffsetValid(pHeader->u64OffL1Table, cbFile, cbTable, pHeader->u32ClusterSize))
1036 {
1037 rc = vdIfError(pImage->pIfError, VERR_VD_GEN_INVALID_HEADER, RT_SRC_POS,
1038 N_("Qed: L1 table offset of image '%s' is corrupt (%llu)"),
1039 pImage->pszFilename, pHeader->u64OffL1Table);
1040 break;
1041 }
1042
1043 /* Read L1 table. */
1044 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pImage->pStorage,
1045 pHeader->u64OffL1Table, paL1Tbl, cbTable);
1046 if (RT_FAILURE(rc))
1047 {
1048 rc = vdIfError(pImage->pIfError, VERR_VD_GEN_INVALID_HEADER, RT_SRC_POS,
1049 N_("Qed: Reading the L1 table from image '%s' failed"),
1050 pImage->pszFilename);
1051 break;
1052 }
1053
1054 /* Mark the L1 table in cluster bitmap. */
1055 ASMBitSet(pvClusterBitmap, 0); /* Header is always in cluster 0. */
1056 offClusterStart = qedByte2Cluster(pImage, pHeader->u64OffL1Table);
1057 bool fSet = qedClusterBitmapCheckAndSet(pvClusterBitmap, offClusterStart, offClusterStart + pHeader->u32TableSize);
1058 Assert(fSet);
1059
1060 /* Scan the L1 and L2 tables for invalid entries. */
1061 qedTableConvertToHostEndianess(paL1Tbl, cTableEntries);
1062
1063 for (unsigned iL1 = 0; iL1 < cTableEntries; iL1++)
1064 {
1065 if (!paL1Tbl[iL1])
1066 continue; /* Skip unallocated clusters. */
1067
1068 if (!qedIsTblOffsetValid(paL1Tbl[iL1], cbFile, cbTable, pHeader->u32ClusterSize))
1069 {
1070 rc = vdIfError(pImage->pIfError, VERR_VD_GEN_INVALID_HEADER, RT_SRC_POS,
1071 N_("Qed: Entry %d of the L1 table from image '%s' is invalid (%llu)"),
1072 iL1, pImage->pszFilename, paL1Tbl[iL1]);
1073 break;
1074 }
1075
1076 /* Now check that the clusters are not allocated already. */
1077 offClusterStart = qedByte2Cluster(pImage, paL1Tbl[iL1]);
1078 fSet = qedClusterBitmapCheckAndSet(pvClusterBitmap, offClusterStart, offClusterStart + pHeader->u32TableSize);
1079 if (!fSet)
1080 {
1081 rc = vdIfError(pImage->pIfError, VERR_VD_GEN_INVALID_HEADER, RT_SRC_POS,
1082 N_("Qed: Entry %d of the L1 table from image '%s' points to a already used cluster (%llu)"),
1083 iL1, pImage->pszFilename, paL1Tbl[iL1]);
1084 break;
1085 }
1086
1087 /* Read the linked L2 table and check it. */
1088 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pImage->pStorage,
1089 paL1Tbl[iL1], paL2Tbl, cbTable);
1090 if (RT_FAILURE(rc))
1091 {
1092 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
1093 N_("Qed: Reading the L2 table from image '%s' failed"),
1094 pImage->pszFilename);
1095 break;
1096 }
1097
1098 /* Check all L2 entries. */
1099 for (unsigned iL2 = 0; iL2 < cTableEntries; iL2++)
1100 {
1101 if (paL2Tbl[iL2])
1102 continue; /* Skip unallocated clusters. */
1103
1104 if (!qedIsClusterOffsetValid(paL2Tbl[iL2], cbFile, pHeader->u32ClusterSize))
1105 {
1106 rc = vdIfError(pImage->pIfError, VERR_VD_GEN_INVALID_HEADER, RT_SRC_POS,
1107 N_("Qed: Entry %d of the L2 table from image '%s' is invalid (%llu)"),
1108 iL2, pImage->pszFilename, paL2Tbl[iL2]);
1109 break;
1110 }
1111
1112 /* Now check that the clusters are not allocated already. */
1113 offClusterStart = qedByte2Cluster(pImage, paL2Tbl[iL2]);
1114 fSet = qedClusterBitmapCheckAndSet(pvClusterBitmap, offClusterStart, offClusterStart + 1);
1115 if (!fSet)
1116 {
1117 rc = vdIfError(pImage->pIfError, VERR_VD_GEN_INVALID_HEADER, RT_SRC_POS,
1118 N_("Qed: Entry %d of the L2 table from image '%s' points to a already used cluster (%llu)"),
1119 iL2, pImage->pszFilename, paL2Tbl[iL2]);
1120 break;
1121 }
1122 }
1123 }
1124 } while(0);
1125
1126 if (paL1Tbl)
1127 RTMemFree(paL1Tbl);
1128 if (paL2Tbl)
1129 RTMemFree(paL2Tbl);
1130 if (pvClusterBitmap)
1131 RTMemFree(pvClusterBitmap);
1132
1133 return rc;
1134}
1135
1136/**
1137 * Internal. Free all allocated space for representing an image except pImage,
1138 * and optionally delete the image from disk.
1139 */
1140static int qedFreeImage(PQEDIMAGE pImage, bool fDelete)
1141{
1142 int rc = VINF_SUCCESS;
1143
1144 /* Freeing a never allocated image (e.g. because the open failed) is
1145 * not signalled as an error. After all nothing bad happens. */
1146 if (pImage)
1147 {
1148 if (pImage->pStorage)
1149 {
1150 /* No point updating the file that is deleted anyway. */
1151 if (!fDelete)
1152 qedFlushImage(pImage);
1153
1154 rc = vdIfIoIntFileClose(pImage->pIfIo, pImage->pStorage);
1155 pImage->pStorage = NULL;
1156 }
1157
1158 if (pImage->paL1Table)
1159 RTMemFree(pImage->paL1Table);
1160
1161 if (pImage->pszBackingFilename)
1162 {
1163 RTMemFree(pImage->pszBackingFilename);
1164 pImage->pszBackingFilename = NULL;
1165 }
1166
1167 qedL2TblCacheDestroy(pImage);
1168
1169 if (fDelete && pImage->pszFilename)
1170 vdIfIoIntFileDelete(pImage->pIfIo, pImage->pszFilename);
1171 }
1172
1173 LogFlowFunc(("returns %Rrc\n", rc));
1174 return rc;
1175}
1176
1177/**
1178 * Internal: Open an image, constructing all necessary data structures.
1179 */
1180static int qedOpenImage(PQEDIMAGE pImage, unsigned uOpenFlags)
1181{
1182 int rc;
1183
1184 pImage->uOpenFlags = uOpenFlags;
1185
1186 pImage->pIfError = VDIfErrorGet(pImage->pVDIfsDisk);
1187 pImage->pIfIo = VDIfIoIntGet(pImage->pVDIfsImage);
1188 AssertPtrReturn(pImage->pIfIo, VERR_INVALID_PARAMETER);
1189
1190 /*
1191 * Create the L2 cache before opening the image so we can call qedFreeImage()
1192 * even if opening the image file fails.
1193 */
1194 rc = qedL2TblCacheCreate(pImage);
1195 if (RT_FAILURE(rc))
1196 {
1197 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
1198 N_("Qed: Creating the L2 table cache for image '%s' failed"),
1199 pImage->pszFilename);
1200
1201 goto out;
1202 }
1203
1204 /*
1205 * Open the image.
1206 */
1207 rc = vdIfIoIntFileOpen(pImage->pIfIo, pImage->pszFilename,
1208 VDOpenFlagsToFileOpenFlags(uOpenFlags,
1209 false /* fCreate */),
1210 &pImage->pStorage);
1211 if (RT_FAILURE(rc))
1212 {
1213 /* Do NOT signal an appropriate error here, as the VD layer has the
1214 * choice of retrying the open if it failed. */
1215 goto out;
1216 }
1217
1218 uint64_t cbFile;
1219 QedHeader Header;
1220 rc = vdIfIoIntFileGetSize(pImage->pIfIo, pImage->pStorage, &cbFile);
1221 if (RT_FAILURE(rc))
1222 goto out;
1223 if (cbFile > sizeof(Header))
1224 {
1225 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pImage->pStorage, 0, &Header, sizeof(Header));
1226 if ( RT_SUCCESS(rc)
1227 && qedHdrConvertToHostEndianess(&Header))
1228 {
1229 if ( !(Header.u64FeatureFlags & ~QED_FEATURE_MASK)
1230 && !(Header.u64FeatureFlags & QED_FEATURE_BACKING_FILE_NO_PROBE))
1231 {
1232 if (Header.u64FeatureFlags & QED_FEATURE_NEED_CHECK)
1233 {
1234 /* Image needs checking. */
1235 if (!(uOpenFlags & VD_OPEN_FLAGS_READONLY))
1236 rc = qedCheckImage(pImage, &Header);
1237 else
1238 rc = vdIfError(pImage->pIfError, VERR_NOT_SUPPORTED, RT_SRC_POS,
1239 N_("Qed: Image '%s' needs checking but is opened readonly"),
1240 pImage->pszFilename);
1241 }
1242
1243 if ( RT_SUCCESS(rc)
1244 && (Header.u64FeatureFlags & QED_FEATURE_BACKING_FILE))
1245 {
1246 /* Load backing filename from image. */
1247 pImage->pszBackingFilename = (char *)RTMemAllocZ(Header.u32BackingFilenameSize + 1); /* +1 for \0 terminator. */
1248 if (pImage->pszBackingFilename)
1249 {
1250 pImage->cbBackingFilename = Header.u32BackingFilenameSize;
1251 pImage->offBackingFilename = Header.u32OffBackingFilename;
1252 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pImage->pStorage,
1253 Header.u32OffBackingFilename, pImage->pszBackingFilename,
1254 Header.u32BackingFilenameSize);
1255 }
1256 else
1257 rc = VERR_NO_MEMORY;
1258 }
1259
1260 if (RT_SUCCESS(rc))
1261 {
1262 pImage->cbImage = cbFile;
1263 pImage->cbCluster = Header.u32ClusterSize;
1264 pImage->cbTable = Header.u32TableSize * pImage->cbCluster;
1265 pImage->cTableEntries = pImage->cbTable / sizeof(uint64_t);
1266 pImage->offL1Table = Header.u64OffL1Table;
1267 pImage->cbSize = Header.u64Size;
1268 qedTableMasksInit(pImage);
1269
1270 /* Allocate L1 table. */
1271 pImage->paL1Table = (uint64_t *)RTMemAllocZ(pImage->cbTable);
1272 if (pImage->paL1Table)
1273 {
1274 /* Read from the image. */
1275 rc = vdIfIoIntFileReadSync(pImage->pIfIo, pImage->pStorage,
1276 pImage->offL1Table, pImage->paL1Table,
1277 pImage->cbTable);
1278 if (RT_SUCCESS(rc))
1279 {
1280 qedTableConvertToHostEndianess(pImage->paL1Table, pImage->cTableEntries);
1281
1282 /* If the consistency check succeeded, clear the flag by flushing the image. */
1283 if (Header.u64FeatureFlags & QED_FEATURE_NEED_CHECK)
1284 rc = qedFlushImage(pImage);
1285 }
1286 else
1287 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
1288 N_("Qed: Reading the L1 table for image '%s' failed"),
1289 pImage->pszFilename);
1290 }
1291 else
1292 rc = vdIfError(pImage->pIfError, VERR_NO_MEMORY, RT_SRC_POS,
1293 N_("Qed: Out of memory allocating L1 table for image '%s'"),
1294 pImage->pszFilename);
1295 }
1296 }
1297 else
1298 rc = vdIfError(pImage->pIfError, VERR_NOT_SUPPORTED, RT_SRC_POS,
1299 N_("Qed: The image '%s' makes use of unsupported features"),
1300 pImage->pszFilename);
1301 }
1302 else if (RT_SUCCESS(rc))
1303 rc = VERR_VD_GEN_INVALID_HEADER;
1304 }
1305 else
1306 rc = VERR_VD_GEN_INVALID_HEADER;
1307
1308out:
1309 if (RT_FAILURE(rc))
1310 qedFreeImage(pImage, false);
1311 return rc;
1312}
1313
1314/**
1315 * Internal: Create a qed image.
1316 */
1317static int qedCreateImage(PQEDIMAGE pImage, uint64_t cbSize,
1318 unsigned uImageFlags, const char *pszComment,
1319 PCVDGEOMETRY pPCHSGeometry,
1320 PCVDGEOMETRY pLCHSGeometry, unsigned uOpenFlags,
1321 PFNVDPROGRESS pfnProgress, void *pvUser,
1322 unsigned uPercentStart, unsigned uPercentSpan)
1323{
1324 RT_NOREF1(pszComment);
1325 int rc;
1326 int32_t fOpen;
1327
1328 if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
1329 {
1330 rc = vdIfError(pImage->pIfError, VERR_VD_INVALID_TYPE, RT_SRC_POS, N_("Qed: cannot create fixed image '%s'"), pImage->pszFilename);
1331 goto out;
1332 }
1333
1334 pImage->uOpenFlags = uOpenFlags & ~VD_OPEN_FLAGS_READONLY;
1335 pImage->uImageFlags = uImageFlags;
1336 pImage->PCHSGeometry = *pPCHSGeometry;
1337 pImage->LCHSGeometry = *pLCHSGeometry;
1338
1339 pImage->pIfError = VDIfErrorGet(pImage->pVDIfsDisk);
1340 pImage->pIfIo = VDIfIoIntGet(pImage->pVDIfsImage);
1341 AssertPtrReturn(pImage->pIfIo, VERR_INVALID_PARAMETER);
1342
1343 /* Create image file. */
1344 fOpen = VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags, true /* fCreate */);
1345 rc = vdIfIoIntFileOpen(pImage->pIfIo, pImage->pszFilename, fOpen, &pImage->pStorage);
1346 if (RT_FAILURE(rc))
1347 {
1348 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("Qed: cannot create image '%s'"), pImage->pszFilename);
1349 goto out;
1350 }
1351
1352 /* Init image state. */
1353 pImage->cbSize = cbSize;
1354 pImage->cbCluster = QED_CLUSTER_SIZE_DEFAULT;
1355 pImage->cbTable = qedCluster2Byte(pImage, QED_TABLE_SIZE_DEFAULT);
1356 pImage->cTableEntries = pImage->cbTable / sizeof(uint64_t);
1357 pImage->offL1Table = qedCluster2Byte(pImage, 1); /* Cluster 0 is the header. */
1358 pImage->cbImage = (1 * pImage->cbCluster) + pImage->cbTable; /* Header + L1 table size. */
1359 pImage->cbBackingFilename = 0;
1360 pImage->offBackingFilename = 0;
1361 qedTableMasksInit(pImage);
1362
1363 /* Init L1 table. */
1364 pImage->paL1Table = (uint64_t *)RTMemAllocZ(pImage->cbTable);
1365 if (!pImage->paL1Table)
1366 {
1367 rc = vdIfError(pImage->pIfError, VERR_NO_MEMORY, RT_SRC_POS, N_("Qed: cannot allocate memory for L1 table of image '%s'"),
1368 pImage->pszFilename);
1369 goto out;
1370 }
1371
1372 rc = qedL2TblCacheCreate(pImage);
1373 if (RT_FAILURE(rc))
1374 {
1375 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("Qed: Failed to create L2 cache for image '%s'"),
1376 pImage->pszFilename);
1377 goto out;
1378 }
1379
1380 if (RT_SUCCESS(rc) && pfnProgress)
1381 pfnProgress(pvUser, uPercentStart + uPercentSpan * 98 / 100);
1382
1383 rc = qedFlushImage(pImage);
1384
1385out:
1386 if (RT_SUCCESS(rc) && pfnProgress)
1387 pfnProgress(pvUser, uPercentStart + uPercentSpan);
1388
1389 if (RT_FAILURE(rc))
1390 qedFreeImage(pImage, rc != VERR_ALREADY_EXISTS);
1391 return rc;
1392}
1393
1394/**
1395 * Rollback anything done during async cluster allocation.
1396 *
1397 * @returns VBox status code.
1398 * @param pImage The image instance data.
1399 * @param pIoCtx The I/O context.
1400 * @param pClusterAlloc The cluster allocation to rollback.
1401 */
1402static int qedAsyncClusterAllocRollback(PQEDIMAGE pImage, PVDIOCTX pIoCtx, PQEDCLUSTERASYNCALLOC pClusterAlloc)
1403{
1404 RT_NOREF1(pIoCtx);
1405 int rc = VINF_SUCCESS;
1406
1407 switch (pClusterAlloc->enmAllocState)
1408 {
1409 case QEDCLUSTERASYNCALLOCSTATE_L2_ALLOC:
1410 case QEDCLUSTERASYNCALLOCSTATE_L2_LINK:
1411 {
1412 /* Assumption right now is that the L1 table is not modified if the link fails. */
1413 rc = vdIfIoIntFileSetSize(pImage->pIfIo, pImage->pStorage, pClusterAlloc->cbImageOld);
1414 qedL2TblCacheEntryRelease(pClusterAlloc->pL2Entry); /* Release L2 cache entry. */
1415 qedL2TblCacheEntryFree(pImage, pClusterAlloc->pL2Entry); /* Free it, it is not in the cache yet. */
1416 break;
1417 }
1418 case QEDCLUSTERASYNCALLOCSTATE_USER_ALLOC:
1419 case QEDCLUSTERASYNCALLOCSTATE_USER_LINK:
1420 {
1421 /* Assumption right now is that the L2 table is not modified if the link fails. */
1422 rc = vdIfIoIntFileSetSize(pImage->pIfIo, pImage->pStorage, pClusterAlloc->cbImageOld);
1423 qedL2TblCacheEntryRelease(pClusterAlloc->pL2Entry); /* Release L2 cache entry. */
1424 break;
1425 }
1426 default:
1427 AssertMsgFailed(("Invalid cluster allocation state %d\n", pClusterAlloc->enmAllocState));
1428 rc = VERR_INVALID_STATE;
1429 }
1430
1431 RTMemFree(pClusterAlloc);
1432 return rc;
1433}
1434
1435/**
1436 * Updates the state of the async cluster allocation.
1437 *
1438 * @returns VBox status code.
1439 * @param pBackendData The opaque backend data.
1440 * @param pIoCtx I/O context associated with this request.
1441 * @param pvUser Opaque user data passed during a read/write request.
1442 * @param rcReq Status code for the completed request.
1443 */
1444static DECLCALLBACK(int) qedAsyncClusterAllocUpdate(void *pBackendData, PVDIOCTX pIoCtx, void *pvUser, int rcReq)
1445{
1446 int rc = VINF_SUCCESS;
1447 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
1448 PQEDCLUSTERASYNCALLOC pClusterAlloc = (PQEDCLUSTERASYNCALLOC)pvUser;
1449
1450 if (RT_FAILURE(rcReq))
1451 return qedAsyncClusterAllocRollback(pImage, pIoCtx, pClusterAlloc);
1452
1453 AssertPtr(pClusterAlloc->pL2Entry);
1454
1455 switch (pClusterAlloc->enmAllocState)
1456 {
1457 case QEDCLUSTERASYNCALLOCSTATE_L2_ALLOC:
1458 {
1459 uint64_t offUpdateLe = RT_H2LE_U64(pClusterAlloc->pL2Entry->offL2Tbl);
1460
1461 /* Update the link in the on disk L1 table now. */
1462 pClusterAlloc->enmAllocState = QEDCLUSTERASYNCALLOCSTATE_L2_LINK;
1463 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pImage->pStorage,
1464 pImage->offL1Table + pClusterAlloc->idxL1*sizeof(uint64_t),
1465 &offUpdateLe, sizeof(uint64_t), pIoCtx,
1466 qedAsyncClusterAllocUpdate, pClusterAlloc);
1467 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
1468 break;
1469 else if (RT_FAILURE(rc))
1470 {
1471 /* Rollback. */
1472 qedAsyncClusterAllocRollback(pImage, pIoCtx, pClusterAlloc);
1473 break;
1474 }
1475 /* Success, fall through. */
1476 }
1477 case QEDCLUSTERASYNCALLOCSTATE_L2_LINK:
1478 {
1479 /* L2 link updated in L1 , save L2 entry in cache and allocate new user data cluster. */
1480 uint64_t offData = qedClusterAllocate(pImage, 1);
1481
1482 /* Update the link in the in memory L1 table now. */
1483 pImage->paL1Table[pClusterAlloc->idxL1] = pClusterAlloc->pL2Entry->offL2Tbl;
1484 qedL2TblCacheEntryInsert(pImage, pClusterAlloc->pL2Entry);
1485
1486 pClusterAlloc->enmAllocState = QEDCLUSTERASYNCALLOCSTATE_USER_ALLOC;
1487 pClusterAlloc->cbImageOld = offData;
1488 pClusterAlloc->offClusterNew = offData;
1489
1490 /* Write data. */
1491 rc = vdIfIoIntFileWriteUser(pImage->pIfIo, pImage->pStorage,
1492 offData, pIoCtx, pClusterAlloc->cbToWrite,
1493 qedAsyncClusterAllocUpdate, pClusterAlloc);
1494 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
1495 break;
1496 else if (RT_FAILURE(rc))
1497 {
1498 qedAsyncClusterAllocRollback(pImage, pIoCtx, pClusterAlloc);
1499 RTMemFree(pClusterAlloc);
1500 break;
1501 }
1502 }
1503 case QEDCLUSTERASYNCALLOCSTATE_USER_ALLOC:
1504 {
1505 uint64_t offUpdateLe = RT_H2LE_U64(pClusterAlloc->offClusterNew);
1506
1507 pClusterAlloc->enmAllocState = QEDCLUSTERASYNCALLOCSTATE_USER_LINK;
1508
1509 /* Link L2 table and update it. */
1510 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pImage->pStorage,
1511 pImage->paL1Table[pClusterAlloc->idxL1] + pClusterAlloc->idxL2*sizeof(uint64_t),
1512 &offUpdateLe, sizeof(uint64_t), pIoCtx,
1513 qedAsyncClusterAllocUpdate, pClusterAlloc);
1514 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
1515 break;
1516 else if (RT_FAILURE(rc))
1517 {
1518 qedAsyncClusterAllocRollback(pImage, pIoCtx, pClusterAlloc);
1519 RTMemFree(pClusterAlloc);
1520 break;
1521 }
1522 }
1523 case QEDCLUSTERASYNCALLOCSTATE_USER_LINK:
1524 {
1525 /* Everything done without errors, signal completion. */
1526 pClusterAlloc->pL2Entry->paL2Tbl[pClusterAlloc->idxL2] = pClusterAlloc->offClusterNew;
1527 qedL2TblCacheEntryRelease(pClusterAlloc->pL2Entry);
1528 RTMemFree(pClusterAlloc);
1529 rc = VINF_SUCCESS;
1530 break;
1531 }
1532 default:
1533 AssertMsgFailed(("Invalid async cluster allocation state %d\n",
1534 pClusterAlloc->enmAllocState));
1535 }
1536
1537 return rc;
1538}
1539
1540/** @copydoc VBOXHDDBACKEND::pfnCheckIfValid */
1541static DECLCALLBACK(int) qedCheckIfValid(const char *pszFilename, PVDINTERFACE pVDIfsDisk,
1542 PVDINTERFACE pVDIfsImage, VDTYPE *penmType)
1543{
1544 RT_NOREF1(pVDIfsDisk);
1545 LogFlowFunc(("pszFilename=\"%s\" pVDIfsDisk=%#p pVDIfsImage=%#p\n", pszFilename, pVDIfsDisk, pVDIfsImage));
1546 PVDIOSTORAGE pStorage = NULL;
1547 uint64_t cbFile;
1548 int rc = VINF_SUCCESS;
1549
1550 /* Get I/O interface. */
1551 PVDINTERFACEIOINT pIfIo = VDIfIoIntGet(pVDIfsImage);
1552 AssertPtrReturn(pIfIo, VERR_INVALID_PARAMETER);
1553
1554 if ( !VALID_PTR(pszFilename)
1555 || !*pszFilename)
1556 {
1557 rc = VERR_INVALID_PARAMETER;
1558 goto out;
1559 }
1560
1561 /*
1562 * Open the file and read the footer.
1563 */
1564 rc = vdIfIoIntFileOpen(pIfIo, pszFilename,
1565 VDOpenFlagsToFileOpenFlags(VD_OPEN_FLAGS_READONLY,
1566 false /* fCreate */),
1567 &pStorage);
1568 if (RT_SUCCESS(rc))
1569 {
1570 rc = vdIfIoIntFileGetSize(pIfIo, pStorage, &cbFile);
1571 if ( RT_SUCCESS(rc)
1572 && cbFile > sizeof(QedHeader))
1573 {
1574 QedHeader Header;
1575
1576 rc = vdIfIoIntFileReadSync(pIfIo, pStorage, 0, &Header, sizeof(Header));
1577 if ( RT_SUCCESS(rc)
1578 && qedHdrConvertToHostEndianess(&Header))
1579 {
1580 *penmType = VDTYPE_HDD;
1581 rc = VINF_SUCCESS;
1582 }
1583 else
1584 rc = VERR_VD_GEN_INVALID_HEADER;
1585 }
1586 else
1587 rc = VERR_VD_GEN_INVALID_HEADER;
1588 }
1589
1590 if (pStorage)
1591 vdIfIoIntFileClose(pIfIo, pStorage);
1592
1593out:
1594 LogFlowFunc(("returns %Rrc\n", rc));
1595 return rc;
1596}
1597
1598/** @copydoc VBOXHDDBACKEND::pfnOpen */
1599static DECLCALLBACK(int) qedOpen(const char *pszFilename, unsigned uOpenFlags,
1600 PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
1601 VDTYPE enmType, void **ppBackendData)
1602{
1603 LogFlowFunc(("pszFilename=\"%s\" uOpenFlags=%#x pVDIfsDisk=%#p pVDIfsImage=%#p enmType=%u ppBackendData=%#p\n", pszFilename, uOpenFlags, pVDIfsDisk, pVDIfsImage, enmType, ppBackendData));
1604 int rc;
1605 PQEDIMAGE pImage;
1606
1607 NOREF(enmType); /**< @todo r=klaus make use of the type info. */
1608
1609 /* Check open flags. All valid flags are supported. */
1610 if (uOpenFlags & ~VD_OPEN_FLAGS_MASK)
1611 {
1612 rc = VERR_INVALID_PARAMETER;
1613 goto out;
1614 }
1615
1616 /* Check remaining arguments. */
1617 if ( !VALID_PTR(pszFilename)
1618 || !*pszFilename)
1619 {
1620 rc = VERR_INVALID_PARAMETER;
1621 goto out;
1622 }
1623
1624
1625 pImage = (PQEDIMAGE)RTMemAllocZ(sizeof(QEDIMAGE));
1626 if (!pImage)
1627 {
1628 rc = VERR_NO_MEMORY;
1629 goto out;
1630 }
1631 pImage->pszFilename = pszFilename;
1632 pImage->pStorage = NULL;
1633 pImage->pVDIfsDisk = pVDIfsDisk;
1634 pImage->pVDIfsImage = pVDIfsImage;
1635
1636 rc = qedOpenImage(pImage, uOpenFlags);
1637 if (RT_SUCCESS(rc))
1638 *ppBackendData = pImage;
1639 else
1640 RTMemFree(pImage);
1641
1642out:
1643 LogFlowFunc(("returns %Rrc (pBackendData=%#p)\n", rc, *ppBackendData));
1644 return rc;
1645}
1646
1647/** @copydoc VBOXHDDBACKEND::pfnCreate */
1648static DECLCALLBACK(int) qedCreate(const char *pszFilename, uint64_t cbSize,
1649 unsigned uImageFlags, const char *pszComment,
1650 PCVDGEOMETRY pPCHSGeometry, PCVDGEOMETRY pLCHSGeometry,
1651 PCRTUUID pUuid, unsigned uOpenFlags,
1652 unsigned uPercentStart, unsigned uPercentSpan,
1653 PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
1654 PVDINTERFACE pVDIfsOperation, VDTYPE enmType,
1655 void **ppBackendData)
1656{
1657 RT_NOREF1(pUuid);
1658 LogFlowFunc(("pszFilename=\"%s\" cbSize=%llu uImageFlags=%#x pszComment=\"%s\" pPCHSGeometry=%#p pLCHSGeometry=%#p Uuid=%RTuuid uOpenFlags=%#x uPercentStart=%u uPercentSpan=%u pVDIfsDisk=%#p pVDIfsImage=%#p pVDIfsOperation=%#p enmType=%d ppBackendData=%#p",
1659 pszFilename, cbSize, uImageFlags, pszComment, pPCHSGeometry, pLCHSGeometry, pUuid, uOpenFlags, uPercentStart, uPercentSpan, pVDIfsDisk, pVDIfsImage, pVDIfsOperation, enmType, ppBackendData));
1660 int rc;
1661 PQEDIMAGE pImage;
1662
1663 PFNVDPROGRESS pfnProgress = NULL;
1664 void *pvUser = NULL;
1665 PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
1666 if (pIfProgress)
1667 {
1668 pfnProgress = pIfProgress->pfnProgress;
1669 pvUser = pIfProgress->Core.pvUser;
1670 }
1671
1672 /* Check the VD container type. */
1673 if (enmType != VDTYPE_HDD)
1674 {
1675 rc = VERR_VD_INVALID_TYPE;
1676 goto out;
1677 }
1678
1679 /* Check open flags. All valid flags are supported. */
1680 if (uOpenFlags & ~VD_OPEN_FLAGS_MASK)
1681 {
1682 rc = VERR_INVALID_PARAMETER;
1683 goto out;
1684 }
1685
1686 /* Check remaining arguments. */
1687 if ( !VALID_PTR(pszFilename)
1688 || !*pszFilename
1689 || !VALID_PTR(pPCHSGeometry)
1690 || !VALID_PTR(pLCHSGeometry))
1691 {
1692 rc = VERR_INVALID_PARAMETER;
1693 goto out;
1694 }
1695
1696 pImage = (PQEDIMAGE)RTMemAllocZ(sizeof(QEDIMAGE));
1697 if (!pImage)
1698 {
1699 rc = VERR_NO_MEMORY;
1700 goto out;
1701 }
1702 pImage->pszFilename = pszFilename;
1703 pImage->pStorage = NULL;
1704 pImage->pVDIfsDisk = pVDIfsDisk;
1705 pImage->pVDIfsImage = pVDIfsImage;
1706
1707 rc = qedCreateImage(pImage, cbSize, uImageFlags, pszComment,
1708 pPCHSGeometry, pLCHSGeometry, uOpenFlags,
1709 pfnProgress, pvUser, uPercentStart, uPercentSpan);
1710 if (RT_SUCCESS(rc))
1711 {
1712 /* So far the image is opened in read/write mode. Make sure the
1713 * image is opened in read-only mode if the caller requested that. */
1714 if (uOpenFlags & VD_OPEN_FLAGS_READONLY)
1715 {
1716 qedFreeImage(pImage, false);
1717 rc = qedOpenImage(pImage, uOpenFlags);
1718 if (RT_FAILURE(rc))
1719 {
1720 RTMemFree(pImage);
1721 goto out;
1722 }
1723 }
1724 *ppBackendData = pImage;
1725 }
1726 else
1727 RTMemFree(pImage);
1728
1729out:
1730 LogFlowFunc(("returns %Rrc (pBackendData=%#p)\n", rc, *ppBackendData));
1731 return rc;
1732}
1733
1734/** @copydoc VBOXHDDBACKEND::pfnRename */
1735static DECLCALLBACK(int) qedRename(void *pBackendData, const char *pszFilename)
1736{
1737 LogFlowFunc(("pBackendData=%#p pszFilename=%#p\n", pBackendData, pszFilename));
1738 int rc = VINF_SUCCESS;
1739 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
1740
1741 /* Check arguments. */
1742 if ( !pImage
1743 || !pszFilename
1744 || !*pszFilename)
1745 {
1746 rc = VERR_INVALID_PARAMETER;
1747 goto out;
1748 }
1749
1750 /* Close the image. */
1751 rc = qedFreeImage(pImage, false);
1752 if (RT_FAILURE(rc))
1753 goto out;
1754
1755 /* Rename the file. */
1756 rc = vdIfIoIntFileMove(pImage->pIfIo, pImage->pszFilename, pszFilename, 0);
1757 if (RT_FAILURE(rc))
1758 {
1759 /* The move failed, try to reopen the original image. */
1760 int rc2 = qedOpenImage(pImage, pImage->uOpenFlags);
1761 if (RT_FAILURE(rc2))
1762 rc = rc2;
1763
1764 goto out;
1765 }
1766
1767 /* Update pImage with the new information. */
1768 pImage->pszFilename = pszFilename;
1769
1770 /* Open the old image with new name. */
1771 rc = qedOpenImage(pImage, pImage->uOpenFlags);
1772 if (RT_FAILURE(rc))
1773 goto out;
1774
1775out:
1776 LogFlowFunc(("returns %Rrc\n", rc));
1777 return rc;
1778}
1779
1780/** @copydoc VBOXHDDBACKEND::pfnClose */
1781static DECLCALLBACK(int) qedClose(void *pBackendData, bool fDelete)
1782{
1783 LogFlowFunc(("pBackendData=%#p fDelete=%d\n", pBackendData, fDelete));
1784 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
1785 int rc;
1786
1787 rc = qedFreeImage(pImage, fDelete);
1788 RTMemFree(pImage);
1789
1790 LogFlowFunc(("returns %Rrc\n", rc));
1791 return rc;
1792}
1793
1794static DECLCALLBACK(int) qedRead(void *pBackendData, uint64_t uOffset, size_t cbToRead,
1795 PVDIOCTX pIoCtx, size_t *pcbActuallyRead)
1796{
1797 LogFlowFunc(("pBackendData=%#p uOffset=%llu pIoCtx=%#p cbToRead=%zu pcbActuallyRead=%#p\n",
1798 pBackendData, uOffset, pIoCtx, cbToRead, pcbActuallyRead));
1799 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
1800 uint32_t offCluster = 0;
1801 uint32_t idxL1 = 0;
1802 uint32_t idxL2 = 0;
1803 uint64_t offFile = 0;
1804 int rc;
1805
1806 AssertPtr(pImage);
1807 Assert(uOffset % 512 == 0);
1808 Assert(cbToRead % 512 == 0);
1809
1810 if (!VALID_PTR(pIoCtx) || !cbToRead)
1811 {
1812 rc = VERR_INVALID_PARAMETER;
1813 goto out;
1814 }
1815
1816 if ( uOffset + cbToRead > pImage->cbSize
1817 || cbToRead == 0)
1818 {
1819 rc = VERR_INVALID_PARAMETER;
1820 goto out;
1821 }
1822
1823 qedConvertLogicalOffset(pImage, uOffset, &idxL1, &idxL2, &offCluster);
1824
1825 /* Clip read size to remain in the cluster. */
1826 cbToRead = RT_MIN(cbToRead, pImage->cbCluster - offCluster);
1827
1828 /* Get offset in image. */
1829 rc = qedConvertToImageOffset(pImage, pIoCtx, idxL1, idxL2, offCluster, &offFile);
1830 if (RT_SUCCESS(rc))
1831 rc = vdIfIoIntFileReadUser(pImage->pIfIo, pImage->pStorage, offFile,
1832 pIoCtx, cbToRead);
1833
1834 if ( ( RT_SUCCESS(rc)
1835 || rc == VERR_VD_BLOCK_FREE
1836 || rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
1837 && pcbActuallyRead)
1838 *pcbActuallyRead = cbToRead;
1839
1840out:
1841 LogFlowFunc(("returns %Rrc\n", rc));
1842 return rc;
1843}
1844
1845static DECLCALLBACK(int) qedWrite(void *pBackendData, uint64_t uOffset, size_t cbToWrite,
1846 PVDIOCTX pIoCtx, size_t *pcbWriteProcess, size_t *pcbPreRead,
1847 size_t *pcbPostRead, unsigned fWrite)
1848{
1849 LogFlowFunc(("pBackendData=%#p uOffset=%llu pIoCtx=%#p cbToWrite=%zu pcbWriteProcess=%#p pcbPreRead=%#p pcbPostRead=%#p\n",
1850 pBackendData, uOffset, pIoCtx, cbToWrite, pcbWriteProcess, pcbPreRead, pcbPostRead));
1851 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
1852 uint32_t offCluster = 0;
1853 uint32_t idxL1 = 0;
1854 uint32_t idxL2 = 0;
1855 uint64_t offImage = 0;
1856 int rc = VINF_SUCCESS;
1857
1858 AssertPtr(pImage);
1859 Assert(!(uOffset % 512));
1860 Assert(!(cbToWrite % 512));
1861
1862 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
1863 {
1864 rc = VERR_VD_IMAGE_READ_ONLY;
1865 goto out;
1866 }
1867
1868 if (!VALID_PTR(pIoCtx) || !cbToWrite)
1869 {
1870 rc = VERR_INVALID_PARAMETER;
1871 goto out;
1872 }
1873
1874 if ( uOffset + cbToWrite > pImage->cbSize
1875 || cbToWrite == 0)
1876 {
1877 rc = VERR_INVALID_PARAMETER;
1878 goto out;
1879 }
1880
1881 /* Convert offset to L1, L2 index and cluster offset. */
1882 qedConvertLogicalOffset(pImage, uOffset, &idxL1, &idxL2, &offCluster);
1883
1884 /* Clip write size to remain in the cluster. */
1885 cbToWrite = RT_MIN(cbToWrite, pImage->cbCluster - offCluster);
1886 Assert(!(cbToWrite % 512));
1887
1888 /* Get offset in image. */
1889 rc = qedConvertToImageOffset(pImage, pIoCtx, idxL1, idxL2, offCluster, &offImage);
1890 if (RT_SUCCESS(rc))
1891 rc = vdIfIoIntFileWriteUser(pImage->pIfIo, pImage->pStorage,
1892 offImage, pIoCtx, cbToWrite, NULL, NULL);
1893 else if (rc == VERR_VD_BLOCK_FREE)
1894 {
1895 if ( cbToWrite == pImage->cbCluster
1896 && !(fWrite & VD_WRITE_NO_ALLOC))
1897 {
1898 PQEDL2CACHEENTRY pL2Entry = NULL;
1899
1900 /* Full cluster write to previously unallocated cluster.
1901 * Allocate cluster and write data. */
1902 Assert(!offCluster);
1903
1904 do
1905 {
1906 /* Check if we have to allocate a new cluster for L2 tables. */
1907 if (!pImage->paL1Table[idxL1])
1908 {
1909 uint64_t offL2Tbl;
1910 PQEDCLUSTERASYNCALLOC pL2ClusterAlloc = NULL;
1911
1912 /* Allocate new async cluster allocation state. */
1913 pL2ClusterAlloc = (PQEDCLUSTERASYNCALLOC)RTMemAllocZ(sizeof(QEDCLUSTERASYNCALLOC));
1914 if (RT_UNLIKELY(!pL2ClusterAlloc))
1915 {
1916 rc = VERR_NO_MEMORY;
1917 break;
1918 }
1919
1920 pL2Entry = qedL2TblCacheEntryAlloc(pImage);
1921 if (!pL2Entry)
1922 {
1923 rc = VERR_NO_MEMORY;
1924 RTMemFree(pL2ClusterAlloc);
1925 break;
1926 }
1927
1928 offL2Tbl = qedClusterAllocate(pImage, qedByte2Cluster(pImage, pImage->cbTable));
1929 pL2Entry->offL2Tbl = offL2Tbl;
1930 memset(pL2Entry->paL2Tbl, 0, pImage->cbTable);
1931
1932 pL2ClusterAlloc->enmAllocState = QEDCLUSTERASYNCALLOCSTATE_L2_ALLOC;
1933 pL2ClusterAlloc->cbImageOld = offL2Tbl;
1934 pL2ClusterAlloc->offClusterNew = offL2Tbl;
1935 pL2ClusterAlloc->idxL1 = idxL1;
1936 pL2ClusterAlloc->idxL2 = idxL2;
1937 pL2ClusterAlloc->cbToWrite = cbToWrite;
1938 pL2ClusterAlloc->pL2Entry = pL2Entry;
1939
1940 /*
1941 * Write the L2 table first and link to the L1 table afterwards.
1942 * If something unexpected happens the worst case which can happen
1943 * is a leak of some clusters.
1944 */
1945 rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pImage->pStorage,
1946 offL2Tbl, pL2Entry->paL2Tbl, pImage->cbTable, pIoCtx,
1947 qedAsyncClusterAllocUpdate, pL2ClusterAlloc);
1948 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
1949 break;
1950 else if (RT_FAILURE(rc))
1951 {
1952 RTMemFree(pL2ClusterAlloc);
1953 qedL2TblCacheEntryFree(pImage, pL2Entry);
1954 break;
1955 }
1956
1957 rc = qedAsyncClusterAllocUpdate(pImage, pIoCtx, pL2ClusterAlloc, rc);
1958 }
1959 else
1960 {
1961 rc = qedL2TblCacheFetchAsync(pImage, pIoCtx, pImage->paL1Table[idxL1],
1962 &pL2Entry);
1963
1964 if (RT_SUCCESS(rc))
1965 {
1966 PQEDCLUSTERASYNCALLOC pDataClusterAlloc = NULL;
1967
1968 /* Allocate new async cluster allocation state. */
1969 pDataClusterAlloc = (PQEDCLUSTERASYNCALLOC)RTMemAllocZ(sizeof(QEDCLUSTERASYNCALLOC));
1970 if (RT_UNLIKELY(!pDataClusterAlloc))
1971 {
1972 rc = VERR_NO_MEMORY;
1973 break;
1974 }
1975
1976 /* Allocate new cluster for the data. */
1977 uint64_t offData = qedClusterAllocate(pImage, 1);
1978
1979 pDataClusterAlloc->enmAllocState = QEDCLUSTERASYNCALLOCSTATE_USER_ALLOC;
1980 pDataClusterAlloc->cbImageOld = offData;
1981 pDataClusterAlloc->offClusterNew = offData;
1982 pDataClusterAlloc->idxL1 = idxL1;
1983 pDataClusterAlloc->idxL2 = idxL2;
1984 pDataClusterAlloc->cbToWrite = cbToWrite;
1985 pDataClusterAlloc->pL2Entry = pL2Entry;
1986
1987 /* Write data. */
1988 rc = vdIfIoIntFileWriteUser(pImage->pIfIo, pImage->pStorage,
1989 offData, pIoCtx, cbToWrite,
1990 qedAsyncClusterAllocUpdate, pDataClusterAlloc);
1991 if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
1992 break;
1993 else if (RT_FAILURE(rc))
1994 {
1995 RTMemFree(pDataClusterAlloc);
1996 break;
1997 }
1998
1999 rc = qedAsyncClusterAllocUpdate(pImage, pIoCtx, pDataClusterAlloc, rc);
2000 }
2001 }
2002
2003 } while (0);
2004
2005 *pcbPreRead = 0;
2006 *pcbPostRead = 0;
2007 }
2008 else
2009 {
2010 /* Trying to do a partial write to an unallocated cluster. Don't do
2011 * anything except letting the upper layer know what to do. */
2012 *pcbPreRead = offCluster;
2013 *pcbPostRead = pImage->cbCluster - cbToWrite - *pcbPreRead;
2014 }
2015 }
2016
2017 if (pcbWriteProcess)
2018 *pcbWriteProcess = cbToWrite;
2019
2020
2021out:
2022 LogFlowFunc(("returns %Rrc\n", rc));
2023 return rc;
2024}
2025
2026static DECLCALLBACK(int) qedFlush(void *pBackendData, PVDIOCTX pIoCtx)
2027{
2028 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
2029 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2030 int rc = VINF_SUCCESS;
2031
2032 Assert(pImage);
2033
2034 if (VALID_PTR(pIoCtx))
2035 rc = qedFlushImageAsync(pImage, pIoCtx);
2036 else
2037 rc = VERR_INVALID_PARAMETER;
2038
2039 LogFlowFunc(("returns %Rrc\n", rc));
2040 return rc;
2041}
2042
2043/** @copydoc VBOXHDDBACKEND::pfnGetVersion */
2044static DECLCALLBACK(unsigned) qedGetVersion(void *pBackendData)
2045{
2046 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
2047 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2048
2049 AssertPtr(pImage);
2050
2051 if (pImage)
2052 return 1;
2053 else
2054 return 0;
2055}
2056
2057/** @copydoc VBOXHDDBACKEND::pfnGetSectorSize */
2058static DECLCALLBACK(uint32_t) qedGetSectorSize(void *pBackendData)
2059{
2060 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
2061 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2062 uint32_t cb = 0;
2063
2064 AssertPtr(pImage);
2065
2066 if (pImage && pImage->pStorage)
2067 cb = 512;
2068
2069 LogFlowFunc(("returns %u\n", cb));
2070 return cb;
2071}
2072
2073/** @copydoc VBOXHDDBACKEND::pfnGetSize */
2074static DECLCALLBACK(uint64_t) qedGetSize(void *pBackendData)
2075{
2076 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
2077 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2078 uint64_t cb = 0;
2079
2080 AssertPtr(pImage);
2081
2082 if (pImage && pImage->pStorage)
2083 cb = pImage->cbSize;
2084
2085 LogFlowFunc(("returns %llu\n", cb));
2086 return cb;
2087}
2088
2089/** @copydoc VBOXHDDBACKEND::pfnGetFileSize */
2090static DECLCALLBACK(uint64_t) qedGetFileSize(void *pBackendData)
2091{
2092 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
2093 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2094 uint64_t cb = 0;
2095
2096 AssertPtr(pImage);
2097
2098 if (pImage)
2099 {
2100 uint64_t cbFile;
2101 if (pImage->pStorage)
2102 {
2103 int rc = vdIfIoIntFileGetSize(pImage->pIfIo, pImage->pStorage, &cbFile);
2104 if (RT_SUCCESS(rc))
2105 cb += cbFile;
2106 }
2107 }
2108
2109 LogFlowFunc(("returns %lld\n", cb));
2110 return cb;
2111}
2112
2113/** @copydoc VBOXHDDBACKEND::pfnGetPCHSGeometry */
2114static DECLCALLBACK(int) qedGetPCHSGeometry(void *pBackendData,
2115 PVDGEOMETRY pPCHSGeometry)
2116{
2117 LogFlowFunc(("pBackendData=%#p pPCHSGeometry=%#p\n", pBackendData, pPCHSGeometry));
2118 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2119 int rc;
2120
2121 AssertPtr(pImage);
2122
2123 if (pImage)
2124 {
2125 if (pImage->PCHSGeometry.cCylinders)
2126 {
2127 *pPCHSGeometry = pImage->PCHSGeometry;
2128 rc = VINF_SUCCESS;
2129 }
2130 else
2131 rc = VERR_VD_GEOMETRY_NOT_SET;
2132 }
2133 else
2134 rc = VERR_VD_NOT_OPENED;
2135
2136 LogFlowFunc(("returns %Rrc (PCHS=%u/%u/%u)\n", rc, pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads, pPCHSGeometry->cSectors));
2137 return rc;
2138}
2139
2140/** @copydoc VBOXHDDBACKEND::pfnSetPCHSGeometry */
2141static DECLCALLBACK(int) qedSetPCHSGeometry(void *pBackendData,
2142 PCVDGEOMETRY pPCHSGeometry)
2143{
2144 LogFlowFunc(("pBackendData=%#p pPCHSGeometry=%#p PCHS=%u/%u/%u\n", pBackendData, pPCHSGeometry, pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads, pPCHSGeometry->cSectors));
2145 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2146 int rc;
2147
2148 AssertPtr(pImage);
2149
2150 if (pImage)
2151 {
2152 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2153 {
2154 rc = VERR_VD_IMAGE_READ_ONLY;
2155 goto out;
2156 }
2157
2158 pImage->PCHSGeometry = *pPCHSGeometry;
2159 rc = VINF_SUCCESS;
2160 }
2161 else
2162 rc = VERR_VD_NOT_OPENED;
2163
2164out:
2165 LogFlowFunc(("returns %Rrc\n", rc));
2166 return rc;
2167}
2168
2169/** @copydoc VBOXHDDBACKEND::pfnGetLCHSGeometry */
2170static DECLCALLBACK(int) qedGetLCHSGeometry(void *pBackendData,
2171 PVDGEOMETRY pLCHSGeometry)
2172{
2173 LogFlowFunc(("pBackendData=%#p pLCHSGeometry=%#p\n", pBackendData, pLCHSGeometry));
2174 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2175 int rc;
2176
2177 AssertPtr(pImage);
2178
2179 if (pImage)
2180 {
2181 if (pImage->LCHSGeometry.cCylinders)
2182 {
2183 *pLCHSGeometry = pImage->LCHSGeometry;
2184 rc = VINF_SUCCESS;
2185 }
2186 else
2187 rc = VERR_VD_GEOMETRY_NOT_SET;
2188 }
2189 else
2190 rc = VERR_VD_NOT_OPENED;
2191
2192 LogFlowFunc(("returns %Rrc (LCHS=%u/%u/%u)\n", rc, pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads, pLCHSGeometry->cSectors));
2193 return rc;
2194}
2195
2196/** @copydoc VBOXHDDBACKEND::pfnSetLCHSGeometry */
2197static DECLCALLBACK(int) qedSetLCHSGeometry(void *pBackendData,
2198 PCVDGEOMETRY pLCHSGeometry)
2199{
2200 LogFlowFunc(("pBackendData=%#p pLCHSGeometry=%#p LCHS=%u/%u/%u\n", pBackendData, pLCHSGeometry, pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads, pLCHSGeometry->cSectors));
2201 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2202 int rc;
2203
2204 AssertPtr(pImage);
2205
2206 if (pImage)
2207 {
2208 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2209 {
2210 rc = VERR_VD_IMAGE_READ_ONLY;
2211 goto out;
2212 }
2213
2214 pImage->LCHSGeometry = *pLCHSGeometry;
2215 rc = VINF_SUCCESS;
2216 }
2217 else
2218 rc = VERR_VD_NOT_OPENED;
2219
2220out:
2221 LogFlowFunc(("returns %Rrc\n", rc));
2222 return rc;
2223}
2224
2225/** @copydoc VBOXHDDBACKEND::pfnGetImageFlags */
2226static DECLCALLBACK(unsigned) qedGetImageFlags(void *pBackendData)
2227{
2228 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
2229 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2230 unsigned uImageFlags;
2231
2232 AssertPtr(pImage);
2233
2234 if (pImage)
2235 uImageFlags = pImage->uImageFlags;
2236 else
2237 uImageFlags = 0;
2238
2239 LogFlowFunc(("returns %#x\n", uImageFlags));
2240 return uImageFlags;
2241}
2242
2243/** @copydoc VBOXHDDBACKEND::pfnGetOpenFlags */
2244static DECLCALLBACK(unsigned) qedGetOpenFlags(void *pBackendData)
2245{
2246 LogFlowFunc(("pBackendData=%#p\n", pBackendData));
2247 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2248 unsigned uOpenFlags;
2249
2250 AssertPtr(pImage);
2251
2252 if (pImage)
2253 uOpenFlags = pImage->uOpenFlags;
2254 else
2255 uOpenFlags = 0;
2256
2257 LogFlowFunc(("returns %#x\n", uOpenFlags));
2258 return uOpenFlags;
2259}
2260
2261/** @copydoc VBOXHDDBACKEND::pfnSetOpenFlags */
2262static DECLCALLBACK(int) qedSetOpenFlags(void *pBackendData, unsigned uOpenFlags)
2263{
2264 LogFlowFunc(("pBackendData=%#p\n uOpenFlags=%#x", pBackendData, uOpenFlags));
2265 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2266 int rc;
2267
2268 /* Image must be opened and the new flags must be valid. */
2269 if (!pImage || (uOpenFlags & ~( VD_OPEN_FLAGS_READONLY | VD_OPEN_FLAGS_INFO
2270 | VD_OPEN_FLAGS_ASYNC_IO | VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS)))
2271 {
2272 rc = VERR_INVALID_PARAMETER;
2273 goto out;
2274 }
2275
2276 /* Implement this operation via reopening the image. */
2277 rc = qedFreeImage(pImage, false);
2278 if (RT_FAILURE(rc))
2279 goto out;
2280 rc = qedOpenImage(pImage, uOpenFlags);
2281
2282out:
2283 LogFlowFunc(("returns %Rrc\n", rc));
2284 return rc;
2285}
2286
2287/** @copydoc VBOXHDDBACKEND::pfnGetComment */
2288static DECLCALLBACK(int) qedGetComment(void *pBackendData, char *pszComment,
2289 size_t cbComment)
2290{
2291 RT_NOREF2(pszComment, cbComment);
2292 LogFlowFunc(("pBackendData=%#p pszComment=%#p cbComment=%zu\n", pBackendData, pszComment, cbComment));
2293 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2294 int rc;
2295
2296 AssertPtr(pImage);
2297
2298 if (pImage)
2299 rc = VERR_NOT_SUPPORTED;
2300 else
2301 rc = VERR_VD_NOT_OPENED;
2302
2303 LogFlowFunc(("returns %Rrc comment='%s'\n", rc, pszComment));
2304 return rc;
2305}
2306
2307/** @copydoc VBOXHDDBACKEND::pfnSetComment */
2308static DECLCALLBACK(int) qedSetComment(void *pBackendData, const char *pszComment)
2309{
2310 RT_NOREF1(pszComment);
2311 LogFlowFunc(("pBackendData=%#p pszComment=\"%s\"\n", pBackendData, pszComment));
2312 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2313 int rc;
2314
2315 AssertPtr(pImage);
2316
2317 if (pImage)
2318 {
2319 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2320 rc = VERR_VD_IMAGE_READ_ONLY;
2321 else
2322 rc = VERR_NOT_SUPPORTED;
2323 }
2324 else
2325 rc = VERR_VD_NOT_OPENED;
2326
2327 LogFlowFunc(("returns %Rrc\n", rc));
2328 return rc;
2329}
2330
2331/** @copydoc VBOXHDDBACKEND::pfnGetUuid */
2332static DECLCALLBACK(int) qedGetUuid(void *pBackendData, PRTUUID pUuid)
2333{
2334 RT_NOREF1(pUuid);
2335 LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
2336 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2337 int rc;
2338
2339 AssertPtr(pImage);
2340
2341 if (pImage)
2342 rc = VERR_NOT_SUPPORTED;
2343 else
2344 rc = VERR_VD_NOT_OPENED;
2345
2346 LogFlowFunc(("returns %Rrc (%RTuuid)\n", rc, pUuid));
2347 return rc;
2348}
2349
2350/** @copydoc VBOXHDDBACKEND::pfnSetUuid */
2351static DECLCALLBACK(int) qedSetUuid(void *pBackendData, PCRTUUID pUuid)
2352{
2353 RT_NOREF1(pUuid);
2354 LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
2355 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2356 int rc;
2357
2358 LogFlowFunc(("%RTuuid\n", pUuid));
2359 AssertPtr(pImage);
2360
2361 if (pImage)
2362 {
2363 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
2364 rc = VERR_NOT_SUPPORTED;
2365 else
2366 rc = VERR_VD_IMAGE_READ_ONLY;
2367 }
2368 else
2369 rc = VERR_VD_NOT_OPENED;
2370
2371 LogFlowFunc(("returns %Rrc\n", rc));
2372 return rc;
2373}
2374
2375/** @copydoc VBOXHDDBACKEND::pfnGetModificationUuid */
2376static DECLCALLBACK(int) qedGetModificationUuid(void *pBackendData, PRTUUID pUuid)
2377{
2378 RT_NOREF1(pUuid);
2379 LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
2380 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2381 int rc;
2382
2383 AssertPtr(pImage);
2384
2385 if (pImage)
2386 rc = VERR_NOT_SUPPORTED;
2387 else
2388 rc = VERR_VD_NOT_OPENED;
2389
2390 LogFlowFunc(("returns %Rrc (%RTuuid)\n", rc, pUuid));
2391 return rc;
2392}
2393
2394/** @copydoc VBOXHDDBACKEND::pfnSetModificationUuid */
2395static DECLCALLBACK(int) qedSetModificationUuid(void *pBackendData, PCRTUUID pUuid)
2396{
2397 RT_NOREF1(pUuid);
2398 LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
2399 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2400 int rc;
2401
2402 AssertPtr(pImage);
2403
2404 if (pImage)
2405 {
2406 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
2407 rc = VERR_NOT_SUPPORTED;
2408 else
2409 rc = VERR_VD_IMAGE_READ_ONLY;
2410 }
2411 else
2412 rc = VERR_VD_NOT_OPENED;
2413
2414 LogFlowFunc(("returns %Rrc\n", rc));
2415 return rc;
2416}
2417
2418/** @copydoc VBOXHDDBACKEND::pfnGetParentUuid */
2419static DECLCALLBACK(int) qedGetParentUuid(void *pBackendData, PRTUUID pUuid)
2420{
2421 RT_NOREF1(pUuid);
2422 LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
2423 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2424 int rc;
2425
2426 AssertPtr(pImage);
2427
2428 if (pImage)
2429 rc = VERR_NOT_SUPPORTED;
2430 else
2431 rc = VERR_VD_NOT_OPENED;
2432
2433 LogFlowFunc(("returns %Rrc (%RTuuid)\n", rc, pUuid));
2434 return rc;
2435}
2436
2437/** @copydoc VBOXHDDBACKEND::pfnSetParentUuid */
2438static DECLCALLBACK(int) qedSetParentUuid(void *pBackendData, PCRTUUID pUuid)
2439{
2440 RT_NOREF1(pUuid);
2441 LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
2442 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2443 int rc;
2444
2445 AssertPtr(pImage);
2446
2447 if (pImage)
2448 {
2449 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
2450 rc = VERR_NOT_SUPPORTED;
2451 else
2452 rc = VERR_VD_IMAGE_READ_ONLY;
2453 }
2454 else
2455 rc = VERR_VD_NOT_OPENED;
2456
2457 LogFlowFunc(("returns %Rrc\n", rc));
2458 return rc;
2459}
2460
2461/** @copydoc VBOXHDDBACKEND::pfnGetParentModificationUuid */
2462static DECLCALLBACK(int) qedGetParentModificationUuid(void *pBackendData, PRTUUID pUuid)
2463{
2464 RT_NOREF1(pUuid);
2465 LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
2466 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2467 int rc;
2468
2469 AssertPtr(pImage);
2470
2471 if (pImage)
2472 rc = VERR_NOT_SUPPORTED;
2473 else
2474 rc = VERR_VD_NOT_OPENED;
2475
2476 LogFlowFunc(("returns %Rrc (%RTuuid)\n", rc, pUuid));
2477 return rc;
2478}
2479
2480/** @copydoc VBOXHDDBACKEND::pfnSetParentModificationUuid */
2481static DECLCALLBACK(int) qedSetParentModificationUuid(void *pBackendData, PCRTUUID pUuid)
2482{
2483 RT_NOREF1(pUuid);
2484 LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
2485 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2486 int rc;
2487
2488 AssertPtr(pImage);
2489
2490 if (pImage)
2491 {
2492 if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
2493 rc = VERR_NOT_SUPPORTED;
2494 else
2495 rc = VERR_VD_IMAGE_READ_ONLY;
2496 }
2497 else
2498 rc = VERR_VD_NOT_OPENED;
2499
2500 LogFlowFunc(("returns %Rrc\n", rc));
2501 return rc;
2502}
2503
2504/** @copydoc VBOXHDDBACKEND::pfnDump */
2505static DECLCALLBACK(void) qedDump(void *pBackendData)
2506{
2507 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2508
2509 AssertPtr(pImage);
2510 if (pImage)
2511 {
2512 vdIfErrorMessage(pImage->pIfError, "Header: Geometry PCHS=%u/%u/%u LCHS=%u/%u/%u cbSector=%llu\n",
2513 pImage->PCHSGeometry.cCylinders, pImage->PCHSGeometry.cHeads, pImage->PCHSGeometry.cSectors,
2514 pImage->LCHSGeometry.cCylinders, pImage->LCHSGeometry.cHeads, pImage->LCHSGeometry.cSectors,
2515 pImage->cbSize / 512);
2516 }
2517}
2518
2519/** @copydoc VBOXHDDBACKEND::pfnGetParentFilename */
2520static DECLCALLBACK(int) qedGetParentFilename(void *pBackendData, char **ppszParentFilename)
2521{
2522 int rc = VINF_SUCCESS;
2523 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2524
2525 AssertPtr(pImage);
2526 if (pImage)
2527 if (pImage->pszBackingFilename)
2528 *ppszParentFilename = RTStrDup(pImage->pszBackingFilename);
2529 else
2530 rc = VERR_NOT_SUPPORTED;
2531 else
2532 rc = VERR_VD_NOT_OPENED;
2533
2534 LogFlowFunc(("returns %Rrc\n", rc));
2535 return rc;
2536}
2537
2538/** @copydoc VBOXHDDBACKEND::pfnSetParentFilename */
2539static DECLCALLBACK(int) qedSetParentFilename(void *pBackendData, const char *pszParentFilename)
2540{
2541 int rc = VINF_SUCCESS;
2542 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2543
2544 AssertPtr(pImage);
2545 if (pImage)
2546 {
2547 if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
2548 rc = VERR_VD_IMAGE_READ_ONLY;
2549 else if ( pImage->pszBackingFilename
2550 && (strlen(pszParentFilename) > pImage->cbBackingFilename))
2551 rc = VERR_NOT_SUPPORTED; /* The new filename is longer than the old one. */
2552 else
2553 {
2554 if (pImage->pszBackingFilename)
2555 RTStrFree(pImage->pszBackingFilename);
2556 pImage->pszBackingFilename = RTStrDup(pszParentFilename);
2557 if (!pImage->pszBackingFilename)
2558 rc = VERR_NO_MEMORY;
2559 else
2560 {
2561 if (!pImage->offBackingFilename)
2562 {
2563 /* Allocate new cluster. */
2564 uint64_t offData = qedClusterAllocate(pImage, 1);
2565
2566 Assert((offData & UINT32_MAX) == offData);
2567 pImage->offBackingFilename = (uint32_t)offData;
2568 pImage->cbBackingFilename = (uint32_t)strlen(pszParentFilename);
2569 rc = vdIfIoIntFileSetSize(pImage->pIfIo, pImage->pStorage,
2570 offData + pImage->cbCluster);
2571 }
2572
2573 if (RT_SUCCESS(rc))
2574 rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pImage->pStorage,
2575 pImage->offBackingFilename,
2576 pImage->pszBackingFilename,
2577 strlen(pImage->pszBackingFilename));
2578 }
2579 }
2580 }
2581 else
2582 rc = VERR_VD_NOT_OPENED;
2583
2584 LogFlowFunc(("returns %Rrc\n", rc));
2585 return rc;
2586}
2587
2588/** @copydoc VBOXHDDBACKEND::pfnResize */
2589static DECLCALLBACK(int) qedResize(void *pBackendData, uint64_t cbSize,
2590 PCVDGEOMETRY pPCHSGeometry, PCVDGEOMETRY pLCHSGeometry,
2591 unsigned uPercentStart, unsigned uPercentSpan,
2592 PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
2593 PVDINTERFACE pVDIfsOperation)
2594{
2595 RT_NOREF7(pPCHSGeometry, pLCHSGeometry, uPercentStart, uPercentSpan, pVDIfsDisk, pVDIfsImage, pVDIfsOperation);
2596 PQEDIMAGE pImage = (PQEDIMAGE)pBackendData;
2597 int rc = VINF_SUCCESS;
2598
2599 /* Making the image smaller is not supported at the moment. */
2600 if (cbSize < pImage->cbSize)
2601 rc = VERR_NOT_SUPPORTED;
2602 else if (cbSize > pImage->cbSize)
2603 {
2604 /*
2605 * It is enough to just update the size field in the header to complete
2606 * growing. With the default cluster and table sizes the image can be expanded
2607 * to 64TB without overflowing the L1 and L2 tables making block relocation
2608 * superfluous.
2609 * @todo: The rare case where block relocation is still required (non default
2610 * table and/or cluster size or images with more than 64TB) is not
2611 * implemented yet and resizing such an image will fail with an error.
2612 */
2613 if (qedByte2Cluster(pImage, pImage->cbTable)*pImage->cTableEntries*pImage->cTableEntries*pImage->cbCluster < cbSize)
2614 rc = vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS,
2615 N_("Qed: Resizing the image '%s' is not supported because it would overflow the L1 and L2 table\n"),
2616 pImage->pszFilename);
2617 else
2618 {
2619 uint64_t cbSizeOld = pImage->cbSize;
2620
2621 pImage->cbSize = cbSize;
2622 rc = qedFlushImage(pImage);
2623 if (RT_FAILURE(rc))
2624 {
2625 pImage->cbSize = cbSizeOld; /* Restore */
2626
2627 rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("Qed: Resizing the image '%s' failed\n"),
2628 pImage->pszFilename);
2629 }
2630 }
2631 }
2632 /* Same size doesn't change the image at all. */
2633
2634 LogFlowFunc(("returns %Rrc\n", rc));
2635 return rc;
2636}
2637
2638
2639const VBOXHDDBACKEND g_QedBackend =
2640{
2641 /* pszBackendName */
2642 "QED",
2643 /* cbSize */
2644 sizeof(VBOXHDDBACKEND),
2645 /* uBackendCaps */
2646 VD_CAP_FILE | VD_CAP_VFS | VD_CAP_CREATE_DYNAMIC | VD_CAP_DIFF | VD_CAP_ASYNC,
2647 /* paFileExtensions */
2648 s_aQedFileExtensions,
2649 /* paConfigInfo */
2650 NULL,
2651 /* pfnCheckIfValid */
2652 qedCheckIfValid,
2653 /* pfnOpen */
2654 qedOpen,
2655 /* pfnCreate */
2656 qedCreate,
2657 /* pfnRename */
2658 qedRename,
2659 /* pfnClose */
2660 qedClose,
2661 /* pfnRead */
2662 qedRead,
2663 /* pfnWrite */
2664 qedWrite,
2665 /* pfnFlush */
2666 qedFlush,
2667 /* pfnDiscard */
2668 NULL,
2669 /* pfnGetVersion */
2670 qedGetVersion,
2671 /* pfnGetSectorSize */
2672 qedGetSectorSize,
2673 /* pfnGetSize */
2674 qedGetSize,
2675 /* pfnGetFileSize */
2676 qedGetFileSize,
2677 /* pfnGetPCHSGeometry */
2678 qedGetPCHSGeometry,
2679 /* pfnSetPCHSGeometry */
2680 qedSetPCHSGeometry,
2681 /* pfnGetLCHSGeometry */
2682 qedGetLCHSGeometry,
2683 /* pfnSetLCHSGeometry */
2684 qedSetLCHSGeometry,
2685 /* pfnGetImageFlags */
2686 qedGetImageFlags,
2687 /* pfnGetOpenFlags */
2688 qedGetOpenFlags,
2689 /* pfnSetOpenFlags */
2690 qedSetOpenFlags,
2691 /* pfnGetComment */
2692 qedGetComment,
2693 /* pfnSetComment */
2694 qedSetComment,
2695 /* pfnGetUuid */
2696 qedGetUuid,
2697 /* pfnSetUuid */
2698 qedSetUuid,
2699 /* pfnGetModificationUuid */
2700 qedGetModificationUuid,
2701 /* pfnSetModificationUuid */
2702 qedSetModificationUuid,
2703 /* pfnGetParentUuid */
2704 qedGetParentUuid,
2705 /* pfnSetParentUuid */
2706 qedSetParentUuid,
2707 /* pfnGetParentModificationUuid */
2708 qedGetParentModificationUuid,
2709 /* pfnSetParentModificationUuid */
2710 qedSetParentModificationUuid,
2711 /* pfnDump */
2712 qedDump,
2713 /* pfnGetTimestamp */
2714 NULL,
2715 /* pfnGetParentTimestamp */
2716 NULL,
2717 /* pfnSetParentTimestamp */
2718 NULL,
2719 /* pfnGetParentFilename */
2720 qedGetParentFilename,
2721 /* pfnSetParentFilename */
2722 qedSetParentFilename,
2723 /* pfnComposeLocation */
2724 genericFileComposeLocation,
2725 /* pfnComposeName */
2726 genericFileComposeName,
2727 /* pfnCompact */
2728 NULL,
2729 /* pfnResize */
2730 qedResize,
2731 /* pfnRepair */
2732 NULL,
2733 /* pfnTraverseMetadata */
2734 NULL
2735};
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette