VirtualBox

source: vbox/trunk/src/VBox/VMM/MM.cpp@ 26234

Last change on this file since 26234 was 26106, checked in by vboxsync, 15 years ago

MM.cpp: Register the reserved page counts with STAM.

  • Property svn:eol-style set to native
  • Property svn:keywords set to Id
File size: 28.8 KB
Line 
1/* $Id: MM.cpp 26106 2010-01-30 01:35:33Z vboxsync $ */
2/** @file
3 * MM - Memory Manager.
4 */
5
6/*
7 * Copyright (C) 2006-2007 Sun Microsystems, Inc.
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 *
17 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa
18 * Clara, CA 95054 USA or visit http://www.sun.com if you need
19 * additional information or have any questions.
20 */
21
22
23/** @page pg_mm MM - The Memory Manager
24 *
25 * The memory manager is in charge of the following memory:
26 * - Hypervisor Memory Area (HMA) - Address space management.
27 * - Hypervisor Heap - A memory heap that lives in all contexts.
28 * - Tagged ring-3 heap.
29 * - Page pools - Primarily used by PGM for shadow page tables.
30 * - Locked process memory - Guest RAM and other. (reduce/obsolete this)
31 * - Physical guest memory (RAM & ROM) - Moving to PGM. (obsolete this)
32 *
33 * The global memory manager (GMM) is the global counter part / partner of MM.
34 * MM will provide therefore ring-3 callable interfaces for some of the GMM APIs
35 * related to resource tracking (PGM is the user).
36 *
37 * @see grp_mm
38 *
39 *
40 * @section sec_mm_hma Hypervisor Memory Area
41 *
42 * The HMA is used when executing in raw-mode. We borrow, with the help of
43 * PGMMap, some unused space (one or more page directory entries to be precise)
44 * in the guest's virtual memory context. PGM will monitor the guest's virtual
45 * address space for changes and relocate the HMA when required.
46 *
47 * To give some idea what's in the HMA, study the 'info hma' output:
48 * @verbatim
49VBoxDbg> info hma
50Hypervisor Memory Area (HMA) Layout: Base 00000000a0000000, 0x00800000 bytes
5100000000a05cc000-00000000a05cd000 DYNAMIC fence
5200000000a05c4000-00000000a05cc000 DYNAMIC Dynamic mapping
5300000000a05c3000-00000000a05c4000 DYNAMIC fence
5400000000a05b8000-00000000a05c3000 DYNAMIC Paging
5500000000a05b6000-00000000a05b8000 MMIO2 0000000000000000 PCNetShMem
5600000000a0536000-00000000a05b6000 MMIO2 0000000000000000 VGA VRam
5700000000a0523000-00000000a0536000 00002aaab3d0c000 LOCKED autofree alloc once (PDM_DEVICE)
5800000000a0522000-00000000a0523000 DYNAMIC fence
5900000000a051e000-00000000a0522000 00002aaab36f5000 LOCKED autofree VBoxDD2GC.gc
6000000000a051d000-00000000a051e000 DYNAMIC fence
6100000000a04eb000-00000000a051d000 00002aaab36c3000 LOCKED autofree VBoxDDGC.gc
6200000000a04ea000-00000000a04eb000 DYNAMIC fence
6300000000a04e9000-00000000a04ea000 00002aaab36c2000 LOCKED autofree ram range (High ROM Region)
6400000000a04e8000-00000000a04e9000 DYNAMIC fence
6500000000a040e000-00000000a04e8000 00002aaab2e6d000 LOCKED autofree VMMGC.gc
6600000000a0208000-00000000a040e000 00002aaab2c67000 LOCKED autofree alloc once (PATM)
6700000000a01f7000-00000000a0208000 00002aaaab92d000 LOCKED autofree alloc once (SELM)
6800000000a01e7000-00000000a01f7000 00002aaaab5e8000 LOCKED autofree alloc once (SELM)
6900000000a01e6000-00000000a01e7000 DYNAMIC fence
7000000000a01e5000-00000000a01e6000 00002aaaab5e7000 HCPHYS 00000000c363c000 Core Code
7100000000a01e4000-00000000a01e5000 DYNAMIC fence
7200000000a01e3000-00000000a01e4000 00002aaaaab26000 HCPHYS 00000000619cf000 GIP
7300000000a01a2000-00000000a01e3000 00002aaaabf32000 LOCKED autofree alloc once (PGM_PHYS)
7400000000a016b000-00000000a01a2000 00002aaab233f000 LOCKED autofree alloc once (PGM_POOL)
7500000000a016a000-00000000a016b000 DYNAMIC fence
7600000000a0165000-00000000a016a000 DYNAMIC CR3 mapping
7700000000a0164000-00000000a0165000 DYNAMIC fence
7800000000a0024000-00000000a0164000 00002aaab215f000 LOCKED autofree Heap
7900000000a0023000-00000000a0024000 DYNAMIC fence
8000000000a0001000-00000000a0023000 00002aaab1d24000 LOCKED pages VM
8100000000a0000000-00000000a0001000 DYNAMIC fence
82 @endverbatim
83 *
84 *
85 * @section sec_mm_hyperheap Hypervisor Heap
86 *
87 * The heap is accessible from ring-3, ring-0 and the raw-mode context. That
88 * said, it's not necessarily mapped into ring-0 on if that's possible since we
89 * don't wish to waste kernel address space without a good reason.
90 *
91 * Allocations within the heap are always in the same relative position in all
92 * contexts, so, it's possible to use offset based linking. In fact, the heap is
93 * internally using offset based linked lists tracking heap blocks. We use
94 * offset linked AVL trees and lists in a lot of places where share structures
95 * between RC, R3 and R0, so this is a strict requirement of the heap. However
96 * this means that we cannot easily extend the heap since the extension won't
97 * necessarily be in the continuation of the current heap memory in all (or any)
98 * context.
99 *
100 * All allocations are tagged. Per tag allocation statistics will be maintaing
101 * and exposed thru STAM when VBOX_WITH_STATISTICS is defined.
102 *
103 *
104 * @section sec_mm_r3heap Tagged Ring-3 Heap
105 *
106 * The ring-3 heap is a wrapper around the RTMem API adding allocation
107 * statistics and automatic cleanup on VM destruction.
108 *
109 * Per tag allocation statistics will be maintaing and exposed thru STAM when
110 * VBOX_WITH_STATISTICS is defined.
111 *
112 *
113 * @section sec_mm_page Page Pool
114 *
115 * The MM manages a page pool from which other components can allocate locked,
116 * page aligned and page sized memory objects. The pool provides facilities to
117 * convert back and forth between (host) physical and virtual addresses (within
118 * the pool of course). Several specialized interfaces are provided for the most
119 * common alloctions and convertions to save the caller from bothersome casting
120 * and extra parameter passing.
121 *
122 *
123 * @section sec_mm_locked Locked Process Memory
124 *
125 * MM manages the locked process memory. This is used for a bunch of things
126 * (count the LOCKED entries in the'info hma' output found in @ref sec_mm_hma),
127 * but the main consumer of memory is currently for guest RAM. There is an
128 * ongoing rewrite that will move all the guest RAM allocation to PGM and
129 * GMM.
130 *
131 * The locking of memory is something doing in cooperation with the VirtualBox
132 * support driver, SUPDrv (aka. VBoxDrv), thru the support library API,
133 * SUPR3 (aka. SUPLib).
134 *
135 *
136 * @section sec_mm_phys Physical Guest Memory
137 *
138 * MM is currently managing the physical memory for the guest. It relies heavily
139 * on PGM for this. There is an ongoing rewrite that will move this to PGM. (The
140 * rewrite is driven by the need for more flexible guest ram allocation, but
141 * also motivated by the fact that MMPhys is just adding stupid bureaucracy and
142 * that MMR3PhysReserve is a totally weird artifact that must go away.)
143 *
144 */
145
146
147/*******************************************************************************
148* Header Files *
149*******************************************************************************/
150#define LOG_GROUP LOG_GROUP_MM
151#include <VBox/mm.h>
152#include <VBox/pgm.h>
153#include <VBox/cfgm.h>
154#include <VBox/ssm.h>
155#include <VBox/gmm.h>
156#include "MMInternal.h"
157#include <VBox/vm.h>
158#include <VBox/uvm.h>
159#include <VBox/err.h>
160#include <VBox/param.h>
161
162#include <VBox/log.h>
163#include <iprt/alloc.h>
164#include <iprt/assert.h>
165#include <iprt/string.h>
166
167
168/*******************************************************************************
169* Defined Constants And Macros *
170*******************************************************************************/
171/** The current saved state versino of MM. */
172#define MM_SAVED_STATE_VERSION 2
173
174
175/*******************************************************************************
176* Internal Functions *
177*******************************************************************************/
178static DECLCALLBACK(int) mmR3Save(PVM pVM, PSSMHANDLE pSSM);
179static DECLCALLBACK(int) mmR3Load(PVM pVM, PSSMHANDLE pSSM, uint32_t uVersion, uint32_t uPass);
180
181
182
183
184/**
185 * Initializes the MM members of the UVM.
186 *
187 * This is currently only the ring-3 heap.
188 *
189 * @returns VBox status code.
190 * @param pUVM Pointer to the user mode VM structure.
191 */
192VMMR3DECL(int) MMR3InitUVM(PUVM pUVM)
193{
194 /*
195 * Assert sizes and order.
196 */
197 AssertCompile(sizeof(pUVM->mm.s) <= sizeof(pUVM->mm.padding));
198 AssertRelease(sizeof(pUVM->mm.s) <= sizeof(pUVM->mm.padding));
199 Assert(!pUVM->mm.s.pHeap);
200
201 /*
202 * Init the heap.
203 */
204 int rc = mmR3HeapCreateU(pUVM, &pUVM->mm.s.pHeap);
205 if (RT_SUCCESS(rc))
206 {
207 rc = mmR3UkHeapCreateU(pUVM, &pUVM->mm.s.pUkHeap);
208 if (RT_SUCCESS(rc))
209 return VINF_SUCCESS;
210 mmR3HeapDestroy(pUVM->mm.s.pHeap);
211 pUVM->mm.s.pHeap = NULL;
212 }
213 return rc;
214}
215
216
217/**
218 * Initializes the MM.
219 *
220 * MM is managing the virtual address space (among other things) and
221 * setup the hypvervisor memory area mapping in the VM structure and
222 * the hypvervisor alloc-only-heap. Assuming the current init order
223 * and components the hypvervisor memory area looks like this:
224 * -# VM Structure.
225 * -# Hypervisor alloc only heap (also call Hypervisor memory region).
226 * -# Core code.
227 *
228 * MM determins the virtual address of the hypvervisor memory area by
229 * checking for location at previous run. If that property isn't available
230 * it will choose a default starting location, currently 0xa0000000.
231 *
232 * @returns VBox status code.
233 * @param pVM The VM to operate on.
234 */
235VMMR3DECL(int) MMR3Init(PVM pVM)
236{
237 LogFlow(("MMR3Init\n"));
238
239 /*
240 * Assert alignment, sizes and order.
241 */
242 AssertRelease(!(RT_OFFSETOF(VM, mm.s) & 31));
243 AssertRelease(sizeof(pVM->mm.s) <= sizeof(pVM->mm.padding));
244 AssertMsg(pVM->mm.s.offVM == 0, ("Already initialized!\n"));
245
246 /*
247 * Init the structure.
248 */
249 pVM->mm.s.offVM = RT_OFFSETOF(VM, mm);
250 pVM->mm.s.offLookupHyper = NIL_OFFSET;
251
252 /*
253 * Init the page pool.
254 */
255 int rc = mmR3PagePoolInit(pVM);
256 if (RT_SUCCESS(rc))
257 {
258 /*
259 * Init the hypervisor related stuff.
260 */
261 rc = mmR3HyperInit(pVM);
262 if (RT_SUCCESS(rc))
263 {
264 /*
265 * Register the saved state data unit.
266 */
267 rc = SSMR3RegisterInternal(pVM, "mm", 1, MM_SAVED_STATE_VERSION, sizeof(uint32_t) * 2,
268 NULL, NULL, NULL,
269 NULL, mmR3Save, NULL,
270 NULL, mmR3Load, NULL);
271 if (RT_SUCCESS(rc))
272 {
273 /*
274 * Statistics.
275 */
276 STAM_REG(pVM, &pVM->mm.s.cBasePages, STAMTYPE_U64, "/MM/Reserved/cBasePages", STAMUNIT_PAGES, "Reserved number of base pages, ROM and Shadow ROM included.");
277 STAM_REG(pVM, &pVM->mm.s.cHandyPages, STAMTYPE_U32, "/MM/Reserved/cHandyPages", STAMUNIT_PAGES, "Reserved number of handy pages.");
278 STAM_REG(pVM, &pVM->mm.s.cShadowPages, STAMTYPE_U32, "/MM/Reserved/cShadowPages", STAMUNIT_PAGES, "Reserved number of shadow paging pages.");
279 STAM_REG(pVM, &pVM->mm.s.cFixedPages, STAMTYPE_U32, "/MM/Reserved/cFixedPages", STAMUNIT_PAGES, "Reserved number of fixed pages (MMIO2).");
280 STAM_REG(pVM, &pVM->mm.s.cbRamBase, STAMTYPE_U64, "/MM/cbRamBase", STAMUNIT_BYTES, "Size of the base RAM.");
281
282 return rc;
283 }
284
285 /* .... failure .... */
286 }
287 }
288 MMR3Term(pVM);
289 return rc;
290}
291
292
293/**
294 * Initializes the MM parts which depends on PGM being initialized.
295 *
296 * @returns VBox status code.
297 * @param pVM The VM to operate on.
298 * @remark No cleanup necessary since MMR3Term() will be called on failure.
299 */
300VMMR3DECL(int) MMR3InitPaging(PVM pVM)
301{
302 LogFlow(("MMR3InitPaging:\n"));
303
304 /*
305 * Query the CFGM values.
306 */
307 int rc;
308 PCFGMNODE pMMCfg = CFGMR3GetChild(CFGMR3GetRoot(pVM), "MM");
309 if (!pMMCfg)
310 {
311 rc = CFGMR3InsertNode(CFGMR3GetRoot(pVM), "MM", &pMMCfg);
312 AssertRCReturn(rc, rc);
313 }
314
315 /** @cfgm{RamSize, uint64_t, 0, 16TB, 0}
316 * Specifies the size of the base RAM that is to be set up during
317 * VM initialization.
318 */
319 uint64_t cbRam;
320 rc = CFGMR3QueryU64(CFGMR3GetRoot(pVM), "RamSize", &cbRam);
321 if (rc == VERR_CFGM_VALUE_NOT_FOUND)
322 cbRam = 0;
323 else
324 AssertMsgRCReturn(rc, ("Configuration error: Failed to query integer \"RamSize\", rc=%Rrc.\n", rc), rc);
325 AssertLogRelMsg(!(cbRam & ~X86_PTE_PAE_PG_MASK), ("%RGp X86_PTE_PAE_PG_MASK=%RX64\n", cbRam, X86_PTE_PAE_PG_MASK));
326 AssertLogRelMsgReturn(cbRam <= GMM_GCPHYS_LAST, ("cbRam=%RGp GMM_GCPHYS_LAST=%RX64\n", cbRam, GMM_GCPHYS_LAST), VERR_OUT_OF_RANGE);
327 cbRam &= X86_PTE_PAE_PG_MASK;
328 pVM->mm.s.cbRamBase = cbRam;
329
330 /** @cfgm{RamHoleSize, uint32_t, 0, 4032MB, 512MB}
331 * Specifies the size of the memory hole. The memory hole is used
332 * to avoid mapping RAM to the range normally used for PCI memory regions.
333 * Must be aligned on a 4MB boundrary. */
334 uint32_t cbRamHole;
335 rc = CFGMR3QueryU32Def(CFGMR3GetRoot(pVM), "RamHoleSize", &cbRamHole, MM_RAM_HOLE_SIZE_DEFAULT);
336 AssertLogRelMsgRCReturn(rc, ("Configuration error: Failed to query integer \"RamHoleSize\", rc=%Rrc.\n", rc), rc);
337 AssertLogRelMsgReturn(cbRamHole <= 4032U * _1M,
338 ("Configuration error: \"RamHoleSize\"=%#RX32 is too large.\n", cbRamHole), VERR_OUT_OF_RANGE);
339 AssertLogRelMsgReturn(cbRamHole > 16 * _1M,
340 ("Configuration error: \"RamHoleSize\"=%#RX32 is too large.\n", cbRamHole), VERR_OUT_OF_RANGE);
341 AssertLogRelMsgReturn(!(cbRamHole & (_4M - 1)),
342 ("Configuration error: \"RamHoleSize\"=%#RX32 is misaligned.\n", cbRamHole), VERR_OUT_OF_RANGE);
343 uint64_t const offRamHole = _4G - cbRamHole;
344 if (cbRam < offRamHole)
345 Log(("MM: %RU64 bytes of RAM\n", cbRam));
346 else
347 Log(("MM: %RU64 bytes of RAM with a hole at %RU64 up to 4GB.\n", cbRam, offRamHole));
348
349 /** @cfgm{MM/Policy, string, no overcommitment}
350 * Specifies the policy to use when reserving memory for this VM. The recognized
351 * value is 'no overcommitment' (default). See GMMPOLICY.
352 */
353 GMMOCPOLICY enmOcPolicy;
354 char sz[64];
355 rc = CFGMR3QueryString(CFGMR3GetRoot(pVM), "Policy", sz, sizeof(sz));
356 if (RT_SUCCESS(rc))
357 {
358 if ( !RTStrICmp(sz, "no_oc")
359 || !RTStrICmp(sz, "no overcommitment"))
360 enmOcPolicy = GMMOCPOLICY_NO_OC;
361 else
362 return VMSetError(pVM, VERR_INVALID_PARAMETER, RT_SRC_POS, "Unknown \"MM/Policy\" value \"%s\"", sz);
363 }
364 else if (rc == VERR_CFGM_VALUE_NOT_FOUND)
365 enmOcPolicy = GMMOCPOLICY_NO_OC;
366 else
367 AssertMsgFailedReturn(("Configuration error: Failed to query string \"MM/Policy\", rc=%Rrc.\n", rc), rc);
368
369 /** @cfgm{MM/Priority, string, normal}
370 * Specifies the memory priority of this VM. The priority comes into play when the
371 * system is overcommitted and the VMs needs to be milked for memory. The recognized
372 * values are 'low', 'normal' (default) and 'high'. See GMMPRIORITY.
373 */
374 GMMPRIORITY enmPriority;
375 rc = CFGMR3QueryString(CFGMR3GetRoot(pVM), "Priority", sz, sizeof(sz));
376 if (RT_SUCCESS(rc))
377 {
378 if (!RTStrICmp(sz, "low"))
379 enmPriority = GMMPRIORITY_LOW;
380 else if (!RTStrICmp(sz, "normal"))
381 enmPriority = GMMPRIORITY_NORMAL;
382 else if (!RTStrICmp(sz, "high"))
383 enmPriority = GMMPRIORITY_HIGH;
384 else
385 return VMSetError(pVM, VERR_INVALID_PARAMETER, RT_SRC_POS, "Unknown \"MM/Priority\" value \"%s\"", sz);
386 }
387 else if (rc == VERR_CFGM_VALUE_NOT_FOUND)
388 enmPriority = GMMPRIORITY_NORMAL;
389 else
390 AssertMsgFailedReturn(("Configuration error: Failed to query string \"MM/Priority\", rc=%Rrc.\n", rc), rc);
391
392 /*
393 * Make the initial memory reservation with GMM.
394 */
395 uint64_t cBasePages = (cbRam >> PAGE_SHIFT) + pVM->mm.s.cBasePages;
396 rc = GMMR3InitialReservation(pVM,
397 RT_MAX(cBasePages + pVM->mm.s.cHandyPages, 1),
398 RT_MAX(pVM->mm.s.cShadowPages, 1),
399 RT_MAX(pVM->mm.s.cFixedPages, 1),
400 enmOcPolicy,
401 enmPriority);
402 if (RT_FAILURE(rc))
403 {
404 if (rc == VERR_GMM_MEMORY_RESERVATION_DECLINED)
405 return VMSetError(pVM, rc, RT_SRC_POS,
406 N_("Insufficient free memory to start the VM (cbRam=%#RX64 enmOcPolicy=%d enmPriority=%d)"),
407 cbRam, enmOcPolicy, enmPriority);
408 return VMSetError(pVM, rc, RT_SRC_POS, "GMMR3InitialReservation(,%#RX64,0,0,%d,%d)",
409 cbRam >> PAGE_SHIFT, enmOcPolicy, enmPriority);
410 }
411
412 /*
413 * If RamSize is 0 we're done now.
414 */
415 if (cbRam < PAGE_SIZE)
416 {
417 Log(("MM: No RAM configured\n"));
418 return VINF_SUCCESS;
419 }
420
421 /*
422 * Setup the base ram (PGM).
423 */
424 if (cbRam > offRamHole)
425 {
426 rc = PGMR3PhysRegisterRam(pVM, 0, offRamHole, "Base RAM");
427 if (RT_SUCCESS(rc))
428 rc = PGMR3PhysRegisterRam(pVM, _4G, cbRam - offRamHole, "Above 4GB Base RAM");
429 }
430 else
431 rc = PGMR3PhysRegisterRam(pVM, 0, RT_MIN(cbRam, offRamHole), "Base RAM");
432
433 /*
434 * Enabled mmR3UpdateReservation here since we don't want the
435 * PGMR3PhysRegisterRam calls above mess things up.
436 */
437 pVM->mm.s.fDoneMMR3InitPaging = true;
438 AssertMsg(pVM->mm.s.cBasePages == cBasePages || RT_FAILURE(rc), ("%RX64 != %RX64\n", pVM->mm.s.cBasePages, cBasePages));
439
440 LogFlow(("MMR3InitPaging: returns %Rrc\n", rc));
441 return rc;
442}
443
444
445/**
446 * Terminates the MM.
447 *
448 * Termination means cleaning up and freeing all resources,
449 * the VM it self is at this point powered off or suspended.
450 *
451 * @returns VBox status code.
452 * @param pVM The VM to operate on.
453 */
454VMMR3DECL(int) MMR3Term(PVM pVM)
455{
456 /*
457 * Destroy the page pool. (first as it used the hyper heap)
458 */
459 mmR3PagePoolTerm(pVM);
460
461 /* Clean up the hypervisor heap. */
462 mmR3HyperTerm(pVM);
463
464 /*
465 * Zero stuff to detect after termination use of the MM interface
466 */
467 pVM->mm.s.offLookupHyper = NIL_OFFSET;
468 pVM->mm.s.pHyperHeapR3 = NULL; /* freed above. */
469 pVM->mm.s.pHyperHeapR0 = NIL_RTR0PTR; /* freed above. */
470 pVM->mm.s.pHyperHeapRC = NIL_RTRCPTR; /* freed above. */
471 pVM->mm.s.offVM = 0; /* init assertion on this */
472
473 /*
474 * Destroy the User-kernel heap here since the support driver session
475 * may have been terminated by the time we get to MMR3TermUVM.
476 */
477 mmR3UkHeapDestroy(pVM->pUVM->mm.s.pUkHeap);
478 pVM->pUVM->mm.s.pUkHeap = NULL;
479
480 return VINF_SUCCESS;
481}
482
483
484/**
485 * Terminates the UVM part of MM.
486 *
487 * Termination means cleaning up and freeing all resources,
488 * the VM it self is at this point powered off or suspended.
489 *
490 * @returns VBox status code.
491 * @param pUVM Pointer to the user mode VM structure.
492 */
493VMMR3DECL(void) MMR3TermUVM(PUVM pUVM)
494{
495 /*
496 * Destroy the heaps.
497 */
498 if (pUVM->mm.s.pUkHeap)
499 {
500 mmR3UkHeapDestroy(pUVM->mm.s.pUkHeap);
501 pUVM->mm.s.pUkHeap = NULL;
502 }
503 mmR3HeapDestroy(pUVM->mm.s.pHeap);
504 pUVM->mm.s.pHeap = NULL;
505}
506
507
508/**
509 * Reset notification.
510 *
511 * @param pVM The VM handle.
512 */
513VMMR3DECL(void) MMR3Reset(PVM pVM)
514{
515 /* nothing to do anylonger. */
516}
517
518
519/**
520 * Execute state save operation.
521 *
522 * @returns VBox status code.
523 * @param pVM VM Handle.
524 * @param pSSM SSM operation handle.
525 */
526static DECLCALLBACK(int) mmR3Save(PVM pVM, PSSMHANDLE pSSM)
527{
528 LogFlow(("mmR3Save:\n"));
529
530 /* (PGM saves the physical memory.) */
531 SSMR3PutU64(pSSM, pVM->mm.s.cBasePages);
532 return SSMR3PutU64(pSSM, pVM->mm.s.cbRamBase);
533}
534
535
536/**
537 * Execute state load operation.
538 *
539 * @returns VBox status code.
540 * @param pVM VM Handle.
541 * @param pSSM SSM operation handle.
542 * @param uVersion Data layout version.
543 * @param uPass The data pass.
544 */
545static DECLCALLBACK(int) mmR3Load(PVM pVM, PSSMHANDLE pSSM, uint32_t uVersion, uint32_t uPass)
546{
547 LogFlow(("mmR3Load:\n"));
548 Assert(uPass == SSM_PASS_FINAL); NOREF(uPass);
549
550 /*
551 * Validate version.
552 */
553 if ( SSM_VERSION_MAJOR_CHANGED(uVersion, MM_SAVED_STATE_VERSION)
554 || !uVersion)
555 {
556 AssertMsgFailed(("mmR3Load: Invalid version uVersion=%d!\n", uVersion));
557 return VERR_SSM_UNSUPPORTED_DATA_UNIT_VERSION;
558 }
559
560 /*
561 * Check the cBasePages and cbRamBase values.
562 */
563 int rc;
564 RTUINT cb1;
565
566 /* cBasePages (ignored) */
567 uint64_t cPages;
568 if (uVersion >= 2)
569 rc = SSMR3GetU64(pSSM, &cPages);
570 else
571 {
572 rc = SSMR3GetUInt(pSSM, &cb1);
573 cPages = cb1 >> PAGE_SHIFT;
574 }
575 if (RT_FAILURE(rc))
576 return rc;
577
578 /* cbRamBase */
579 uint64_t cb;
580 if (uVersion != 1)
581 rc = SSMR3GetU64(pSSM, &cb);
582 else
583 {
584 rc = SSMR3GetUInt(pSSM, &cb1);
585 cb = cb1;
586 }
587 if (RT_FAILURE(rc))
588 return rc;
589 AssertLogRelMsgReturn(cb == pVM->mm.s.cbRamBase,
590 ("Memory configuration has changed. cbRamBase=%#RX64 save=%#RX64\n", pVM->mm.s.cbRamBase, cb),
591 VERR_SSM_LOAD_MEMORY_SIZE_MISMATCH);
592
593 /* (PGM restores the physical memory.) */
594 return rc;
595}
596
597
598/**
599 * Updates GMM with memory reservation changes.
600 *
601 * Called when MM::cbRamRegistered, MM::cShadowPages or MM::cFixedPages changes.
602 *
603 * @returns VBox status code - see GMMR0UpdateReservation.
604 * @param pVM The shared VM structure.
605 */
606int mmR3UpdateReservation(PVM pVM)
607{
608 VM_ASSERT_EMT(pVM);
609 if (pVM->mm.s.fDoneMMR3InitPaging)
610 return GMMR3UpdateReservation(pVM,
611 RT_MAX(pVM->mm.s.cBasePages + pVM->mm.s.cHandyPages, 1),
612 RT_MAX(pVM->mm.s.cShadowPages, 1),
613 RT_MAX(pVM->mm.s.cFixedPages, 1));
614 return VINF_SUCCESS;
615}
616
617
618/**
619 * Interface for PGM to increase the reservation of RAM and ROM pages.
620 *
621 * This can be called before MMR3InitPaging.
622 *
623 * @returns VBox status code. Will set VM error on failure.
624 * @param pVM The shared VM structure.
625 * @param cAddBasePages The number of pages to add.
626 */
627VMMR3DECL(int) MMR3IncreaseBaseReservation(PVM pVM, uint64_t cAddBasePages)
628{
629 uint64_t cOld = pVM->mm.s.cBasePages;
630 pVM->mm.s.cBasePages += cAddBasePages;
631 LogFlow(("MMR3IncreaseBaseReservation: +%RU64 (%RU64 -> %RU64\n", cAddBasePages, cOld, pVM->mm.s.cBasePages));
632 int rc = mmR3UpdateReservation(pVM);
633 if (RT_FAILURE(rc))
634 {
635 VMSetError(pVM, rc, RT_SRC_POS, N_("Failed to reserved physical memory for the RAM (%#RX64 -> %#RX64 + %#RX32)"),
636 cOld, pVM->mm.s.cBasePages, pVM->mm.s.cHandyPages);
637 pVM->mm.s.cBasePages = cOld;
638 }
639 return rc;
640}
641
642
643/**
644 * Interface for PGM to make reservations for handy pages in addition to the
645 * base memory.
646 *
647 * This can be called before MMR3InitPaging.
648 *
649 * @returns VBox status code. Will set VM error on failure.
650 * @param pVM The shared VM structure.
651 * @param cHandyPages The number of handy pages.
652 */
653VMMR3DECL(int) MMR3ReserveHandyPages(PVM pVM, uint32_t cHandyPages)
654{
655 AssertReturn(!pVM->mm.s.cHandyPages, VERR_WRONG_ORDER);
656
657 pVM->mm.s.cHandyPages = cHandyPages;
658 LogFlow(("MMR3ReserveHandyPages: %RU32 (base %RU64)\n", pVM->mm.s.cHandyPages, pVM->mm.s.cBasePages));
659 int rc = mmR3UpdateReservation(pVM);
660 if (RT_FAILURE(rc))
661 {
662 VMSetError(pVM, rc, RT_SRC_POS, N_("Failed to reserved physical memory for the RAM (%#RX64 + %#RX32)"),
663 pVM->mm.s.cBasePages, pVM->mm.s.cHandyPages);
664 pVM->mm.s.cHandyPages = 0;
665 }
666 return rc;
667}
668
669
670/**
671 * Interface for PGM to adjust the reservation of fixed pages.
672 *
673 * This can be called before MMR3InitPaging.
674 *
675 * @returns VBox status code. Will set VM error on failure.
676 * @param pVM The shared VM structure.
677 * @param cDeltaFixedPages The number of pages to add (positive) or subtract (negative).
678 * @param pszDesc Some description associated with the reservation.
679 */
680VMMR3DECL(int) MMR3AdjustFixedReservation(PVM pVM, int32_t cDeltaFixedPages, const char *pszDesc)
681{
682 const uint32_t cOld = pVM->mm.s.cFixedPages;
683 pVM->mm.s.cFixedPages += cDeltaFixedPages;
684 LogFlow(("MMR3AdjustFixedReservation: %d (%u -> %u)\n", cDeltaFixedPages, cOld, pVM->mm.s.cFixedPages));
685 int rc = mmR3UpdateReservation(pVM);
686 if (RT_FAILURE(rc))
687 {
688 VMSetError(pVM, rc, RT_SRC_POS, N_("Failed to reserve physical memory (%#x -> %#x; %s)"),
689 cOld, pVM->mm.s.cFixedPages, pszDesc);
690 pVM->mm.s.cFixedPages = cOld;
691 }
692 return rc;
693}
694
695
696/**
697 * Interface for PGM to update the reservation of shadow pages.
698 *
699 * This can be called before MMR3InitPaging.
700 *
701 * @returns VBox status code. Will set VM error on failure.
702 * @param pVM The shared VM structure.
703 * @param cShadowPages The new page count.
704 */
705VMMR3DECL(int) MMR3UpdateShadowReservation(PVM pVM, uint32_t cShadowPages)
706{
707 const uint32_t cOld = pVM->mm.s.cShadowPages;
708 pVM->mm.s.cShadowPages = cShadowPages;
709 LogFlow(("MMR3UpdateShadowReservation: %u -> %u\n", cOld, pVM->mm.s.cShadowPages));
710 int rc = mmR3UpdateReservation(pVM);
711 if (RT_FAILURE(rc))
712 {
713 VMSetError(pVM, rc, RT_SRC_POS, N_("Failed to reserve physical memory for shadow page tables (%#x -> %#x)"), cOld, pVM->mm.s.cShadowPages);
714 pVM->mm.s.cShadowPages = cOld;
715 }
716 return rc;
717}
718
719
720/**
721 * Convert HC Physical address to HC Virtual address.
722 *
723 * @returns VBox status.
724 * @param pVM VM handle.
725 * @param HCPhys The host context virtual address.
726 * @param ppv Where to store the resulting address.
727 * @thread The Emulation Thread.
728 *
729 * @remarks Avoid whenever possible.
730 * Intended for the debugger facility only.
731 * @todo Rename to indicate the special usage.
732 */
733VMMR3DECL(int) MMR3HCPhys2HCVirt(PVM pVM, RTHCPHYS HCPhys, void **ppv)
734{
735 /*
736 * Try page tables.
737 */
738 int rc = MMPagePhys2PageTry(pVM, HCPhys, ppv);
739 if (RT_SUCCESS(rc))
740 return rc;
741
742 /*
743 * Iterate thru the lookup records for HMA.
744 */
745 uint32_t off = HCPhys & PAGE_OFFSET_MASK;
746 HCPhys &= X86_PTE_PAE_PG_MASK;
747 PMMLOOKUPHYPER pCur = (PMMLOOKUPHYPER)((uint8_t *)pVM->mm.s.CTX_SUFF(pHyperHeap) + pVM->mm.s.offLookupHyper);
748 for (;;)
749 {
750 switch (pCur->enmType)
751 {
752 case MMLOOKUPHYPERTYPE_LOCKED:
753 {
754 PCRTHCPHYS paHCPhysPages = pCur->u.Locked.paHCPhysPages;
755 size_t iPage = pCur->cb >> PAGE_SHIFT;
756 while (iPage-- > 0)
757 if (paHCPhysPages[iPage] == HCPhys)
758 {
759 *ppv = (char *)pCur->u.Locked.pvR3 + (iPage << PAGE_SHIFT) + off;
760 return VINF_SUCCESS;
761 }
762 break;
763 }
764
765 case MMLOOKUPHYPERTYPE_HCPHYS:
766 if (pCur->u.HCPhys.HCPhys - HCPhys < pCur->cb)
767 {
768 *ppv = (uint8_t *)pCur->u.HCPhys.pvR3 + pCur->u.HCPhys.HCPhys - HCPhys + off;
769 return VINF_SUCCESS;
770 }
771 break;
772
773 case MMLOOKUPHYPERTYPE_GCPHYS: /* (for now we'll not allow these kind of conversions) */
774 case MMLOOKUPHYPERTYPE_MMIO2:
775 case MMLOOKUPHYPERTYPE_DYNAMIC:
776 break;
777
778 default:
779 AssertMsgFailed(("enmType=%d\n", pCur->enmType));
780 break;
781 }
782
783 /* next */
784 if (pCur->offNext == (int32_t)NIL_OFFSET)
785 break;
786 pCur = (PMMLOOKUPHYPER)((uint8_t *)pCur + pCur->offNext);
787 }
788 /* give up */
789 return VERR_INVALID_POINTER;
790}
791
792
793
794/**
795 * Get the size of the base RAM.
796 * This usually means the size of the first contigous block of physical memory.
797 *
798 * @returns The guest base RAM size.
799 * @param pVM The VM handle.
800 * @thread Any.
801 *
802 * @deprecated
803 */
804VMMR3DECL(uint64_t) MMR3PhysGetRamSize(PVM pVM)
805{
806 return pVM->mm.s.cbRamBase;
807}
808
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette