/** @file * GMM - The Global Memory Manager. (VMM) */ /* * Copyright (C) 2007 Oracle Corporation * * This file is part of VirtualBox Open Source Edition (OSE), as * available from http://www.virtualbox.org. This file is free software; * you can redistribute it and/or modify it under the terms of the GNU * General Public License (GPL) as published by the Free Software * Foundation, in version 2 as it comes in the "COPYING" file of the * VirtualBox OSE distribution. VirtualBox OSE is distributed in the * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind. * * The contents of this file may alternatively be used under the terms * of the Common Development and Distribution License Version 1.0 * (CDDL) only, as it comes in the "COPYING.CDDL" file of the * VirtualBox OSE distribution, in which case the provisions of the * CDDL are applicable instead of those of the GPL. * * You may elect to license modified versions of this file under the * terms and conditions of either the GPL or the CDDL or both. */ #ifndef ___VBox_gmm_h #define ___VBox_gmm_h #include #include #include #include /* for VMMDEVSHAREDREGIONDESC */ RT_C_DECLS_BEGIN /** @defgroup grp_gmm GMM - The Global Memory Manager * @{ */ /** @def IN_GMM_R0 * Used to indicate whether we're inside the same link module as the ring 0 * part of the Global Memory Manager or not. */ #ifdef DOXYGEN_RUNNING # define IN_GMM_R0 #endif /** @def GMMR0DECL * Ring 0 GMM export or import declaration. * @param type The return type of the function declaration. */ #ifdef IN_GMM_R0 # define GMMR0DECL(type) DECLEXPORT(type) VBOXCALL #else # define GMMR0DECL(type) DECLIMPORT(type) VBOXCALL #endif /** @def IN_GMM_R3 * Used to indicate whether we're inside the same link module as the ring 3 * part of the Global Memory Manager or not. */ #ifdef DOXYGEN_RUNNING # define IN_GMM_R3 #endif /** @def GMMR3DECL * Ring 3 GMM export or import declaration. * @param type The return type of the function declaration. */ #ifdef IN_GMM_R3 # define GMMR3DECL(type) DECLEXPORT(type) VBOXCALL #else # define GMMR3DECL(type) DECLIMPORT(type) VBOXCALL #endif /** The chunk shift. (2^21 = 2 MB) */ #define GMM_CHUNK_SHIFT 21 /** The allocation chunk size. */ #define GMM_CHUNK_SIZE (1U << GMM_CHUNK_SHIFT) /** The allocation chunk size in pages. */ #define GMM_CHUNK_NUM_PAGES (1U << (GMM_CHUNK_SHIFT - PAGE_SHIFT)) /** The shift factor for converting a page id into a chunk id. */ #define GMM_CHUNKID_SHIFT (GMM_CHUNK_SHIFT - PAGE_SHIFT) /** The last valid Chunk ID value. */ #define GMM_CHUNKID_LAST (GMM_PAGEID_LAST >> GMM_CHUNKID_SHIFT) /** The last valid Page ID value. * The current limit is 2^28 - 1, or almost 1TB if you like. * The constraints are currently dictated by PGMPAGE. */ #define GMM_PAGEID_LAST (RT_BIT_32(28) - 1) /** Mask out the page index from the Page ID. */ #define GMM_PAGEID_IDX_MASK ((1U << GMM_CHUNKID_SHIFT) - 1) /** The NIL Chunk ID value. */ #define NIL_GMM_CHUNKID 0 /** The NIL Page ID value. */ #define NIL_GMM_PAGEID 0 #if 0 /* wrong - these are guest page pfns and not page ids! */ /** Special Page ID used by unassigned pages. */ #define GMM_PAGEID_UNASSIGNED 0x0fffffffU /** Special Page ID used by unsharable pages. * Like MMIO2, shadow and heap. This is for later, obviously. */ #define GMM_PAGEID_UNSHARABLE 0x0ffffffeU /** The end of the valid Page IDs. This is the first special one. */ #define GMM_PAGEID_END 0x0ffffff0U #endif /** @def GMM_GCPHYS_LAST * The last of the valid guest physical address as it applies to GMM pages. * * This must reflect the constraints imposed by the RTGCPHYS type and * the guest page frame number used internally in GMMPAGE. * * @note Note this corresponds to GMM_PAGE_PFN_LAST. */ #if HC_ARCH_BITS == 64 # define GMM_GCPHYS_LAST UINT64_C(0x00000fffffff0000) /* 2^44 (16TB) - 0x10000 */ #else # define GMM_GCPHYS_LAST UINT64_C(0x0000000fffff0000) /* 2^36 (64GB) - 0x10000 */ #endif /** * Over-commitment policy. */ typedef enum GMMOCPOLICY { /** The usual invalid 0 value. */ GMMOCPOLICY_INVALID = 0, /** No over-commitment, fully backed. * The GMM guarantees that it will be able to allocate all of the * guest RAM for a VM with OC policy. */ GMMOCPOLICY_NO_OC, /** to-be-determined. */ GMMOCPOLICY_TBD, /** The end of the valid policy range. */ GMMOCPOLICY_END, /** The usual 32-bit hack. */ GMMOCPOLICY_32BIT_HACK = 0x7fffffff } GMMOCPOLICY; /** * VM / Memory priority. */ typedef enum GMMPRIORITY { /** The usual invalid 0 value. */ GMMPRIORITY_INVALID = 0, /** High. * When ballooning, ask these VMs last. * When running out of memory, try not to interrupt these VMs. */ GMMPRIORITY_HIGH, /** Normal. * When ballooning, don't wait to ask these. * When running out of memory, pause, save and/or kill these VMs. */ GMMPRIORITY_NORMAL, /** Low. * When ballooning, maximize these first. * When running out of memory, save or kill these VMs. */ GMMPRIORITY_LOW, /** The end of the valid priority range. */ GMMPRIORITY_END, /** The custom 32-bit type blowup. */ GMMPRIORITY_32BIT_HACK = 0x7fffffff } GMMPRIORITY; /** * GMM Memory Accounts. */ typedef enum GMMACCOUNT { /** The customary invalid zero entry. */ GMMACCOUNT_INVALID = 0, /** Account with the base allocations. */ GMMACCOUNT_BASE, /** Account with the shadow allocations. */ GMMACCOUNT_SHADOW, /** Account with the fixed allocations. */ GMMACCOUNT_FIXED, /** The end of the valid values. */ GMMACCOUNT_END, /** The usual 32-bit value to finish it off. */ GMMACCOUNT_32BIT_HACK = 0x7fffffff } GMMACCOUNT; /** * Balloon actions. */ typedef enum { /** Invalid zero entry. */ GMMBALLOONACTION_INVALID = 0, /** Inflate the balloon. */ GMMBALLOONACTION_INFLATE, /** Deflate the balloon. */ GMMBALLOONACTION_DEFLATE, /** Puncture the balloon because of VM reset. */ GMMBALLOONACTION_RESET, /** End of the valid actions. */ GMMBALLOONACTION_END, /** hack forcing the size of the enum to 32-bits. */ GMMBALLOONACTION_MAKE_32BIT_HACK = 0x7fffffff } GMMBALLOONACTION; /** * A page descriptor for use when freeing pages. * See GMMR0FreePages, GMMR0BalloonedPages. */ typedef struct GMMFREEPAGEDESC { /** The Page ID of the page to be freed. */ uint32_t idPage; } GMMFREEPAGEDESC; /** Pointer to a page descriptor for freeing pages. */ typedef GMMFREEPAGEDESC *PGMMFREEPAGEDESC; /** * A page descriptor for use when updating and allocating pages. * * This is a bit complicated because we want to do as much as possible * with the same structure. */ typedef struct GMMPAGEDESC { /** The physical address of the page. * * @input GMMR0AllocateHandyPages expects the guest physical address * to update the GMMPAGE structure with. Pass GMM_GCPHYS_UNSHAREABLE * when appropriate and NIL_RTHCPHYS when the page wasn't used * for any specific guest address. * * GMMR0AllocatePage expects the guest physical address to put in * the GMMPAGE structure for the page it allocates for this entry. * Pass NIL_RTHCPHYS and GMM_GCPHYS_UNSHAREABLE as above. * * @output The host physical address of the allocated page. * NIL_RTHCPHYS on allocation failure. * * ASSUMES: sizeof(RTHCPHYS) >= sizeof(RTGCPHYS). */ RTHCPHYS HCPhysGCPhys; /** The Page ID. * * @intput GMMR0AllocateHandyPages expects the Page ID of the page to * update here. NIL_GMM_PAGEID means no page should be updated. * * GMMR0AllocatePages requires this to be initialized to * NIL_GMM_PAGEID currently. * * @output The ID of the page, NIL_GMM_PAGEID if the allocation failed. */ uint32_t idPage; /** The Page ID of the shared page was replaced by this page. * * @input GMMR0AllocateHandyPages expects this to indicate a shared * page that has been replaced by this page and should have its * reference counter decremented and perhaps be freed up. Use * NIL_GMM_PAGEID if no shared page was involved. * * All other APIs expects NIL_GMM_PAGEID here. * * @output All APIs sets this to NIL_GMM_PAGEID. */ uint32_t idSharedPage; } GMMPAGEDESC; AssertCompileSize(GMMPAGEDESC, 16); /** Pointer to a page allocation. */ typedef GMMPAGEDESC *PGMMPAGEDESC; /** GMMPAGEDESC::HCPhysGCPhys value that indicates that the page is unsharable. * @note This corresponds to GMM_PAGE_PFN_UNSHAREABLE. */ #if HC_ARCH_BITS == 64 # define GMM_GCPHYS_UNSHAREABLE UINT64_C(0x00000fffffff1000) #else # define GMM_GCPHYS_UNSHAREABLE UINT64_C(0x0000000fffff1000) #endif GMMR0DECL(int) GMMR0Init(void); GMMR0DECL(void) GMMR0Term(void); GMMR0DECL(void) GMMR0InitPerVMData(PGVM pGVM); GMMR0DECL(void) GMMR0CleanupVM(PGVM pGVM); GMMR0DECL(int) GMMR0InitialReservation(PVM pVM, VMCPUID idCpu, uint64_t cBasePages, uint32_t cShadowPages, uint32_t cFixedPages, GMMOCPOLICY enmPolicy, GMMPRIORITY enmPriority); GMMR0DECL(int) GMMR0UpdateReservation(PVM pVM, VMCPUID idCpu, uint64_t cBasePages, uint32_t cShadowPages, uint32_t cFixedPages); GMMR0DECL(int) GMMR0AllocateHandyPages(PVM pVM, VMCPUID idCpu, uint32_t cPagesToUpdate, uint32_t cPagesToAlloc, PGMMPAGEDESC paPages); GMMR0DECL(int) GMMR0AllocatePages(PVM pVM, VMCPUID idCpu, uint32_t cPages, PGMMPAGEDESC paPages, GMMACCOUNT enmAccount); GMMR0DECL(int) GMMR0AllocateLargePage(PVM pVM, VMCPUID idCpu, uint32_t cbPage, uint32_t *pIdPage, RTHCPHYS *pHCPhys); GMMR0DECL(int) GMMR0FreePages(PVM pVM, VMCPUID idCpu, uint32_t cPages, PGMMFREEPAGEDESC paPages, GMMACCOUNT enmAccount); GMMR0DECL(int) GMMR0FreeLargePage(PVM pVM, VMCPUID idCpu, uint32_t idPage); GMMR0DECL(int) GMMR0BalloonedPages(PVM pVM, VMCPUID idCpu, GMMBALLOONACTION enmAction, uint32_t cBalloonedPages); GMMR0DECL(int) GMMR0MapUnmapChunk(PVM pVM, VMCPUID idCpu, uint32_t idChunkMap, uint32_t idChunkUnmap, PRTR3PTR ppvR3); GMMR0DECL(int) GMMR0SeedChunk(PVM pVM, VMCPUID idCpu, RTR3PTR pvR3); GMMR0DECL(int) GMMR0RegisterSharedModule(PVM pVM, VMCPUID idCpu, char *pszModuleName, char *pszVersion, RTGCPTR GCBaseAddr, uint32_t cbModule, unsigned cRegions, VMMDEVSHAREDREGIONDESC *pRegions); GMMR0DECL(int) GMMR0UnregisterSharedModule(PVM pVM, VMCPUID idCpu, char *pszModuleName, char *pszVersion, RTGCPTR GCBaseAddr, uint32_t cbModule); GMMR0DECL(int) GMMR0UnregisterAllSharedModules(PVM pVM, VMCPUID idCpu); GMMR0DECL(int) GMMR0ResetSharedModules(PVM pVM, VMCPUID idCpu); /** * Request buffer for GMMR0InitialReservationReq / VMMR0_DO_GMM_INITIAL_RESERVATION. * @see GMMR0InitialReservation */ typedef struct GMMINITIALRESERVATIONREQ { /** The header. */ SUPVMMR0REQHDR Hdr; uint64_t cBasePages; /**< @see GMMR0InitialReservation */ uint32_t cShadowPages; /**< @see GMMR0InitialReservation */ uint32_t cFixedPages; /**< @see GMMR0InitialReservation */ GMMOCPOLICY enmPolicy; /**< @see GMMR0InitialReservation */ GMMPRIORITY enmPriority; /**< @see GMMR0InitialReservation */ } GMMINITIALRESERVATIONREQ; /** Pointer to a GMMR0InitialReservationReq / VMMR0_DO_GMM_INITIAL_RESERVATION request buffer. */ typedef GMMINITIALRESERVATIONREQ *PGMMINITIALRESERVATIONREQ; GMMR0DECL(int) GMMR0InitialReservationReq(PVM pVM, VMCPUID idCpu, PGMMINITIALRESERVATIONREQ pReq); /** * Request buffer for GMMR0UpdateReservationReq / VMMR0_DO_GMM_UPDATE_RESERVATION. * @see GMMR0UpdateReservation */ typedef struct GMMUPDATERESERVATIONREQ { /** The header. */ SUPVMMR0REQHDR Hdr; uint64_t cBasePages; /**< @see GMMR0UpdateReservation */ uint32_t cShadowPages; /**< @see GMMR0UpdateReservation */ uint32_t cFixedPages; /**< @see GMMR0UpdateReservation */ } GMMUPDATERESERVATIONREQ; /** Pointer to a GMMR0InitialReservationReq / VMMR0_DO_GMM_INITIAL_RESERVATION request buffer. */ typedef GMMUPDATERESERVATIONREQ *PGMMUPDATERESERVATIONREQ; GMMR0DECL(int) GMMR0UpdateReservationReq(PVM pVM, VMCPUID idCpu, PGMMUPDATERESERVATIONREQ pReq); /** * Request buffer for GMMR0AllocatePagesReq / VMMR0_DO_GMM_ALLOCATE_PAGES. * @see GMMR0AllocatePages. */ typedef struct GMMALLOCATEPAGESREQ { /** The header. */ SUPVMMR0REQHDR Hdr; /** The account to charge the allocation to. */ GMMACCOUNT enmAccount; /** The number of pages to allocate. */ uint32_t cPages; /** Array of page descriptors. */ GMMPAGEDESC aPages[1]; } GMMALLOCATEPAGESREQ; /** Pointer to a GMMR0AllocatePagesReq / VMMR0_DO_GMM_ALLOCATE_PAGES request buffer. */ typedef GMMALLOCATEPAGESREQ *PGMMALLOCATEPAGESREQ; GMMR0DECL(int) GMMR0AllocatePagesReq(PVM pVM, VMCPUID idCpu, PGMMALLOCATEPAGESREQ pReq); /** * Request buffer for GMMR0FreePagesReq / VMMR0_DO_GMM_FREE_PAGES. * @see GMMR0FreePages. */ typedef struct GMMFREEPAGESREQ { /** The header. */ SUPVMMR0REQHDR Hdr; /** The account this relates to. */ GMMACCOUNT enmAccount; /** The number of pages to free. */ uint32_t cPages; /** Array of free page descriptors. */ GMMFREEPAGEDESC aPages[1]; } GMMFREEPAGESREQ; /** Pointer to a GMMR0FreePagesReq / VMMR0_DO_GMM_FREE_PAGES request buffer. */ typedef GMMFREEPAGESREQ *PGMMFREEPAGESREQ; GMMR0DECL(int) GMMR0FreePagesReq(PVM pVM, VMCPUID idCpu, PGMMFREEPAGESREQ pReq); /** * Request buffer for GMMR0BalloonedPagesReq / VMMR0_DO_GMM_BALLOONED_PAGES. * @see GMMR0BalloonedPages. */ typedef struct GMMBALLOONEDPAGESREQ { /** The header. */ SUPVMMR0REQHDR Hdr; /** The number of ballooned pages. */ uint32_t cBalloonedPages; /** Inflate or deflate the balloon. */ GMMBALLOONACTION enmAction; } GMMBALLOONEDPAGESREQ; /** Pointer to a GMMR0BalloonedPagesReq / VMMR0_DO_GMM_BALLOONED_PAGES request buffer. */ typedef GMMBALLOONEDPAGESREQ *PGMMBALLOONEDPAGESREQ; GMMR0DECL(int) GMMR0BalloonedPagesReq(PVM pVM, VMCPUID idCpu, PGMMBALLOONEDPAGESREQ pReq); /** * Request buffer for GMMR0QueryHypervisorMemoryStatsReq / VMMR0_DO_GMM_QUERY_VMM_MEM_STATS. * @see GMMR0QueryHypervisorMemoryStatsReq. */ typedef struct GMMMEMSTATSREQ { /** The header. */ SUPVMMR0REQHDR Hdr; /** The number of allocated pages (out). */ uint64_t cAllocPages; /** The number of free pages (out). */ uint64_t cFreePages; /** The number of ballooned pages (out). */ uint64_t cBalloonedPages; /** Maximum nr of pages (out). */ uint64_t cMaxPages; } GMMMEMSTATSREQ; /** Pointer to a GMMR0QueryHypervisorMemoryStatsReq / VMMR0_DO_GMM_QUERY_HYPERVISOR_MEM_STATS request buffer. */ typedef GMMMEMSTATSREQ *PGMMMEMSTATSREQ; GMMR0DECL(int) GMMR0QueryHypervisorMemoryStatsReq(PVM pVM, PGMMMEMSTATSREQ pReq); GMMR0DECL(int) GMMR0QueryMemoryStatsReq(PVM pVM, VMCPUID idCpu, PGMMMEMSTATSREQ pReq); /** * Request buffer for GMMR0MapUnmapChunkReq / VMMR0_DO_GMM_MAP_UNMAP_CHUNK. * @see GMMR0MapUnmapChunk */ typedef struct GMMMAPUNMAPCHUNKREQ { /** The header. */ SUPVMMR0REQHDR Hdr; /** The chunk to map, NIL_GMM_CHUNKID if unmap only. (IN) */ uint32_t idChunkMap; /** The chunk to unmap, NIL_GMM_CHUNKID if map only. (IN) */ uint32_t idChunkUnmap; /** Where the mapping address is returned. (OUT) */ RTR3PTR pvR3; } GMMMAPUNMAPCHUNKREQ; /** Pointer to a GMMR0MapUnmapChunkReq / VMMR0_DO_GMM_MAP_UNMAP_CHUNK request buffer. */ typedef GMMMAPUNMAPCHUNKREQ *PGMMMAPUNMAPCHUNKREQ; GMMR0DECL(int) GMMR0MapUnmapChunkReq(PVM pVM, VMCPUID idCpu, PGMMMAPUNMAPCHUNKREQ pReq); /** * Request buffer for GMMR0FreeLargePageReq / VMMR0_DO_GMM_FREE_LARGE_PAGE. * @see GMMR0FreeLargePage. */ typedef struct GMMFREELARGEPAGEREQ { /** The header. */ SUPVMMR0REQHDR Hdr; /** The Page ID. */ uint32_t idPage; } GMMFREELARGEPAGEREQ; /** Pointer to a GMMR0FreePagesReq / VMMR0_DO_GMM_FREE_PAGES request buffer. */ typedef GMMFREELARGEPAGEREQ *PGMMFREELARGEPAGEREQ; GMMR0DECL(int) GMMR0FreeLargePageReq(PVM pVM, VMCPUID idCpu, PGMMFREELARGEPAGEREQ pReq); /** Maximum length of the shared module name string. */ #define GMM_SHARED_MODULE_MAX_NAME_STRING 128 /** Maximum length of the shared module version string. */ #define GMM_SHARED_MODULE_MAX_VERSION_STRING 16 /** * Request buffer for GMMR0RegisterSharedModuleReq / VMMR0_DO_GMM_REGISTER_SHARED_MODULE. * @see GMMR0RegisterSharedModule. */ typedef struct GMMREGISTERSHAREDMODULEREQ { /** The header. */ SUPVMMR0REQHDR Hdr; /** Shared module size. */ uint32_t cbModule; /** Number of included region descriptors */ uint32_t cRegions; /** Base address of the shared module. */ RTGCPTR64 GCBaseAddr; /** Module name */ char szName[GMM_SHARED_MODULE_MAX_NAME_STRING]; /** Module version */ char szVersion[GMM_SHARED_MODULE_MAX_VERSION_STRING]; /** Shared region descriptor(s). */ VMMDEVSHAREDREGIONDESC aRegions[1]; } GMMREGISTERSHAREDMODULEREQ; /** Pointer to a GMMR0RegisterSharedModuleReq / VMMR0_DO_GMM_REGISTER_SHARED_MODULE request buffer. */ typedef GMMREGISTERSHAREDMODULEREQ *PGMMREGISTERSHAREDMODULEREQ; GMMR0DECL(int) GMMR0RegisterSharedModuleReq(PVM pVM, VMCPUID idCpu, PGMMREGISTERSHAREDMODULEREQ pReq); /** * Page descriptor for GMMR0SharedModuleCheckRange */ typedef struct GMMSHAREDPAGEDESC { /** HC Physical address (in/out) */ RTHCPHYS HCPhys; /** GC Physical address (in) */ RTGCPHYS GCPhys; /** GMM page id. (in/out) */ uint32_t uHCPhysPageId; /** Align at 8 byte boundary. */ uint32_t uAlignment; } GMMSHAREDPAGEDESC; /** Pointer to a GMMSHAREDPAGEDESC. */ typedef GMMSHAREDPAGEDESC *PGMMSHAREDPAGEDESC; GMMR0DECL(int) GMMR0SharedModuleCheckRange(PVM pVM, VMCPUID idCpu, PGMMREGISTERSHAREDMODULEREQ pReq, unsigned idxRegion, unsigned cPages, PGMMSHAREDPAGEDESC paPageDesc); /** * Request buffer for GMMR0UnregisterSharedModuleReq / VMMR0_DO_GMM_UNREGISTER_SHARED_MODULE. * @see GMMR0UnregisterSharedModule. */ typedef struct GMMUNREGISTERSHAREDMODULEREQ { /** The header. */ SUPVMMR0REQHDR Hdr; /** Shared module size. */ uint32_t cbModule; /** Align at 8 byte boundary. */ uint32_t u32Alignment; /** Base address of the shared module. */ RTGCPTR64 GCBaseAddr; /** Module name */ char szName[GMM_SHARED_MODULE_MAX_NAME_STRING]; /** Module version */ char szVersion[GMM_SHARED_MODULE_MAX_VERSION_STRING]; } GMMUNREGISTERSHAREDMODULEREQ; /** Pointer to a GMMR0UnregisterSharedModuleReq / VMMR0_DO_GMM_UNREGISTER_SHARED_MODULE request buffer. */ typedef GMMUNREGISTERSHAREDMODULEREQ *PGMMUNREGISTERSHAREDMODULEREQ; GMMR0DECL(int) GMMR0UnregisterSharedModuleReq(PVM pVM, VMCPUID idCpu, PGMMUNREGISTERSHAREDMODULEREQ pReq); #ifdef IN_RING3 /** @defgroup grp_gmm_r3 The Global Memory Manager Ring-3 API Wrappers * @ingroup grp_gmm * @{ */ GMMR3DECL(int) GMMR3InitialReservation(PVM pVM, uint64_t cBasePages, uint32_t cShadowPages, uint32_t cFixedPages, GMMOCPOLICY enmPolicy, GMMPRIORITY enmPriority); GMMR3DECL(int) GMMR3UpdateReservation(PVM pVM, uint64_t cBasePages, uint32_t cShadowPages, uint32_t cFixedPages); GMMR3DECL(int) GMMR3AllocatePagesPrepare(PVM pVM, PGMMALLOCATEPAGESREQ *ppReq, uint32_t cPages, GMMACCOUNT enmAccount); GMMR3DECL(int) GMMR3AllocatePagesPerform(PVM pVM, PGMMALLOCATEPAGESREQ pReq); GMMR3DECL(void) GMMR3AllocatePagesCleanup(PGMMALLOCATEPAGESREQ pReq); GMMR3DECL(int) GMMR3FreePagesPrepare(PVM pVM, PGMMFREEPAGESREQ *ppReq, uint32_t cPages, GMMACCOUNT enmAccount); GMMR3DECL(void) GMMR3FreePagesRePrep(PVM pVM, PGMMFREEPAGESREQ pReq, uint32_t cPages, GMMACCOUNT enmAccount); GMMR3DECL(int) GMMR3FreePagesPerform(PVM pVM, PGMMFREEPAGESREQ pReq, uint32_t cActualPages); GMMR3DECL(void) GMMR3FreePagesCleanup(PGMMFREEPAGESREQ pReq); GMMR3DECL(void) GMMR3FreeAllocatedPages(PVM pVM, GMMALLOCATEPAGESREQ const *pAllocReq); GMMR3DECL(int) GMMR3AllocateLargePage(PVM pVM, uint32_t cbPage); GMMR3DECL(int) GMMR3FreeLargePage(PVM pVM, uint32_t idPage); GMMR3DECL(int) GMMR3MapUnmapChunk(PVM pVM, uint32_t idChunkMap, uint32_t idChunkUnmap, PRTR3PTR ppvR3); GMMR3DECL(int) GMMR3SeedChunk(PVM pVM, RTR3PTR pvR3); GMMR3DECL(int) GMMR3QueryHypervisorMemoryStats(PVM pVM, uint64_t *pcTotalAllocPages, uint64_t *pcTotalFreePages, uint64_t *pcTotalBalloonPages); GMMR3DECL(int) GMMR3QueryMemoryStats(PVM pVM, uint64_t *pcAllocPages, uint64_t *pcMaxPages, uint64_t *pcBalloonPages); GMMR3DECL(int) GMMR3BalloonedPages(PVM pVM, GMMBALLOONACTION enmAction, uint32_t cBalloonedPages); GMMR3DECL(int) GMMR3RegisterSharedModule(PVM pVM, PGMMREGISTERSHAREDMODULEREQ pReq); GMMR3DECL(int) GMMR3UnregisterSharedModule(PVM pVM, PGMMREGISTERSHAREDMODULEREQ pReq); GMMR3DECL(int) GMMR3ResetSharedModules(PVM pVM); /** @} */ #endif /* IN_RING3 */ /** @} */ RT_C_DECLS_END #endif