/* $Id: HWACCMAll.cpp 26685 2010-02-22 17:48:23Z vboxsync $ */ /** @file * HWACCM - All contexts. */ /* * Copyright (C) 2006-2007 Sun Microsystems, Inc. * * This file is part of VirtualBox Open Source Edition (OSE), as * available from http://www.virtualbox.org. This file is free software; * you can redistribute it and/or modify it under the terms of the GNU * General Public License (GPL) as published by the Free Software * Foundation, in version 2 as it comes in the "COPYING" file of the * VirtualBox OSE distribution. VirtualBox OSE is distributed in the * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind. * * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa * Clara, CA 95054 USA or visit http://www.sun.com if you need * additional information or have any questions. */ /******************************************************************************* * Header Files * *******************************************************************************/ #define LOG_GROUP LOG_GROUP_HWACCM #include #include #include "HWACCMInternal.h" #include #include #include #include #include #include #include #include #include #include /** * Queues a page for invalidation * * @returns VBox status code. * @param pVCpu The VMCPU to operate on. * @param GCVirt Page to invalidate */ void hwaccmQueueInvlPage(PVMCPU pVCpu, RTGCPTR GCVirt) { /* Nothing to do if a TLB flush is already pending */ if (VMCPU_FF_ISSET(pVCpu, VMCPU_FF_TLB_FLUSH)) return; #if 1 VMCPU_FF_SET(pVCpu, VMCPU_FF_TLB_FLUSH); #else Be very careful when activating this code! if (iPage == RT_ELEMENTS(pVCpu->hwaccm.s.TlbShootdown.aPages)) VMCPU_FF_SET(pVCpu, VMCPU_FF_TLB_FLUSH); else VMCPU_FF_SET(pVCpu, VMCPU_FF_TLB_SHOOTDOWN); #endif } /** * Invalidates a guest page * * @returns VBox status code. * @param pVCpu The VMCPU to operate on. * @param GCVirt Page to invalidate */ VMMDECL(int) HWACCMInvalidatePage(PVMCPU pVCpu, RTGCPTR GCVirt) { STAM_COUNTER_INC(&pVCpu->hwaccm.s.StatFlushPageManual); #ifdef IN_RING0 PVM pVM = pVCpu->CTX_SUFF(pVM); if (pVM->hwaccm.s.vmx.fSupported) return VMXR0InvalidatePage(pVM, pVCpu, GCVirt); Assert(pVM->hwaccm.s.svm.fSupported); return SVMR0InvalidatePage(pVM, pVCpu, GCVirt); #endif hwaccmQueueInvlPage(pVCpu, GCVirt); return VINF_SUCCESS; } /** * Flushes the guest TLB * * @returns VBox status code. * @param pVCpu The VMCPU to operate on. */ VMMDECL(int) HWACCMFlushTLB(PVMCPU pVCpu) { LogFlow(("HWACCMFlushTLB\n")); VMCPU_FF_SET(pVCpu, VMCPU_FF_TLB_FLUSH); STAM_COUNTER_INC(&pVCpu->hwaccm.s.StatFlushTLBManual); return VINF_SUCCESS; } #ifdef IN_RING0 /** * Dummy RTMpOnSpecific handler since RTMpPokeCpu couldn't be used. * */ static DECLCALLBACK(void) hwaccmFlushHandler(RTCPUID idCpu, void *pvUser1, void *pvUser2) { return; } /** * Wrapper for RTMpPokeCpu to deal with VERR_NOT_SUPPORTED * */ void hwaccmMpPokeCpu(PVMCPU pVCpu, RTCPUID idHostCpu) { uint32_t cWorldSwitchExit = pVCpu->hwaccm.s.cWorldSwitchExit; STAM_PROFILE_ADV_START(&pVCpu->hwaccm.s.StatPoke, x); int rc = RTMpPokeCpu(idHostCpu); STAM_PROFILE_ADV_STOP(&pVCpu->hwaccm.s.StatPoke, x); /* Not implemented on some platforms (Darwin, Linux kernel < 2.6.19); fall back to a less efficient implementation (broadcast). */ if (rc == VERR_NOT_SUPPORTED) { STAM_PROFILE_ADV_START(&pVCpu->hwaccm.s.StatSpinPoke, z); /* synchronous. */ RTMpOnSpecific(idHostCpu, hwaccmFlushHandler, 0, 0); STAM_PROFILE_ADV_STOP(&pVCpu->hwaccm.s.StatSpinPoke, z); } else { if (rc == VINF_SUCCESS) STAM_PROFILE_ADV_START(&pVCpu->hwaccm.s.StatSpinPoke, z); else STAM_PROFILE_ADV_START(&pVCpu->hwaccm.s.StatSpinPokeFailed, z); /* Spin until the VCPU has switched back. */ while ( pVCpu->hwaccm.s.fCheckedTLBFlush && cWorldSwitchExit == pVCpu->hwaccm.s.cWorldSwitchExit) { ASMNopPause(); } if (rc == VINF_SUCCESS) STAM_PROFILE_ADV_STOP(&pVCpu->hwaccm.s.StatSpinPoke, z); else STAM_PROFILE_ADV_STOP(&pVCpu->hwaccm.s.StatSpinPokeFailed, z); } } #endif #ifndef IN_RC /** * Invalidates a guest page on all VCPUs. * * @returns VBox status code. * @param pVM The VM to operate on. * @param GCVirt Page to invalidate */ VMMDECL(int) HWACCMInvalidatePageOnAllVCpus(PVM pVM, RTGCPTR GCPtr) { VMCPUID idCurCpu = VMMGetCpuId(pVM); STAM_COUNTER_INC(&pVM->aCpus[idCurCpu].hwaccm.s.StatFlushPage); for (VMCPUID idCpu = 0; idCpu < pVM->cCpus; idCpu++) { PVMCPU pVCpu = &pVM->aCpus[idCpu]; /* Nothing to do if a TLB flush is already pending; the VCPU should have already been poked if it were active */ if (VMCPU_FF_ISSET(pVCpu, VMCPU_FF_TLB_FLUSH)) continue; if (pVCpu->idCpu == idCurCpu) { HWACCMInvalidatePage(pVCpu, GCPtr); } else { hwaccmQueueInvlPage(pVCpu, GCPtr); if (pVCpu->hwaccm.s.fCheckedTLBFlush) { STAM_COUNTER_INC(&pVCpu->hwaccm.s.StatTlbShootdown); #ifdef IN_RING0 RTCPUID idHostCpu = pVCpu->hwaccm.s.idEnteredCpu; if (idHostCpu != NIL_RTCPUID) hwaccmMpPokeCpu(pVCpu, idHostCpu); #else VMR3NotifyCpuFFU(pVCpu->pUVCpu, VMNOTIFYFF_FLAGS_POKE); #endif } else STAM_COUNTER_INC(&pVCpu->hwaccm.s.StatFlushPageManual); } } return VINF_SUCCESS; } /** * Flush the TLBs of all VCPUs * * @returns VBox status code. * @param pVM The VM to operate on. */ VMMDECL(int) HWACCMFlushTLBOnAllVCpus(PVM pVM) { if (pVM->cCpus == 1) return HWACCMFlushTLB(&pVM->aCpus[0]); VMCPUID idThisCpu = VMMGetCpuId(pVM); STAM_COUNTER_INC(&pVM->aCpus[idThisCpu].hwaccm.s.StatFlushTLB); for (VMCPUID idCpu = 0; idCpu < pVM->cCpus; idCpu++) { PVMCPU pVCpu = &pVM->aCpus[idCpu]; /* Nothing to do if a TLB flush is already pending; the VCPU should have already been poked if it were active */ if (VMCPU_FF_ISSET(pVCpu, VMCPU_FF_TLB_FLUSH)) continue; VMCPU_FF_SET(pVCpu, VMCPU_FF_TLB_FLUSH); if (idThisCpu == idCpu) continue; if (pVCpu->hwaccm.s.fCheckedTLBFlush) { STAM_COUNTER_INC(&pVCpu->hwaccm.s.StatTlbShootdownFlush); #ifdef IN_RING0 RTCPUID idHostCpu = pVCpu->hwaccm.s.idEnteredCpu; if (idHostCpu != NIL_RTCPUID) hwaccmMpPokeCpu(pVCpu, idHostCpu); #else VMR3NotifyCpuFFU(pVCpu->pUVCpu, VMNOTIFYFF_FLAGS_POKE); #endif } else STAM_COUNTER_INC(&pVCpu->hwaccm.s.StatFlushTLBManual); } return VINF_SUCCESS; } #endif /** * Checks if nested paging is enabled * * @returns boolean * @param pVM The VM to operate on. */ VMMDECL(bool) HWACCMIsNestedPagingActive(PVM pVM) { return HWACCMIsEnabled(pVM) && pVM->hwaccm.s.fNestedPaging; } /** * Return the shadow paging mode for nested paging/ept * * @returns shadow paging mode * @param pVM The VM to operate on. */ VMMDECL(PGMMODE) HWACCMGetShwPagingMode(PVM pVM) { Assert(HWACCMIsNestedPagingActive(pVM)); if (pVM->hwaccm.s.svm.fSupported) return PGMMODE_NESTED; Assert(pVM->hwaccm.s.vmx.fSupported); return PGMMODE_EPT; } /** * Invalidates a guest page by physical address * * NOTE: Assumes the current instruction references this physical page though a virtual address!! * * @returns VBox status code. * @param pVM The VM to operate on. * @param GCPhys Page to invalidate */ VMMDECL(int) HWACCMInvalidatePhysPage(PVM pVM, RTGCPHYS GCPhys) { if (!HWACCMIsNestedPagingActive(pVM)) return VINF_SUCCESS; #ifdef IN_RING0 if (pVM->hwaccm.s.vmx.fSupported) { VMCPUID idThisCpu = VMMGetCpuId(pVM); for (VMCPUID idCpu = 0; idCpu < pVM->cCpus; idCpu++) { PVMCPU pVCpu = &pVM->aCpus[idCpu]; if (idThisCpu == idCpu) { VMXR0InvalidatePhysPage(pVM, pVCpu, GCPhys); continue; } VMCPU_FF_SET(pVCpu, VMCPU_FF_TLB_FLUSH); if (pVCpu->hwaccm.s.fCheckedTLBFlush) { STAM_COUNTER_INC(&pVCpu->hwaccm.s.StatTlbShootdownFlush); # ifdef IN_RING0 RTCPUID idHostCpu = pVCpu->hwaccm.s.idEnteredCpu; if (idHostCpu != NIL_RTCPUID) hwaccmMpPokeCpu(pVCpu, idHostCpu); # else VMR3NotifyCpuFFU(pVCpu->pUVCpu, VMNOTIFYFF_FLAGS_POKE); # endif } else STAM_COUNTER_INC(&pVCpu->hwaccm.s.StatFlushTLBManual); } return VINF_SUCCESS; } Assert(pVM->hwaccm.s.svm.fSupported); /* AMD-V doesn't support invalidation with guest physical addresses; see comment in SVMR0InvalidatePhysPage. */ HWACCMFlushTLBOnAllVCpus(pVM); #else HWACCMFlushTLBOnAllVCpus(pVM); #endif return VINF_SUCCESS; } /** * Checks if an interrupt event is currently pending. * * @returns Interrupt event pending state. * @param pVM The VM to operate on. */ VMMDECL(bool) HWACCMHasPendingIrq(PVM pVM) { PVMCPU pVCpu = VMMGetCpu(pVM); return !!pVCpu->hwaccm.s.Event.fPending; }