VirtualBox

source: vbox/trunk/src/VBox/VMM/VMMR3/EMHM.cpp@ 63495

Last change on this file since 63495 was 62637, checked in by vboxsync, 8 years ago

VMMR3: warnings

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 18.5 KB
Line 
1/* $Id: EMHM.cpp 62637 2016-07-28 17:12:17Z vboxsync $ */
2/** @file
3 * EM - Execution Monitor / Manager - hardware virtualization
4 */
5
6/*
7 * Copyright (C) 2006-2016 Oracle Corporation
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 */
17
18
19/*********************************************************************************************************************************
20* Header Files *
21*********************************************************************************************************************************/
22#define LOG_GROUP LOG_GROUP_EM
23#include <VBox/vmm/em.h>
24#include <VBox/vmm/vmm.h>
25#include <VBox/vmm/csam.h>
26#include <VBox/vmm/selm.h>
27#include <VBox/vmm/trpm.h>
28#include <VBox/vmm/iem.h>
29#include <VBox/vmm/iom.h>
30#include <VBox/vmm/dbgf.h>
31#include <VBox/vmm/pgm.h>
32#ifdef VBOX_WITH_REM
33# include <VBox/vmm/rem.h>
34#endif
35#include <VBox/vmm/tm.h>
36#include <VBox/vmm/mm.h>
37#include <VBox/vmm/ssm.h>
38#include <VBox/vmm/pdmapi.h>
39#include <VBox/vmm/pdmcritsect.h>
40#include <VBox/vmm/pdmqueue.h>
41#include <VBox/vmm/hm.h>
42#include "EMInternal.h"
43#include <VBox/vmm/vm.h>
44#include <VBox/vmm/gim.h>
45#include <VBox/vmm/cpumdis.h>
46#include <VBox/dis.h>
47#include <VBox/disopcode.h>
48#include <VBox/vmm/dbgf.h>
49#include "VMMTracing.h"
50
51#include <iprt/asm.h>
52
53
54/*********************************************************************************************************************************
55* Defined Constants And Macros *
56*********************************************************************************************************************************/
57#if 0 /* Disabled till after 2.1.0 when we've time to test it. */
58#define EM_NOTIFY_HM
59#endif
60
61
62/*********************************************************************************************************************************
63* Internal Functions *
64*********************************************************************************************************************************/
65DECLINLINE(int) emR3HmExecuteInstruction(PVM pVM, PVMCPU pVCpu, const char *pszPrefix, int rcGC = VINF_SUCCESS);
66static int emR3HmExecuteIOInstruction(PVM pVM, PVMCPU pVCpu);
67static int emR3HmForcedActions(PVM pVM, PVMCPU pVCpu, PCPUMCTX pCtx);
68
69#define EMHANDLERC_WITH_HM
70#define emR3ExecuteInstruction emR3HmExecuteInstruction
71#define emR3ExecuteIOInstruction emR3HmExecuteIOInstruction
72#include "EMHandleRCTmpl.h"
73
74
75/**
76 * Executes instruction in HM mode if we can.
77 *
78 * This is somewhat comparable to REMR3EmulateInstruction.
79 *
80 * @returns VBox strict status code.
81 * @retval VINF_EM_DBG_STEPPED on success.
82 * @retval VERR_EM_CANNOT_EXEC_GUEST if we cannot execute guest instructions in
83 * HM right now.
84 *
85 * @param pVM The cross context VM structure.
86 * @param pVCpu The cross context virtual CPU structure for the calling EMT.
87 * @param fFlags Combinations of EM_ONE_INS_FLAGS_XXX.
88 * @thread EMT.
89 */
90VMMR3_INT_DECL(VBOXSTRICTRC) EMR3HmSingleInstruction(PVM pVM, PVMCPU pVCpu, uint32_t fFlags)
91{
92 PCPUMCTX pCtx = pVCpu->em.s.pCtx;
93 Assert(!(fFlags & ~EM_ONE_INS_FLAGS_MASK));
94
95 if (!HMR3CanExecuteGuest(pVM, pCtx))
96 return VINF_EM_RESCHEDULE;
97
98 uint64_t const uOldRip = pCtx->rip;
99 for (;;)
100 {
101 /*
102 * Service necessary FFs before going into HM.
103 */
104 if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_PRE_RAW_MASK)
105 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_PRE_RAW_MASK))
106 {
107 VBOXSTRICTRC rcStrict = emR3HmForcedActions(pVM, pVCpu, pCtx);
108 if (rcStrict != VINF_SUCCESS)
109 {
110 Log(("EMR3HmSingleInstruction: FFs before -> %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
111 return rcStrict;
112 }
113 }
114
115 /*
116 * Go execute it.
117 */
118 bool fOld = HMSetSingleInstruction(pVM, pVCpu, true);
119 VBOXSTRICTRC rcStrict = VMMR3HmRunGC(pVM, pVCpu);
120 HMSetSingleInstruction(pVM, pVCpu, fOld);
121 LogFlow(("EMR3HmSingleInstruction: %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
122
123 /*
124 * Handle high priority FFs and informational status codes. We don't do
125 * normal FF processing the caller or the next call can deal with them.
126 */
127 VMCPU_FF_CLEAR(pVCpu, VMCPU_FF_RESUME_GUEST_MASK);
128 if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_POST_MASK)
129 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_POST_MASK))
130 {
131 rcStrict = emR3HighPriorityPostForcedActions(pVM, pVCpu, VBOXSTRICTRC_TODO(rcStrict));
132 LogFlow(("EMR3HmSingleInstruction: FFs after -> %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
133 }
134
135 if (rcStrict != VINF_SUCCESS && (rcStrict < VINF_EM_FIRST || rcStrict > VINF_EM_LAST))
136 {
137 rcStrict = emR3HmHandleRC(pVM, pVCpu, pCtx, VBOXSTRICTRC_TODO(rcStrict));
138 Log(("EMR3HmSingleInstruction: emR3HmHandleRC -> %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
139 }
140
141 /*
142 * Done?
143 */
144 if ( (rcStrict != VINF_SUCCESS && rcStrict != VINF_EM_DBG_STEPPED)
145 || !(fFlags & EM_ONE_INS_FLAGS_RIP_CHANGE)
146 || pCtx->rip != uOldRip)
147 {
148 if (rcStrict == VINF_SUCCESS && pCtx->rip != uOldRip)
149 rcStrict = VINF_EM_DBG_STEPPED;
150 Log(("EMR3HmSingleInstruction: returns %Rrc (rip %llx -> %llx)\n", VBOXSTRICTRC_VAL(rcStrict), uOldRip, pCtx->rip));
151 return rcStrict;
152 }
153 }
154}
155
156
157/**
158 * Executes one (or perhaps a few more) instruction(s).
159 *
160 * @returns VBox status code suitable for EM.
161 *
162 * @param pVM The cross context VM structure.
163 * @param pVCpu The cross context virtual CPU structure.
164 * @param rcRC Return code from RC.
165 * @param pszPrefix Disassembly prefix. If not NULL we'll disassemble the
166 * instruction and prefix the log output with this text.
167 */
168#if defined(LOG_ENABLED) || defined(DOXYGEN_RUNNING)
169static int emR3HmExecuteInstructionWorker(PVM pVM, PVMCPU pVCpu, int rcRC, const char *pszPrefix)
170#else
171static int emR3HmExecuteInstructionWorker(PVM pVM, PVMCPU pVCpu, int rcRC)
172#endif
173{
174#ifdef LOG_ENABLED
175 PCPUMCTX pCtx = pVCpu->em.s.pCtx;
176#endif
177 int rc;
178 NOREF(rcRC);
179
180#ifdef LOG_ENABLED
181 /*
182 * Log it.
183 */
184 Log(("EMINS: %04x:%RGv RSP=%RGv\n", pCtx->cs.Sel, (RTGCPTR)pCtx->rip, (RTGCPTR)pCtx->rsp));
185 if (pszPrefix)
186 {
187 DBGFR3_INFO_LOG(pVM, pVCpu, "cpumguest", pszPrefix);
188 DBGFR3_DISAS_INSTR_CUR_LOG(pVCpu, pszPrefix);
189 }
190#endif
191
192 /*
193 * Use IEM and fallback on REM if the functionality is missing.
194 * Once IEM gets mature enough, nothing should ever fall back.
195 */
196 STAM_PROFILE_START(&pVCpu->em.s.StatIEMEmu, a);
197 rc = VBOXSTRICTRC_TODO(IEMExecOne(pVCpu));
198 STAM_PROFILE_STOP(&pVCpu->em.s.StatIEMEmu, a);
199
200 if ( rc == VERR_IEM_ASPECT_NOT_IMPLEMENTED
201 || rc == VERR_IEM_INSTR_NOT_IMPLEMENTED)
202 {
203#ifdef VBOX_WITH_REM
204 STAM_PROFILE_START(&pVCpu->em.s.StatREMEmu, b);
205 EMRemLock(pVM);
206 /* Flush the recompiler TLB if the VCPU has changed. */
207 if (pVM->em.s.idLastRemCpu != pVCpu->idCpu)
208 CPUMSetChangedFlags(pVCpu, CPUM_CHANGED_ALL);
209 pVM->em.s.idLastRemCpu = pVCpu->idCpu;
210
211 rc = REMR3EmulateInstruction(pVM, pVCpu);
212 EMRemUnlock(pVM);
213 STAM_PROFILE_STOP(&pVCpu->em.s.StatREMEmu, b);
214#else /* !VBOX_WITH_REM */
215 NOREF(pVM);
216#endif /* !VBOX_WITH_REM */
217 }
218
219#ifdef EM_NOTIFY_HM
220 if (pVCpu->em.s.enmState == EMSTATE_DEBUG_GUEST_HM)
221 HMR3NotifyEmulated(pVCpu);
222#endif
223 return rc;
224}
225
226
227/**
228 * Executes one (or perhaps a few more) instruction(s).
229 * This is just a wrapper for discarding pszPrefix in non-logging builds.
230 *
231 * @returns VBox status code suitable for EM.
232 * @param pVM The cross context VM structure.
233 * @param pVCpu The cross context virtual CPU structure.
234 * @param pszPrefix Disassembly prefix. If not NULL we'll disassemble the
235 * instruction and prefix the log output with this text.
236 * @param rcGC GC return code
237 */
238DECLINLINE(int) emR3HmExecuteInstruction(PVM pVM, PVMCPU pVCpu, const char *pszPrefix, int rcGC)
239{
240#ifdef LOG_ENABLED
241 return emR3HmExecuteInstructionWorker(pVM, pVCpu, rcGC, pszPrefix);
242#else
243 RT_NOREF_PV(pszPrefix);
244 return emR3HmExecuteInstructionWorker(pVM, pVCpu, rcGC);
245#endif
246}
247
248/**
249 * Executes one (or perhaps a few more) IO instruction(s).
250 *
251 * @returns VBox status code suitable for EM.
252 * @param pVM The cross context VM structure.
253 * @param pVCpu The cross context virtual CPU structure.
254 */
255static int emR3HmExecuteIOInstruction(PVM pVM, PVMCPU pVCpu)
256{
257 PCPUMCTX pCtx = pVCpu->em.s.pCtx;
258
259 STAM_PROFILE_START(&pVCpu->em.s.StatIOEmu, a);
260
261 /*
262 * Try to restart the io instruction that was refused in ring-0.
263 */
264 VBOXSTRICTRC rcStrict = HMR3RestartPendingIOInstr(pVM, pVCpu, pCtx);
265 if (IOM_SUCCESS(rcStrict))
266 {
267 STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatIoRestarted);
268 STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
269 return VBOXSTRICTRC_TODO(rcStrict); /* rip already updated. */
270 }
271 AssertMsgReturn(rcStrict == VERR_NOT_FOUND, ("%Rrc\n", VBOXSTRICTRC_VAL(rcStrict)),
272 RT_SUCCESS_NP(rcStrict) ? VERR_IPE_UNEXPECTED_INFO_STATUS : VBOXSTRICTRC_TODO(rcStrict));
273
274 /*
275 * Hand it over to the interpreter.
276 */
277 rcStrict = IEMExecOne(pVCpu);
278 LogFlow(("emR3HmExecuteIOInstruction: %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
279 STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatIoIem);
280 STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
281 return VBOXSTRICTRC_TODO(rcStrict);
282}
283
284
285/**
286 * Process raw-mode specific forced actions.
287 *
288 * This function is called when any FFs in the VM_FF_HIGH_PRIORITY_PRE_RAW_MASK is pending.
289 *
290 * @returns VBox status code. May return VINF_EM_NO_MEMORY but none of the other
291 * EM statuses.
292 * @param pVM The cross context VM structure.
293 * @param pVCpu The cross context virtual CPU structure.
294 * @param pCtx Pointer to the guest CPU context.
295 */
296static int emR3HmForcedActions(PVM pVM, PVMCPU pVCpu, PCPUMCTX pCtx)
297{
298 /*
299 * Sync page directory.
300 */
301 if (VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_PGM_SYNC_CR3 | VMCPU_FF_PGM_SYNC_CR3_NON_GLOBAL))
302 {
303 Assert(pVCpu->em.s.enmState != EMSTATE_WAIT_SIPI);
304 int rc = PGMSyncCR3(pVCpu, pCtx->cr0, pCtx->cr3, pCtx->cr4, VMCPU_FF_IS_SET(pVCpu, VMCPU_FF_PGM_SYNC_CR3));
305 if (RT_FAILURE(rc))
306 return rc;
307
308#ifdef VBOX_WITH_RAW_MODE
309 Assert(!VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT));
310#endif
311
312 /* Prefetch pages for EIP and ESP. */
313 /** @todo This is rather expensive. Should investigate if it really helps at all. */
314 rc = PGMPrefetchPage(pVCpu, SELMToFlat(pVM, DISSELREG_CS, CPUMCTX2CORE(pCtx), pCtx->rip));
315 if (rc == VINF_SUCCESS)
316 rc = PGMPrefetchPage(pVCpu, SELMToFlat(pVM, DISSELREG_SS, CPUMCTX2CORE(pCtx), pCtx->rsp));
317 if (rc != VINF_SUCCESS)
318 {
319 if (rc != VINF_PGM_SYNC_CR3)
320 {
321 AssertLogRelMsgReturn(RT_FAILURE(rc), ("%Rrc\n", rc), VERR_IPE_UNEXPECTED_INFO_STATUS);
322 return rc;
323 }
324 rc = PGMSyncCR3(pVCpu, pCtx->cr0, pCtx->cr3, pCtx->cr4, VMCPU_FF_IS_SET(pVCpu, VMCPU_FF_PGM_SYNC_CR3));
325 if (RT_FAILURE(rc))
326 return rc;
327 }
328 /** @todo maybe prefetch the supervisor stack page as well */
329#ifdef VBOX_WITH_RAW_MODE
330 Assert(!VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT));
331#endif
332 }
333
334 /*
335 * Allocate handy pages (just in case the above actions have consumed some pages).
336 */
337 if (VM_FF_IS_PENDING_EXCEPT(pVM, VM_FF_PGM_NEED_HANDY_PAGES, VM_FF_PGM_NO_MEMORY))
338 {
339 int rc = PGMR3PhysAllocateHandyPages(pVM);
340 if (RT_FAILURE(rc))
341 return rc;
342 }
343
344 /*
345 * Check whether we're out of memory now.
346 *
347 * This may stem from some of the above actions or operations that has been executed
348 * since we ran FFs. The allocate handy pages must for instance always be followed by
349 * this check.
350 */
351 if (VM_FF_IS_PENDING(pVM, VM_FF_PGM_NO_MEMORY))
352 return VINF_EM_NO_MEMORY;
353
354 return VINF_SUCCESS;
355}
356
357
358/**
359 * Executes hardware accelerated raw code. (Intel VT-x & AMD-V)
360 *
361 * This function contains the raw-mode version of the inner
362 * execution loop (the outer loop being in EMR3ExecuteVM()).
363 *
364 * @returns VBox status code. The most important ones are: VINF_EM_RESCHEDULE, VINF_EM_RESCHEDULE_RAW,
365 * VINF_EM_RESCHEDULE_REM, VINF_EM_SUSPEND, VINF_EM_RESET and VINF_EM_TERMINATE.
366 *
367 * @param pVM The cross context VM structure.
368 * @param pVCpu The cross context virtual CPU structure.
369 * @param pfFFDone Where to store an indicator telling whether or not
370 * FFs were done before returning.
371 */
372int emR3HmExecute(PVM pVM, PVMCPU pVCpu, bool *pfFFDone)
373{
374 int rc = VERR_IPE_UNINITIALIZED_STATUS;
375 PCPUMCTX pCtx = pVCpu->em.s.pCtx;
376
377 LogFlow(("emR3HmExecute%d: (cs:eip=%04x:%RGv)\n", pVCpu->idCpu, pCtx->cs.Sel, (RTGCPTR)pCtx->rip));
378 *pfFFDone = false;
379
380 STAM_COUNTER_INC(&pVCpu->em.s.StatHmExecuteEntry);
381
382#ifdef EM_NOTIFY_HM
383 HMR3NotifyScheduled(pVCpu);
384#endif
385
386 /*
387 * Spin till we get a forced action which returns anything but VINF_SUCCESS.
388 */
389 for (;;)
390 {
391 STAM_PROFILE_ADV_START(&pVCpu->em.s.StatHmEntry, a);
392
393 /* Check if a forced reschedule is pending. */
394 if (HMR3IsRescheduleRequired(pVM, pCtx))
395 {
396 rc = VINF_EM_RESCHEDULE;
397 break;
398 }
399
400 /*
401 * Process high priority pre-execution raw-mode FFs.
402 */
403#ifdef VBOX_WITH_RAW_MODE
404 Assert(!VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_SELM_SYNC_TSS | VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT));
405#endif
406 if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_PRE_RAW_MASK)
407 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_PRE_RAW_MASK))
408 {
409 rc = emR3HmForcedActions(pVM, pVCpu, pCtx);
410 if (rc != VINF_SUCCESS)
411 break;
412 }
413
414#ifdef LOG_ENABLED
415 /*
416 * Log important stuff before entering GC.
417 */
418 if (TRPMHasTrap(pVCpu))
419 Log(("CPU%d: Pending hardware interrupt=0x%x cs:rip=%04X:%RGv\n", pVCpu->idCpu, TRPMGetTrapNo(pVCpu), pCtx->cs.Sel, (RTGCPTR)pCtx->rip));
420
421 uint32_t cpl = CPUMGetGuestCPL(pVCpu);
422
423 if (pVM->cCpus == 1)
424 {
425 if (pCtx->eflags.Bits.u1VM)
426 Log(("HWV86: %08X IF=%d\n", pCtx->eip, pCtx->eflags.Bits.u1IF));
427 else if (CPUMIsGuestIn64BitCodeEx(pCtx))
428 Log(("HWR%d: %04X:%RGv ESP=%RGv IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pCtx->cs.Sel, (RTGCPTR)pCtx->rip, pCtx->rsp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
429 else
430 Log(("HWR%d: %04X:%08X ESP=%08X IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pCtx->cs.Sel, pCtx->eip, pCtx->esp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
431 }
432 else
433 {
434 if (pCtx->eflags.Bits.u1VM)
435 Log(("HWV86-CPU%d: %08X IF=%d\n", pVCpu->idCpu, pCtx->eip, pCtx->eflags.Bits.u1IF));
436 else if (CPUMIsGuestIn64BitCodeEx(pCtx))
437 Log(("HWR%d-CPU%d: %04X:%RGv ESP=%RGv IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pVCpu->idCpu, pCtx->cs.Sel, (RTGCPTR)pCtx->rip, pCtx->rsp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
438 else
439 Log(("HWR%d-CPU%d: %04X:%08X ESP=%08X IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pVCpu->idCpu, pCtx->cs.Sel, pCtx->eip, pCtx->esp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
440 }
441#endif /* LOG_ENABLED */
442
443 /*
444 * Execute the code.
445 */
446 STAM_PROFILE_ADV_STOP(&pVCpu->em.s.StatHmEntry, a);
447
448 if (RT_LIKELY(emR3IsExecutionAllowed(pVM, pVCpu)))
449 {
450 STAM_PROFILE_START(&pVCpu->em.s.StatHmExec, x);
451 rc = VMMR3HmRunGC(pVM, pVCpu);
452 STAM_PROFILE_STOP(&pVCpu->em.s.StatHmExec, x);
453 }
454 else
455 {
456 /* Give up this time slice; virtual time continues */
457 STAM_REL_PROFILE_ADV_START(&pVCpu->em.s.StatCapped, u);
458 RTThreadSleep(5);
459 STAM_REL_PROFILE_ADV_STOP(&pVCpu->em.s.StatCapped, u);
460 rc = VINF_SUCCESS;
461 }
462
463
464 /*
465 * Deal with high priority post execution FFs before doing anything else.
466 */
467 VMCPU_FF_CLEAR(pVCpu, VMCPU_FF_RESUME_GUEST_MASK);
468 if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_POST_MASK)
469 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_POST_MASK))
470 rc = emR3HighPriorityPostForcedActions(pVM, pVCpu, rc);
471
472 /*
473 * Process the returned status code.
474 */
475 if (rc >= VINF_EM_FIRST && rc <= VINF_EM_LAST)
476 break;
477
478 rc = emR3HmHandleRC(pVM, pVCpu, pCtx, rc);
479 if (rc != VINF_SUCCESS)
480 break;
481
482 /*
483 * Check and execute forced actions.
484 */
485#ifdef VBOX_HIGH_RES_TIMERS_HACK
486 TMTimerPollVoid(pVM, pVCpu);
487#endif
488 if ( VM_FF_IS_PENDING(pVM, VM_FF_ALL_MASK)
489 || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_ALL_MASK))
490 {
491 rc = emR3ForcedActions(pVM, pVCpu, rc);
492 VBOXVMM_EM_FF_ALL_RET(pVCpu, rc);
493 if ( rc != VINF_SUCCESS
494 && rc != VINF_EM_RESCHEDULE_HM)
495 {
496 *pfFFDone = true;
497 break;
498 }
499 }
500 }
501
502 /*
503 * Return to outer loop.
504 */
505#if defined(LOG_ENABLED) && defined(DEBUG)
506 RTLogFlush(NULL);
507#endif
508 return rc;
509}
510
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette