VirtualBox

source: vbox/trunk/src/VBox/VMM/VMMR0/VMMR0.cpp@ 672

Last change on this file since 672 was 672, checked in by vboxsync, 18 years ago

#pragma intrinsic can't be in a function, added missing ().

  • Property svn:eol-style set to native
  • Property svn:keywords set to Id
File size: 24.4 KB
Line 
1/* $Id: VMMR0.cpp 672 2007-02-06 04:24:06Z vboxsync $ */
2/** @file
3 * VMM - Host Context Ring 0.
4 */
5
6/*
7 * Copyright (C) 2006 InnoTek Systemberatung GmbH
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License as published by the Free Software Foundation,
13 * in version 2 as it comes in the "COPYING" file of the VirtualBox OSE
14 * distribution. VirtualBox OSE is distributed in the hope that it will
15 * be useful, but WITHOUT ANY WARRANTY of any kind.
16 *
17 * If you received this file as part of a commercial VirtualBox
18 * distribution, then only the terms of your commercial VirtualBox
19 * license agreement apply instead of the previous paragraph.
20 */
21
22
23/*******************************************************************************
24* Header Files *
25*******************************************************************************/
26#define LOG_GROUP LOG_GROUP_VMM
27#ifdef __AMD64__ /** @todo fix logging on __AMD64__ (swapgs) */
28# define LOG_DISABLED
29#endif
30#include <VBox/vmm.h>
31#include <VBox/sup.h>
32#include <VBox/trpm.h>
33#include <VBox/cpum.h>
34#include <VBox/stam.h>
35#include <VBox/tm.h>
36#include "VMMInternal.h"
37#include <VBox/vm.h>
38#include <VBox/intnet.h>
39#include <VBox/hwaccm.h>
40
41#include <VBox/err.h>
42#include <VBox/version.h>
43#include <VBox/log.h>
44#include <iprt/assert.h>
45
46#if defined(_MSC_VER) && defined(__AMD64__) /** @todo check this with with VC7! */
47# pragma intrinsic(_AddressOfReturnAddress)
48#endif
49
50
51/*******************************************************************************
52* Internal Functions *
53*******************************************************************************/
54static int VMMR0Init(PVM pVM, unsigned uVersion);
55static int VMMR0Term(PVM pVM);
56__BEGIN_DECLS
57VMMR0DECL(int) ModuleInit(void);
58VMMR0DECL(void) ModuleTerm(void);
59__END_DECLS
60
61
62//#define DEBUG_NO_RING0_ASSERTIONS
63#ifdef DEBUG_NO_RING0_ASSERTIONS
64static PVM g_pVMAssert = 0;
65#endif
66
67/*******************************************************************************
68* Global Variables *
69*******************************************************************************/
70#ifndef __AMD64__ /* doesn't link here */
71/** Pointer to the internal networking service instance. */
72PINTNET g_pIntNet = 0;
73#endif
74
75
76/**
77 * Initialize the module.
78 * This is called when we're first loaded.
79 *
80 * @returns 0 on success.
81 * @returns VBox status on failure.
82 */
83VMMR0DECL(int) ModuleInit(void)
84{
85#ifndef __AMD64__ /* doesn't link here */
86 LogFlow(("ModuleInit: g_pIntNet=%p\n", g_pIntNet));
87 g_pIntNet = NULL;
88 LogFlow(("ModuleInit: g_pIntNet=%p should be NULL now...\n", g_pIntNet));
89 int rc = INTNETR0Create(&g_pIntNet);
90 if (VBOX_SUCCESS(rc))
91 {
92 LogFlow(("ModuleInit: returns success. g_pIntNet=%p\n", g_pIntNet));
93 return 0;
94 }
95 g_pIntNet = NULL;
96 LogFlow(("ModuleTerm: returns %Vrc\n", rc));
97 return rc;
98#else
99 return 0;
100#endif
101}
102
103
104/**
105 * Terminate the module.
106 * This is called when we're finally unloaded.
107 */
108VMMR0DECL(void) ModuleTerm(void)
109{
110#ifndef __AMD64__ /* doesn't link here */
111 LogFlow(("ModuleTerm:\n"));
112 if (g_pIntNet)
113 {
114 INTNETR0Destroy(g_pIntNet);
115 g_pIntNet = NULL;
116 }
117 LogFlow(("ModuleTerm: returns\n"));
118#endif
119}
120
121
122/**
123 * Initaties the R0 driver for a particular VM instance.
124 *
125 * @returns VBox status code.
126 *
127 * @param pVM The VM instance in question.
128 * @param uVersion The minimum module version required.
129 */
130static int VMMR0Init(PVM pVM, unsigned uVersion)
131{
132 /*
133 * Check if compatible version.
134 */
135 if ( uVersion != VBOX_VERSION
136 && ( VBOX_GET_VERSION_MAJOR(uVersion) != VBOX_VERSION_MAJOR
137 || VBOX_GET_VERSION_MINOR(uVersion) < VBOX_VERSION_MINOR))
138 return VERR_VERSION_MISMATCH;
139
140 /*
141 * Register the EMT R0 logger instance.
142 */
143 PVMMR0LOGGER pR0Logger = pVM->vmm.s.pR0Logger;
144 if (pR0Logger)
145 {
146#if 0 /* testing of the logger. */
147 LogCom(("VMMR0Init: before %p\n", RTLogDefaultInstance()));
148 LogCom(("VMMR0Init: pfnFlush=%p actual=%p\n", pR0Logger->Logger.pfnFlush, vmmR0LoggerFlush));
149 LogCom(("VMMR0Init: pfnLogger=%p actual=%p\n", pR0Logger->Logger.pfnLogger, vmmR0LoggerWrapper));
150 LogCom(("VMMR0Init: offScratch=%d fFlags=%#x fDestFlags=%#x\n", pR0Logger->Logger.offScratch, pR0Logger->Logger.fFlags, pR0Logger->Logger.fDestFlags));
151
152 RTLogSetDefaultInstanceThread(&pR0Logger->Logger, (uintptr_t)pVM->pSession);
153 LogCom(("VMMR0Init: after %p reg\n", RTLogDefaultInstance()));
154 RTLogSetDefaultInstanceThread(NULL, 0);
155 LogCom(("VMMR0Init: after %p dereg\n", RTLogDefaultInstance()));
156
157 pR0Logger->Logger.pfnLogger("hello ring-0 logger\n");
158 LogCom(("VMMR0Init: returned succesfully from direct logger call.\n"));
159 pR0Logger->Logger.pfnFlush(&pR0Logger->Logger);
160 LogCom(("VMMR0Init: returned succesfully from direct flush call.\n"));
161
162 RTLogSetDefaultInstanceThread(&pR0Logger->Logger, (uintptr_t)pVM->pSession);
163 LogCom(("VMMR0Init: after %p reg2\n", RTLogDefaultInstance()));
164 pR0Logger->Logger.pfnLogger("hello ring-0 logger\n");
165 LogCom(("VMMR0Init: returned succesfully from direct logger call (2). offScratch=%d\n", pR0Logger->Logger.offScratch));
166 RTLogSetDefaultInstanceThread(NULL, 0);
167 LogCom(("VMMR0Init: after %p dereg2\n", RTLogDefaultInstance()));
168
169 RTLogLoggerEx(&pR0Logger->Logger, 0, ~0U, "hello ring-0 logger (RTLogLoggerEx)\n");
170 LogCom(("VMMR0Init: RTLogLoggerEx returned fine offScratch=%d\n", pR0Logger->Logger.offScratch));
171#endif
172 RTLogSetDefaultInstanceThread(&pR0Logger->Logger, (uintptr_t)pVM->pSession);
173 }
174
175
176 /*
177 * Init VMXM.
178 */
179 HWACCMR0Init(pVM);
180
181 /*
182 * Init CPUM.
183 */
184 int rc = CPUMR0Init(pVM);
185
186 if (RT_FAILURE(rc))
187 RTLogSetDefaultInstanceThread(NULL, 0);
188 return rc;
189}
190
191
192/**
193 * Terminates the R0 driver for a particular VM instance.
194 *
195 * @returns VBox status code.
196 *
197 * @param pVM The VM instance in question.
198 */
199static int VMMR0Term(PVM pVM)
200{
201 /*
202 * Deregister the logger.
203 */
204 RTLogSetDefaultInstanceThread(NULL, 0);
205 return VINF_SUCCESS;
206}
207
208
209/**
210 * Calls the ring-3 host code.
211 *
212 * @returns VBox status code of the ring-3 call.
213 * @param pVM The VM handle.
214 * @param enmOperation The operation.
215 * @param uArg The argument to the operation.
216 */
217VMMR0DECL(int) VMMR0CallHost(PVM pVM, VMMCALLHOST enmOperation, uint64_t uArg)
218{
219/** @todo profile this! */
220 pVM->vmm.s.enmCallHostOperation = enmOperation;
221 pVM->vmm.s.u64CallHostArg = uArg;
222 pVM->vmm.s.rcCallHost = VERR_INTERNAL_ERROR;
223 int rc = vmmR0CallHostLongJmp(&pVM->vmm.s.CallHostR0JmpBuf, VINF_VMM_CALL_HOST);
224 if (rc == VINF_SUCCESS)
225 rc = pVM->vmm.s.rcCallHost;
226 return rc;
227}
228
229
230#ifdef VBOX_WITH_STATISTICS
231/**
232 * Record return code statistics
233 * @param pVM The VM handle.
234 * @param rc The status code.
235 */
236static void vmmR0RecordRC(PVM pVM, int rc)
237{
238 /*
239 * Collect statistics.
240 */
241 switch (rc)
242 {
243 case VINF_SUCCESS:
244 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetNormal);
245 break;
246 case VINF_EM_RAW_INTERRUPT:
247 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetInterrupt);
248 break;
249 case VINF_EM_RAW_INTERRUPT_HYPER:
250 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetInterruptHyper);
251 break;
252 case VINF_EM_RAW_GUEST_TRAP:
253 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetGuestTrap);
254 break;
255 case VINF_EM_RAW_RING_SWITCH:
256 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetRingSwitch);
257 break;
258 case VINF_EM_RAW_RING_SWITCH_INT:
259 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetRingSwitchInt);
260 break;
261 case VINF_EM_RAW_EXCEPTION_PRIVILEGED:
262 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetExceptionPrivilege);
263 break;
264 case VINF_EM_RAW_STALE_SELECTOR:
265 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetStaleSelector);
266 break;
267 case VINF_EM_RAW_IRET_TRAP:
268 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetIRETTrap);
269 break;
270 case VINF_IOM_HC_IOPORT_READ:
271 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetIORead);
272 break;
273 case VINF_IOM_HC_IOPORT_WRITE:
274 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetIOWrite);
275 break;
276 case VINF_IOM_HC_IOPORT_READWRITE:
277 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetIOReadWrite);
278 break;
279 case VINF_IOM_HC_MMIO_READ:
280 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetMMIORead);
281 break;
282 case VINF_IOM_HC_MMIO_WRITE:
283 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetMMIOWrite);
284 break;
285 case VINF_IOM_HC_MMIO_READ_WRITE:
286 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetMMIOReadWrite);
287 break;
288 case VINF_PATM_HC_MMIO_PATCH_READ:
289 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetMMIOPatchRead);
290 break;
291 case VINF_PATM_HC_MMIO_PATCH_WRITE:
292 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetMMIOPatchWrite);
293 break;
294 case VINF_EM_RAW_EMULATE_INSTR:
295 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetEmulate);
296 break;
297 case VINF_PATCH_EMULATE_INSTR:
298 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPatchEmulate);
299 break;
300 case VINF_EM_RAW_EMULATE_INSTR_LDT_FAULT:
301 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetLDTFault);
302 break;
303 case VINF_EM_RAW_EMULATE_INSTR_GDT_FAULT:
304 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetGDTFault);
305 break;
306 case VINF_EM_RAW_EMULATE_INSTR_IDT_FAULT:
307 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetIDTFault);
308 break;
309 case VINF_EM_RAW_EMULATE_INSTR_TSS_FAULT:
310 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetTSSFault);
311 break;
312 case VINF_EM_RAW_EMULATE_INSTR_PD_FAULT:
313 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPDFault);
314 break;
315 case VINF_CSAM_PENDING_ACTION:
316 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetCSAMTask);
317 break;
318 case VINF_PGM_SYNC_CR3:
319 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetSyncCR3);
320 break;
321 case VINF_PATM_PATCH_INT3:
322 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPatchInt3);
323 break;
324 case VINF_PATM_PATCH_TRAP_PF:
325 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPatchPF);
326 break;
327 case VINF_PATM_PATCH_TRAP_GP:
328 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPatchGP);
329 break;
330 case VINF_PATM_PENDING_IRQ_AFTER_IRET:
331 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPatchIretIRQ);
332 break;
333 case VERR_REM_FLUSHED_PAGES_OVERFLOW:
334 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPageOverflow);
335 break;
336 case VINF_EM_RESCHEDULE_REM:
337 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetRescheduleREM);
338 break;
339 case VINF_EM_RAW_TO_R3:
340 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetToR3);
341 break;
342 case VINF_EM_RAW_TIMER_PENDING:
343 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetTimerPending);
344 break;
345 case VINF_EM_RAW_INTERRUPT_PENDING:
346 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetInterruptPending);
347 break;
348 case VINF_VMM_CALL_HOST:
349 switch (pVM->vmm.s.enmCallHostOperation)
350 {
351 case VMMCALLHOST_PDM_LOCK:
352 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPDMLock);
353 break;
354 case VMMCALLHOST_PDM_QUEUE_FLUSH:
355 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPDMQueueFlush);
356 break;
357 case VMMCALLHOST_PGM_POOL_GROW:
358 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPGMPoolGrow);
359 break;
360 case VMMCALLHOST_PGM_LOCK:
361 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPGMLock);
362 break;
363 case VMMCALLHOST_REM_REPLAY_HANDLER_NOTIFICATIONS:
364 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetRemReplay);
365 break;
366 case VMMCALLHOST_PGM_RAM_GROW_RANGE:
367 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPGMGrowRAM);
368 break;
369 case VMMCALLHOST_VMM_LOGGER_FLUSH:
370 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetLogFlush);
371 break;
372 case VMMCALLHOST_VM_SET_ERROR:
373 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetVMSetError);
374 break;
375 case VMMCALLHOST_VM_SET_RUNTIME_ERROR:
376 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetVMSetRuntimeError);
377 break;
378 default:
379 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetCallHost);
380 break;
381 }
382 break;
383 case VINF_PATM_DUPLICATE_FUNCTION:
384 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPATMDuplicateFn);
385 break;
386 case VINF_PGM_CHANGE_MODE:
387 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPGMChangeMode);
388 break;
389 case VINF_EM_RAW_EMULATE_INSTR_HLT:
390 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetEmulHlt);
391 break;
392 case VINF_EM_PENDING_REQUEST:
393 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetPendingRequest);
394 break;
395 default:
396 STAM_COUNTER_INC(&pVM->vmm.s.StatGCRetMisc);
397 break;
398 }
399}
400#endif /* VBOX_WITH_STATISTICS */
401
402
403/**
404 * The Ring 0 entry point, called by the support library (SUP).
405 *
406 * @returns VBox status code.
407 * @param pVM The VM to operate on.
408 * @param uOperation Which operation to execute. (VMMR0OPERATION)
409 * @param pvArg Argument to the operation.
410 */
411VMMR0DECL(int) VMMR0Entry(PVM pVM, unsigned /* make me an enum */ uOperation, void *pvArg)
412{
413 switch (uOperation)
414 {
415 /*
416 * Switch to GC.
417 * These calls return whatever the GC returns.
418 */
419 case VMMR0_DO_RAW_RUN:
420 {
421 /* Safety precaution as VMX disables the switcher. */
422 Assert(!pVM->vmm.s.fSwitcherDisabled);
423 if (pVM->vmm.s.fSwitcherDisabled)
424 return VERR_NOT_SUPPORTED;
425
426 STAM_COUNTER_INC(&pVM->vmm.s.StatRunGC);
427 register int rc;
428 TMCpuTickResume(pVM);
429 pVM->vmm.s.iLastGCRc = rc = pVM->vmm.s.pfnR0HostToGuest(pVM);
430 TMCpuTickPause(pVM);
431
432#ifdef VBOX_WITH_STATISTICS
433 vmmR0RecordRC(pVM, rc);
434#endif
435
436 /*
437 * Check if there is an exit R0 action associated with the return code.
438 */
439 switch (rc)
440 {
441 /*
442 * Default - no action, just return.
443 */
444 default:
445 return rc;
446
447 /*
448 * We'll let TRPM change the stack frame so our return is different.
449 * Just keep in mind that after the call, things have changed!
450 */
451 case VINF_EM_RAW_INTERRUPT:
452 case VINF_EM_RAW_INTERRUPT_HYPER:
453 {
454#ifdef VBOX_WITHOUT_IDT_PATCHING
455 TRPMR0DispatchHostInterrupt(pVM);
456#else /* !VBOX_WITHOUT_IDT_PATCHING */
457 /*
458 * Don't trust the compiler to get this right.
459 * gcc -fomit-frame-pointer screws up big time here. This works fine in 64-bit
460 * mode too because we push the arguments on the stack in the IDT patch code.
461 */
462# if defined(__GNUC__)
463 void *pvRet = (uint8_t *)__builtin_frame_address(0) + sizeof(void *);
464# elif defined(_MSC_VER) && defined(__AMD64__) /** @todo check this with with VC7! */
465 void *pvRet = (uint8_t *)_AddressOfReturnAddress();
466# elif defined(__X86__)
467 void *pvRet = (uint8_t *)&pVM - sizeof(pVM);
468# else
469# error "huh?"
470# endif
471 if ( ((uintptr_t *)pvRet)[1] == (uintptr_t)pVM
472 && ((uintptr_t *)pvRet)[2] == (uintptr_t)uOperation
473 && ((uintptr_t *)pvRet)[3] == (uintptr_t)pvArg)
474 TRPMR0SetupInterruptDispatcherFrame(pVM, pvRet);
475 else
476 {
477# if defined(DEBUG) || defined(LOG_ENABLED)
478 static bool s_fHaveWarned = false;
479 if (!s_fHaveWarned)
480 {
481 s_fHaveWarned = true;
482 //RTLogPrintf("VMMR0.r0: The compiler can't find the stack frame!\n"); -- @todo export me!
483 RTLogComPrintf("VMMR0.r0: The compiler can't find the stack frame!\n");
484 }
485# endif
486 TRPMR0DispatchHostInterrupt(pVM);
487 }
488#endif /* !VBOX_WITHOUT_IDT_PATCHING */
489 return rc;
490 }
491 }
492 /* Won't get here! */
493 break;
494 }
495
496 /*
497 * Run guest code using the available hardware acceleration technology.
498 */
499 case VMMR0_DO_HWACC_RUN:
500 {
501 int rc;
502
503 STAM_COUNTER_INC(&pVM->vmm.s.StatRunGC);
504 TMCpuTickResume(pVM);
505 rc = HWACCMR0Enable(pVM);
506 if (VBOX_SUCCESS(rc))
507 {
508#ifdef DEBUG_NO_RING0_ASSERTIONS
509 g_pVMAssert = pVM;
510#endif
511 rc = vmmR0CallHostSetJmp(&pVM->vmm.s.CallHostR0JmpBuf, HWACCMR0RunGuestCode, pVM); /* this may resume code. */
512#ifdef DEBUG_NO_RING0_ASSERTIONS
513 g_pVMAssert = 0;
514#endif
515 int rc2 = HWACCMR0Disable(pVM);
516 AssertRC(rc2);
517 }
518 TMCpuTickPause(pVM);
519 pVM->vmm.s.iLastGCRc = rc;
520
521#ifdef VBOX_WITH_STATISTICS
522 vmmR0RecordRC(pVM, rc);
523#endif
524 /* No special action required for external interrupts, just return. */
525 return rc;
526 }
527
528 /*
529 * Initialize the R0 part of a VM instance.
530 */
531 case VMMR0_DO_VMMR0_INIT:
532 return VMMR0Init(pVM, (unsigned)(uintptr_t)pvArg);
533
534 /*
535 * Terminate the R0 part of a VM instance.
536 */
537 case VMMR0_DO_VMMR0_TERM:
538 return VMMR0Term(pVM);
539
540 /*
541 * Setup the hardware accelerated raw-mode session.
542 */
543 case VMMR0_DO_HWACC_SETUP_VM:
544 return HWACCMR0SetupVMX(pVM);
545
546 /*
547 * Switch to GC to execute Hypervisor function.
548 */
549 case VMMR0_DO_CALL_HYPERVISOR:
550 {
551 /* Safety precaution as VMX disables the switcher. */
552 Assert(!pVM->vmm.s.fSwitcherDisabled);
553 if (pVM->vmm.s.fSwitcherDisabled)
554 return VERR_NOT_SUPPORTED;
555
556 int rc = pVM->vmm.s.pfnR0HostToGuest(pVM);
557 return rc;
558 }
559
560#if !defined(__L4__) && !defined(__AMD64__) /** @todo Port this to L4. */ /** @todo fix logging and other services problems on AMD64. */
561 /*
562 * Services.
563 */
564 case VMMR0_DO_INTNET_OPEN:
565 case VMMR0_DO_INTNET_IF_CLOSE:
566 case VMMR0_DO_INTNET_IF_GET_RING3_BUFFER:
567 case VMMR0_DO_INTNET_IF_SET_PROMISCUOUS_MODE:
568 case VMMR0_DO_INTNET_IF_SEND:
569 case VMMR0_DO_INTNET_IF_WAIT:
570 {
571 /*
572 * Validate arguments a bit first.
573 */
574 if (!VALID_PTR(pvArg))
575 return VERR_INVALID_POINTER;
576 if (!VALID_PTR(pVM))
577 return VERR_INVALID_POINTER;
578 if (pVM->pVMHC != pVM)
579 return VERR_INVALID_POINTER;
580 if (!VALID_PTR(pVM->pSession))
581 return VERR_INVALID_POINTER;
582 if (!g_pIntNet)
583 return VERR_FILE_NOT_FOUND; ///@todo fix this status code!
584
585 /*
586 * Unpack the arguments and call the service.
587 */
588 switch (uOperation)
589 {
590 case VMMR0_DO_INTNET_OPEN:
591 {
592 PINTNETOPENARGS pArgs = (PINTNETOPENARGS)pvArg;
593 return INTNETR0Open(g_pIntNet, pVM->pSession, &pArgs->szNetwork[0], pArgs->cbSend, pArgs->cbRecv, &pArgs->hIf);
594 }
595
596 case VMMR0_DO_INTNET_IF_CLOSE:
597 {
598 PINTNETIFCLOSEARGS pArgs = (PINTNETIFCLOSEARGS)pvArg;
599 return INTNETR0IfClose(g_pIntNet, pArgs->hIf);
600 }
601
602 case VMMR0_DO_INTNET_IF_GET_RING3_BUFFER:
603 {
604 PINTNETIFGETRING3BUFFERARGS pArgs = (PINTNETIFGETRING3BUFFERARGS)pvArg;
605 return INTNETR0IfGetRing3Buffer(g_pIntNet, pArgs->hIf, &pArgs->pRing3Buf);
606 }
607
608 case VMMR0_DO_INTNET_IF_SET_PROMISCUOUS_MODE:
609 {
610 PINTNETIFSETPROMISCUOUSMODEARGS pArgs = (PINTNETIFSETPROMISCUOUSMODEARGS)pvArg;
611 return INTNETR0IfSetPromiscuousMode(g_pIntNet, pArgs->hIf, pArgs->fPromiscuous);
612 }
613
614 case VMMR0_DO_INTNET_IF_SEND:
615 {
616 PINTNETIFSENDARGS pArgs = (PINTNETIFSENDARGS)pvArg;
617 return INTNETR0IfSend(g_pIntNet, pArgs->hIf, pArgs->pvFrame, pArgs->cbFrame);
618 }
619
620 case VMMR0_DO_INTNET_IF_WAIT:
621 {
622 PINTNETIFWAITARGS pArgs = (PINTNETIFWAITARGS)pvArg;
623 return INTNETR0IfWait(g_pIntNet, pArgs->hIf, pArgs->cMillies);
624 }
625
626 default:
627 return VERR_NOT_SUPPORTED;
628 }
629 }
630#endif /* !__L4__ */
631
632 /*
633 * For profiling.
634 */
635 case VMMR0_DO_NOP:
636 return VINF_SUCCESS;
637
638 /*
639 * For testing Ring-0 APIs invoked in this environment.
640 */
641 case VMMR0_DO_TESTS:
642 /** @todo make new test */
643 return VINF_SUCCESS;
644
645
646 default:
647 /*
648 * We're returning VERR_NOT_SUPPORT here so we've got something else
649 * than -1 which the interrupt gate glue code might return.
650 */
651 Log(("operation %#x is not supported\n", uOperation));
652 return VERR_NOT_SUPPORTED;
653 }
654}
655
656
657/**
658 * Internal R0 logger worker: Flush logger.
659 *
660 * @param pLogger The logger instance to flush.
661 * @remark This function must be exported!
662 */
663VMMR0DECL(void) vmmR0LoggerFlush(PRTLOGGER pLogger)
664{
665 /*
666 * Convert the pLogger into a VM handle and 'call' back to Ring-3.
667 * (This is a bit paranoid code.)
668 */
669 PVMMR0LOGGER pR0Logger = (PVMMR0LOGGER)((uintptr_t)pLogger - RT_OFFSETOF(VMMR0LOGGER, Logger));
670 if ( !VALID_PTR(pR0Logger)
671 || !VALID_PTR(pR0Logger + 1)
672 || !VALID_PTR(pLogger)
673 || pLogger->u32Magic != RTLOGGER_MAGIC)
674 {
675 LogCom(("vmmR0LoggerFlush: pLogger=%p!\n", pLogger));
676 return;
677 }
678
679 PVM pVM = pR0Logger->pVM;
680 if ( !VALID_PTR(pVM)
681 || pVM->pVMHC != pVM)
682 {
683 LogCom(("vmmR0LoggerFlush: pVM=%p! pLogger=%p\n", pVM, pLogger));
684 return;
685 }
686
687 /*
688 * Check that the jump buffer is armed.
689 */
690#ifdef __X86__
691 if (!pVM->vmm.s.CallHostR0JmpBuf.eip)
692#else
693 if (!pVM->vmm.s.CallHostR0JmpBuf.rip)
694#endif
695 {
696 LogCom(("vmmR0LoggerFlush: Jump buffer isn't armed!\n"));
697 pLogger->offScratch = 0;
698 return;
699 }
700
701 VMMR0CallHost(pVM, VMMCALLHOST_VMM_LOGGER_FLUSH, 0);
702}
703
704#ifdef DEBUG_NO_RING0_ASSERTIONS
705/**
706 * Check if we really want to hit a breakpoint.
707 * Can jump back to ring-3 when the longjmp is armed.
708 */
709DECLEXPORT(bool) RTCALL RTAssertDoBreakpoint()
710{
711 if (g_pVMAssert)
712 {
713 g_pVMAssert->vmm.s.enmCallHostOperation = VMMCALLHOST_VMM_LOGGER_FLUSH;
714 g_pVMAssert->vmm.s.u64CallHostArg = 0;
715 g_pVMAssert->vmm.s.rcCallHost = VERR_INTERNAL_ERROR;
716 int rc = vmmR0CallHostLongJmp(&g_pVMAssert->vmm.s.CallHostR0JmpBuf, VERR_INTERNAL_ERROR);
717 if (rc == VINF_SUCCESS)
718 rc = g_pVMAssert->vmm.s.rcCallHost;
719 }
720
721 return true;
722}
723
724
725#undef LOG_GROUP
726#define LOG_GROUP LOG_GROUP_EM
727
728/** Runtime assert implementation for Native Win32 Ring-0. */
729DECLEXPORT(void) RTCALL AssertMsg1(const char *pszExpr, unsigned uLine, const char *pszFile, const char *pszFunction)
730{
731 Log(("\n!!Assertion Failed!!\n"
732 "Expression: %s\n"
733 "Location : %s(%d) %s\n",
734 pszExpr, pszFile, uLine, pszFunction));
735}
736
737#endif
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette