aboutsummaryrefslogtreecommitdiff
path: root/emulators/virtualbox-ose-kmod-legacy/files/patch-src-VBox-Runtime-r0drv-freebsd-memobj-r0drv-freebsd.c
diff options
context:
space:
mode:
Diffstat (limited to 'emulators/virtualbox-ose-kmod-legacy/files/patch-src-VBox-Runtime-r0drv-freebsd-memobj-r0drv-freebsd.c')
-rw-r--r--emulators/virtualbox-ose-kmod-legacy/files/patch-src-VBox-Runtime-r0drv-freebsd-memobj-r0drv-freebsd.c870
1 files changed, 870 insertions, 0 deletions
diff --git a/emulators/virtualbox-ose-kmod-legacy/files/patch-src-VBox-Runtime-r0drv-freebsd-memobj-r0drv-freebsd.c b/emulators/virtualbox-ose-kmod-legacy/files/patch-src-VBox-Runtime-r0drv-freebsd-memobj-r0drv-freebsd.c
new file mode 100644
index 000000000000..5dca2a1e2cd0
--- /dev/null
+++ b/emulators/virtualbox-ose-kmod-legacy/files/patch-src-VBox-Runtime-r0drv-freebsd-memobj-r0drv-freebsd.c
@@ -0,0 +1,870 @@
+Index: src/VBox/Runtime/r0drv/freebsd/memobj-r0drv-freebsd.c
+===================================================================
+--- src/VBox/Runtime/r0drv/freebsd/memobj-r0drv-freebsd.c (Revision 58727)
++++ src/VBox/Runtime/r0drv/freebsd/memobj-r0drv-freebsd.c (Arbeitskopie)
+@@ -43,6 +43,16 @@
+ #include <iprt/process.h>
+ #include "internal/memobj.h"
+
++/**
++ * Our pmap_enter version
++ */
++#if __FreeBSD_version >= 701105
++# define MY_PMAP_ENTER(pPhysMap, AddrR3, pPage, fProt, fWired) \
++ pmap_enter(pPhysMap, AddrR3, VM_PROT_NONE, pPage, fProt, fWired)
++#else
++# define MY_PMAP_ENTER(pPhysMap, AddrR3, pPage, fProt, fWired) \
++ pmap_enter(pPhysMap, AddrR3, pPage, fProt, fWired)
++#endif
+
+ /*******************************************************************************
+ * Structures and Typedefs *
+@@ -54,12 +64,24 @@
+ {
+ /** The core structure. */
+ RTR0MEMOBJINTERNAL Core;
+- /** The VM object associated with the allocation. */
+- vm_object_t pObject;
+- /** the VM object associated with the mapping.
+- * In mapping mem object, this is the shadow object?
+- * In a allocation/enter mem object, this is the shared object we constructed (contig, perhaps alloc). */
+- vm_object_t pMappingObject;
++ /** Type dependent data */
++ union
++ {
++ /** Non physical memory allocations */
++ struct
++ {
++ /** The VM object associated with the allocation. */
++ vm_object_t pObject;
++ } NonPhys;
++ /** Physical memory allocations */
++ struct
++ {
++ /** Number of pages */
++ uint32_t cPages;
++ /** Array of pages - variable */
++ vm_page_t apPages[1];
++ } Phys;
++ } u;
+ } RTR0MEMOBJFREEBSD, *PRTR0MEMOBJFREEBSD;
+
+
+@@ -69,7 +91,45 @@
+ * Internal Functions *
+ *******************************************************************************/
+
++/**
++ * Gets the virtual memory map the specified object is mapped into.
++ *
++ * @returns VM map handle on success, NULL if no map.
++ * @param pMem The memory object.
++ */
++static vm_map_t rtR0MemObjFreeBSDGetMap(PRTR0MEMOBJINTERNAL pMem)
++{
++ switch (pMem->enmType)
++ {
++ case RTR0MEMOBJTYPE_PAGE:
++ case RTR0MEMOBJTYPE_LOW:
++ case RTR0MEMOBJTYPE_CONT:
++ return kernel_map;
+
++ case RTR0MEMOBJTYPE_PHYS:
++ case RTR0MEMOBJTYPE_PHYS_NC:
++ return NULL; /* pretend these have no mapping atm. */
++
++ case RTR0MEMOBJTYPE_LOCK:
++ return pMem->u.Lock.R0Process == NIL_RTR0PROCESS
++ ? kernel_map
++ : &((struct proc *)pMem->u.Lock.R0Process)->p_vmspace->vm_map;
++
++ case RTR0MEMOBJTYPE_RES_VIRT:
++ return pMem->u.ResVirt.R0Process == NIL_RTR0PROCESS
++ ? kernel_map
++ : &((struct proc *)pMem->u.ResVirt.R0Process)->p_vmspace->vm_map;
++
++ case RTR0MEMOBJTYPE_MAPPING:
++ return pMem->u.Mapping.R0Process == NIL_RTR0PROCESS
++ ? kernel_map
++ : &((struct proc *)pMem->u.Mapping.R0Process)->p_vmspace->vm_map;
++
++ default:
++ return NULL;
++ }
++}
++
+ int rtR0MemObjNativeFree(RTR0MEMOBJ pMem)
+ {
+ PRTR0MEMOBJFREEBSD pMemFreeBSD = (PRTR0MEMOBJFREEBSD)pMem;
+@@ -79,53 +139,37 @@
+ {
+ case RTR0MEMOBJTYPE_CONT:
+ contigfree(pMemFreeBSD->Core.pv, pMemFreeBSD->Core.cb, M_IPRTMOBJ);
+- if (pMemFreeBSD->pMappingObject)
+- {
+- rc = vm_map_remove(kernel_map,
+- (vm_offset_t)pMemFreeBSD->Core.pv,
+- (vm_offset_t)pMemFreeBSD->Core.pv + pMemFreeBSD->Core.cb);
+- AssertMsg(rc == KERN_SUCCESS, ("%#x", rc));
+- }
+ break;
+
+ case RTR0MEMOBJTYPE_PAGE:
+- if (pMemFreeBSD->pObject)
++ {
++ rc = vm_map_remove(kernel_map,
++ (vm_offset_t)pMemFreeBSD->Core.pv,
++ (vm_offset_t)pMemFreeBSD->Core.pv + pMemFreeBSD->Core.cb);
++ AssertMsg(rc == KERN_SUCCESS, ("%#x", rc));
++
++ vm_page_lock_queues();
++ for (uint32_t iPage = 0; iPage < pMemFreeBSD->u.Phys.cPages; iPage++)
+ {
+- rc = vm_map_remove(kernel_map,
+- (vm_offset_t)pMemFreeBSD->Core.pv,
+- (vm_offset_t)pMemFreeBSD->Core.pv + pMemFreeBSD->Core.cb);
+- AssertMsg(rc == KERN_SUCCESS, ("%#x", rc));
++ vm_page_t pPage = pMemFreeBSD->u.Phys.apPages[iPage];
++ vm_page_unwire(pPage, 0);
++ vm_page_free(pPage);
+ }
+- else
+- {
+- contigfree(pMemFreeBSD->Core.pv, pMemFreeBSD->Core.cb, M_IPRTMOBJ);
+- if (pMemFreeBSD->pMappingObject)
+- {
+- rc = vm_map_remove(kernel_map,
+- (vm_offset_t)pMemFreeBSD->Core.pv,
+- (vm_offset_t)pMemFreeBSD->Core.pv + pMemFreeBSD->Core.cb);
+- AssertMsg(rc == KERN_SUCCESS, ("%#x", rc));
+- }
+- }
++ vm_page_unlock_queues();
+ break;
++ }
+
+ case RTR0MEMOBJTYPE_LOCK:
+ {
+- int fFlags = VM_MAP_WIRE_NOHOLES;
+ vm_map_t pMap = kernel_map;
+
+ if (pMemFreeBSD->Core.u.Lock.R0Process != NIL_RTR0PROCESS)
+- {
+ pMap = &((struct proc *)pMemFreeBSD->Core.u.Lock.R0Process)->p_vmspace->vm_map;
+- fFlags |= VM_MAP_WIRE_USER;
+- }
+- else
+- fFlags |= VM_MAP_WIRE_SYSTEM;
+
+ rc = vm_map_unwire(pMap,
+ (vm_offset_t)pMemFreeBSD->Core.pv,
+ (vm_offset_t)pMemFreeBSD->Core.pv + pMemFreeBSD->Core.cb,
+- fFlags);
++ VM_MAP_WIRE_SYSTEM | VM_MAP_WIRE_NOHOLES);
+ AssertMsg(rc == KERN_SUCCESS, ("%#x", rc));
+ break;
+ }
+@@ -146,8 +190,6 @@
+ {
+ vm_map_t pMap = kernel_map;
+
+- /* vm_map_remove will unmap the pages we inserted with pmap_enter */
+- AssertMsg(pMemFreeBSD->pMappingObject != NULL, ("MappingObject is NULL\n"));
+ if (pMemFreeBSD->Core.u.Mapping.R0Process != NIL_RTR0PROCESS)
+ pMap = &((struct proc *)pMemFreeBSD->Core.u.Mapping.R0Process)->p_vmspace->vm_map;
+
+@@ -158,10 +200,22 @@
+ break;
+ }
+
++ case RTR0MEMOBJTYPE_PHYS:
++ case RTR0MEMOBJTYPE_PHYS_NC:
++ {
++ vm_page_lock_queues();
++ for (uint32_t iPage = 0; iPage < pMemFreeBSD->u.Phys.cPages; iPage++)
++ {
++ vm_page_t pPage = pMemFreeBSD->u.Phys.apPages[iPage];
++ vm_page_unwire(pPage, 0);
++ vm_page_free(pPage);
++ }
++ vm_page_unlock_queues();
++ break;
++ }
++
+ /* unused: */
+ case RTR0MEMOBJTYPE_LOW:
+- case RTR0MEMOBJTYPE_PHYS:
+- case RTR0MEMOBJTYPE_PHYS_NC:
+ default:
+ AssertMsgFailed(("enmType=%d\n", pMemFreeBSD->Core.enmType));
+ return VERR_INTERNAL_ERROR;
+@@ -177,95 +231,83 @@
+ size_t cPages = cb >> PAGE_SHIFT;
+
+ /* create the object. */
+- PRTR0MEMOBJFREEBSD pMemFreeBSD = (PRTR0MEMOBJFREEBSD)rtR0MemObjNew(sizeof(*pMemFreeBSD), RTR0MEMOBJTYPE_PAGE, NULL, cb);
++ PRTR0MEMOBJFREEBSD pMemFreeBSD = (PRTR0MEMOBJFREEBSD)rtR0MemObjNew(RT_OFFSETOF(RTR0MEMOBJFREEBSD, u.Phys.apPages[cPages]),
++ RTR0MEMOBJTYPE_PAGE, NULL, cb);
+ if (!pMemFreeBSD)
+ return VERR_NO_MEMORY;
+
+- pMemFreeBSD->pObject = vm_object_allocate(OBJT_DEFAULT, cPages);
+- if (pMemFreeBSD->pObject)
++ pMemFreeBSD->u.Phys.cPages = cPages;
++
++ vm_offset_t MapAddress = vm_map_min(kernel_map);
++ rc = vm_map_find(kernel_map, /* map */
++ NULL, /* object */
++ 0, /* offset */
++ &MapAddress, /* addr (IN/OUT) */
++ cb, /* length */
++ TRUE, /* find_space */
++ fExecutable /* protection */
++ ? VM_PROT_ALL
++ : VM_PROT_RW,
++ VM_PROT_ALL, /* max(_prot) */
++ 0); /* cow (copy-on-write) */
++ if (rc == KERN_SUCCESS)
+ {
+- vm_offset_t MapAddress = vm_map_min(kernel_map);
+- rc = vm_map_find(kernel_map, /* map */
+- pMemFreeBSD->pObject, /* object */
+- 0, /* offset */
+- &MapAddress, /* addr (IN/OUT) */
+- cb, /* length */
+- TRUE, /* find_space */
+- fExecutable /* protection */
+- ? VM_PROT_ALL
+- : VM_PROT_RW,
+- VM_PROT_ALL, /* max(_prot) */
+- FALSE); /* cow (copy-on-write) */
+- if (rc == KERN_SUCCESS)
++ rc = VINF_SUCCESS;
++
++ for (size_t iPage = 0; iPage < cPages; iPage++)
+ {
+- vm_offset_t AddressDst = MapAddress;
++ vm_page_t pPage;
+
+- rc = VINF_SUCCESS;
++ pPage = vm_page_alloc(NULL, iPage,
++ VM_ALLOC_SYSTEM |
++ VM_ALLOC_WIRED | VM_ALLOC_NOOBJ);
+
+- VM_OBJECT_LOCK(pMemFreeBSD->pObject);
+- for (size_t iPage = 0; iPage < cPages; iPage++)
++ if (!pPage)
+ {
+- vm_pindex_t PageIndex = OFF_TO_IDX(AddressDst);
+- vm_page_t pPage;
+-
+- pPage = vm_page_alloc(pMemFreeBSD->pObject, PageIndex,
+- VM_ALLOC_NOBUSY | VM_ALLOC_SYSTEM |
+- VM_ALLOC_WIRED);
+-
+-#if __FreeBSD_version >= 800000 /** @todo Find exact version number */
+- /* Fixes crashes during VM termination on FreeBSD8-CURRENT amd64
+- * with kernel debugging enabled. */
+- vm_page_set_valid(pPage, 0, PAGE_SIZE);
+-#endif
+-
+- if (pPage)
++ /*
++ * Out of pages
++ * Remove already allocated pages
++ */
++ while (iPage-- > 0)
+ {
++ pPage = pMemFreeBSD->u.Phys.apPages[iPage];
+ vm_page_lock_queues();
+- vm_page_wire(pPage);
++ vm_page_unwire(pPage, 0);
++ vm_page_free(pPage);
+ vm_page_unlock_queues();
+- /* Put the page into the page table now. */
+-#if __FreeBSD_version >= 701105
+- pmap_enter(kernel_map->pmap, AddressDst, VM_PROT_NONE, pPage,
+- fExecutable
+- ? VM_PROT_ALL
+- : VM_PROT_RW,
+- TRUE);
+-#else
+- pmap_enter(kernel_map->pmap, AddressDst, pPage,
+- fExecutable
+- ? VM_PROT_ALL
+- : VM_PROT_RW,
+- TRUE);
+-#endif
+ }
+- else
+- {
+- /*
+- * Allocation failed. vm_map_remove will remove any
+- * page already alocated.
+- */
+- rc = VERR_NO_MEMORY;
+- break;
+- }
+- AddressDst += PAGE_SIZE;
++ rc = VERR_NO_MEMORY;
++ break;
+ }
+- VM_OBJECT_UNLOCK(pMemFreeBSD->pObject);
+
+- if (rc == VINF_SUCCESS)
++ pPage->valid = VM_PAGE_BITS_ALL;
++ pMemFreeBSD->u.Phys.apPages[iPage] = pPage;
++ }
++
++ if (rc == VINF_SUCCESS)
++ {
++ vm_offset_t AddressDst = MapAddress;
++
++ for (size_t iPage = 0; iPage < cPages; iPage++)
+ {
+- pMemFreeBSD->Core.pv = (void *)MapAddress;
+- *ppMem = &pMemFreeBSD->Core;
+- return VINF_SUCCESS;
++ vm_page_t pPage = pMemFreeBSD->u.Phys.apPages[iPage];
++
++ MY_PMAP_ENTER(kernel_map->pmap, AddressDst, pPage,
++ fExecutable
++ ? VM_PROT_ALL
++ : VM_PROT_RW,
++ TRUE);
++
++ AddressDst += PAGE_SIZE;
+ }
+
+- vm_map_remove(kernel_map,
+- MapAddress,
+- MapAddress + cb);
++ /* Store start address */
++ pMemFreeBSD->Core.pv = (void *)MapAddress;
++ *ppMem = &pMemFreeBSD->Core;
++ return VINF_SUCCESS;
+ }
+- rc = VERR_NO_MEMORY; /** @todo fix translation (borrow from darwin) */
+ }
+- else
+- rc = VERR_NO_MEMORY;
++ rc = VERR_NO_MEMORY; /** @todo fix translation (borrow from darwin) */
+
+ rtR0MemObjDelete(&pMemFreeBSD->Core);
+ return rc;
+@@ -325,12 +367,99 @@
+ return VERR_NO_MEMORY;
+ }
+
++static void rtR0MemObjFreeBSDPhysPageInit(vm_page_t pPage, vm_pindex_t iPage)
++{
++ pPage->wire_count = 1;
++ pPage->pindex = iPage;
++ pPage->act_count = 0;
++ pPage->oflags = 0;
++ pPage->flags = PG_UNMANAGED;
++ atomic_add_int(&cnt.v_wire_count, 1);
++}
+
+-int rtR0MemObjNativeAllocPhys(PPRTR0MEMOBJINTERNAL ppMem, size_t cb, RTHCPHYS PhysHighest)
++static int rtR0MemObjFreeBSDAllocPhysPages(PPRTR0MEMOBJINTERNAL ppMem, RTR0MEMOBJTYPE enmType,
++ size_t cb,
++ RTHCPHYS PhysHighest, size_t uAlignment,
++ bool fContiguous)
+ {
+- /** @todo check if there is a more appropriate API somewhere.. */
++ int rc = VINF_SUCCESS;
++ uint32_t cPages = cb >> PAGE_SHIFT;
++ vm_paddr_t VmPhysAddrHigh;
+
+ /* create the object. */
++ PRTR0MEMOBJFREEBSD pMemFreeBSD = (PRTR0MEMOBJFREEBSD)rtR0MemObjNew(RT_OFFSETOF(RTR0MEMOBJFREEBSD, u.Phys.apPages[cPages]),
++ enmType, NULL, cb);
++ if (!pMemFreeBSD)
++ return VERR_NO_MEMORY;
++
++ pMemFreeBSD->u.Phys.cPages = cPages;
++
++ if (PhysHighest != NIL_RTHCPHYS)
++ VmPhysAddrHigh = PhysHighest;
++ else
++ VmPhysAddrHigh = ~(vm_paddr_t)0;
++
++ if (fContiguous)
++ {
++ vm_page_t pPage = vm_phys_alloc_contig(cPages, 0, VmPhysAddrHigh, uAlignment, 0);
++
++ if (pPage)
++ for (uint32_t iPage = 0; iPage < cPages; iPage++)
++ {
++ rtR0MemObjFreeBSDPhysPageInit(&pPage[iPage], iPage);
++ pMemFreeBSD->u.Phys.apPages[iPage] = &pPage[iPage];
++ }
++ else
++ rc = VERR_NO_MEMORY;
++ }
++ else
++ {
++ /* Allocate page by page */
++ for (uint32_t iPage = 0; iPage < cPages; iPage++)
++ {
++ vm_page_t pPage = vm_phys_alloc_contig(1, 0, VmPhysAddrHigh, uAlignment, 0);
++
++ if (!pPage)
++ {
++ /* Free all allocated pages */
++ while (iPage-- > 0)
++ {
++ pPage = pMemFreeBSD->u.Phys.apPages[iPage];
++ vm_page_lock_queues();
++ vm_page_unwire(pPage, 0);
++ vm_page_free(pPage);
++ vm_page_unlock_queues();
++ }
++ rc = VERR_NO_MEMORY;
++ break;
++ }
++ rtR0MemObjFreeBSDPhysPageInit(pPage, iPage);
++ pMemFreeBSD->u.Phys.apPages[iPage] = pPage;
++ }
++ }
++
++ if (RT_FAILURE(rc))
++ rtR0MemObjDelete(&pMemFreeBSD->Core);
++ else
++ {
++ if (enmType == RTR0MEMOBJTYPE_PHYS)
++ {
++ pMemFreeBSD->Core.u.Phys.PhysBase = VM_PAGE_TO_PHYS(pMemFreeBSD->u.Phys.apPages[0]);
++ pMemFreeBSD->Core.u.Phys.fAllocated = true;
++ }
++
++ *ppMem = &pMemFreeBSD->Core;
++ }
++
++ return rc;
++}
++
++int rtR0MemObjNativeAllocPhys(PPRTR0MEMOBJINTERNAL ppMem, size_t cb, RTHCPHYS PhysHighest)
++{
++#if 1
++ return rtR0MemObjFreeBSDAllocPhysPages(ppMem, RTR0MEMOBJTYPE_PHYS, cb, PhysHighest, PAGE_SIZE, true);
++#else
++ /* create the object. */
+ PRTR0MEMOBJFREEBSD pMemFreeBSD = (PRTR0MEMOBJFREEBSD)rtR0MemObjNew(sizeof(*pMemFreeBSD), RTR0MEMOBJTYPE_CONT, NULL, cb);
+ if (!pMemFreeBSD)
+ return VERR_NO_MEMORY;
+@@ -340,7 +469,7 @@
+ M_IPRTMOBJ, /* type */
+ M_NOWAIT | M_ZERO, /* flags */
+ 0, /* lowest physical address*/
+- PhysHighest, /* highest physical address */
++ _4G-1, /* highest physical address */
+ PAGE_SIZE, /* alignment. */
+ 0); /* boundrary */
+ if (pMemFreeBSD->Core.pv)
+@@ -352,13 +481,17 @@
+
+ rtR0MemObjDelete(&pMemFreeBSD->Core);
+ return VERR_NO_MEMORY;
++#endif
+ }
+
+
+ int rtR0MemObjNativeAllocPhysNC(PPRTR0MEMOBJINTERNAL ppMem, size_t cb, RTHCPHYS PhysHighest)
+ {
+- /** @todo rtR0MemObjNativeAllocPhys / freebsd */
++#if 1
++ return rtR0MemObjFreeBSDAllocPhysPages(ppMem, RTR0MEMOBJTYPE_PHYS_NC, cb, PhysHighest, PAGE_SIZE, false);
++#else
+ return VERR_NOT_SUPPORTED;
++#endif
+ }
+
+
+@@ -377,13 +510,18 @@
+ }
+
+
+-int rtR0MemObjNativeLockUser(PPRTR0MEMOBJINTERNAL ppMem, RTR3PTR R3Ptr, size_t cb, uint32_t fAccess, RTR0PROCESS R0Process)
++/**
++ * Worker locking the memory in either kernel or user maps.
++ */
++static int rtR0MemObjNativeLockInMap(PPRTR0MEMOBJINTERNAL ppMem, vm_map_t pVmMap,
++ vm_offset_t AddrStart, size_t cb, uint32_t fAccess,
++ RTR0PROCESS R0Process)
+ {
+ int rc;
+ NOREF(fAccess);
+
+ /* create the object. */
+- PRTR0MEMOBJFREEBSD pMemFreeBSD = (PRTR0MEMOBJFREEBSD)rtR0MemObjNew(sizeof(*pMemFreeBSD), RTR0MEMOBJTYPE_LOCK, (void *)R3Ptr, cb);
++ PRTR0MEMOBJFREEBSD pMemFreeBSD = (PRTR0MEMOBJFREEBSD)rtR0MemObjNew(sizeof(*pMemFreeBSD), RTR0MEMOBJTYPE_LOCK, (void *)AddrStart, cb);
+ if (!pMemFreeBSD)
+ return VERR_NO_MEMORY;
+
+@@ -391,10 +529,10 @@
+ * We could've used vslock here, but we don't wish to be subject to
+ * resource usage restrictions, so we'll call vm_map_wire directly.
+ */
+- rc = vm_map_wire(&((struct proc *)R0Process)->p_vmspace->vm_map, /* the map */
+- (vm_offset_t)R3Ptr, /* start */
+- (vm_offset_t)R3Ptr + cb, /* end */
+- VM_MAP_WIRE_USER | VM_MAP_WIRE_NOHOLES); /* flags */
++ rc = vm_map_wire(pVmMap, /* the map */
++ AddrStart, /* start */
++ AddrStart + cb, /* end */
++ VM_MAP_WIRE_SYSTEM | VM_MAP_WIRE_NOHOLES); /* flags */
+ if (rc == KERN_SUCCESS)
+ {
+ pMemFreeBSD->Core.u.Lock.R0Process = R0Process;
+@@ -406,29 +544,25 @@
+ }
+
+
+-int rtR0MemObjNativeLockKernel(PPRTR0MEMOBJINTERNAL ppMem, void *pv, size_t cb, uint32_t fAccess)
++int rtR0MemObjNativeLockUser(PPRTR0MEMOBJINTERNAL ppMem, RTR3PTR R3Ptr, size_t cb, uint32_t fAccess, RTR0PROCESS R0Process)
+ {
+- int rc;
+- NOREF(fAccess);
++ return rtR0MemObjNativeLockInMap(ppMem,
++ &((struct proc *)R0Process)->p_vmspace->vm_map,
++ (vm_offset_t)R3Ptr,
++ cb,
++ fAccess,
++ R0Process);
++}
+
+- /* create the object. */
+- PRTR0MEMOBJFREEBSD pMemFreeBSD = (PRTR0MEMOBJFREEBSD)rtR0MemObjNew(sizeof(*pMemFreeBSD), RTR0MEMOBJTYPE_LOCK, pv, cb);
+- if (!pMemFreeBSD)
+- return VERR_NO_MEMORY;
+
+- /* lock the memory */
+- rc = vm_map_wire(kernel_map, /* the map */
+- (vm_offset_t)pv, /* start */
+- (vm_offset_t)pv + cb, /* end */
+- VM_MAP_WIRE_SYSTEM | VM_MAP_WIRE_NOHOLES); /* flags - SYSTEM? */
+- if (rc == KERN_SUCCESS)
+- {
+- pMemFreeBSD->Core.u.Lock.R0Process = NIL_RTR0PROCESS;
+- *ppMem = &pMemFreeBSD->Core;
+- return VINF_SUCCESS;
+- }
+- rtR0MemObjDelete(&pMemFreeBSD->Core);
+- return VERR_NO_MEMORY;/** @todo fix mach -> vbox error conversion for freebsd. */
++int rtR0MemObjNativeLockKernel(PPRTR0MEMOBJINTERNAL ppMem, void *pv, size_t cb, uint32_t fAccess)
++{
++ return rtR0MemObjNativeLockInMap(ppMem,
++ kernel_map,
++ (vm_offset_t)pv,
++ cb,
++ fAccess,
++ NIL_RTR0PROCESS);
+ }
+
+
+@@ -465,8 +599,8 @@
+ /*
+ * Allocate an empty VM object and map it into the requested map.
+ */
+- pMemFreeBSD->pObject = vm_object_allocate(OBJT_DEFAULT, cb >> PAGE_SHIFT);
+- if (pMemFreeBSD->pObject)
++ pMemFreeBSD->u.NonPhys.pObject = vm_object_allocate(OBJT_DEFAULT, cb >> PAGE_SHIFT);
++ if (pMemFreeBSD->u.NonPhys.pObject)
+ {
+ vm_offset_t MapAddress = pvFixed != (void *)-1
+ ? (vm_offset_t)pvFixed
+@@ -477,7 +611,7 @@
+ MapAddress + cb);
+
+ rc = vm_map_find(pMap, /* map */
+- pMemFreeBSD->pObject, /* object */
++ pMemFreeBSD->u.NonPhys.pObject, /* object */
+ 0, /* offset */
+ &MapAddress, /* addr (IN/OUT) */
+ cb, /* length */
+@@ -500,7 +634,7 @@
+ *ppMem = &pMemFreeBSD->Core;
+ return VINF_SUCCESS;
+ }
+- vm_object_deallocate(pMemFreeBSD->pObject);
++ vm_object_deallocate(pMemFreeBSD->u.NonPhys.pObject);
+ rc = VERR_NO_MEMORY; /** @todo fix translation (borrow from darwin) */
+ }
+ else
+@@ -535,74 +669,9 @@
+ if (uAlignment > PAGE_SIZE)
+ return VERR_NOT_SUPPORTED;
+
+-
+-
+ /* Phys: see pmap_mapdev in i386/i386/pmap.c (http://fxr.watson.org/fxr/source/i386/i386/pmap.c?v=RELENG62#L2860) */
+-
+-#if 0
+ /** @todo finish the implementation. */
+
+- int rc;
+- void *pvR0 = NULL;
+- PRTR0MEMOBJFREEBSD pMemToMapOs2 = (PRTR0MEMOBJFREEBSD)pMemToMap;
+- switch (pMemToMapOs2->Core.enmType)
+- {
+- /*
+- * These has kernel mappings.
+- */
+- case RTR0MEMOBJTYPE_PAGE:
+- case RTR0MEMOBJTYPE_LOW:
+- case RTR0MEMOBJTYPE_CONT:
+- pvR0 = pMemToMapOs2->Core.pv;
+- break;
+-
+- case RTR0MEMOBJTYPE_PHYS_NC:
+- case RTR0MEMOBJTYPE_PHYS:
+- pvR0 = pMemToMapOs2->Core.pv;
+- if (!pvR0)
+- {
+- /* no ring-0 mapping, so allocate a mapping in the process. */
+- AssertMsgReturn(uAlignment == PAGE_SIZE, ("%#zx\n", uAlignment), VERR_NOT_SUPPORTED);
+- AssertMsgReturn(fProt & RTMEM_PROT_WRITE, ("%#x\n", fProt), VERR_NOT_SUPPORTED);
+- Assert(!pMemToMapOs2->Core.u.Phys.fAllocated);
+- ULONG ulPhys = pMemToMapOs2->Core.u.Phys.PhysBase;
+- rc = KernVMAlloc(pMemToMapOs2->Core.cb, VMDHA_PHYS, &pvR0, (PPVOID)&ulPhys, NULL);
+- if (rc)
+- return RTErrConvertFromOS2(rc);
+- pMemToMapOs2->Core.pv = pvR0;
+- }
+- break;
+-
+- case RTR0MEMOBJTYPE_LOCK:
+- if (pMemToMapOs2->Core.u.Lock.R0Process != NIL_RTR0PROCESS)
+- return VERR_NOT_SUPPORTED; /** @todo implement this... */
+- pvR0 = pMemToMapOs2->Core.pv;
+- break;
+-
+- case RTR0MEMOBJTYPE_RES_VIRT:
+- case RTR0MEMOBJTYPE_MAPPING:
+- default:
+- AssertMsgFailed(("enmType=%d\n", pMemToMapOs2->Core.enmType));
+- return VERR_INTERNAL_ERROR;
+- }
+-
+- /*
+- * Create a dummy mapping object for it.
+- *
+- * All mappings are read/write/execute in OS/2 and there isn't
+- * any cache options, so sharing is ok. And the main memory object
+- * isn't actually freed until all the mappings have been freed up
+- * (reference counting).
+- */
+- PRTR0MEMOBJFREEBSD pMemFreeBSD = (PRTR0MEMOBJFREEBSD)rtR0MemObjNew(RT_OFFSETOF(RTR0MEMOBJOS2, Lock), RTR0MEMOBJTYPE_MAPPING, pvR0, pMemToMapOs2->Core.cb);
+- if (pMemFreeBSD)
+- {
+- pMemFreeBSD->Core.u.Mapping.R0Process = NIL_RTR0PROCESS;
+- *ppMem = &pMemFreeBSD->Core;
+- return VINF_SUCCESS;
+- }
+- return VERR_NO_MEMORY;
+-#endif
+ return VERR_NOT_IMPLEMENTED;
+ }
+
+@@ -618,10 +687,10 @@
+ if (uAlignment > PAGE_SIZE)
+ return VERR_NOT_SUPPORTED;
+
+- int rc;
+- vm_object_t pObjectToMap = ((PRTR0MEMOBJFREEBSD)pMemToMap)->pObject;
+- struct proc *pProc = (struct proc *)R0Process;
+- struct vm_map *pProcMap = &pProc->p_vmspace->vm_map;
++ int rc;
++ PRTR0MEMOBJFREEBSD pMemToMapFreeBSD = (PRTR0MEMOBJFREEBSD)pMemToMap;
++ struct proc *pProc = (struct proc *)R0Process;
++ struct vm_map *pProcMap = &pProc->p_vmspace->vm_map;
+
+ /* calc protection */
+ vm_prot_t ProtectionFlags = 0;
+@@ -639,48 +708,57 @@
+ vm_offset_t AddrR3 = round_page((vm_offset_t)pProc->p_vmspace->vm_daddr + lim_max(pProc, RLIMIT_DATA));
+ PROC_UNLOCK(pProc);
+
+- vm_object_t pObjectNew = vm_object_allocate(OBJT_PHYS, pMemToMap->cb >> PAGE_SHIFT);
+- if (!RT_UNLIKELY(pObjectNew))
+- return VERR_NO_MEMORY;
+-
+ /* Insert the object in the map. */
+ rc = vm_map_find(pProcMap, /* Map to insert the object in */
+- pObjectNew , /* Object to map */
+- 0, /* Start offset in the object */
+- &AddrR3, /* Start address IN/OUT */
+- pMemToMap->cb, /* Size of the mapping */
+- TRUE, /* Whether a suitable address should be searched for first */
+- ProtectionFlags, /* protection flags */
+- VM_PROT_ALL, /* Maximum protection flags */
+- 0); /* Copy on write */
++ NULL, /* Object to map */
++ 0, /* Start offset in the object */
++ &AddrR3, /* Start address IN/OUT */
++ pMemToMap->cb, /* Size of the mapping */
++ TRUE, /* Whether a suitable address should be searched for first */
++ ProtectionFlags, /* protection flags */
++ VM_PROT_ALL, /* Maximum protection flags */
++ 0); /* Copy on write */
+
+ /* Map the memory page by page into the destination map. */
+ if (rc == KERN_SUCCESS)
+ {
+- size_t cLeft = pMemToMap->cb >> PAGE_SHIFT;
+- vm_offset_t AddrToMap = (vm_offset_t)pMemToMap->pv;
++ size_t cPages = pMemToMap->cb >> PAGE_SHIFT;;
+ pmap_t pPhysicalMap = pProcMap->pmap;
+ vm_offset_t AddrR3Dst = AddrR3;
+
+- /* Insert the memory page by page into the mapping. */
+- while (cLeft-- > 0)
++ if ( pMemToMap->enmType == RTR0MEMOBJTYPE_PHYS
++ || pMemToMap->enmType == RTR0MEMOBJTYPE_PHYS_NC
++ || pMemToMap->enmType == RTR0MEMOBJTYPE_PAGE)
+ {
+- vm_page_t Page = PHYS_TO_VM_PAGE(vtophys(AddrToMap));
++ /* Mapping physical allocations */
++ Assert(cPages == pMemToMapFreeBSD->u.Phys.cPages);
+
+-#if __FreeBSD_version >= 701105
+- pmap_enter(pPhysicalMap, AddrR3Dst, VM_PROT_NONE, Page, ProtectionFlags, TRUE);
+-#else
+- pmap_enter(pPhysicalMap, AddrR3Dst, Page, ProtectionFlags, TRUE);
+-#endif
+- AddrToMap += PAGE_SIZE;
+- AddrR3Dst += PAGE_SIZE;
++ /* Insert the memory page by page into the mapping. */
++ for (uint32_t iPage = 0; iPage < cPages; iPage++)
++ {
++ vm_page_t pPage = pMemToMapFreeBSD->u.Phys.apPages[iPage];
++
++ MY_PMAP_ENTER(pPhysicalMap, AddrR3Dst, pPage, ProtectionFlags, TRUE);
++ AddrR3Dst += PAGE_SIZE;
++ }
+ }
+- pObjectToMap = pObjectNew;
++ else
++ {
++ /* Mapping cont or low memory types */
++ vm_offset_t AddrToMap = (vm_offset_t)pMemToMap->pv;
++
++ for (uint32_t iPage = 0; iPage < cPages; iPage++)
++ {
++ vm_page_t pPage = PHYS_TO_VM_PAGE(vtophys(AddrToMap));
++
++ MY_PMAP_ENTER(pPhysicalMap, AddrR3Dst, pPage, ProtectionFlags, TRUE);
++ AddrR3Dst += PAGE_SIZE;
++ AddrToMap += PAGE_SIZE;
++ }
++ }
+ }
+- else
+- vm_object_deallocate(pObjectNew);
+
+- if (rc == KERN_SUCCESS)
++ if (RT_SUCCESS(rc))
+ {
+ /*
+ * Create a mapping object for it.
+@@ -693,7 +771,6 @@
+ {
+ Assert((vm_offset_t)pMemFreeBSD->Core.pv == AddrR3);
+ pMemFreeBSD->Core.u.Mapping.R0Process = R0Process;
+- pMemFreeBSD->pMappingObject = pObjectToMap;
+ *ppMem = &pMemFreeBSD->Core;
+ return VINF_SUCCESS;
+ }
+@@ -702,19 +779,33 @@
+ AssertMsg(rc == KERN_SUCCESS, ("Deleting mapping failed\n"));
+ }
+
+- if (pObjectToMap)
+- vm_object_deallocate(pObjectToMap);
+-
+ return VERR_NO_MEMORY;
+ }
+
+
+ int rtR0MemObjNativeProtect(PRTR0MEMOBJINTERNAL pMem, size_t offSub, size_t cbSub, uint32_t fProt)
+ {
+- NOREF(pMem);
+- NOREF(offSub);
+- NOREF(cbSub);
+- NOREF(fProt);
++ vm_prot_t ProtectionFlags = 0;
++ vm_offset_t AddrStart = (uintptr_t)pMem->pv + offSub;
++ vm_offset_t AddrEnd = AddrStart + cbSub;
++ vm_map_t pVmMap = rtR0MemObjFreeBSDGetMap(pMem);
++
++ if (!pVmMap)
++ return VERR_NOT_SUPPORTED;
++
++ if ((fProt & RTMEM_PROT_NONE) == RTMEM_PROT_NONE)
++ ProtectionFlags = VM_PROT_NONE;
++ if ((fProt & RTMEM_PROT_READ) == RTMEM_PROT_READ)
++ ProtectionFlags |= VM_PROT_READ;
++ if ((fProt & RTMEM_PROT_WRITE) == RTMEM_PROT_WRITE)
++ ProtectionFlags |= VM_PROT_WRITE;
++ if ((fProt & RTMEM_PROT_EXEC) == RTMEM_PROT_EXEC)
++ ProtectionFlags |= VM_PROT_EXECUTE;
++
++ int krc = vm_map_protect(pVmMap, AddrStart, AddrEnd, ProtectionFlags, FALSE);
++ if (krc == KERN_SUCCESS)
++ return VINF_SUCCESS;
++
+ return VERR_NOT_SUPPORTED;
+ }
+
+@@ -726,17 +817,35 @@
+ switch (pMemFreeBSD->Core.enmType)
+ {
+ case RTR0MEMOBJTYPE_LOCK:
++ {
+ if ( pMemFreeBSD->Core.u.Lock.R0Process != NIL_RTR0PROCESS
+ && pMemFreeBSD->Core.u.Lock.R0Process != (RTR0PROCESS)curproc)
+ {
+ /* later */
+ return NIL_RTHCPHYS;
+ }
+- /* fall thru*/
+- case RTR0MEMOBJTYPE_PAGE:
++
++ vm_offset_t pb = (vm_offset_t)pMemFreeBSD->Core.pv + (iPage << PAGE_SHIFT);
++
++ struct proc *pProc = (struct proc *)pMemFreeBSD->Core.u.Lock.R0Process;
++ struct vm_map *pProcMap = &pProc->p_vmspace->vm_map;
++ pmap_t pPhysicalMap = pProcMap->pmap;
++
++ return pmap_extract(pPhysicalMap, pb);
++ }
++
+ case RTR0MEMOBJTYPE_MAPPING:
+ {
+- uint8_t *pb = (uint8_t *)pMemFreeBSD->Core.pv + (iPage << PAGE_SHIFT);
++ vm_offset_t pb = (vm_offset_t)pMemFreeBSD->Core.pv + (iPage << PAGE_SHIFT);
++
++ if (pMemFreeBSD->Core.u.Mapping.R0Process != NIL_RTR0PROCESS)
++ {
++ struct proc *pProc = (struct proc *)pMemFreeBSD->Core.u.Mapping.R0Process;
++ struct vm_map *pProcMap = &pProc->p_vmspace->vm_map;
++ pmap_t pPhysicalMap = pProcMap->pmap;
++
++ return pmap_extract(pPhysicalMap, pb);
++ }
+ return vtophys(pb);
+ }
+
+@@ -746,7 +855,10 @@
+ case RTR0MEMOBJTYPE_PHYS:
+ return pMemFreeBSD->Core.u.Phys.PhysBase + (iPage << PAGE_SHIFT);
+
++ case RTR0MEMOBJTYPE_PAGE:
+ case RTR0MEMOBJTYPE_PHYS_NC:
++ return VM_PAGE_TO_PHYS(pMemFreeBSD->u.Phys.apPages[iPage]);
++
+ case RTR0MEMOBJTYPE_RES_VIRT:
+ case RTR0MEMOBJTYPE_LOW:
+ default: