Lines 1448-1454
DECLHIDDEN(int) rtR0MemObjNativeMapUser(PPRTR0MEMOBJINTERNAL ppMem, RTR0MEMOBJ p
Link Here
|
1448 |
|
1448 |
|
1449 |
#if defined(VBOX_USE_INSERT_PAGE) && LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 22) |
1449 |
#if defined(VBOX_USE_INSERT_PAGE) && LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 22) |
1450 |
rc = vm_insert_page(vma, ulAddrCur, pMemLnxToMap->apPages[iPage]); |
1450 |
rc = vm_insert_page(vma, ulAddrCur, pMemLnxToMap->apPages[iPage]); |
1451 |
vma->vm_flags |= VM_RESERVED; /* This flag helps making 100% sure some bad stuff wont happen (swap, core, ++). */ |
1451 |
vma->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; /* These flags help making 100% sure some bad stuff wont happen (swap, core, ++). */ |
1452 |
#elif LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 11) |
1452 |
#elif LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 11) |
1453 |
rc = remap_pfn_range(vma, ulAddrCur, page_to_pfn(pMemLnxToMap->apPages[iPage]), PAGE_SIZE, fPg); |
1453 |
rc = remap_pfn_range(vma, ulAddrCur, page_to_pfn(pMemLnxToMap->apPages[iPage]), PAGE_SIZE, fPg); |
1454 |
#elif defined(VBOX_USE_PAE_HACK) |
1454 |
#elif defined(VBOX_USE_PAE_HACK) |
1455 |
- |
|
|