Lines Matching refs:PAGE_SIZE
88 if (cb <= PAGE_SIZE)
90 cb -= PAGE_SIZE;
91 pu32 += PAGE_SIZE / sizeof(uint32_t);
114 if (cb <= PAGE_SIZE)
116 cb -= PAGE_SIZE;
117 pu32 += PAGE_SIZE / sizeof(uint32_t);
224 IOAddressRange aRanges[1] = { { (mach_vm_address_t)HCPhys, RT_ALIGN_Z(cb, PAGE_SIZE) } };
460 cbFudged += PAGE_SIZE;
470 uint64_t uAlignment = PAGE_SIZE;
498 for (IOByteCount off = 0; off < cb; off += PAGE_SIZE)
510 && Addr == AddrPrev + PAGE_SIZE))
627 ~(uint32_t)PAGE_OFFSET_MASK, _4G - PAGE_SIZE, RTR0MEMOBJTYPE_LOW);
630 0 /* PhysMask */, _4G - PAGE_SIZE, RTR0MEMOBJTYPE_LOW);
638 ~(uint32_t)PAGE_OFFSET_MASK, _4G - PAGE_SIZE,
643 * cb <= PAGE_SIZE allocations take a different path, using a different allocator.
645 if (RT_FAILURE(rc) && cb <= PAGE_SIZE)
646 rc = rtR0MemObjNativeAllocWorker(ppMem, cb + PAGE_SIZE, fExecutable, true /* fContiguous */,
647 ~(uint32_t)PAGE_OFFSET_MASK, _4G - PAGE_SIZE,
656 if (uAlignment != PAGE_SIZE)
856 if (uAlignment > PAGE_SIZE)
959 if (uAlignment > PAGE_SIZE)
1120 PgNo = pmap_find_phys(kernel_pmap, (uintptr_t)pMemDarwin->Core.pv + iPage * PAGE_SIZE);
1149 PgNo = pmap_find_phys(Pmap, (uintptr_t)pMemDarwin->Core.pv + iPage * PAGE_SIZE);
1171 addr64_t Addr = pMemDesc->getPhysicalSegment(iPage * PAGE_SIZE, NULL, kIOMemoryMapperNone);
1173 addr64_t Addr = pMemDesc->getPhysicalSegment64(iPage * PAGE_SIZE, NULL);