/dragonfly/sys/platform/vkernel64/platform/ |
H A D | copyio.c | 67 kva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)); in casu64() 105 kva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)); in casu32() 142 kva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)); in swapu64() 174 kva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)); in swapu32() 206 kva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)); in fuwordadd64() 238 kva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)); in fuwordadd32()
|
H A D | pmap.c | 366 pa = VM_PAGE_TO_PHYS(m) | (va & PAGE_MASK); in uservtophys() 903 atomic_swap_long(ptep, VM_PAGE_TO_PHYS(*m) | in _pmap_qenter() 1275 pml4 = (pml4_entry_t *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m4)); in pmap_release_free_page() 1295 pdp = (pdp_entry_t *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m3)); in pmap_release_free_page() 1313 pd = (pd_entry_t *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m2)); in pmap_release_free_page() 1326 pmap, p, (void *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(p)), in pmap_release_free_page() 1381 data = VM_PAGE_TO_PHYS(m) | in _pmap_allocpte() 1573 paddr = VM_PAGE_TO_PHYS(nkpg); in pmap_growkernel() 1605 ptppaddr = VM_PAGE_TO_PHYS(nkpg); in pmap_growkernel() 1665 vpte = VM_PAGE_TO_PHYS(vmspace_pmap(vm)->pm_pdirm) | in cpu_vmspace_alloc() [all …]
|
/dragonfly/sys/dev/drm/include/linux/ |
H A D | highmem.h | 51 return (void *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS( (struct vm_page *)pg )); in kmap() 61 return (void *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS( (struct vm_page *)pg )); in kmap_atomic()
|
H A D | dma-mapping.h | 45 return VM_PAGE_TO_PHYS((struct vm_page *)page) + offset; in dma_map_page()
|
H A D | mm.h | 182 return (void *)VM_PAGE_TO_PHYS((const struct vm_page *)page); in page_address()
|
/dragonfly/sys/vm/ |
H A D | vm_contig.c | 305 phys = VM_PAGE_TO_PHYS(m); in vm_contig_pg_alloc() 327 ((VM_PAGE_TO_PHYS(&pga[i]) + size) > high)) { in vm_contig_pg_alloc() 375 if ((VM_PAGE_TO_PHYS(&m[0]) != in vm_contig_pg_alloc() 376 (VM_PAGE_TO_PHYS(&m[-1]) + PAGE_SIZE)) || in vm_contig_pg_alloc() 524 pa = VM_PAGE_TO_PHYS(&pga[start]); in vm_contig_pg_kmap()
|
H A D | vm_page2.h | 542 pmap_zero_page(VM_PAGE_TO_PHYS(m)); in vm_page_zero_fill() 553 pmap_copy_page(VM_PAGE_TO_PHYS(src_m), VM_PAGE_TO_PHYS(dest_m)); in vm_page_copy() 576 char *p = (char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)); in vm_page_free_zero()
|
H A D | vm_object.c | 1912 (u_long) p->pindex, (u_long) VM_PAGE_TO_PHYS(p)); in DB_SHOW_COMMAND() 1992 (VM_PAGE_TO_PHYS(m) == pa + rcount * PAGE_SIZE)) { in DB_SHOW_COMMAND() 1997 padiff = pa + rcount * PAGE_SIZE - VM_PAGE_TO_PHYS(m); in DB_SHOW_COMMAND() 2001 pa = VM_PAGE_TO_PHYS(m) - rcount * PAGE_SIZE; in DB_SHOW_COMMAND() 2017 pa = VM_PAGE_TO_PHYS(m); in DB_SHOW_COMMAND()
|
H A D | vm_page.c | 826 vm_low_phys_reserved = VM_PAGE_TO_PHYS(m); in vm_page_startup_finish() 2860 vm_paddr_t pa = VM_PAGE_TO_PHYS(m); in vm_page_free_contig() 3755 pmap_zero_page(VM_PAGE_TO_PHYS(m)); in vm_page_alloczwq() 3881 pmap_zero_page(VM_PAGE_TO_PHYS(m)); in vm_page_grab() 3885 pmap_zero_page(VM_PAGE_TO_PHYS(m)); in vm_page_grab() 3956 VM_PAGE_TO_PHYS(m), in _vm_page_zero_valid() 3974 VM_PAGE_TO_PHYS(m), in _vm_page_zero_valid() 4139 VM_PAGE_TO_PHYS(m), in vm_page_zero_invalid()
|
H A D | vm_page.h | 442 #define VM_PAGE_TO_PHYS(entry) \ macro
|
/dragonfly/sys/dev/drm/include/asm/ |
H A D | cacheflush.h | 58 pmap_change_attr(PHYS_TO_DMAP(VM_PAGE_TO_PHYS(p)), in set_pages_uc() 68 pmap_change_attr(PHYS_TO_DMAP(VM_PAGE_TO_PHYS(p)), in set_pages_wb()
|
H A D | memory_model.h | 37 return OFF_TO_IDX(VM_PAGE_TO_PHYS(p)); in page_to_pfn()
|
H A D | page.h | 39 return VM_PAGE_TO_PHYS(p); in page_to_phys()
|
H A D | pci-dma-compat.h | 37 return VM_PAGE_TO_PHYS((struct vm_page *)page) + offset; in pci_map_page()
|
/dragonfly/sys/dev/drm/ttm/ |
H A D | ttm_tt.c | 386 pmap_copy_page(VM_PAGE_TO_PHYS((struct vm_page *)from_page), in ttm_tt_swapin() 387 VM_PAGE_TO_PHYS((struct vm_page *)to_page)); in ttm_tt_swapin() 433 pmap_copy_page(VM_PAGE_TO_PHYS(from_page), in ttm_tt_swapout() 434 VM_PAGE_TO_PHYS(to_page)); in ttm_tt_swapout()
|
/dragonfly/sys/cpu/x86_64/include/ |
H A D | lwbuf.h | 83 lwb->kva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(lwb->m)); in lwbuf_alloc()
|
/dragonfly/sys/cpu/x86_64/misc/ |
H A D | lwbuf.c | 95 lwb->kva = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(lwb->m));
|
/dragonfly/sys/dev/virtual/nvmm/ |
H A D | nvmm_netbsd.c | 173 return VM_PAGE_TO_PHYS(pg); in os_pa_zalloc() 195 _pa = VM_PAGE_TO_PHYS(TAILQ_FIRST(&pglist)); in os_contigpa_zalloc()
|
H A D | nvmm_dragonfly.c | 242 return VM_PAGE_TO_PHYS(pg); in os_pa_zalloc()
|
/dragonfly/sys/vfs/procfs/ |
H A D | procfs_mem.c | 147 pmap_kenter_quick(kva, VM_PAGE_TO_PHYS(m)); in procfs_rwmem()
|
/dragonfly/sys/dev/drm/ |
H A D | linux_shmem.c | 76 pmap_zero_page(VM_PAGE_TO_PHYS(m)); in shmem_read_mapping_page()
|
/dragonfly/sys/platform/pc64/x86_64/ |
H A D | pmap.c | 749 phys = VM_PAGE_TO_PHYS(pv->pv_m); in pmap_pt() 2082 daddr = PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pages[i])); in pmap_invalidate_cache_pages() 2141 pte = VM_PAGE_TO_PHYS(*m) | in _pmap_qenter() 2382 VM_PAGE_TO_PHYS(pv->pv_m)); in pmap_pinit() 2431 VM_PAGE_TO_PHYS(pv->pv_m)); in pmap_pinit() 2754 pmap_zero_page(VM_PAGE_TO_PHYS(m)); in pmap_allocpte() 2778 v = VM_PAGE_TO_PHYS(m) | in pmap_allocpte() 2816 v = VM_PAGE_TO_PHYS(pv->pv_m) | in pmap_allocpte() 3389 paddr = VM_PAGE_TO_PHYS(nkpg); in pmap_growkernel() 3431 ptppaddr = VM_PAGE_TO_PHYS(nkpg); in pmap_growkernel() [all …]
|
/dragonfly/sys/dev/agp/ |
H A D | intel-gtt.c | 1196 mem->am_physical = VM_PAGE_TO_PHYS(m); in agp_i810_alloc_memory() 1385 VM_PAGE_TO_PHYS(bogus_page), 0); in agp_intel_gtt_clear_range() 1401 VM_PAGE_TO_PHYS(pages[i]), flags); in agp_intel_gtt_insert_pages() 1449 res.scratch_page_dma = VM_PAGE_TO_PHYS(bogus_page); in agp_intel_gtt_get()
|
H A D | agp.c | 546 AGP_DPF("found page pa=%#jx\n", (uintmax_t)VM_PAGE_TO_PHYS(m)); in agp_generic_bind_memory() 557 vm_offset_t pa = VM_PAGE_TO_PHYS(m) + j; in agp_generic_bind_memory()
|
/dragonfly/sys/dev/virtual/virtio/balloon/ |
H A D | virtio_balloon.c | 643 VM_PAGE_TO_PHYS(m) >> VIRTIO_BALLOON_PFN_SHIFT; in vtballoon_inflate() 673 VM_PAGE_TO_PHYS(m) >> VIRTIO_BALLOON_PFN_SHIFT; in vtballoon_deflate()
|