/freebsd/sys/arm64/arm64/ |
H A D | pmap.c | 9198 pmap_san_enter_bootstrap_alloc_pages(int npages) in pmap_san_enter_bootstrap_alloc_pages() argument 9204 if (offset + (npages * PAGE_SIZE) > sizeof(bootstrap_data)) { in pmap_san_enter_bootstrap_alloc_pages() 9210 offset += (npages * PAGE_SIZE); in pmap_san_enter_bootstrap_alloc_pages()
|
H A D | gicv3_its.c | 507 size_t its_tbl_size, nitspages, npages; in gicv3_its_table_init() local 601 npages = howmany(its_tbl_size, PAGE_SIZE); in gicv3_its_table_init() 604 table = contigmalloc_domainset(npages * PAGE_SIZE, in gicv3_its_table_init()
|
/freebsd/sys/vm/ |
H A D | swap_pager.c | 826 int mpages, npages; in swp_pager_getswapspace() local 844 if (npages == 1) in swp_pager_getswapspace() 846 mpages = npages - 1; in swp_pager_getswapspace() 847 npages >>= 1; in swp_pager_getswapspace() 851 *io_npages = npages; in swp_pager_getswapspace() 853 sp->sw_used += npages; in swp_pager_getswapspace() 854 swap_pager_avail -= npages; in swp_pager_getswapspace() 913 if (npages == 0) in swp_pager_freeswapspace() 918 sp->sw_used -= npages; in swp_pager_freeswapspace() 926 npages); in swp_pager_freeswapspace() [all …]
|
H A D | device_pager.c | 465 unsigned int npages; in old_dev_pager_ctor() local 482 npages = OFF_TO_IDX(size); in old_dev_pager_ctor() 484 for (off = foff; npages--; off += PAGE_SIZE) { in old_dev_pager_ctor()
|
H A D | vm_extern.h | 130 vm_pindex_t vm_kstack_pindex(vm_offset_t ks, int npages); 131 vm_object_t vm_thread_kstack_size_to_obj(int npages); 132 int vm_thread_stack_back(vm_offset_t kaddr, vm_page_t ma[], int npages,
|
H A D | vm_glue.c | 183 vm_size_t npages; in vslock() local 191 npages = atop(end - start); in vslock() 192 if (npages > vm_page_max_user_wired) in vslock() 605 vm_thread_stack_back(vm_offset_t ks, vm_page_t ma[], int npages, int req_class, in vm_thread_stack_back() argument 608 vm_object_t obj = vm_thread_kstack_size_to_obj(npages); in vm_thread_stack_back() 613 pindex = vm_kstack_pindex(ks, npages); in vm_thread_stack_back() 616 for (n = 0; n < npages;) { in vm_thread_stack_back() 627 if (n < npages) in vm_thread_stack_back() 644 vm_thread_kstack_size_to_obj(int npages) in vm_thread_kstack_size_to_obj() argument 646 return (npages == kstack_pages ? kstack_object : kstack_alt_object); in vm_thread_kstack_size_to_obj()
|
H A D | vm_page.h | 611 u_long npages, vm_paddr_t low, vm_paddr_t high, u_long alignment, 614 vm_pindex_t pindex, int domain, int req, u_long npages, vm_paddr_t low, 621 vm_page_t vm_page_alloc_noobj_contig(int req, u_long npages, vm_paddr_t low, 624 vm_page_t vm_page_alloc_noobj_contig_domain(int domain, int req, u_long npages, 663 int vm_page_reclaim_contig(int req, u_long npages, vm_paddr_t low, 665 int vm_page_reclaim_contig_domain(int domain, int req, u_long npages, 667 int vm_page_reclaim_contig_domain_ext(int domain, int req, u_long npages,
|
H A D | vm_phys.c | 498 u_long npages; in vm_phys_init() local 510 npages = 0; in vm_phys_init() 556 npages = 0; in vm_phys_init() 740 m, npages)); in vm_phys_enq_range() 741 while (npages > 0) { in vm_phys_enq_range() 806 if (i == npages) { in vm_phys_alloc_npages() 812 return (npages); in vm_phys_alloc_npages() 835 return (npages); in vm_phys_alloc_npages() 1191 m_end = m + npages; in vm_phys_enqueue_contig() 1234 m_end = m + npages; in vm_phys_free_contig() [all …]
|
H A D | vm_phys.h | 62 vm_page_t vm_phys_alloc_contig(int domain, u_long npages, vm_paddr_t low, 66 int vm_phys_alloc_npages(int domain, int pool, int npages, vm_page_t ma[]); 69 void vm_phys_enqueue_contig(vm_page_t m, u_long npages); 75 u_long npages, vm_paddr_t low, vm_paddr_t high); 76 void vm_phys_free_contig(vm_page_t m, u_long npages);
|
H A D | vm_reserv.c | 588 pindex + npages > object->size) in vm_reserv_alloc_contig() 602 size = npages << PAGE_SHIFT; in vm_reserv_alloc_contig() 633 for (i = 0; i < npages; i++) in vm_reserv_alloc_contig() 717 vm_domain_freecnt_inc(vmd, npages); in vm_reserv_alloc_contig() 743 npages -= n; in vm_reserv_alloc_contig() 1071 KASSERT(npages <= VM_LEVEL_0_NPAGES, in vm_reserv_is_populated() 1073 KASSERT(powerof2(npages), in vm_reserv_is_populated() 1192 KASSERT(npages <= ppn_bound, in vm_reserv_find_contig() 1231 if (npages > VM_LEVEL_0_NPAGES - 1) in vm_reserv_reclaim_contig() 1233 size = npages << PAGE_SHIFT; in vm_reserv_reclaim_contig() [all …]
|
H A D | vm_reserv.h | 49 int domain, int req, vm_page_t mpred, u_long npages, 58 bool vm_reserv_is_populated(vm_page_t m, int npages); 61 vm_page_t vm_reserv_reclaim_contig(int domain, u_long npages,
|
/freebsd/sys/kern/ |
H A D | kern_physio.c | 48 int error, i, npages, maxpages; in physio() local 52 npages = 0; in physio() 158 if ((npages = vm_fault_quick_hold_pages( in physio() 168 pages, npages); in physio() 172 bp->bio_ma_n = npages; in physio() 188 pmap_qremove((vm_offset_t)sa, npages); in physio() 189 vm_page_unhold_pages(pages, npages); in physio()
|
H A D | subr_asan.c | 106 size_t sz, npages, i; in kasan_shadow_map() local 120 npages = (eva - sva) / PAGE_SIZE; in kasan_shadow_map() 125 for (i = 0; i < npages; i++) in kasan_shadow_map()
|
H A D | uipc_ktls.c | 3010 int error, npages, total_pages; in ktls_encrypt() local 3020 npages = 0; in ktls_encrypt() 3057 npages += m->m_epg_nrdy; in ktls_encrypt() 3088 int npages; in ktls_encrypt_cb() local 3112 npages = m->m_epg_nrdy; in ktls_encrypt_cb() 3115 (void)so->so_proto->pr_ready(so, m, npages); in ktls_encrypt_cb() 3118 mb_free_notready(m, npages); in ktls_encrypt_cb() 3137 int error, mpages, npages, total_pages; in ktls_encrypt_async() local 3147 npages = 0; in ktls_encrypt_async() 3176 npages += mpages; in ktls_encrypt_async() [all …]
|
H A D | vfs_aio.c | 1287 int npages; in aio_qbio() local 1318 npages = vm_fault_quick_hold_pages(&curproc->p_vmspace->vm_map, in aio_qbio() 1321 if (npages < 0) { in aio_qbio() 1332 pmap_qenter((vm_offset_t)pbuf->b_data, pages, npages); in aio_qbio() 1334 pbuf->b_npages = npages; in aio_qbio() 1338 bp->bio_ma_n = npages; in aio_qbio()
|
H A D | vfs_bio.c | 178 static void vfs_vmio_truncate(struct buf *bp, int npages); 179 static void vfs_vmio_extend(struct buf *bp, int npages, int size); 5101 int npages; in vunmapbuf() local 5103 npages = bp->b_npages; in vunmapbuf() 5105 pmap_qremove(trunc_page((vm_offset_t)bp->b_data), npages); in vunmapbuf() 5106 vm_page_unhold_pages(bp->b_pages, npages); in vunmapbuf()
|
/freebsd/sys/dev/md/ |
H A D | md.c | 887 int ma_offs, npages; in mdstart_vnode() local 960 npages = atop(min(maxphys, round_page(len + (ma_offs & in mdstart_vnode() 962 iolen = min(ptoa(npages) - (ma_offs & PAGE_MASK), len); in mdstart_vnode() 964 KASSERT(npages <= atop(maxphys + PAGE_SIZE), in mdstart_vnode() 965 ("npages %d too large", npages)); in mdstart_vnode() 966 pmap_qenter(sc->kva, &bp->bio_ma[atop(ma_offs)], npages); in mdstart_vnode() 1001 pmap_qremove(sc->kva, npages); in mdstart_vnode()
|
/freebsd/sys/fs/nfsclient/ |
H A D | nfs_clbio.c | 116 int i, error, nextoff, size, toff, count, npages; in ncl_getpages() local 135 npages = ap->a_count; in ncl_getpages() 173 if (!vm_page_none_valid(pages[npages - 1]) && --npages == 0) in ncl_getpages() 184 pmap_qenter(kva, pages, npages); in ncl_getpages() 186 VM_CNT_ADD(v_vnodepgsin, npages); in ncl_getpages() 188 count = npages << PAGE_SHIFT; in ncl_getpages() 200 pmap_qremove(kva, npages); in ncl_getpages() 267 int i, error, npages, count; in ncl_putpages() local 289 npages = btoc(count); in ncl_putpages() 317 for (i = 0; i < npages; i++) in ncl_putpages() [all …]
|
/freebsd/sys/dev/mlx5/mlx5_en/ |
H A D | mlx5_en_main.c | 1347 sizeof(u64) * rq->wq_ctrl.buf.npages; in mlx5e_enable_rq() 1739 sizeof(u64) * sq->wq_ctrl.buf.npages; in mlx5e_enable_sq() 2035 sizeof(u64) * cq->wq_ctrl.buf.npages; in mlx5e_enable_cq()
|
/freebsd/sys/contrib/openzfs/module/os/linux/zfs/ |
H A D | vdev_disk.c | 867 uint_t npages; member 891 if (s->npages != 0 && (off & s->bmask) != 0) in vdev_disk_check_pages_cb() 894 s->npages++; in vdev_disk_check_pages_cb() 907 .npages = 0, in vdev_disk_check_pages()
|
/freebsd/sys/contrib/openzfs/tests/zfs-tests/tests/functional/vdev_disk/ |
H A D | page_alignment.c | 43 uint32_t npages; member 67 if (s->npages != 0 && (off & s->bmask) != 0) in vdev_disk_check_pages_cb() 70 s->npages++; in vdev_disk_check_pages_cb() 350 .npages = 0, in run_test()
|
/freebsd/sys/compat/linuxkpi/common/include/linux/ |
H A D | mm.h | 317 #define unpin_user_pages(pages, npages) release_pages(pages, npages) argument
|
/freebsd/sys/dev/cxgbe/tom/ |
H A D | t4_ddp.c | 121 for (i = 0; i < ps->npages; i++) { in free_pageset() 1429 for (i = 0; i < npages; i++) { in t4_alloc_page_pods_for_vmpages() 1431 while (i < npages - 1 && in t4_alloc_page_pods_for_vmpages() 1491 int hcf, seglen, idx, npages, nppods; in t4_alloc_page_pods_for_buf() local 1535 npages = 1; in t4_alloc_page_pods_for_buf() 1613 npages = 0; in t4_alloc_page_pods_for_sgl() 1615 npages++; in t4_alloc_page_pods_for_sgl() 1707 if (idx < ps->npages) { in t4_write_page_pods_for_ps() 2189 if (ps->start != start || ps->npages != npages || in pscmp() 2279 if (ps->npages < 0) { in hold_aio() [all …]
|
H A D | t4_tom.h | 148 int npages; member
|
H A D | t4_cpl_io.c | 2098 int i, mlen, npages, pgoff; in alloc_aiotx_mbuf() local 2122 npages = vm_fault_quick_hold_pages(map, start, mlen, in alloc_aiotx_mbuf() 2124 if (npages < 0) in alloc_aiotx_mbuf() 2129 vm_page_unhold_pages(pgs, npages); in alloc_aiotx_mbuf() 2134 m->m_epg_npgs = npages; in alloc_aiotx_mbuf() 2135 if (npages == 1) { in alloc_aiotx_mbuf() 2142 (npages - 2) * PAGE_SIZE; in alloc_aiotx_mbuf() 2144 for (i = 0; i < npages; i++) in alloc_aiotx_mbuf() 2148 m->m_ext.ext_size = npages * PAGE_SIZE; in alloc_aiotx_mbuf() 2154 __func__, jobtotid(job), m, job, npages); in alloc_aiotx_mbuf()
|