/dragonfly/lib/libu4bhid/ |
H A D | usage.c | 53 } *pages; variable 63 printf("%d\t%s\n", pages[i].usage, pages[i].name); in dump_hid_table() 126 if (pages == NULL) { in hid_init() 132 pages = realloc(pages, in hid_init() 136 if (!pages) in hid_init() 163 if (!pages) in hid_usage_page() 167 if (pages[k].usage == i) in hid_usage_page() 168 return pages[k].name; in hid_usage_page() 207 if (!pages) in hid_parse_usage_page() 212 return pages[k].usage; in hid_parse_usage_page() [all …]
|
/dragonfly/contrib/binutils-2.27/libiberty/ |
H A D | physmem.c | 90 return pages * pagesize; in physmem_total() 101 if (0 <= pages && 0 <= pagesize) in physmem_total() 102 return pages * pagesize; in physmem_total() 113 double pages = realmem.physmem; in physmem_total() local 114 if (0 <= pages && 0 <= pagesize) in physmem_total() 115 return pages * pagesize; in physmem_total() 194 return pages * pagesize; in physmem_available() 205 double pages = psd.psd_free; in physmem_available() local 208 return pages * pagesize; in physmem_available() 221 return pages * pagesize; in physmem_available() [all …]
|
/dragonfly/contrib/binutils-2.34/libiberty/ |
H A D | physmem.c | 90 return pages * pagesize; in physmem_total() 101 if (0 <= pages && 0 <= pagesize) in physmem_total() 102 return pages * pagesize; in physmem_total() 113 double pages = realmem.physmem; in physmem_total() local 114 if (0 <= pages && 0 <= pagesize) in physmem_total() 115 return pages * pagesize; in physmem_total() 194 return pages * pagesize; in physmem_available() 205 double pages = psd.psd_free; in physmem_available() local 208 return pages * pagesize; in physmem_available() 221 return pages * pagesize; in physmem_available() [all …]
|
/dragonfly/contrib/gcc-4.7/libiberty/ |
H A D | physmem.c | 90 return pages * pagesize; in physmem_total() 101 if (0 <= pages && 0 <= pagesize) in physmem_total() 102 return pages * pagesize; in physmem_total() 113 double pages = realmem.physmem; in physmem_total() local 114 if (0 <= pages && 0 <= pagesize) in physmem_total() 115 return pages * pagesize; in physmem_total() 194 return pages * pagesize; in physmem_available() 205 double pages = psd.psd_free; in physmem_available() local 208 return pages * pagesize; in physmem_available() 221 return pages * pagesize; in physmem_available() [all …]
|
/dragonfly/contrib/gcc-8.0/libiberty/ |
H A D | physmem.c | 90 return pages * pagesize; in physmem_total() 101 if (0 <= pages && 0 <= pagesize) in physmem_total() 102 return pages * pagesize; in physmem_total() 113 double pages = realmem.physmem; in physmem_total() local 114 if (0 <= pages && 0 <= pagesize) in physmem_total() 115 return pages * pagesize; in physmem_total() 194 return pages * pagesize; in physmem_available() 205 double pages = psd.psd_free; in physmem_available() local 208 return pages * pagesize; in physmem_available() 221 return pages * pagesize; in physmem_available() [all …]
|
/dragonfly/sys/dev/drm/include/drm/ttm/ |
H A D | ttm_set_memory.h | 40 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() argument 42 return set_pages_array_wb(pages, addrinarray); in ttm_set_pages_array_wb() 45 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc() argument 47 return set_pages_array_wc(pages, addrinarray); in ttm_set_pages_array_wc() 50 static inline int ttm_set_pages_array_uc(struct page **pages, int addrinarray) in ttm_set_pages_array_uc() argument 52 return set_pages_array_uc(pages, addrinarray); in ttm_set_pages_array_uc() 78 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() argument 83 unmap_page_from_agp(pages[i]); in ttm_set_pages_array_wb() 87 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc() argument 92 map_page_into_agp(pages[i]); in ttm_set_pages_array_wc() [all …]
|
/dragonfly/sys/dev/drm/ttm/ |
H A D | ttm_page_alloc.c | 269 __free_pages(pages[i], order); in ttm_pages_put() 643 struct pglist *pages, in ttm_page_pool_get_pages() argument 688 if (pages[i]) { in ttm_put_pages() 693 __free_page(pages[i]); in ttm_put_pages() 694 pages[i] = NULL; in ttm_put_pages() 702 if (pages[i]) { in ttm_put_pages() 705 pages[i] = NULL; in ttm_put_pages() 755 pages[r] = (struct page *)p; in ttm_get_pages() 878 if (!ttm->pages[i]) in ttm_pool_unpopulate_helper() 952 if (++p != tt->ttm.pages[j]) in ttm_populate_and_map_pages() [all …]
|
H A D | ttm_tt.c | 94 if (!ttm->pages) in ttm_tt_alloc_page_directory() 102 sizeof(*ttm->ttm.pages) + in ttm_dma_tt_alloc_page_directory() 105 if (!ttm->ttm.pages) in ttm_dma_tt_alloc_page_directory() 174 cur_page = ttm->pages[i]; in ttm_tt_set_caching() 190 cur_page = ttm->pages[j]; in ttm_tt_set_caching() 261 kvfree(ttm->pages); in ttm_tt_fini() 262 ttm->pages = NULL; in ttm_tt_fini() 307 if (ttm->pages) in ttm_dma_tt_fini() 308 kvfree(ttm->pages); in ttm_dma_tt_fini() 311 ttm->pages = NULL; in ttm_dma_tt_fini() [all …]
|
/dragonfly/contrib/mdocml/ |
H A D | dbm.c | 46 int32_t pages; member 68 static struct page *pages; variable 152 res.name = dbm_get(pages[ip].name); in dbm_page_get() 155 res.sect = dbm_get(pages[ip].sect); in dbm_page_get() 158 res.arch = pages[ip].arch ? dbm_get(pages[ip].arch) : NULL; in dbm_page_get() 165 res.addr = dbm_addr(pages + ip); in dbm_page_get() 250 cp = dbm_get(pages[0].name); in page_bytitle() 253 cp = dbm_get(pages[0].sect); in page_bytitle() 256 cp = dbm_get(pages[0].desc); in page_bytitle() 332 if (pages[ip].arch) in page_byarch() [all …]
|
H A D | dba.c | 45 struct dba_array *pages; member 93 dba_array_free(entry->pages); in dba_free() 101 dba_array_undel(dba->pages); in dba_free() 110 dba_array_free(dba->pages); in dba_free() 134 dba_pages_write(dba->pages); in dba_write() 178 dba_array_add(pages, page); in dba_page_new() 277 dba_array_FOREACH(pages, page) { in dba_pages_write() 283 dba_array_FOREACH(pages, page) { in dba_pages_write() 289 dba_array_FOREACH(pages, page) { in dba_pages_write() 307 dba_array_FOREACH(pages, page) in dba_pages_write() [all …]
|
/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_gart.c | 220 int pages) in amdgpu_gart_unbind() argument 236 for (i = 0; i < pages; i++, p++) { in amdgpu_gart_unbind() 238 adev->gart.pages[p] = NULL; in amdgpu_gart_unbind() 268 int pages, dma_addr_t *dma_addr, uint64_t flags, in amdgpu_gart_map() argument 281 for (i = 0; i < pages; i++) { in amdgpu_gart_map() 321 for (i = 0; i < pages; i++, p++) in amdgpu_gart_bind() 322 adev->gart.pages[p] = pagelist ? pagelist[i] : NULL; in amdgpu_gart_bind() 370 adev->gart.pages = vzalloc(array_size(sizeof(void *), in amdgpu_gart_init() 372 if (adev->gart.pages == NULL) in amdgpu_gart_init() 389 vfree(adev->gart.pages); in amdgpu_gart_fini() [all …]
|
H A D | amdgpu_vram_mgr.c | 112 unsigned pages = mem->num_pages; in amdgpu_vram_mgr_bo_visible_size() local 121 for (usage = 0; nodes && pages; pages -= nodes->size, nodes++) in amdgpu_vram_mgr_bo_visible_size() 180 unsigned long pages = min(pages_left, pages_per_node); in amdgpu_vram_mgr_new() local 184 if (pages == pages_per_node) in amdgpu_vram_mgr_new() 188 pages, alignment, 0, in amdgpu_vram_mgr_new() 206 pages_left -= pages; in amdgpu_vram_mgr_new() 243 unsigned pages = mem->num_pages; in amdgpu_vram_mgr_del() local 249 while (pages) { in amdgpu_vram_mgr_del() 250 pages -= nodes->size; in amdgpu_vram_mgr_del()
|
H A D | amdgpu_gart.h | 50 struct page **pages; member 65 int pages); 67 int pages, dma_addr_t *dma_addr, uint64_t flags, 70 int pages, struct page **pagelist,
|
/dragonfly/sys/dev/drm/radeon/ |
H A D | radeon_gart.c | 242 int pages) in radeon_gart_unbind() argument 254 for (i = 0; i < pages; i++, p++) { in radeon_gart_unbind() 255 if (rdev->gart.pages[p]) { in radeon_gart_unbind() 256 rdev->gart.pages[p] = NULL; in radeon_gart_unbind() 287 int pages, struct page **pagelist, dma_addr_t *dma_addr, in radeon_gart_bind() argument 302 for (i = 0; i < pages; i++, p++) { in radeon_gart_bind() 303 rdev->gart.pages[p] = pagelist[i]; in radeon_gart_bind() 333 if (rdev->gart.pages) { in radeon_gart_init() 351 if (rdev->gart.pages == NULL) { in radeon_gart_init() 381 vfree(rdev->gart.pages); in radeon_gart_fini() [all …]
|
/dragonfly/sys/dev/drm/include/asm/ |
H A D | cacheflush.h | 75 set_pages_array_uc(struct page **pages, int addrinarray) in set_pages_array_uc() argument 78 pmap_page_set_memattr((struct vm_page *)pages[i], VM_MEMATTR_UNCACHEABLE); in set_pages_array_uc() 84 set_pages_array_wb(struct page **pages, int addrinarray) in set_pages_array_wb() argument 87 pmap_page_set_memattr((struct vm_page *)pages[i], VM_MEMATTR_WRITE_BACK); in set_pages_array_wb() 93 set_pages_array_wc(struct page **pages, int addrinarray) in set_pages_array_wc() argument 96 pmap_page_set_memattr((struct vm_page *)pages[i], VM_MEMATTR_WRITE_COMBINING); in set_pages_array_wc()
|
/dragonfly/sys/dev/drm/ |
H A D | linux_shmem.c | 123 unsigned int gup_flags, struct page **pages, in get_user_pages() argument 158 pages[i] = (void *)m; in get_user_pages() 162 put_page(pages[i]); in get_user_pages() 163 pages[i] = NULL; in get_user_pages() 171 release_pages(struct page **pages, unsigned long nr_pages) in release_pages() argument 175 put_page(pages[nr_pages]); in release_pages() 176 pages[nr_pages] = NULL; in release_pages()
|
H A D | drm_cache.c | 60 static void drm_cache_flush_clflush(struct page *pages[], in drm_cache_flush_clflush() argument 67 drm_clflush_page(*pages++); in drm_cache_flush_clflush() 73 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages() argument 78 drm_cache_flush_clflush(pages, num_pages); in drm_clflush_pages() 87 struct page *page = pages[i]; in drm_clflush_pages()
|
H A D | drm_scatter.c | 44 kmem_free(kernel_map, entry->vaddr, IDX_TO_OFF(entry->pages)); in drm_sg_cleanup() 80 entry->pages = OFF_TO_IDX(size); in drm_legacy_sg_alloc() 81 entry->busaddr = kmalloc(entry->pages * sizeof(*entry->busaddr), in drm_legacy_sg_alloc() 93 for(pindex = 0; pindex < entry->pages; pindex++) { in drm_legacy_sg_alloc() 110 entry->pages, (uintmax_t)entry->vaddr, in drm_legacy_sg_alloc()
|
/dragonfly/sys/dev/drm/i915/ |
H A D | i915_gem_dmabuf.c | 58 ret = sg_alloc_table(st, obj->mm.pages->nents, GFP_KERNEL); in i915_gem_map_dma_buf() 62 src = obj->mm.pages->sgl; in i915_gem_map_dma_buf() 64 for (i = 0; i < obj->mm.pages->nents; i++) { in i915_gem_map_dma_buf() 265 struct sg_table *pages; in i915_gem_object_get_pages_dmabuf() local 268 pages = dma_buf_map_attachment(obj->base.import_attach, in i915_gem_object_get_pages_dmabuf() 270 if (IS_ERR(pages)) in i915_gem_object_get_pages_dmabuf() 271 return PTR_ERR(pages); in i915_gem_object_get_pages_dmabuf() 273 sg_page_sizes = i915_sg_page_sizes(pages->sgl); in i915_gem_object_get_pages_dmabuf() 275 __i915_gem_object_set_pages(obj, pages, sg_page_sizes); in i915_gem_object_get_pages_dmabuf() 281 struct sg_table *pages) in i915_gem_object_put_pages_dmabuf() argument [all …]
|
H A D | i915_gem_userptr.c | 545 if (!IS_ERR(pages)) { 547 pages = NULL; 552 if (IS_ERR(pages)) 613 struct sg_table *pages; in i915_gem_userptr_get_pages() 659 pages = ERR_PTR(pinned); in i915_gem_userptr_get_pages() 666 active = !IS_ERR(pages); in i915_gem_userptr_get_pages() 671 if (IS_ERR(pages)) in i915_gem_userptr_get_pages() 675 return PTR_ERR_OR_ZERO(pages); in i915_gem_userptr_get_pages() 682 struct sg_table *pages) in i915_gem_userptr_put_pages() argument 704 sg_free_table(pages); in i915_gem_userptr_put_pages() [all …]
|
/dragonfly/lib/libnvmm/ |
H A D | libnvmm.c | 223 struct nvmm_comm_page **pages; in nvmm_machine_create() local 231 pages = calloc(__capability.max_vcpus, sizeof(*pages)); in nvmm_machine_create() 232 if (pages == NULL) { in nvmm_machine_create() 240 free(pages); in nvmm_machine_create() 248 mach->pages = pages; in nvmm_machine_create() 267 free(mach->pages); in nvmm_machine_destroy() 304 mach->pages[cpuid] = args.comm; in nvmm_vcpu_create() 333 comm = mach->pages[vcpu->cpuid]; in nvmm_vcpu_destroy() 372 comm = mach->pages[vcpu->cpuid]; in nvmm_vcpu_setstate() 387 comm = mach->pages[vcpu->cpuid]; in nvmm_vcpu_getstate() [all …]
|
/dragonfly/sys/platform/vkernel64/platform/ |
H A D | busdma_machdep.c | 263 int pages; in bus_dma_tag_create() local 266 pages = 1; in bus_dma_tag_create() 270 pages = MAX(pages, 1); in bus_dma_tag_create() 274 if (alloc_bounce_pages(newtag, pages, flags) < pages) in bus_dma_tag_create() 351 int pages; in bus_dmamap_create() local 354 pages = 1; in bus_dmamap_create() 357 pages = MIN(maxpages - bz->total_bpages, pages); in bus_dmamap_create() 358 pages = MAX(pages, 1); in bus_dmamap_create() 360 if (alloc_bounce_pages(dmat, pages, flags) < pages) in bus_dmamap_create() 1075 int pages; in reserve_bounce_pages() local [all …]
|
/dragonfly/stand/boot/efi/loader/ |
H A D | copy.c | 56 size_t pages = STAGE_PAGES; in efi_copy_init() local 59 pages, &staging); in efi_copy_init() 65 pages /= 2; in efi_copy_init() 67 pages, &staging); in efi_copy_init() 74 staging_end = staging + pages * EFI_PAGE_SIZE; in efi_copy_init()
|
/dragonfly/sys/platform/pc64/x86_64/ |
H A D | busdma_machdep.c | 305 int pages; in bus_dma_tag_create() local 308 pages = 1; in bus_dma_tag_create() 312 pages = MAX(pages, 1); in bus_dma_tag_create() 316 if (alloc_bounce_pages(newtag, pages, flags) < pages) in bus_dma_tag_create() 403 int pages; in bus_dmamap_create() local 406 pages = 1; in bus_dmamap_create() 409 pages = MIN(maxpages - bz->total_bpages, pages); in bus_dmamap_create() 410 pages = MAX(pages, 1); in bus_dmamap_create() 412 if (alloc_bounce_pages(dmat, pages, flags) < pages) in bus_dmamap_create() 1266 int pages; in reserve_bounce_pages() local [all …]
|
/dragonfly/sys/dev/drm/include/drm/ |
H A D | drm_vma_manager.h | 59 unsigned long pages); 61 struct drm_vma_offset_node *node, unsigned long pages); 86 unsigned long pages) in drm_vma_offset_exact_lookup_locked() argument 90 node = drm_vma_offset_lookup_locked(mgr, start, pages); in drm_vma_offset_exact_lookup_locked()
|