Home
last modified time | relevance | path

Searched refs:num_pages (Results 1 – 25 of 39) sorted by relevance

12

/dragonfly/sys/dev/drm/ttm/
H A Dttm_tt.c92 ttm->pages = kvmalloc_array(ttm->num_pages, sizeof(void*), in ttm_tt_alloc_page_directory()
101 ttm->ttm.pages = kvmalloc_array(ttm->ttm.num_pages, in ttm_dma_tt_alloc_page_directory()
113 ttm->dma_address = kvmalloc_array(ttm->ttm.num_pages, in ttm_sg_tt_alloc_page_directory()
171 drm_clflush_pages(ttm->pages, ttm->num_pages); in ttm_tt_set_caching()
173 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_set_caching()
238 ttm->num_pages = bo->num_pages; in ttm_tt_init_fields()
365 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapin()
417 IDX_TO_OFF(ttm->num_pages), VM_PROT_DEFAULT, 0); in ttm_tt_swapout()
427 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapout()
460 for (i = 0; i < ttm->num_pages; ++i) in ttm_tt_add_mapping()
[all …]
H A Dttm_page_alloc.c909 for (i = 0; i < ttm->num_pages; ++i) { in ttm_pool_populate()
947 for (i = 0; i < tt->ttm.num_pages; ++i) { in ttm_populate_and_map_pages()
949 size_t num_pages = 1; in ttm_populate_and_map_pages() local
951 for (j = i + 1; j < tt->ttm.num_pages; ++j) { in ttm_populate_and_map_pages()
955 ++num_pages; in ttm_populate_and_map_pages()
959 0, num_pages * PAGE_SIZE, in ttm_populate_and_map_pages()
971 for (j = 1; j < num_pages; ++j) { in ttm_populate_and_map_pages()
984 for (i = 0; i < tt->ttm.num_pages;) { in ttm_unmap_and_unpopulate_pages()
986 size_t num_pages = 1; in ttm_unmap_and_unpopulate_pages() local
997 ++num_pages; in ttm_unmap_and_unpopulate_pages()
[all …]
H A Dttm_bo_util.c376 memset_io(new_iomap, 0, new_mem->num_pages*PAGE_SIZE); in ttm_bo_move_memcpy()
395 add = new_mem->num_pages - 1; in ttm_bo_move_memcpy()
398 for (i = 0; i < new_mem->num_pages; ++i) { in ttm_bo_move_memcpy()
559 unsigned long num_pages, in ttm_bo_kmap_ttm() argument
593 map->virtual = vmap(ttm->pages + start_page, num_pages, in ttm_bo_kmap_ttm()
600 unsigned long start_page, unsigned long num_pages, in ttm_bo_kmap() argument
610 if (num_pages > bo->num_pages) in ttm_bo_kmap()
612 if (start_page > bo->num_pages) in ttm_bo_kmap()
615 if (num_pages > 1 && !capable(CAP_SYS_ADMIN)) in ttm_bo_kmap()
624 return ttm_bo_kmap_ttm(bo, start_page, num_pages, map); in ttm_bo_kmap()
[all …]
H A Dttm_page_alloc_dma.c352 unsigned num_pages; in ttm_dma_page_put() local
356 num_pages = pool->size / PAGE_SIZE; in ttm_dma_page_put()
357 if (ttm_set_pages_wb(page, num_pages)) in ttm_dma_page_put()
359 pool->dev_name, num_pages); in ttm_dma_page_put()
893 unsigned long num_pages = ttm->num_pages; in ttm_dma_populate() local
903 if (ttm_check_under_lowerlimit(mem_glob, num_pages, ctx)) in ttm_dma_populate()
924 while (num_pages >= HPAGE_PMD_NR) { in ttm_dma_populate()
946 num_pages -= HPAGE_PMD_NR; in ttm_dma_populate()
961 while (num_pages) { in ttm_dma_populate()
977 --num_pages; in ttm_dma_populate()
[all …]
H A Dttm_bo.c97 bo, bo->mem.num_pages, bo->mem.size >> 10, in ttm_bo_mem_space_debug()
321 ctx->bytes_moved += bo->num_pages << PAGE_SHIFT; in ttm_bo_handle_move_mem()
1023 mem.num_pages = bo->num_pages; in ttm_bo_move_buffer()
1024 mem.size = mem.num_pages << PAGE_SHIFT; in ttm_bo_move_buffer()
1132 unsigned long num_pages; in ttm_bo_init_reserved() local
1146 num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in ttm_bo_init_reserved()
1147 if (num_pages == 0) { in ttm_bo_init_reserved()
1168 bo->num_pages = num_pages; in ttm_bo_init_reserved()
1169 bo->mem.size = num_pages << PAGE_SHIFT; in ttm_bo_init_reserved()
1171 bo->mem.num_pages = bo->num_pages; in ttm_bo_init_reserved()
[all …]
H A Dttm_agp_backend.c59 mem = agp_alloc_memory(agp_be->bridge, AGP_USER_MEMORY, ttm->num_pages); in ttm_agp_bind()
64 for (i = 0; i < ttm->num_pages; i++) { in ttm_agp_bind()
H A Dttm_bo_vm.c220 if (unlikely(page_offset >= bo->num_pages)) { in ttm_bo_vm_fault()
375 if (len < 1 || (offset + len) >> PAGE_SHIFT > bo->num_pages) in ttm_bo_vm_access()
638 if (unlikely(OFF_TO_IDX(offset) >= bo->num_pages)) { in ttm_bo_vm_fault_dfly()
848 for (i = 0; i < bo->num_pages; i++) { in ttm_bo_release_mmap()
/dragonfly/sys/dev/drm/
H A Ddrm_cache.c61 unsigned long num_pages) in drm_cache_flush_clflush() argument
66 for (i = 0; i < num_pages; i++) in drm_cache_flush_clflush()
73 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages() argument
78 drm_cache_flush_clflush(pages, num_pages); in drm_clflush_pages()
86 for (i = 0; i < num_pages; i++) { in drm_clflush_pages()
/dragonfly/sys/dev/drm/amd/amdgpu/
H A Damdgpu_ttm.c845 unsigned num_pages = ttm->num_pages - pinned; in amdgpu_ttm_tt_get_user_pages() local
1059 if (!ttm->num_pages) { in amdgpu_ttm_backend_bind()
1262 ttm->num_pages); in amdgpu_ttm_tt_populate()
1489 unsigned long num_pages = bo->mem.num_pages; in amdgpu_ttm_bo_eviction_valuable() local
1515 while (num_pages) { in amdgpu_ttm_bo_eviction_valuable()
1520 num_pages -= node->size; in amdgpu_ttm_bo_eviction_valuable()
2122 unsigned long num_pages; in amdgpu_fill_buffer() local
2139 num_pages = bo->tbo.num_pages; in amdgpu_fill_buffer()
2142 while (num_pages) { in amdgpu_fill_buffer()
2167 num_pages = bo->tbo.num_pages; in amdgpu_fill_buffer()
[all …]
H A Damdgpu_gtt_mgr.c140 r = drm_mm_insert_node_in_range(&mgr->mm, &node->node, mem->num_pages, in amdgpu_gtt_mgr_alloc()
172 atomic64_read(&mgr->available) < mem->num_pages) { in amdgpu_gtt_mgr_new()
176 atomic64_sub(mem->num_pages, &mgr->available); in amdgpu_gtt_mgr_new()
186 node->node.size = mem->num_pages; in amdgpu_gtt_mgr_new()
204 atomic64_add(mem->num_pages, &mgr->available); in amdgpu_gtt_mgr_new()
232 atomic64_add(mem->num_pages, &mgr->available); in amdgpu_gtt_mgr_del()
H A Damdgpu_vram_mgr.c112 unsigned pages = mem->num_pages; in amdgpu_vram_mgr_bo_visible_size()
163 num_nodes = DIV_ROUND_UP(mem->num_pages, pages_per_node); in amdgpu_vram_mgr_new()
176 pages_left = mem->num_pages; in amdgpu_vram_mgr_new()
201 if (start > mem->num_pages) in amdgpu_vram_mgr_new()
202 start -= mem->num_pages; in amdgpu_vram_mgr_new()
243 unsigned pages = mem->num_pages; in amdgpu_vram_mgr_del()
H A Damdgpu_object.h172 return bo->tbo.num_pages << PAGE_SHIFT; in amdgpu_bo_size()
177 return (bo->tbo.num_pages << PAGE_SHIFT) / AMDGPU_GPU_PAGE_SIZE; in amdgpu_bo_ngpu_pages()
222 for (pages_left = bo->tbo.mem.num_pages; pages_left; in amdgpu_bo_in_cpu_visible_vram()
/dragonfly/sys/dev/drm/include/asm/
H A Dcacheflush.h54 static inline int set_pages_uc(struct page *page, int num_pages) in set_pages_uc() argument
59 num_pages, PAT_UNCACHED); in set_pages_uc()
64 static inline int set_pages_wb(struct page *page, int num_pages) in set_pages_wb() argument
69 num_pages, PAT_WRITE_BACK); in set_pages_wb()
/dragonfly/sys/dev/drm/radeon/
H A Dradeon_ttm.c264 unsigned num_pages; in radeon_move_blit() local
301 num_pages = new_mem->num_pages * (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); in radeon_move_blit()
570 unsigned num_pages = ttm->num_pages - pinned;
581 } while (pinned < ttm->num_pages);
584 ttm->num_pages << PAGE_SHIFT,
595 gtt->ttm.dma_address, ttm->num_pages);
653 if (!ttm->num_pages) { in radeon_ttm_backend_bind()
655 ttm->num_pages, bo_mem, ttm); in radeon_ttm_backend_bind()
663 ttm->num_pages, (unsigned)gtt->offset); in radeon_ttm_backend_bind()
787 for (i = 0; i < ttm->num_pages; i++) { in radeon_ttm_tt_populate()
[all …]
H A Dradeon_prime.c48 int npages = bo->tbo.num_pages; in radeon_gem_prime_get_sg_table()
58 ret = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, in radeon_gem_prime_vmap()
H A Dradeon_object.h98 return bo->tbo.num_pages << PAGE_SHIFT; in radeon_bo_size()
103 return (bo->tbo.num_pages << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE; in radeon_bo_ngpu_pages()
/dragonfly/sys/dev/drm/i915/
H A Di915_gem_userptr.c419 struct page **pvec, int num_pages)
431 ret = __sg_alloc_table_from_pages(st, pvec, num_pages,
432 0, num_pages << PAGE_SHIFT,
610 const int num_pages = obj->base.size >> PAGE_SHIFT; in i915_gem_userptr_get_pages()
646 pvec = kvmalloc_array(num_pages, sizeof(struct page *), in i915_gem_userptr_get_pages()
652 num_pages, in i915_gem_userptr_get_pages()
661 } else if (pinned < num_pages) { in i915_gem_userptr_get_pages()
665 pages = __i915_gem_userptr_alloc_pages(obj, pvec, num_pages); in i915_gem_userptr_get_pages()
/dragonfly/sys/dev/drm/include/drm/ttm/
H A Dttm_bo_api.h97 unsigned long num_pages; member
178 unsigned long num_pages; member
697 unsigned long num_pages, struct ttm_bo_kmap_obj *map);
H A Dttm_memory.h97 uint64_t num_pages, struct ttm_operation_ctx *ctx);
/dragonfly/sys/dev/netif/oce/
H A Doce_mbox.c1615 int rc, num_pages = 0; in oce_mbox_create_rq() local
1630 num_pages = oce_page_list(rq->ring, &fwcmd->params.req.pages[0]); in oce_mbox_create_rq()
1638 fwcmd->params.req.num_pages = num_pages; in oce_mbox_create_rq()
1670 int rc = 0, version, num_pages; in oce_mbox_create_wq() local
1690 num_pages = oce_page_list(wq->ring, &fwcmd->params.req.pages[0]); in oce_mbox_create_wq()
1693 fwcmd->params.req.num_pages = num_pages; in oce_mbox_create_wq()
1728 uint32_t num_pages; in oce_mbox_create_eq() local
1741 fwcmd->params.req.ctx.num_pages = num_pages; in oce_mbox_create_eq()
1775 uint32_t num_pages, page_size; in oce_mbox_cq_create() local
1801 ctx->v2.num_pages = LE_16(num_pages); in oce_mbox_cq_create()
[all …]
H A Doce_hw.h729 uint32_t num_pages:16; member
746 uint32_t num_pages:16;
1194 uint32_t num_pages:16; member
1212 uint32_t num_pages:16;
1276 uint32_t num_pages:16; member
1293 uint32_t num_pages:16;
1320 uint32_t num_pages:16; member
1340 uint32_t num_pages:16;
1408 uint32_t num_pages:16; member
2772 uint8_t num_pages; member
[all …]
H A Doce_queue.c671 uint32_t num_pages; in oce_mq_create() local
704 num_pages = oce_page_list(mq->ring, &fwcmd->params.req.pages[0]); in oce_mq_create()
710 ctx->v1.num_pages = num_pages; in oce_mq_create()
723 ctx->v0.num_pages = num_pages; in oce_mq_create()
/dragonfly/sys/dev/drm/include/drm/
H A Ddrm_agpsupport.h43 unsigned long num_pages,
92 unsigned long num_pages, in drm_agp_bind_pages() argument
H A Ddrm_cache.h38 void drm_clflush_pages(struct page *pages[], unsigned long num_pages);
/dragonfly/sys/dev/virtual/virtio/balloon/
H A Dvirtio_balloon.h81 uint32_t num_pages; member

12