/dragonfly/sys/dev/drm/ttm/ |
H A D | ttm_tt.c | 92 ttm->pages = kvmalloc_array(ttm->num_pages, sizeof(void*), in ttm_tt_alloc_page_directory() 101 ttm->ttm.pages = kvmalloc_array(ttm->ttm.num_pages, in ttm_dma_tt_alloc_page_directory() 113 ttm->dma_address = kvmalloc_array(ttm->ttm.num_pages, in ttm_sg_tt_alloc_page_directory() 171 drm_clflush_pages(ttm->pages, ttm->num_pages); in ttm_tt_set_caching() 173 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_set_caching() 238 ttm->num_pages = bo->num_pages; in ttm_tt_init_fields() 365 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapin() 417 IDX_TO_OFF(ttm->num_pages), VM_PROT_DEFAULT, 0); in ttm_tt_swapout() 427 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapout() 460 for (i = 0; i < ttm->num_pages; ++i) in ttm_tt_add_mapping() [all …]
|
H A D | ttm_page_alloc.c | 909 for (i = 0; i < ttm->num_pages; ++i) { in ttm_pool_populate() 947 for (i = 0; i < tt->ttm.num_pages; ++i) { in ttm_populate_and_map_pages() 949 size_t num_pages = 1; in ttm_populate_and_map_pages() local 951 for (j = i + 1; j < tt->ttm.num_pages; ++j) { in ttm_populate_and_map_pages() 955 ++num_pages; in ttm_populate_and_map_pages() 959 0, num_pages * PAGE_SIZE, in ttm_populate_and_map_pages() 971 for (j = 1; j < num_pages; ++j) { in ttm_populate_and_map_pages() 984 for (i = 0; i < tt->ttm.num_pages;) { in ttm_unmap_and_unpopulate_pages() 986 size_t num_pages = 1; in ttm_unmap_and_unpopulate_pages() local 997 ++num_pages; in ttm_unmap_and_unpopulate_pages() [all …]
|
H A D | ttm_bo_util.c | 376 memset_io(new_iomap, 0, new_mem->num_pages*PAGE_SIZE); in ttm_bo_move_memcpy() 395 add = new_mem->num_pages - 1; in ttm_bo_move_memcpy() 398 for (i = 0; i < new_mem->num_pages; ++i) { in ttm_bo_move_memcpy() 559 unsigned long num_pages, in ttm_bo_kmap_ttm() argument 593 map->virtual = vmap(ttm->pages + start_page, num_pages, in ttm_bo_kmap_ttm() 600 unsigned long start_page, unsigned long num_pages, in ttm_bo_kmap() argument 610 if (num_pages > bo->num_pages) in ttm_bo_kmap() 612 if (start_page > bo->num_pages) in ttm_bo_kmap() 615 if (num_pages > 1 && !capable(CAP_SYS_ADMIN)) in ttm_bo_kmap() 624 return ttm_bo_kmap_ttm(bo, start_page, num_pages, map); in ttm_bo_kmap() [all …]
|
H A D | ttm_page_alloc_dma.c | 352 unsigned num_pages; in ttm_dma_page_put() local 356 num_pages = pool->size / PAGE_SIZE; in ttm_dma_page_put() 357 if (ttm_set_pages_wb(page, num_pages)) in ttm_dma_page_put() 359 pool->dev_name, num_pages); in ttm_dma_page_put() 893 unsigned long num_pages = ttm->num_pages; in ttm_dma_populate() local 903 if (ttm_check_under_lowerlimit(mem_glob, num_pages, ctx)) in ttm_dma_populate() 924 while (num_pages >= HPAGE_PMD_NR) { in ttm_dma_populate() 946 num_pages -= HPAGE_PMD_NR; in ttm_dma_populate() 961 while (num_pages) { in ttm_dma_populate() 977 --num_pages; in ttm_dma_populate() [all …]
|
H A D | ttm_bo.c | 97 bo, bo->mem.num_pages, bo->mem.size >> 10, in ttm_bo_mem_space_debug() 321 ctx->bytes_moved += bo->num_pages << PAGE_SHIFT; in ttm_bo_handle_move_mem() 1023 mem.num_pages = bo->num_pages; in ttm_bo_move_buffer() 1024 mem.size = mem.num_pages << PAGE_SHIFT; in ttm_bo_move_buffer() 1132 unsigned long num_pages; in ttm_bo_init_reserved() local 1146 num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in ttm_bo_init_reserved() 1147 if (num_pages == 0) { in ttm_bo_init_reserved() 1168 bo->num_pages = num_pages; in ttm_bo_init_reserved() 1169 bo->mem.size = num_pages << PAGE_SHIFT; in ttm_bo_init_reserved() 1171 bo->mem.num_pages = bo->num_pages; in ttm_bo_init_reserved() [all …]
|
H A D | ttm_agp_backend.c | 59 mem = agp_alloc_memory(agp_be->bridge, AGP_USER_MEMORY, ttm->num_pages); in ttm_agp_bind() 64 for (i = 0; i < ttm->num_pages; i++) { in ttm_agp_bind()
|
H A D | ttm_bo_vm.c | 220 if (unlikely(page_offset >= bo->num_pages)) { in ttm_bo_vm_fault() 375 if (len < 1 || (offset + len) >> PAGE_SHIFT > bo->num_pages) in ttm_bo_vm_access() 638 if (unlikely(OFF_TO_IDX(offset) >= bo->num_pages)) { in ttm_bo_vm_fault_dfly() 848 for (i = 0; i < bo->num_pages; i++) { in ttm_bo_release_mmap()
|
/dragonfly/sys/dev/drm/ |
H A D | drm_cache.c | 61 unsigned long num_pages) in drm_cache_flush_clflush() argument 66 for (i = 0; i < num_pages; i++) in drm_cache_flush_clflush() 73 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages() argument 78 drm_cache_flush_clflush(pages, num_pages); in drm_clflush_pages() 86 for (i = 0; i < num_pages; i++) { in drm_clflush_pages()
|
/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_ttm.c | 845 unsigned num_pages = ttm->num_pages - pinned; in amdgpu_ttm_tt_get_user_pages() local 1059 if (!ttm->num_pages) { in amdgpu_ttm_backend_bind() 1262 ttm->num_pages); in amdgpu_ttm_tt_populate() 1489 unsigned long num_pages = bo->mem.num_pages; in amdgpu_ttm_bo_eviction_valuable() local 1515 while (num_pages) { in amdgpu_ttm_bo_eviction_valuable() 1520 num_pages -= node->size; in amdgpu_ttm_bo_eviction_valuable() 2122 unsigned long num_pages; in amdgpu_fill_buffer() local 2139 num_pages = bo->tbo.num_pages; in amdgpu_fill_buffer() 2142 while (num_pages) { in amdgpu_fill_buffer() 2167 num_pages = bo->tbo.num_pages; in amdgpu_fill_buffer() [all …]
|
H A D | amdgpu_gtt_mgr.c | 140 r = drm_mm_insert_node_in_range(&mgr->mm, &node->node, mem->num_pages, in amdgpu_gtt_mgr_alloc() 172 atomic64_read(&mgr->available) < mem->num_pages) { in amdgpu_gtt_mgr_new() 176 atomic64_sub(mem->num_pages, &mgr->available); in amdgpu_gtt_mgr_new() 186 node->node.size = mem->num_pages; in amdgpu_gtt_mgr_new() 204 atomic64_add(mem->num_pages, &mgr->available); in amdgpu_gtt_mgr_new() 232 atomic64_add(mem->num_pages, &mgr->available); in amdgpu_gtt_mgr_del()
|
H A D | amdgpu_vram_mgr.c | 112 unsigned pages = mem->num_pages; in amdgpu_vram_mgr_bo_visible_size() 163 num_nodes = DIV_ROUND_UP(mem->num_pages, pages_per_node); in amdgpu_vram_mgr_new() 176 pages_left = mem->num_pages; in amdgpu_vram_mgr_new() 201 if (start > mem->num_pages) in amdgpu_vram_mgr_new() 202 start -= mem->num_pages; in amdgpu_vram_mgr_new() 243 unsigned pages = mem->num_pages; in amdgpu_vram_mgr_del()
|
H A D | amdgpu_object.h | 172 return bo->tbo.num_pages << PAGE_SHIFT; in amdgpu_bo_size() 177 return (bo->tbo.num_pages << PAGE_SHIFT) / AMDGPU_GPU_PAGE_SIZE; in amdgpu_bo_ngpu_pages() 222 for (pages_left = bo->tbo.mem.num_pages; pages_left; in amdgpu_bo_in_cpu_visible_vram()
|
/dragonfly/sys/dev/drm/include/asm/ |
H A D | cacheflush.h | 54 static inline int set_pages_uc(struct page *page, int num_pages) in set_pages_uc() argument 59 num_pages, PAT_UNCACHED); in set_pages_uc() 64 static inline int set_pages_wb(struct page *page, int num_pages) in set_pages_wb() argument 69 num_pages, PAT_WRITE_BACK); in set_pages_wb()
|
/dragonfly/sys/dev/drm/radeon/ |
H A D | radeon_ttm.c | 264 unsigned num_pages; in radeon_move_blit() local 301 num_pages = new_mem->num_pages * (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); in radeon_move_blit() 570 unsigned num_pages = ttm->num_pages - pinned; 581 } while (pinned < ttm->num_pages); 584 ttm->num_pages << PAGE_SHIFT, 595 gtt->ttm.dma_address, ttm->num_pages); 653 if (!ttm->num_pages) { in radeon_ttm_backend_bind() 655 ttm->num_pages, bo_mem, ttm); in radeon_ttm_backend_bind() 663 ttm->num_pages, (unsigned)gtt->offset); in radeon_ttm_backend_bind() 787 for (i = 0; i < ttm->num_pages; i++) { in radeon_ttm_tt_populate() [all …]
|
H A D | radeon_prime.c | 48 int npages = bo->tbo.num_pages; in radeon_gem_prime_get_sg_table() 58 ret = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, in radeon_gem_prime_vmap()
|
H A D | radeon_object.h | 98 return bo->tbo.num_pages << PAGE_SHIFT; in radeon_bo_size() 103 return (bo->tbo.num_pages << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE; in radeon_bo_ngpu_pages()
|
/dragonfly/sys/dev/drm/i915/ |
H A D | i915_gem_userptr.c | 419 struct page **pvec, int num_pages) 431 ret = __sg_alloc_table_from_pages(st, pvec, num_pages, 432 0, num_pages << PAGE_SHIFT, 610 const int num_pages = obj->base.size >> PAGE_SHIFT; in i915_gem_userptr_get_pages() 646 pvec = kvmalloc_array(num_pages, sizeof(struct page *), in i915_gem_userptr_get_pages() 652 num_pages, in i915_gem_userptr_get_pages() 661 } else if (pinned < num_pages) { in i915_gem_userptr_get_pages() 665 pages = __i915_gem_userptr_alloc_pages(obj, pvec, num_pages); in i915_gem_userptr_get_pages()
|
/dragonfly/sys/dev/drm/include/drm/ttm/ |
H A D | ttm_bo_api.h | 97 unsigned long num_pages; member 178 unsigned long num_pages; member 697 unsigned long num_pages, struct ttm_bo_kmap_obj *map);
|
H A D | ttm_memory.h | 97 uint64_t num_pages, struct ttm_operation_ctx *ctx);
|
/dragonfly/sys/dev/netif/oce/ |
H A D | oce_mbox.c | 1615 int rc, num_pages = 0; in oce_mbox_create_rq() local 1630 num_pages = oce_page_list(rq->ring, &fwcmd->params.req.pages[0]); in oce_mbox_create_rq() 1638 fwcmd->params.req.num_pages = num_pages; in oce_mbox_create_rq() 1670 int rc = 0, version, num_pages; in oce_mbox_create_wq() local 1690 num_pages = oce_page_list(wq->ring, &fwcmd->params.req.pages[0]); in oce_mbox_create_wq() 1693 fwcmd->params.req.num_pages = num_pages; in oce_mbox_create_wq() 1728 uint32_t num_pages; in oce_mbox_create_eq() local 1741 fwcmd->params.req.ctx.num_pages = num_pages; in oce_mbox_create_eq() 1775 uint32_t num_pages, page_size; in oce_mbox_cq_create() local 1801 ctx->v2.num_pages = LE_16(num_pages); in oce_mbox_cq_create() [all …]
|
H A D | oce_hw.h | 729 uint32_t num_pages:16; member 746 uint32_t num_pages:16; 1194 uint32_t num_pages:16; member 1212 uint32_t num_pages:16; 1276 uint32_t num_pages:16; member 1293 uint32_t num_pages:16; 1320 uint32_t num_pages:16; member 1340 uint32_t num_pages:16; 1408 uint32_t num_pages:16; member 2772 uint8_t num_pages; member [all …]
|
H A D | oce_queue.c | 671 uint32_t num_pages; in oce_mq_create() local 704 num_pages = oce_page_list(mq->ring, &fwcmd->params.req.pages[0]); in oce_mq_create() 710 ctx->v1.num_pages = num_pages; in oce_mq_create() 723 ctx->v0.num_pages = num_pages; in oce_mq_create()
|
/dragonfly/sys/dev/drm/include/drm/ |
H A D | drm_agpsupport.h | 43 unsigned long num_pages, 92 unsigned long num_pages, in drm_agp_bind_pages() argument
|
H A D | drm_cache.h | 38 void drm_clflush_pages(struct page *pages[], unsigned long num_pages);
|
/dragonfly/sys/dev/virtual/virtio/balloon/ |
H A D | virtio_balloon.h | 81 uint32_t num_pages; member
|