/netbsd/sys/external/bsd/drm2/dist/drm/ttm/ |
H A D | ttm_tt.c | 116 kvfree(ttm->ttm.pages); in ttm_dma_tt_alloc_page_directory() 117 ttm->ttm.pages = NULL; in ttm_dma_tt_alloc_page_directory() 124 ttm->ttm.pages = kvmalloc_array(ttm->ttm.num_pages, in ttm_dma_tt_alloc_page_directory() 128 if (!ttm->ttm.pages) in ttm_dma_tt_alloc_page_directory() 130 ttm->dma_address = (void *) (ttm->ttm.pages + ttm->ttm.num_pages); in ttm_dma_tt_alloc_page_directory() 141 ttm->ttm.num_pages << PAGE_SHIFT, ttm->ttm.num_pages, PAGE_SIZE, 0, in ttm_sg_tt_alloc_page_directory() 144 ttm->dma_address = kvmalloc_array(ttm->ttm.num_pages, in ttm_sg_tt_alloc_page_directory() 266 ttm->func->destroy(ttm); in ttm_tt_destroy() 430 ttm, (int)ttm->state); in ttm_tt_wire() 467 ttm, (int)ttm->state); in ttm_tt_unwire() [all …]
|
H A D | ttm_agp_backend.c | 51 struct ttm_tt ttm; member 58 struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm); in ttm_agp_bind() 71 struct page *page = ttm->pages[i]; in ttm_agp_bind() 92 struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm); in ttm_agp_unbind() 105 struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm); in ttm_agp_destroy() 108 ttm_agp_unbind(ttm); in ttm_agp_destroy() 109 ttm_tt_fini(ttm); in ttm_agp_destroy() 131 agp_be->ttm.func = &ttm_agp_func; in ttm_agp_tt_create() 138 return &agp_be->ttm; in ttm_agp_tt_create() 144 if (ttm->state != tt_unpopulated) in ttm_agp_tt_populate() [all …]
|
H A D | ttm_page_alloc.c | 1043 if (!ttm->pages[i]) in ttm_pool_unpopulate_helper() 1050 ttm_put_pages(ttm->pages, ttm->num_pages, ttm->page_flags, in ttm_pool_unpopulate_helper() 1051 ttm->caching_state); in ttm_pool_unpopulate_helper() 1052 ttm->state = tt_unpopulated; in ttm_pool_unpopulate_helper() 1067 ret = ttm_get_pages(ttm->pages, ttm->num_pages, ttm->page_flags, in ttm_pool_populate() 1068 ttm->caching_state); in ttm_pool_populate() 1084 ret = ttm_tt_swapin(ttm); in ttm_pool_populate() 1086 ttm_pool_unpopulate(ttm); in ttm_pool_populate() 1091 ttm->state = tt_unbound; in ttm_pool_populate() 1098 ttm_pool_unpopulate_helper(ttm, ttm->num_pages); in ttm_pool_unpopulate() [all …]
|
H A D | ttm_bo_util.c | 67 struct ttm_tt *ttm = bo->ttm; in ttm_bo_move_ttm() local 80 ttm_tt_unbind(ttm); in ttm_bo_move_ttm() 419 struct ttm_tt *ttm = bo->ttm; in ttm_bo_move_memcpy() local 451 (ttm == NULL || (ttm->state == tt_unpopulated && in ttm_bo_move_memcpy() 460 if (ttm) { in ttm_bo_move_memcpy() 501 bo->ttm = NULL; in ttm_bo_move_memcpy() 690 struct ttm_tt *ttm = bo->ttm; in ttm_bo_kmap_ttm() local 694 BUG_ON(!ttm); in ttm_bo_kmap_ttm() 826 bo->ttm = NULL; in ttm_bo_move_accel_cleanup() 856 bo->ttm = NULL; in ttm_bo_move_accel_cleanup() [all …]
|
H A D | ttm_page_alloc_dma.c | 843 struct ttm_tt *ttm = &ttm_dma->ttm; in ttm_dma_pool_get_pages() local 863 struct ttm_tt *ttm = &ttm_dma->ttm; in ttm_dma_pool_gfp_flags() local 894 struct ttm_tt *ttm = &ttm_dma->ttm; in ttm_dma_populate() local 911 type = ttm_to_type(ttm->page_flags, ttm->caching_state); in ttm_dma_populate() 942 ttm->pages[j] = ttm->pages[j - 1] + 1; in ttm_dma_populate() 983 ret = ttm_tt_swapin(ttm); in ttm_dma_populate() 990 ttm->state = tt_unbound; in ttm_dma_populate() 999 struct ttm_tt *ttm = &ttm_dma->ttm; in ttm_dma_unpopulate() local 1007 type = ttm_to_type(ttm->page_flags, ttm->caching_state); in ttm_dma_unpopulate() 1077 ttm->pages[i] = NULL; in ttm_dma_unpopulate() [all …]
|
H A D | Makefile | 5 ttm-y := ttm_memory.o ttm_tt.o ttm_bo.o \ 8 ttm-$(CONFIG_AGP) += ttm_agp_backend.o 9 ttm-$(CONFIG_DRM_TTM_DMA_PAGE_POOL) += ttm_page_alloc_dma.o 11 obj-$(CONFIG_DRM_TTM) += ttm.o
|
H A D | ttm_bo_vm.c | 192 struct ttm_tt *ttm = NULL; in ttm_bo_vm_fault_reserved() local 205 if (bo->ttm && (bo->ttm->page_flags & TTM_PAGE_FLAG_SG)) in ttm_bo_vm_fault_reserved() 266 ttm = bo->ttm; in ttm_bo_vm_fault_reserved() 267 if (ttm_tt_populate(bo->ttm, &ctx)) { in ttm_bo_vm_fault_reserved() 284 page = ttm->pages[page_offset]; in ttm_bo_vm_fault_reserved() 426 if (unlikely(bo->ttm->page_flags & TTM_PAGE_FLAG_SWAPPED)) { in ttm_bo_vm_access() 427 ret = ttm_tt_swapin(bo->ttm); in ttm_bo_vm_access()
|
/netbsd/sys/external/bsd/drm2/ttm/ |
H A D | ttm_agp_backend.c | 71 return &ttm_agp->ttm_dma.ttm; in ttm_agp_tt_create() 83 ttm, (int)ttm->state); in ttm_agp_tt_populate() 84 return ttm_bus_dma_populate(container_of(ttm, struct ttm_dma_tt, ttm)); in ttm_agp_tt_populate() 93 ttm, (int)ttm->state); in ttm_agp_tt_unpopulate() 94 ttm_bus_dma_unpopulate(container_of(ttm, struct ttm_dma_tt, ttm)); in ttm_agp_tt_unpopulate() 101 ttm_dma.ttm); in ttm_agp_bind() 134 ttm_agp_unbind(struct ttm_tt *ttm) in ttm_agp_unbind() argument 137 ttm_dma.ttm); in ttm_agp_unbind() 158 ttm_dma.ttm); in ttm_agp_destroy() 162 ttm_agp_unbind(ttm); in ttm_agp_destroy() [all …]
|
H A D | ttm_bus_dma.c | 58 KASSERT(ttm_dma->ttm.state == tt_unpopulated); in ttm_bus_dma_populate() 66 ret = ttm_tt_wire(&ttm_dma->ttm); in ttm_bus_dma_populate() 71 ttm_dma->ttm.state = tt_unbound; in ttm_bus_dma_populate() 79 ttm_dma->dma_address, ttm_dma->ttm.pages, in ttm_bus_dma_populate() 90 ttm_tt_unwire(&ttm_dma->ttm); in ttm_bus_dma_populate() 91 ttm_dma->ttm.state = tt_unpopulated; in ttm_bus_dma_populate() 107 KASSERTMSG((ttm_dma->ttm.state == tt_unbound), in ttm_bus_dma_put() 109 &ttm_dma->ttm, (int)ttm_dma->ttm.state); in ttm_bus_dma_put() 113 bus_dmamap_unload(ttm_dma->ttm.bdev->dmat, in ttm_bus_dma_put() 115 ttm_tt_unwire(&ttm_dma->ttm); in ttm_bus_dma_put() [all …]
|
H A D | files.ttm | 1 # $NetBSD: files.ttm,v 1.10 2023/06/03 21:30:22 lukem Exp $ 15 file external/bsd/drm2/ttm/ttm_agp_backend.c drmkms_ttm 16 file external/bsd/drm2/dist/drm/ttm/ttm_memory.c drmkms_ttm 17 file external/bsd/drm2/dist/drm/ttm/ttm_tt.c drmkms_ttm 18 file external/bsd/drm2/dist/drm/ttm/ttm_bo.c drmkms_ttm 19 file external/bsd/drm2/dist/drm/ttm/ttm_bo_util.c drmkms_ttm 20 file external/bsd/drm2/ttm/ttm_bo_vm.c drmkms_ttm 21 file external/bsd/drm2/ttm/ttm_module.c drmkms_ttm 23 #file external/bsd/drm2/dist/drm/ttm/ttm_object.c drmkms_ttm 25 #file external/bsd/drm2/dist/drm/ttm/ttm_lock.c drmkms_ttm [all …]
|
H A D | ttm_bo_vm.c | 83 struct ttm_tt *ttm; in ttm_bo_uvm_fault() member 119 if (bo->ttm && ISSET(bo->ttm->page_flags, TTM_PAGE_FLAG_SG)) { in ttm_bo_uvm_fault() 165 u.ttm = bo->ttm; in ttm_bo_uvm_fault() 166 size = (size_t)bo->ttm->num_pages << PAGE_SHIFT; in ttm_bo_uvm_fault() 171 if (ttm_tt_populate(u.ttm, &ctx)) { in ttm_bo_uvm_fault() 196 paddr = page_to_phys(u.ttm->pages[startpage + i]); in ttm_bo_uvm_fault()
|
/netbsd/sys/external/bsd/drm2/dist/drm/radeon/ |
H A D | radeon_ttm.c | 633 r = sg_alloc_table_from_pages(ttm->sg, ttm->pages, ttm->num_pages, 0, in radeon_ttm_tt_pin_userptr() 644 drm_prime_sg_to_page_addr_arrays(ttm->sg, ttm->pages, in radeon_ttm_tt_pin_userptr() 645 gtt->ttm.dma_address, ttm->num_pages); in radeon_ttm_tt_pin_userptr() 712 ttm->num_pages, bo_mem, ttm); in radeon_ttm_backend_bind() 717 ttm->pages, gtt->ttm.dma_address, flags); in radeon_ttm_backend_bind() 770 gtt->ttm.ttm.func = &radeon_backend_func; in radeon_ttm_tt_create() 776 return >t->ttm.ttm; in radeon_ttm_tt_create() 781 if (!ttm || ttm->func != &radeon_backend_func) in radeon_ttm_tt_to_gtt() 812 gtt->ttm.dma_address, ttm->sg); in radeon_ttm_tt_populate() 817 gtt->ttm.dma_address, ttm->num_pages); in radeon_ttm_tt_populate() [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/nouveau/ |
H A D | nouveau_sgdma.c | 18 struct ttm_dma_tt ttm; member 23 nouveau_sgdma_destroy(struct ttm_tt *ttm) in nouveau_sgdma_destroy() argument 27 if (ttm) { in nouveau_sgdma_destroy() 28 ttm_dma_tt_fini(&nvbe->ttm); in nouveau_sgdma_destroy() 40 ret = nouveau_mem_host(reg, &nvbe->ttm); in nv04_sgdma_bind() 55 nv04_sgdma_unbind(struct ttm_tt *ttm) in nv04_sgdma_unbind() argument 75 ret = nouveau_mem_host(reg, &nvbe->ttm); in nv50_sgdma_bind() 100 nvbe->ttm.ttm.func = &nv04_sgdma_backend; in nouveau_sgdma_create_ttm() 102 nvbe->ttm.ttm.func = &nv50_sgdma_backend; in nouveau_sgdma_create_ttm() 104 if (ttm_dma_tt_init(&nvbe->ttm, bo, page_flags)) in nouveau_sgdma_create_ttm() [all …]
|
H A D | nouveau_ttm.c | 192 return ttm_bo_mmap(filp, vma, &drm->ttm.bdev); in nouveau_ttm_mmap() 208 drm->ttm.type_host[!!kind] = typei; in nouveau_ttm_init_host() 214 drm->ttm.type_ncoh[!!kind] = typei; in nouveau_ttm_init_host() 247 drm->ttm.type_vram = typei; in nouveau_ttm_init() 249 drm->ttm.type_vram = -1; in nouveau_ttm_init() 259 ret = ttm_bo_device_init(&drm->ttm.bdev, in nouveau_ttm_init() 319 ttm_bo_clean_mm(&drm->ttm.bdev, TTM_PL_VRAM); in nouveau_ttm_fini() 320 ttm_bo_clean_mm(&drm->ttm.bdev, TTM_PL_TT); in nouveau_ttm_fini() 322 ttm_bo_device_release(&drm->ttm.bdev); in nouveau_ttm_fini() 324 arch_phys_wc_del(drm->ttm.mtrr); in nouveau_ttm_fini() [all …]
|
H A D | nouveau_bo.c | 220 nvbo->bo.bdev = &drm->ttm.bdev; in nouveau_bo_alloc() 1300 &drm->ttm.copy); in nouveau_bo_move_init() 1309 drm->ttm.chan = chan; in nouveau_bo_move_init() 1483 if (drm->ttm.move) { in nouveau_bo_move() 1691 if (slave && ttm->sg) { in nouveau_ttm_tt_populate() 1699 drm_prime_sg_to_page_addr_arrays(ttm->sg, ttm->pages, in nouveau_ttm_tt_populate() 1702 ttm->state = tt_unbound; in nouveau_ttm_tt_populate() 1743 ttm_pool_unpopulate(ttm); in nouveau_ttm_tt_populate() 1770 ttm_agp_tt_unpopulate(ttm); in nouveau_ttm_tt_unpopulate() 1794 ttm_pool_unpopulate(ttm); in nouveau_ttm_tt_unpopulate() [all …]
|
H A D | nouveau_mem.c | 114 type = drm->ttm.type_ncoh[!!mem->kind]; in nouveau_mem_host() 116 type = drm->ttm.type_host[0]; in nouveau_mem_host() 129 if (tt->ttm.sg) args.sgl = tt->ttm.sg->sgl; in nouveau_mem_host() 159 drm->ttm.type_vram, page, size, in nouveau_mem_vram() 167 drm->ttm.type_vram, page, size, in nouveau_mem_vram()
|
/netbsd/sys/external/bsd/drm2/dist/include/drm/ttm/ |
H A D | ttm_tt.h | 75 int (*unbind) (struct ttm_tt *ttm); 85 void (*destroy) (struct ttm_tt *ttm); 139 struct ttm_tt ttm; member 185 void ttm_tt_fini(struct ttm_tt *ttm); 206 void ttm_tt_destroy(struct ttm_tt *ttm); 215 void ttm_tt_unbind(struct ttm_tt *ttm); 225 extern int ttm_tt_wire(struct ttm_tt *ttm); 234 extern void ttm_tt_unwire(struct ttm_tt *ttm); 243 int ttm_tt_swapin(struct ttm_tt *ttm); 278 void ttm_tt_unpopulate(struct ttm_tt *ttm); [all …]
|
/netbsd/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
H A D | amdgpu_ttm.c | 829 struct ttm_tt *ttm = bo->tbo.ttm; in amdgpu_ttm_tt_get_user_pages() local 1003 r = sg_alloc_table_from_pages(ttm->sg, ttm->pages, ttm->num_pages, 0, in amdgpu_ttm_tt_pin_userptr() 1071 struct ttm_tt *ttm = tbo->ttm; in amdgpu_ttm_gart_bind() local 1128 ttm->num_pages, bo_mem, ttm); in amdgpu_ttm_backend_bind() 1147 ttm->pages, gtt->ttm.dma_address, flags); in amdgpu_ttm_backend_bind() 1258 gtt->ttm.ttm.num_pages, gtt->offset); in amdgpu_ttm_backend_unbind() 1297 gtt->ttm.ttm.func = &amdgpu_backend_func; in amdgpu_ttm_tt_create() 1305 return >t->ttm.ttm; in amdgpu_ttm_tt_create() 1352 gtt->ttm.dma_address, ttm->sg); in amdgpu_ttm_tt_populate() 1434 ttm.ttm); in amdgpu_ttm_tt_swapout() [all …]
|
H A D | amdgpu_ttm.h | 129 bool amdgpu_ttm_tt_get_user_pages_done(struct ttm_tt *ttm); 136 static inline bool amdgpu_ttm_tt_get_user_pages_done(struct ttm_tt *ttm) in amdgpu_ttm_tt_get_user_pages_done() argument 142 void amdgpu_ttm_tt_set_user_pages(struct ttm_tt *ttm, struct page **pages); 143 int amdgpu_ttm_tt_set_userptr(struct ttm_tt *ttm, uint64_t addr, 145 bool amdgpu_ttm_tt_has_userptr(struct ttm_tt *ttm); 147 struct vmspace *amdgpu_ttm_tt_get_usermm(struct ttm_tt *ttm); 149 struct mm_struct *amdgpu_ttm_tt_get_usermm(struct ttm_tt *ttm); 151 bool amdgpu_ttm_tt_affect_userptr(struct ttm_tt *ttm, unsigned long start, 153 bool amdgpu_ttm_tt_userptr_invalidated(struct ttm_tt *ttm, 155 bool amdgpu_ttm_tt_is_userptr(struct ttm_tt *ttm); [all …]
|
H A D | amdgpu_gmc.c | 54 struct ttm_dma_tt *ttm; in amdgpu_gmc_get_pde_for_bo() local 58 ttm = container_of(bo->tbo.ttm, struct ttm_dma_tt, ttm); in amdgpu_gmc_get_pde_for_bo() 60 *addr = ttm->dma_address->dm_segs[0].ds_addr; in amdgpu_gmc_get_pde_for_bo() 62 *addr = ttm->dma_address[0]; in amdgpu_gmc_get_pde_for_bo() 72 *flags = amdgpu_ttm_tt_pde_flags(bo->tbo.ttm, &bo->tbo.mem); in amdgpu_gmc_get_pde_for_bo() 142 struct ttm_dma_tt *ttm; in amdgpu_gmc_agp_addr() local 145 if (bo->num_pages != 1 || bo->ttm->caching_state == tt_cached) in amdgpu_gmc_agp_addr() 148 ttm = container_of(bo->ttm, struct ttm_dma_tt, ttm); in amdgpu_gmc_agp_addr() 150 addr = ttm->dma_address->dm_segs[0].ds_addr; in amdgpu_gmc_agp_addr() 152 addr = ttm->dma_address[0]; in amdgpu_gmc_agp_addr()
|
/netbsd/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
H A D | vmwgfx_ttm_buffer.c | 541 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_bo_map_dma() 558 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_bo_unmap_dma() 578 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_bo_sg_table() 587 container_of(ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_ttm_bind() 621 container_of(ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_ttm_unbind() 644 container_of(ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_ttm_destroy() 650 ttm_tt_fini(ttm); in vmw_ttm_destroy() 662 container_of(ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_ttm_populate() 690 dma_ttm.ttm); in vmw_ttm_unpopulate() 708 ttm_pool_unpopulate(ttm); in vmw_ttm_unpopulate() [all …]
|
H A D | vmwgfx_blit.c | 471 if (dst->ttm->state == tt_unpopulated) { in vmw_bo_cpu_blit() 472 ret = dst->ttm->bdev->driver->ttm_tt_populate(dst->ttm, &ctx); in vmw_bo_cpu_blit() 477 if (src->ttm->state == tt_unpopulated) { in vmw_bo_cpu_blit() 478 ret = src->ttm->bdev->driver->ttm_tt_populate(src->ttm, &ctx); in vmw_bo_cpu_blit() 487 d.dst_pages = dst->ttm->pages; in vmw_bo_cpu_blit() 488 d.src_pages = src->ttm->pages; in vmw_bo_cpu_blit()
|
/netbsd/sys/external/bsd/drm2/dist/drm/qxl/ |
H A D | qxl_ttm.c | 160 struct ttm_tt ttm; member 165 static int qxl_ttm_backend_bind(struct ttm_tt *ttm, in qxl_ttm_backend_bind() argument 168 struct qxl_ttm_tt *gtt = (void *)ttm; in qxl_ttm_backend_bind() 171 if (!ttm->num_pages) { in qxl_ttm_backend_bind() 173 ttm->num_pages, bo_mem, ttm); in qxl_ttm_backend_bind() 179 static int qxl_ttm_backend_unbind(struct ttm_tt *ttm) in qxl_ttm_backend_unbind() argument 187 struct qxl_ttm_tt *gtt = (void *)ttm; in qxl_ttm_backend_destroy() 189 ttm_tt_fini(>t->ttm); in qxl_ttm_backend_destroy() 209 gtt->ttm.func = &qxl_backend_func; in qxl_ttm_tt_create() 211 if (ttm_tt_init(>t->ttm, bo, page_flags)) { in qxl_ttm_tt_create() [all …]
|
/netbsd/usr.bin/calendar/ |
H A D | calendar.c | 688 struct tm ttm; in getmmdd() local 690 ttm = *ptm; in getmmdd() 691 ttm.tm_isdst = -1; in getmmdd() 694 ttm.tm_mon = ATOI2(ds) - 1; in getmmdd() 698 ttm.tm_mday = ATOI2(ds); in getmmdd() 706 ttm.tm_year += ATOI2(ds); in getmmdd() 708 ttm.tm_year = ATOI2(ds); in getmmdd() 709 if (ttm.tm_year < 69) in getmmdd() 710 ttm.tm_year += 2000 - TM_YEAR_BASE; in getmmdd() 715 if (ok && mktime(&ttm) == -1) in getmmdd() [all …]
|
/netbsd/crypto/external/bsd/openssl.old/dist/crypto/asn1/ |
H A D | a_utctm.c | 74 struct tm stm, ttm; in ASN1_UTCTIME_cmp_time_t() local 80 if (OPENSSL_gmtime(&t, &ttm) == NULL) in ASN1_UTCTIME_cmp_time_t() 83 if (!OPENSSL_gmtime_diff(&day, &sec, &ttm, &stm)) in ASN1_UTCTIME_cmp_time_t()
|