/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_object.c | 88 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in amdgpu_bo_destroy() 389 ttm_resource_free(&(*bo_ptr)->tbo, &(*bo_ptr)->tbo.mem); in amdgpu_bo_create_kernel_at() 580 bo->tbo.priority = 1; in amdgpu_bo_do_create() 748 if (bo->tbo.pin_count) in amdgpu_bo_validate() 823 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.mem.num_pages, &bo->kmap); in amdgpu_bo_kmap() 875 ttm_bo_get(&bo->tbo); in amdgpu_bo_ref() 892 tbo = &((*bo)->tbo); in amdgpu_bo_unref() 893 ttm_bo_put(tbo); in amdgpu_bo_unref() 995 ttm_bo_pin(&bo->tbo); in amdgpu_bo_pin_restricted() 1291 if (abo->tbo.base.dma_buf && !abo->tbo.base.import_attach && in amdgpu_bo_move_notify() [all …]
|
H A D | amdgpu_dma_buf.c | 74 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) || in amdgpu_gem_prime_mmap() 171 r = __dma_resv_make_exclusive(bo->tbo.base.resv); in amdgpu_dma_buf_attach() 260 if (!bo->tbo.pin_count) { in amdgpu_dma_buf_map() 280 switch (bo->tbo.mem.mem_type) { in amdgpu_dma_buf_map() 283 bo->tbo.ttm->pages, in amdgpu_dma_buf_map() 284 bo->tbo.ttm->num_pages); in amdgpu_dma_buf_map() 295 bo->tbo.base.size, attach->dev, dir, &sgt); in amdgpu_dma_buf_map() 364 if (!bo->tbo.pin_count && in amdgpu_dma_buf_begin_cpu_access() 404 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) || in amdgpu_gem_prime_export() 485 if (bo->tbo.mem.mem_type == TTM_PL_SYSTEM) in amdgpu_dma_buf_move_notify() [all …]
|
H A D | amdgpu_object.h | 92 struct ttm_buffer_object tbo; member 124 return container_of(tbo, struct amdgpu_bo, tbo); in ttm_to_amdgpu_bo() 165 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_bo_reserve() 168 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in amdgpu_bo_reserve() 179 ttm_bo_unreserve(&bo->tbo); in amdgpu_bo_unreserve() 184 return bo->tbo.base.size; in amdgpu_bo_size() 189 return bo->tbo.base.size / AMDGPU_GPU_PAGE_SIZE; in amdgpu_bo_ngpu_pages() 205 return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); in amdgpu_bo_mmap_offset() 215 struct drm_mm_node *node = bo->tbo.mem.mm_node; in amdgpu_bo_in_cpu_visible_vram() 218 if (bo->tbo.mem.mem_type != TTM_PL_VRAM) in amdgpu_bo_in_cpu_visible_vram() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_object.c | 88 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in amdgpu_bo_destroy() 389 ttm_resource_free(&(*bo_ptr)->tbo, &(*bo_ptr)->tbo.mem); in amdgpu_bo_create_kernel_at() 580 bo->tbo.priority = 1; in amdgpu_bo_do_create() 748 if (bo->tbo.pin_count) in amdgpu_bo_validate() 823 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.mem.num_pages, &bo->kmap); in amdgpu_bo_kmap() 875 ttm_bo_get(&bo->tbo); in amdgpu_bo_ref() 892 tbo = &((*bo)->tbo); in amdgpu_bo_unref() 893 ttm_bo_put(tbo); in amdgpu_bo_unref() 995 ttm_bo_pin(&bo->tbo); in amdgpu_bo_pin_restricted() 1291 if (abo->tbo.base.dma_buf && !abo->tbo.base.import_attach && in amdgpu_bo_move_notify() [all …]
|
H A D | amdgpu_dma_buf.c | 74 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) || in amdgpu_gem_prime_mmap() 171 r = __dma_resv_make_exclusive(bo->tbo.base.resv); in amdgpu_dma_buf_attach() 260 if (!bo->tbo.pin_count) { in amdgpu_dma_buf_map() 280 switch (bo->tbo.mem.mem_type) { in amdgpu_dma_buf_map() 283 bo->tbo.ttm->pages, in amdgpu_dma_buf_map() 284 bo->tbo.ttm->num_pages); in amdgpu_dma_buf_map() 295 bo->tbo.base.size, attach->dev, dir, &sgt); in amdgpu_dma_buf_map() 364 if (!bo->tbo.pin_count && in amdgpu_dma_buf_begin_cpu_access() 404 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) || in amdgpu_gem_prime_export() 485 if (bo->tbo.mem.mem_type == TTM_PL_SYSTEM) in amdgpu_dma_buf_move_notify() [all …]
|
H A D | amdgpu_object.h | 92 struct ttm_buffer_object tbo; member 124 return container_of(tbo, struct amdgpu_bo, tbo); in ttm_to_amdgpu_bo() 165 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_bo_reserve() 168 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in amdgpu_bo_reserve() 179 ttm_bo_unreserve(&bo->tbo); in amdgpu_bo_unreserve() 184 return bo->tbo.base.size; in amdgpu_bo_size() 189 return bo->tbo.base.size / AMDGPU_GPU_PAGE_SIZE; in amdgpu_bo_ngpu_pages() 205 return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); in amdgpu_bo_mmap_offset() 215 struct drm_mm_node *node = bo->tbo.mem.mm_node; in amdgpu_bo_in_cpu_visible_vram() 218 if (bo->tbo.mem.mem_type != TTM_PL_VRAM) in amdgpu_bo_in_cpu_visible_vram() [all …]
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_object.c | 88 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in amdgpu_bo_destroy() 389 ttm_resource_free(&(*bo_ptr)->tbo, &(*bo_ptr)->tbo.mem); in amdgpu_bo_create_kernel_at() 580 bo->tbo.priority = 1; in amdgpu_bo_do_create() 748 if (bo->tbo.pin_count) in amdgpu_bo_validate() 823 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.mem.num_pages, &bo->kmap); in amdgpu_bo_kmap() 875 ttm_bo_get(&bo->tbo); in amdgpu_bo_ref() 892 tbo = &((*bo)->tbo); in amdgpu_bo_unref() 893 ttm_bo_put(tbo); in amdgpu_bo_unref() 995 ttm_bo_pin(&bo->tbo); in amdgpu_bo_pin_restricted() 1291 if (abo->tbo.base.dma_buf && !abo->tbo.base.import_attach && in amdgpu_bo_move_notify() [all …]
|
H A D | amdgpu_dma_buf.c | 74 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) || in amdgpu_gem_prime_mmap() 171 r = __dma_resv_make_exclusive(bo->tbo.base.resv); in amdgpu_dma_buf_attach() 260 if (!bo->tbo.pin_count) { in amdgpu_dma_buf_map() 280 switch (bo->tbo.mem.mem_type) { in amdgpu_dma_buf_map() 283 bo->tbo.ttm->pages, in amdgpu_dma_buf_map() 284 bo->tbo.ttm->num_pages); in amdgpu_dma_buf_map() 295 bo->tbo.base.size, attach->dev, dir, &sgt); in amdgpu_dma_buf_map() 364 if (!bo->tbo.pin_count && in amdgpu_dma_buf_begin_cpu_access() 404 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) || in amdgpu_gem_prime_export() 485 if (bo->tbo.mem.mem_type == TTM_PL_SYSTEM) in amdgpu_dma_buf_move_notify() [all …]
|
H A D | amdgpu_object.h | 92 struct ttm_buffer_object tbo; member 124 return container_of(tbo, struct amdgpu_bo, tbo); in ttm_to_amdgpu_bo() 165 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_bo_reserve() 168 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in amdgpu_bo_reserve() 179 ttm_bo_unreserve(&bo->tbo); in amdgpu_bo_unreserve() 184 return bo->tbo.base.size; in amdgpu_bo_size() 189 return bo->tbo.base.size / AMDGPU_GPU_PAGE_SIZE; in amdgpu_bo_ngpu_pages() 205 return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); in amdgpu_bo_mmap_offset() 215 struct drm_mm_node *node = bo->tbo.mem.mm_node; in amdgpu_bo_in_cpu_visible_vram() 218 if (bo->tbo.mem.mem_type != TTM_PL_VRAM) in amdgpu_bo_in_cpu_visible_vram() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/qxl/ |
H A D | qxl_object.c | 40 bo = to_qxl_bo(tbo); in qxl_ttm_bo_destroy() 41 qdev = to_qxl(bo->tbo.base.dev); in qxl_ttm_bo_destroy() 65 if (qbo->tbo.base.size <= PAGE_SIZE) in qxl_ttm_placement_from_domain() 143 bo->tbo.priority = priority; in qxl_bo_create() 155 ttm_bo_pin(&bo->tbo); in qxl_bo_create() 156 ttm_bo_unreserve(&bo->tbo); in qxl_bo_create() 300 if (bo->tbo.pin_count) { in __qxl_bo_pin() 301 ttm_bo_pin(&bo->tbo); in __qxl_bo_pin() 307 ttm_bo_pin(&bo->tbo); in __qxl_bo_pin() 315 ttm_bo_unpin(&bo->tbo); in __qxl_bo_unpin() [all …]
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/qxl/ |
H A D | qxl_object.c | 40 bo = to_qxl_bo(tbo); in qxl_ttm_bo_destroy() 41 qdev = to_qxl(bo->tbo.base.dev); in qxl_ttm_bo_destroy() 65 if (qbo->tbo.base.size <= PAGE_SIZE) in qxl_ttm_placement_from_domain() 143 bo->tbo.priority = priority; in qxl_bo_create() 155 ttm_bo_pin(&bo->tbo); in qxl_bo_create() 156 ttm_bo_unreserve(&bo->tbo); in qxl_bo_create() 300 if (bo->tbo.pin_count) { in __qxl_bo_pin() 301 ttm_bo_pin(&bo->tbo); in __qxl_bo_pin() 307 ttm_bo_pin(&bo->tbo); in __qxl_bo_pin() 315 ttm_bo_unpin(&bo->tbo); in __qxl_bo_unpin() [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/qxl/ |
H A D | qxl_object.c | 40 bo = to_qxl_bo(tbo); in qxl_ttm_bo_destroy() 41 qdev = to_qxl(bo->tbo.base.dev); in qxl_ttm_bo_destroy() 65 if (qbo->tbo.base.size <= PAGE_SIZE) in qxl_ttm_placement_from_domain() 143 bo->tbo.priority = priority; in qxl_bo_create() 155 ttm_bo_pin(&bo->tbo); in qxl_bo_create() 156 ttm_bo_unreserve(&bo->tbo); in qxl_bo_create() 300 if (bo->tbo.pin_count) { in __qxl_bo_pin() 301 ttm_bo_pin(&bo->tbo); in __qxl_bo_pin() 307 ttm_bo_pin(&bo->tbo); in __qxl_bo_pin() 315 ttm_bo_unpin(&bo->tbo); in __qxl_bo_unpin() [all …]
|
/dports/misc/rump/buildrump.sh-b914579/src/sys/external/bsd/drm2/dist/drm/qxl/ |
H A D | qxl_object.c | 35 bo = container_of(tbo, struct qxl_bo, tbo); in qxl_ttm_bo_destroy() 134 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); in qxl_bo_kmap() 146 struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; in qxl_bo_kmap_atomic_page() 159 ret = ttm_mem_io_reserve(bo->tbo.bdev, &bo->tbo.mem); in qxl_bo_kmap_atomic_page() 188 struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; in qxl_bo_kunmap_atomic_page() 201 ttm_mem_io_free(bo->tbo.bdev, &bo->tbo.mem); in qxl_bo_kunmap_atomic_page() 210 struct ttm_buffer_object *tbo; in qxl_bo_unref() local 214 tbo = &((*bo)->tbo); in qxl_bo_unref() 215 ttm_bo_unref(&tbo); in qxl_bo_unref() 216 if (tbo == NULL) in qxl_bo_unref() [all …]
|
H A D | qxl_object.h | 34 r = ttm_bo_reserve(&bo->tbo, true, no_wait, false, 0); in qxl_bo_reserve() 47 ttm_bo_unreserve(&bo->tbo); in qxl_bo_unreserve() 52 return bo->tbo.offset; in qxl_bo_gpu_offset() 57 return bo->tbo.num_pages << PAGE_SHIFT; in qxl_bo_size() 62 return drm_vma_node_offset_addr(&bo->tbo.vma_node); in qxl_bo_mmap_offset() 79 spin_lock(&bo->tbo.bdev->fence_lock); in qxl_bo_wait() 81 *mem_type = bo->tbo.mem.mem_type; in qxl_bo_wait() 82 if (bo->tbo.sync_obj) in qxl_bo_wait() 83 r = ttm_bo_wait(&bo->tbo, true, true, no_wait); in qxl_bo_wait() 84 spin_unlock(&bo->tbo.bdev->fence_lock); in qxl_bo_wait() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | radeon_object.c | 77 bo = container_of(tbo, struct radeon_bo, tbo); in radeon_ttm_bo_destroy() 87 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in radeon_ttm_bo_destroy() 253 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.mem.num_pages, &bo->kmap); in radeon_bo_kmap() 279 ttm_bo_get(&bo->tbo); in radeon_bo_ref() 289 tbo = &((*bo)->tbo); in radeon_bo_unref() 290 ttm_bo_put(tbo); in radeon_bo_unref() 303 if (bo->tbo.pin_count) { in radeon_bo_pin_restricted() 304 ttm_bo_pin(&bo->tbo); in radeon_bo_pin_restricted() 340 ttm_bo_pin(&bo->tbo); in radeon_bo_pin_restricted() 360 ttm_bo_unpin(&bo->tbo); in radeon_bo_unpin() [all …]
|
H A D | radeon_object.h | 68 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in radeon_bo_reserve() 79 ttm_bo_unreserve(&bo->tbo); in radeon_bo_unreserve() 96 rdev = radeon_get_rdev(bo->tbo.bdev); in radeon_bo_gpu_offset() 98 switch (bo->tbo.mem.mem_type) { in radeon_bo_gpu_offset() 107 return (bo->tbo.mem.start << PAGE_SHIFT) + start; in radeon_bo_gpu_offset() 112 return bo->tbo.base.size; in radeon_bo_size() 117 return bo->tbo.base.size / RADEON_GPU_PAGE_SIZE; in radeon_bo_ngpu_pages() 122 return (bo->tbo.mem.page_alignment << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE; in radeon_bo_gpu_page_alignment() 133 return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); in radeon_bo_mmap_offset()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | radeon_object.c | 77 bo = container_of(tbo, struct radeon_bo, tbo); in radeon_ttm_bo_destroy() 87 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in radeon_ttm_bo_destroy() 253 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.mem.num_pages, &bo->kmap); in radeon_bo_kmap() 279 ttm_bo_get(&bo->tbo); in radeon_bo_ref() 289 tbo = &((*bo)->tbo); in radeon_bo_unref() 290 ttm_bo_put(tbo); in radeon_bo_unref() 303 if (bo->tbo.pin_count) { in radeon_bo_pin_restricted() 304 ttm_bo_pin(&bo->tbo); in radeon_bo_pin_restricted() 340 ttm_bo_pin(&bo->tbo); in radeon_bo_pin_restricted() 360 ttm_bo_unpin(&bo->tbo); in radeon_bo_unpin() [all …]
|
H A D | radeon_object.h | 68 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in radeon_bo_reserve() 79 ttm_bo_unreserve(&bo->tbo); in radeon_bo_unreserve() 96 rdev = radeon_get_rdev(bo->tbo.bdev); in radeon_bo_gpu_offset() 98 switch (bo->tbo.mem.mem_type) { in radeon_bo_gpu_offset() 107 return (bo->tbo.mem.start << PAGE_SHIFT) + start; in radeon_bo_gpu_offset() 112 return bo->tbo.base.size; in radeon_bo_size() 117 return bo->tbo.base.size / RADEON_GPU_PAGE_SIZE; in radeon_bo_ngpu_pages() 122 return (bo->tbo.mem.page_alignment << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE; in radeon_bo_gpu_page_alignment() 133 return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); in radeon_bo_mmap_offset()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | radeon_object.c | 77 bo = container_of(tbo, struct radeon_bo, tbo); in radeon_ttm_bo_destroy() 87 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in radeon_ttm_bo_destroy() 253 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.mem.num_pages, &bo->kmap); in radeon_bo_kmap() 279 ttm_bo_get(&bo->tbo); in radeon_bo_ref() 289 tbo = &((*bo)->tbo); in radeon_bo_unref() 290 ttm_bo_put(tbo); in radeon_bo_unref() 303 if (bo->tbo.pin_count) { in radeon_bo_pin_restricted() 304 ttm_bo_pin(&bo->tbo); in radeon_bo_pin_restricted() 340 ttm_bo_pin(&bo->tbo); in radeon_bo_pin_restricted() 360 ttm_bo_unpin(&bo->tbo); in radeon_bo_unpin() [all …]
|
/dports/graphics/piglit/piglit-136a7f5fa0703603be1ffb338abe4865e76a8058/tests/spec/arb_texture_buffer_object/ |
H A D | re-init.c | 69 void destroy_tbo(struct texbo *tbo); 149 struct texbo tbo; in create_tbo() local 150 glGenBuffers(1, &tbo.bo); in create_tbo() 151 glBindBuffer(GL_TEXTURE_BUFFER, tbo.bo); in create_tbo() 153 glGenTextures(1, &tbo.tex); in create_tbo() 154 glBindTexture(GL_TEXTURE_BUFFER, tbo.tex); in create_tbo() 159 return tbo; in create_tbo() 164 glBindBuffer(GL_TEXTURE_BUFFER, tbo->bo); in init_tbo_data() 185 void destroy_tbo(struct texbo *tbo) in destroy_tbo() argument 187 glDeleteBuffers(1, &tbo->bo); in destroy_tbo() [all …]
|
/dports/misc/rump/buildrump.sh-b914579/src/sys/external/bsd/drm2/dist/drm/radeon/ |
H A D | radeon_object.c | 85 bo = container_of(tbo, struct radeon_bo, tbo); in radeon_ttm_bo_destroy() 140 if (rbo->tbo.mem.size > 512 * 1024) { in radeon_ttm_placement_from_domain() 218 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); in radeon_bo_kmap() 241 struct ttm_buffer_object *tbo; in radeon_bo_unref() local 247 tbo = &((*bo)->tbo); in radeon_bo_unref() 249 ttm_bo_unref(&tbo); in radeon_bo_unref() 251 if (tbo == NULL) in radeon_bo_unref() 511 return ttm_fbdev_mmap(vma, &bo->tbo); in radeon_bo_fbdev_mmap() 754 *mem_type = bo->tbo.mem.mem_type; in radeon_bo_wait() 755 if (bo->tbo.sync_obj) in radeon_bo_wait() [all …]
|
/dports/cad/horizon-eda/horizon-2.1.0/src/widgets/ |
H A D | sheet_box.cpp | 87 tbo->set_icon_name("list-add-symbolic"); in SheetBox() 89 tb->insert(*tbo, -1); in SheetBox() 93 tbo->set_icon_name("list-remove-symbolic"); in SheetBox() 95 tb->insert(*tbo, -1); in SheetBox() 96 remove_button = tbo; in SheetBox() 100 tbo->set_icon_name("go-up-symbolic"); in SheetBox() 102 tb->insert(*tbo, -1); in SheetBox() 103 move_up_button = tbo; in SheetBox() 107 tbo->set_icon_name("go-down-symbolic"); in SheetBox() 109 tb->insert(*tbo, -1); in SheetBox() [all …]
|
H A D | title_block_values_editor.cpp | 71 auto tbo = Gtk::manage(new Gtk::ToolButton()); in TitleBlockValuesEditor() local 72 tbo->set_icon_name("list-add-symbolic"); in TitleBlockValuesEditor() 73 tbo->signal_clicked().connect([this] { in TitleBlockValuesEditor() 85 tb->insert(*tbo, -1); in TitleBlockValuesEditor() 88 auto tbo = Gtk::manage(new Gtk::ToolButton()); in TitleBlockValuesEditor() local 89 tbo->set_icon_name("list-remove-symbolic"); in TitleBlockValuesEditor() 90 tbo->signal_clicked().connect([this] { in TitleBlockValuesEditor() 100 tb->insert(*tbo, -1); in TitleBlockValuesEditor() 101 tb_remove = tbo; in TitleBlockValuesEditor()
|
/dports/net-mgmt/rtrlib/rtrlib-0.6.3/rtrlib/lib/ |
H A D | convert_byte_order.c | 15 uint16_t lrtr_convert_short(const enum target_byte_order tbo, in lrtr_convert_short() argument 18 if (tbo == TO_NETWORK_BYTE_ORDER) in lrtr_convert_short() 20 else if (tbo == TO_HOST_HOST_BYTE_ORDER) in lrtr_convert_short() 26 uint32_t lrtr_convert_long(const enum target_byte_order tbo, in lrtr_convert_long() argument 29 if (tbo == TO_NETWORK_BYTE_ORDER) in lrtr_convert_long() 31 else if (tbo == TO_HOST_HOST_BYTE_ORDER) in lrtr_convert_long()
|
/dports/graphics/piglit/piglit-136a7f5fa0703603be1ffb338abe4865e76a8058/tests/spec/arb_bindless_texture/ |
H A D | illegal.c | 169 GLuint tex, tbo; in call_TexBuffer_when_texture_is_referenced() local 171 glGenBuffers(1, &tbo); in call_TexBuffer_when_texture_is_referenced() 172 glBindBuffer(GL_TEXTURE_BUFFER, tbo); in call_TexBuffer_when_texture_is_referenced() 177 glTexBuffer(GL_TEXTURE_BUFFER, GL_RGBA32F, tbo); in call_TexBuffer_when_texture_is_referenced() 179 glTexBuffer(GL_TEXTURE_BUFFER, GL_RGBA32F, tbo); in call_TexBuffer_when_texture_is_referenced() 192 glTexBuffer(GL_TEXTURE_BUFFER, GL_RGBA32F, tbo); in call_TexBuffer_when_texture_is_referenced() 203 GLuint tex, tbo; in call_BufferData_when_texture_is_referenced() local 205 glGenBuffers(1, &tbo); in call_BufferData_when_texture_is_referenced() 206 glBindBuffer(GL_TEXTURE_BUFFER, tbo); in call_BufferData_when_texture_is_referenced() 211 glTexBuffer(GL_TEXTURE_BUFFER, GL_RGBA32F, tbo); in call_BufferData_when_texture_is_referenced()
|