/openbsd/sys/dev/pci/drm/amd/amdgpu/ |
H A D | amdgpu_vce.c | 471 ib->ptr[ib->length_dw++] = handle; in amdgpu_vce_get_create_msg() 478 ib->ptr[ib->length_dw++] = 0x00000000; in amdgpu_vce_get_create_msg() 479 ib->ptr[ib->length_dw++] = 0x00000042; in amdgpu_vce_get_create_msg() 480 ib->ptr[ib->length_dw++] = 0x0000000a; in amdgpu_vce_get_create_msg() 481 ib->ptr[ib->length_dw++] = 0x00000001; in amdgpu_vce_get_create_msg() 482 ib->ptr[ib->length_dw++] = 0x00000080; in amdgpu_vce_get_create_msg() 483 ib->ptr[ib->length_dw++] = 0x00000060; in amdgpu_vce_get_create_msg() 484 ib->ptr[ib->length_dw++] = 0x00000100; in amdgpu_vce_get_create_msg() 485 ib->ptr[ib->length_dw++] = 0x00000100; in amdgpu_vce_get_create_msg() 498 ib->ptr[ib->length_dw++] = addr; in amdgpu_vce_get_create_msg() [all …]
|
H A D | amdgpu_si_dma.c | 265 memset(&ib, 0, sizeof(ib)); in si_dma_ring_test_ib() 319 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pte() 320 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pte() 343 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_write_pte() 344 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in si_dma_vm_write_pte() 346 ib->ptr[ib->length_dw++] = lower_32_bits(value); in si_dma_vm_write_pte() 347 ib->ptr[ib->length_dw++] = upper_32_bits(value); in si_dma_vm_write_pte() 384 ib->ptr[ib->length_dw++] = pe; /* dst addr */ in si_dma_vm_set_pte_pde() 388 ib->ptr[ib->length_dw++] = value; /* value */ in si_dma_vm_set_pte_pde() 391 ib->ptr[ib->length_dw++] = 0; in si_dma_vm_set_pte_pde() [all …]
|
H A D | amdgpu_vcn.c | 582 memset(ib, 0, sizeof(*ib)); in amdgpu_vcn_dec_get_create_msg() 617 memset(ib, 0, sizeof(*ib)); in amdgpu_vcn_dec_get_destroy_msg() 675 ib->ptr[ib->length_dw++] = 0x30000002; in amdgpu_vcn_unified_ring_ib_header() 680 ib->ptr[ib->length_dw++] = 0x30000001; in amdgpu_vcn_unified_ring_ib_header() 861 ib->ptr[ib->length_dw++] = 0x00000018; in amdgpu_vcn_enc_get_create_msg() 863 ib->ptr[ib->length_dw++] = handle; in amdgpu_vcn_enc_get_create_msg() 865 ib->ptr[ib->length_dw++] = addr; in amdgpu_vcn_enc_get_create_msg() 866 ib->ptr[ib->length_dw++] = 0x0000000b; in amdgpu_vcn_enc_get_create_msg() 930 ib->ptr[ib->length_dw++] = handle; in amdgpu_vcn_enc_get_destroy_msg() 932 ib->ptr[ib->length_dw++] = addr; in amdgpu_vcn_enc_get_destroy_msg() [all …]
|
H A D | sdma_v2_4.c | 605 memset(&ib, 0, sizeof(ib)); in sdma_v2_4_ring_test_ib() 665 ib->ptr[ib->length_dw++] = bytes; in sdma_v2_4_vm_copy_pte() 669 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v2_4_vm_copy_pte() 692 ib->ptr[ib->length_dw++] = pe; in sdma_v2_4_vm_write_pte() 694 ib->ptr[ib->length_dw++] = ndw; in sdma_v2_4_vm_write_pte() 727 ib->ptr[ib->length_dw++] = 0; in sdma_v2_4_vm_set_pte_pde() 747 ib->ptr[ib->length_dw++] = in sdma_v2_4_ring_pad_ib() 751 ib->ptr[ib->length_dw++] = in sdma_v2_4_ring_pad_ib() 1201 ib->ptr[ib->length_dw++] = byte_count; in sdma_v2_4_emit_copy_buffer() 1227 ib->ptr[ib->length_dw++] = src_data; in sdma_v2_4_emit_fill_buffer() [all …]
|
H A D | amdgpu_cik_sdma.c | 672 memset(&ib, 0, sizeof(ib)); in cik_sdma_ring_test_ib() 728 ib->ptr[ib->length_dw++] = bytes; in cik_sdma_vm_copy_pte() 732 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cik_sdma_vm_copy_pte() 733 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in cik_sdma_vm_copy_pte() 757 ib->ptr[ib->length_dw++] = ndw; in cik_sdma_vm_write_pte() 790 ib->ptr[ib->length_dw++] = 0; in cik_sdma_vm_set_pte_pde() 810 ib->ptr[ib->length_dw++] = in cik_sdma_ring_pad_ib() 814 ib->ptr[ib->length_dw++] = in cik_sdma_ring_pad_ib() 1312 ib->ptr[ib->length_dw++] = byte_count; in cik_sdma_emit_copy_buffer() 1338 ib->ptr[ib->length_dw++] = src_data; in cik_sdma_emit_fill_buffer() [all …]
|
H A D | sdma_v3_0.c | 877 memset(&ib, 0, sizeof(ib)); in sdma_v3_0_ring_test_ib() 936 ib->ptr[ib->length_dw++] = bytes; in sdma_v3_0_vm_copy_pte() 940 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v3_0_vm_copy_pte() 941 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v3_0_vm_copy_pte() 965 ib->ptr[ib->length_dw++] = ndw; in sdma_v3_0_vm_write_pte() 998 ib->ptr[ib->length_dw++] = 0; in sdma_v3_0_vm_set_pte_pde() 1018 ib->ptr[ib->length_dw++] = in sdma_v3_0_ring_pad_ib() 1022 ib->ptr[ib->length_dw++] = in sdma_v3_0_ring_pad_ib() 1639 ib->ptr[ib->length_dw++] = byte_count; in sdma_v3_0_emit_copy_buffer() 1665 ib->ptr[ib->length_dw++] = src_data; in sdma_v3_0_emit_fill_buffer() [all …]
|
H A D | sdma_v6_0.c | 961 memset(&ib, 0, sizeof(ib)); in sdma_v6_0_ring_test_ib() 1054 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v6_0_vm_copy_pte() 1058 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v6_0_vm_copy_pte() 1059 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v6_0_vm_copy_pte() 1084 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v6_0_vm_write_pte() 1118 ib->ptr[ib->length_dw++] = 0; in sdma_v6_0_vm_set_pte_pde() 1138 ib->ptr[ib->length_dw++] = in sdma_v6_0_ring_pad_ib() 1142 ib->ptr[ib->length_dw++] = in sdma_v6_0_ring_pad_ib() 1640 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v6_0_emit_copy_buffer() 1666 ib->ptr[ib->length_dw++] = src_data; in sdma_v6_0_emit_fill_buffer() [all …]
|
H A D | sdma_v5_2.c | 922 memset(&ib, 0, sizeof(ib)); in sdma_v5_2_ring_test_ib() 1015 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v5_2_vm_copy_pte() 1019 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v5_2_vm_copy_pte() 1020 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v5_2_vm_copy_pte() 1045 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v5_2_vm_write_pte() 1079 ib->ptr[ib->length_dw++] = 0; in sdma_v5_2_vm_set_pte_pde() 1100 ib->ptr[ib->length_dw++] = in sdma_v5_2_ring_pad_ib() 1104 ib->ptr[ib->length_dw++] = in sdma_v5_2_ring_pad_ib() 1802 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v5_2_emit_copy_buffer() 1828 ib->ptr[ib->length_dw++] = src_data; in sdma_v5_2_emit_fill_buffer() [all …]
|
H A D | sdma_v5_0.c | 1074 memset(&ib, 0, sizeof(ib)); in sdma_v5_0_ring_test_ib() 1168 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v5_0_vm_copy_pte() 1172 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v5_0_vm_copy_pte() 1173 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v5_0_vm_copy_pte() 1198 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v5_0_vm_write_pte() 1232 ib->ptr[ib->length_dw++] = 0; in sdma_v5_0_vm_set_pte_pde() 1252 ib->ptr[ib->length_dw++] = in sdma_v5_0_ring_pad_ib() 1256 ib->ptr[ib->length_dw++] = in sdma_v5_0_ring_pad_ib() 1845 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v5_0_emit_copy_buffer() 1871 ib->ptr[ib->length_dw++] = src_data; in sdma_v5_0_emit_fill_buffer() [all …]
|
H A D | sdma_v4_4_2.c | 1029 memset(&ib, 0, sizeof(ib)); in sdma_v4_4_2_ring_test_ib() 1090 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v4_4_2_vm_copy_pte() 1094 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v4_4_2_vm_copy_pte() 1095 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v4_4_2_vm_copy_pte() 1120 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v4_4_2_vm_write_pte() 1154 ib->ptr[ib->length_dw++] = 0; in sdma_v4_4_2_vm_set_pte_pde() 1173 ib->ptr[ib->length_dw++] = in sdma_v4_4_2_ring_pad_ib() 1177 ib->ptr[ib->length_dw++] = in sdma_v4_4_2_ring_pad_ib() 1969 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v4_4_2_emit_copy_buffer() 1995 ib->ptr[ib->length_dw++] = src_data; in sdma_v4_4_2_emit_fill_buffer() [all …]
|
H A D | uvd_v6_0.c | 228 ib->ptr[ib->length_dw++] = 0x00000018; in uvd_v6_0_enc_get_create_msg() 230 ib->ptr[ib->length_dw++] = handle; in uvd_v6_0_enc_get_create_msg() 231 ib->ptr[ib->length_dw++] = 0x00010000; in uvd_v6_0_enc_get_create_msg() 233 ib->ptr[ib->length_dw++] = addr; in uvd_v6_0_enc_get_create_msg() 235 ib->ptr[ib->length_dw++] = 0x00000014; in uvd_v6_0_enc_get_create_msg() 237 ib->ptr[ib->length_dw++] = 0x0000001c; in uvd_v6_0_enc_get_create_msg() 238 ib->ptr[ib->length_dw++] = 0x00000001; in uvd_v6_0_enc_get_create_msg() 239 ib->ptr[ib->length_dw++] = 0x00000000; in uvd_v6_0_enc_get_create_msg() 241 ib->ptr[ib->length_dw++] = 0x00000008; in uvd_v6_0_enc_get_create_msg() 294 ib->ptr[ib->length_dw++] = handle; in uvd_v6_0_enc_get_destroy_msg() [all …]
|
H A D | sdma_v4_0.c | 1493 memset(&ib, 0, sizeof(ib)); in sdma_v4_0_ring_test_ib() 1554 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v4_0_vm_copy_pte() 1558 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v4_0_vm_copy_pte() 1559 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v4_0_vm_copy_pte() 1584 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v4_0_vm_write_pte() 1618 ib->ptr[ib->length_dw++] = 0; in sdma_v4_0_vm_set_pte_pde() 1637 ib->ptr[ib->length_dw++] = in sdma_v4_0_ring_pad_ib() 1641 ib->ptr[ib->length_dw++] = in sdma_v4_0_ring_pad_ib() 2472 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v4_0_emit_copy_buffer() 2498 ib->ptr[ib->length_dw++] = src_data; in sdma_v4_0_emit_fill_buffer() [all …]
|
H A D | uvd_v7_0.c | 236 ib->ptr[ib->length_dw++] = 0x00000018; in uvd_v7_0_enc_get_create_msg() 238 ib->ptr[ib->length_dw++] = handle; in uvd_v7_0_enc_get_create_msg() 239 ib->ptr[ib->length_dw++] = 0x00000000; in uvd_v7_0_enc_get_create_msg() 241 ib->ptr[ib->length_dw++] = addr; in uvd_v7_0_enc_get_create_msg() 243 ib->ptr[ib->length_dw++] = 0x00000014; in uvd_v7_0_enc_get_create_msg() 245 ib->ptr[ib->length_dw++] = 0x0000001c; in uvd_v7_0_enc_get_create_msg() 246 ib->ptr[ib->length_dw++] = 0x00000000; in uvd_v7_0_enc_get_create_msg() 247 ib->ptr[ib->length_dw++] = 0x00000000; in uvd_v7_0_enc_get_create_msg() 249 ib->ptr[ib->length_dw++] = 0x00000008; in uvd_v7_0_enc_get_create_msg() 301 ib->ptr[ib->length_dw++] = handle; in uvd_v7_0_enc_get_destroy_msg() [all …]
|
H A D | amdgpu_ib.c | 66 struct amdgpu_ib *ib) in amdgpu_ib_get() argument 72 &ib->sa_bo, size); in amdgpu_ib_get() 78 ib->ptr = amdgpu_sa_bo_cpu_addr(ib->sa_bo); in amdgpu_ib_get() 80 ib->flags = AMDGPU_IB_FLAG_EMIT_MEM_SYNC; in amdgpu_ib_get() 83 ib->gpu_addr = amdgpu_sa_bo_gpu_addr(ib->sa_bo); in amdgpu_ib_get() 101 amdgpu_sa_bo_free(adev, &ib->sa_bo, f); in amdgpu_ib_free() 131 struct amdgpu_ib *ib = &ibs[0]; in amdgpu_ib_schedule() local 178 if ((ib->flags & AMDGPU_IB_FLAGS_SECURE) && in amdgpu_ib_schedule() 248 secure = ib->flags & AMDGPU_IB_FLAGS_SECURE; in amdgpu_ib_schedule() 253 ib = &ibs[i]; in amdgpu_ib_schedule() [all …]
|
H A D | amdgpu_ring.h | 178 struct amdgpu_ib *ib); 181 struct amdgpu_ib *ib); 188 struct amdgpu_ib *ib, 208 void (*pad_ib)(struct amdgpu_ring *ring, struct amdgpu_ib *ib); 305 #define amdgpu_ring_parse_cs(r, p, job, ib) ((r)->funcs->parse_cs((p), (job), (ib))) argument 306 #define amdgpu_ring_patch_cs_in_place(r, p, job, ib) ((r)->funcs->patch_cs_in_place((p), (job), (ib… argument 312 #define amdgpu_ring_emit_ib(r, job, ib, flags) ((r)->funcs->emit_ib((r), (job), (ib), (flags))) argument 326 #define amdgpu_ring_pad_ib(r, ib) ((r)->funcs->pad_ib((r), (ib))) argument 429 return ib->ptr[idx]; in amdgpu_ib_get_value() 435 ib->ptr[idx] = value; in amdgpu_ib_set_value() [all …]
|
/openbsd/sys/dev/pci/drm/radeon/ |
H A D | radeon_ib.c | 74 ib->ptr = radeon_sa_bo_cpu_addr(ib->sa_bo); in radeon_ib_get() 75 ib->vm = vm; in radeon_ib_get() 82 ib->gpu_addr = radeon_sa_bo_gpu_addr(ib->sa_bo); in radeon_ib_get() 99 radeon_sync_free(rdev, &ib->sync, ib->fence); in radeon_ib_free() 100 radeon_sa_bo_free(&ib->sa_bo, ib->fence); in radeon_ib_free() 152 r = radeon_sync_rings(rdev, &ib->sync, ib->ring); in radeon_ib_schedule() 159 if (ib->vm) in radeon_ib_schedule() 160 radeon_vm_flush(rdev, ib->vm, ib->ring, in radeon_ib_schedule() 167 radeon_ring_ib_execute(rdev, ib->ring, ib); in radeon_ib_schedule() 168 r = radeon_fence_emit(rdev, &ib->fence, ib->ring); in radeon_ib_schedule() [all …]
|
H A D | si_dma.c | 80 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pages() 81 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pages() 120 ib->ptr[ib->length_dw++] = pe; in si_dma_vm_write_pages() 132 ib->ptr[ib->length_dw++] = value; in si_dma_vm_write_pages() 133 ib->ptr[ib->length_dw++] = upper_32_bits(value); in si_dma_vm_write_pages() 172 ib->ptr[ib->length_dw++] = pe; /* dst addr */ in si_dma_vm_set_pages() 174 ib->ptr[ib->length_dw++] = flags; /* mask */ in si_dma_vm_set_pages() 175 ib->ptr[ib->length_dw++] = 0; in si_dma_vm_set_pages() 176 ib->ptr[ib->length_dw++] = value; /* value */ in si_dma_vm_set_pages() 177 ib->ptr[ib->length_dw++] = upper_32_bits(value); in si_dma_vm_set_pages() [all …]
|
H A D | ni_dma.c | 125 unsigned vm_id = ib->vm ? ib->vm->ids[ib->ring].id : 0; in cayman_dma_ring_ib_execute() 328 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cayman_dma_vm_copy_pages() 329 ib->ptr[ib->length_dw++] = lower_32_bits(src); in cayman_dma_vm_copy_pages() 369 ib->ptr[ib->length_dw++] = pe; in cayman_dma_vm_write_pages() 381 ib->ptr[ib->length_dw++] = value; in cayman_dma_vm_write_pages() 421 ib->ptr[ib->length_dw++] = pe; /* dst addr */ in cayman_dma_vm_set_pages() 423 ib->ptr[ib->length_dw++] = flags; /* mask */ in cayman_dma_vm_set_pages() 424 ib->ptr[ib->length_dw++] = 0; in cayman_dma_vm_set_pages() 425 ib->ptr[ib->length_dw++] = value; /* value */ in cayman_dma_vm_set_pages() 426 ib->ptr[ib->length_dw++] = upper_32_bits(value); in cayman_dma_vm_set_pages() [all …]
|
H A D | radeon_vce.c | 395 ib.ptr[ib.length_dw++] = cpu_to_le32(handle); in radeon_vce_get_create_msg() 399 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000000); in radeon_vce_get_create_msg() 400 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000042); in radeon_vce_get_create_msg() 401 ib.ptr[ib.length_dw++] = cpu_to_le32(0x0000000a); in radeon_vce_get_create_msg() 402 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000001); in radeon_vce_get_create_msg() 403 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000080); in radeon_vce_get_create_msg() 404 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000060); in radeon_vce_get_create_msg() 405 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000100); in radeon_vce_get_create_msg() 413 ib.ptr[ib.length_dw++] = cpu_to_le32(dummy); in radeon_vce_get_create_msg() 462 ib.ptr[ib.length_dw++] = cpu_to_le32(handle); in radeon_vce_get_destroy_msg() [all …]
|
H A D | cik_sdma.c | 136 u32 extra_bits = (ib->vm ? ib->vm->ids[ib->ring].id : 0) & 0xf; in cik_sdma_ring_ib_execute() 814 ib->ptr[ib->length_dw++] = bytes; in cik_sdma_vm_copy_pages() 818 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cik_sdma_vm_copy_pages() 819 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in cik_sdma_vm_copy_pages() 857 ib->ptr[ib->length_dw++] = pe; in cik_sdma_vm_write_pages() 858 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in cik_sdma_vm_write_pages() 859 ib->ptr[ib->length_dw++] = ndw; in cik_sdma_vm_write_pages() 870 ib->ptr[ib->length_dw++] = value; in cik_sdma_vm_write_pages() 912 ib->ptr[ib->length_dw++] = flags; /* mask */ in cik_sdma_vm_set_pages() 913 ib->ptr[ib->length_dw++] = 0; in cik_sdma_vm_set_pages() [all …]
|
H A D | evergreen_cs.c | 450 uint32_t *ib = p->ib.ptr; in evergreen_cs_track_validate_cb() local 1097 u32 tmp, *ib; in evergreen_cs_handle_reg() local 1100 ib = p->ib.ptr; in evergreen_cs_handle_reg() 1598 ib[idx] |= 3; in evergreen_cs_handle_reg() 1776 uint32_t *ib; in evergreen_packet3_check() local 1784 ib = p->ib.ptr; in evergreen_packet3_check() 1868 ib[idx+0] = offset; in evergreen_packet3_check() 1903 ib[idx+0] = offset; in evergreen_packet3_check() 2104 ib[idx+1] = (ib[idx+1] & 0x3) | (offset & 0xfffffffc); in evergreen_packet3_check() 2425 ib[idx+1+(i*8)+2] = (ib[idx+1+(i*8)+2] & 0xffffff00) | in evergreen_packet3_check() [all …]
|
H A D | radeon_vm.c | 361 struct radeon_ib *ib, in radeon_vm_set_pages() argument 392 struct radeon_ib ib; in radeon_vm_clear_bo() local 412 ib.length_dw = 0; in radeon_vm_clear_bo() 426 radeon_ib_free(rdev, &ib); in radeon_vm_clear_bo() 648 struct radeon_ib ib; in radeon_vm_update_page_directory() local 664 ib.length_dw = 0; in radeon_vm_update_page_directory() 701 if (ib.length_dw != 0) { in radeon_vm_update_page_directory() 732 struct radeon_ib *ib, in radeon_vm_frag_ptes() argument 816 struct radeon_ib *ib, in radeon_vm_update_ptes() argument 916 struct radeon_ib ib; in radeon_vm_bo_update() local [all …]
|
H A D | r600_cs.c | 356 volatile u32 *ib = p->ib.ptr; in r600_cs_track_validate_cb() local 525 volatile u32 *ib = p->ib.ptr; in r600_cs_track_validate_db() local 834 ib = p->ib.ptr; in r600_cs_common_vline_parse() 970 u32 m, i, tmp, *ib; in r600_cs_check_reg() local 981 ib = p->ib.ptr; in r600_cs_check_reg() 1313 ib[idx] |= 3; in r600_cs_check_reg() 1631 volatile u32 *ib; in r600_packet3_check() local 1639 ib = p->ib.ptr; in r600_packet3_check() 1770 ib[idx+1] = (ib[idx+1] & 0x3) | (offset & 0xfffffff0); in r600_packet3_check() 2011 ib[idx+1+(i*8)+2] = (ib[idx+1+(i*8)+2] & 0xffffff00) | in r600_packet3_check() [all …]
|
H A D | r600_dma.c | 338 struct radeon_ib ib; in r600_dma_ib_test() local 358 ib.ptr[0] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 1); in r600_dma_ib_test() 359 ib.ptr[1] = lower_32_bits(gpu_addr); in r600_dma_ib_test() 360 ib.ptr[2] = upper_32_bits(gpu_addr) & 0xff; in r600_dma_ib_test() 361 ib.ptr[3] = 0xDEADBEEF; in r600_dma_ib_test() 362 ib.length_dw = 4; in r600_dma_ib_test() 364 r = radeon_ib_schedule(rdev, &ib, NULL, false); in r600_dma_ib_test() 366 radeon_ib_free(rdev, &ib); in r600_dma_ib_test() 392 radeon_ib_free(rdev, &ib); in r600_dma_ib_test() 406 struct radeon_ring *ring = &rdev->ring[ib->ring]; in r600_dma_ring_ib_execute() [all …]
|
/openbsd/gnu/llvm/lld/MachO/ |
H A D | ICF.cpp | 45 const ConcatInputSection *ib); 47 const ConcatInputSection *ib); 100 if (ia->parent != ib->parent) in equalsConstant() 102 if (ia->data.size() != ib->data.size()) in equalsConstant() 104 if (ia->data != ib->data) in equalsConstant() 106 if (ia->relocs.size() != ib->relocs.size()) in equalsConstant() 180 assert(ia->relocs.size() == ib->relocs.size()); in equalsVariable() 217 auto itB = std::find_if(ib->symbols.begin(), ib->symbols.end(), hasUnwind); in equalsVariable() 219 return itB == ib->symbols.end(); in equalsVariable() 220 if (itB == ib->symbols.end()) in equalsVariable() [all …]
|