/dragonfly/sys/dev/drm/radeon/ |
H A D | r300.c | 205 struct radeon_fence *fence) in r300_fence_ring_emit() argument 207 struct radeon_ring *ring = &rdev->ring[fence->ring]; in r300_fence_ring_emit() 232 radeon_ring_write(ring, PACKET0(rdev->fence_drv[fence->ring].scratch_reg, 0)); in r300_fence_ring_emit() 233 radeon_ring_write(ring, fence->seq); in r300_fence_ring_emit()
|
H A D | r600.c | 2877 struct radeon_fence *fence) in r600_fence_ring_emit() argument 2879 struct radeon_ring *ring = &rdev->ring[fence->ring]; in r600_fence_ring_emit() 2887 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in r600_fence_ring_emit() 2899 radeon_ring_write(ring, fence->seq); in r600_fence_ring_emit() 2917 radeon_ring_write(ring, fence->seq); in r600_fence_ring_emit() 2978 struct radeon_fence *fence; in r600_copy_cpdma() local 3024 r = radeon_fence_emit(rdev, &fence, ring->idx); in r600_copy_cpdma() 3032 radeon_sync_free(rdev, &sync, fence); in r600_copy_cpdma() 3034 return fence; in r600_copy_cpdma() 3438 r = radeon_fence_wait_timeout(ib.fence, false, usecs_to_jiffies( in r600_ib_test() [all …]
|
H A D | cik.c | 3581 struct radeon_fence *fence) in cik_fence_gfx_ring_emit() argument 3584 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in cik_fence_gfx_ring_emit() 3597 radeon_ring_write(ring, fence->seq - 1); in cik_fence_gfx_ring_emit() 3608 radeon_ring_write(ring, fence->seq); in cik_fence_gfx_ring_emit() 3622 struct radeon_fence *fence) in cik_fence_compute_ring_emit() argument 3625 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in cik_fence_compute_ring_emit() 3636 radeon_ring_write(ring, fence->seq); in cik_fence_compute_ring_emit() 3690 struct radeon_fence *fence; in cik_copy_cpdma() local 3731 r = radeon_fence_emit(rdev, &fence, ring->idx); in cik_copy_cpdma() 3739 radeon_sync_free(rdev, &sync, fence); in cik_copy_cpdma() [all …]
|
H A D | ni.c | 1412 struct radeon_fence *fence) in cayman_fence_ring_emit() argument 1414 struct radeon_ring *ring = &rdev->ring[fence->ring]; in cayman_fence_ring_emit() 1415 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in cayman_fence_ring_emit() 1430 radeon_ring_write(ring, fence->seq); in cayman_fence_ring_emit()
|
H A D | r100.c | 853 struct radeon_fence *fence) in r100_fence_ring_emit() argument 855 struct radeon_ring *ring = &rdev->ring[fence->ring]; in r100_fence_ring_emit() 868 radeon_ring_write(ring, PACKET0(rdev->fence_drv[fence->ring].scratch_reg, 0)); in r100_fence_ring_emit() 869 radeon_ring_write(ring, fence->seq); in r100_fence_ring_emit() 891 struct radeon_fence *fence; in r100_copy_blit() local 952 r = radeon_fence_emit(rdev, &fence, RADEON_RING_TYPE_GFX_INDEX); in r100_copy_blit() 958 return fence; in r100_copy_blit() 3749 r = radeon_fence_wait_timeout(ib.fence, false, usecs_to_jiffies( in r100_ib_test()
|
H A D | radeon_cs.c | 429 &parser->ib.fence->base); in radeon_cs_parser_fini()
|
/dragonfly/contrib/gcc-4.7/gcc/config/i386/ |
H A D | sync.md | 71 (set_attr "atom_sse_attr" "fence") 90 (set_attr "atom_sse_attr" "fence") 105 /* Unless this is a SEQ_CST fence, the i386 memory model is strong
|
H A D | atom.md | 26 (define_attr "atom_sse_attr" "rcp,movdup,lfence,fence,prefetch,sqrt,mxcsr,other" 547 (ior (eq_attr "atom_sse_attr" "fence")
|
/dragonfly/sys/dev/drm/i915/ |
H A D | intel_breadcrumbs.c | 610 &request->fence.flags)) { in intel_breadcrumbs_signaler() 612 dma_fence_signal(&request->fence); in intel_breadcrumbs_signaler()
|
H A D | intel_guc_fwif.h | 341 u32 fence; /* fence updated by GuC */ member
|
H A D | i915_vma.c | 672 GEM_BUG_ON(vma->fence); in i915_vma_destroy() 839 GEM_BUG_ON(vma->fence); in i915_vma_unbind()
|
H A D | intel_lrc.c | 727 dma_fence_set_error(&rq->fence, -EIO); in execlists_cancel_requests() 738 dma_fence_set_error(&rq->fence, -EIO); in execlists_cancel_requests() 1566 if (!request || request->fence.error != -EIO) in reset_common_ring()
|
H A D | intel_fbdev.c | 51 unsigned int origin = ifbdev->vma->fence ? ORIGIN_GTT : ORIGIN_CPU; in intel_fbdev_invalidate()
|
H A D | intel_ringbuffer.c | 636 if (request->fence.error == -EIO) in reset_ring_common() 803 dma_fence_set_error(&request->fence, -EIO); in cancel_requests()
|
H A D | intel_display.c | 12433 intel_atomic_commit_ready(struct i915_sw_fence *fence, in intel_atomic_commit_ready() argument 12601 struct dma_fence *fence) in add_rps_boost_after_vblank() argument 12605 if (!dma_fence_is_i915(fence)) in add_rps_boost_after_vblank() 12620 wait->request = to_request(dma_fence_get(fence)); in add_rps_boost_after_vblank() 12681 if (new_state->fence) { /* explicit fencing */ in intel_prepare_plane_fb() 12683 new_state->fence, in intel_prepare_plane_fb() 12725 if (!new_state->fence) { /* implicit fencing */ in intel_prepare_plane_fb() 12726 struct dma_fence *fence; in intel_prepare_plane_fb() local 12735 fence = reservation_object_get_excl_rcu(obj->resv); in intel_prepare_plane_fb() 12736 if (fence) { in intel_prepare_plane_fb() [all …]
|
H A D | i915_drv.h | 933 u64 fence[I915_MAX_NUM_FENCES]; member 3797 void i915_unreserve_fence(struct drm_i915_fence_reg *fence); 4387 if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &req->fence.flags)) in __i915_request_irq_complete()
|
/dragonfly/sys/dev/drm/amd/include/ |
H A D | kgd_kfd_interface.h | 457 struct dma_fence *fence);
|
/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_ring.h | 97 int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **fence,
|
H A D | amdgpu.h | 442 struct dma_fence *fence; member 668 struct dma_fence *fence, uint64_t *seq); 1002 struct dma_fence *fence; member
|
/dragonfly/contrib/gcc-8.0/gcc/config/i386/ |
H A D | sync.md | 77 (set_attr "atom_sse_attr" "fence") 96 (set_attr "atom_sse_attr" "fence") 113 /* Unless this is a SEQ_CST fence, the i386 memory model is strong
|
H A D | atom.md | 26 (define_attr "atom_sse_attr" "rcp,movdup,lfence,fence,prefetch,sqrt,mxcsr,other" 547 (eq_attr "atom_sse_attr" "fence,prefetch")))
|
/dragonfly/sys/dev/drm/include/drm/ |
H A D | drm_atomic.h | 561 struct dma_fence *fence);
|
/dragonfly/sys/dev/drm/include/uapi/drm/ |
H A D | amdgpu_drm.h | 601 struct drm_amdgpu_fence fence; member
|
/dragonfly/contrib/gcc-8.0/gcc/ |
H A D | sel-sched-ir.h | 325 #define FLIST_FENCE(L) (&(L)->u.fence) 350 struct _fence fence; member
|
/dragonfly/contrib/gcc-4.7/gcc/ |
H A D | sel-sched-ir.h | 340 #define FLIST_FENCE(L) (&(L)->u.fence) 364 struct _fence fence; member
|