Lines Matching refs:fence
135 struct radeon_fence **fence, in radeon_fence_emit() argument
141 *fence = kmalloc(sizeof(struct radeon_fence), GFP_KERNEL); in radeon_fence_emit()
142 if ((*fence) == NULL) { in radeon_fence_emit()
145 (*fence)->rdev = rdev; in radeon_fence_emit()
146 (*fence)->seq = seq = ++rdev->fence_drv[ring].sync_seq[ring]; in radeon_fence_emit()
147 (*fence)->ring = ring; in radeon_fence_emit()
148 (*fence)->is_vm_update = false; in radeon_fence_emit()
149 dma_fence_init(&(*fence)->base, &radeon_fence_ops, in radeon_fence_emit()
153 radeon_fence_ring_emit(rdev, ring, *fence); in radeon_fence_emit()
154 trace_radeon_fence_emit(rdev_to_drm(rdev), ring, (*fence)->seq); in radeon_fence_emit()
168 struct radeon_fence *fence; in radeon_fence_check_signaled() local
171 fence = container_of(wait, struct radeon_fence, fence_wake); in radeon_fence_check_signaled()
177 seq = atomic64_read(&fence->rdev->fence_drv[fence->ring].last_seq); in radeon_fence_check_signaled()
178 if (seq >= fence->seq) { in radeon_fence_check_signaled()
179 dma_fence_signal_locked(&fence->base); in radeon_fence_check_signaled()
180 radeon_irq_kms_sw_irq_put(fence->rdev, fence->ring); in radeon_fence_check_signaled()
181 __remove_wait_queue(&fence->rdev->fence_queue, &fence->fence_wake); in radeon_fence_check_signaled()
182 dma_fence_put(&fence->base); in radeon_fence_check_signaled()
356 struct radeon_fence *fence = to_radeon_fence(f); in radeon_fence_is_signaled() local
357 struct radeon_device *rdev = fence->rdev; in radeon_fence_is_signaled()
358 unsigned ring = fence->ring; in radeon_fence_is_signaled()
359 u64 seq = fence->seq; in radeon_fence_is_signaled()
386 struct radeon_fence *fence = to_radeon_fence(f); in radeon_fence_enable_signaling() local
387 struct radeon_device *rdev = fence->rdev; in radeon_fence_enable_signaling()
389 if (atomic64_read(&rdev->fence_drv[fence->ring].last_seq) >= fence->seq) in radeon_fence_enable_signaling()
393 radeon_irq_kms_sw_irq_get(rdev, fence->ring); in radeon_fence_enable_signaling()
395 if (radeon_fence_activity(rdev, fence->ring)) in radeon_fence_enable_signaling()
399 if (atomic64_read(&rdev->fence_drv[fence->ring].last_seq) >= fence->seq) { in radeon_fence_enable_signaling()
400 radeon_irq_kms_sw_irq_put(rdev, fence->ring); in radeon_fence_enable_signaling()
408 if (radeon_irq_kms_sw_irq_get_delayed(rdev, fence->ring)) in radeon_fence_enable_signaling()
409 rdev->fence_drv[fence->ring].delayed_irq = true; in radeon_fence_enable_signaling()
410 radeon_fence_schedule_check(rdev, fence->ring); in radeon_fence_enable_signaling()
413 fence->fence_wake.flags = 0; in radeon_fence_enable_signaling()
414 fence->fence_wake.private = NULL; in radeon_fence_enable_signaling()
415 fence->fence_wake.func = radeon_fence_check_signaled; in radeon_fence_enable_signaling()
416 __add_wait_queue(&rdev->fence_queue, &fence->fence_wake); in radeon_fence_enable_signaling()
429 bool radeon_fence_signaled(struct radeon_fence *fence) in radeon_fence_signaled() argument
431 if (!fence) in radeon_fence_signaled()
434 if (radeon_fence_seq_signaled(fence->rdev, fence->seq, fence->ring)) { in radeon_fence_signaled()
435 dma_fence_signal(&fence->base); in radeon_fence_signaled()
536 long radeon_fence_wait_timeout(struct radeon_fence *fence, bool intr, long timeout) in radeon_fence_wait_timeout() argument
547 if (WARN_ON_ONCE(!to_radeon_fence(&fence->base))) in radeon_fence_wait_timeout()
548 return dma_fence_wait(&fence->base, intr); in radeon_fence_wait_timeout()
550 seq[fence->ring] = fence->seq; in radeon_fence_wait_timeout()
551 r = radeon_fence_wait_seq_timeout(fence->rdev, seq, intr, timeout); in radeon_fence_wait_timeout()
556 dma_fence_signal(&fence->base); in radeon_fence_wait_timeout()
571 int radeon_fence_wait(struct radeon_fence *fence, bool intr) in radeon_fence_wait() argument
573 long r = radeon_fence_wait_timeout(fence, intr, MAX_SCHEDULE_TIMEOUT); in radeon_fence_wait()
689 struct radeon_fence *radeon_fence_ref(struct radeon_fence *fence) in radeon_fence_ref() argument
691 dma_fence_get(&fence->base); in radeon_fence_ref()
692 return fence; in radeon_fence_ref()
702 void radeon_fence_unref(struct radeon_fence **fence) in radeon_fence_unref() argument
704 struct radeon_fence *tmp = *fence; in radeon_fence_unref()
706 *fence = NULL; in radeon_fence_unref()
750 bool radeon_fence_need_sync(struct radeon_fence *fence, int dst_ring) in radeon_fence_need_sync() argument
754 if (!fence) { in radeon_fence_need_sync()
758 if (fence->ring == dst_ring) { in radeon_fence_need_sync()
763 fdrv = &fence->rdev->fence_drv[dst_ring]; in radeon_fence_need_sync()
764 if (fence->seq <= fdrv->sync_seq[fence->ring]) { in radeon_fence_need_sync()
780 void radeon_fence_note_sync(struct radeon_fence *fence, int dst_ring) in radeon_fence_note_sync() argument
785 if (!fence) { in radeon_fence_note_sync()
789 if (fence->ring == dst_ring) { in radeon_fence_note_sync()
794 src = &fence->rdev->fence_drv[fence->ring]; in radeon_fence_note_sync()
795 dst = &fence->rdev->fence_drv[dst_ring]; in radeon_fence_note_sync()
1018 static const char *radeon_fence_get_driver_name(struct dma_fence *fence) in radeon_fence_get_driver_name() argument
1025 struct radeon_fence *fence = to_radeon_fence(f); in radeon_fence_get_timeline_name() local
1026 switch (fence->ring) { in radeon_fence_get_timeline_name()
1039 static inline bool radeon_test_signaled(struct radeon_fence *fence) in radeon_test_signaled() argument
1041 return test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->base.flags); in radeon_test_signaled()
1050 radeon_fence_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb) in radeon_fence_wait_cb() argument
1061 struct radeon_fence *fence = to_radeon_fence(f); in radeon_fence_default_wait() local
1062 struct radeon_device *rdev = fence->rdev; in radeon_fence_default_wait()
1080 if (radeon_test_signaled(fence)) in radeon_fence_default_wait()