xref: /openbsd/sys/dev/pci/drm/radeon/rv770_dma.c (revision c349dbc7)
17ccd5a2cSjsg /*
27ccd5a2cSjsg  * Copyright 2013 Advanced Micro Devices, Inc.
37ccd5a2cSjsg  *
47ccd5a2cSjsg  * Permission is hereby granted, free of charge, to any person obtaining a
57ccd5a2cSjsg  * copy of this software and associated documentation files (the "Software"),
67ccd5a2cSjsg  * to deal in the Software without restriction, including without limitation
77ccd5a2cSjsg  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
87ccd5a2cSjsg  * and/or sell copies of the Software, and to permit persons to whom the
97ccd5a2cSjsg  * Software is furnished to do so, subject to the following conditions:
107ccd5a2cSjsg  *
117ccd5a2cSjsg  * The above copyright notice and this permission notice shall be included in
127ccd5a2cSjsg  * all copies or substantial portions of the Software.
137ccd5a2cSjsg  *
147ccd5a2cSjsg  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
157ccd5a2cSjsg  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
167ccd5a2cSjsg  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
177ccd5a2cSjsg  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
187ccd5a2cSjsg  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
197ccd5a2cSjsg  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
207ccd5a2cSjsg  * OTHER DEALINGS IN THE SOFTWARE.
217ccd5a2cSjsg  *
227ccd5a2cSjsg  * Authors: Alex Deucher
237ccd5a2cSjsg  */
24*c349dbc7Sjsg 
257ccd5a2cSjsg #include "radeon.h"
267ccd5a2cSjsg #include "radeon_asic.h"
277ccd5a2cSjsg #include "rv770d.h"
287ccd5a2cSjsg 
297ccd5a2cSjsg /**
307ccd5a2cSjsg  * rv770_copy_dma - copy pages using the DMA engine
317ccd5a2cSjsg  *
327ccd5a2cSjsg  * @rdev: radeon_device pointer
337ccd5a2cSjsg  * @src_offset: src GPU address
347ccd5a2cSjsg  * @dst_offset: dst GPU address
357ccd5a2cSjsg  * @num_gpu_pages: number of GPU pages to xfer
367ccd5a2cSjsg  * @resv: reservation object to sync to
377ccd5a2cSjsg  *
387ccd5a2cSjsg  * Copy GPU paging using the DMA engine (r7xx).
397ccd5a2cSjsg  * Used by the radeon ttm implementation to move pages if
407ccd5a2cSjsg  * registered as the asic copy callback.
417ccd5a2cSjsg  */
rv770_copy_dma(struct radeon_device * rdev,uint64_t src_offset,uint64_t dst_offset,unsigned num_gpu_pages,struct dma_resv * resv)427ccd5a2cSjsg struct radeon_fence *rv770_copy_dma(struct radeon_device *rdev,
437ccd5a2cSjsg 				    uint64_t src_offset, uint64_t dst_offset,
447ccd5a2cSjsg 				    unsigned num_gpu_pages,
45*c349dbc7Sjsg 				    struct dma_resv *resv)
467ccd5a2cSjsg {
477ccd5a2cSjsg 	struct radeon_fence *fence;
487ccd5a2cSjsg 	struct radeon_sync sync;
497ccd5a2cSjsg 	int ring_index = rdev->asic->copy.dma_ring_index;
507ccd5a2cSjsg 	struct radeon_ring *ring = &rdev->ring[ring_index];
517ccd5a2cSjsg 	u32 size_in_dw, cur_size_in_dw;
527ccd5a2cSjsg 	int i, num_loops;
537ccd5a2cSjsg 	int r = 0;
547ccd5a2cSjsg 
557ccd5a2cSjsg 	radeon_sync_create(&sync);
567ccd5a2cSjsg 
577ccd5a2cSjsg 	size_in_dw = (num_gpu_pages << RADEON_GPU_PAGE_SHIFT) / 4;
587ccd5a2cSjsg 	num_loops = DIV_ROUND_UP(size_in_dw, 0xFFFF);
597ccd5a2cSjsg 	r = radeon_ring_lock(rdev, ring, num_loops * 5 + 8);
607ccd5a2cSjsg 	if (r) {
617ccd5a2cSjsg 		DRM_ERROR("radeon: moving bo (%d).\n", r);
627ccd5a2cSjsg 		radeon_sync_free(rdev, &sync, NULL);
637ccd5a2cSjsg 		return ERR_PTR(r);
647ccd5a2cSjsg 	}
657ccd5a2cSjsg 
667ccd5a2cSjsg 	radeon_sync_resv(rdev, &sync, resv, false);
677ccd5a2cSjsg 	radeon_sync_rings(rdev, &sync, ring->idx);
687ccd5a2cSjsg 
697ccd5a2cSjsg 	for (i = 0; i < num_loops; i++) {
707ccd5a2cSjsg 		cur_size_in_dw = size_in_dw;
717ccd5a2cSjsg 		if (cur_size_in_dw > 0xFFFF)
727ccd5a2cSjsg 			cur_size_in_dw = 0xFFFF;
737ccd5a2cSjsg 		size_in_dw -= cur_size_in_dw;
747ccd5a2cSjsg 		radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_COPY, 0, 0, cur_size_in_dw));
757ccd5a2cSjsg 		radeon_ring_write(ring, dst_offset & 0xfffffffc);
767ccd5a2cSjsg 		radeon_ring_write(ring, src_offset & 0xfffffffc);
777ccd5a2cSjsg 		radeon_ring_write(ring, upper_32_bits(dst_offset) & 0xff);
787ccd5a2cSjsg 		radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff);
797ccd5a2cSjsg 		src_offset += cur_size_in_dw * 4;
807ccd5a2cSjsg 		dst_offset += cur_size_in_dw * 4;
817ccd5a2cSjsg 	}
827ccd5a2cSjsg 
837ccd5a2cSjsg 	r = radeon_fence_emit(rdev, &fence, ring->idx);
847ccd5a2cSjsg 	if (r) {
857ccd5a2cSjsg 		radeon_ring_unlock_undo(rdev, ring);
867ccd5a2cSjsg 		radeon_sync_free(rdev, &sync, NULL);
877ccd5a2cSjsg 		return ERR_PTR(r);
887ccd5a2cSjsg 	}
897ccd5a2cSjsg 
907ccd5a2cSjsg 	radeon_ring_unlock_commit(rdev, ring, false);
917ccd5a2cSjsg 	radeon_sync_free(rdev, &sync, fence);
927ccd5a2cSjsg 
937ccd5a2cSjsg 	return fence;
947ccd5a2cSjsg }
95