xref: /freebsd/sys/arm/allwinner/a10_dmac.c (revision 42249ef2)
1 /*-
2  * Copyright (c) 2014-2016 Jared D. McNeill <jmcneill@invisible.ca>
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
15  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
16  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
17  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
18  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
19  * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
20  * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
21  * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
22  * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  *
26  */
27 
28 /*
29  * Allwinner A10/A20 DMA controller
30  */
31 
32 #include <sys/cdefs.h>
33 __FBSDID("$FreeBSD$");
34 
35 #include <sys/param.h>
36 #include <sys/systm.h>
37 #include <sys/bus.h>
38 #include <sys/rman.h>
39 #include <sys/condvar.h>
40 #include <sys/kernel.h>
41 #include <sys/lock.h>
42 #include <sys/module.h>
43 #include <sys/mutex.h>
44 
45 #include <machine/bus.h>
46 
47 #include <dev/ofw/ofw_bus.h>
48 #include <dev/ofw/ofw_bus_subr.h>
49 
50 #include <arm/allwinner/a10_dmac.h>
51 #include <dev/extres/clk/clk.h>
52 
53 #include "sunxi_dma_if.h"
54 
55 #define	NDMA_CHANNELS	8
56 #define	DDMA_CHANNELS	8
57 
58 enum a10dmac_type {
59 	CH_NDMA,
60 	CH_DDMA
61 };
62 
63 struct a10dmac_softc;
64 
65 struct a10dmac_channel {
66 	struct a10dmac_softc *	ch_sc;
67 	uint8_t			ch_index;
68 	enum a10dmac_type	ch_type;
69 	void			(*ch_callback)(void *);
70 	void *			ch_callbackarg;
71 	uint32_t		ch_regoff;
72 };
73 
74 struct a10dmac_softc {
75 	struct resource *	sc_res[2];
76 	struct mtx		sc_mtx;
77 	void *			sc_ih;
78 
79 	struct a10dmac_channel	sc_ndma_channels[NDMA_CHANNELS];
80 	struct a10dmac_channel	sc_ddma_channels[DDMA_CHANNELS];
81 };
82 
83 static struct resource_spec a10dmac_spec[] = {
84 	{ SYS_RES_MEMORY,	0,	RF_ACTIVE },
85 	{ SYS_RES_IRQ,		0,	RF_ACTIVE },
86 	{ -1, 0 }
87 };
88 
89 #define	DMA_READ(sc, reg)	bus_read_4((sc)->sc_res[0], (reg))
90 #define	DMA_WRITE(sc, reg, val)	bus_write_4((sc)->sc_res[0], (reg), (val))
91 #define	DMACH_READ(ch, reg)		\
92     DMA_READ((ch)->ch_sc, (reg) + (ch)->ch_regoff)
93 #define	DMACH_WRITE(ch, reg, val)	\
94     DMA_WRITE((ch)->ch_sc, (reg) + (ch)->ch_regoff, (val))
95 
96 static void a10dmac_intr(void *);
97 
98 static int
99 a10dmac_probe(device_t dev)
100 {
101 	if (!ofw_bus_status_okay(dev))
102 		return (ENXIO);
103 
104 	if (!ofw_bus_is_compatible(dev, "allwinner,sun4i-a10-dma"))
105 		return (ENXIO);
106 
107 	device_set_desc(dev, "Allwinner DMA controller");
108 	return (BUS_PROBE_DEFAULT);
109 }
110 
111 static int
112 a10dmac_attach(device_t dev)
113 {
114 	struct a10dmac_softc *sc;
115 	unsigned int index;
116 	clk_t clk;
117 	int error;
118 
119 	sc = device_get_softc(dev);
120 
121 	if (bus_alloc_resources(dev, a10dmac_spec, sc->sc_res)) {
122 		device_printf(dev, "cannot allocate resources for device\n");
123 		return (ENXIO);
124 	}
125 
126 	mtx_init(&sc->sc_mtx, "a10 dmac", NULL, MTX_SPIN);
127 
128 	/* Activate DMA controller clock */
129 	error = clk_get_by_ofw_index(dev, 0, 0, &clk);
130 	if (error != 0) {
131 		device_printf(dev, "cannot get clock\n");
132 		return (error);
133 	}
134 	error = clk_enable(clk);
135 	if (error != 0) {
136 		device_printf(dev, "cannot enable clock\n");
137 		return (error);
138 	}
139 
140 	/* Disable all interrupts and clear pending status */
141 	DMA_WRITE(sc, AWIN_DMA_IRQ_EN_REG, 0);
142 	DMA_WRITE(sc, AWIN_DMA_IRQ_PEND_STA_REG, ~0);
143 
144 	/* Initialize channels */
145 	for (index = 0; index < NDMA_CHANNELS; index++) {
146 		sc->sc_ndma_channels[index].ch_sc = sc;
147 		sc->sc_ndma_channels[index].ch_index = index;
148 		sc->sc_ndma_channels[index].ch_type = CH_NDMA;
149 		sc->sc_ndma_channels[index].ch_callback = NULL;
150 		sc->sc_ndma_channels[index].ch_callbackarg = NULL;
151 		sc->sc_ndma_channels[index].ch_regoff = AWIN_NDMA_REG(index);
152 		DMACH_WRITE(&sc->sc_ndma_channels[index], AWIN_NDMA_CTL_REG, 0);
153 	}
154 	for (index = 0; index < DDMA_CHANNELS; index++) {
155 		sc->sc_ddma_channels[index].ch_sc = sc;
156 		sc->sc_ddma_channels[index].ch_index = index;
157 		sc->sc_ddma_channels[index].ch_type = CH_DDMA;
158 		sc->sc_ddma_channels[index].ch_callback = NULL;
159 		sc->sc_ddma_channels[index].ch_callbackarg = NULL;
160 		sc->sc_ddma_channels[index].ch_regoff = AWIN_DDMA_REG(index);
161 		DMACH_WRITE(&sc->sc_ddma_channels[index], AWIN_DDMA_CTL_REG, 0);
162 	}
163 
164 	error = bus_setup_intr(dev, sc->sc_res[1], INTR_MPSAFE | INTR_TYPE_MISC,
165 	    NULL, a10dmac_intr, sc, &sc->sc_ih);
166 	if (error != 0) {
167 		device_printf(dev, "could not setup interrupt handler\n");
168 		bus_release_resources(dev, a10dmac_spec, sc->sc_res);
169 		mtx_destroy(&sc->sc_mtx);
170 		return (ENXIO);
171 	}
172 
173 	OF_device_register_xref(OF_xref_from_node(ofw_bus_get_node(dev)), dev);
174 	return (0);
175 }
176 
177 static void
178 a10dmac_intr(void *priv)
179 {
180 	struct a10dmac_softc *sc = priv;
181 	uint32_t sta, bit, mask;
182 	uint8_t index;
183 
184 	sta = DMA_READ(sc, AWIN_DMA_IRQ_PEND_STA_REG);
185 	DMA_WRITE(sc, AWIN_DMA_IRQ_PEND_STA_REG, sta);
186 
187 	while ((bit = ffs(sta & AWIN_DMA_IRQ_END_MASK)) != 0) {
188 		mask = (1U << (bit - 1));
189 		sta &= ~mask;
190 		/*
191 		 * Map status bit to channel number. The status register is
192 		 * encoded with two bits of status per channel (lowest bit
193 		 * is half transfer pending, highest bit is end transfer
194 		 * pending). The 8 normal DMA channel status are in the lower
195 		 * 16 bits and the 8 dedicated DMA channel status are in
196 		 * the upper 16 bits. The output is a channel number from 0-7.
197 		 */
198 		index = ((bit - 1) / 2) & 7;
199 		if (mask & AWIN_DMA_IRQ_NDMA) {
200 			if (sc->sc_ndma_channels[index].ch_callback == NULL)
201 				continue;
202 			sc->sc_ndma_channels[index].ch_callback(
203 			    sc->sc_ndma_channels[index].ch_callbackarg);
204 		} else {
205 			if (sc->sc_ddma_channels[index].ch_callback == NULL)
206 				continue;
207 			sc->sc_ddma_channels[index].ch_callback(
208 			    sc->sc_ddma_channels[index].ch_callbackarg);
209 		}
210 	}
211 }
212 
213 static uint32_t
214 a10dmac_read_ctl(struct a10dmac_channel *ch)
215 {
216 	if (ch->ch_type == CH_NDMA) {
217 		return (DMACH_READ(ch, AWIN_NDMA_CTL_REG));
218 	} else {
219 		return (DMACH_READ(ch, AWIN_DDMA_CTL_REG));
220 	}
221 }
222 
223 static void
224 a10dmac_write_ctl(struct a10dmac_channel *ch, uint32_t val)
225 {
226 	if (ch->ch_type == CH_NDMA) {
227 		DMACH_WRITE(ch, AWIN_NDMA_CTL_REG, val);
228 	} else {
229 		DMACH_WRITE(ch, AWIN_DDMA_CTL_REG, val);
230 	}
231 }
232 
233 static int
234 a10dmac_set_config(device_t dev, void *priv, const struct sunxi_dma_config *cfg)
235 {
236 	struct a10dmac_channel *ch = priv;
237 	uint32_t val;
238 	unsigned int dst_dw, dst_bl, dst_bs, dst_wc, dst_am;
239 	unsigned int src_dw, src_bl, src_bs, src_wc, src_am;
240 
241 	switch (cfg->dst_width) {
242 	case 8:
243 		dst_dw = AWIN_DMA_CTL_DATA_WIDTH_8;
244 		break;
245 	case 16:
246 		dst_dw = AWIN_DMA_CTL_DATA_WIDTH_16;
247 		break;
248 	case 32:
249 		dst_dw = AWIN_DMA_CTL_DATA_WIDTH_32;
250 		break;
251 	default:
252 		return (EINVAL);
253 	}
254 	switch (cfg->dst_burst_len) {
255 	case 1:
256 		dst_bl = AWIN_DMA_CTL_BURST_LEN_1;
257 		break;
258 	case 4:
259 		dst_bl = AWIN_DMA_CTL_BURST_LEN_4;
260 		break;
261 	case 8:
262 		dst_bl = AWIN_DMA_CTL_BURST_LEN_8;
263 		break;
264 	default:
265 		return (EINVAL);
266 	}
267 	switch (cfg->src_width) {
268 	case 8:
269 		src_dw = AWIN_DMA_CTL_DATA_WIDTH_8;
270 		break;
271 	case 16:
272 		src_dw = AWIN_DMA_CTL_DATA_WIDTH_16;
273 		break;
274 	case 32:
275 		src_dw = AWIN_DMA_CTL_DATA_WIDTH_32;
276 		break;
277 	default:
278 		return (EINVAL);
279 	}
280 	switch (cfg->src_burst_len) {
281 	case 1:
282 		src_bl = AWIN_DMA_CTL_BURST_LEN_1;
283 		break;
284 	case 4:
285 		src_bl = AWIN_DMA_CTL_BURST_LEN_4;
286 		break;
287 	case 8:
288 		src_bl = AWIN_DMA_CTL_BURST_LEN_8;
289 		break;
290 	default:
291 		return (EINVAL);
292 	}
293 
294 	val = (dst_dw << AWIN_DMA_CTL_DST_DATA_WIDTH_SHIFT) |
295 	      (dst_bl << AWIN_DMA_CTL_DST_BURST_LEN_SHIFT) |
296 	      (cfg->dst_drqtype << AWIN_DMA_CTL_DST_DRQ_TYPE_SHIFT) |
297 	      (src_dw << AWIN_DMA_CTL_SRC_DATA_WIDTH_SHIFT) |
298 	      (src_bl << AWIN_DMA_CTL_SRC_BURST_LEN_SHIFT) |
299 	      (cfg->src_drqtype << AWIN_DMA_CTL_SRC_DRQ_TYPE_SHIFT);
300 
301 	if (ch->ch_type == CH_NDMA) {
302 		if (cfg->dst_noincr)
303 			val |= AWIN_NDMA_CTL_DST_ADDR_NOINCR;
304 		if (cfg->src_noincr)
305 			val |= AWIN_NDMA_CTL_SRC_ADDR_NOINCR;
306 
307 		DMACH_WRITE(ch, AWIN_NDMA_CTL_REG, val);
308 	} else {
309 		dst_am = cfg->dst_noincr ? AWIN_DDMA_CTL_DMA_ADDR_IO :
310 		    AWIN_DDMA_CTL_DMA_ADDR_LINEAR;
311 		src_am = cfg->src_noincr ? AWIN_DDMA_CTL_DMA_ADDR_IO :
312 		    AWIN_DDMA_CTL_DMA_ADDR_LINEAR;
313 
314 		val |= (dst_am << AWIN_DDMA_CTL_DST_ADDR_MODE_SHIFT);
315 		val |= (src_am << AWIN_DDMA_CTL_SRC_ADDR_MODE_SHIFT);
316 
317 		DMACH_WRITE(ch, AWIN_DDMA_CTL_REG, val);
318 
319 		dst_bs = cfg->dst_blksize - 1;
320 		dst_wc = cfg->dst_wait_cyc - 1;
321 		src_bs = cfg->src_blksize - 1;
322 		src_wc = cfg->src_wait_cyc - 1;
323 
324 		DMACH_WRITE(ch, AWIN_DDMA_PARA_REG,
325 		    (dst_bs << AWIN_DDMA_PARA_DST_DATA_BLK_SIZ_SHIFT) |
326 		    (dst_wc << AWIN_DDMA_PARA_DST_WAIT_CYC_SHIFT) |
327 		    (src_bs << AWIN_DDMA_PARA_SRC_DATA_BLK_SIZ_SHIFT) |
328 		    (src_wc << AWIN_DDMA_PARA_SRC_WAIT_CYC_SHIFT));
329 	}
330 
331 	return (0);
332 }
333 
334 static void *
335 a10dmac_alloc(device_t dev, bool dedicated, void (*cb)(void *), void *cbarg)
336 {
337 	struct a10dmac_softc *sc = device_get_softc(dev);
338 	struct a10dmac_channel *ch_list;
339 	struct a10dmac_channel *ch = NULL;
340 	uint32_t irqen;
341 	uint8_t ch_count, index;
342 
343 	if (dedicated) {
344 		ch_list = sc->sc_ddma_channels;
345 		ch_count = DDMA_CHANNELS;
346 	} else {
347 		ch_list = sc->sc_ndma_channels;
348 		ch_count = NDMA_CHANNELS;
349 	}
350 
351 	mtx_lock_spin(&sc->sc_mtx);
352 	for (index = 0; index < ch_count; index++) {
353 		if (ch_list[index].ch_callback == NULL) {
354 			ch = &ch_list[index];
355 			ch->ch_callback = cb;
356 			ch->ch_callbackarg = cbarg;
357 
358 			irqen = DMA_READ(sc, AWIN_DMA_IRQ_EN_REG);
359 			if (ch->ch_type == CH_NDMA)
360 				irqen |= AWIN_DMA_IRQ_NDMA_END(index);
361 			else
362 				irqen |= AWIN_DMA_IRQ_DDMA_END(index);
363 			DMA_WRITE(sc, AWIN_DMA_IRQ_EN_REG, irqen);
364 
365 			break;
366 		}
367 	}
368 	mtx_unlock_spin(&sc->sc_mtx);
369 
370 	return (ch);
371 }
372 
373 static void
374 a10dmac_free(device_t dev, void *priv)
375 {
376 	struct a10dmac_channel *ch = priv;
377 	struct a10dmac_softc *sc = ch->ch_sc;
378 	uint32_t irqen, sta, cfg;
379 
380 	mtx_lock_spin(&sc->sc_mtx);
381 
382 	irqen = DMA_READ(sc, AWIN_DMA_IRQ_EN_REG);
383 	cfg = a10dmac_read_ctl(ch);
384 	if (ch->ch_type == CH_NDMA) {
385 		sta = AWIN_DMA_IRQ_NDMA_END(ch->ch_index);
386 		cfg &= ~AWIN_NDMA_CTL_DMA_LOADING;
387 	} else {
388 		sta = AWIN_DMA_IRQ_DDMA_END(ch->ch_index);
389 		cfg &= ~AWIN_DDMA_CTL_DMA_LOADING;
390 	}
391 	irqen &= ~sta;
392 	a10dmac_write_ctl(ch, cfg);
393 	DMA_WRITE(sc, AWIN_DMA_IRQ_EN_REG, irqen);
394 	DMA_WRITE(sc, AWIN_DMA_IRQ_PEND_STA_REG, sta);
395 
396 	ch->ch_callback = NULL;
397 	ch->ch_callbackarg = NULL;
398 
399 	mtx_unlock_spin(&sc->sc_mtx);
400 }
401 
402 static int
403 a10dmac_transfer(device_t dev, void *priv, bus_addr_t src, bus_addr_t dst,
404     size_t nbytes)
405 {
406 	struct a10dmac_channel *ch = priv;
407 	uint32_t cfg;
408 
409 	cfg = a10dmac_read_ctl(ch);
410 	if (ch->ch_type == CH_NDMA) {
411 		if (cfg & AWIN_NDMA_CTL_DMA_LOADING)
412 			return (EBUSY);
413 
414 		DMACH_WRITE(ch, AWIN_NDMA_SRC_ADDR_REG, src);
415 		DMACH_WRITE(ch, AWIN_NDMA_DEST_ADDR_REG, dst);
416 		DMACH_WRITE(ch, AWIN_NDMA_BC_REG, nbytes);
417 
418 		cfg |= AWIN_NDMA_CTL_DMA_LOADING;
419 		a10dmac_write_ctl(ch, cfg);
420 	} else {
421 		if (cfg & AWIN_DDMA_CTL_DMA_LOADING)
422 			return (EBUSY);
423 
424 		DMACH_WRITE(ch, AWIN_DDMA_SRC_START_ADDR_REG, src);
425 		DMACH_WRITE(ch, AWIN_DDMA_DEST_START_ADDR_REG, dst);
426 		DMACH_WRITE(ch, AWIN_DDMA_BC_REG, nbytes);
427 
428 		cfg |= AWIN_DDMA_CTL_DMA_LOADING;
429 		a10dmac_write_ctl(ch, cfg);
430 	}
431 
432 	return (0);
433 }
434 
435 static void
436 a10dmac_halt(device_t dev, void *priv)
437 {
438 	struct a10dmac_channel *ch = priv;
439 	uint32_t cfg;
440 
441 	cfg = a10dmac_read_ctl(ch);
442 	if (ch->ch_type == CH_NDMA) {
443 		cfg &= ~AWIN_NDMA_CTL_DMA_LOADING;
444 	} else {
445 		cfg &= ~AWIN_DDMA_CTL_DMA_LOADING;
446 	}
447 	a10dmac_write_ctl(ch, cfg);
448 }
449 
450 static device_method_t a10dmac_methods[] = {
451 	/* Device interface */
452 	DEVMETHOD(device_probe,		a10dmac_probe),
453 	DEVMETHOD(device_attach,	a10dmac_attach),
454 
455 	/* sunxi DMA interface */
456 	DEVMETHOD(sunxi_dma_alloc,	a10dmac_alloc),
457 	DEVMETHOD(sunxi_dma_free,	a10dmac_free),
458 	DEVMETHOD(sunxi_dma_set_config,	a10dmac_set_config),
459 	DEVMETHOD(sunxi_dma_transfer,	a10dmac_transfer),
460 	DEVMETHOD(sunxi_dma_halt,	a10dmac_halt),
461 
462 	DEVMETHOD_END
463 };
464 
465 static driver_t a10dmac_driver = {
466 	"a10dmac",
467 	a10dmac_methods,
468 	sizeof(struct a10dmac_softc)
469 };
470 
471 static devclass_t a10dmac_devclass;
472 
473 DRIVER_MODULE(a10dmac, simplebus, a10dmac_driver, a10dmac_devclass, 0, 0);
474