xref: /linux/drivers/gpu/drm/nouveau/nvkm/falcon/v1.c (revision 52338415)
1 /*
2  * Copyright (c) 2016, NVIDIA CORPORATION. All rights reserved.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
18  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
19  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
20  * DEALINGS IN THE SOFTWARE.
21  */
22 #include "priv.h"
23 
24 #include <core/gpuobj.h>
25 #include <core/memory.h>
26 #include <subdev/timer.h>
27 
28 static void
29 nvkm_falcon_v1_load_imem(struct nvkm_falcon *falcon, void *data, u32 start,
30 			 u32 size, u16 tag, u8 port, bool secure)
31 {
32 	u8 rem = size % 4;
33 	u32 reg;
34 	int i;
35 
36 	size -= rem;
37 
38 	reg = start | BIT(24) | (secure ? BIT(28) : 0);
39 	nvkm_falcon_wr32(falcon, 0x180 + (port * 16), reg);
40 	for (i = 0; i < size / 4; i++) {
41 		/* write new tag every 256B */
42 		if ((i & 0x3f) == 0)
43 			nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
44 		nvkm_falcon_wr32(falcon, 0x184 + (port * 16), ((u32 *)data)[i]);
45 	}
46 
47 	/*
48 	 * If size is not a multiple of 4, mask the last work to ensure garbage
49 	 * does not get written
50 	 */
51 	if (rem) {
52 		u32 extra = ((u32 *)data)[i];
53 
54 		/* write new tag every 256B */
55 		if ((i & 0x3f) == 0)
56 			nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
57 		nvkm_falcon_wr32(falcon, 0x184 + (port * 16),
58 				 extra & (BIT(rem * 8) - 1));
59 		++i;
60 	}
61 
62 	/* code must be padded to 0x40 words */
63 	for (; i & 0x3f; i++)
64 		nvkm_falcon_wr32(falcon, 0x184 + (port * 16), 0);
65 }
66 
67 static void
68 nvkm_falcon_v1_load_emem(struct nvkm_falcon *falcon, void *data, u32 start,
69 			 u32 size, u8 port)
70 {
71 	u8 rem = size % 4;
72 	int i;
73 
74 	size -= rem;
75 
76 	nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 24));
77 	for (i = 0; i < size / 4; i++)
78 		nvkm_falcon_wr32(falcon, 0xac4 + (port * 8), ((u32 *)data)[i]);
79 
80 	/*
81 	 * If size is not a multiple of 4, mask the last word to ensure garbage
82 	 * does not get written
83 	 */
84 	if (rem) {
85 		u32 extra = ((u32 *)data)[i];
86 
87 		nvkm_falcon_wr32(falcon, 0xac4 + (port * 8),
88 				 extra & (BIT(rem * 8) - 1));
89 	}
90 }
91 
92 static const u32 EMEM_START_ADDR = 0x1000000;
93 
94 static void
95 nvkm_falcon_v1_load_dmem(struct nvkm_falcon *falcon, void *data, u32 start,
96 		      u32 size, u8 port)
97 {
98 	u8 rem = size % 4;
99 	int i;
100 
101 	if (start >= EMEM_START_ADDR && falcon->has_emem)
102 		return nvkm_falcon_v1_load_emem(falcon, data,
103 						start - EMEM_START_ADDR, size,
104 						port);
105 
106 	size -= rem;
107 
108 	nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 24));
109 	for (i = 0; i < size / 4; i++)
110 		nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8), ((u32 *)data)[i]);
111 
112 	/*
113 	 * If size is not a multiple of 4, mask the last word to ensure garbage
114 	 * does not get written
115 	 */
116 	if (rem) {
117 		u32 extra = ((u32 *)data)[i];
118 
119 		nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8),
120 				 extra & (BIT(rem * 8) - 1));
121 	}
122 }
123 
124 static void
125 nvkm_falcon_v1_read_emem(struct nvkm_falcon *falcon, u32 start, u32 size,
126 			 u8 port, void *data)
127 {
128 	u8 rem = size % 4;
129 	int i;
130 
131 	size -= rem;
132 
133 	nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 25));
134 	for (i = 0; i < size / 4; i++)
135 		((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
136 
137 	/*
138 	 * If size is not a multiple of 4, mask the last word to ensure garbage
139 	 * does not get read
140 	 */
141 	if (rem) {
142 		u32 extra = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
143 
144 		for (i = size; i < size + rem; i++) {
145 			((u8 *)data)[i] = (u8)(extra & 0xff);
146 			extra >>= 8;
147 		}
148 	}
149 }
150 
151 static void
152 nvkm_falcon_v1_read_dmem(struct nvkm_falcon *falcon, u32 start, u32 size,
153 			 u8 port, void *data)
154 {
155 	u8 rem = size % 4;
156 	int i;
157 
158 	if (start >= EMEM_START_ADDR && falcon->has_emem)
159 		return nvkm_falcon_v1_read_emem(falcon, start - EMEM_START_ADDR,
160 						size, port, data);
161 
162 	size -= rem;
163 
164 	nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 25));
165 	for (i = 0; i < size / 4; i++)
166 		((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
167 
168 	/*
169 	 * If size is not a multiple of 4, mask the last word to ensure garbage
170 	 * does not get read
171 	 */
172 	if (rem) {
173 		u32 extra = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
174 
175 		for (i = size; i < size + rem; i++) {
176 			((u8 *)data)[i] = (u8)(extra & 0xff);
177 			extra >>= 8;
178 		}
179 	}
180 }
181 
182 static void
183 nvkm_falcon_v1_bind_context(struct nvkm_falcon *falcon, struct nvkm_memory *ctx)
184 {
185 	struct nvkm_device *device = falcon->owner->device;
186 	u32 inst_loc;
187 	u32 fbif;
188 
189 	/* disable instance block binding */
190 	if (ctx == NULL) {
191 		nvkm_falcon_wr32(falcon, 0x10c, 0x0);
192 		return;
193 	}
194 
195 	switch (falcon->owner->index) {
196 	case NVKM_ENGINE_NVENC0:
197 	case NVKM_ENGINE_NVENC1:
198 	case NVKM_ENGINE_NVENC2:
199 		fbif = 0x800;
200 		break;
201 	case NVKM_SUBDEV_PMU:
202 		fbif = 0xe00;
203 		break;
204 	default:
205 		fbif = 0x600;
206 		break;
207 	}
208 
209 	nvkm_falcon_wr32(falcon, 0x10c, 0x1);
210 
211 	/* setup apertures - virtual */
212 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_UCODE, 0x4);
213 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_VIRT, 0x0);
214 	/* setup apertures - physical */
215 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_VID, 0x4);
216 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_COH, 0x5);
217 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_NCOH, 0x6);
218 
219 	/* Set context */
220 	switch (nvkm_memory_target(ctx)) {
221 	case NVKM_MEM_TARGET_VRAM: inst_loc = 0; break;
222 	case NVKM_MEM_TARGET_HOST: inst_loc = 2; break;
223 	case NVKM_MEM_TARGET_NCOH: inst_loc = 3; break;
224 	default:
225 		WARN_ON(1);
226 		return;
227 	}
228 
229 	/* Enable context */
230 	nvkm_falcon_mask(falcon, 0x048, 0x1, 0x1);
231 	nvkm_falcon_wr32(falcon, 0x054,
232 			 ((nvkm_memory_addr(ctx) >> 12) & 0xfffffff) |
233 			 (inst_loc << 28) | (1 << 30));
234 
235 	nvkm_falcon_mask(falcon, 0x090, 0x10000, 0x10000);
236 	nvkm_falcon_mask(falcon, 0x0a4, 0x8, 0x8);
237 
238 	/* Not sure if this is a WAR for a HW issue, or some additional
239 	 * programming sequence that's needed to properly complete the
240 	 * context switch we trigger above.
241 	 *
242 	 * Fixes unreliability of booting the SEC2 RTOS on Quadro P620,
243 	 * particularly when resuming from suspend.
244 	 *
245 	 * Also removes the need for an odd workaround where we needed
246 	 * to program SEC2's FALCON_CPUCTL_ALIAS_STARTCPU twice before
247 	 * the SEC2 RTOS would begin executing.
248 	 */
249 	switch (falcon->owner->index) {
250 	case NVKM_SUBDEV_GSP:
251 	case NVKM_ENGINE_SEC2:
252 		nvkm_msec(device, 10,
253 			u32 irqstat = nvkm_falcon_rd32(falcon, 0x008);
254 			u32 flcn0dc = nvkm_falcon_rd32(falcon, 0x0dc);
255 			if ((irqstat & 0x00000008) &&
256 			    (flcn0dc & 0x00007000) == 0x00005000)
257 				break;
258 		);
259 
260 		nvkm_falcon_mask(falcon, 0x004, 0x00000008, 0x00000008);
261 		nvkm_falcon_mask(falcon, 0x058, 0x00000002, 0x00000002);
262 
263 		nvkm_msec(device, 10,
264 			u32 flcn0dc = nvkm_falcon_rd32(falcon, 0x0dc);
265 			if ((flcn0dc & 0x00007000) == 0x00000000)
266 				break;
267 		);
268 		break;
269 	default:
270 		break;
271 	}
272 }
273 
274 static void
275 nvkm_falcon_v1_set_start_addr(struct nvkm_falcon *falcon, u32 start_addr)
276 {
277 	nvkm_falcon_wr32(falcon, 0x104, start_addr);
278 }
279 
280 static void
281 nvkm_falcon_v1_start(struct nvkm_falcon *falcon)
282 {
283 	u32 reg = nvkm_falcon_rd32(falcon, 0x100);
284 
285 	if (reg & BIT(6))
286 		nvkm_falcon_wr32(falcon, 0x130, 0x2);
287 	else
288 		nvkm_falcon_wr32(falcon, 0x100, 0x2);
289 }
290 
291 static int
292 nvkm_falcon_v1_wait_for_halt(struct nvkm_falcon *falcon, u32 ms)
293 {
294 	struct nvkm_device *device = falcon->owner->device;
295 	int ret;
296 
297 	ret = nvkm_wait_msec(device, ms, falcon->addr + 0x100, 0x10, 0x10);
298 	if (ret < 0)
299 		return ret;
300 
301 	return 0;
302 }
303 
304 static int
305 nvkm_falcon_v1_clear_interrupt(struct nvkm_falcon *falcon, u32 mask)
306 {
307 	struct nvkm_device *device = falcon->owner->device;
308 	int ret;
309 
310 	/* clear interrupt(s) */
311 	nvkm_falcon_mask(falcon, 0x004, mask, mask);
312 	/* wait until interrupts are cleared */
313 	ret = nvkm_wait_msec(device, 10, falcon->addr + 0x008, mask, 0x0);
314 	if (ret < 0)
315 		return ret;
316 
317 	return 0;
318 }
319 
320 static int
321 falcon_v1_wait_idle(struct nvkm_falcon *falcon)
322 {
323 	struct nvkm_device *device = falcon->owner->device;
324 	int ret;
325 
326 	ret = nvkm_wait_msec(device, 10, falcon->addr + 0x04c, 0xffff, 0x0);
327 	if (ret < 0)
328 		return ret;
329 
330 	return 0;
331 }
332 
333 static int
334 nvkm_falcon_v1_enable(struct nvkm_falcon *falcon)
335 {
336 	struct nvkm_device *device = falcon->owner->device;
337 	int ret;
338 
339 	ret = nvkm_wait_msec(device, 10, falcon->addr + 0x10c, 0x6, 0x0);
340 	if (ret < 0) {
341 		nvkm_error(falcon->user, "Falcon mem scrubbing timeout\n");
342 		return ret;
343 	}
344 
345 	ret = falcon_v1_wait_idle(falcon);
346 	if (ret)
347 		return ret;
348 
349 	/* enable IRQs */
350 	nvkm_falcon_wr32(falcon, 0x010, 0xff);
351 
352 	return 0;
353 }
354 
355 static void
356 nvkm_falcon_v1_disable(struct nvkm_falcon *falcon)
357 {
358 	/* disable IRQs and wait for any previous code to complete */
359 	nvkm_falcon_wr32(falcon, 0x014, 0xff);
360 	falcon_v1_wait_idle(falcon);
361 }
362 
363 static const struct nvkm_falcon_func
364 nvkm_falcon_v1 = {
365 	.load_imem = nvkm_falcon_v1_load_imem,
366 	.load_dmem = nvkm_falcon_v1_load_dmem,
367 	.read_dmem = nvkm_falcon_v1_read_dmem,
368 	.bind_context = nvkm_falcon_v1_bind_context,
369 	.start = nvkm_falcon_v1_start,
370 	.wait_for_halt = nvkm_falcon_v1_wait_for_halt,
371 	.clear_interrupt = nvkm_falcon_v1_clear_interrupt,
372 	.enable = nvkm_falcon_v1_enable,
373 	.disable = nvkm_falcon_v1_disable,
374 	.set_start_addr = nvkm_falcon_v1_set_start_addr,
375 };
376 
377 int
378 nvkm_falcon_v1_new(struct nvkm_subdev *owner, const char *name, u32 addr,
379 		   struct nvkm_falcon **pfalcon)
380 {
381 	struct nvkm_falcon *falcon;
382 	if (!(falcon = *pfalcon = kzalloc(sizeof(*falcon), GFP_KERNEL)))
383 		return -ENOMEM;
384 	nvkm_falcon_ctor(&nvkm_falcon_v1, owner, name, addr, falcon);
385 	return 0;
386 }
387