xref: /netbsd-src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/falcon/nouveau_nvkm_falcon_v1.c (revision 41ec02673d281bbb3d38e6c78504ce6e30c228c1)
1 /*	$NetBSD: nouveau_nvkm_falcon_v1.c,v 1.2 2021/12/18 23:45:38 riastradh Exp $	*/
2 
3 /*
4  * Copyright (c) 2016, NVIDIA CORPORATION. All rights reserved.
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a
7  * copy of this software and associated documentation files (the "Software"),
8  * to deal in the Software without restriction, including without limitation
9  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10  * and/or sell copies of the Software, and to permit persons to whom the
11  * Software is furnished to do so, subject to the following conditions:
12  *
13  * The above copyright notice and this permission notice shall be included in
14  * all copies or substantial portions of the Software.
15  *
16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
19  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
21  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
22  * DEALINGS IN THE SOFTWARE.
23  */
24 #include <sys/cdefs.h>
25 __KERNEL_RCSID(0, "$NetBSD: nouveau_nvkm_falcon_v1.c,v 1.2 2021/12/18 23:45:38 riastradh Exp $");
26 
27 #include "priv.h"
28 
29 #include <core/gpuobj.h>
30 #include <core/memory.h>
31 #include <subdev/timer.h>
32 
33 void
nvkm_falcon_v1_load_imem(struct nvkm_falcon * falcon,void * data,u32 start,u32 size,u16 tag,u8 port,bool secure)34 nvkm_falcon_v1_load_imem(struct nvkm_falcon *falcon, void *data, u32 start,
35 			 u32 size, u16 tag, u8 port, bool secure)
36 {
37 	u8 rem = size % 4;
38 	u32 reg;
39 	int i;
40 
41 	size -= rem;
42 
43 	reg = start | BIT(24) | (secure ? BIT(28) : 0);
44 	nvkm_falcon_wr32(falcon, 0x180 + (port * 16), reg);
45 	for (i = 0; i < size / 4; i++) {
46 		/* write new tag every 256B */
47 		if ((i & 0x3f) == 0)
48 			nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
49 		nvkm_falcon_wr32(falcon, 0x184 + (port * 16), ((u32 *)data)[i]);
50 	}
51 
52 	/*
53 	 * If size is not a multiple of 4, mask the last work to ensure garbage
54 	 * does not get written
55 	 */
56 	if (rem) {
57 		u32 extra = ((u32 *)data)[i];
58 
59 		/* write new tag every 256B */
60 		if ((i & 0x3f) == 0)
61 			nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
62 		nvkm_falcon_wr32(falcon, 0x184 + (port * 16),
63 				 extra & (BIT(rem * 8) - 1));
64 		++i;
65 	}
66 
67 	/* code must be padded to 0x40 words */
68 	for (; i & 0x3f; i++)
69 		nvkm_falcon_wr32(falcon, 0x184 + (port * 16), 0);
70 }
71 
72 static void
nvkm_falcon_v1_load_emem(struct nvkm_falcon * falcon,void * data,u32 start,u32 size,u8 port)73 nvkm_falcon_v1_load_emem(struct nvkm_falcon *falcon, void *data, u32 start,
74 			 u32 size, u8 port)
75 {
76 	u8 rem = size % 4;
77 	int i;
78 
79 	size -= rem;
80 
81 	nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 24));
82 	for (i = 0; i < size / 4; i++)
83 		nvkm_falcon_wr32(falcon, 0xac4 + (port * 8), ((u32 *)data)[i]);
84 
85 	/*
86 	 * If size is not a multiple of 4, mask the last word to ensure garbage
87 	 * does not get written
88 	 */
89 	if (rem) {
90 		u32 extra = ((u32 *)data)[i];
91 
92 		nvkm_falcon_wr32(falcon, 0xac4 + (port * 8),
93 				 extra & (BIT(rem * 8) - 1));
94 	}
95 }
96 
97 void
nvkm_falcon_v1_load_dmem(struct nvkm_falcon * falcon,void * data,u32 start,u32 size,u8 port)98 nvkm_falcon_v1_load_dmem(struct nvkm_falcon *falcon, void *data, u32 start,
99 			 u32 size, u8 port)
100 {
101 	const struct nvkm_falcon_func *func = falcon->func;
102 	u8 rem = size % 4;
103 	int i;
104 
105 	if (func->emem_addr && start >= func->emem_addr)
106 		return nvkm_falcon_v1_load_emem(falcon, data,
107 						start - func->emem_addr, size,
108 						port);
109 
110 	size -= rem;
111 
112 	nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 24));
113 	for (i = 0; i < size / 4; i++)
114 		nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8), ((u32 *)data)[i]);
115 
116 	/*
117 	 * If size is not a multiple of 4, mask the last word to ensure garbage
118 	 * does not get written
119 	 */
120 	if (rem) {
121 		u32 extra = ((u32 *)data)[i];
122 
123 		nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8),
124 				 extra & (BIT(rem * 8) - 1));
125 	}
126 }
127 
128 static void
nvkm_falcon_v1_read_emem(struct nvkm_falcon * falcon,u32 start,u32 size,u8 port,void * data)129 nvkm_falcon_v1_read_emem(struct nvkm_falcon *falcon, u32 start, u32 size,
130 			 u8 port, void *data)
131 {
132 	u8 rem = size % 4;
133 	int i;
134 
135 	size -= rem;
136 
137 	nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 25));
138 	for (i = 0; i < size / 4; i++)
139 		((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
140 
141 	/*
142 	 * If size is not a multiple of 4, mask the last word to ensure garbage
143 	 * does not get read
144 	 */
145 	if (rem) {
146 		u32 extra = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
147 
148 		for (i = size; i < size + rem; i++) {
149 			((u8 *)data)[i] = (u8)(extra & 0xff);
150 			extra >>= 8;
151 		}
152 	}
153 }
154 
155 void
nvkm_falcon_v1_read_dmem(struct nvkm_falcon * falcon,u32 start,u32 size,u8 port,void * data)156 nvkm_falcon_v1_read_dmem(struct nvkm_falcon *falcon, u32 start, u32 size,
157 			 u8 port, void *data)
158 {
159 	const struct nvkm_falcon_func *func = falcon->func;
160 	u8 rem = size % 4;
161 	int i;
162 
163 	if (func->emem_addr && start >= func->emem_addr)
164 		return nvkm_falcon_v1_read_emem(falcon, start - func->emem_addr,
165 						size, port, data);
166 
167 	size -= rem;
168 
169 	nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 25));
170 	for (i = 0; i < size / 4; i++)
171 		((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
172 
173 	/*
174 	 * If size is not a multiple of 4, mask the last word to ensure garbage
175 	 * does not get read
176 	 */
177 	if (rem) {
178 		u32 extra = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
179 
180 		for (i = size; i < size + rem; i++) {
181 			((u8 *)data)[i] = (u8)(extra & 0xff);
182 			extra >>= 8;
183 		}
184 	}
185 }
186 
187 void
nvkm_falcon_v1_bind_context(struct nvkm_falcon * falcon,struct nvkm_memory * ctx)188 nvkm_falcon_v1_bind_context(struct nvkm_falcon *falcon, struct nvkm_memory *ctx)
189 {
190 	const u32 fbif = falcon->func->fbif;
191 	u32 inst_loc;
192 
193 	/* disable instance block binding */
194 	if (ctx == NULL) {
195 		nvkm_falcon_wr32(falcon, 0x10c, 0x0);
196 		return;
197 	}
198 
199 	nvkm_falcon_wr32(falcon, 0x10c, 0x1);
200 
201 	/* setup apertures - virtual */
202 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_UCODE, 0x4);
203 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_VIRT, 0x0);
204 	/* setup apertures - physical */
205 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_VID, 0x4);
206 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_COH, 0x5);
207 	nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_NCOH, 0x6);
208 
209 	/* Set context */
210 	switch (nvkm_memory_target(ctx)) {
211 	case NVKM_MEM_TARGET_VRAM: inst_loc = 0; break;
212 	case NVKM_MEM_TARGET_HOST: inst_loc = 2; break;
213 	case NVKM_MEM_TARGET_NCOH: inst_loc = 3; break;
214 	default:
215 		WARN_ON(1);
216 		return;
217 	}
218 
219 	/* Enable context */
220 	nvkm_falcon_mask(falcon, 0x048, 0x1, 0x1);
221 	nvkm_falcon_wr32(falcon, 0x054,
222 			 ((nvkm_memory_addr(ctx) >> 12) & 0xfffffff) |
223 			 (inst_loc << 28) | (1 << 30));
224 
225 	nvkm_falcon_mask(falcon, 0x090, 0x10000, 0x10000);
226 	nvkm_falcon_mask(falcon, 0x0a4, 0x8, 0x8);
227 }
228 
229 void
nvkm_falcon_v1_set_start_addr(struct nvkm_falcon * falcon,u32 start_addr)230 nvkm_falcon_v1_set_start_addr(struct nvkm_falcon *falcon, u32 start_addr)
231 {
232 	nvkm_falcon_wr32(falcon, 0x104, start_addr);
233 }
234 
235 void
nvkm_falcon_v1_start(struct nvkm_falcon * falcon)236 nvkm_falcon_v1_start(struct nvkm_falcon *falcon)
237 {
238 	u32 reg = nvkm_falcon_rd32(falcon, 0x100);
239 
240 	if (reg & BIT(6))
241 		nvkm_falcon_wr32(falcon, 0x130, 0x2);
242 	else
243 		nvkm_falcon_wr32(falcon, 0x100, 0x2);
244 }
245 
246 int
nvkm_falcon_v1_wait_for_halt(struct nvkm_falcon * falcon,u32 ms)247 nvkm_falcon_v1_wait_for_halt(struct nvkm_falcon *falcon, u32 ms)
248 {
249 	struct nvkm_device *device = falcon->owner->device;
250 	int ret;
251 
252 	ret = nvkm_wait_msec(device, ms, falcon->addr + 0x100, 0x10, 0x10);
253 	if (ret < 0)
254 		return ret;
255 
256 	return 0;
257 }
258 
259 int
nvkm_falcon_v1_clear_interrupt(struct nvkm_falcon * falcon,u32 mask)260 nvkm_falcon_v1_clear_interrupt(struct nvkm_falcon *falcon, u32 mask)
261 {
262 	struct nvkm_device *device = falcon->owner->device;
263 	int ret;
264 
265 	/* clear interrupt(s) */
266 	nvkm_falcon_mask(falcon, 0x004, mask, mask);
267 	/* wait until interrupts are cleared */
268 	ret = nvkm_wait_msec(device, 10, falcon->addr + 0x008, mask, 0x0);
269 	if (ret < 0)
270 		return ret;
271 
272 	return 0;
273 }
274 
275 static int
falcon_v1_wait_idle(struct nvkm_falcon * falcon)276 falcon_v1_wait_idle(struct nvkm_falcon *falcon)
277 {
278 	struct nvkm_device *device = falcon->owner->device;
279 	int ret;
280 
281 	ret = nvkm_wait_msec(device, 10, falcon->addr + 0x04c, 0xffff, 0x0);
282 	if (ret < 0)
283 		return ret;
284 
285 	return 0;
286 }
287 
288 int
nvkm_falcon_v1_enable(struct nvkm_falcon * falcon)289 nvkm_falcon_v1_enable(struct nvkm_falcon *falcon)
290 {
291 	struct nvkm_device *device = falcon->owner->device;
292 	int ret;
293 
294 	ret = nvkm_wait_msec(device, 10, falcon->addr + 0x10c, 0x6, 0x0);
295 	if (ret < 0) {
296 		nvkm_error(falcon->user, "Falcon mem scrubbing timeout\n");
297 		return ret;
298 	}
299 
300 	ret = falcon_v1_wait_idle(falcon);
301 	if (ret)
302 		return ret;
303 
304 	/* enable IRQs */
305 	nvkm_falcon_wr32(falcon, 0x010, 0xff);
306 
307 	return 0;
308 }
309 
310 void
nvkm_falcon_v1_disable(struct nvkm_falcon * falcon)311 nvkm_falcon_v1_disable(struct nvkm_falcon *falcon)
312 {
313 	/* disable IRQs and wait for any previous code to complete */
314 	nvkm_falcon_wr32(falcon, 0x014, 0xff);
315 	falcon_v1_wait_idle(falcon);
316 }
317 
318 static const struct nvkm_falcon_func
319 nvkm_falcon_v1 = {
320 	.load_imem = nvkm_falcon_v1_load_imem,
321 	.load_dmem = nvkm_falcon_v1_load_dmem,
322 	.read_dmem = nvkm_falcon_v1_read_dmem,
323 	.bind_context = nvkm_falcon_v1_bind_context,
324 	.start = nvkm_falcon_v1_start,
325 	.wait_for_halt = nvkm_falcon_v1_wait_for_halt,
326 	.clear_interrupt = nvkm_falcon_v1_clear_interrupt,
327 	.enable = nvkm_falcon_v1_enable,
328 	.disable = nvkm_falcon_v1_disable,
329 	.set_start_addr = nvkm_falcon_v1_set_start_addr,
330 };
331 
332 int
nvkm_falcon_v1_new(struct nvkm_subdev * owner,const char * name,u32 addr,struct nvkm_falcon ** pfalcon)333 nvkm_falcon_v1_new(struct nvkm_subdev *owner, const char *name, u32 addr,
334 		   struct nvkm_falcon **pfalcon)
335 {
336 	struct nvkm_falcon *falcon;
337 	if (!(falcon = *pfalcon = kzalloc(sizeof(*falcon), GFP_KERNEL)))
338 		return -ENOMEM;
339 	nvkm_falcon_ctor(&nvkm_falcon_v1, owner, name, addr, falcon);
340 	return 0;
341 }
342