xref: /openbsd-src/sys/dev/pci/drm/drm_exec.c (revision f005ef32267c16bdb134f0e9fa4477dbe07c263a)
1*f005ef32Sjsg // SPDX-License-Identifier: GPL-2.0 OR MIT
2*f005ef32Sjsg 
3*f005ef32Sjsg #include <drm/drm_exec.h>
4*f005ef32Sjsg #include <drm/drm_gem.h>
5*f005ef32Sjsg #include <linux/dma-resv.h>
6*f005ef32Sjsg 
7*f005ef32Sjsg /**
8*f005ef32Sjsg  * DOC: Overview
9*f005ef32Sjsg  *
10*f005ef32Sjsg  * This component mainly abstracts the retry loop necessary for locking
11*f005ef32Sjsg  * multiple GEM objects while preparing hardware operations (e.g. command
12*f005ef32Sjsg  * submissions, page table updates etc..).
13*f005ef32Sjsg  *
14*f005ef32Sjsg  * If a contention is detected while locking a GEM object the cleanup procedure
15*f005ef32Sjsg  * unlocks all previously locked GEM objects and locks the contended one first
16*f005ef32Sjsg  * before locking any further objects.
17*f005ef32Sjsg  *
18*f005ef32Sjsg  * After an object is locked fences slots can optionally be reserved on the
19*f005ef32Sjsg  * dma_resv object inside the GEM object.
20*f005ef32Sjsg  *
21*f005ef32Sjsg  * A typical usage pattern should look like this::
22*f005ef32Sjsg  *
23*f005ef32Sjsg  *	struct drm_gem_object *obj;
24*f005ef32Sjsg  *	struct drm_exec exec;
25*f005ef32Sjsg  *	unsigned long index;
26*f005ef32Sjsg  *	int ret;
27*f005ef32Sjsg  *
28*f005ef32Sjsg  *	drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT);
29*f005ef32Sjsg  *	drm_exec_until_all_locked(&exec) {
30*f005ef32Sjsg  *		ret = drm_exec_prepare_obj(&exec, boA, 1);
31*f005ef32Sjsg  *		drm_exec_retry_on_contention(&exec);
32*f005ef32Sjsg  *		if (ret)
33*f005ef32Sjsg  *			goto error;
34*f005ef32Sjsg  *
35*f005ef32Sjsg  *		ret = drm_exec_prepare_obj(&exec, boB, 1);
36*f005ef32Sjsg  *		drm_exec_retry_on_contention(&exec);
37*f005ef32Sjsg  *		if (ret)
38*f005ef32Sjsg  *			goto error;
39*f005ef32Sjsg  *	}
40*f005ef32Sjsg  *
41*f005ef32Sjsg  *	drm_exec_for_each_locked_object(&exec, index, obj) {
42*f005ef32Sjsg  *		dma_resv_add_fence(obj->resv, fence, DMA_RESV_USAGE_READ);
43*f005ef32Sjsg  *		...
44*f005ef32Sjsg  *	}
45*f005ef32Sjsg  *	drm_exec_fini(&exec);
46*f005ef32Sjsg  *
47*f005ef32Sjsg  * See struct dma_exec for more details.
48*f005ef32Sjsg  */
49*f005ef32Sjsg 
50*f005ef32Sjsg /* Dummy value used to initially enter the retry loop */
51*f005ef32Sjsg #define DRM_EXEC_DUMMY ((void *)~0)
52*f005ef32Sjsg 
53*f005ef32Sjsg /* Unlock all objects and drop references */
drm_exec_unlock_all(struct drm_exec * exec)54*f005ef32Sjsg static void drm_exec_unlock_all(struct drm_exec *exec)
55*f005ef32Sjsg {
56*f005ef32Sjsg 	struct drm_gem_object *obj;
57*f005ef32Sjsg 	unsigned long index;
58*f005ef32Sjsg 
59*f005ef32Sjsg 	drm_exec_for_each_locked_object_reverse(exec, index, obj) {
60*f005ef32Sjsg 		dma_resv_unlock(obj->resv);
61*f005ef32Sjsg 		drm_gem_object_put(obj);
62*f005ef32Sjsg 	}
63*f005ef32Sjsg 
64*f005ef32Sjsg 	drm_gem_object_put(exec->prelocked);
65*f005ef32Sjsg 	exec->prelocked = NULL;
66*f005ef32Sjsg }
67*f005ef32Sjsg 
68*f005ef32Sjsg /**
69*f005ef32Sjsg  * drm_exec_init - initialize a drm_exec object
70*f005ef32Sjsg  * @exec: the drm_exec object to initialize
71*f005ef32Sjsg  * @flags: controls locking behavior, see DRM_EXEC_* defines
72*f005ef32Sjsg  *
73*f005ef32Sjsg  * Initialize the object and make sure that we can track locked objects.
74*f005ef32Sjsg  */
drm_exec_init(struct drm_exec * exec,uint32_t flags)75*f005ef32Sjsg void drm_exec_init(struct drm_exec *exec, uint32_t flags)
76*f005ef32Sjsg {
77*f005ef32Sjsg 	exec->flags = flags;
78*f005ef32Sjsg 	exec->objects = kmalloc(PAGE_SIZE, GFP_KERNEL);
79*f005ef32Sjsg 
80*f005ef32Sjsg 	/* If allocation here fails, just delay that till the first use */
81*f005ef32Sjsg 	exec->max_objects = exec->objects ? PAGE_SIZE / sizeof(void *) : 0;
82*f005ef32Sjsg 	exec->num_objects = 0;
83*f005ef32Sjsg 	exec->contended = DRM_EXEC_DUMMY;
84*f005ef32Sjsg 	exec->prelocked = NULL;
85*f005ef32Sjsg }
86*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_init);
87*f005ef32Sjsg 
88*f005ef32Sjsg /**
89*f005ef32Sjsg  * drm_exec_fini - finalize a drm_exec object
90*f005ef32Sjsg  * @exec: the drm_exec object to finalize
91*f005ef32Sjsg  *
92*f005ef32Sjsg  * Unlock all locked objects, drop the references to objects and free all memory
93*f005ef32Sjsg  * used for tracking the state.
94*f005ef32Sjsg  */
drm_exec_fini(struct drm_exec * exec)95*f005ef32Sjsg void drm_exec_fini(struct drm_exec *exec)
96*f005ef32Sjsg {
97*f005ef32Sjsg 	drm_exec_unlock_all(exec);
98*f005ef32Sjsg 	kvfree(exec->objects);
99*f005ef32Sjsg 	if (exec->contended != DRM_EXEC_DUMMY) {
100*f005ef32Sjsg 		drm_gem_object_put(exec->contended);
101*f005ef32Sjsg 		ww_acquire_fini(&exec->ticket);
102*f005ef32Sjsg 	}
103*f005ef32Sjsg }
104*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_fini);
105*f005ef32Sjsg 
106*f005ef32Sjsg /**
107*f005ef32Sjsg  * drm_exec_cleanup - cleanup when contention is detected
108*f005ef32Sjsg  * @exec: the drm_exec object to cleanup
109*f005ef32Sjsg  *
110*f005ef32Sjsg  * Cleanup the current state and return true if we should stay inside the retry
111*f005ef32Sjsg  * loop, false if there wasn't any contention detected and we can keep the
112*f005ef32Sjsg  * objects locked.
113*f005ef32Sjsg  */
drm_exec_cleanup(struct drm_exec * exec)114*f005ef32Sjsg bool drm_exec_cleanup(struct drm_exec *exec)
115*f005ef32Sjsg {
116*f005ef32Sjsg 	if (likely(!exec->contended)) {
117*f005ef32Sjsg 		ww_acquire_done(&exec->ticket);
118*f005ef32Sjsg 		return false;
119*f005ef32Sjsg 	}
120*f005ef32Sjsg 
121*f005ef32Sjsg 	if (likely(exec->contended == DRM_EXEC_DUMMY)) {
122*f005ef32Sjsg 		exec->contended = NULL;
123*f005ef32Sjsg 		ww_acquire_init(&exec->ticket, &reservation_ww_class);
124*f005ef32Sjsg 		return true;
125*f005ef32Sjsg 	}
126*f005ef32Sjsg 
127*f005ef32Sjsg 	drm_exec_unlock_all(exec);
128*f005ef32Sjsg 	exec->num_objects = 0;
129*f005ef32Sjsg 	return true;
130*f005ef32Sjsg }
131*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_cleanup);
132*f005ef32Sjsg 
133*f005ef32Sjsg /* Track the locked object in the array */
drm_exec_obj_locked(struct drm_exec * exec,struct drm_gem_object * obj)134*f005ef32Sjsg static int drm_exec_obj_locked(struct drm_exec *exec,
135*f005ef32Sjsg 			       struct drm_gem_object *obj)
136*f005ef32Sjsg {
137*f005ef32Sjsg 	if (unlikely(exec->num_objects == exec->max_objects)) {
138*f005ef32Sjsg 		size_t size = exec->max_objects * sizeof(void *);
139*f005ef32Sjsg 		void *tmp;
140*f005ef32Sjsg 
141*f005ef32Sjsg #ifdef __linux__
142*f005ef32Sjsg 		tmp = kvrealloc(exec->objects, size, size + PAGE_SIZE,
143*f005ef32Sjsg 				GFP_KERNEL);
144*f005ef32Sjsg 		if (!tmp)
145*f005ef32Sjsg 			return -ENOMEM;
146*f005ef32Sjsg #else
147*f005ef32Sjsg 		tmp = kvmalloc(size + PAGE_SIZE, GFP_KERNEL);
148*f005ef32Sjsg 		if (!tmp)
149*f005ef32Sjsg 			return -ENOMEM;
150*f005ef32Sjsg 		memcpy(tmp, exec->objects, size);
151*f005ef32Sjsg 		kvfree(exec->objects);
152*f005ef32Sjsg #endif
153*f005ef32Sjsg 
154*f005ef32Sjsg 		exec->objects = tmp;
155*f005ef32Sjsg 		exec->max_objects += PAGE_SIZE / sizeof(void *);
156*f005ef32Sjsg 	}
157*f005ef32Sjsg 	drm_gem_object_get(obj);
158*f005ef32Sjsg 	exec->objects[exec->num_objects++] = obj;
159*f005ef32Sjsg 
160*f005ef32Sjsg 	return 0;
161*f005ef32Sjsg }
162*f005ef32Sjsg 
163*f005ef32Sjsg /* Make sure the contended object is locked first */
drm_exec_lock_contended(struct drm_exec * exec)164*f005ef32Sjsg static int drm_exec_lock_contended(struct drm_exec *exec)
165*f005ef32Sjsg {
166*f005ef32Sjsg 	struct drm_gem_object *obj = exec->contended;
167*f005ef32Sjsg 	int ret;
168*f005ef32Sjsg 
169*f005ef32Sjsg 	if (likely(!obj))
170*f005ef32Sjsg 		return 0;
171*f005ef32Sjsg 
172*f005ef32Sjsg 	/* Always cleanup the contention so that error handling can kick in */
173*f005ef32Sjsg 	exec->contended = NULL;
174*f005ef32Sjsg 	if (exec->flags & DRM_EXEC_INTERRUPTIBLE_WAIT) {
175*f005ef32Sjsg 		ret = dma_resv_lock_slow_interruptible(obj->resv,
176*f005ef32Sjsg 						       &exec->ticket);
177*f005ef32Sjsg 		if (unlikely(ret))
178*f005ef32Sjsg 			goto error_dropref;
179*f005ef32Sjsg 	} else {
180*f005ef32Sjsg 		dma_resv_lock_slow(obj->resv, &exec->ticket);
181*f005ef32Sjsg 	}
182*f005ef32Sjsg 
183*f005ef32Sjsg 	ret = drm_exec_obj_locked(exec, obj);
184*f005ef32Sjsg 	if (unlikely(ret))
185*f005ef32Sjsg 		goto error_unlock;
186*f005ef32Sjsg 
187*f005ef32Sjsg 	exec->prelocked = obj;
188*f005ef32Sjsg 	return 0;
189*f005ef32Sjsg 
190*f005ef32Sjsg error_unlock:
191*f005ef32Sjsg 	dma_resv_unlock(obj->resv);
192*f005ef32Sjsg 
193*f005ef32Sjsg error_dropref:
194*f005ef32Sjsg 	drm_gem_object_put(obj);
195*f005ef32Sjsg 	return ret;
196*f005ef32Sjsg }
197*f005ef32Sjsg 
198*f005ef32Sjsg /**
199*f005ef32Sjsg  * drm_exec_lock_obj - lock a GEM object for use
200*f005ef32Sjsg  * @exec: the drm_exec object with the state
201*f005ef32Sjsg  * @obj: the GEM object to lock
202*f005ef32Sjsg  *
203*f005ef32Sjsg  * Lock a GEM object for use and grab a reference to it.
204*f005ef32Sjsg  *
205*f005ef32Sjsg  * Returns: -EDEADLK if a contention is detected, -EALREADY when object is
206*f005ef32Sjsg  * already locked (can be suppressed by setting the DRM_EXEC_IGNORE_DUPLICATES
207*f005ef32Sjsg  * flag), -ENOMEM when memory allocation failed and zero for success.
208*f005ef32Sjsg  */
drm_exec_lock_obj(struct drm_exec * exec,struct drm_gem_object * obj)209*f005ef32Sjsg int drm_exec_lock_obj(struct drm_exec *exec, struct drm_gem_object *obj)
210*f005ef32Sjsg {
211*f005ef32Sjsg 	int ret;
212*f005ef32Sjsg 
213*f005ef32Sjsg 	ret = drm_exec_lock_contended(exec);
214*f005ef32Sjsg 	if (unlikely(ret))
215*f005ef32Sjsg 		return ret;
216*f005ef32Sjsg 
217*f005ef32Sjsg 	if (exec->prelocked == obj) {
218*f005ef32Sjsg 		drm_gem_object_put(exec->prelocked);
219*f005ef32Sjsg 		exec->prelocked = NULL;
220*f005ef32Sjsg 		return 0;
221*f005ef32Sjsg 	}
222*f005ef32Sjsg 
223*f005ef32Sjsg 	if (exec->flags & DRM_EXEC_INTERRUPTIBLE_WAIT)
224*f005ef32Sjsg 		ret = dma_resv_lock_interruptible(obj->resv, &exec->ticket);
225*f005ef32Sjsg 	else
226*f005ef32Sjsg 		ret = dma_resv_lock(obj->resv, &exec->ticket);
227*f005ef32Sjsg 
228*f005ef32Sjsg 	if (unlikely(ret == -EDEADLK)) {
229*f005ef32Sjsg 		drm_gem_object_get(obj);
230*f005ef32Sjsg 		exec->contended = obj;
231*f005ef32Sjsg 		return -EDEADLK;
232*f005ef32Sjsg 	}
233*f005ef32Sjsg 
234*f005ef32Sjsg 	if (unlikely(ret == -EALREADY) &&
235*f005ef32Sjsg 	    exec->flags & DRM_EXEC_IGNORE_DUPLICATES)
236*f005ef32Sjsg 		return 0;
237*f005ef32Sjsg 
238*f005ef32Sjsg 	if (unlikely(ret))
239*f005ef32Sjsg 		return ret;
240*f005ef32Sjsg 
241*f005ef32Sjsg 	ret = drm_exec_obj_locked(exec, obj);
242*f005ef32Sjsg 	if (ret)
243*f005ef32Sjsg 		goto error_unlock;
244*f005ef32Sjsg 
245*f005ef32Sjsg 	return 0;
246*f005ef32Sjsg 
247*f005ef32Sjsg error_unlock:
248*f005ef32Sjsg 	dma_resv_unlock(obj->resv);
249*f005ef32Sjsg 	return ret;
250*f005ef32Sjsg }
251*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_lock_obj);
252*f005ef32Sjsg 
253*f005ef32Sjsg /**
254*f005ef32Sjsg  * drm_exec_unlock_obj - unlock a GEM object in this exec context
255*f005ef32Sjsg  * @exec: the drm_exec object with the state
256*f005ef32Sjsg  * @obj: the GEM object to unlock
257*f005ef32Sjsg  *
258*f005ef32Sjsg  * Unlock the GEM object and remove it from the collection of locked objects.
259*f005ef32Sjsg  * Should only be used to unlock the most recently locked objects. It's not time
260*f005ef32Sjsg  * efficient to unlock objects locked long ago.
261*f005ef32Sjsg  */
drm_exec_unlock_obj(struct drm_exec * exec,struct drm_gem_object * obj)262*f005ef32Sjsg void drm_exec_unlock_obj(struct drm_exec *exec, struct drm_gem_object *obj)
263*f005ef32Sjsg {
264*f005ef32Sjsg 	unsigned int i;
265*f005ef32Sjsg 
266*f005ef32Sjsg 	for (i = exec->num_objects; i--;) {
267*f005ef32Sjsg 		if (exec->objects[i] == obj) {
268*f005ef32Sjsg 			dma_resv_unlock(obj->resv);
269*f005ef32Sjsg 			for (++i; i < exec->num_objects; ++i)
270*f005ef32Sjsg 				exec->objects[i - 1] = exec->objects[i];
271*f005ef32Sjsg 			--exec->num_objects;
272*f005ef32Sjsg 			drm_gem_object_put(obj);
273*f005ef32Sjsg 			return;
274*f005ef32Sjsg 		}
275*f005ef32Sjsg 
276*f005ef32Sjsg 	}
277*f005ef32Sjsg }
278*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_unlock_obj);
279*f005ef32Sjsg 
280*f005ef32Sjsg /**
281*f005ef32Sjsg  * drm_exec_prepare_obj - prepare a GEM object for use
282*f005ef32Sjsg  * @exec: the drm_exec object with the state
283*f005ef32Sjsg  * @obj: the GEM object to prepare
284*f005ef32Sjsg  * @num_fences: how many fences to reserve
285*f005ef32Sjsg  *
286*f005ef32Sjsg  * Prepare a GEM object for use by locking it and reserving fence slots.
287*f005ef32Sjsg  *
288*f005ef32Sjsg  * Returns: -EDEADLK if a contention is detected, -EALREADY when object is
289*f005ef32Sjsg  * already locked, -ENOMEM when memory allocation failed and zero for success.
290*f005ef32Sjsg  */
drm_exec_prepare_obj(struct drm_exec * exec,struct drm_gem_object * obj,unsigned int num_fences)291*f005ef32Sjsg int drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj,
292*f005ef32Sjsg 			 unsigned int num_fences)
293*f005ef32Sjsg {
294*f005ef32Sjsg 	int ret;
295*f005ef32Sjsg 
296*f005ef32Sjsg 	ret = drm_exec_lock_obj(exec, obj);
297*f005ef32Sjsg 	if (ret)
298*f005ef32Sjsg 		return ret;
299*f005ef32Sjsg 
300*f005ef32Sjsg 	ret = dma_resv_reserve_fences(obj->resv, num_fences);
301*f005ef32Sjsg 	if (ret) {
302*f005ef32Sjsg 		drm_exec_unlock_obj(exec, obj);
303*f005ef32Sjsg 		return ret;
304*f005ef32Sjsg 	}
305*f005ef32Sjsg 
306*f005ef32Sjsg 	return 0;
307*f005ef32Sjsg }
308*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_prepare_obj);
309*f005ef32Sjsg 
310*f005ef32Sjsg /**
311*f005ef32Sjsg  * drm_exec_prepare_array - helper to prepare an array of objects
312*f005ef32Sjsg  * @exec: the drm_exec object with the state
313*f005ef32Sjsg  * @objects: array of GEM object to prepare
314*f005ef32Sjsg  * @num_objects: number of GEM objects in the array
315*f005ef32Sjsg  * @num_fences: number of fences to reserve on each GEM object
316*f005ef32Sjsg  *
317*f005ef32Sjsg  * Prepares all GEM objects in an array, aborts on first error.
318*f005ef32Sjsg  * Reserves @num_fences on each GEM object after locking it.
319*f005ef32Sjsg  *
320*f005ef32Sjsg  * Returns: -EDEADLOCK on contention, -EALREADY when object is already locked,
321*f005ef32Sjsg  * -ENOMEM when memory allocation failed and zero for success.
322*f005ef32Sjsg  */
drm_exec_prepare_array(struct drm_exec * exec,struct drm_gem_object ** objects,unsigned int num_objects,unsigned int num_fences)323*f005ef32Sjsg int drm_exec_prepare_array(struct drm_exec *exec,
324*f005ef32Sjsg 			   struct drm_gem_object **objects,
325*f005ef32Sjsg 			   unsigned int num_objects,
326*f005ef32Sjsg 			   unsigned int num_fences)
327*f005ef32Sjsg {
328*f005ef32Sjsg 	int ret;
329*f005ef32Sjsg 
330*f005ef32Sjsg 	for (unsigned int i = 0; i < num_objects; ++i) {
331*f005ef32Sjsg 		ret = drm_exec_prepare_obj(exec, objects[i], num_fences);
332*f005ef32Sjsg 		if (unlikely(ret))
333*f005ef32Sjsg 			return ret;
334*f005ef32Sjsg 	}
335*f005ef32Sjsg 
336*f005ef32Sjsg 	return 0;
337*f005ef32Sjsg }
338*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_prepare_array);
339*f005ef32Sjsg 
340*f005ef32Sjsg MODULE_DESCRIPTION("DRM execution context");
341*f005ef32Sjsg MODULE_LICENSE("Dual MIT/GPL");
342