1*f005ef32Sjsg // SPDX-License-Identifier: GPL-2.0 OR MIT
2*f005ef32Sjsg
3*f005ef32Sjsg #include <drm/drm_exec.h>
4*f005ef32Sjsg #include <drm/drm_gem.h>
5*f005ef32Sjsg #include <linux/dma-resv.h>
6*f005ef32Sjsg
7*f005ef32Sjsg /**
8*f005ef32Sjsg * DOC: Overview
9*f005ef32Sjsg *
10*f005ef32Sjsg * This component mainly abstracts the retry loop necessary for locking
11*f005ef32Sjsg * multiple GEM objects while preparing hardware operations (e.g. command
12*f005ef32Sjsg * submissions, page table updates etc..).
13*f005ef32Sjsg *
14*f005ef32Sjsg * If a contention is detected while locking a GEM object the cleanup procedure
15*f005ef32Sjsg * unlocks all previously locked GEM objects and locks the contended one first
16*f005ef32Sjsg * before locking any further objects.
17*f005ef32Sjsg *
18*f005ef32Sjsg * After an object is locked fences slots can optionally be reserved on the
19*f005ef32Sjsg * dma_resv object inside the GEM object.
20*f005ef32Sjsg *
21*f005ef32Sjsg * A typical usage pattern should look like this::
22*f005ef32Sjsg *
23*f005ef32Sjsg * struct drm_gem_object *obj;
24*f005ef32Sjsg * struct drm_exec exec;
25*f005ef32Sjsg * unsigned long index;
26*f005ef32Sjsg * int ret;
27*f005ef32Sjsg *
28*f005ef32Sjsg * drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT);
29*f005ef32Sjsg * drm_exec_until_all_locked(&exec) {
30*f005ef32Sjsg * ret = drm_exec_prepare_obj(&exec, boA, 1);
31*f005ef32Sjsg * drm_exec_retry_on_contention(&exec);
32*f005ef32Sjsg * if (ret)
33*f005ef32Sjsg * goto error;
34*f005ef32Sjsg *
35*f005ef32Sjsg * ret = drm_exec_prepare_obj(&exec, boB, 1);
36*f005ef32Sjsg * drm_exec_retry_on_contention(&exec);
37*f005ef32Sjsg * if (ret)
38*f005ef32Sjsg * goto error;
39*f005ef32Sjsg * }
40*f005ef32Sjsg *
41*f005ef32Sjsg * drm_exec_for_each_locked_object(&exec, index, obj) {
42*f005ef32Sjsg * dma_resv_add_fence(obj->resv, fence, DMA_RESV_USAGE_READ);
43*f005ef32Sjsg * ...
44*f005ef32Sjsg * }
45*f005ef32Sjsg * drm_exec_fini(&exec);
46*f005ef32Sjsg *
47*f005ef32Sjsg * See struct dma_exec for more details.
48*f005ef32Sjsg */
49*f005ef32Sjsg
50*f005ef32Sjsg /* Dummy value used to initially enter the retry loop */
51*f005ef32Sjsg #define DRM_EXEC_DUMMY ((void *)~0)
52*f005ef32Sjsg
53*f005ef32Sjsg /* Unlock all objects and drop references */
drm_exec_unlock_all(struct drm_exec * exec)54*f005ef32Sjsg static void drm_exec_unlock_all(struct drm_exec *exec)
55*f005ef32Sjsg {
56*f005ef32Sjsg struct drm_gem_object *obj;
57*f005ef32Sjsg unsigned long index;
58*f005ef32Sjsg
59*f005ef32Sjsg drm_exec_for_each_locked_object_reverse(exec, index, obj) {
60*f005ef32Sjsg dma_resv_unlock(obj->resv);
61*f005ef32Sjsg drm_gem_object_put(obj);
62*f005ef32Sjsg }
63*f005ef32Sjsg
64*f005ef32Sjsg drm_gem_object_put(exec->prelocked);
65*f005ef32Sjsg exec->prelocked = NULL;
66*f005ef32Sjsg }
67*f005ef32Sjsg
68*f005ef32Sjsg /**
69*f005ef32Sjsg * drm_exec_init - initialize a drm_exec object
70*f005ef32Sjsg * @exec: the drm_exec object to initialize
71*f005ef32Sjsg * @flags: controls locking behavior, see DRM_EXEC_* defines
72*f005ef32Sjsg *
73*f005ef32Sjsg * Initialize the object and make sure that we can track locked objects.
74*f005ef32Sjsg */
drm_exec_init(struct drm_exec * exec,uint32_t flags)75*f005ef32Sjsg void drm_exec_init(struct drm_exec *exec, uint32_t flags)
76*f005ef32Sjsg {
77*f005ef32Sjsg exec->flags = flags;
78*f005ef32Sjsg exec->objects = kmalloc(PAGE_SIZE, GFP_KERNEL);
79*f005ef32Sjsg
80*f005ef32Sjsg /* If allocation here fails, just delay that till the first use */
81*f005ef32Sjsg exec->max_objects = exec->objects ? PAGE_SIZE / sizeof(void *) : 0;
82*f005ef32Sjsg exec->num_objects = 0;
83*f005ef32Sjsg exec->contended = DRM_EXEC_DUMMY;
84*f005ef32Sjsg exec->prelocked = NULL;
85*f005ef32Sjsg }
86*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_init);
87*f005ef32Sjsg
88*f005ef32Sjsg /**
89*f005ef32Sjsg * drm_exec_fini - finalize a drm_exec object
90*f005ef32Sjsg * @exec: the drm_exec object to finalize
91*f005ef32Sjsg *
92*f005ef32Sjsg * Unlock all locked objects, drop the references to objects and free all memory
93*f005ef32Sjsg * used for tracking the state.
94*f005ef32Sjsg */
drm_exec_fini(struct drm_exec * exec)95*f005ef32Sjsg void drm_exec_fini(struct drm_exec *exec)
96*f005ef32Sjsg {
97*f005ef32Sjsg drm_exec_unlock_all(exec);
98*f005ef32Sjsg kvfree(exec->objects);
99*f005ef32Sjsg if (exec->contended != DRM_EXEC_DUMMY) {
100*f005ef32Sjsg drm_gem_object_put(exec->contended);
101*f005ef32Sjsg ww_acquire_fini(&exec->ticket);
102*f005ef32Sjsg }
103*f005ef32Sjsg }
104*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_fini);
105*f005ef32Sjsg
106*f005ef32Sjsg /**
107*f005ef32Sjsg * drm_exec_cleanup - cleanup when contention is detected
108*f005ef32Sjsg * @exec: the drm_exec object to cleanup
109*f005ef32Sjsg *
110*f005ef32Sjsg * Cleanup the current state and return true if we should stay inside the retry
111*f005ef32Sjsg * loop, false if there wasn't any contention detected and we can keep the
112*f005ef32Sjsg * objects locked.
113*f005ef32Sjsg */
drm_exec_cleanup(struct drm_exec * exec)114*f005ef32Sjsg bool drm_exec_cleanup(struct drm_exec *exec)
115*f005ef32Sjsg {
116*f005ef32Sjsg if (likely(!exec->contended)) {
117*f005ef32Sjsg ww_acquire_done(&exec->ticket);
118*f005ef32Sjsg return false;
119*f005ef32Sjsg }
120*f005ef32Sjsg
121*f005ef32Sjsg if (likely(exec->contended == DRM_EXEC_DUMMY)) {
122*f005ef32Sjsg exec->contended = NULL;
123*f005ef32Sjsg ww_acquire_init(&exec->ticket, &reservation_ww_class);
124*f005ef32Sjsg return true;
125*f005ef32Sjsg }
126*f005ef32Sjsg
127*f005ef32Sjsg drm_exec_unlock_all(exec);
128*f005ef32Sjsg exec->num_objects = 0;
129*f005ef32Sjsg return true;
130*f005ef32Sjsg }
131*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_cleanup);
132*f005ef32Sjsg
133*f005ef32Sjsg /* Track the locked object in the array */
drm_exec_obj_locked(struct drm_exec * exec,struct drm_gem_object * obj)134*f005ef32Sjsg static int drm_exec_obj_locked(struct drm_exec *exec,
135*f005ef32Sjsg struct drm_gem_object *obj)
136*f005ef32Sjsg {
137*f005ef32Sjsg if (unlikely(exec->num_objects == exec->max_objects)) {
138*f005ef32Sjsg size_t size = exec->max_objects * sizeof(void *);
139*f005ef32Sjsg void *tmp;
140*f005ef32Sjsg
141*f005ef32Sjsg #ifdef __linux__
142*f005ef32Sjsg tmp = kvrealloc(exec->objects, size, size + PAGE_SIZE,
143*f005ef32Sjsg GFP_KERNEL);
144*f005ef32Sjsg if (!tmp)
145*f005ef32Sjsg return -ENOMEM;
146*f005ef32Sjsg #else
147*f005ef32Sjsg tmp = kvmalloc(size + PAGE_SIZE, GFP_KERNEL);
148*f005ef32Sjsg if (!tmp)
149*f005ef32Sjsg return -ENOMEM;
150*f005ef32Sjsg memcpy(tmp, exec->objects, size);
151*f005ef32Sjsg kvfree(exec->objects);
152*f005ef32Sjsg #endif
153*f005ef32Sjsg
154*f005ef32Sjsg exec->objects = tmp;
155*f005ef32Sjsg exec->max_objects += PAGE_SIZE / sizeof(void *);
156*f005ef32Sjsg }
157*f005ef32Sjsg drm_gem_object_get(obj);
158*f005ef32Sjsg exec->objects[exec->num_objects++] = obj;
159*f005ef32Sjsg
160*f005ef32Sjsg return 0;
161*f005ef32Sjsg }
162*f005ef32Sjsg
163*f005ef32Sjsg /* Make sure the contended object is locked first */
drm_exec_lock_contended(struct drm_exec * exec)164*f005ef32Sjsg static int drm_exec_lock_contended(struct drm_exec *exec)
165*f005ef32Sjsg {
166*f005ef32Sjsg struct drm_gem_object *obj = exec->contended;
167*f005ef32Sjsg int ret;
168*f005ef32Sjsg
169*f005ef32Sjsg if (likely(!obj))
170*f005ef32Sjsg return 0;
171*f005ef32Sjsg
172*f005ef32Sjsg /* Always cleanup the contention so that error handling can kick in */
173*f005ef32Sjsg exec->contended = NULL;
174*f005ef32Sjsg if (exec->flags & DRM_EXEC_INTERRUPTIBLE_WAIT) {
175*f005ef32Sjsg ret = dma_resv_lock_slow_interruptible(obj->resv,
176*f005ef32Sjsg &exec->ticket);
177*f005ef32Sjsg if (unlikely(ret))
178*f005ef32Sjsg goto error_dropref;
179*f005ef32Sjsg } else {
180*f005ef32Sjsg dma_resv_lock_slow(obj->resv, &exec->ticket);
181*f005ef32Sjsg }
182*f005ef32Sjsg
183*f005ef32Sjsg ret = drm_exec_obj_locked(exec, obj);
184*f005ef32Sjsg if (unlikely(ret))
185*f005ef32Sjsg goto error_unlock;
186*f005ef32Sjsg
187*f005ef32Sjsg exec->prelocked = obj;
188*f005ef32Sjsg return 0;
189*f005ef32Sjsg
190*f005ef32Sjsg error_unlock:
191*f005ef32Sjsg dma_resv_unlock(obj->resv);
192*f005ef32Sjsg
193*f005ef32Sjsg error_dropref:
194*f005ef32Sjsg drm_gem_object_put(obj);
195*f005ef32Sjsg return ret;
196*f005ef32Sjsg }
197*f005ef32Sjsg
198*f005ef32Sjsg /**
199*f005ef32Sjsg * drm_exec_lock_obj - lock a GEM object for use
200*f005ef32Sjsg * @exec: the drm_exec object with the state
201*f005ef32Sjsg * @obj: the GEM object to lock
202*f005ef32Sjsg *
203*f005ef32Sjsg * Lock a GEM object for use and grab a reference to it.
204*f005ef32Sjsg *
205*f005ef32Sjsg * Returns: -EDEADLK if a contention is detected, -EALREADY when object is
206*f005ef32Sjsg * already locked (can be suppressed by setting the DRM_EXEC_IGNORE_DUPLICATES
207*f005ef32Sjsg * flag), -ENOMEM when memory allocation failed and zero for success.
208*f005ef32Sjsg */
drm_exec_lock_obj(struct drm_exec * exec,struct drm_gem_object * obj)209*f005ef32Sjsg int drm_exec_lock_obj(struct drm_exec *exec, struct drm_gem_object *obj)
210*f005ef32Sjsg {
211*f005ef32Sjsg int ret;
212*f005ef32Sjsg
213*f005ef32Sjsg ret = drm_exec_lock_contended(exec);
214*f005ef32Sjsg if (unlikely(ret))
215*f005ef32Sjsg return ret;
216*f005ef32Sjsg
217*f005ef32Sjsg if (exec->prelocked == obj) {
218*f005ef32Sjsg drm_gem_object_put(exec->prelocked);
219*f005ef32Sjsg exec->prelocked = NULL;
220*f005ef32Sjsg return 0;
221*f005ef32Sjsg }
222*f005ef32Sjsg
223*f005ef32Sjsg if (exec->flags & DRM_EXEC_INTERRUPTIBLE_WAIT)
224*f005ef32Sjsg ret = dma_resv_lock_interruptible(obj->resv, &exec->ticket);
225*f005ef32Sjsg else
226*f005ef32Sjsg ret = dma_resv_lock(obj->resv, &exec->ticket);
227*f005ef32Sjsg
228*f005ef32Sjsg if (unlikely(ret == -EDEADLK)) {
229*f005ef32Sjsg drm_gem_object_get(obj);
230*f005ef32Sjsg exec->contended = obj;
231*f005ef32Sjsg return -EDEADLK;
232*f005ef32Sjsg }
233*f005ef32Sjsg
234*f005ef32Sjsg if (unlikely(ret == -EALREADY) &&
235*f005ef32Sjsg exec->flags & DRM_EXEC_IGNORE_DUPLICATES)
236*f005ef32Sjsg return 0;
237*f005ef32Sjsg
238*f005ef32Sjsg if (unlikely(ret))
239*f005ef32Sjsg return ret;
240*f005ef32Sjsg
241*f005ef32Sjsg ret = drm_exec_obj_locked(exec, obj);
242*f005ef32Sjsg if (ret)
243*f005ef32Sjsg goto error_unlock;
244*f005ef32Sjsg
245*f005ef32Sjsg return 0;
246*f005ef32Sjsg
247*f005ef32Sjsg error_unlock:
248*f005ef32Sjsg dma_resv_unlock(obj->resv);
249*f005ef32Sjsg return ret;
250*f005ef32Sjsg }
251*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_lock_obj);
252*f005ef32Sjsg
253*f005ef32Sjsg /**
254*f005ef32Sjsg * drm_exec_unlock_obj - unlock a GEM object in this exec context
255*f005ef32Sjsg * @exec: the drm_exec object with the state
256*f005ef32Sjsg * @obj: the GEM object to unlock
257*f005ef32Sjsg *
258*f005ef32Sjsg * Unlock the GEM object and remove it from the collection of locked objects.
259*f005ef32Sjsg * Should only be used to unlock the most recently locked objects. It's not time
260*f005ef32Sjsg * efficient to unlock objects locked long ago.
261*f005ef32Sjsg */
drm_exec_unlock_obj(struct drm_exec * exec,struct drm_gem_object * obj)262*f005ef32Sjsg void drm_exec_unlock_obj(struct drm_exec *exec, struct drm_gem_object *obj)
263*f005ef32Sjsg {
264*f005ef32Sjsg unsigned int i;
265*f005ef32Sjsg
266*f005ef32Sjsg for (i = exec->num_objects; i--;) {
267*f005ef32Sjsg if (exec->objects[i] == obj) {
268*f005ef32Sjsg dma_resv_unlock(obj->resv);
269*f005ef32Sjsg for (++i; i < exec->num_objects; ++i)
270*f005ef32Sjsg exec->objects[i - 1] = exec->objects[i];
271*f005ef32Sjsg --exec->num_objects;
272*f005ef32Sjsg drm_gem_object_put(obj);
273*f005ef32Sjsg return;
274*f005ef32Sjsg }
275*f005ef32Sjsg
276*f005ef32Sjsg }
277*f005ef32Sjsg }
278*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_unlock_obj);
279*f005ef32Sjsg
280*f005ef32Sjsg /**
281*f005ef32Sjsg * drm_exec_prepare_obj - prepare a GEM object for use
282*f005ef32Sjsg * @exec: the drm_exec object with the state
283*f005ef32Sjsg * @obj: the GEM object to prepare
284*f005ef32Sjsg * @num_fences: how many fences to reserve
285*f005ef32Sjsg *
286*f005ef32Sjsg * Prepare a GEM object for use by locking it and reserving fence slots.
287*f005ef32Sjsg *
288*f005ef32Sjsg * Returns: -EDEADLK if a contention is detected, -EALREADY when object is
289*f005ef32Sjsg * already locked, -ENOMEM when memory allocation failed and zero for success.
290*f005ef32Sjsg */
drm_exec_prepare_obj(struct drm_exec * exec,struct drm_gem_object * obj,unsigned int num_fences)291*f005ef32Sjsg int drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj,
292*f005ef32Sjsg unsigned int num_fences)
293*f005ef32Sjsg {
294*f005ef32Sjsg int ret;
295*f005ef32Sjsg
296*f005ef32Sjsg ret = drm_exec_lock_obj(exec, obj);
297*f005ef32Sjsg if (ret)
298*f005ef32Sjsg return ret;
299*f005ef32Sjsg
300*f005ef32Sjsg ret = dma_resv_reserve_fences(obj->resv, num_fences);
301*f005ef32Sjsg if (ret) {
302*f005ef32Sjsg drm_exec_unlock_obj(exec, obj);
303*f005ef32Sjsg return ret;
304*f005ef32Sjsg }
305*f005ef32Sjsg
306*f005ef32Sjsg return 0;
307*f005ef32Sjsg }
308*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_prepare_obj);
309*f005ef32Sjsg
310*f005ef32Sjsg /**
311*f005ef32Sjsg * drm_exec_prepare_array - helper to prepare an array of objects
312*f005ef32Sjsg * @exec: the drm_exec object with the state
313*f005ef32Sjsg * @objects: array of GEM object to prepare
314*f005ef32Sjsg * @num_objects: number of GEM objects in the array
315*f005ef32Sjsg * @num_fences: number of fences to reserve on each GEM object
316*f005ef32Sjsg *
317*f005ef32Sjsg * Prepares all GEM objects in an array, aborts on first error.
318*f005ef32Sjsg * Reserves @num_fences on each GEM object after locking it.
319*f005ef32Sjsg *
320*f005ef32Sjsg * Returns: -EDEADLOCK on contention, -EALREADY when object is already locked,
321*f005ef32Sjsg * -ENOMEM when memory allocation failed and zero for success.
322*f005ef32Sjsg */
drm_exec_prepare_array(struct drm_exec * exec,struct drm_gem_object ** objects,unsigned int num_objects,unsigned int num_fences)323*f005ef32Sjsg int drm_exec_prepare_array(struct drm_exec *exec,
324*f005ef32Sjsg struct drm_gem_object **objects,
325*f005ef32Sjsg unsigned int num_objects,
326*f005ef32Sjsg unsigned int num_fences)
327*f005ef32Sjsg {
328*f005ef32Sjsg int ret;
329*f005ef32Sjsg
330*f005ef32Sjsg for (unsigned int i = 0; i < num_objects; ++i) {
331*f005ef32Sjsg ret = drm_exec_prepare_obj(exec, objects[i], num_fences);
332*f005ef32Sjsg if (unlikely(ret))
333*f005ef32Sjsg return ret;
334*f005ef32Sjsg }
335*f005ef32Sjsg
336*f005ef32Sjsg return 0;
337*f005ef32Sjsg }
338*f005ef32Sjsg EXPORT_SYMBOL(drm_exec_prepare_array);
339*f005ef32Sjsg
340*f005ef32Sjsg MODULE_DESCRIPTION("DRM execution context");
341*f005ef32Sjsg MODULE_LICENSE("Dual MIT/GPL");
342