1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun * SPDX-License-Identifier: MIT
3*4882a593Smuzhiyun *
4*4882a593Smuzhiyun * (C) Copyright 2016 Intel Corporation
5*4882a593Smuzhiyun */
6*4882a593Smuzhiyun
7*4882a593Smuzhiyun #include <linux/slab.h>
8*4882a593Smuzhiyun #include <linux/dma-fence.h>
9*4882a593Smuzhiyun #include <linux/irq_work.h>
10*4882a593Smuzhiyun #include <linux/dma-resv.h>
11*4882a593Smuzhiyun
12*4882a593Smuzhiyun #include "i915_sw_fence.h"
13*4882a593Smuzhiyun #include "i915_selftest.h"
14*4882a593Smuzhiyun
15*4882a593Smuzhiyun #if IS_ENABLED(CONFIG_DRM_I915_DEBUG)
16*4882a593Smuzhiyun #define I915_SW_FENCE_BUG_ON(expr) BUG_ON(expr)
17*4882a593Smuzhiyun #else
18*4882a593Smuzhiyun #define I915_SW_FENCE_BUG_ON(expr) BUILD_BUG_ON_INVALID(expr)
19*4882a593Smuzhiyun #endif
20*4882a593Smuzhiyun
21*4882a593Smuzhiyun #define I915_SW_FENCE_FLAG_ALLOC BIT(3) /* after WQ_FLAG_* for safety */
22*4882a593Smuzhiyun
23*4882a593Smuzhiyun static DEFINE_SPINLOCK(i915_sw_fence_lock);
24*4882a593Smuzhiyun
25*4882a593Smuzhiyun enum {
26*4882a593Smuzhiyun DEBUG_FENCE_IDLE = 0,
27*4882a593Smuzhiyun DEBUG_FENCE_NOTIFY,
28*4882a593Smuzhiyun };
29*4882a593Smuzhiyun
i915_sw_fence_debug_hint(void * addr)30*4882a593Smuzhiyun static void *i915_sw_fence_debug_hint(void *addr)
31*4882a593Smuzhiyun {
32*4882a593Smuzhiyun return (void *)(((struct i915_sw_fence *)addr)->flags & I915_SW_FENCE_MASK);
33*4882a593Smuzhiyun }
34*4882a593Smuzhiyun
35*4882a593Smuzhiyun #ifdef CONFIG_DRM_I915_SW_FENCE_DEBUG_OBJECTS
36*4882a593Smuzhiyun
37*4882a593Smuzhiyun static const struct debug_obj_descr i915_sw_fence_debug_descr = {
38*4882a593Smuzhiyun .name = "i915_sw_fence",
39*4882a593Smuzhiyun .debug_hint = i915_sw_fence_debug_hint,
40*4882a593Smuzhiyun };
41*4882a593Smuzhiyun
debug_fence_init(struct i915_sw_fence * fence)42*4882a593Smuzhiyun static inline void debug_fence_init(struct i915_sw_fence *fence)
43*4882a593Smuzhiyun {
44*4882a593Smuzhiyun debug_object_init(fence, &i915_sw_fence_debug_descr);
45*4882a593Smuzhiyun }
46*4882a593Smuzhiyun
debug_fence_init_onstack(struct i915_sw_fence * fence)47*4882a593Smuzhiyun static inline void debug_fence_init_onstack(struct i915_sw_fence *fence)
48*4882a593Smuzhiyun {
49*4882a593Smuzhiyun debug_object_init_on_stack(fence, &i915_sw_fence_debug_descr);
50*4882a593Smuzhiyun }
51*4882a593Smuzhiyun
debug_fence_activate(struct i915_sw_fence * fence)52*4882a593Smuzhiyun static inline void debug_fence_activate(struct i915_sw_fence *fence)
53*4882a593Smuzhiyun {
54*4882a593Smuzhiyun debug_object_activate(fence, &i915_sw_fence_debug_descr);
55*4882a593Smuzhiyun }
56*4882a593Smuzhiyun
debug_fence_set_state(struct i915_sw_fence * fence,int old,int new)57*4882a593Smuzhiyun static inline void debug_fence_set_state(struct i915_sw_fence *fence,
58*4882a593Smuzhiyun int old, int new)
59*4882a593Smuzhiyun {
60*4882a593Smuzhiyun debug_object_active_state(fence, &i915_sw_fence_debug_descr, old, new);
61*4882a593Smuzhiyun }
62*4882a593Smuzhiyun
debug_fence_deactivate(struct i915_sw_fence * fence)63*4882a593Smuzhiyun static inline void debug_fence_deactivate(struct i915_sw_fence *fence)
64*4882a593Smuzhiyun {
65*4882a593Smuzhiyun debug_object_deactivate(fence, &i915_sw_fence_debug_descr);
66*4882a593Smuzhiyun }
67*4882a593Smuzhiyun
debug_fence_destroy(struct i915_sw_fence * fence)68*4882a593Smuzhiyun static inline void debug_fence_destroy(struct i915_sw_fence *fence)
69*4882a593Smuzhiyun {
70*4882a593Smuzhiyun debug_object_destroy(fence, &i915_sw_fence_debug_descr);
71*4882a593Smuzhiyun }
72*4882a593Smuzhiyun
debug_fence_free(struct i915_sw_fence * fence)73*4882a593Smuzhiyun static inline void debug_fence_free(struct i915_sw_fence *fence)
74*4882a593Smuzhiyun {
75*4882a593Smuzhiyun debug_object_free(fence, &i915_sw_fence_debug_descr);
76*4882a593Smuzhiyun smp_wmb(); /* flush the change in state before reallocation */
77*4882a593Smuzhiyun }
78*4882a593Smuzhiyun
debug_fence_assert(struct i915_sw_fence * fence)79*4882a593Smuzhiyun static inline void debug_fence_assert(struct i915_sw_fence *fence)
80*4882a593Smuzhiyun {
81*4882a593Smuzhiyun debug_object_assert_init(fence, &i915_sw_fence_debug_descr);
82*4882a593Smuzhiyun }
83*4882a593Smuzhiyun
84*4882a593Smuzhiyun #else
85*4882a593Smuzhiyun
debug_fence_init(struct i915_sw_fence * fence)86*4882a593Smuzhiyun static inline void debug_fence_init(struct i915_sw_fence *fence)
87*4882a593Smuzhiyun {
88*4882a593Smuzhiyun }
89*4882a593Smuzhiyun
debug_fence_init_onstack(struct i915_sw_fence * fence)90*4882a593Smuzhiyun static inline void debug_fence_init_onstack(struct i915_sw_fence *fence)
91*4882a593Smuzhiyun {
92*4882a593Smuzhiyun }
93*4882a593Smuzhiyun
debug_fence_activate(struct i915_sw_fence * fence)94*4882a593Smuzhiyun static inline void debug_fence_activate(struct i915_sw_fence *fence)
95*4882a593Smuzhiyun {
96*4882a593Smuzhiyun }
97*4882a593Smuzhiyun
debug_fence_set_state(struct i915_sw_fence * fence,int old,int new)98*4882a593Smuzhiyun static inline void debug_fence_set_state(struct i915_sw_fence *fence,
99*4882a593Smuzhiyun int old, int new)
100*4882a593Smuzhiyun {
101*4882a593Smuzhiyun }
102*4882a593Smuzhiyun
debug_fence_deactivate(struct i915_sw_fence * fence)103*4882a593Smuzhiyun static inline void debug_fence_deactivate(struct i915_sw_fence *fence)
104*4882a593Smuzhiyun {
105*4882a593Smuzhiyun }
106*4882a593Smuzhiyun
debug_fence_destroy(struct i915_sw_fence * fence)107*4882a593Smuzhiyun static inline void debug_fence_destroy(struct i915_sw_fence *fence)
108*4882a593Smuzhiyun {
109*4882a593Smuzhiyun }
110*4882a593Smuzhiyun
debug_fence_free(struct i915_sw_fence * fence)111*4882a593Smuzhiyun static inline void debug_fence_free(struct i915_sw_fence *fence)
112*4882a593Smuzhiyun {
113*4882a593Smuzhiyun }
114*4882a593Smuzhiyun
debug_fence_assert(struct i915_sw_fence * fence)115*4882a593Smuzhiyun static inline void debug_fence_assert(struct i915_sw_fence *fence)
116*4882a593Smuzhiyun {
117*4882a593Smuzhiyun }
118*4882a593Smuzhiyun
119*4882a593Smuzhiyun #endif
120*4882a593Smuzhiyun
__i915_sw_fence_notify(struct i915_sw_fence * fence,enum i915_sw_fence_notify state)121*4882a593Smuzhiyun static int __i915_sw_fence_notify(struct i915_sw_fence *fence,
122*4882a593Smuzhiyun enum i915_sw_fence_notify state)
123*4882a593Smuzhiyun {
124*4882a593Smuzhiyun i915_sw_fence_notify_t fn;
125*4882a593Smuzhiyun
126*4882a593Smuzhiyun fn = (i915_sw_fence_notify_t)(fence->flags & I915_SW_FENCE_MASK);
127*4882a593Smuzhiyun return fn(fence, state);
128*4882a593Smuzhiyun }
129*4882a593Smuzhiyun
130*4882a593Smuzhiyun #ifdef CONFIG_DRM_I915_SW_FENCE_DEBUG_OBJECTS
i915_sw_fence_fini(struct i915_sw_fence * fence)131*4882a593Smuzhiyun void i915_sw_fence_fini(struct i915_sw_fence *fence)
132*4882a593Smuzhiyun {
133*4882a593Smuzhiyun debug_fence_free(fence);
134*4882a593Smuzhiyun }
135*4882a593Smuzhiyun #endif
136*4882a593Smuzhiyun
__i915_sw_fence_wake_up_all(struct i915_sw_fence * fence,struct list_head * continuation)137*4882a593Smuzhiyun static void __i915_sw_fence_wake_up_all(struct i915_sw_fence *fence,
138*4882a593Smuzhiyun struct list_head *continuation)
139*4882a593Smuzhiyun {
140*4882a593Smuzhiyun wait_queue_head_t *x = &fence->wait;
141*4882a593Smuzhiyun wait_queue_entry_t *pos, *next;
142*4882a593Smuzhiyun unsigned long flags;
143*4882a593Smuzhiyun
144*4882a593Smuzhiyun debug_fence_deactivate(fence);
145*4882a593Smuzhiyun atomic_set_release(&fence->pending, -1); /* 0 -> -1 [done] */
146*4882a593Smuzhiyun
147*4882a593Smuzhiyun /*
148*4882a593Smuzhiyun * To prevent unbounded recursion as we traverse the graph of
149*4882a593Smuzhiyun * i915_sw_fences, we move the entry list from this, the next ready
150*4882a593Smuzhiyun * fence, to the tail of the original fence's entry list
151*4882a593Smuzhiyun * (and so added to the list to be woken).
152*4882a593Smuzhiyun */
153*4882a593Smuzhiyun
154*4882a593Smuzhiyun spin_lock_irqsave_nested(&x->lock, flags, 1 + !!continuation);
155*4882a593Smuzhiyun if (continuation) {
156*4882a593Smuzhiyun list_for_each_entry_safe(pos, next, &x->head, entry) {
157*4882a593Smuzhiyun if (pos->func == autoremove_wake_function)
158*4882a593Smuzhiyun pos->func(pos, TASK_NORMAL, 0, continuation);
159*4882a593Smuzhiyun else
160*4882a593Smuzhiyun list_move_tail(&pos->entry, continuation);
161*4882a593Smuzhiyun }
162*4882a593Smuzhiyun } else {
163*4882a593Smuzhiyun LIST_HEAD(extra);
164*4882a593Smuzhiyun
165*4882a593Smuzhiyun do {
166*4882a593Smuzhiyun list_for_each_entry_safe(pos, next, &x->head, entry) {
167*4882a593Smuzhiyun int wake_flags;
168*4882a593Smuzhiyun
169*4882a593Smuzhiyun wake_flags = fence->error;
170*4882a593Smuzhiyun if (pos->func == autoremove_wake_function)
171*4882a593Smuzhiyun wake_flags = 0;
172*4882a593Smuzhiyun
173*4882a593Smuzhiyun pos->func(pos, TASK_NORMAL, wake_flags, &extra);
174*4882a593Smuzhiyun }
175*4882a593Smuzhiyun
176*4882a593Smuzhiyun if (list_empty(&extra))
177*4882a593Smuzhiyun break;
178*4882a593Smuzhiyun
179*4882a593Smuzhiyun list_splice_tail_init(&extra, &x->head);
180*4882a593Smuzhiyun } while (1);
181*4882a593Smuzhiyun }
182*4882a593Smuzhiyun spin_unlock_irqrestore(&x->lock, flags);
183*4882a593Smuzhiyun
184*4882a593Smuzhiyun debug_fence_assert(fence);
185*4882a593Smuzhiyun }
186*4882a593Smuzhiyun
__i915_sw_fence_complete(struct i915_sw_fence * fence,struct list_head * continuation)187*4882a593Smuzhiyun static void __i915_sw_fence_complete(struct i915_sw_fence *fence,
188*4882a593Smuzhiyun struct list_head *continuation)
189*4882a593Smuzhiyun {
190*4882a593Smuzhiyun debug_fence_assert(fence);
191*4882a593Smuzhiyun
192*4882a593Smuzhiyun if (!atomic_dec_and_test(&fence->pending))
193*4882a593Smuzhiyun return;
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun debug_fence_set_state(fence, DEBUG_FENCE_IDLE, DEBUG_FENCE_NOTIFY);
196*4882a593Smuzhiyun
197*4882a593Smuzhiyun if (__i915_sw_fence_notify(fence, FENCE_COMPLETE) != NOTIFY_DONE)
198*4882a593Smuzhiyun return;
199*4882a593Smuzhiyun
200*4882a593Smuzhiyun debug_fence_set_state(fence, DEBUG_FENCE_NOTIFY, DEBUG_FENCE_IDLE);
201*4882a593Smuzhiyun
202*4882a593Smuzhiyun __i915_sw_fence_wake_up_all(fence, continuation);
203*4882a593Smuzhiyun
204*4882a593Smuzhiyun debug_fence_destroy(fence);
205*4882a593Smuzhiyun __i915_sw_fence_notify(fence, FENCE_FREE);
206*4882a593Smuzhiyun }
207*4882a593Smuzhiyun
i915_sw_fence_complete(struct i915_sw_fence * fence)208*4882a593Smuzhiyun void i915_sw_fence_complete(struct i915_sw_fence *fence)
209*4882a593Smuzhiyun {
210*4882a593Smuzhiyun debug_fence_assert(fence);
211*4882a593Smuzhiyun
212*4882a593Smuzhiyun if (WARN_ON(i915_sw_fence_done(fence)))
213*4882a593Smuzhiyun return;
214*4882a593Smuzhiyun
215*4882a593Smuzhiyun __i915_sw_fence_complete(fence, NULL);
216*4882a593Smuzhiyun }
217*4882a593Smuzhiyun
i915_sw_fence_await(struct i915_sw_fence * fence)218*4882a593Smuzhiyun bool i915_sw_fence_await(struct i915_sw_fence *fence)
219*4882a593Smuzhiyun {
220*4882a593Smuzhiyun int pending;
221*4882a593Smuzhiyun
222*4882a593Smuzhiyun /*
223*4882a593Smuzhiyun * It is only safe to add a new await to the fence while it has
224*4882a593Smuzhiyun * not yet been signaled (i.e. there are still existing signalers).
225*4882a593Smuzhiyun */
226*4882a593Smuzhiyun pending = atomic_read(&fence->pending);
227*4882a593Smuzhiyun do {
228*4882a593Smuzhiyun if (pending < 1)
229*4882a593Smuzhiyun return false;
230*4882a593Smuzhiyun } while (!atomic_try_cmpxchg(&fence->pending, &pending, pending + 1));
231*4882a593Smuzhiyun
232*4882a593Smuzhiyun return true;
233*4882a593Smuzhiyun }
234*4882a593Smuzhiyun
__i915_sw_fence_init(struct i915_sw_fence * fence,i915_sw_fence_notify_t fn,const char * name,struct lock_class_key * key)235*4882a593Smuzhiyun void __i915_sw_fence_init(struct i915_sw_fence *fence,
236*4882a593Smuzhiyun i915_sw_fence_notify_t fn,
237*4882a593Smuzhiyun const char *name,
238*4882a593Smuzhiyun struct lock_class_key *key)
239*4882a593Smuzhiyun {
240*4882a593Smuzhiyun BUG_ON(!fn || (unsigned long)fn & ~I915_SW_FENCE_MASK);
241*4882a593Smuzhiyun
242*4882a593Smuzhiyun __init_waitqueue_head(&fence->wait, name, key);
243*4882a593Smuzhiyun fence->flags = (unsigned long)fn;
244*4882a593Smuzhiyun
245*4882a593Smuzhiyun i915_sw_fence_reinit(fence);
246*4882a593Smuzhiyun }
247*4882a593Smuzhiyun
i915_sw_fence_reinit(struct i915_sw_fence * fence)248*4882a593Smuzhiyun void i915_sw_fence_reinit(struct i915_sw_fence *fence)
249*4882a593Smuzhiyun {
250*4882a593Smuzhiyun debug_fence_init(fence);
251*4882a593Smuzhiyun
252*4882a593Smuzhiyun atomic_set(&fence->pending, 1);
253*4882a593Smuzhiyun fence->error = 0;
254*4882a593Smuzhiyun
255*4882a593Smuzhiyun I915_SW_FENCE_BUG_ON(!fence->flags);
256*4882a593Smuzhiyun I915_SW_FENCE_BUG_ON(!list_empty(&fence->wait.head));
257*4882a593Smuzhiyun }
258*4882a593Smuzhiyun
i915_sw_fence_commit(struct i915_sw_fence * fence)259*4882a593Smuzhiyun void i915_sw_fence_commit(struct i915_sw_fence *fence)
260*4882a593Smuzhiyun {
261*4882a593Smuzhiyun debug_fence_activate(fence);
262*4882a593Smuzhiyun i915_sw_fence_complete(fence);
263*4882a593Smuzhiyun }
264*4882a593Smuzhiyun
i915_sw_fence_wake(wait_queue_entry_t * wq,unsigned mode,int flags,void * key)265*4882a593Smuzhiyun static int i915_sw_fence_wake(wait_queue_entry_t *wq, unsigned mode, int flags, void *key)
266*4882a593Smuzhiyun {
267*4882a593Smuzhiyun i915_sw_fence_set_error_once(wq->private, flags);
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun list_del(&wq->entry);
270*4882a593Smuzhiyun __i915_sw_fence_complete(wq->private, key);
271*4882a593Smuzhiyun
272*4882a593Smuzhiyun if (wq->flags & I915_SW_FENCE_FLAG_ALLOC)
273*4882a593Smuzhiyun kfree(wq);
274*4882a593Smuzhiyun return 0;
275*4882a593Smuzhiyun }
276*4882a593Smuzhiyun
__i915_sw_fence_check_if_after(struct i915_sw_fence * fence,const struct i915_sw_fence * const signaler)277*4882a593Smuzhiyun static bool __i915_sw_fence_check_if_after(struct i915_sw_fence *fence,
278*4882a593Smuzhiyun const struct i915_sw_fence * const signaler)
279*4882a593Smuzhiyun {
280*4882a593Smuzhiyun wait_queue_entry_t *wq;
281*4882a593Smuzhiyun
282*4882a593Smuzhiyun if (__test_and_set_bit(I915_SW_FENCE_CHECKED_BIT, &fence->flags))
283*4882a593Smuzhiyun return false;
284*4882a593Smuzhiyun
285*4882a593Smuzhiyun if (fence == signaler)
286*4882a593Smuzhiyun return true;
287*4882a593Smuzhiyun
288*4882a593Smuzhiyun list_for_each_entry(wq, &fence->wait.head, entry) {
289*4882a593Smuzhiyun if (wq->func != i915_sw_fence_wake)
290*4882a593Smuzhiyun continue;
291*4882a593Smuzhiyun
292*4882a593Smuzhiyun if (__i915_sw_fence_check_if_after(wq->private, signaler))
293*4882a593Smuzhiyun return true;
294*4882a593Smuzhiyun }
295*4882a593Smuzhiyun
296*4882a593Smuzhiyun return false;
297*4882a593Smuzhiyun }
298*4882a593Smuzhiyun
__i915_sw_fence_clear_checked_bit(struct i915_sw_fence * fence)299*4882a593Smuzhiyun static void __i915_sw_fence_clear_checked_bit(struct i915_sw_fence *fence)
300*4882a593Smuzhiyun {
301*4882a593Smuzhiyun wait_queue_entry_t *wq;
302*4882a593Smuzhiyun
303*4882a593Smuzhiyun if (!__test_and_clear_bit(I915_SW_FENCE_CHECKED_BIT, &fence->flags))
304*4882a593Smuzhiyun return;
305*4882a593Smuzhiyun
306*4882a593Smuzhiyun list_for_each_entry(wq, &fence->wait.head, entry) {
307*4882a593Smuzhiyun if (wq->func != i915_sw_fence_wake)
308*4882a593Smuzhiyun continue;
309*4882a593Smuzhiyun
310*4882a593Smuzhiyun __i915_sw_fence_clear_checked_bit(wq->private);
311*4882a593Smuzhiyun }
312*4882a593Smuzhiyun }
313*4882a593Smuzhiyun
i915_sw_fence_check_if_after(struct i915_sw_fence * fence,const struct i915_sw_fence * const signaler)314*4882a593Smuzhiyun static bool i915_sw_fence_check_if_after(struct i915_sw_fence *fence,
315*4882a593Smuzhiyun const struct i915_sw_fence * const signaler)
316*4882a593Smuzhiyun {
317*4882a593Smuzhiyun unsigned long flags;
318*4882a593Smuzhiyun bool err;
319*4882a593Smuzhiyun
320*4882a593Smuzhiyun if (!IS_ENABLED(CONFIG_DRM_I915_SW_FENCE_CHECK_DAG))
321*4882a593Smuzhiyun return false;
322*4882a593Smuzhiyun
323*4882a593Smuzhiyun spin_lock_irqsave(&i915_sw_fence_lock, flags);
324*4882a593Smuzhiyun err = __i915_sw_fence_check_if_after(fence, signaler);
325*4882a593Smuzhiyun __i915_sw_fence_clear_checked_bit(fence);
326*4882a593Smuzhiyun spin_unlock_irqrestore(&i915_sw_fence_lock, flags);
327*4882a593Smuzhiyun
328*4882a593Smuzhiyun return err;
329*4882a593Smuzhiyun }
330*4882a593Smuzhiyun
__i915_sw_fence_await_sw_fence(struct i915_sw_fence * fence,struct i915_sw_fence * signaler,wait_queue_entry_t * wq,gfp_t gfp)331*4882a593Smuzhiyun static int __i915_sw_fence_await_sw_fence(struct i915_sw_fence *fence,
332*4882a593Smuzhiyun struct i915_sw_fence *signaler,
333*4882a593Smuzhiyun wait_queue_entry_t *wq, gfp_t gfp)
334*4882a593Smuzhiyun {
335*4882a593Smuzhiyun unsigned long flags;
336*4882a593Smuzhiyun int pending;
337*4882a593Smuzhiyun
338*4882a593Smuzhiyun debug_fence_assert(fence);
339*4882a593Smuzhiyun might_sleep_if(gfpflags_allow_blocking(gfp));
340*4882a593Smuzhiyun
341*4882a593Smuzhiyun if (i915_sw_fence_done(signaler)) {
342*4882a593Smuzhiyun i915_sw_fence_set_error_once(fence, signaler->error);
343*4882a593Smuzhiyun return 0;
344*4882a593Smuzhiyun }
345*4882a593Smuzhiyun
346*4882a593Smuzhiyun debug_fence_assert(signaler);
347*4882a593Smuzhiyun
348*4882a593Smuzhiyun /* The dependency graph must be acyclic. */
349*4882a593Smuzhiyun if (unlikely(i915_sw_fence_check_if_after(fence, signaler)))
350*4882a593Smuzhiyun return -EINVAL;
351*4882a593Smuzhiyun
352*4882a593Smuzhiyun pending = 0;
353*4882a593Smuzhiyun if (!wq) {
354*4882a593Smuzhiyun wq = kmalloc(sizeof(*wq), gfp);
355*4882a593Smuzhiyun if (!wq) {
356*4882a593Smuzhiyun if (!gfpflags_allow_blocking(gfp))
357*4882a593Smuzhiyun return -ENOMEM;
358*4882a593Smuzhiyun
359*4882a593Smuzhiyun i915_sw_fence_wait(signaler);
360*4882a593Smuzhiyun i915_sw_fence_set_error_once(fence, signaler->error);
361*4882a593Smuzhiyun return 0;
362*4882a593Smuzhiyun }
363*4882a593Smuzhiyun
364*4882a593Smuzhiyun pending |= I915_SW_FENCE_FLAG_ALLOC;
365*4882a593Smuzhiyun }
366*4882a593Smuzhiyun
367*4882a593Smuzhiyun INIT_LIST_HEAD(&wq->entry);
368*4882a593Smuzhiyun wq->flags = pending;
369*4882a593Smuzhiyun wq->func = i915_sw_fence_wake;
370*4882a593Smuzhiyun wq->private = fence;
371*4882a593Smuzhiyun
372*4882a593Smuzhiyun i915_sw_fence_await(fence);
373*4882a593Smuzhiyun
374*4882a593Smuzhiyun spin_lock_irqsave(&signaler->wait.lock, flags);
375*4882a593Smuzhiyun if (likely(!i915_sw_fence_done(signaler))) {
376*4882a593Smuzhiyun __add_wait_queue_entry_tail(&signaler->wait, wq);
377*4882a593Smuzhiyun pending = 1;
378*4882a593Smuzhiyun } else {
379*4882a593Smuzhiyun i915_sw_fence_wake(wq, 0, signaler->error, NULL);
380*4882a593Smuzhiyun pending = 0;
381*4882a593Smuzhiyun }
382*4882a593Smuzhiyun spin_unlock_irqrestore(&signaler->wait.lock, flags);
383*4882a593Smuzhiyun
384*4882a593Smuzhiyun return pending;
385*4882a593Smuzhiyun }
386*4882a593Smuzhiyun
i915_sw_fence_await_sw_fence(struct i915_sw_fence * fence,struct i915_sw_fence * signaler,wait_queue_entry_t * wq)387*4882a593Smuzhiyun int i915_sw_fence_await_sw_fence(struct i915_sw_fence *fence,
388*4882a593Smuzhiyun struct i915_sw_fence *signaler,
389*4882a593Smuzhiyun wait_queue_entry_t *wq)
390*4882a593Smuzhiyun {
391*4882a593Smuzhiyun return __i915_sw_fence_await_sw_fence(fence, signaler, wq, 0);
392*4882a593Smuzhiyun }
393*4882a593Smuzhiyun
i915_sw_fence_await_sw_fence_gfp(struct i915_sw_fence * fence,struct i915_sw_fence * signaler,gfp_t gfp)394*4882a593Smuzhiyun int i915_sw_fence_await_sw_fence_gfp(struct i915_sw_fence *fence,
395*4882a593Smuzhiyun struct i915_sw_fence *signaler,
396*4882a593Smuzhiyun gfp_t gfp)
397*4882a593Smuzhiyun {
398*4882a593Smuzhiyun return __i915_sw_fence_await_sw_fence(fence, signaler, NULL, gfp);
399*4882a593Smuzhiyun }
400*4882a593Smuzhiyun
401*4882a593Smuzhiyun struct i915_sw_dma_fence_cb_timer {
402*4882a593Smuzhiyun struct i915_sw_dma_fence_cb base;
403*4882a593Smuzhiyun struct dma_fence *dma;
404*4882a593Smuzhiyun struct timer_list timer;
405*4882a593Smuzhiyun struct irq_work work;
406*4882a593Smuzhiyun struct rcu_head rcu;
407*4882a593Smuzhiyun };
408*4882a593Smuzhiyun
dma_i915_sw_fence_wake(struct dma_fence * dma,struct dma_fence_cb * data)409*4882a593Smuzhiyun static void dma_i915_sw_fence_wake(struct dma_fence *dma,
410*4882a593Smuzhiyun struct dma_fence_cb *data)
411*4882a593Smuzhiyun {
412*4882a593Smuzhiyun struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base);
413*4882a593Smuzhiyun
414*4882a593Smuzhiyun i915_sw_fence_set_error_once(cb->fence, dma->error);
415*4882a593Smuzhiyun i915_sw_fence_complete(cb->fence);
416*4882a593Smuzhiyun kfree(cb);
417*4882a593Smuzhiyun }
418*4882a593Smuzhiyun
timer_i915_sw_fence_wake(struct timer_list * t)419*4882a593Smuzhiyun static void timer_i915_sw_fence_wake(struct timer_list *t)
420*4882a593Smuzhiyun {
421*4882a593Smuzhiyun struct i915_sw_dma_fence_cb_timer *cb = from_timer(cb, t, timer);
422*4882a593Smuzhiyun struct i915_sw_fence *fence;
423*4882a593Smuzhiyun
424*4882a593Smuzhiyun fence = xchg(&cb->base.fence, NULL);
425*4882a593Smuzhiyun if (!fence)
426*4882a593Smuzhiyun return;
427*4882a593Smuzhiyun
428*4882a593Smuzhiyun pr_notice("Asynchronous wait on fence %s:%s:%llx timed out (hint:%ps)\n",
429*4882a593Smuzhiyun cb->dma->ops->get_driver_name(cb->dma),
430*4882a593Smuzhiyun cb->dma->ops->get_timeline_name(cb->dma),
431*4882a593Smuzhiyun cb->dma->seqno,
432*4882a593Smuzhiyun i915_sw_fence_debug_hint(fence));
433*4882a593Smuzhiyun
434*4882a593Smuzhiyun i915_sw_fence_set_error_once(fence, -ETIMEDOUT);
435*4882a593Smuzhiyun i915_sw_fence_complete(fence);
436*4882a593Smuzhiyun }
437*4882a593Smuzhiyun
dma_i915_sw_fence_wake_timer(struct dma_fence * dma,struct dma_fence_cb * data)438*4882a593Smuzhiyun static void dma_i915_sw_fence_wake_timer(struct dma_fence *dma,
439*4882a593Smuzhiyun struct dma_fence_cb *data)
440*4882a593Smuzhiyun {
441*4882a593Smuzhiyun struct i915_sw_dma_fence_cb_timer *cb =
442*4882a593Smuzhiyun container_of(data, typeof(*cb), base.base);
443*4882a593Smuzhiyun struct i915_sw_fence *fence;
444*4882a593Smuzhiyun
445*4882a593Smuzhiyun fence = xchg(&cb->base.fence, NULL);
446*4882a593Smuzhiyun if (fence) {
447*4882a593Smuzhiyun i915_sw_fence_set_error_once(fence, dma->error);
448*4882a593Smuzhiyun i915_sw_fence_complete(fence);
449*4882a593Smuzhiyun }
450*4882a593Smuzhiyun
451*4882a593Smuzhiyun irq_work_queue(&cb->work);
452*4882a593Smuzhiyun }
453*4882a593Smuzhiyun
irq_i915_sw_fence_work(struct irq_work * wrk)454*4882a593Smuzhiyun static void irq_i915_sw_fence_work(struct irq_work *wrk)
455*4882a593Smuzhiyun {
456*4882a593Smuzhiyun struct i915_sw_dma_fence_cb_timer *cb =
457*4882a593Smuzhiyun container_of(wrk, typeof(*cb), work);
458*4882a593Smuzhiyun
459*4882a593Smuzhiyun del_timer_sync(&cb->timer);
460*4882a593Smuzhiyun dma_fence_put(cb->dma);
461*4882a593Smuzhiyun
462*4882a593Smuzhiyun kfree_rcu(cb, rcu);
463*4882a593Smuzhiyun }
464*4882a593Smuzhiyun
i915_sw_fence_await_dma_fence(struct i915_sw_fence * fence,struct dma_fence * dma,unsigned long timeout,gfp_t gfp)465*4882a593Smuzhiyun int i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence,
466*4882a593Smuzhiyun struct dma_fence *dma,
467*4882a593Smuzhiyun unsigned long timeout,
468*4882a593Smuzhiyun gfp_t gfp)
469*4882a593Smuzhiyun {
470*4882a593Smuzhiyun struct i915_sw_dma_fence_cb *cb;
471*4882a593Smuzhiyun dma_fence_func_t func;
472*4882a593Smuzhiyun int ret;
473*4882a593Smuzhiyun
474*4882a593Smuzhiyun debug_fence_assert(fence);
475*4882a593Smuzhiyun might_sleep_if(gfpflags_allow_blocking(gfp));
476*4882a593Smuzhiyun
477*4882a593Smuzhiyun if (dma_fence_is_signaled(dma)) {
478*4882a593Smuzhiyun i915_sw_fence_set_error_once(fence, dma->error);
479*4882a593Smuzhiyun return 0;
480*4882a593Smuzhiyun }
481*4882a593Smuzhiyun
482*4882a593Smuzhiyun cb = kmalloc(timeout ?
483*4882a593Smuzhiyun sizeof(struct i915_sw_dma_fence_cb_timer) :
484*4882a593Smuzhiyun sizeof(struct i915_sw_dma_fence_cb),
485*4882a593Smuzhiyun gfp);
486*4882a593Smuzhiyun if (!cb) {
487*4882a593Smuzhiyun if (!gfpflags_allow_blocking(gfp))
488*4882a593Smuzhiyun return -ENOMEM;
489*4882a593Smuzhiyun
490*4882a593Smuzhiyun ret = dma_fence_wait(dma, false);
491*4882a593Smuzhiyun if (ret)
492*4882a593Smuzhiyun return ret;
493*4882a593Smuzhiyun
494*4882a593Smuzhiyun i915_sw_fence_set_error_once(fence, dma->error);
495*4882a593Smuzhiyun return 0;
496*4882a593Smuzhiyun }
497*4882a593Smuzhiyun
498*4882a593Smuzhiyun cb->fence = fence;
499*4882a593Smuzhiyun i915_sw_fence_await(fence);
500*4882a593Smuzhiyun
501*4882a593Smuzhiyun func = dma_i915_sw_fence_wake;
502*4882a593Smuzhiyun if (timeout) {
503*4882a593Smuzhiyun struct i915_sw_dma_fence_cb_timer *timer =
504*4882a593Smuzhiyun container_of(cb, typeof(*timer), base);
505*4882a593Smuzhiyun
506*4882a593Smuzhiyun timer->dma = dma_fence_get(dma);
507*4882a593Smuzhiyun init_irq_work(&timer->work, irq_i915_sw_fence_work);
508*4882a593Smuzhiyun
509*4882a593Smuzhiyun timer_setup(&timer->timer,
510*4882a593Smuzhiyun timer_i915_sw_fence_wake, TIMER_IRQSAFE);
511*4882a593Smuzhiyun mod_timer(&timer->timer, round_jiffies_up(jiffies + timeout));
512*4882a593Smuzhiyun
513*4882a593Smuzhiyun func = dma_i915_sw_fence_wake_timer;
514*4882a593Smuzhiyun }
515*4882a593Smuzhiyun
516*4882a593Smuzhiyun ret = dma_fence_add_callback(dma, &cb->base, func);
517*4882a593Smuzhiyun if (ret == 0) {
518*4882a593Smuzhiyun ret = 1;
519*4882a593Smuzhiyun } else {
520*4882a593Smuzhiyun func(dma, &cb->base);
521*4882a593Smuzhiyun if (ret == -ENOENT) /* fence already signaled */
522*4882a593Smuzhiyun ret = 0;
523*4882a593Smuzhiyun }
524*4882a593Smuzhiyun
525*4882a593Smuzhiyun return ret;
526*4882a593Smuzhiyun }
527*4882a593Smuzhiyun
__dma_i915_sw_fence_wake(struct dma_fence * dma,struct dma_fence_cb * data)528*4882a593Smuzhiyun static void __dma_i915_sw_fence_wake(struct dma_fence *dma,
529*4882a593Smuzhiyun struct dma_fence_cb *data)
530*4882a593Smuzhiyun {
531*4882a593Smuzhiyun struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base);
532*4882a593Smuzhiyun
533*4882a593Smuzhiyun i915_sw_fence_set_error_once(cb->fence, dma->error);
534*4882a593Smuzhiyun i915_sw_fence_complete(cb->fence);
535*4882a593Smuzhiyun }
536*4882a593Smuzhiyun
__i915_sw_fence_await_dma_fence(struct i915_sw_fence * fence,struct dma_fence * dma,struct i915_sw_dma_fence_cb * cb)537*4882a593Smuzhiyun int __i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence,
538*4882a593Smuzhiyun struct dma_fence *dma,
539*4882a593Smuzhiyun struct i915_sw_dma_fence_cb *cb)
540*4882a593Smuzhiyun {
541*4882a593Smuzhiyun int ret;
542*4882a593Smuzhiyun
543*4882a593Smuzhiyun debug_fence_assert(fence);
544*4882a593Smuzhiyun
545*4882a593Smuzhiyun if (dma_fence_is_signaled(dma)) {
546*4882a593Smuzhiyun i915_sw_fence_set_error_once(fence, dma->error);
547*4882a593Smuzhiyun return 0;
548*4882a593Smuzhiyun }
549*4882a593Smuzhiyun
550*4882a593Smuzhiyun cb->fence = fence;
551*4882a593Smuzhiyun i915_sw_fence_await(fence);
552*4882a593Smuzhiyun
553*4882a593Smuzhiyun ret = 1;
554*4882a593Smuzhiyun if (dma_fence_add_callback(dma, &cb->base, __dma_i915_sw_fence_wake)) {
555*4882a593Smuzhiyun /* fence already signaled */
556*4882a593Smuzhiyun __dma_i915_sw_fence_wake(dma, &cb->base);
557*4882a593Smuzhiyun ret = 0;
558*4882a593Smuzhiyun }
559*4882a593Smuzhiyun
560*4882a593Smuzhiyun return ret;
561*4882a593Smuzhiyun }
562*4882a593Smuzhiyun
i915_sw_fence_await_reservation(struct i915_sw_fence * fence,struct dma_resv * resv,const struct dma_fence_ops * exclude,bool write,unsigned long timeout,gfp_t gfp)563*4882a593Smuzhiyun int i915_sw_fence_await_reservation(struct i915_sw_fence *fence,
564*4882a593Smuzhiyun struct dma_resv *resv,
565*4882a593Smuzhiyun const struct dma_fence_ops *exclude,
566*4882a593Smuzhiyun bool write,
567*4882a593Smuzhiyun unsigned long timeout,
568*4882a593Smuzhiyun gfp_t gfp)
569*4882a593Smuzhiyun {
570*4882a593Smuzhiyun struct dma_fence *excl;
571*4882a593Smuzhiyun int ret = 0, pending;
572*4882a593Smuzhiyun
573*4882a593Smuzhiyun debug_fence_assert(fence);
574*4882a593Smuzhiyun might_sleep_if(gfpflags_allow_blocking(gfp));
575*4882a593Smuzhiyun
576*4882a593Smuzhiyun if (write) {
577*4882a593Smuzhiyun struct dma_fence **shared;
578*4882a593Smuzhiyun unsigned int count, i;
579*4882a593Smuzhiyun
580*4882a593Smuzhiyun ret = dma_resv_get_fences_rcu(resv, &excl, &count, &shared);
581*4882a593Smuzhiyun if (ret)
582*4882a593Smuzhiyun return ret;
583*4882a593Smuzhiyun
584*4882a593Smuzhiyun for (i = 0; i < count; i++) {
585*4882a593Smuzhiyun if (shared[i]->ops == exclude)
586*4882a593Smuzhiyun continue;
587*4882a593Smuzhiyun
588*4882a593Smuzhiyun pending = i915_sw_fence_await_dma_fence(fence,
589*4882a593Smuzhiyun shared[i],
590*4882a593Smuzhiyun timeout,
591*4882a593Smuzhiyun gfp);
592*4882a593Smuzhiyun if (pending < 0) {
593*4882a593Smuzhiyun ret = pending;
594*4882a593Smuzhiyun break;
595*4882a593Smuzhiyun }
596*4882a593Smuzhiyun
597*4882a593Smuzhiyun ret |= pending;
598*4882a593Smuzhiyun }
599*4882a593Smuzhiyun
600*4882a593Smuzhiyun for (i = 0; i < count; i++)
601*4882a593Smuzhiyun dma_fence_put(shared[i]);
602*4882a593Smuzhiyun kfree(shared);
603*4882a593Smuzhiyun } else {
604*4882a593Smuzhiyun excl = dma_resv_get_excl_rcu(resv);
605*4882a593Smuzhiyun }
606*4882a593Smuzhiyun
607*4882a593Smuzhiyun if (ret >= 0 && excl && excl->ops != exclude) {
608*4882a593Smuzhiyun pending = i915_sw_fence_await_dma_fence(fence,
609*4882a593Smuzhiyun excl,
610*4882a593Smuzhiyun timeout,
611*4882a593Smuzhiyun gfp);
612*4882a593Smuzhiyun if (pending < 0)
613*4882a593Smuzhiyun ret = pending;
614*4882a593Smuzhiyun else
615*4882a593Smuzhiyun ret |= pending;
616*4882a593Smuzhiyun }
617*4882a593Smuzhiyun
618*4882a593Smuzhiyun dma_fence_put(excl);
619*4882a593Smuzhiyun
620*4882a593Smuzhiyun return ret;
621*4882a593Smuzhiyun }
622*4882a593Smuzhiyun
623*4882a593Smuzhiyun #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST)
624*4882a593Smuzhiyun #include "selftests/lib_sw_fence.c"
625*4882a593Smuzhiyun #include "selftests/i915_sw_fence.c"
626*4882a593Smuzhiyun #endif
627