1*4882a593Smuzhiyun // SPDX-License-Identifier: MIT
2*4882a593Smuzhiyun
3*4882a593Smuzhiyun #include <uapi/linux/sched/types.h>
4*4882a593Smuzhiyun
5*4882a593Smuzhiyun #include <drm/drm_print.h>
6*4882a593Smuzhiyun #include <drm/drm_vblank.h>
7*4882a593Smuzhiyun #include <drm/drm_vblank_work.h>
8*4882a593Smuzhiyun #include <drm/drm_crtc.h>
9*4882a593Smuzhiyun
10*4882a593Smuzhiyun #include "drm_internal.h"
11*4882a593Smuzhiyun
12*4882a593Smuzhiyun /**
13*4882a593Smuzhiyun * DOC: vblank works
14*4882a593Smuzhiyun *
15*4882a593Smuzhiyun * Many DRM drivers need to program hardware in a time-sensitive manner, many
16*4882a593Smuzhiyun * times with a deadline of starting and finishing within a certain region of
17*4882a593Smuzhiyun * the scanout. Most of the time the safest way to accomplish this is to
18*4882a593Smuzhiyun * simply do said time-sensitive programming in the driver's IRQ handler,
19*4882a593Smuzhiyun * which allows drivers to avoid being preempted during these critical
20*4882a593Smuzhiyun * regions. Or even better, the hardware may even handle applying such
21*4882a593Smuzhiyun * time-critical programming independently of the CPU.
22*4882a593Smuzhiyun *
23*4882a593Smuzhiyun * While there's a decent amount of hardware that's designed so that the CPU
24*4882a593Smuzhiyun * doesn't need to be concerned with extremely time-sensitive programming,
25*4882a593Smuzhiyun * there's a few situations where it can't be helped. Some unforgiving
26*4882a593Smuzhiyun * hardware may require that certain time-sensitive programming be handled
27*4882a593Smuzhiyun * completely by the CPU, and said programming may even take too long to
28*4882a593Smuzhiyun * handle in an IRQ handler. Another such situation would be where the driver
29*4882a593Smuzhiyun * needs to perform a task that needs to complete within a specific scanout
30*4882a593Smuzhiyun * period, but might possibly block and thus cannot be handled in an IRQ
31*4882a593Smuzhiyun * context. Both of these situations can't be solved perfectly in Linux since
32*4882a593Smuzhiyun * we're not a realtime kernel, and thus the scheduler may cause us to miss
33*4882a593Smuzhiyun * our deadline if it decides to preempt us. But for some drivers, it's good
34*4882a593Smuzhiyun * enough if we can lower our chance of being preempted to an absolute
35*4882a593Smuzhiyun * minimum.
36*4882a593Smuzhiyun *
37*4882a593Smuzhiyun * This is where &drm_vblank_work comes in. &drm_vblank_work provides a simple
38*4882a593Smuzhiyun * generic delayed work implementation which delays work execution until a
39*4882a593Smuzhiyun * particular vblank has passed, and then executes the work at realtime
40*4882a593Smuzhiyun * priority. This provides the best possible chance at performing
41*4882a593Smuzhiyun * time-sensitive hardware programming on time, even when the system is under
42*4882a593Smuzhiyun * heavy load. &drm_vblank_work also supports rescheduling, so that self
43*4882a593Smuzhiyun * re-arming work items can be easily implemented.
44*4882a593Smuzhiyun */
45*4882a593Smuzhiyun
drm_handle_vblank_works(struct drm_vblank_crtc * vblank)46*4882a593Smuzhiyun void drm_handle_vblank_works(struct drm_vblank_crtc *vblank)
47*4882a593Smuzhiyun {
48*4882a593Smuzhiyun struct drm_vblank_work *work, *next;
49*4882a593Smuzhiyun u64 count = atomic64_read(&vblank->count);
50*4882a593Smuzhiyun bool wake = false;
51*4882a593Smuzhiyun
52*4882a593Smuzhiyun assert_spin_locked(&vblank->dev->event_lock);
53*4882a593Smuzhiyun
54*4882a593Smuzhiyun list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
55*4882a593Smuzhiyun if (!drm_vblank_passed(count, work->count))
56*4882a593Smuzhiyun continue;
57*4882a593Smuzhiyun
58*4882a593Smuzhiyun list_del_init(&work->node);
59*4882a593Smuzhiyun drm_vblank_put(vblank->dev, vblank->pipe);
60*4882a593Smuzhiyun kthread_queue_work(vblank->worker, &work->base);
61*4882a593Smuzhiyun wake = true;
62*4882a593Smuzhiyun }
63*4882a593Smuzhiyun if (wake)
64*4882a593Smuzhiyun wake_up_all(&vblank->work_wait_queue);
65*4882a593Smuzhiyun }
66*4882a593Smuzhiyun
67*4882a593Smuzhiyun /* Handle cancelling any pending vblank work items and drop respective vblank
68*4882a593Smuzhiyun * references in response to vblank interrupts being disabled.
69*4882a593Smuzhiyun */
drm_vblank_cancel_pending_works(struct drm_vblank_crtc * vblank)70*4882a593Smuzhiyun void drm_vblank_cancel_pending_works(struct drm_vblank_crtc *vblank)
71*4882a593Smuzhiyun {
72*4882a593Smuzhiyun struct drm_vblank_work *work, *next;
73*4882a593Smuzhiyun
74*4882a593Smuzhiyun assert_spin_locked(&vblank->dev->event_lock);
75*4882a593Smuzhiyun
76*4882a593Smuzhiyun list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
77*4882a593Smuzhiyun list_del_init(&work->node);
78*4882a593Smuzhiyun drm_vblank_put(vblank->dev, vblank->pipe);
79*4882a593Smuzhiyun }
80*4882a593Smuzhiyun
81*4882a593Smuzhiyun wake_up_all(&vblank->work_wait_queue);
82*4882a593Smuzhiyun }
83*4882a593Smuzhiyun
84*4882a593Smuzhiyun /**
85*4882a593Smuzhiyun * drm_vblank_work_schedule - schedule a vblank work
86*4882a593Smuzhiyun * @work: vblank work to schedule
87*4882a593Smuzhiyun * @count: target vblank count
88*4882a593Smuzhiyun * @nextonmiss: defer until the next vblank if target vblank was missed
89*4882a593Smuzhiyun *
90*4882a593Smuzhiyun * Schedule @work for execution once the crtc vblank count reaches @count.
91*4882a593Smuzhiyun *
92*4882a593Smuzhiyun * If the crtc vblank count has already reached @count and @nextonmiss is
93*4882a593Smuzhiyun * %false the work starts to execute immediately.
94*4882a593Smuzhiyun *
95*4882a593Smuzhiyun * If the crtc vblank count has already reached @count and @nextonmiss is
96*4882a593Smuzhiyun * %true the work is deferred until the next vblank (as if @count has been
97*4882a593Smuzhiyun * specified as crtc vblank count + 1).
98*4882a593Smuzhiyun *
99*4882a593Smuzhiyun * If @work is already scheduled, this function will reschedule said work
100*4882a593Smuzhiyun * using the new @count. This can be used for self-rearming work items.
101*4882a593Smuzhiyun *
102*4882a593Smuzhiyun * Returns:
103*4882a593Smuzhiyun * %1 if @work was successfully (re)scheduled, %0 if it was either already
104*4882a593Smuzhiyun * scheduled or cancelled, or a negative error code on failure.
105*4882a593Smuzhiyun */
drm_vblank_work_schedule(struct drm_vblank_work * work,u64 count,bool nextonmiss)106*4882a593Smuzhiyun int drm_vblank_work_schedule(struct drm_vblank_work *work,
107*4882a593Smuzhiyun u64 count, bool nextonmiss)
108*4882a593Smuzhiyun {
109*4882a593Smuzhiyun struct drm_vblank_crtc *vblank = work->vblank;
110*4882a593Smuzhiyun struct drm_device *dev = vblank->dev;
111*4882a593Smuzhiyun u64 cur_vbl;
112*4882a593Smuzhiyun unsigned long irqflags;
113*4882a593Smuzhiyun bool passed, inmodeset, rescheduling = false, wake = false;
114*4882a593Smuzhiyun int ret = 0;
115*4882a593Smuzhiyun
116*4882a593Smuzhiyun spin_lock_irqsave(&dev->event_lock, irqflags);
117*4882a593Smuzhiyun if (work->cancelling)
118*4882a593Smuzhiyun goto out;
119*4882a593Smuzhiyun
120*4882a593Smuzhiyun spin_lock(&dev->vbl_lock);
121*4882a593Smuzhiyun inmodeset = vblank->inmodeset;
122*4882a593Smuzhiyun spin_unlock(&dev->vbl_lock);
123*4882a593Smuzhiyun if (inmodeset)
124*4882a593Smuzhiyun goto out;
125*4882a593Smuzhiyun
126*4882a593Smuzhiyun if (list_empty(&work->node)) {
127*4882a593Smuzhiyun ret = drm_vblank_get(dev, vblank->pipe);
128*4882a593Smuzhiyun if (ret < 0)
129*4882a593Smuzhiyun goto out;
130*4882a593Smuzhiyun } else if (work->count == count) {
131*4882a593Smuzhiyun /* Already scheduled w/ same vbl count */
132*4882a593Smuzhiyun goto out;
133*4882a593Smuzhiyun } else {
134*4882a593Smuzhiyun rescheduling = true;
135*4882a593Smuzhiyun }
136*4882a593Smuzhiyun
137*4882a593Smuzhiyun work->count = count;
138*4882a593Smuzhiyun cur_vbl = drm_vblank_count(dev, vblank->pipe);
139*4882a593Smuzhiyun passed = drm_vblank_passed(cur_vbl, count);
140*4882a593Smuzhiyun if (passed)
141*4882a593Smuzhiyun drm_dbg_core(dev,
142*4882a593Smuzhiyun "crtc %d vblank %llu already passed (current %llu)\n",
143*4882a593Smuzhiyun vblank->pipe, count, cur_vbl);
144*4882a593Smuzhiyun
145*4882a593Smuzhiyun if (!nextonmiss && passed) {
146*4882a593Smuzhiyun drm_vblank_put(dev, vblank->pipe);
147*4882a593Smuzhiyun ret = kthread_queue_work(vblank->worker, &work->base);
148*4882a593Smuzhiyun
149*4882a593Smuzhiyun if (rescheduling) {
150*4882a593Smuzhiyun list_del_init(&work->node);
151*4882a593Smuzhiyun wake = true;
152*4882a593Smuzhiyun }
153*4882a593Smuzhiyun } else {
154*4882a593Smuzhiyun if (!rescheduling)
155*4882a593Smuzhiyun list_add_tail(&work->node, &vblank->pending_work);
156*4882a593Smuzhiyun ret = true;
157*4882a593Smuzhiyun }
158*4882a593Smuzhiyun
159*4882a593Smuzhiyun out:
160*4882a593Smuzhiyun spin_unlock_irqrestore(&dev->event_lock, irqflags);
161*4882a593Smuzhiyun if (wake)
162*4882a593Smuzhiyun wake_up_all(&vblank->work_wait_queue);
163*4882a593Smuzhiyun return ret;
164*4882a593Smuzhiyun }
165*4882a593Smuzhiyun EXPORT_SYMBOL(drm_vblank_work_schedule);
166*4882a593Smuzhiyun
167*4882a593Smuzhiyun /**
168*4882a593Smuzhiyun * drm_vblank_work_cancel_sync - cancel a vblank work and wait for it to
169*4882a593Smuzhiyun * finish executing
170*4882a593Smuzhiyun * @work: vblank work to cancel
171*4882a593Smuzhiyun *
172*4882a593Smuzhiyun * Cancel an already scheduled vblank work and wait for its
173*4882a593Smuzhiyun * execution to finish.
174*4882a593Smuzhiyun *
175*4882a593Smuzhiyun * On return, @work is guaranteed to no longer be scheduled or running, even
176*4882a593Smuzhiyun * if it's self-arming.
177*4882a593Smuzhiyun *
178*4882a593Smuzhiyun * Returns:
179*4882a593Smuzhiyun * %True if the work was cancelled before it started to execute, %false
180*4882a593Smuzhiyun * otherwise.
181*4882a593Smuzhiyun */
drm_vblank_work_cancel_sync(struct drm_vblank_work * work)182*4882a593Smuzhiyun bool drm_vblank_work_cancel_sync(struct drm_vblank_work *work)
183*4882a593Smuzhiyun {
184*4882a593Smuzhiyun struct drm_vblank_crtc *vblank = work->vblank;
185*4882a593Smuzhiyun struct drm_device *dev = vblank->dev;
186*4882a593Smuzhiyun bool ret = false;
187*4882a593Smuzhiyun
188*4882a593Smuzhiyun spin_lock_irq(&dev->event_lock);
189*4882a593Smuzhiyun if (!list_empty(&work->node)) {
190*4882a593Smuzhiyun list_del_init(&work->node);
191*4882a593Smuzhiyun drm_vblank_put(vblank->dev, vblank->pipe);
192*4882a593Smuzhiyun ret = true;
193*4882a593Smuzhiyun }
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun work->cancelling++;
196*4882a593Smuzhiyun spin_unlock_irq(&dev->event_lock);
197*4882a593Smuzhiyun
198*4882a593Smuzhiyun wake_up_all(&vblank->work_wait_queue);
199*4882a593Smuzhiyun
200*4882a593Smuzhiyun if (kthread_cancel_work_sync(&work->base))
201*4882a593Smuzhiyun ret = true;
202*4882a593Smuzhiyun
203*4882a593Smuzhiyun spin_lock_irq(&dev->event_lock);
204*4882a593Smuzhiyun work->cancelling--;
205*4882a593Smuzhiyun spin_unlock_irq(&dev->event_lock);
206*4882a593Smuzhiyun
207*4882a593Smuzhiyun return ret;
208*4882a593Smuzhiyun }
209*4882a593Smuzhiyun EXPORT_SYMBOL(drm_vblank_work_cancel_sync);
210*4882a593Smuzhiyun
211*4882a593Smuzhiyun /**
212*4882a593Smuzhiyun * drm_vblank_work_flush - wait for a scheduled vblank work to finish
213*4882a593Smuzhiyun * executing
214*4882a593Smuzhiyun * @work: vblank work to flush
215*4882a593Smuzhiyun *
216*4882a593Smuzhiyun * Wait until @work has finished executing once.
217*4882a593Smuzhiyun */
drm_vblank_work_flush(struct drm_vblank_work * work)218*4882a593Smuzhiyun void drm_vblank_work_flush(struct drm_vblank_work *work)
219*4882a593Smuzhiyun {
220*4882a593Smuzhiyun struct drm_vblank_crtc *vblank = work->vblank;
221*4882a593Smuzhiyun struct drm_device *dev = vblank->dev;
222*4882a593Smuzhiyun
223*4882a593Smuzhiyun spin_lock_irq(&dev->event_lock);
224*4882a593Smuzhiyun wait_event_lock_irq(vblank->work_wait_queue, list_empty(&work->node),
225*4882a593Smuzhiyun dev->event_lock);
226*4882a593Smuzhiyun spin_unlock_irq(&dev->event_lock);
227*4882a593Smuzhiyun
228*4882a593Smuzhiyun kthread_flush_work(&work->base);
229*4882a593Smuzhiyun }
230*4882a593Smuzhiyun EXPORT_SYMBOL(drm_vblank_work_flush);
231*4882a593Smuzhiyun
232*4882a593Smuzhiyun /**
233*4882a593Smuzhiyun * drm_vblank_work_init - initialize a vblank work item
234*4882a593Smuzhiyun * @work: vblank work item
235*4882a593Smuzhiyun * @crtc: CRTC whose vblank will trigger the work execution
236*4882a593Smuzhiyun * @func: work function to be executed
237*4882a593Smuzhiyun *
238*4882a593Smuzhiyun * Initialize a vblank work item for a specific crtc.
239*4882a593Smuzhiyun */
drm_vblank_work_init(struct drm_vblank_work * work,struct drm_crtc * crtc,void (* func)(struct kthread_work * work))240*4882a593Smuzhiyun void drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc,
241*4882a593Smuzhiyun void (*func)(struct kthread_work *work))
242*4882a593Smuzhiyun {
243*4882a593Smuzhiyun kthread_init_work(&work->base, func);
244*4882a593Smuzhiyun INIT_LIST_HEAD(&work->node);
245*4882a593Smuzhiyun work->vblank = &crtc->dev->vblank[drm_crtc_index(crtc)];
246*4882a593Smuzhiyun }
247*4882a593Smuzhiyun EXPORT_SYMBOL(drm_vblank_work_init);
248*4882a593Smuzhiyun
drm_vblank_worker_init(struct drm_vblank_crtc * vblank)249*4882a593Smuzhiyun int drm_vblank_worker_init(struct drm_vblank_crtc *vblank)
250*4882a593Smuzhiyun {
251*4882a593Smuzhiyun struct kthread_worker *worker;
252*4882a593Smuzhiyun
253*4882a593Smuzhiyun INIT_LIST_HEAD(&vblank->pending_work);
254*4882a593Smuzhiyun init_waitqueue_head(&vblank->work_wait_queue);
255*4882a593Smuzhiyun worker = kthread_create_worker(0, "card%d-crtc%d",
256*4882a593Smuzhiyun vblank->dev->primary->index,
257*4882a593Smuzhiyun vblank->pipe);
258*4882a593Smuzhiyun if (IS_ERR(worker))
259*4882a593Smuzhiyun return PTR_ERR(worker);
260*4882a593Smuzhiyun
261*4882a593Smuzhiyun vblank->worker = worker;
262*4882a593Smuzhiyun
263*4882a593Smuzhiyun sched_set_fifo(worker->task);
264*4882a593Smuzhiyun return 0;
265*4882a593Smuzhiyun }
266