linux/drivers/gpu/drm/drm_vblank_work.c
<<
>>
Prefs
   1// SPDX-License-Identifier: MIT
   2
   3#include <uapi/linux/sched/types.h>
   4
   5#include <drm/drm_print.h>
   6#include <drm/drm_vblank.h>
   7#include <drm/drm_vblank_work.h>
   8#include <drm/drm_crtc.h>
   9
  10#include "drm_internal.h"
  11
  12/**
  13 * DOC: vblank works
  14 *
  15 * Many DRM drivers need to program hardware in a time-sensitive manner, many
  16 * times with a deadline of starting and finishing within a certain region of
  17 * the scanout. Most of the time the safest way to accomplish this is to
  18 * simply do said time-sensitive programming in the driver's IRQ handler,
  19 * which allows drivers to avoid being preempted during these critical
  20 * regions. Or even better, the hardware may even handle applying such
  21 * time-critical programming independently of the CPU.
  22 *
  23 * While there's a decent amount of hardware that's designed so that the CPU
  24 * doesn't need to be concerned with extremely time-sensitive programming,
  25 * there's a few situations where it can't be helped. Some unforgiving
  26 * hardware may require that certain time-sensitive programming be handled
  27 * completely by the CPU, and said programming may even take too long to
  28 * handle in an IRQ handler. Another such situation would be where the driver
  29 * needs to perform a task that needs to complete within a specific scanout
  30 * period, but might possibly block and thus cannot be handled in an IRQ
  31 * context. Both of these situations can't be solved perfectly in Linux since
  32 * we're not a realtime kernel, and thus the scheduler may cause us to miss
  33 * our deadline if it decides to preempt us. But for some drivers, it's good
  34 * enough if we can lower our chance of being preempted to an absolute
  35 * minimum.
  36 *
  37 * This is where &drm_vblank_work comes in. &drm_vblank_work provides a simple
  38 * generic delayed work implementation which delays work execution until a
  39 * particular vblank has passed, and then executes the work at realtime
  40 * priority. This provides the best possible chance at performing
  41 * time-sensitive hardware programming on time, even when the system is under
  42 * heavy load. &drm_vblank_work also supports rescheduling, so that self
  43 * re-arming work items can be easily implemented.
  44 */
  45
  46void drm_handle_vblank_works(struct drm_vblank_crtc *vblank)
  47{
  48        struct drm_vblank_work *work, *next;
  49        u64 count = atomic64_read(&vblank->count);
  50        bool wake = false;
  51
  52        assert_spin_locked(&vblank->dev->event_lock);
  53
  54        list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
  55                if (!drm_vblank_passed(count, work->count))
  56                        continue;
  57
  58                list_del_init(&work->node);
  59                drm_vblank_put(vblank->dev, vblank->pipe);
  60                kthread_queue_work(vblank->worker, &work->base);
  61                wake = true;
  62        }
  63        if (wake)
  64                wake_up_all(&vblank->work_wait_queue);
  65}
  66
  67/* Handle cancelling any pending vblank work items and drop respective vblank
  68 * references in response to vblank interrupts being disabled.
  69 */
  70void drm_vblank_cancel_pending_works(struct drm_vblank_crtc *vblank)
  71{
  72        struct drm_vblank_work *work, *next;
  73
  74        assert_spin_locked(&vblank->dev->event_lock);
  75
  76        list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
  77                list_del_init(&work->node);
  78                drm_vblank_put(vblank->dev, vblank->pipe);
  79        }
  80
  81        wake_up_all(&vblank->work_wait_queue);
  82}
  83
  84/**
  85 * drm_vblank_work_schedule - schedule a vblank work
  86 * @work: vblank work to schedule
  87 * @count: target vblank count
  88 * @nextonmiss: defer until the next vblank if target vblank was missed
  89 *
  90 * Schedule @work for execution once the crtc vblank count reaches @count.
  91 *
  92 * If the crtc vblank count has already reached @count and @nextonmiss is
  93 * %false the work starts to execute immediately.
  94 *
  95 * If the crtc vblank count has already reached @count and @nextonmiss is
  96 * %true the work is deferred until the next vblank (as if @count has been
  97 * specified as crtc vblank count + 1).
  98 *
  99 * If @work is already scheduled, this function will reschedule said work
 100 * using the new @count. This can be used for self-rearming work items.
 101 *
 102 * Returns:
 103 * %1 if @work was successfully (re)scheduled, %0 if it was either already
 104 * scheduled or cancelled, or a negative error code on failure.
 105 */
 106int drm_vblank_work_schedule(struct drm_vblank_work *work,
 107                             u64 count, bool nextonmiss)
 108{
 109        struct drm_vblank_crtc *vblank = work->vblank;
 110        struct drm_device *dev = vblank->dev;
 111        u64 cur_vbl;
 112        unsigned long irqflags;
 113        bool passed, inmodeset, rescheduling = false, wake = false;
 114        int ret = 0;
 115
 116        spin_lock_irqsave(&dev->event_lock, irqflags);
 117        if (work->cancelling)
 118                goto out;
 119
 120        spin_lock(&dev->vbl_lock);
 121        inmodeset = vblank->inmodeset;
 122        spin_unlock(&dev->vbl_lock);
 123        if (inmodeset)
 124                goto out;
 125
 126        if (list_empty(&work->node)) {
 127                ret = drm_vblank_get(dev, vblank->pipe);
 128                if (ret < 0)
 129                        goto out;
 130        } else if (work->count == count) {
 131                /* Already scheduled w/ same vbl count */
 132                goto out;
 133        } else {
 134                rescheduling = true;
 135        }
 136
 137        work->count = count;
 138        cur_vbl = drm_vblank_count(dev, vblank->pipe);
 139        passed = drm_vblank_passed(cur_vbl, count);
 140        if (passed)
 141                drm_dbg_core(dev,
 142                             "crtc %d vblank %llu already passed (current %llu)\n",
 143                             vblank->pipe, count, cur_vbl);
 144
 145        if (!nextonmiss && passed) {
 146                drm_vblank_put(dev, vblank->pipe);
 147                ret = kthread_queue_work(vblank->worker, &work->base);
 148
 149                if (rescheduling) {
 150                        list_del_init(&work->node);
 151                        wake = true;
 152                }
 153        } else {
 154                if (!rescheduling)
 155                        list_add_tail(&work->node, &vblank->pending_work);
 156                ret = true;
 157        }
 158
 159out:
 160        spin_unlock_irqrestore(&dev->event_lock, irqflags);
 161        if (wake)
 162                wake_up_all(&vblank->work_wait_queue);
 163        return ret;
 164}
 165EXPORT_SYMBOL(drm_vblank_work_schedule);
 166
 167/**
 168 * drm_vblank_work_cancel_sync - cancel a vblank work and wait for it to
 169 * finish executing
 170 * @work: vblank work to cancel
 171 *
 172 * Cancel an already scheduled vblank work and wait for its
 173 * execution to finish.
 174 *
 175 * On return, @work is guaranteed to no longer be scheduled or running, even
 176 * if it's self-arming.
 177 *
 178 * Returns:
 179 * %True if the work was cancelled before it started to execute, %false
 180 * otherwise.
 181 */
 182bool drm_vblank_work_cancel_sync(struct drm_vblank_work *work)
 183{
 184        struct drm_vblank_crtc *vblank = work->vblank;
 185        struct drm_device *dev = vblank->dev;
 186        bool ret = false;
 187
 188        spin_lock_irq(&dev->event_lock);
 189        if (!list_empty(&work->node)) {
 190                list_del_init(&work->node);
 191                drm_vblank_put(vblank->dev, vblank->pipe);
 192                ret = true;
 193        }
 194
 195        work->cancelling++;
 196        spin_unlock_irq(&dev->event_lock);
 197
 198        wake_up_all(&vblank->work_wait_queue);
 199
 200        if (kthread_cancel_work_sync(&work->base))
 201                ret = true;
 202
 203        spin_lock_irq(&dev->event_lock);
 204        work->cancelling--;
 205        spin_unlock_irq(&dev->event_lock);
 206
 207        return ret;
 208}
 209EXPORT_SYMBOL(drm_vblank_work_cancel_sync);
 210
 211/**
 212 * drm_vblank_work_flush - wait for a scheduled vblank work to finish
 213 * executing
 214 * @work: vblank work to flush
 215 *
 216 * Wait until @work has finished executing once.
 217 */
 218void drm_vblank_work_flush(struct drm_vblank_work *work)
 219{
 220        struct drm_vblank_crtc *vblank = work->vblank;
 221        struct drm_device *dev = vblank->dev;
 222
 223        spin_lock_irq(&dev->event_lock);
 224        wait_event_lock_irq(vblank->work_wait_queue, list_empty(&work->node),
 225                            dev->event_lock);
 226        spin_unlock_irq(&dev->event_lock);
 227
 228        kthread_flush_work(&work->base);
 229}
 230EXPORT_SYMBOL(drm_vblank_work_flush);
 231
 232/**
 233 * drm_vblank_work_init - initialize a vblank work item
 234 * @work: vblank work item
 235 * @crtc: CRTC whose vblank will trigger the work execution
 236 * @func: work function to be executed
 237 *
 238 * Initialize a vblank work item for a specific crtc.
 239 */
 240void drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc,
 241                          void (*func)(struct kthread_work *work))
 242{
 243        kthread_init_work(&work->base, func);
 244        INIT_LIST_HEAD(&work->node);
 245        work->vblank = &crtc->dev->vblank[drm_crtc_index(crtc)];
 246}
 247EXPORT_SYMBOL(drm_vblank_work_init);
 248
 249int drm_vblank_worker_init(struct drm_vblank_crtc *vblank)
 250{
 251        struct kthread_worker *worker;
 252
 253        INIT_LIST_HEAD(&vblank->pending_work);
 254        init_waitqueue_head(&vblank->work_wait_queue);
 255        worker = kthread_create_worker(0, "card%d-crtc%d",
 256                                       vblank->dev->primary->index,
 257                                       vblank->pipe);
 258        if (IS_ERR(worker))
 259                return PTR_ERR(worker);
 260
 261        vblank->worker = worker;
 262
 263        sched_set_fifo(worker->task);
 264        return 0;
 265}
 266