mirror of
https://github.com/torvalds/linux.git
synced 2024-11-25 13:41:51 +00:00
653e248645
Silently cancelling vblank works is a bit rude, especially if said works do any resource management (eg. free memory). WARN if we ever hit this. TODO: Maybe drm_crtc_vblank_off() should wait for any pending work to reach its target vblank before actually doing anything drastic? Cc: Lyude Paul <lyude@redhat.com> Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20230920140339.28322-2-ville.syrjala@linux.intel.com Reviewed-by: Lyude Paul <lyude@redhat.com>
269 lines
8.1 KiB
C
269 lines
8.1 KiB
C
// SPDX-License-Identifier: MIT
|
|
|
|
#include <uapi/linux/sched/types.h>
|
|
|
|
#include <drm/drm_print.h>
|
|
#include <drm/drm_vblank.h>
|
|
#include <drm/drm_vblank_work.h>
|
|
#include <drm/drm_crtc.h>
|
|
|
|
#include "drm_internal.h"
|
|
|
|
/**
|
|
* DOC: vblank works
|
|
*
|
|
* Many DRM drivers need to program hardware in a time-sensitive manner, many
|
|
* times with a deadline of starting and finishing within a certain region of
|
|
* the scanout. Most of the time the safest way to accomplish this is to
|
|
* simply do said time-sensitive programming in the driver's IRQ handler,
|
|
* which allows drivers to avoid being preempted during these critical
|
|
* regions. Or even better, the hardware may even handle applying such
|
|
* time-critical programming independently of the CPU.
|
|
*
|
|
* While there's a decent amount of hardware that's designed so that the CPU
|
|
* doesn't need to be concerned with extremely time-sensitive programming,
|
|
* there's a few situations where it can't be helped. Some unforgiving
|
|
* hardware may require that certain time-sensitive programming be handled
|
|
* completely by the CPU, and said programming may even take too long to
|
|
* handle in an IRQ handler. Another such situation would be where the driver
|
|
* needs to perform a task that needs to complete within a specific scanout
|
|
* period, but might possibly block and thus cannot be handled in an IRQ
|
|
* context. Both of these situations can't be solved perfectly in Linux since
|
|
* we're not a realtime kernel, and thus the scheduler may cause us to miss
|
|
* our deadline if it decides to preempt us. But for some drivers, it's good
|
|
* enough if we can lower our chance of being preempted to an absolute
|
|
* minimum.
|
|
*
|
|
* This is where &drm_vblank_work comes in. &drm_vblank_work provides a simple
|
|
* generic delayed work implementation which delays work execution until a
|
|
* particular vblank has passed, and then executes the work at realtime
|
|
* priority. This provides the best possible chance at performing
|
|
* time-sensitive hardware programming on time, even when the system is under
|
|
* heavy load. &drm_vblank_work also supports rescheduling, so that self
|
|
* re-arming work items can be easily implemented.
|
|
*/
|
|
|
|
void drm_handle_vblank_works(struct drm_vblank_crtc *vblank)
|
|
{
|
|
struct drm_vblank_work *work, *next;
|
|
u64 count = atomic64_read(&vblank->count);
|
|
bool wake = false;
|
|
|
|
assert_spin_locked(&vblank->dev->event_lock);
|
|
|
|
list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
|
|
if (!drm_vblank_passed(count, work->count))
|
|
continue;
|
|
|
|
list_del_init(&work->node);
|
|
drm_vblank_put(vblank->dev, vblank->pipe);
|
|
kthread_queue_work(vblank->worker, &work->base);
|
|
wake = true;
|
|
}
|
|
if (wake)
|
|
wake_up_all(&vblank->work_wait_queue);
|
|
}
|
|
|
|
/* Handle cancelling any pending vblank work items and drop respective vblank
|
|
* references in response to vblank interrupts being disabled.
|
|
*/
|
|
void drm_vblank_cancel_pending_works(struct drm_vblank_crtc *vblank)
|
|
{
|
|
struct drm_vblank_work *work, *next;
|
|
|
|
assert_spin_locked(&vblank->dev->event_lock);
|
|
|
|
drm_WARN_ONCE(vblank->dev, !list_empty(&vblank->pending_work),
|
|
"Cancelling pending vblank works!\n");
|
|
|
|
list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
|
|
list_del_init(&work->node);
|
|
drm_vblank_put(vblank->dev, vblank->pipe);
|
|
}
|
|
|
|
wake_up_all(&vblank->work_wait_queue);
|
|
}
|
|
|
|
/**
|
|
* drm_vblank_work_schedule - schedule a vblank work
|
|
* @work: vblank work to schedule
|
|
* @count: target vblank count
|
|
* @nextonmiss: defer until the next vblank if target vblank was missed
|
|
*
|
|
* Schedule @work for execution once the crtc vblank count reaches @count.
|
|
*
|
|
* If the crtc vblank count has already reached @count and @nextonmiss is
|
|
* %false the work starts to execute immediately.
|
|
*
|
|
* If the crtc vblank count has already reached @count and @nextonmiss is
|
|
* %true the work is deferred until the next vblank (as if @count has been
|
|
* specified as crtc vblank count + 1).
|
|
*
|
|
* If @work is already scheduled, this function will reschedule said work
|
|
* using the new @count. This can be used for self-rearming work items.
|
|
*
|
|
* Returns:
|
|
* %1 if @work was successfully (re)scheduled, %0 if it was either already
|
|
* scheduled or cancelled, or a negative error code on failure.
|
|
*/
|
|
int drm_vblank_work_schedule(struct drm_vblank_work *work,
|
|
u64 count, bool nextonmiss)
|
|
{
|
|
struct drm_vblank_crtc *vblank = work->vblank;
|
|
struct drm_device *dev = vblank->dev;
|
|
u64 cur_vbl;
|
|
unsigned long irqflags;
|
|
bool passed, inmodeset, rescheduling = false, wake = false;
|
|
int ret = 0;
|
|
|
|
spin_lock_irqsave(&dev->event_lock, irqflags);
|
|
if (work->cancelling)
|
|
goto out;
|
|
|
|
spin_lock(&dev->vbl_lock);
|
|
inmodeset = vblank->inmodeset;
|
|
spin_unlock(&dev->vbl_lock);
|
|
if (inmodeset)
|
|
goto out;
|
|
|
|
if (list_empty(&work->node)) {
|
|
ret = drm_vblank_get(dev, vblank->pipe);
|
|
if (ret < 0)
|
|
goto out;
|
|
} else if (work->count == count) {
|
|
/* Already scheduled w/ same vbl count */
|
|
goto out;
|
|
} else {
|
|
rescheduling = true;
|
|
}
|
|
|
|
work->count = count;
|
|
cur_vbl = drm_vblank_count(dev, vblank->pipe);
|
|
passed = drm_vblank_passed(cur_vbl, count);
|
|
if (passed)
|
|
drm_dbg_core(dev,
|
|
"crtc %d vblank %llu already passed (current %llu)\n",
|
|
vblank->pipe, count, cur_vbl);
|
|
|
|
if (!nextonmiss && passed) {
|
|
drm_vblank_put(dev, vblank->pipe);
|
|
ret = kthread_queue_work(vblank->worker, &work->base);
|
|
|
|
if (rescheduling) {
|
|
list_del_init(&work->node);
|
|
wake = true;
|
|
}
|
|
} else {
|
|
if (!rescheduling)
|
|
list_add_tail(&work->node, &vblank->pending_work);
|
|
ret = true;
|
|
}
|
|
|
|
out:
|
|
spin_unlock_irqrestore(&dev->event_lock, irqflags);
|
|
if (wake)
|
|
wake_up_all(&vblank->work_wait_queue);
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(drm_vblank_work_schedule);
|
|
|
|
/**
|
|
* drm_vblank_work_cancel_sync - cancel a vblank work and wait for it to
|
|
* finish executing
|
|
* @work: vblank work to cancel
|
|
*
|
|
* Cancel an already scheduled vblank work and wait for its
|
|
* execution to finish.
|
|
*
|
|
* On return, @work is guaranteed to no longer be scheduled or running, even
|
|
* if it's self-arming.
|
|
*
|
|
* Returns:
|
|
* %True if the work was cancelled before it started to execute, %false
|
|
* otherwise.
|
|
*/
|
|
bool drm_vblank_work_cancel_sync(struct drm_vblank_work *work)
|
|
{
|
|
struct drm_vblank_crtc *vblank = work->vblank;
|
|
struct drm_device *dev = vblank->dev;
|
|
bool ret = false;
|
|
|
|
spin_lock_irq(&dev->event_lock);
|
|
if (!list_empty(&work->node)) {
|
|
list_del_init(&work->node);
|
|
drm_vblank_put(vblank->dev, vblank->pipe);
|
|
ret = true;
|
|
}
|
|
|
|
work->cancelling++;
|
|
spin_unlock_irq(&dev->event_lock);
|
|
|
|
wake_up_all(&vblank->work_wait_queue);
|
|
|
|
if (kthread_cancel_work_sync(&work->base))
|
|
ret = true;
|
|
|
|
spin_lock_irq(&dev->event_lock);
|
|
work->cancelling--;
|
|
spin_unlock_irq(&dev->event_lock);
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(drm_vblank_work_cancel_sync);
|
|
|
|
/**
|
|
* drm_vblank_work_flush - wait for a scheduled vblank work to finish
|
|
* executing
|
|
* @work: vblank work to flush
|
|
*
|
|
* Wait until @work has finished executing once.
|
|
*/
|
|
void drm_vblank_work_flush(struct drm_vblank_work *work)
|
|
{
|
|
struct drm_vblank_crtc *vblank = work->vblank;
|
|
struct drm_device *dev = vblank->dev;
|
|
|
|
spin_lock_irq(&dev->event_lock);
|
|
wait_event_lock_irq(vblank->work_wait_queue, list_empty(&work->node),
|
|
dev->event_lock);
|
|
spin_unlock_irq(&dev->event_lock);
|
|
|
|
kthread_flush_work(&work->base);
|
|
}
|
|
EXPORT_SYMBOL(drm_vblank_work_flush);
|
|
|
|
/**
|
|
* drm_vblank_work_init - initialize a vblank work item
|
|
* @work: vblank work item
|
|
* @crtc: CRTC whose vblank will trigger the work execution
|
|
* @func: work function to be executed
|
|
*
|
|
* Initialize a vblank work item for a specific crtc.
|
|
*/
|
|
void drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc,
|
|
void (*func)(struct kthread_work *work))
|
|
{
|
|
kthread_init_work(&work->base, func);
|
|
INIT_LIST_HEAD(&work->node);
|
|
work->vblank = &crtc->dev->vblank[drm_crtc_index(crtc)];
|
|
}
|
|
EXPORT_SYMBOL(drm_vblank_work_init);
|
|
|
|
int drm_vblank_worker_init(struct drm_vblank_crtc *vblank)
|
|
{
|
|
struct kthread_worker *worker;
|
|
|
|
INIT_LIST_HEAD(&vblank->pending_work);
|
|
init_waitqueue_head(&vblank->work_wait_queue);
|
|
worker = kthread_create_worker(0, "card%d-crtc%d",
|
|
vblank->dev->primary->index,
|
|
vblank->pipe);
|
|
if (IS_ERR(worker))
|
|
return PTR_ERR(worker);
|
|
|
|
vblank->worker = worker;
|
|
|
|
sched_set_fifo(worker->task);
|
|
return 0;
|
|
}
|