forked from Minki/linux
c0fe07aa50
Completely different approach: Instead of encoding each and every framebuffer update as spice operation simply update the shadow framebuffer and maintain a dirty rectangle. Also schedule a worker to push an update for the dirty rectangle as spice operation. Usually a bunch of dirty rectangle updates are collected before the worker actually runs. What changes: Updates get batched now. Instead of sending tons of small updates a few large ones are sent. When the same region is updated multiple times within a short timeframe (scrolling multiple lines for example) we send a single update only. Spice server has an easier job now: The dependency tree for display operations which spice server maintains for lazy rendering is alot smaller now. Spice server's image compression probably works better too with the larger image blits. Net effect: framebuffer console @ qxldrmfb is an order of magnitude faster now. Signed-off-by: Gerd Hoffmann <kraxel@redhat.com> Signed-off-by: Dave Airlie <airlied@redhat.com>
569 lines
14 KiB
C
569 lines
14 KiB
C
/*
|
|
* Copyright © 2013 Red Hat
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*
|
|
* Authors:
|
|
* David Airlie
|
|
*/
|
|
#include <linux/module.h>
|
|
#include <linux/fb.h>
|
|
|
|
#include "drmP.h"
|
|
#include "drm/drm.h"
|
|
#include "drm/drm_crtc.h"
|
|
#include "drm/drm_crtc_helper.h"
|
|
#include "qxl_drv.h"
|
|
|
|
#include "qxl_object.h"
|
|
#include "drm_fb_helper.h"
|
|
|
|
#define QXL_DIRTY_DELAY (HZ / 30)
|
|
|
|
struct qxl_fbdev {
|
|
struct drm_fb_helper helper;
|
|
struct qxl_framebuffer qfb;
|
|
struct list_head fbdev_list;
|
|
struct qxl_device *qdev;
|
|
|
|
spinlock_t delayed_ops_lock;
|
|
struct list_head delayed_ops;
|
|
void *shadow;
|
|
int size;
|
|
|
|
/* dirty memory logging */
|
|
struct {
|
|
spinlock_t lock;
|
|
unsigned x1;
|
|
unsigned y1;
|
|
unsigned x2;
|
|
unsigned y2;
|
|
} dirty;
|
|
};
|
|
|
|
static void qxl_fb_image_init(struct qxl_fb_image *qxl_fb_image,
|
|
struct qxl_device *qdev, struct fb_info *info,
|
|
const struct fb_image *image)
|
|
{
|
|
qxl_fb_image->qdev = qdev;
|
|
if (info) {
|
|
qxl_fb_image->visual = info->fix.visual;
|
|
if (qxl_fb_image->visual == FB_VISUAL_TRUECOLOR ||
|
|
qxl_fb_image->visual == FB_VISUAL_DIRECTCOLOR)
|
|
memcpy(&qxl_fb_image->pseudo_palette,
|
|
info->pseudo_palette,
|
|
sizeof(qxl_fb_image->pseudo_palette));
|
|
} else {
|
|
/* fallback */
|
|
if (image->depth == 1)
|
|
qxl_fb_image->visual = FB_VISUAL_MONO10;
|
|
else
|
|
qxl_fb_image->visual = FB_VISUAL_DIRECTCOLOR;
|
|
}
|
|
if (image) {
|
|
memcpy(&qxl_fb_image->fb_image, image,
|
|
sizeof(qxl_fb_image->fb_image));
|
|
}
|
|
}
|
|
|
|
static void qxl_fb_dirty_flush(struct fb_info *info)
|
|
{
|
|
struct qxl_fbdev *qfbdev = info->par;
|
|
struct qxl_device *qdev = qfbdev->qdev;
|
|
struct qxl_fb_image qxl_fb_image;
|
|
struct fb_image *image = &qxl_fb_image.fb_image;
|
|
unsigned long flags;
|
|
u32 x1, x2, y1, y2;
|
|
|
|
/* TODO: hard coding 32 bpp */
|
|
int stride = qfbdev->qfb.base.pitches[0];
|
|
|
|
spin_lock_irqsave(&qfbdev->dirty.lock, flags);
|
|
|
|
x1 = qfbdev->dirty.x1;
|
|
x2 = qfbdev->dirty.x2;
|
|
y1 = qfbdev->dirty.y1;
|
|
y2 = qfbdev->dirty.y2;
|
|
qfbdev->dirty.x1 = 0;
|
|
qfbdev->dirty.x2 = 0;
|
|
qfbdev->dirty.y1 = 0;
|
|
qfbdev->dirty.y2 = 0;
|
|
|
|
spin_unlock_irqrestore(&qfbdev->dirty.lock, flags);
|
|
|
|
/*
|
|
* we are using a shadow draw buffer, at qdev->surface0_shadow
|
|
*/
|
|
qxl_io_log(qdev, "dirty x[%d, %d], y[%d, %d]", x1, x2, y1, y2);
|
|
image->dx = x1;
|
|
image->dy = y1;
|
|
image->width = x2 - x1 + 1;
|
|
image->height = y2 - y1 + 1;
|
|
image->fg_color = 0xffffffff; /* unused, just to avoid uninitialized
|
|
warnings */
|
|
image->bg_color = 0;
|
|
image->depth = 32; /* TODO: take from somewhere? */
|
|
image->cmap.start = 0;
|
|
image->cmap.len = 0;
|
|
image->cmap.red = NULL;
|
|
image->cmap.green = NULL;
|
|
image->cmap.blue = NULL;
|
|
image->cmap.transp = NULL;
|
|
image->data = qfbdev->shadow + (x1 * 4) + (stride * y1);
|
|
|
|
qxl_fb_image_init(&qxl_fb_image, qdev, info, NULL);
|
|
qxl_draw_opaque_fb(&qxl_fb_image, stride);
|
|
}
|
|
|
|
static void qxl_dirty_update(struct qxl_fbdev *qfbdev,
|
|
int x, int y, int width, int height)
|
|
{
|
|
struct qxl_device *qdev = qfbdev->qdev;
|
|
unsigned long flags;
|
|
int x2, y2;
|
|
|
|
x2 = x + width - 1;
|
|
y2 = y + height - 1;
|
|
|
|
spin_lock_irqsave(&qfbdev->dirty.lock, flags);
|
|
|
|
if (qfbdev->dirty.y1 < y)
|
|
y = qfbdev->dirty.y1;
|
|
if (qfbdev->dirty.y2 > y2)
|
|
y2 = qfbdev->dirty.y2;
|
|
if (qfbdev->dirty.x1 < x)
|
|
x = qfbdev->dirty.x1;
|
|
if (qfbdev->dirty.x2 > x2)
|
|
x2 = qfbdev->dirty.x2;
|
|
|
|
qfbdev->dirty.x1 = x;
|
|
qfbdev->dirty.x2 = x2;
|
|
qfbdev->dirty.y1 = y;
|
|
qfbdev->dirty.y2 = y2;
|
|
|
|
spin_unlock_irqrestore(&qfbdev->dirty.lock, flags);
|
|
|
|
schedule_work(&qdev->fb_work);
|
|
}
|
|
|
|
static void qxl_deferred_io(struct fb_info *info,
|
|
struct list_head *pagelist)
|
|
{
|
|
struct qxl_fbdev *qfbdev = info->par;
|
|
unsigned long start, end, min, max;
|
|
struct page *page;
|
|
int y1, y2;
|
|
|
|
min = ULONG_MAX;
|
|
max = 0;
|
|
list_for_each_entry(page, pagelist, lru) {
|
|
start = page->index << PAGE_SHIFT;
|
|
end = start + PAGE_SIZE - 1;
|
|
min = min(min, start);
|
|
max = max(max, end);
|
|
}
|
|
|
|
if (min < max) {
|
|
y1 = min / info->fix.line_length;
|
|
y2 = (max / info->fix.line_length) + 1;
|
|
qxl_dirty_update(qfbdev, 0, y1, info->var.xres, y2 - y1);
|
|
}
|
|
};
|
|
|
|
static struct fb_deferred_io qxl_defio = {
|
|
.delay = QXL_DIRTY_DELAY,
|
|
.deferred_io = qxl_deferred_io,
|
|
};
|
|
|
|
static void qxl_fb_fillrect(struct fb_info *info,
|
|
const struct fb_fillrect *rect)
|
|
{
|
|
struct qxl_fbdev *qfbdev = info->par;
|
|
|
|
sys_fillrect(info, rect);
|
|
qxl_dirty_update(qfbdev, rect->dx, rect->dy, rect->width,
|
|
rect->height);
|
|
}
|
|
|
|
static void qxl_fb_copyarea(struct fb_info *info,
|
|
const struct fb_copyarea *area)
|
|
{
|
|
struct qxl_fbdev *qfbdev = info->par;
|
|
|
|
sys_copyarea(info, area);
|
|
qxl_dirty_update(qfbdev, area->dx, area->dy, area->width,
|
|
area->height);
|
|
}
|
|
|
|
static void qxl_fb_imageblit(struct fb_info *info,
|
|
const struct fb_image *image)
|
|
{
|
|
struct qxl_fbdev *qfbdev = info->par;
|
|
|
|
sys_imageblit(info, image);
|
|
qxl_dirty_update(qfbdev, image->dx, image->dy, image->width,
|
|
image->height);
|
|
}
|
|
|
|
static void qxl_fb_work(struct work_struct *work)
|
|
{
|
|
struct qxl_device *qdev = container_of(work, struct qxl_device, fb_work);
|
|
struct qxl_fbdev *qfbdev = qdev->mode_info.qfbdev;
|
|
|
|
qxl_fb_dirty_flush(qfbdev->helper.fbdev);
|
|
}
|
|
|
|
int qxl_fb_init(struct qxl_device *qdev)
|
|
{
|
|
INIT_WORK(&qdev->fb_work, qxl_fb_work);
|
|
return 0;
|
|
}
|
|
|
|
static struct fb_ops qxlfb_ops = {
|
|
.owner = THIS_MODULE,
|
|
.fb_check_var = drm_fb_helper_check_var,
|
|
.fb_set_par = drm_fb_helper_set_par, /* TODO: copy vmwgfx */
|
|
.fb_fillrect = qxl_fb_fillrect,
|
|
.fb_copyarea = qxl_fb_copyarea,
|
|
.fb_imageblit = qxl_fb_imageblit,
|
|
.fb_pan_display = drm_fb_helper_pan_display,
|
|
.fb_blank = drm_fb_helper_blank,
|
|
.fb_setcmap = drm_fb_helper_setcmap,
|
|
.fb_debug_enter = drm_fb_helper_debug_enter,
|
|
.fb_debug_leave = drm_fb_helper_debug_leave,
|
|
};
|
|
|
|
static void qxlfb_destroy_pinned_object(struct drm_gem_object *gobj)
|
|
{
|
|
struct qxl_bo *qbo = gem_to_qxl_bo(gobj);
|
|
int ret;
|
|
|
|
ret = qxl_bo_reserve(qbo, false);
|
|
if (likely(ret == 0)) {
|
|
qxl_bo_kunmap(qbo);
|
|
qxl_bo_unpin(qbo);
|
|
qxl_bo_unreserve(qbo);
|
|
}
|
|
drm_gem_object_unreference_unlocked(gobj);
|
|
}
|
|
|
|
int qxl_get_handle_for_primary_fb(struct qxl_device *qdev,
|
|
struct drm_file *file_priv,
|
|
uint32_t *handle)
|
|
{
|
|
int r;
|
|
struct drm_gem_object *gobj = qdev->fbdev_qfb->obj;
|
|
|
|
BUG_ON(!gobj);
|
|
/* drm_get_handle_create adds a reference - good */
|
|
r = drm_gem_handle_create(file_priv, gobj, handle);
|
|
if (r)
|
|
return r;
|
|
return 0;
|
|
}
|
|
|
|
static int qxlfb_create_pinned_object(struct qxl_fbdev *qfbdev,
|
|
struct drm_mode_fb_cmd2 *mode_cmd,
|
|
struct drm_gem_object **gobj_p)
|
|
{
|
|
struct qxl_device *qdev = qfbdev->qdev;
|
|
struct drm_gem_object *gobj = NULL;
|
|
struct qxl_bo *qbo = NULL;
|
|
int ret;
|
|
int aligned_size, size;
|
|
int height = mode_cmd->height;
|
|
int bpp;
|
|
int depth;
|
|
|
|
drm_fb_get_bpp_depth(mode_cmd->pixel_format, &bpp, &depth);
|
|
|
|
size = mode_cmd->pitches[0] * height;
|
|
aligned_size = ALIGN(size, PAGE_SIZE);
|
|
/* TODO: unallocate and reallocate surface0 for real. Hack to just
|
|
* have a large enough surface0 for 1024x768 Xorg 32bpp mode */
|
|
ret = qxl_gem_object_create(qdev, aligned_size, 0,
|
|
QXL_GEM_DOMAIN_SURFACE,
|
|
false, /* is discardable */
|
|
false, /* is kernel (false means device) */
|
|
NULL,
|
|
&gobj);
|
|
if (ret) {
|
|
pr_err("failed to allocate framebuffer (%d)\n",
|
|
aligned_size);
|
|
return -ENOMEM;
|
|
}
|
|
qbo = gem_to_qxl_bo(gobj);
|
|
|
|
qbo->surf.width = mode_cmd->width;
|
|
qbo->surf.height = mode_cmd->height;
|
|
qbo->surf.stride = mode_cmd->pitches[0];
|
|
qbo->surf.format = SPICE_SURFACE_FMT_32_xRGB;
|
|
ret = qxl_bo_reserve(qbo, false);
|
|
if (unlikely(ret != 0))
|
|
goto out_unref;
|
|
ret = qxl_bo_pin(qbo, QXL_GEM_DOMAIN_SURFACE, NULL);
|
|
if (ret) {
|
|
qxl_bo_unreserve(qbo);
|
|
goto out_unref;
|
|
}
|
|
ret = qxl_bo_kmap(qbo, NULL);
|
|
qxl_bo_unreserve(qbo); /* unreserve, will be mmaped */
|
|
if (ret)
|
|
goto out_unref;
|
|
|
|
*gobj_p = gobj;
|
|
return 0;
|
|
out_unref:
|
|
qxlfb_destroy_pinned_object(gobj);
|
|
*gobj_p = NULL;
|
|
return ret;
|
|
}
|
|
|
|
static int qxlfb_create(struct qxl_fbdev *qfbdev,
|
|
struct drm_fb_helper_surface_size *sizes)
|
|
{
|
|
struct qxl_device *qdev = qfbdev->qdev;
|
|
struct fb_info *info;
|
|
struct drm_framebuffer *fb = NULL;
|
|
struct drm_mode_fb_cmd2 mode_cmd;
|
|
struct drm_gem_object *gobj = NULL;
|
|
struct qxl_bo *qbo = NULL;
|
|
struct device *device = &qdev->pdev->dev;
|
|
int ret;
|
|
int size;
|
|
int bpp = sizes->surface_bpp;
|
|
int depth = sizes->surface_depth;
|
|
void *shadow;
|
|
|
|
mode_cmd.width = sizes->surface_width;
|
|
mode_cmd.height = sizes->surface_height;
|
|
|
|
mode_cmd.pitches[0] = ALIGN(mode_cmd.width * ((bpp + 1) / 8), 64);
|
|
mode_cmd.pixel_format = drm_mode_legacy_fb_format(bpp, depth);
|
|
|
|
ret = qxlfb_create_pinned_object(qfbdev, &mode_cmd, &gobj);
|
|
qbo = gem_to_qxl_bo(gobj);
|
|
QXL_INFO(qdev, "%s: %dx%d %d\n", __func__, mode_cmd.width,
|
|
mode_cmd.height, mode_cmd.pitches[0]);
|
|
|
|
shadow = vmalloc(mode_cmd.pitches[0] * mode_cmd.height);
|
|
/* TODO: what's the usual response to memory allocation errors? */
|
|
BUG_ON(!shadow);
|
|
QXL_INFO(qdev,
|
|
"surface0 at gpu offset %lld, mmap_offset %lld (virt %p, shadow %p)\n",
|
|
qxl_bo_gpu_offset(qbo),
|
|
qxl_bo_mmap_offset(qbo),
|
|
qbo->kptr,
|
|
shadow);
|
|
size = mode_cmd.pitches[0] * mode_cmd.height;
|
|
|
|
info = framebuffer_alloc(0, device);
|
|
if (info == NULL) {
|
|
ret = -ENOMEM;
|
|
goto out_unref;
|
|
}
|
|
|
|
info->par = qfbdev;
|
|
|
|
qxl_framebuffer_init(qdev->ddev, &qfbdev->qfb, &mode_cmd, gobj);
|
|
|
|
fb = &qfbdev->qfb.base;
|
|
|
|
/* setup helper with fb data */
|
|
qfbdev->helper.fb = fb;
|
|
qfbdev->helper.fbdev = info;
|
|
qfbdev->shadow = shadow;
|
|
strcpy(info->fix.id, "qxldrmfb");
|
|
|
|
drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
|
|
|
|
info->flags = FBINFO_DEFAULT | FBINFO_HWACCEL_COPYAREA | FBINFO_HWACCEL_FILLRECT;
|
|
info->fbops = &qxlfb_ops;
|
|
|
|
/*
|
|
* TODO: using gobj->size in various places in this function. Not sure
|
|
* what the difference between the different sizes is.
|
|
*/
|
|
info->fix.smem_start = qdev->vram_base; /* TODO - correct? */
|
|
info->fix.smem_len = gobj->size;
|
|
info->screen_base = qfbdev->shadow;
|
|
info->screen_size = gobj->size;
|
|
|
|
drm_fb_helper_fill_var(info, &qfbdev->helper, sizes->fb_width,
|
|
sizes->fb_height);
|
|
|
|
/* setup aperture base/size for vesafb takeover */
|
|
info->apertures = alloc_apertures(1);
|
|
if (!info->apertures) {
|
|
ret = -ENOMEM;
|
|
goto out_unref;
|
|
}
|
|
info->apertures->ranges[0].base = qdev->ddev->mode_config.fb_base;
|
|
info->apertures->ranges[0].size = qdev->vram_size;
|
|
|
|
info->fix.mmio_start = 0;
|
|
info->fix.mmio_len = 0;
|
|
|
|
if (info->screen_base == NULL) {
|
|
ret = -ENOSPC;
|
|
goto out_unref;
|
|
}
|
|
|
|
ret = fb_alloc_cmap(&info->cmap, 256, 0);
|
|
if (ret) {
|
|
ret = -ENOMEM;
|
|
goto out_unref;
|
|
}
|
|
|
|
info->fbdefio = &qxl_defio;
|
|
fb_deferred_io_init(info);
|
|
|
|
qdev->fbdev_info = info;
|
|
qdev->fbdev_qfb = &qfbdev->qfb;
|
|
DRM_INFO("fb mappable at 0x%lX, size %lu\n", info->fix.smem_start, (unsigned long)info->screen_size);
|
|
DRM_INFO("fb: depth %d, pitch %d, width %d, height %d\n", fb->depth, fb->pitches[0], fb->width, fb->height);
|
|
return 0;
|
|
|
|
out_unref:
|
|
if (qbo) {
|
|
ret = qxl_bo_reserve(qbo, false);
|
|
if (likely(ret == 0)) {
|
|
qxl_bo_kunmap(qbo);
|
|
qxl_bo_unpin(qbo);
|
|
qxl_bo_unreserve(qbo);
|
|
}
|
|
}
|
|
if (fb && ret) {
|
|
drm_gem_object_unreference(gobj);
|
|
drm_framebuffer_cleanup(fb);
|
|
kfree(fb);
|
|
}
|
|
drm_gem_object_unreference(gobj);
|
|
return ret;
|
|
}
|
|
|
|
static int qxl_fb_find_or_create_single(
|
|
struct drm_fb_helper *helper,
|
|
struct drm_fb_helper_surface_size *sizes)
|
|
{
|
|
struct qxl_fbdev *qfbdev =
|
|
container_of(helper, struct qxl_fbdev, helper);
|
|
int new_fb = 0;
|
|
int ret;
|
|
|
|
if (!helper->fb) {
|
|
ret = qxlfb_create(qfbdev, sizes);
|
|
if (ret)
|
|
return ret;
|
|
new_fb = 1;
|
|
}
|
|
return new_fb;
|
|
}
|
|
|
|
static int qxl_fbdev_destroy(struct drm_device *dev, struct qxl_fbdev *qfbdev)
|
|
{
|
|
struct fb_info *info;
|
|
struct qxl_framebuffer *qfb = &qfbdev->qfb;
|
|
|
|
if (qfbdev->helper.fbdev) {
|
|
info = qfbdev->helper.fbdev;
|
|
|
|
unregister_framebuffer(info);
|
|
framebuffer_release(info);
|
|
}
|
|
if (qfb->obj) {
|
|
qxlfb_destroy_pinned_object(qfb->obj);
|
|
qfb->obj = NULL;
|
|
}
|
|
drm_fb_helper_fini(&qfbdev->helper);
|
|
vfree(qfbdev->shadow);
|
|
drm_framebuffer_cleanup(&qfb->base);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct drm_fb_helper_funcs qxl_fb_helper_funcs = {
|
|
.fb_probe = qxl_fb_find_or_create_single,
|
|
};
|
|
|
|
int qxl_fbdev_init(struct qxl_device *qdev)
|
|
{
|
|
struct qxl_fbdev *qfbdev;
|
|
int bpp_sel = 32; /* TODO: parameter from somewhere? */
|
|
int ret;
|
|
|
|
qfbdev = kzalloc(sizeof(struct qxl_fbdev), GFP_KERNEL);
|
|
if (!qfbdev)
|
|
return -ENOMEM;
|
|
|
|
qfbdev->qdev = qdev;
|
|
qdev->mode_info.qfbdev = qfbdev;
|
|
spin_lock_init(&qfbdev->delayed_ops_lock);
|
|
spin_lock_init(&qfbdev->dirty.lock);
|
|
INIT_LIST_HEAD(&qfbdev->delayed_ops);
|
|
|
|
drm_fb_helper_prepare(qdev->ddev, &qfbdev->helper,
|
|
&qxl_fb_helper_funcs);
|
|
|
|
ret = drm_fb_helper_init(qdev->ddev, &qfbdev->helper,
|
|
qxl_num_crtc /* num_crtc - QXL supports just 1 */,
|
|
QXLFB_CONN_LIMIT);
|
|
if (ret)
|
|
goto free;
|
|
|
|
ret = drm_fb_helper_single_add_all_connectors(&qfbdev->helper);
|
|
if (ret)
|
|
goto fini;
|
|
|
|
ret = drm_fb_helper_initial_config(&qfbdev->helper, bpp_sel);
|
|
if (ret)
|
|
goto fini;
|
|
|
|
return 0;
|
|
|
|
fini:
|
|
drm_fb_helper_fini(&qfbdev->helper);
|
|
free:
|
|
kfree(qfbdev);
|
|
return ret;
|
|
}
|
|
|
|
void qxl_fbdev_fini(struct qxl_device *qdev)
|
|
{
|
|
if (!qdev->mode_info.qfbdev)
|
|
return;
|
|
|
|
qxl_fbdev_destroy(qdev->ddev, qdev->mode_info.qfbdev);
|
|
kfree(qdev->mode_info.qfbdev);
|
|
qdev->mode_info.qfbdev = NULL;
|
|
}
|
|
|
|
void qxl_fbdev_set_suspend(struct qxl_device *qdev, int state)
|
|
{
|
|
fb_set_suspend(qdev->mode_info.qfbdev->helper.fbdev, state);
|
|
}
|
|
|
|
bool qxl_fbdev_qobj_is_fb(struct qxl_device *qdev, struct qxl_bo *qobj)
|
|
{
|
|
if (qobj == gem_to_qxl_bo(qdev->mode_info.qfbdev->qfb.obj))
|
|
return true;
|
|
return false;
|
|
}
|