forked from Minki/linux
0c756134cf
For reasons that entirely elude me fb.h exposes all the structures, even when it is not enabled. Except for special stuff like fb_defio. Which means all the drivers which haven't yet switched over to the defio support in the helpers and still roll their own, will fail to compile when fbdev emulation is disabled. Protect just those bits, as a gnarly reminder that conversion to the core defio helpers would be good. Cc: Dave Airlie <airlied@redhat.com> Reviewed-by: Alex Deucher <alexander.deucher@amd.com> Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch> Link: http://patchwork.freedesktop.org/patch/msgid/1470847958-28465-6-git-send-email-daniel.vetter@ffwll.ch
450 lines
12 KiB
C
450 lines
12 KiB
C
/*
|
|
* Copyright © 2013 Red Hat
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*
|
|
* Authors:
|
|
* David Airlie
|
|
*/
|
|
#include <linux/module.h>
|
|
|
|
#include "drmP.h"
|
|
#include "drm/drm.h"
|
|
#include "drm/drm_crtc.h"
|
|
#include "drm/drm_crtc_helper.h"
|
|
#include "qxl_drv.h"
|
|
|
|
#include "qxl_object.h"
|
|
#include "drm_fb_helper.h"
|
|
|
|
#define QXL_DIRTY_DELAY (HZ / 30)
|
|
|
|
struct qxl_fbdev {
|
|
struct drm_fb_helper helper;
|
|
struct qxl_framebuffer qfb;
|
|
struct qxl_device *qdev;
|
|
|
|
spinlock_t delayed_ops_lock;
|
|
struct list_head delayed_ops;
|
|
void *shadow;
|
|
int size;
|
|
};
|
|
|
|
static void qxl_fb_image_init(struct qxl_fb_image *qxl_fb_image,
|
|
struct qxl_device *qdev, struct fb_info *info,
|
|
const struct fb_image *image)
|
|
{
|
|
qxl_fb_image->qdev = qdev;
|
|
if (info) {
|
|
qxl_fb_image->visual = info->fix.visual;
|
|
if (qxl_fb_image->visual == FB_VISUAL_TRUECOLOR ||
|
|
qxl_fb_image->visual == FB_VISUAL_DIRECTCOLOR)
|
|
memcpy(&qxl_fb_image->pseudo_palette,
|
|
info->pseudo_palette,
|
|
sizeof(qxl_fb_image->pseudo_palette));
|
|
} else {
|
|
/* fallback */
|
|
if (image->depth == 1)
|
|
qxl_fb_image->visual = FB_VISUAL_MONO10;
|
|
else
|
|
qxl_fb_image->visual = FB_VISUAL_DIRECTCOLOR;
|
|
}
|
|
if (image) {
|
|
memcpy(&qxl_fb_image->fb_image, image,
|
|
sizeof(qxl_fb_image->fb_image));
|
|
}
|
|
}
|
|
|
|
#ifdef CONFIG_DRM_FBDEV_EMULATION
|
|
static struct fb_deferred_io qxl_defio = {
|
|
.delay = QXL_DIRTY_DELAY,
|
|
.deferred_io = drm_fb_helper_deferred_io,
|
|
};
|
|
#endif
|
|
|
|
static struct fb_ops qxlfb_ops = {
|
|
.owner = THIS_MODULE,
|
|
.fb_check_var = drm_fb_helper_check_var,
|
|
.fb_set_par = drm_fb_helper_set_par, /* TODO: copy vmwgfx */
|
|
.fb_fillrect = drm_fb_helper_sys_fillrect,
|
|
.fb_copyarea = drm_fb_helper_sys_copyarea,
|
|
.fb_imageblit = drm_fb_helper_sys_imageblit,
|
|
.fb_pan_display = drm_fb_helper_pan_display,
|
|
.fb_blank = drm_fb_helper_blank,
|
|
.fb_setcmap = drm_fb_helper_setcmap,
|
|
.fb_debug_enter = drm_fb_helper_debug_enter,
|
|
.fb_debug_leave = drm_fb_helper_debug_leave,
|
|
};
|
|
|
|
static void qxlfb_destroy_pinned_object(struct drm_gem_object *gobj)
|
|
{
|
|
struct qxl_bo *qbo = gem_to_qxl_bo(gobj);
|
|
int ret;
|
|
|
|
ret = qxl_bo_reserve(qbo, false);
|
|
if (likely(ret == 0)) {
|
|
qxl_bo_kunmap(qbo);
|
|
qxl_bo_unpin(qbo);
|
|
qxl_bo_unreserve(qbo);
|
|
}
|
|
drm_gem_object_unreference_unlocked(gobj);
|
|
}
|
|
|
|
int qxl_get_handle_for_primary_fb(struct qxl_device *qdev,
|
|
struct drm_file *file_priv,
|
|
uint32_t *handle)
|
|
{
|
|
int r;
|
|
struct drm_gem_object *gobj = qdev->fbdev_qfb->obj;
|
|
|
|
BUG_ON(!gobj);
|
|
/* drm_get_handle_create adds a reference - good */
|
|
r = drm_gem_handle_create(file_priv, gobj, handle);
|
|
if (r)
|
|
return r;
|
|
return 0;
|
|
}
|
|
|
|
static int qxlfb_create_pinned_object(struct qxl_fbdev *qfbdev,
|
|
const struct drm_mode_fb_cmd2 *mode_cmd,
|
|
struct drm_gem_object **gobj_p)
|
|
{
|
|
struct qxl_device *qdev = qfbdev->qdev;
|
|
struct drm_gem_object *gobj = NULL;
|
|
struct qxl_bo *qbo = NULL;
|
|
int ret;
|
|
int aligned_size, size;
|
|
int height = mode_cmd->height;
|
|
|
|
size = mode_cmd->pitches[0] * height;
|
|
aligned_size = ALIGN(size, PAGE_SIZE);
|
|
/* TODO: unallocate and reallocate surface0 for real. Hack to just
|
|
* have a large enough surface0 for 1024x768 Xorg 32bpp mode */
|
|
ret = qxl_gem_object_create(qdev, aligned_size, 0,
|
|
QXL_GEM_DOMAIN_SURFACE,
|
|
false, /* is discardable */
|
|
false, /* is kernel (false means device) */
|
|
NULL,
|
|
&gobj);
|
|
if (ret) {
|
|
pr_err("failed to allocate framebuffer (%d)\n",
|
|
aligned_size);
|
|
return -ENOMEM;
|
|
}
|
|
qbo = gem_to_qxl_bo(gobj);
|
|
|
|
qbo->surf.width = mode_cmd->width;
|
|
qbo->surf.height = mode_cmd->height;
|
|
qbo->surf.stride = mode_cmd->pitches[0];
|
|
qbo->surf.format = SPICE_SURFACE_FMT_32_xRGB;
|
|
ret = qxl_bo_reserve(qbo, false);
|
|
if (unlikely(ret != 0))
|
|
goto out_unref;
|
|
ret = qxl_bo_pin(qbo, QXL_GEM_DOMAIN_SURFACE, NULL);
|
|
if (ret) {
|
|
qxl_bo_unreserve(qbo);
|
|
goto out_unref;
|
|
}
|
|
ret = qxl_bo_kmap(qbo, NULL);
|
|
qxl_bo_unreserve(qbo); /* unreserve, will be mmaped */
|
|
if (ret)
|
|
goto out_unref;
|
|
|
|
*gobj_p = gobj;
|
|
return 0;
|
|
out_unref:
|
|
qxlfb_destroy_pinned_object(gobj);
|
|
*gobj_p = NULL;
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* FIXME
|
|
* It should not be necessary to have a special dirty() callback for fbdev.
|
|
*/
|
|
static int qxlfb_framebuffer_dirty(struct drm_framebuffer *fb,
|
|
struct drm_file *file_priv,
|
|
unsigned flags, unsigned color,
|
|
struct drm_clip_rect *clips,
|
|
unsigned num_clips)
|
|
{
|
|
struct qxl_device *qdev = fb->dev->dev_private;
|
|
struct fb_info *info = qdev->fbdev_info;
|
|
struct qxl_fbdev *qfbdev = info->par;
|
|
struct qxl_fb_image qxl_fb_image;
|
|
struct fb_image *image = &qxl_fb_image.fb_image;
|
|
|
|
/* TODO: hard coding 32 bpp */
|
|
int stride = qfbdev->qfb.base.pitches[0];
|
|
|
|
/*
|
|
* we are using a shadow draw buffer, at qdev->surface0_shadow
|
|
*/
|
|
qxl_io_log(qdev, "dirty x[%d, %d], y[%d, %d]", clips->x1, clips->x2,
|
|
clips->y1, clips->y2);
|
|
image->dx = clips->x1;
|
|
image->dy = clips->y1;
|
|
image->width = clips->x2 - clips->x1;
|
|
image->height = clips->y2 - clips->y1;
|
|
image->fg_color = 0xffffffff; /* unused, just to avoid uninitialized
|
|
warnings */
|
|
image->bg_color = 0;
|
|
image->depth = 32; /* TODO: take from somewhere? */
|
|
image->cmap.start = 0;
|
|
image->cmap.len = 0;
|
|
image->cmap.red = NULL;
|
|
image->cmap.green = NULL;
|
|
image->cmap.blue = NULL;
|
|
image->cmap.transp = NULL;
|
|
image->data = qfbdev->shadow + (clips->x1 * 4) + (stride * clips->y1);
|
|
|
|
qxl_fb_image_init(&qxl_fb_image, qdev, info, NULL);
|
|
qxl_draw_opaque_fb(&qxl_fb_image, stride);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct drm_framebuffer_funcs qxlfb_fb_funcs = {
|
|
.destroy = qxl_user_framebuffer_destroy,
|
|
.dirty = qxlfb_framebuffer_dirty,
|
|
};
|
|
|
|
static int qxlfb_create(struct qxl_fbdev *qfbdev,
|
|
struct drm_fb_helper_surface_size *sizes)
|
|
{
|
|
struct qxl_device *qdev = qfbdev->qdev;
|
|
struct fb_info *info;
|
|
struct drm_framebuffer *fb = NULL;
|
|
struct drm_mode_fb_cmd2 mode_cmd;
|
|
struct drm_gem_object *gobj = NULL;
|
|
struct qxl_bo *qbo = NULL;
|
|
int ret;
|
|
int size;
|
|
int bpp = sizes->surface_bpp;
|
|
int depth = sizes->surface_depth;
|
|
void *shadow;
|
|
|
|
mode_cmd.width = sizes->surface_width;
|
|
mode_cmd.height = sizes->surface_height;
|
|
|
|
mode_cmd.pitches[0] = ALIGN(mode_cmd.width * ((bpp + 1) / 8), 64);
|
|
mode_cmd.pixel_format = drm_mode_legacy_fb_format(bpp, depth);
|
|
|
|
ret = qxlfb_create_pinned_object(qfbdev, &mode_cmd, &gobj);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
qbo = gem_to_qxl_bo(gobj);
|
|
QXL_INFO(qdev, "%s: %dx%d %d\n", __func__, mode_cmd.width,
|
|
mode_cmd.height, mode_cmd.pitches[0]);
|
|
|
|
shadow = vmalloc(mode_cmd.pitches[0] * mode_cmd.height);
|
|
/* TODO: what's the usual response to memory allocation errors? */
|
|
BUG_ON(!shadow);
|
|
QXL_INFO(qdev,
|
|
"surface0 at gpu offset %lld, mmap_offset %lld (virt %p, shadow %p)\n",
|
|
qxl_bo_gpu_offset(qbo),
|
|
qxl_bo_mmap_offset(qbo),
|
|
qbo->kptr,
|
|
shadow);
|
|
size = mode_cmd.pitches[0] * mode_cmd.height;
|
|
|
|
info = drm_fb_helper_alloc_fbi(&qfbdev->helper);
|
|
if (IS_ERR(info)) {
|
|
ret = PTR_ERR(info);
|
|
goto out_unref;
|
|
}
|
|
|
|
info->par = qfbdev;
|
|
|
|
qxl_framebuffer_init(qdev->ddev, &qfbdev->qfb, &mode_cmd, gobj,
|
|
&qxlfb_fb_funcs);
|
|
|
|
fb = &qfbdev->qfb.base;
|
|
|
|
/* setup helper with fb data */
|
|
qfbdev->helper.fb = fb;
|
|
|
|
qfbdev->shadow = shadow;
|
|
strcpy(info->fix.id, "qxldrmfb");
|
|
|
|
drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
|
|
|
|
info->flags = FBINFO_DEFAULT | FBINFO_HWACCEL_COPYAREA | FBINFO_HWACCEL_FILLRECT;
|
|
info->fbops = &qxlfb_ops;
|
|
|
|
/*
|
|
* TODO: using gobj->size in various places in this function. Not sure
|
|
* what the difference between the different sizes is.
|
|
*/
|
|
info->fix.smem_start = qdev->vram_base; /* TODO - correct? */
|
|
info->fix.smem_len = gobj->size;
|
|
info->screen_base = qfbdev->shadow;
|
|
info->screen_size = gobj->size;
|
|
|
|
drm_fb_helper_fill_var(info, &qfbdev->helper, sizes->fb_width,
|
|
sizes->fb_height);
|
|
|
|
/* setup aperture base/size for vesafb takeover */
|
|
info->apertures->ranges[0].base = qdev->ddev->mode_config.fb_base;
|
|
info->apertures->ranges[0].size = qdev->vram_size;
|
|
|
|
info->fix.mmio_start = 0;
|
|
info->fix.mmio_len = 0;
|
|
|
|
if (info->screen_base == NULL) {
|
|
ret = -ENOSPC;
|
|
goto out_destroy_fbi;
|
|
}
|
|
|
|
#ifdef CONFIG_DRM_FBDEV_EMULATION
|
|
info->fbdefio = &qxl_defio;
|
|
fb_deferred_io_init(info);
|
|
#endif
|
|
|
|
qdev->fbdev_info = info;
|
|
qdev->fbdev_qfb = &qfbdev->qfb;
|
|
DRM_INFO("fb mappable at 0x%lX, size %lu\n", info->fix.smem_start, (unsigned long)info->screen_size);
|
|
DRM_INFO("fb: depth %d, pitch %d, width %d, height %d\n", fb->depth, fb->pitches[0], fb->width, fb->height);
|
|
return 0;
|
|
|
|
out_destroy_fbi:
|
|
drm_fb_helper_release_fbi(&qfbdev->helper);
|
|
out_unref:
|
|
if (qbo) {
|
|
ret = qxl_bo_reserve(qbo, false);
|
|
if (likely(ret == 0)) {
|
|
qxl_bo_kunmap(qbo);
|
|
qxl_bo_unpin(qbo);
|
|
qxl_bo_unreserve(qbo);
|
|
}
|
|
}
|
|
if (fb && ret) {
|
|
drm_gem_object_unreference_unlocked(gobj);
|
|
drm_framebuffer_cleanup(fb);
|
|
kfree(fb);
|
|
}
|
|
drm_gem_object_unreference_unlocked(gobj);
|
|
return ret;
|
|
}
|
|
|
|
static int qxl_fb_find_or_create_single(
|
|
struct drm_fb_helper *helper,
|
|
struct drm_fb_helper_surface_size *sizes)
|
|
{
|
|
struct qxl_fbdev *qfbdev =
|
|
container_of(helper, struct qxl_fbdev, helper);
|
|
int new_fb = 0;
|
|
int ret;
|
|
|
|
if (!helper->fb) {
|
|
ret = qxlfb_create(qfbdev, sizes);
|
|
if (ret)
|
|
return ret;
|
|
new_fb = 1;
|
|
}
|
|
return new_fb;
|
|
}
|
|
|
|
static int qxl_fbdev_destroy(struct drm_device *dev, struct qxl_fbdev *qfbdev)
|
|
{
|
|
struct qxl_framebuffer *qfb = &qfbdev->qfb;
|
|
|
|
drm_fb_helper_unregister_fbi(&qfbdev->helper);
|
|
drm_fb_helper_release_fbi(&qfbdev->helper);
|
|
|
|
if (qfb->obj) {
|
|
qxlfb_destroy_pinned_object(qfb->obj);
|
|
qfb->obj = NULL;
|
|
}
|
|
drm_fb_helper_fini(&qfbdev->helper);
|
|
vfree(qfbdev->shadow);
|
|
drm_framebuffer_cleanup(&qfb->base);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct drm_fb_helper_funcs qxl_fb_helper_funcs = {
|
|
.fb_probe = qxl_fb_find_or_create_single,
|
|
};
|
|
|
|
int qxl_fbdev_init(struct qxl_device *qdev)
|
|
{
|
|
struct qxl_fbdev *qfbdev;
|
|
int bpp_sel = 32; /* TODO: parameter from somewhere? */
|
|
int ret;
|
|
|
|
qfbdev = kzalloc(sizeof(struct qxl_fbdev), GFP_KERNEL);
|
|
if (!qfbdev)
|
|
return -ENOMEM;
|
|
|
|
qfbdev->qdev = qdev;
|
|
qdev->mode_info.qfbdev = qfbdev;
|
|
spin_lock_init(&qfbdev->delayed_ops_lock);
|
|
INIT_LIST_HEAD(&qfbdev->delayed_ops);
|
|
|
|
drm_fb_helper_prepare(qdev->ddev, &qfbdev->helper,
|
|
&qxl_fb_helper_funcs);
|
|
|
|
ret = drm_fb_helper_init(qdev->ddev, &qfbdev->helper,
|
|
qxl_num_crtc /* num_crtc - QXL supports just 1 */,
|
|
QXLFB_CONN_LIMIT);
|
|
if (ret)
|
|
goto free;
|
|
|
|
ret = drm_fb_helper_single_add_all_connectors(&qfbdev->helper);
|
|
if (ret)
|
|
goto fini;
|
|
|
|
ret = drm_fb_helper_initial_config(&qfbdev->helper, bpp_sel);
|
|
if (ret)
|
|
goto fini;
|
|
|
|
return 0;
|
|
|
|
fini:
|
|
drm_fb_helper_fini(&qfbdev->helper);
|
|
free:
|
|
kfree(qfbdev);
|
|
return ret;
|
|
}
|
|
|
|
void qxl_fbdev_fini(struct qxl_device *qdev)
|
|
{
|
|
if (!qdev->mode_info.qfbdev)
|
|
return;
|
|
|
|
qxl_fbdev_destroy(qdev->ddev, qdev->mode_info.qfbdev);
|
|
kfree(qdev->mode_info.qfbdev);
|
|
qdev->mode_info.qfbdev = NULL;
|
|
}
|
|
|
|
void qxl_fbdev_set_suspend(struct qxl_device *qdev, int state)
|
|
{
|
|
drm_fb_helper_set_suspend(&qdev->mode_info.qfbdev->helper, state);
|
|
}
|
|
|
|
bool qxl_fbdev_qobj_is_fb(struct qxl_device *qdev, struct qxl_bo *qobj)
|
|
{
|
|
if (qobj == gem_to_qxl_bo(qdev->mode_info.qfbdev->qfb.obj))
|
|
return true;
|
|
return false;
|
|
}
|