forked from Minki/linux
1169af2180
This patch fixes memory alloction(contiguous or not) and cache mapping types(cachable or not). For this, it converts each type from user request into dma attribute properly. Changelog v2: - just code cleanup. Signed-off-by: Inki Dae <inki.dae@samsung.com> Signed-off-by: Kyungmin Park <kyungmin.park@samsung.com>
174 lines
4.4 KiB
C
174 lines
4.4 KiB
C
/* exynos_drm_buf.c
|
|
*
|
|
* Copyright (c) 2011 Samsung Electronics Co., Ltd.
|
|
* Author: Inki Dae <inki.dae@samsung.com>
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
|
|
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
|
|
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
|
* OTHER DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <drm/drmP.h>
|
|
#include <drm/exynos_drm.h>
|
|
|
|
#include "exynos_drm_drv.h"
|
|
#include "exynos_drm_gem.h"
|
|
#include "exynos_drm_buf.h"
|
|
|
|
static int lowlevel_buffer_allocate(struct drm_device *dev,
|
|
unsigned int flags, struct exynos_drm_gem_buf *buf)
|
|
{
|
|
int ret = 0;
|
|
enum dma_attr attr;
|
|
unsigned int nr_pages;
|
|
|
|
DRM_DEBUG_KMS("%s\n", __FILE__);
|
|
|
|
if (buf->dma_addr) {
|
|
DRM_DEBUG_KMS("already allocated.\n");
|
|
return 0;
|
|
}
|
|
|
|
init_dma_attrs(&buf->dma_attrs);
|
|
|
|
/*
|
|
* if EXYNOS_BO_CONTIG, fully physically contiguous memory
|
|
* region will be allocated else physically contiguous
|
|
* as possible.
|
|
*/
|
|
if (flags & EXYNOS_BO_CONTIG)
|
|
dma_set_attr(DMA_ATTR_FORCE_CONTIGUOUS, &buf->dma_attrs);
|
|
|
|
/*
|
|
* if EXYNOS_BO_WC or EXYNOS_BO_NONCACHABLE, writecombine mapping
|
|
* else cachable mapping.
|
|
*/
|
|
if (flags & EXYNOS_BO_WC || !(flags & EXYNOS_BO_CACHABLE))
|
|
attr = DMA_ATTR_WRITE_COMBINE;
|
|
else
|
|
attr = DMA_ATTR_NON_CONSISTENT;
|
|
|
|
dma_set_attr(attr, &buf->dma_attrs);
|
|
dma_set_attr(DMA_ATTR_NO_KERNEL_MAPPING, &buf->dma_attrs);
|
|
|
|
buf->pages = dma_alloc_attrs(dev->dev, buf->size,
|
|
&buf->dma_addr, GFP_KERNEL, &buf->dma_attrs);
|
|
if (!buf->pages) {
|
|
DRM_ERROR("failed to allocate buffer.\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
nr_pages = buf->size >> PAGE_SHIFT;
|
|
buf->sgt = drm_prime_pages_to_sg(buf->pages, nr_pages);
|
|
if (!buf->sgt) {
|
|
DRM_ERROR("failed to get sg table.\n");
|
|
ret = -ENOMEM;
|
|
goto err_free_attrs;
|
|
}
|
|
|
|
DRM_DEBUG_KMS("dma_addr(0x%lx), size(0x%lx)\n",
|
|
(unsigned long)buf->dma_addr,
|
|
buf->size);
|
|
|
|
return ret;
|
|
|
|
err_free_attrs:
|
|
dma_free_attrs(dev->dev, buf->size, buf->pages,
|
|
(dma_addr_t)buf->dma_addr, &buf->dma_attrs);
|
|
buf->dma_addr = (dma_addr_t)NULL;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void lowlevel_buffer_deallocate(struct drm_device *dev,
|
|
unsigned int flags, struct exynos_drm_gem_buf *buf)
|
|
{
|
|
DRM_DEBUG_KMS("%s.\n", __FILE__);
|
|
|
|
if (!buf->dma_addr) {
|
|
DRM_DEBUG_KMS("dma_addr is invalid.\n");
|
|
return;
|
|
}
|
|
|
|
DRM_DEBUG_KMS("dma_addr(0x%lx), size(0x%lx)\n",
|
|
(unsigned long)buf->dma_addr,
|
|
buf->size);
|
|
|
|
sg_free_table(buf->sgt);
|
|
|
|
kfree(buf->sgt);
|
|
buf->sgt = NULL;
|
|
|
|
dma_free_attrs(dev->dev, buf->size, buf->pages,
|
|
(dma_addr_t)buf->dma_addr, &buf->dma_attrs);
|
|
buf->dma_addr = (dma_addr_t)NULL;
|
|
}
|
|
|
|
struct exynos_drm_gem_buf *exynos_drm_init_buf(struct drm_device *dev,
|
|
unsigned int size)
|
|
{
|
|
struct exynos_drm_gem_buf *buffer;
|
|
|
|
DRM_DEBUG_KMS("%s.\n", __FILE__);
|
|
DRM_DEBUG_KMS("desired size = 0x%x\n", size);
|
|
|
|
buffer = kzalloc(sizeof(*buffer), GFP_KERNEL);
|
|
if (!buffer) {
|
|
DRM_ERROR("failed to allocate exynos_drm_gem_buf.\n");
|
|
return NULL;
|
|
}
|
|
|
|
buffer->size = size;
|
|
return buffer;
|
|
}
|
|
|
|
void exynos_drm_fini_buf(struct drm_device *dev,
|
|
struct exynos_drm_gem_buf *buffer)
|
|
{
|
|
DRM_DEBUG_KMS("%s.\n", __FILE__);
|
|
|
|
if (!buffer) {
|
|
DRM_DEBUG_KMS("buffer is null.\n");
|
|
return;
|
|
}
|
|
|
|
kfree(buffer);
|
|
buffer = NULL;
|
|
}
|
|
|
|
int exynos_drm_alloc_buf(struct drm_device *dev,
|
|
struct exynos_drm_gem_buf *buf, unsigned int flags)
|
|
{
|
|
|
|
/*
|
|
* allocate memory region and set the memory information
|
|
* to vaddr and dma_addr of a buffer object.
|
|
*/
|
|
if (lowlevel_buffer_allocate(dev, flags, buf) < 0)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void exynos_drm_free_buf(struct drm_device *dev,
|
|
unsigned int flags, struct exynos_drm_gem_buf *buffer)
|
|
{
|
|
|
|
lowlevel_buffer_deallocate(dev, flags, buffer);
|
|
}
|