This patch updates dma_buf_vunmap() and dma-buf's vunmap callback to use struct dma_buf_map. The interfaces used to receive a buffer address. This address is now given in an instance of the structure. Users of the functions are updated accordingly. This is only an interface change. It is currently expected that dma-buf memory can be accessed with system memory load/store operations. v2: * include dma-buf-heaps and i915 selftests (kernel test robot) * initialize cma_obj before using it in drm_gem_cma_free_object() (kernel test robot) Signed-off-by: Thomas Zimmermann <tzimmermann@suse.de> Acked-by: Sumit Semwal <sumit.semwal@linaro.org> Acked-by: Christian König <christian.koenig@amd.com> Acked-by: Daniel Vetter <daniel.vetter@ffwll.ch> Acked-by: Tomasz Figa <tfiga@chromium.org> Acked-by: Sam Ravnborg <sam@ravnborg.org> Link: https://patchwork.freedesktop.org/patch/msgid/20200925115601.23955-4-tzimmermann@suse.de
293 lines
6.1 KiB
C
293 lines
6.1 KiB
C
/*
|
|
* SPDX-License-Identifier: MIT
|
|
*
|
|
* Copyright © 2016 Intel Corporation
|
|
*/
|
|
|
|
#include "i915_drv.h"
|
|
#include "i915_selftest.h"
|
|
|
|
#include "mock_dmabuf.h"
|
|
#include "selftests/mock_gem_device.h"
|
|
|
|
static int igt_dmabuf_export(void *arg)
|
|
{
|
|
struct drm_i915_private *i915 = arg;
|
|
struct drm_i915_gem_object *obj;
|
|
struct dma_buf *dmabuf;
|
|
|
|
obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
|
|
if (IS_ERR(obj))
|
|
return PTR_ERR(obj);
|
|
|
|
dmabuf = i915_gem_prime_export(&obj->base, 0);
|
|
i915_gem_object_put(obj);
|
|
if (IS_ERR(dmabuf)) {
|
|
pr_err("i915_gem_prime_export failed with err=%d\n",
|
|
(int)PTR_ERR(dmabuf));
|
|
return PTR_ERR(dmabuf);
|
|
}
|
|
|
|
dma_buf_put(dmabuf);
|
|
return 0;
|
|
}
|
|
|
|
static int igt_dmabuf_import_self(void *arg)
|
|
{
|
|
struct drm_i915_private *i915 = arg;
|
|
struct drm_i915_gem_object *obj;
|
|
struct drm_gem_object *import;
|
|
struct dma_buf *dmabuf;
|
|
int err;
|
|
|
|
obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
|
|
if (IS_ERR(obj))
|
|
return PTR_ERR(obj);
|
|
|
|
dmabuf = i915_gem_prime_export(&obj->base, 0);
|
|
if (IS_ERR(dmabuf)) {
|
|
pr_err("i915_gem_prime_export failed with err=%d\n",
|
|
(int)PTR_ERR(dmabuf));
|
|
err = PTR_ERR(dmabuf);
|
|
goto out;
|
|
}
|
|
|
|
import = i915_gem_prime_import(&i915->drm, dmabuf);
|
|
if (IS_ERR(import)) {
|
|
pr_err("i915_gem_prime_import failed with err=%d\n",
|
|
(int)PTR_ERR(import));
|
|
err = PTR_ERR(import);
|
|
goto out_dmabuf;
|
|
}
|
|
|
|
if (import != &obj->base) {
|
|
pr_err("i915_gem_prime_import created a new object!\n");
|
|
err = -EINVAL;
|
|
goto out_import;
|
|
}
|
|
|
|
err = 0;
|
|
out_import:
|
|
i915_gem_object_put(to_intel_bo(import));
|
|
out_dmabuf:
|
|
dma_buf_put(dmabuf);
|
|
out:
|
|
i915_gem_object_put(obj);
|
|
return err;
|
|
}
|
|
|
|
static int igt_dmabuf_import(void *arg)
|
|
{
|
|
struct drm_i915_private *i915 = arg;
|
|
struct drm_i915_gem_object *obj;
|
|
struct dma_buf *dmabuf;
|
|
void *obj_map, *dma_map;
|
|
struct dma_buf_map map;
|
|
u32 pattern[] = { 0, 0xaa, 0xcc, 0x55, 0xff };
|
|
int err, i;
|
|
|
|
dmabuf = mock_dmabuf(1);
|
|
if (IS_ERR(dmabuf))
|
|
return PTR_ERR(dmabuf);
|
|
|
|
obj = to_intel_bo(i915_gem_prime_import(&i915->drm, dmabuf));
|
|
if (IS_ERR(obj)) {
|
|
pr_err("i915_gem_prime_import failed with err=%d\n",
|
|
(int)PTR_ERR(obj));
|
|
err = PTR_ERR(obj);
|
|
goto out_dmabuf;
|
|
}
|
|
|
|
if (obj->base.dev != &i915->drm) {
|
|
pr_err("i915_gem_prime_import created a non-i915 object!\n");
|
|
err = -EINVAL;
|
|
goto out_obj;
|
|
}
|
|
|
|
if (obj->base.size != PAGE_SIZE) {
|
|
pr_err("i915_gem_prime_import is wrong size found %lld, expected %ld\n",
|
|
(long long)obj->base.size, PAGE_SIZE);
|
|
err = -EINVAL;
|
|
goto out_obj;
|
|
}
|
|
|
|
err = dma_buf_vmap(dmabuf, &map);
|
|
dma_map = err ? NULL : map.vaddr;
|
|
if (!dma_map) {
|
|
pr_err("dma_buf_vmap failed\n");
|
|
err = -ENOMEM;
|
|
goto out_obj;
|
|
}
|
|
|
|
if (0) { /* Can not yet map dmabuf */
|
|
obj_map = i915_gem_object_pin_map(obj, I915_MAP_WB);
|
|
if (IS_ERR(obj_map)) {
|
|
err = PTR_ERR(obj_map);
|
|
pr_err("i915_gem_object_pin_map failed with err=%d\n", err);
|
|
goto out_dma_map;
|
|
}
|
|
|
|
for (i = 0; i < ARRAY_SIZE(pattern); i++) {
|
|
memset(dma_map, pattern[i], PAGE_SIZE);
|
|
if (memchr_inv(obj_map, pattern[i], PAGE_SIZE)) {
|
|
err = -EINVAL;
|
|
pr_err("imported vmap not all set to %x!\n", pattern[i]);
|
|
i915_gem_object_unpin_map(obj);
|
|
goto out_dma_map;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < ARRAY_SIZE(pattern); i++) {
|
|
memset(obj_map, pattern[i], PAGE_SIZE);
|
|
if (memchr_inv(dma_map, pattern[i], PAGE_SIZE)) {
|
|
err = -EINVAL;
|
|
pr_err("exported vmap not all set to %x!\n", pattern[i]);
|
|
i915_gem_object_unpin_map(obj);
|
|
goto out_dma_map;
|
|
}
|
|
}
|
|
|
|
i915_gem_object_unpin_map(obj);
|
|
}
|
|
|
|
err = 0;
|
|
out_dma_map:
|
|
dma_buf_vunmap(dmabuf, &map);
|
|
out_obj:
|
|
i915_gem_object_put(obj);
|
|
out_dmabuf:
|
|
dma_buf_put(dmabuf);
|
|
return err;
|
|
}
|
|
|
|
static int igt_dmabuf_import_ownership(void *arg)
|
|
{
|
|
struct drm_i915_private *i915 = arg;
|
|
struct drm_i915_gem_object *obj;
|
|
struct dma_buf *dmabuf;
|
|
struct dma_buf_map map;
|
|
void *ptr;
|
|
int err;
|
|
|
|
dmabuf = mock_dmabuf(1);
|
|
if (IS_ERR(dmabuf))
|
|
return PTR_ERR(dmabuf);
|
|
|
|
err = dma_buf_vmap(dmabuf, &map);
|
|
ptr = err ? NULL : map.vaddr;
|
|
if (!ptr) {
|
|
pr_err("dma_buf_vmap failed\n");
|
|
err = -ENOMEM;
|
|
goto err_dmabuf;
|
|
}
|
|
|
|
memset(ptr, 0xc5, PAGE_SIZE);
|
|
dma_buf_vunmap(dmabuf, &map);
|
|
|
|
obj = to_intel_bo(i915_gem_prime_import(&i915->drm, dmabuf));
|
|
if (IS_ERR(obj)) {
|
|
pr_err("i915_gem_prime_import failed with err=%d\n",
|
|
(int)PTR_ERR(obj));
|
|
err = PTR_ERR(obj);
|
|
goto err_dmabuf;
|
|
}
|
|
|
|
dma_buf_put(dmabuf);
|
|
|
|
err = i915_gem_object_pin_pages(obj);
|
|
if (err) {
|
|
pr_err("i915_gem_object_pin_pages failed with err=%d\n", err);
|
|
goto out_obj;
|
|
}
|
|
|
|
err = 0;
|
|
i915_gem_object_unpin_pages(obj);
|
|
out_obj:
|
|
i915_gem_object_put(obj);
|
|
return err;
|
|
|
|
err_dmabuf:
|
|
dma_buf_put(dmabuf);
|
|
return err;
|
|
}
|
|
|
|
static int igt_dmabuf_export_vmap(void *arg)
|
|
{
|
|
struct drm_i915_private *i915 = arg;
|
|
struct drm_i915_gem_object *obj;
|
|
struct dma_buf *dmabuf;
|
|
struct dma_buf_map map;
|
|
void *ptr;
|
|
int err;
|
|
|
|
obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
|
|
if (IS_ERR(obj))
|
|
return PTR_ERR(obj);
|
|
|
|
dmabuf = i915_gem_prime_export(&obj->base, 0);
|
|
if (IS_ERR(dmabuf)) {
|
|
pr_err("i915_gem_prime_export failed with err=%d\n",
|
|
(int)PTR_ERR(dmabuf));
|
|
err = PTR_ERR(dmabuf);
|
|
goto err_obj;
|
|
}
|
|
i915_gem_object_put(obj);
|
|
|
|
err = dma_buf_vmap(dmabuf, &map);
|
|
ptr = err ? NULL : map.vaddr;
|
|
if (!ptr) {
|
|
pr_err("dma_buf_vmap failed\n");
|
|
err = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
if (memchr_inv(ptr, 0, dmabuf->size)) {
|
|
pr_err("Exported object not initialiased to zero!\n");
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
memset(ptr, 0xc5, dmabuf->size);
|
|
|
|
err = 0;
|
|
dma_buf_vunmap(dmabuf, &map);
|
|
out:
|
|
dma_buf_put(dmabuf);
|
|
return err;
|
|
|
|
err_obj:
|
|
i915_gem_object_put(obj);
|
|
return err;
|
|
}
|
|
|
|
int i915_gem_dmabuf_mock_selftests(void)
|
|
{
|
|
static const struct i915_subtest tests[] = {
|
|
SUBTEST(igt_dmabuf_export),
|
|
SUBTEST(igt_dmabuf_import_self),
|
|
SUBTEST(igt_dmabuf_import),
|
|
SUBTEST(igt_dmabuf_import_ownership),
|
|
SUBTEST(igt_dmabuf_export_vmap),
|
|
};
|
|
struct drm_i915_private *i915;
|
|
int err;
|
|
|
|
i915 = mock_gem_device();
|
|
if (!i915)
|
|
return -ENOMEM;
|
|
|
|
err = i915_subtests(tests, i915);
|
|
|
|
mock_destroy_device(i915);
|
|
return err;
|
|
}
|
|
|
|
int i915_gem_dmabuf_live_selftests(struct drm_i915_private *i915)
|
|
{
|
|
static const struct i915_subtest tests[] = {
|
|
SUBTEST(igt_dmabuf_export),
|
|
};
|
|
|
|
return i915_subtests(tests, i915);
|
|
}
|