drm/msm: implement a2xx mmu
A2XX has its own very simple MMU. Added a msm_use_mmu() function because we can't rely on iommu_present to decide to use MMU or not. Signed-off-by: Jonathan Marek <jonathan@marek.ca> Signed-off-by: Rob Clark <robdclark@gmail.com>
This commit is contained in:
parent
d1d9d0e172
commit
c2052a4e5c
@ -93,7 +93,8 @@ msm-y := \
|
|||||||
msm_rd.o \
|
msm_rd.o \
|
||||||
msm_ringbuffer.o \
|
msm_ringbuffer.o \
|
||||||
msm_submitqueue.o \
|
msm_submitqueue.o \
|
||||||
msm_gpu_tracepoints.o
|
msm_gpu_tracepoints.o \
|
||||||
|
msm_gpummu.o
|
||||||
|
|
||||||
msm-$(CONFIG_DEBUG_FS) += adreno/a5xx_debugfs.o \
|
msm-$(CONFIG_DEBUG_FS) += adreno/a5xx_debugfs.o \
|
||||||
disp/dpu1/dpu_dbg.o
|
disp/dpu1/dpu_dbg.o
|
||||||
|
@ -2,6 +2,8 @@
|
|||||||
/* Copyright (c) 2018 The Linux Foundation. All rights reserved. */
|
/* Copyright (c) 2018 The Linux Foundation. All rights reserved. */
|
||||||
|
|
||||||
#include "a2xx_gpu.h"
|
#include "a2xx_gpu.h"
|
||||||
|
#include "msm_gem.h"
|
||||||
|
#include "msm_mmu.h"
|
||||||
|
|
||||||
extern bool hang_debug;
|
extern bool hang_debug;
|
||||||
|
|
||||||
@ -58,9 +60,12 @@ static bool a2xx_me_init(struct msm_gpu *gpu)
|
|||||||
static int a2xx_hw_init(struct msm_gpu *gpu)
|
static int a2xx_hw_init(struct msm_gpu *gpu)
|
||||||
{
|
{
|
||||||
struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
|
struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
|
||||||
|
dma_addr_t pt_base, tran_error;
|
||||||
uint32_t *ptr, len;
|
uint32_t *ptr, len;
|
||||||
int i, ret;
|
int i, ret;
|
||||||
|
|
||||||
|
msm_gpummu_params(gpu->aspace->mmu, &pt_base, &tran_error);
|
||||||
|
|
||||||
DBG("%s", gpu->name);
|
DBG("%s", gpu->name);
|
||||||
|
|
||||||
/* halt ME to avoid ucode upload issues on a20x */
|
/* halt ME to avoid ucode upload issues on a20x */
|
||||||
@ -80,9 +85,34 @@ static int a2xx_hw_init(struct msm_gpu *gpu)
|
|||||||
/* note: kgsl uses 0x0000ffff for a20x */
|
/* note: kgsl uses 0x0000ffff for a20x */
|
||||||
gpu_write(gpu, REG_A2XX_RBBM_CNTL, 0x00004442);
|
gpu_write(gpu, REG_A2XX_RBBM_CNTL, 0x00004442);
|
||||||
|
|
||||||
gpu_write(gpu, REG_A2XX_MH_MMU_CONFIG, 0);
|
/* MPU: physical range */
|
||||||
gpu_write(gpu, REG_A2XX_MH_MMU_MPU_BASE, 0);
|
gpu_write(gpu, REG_A2XX_MH_MMU_MPU_BASE, 0x00000000);
|
||||||
gpu_write(gpu, REG_A2XX_MH_MMU_MPU_END, 0xfffff000);
|
gpu_write(gpu, REG_A2XX_MH_MMU_MPU_END, 0xfffff000);
|
||||||
|
|
||||||
|
gpu_write(gpu, REG_A2XX_MH_MMU_CONFIG, A2XX_MH_MMU_CONFIG_MMU_ENABLE |
|
||||||
|
A2XX_MH_MMU_CONFIG_RB_W_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_CP_W_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_CP_R0_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_CP_R1_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_CP_R2_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_CP_R3_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_CP_R4_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_VGT_R0_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_VGT_R1_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_TC_R_CLNT_BEHAVIOR(BEH_TRAN_RNG) |
|
||||||
|
A2XX_MH_MMU_CONFIG_PA_W_CLNT_BEHAVIOR(BEH_TRAN_RNG));
|
||||||
|
|
||||||
|
/* same as parameters in adreno_gpu */
|
||||||
|
gpu_write(gpu, REG_A2XX_MH_MMU_VA_RANGE, SZ_16M |
|
||||||
|
A2XX_MH_MMU_VA_RANGE_NUM_64KB_REGIONS(0xfff));
|
||||||
|
|
||||||
|
gpu_write(gpu, REG_A2XX_MH_MMU_PT_BASE, pt_base);
|
||||||
|
gpu_write(gpu, REG_A2XX_MH_MMU_TRAN_ERROR, tran_error);
|
||||||
|
|
||||||
|
gpu_write(gpu, REG_A2XX_MH_MMU_INVALIDATE,
|
||||||
|
A2XX_MH_MMU_INVALIDATE_INVALIDATE_ALL |
|
||||||
|
A2XX_MH_MMU_INVALIDATE_INVALIDATE_TC);
|
||||||
|
|
||||||
gpu_write(gpu, REG_A2XX_MH_ARBITER_CONFIG,
|
gpu_write(gpu, REG_A2XX_MH_ARBITER_CONFIG,
|
||||||
A2XX_MH_ARBITER_CONFIG_SAME_PAGE_LIMIT(16) |
|
A2XX_MH_ARBITER_CONFIG_SAME_PAGE_LIMIT(16) |
|
||||||
A2XX_MH_ARBITER_CONFIG_L1_ARB_ENABLE |
|
A2XX_MH_ARBITER_CONFIG_L1_ARB_ENABLE |
|
||||||
@ -109,9 +139,21 @@ static int a2xx_hw_init(struct msm_gpu *gpu)
|
|||||||
/* note: gsl doesn't set this */
|
/* note: gsl doesn't set this */
|
||||||
gpu_write(gpu, REG_A2XX_RBBM_DEBUG, 0x00080000);
|
gpu_write(gpu, REG_A2XX_RBBM_DEBUG, 0x00080000);
|
||||||
|
|
||||||
gpu_write(gpu, REG_A2XX_RBBM_INT_CNTL, 0);
|
gpu_write(gpu, REG_A2XX_RBBM_INT_CNTL,
|
||||||
gpu_write(gpu, REG_AXXX_CP_INT_CNTL, 0x80000000); /* RB INT */
|
A2XX_RBBM_INT_CNTL_RDERR_INT_MASK);
|
||||||
|
gpu_write(gpu, REG_AXXX_CP_INT_CNTL,
|
||||||
|
AXXX_CP_INT_CNTL_T0_PACKET_IN_IB_MASK |
|
||||||
|
AXXX_CP_INT_CNTL_OPCODE_ERROR_MASK |
|
||||||
|
AXXX_CP_INT_CNTL_PROTECTED_MODE_ERROR_MASK |
|
||||||
|
AXXX_CP_INT_CNTL_RESERVED_BIT_ERROR_MASK |
|
||||||
|
AXXX_CP_INT_CNTL_IB_ERROR_MASK |
|
||||||
|
AXXX_CP_INT_CNTL_IB1_INT_MASK |
|
||||||
|
AXXX_CP_INT_CNTL_RB_INT_MASK);
|
||||||
gpu_write(gpu, REG_A2XX_SQ_INT_CNTL, 0);
|
gpu_write(gpu, REG_A2XX_SQ_INT_CNTL, 0);
|
||||||
|
gpu_write(gpu, REG_A2XX_MH_INTERRUPT_MASK,
|
||||||
|
A2XX_MH_INTERRUPT_MASK_AXI_READ_ERROR |
|
||||||
|
A2XX_MH_INTERRUPT_MASK_AXI_WRITE_ERROR |
|
||||||
|
A2XX_MH_INTERRUPT_MASK_MMU_PAGE_FAULT);
|
||||||
|
|
||||||
for (i = 3; i <= 5; i++)
|
for (i = 3; i <= 5; i++)
|
||||||
if ((SZ_16K << i) == adreno_gpu->gmem)
|
if ((SZ_16K << i) == adreno_gpu->gmem)
|
||||||
|
@ -307,6 +307,7 @@ static int adreno_bind(struct device *dev, struct device *master, void *data)
|
|||||||
static struct adreno_platform_config config = {};
|
static struct adreno_platform_config config = {};
|
||||||
const struct adreno_info *info;
|
const struct adreno_info *info;
|
||||||
struct drm_device *drm = dev_get_drvdata(master);
|
struct drm_device *drm = dev_get_drvdata(master);
|
||||||
|
struct msm_drm_private *priv = drm->dev_private;
|
||||||
struct msm_gpu *gpu;
|
struct msm_gpu *gpu;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
@ -329,6 +330,8 @@ static int adreno_bind(struct device *dev, struct device *master, void *data)
|
|||||||
DBG("Found GPU: %u.%u.%u.%u", config.rev.core, config.rev.major,
|
DBG("Found GPU: %u.%u.%u.%u", config.rev.core, config.rev.major,
|
||||||
config.rev.minor, config.rev.patchid);
|
config.rev.minor, config.rev.patchid);
|
||||||
|
|
||||||
|
priv->is_a2xx = config.rev.core == 2;
|
||||||
|
|
||||||
gpu = info->init(drm);
|
gpu = info->init(drm);
|
||||||
if (IS_ERR(gpu)) {
|
if (IS_ERR(gpu)) {
|
||||||
dev_warn(drm->dev, "failed to load adreno gpu\n");
|
dev_warn(drm->dev, "failed to load adreno gpu\n");
|
||||||
|
@ -769,6 +769,9 @@ int adreno_gpu_init(struct drm_device *drm, struct platform_device *pdev,
|
|||||||
|
|
||||||
adreno_gpu_config.va_start = SZ_16M;
|
adreno_gpu_config.va_start = SZ_16M;
|
||||||
adreno_gpu_config.va_end = 0xffffffff;
|
adreno_gpu_config.va_end = 0xffffffff;
|
||||||
|
/* maximum range of a2xx mmu */
|
||||||
|
if (adreno_is_a2xx(adreno_gpu))
|
||||||
|
adreno_gpu_config.va_end = SZ_16M + 0xfff * SZ_64K;
|
||||||
|
|
||||||
adreno_gpu_config.nr_rings = nr_rings;
|
adreno_gpu_config.nr_rings = nr_rings;
|
||||||
|
|
||||||
|
@ -26,6 +26,7 @@
|
|||||||
#include "msm_gem.h"
|
#include "msm_gem.h"
|
||||||
#include "msm_gpu.h"
|
#include "msm_gpu.h"
|
||||||
#include "msm_kms.h"
|
#include "msm_kms.h"
|
||||||
|
#include "adreno/adreno_gpu.h"
|
||||||
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -361,6 +362,14 @@ static int get_mdp_ver(struct platform_device *pdev)
|
|||||||
|
|
||||||
#include <linux/of_address.h>
|
#include <linux/of_address.h>
|
||||||
|
|
||||||
|
bool msm_use_mmu(struct drm_device *dev)
|
||||||
|
{
|
||||||
|
struct msm_drm_private *priv = dev->dev_private;
|
||||||
|
|
||||||
|
/* a2xx comes with its own MMU */
|
||||||
|
return priv->is_a2xx || iommu_present(&platform_bus_type);
|
||||||
|
}
|
||||||
|
|
||||||
static int msm_init_vram(struct drm_device *dev)
|
static int msm_init_vram(struct drm_device *dev)
|
||||||
{
|
{
|
||||||
struct msm_drm_private *priv = dev->dev_private;
|
struct msm_drm_private *priv = dev->dev_private;
|
||||||
@ -399,7 +408,7 @@ static int msm_init_vram(struct drm_device *dev)
|
|||||||
* Grab the entire CMA chunk carved out in early startup in
|
* Grab the entire CMA chunk carved out in early startup in
|
||||||
* mach-msm:
|
* mach-msm:
|
||||||
*/
|
*/
|
||||||
} else if (!iommu_present(&platform_bus_type)) {
|
} else if (!msm_use_mmu(dev)) {
|
||||||
DRM_INFO("using %s VRAM carveout\n", vram);
|
DRM_INFO("using %s VRAM carveout\n", vram);
|
||||||
size = memparse(vram, NULL);
|
size = memparse(vram, NULL);
|
||||||
}
|
}
|
||||||
|
@ -179,6 +179,8 @@ struct msm_drm_private {
|
|||||||
/* when we have more than one 'msm_gpu' these need to be an array: */
|
/* when we have more than one 'msm_gpu' these need to be an array: */
|
||||||
struct msm_gpu *gpu;
|
struct msm_gpu *gpu;
|
||||||
struct msm_file_private *lastctx;
|
struct msm_file_private *lastctx;
|
||||||
|
/* gpu is only set on open(), but we need this info earlier */
|
||||||
|
bool is_a2xx;
|
||||||
|
|
||||||
struct drm_fb_helper *fbdev;
|
struct drm_fb_helper *fbdev;
|
||||||
|
|
||||||
@ -258,9 +260,15 @@ struct msm_gem_address_space *
|
|||||||
msm_gem_address_space_create(struct device *dev, struct iommu_domain *domain,
|
msm_gem_address_space_create(struct device *dev, struct iommu_domain *domain,
|
||||||
const char *name);
|
const char *name);
|
||||||
|
|
||||||
|
struct msm_gem_address_space *
|
||||||
|
msm_gem_address_space_create_a2xx(struct device *dev, struct msm_gpu *gpu,
|
||||||
|
const char *name, uint64_t va_start, uint64_t va_end);
|
||||||
|
|
||||||
int msm_register_mmu(struct drm_device *dev, struct msm_mmu *mmu);
|
int msm_register_mmu(struct drm_device *dev, struct msm_mmu *mmu);
|
||||||
void msm_unregister_mmu(struct drm_device *dev, struct msm_mmu *mmu);
|
void msm_unregister_mmu(struct drm_device *dev, struct msm_mmu *mmu);
|
||||||
|
|
||||||
|
bool msm_use_mmu(struct drm_device *dev);
|
||||||
|
|
||||||
void msm_gem_submit_free(struct msm_gem_submit *submit);
|
void msm_gem_submit_free(struct msm_gem_submit *submit);
|
||||||
int msm_ioctl_gem_submit(struct drm_device *dev, void *data,
|
int msm_ioctl_gem_submit(struct drm_device *dev, void *data,
|
||||||
struct drm_file *file);
|
struct drm_file *file);
|
||||||
|
@ -975,7 +975,7 @@ static struct drm_gem_object *_msm_gem_new(struct drm_device *dev,
|
|||||||
|
|
||||||
size = PAGE_ALIGN(size);
|
size = PAGE_ALIGN(size);
|
||||||
|
|
||||||
if (!iommu_present(&platform_bus_type))
|
if (!msm_use_mmu(dev))
|
||||||
use_vram = true;
|
use_vram = true;
|
||||||
else if ((flags & (MSM_BO_STOLEN | MSM_BO_SCANOUT)) && priv->vram.size)
|
else if ((flags & (MSM_BO_STOLEN | MSM_BO_SCANOUT)) && priv->vram.size)
|
||||||
use_vram = true;
|
use_vram = true;
|
||||||
@ -1052,7 +1052,7 @@ struct drm_gem_object *msm_gem_import(struct drm_device *dev,
|
|||||||
int ret, npages;
|
int ret, npages;
|
||||||
|
|
||||||
/* if we don't have IOMMU, don't bother pretending we can import: */
|
/* if we don't have IOMMU, don't bother pretending we can import: */
|
||||||
if (!iommu_present(&platform_bus_type)) {
|
if (!msm_use_mmu(dev)) {
|
||||||
DRM_DEV_ERROR(dev->dev, "cannot import without IOMMU\n");
|
DRM_DEV_ERROR(dev->dev, "cannot import without IOMMU\n");
|
||||||
return ERR_PTR(-EINVAL);
|
return ERR_PTR(-EINVAL);
|
||||||
}
|
}
|
||||||
|
@ -159,3 +159,26 @@ msm_gem_address_space_create(struct device *dev, struct iommu_domain *domain,
|
|||||||
|
|
||||||
return aspace;
|
return aspace;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
struct msm_gem_address_space *
|
||||||
|
msm_gem_address_space_create_a2xx(struct device *dev, struct msm_gpu *gpu,
|
||||||
|
const char *name, uint64_t va_start, uint64_t va_end)
|
||||||
|
{
|
||||||
|
struct msm_gem_address_space *aspace;
|
||||||
|
u64 size = va_end - va_start;
|
||||||
|
|
||||||
|
aspace = kzalloc(sizeof(*aspace), GFP_KERNEL);
|
||||||
|
if (!aspace)
|
||||||
|
return ERR_PTR(-ENOMEM);
|
||||||
|
|
||||||
|
spin_lock_init(&aspace->lock);
|
||||||
|
aspace->name = name;
|
||||||
|
aspace->mmu = msm_gpummu_new(dev, gpu);
|
||||||
|
|
||||||
|
drm_mm_init(&aspace->mm, (va_start >> PAGE_SHIFT),
|
||||||
|
size >> PAGE_SHIFT);
|
||||||
|
|
||||||
|
kref_init(&aspace->kref);
|
||||||
|
|
||||||
|
return aspace;
|
||||||
|
}
|
||||||
|
@ -20,6 +20,7 @@
|
|||||||
#include "msm_mmu.h"
|
#include "msm_mmu.h"
|
||||||
#include "msm_fence.h"
|
#include "msm_fence.h"
|
||||||
#include "msm_gpu_trace.h"
|
#include "msm_gpu_trace.h"
|
||||||
|
#include "adreno/adreno_gpu.h"
|
||||||
|
|
||||||
#include <generated/utsrelease.h>
|
#include <generated/utsrelease.h>
|
||||||
#include <linux/string_helpers.h>
|
#include <linux/string_helpers.h>
|
||||||
@ -822,7 +823,6 @@ static struct msm_gem_address_space *
|
|||||||
msm_gpu_create_address_space(struct msm_gpu *gpu, struct platform_device *pdev,
|
msm_gpu_create_address_space(struct msm_gpu *gpu, struct platform_device *pdev,
|
||||||
uint64_t va_start, uint64_t va_end)
|
uint64_t va_start, uint64_t va_end)
|
||||||
{
|
{
|
||||||
struct iommu_domain *iommu;
|
|
||||||
struct msm_gem_address_space *aspace;
|
struct msm_gem_address_space *aspace;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
@ -831,7 +831,8 @@ msm_gpu_create_address_space(struct msm_gpu *gpu, struct platform_device *pdev,
|
|||||||
* and have separate page tables per context. For now, to keep things
|
* and have separate page tables per context. For now, to keep things
|
||||||
* simple and to get something working, just use a single address space:
|
* simple and to get something working, just use a single address space:
|
||||||
*/
|
*/
|
||||||
iommu = iommu_domain_alloc(&platform_bus_type);
|
if (!adreno_is_a2xx(to_adreno_gpu(gpu))) {
|
||||||
|
struct iommu_domain *iommu = iommu_domain_alloc(&platform_bus_type);
|
||||||
if (!iommu)
|
if (!iommu)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
@ -841,10 +842,16 @@ msm_gpu_create_address_space(struct msm_gpu *gpu, struct platform_device *pdev,
|
|||||||
DRM_DEV_INFO(gpu->dev->dev, "%s: using IOMMU\n", gpu->name);
|
DRM_DEV_INFO(gpu->dev->dev, "%s: using IOMMU\n", gpu->name);
|
||||||
|
|
||||||
aspace = msm_gem_address_space_create(&pdev->dev, iommu, "gpu");
|
aspace = msm_gem_address_space_create(&pdev->dev, iommu, "gpu");
|
||||||
if (IS_ERR(aspace)) {
|
if (IS_ERR(aspace))
|
||||||
DRM_DEV_ERROR(gpu->dev->dev, "failed to init iommu: %ld\n",
|
|
||||||
PTR_ERR(aspace));
|
|
||||||
iommu_domain_free(iommu);
|
iommu_domain_free(iommu);
|
||||||
|
} else {
|
||||||
|
aspace = msm_gem_address_space_create_a2xx(&pdev->dev, gpu, "gpu",
|
||||||
|
va_start, va_end);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (IS_ERR(aspace)) {
|
||||||
|
DRM_DEV_ERROR(gpu->dev->dev, "failed to init mmu: %ld\n",
|
||||||
|
PTR_ERR(aspace));
|
||||||
return ERR_CAST(aspace);
|
return ERR_CAST(aspace);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
123
drivers/gpu/drm/msm/msm_gpummu.c
Normal file
123
drivers/gpu/drm/msm/msm_gpummu.c
Normal file
@ -0,0 +1,123 @@
|
|||||||
|
// SPDX-License-Identifier: GPL-2.0
|
||||||
|
/* Copyright (c) 2018 The Linux Foundation. All rights reserved. */
|
||||||
|
|
||||||
|
#include "msm_drv.h"
|
||||||
|
#include "msm_mmu.h"
|
||||||
|
#include "adreno/adreno_gpu.h"
|
||||||
|
#include "adreno/a2xx.xml.h"
|
||||||
|
|
||||||
|
struct msm_gpummu {
|
||||||
|
struct msm_mmu base;
|
||||||
|
struct msm_gpu *gpu;
|
||||||
|
dma_addr_t pt_base;
|
||||||
|
uint32_t *table;
|
||||||
|
};
|
||||||
|
#define to_msm_gpummu(x) container_of(x, struct msm_gpummu, base)
|
||||||
|
|
||||||
|
#define GPUMMU_VA_START SZ_16M
|
||||||
|
#define GPUMMU_VA_RANGE (0xfff * SZ_64K)
|
||||||
|
#define GPUMMU_PAGE_SIZE SZ_4K
|
||||||
|
#define TABLE_SIZE (sizeof(uint32_t) * GPUMMU_VA_RANGE / GPUMMU_PAGE_SIZE)
|
||||||
|
|
||||||
|
static int msm_gpummu_attach(struct msm_mmu *mmu, const char * const *names,
|
||||||
|
int cnt)
|
||||||
|
{
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void msm_gpummu_detach(struct msm_mmu *mmu, const char * const *names,
|
||||||
|
int cnt)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
static int msm_gpummu_map(struct msm_mmu *mmu, uint64_t iova,
|
||||||
|
struct sg_table *sgt, unsigned len, int prot)
|
||||||
|
{
|
||||||
|
struct msm_gpummu *gpummu = to_msm_gpummu(mmu);
|
||||||
|
unsigned idx = (iova - GPUMMU_VA_START) / GPUMMU_PAGE_SIZE;
|
||||||
|
struct scatterlist *sg;
|
||||||
|
unsigned prot_bits = 0;
|
||||||
|
unsigned i, j;
|
||||||
|
|
||||||
|
if (prot & IOMMU_WRITE)
|
||||||
|
prot_bits |= 1;
|
||||||
|
if (prot & IOMMU_READ)
|
||||||
|
prot_bits |= 2;
|
||||||
|
|
||||||
|
for_each_sg(sgt->sgl, sg, sgt->nents, i) {
|
||||||
|
dma_addr_t addr = sg->dma_address;
|
||||||
|
for (j = 0; j < sg->length / GPUMMU_PAGE_SIZE; j++, idx++) {
|
||||||
|
gpummu->table[idx] = addr | prot_bits;
|
||||||
|
addr += GPUMMU_PAGE_SIZE;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* we can improve by deferring flush for multiple map() */
|
||||||
|
gpu_write(gpummu->gpu, REG_A2XX_MH_MMU_INVALIDATE,
|
||||||
|
A2XX_MH_MMU_INVALIDATE_INVALIDATE_ALL |
|
||||||
|
A2XX_MH_MMU_INVALIDATE_INVALIDATE_TC);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int msm_gpummu_unmap(struct msm_mmu *mmu, uint64_t iova, unsigned len)
|
||||||
|
{
|
||||||
|
struct msm_gpummu *gpummu = to_msm_gpummu(mmu);
|
||||||
|
unsigned idx = (iova - GPUMMU_VA_START) / GPUMMU_PAGE_SIZE;
|
||||||
|
unsigned i;
|
||||||
|
|
||||||
|
for (i = 0; i < len / GPUMMU_PAGE_SIZE; i++, idx++)
|
||||||
|
gpummu->table[idx] = 0;
|
||||||
|
|
||||||
|
gpu_write(gpummu->gpu, REG_A2XX_MH_MMU_INVALIDATE,
|
||||||
|
A2XX_MH_MMU_INVALIDATE_INVALIDATE_ALL |
|
||||||
|
A2XX_MH_MMU_INVALIDATE_INVALIDATE_TC);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void msm_gpummu_destroy(struct msm_mmu *mmu)
|
||||||
|
{
|
||||||
|
struct msm_gpummu *gpummu = to_msm_gpummu(mmu);
|
||||||
|
|
||||||
|
dma_free_attrs(mmu->dev, TABLE_SIZE, gpummu->table, gpummu->pt_base,
|
||||||
|
DMA_ATTR_FORCE_CONTIGUOUS);
|
||||||
|
|
||||||
|
kfree(gpummu);
|
||||||
|
}
|
||||||
|
|
||||||
|
static const struct msm_mmu_funcs funcs = {
|
||||||
|
.attach = msm_gpummu_attach,
|
||||||
|
.detach = msm_gpummu_detach,
|
||||||
|
.map = msm_gpummu_map,
|
||||||
|
.unmap = msm_gpummu_unmap,
|
||||||
|
.destroy = msm_gpummu_destroy,
|
||||||
|
};
|
||||||
|
|
||||||
|
struct msm_mmu *msm_gpummu_new(struct device *dev, struct msm_gpu *gpu)
|
||||||
|
{
|
||||||
|
struct msm_gpummu *gpummu;
|
||||||
|
|
||||||
|
gpummu = kzalloc(sizeof(*gpummu), GFP_KERNEL);
|
||||||
|
if (!gpummu)
|
||||||
|
return ERR_PTR(-ENOMEM);
|
||||||
|
|
||||||
|
gpummu->table = dma_alloc_attrs(dev, TABLE_SIZE + 32, &gpummu->pt_base,
|
||||||
|
GFP_KERNEL | __GFP_ZERO, DMA_ATTR_FORCE_CONTIGUOUS);
|
||||||
|
if (!gpummu->table) {
|
||||||
|
kfree(gpummu);
|
||||||
|
return ERR_PTR(-ENOMEM);
|
||||||
|
}
|
||||||
|
|
||||||
|
gpummu->gpu = gpu;
|
||||||
|
msm_mmu_init(&gpummu->base, dev, &funcs);
|
||||||
|
|
||||||
|
return &gpummu->base;
|
||||||
|
}
|
||||||
|
|
||||||
|
void msm_gpummu_params(struct msm_mmu *mmu, dma_addr_t *pt_base,
|
||||||
|
dma_addr_t *tran_error)
|
||||||
|
{
|
||||||
|
dma_addr_t base = to_msm_gpummu(mmu)->pt_base;
|
||||||
|
|
||||||
|
*pt_base = base;
|
||||||
|
*tran_error = base + TABLE_SIZE; /* 32-byte aligned */
|
||||||
|
}
|
@ -53,4 +53,7 @@ static inline void msm_mmu_set_fault_handler(struct msm_mmu *mmu, void *arg,
|
|||||||
mmu->handler = handler;
|
mmu->handler = handler;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void msm_gpummu_params(struct msm_mmu *mmu, dma_addr_t *pt_base,
|
||||||
|
dma_addr_t *tran_error);
|
||||||
|
|
||||||
#endif /* __MSM_MMU_H__ */
|
#endif /* __MSM_MMU_H__ */
|
||||||
|
Loading…
Reference in New Issue
Block a user