forked from Minki/linux
944fc36c31
Downstream kernel IOMMU had a non-standard way of dealing with multiple devices and multiple ports/contexts. We don't need that on upstream kernel, so rip out the crazy. Note that we have to move the pinning of the ringbuffer to after the IOMMU is attached. No idea how that managed to work properly on the downstream kernel. For now, I am leaving the IOMMU port name stuff in place, to simplify things for folks trying to backport latest drm/msm to device kernels. Once we no longer have to care about pre-DT kernels, we can drop this and instead backport upstream IOMMU driver. Signed-off-by: Rob Clark <robdclark@gmail.com>
141 lines
3.3 KiB
C
141 lines
3.3 KiB
C
/*
|
|
* Copyright (C) 2013 Red Hat
|
|
* Author: Rob Clark <robdclark@gmail.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 as published by
|
|
* the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along with
|
|
* this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include "msm_drv.h"
|
|
#include "msm_mmu.h"
|
|
|
|
struct msm_iommu {
|
|
struct msm_mmu base;
|
|
struct iommu_domain *domain;
|
|
};
|
|
#define to_msm_iommu(x) container_of(x, struct msm_iommu, base)
|
|
|
|
static int msm_fault_handler(struct iommu_domain *iommu, struct device *dev,
|
|
unsigned long iova, int flags, void *arg)
|
|
{
|
|
DBG("*** fault: iova=%08lx, flags=%d", iova, flags);
|
|
return -ENOSYS;
|
|
}
|
|
|
|
static int msm_iommu_attach(struct msm_mmu *mmu, const char **names, int cnt)
|
|
{
|
|
struct msm_iommu *iommu = to_msm_iommu(mmu);
|
|
return iommu_attach_device(iommu->domain, mmu->dev);
|
|
}
|
|
|
|
static void msm_iommu_detach(struct msm_mmu *mmu, const char **names, int cnt)
|
|
{
|
|
struct msm_iommu *iommu = to_msm_iommu(mmu);
|
|
iommu_detach_device(iommu->domain, mmu->dev);
|
|
}
|
|
|
|
static int msm_iommu_map(struct msm_mmu *mmu, uint32_t iova,
|
|
struct sg_table *sgt, unsigned len, int prot)
|
|
{
|
|
struct msm_iommu *iommu = to_msm_iommu(mmu);
|
|
struct iommu_domain *domain = iommu->domain;
|
|
struct scatterlist *sg;
|
|
unsigned int da = iova;
|
|
unsigned int i, j;
|
|
int ret;
|
|
|
|
if (!domain || !sgt)
|
|
return -EINVAL;
|
|
|
|
for_each_sg(sgt->sgl, sg, sgt->nents, i) {
|
|
u32 pa = sg_phys(sg) - sg->offset;
|
|
size_t bytes = sg->length + sg->offset;
|
|
|
|
VERB("map[%d]: %08x %08x(%x)", i, iova, pa, bytes);
|
|
|
|
ret = iommu_map(domain, da, pa, bytes, prot);
|
|
if (ret)
|
|
goto fail;
|
|
|
|
da += bytes;
|
|
}
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
da = iova;
|
|
|
|
for_each_sg(sgt->sgl, sg, i, j) {
|
|
size_t bytes = sg->length + sg->offset;
|
|
iommu_unmap(domain, da, bytes);
|
|
da += bytes;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int msm_iommu_unmap(struct msm_mmu *mmu, uint32_t iova,
|
|
struct sg_table *sgt, unsigned len)
|
|
{
|
|
struct msm_iommu *iommu = to_msm_iommu(mmu);
|
|
struct iommu_domain *domain = iommu->domain;
|
|
struct scatterlist *sg;
|
|
unsigned int da = iova;
|
|
int i;
|
|
|
|
for_each_sg(sgt->sgl, sg, sgt->nents, i) {
|
|
size_t bytes = sg->length + sg->offset;
|
|
size_t unmapped;
|
|
|
|
unmapped = iommu_unmap(domain, da, bytes);
|
|
if (unmapped < bytes)
|
|
return unmapped;
|
|
|
|
VERB("unmap[%d]: %08x(%x)", i, iova, bytes);
|
|
|
|
BUG_ON(!PAGE_ALIGNED(bytes));
|
|
|
|
da += bytes;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void msm_iommu_destroy(struct msm_mmu *mmu)
|
|
{
|
|
struct msm_iommu *iommu = to_msm_iommu(mmu);
|
|
iommu_domain_free(iommu->domain);
|
|
kfree(iommu);
|
|
}
|
|
|
|
static const struct msm_mmu_funcs funcs = {
|
|
.attach = msm_iommu_attach,
|
|
.detach = msm_iommu_detach,
|
|
.map = msm_iommu_map,
|
|
.unmap = msm_iommu_unmap,
|
|
.destroy = msm_iommu_destroy,
|
|
};
|
|
|
|
struct msm_mmu *msm_iommu_new(struct device *dev, struct iommu_domain *domain)
|
|
{
|
|
struct msm_iommu *iommu;
|
|
|
|
iommu = kzalloc(sizeof(*iommu), GFP_KERNEL);
|
|
if (!iommu)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
iommu->domain = domain;
|
|
msm_mmu_init(&iommu->base, dev, &funcs);
|
|
iommu_set_fault_handler(domain, msm_fault_handler, dev);
|
|
|
|
return &iommu->base;
|
|
}
|