mirror of
https://github.com/torvalds/linux.git
synced 2024-12-27 05:11:48 +00:00
e7ba5f1d0f
This patch make the dma dev request operation compatible for both DT and non-DT cases. It takes the all the arguments required for dma_request_slave_channel and dma_request_channel. If the driver is initiated via DT or non-DT the corresponding call will be made. Signed-off-by: Padmavathi Venna <padma.v@samsung.com> Signed-off-by: Mark Brown <broonie@opensource.wolfsonmicro.com>
149 lines
3.7 KiB
C
149 lines
3.7 KiB
C
/* linux/arch/arm/plat-samsung/dma-ops.c
|
|
*
|
|
* Copyright (c) 2011 Samsung Electronics Co., Ltd.
|
|
* http://www.samsung.com
|
|
*
|
|
* Samsung DMA Operations
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/amba/pl330.h>
|
|
#include <linux/scatterlist.h>
|
|
#include <linux/export.h>
|
|
|
|
#include <mach/dma.h>
|
|
|
|
static unsigned samsung_dmadev_request(enum dma_ch dma_ch,
|
|
struct samsung_dma_req *param,
|
|
struct device *dev, char *ch_name)
|
|
{
|
|
dma_cap_mask_t mask;
|
|
void *filter_param;
|
|
|
|
dma_cap_zero(mask);
|
|
dma_cap_set(param->cap, mask);
|
|
|
|
/*
|
|
* If a dma channel property of a device node from device tree is
|
|
* specified, use that as the fliter parameter.
|
|
*/
|
|
filter_param = (dma_ch == DMACH_DT_PROP) ?
|
|
(void *)param->dt_dmach_prop : (void *)dma_ch;
|
|
|
|
if (dev->of_node)
|
|
return (unsigned)dma_request_slave_channel(dev, ch_name);
|
|
else
|
|
return (unsigned)dma_request_channel(mask, pl330_filter,
|
|
filter_param);
|
|
}
|
|
|
|
static int samsung_dmadev_release(unsigned ch, void *param)
|
|
{
|
|
dma_release_channel((struct dma_chan *)ch);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int samsung_dmadev_config(unsigned ch,
|
|
struct samsung_dma_config *param)
|
|
{
|
|
struct dma_chan *chan = (struct dma_chan *)ch;
|
|
struct dma_slave_config slave_config;
|
|
|
|
if (param->direction == DMA_DEV_TO_MEM) {
|
|
memset(&slave_config, 0, sizeof(struct dma_slave_config));
|
|
slave_config.direction = param->direction;
|
|
slave_config.src_addr = param->fifo;
|
|
slave_config.src_addr_width = param->width;
|
|
slave_config.src_maxburst = 1;
|
|
dmaengine_slave_config(chan, &slave_config);
|
|
} else if (param->direction == DMA_MEM_TO_DEV) {
|
|
memset(&slave_config, 0, sizeof(struct dma_slave_config));
|
|
slave_config.direction = param->direction;
|
|
slave_config.dst_addr = param->fifo;
|
|
slave_config.dst_addr_width = param->width;
|
|
slave_config.dst_maxburst = 1;
|
|
dmaengine_slave_config(chan, &slave_config);
|
|
} else {
|
|
pr_warn("unsupported direction\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int samsung_dmadev_prepare(unsigned ch,
|
|
struct samsung_dma_prep *param)
|
|
{
|
|
struct scatterlist sg;
|
|
struct dma_chan *chan = (struct dma_chan *)ch;
|
|
struct dma_async_tx_descriptor *desc;
|
|
|
|
switch (param->cap) {
|
|
case DMA_SLAVE:
|
|
sg_init_table(&sg, 1);
|
|
sg_dma_len(&sg) = param->len;
|
|
sg_set_page(&sg, pfn_to_page(PFN_DOWN(param->buf)),
|
|
param->len, offset_in_page(param->buf));
|
|
sg_dma_address(&sg) = param->buf;
|
|
|
|
desc = dmaengine_prep_slave_sg(chan,
|
|
&sg, 1, param->direction, DMA_PREP_INTERRUPT);
|
|
break;
|
|
case DMA_CYCLIC:
|
|
desc = dmaengine_prep_dma_cyclic(chan, param->buf,
|
|
param->len, param->period, param->direction,
|
|
DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
|
break;
|
|
default:
|
|
dev_err(&chan->dev->device, "unsupported format\n");
|
|
return -EFAULT;
|
|
}
|
|
|
|
if (!desc) {
|
|
dev_err(&chan->dev->device, "cannot prepare cyclic dma\n");
|
|
return -EFAULT;
|
|
}
|
|
|
|
desc->callback = param->fp;
|
|
desc->callback_param = param->fp_param;
|
|
|
|
dmaengine_submit((struct dma_async_tx_descriptor *)desc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline int samsung_dmadev_trigger(unsigned ch)
|
|
{
|
|
dma_async_issue_pending((struct dma_chan *)ch);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline int samsung_dmadev_flush(unsigned ch)
|
|
{
|
|
return dmaengine_terminate_all((struct dma_chan *)ch);
|
|
}
|
|
|
|
static struct samsung_dma_ops dmadev_ops = {
|
|
.request = samsung_dmadev_request,
|
|
.release = samsung_dmadev_release,
|
|
.config = samsung_dmadev_config,
|
|
.prepare = samsung_dmadev_prepare,
|
|
.trigger = samsung_dmadev_trigger,
|
|
.started = NULL,
|
|
.flush = samsung_dmadev_flush,
|
|
.stop = samsung_dmadev_flush,
|
|
};
|
|
|
|
void *samsung_dmadev_get_ops(void)
|
|
{
|
|
return &dmadev_ops;
|
|
}
|
|
EXPORT_SYMBOL(samsung_dmadev_get_ops);
|