2009-08-07 19:01:13 +00:00
|
|
|
/*****************************************************************************
|
|
|
|
* Copyright 2004 - 2008 Broadcom Corporation. All rights reserved.
|
|
|
|
*
|
|
|
|
* Unless you and Broadcom execute a separate written software license
|
|
|
|
* agreement governing use of this software, this software is licensed to you
|
|
|
|
* under the terms of the GNU General Public License version 2, available at
|
|
|
|
* http://www.broadcom.com/licenses/GPLv2.php (the "GPL").
|
|
|
|
*
|
|
|
|
* Notwithstanding the above, under no circumstances may you combine this
|
|
|
|
* software in any way with any other Broadcom software provided under a
|
|
|
|
* license other than the GPL, without Broadcom's express prior written
|
|
|
|
* consent.
|
|
|
|
*****************************************************************************/
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* @file dma.c
|
|
|
|
*
|
|
|
|
* @brief Implements the DMA interface.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
/* ---- Include Files ---------------------------------------------------- */
|
|
|
|
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/device.h>
|
|
|
|
#include <linux/dma-mapping.h>
|
|
|
|
#include <linux/interrupt.h>
|
2011-07-31 21:04:15 +00:00
|
|
|
#include <linux/sched.h>
|
2009-08-07 19:01:13 +00:00
|
|
|
#include <linux/irqreturn.h>
|
|
|
|
#include <linux/proc_fs.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 08:04:11 +00:00
|
|
|
#include <linux/slab.h>
|
2009-08-07 19:01:13 +00:00
|
|
|
|
|
|
|
#include <mach/timer.h>
|
|
|
|
|
|
|
|
#include <linux/mm.h>
|
|
|
|
#include <linux/pfn.h>
|
2011-07-26 23:09:06 +00:00
|
|
|
#include <linux/atomic.h>
|
2011-11-09 10:27:36 +00:00
|
|
|
#include <linux/sched.h>
|
2009-08-07 19:01:13 +00:00
|
|
|
#include <mach/dma.h>
|
|
|
|
|
|
|
|
/* I don't quite understand why dc4 fails when this is set to 1 and DMA is enabled */
|
|
|
|
/* especially since dc4 doesn't use kmalloc'd memory. */
|
|
|
|
|
|
|
|
#define ALLOW_MAP_OF_KMALLOC_MEMORY 0
|
|
|
|
|
|
|
|
/* ---- Public Variables ------------------------------------------------- */
|
|
|
|
|
|
|
|
/* ---- Private Constants and Types -------------------------------------- */
|
|
|
|
|
|
|
|
#define MAKE_HANDLE(controllerIdx, channelIdx) (((controllerIdx) << 4) | (channelIdx))
|
|
|
|
|
|
|
|
#define CONTROLLER_FROM_HANDLE(handle) (((handle) >> 4) & 0x0f)
|
|
|
|
#define CHANNEL_FROM_HANDLE(handle) ((handle) & 0x0f)
|
|
|
|
|
|
|
|
#define DMA_MAP_DEBUG 0
|
|
|
|
|
|
|
|
#if DMA_MAP_DEBUG
|
|
|
|
# define DMA_MAP_PRINT(fmt, args...) printk("%s: " fmt, __func__, ## args)
|
|
|
|
#else
|
|
|
|
# define DMA_MAP_PRINT(fmt, args...)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* ---- Private Variables ------------------------------------------------ */
|
|
|
|
|
|
|
|
static DMA_Global_t gDMA;
|
|
|
|
static struct proc_dir_entry *gDmaDir;
|
|
|
|
|
|
|
|
static atomic_t gDmaStatMemTypeKmalloc = ATOMIC_INIT(0);
|
|
|
|
static atomic_t gDmaStatMemTypeVmalloc = ATOMIC_INIT(0);
|
|
|
|
static atomic_t gDmaStatMemTypeUser = ATOMIC_INIT(0);
|
|
|
|
static atomic_t gDmaStatMemTypeCoherent = ATOMIC_INIT(0);
|
|
|
|
|
|
|
|
#include "dma_device.c"
|
|
|
|
|
|
|
|
/* ---- Private Function Prototypes -------------------------------------- */
|
|
|
|
|
|
|
|
/* ---- Functions ------------------------------------------------------- */
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Displays information for /proc/dma/mem-type
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
static int dma_proc_read_mem_type(char *buf, char **start, off_t offset,
|
|
|
|
int count, int *eof, void *data)
|
|
|
|
{
|
|
|
|
int len = 0;
|
|
|
|
|
|
|
|
len += sprintf(buf + len, "dma_map_mem statistics\n");
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, "coherent: %d\n",
|
|
|
|
atomic_read(&gDmaStatMemTypeCoherent));
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, "kmalloc: %d\n",
|
|
|
|
atomic_read(&gDmaStatMemTypeKmalloc));
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, "vmalloc: %d\n",
|
|
|
|
atomic_read(&gDmaStatMemTypeVmalloc));
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, "user: %d\n",
|
|
|
|
atomic_read(&gDmaStatMemTypeUser));
|
|
|
|
|
|
|
|
return len;
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Displays information for /proc/dma/channels
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
static int dma_proc_read_channels(char *buf, char **start, off_t offset,
|
|
|
|
int count, int *eof, void *data)
|
|
|
|
{
|
|
|
|
int controllerIdx;
|
|
|
|
int channelIdx;
|
|
|
|
int limit = count - 200;
|
|
|
|
int len = 0;
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
|
|
|
|
if (down_interruptible(&gDMA.lock) < 0) {
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (controllerIdx = 0; controllerIdx < DMA_NUM_CONTROLLERS;
|
|
|
|
controllerIdx++) {
|
|
|
|
for (channelIdx = 0; channelIdx < DMA_NUM_CHANNELS;
|
|
|
|
channelIdx++) {
|
|
|
|
if (len >= limit) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
channel =
|
|
|
|
&gDMA.controller[controllerIdx].channel[channelIdx];
|
|
|
|
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, "%d:%d ", controllerIdx,
|
|
|
|
channelIdx);
|
|
|
|
|
|
|
|
if ((channel->flags & DMA_CHANNEL_FLAG_IS_DEDICATED) !=
|
|
|
|
0) {
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, "Dedicated for %s ",
|
|
|
|
DMA_gDeviceAttribute[channel->
|
|
|
|
devType].name);
|
|
|
|
} else {
|
|
|
|
len += sprintf(buf + len, "Shared ");
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((channel->flags & DMA_CHANNEL_FLAG_NO_ISR) != 0) {
|
|
|
|
len += sprintf(buf + len, "No ISR ");
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((channel->flags & DMA_CHANNEL_FLAG_LARGE_FIFO) != 0) {
|
|
|
|
len += sprintf(buf + len, "Fifo: 128 ");
|
|
|
|
} else {
|
|
|
|
len += sprintf(buf + len, "Fifo: 64 ");
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((channel->flags & DMA_CHANNEL_FLAG_IN_USE) != 0) {
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, "InUse by %s",
|
|
|
|
DMA_gDeviceAttribute[channel->
|
|
|
|
devType].name);
|
|
|
|
#if (DMA_DEBUG_TRACK_RESERVATION)
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, " (%s:%d)",
|
|
|
|
channel->fileName,
|
|
|
|
channel->lineNum);
|
|
|
|
#endif
|
|
|
|
} else {
|
|
|
|
len += sprintf(buf + len, "Avail ");
|
|
|
|
}
|
|
|
|
|
|
|
|
if (channel->lastDevType != DMA_DEVICE_NONE) {
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, "Last use: %s ",
|
|
|
|
DMA_gDeviceAttribute[channel->
|
|
|
|
lastDevType].
|
|
|
|
name);
|
|
|
|
}
|
|
|
|
|
|
|
|
len += sprintf(buf + len, "\n");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
up(&gDMA.lock);
|
|
|
|
*eof = 1;
|
|
|
|
|
|
|
|
return len;
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Displays information for /proc/dma/devices
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
static int dma_proc_read_devices(char *buf, char **start, off_t offset,
|
|
|
|
int count, int *eof, void *data)
|
|
|
|
{
|
|
|
|
int limit = count - 200;
|
|
|
|
int len = 0;
|
|
|
|
int devIdx;
|
|
|
|
|
|
|
|
if (down_interruptible(&gDMA.lock) < 0) {
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (devIdx = 0; devIdx < DMA_NUM_DEVICE_ENTRIES; devIdx++) {
|
|
|
|
DMA_DeviceAttribute_t *devAttr = &DMA_gDeviceAttribute[devIdx];
|
|
|
|
|
|
|
|
if (devAttr->name == NULL) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (len >= limit) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
len += sprintf(buf + len, "%-12s ", devAttr->name);
|
|
|
|
|
|
|
|
if ((devAttr->flags & DMA_DEVICE_FLAG_IS_DEDICATED) != 0) {
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len, "Dedicated %d:%d ",
|
|
|
|
devAttr->dedicatedController,
|
|
|
|
devAttr->dedicatedChannel);
|
|
|
|
} else {
|
|
|
|
len += sprintf(buf + len, "Shared DMA:");
|
|
|
|
if ((devAttr->flags & DMA_DEVICE_FLAG_ON_DMA0) != 0) {
|
|
|
|
len += sprintf(buf + len, "0");
|
|
|
|
}
|
|
|
|
if ((devAttr->flags & DMA_DEVICE_FLAG_ON_DMA1) != 0) {
|
|
|
|
len += sprintf(buf + len, "1");
|
|
|
|
}
|
|
|
|
len += sprintf(buf + len, " ");
|
|
|
|
}
|
|
|
|
if ((devAttr->flags & DMA_DEVICE_FLAG_NO_ISR) != 0) {
|
|
|
|
len += sprintf(buf + len, "NoISR ");
|
|
|
|
}
|
|
|
|
if ((devAttr->flags & DMA_DEVICE_FLAG_ALLOW_LARGE_FIFO) != 0) {
|
|
|
|
len += sprintf(buf + len, "Allow-128 ");
|
|
|
|
}
|
|
|
|
|
|
|
|
len +=
|
|
|
|
sprintf(buf + len,
|
|
|
|
"Xfer #: %Lu Ticks: %Lu Bytes: %Lu DescLen: %u\n",
|
|
|
|
devAttr->numTransfers, devAttr->transferTicks,
|
|
|
|
devAttr->transferBytes,
|
|
|
|
devAttr->ring.bytesAllocated);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
up(&gDMA.lock);
|
|
|
|
*eof = 1;
|
|
|
|
|
|
|
|
return len;
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Determines if a DMA_Device_t is "valid".
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* TRUE - dma device is valid
|
|
|
|
* FALSE - dma device isn't valid
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
static inline int IsDeviceValid(DMA_Device_t device)
|
|
|
|
{
|
|
|
|
return (device >= 0) && (device < DMA_NUM_DEVICE_ENTRIES);
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Translates a DMA handle into a pointer to a channel.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* non-NULL - pointer to DMA_Channel_t
|
|
|
|
* NULL - DMA Handle was invalid
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
static inline DMA_Channel_t *HandleToChannel(DMA_Handle_t handle)
|
|
|
|
{
|
|
|
|
int controllerIdx;
|
|
|
|
int channelIdx;
|
|
|
|
|
|
|
|
controllerIdx = CONTROLLER_FROM_HANDLE(handle);
|
|
|
|
channelIdx = CHANNEL_FROM_HANDLE(handle);
|
|
|
|
|
|
|
|
if ((controllerIdx > DMA_NUM_CONTROLLERS)
|
|
|
|
|| (channelIdx > DMA_NUM_CHANNELS)) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return &gDMA.controller[controllerIdx].channel[channelIdx];
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Interrupt handler which is called to process DMA interrupts.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
static irqreturn_t dma_interrupt_handler(int irq, void *dev_id)
|
|
|
|
{
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
int irqStatus;
|
|
|
|
|
|
|
|
channel = (DMA_Channel_t *) dev_id;
|
|
|
|
|
|
|
|
/* Figure out why we were called, and knock down the interrupt */
|
|
|
|
|
|
|
|
irqStatus = dmacHw_getInterruptStatus(channel->dmacHwHandle);
|
|
|
|
dmacHw_clearInterrupt(channel->dmacHwHandle);
|
|
|
|
|
|
|
|
if ((channel->devType < 0)
|
|
|
|
|| (channel->devType > DMA_NUM_DEVICE_ENTRIES)) {
|
|
|
|
printk(KERN_ERR "dma_interrupt_handler: Invalid devType: %d\n",
|
|
|
|
channel->devType);
|
|
|
|
return IRQ_NONE;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[channel->devType];
|
|
|
|
|
|
|
|
/* Update stats */
|
|
|
|
|
|
|
|
if ((irqStatus & dmacHw_INTERRUPT_STATUS_TRANS) != 0) {
|
|
|
|
devAttr->transferTicks +=
|
|
|
|
(timer_get_tick_count() - devAttr->transferStartTime);
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((irqStatus & dmacHw_INTERRUPT_STATUS_ERROR) != 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"dma_interrupt_handler: devType :%d DMA error (%s)\n",
|
|
|
|
channel->devType, devAttr->name);
|
|
|
|
} else {
|
|
|
|
devAttr->numTransfers++;
|
|
|
|
devAttr->transferBytes += devAttr->numBytes;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Call any installed handler */
|
|
|
|
|
|
|
|
if (devAttr->devHandler != NULL) {
|
|
|
|
devAttr->devHandler(channel->devType, irqStatus,
|
|
|
|
devAttr->userData);
|
|
|
|
}
|
|
|
|
|
|
|
|
return IRQ_HANDLED;
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Allocates memory to hold a descriptor ring. The descriptor ring then
|
|
|
|
* needs to be populated by making one or more calls to
|
|
|
|
* dna_add_descriptors.
|
|
|
|
*
|
|
|
|
* The returned descriptor ring will be automatically initialized.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Descriptor ring was allocated successfully
|
|
|
|
* -EINVAL Invalid parameters passed in
|
|
|
|
* -ENOMEM Unable to allocate memory for the desired number of descriptors.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_alloc_descriptor_ring(DMA_DescriptorRing_t *ring, /* Descriptor ring to populate */
|
|
|
|
int numDescriptors /* Number of descriptors that need to be allocated. */
|
|
|
|
) {
|
|
|
|
size_t bytesToAlloc = dmacHw_descriptorLen(numDescriptors);
|
|
|
|
|
|
|
|
if ((ring == NULL) || (numDescriptors <= 0)) {
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
ring->physAddr = 0;
|
|
|
|
ring->descriptorsAllocated = 0;
|
|
|
|
ring->bytesAllocated = 0;
|
|
|
|
|
|
|
|
ring->virtAddr = dma_alloc_writecombine(NULL,
|
|
|
|
bytesToAlloc,
|
|
|
|
&ring->physAddr,
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (ring->virtAddr == NULL) {
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
ring->bytesAllocated = bytesToAlloc;
|
|
|
|
ring->descriptorsAllocated = numDescriptors;
|
|
|
|
|
|
|
|
return dma_init_descriptor_ring(ring, numDescriptors);
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_alloc_descriptor_ring);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Releases the memory which was previously allocated for a descriptor ring.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
void dma_free_descriptor_ring(DMA_DescriptorRing_t *ring /* Descriptor to release */
|
|
|
|
) {
|
|
|
|
if (ring->virtAddr != NULL) {
|
|
|
|
dma_free_writecombine(NULL,
|
|
|
|
ring->bytesAllocated,
|
|
|
|
ring->virtAddr, ring->physAddr);
|
|
|
|
}
|
|
|
|
|
|
|
|
ring->bytesAllocated = 0;
|
|
|
|
ring->descriptorsAllocated = 0;
|
|
|
|
ring->virtAddr = NULL;
|
|
|
|
ring->physAddr = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_free_descriptor_ring);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Initializes a descriptor ring, so that descriptors can be added to it.
|
|
|
|
* Once a descriptor ring has been allocated, it may be reinitialized for
|
|
|
|
* use with additional/different regions of memory.
|
|
|
|
*
|
|
|
|
* Note that if 7 descriptors are allocated, it's perfectly acceptable to
|
|
|
|
* initialize the ring with a smaller number of descriptors. The amount
|
|
|
|
* of memory allocated for the descriptor ring will not be reduced, and
|
|
|
|
* the descriptor ring may be reinitialized later
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Descriptor ring was initialized successfully
|
|
|
|
* -ENOMEM The descriptor which was passed in has insufficient space
|
|
|
|
* to hold the desired number of descriptors.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_init_descriptor_ring(DMA_DescriptorRing_t *ring, /* Descriptor ring to initialize */
|
|
|
|
int numDescriptors /* Number of descriptors to initialize. */
|
|
|
|
) {
|
|
|
|
if (ring->virtAddr == NULL) {
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
if (dmacHw_initDescriptor(ring->virtAddr,
|
|
|
|
ring->physAddr,
|
|
|
|
ring->bytesAllocated, numDescriptors) < 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"dma_init_descriptor_ring: dmacHw_initDescriptor failed\n");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_init_descriptor_ring);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Determines the number of descriptors which would be required for a
|
|
|
|
* transfer of the indicated memory region.
|
|
|
|
*
|
|
|
|
* This function also needs to know which DMA device this transfer will
|
|
|
|
* be destined for, so that the appropriate DMA configuration can be retrieved.
|
|
|
|
* DMA parameters such as transfer width, and whether this is a memory-to-memory
|
|
|
|
* or memory-to-peripheral, etc can all affect the actual number of descriptors
|
|
|
|
* required.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* > 0 Returns the number of descriptors required for the indicated transfer
|
|
|
|
* -ENODEV - Device handed in is invalid.
|
|
|
|
* -EINVAL Invalid parameters
|
|
|
|
* -ENOMEM Memory exhausted
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_calculate_descriptor_count(DMA_Device_t device, /* DMA Device that this will be associated with */
|
|
|
|
dma_addr_t srcData, /* Place to get data to write to device */
|
|
|
|
dma_addr_t dstData, /* Pointer to device data address */
|
|
|
|
size_t numBytes /* Number of bytes to transfer to the device */
|
|
|
|
) {
|
|
|
|
int numDescriptors;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
|
|
|
|
if (!IsDeviceValid(device)) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[device];
|
|
|
|
|
|
|
|
numDescriptors = dmacHw_calculateDescriptorCount(&devAttr->config,
|
|
|
|
(void *)srcData,
|
|
|
|
(void *)dstData,
|
|
|
|
numBytes);
|
|
|
|
if (numDescriptors < 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"dma_calculate_descriptor_count: dmacHw_calculateDescriptorCount failed\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return numDescriptors;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_calculate_descriptor_count);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Adds a region of memory to the descriptor ring. Note that it may take
|
|
|
|
* multiple descriptors for each region of memory. It is the callers
|
|
|
|
* responsibility to allocate a sufficiently large descriptor ring.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Descriptors were added successfully
|
|
|
|
* -ENODEV Device handed in is invalid.
|
|
|
|
* -EINVAL Invalid parameters
|
|
|
|
* -ENOMEM Memory exhausted
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_add_descriptors(DMA_DescriptorRing_t *ring, /* Descriptor ring to add descriptors to */
|
|
|
|
DMA_Device_t device, /* DMA Device that descriptors are for */
|
|
|
|
dma_addr_t srcData, /* Place to get data (memory or device) */
|
|
|
|
dma_addr_t dstData, /* Place to put data (memory or device) */
|
|
|
|
size_t numBytes /* Number of bytes to transfer to the device */
|
|
|
|
) {
|
|
|
|
int rc;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
|
|
|
|
if (!IsDeviceValid(device)) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[device];
|
|
|
|
|
|
|
|
rc = dmacHw_setDataDescriptor(&devAttr->config,
|
|
|
|
ring->virtAddr,
|
|
|
|
(void *)srcData,
|
|
|
|
(void *)dstData, numBytes);
|
|
|
|
if (rc < 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"dma_add_descriptors: dmacHw_setDataDescriptor failed with code: %d\n",
|
|
|
|
rc);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_add_descriptors);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Sets the descriptor ring associated with a device.
|
|
|
|
*
|
|
|
|
* Once set, the descriptor ring will be associated with the device, even
|
|
|
|
* across channel request/free calls. Passing in a NULL descriptor ring
|
|
|
|
* will release any descriptor ring currently associated with the device.
|
|
|
|
*
|
|
|
|
* Note: If you call dma_transfer, or one of the other dma_alloc_ functions
|
|
|
|
* the descriptor ring may be released and reallocated.
|
|
|
|
*
|
|
|
|
* Note: This function will release the descriptor memory for any current
|
|
|
|
* descriptor ring associated with this device.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Descriptors were added successfully
|
|
|
|
* -ENODEV Device handed in is invalid.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_set_device_descriptor_ring(DMA_Device_t device, /* Device to update the descriptor ring for. */
|
|
|
|
DMA_DescriptorRing_t *ring /* Descriptor ring to add descriptors to */
|
|
|
|
) {
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
|
|
|
|
if (!IsDeviceValid(device)) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[device];
|
|
|
|
|
|
|
|
/* Free the previously allocated descriptor ring */
|
|
|
|
|
|
|
|
dma_free_descriptor_ring(&devAttr->ring);
|
|
|
|
|
|
|
|
if (ring != NULL) {
|
|
|
|
/* Copy in the new one */
|
|
|
|
|
|
|
|
devAttr->ring = *ring;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Set things up so that if dma_transfer is called then this descriptor */
|
|
|
|
/* ring will get freed. */
|
|
|
|
|
|
|
|
devAttr->prevSrcData = 0;
|
|
|
|
devAttr->prevDstData = 0;
|
|
|
|
devAttr->prevNumBytes = 0;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_set_device_descriptor_ring);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Retrieves the descriptor ring associated with a device.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Descriptors were added successfully
|
|
|
|
* -ENODEV Device handed in is invalid.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_get_device_descriptor_ring(DMA_Device_t device, /* Device to retrieve the descriptor ring for. */
|
|
|
|
DMA_DescriptorRing_t *ring /* Place to store retrieved ring */
|
|
|
|
) {
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
|
|
|
|
memset(ring, 0, sizeof(*ring));
|
|
|
|
|
|
|
|
if (!IsDeviceValid(device)) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[device];
|
|
|
|
|
|
|
|
*ring = devAttr->ring;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_get_device_descriptor_ring);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Configures a DMA channel.
|
|
|
|
*
|
|
|
|
* @return
|
2011-03-31 01:57:33 +00:00
|
|
|
* >= 0 - Initialization was successful.
|
2009-08-07 19:01:13 +00:00
|
|
|
*
|
|
|
|
* -EBUSY - Device is currently being used.
|
|
|
|
* -ENODEV - Device handed in is invalid.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
static int ConfigChannel(DMA_Handle_t handle)
|
|
|
|
{
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
int controllerIdx;
|
|
|
|
|
|
|
|
channel = HandleToChannel(handle);
|
|
|
|
if (channel == NULL) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[channel->devType];
|
|
|
|
controllerIdx = CONTROLLER_FROM_HANDLE(handle);
|
|
|
|
|
|
|
|
if ((devAttr->flags & DMA_DEVICE_FLAG_PORT_PER_DMAC) != 0) {
|
|
|
|
if (devAttr->config.transferType ==
|
|
|
|
dmacHw_TRANSFER_TYPE_MEM_TO_PERIPHERAL) {
|
|
|
|
devAttr->config.dstPeripheralPort =
|
|
|
|
devAttr->dmacPort[controllerIdx];
|
|
|
|
} else if (devAttr->config.transferType ==
|
|
|
|
dmacHw_TRANSFER_TYPE_PERIPHERAL_TO_MEM) {
|
|
|
|
devAttr->config.srcPeripheralPort =
|
|
|
|
devAttr->dmacPort[controllerIdx];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (dmacHw_configChannel(channel->dmacHwHandle, &devAttr->config) != 0) {
|
|
|
|
printk(KERN_ERR "ConfigChannel: dmacHw_configChannel failed\n");
|
|
|
|
return -EIO;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
tree-wide: fix comment/printk typos
"gadget", "through", "command", "maintain", "maintain", "controller", "address",
"between", "initiali[zs]e", "instead", "function", "select", "already",
"equal", "access", "management", "hierarchy", "registration", "interest",
"relative", "memory", "offset", "already",
Signed-off-by: Uwe Kleine-König <u.kleine-koenig@pengutronix.de>
Signed-off-by: Jiri Kosina <jkosina@suse.cz>
2010-11-01 19:38:34 +00:00
|
|
|
* Initializes all of the data structures associated with the DMA.
|
2009-08-07 19:01:13 +00:00
|
|
|
* @return
|
2011-03-31 01:57:33 +00:00
|
|
|
* >= 0 - Initialization was successful.
|
2009-08-07 19:01:13 +00:00
|
|
|
*
|
|
|
|
* -EBUSY - Device is currently being used.
|
|
|
|
* -ENODEV - Device handed in is invalid.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_init(void)
|
|
|
|
{
|
|
|
|
int rc = 0;
|
|
|
|
int controllerIdx;
|
|
|
|
int channelIdx;
|
|
|
|
DMA_Device_t devIdx;
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
DMA_Handle_t dedicatedHandle;
|
|
|
|
|
|
|
|
memset(&gDMA, 0, sizeof(gDMA));
|
|
|
|
|
2010-09-07 14:33:52 +00:00
|
|
|
sema_init(&gDMA.lock, 0);
|
2009-08-07 19:01:13 +00:00
|
|
|
init_waitqueue_head(&gDMA.freeChannelQ);
|
|
|
|
|
|
|
|
/* Initialize the Hardware */
|
|
|
|
|
|
|
|
dmacHw_initDma();
|
|
|
|
|
|
|
|
/* Start off by marking all of the DMA channels as shared. */
|
|
|
|
|
|
|
|
for (controllerIdx = 0; controllerIdx < DMA_NUM_CONTROLLERS;
|
|
|
|
controllerIdx++) {
|
|
|
|
for (channelIdx = 0; channelIdx < DMA_NUM_CHANNELS;
|
|
|
|
channelIdx++) {
|
|
|
|
channel =
|
|
|
|
&gDMA.controller[controllerIdx].channel[channelIdx];
|
|
|
|
|
|
|
|
channel->flags = 0;
|
|
|
|
channel->devType = DMA_DEVICE_NONE;
|
|
|
|
channel->lastDevType = DMA_DEVICE_NONE;
|
|
|
|
|
|
|
|
#if (DMA_DEBUG_TRACK_RESERVATION)
|
|
|
|
channel->fileName = "";
|
|
|
|
channel->lineNum = 0;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
channel->dmacHwHandle =
|
|
|
|
dmacHw_getChannelHandle(dmacHw_MAKE_CHANNEL_ID
|
|
|
|
(controllerIdx,
|
|
|
|
channelIdx));
|
|
|
|
dmacHw_initChannel(channel->dmacHwHandle);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Record any special attributes that channels may have */
|
|
|
|
|
|
|
|
gDMA.controller[0].channel[0].flags |= DMA_CHANNEL_FLAG_LARGE_FIFO;
|
|
|
|
gDMA.controller[0].channel[1].flags |= DMA_CHANNEL_FLAG_LARGE_FIFO;
|
|
|
|
gDMA.controller[1].channel[0].flags |= DMA_CHANNEL_FLAG_LARGE_FIFO;
|
|
|
|
gDMA.controller[1].channel[1].flags |= DMA_CHANNEL_FLAG_LARGE_FIFO;
|
|
|
|
|
|
|
|
/* Now walk through and record the dedicated channels. */
|
|
|
|
|
|
|
|
for (devIdx = 0; devIdx < DMA_NUM_DEVICE_ENTRIES; devIdx++) {
|
|
|
|
DMA_DeviceAttribute_t *devAttr = &DMA_gDeviceAttribute[devIdx];
|
|
|
|
|
|
|
|
if (((devAttr->flags & DMA_DEVICE_FLAG_NO_ISR) != 0)
|
|
|
|
&& ((devAttr->flags & DMA_DEVICE_FLAG_IS_DEDICATED) == 0)) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"DMA Device: %s Can only request NO_ISR for dedicated devices\n",
|
|
|
|
devAttr->name);
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((devAttr->flags & DMA_DEVICE_FLAG_IS_DEDICATED) != 0) {
|
|
|
|
/* This is a dedicated device. Mark the channel as being reserved. */
|
|
|
|
|
|
|
|
if (devAttr->dedicatedController >= DMA_NUM_CONTROLLERS) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"DMA Device: %s DMA Controller %d is out of range\n",
|
|
|
|
devAttr->name,
|
|
|
|
devAttr->dedicatedController);
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (devAttr->dedicatedChannel >= DMA_NUM_CHANNELS) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"DMA Device: %s DMA Channel %d is out of range\n",
|
|
|
|
devAttr->name,
|
|
|
|
devAttr->dedicatedChannel);
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
dedicatedHandle =
|
|
|
|
MAKE_HANDLE(devAttr->dedicatedController,
|
|
|
|
devAttr->dedicatedChannel);
|
|
|
|
channel = HandleToChannel(dedicatedHandle);
|
|
|
|
|
|
|
|
if ((channel->flags & DMA_CHANNEL_FLAG_IS_DEDICATED) !=
|
|
|
|
0) {
|
|
|
|
printk
|
|
|
|
("DMA Device: %s attempting to use same DMA Controller:Channel (%d:%d) as %s\n",
|
|
|
|
devAttr->name,
|
|
|
|
devAttr->dedicatedController,
|
|
|
|
devAttr->dedicatedChannel,
|
|
|
|
DMA_gDeviceAttribute[channel->devType].
|
|
|
|
name);
|
|
|
|
rc = -EBUSY;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
channel->flags |= DMA_CHANNEL_FLAG_IS_DEDICATED;
|
|
|
|
channel->devType = devIdx;
|
|
|
|
|
|
|
|
if (devAttr->flags & DMA_DEVICE_FLAG_NO_ISR) {
|
|
|
|
channel->flags |= DMA_CHANNEL_FLAG_NO_ISR;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* For dedicated channels, we can go ahead and configure the DMA channel now */
|
|
|
|
/* as well. */
|
|
|
|
|
|
|
|
ConfigChannel(dedicatedHandle);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Go through and register the interrupt handlers */
|
|
|
|
|
|
|
|
for (controllerIdx = 0; controllerIdx < DMA_NUM_CONTROLLERS;
|
|
|
|
controllerIdx++) {
|
|
|
|
for (channelIdx = 0; channelIdx < DMA_NUM_CHANNELS;
|
|
|
|
channelIdx++) {
|
|
|
|
channel =
|
|
|
|
&gDMA.controller[controllerIdx].channel[channelIdx];
|
|
|
|
|
|
|
|
if ((channel->flags & DMA_CHANNEL_FLAG_NO_ISR) == 0) {
|
|
|
|
snprintf(channel->name, sizeof(channel->name),
|
|
|
|
"dma %d:%d %s", controllerIdx,
|
|
|
|
channelIdx,
|
|
|
|
channel->devType ==
|
|
|
|
DMA_DEVICE_NONE ? "" :
|
|
|
|
DMA_gDeviceAttribute[channel->devType].
|
|
|
|
name);
|
|
|
|
|
|
|
|
rc =
|
|
|
|
request_irq(IRQ_DMA0C0 +
|
|
|
|
(controllerIdx *
|
|
|
|
DMA_NUM_CHANNELS) +
|
|
|
|
channelIdx,
|
|
|
|
dma_interrupt_handler,
|
|
|
|
IRQF_DISABLED, channel->name,
|
|
|
|
channel);
|
|
|
|
if (rc != 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"request_irq for IRQ_DMA%dC%d failed\n",
|
|
|
|
controllerIdx, channelIdx);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Create /proc/dma/channels and /proc/dma/devices */
|
|
|
|
|
2011-07-24 06:07:46 +00:00
|
|
|
gDmaDir = proc_mkdir("dma", NULL);
|
2009-08-07 19:01:13 +00:00
|
|
|
|
|
|
|
if (gDmaDir == NULL) {
|
|
|
|
printk(KERN_ERR "Unable to create /proc/dma\n");
|
|
|
|
} else {
|
|
|
|
create_proc_read_entry("channels", 0, gDmaDir,
|
|
|
|
dma_proc_read_channels, NULL);
|
|
|
|
create_proc_read_entry("devices", 0, gDmaDir,
|
|
|
|
dma_proc_read_devices, NULL);
|
|
|
|
create_proc_read_entry("mem-type", 0, gDmaDir,
|
|
|
|
dma_proc_read_mem_type, NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
|
|
|
|
up(&gDMA.lock);
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Reserves a channel for use with @a dev. If the device is setup to use
|
|
|
|
* a shared channel, then this function will block until a free channel
|
|
|
|
* becomes available.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* >= 0 - A valid DMA Handle.
|
|
|
|
* -EBUSY - Device is currently being used.
|
|
|
|
* -ENODEV - Device handed in is invalid.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
#if (DMA_DEBUG_TRACK_RESERVATION)
|
|
|
|
DMA_Handle_t dma_request_channel_dbg
|
|
|
|
(DMA_Device_t dev, const char *fileName, int lineNum)
|
|
|
|
#else
|
|
|
|
DMA_Handle_t dma_request_channel(DMA_Device_t dev)
|
|
|
|
#endif
|
|
|
|
{
|
|
|
|
DMA_Handle_t handle;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
int controllerIdx;
|
|
|
|
int controllerIdx2;
|
|
|
|
int channelIdx;
|
|
|
|
|
|
|
|
if (down_interruptible(&gDMA.lock) < 0) {
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((dev < 0) || (dev >= DMA_NUM_DEVICE_ENTRIES)) {
|
|
|
|
handle = -ENODEV;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[dev];
|
|
|
|
|
|
|
|
#if (DMA_DEBUG_TRACK_RESERVATION)
|
|
|
|
{
|
|
|
|
char *s;
|
|
|
|
|
|
|
|
s = strrchr(fileName, '/');
|
|
|
|
if (s != NULL) {
|
|
|
|
fileName = s + 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
if ((devAttr->flags & DMA_DEVICE_FLAG_IN_USE) != 0) {
|
|
|
|
/* This device has already been requested and not been freed */
|
|
|
|
|
|
|
|
printk(KERN_ERR "%s: device %s is already requested\n",
|
|
|
|
__func__, devAttr->name);
|
|
|
|
handle = -EBUSY;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((devAttr->flags & DMA_DEVICE_FLAG_IS_DEDICATED) != 0) {
|
|
|
|
/* This device has a dedicated channel. */
|
|
|
|
|
|
|
|
channel =
|
|
|
|
&gDMA.controller[devAttr->dedicatedController].
|
|
|
|
channel[devAttr->dedicatedChannel];
|
|
|
|
if ((channel->flags & DMA_CHANNEL_FLAG_IN_USE) != 0) {
|
|
|
|
handle = -EBUSY;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
channel->flags |= DMA_CHANNEL_FLAG_IN_USE;
|
|
|
|
devAttr->flags |= DMA_DEVICE_FLAG_IN_USE;
|
|
|
|
|
|
|
|
#if (DMA_DEBUG_TRACK_RESERVATION)
|
|
|
|
channel->fileName = fileName;
|
|
|
|
channel->lineNum = lineNum;
|
|
|
|
#endif
|
|
|
|
handle =
|
|
|
|
MAKE_HANDLE(devAttr->dedicatedController,
|
|
|
|
devAttr->dedicatedChannel);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* This device needs to use one of the shared channels. */
|
|
|
|
|
|
|
|
handle = DMA_INVALID_HANDLE;
|
|
|
|
while (handle == DMA_INVALID_HANDLE) {
|
|
|
|
/* Scan through the shared channels and see if one is available */
|
|
|
|
|
|
|
|
for (controllerIdx2 = 0; controllerIdx2 < DMA_NUM_CONTROLLERS;
|
|
|
|
controllerIdx2++) {
|
|
|
|
/* Check to see if we should try on controller 1 first. */
|
|
|
|
|
|
|
|
controllerIdx = controllerIdx2;
|
|
|
|
if ((devAttr->
|
|
|
|
flags & DMA_DEVICE_FLAG_ALLOC_DMA1_FIRST) != 0) {
|
|
|
|
controllerIdx = 1 - controllerIdx;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* See if the device is available on the controller being tested */
|
|
|
|
|
|
|
|
if ((devAttr->
|
|
|
|
flags & (DMA_DEVICE_FLAG_ON_DMA0 << controllerIdx))
|
|
|
|
!= 0) {
|
|
|
|
for (channelIdx = 0;
|
|
|
|
channelIdx < DMA_NUM_CHANNELS;
|
|
|
|
channelIdx++) {
|
|
|
|
channel =
|
|
|
|
&gDMA.controller[controllerIdx].
|
|
|
|
channel[channelIdx];
|
|
|
|
|
|
|
|
if (((channel->
|
|
|
|
flags &
|
|
|
|
DMA_CHANNEL_FLAG_IS_DEDICATED) ==
|
|
|
|
0)
|
|
|
|
&&
|
|
|
|
((channel->
|
|
|
|
flags & DMA_CHANNEL_FLAG_IN_USE)
|
|
|
|
== 0)) {
|
|
|
|
if (((channel->
|
|
|
|
flags &
|
|
|
|
DMA_CHANNEL_FLAG_LARGE_FIFO)
|
|
|
|
!= 0)
|
|
|
|
&&
|
|
|
|
((devAttr->
|
|
|
|
flags &
|
|
|
|
DMA_DEVICE_FLAG_ALLOW_LARGE_FIFO)
|
|
|
|
== 0)) {
|
|
|
|
/* This channel is a large fifo - don't tie it up */
|
|
|
|
/* with devices that we don't want using it. */
|
|
|
|
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
channel->flags |=
|
|
|
|
DMA_CHANNEL_FLAG_IN_USE;
|
|
|
|
channel->devType = dev;
|
|
|
|
devAttr->flags |=
|
|
|
|
DMA_DEVICE_FLAG_IN_USE;
|
|
|
|
|
|
|
|
#if (DMA_DEBUG_TRACK_RESERVATION)
|
|
|
|
channel->fileName = fileName;
|
|
|
|
channel->lineNum = lineNum;
|
|
|
|
#endif
|
|
|
|
handle =
|
|
|
|
MAKE_HANDLE(controllerIdx,
|
|
|
|
channelIdx);
|
|
|
|
|
|
|
|
/* Now that we've reserved the channel - we can go ahead and configure it */
|
|
|
|
|
|
|
|
if (ConfigChannel(handle) != 0) {
|
|
|
|
handle = -EIO;
|
|
|
|
printk(KERN_ERR
|
|
|
|
"dma_request_channel: ConfigChannel failed\n");
|
|
|
|
}
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* No channels are currently available. Let's wait for one to free up. */
|
|
|
|
|
|
|
|
{
|
|
|
|
DEFINE_WAIT(wait);
|
|
|
|
|
|
|
|
prepare_to_wait(&gDMA.freeChannelQ, &wait,
|
|
|
|
TASK_INTERRUPTIBLE);
|
|
|
|
up(&gDMA.lock);
|
|
|
|
schedule();
|
|
|
|
finish_wait(&gDMA.freeChannelQ, &wait);
|
|
|
|
|
|
|
|
if (signal_pending(current)) {
|
|
|
|
/* We don't currently hold gDMA.lock, so we return directly */
|
|
|
|
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (down_interruptible(&gDMA.lock)) {
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
up(&gDMA.lock);
|
|
|
|
|
|
|
|
return handle;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Create both _dbg and non _dbg functions for modules. */
|
|
|
|
|
|
|
|
#if (DMA_DEBUG_TRACK_RESERVATION)
|
|
|
|
#undef dma_request_channel
|
|
|
|
DMA_Handle_t dma_request_channel(DMA_Device_t dev)
|
|
|
|
{
|
|
|
|
return dma_request_channel_dbg(dev, __FILE__, __LINE__);
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_request_channel_dbg);
|
|
|
|
#endif
|
|
|
|
EXPORT_SYMBOL(dma_request_channel);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Frees a previously allocated DMA Handle.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_free_channel(DMA_Handle_t handle /* DMA handle. */
|
|
|
|
) {
|
|
|
|
int rc = 0;
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
|
|
|
|
if (down_interruptible(&gDMA.lock) < 0) {
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
}
|
|
|
|
|
|
|
|
channel = HandleToChannel(handle);
|
|
|
|
if (channel == NULL) {
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
devAttr = &DMA_gDeviceAttribute[channel->devType];
|
|
|
|
|
|
|
|
if ((channel->flags & DMA_CHANNEL_FLAG_IS_DEDICATED) == 0) {
|
|
|
|
channel->lastDevType = channel->devType;
|
|
|
|
channel->devType = DMA_DEVICE_NONE;
|
|
|
|
}
|
|
|
|
channel->flags &= ~DMA_CHANNEL_FLAG_IN_USE;
|
|
|
|
devAttr->flags &= ~DMA_DEVICE_FLAG_IN_USE;
|
|
|
|
|
|
|
|
out:
|
|
|
|
up(&gDMA.lock);
|
|
|
|
|
|
|
|
wake_up_interruptible(&gDMA.freeChannelQ);
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_free_channel);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Determines if a given device has been configured as using a shared
|
|
|
|
* channel.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Device uses a dedicated channel
|
|
|
|
* > zero Device uses a shared channel
|
|
|
|
* < zero Error code
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_device_is_channel_shared(DMA_Device_t device /* Device to check. */
|
|
|
|
) {
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
|
|
|
|
if (!IsDeviceValid(device)) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[device];
|
|
|
|
|
|
|
|
return ((devAttr->flags & DMA_DEVICE_FLAG_IS_DEDICATED) == 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_device_is_channel_shared);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Allocates buffers for the descriptors. This is normally done automatically
|
|
|
|
* but needs to be done explicitly when initiating a dma from interrupt
|
|
|
|
* context.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Descriptors were allocated successfully
|
|
|
|
* -EINVAL Invalid device type for this kind of transfer
|
|
|
|
* (i.e. the device is _MEM_TO_DEV and not _DEV_TO_MEM)
|
|
|
|
* -ENOMEM Memory exhausted
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_alloc_descriptors(DMA_Handle_t handle, /* DMA Handle */
|
|
|
|
dmacHw_TRANSFER_TYPE_e transferType, /* Type of transfer being performed */
|
|
|
|
dma_addr_t srcData, /* Place to get data to write to device */
|
|
|
|
dma_addr_t dstData, /* Pointer to device data address */
|
|
|
|
size_t numBytes /* Number of bytes to transfer to the device */
|
|
|
|
) {
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
int numDescriptors;
|
|
|
|
size_t ringBytesRequired;
|
|
|
|
int rc = 0;
|
|
|
|
|
|
|
|
channel = HandleToChannel(handle);
|
|
|
|
if (channel == NULL) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
|
|
|
devAttr = &DMA_gDeviceAttribute[channel->devType];
|
|
|
|
|
|
|
|
if (devAttr->config.transferType != transferType) {
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Figure out how many descriptors we need. */
|
|
|
|
|
|
|
|
/* printk("srcData: 0x%08x dstData: 0x%08x, numBytes: %d\n", */
|
|
|
|
/* srcData, dstData, numBytes); */
|
|
|
|
|
|
|
|
numDescriptors = dmacHw_calculateDescriptorCount(&devAttr->config,
|
|
|
|
(void *)srcData,
|
|
|
|
(void *)dstData,
|
|
|
|
numBytes);
|
|
|
|
if (numDescriptors < 0) {
|
|
|
|
printk(KERN_ERR "%s: dmacHw_calculateDescriptorCount failed\n",
|
|
|
|
__func__);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Check to see if we can reuse the existing descriptor ring, or if we need to allocate */
|
|
|
|
/* a new one. */
|
|
|
|
|
|
|
|
ringBytesRequired = dmacHw_descriptorLen(numDescriptors);
|
|
|
|
|
|
|
|
/* printk("ringBytesRequired: %d\n", ringBytesRequired); */
|
|
|
|
|
|
|
|
if (ringBytesRequired > devAttr->ring.bytesAllocated) {
|
|
|
|
/* Make sure that this code path is never taken from interrupt context. */
|
|
|
|
/* It's OK for an interrupt to initiate a DMA transfer, but the descriptor */
|
|
|
|
/* allocation needs to have already been done. */
|
|
|
|
|
|
|
|
might_sleep();
|
|
|
|
|
|
|
|
/* Free the old descriptor ring and allocate a new one. */
|
|
|
|
|
|
|
|
dma_free_descriptor_ring(&devAttr->ring);
|
|
|
|
|
|
|
|
/* And allocate a new one. */
|
|
|
|
|
|
|
|
rc =
|
|
|
|
dma_alloc_descriptor_ring(&devAttr->ring,
|
|
|
|
numDescriptors);
|
|
|
|
if (rc < 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: dma_alloc_descriptor_ring(%d) failed\n",
|
|
|
|
__func__, numDescriptors);
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
/* Setup the descriptor for this transfer */
|
|
|
|
|
|
|
|
if (dmacHw_initDescriptor(devAttr->ring.virtAddr,
|
|
|
|
devAttr->ring.physAddr,
|
|
|
|
devAttr->ring.bytesAllocated,
|
|
|
|
numDescriptors) < 0) {
|
|
|
|
printk(KERN_ERR "%s: dmacHw_initDescriptor failed\n",
|
|
|
|
__func__);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
/* We've already got enough ring buffer allocated. All we need to do is reset */
|
|
|
|
/* any control information, just in case the previous DMA was stopped. */
|
|
|
|
|
|
|
|
dmacHw_resetDescriptorControl(devAttr->ring.virtAddr);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* dma_alloc/free both set the prevSrc/DstData to 0. If they happen to be the same */
|
|
|
|
/* as last time, then we don't need to call setDataDescriptor again. */
|
|
|
|
|
|
|
|
if (dmacHw_setDataDescriptor(&devAttr->config,
|
|
|
|
devAttr->ring.virtAddr,
|
|
|
|
(void *)srcData,
|
|
|
|
(void *)dstData, numBytes) < 0) {
|
|
|
|
printk(KERN_ERR "%s: dmacHw_setDataDescriptor failed\n",
|
|
|
|
__func__);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Remember the critical information for this transfer so that we can eliminate */
|
|
|
|
/* another call to dma_alloc_descriptors if the caller reuses the same buffers */
|
|
|
|
|
|
|
|
devAttr->prevSrcData = srcData;
|
|
|
|
devAttr->prevDstData = dstData;
|
|
|
|
devAttr->prevNumBytes = numBytes;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_alloc_descriptors);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Allocates and sets up descriptors for a double buffered circular buffer.
|
|
|
|
*
|
|
|
|
* This is primarily intended to be used for things like the ingress samples
|
|
|
|
* from a microphone.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* > 0 Number of descriptors actually allocated.
|
|
|
|
* -EINVAL Invalid device type for this kind of transfer
|
|
|
|
* (i.e. the device is _MEM_TO_DEV and not _DEV_TO_MEM)
|
|
|
|
* -ENOMEM Memory exhausted
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_alloc_double_dst_descriptors(DMA_Handle_t handle, /* DMA Handle */
|
|
|
|
dma_addr_t srcData, /* Physical address of source data */
|
|
|
|
dma_addr_t dstData1, /* Physical address of first destination buffer */
|
|
|
|
dma_addr_t dstData2, /* Physical address of second destination buffer */
|
|
|
|
size_t numBytes /* Number of bytes in each destination buffer */
|
|
|
|
) {
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
int numDst1Descriptors;
|
|
|
|
int numDst2Descriptors;
|
|
|
|
int numDescriptors;
|
|
|
|
size_t ringBytesRequired;
|
|
|
|
int rc = 0;
|
|
|
|
|
|
|
|
channel = HandleToChannel(handle);
|
|
|
|
if (channel == NULL) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
|
|
|
devAttr = &DMA_gDeviceAttribute[channel->devType];
|
|
|
|
|
|
|
|
/* Figure out how many descriptors we need. */
|
|
|
|
|
|
|
|
/* printk("srcData: 0x%08x dstData: 0x%08x, numBytes: %d\n", */
|
|
|
|
/* srcData, dstData, numBytes); */
|
|
|
|
|
|
|
|
numDst1Descriptors =
|
|
|
|
dmacHw_calculateDescriptorCount(&devAttr->config, (void *)srcData,
|
|
|
|
(void *)dstData1, numBytes);
|
|
|
|
if (numDst1Descriptors < 0) {
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
numDst2Descriptors =
|
|
|
|
dmacHw_calculateDescriptorCount(&devAttr->config, (void *)srcData,
|
|
|
|
(void *)dstData2, numBytes);
|
|
|
|
if (numDst2Descriptors < 0) {
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
numDescriptors = numDst1Descriptors + numDst2Descriptors;
|
|
|
|
/* printk("numDescriptors: %d\n", numDescriptors); */
|
|
|
|
|
|
|
|
/* Check to see if we can reuse the existing descriptor ring, or if we need to allocate */
|
|
|
|
/* a new one. */
|
|
|
|
|
|
|
|
ringBytesRequired = dmacHw_descriptorLen(numDescriptors);
|
|
|
|
|
|
|
|
/* printk("ringBytesRequired: %d\n", ringBytesRequired); */
|
|
|
|
|
|
|
|
if (ringBytesRequired > devAttr->ring.bytesAllocated) {
|
|
|
|
/* Make sure that this code path is never taken from interrupt context. */
|
|
|
|
/* It's OK for an interrupt to initiate a DMA transfer, but the descriptor */
|
|
|
|
/* allocation needs to have already been done. */
|
|
|
|
|
|
|
|
might_sleep();
|
|
|
|
|
|
|
|
/* Free the old descriptor ring and allocate a new one. */
|
|
|
|
|
|
|
|
dma_free_descriptor_ring(&devAttr->ring);
|
|
|
|
|
|
|
|
/* And allocate a new one. */
|
|
|
|
|
|
|
|
rc =
|
|
|
|
dma_alloc_descriptor_ring(&devAttr->ring,
|
|
|
|
numDescriptors);
|
|
|
|
if (rc < 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: dma_alloc_descriptor_ring(%d) failed\n",
|
|
|
|
__func__, ringBytesRequired);
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Setup the descriptor for this transfer. Since this function is used with */
|
|
|
|
/* CONTINUOUS DMA operations, we need to reinitialize every time, otherwise */
|
|
|
|
/* setDataDescriptor will keep trying to append onto the end. */
|
|
|
|
|
|
|
|
if (dmacHw_initDescriptor(devAttr->ring.virtAddr,
|
|
|
|
devAttr->ring.physAddr,
|
|
|
|
devAttr->ring.bytesAllocated,
|
|
|
|
numDescriptors) < 0) {
|
|
|
|
printk(KERN_ERR "%s: dmacHw_initDescriptor failed\n", __func__);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* dma_alloc/free both set the prevSrc/DstData to 0. If they happen to be the same */
|
|
|
|
/* as last time, then we don't need to call setDataDescriptor again. */
|
|
|
|
|
|
|
|
if (dmacHw_setDataDescriptor(&devAttr->config,
|
|
|
|
devAttr->ring.virtAddr,
|
|
|
|
(void *)srcData,
|
|
|
|
(void *)dstData1, numBytes) < 0) {
|
|
|
|
printk(KERN_ERR "%s: dmacHw_setDataDescriptor 1 failed\n",
|
|
|
|
__func__);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
if (dmacHw_setDataDescriptor(&devAttr->config,
|
|
|
|
devAttr->ring.virtAddr,
|
|
|
|
(void *)srcData,
|
|
|
|
(void *)dstData2, numBytes) < 0) {
|
|
|
|
printk(KERN_ERR "%s: dmacHw_setDataDescriptor 2 failed\n",
|
|
|
|
__func__);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* You should use dma_start_transfer rather than dma_transfer_xxx so we don't */
|
|
|
|
/* try to make the 'prev' variables right. */
|
|
|
|
|
|
|
|
devAttr->prevSrcData = 0;
|
|
|
|
devAttr->prevDstData = 0;
|
|
|
|
devAttr->prevNumBytes = 0;
|
|
|
|
|
|
|
|
return numDescriptors;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_alloc_double_dst_descriptors);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Initiates a transfer when the descriptors have already been setup.
|
|
|
|
*
|
|
|
|
* This is a special case, and normally, the dma_transfer_xxx functions should
|
|
|
|
* be used.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Transfer was started successfully
|
|
|
|
* -ENODEV Invalid handle
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_start_transfer(DMA_Handle_t handle)
|
|
|
|
{
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
|
|
|
|
channel = HandleToChannel(handle);
|
|
|
|
if (channel == NULL) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[channel->devType];
|
|
|
|
|
|
|
|
dmacHw_initiateTransfer(channel->dmacHwHandle, &devAttr->config,
|
|
|
|
devAttr->ring.virtAddr);
|
|
|
|
|
|
|
|
/* Since we got this far, everything went successfully */
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_start_transfer);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Stops a previously started DMA transfer.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Transfer was stopped successfully
|
|
|
|
* -ENODEV Invalid handle
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_stop_transfer(DMA_Handle_t handle)
|
|
|
|
{
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
|
|
|
|
channel = HandleToChannel(handle);
|
|
|
|
if (channel == NULL) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
|
|
|
dmacHw_stopTransfer(channel->dmacHwHandle);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_stop_transfer);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Waits for a DMA to complete by polling. This function is only intended
|
|
|
|
* to be used for testing. Interrupts should be used for most DMA operations.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_wait_transfer_done(DMA_Handle_t handle)
|
|
|
|
{
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
dmacHw_TRANSFER_STATUS_e status;
|
|
|
|
|
|
|
|
channel = HandleToChannel(handle);
|
|
|
|
if (channel == NULL) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
|
|
|
while ((status =
|
|
|
|
dmacHw_transferCompleted(channel->dmacHwHandle)) ==
|
|
|
|
dmacHw_TRANSFER_STATUS_BUSY) {
|
|
|
|
;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (status == dmacHw_TRANSFER_STATUS_ERROR) {
|
|
|
|
printk(KERN_ERR "%s: DMA transfer failed\n", __func__);
|
|
|
|
return -EIO;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_wait_transfer_done);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Initiates a DMA, allocating the descriptors as required.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 Transfer was started successfully
|
|
|
|
* -EINVAL Invalid device type for this kind of transfer
|
|
|
|
* (i.e. the device is _DEV_TO_MEM and not _MEM_TO_DEV)
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_transfer(DMA_Handle_t handle, /* DMA Handle */
|
|
|
|
dmacHw_TRANSFER_TYPE_e transferType, /* Type of transfer being performed */
|
|
|
|
dma_addr_t srcData, /* Place to get data to write to device */
|
|
|
|
dma_addr_t dstData, /* Pointer to device data address */
|
|
|
|
size_t numBytes /* Number of bytes to transfer to the device */
|
|
|
|
) {
|
|
|
|
DMA_Channel_t *channel;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
int rc = 0;
|
|
|
|
|
|
|
|
channel = HandleToChannel(handle);
|
|
|
|
if (channel == NULL) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
|
|
|
devAttr = &DMA_gDeviceAttribute[channel->devType];
|
|
|
|
|
|
|
|
if (devAttr->config.transferType != transferType) {
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* We keep track of the information about the previous request for this */
|
|
|
|
/* device, and if the attributes match, then we can use the descriptors we setup */
|
|
|
|
/* the last time, and not have to reinitialize everything. */
|
|
|
|
|
|
|
|
{
|
|
|
|
rc =
|
|
|
|
dma_alloc_descriptors(handle, transferType, srcData,
|
|
|
|
dstData, numBytes);
|
|
|
|
if (rc != 0) {
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* And kick off the transfer */
|
|
|
|
|
|
|
|
devAttr->numBytes = numBytes;
|
|
|
|
devAttr->transferStartTime = timer_get_tick_count();
|
|
|
|
|
|
|
|
dmacHw_initiateTransfer(channel->dmacHwHandle, &devAttr->config,
|
|
|
|
devAttr->ring.virtAddr);
|
|
|
|
|
|
|
|
/* Since we got this far, everything went successfully */
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_transfer);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Set the callback function which will be called when a transfer completes.
|
|
|
|
* If a NULL callback function is set, then no callback will occur.
|
|
|
|
*
|
|
|
|
* @note @a devHandler will be called from IRQ context.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 - Success
|
|
|
|
* -ENODEV - Device handed in is invalid.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_set_device_handler(DMA_Device_t dev, /* Device to set the callback for. */
|
|
|
|
DMA_DeviceHandler_t devHandler, /* Function to call when the DMA completes */
|
|
|
|
void *userData /* Pointer which will be passed to devHandler. */
|
|
|
|
) {
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
if (!IsDeviceValid(dev)) {
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
devAttr = &DMA_gDeviceAttribute[dev];
|
|
|
|
|
|
|
|
local_irq_save(flags);
|
|
|
|
|
|
|
|
devAttr->userData = userData;
|
|
|
|
devAttr->devHandler = devHandler;
|
|
|
|
|
|
|
|
local_irq_restore(flags);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_set_device_handler);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Initializes a memory mapping structure
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_init_mem_map(DMA_MemMap_t *memMap)
|
|
|
|
{
|
|
|
|
memset(memMap, 0, sizeof(*memMap));
|
|
|
|
|
2010-09-07 14:33:52 +00:00
|
|
|
sema_init(&memMap->lock, 1);
|
2009-08-07 19:01:13 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_init_mem_map);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Releases any memory currently being held by a memory mapping structure.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_term_mem_map(DMA_MemMap_t *memMap)
|
|
|
|
{
|
|
|
|
down(&memMap->lock); /* Just being paranoid */
|
|
|
|
|
|
|
|
/* Free up any allocated memory */
|
|
|
|
|
|
|
|
up(&memMap->lock);
|
|
|
|
memset(memMap, 0, sizeof(*memMap));
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_term_mem_map);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Looks at a memory address and categorizes it.
|
|
|
|
*
|
|
|
|
* @return One of the values from the DMA_MemType_t enumeration.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
DMA_MemType_t dma_mem_type(void *addr)
|
|
|
|
{
|
|
|
|
unsigned long addrVal = (unsigned long)addr;
|
|
|
|
|
|
|
|
if (addrVal >= VMALLOC_END) {
|
|
|
|
/* NOTE: DMA virtual memory space starts at 0xFFxxxxxx */
|
|
|
|
|
|
|
|
/* dma_alloc_xxx pages are physically and virtually contiguous */
|
|
|
|
|
|
|
|
return DMA_MEM_TYPE_DMA;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Technically, we could add one more classification. Addresses between VMALLOC_END */
|
|
|
|
/* and the beginning of the DMA virtual address could be considered to be I/O space. */
|
|
|
|
/* Right now, nobody cares about this particular classification, so we ignore it. */
|
|
|
|
|
|
|
|
if (is_vmalloc_addr(addr)) {
|
|
|
|
/* Address comes from the vmalloc'd region. Pages are virtually */
|
|
|
|
/* contiguous but NOT physically contiguous */
|
|
|
|
|
|
|
|
return DMA_MEM_TYPE_VMALLOC;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (addrVal >= PAGE_OFFSET) {
|
|
|
|
/* PAGE_OFFSET is typically 0xC0000000 */
|
|
|
|
|
|
|
|
/* kmalloc'd pages are physically contiguous */
|
|
|
|
|
|
|
|
return DMA_MEM_TYPE_KMALLOC;
|
|
|
|
}
|
|
|
|
|
|
|
|
return DMA_MEM_TYPE_USER;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_mem_type);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Looks at a memory address and determines if we support DMA'ing to/from
|
|
|
|
* that type of memory.
|
|
|
|
*
|
|
|
|
* @return boolean -
|
|
|
|
* return value != 0 means dma supported
|
|
|
|
* return value == 0 means dma not supported
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_mem_supports_dma(void *addr)
|
|
|
|
{
|
|
|
|
DMA_MemType_t memType = dma_mem_type(addr);
|
|
|
|
|
|
|
|
return (memType == DMA_MEM_TYPE_DMA)
|
|
|
|
#if ALLOW_MAP_OF_KMALLOC_MEMORY
|
|
|
|
|| (memType == DMA_MEM_TYPE_KMALLOC)
|
|
|
|
#endif
|
|
|
|
|| (memType == DMA_MEM_TYPE_USER);
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_mem_supports_dma);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Maps in a memory region such that it can be used for performing a DMA.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_map_start(DMA_MemMap_t *memMap, /* Stores state information about the map */
|
|
|
|
enum dma_data_direction dir /* Direction that the mapping will be going */
|
|
|
|
) {
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
down(&memMap->lock);
|
|
|
|
|
|
|
|
DMA_MAP_PRINT("memMap: %p\n", memMap);
|
|
|
|
|
|
|
|
if (memMap->inUse) {
|
|
|
|
printk(KERN_ERR "%s: memory map %p is already being used\n",
|
|
|
|
__func__, memMap);
|
|
|
|
rc = -EBUSY;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
memMap->inUse = 1;
|
|
|
|
memMap->dir = dir;
|
|
|
|
memMap->numRegionsUsed = 0;
|
|
|
|
|
|
|
|
rc = 0;
|
|
|
|
|
|
|
|
out:
|
|
|
|
|
|
|
|
DMA_MAP_PRINT("returning %d", rc);
|
|
|
|
|
|
|
|
up(&memMap->lock);
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_map_start);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Adds a segment of memory to a memory map. Each segment is both
|
|
|
|
* physically and virtually contiguous.
|
|
|
|
*
|
|
|
|
* @return 0 on success, error code otherwise.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
static int dma_map_add_segment(DMA_MemMap_t *memMap, /* Stores state information about the map */
|
|
|
|
DMA_Region_t *region, /* Region that the segment belongs to */
|
|
|
|
void *virtAddr, /* Virtual address of the segment being added */
|
|
|
|
dma_addr_t physAddr, /* Physical address of the segment being added */
|
|
|
|
size_t numBytes /* Number of bytes of the segment being added */
|
|
|
|
) {
|
|
|
|
DMA_Segment_t *segment;
|
|
|
|
|
|
|
|
DMA_MAP_PRINT("memMap:%p va:%p pa:0x%x #:%d\n", memMap, virtAddr,
|
|
|
|
physAddr, numBytes);
|
|
|
|
|
|
|
|
/* Sanity check */
|
|
|
|
|
|
|
|
if (((unsigned long)virtAddr < (unsigned long)region->virtAddr)
|
|
|
|
|| (((unsigned long)virtAddr + numBytes)) >
|
|
|
|
((unsigned long)region->virtAddr + region->numBytes)) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: virtAddr %p is outside region @ %p len: %d\n",
|
|
|
|
__func__, virtAddr, region->virtAddr, region->numBytes);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (region->numSegmentsUsed > 0) {
|
|
|
|
/* Check to see if this segment is physically contiguous with the previous one */
|
|
|
|
|
|
|
|
segment = ®ion->segment[region->numSegmentsUsed - 1];
|
|
|
|
|
|
|
|
if ((segment->physAddr + segment->numBytes) == physAddr) {
|
|
|
|
/* It is - just add on to the end */
|
|
|
|
|
|
|
|
DMA_MAP_PRINT("appending %d bytes to last segment\n",
|
|
|
|
numBytes);
|
|
|
|
|
|
|
|
segment->numBytes += numBytes;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Reallocate to hold more segments, if required. */
|
|
|
|
|
|
|
|
if (region->numSegmentsUsed >= region->numSegmentsAllocated) {
|
|
|
|
DMA_Segment_t *newSegment;
|
|
|
|
size_t oldSize =
|
|
|
|
region->numSegmentsAllocated * sizeof(*newSegment);
|
|
|
|
int newAlloc = region->numSegmentsAllocated + 4;
|
|
|
|
size_t newSize = newAlloc * sizeof(*newSegment);
|
|
|
|
|
|
|
|
newSegment = kmalloc(newSize, GFP_KERNEL);
|
|
|
|
if (newSegment == NULL) {
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
memcpy(newSegment, region->segment, oldSize);
|
|
|
|
memset(&((uint8_t *) newSegment)[oldSize], 0,
|
|
|
|
newSize - oldSize);
|
|
|
|
kfree(region->segment);
|
|
|
|
|
|
|
|
region->numSegmentsAllocated = newAlloc;
|
|
|
|
region->segment = newSegment;
|
|
|
|
}
|
|
|
|
|
|
|
|
segment = ®ion->segment[region->numSegmentsUsed];
|
|
|
|
region->numSegmentsUsed++;
|
|
|
|
|
|
|
|
segment->virtAddr = virtAddr;
|
|
|
|
segment->physAddr = physAddr;
|
|
|
|
segment->numBytes = numBytes;
|
|
|
|
|
|
|
|
DMA_MAP_PRINT("returning success\n");
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Adds a region of memory to a memory map. Each region is virtually
|
|
|
|
* contiguous, but not necessarily physically contiguous.
|
|
|
|
*
|
|
|
|
* @return 0 on success, error code otherwise.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_map_add_region(DMA_MemMap_t *memMap, /* Stores state information about the map */
|
|
|
|
void *mem, /* Virtual address that we want to get a map of */
|
|
|
|
size_t numBytes /* Number of bytes being mapped */
|
|
|
|
) {
|
|
|
|
unsigned long addr = (unsigned long)mem;
|
|
|
|
unsigned int offset;
|
|
|
|
int rc = 0;
|
|
|
|
DMA_Region_t *region;
|
|
|
|
dma_addr_t physAddr;
|
|
|
|
|
|
|
|
down(&memMap->lock);
|
|
|
|
|
|
|
|
DMA_MAP_PRINT("memMap:%p va:%p #:%d\n", memMap, mem, numBytes);
|
|
|
|
|
|
|
|
if (!memMap->inUse) {
|
|
|
|
printk(KERN_ERR "%s: Make sure you call dma_map_start first\n",
|
|
|
|
__func__);
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Reallocate to hold more regions. */
|
|
|
|
|
|
|
|
if (memMap->numRegionsUsed >= memMap->numRegionsAllocated) {
|
|
|
|
DMA_Region_t *newRegion;
|
|
|
|
size_t oldSize =
|
|
|
|
memMap->numRegionsAllocated * sizeof(*newRegion);
|
|
|
|
int newAlloc = memMap->numRegionsAllocated + 4;
|
|
|
|
size_t newSize = newAlloc * sizeof(*newRegion);
|
|
|
|
|
|
|
|
newRegion = kmalloc(newSize, GFP_KERNEL);
|
|
|
|
if (newRegion == NULL) {
|
|
|
|
rc = -ENOMEM;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
memcpy(newRegion, memMap->region, oldSize);
|
|
|
|
memset(&((uint8_t *) newRegion)[oldSize], 0, newSize - oldSize);
|
|
|
|
|
|
|
|
kfree(memMap->region);
|
|
|
|
|
|
|
|
memMap->numRegionsAllocated = newAlloc;
|
|
|
|
memMap->region = newRegion;
|
|
|
|
}
|
|
|
|
|
|
|
|
region = &memMap->region[memMap->numRegionsUsed];
|
|
|
|
memMap->numRegionsUsed++;
|
|
|
|
|
|
|
|
offset = addr & ~PAGE_MASK;
|
|
|
|
|
|
|
|
region->memType = dma_mem_type(mem);
|
|
|
|
region->virtAddr = mem;
|
|
|
|
region->numBytes = numBytes;
|
|
|
|
region->numSegmentsUsed = 0;
|
|
|
|
region->numLockedPages = 0;
|
|
|
|
region->lockedPages = NULL;
|
|
|
|
|
|
|
|
switch (region->memType) {
|
|
|
|
case DMA_MEM_TYPE_VMALLOC:
|
|
|
|
{
|
|
|
|
atomic_inc(&gDmaStatMemTypeVmalloc);
|
|
|
|
|
|
|
|
/* printk(KERN_ERR "%s: vmalloc'd pages are not supported\n", __func__); */
|
|
|
|
|
|
|
|
/* vmalloc'd pages are not physically contiguous */
|
|
|
|
|
|
|
|
rc = -EINVAL;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case DMA_MEM_TYPE_KMALLOC:
|
|
|
|
{
|
|
|
|
atomic_inc(&gDmaStatMemTypeKmalloc);
|
|
|
|
|
|
|
|
/* kmalloc'd pages are physically contiguous, so they'll have exactly */
|
|
|
|
/* one segment */
|
|
|
|
|
|
|
|
#if ALLOW_MAP_OF_KMALLOC_MEMORY
|
|
|
|
physAddr =
|
|
|
|
dma_map_single(NULL, mem, numBytes, memMap->dir);
|
|
|
|
rc = dma_map_add_segment(memMap, region, mem, physAddr,
|
|
|
|
numBytes);
|
|
|
|
#else
|
|
|
|
rc = -EINVAL;
|
|
|
|
#endif
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case DMA_MEM_TYPE_DMA:
|
|
|
|
{
|
|
|
|
/* dma_alloc_xxx pages are physically contiguous */
|
|
|
|
|
|
|
|
atomic_inc(&gDmaStatMemTypeCoherent);
|
|
|
|
|
|
|
|
physAddr = (vmalloc_to_pfn(mem) << PAGE_SHIFT) + offset;
|
|
|
|
|
|
|
|
dma_sync_single_for_cpu(NULL, physAddr, numBytes,
|
|
|
|
memMap->dir);
|
|
|
|
rc = dma_map_add_segment(memMap, region, mem, physAddr,
|
|
|
|
numBytes);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case DMA_MEM_TYPE_USER:
|
|
|
|
{
|
|
|
|
size_t firstPageOffset;
|
|
|
|
size_t firstPageSize;
|
|
|
|
struct page **pages;
|
|
|
|
struct task_struct *userTask;
|
|
|
|
|
|
|
|
atomic_inc(&gDmaStatMemTypeUser);
|
|
|
|
|
|
|
|
#if 1
|
|
|
|
/* If the pages are user pages, then the dma_mem_map_set_user_task function */
|
|
|
|
/* must have been previously called. */
|
|
|
|
|
|
|
|
if (memMap->userTask == NULL) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: must call dma_mem_map_set_user_task when using user-mode memory\n",
|
|
|
|
__func__);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* User pages need to be locked. */
|
|
|
|
|
|
|
|
firstPageOffset =
|
|
|
|
(unsigned long)region->virtAddr & (PAGE_SIZE - 1);
|
|
|
|
firstPageSize = PAGE_SIZE - firstPageOffset;
|
|
|
|
|
|
|
|
region->numLockedPages = (firstPageOffset
|
|
|
|
+ region->numBytes +
|
|
|
|
PAGE_SIZE - 1) / PAGE_SIZE;
|
|
|
|
pages =
|
|
|
|
kmalloc(region->numLockedPages *
|
|
|
|
sizeof(struct page *), GFP_KERNEL);
|
|
|
|
|
|
|
|
if (pages == NULL) {
|
|
|
|
region->numLockedPages = 0;
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
userTask = memMap->userTask;
|
|
|
|
|
|
|
|
down_read(&userTask->mm->mmap_sem);
|
|
|
|
rc = get_user_pages(userTask, /* task */
|
|
|
|
userTask->mm, /* mm */
|
|
|
|
(unsigned long)region->virtAddr, /* start */
|
|
|
|
region->numLockedPages, /* len */
|
|
|
|
memMap->dir == DMA_FROM_DEVICE, /* write */
|
|
|
|
0, /* force */
|
|
|
|
pages, /* pages (array of pointers to page) */
|
|
|
|
NULL); /* vmas */
|
|
|
|
up_read(&userTask->mm->mmap_sem);
|
|
|
|
|
|
|
|
if (rc != region->numLockedPages) {
|
|
|
|
kfree(pages);
|
|
|
|
region->numLockedPages = 0;
|
|
|
|
|
|
|
|
if (rc >= 0) {
|
|
|
|
rc = -EINVAL;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
uint8_t *virtAddr = region->virtAddr;
|
|
|
|
size_t bytesRemaining;
|
|
|
|
int pageIdx;
|
|
|
|
|
|
|
|
rc = 0; /* Since get_user_pages returns +ve number */
|
|
|
|
|
|
|
|
region->lockedPages = pages;
|
|
|
|
|
|
|
|
/* We've locked the user pages. Now we need to walk them and figure */
|
|
|
|
/* out the physical addresses. */
|
|
|
|
|
|
|
|
/* The first page may be partial */
|
|
|
|
|
|
|
|
dma_map_add_segment(memMap,
|
|
|
|
region,
|
|
|
|
virtAddr,
|
|
|
|
PFN_PHYS(page_to_pfn
|
|
|
|
(pages[0])) +
|
|
|
|
firstPageOffset,
|
|
|
|
firstPageSize);
|
|
|
|
|
|
|
|
virtAddr += firstPageSize;
|
|
|
|
bytesRemaining =
|
|
|
|
region->numBytes - firstPageSize;
|
|
|
|
|
|
|
|
for (pageIdx = 1;
|
|
|
|
pageIdx < region->numLockedPages;
|
|
|
|
pageIdx++) {
|
|
|
|
size_t bytesThisPage =
|
|
|
|
(bytesRemaining >
|
|
|
|
PAGE_SIZE ? PAGE_SIZE :
|
|
|
|
bytesRemaining);
|
|
|
|
|
|
|
|
DMA_MAP_PRINT
|
|
|
|
("pageIdx:%d pages[pageIdx]=%p pfn=%u phys=%u\n",
|
|
|
|
pageIdx, pages[pageIdx],
|
|
|
|
page_to_pfn(pages[pageIdx]),
|
|
|
|
PFN_PHYS(page_to_pfn
|
|
|
|
(pages[pageIdx])));
|
|
|
|
|
|
|
|
dma_map_add_segment(memMap,
|
|
|
|
region,
|
|
|
|
virtAddr,
|
|
|
|
PFN_PHYS(page_to_pfn
|
|
|
|
(pages
|
|
|
|
[pageIdx])),
|
|
|
|
bytesThisPage);
|
|
|
|
|
|
|
|
virtAddr += bytesThisPage;
|
|
|
|
bytesRemaining -= bytesThisPage;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: User mode pages are not yet supported\n",
|
|
|
|
__func__);
|
|
|
|
|
|
|
|
/* user pages are not physically contiguous */
|
|
|
|
|
|
|
|
rc = -EINVAL;
|
|
|
|
#endif
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
default:
|
|
|
|
{
|
|
|
|
printk(KERN_ERR "%s: Unsupported memory type: %d\n",
|
|
|
|
__func__, region->memType);
|
|
|
|
|
|
|
|
rc = -EINVAL;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rc != 0) {
|
|
|
|
memMap->numRegionsUsed--;
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
|
|
|
|
DMA_MAP_PRINT("returning %d\n", rc);
|
|
|
|
|
|
|
|
up(&memMap->lock);
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_map_add_segment);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Maps in a memory region such that it can be used for performing a DMA.
|
|
|
|
*
|
|
|
|
* @return 0 on success, error code otherwise.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_map_mem(DMA_MemMap_t *memMap, /* Stores state information about the map */
|
|
|
|
void *mem, /* Virtual address that we want to get a map of */
|
|
|
|
size_t numBytes, /* Number of bytes being mapped */
|
|
|
|
enum dma_data_direction dir /* Direction that the mapping will be going */
|
|
|
|
) {
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
rc = dma_map_start(memMap, dir);
|
|
|
|
if (rc == 0) {
|
|
|
|
rc = dma_map_add_region(memMap, mem, numBytes);
|
|
|
|
if (rc < 0) {
|
|
|
|
/* Since the add fails, this function will fail, and the caller won't */
|
|
|
|
/* call unmap, so we need to do it here. */
|
|
|
|
|
|
|
|
dma_unmap(memMap, 0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_map_mem);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Setup a descriptor ring for a given memory map.
|
|
|
|
*
|
|
|
|
* It is assumed that the descriptor ring has already been initialized, and
|
|
|
|
* this routine will only reallocate a new descriptor ring if the existing
|
|
|
|
* one is too small.
|
|
|
|
*
|
|
|
|
* @return 0 on success, error code otherwise.
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_map_create_descriptor_ring(DMA_Device_t dev, /* DMA device (where the ring is stored) */
|
|
|
|
DMA_MemMap_t *memMap, /* Memory map that will be used */
|
|
|
|
dma_addr_t devPhysAddr /* Physical address of device */
|
|
|
|
) {
|
|
|
|
int rc;
|
|
|
|
int numDescriptors;
|
|
|
|
DMA_DeviceAttribute_t *devAttr;
|
|
|
|
DMA_Region_t *region;
|
|
|
|
DMA_Segment_t *segment;
|
|
|
|
dma_addr_t srcPhysAddr;
|
|
|
|
dma_addr_t dstPhysAddr;
|
|
|
|
int regionIdx;
|
|
|
|
int segmentIdx;
|
|
|
|
|
|
|
|
devAttr = &DMA_gDeviceAttribute[dev];
|
|
|
|
|
|
|
|
down(&memMap->lock);
|
|
|
|
|
|
|
|
/* Figure out how many descriptors we need */
|
|
|
|
|
|
|
|
numDescriptors = 0;
|
|
|
|
for (regionIdx = 0; regionIdx < memMap->numRegionsUsed; regionIdx++) {
|
|
|
|
region = &memMap->region[regionIdx];
|
|
|
|
|
|
|
|
for (segmentIdx = 0; segmentIdx < region->numSegmentsUsed;
|
|
|
|
segmentIdx++) {
|
|
|
|
segment = ®ion->segment[segmentIdx];
|
|
|
|
|
|
|
|
if (memMap->dir == DMA_TO_DEVICE) {
|
|
|
|
srcPhysAddr = segment->physAddr;
|
|
|
|
dstPhysAddr = devPhysAddr;
|
|
|
|
} else {
|
|
|
|
srcPhysAddr = devPhysAddr;
|
|
|
|
dstPhysAddr = segment->physAddr;
|
|
|
|
}
|
|
|
|
|
|
|
|
rc =
|
|
|
|
dma_calculate_descriptor_count(dev, srcPhysAddr,
|
|
|
|
dstPhysAddr,
|
|
|
|
segment->
|
|
|
|
numBytes);
|
|
|
|
if (rc < 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: dma_calculate_descriptor_count failed: %d\n",
|
|
|
|
__func__, rc);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
numDescriptors += rc;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Adjust the size of the ring, if it isn't big enough */
|
|
|
|
|
|
|
|
if (numDescriptors > devAttr->ring.descriptorsAllocated) {
|
|
|
|
dma_free_descriptor_ring(&devAttr->ring);
|
|
|
|
rc =
|
|
|
|
dma_alloc_descriptor_ring(&devAttr->ring,
|
|
|
|
numDescriptors);
|
|
|
|
if (rc < 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: dma_alloc_descriptor_ring failed: %d\n",
|
|
|
|
__func__, rc);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
rc =
|
|
|
|
dma_init_descriptor_ring(&devAttr->ring,
|
|
|
|
numDescriptors);
|
|
|
|
if (rc < 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: dma_init_descriptor_ring failed: %d\n",
|
|
|
|
__func__, rc);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Populate the descriptors */
|
|
|
|
|
|
|
|
for (regionIdx = 0; regionIdx < memMap->numRegionsUsed; regionIdx++) {
|
|
|
|
region = &memMap->region[regionIdx];
|
|
|
|
|
|
|
|
for (segmentIdx = 0; segmentIdx < region->numSegmentsUsed;
|
|
|
|
segmentIdx++) {
|
|
|
|
segment = ®ion->segment[segmentIdx];
|
|
|
|
|
|
|
|
if (memMap->dir == DMA_TO_DEVICE) {
|
|
|
|
srcPhysAddr = segment->physAddr;
|
|
|
|
dstPhysAddr = devPhysAddr;
|
|
|
|
} else {
|
|
|
|
srcPhysAddr = devPhysAddr;
|
|
|
|
dstPhysAddr = segment->physAddr;
|
|
|
|
}
|
|
|
|
|
|
|
|
rc =
|
|
|
|
dma_add_descriptors(&devAttr->ring, dev,
|
|
|
|
srcPhysAddr, dstPhysAddr,
|
|
|
|
segment->numBytes);
|
|
|
|
if (rc < 0) {
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: dma_add_descriptors failed: %d\n",
|
|
|
|
__func__, rc);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
rc = 0;
|
|
|
|
|
|
|
|
out:
|
|
|
|
|
|
|
|
up(&memMap->lock);
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_map_create_descriptor_ring);
|
|
|
|
|
|
|
|
/****************************************************************************/
|
|
|
|
/**
|
|
|
|
* Maps in a memory region such that it can be used for performing a DMA.
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
/****************************************************************************/
|
|
|
|
|
|
|
|
int dma_unmap(DMA_MemMap_t *memMap, /* Stores state information about the map */
|
|
|
|
int dirtied /* non-zero if any of the pages were modified */
|
|
|
|
) {
|
2010-04-01 18:13:19 +00:00
|
|
|
|
|
|
|
int rc = 0;
|
2009-08-07 19:01:13 +00:00
|
|
|
int regionIdx;
|
|
|
|
int segmentIdx;
|
|
|
|
DMA_Region_t *region;
|
|
|
|
DMA_Segment_t *segment;
|
|
|
|
|
2010-04-01 18:13:19 +00:00
|
|
|
down(&memMap->lock);
|
|
|
|
|
2009-08-07 19:01:13 +00:00
|
|
|
for (regionIdx = 0; regionIdx < memMap->numRegionsUsed; regionIdx++) {
|
|
|
|
region = &memMap->region[regionIdx];
|
|
|
|
|
|
|
|
for (segmentIdx = 0; segmentIdx < region->numSegmentsUsed;
|
|
|
|
segmentIdx++) {
|
|
|
|
segment = ®ion->segment[segmentIdx];
|
|
|
|
|
|
|
|
switch (region->memType) {
|
|
|
|
case DMA_MEM_TYPE_VMALLOC:
|
|
|
|
{
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: vmalloc'd pages are not yet supported\n",
|
|
|
|
__func__);
|
2010-04-01 18:13:19 +00:00
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
2009-08-07 19:01:13 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
case DMA_MEM_TYPE_KMALLOC:
|
|
|
|
{
|
|
|
|
#if ALLOW_MAP_OF_KMALLOC_MEMORY
|
|
|
|
dma_unmap_single(NULL,
|
|
|
|
segment->physAddr,
|
|
|
|
segment->numBytes,
|
|
|
|
memMap->dir);
|
|
|
|
#endif
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case DMA_MEM_TYPE_DMA:
|
|
|
|
{
|
|
|
|
dma_sync_single_for_cpu(NULL,
|
|
|
|
segment->
|
|
|
|
physAddr,
|
|
|
|
segment->
|
|
|
|
numBytes,
|
|
|
|
memMap->dir);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case DMA_MEM_TYPE_USER:
|
|
|
|
{
|
|
|
|
/* Nothing to do here. */
|
|
|
|
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
default:
|
|
|
|
{
|
|
|
|
printk(KERN_ERR
|
|
|
|
"%s: Unsupported memory type: %d\n",
|
|
|
|
__func__, region->memType);
|
2010-04-01 18:13:19 +00:00
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
2009-08-07 19:01:13 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
segment->virtAddr = NULL;
|
|
|
|
segment->physAddr = 0;
|
|
|
|
segment->numBytes = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (region->numLockedPages > 0) {
|
|
|
|
int pageIdx;
|
|
|
|
|
|
|
|
/* Some user pages were locked. We need to go and unlock them now. */
|
|
|
|
|
|
|
|
for (pageIdx = 0; pageIdx < region->numLockedPages;
|
|
|
|
pageIdx++) {
|
|
|
|
struct page *page =
|
|
|
|
region->lockedPages[pageIdx];
|
|
|
|
|
|
|
|
if (memMap->dir == DMA_FROM_DEVICE) {
|
|
|
|
SetPageDirty(page);
|
|
|
|
}
|
|
|
|
page_cache_release(page);
|
|
|
|
}
|
|
|
|
kfree(region->lockedPages);
|
|
|
|
region->numLockedPages = 0;
|
|
|
|
region->lockedPages = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
region->memType = DMA_MEM_TYPE_NONE;
|
|
|
|
region->virtAddr = NULL;
|
|
|
|
region->numBytes = 0;
|
|
|
|
region->numSegmentsUsed = 0;
|
|
|
|
}
|
|
|
|
memMap->userTask = NULL;
|
|
|
|
memMap->numRegionsUsed = 0;
|
|
|
|
memMap->inUse = 0;
|
|
|
|
|
2010-04-01 18:13:19 +00:00
|
|
|
out:
|
2009-08-07 19:01:13 +00:00
|
|
|
up(&memMap->lock);
|
|
|
|
|
2010-04-01 18:13:19 +00:00
|
|
|
return rc;
|
2009-08-07 19:01:13 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(dma_unmap);
|