mirror of
https://github.com/torvalds/linux.git
synced 2024-12-11 05:33:09 +00:00
37af81c599
This patch introduces the ops table to each memory allocation type (SNDRV_DMA_TYPE_XXX) and abstract the handling for the better code management. Then we get separate the page allocation, release and other tasks for each type, especially for the SG buffer. Each buffer type has now callbacks in the struct snd_malloc_ops, and the common helper functions call those ops accordingly. The former inline code that is specific to SG-buffer is moved into the local sgbuf.c, and we can simplify the PCM code without details of memory handling. Link: https://lore.kernel.org/r/20210609162551.7842-4-tiwai@suse.de Signed-off-by: Takashi Iwai <tiwai@suse.de>
192 lines
4.6 KiB
C
192 lines
4.6 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Scatter-Gather buffer
|
|
*
|
|
* Copyright (c) by Takashi Iwai <tiwai@suse.de>
|
|
*/
|
|
|
|
#include <linux/slab.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/export.h>
|
|
#include <sound/memalloc.h>
|
|
#include "memalloc_local.h"
|
|
|
|
struct snd_sg_page {
|
|
void *buf;
|
|
dma_addr_t addr;
|
|
};
|
|
|
|
struct snd_sg_buf {
|
|
int size; /* allocated byte size */
|
|
int pages; /* allocated pages */
|
|
int tblsize; /* allocated table size */
|
|
struct snd_sg_page *table; /* address table */
|
|
struct page **page_table; /* page table (for vmap/vunmap) */
|
|
struct device *dev;
|
|
};
|
|
|
|
/* table entries are align to 32 */
|
|
#define SGBUF_TBL_ALIGN 32
|
|
#define sgbuf_align_table(tbl) ALIGN((tbl), SGBUF_TBL_ALIGN)
|
|
|
|
static void snd_dma_sg_free(struct snd_dma_buffer *dmab)
|
|
{
|
|
struct snd_sg_buf *sgbuf = dmab->private_data;
|
|
struct snd_dma_buffer tmpb;
|
|
int i;
|
|
|
|
if (!sgbuf)
|
|
return;
|
|
|
|
vunmap(dmab->area);
|
|
dmab->area = NULL;
|
|
|
|
tmpb.dev.type = SNDRV_DMA_TYPE_DEV;
|
|
if (dmab->dev.type == SNDRV_DMA_TYPE_DEV_UC_SG)
|
|
tmpb.dev.type = SNDRV_DMA_TYPE_DEV_UC;
|
|
tmpb.dev.dev = sgbuf->dev;
|
|
for (i = 0; i < sgbuf->pages; i++) {
|
|
if (!(sgbuf->table[i].addr & ~PAGE_MASK))
|
|
continue; /* continuous pages */
|
|
tmpb.area = sgbuf->table[i].buf;
|
|
tmpb.addr = sgbuf->table[i].addr & PAGE_MASK;
|
|
tmpb.bytes = (sgbuf->table[i].addr & ~PAGE_MASK) << PAGE_SHIFT;
|
|
snd_dma_free_pages(&tmpb);
|
|
}
|
|
|
|
kfree(sgbuf->table);
|
|
kfree(sgbuf->page_table);
|
|
kfree(sgbuf);
|
|
dmab->private_data = NULL;
|
|
}
|
|
|
|
#define MAX_ALLOC_PAGES 32
|
|
|
|
static int snd_dma_sg_alloc(struct snd_dma_buffer *dmab, size_t size)
|
|
{
|
|
struct snd_sg_buf *sgbuf;
|
|
unsigned int i, pages, chunk, maxpages;
|
|
struct snd_dma_buffer tmpb;
|
|
struct snd_sg_page *table;
|
|
struct page **pgtable;
|
|
int type = SNDRV_DMA_TYPE_DEV;
|
|
pgprot_t prot = PAGE_KERNEL;
|
|
|
|
dmab->private_data = sgbuf = kzalloc(sizeof(*sgbuf), GFP_KERNEL);
|
|
if (!sgbuf)
|
|
return -ENOMEM;
|
|
if (dmab->dev.type == SNDRV_DMA_TYPE_DEV_UC_SG) {
|
|
type = SNDRV_DMA_TYPE_DEV_UC;
|
|
#ifdef pgprot_noncached
|
|
prot = pgprot_noncached(PAGE_KERNEL);
|
|
#endif
|
|
}
|
|
sgbuf->dev = dmab->dev.dev;
|
|
pages = snd_sgbuf_aligned_pages(size);
|
|
sgbuf->tblsize = sgbuf_align_table(pages);
|
|
table = kcalloc(sgbuf->tblsize, sizeof(*table), GFP_KERNEL);
|
|
if (!table)
|
|
goto _failed;
|
|
sgbuf->table = table;
|
|
pgtable = kcalloc(sgbuf->tblsize, sizeof(*pgtable), GFP_KERNEL);
|
|
if (!pgtable)
|
|
goto _failed;
|
|
sgbuf->page_table = pgtable;
|
|
|
|
/* allocate pages */
|
|
maxpages = MAX_ALLOC_PAGES;
|
|
while (pages > 0) {
|
|
chunk = pages;
|
|
/* don't be too eager to take a huge chunk */
|
|
if (chunk > maxpages)
|
|
chunk = maxpages;
|
|
chunk <<= PAGE_SHIFT;
|
|
if (snd_dma_alloc_pages_fallback(type, dmab->dev.dev,
|
|
chunk, &tmpb) < 0) {
|
|
if (!sgbuf->pages)
|
|
goto _failed;
|
|
size = sgbuf->pages * PAGE_SIZE;
|
|
break;
|
|
}
|
|
chunk = tmpb.bytes >> PAGE_SHIFT;
|
|
for (i = 0; i < chunk; i++) {
|
|
table->buf = tmpb.area;
|
|
table->addr = tmpb.addr;
|
|
if (!i)
|
|
table->addr |= chunk; /* mark head */
|
|
table++;
|
|
*pgtable++ = virt_to_page(tmpb.area);
|
|
tmpb.area += PAGE_SIZE;
|
|
tmpb.addr += PAGE_SIZE;
|
|
}
|
|
sgbuf->pages += chunk;
|
|
pages -= chunk;
|
|
if (chunk < maxpages)
|
|
maxpages = chunk;
|
|
}
|
|
|
|
sgbuf->size = size;
|
|
dmab->area = vmap(sgbuf->page_table, sgbuf->pages, VM_MAP, prot);
|
|
if (! dmab->area)
|
|
goto _failed;
|
|
return 0;
|
|
|
|
_failed:
|
|
snd_dma_sg_free(dmab); /* free the table */
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static dma_addr_t snd_dma_sg_get_addr(struct snd_dma_buffer *dmab,
|
|
size_t offset)
|
|
{
|
|
struct snd_sg_buf *sgbuf = dmab->private_data;
|
|
dma_addr_t addr;
|
|
|
|
addr = sgbuf->table[offset >> PAGE_SHIFT].addr;
|
|
addr &= ~((dma_addr_t)PAGE_SIZE - 1);
|
|
return addr + offset % PAGE_SIZE;
|
|
}
|
|
|
|
static struct page *snd_dma_sg_get_page(struct snd_dma_buffer *dmab,
|
|
size_t offset)
|
|
{
|
|
struct snd_sg_buf *sgbuf = dmab->private_data;
|
|
unsigned int idx = offset >> PAGE_SHIFT;
|
|
|
|
if (idx >= (unsigned int)sgbuf->pages)
|
|
return NULL;
|
|
return sgbuf->page_table[idx];
|
|
}
|
|
|
|
static unsigned int snd_dma_sg_get_chunk_size(struct snd_dma_buffer *dmab,
|
|
unsigned int ofs,
|
|
unsigned int size)
|
|
{
|
|
struct snd_sg_buf *sg = dmab->private_data;
|
|
unsigned int start, end, pg;
|
|
|
|
start = ofs >> PAGE_SHIFT;
|
|
end = (ofs + size - 1) >> PAGE_SHIFT;
|
|
/* check page continuity */
|
|
pg = sg->table[start].addr >> PAGE_SHIFT;
|
|
for (;;) {
|
|
start++;
|
|
if (start > end)
|
|
break;
|
|
pg++;
|
|
if ((sg->table[start].addr >> PAGE_SHIFT) != pg)
|
|
return (start << PAGE_SHIFT) - ofs;
|
|
}
|
|
/* ok, all on continuous pages */
|
|
return size;
|
|
}
|
|
|
|
const struct snd_malloc_ops snd_dma_sg_ops = {
|
|
.alloc = snd_dma_sg_alloc,
|
|
.free = snd_dma_sg_free,
|
|
.get_addr = snd_dma_sg_get_addr,
|
|
.get_page = snd_dma_sg_get_page,
|
|
.get_chunk_size = snd_dma_sg_get_chunk_size,
|
|
};
|