mirror of
https://github.com/torvalds/linux.git
synced 2024-12-01 16:41:39 +00:00
1b13d196d2
The .remove() callback for a platform driver returns an int which makes many driver authors wrongly assume it's possible to do error handling by returning an error code. However the value returned is (mostly) ignored and this typically results in resource leaks. To improve here there is a quest to make the remove callback return void. In the first step of this quest all drivers are converted to .remove_new() which already returns void. Trivially convert this driver from always returning zero in the remove callback to the void returning variant. Signed-off-by: Uwe Kleine-König <u.kleine-koenig@pengutronix.de> Link: https://lore.kernel.org/r/20230303172041.2103336-84-u.kleine-koenig@pengutronix.de Signed-off-by: Mark Brown <broonie@kernel.org>
811 lines
20 KiB
C
811 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
// spi-uniphier.c - Socionext UniPhier SPI controller driver
|
|
// Copyright 2012 Panasonic Corporation
|
|
// Copyright 2016-2018 Socionext Inc.
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/bitfield.h>
|
|
#include <linux/bitops.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/dmaengine.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/io.h>
|
|
#include <linux/module.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/spi/spi.h>
|
|
|
|
#include <asm/unaligned.h>
|
|
|
|
#define SSI_TIMEOUT_MS 2000
|
|
#define SSI_POLL_TIMEOUT_US 200
|
|
#define SSI_MAX_CLK_DIVIDER 254
|
|
#define SSI_MIN_CLK_DIVIDER 4
|
|
|
|
struct uniphier_spi_priv {
|
|
void __iomem *base;
|
|
dma_addr_t base_dma_addr;
|
|
struct clk *clk;
|
|
struct spi_master *master;
|
|
struct completion xfer_done;
|
|
|
|
int error;
|
|
unsigned int tx_bytes;
|
|
unsigned int rx_bytes;
|
|
const u8 *tx_buf;
|
|
u8 *rx_buf;
|
|
atomic_t dma_busy;
|
|
|
|
bool is_save_param;
|
|
u8 bits_per_word;
|
|
u16 mode;
|
|
u32 speed_hz;
|
|
};
|
|
|
|
#define SSI_CTL 0x00
|
|
#define SSI_CTL_EN BIT(0)
|
|
|
|
#define SSI_CKS 0x04
|
|
#define SSI_CKS_CKRAT_MASK GENMASK(7, 0)
|
|
#define SSI_CKS_CKPHS BIT(14)
|
|
#define SSI_CKS_CKINIT BIT(13)
|
|
#define SSI_CKS_CKDLY BIT(12)
|
|
|
|
#define SSI_TXWDS 0x08
|
|
#define SSI_TXWDS_WDLEN_MASK GENMASK(13, 8)
|
|
#define SSI_TXWDS_TDTF_MASK GENMASK(7, 6)
|
|
#define SSI_TXWDS_DTLEN_MASK GENMASK(5, 0)
|
|
|
|
#define SSI_RXWDS 0x0c
|
|
#define SSI_RXWDS_DTLEN_MASK GENMASK(5, 0)
|
|
|
|
#define SSI_FPS 0x10
|
|
#define SSI_FPS_FSPOL BIT(15)
|
|
#define SSI_FPS_FSTRT BIT(14)
|
|
|
|
#define SSI_SR 0x14
|
|
#define SSI_SR_BUSY BIT(7)
|
|
#define SSI_SR_RNE BIT(0)
|
|
|
|
#define SSI_IE 0x18
|
|
#define SSI_IE_TCIE BIT(4)
|
|
#define SSI_IE_RCIE BIT(3)
|
|
#define SSI_IE_TXRE BIT(2)
|
|
#define SSI_IE_RXRE BIT(1)
|
|
#define SSI_IE_RORIE BIT(0)
|
|
#define SSI_IE_ALL_MASK GENMASK(4, 0)
|
|
|
|
#define SSI_IS 0x1c
|
|
#define SSI_IS_RXRS BIT(9)
|
|
#define SSI_IS_RCID BIT(3)
|
|
#define SSI_IS_RORID BIT(0)
|
|
|
|
#define SSI_IC 0x1c
|
|
#define SSI_IC_TCIC BIT(4)
|
|
#define SSI_IC_RCIC BIT(3)
|
|
#define SSI_IC_RORIC BIT(0)
|
|
|
|
#define SSI_FC 0x20
|
|
#define SSI_FC_TXFFL BIT(12)
|
|
#define SSI_FC_TXFTH_MASK GENMASK(11, 8)
|
|
#define SSI_FC_RXFFL BIT(4)
|
|
#define SSI_FC_RXFTH_MASK GENMASK(3, 0)
|
|
|
|
#define SSI_TXDR 0x24
|
|
#define SSI_RXDR 0x24
|
|
|
|
#define SSI_FIFO_DEPTH 8U
|
|
#define SSI_FIFO_BURST_NUM 1
|
|
|
|
#define SSI_DMA_RX_BUSY BIT(1)
|
|
#define SSI_DMA_TX_BUSY BIT(0)
|
|
|
|
static inline unsigned int bytes_per_word(unsigned int bits)
|
|
{
|
|
return bits <= 8 ? 1 : (bits <= 16 ? 2 : 4);
|
|
}
|
|
|
|
static inline void uniphier_spi_irq_enable(struct uniphier_spi_priv *priv,
|
|
u32 mask)
|
|
{
|
|
u32 val;
|
|
|
|
val = readl(priv->base + SSI_IE);
|
|
val |= mask;
|
|
writel(val, priv->base + SSI_IE);
|
|
}
|
|
|
|
static inline void uniphier_spi_irq_disable(struct uniphier_spi_priv *priv,
|
|
u32 mask)
|
|
{
|
|
u32 val;
|
|
|
|
val = readl(priv->base + SSI_IE);
|
|
val &= ~mask;
|
|
writel(val, priv->base + SSI_IE);
|
|
}
|
|
|
|
static void uniphier_spi_set_mode(struct spi_device *spi)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(spi->master);
|
|
u32 val1, val2;
|
|
|
|
/*
|
|
* clock setting
|
|
* CKPHS capture timing. 0:rising edge, 1:falling edge
|
|
* CKINIT clock initial level. 0:low, 1:high
|
|
* CKDLY clock delay. 0:no delay, 1:delay depending on FSTRT
|
|
* (FSTRT=0: 1 clock, FSTRT=1: 0.5 clock)
|
|
*
|
|
* frame setting
|
|
* FSPOL frame signal porarity. 0: low, 1: high
|
|
* FSTRT start frame timing
|
|
* 0: rising edge of clock, 1: falling edge of clock
|
|
*/
|
|
switch (spi->mode & SPI_MODE_X_MASK) {
|
|
case SPI_MODE_0:
|
|
/* CKPHS=1, CKINIT=0, CKDLY=1, FSTRT=0 */
|
|
val1 = SSI_CKS_CKPHS | SSI_CKS_CKDLY;
|
|
val2 = 0;
|
|
break;
|
|
case SPI_MODE_1:
|
|
/* CKPHS=0, CKINIT=0, CKDLY=0, FSTRT=1 */
|
|
val1 = 0;
|
|
val2 = SSI_FPS_FSTRT;
|
|
break;
|
|
case SPI_MODE_2:
|
|
/* CKPHS=0, CKINIT=1, CKDLY=1, FSTRT=1 */
|
|
val1 = SSI_CKS_CKINIT | SSI_CKS_CKDLY;
|
|
val2 = SSI_FPS_FSTRT;
|
|
break;
|
|
case SPI_MODE_3:
|
|
/* CKPHS=1, CKINIT=1, CKDLY=0, FSTRT=0 */
|
|
val1 = SSI_CKS_CKPHS | SSI_CKS_CKINIT;
|
|
val2 = 0;
|
|
break;
|
|
}
|
|
|
|
if (!(spi->mode & SPI_CS_HIGH))
|
|
val2 |= SSI_FPS_FSPOL;
|
|
|
|
writel(val1, priv->base + SSI_CKS);
|
|
writel(val2, priv->base + SSI_FPS);
|
|
|
|
val1 = 0;
|
|
if (spi->mode & SPI_LSB_FIRST)
|
|
val1 |= FIELD_PREP(SSI_TXWDS_TDTF_MASK, 1);
|
|
writel(val1, priv->base + SSI_TXWDS);
|
|
writel(val1, priv->base + SSI_RXWDS);
|
|
}
|
|
|
|
static void uniphier_spi_set_transfer_size(struct spi_device *spi, int size)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(spi->master);
|
|
u32 val;
|
|
|
|
val = readl(priv->base + SSI_TXWDS);
|
|
val &= ~(SSI_TXWDS_WDLEN_MASK | SSI_TXWDS_DTLEN_MASK);
|
|
val |= FIELD_PREP(SSI_TXWDS_WDLEN_MASK, size);
|
|
val |= FIELD_PREP(SSI_TXWDS_DTLEN_MASK, size);
|
|
writel(val, priv->base + SSI_TXWDS);
|
|
|
|
val = readl(priv->base + SSI_RXWDS);
|
|
val &= ~SSI_RXWDS_DTLEN_MASK;
|
|
val |= FIELD_PREP(SSI_RXWDS_DTLEN_MASK, size);
|
|
writel(val, priv->base + SSI_RXWDS);
|
|
}
|
|
|
|
static void uniphier_spi_set_baudrate(struct spi_device *spi,
|
|
unsigned int speed)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(spi->master);
|
|
u32 val, ckdiv;
|
|
|
|
/*
|
|
* the supported rates are even numbers from 4 to 254. (4,6,8...254)
|
|
* round up as we look for equal or less speed
|
|
*/
|
|
ckdiv = DIV_ROUND_UP(clk_get_rate(priv->clk), speed);
|
|
ckdiv = round_up(ckdiv, 2);
|
|
|
|
val = readl(priv->base + SSI_CKS);
|
|
val &= ~SSI_CKS_CKRAT_MASK;
|
|
val |= ckdiv & SSI_CKS_CKRAT_MASK;
|
|
writel(val, priv->base + SSI_CKS);
|
|
}
|
|
|
|
static void uniphier_spi_setup_transfer(struct spi_device *spi,
|
|
struct spi_transfer *t)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(spi->master);
|
|
u32 val;
|
|
|
|
priv->error = 0;
|
|
priv->tx_buf = t->tx_buf;
|
|
priv->rx_buf = t->rx_buf;
|
|
priv->tx_bytes = priv->rx_bytes = t->len;
|
|
|
|
if (!priv->is_save_param || priv->mode != spi->mode) {
|
|
uniphier_spi_set_mode(spi);
|
|
priv->mode = spi->mode;
|
|
priv->is_save_param = false;
|
|
}
|
|
|
|
if (!priv->is_save_param || priv->bits_per_word != t->bits_per_word) {
|
|
uniphier_spi_set_transfer_size(spi, t->bits_per_word);
|
|
priv->bits_per_word = t->bits_per_word;
|
|
}
|
|
|
|
if (!priv->is_save_param || priv->speed_hz != t->speed_hz) {
|
|
uniphier_spi_set_baudrate(spi, t->speed_hz);
|
|
priv->speed_hz = t->speed_hz;
|
|
}
|
|
|
|
priv->is_save_param = true;
|
|
|
|
/* reset FIFOs */
|
|
val = SSI_FC_TXFFL | SSI_FC_RXFFL;
|
|
writel(val, priv->base + SSI_FC);
|
|
}
|
|
|
|
static void uniphier_spi_send(struct uniphier_spi_priv *priv)
|
|
{
|
|
int wsize;
|
|
u32 val = 0;
|
|
|
|
wsize = min(bytes_per_word(priv->bits_per_word), priv->tx_bytes);
|
|
priv->tx_bytes -= wsize;
|
|
|
|
if (priv->tx_buf) {
|
|
switch (wsize) {
|
|
case 1:
|
|
val = *priv->tx_buf;
|
|
break;
|
|
case 2:
|
|
val = get_unaligned_le16(priv->tx_buf);
|
|
break;
|
|
case 4:
|
|
val = get_unaligned_le32(priv->tx_buf);
|
|
break;
|
|
}
|
|
|
|
priv->tx_buf += wsize;
|
|
}
|
|
|
|
writel(val, priv->base + SSI_TXDR);
|
|
}
|
|
|
|
static void uniphier_spi_recv(struct uniphier_spi_priv *priv)
|
|
{
|
|
int rsize;
|
|
u32 val;
|
|
|
|
rsize = min(bytes_per_word(priv->bits_per_word), priv->rx_bytes);
|
|
priv->rx_bytes -= rsize;
|
|
|
|
val = readl(priv->base + SSI_RXDR);
|
|
|
|
if (priv->rx_buf) {
|
|
switch (rsize) {
|
|
case 1:
|
|
*priv->rx_buf = val;
|
|
break;
|
|
case 2:
|
|
put_unaligned_le16(val, priv->rx_buf);
|
|
break;
|
|
case 4:
|
|
put_unaligned_le32(val, priv->rx_buf);
|
|
break;
|
|
}
|
|
|
|
priv->rx_buf += rsize;
|
|
}
|
|
}
|
|
|
|
static void uniphier_spi_set_fifo_threshold(struct uniphier_spi_priv *priv,
|
|
unsigned int threshold)
|
|
{
|
|
u32 val;
|
|
|
|
val = readl(priv->base + SSI_FC);
|
|
val &= ~(SSI_FC_TXFTH_MASK | SSI_FC_RXFTH_MASK);
|
|
val |= FIELD_PREP(SSI_FC_TXFTH_MASK, SSI_FIFO_DEPTH - threshold);
|
|
val |= FIELD_PREP(SSI_FC_RXFTH_MASK, threshold);
|
|
writel(val, priv->base + SSI_FC);
|
|
}
|
|
|
|
static void uniphier_spi_fill_tx_fifo(struct uniphier_spi_priv *priv)
|
|
{
|
|
unsigned int fifo_threshold, fill_words;
|
|
unsigned int bpw = bytes_per_word(priv->bits_per_word);
|
|
|
|
fifo_threshold = DIV_ROUND_UP(priv->rx_bytes, bpw);
|
|
fifo_threshold = min(fifo_threshold, SSI_FIFO_DEPTH);
|
|
|
|
uniphier_spi_set_fifo_threshold(priv, fifo_threshold);
|
|
|
|
fill_words = fifo_threshold -
|
|
DIV_ROUND_UP(priv->rx_bytes - priv->tx_bytes, bpw);
|
|
|
|
while (fill_words--)
|
|
uniphier_spi_send(priv);
|
|
}
|
|
|
|
static void uniphier_spi_set_cs(struct spi_device *spi, bool enable)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(spi->master);
|
|
u32 val;
|
|
|
|
val = readl(priv->base + SSI_FPS);
|
|
|
|
if (enable)
|
|
val |= SSI_FPS_FSPOL;
|
|
else
|
|
val &= ~SSI_FPS_FSPOL;
|
|
|
|
writel(val, priv->base + SSI_FPS);
|
|
}
|
|
|
|
static bool uniphier_spi_can_dma(struct spi_master *master,
|
|
struct spi_device *spi,
|
|
struct spi_transfer *t)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
unsigned int bpw = bytes_per_word(priv->bits_per_word);
|
|
|
|
if ((!master->dma_tx && !master->dma_rx)
|
|
|| (!master->dma_tx && t->tx_buf)
|
|
|| (!master->dma_rx && t->rx_buf))
|
|
return false;
|
|
|
|
return DIV_ROUND_UP(t->len, bpw) > SSI_FIFO_DEPTH;
|
|
}
|
|
|
|
static void uniphier_spi_dma_rxcb(void *data)
|
|
{
|
|
struct spi_master *master = data;
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
int state = atomic_fetch_andnot(SSI_DMA_RX_BUSY, &priv->dma_busy);
|
|
|
|
uniphier_spi_irq_disable(priv, SSI_IE_RXRE);
|
|
|
|
if (!(state & SSI_DMA_TX_BUSY))
|
|
spi_finalize_current_transfer(master);
|
|
}
|
|
|
|
static void uniphier_spi_dma_txcb(void *data)
|
|
{
|
|
struct spi_master *master = data;
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
int state = atomic_fetch_andnot(SSI_DMA_TX_BUSY, &priv->dma_busy);
|
|
|
|
uniphier_spi_irq_disable(priv, SSI_IE_TXRE);
|
|
|
|
if (!(state & SSI_DMA_RX_BUSY))
|
|
spi_finalize_current_transfer(master);
|
|
}
|
|
|
|
static int uniphier_spi_transfer_one_dma(struct spi_master *master,
|
|
struct spi_device *spi,
|
|
struct spi_transfer *t)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
struct dma_async_tx_descriptor *rxdesc = NULL, *txdesc = NULL;
|
|
int buswidth;
|
|
|
|
atomic_set(&priv->dma_busy, 0);
|
|
|
|
uniphier_spi_set_fifo_threshold(priv, SSI_FIFO_BURST_NUM);
|
|
|
|
if (priv->bits_per_word <= 8)
|
|
buswidth = DMA_SLAVE_BUSWIDTH_1_BYTE;
|
|
else if (priv->bits_per_word <= 16)
|
|
buswidth = DMA_SLAVE_BUSWIDTH_2_BYTES;
|
|
else
|
|
buswidth = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
|
|
|
if (priv->rx_buf) {
|
|
struct dma_slave_config rxconf = {
|
|
.direction = DMA_DEV_TO_MEM,
|
|
.src_addr = priv->base_dma_addr + SSI_RXDR,
|
|
.src_addr_width = buswidth,
|
|
.src_maxburst = SSI_FIFO_BURST_NUM,
|
|
};
|
|
|
|
dmaengine_slave_config(master->dma_rx, &rxconf);
|
|
|
|
rxdesc = dmaengine_prep_slave_sg(
|
|
master->dma_rx,
|
|
t->rx_sg.sgl, t->rx_sg.nents,
|
|
DMA_DEV_TO_MEM, DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
|
if (!rxdesc)
|
|
goto out_err_prep;
|
|
|
|
rxdesc->callback = uniphier_spi_dma_rxcb;
|
|
rxdesc->callback_param = master;
|
|
|
|
uniphier_spi_irq_enable(priv, SSI_IE_RXRE);
|
|
atomic_or(SSI_DMA_RX_BUSY, &priv->dma_busy);
|
|
|
|
dmaengine_submit(rxdesc);
|
|
dma_async_issue_pending(master->dma_rx);
|
|
}
|
|
|
|
if (priv->tx_buf) {
|
|
struct dma_slave_config txconf = {
|
|
.direction = DMA_MEM_TO_DEV,
|
|
.dst_addr = priv->base_dma_addr + SSI_TXDR,
|
|
.dst_addr_width = buswidth,
|
|
.dst_maxburst = SSI_FIFO_BURST_NUM,
|
|
};
|
|
|
|
dmaengine_slave_config(master->dma_tx, &txconf);
|
|
|
|
txdesc = dmaengine_prep_slave_sg(
|
|
master->dma_tx,
|
|
t->tx_sg.sgl, t->tx_sg.nents,
|
|
DMA_MEM_TO_DEV, DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
|
if (!txdesc)
|
|
goto out_err_prep;
|
|
|
|
txdesc->callback = uniphier_spi_dma_txcb;
|
|
txdesc->callback_param = master;
|
|
|
|
uniphier_spi_irq_enable(priv, SSI_IE_TXRE);
|
|
atomic_or(SSI_DMA_TX_BUSY, &priv->dma_busy);
|
|
|
|
dmaengine_submit(txdesc);
|
|
dma_async_issue_pending(master->dma_tx);
|
|
}
|
|
|
|
/* signal that we need to wait for completion */
|
|
return (priv->tx_buf || priv->rx_buf);
|
|
|
|
out_err_prep:
|
|
if (rxdesc)
|
|
dmaengine_terminate_sync(master->dma_rx);
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static int uniphier_spi_transfer_one_irq(struct spi_master *master,
|
|
struct spi_device *spi,
|
|
struct spi_transfer *t)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
struct device *dev = master->dev.parent;
|
|
unsigned long time_left;
|
|
|
|
reinit_completion(&priv->xfer_done);
|
|
|
|
uniphier_spi_fill_tx_fifo(priv);
|
|
|
|
uniphier_spi_irq_enable(priv, SSI_IE_RCIE | SSI_IE_RORIE);
|
|
|
|
time_left = wait_for_completion_timeout(&priv->xfer_done,
|
|
msecs_to_jiffies(SSI_TIMEOUT_MS));
|
|
|
|
uniphier_spi_irq_disable(priv, SSI_IE_RCIE | SSI_IE_RORIE);
|
|
|
|
if (!time_left) {
|
|
dev_err(dev, "transfer timeout.\n");
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
return priv->error;
|
|
}
|
|
|
|
static int uniphier_spi_transfer_one_poll(struct spi_master *master,
|
|
struct spi_device *spi,
|
|
struct spi_transfer *t)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
int loop = SSI_POLL_TIMEOUT_US * 10;
|
|
|
|
while (priv->tx_bytes) {
|
|
uniphier_spi_fill_tx_fifo(priv);
|
|
|
|
while ((priv->rx_bytes - priv->tx_bytes) > 0) {
|
|
while (!(readl(priv->base + SSI_SR) & SSI_SR_RNE)
|
|
&& loop--)
|
|
ndelay(100);
|
|
|
|
if (loop == -1)
|
|
goto irq_transfer;
|
|
|
|
uniphier_spi_recv(priv);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
irq_transfer:
|
|
return uniphier_spi_transfer_one_irq(master, spi, t);
|
|
}
|
|
|
|
static int uniphier_spi_transfer_one(struct spi_master *master,
|
|
struct spi_device *spi,
|
|
struct spi_transfer *t)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
unsigned long threshold;
|
|
bool use_dma;
|
|
|
|
/* Terminate and return success for 0 byte length transfer */
|
|
if (!t->len)
|
|
return 0;
|
|
|
|
uniphier_spi_setup_transfer(spi, t);
|
|
|
|
use_dma = master->can_dma ? master->can_dma(master, spi, t) : false;
|
|
if (use_dma)
|
|
return uniphier_spi_transfer_one_dma(master, spi, t);
|
|
|
|
/*
|
|
* If the transfer operation will take longer than
|
|
* SSI_POLL_TIMEOUT_US, it should use irq.
|
|
*/
|
|
threshold = DIV_ROUND_UP(SSI_POLL_TIMEOUT_US * priv->speed_hz,
|
|
USEC_PER_SEC * BITS_PER_BYTE);
|
|
if (t->len > threshold)
|
|
return uniphier_spi_transfer_one_irq(master, spi, t);
|
|
else
|
|
return uniphier_spi_transfer_one_poll(master, spi, t);
|
|
}
|
|
|
|
static int uniphier_spi_prepare_transfer_hardware(struct spi_master *master)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
|
|
writel(SSI_CTL_EN, priv->base + SSI_CTL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int uniphier_spi_unprepare_transfer_hardware(struct spi_master *master)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
|
|
writel(0, priv->base + SSI_CTL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void uniphier_spi_handle_err(struct spi_master *master,
|
|
struct spi_message *msg)
|
|
{
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
u32 val;
|
|
|
|
/* stop running spi transfer */
|
|
writel(0, priv->base + SSI_CTL);
|
|
|
|
/* reset FIFOs */
|
|
val = SSI_FC_TXFFL | SSI_FC_RXFFL;
|
|
writel(val, priv->base + SSI_FC);
|
|
|
|
uniphier_spi_irq_disable(priv, SSI_IE_ALL_MASK);
|
|
|
|
if (atomic_read(&priv->dma_busy) & SSI_DMA_TX_BUSY) {
|
|
dmaengine_terminate_async(master->dma_tx);
|
|
atomic_andnot(SSI_DMA_TX_BUSY, &priv->dma_busy);
|
|
}
|
|
|
|
if (atomic_read(&priv->dma_busy) & SSI_DMA_RX_BUSY) {
|
|
dmaengine_terminate_async(master->dma_rx);
|
|
atomic_andnot(SSI_DMA_RX_BUSY, &priv->dma_busy);
|
|
}
|
|
}
|
|
|
|
static irqreturn_t uniphier_spi_handler(int irq, void *dev_id)
|
|
{
|
|
struct uniphier_spi_priv *priv = dev_id;
|
|
u32 val, stat;
|
|
|
|
stat = readl(priv->base + SSI_IS);
|
|
val = SSI_IC_TCIC | SSI_IC_RCIC | SSI_IC_RORIC;
|
|
writel(val, priv->base + SSI_IC);
|
|
|
|
/* rx fifo overrun */
|
|
if (stat & SSI_IS_RORID) {
|
|
priv->error = -EIO;
|
|
goto done;
|
|
}
|
|
|
|
/* rx complete */
|
|
if ((stat & SSI_IS_RCID) && (stat & SSI_IS_RXRS)) {
|
|
while ((readl(priv->base + SSI_SR) & SSI_SR_RNE) &&
|
|
(priv->rx_bytes - priv->tx_bytes) > 0)
|
|
uniphier_spi_recv(priv);
|
|
|
|
if ((readl(priv->base + SSI_SR) & SSI_SR_RNE) ||
|
|
(priv->rx_bytes != priv->tx_bytes)) {
|
|
priv->error = -EIO;
|
|
goto done;
|
|
} else if (priv->rx_bytes == 0)
|
|
goto done;
|
|
|
|
/* next tx transfer */
|
|
uniphier_spi_fill_tx_fifo(priv);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
return IRQ_NONE;
|
|
|
|
done:
|
|
complete(&priv->xfer_done);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static int uniphier_spi_probe(struct platform_device *pdev)
|
|
{
|
|
struct uniphier_spi_priv *priv;
|
|
struct spi_master *master;
|
|
struct resource *res;
|
|
struct dma_slave_caps caps;
|
|
u32 dma_tx_burst = 0, dma_rx_burst = 0;
|
|
unsigned long clk_rate;
|
|
int irq;
|
|
int ret;
|
|
|
|
master = spi_alloc_master(&pdev->dev, sizeof(*priv));
|
|
if (!master)
|
|
return -ENOMEM;
|
|
|
|
platform_set_drvdata(pdev, master);
|
|
|
|
priv = spi_master_get_devdata(master);
|
|
priv->master = master;
|
|
priv->is_save_param = false;
|
|
|
|
priv->base = devm_platform_get_and_ioremap_resource(pdev, 0, &res);
|
|
if (IS_ERR(priv->base)) {
|
|
ret = PTR_ERR(priv->base);
|
|
goto out_master_put;
|
|
}
|
|
priv->base_dma_addr = res->start;
|
|
|
|
priv->clk = devm_clk_get(&pdev->dev, NULL);
|
|
if (IS_ERR(priv->clk)) {
|
|
dev_err(&pdev->dev, "failed to get clock\n");
|
|
ret = PTR_ERR(priv->clk);
|
|
goto out_master_put;
|
|
}
|
|
|
|
ret = clk_prepare_enable(priv->clk);
|
|
if (ret)
|
|
goto out_master_put;
|
|
|
|
irq = platform_get_irq(pdev, 0);
|
|
if (irq < 0) {
|
|
ret = irq;
|
|
goto out_disable_clk;
|
|
}
|
|
|
|
ret = devm_request_irq(&pdev->dev, irq, uniphier_spi_handler,
|
|
0, "uniphier-spi", priv);
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "failed to request IRQ\n");
|
|
goto out_disable_clk;
|
|
}
|
|
|
|
init_completion(&priv->xfer_done);
|
|
|
|
clk_rate = clk_get_rate(priv->clk);
|
|
|
|
master->max_speed_hz = DIV_ROUND_UP(clk_rate, SSI_MIN_CLK_DIVIDER);
|
|
master->min_speed_hz = DIV_ROUND_UP(clk_rate, SSI_MAX_CLK_DIVIDER);
|
|
master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_CS_HIGH | SPI_LSB_FIRST;
|
|
master->dev.of_node = pdev->dev.of_node;
|
|
master->bus_num = pdev->id;
|
|
master->bits_per_word_mask = SPI_BPW_RANGE_MASK(1, 32);
|
|
|
|
master->set_cs = uniphier_spi_set_cs;
|
|
master->transfer_one = uniphier_spi_transfer_one;
|
|
master->prepare_transfer_hardware
|
|
= uniphier_spi_prepare_transfer_hardware;
|
|
master->unprepare_transfer_hardware
|
|
= uniphier_spi_unprepare_transfer_hardware;
|
|
master->handle_err = uniphier_spi_handle_err;
|
|
master->can_dma = uniphier_spi_can_dma;
|
|
|
|
master->num_chipselect = 1;
|
|
master->flags = SPI_CONTROLLER_MUST_RX | SPI_CONTROLLER_MUST_TX;
|
|
|
|
master->dma_tx = dma_request_chan(&pdev->dev, "tx");
|
|
if (IS_ERR_OR_NULL(master->dma_tx)) {
|
|
if (PTR_ERR(master->dma_tx) == -EPROBE_DEFER) {
|
|
ret = -EPROBE_DEFER;
|
|
goto out_disable_clk;
|
|
}
|
|
master->dma_tx = NULL;
|
|
dma_tx_burst = INT_MAX;
|
|
} else {
|
|
ret = dma_get_slave_caps(master->dma_tx, &caps);
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "failed to get TX DMA capacities: %d\n",
|
|
ret);
|
|
goto out_release_dma;
|
|
}
|
|
dma_tx_burst = caps.max_burst;
|
|
}
|
|
|
|
master->dma_rx = dma_request_chan(&pdev->dev, "rx");
|
|
if (IS_ERR_OR_NULL(master->dma_rx)) {
|
|
if (PTR_ERR(master->dma_rx) == -EPROBE_DEFER) {
|
|
ret = -EPROBE_DEFER;
|
|
goto out_release_dma;
|
|
}
|
|
master->dma_rx = NULL;
|
|
dma_rx_burst = INT_MAX;
|
|
} else {
|
|
ret = dma_get_slave_caps(master->dma_rx, &caps);
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "failed to get RX DMA capacities: %d\n",
|
|
ret);
|
|
goto out_release_dma;
|
|
}
|
|
dma_rx_burst = caps.max_burst;
|
|
}
|
|
|
|
master->max_dma_len = min(dma_tx_burst, dma_rx_burst);
|
|
|
|
ret = devm_spi_register_master(&pdev->dev, master);
|
|
if (ret)
|
|
goto out_release_dma;
|
|
|
|
return 0;
|
|
|
|
out_release_dma:
|
|
if (!IS_ERR_OR_NULL(master->dma_rx)) {
|
|
dma_release_channel(master->dma_rx);
|
|
master->dma_rx = NULL;
|
|
}
|
|
if (!IS_ERR_OR_NULL(master->dma_tx)) {
|
|
dma_release_channel(master->dma_tx);
|
|
master->dma_tx = NULL;
|
|
}
|
|
|
|
out_disable_clk:
|
|
clk_disable_unprepare(priv->clk);
|
|
|
|
out_master_put:
|
|
spi_master_put(master);
|
|
return ret;
|
|
}
|
|
|
|
static void uniphier_spi_remove(struct platform_device *pdev)
|
|
{
|
|
struct spi_master *master = platform_get_drvdata(pdev);
|
|
struct uniphier_spi_priv *priv = spi_master_get_devdata(master);
|
|
|
|
if (master->dma_tx)
|
|
dma_release_channel(master->dma_tx);
|
|
if (master->dma_rx)
|
|
dma_release_channel(master->dma_rx);
|
|
|
|
clk_disable_unprepare(priv->clk);
|
|
}
|
|
|
|
static const struct of_device_id uniphier_spi_match[] = {
|
|
{ .compatible = "socionext,uniphier-scssi" },
|
|
{ /* sentinel */ }
|
|
};
|
|
MODULE_DEVICE_TABLE(of, uniphier_spi_match);
|
|
|
|
static struct platform_driver uniphier_spi_driver = {
|
|
.probe = uniphier_spi_probe,
|
|
.remove_new = uniphier_spi_remove,
|
|
.driver = {
|
|
.name = "uniphier-spi",
|
|
.of_match_table = uniphier_spi_match,
|
|
},
|
|
};
|
|
module_platform_driver(uniphier_spi_driver);
|
|
|
|
MODULE_AUTHOR("Kunihiko Hayashi <hayashi.kunihiko@socionext.com>");
|
|
MODULE_AUTHOR("Keiji Hayashibara <hayashibara.keiji@socionext.com>");
|
|
MODULE_DESCRIPTION("Socionext UniPhier SPI controller driver");
|
|
MODULE_LICENSE("GPL v2");
|