forked from Minki/linux
48863ce594
DMA behavior is linked to descriptor management: -descriptor mechanism (Tx for example, but it is exactly the same for RX): -useful registers: -DMA_CH#_TxDesc_Ring_Len: length of transmit descriptor ring -DMA_CH#_TxDesc_List_Address: start address of the ring -DMA_CH#_TxDesc_Tail_Pointer: address of the last descriptor to send + 1. -DMA_CH#_TxDesc_Current_App_TxDesc: address of the current descriptor -The descriptor Tail Pointer register contains the pointer to the descriptor address (N). The base address and the current descriptor decide the address of the current descriptor that the DMA can process. The descriptors up to one location less than the one indicated by the descriptor tail pointer (N-1) are owned by the DMA. The DMA continues to process the descriptors until the following condition occurs: "current descriptor pointer == Descriptor Tail pointer" Then the DMA goes into suspend mode. The application must perform a write to descriptor tail pointer register and update the tail pointer to have the following condition and to start a new transfer: "current descriptor pointer < Descriptor tail pointer" The DMA automatically wraps around the base address when the end of ring is reached. Up to 8 DMA could be use but currently we only use one (channel0) Signed-off-by: Alexandre TORGUE <alexandre.torgue@st.com> Signed-off-by: Giuseppe Cavallaro <peppe.cavallaro@st.com> Signed-off-by: David S. Miller <davem@davemloft.net>
226 lines
5.7 KiB
C
226 lines
5.7 KiB
C
/*
|
|
* Copyright (C) 2007-2015 STMicroelectronics Ltd
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* Author: Alexandre Torgue <alexandre.torgue@st.com>
|
|
*/
|
|
|
|
#include <linux/io.h>
|
|
#include <linux/delay.h>
|
|
#include "common.h"
|
|
#include "dwmac4_dma.h"
|
|
#include "dwmac4.h"
|
|
|
|
int dwmac4_dma_reset(void __iomem *ioaddr)
|
|
{
|
|
u32 value = readl(ioaddr + DMA_BUS_MODE);
|
|
int limit;
|
|
|
|
/* DMA SW reset */
|
|
value |= DMA_BUS_MODE_SFT_RESET;
|
|
writel(value, ioaddr + DMA_BUS_MODE);
|
|
limit = 10;
|
|
while (limit--) {
|
|
if (!(readl(ioaddr + DMA_BUS_MODE) & DMA_BUS_MODE_SFT_RESET))
|
|
break;
|
|
mdelay(10);
|
|
}
|
|
|
|
if (limit < 0)
|
|
return -EBUSY;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void dwmac4_set_rx_tail_ptr(void __iomem *ioaddr, u32 tail_ptr, u32 chan)
|
|
{
|
|
writel(tail_ptr, ioaddr + DMA_CHAN_RX_END_ADDR(0));
|
|
}
|
|
|
|
void dwmac4_set_tx_tail_ptr(void __iomem *ioaddr, u32 tail_ptr, u32 chan)
|
|
{
|
|
writel(tail_ptr, ioaddr + DMA_CHAN_TX_END_ADDR(0));
|
|
}
|
|
|
|
void dwmac4_dma_start_tx(void __iomem *ioaddr)
|
|
{
|
|
u32 value = readl(ioaddr + DMA_CHAN_TX_CONTROL(STMMAC_CHAN0));
|
|
|
|
value |= DMA_CONTROL_ST;
|
|
writel(value, ioaddr + DMA_CHAN_TX_CONTROL(STMMAC_CHAN0));
|
|
|
|
value = readl(ioaddr + GMAC_CONFIG);
|
|
value |= GMAC_CONFIG_TE;
|
|
writel(value, ioaddr + GMAC_CONFIG);
|
|
}
|
|
|
|
void dwmac4_dma_stop_tx(void __iomem *ioaddr)
|
|
{
|
|
u32 value = readl(ioaddr + DMA_CHAN_TX_CONTROL(STMMAC_CHAN0));
|
|
|
|
value &= ~DMA_CONTROL_ST;
|
|
writel(value, ioaddr + DMA_CHAN_TX_CONTROL(STMMAC_CHAN0));
|
|
|
|
value = readl(ioaddr + GMAC_CONFIG);
|
|
value &= ~GMAC_CONFIG_TE;
|
|
writel(value, ioaddr + GMAC_CONFIG);
|
|
}
|
|
|
|
void dwmac4_dma_start_rx(void __iomem *ioaddr)
|
|
{
|
|
u32 value = readl(ioaddr + DMA_CHAN_RX_CONTROL(STMMAC_CHAN0));
|
|
|
|
value |= DMA_CONTROL_SR;
|
|
|
|
writel(value, ioaddr + DMA_CHAN_RX_CONTROL(STMMAC_CHAN0));
|
|
|
|
value = readl(ioaddr + GMAC_CONFIG);
|
|
value |= GMAC_CONFIG_RE;
|
|
writel(value, ioaddr + GMAC_CONFIG);
|
|
}
|
|
|
|
void dwmac4_dma_stop_rx(void __iomem *ioaddr)
|
|
{
|
|
u32 value = readl(ioaddr + DMA_CHAN_RX_CONTROL(STMMAC_CHAN0));
|
|
|
|
value &= ~DMA_CONTROL_SR;
|
|
writel(value, ioaddr + DMA_CHAN_RX_CONTROL(STMMAC_CHAN0));
|
|
|
|
value = readl(ioaddr + GMAC_CONFIG);
|
|
value &= ~GMAC_CONFIG_RE;
|
|
writel(value, ioaddr + GMAC_CONFIG);
|
|
}
|
|
|
|
void dwmac4_set_tx_ring_len(void __iomem *ioaddr, u32 len)
|
|
{
|
|
writel(len, ioaddr + DMA_CHAN_TX_RING_LEN(STMMAC_CHAN0));
|
|
}
|
|
|
|
void dwmac4_set_rx_ring_len(void __iomem *ioaddr, u32 len)
|
|
{
|
|
writel(len, ioaddr + DMA_CHAN_RX_RING_LEN(STMMAC_CHAN0));
|
|
}
|
|
|
|
void dwmac4_enable_dma_irq(void __iomem *ioaddr)
|
|
{
|
|
writel(DMA_CHAN_INTR_DEFAULT_MASK, ioaddr +
|
|
DMA_CHAN_INTR_ENA(STMMAC_CHAN0));
|
|
}
|
|
|
|
void dwmac410_enable_dma_irq(void __iomem *ioaddr)
|
|
{
|
|
writel(DMA_CHAN_INTR_DEFAULT_MASK_4_10,
|
|
ioaddr + DMA_CHAN_INTR_ENA(STMMAC_CHAN0));
|
|
}
|
|
|
|
void dwmac4_disable_dma_irq(void __iomem *ioaddr)
|
|
{
|
|
writel(0, ioaddr + DMA_CHAN_INTR_ENA(STMMAC_CHAN0));
|
|
}
|
|
|
|
int dwmac4_dma_interrupt(void __iomem *ioaddr,
|
|
struct stmmac_extra_stats *x)
|
|
{
|
|
int ret = 0;
|
|
|
|
u32 intr_status = readl(ioaddr + DMA_CHAN_STATUS(0));
|
|
|
|
/* ABNORMAL interrupts */
|
|
if (unlikely(intr_status & DMA_CHAN_STATUS_AIS)) {
|
|
if (unlikely(intr_status & DMA_CHAN_STATUS_RBU))
|
|
x->rx_buf_unav_irq++;
|
|
if (unlikely(intr_status & DMA_CHAN_STATUS_RPS))
|
|
x->rx_process_stopped_irq++;
|
|
if (unlikely(intr_status & DMA_CHAN_STATUS_RWT))
|
|
x->rx_watchdog_irq++;
|
|
if (unlikely(intr_status & DMA_CHAN_STATUS_ETI))
|
|
x->tx_early_irq++;
|
|
if (unlikely(intr_status & DMA_CHAN_STATUS_TPS)) {
|
|
x->tx_process_stopped_irq++;
|
|
ret = tx_hard_error;
|
|
}
|
|
if (unlikely(intr_status & DMA_CHAN_STATUS_FBE)) {
|
|
x->fatal_bus_error_irq++;
|
|
ret = tx_hard_error;
|
|
}
|
|
}
|
|
/* TX/RX NORMAL interrupts */
|
|
if (likely(intr_status & DMA_CHAN_STATUS_NIS)) {
|
|
x->normal_irq_n++;
|
|
if (likely(intr_status & DMA_CHAN_STATUS_RI)) {
|
|
u32 value;
|
|
|
|
value = readl(ioaddr + DMA_CHAN_INTR_ENA(STMMAC_CHAN0));
|
|
/* to schedule NAPI on real RIE event. */
|
|
if (likely(value & DMA_CHAN_INTR_ENA_RIE)) {
|
|
x->rx_normal_irq_n++;
|
|
ret |= handle_rx;
|
|
}
|
|
}
|
|
if (likely(intr_status & DMA_CHAN_STATUS_TI)) {
|
|
x->tx_normal_irq_n++;
|
|
ret |= handle_tx;
|
|
}
|
|
if (unlikely(intr_status & DMA_CHAN_STATUS_ERI))
|
|
x->rx_early_irq++;
|
|
}
|
|
|
|
/* Clear the interrupt by writing a logic 1 to the chanX interrupt
|
|
* status [21-0] expect reserved bits [5-3]
|
|
*/
|
|
writel((intr_status & 0x3fffc7),
|
|
ioaddr + DMA_CHAN_STATUS(STMMAC_CHAN0));
|
|
|
|
return ret;
|
|
}
|
|
|
|
void stmmac_dwmac4_set_mac_addr(void __iomem *ioaddr, u8 addr[6],
|
|
unsigned int high, unsigned int low)
|
|
{
|
|
unsigned long data;
|
|
|
|
data = (addr[5] << 8) | addr[4];
|
|
/* For MAC Addr registers se have to set the Address Enable (AE)
|
|
* bit that has no effect on the High Reg 0 where the bit 31 (MO)
|
|
* is RO.
|
|
*/
|
|
data |= (STMMAC_CHAN0 << GMAC_HI_DCS_SHIFT);
|
|
writel(data | GMAC_HI_REG_AE, ioaddr + high);
|
|
data = (addr[3] << 24) | (addr[2] << 16) | (addr[1] << 8) | addr[0];
|
|
writel(data, ioaddr + low);
|
|
}
|
|
|
|
/* Enable disable MAC RX/TX */
|
|
void stmmac_dwmac4_set_mac(void __iomem *ioaddr, bool enable)
|
|
{
|
|
u32 value = readl(ioaddr + GMAC_CONFIG);
|
|
|
|
if (enable)
|
|
value |= GMAC_CONFIG_RE | GMAC_CONFIG_TE;
|
|
else
|
|
value &= ~(GMAC_CONFIG_TE | GMAC_CONFIG_RE);
|
|
|
|
writel(value, ioaddr + GMAC_CONFIG);
|
|
}
|
|
|
|
void stmmac_dwmac4_get_mac_addr(void __iomem *ioaddr, unsigned char *addr,
|
|
unsigned int high, unsigned int low)
|
|
{
|
|
unsigned int hi_addr, lo_addr;
|
|
|
|
/* Read the MAC address from the hardware */
|
|
hi_addr = readl(ioaddr + high);
|
|
lo_addr = readl(ioaddr + low);
|
|
|
|
/* Extract the MAC address from the high and low words */
|
|
addr[0] = lo_addr & 0xff;
|
|
addr[1] = (lo_addr >> 8) & 0xff;
|
|
addr[2] = (lo_addr >> 16) & 0xff;
|
|
addr[3] = (lo_addr >> 24) & 0xff;
|
|
addr[4] = hi_addr & 0xff;
|
|
addr[5] = (hi_addr >> 8) & 0xff;
|
|
}
|