mirror of
https://github.com/torvalds/linux.git
synced 2024-11-26 14:12:06 +00:00
18836b48eb
The switch to the generic dma ops made dma masks mandatory, breaking
devices having them not set. In case of bcm63xx, it broke ethernet with
the following warning when trying to up the device:
[ 2.633123] ------------[ cut here ]------------
[ 2.637949] WARNING: CPU: 0 PID: 325 at ./include/linux/dma-mapping.h:516 bcm_enetsw_open+0x160/0xbbc
[ 2.647423] Modules linked in: gpio_button_hotplug
[ 2.652361] CPU: 0 PID: 325 Comm: ip Not tainted 4.19.16 #0
[ 2.658080] Stack : 80520000 804cd3ec 00000000 00000000 804ccc00 87085bdc 87d3f9d4 804f9a17
[ 2.666707] 8049cf18 00000145 80a942a0 00000204 80ac0000 10008400 87085b90 eb3d5ab7
[ 2.675325] 00000000 00000000 80ac0000 000022b0 00000000 00000000 00000007 00000000
[ 2.683954] 0000007a 80500000 0013b381 00000000 80000000 00000000 804a1664 80289878
[ 2.692572] 00000009 00000204 80ac0000 00000200 00000002 00000000 00000000 80a90000
[ 2.701191] ...
[ 2.703701] Call Trace:
[ 2.706244] [<8001f3c8>] show_stack+0x58/0x100
[ 2.710840] [<800336e4>] __warn+0xe4/0x118
[ 2.715049] [<800337d4>] warn_slowpath_null+0x48/0x64
[ 2.720237] [<80289878>] bcm_enetsw_open+0x160/0xbbc
[ 2.725347] [<802d1d4c>] __dev_open+0xf8/0x16c
[ 2.729913] [<802d20cc>] __dev_change_flags+0x100/0x1c4
[ 2.735290] [<802d21b8>] dev_change_flags+0x28/0x70
[ 2.740326] [<803539e0>] devinet_ioctl+0x310/0x7b0
[ 2.745250] [<80355fd8>] inet_ioctl+0x1f8/0x224
[ 2.749939] [<802af290>] sock_ioctl+0x30c/0x488
[ 2.754632] [<80112b34>] do_vfs_ioctl+0x740/0x7dc
[ 2.759459] [<80112c20>] ksys_ioctl+0x50/0x94
[ 2.763955] [<800240b8>] syscall_common+0x34/0x58
[ 2.768782] ---[ end trace fb1a6b14d74e28b6 ]---
[ 2.773544] bcm63xx_enetsw bcm63xx_enetsw.0: cannot allocate rx ring 512
Fix this by adding appropriate DMA masks for the platform devices.
Fixes: f8c55dc6e8
("MIPS: use generic dma noncoherent ops for simple noncoherent platforms")
Signed-off-by: Jonas Gorski <jonas.gorski@gmail.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Florian Fainelli <f.fainelli@gmail.com>
Signed-off-by: Paul Burton <paul.burton@mips.com>
Cc: linux-mips@linux-mips.org
Cc: linux-kernel@vger.kernel.org
Cc: Ralf Baechle <ralf@linux-mips.org>
Cc: James Hogan <jhogan@kernel.org>
Cc: stable@vger.kernel.org # v4.19+
328 lines
8.1 KiB
C
328 lines
8.1 KiB
C
/*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* Copyright (C) 2008 Maxime Bizon <mbizon@freebox.fr>
|
|
*/
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/export.h>
|
|
#include <bcm63xx_dev_enet.h>
|
|
#include <bcm63xx_io.h>
|
|
#include <bcm63xx_regs.h>
|
|
|
|
static const unsigned long bcm6348_regs_enetdmac[] = {
|
|
[ENETDMAC_CHANCFG] = ENETDMAC_CHANCFG_REG,
|
|
[ENETDMAC_IR] = ENETDMAC_IR_REG,
|
|
[ENETDMAC_IRMASK] = ENETDMAC_IRMASK_REG,
|
|
[ENETDMAC_MAXBURST] = ENETDMAC_MAXBURST_REG,
|
|
};
|
|
|
|
static const unsigned long bcm6345_regs_enetdmac[] = {
|
|
[ENETDMAC_CHANCFG] = ENETDMA_6345_CHANCFG_REG,
|
|
[ENETDMAC_IR] = ENETDMA_6345_IR_REG,
|
|
[ENETDMAC_IRMASK] = ENETDMA_6345_IRMASK_REG,
|
|
[ENETDMAC_MAXBURST] = ENETDMA_6345_MAXBURST_REG,
|
|
[ENETDMAC_BUFALLOC] = ENETDMA_6345_BUFALLOC_REG,
|
|
[ENETDMAC_RSTART] = ENETDMA_6345_RSTART_REG,
|
|
[ENETDMAC_FC] = ENETDMA_6345_FC_REG,
|
|
[ENETDMAC_LEN] = ENETDMA_6345_LEN_REG,
|
|
};
|
|
|
|
const unsigned long *bcm63xx_regs_enetdmac;
|
|
EXPORT_SYMBOL(bcm63xx_regs_enetdmac);
|
|
|
|
static __init void bcm63xx_enetdmac_regs_init(void)
|
|
{
|
|
if (BCMCPU_IS_6345())
|
|
bcm63xx_regs_enetdmac = bcm6345_regs_enetdmac;
|
|
else
|
|
bcm63xx_regs_enetdmac = bcm6348_regs_enetdmac;
|
|
}
|
|
|
|
static struct resource shared_res[] = {
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.end = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_MEM,
|
|
},
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.end = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_MEM,
|
|
},
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.end = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_MEM,
|
|
},
|
|
};
|
|
|
|
static struct platform_device bcm63xx_enet_shared_device = {
|
|
.name = "bcm63xx_enet_shared",
|
|
.id = 0,
|
|
.num_resources = ARRAY_SIZE(shared_res),
|
|
.resource = shared_res,
|
|
};
|
|
|
|
static int shared_device_registered;
|
|
|
|
static u64 enet_dmamask = DMA_BIT_MASK(32);
|
|
|
|
static struct resource enet0_res[] = {
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.end = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_MEM,
|
|
},
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_IRQ,
|
|
},
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_IRQ,
|
|
},
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_IRQ,
|
|
},
|
|
};
|
|
|
|
static struct bcm63xx_enet_platform_data enet0_pd;
|
|
|
|
static struct platform_device bcm63xx_enet0_device = {
|
|
.name = "bcm63xx_enet",
|
|
.id = 0,
|
|
.num_resources = ARRAY_SIZE(enet0_res),
|
|
.resource = enet0_res,
|
|
.dev = {
|
|
.platform_data = &enet0_pd,
|
|
.dma_mask = &enet_dmamask,
|
|
.coherent_dma_mask = DMA_BIT_MASK(32),
|
|
},
|
|
};
|
|
|
|
static struct resource enet1_res[] = {
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.end = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_MEM,
|
|
},
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_IRQ,
|
|
},
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_IRQ,
|
|
},
|
|
{
|
|
.start = -1, /* filled at runtime */
|
|
.flags = IORESOURCE_IRQ,
|
|
},
|
|
};
|
|
|
|
static struct bcm63xx_enet_platform_data enet1_pd;
|
|
|
|
static struct platform_device bcm63xx_enet1_device = {
|
|
.name = "bcm63xx_enet",
|
|
.id = 1,
|
|
.num_resources = ARRAY_SIZE(enet1_res),
|
|
.resource = enet1_res,
|
|
.dev = {
|
|
.platform_data = &enet1_pd,
|
|
.dma_mask = &enet_dmamask,
|
|
.coherent_dma_mask = DMA_BIT_MASK(32),
|
|
},
|
|
};
|
|
|
|
static struct resource enetsw_res[] = {
|
|
{
|
|
/* start & end filled at runtime */
|
|
.flags = IORESOURCE_MEM,
|
|
},
|
|
{
|
|
/* start filled at runtime */
|
|
.flags = IORESOURCE_IRQ,
|
|
},
|
|
{
|
|
/* start filled at runtime */
|
|
.flags = IORESOURCE_IRQ,
|
|
},
|
|
};
|
|
|
|
static struct bcm63xx_enetsw_platform_data enetsw_pd;
|
|
|
|
static struct platform_device bcm63xx_enetsw_device = {
|
|
.name = "bcm63xx_enetsw",
|
|
.num_resources = ARRAY_SIZE(enetsw_res),
|
|
.resource = enetsw_res,
|
|
.dev = {
|
|
.platform_data = &enetsw_pd,
|
|
.dma_mask = &enet_dmamask,
|
|
.coherent_dma_mask = DMA_BIT_MASK(32),
|
|
},
|
|
};
|
|
|
|
static int __init register_shared(void)
|
|
{
|
|
int ret, chan_count;
|
|
|
|
if (shared_device_registered)
|
|
return 0;
|
|
|
|
bcm63xx_enetdmac_regs_init();
|
|
|
|
shared_res[0].start = bcm63xx_regset_address(RSET_ENETDMA);
|
|
shared_res[0].end = shared_res[0].start;
|
|
if (BCMCPU_IS_6345())
|
|
shared_res[0].end += (RSET_6345_ENETDMA_SIZE) - 1;
|
|
else
|
|
shared_res[0].end += (RSET_ENETDMA_SIZE) - 1;
|
|
|
|
if (BCMCPU_IS_6328() || BCMCPU_IS_6362() || BCMCPU_IS_6368())
|
|
chan_count = 32;
|
|
else if (BCMCPU_IS_6345())
|
|
chan_count = 8;
|
|
else
|
|
chan_count = 16;
|
|
|
|
shared_res[1].start = bcm63xx_regset_address(RSET_ENETDMAC);
|
|
shared_res[1].end = shared_res[1].start;
|
|
shared_res[1].end += RSET_ENETDMAC_SIZE(chan_count) - 1;
|
|
|
|
shared_res[2].start = bcm63xx_regset_address(RSET_ENETDMAS);
|
|
shared_res[2].end = shared_res[2].start;
|
|
shared_res[2].end += RSET_ENETDMAS_SIZE(chan_count) - 1;
|
|
|
|
ret = platform_device_register(&bcm63xx_enet_shared_device);
|
|
if (ret)
|
|
return ret;
|
|
shared_device_registered = 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int __init bcm63xx_enet_register(int unit,
|
|
const struct bcm63xx_enet_platform_data *pd)
|
|
{
|
|
struct platform_device *pdev;
|
|
struct bcm63xx_enet_platform_data *dpd;
|
|
int ret;
|
|
|
|
if (unit > 1)
|
|
return -ENODEV;
|
|
|
|
if (unit == 1 && (BCMCPU_IS_6338() || BCMCPU_IS_6345()))
|
|
return -ENODEV;
|
|
|
|
ret = register_shared();
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (unit == 0) {
|
|
enet0_res[0].start = bcm63xx_regset_address(RSET_ENET0);
|
|
enet0_res[0].end = enet0_res[0].start;
|
|
enet0_res[0].end += RSET_ENET_SIZE - 1;
|
|
enet0_res[1].start = bcm63xx_get_irq_number(IRQ_ENET0);
|
|
enet0_res[2].start = bcm63xx_get_irq_number(IRQ_ENET0_RXDMA);
|
|
enet0_res[3].start = bcm63xx_get_irq_number(IRQ_ENET0_TXDMA);
|
|
pdev = &bcm63xx_enet0_device;
|
|
} else {
|
|
enet1_res[0].start = bcm63xx_regset_address(RSET_ENET1);
|
|
enet1_res[0].end = enet1_res[0].start;
|
|
enet1_res[0].end += RSET_ENET_SIZE - 1;
|
|
enet1_res[1].start = bcm63xx_get_irq_number(IRQ_ENET1);
|
|
enet1_res[2].start = bcm63xx_get_irq_number(IRQ_ENET1_RXDMA);
|
|
enet1_res[3].start = bcm63xx_get_irq_number(IRQ_ENET1_TXDMA);
|
|
pdev = &bcm63xx_enet1_device;
|
|
}
|
|
|
|
/* copy given platform data */
|
|
dpd = pdev->dev.platform_data;
|
|
memcpy(dpd, pd, sizeof(*pd));
|
|
|
|
/* adjust them in case internal phy is used */
|
|
if (dpd->use_internal_phy) {
|
|
|
|
/* internal phy only exists for enet0 */
|
|
if (unit == 1)
|
|
return -ENODEV;
|
|
|
|
dpd->phy_id = 1;
|
|
dpd->has_phy_interrupt = 1;
|
|
dpd->phy_interrupt = bcm63xx_get_irq_number(IRQ_ENET_PHY);
|
|
}
|
|
|
|
dpd->dma_chan_en_mask = ENETDMAC_CHANCFG_EN_MASK;
|
|
dpd->dma_chan_int_mask = ENETDMAC_IR_PKTDONE_MASK;
|
|
if (BCMCPU_IS_6345()) {
|
|
dpd->dma_chan_en_mask |= ENETDMAC_CHANCFG_CHAINING_MASK;
|
|
dpd->dma_chan_en_mask |= ENETDMAC_CHANCFG_WRAP_EN_MASK;
|
|
dpd->dma_chan_en_mask |= ENETDMAC_CHANCFG_FLOWC_EN_MASK;
|
|
dpd->dma_chan_int_mask |= ENETDMA_IR_BUFDONE_MASK;
|
|
dpd->dma_chan_int_mask |= ENETDMA_IR_NOTOWNER_MASK;
|
|
dpd->dma_chan_width = ENETDMA_6345_CHAN_WIDTH;
|
|
dpd->dma_desc_shift = ENETDMA_6345_DESC_SHIFT;
|
|
} else {
|
|
dpd->dma_has_sram = true;
|
|
dpd->dma_chan_width = ENETDMA_CHAN_WIDTH;
|
|
}
|
|
|
|
if (unit == 0) {
|
|
dpd->rx_chan = 0;
|
|
dpd->tx_chan = 1;
|
|
} else {
|
|
dpd->rx_chan = 2;
|
|
dpd->tx_chan = 3;
|
|
}
|
|
|
|
ret = platform_device_register(pdev);
|
|
if (ret)
|
|
return ret;
|
|
return 0;
|
|
}
|
|
|
|
int __init
|
|
bcm63xx_enetsw_register(const struct bcm63xx_enetsw_platform_data *pd)
|
|
{
|
|
int ret;
|
|
|
|
if (!BCMCPU_IS_6328() && !BCMCPU_IS_6362() && !BCMCPU_IS_6368())
|
|
return -ENODEV;
|
|
|
|
ret = register_shared();
|
|
if (ret)
|
|
return ret;
|
|
|
|
enetsw_res[0].start = bcm63xx_regset_address(RSET_ENETSW);
|
|
enetsw_res[0].end = enetsw_res[0].start;
|
|
enetsw_res[0].end += RSET_ENETSW_SIZE - 1;
|
|
enetsw_res[1].start = bcm63xx_get_irq_number(IRQ_ENETSW_RXDMA0);
|
|
enetsw_res[2].start = bcm63xx_get_irq_number(IRQ_ENETSW_TXDMA0);
|
|
if (!enetsw_res[2].start)
|
|
enetsw_res[2].start = -1;
|
|
|
|
memcpy(bcm63xx_enetsw_device.dev.platform_data, pd, sizeof(*pd));
|
|
|
|
if (BCMCPU_IS_6328())
|
|
enetsw_pd.num_ports = ENETSW_PORTS_6328;
|
|
else if (BCMCPU_IS_6362() || BCMCPU_IS_6368())
|
|
enetsw_pd.num_ports = ENETSW_PORTS_6368;
|
|
|
|
enetsw_pd.dma_has_sram = true;
|
|
enetsw_pd.dma_chan_width = ENETDMA_CHAN_WIDTH;
|
|
enetsw_pd.dma_chan_en_mask = ENETDMAC_CHANCFG_EN_MASK;
|
|
enetsw_pd.dma_chan_int_mask = ENETDMAC_IR_PKTDONE_MASK;
|
|
|
|
ret = platform_device_register(&bcm63xx_enetsw_device);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return 0;
|
|
}
|