net: mvpp2: map the CPUs to threads
This patch maps all uses of the CPU to threads. All this_cpu calls are replaced, and all smp_processor_id() calls are wrapped into the indirection. Signed-off-by: Antoine Tenart <antoine.tenart@bootlin.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
543ec376e3
commit
074c74dfcb
@ -814,6 +814,9 @@ struct mvpp2_port {
|
||||
void __iomem *base;
|
||||
void __iomem *stats_base;
|
||||
|
||||
/* Number of threads used on the port */
|
||||
unsigned int nthreads;
|
||||
|
||||
struct mvpp2_rx_queue **rxqs;
|
||||
unsigned int nrxqs;
|
||||
struct mvpp2_tx_queue **txqs;
|
||||
@ -971,7 +974,7 @@ struct mvpp2_txq_pcpu_buf {
|
||||
|
||||
/* Per-CPU Tx queue control */
|
||||
struct mvpp2_txq_pcpu {
|
||||
unsigned int cpu;
|
||||
unsigned int thread;
|
||||
|
||||
/* Number of Tx DMA descriptors in the descriptor ring */
|
||||
int size;
|
||||
|
@ -1680,13 +1680,14 @@ static int mvpp2_txq_alloc_reserved_desc(struct mvpp2 *priv,
|
||||
/* Check if there are enough reserved descriptors for transmission.
|
||||
* If not, request chunk of reserved descriptors and check again.
|
||||
*/
|
||||
static int mvpp2_txq_reserved_desc_num_proc(struct mvpp2 *priv,
|
||||
static int mvpp2_txq_reserved_desc_num_proc(struct mvpp2_port *port,
|
||||
struct mvpp2_tx_queue *txq,
|
||||
struct mvpp2_txq_pcpu *txq_pcpu,
|
||||
int num)
|
||||
{
|
||||
struct mvpp2 *priv = port->priv;
|
||||
int req, desc_count;
|
||||
unsigned int cpu;
|
||||
unsigned int thread;
|
||||
|
||||
if (txq_pcpu->reserved_num >= num)
|
||||
return 0;
|
||||
@ -1697,10 +1698,10 @@ static int mvpp2_txq_reserved_desc_num_proc(struct mvpp2 *priv,
|
||||
|
||||
desc_count = 0;
|
||||
/* Compute total of used descriptors */
|
||||
for_each_present_cpu(cpu) {
|
||||
for (thread = 0; thread < port->nthreads; thread++) {
|
||||
struct mvpp2_txq_pcpu *txq_pcpu_aux;
|
||||
|
||||
txq_pcpu_aux = per_cpu_ptr(txq->pcpu, cpu);
|
||||
txq_pcpu_aux = per_cpu_ptr(txq->pcpu, thread);
|
||||
desc_count += txq_pcpu_aux->count;
|
||||
desc_count += txq_pcpu_aux->reserved_num;
|
||||
}
|
||||
@ -1709,7 +1710,7 @@ static int mvpp2_txq_reserved_desc_num_proc(struct mvpp2 *priv,
|
||||
desc_count += req;
|
||||
|
||||
if (desc_count >
|
||||
(txq->size - (num_present_cpus() * MVPP2_CPU_DESC_CHUNK)))
|
||||
(txq->size - (MVPP2_MAX_THREADS * MVPP2_CPU_DESC_CHUNK)))
|
||||
return -ENOMEM;
|
||||
|
||||
txq_pcpu->reserved_num += mvpp2_txq_alloc_reserved_desc(priv, txq, req);
|
||||
@ -1983,7 +1984,7 @@ static void mvpp2_txq_done(struct mvpp2_port *port, struct mvpp2_tx_queue *txq,
|
||||
struct netdev_queue *nq = netdev_get_tx_queue(port->dev, txq->log_id);
|
||||
int tx_done;
|
||||
|
||||
if (txq_pcpu->cpu != smp_processor_id())
|
||||
if (txq_pcpu->thread != mvpp2_cpu_to_thread(smp_processor_id()))
|
||||
netdev_err(port->dev, "wrong cpu on the end of Tx processing\n");
|
||||
|
||||
tx_done = mvpp2_txq_sent_desc_proc(port, txq);
|
||||
@ -2167,7 +2168,7 @@ static int mvpp2_txq_init(struct mvpp2_port *port,
|
||||
struct mvpp2_tx_queue *txq)
|
||||
{
|
||||
u32 val;
|
||||
unsigned int cpu, thread;
|
||||
unsigned int thread;
|
||||
int desc, desc_per_txq, tx_port_num;
|
||||
struct mvpp2_txq_pcpu *txq_pcpu;
|
||||
|
||||
@ -2224,8 +2225,8 @@ static int mvpp2_txq_init(struct mvpp2_port *port,
|
||||
mvpp2_write(port->priv, MVPP2_TXQ_SCHED_TOKEN_SIZE_REG(txq->log_id),
|
||||
val);
|
||||
|
||||
for_each_present_cpu(cpu) {
|
||||
txq_pcpu = per_cpu_ptr(txq->pcpu, cpu);
|
||||
for (thread = 0; thread < port->nthreads; thread++) {
|
||||
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
|
||||
txq_pcpu->size = txq->size;
|
||||
txq_pcpu->buffs = kmalloc_array(txq_pcpu->size,
|
||||
sizeof(*txq_pcpu->buffs),
|
||||
@ -2259,10 +2260,10 @@ static void mvpp2_txq_deinit(struct mvpp2_port *port,
|
||||
struct mvpp2_tx_queue *txq)
|
||||
{
|
||||
struct mvpp2_txq_pcpu *txq_pcpu;
|
||||
unsigned int cpu, thread;
|
||||
unsigned int thread;
|
||||
|
||||
for_each_present_cpu(cpu) {
|
||||
txq_pcpu = per_cpu_ptr(txq->pcpu, cpu);
|
||||
for (thread = 0; thread < port->nthreads; thread++) {
|
||||
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
|
||||
kfree(txq_pcpu->buffs);
|
||||
|
||||
if (txq_pcpu->tso_headers)
|
||||
@ -2300,7 +2301,7 @@ static void mvpp2_txq_clean(struct mvpp2_port *port, struct mvpp2_tx_queue *txq)
|
||||
{
|
||||
struct mvpp2_txq_pcpu *txq_pcpu;
|
||||
int delay, pending;
|
||||
unsigned int cpu, thread = mvpp2_cpu_to_thread(get_cpu());
|
||||
unsigned int thread = mvpp2_cpu_to_thread(get_cpu());
|
||||
u32 val;
|
||||
|
||||
mvpp2_percpu_write(port->priv, thread, MVPP2_TXQ_NUM_REG, txq->id);
|
||||
@ -2331,8 +2332,8 @@ static void mvpp2_txq_clean(struct mvpp2_port *port, struct mvpp2_tx_queue *txq)
|
||||
mvpp2_percpu_write(port->priv, thread, MVPP2_TXQ_PREF_BUF_REG, val);
|
||||
put_cpu();
|
||||
|
||||
for_each_present_cpu(cpu) {
|
||||
txq_pcpu = per_cpu_ptr(txq->pcpu, cpu);
|
||||
for (thread = 0; thread < port->nthreads; thread++) {
|
||||
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
|
||||
|
||||
/* Release all packets */
|
||||
mvpp2_txq_bufs_free(port, txq, txq_pcpu, txq_pcpu->count);
|
||||
@ -2513,16 +2514,20 @@ static void mvpp2_tx_proc_cb(unsigned long data)
|
||||
{
|
||||
struct net_device *dev = (struct net_device *)data;
|
||||
struct mvpp2_port *port = netdev_priv(dev);
|
||||
struct mvpp2_port_pcpu *port_pcpu = this_cpu_ptr(port->pcpu);
|
||||
struct mvpp2_port_pcpu *port_pcpu;
|
||||
unsigned int tx_todo, cause;
|
||||
|
||||
port_pcpu = per_cpu_ptr(port->pcpu,
|
||||
mvpp2_cpu_to_thread(smp_processor_id()));
|
||||
|
||||
if (!netif_running(dev))
|
||||
return;
|
||||
port_pcpu->timer_scheduled = false;
|
||||
|
||||
/* Process all the Tx queues */
|
||||
cause = (1 << port->ntxqs) - 1;
|
||||
tx_todo = mvpp2_tx_done(port, cause, smp_processor_id());
|
||||
tx_todo = mvpp2_tx_done(port, cause,
|
||||
mvpp2_cpu_to_thread(smp_processor_id()));
|
||||
|
||||
/* Set the timer in case not all the packets were processed */
|
||||
if (tx_todo)
|
||||
@ -2738,7 +2743,8 @@ static inline void
|
||||
tx_desc_unmap_put(struct mvpp2_port *port, struct mvpp2_tx_queue *txq,
|
||||
struct mvpp2_tx_desc *desc)
|
||||
{
|
||||
struct mvpp2_txq_pcpu *txq_pcpu = this_cpu_ptr(txq->pcpu);
|
||||
unsigned int thread = mvpp2_cpu_to_thread(smp_processor_id());
|
||||
struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
|
||||
|
||||
dma_addr_t buf_dma_addr =
|
||||
mvpp2_txdesc_dma_addr_get(port, desc);
|
||||
@ -2755,7 +2761,8 @@ static int mvpp2_tx_frag_process(struct mvpp2_port *port, struct sk_buff *skb,
|
||||
struct mvpp2_tx_queue *aggr_txq,
|
||||
struct mvpp2_tx_queue *txq)
|
||||
{
|
||||
struct mvpp2_txq_pcpu *txq_pcpu = this_cpu_ptr(txq->pcpu);
|
||||
unsigned int thread = mvpp2_cpu_to_thread(smp_processor_id());
|
||||
struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
|
||||
struct mvpp2_tx_desc *tx_desc;
|
||||
int i;
|
||||
dma_addr_t buf_dma_addr;
|
||||
@ -2876,7 +2883,7 @@ static int mvpp2_tx_tso(struct sk_buff *skb, struct net_device *dev,
|
||||
/* Check number of available descriptors */
|
||||
if (mvpp2_aggr_desc_num_check(port->priv, aggr_txq,
|
||||
tso_count_descs(skb)) ||
|
||||
mvpp2_txq_reserved_desc_num_proc(port->priv, txq, txq_pcpu,
|
||||
mvpp2_txq_reserved_desc_num_proc(port, txq, txq_pcpu,
|
||||
tso_count_descs(skb)))
|
||||
return 0;
|
||||
|
||||
@ -2923,14 +2930,17 @@ static int mvpp2_tx(struct sk_buff *skb, struct net_device *dev)
|
||||
struct mvpp2_txq_pcpu *txq_pcpu;
|
||||
struct mvpp2_tx_desc *tx_desc;
|
||||
dma_addr_t buf_dma_addr;
|
||||
unsigned int thread;
|
||||
int frags = 0;
|
||||
u16 txq_id;
|
||||
u32 tx_cmd;
|
||||
|
||||
thread = mvpp2_cpu_to_thread(smp_processor_id());
|
||||
|
||||
txq_id = skb_get_queue_mapping(skb);
|
||||
txq = port->txqs[txq_id];
|
||||
txq_pcpu = this_cpu_ptr(txq->pcpu);
|
||||
aggr_txq = &port->priv->aggr_txqs[smp_processor_id()];
|
||||
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
|
||||
aggr_txq = &port->priv->aggr_txqs[thread];
|
||||
|
||||
if (skb_is_gso(skb)) {
|
||||
frags = mvpp2_tx_tso(skb, dev, txq, aggr_txq, txq_pcpu);
|
||||
@ -2940,8 +2950,7 @@ static int mvpp2_tx(struct sk_buff *skb, struct net_device *dev)
|
||||
|
||||
/* Check number of available descriptors */
|
||||
if (mvpp2_aggr_desc_num_check(port->priv, aggr_txq, frags) ||
|
||||
mvpp2_txq_reserved_desc_num_proc(port->priv, txq,
|
||||
txq_pcpu, frags)) {
|
||||
mvpp2_txq_reserved_desc_num_proc(port, txq, txq_pcpu, frags)) {
|
||||
frags = 0;
|
||||
goto out;
|
||||
}
|
||||
@ -2983,7 +2992,7 @@ static int mvpp2_tx(struct sk_buff *skb, struct net_device *dev)
|
||||
|
||||
out:
|
||||
if (frags > 0) {
|
||||
struct mvpp2_pcpu_stats *stats = this_cpu_ptr(port->stats);
|
||||
struct mvpp2_pcpu_stats *stats = per_cpu_ptr(port->stats, thread);
|
||||
struct netdev_queue *nq = netdev_get_tx_queue(dev, txq_id);
|
||||
|
||||
txq_pcpu->reserved_num -= frags;
|
||||
@ -3013,7 +3022,7 @@ out:
|
||||
/* Set the timer in case not all frags were processed */
|
||||
if (!port->has_tx_irqs && txq_pcpu->count <= frags &&
|
||||
txq_pcpu->count > 0) {
|
||||
struct mvpp2_port_pcpu *port_pcpu = this_cpu_ptr(port->pcpu);
|
||||
struct mvpp2_port_pcpu *port_pcpu = per_cpu_ptr(port->pcpu, thread);
|
||||
|
||||
mvpp2_timer_set(port_pcpu);
|
||||
}
|
||||
@ -3404,7 +3413,7 @@ static int mvpp2_stop(struct net_device *dev)
|
||||
{
|
||||
struct mvpp2_port *port = netdev_priv(dev);
|
||||
struct mvpp2_port_pcpu *port_pcpu;
|
||||
unsigned int cpu;
|
||||
unsigned int thread;
|
||||
|
||||
mvpp2_stop_dev(port);
|
||||
|
||||
@ -3419,8 +3428,8 @@ static int mvpp2_stop(struct net_device *dev)
|
||||
|
||||
mvpp2_irqs_deinit(port);
|
||||
if (!port->has_tx_irqs) {
|
||||
for_each_present_cpu(cpu) {
|
||||
port_pcpu = per_cpu_ptr(port->pcpu, cpu);
|
||||
for (thread = 0; thread < port->nthreads; thread++) {
|
||||
port_pcpu = per_cpu_ptr(port->pcpu, thread);
|
||||
|
||||
hrtimer_cancel(&port_pcpu->tx_done_timer);
|
||||
port_pcpu->timer_scheduled = false;
|
||||
@ -4101,7 +4110,7 @@ static int mvpp2_port_init(struct mvpp2_port *port)
|
||||
struct device *dev = port->dev->dev.parent;
|
||||
struct mvpp2 *priv = port->priv;
|
||||
struct mvpp2_txq_pcpu *txq_pcpu;
|
||||
unsigned int cpu;
|
||||
unsigned int thread;
|
||||
int queue, err;
|
||||
|
||||
/* Checks for hardware constraints */
|
||||
@ -4146,9 +4155,9 @@ static int mvpp2_port_init(struct mvpp2_port *port)
|
||||
txq->id = queue_phy_id;
|
||||
txq->log_id = queue;
|
||||
txq->done_pkts_coal = MVPP2_TXDONE_COAL_PKTS_THRESH;
|
||||
for_each_present_cpu(cpu) {
|
||||
txq_pcpu = per_cpu_ptr(txq->pcpu, cpu);
|
||||
txq_pcpu->cpu = cpu;
|
||||
for (thread = 0; thread < port->nthreads; thread++) {
|
||||
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
|
||||
txq_pcpu->thread = thread;
|
||||
}
|
||||
|
||||
port->txqs[queue] = txq;
|
||||
@ -4642,7 +4651,7 @@ static int mvpp2_port_probe(struct platform_device *pdev,
|
||||
struct resource *res;
|
||||
struct phylink *phylink;
|
||||
char *mac_from = "";
|
||||
unsigned int ntxqs, nrxqs, cpu;
|
||||
unsigned int ntxqs, nrxqs, thread;
|
||||
unsigned long flags = 0;
|
||||
bool has_tx_irqs;
|
||||
u32 id;
|
||||
@ -4706,6 +4715,9 @@ static int mvpp2_port_probe(struct platform_device *pdev,
|
||||
port->has_tx_irqs = has_tx_irqs;
|
||||
port->flags = flags;
|
||||
|
||||
port->nthreads = min_t(unsigned int, num_present_cpus(),
|
||||
MVPP2_MAX_THREADS);
|
||||
|
||||
err = mvpp2_queue_vectors_init(port, port_node);
|
||||
if (err)
|
||||
goto err_free_netdev;
|
||||
@ -4801,8 +4813,8 @@ static int mvpp2_port_probe(struct platform_device *pdev,
|
||||
}
|
||||
|
||||
if (!port->has_tx_irqs) {
|
||||
for_each_present_cpu(cpu) {
|
||||
port_pcpu = per_cpu_ptr(port->pcpu, cpu);
|
||||
for (thread = 0; thread < port->nthreads; thread++) {
|
||||
port_pcpu = per_cpu_ptr(port->pcpu, thread);
|
||||
|
||||
hrtimer_init(&port_pcpu->tx_done_timer, CLOCK_MONOTONIC,
|
||||
HRTIMER_MODE_REL_PINNED);
|
||||
@ -5086,13 +5098,13 @@ static int mvpp2_init(struct platform_device *pdev, struct mvpp2 *priv)
|
||||
}
|
||||
|
||||
/* Allocate and initialize aggregated TXQs */
|
||||
priv->aggr_txqs = devm_kcalloc(&pdev->dev, num_present_cpus(),
|
||||
priv->aggr_txqs = devm_kcalloc(&pdev->dev, MVPP2_MAX_THREADS,
|
||||
sizeof(*priv->aggr_txqs),
|
||||
GFP_KERNEL);
|
||||
if (!priv->aggr_txqs)
|
||||
return -ENOMEM;
|
||||
|
||||
for_each_present_cpu(i) {
|
||||
for (i = 0; i < MVPP2_MAX_THREADS; i++) {
|
||||
priv->aggr_txqs[i].id = i;
|
||||
priv->aggr_txqs[i].size = MVPP2_AGGR_TXQ_SIZE;
|
||||
err = mvpp2_aggr_txq_init(pdev, &priv->aggr_txqs[i], i, priv);
|
||||
@ -5378,7 +5390,7 @@ static int mvpp2_remove(struct platform_device *pdev)
|
||||
mvpp2_bm_pool_destroy(pdev, priv, bm_pool);
|
||||
}
|
||||
|
||||
for_each_present_cpu(i) {
|
||||
for (i = 0; i < MVPP2_MAX_THREADS; i++) {
|
||||
struct mvpp2_tx_queue *aggr_txq = &priv->aggr_txqs[i];
|
||||
|
||||
dma_free_coherent(&pdev->dev,
|
||||
|
Loading…
Reference in New Issue
Block a user