mirror of
https://github.com/torvalds/linux.git
synced 2024-12-13 06:32:50 +00:00
e9b1a4f867
After modifying the QP to the Error state, all RX WR would be completed
with WC in IB_WC_WR_FLUSH_ERR status. Current implementation does not
wait for it is done, but destroy the QP and free the link group directly.
So there is a risk that accessing the freed memory in tasklet context.
Here is a crash example:
BUG: unable to handle page fault for address: ffffffff8f220860
#PF: supervisor write access in kernel mode
#PF: error_code(0x0002) - not-present page
PGD f7300e067 P4D f7300e067 PUD f7300f063 PMD 8c4e45063 PTE 800ffff08c9df060
Oops: 0002 [#1] SMP PTI
CPU: 1 PID: 0 Comm: swapper/1 Kdump: loaded Tainted: G S OE 5.10.0-0607+ #23
Hardware name: Inspur NF5280M4/YZMB-00689-101, BIOS 4.1.20 07/09/2018
RIP: 0010:native_queued_spin_lock_slowpath+0x176/0x1b0
Code: f3 90 48 8b 32 48 85 f6 74 f6 eb d5 c1 ee 12 83 e0 03 83 ee 01 48 c1 e0 05 48 63 f6 48 05 00 c8 02 00 48 03 04 f5 00 09 98 8e <48> 89 10 8b 42 08 85 c0 75 09 f3 90 8b 42 08 85 c0 74 f7 48 8b 32
RSP: 0018:ffffb3b6c001ebd8 EFLAGS: 00010086
RAX: ffffffff8f220860 RBX: 0000000000000246 RCX: 0000000000080000
RDX: ffff91db1f86c800 RSI: 000000000000173c RDI: ffff91db62bace00
RBP: ffff91db62bacc00 R08: 0000000000000000 R09: c00000010000028b
R10: 0000000000055198 R11: ffffb3b6c001ea58 R12: ffff91db80e05010
R13: 000000000000000a R14: 0000000000000006 R15: 0000000000000040
FS: 0000000000000000(0000) GS:ffff91db1f840000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: ffffffff8f220860 CR3: 00000001f9580004 CR4: 00000000003706e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
Call Trace:
<IRQ>
_raw_spin_lock_irqsave+0x30/0x40
mlx5_ib_poll_cq+0x4c/0xc50 [mlx5_ib]
smc_wr_rx_tasklet_fn+0x56/0xa0 [smc]
tasklet_action_common.isra.21+0x66/0x100
__do_softirq+0xd5/0x29c
asm_call_irq_on_stack+0x12/0x20
</IRQ>
do_softirq_own_stack+0x37/0x40
irq_exit_rcu+0x9d/0xa0
sysvec_call_function_single+0x34/0x80
asm_sysvec_call_function_single+0x12/0x20
Fixes: bd4ad57718
("smc: initialize IB transport incl. PD, MR, QP, CQ, event, WR")
Signed-off-by: Yacan Liu <liuyacan@corp.netease.com>
Reviewed-by: Tony Lu <tonylu@linux.alibaba.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
141 lines
4.0 KiB
C
141 lines
4.0 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Shared Memory Communications over RDMA (SMC-R) and RoCE
|
|
*
|
|
* Work Requests exploiting Infiniband API
|
|
*
|
|
* Copyright IBM Corp. 2016
|
|
*
|
|
* Author(s): Steffen Maier <maier@linux.vnet.ibm.com>
|
|
*/
|
|
|
|
#ifndef SMC_WR_H
|
|
#define SMC_WR_H
|
|
|
|
#include <linux/atomic.h>
|
|
#include <rdma/ib_verbs.h>
|
|
#include <asm/div64.h>
|
|
|
|
#include "smc.h"
|
|
#include "smc_core.h"
|
|
|
|
#define SMC_WR_BUF_CNT 16 /* # of ctrl buffers per link */
|
|
|
|
#define SMC_WR_TX_WAIT_FREE_SLOT_TIME (10 * HZ)
|
|
|
|
#define SMC_WR_TX_SIZE 44 /* actual size of wr_send data (<=SMC_WR_BUF_SIZE) */
|
|
|
|
#define SMC_WR_TX_PEND_PRIV_SIZE 32
|
|
|
|
struct smc_wr_tx_pend_priv {
|
|
u8 priv[SMC_WR_TX_PEND_PRIV_SIZE];
|
|
};
|
|
|
|
typedef void (*smc_wr_tx_handler)(struct smc_wr_tx_pend_priv *,
|
|
struct smc_link *,
|
|
enum ib_wc_status);
|
|
|
|
typedef bool (*smc_wr_tx_filter)(struct smc_wr_tx_pend_priv *,
|
|
unsigned long);
|
|
|
|
typedef void (*smc_wr_tx_dismisser)(struct smc_wr_tx_pend_priv *);
|
|
|
|
struct smc_wr_rx_handler {
|
|
struct hlist_node list; /* hash table collision resolution */
|
|
void (*handler)(struct ib_wc *, void *);
|
|
u8 type;
|
|
};
|
|
|
|
/* Only used by RDMA write WRs.
|
|
* All other WRs (CDC/LLC) use smc_wr_tx_send handling WR_ID implicitly
|
|
*/
|
|
static inline long smc_wr_tx_get_next_wr_id(struct smc_link *link)
|
|
{
|
|
return atomic_long_inc_return(&link->wr_tx_id);
|
|
}
|
|
|
|
static inline void smc_wr_tx_set_wr_id(atomic_long_t *wr_tx_id, long val)
|
|
{
|
|
atomic_long_set(wr_tx_id, val);
|
|
}
|
|
|
|
static inline bool smc_wr_tx_link_hold(struct smc_link *link)
|
|
{
|
|
if (!smc_link_sendable(link))
|
|
return false;
|
|
atomic_inc(&link->wr_tx_refcnt);
|
|
return true;
|
|
}
|
|
|
|
static inline void smc_wr_tx_link_put(struct smc_link *link)
|
|
{
|
|
if (atomic_dec_and_test(&link->wr_tx_refcnt))
|
|
wake_up_all(&link->wr_tx_wait);
|
|
}
|
|
|
|
static inline void smc_wr_drain_cq(struct smc_link *lnk)
|
|
{
|
|
wait_event(lnk->wr_rx_empty_wait, lnk->wr_rx_id_compl == lnk->wr_rx_id);
|
|
}
|
|
|
|
static inline void smc_wr_wakeup_tx_wait(struct smc_link *lnk)
|
|
{
|
|
wake_up_all(&lnk->wr_tx_wait);
|
|
}
|
|
|
|
static inline void smc_wr_wakeup_reg_wait(struct smc_link *lnk)
|
|
{
|
|
wake_up(&lnk->wr_reg_wait);
|
|
}
|
|
|
|
/* post a new receive work request to fill a completed old work request entry */
|
|
static inline int smc_wr_rx_post(struct smc_link *link)
|
|
{
|
|
int rc;
|
|
u64 wr_id, temp_wr_id;
|
|
u32 index;
|
|
|
|
wr_id = ++link->wr_rx_id; /* tasklet context, thus not atomic */
|
|
temp_wr_id = wr_id;
|
|
index = do_div(temp_wr_id, link->wr_rx_cnt);
|
|
link->wr_rx_ibs[index].wr_id = wr_id;
|
|
rc = ib_post_recv(link->roce_qp, &link->wr_rx_ibs[index], NULL);
|
|
return rc;
|
|
}
|
|
|
|
int smc_wr_create_link(struct smc_link *lnk);
|
|
int smc_wr_alloc_link_mem(struct smc_link *lnk);
|
|
int smc_wr_alloc_lgr_mem(struct smc_link_group *lgr);
|
|
void smc_wr_free_link(struct smc_link *lnk);
|
|
void smc_wr_free_link_mem(struct smc_link *lnk);
|
|
void smc_wr_free_lgr_mem(struct smc_link_group *lgr);
|
|
void smc_wr_remember_qp_attr(struct smc_link *lnk);
|
|
void smc_wr_remove_dev(struct smc_ib_device *smcibdev);
|
|
void smc_wr_add_dev(struct smc_ib_device *smcibdev);
|
|
|
|
int smc_wr_tx_get_free_slot(struct smc_link *link, smc_wr_tx_handler handler,
|
|
struct smc_wr_buf **wr_buf,
|
|
struct smc_rdma_wr **wrs,
|
|
struct smc_wr_tx_pend_priv **wr_pend_priv);
|
|
int smc_wr_tx_get_v2_slot(struct smc_link *link,
|
|
smc_wr_tx_handler handler,
|
|
struct smc_wr_v2_buf **wr_buf,
|
|
struct smc_wr_tx_pend_priv **wr_pend_priv);
|
|
int smc_wr_tx_put_slot(struct smc_link *link,
|
|
struct smc_wr_tx_pend_priv *wr_pend_priv);
|
|
int smc_wr_tx_send(struct smc_link *link,
|
|
struct smc_wr_tx_pend_priv *wr_pend_priv);
|
|
int smc_wr_tx_v2_send(struct smc_link *link,
|
|
struct smc_wr_tx_pend_priv *priv, int len);
|
|
int smc_wr_tx_send_wait(struct smc_link *link, struct smc_wr_tx_pend_priv *priv,
|
|
unsigned long timeout);
|
|
void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context);
|
|
void smc_wr_tx_wait_no_pending_sends(struct smc_link *link);
|
|
|
|
int smc_wr_rx_register_handler(struct smc_wr_rx_handler *handler);
|
|
int smc_wr_rx_post_init(struct smc_link *link);
|
|
void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context);
|
|
int smc_wr_reg_send(struct smc_link *link, struct ib_mr *mr);
|
|
|
|
#endif /* SMC_WR_H */
|