forked from Minki/linux
vdpa/mlx5: Add shared memory registration code
Add code to support registering address space region for the device. The virtio driver can run as either: 1. Guest virtio driver 2. Userspace virtio driver on the host 3. Kernel virtio driver on the host In any case a memory key object is required to provide access to memory for the device. This code will be shared by network or block driver implementations. Reviewed-by: Parav Pandit <parav@mellanox.com> Signed-off-by: Eli Cohen <eli@mellanox.com> Link: https://lore.kernel.org/r/20200804162048.22587-12-eli@mellanox.com Signed-off-by: Michael S. Tsirkin <mst@redhat.com>
This commit is contained in:
parent
29064bfdab
commit
94abbccdf2
@ -1 +1 @@
|
||||
obj-$(CONFIG_MLX5_VDPA) += core/resources.o
|
||||
obj-$(CONFIG_MLX5_VDPA) += core/resources.o core/mr.o
|
||||
|
@ -1 +0,0 @@
|
||||
obj-y += resources.o
|
@ -7,6 +7,31 @@
|
||||
#include <linux/vdpa.h>
|
||||
#include <linux/mlx5/driver.h>
|
||||
|
||||
struct mlx5_vdpa_direct_mr {
|
||||
u64 start;
|
||||
u64 end;
|
||||
u32 perm;
|
||||
struct mlx5_core_mkey mr;
|
||||
struct sg_table sg_head;
|
||||
int log_size;
|
||||
int nsg;
|
||||
struct list_head list;
|
||||
u64 offset;
|
||||
};
|
||||
|
||||
struct mlx5_vdpa_mr {
|
||||
struct mlx5_core_mkey mkey;
|
||||
|
||||
/* list of direct MRs descendants of this indirect mr */
|
||||
struct list_head head;
|
||||
unsigned long num_directs;
|
||||
unsigned long num_klms;
|
||||
bool initialized;
|
||||
|
||||
/* serialize mkey creation and destruction */
|
||||
struct mutex mkey_mtx;
|
||||
};
|
||||
|
||||
struct mlx5_vdpa_resources {
|
||||
u32 pdn;
|
||||
struct mlx5_uars_page *uar;
|
||||
@ -26,6 +51,8 @@ struct mlx5_vdpa_dev {
|
||||
u8 status;
|
||||
u32 max_vqs;
|
||||
u32 generation;
|
||||
|
||||
struct mlx5_vdpa_mr mr;
|
||||
};
|
||||
|
||||
int mlx5_vdpa_alloc_pd(struct mlx5_vdpa_dev *dev, u32 *pdn, u16 uid);
|
||||
@ -41,6 +68,13 @@ int mlx5_vdpa_alloc_transport_domain(struct mlx5_vdpa_dev *mvdev, u32 *tdn);
|
||||
void mlx5_vdpa_dealloc_transport_domain(struct mlx5_vdpa_dev *mvdev, u32 tdn);
|
||||
int mlx5_vdpa_alloc_resources(struct mlx5_vdpa_dev *mvdev);
|
||||
void mlx5_vdpa_free_resources(struct mlx5_vdpa_dev *mvdev);
|
||||
int mlx5_vdpa_create_mkey(struct mlx5_vdpa_dev *mvdev, struct mlx5_core_mkey *mkey, u32 *in,
|
||||
int inlen);
|
||||
int mlx5_vdpa_destroy_mkey(struct mlx5_vdpa_dev *mvdev, struct mlx5_core_mkey *mkey);
|
||||
int mlx5_vdpa_handle_set_map(struct mlx5_vdpa_dev *mvdev, struct vhost_iotlb *iotlb,
|
||||
bool *change_map);
|
||||
int mlx5_vdpa_create_mr(struct mlx5_vdpa_dev *mvdev, struct vhost_iotlb *iotlb);
|
||||
void mlx5_vdpa_destroy_mr(struct mlx5_vdpa_dev *mvdev);
|
||||
|
||||
#define mlx5_vdpa_warn(__dev, format, ...) \
|
||||
dev_warn((__dev)->mdev->device, "%s:%d:(pid %d) warning: " format, __func__, __LINE__, \
|
||||
|
484
drivers/vdpa/mlx5/core/mr.c
Normal file
484
drivers/vdpa/mlx5/core/mr.c
Normal file
@ -0,0 +1,484 @@
|
||||
// SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
|
||||
/* Copyright (c) 2020 Mellanox Technologies Ltd. */
|
||||
|
||||
#include <linux/vdpa.h>
|
||||
#include <linux/gcd.h>
|
||||
#include <linux/string.h>
|
||||
#include <linux/mlx5/qp.h>
|
||||
#include "mlx5_vdpa.h"
|
||||
|
||||
/* DIV_ROUND_UP where the divider is a power of 2 give by its log base 2 value */
|
||||
#define MLX5_DIV_ROUND_UP_POW2(_n, _s) \
|
||||
({ \
|
||||
u64 __s = _s; \
|
||||
u64 _res; \
|
||||
_res = (((_n) + (1 << (__s)) - 1) >> (__s)); \
|
||||
_res; \
|
||||
})
|
||||
|
||||
static int get_octo_len(u64 len, int page_shift)
|
||||
{
|
||||
u64 page_size = 1ULL << page_shift;
|
||||
int npages;
|
||||
|
||||
npages = ALIGN(len, page_size) >> page_shift;
|
||||
return (npages + 1) / 2;
|
||||
}
|
||||
|
||||
static void fill_sg(struct mlx5_vdpa_direct_mr *mr, void *in)
|
||||
{
|
||||
struct scatterlist *sg;
|
||||
__be64 *pas;
|
||||
int i;
|
||||
|
||||
pas = MLX5_ADDR_OF(create_mkey_in, in, klm_pas_mtt);
|
||||
for_each_sg(mr->sg_head.sgl, sg, mr->nsg, i)
|
||||
(*pas) = cpu_to_be64(sg_dma_address(sg));
|
||||
}
|
||||
|
||||
static void mlx5_set_access_mode(void *mkc, int mode)
|
||||
{
|
||||
MLX5_SET(mkc, mkc, access_mode_1_0, mode & 0x3);
|
||||
MLX5_SET(mkc, mkc, access_mode_4_2, mode >> 2);
|
||||
}
|
||||
|
||||
static void populate_mtts(struct mlx5_vdpa_direct_mr *mr, __be64 *mtt)
|
||||
{
|
||||
struct scatterlist *sg;
|
||||
int i;
|
||||
|
||||
for_each_sg(mr->sg_head.sgl, sg, mr->nsg, i)
|
||||
mtt[i] = cpu_to_be64(sg_dma_address(sg));
|
||||
}
|
||||
|
||||
static int create_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr)
|
||||
{
|
||||
int inlen;
|
||||
void *mkc;
|
||||
void *in;
|
||||
int err;
|
||||
|
||||
inlen = MLX5_ST_SZ_BYTES(create_mkey_in) + roundup(MLX5_ST_SZ_BYTES(mtt) * mr->nsg, 16);
|
||||
in = kvzalloc(inlen, GFP_KERNEL);
|
||||
if (!in)
|
||||
return -ENOMEM;
|
||||
|
||||
MLX5_SET(create_mkey_in, in, uid, mvdev->res.uid);
|
||||
fill_sg(mr, in);
|
||||
mkc = MLX5_ADDR_OF(create_mkey_in, in, memory_key_mkey_entry);
|
||||
MLX5_SET(mkc, mkc, lw, !!(mr->perm & VHOST_MAP_WO));
|
||||
MLX5_SET(mkc, mkc, lr, !!(mr->perm & VHOST_MAP_RO));
|
||||
mlx5_set_access_mode(mkc, MLX5_MKC_ACCESS_MODE_MTT);
|
||||
MLX5_SET(mkc, mkc, qpn, 0xffffff);
|
||||
MLX5_SET(mkc, mkc, pd, mvdev->res.pdn);
|
||||
MLX5_SET64(mkc, mkc, start_addr, mr->offset);
|
||||
MLX5_SET64(mkc, mkc, len, mr->end - mr->start);
|
||||
MLX5_SET(mkc, mkc, log_page_size, mr->log_size);
|
||||
MLX5_SET(mkc, mkc, translations_octword_size,
|
||||
get_octo_len(mr->end - mr->start, mr->log_size));
|
||||
MLX5_SET(create_mkey_in, in, translations_octword_actual_size,
|
||||
get_octo_len(mr->end - mr->start, mr->log_size));
|
||||
populate_mtts(mr, MLX5_ADDR_OF(create_mkey_in, in, klm_pas_mtt));
|
||||
err = mlx5_vdpa_create_mkey(mvdev, &mr->mr, in, inlen);
|
||||
kvfree(in);
|
||||
if (err) {
|
||||
mlx5_vdpa_warn(mvdev, "Failed to create direct MR\n");
|
||||
return err;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void destroy_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr)
|
||||
{
|
||||
mlx5_vdpa_destroy_mkey(mvdev, &mr->mr);
|
||||
}
|
||||
|
||||
static u64 map_start(struct vhost_iotlb_map *map, struct mlx5_vdpa_direct_mr *mr)
|
||||
{
|
||||
return max_t(u64, map->start, mr->start);
|
||||
}
|
||||
|
||||
static u64 map_end(struct vhost_iotlb_map *map, struct mlx5_vdpa_direct_mr *mr)
|
||||
{
|
||||
return min_t(u64, map->last + 1, mr->end);
|
||||
}
|
||||
|
||||
static u64 maplen(struct vhost_iotlb_map *map, struct mlx5_vdpa_direct_mr *mr)
|
||||
{
|
||||
return map_end(map, mr) - map_start(map, mr);
|
||||
}
|
||||
|
||||
#define MLX5_VDPA_INVALID_START_ADDR ((u64)-1)
|
||||
#define MLX5_VDPA_INVALID_LEN ((u64)-1)
|
||||
|
||||
static u64 indir_start_addr(struct mlx5_vdpa_mr *mkey)
|
||||
{
|
||||
struct mlx5_vdpa_direct_mr *s;
|
||||
|
||||
s = list_first_entry_or_null(&mkey->head, struct mlx5_vdpa_direct_mr, list);
|
||||
if (!s)
|
||||
return MLX5_VDPA_INVALID_START_ADDR;
|
||||
|
||||
return s->start;
|
||||
}
|
||||
|
||||
static u64 indir_len(struct mlx5_vdpa_mr *mkey)
|
||||
{
|
||||
struct mlx5_vdpa_direct_mr *s;
|
||||
struct mlx5_vdpa_direct_mr *e;
|
||||
|
||||
s = list_first_entry_or_null(&mkey->head, struct mlx5_vdpa_direct_mr, list);
|
||||
if (!s)
|
||||
return MLX5_VDPA_INVALID_LEN;
|
||||
|
||||
e = list_last_entry(&mkey->head, struct mlx5_vdpa_direct_mr, list);
|
||||
|
||||
return e->end - s->start;
|
||||
}
|
||||
|
||||
#define LOG_MAX_KLM_SIZE 30
|
||||
#define MAX_KLM_SIZE BIT(LOG_MAX_KLM_SIZE)
|
||||
|
||||
static u32 klm_bcount(u64 size)
|
||||
{
|
||||
return (u32)size;
|
||||
}
|
||||
|
||||
static void fill_indir(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mkey, void *in)
|
||||
{
|
||||
struct mlx5_vdpa_direct_mr *dmr;
|
||||
struct mlx5_klm *klmarr;
|
||||
struct mlx5_klm *klm;
|
||||
bool first = true;
|
||||
u64 preve;
|
||||
int i;
|
||||
|
||||
klmarr = MLX5_ADDR_OF(create_mkey_in, in, klm_pas_mtt);
|
||||
i = 0;
|
||||
list_for_each_entry(dmr, &mkey->head, list) {
|
||||
again:
|
||||
klm = &klmarr[i++];
|
||||
if (first) {
|
||||
preve = dmr->start;
|
||||
first = false;
|
||||
}
|
||||
|
||||
if (preve == dmr->start) {
|
||||
klm->key = cpu_to_be32(dmr->mr.key);
|
||||
klm->bcount = cpu_to_be32(klm_bcount(dmr->end - dmr->start));
|
||||
preve = dmr->end;
|
||||
} else {
|
||||
klm->key = cpu_to_be32(mvdev->res.null_mkey);
|
||||
klm->bcount = cpu_to_be32(klm_bcount(dmr->start - preve));
|
||||
preve = dmr->start;
|
||||
goto again;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static int klm_byte_size(int nklms)
|
||||
{
|
||||
return 16 * ALIGN(nklms, 4);
|
||||
}
|
||||
|
||||
static int create_indirect_key(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr)
|
||||
{
|
||||
int inlen;
|
||||
void *mkc;
|
||||
void *in;
|
||||
int err;
|
||||
u64 start;
|
||||
u64 len;
|
||||
|
||||
start = indir_start_addr(mr);
|
||||
len = indir_len(mr);
|
||||
if (start == MLX5_VDPA_INVALID_START_ADDR || len == MLX5_VDPA_INVALID_LEN)
|
||||
return -EINVAL;
|
||||
|
||||
inlen = MLX5_ST_SZ_BYTES(create_mkey_in) + klm_byte_size(mr->num_klms);
|
||||
in = kzalloc(inlen, GFP_KERNEL);
|
||||
if (!in)
|
||||
return -ENOMEM;
|
||||
|
||||
MLX5_SET(create_mkey_in, in, uid, mvdev->res.uid);
|
||||
mkc = MLX5_ADDR_OF(create_mkey_in, in, memory_key_mkey_entry);
|
||||
MLX5_SET(mkc, mkc, lw, 1);
|
||||
MLX5_SET(mkc, mkc, lr, 1);
|
||||
mlx5_set_access_mode(mkc, MLX5_MKC_ACCESS_MODE_KLMS);
|
||||
MLX5_SET(mkc, mkc, qpn, 0xffffff);
|
||||
MLX5_SET(mkc, mkc, pd, mvdev->res.pdn);
|
||||
MLX5_SET64(mkc, mkc, start_addr, start);
|
||||
MLX5_SET64(mkc, mkc, len, len);
|
||||
MLX5_SET(mkc, mkc, translations_octword_size, klm_byte_size(mr->num_klms) / 16);
|
||||
MLX5_SET(create_mkey_in, in, translations_octword_actual_size, mr->num_klms);
|
||||
fill_indir(mvdev, mr, in);
|
||||
err = mlx5_vdpa_create_mkey(mvdev, &mr->mkey, in, inlen);
|
||||
kfree(in);
|
||||
return err;
|
||||
}
|
||||
|
||||
static void destroy_indirect_key(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mkey)
|
||||
{
|
||||
mlx5_vdpa_destroy_mkey(mvdev, &mkey->mkey);
|
||||
}
|
||||
|
||||
static int map_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr,
|
||||
struct vhost_iotlb *iotlb)
|
||||
{
|
||||
struct vhost_iotlb_map *map;
|
||||
unsigned long lgcd = 0;
|
||||
int log_entity_size;
|
||||
unsigned long size;
|
||||
u64 start = 0;
|
||||
int err;
|
||||
struct page *pg;
|
||||
unsigned int nsg;
|
||||
int sglen;
|
||||
u64 pa;
|
||||
u64 paend;
|
||||
struct scatterlist *sg;
|
||||
struct device *dma = mvdev->mdev->device;
|
||||
int ret;
|
||||
|
||||
for (map = vhost_iotlb_itree_first(iotlb, mr->start, mr->end - 1);
|
||||
map; map = vhost_iotlb_itree_next(map, start, mr->end - 1)) {
|
||||
size = maplen(map, mr);
|
||||
lgcd = gcd(lgcd, size);
|
||||
start += size;
|
||||
}
|
||||
log_entity_size = ilog2(lgcd);
|
||||
|
||||
sglen = 1 << log_entity_size;
|
||||
nsg = MLX5_DIV_ROUND_UP_POW2(mr->end - mr->start, log_entity_size);
|
||||
|
||||
err = sg_alloc_table(&mr->sg_head, nsg, GFP_KERNEL);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
sg = mr->sg_head.sgl;
|
||||
for (map = vhost_iotlb_itree_first(iotlb, mr->start, mr->end - 1);
|
||||
map; map = vhost_iotlb_itree_next(map, mr->start, mr->end - 1)) {
|
||||
paend = map->addr + maplen(map, mr);
|
||||
for (pa = map->addr; pa < paend; pa += sglen) {
|
||||
pg = pfn_to_page(__phys_to_pfn(pa));
|
||||
if (!sg) {
|
||||
mlx5_vdpa_warn(mvdev, "sg null. start 0x%llx, end 0x%llx\n",
|
||||
map->start, map->last + 1);
|
||||
err = -ENOMEM;
|
||||
goto err_map;
|
||||
}
|
||||
sg_set_page(sg, pg, sglen, 0);
|
||||
sg = sg_next(sg);
|
||||
if (!sg)
|
||||
goto done;
|
||||
}
|
||||
}
|
||||
done:
|
||||
mr->log_size = log_entity_size;
|
||||
mr->nsg = nsg;
|
||||
ret = dma_map_sg_attrs(dma, mr->sg_head.sgl, mr->nsg, DMA_BIDIRECTIONAL, 0);
|
||||
if (!ret)
|
||||
goto err_map;
|
||||
|
||||
err = create_direct_mr(mvdev, mr);
|
||||
if (err)
|
||||
goto err_direct;
|
||||
|
||||
return 0;
|
||||
|
||||
err_direct:
|
||||
dma_unmap_sg_attrs(dma, mr->sg_head.sgl, mr->nsg, DMA_BIDIRECTIONAL, 0);
|
||||
err_map:
|
||||
sg_free_table(&mr->sg_head);
|
||||
return err;
|
||||
}
|
||||
|
||||
static void unmap_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr)
|
||||
{
|
||||
struct device *dma = mvdev->mdev->device;
|
||||
|
||||
destroy_direct_mr(mvdev, mr);
|
||||
dma_unmap_sg_attrs(dma, mr->sg_head.sgl, mr->nsg, DMA_BIDIRECTIONAL, 0);
|
||||
sg_free_table(&mr->sg_head);
|
||||
}
|
||||
|
||||
static int add_direct_chain(struct mlx5_vdpa_dev *mvdev, u64 start, u64 size, u8 perm,
|
||||
struct vhost_iotlb *iotlb)
|
||||
{
|
||||
struct mlx5_vdpa_mr *mr = &mvdev->mr;
|
||||
struct mlx5_vdpa_direct_mr *dmr;
|
||||
struct mlx5_vdpa_direct_mr *n;
|
||||
LIST_HEAD(tmp);
|
||||
u64 st;
|
||||
u64 sz;
|
||||
int err;
|
||||
int i = 0;
|
||||
|
||||
st = start;
|
||||
while (size) {
|
||||
sz = (u32)min_t(u64, MAX_KLM_SIZE, size);
|
||||
dmr = kzalloc(sizeof(*dmr), GFP_KERNEL);
|
||||
if (!dmr)
|
||||
goto err_alloc;
|
||||
|
||||
dmr->start = st;
|
||||
dmr->end = st + sz;
|
||||
dmr->perm = perm;
|
||||
err = map_direct_mr(mvdev, dmr, iotlb);
|
||||
if (err) {
|
||||
kfree(dmr);
|
||||
goto err_alloc;
|
||||
}
|
||||
|
||||
list_add_tail(&dmr->list, &tmp);
|
||||
size -= sz;
|
||||
mr->num_directs++;
|
||||
mr->num_klms++;
|
||||
st += sz;
|
||||
i++;
|
||||
}
|
||||
list_splice_tail(&tmp, &mr->head);
|
||||
return 0;
|
||||
|
||||
err_alloc:
|
||||
list_for_each_entry_safe(dmr, n, &mr->head, list) {
|
||||
list_del_init(&dmr->list);
|
||||
unmap_direct_mr(mvdev, dmr);
|
||||
kfree(dmr);
|
||||
}
|
||||
return err;
|
||||
}
|
||||
|
||||
/* The iotlb pointer contains a list of maps. Go over the maps, possibly
|
||||
* merging mergeable maps, and create direct memory keys that provide the
|
||||
* device access to memory. The direct mkeys are then referred to by the
|
||||
* indirect memory key that provides access to the enitre address space given
|
||||
* by iotlb.
|
||||
*/
|
||||
static int _mlx5_vdpa_create_mr(struct mlx5_vdpa_dev *mvdev, struct vhost_iotlb *iotlb)
|
||||
{
|
||||
struct mlx5_vdpa_mr *mr = &mvdev->mr;
|
||||
struct mlx5_vdpa_direct_mr *dmr;
|
||||
struct mlx5_vdpa_direct_mr *n;
|
||||
struct vhost_iotlb_map *map;
|
||||
u32 pperm = U16_MAX;
|
||||
u64 last = U64_MAX;
|
||||
u64 ps = U64_MAX;
|
||||
u64 pe = U64_MAX;
|
||||
u64 start = 0;
|
||||
int err = 0;
|
||||
int nnuls;
|
||||
|
||||
if (mr->initialized)
|
||||
return 0;
|
||||
|
||||
INIT_LIST_HEAD(&mr->head);
|
||||
for (map = vhost_iotlb_itree_first(iotlb, start, last); map;
|
||||
map = vhost_iotlb_itree_next(map, start, last)) {
|
||||
start = map->start;
|
||||
if (pe == map->start && pperm == map->perm) {
|
||||
pe = map->last + 1;
|
||||
} else {
|
||||
if (ps != U64_MAX) {
|
||||
if (pe < map->start) {
|
||||
/* We have a hole in the map. Check how
|
||||
* many null keys are required to fill it.
|
||||
*/
|
||||
nnuls = MLX5_DIV_ROUND_UP_POW2(map->start - pe,
|
||||
LOG_MAX_KLM_SIZE);
|
||||
mr->num_klms += nnuls;
|
||||
}
|
||||
err = add_direct_chain(mvdev, ps, pe - ps, pperm, iotlb);
|
||||
if (err)
|
||||
goto err_chain;
|
||||
}
|
||||
ps = map->start;
|
||||
pe = map->last + 1;
|
||||
pperm = map->perm;
|
||||
}
|
||||
}
|
||||
err = add_direct_chain(mvdev, ps, pe - ps, pperm, iotlb);
|
||||
if (err)
|
||||
goto err_chain;
|
||||
|
||||
/* Create the memory key that defines the guests's address space. This
|
||||
* memory key refers to the direct keys that contain the MTT
|
||||
* translations
|
||||
*/
|
||||
err = create_indirect_key(mvdev, mr);
|
||||
if (err)
|
||||
goto err_chain;
|
||||
|
||||
mr->initialized = true;
|
||||
return 0;
|
||||
|
||||
err_chain:
|
||||
list_for_each_entry_safe_reverse(dmr, n, &mr->head, list) {
|
||||
list_del_init(&dmr->list);
|
||||
unmap_direct_mr(mvdev, dmr);
|
||||
kfree(dmr);
|
||||
}
|
||||
return err;
|
||||
}
|
||||
|
||||
int mlx5_vdpa_create_mr(struct mlx5_vdpa_dev *mvdev, struct vhost_iotlb *iotlb)
|
||||
{
|
||||
struct mlx5_vdpa_mr *mr = &mvdev->mr;
|
||||
int err;
|
||||
|
||||
mutex_lock(&mr->mkey_mtx);
|
||||
err = _mlx5_vdpa_create_mr(mvdev, iotlb);
|
||||
mutex_unlock(&mr->mkey_mtx);
|
||||
return err;
|
||||
}
|
||||
|
||||
void mlx5_vdpa_destroy_mr(struct mlx5_vdpa_dev *mvdev)
|
||||
{
|
||||
struct mlx5_vdpa_mr *mr = &mvdev->mr;
|
||||
struct mlx5_vdpa_direct_mr *dmr;
|
||||
struct mlx5_vdpa_direct_mr *n;
|
||||
|
||||
mutex_lock(&mr->mkey_mtx);
|
||||
if (!mr->initialized)
|
||||
goto out;
|
||||
|
||||
destroy_indirect_key(mvdev, mr);
|
||||
list_for_each_entry_safe_reverse(dmr, n, &mr->head, list) {
|
||||
list_del_init(&dmr->list);
|
||||
unmap_direct_mr(mvdev, dmr);
|
||||
kfree(dmr);
|
||||
}
|
||||
memset(mr, 0, sizeof(*mr));
|
||||
mr->initialized = false;
|
||||
out:
|
||||
mutex_unlock(&mr->mkey_mtx);
|
||||
}
|
||||
|
||||
static bool map_empty(struct vhost_iotlb *iotlb)
|
||||
{
|
||||
return !vhost_iotlb_itree_first(iotlb, 0, U64_MAX);
|
||||
}
|
||||
|
||||
int mlx5_vdpa_handle_set_map(struct mlx5_vdpa_dev *mvdev, struct vhost_iotlb *iotlb,
|
||||
bool *change_map)
|
||||
{
|
||||
struct mlx5_vdpa_mr *mr = &mvdev->mr;
|
||||
int err;
|
||||
|
||||
*change_map = false;
|
||||
if (map_empty(iotlb)) {
|
||||
mlx5_vdpa_destroy_mr(mvdev);
|
||||
return 0;
|
||||
}
|
||||
mutex_lock(&mr->mkey_mtx);
|
||||
if (mr->initialized) {
|
||||
mlx5_vdpa_info(mvdev, "memory map update\n");
|
||||
*change_map = true;
|
||||
}
|
||||
if (!*change_map)
|
||||
err = _mlx5_vdpa_create_mr(mvdev, iotlb);
|
||||
mutex_unlock(&mr->mkey_mtx);
|
||||
|
||||
return err;
|
||||
}
|
@ -227,6 +227,7 @@ int mlx5_vdpa_alloc_resources(struct mlx5_vdpa_dev *mvdev)
|
||||
mlx5_vdpa_warn(mvdev, "resources already allocated\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
mutex_init(&mvdev->mr.mkey_mtx);
|
||||
res->uar = mlx5_get_uars_page(mdev);
|
||||
if (IS_ERR(res->uar)) {
|
||||
err = PTR_ERR(res->uar);
|
||||
@ -262,6 +263,7 @@ err_pd:
|
||||
err_uctx:
|
||||
mlx5_put_uars_page(mdev, res->uar);
|
||||
err_uars:
|
||||
mutex_destroy(&mvdev->mr.mkey_mtx);
|
||||
return err;
|
||||
}
|
||||
|
||||
@ -277,5 +279,6 @@ void mlx5_vdpa_free_resources(struct mlx5_vdpa_dev *mvdev)
|
||||
dealloc_pd(mvdev, res->pdn, res->uid);
|
||||
destroy_uctx(mvdev, res->uid);
|
||||
mlx5_put_uars_page(mvdev->mdev, res->uar);
|
||||
mutex_destroy(&mvdev->mr.mkey_mtx);
|
||||
res->valid = false;
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user