mirror of
https://github.com/torvalds/linux.git
synced 2024-11-21 19:41:42 +00:00
netdev: support binding dma-buf to netdevice
Add a netdev_dmabuf_binding struct which represents the dma-buf-to-netdevice binding. The netlink API will bind the dma-buf to rx queues on the netdevice. On the binding, the dma_buf_attach & dma_buf_map_attachment will occur. The entries in the sg_table from mapping will be inserted into a genpool to make it ready for allocation. The chunks in the genpool are owned by a dmabuf_chunk_owner struct which holds the dma-buf offset of the base of the chunk and the dma_addr of the chunk. Both are needed to use allocations that come from this chunk. We create a new type that represents an allocation from the genpool: net_iov. We setup the net_iov allocation size in the genpool to PAGE_SIZE for simplicity: to match the PAGE_SIZE normally allocated by the page pool and given to the drivers. The user can unbind the dmabuf from the netdevice by closing the netlink socket that established the binding. We do this so that the binding is automatically unbound even if the userspace process crashes. The binding and unbinding leaves an indicator in struct netdev_rx_queue that the given queue is bound, and the binding is actuated by resetting the rx queue using the queue API. The netdev_dmabuf_binding struct is refcounted, and releases its resources only when all the refs are released. Signed-off-by: Willem de Bruijn <willemb@google.com> Signed-off-by: Kaiyuan Zhang <kaiyuanz@google.com> Signed-off-by: Mina Almasry <almasrymina@google.com> Reviewed-by: Pavel Begunkov <asml.silence@gmail.com> # excluding netlink Acked-by: Daniel Vetter <daniel.vetter@ffwll.ch> Reviewed-by: Jakub Kicinski <kuba@kernel.org> Link: https://patch.msgid.link/20240910171458.219195-4-almasrymina@google.com Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
parent
3efd7ab46d
commit
170aafe35c
@ -667,6 +667,10 @@ operations:
|
||||
attributes:
|
||||
- id
|
||||
|
||||
kernel-family:
|
||||
headers: [ "linux/list.h"]
|
||||
sock-priv: struct list_head
|
||||
|
||||
mcast-groups:
|
||||
list:
|
||||
-
|
||||
|
@ -3953,6 +3953,8 @@ u8 dev_xdp_prog_count(struct net_device *dev);
|
||||
int dev_xdp_propagate(struct net_device *dev, struct netdev_bpf *bpf);
|
||||
u32 dev_xdp_prog_id(struct net_device *dev, enum bpf_xdp_mode mode);
|
||||
|
||||
u32 dev_get_min_mp_channel_count(const struct net_device *dev);
|
||||
|
||||
int __dev_forward_skb(struct net_device *dev, struct sk_buff *skb);
|
||||
int dev_forward_skb(struct net_device *dev, struct sk_buff *skb);
|
||||
int dev_forward_skb_nomtu(struct net_device *dev, struct sk_buff *skb);
|
||||
|
@ -6,6 +6,7 @@
|
||||
#include <linux/netdevice.h>
|
||||
#include <linux/sysfs.h>
|
||||
#include <net/xdp.h>
|
||||
#include <net/page_pool/types.h>
|
||||
|
||||
/* This structure contains an instance of an RX queue. */
|
||||
struct netdev_rx_queue {
|
||||
@ -25,6 +26,7 @@ struct netdev_rx_queue {
|
||||
* Readers and writers must hold RTNL
|
||||
*/
|
||||
struct napi_struct *napi;
|
||||
struct pp_memory_provider_params mp_params;
|
||||
} ____cacheline_aligned_in_smp;
|
||||
|
||||
/*
|
||||
|
@ -8,6 +8,14 @@
|
||||
#ifndef _NET_NETMEM_H
|
||||
#define _NET_NETMEM_H
|
||||
|
||||
/* net_iov */
|
||||
|
||||
struct net_iov {
|
||||
struct dmabuf_genpool_chunk_owner *owner;
|
||||
};
|
||||
|
||||
/* netmem */
|
||||
|
||||
/**
|
||||
* typedef netmem_ref - a nonexistent type marking a reference to generic
|
||||
* network memory.
|
||||
|
@ -139,6 +139,10 @@ struct page_pool_stats {
|
||||
*/
|
||||
#define PAGE_POOL_FRAG_GROUP_ALIGN (4 * sizeof(long))
|
||||
|
||||
struct pp_memory_provider_params {
|
||||
void *mp_priv;
|
||||
};
|
||||
|
||||
struct page_pool {
|
||||
struct page_pool_params_fast p;
|
||||
|
||||
@ -197,6 +201,8 @@ struct page_pool {
|
||||
*/
|
||||
struct ptr_ring ring;
|
||||
|
||||
void *mp_priv;
|
||||
|
||||
#ifdef CONFIG_PAGE_POOL_STATS
|
||||
/* recycle stats are per-cpu to avoid locking */
|
||||
struct page_pool_recycle_stats __percpu *recycle_stats;
|
||||
|
@ -66,6 +66,11 @@ config SKB_DECRYPTED
|
||||
config SKB_EXTENSIONS
|
||||
bool
|
||||
|
||||
config NET_DEVMEM
|
||||
def_bool y
|
||||
depends on DMA_SHARED_BUFFER
|
||||
depends on GENERIC_ALLOCATOR
|
||||
|
||||
menu "Networking options"
|
||||
|
||||
source "net/packet/Kconfig"
|
||||
|
@ -44,3 +44,4 @@ obj-$(CONFIG_BPF_SYSCALL) += sock_map.o
|
||||
obj-$(CONFIG_BPF_SYSCALL) += bpf_sk_storage.o
|
||||
obj-$(CONFIG_OF) += of_net.o
|
||||
obj-$(CONFIG_NET_TEST) += net_test.o
|
||||
obj-$(CONFIG_NET_DEVMEM) += devmem.o
|
||||
|
@ -161,6 +161,7 @@
|
||||
#include <linux/phy_link_topology.h>
|
||||
|
||||
#include "dev.h"
|
||||
#include "devmem.h"
|
||||
#include "net-sysfs.h"
|
||||
|
||||
static DEFINE_SPINLOCK(ptype_lock);
|
||||
@ -9367,6 +9368,11 @@ int dev_xdp_propagate(struct net_device *dev, struct netdev_bpf *bpf)
|
||||
if (!dev->netdev_ops->ndo_bpf)
|
||||
return -EOPNOTSUPP;
|
||||
|
||||
if (dev_get_min_mp_channel_count(dev)) {
|
||||
NL_SET_ERR_MSG(bpf->extack, "unable to propagate XDP to device using memory provider");
|
||||
return -EBUSY;
|
||||
}
|
||||
|
||||
return dev->netdev_ops->ndo_bpf(dev, bpf);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(dev_xdp_propagate);
|
||||
@ -9399,6 +9405,11 @@ static int dev_xdp_install(struct net_device *dev, enum bpf_xdp_mode mode,
|
||||
struct netdev_bpf xdp;
|
||||
int err;
|
||||
|
||||
if (dev_get_min_mp_channel_count(dev)) {
|
||||
NL_SET_ERR_MSG(extack, "unable to install XDP to device using memory provider");
|
||||
return -EBUSY;
|
||||
}
|
||||
|
||||
memset(&xdp, 0, sizeof(xdp));
|
||||
xdp.command = mode == XDP_MODE_HW ? XDP_SETUP_PROG_HW : XDP_SETUP_PROG;
|
||||
xdp.extack = extack;
|
||||
@ -9823,6 +9834,20 @@ err_out:
|
||||
return err;
|
||||
}
|
||||
|
||||
u32 dev_get_min_mp_channel_count(const struct net_device *dev)
|
||||
{
|
||||
int i;
|
||||
|
||||
ASSERT_RTNL();
|
||||
|
||||
for (i = dev->real_num_rx_queues - 1; i >= 0; i--)
|
||||
if (dev->_rx[i].mp_params.mp_priv)
|
||||
/* The channel count is the idx plus 1. */
|
||||
return i + 1;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* dev_index_reserve() - allocate an ifindex in a namespace
|
||||
* @net: the applicable net namespace
|
||||
@ -11359,6 +11384,7 @@ void unregister_netdevice_many_notify(struct list_head *head,
|
||||
dev_tcx_uninstall(dev);
|
||||
dev_xdp_uninstall(dev);
|
||||
bpf_dev_bound_netdev_unregister(dev);
|
||||
dev_dmabuf_uninstall(dev);
|
||||
|
||||
netdev_offload_xstats_disable_all(dev);
|
||||
|
||||
|
274
net/core/devmem.c
Normal file
274
net/core/devmem.c
Normal file
@ -0,0 +1,274 @@
|
||||
// SPDX-License-Identifier: GPL-2.0-or-later
|
||||
/*
|
||||
* Devmem TCP
|
||||
*
|
||||
* Authors: Mina Almasry <almasrymina@google.com>
|
||||
* Willem de Bruijn <willemdebruijn.kernel@gmail.com>
|
||||
* Kaiyuan Zhang <kaiyuanz@google.com
|
||||
*/
|
||||
|
||||
#include <linux/dma-buf.h>
|
||||
#include <linux/genalloc.h>
|
||||
#include <linux/mm.h>
|
||||
#include <linux/netdevice.h>
|
||||
#include <linux/types.h>
|
||||
#include <net/netdev_queues.h>
|
||||
#include <net/netdev_rx_queue.h>
|
||||
#include <net/page_pool/helpers.h>
|
||||
#include <trace/events/page_pool.h>
|
||||
|
||||
#include "devmem.h"
|
||||
|
||||
/* Device memory support */
|
||||
|
||||
/* Protected by rtnl_lock() */
|
||||
static DEFINE_XARRAY_FLAGS(net_devmem_dmabuf_bindings, XA_FLAGS_ALLOC1);
|
||||
|
||||
static void net_devmem_dmabuf_free_chunk_owner(struct gen_pool *genpool,
|
||||
struct gen_pool_chunk *chunk,
|
||||
void *not_used)
|
||||
{
|
||||
struct dmabuf_genpool_chunk_owner *owner = chunk->owner;
|
||||
|
||||
kvfree(owner->niovs);
|
||||
kfree(owner);
|
||||
}
|
||||
|
||||
void __net_devmem_dmabuf_binding_free(struct net_devmem_dmabuf_binding *binding)
|
||||
{
|
||||
size_t size, avail;
|
||||
|
||||
gen_pool_for_each_chunk(binding->chunk_pool,
|
||||
net_devmem_dmabuf_free_chunk_owner, NULL);
|
||||
|
||||
size = gen_pool_size(binding->chunk_pool);
|
||||
avail = gen_pool_avail(binding->chunk_pool);
|
||||
|
||||
if (!WARN(size != avail, "can't destroy genpool. size=%zu, avail=%zu",
|
||||
size, avail))
|
||||
gen_pool_destroy(binding->chunk_pool);
|
||||
|
||||
dma_buf_unmap_attachment_unlocked(binding->attachment, binding->sgt,
|
||||
DMA_FROM_DEVICE);
|
||||
dma_buf_detach(binding->dmabuf, binding->attachment);
|
||||
dma_buf_put(binding->dmabuf);
|
||||
xa_destroy(&binding->bound_rxqs);
|
||||
kfree(binding);
|
||||
}
|
||||
|
||||
void net_devmem_unbind_dmabuf(struct net_devmem_dmabuf_binding *binding)
|
||||
{
|
||||
struct netdev_rx_queue *rxq;
|
||||
unsigned long xa_idx;
|
||||
unsigned int rxq_idx;
|
||||
|
||||
if (binding->list.next)
|
||||
list_del(&binding->list);
|
||||
|
||||
xa_for_each(&binding->bound_rxqs, xa_idx, rxq) {
|
||||
WARN_ON(rxq->mp_params.mp_priv != binding);
|
||||
|
||||
rxq->mp_params.mp_priv = NULL;
|
||||
|
||||
rxq_idx = get_netdev_rx_queue_index(rxq);
|
||||
|
||||
WARN_ON(netdev_rx_queue_restart(binding->dev, rxq_idx));
|
||||
}
|
||||
|
||||
xa_erase(&net_devmem_dmabuf_bindings, binding->id);
|
||||
|
||||
net_devmem_dmabuf_binding_put(binding);
|
||||
}
|
||||
|
||||
int net_devmem_bind_dmabuf_to_queue(struct net_device *dev, u32 rxq_idx,
|
||||
struct net_devmem_dmabuf_binding *binding,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
struct netdev_rx_queue *rxq;
|
||||
u32 xa_idx;
|
||||
int err;
|
||||
|
||||
if (rxq_idx >= dev->real_num_rx_queues) {
|
||||
NL_SET_ERR_MSG(extack, "rx queue index out of range");
|
||||
return -ERANGE;
|
||||
}
|
||||
|
||||
rxq = __netif_get_rx_queue(dev, rxq_idx);
|
||||
if (rxq->mp_params.mp_priv) {
|
||||
NL_SET_ERR_MSG(extack, "designated queue already memory provider bound");
|
||||
return -EEXIST;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_XDP_SOCKETS
|
||||
if (rxq->pool) {
|
||||
NL_SET_ERR_MSG(extack, "designated queue already in use by AF_XDP");
|
||||
return -EBUSY;
|
||||
}
|
||||
#endif
|
||||
|
||||
err = xa_alloc(&binding->bound_rxqs, &xa_idx, rxq, xa_limit_32b,
|
||||
GFP_KERNEL);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
rxq->mp_params.mp_priv = binding;
|
||||
|
||||
err = netdev_rx_queue_restart(dev, rxq_idx);
|
||||
if (err)
|
||||
goto err_xa_erase;
|
||||
|
||||
return 0;
|
||||
|
||||
err_xa_erase:
|
||||
rxq->mp_params.mp_priv = NULL;
|
||||
xa_erase(&binding->bound_rxqs, xa_idx);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
struct net_devmem_dmabuf_binding *
|
||||
net_devmem_bind_dmabuf(struct net_device *dev, unsigned int dmabuf_fd,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
struct net_devmem_dmabuf_binding *binding;
|
||||
static u32 id_alloc_next;
|
||||
struct scatterlist *sg;
|
||||
struct dma_buf *dmabuf;
|
||||
unsigned int sg_idx, i;
|
||||
unsigned long virtual;
|
||||
int err;
|
||||
|
||||
dmabuf = dma_buf_get(dmabuf_fd);
|
||||
if (IS_ERR(dmabuf))
|
||||
return ERR_CAST(dmabuf);
|
||||
|
||||
binding = kzalloc_node(sizeof(*binding), GFP_KERNEL,
|
||||
dev_to_node(&dev->dev));
|
||||
if (!binding) {
|
||||
err = -ENOMEM;
|
||||
goto err_put_dmabuf;
|
||||
}
|
||||
|
||||
binding->dev = dev;
|
||||
|
||||
err = xa_alloc_cyclic(&net_devmem_dmabuf_bindings, &binding->id,
|
||||
binding, xa_limit_32b, &id_alloc_next,
|
||||
GFP_KERNEL);
|
||||
if (err < 0)
|
||||
goto err_free_binding;
|
||||
|
||||
xa_init_flags(&binding->bound_rxqs, XA_FLAGS_ALLOC);
|
||||
|
||||
refcount_set(&binding->ref, 1);
|
||||
|
||||
binding->dmabuf = dmabuf;
|
||||
|
||||
binding->attachment = dma_buf_attach(binding->dmabuf, dev->dev.parent);
|
||||
if (IS_ERR(binding->attachment)) {
|
||||
err = PTR_ERR(binding->attachment);
|
||||
NL_SET_ERR_MSG(extack, "Failed to bind dmabuf to device");
|
||||
goto err_free_id;
|
||||
}
|
||||
|
||||
binding->sgt = dma_buf_map_attachment_unlocked(binding->attachment,
|
||||
DMA_FROM_DEVICE);
|
||||
if (IS_ERR(binding->sgt)) {
|
||||
err = PTR_ERR(binding->sgt);
|
||||
NL_SET_ERR_MSG(extack, "Failed to map dmabuf attachment");
|
||||
goto err_detach;
|
||||
}
|
||||
|
||||
/* For simplicity we expect to make PAGE_SIZE allocations, but the
|
||||
* binding can be much more flexible than that. We may be able to
|
||||
* allocate MTU sized chunks here. Leave that for future work...
|
||||
*/
|
||||
binding->chunk_pool =
|
||||
gen_pool_create(PAGE_SHIFT, dev_to_node(&dev->dev));
|
||||
if (!binding->chunk_pool) {
|
||||
err = -ENOMEM;
|
||||
goto err_unmap;
|
||||
}
|
||||
|
||||
virtual = 0;
|
||||
for_each_sgtable_dma_sg(binding->sgt, sg, sg_idx) {
|
||||
dma_addr_t dma_addr = sg_dma_address(sg);
|
||||
struct dmabuf_genpool_chunk_owner *owner;
|
||||
size_t len = sg_dma_len(sg);
|
||||
struct net_iov *niov;
|
||||
|
||||
owner = kzalloc_node(sizeof(*owner), GFP_KERNEL,
|
||||
dev_to_node(&dev->dev));
|
||||
if (!owner) {
|
||||
err = -ENOMEM;
|
||||
goto err_free_chunks;
|
||||
}
|
||||
|
||||
owner->base_virtual = virtual;
|
||||
owner->base_dma_addr = dma_addr;
|
||||
owner->num_niovs = len / PAGE_SIZE;
|
||||
owner->binding = binding;
|
||||
|
||||
err = gen_pool_add_owner(binding->chunk_pool, dma_addr,
|
||||
dma_addr, len, dev_to_node(&dev->dev),
|
||||
owner);
|
||||
if (err) {
|
||||
kfree(owner);
|
||||
err = -EINVAL;
|
||||
goto err_free_chunks;
|
||||
}
|
||||
|
||||
owner->niovs = kvmalloc_array(owner->num_niovs,
|
||||
sizeof(*owner->niovs),
|
||||
GFP_KERNEL);
|
||||
if (!owner->niovs) {
|
||||
err = -ENOMEM;
|
||||
goto err_free_chunks;
|
||||
}
|
||||
|
||||
for (i = 0; i < owner->num_niovs; i++) {
|
||||
niov = &owner->niovs[i];
|
||||
niov->owner = owner;
|
||||
}
|
||||
|
||||
virtual += len;
|
||||
}
|
||||
|
||||
return binding;
|
||||
|
||||
err_free_chunks:
|
||||
gen_pool_for_each_chunk(binding->chunk_pool,
|
||||
net_devmem_dmabuf_free_chunk_owner, NULL);
|
||||
gen_pool_destroy(binding->chunk_pool);
|
||||
err_unmap:
|
||||
dma_buf_unmap_attachment_unlocked(binding->attachment, binding->sgt,
|
||||
DMA_FROM_DEVICE);
|
||||
err_detach:
|
||||
dma_buf_detach(dmabuf, binding->attachment);
|
||||
err_free_id:
|
||||
xa_erase(&net_devmem_dmabuf_bindings, binding->id);
|
||||
err_free_binding:
|
||||
kfree(binding);
|
||||
err_put_dmabuf:
|
||||
dma_buf_put(dmabuf);
|
||||
return ERR_PTR(err);
|
||||
}
|
||||
|
||||
void dev_dmabuf_uninstall(struct net_device *dev)
|
||||
{
|
||||
struct net_devmem_dmabuf_binding *binding;
|
||||
struct netdev_rx_queue *rxq;
|
||||
unsigned long xa_idx;
|
||||
unsigned int i;
|
||||
|
||||
for (i = 0; i < dev->real_num_rx_queues; i++) {
|
||||
binding = dev->_rx[i].mp_params.mp_priv;
|
||||
if (!binding)
|
||||
continue;
|
||||
|
||||
xa_for_each(&binding->bound_rxqs, xa_idx, rxq)
|
||||
if (rxq == &dev->_rx[i]) {
|
||||
xa_erase(&binding->bound_rxqs, xa_idx);
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
124
net/core/devmem.h
Normal file
124
net/core/devmem.h
Normal file
@ -0,0 +1,124 @@
|
||||
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
||||
/*
|
||||
* Device memory TCP support
|
||||
*
|
||||
* Authors: Mina Almasry <almasrymina@google.com>
|
||||
* Willem de Bruijn <willemb@google.com>
|
||||
* Kaiyuan Zhang <kaiyuanz@google.com>
|
||||
*
|
||||
*/
|
||||
#ifndef _NET_DEVMEM_H
|
||||
#define _NET_DEVMEM_H
|
||||
|
||||
struct netlink_ext_ack;
|
||||
|
||||
struct net_devmem_dmabuf_binding {
|
||||
struct dma_buf *dmabuf;
|
||||
struct dma_buf_attachment *attachment;
|
||||
struct sg_table *sgt;
|
||||
struct net_device *dev;
|
||||
struct gen_pool *chunk_pool;
|
||||
|
||||
/* The user holds a ref (via the netlink API) for as long as they want
|
||||
* the binding to remain alive. Each page pool using this binding holds
|
||||
* a ref to keep the binding alive. Each allocated net_iov holds a
|
||||
* ref.
|
||||
*
|
||||
* The binding undos itself and unmaps the underlying dmabuf once all
|
||||
* those refs are dropped and the binding is no longer desired or in
|
||||
* use.
|
||||
*/
|
||||
refcount_t ref;
|
||||
|
||||
/* The list of bindings currently active. Used for netlink to notify us
|
||||
* of the user dropping the bind.
|
||||
*/
|
||||
struct list_head list;
|
||||
|
||||
/* rxq's this binding is active on. */
|
||||
struct xarray bound_rxqs;
|
||||
|
||||
/* ID of this binding. Globally unique to all bindings currently
|
||||
* active.
|
||||
*/
|
||||
u32 id;
|
||||
};
|
||||
|
||||
#if defined(CONFIG_NET_DEVMEM)
|
||||
/* Owner of the dma-buf chunks inserted into the gen pool. Each scatterlist
|
||||
* entry from the dmabuf is inserted into the genpool as a chunk, and needs
|
||||
* this owner struct to keep track of some metadata necessary to create
|
||||
* allocations from this chunk.
|
||||
*/
|
||||
struct dmabuf_genpool_chunk_owner {
|
||||
/* Offset into the dma-buf where this chunk starts. */
|
||||
unsigned long base_virtual;
|
||||
|
||||
/* dma_addr of the start of the chunk. */
|
||||
dma_addr_t base_dma_addr;
|
||||
|
||||
/* Array of net_iovs for this chunk. */
|
||||
struct net_iov *niovs;
|
||||
size_t num_niovs;
|
||||
|
||||
struct net_devmem_dmabuf_binding *binding;
|
||||
};
|
||||
|
||||
void __net_devmem_dmabuf_binding_free(struct net_devmem_dmabuf_binding *binding);
|
||||
struct net_devmem_dmabuf_binding *
|
||||
net_devmem_bind_dmabuf(struct net_device *dev, unsigned int dmabuf_fd,
|
||||
struct netlink_ext_ack *extack);
|
||||
void net_devmem_unbind_dmabuf(struct net_devmem_dmabuf_binding *binding);
|
||||
int net_devmem_bind_dmabuf_to_queue(struct net_device *dev, u32 rxq_idx,
|
||||
struct net_devmem_dmabuf_binding *binding,
|
||||
struct netlink_ext_ack *extack);
|
||||
void dev_dmabuf_uninstall(struct net_device *dev);
|
||||
|
||||
static inline void
|
||||
net_devmem_dmabuf_binding_get(struct net_devmem_dmabuf_binding *binding)
|
||||
{
|
||||
refcount_inc(&binding->ref);
|
||||
}
|
||||
|
||||
static inline void
|
||||
net_devmem_dmabuf_binding_put(struct net_devmem_dmabuf_binding *binding)
|
||||
{
|
||||
if (!refcount_dec_and_test(&binding->ref))
|
||||
return;
|
||||
|
||||
__net_devmem_dmabuf_binding_free(binding);
|
||||
}
|
||||
|
||||
#else
|
||||
static inline void
|
||||
__net_devmem_dmabuf_binding_free(struct net_devmem_dmabuf_binding *binding)
|
||||
{
|
||||
}
|
||||
|
||||
static inline struct net_devmem_dmabuf_binding *
|
||||
net_devmem_bind_dmabuf(struct net_device *dev, unsigned int dmabuf_fd,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
return ERR_PTR(-EOPNOTSUPP);
|
||||
}
|
||||
|
||||
static inline void
|
||||
net_devmem_unbind_dmabuf(struct net_devmem_dmabuf_binding *binding)
|
||||
{
|
||||
}
|
||||
|
||||
static inline int
|
||||
net_devmem_bind_dmabuf_to_queue(struct net_device *dev, u32 rxq_idx,
|
||||
struct net_devmem_dmabuf_binding *binding,
|
||||
struct netlink_ext_ack *extack)
|
||||
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
||||
static inline void dev_dmabuf_uninstall(struct net_device *dev)
|
||||
{
|
||||
}
|
||||
#endif
|
||||
|
||||
#endif /* _NET_DEVMEM_H */
|
@ -9,6 +9,7 @@
|
||||
#include "netdev-genl-gen.h"
|
||||
|
||||
#include <uapi/linux/netdev.h>
|
||||
#include <linux/list.h>
|
||||
|
||||
/* Integer value ranges */
|
||||
static const struct netlink_range_validation netdev_a_page_pool_id_range = {
|
||||
@ -187,4 +188,7 @@ struct genl_family netdev_nl_family __ro_after_init = {
|
||||
.n_split_ops = ARRAY_SIZE(netdev_nl_ops),
|
||||
.mcgrps = netdev_nl_mcgrps,
|
||||
.n_mcgrps = ARRAY_SIZE(netdev_nl_mcgrps),
|
||||
.sock_priv_size = sizeof(struct list_head),
|
||||
.sock_priv_init = (void *)netdev_nl_sock_priv_init,
|
||||
.sock_priv_destroy = (void *)netdev_nl_sock_priv_destroy,
|
||||
};
|
||||
|
@ -10,6 +10,7 @@
|
||||
#include <net/genetlink.h>
|
||||
|
||||
#include <uapi/linux/netdev.h>
|
||||
#include <linux/list.h>
|
||||
|
||||
/* Common nested types */
|
||||
extern const struct nla_policy netdev_page_pool_info_nl_policy[NETDEV_A_PAGE_POOL_IFINDEX + 1];
|
||||
@ -40,4 +41,7 @@ enum {
|
||||
|
||||
extern struct genl_family netdev_nl_family;
|
||||
|
||||
void netdev_nl_sock_priv_init(struct list_head *priv);
|
||||
void netdev_nl_sock_priv_destroy(struct list_head *priv);
|
||||
|
||||
#endif /* _LINUX_NETDEV_GEN_H */
|
||||
|
@ -3,16 +3,17 @@
|
||||
#include <linux/netdevice.h>
|
||||
#include <linux/notifier.h>
|
||||
#include <linux/rtnetlink.h>
|
||||
#include <net/busy_poll.h>
|
||||
#include <net/net_namespace.h>
|
||||
#include <net/netdev_queues.h>
|
||||
#include <net/netdev_rx_queue.h>
|
||||
#include <net/sock.h>
|
||||
#include <net/xdp.h>
|
||||
#include <net/xdp_sock.h>
|
||||
#include <net/netdev_rx_queue.h>
|
||||
#include <net/netdev_queues.h>
|
||||
#include <net/busy_poll.h>
|
||||
|
||||
#include "netdev-genl-gen.h"
|
||||
#include "dev.h"
|
||||
#include "devmem.h"
|
||||
#include "netdev-genl-gen.h"
|
||||
|
||||
struct netdev_nl_dump_ctx {
|
||||
unsigned long ifindex;
|
||||
@ -723,10 +724,127 @@ int netdev_nl_qstats_get_dumpit(struct sk_buff *skb,
|
||||
return err;
|
||||
}
|
||||
|
||||
/* Stub */
|
||||
int netdev_nl_bind_rx_doit(struct sk_buff *skb, struct genl_info *info)
|
||||
{
|
||||
struct nlattr *tb[ARRAY_SIZE(netdev_queue_id_nl_policy)];
|
||||
struct net_devmem_dmabuf_binding *binding;
|
||||
struct list_head *sock_binding_list;
|
||||
u32 ifindex, dmabuf_fd, rxq_idx;
|
||||
struct net_device *netdev;
|
||||
struct sk_buff *rsp;
|
||||
struct nlattr *attr;
|
||||
int rem, err = 0;
|
||||
void *hdr;
|
||||
|
||||
if (GENL_REQ_ATTR_CHECK(info, NETDEV_A_DEV_IFINDEX) ||
|
||||
GENL_REQ_ATTR_CHECK(info, NETDEV_A_DMABUF_FD) ||
|
||||
GENL_REQ_ATTR_CHECK(info, NETDEV_A_DMABUF_QUEUES))
|
||||
return -EINVAL;
|
||||
|
||||
ifindex = nla_get_u32(info->attrs[NETDEV_A_DEV_IFINDEX]);
|
||||
dmabuf_fd = nla_get_u32(info->attrs[NETDEV_A_DMABUF_FD]);
|
||||
|
||||
sock_binding_list = genl_sk_priv_get(&netdev_nl_family,
|
||||
NETLINK_CB(skb).sk);
|
||||
if (IS_ERR(sock_binding_list))
|
||||
return PTR_ERR(sock_binding_list);
|
||||
|
||||
rsp = genlmsg_new(GENLMSG_DEFAULT_SIZE, GFP_KERNEL);
|
||||
if (!rsp)
|
||||
return -ENOMEM;
|
||||
|
||||
hdr = genlmsg_iput(rsp, info);
|
||||
if (!hdr) {
|
||||
err = -EMSGSIZE;
|
||||
goto err_genlmsg_free;
|
||||
}
|
||||
|
||||
rtnl_lock();
|
||||
|
||||
netdev = __dev_get_by_index(genl_info_net(info), ifindex);
|
||||
if (!netdev || !netif_device_present(netdev)) {
|
||||
err = -ENODEV;
|
||||
goto err_unlock;
|
||||
}
|
||||
|
||||
if (dev_xdp_prog_count(netdev)) {
|
||||
NL_SET_ERR_MSG(info->extack, "unable to bind dmabuf to device with XDP program attached");
|
||||
err = -EEXIST;
|
||||
goto err_unlock;
|
||||
}
|
||||
|
||||
binding = net_devmem_bind_dmabuf(netdev, dmabuf_fd, info->extack);
|
||||
if (IS_ERR(binding)) {
|
||||
err = PTR_ERR(binding);
|
||||
goto err_unlock;
|
||||
}
|
||||
|
||||
nla_for_each_attr_type(attr, NETDEV_A_DMABUF_QUEUES,
|
||||
genlmsg_data(info->genlhdr),
|
||||
genlmsg_len(info->genlhdr), rem) {
|
||||
err = nla_parse_nested(
|
||||
tb, ARRAY_SIZE(netdev_queue_id_nl_policy) - 1, attr,
|
||||
netdev_queue_id_nl_policy, info->extack);
|
||||
if (err < 0)
|
||||
goto err_unbind;
|
||||
|
||||
if (NL_REQ_ATTR_CHECK(info->extack, attr, tb, NETDEV_A_QUEUE_ID) ||
|
||||
NL_REQ_ATTR_CHECK(info->extack, attr, tb, NETDEV_A_QUEUE_TYPE)) {
|
||||
err = -EINVAL;
|
||||
goto err_unbind;
|
||||
}
|
||||
|
||||
if (nla_get_u32(tb[NETDEV_A_QUEUE_TYPE]) != NETDEV_QUEUE_TYPE_RX) {
|
||||
NL_SET_BAD_ATTR(info->extack, tb[NETDEV_A_QUEUE_TYPE]);
|
||||
err = -EINVAL;
|
||||
goto err_unbind;
|
||||
}
|
||||
|
||||
rxq_idx = nla_get_u32(tb[NETDEV_A_QUEUE_ID]);
|
||||
|
||||
err = net_devmem_bind_dmabuf_to_queue(netdev, rxq_idx, binding,
|
||||
info->extack);
|
||||
if (err)
|
||||
goto err_unbind;
|
||||
}
|
||||
|
||||
list_add(&binding->list, sock_binding_list);
|
||||
|
||||
nla_put_u32(rsp, NETDEV_A_DMABUF_ID, binding->id);
|
||||
genlmsg_end(rsp, hdr);
|
||||
|
||||
err = genlmsg_reply(rsp, info);
|
||||
if (err)
|
||||
goto err_unbind;
|
||||
|
||||
rtnl_unlock();
|
||||
|
||||
return 0;
|
||||
|
||||
err_unbind:
|
||||
net_devmem_unbind_dmabuf(binding);
|
||||
err_unlock:
|
||||
rtnl_unlock();
|
||||
err_genlmsg_free:
|
||||
nlmsg_free(rsp);
|
||||
return err;
|
||||
}
|
||||
|
||||
void netdev_nl_sock_priv_init(struct list_head *priv)
|
||||
{
|
||||
INIT_LIST_HEAD(priv);
|
||||
}
|
||||
|
||||
void netdev_nl_sock_priv_destroy(struct list_head *priv)
|
||||
{
|
||||
struct net_devmem_dmabuf_binding *binding;
|
||||
struct net_devmem_dmabuf_binding *temp;
|
||||
|
||||
list_for_each_entry_safe(binding, temp, priv, list) {
|
||||
rtnl_lock();
|
||||
net_devmem_unbind_dmabuf(binding);
|
||||
rtnl_unlock();
|
||||
}
|
||||
}
|
||||
|
||||
static int netdev_genl_netdevice_event(struct notifier_block *nb,
|
||||
|
@ -655,6 +655,7 @@ int ethtool_check_max_channel(struct net_device *dev,
|
||||
{
|
||||
u64 max_rxnfc_in_use;
|
||||
u32 max_rxfh_in_use;
|
||||
int max_mp_in_use;
|
||||
|
||||
/* ensure the new Rx count fits within the configured Rx flow
|
||||
* indirection table/rxnfc settings
|
||||
@ -673,6 +674,13 @@ int ethtool_check_max_channel(struct net_device *dev,
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
max_mp_in_use = dev_get_min_mp_channel_count(dev);
|
||||
if (channels.combined_count + channels.rx_count <= max_mp_in_use) {
|
||||
if (info)
|
||||
GENL_SET_ERR_MSG_FMT(info, "requested channel counts are too low for existing memory provider setting (%d)", max_mp_in_use);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -211,6 +211,11 @@ int xp_assign_dev(struct xsk_buff_pool *pool,
|
||||
goto err_unreg_pool;
|
||||
}
|
||||
|
||||
if (dev_get_min_mp_channel_count(netdev)) {
|
||||
err = -EBUSY;
|
||||
goto err_unreg_pool;
|
||||
}
|
||||
|
||||
bpf.command = XDP_SETUP_XSK_POOL;
|
||||
bpf.xsk.pool = pool;
|
||||
bpf.xsk.queue_id = queue_id;
|
||||
|
Loading…
Reference in New Issue
Block a user