From: Pavel Begunkov <[email protected]>

There are currently no easy ways for the user to know if zcrx is out of
buffers and page pool fails to allocate. Add uapi for zcrx to communicate
it back.

It's implemented as a separate CQE, which for now is posted to the creator
ctx. To use it, on registration the user space needs to pass an instance
of struct zcrx_notification_desc, which tells the kernel the user_data
for resulting CQEs and which event types are expected / allowed.

When an allowed event happens, zcrx will post a CQE containing the
specified user_data, and lower bits of cqe->res will be set to the event
mask. Before the kernel could post another notification of the given
type, the user needs to acknowledge that it processed the previous one
by issuing IORING_REGISTER_ZCRX_CTRL with ZCRX_CTRL_ARM_NOTIFICATION.

The only notification type the patch implements yet is
ZCRX_NOTIF_NO_BUFFERS. Next commit adds copy fallback signaling.

Co-developed-by: Vishwanath Seshagiri <[email protected]>
Signed-off-by: Vishwanath Seshagiri <[email protected]>
Signed-off-by: Pavel Begunkov <[email protected]>
---
 include/uapi/linux/io_uring/zcrx.h | 22 ++++++-
 io_uring/zcrx.c                    | 98 +++++++++++++++++++++++++++++-
 io_uring/zcrx.h                    | 11 +++-
 3 files changed, 128 insertions(+), 3 deletions(-)

diff --git a/include/uapi/linux/io_uring/zcrx.h 
b/include/uapi/linux/io_uring/zcrx.h
index 5ce02c7a6096..b8596d7d47b6 100644
--- a/include/uapi/linux/io_uring/zcrx.h
+++ b/include/uapi/linux/io_uring/zcrx.h
@@ -65,6 +65,18 @@ enum zcrx_features {
         * value in struct io_uring_zcrx_ifq_reg::rx_buf_len.
         */
        ZCRX_FEATURE_RX_PAGE_SIZE       = 1 << 0,
+       ZCRX_FEATURE_NOTIFICATION       = 1 << 1,
+};
+
+enum zcrx_notification_type {
+       ZCRX_NOTIF_NO_BUFFERS = 1 << 0,
+};
+
+struct zcrx_notification_desc {
+       __u64   user_data;
+       __u32   type_mask;
+       __u32   __resv1;
+       __u64   __resv2[10];
 };
 
 /*
@@ -82,12 +94,14 @@ struct io_uring_zcrx_ifq_reg {
        struct io_uring_zcrx_offsets offsets;
        __u32   zcrx_id;
        __u32   rx_buf_len;
-       __u64   __resv[3];
+       __u64   notif_desc; /* see struct zcrx_notification_desc */
+       __u64   __resv[2];
 };
 
 enum zcrx_ctrl_op {
        ZCRX_CTRL_FLUSH_RQ,
        ZCRX_CTRL_EXPORT,
+       ZCRX_CTRL_ARM_NOTIFICATION,
 
        __ZCRX_CTRL_LAST,
 };
@@ -101,6 +115,11 @@ struct zcrx_ctrl_export {
        __u32           __resv1[11];
 };
 
+struct zcrx_ctrl_arm_notif {
+       __u32           type_mask;
+       __u32           __resv[11];
+};
+
 struct zcrx_ctrl {
        __u32   zcrx_id;
        __u32   op; /* see enum zcrx_ctrl_op */
@@ -109,6 +128,7 @@ struct zcrx_ctrl {
        union {
                struct zcrx_ctrl_export         zc_export;
                struct zcrx_ctrl_flush_rq       zc_flush;
+               struct zcrx_ctrl_arm_notif      zc_arm_notif;
        };
 };
 
diff --git a/io_uring/zcrx.c b/io_uring/zcrx.c
index 9a83d7eb4210..35ca28cb6583 100644
--- a/io_uring/zcrx.c
+++ b/io_uring/zcrx.c
@@ -44,6 +44,16 @@ static inline struct io_zcrx_area *io_zcrx_iov_to_area(const 
struct net_iov *nio
        return container_of(owner, struct io_zcrx_area, nia);
 }
 
+static bool zcrx_set_ring_ctx(struct io_zcrx_ifq *zcrx, struct io_ring_ctx 
*ctx)
+{
+       guard(spinlock_bh)(&zcrx->ctx_lock);
+       if (zcrx->master_ctx)
+               return false;
+       percpu_ref_get(&ctx->refs);
+       zcrx->master_ctx = ctx;
+       return true;
+}
+
 static inline struct page *io_zcrx_iov_page(const struct net_iov *niov)
 {
        struct io_zcrx_area *area = io_zcrx_iov_to_area(niov);
@@ -531,6 +541,7 @@ static struct io_zcrx_ifq *io_zcrx_ifq_alloc(struct 
io_ring_ctx *ctx)
 
        ifq->if_rxq = -1;
        spin_lock_init(&ifq->rq.lock);
+       spin_lock_init(&ifq->ctx_lock);
        mutex_init(&ifq->pp_lock);
        refcount_set(&ifq->refs, 1);
        refcount_set(&ifq->user_refs, 1);
@@ -585,6 +596,11 @@ static void io_zcrx_ifq_free(struct io_zcrx_ifq *ifq)
        if (ifq->dev)
                put_device(ifq->dev);
 
+       scoped_guard(spinlock_bh, &ifq->ctx_lock) {
+               if (ifq->master_ctx)
+                       percpu_ref_put(&ifq->master_ctx->refs);
+       }
+
        io_free_rbuf_ring(ifq);
        mutex_destroy(&ifq->pp_lock);
        kfree(ifq);
@@ -738,6 +754,8 @@ static int import_zcrx(struct io_ring_ctx *ctx,
                return -EINVAL;
        if (reg->if_rxq || reg->rq_entries || reg->area_ptr || reg->region_ptr)
                return -EINVAL;
+       if (reg->notif_desc)
+               return -EINVAL;
        if (reg->flags & ~ZCRX_REG_IMPORT)
                return -EINVAL;
 
@@ -826,6 +844,7 @@ static int zcrx_register_netdev(struct io_zcrx_ifq *ifq,
 int io_register_zcrx(struct io_ring_ctx *ctx,
                     struct io_uring_zcrx_ifq_reg __user *arg)
 {
+       struct zcrx_notification_desc notif;
        struct io_uring_zcrx_area_reg area;
        struct io_uring_zcrx_ifq_reg reg;
        struct io_uring_region_desc rd;
@@ -869,10 +888,22 @@ int io_register_zcrx(struct io_ring_ctx *ctx,
        if (copy_from_user(&area, u64_to_user_ptr(reg.area_ptr), sizeof(area)))
                return -EFAULT;
 
+       memset(&notif, 0, sizeof(notif));
+       if (reg.notif_desc && copy_from_user(&notif, 
u64_to_user_ptr(reg.notif_desc),
+                                            sizeof(notif)))
+               return -EFAULT;
+       if (notif.type_mask & ~ZCRX_NOTIF_TYPE_MASK)
+               return -EINVAL;
+       if (notif.__resv1 || !mem_is_zero(&notif.__resv2, 
sizeof(notif.__resv2)))
+               return -EINVAL;
+
        ifq = io_zcrx_ifq_alloc(ctx);
        if (!ifq)
                return -ENOMEM;
 
+       ifq->notif_data = notif.user_data;
+       ifq->allowed_notif_mask = notif.type_mask;
+
        if (ctx->user) {
                get_uid(ctx->user);
                ifq->user = ctx->user;
@@ -923,6 +954,9 @@ int io_register_zcrx(struct io_ring_ctx *ctx,
                ret = -EFAULT;
                goto err;
        }
+
+       if (notif.type_mask)
+               zcrx_set_ring_ctx(ifq, ctx);
        return 0;
 err:
        scoped_guard(mutex, &ctx->mmap_lock)
@@ -1089,6 +1123,46 @@ static unsigned io_zcrx_refill_slow(struct page_pool 
*pp, struct io_zcrx_ifq *if
        return allocated;
 }
 
+static void zcrx_notif_tw(struct io_tw_req tw_req, io_tw_token_t tw)
+{
+       struct io_kiocb *req = tw_req.req;
+       struct io_ring_ctx *ctx = req->ctx;
+
+       io_post_aux_cqe(ctx, req->cqe.user_data, req->cqe.res, 0);
+       percpu_ref_put(&ctx->refs);
+       kfree_rcu(req, rcu_head);
+}
+
+static void zcrx_send_notif(struct io_zcrx_ifq *ifq, u32 type_mask)
+{
+       gfp_t gfp = GFP_ATOMIC | __GFP_NOWARN | __GFP_ZERO;
+       struct io_kiocb *req;
+
+       if (!(type_mask & ifq->allowed_notif_mask))
+               return;
+
+       guard(spinlock_bh)(&ifq->ctx_lock);
+       if (!ifq->master_ctx)
+               return;
+       if (type_mask & ifq->fired_notifs)
+               return;
+
+       req = kmem_cache_alloc(req_cachep, gfp);
+       if (unlikely(!req))
+               return;
+
+       ifq->fired_notifs |= type_mask;
+
+       req->opcode = IORING_OP_NOP;
+       req->cqe.user_data = ifq->notif_data;
+       req->cqe.res = type_mask;
+       req->ctx = ifq->master_ctx;
+       percpu_ref_get(&req->ctx->refs);
+       req->tctx = NULL;
+       req->io_task_work.func = zcrx_notif_tw;
+       io_req_task_work_add(req);
+}
+
 static netmem_ref io_pp_zc_alloc_netmems(struct page_pool *pp, gfp_t gfp)
 {
        struct io_zcrx_ifq *ifq = io_pp_to_ifq(pp);
@@ -1105,8 +1179,10 @@ static netmem_ref io_pp_zc_alloc_netmems(struct 
page_pool *pp, gfp_t gfp)
                goto out_return;
 
        allocated = io_zcrx_refill_slow(pp, ifq, netmems, to_alloc);
-       if (!allocated)
+       if (!allocated) {
+               zcrx_send_notif(ifq, ZCRX_NOTIF_NO_BUFFERS);
                return 0;
+       }
 out_return:
        zcrx_sync_for_device(pp, ifq, netmems, allocated);
        allocated--;
@@ -1255,12 +1331,30 @@ static int zcrx_flush_rq(struct io_ring_ctx *ctx, 
struct io_zcrx_ifq *zcrx,
        return 0;
 }
 
+static int zcrx_arm_notif(struct io_ring_ctx *ctx, struct io_zcrx_ifq *zcrx,
+                         struct zcrx_ctrl *ctrl)
+{
+       const struct zcrx_ctrl_arm_notif *an = &ctrl->zc_arm_notif;
+
+       if (an->type_mask & ~ZCRX_NOTIF_TYPE_MASK)
+               return -EINVAL;
+       if (!mem_is_zero(&an->__resv, sizeof(an->__resv)))
+               return -EINVAL;
+
+       guard(spinlock_bh)(&zcrx->ctx_lock);
+       if (an->type_mask & ~zcrx->fired_notifs)
+               return -EINVAL;
+       zcrx->fired_notifs &= ~an->type_mask;
+       return 0;
+}
+
 int io_zcrx_ctrl(struct io_ring_ctx *ctx, void __user *arg, unsigned nr_args)
 {
        struct zcrx_ctrl ctrl;
        struct io_zcrx_ifq *zcrx;
 
        BUILD_BUG_ON(sizeof(ctrl.zc_export) != sizeof(ctrl.zc_flush));
+       BUILD_BUG_ON(sizeof(ctrl.zc_export) != sizeof(ctrl.zc_arm_notif));
 
        if (nr_args)
                return -EINVAL;
@@ -1278,6 +1372,8 @@ int io_zcrx_ctrl(struct io_ring_ctx *ctx, void __user 
*arg, unsigned nr_args)
                return zcrx_flush_rq(ctx, zcrx, &ctrl);
        case ZCRX_CTRL_EXPORT:
                return zcrx_export(ctx, zcrx, &ctrl, arg);
+       case ZCRX_CTRL_ARM_NOTIFICATION:
+               return zcrx_arm_notif(ctx, zcrx, &ctrl);
        }
 
        return -EOPNOTSUPP;
diff --git a/io_uring/zcrx.h b/io_uring/zcrx.h
index 75e0a4e6ef6e..3ddebed06d57 100644
--- a/io_uring/zcrx.h
+++ b/io_uring/zcrx.h
@@ -9,7 +9,9 @@
 #include <net/net_trackers.h>
 
 #define ZCRX_SUPPORTED_REG_FLAGS       (ZCRX_REG_IMPORT | ZCRX_REG_NODEV)
-#define ZCRX_FEATURES                  (ZCRX_FEATURE_RX_PAGE_SIZE)
+#define ZCRX_FEATURES                  (ZCRX_FEATURE_RX_PAGE_SIZE |\
+                                        ZCRX_FEATURE_NOTIFICATION)
+#define ZCRX_NOTIF_TYPE_MASK           (ZCRX_NOTIF_NO_BUFFERS)
 
 struct io_zcrx_mem {
        unsigned long                   size;
@@ -72,6 +74,13 @@ struct io_zcrx_ifq {
         */
        struct mutex                    pp_lock;
        struct io_mapped_region         rq_region;
+
+       /* Locks the access to notifification context data */
+       spinlock_t                      ctx_lock;
+       struct io_ring_ctx              *master_ctx;
+       u32                             allowed_notif_mask;
+       u32                             fired_notifs;
+       u64                             notif_data;
 };
 
 #if defined(CONFIG_IO_URING_ZCRX)
-- 
2.52.0


Reply via email to