summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMing Lei <ming.lei@redhat.com>2026-03-31 18:31:53 +0300
committerJens Axboe <axboe@kernel.dk>2026-04-07 16:38:36 +0300
commit4d4a512a1f87b156f694d25c800e3d525aa56e8a (patch)
tree6bca1b099bb2abdf1cad3541f6ac705a9ce9e076
parent2fb0ded237bb55dae45bc076666b348fc948ac9e (diff)
downloadlinux-4d4a512a1f87b156f694d25c800e3d525aa56e8a.tar.xz
ublk: add PFN-based buffer matching in I/O path
Add ublk_try_buf_match() which walks a request's bio_vecs, looks up each page's PFN in the per-device maple tree, and verifies all pages belong to the same registered buffer at contiguous offsets. Add ublk_iod_is_shmem_zc() inline helper for checking whether a request uses the shmem zero-copy path. Integrate into the I/O path: - ublk_setup_iod(): if pages match a registered buffer, set UBLK_IO_F_SHMEM_ZC and encode buffer index + offset in addr - ublk_start_io(): skip ublk_map_io() for zero-copy requests - __ublk_complete_rq(): skip ublk_unmap_io() for zero-copy requests The feature remains disabled (ublk_support_shmem_zc() returns false) until the UBLK_F_SHMEM_ZC flag is enabled in the next patch. Signed-off-by: Ming Lei <ming.lei@redhat.com> Link: https://patch.msgid.link/20260331153207.3635125-3-ming.lei@redhat.com Signed-off-by: Jens Axboe <axboe@kernel.dk>
-rw-r--r--drivers/block/ublk_drv.c77
1 files changed, 76 insertions, 1 deletions
diff --git a/drivers/block/ublk_drv.c b/drivers/block/ublk_drv.c
index e6a10a1c8cdb..264b41ceedd8 100644
--- a/drivers/block/ublk_drv.c
+++ b/drivers/block/ublk_drv.c
@@ -356,6 +356,8 @@ struct ublk_params_header {
static void ublk_io_release(void *priv);
static void ublk_stop_dev_unlocked(struct ublk_device *ub);
+static bool ublk_try_buf_match(struct ublk_device *ub, struct request *rq,
+ u32 *buf_idx, u32 *buf_off);
static void ublk_buf_cleanup(struct ublk_device *ub);
static void ublk_abort_queue(struct ublk_device *ub, struct ublk_queue *ubq);
static inline struct request *__ublk_check_and_get_req(struct ublk_device *ub,
@@ -426,6 +428,12 @@ static inline bool ublk_support_shmem_zc(const struct ublk_queue *ubq)
return false;
}
+static inline bool ublk_iod_is_shmem_zc(const struct ublk_queue *ubq,
+ unsigned int tag)
+{
+ return ublk_get_iod(ubq, tag)->op_flags & UBLK_IO_F_SHMEM_ZC;
+}
+
static inline bool ublk_dev_support_shmem_zc(const struct ublk_device *ub)
{
return false;
@@ -1494,6 +1502,18 @@ static blk_status_t ublk_setup_iod(struct ublk_queue *ubq, struct request *req)
iod->nr_sectors = blk_rq_sectors(req);
iod->start_sector = blk_rq_pos(req);
+ /* Try shmem zero-copy match before setting addr */
+ if (ublk_support_shmem_zc(ubq) && ublk_rq_has_data(req)) {
+ u32 buf_idx, buf_off;
+
+ if (ublk_try_buf_match(ubq->dev, req,
+ &buf_idx, &buf_off)) {
+ iod->op_flags |= UBLK_IO_F_SHMEM_ZC;
+ iod->addr = ublk_shmem_zc_addr(buf_idx, buf_off);
+ return BLK_STS_OK;
+ }
+ }
+
iod->addr = io->buf.addr;
return BLK_STS_OK;
@@ -1539,6 +1559,10 @@ static inline void __ublk_complete_rq(struct request *req, struct ublk_io *io,
req_op(req) != REQ_OP_DRV_IN)
goto exit;
+ /* shmem zero copy: no data to unmap, pages already shared */
+ if (ublk_iod_is_shmem_zc(req->mq_hctx->driver_data, req->tag))
+ goto exit;
+
/* for READ request, writing data in iod->addr to rq buffers */
unmapped_bytes = ublk_unmap_io(need_map, req, io);
@@ -1697,8 +1721,13 @@ static void ublk_auto_buf_dispatch(const struct ublk_queue *ubq,
static bool ublk_start_io(const struct ublk_queue *ubq, struct request *req,
struct ublk_io *io)
{
- unsigned mapped_bytes = ublk_map_io(ubq, req, io);
+ unsigned mapped_bytes;
+ /* shmem zero copy: skip data copy, pages already shared */
+ if (ublk_iod_is_shmem_zc(ubq, req->tag))
+ return true;
+
+ mapped_bytes = ublk_map_io(ubq, req, io);
/* partially mapped, update io descriptor */
if (unlikely(mapped_bytes != blk_rq_bytes(req))) {
@@ -5458,7 +5487,53 @@ static void ublk_buf_cleanup(struct ublk_device *ub)
mtree_destroy(&ub->buf_tree);
}
+/* Check if request pages match a registered shared memory buffer */
+static bool ublk_try_buf_match(struct ublk_device *ub,
+ struct request *rq,
+ u32 *buf_idx, u32 *buf_off)
+{
+ struct req_iterator iter;
+ struct bio_vec bv;
+ int index = -1;
+ unsigned long expected_offset = 0;
+ bool first = true;
+
+ rq_for_each_bvec(bv, rq, iter) {
+ unsigned long pfn = page_to_pfn(bv.bv_page);
+ struct ublk_buf_range *range;
+ unsigned long off;
+ range = mtree_load(&ub->buf_tree, pfn);
+ if (!range)
+ return false;
+
+ off = range->base_offset +
+ (pfn - range->base_pfn) * PAGE_SIZE + bv.bv_offset;
+
+ if (first) {
+ /* Read-only buffer can't serve READ (kernel writes) */
+ if ((range->flags & UBLK_SHMEM_BUF_READ_ONLY) &&
+ req_op(rq) != REQ_OP_WRITE)
+ return false;
+ index = range->buf_index;
+ expected_offset = off;
+ *buf_off = off;
+ first = false;
+ } else {
+ if (range->buf_index != index)
+ return false;
+ if (off != expected_offset)
+ return false;
+ }
+ expected_offset += bv.bv_len;
+ }
+
+ if (first)
+ return false;
+
+ *buf_idx = index;
+ return true;
+}
static int ublk_ctrl_uring_cmd_permission(struct ublk_device *ub,
u32 cmd_op, struct ublksrv_ctrl_cmd *header)