io_uring/rsrc: don't use blk_rq_nr_phys_segments() as number of bvecs

io_buffer_register_bvec() currently uses blk_rq_nr_phys_segments() as
the number of bvecs in the request. However, bvecs may be split into
multiple segments depending on the queue limits. Thus, the number of
segments may overestimate the number of bvecs. For ublk devices, the
only current users of io_buffer_register_bvec(), virt_boundary_mask,
seg_boundary_mask, max_segments, and max_segment_size can all be set
arbitrarily by the ublk server process.
Set imu->nr_bvecs based on the number of bvecs the rq_for_each_bvec()
loop actually yields. However, continue using blk_rq_nr_phys_segments()
as an upper bound on the number of bvecs when allocating imu to avoid
needing to iterate the bvecs a second time.

Link: https://lore.kernel.org/io-uring/20251111191530.1268875-1-csander@purestorage.com/
Signed-off-by: Caleb Sander Mateos <csander@purestorage.com>
Fixes: 27cb27b6d5 ("io_uring: add support for kernel registered bvecs")
Reviewed-by: Ming Lei <ming.lei@redhat.com>
Reviewed-by: Chaitanya Kulkarni <kch@nvidia.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
Caleb Sander Mateos 2025-11-11 12:15:29 -07:00 committed by Jens Axboe
parent 6a77267d97
commit 2d0e88f3fd

View file

@ -943,8 +943,8 @@ int io_buffer_register_bvec(struct io_uring_cmd *cmd, struct request *rq,
struct req_iterator rq_iter; struct req_iterator rq_iter;
struct io_mapped_ubuf *imu; struct io_mapped_ubuf *imu;
struct io_rsrc_node *node; struct io_rsrc_node *node;
struct bio_vec bv, *bvec; struct bio_vec bv;
u16 nr_bvecs; unsigned int nr_bvecs = 0;
int ret = 0; int ret = 0;
io_ring_submit_lock(ctx, issue_flags); io_ring_submit_lock(ctx, issue_flags);
@ -965,8 +965,11 @@ int io_buffer_register_bvec(struct io_uring_cmd *cmd, struct request *rq,
goto unlock; goto unlock;
} }
nr_bvecs = blk_rq_nr_phys_segments(rq); /*
imu = io_alloc_imu(ctx, nr_bvecs); * blk_rq_nr_phys_segments() may overestimate the number of bvecs
* but avoids needing to iterate over the bvecs
*/
imu = io_alloc_imu(ctx, blk_rq_nr_phys_segments(rq));
if (!imu) { if (!imu) {
kfree(node); kfree(node);
ret = -ENOMEM; ret = -ENOMEM;
@ -977,16 +980,15 @@ int io_buffer_register_bvec(struct io_uring_cmd *cmd, struct request *rq,
imu->len = blk_rq_bytes(rq); imu->len = blk_rq_bytes(rq);
imu->acct_pages = 0; imu->acct_pages = 0;
imu->folio_shift = PAGE_SHIFT; imu->folio_shift = PAGE_SHIFT;
imu->nr_bvecs = nr_bvecs;
refcount_set(&imu->refs, 1); refcount_set(&imu->refs, 1);
imu->release = release; imu->release = release;
imu->priv = rq; imu->priv = rq;
imu->is_kbuf = true; imu->is_kbuf = true;
imu->dir = 1 << rq_data_dir(rq); imu->dir = 1 << rq_data_dir(rq);
bvec = imu->bvec;
rq_for_each_bvec(bv, rq, rq_iter) rq_for_each_bvec(bv, rq, rq_iter)
*bvec++ = bv; imu->bvec[nr_bvecs++] = bv;
imu->nr_bvecs = nr_bvecs;
node->buf = imu; node->buf = imu;
data->nodes[index] = node; data->nodes[index] = node;