Construct bio_vec from folios, then call the other file's r/w callbacks for IO operations. Test data shows direct I/O copy_file_range improves performance by over 50% vs direct I/O mmap&read (2557 vs 1534).
Test data: | 32x32MB Read 1024MB |Creat-ms|Close-ms| I/O-ms|I/O-MB/s| I/O% |-------------------------|--------|--------|--------|--------|----- | 1)Beg udmabuf buffer R/W| 580 | 323 | 1238 | 867 | 100% | 2) dmabuf buffer R/W| 48 | 5 | 1149 | 934 | 107% | 3) udma+memfd buffer R/W| 597 | 340 | 2157 | 497 | 57% | 4) udma+memfd direct R/W| 573 | 340 | 700 | 1534 | 176% | 5) u+mfd buffer sendfile| 577 | 340 | 1204 | 891 | 102% | 6) u+mfd direct sendfile| 567 | 339 | 2272 | 472 | 54% | 7) u+mfd buffer splice| 570 | 337 | 1114 | 964 | 111% | 8) u+mfd direct splice| 564 | 335 | 793 | 1355 | 156% | 9) udmabuf buffer c_f_r| 577 | 323 | 1059 | 1014 | 116% |10) udmabuf direct c_f_r| 582 | 325 | 420 | 2557 | 294% |11)End udmabuf buffer R/W| 586 | 323 | 1188 | 903 | 104% Signed-off-by: wangtao <tao.wang...@honor.com> --- drivers/dma-buf/udmabuf.c | 59 +++++++++++++++++++++++++++++++++++++++ 1 file changed, 59 insertions(+) diff --git a/drivers/dma-buf/udmabuf.c b/drivers/dma-buf/udmabuf.c index e74e36a8ecda..573275a51674 100644 --- a/drivers/dma-buf/udmabuf.c +++ b/drivers/dma-buf/udmabuf.c @@ -284,6 +284,64 @@ static int end_cpu_udmabuf(struct dma_buf *buf, return 0; } +static ssize_t udmabuf_rw_file(struct dma_buf *dmabuf, loff_t my_pos, + struct file *other, loff_t pos, + size_t count, bool is_write) +{ + struct udmabuf *ubuf = dmabuf->priv; + loff_t my_end = my_pos + count, bv_beg, bv_end = 0; + pgoff_t pg_idx = my_pos / PAGE_SIZE; + pgoff_t pg_end = DIV_ROUND_UP(my_end, PAGE_SIZE); + size_t i, bv_off, bv_len, bv_num, bv_idx = 0, bv_total = 0; + struct bio_vec *bvec; + struct kiocb kiocb; + struct iov_iter iter; + unsigned int direction = is_write ? ITER_SOURCE : ITER_DEST; + ssize_t ret = 0, rw_total = 0; + struct folio *folio; + + bv_num = min_t(size_t, pg_end - pg_idx + 1, 1024); + bvec = kvcalloc(bv_num, sizeof(*bvec), GFP_KERNEL); + if (!bvec) + return -ENOMEM; + + init_sync_kiocb(&kiocb, other); + kiocb.ki_pos = pos; + + for (i = 0; i < ubuf->nr_pinned && my_pos < my_end; i++) { + folio = ubuf->pinned_folios[i]; + bv_beg = bv_end; + bv_end += folio_size(folio); + if (bv_end <= my_pos) + continue; + + bv_len = min(bv_end, my_end) - my_pos; + bv_off = my_pos - bv_beg; + my_pos += bv_len; + bv_total += bv_len; + bvec_set_page(&bvec[bv_idx], &folio->page, bv_len, bv_off); + if (++bv_idx < bv_num && my_pos < my_end) + continue; + + /* start R/W if bvec is full or count reaches zero. */ + iov_iter_bvec(&iter, direction, bvec, bv_idx, bv_total); + if (is_write) + ret = other->f_op->write_iter(&kiocb, &iter); + else + ret = other->f_op->read_iter(&kiocb, &iter); + if (ret <= 0) + break; + rw_total += ret; + if (ret < bv_total || fatal_signal_pending(current)) + break; + + bv_idx = bv_total = 0; + } + kvfree(bvec); + + return rw_total > 0 ? rw_total : ret; +} + static const struct dma_buf_ops udmabuf_ops = { .cache_sgt_mapping = true, .map_dma_buf = map_udmabuf, @@ -294,6 +352,7 @@ static const struct dma_buf_ops udmabuf_ops = { .vunmap = vunmap_udmabuf, .begin_cpu_access = begin_cpu_udmabuf, .end_cpu_access = end_cpu_udmabuf, + .rw_file = udmabuf_rw_file, }; #define SEALS_WANTED (F_SEAL_SHRINK) -- 2.17.1