On Sat, Sep 23, 2017 at 05:55:37PM +0100, Al Viro wrote:

> IOW, the loop on failure exit should go through the bio, like 
> __bio_unmap_user()
> does.  We *also* need to put everything left unused in pages[], but only from 
> the
> last iteration through iov_for_each().
> 
> Frankly, I would prefer to reuse the pages[], rather than append to it on each
> iteration.  Used iov_iter_get_pages_alloc(), actually.

Something like completely untested diff below, perhaps...

diff --git a/block/bio.c b/block/bio.c
index b38e962fa83e..b5fe23597b41 100644
--- a/block/bio.c
+++ b/block/bio.c
@@ -1323,94 +1323,60 @@ struct bio *bio_map_user_iov(struct request_queue *q,
                             const struct iov_iter *iter,
                             gfp_t gfp_mask)
 {
-       int j;
-       int nr_pages = 0;
-       struct page **pages;
        struct bio *bio;
-       int cur_page = 0;
-       int ret, offset;
+       struct bio_vec *bvec;
        struct iov_iter i;
-       struct iovec iov;
-
-       iov_for_each(iov, i, *iter) {
-               unsigned long uaddr = (unsigned long) iov.iov_base;
-               unsigned long len = iov.iov_len;
-               unsigned long end = (uaddr + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
-               unsigned long start = uaddr >> PAGE_SHIFT;
-
-               /*
-                * Overflow, abort
-                */
-               if (end < start)
-                       return ERR_PTR(-EINVAL);
-
-               nr_pages += end - start;
-               /*
-                * buffer must be aligned to at least logical block size for now
-                */
-               if (uaddr & queue_dma_alignment(q))
-                       return ERR_PTR(-EINVAL);
-       }
+       int ret, j;
 
-       if (!nr_pages)
+       if (!iov_iter_count(iter))
                return ERR_PTR(-EINVAL);
 
-       bio = bio_kmalloc(gfp_mask, nr_pages);
+       bio = bio_kmalloc(gfp_mask, iov_iter_npages(iter, BIO_MAX_PAGES));
        if (!bio)
                return ERR_PTR(-ENOMEM);
 
-       ret = -ENOMEM;
-       pages = kcalloc(nr_pages, sizeof(struct page *), gfp_mask);
-       if (!pages)
-               goto out;
+       i = *iter;
+       while (iov_iter_count(&i)) {
+               struct page **pages;
+               size_t offs;
+               ssize_t bytes;
+               int npages, j;
 
-       iov_for_each(iov, i, *iter) {
-               unsigned long uaddr = (unsigned long) iov.iov_base;
-               unsigned long len = iov.iov_len;
-               unsigned long end = (uaddr + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
-               unsigned long start = uaddr >> PAGE_SHIFT;
-               const int local_nr_pages = end - start;
-               const int page_limit = cur_page + local_nr_pages;
-
-               ret = get_user_pages_fast(uaddr, local_nr_pages,
-                               (iter->type & WRITE) != WRITE,
-                               &pages[cur_page]);
-               if (ret < local_nr_pages) {
-                       ret = -EFAULT;
+               bytes = iov_iter_get_pages_alloc(&i, &pages, LONG_MAX, &offs);
+               if (bytes <= 0) {
+                       ret = bytes ? bytes : -EFAULT;
                        goto out_unmap;
                }
 
-               offset = offset_in_page(uaddr);
-               for (j = cur_page; j < page_limit; j++) {
-                       unsigned int bytes = PAGE_SIZE - offset;
+               npages = DIV_ROUND_UP(offs + bytes, PAGE_SIZE);
 
-                       if (len <= 0)
-                               break;
+               if (unlikely(offs & queue_dma_alignment(q))) {
+                       ret = -EINVAL;
+                       j = 0;
+               } else {
+                       for (j = 0; j < npages; j++) {
+                               struct page *page = pages[j];
+                               unsigned n = PAGE_SIZE - offs;
+                               unsigned prev_bi_vcnt = bio->bi_vcnt;
+
+                               if (!bio_add_pc_page(q, bio, page, n, offs)) {
+                                       iov_iter_truncate(&i, 0);
+                                       break;
+                               }
+
+                               if (bio->bi_vcnt == prev_bi_vcnt)
+                                       put_page(page);
                        
-                       if (bytes > len)
-                               bytes = len;
-
-                       /*
-                        * sorry...
-                        */
-                       if (bio_add_pc_page(q, bio, pages[j], bytes, offset) <
-                                           bytes)
-                               break;
-
-                       len -= bytes;
-                       offset = 0;
+                               iov_iter_advance(&i, n);
+                               bytes -= n;
+                               offs = 0;
+                       }
                }
-
-               cur_page = j;
-               /*
-                * release the pages we didn't map into the bio, if any
-                */
-               while (j < page_limit)
+               while (j < npages)
                        put_page(pages[j++]);
+               kvfree(pages);
        }
 
-       kfree(pages);
-
        bio_set_flag(bio, BIO_USER_MAPPED);
 
        /*
@@ -1423,13 +1389,9 @@ struct bio *bio_map_user_iov(struct request_queue *q,
        return bio;
 
  out_unmap:
-       for (j = 0; j < nr_pages; j++) {
-               if (!pages[j])
-                       break;
-               put_page(pages[j]);
+       bio_for_each_segment_all(bvec, bio, j) {
+               put_page(bvec->bv_page);
        }
- out:
-       kfree(pages);
        bio_put(bio);
        return ERR_PTR(ret);
 }

Reply via email to