From: Daeho Jeong <daehoje...@google.com>

Now decompression is being handled in workqueue and it makes read I/O
latency non-deterministic, because of the non-deterministic scheduling
nature of workqueues. So, I made it handled in softirq context only if
possible.

Signed-off-by: Daeho Jeong <daehoje...@google.com>
---
 fs/f2fs/compress.c | 145 +++++++++++++++++++++++++--------------------
 fs/f2fs/data.c     |  50 ++++++++++------
 fs/f2fs/f2fs.h     |  10 ++--
 3 files changed, 119 insertions(+), 86 deletions(-)

diff --git a/fs/f2fs/compress.c b/fs/f2fs/compress.c
index 24824cd96f36..8bc4cd0390dc 100644
--- a/fs/f2fs/compress.c
+++ b/fs/f2fs/compress.c
@@ -729,14 +729,13 @@ static int f2fs_compress_pages(struct compress_ctx *cc)
        return ret;
 }
 
-void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
+void f2fs_decompress_cluster(struct decompress_io_ctx *dic, bool in_softirq)
 {
        struct f2fs_sb_info *sbi = F2FS_I_SB(dic->inode);
        struct f2fs_inode_info *fi = F2FS_I(dic->inode);
        const struct f2fs_compress_ops *cops =
                        f2fs_cops[fi->i_compress_algorithm];
        int ret;
-       int i;
 
        trace_f2fs_decompress_pages_start(dic->inode, dic->cluster_idx,
                                dic->cluster_size, fi->i_compress_algorithm);
@@ -746,49 +745,12 @@ void f2fs_decompress_cluster(struct decompress_io_ctx 
*dic)
                goto out_end_io;
        }
 
-       dic->tpages = page_array_alloc(dic->inode, dic->cluster_size);
-       if (!dic->tpages) {
-               ret = -ENOMEM;
-               goto out_end_io;
-       }
-
-       for (i = 0; i < dic->cluster_size; i++) {
-               if (dic->rpages[i]) {
-                       dic->tpages[i] = dic->rpages[i];
-                       continue;
-               }
-
-               dic->tpages[i] = f2fs_compress_alloc_page();
-               if (!dic->tpages[i]) {
-                       ret = -ENOMEM;
-                       goto out_end_io;
-               }
-       }
-
-       if (cops->init_decompress_ctx) {
-               ret = cops->init_decompress_ctx(dic);
-               if (ret)
-                       goto out_end_io;
-       }
-
-       dic->rbuf = f2fs_vmap(dic->tpages, dic->cluster_size);
-       if (!dic->rbuf) {
-               ret = -ENOMEM;
-               goto out_destroy_decompress_ctx;
-       }
-
-       dic->cbuf = f2fs_vmap(dic->cpages, dic->nr_cpages);
-       if (!dic->cbuf) {
-               ret = -ENOMEM;
-               goto out_vunmap_rbuf;
-       }
-
        dic->clen = le32_to_cpu(dic->cbuf->clen);
        dic->rlen = PAGE_SIZE << dic->log_cluster_size;
 
        if (dic->clen > PAGE_SIZE * dic->nr_cpages - COMPRESS_HEADER_SIZE) {
                ret = -EFSCORRUPTED;
-               goto out_vunmap_cbuf;
+               goto out_end_io;
        }
 
        ret = cops->decompress_pages(dic);
@@ -809,17 +771,10 @@ void f2fs_decompress_cluster(struct decompress_io_ctx 
*dic)
                }
        }
 
-out_vunmap_cbuf:
-       vm_unmap_ram(dic->cbuf, dic->nr_cpages);
-out_vunmap_rbuf:
-       vm_unmap_ram(dic->rbuf, dic->cluster_size);
-out_destroy_decompress_ctx:
-       if (cops->destroy_decompress_ctx)
-               cops->destroy_decompress_ctx(dic);
 out_end_io:
        trace_f2fs_decompress_pages_end(dic->inode, dic->cluster_idx,
                                                        dic->clen, ret);
-       f2fs_decompress_end_io(dic, ret);
+       f2fs_decompress_end_io(dic, ret, in_softirq);
 }
 
 /*
@@ -829,7 +784,7 @@ void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
  * (or in the case of a failure, cleans up without actually decompressing).
  */
 void f2fs_end_read_compressed_page(struct page *page, bool failed,
-                                               block_t blkaddr)
+               block_t blkaddr, bool in_softirq)
 {
        struct decompress_io_ctx *dic =
                        (struct decompress_io_ctx *)page_private(page);
@@ -839,12 +794,12 @@ void f2fs_end_read_compressed_page(struct page *page, 
bool failed,
 
        if (failed)
                WRITE_ONCE(dic->failed, true);
-       else if (blkaddr)
+       else if (blkaddr && !in_softirq)
                f2fs_cache_compressed_page(sbi, page,
                                        dic->inode->i_ino, blkaddr);
 
        if (atomic_dec_and_test(&dic->remaining_pages))
-               f2fs_decompress_cluster(dic);
+               f2fs_decompress_cluster(dic, in_softirq);
 }
 
 static bool is_page_in_cluster(struct compress_ctx *cc, pgoff_t index)
@@ -1552,12 +1507,14 @@ int f2fs_write_multi_pages(struct compress_ctx *cc,
        return err;
 }
 
-static void f2fs_free_dic(struct decompress_io_ctx *dic);
+static void f2fs_free_dic(struct decompress_io_ctx *dic,
+               bool bypass_destroy_callback);
 
 struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc)
 {
        struct decompress_io_ctx *dic;
        pgoff_t start_idx = start_idx_of_cluster(cc);
+       const struct f2fs_compress_ops *cops;
        int i;
 
        dic = f2fs_kmem_cache_alloc(dic_entry_slab, GFP_F2FS_ZERO,
@@ -1602,17 +1559,60 @@ struct decompress_io_ctx *f2fs_alloc_dic(struct 
compress_ctx *cc)
                dic->cpages[i] = page;
        }
 
+       dic->tpages = page_array_alloc(dic->inode, dic->cluster_size);
+       if (!dic->tpages)
+               goto out_free;
+
+       for (i = 0; i < dic->cluster_size; i++) {
+               if (dic->rpages[i]) {
+                       dic->tpages[i] = dic->rpages[i];
+                       continue;
+               }
+
+               dic->tpages[i] = f2fs_compress_alloc_page();
+               if (!dic->tpages[i])
+                       goto out_free;
+       }
+
+       dic->rbuf = f2fs_vmap(dic->tpages, dic->cluster_size);
+       if (!dic->rbuf)
+               goto out_free;
+
+       dic->cbuf = f2fs_vmap(dic->cpages, dic->nr_cpages);
+       if (!dic->cbuf)
+               goto out_free;
+
+       cops = f2fs_cops[F2FS_I(dic->inode)->i_compress_algorithm];
+       if (cops->init_decompress_ctx) {
+               int ret = cops->init_decompress_ctx(dic);
+
+               if (ret)
+                       goto out_free;
+       }
+
        return dic;
 
 out_free:
-       f2fs_free_dic(dic);
+       f2fs_free_dic(dic, true);
        return ERR_PTR(-ENOMEM);
 }
 
-static void f2fs_free_dic(struct decompress_io_ctx *dic)
+static void f2fs_free_dic(struct decompress_io_ctx *dic,
+               bool bypass_destroy_callback)
 {
+       const struct f2fs_compress_ops *cops =
+               f2fs_cops[F2FS_I(dic->inode)->i_compress_algorithm];
        int i;
 
+       if (!bypass_destroy_callback && cops->destroy_decompress_ctx)
+               cops->destroy_decompress_ctx(dic);
+
+       if (dic->cbuf)
+               vm_unmap_ram(dic->cbuf, dic->nr_cpages);
+
+       if (dic->rbuf)
+               vm_unmap_ram(dic->rbuf, dic->cluster_size);
+
        if (dic->tpages) {
                for (i = 0; i < dic->cluster_size; i++) {
                        if (dic->rpages[i])
@@ -1637,17 +1637,33 @@ static void f2fs_free_dic(struct decompress_io_ctx *dic)
        kmem_cache_free(dic_entry_slab, dic);
 }
 
-static void f2fs_put_dic(struct decompress_io_ctx *dic)
+static void f2fs_late_free_dic(struct work_struct *work)
 {
-       if (refcount_dec_and_test(&dic->refcnt))
-               f2fs_free_dic(dic);
+       struct decompress_io_ctx *dic =
+               container_of(work, struct decompress_io_ctx, free_work);
+
+       f2fs_free_dic(dic, false);
+}
+
+static void f2fs_put_dic(struct decompress_io_ctx *dic, bool in_softirq)
+{
+       if (refcount_dec_and_test(&dic->refcnt)) {
+               if (in_softirq) {
+                       INIT_WORK(&dic->free_work, f2fs_late_free_dic);
+                       queue_work(F2FS_I_SB(dic->inode)->post_read_wq,
+                                       &dic->free_work);
+               } else {
+                       f2fs_free_dic(dic, false);
+               }
+       }
 }
 
 /*
  * Update and unlock the cluster's pagecache pages, and release the reference 
to
  * the decompress_io_ctx that was being held for I/O completion.
  */
-static void __f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool 
failed)
+static void __f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool 
failed,
+                               bool in_softirq)
 {
        int i;
 
@@ -1668,7 +1684,7 @@ static void __f2fs_decompress_end_io(struct 
decompress_io_ctx *dic, bool failed)
                unlock_page(rpage);
        }
 
-       f2fs_put_dic(dic);
+       f2fs_put_dic(dic, in_softirq);
 }
 
 static void f2fs_verify_cluster(struct work_struct *work)
@@ -1685,14 +1701,15 @@ static void f2fs_verify_cluster(struct work_struct 
*work)
                        SetPageError(rpage);
        }
 
-       __f2fs_decompress_end_io(dic, false);
+       __f2fs_decompress_end_io(dic, false, false);
 }
 
 /*
  * This is called when a compressed cluster has been decompressed
  * (or failed to be read and/or decompressed).
  */
-void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed)
+void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed,
+                               bool in_softirq)
 {
        if (!failed && dic->need_verity) {
                /*
@@ -1704,7 +1721,7 @@ void f2fs_decompress_end_io(struct decompress_io_ctx 
*dic, bool failed)
                INIT_WORK(&dic->verity_work, f2fs_verify_cluster);
                fsverity_enqueue_verify_work(&dic->verity_work);
        } else {
-               __f2fs_decompress_end_io(dic, failed);
+               __f2fs_decompress_end_io(dic, failed, in_softirq);
        }
 }
 
@@ -1713,12 +1730,12 @@ void f2fs_decompress_end_io(struct decompress_io_ctx 
*dic, bool failed)
  *
  * This is called when the page is no longer needed and can be freed.
  */
-void f2fs_put_page_dic(struct page *page)
+void f2fs_put_page_dic(struct page *page, bool in_softirq)
 {
        struct decompress_io_ctx *dic =
                        (struct decompress_io_ctx *)page_private(page);
 
-       f2fs_put_dic(dic);
+       f2fs_put_dic(dic, in_softirq);
 }
 
 /*
diff --git a/fs/f2fs/data.c b/fs/f2fs/data.c
index 7fcbcf979737..87455350c9e4 100644
--- a/fs/f2fs/data.c
+++ b/fs/f2fs/data.c
@@ -119,7 +119,7 @@ struct bio_post_read_ctx {
        block_t fs_blkaddr;
 };
 
-static void f2fs_finish_read_bio(struct bio *bio)
+static void f2fs_finish_read_bio(struct bio *bio, bool in_softirq)
 {
        struct bio_vec *bv;
        struct bvec_iter_all iter_all;
@@ -133,8 +133,9 @@ static void f2fs_finish_read_bio(struct bio *bio)
 
                if (f2fs_is_compressed_page(page)) {
                        if (bio->bi_status)
-                               f2fs_end_read_compressed_page(page, true, 0);
-                       f2fs_put_page_dic(page);
+                               f2fs_end_read_compressed_page(page, true, 0,
+                                                       in_softirq);
+                       f2fs_put_page_dic(page, in_softirq);
                        continue;
                }
 
@@ -191,7 +192,7 @@ static void f2fs_verify_bio(struct work_struct *work)
                fsverity_verify_bio(bio);
        }
 
-       f2fs_finish_read_bio(bio);
+       f2fs_finish_read_bio(bio, false);
 }
 
 /*
@@ -203,7 +204,7 @@ static void f2fs_verify_bio(struct work_struct *work)
  * can involve reading verity metadata pages from the file, and these verity
  * metadata pages may be encrypted and/or compressed.
  */
-static void f2fs_verify_and_finish_bio(struct bio *bio)
+static void f2fs_verify_and_finish_bio(struct bio *bio, bool in_softirq)
 {
        struct bio_post_read_ctx *ctx = bio->bi_private;
 
@@ -211,7 +212,7 @@ static void f2fs_verify_and_finish_bio(struct bio *bio)
                INIT_WORK(&ctx->work, f2fs_verify_bio);
                fsverity_enqueue_verify_work(&ctx->work);
        } else {
-               f2fs_finish_read_bio(bio);
+               f2fs_finish_read_bio(bio, in_softirq);
        }
 }
 
@@ -224,7 +225,8 @@ static void f2fs_verify_and_finish_bio(struct bio *bio)
  * that the bio includes at least one compressed page.  The actual 
decompression
  * is done on a per-cluster basis, not a per-bio basis.
  */
-static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx)
+static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx,
+               bool in_softirq)
 {
        struct bio_vec *bv;
        struct bvec_iter_all iter_all;
@@ -237,7 +239,7 @@ static void f2fs_handle_step_decompress(struct 
bio_post_read_ctx *ctx)
                /* PG_error was set if decryption failed. */
                if (f2fs_is_compressed_page(page))
                        f2fs_end_read_compressed_page(page, PageError(page),
-                                               blkaddr);
+                                               blkaddr, in_softirq);
                else
                        all_compressed = false;
 
@@ -262,9 +264,9 @@ static void f2fs_post_read_work(struct work_struct *work)
                fscrypt_decrypt_bio(ctx->bio);
 
        if (ctx->enabled_steps & STEP_DECOMPRESS)
-               f2fs_handle_step_decompress(ctx);
+               f2fs_handle_step_decompress(ctx, false);
 
-       f2fs_verify_and_finish_bio(ctx->bio);
+       f2fs_verify_and_finish_bio(ctx->bio, false);
 }
 
 static void f2fs_read_end_io(struct bio *bio)
@@ -281,16 +283,28 @@ static void f2fs_read_end_io(struct bio *bio)
        }
 
        if (bio->bi_status) {
-               f2fs_finish_read_bio(bio);
+               f2fs_finish_read_bio(bio, true);
                return;
        }
 
-       if (ctx && (ctx->enabled_steps & (STEP_DECRYPT | STEP_DECOMPRESS))) {
-               INIT_WORK(&ctx->work, f2fs_post_read_work);
-               queue_work(ctx->sbi->post_read_wq, &ctx->work);
-       } else {
-               f2fs_verify_and_finish_bio(bio);
+       if (ctx) {
+               unsigned int enabled_steps = ctx->enabled_steps &
+                                       (STEP_DECRYPT | STEP_DECOMPRESS);
+
+               /*
+                * If we have only decompression step between decompression and
+                * decrypt, we don't need post processing for this.
+                */
+               if (enabled_steps == STEP_DECOMPRESS) {
+                       f2fs_handle_step_decompress(ctx, true);
+               } else if (enabled_steps) {
+                       INIT_WORK(&ctx->work, f2fs_post_read_work);
+                       queue_work(ctx->sbi->post_read_wq, &ctx->work);
+                       return;
+               }
        }
+
+       f2fs_verify_and_finish_bio(bio, true);
 }
 
 static void f2fs_write_end_io(struct bio *bio)
@@ -2222,7 +2236,7 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct 
bio **bio_ret,
 
                if (f2fs_load_compressed_page(sbi, page, blkaddr)) {
                        if (atomic_dec_and_test(&dic->remaining_pages))
-                               f2fs_decompress_cluster(dic);
+                               f2fs_decompress_cluster(dic, false);
                        continue;
                }
 
@@ -2240,7 +2254,7 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct 
bio **bio_ret,
                                        page->index, for_write);
                        if (IS_ERR(bio)) {
                                ret = PTR_ERR(bio);
-                               f2fs_decompress_end_io(dic, ret);
+                               f2fs_decompress_end_io(dic, ret, false);
                                f2fs_put_dnode(&dn);
                                *bio_ret = NULL;
                                return ret;
diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
index d9bbecd008d2..be118d332536 100644
--- a/fs/f2fs/f2fs.h
+++ b/fs/f2fs/f2fs.h
@@ -1580,6 +1580,7 @@ struct decompress_io_ctx {
        void *private;                  /* payload buffer for specified 
decompression algorithm */
        void *private2;                 /* extra payload buffer */
        struct work_struct verity_work; /* work to verify the decompressed 
pages */
+       struct work_struct free_work;   /* work for late free this structure 
itself */
 };
 
 #define NULL_CLUSTER                   ((unsigned int)(~0))
@@ -4158,9 +4159,9 @@ void f2fs_compress_write_end_io(struct bio *bio, struct 
page *page);
 bool f2fs_is_compress_backend_ready(struct inode *inode);
 int f2fs_init_compress_mempool(void);
 void f2fs_destroy_compress_mempool(void);
-void f2fs_decompress_cluster(struct decompress_io_ctx *dic);
+void f2fs_decompress_cluster(struct decompress_io_ctx *dic, bool in_softirq);
 void f2fs_end_read_compressed_page(struct page *page, bool failed,
-                                                       block_t blkaddr);
+                               block_t blkaddr, bool in_softirq);
 bool f2fs_cluster_is_empty(struct compress_ctx *cc);
 bool f2fs_cluster_can_merge_page(struct compress_ctx *cc, pgoff_t index);
 bool f2fs_all_cluster_page_loaded(struct compress_ctx *cc, struct pagevec 
*pvec,
@@ -4179,8 +4180,9 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct 
bio **bio_ret,
                                unsigned nr_pages, sector_t *last_block_in_bio,
                                bool is_readahead, bool for_write);
 struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc);
-void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed);
-void f2fs_put_page_dic(struct page *page);
+void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed,
+                               bool in_softirq);
+void f2fs_put_page_dic(struct page *page, bool in_softirq);
 unsigned int f2fs_cluster_blocks_are_contiguous(struct dnode_of_data *dn);
 int f2fs_init_compress_ctx(struct compress_ctx *cc);
 void f2fs_destroy_compress_ctx(struct compress_ctx *cc, bool reuse);
-- 
2.36.1.476.g0c4daa206d-goog



_______________________________________________
Linux-f2fs-devel mailing list
Linux-f2fs-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel

Reply via email to