On Mon, Aug 17, 2020 at 08:31:53AM +0200, Christoph Hellwig wrote:
> On Mon, Aug 17, 2020 at 12:09:18PM +0800, Baolin Wang wrote:
> >             unsigned int nr_segs)
> >  {
> > @@ -447,7 +425,16 @@ bool __blk_mq_sched_bio_merge(struct request_queue *q, 
> > struct bio *bio,
> >                     !list_empty_careful(&ctx->rq_lists[type])) {
> >             /* default per sw-queue merge */
> >             spin_lock(&ctx->lock);
> > -           ret = blk_mq_attempt_merge(q, hctx, ctx, bio, nr_segs);
> > +           /*
> > +            * Reverse check our software queue for entries that we could
> > +            * potentially merge with. Currently includes a hand-wavy stop
> > +            * count of 8, to not spend too much time checking for merges.
> > +            */
> > +           if (blk_mq_bio_list_merge(q, &ctx->rq_lists[type], bio, 
> > nr_segs)) {
> > +                   ctx->rq_merged++;
> > +                   ret = true;
> > +           }
> > +
> >             spin_unlock(&ctx->lock);
> 
> This adds an overly long line.  That being said the whole thing could
> be nicely simplified to:
> 
>       ...
> 
>       if (e && e->type->ops.bio_merge)
>               return e->type->ops.bio_merge(hctx, bio, nr_segs);
> 
>       if (!(hctx->flags & BLK_MQ_F_SHOULD_MERGE) ||
>           list_empty_careful(&ctx->rq_lists[hctx->type]))
>               return false;
> 
>       /*
>        * Reverse check our software queue for entries that we could
>        * potentially merge with. Currently includes a hand-wavy stop count of
>        * 8, to not spend too much time checking for merges.
>        */
>       spin_lock(&ctx->lock);
>       ret = blk_mq_bio_list_merge(q, &ctx->rq_lists[type], bio, nr_segs);
>       if (ret)
>               ctx->rq_merged++;
>       spin_unlock(&ctx->lock);
> 
> Also I think it would make sense to move the locking into
> blk_mq_bio_list_merge.

Sure, will do in next version.

Reply via email to