On Mon, Aug 26, 2019 at 09:24:03AM -0700, Bart Van Assche wrote:
> On 8/25/19 7:51 PM, Ming Lei wrote:
> > diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c
> > index 5b0b5224cfd4..5941a0176f87 100644
> > --- a/block/blk-sysfs.c
> > +++ b/block/blk-sysfs.c
> > @@ -938,6 +938,7 @@ int blk_register_queue(struct gendisk *disk)
> >     int ret;
> >     struct device *dev = disk_to_dev(disk);
> >     struct request_queue *q = disk->queue;
> > +   bool has_elevator = false;
> >     if (WARN_ON(!q))
> >             return -ENXIO;
> > @@ -945,7 +946,6 @@ int blk_register_queue(struct gendisk *disk)
> >     WARN_ONCE(blk_queue_registered(q),
> >               "%s is registering an already registered queue\n",
> >               kobject_name(&dev->kobj));
> > -   blk_queue_flag_set(QUEUE_FLAG_REGISTERED, q);
> >     /*
> >      * SCSI probing may synchronously create and destroy a lot of
> > @@ -966,7 +966,7 @@ int blk_register_queue(struct gendisk *disk)
> >             return ret;
> >     /* Prevent changes through sysfs until registration is completed. */
> > -   mutex_lock(&q->sysfs_lock);
> > +   mutex_lock(&q->sysfs_dir_lock);
> 
> Does mutex_lock(&q->sysfs_dir_lock) really protect against changes of the
> I/O scheduler through sysfs or does it only protect against concurrent sysfs
> object creation and removal?

It is only for protecting against concurrent sysfs object creation and removal.

> In other words, should the comment above this
> mutex lock call be updated?

Yeah, it should be removed.

> 
> > @@ -987,26 +987,37 @@ int blk_register_queue(struct gendisk *disk)
> >             blk_mq_debugfs_register(q);
> >     }
> > -   kobject_uevent(&q->kobj, KOBJ_ADD);
> > -
> > -   wbt_enable_default(q);
> > -
> > -   blk_throtl_register_queue(q);
> > -
> > +   /*
> > +    * The queue's kobject ADD uevent isn't sent out, also the
> > +    * flag of QUEUE_FLAG_REGISTERED isn't set yet, so elevator
> > +    * switch won't happen at all.
> > +    */
> >     if (q->elevator) {
> > -           ret = elv_register_queue(q);
> > +           ret = elv_register_queue(q, false);
> >             if (ret) {
> > -                   mutex_unlock(&q->sysfs_lock);
> > -                   kobject_uevent(&q->kobj, KOBJ_REMOVE);
> > +                   mutex_unlock(&q->sysfs_dir_lock);
> >                     kobject_del(&q->kobj);
> >                     blk_trace_remove_sysfs(dev);
> >                     kobject_put(&dev->kobj);
> >                     return ret;
> >             }
> > +           has_elevator = true;
> >     }
> 
> I think the reference to the kobject ADD event in the comment is misleading.
> If e.g. a request queue is registered, unregistered and reregistered
> quickly, can it happen that a udev rule for the ADD event triggered by the
> first registration is executed in the middle of the second registration? Is

It should happen, but this patch doesn't change anything about this
behavior.

> setting the REGISTERED flag later sufficient to fix the race against
> scheduler changes through sysfs?

Yes, it is enough. 

> If so, how about leaving out the reference
> to the kobject ADD event from the above comment?

OK.

> 
> > +   mutex_lock(&q->sysfs_lock);
> > +   blk_queue_flag_set(QUEUE_FLAG_REGISTERED, q);
> > +   wbt_enable_default(q);
> > +   blk_throtl_register_queue(q);
> > +   mutex_unlock(&q->sysfs_lock);
> > +
> > +   /* Now everything is ready and send out KOBJ_ADD uevent */
> > +   kobject_uevent(&q->kobj, KOBJ_ADD);
> > +   if (has_elevator)
> > +           kobject_uevent(&q->elevator->kobj, KOBJ_ADD);
> 
> Can it happen that immediately after mutex_unlock(&q->sysfs_lock) a script
> removes the I/O scheduler and hence makes the value of the 'has_elevator'
> variable stale? In other words, should emitting KOBJ_ADD also be protected
> by sysfs_lock?

Good catch, it could be fine to hold syfs_lock for emitting KOBJ_ADD.

> 
> > @@ -1021,6 +1032,7 @@ EXPORT_SYMBOL_GPL(blk_register_queue);
> >   void blk_unregister_queue(struct gendisk *disk)
> >   {
> >     struct request_queue *q = disk->queue;
> > +   bool has_elevator;
> >     if (WARN_ON(!q))
> >             return;
> > @@ -1035,25 +1047,25 @@ void blk_unregister_queue(struct gendisk *disk)
> >      * concurrent elv_iosched_store() calls.
> >      */
> >     mutex_lock(&q->sysfs_lock);
> > -
> >     blk_queue_flag_clear(QUEUE_FLAG_REGISTERED, q);
> > +   has_elevator = !!q->elevator;
> > +   mutex_unlock(&q->sysfs_lock);
> > +   mutex_lock(&q->sysfs_dir_lock);
> >     /*
> >      * Remove the sysfs attributes before unregistering the queue data
> >      * structures that can be modified through sysfs.
> >      */
> >     if (queue_is_mq(q))
> >             blk_mq_unregister_dev(disk_to_dev(disk), q);
> > -   mutex_unlock(&q->sysfs_lock);
> >     kobject_uevent(&q->kobj, KOBJ_REMOVE);
> >     kobject_del(&q->kobj);
> >     blk_trace_remove_sysfs(disk_to_dev(disk));
> > -   mutex_lock(&q->sysfs_lock);
> > -   if (q->elevator)
> > +   if (has_elevator)
> >             elv_unregister_queue(q);
> > -   mutex_unlock(&q->sysfs_lock);
> > +   mutex_unlock(&q->sysfs_dir_lock);
> 
> Is it safe to call elv_unregister_queue() if no I/O scheduler is associated
> with a request queue?

No, q->elevator has to be valid for elv_unregister_queue().

>If so, have you considered to leave out the
> 'has_elevator' variable from this function?
> 
> > @@ -567,10 +580,23 @@ int elevator_switch_mq(struct request_queue *q,
> >     lockdep_assert_held(&q->sysfs_lock);
> >     if (q->elevator) {
> > -           if (q->elevator->registered)
> > +           if (q->elevator->registered) {
> > +                   mutex_unlock(&q->sysfs_lock);
> > +
> >                     elv_unregister_queue(q);
> > +
> > +                   mutex_lock(&q->sysfs_lock);
> > +           }
> >             ioc_clear_queue(q);
> >             elevator_exit(q, q->elevator);
> > +
> > +           /*
> > +            * sysfs_lock may be dropped, so re-check if queue is
> > +            * unregistered. If yes, don't switch to new elevator
> > +            * any more
> > +            */
> > +           if (!blk_queue_registered(q))
> > +                   return 0;
> >     }
> 
> So elevator_switch_mq() is called with sysfs_lock held and releases and
> reacquires that mutex?

Yes.

> What will happen if e.g. syzbot writes into
> /sys/block/*/queue/scheduler from multiple threads concurrently? Can that

It can't happen, sysfs's write on same file is always exclusively protected
by one mutex, see kernfs_fop_write(), and should be same for normal fs too.

> lead to multiple concurrent calls of elv_register_queue() and
> elv_unregister_queue()? Can that e.g. cause concurrent calls of the
> following code from elv_register_queue(): kobject_add(&e->kobj, &q->kobj,
> "%s", "iosched")?

No, it won't happen.

> 
> Is it even possible to fix this lock inversion by introducing only one new
> mutex? I think the sysfs directories and attributes referenced by this patch
> are as follows:
> 
> /sys/block/<q>/queue
> /sys/block/<q>/queue/attr
> /sys/block/<q>/queue/iosched/attr
> /sys/block/<q>/mq
> /sys/block/<q>/mq/<n>
> /sys/block/<q>/mq/<n>/attr
> 
> Isn't the traditional approach to protect such a hierarchy to use one mutex
> per level? E.g. one mutex to serialize "queue" and "mq" manipulations
> (sysfs_dir_lock?), one mutex to protect the queue/attr attributes
> (sysfs_lock?), one mutex to serialize kobj creation in the mq directory, one
> mutex to protect the mq/<n>/attr attributes and one mutex to protect the I/O
> scheduler attributes?

This patch keeps to use sysfs_lock for protecting attributes show/write,
meantime don't use it for serializing kobj creation & removal, so far
looks good.

I will address your above comments and post V4 for further review.

Thanks,
Ming

Reply via email to