On Wed, 07/29 13:02, Paolo Bonzini wrote: > > > On 29/07/2015 12:57, Fam Zheng wrote: > > On Wed, 07/29 09:37, Paolo Bonzini wrote: > >> > >> > >> On 29/07/2015 06:42, Fam Zheng wrote: > >>> @@ -2613,6 +2613,8 @@ bool bdrv_aio_poll(AioContext *ctx, bool blocking) > >>> { > >>> bool ret; > >>> > >>> + aio_disable_clients(ctx, AIO_CLIENT_DATAPLANE | > >>> AIO_CLIENT_NBD_SERVER); > >>> ret = aio_poll(ctx, blocking); > >>> + aio_enable_clients(ctx, AIO_CLIENT_DATAPLANE | > >>> AIO_CLIENT_NBD_SERVER); > >>> return ret; > >> > >> This is not enough, because another thread's aio_poll can sneak in > >> between calls to bdrv_aio_poll if the AioContext lock is released, and > >> they will use the full set of clients. > >> > >> Similar to your v1, it works with the large critical sections we > >> currently have, but it has the same problem in the longer term. > > > > Can we add more disable/enable pairs around bdrv_drain on top? > > Yes, though I think you'd end up reverting patches 10 and 11 in the end.
We will add outer disable/enable pairs to prevent another threads's aio_poll from sneaking in between bdrv_aio_poll calls, but we needn't obsolete bdrv_aio_poll() because of that - it can be useful by itself. For example bdrv_aio_cancel shouldn't look at ioeventfd, otherwise it could spin for too long on high load. Does that make sense? Fam > > All 11 patches are okay, though it would be great if you could post the > full series before this is applied. > > Paolo