On Fri, Aug 21, 2020 at 03:09:13PM +0200, Philippe Mathieu-Daudé wrote: > On 8/21/20 12:29 PM, Stefano Garzarella wrote: > > On Thu, Aug 20, 2020 at 06:59:01PM +0200, Philippe Mathieu-Daudé wrote: > >> In preparation of using multiple IRQ (thus multiple eventfds) > >> make BDRVNVMeState::irq_notifier an array (for now of a single > >> element, the admin queue notifier). > >> > >> Reviewed-by: Stefan Hajnoczi <stefa...@redhat.com> > >> Signed-off-by: Philippe Mathieu-Daudé <phi...@redhat.com> > >> --- > >> block/nvme.c | 31 +++++++++++++++++++++---------- > >> 1 file changed, 21 insertions(+), 10 deletions(-) > >> > >> diff --git a/block/nvme.c b/block/nvme.c > >> index a61e86a83eb..fe8a40b7ede 100644 > >> --- a/block/nvme.c > >> +++ b/block/nvme.c > >> @@ -106,6 +106,12 @@ QEMU_BUILD_BUG_ON(offsetof(NVMeRegs, doorbells) != > >> 0x1000); > >> #define INDEX_ADMIN 0 > >> #define INDEX_IO(n) (1 + n) > >> > >> +/* This driver shares a single MSIX IRQ for the admin and I/O queues */ > >> +enum { > >> + MSIX_SHARED_IRQ_IDX = 0, > >> + MSIX_IRQ_COUNT = 1 > >> +}; > >> + > >> struct BDRVNVMeState { > >> AioContext *aio_context; > >> QEMUVFIOState *vfio; > >> @@ -120,7 +126,7 @@ struct BDRVNVMeState { > >> /* How many uint32_t elements does each doorbell entry take. */ > >> size_t doorbell_scale; > >> bool write_cache_supported; > >> - EventNotifier irq_notifier; > >> + EventNotifier irq_notifier[MSIX_IRQ_COUNT]; > >> > >> uint64_t nsze; /* Namespace size reported by identify command */ > >> int nsid; /* The namespace id to read/write data. */ > >> @@ -631,7 +637,8 @@ static bool nvme_poll_queues(BDRVNVMeState *s) > >> > >> static void nvme_handle_event(EventNotifier *n) > >> { > >> - BDRVNVMeState *s = container_of(n, BDRVNVMeState, irq_notifier); > >> + BDRVNVMeState *s = container_of(n, BDRVNVMeState, > >> + irq_notifier[MSIX_SHARED_IRQ_IDX]); > >> > >> trace_nvme_handle_event(s); > >> event_notifier_test_and_clear(n); > >> @@ -683,7 +690,8 @@ out_error: > >> static bool nvme_poll_cb(void *opaque) > >> { > >> EventNotifier *e = opaque; > >> - BDRVNVMeState *s = container_of(e, BDRVNVMeState, irq_notifier); > >> + BDRVNVMeState *s = container_of(e, BDRVNVMeState, > >> + irq_notifier[MSIX_SHARED_IRQ_IDX]); > >> > >> trace_nvme_poll_cb(s); > >> return nvme_poll_queues(s); > >> @@ -705,7 +713,7 @@ static int nvme_init(BlockDriverState *bs, const char > >> *device, int namespace, > >> s->device = g_strdup(device); > >> s->nsid = namespace; > >> s->aio_context = bdrv_get_aio_context(bs); > >> - ret = event_notifier_init(&s->irq_notifier, 0); > >> + ret = event_notifier_init(&s->irq_notifier[MSIX_SHARED_IRQ_IDX], 0); > >> if (ret) { > >> error_setg(errp, "Failed to init event notifier"); > >> return ret; > >> @@ -784,12 +792,13 @@ static int nvme_init(BlockDriverState *bs, const > >> char *device, int namespace, > >> } > >> } > >> > >> - ret = qemu_vfio_pci_init_irq(s->vfio, &s->irq_notifier, > >> + ret = qemu_vfio_pci_init_irq(s->vfio, s->irq_notifier, > > > > Maybe we can use '&s->irq_notifier[MSIX_SHARED_IRQ_IDX]' to match the other > > changes. > > This makes the following patch in the next series (using multiple > queues) simpler, but if you prefer I don't mind using your suggestion > here, then adding another patch to directly use the array address > (instead of the address of the 1st element in that array). As you > wish :)
If it simplifies the next patches, it's fine for me ;-) Reviewed-by: Stefano Garzarella <sgarz...@redhat.com> > > > > >> VFIO_PCI_MSIX_IRQ_INDEX, errp); > >> if (ret) { > >> goto out; > >> } > >> - aio_set_event_notifier(bdrv_get_aio_context(bs), &s->irq_notifier, > >> + aio_set_event_notifier(bdrv_get_aio_context(bs), > >> + &s->irq_notifier[MSIX_SHARED_IRQ_IDX], > >> false, nvme_handle_event, nvme_poll_cb); > >> > >> nvme_identify(bs, namespace, &local_err); > >> @@ -872,9 +881,10 @@ static void nvme_close(BlockDriverState *bs) > >> nvme_free_queue_pair(s->queues[i]); > >> } > >> g_free(s->queues); > >> - aio_set_event_notifier(bdrv_get_aio_context(bs), &s->irq_notifier, > >> + aio_set_event_notifier(bdrv_get_aio_context(bs), > >> + &s->irq_notifier[MSIX_SHARED_IRQ_IDX], > >> false, NULL, NULL); > >> - event_notifier_cleanup(&s->irq_notifier); > >> + event_notifier_cleanup(&s->irq_notifier[MSIX_SHARED_IRQ_IDX]); > >> qemu_vfio_pci_unmap_bar(s->vfio, 0, (void *)s->regs, 0, > >> NVME_BAR_SIZE); > >> qemu_vfio_close(s->vfio); > >> > >> @@ -1381,7 +1391,8 @@ static void nvme_detach_aio_context(BlockDriverState > >> *bs) > >> q->completion_bh = NULL; > >> } > >> > >> - aio_set_event_notifier(bdrv_get_aio_context(bs), &s->irq_notifier, > >> + aio_set_event_notifier(bdrv_get_aio_context(bs), > >> + &s->irq_notifier[MSIX_SHARED_IRQ_IDX], > >> false, NULL, NULL); > >> } > >> > >> @@ -1391,7 +1402,7 @@ static void nvme_attach_aio_context(BlockDriverState > >> *bs, > >> BDRVNVMeState *s = bs->opaque; > >> > >> s->aio_context = new_context; > >> - aio_set_event_notifier(new_context, &s->irq_notifier, > >> + aio_set_event_notifier(new_context, > >> &s->irq_notifier[MSIX_SHARED_IRQ_IDX], > >> false, nvme_handle_event, nvme_poll_cb); > >> > >> for (int i = 0; i < s->nr_queues; i++) { > >> -- > >> 2.26.2 > >> > >> > > >