this patch adds a new aio readv compatible function which copies all data through a bounce buffer. The benefit is that these requests can be flagged as canceled to avoid guest memory corruption when a canceled request is completed by the backend at a later stage.
If an IDE protocol wants to use this function it has to pipe all read requests through ide_readv_cancelable and it may then enable requests_cancelable in the IDEState. If this state is enable we can avoid the blocking blk_drain_all in case of a BMDMA reset. Currently only read operations are cancelable thus we can only use this logic for read-only devices. Signed-off-by: Peter Lieven <p...@kamp.de> --- hw/ide/core.c | 54 ++++++++++++++++++++++++++++++++++++++++++++++++++++++ hw/ide/internal.h | 16 ++++++++++++++++ hw/ide/pci.c | 42 ++++++++++++++++++++++++++++-------------- 3 files changed, 98 insertions(+), 14 deletions(-) diff --git a/hw/ide/core.c b/hw/ide/core.c index 317406d..24547ce 100644 --- a/hw/ide/core.c +++ b/hw/ide/core.c @@ -561,6 +561,59 @@ static bool ide_sect_range_ok(IDEState *s, return true; } +static void ide_readv_cancelable_cb(void *opaque, int ret) +{ + IDECancelableRequest *req = opaque; + if (!req->canceled) { + if (!ret) { + qemu_iovec_from_buf(req->org_qiov, 0, req->buf, req->org_qiov->size); + } + req->org_cb(req->org_opaque, ret); + } + QLIST_REMOVE(req, list); + qemu_vfree(req->buf); + qemu_iovec_destroy(&req->qiov); + g_free(req); +} + +#define MAX_CANCELABLE_REQS 16 + +BlockAIOCB *ide_readv_cancelable(IDEState *s, int64_t sector_num, + QEMUIOVector *iov, int nb_sectors, + BlockCompletionFunc *cb, void *opaque) +{ + BlockAIOCB *aioreq; + IDECancelableRequest *req; + int c = 0; + + QLIST_FOREACH(req, &s->cancelable_requests, list) { + c++; + } + if (c > MAX_CANCELABLE_REQS) { + return NULL; + } + + req = g_new0(IDECancelableRequest, 1); + qemu_iovec_init(&req->qiov, 1); + req->buf = qemu_blockalign(blk_bs(s->blk), iov->size); + qemu_iovec_add(&req->qiov, req->buf, iov->size); + req->org_qiov = iov; + req->org_cb = cb; + req->org_opaque = opaque; + + aioreq = blk_aio_readv(s->blk, sector_num, &req->qiov, nb_sectors, + ide_readv_cancelable_cb, req); + if (aioreq == NULL) { + qemu_vfree(req->buf); + qemu_iovec_destroy(&req->qiov); + g_free(req); + } else { + QLIST_INSERT_HEAD(&s->cancelable_requests, req, list); + } + + return aioreq; +} + static void ide_sector_read(IDEState *s); static void ide_sector_read_cb(void *opaque, int ret) @@ -805,6 +858,7 @@ void ide_start_dma(IDEState *s, BlockCompletionFunc *cb) s->bus->retry_unit = s->unit; s->bus->retry_sector_num = ide_get_sector(s); s->bus->retry_nsector = s->nsector; + s->bus->s = s; if (s->bus->dma->ops->start_dma) { s->bus->dma->ops->start_dma(s->bus->dma, s, cb); } diff --git a/hw/ide/internal.h b/hw/ide/internal.h index 05e93ff..ad188c2 100644 --- a/hw/ide/internal.h +++ b/hw/ide/internal.h @@ -343,6 +343,16 @@ enum ide_dma_cmd { #define ide_cmd_is_read(s) \ ((s)->dma_cmd == IDE_DMA_READ) +typedef struct IDECancelableRequest { + QLIST_ENTRY(IDECancelableRequest) list; + QEMUIOVector qiov; + uint8_t *buf; + QEMUIOVector *org_qiov; + BlockCompletionFunc *org_cb; + void *org_opaque; + bool canceled; +} IDECancelableRequest; + /* NOTE: IDEState represents in fact one drive */ struct IDEState { IDEBus *bus; @@ -396,6 +406,8 @@ struct IDEState { BlockAIOCB *pio_aiocb; struct iovec iov; QEMUIOVector qiov; + QLIST_HEAD(, IDECancelableRequest) cancelable_requests; + bool requests_cancelable; /* ATA DMA state */ int32_t io_buffer_offset; int32_t io_buffer_size; @@ -468,6 +480,7 @@ struct IDEBus { uint8_t retry_unit; int64_t retry_sector_num; uint32_t retry_nsector; + IDEState *s; }; #define TYPE_IDE_DEVICE "ide-device" @@ -572,6 +585,9 @@ void ide_set_inactive(IDEState *s, bool more); BlockAIOCB *ide_issue_trim(BlockBackend *blk, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockCompletionFunc *cb, void *opaque); +BlockAIOCB *ide_readv_cancelable(IDEState *s, int64_t sector_num, + QEMUIOVector *iov, int nb_sectors, + BlockCompletionFunc *cb, void *opaque); /* hw/ide/atapi.c */ void ide_atapi_cmd(IDEState *s); diff --git a/hw/ide/pci.c b/hw/ide/pci.c index d31ff88..5587183 100644 --- a/hw/ide/pci.c +++ b/hw/ide/pci.c @@ -240,21 +240,35 @@ void bmdma_cmd_writeb(BMDMAState *bm, uint32_t val) /* Ignore writes to SSBM if it keeps the old value */ if ((val & BM_CMD_START) != (bm->cmd & BM_CMD_START)) { if (!(val & BM_CMD_START)) { - /* - * We can't cancel Scatter Gather DMA in the middle of the - * operation or a partial (not full) DMA transfer would reach - * the storage so we wait for completion instead (we beahve - * like if the DMA was completed by the time the guest trying - * to cancel dma with bmdma_cmd_writeb with BM_CMD_START not - * set). - * - * In the future we'll be able to safely cancel the I/O if the - * whole DMA operation will be submitted to disk with a single - * aio operation with preadv/pwritev. - */ if (bm->bus->dma->aiocb) { - blk_drain_all(); - assert(bm->bus->dma->aiocb == NULL); + if (bm->bus->s && bm->bus->s->requests_cancelable) { + /* + * If the used IDE protocol supports request cancelation we + * can flag requests as canceled here and disable DMA. + * The IDE protocol used MUST use ide_readv_cancelable for all + * read operations and then subsequently can enable this code + * path. Currently this is only supported for read-only + * devices. + */ + IDECancelableRequest *req; + QLIST_FOREACH(req, &bm->bus->s->cancelable_requests, list) { + if (!req->canceled) { + req->org_cb(req->org_opaque, -ECANCELED); + } + req->canceled = true; + } + } else { + /* + * We can't cancel Scatter Gather DMA in the middle of the + * operation or a partial (not full) DMA transfer would reach + * the storage so we wait for completion instead (we beahve + * like if the DMA was completed by the time the guest trying + * to cancel dma with bmdma_cmd_writeb with BM_CMD_START not + * set). + */ + blk_drain_all(); + assert(bm->bus->dma->aiocb == NULL); + } } bm->status &= ~BM_STATUS_DMAING; } else { -- 1.9.1