Add data path functions for enqueuing and submitting operations to
IOAT devices.

Signed-off-by: Conor Walsh <conor.wa...@intel.com>
Reviewed-by: Kevin Laatz <kevin.la...@intel.com>
Reviewed-by: Chengwen Feng <fengcheng...@huawei.com>
---
 doc/guides/dmadevs/ioat.rst    |  9 ++++
 drivers/dma/ioat/ioat_dmadev.c | 92 ++++++++++++++++++++++++++++++++++
 2 files changed, 101 insertions(+)

diff --git a/doc/guides/dmadevs/ioat.rst b/doc/guides/dmadevs/ioat.rst
index d93d28023f..ec8ce5a8e5 100644
--- a/doc/guides/dmadevs/ioat.rst
+++ b/doc/guides/dmadevs/ioat.rst
@@ -89,3 +89,12 @@ IOAT configuration requirements:
 
 Once configured, the device can then be made ready for use by calling the
 ``rte_dma_start()`` API.
+
+Performing Data Copies
+~~~~~~~~~~~~~~~~~~~~~~~
+
+Refer to the :ref:`Enqueue / Dequeue APIs <dmadev_enqueue_dequeue>` section of 
the dmadev library
+documentation for details on operation enqueue and submission API usage.
+
+It is expected that, for efficiency reasons, a burst of operations will be 
enqueued to the
+device via multiple enqueue calls between calls to the ``rte_dma_submit()`` 
function.
diff --git a/drivers/dma/ioat/ioat_dmadev.c b/drivers/dma/ioat/ioat_dmadev.c
index 96bf55135f..0e92c80fb0 100644
--- a/drivers/dma/ioat/ioat_dmadev.c
+++ b/drivers/dma/ioat/ioat_dmadev.c
@@ -5,6 +5,7 @@
 #include <rte_bus_pci.h>
 #include <rte_dmadev_pmd.h>
 #include <rte_malloc.h>
+#include <rte_prefetch.h>
 
 #include "ioat_internal.h"
 
@@ -17,6 +18,12 @@ RTE_LOG_REGISTER_DEFAULT(ioat_pmd_logtype, INFO);
 #define IOAT_PMD_NAME dmadev_ioat
 #define IOAT_PMD_NAME_STR RTE_STR(IOAT_PMD_NAME)
 
+/* IOAT operations. */
+enum rte_ioat_ops {
+       ioat_op_copy = 0,       /* Standard DMA Operation */
+       ioat_op_fill            /* Block Fill */
+};
+
 /* Configure a device. */
 static int
 ioat_dev_configure(struct rte_dma_dev *dev __rte_unused, const struct 
rte_dma_conf *dev_conf,
@@ -208,6 +215,87 @@ ioat_dev_close(struct rte_dma_dev *dev)
        return 0;
 }
 
+/* Trigger hardware to begin performing enqueued operations. */
+static inline void
+__submit(struct ioat_dmadev *ioat)
+{
+       *ioat->doorbell = ioat->next_write - ioat->offset;
+
+       ioat->last_write = ioat->next_write;
+}
+
+/* External submit function wrapper. */
+static int
+ioat_submit(struct rte_dma_dev *dev, uint16_t qid __rte_unused)
+{
+       struct ioat_dmadev *ioat = (struct ioat_dmadev *)dev->dev_private;
+
+       __submit(ioat);
+
+       return 0;
+}
+
+/* Write descriptor for enqueue. */
+static inline int
+__write_desc(struct rte_dma_dev *dev, uint32_t op, uint64_t src, phys_addr_t 
dst,
+               unsigned int length, uint64_t flags)
+{
+       struct ioat_dmadev *ioat = dev->dev_private;
+       uint16_t ret;
+       const unsigned short mask = ioat->qcfg.nb_desc - 1;
+       const unsigned short read = ioat->next_read;
+       unsigned short write = ioat->next_write;
+       const unsigned short space = mask + read - write;
+       struct ioat_dma_hw_desc *desc;
+
+       if (space == 0)
+               return -ENOSPC;
+
+       ioat->next_write = write + 1;
+       write &= mask;
+
+       desc = &ioat->desc_ring[write];
+       desc->size = length;
+       desc->u.control_raw = (uint32_t)((op << IOAT_CMD_OP_SHIFT) |
+                       (1 << IOAT_COMP_UPDATE_SHIFT));
+
+       /* In IOAT the fence ensures that all operations including the current 
one
+        * are completed before moving on, DMAdev assumes that the fence ensures
+        * all operations before the current one are completed before starting
+        * the current one, so in IOAT we set the fence for the previous 
descriptor.
+        */
+       if (flags & RTE_DMA_OP_FLAG_FENCE)
+               ioat->desc_ring[(write - 1) & mask].u.control.fence = 1;
+
+       desc->src_addr = src;
+       desc->dest_addr = dst;
+
+       rte_prefetch0(&ioat->desc_ring[ioat->next_write & mask]);
+
+       ret = (uint16_t)(ioat->next_write - 1);
+
+       if (flags & RTE_DMA_OP_FLAG_SUBMIT)
+               __submit(ioat);
+
+       return ret;
+}
+
+/* Enqueue a fill operation onto the ioat device. */
+static int
+ioat_enqueue_fill(struct rte_dma_dev *dev, uint16_t qid __rte_unused, uint64_t 
pattern,
+               rte_iova_t dst, unsigned int length, uint64_t flags)
+{
+       return __write_desc(dev, ioat_op_fill, pattern, dst, length, flags);
+}
+
+/* Enqueue a copy operation onto the ioat device. */
+static int
+ioat_enqueue_copy(struct rte_dma_dev *dev, uint16_t qid __rte_unused, 
rte_iova_t src,
+               rte_iova_t dst, unsigned int length, uint64_t flags)
+{
+       return __write_desc(dev, ioat_op_copy, src, dst, length, flags);
+}
+
 /* Dump DMA device info. */
 static int
 ioat_dev_dump(const struct rte_dma_dev *dev, FILE *f)
@@ -303,6 +391,10 @@ ioat_dmadev_create(const char *name, struct rte_pci_device 
*dev)
 
        dmadev->dev_ops = &ioat_dmadev_ops;
 
+       dmadev->copy = ioat_enqueue_copy;
+       dmadev->fill = ioat_enqueue_fill;
+       dmadev->submit = ioat_submit;
+
        ioat = dmadev->data->dev_private;
        ioat->regs = dev->mem_resource[0].addr;
        ioat->doorbell = &ioat->regs->dmacount;
-- 
2.25.1

Reply via email to