For each dmadev instance, perform some basic copy tests to validate that functionality.
Signed-off-by: Bruce Richardson <bruce.richard...@intel.com> --- app/test/test_dmadev.c | 174 +++++++++++++++++++++++++++++++++++++++++ 1 file changed, 174 insertions(+) diff --git a/app/test/test_dmadev.c b/app/test/test_dmadev.c index 12f7c69629..261f45db71 100644 --- a/app/test/test_dmadev.c +++ b/app/test/test_dmadev.c @@ -2,12 +2,15 @@ * Copyright(c) 2021 HiSilicon Limited. * Copyright(c) 2021 Intel Corporation. */ +#include <unistd.h> #include <inttypes.h> #include <rte_common.h> #include <rte_dev.h> #include <rte_dmadev.h> #include <rte_bus_vdev.h> +#include <rte_mbuf.h> +#include <rte_random.h> #include "test.h" @@ -16,6 +19,11 @@ extern int test_dmadev_api(uint16_t dev_id); #define PRINT_ERR(...) print_err(__func__, __LINE__, __VA_ARGS__) +#define COPY_LEN 1024 + +static struct rte_mempool *pool; +static uint16_t id_count; + static inline int __rte_format_printf(3, 4) print_err(const char *func, int lineno, const char *format, ...) @@ -31,6 +39,134 @@ print_err(const char *func, int lineno, const char *format, ...) return ret; } +static inline void +await_hw(int dev_id, uint16_t vchan) +{ + int idle = rte_dmadev_vchan_idle(dev_id, vchan); + if (idle < 0) { + /* for drivers that don't support this op, just sleep for 25 microseconds */ + usleep(25); + return; + } + + /* for those that do, *max* end time is one second from now, but all should be faster */ + const uint64_t end_cycles = rte_get_timer_cycles() + rte_get_timer_hz(); + while (!idle && rte_get_timer_cycles() < end_cycles) { + rte_pause(); + idle = rte_dmadev_vchan_idle(dev_id, vchan); + } +} + +static int +test_enqueue_copies(int dev_id, uint16_t vchan) +{ + unsigned int i; + uint16_t id; + + /* test doing a single copy */ + do { + struct rte_mbuf *src, *dst; + char *src_data, *dst_data; + + src = rte_pktmbuf_alloc(pool); + dst = rte_pktmbuf_alloc(pool); + src_data = rte_pktmbuf_mtod(src, char *); + dst_data = rte_pktmbuf_mtod(dst, char *); + + for (i = 0; i < COPY_LEN; i++) + src_data[i] = rte_rand() & 0xFF; + + id = rte_dmadev_copy(dev_id, vchan, src->buf_iova + src->data_off, + dst->buf_iova + dst->data_off, COPY_LEN, RTE_DMA_OP_FLAG_SUBMIT); + if (id != id_count) { + PRINT_ERR("Error with rte_dmadev_copy, got %u, expected %u\n", + id, id_count); + return -1; + } + + /* give time for copy to finish, then check it was done */ + await_hw(dev_id, vchan); + + for (i = 0; i < COPY_LEN; i++) { + if (dst_data[i] != src_data[i]) { + PRINT_ERR("Data mismatch at char %u [Got %02x not %02x]\n", i, + dst_data[i], src_data[i]); + rte_dmadev_dump(dev_id, stderr); + return -1; + } + } + + /* now check completion works */ + if (rte_dmadev_completed(dev_id, vchan, 1, &id, NULL) != 1) { + PRINT_ERR("Error with rte_dmadev_completed\n"); + return -1; + } + if (id != id_count) { + PRINT_ERR("Error:incorrect job id received, %u [expected %u]\n", + id, id_count); + return -1; + } + + rte_pktmbuf_free(src); + rte_pktmbuf_free(dst); + + /* now check completion works */ + if (rte_dmadev_completed(dev_id, 0, 1, NULL, NULL) != 0) { + PRINT_ERR("Error with rte_dmadev_completed in empty check\n"); + return -1; + } + id_count++; + + } while (0); + + /* test doing a multiple single copies */ + do { + const uint16_t max_ops = 4; + struct rte_mbuf *src, *dst; + char *src_data, *dst_data; + + src = rte_pktmbuf_alloc(pool); + dst = rte_pktmbuf_alloc(pool); + src_data = rte_pktmbuf_mtod(src, char *); + dst_data = rte_pktmbuf_mtod(dst, char *); + + for (i = 0; i < COPY_LEN; i++) + src_data[i] = rte_rand() & 0xFF; + + /* perform the same copy <max_ops> times */ + for (i = 0; i < max_ops; i++) { + if (rte_dmadev_copy(dev_id, vchan, + src->buf_iova + src->data_off, + dst->buf_iova + dst->data_off, + COPY_LEN, RTE_DMA_OP_FLAG_SUBMIT) != id_count++) { + PRINT_ERR("Error with rte_dmadev_copy\n"); + return -1; + } + } + await_hw(dev_id, vchan); + + if ((i = rte_dmadev_completed(dev_id, vchan, max_ops * 2, &id, NULL)) != max_ops) { + PRINT_ERR("Error with rte_dmadev_completed, got %u not %u\n", i, max_ops); + return -1; + } + if (id != id_count - 1) { + PRINT_ERR("Error, incorrect job id returned: got %u not %u\n", + id, id_count - 1); + return -1; + } + for (i = 0; i < COPY_LEN; i++) { + if (dst_data[i] != src_data[i]) { + PRINT_ERR("Data mismatch at char %u\n", i); + return -1; + } + } + rte_pktmbuf_free(src); + rte_pktmbuf_free(dst); + } while (0); + + return 0; +} + static int test_dmadev_instance(uint16_t dev_id) { @@ -43,6 +179,7 @@ test_dmadev_instance(uint16_t dev_id) .nb_desc = TEST_RINGSIZE, }; const int vchan = 0; + int i; printf("\n### Test dmadev instance %u\n", dev_id); @@ -79,10 +216,47 @@ test_dmadev_instance(uint16_t dev_id) stats.completed, stats.submitted, stats.errors); return -1; } + id_count = 0; + + /* create a mempool for running tests */ + pool = rte_pktmbuf_pool_create("TEST_DMADEV_POOL", + TEST_RINGSIZE * 2, /* n == num elements */ + 32, /* cache size */ + 0, /* priv size */ + 2048, /* data room size */ + info.device->numa_node); + if (pool == NULL) { + PRINT_ERR("Error with mempool creation\n"); + return -1; + } + + /* run the test cases, use many iterations to ensure UINT16_MAX id wraparound */ + printf("DMA Dev: %u, Running Copy Tests\n", dev_id); + for (i = 0; i < 640; i++) { + + if (test_enqueue_copies(dev_id, vchan) != 0) { + printf("Error with iteration %d\n", i); + rte_dmadev_dump(dev_id, stdout); + goto err; + } + rte_dmadev_stats_get(dev_id, 0, &stats); + printf("Ops submitted: %"PRIu64"\t", stats.submitted); + printf("Ops completed: %"PRIu64"\t", stats.completed); + printf("Errors: %"PRIu64"\r", stats.errors); + } + printf("\n"); + + + rte_mempool_free(pool); rte_dmadev_stop(dev_id); rte_dmadev_stats_reset(dev_id, vchan); return 0; + +err: + rte_mempool_free(pool); + rte_dmadev_stop(dev_id); + return -1; } static int -- 2.30.2