On Thu, Nov 20, 2014 at 09:02:24PM +0200, Ciprian Barbu wrote:
> Signed-off-by: Ciprian Barbu <ciprian.ba...@linaro.org>
> ---
> The testcases are based almost entirely on the odp_example.
> There are no alloc tests and I added a test case for odp_schedule_wait_time.
> The major differencs between the odp_example and this cunit is the partition
> into testcases, the odp_example calls every test case from one big function.
> 
> I had to work some magic in order to be able to pass arguments to test cases,
> I hope is not too hard to follow.
> 
>  configure.ac                                  |   1 +
>  test/cunit/Makefile.am                        |   2 +
>  test/cunit/schedule/Makefile.am               |  10 +
>  test/cunit/schedule/odp_schedule_test.c       | 844 
> ++++++++++++++++++++++++++
>  test/cunit/schedule/odp_schedule_testsuites.c |  35 ++
>  test/cunit/schedule/odp_schedule_testsuites.h |  21 +
>  6 files changed, 913 insertions(+)
>  create mode 100644 test/cunit/schedule/Makefile.am
>  create mode 100644 test/cunit/schedule/odp_schedule_test.c
>  create mode 100644 test/cunit/schedule/odp_schedule_testsuites.c
>  create mode 100644 test/cunit/schedule/odp_schedule_testsuites.h
> 
> diff --git a/configure.ac b/configure.ac
> index fcd7279..a47db72 100644
> --- a/configure.ac
> +++ b/configure.ac
> @@ -173,6 +173,7 @@ AC_CONFIG_FILES([Makefile
>                test/Makefile
>                test/api_test/Makefile
>                   test/cunit/Makefile
> +                 test/cunit/schedule/Makefile
>                pkgconfig/libodp.pc])
>  
>  AC_SEARCH_LIBS([timer_create],[rt posix4])
> diff --git a/test/cunit/Makefile.am b/test/cunit/Makefile.am
> index 439e134..b6033ee 100644
> --- a/test/cunit/Makefile.am
> +++ b/test/cunit/Makefile.am
> @@ -3,6 +3,8 @@ include $(top_srcdir)/test/Makefile.inc
>  AM_CFLAGS += -I$(CUNIT_PATH)/include
>  AM_LDFLAGS += -L$(CUNIT_PATH)/lib -static -lcunit
>  
> +SUBDIRS = schedule
> +
>  if ODP_CUNIT_ENABLED
>  TESTS = ${bin_PROGRAMS}
>  check_PROGRAMS = ${bin_PROGRAMS}
> diff --git a/test/cunit/schedule/Makefile.am b/test/cunit/schedule/Makefile.am
> new file mode 100644
> index 0000000..ad68b03
> --- /dev/null
> +++ b/test/cunit/schedule/Makefile.am
> @@ -0,0 +1,10 @@
> +include $(top_srcdir)/test/Makefile.inc
> +
> +if ODP_CUNIT_ENABLED
> +bin_PROGRAMS = odp_schedule_test
> +odp_schedule_test_LDFLAGS = $(AM_LDFLAGS) -L$(CUNIT_PATH)/lib -static -lcunit
> +odp_schedule_test_CFLAGS = $(AM_CFLAGS) -I$(CUNIT_PATH)/include
> +endif
> +
> +dist_odp_schedule_test_SOURCES = odp_schedule_test.c \
> +                              odp_schedule_testsuites.c
> diff --git a/test/cunit/schedule/odp_schedule_test.c 
> b/test/cunit/schedule/odp_schedule_test.c
> new file mode 100644
> index 0000000..fa67f6e
> --- /dev/null
> +++ b/test/cunit/schedule/odp_schedule_test.c
> @@ -0,0 +1,844 @@
> +/* Copyright (c) 2014, Linaro Limited
> + * All rights reserved.
> + *
> + * SPDX-License-Identifier:     BSD-3-Clause
> + */
> +
> +#include "odp_schedule_testsuites.h"
> +#include <odph_linux.h>
> +
> +#define MAX_WORKERS          32            /**< Max worker threads */
> +#define MSG_POOL_SIZE           (4*1024*1024)
> +#define QUEUES_PER_PRIO              64            /**< Queue per priority */
> +#define QUEUE_ROUNDS         (512*1024)    /**< Queue test rounds */
> +#define MULTI_BUFS_MAX               4             /**< Buffer burst size */
> +#define BUF_SIZE             64
> +
> +#define SCHED_MSG "Test_buff_FOR_simple_schedule"
> +
> +/** Test arguments */
> +typedef struct {
> +     int core_count; /**< Core count */
> +     int proc_mode;  /**< Process mode */
> +} test_args_t;
> +
> +typedef int (*test_case_routine)(const char *, int, odp_buffer_pool_t,
> +                              int, odp_barrier_t *);
> +
> +/** Scheduler test case arguments */
> +typedef struct {
> +     char name[64];  /**< test case name */
> +     int prio;
> +     test_case_routine func;
> +} test_case_args_t;
> +
> +/** Test global variables */
> +typedef struct {
> +     odp_barrier_t barrier;/**< @private Barrier for test synchronisation */
> +     test_args_t test_args;/**< @private Test case function and arguments */
> +} test_globals_t;
> +
> +static void execute_parallel(void *(*func) (void *), test_case_args_t *);
> +static int num_workers;
> +
> +/**
> + * @internal CUnit test case for verifying functionality of
> + *           schedule_wait_time
> + */
> +static void schedule_wait_time(void)
> +{
> +     uint64_t wait_time;
> +
> +     wait_time = odp_schedule_wait_time(0);
> +     CU_ASSERT(wait_time > 0);

wait_time can be zero for odp_schedule_wait_time(0).

> +     CU_PASS("schedule_wait_time(0)");
> +
> +     wait_time = odp_schedule_wait_time(1);
> +     CU_ASSERT(wait_time > 0);
> +     CU_PASS("schedule_wait_time(1)");
> +
> +     wait_time = odp_schedule_wait_time((uint64_t)-1LL);
> +     CU_ASSERT(wait_time > 0);
> +     CU_PASS("schedule_wait_time(MAX_LONG_INT)");
> +}
> +
> +/**
> + * @internal Clear all scheduled queues. Retry to be sure that all
> + * buffers have been scheduled.
> + */
> +static void clear_sched_queues(void)
> +{
> +     odp_buffer_t buf;
> +
> +     while (1) {
> +             buf = odp_schedule(NULL, ODP_SCHED_NO_WAIT);
> +
> +             if (buf == ODP_BUFFER_INVALID)
> +                     break;
> +
> +             odp_buffer_free(buf);
> +     }
> +}
> +
> +/**
> + * @internal Create multiple queues from a pool of buffers
> + *
> + * @param thr  Thread
> + * @param msg_pool  Buffer pool
> + * @param prio   Queue priority
> + *
> + * @return 0 if successful
> + */
> +static int create_queues(int thr, odp_buffer_pool_t msg_pool, int prio)
> +{
> +     char name[] = "sched_XX_YY";
> +     odp_buffer_t buf;
> +     odp_queue_t queue;
> +     int i;
> +
> +     name[6] = '0' + prio/10;
> +     name[7] = '0' + prio - 10*(prio/10);
> +
> +     /* Alloc and enqueue a buffer per queue */
> +     for (i = 0; i < QUEUES_PER_PRIO; i++) {
> +             name[9]  = '0' + i/10;
> +             name[10] = '0' + i - 10*(i/10);
> +
> +             queue = odp_queue_lookup(name);
> +
> +             if (queue == ODP_QUEUE_INVALID) {
> +                     ODP_ERR("  [%i] Queue %s lookup failed.\n", thr, name);
> +                     return -1;
> +             }
> +
> +             buf = odp_buffer_alloc(msg_pool);
> +
> +             if (!odp_buffer_is_valid(buf)) {
> +                     ODP_ERR("  [%i] msg_pool alloc failed\n", thr);
> +                     return -1;
> +             }
> +
> +             if (odp_queue_enq(queue, buf)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     return 0;
> +}
> +
> +/**
> + * @internal Create a single queue from a pool of buffers
> + *
> + * @param thr  Thread
> + * @param msg_pool  Buffer pool
> + * @param prio   Queue priority
> + *
> + * @return 0 if successful
> + */
> +static int create_queue(int thr, odp_buffer_pool_t msg_pool, int prio)
> +{
> +     char name[] = "sched_XX_00";
> +     odp_buffer_t buf;
> +     odp_queue_t queue;
> +
> +     buf = odp_buffer_alloc(msg_pool);
> +
> +     if (!odp_buffer_is_valid(buf)) {
> +             ODP_ERR("  [%i] msg_pool alloc failed\n", thr);
> +             return -1;
> +     }
> +
> +     name[6] = '0' + prio/10;
> +     name[7] = '0' + prio - 10*(prio/10);
> +
> +     queue = odp_queue_lookup(name);
> +
> +     if (queue == ODP_QUEUE_INVALID) {
> +             ODP_ERR("  [%i] Queue %s lookup failed.\n", thr, name);
> +             return -1;
> +     }
> +
> +     if (odp_queue_enq(queue, buf)) {
> +             ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +             return -1;
> +     }
> +
> +     return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of a single queue - with odp_schedule_one()
> + *
> + * Enqueue a buffer to the shared queue. Schedule and enqueue the received
> + * buffer back into the queue.
> + *
> + * @param str      Test case name string
> + * @param thr      Thread
> + * @param msg_pool Buffer pool
> + * @param prio     Priority
> + * @param barrier  Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_one_single(const char *str, int thr,
> +                                 odp_buffer_pool_t msg_pool,
> +                                 int prio, odp_barrier_t *barrier)
> +{
> +     odp_buffer_t buf;
> +     odp_queue_t queue;
> +     uint64_t t1, t2, cycles, ns;
> +     uint32_t i;
> +     uint32_t tot = 0;
> +
> +     if (create_queue(thr, msg_pool, prio)) {
> +             CU_FAIL_FATAL("lookup queue");
> +             return -1;
> +     }
> +
> +     t1 = odp_time_get_cycles();
> +
> +     for (i = 0; i < QUEUE_ROUNDS; i++) {
> +             buf = odp_schedule_one(&queue, ODP_SCHED_WAIT);
> +
> +             if (odp_queue_enq(queue, buf)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     if (odp_queue_sched_type(queue) == ODP_SCHED_SYNC_ATOMIC)
> +             odp_schedule_release_atomic();
> +
> +     t2     = odp_time_get_cycles();
> +     cycles = odp_time_diff_cycles(t1, t2);
> +     ns     = odp_time_cycles_to_ns(cycles);
> +     tot    = i;
> +
> +     odp_barrier_sync(barrier);
> +     clear_sched_queues();
> +
> +     cycles = cycles/tot;
> +     ns     = ns/tot;
> +
> +     printf("  [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> +            thr, str, cycles, ns);
> +
> +     return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of multiple queues - with odp_schedule_one()
> + *
> + * Enqueue a buffer to each queue. Schedule and enqueue the received
> + * buffer back into the queue it came from.
> + *
> + * @param str      Test case name string
> + * @param thr      Thread
> + * @param msg_pool Buffer pool
> + * @param prio     Priority
> + * @param barrier  Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_one_many(const char *str, int thr,
> +                               odp_buffer_pool_t msg_pool,
> +                               int prio, odp_barrier_t *barrier)
> +{
> +     odp_buffer_t buf;
> +     odp_queue_t queue;
> +     uint64_t t1 = 0;
> +     uint64_t t2 = 0;
> +     uint64_t cycles, ns;
> +     uint32_t i;
> +     uint32_t tot = 0;
> +
> +     if (create_queues(thr, msg_pool, prio))
> +             return -1;
> +
> +     /* Start sched-enq loop */
> +     t1 = odp_time_get_cycles();
> +
> +     for (i = 0; i < QUEUE_ROUNDS; i++) {
> +             buf = odp_schedule_one(&queue, ODP_SCHED_WAIT);
> +
> +             if (odp_queue_enq(queue, buf)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     if (odp_queue_sched_type(queue) == ODP_SCHED_SYNC_ATOMIC)
> +             odp_schedule_release_atomic();
> +
> +     t2     = odp_time_get_cycles();
> +     cycles = odp_time_diff_cycles(t1, t2);
> +     ns     = odp_time_cycles_to_ns(cycles);
> +     tot    = i;
> +
> +     odp_barrier_sync(barrier);
> +     clear_sched_queues();
> +
> +     cycles = cycles/tot;
> +     ns     = ns/tot;
> +
> +     printf("  [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> +            thr, str, cycles, ns);
> +
> +     return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of a single queue - with odp_schedule()
> + *
> + * Enqueue a buffer to the shared queue. Schedule and enqueue the received
> + * buffer back into the queue.
> + *
> + * @param str      Test case name string
> + * @param thr      Thread
> + * @param msg_pool Buffer pool
> + * @param prio     Priority
> + * @param barrier  Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_single(const char *str, int thr,
> +                             odp_buffer_pool_t msg_pool,
> +                             int prio, odp_barrier_t *barrier)
> +{
> +     odp_buffer_t buf;
> +     odp_queue_t queue;
> +     uint64_t t1, t2, cycles, ns;
> +     uint32_t i;
> +     uint32_t tot = 0;
> +
> +     if (create_queue(thr, msg_pool, prio))
> +             return -1;
> +
> +     t1 = odp_time_get_cycles();
> +
> +     for (i = 0; i < QUEUE_ROUNDS; i++) {
> +             buf = odp_schedule(&queue, ODP_SCHED_WAIT);
> +
> +             if (odp_queue_enq(queue, buf)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     /* Clear possible locally stored buffers */
> +     odp_schedule_pause();
> +
> +     tot = i;
> +
> +     while (1) {
> +             buf = odp_schedule(&queue, ODP_SCHED_NO_WAIT);
> +
> +             if (buf == ODP_BUFFER_INVALID)
> +                     break;
> +
> +             tot++;
> +
> +             if (odp_queue_enq(queue, buf)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     odp_schedule_resume();
> +
> +     t2     = odp_time_get_cycles();
> +     cycles = odp_time_diff_cycles(t1, t2);
> +     ns     = odp_time_cycles_to_ns(cycles);
> +
> +     odp_barrier_sync(barrier);
> +     clear_sched_queues();
> +
> +     cycles = cycles/tot;
> +     ns     = ns/tot;
> +
> +     printf("  [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> +            thr, str, cycles, ns);
> +
> +     return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of multiple queues - with odp_schedule()
> + *
> + * Enqueue a buffer to each queue. Schedule and enqueue the received
> + * buffer back into the queue it came from.
> + *
> + * @param str      Test case name string
> + * @param thr      Thread
> + * @param msg_pool Buffer pool
> + * @param prio     Priority
> + * @param barrier  Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_many(const char *str, int thr,
> +                           odp_buffer_pool_t msg_pool,
> +                           int prio, odp_barrier_t *barrier)
> +{
> +     odp_buffer_t buf;
> +     odp_queue_t queue;
> +     uint64_t t1 = 0;
> +     uint64_t t2 = 0;
> +     uint64_t cycles, ns;
> +     uint32_t i;
> +     uint32_t tot = 0;
> +
> +     if (create_queues(thr, msg_pool, prio))
> +             return -1;
> +
> +     /* Start sched-enq loop */
> +     t1 = odp_time_get_cycles();
> +
> +     for (i = 0; i < QUEUE_ROUNDS; i++) {
> +             buf = odp_schedule(&queue, ODP_SCHED_WAIT);
> +
> +             if (odp_queue_enq(queue, buf)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     /* Clear possible locally stored buffers */
> +     odp_schedule_pause();
> +
> +     tot = i;
> +
> +     while (1) {
> +             buf = odp_schedule(&queue, ODP_SCHED_NO_WAIT);
> +
> +             if (buf == ODP_BUFFER_INVALID)
> +                     break;
> +
> +             tot++;
> +
> +             if (odp_queue_enq(queue, buf)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     odp_schedule_resume();
> +
> +     t2     = odp_time_get_cycles();
> +     cycles = odp_time_diff_cycles(t1, t2);
> +     ns     = odp_time_cycles_to_ns(cycles);
> +
> +     odp_barrier_sync(barrier);
> +     clear_sched_queues();
> +
> +     cycles = cycles/tot;
> +     ns     = ns/tot;
> +
> +     printf("  [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> +            thr, str, cycles, ns);
> +
> +     return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of multiple queues with multi_sched and 
> multi_enq
> + *
> + * @param str      Test case name string
> + * @param thr      Thread
> + * @param msg_pool Buffer pool
> + * @param prio     Priority
> + * @param barrier  Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_multi(const char *str, int thr,
> +                            odp_buffer_pool_t msg_pool,
> +                            int prio, odp_barrier_t *barrier)
> +{
> +     odp_buffer_t buf[MULTI_BUFS_MAX];
> +     odp_queue_t queue;
> +     uint64_t t1 = 0;
> +     uint64_t t2 = 0;
> +     uint64_t cycles, ns;
> +     int i, j;
> +     int num;
> +     uint32_t tot = 0;
> +     char name[] = "sched_XX_YY";
> +
> +     name[6] = '0' + prio/10;
> +     name[7] = '0' + prio - 10*(prio/10);
> +
> +     /* Alloc and enqueue a buffer per queue */
> +     for (i = 0; i < QUEUES_PER_PRIO; i++) {
> +             name[9]  = '0' + i/10;
> +             name[10] = '0' + i - 10*(i/10);
> +
> +             queue = odp_queue_lookup(name);
> +
> +             if (queue == ODP_QUEUE_INVALID) {
> +                     ODP_ERR("  [%i] Queue %s lookup failed.\n", thr, name);
> +                     return -1;
> +             }
> +
> +             for (j = 0; j < MULTI_BUFS_MAX; j++) {
> +                     buf[j] = odp_buffer_alloc(msg_pool);
> +
> +                     if (!odp_buffer_is_valid(buf[j])) {
> +                             ODP_ERR("  [%i] msg_pool alloc failed\n", thr);
> +                             return -1;
> +                     }
> +             }
> +
> +             if (odp_queue_enq_multi(queue, buf, MULTI_BUFS_MAX)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     /* Start sched-enq loop */
> +     t1 = odp_time_get_cycles();
> +
> +     for (i = 0; i < QUEUE_ROUNDS; i++) {
> +             num = odp_schedule_multi(&queue, ODP_SCHED_WAIT, buf,
> +                                      MULTI_BUFS_MAX);
> +
> +             tot += num;
> +
> +             if (odp_queue_enq_multi(queue, buf, num)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     /* Clear possible locally stored buffers */
> +     odp_schedule_pause();
> +
> +     while (1) {
> +             num = odp_schedule_multi(&queue, ODP_SCHED_NO_WAIT, buf,
> +                                      MULTI_BUFS_MAX);
> +
> +             if (num == 0)
> +                     break;
> +
> +             tot += num;
> +
> +             if (odp_queue_enq_multi(queue, buf, num)) {
> +                     ODP_ERR("  [%i] Queue enqueue failed.\n", thr);
> +                     return -1;
> +             }
> +     }
> +
> +     odp_schedule_resume();
> +
> +
> +     t2     = odp_time_get_cycles();
> +     cycles = odp_time_diff_cycles(t1, t2);
> +     ns     = odp_time_cycles_to_ns(cycles);
> +
> +     odp_barrier_sync(barrier);
> +     clear_sched_queues();
> +
> +     if (tot) {
> +             cycles = cycles/tot;
> +             ns     = ns/tot;
> +     } else {
> +             cycles = 0;
> +             ns     = 0;
> +     }
> +
> +     printf("  [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> +            thr, str, cycles, ns);
> +
> +     return 0;
> +}
> +
> +/**
> + * Template function for running the scheduler tests.
> + * The main reason for having this function is that CUnit does not offer a 
> way
> + * to pass arguments to a testcase function.
> + * The other reason is that there are common steps for all testcases.
> + */
> +static void *exec_template(void *arg)
> +{
> +     odp_buffer_pool_t msg_pool;
> +     odp_shm_t shm;
> +     test_globals_t *globals;
> +     odp_barrier_t *barrier;
> +     test_case_args_t *args = (test_case_args_t*) arg;
> +
> +     shm     = odp_shm_lookup("test_globals");
> +     globals = odp_shm_addr(shm);
> +
> +     CU_ASSERT(globals != NULL);
> +
> +     barrier = &globals->barrier;
> +
> +     /*
> +      * Sync before start
> +      */
> +     odp_barrier_sync(barrier);
> +     
> +     /*
> +      * Find the buffer pool
> +      */
> +     msg_pool = odp_buffer_pool_lookup("msg_pool");
> +
> +     CU_ASSERT(msg_pool != ODP_BUFFER_POOL_INVALID);
> +
> +     odp_barrier_sync(barrier);
> +
> +     /*
> +      * Now run the testcase routine passing the arguments
> +      */
> +     args->func(args->name, odp_thread_id(), msg_pool,
> +                args->prio, barrier);
> +
> +     return arg;
> +}
> +
> +/* Low prio */
> +
> +static void schedule_one_single_lo(void)
> +{
> +     test_case_args_t args;
> +     snprintf(args.name, sizeof(args.name), "sched_one_s_lo");
> +     args.prio = ODP_SCHED_PRIO_LOWEST;
> +     args.func = test_schedule_one_single;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_single_lo(void)
> +{
> +     test_case_args_t args;
> +     snprintf(args.name, sizeof(args.name), "sched_____s_lo");
> +     args.prio = ODP_SCHED_PRIO_LOWEST;
> +     args.func = test_schedule_single;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_one_many_lo(void)
> +{
> +     test_case_args_t args;
> +     snprintf(args.name, sizeof(args.name), "sched_one_m_lo");
> +     args.prio = ODP_SCHED_PRIO_LOWEST;
> +     args.func = test_schedule_one_many;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_many_lo(void)
> +{
> +     test_case_args_t args;
> +     snprintf(args.name, sizeof(args.name), "sched_____m_lo");
> +     args.prio = ODP_SCHED_PRIO_LOWEST;
> +     args.func = test_schedule_many;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_multi_lo(void)
> +{
> +     test_case_args_t args;
> +     snprintf(args.name, sizeof(args.name), "sched_____m_lo");
> +     args.prio = ODP_SCHED_PRIO_LOWEST;
> +     args.func = test_schedule_multi;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +/* High prio */
> +
> +static void schedule_one_single_hi(void)
> +{
> +     test_case_args_t args;
> +     snprintf(args.name, sizeof(args.name), "sched_one_s_hi");
> +     args.prio = ODP_SCHED_PRIO_HIGHEST;
> +     args.func = test_schedule_single;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_single_hi(void)
> +{
> +     test_case_args_t args;
> +     snprintf(args.name, sizeof(args.name), "sched_____s_hi");
> +     args.prio = ODP_SCHED_PRIO_HIGHEST;
> +     args.func = test_schedule_single;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_one_many_hi(void)
> +{
> +     test_case_args_t args;
> +     snprintf(args.name, sizeof(args.name), "sched_one_m_hi");
> +     args.prio = ODP_SCHED_PRIO_HIGHEST;
> +     args.func = test_schedule_one_many;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_many_hi(void)
> +{
> +     test_case_args_t args;
> +     snprintf(args.name, sizeof(args.name), "sched_____m_hi");
> +     args.prio = ODP_SCHED_PRIO_HIGHEST;
> +     args.func = test_schedule_many;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_multi_hi(void)
> +{
> +     test_case_args_t args;

IMO, We should avoid using variables allocated from stack to share
the data between different execution context.
It will work in linux threaded run-time environment.But it will
have issues in different run-time environment like bare-metal.
IMO if any memory needs to be shared across different
execution environment should use the memory allocated from odp shared mem.


> +     snprintf(args.name, sizeof(args.name), "sched_multi_hi");
> +     args.prio = ODP_SCHED_PRIO_HIGHEST;
> +     args.func = test_schedule_multi;
> +     execute_parallel(exec_template, &args);
> +}
> +
> +static void execute_parallel(void *(*start_routine) (void *),
> +                          test_case_args_t *test_case_args)
> +{
> +     odph_linux_pthread_t thread_tbl[MAX_WORKERS];
> +     int first_core;
> +
> +     memset(thread_tbl, 0, sizeof(thread_tbl));
> +
> +     /*
> +      * By default core #0 runs Linux kernel background tasks.
> +      * Start mapping thread from core #1
> +      */
> +     first_core = 1;
> +
> +     if (odp_sys_core_count() == 1)
> +             first_core = 0;
> +
> +     odph_linux_pthread_create(thread_tbl, num_workers, first_core,
> +                                     start_routine, test_case_args);
> +
> +     /* Wait for worker threads to terminate */
> +     odph_linux_pthread_join(thread_tbl, num_workers);
> +}
> +
> +static odp_buffer_pool_t test_odp_buffer_pool_init(void)
> +{
> +     void *pool_base;
> +     odp_shm_t shm;
> +     odp_buffer_pool_t pool;
> +
> +     shm = odp_shm_reserve("msg_pool",
> +                           MSG_POOL_SIZE, ODP_CACHE_LINE_SIZE, 0);
> +
> +     pool_base = odp_shm_addr(shm);
> +
> +     if (NULL == pool_base) {
> +             printf("Shared memory reserve failed.\n");
> +             return -1;
> +     }
> +
> +     pool = odp_buffer_pool_create("msg_pool", pool_base, MSG_POOL_SIZE,
> +                                   BUF_SIZE, ODP_CACHE_LINE_SIZE,
> +                                   ODP_BUFFER_TYPE_RAW);
> +
> +     if (ODP_BUFFER_POOL_INVALID == pool) {
> +             printf("Pool create failed.\n");
> +             return -1;
> +     }
> +     return pool;
> +}
> +
> +int schedule_test_init(void)
> +{
> +     test_args_t args;
> +     odp_shm_t shm;
> +     test_globals_t *globals;
> +     int i, j;
> +     int prios;
> +
> +     if (0 != odp_init_global(NULL, NULL)) {
> +             printf("odp_init_global fail.\n");
> +             return -1;
> +     }
> +     if (0 != odp_init_local()) {
> +             printf("odp_init_local fail.\n");
> +             return -1;
> +     }
> +     if (ODP_BUFFER_POOL_INVALID == test_odp_buffer_pool_init()) {
> +             printf("test_odp_buffer_pool_init fail.\n");
> +             return -1;
> +     }
> +
> +     /* A worker thread per core */
> +     num_workers = odp_sys_core_count();
> +
> +     if (args.core_count)

args.core_count is uninitialized

> +             num_workers = args.core_count;
> +
> +     /* force to max core count */
> +     if (num_workers > MAX_WORKERS)
> +             num_workers = MAX_WORKERS;
> +     shm = odp_shm_reserve("test_globals",
> +                           sizeof(test_globals_t), ODP_CACHE_LINE_SIZE, 0);
> +
> +     globals = odp_shm_addr(shm);
> +
> +     if (globals == NULL) {
> +             ODP_ERR("Shared memory reserve failed.\n");
> +             return -1;
> +     }
> +
> +     memset(globals, 0, sizeof(test_globals_t));
> +
> +     /* Barrier to sync test case execution */
> +     odp_barrier_init_count(&globals->barrier, num_workers);
> +
> +     prios = odp_schedule_num_prio();
> +
> +     for (i = 0; i < prios; i++) {
> +             odp_queue_param_t param;
> +             odp_queue_t queue;
> +             char name[] = "sched_XX_YY";
> +
> +             if (i != ODP_SCHED_PRIO_HIGHEST &&
> +                 i != ODP_SCHED_PRIO_LOWEST)
> +                     continue;
> +
> +             name[6] = '0' + i/10;
> +             name[7] = '0' + i - 10*(i/10);
> +
> +             param.sched.prio  = i;
> +             param.sched.sync  = ODP_SCHED_SYNC_ATOMIC;
> +             param.sched.group = ODP_SCHED_GROUP_DEFAULT;
> +
> +             for (j = 0; j < QUEUES_PER_PRIO; j++) {
> +                     name[9]  = '0' + j/10;
> +                     name[10] = '0' + j - 10*(j/10);
> +
> +                     queue = odp_queue_create(name, ODP_QUEUE_TYPE_SCHED,
> +                                              &param);
> +
> +                     if (queue == ODP_QUEUE_INVALID) {
> +                             ODP_ERR("Schedule queue create failed.\n");
> +                             return -1;
> +                     }
> +             }
> +     }
> +     return 0;
> +}
> +
> +int schedule_test_finalize(void)
> +{
> +     odp_term_local();
> +     odp_term_global();
> +     return 0;
> +}
> +
> +struct CU_TestInfo schedule_tests[] = {
> +     _CU_TEST_INFO(schedule_wait_time),
> +     _CU_TEST_INFO(schedule_one_single_lo),
> +     _CU_TEST_INFO(schedule_single_lo),
> +     _CU_TEST_INFO(schedule_one_many_lo),
> +     _CU_TEST_INFO(schedule_many_lo),
> +     _CU_TEST_INFO(schedule_multi_lo),
> +     _CU_TEST_INFO(schedule_one_single_hi),
> +     _CU_TEST_INFO(schedule_single_hi),
> +     _CU_TEST_INFO(schedule_one_many_hi),
> +     _CU_TEST_INFO(schedule_many_hi),
> +     _CU_TEST_INFO(schedule_multi_hi),
> +     CU_TEST_INFO_NULL,
> +};
> diff --git a/test/cunit/schedule/odp_schedule_testsuites.c 
> b/test/cunit/schedule/odp_schedule_testsuites.c
> new file mode 100644
> index 0000000..1053069
> --- /dev/null
> +++ b/test/cunit/schedule/odp_schedule_testsuites.c
> @@ -0,0 +1,35 @@
> +/* Copyright (c) 2014, Linaro Limited
> + * All rights reserved.
> + *
> + * SPDX-License-Identifier:     BSD-3-Clause
> + */
> +
> +#include "odp_schedule_testsuites.h"
> +
> +static CU_SuiteInfo suites[] = {
> +     {
> +             "Scheduler tests" ,
> +             schedule_test_init,
> +             schedule_test_finalize,
> +             NULL,
> +             NULL,
> +             schedule_tests
> +     },
> +     CU_SUITE_INFO_NULL,
> +};
> +
> +int main(void)
> +{
> +     /* initialize the CUnit test registry */
> +     if (CUE_SUCCESS != CU_initialize_registry())
> +             return CU_get_error();
> +
> +     /* register suites */
> +     CU_register_suites(suites);
> +     /* Run all tests using the CUnit Basic interface */
> +     CU_basic_set_mode(CU_BRM_VERBOSE);
> +     CU_basic_run_tests();
> +     CU_cleanup_registry();
> +
> +     return CU_get_error();
> +}
> diff --git a/test/cunit/schedule/odp_schedule_testsuites.h 
> b/test/cunit/schedule/odp_schedule_testsuites.h
> new file mode 100644
> index 0000000..67a2a69
> --- /dev/null
> +++ b/test/cunit/schedule/odp_schedule_testsuites.h
> @@ -0,0 +1,21 @@
> +/* Copyright (c) 2014, Linaro Limited
> + * All rights reserved.
> + *
> + * SPDX-License-Identifier:     BSD-3-Clause
> + */
> +
> +#ifndef ODP_SCHEDULE_TESTSUITES_H_
> +#define ODP_SCHEDULE_TESTSUITES_H_
> +
> +#include "odp.h"
> +#include <CUnit/Basic.h>
> +
> +/* Helper macro for CU_TestInfo initialization */
> +#define _CU_TEST_INFO(test_func) {#test_func, test_func}
> +
> +extern struct CU_TestInfo schedule_tests[];
> +
> +extern int schedule_test_init(void);
> +extern int schedule_test_finalize(void);
> +
> +#endif /* ODP_SCHEDULE_TESTSUITES_H_ */
> -- 
> 1.8.3.2
> 
> 
> _______________________________________________
> lng-odp mailing list
> lng-odp@lists.linaro.org
> http://lists.linaro.org/mailman/listinfo/lng-odp

_______________________________________________
lng-odp mailing list
lng-odp@lists.linaro.org
http://lists.linaro.org/mailman/listinfo/lng-odp

Reply via email to