On Thu, Nov 20, 2014 at 09:02:24PM +0200, Ciprian Barbu wrote: > Signed-off-by: Ciprian Barbu <ciprian.ba...@linaro.org> > --- > The testcases are based almost entirely on the odp_example. > There are no alloc tests and I added a test case for odp_schedule_wait_time. > The major differencs between the odp_example and this cunit is the partition > into testcases, the odp_example calls every test case from one big function. > > I had to work some magic in order to be able to pass arguments to test cases, > I hope is not too hard to follow. > > configure.ac | 1 + > test/cunit/Makefile.am | 2 + > test/cunit/schedule/Makefile.am | 10 + > test/cunit/schedule/odp_schedule_test.c | 844 > ++++++++++++++++++++++++++ > test/cunit/schedule/odp_schedule_testsuites.c | 35 ++ > test/cunit/schedule/odp_schedule_testsuites.h | 21 + > 6 files changed, 913 insertions(+) > create mode 100644 test/cunit/schedule/Makefile.am > create mode 100644 test/cunit/schedule/odp_schedule_test.c > create mode 100644 test/cunit/schedule/odp_schedule_testsuites.c > create mode 100644 test/cunit/schedule/odp_schedule_testsuites.h > > diff --git a/configure.ac b/configure.ac > index fcd7279..a47db72 100644 > --- a/configure.ac > +++ b/configure.ac > @@ -173,6 +173,7 @@ AC_CONFIG_FILES([Makefile > test/Makefile > test/api_test/Makefile > test/cunit/Makefile > + test/cunit/schedule/Makefile > pkgconfig/libodp.pc]) > > AC_SEARCH_LIBS([timer_create],[rt posix4]) > diff --git a/test/cunit/Makefile.am b/test/cunit/Makefile.am > index 439e134..b6033ee 100644 > --- a/test/cunit/Makefile.am > +++ b/test/cunit/Makefile.am > @@ -3,6 +3,8 @@ include $(top_srcdir)/test/Makefile.inc > AM_CFLAGS += -I$(CUNIT_PATH)/include > AM_LDFLAGS += -L$(CUNIT_PATH)/lib -static -lcunit > > +SUBDIRS = schedule > + > if ODP_CUNIT_ENABLED > TESTS = ${bin_PROGRAMS} > check_PROGRAMS = ${bin_PROGRAMS} > diff --git a/test/cunit/schedule/Makefile.am b/test/cunit/schedule/Makefile.am > new file mode 100644 > index 0000000..ad68b03 > --- /dev/null > +++ b/test/cunit/schedule/Makefile.am > @@ -0,0 +1,10 @@ > +include $(top_srcdir)/test/Makefile.inc > + > +if ODP_CUNIT_ENABLED > +bin_PROGRAMS = odp_schedule_test > +odp_schedule_test_LDFLAGS = $(AM_LDFLAGS) -L$(CUNIT_PATH)/lib -static -lcunit > +odp_schedule_test_CFLAGS = $(AM_CFLAGS) -I$(CUNIT_PATH)/include > +endif > + > +dist_odp_schedule_test_SOURCES = odp_schedule_test.c \ > + odp_schedule_testsuites.c > diff --git a/test/cunit/schedule/odp_schedule_test.c > b/test/cunit/schedule/odp_schedule_test.c > new file mode 100644 > index 0000000..fa67f6e > --- /dev/null > +++ b/test/cunit/schedule/odp_schedule_test.c > @@ -0,0 +1,844 @@ > +/* Copyright (c) 2014, Linaro Limited > + * All rights reserved. > + * > + * SPDX-License-Identifier: BSD-3-Clause > + */ > + > +#include "odp_schedule_testsuites.h" > +#include <odph_linux.h> > + > +#define MAX_WORKERS 32 /**< Max worker threads */ > +#define MSG_POOL_SIZE (4*1024*1024) > +#define QUEUES_PER_PRIO 64 /**< Queue per priority */ > +#define QUEUE_ROUNDS (512*1024) /**< Queue test rounds */ > +#define MULTI_BUFS_MAX 4 /**< Buffer burst size */ > +#define BUF_SIZE 64 > + > +#define SCHED_MSG "Test_buff_FOR_simple_schedule" > + > +/** Test arguments */ > +typedef struct { > + int core_count; /**< Core count */ > + int proc_mode; /**< Process mode */ > +} test_args_t; > + > +typedef int (*test_case_routine)(const char *, int, odp_buffer_pool_t, > + int, odp_barrier_t *); > + > +/** Scheduler test case arguments */ > +typedef struct { > + char name[64]; /**< test case name */ > + int prio; > + test_case_routine func; > +} test_case_args_t; > + > +/** Test global variables */ > +typedef struct { > + odp_barrier_t barrier;/**< @private Barrier for test synchronisation */ > + test_args_t test_args;/**< @private Test case function and arguments */ > +} test_globals_t; > + > +static void execute_parallel(void *(*func) (void *), test_case_args_t *); > +static int num_workers; > + > +/** > + * @internal CUnit test case for verifying functionality of > + * schedule_wait_time > + */ > +static void schedule_wait_time(void) > +{ > + uint64_t wait_time; > + > + wait_time = odp_schedule_wait_time(0); > + CU_ASSERT(wait_time > 0);
wait_time can be zero for odp_schedule_wait_time(0). > + CU_PASS("schedule_wait_time(0)"); > + > + wait_time = odp_schedule_wait_time(1); > + CU_ASSERT(wait_time > 0); > + CU_PASS("schedule_wait_time(1)"); > + > + wait_time = odp_schedule_wait_time((uint64_t)-1LL); > + CU_ASSERT(wait_time > 0); > + CU_PASS("schedule_wait_time(MAX_LONG_INT)"); > +} > + > +/** > + * @internal Clear all scheduled queues. Retry to be sure that all > + * buffers have been scheduled. > + */ > +static void clear_sched_queues(void) > +{ > + odp_buffer_t buf; > + > + while (1) { > + buf = odp_schedule(NULL, ODP_SCHED_NO_WAIT); > + > + if (buf == ODP_BUFFER_INVALID) > + break; > + > + odp_buffer_free(buf); > + } > +} > + > +/** > + * @internal Create multiple queues from a pool of buffers > + * > + * @param thr Thread > + * @param msg_pool Buffer pool > + * @param prio Queue priority > + * > + * @return 0 if successful > + */ > +static int create_queues(int thr, odp_buffer_pool_t msg_pool, int prio) > +{ > + char name[] = "sched_XX_YY"; > + odp_buffer_t buf; > + odp_queue_t queue; > + int i; > + > + name[6] = '0' + prio/10; > + name[7] = '0' + prio - 10*(prio/10); > + > + /* Alloc and enqueue a buffer per queue */ > + for (i = 0; i < QUEUES_PER_PRIO; i++) { > + name[9] = '0' + i/10; > + name[10] = '0' + i - 10*(i/10); > + > + queue = odp_queue_lookup(name); > + > + if (queue == ODP_QUEUE_INVALID) { > + ODP_ERR(" [%i] Queue %s lookup failed.\n", thr, name); > + return -1; > + } > + > + buf = odp_buffer_alloc(msg_pool); > + > + if (!odp_buffer_is_valid(buf)) { > + ODP_ERR(" [%i] msg_pool alloc failed\n", thr); > + return -1; > + } > + > + if (odp_queue_enq(queue, buf)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + return 0; > +} > + > +/** > + * @internal Create a single queue from a pool of buffers > + * > + * @param thr Thread > + * @param msg_pool Buffer pool > + * @param prio Queue priority > + * > + * @return 0 if successful > + */ > +static int create_queue(int thr, odp_buffer_pool_t msg_pool, int prio) > +{ > + char name[] = "sched_XX_00"; > + odp_buffer_t buf; > + odp_queue_t queue; > + > + buf = odp_buffer_alloc(msg_pool); > + > + if (!odp_buffer_is_valid(buf)) { > + ODP_ERR(" [%i] msg_pool alloc failed\n", thr); > + return -1; > + } > + > + name[6] = '0' + prio/10; > + name[7] = '0' + prio - 10*(prio/10); > + > + queue = odp_queue_lookup(name); > + > + if (queue == ODP_QUEUE_INVALID) { > + ODP_ERR(" [%i] Queue %s lookup failed.\n", thr, name); > + return -1; > + } > + > + if (odp_queue_enq(queue, buf)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + > + return 0; > +} > + > +/** > + * @internal Test scheduling of a single queue - with odp_schedule_one() > + * > + * Enqueue a buffer to the shared queue. Schedule and enqueue the received > + * buffer back into the queue. > + * > + * @param str Test case name string > + * @param thr Thread > + * @param msg_pool Buffer pool > + * @param prio Priority > + * @param barrier Barrier > + * > + * @return 0 if successful > + */ > +static int test_schedule_one_single(const char *str, int thr, > + odp_buffer_pool_t msg_pool, > + int prio, odp_barrier_t *barrier) > +{ > + odp_buffer_t buf; > + odp_queue_t queue; > + uint64_t t1, t2, cycles, ns; > + uint32_t i; > + uint32_t tot = 0; > + > + if (create_queue(thr, msg_pool, prio)) { > + CU_FAIL_FATAL("lookup queue"); > + return -1; > + } > + > + t1 = odp_time_get_cycles(); > + > + for (i = 0; i < QUEUE_ROUNDS; i++) { > + buf = odp_schedule_one(&queue, ODP_SCHED_WAIT); > + > + if (odp_queue_enq(queue, buf)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + if (odp_queue_sched_type(queue) == ODP_SCHED_SYNC_ATOMIC) > + odp_schedule_release_atomic(); > + > + t2 = odp_time_get_cycles(); > + cycles = odp_time_diff_cycles(t1, t2); > + ns = odp_time_cycles_to_ns(cycles); > + tot = i; > + > + odp_barrier_sync(barrier); > + clear_sched_queues(); > + > + cycles = cycles/tot; > + ns = ns/tot; > + > + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n", > + thr, str, cycles, ns); > + > + return 0; > +} > + > +/** > + * @internal Test scheduling of multiple queues - with odp_schedule_one() > + * > + * Enqueue a buffer to each queue. Schedule and enqueue the received > + * buffer back into the queue it came from. > + * > + * @param str Test case name string > + * @param thr Thread > + * @param msg_pool Buffer pool > + * @param prio Priority > + * @param barrier Barrier > + * > + * @return 0 if successful > + */ > +static int test_schedule_one_many(const char *str, int thr, > + odp_buffer_pool_t msg_pool, > + int prio, odp_barrier_t *barrier) > +{ > + odp_buffer_t buf; > + odp_queue_t queue; > + uint64_t t1 = 0; > + uint64_t t2 = 0; > + uint64_t cycles, ns; > + uint32_t i; > + uint32_t tot = 0; > + > + if (create_queues(thr, msg_pool, prio)) > + return -1; > + > + /* Start sched-enq loop */ > + t1 = odp_time_get_cycles(); > + > + for (i = 0; i < QUEUE_ROUNDS; i++) { > + buf = odp_schedule_one(&queue, ODP_SCHED_WAIT); > + > + if (odp_queue_enq(queue, buf)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + if (odp_queue_sched_type(queue) == ODP_SCHED_SYNC_ATOMIC) > + odp_schedule_release_atomic(); > + > + t2 = odp_time_get_cycles(); > + cycles = odp_time_diff_cycles(t1, t2); > + ns = odp_time_cycles_to_ns(cycles); > + tot = i; > + > + odp_barrier_sync(barrier); > + clear_sched_queues(); > + > + cycles = cycles/tot; > + ns = ns/tot; > + > + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n", > + thr, str, cycles, ns); > + > + return 0; > +} > + > +/** > + * @internal Test scheduling of a single queue - with odp_schedule() > + * > + * Enqueue a buffer to the shared queue. Schedule and enqueue the received > + * buffer back into the queue. > + * > + * @param str Test case name string > + * @param thr Thread > + * @param msg_pool Buffer pool > + * @param prio Priority > + * @param barrier Barrier > + * > + * @return 0 if successful > + */ > +static int test_schedule_single(const char *str, int thr, > + odp_buffer_pool_t msg_pool, > + int prio, odp_barrier_t *barrier) > +{ > + odp_buffer_t buf; > + odp_queue_t queue; > + uint64_t t1, t2, cycles, ns; > + uint32_t i; > + uint32_t tot = 0; > + > + if (create_queue(thr, msg_pool, prio)) > + return -1; > + > + t1 = odp_time_get_cycles(); > + > + for (i = 0; i < QUEUE_ROUNDS; i++) { > + buf = odp_schedule(&queue, ODP_SCHED_WAIT); > + > + if (odp_queue_enq(queue, buf)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + /* Clear possible locally stored buffers */ > + odp_schedule_pause(); > + > + tot = i; > + > + while (1) { > + buf = odp_schedule(&queue, ODP_SCHED_NO_WAIT); > + > + if (buf == ODP_BUFFER_INVALID) > + break; > + > + tot++; > + > + if (odp_queue_enq(queue, buf)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + odp_schedule_resume(); > + > + t2 = odp_time_get_cycles(); > + cycles = odp_time_diff_cycles(t1, t2); > + ns = odp_time_cycles_to_ns(cycles); > + > + odp_barrier_sync(barrier); > + clear_sched_queues(); > + > + cycles = cycles/tot; > + ns = ns/tot; > + > + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n", > + thr, str, cycles, ns); > + > + return 0; > +} > + > +/** > + * @internal Test scheduling of multiple queues - with odp_schedule() > + * > + * Enqueue a buffer to each queue. Schedule and enqueue the received > + * buffer back into the queue it came from. > + * > + * @param str Test case name string > + * @param thr Thread > + * @param msg_pool Buffer pool > + * @param prio Priority > + * @param barrier Barrier > + * > + * @return 0 if successful > + */ > +static int test_schedule_many(const char *str, int thr, > + odp_buffer_pool_t msg_pool, > + int prio, odp_barrier_t *barrier) > +{ > + odp_buffer_t buf; > + odp_queue_t queue; > + uint64_t t1 = 0; > + uint64_t t2 = 0; > + uint64_t cycles, ns; > + uint32_t i; > + uint32_t tot = 0; > + > + if (create_queues(thr, msg_pool, prio)) > + return -1; > + > + /* Start sched-enq loop */ > + t1 = odp_time_get_cycles(); > + > + for (i = 0; i < QUEUE_ROUNDS; i++) { > + buf = odp_schedule(&queue, ODP_SCHED_WAIT); > + > + if (odp_queue_enq(queue, buf)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + /* Clear possible locally stored buffers */ > + odp_schedule_pause(); > + > + tot = i; > + > + while (1) { > + buf = odp_schedule(&queue, ODP_SCHED_NO_WAIT); > + > + if (buf == ODP_BUFFER_INVALID) > + break; > + > + tot++; > + > + if (odp_queue_enq(queue, buf)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + odp_schedule_resume(); > + > + t2 = odp_time_get_cycles(); > + cycles = odp_time_diff_cycles(t1, t2); > + ns = odp_time_cycles_to_ns(cycles); > + > + odp_barrier_sync(barrier); > + clear_sched_queues(); > + > + cycles = cycles/tot; > + ns = ns/tot; > + > + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n", > + thr, str, cycles, ns); > + > + return 0; > +} > + > +/** > + * @internal Test scheduling of multiple queues with multi_sched and > multi_enq > + * > + * @param str Test case name string > + * @param thr Thread > + * @param msg_pool Buffer pool > + * @param prio Priority > + * @param barrier Barrier > + * > + * @return 0 if successful > + */ > +static int test_schedule_multi(const char *str, int thr, > + odp_buffer_pool_t msg_pool, > + int prio, odp_barrier_t *barrier) > +{ > + odp_buffer_t buf[MULTI_BUFS_MAX]; > + odp_queue_t queue; > + uint64_t t1 = 0; > + uint64_t t2 = 0; > + uint64_t cycles, ns; > + int i, j; > + int num; > + uint32_t tot = 0; > + char name[] = "sched_XX_YY"; > + > + name[6] = '0' + prio/10; > + name[7] = '0' + prio - 10*(prio/10); > + > + /* Alloc and enqueue a buffer per queue */ > + for (i = 0; i < QUEUES_PER_PRIO; i++) { > + name[9] = '0' + i/10; > + name[10] = '0' + i - 10*(i/10); > + > + queue = odp_queue_lookup(name); > + > + if (queue == ODP_QUEUE_INVALID) { > + ODP_ERR(" [%i] Queue %s lookup failed.\n", thr, name); > + return -1; > + } > + > + for (j = 0; j < MULTI_BUFS_MAX; j++) { > + buf[j] = odp_buffer_alloc(msg_pool); > + > + if (!odp_buffer_is_valid(buf[j])) { > + ODP_ERR(" [%i] msg_pool alloc failed\n", thr); > + return -1; > + } > + } > + > + if (odp_queue_enq_multi(queue, buf, MULTI_BUFS_MAX)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + /* Start sched-enq loop */ > + t1 = odp_time_get_cycles(); > + > + for (i = 0; i < QUEUE_ROUNDS; i++) { > + num = odp_schedule_multi(&queue, ODP_SCHED_WAIT, buf, > + MULTI_BUFS_MAX); > + > + tot += num; > + > + if (odp_queue_enq_multi(queue, buf, num)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + /* Clear possible locally stored buffers */ > + odp_schedule_pause(); > + > + while (1) { > + num = odp_schedule_multi(&queue, ODP_SCHED_NO_WAIT, buf, > + MULTI_BUFS_MAX); > + > + if (num == 0) > + break; > + > + tot += num; > + > + if (odp_queue_enq_multi(queue, buf, num)) { > + ODP_ERR(" [%i] Queue enqueue failed.\n", thr); > + return -1; > + } > + } > + > + odp_schedule_resume(); > + > + > + t2 = odp_time_get_cycles(); > + cycles = odp_time_diff_cycles(t1, t2); > + ns = odp_time_cycles_to_ns(cycles); > + > + odp_barrier_sync(barrier); > + clear_sched_queues(); > + > + if (tot) { > + cycles = cycles/tot; > + ns = ns/tot; > + } else { > + cycles = 0; > + ns = 0; > + } > + > + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n", > + thr, str, cycles, ns); > + > + return 0; > +} > + > +/** > + * Template function for running the scheduler tests. > + * The main reason for having this function is that CUnit does not offer a > way > + * to pass arguments to a testcase function. > + * The other reason is that there are common steps for all testcases. > + */ > +static void *exec_template(void *arg) > +{ > + odp_buffer_pool_t msg_pool; > + odp_shm_t shm; > + test_globals_t *globals; > + odp_barrier_t *barrier; > + test_case_args_t *args = (test_case_args_t*) arg; > + > + shm = odp_shm_lookup("test_globals"); > + globals = odp_shm_addr(shm); > + > + CU_ASSERT(globals != NULL); > + > + barrier = &globals->barrier; > + > + /* > + * Sync before start > + */ > + odp_barrier_sync(barrier); > + > + /* > + * Find the buffer pool > + */ > + msg_pool = odp_buffer_pool_lookup("msg_pool"); > + > + CU_ASSERT(msg_pool != ODP_BUFFER_POOL_INVALID); > + > + odp_barrier_sync(barrier); > + > + /* > + * Now run the testcase routine passing the arguments > + */ > + args->func(args->name, odp_thread_id(), msg_pool, > + args->prio, barrier); > + > + return arg; > +} > + > +/* Low prio */ > + > +static void schedule_one_single_lo(void) > +{ > + test_case_args_t args; > + snprintf(args.name, sizeof(args.name), "sched_one_s_lo"); > + args.prio = ODP_SCHED_PRIO_LOWEST; > + args.func = test_schedule_one_single; > + execute_parallel(exec_template, &args); > +} > + > +static void schedule_single_lo(void) > +{ > + test_case_args_t args; > + snprintf(args.name, sizeof(args.name), "sched_____s_lo"); > + args.prio = ODP_SCHED_PRIO_LOWEST; > + args.func = test_schedule_single; > + execute_parallel(exec_template, &args); > +} > + > +static void schedule_one_many_lo(void) > +{ > + test_case_args_t args; > + snprintf(args.name, sizeof(args.name), "sched_one_m_lo"); > + args.prio = ODP_SCHED_PRIO_LOWEST; > + args.func = test_schedule_one_many; > + execute_parallel(exec_template, &args); > +} > + > +static void schedule_many_lo(void) > +{ > + test_case_args_t args; > + snprintf(args.name, sizeof(args.name), "sched_____m_lo"); > + args.prio = ODP_SCHED_PRIO_LOWEST; > + args.func = test_schedule_many; > + execute_parallel(exec_template, &args); > +} > + > +static void schedule_multi_lo(void) > +{ > + test_case_args_t args; > + snprintf(args.name, sizeof(args.name), "sched_____m_lo"); > + args.prio = ODP_SCHED_PRIO_LOWEST; > + args.func = test_schedule_multi; > + execute_parallel(exec_template, &args); > +} > + > +/* High prio */ > + > +static void schedule_one_single_hi(void) > +{ > + test_case_args_t args; > + snprintf(args.name, sizeof(args.name), "sched_one_s_hi"); > + args.prio = ODP_SCHED_PRIO_HIGHEST; > + args.func = test_schedule_single; > + execute_parallel(exec_template, &args); > +} > + > +static void schedule_single_hi(void) > +{ > + test_case_args_t args; > + snprintf(args.name, sizeof(args.name), "sched_____s_hi"); > + args.prio = ODP_SCHED_PRIO_HIGHEST; > + args.func = test_schedule_single; > + execute_parallel(exec_template, &args); > +} > + > +static void schedule_one_many_hi(void) > +{ > + test_case_args_t args; > + snprintf(args.name, sizeof(args.name), "sched_one_m_hi"); > + args.prio = ODP_SCHED_PRIO_HIGHEST; > + args.func = test_schedule_one_many; > + execute_parallel(exec_template, &args); > +} > + > +static void schedule_many_hi(void) > +{ > + test_case_args_t args; > + snprintf(args.name, sizeof(args.name), "sched_____m_hi"); > + args.prio = ODP_SCHED_PRIO_HIGHEST; > + args.func = test_schedule_many; > + execute_parallel(exec_template, &args); > +} > + > +static void schedule_multi_hi(void) > +{ > + test_case_args_t args; IMO, We should avoid using variables allocated from stack to share the data between different execution context. It will work in linux threaded run-time environment.But it will have issues in different run-time environment like bare-metal. IMO if any memory needs to be shared across different execution environment should use the memory allocated from odp shared mem. > + snprintf(args.name, sizeof(args.name), "sched_multi_hi"); > + args.prio = ODP_SCHED_PRIO_HIGHEST; > + args.func = test_schedule_multi; > + execute_parallel(exec_template, &args); > +} > + > +static void execute_parallel(void *(*start_routine) (void *), > + test_case_args_t *test_case_args) > +{ > + odph_linux_pthread_t thread_tbl[MAX_WORKERS]; > + int first_core; > + > + memset(thread_tbl, 0, sizeof(thread_tbl)); > + > + /* > + * By default core #0 runs Linux kernel background tasks. > + * Start mapping thread from core #1 > + */ > + first_core = 1; > + > + if (odp_sys_core_count() == 1) > + first_core = 0; > + > + odph_linux_pthread_create(thread_tbl, num_workers, first_core, > + start_routine, test_case_args); > + > + /* Wait for worker threads to terminate */ > + odph_linux_pthread_join(thread_tbl, num_workers); > +} > + > +static odp_buffer_pool_t test_odp_buffer_pool_init(void) > +{ > + void *pool_base; > + odp_shm_t shm; > + odp_buffer_pool_t pool; > + > + shm = odp_shm_reserve("msg_pool", > + MSG_POOL_SIZE, ODP_CACHE_LINE_SIZE, 0); > + > + pool_base = odp_shm_addr(shm); > + > + if (NULL == pool_base) { > + printf("Shared memory reserve failed.\n"); > + return -1; > + } > + > + pool = odp_buffer_pool_create("msg_pool", pool_base, MSG_POOL_SIZE, > + BUF_SIZE, ODP_CACHE_LINE_SIZE, > + ODP_BUFFER_TYPE_RAW); > + > + if (ODP_BUFFER_POOL_INVALID == pool) { > + printf("Pool create failed.\n"); > + return -1; > + } > + return pool; > +} > + > +int schedule_test_init(void) > +{ > + test_args_t args; > + odp_shm_t shm; > + test_globals_t *globals; > + int i, j; > + int prios; > + > + if (0 != odp_init_global(NULL, NULL)) { > + printf("odp_init_global fail.\n"); > + return -1; > + } > + if (0 != odp_init_local()) { > + printf("odp_init_local fail.\n"); > + return -1; > + } > + if (ODP_BUFFER_POOL_INVALID == test_odp_buffer_pool_init()) { > + printf("test_odp_buffer_pool_init fail.\n"); > + return -1; > + } > + > + /* A worker thread per core */ > + num_workers = odp_sys_core_count(); > + > + if (args.core_count) args.core_count is uninitialized > + num_workers = args.core_count; > + > + /* force to max core count */ > + if (num_workers > MAX_WORKERS) > + num_workers = MAX_WORKERS; > + shm = odp_shm_reserve("test_globals", > + sizeof(test_globals_t), ODP_CACHE_LINE_SIZE, 0); > + > + globals = odp_shm_addr(shm); > + > + if (globals == NULL) { > + ODP_ERR("Shared memory reserve failed.\n"); > + return -1; > + } > + > + memset(globals, 0, sizeof(test_globals_t)); > + > + /* Barrier to sync test case execution */ > + odp_barrier_init_count(&globals->barrier, num_workers); > + > + prios = odp_schedule_num_prio(); > + > + for (i = 0; i < prios; i++) { > + odp_queue_param_t param; > + odp_queue_t queue; > + char name[] = "sched_XX_YY"; > + > + if (i != ODP_SCHED_PRIO_HIGHEST && > + i != ODP_SCHED_PRIO_LOWEST) > + continue; > + > + name[6] = '0' + i/10; > + name[7] = '0' + i - 10*(i/10); > + > + param.sched.prio = i; > + param.sched.sync = ODP_SCHED_SYNC_ATOMIC; > + param.sched.group = ODP_SCHED_GROUP_DEFAULT; > + > + for (j = 0; j < QUEUES_PER_PRIO; j++) { > + name[9] = '0' + j/10; > + name[10] = '0' + j - 10*(j/10); > + > + queue = odp_queue_create(name, ODP_QUEUE_TYPE_SCHED, > + ¶m); > + > + if (queue == ODP_QUEUE_INVALID) { > + ODP_ERR("Schedule queue create failed.\n"); > + return -1; > + } > + } > + } > + return 0; > +} > + > +int schedule_test_finalize(void) > +{ > + odp_term_local(); > + odp_term_global(); > + return 0; > +} > + > +struct CU_TestInfo schedule_tests[] = { > + _CU_TEST_INFO(schedule_wait_time), > + _CU_TEST_INFO(schedule_one_single_lo), > + _CU_TEST_INFO(schedule_single_lo), > + _CU_TEST_INFO(schedule_one_many_lo), > + _CU_TEST_INFO(schedule_many_lo), > + _CU_TEST_INFO(schedule_multi_lo), > + _CU_TEST_INFO(schedule_one_single_hi), > + _CU_TEST_INFO(schedule_single_hi), > + _CU_TEST_INFO(schedule_one_many_hi), > + _CU_TEST_INFO(schedule_many_hi), > + _CU_TEST_INFO(schedule_multi_hi), > + CU_TEST_INFO_NULL, > +}; > diff --git a/test/cunit/schedule/odp_schedule_testsuites.c > b/test/cunit/schedule/odp_schedule_testsuites.c > new file mode 100644 > index 0000000..1053069 > --- /dev/null > +++ b/test/cunit/schedule/odp_schedule_testsuites.c > @@ -0,0 +1,35 @@ > +/* Copyright (c) 2014, Linaro Limited > + * All rights reserved. > + * > + * SPDX-License-Identifier: BSD-3-Clause > + */ > + > +#include "odp_schedule_testsuites.h" > + > +static CU_SuiteInfo suites[] = { > + { > + "Scheduler tests" , > + schedule_test_init, > + schedule_test_finalize, > + NULL, > + NULL, > + schedule_tests > + }, > + CU_SUITE_INFO_NULL, > +}; > + > +int main(void) > +{ > + /* initialize the CUnit test registry */ > + if (CUE_SUCCESS != CU_initialize_registry()) > + return CU_get_error(); > + > + /* register suites */ > + CU_register_suites(suites); > + /* Run all tests using the CUnit Basic interface */ > + CU_basic_set_mode(CU_BRM_VERBOSE); > + CU_basic_run_tests(); > + CU_cleanup_registry(); > + > + return CU_get_error(); > +} > diff --git a/test/cunit/schedule/odp_schedule_testsuites.h > b/test/cunit/schedule/odp_schedule_testsuites.h > new file mode 100644 > index 0000000..67a2a69 > --- /dev/null > +++ b/test/cunit/schedule/odp_schedule_testsuites.h > @@ -0,0 +1,21 @@ > +/* Copyright (c) 2014, Linaro Limited > + * All rights reserved. > + * > + * SPDX-License-Identifier: BSD-3-Clause > + */ > + > +#ifndef ODP_SCHEDULE_TESTSUITES_H_ > +#define ODP_SCHEDULE_TESTSUITES_H_ > + > +#include "odp.h" > +#include <CUnit/Basic.h> > + > +/* Helper macro for CU_TestInfo initialization */ > +#define _CU_TEST_INFO(test_func) {#test_func, test_func} > + > +extern struct CU_TestInfo schedule_tests[]; > + > +extern int schedule_test_init(void); > +extern int schedule_test_finalize(void); > + > +#endif /* ODP_SCHEDULE_TESTSUITES_H_ */ > -- > 1.8.3.2 > > > _______________________________________________ > lng-odp mailing list > lng-odp@lists.linaro.org > http://lists.linaro.org/mailman/listinfo/lng-odp _______________________________________________ lng-odp mailing list lng-odp@lists.linaro.org http://lists.linaro.org/mailman/listinfo/lng-odp