On Thu, Nov 20, 2014 at 09:02:24PM +0200, Ciprian Barbu wrote:
> Signed-off-by: Ciprian Barbu <[email protected]>
> ---
> The testcases are based almost entirely on the odp_example.
> There are no alloc tests and I added a test case for odp_schedule_wait_time.
> The major differencs between the odp_example and this cunit is the partition
> into testcases, the odp_example calls every test case from one big function.
>
> I had to work some magic in order to be able to pass arguments to test cases,
> I hope is not too hard to follow.
>
> configure.ac | 1 +
> test/cunit/Makefile.am | 2 +
> test/cunit/schedule/Makefile.am | 10 +
> test/cunit/schedule/odp_schedule_test.c | 844
> ++++++++++++++++++++++++++
> test/cunit/schedule/odp_schedule_testsuites.c | 35 ++
> test/cunit/schedule/odp_schedule_testsuites.h | 21 +
> 6 files changed, 913 insertions(+)
> create mode 100644 test/cunit/schedule/Makefile.am
> create mode 100644 test/cunit/schedule/odp_schedule_test.c
> create mode 100644 test/cunit/schedule/odp_schedule_testsuites.c
> create mode 100644 test/cunit/schedule/odp_schedule_testsuites.h
>
> diff --git a/configure.ac b/configure.ac
> index fcd7279..a47db72 100644
> --- a/configure.ac
> +++ b/configure.ac
> @@ -173,6 +173,7 @@ AC_CONFIG_FILES([Makefile
> test/Makefile
> test/api_test/Makefile
> test/cunit/Makefile
> + test/cunit/schedule/Makefile
> pkgconfig/libodp.pc])
>
> AC_SEARCH_LIBS([timer_create],[rt posix4])
> diff --git a/test/cunit/Makefile.am b/test/cunit/Makefile.am
> index 439e134..b6033ee 100644
> --- a/test/cunit/Makefile.am
> +++ b/test/cunit/Makefile.am
> @@ -3,6 +3,8 @@ include $(top_srcdir)/test/Makefile.inc
> AM_CFLAGS += -I$(CUNIT_PATH)/include
> AM_LDFLAGS += -L$(CUNIT_PATH)/lib -static -lcunit
>
> +SUBDIRS = schedule
> +
> if ODP_CUNIT_ENABLED
> TESTS = ${bin_PROGRAMS}
> check_PROGRAMS = ${bin_PROGRAMS}
> diff --git a/test/cunit/schedule/Makefile.am b/test/cunit/schedule/Makefile.am
> new file mode 100644
> index 0000000..ad68b03
> --- /dev/null
> +++ b/test/cunit/schedule/Makefile.am
> @@ -0,0 +1,10 @@
> +include $(top_srcdir)/test/Makefile.inc
> +
> +if ODP_CUNIT_ENABLED
> +bin_PROGRAMS = odp_schedule_test
> +odp_schedule_test_LDFLAGS = $(AM_LDFLAGS) -L$(CUNIT_PATH)/lib -static -lcunit
> +odp_schedule_test_CFLAGS = $(AM_CFLAGS) -I$(CUNIT_PATH)/include
> +endif
> +
> +dist_odp_schedule_test_SOURCES = odp_schedule_test.c \
> + odp_schedule_testsuites.c
> diff --git a/test/cunit/schedule/odp_schedule_test.c
> b/test/cunit/schedule/odp_schedule_test.c
> new file mode 100644
> index 0000000..fa67f6e
> --- /dev/null
> +++ b/test/cunit/schedule/odp_schedule_test.c
> @@ -0,0 +1,844 @@
> +/* Copyright (c) 2014, Linaro Limited
> + * All rights reserved.
> + *
> + * SPDX-License-Identifier: BSD-3-Clause
> + */
> +
> +#include "odp_schedule_testsuites.h"
> +#include <odph_linux.h>
> +
> +#define MAX_WORKERS 32 /**< Max worker threads */
> +#define MSG_POOL_SIZE (4*1024*1024)
> +#define QUEUES_PER_PRIO 64 /**< Queue per priority */
> +#define QUEUE_ROUNDS (512*1024) /**< Queue test rounds */
> +#define MULTI_BUFS_MAX 4 /**< Buffer burst size */
> +#define BUF_SIZE 64
> +
> +#define SCHED_MSG "Test_buff_FOR_simple_schedule"
> +
> +/** Test arguments */
> +typedef struct {
> + int core_count; /**< Core count */
> + int proc_mode; /**< Process mode */
> +} test_args_t;
> +
> +typedef int (*test_case_routine)(const char *, int, odp_buffer_pool_t,
> + int, odp_barrier_t *);
> +
> +/** Scheduler test case arguments */
> +typedef struct {
> + char name[64]; /**< test case name */
> + int prio;
> + test_case_routine func;
> +} test_case_args_t;
> +
> +/** Test global variables */
> +typedef struct {
> + odp_barrier_t barrier;/**< @private Barrier for test synchronisation */
> + test_args_t test_args;/**< @private Test case function and arguments */
> +} test_globals_t;
> +
> +static void execute_parallel(void *(*func) (void *), test_case_args_t *);
> +static int num_workers;
> +
> +/**
> + * @internal CUnit test case for verifying functionality of
> + * schedule_wait_time
> + */
> +static void schedule_wait_time(void)
> +{
> + uint64_t wait_time;
> +
> + wait_time = odp_schedule_wait_time(0);
> + CU_ASSERT(wait_time > 0);
wait_time can be zero for odp_schedule_wait_time(0).
> + CU_PASS("schedule_wait_time(0)");
> +
> + wait_time = odp_schedule_wait_time(1);
> + CU_ASSERT(wait_time > 0);
> + CU_PASS("schedule_wait_time(1)");
> +
> + wait_time = odp_schedule_wait_time((uint64_t)-1LL);
> + CU_ASSERT(wait_time > 0);
> + CU_PASS("schedule_wait_time(MAX_LONG_INT)");
> +}
> +
> +/**
> + * @internal Clear all scheduled queues. Retry to be sure that all
> + * buffers have been scheduled.
> + */
> +static void clear_sched_queues(void)
> +{
> + odp_buffer_t buf;
> +
> + while (1) {
> + buf = odp_schedule(NULL, ODP_SCHED_NO_WAIT);
> +
> + if (buf == ODP_BUFFER_INVALID)
> + break;
> +
> + odp_buffer_free(buf);
> + }
> +}
> +
> +/**
> + * @internal Create multiple queues from a pool of buffers
> + *
> + * @param thr Thread
> + * @param msg_pool Buffer pool
> + * @param prio Queue priority
> + *
> + * @return 0 if successful
> + */
> +static int create_queues(int thr, odp_buffer_pool_t msg_pool, int prio)
> +{
> + char name[] = "sched_XX_YY";
> + odp_buffer_t buf;
> + odp_queue_t queue;
> + int i;
> +
> + name[6] = '0' + prio/10;
> + name[7] = '0' + prio - 10*(prio/10);
> +
> + /* Alloc and enqueue a buffer per queue */
> + for (i = 0; i < QUEUES_PER_PRIO; i++) {
> + name[9] = '0' + i/10;
> + name[10] = '0' + i - 10*(i/10);
> +
> + queue = odp_queue_lookup(name);
> +
> + if (queue == ODP_QUEUE_INVALID) {
> + ODP_ERR(" [%i] Queue %s lookup failed.\n", thr, name);
> + return -1;
> + }
> +
> + buf = odp_buffer_alloc(msg_pool);
> +
> + if (!odp_buffer_is_valid(buf)) {
> + ODP_ERR(" [%i] msg_pool alloc failed\n", thr);
> + return -1;
> + }
> +
> + if (odp_queue_enq(queue, buf)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + return 0;
> +}
> +
> +/**
> + * @internal Create a single queue from a pool of buffers
> + *
> + * @param thr Thread
> + * @param msg_pool Buffer pool
> + * @param prio Queue priority
> + *
> + * @return 0 if successful
> + */
> +static int create_queue(int thr, odp_buffer_pool_t msg_pool, int prio)
> +{
> + char name[] = "sched_XX_00";
> + odp_buffer_t buf;
> + odp_queue_t queue;
> +
> + buf = odp_buffer_alloc(msg_pool);
> +
> + if (!odp_buffer_is_valid(buf)) {
> + ODP_ERR(" [%i] msg_pool alloc failed\n", thr);
> + return -1;
> + }
> +
> + name[6] = '0' + prio/10;
> + name[7] = '0' + prio - 10*(prio/10);
> +
> + queue = odp_queue_lookup(name);
> +
> + if (queue == ODP_QUEUE_INVALID) {
> + ODP_ERR(" [%i] Queue %s lookup failed.\n", thr, name);
> + return -1;
> + }
> +
> + if (odp_queue_enq(queue, buf)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> +
> + return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of a single queue - with odp_schedule_one()
> + *
> + * Enqueue a buffer to the shared queue. Schedule and enqueue the received
> + * buffer back into the queue.
> + *
> + * @param str Test case name string
> + * @param thr Thread
> + * @param msg_pool Buffer pool
> + * @param prio Priority
> + * @param barrier Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_one_single(const char *str, int thr,
> + odp_buffer_pool_t msg_pool,
> + int prio, odp_barrier_t *barrier)
> +{
> + odp_buffer_t buf;
> + odp_queue_t queue;
> + uint64_t t1, t2, cycles, ns;
> + uint32_t i;
> + uint32_t tot = 0;
> +
> + if (create_queue(thr, msg_pool, prio)) {
> + CU_FAIL_FATAL("lookup queue");
> + return -1;
> + }
> +
> + t1 = odp_time_get_cycles();
> +
> + for (i = 0; i < QUEUE_ROUNDS; i++) {
> + buf = odp_schedule_one(&queue, ODP_SCHED_WAIT);
> +
> + if (odp_queue_enq(queue, buf)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + if (odp_queue_sched_type(queue) == ODP_SCHED_SYNC_ATOMIC)
> + odp_schedule_release_atomic();
> +
> + t2 = odp_time_get_cycles();
> + cycles = odp_time_diff_cycles(t1, t2);
> + ns = odp_time_cycles_to_ns(cycles);
> + tot = i;
> +
> + odp_barrier_sync(barrier);
> + clear_sched_queues();
> +
> + cycles = cycles/tot;
> + ns = ns/tot;
> +
> + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> + thr, str, cycles, ns);
> +
> + return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of multiple queues - with odp_schedule_one()
> + *
> + * Enqueue a buffer to each queue. Schedule and enqueue the received
> + * buffer back into the queue it came from.
> + *
> + * @param str Test case name string
> + * @param thr Thread
> + * @param msg_pool Buffer pool
> + * @param prio Priority
> + * @param barrier Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_one_many(const char *str, int thr,
> + odp_buffer_pool_t msg_pool,
> + int prio, odp_barrier_t *barrier)
> +{
> + odp_buffer_t buf;
> + odp_queue_t queue;
> + uint64_t t1 = 0;
> + uint64_t t2 = 0;
> + uint64_t cycles, ns;
> + uint32_t i;
> + uint32_t tot = 0;
> +
> + if (create_queues(thr, msg_pool, prio))
> + return -1;
> +
> + /* Start sched-enq loop */
> + t1 = odp_time_get_cycles();
> +
> + for (i = 0; i < QUEUE_ROUNDS; i++) {
> + buf = odp_schedule_one(&queue, ODP_SCHED_WAIT);
> +
> + if (odp_queue_enq(queue, buf)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + if (odp_queue_sched_type(queue) == ODP_SCHED_SYNC_ATOMIC)
> + odp_schedule_release_atomic();
> +
> + t2 = odp_time_get_cycles();
> + cycles = odp_time_diff_cycles(t1, t2);
> + ns = odp_time_cycles_to_ns(cycles);
> + tot = i;
> +
> + odp_barrier_sync(barrier);
> + clear_sched_queues();
> +
> + cycles = cycles/tot;
> + ns = ns/tot;
> +
> + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> + thr, str, cycles, ns);
> +
> + return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of a single queue - with odp_schedule()
> + *
> + * Enqueue a buffer to the shared queue. Schedule and enqueue the received
> + * buffer back into the queue.
> + *
> + * @param str Test case name string
> + * @param thr Thread
> + * @param msg_pool Buffer pool
> + * @param prio Priority
> + * @param barrier Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_single(const char *str, int thr,
> + odp_buffer_pool_t msg_pool,
> + int prio, odp_barrier_t *barrier)
> +{
> + odp_buffer_t buf;
> + odp_queue_t queue;
> + uint64_t t1, t2, cycles, ns;
> + uint32_t i;
> + uint32_t tot = 0;
> +
> + if (create_queue(thr, msg_pool, prio))
> + return -1;
> +
> + t1 = odp_time_get_cycles();
> +
> + for (i = 0; i < QUEUE_ROUNDS; i++) {
> + buf = odp_schedule(&queue, ODP_SCHED_WAIT);
> +
> + if (odp_queue_enq(queue, buf)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + /* Clear possible locally stored buffers */
> + odp_schedule_pause();
> +
> + tot = i;
> +
> + while (1) {
> + buf = odp_schedule(&queue, ODP_SCHED_NO_WAIT);
> +
> + if (buf == ODP_BUFFER_INVALID)
> + break;
> +
> + tot++;
> +
> + if (odp_queue_enq(queue, buf)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + odp_schedule_resume();
> +
> + t2 = odp_time_get_cycles();
> + cycles = odp_time_diff_cycles(t1, t2);
> + ns = odp_time_cycles_to_ns(cycles);
> +
> + odp_barrier_sync(barrier);
> + clear_sched_queues();
> +
> + cycles = cycles/tot;
> + ns = ns/tot;
> +
> + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> + thr, str, cycles, ns);
> +
> + return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of multiple queues - with odp_schedule()
> + *
> + * Enqueue a buffer to each queue. Schedule and enqueue the received
> + * buffer back into the queue it came from.
> + *
> + * @param str Test case name string
> + * @param thr Thread
> + * @param msg_pool Buffer pool
> + * @param prio Priority
> + * @param barrier Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_many(const char *str, int thr,
> + odp_buffer_pool_t msg_pool,
> + int prio, odp_barrier_t *barrier)
> +{
> + odp_buffer_t buf;
> + odp_queue_t queue;
> + uint64_t t1 = 0;
> + uint64_t t2 = 0;
> + uint64_t cycles, ns;
> + uint32_t i;
> + uint32_t tot = 0;
> +
> + if (create_queues(thr, msg_pool, prio))
> + return -1;
> +
> + /* Start sched-enq loop */
> + t1 = odp_time_get_cycles();
> +
> + for (i = 0; i < QUEUE_ROUNDS; i++) {
> + buf = odp_schedule(&queue, ODP_SCHED_WAIT);
> +
> + if (odp_queue_enq(queue, buf)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + /* Clear possible locally stored buffers */
> + odp_schedule_pause();
> +
> + tot = i;
> +
> + while (1) {
> + buf = odp_schedule(&queue, ODP_SCHED_NO_WAIT);
> +
> + if (buf == ODP_BUFFER_INVALID)
> + break;
> +
> + tot++;
> +
> + if (odp_queue_enq(queue, buf)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + odp_schedule_resume();
> +
> + t2 = odp_time_get_cycles();
> + cycles = odp_time_diff_cycles(t1, t2);
> + ns = odp_time_cycles_to_ns(cycles);
> +
> + odp_barrier_sync(barrier);
> + clear_sched_queues();
> +
> + cycles = cycles/tot;
> + ns = ns/tot;
> +
> + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> + thr, str, cycles, ns);
> +
> + return 0;
> +}
> +
> +/**
> + * @internal Test scheduling of multiple queues with multi_sched and
> multi_enq
> + *
> + * @param str Test case name string
> + * @param thr Thread
> + * @param msg_pool Buffer pool
> + * @param prio Priority
> + * @param barrier Barrier
> + *
> + * @return 0 if successful
> + */
> +static int test_schedule_multi(const char *str, int thr,
> + odp_buffer_pool_t msg_pool,
> + int prio, odp_barrier_t *barrier)
> +{
> + odp_buffer_t buf[MULTI_BUFS_MAX];
> + odp_queue_t queue;
> + uint64_t t1 = 0;
> + uint64_t t2 = 0;
> + uint64_t cycles, ns;
> + int i, j;
> + int num;
> + uint32_t tot = 0;
> + char name[] = "sched_XX_YY";
> +
> + name[6] = '0' + prio/10;
> + name[7] = '0' + prio - 10*(prio/10);
> +
> + /* Alloc and enqueue a buffer per queue */
> + for (i = 0; i < QUEUES_PER_PRIO; i++) {
> + name[9] = '0' + i/10;
> + name[10] = '0' + i - 10*(i/10);
> +
> + queue = odp_queue_lookup(name);
> +
> + if (queue == ODP_QUEUE_INVALID) {
> + ODP_ERR(" [%i] Queue %s lookup failed.\n", thr, name);
> + return -1;
> + }
> +
> + for (j = 0; j < MULTI_BUFS_MAX; j++) {
> + buf[j] = odp_buffer_alloc(msg_pool);
> +
> + if (!odp_buffer_is_valid(buf[j])) {
> + ODP_ERR(" [%i] msg_pool alloc failed\n", thr);
> + return -1;
> + }
> + }
> +
> + if (odp_queue_enq_multi(queue, buf, MULTI_BUFS_MAX)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + /* Start sched-enq loop */
> + t1 = odp_time_get_cycles();
> +
> + for (i = 0; i < QUEUE_ROUNDS; i++) {
> + num = odp_schedule_multi(&queue, ODP_SCHED_WAIT, buf,
> + MULTI_BUFS_MAX);
> +
> + tot += num;
> +
> + if (odp_queue_enq_multi(queue, buf, num)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + /* Clear possible locally stored buffers */
> + odp_schedule_pause();
> +
> + while (1) {
> + num = odp_schedule_multi(&queue, ODP_SCHED_NO_WAIT, buf,
> + MULTI_BUFS_MAX);
> +
> + if (num == 0)
> + break;
> +
> + tot += num;
> +
> + if (odp_queue_enq_multi(queue, buf, num)) {
> + ODP_ERR(" [%i] Queue enqueue failed.\n", thr);
> + return -1;
> + }
> + }
> +
> + odp_schedule_resume();
> +
> +
> + t2 = odp_time_get_cycles();
> + cycles = odp_time_diff_cycles(t1, t2);
> + ns = odp_time_cycles_to_ns(cycles);
> +
> + odp_barrier_sync(barrier);
> + clear_sched_queues();
> +
> + if (tot) {
> + cycles = cycles/tot;
> + ns = ns/tot;
> + } else {
> + cycles = 0;
> + ns = 0;
> + }
> +
> + printf(" [%i] %s enq+deq %"PRIu64" cycles, %"PRIu64" ns\n",
> + thr, str, cycles, ns);
> +
> + return 0;
> +}
> +
> +/**
> + * Template function for running the scheduler tests.
> + * The main reason for having this function is that CUnit does not offer a
> way
> + * to pass arguments to a testcase function.
> + * The other reason is that there are common steps for all testcases.
> + */
> +static void *exec_template(void *arg)
> +{
> + odp_buffer_pool_t msg_pool;
> + odp_shm_t shm;
> + test_globals_t *globals;
> + odp_barrier_t *barrier;
> + test_case_args_t *args = (test_case_args_t*) arg;
> +
> + shm = odp_shm_lookup("test_globals");
> + globals = odp_shm_addr(shm);
> +
> + CU_ASSERT(globals != NULL);
> +
> + barrier = &globals->barrier;
> +
> + /*
> + * Sync before start
> + */
> + odp_barrier_sync(barrier);
> +
> + /*
> + * Find the buffer pool
> + */
> + msg_pool = odp_buffer_pool_lookup("msg_pool");
> +
> + CU_ASSERT(msg_pool != ODP_BUFFER_POOL_INVALID);
> +
> + odp_barrier_sync(barrier);
> +
> + /*
> + * Now run the testcase routine passing the arguments
> + */
> + args->func(args->name, odp_thread_id(), msg_pool,
> + args->prio, barrier);
> +
> + return arg;
> +}
> +
> +/* Low prio */
> +
> +static void schedule_one_single_lo(void)
> +{
> + test_case_args_t args;
> + snprintf(args.name, sizeof(args.name), "sched_one_s_lo");
> + args.prio = ODP_SCHED_PRIO_LOWEST;
> + args.func = test_schedule_one_single;
> + execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_single_lo(void)
> +{
> + test_case_args_t args;
> + snprintf(args.name, sizeof(args.name), "sched_____s_lo");
> + args.prio = ODP_SCHED_PRIO_LOWEST;
> + args.func = test_schedule_single;
> + execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_one_many_lo(void)
> +{
> + test_case_args_t args;
> + snprintf(args.name, sizeof(args.name), "sched_one_m_lo");
> + args.prio = ODP_SCHED_PRIO_LOWEST;
> + args.func = test_schedule_one_many;
> + execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_many_lo(void)
> +{
> + test_case_args_t args;
> + snprintf(args.name, sizeof(args.name), "sched_____m_lo");
> + args.prio = ODP_SCHED_PRIO_LOWEST;
> + args.func = test_schedule_many;
> + execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_multi_lo(void)
> +{
> + test_case_args_t args;
> + snprintf(args.name, sizeof(args.name), "sched_____m_lo");
> + args.prio = ODP_SCHED_PRIO_LOWEST;
> + args.func = test_schedule_multi;
> + execute_parallel(exec_template, &args);
> +}
> +
> +/* High prio */
> +
> +static void schedule_one_single_hi(void)
> +{
> + test_case_args_t args;
> + snprintf(args.name, sizeof(args.name), "sched_one_s_hi");
> + args.prio = ODP_SCHED_PRIO_HIGHEST;
> + args.func = test_schedule_single;
> + execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_single_hi(void)
> +{
> + test_case_args_t args;
> + snprintf(args.name, sizeof(args.name), "sched_____s_hi");
> + args.prio = ODP_SCHED_PRIO_HIGHEST;
> + args.func = test_schedule_single;
> + execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_one_many_hi(void)
> +{
> + test_case_args_t args;
> + snprintf(args.name, sizeof(args.name), "sched_one_m_hi");
> + args.prio = ODP_SCHED_PRIO_HIGHEST;
> + args.func = test_schedule_one_many;
> + execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_many_hi(void)
> +{
> + test_case_args_t args;
> + snprintf(args.name, sizeof(args.name), "sched_____m_hi");
> + args.prio = ODP_SCHED_PRIO_HIGHEST;
> + args.func = test_schedule_many;
> + execute_parallel(exec_template, &args);
> +}
> +
> +static void schedule_multi_hi(void)
> +{
> + test_case_args_t args;
IMO, We should avoid using variables allocated from stack to share
the data between different execution context.
It will work in linux threaded run-time environment.But it will
have issues in different run-time environment like bare-metal.
IMO if any memory needs to be shared across different
execution environment should use the memory allocated from odp shared mem.
> + snprintf(args.name, sizeof(args.name), "sched_multi_hi");
> + args.prio = ODP_SCHED_PRIO_HIGHEST;
> + args.func = test_schedule_multi;
> + execute_parallel(exec_template, &args);
> +}
> +
> +static void execute_parallel(void *(*start_routine) (void *),
> + test_case_args_t *test_case_args)
> +{
> + odph_linux_pthread_t thread_tbl[MAX_WORKERS];
> + int first_core;
> +
> + memset(thread_tbl, 0, sizeof(thread_tbl));
> +
> + /*
> + * By default core #0 runs Linux kernel background tasks.
> + * Start mapping thread from core #1
> + */
> + first_core = 1;
> +
> + if (odp_sys_core_count() == 1)
> + first_core = 0;
> +
> + odph_linux_pthread_create(thread_tbl, num_workers, first_core,
> + start_routine, test_case_args);
> +
> + /* Wait for worker threads to terminate */
> + odph_linux_pthread_join(thread_tbl, num_workers);
> +}
> +
> +static odp_buffer_pool_t test_odp_buffer_pool_init(void)
> +{
> + void *pool_base;
> + odp_shm_t shm;
> + odp_buffer_pool_t pool;
> +
> + shm = odp_shm_reserve("msg_pool",
> + MSG_POOL_SIZE, ODP_CACHE_LINE_SIZE, 0);
> +
> + pool_base = odp_shm_addr(shm);
> +
> + if (NULL == pool_base) {
> + printf("Shared memory reserve failed.\n");
> + return -1;
> + }
> +
> + pool = odp_buffer_pool_create("msg_pool", pool_base, MSG_POOL_SIZE,
> + BUF_SIZE, ODP_CACHE_LINE_SIZE,
> + ODP_BUFFER_TYPE_RAW);
> +
> + if (ODP_BUFFER_POOL_INVALID == pool) {
> + printf("Pool create failed.\n");
> + return -1;
> + }
> + return pool;
> +}
> +
> +int schedule_test_init(void)
> +{
> + test_args_t args;
> + odp_shm_t shm;
> + test_globals_t *globals;
> + int i, j;
> + int prios;
> +
> + if (0 != odp_init_global(NULL, NULL)) {
> + printf("odp_init_global fail.\n");
> + return -1;
> + }
> + if (0 != odp_init_local()) {
> + printf("odp_init_local fail.\n");
> + return -1;
> + }
> + if (ODP_BUFFER_POOL_INVALID == test_odp_buffer_pool_init()) {
> + printf("test_odp_buffer_pool_init fail.\n");
> + return -1;
> + }
> +
> + /* A worker thread per core */
> + num_workers = odp_sys_core_count();
> +
> + if (args.core_count)
args.core_count is uninitialized
> + num_workers = args.core_count;
> +
> + /* force to max core count */
> + if (num_workers > MAX_WORKERS)
> + num_workers = MAX_WORKERS;
> + shm = odp_shm_reserve("test_globals",
> + sizeof(test_globals_t), ODP_CACHE_LINE_SIZE, 0);
> +
> + globals = odp_shm_addr(shm);
> +
> + if (globals == NULL) {
> + ODP_ERR("Shared memory reserve failed.\n");
> + return -1;
> + }
> +
> + memset(globals, 0, sizeof(test_globals_t));
> +
> + /* Barrier to sync test case execution */
> + odp_barrier_init_count(&globals->barrier, num_workers);
> +
> + prios = odp_schedule_num_prio();
> +
> + for (i = 0; i < prios; i++) {
> + odp_queue_param_t param;
> + odp_queue_t queue;
> + char name[] = "sched_XX_YY";
> +
> + if (i != ODP_SCHED_PRIO_HIGHEST &&
> + i != ODP_SCHED_PRIO_LOWEST)
> + continue;
> +
> + name[6] = '0' + i/10;
> + name[7] = '0' + i - 10*(i/10);
> +
> + param.sched.prio = i;
> + param.sched.sync = ODP_SCHED_SYNC_ATOMIC;
> + param.sched.group = ODP_SCHED_GROUP_DEFAULT;
> +
> + for (j = 0; j < QUEUES_PER_PRIO; j++) {
> + name[9] = '0' + j/10;
> + name[10] = '0' + j - 10*(j/10);
> +
> + queue = odp_queue_create(name, ODP_QUEUE_TYPE_SCHED,
> + ¶m);
> +
> + if (queue == ODP_QUEUE_INVALID) {
> + ODP_ERR("Schedule queue create failed.\n");
> + return -1;
> + }
> + }
> + }
> + return 0;
> +}
> +
> +int schedule_test_finalize(void)
> +{
> + odp_term_local();
> + odp_term_global();
> + return 0;
> +}
> +
> +struct CU_TestInfo schedule_tests[] = {
> + _CU_TEST_INFO(schedule_wait_time),
> + _CU_TEST_INFO(schedule_one_single_lo),
> + _CU_TEST_INFO(schedule_single_lo),
> + _CU_TEST_INFO(schedule_one_many_lo),
> + _CU_TEST_INFO(schedule_many_lo),
> + _CU_TEST_INFO(schedule_multi_lo),
> + _CU_TEST_INFO(schedule_one_single_hi),
> + _CU_TEST_INFO(schedule_single_hi),
> + _CU_TEST_INFO(schedule_one_many_hi),
> + _CU_TEST_INFO(schedule_many_hi),
> + _CU_TEST_INFO(schedule_multi_hi),
> + CU_TEST_INFO_NULL,
> +};
> diff --git a/test/cunit/schedule/odp_schedule_testsuites.c
> b/test/cunit/schedule/odp_schedule_testsuites.c
> new file mode 100644
> index 0000000..1053069
> --- /dev/null
> +++ b/test/cunit/schedule/odp_schedule_testsuites.c
> @@ -0,0 +1,35 @@
> +/* Copyright (c) 2014, Linaro Limited
> + * All rights reserved.
> + *
> + * SPDX-License-Identifier: BSD-3-Clause
> + */
> +
> +#include "odp_schedule_testsuites.h"
> +
> +static CU_SuiteInfo suites[] = {
> + {
> + "Scheduler tests" ,
> + schedule_test_init,
> + schedule_test_finalize,
> + NULL,
> + NULL,
> + schedule_tests
> + },
> + CU_SUITE_INFO_NULL,
> +};
> +
> +int main(void)
> +{
> + /* initialize the CUnit test registry */
> + if (CUE_SUCCESS != CU_initialize_registry())
> + return CU_get_error();
> +
> + /* register suites */
> + CU_register_suites(suites);
> + /* Run all tests using the CUnit Basic interface */
> + CU_basic_set_mode(CU_BRM_VERBOSE);
> + CU_basic_run_tests();
> + CU_cleanup_registry();
> +
> + return CU_get_error();
> +}
> diff --git a/test/cunit/schedule/odp_schedule_testsuites.h
> b/test/cunit/schedule/odp_schedule_testsuites.h
> new file mode 100644
> index 0000000..67a2a69
> --- /dev/null
> +++ b/test/cunit/schedule/odp_schedule_testsuites.h
> @@ -0,0 +1,21 @@
> +/* Copyright (c) 2014, Linaro Limited
> + * All rights reserved.
> + *
> + * SPDX-License-Identifier: BSD-3-Clause
> + */
> +
> +#ifndef ODP_SCHEDULE_TESTSUITES_H_
> +#define ODP_SCHEDULE_TESTSUITES_H_
> +
> +#include "odp.h"
> +#include <CUnit/Basic.h>
> +
> +/* Helper macro for CU_TestInfo initialization */
> +#define _CU_TEST_INFO(test_func) {#test_func, test_func}
> +
> +extern struct CU_TestInfo schedule_tests[];
> +
> +extern int schedule_test_init(void);
> +extern int schedule_test_finalize(void);
> +
> +#endif /* ODP_SCHEDULE_TESTSUITES_H_ */
> --
> 1.8.3.2
>
>
> _______________________________________________
> lng-odp mailing list
> [email protected]
> http://lists.linaro.org/mailman/listinfo/lng-odp
_______________________________________________
lng-odp mailing list
[email protected]
http://lists.linaro.org/mailman/listinfo/lng-odp