On Wed, Dec 10, 2014 at 07:12:43PM +0200, Ciprian Barbu wrote:
> Signed-off-by: Ciprian Barbu <ciprian.ba...@linaro.org>
> ---
> v3:
>  - changes after Mike's review
>  - removed duplicate check of end of test in schedule_common_
> v2:
>  - rebased against ODP tip
>  - fixed some bugs
>  - added some defines to clearly see the testcase parameters
> 
>  test/validation/Makefile.am    |   5 +-
>  test/validation/odp_schedule.c | 608 
> +++++++++++++++++++++++++++++++++++++++++


Add odp_schedule in .gitignore


>  2 files changed, 612 insertions(+), 1 deletion(-)
>  create mode 100644 test/validation/odp_schedule.c
> 
> diff --git a/test/validation/Makefile.am b/test/validation/Makefile.am
> index 8547085..3670c76 100644
> --- a/test/validation/Makefile.am
> +++ b/test/validation/Makefile.am
> @@ -6,13 +6,15 @@ AM_LDFLAGS += -static
>  if ODP_CUNIT_ENABLED
>  TESTS = ${bin_PROGRAMS}
>  check_PROGRAMS = ${bin_PROGRAMS}
> -bin_PROGRAMS = odp_init odp_queue odp_crypto odp_shm
> +bin_PROGRAMS = odp_init odp_queue odp_crypto odp_shm odp_schedule
>  odp_init_LDFLAGS = $(AM_LDFLAGS)
>  odp_queue_LDFLAGS = $(AM_LDFLAGS)
>  odp_crypto_CFLAGS = $(AM_CFLAGS) -I$(srcdir)/crypto
>  odp_crypto_LDFLAGS = $(AM_LDFLAGS)
>  odp_shm_CFLAGS = $(AM_CFLAGS)
>  odp_shm_LDFLAGS = $(AM_LDFLAGS)
> +odp_schedule_CFLAGS = $(AM_CFLAGS)
> +odp_schedule_LDFLAGS = $(AM_LDFLAGS)
>  endif
>  
>  dist_odp_init_SOURCES = odp_init.c
> @@ -22,3 +24,4 @@ dist_odp_crypto_SOURCES = 
> crypto/odp_crypto_test_async_inp.c \
>                         crypto/odp_crypto_test_rng.c \
>                         odp_crypto.c common/odp_cunit_common.c
>  dist_odp_shm_SOURCES = odp_shm.c common/odp_cunit_common.c
> +dist_odp_schedule_SOURCES = odp_schedule.c common/odp_cunit_common.c
> diff --git a/test/validation/odp_schedule.c b/test/validation/odp_schedule.c
> new file mode 100644
> index 0000000..faca497
> --- /dev/null
> +++ b/test/validation/odp_schedule.c
> @@ -0,0 +1,608 @@
> +/* Copyright (c) 2014, Linaro Limited
> + * All rights reserved.
> + *
> + * SPDX-License-Identifier:     BSD-3-Clause
> + */
> +
> +#include <odp.h>
> +#include "odp_cunit_common.h"
> +
> +#define MAX_WORKERS_THREADS  32
> +#define MSG_POOL_SIZE                (4*1024*1024)
> +#define QUEUES_PER_PRIO              16
> +#define BUF_SIZE             64
> +#define TEST_NUM_BUFS                100
> +#define BURST_BUF_SIZE               4
> +#define TEST_NUM_BUFS_EXCL   10000
> +
> +#define GLOBALS_SHM_NAME     "test_globals"
> +#define MSG_POOL_NAME                "msg_pool"
> +#define SHM_MSG_POOL_NAME    "shm_msg_pool"
> +#define SHM_THR_ARGS_NAME    "shm_thr_args"
> +
> +#define ONE_Q                        1
> +#define MANY_QS                      QUEUES_PER_PRIO
> +
> +#define ONE_PRIO             1
> +
> +#define SCHD_ONE             0
> +#define SCHD_MULTI           1
> +
> +#define DISABLE_EXCL_ATOMIC  0
> +#define ENABLE_EXCL_ATOMIC   1
> +
> +
> +/* Test global variables */
> +typedef struct {
> +     int core_count;
> +     odp_barrier_t barrier;
> +     odp_schedule_prio_t current_prio;
> +     int prio_buf_count;
> +     odp_spinlock_t count_lock;
> +     odp_spinlock_t atomic_lock;
> +} test_globals_t;
> +
> +typedef struct ODP_PACKED {
> +     pthrd_arg thrdarg;
> +     odp_schedule_sync_t sync;
> +     int num_queues;
> +     int num_prio;
> +     int num_bufs;
> +     int num_cores;
> +     int enable_schd_multi;
> +     int enable_excl_atomic;
> +} thread_args_t;
> +
> +odp_buffer_pool_t pool;
> +
> +static void test_schedule_wait_time(void)
> +{
> +     uint64_t wait_time;
> +
> +     wait_time = odp_schedule_wait_time(0);
> +     CU_ASSERT(wait_time > 0);

IMO, Its an invalid assert as odp_schedule_wait_time(0) very well can be 0
and wait_time is unsigned.

> +
> +     wait_time = odp_schedule_wait_time(1);
> +     CU_ASSERT(wait_time > 0);
> +
> +     wait_time = odp_schedule_wait_time((uint64_t)-1LL);
> +     CU_ASSERT(wait_time > 0);
> +}
> +
> +static void test_schedule_num_prio(void)
> +{
> +     int prio;
> +
> +     prio = odp_schedule_num_prio();
> +
> +     CU_ASSERT(prio > 0);
> +     CU_ASSERT(prio == odp_schedule_num_prio());
> +}
> +
> +static void *schedule_common_(void *arg)
> +{
> +     thread_args_t *args = (thread_args_t *)arg;
> +     odp_schedule_sync_t sync;
> +     int num_queues, num_prio, num_bufs, num_cores;
> +     odp_shm_t shm;
> +     test_globals_t *globals;
> +
> +     sync = args->sync;
> +     num_queues = args->num_queues;
> +     num_prio = args->num_prio;
> +     num_bufs = args->num_bufs;
> +     num_cores = args->num_cores;
> +
> +     shm = odp_shm_lookup(GLOBALS_SHM_NAME);
> +     CU_ASSERT_FATAL(shm != ODP_SHM_INVALID);
> +     globals = odp_shm_addr(shm);
> +     CU_ASSERT_FATAL(globals != NULL);
> +
> +
> +     if (num_cores == globals->core_count)
> +             odp_barrier_wait(&globals->barrier);
> +
> +     while (1) {
> +             odp_buffer_t buf;
> +             odp_queue_t from;
> +             int num = 0;
> +             int locked;
> +
> +             odp_spinlock_lock(&globals->count_lock);
> +             if (globals->prio_buf_count ==
> +                 num_bufs * num_queues * num_prio) {
> +                     odp_spinlock_unlock(&globals->count_lock);
> +                     break;
> +             }
> +             odp_spinlock_unlock(&globals->count_lock);

IMO, the above "if" check need not to be in critical section for this 
application logic.

Actually using spinlock in this specific application logic is not good idea
as "globals->prio_buf_count += num" below(in critical section) will starve badly
when globals->prio_buf_count reaching to num. Based on my testing it will be 
prominent when
the test run on >21 cores(on octeon), Its like when  globals->prio_buf_count 
reaches num - 1
1 core will be trying to acquire the below "globals->prio_buf_count += num"
critical section lock and rest n-1 cores will be  busy in not letting the 1 core
to provide the below critical section for no reason.(1 vs n-1)

If you have any _specific_ reason for the above "if" needs to be in critical 
section then use
ticket lock, it will scale properly else remove the critical section above.

> +
> +             if (args->enable_schd_multi) {
> +                     odp_buffer_t bufs[BURST_BUF_SIZE];
> +                     int j;
> +                     num = odp_schedule_multi(&from, ODP_SCHED_NO_WAIT, bufs,
> +                                              BURST_BUF_SIZE);
> +                     CU_ASSERT(num >= 0);
> +                     CU_ASSERT(num <= BURST_BUF_SIZE);
> +                     if (num == 0)
> +                             continue;
> +                     for (j = 0; j < num; j++)
> +                             odp_buffer_free(bufs[j]);
> +             } else {
> +                     buf = odp_schedule(&from, ODP_SCHED_NO_WAIT);
> +                     if (buf == ODP_BUFFER_INVALID)
> +                             continue;
> +                     num = 1;
> +                     odp_buffer_free(buf);
> +             }
> +
> +             if (args->enable_excl_atomic) {
> +                     locked = odp_spinlock_trylock(&globals->atomic_lock);
> +                     CU_ASSERT(locked == 1);
> +                     CU_ASSERT(from != ODP_QUEUE_INVALID);
> +                     if (locked) {
> +                             int cnt;
> +                             uint64_t cycles = 0;
> +                             /* Do some work here to keep the thread busy */
> +                             for (cnt = 0; cnt < 1000; cnt++)
> +                                     cycles += odp_time_cycles();
> +
> +                             odp_spinlock_unlock(&globals->atomic_lock);
> +                     }
> +             }
> +
> +             odp_spinlock_lock(&globals->count_lock);
> +             globals->prio_buf_count += num;
> +
> +             if (sync == ODP_SCHED_SYNC_ATOMIC)
> +                     odp_schedule_release_atomic();
> +
> +             odp_spinlock_unlock(&globals->count_lock);
> +     }
> +
> +     return NULL;
> +}
> +
> +static void fill_queues(thread_args_t *args)
> +{
> +     odp_schedule_sync_t sync;
> +     int num_queues, num_prio;
> +     odp_buffer_pool_t pool;
> +     int i, j, k;
> +     char name[32];
> +
> +     sync = args->sync;
> +     num_queues = args->num_queues;
> +     num_prio = args->num_prio;
> +
> +     pool = odp_buffer_pool_lookup(MSG_POOL_NAME);
> +     CU_ASSERT_FATAL(pool != ODP_BUFFER_POOL_INVALID);
> +
> +     for (i = 0; i < num_prio; i++) {
> +             for (j = 0; j < num_queues; j++) {
> +                     odp_queue_t queue;
> +
> +                     switch (sync) {
> +                     case ODP_SCHED_SYNC_NONE:
> +                             snprintf(name, sizeof(name),
> +                                      "sched_%d_%d_n", i, j);
> +                             break;
> +                     case ODP_SCHED_SYNC_ATOMIC:
> +                             snprintf(name, sizeof(name),
> +                                      "sched_%d_%d_a", i, j);
> +                             break;
> +                     case ODP_SCHED_SYNC_ORDERED:
> +                             snprintf(name, sizeof(name),
> +                                      "sched_%d_%d_o", i, j);
> +                             break;
> +                     default:
> +                             CU_ASSERT(0);
> +                             break;
> +                     }
> +
> +                     queue = odp_queue_lookup(name);
> +                     CU_ASSERT_FATAL(queue != ODP_QUEUE_INVALID);
> +
> +                     for (k = 0; k < args->num_bufs; k++) {
> +                             odp_buffer_t buf;
> +                             buf = odp_buffer_alloc(pool);
> +                             CU_ASSERT(buf != ODP_BUFFER_INVALID);
> +                             CU_ASSERT(odp_queue_enq(queue, buf) == 0);
> +                     }
> +             }
> +     }
> +}
> +
> +static void schedule_common(odp_schedule_sync_t sync, int num_queues,
> +                         int num_prio, int enable_schd_multi)
> +{
> +     thread_args_t args;
> +     odp_shm_t shm;
> +     test_globals_t *globals;
> +
> +     shm = odp_shm_lookup(GLOBALS_SHM_NAME);
> +     CU_ASSERT_FATAL(shm != ODP_SHM_INVALID);
> +     globals = odp_shm_addr(shm);
> +     CU_ASSERT_FATAL(globals != NULL);
> +
> +     globals->current_prio = ODP_SCHED_PRIO_HIGHEST;
> +     globals->prio_buf_count = 0;
> +
> +     args.sync = sync;
> +     args.num_queues = num_queues;
> +     args.num_prio = num_prio;
> +     args.num_bufs = TEST_NUM_BUFS;
> +     args.num_cores = 1;
> +     args.enable_schd_multi = enable_schd_multi;
> +     args.enable_excl_atomic = 0;    /* Not needed with a single core */
> +
> +     fill_queues(&args);
> +
> +     schedule_common_(&args);
> +}
> +
> +static void parallel_execute(odp_schedule_sync_t sync, int num_queues,
> +                          int num_prio, int enable_schd_multi,
> +                          int enable_excl_atomic)
> +{
> +     odp_shm_t shm;
> +     test_globals_t *globals;
> +     thread_args_t *thr_args;
> +
> +     shm = odp_shm_lookup(GLOBALS_SHM_NAME);
> +     CU_ASSERT_FATAL(shm != ODP_SHM_INVALID);
> +     globals = odp_shm_addr(shm);
> +     CU_ASSERT_FATAL(globals != NULL);
> +
> +     shm = odp_shm_lookup(SHM_THR_ARGS_NAME);
> +     CU_ASSERT_FATAL(shm != ODP_SHM_INVALID);
> +     thr_args = odp_shm_addr(shm);
> +     CU_ASSERT_FATAL(thr_args != NULL);
> +
> +     thr_args->sync = sync;
> +     thr_args->num_queues = num_queues;
> +     thr_args->num_prio = num_prio;
> +     if (enable_excl_atomic)
> +             thr_args->num_bufs = TEST_NUM_BUFS_EXCL;
> +     else
> +             thr_args->num_bufs = TEST_NUM_BUFS;
> +     thr_args->num_cores = globals->core_count;
> +     thr_args->enable_schd_multi = enable_schd_multi;
> +     thr_args->enable_excl_atomic = enable_excl_atomic;
> +
> +     fill_queues(thr_args);
> +
> +     /* Reset buffer counters from the main thread */
> +     globals->current_prio = ODP_SCHED_PRIO_HIGHEST;
> +     globals->prio_buf_count = 0;
> +
> +     /* Create and launch worker threads */
> +     thr_args->thrdarg.numthrds = globals->core_count;
> +     odp_cunit_thread_create(schedule_common_, &thr_args->thrdarg);
> +
> +     /* Wait for worker threads to terminate */
> +     odp_cunit_thread_exit(&thr_args->thrdarg);
> +}
> +
> +/* 1 queue 1 thread ODP_SCHED_SYNC_NONE */
> +static void test_schedule_1q_1t_n(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_NONE, ONE_Q, ONE_PRIO, SCHD_ONE);
> +}
> +
> +/* 1 queue 1 thread ODP_SCHED_SYNC_ATOMIC */
> +static void test_schedule_1q_1t_a(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_ATOMIC, ONE_Q, ONE_PRIO, SCHD_ONE);
> +}
> +
> +/* 1 queue 1 thread ODP_SCHED_SYNC_ORDERED */
> +static void test_schedule_1q_1t_o(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_ORDERED, ONE_Q, ONE_PRIO, SCHD_ONE);
> +}
> +
> +/* Many queues 1 thread ODP_SCHED_SYNC_NONE */
> +static void test_schedule_mq_1t_n(void)
> +{
> +     /* Only one priority involved in these tests, but use
> +        the same number of queues the more general case uses */
> +     schedule_common(ODP_SCHED_SYNC_NONE, MANY_QS, ONE_PRIO, SCHD_ONE);
> +}
> +
> +/* Many queues 1 thread ODP_SCHED_SYNC_ATOMIC */
> +static void test_schedule_mq_1t_a(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_ATOMIC, MANY_QS, ONE_PRIO, SCHD_ONE);
> +}
> +
> +/* Many queues 1 thread ODP_SCHED_SYNC_ORDERED */
> +static void test_schedule_mq_1t_o(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_ORDERED, MANY_QS, ONE_PRIO, SCHD_ONE);
> +}
> +
> +/* Many queues 1 thread check priority ODP_SCHED_SYNC_NONE */
> +static void test_schedule_mq_1t_prio_n(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     schedule_common(ODP_SCHED_SYNC_NONE, MANY_QS, prio, SCHD_ONE);
> +}
> +
> +/* Many queues 1 thread check priority ODP_SCHED_SYNC_ATOMIC */
> +static void test_schedule_mq_1t_prio_a(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     schedule_common(ODP_SCHED_SYNC_ATOMIC, MANY_QS, prio, SCHD_ONE);
> +}
> +
> +/* Many queues 1 thread check priority ODP_SCHED_SYNC_ORDERED */
> +static void test_schedule_mq_1t_prio_o(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     schedule_common(ODP_SCHED_SYNC_ORDERED, MANY_QS, prio, SCHD_ONE);
> +}
> +
> +/* Many queues many threads check priority ODP_SCHED_SYNC_NONE */
> +static void test_schedule_mq_mt_prio_n(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     parallel_execute(ODP_SCHED_SYNC_NONE, MANY_QS, prio, SCHD_ONE,
> +                      DISABLE_EXCL_ATOMIC);
> +}
> +
> +/* Many queues many threads check priority ODP_SCHED_SYNC_ATOMIC */
> +static void test_schedule_mq_mt_prio_a(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     parallel_execute(ODP_SCHED_SYNC_ATOMIC, MANY_QS, prio, SCHD_ONE,
> +                      DISABLE_EXCL_ATOMIC);
> +}
> +
> +/* Many queues many threads check priority ODP_SCHED_SYNC_ORDERED */
> +static void test_schedule_mq_mt_prio_o(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     parallel_execute(ODP_SCHED_SYNC_ORDERED, MANY_QS, prio, SCHD_ONE,
> +                      DISABLE_EXCL_ATOMIC);
> +}
> +
> +/* 1 queue many threads check exclusive access on ATOMIC queues */
> +static void test_schedule_1q_mt_a_excl(void)
> +{
> +     parallel_execute(ODP_SCHED_SYNC_ATOMIC, ONE_Q, ONE_PRIO, SCHD_ONE,
> +                      ENABLE_EXCL_ATOMIC);
> +}
> +
> +/* 1 queue 1 thread ODP_SCHED_SYNC_NONE multi */
> +static void test_schedule_multi_1q_1t_n(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_NONE, ONE_Q, ONE_PRIO, SCHD_MULTI);
> +}
> +
> +/* 1 queue 1 thread ODP_SCHED_SYNC_ATOMIC multi */
> +static void test_schedule_multi_1q_1t_a(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_ATOMIC, ONE_Q, ONE_PRIO, SCHD_MULTI);
> +}
> +
> +/* 1 queue 1 thread ODP_SCHED_SYNC_ORDERED multi */
> +static void test_schedule_multi_1q_1t_o(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_ORDERED, ONE_Q, ONE_PRIO, SCHD_MULTI);
> +}
> +
> +/* Many queues 1 thread ODP_SCHED_SYNC_NONE multi */
> +static void test_schedule_multi_mq_1t_n(void)
> +{
> +     /* Only one priority involved in these tests, but use
> +        the same number of queues the more general case uses */
> +     schedule_common(ODP_SCHED_SYNC_NONE, MANY_QS, ONE_PRIO, SCHD_MULTI);
> +}
> +
> +/* Many queues 1 thread ODP_SCHED_SYNC_ATOMIC multi */
> +static void test_schedule_multi_mq_1t_a(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_ATOMIC, MANY_QS, ONE_PRIO, SCHD_MULTI);
> +}
> +
> +/* Many queues 1 thread ODP_SCHED_SYNC_ORDERED multi */
> +static void test_schedule_multi_mq_1t_o(void)
> +{
> +     schedule_common(ODP_SCHED_SYNC_ORDERED, MANY_QS, ONE_PRIO, SCHD_MULTI);
> +}
> +
> +/* Many queues 1 thread check priority ODP_SCHED_SYNC_NONE multi */
> +static void test_schedule_multi_mq_1t_prio_n(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     schedule_common(ODP_SCHED_SYNC_NONE, MANY_QS, prio, SCHD_MULTI);
> +}
> +
> +/* Many queues 1 thread check priority ODP_SCHED_SYNC_ATOMIC multi */
> +static void test_schedule_multi_mq_1t_prio_a(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     schedule_common(ODP_SCHED_SYNC_ATOMIC, MANY_QS, prio, SCHD_MULTI);
> +}
> +
> +/* Many queues 1 thread check priority ODP_SCHED_SYNC_ORDERED multi */
> +static void test_schedule_multi_mq_1t_prio_o(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     schedule_common(ODP_SCHED_SYNC_ORDERED, MANY_QS, prio, SCHD_MULTI);
> +}
> +
> +/* Many queues many threads check priority ODP_SCHED_SYNC_NONE multi */
> +static void test_schedule_multi_mq_mt_prio_n(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     parallel_execute(ODP_SCHED_SYNC_NONE, MANY_QS, prio, SCHD_MULTI, 0);
> +}
> +
> +/* Many queues many threads check priority ODP_SCHED_SYNC_ATOMIC multi */
> +static void test_schedule_multi_mq_mt_prio_a(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     parallel_execute(ODP_SCHED_SYNC_ATOMIC, MANY_QS, prio, SCHD_MULTI, 0);
> +}
> +
> +/* Many queues many threads check priority ODP_SCHED_SYNC_ORDERED multi */
> +static void test_schedule_multi_mq_mt_prio_o(void)
> +{
> +     int prio = odp_schedule_num_prio();
> +     parallel_execute(ODP_SCHED_SYNC_ORDERED, MANY_QS, prio, SCHD_MULTI, 0);
> +}
> +
> +/* 1 queue many threads check exclusive access on ATOMIC queues multi */
> +static void test_schedule_multi_1q_mt_a_excl(void)
> +{
> +     parallel_execute(ODP_SCHED_SYNC_ATOMIC, ONE_Q, ONE_PRIO, SCHD_MULTI,
> +                      ENABLE_EXCL_ATOMIC);
> +}
> +
> +static int create_queues(void)
> +{
> +     int i, j, prios;
> +
> +     prios = odp_schedule_num_prio();
> +
> +     for (i = 0; i < prios; i++) {
> +             odp_queue_param_t p;
> +             p.sched.prio  = i;
> +             p.sched.group = ODP_SCHED_GROUP_DEFAULT;
> +
> +             for (j = 0; j < QUEUES_PER_PRIO; j++) {
> +                     /* Per sched sync type */
> +                     char name[32];
> +                     odp_queue_t q;
> +
> +                     snprintf(name, sizeof(name), "sched_%d_%d_n", i, j);
> +                     p.sched.sync = ODP_SCHED_SYNC_NONE;
> +                     q = odp_queue_create(name, ODP_QUEUE_TYPE_SCHED, &p);
> +
> +                     if (q == ODP_QUEUE_INVALID) {
> +                             printf("Schedule queue create failed.\n");
> +                             return -1;
> +                     }
> +
> +                     snprintf(name, sizeof(name), "sched_%d_%d_a", i, j);
> +                     p.sched.sync = ODP_SCHED_SYNC_ATOMIC;
> +                     q = odp_queue_create(name, ODP_QUEUE_TYPE_SCHED, &p);
> +
> +                     if (q == ODP_QUEUE_INVALID) {
> +                             printf("Schedule queue create failed.\n");
> +                             return -1;
> +                     }
> +
> +                     snprintf(name, sizeof(name), "sched_%d_%d_o", i, j);
> +                     p.sched.sync = ODP_SCHED_SYNC_ORDERED;
> +                     q = odp_queue_create(name, ODP_QUEUE_TYPE_SCHED, &p);
> +
> +                     if (q == ODP_QUEUE_INVALID) {
> +                             printf("Schedule queue create failed.\n");
> +                             return -1;
> +                     }
> +             }
> +     }
> +
> +     return 0;
> +}
> +
> +int tests_global_init(void)
> +{
> +     odp_shm_t shm;
> +     void *pool_base;
> +     odp_buffer_pool_t pool;
> +     test_globals_t *globals;
> +     thread_args_t *thr_args;
> +
> +     shm = odp_shm_reserve(SHM_MSG_POOL_NAME, MSG_POOL_SIZE,
> +                           ODP_CACHE_LINE_SIZE, 0);
> +     pool_base = odp_shm_addr(shm);
> +     if (pool_base == NULL) {
> +             printf("Shared memory reserve failed.\n");
> +             return -1;
> +     }
> +
> +     pool = odp_buffer_pool_create(MSG_POOL_NAME, pool_base, MSG_POOL_SIZE,
> +                                   BUF_SIZE, ODP_CACHE_LINE_SIZE,
> +                                   ODP_BUFFER_TYPE_RAW);
> +     if (pool == ODP_BUFFER_POOL_INVALID) {
> +             printf("Pool creation failed (msg).\n");
> +             return -1;
> +     }
> +
> +     shm = odp_shm_reserve(GLOBALS_SHM_NAME,
> +                           sizeof(test_globals_t), ODP_CACHE_LINE_SIZE, 0);
> +
> +     globals = odp_shm_addr(shm);
> +
> +     if (globals == NULL) {
> +             printf("Shared memory reserve failed (globals).\n");
> +             return -1;
> +     }
> +
> +     memset(globals, 0, sizeof(test_globals_t));
> +
> +     globals->core_count = odp_sys_core_count();
> +     if (globals->core_count > MAX_WORKERS)
> +             globals->core_count = MAX_WORKERS;
> +
> +     shm = odp_shm_reserve(SHM_THR_ARGS_NAME, sizeof(thread_args_t),
> +                           ODP_CACHE_LINE_SIZE, 0);
> +     thr_args = odp_shm_addr(shm);
> +
> +     if (thr_args == NULL) {
> +             printf("Shared memory reserve failed (thr_args).\n");
> +             return -1;
> +     }
> +
> +     memset(thr_args, 0, sizeof(thread_args_t));
> +
> +     /* Barrier to sync test case execution */
> +     odp_barrier_init(&globals->barrier, globals->core_count);
> +     odp_spinlock_init(&globals->count_lock);
> +     odp_spinlock_init(&globals->atomic_lock);
> +
> +     if (!create_queues())
> +             return -1;

Broken logic, create_queues() returns 0 for success case.

You can add my Reviewed and Tested by: after fixing the above issues.
Reviewed-and-Tested-by: Jerin Jacob <jerin.ja...@caviumnetworks.com>

Its a good testsuite, I have verified it on octeon.

> +
> +     return 0;
> +}
> +
> +struct CU_TestInfo test_odp_schedule[] = {
> +     {"schedule_wait_time",          test_schedule_wait_time},
> +     {"schedule_num_prio",           test_schedule_num_prio},
> +     {"schedule_1q_1t_n",            test_schedule_1q_1t_n},
> +     {"schedule_1q_1t_a",            test_schedule_1q_1t_a},
> +     {"schedule_1q_1t_o",            test_schedule_1q_1t_o},
> +     {"schedule_mq_1t_n",            test_schedule_mq_1t_n},
> +     {"schedule_mq_1t_a",            test_schedule_mq_1t_a},
> +     {"schedule_mq_1t_o",            test_schedule_mq_1t_o},
> +     {"schedule_mq_1t_prio_n",       test_schedule_mq_1t_prio_n},
> +     {"schedule_mq_1t_prio_a",       test_schedule_mq_1t_prio_a},
> +     {"schedule_mq_1t_prio_o",       test_schedule_mq_1t_prio_o},
> +     {"schedule_mq_mt_prio_n",       test_schedule_mq_mt_prio_n},
> +     {"schedule_mq_mt_prio_a",       test_schedule_mq_mt_prio_a},
> +     {"schedule_mq_mt_prio_o",       test_schedule_mq_mt_prio_o},
> +     {"schedule_1q_mt_a_excl",       test_schedule_1q_mt_a_excl},
> +     {"schedule_multi_1q_1t_n",      test_schedule_multi_1q_1t_n},
> +     {"schedule_multi_1q_1t_a",      test_schedule_multi_1q_1t_a},
> +     {"schedule_multi_1q_1t_o",      test_schedule_multi_1q_1t_o},
> +     {"schedule_multi_mq_1t_n",      test_schedule_multi_mq_1t_n},
> +     {"schedule_multi_mq_1t_a",      test_schedule_multi_mq_1t_a},
> +     {"schedule_multi_mq_1t_o",      test_schedule_multi_mq_1t_o},
> +     {"schedule_multi_mq_1t_prio_n", test_schedule_multi_mq_1t_prio_n},
> +     {"schedule_multi_mq_1t_prio_a", test_schedule_multi_mq_1t_prio_a},
> +     {"schedule_multi_mq_1t_prio_o", test_schedule_multi_mq_1t_prio_o},
> +     {"schedule_multi_mq_mt_prio_n", test_schedule_multi_mq_mt_prio_n},
> +     {"schedule_multi_mq_mt_prio_a", test_schedule_multi_mq_mt_prio_a},
> +     {"schedule_multi_mq_mt_prio_o", test_schedule_multi_mq_mt_prio_o},
> +     {"schedule_multi_1q_mt_a_excl", test_schedule_multi_1q_mt_a_excl},
> +     CU_TEST_INFO_NULL,
> +};
> +
> +CU_SuiteInfo odp_testsuites[] = {
> +     {"Scheduler", NULL, NULL, NULL, NULL, test_odp_schedule},
> +     CU_SUITE_INFO_NULL,
> +};
> -- 
> 1.8.3.2
> 
> 
> _______________________________________________
> lng-odp mailing list
> lng-odp@lists.linaro.org
> http://lists.linaro.org/mailman/listinfo/lng-odp

_______________________________________________
lng-odp mailing list
lng-odp@lists.linaro.org
http://lists.linaro.org/mailman/listinfo/lng-odp

Reply via email to