Not sure why this is needed, but:

Reviewed-by: Bill Fischofer <[email protected]>

On Mon, Mar 23, 2015 at 8:04 AM, Maxim Uvarov <[email protected]>
wrote:

> global_mem is global static variable usage same name variable
> inside function is confusing. Rename local vars from global_mem
> to mem.
>
> Signed-off-by: Maxim Uvarov <[email protected]>
> ---
>  test/validation/odp_synchronizers.c | 158
> ++++++++++++++++++------------------
>  1 file changed, 79 insertions(+), 79 deletions(-)
>
> diff --git a/test/validation/odp_synchronizers.c
> b/test/validation/odp_synchronizers.c
> index ab9164f..b8f4e6a 100644
> --- a/test/validation/odp_synchronizers.c
> +++ b/test/validation/odp_synchronizers.c
> @@ -107,7 +107,7 @@ static void thread_delay(per_thread_mem_t
> *per_thread_mem, uint32_t iterations)
>  /* Initialise per-thread memory */
>  static per_thread_mem_t *thread_init(void)
>  {
> -       global_shared_mem_t *global_mem;
> +       global_shared_mem_t *mem;
>         per_thread_mem_t *per_thread_mem;
>         odp_shm_t global_shm;
>         uint32_t per_thread_mem_len;
> @@ -122,10 +122,10 @@ static per_thread_mem_t *thread_init(void)
>         per_thread_mem->thread_core = odp_cpu_id();
>
>         global_shm = odp_shm_lookup(GLOBAL_SHM_NAME);
> -       global_mem = odp_shm_addr(global_shm);
> +       mem = odp_shm_addr(global_shm);
>         CU_ASSERT(global_mem != NULL);
>
> -       per_thread_mem->global_mem = global_mem;
> +       per_thread_mem->global_mem = mem;
>
>         return per_thread_mem;
>  }
> @@ -160,13 +160,13 @@ static void custom_barrier_wait(custom_barrier_t
> *custom_barrier)
>  static uint32_t barrier_test(per_thread_mem_t *per_thread_mem,
>                              odp_bool_t no_barrier_test)
>  {
> -       global_shared_mem_t *global_mem;
> +       global_shared_mem_t *mem;
>         uint32_t barrier_errs, iterations, cnt, i_am_slow_thread;
>         uint32_t thread_num, slow_thread_num, next_slow_thread,
> num_threads;
>         uint32_t lock_owner_delay, barrier_cnt1, barrier_cnt2;
>
>         thread_num = odp_cpu_id() + 1;
> -       global_mem = per_thread_mem->global_mem;
> +       mem = per_thread_mem->global_mem;
>         num_threads = global_mem->g_num_threads;
>         iterations = BARRIER_ITERATIONS;
>
> @@ -175,10 +175,10 @@ static uint32_t barrier_test(per_thread_mem_t
> *per_thread_mem,
>
>         for (cnt = 1; cnt < iterations; cnt++) {
>                 /* Wait here until all of the threads reach this point */
> -               custom_barrier_wait(&global_mem->custom_barrier1[cnt]);
> +               custom_barrier_wait(&mem->custom_barrier1[cnt]);
>
> -               barrier_cnt1 = global_mem->barrier_cnt1;
> -               barrier_cnt2 = global_mem->barrier_cnt2;
> +               barrier_cnt1 = mem->barrier_cnt1;
> +               barrier_cnt2 = mem->barrier_cnt2;
>
>                 if ((barrier_cnt1 != cnt) || (barrier_cnt2 != cnt)) {
>                         printf("thread_num=%u barrier_cnts of %u %u
> cnt=%u\n",
> @@ -187,9 +187,9 @@ static uint32_t barrier_test(per_thread_mem_t
> *per_thread_mem,
>                 }
>
>                 /* Wait here until all of the threads reach this point */
> -               custom_barrier_wait(&global_mem->custom_barrier2[cnt]);
> +               custom_barrier_wait(&mem->custom_barrier2[cnt]);
>
> -               slow_thread_num = global_mem->slow_thread_num;
> +               slow_thread_num = mem->slow_thread_num;
>                 i_am_slow_thread = thread_num == slow_thread_num;
>                 next_slow_thread = slow_thread_num + 1;
>                 if (num_threads < next_slow_thread)
> @@ -206,30 +206,30 @@ static uint32_t barrier_test(per_thread_mem_t
> *per_thread_mem,
>                 if (i_am_slow_thread) {
>                         thread_delay(per_thread_mem, lock_owner_delay);
>                         lock_owner_delay += BASE_DELAY;
> -                       if ((global_mem->barrier_cnt1 != cnt) ||
> -                           (global_mem->barrier_cnt2 != cnt) ||
> -                           (global_mem->slow_thread_num
> +                       if ((mem->barrier_cnt1 != cnt) ||
> +                           (mem->barrier_cnt2 != cnt) ||
> +                           (mem->slow_thread_num
>                                         != slow_thread_num))
>                                 barrier_errs++;
>                 }
>
>                 if (no_barrier_test == 0)
> -                       odp_barrier_wait(&global_mem->test_barriers[cnt]);
> +                       odp_barrier_wait(&mem->test_barriers[cnt]);
>
> -               global_mem->barrier_cnt1 = cnt + 1;
> +               mem->barrier_cnt1 = cnt + 1;
>                 odp_sync_stores();
>
>                 if (i_am_slow_thread) {
> -                       global_mem->slow_thread_num = next_slow_thread;
> -                       global_mem->barrier_cnt2 = cnt + 1;
> +                       mem->slow_thread_num = next_slow_thread;
> +                       mem->barrier_cnt2 = cnt + 1;
>                         odp_sync_stores();
>                 } else {
> -                       while (global_mem->barrier_cnt2 != (cnt + 1))
> +                       while (mem->barrier_cnt2 != (cnt + 1))
>                                 thread_delay(per_thread_mem, BASE_DELAY);
>                 }
>         }
>
> -       if ((global_mem->g_verbose) && (barrier_errs != 0))
> +       if ((mem->g_verbose) && (barrier_errs != 0))
>                 printf("\nThread %u (id=%d core=%d) had %u barrier_errs"
>                        " in %u iterations\n", thread_num,
>                         per_thread_mem->thread_id,
> @@ -293,14 +293,14 @@ static void spinlock_api_test(odp_spinlock_t
> *spinlock)
>
>  static void *spinlock_api_tests(void *arg UNUSED)
>  {
> -       global_shared_mem_t *global_mem;
> +       global_shared_mem_t *mem;
>         per_thread_mem_t *per_thread_mem;
>         odp_spinlock_t local_spin_lock;
>
>         per_thread_mem = thread_init();
> -       global_mem = per_thread_mem->global_mem;
> +       mem = per_thread_mem->global_mem;
>
> -       odp_barrier_wait(&global_mem->global_barrier);
> +       odp_barrier_wait(&mem->global_barrier);
>
>         spinlock_api_test(&local_spin_lock);
>         spinlock_api_test(&per_thread_mem->per_thread_spinlock);
> @@ -331,14 +331,14 @@ static void ticketlock_api_test(odp_ticketlock_t
> *ticketlock)
>
>  static void *ticketlock_api_tests(void *arg UNUSED)
>  {
> -       global_shared_mem_t *global_mem;
> +       global_shared_mem_t *mem;
>         per_thread_mem_t *per_thread_mem;
>         odp_ticketlock_t local_ticket_lock;
>
>         per_thread_mem = thread_init();
> -       global_mem = per_thread_mem->global_mem;
> +       mem = per_thread_mem->global_mem;
>
> -       odp_barrier_wait(&global_mem->global_barrier);
> +       odp_barrier_wait(&mem->global_barrier);
>
>         ticketlock_api_test(&local_ticket_lock);
>         ticketlock_api_test(&per_thread_mem->per_thread_ticketlock);
> @@ -365,14 +365,14 @@ static void rwlock_api_test(odp_rwlock_t *rw_lock)
>
>  static void *rwlock_api_tests(void *arg UNUSED)
>  {
> -       global_shared_mem_t *global_mem;
> +       global_shared_mem_t *mem;
>         per_thread_mem_t *per_thread_mem;
>         odp_rwlock_t local_rwlock;
>
>         per_thread_mem = thread_init();
> -       global_mem = per_thread_mem->global_mem;
> +       mem = per_thread_mem->global_mem;
>
> -       odp_barrier_wait(&global_mem->global_barrier);
> +       odp_barrier_wait(&mem->global_barrier);
>
>         rwlock_api_test(&local_rwlock);
>         rwlock_api_test(&per_thread_mem->per_thread_rwlock);
> @@ -384,17 +384,17 @@ static void *rwlock_api_tests(void *arg UNUSED)
>
>  static void *no_lock_functional_test(void *arg UNUSED)
>  {
> -       global_shared_mem_t *global_mem;
> +       global_shared_mem_t *mem;
>         per_thread_mem_t *per_thread_mem;
>         uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt;
>         uint32_t sync_failures, current_errs, lock_owner_delay;
>
>         thread_num = odp_cpu_id() + 1;
>         per_thread_mem = thread_init();
> -       global_mem = per_thread_mem->global_mem;
> -       iterations = global_mem->g_iterations;
> +       mem = per_thread_mem->global_mem;
> +       iterations = mem->g_iterations;
>
> -       odp_barrier_wait(&global_mem->global_barrier);
> +       odp_barrier_wait(&mem->global_barrier);
>
>         sync_failures = 0;
>         current_errs = 0;
> @@ -403,20 +403,20 @@ static void *no_lock_functional_test(void *arg
> UNUSED)
>         lock_owner_delay = BASE_DELAY;
>
>         for (cnt = 1; cnt <= iterations; cnt++) {
> -               global_mem->global_lock_owner = thread_num;
> +               mem->global_lock_owner = thread_num;
>                 odp_sync_stores();
>                 thread_delay(per_thread_mem, lock_owner_delay);
>
> -               if (global_mem->global_lock_owner != thread_num) {
> +               if (mem->global_lock_owner != thread_num) {
>                         current_errs++;
>                         sync_failures++;
>                 }
>
> -               global_mem->global_lock_owner = 0;
> +               mem->global_lock_owner = 0;
>                 odp_sync_stores();
>                 thread_delay(per_thread_mem, MIN_DELAY);
>
> -               if (global_mem->global_lock_owner == thread_num) {
> +               if (mem->global_lock_owner == thread_num) {
>                         current_errs++;
>                         sync_failures++;
>                 }
> @@ -430,10 +430,10 @@ static void *no_lock_functional_test(void *arg
> UNUSED)
>                 /* Try to resync all of the threads to increase contention
> */
>                 if ((rs_idx < NUM_RESYNC_BARRIERS) &&
>                     ((cnt % resync_cnt) == (resync_cnt - 1)))
> -
>  odp_barrier_wait(&global_mem->barrier_array[rs_idx++]);
> +                       odp_barrier_wait(&mem->barrier_array[rs_idx++]);
>         }
>
> -       if (global_mem->g_verbose)
> +       if (mem->g_verbose)
>                 printf("\nThread %u (id=%d core=%d) had %u sync_failures"
>                        " in %u iterations\n", thread_num,
>                        per_thread_mem->thread_id,
> @@ -454,7 +454,7 @@ static void *no_lock_functional_test(void *arg UNUSED)
>
>  static void *spinlock_functional_test(void *arg UNUSED)
>  {
> -       global_shared_mem_t *global_mem;
> +       global_shared_mem_t *mem;
>         per_thread_mem_t *per_thread_mem;
>         uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt;
>         uint32_t sync_failures, is_locked_errs, current_errs;
> @@ -462,10 +462,10 @@ static void *spinlock_functional_test(void *arg
> UNUSED)
>
>         thread_num = odp_cpu_id() + 1;
>         per_thread_mem = thread_init();
> -       global_mem = per_thread_mem->global_mem;
> -       iterations = global_mem->g_iterations;
> +       mem = per_thread_mem->global_mem;
> +       iterations = mem->g_iterations;
>
> -       odp_barrier_wait(&global_mem->global_barrier);
> +       odp_barrier_wait(&mem->global_barrier);
>
>         sync_failures = 0;
>         is_locked_errs = 0;
> @@ -476,13 +476,13 @@ static void *spinlock_functional_test(void *arg
> UNUSED)
>
>         for (cnt = 1; cnt <= iterations; cnt++) {
>                 /* Acquire the shared global lock */
> -               odp_spinlock_lock(&global_mem->global_spinlock);
> +               odp_spinlock_lock(&mem->global_spinlock);
>
>                 /* Make sure we have the lock AND didn't previously own it
> */
> -               if (odp_spinlock_is_locked(&global_mem->global_spinlock)
> != 1)
> +               if (odp_spinlock_is_locked(&mem->global_spinlock) != 1)
>                         is_locked_errs++;
>
> -               if (global_mem->global_lock_owner != 0) {
> +               if (mem->global_lock_owner != 0) {
>                         current_errs++;
>                         sync_failures++;
>                 }
> @@ -491,19 +491,19 @@ static void *spinlock_functional_test(void *arg
> UNUSED)
>                 * then we see if anyone else has snuck in and changed the
>                 * global_lock_owner to be themselves
>                 */
> -               global_mem->global_lock_owner = thread_num;
> +               mem->global_lock_owner = thread_num;
>                 odp_sync_stores();
>                 thread_delay(per_thread_mem, lock_owner_delay);
> -               if (global_mem->global_lock_owner != thread_num) {
> +               if (mem->global_lock_owner != thread_num) {
>                         current_errs++;
>                         sync_failures++;
>                 }
>
>                 /* Release shared lock, and make sure we no longer have it
> */
> -               global_mem->global_lock_owner = 0;
> +               mem->global_lock_owner = 0;
>                 odp_sync_stores();
> -               odp_spinlock_unlock(&global_mem->global_spinlock);
> -               if (global_mem->global_lock_owner == thread_num) {
> +               odp_spinlock_unlock(&mem->global_spinlock);
> +               if (mem->global_lock_owner == thread_num) {
>                         current_errs++;
>                         sync_failures++;
>                 }
> @@ -517,10 +517,10 @@ static void *spinlock_functional_test(void *arg
> UNUSED)
>                 /* Try to resync all of the threads to increase contention
> */
>                 if ((rs_idx < NUM_RESYNC_BARRIERS) &&
>                     ((cnt % resync_cnt) == (resync_cnt - 1)))
> -
>  odp_barrier_wait(&global_mem->barrier_array[rs_idx++]);
> +                       odp_barrier_wait(&mem->barrier_array[rs_idx++]);
>         }
>
> -       if ((global_mem->g_verbose) &&
> +       if ((mem->g_verbose) &&
>             ((sync_failures != 0) || (is_locked_errs != 0)))
>                 printf("\nThread %u (id=%d core=%d) had %u sync_failures"
>                        " and %u is_locked_errs in %u iterations\n",
> thread_num,
> @@ -537,7 +537,7 @@ static void *spinlock_functional_test(void *arg UNUSED)
>
>  static void *ticketlock_functional_test(void *arg UNUSED)
>  {
> -       global_shared_mem_t *global_mem;
> +       global_shared_mem_t *mem;
>         per_thread_mem_t *per_thread_mem;
>         uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt;
>         uint32_t sync_failures, is_locked_errs, current_errs;
> @@ -545,11 +545,11 @@ static void *ticketlock_functional_test(void *arg
> UNUSED)
>
>         thread_num = odp_cpu_id() + 1;
>         per_thread_mem = thread_init();
> -       global_mem = per_thread_mem->global_mem;
> -       iterations = global_mem->g_iterations;
> +       mem = per_thread_mem->global_mem;
> +       iterations = mem->g_iterations;
>
>         /* Wait here until all of the threads have also reached this point
> */
> -       odp_barrier_wait(&global_mem->global_barrier);
> +       odp_barrier_wait(&mem->global_barrier);
>
>         sync_failures = 0;
>         is_locked_errs = 0;
> @@ -560,14 +560,14 @@ static void *ticketlock_functional_test(void *arg
> UNUSED)
>
>         for (cnt = 1; cnt <= iterations; cnt++) {
>                 /* Acquire the shared global lock */
> -               odp_ticketlock_lock(&global_mem->global_ticketlock);
> +               odp_ticketlock_lock(&mem->global_ticketlock);
>
>                 /* Make sure we have the lock AND didn't previously own it
> */
> -               if
> (odp_ticketlock_is_locked(&global_mem->global_ticketlock)
> +               if (odp_ticketlock_is_locked(&mem->global_ticketlock)
>                                 != 1)
>                         is_locked_errs++;
>
> -               if (global_mem->global_lock_owner != 0) {
> +               if (mem->global_lock_owner != 0) {
>                         current_errs++;
>                         sync_failures++;
>                 }
> @@ -576,19 +576,19 @@ static void *ticketlock_functional_test(void *arg
> UNUSED)
>                 * then we see if anyone else has snuck in and changed the
>                 * global_lock_owner to be themselves
>                 */
> -               global_mem->global_lock_owner = thread_num;
> +               mem->global_lock_owner = thread_num;
>                 odp_sync_stores();
>                 thread_delay(per_thread_mem, lock_owner_delay);
> -               if (global_mem->global_lock_owner != thread_num) {
> +               if (mem->global_lock_owner != thread_num) {
>                         current_errs++;
>                         sync_failures++;
>                 }
>
>                 /* Release shared lock, and make sure we no longer have it
> */
> -               global_mem->global_lock_owner = 0;
> +               mem->global_lock_owner = 0;
>                 odp_sync_stores();
> -               odp_ticketlock_unlock(&global_mem->global_ticketlock);
> -               if (global_mem->global_lock_owner == thread_num) {
> +               odp_ticketlock_unlock(&mem->global_ticketlock);
> +               if (mem->global_lock_owner == thread_num) {
>                         current_errs++;
>                         sync_failures++;
>                 }
> @@ -602,10 +602,10 @@ static void *ticketlock_functional_test(void *arg
> UNUSED)
>                 /* Try to resync all of the threads to increase contention
> */
>                 if ((rs_idx < NUM_RESYNC_BARRIERS) &&
>                     ((cnt % resync_cnt) == (resync_cnt - 1)))
> -
>  odp_barrier_wait(&global_mem->barrier_array[rs_idx++]);
> +                       odp_barrier_wait(&mem->barrier_array[rs_idx++]);
>         }
>
> -       if ((global_mem->g_verbose) &&
> +       if ((mem->g_verbose) &&
>             ((sync_failures != 0) || (is_locked_errs != 0)))
>                 printf("\nThread %u (id=%d core=%d) had %u sync_failures"
>                        " and %u is_locked_errs in %u iterations\n",
> thread_num,
> @@ -622,18 +622,18 @@ static void *ticketlock_functional_test(void *arg
> UNUSED)
>
>  static void *rwlock_functional_test(void *arg UNUSED)
>  {
> -       global_shared_mem_t *global_mem;
> +       global_shared_mem_t *mem;
>         per_thread_mem_t *per_thread_mem;
>         uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt;
>         uint32_t sync_failures, current_errs, lock_owner_delay;
>
>         thread_num = odp_cpu_id() + 1;
>         per_thread_mem = thread_init();
> -       global_mem = per_thread_mem->global_mem;
> -       iterations = global_mem->g_iterations;
> +       mem = per_thread_mem->global_mem;
> +       iterations = mem->g_iterations;
>
>         /* Wait here until all of the threads have also reached this point
> */
> -       odp_barrier_wait(&global_mem->global_barrier);
> +       odp_barrier_wait(&mem->global_barrier);
>
>         sync_failures = 0;
>         current_errs = 0;
> @@ -643,10 +643,10 @@ static void *rwlock_functional_test(void *arg UNUSED)
>
>         for (cnt = 1; cnt <= iterations; cnt++) {
>                 /* Acquire the shared global lock */
> -               odp_rwlock_write_lock(&global_mem->global_rwlock);
> +               odp_rwlock_write_lock(&mem->global_rwlock);
>
>                 /* Make sure we have lock now AND didn't previously own it
> */
> -               if (global_mem->global_lock_owner != 0) {
> +               if (mem->global_lock_owner != 0) {
>                         current_errs++;
>                         sync_failures++;
>                 }
> @@ -655,19 +655,19 @@ static void *rwlock_functional_test(void *arg UNUSED)
>                 * then we see if anyone else has snuck in and changed the
>                 * global_lock_owner to be themselves
>                 */
> -               global_mem->global_lock_owner = thread_num;
> +               mem->global_lock_owner = thread_num;
>                 odp_sync_stores();
>                 thread_delay(per_thread_mem, lock_owner_delay);
> -               if (global_mem->global_lock_owner != thread_num) {
> +               if (mem->global_lock_owner != thread_num) {
>                         current_errs++;
>                         sync_failures++;
>                 }
>
>                 /* Release shared lock, and make sure we no longer have it
> */
> -               global_mem->global_lock_owner = 0;
> +               mem->global_lock_owner = 0;
>                 odp_sync_stores();
> -               odp_rwlock_write_unlock(&global_mem->global_rwlock);
> -               if (global_mem->global_lock_owner == thread_num) {
> +               odp_rwlock_write_unlock(&mem->global_rwlock);
> +               if (mem->global_lock_owner == thread_num) {
>                         current_errs++;
>                         sync_failures++;
>                 }
> @@ -681,10 +681,10 @@ static void *rwlock_functional_test(void *arg UNUSED)
>                 /* Try to resync all of the threads to increase contention
> */
>                 if ((rs_idx < NUM_RESYNC_BARRIERS) &&
>                     ((cnt % resync_cnt) == (resync_cnt - 1)))
> -
>  odp_barrier_wait(&global_mem->barrier_array[rs_idx++]);
> +                       odp_barrier_wait(&mem->barrier_array[rs_idx++]);
>         }
>
> -       if ((global_mem->g_verbose) && (sync_failures != 0))
> +       if ((mem->g_verbose) && (sync_failures != 0))
>                 printf("\nThread %u (id=%d core=%d) had %u sync_failures"
>                        " in %u iterations\n", thread_num,
>                        per_thread_mem->thread_id,
> --
> 1.9.1
>
>
> _______________________________________________
> lng-odp mailing list
> [email protected]
> http://lists.linaro.org/mailman/listinfo/lng-odp
>
_______________________________________________
lng-odp mailing list
[email protected]
http://lists.linaro.org/mailman/listinfo/lng-odp

Reply via email to