global_mem is global static variable usage same name variable
inside function is confusing. Rename local vars from global_mem
to mem.

Signed-off-by: Maxim Uvarov <[email protected]>
---
 test/validation/odp_synchronizers.c | 158 ++++++++++++++++++------------------
 1 file changed, 79 insertions(+), 79 deletions(-)

diff --git a/test/validation/odp_synchronizers.c 
b/test/validation/odp_synchronizers.c
index ab9164f..b8f4e6a 100644
--- a/test/validation/odp_synchronizers.c
+++ b/test/validation/odp_synchronizers.c
@@ -107,7 +107,7 @@ static void thread_delay(per_thread_mem_t *per_thread_mem, 
uint32_t iterations)
 /* Initialise per-thread memory */
 static per_thread_mem_t *thread_init(void)
 {
-       global_shared_mem_t *global_mem;
+       global_shared_mem_t *mem;
        per_thread_mem_t *per_thread_mem;
        odp_shm_t global_shm;
        uint32_t per_thread_mem_len;
@@ -122,10 +122,10 @@ static per_thread_mem_t *thread_init(void)
        per_thread_mem->thread_core = odp_cpu_id();
 
        global_shm = odp_shm_lookup(GLOBAL_SHM_NAME);
-       global_mem = odp_shm_addr(global_shm);
+       mem = odp_shm_addr(global_shm);
        CU_ASSERT(global_mem != NULL);
 
-       per_thread_mem->global_mem = global_mem;
+       per_thread_mem->global_mem = mem;
 
        return per_thread_mem;
 }
@@ -160,13 +160,13 @@ static void custom_barrier_wait(custom_barrier_t 
*custom_barrier)
 static uint32_t barrier_test(per_thread_mem_t *per_thread_mem,
                             odp_bool_t no_barrier_test)
 {
-       global_shared_mem_t *global_mem;
+       global_shared_mem_t *mem;
        uint32_t barrier_errs, iterations, cnt, i_am_slow_thread;
        uint32_t thread_num, slow_thread_num, next_slow_thread, num_threads;
        uint32_t lock_owner_delay, barrier_cnt1, barrier_cnt2;
 
        thread_num = odp_cpu_id() + 1;
-       global_mem = per_thread_mem->global_mem;
+       mem = per_thread_mem->global_mem;
        num_threads = global_mem->g_num_threads;
        iterations = BARRIER_ITERATIONS;
 
@@ -175,10 +175,10 @@ static uint32_t barrier_test(per_thread_mem_t 
*per_thread_mem,
 
        for (cnt = 1; cnt < iterations; cnt++) {
                /* Wait here until all of the threads reach this point */
-               custom_barrier_wait(&global_mem->custom_barrier1[cnt]);
+               custom_barrier_wait(&mem->custom_barrier1[cnt]);
 
-               barrier_cnt1 = global_mem->barrier_cnt1;
-               barrier_cnt2 = global_mem->barrier_cnt2;
+               barrier_cnt1 = mem->barrier_cnt1;
+               barrier_cnt2 = mem->barrier_cnt2;
 
                if ((barrier_cnt1 != cnt) || (barrier_cnt2 != cnt)) {
                        printf("thread_num=%u barrier_cnts of %u %u cnt=%u\n",
@@ -187,9 +187,9 @@ static uint32_t barrier_test(per_thread_mem_t 
*per_thread_mem,
                }
 
                /* Wait here until all of the threads reach this point */
-               custom_barrier_wait(&global_mem->custom_barrier2[cnt]);
+               custom_barrier_wait(&mem->custom_barrier2[cnt]);
 
-               slow_thread_num = global_mem->slow_thread_num;
+               slow_thread_num = mem->slow_thread_num;
                i_am_slow_thread = thread_num == slow_thread_num;
                next_slow_thread = slow_thread_num + 1;
                if (num_threads < next_slow_thread)
@@ -206,30 +206,30 @@ static uint32_t barrier_test(per_thread_mem_t 
*per_thread_mem,
                if (i_am_slow_thread) {
                        thread_delay(per_thread_mem, lock_owner_delay);
                        lock_owner_delay += BASE_DELAY;
-                       if ((global_mem->barrier_cnt1 != cnt) ||
-                           (global_mem->barrier_cnt2 != cnt) ||
-                           (global_mem->slow_thread_num
+                       if ((mem->barrier_cnt1 != cnt) ||
+                           (mem->barrier_cnt2 != cnt) ||
+                           (mem->slow_thread_num
                                        != slow_thread_num))
                                barrier_errs++;
                }
 
                if (no_barrier_test == 0)
-                       odp_barrier_wait(&global_mem->test_barriers[cnt]);
+                       odp_barrier_wait(&mem->test_barriers[cnt]);
 
-               global_mem->barrier_cnt1 = cnt + 1;
+               mem->barrier_cnt1 = cnt + 1;
                odp_sync_stores();
 
                if (i_am_slow_thread) {
-                       global_mem->slow_thread_num = next_slow_thread;
-                       global_mem->barrier_cnt2 = cnt + 1;
+                       mem->slow_thread_num = next_slow_thread;
+                       mem->barrier_cnt2 = cnt + 1;
                        odp_sync_stores();
                } else {
-                       while (global_mem->barrier_cnt2 != (cnt + 1))
+                       while (mem->barrier_cnt2 != (cnt + 1))
                                thread_delay(per_thread_mem, BASE_DELAY);
                }
        }
 
-       if ((global_mem->g_verbose) && (barrier_errs != 0))
+       if ((mem->g_verbose) && (barrier_errs != 0))
                printf("\nThread %u (id=%d core=%d) had %u barrier_errs"
                       " in %u iterations\n", thread_num,
                        per_thread_mem->thread_id,
@@ -293,14 +293,14 @@ static void spinlock_api_test(odp_spinlock_t *spinlock)
 
 static void *spinlock_api_tests(void *arg UNUSED)
 {
-       global_shared_mem_t *global_mem;
+       global_shared_mem_t *mem;
        per_thread_mem_t *per_thread_mem;
        odp_spinlock_t local_spin_lock;
 
        per_thread_mem = thread_init();
-       global_mem = per_thread_mem->global_mem;
+       mem = per_thread_mem->global_mem;
 
-       odp_barrier_wait(&global_mem->global_barrier);
+       odp_barrier_wait(&mem->global_barrier);
 
        spinlock_api_test(&local_spin_lock);
        spinlock_api_test(&per_thread_mem->per_thread_spinlock);
@@ -331,14 +331,14 @@ static void ticketlock_api_test(odp_ticketlock_t 
*ticketlock)
 
 static void *ticketlock_api_tests(void *arg UNUSED)
 {
-       global_shared_mem_t *global_mem;
+       global_shared_mem_t *mem;
        per_thread_mem_t *per_thread_mem;
        odp_ticketlock_t local_ticket_lock;
 
        per_thread_mem = thread_init();
-       global_mem = per_thread_mem->global_mem;
+       mem = per_thread_mem->global_mem;
 
-       odp_barrier_wait(&global_mem->global_barrier);
+       odp_barrier_wait(&mem->global_barrier);
 
        ticketlock_api_test(&local_ticket_lock);
        ticketlock_api_test(&per_thread_mem->per_thread_ticketlock);
@@ -365,14 +365,14 @@ static void rwlock_api_test(odp_rwlock_t *rw_lock)
 
 static void *rwlock_api_tests(void *arg UNUSED)
 {
-       global_shared_mem_t *global_mem;
+       global_shared_mem_t *mem;
        per_thread_mem_t *per_thread_mem;
        odp_rwlock_t local_rwlock;
 
        per_thread_mem = thread_init();
-       global_mem = per_thread_mem->global_mem;
+       mem = per_thread_mem->global_mem;
 
-       odp_barrier_wait(&global_mem->global_barrier);
+       odp_barrier_wait(&mem->global_barrier);
 
        rwlock_api_test(&local_rwlock);
        rwlock_api_test(&per_thread_mem->per_thread_rwlock);
@@ -384,17 +384,17 @@ static void *rwlock_api_tests(void *arg UNUSED)
 
 static void *no_lock_functional_test(void *arg UNUSED)
 {
-       global_shared_mem_t *global_mem;
+       global_shared_mem_t *mem;
        per_thread_mem_t *per_thread_mem;
        uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt;
        uint32_t sync_failures, current_errs, lock_owner_delay;
 
        thread_num = odp_cpu_id() + 1;
        per_thread_mem = thread_init();
-       global_mem = per_thread_mem->global_mem;
-       iterations = global_mem->g_iterations;
+       mem = per_thread_mem->global_mem;
+       iterations = mem->g_iterations;
 
-       odp_barrier_wait(&global_mem->global_barrier);
+       odp_barrier_wait(&mem->global_barrier);
 
        sync_failures = 0;
        current_errs = 0;
@@ -403,20 +403,20 @@ static void *no_lock_functional_test(void *arg UNUSED)
        lock_owner_delay = BASE_DELAY;
 
        for (cnt = 1; cnt <= iterations; cnt++) {
-               global_mem->global_lock_owner = thread_num;
+               mem->global_lock_owner = thread_num;
                odp_sync_stores();
                thread_delay(per_thread_mem, lock_owner_delay);
 
-               if (global_mem->global_lock_owner != thread_num) {
+               if (mem->global_lock_owner != thread_num) {
                        current_errs++;
                        sync_failures++;
                }
 
-               global_mem->global_lock_owner = 0;
+               mem->global_lock_owner = 0;
                odp_sync_stores();
                thread_delay(per_thread_mem, MIN_DELAY);
 
-               if (global_mem->global_lock_owner == thread_num) {
+               if (mem->global_lock_owner == thread_num) {
                        current_errs++;
                        sync_failures++;
                }
@@ -430,10 +430,10 @@ static void *no_lock_functional_test(void *arg UNUSED)
                /* Try to resync all of the threads to increase contention */
                if ((rs_idx < NUM_RESYNC_BARRIERS) &&
                    ((cnt % resync_cnt) == (resync_cnt - 1)))
-                       odp_barrier_wait(&global_mem->barrier_array[rs_idx++]);
+                       odp_barrier_wait(&mem->barrier_array[rs_idx++]);
        }
 
-       if (global_mem->g_verbose)
+       if (mem->g_verbose)
                printf("\nThread %u (id=%d core=%d) had %u sync_failures"
                       " in %u iterations\n", thread_num,
                       per_thread_mem->thread_id,
@@ -454,7 +454,7 @@ static void *no_lock_functional_test(void *arg UNUSED)
 
 static void *spinlock_functional_test(void *arg UNUSED)
 {
-       global_shared_mem_t *global_mem;
+       global_shared_mem_t *mem;
        per_thread_mem_t *per_thread_mem;
        uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt;
        uint32_t sync_failures, is_locked_errs, current_errs;
@@ -462,10 +462,10 @@ static void *spinlock_functional_test(void *arg UNUSED)
 
        thread_num = odp_cpu_id() + 1;
        per_thread_mem = thread_init();
-       global_mem = per_thread_mem->global_mem;
-       iterations = global_mem->g_iterations;
+       mem = per_thread_mem->global_mem;
+       iterations = mem->g_iterations;
 
-       odp_barrier_wait(&global_mem->global_barrier);
+       odp_barrier_wait(&mem->global_barrier);
 
        sync_failures = 0;
        is_locked_errs = 0;
@@ -476,13 +476,13 @@ static void *spinlock_functional_test(void *arg UNUSED)
 
        for (cnt = 1; cnt <= iterations; cnt++) {
                /* Acquire the shared global lock */
-               odp_spinlock_lock(&global_mem->global_spinlock);
+               odp_spinlock_lock(&mem->global_spinlock);
 
                /* Make sure we have the lock AND didn't previously own it */
-               if (odp_spinlock_is_locked(&global_mem->global_spinlock) != 1)
+               if (odp_spinlock_is_locked(&mem->global_spinlock) != 1)
                        is_locked_errs++;
 
-               if (global_mem->global_lock_owner != 0) {
+               if (mem->global_lock_owner != 0) {
                        current_errs++;
                        sync_failures++;
                }
@@ -491,19 +491,19 @@ static void *spinlock_functional_test(void *arg UNUSED)
                * then we see if anyone else has snuck in and changed the
                * global_lock_owner to be themselves
                */
-               global_mem->global_lock_owner = thread_num;
+               mem->global_lock_owner = thread_num;
                odp_sync_stores();
                thread_delay(per_thread_mem, lock_owner_delay);
-               if (global_mem->global_lock_owner != thread_num) {
+               if (mem->global_lock_owner != thread_num) {
                        current_errs++;
                        sync_failures++;
                }
 
                /* Release shared lock, and make sure we no longer have it */
-               global_mem->global_lock_owner = 0;
+               mem->global_lock_owner = 0;
                odp_sync_stores();
-               odp_spinlock_unlock(&global_mem->global_spinlock);
-               if (global_mem->global_lock_owner == thread_num) {
+               odp_spinlock_unlock(&mem->global_spinlock);
+               if (mem->global_lock_owner == thread_num) {
                        current_errs++;
                        sync_failures++;
                }
@@ -517,10 +517,10 @@ static void *spinlock_functional_test(void *arg UNUSED)
                /* Try to resync all of the threads to increase contention */
                if ((rs_idx < NUM_RESYNC_BARRIERS) &&
                    ((cnt % resync_cnt) == (resync_cnt - 1)))
-                       odp_barrier_wait(&global_mem->barrier_array[rs_idx++]);
+                       odp_barrier_wait(&mem->barrier_array[rs_idx++]);
        }
 
-       if ((global_mem->g_verbose) &&
+       if ((mem->g_verbose) &&
            ((sync_failures != 0) || (is_locked_errs != 0)))
                printf("\nThread %u (id=%d core=%d) had %u sync_failures"
                       " and %u is_locked_errs in %u iterations\n", thread_num,
@@ -537,7 +537,7 @@ static void *spinlock_functional_test(void *arg UNUSED)
 
 static void *ticketlock_functional_test(void *arg UNUSED)
 {
-       global_shared_mem_t *global_mem;
+       global_shared_mem_t *mem;
        per_thread_mem_t *per_thread_mem;
        uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt;
        uint32_t sync_failures, is_locked_errs, current_errs;
@@ -545,11 +545,11 @@ static void *ticketlock_functional_test(void *arg UNUSED)
 
        thread_num = odp_cpu_id() + 1;
        per_thread_mem = thread_init();
-       global_mem = per_thread_mem->global_mem;
-       iterations = global_mem->g_iterations;
+       mem = per_thread_mem->global_mem;
+       iterations = mem->g_iterations;
 
        /* Wait here until all of the threads have also reached this point */
-       odp_barrier_wait(&global_mem->global_barrier);
+       odp_barrier_wait(&mem->global_barrier);
 
        sync_failures = 0;
        is_locked_errs = 0;
@@ -560,14 +560,14 @@ static void *ticketlock_functional_test(void *arg UNUSED)
 
        for (cnt = 1; cnt <= iterations; cnt++) {
                /* Acquire the shared global lock */
-               odp_ticketlock_lock(&global_mem->global_ticketlock);
+               odp_ticketlock_lock(&mem->global_ticketlock);
 
                /* Make sure we have the lock AND didn't previously own it */
-               if (odp_ticketlock_is_locked(&global_mem->global_ticketlock)
+               if (odp_ticketlock_is_locked(&mem->global_ticketlock)
                                != 1)
                        is_locked_errs++;
 
-               if (global_mem->global_lock_owner != 0) {
+               if (mem->global_lock_owner != 0) {
                        current_errs++;
                        sync_failures++;
                }
@@ -576,19 +576,19 @@ static void *ticketlock_functional_test(void *arg UNUSED)
                * then we see if anyone else has snuck in and changed the
                * global_lock_owner to be themselves
                */
-               global_mem->global_lock_owner = thread_num;
+               mem->global_lock_owner = thread_num;
                odp_sync_stores();
                thread_delay(per_thread_mem, lock_owner_delay);
-               if (global_mem->global_lock_owner != thread_num) {
+               if (mem->global_lock_owner != thread_num) {
                        current_errs++;
                        sync_failures++;
                }
 
                /* Release shared lock, and make sure we no longer have it */
-               global_mem->global_lock_owner = 0;
+               mem->global_lock_owner = 0;
                odp_sync_stores();
-               odp_ticketlock_unlock(&global_mem->global_ticketlock);
-               if (global_mem->global_lock_owner == thread_num) {
+               odp_ticketlock_unlock(&mem->global_ticketlock);
+               if (mem->global_lock_owner == thread_num) {
                        current_errs++;
                        sync_failures++;
                }
@@ -602,10 +602,10 @@ static void *ticketlock_functional_test(void *arg UNUSED)
                /* Try to resync all of the threads to increase contention */
                if ((rs_idx < NUM_RESYNC_BARRIERS) &&
                    ((cnt % resync_cnt) == (resync_cnt - 1)))
-                       odp_barrier_wait(&global_mem->barrier_array[rs_idx++]);
+                       odp_barrier_wait(&mem->barrier_array[rs_idx++]);
        }
 
-       if ((global_mem->g_verbose) &&
+       if ((mem->g_verbose) &&
            ((sync_failures != 0) || (is_locked_errs != 0)))
                printf("\nThread %u (id=%d core=%d) had %u sync_failures"
                       " and %u is_locked_errs in %u iterations\n", thread_num,
@@ -622,18 +622,18 @@ static void *ticketlock_functional_test(void *arg UNUSED)
 
 static void *rwlock_functional_test(void *arg UNUSED)
 {
-       global_shared_mem_t *global_mem;
+       global_shared_mem_t *mem;
        per_thread_mem_t *per_thread_mem;
        uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt;
        uint32_t sync_failures, current_errs, lock_owner_delay;
 
        thread_num = odp_cpu_id() + 1;
        per_thread_mem = thread_init();
-       global_mem = per_thread_mem->global_mem;
-       iterations = global_mem->g_iterations;
+       mem = per_thread_mem->global_mem;
+       iterations = mem->g_iterations;
 
        /* Wait here until all of the threads have also reached this point */
-       odp_barrier_wait(&global_mem->global_barrier);
+       odp_barrier_wait(&mem->global_barrier);
 
        sync_failures = 0;
        current_errs = 0;
@@ -643,10 +643,10 @@ static void *rwlock_functional_test(void *arg UNUSED)
 
        for (cnt = 1; cnt <= iterations; cnt++) {
                /* Acquire the shared global lock */
-               odp_rwlock_write_lock(&global_mem->global_rwlock);
+               odp_rwlock_write_lock(&mem->global_rwlock);
 
                /* Make sure we have lock now AND didn't previously own it */
-               if (global_mem->global_lock_owner != 0) {
+               if (mem->global_lock_owner != 0) {
                        current_errs++;
                        sync_failures++;
                }
@@ -655,19 +655,19 @@ static void *rwlock_functional_test(void *arg UNUSED)
                * then we see if anyone else has snuck in and changed the
                * global_lock_owner to be themselves
                */
-               global_mem->global_lock_owner = thread_num;
+               mem->global_lock_owner = thread_num;
                odp_sync_stores();
                thread_delay(per_thread_mem, lock_owner_delay);
-               if (global_mem->global_lock_owner != thread_num) {
+               if (mem->global_lock_owner != thread_num) {
                        current_errs++;
                        sync_failures++;
                }
 
                /* Release shared lock, and make sure we no longer have it */
-               global_mem->global_lock_owner = 0;
+               mem->global_lock_owner = 0;
                odp_sync_stores();
-               odp_rwlock_write_unlock(&global_mem->global_rwlock);
-               if (global_mem->global_lock_owner == thread_num) {
+               odp_rwlock_write_unlock(&mem->global_rwlock);
+               if (mem->global_lock_owner == thread_num) {
                        current_errs++;
                        sync_failures++;
                }
@@ -681,10 +681,10 @@ static void *rwlock_functional_test(void *arg UNUSED)
                /* Try to resync all of the threads to increase contention */
                if ((rs_idx < NUM_RESYNC_BARRIERS) &&
                    ((cnt % resync_cnt) == (resync_cnt - 1)))
-                       odp_barrier_wait(&global_mem->barrier_array[rs_idx++]);
+                       odp_barrier_wait(&mem->barrier_array[rs_idx++]);
        }
 
-       if ((global_mem->g_verbose) && (sync_failures != 0))
+       if ((mem->g_verbose) && (sync_failures != 0))
                printf("\nThread %u (id=%d core=%d) had %u sync_failures"
                       " in %u iterations\n", thread_num,
                       per_thread_mem->thread_id,
-- 
1.9.1


_______________________________________________
lng-odp mailing list
[email protected]
http://lists.linaro.org/mailman/listinfo/lng-odp

Reply via email to