On 23 March 2015 at 20:39, Bill Fischofer <[email protected]> wrote:
> Not sure why this is needed, but: > > Reviewed-by: Bill Fischofer <[email protected]> > Bill is this the same as saying Whatevered-by: Bill Fischofer <[email protected]> Cheers, Anders > > On Mon, Mar 23, 2015 at 8:04 AM, Maxim Uvarov <[email protected]> > wrote: > >> global_mem is global static variable usage same name variable >> inside function is confusing. Rename local vars from global_mem >> to mem. >> >> Signed-off-by: Maxim Uvarov <[email protected]> >> --- >> test/validation/odp_synchronizers.c | 158 >> ++++++++++++++++++------------------ >> 1 file changed, 79 insertions(+), 79 deletions(-) >> >> diff --git a/test/validation/odp_synchronizers.c >> b/test/validation/odp_synchronizers.c >> index ab9164f..b8f4e6a 100644 >> --- a/test/validation/odp_synchronizers.c >> +++ b/test/validation/odp_synchronizers.c >> @@ -107,7 +107,7 @@ static void thread_delay(per_thread_mem_t >> *per_thread_mem, uint32_t iterations) >> /* Initialise per-thread memory */ >> static per_thread_mem_t *thread_init(void) >> { >> - global_shared_mem_t *global_mem; >> + global_shared_mem_t *mem; >> per_thread_mem_t *per_thread_mem; >> odp_shm_t global_shm; >> uint32_t per_thread_mem_len; >> @@ -122,10 +122,10 @@ static per_thread_mem_t *thread_init(void) >> per_thread_mem->thread_core = odp_cpu_id(); >> >> global_shm = odp_shm_lookup(GLOBAL_SHM_NAME); >> - global_mem = odp_shm_addr(global_shm); >> + mem = odp_shm_addr(global_shm); >> CU_ASSERT(global_mem != NULL); >> >> - per_thread_mem->global_mem = global_mem; >> + per_thread_mem->global_mem = mem; >> >> return per_thread_mem; >> } >> @@ -160,13 +160,13 @@ static void custom_barrier_wait(custom_barrier_t >> *custom_barrier) >> static uint32_t barrier_test(per_thread_mem_t *per_thread_mem, >> odp_bool_t no_barrier_test) >> { >> - global_shared_mem_t *global_mem; >> + global_shared_mem_t *mem; >> uint32_t barrier_errs, iterations, cnt, i_am_slow_thread; >> uint32_t thread_num, slow_thread_num, next_slow_thread, >> num_threads; >> uint32_t lock_owner_delay, barrier_cnt1, barrier_cnt2; >> >> thread_num = odp_cpu_id() + 1; >> - global_mem = per_thread_mem->global_mem; >> + mem = per_thread_mem->global_mem; >> num_threads = global_mem->g_num_threads; >> iterations = BARRIER_ITERATIONS; >> >> @@ -175,10 +175,10 @@ static uint32_t barrier_test(per_thread_mem_t >> *per_thread_mem, >> >> for (cnt = 1; cnt < iterations; cnt++) { >> /* Wait here until all of the threads reach this point */ >> - custom_barrier_wait(&global_mem->custom_barrier1[cnt]); >> + custom_barrier_wait(&mem->custom_barrier1[cnt]); >> >> - barrier_cnt1 = global_mem->barrier_cnt1; >> - barrier_cnt2 = global_mem->barrier_cnt2; >> + barrier_cnt1 = mem->barrier_cnt1; >> + barrier_cnt2 = mem->barrier_cnt2; >> >> if ((barrier_cnt1 != cnt) || (barrier_cnt2 != cnt)) { >> printf("thread_num=%u barrier_cnts of %u %u >> cnt=%u\n", >> @@ -187,9 +187,9 @@ static uint32_t barrier_test(per_thread_mem_t >> *per_thread_mem, >> } >> >> /* Wait here until all of the threads reach this point */ >> - custom_barrier_wait(&global_mem->custom_barrier2[cnt]); >> + custom_barrier_wait(&mem->custom_barrier2[cnt]); >> >> - slow_thread_num = global_mem->slow_thread_num; >> + slow_thread_num = mem->slow_thread_num; >> i_am_slow_thread = thread_num == slow_thread_num; >> next_slow_thread = slow_thread_num + 1; >> if (num_threads < next_slow_thread) >> @@ -206,30 +206,30 @@ static uint32_t barrier_test(per_thread_mem_t >> *per_thread_mem, >> if (i_am_slow_thread) { >> thread_delay(per_thread_mem, lock_owner_delay); >> lock_owner_delay += BASE_DELAY; >> - if ((global_mem->barrier_cnt1 != cnt) || >> - (global_mem->barrier_cnt2 != cnt) || >> - (global_mem->slow_thread_num >> + if ((mem->barrier_cnt1 != cnt) || >> + (mem->barrier_cnt2 != cnt) || >> + (mem->slow_thread_num >> != slow_thread_num)) >> barrier_errs++; >> } >> >> if (no_barrier_test == 0) >> - odp_barrier_wait(&global_mem->test_barriers[cnt]); >> + odp_barrier_wait(&mem->test_barriers[cnt]); >> >> - global_mem->barrier_cnt1 = cnt + 1; >> + mem->barrier_cnt1 = cnt + 1; >> odp_sync_stores(); >> >> if (i_am_slow_thread) { >> - global_mem->slow_thread_num = next_slow_thread; >> - global_mem->barrier_cnt2 = cnt + 1; >> + mem->slow_thread_num = next_slow_thread; >> + mem->barrier_cnt2 = cnt + 1; >> odp_sync_stores(); >> } else { >> - while (global_mem->barrier_cnt2 != (cnt + 1)) >> + while (mem->barrier_cnt2 != (cnt + 1)) >> thread_delay(per_thread_mem, BASE_DELAY); >> } >> } >> >> - if ((global_mem->g_verbose) && (barrier_errs != 0)) >> + if ((mem->g_verbose) && (barrier_errs != 0)) >> printf("\nThread %u (id=%d core=%d) had %u barrier_errs" >> " in %u iterations\n", thread_num, >> per_thread_mem->thread_id, >> @@ -293,14 +293,14 @@ static void spinlock_api_test(odp_spinlock_t >> *spinlock) >> >> static void *spinlock_api_tests(void *arg UNUSED) >> { >> - global_shared_mem_t *global_mem; >> + global_shared_mem_t *mem; >> per_thread_mem_t *per_thread_mem; >> odp_spinlock_t local_spin_lock; >> >> per_thread_mem = thread_init(); >> - global_mem = per_thread_mem->global_mem; >> + mem = per_thread_mem->global_mem; >> >> - odp_barrier_wait(&global_mem->global_barrier); >> + odp_barrier_wait(&mem->global_barrier); >> >> spinlock_api_test(&local_spin_lock); >> spinlock_api_test(&per_thread_mem->per_thread_spinlock); >> @@ -331,14 +331,14 @@ static void ticketlock_api_test(odp_ticketlock_t >> *ticketlock) >> >> static void *ticketlock_api_tests(void *arg UNUSED) >> { >> - global_shared_mem_t *global_mem; >> + global_shared_mem_t *mem; >> per_thread_mem_t *per_thread_mem; >> odp_ticketlock_t local_ticket_lock; >> >> per_thread_mem = thread_init(); >> - global_mem = per_thread_mem->global_mem; >> + mem = per_thread_mem->global_mem; >> >> - odp_barrier_wait(&global_mem->global_barrier); >> + odp_barrier_wait(&mem->global_barrier); >> >> ticketlock_api_test(&local_ticket_lock); >> ticketlock_api_test(&per_thread_mem->per_thread_ticketlock); >> @@ -365,14 +365,14 @@ static void rwlock_api_test(odp_rwlock_t *rw_lock) >> >> static void *rwlock_api_tests(void *arg UNUSED) >> { >> - global_shared_mem_t *global_mem; >> + global_shared_mem_t *mem; >> per_thread_mem_t *per_thread_mem; >> odp_rwlock_t local_rwlock; >> >> per_thread_mem = thread_init(); >> - global_mem = per_thread_mem->global_mem; >> + mem = per_thread_mem->global_mem; >> >> - odp_barrier_wait(&global_mem->global_barrier); >> + odp_barrier_wait(&mem->global_barrier); >> >> rwlock_api_test(&local_rwlock); >> rwlock_api_test(&per_thread_mem->per_thread_rwlock); >> @@ -384,17 +384,17 @@ static void *rwlock_api_tests(void *arg UNUSED) >> >> static void *no_lock_functional_test(void *arg UNUSED) >> { >> - global_shared_mem_t *global_mem; >> + global_shared_mem_t *mem; >> per_thread_mem_t *per_thread_mem; >> uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt; >> uint32_t sync_failures, current_errs, lock_owner_delay; >> >> thread_num = odp_cpu_id() + 1; >> per_thread_mem = thread_init(); >> - global_mem = per_thread_mem->global_mem; >> - iterations = global_mem->g_iterations; >> + mem = per_thread_mem->global_mem; >> + iterations = mem->g_iterations; >> >> - odp_barrier_wait(&global_mem->global_barrier); >> + odp_barrier_wait(&mem->global_barrier); >> >> sync_failures = 0; >> current_errs = 0; >> @@ -403,20 +403,20 @@ static void *no_lock_functional_test(void *arg >> UNUSED) >> lock_owner_delay = BASE_DELAY; >> >> for (cnt = 1; cnt <= iterations; cnt++) { >> - global_mem->global_lock_owner = thread_num; >> + mem->global_lock_owner = thread_num; >> odp_sync_stores(); >> thread_delay(per_thread_mem, lock_owner_delay); >> >> - if (global_mem->global_lock_owner != thread_num) { >> + if (mem->global_lock_owner != thread_num) { >> current_errs++; >> sync_failures++; >> } >> >> - global_mem->global_lock_owner = 0; >> + mem->global_lock_owner = 0; >> odp_sync_stores(); >> thread_delay(per_thread_mem, MIN_DELAY); >> >> - if (global_mem->global_lock_owner == thread_num) { >> + if (mem->global_lock_owner == thread_num) { >> current_errs++; >> sync_failures++; >> } >> @@ -430,10 +430,10 @@ static void *no_lock_functional_test(void *arg >> UNUSED) >> /* Try to resync all of the threads to increase >> contention */ >> if ((rs_idx < NUM_RESYNC_BARRIERS) && >> ((cnt % resync_cnt) == (resync_cnt - 1))) >> - >> odp_barrier_wait(&global_mem->barrier_array[rs_idx++]); >> + odp_barrier_wait(&mem->barrier_array[rs_idx++]); >> } >> >> - if (global_mem->g_verbose) >> + if (mem->g_verbose) >> printf("\nThread %u (id=%d core=%d) had %u sync_failures" >> " in %u iterations\n", thread_num, >> per_thread_mem->thread_id, >> @@ -454,7 +454,7 @@ static void *no_lock_functional_test(void *arg UNUSED) >> >> static void *spinlock_functional_test(void *arg UNUSED) >> { >> - global_shared_mem_t *global_mem; >> + global_shared_mem_t *mem; >> per_thread_mem_t *per_thread_mem; >> uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt; >> uint32_t sync_failures, is_locked_errs, current_errs; >> @@ -462,10 +462,10 @@ static void *spinlock_functional_test(void *arg >> UNUSED) >> >> thread_num = odp_cpu_id() + 1; >> per_thread_mem = thread_init(); >> - global_mem = per_thread_mem->global_mem; >> - iterations = global_mem->g_iterations; >> + mem = per_thread_mem->global_mem; >> + iterations = mem->g_iterations; >> >> - odp_barrier_wait(&global_mem->global_barrier); >> + odp_barrier_wait(&mem->global_barrier); >> >> sync_failures = 0; >> is_locked_errs = 0; >> @@ -476,13 +476,13 @@ static void *spinlock_functional_test(void *arg >> UNUSED) >> >> for (cnt = 1; cnt <= iterations; cnt++) { >> /* Acquire the shared global lock */ >> - odp_spinlock_lock(&global_mem->global_spinlock); >> + odp_spinlock_lock(&mem->global_spinlock); >> >> /* Make sure we have the lock AND didn't previously own >> it */ >> - if (odp_spinlock_is_locked(&global_mem->global_spinlock) >> != 1) >> + if (odp_spinlock_is_locked(&mem->global_spinlock) != 1) >> is_locked_errs++; >> >> - if (global_mem->global_lock_owner != 0) { >> + if (mem->global_lock_owner != 0) { >> current_errs++; >> sync_failures++; >> } >> @@ -491,19 +491,19 @@ static void *spinlock_functional_test(void *arg >> UNUSED) >> * then we see if anyone else has snuck in and changed the >> * global_lock_owner to be themselves >> */ >> - global_mem->global_lock_owner = thread_num; >> + mem->global_lock_owner = thread_num; >> odp_sync_stores(); >> thread_delay(per_thread_mem, lock_owner_delay); >> - if (global_mem->global_lock_owner != thread_num) { >> + if (mem->global_lock_owner != thread_num) { >> current_errs++; >> sync_failures++; >> } >> >> /* Release shared lock, and make sure we no longer have >> it */ >> - global_mem->global_lock_owner = 0; >> + mem->global_lock_owner = 0; >> odp_sync_stores(); >> - odp_spinlock_unlock(&global_mem->global_spinlock); >> - if (global_mem->global_lock_owner == thread_num) { >> + odp_spinlock_unlock(&mem->global_spinlock); >> + if (mem->global_lock_owner == thread_num) { >> current_errs++; >> sync_failures++; >> } >> @@ -517,10 +517,10 @@ static void *spinlock_functional_test(void *arg >> UNUSED) >> /* Try to resync all of the threads to increase >> contention */ >> if ((rs_idx < NUM_RESYNC_BARRIERS) && >> ((cnt % resync_cnt) == (resync_cnt - 1))) >> - >> odp_barrier_wait(&global_mem->barrier_array[rs_idx++]); >> + odp_barrier_wait(&mem->barrier_array[rs_idx++]); >> } >> >> - if ((global_mem->g_verbose) && >> + if ((mem->g_verbose) && >> ((sync_failures != 0) || (is_locked_errs != 0))) >> printf("\nThread %u (id=%d core=%d) had %u sync_failures" >> " and %u is_locked_errs in %u iterations\n", >> thread_num, >> @@ -537,7 +537,7 @@ static void *spinlock_functional_test(void *arg >> UNUSED) >> >> static void *ticketlock_functional_test(void *arg UNUSED) >> { >> - global_shared_mem_t *global_mem; >> + global_shared_mem_t *mem; >> per_thread_mem_t *per_thread_mem; >> uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt; >> uint32_t sync_failures, is_locked_errs, current_errs; >> @@ -545,11 +545,11 @@ static void *ticketlock_functional_test(void *arg >> UNUSED) >> >> thread_num = odp_cpu_id() + 1; >> per_thread_mem = thread_init(); >> - global_mem = per_thread_mem->global_mem; >> - iterations = global_mem->g_iterations; >> + mem = per_thread_mem->global_mem; >> + iterations = mem->g_iterations; >> >> /* Wait here until all of the threads have also reached this >> point */ >> - odp_barrier_wait(&global_mem->global_barrier); >> + odp_barrier_wait(&mem->global_barrier); >> >> sync_failures = 0; >> is_locked_errs = 0; >> @@ -560,14 +560,14 @@ static void *ticketlock_functional_test(void *arg >> UNUSED) >> >> for (cnt = 1; cnt <= iterations; cnt++) { >> /* Acquire the shared global lock */ >> - odp_ticketlock_lock(&global_mem->global_ticketlock); >> + odp_ticketlock_lock(&mem->global_ticketlock); >> >> /* Make sure we have the lock AND didn't previously own >> it */ >> - if >> (odp_ticketlock_is_locked(&global_mem->global_ticketlock) >> + if (odp_ticketlock_is_locked(&mem->global_ticketlock) >> != 1) >> is_locked_errs++; >> >> - if (global_mem->global_lock_owner != 0) { >> + if (mem->global_lock_owner != 0) { >> current_errs++; >> sync_failures++; >> } >> @@ -576,19 +576,19 @@ static void *ticketlock_functional_test(void *arg >> UNUSED) >> * then we see if anyone else has snuck in and changed the >> * global_lock_owner to be themselves >> */ >> - global_mem->global_lock_owner = thread_num; >> + mem->global_lock_owner = thread_num; >> odp_sync_stores(); >> thread_delay(per_thread_mem, lock_owner_delay); >> - if (global_mem->global_lock_owner != thread_num) { >> + if (mem->global_lock_owner != thread_num) { >> current_errs++; >> sync_failures++; >> } >> >> /* Release shared lock, and make sure we no longer have >> it */ >> - global_mem->global_lock_owner = 0; >> + mem->global_lock_owner = 0; >> odp_sync_stores(); >> - odp_ticketlock_unlock(&global_mem->global_ticketlock); >> - if (global_mem->global_lock_owner == thread_num) { >> + odp_ticketlock_unlock(&mem->global_ticketlock); >> + if (mem->global_lock_owner == thread_num) { >> current_errs++; >> sync_failures++; >> } >> @@ -602,10 +602,10 @@ static void *ticketlock_functional_test(void *arg >> UNUSED) >> /* Try to resync all of the threads to increase >> contention */ >> if ((rs_idx < NUM_RESYNC_BARRIERS) && >> ((cnt % resync_cnt) == (resync_cnt - 1))) >> - >> odp_barrier_wait(&global_mem->barrier_array[rs_idx++]); >> + odp_barrier_wait(&mem->barrier_array[rs_idx++]); >> } >> >> - if ((global_mem->g_verbose) && >> + if ((mem->g_verbose) && >> ((sync_failures != 0) || (is_locked_errs != 0))) >> printf("\nThread %u (id=%d core=%d) had %u sync_failures" >> " and %u is_locked_errs in %u iterations\n", >> thread_num, >> @@ -622,18 +622,18 @@ static void *ticketlock_functional_test(void *arg >> UNUSED) >> >> static void *rwlock_functional_test(void *arg UNUSED) >> { >> - global_shared_mem_t *global_mem; >> + global_shared_mem_t *mem; >> per_thread_mem_t *per_thread_mem; >> uint32_t thread_num, resync_cnt, rs_idx, iterations, cnt; >> uint32_t sync_failures, current_errs, lock_owner_delay; >> >> thread_num = odp_cpu_id() + 1; >> per_thread_mem = thread_init(); >> - global_mem = per_thread_mem->global_mem; >> - iterations = global_mem->g_iterations; >> + mem = per_thread_mem->global_mem; >> + iterations = mem->g_iterations; >> >> /* Wait here until all of the threads have also reached this >> point */ >> - odp_barrier_wait(&global_mem->global_barrier); >> + odp_barrier_wait(&mem->global_barrier); >> >> sync_failures = 0; >> current_errs = 0; >> @@ -643,10 +643,10 @@ static void *rwlock_functional_test(void *arg >> UNUSED) >> >> for (cnt = 1; cnt <= iterations; cnt++) { >> /* Acquire the shared global lock */ >> - odp_rwlock_write_lock(&global_mem->global_rwlock); >> + odp_rwlock_write_lock(&mem->global_rwlock); >> >> /* Make sure we have lock now AND didn't previously own >> it */ >> - if (global_mem->global_lock_owner != 0) { >> + if (mem->global_lock_owner != 0) { >> current_errs++; >> sync_failures++; >> } >> @@ -655,19 +655,19 @@ static void *rwlock_functional_test(void *arg >> UNUSED) >> * then we see if anyone else has snuck in and changed the >> * global_lock_owner to be themselves >> */ >> - global_mem->global_lock_owner = thread_num; >> + mem->global_lock_owner = thread_num; >> odp_sync_stores(); >> thread_delay(per_thread_mem, lock_owner_delay); >> - if (global_mem->global_lock_owner != thread_num) { >> + if (mem->global_lock_owner != thread_num) { >> current_errs++; >> sync_failures++; >> } >> >> /* Release shared lock, and make sure we no longer have >> it */ >> - global_mem->global_lock_owner = 0; >> + mem->global_lock_owner = 0; >> odp_sync_stores(); >> - odp_rwlock_write_unlock(&global_mem->global_rwlock); >> - if (global_mem->global_lock_owner == thread_num) { >> + odp_rwlock_write_unlock(&mem->global_rwlock); >> + if (mem->global_lock_owner == thread_num) { >> current_errs++; >> sync_failures++; >> } >> @@ -681,10 +681,10 @@ static void *rwlock_functional_test(void *arg >> UNUSED) >> /* Try to resync all of the threads to increase >> contention */ >> if ((rs_idx < NUM_RESYNC_BARRIERS) && >> ((cnt % resync_cnt) == (resync_cnt - 1))) >> - >> odp_barrier_wait(&global_mem->barrier_array[rs_idx++]); >> + odp_barrier_wait(&mem->barrier_array[rs_idx++]); >> } >> >> - if ((global_mem->g_verbose) && (sync_failures != 0)) >> + if ((mem->g_verbose) && (sync_failures != 0)) >> printf("\nThread %u (id=%d core=%d) had %u sync_failures" >> " in %u iterations\n", thread_num, >> per_thread_mem->thread_id, >> -- >> 1.9.1 >> >> >> _______________________________________________ >> lng-odp mailing list >> [email protected] >> http://lists.linaro.org/mailman/listinfo/lng-odp >> > > > _______________________________________________ > lng-odp mailing list > [email protected] > http://lists.linaro.org/mailman/listinfo/lng-odp > >
_______________________________________________ lng-odp mailing list [email protected] http://lists.linaro.org/mailman/listinfo/lng-odp
