diff mbox

[PATCHv3] validation: remove MAX_WORKERS

Message ID 1441789231-31070-1-git-send-email-maxim.uvarov@linaro.org
State New
Headers show

Commit Message

Maxim Uvarov Sept. 9, 2015, 9 a.m. UTC
ODP has api to request available number of workers. Now
no need limit that inside application.

Signed-off-by: Maxim Uvarov <maxim.uvarov@linaro.org>
---
 v3: account that thread_tbl calloc can fail.
 v2: timer.c: do not substract -1 from number of workers.

 test/validation/common/odp_cunit_common.c     | 14 ++++++--
 test/validation/common/odp_cunit_common.h     |  2 --
 test/validation/scheduler/scheduler.c         |  3 --
 test/validation/shmem/shmem.c                 |  6 ++--
 test/validation/synchronizers/synchronizers.c | 48 ++++++++++-----------------
 test/validation/timer/timer.c                 | 10 ++----
 6 files changed, 34 insertions(+), 49 deletions(-)

Comments

Nicolas Morey-Chaisemartin Sept. 9, 2015, 9:09 a.m. UTC | #1
On 09/09/2015 11:00 AM, Maxim Uvarov wrote:
> ODP has api to request available number of workers. Now
> no need limit that inside application.
>
> Signed-off-by: Maxim Uvarov <maxim.uvarov@linaro.org>
Reviewed-by: Nicolas Morey-Chaisemartin <nmorey@kalray.eu>
> ---
>  v3: account that thread_tbl calloc can fail.
>  v2: timer.c: do not substract -1 from number of workers.
>
>  test/validation/common/odp_cunit_common.c     | 14 ++++++--
>  test/validation/common/odp_cunit_common.h     |  2 --
>  test/validation/scheduler/scheduler.c         |  3 --
>  test/validation/shmem/shmem.c                 |  6 ++--
>  test/validation/synchronizers/synchronizers.c | 48 ++++++++++-----------------
>  test/validation/timer/timer.c                 | 10 ++----
>  6 files changed, 34 insertions(+), 49 deletions(-)
>
> diff --git a/test/validation/common/odp_cunit_common.c b/test/validation/common/odp_cunit_common.c
> index d995ad3..4d61024 100644
> --- a/test/validation/common/odp_cunit_common.c
> +++ b/test/validation/common/odp_cunit_common.c
> @@ -9,7 +9,7 @@
>  #include <odp_cunit_common.h>
>  #include <odp/helper/linux.h>
>  /* Globals */
> -static odph_linux_pthread_t thread_tbl[MAX_WORKERS];
> +static odph_linux_pthread_t *thread_tbl;
>  
>  /*
>   * global init/term functions which may be registered
> @@ -26,9 +26,15 @@ static struct {
>  int odp_cunit_thread_create(void *func_ptr(void *), pthrd_arg *arg)
>  {
>  	odp_cpumask_t cpumask;
> +	int num;
>  
>  	/* Create and init additional threads */
> -	odp_cpumask_def_worker(&cpumask, arg->numthrds);
> +	num = odp_cpumask_def_worker(&cpumask, arg->numthrds);
> +	thread_tbl = calloc(sizeof(odph_linux_pthread_t), num);
> +	if (!thread_tbl) {
> +		fprintf(stderr, "error: thread_tbl memory alloc.\n");
> +		return -1;
> +	}
>  
>  	return odph_linux_pthread_create(thread_tbl, &cpumask, func_ptr,
>  					 (void *)arg);
> @@ -39,6 +45,10 @@ int odp_cunit_thread_exit(pthrd_arg *arg)
>  {
>  	/* Wait for other threads to exit */
>  	odph_linux_pthread_join(thread_tbl, arg->numthrds);
> +	if (thread_tbl) {
> +		free(thread_tbl);
> +		thread_tbl = NULL;
> +	}
>  
>  	return 0;
>  }
> diff --git a/test/validation/common/odp_cunit_common.h b/test/validation/common/odp_cunit_common.h
> index 6cafaaa..f94b44e 100644
> --- a/test/validation/common/odp_cunit_common.h
> +++ b/test/validation/common/odp_cunit_common.h
> @@ -16,8 +16,6 @@
>  #include <stdint.h>
>  #include "CUnit/Basic.h"
>  
> -#define MAX_WORKERS 32 /**< Maximum number of work threads */
> -
>  /* the function, called by module main(), to run the testsuites: */
>  int odp_cunit_run(CU_SuiteInfo testsuites[]);
>  
> diff --git a/test/validation/scheduler/scheduler.c b/test/validation/scheduler/scheduler.c
> index 1874889..e12895d 100644
> --- a/test/validation/scheduler/scheduler.c
> +++ b/test/validation/scheduler/scheduler.c
> @@ -8,7 +8,6 @@
>  #include "odp_cunit_common.h"
>  #include "scheduler.h"
>  
> -#define MAX_WORKERS_THREADS	32
>  #define MSG_POOL_SIZE		(4 * 1024 * 1024)
>  #define QUEUES_PER_PRIO		16
>  #define BUF_SIZE		64
> @@ -1018,8 +1017,6 @@ int scheduler_suite_init(void)
>  	memset(globals, 0, sizeof(test_globals_t));
>  
>  	globals->num_workers = odp_cpumask_def_worker(&mask, 0);
> -	if (globals->num_workers > MAX_WORKERS)
> -		globals->num_workers = MAX_WORKERS;
>  
>  	shm = odp_shm_reserve(SHM_THR_ARGS_NAME, sizeof(thread_args_t),
>  			      ODP_CACHE_LINE_SIZE, 0);
> diff --git a/test/validation/shmem/shmem.c b/test/validation/shmem/shmem.c
> index 6dc579a..dfa5310 100644
> --- a/test/validation/shmem/shmem.c
> +++ b/test/validation/shmem/shmem.c
> @@ -49,6 +49,7 @@ void shmem_test_odp_shm_sunnyday(void)
>  	pthrd_arg thrdarg;
>  	odp_shm_t shm;
>  	test_shared_data_t *test_shared_data;
> +	odp_cpumask_t mask;
>  
>  	shm = odp_shm_reserve(TESTNAME,
>  			      sizeof(test_shared_data_t), ALIGE_SIZE, 0);
> @@ -67,10 +68,7 @@ void shmem_test_odp_shm_sunnyday(void)
>  	test_shared_data->foo = TEST_SHARE_FOO;
>  	test_shared_data->bar = TEST_SHARE_BAR;
>  
> -	thrdarg.numthrds = odp_cpu_count();
> -
> -	if (thrdarg.numthrds > MAX_WORKERS)
> -		thrdarg.numthrds = MAX_WORKERS;
> +	thrdarg.numthrds = odp_cpumask_def_worker(&mask, 0);
>  
>  	odp_cunit_thread_create(run_shm_thread, &thrdarg);
>  	odp_cunit_thread_exit(&thrdarg);
> diff --git a/test/validation/synchronizers/synchronizers.c b/test/validation/synchronizers/synchronizers.c
> index 0a31a40..914b37e 100644
> --- a/test/validation/synchronizers/synchronizers.c
> +++ b/test/validation/synchronizers/synchronizers.c
> @@ -45,7 +45,7 @@ typedef struct {
>  
>  typedef struct {
>  	/* Global variables */
> -	uint32_t g_num_threads;
> +	uint32_t g_num_workers;
>  	uint32_t g_iterations;
>  	uint32_t g_verbose;
>  	uint32_t g_max_num_cores;
> @@ -169,7 +169,7 @@ static uint32_t barrier_test(per_thread_mem_t *per_thread_mem,
>  
>  	thread_num = odp_thread_id();
>  	global_mem = per_thread_mem->global_mem;
> -	num_threads = global_mem->g_num_threads;
> +	num_threads = global_mem->g_num_workers;
>  	iterations = BARRIER_ITERATIONS;
>  
>  	barrier_errs = 0;
> @@ -710,7 +710,7 @@ static void barrier_test_init(void)
>  {
>  	uint32_t num_threads, idx;
>  
> -	num_threads = global_mem->g_num_threads;
> +	num_threads = global_mem->g_num_workers;
>  
>  	for (idx = 0; idx < NUM_TEST_BARRIERS; idx++) {
>  		odp_barrier_init(&global_mem->test_barriers[idx], num_threads);
> @@ -924,7 +924,7 @@ void synchronizers_test_no_barrier_functional(void)
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	barrier_test_init();
>  	odp_cunit_thread_create(no_barrier_functional_test, &arg);
>  	odp_cunit_thread_exit(&arg);
> @@ -934,7 +934,7 @@ void synchronizers_test_barrier_functional(void)
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	barrier_test_init();
>  	odp_cunit_thread_create(barrier_functional_test, &arg);
>  	odp_cunit_thread_exit(&arg);
> @@ -951,7 +951,7 @@ void synchronizers_test_no_lock_functional(void)
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	odp_cunit_thread_create(no_lock_functional_test, &arg);
>  	odp_cunit_thread_exit(&arg);
>  }
> @@ -966,7 +966,7 @@ void synchronizers_test_spinlock_api(void)
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	odp_cunit_thread_create(spinlock_api_tests, &arg);
>  	odp_cunit_thread_exit(&arg);
>  }
> @@ -975,7 +975,7 @@ void synchronizers_test_spinlock_functional(void)
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	odp_spinlock_init(&global_mem->global_spinlock);
>  	odp_cunit_thread_create(spinlock_functional_test, &arg);
>  	odp_cunit_thread_exit(&arg);
> @@ -992,7 +992,7 @@ void synchronizers_test_ticketlock_api(void)
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	odp_cunit_thread_create(ticketlock_api_tests, &arg);
>  	odp_cunit_thread_exit(&arg);
>  }
> @@ -1001,7 +1001,7 @@ void synchronizers_test_ticketlock_functional(void)
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	odp_ticketlock_init(&global_mem->global_ticketlock);
>  
>  	odp_cunit_thread_create(ticketlock_functional_test, &arg);
> @@ -1019,7 +1019,7 @@ void synchronizers_test_rwlock_api(void)
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	odp_cunit_thread_create(rwlock_api_tests, &arg);
>  	odp_cunit_thread_exit(&arg);
>  }
> @@ -1028,7 +1028,7 @@ void synchronizers_test_rwlock_functional(void)
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	odp_rwlock_init(&global_mem->global_rwlock);
>  	odp_cunit_thread_create(rwlock_functional_test, &arg);
>  	odp_cunit_thread_exit(&arg);
> @@ -1044,7 +1044,7 @@ int synchronizers_suite_init(void)
>  {
>  	uint32_t num_threads, idx;
>  
> -	num_threads = global_mem->g_num_threads;
> +	num_threads = global_mem->g_num_workers;
>  	odp_barrier_init(&global_mem->global_barrier, num_threads);
>  	for (idx = 0; idx < NUM_RESYNC_BARRIERS; idx++)
>  		odp_barrier_init(&global_mem->barrier_array[idx], num_threads);
> @@ -1054,7 +1054,6 @@ int synchronizers_suite_init(void)
>  
>  int synchronizers_init(void)
>  {
> -	uint32_t workers_count, max_threads;
>  	int ret = 0;
>  	odp_cpumask_t mask;
>  
> @@ -1078,25 +1077,12 @@ int synchronizers_init(void)
>  	global_mem = odp_shm_addr(global_shm);
>  	memset(global_mem, 0, sizeof(global_shared_mem_t));
>  
> -	global_mem->g_num_threads = MAX_WORKERS;
> +	global_mem->g_num_workers = odp_cpumask_def_worker(&mask, 0);
>  	global_mem->g_iterations = MAX_ITERATIONS;
>  	global_mem->g_verbose = VERBOSE;
>  
> -	workers_count = odp_cpumask_def_worker(&mask, 0);
> -
> -	max_threads = (workers_count >= MAX_WORKERS) ?
> -			MAX_WORKERS : workers_count;
> -
> -	if (max_threads < global_mem->g_num_threads) {
> -		printf("Requested num of threads is too large\n");
> -		printf("reducing from %" PRIu32 " to %" PRIu32 "\n",
> -		       global_mem->g_num_threads,
> -		       max_threads);
> -		global_mem->g_num_threads = max_threads;
> -	}
> -
> -	printf("Num of threads used = %" PRIu32 "\n",
> -	       global_mem->g_num_threads);
> +	printf("Num of workers used = %" PRIu32 "\n",
> +	       global_mem->g_num_workers);
>  
>  	return ret;
>  }
> @@ -1158,7 +1144,7 @@ static void test_atomic_functional(void *func_ptr(void *))
>  {
>  	pthrd_arg arg;
>  
> -	arg.numthrds = global_mem->g_num_threads;
> +	arg.numthrds = global_mem->g_num_workers;
>  	test_atomic_init();
>  	test_atomic_store();
>  	odp_cunit_thread_create(func_ptr, &arg);
> diff --git a/test/validation/timer/timer.c b/test/validation/timer/timer.c
> index 7a8b98a..bcba3d4 100644
> --- a/test/validation/timer/timer.c
> +++ b/test/validation/timer/timer.c
> @@ -34,12 +34,6 @@ static odp_timer_pool_t tp;
>  /** @private Count of timeouts delivered too late */
>  static odp_atomic_u32_t ndelivtoolate;
>  
> -/** @private min() function */
> -static int min(int a, int b)
> -{
> -	return a < b ? a : b;
> -}
> -
>  /* @private Timer helper structure */
>  struct test_timer {
>  	odp_timer_t tim; /* Timer handle */
> @@ -441,10 +435,12 @@ void timer_test_odp_timer_all(void)
>  	int rc;
>  	odp_pool_param_t params;
>  	odp_timer_pool_param_t tparam;
> +	odp_cpumask_t mask;
> +
>  	/* Reserve at least one core for running other processes so the timer
>  	 * test hopefully can run undisturbed and thus get better timing
>  	 * results. */
> -	int num_workers = min(odp_cpu_count() - 1, MAX_WORKERS);
> +	int num_workers = odp_cpumask_def_worker(&mask, 0) - 1;
>  	/* On a single-CPU machine run at least one thread */
>  	if (num_workers < 1)
>  		num_workers = 1;
Maxim Uvarov Sept. 9, 2015, 9:15 a.m. UTC | #2
opps, somehow change for timer is not in. Will send v4.

Maxim.

On 09/09/15 12:09, Nicolas Morey-Chaisemartin wrote:
> On 09/09/2015 11:00 AM, Maxim Uvarov wrote:
>> ODP has api to request available number of workers. Now
>> no need limit that inside application.
>>
>> Signed-off-by: Maxim Uvarov <maxim.uvarov@linaro.org>
> Reviewed-by: Nicolas Morey-Chaisemartin <nmorey@kalray.eu>
>> ---
>>   v3: account that thread_tbl calloc can fail.
>>   v2: timer.c: do not substract -1 from number of workers.
>>
>>   test/validation/common/odp_cunit_common.c     | 14 ++++++--
>>   test/validation/common/odp_cunit_common.h     |  2 --
>>   test/validation/scheduler/scheduler.c         |  3 --
>>   test/validation/shmem/shmem.c                 |  6 ++--
>>   test/validation/synchronizers/synchronizers.c | 48 ++++++++++-----------------
>>   test/validation/timer/timer.c                 | 10 ++----
>>   6 files changed, 34 insertions(+), 49 deletions(-)
>>
>> diff --git a/test/validation/common/odp_cunit_common.c b/test/validation/common/odp_cunit_common.c
>> index d995ad3..4d61024 100644
>> --- a/test/validation/common/odp_cunit_common.c
>> +++ b/test/validation/common/odp_cunit_common.c
>> @@ -9,7 +9,7 @@
>>   #include <odp_cunit_common.h>
>>   #include <odp/helper/linux.h>
>>   /* Globals */
>> -static odph_linux_pthread_t thread_tbl[MAX_WORKERS];
>> +static odph_linux_pthread_t *thread_tbl;
>>   
>>   /*
>>    * global init/term functions which may be registered
>> @@ -26,9 +26,15 @@ static struct {
>>   int odp_cunit_thread_create(void *func_ptr(void *), pthrd_arg *arg)
>>   {
>>   	odp_cpumask_t cpumask;
>> +	int num;
>>   
>>   	/* Create and init additional threads */
>> -	odp_cpumask_def_worker(&cpumask, arg->numthrds);
>> +	num = odp_cpumask_def_worker(&cpumask, arg->numthrds);
>> +	thread_tbl = calloc(sizeof(odph_linux_pthread_t), num);
>> +	if (!thread_tbl) {
>> +		fprintf(stderr, "error: thread_tbl memory alloc.\n");
>> +		return -1;
>> +	}
>>   
>>   	return odph_linux_pthread_create(thread_tbl, &cpumask, func_ptr,
>>   					 (void *)arg);
>> @@ -39,6 +45,10 @@ int odp_cunit_thread_exit(pthrd_arg *arg)
>>   {
>>   	/* Wait for other threads to exit */
>>   	odph_linux_pthread_join(thread_tbl, arg->numthrds);
>> +	if (thread_tbl) {
>> +		free(thread_tbl);
>> +		thread_tbl = NULL;
>> +	}
>>   
>>   	return 0;
>>   }
>> diff --git a/test/validation/common/odp_cunit_common.h b/test/validation/common/odp_cunit_common.h
>> index 6cafaaa..f94b44e 100644
>> --- a/test/validation/common/odp_cunit_common.h
>> +++ b/test/validation/common/odp_cunit_common.h
>> @@ -16,8 +16,6 @@
>>   #include <stdint.h>
>>   #include "CUnit/Basic.h"
>>   
>> -#define MAX_WORKERS 32 /**< Maximum number of work threads */
>> -
>>   /* the function, called by module main(), to run the testsuites: */
>>   int odp_cunit_run(CU_SuiteInfo testsuites[]);
>>   
>> diff --git a/test/validation/scheduler/scheduler.c b/test/validation/scheduler/scheduler.c
>> index 1874889..e12895d 100644
>> --- a/test/validation/scheduler/scheduler.c
>> +++ b/test/validation/scheduler/scheduler.c
>> @@ -8,7 +8,6 @@
>>   #include "odp_cunit_common.h"
>>   #include "scheduler.h"
>>   
>> -#define MAX_WORKERS_THREADS	32
>>   #define MSG_POOL_SIZE		(4 * 1024 * 1024)
>>   #define QUEUES_PER_PRIO		16
>>   #define BUF_SIZE		64
>> @@ -1018,8 +1017,6 @@ int scheduler_suite_init(void)
>>   	memset(globals, 0, sizeof(test_globals_t));
>>   
>>   	globals->num_workers = odp_cpumask_def_worker(&mask, 0);
>> -	if (globals->num_workers > MAX_WORKERS)
>> -		globals->num_workers = MAX_WORKERS;
>>   
>>   	shm = odp_shm_reserve(SHM_THR_ARGS_NAME, sizeof(thread_args_t),
>>   			      ODP_CACHE_LINE_SIZE, 0);
>> diff --git a/test/validation/shmem/shmem.c b/test/validation/shmem/shmem.c
>> index 6dc579a..dfa5310 100644
>> --- a/test/validation/shmem/shmem.c
>> +++ b/test/validation/shmem/shmem.c
>> @@ -49,6 +49,7 @@ void shmem_test_odp_shm_sunnyday(void)
>>   	pthrd_arg thrdarg;
>>   	odp_shm_t shm;
>>   	test_shared_data_t *test_shared_data;
>> +	odp_cpumask_t mask;
>>   
>>   	shm = odp_shm_reserve(TESTNAME,
>>   			      sizeof(test_shared_data_t), ALIGE_SIZE, 0);
>> @@ -67,10 +68,7 @@ void shmem_test_odp_shm_sunnyday(void)
>>   	test_shared_data->foo = TEST_SHARE_FOO;
>>   	test_shared_data->bar = TEST_SHARE_BAR;
>>   
>> -	thrdarg.numthrds = odp_cpu_count();
>> -
>> -	if (thrdarg.numthrds > MAX_WORKERS)
>> -		thrdarg.numthrds = MAX_WORKERS;
>> +	thrdarg.numthrds = odp_cpumask_def_worker(&mask, 0);
>>   
>>   	odp_cunit_thread_create(run_shm_thread, &thrdarg);
>>   	odp_cunit_thread_exit(&thrdarg);
>> diff --git a/test/validation/synchronizers/synchronizers.c b/test/validation/synchronizers/synchronizers.c
>> index 0a31a40..914b37e 100644
>> --- a/test/validation/synchronizers/synchronizers.c
>> +++ b/test/validation/synchronizers/synchronizers.c
>> @@ -45,7 +45,7 @@ typedef struct {
>>   
>>   typedef struct {
>>   	/* Global variables */
>> -	uint32_t g_num_threads;
>> +	uint32_t g_num_workers;
>>   	uint32_t g_iterations;
>>   	uint32_t g_verbose;
>>   	uint32_t g_max_num_cores;
>> @@ -169,7 +169,7 @@ static uint32_t barrier_test(per_thread_mem_t *per_thread_mem,
>>   
>>   	thread_num = odp_thread_id();
>>   	global_mem = per_thread_mem->global_mem;
>> -	num_threads = global_mem->g_num_threads;
>> +	num_threads = global_mem->g_num_workers;
>>   	iterations = BARRIER_ITERATIONS;
>>   
>>   	barrier_errs = 0;
>> @@ -710,7 +710,7 @@ static void barrier_test_init(void)
>>   {
>>   	uint32_t num_threads, idx;
>>   
>> -	num_threads = global_mem->g_num_threads;
>> +	num_threads = global_mem->g_num_workers;
>>   
>>   	for (idx = 0; idx < NUM_TEST_BARRIERS; idx++) {
>>   		odp_barrier_init(&global_mem->test_barriers[idx], num_threads);
>> @@ -924,7 +924,7 @@ void synchronizers_test_no_barrier_functional(void)
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	barrier_test_init();
>>   	odp_cunit_thread_create(no_barrier_functional_test, &arg);
>>   	odp_cunit_thread_exit(&arg);
>> @@ -934,7 +934,7 @@ void synchronizers_test_barrier_functional(void)
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	barrier_test_init();
>>   	odp_cunit_thread_create(barrier_functional_test, &arg);
>>   	odp_cunit_thread_exit(&arg);
>> @@ -951,7 +951,7 @@ void synchronizers_test_no_lock_functional(void)
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	odp_cunit_thread_create(no_lock_functional_test, &arg);
>>   	odp_cunit_thread_exit(&arg);
>>   }
>> @@ -966,7 +966,7 @@ void synchronizers_test_spinlock_api(void)
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	odp_cunit_thread_create(spinlock_api_tests, &arg);
>>   	odp_cunit_thread_exit(&arg);
>>   }
>> @@ -975,7 +975,7 @@ void synchronizers_test_spinlock_functional(void)
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	odp_spinlock_init(&global_mem->global_spinlock);
>>   	odp_cunit_thread_create(spinlock_functional_test, &arg);
>>   	odp_cunit_thread_exit(&arg);
>> @@ -992,7 +992,7 @@ void synchronizers_test_ticketlock_api(void)
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	odp_cunit_thread_create(ticketlock_api_tests, &arg);
>>   	odp_cunit_thread_exit(&arg);
>>   }
>> @@ -1001,7 +1001,7 @@ void synchronizers_test_ticketlock_functional(void)
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	odp_ticketlock_init(&global_mem->global_ticketlock);
>>   
>>   	odp_cunit_thread_create(ticketlock_functional_test, &arg);
>> @@ -1019,7 +1019,7 @@ void synchronizers_test_rwlock_api(void)
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	odp_cunit_thread_create(rwlock_api_tests, &arg);
>>   	odp_cunit_thread_exit(&arg);
>>   }
>> @@ -1028,7 +1028,7 @@ void synchronizers_test_rwlock_functional(void)
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	odp_rwlock_init(&global_mem->global_rwlock);
>>   	odp_cunit_thread_create(rwlock_functional_test, &arg);
>>   	odp_cunit_thread_exit(&arg);
>> @@ -1044,7 +1044,7 @@ int synchronizers_suite_init(void)
>>   {
>>   	uint32_t num_threads, idx;
>>   
>> -	num_threads = global_mem->g_num_threads;
>> +	num_threads = global_mem->g_num_workers;
>>   	odp_barrier_init(&global_mem->global_barrier, num_threads);
>>   	for (idx = 0; idx < NUM_RESYNC_BARRIERS; idx++)
>>   		odp_barrier_init(&global_mem->barrier_array[idx], num_threads);
>> @@ -1054,7 +1054,6 @@ int synchronizers_suite_init(void)
>>   
>>   int synchronizers_init(void)
>>   {
>> -	uint32_t workers_count, max_threads;
>>   	int ret = 0;
>>   	odp_cpumask_t mask;
>>   
>> @@ -1078,25 +1077,12 @@ int synchronizers_init(void)
>>   	global_mem = odp_shm_addr(global_shm);
>>   	memset(global_mem, 0, sizeof(global_shared_mem_t));
>>   
>> -	global_mem->g_num_threads = MAX_WORKERS;
>> +	global_mem->g_num_workers = odp_cpumask_def_worker(&mask, 0);
>>   	global_mem->g_iterations = MAX_ITERATIONS;
>>   	global_mem->g_verbose = VERBOSE;
>>   
>> -	workers_count = odp_cpumask_def_worker(&mask, 0);
>> -
>> -	max_threads = (workers_count >= MAX_WORKERS) ?
>> -			MAX_WORKERS : workers_count;
>> -
>> -	if (max_threads < global_mem->g_num_threads) {
>> -		printf("Requested num of threads is too large\n");
>> -		printf("reducing from %" PRIu32 " to %" PRIu32 "\n",
>> -		       global_mem->g_num_threads,
>> -		       max_threads);
>> -		global_mem->g_num_threads = max_threads;
>> -	}
>> -
>> -	printf("Num of threads used = %" PRIu32 "\n",
>> -	       global_mem->g_num_threads);
>> +	printf("Num of workers used = %" PRIu32 "\n",
>> +	       global_mem->g_num_workers);
>>   
>>   	return ret;
>>   }
>> @@ -1158,7 +1144,7 @@ static void test_atomic_functional(void *func_ptr(void *))
>>   {
>>   	pthrd_arg arg;
>>   
>> -	arg.numthrds = global_mem->g_num_threads;
>> +	arg.numthrds = global_mem->g_num_workers;
>>   	test_atomic_init();
>>   	test_atomic_store();
>>   	odp_cunit_thread_create(func_ptr, &arg);
>> diff --git a/test/validation/timer/timer.c b/test/validation/timer/timer.c
>> index 7a8b98a..bcba3d4 100644
>> --- a/test/validation/timer/timer.c
>> +++ b/test/validation/timer/timer.c
>> @@ -34,12 +34,6 @@ static odp_timer_pool_t tp;
>>   /** @private Count of timeouts delivered too late */
>>   static odp_atomic_u32_t ndelivtoolate;
>>   
>> -/** @private min() function */
>> -static int min(int a, int b)
>> -{
>> -	return a < b ? a : b;
>> -}
>> -
>>   /* @private Timer helper structure */
>>   struct test_timer {
>>   	odp_timer_t tim; /* Timer handle */
>> @@ -441,10 +435,12 @@ void timer_test_odp_timer_all(void)
>>   	int rc;
>>   	odp_pool_param_t params;
>>   	odp_timer_pool_param_t tparam;
>> +	odp_cpumask_t mask;
>> +
>>   	/* Reserve at least one core for running other processes so the timer
>>   	 * test hopefully can run undisturbed and thus get better timing
>>   	 * results. */
>> -	int num_workers = min(odp_cpu_count() - 1, MAX_WORKERS);
>> +	int num_workers = odp_cpumask_def_worker(&mask, 0) - 1;
>>   	/* On a single-CPU machine run at least one thread */
>>   	if (num_workers < 1)
>>   		num_workers = 1;
diff mbox

Patch

diff --git a/test/validation/common/odp_cunit_common.c b/test/validation/common/odp_cunit_common.c
index d995ad3..4d61024 100644
--- a/test/validation/common/odp_cunit_common.c
+++ b/test/validation/common/odp_cunit_common.c
@@ -9,7 +9,7 @@ 
 #include <odp_cunit_common.h>
 #include <odp/helper/linux.h>
 /* Globals */
-static odph_linux_pthread_t thread_tbl[MAX_WORKERS];
+static odph_linux_pthread_t *thread_tbl;
 
 /*
  * global init/term functions which may be registered
@@ -26,9 +26,15 @@  static struct {
 int odp_cunit_thread_create(void *func_ptr(void *), pthrd_arg *arg)
 {
 	odp_cpumask_t cpumask;
+	int num;
 
 	/* Create and init additional threads */
-	odp_cpumask_def_worker(&cpumask, arg->numthrds);
+	num = odp_cpumask_def_worker(&cpumask, arg->numthrds);
+	thread_tbl = calloc(sizeof(odph_linux_pthread_t), num);
+	if (!thread_tbl) {
+		fprintf(stderr, "error: thread_tbl memory alloc.\n");
+		return -1;
+	}
 
 	return odph_linux_pthread_create(thread_tbl, &cpumask, func_ptr,
 					 (void *)arg);
@@ -39,6 +45,10 @@  int odp_cunit_thread_exit(pthrd_arg *arg)
 {
 	/* Wait for other threads to exit */
 	odph_linux_pthread_join(thread_tbl, arg->numthrds);
+	if (thread_tbl) {
+		free(thread_tbl);
+		thread_tbl = NULL;
+	}
 
 	return 0;
 }
diff --git a/test/validation/common/odp_cunit_common.h b/test/validation/common/odp_cunit_common.h
index 6cafaaa..f94b44e 100644
--- a/test/validation/common/odp_cunit_common.h
+++ b/test/validation/common/odp_cunit_common.h
@@ -16,8 +16,6 @@ 
 #include <stdint.h>
 #include "CUnit/Basic.h"
 
-#define MAX_WORKERS 32 /**< Maximum number of work threads */
-
 /* the function, called by module main(), to run the testsuites: */
 int odp_cunit_run(CU_SuiteInfo testsuites[]);
 
diff --git a/test/validation/scheduler/scheduler.c b/test/validation/scheduler/scheduler.c
index 1874889..e12895d 100644
--- a/test/validation/scheduler/scheduler.c
+++ b/test/validation/scheduler/scheduler.c
@@ -8,7 +8,6 @@ 
 #include "odp_cunit_common.h"
 #include "scheduler.h"
 
-#define MAX_WORKERS_THREADS	32
 #define MSG_POOL_SIZE		(4 * 1024 * 1024)
 #define QUEUES_PER_PRIO		16
 #define BUF_SIZE		64
@@ -1018,8 +1017,6 @@  int scheduler_suite_init(void)
 	memset(globals, 0, sizeof(test_globals_t));
 
 	globals->num_workers = odp_cpumask_def_worker(&mask, 0);
-	if (globals->num_workers > MAX_WORKERS)
-		globals->num_workers = MAX_WORKERS;
 
 	shm = odp_shm_reserve(SHM_THR_ARGS_NAME, sizeof(thread_args_t),
 			      ODP_CACHE_LINE_SIZE, 0);
diff --git a/test/validation/shmem/shmem.c b/test/validation/shmem/shmem.c
index 6dc579a..dfa5310 100644
--- a/test/validation/shmem/shmem.c
+++ b/test/validation/shmem/shmem.c
@@ -49,6 +49,7 @@  void shmem_test_odp_shm_sunnyday(void)
 	pthrd_arg thrdarg;
 	odp_shm_t shm;
 	test_shared_data_t *test_shared_data;
+	odp_cpumask_t mask;
 
 	shm = odp_shm_reserve(TESTNAME,
 			      sizeof(test_shared_data_t), ALIGE_SIZE, 0);
@@ -67,10 +68,7 @@  void shmem_test_odp_shm_sunnyday(void)
 	test_shared_data->foo = TEST_SHARE_FOO;
 	test_shared_data->bar = TEST_SHARE_BAR;
 
-	thrdarg.numthrds = odp_cpu_count();
-
-	if (thrdarg.numthrds > MAX_WORKERS)
-		thrdarg.numthrds = MAX_WORKERS;
+	thrdarg.numthrds = odp_cpumask_def_worker(&mask, 0);
 
 	odp_cunit_thread_create(run_shm_thread, &thrdarg);
 	odp_cunit_thread_exit(&thrdarg);
diff --git a/test/validation/synchronizers/synchronizers.c b/test/validation/synchronizers/synchronizers.c
index 0a31a40..914b37e 100644
--- a/test/validation/synchronizers/synchronizers.c
+++ b/test/validation/synchronizers/synchronizers.c
@@ -45,7 +45,7 @@  typedef struct {
 
 typedef struct {
 	/* Global variables */
-	uint32_t g_num_threads;
+	uint32_t g_num_workers;
 	uint32_t g_iterations;
 	uint32_t g_verbose;
 	uint32_t g_max_num_cores;
@@ -169,7 +169,7 @@  static uint32_t barrier_test(per_thread_mem_t *per_thread_mem,
 
 	thread_num = odp_thread_id();
 	global_mem = per_thread_mem->global_mem;
-	num_threads = global_mem->g_num_threads;
+	num_threads = global_mem->g_num_workers;
 	iterations = BARRIER_ITERATIONS;
 
 	barrier_errs = 0;
@@ -710,7 +710,7 @@  static void barrier_test_init(void)
 {
 	uint32_t num_threads, idx;
 
-	num_threads = global_mem->g_num_threads;
+	num_threads = global_mem->g_num_workers;
 
 	for (idx = 0; idx < NUM_TEST_BARRIERS; idx++) {
 		odp_barrier_init(&global_mem->test_barriers[idx], num_threads);
@@ -924,7 +924,7 @@  void synchronizers_test_no_barrier_functional(void)
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	barrier_test_init();
 	odp_cunit_thread_create(no_barrier_functional_test, &arg);
 	odp_cunit_thread_exit(&arg);
@@ -934,7 +934,7 @@  void synchronizers_test_barrier_functional(void)
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	barrier_test_init();
 	odp_cunit_thread_create(barrier_functional_test, &arg);
 	odp_cunit_thread_exit(&arg);
@@ -951,7 +951,7 @@  void synchronizers_test_no_lock_functional(void)
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	odp_cunit_thread_create(no_lock_functional_test, &arg);
 	odp_cunit_thread_exit(&arg);
 }
@@ -966,7 +966,7 @@  void synchronizers_test_spinlock_api(void)
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	odp_cunit_thread_create(spinlock_api_tests, &arg);
 	odp_cunit_thread_exit(&arg);
 }
@@ -975,7 +975,7 @@  void synchronizers_test_spinlock_functional(void)
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	odp_spinlock_init(&global_mem->global_spinlock);
 	odp_cunit_thread_create(spinlock_functional_test, &arg);
 	odp_cunit_thread_exit(&arg);
@@ -992,7 +992,7 @@  void synchronizers_test_ticketlock_api(void)
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	odp_cunit_thread_create(ticketlock_api_tests, &arg);
 	odp_cunit_thread_exit(&arg);
 }
@@ -1001,7 +1001,7 @@  void synchronizers_test_ticketlock_functional(void)
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	odp_ticketlock_init(&global_mem->global_ticketlock);
 
 	odp_cunit_thread_create(ticketlock_functional_test, &arg);
@@ -1019,7 +1019,7 @@  void synchronizers_test_rwlock_api(void)
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	odp_cunit_thread_create(rwlock_api_tests, &arg);
 	odp_cunit_thread_exit(&arg);
 }
@@ -1028,7 +1028,7 @@  void synchronizers_test_rwlock_functional(void)
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	odp_rwlock_init(&global_mem->global_rwlock);
 	odp_cunit_thread_create(rwlock_functional_test, &arg);
 	odp_cunit_thread_exit(&arg);
@@ -1044,7 +1044,7 @@  int synchronizers_suite_init(void)
 {
 	uint32_t num_threads, idx;
 
-	num_threads = global_mem->g_num_threads;
+	num_threads = global_mem->g_num_workers;
 	odp_barrier_init(&global_mem->global_barrier, num_threads);
 	for (idx = 0; idx < NUM_RESYNC_BARRIERS; idx++)
 		odp_barrier_init(&global_mem->barrier_array[idx], num_threads);
@@ -1054,7 +1054,6 @@  int synchronizers_suite_init(void)
 
 int synchronizers_init(void)
 {
-	uint32_t workers_count, max_threads;
 	int ret = 0;
 	odp_cpumask_t mask;
 
@@ -1078,25 +1077,12 @@  int synchronizers_init(void)
 	global_mem = odp_shm_addr(global_shm);
 	memset(global_mem, 0, sizeof(global_shared_mem_t));
 
-	global_mem->g_num_threads = MAX_WORKERS;
+	global_mem->g_num_workers = odp_cpumask_def_worker(&mask, 0);
 	global_mem->g_iterations = MAX_ITERATIONS;
 	global_mem->g_verbose = VERBOSE;
 
-	workers_count = odp_cpumask_def_worker(&mask, 0);
-
-	max_threads = (workers_count >= MAX_WORKERS) ?
-			MAX_WORKERS : workers_count;
-
-	if (max_threads < global_mem->g_num_threads) {
-		printf("Requested num of threads is too large\n");
-		printf("reducing from %" PRIu32 " to %" PRIu32 "\n",
-		       global_mem->g_num_threads,
-		       max_threads);
-		global_mem->g_num_threads = max_threads;
-	}
-
-	printf("Num of threads used = %" PRIu32 "\n",
-	       global_mem->g_num_threads);
+	printf("Num of workers used = %" PRIu32 "\n",
+	       global_mem->g_num_workers);
 
 	return ret;
 }
@@ -1158,7 +1144,7 @@  static void test_atomic_functional(void *func_ptr(void *))
 {
 	pthrd_arg arg;
 
-	arg.numthrds = global_mem->g_num_threads;
+	arg.numthrds = global_mem->g_num_workers;
 	test_atomic_init();
 	test_atomic_store();
 	odp_cunit_thread_create(func_ptr, &arg);
diff --git a/test/validation/timer/timer.c b/test/validation/timer/timer.c
index 7a8b98a..bcba3d4 100644
--- a/test/validation/timer/timer.c
+++ b/test/validation/timer/timer.c
@@ -34,12 +34,6 @@  static odp_timer_pool_t tp;
 /** @private Count of timeouts delivered too late */
 static odp_atomic_u32_t ndelivtoolate;
 
-/** @private min() function */
-static int min(int a, int b)
-{
-	return a < b ? a : b;
-}
-
 /* @private Timer helper structure */
 struct test_timer {
 	odp_timer_t tim; /* Timer handle */
@@ -441,10 +435,12 @@  void timer_test_odp_timer_all(void)
 	int rc;
 	odp_pool_param_t params;
 	odp_timer_pool_param_t tparam;
+	odp_cpumask_t mask;
+
 	/* Reserve at least one core for running other processes so the timer
 	 * test hopefully can run undisturbed and thus get better timing
 	 * results. */
-	int num_workers = min(odp_cpu_count() - 1, MAX_WORKERS);
+	int num_workers = odp_cpumask_def_worker(&mask, 0) - 1;
 	/* On a single-CPU machine run at least one thread */
 	if (num_workers < 1)
 		num_workers = 1;