Message ID | 1453916193-31377-1-git-send-email-zoltan.kiss@linaro.org |
---|---|
State | Accepted |
Commit | e5c85d3fe092d5eab3a2e02547f4fc8d3ec3391a |
Headers | show |
On 27 January 2016 at 18:36, Zoltan Kiss <zoltan.kiss@linaro.org> wrote: > As per-thread caches might retain some elements, no particular thread > should assume that a certain amount of elements are available at any > time. > Also, to make the high watermark test reliable we should avoid releasing > timers. > > Signed-off-by: Zoltan Kiss <zoltan.kiss@linaro.org> > Reviewed-by: Ola Liljedahl <ola.liljedahl@linaro.org> > --- > > v2: > - keep high watermark test by bookkeeping the exact amounts of allocation. > It > needs a change in order of allocation to make sure no timer is released, > otherwise the watermark becomes unpredictable > - use Ola's subject recommendation > > test/validation/timer/timer.c | 38 +++++++++++++++++++++++++++----------- > 1 file changed, 27 insertions(+), 11 deletions(-) > > diff --git a/test/validation/timer/timer.c b/test/validation/timer/timer.c > index 5d89700..8f00788 100644 > --- a/test/validation/timer/timer.c > +++ b/test/validation/timer/timer.c > @@ -37,6 +37,10 @@ static odp_timer_pool_t tp; > /** @private Count of timeouts delivered too late */ > static odp_atomic_u32_t ndelivtoolate; > > +/** @private Sum of all allocated timers from all threads. Thread-local > + * caches may make this number lower than the capacity of the pool */ > +static odp_atomic_u32_t timers_allocated; > + > /** @private min() function */ > static int min(int a, int b) > { > @@ -274,7 +278,7 @@ static void handle_tmo(odp_event_t ev, bool stale, > uint64_t prev_tick) > static void *worker_entrypoint(void *arg TEST_UNUSED) > { > int thr = odp_thread_id(); > - uint32_t i; > + uint32_t i, allocated; > unsigned seed = thr; > int rc; > > @@ -290,21 +294,30 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) > > /* Prepare all timers */ > for (i = 0; i < NTIMERS; i++) { > - tt[i].tim = odp_timer_alloc(tp, queue, &tt[i]); > - if (tt[i].tim == ODP_TIMER_INVALID) > - CU_FAIL_FATAL("Failed to allocate timer"); > tt[i].ev = odp_timeout_to_event(odp_timeout_alloc(tbp)); > - if (tt[i].ev == ODP_EVENT_INVALID) > - CU_FAIL_FATAL("Failed to allocate timeout"); > + if (tt[i].ev == ODP_EVENT_INVALID) { > + LOG_DBG("Failed to allocate timeout (%d/%d)\n", > + i, NTIMERS); > + break; > + } > + tt[i].tim = odp_timer_alloc(tp, queue, &tt[i]); > + if (tt[i].tim == ODP_TIMER_INVALID) { > + LOG_DBG("Failed to allocate timer (%d/%d)\n", > + i, NTIMERS); > + odp_timeout_free(tt[i].ev); > + break; > + } > tt[i].ev2 = tt[i].ev; > tt[i].tick = TICK_INVALID; > } > + allocated = i; > + odp_atomic_fetch_add_u32(&timers_allocated, allocated); > > odp_barrier_wait(&test_barrier); > > /* Initial set all timers with a random expiration time */ > uint32_t nset = 0; > - for (i = 0; i < NTIMERS; i++) { > + for (i = 0; i < allocated; i++) { > uint64_t tck = odp_timer_current_tick(tp) + 1 + > odp_timer_ns_to_tick(tp, > (rand_r(&seed) % > RANGE_MS) > @@ -336,7 +349,7 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) > nrcv++; > } > prev_tick = odp_timer_current_tick(tp); > - i = rand_r(&seed) % NTIMERS; > + i = rand_r(&seed) % allocated; > if (tt[i].ev == ODP_EVENT_INVALID && > (rand_r(&seed) % 2 == 0)) { > /* Timer active, cancel it */ > @@ -384,7 +397,7 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) > > /* Cancel and free all timers */ > uint32_t nstale = 0; > - for (i = 0; i < NTIMERS; i++) { > + for (i = 0; i < allocated; i++) { > (void)odp_timer_cancel(tt[i].tim, &tt[i].ev); > tt[i].tick = TICK_INVALID; > if (tt[i].ev == ODP_EVENT_INVALID) > @@ -428,7 +441,7 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) > > rc = odp_queue_destroy(queue); > CU_ASSERT(rc == 0); > - for (i = 0; i < NTIMERS; i++) { > + for (i = 0; i < allocated; i++) { > if (tt[i].ev != ODP_EVENT_INVALID) > odp_event_free(tt[i].ev); > } > @@ -504,6 +517,9 @@ void timer_test_odp_timer_all(void) > /* Initialize the shared timeout counter */ > odp_atomic_init_u32(&ndelivtoolate, 0); > > + /* Initialize the number of finally allocated elements */ > + odp_atomic_init_u32(&timers_allocated, 0); > + > /* Create and start worker threads */ > pthrd_arg thrdarg; > thrdarg.testcase = 0; > @@ -520,7 +536,7 @@ void timer_test_odp_timer_all(void) > CU_FAIL("odp_timer_pool_info"); > CU_ASSERT(tpinfo.param.num_timers == (unsigned)num_workers * > NTIMERS); > CU_ASSERT(tpinfo.cur_timers == 0); > - CU_ASSERT(tpinfo.hwm_timers == (unsigned)num_workers * NTIMERS); > + CU_ASSERT(tpinfo.hwm_timers == > odp_atomic_load_u32(&timers_allocated)); > > /* Destroy timer pool, all timers must have been freed */ > odp_timer_pool_destroy(tp); > -- > 1.9.1 > > _______________________________________________ > lng-odp mailing list > lng-odp@lists.linaro.org > https://lists.linaro.org/mailman/listinfo/lng-odp >
Merged, both patches. Maxim. On 01/27/2016 20:36, Zoltan Kiss wrote: > As per-thread caches might retain some elements, no particular thread > should assume that a certain amount of elements are available at any > time. > Also, to make the high watermark test reliable we should avoid releasing > timers. > > Signed-off-by: Zoltan Kiss <zoltan.kiss@linaro.org> > --- > > v2: > - keep high watermark test by bookkeeping the exact amounts of allocation. It > needs a change in order of allocation to make sure no timer is released, > otherwise the watermark becomes unpredictable > - use Ola's subject recommendation > > test/validation/timer/timer.c | 38 +++++++++++++++++++++++++++----------- > 1 file changed, 27 insertions(+), 11 deletions(-) > > diff --git a/test/validation/timer/timer.c b/test/validation/timer/timer.c > index 5d89700..8f00788 100644 > --- a/test/validation/timer/timer.c > +++ b/test/validation/timer/timer.c > @@ -37,6 +37,10 @@ static odp_timer_pool_t tp; > /** @private Count of timeouts delivered too late */ > static odp_atomic_u32_t ndelivtoolate; > > +/** @private Sum of all allocated timers from all threads. Thread-local > + * caches may make this number lower than the capacity of the pool */ > +static odp_atomic_u32_t timers_allocated; > + > /** @private min() function */ > static int min(int a, int b) > { > @@ -274,7 +278,7 @@ static void handle_tmo(odp_event_t ev, bool stale, uint64_t prev_tick) > static void *worker_entrypoint(void *arg TEST_UNUSED) > { > int thr = odp_thread_id(); > - uint32_t i; > + uint32_t i, allocated; > unsigned seed = thr; > int rc; > > @@ -290,21 +294,30 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) > > /* Prepare all timers */ > for (i = 0; i < NTIMERS; i++) { > - tt[i].tim = odp_timer_alloc(tp, queue, &tt[i]); > - if (tt[i].tim == ODP_TIMER_INVALID) > - CU_FAIL_FATAL("Failed to allocate timer"); > tt[i].ev = odp_timeout_to_event(odp_timeout_alloc(tbp)); > - if (tt[i].ev == ODP_EVENT_INVALID) > - CU_FAIL_FATAL("Failed to allocate timeout"); > + if (tt[i].ev == ODP_EVENT_INVALID) { > + LOG_DBG("Failed to allocate timeout (%d/%d)\n", > + i, NTIMERS); > + break; > + } > + tt[i].tim = odp_timer_alloc(tp, queue, &tt[i]); > + if (tt[i].tim == ODP_TIMER_INVALID) { > + LOG_DBG("Failed to allocate timer (%d/%d)\n", > + i, NTIMERS); > + odp_timeout_free(tt[i].ev); > + break; > + } > tt[i].ev2 = tt[i].ev; > tt[i].tick = TICK_INVALID; > } > + allocated = i; > + odp_atomic_fetch_add_u32(&timers_allocated, allocated); > > odp_barrier_wait(&test_barrier); > > /* Initial set all timers with a random expiration time */ > uint32_t nset = 0; > - for (i = 0; i < NTIMERS; i++) { > + for (i = 0; i < allocated; i++) { > uint64_t tck = odp_timer_current_tick(tp) + 1 + > odp_timer_ns_to_tick(tp, > (rand_r(&seed) % RANGE_MS) > @@ -336,7 +349,7 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) > nrcv++; > } > prev_tick = odp_timer_current_tick(tp); > - i = rand_r(&seed) % NTIMERS; > + i = rand_r(&seed) % allocated; > if (tt[i].ev == ODP_EVENT_INVALID && > (rand_r(&seed) % 2 == 0)) { > /* Timer active, cancel it */ > @@ -384,7 +397,7 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) > > /* Cancel and free all timers */ > uint32_t nstale = 0; > - for (i = 0; i < NTIMERS; i++) { > + for (i = 0; i < allocated; i++) { > (void)odp_timer_cancel(tt[i].tim, &tt[i].ev); > tt[i].tick = TICK_INVALID; > if (tt[i].ev == ODP_EVENT_INVALID) > @@ -428,7 +441,7 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) > > rc = odp_queue_destroy(queue); > CU_ASSERT(rc == 0); > - for (i = 0; i < NTIMERS; i++) { > + for (i = 0; i < allocated; i++) { > if (tt[i].ev != ODP_EVENT_INVALID) > odp_event_free(tt[i].ev); > } > @@ -504,6 +517,9 @@ void timer_test_odp_timer_all(void) > /* Initialize the shared timeout counter */ > odp_atomic_init_u32(&ndelivtoolate, 0); > > + /* Initialize the number of finally allocated elements */ > + odp_atomic_init_u32(&timers_allocated, 0); > + > /* Create and start worker threads */ > pthrd_arg thrdarg; > thrdarg.testcase = 0; > @@ -520,7 +536,7 @@ void timer_test_odp_timer_all(void) > CU_FAIL("odp_timer_pool_info"); > CU_ASSERT(tpinfo.param.num_timers == (unsigned)num_workers * NTIMERS); > CU_ASSERT(tpinfo.cur_timers == 0); > - CU_ASSERT(tpinfo.hwm_timers == (unsigned)num_workers * NTIMERS); > + CU_ASSERT(tpinfo.hwm_timers == odp_atomic_load_u32(&timers_allocated)); > > /* Destroy timer pool, all timers must have been freed */ > odp_timer_pool_destroy(tp);
diff --git a/test/validation/timer/timer.c b/test/validation/timer/timer.c index 5d89700..8f00788 100644 --- a/test/validation/timer/timer.c +++ b/test/validation/timer/timer.c @@ -37,6 +37,10 @@ static odp_timer_pool_t tp; /** @private Count of timeouts delivered too late */ static odp_atomic_u32_t ndelivtoolate; +/** @private Sum of all allocated timers from all threads. Thread-local + * caches may make this number lower than the capacity of the pool */ +static odp_atomic_u32_t timers_allocated; + /** @private min() function */ static int min(int a, int b) { @@ -274,7 +278,7 @@ static void handle_tmo(odp_event_t ev, bool stale, uint64_t prev_tick) static void *worker_entrypoint(void *arg TEST_UNUSED) { int thr = odp_thread_id(); - uint32_t i; + uint32_t i, allocated; unsigned seed = thr; int rc; @@ -290,21 +294,30 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) /* Prepare all timers */ for (i = 0; i < NTIMERS; i++) { - tt[i].tim = odp_timer_alloc(tp, queue, &tt[i]); - if (tt[i].tim == ODP_TIMER_INVALID) - CU_FAIL_FATAL("Failed to allocate timer"); tt[i].ev = odp_timeout_to_event(odp_timeout_alloc(tbp)); - if (tt[i].ev == ODP_EVENT_INVALID) - CU_FAIL_FATAL("Failed to allocate timeout"); + if (tt[i].ev == ODP_EVENT_INVALID) { + LOG_DBG("Failed to allocate timeout (%d/%d)\n", + i, NTIMERS); + break; + } + tt[i].tim = odp_timer_alloc(tp, queue, &tt[i]); + if (tt[i].tim == ODP_TIMER_INVALID) { + LOG_DBG("Failed to allocate timer (%d/%d)\n", + i, NTIMERS); + odp_timeout_free(tt[i].ev); + break; + } tt[i].ev2 = tt[i].ev; tt[i].tick = TICK_INVALID; } + allocated = i; + odp_atomic_fetch_add_u32(&timers_allocated, allocated); odp_barrier_wait(&test_barrier); /* Initial set all timers with a random expiration time */ uint32_t nset = 0; - for (i = 0; i < NTIMERS; i++) { + for (i = 0; i < allocated; i++) { uint64_t tck = odp_timer_current_tick(tp) + 1 + odp_timer_ns_to_tick(tp, (rand_r(&seed) % RANGE_MS) @@ -336,7 +349,7 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) nrcv++; } prev_tick = odp_timer_current_tick(tp); - i = rand_r(&seed) % NTIMERS; + i = rand_r(&seed) % allocated; if (tt[i].ev == ODP_EVENT_INVALID && (rand_r(&seed) % 2 == 0)) { /* Timer active, cancel it */ @@ -384,7 +397,7 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) /* Cancel and free all timers */ uint32_t nstale = 0; - for (i = 0; i < NTIMERS; i++) { + for (i = 0; i < allocated; i++) { (void)odp_timer_cancel(tt[i].tim, &tt[i].ev); tt[i].tick = TICK_INVALID; if (tt[i].ev == ODP_EVENT_INVALID) @@ -428,7 +441,7 @@ static void *worker_entrypoint(void *arg TEST_UNUSED) rc = odp_queue_destroy(queue); CU_ASSERT(rc == 0); - for (i = 0; i < NTIMERS; i++) { + for (i = 0; i < allocated; i++) { if (tt[i].ev != ODP_EVENT_INVALID) odp_event_free(tt[i].ev); } @@ -504,6 +517,9 @@ void timer_test_odp_timer_all(void) /* Initialize the shared timeout counter */ odp_atomic_init_u32(&ndelivtoolate, 0); + /* Initialize the number of finally allocated elements */ + odp_atomic_init_u32(&timers_allocated, 0); + /* Create and start worker threads */ pthrd_arg thrdarg; thrdarg.testcase = 0; @@ -520,7 +536,7 @@ void timer_test_odp_timer_all(void) CU_FAIL("odp_timer_pool_info"); CU_ASSERT(tpinfo.param.num_timers == (unsigned)num_workers * NTIMERS); CU_ASSERT(tpinfo.cur_timers == 0); - CU_ASSERT(tpinfo.hwm_timers == (unsigned)num_workers * NTIMERS); + CU_ASSERT(tpinfo.hwm_timers == odp_atomic_load_u32(&timers_allocated)); /* Destroy timer pool, all timers must have been freed */ odp_timer_pool_destroy(tp);
As per-thread caches might retain some elements, no particular thread should assume that a certain amount of elements are available at any time. Also, to make the high watermark test reliable we should avoid releasing timers. Signed-off-by: Zoltan Kiss <zoltan.kiss@linaro.org> --- v2: - keep high watermark test by bookkeeping the exact amounts of allocation. It needs a change in order of allocation to make sure no timer is released, otherwise the watermark becomes unpredictable - use Ola's subject recommendation test/validation/timer/timer.c | 38 +++++++++++++++++++++++++++----------- 1 file changed, 27 insertions(+), 11 deletions(-)