diff mbox series

[2/5] lib/ring: add template to support different element sizes

Message ID 20190828144614.25284-3-honnappa.nagarahalli@arm.com
State Superseded
Headers show
Series lib/ring: templates to support custom element size | expand

Commit Message

Honnappa Nagarahalli Aug. 28, 2019, 2:46 p.m. UTC
Add templates to support creating ring APIs with different
ring element sizes.

Signed-off-by: Honnappa Nagarahalli <honnappa.nagarahalli@arm.com>

Reviewed-by: Dharmik Thakkar <dharmik.thakkar@arm.com>

Reviewed-by: Gavin Hu <gavin.hu@arm.com>

Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>

---
 lib/librte_ring/Makefile            |   4 +-
 lib/librte_ring/meson.build         |   4 +-
 lib/librte_ring/rte_ring_template.c |  46 ++++
 lib/librte_ring/rte_ring_template.h | 330 ++++++++++++++++++++++++++++
 4 files changed, 382 insertions(+), 2 deletions(-)
 create mode 100644 lib/librte_ring/rte_ring_template.c
 create mode 100644 lib/librte_ring/rte_ring_template.h

-- 
2.17.1

Comments

Ananyev, Konstantin Oct. 1, 2019, 11:47 a.m. UTC | #1
> 

> 

> Add templates to support creating ring APIs with different

> ring element sizes.

> 

> Signed-off-by: Honnappa Nagarahalli <honnappa.nagarahalli@arm.com>

> Reviewed-by: Dharmik Thakkar <dharmik.thakkar@arm.com>

> Reviewed-by: Gavin Hu <gavin.hu@arm.com>

> Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>

> ---

>  lib/librte_ring/Makefile            |   4 +-

>  lib/librte_ring/meson.build         |   4 +-

>  lib/librte_ring/rte_ring_template.c |  46 ++++

>  lib/librte_ring/rte_ring_template.h | 330 ++++++++++++++++++++++++++++

>  4 files changed, 382 insertions(+), 2 deletions(-)

>  create mode 100644 lib/librte_ring/rte_ring_template.c

>  create mode 100644 lib/librte_ring/rte_ring_template.h

> 

> diff --git a/lib/librte_ring/Makefile b/lib/librte_ring/Makefile

> index 4c8410229..818898110 100644

> --- a/lib/librte_ring/Makefile

> +++ b/lib/librte_ring/Makefile

> @@ -19,6 +19,8 @@ SRCS-$(CONFIG_RTE_LIBRTE_RING) := rte_ring.c

>  # install includes

>  SYMLINK-$(CONFIG_RTE_LIBRTE_RING)-include := rte_ring.h \

>  					rte_ring_generic.h \

> -					rte_ring_c11_mem.h

> +					rte_ring_c11_mem.h \

> +					rte_ring_template.h \

> +					rte_ring_template.c

> 

>  include $(RTE_SDK)/mk/rte.lib.mk

> diff --git a/lib/librte_ring/meson.build b/lib/librte_ring/meson.build

> index 74219840a..e4e208a7c 100644

> --- a/lib/librte_ring/meson.build

> +++ b/lib/librte_ring/meson.build

> @@ -5,7 +5,9 @@ version = 2

>  sources = files('rte_ring.c')

>  headers = files('rte_ring.h',

>  		'rte_ring_c11_mem.h',

> -		'rte_ring_generic.h')

> +		'rte_ring_generic.h',

> +		'rte_ring_template.h',

> +		'rte_ring_template.c')

> 

>  # rte_ring_create_elem and rte_ring_get_memsize_elem are experimental

>  allow_experimental_apis = true

> diff --git a/lib/librte_ring/rte_ring_template.c b/lib/librte_ring/rte_ring_template.c

> new file mode 100644

> index 000000000..1ca593f95

> --- /dev/null

> +++ b/lib/librte_ring/rte_ring_template.c

> @@ -0,0 +1,46 @@

> +/* SPDX-License-Identifier: BSD-3-Clause

> + * Copyright (c) 2019 Arm Limited

> + */

> +

> +#include <stdio.h>

> +#include <stdarg.h>

> +#include <string.h>

> +#include <stdint.h>

> +#include <inttypes.h>

> +#include <errno.h>

> +#include <sys/queue.h>

> +

> +#include <rte_common.h>

> +#include <rte_log.h>

> +#include <rte_memory.h>

> +#include <rte_memzone.h>

> +#include <rte_malloc.h>

> +#include <rte_launch.h>

> +#include <rte_eal.h>

> +#include <rte_eal_memconfig.h>

> +#include <rte_atomic.h>

> +#include <rte_per_lcore.h>

> +#include <rte_lcore.h>

> +#include <rte_branch_prediction.h>

> +#include <rte_errno.h>

> +#include <rte_string_fns.h>

> +#include <rte_spinlock.h>

> +#include <rte_tailq.h>

> +

> +#include "rte_ring.h"

> +

> +/* return the size of memory occupied by a ring */

> +ssize_t

> +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count)

> +{

> +	return rte_ring_get_memsize_elem(count, RTE_RING_TMPLT_ELEM_SIZE);

> +}

> +

> +/* create the ring */

> +struct rte_ring *

> +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned count,

> +		int socket_id, unsigned flags)

> +{

> +	return rte_ring_create_elem(name, count, RTE_RING_TMPLT_ELEM_SIZE,

> +		socket_id, flags);

> +}

> diff --git a/lib/librte_ring/rte_ring_template.h b/lib/librte_ring/rte_ring_template.h

> new file mode 100644

> index 000000000..b9b14dfbb

> --- /dev/null

> +++ b/lib/librte_ring/rte_ring_template.h

> @@ -0,0 +1,330 @@

> +/* SPDX-License-Identifier: BSD-3-Clause

> + * Copyright (c) 2019 Arm Limited

> + */

> +

> +#ifndef _RTE_RING_TEMPLATE_H_

> +#define _RTE_RING_TEMPLATE_H_

> +

> +#ifdef __cplusplus

> +extern "C" {

> +#endif

> +

> +#include <stdio.h>

> +#include <stdint.h>

> +#include <sys/queue.h>

> +#include <errno.h>

> +#include <rte_common.h>

> +#include <rte_config.h>

> +#include <rte_memory.h>

> +#include <rte_lcore.h>

> +#include <rte_atomic.h>

> +#include <rte_branch_prediction.h>

> +#include <rte_memzone.h>

> +#include <rte_pause.h>

> +#include <rte_ring.h>

> +

> +/* Ring API suffix name - used to append to API names */

> +#ifndef RTE_RING_TMPLT_API_SUFFIX

> +#error RTE_RING_TMPLT_API_SUFFIX not defined

> +#endif

> +

> +/* Ring's element size in bits, should be a power of 2 */

> +#ifndef RTE_RING_TMPLT_ELEM_SIZE

> +#error RTE_RING_TMPLT_ELEM_SIZE not defined

> +#endif

> +

> +/* Type of ring elements */

> +#ifndef RTE_RING_TMPLT_ELEM_TYPE

> +#error RTE_RING_TMPLT_ELEM_TYPE not defined

> +#endif

> +

> +#define _rte_fuse(a, b) a##_##b

> +#define __rte_fuse(a, b) _rte_fuse(a, b)

> +#define __RTE_RING_CONCAT(a) __rte_fuse(a, RTE_RING_TMPLT_API_SUFFIX)

> +

> +/* Calculate the memory size needed for a ring */

> +RTE_RING_TMPLT_EXPERIMENTAL

> +ssize_t __RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count);

> +

> +/* Create a new ring named *name* in memory. */

> +RTE_RING_TMPLT_EXPERIMENTAL

> +struct rte_ring *

> +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned count,

> +					int socket_id, unsigned flags);



Just an idea - probably same thing can be achieved in a different way.
Instead of all these defines - replace ENQUEUE_PTRS/DEQUEUE_PTRS macros
with static inline functions and then make all internal functions, i.e. __rte_ring_do_dequeue()
to accept enqueue/dequeue function pointer as a parameter.
Then let say default rte_ring_mc_dequeue_bulk will do:

rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,
                unsigned int n, unsigned int *available)
{
        return __rte_ring_do_dequeue(r, obj_table, n, RTE_RING_QUEUE_FIXED,
                        __IS_MC, available, dequeue_ptr_default);
}

Then if someone will like to define ring functions forelt_size==X, all he would need to do:
1. define his own enqueue/dequeuer functions.
2. do something like:
rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,
                unsigned int n, unsigned int *available)
{
        return __rte_ring_do_dequeue(r, obj_table, n, RTE_RING_QUEUE_FIXED,
                        __IS_MC, available, dequeue_X);
}

Konstantin


> +

> +/**

> + * @internal Enqueue several objects on the ring

> + */

> +static __rte_always_inline unsigned int

> +__RTE_RING_CONCAT(__rte_ring_do_enqueue)(struct rte_ring *r,

> +		RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> +		enum rte_ring_queue_behavior behavior, unsigned int is_sp,

> +		unsigned int *free_space)

> +{

> +	uint32_t prod_head, prod_next;

> +	uint32_t free_entries;

> +

> +	n = __rte_ring_move_prod_head(r, is_sp, n, behavior,

> +			&prod_head, &prod_next, &free_entries);

> +	if (n == 0)

> +		goto end;

> +

> +	ENQUEUE_PTRS(r, &r[1], prod_head, obj_table, n,

> +		RTE_RING_TMPLT_ELEM_TYPE);

> +

> +	update_tail(&r->prod, prod_head, prod_next, is_sp, 1);

> +end:

> +	if (free_space != NULL)

> +		*free_space = free_entries - n;

> +	return n;

> +}

> +

> +/**

> + * @internal Dequeue several objects from the ring

> + */

> +static __rte_always_inline unsigned int

> +__RTE_RING_CONCAT(__rte_ring_do_dequeue)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> +	enum rte_ring_queue_behavior behavior, unsigned int is_sc,

> +	unsigned int *available)

> +{

> +	uint32_t cons_head, cons_next;

> +	uint32_t entries;

> +

> +	n = __rte_ring_move_cons_head(r, (int)is_sc, n, behavior,

> +			&cons_head, &cons_next, &entries);

> +	if (n == 0)

> +		goto end;

> +

> +	DEQUEUE_PTRS(r, &r[1], cons_head, obj_table, n,

> +		RTE_RING_TMPLT_ELEM_TYPE);

> +

> +	update_tail(&r->cons, cons_head, cons_next, is_sc, 0);

> +

> +end:

> +	if (available != NULL)

> +		*available = entries - n;

> +	return n;

> +}

> +

> +

> +/**

> + * Enqueue several objects on the ring (multi-producers safe).

> + */

> +static __rte_always_inline unsigned int

> +__RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> +	unsigned int *free_space)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> +			RTE_RING_QUEUE_FIXED, __IS_MP, free_space);

> +}

> +

> +/**

> + * Enqueue several objects on a ring (NOT multi-producers safe).

> + */

> +static __rte_always_inline unsigned int

> +__RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> +	unsigned int *free_space)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> +			RTE_RING_QUEUE_FIXED, __IS_SP, free_space);

> +}

> +

> +/**

> + * Enqueue several objects on a ring.

> + */

> +static __rte_always_inline unsigned int

> +__RTE_RING_CONCAT(rte_ring_enqueue_bulk)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> +	unsigned int *free_space)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> +			RTE_RING_QUEUE_FIXED, r->prod.single, free_space);

> +}

> +

> +/**

> + * Enqueue one object on a ring (multi-producers safe).

> + */

> +static __rte_always_inline int

> +__RTE_RING_CONCAT(rte_ring_mp_enqueue)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE obj)

> +{

> +	return __RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(r, &obj, 1, NULL) ?

> +			0 : -ENOBUFS;

> +}

> +

> +/**

> + * Enqueue one object on a ring (NOT multi-producers safe).

> + */

> +static __rte_always_inline int

> +__RTE_RING_CONCAT(rte_ring_sp_enqueue)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE obj)

> +{

> +	return __RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(r, &obj, 1, NULL) ?

> +			0 : -ENOBUFS;

> +}

> +

> +/**

> + * Enqueue one object on a ring.

> + */

> +static __rte_always_inline int

> +__RTE_RING_CONCAT(rte_ring_enqueue)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj)

> +{

> +	return __RTE_RING_CONCAT(rte_ring_enqueue_bulk)(r, obj, 1, NULL) ?

> +			0 : -ENOBUFS;

> +}

> +

> +/**

> + * Dequeue several objects from a ring (multi-consumers safe).

> + */

> +static __rte_always_inline unsigned int

> +__RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> +	unsigned int *available)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> +			RTE_RING_QUEUE_FIXED, __IS_MC, available);

> +}

> +

> +/**

> + * Dequeue several objects from a ring (NOT multi-consumers safe).

> + */

> +static __rte_always_inline unsigned int

> +__RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> +	unsigned int *available)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> +			RTE_RING_QUEUE_FIXED, __IS_SC, available);

> +}

> +

> +/**

> + * Dequeue several objects from a ring.

> + */

> +static __rte_always_inline unsigned int

> +__RTE_RING_CONCAT(rte_ring_dequeue_bulk)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> +	unsigned int *available)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> +			RTE_RING_QUEUE_FIXED, r->cons.single, available);

> +}

> +

> +/**

> + * Dequeue one object from a ring (multi-consumers safe).

> + */

> +static __rte_always_inline int

> +__RTE_RING_CONCAT(rte_ring_mc_dequeue)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_p)

> +{

> +	return __RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(r, obj_p, 1, NULL) ?

> +			0 : -ENOENT;

> +}

> +

> +/**

> + * Dequeue one object from a ring (NOT multi-consumers safe).

> + */

> +static __rte_always_inline int

> +__RTE_RING_CONCAT(rte_ring_sc_dequeue)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_p)

> +{

> +	return __RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(r, obj_p, 1, NULL) ?

> +			0 : -ENOENT;

> +}

> +

> +/**

> + * Dequeue one object from a ring.

> + */

> +static __rte_always_inline int

> +__RTE_RING_CONCAT(rte_ring_dequeue)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_p)

> +{

> +	return __RTE_RING_CONCAT(rte_ring_dequeue_bulk)(r, obj_p, 1, NULL) ?

> +			0 : -ENOENT;

> +}

> +

> +/**

> + * Enqueue several objects on the ring (multi-producers safe).

> + */

> +static __rte_always_inline unsigned

> +__RTE_RING_CONCAT(rte_ring_mp_enqueue_burst)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table,

> +			 unsigned int n, unsigned int *free_space)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> +			RTE_RING_QUEUE_VARIABLE, __IS_MP, free_space);

> +}

> +

> +/**

> + * Enqueue several objects on a ring (NOT multi-producers safe).

> + */

> +static __rte_always_inline unsigned

> +__RTE_RING_CONCAT(rte_ring_sp_enqueue_burst)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table,

> +			 unsigned int n, unsigned int *free_space)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> +			RTE_RING_QUEUE_VARIABLE, __IS_SP, free_space);

> +}

> +

> +/**

> + * Enqueue several objects on a ring.

> + */

> +static __rte_always_inline unsigned

> +__RTE_RING_CONCAT(rte_ring_enqueue_burst)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> +	unsigned int *free_space)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> +			RTE_RING_QUEUE_VARIABLE, r->prod.single, free_space);

> +}

> +

> +/**

> + * Dequeue several objects from a ring (multi-consumers safe). When the request

> + * objects are more than the available objects, only dequeue the actual number

> + * of objects

> + */

> +static __rte_always_inline unsigned

> +__RTE_RING_CONCAT(rte_ring_mc_dequeue_burst)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> +	unsigned int *available)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> +			RTE_RING_QUEUE_VARIABLE, __IS_MC, available);

> +}

> +

> +/**

> + * Dequeue several objects from a ring (NOT multi-consumers safe).When the

> + * request objects are more than the available objects, only dequeue the

> + * actual number of objects

> + */

> +static __rte_always_inline unsigned

> +__RTE_RING_CONCAT(rte_ring_sc_dequeue_burst)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> +	unsigned int *available)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> +			RTE_RING_QUEUE_VARIABLE, __IS_SC, available);

> +}

> +

> +/**

> + * Dequeue multiple objects from a ring up to a maximum number.

> + */

> +static __rte_always_inline unsigned

> +__RTE_RING_CONCAT(rte_ring_dequeue_burst)(struct rte_ring *r,

> +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> +	unsigned int *available)

> +{

> +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> +				RTE_RING_QUEUE_VARIABLE,

> +				r->cons.single, available);

> +}

> +

> +#ifdef __cplusplus

> +}

> +#endif

> +

> +#endif /* _RTE_RING_TEMPLATE_H_ */

> --

> 2.17.1
Honnappa Nagarahalli Oct. 2, 2019, 4:21 a.m. UTC | #2
> > Add templates to support creating ring APIs with different ring

> > element sizes.

> >

> > Signed-off-by: Honnappa Nagarahalli <honnappa.nagarahalli@arm.com>

> > Reviewed-by: Dharmik Thakkar <dharmik.thakkar@arm.com>

> > Reviewed-by: Gavin Hu <gavin.hu@arm.com>

> > Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>

> > ---

> >  lib/librte_ring/Makefile            |   4 +-

> >  lib/librte_ring/meson.build         |   4 +-

> >  lib/librte_ring/rte_ring_template.c |  46 ++++

> > lib/librte_ring/rte_ring_template.h | 330 ++++++++++++++++++++++++++++

> >  4 files changed, 382 insertions(+), 2 deletions(-)  create mode

> > 100644 lib/librte_ring/rte_ring_template.c

> >  create mode 100644 lib/librte_ring/rte_ring_template.h

> >

> > diff --git a/lib/librte_ring/Makefile b/lib/librte_ring/Makefile index

> > 4c8410229..818898110 100644

> > --- a/lib/librte_ring/Makefile

> > +++ b/lib/librte_ring/Makefile

> > @@ -19,6 +19,8 @@ SRCS-$(CONFIG_RTE_LIBRTE_RING) := rte_ring.c  #

> > install includes  SYMLINK-$(CONFIG_RTE_LIBRTE_RING)-include :=

> > rte_ring.h \

> >  					rte_ring_generic.h \

> > -					rte_ring_c11_mem.h

> > +					rte_ring_c11_mem.h \

> > +					rte_ring_template.h \

> > +					rte_ring_template.c

> >

> >  include $(RTE_SDK)/mk/rte.lib.mk

> > diff --git a/lib/librte_ring/meson.build b/lib/librte_ring/meson.build

> > index 74219840a..e4e208a7c 100644

> > --- a/lib/librte_ring/meson.build

> > +++ b/lib/librte_ring/meson.build

> > @@ -5,7 +5,9 @@ version = 2

> >  sources = files('rte_ring.c')

> >  headers = files('rte_ring.h',

> >  		'rte_ring_c11_mem.h',

> > -		'rte_ring_generic.h')

> > +		'rte_ring_generic.h',

> > +		'rte_ring_template.h',

> > +		'rte_ring_template.c')

> >

> >  # rte_ring_create_elem and rte_ring_get_memsize_elem are experimental

> > allow_experimental_apis = true diff --git

> > a/lib/librte_ring/rte_ring_template.c

> > b/lib/librte_ring/rte_ring_template.c

> > new file mode 100644

> > index 000000000..1ca593f95

> > --- /dev/null

> > +++ b/lib/librte_ring/rte_ring_template.c

> > @@ -0,0 +1,46 @@

> > +/* SPDX-License-Identifier: BSD-3-Clause

> > + * Copyright (c) 2019 Arm Limited

> > + */

> > +

> > +#include <stdio.h>

> > +#include <stdarg.h>

> > +#include <string.h>

> > +#include <stdint.h>

> > +#include <inttypes.h>

> > +#include <errno.h>

> > +#include <sys/queue.h>

> > +

> > +#include <rte_common.h>

> > +#include <rte_log.h>

> > +#include <rte_memory.h>

> > +#include <rte_memzone.h>

> > +#include <rte_malloc.h>

> > +#include <rte_launch.h>

> > +#include <rte_eal.h>

> > +#include <rte_eal_memconfig.h>

> > +#include <rte_atomic.h>

> > +#include <rte_per_lcore.h>

> > +#include <rte_lcore.h>

> > +#include <rte_branch_prediction.h>

> > +#include <rte_errno.h>

> > +#include <rte_string_fns.h>

> > +#include <rte_spinlock.h>

> > +#include <rte_tailq.h>

> > +

> > +#include "rte_ring.h"

> > +

> > +/* return the size of memory occupied by a ring */ ssize_t

> > +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count) {

> > +	return rte_ring_get_memsize_elem(count,

> RTE_RING_TMPLT_ELEM_SIZE); }

> > +

> > +/* create the ring */

> > +struct rte_ring *

> > +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned count,

> > +		int socket_id, unsigned flags)

> > +{

> > +	return rte_ring_create_elem(name, count,

> RTE_RING_TMPLT_ELEM_SIZE,

> > +		socket_id, flags);

> > +}

> > diff --git a/lib/librte_ring/rte_ring_template.h

> > b/lib/librte_ring/rte_ring_template.h

> > new file mode 100644

> > index 000000000..b9b14dfbb

> > --- /dev/null

> > +++ b/lib/librte_ring/rte_ring_template.h

> > @@ -0,0 +1,330 @@

> > +/* SPDX-License-Identifier: BSD-3-Clause

> > + * Copyright (c) 2019 Arm Limited

> > + */

> > +

> > +#ifndef _RTE_RING_TEMPLATE_H_

> > +#define _RTE_RING_TEMPLATE_H_

> > +

> > +#ifdef __cplusplus

> > +extern "C" {

> > +#endif

> > +

> > +#include <stdio.h>

> > +#include <stdint.h>

> > +#include <sys/queue.h>

> > +#include <errno.h>

> > +#include <rte_common.h>

> > +#include <rte_config.h>

> > +#include <rte_memory.h>

> > +#include <rte_lcore.h>

> > +#include <rte_atomic.h>

> > +#include <rte_branch_prediction.h>

> > +#include <rte_memzone.h>

> > +#include <rte_pause.h>

> > +#include <rte_ring.h>

> > +

> > +/* Ring API suffix name - used to append to API names */ #ifndef

> > +RTE_RING_TMPLT_API_SUFFIX #error RTE_RING_TMPLT_API_SUFFIX not

> > +defined #endif

> > +

> > +/* Ring's element size in bits, should be a power of 2 */ #ifndef

> > +RTE_RING_TMPLT_ELEM_SIZE #error RTE_RING_TMPLT_ELEM_SIZE not

> defined

> > +#endif

> > +

> > +/* Type of ring elements */

> > +#ifndef RTE_RING_TMPLT_ELEM_TYPE

> > +#error RTE_RING_TMPLT_ELEM_TYPE not defined #endif

> > +

> > +#define _rte_fuse(a, b) a##_##b

> > +#define __rte_fuse(a, b) _rte_fuse(a, b) #define __RTE_RING_CONCAT(a)

> > +__rte_fuse(a, RTE_RING_TMPLT_API_SUFFIX)

> > +

> > +/* Calculate the memory size needed for a ring */

> > +RTE_RING_TMPLT_EXPERIMENTAL ssize_t

> > +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count);

> > +

> > +/* Create a new ring named *name* in memory. */

> > +RTE_RING_TMPLT_EXPERIMENTAL struct rte_ring *

> > +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned count,

> > +					int socket_id, unsigned flags);

> 

> 

> Just an idea - probably same thing can be achieved in a different way.

> Instead of all these defines - replace ENQUEUE_PTRS/DEQUEUE_PTRS macros

> with static inline functions and then make all internal functions, i.e.

> __rte_ring_do_dequeue()

> to accept enqueue/dequeue function pointer as a parameter.

> Then let say default rte_ring_mc_dequeue_bulk will do:

> 

> rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

>                 unsigned int n, unsigned int *available)

> {

>         return __rte_ring_do_dequeue(r, obj_table, n, RTE_RING_QUEUE_FIXED,

>                         __IS_MC, available, dequeue_ptr_default);

> }

> 

> Then if someone will like to define ring functions forelt_size==X, all he would

> need to do:

> 1. define his own enqueue/dequeuer functions.

> 2. do something like:

> rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

>                 unsigned int n, unsigned int *available)

> {

>         return __rte_ring_do_dequeue(r, obj_table, n, RTE_RING_QUEUE_FIXED,

>                         __IS_MC, available, dequeue_X);

> }

> 

> Konstantin

Thanks for the feedback/idea. The goal of this patch was to make it simple enough to define APIs to store any element size without code duplication. With this patch, the user has to write ~4 lines of code to get APIs for any element size. I would like to keep the goal still the same.

If we have to avoid the macro-fest, the main problem that needs to be addressed is - how to represent different sizes of element types in a generic way? IMO, we can do this by defining the element type to be a multiple of uint32_t (I do not think we need to go to uint16_t).

For ex:
rte_ring_mp_enqueue_bulk_objs(struct rte_ring *r,
                uint32_t *obj_table, unsigned int num_objs,
                unsigned int n,
                enum rte_ring_queue_behavior behavior, unsigned int is_sp,
                unsigned int *free_space)
{
}

This approach would ensure that we have generic enough APIs and they can be used for elements of any size. But the element itself needs to be a multiple of 32b - I think this should not be a concern.

The API suffix definitely needs to be better, any suggestions?

> 

> 

> > +

> > +/**

> > + * @internal Enqueue several objects on the ring

> > + */

> > +static __rte_always_inline unsigned int

> > +__RTE_RING_CONCAT(__rte_ring_do_enqueue)(struct rte_ring *r,

> > +		RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int

> n,

> > +		enum rte_ring_queue_behavior behavior, unsigned int is_sp,

> > +		unsigned int *free_space)

> > +{

> > +	uint32_t prod_head, prod_next;

> > +	uint32_t free_entries;

> > +

> > +	n = __rte_ring_move_prod_head(r, is_sp, n, behavior,

> > +			&prod_head, &prod_next, &free_entries);

> > +	if (n == 0)

> > +		goto end;

> > +

> > +	ENQUEUE_PTRS(r, &r[1], prod_head, obj_table, n,

> > +		RTE_RING_TMPLT_ELEM_TYPE);

> > +

> > +	update_tail(&r->prod, prod_head, prod_next, is_sp, 1);

> > +end:

> > +	if (free_space != NULL)

> > +		*free_space = free_entries - n;

> > +	return n;

> > +}

> > +

> > +/**

> > + * @internal Dequeue several objects from the ring

> > + */

> > +static __rte_always_inline unsigned int

> > +__RTE_RING_CONCAT(__rte_ring_do_dequeue)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > +	enum rte_ring_queue_behavior behavior, unsigned int is_sc,

> > +	unsigned int *available)

> > +{

> > +	uint32_t cons_head, cons_next;

> > +	uint32_t entries;

> > +

> > +	n = __rte_ring_move_cons_head(r, (int)is_sc, n, behavior,

> > +			&cons_head, &cons_next, &entries);

> > +	if (n == 0)

> > +		goto end;

> > +

> > +	DEQUEUE_PTRS(r, &r[1], cons_head, obj_table, n,

> > +		RTE_RING_TMPLT_ELEM_TYPE);

> > +

> > +	update_tail(&r->cons, cons_head, cons_next, is_sc, 0);

> > +

> > +end:

> > +	if (available != NULL)

> > +		*available = entries - n;

> > +	return n;

> > +}

> > +

> > +

> > +/**

> > + * Enqueue several objects on the ring (multi-producers safe).

> > + */

> > +static __rte_always_inline unsigned int

> > +__RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> > +	unsigned int *free_space)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_FIXED, __IS_MP, free_space);

> > +}

> > +

> > +/**

> > + * Enqueue several objects on a ring (NOT multi-producers safe).

> > + */

> > +static __rte_always_inline unsigned int

> > +__RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> > +	unsigned int *free_space)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_FIXED, __IS_SP, free_space);

> > +}

> > +

> > +/**

> > + * Enqueue several objects on a ring.

> > + */

> > +static __rte_always_inline unsigned int

> > +__RTE_RING_CONCAT(rte_ring_enqueue_bulk)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> > +	unsigned int *free_space)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_FIXED, r->prod.single, free_space);

> > +}

> > +

> > +/**

> > + * Enqueue one object on a ring (multi-producers safe).

> > + */

> > +static __rte_always_inline int

> > +__RTE_RING_CONCAT(rte_ring_mp_enqueue)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE obj)

> > +{

> > +	return __RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(r, &obj, 1,

> NULL) ?

> > +			0 : -ENOBUFS;

> > +}

> > +

> > +/**

> > + * Enqueue one object on a ring (NOT multi-producers safe).

> > + */

> > +static __rte_always_inline int

> > +__RTE_RING_CONCAT(rte_ring_sp_enqueue)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE obj)

> > +{

> > +	return __RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(r, &obj, 1,

> NULL) ?

> > +			0 : -ENOBUFS;

> > +}

> > +

> > +/**

> > + * Enqueue one object on a ring.

> > + */

> > +static __rte_always_inline int

> > +__RTE_RING_CONCAT(rte_ring_enqueue)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj)

> > +{

> > +	return __RTE_RING_CONCAT(rte_ring_enqueue_bulk)(r, obj, 1,

> NULL) ?

> > +			0 : -ENOBUFS;

> > +}

> > +

> > +/**

> > + * Dequeue several objects from a ring (multi-consumers safe).

> > + */

> > +static __rte_always_inline unsigned int

> > +__RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > +	unsigned int *available)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_FIXED, __IS_MC, available);

> > +}

> > +

> > +/**

> > + * Dequeue several objects from a ring (NOT multi-consumers safe).

> > + */

> > +static __rte_always_inline unsigned int

> > +__RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > +	unsigned int *available)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_FIXED, __IS_SC, available);

> > +}

> > +

> > +/**

> > + * Dequeue several objects from a ring.

> > + */

> > +static __rte_always_inline unsigned int

> > +__RTE_RING_CONCAT(rte_ring_dequeue_bulk)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > +	unsigned int *available)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_FIXED, r->cons.single, available);

> > +}

> > +

> > +/**

> > + * Dequeue one object from a ring (multi-consumers safe).

> > + */

> > +static __rte_always_inline int

> > +__RTE_RING_CONCAT(rte_ring_mc_dequeue)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_p)

> > +{

> > +	return __RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(r, obj_p, 1,

> NULL) ?

> > +			0 : -ENOENT;

> > +}

> > +

> > +/**

> > + * Dequeue one object from a ring (NOT multi-consumers safe).

> > + */

> > +static __rte_always_inline int

> > +__RTE_RING_CONCAT(rte_ring_sc_dequeue)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_p)

> > +{

> > +	return __RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(r, obj_p, 1,

> NULL) ?

> > +			0 : -ENOENT;

> > +}

> > +

> > +/**

> > + * Dequeue one object from a ring.

> > + */

> > +static __rte_always_inline int

> > +__RTE_RING_CONCAT(rte_ring_dequeue)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_p)

> > +{

> > +	return __RTE_RING_CONCAT(rte_ring_dequeue_bulk)(r, obj_p, 1,

> NULL) ?

> > +			0 : -ENOENT;

> > +}

> > +

> > +/**

> > + * Enqueue several objects on the ring (multi-producers safe).

> > + */

> > +static __rte_always_inline unsigned

> > +__RTE_RING_CONCAT(rte_ring_mp_enqueue_burst)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table,

> > +			 unsigned int n, unsigned int *free_space)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_VARIABLE, __IS_MP, free_space);

> > +}

> > +

> > +/**

> > + * Enqueue several objects on a ring (NOT multi-producers safe).

> > + */

> > +static __rte_always_inline unsigned

> > +__RTE_RING_CONCAT(rte_ring_sp_enqueue_burst)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table,

> > +			 unsigned int n, unsigned int *free_space)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_VARIABLE, __IS_SP, free_space);

> > +}

> > +

> > +/**

> > + * Enqueue several objects on a ring.

> > + */

> > +static __rte_always_inline unsigned

> > +__RTE_RING_CONCAT(rte_ring_enqueue_burst)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > +	unsigned int *free_space)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_VARIABLE, r->prod.single,

> free_space);

> > +}

> > +

> > +/**

> > + * Dequeue several objects from a ring (multi-consumers safe). When the

> request

> > + * objects are more than the available objects, only dequeue the actual

> number

> > + * of objects

> > + */

> > +static __rte_always_inline unsigned

> > +__RTE_RING_CONCAT(rte_ring_mc_dequeue_burst)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > +	unsigned int *available)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_VARIABLE, __IS_MC, available);

> > +}

> > +

> > +/**

> > + * Dequeue several objects from a ring (NOT multi-consumers safe).When

> the

> > + * request objects are more than the available objects, only dequeue the

> > + * actual number of objects

> > + */

> > +static __rte_always_inline unsigned

> > +__RTE_RING_CONCAT(rte_ring_sc_dequeue_burst)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > +	unsigned int *available)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > +			RTE_RING_QUEUE_VARIABLE, __IS_SC, available);

> > +}

> > +

> > +/**

> > + * Dequeue multiple objects from a ring up to a maximum number.

> > + */

> > +static __rte_always_inline unsigned

> > +__RTE_RING_CONCAT(rte_ring_dequeue_burst)(struct rte_ring *r,

> > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > +	unsigned int *available)

> > +{

> > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > +				RTE_RING_QUEUE_VARIABLE,

> > +				r->cons.single, available);

> > +}

> > +

> > +#ifdef __cplusplus

> > +}

> > +#endif

> > +

> > +#endif /* _RTE_RING_TEMPLATE_H_ */

> > --

> > 2.17.1
Ananyev, Konstantin Oct. 2, 2019, 8:39 a.m. UTC | #3
> -----Original Message-----

> From: Honnappa Nagarahalli [mailto:Honnappa.Nagarahalli@arm.com]

> Sent: Wednesday, October 2, 2019 5:22 AM

> To: Ananyev, Konstantin <konstantin.ananyev@intel.com>; olivier.matz@6wind.com; Wang, Yipeng1 <yipeng1.wang@intel.com>; Gobriel,

> Sameh <sameh.gobriel@intel.com>; Richardson, Bruce <bruce.richardson@intel.com>; De Lara Guarch, Pablo

> <pablo.de.lara.guarch@intel.com>

> Cc: dev@dpdk.org; Dharmik Thakkar <Dharmik.Thakkar@arm.com>; Gavin Hu (Arm Technology China) <Gavin.Hu@arm.com>; Ruifeng

> Wang (Arm Technology China) <Ruifeng.Wang@arm.com>; Honnappa Nagarahalli <Honnappa.Nagarahalli@arm.com>; nd

> <nd@arm.com>; nd <nd@arm.com>

> Subject: RE: [dpdk-dev] [PATCH 2/5] lib/ring: add template to support different element sizes

> 

> > > Add templates to support creating ring APIs with different ring

> > > element sizes.

> > >

> > > Signed-off-by: Honnappa Nagarahalli <honnappa.nagarahalli@arm.com>

> > > Reviewed-by: Dharmik Thakkar <dharmik.thakkar@arm.com>

> > > Reviewed-by: Gavin Hu <gavin.hu@arm.com>

> > > Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>

> > > ---

> > >  lib/librte_ring/Makefile            |   4 +-

> > >  lib/librte_ring/meson.build         |   4 +-

> > >  lib/librte_ring/rte_ring_template.c |  46 ++++

> > > lib/librte_ring/rte_ring_template.h | 330 ++++++++++++++++++++++++++++

> > >  4 files changed, 382 insertions(+), 2 deletions(-)  create mode

> > > 100644 lib/librte_ring/rte_ring_template.c

> > >  create mode 100644 lib/librte_ring/rte_ring_template.h

> > >

> > > diff --git a/lib/librte_ring/Makefile b/lib/librte_ring/Makefile index

> > > 4c8410229..818898110 100644

> > > --- a/lib/librte_ring/Makefile

> > > +++ b/lib/librte_ring/Makefile

> > > @@ -19,6 +19,8 @@ SRCS-$(CONFIG_RTE_LIBRTE_RING) := rte_ring.c  #

> > > install includes  SYMLINK-$(CONFIG_RTE_LIBRTE_RING)-include :=

> > > rte_ring.h \

> > >  					rte_ring_generic.h \

> > > -					rte_ring_c11_mem.h

> > > +					rte_ring_c11_mem.h \

> > > +					rte_ring_template.h \

> > > +					rte_ring_template.c

> > >

> > >  include $(RTE_SDK)/mk/rte.lib.mk

> > > diff --git a/lib/librte_ring/meson.build b/lib/librte_ring/meson.build

> > > index 74219840a..e4e208a7c 100644

> > > --- a/lib/librte_ring/meson.build

> > > +++ b/lib/librte_ring/meson.build

> > > @@ -5,7 +5,9 @@ version = 2

> > >  sources = files('rte_ring.c')

> > >  headers = files('rte_ring.h',

> > >  		'rte_ring_c11_mem.h',

> > > -		'rte_ring_generic.h')

> > > +		'rte_ring_generic.h',

> > > +		'rte_ring_template.h',

> > > +		'rte_ring_template.c')

> > >

> > >  # rte_ring_create_elem and rte_ring_get_memsize_elem are experimental

> > > allow_experimental_apis = true diff --git

> > > a/lib/librte_ring/rte_ring_template.c

> > > b/lib/librte_ring/rte_ring_template.c

> > > new file mode 100644

> > > index 000000000..1ca593f95

> > > --- /dev/null

> > > +++ b/lib/librte_ring/rte_ring_template.c

> > > @@ -0,0 +1,46 @@

> > > +/* SPDX-License-Identifier: BSD-3-Clause

> > > + * Copyright (c) 2019 Arm Limited

> > > + */

> > > +

> > > +#include <stdio.h>

> > > +#include <stdarg.h>

> > > +#include <string.h>

> > > +#include <stdint.h>

> > > +#include <inttypes.h>

> > > +#include <errno.h>

> > > +#include <sys/queue.h>

> > > +

> > > +#include <rte_common.h>

> > > +#include <rte_log.h>

> > > +#include <rte_memory.h>

> > > +#include <rte_memzone.h>

> > > +#include <rte_malloc.h>

> > > +#include <rte_launch.h>

> > > +#include <rte_eal.h>

> > > +#include <rte_eal_memconfig.h>

> > > +#include <rte_atomic.h>

> > > +#include <rte_per_lcore.h>

> > > +#include <rte_lcore.h>

> > > +#include <rte_branch_prediction.h>

> > > +#include <rte_errno.h>

> > > +#include <rte_string_fns.h>

> > > +#include <rte_spinlock.h>

> > > +#include <rte_tailq.h>

> > > +

> > > +#include "rte_ring.h"

> > > +

> > > +/* return the size of memory occupied by a ring */ ssize_t

> > > +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count) {

> > > +	return rte_ring_get_memsize_elem(count,

> > RTE_RING_TMPLT_ELEM_SIZE); }

> > > +

> > > +/* create the ring */

> > > +struct rte_ring *

> > > +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned count,

> > > +		int socket_id, unsigned flags)

> > > +{

> > > +	return rte_ring_create_elem(name, count,

> > RTE_RING_TMPLT_ELEM_SIZE,

> > > +		socket_id, flags);

> > > +}

> > > diff --git a/lib/librte_ring/rte_ring_template.h

> > > b/lib/librte_ring/rte_ring_template.h

> > > new file mode 100644

> > > index 000000000..b9b14dfbb

> > > --- /dev/null

> > > +++ b/lib/librte_ring/rte_ring_template.h

> > > @@ -0,0 +1,330 @@

> > > +/* SPDX-License-Identifier: BSD-3-Clause

> > > + * Copyright (c) 2019 Arm Limited

> > > + */

> > > +

> > > +#ifndef _RTE_RING_TEMPLATE_H_

> > > +#define _RTE_RING_TEMPLATE_H_

> > > +

> > > +#ifdef __cplusplus

> > > +extern "C" {

> > > +#endif

> > > +

> > > +#include <stdio.h>

> > > +#include <stdint.h>

> > > +#include <sys/queue.h>

> > > +#include <errno.h>

> > > +#include <rte_common.h>

> > > +#include <rte_config.h>

> > > +#include <rte_memory.h>

> > > +#include <rte_lcore.h>

> > > +#include <rte_atomic.h>

> > > +#include <rte_branch_prediction.h>

> > > +#include <rte_memzone.h>

> > > +#include <rte_pause.h>

> > > +#include <rte_ring.h>

> > > +

> > > +/* Ring API suffix name - used to append to API names */ #ifndef

> > > +RTE_RING_TMPLT_API_SUFFIX #error RTE_RING_TMPLT_API_SUFFIX not

> > > +defined #endif

> > > +

> > > +/* Ring's element size in bits, should be a power of 2 */ #ifndef

> > > +RTE_RING_TMPLT_ELEM_SIZE #error RTE_RING_TMPLT_ELEM_SIZE not

> > defined

> > > +#endif

> > > +

> > > +/* Type of ring elements */

> > > +#ifndef RTE_RING_TMPLT_ELEM_TYPE

> > > +#error RTE_RING_TMPLT_ELEM_TYPE not defined #endif

> > > +

> > > +#define _rte_fuse(a, b) a##_##b

> > > +#define __rte_fuse(a, b) _rte_fuse(a, b) #define __RTE_RING_CONCAT(a)

> > > +__rte_fuse(a, RTE_RING_TMPLT_API_SUFFIX)

> > > +

> > > +/* Calculate the memory size needed for a ring */

> > > +RTE_RING_TMPLT_EXPERIMENTAL ssize_t

> > > +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count);

> > > +

> > > +/* Create a new ring named *name* in memory. */

> > > +RTE_RING_TMPLT_EXPERIMENTAL struct rte_ring *

> > > +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned count,

> > > +					int socket_id, unsigned flags);

> >

> >

> > Just an idea - probably same thing can be achieved in a different way.

> > Instead of all these defines - replace ENQUEUE_PTRS/DEQUEUE_PTRS macros

> > with static inline functions and then make all internal functions, i.e.

> > __rte_ring_do_dequeue()

> > to accept enqueue/dequeue function pointer as a parameter.

> > Then let say default rte_ring_mc_dequeue_bulk will do:

> >

> > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> >                 unsigned int n, unsigned int *available)

> > {

> >         return __rte_ring_do_dequeue(r, obj_table, n, RTE_RING_QUEUE_FIXED,

> >                         __IS_MC, available, dequeue_ptr_default);

> > }

> >

> > Then if someone will like to define ring functions forelt_size==X, all he would

> > need to do:

> > 1. define his own enqueue/dequeuer functions.

> > 2. do something like:

> > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> >                 unsigned int n, unsigned int *available)

> > {

> >         return __rte_ring_do_dequeue(r, obj_table, n, RTE_RING_QUEUE_FIXED,

> >                         __IS_MC, available, dequeue_X);

> > }

> >

> > Konstantin

> Thanks for the feedback/idea. The goal of this patch was to make it simple enough to define APIs to store any element size without code

> duplication. 


Well, then if we store elt_size inside the ring, it should be easy enough
to add  to the API generic functions that would use memcpy(or rte_memcpy) for enqueue/dequeue.
Yes, it might be slower than existing (8B per elem), but might be still acceptable.

>With this patch, the user has to write ~4 lines of code to get APIs for any element size. I would like to keep the goal still the

> same.

> 

> If we have to avoid the macro-fest, the main problem that needs to be addressed is - how to represent different sizes of element types in a

> generic way? IMO, we can do this by defining the element type to be a multiple of uint32_t (I do not think we need to go to uint16_t).

> 

> For ex:

> rte_ring_mp_enqueue_bulk_objs(struct rte_ring *r,

>                 uint32_t *obj_table, unsigned int num_objs,

>                 unsigned int n,

>                 enum rte_ring_queue_behavior behavior, unsigned int is_sp,

>                 unsigned int *free_space)

> {

> }

> 

> This approach would ensure that we have generic enough APIs and they can be used for elements of any size. But the element itself needs

> to be a multiple of 32b - I think this should not be a concern.

> 

> The API suffix definitely needs to be better, any suggestions?


> 

> >

> >

> > > +

> > > +/**

> > > + * @internal Enqueue several objects on the ring

> > > + */

> > > +static __rte_always_inline unsigned int

> > > +__RTE_RING_CONCAT(__rte_ring_do_enqueue)(struct rte_ring *r,

> > > +		RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int

> > n,

> > > +		enum rte_ring_queue_behavior behavior, unsigned int is_sp,

> > > +		unsigned int *free_space)

> > > +{

> > > +	uint32_t prod_head, prod_next;

> > > +	uint32_t free_entries;

> > > +

> > > +	n = __rte_ring_move_prod_head(r, is_sp, n, behavior,

> > > +			&prod_head, &prod_next, &free_entries);

> > > +	if (n == 0)

> > > +		goto end;

> > > +

> > > +	ENQUEUE_PTRS(r, &r[1], prod_head, obj_table, n,

> > > +		RTE_RING_TMPLT_ELEM_TYPE);

> > > +

> > > +	update_tail(&r->prod, prod_head, prod_next, is_sp, 1);

> > > +end:

> > > +	if (free_space != NULL)

> > > +		*free_space = free_entries - n;

> > > +	return n;

> > > +}

> > > +

> > > +/**

> > > + * @internal Dequeue several objects from the ring

> > > + */

> > > +static __rte_always_inline unsigned int

> > > +__RTE_RING_CONCAT(__rte_ring_do_dequeue)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > +	enum rte_ring_queue_behavior behavior, unsigned int is_sc,

> > > +	unsigned int *available)

> > > +{

> > > +	uint32_t cons_head, cons_next;

> > > +	uint32_t entries;

> > > +

> > > +	n = __rte_ring_move_cons_head(r, (int)is_sc, n, behavior,

> > > +			&cons_head, &cons_next, &entries);

> > > +	if (n == 0)

> > > +		goto end;

> > > +

> > > +	DEQUEUE_PTRS(r, &r[1], cons_head, obj_table, n,

> > > +		RTE_RING_TMPLT_ELEM_TYPE);

> > > +

> > > +	update_tail(&r->cons, cons_head, cons_next, is_sc, 0);

> > > +

> > > +end:

> > > +	if (available != NULL)

> > > +		*available = entries - n;

> > > +	return n;

> > > +}

> > > +

> > > +

> > > +/**

> > > + * Enqueue several objects on the ring (multi-producers safe).

> > > + */

> > > +static __rte_always_inline unsigned int

> > > +__RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> > > +	unsigned int *free_space)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_FIXED, __IS_MP, free_space);

> > > +}

> > > +

> > > +/**

> > > + * Enqueue several objects on a ring (NOT multi-producers safe).

> > > + */

> > > +static __rte_always_inline unsigned int

> > > +__RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> > > +	unsigned int *free_space)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_FIXED, __IS_SP, free_space);

> > > +}

> > > +

> > > +/**

> > > + * Enqueue several objects on a ring.

> > > + */

> > > +static __rte_always_inline unsigned int

> > > +__RTE_RING_CONCAT(rte_ring_enqueue_bulk)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> > > +	unsigned int *free_space)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_FIXED, r->prod.single, free_space);

> > > +}

> > > +

> > > +/**

> > > + * Enqueue one object on a ring (multi-producers safe).

> > > + */

> > > +static __rte_always_inline int

> > > +__RTE_RING_CONCAT(rte_ring_mp_enqueue)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE obj)

> > > +{

> > > +	return __RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(r, &obj, 1,

> > NULL) ?

> > > +			0 : -ENOBUFS;

> > > +}

> > > +

> > > +/**

> > > + * Enqueue one object on a ring (NOT multi-producers safe).

> > > + */

> > > +static __rte_always_inline int

> > > +__RTE_RING_CONCAT(rte_ring_sp_enqueue)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE obj)

> > > +{

> > > +	return __RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(r, &obj, 1,

> > NULL) ?

> > > +			0 : -ENOBUFS;

> > > +}

> > > +

> > > +/**

> > > + * Enqueue one object on a ring.

> > > + */

> > > +static __rte_always_inline int

> > > +__RTE_RING_CONCAT(rte_ring_enqueue)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj)

> > > +{

> > > +	return __RTE_RING_CONCAT(rte_ring_enqueue_bulk)(r, obj, 1,

> > NULL) ?

> > > +			0 : -ENOBUFS;

> > > +}

> > > +

> > > +/**

> > > + * Dequeue several objects from a ring (multi-consumers safe).

> > > + */

> > > +static __rte_always_inline unsigned int

> > > +__RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > +	unsigned int *available)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_FIXED, __IS_MC, available);

> > > +}

> > > +

> > > +/**

> > > + * Dequeue several objects from a ring (NOT multi-consumers safe).

> > > + */

> > > +static __rte_always_inline unsigned int

> > > +__RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > +	unsigned int *available)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_FIXED, __IS_SC, available);

> > > +}

> > > +

> > > +/**

> > > + * Dequeue several objects from a ring.

> > > + */

> > > +static __rte_always_inline unsigned int

> > > +__RTE_RING_CONCAT(rte_ring_dequeue_bulk)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > +	unsigned int *available)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_FIXED, r->cons.single, available);

> > > +}

> > > +

> > > +/**

> > > + * Dequeue one object from a ring (multi-consumers safe).

> > > + */

> > > +static __rte_always_inline int

> > > +__RTE_RING_CONCAT(rte_ring_mc_dequeue)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_p)

> > > +{

> > > +	return __RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(r, obj_p, 1,

> > NULL) ?

> > > +			0 : -ENOENT;

> > > +}

> > > +

> > > +/**

> > > + * Dequeue one object from a ring (NOT multi-consumers safe).

> > > + */

> > > +static __rte_always_inline int

> > > +__RTE_RING_CONCAT(rte_ring_sc_dequeue)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_p)

> > > +{

> > > +	return __RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(r, obj_p, 1,

> > NULL) ?

> > > +			0 : -ENOENT;

> > > +}

> > > +

> > > +/**

> > > + * Dequeue one object from a ring.

> > > + */

> > > +static __rte_always_inline int

> > > +__RTE_RING_CONCAT(rte_ring_dequeue)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_p)

> > > +{

> > > +	return __RTE_RING_CONCAT(rte_ring_dequeue_bulk)(r, obj_p, 1,

> > NULL) ?

> > > +			0 : -ENOENT;

> > > +}

> > > +

> > > +/**

> > > + * Enqueue several objects on the ring (multi-producers safe).

> > > + */

> > > +static __rte_always_inline unsigned

> > > +__RTE_RING_CONCAT(rte_ring_mp_enqueue_burst)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table,

> > > +			 unsigned int n, unsigned int *free_space)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_VARIABLE, __IS_MP, free_space);

> > > +}

> > > +

> > > +/**

> > > + * Enqueue several objects on a ring (NOT multi-producers safe).

> > > + */

> > > +static __rte_always_inline unsigned

> > > +__RTE_RING_CONCAT(rte_ring_sp_enqueue_burst)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table,

> > > +			 unsigned int n, unsigned int *free_space)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_VARIABLE, __IS_SP, free_space);

> > > +}

> > > +

> > > +/**

> > > + * Enqueue several objects on a ring.

> > > + */

> > > +static __rte_always_inline unsigned

> > > +__RTE_RING_CONCAT(rte_ring_enqueue_burst)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > +	unsigned int *free_space)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_VARIABLE, r->prod.single,

> > free_space);

> > > +}

> > > +

> > > +/**

> > > + * Dequeue several objects from a ring (multi-consumers safe). When the

> > request

> > > + * objects are more than the available objects, only dequeue the actual

> > number

> > > + * of objects

> > > + */

> > > +static __rte_always_inline unsigned

> > > +__RTE_RING_CONCAT(rte_ring_mc_dequeue_burst)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > +	unsigned int *available)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_VARIABLE, __IS_MC, available);

> > > +}

> > > +

> > > +/**

> > > + * Dequeue several objects from a ring (NOT multi-consumers safe).When

> > the

> > > + * request objects are more than the available objects, only dequeue the

> > > + * actual number of objects

> > > + */

> > > +static __rte_always_inline unsigned

> > > +__RTE_RING_CONCAT(rte_ring_sc_dequeue_burst)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > +	unsigned int *available)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > +			RTE_RING_QUEUE_VARIABLE, __IS_SC, available);

> > > +}

> > > +

> > > +/**

> > > + * Dequeue multiple objects from a ring up to a maximum number.

> > > + */

> > > +static __rte_always_inline unsigned

> > > +__RTE_RING_CONCAT(rte_ring_dequeue_burst)(struct rte_ring *r,

> > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > +	unsigned int *available)

> > > +{

> > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > +				RTE_RING_QUEUE_VARIABLE,

> > > +				r->cons.single, available);

> > > +}

> > > +

> > > +#ifdef __cplusplus

> > > +}

> > > +#endif

> > > +

> > > +#endif /* _RTE_RING_TEMPLATE_H_ */

> > > --

> > > 2.17.1
Honnappa Nagarahalli Oct. 3, 2019, 3:33 a.m. UTC | #4
<snip>

> >

> > > > Add templates to support creating ring APIs with different ring

> > > > element sizes.

> > > >

> > > > Signed-off-by: Honnappa Nagarahalli <honnappa.nagarahalli@arm.com>

> > > > Reviewed-by: Dharmik Thakkar <dharmik.thakkar@arm.com>

> > > > Reviewed-by: Gavin Hu <gavin.hu@arm.com>

> > > > Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>

> > > > ---

> > > >  lib/librte_ring/Makefile            |   4 +-

> > > >  lib/librte_ring/meson.build         |   4 +-

> > > >  lib/librte_ring/rte_ring_template.c |  46 ++++

> > > > lib/librte_ring/rte_ring_template.h | 330

> > > > ++++++++++++++++++++++++++++

> > > >  4 files changed, 382 insertions(+), 2 deletions(-)  create mode

> > > > 100644 lib/librte_ring/rte_ring_template.c

> > > >  create mode 100644 lib/librte_ring/rte_ring_template.h

> > > >

> > > > diff --git a/lib/librte_ring/Makefile b/lib/librte_ring/Makefile

> > > > index

> > > > 4c8410229..818898110 100644

> > > > --- a/lib/librte_ring/Makefile

> > > > +++ b/lib/librte_ring/Makefile

> > > > @@ -19,6 +19,8 @@ SRCS-$(CONFIG_RTE_LIBRTE_RING) := rte_ring.c  #

> > > > install includes  SYMLINK-$(CONFIG_RTE_LIBRTE_RING)-include :=

> > > > rte_ring.h \

> > > >  					rte_ring_generic.h \

> > > > -					rte_ring_c11_mem.h

> > > > +					rte_ring_c11_mem.h \

> > > > +					rte_ring_template.h \

> > > > +					rte_ring_template.c

> > > >

> > > >  include $(RTE_SDK)/mk/rte.lib.mk

> > > > diff --git a/lib/librte_ring/meson.build

> > > > b/lib/librte_ring/meson.build index 74219840a..e4e208a7c 100644

> > > > --- a/lib/librte_ring/meson.build

> > > > +++ b/lib/librte_ring/meson.build

> > > > @@ -5,7 +5,9 @@ version = 2

> > > >  sources = files('rte_ring.c')

> > > >  headers = files('rte_ring.h',

> > > >  		'rte_ring_c11_mem.h',

> > > > -		'rte_ring_generic.h')

> > > > +		'rte_ring_generic.h',

> > > > +		'rte_ring_template.h',

> > > > +		'rte_ring_template.c')

> > > >

> > > >  # rte_ring_create_elem and rte_ring_get_memsize_elem are

> > > > experimental allow_experimental_apis = true diff --git

> > > > a/lib/librte_ring/rte_ring_template.c

> > > > b/lib/librte_ring/rte_ring_template.c

> > > > new file mode 100644

> > > > index 000000000..1ca593f95

> > > > --- /dev/null

> > > > +++ b/lib/librte_ring/rte_ring_template.c

> > > > @@ -0,0 +1,46 @@

> > > > +/* SPDX-License-Identifier: BSD-3-Clause

> > > > + * Copyright (c) 2019 Arm Limited  */

> > > > +

> > > > +#include <stdio.h>

> > > > +#include <stdarg.h>

> > > > +#include <string.h>

> > > > +#include <stdint.h>

> > > > +#include <inttypes.h>

> > > > +#include <errno.h>

> > > > +#include <sys/queue.h>

> > > > +

> > > > +#include <rte_common.h>

> > > > +#include <rte_log.h>

> > > > +#include <rte_memory.h>

> > > > +#include <rte_memzone.h>

> > > > +#include <rte_malloc.h>

> > > > +#include <rte_launch.h>

> > > > +#include <rte_eal.h>

> > > > +#include <rte_eal_memconfig.h>

> > > > +#include <rte_atomic.h>

> > > > +#include <rte_per_lcore.h>

> > > > +#include <rte_lcore.h>

> > > > +#include <rte_branch_prediction.h> #include <rte_errno.h>

> > > > +#include <rte_string_fns.h> #include <rte_spinlock.h> #include

> > > > +<rte_tailq.h>

> > > > +

> > > > +#include "rte_ring.h"

> > > > +

> > > > +/* return the size of memory occupied by a ring */ ssize_t

> > > > +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count) {

> > > > +	return rte_ring_get_memsize_elem(count,

> > > RTE_RING_TMPLT_ELEM_SIZE); }

> > > > +

> > > > +/* create the ring */

> > > > +struct rte_ring *

> > > > +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned

> count,

> > > > +		int socket_id, unsigned flags)

> > > > +{

> > > > +	return rte_ring_create_elem(name, count,

> > > RTE_RING_TMPLT_ELEM_SIZE,

> > > > +		socket_id, flags);

> > > > +}

> > > > diff --git a/lib/librte_ring/rte_ring_template.h

> > > > b/lib/librte_ring/rte_ring_template.h

> > > > new file mode 100644

> > > > index 000000000..b9b14dfbb

> > > > --- /dev/null

> > > > +++ b/lib/librte_ring/rte_ring_template.h

> > > > @@ -0,0 +1,330 @@

> > > > +/* SPDX-License-Identifier: BSD-3-Clause

> > > > + * Copyright (c) 2019 Arm Limited  */

> > > > +

> > > > +#ifndef _RTE_RING_TEMPLATE_H_

> > > > +#define _RTE_RING_TEMPLATE_H_

> > > > +

> > > > +#ifdef __cplusplus

> > > > +extern "C" {

> > > > +#endif

> > > > +

> > > > +#include <stdio.h>

> > > > +#include <stdint.h>

> > > > +#include <sys/queue.h>

> > > > +#include <errno.h>

> > > > +#include <rte_common.h>

> > > > +#include <rte_config.h>

> > > > +#include <rte_memory.h>

> > > > +#include <rte_lcore.h>

> > > > +#include <rte_atomic.h>

> > > > +#include <rte_branch_prediction.h> #include <rte_memzone.h>

> > > > +#include <rte_pause.h> #include <rte_ring.h>

> > > > +

> > > > +/* Ring API suffix name - used to append to API names */ #ifndef

> > > > +RTE_RING_TMPLT_API_SUFFIX #error RTE_RING_TMPLT_API_SUFFIX

> not

> > > > +defined #endif

> > > > +

> > > > +/* Ring's element size in bits, should be a power of 2 */ #ifndef

> > > > +RTE_RING_TMPLT_ELEM_SIZE #error RTE_RING_TMPLT_ELEM_SIZE

> not

> > > defined

> > > > +#endif

> > > > +

> > > > +/* Type of ring elements */

> > > > +#ifndef RTE_RING_TMPLT_ELEM_TYPE

> > > > +#error RTE_RING_TMPLT_ELEM_TYPE not defined #endif

> > > > +

> > > > +#define _rte_fuse(a, b) a##_##b

> > > > +#define __rte_fuse(a, b) _rte_fuse(a, b) #define

> > > > +__RTE_RING_CONCAT(a) __rte_fuse(a, RTE_RING_TMPLT_API_SUFFIX)

> > > > +

> > > > +/* Calculate the memory size needed for a ring */

> > > > +RTE_RING_TMPLT_EXPERIMENTAL ssize_t

> > > > +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count);

> > > > +

> > > > +/* Create a new ring named *name* in memory. */

> > > > +RTE_RING_TMPLT_EXPERIMENTAL struct rte_ring *

> > > > +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned

> count,

> > > > +					int socket_id, unsigned flags);

> > >

> > >

> > > Just an idea - probably same thing can be achieved in a different way.

> > > Instead of all these defines - replace ENQUEUE_PTRS/DEQUEUE_PTRS

> > > macros with static inline functions and then make all internal functions,

> i.e.

> > > __rte_ring_do_dequeue()

> > > to accept enqueue/dequeue function pointer as a parameter.

> > > Then let say default rte_ring_mc_dequeue_bulk will do:

> > >

> > > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> > >                 unsigned int n, unsigned int *available) {

> > >         return __rte_ring_do_dequeue(r, obj_table, n,

> RTE_RING_QUEUE_FIXED,

> > >                         __IS_MC, available, dequeue_ptr_default); }

> > >

> > > Then if someone will like to define ring functions forelt_size==X,

> > > all he would need to do:

> > > 1. define his own enqueue/dequeuer functions.

> > > 2. do something like:

> > > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> > >                 unsigned int n, unsigned int *available) {

> > >         return __rte_ring_do_dequeue(r, obj_table, n,

> RTE_RING_QUEUE_FIXED,

> > >                         __IS_MC, available, dequeue_X); }

> > >

> > > Konstantin

> > Thanks for the feedback/idea. The goal of this patch was to make it

> > simple enough to define APIs to store any element size without code

> duplication.

> 

> Well, then if we store elt_size inside the ring, it should be easy enough to add

> to the API generic functions that would use memcpy(or rte_memcpy) for

> enqueue/dequeue.

> Yes, it might be slower than existing (8B per elem), but might be still

> acceptable.

The element size will be a constant in most use cases. If we keep the element size as a parameter, it allows the compiler to do any loop unrolling and auto-vectorization optimizations on copying.
Storing the element size will result in additional memory access.

> 

> >With this patch, the user has to write ~4 lines of code to get APIs for

> >any element size. I would like to keep the goal still the  same.

> >

> > If we have to avoid the macro-fest, the main problem that needs to be

> > addressed is - how to represent different sizes of element types in a generic

> way? IMO, we can do this by defining the element type to be a multiple of

> uint32_t (I do not think we need to go to uint16_t).

> >

> > For ex:

> > rte_ring_mp_enqueue_bulk_objs(struct rte_ring *r,

> >                 uint32_t *obj_table, unsigned int num_objs,

> >                 unsigned int n,

> >                 enum rte_ring_queue_behavior behavior, unsigned int is_sp,

> >                 unsigned int *free_space) { }

> >

> > This approach would ensure that we have generic enough APIs and they

> > can be used for elements of any size. But the element itself needs to be a

> multiple of 32b - I think this should not be a concern.

> >

> > The API suffix definitely needs to be better, any suggestions?

> 

> >

> > >

> > >

> > > > +

> > > > +/**

> > > > + * @internal Enqueue several objects on the ring  */ static

> > > > +__rte_always_inline unsigned int

> > > > +__RTE_RING_CONCAT(__rte_ring_do_enqueue)(struct rte_ring *r,

> > > > +		RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int

> > > n,

> > > > +		enum rte_ring_queue_behavior behavior, unsigned int is_sp,

> > > > +		unsigned int *free_space)

> > > > +{

> > > > +	uint32_t prod_head, prod_next;

> > > > +	uint32_t free_entries;

> > > > +

> > > > +	n = __rte_ring_move_prod_head(r, is_sp, n, behavior,

> > > > +			&prod_head, &prod_next, &free_entries);

> > > > +	if (n == 0)

> > > > +		goto end;

> > > > +

> > > > +	ENQUEUE_PTRS(r, &r[1], prod_head, obj_table, n,

> > > > +		RTE_RING_TMPLT_ELEM_TYPE);

> > > > +

> > > > +	update_tail(&r->prod, prod_head, prod_next, is_sp, 1);

> > > > +end:

> > > > +	if (free_space != NULL)

> > > > +		*free_space = free_entries - n;

> > > > +	return n;

> > > > +}

> > > > +

> > > > +/**

> > > > + * @internal Dequeue several objects from the ring  */ static

> > > > +__rte_always_inline unsigned int

> > > > +__RTE_RING_CONCAT(__rte_ring_do_dequeue)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > > +	enum rte_ring_queue_behavior behavior, unsigned int is_sc,

> > > > +	unsigned int *available)

> > > > +{

> > > > +	uint32_t cons_head, cons_next;

> > > > +	uint32_t entries;

> > > > +

> > > > +	n = __rte_ring_move_cons_head(r, (int)is_sc, n, behavior,

> > > > +			&cons_head, &cons_next, &entries);

> > > > +	if (n == 0)

> > > > +		goto end;

> > > > +

> > > > +	DEQUEUE_PTRS(r, &r[1], cons_head, obj_table, n,

> > > > +		RTE_RING_TMPLT_ELEM_TYPE);

> > > > +

> > > > +	update_tail(&r->cons, cons_head, cons_next, is_sc, 0);

> > > > +

> > > > +end:

> > > > +	if (available != NULL)

> > > > +		*available = entries - n;

> > > > +	return n;

> > > > +}

> > > > +

> > > > +

> > > > +/**

> > > > + * Enqueue several objects on the ring (multi-producers safe).

> > > > + */

> > > > +static __rte_always_inline unsigned int

> > > > +__RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> > > > +	unsigned int *free_space)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_FIXED, __IS_MP, free_space); }

> > > > +

> > > > +/**

> > > > + * Enqueue several objects on a ring (NOT multi-producers safe).

> > > > + */

> > > > +static __rte_always_inline unsigned int

> > > > +__RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> > > > +	unsigned int *free_space)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_FIXED, __IS_SP, free_space); }

> > > > +

> > > > +/**

> > > > + * Enqueue several objects on a ring.

> > > > + */

> > > > +static __rte_always_inline unsigned int

> > > > +__RTE_RING_CONCAT(rte_ring_enqueue_bulk)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,

> > > > +	unsigned int *free_space)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_FIXED, r->prod.single,

> free_space); }

> > > > +

> > > > +/**

> > > > + * Enqueue one object on a ring (multi-producers safe).

> > > > + */

> > > > +static __rte_always_inline int

> > > > +__RTE_RING_CONCAT(rte_ring_mp_enqueue)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE obj)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(r, &obj, 1,

> > > NULL) ?

> > > > +			0 : -ENOBUFS;

> > > > +}

> > > > +

> > > > +/**

> > > > + * Enqueue one object on a ring (NOT multi-producers safe).

> > > > + */

> > > > +static __rte_always_inline int

> > > > +__RTE_RING_CONCAT(rte_ring_sp_enqueue)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE obj)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(r, &obj, 1,

> > > NULL) ?

> > > > +			0 : -ENOBUFS;

> > > > +}

> > > > +

> > > > +/**

> > > > + * Enqueue one object on a ring.

> > > > + */

> > > > +static __rte_always_inline int

> > > > +__RTE_RING_CONCAT(rte_ring_enqueue)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(rte_ring_enqueue_bulk)(r, obj, 1,

> > > NULL) ?

> > > > +			0 : -ENOBUFS;

> > > > +}

> > > > +

> > > > +/**

> > > > + * Dequeue several objects from a ring (multi-consumers safe).

> > > > + */

> > > > +static __rte_always_inline unsigned int

> > > > +__RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > > +	unsigned int *available)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_FIXED, __IS_MC, available); }

> > > > +

> > > > +/**

> > > > + * Dequeue several objects from a ring (NOT multi-consumers safe).

> > > > + */

> > > > +static __rte_always_inline unsigned int

> > > > +__RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > > +	unsigned int *available)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_FIXED, __IS_SC, available); }

> > > > +

> > > > +/**

> > > > + * Dequeue several objects from a ring.

> > > > + */

> > > > +static __rte_always_inline unsigned int

> > > > +__RTE_RING_CONCAT(rte_ring_dequeue_bulk)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > > +	unsigned int *available)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_FIXED, r->cons.single, available); }

> > > > +

> > > > +/**

> > > > + * Dequeue one object from a ring (multi-consumers safe).

> > > > + */

> > > > +static __rte_always_inline int

> > > > +__RTE_RING_CONCAT(rte_ring_mc_dequeue)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_p) {

> > > > +	return __RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(r, obj_p, 1,

> > > NULL) ?

> > > > +			0 : -ENOENT;

> > > > +}

> > > > +

> > > > +/**

> > > > + * Dequeue one object from a ring (NOT multi-consumers safe).

> > > > + */

> > > > +static __rte_always_inline int

> > > > +__RTE_RING_CONCAT(rte_ring_sc_dequeue)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_p) {

> > > > +	return __RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(r, obj_p, 1,

> > > NULL) ?

> > > > +			0 : -ENOENT;

> > > > +}

> > > > +

> > > > +/**

> > > > + * Dequeue one object from a ring.

> > > > + */

> > > > +static __rte_always_inline int

> > > > +__RTE_RING_CONCAT(rte_ring_dequeue)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_p) {

> > > > +	return __RTE_RING_CONCAT(rte_ring_dequeue_bulk)(r, obj_p, 1,

> > > NULL) ?

> > > > +			0 : -ENOENT;

> > > > +}

> > > > +

> > > > +/**

> > > > + * Enqueue several objects on the ring (multi-producers safe).

> > > > + */

> > > > +static __rte_always_inline unsigned

> > > > +__RTE_RING_CONCAT(rte_ring_mp_enqueue_burst)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table,

> > > > +			 unsigned int n, unsigned int *free_space) {

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_VARIABLE, __IS_MP, free_space); }

> > > > +

> > > > +/**

> > > > + * Enqueue several objects on a ring (NOT multi-producers safe).

> > > > + */

> > > > +static __rte_always_inline unsigned

> > > > +__RTE_RING_CONCAT(rte_ring_sp_enqueue_burst)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table,

> > > > +			 unsigned int n, unsigned int *free_space) {

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_VARIABLE, __IS_SP, free_space); }

> > > > +

> > > > +/**

> > > > + * Enqueue several objects on a ring.

> > > > + */

> > > > +static __rte_always_inline unsigned

> > > > +__RTE_RING_CONCAT(rte_ring_enqueue_burst)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > > +	unsigned int *free_space)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_VARIABLE, r->prod.single,

> > > free_space);

> > > > +}

> > > > +

> > > > +/**

> > > > + * Dequeue several objects from a ring (multi-consumers safe).

> > > > +When the

> > > request

> > > > + * objects are more than the available objects, only dequeue the

> > > > + actual

> > > number

> > > > + * of objects

> > > > + */

> > > > +static __rte_always_inline unsigned

> > > > +__RTE_RING_CONCAT(rte_ring_mc_dequeue_burst)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > > +	unsigned int *available)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_VARIABLE, __IS_MC, available); }

> > > > +

> > > > +/**

> > > > + * Dequeue several objects from a ring (NOT multi-consumers

> > > > +safe).When

> > > the

> > > > + * request objects are more than the available objects, only

> > > > +dequeue the

> > > > + * actual number of objects

> > > > + */

> > > > +static __rte_always_inline unsigned

> > > > +__RTE_RING_CONCAT(rte_ring_sc_dequeue_burst)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > > +	unsigned int *available)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > > +			RTE_RING_QUEUE_VARIABLE, __IS_SC, available); }

> > > > +

> > > > +/**

> > > > + * Dequeue multiple objects from a ring up to a maximum number.

> > > > + */

> > > > +static __rte_always_inline unsigned

> > > > +__RTE_RING_CONCAT(rte_ring_dequeue_burst)(struct rte_ring *r,

> > > > +	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,

> > > > +	unsigned int *available)

> > > > +{

> > > > +	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,

> > > > +				RTE_RING_QUEUE_VARIABLE,

> > > > +				r->cons.single, available);

> > > > +}

> > > > +

> > > > +#ifdef __cplusplus

> > > > +}

> > > > +#endif

> > > > +

> > > > +#endif /* _RTE_RING_TEMPLATE_H_ */

> > > > --

> > > > 2.17.1
Ananyev, Konstantin Oct. 3, 2019, 11:51 a.m. UTC | #5
> > > > > +++ b/lib/librte_ring/rte_ring_template.h

> > > > > @@ -0,0 +1,330 @@

> > > > > +/* SPDX-License-Identifier: BSD-3-Clause

> > > > > + * Copyright (c) 2019 Arm Limited  */

> > > > > +

> > > > > +#ifndef _RTE_RING_TEMPLATE_H_

> > > > > +#define _RTE_RING_TEMPLATE_H_

> > > > > +

> > > > > +#ifdef __cplusplus

> > > > > +extern "C" {

> > > > > +#endif

> > > > > +

> > > > > +#include <stdio.h>

> > > > > +#include <stdint.h>

> > > > > +#include <sys/queue.h>

> > > > > +#include <errno.h>

> > > > > +#include <rte_common.h>

> > > > > +#include <rte_config.h>

> > > > > +#include <rte_memory.h>

> > > > > +#include <rte_lcore.h>

> > > > > +#include <rte_atomic.h>

> > > > > +#include <rte_branch_prediction.h> #include <rte_memzone.h>

> > > > > +#include <rte_pause.h> #include <rte_ring.h>

> > > > > +

> > > > > +/* Ring API suffix name - used to append to API names */ #ifndef

> > > > > +RTE_RING_TMPLT_API_SUFFIX #error RTE_RING_TMPLT_API_SUFFIX

> > not

> > > > > +defined #endif

> > > > > +

> > > > > +/* Ring's element size in bits, should be a power of 2 */ #ifndef

> > > > > +RTE_RING_TMPLT_ELEM_SIZE #error RTE_RING_TMPLT_ELEM_SIZE

> > not

> > > > defined

> > > > > +#endif

> > > > > +

> > > > > +/* Type of ring elements */

> > > > > +#ifndef RTE_RING_TMPLT_ELEM_TYPE

> > > > > +#error RTE_RING_TMPLT_ELEM_TYPE not defined #endif

> > > > > +

> > > > > +#define _rte_fuse(a, b) a##_##b

> > > > > +#define __rte_fuse(a, b) _rte_fuse(a, b) #define

> > > > > +__RTE_RING_CONCAT(a) __rte_fuse(a, RTE_RING_TMPLT_API_SUFFIX)

> > > > > +

> > > > > +/* Calculate the memory size needed for a ring */

> > > > > +RTE_RING_TMPLT_EXPERIMENTAL ssize_t

> > > > > +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count);

> > > > > +

> > > > > +/* Create a new ring named *name* in memory. */

> > > > > +RTE_RING_TMPLT_EXPERIMENTAL struct rte_ring *

> > > > > +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned

> > count,

> > > > > +					int socket_id, unsigned flags);

> > > >

> > > >

> > > > Just an idea - probably same thing can be achieved in a different way.

> > > > Instead of all these defines - replace ENQUEUE_PTRS/DEQUEUE_PTRS

> > > > macros with static inline functions and then make all internal functions,

> > i.e.

> > > > __rte_ring_do_dequeue()

> > > > to accept enqueue/dequeue function pointer as a parameter.

> > > > Then let say default rte_ring_mc_dequeue_bulk will do:

> > > >

> > > > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> > > >                 unsigned int n, unsigned int *available) {

> > > >         return __rte_ring_do_dequeue(r, obj_table, n,

> > RTE_RING_QUEUE_FIXED,

> > > >                         __IS_MC, available, dequeue_ptr_default); }

> > > >

> > > > Then if someone will like to define ring functions forelt_size==X,

> > > > all he would need to do:

> > > > 1. define his own enqueue/dequeuer functions.

> > > > 2. do something like:

> > > > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> > > >                 unsigned int n, unsigned int *available) {

> > > >         return __rte_ring_do_dequeue(r, obj_table, n,

> > RTE_RING_QUEUE_FIXED,

> > > >                         __IS_MC, available, dequeue_X); }

> > > >

> > > > Konstantin

> > > Thanks for the feedback/idea. The goal of this patch was to make it

> > > simple enough to define APIs to store any element size without code

> > duplication.

> >

> > Well, then if we store elt_size inside the ring, it should be easy enough to add

> > to the API generic functions that would use memcpy(or rte_memcpy) for

> > enqueue/dequeue.

> > Yes, it might be slower than existing (8B per elem), but might be still

> > acceptable.

> The element size will be a constant in most use cases. If we keep the element size as a parameter, it allows the compiler to do any loop

> unrolling and auto-vectorization optimizations on copying.

> Storing the element size will result in additional memory access.


I understand that, but for you case (rcu defer queue) you probably need highest possible performance, right?
I am sure there will be other cases where such slight perf degradation is acceptatble.

> 

> >

> > >With this patch, the user has to write ~4 lines of code to get APIs for

> > >any element size. I would like to keep the goal still the  same.

> > >

> > > If we have to avoid the macro-fest, the main problem that needs to be

> > > addressed is - how to represent different sizes of element types in a generic

> > way? IMO, we can do this by defining the element type to be a multiple of

> > uint32_t (I do not think we need to go to uint16_t).

> > >

> > > For ex:

> > > rte_ring_mp_enqueue_bulk_objs(struct rte_ring *r,

> > >                 uint32_t *obj_table, unsigned int num_objs,

> > >                 unsigned int n,

> > >                 enum rte_ring_queue_behavior behavior, unsigned int is_sp,

> > >                 unsigned int *free_space) { }

> > >

> > > This approach would ensure that we have generic enough APIs and they

> > > can be used for elements of any size. But the element itself needs to be a

> > multiple of 32b - I think this should not be a concern.

> > >

> > > The API suffix definitely needs to be better, any suggestions?

> >
Ananyev, Konstantin Oct. 3, 2019, 12:27 p.m. UTC | #6
> 

> > > > > > +++ b/lib/librte_ring/rte_ring_template.h

> > > > > > @@ -0,0 +1,330 @@

> > > > > > +/* SPDX-License-Identifier: BSD-3-Clause

> > > > > > + * Copyright (c) 2019 Arm Limited  */

> > > > > > +

> > > > > > +#ifndef _RTE_RING_TEMPLATE_H_

> > > > > > +#define _RTE_RING_TEMPLATE_H_

> > > > > > +

> > > > > > +#ifdef __cplusplus

> > > > > > +extern "C" {

> > > > > > +#endif

> > > > > > +

> > > > > > +#include <stdio.h>

> > > > > > +#include <stdint.h>

> > > > > > +#include <sys/queue.h>

> > > > > > +#include <errno.h>

> > > > > > +#include <rte_common.h>

> > > > > > +#include <rte_config.h>

> > > > > > +#include <rte_memory.h>

> > > > > > +#include <rte_lcore.h>

> > > > > > +#include <rte_atomic.h>

> > > > > > +#include <rte_branch_prediction.h> #include <rte_memzone.h>

> > > > > > +#include <rte_pause.h> #include <rte_ring.h>

> > > > > > +

> > > > > > +/* Ring API suffix name - used to append to API names */ #ifndef

> > > > > > +RTE_RING_TMPLT_API_SUFFIX #error RTE_RING_TMPLT_API_SUFFIX

> > > not

> > > > > > +defined #endif

> > > > > > +

> > > > > > +/* Ring's element size in bits, should be a power of 2 */ #ifndef

> > > > > > +RTE_RING_TMPLT_ELEM_SIZE #error RTE_RING_TMPLT_ELEM_SIZE

> > > not

> > > > > defined

> > > > > > +#endif

> > > > > > +

> > > > > > +/* Type of ring elements */

> > > > > > +#ifndef RTE_RING_TMPLT_ELEM_TYPE

> > > > > > +#error RTE_RING_TMPLT_ELEM_TYPE not defined #endif

> > > > > > +

> > > > > > +#define _rte_fuse(a, b) a##_##b

> > > > > > +#define __rte_fuse(a, b) _rte_fuse(a, b) #define

> > > > > > +__RTE_RING_CONCAT(a) __rte_fuse(a, RTE_RING_TMPLT_API_SUFFIX)

> > > > > > +

> > > > > > +/* Calculate the memory size needed for a ring */

> > > > > > +RTE_RING_TMPLT_EXPERIMENTAL ssize_t

> > > > > > +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count);

> > > > > > +

> > > > > > +/* Create a new ring named *name* in memory. */

> > > > > > +RTE_RING_TMPLT_EXPERIMENTAL struct rte_ring *

> > > > > > +__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned

> > > count,

> > > > > > +					int socket_id, unsigned flags);

> > > > >

> > > > >

> > > > > Just an idea - probably same thing can be achieved in a different way.

> > > > > Instead of all these defines - replace ENQUEUE_PTRS/DEQUEUE_PTRS

> > > > > macros with static inline functions and then make all internal functions,

> > > i.e.

> > > > > __rte_ring_do_dequeue()

> > > > > to accept enqueue/dequeue function pointer as a parameter.

> > > > > Then let say default rte_ring_mc_dequeue_bulk will do:

> > > > >

> > > > > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> > > > >                 unsigned int n, unsigned int *available) {

> > > > >         return __rte_ring_do_dequeue(r, obj_table, n,

> > > RTE_RING_QUEUE_FIXED,

> > > > >                         __IS_MC, available, dequeue_ptr_default); }

> > > > >

> > > > > Then if someone will like to define ring functions forelt_size==X,

> > > > > all he would need to do:

> > > > > 1. define his own enqueue/dequeuer functions.

> > > > > 2. do something like:

> > > > > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> > > > >                 unsigned int n, unsigned int *available) {

> > > > >         return __rte_ring_do_dequeue(r, obj_table, n,

> > > RTE_RING_QUEUE_FIXED,

> > > > >                         __IS_MC, available, dequeue_X); }

> > > > >

> > > > > Konstantin

> > > > Thanks for the feedback/idea. The goal of this patch was to make it

> > > > simple enough to define APIs to store any element size without code

> > > duplication.

> > >

> > > Well, then if we store elt_size inside the ring, it should be easy enough to add

> > > to the API generic functions that would use memcpy(or rte_memcpy) for

> > > enqueue/dequeue.

> > > Yes, it might be slower than existing (8B per elem), but might be still

> > > acceptable.

> > The element size will be a constant in most use cases. If we keep the element size as a parameter, it allows the compiler to do any loop

> > unrolling and auto-vectorization optimizations on copying.

> > Storing the element size will result in additional memory access.

> 

> I understand that, but for you case (rcu defer queue) you probably need highest possible performance, right?


Meant 'don't need' of course :)

> I am sure there will be other cases where such slight perf degradation is acceptatble.

> 

> >

> > >

> > > >With this patch, the user has to write ~4 lines of code to get APIs for

> > > >any element size. I would like to keep the goal still the  same.

> > > >

> > > > If we have to avoid the macro-fest, the main problem that needs to be

> > > > addressed is - how to represent different sizes of element types in a generic

> > > way? IMO, we can do this by defining the element type to be a multiple of

> > > uint32_t (I do not think we need to go to uint16_t).

> > > >

> > > > For ex:

> > > > rte_ring_mp_enqueue_bulk_objs(struct rte_ring *r,

> > > >                 uint32_t *obj_table, unsigned int num_objs,

> > > >                 unsigned int n,

> > > >                 enum rte_ring_queue_behavior behavior, unsigned int is_sp,

> > > >                 unsigned int *free_space) { }

> > > >

> > > > This approach would ensure that we have generic enough APIs and they

> > > > can be used for elements of any size. But the element itself needs to be a

> > > multiple of 32b - I think this should not be a concern.

> > > >

> > > > The API suffix definitely needs to be better, any suggestions?

> > >
Honnappa Nagarahalli Oct. 3, 2019, 10:49 p.m. UTC | #7
<snip>

> > > > > > > +++ b/lib/librte_ring/rte_ring_template.h

> > > > > > > @@ -0,0 +1,330 @@

> > > > > > > +/* SPDX-License-Identifier: BSD-3-Clause

> > > > > > > + * Copyright (c) 2019 Arm Limited  */

> > > > > > > +

> > > > > > > +#ifndef _RTE_RING_TEMPLATE_H_ #define

> _RTE_RING_TEMPLATE_H_

> > > > > > > +

> > > > > > > +#ifdef __cplusplus

> > > > > > > +extern "C" {

> > > > > > > +#endif

> > > > > > > +

> > > > > > > +#include <stdio.h>

> > > > > > > +#include <stdint.h>

> > > > > > > +#include <sys/queue.h>

> > > > > > > +#include <errno.h>

> > > > > > > +#include <rte_common.h>

> > > > > > > +#include <rte_config.h>

> > > > > > > +#include <rte_memory.h>

> > > > > > > +#include <rte_lcore.h>

> > > > > > > +#include <rte_atomic.h>

> > > > > > > +#include <rte_branch_prediction.h> #include <rte_memzone.h>

> > > > > > > +#include <rte_pause.h> #include <rte_ring.h>

> > > > > > > +

> > > > > > > +/* Ring API suffix name - used to append to API names */

> > > > > > > +#ifndef RTE_RING_TMPLT_API_SUFFIX #error

> > > > > > > +RTE_RING_TMPLT_API_SUFFIX

> > > > not

> > > > > > > +defined #endif

> > > > > > > +

> > > > > > > +/* Ring's element size in bits, should be a power of 2 */

> > > > > > > +#ifndef RTE_RING_TMPLT_ELEM_SIZE #error

> > > > > > > +RTE_RING_TMPLT_ELEM_SIZE

> > > > not

> > > > > > defined

> > > > > > > +#endif

> > > > > > > +

> > > > > > > +/* Type of ring elements */ #ifndef

> > > > > > > +RTE_RING_TMPLT_ELEM_TYPE #error

> RTE_RING_TMPLT_ELEM_TYPE

> > > > > > > +not defined #endif

> > > > > > > +

> > > > > > > +#define _rte_fuse(a, b) a##_##b #define __rte_fuse(a, b)

> > > > > > > +_rte_fuse(a, b) #define

> > > > > > > +__RTE_RING_CONCAT(a) __rte_fuse(a,

> > > > > > > +RTE_RING_TMPLT_API_SUFFIX)

> > > > > > > +

> > > > > > > +/* Calculate the memory size needed for a ring */

> > > > > > > +RTE_RING_TMPLT_EXPERIMENTAL ssize_t

> > > > > > > +__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count);

> > > > > > > +

> > > > > > > +/* Create a new ring named *name* in memory. */

> > > > > > > +RTE_RING_TMPLT_EXPERIMENTAL struct rte_ring *

> > > > > > > +__RTE_RING_CONCAT(rte_ring_create)(const char *name,

> > > > > > > +unsigned

> > > > count,

> > > > > > > +					int socket_id, unsigned flags);

> > > > > >

> > > > > >

> > > > > > Just an idea - probably same thing can be achieved in a different

> way.

> > > > > > Instead of all these defines - replace

> > > > > > ENQUEUE_PTRS/DEQUEUE_PTRS macros with static inline functions

> > > > > > and then make all internal functions,

> > > > i.e.

> > > > > > __rte_ring_do_dequeue()

> > > > > > to accept enqueue/dequeue function pointer as a parameter.

> > > > > > Then let say default rte_ring_mc_dequeue_bulk will do:

> > > > > >

> > > > > > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> > > > > >                 unsigned int n, unsigned int *available) {

> > > > > >         return __rte_ring_do_dequeue(r, obj_table, n,

> > > > RTE_RING_QUEUE_FIXED,

> > > > > >                         __IS_MC, available,

> > > > > > dequeue_ptr_default); }

> > > > > >

> > > > > > Then if someone will like to define ring functions

> > > > > > forelt_size==X, all he would need to do:

> > > > > > 1. define his own enqueue/dequeuer functions.

> > > > > > 2. do something like:

> > > > > > rte_ring_mc_dequeue_bulk(struct rte_ring *r, void **obj_table,

> > > > > >                 unsigned int n, unsigned int *available) {

> > > > > >         return __rte_ring_do_dequeue(r, obj_table, n,

> > > > RTE_RING_QUEUE_FIXED,

> > > > > >                         __IS_MC, available, dequeue_X); }

> > > > > >

> > > > > > Konstantin

> > > > > Thanks for the feedback/idea. The goal of this patch was to make

> > > > > it simple enough to define APIs to store any element size

> > > > > without code

> > > > duplication.

> > > >

> > > > Well, then if we store elt_size inside the ring, it should be easy

> > > > enough to add to the API generic functions that would use

> > > > memcpy(or rte_memcpy) for enqueue/dequeue.

> > > > Yes, it might be slower than existing (8B per elem), but might be

> > > > still acceptable.

> > > The element size will be a constant in most use cases. If we keep

> > > the element size as a parameter, it allows the compiler to do any loop

> unrolling and auto-vectorization optimizations on copying.

> > > Storing the element size will result in additional memory access.

> >

> > I understand that, but for you case (rcu defer queue) you probably need

> highest possible performance, right?

> 

> Meant 'don't need' of course :)

😊 understood. that is just one use case. It actually started as an option to reduce memory usage in different places. You can look at the rte_hash changes in this patch. I also have plans for further changes.

> 

> > I am sure there will be other cases where such slight perf degradation is

> acceptatble.

> >

> > >

> > > >

> > > > >With this patch, the user has to write ~4 lines of code to get

> > > > >APIs for any element size. I would like to keep the goal still the  same.

> > > > >

> > > > > If we have to avoid the macro-fest, the main problem that needs

> > > > > to be addressed is - how to represent different sizes of element

> > > > > types in a generic

> > > > way? IMO, we can do this by defining the element type to be a

> > > > multiple of uint32_t (I do not think we need to go to uint16_t).

> > > > >

> > > > > For ex:

> > > > > rte_ring_mp_enqueue_bulk_objs(struct rte_ring *r,

> > > > >                 uint32_t *obj_table, unsigned int num_objs,

> > > > >                 unsigned int n,

> > > > >                 enum rte_ring_queue_behavior behavior, unsigned int is_sp,

> > > > >                 unsigned int *free_space) { }

> > > > >

> > > > > This approach would ensure that we have generic enough APIs and

> > > > > they can be used for elements of any size. But the element

> > > > > itself needs to be a

> > > > multiple of 32b - I think this should not be a concern.

> > > > >

> > > > > The API suffix definitely needs to be better, any suggestions?

> > > >
diff mbox series

Patch

diff --git a/lib/librte_ring/Makefile b/lib/librte_ring/Makefile
index 4c8410229..818898110 100644
--- a/lib/librte_ring/Makefile
+++ b/lib/librte_ring/Makefile
@@ -19,6 +19,8 @@  SRCS-$(CONFIG_RTE_LIBRTE_RING) := rte_ring.c
 # install includes
 SYMLINK-$(CONFIG_RTE_LIBRTE_RING)-include := rte_ring.h \
 					rte_ring_generic.h \
-					rte_ring_c11_mem.h
+					rte_ring_c11_mem.h \
+					rte_ring_template.h \
+					rte_ring_template.c
 
 include $(RTE_SDK)/mk/rte.lib.mk
diff --git a/lib/librte_ring/meson.build b/lib/librte_ring/meson.build
index 74219840a..e4e208a7c 100644
--- a/lib/librte_ring/meson.build
+++ b/lib/librte_ring/meson.build
@@ -5,7 +5,9 @@  version = 2
 sources = files('rte_ring.c')
 headers = files('rte_ring.h',
 		'rte_ring_c11_mem.h',
-		'rte_ring_generic.h')
+		'rte_ring_generic.h',
+		'rte_ring_template.h',
+		'rte_ring_template.c')
 
 # rte_ring_create_elem and rte_ring_get_memsize_elem are experimental
 allow_experimental_apis = true
diff --git a/lib/librte_ring/rte_ring_template.c b/lib/librte_ring/rte_ring_template.c
new file mode 100644
index 000000000..1ca593f95
--- /dev/null
+++ b/lib/librte_ring/rte_ring_template.c
@@ -0,0 +1,46 @@ 
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (c) 2019 Arm Limited
+ */
+
+#include <stdio.h>
+#include <stdarg.h>
+#include <string.h>
+#include <stdint.h>
+#include <inttypes.h>
+#include <errno.h>
+#include <sys/queue.h>
+
+#include <rte_common.h>
+#include <rte_log.h>
+#include <rte_memory.h>
+#include <rte_memzone.h>
+#include <rte_malloc.h>
+#include <rte_launch.h>
+#include <rte_eal.h>
+#include <rte_eal_memconfig.h>
+#include <rte_atomic.h>
+#include <rte_per_lcore.h>
+#include <rte_lcore.h>
+#include <rte_branch_prediction.h>
+#include <rte_errno.h>
+#include <rte_string_fns.h>
+#include <rte_spinlock.h>
+#include <rte_tailq.h>
+
+#include "rte_ring.h"
+
+/* return the size of memory occupied by a ring */
+ssize_t
+__RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count)
+{
+	return rte_ring_get_memsize_elem(count, RTE_RING_TMPLT_ELEM_SIZE);
+}
+
+/* create the ring */
+struct rte_ring *
+__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned count,
+		int socket_id, unsigned flags)
+{
+	return rte_ring_create_elem(name, count, RTE_RING_TMPLT_ELEM_SIZE,
+		socket_id, flags);
+}
diff --git a/lib/librte_ring/rte_ring_template.h b/lib/librte_ring/rte_ring_template.h
new file mode 100644
index 000000000..b9b14dfbb
--- /dev/null
+++ b/lib/librte_ring/rte_ring_template.h
@@ -0,0 +1,330 @@ 
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (c) 2019 Arm Limited
+ */
+
+#ifndef _RTE_RING_TEMPLATE_H_
+#define _RTE_RING_TEMPLATE_H_
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+
+#include <stdio.h>
+#include <stdint.h>
+#include <sys/queue.h>
+#include <errno.h>
+#include <rte_common.h>
+#include <rte_config.h>
+#include <rte_memory.h>
+#include <rte_lcore.h>
+#include <rte_atomic.h>
+#include <rte_branch_prediction.h>
+#include <rte_memzone.h>
+#include <rte_pause.h>
+#include <rte_ring.h>
+
+/* Ring API suffix name - used to append to API names */
+#ifndef RTE_RING_TMPLT_API_SUFFIX
+#error RTE_RING_TMPLT_API_SUFFIX not defined
+#endif
+
+/* Ring's element size in bits, should be a power of 2 */
+#ifndef RTE_RING_TMPLT_ELEM_SIZE
+#error RTE_RING_TMPLT_ELEM_SIZE not defined
+#endif
+
+/* Type of ring elements */
+#ifndef RTE_RING_TMPLT_ELEM_TYPE
+#error RTE_RING_TMPLT_ELEM_TYPE not defined
+#endif
+
+#define _rte_fuse(a, b) a##_##b
+#define __rte_fuse(a, b) _rte_fuse(a, b)
+#define __RTE_RING_CONCAT(a) __rte_fuse(a, RTE_RING_TMPLT_API_SUFFIX)
+
+/* Calculate the memory size needed for a ring */
+RTE_RING_TMPLT_EXPERIMENTAL
+ssize_t __RTE_RING_CONCAT(rte_ring_get_memsize)(unsigned count);
+
+/* Create a new ring named *name* in memory. */
+RTE_RING_TMPLT_EXPERIMENTAL
+struct rte_ring *
+__RTE_RING_CONCAT(rte_ring_create)(const char *name, unsigned count,
+					int socket_id, unsigned flags);
+
+/**
+ * @internal Enqueue several objects on the ring
+ */
+static __rte_always_inline unsigned int
+__RTE_RING_CONCAT(__rte_ring_do_enqueue)(struct rte_ring *r,
+		RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,
+		enum rte_ring_queue_behavior behavior, unsigned int is_sp,
+		unsigned int *free_space)
+{
+	uint32_t prod_head, prod_next;
+	uint32_t free_entries;
+
+	n = __rte_ring_move_prod_head(r, is_sp, n, behavior,
+			&prod_head, &prod_next, &free_entries);
+	if (n == 0)
+		goto end;
+
+	ENQUEUE_PTRS(r, &r[1], prod_head, obj_table, n,
+		RTE_RING_TMPLT_ELEM_TYPE);
+
+	update_tail(&r->prod, prod_head, prod_next, is_sp, 1);
+end:
+	if (free_space != NULL)
+		*free_space = free_entries - n;
+	return n;
+}
+
+/**
+ * @internal Dequeue several objects from the ring
+ */
+static __rte_always_inline unsigned int
+__RTE_RING_CONCAT(__rte_ring_do_dequeue)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,
+	enum rte_ring_queue_behavior behavior, unsigned int is_sc,
+	unsigned int *available)
+{
+	uint32_t cons_head, cons_next;
+	uint32_t entries;
+
+	n = __rte_ring_move_cons_head(r, (int)is_sc, n, behavior,
+			&cons_head, &cons_next, &entries);
+	if (n == 0)
+		goto end;
+
+	DEQUEUE_PTRS(r, &r[1], cons_head, obj_table, n,
+		RTE_RING_TMPLT_ELEM_TYPE);
+
+	update_tail(&r->cons, cons_head, cons_next, is_sc, 0);
+
+end:
+	if (available != NULL)
+		*available = entries - n;
+	return n;
+}
+
+
+/**
+ * Enqueue several objects on the ring (multi-producers safe).
+ */
+static __rte_always_inline unsigned int
+__RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,
+	unsigned int *free_space)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,
+			RTE_RING_QUEUE_FIXED, __IS_MP, free_space);
+}
+
+/**
+ * Enqueue several objects on a ring (NOT multi-producers safe).
+ */
+static __rte_always_inline unsigned int
+__RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,
+	unsigned int *free_space)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,
+			RTE_RING_QUEUE_FIXED, __IS_SP, free_space);
+}
+
+/**
+ * Enqueue several objects on a ring.
+ */
+static __rte_always_inline unsigned int
+__RTE_RING_CONCAT(rte_ring_enqueue_bulk)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE const *obj_table, unsigned int n,
+	unsigned int *free_space)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,
+			RTE_RING_QUEUE_FIXED, r->prod.single, free_space);
+}
+
+/**
+ * Enqueue one object on a ring (multi-producers safe).
+ */
+static __rte_always_inline int
+__RTE_RING_CONCAT(rte_ring_mp_enqueue)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE obj)
+{
+	return __RTE_RING_CONCAT(rte_ring_mp_enqueue_bulk)(r, &obj, 1, NULL) ?
+			0 : -ENOBUFS;
+}
+
+/**
+ * Enqueue one object on a ring (NOT multi-producers safe).
+ */
+static __rte_always_inline int
+__RTE_RING_CONCAT(rte_ring_sp_enqueue)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE obj)
+{
+	return __RTE_RING_CONCAT(rte_ring_sp_enqueue_bulk)(r, &obj, 1, NULL) ?
+			0 : -ENOBUFS;
+}
+
+/**
+ * Enqueue one object on a ring.
+ */
+static __rte_always_inline int
+__RTE_RING_CONCAT(rte_ring_enqueue)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj)
+{
+	return __RTE_RING_CONCAT(rte_ring_enqueue_bulk)(r, obj, 1, NULL) ?
+			0 : -ENOBUFS;
+}
+
+/**
+ * Dequeue several objects from a ring (multi-consumers safe).
+ */
+static __rte_always_inline unsigned int
+__RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,
+	unsigned int *available)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,
+			RTE_RING_QUEUE_FIXED, __IS_MC, available);
+}
+
+/**
+ * Dequeue several objects from a ring (NOT multi-consumers safe).
+ */
+static __rte_always_inline unsigned int
+__RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,
+	unsigned int *available)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,
+			RTE_RING_QUEUE_FIXED, __IS_SC, available);
+}
+
+/**
+ * Dequeue several objects from a ring.
+ */
+static __rte_always_inline unsigned int
+__RTE_RING_CONCAT(rte_ring_dequeue_bulk)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,
+	unsigned int *available)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,
+			RTE_RING_QUEUE_FIXED, r->cons.single, available);
+}
+
+/**
+ * Dequeue one object from a ring (multi-consumers safe).
+ */
+static __rte_always_inline int
+__RTE_RING_CONCAT(rte_ring_mc_dequeue)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_p)
+{
+	return __RTE_RING_CONCAT(rte_ring_mc_dequeue_bulk)(r, obj_p, 1, NULL) ?
+			0 : -ENOENT;
+}
+
+/**
+ * Dequeue one object from a ring (NOT multi-consumers safe).
+ */
+static __rte_always_inline int
+__RTE_RING_CONCAT(rte_ring_sc_dequeue)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_p)
+{
+	return __RTE_RING_CONCAT(rte_ring_sc_dequeue_bulk)(r, obj_p, 1, NULL) ?
+			0 : -ENOENT;
+}
+
+/**
+ * Dequeue one object from a ring.
+ */
+static __rte_always_inline int
+__RTE_RING_CONCAT(rte_ring_dequeue)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_p)
+{
+	return __RTE_RING_CONCAT(rte_ring_dequeue_bulk)(r, obj_p, 1, NULL) ?
+			0 : -ENOENT;
+}
+
+/**
+ * Enqueue several objects on the ring (multi-producers safe).
+ */
+static __rte_always_inline unsigned
+__RTE_RING_CONCAT(rte_ring_mp_enqueue_burst)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table,
+			 unsigned int n, unsigned int *free_space)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,
+			RTE_RING_QUEUE_VARIABLE, __IS_MP, free_space);
+}
+
+/**
+ * Enqueue several objects on a ring (NOT multi-producers safe).
+ */
+static __rte_always_inline unsigned
+__RTE_RING_CONCAT(rte_ring_sp_enqueue_burst)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table,
+			 unsigned int n, unsigned int *free_space)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,
+			RTE_RING_QUEUE_VARIABLE, __IS_SP, free_space);
+}
+
+/**
+ * Enqueue several objects on a ring.
+ */
+static __rte_always_inline unsigned
+__RTE_RING_CONCAT(rte_ring_enqueue_burst)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,
+	unsigned int *free_space)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_enqueue)(r, obj_table, n,
+			RTE_RING_QUEUE_VARIABLE, r->prod.single, free_space);
+}
+
+/**
+ * Dequeue several objects from a ring (multi-consumers safe). When the request
+ * objects are more than the available objects, only dequeue the actual number
+ * of objects
+ */
+static __rte_always_inline unsigned
+__RTE_RING_CONCAT(rte_ring_mc_dequeue_burst)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,
+	unsigned int *available)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,
+			RTE_RING_QUEUE_VARIABLE, __IS_MC, available);
+}
+
+/**
+ * Dequeue several objects from a ring (NOT multi-consumers safe).When the
+ * request objects are more than the available objects, only dequeue the
+ * actual number of objects
+ */
+static __rte_always_inline unsigned
+__RTE_RING_CONCAT(rte_ring_sc_dequeue_burst)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,
+	unsigned int *available)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,
+			RTE_RING_QUEUE_VARIABLE, __IS_SC, available);
+}
+
+/**
+ * Dequeue multiple objects from a ring up to a maximum number.
+ */
+static __rte_always_inline unsigned
+__RTE_RING_CONCAT(rte_ring_dequeue_burst)(struct rte_ring *r,
+	RTE_RING_TMPLT_ELEM_TYPE *obj_table, unsigned int n,
+	unsigned int *available)
+{
+	return __RTE_RING_CONCAT(__rte_ring_do_dequeue)(r, obj_table, n,
+				RTE_RING_QUEUE_VARIABLE,
+				r->cons.single, available);
+}
+
+#ifdef __cplusplus
+}
+#endif
+
+#endif /* _RTE_RING_TEMPLATE_H_ */