diff mbox series

[v2,01/15] crypto: change sgl to src_sgl in vector

Message ID 20210907075957.28848-2-hemant.agrawal@nxp.com
State New
Headers show
Series crypto: add raw vector support in DPAAx | expand

Commit Message

Hemant Agrawal Sept. 7, 2021, 7:59 a.m. UTC
This patch renames the sgl to src_sgl to help differentiating
between source and destination sgl.

Signed-off-by: Hemant Agrawal <hemant.agrawal@nxp.com>

Acked-by: Akhil Goyal <gakhil@marvell.com>

---
 app/test/test_cryptodev.c                  |  6 ++---
 drivers/crypto/aesni_gcm/aesni_gcm_pmd.c   | 12 +++++-----
 drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c |  6 ++---
 drivers/crypto/qat/qat_sym_hw_dp.c         | 27 +++++++++++++---------
 lib/cryptodev/rte_crypto_sym.h             |  2 +-
 lib/ipsec/misc.h                           |  4 ++--
 6 files changed, 31 insertions(+), 26 deletions(-)

-- 
2.17.1

Comments

Ananyev, Konstantin Sept. 16, 2021, 11:38 a.m. UTC | #1
> This patch renames the sgl to src_sgl to help differentiating

> between source and destination sgl.

> 

> Signed-off-by: Hemant Agrawal <hemant.agrawal@nxp.com>

> Acked-by: Akhil Goyal <gakhil@marvell.com>

> ---

>  app/test/test_cryptodev.c                  |  6 ++---

>  drivers/crypto/aesni_gcm/aesni_gcm_pmd.c   | 12 +++++-----

>  drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c |  6 ++---

>  drivers/crypto/qat/qat_sym_hw_dp.c         | 27 +++++++++++++---------

>  lib/cryptodev/rte_crypto_sym.h             |  2 +-

>  lib/ipsec/misc.h                           |  4 ++--

>  6 files changed, 31 insertions(+), 26 deletions(-)

> 

> diff --git a/app/test/test_cryptodev.c b/app/test/test_cryptodev.c

> index 843d07ba37..ed63524edc 100644

> --- a/app/test/test_cryptodev.c

> +++ b/app/test/test_cryptodev.c

> @@ -221,7 +221,7 @@ process_sym_raw_dp_op(uint8_t dev_id, uint16_t qp_id,

>  	digest.va = NULL;

>  	sgl.vec = data_vec;

>  	vec.num = 1;

> -	vec.sgl = &sgl;

> +	vec.src_sgl = &sgl;

>  	vec.iv = &cipher_iv;

>  	vec.digest = &digest;

>  	vec.aad = &aad_auth_iv;

> @@ -385,7 +385,7 @@ process_cpu_aead_op(uint8_t dev_id, struct rte_crypto_op *op)

> 

>  	sgl.vec = vec;

>  	sgl.num = n;

> -	symvec.sgl = &sgl;

> +	symvec.src_sgl = &sgl;

>  	symvec.iv = &iv_ptr;

>  	symvec.digest = &digest_ptr;

>  	symvec.aad = &aad_ptr;

> @@ -431,7 +431,7 @@ process_cpu_crypt_auth_op(uint8_t dev_id, struct rte_crypto_op *op)

> 

>  	sgl.vec = vec;

>  	sgl.num = n;

> -	symvec.sgl = &sgl;

> +	symvec.src_sgl = &sgl;

>  	symvec.iv = &iv_ptr;

>  	symvec.digest = &digest_ptr;

>  	symvec.status = &st;

> diff --git a/drivers/crypto/aesni_gcm/aesni_gcm_pmd.c b/drivers/crypto/aesni_gcm/aesni_gcm_pmd.c

> index 886e2a5aaa..5fbb9b79f8 100644

> --- a/drivers/crypto/aesni_gcm/aesni_gcm_pmd.c

> +++ b/drivers/crypto/aesni_gcm/aesni_gcm_pmd.c

> @@ -535,7 +535,7 @@ aesni_gcm_sgl_encrypt(struct aesni_gcm_session *s,

>  	processed = 0;

>  	for (i = 0; i < vec->num; ++i) {

>  		aesni_gcm_process_gcm_sgl_op(s, gdata_ctx,

> -			&vec->sgl[i], vec->iv[i].va,

> +			&vec->src_sgl[i], vec->iv[i].va,

>  			vec->aad[i].va);

>  		vec->status[i] = aesni_gcm_sgl_op_finalize_encryption(s,

>  			gdata_ctx, vec->digest[i].va);

> @@ -554,7 +554,7 @@ aesni_gcm_sgl_decrypt(struct aesni_gcm_session *s,

>  	processed = 0;

>  	for (i = 0; i < vec->num; ++i) {

>  		aesni_gcm_process_gcm_sgl_op(s, gdata_ctx,

> -			&vec->sgl[i], vec->iv[i].va,

> +			&vec->src_sgl[i], vec->iv[i].va,

>  			vec->aad[i].va);

>  		 vec->status[i] = aesni_gcm_sgl_op_finalize_decryption(s,

>  			gdata_ctx, vec->digest[i].va);

> @@ -572,13 +572,13 @@ aesni_gmac_sgl_generate(struct aesni_gcm_session *s,

> 

>  	processed = 0;

>  	for (i = 0; i < vec->num; ++i) {

> -		if (vec->sgl[i].num != 1) {

> +		if (vec->src_sgl[i].num != 1) {

>  			vec->status[i] = ENOTSUP;

>  			continue;

>  		}

> 

>  		aesni_gcm_process_gmac_sgl_op(s, gdata_ctx,

> -			&vec->sgl[i], vec->iv[i].va);

> +			&vec->src_sgl[i], vec->iv[i].va);

>  		vec->status[i] = aesni_gcm_sgl_op_finalize_encryption(s,

>  			gdata_ctx, vec->digest[i].va);

>  		processed += (vec->status[i] == 0);

> @@ -595,13 +595,13 @@ aesni_gmac_sgl_verify(struct aesni_gcm_session *s,

> 

>  	processed = 0;

>  	for (i = 0; i < vec->num; ++i) {

> -		if (vec->sgl[i].num != 1) {

> +		if (vec->src_sgl[i].num != 1) {

>  			vec->status[i] = ENOTSUP;

>  			continue;

>  		}

> 

>  		aesni_gcm_process_gmac_sgl_op(s, gdata_ctx,

> -			&vec->sgl[i], vec->iv[i].va);

> +			&vec->src_sgl[i], vec->iv[i].va);

>  		vec->status[i] = aesni_gcm_sgl_op_finalize_decryption(s,

>  			gdata_ctx, vec->digest[i].va);

>  		processed += (vec->status[i] == 0);

> diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c

> index a01c826a3c..1b05099446 100644

> --- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c

> +++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c

> @@ -2002,14 +2002,14 @@ aesni_mb_cpu_crypto_process_bulk(struct rte_cryptodev *dev,

>  	for (i = 0, j = 0, k = 0; i != vec->num; i++) {

> 

> 

> -		ret = check_crypto_sgl(sofs, vec->sgl + i);

> +		ret = check_crypto_sgl(sofs, vec->src_sgl + i);

>  		if (ret != 0) {

>  			vec->status[i] = ret;

>  			continue;

>  		}

> 

> -		buf = vec->sgl[i].vec[0].base;

> -		len = vec->sgl[i].vec[0].len;

> +		buf = vec->src_sgl[i].vec[0].base;

> +		len = vec->src_sgl[i].vec[0].len;

> 

>  		job = IMB_GET_NEXT_JOB(mb_mgr);

>  		if (job == NULL) {

> diff --git a/drivers/crypto/qat/qat_sym_hw_dp.c b/drivers/crypto/qat/qat_sym_hw_dp.c

> index ac9ac05363..4870ebf66a 100644

> --- a/drivers/crypto/qat/qat_sym_hw_dp.c

> +++ b/drivers/crypto/qat/qat_sym_hw_dp.c

> @@ -181,8 +181,9 @@ qat_sym_dp_enqueue_cipher_jobs(void *qp_data, uint8_t *drv_ctx,

>  			(uint8_t *)tx_queue->base_addr + tail);

>  		rte_mov128((uint8_t *)req, (const uint8_t *)&(ctx->fw_req));

> 

> -		data_len = qat_sym_dp_parse_data_vec(qp, req, vec->sgl[i].vec,

> -			vec->sgl[i].num);

> +		data_len = qat_sym_dp_parse_data_vec(qp, req,

> +			vec->src_sgl[i].vec,

> +			vec->src_sgl[i].num);

>  		if (unlikely(data_len < 0))

>  			break;

>  		req->comn_mid.opaque_data = (uint64_t)(uintptr_t)user_data[i];

> @@ -302,8 +303,9 @@ qat_sym_dp_enqueue_auth_jobs(void *qp_data, uint8_t *drv_ctx,

>  			(uint8_t *)tx_queue->base_addr + tail);

>  		rte_mov128((uint8_t *)req, (const uint8_t *)&(ctx->fw_req));

> 

> -		data_len = qat_sym_dp_parse_data_vec(qp, req, vec->sgl[i].vec,

> -			vec->sgl[i].num);

> +		data_len = qat_sym_dp_parse_data_vec(qp, req,

> +			vec->src_sgl[i].vec,

> +			vec->src_sgl[i].num);

>  		if (unlikely(data_len < 0))

>  			break;

>  		req->comn_mid.opaque_data = (uint64_t)(uintptr_t)user_data[i];

> @@ -484,14 +486,16 @@ qat_sym_dp_enqueue_chain_jobs(void *qp_data, uint8_t *drv_ctx,

>  			(uint8_t *)tx_queue->base_addr + tail);

>  		rte_mov128((uint8_t *)req, (const uint8_t *)&(ctx->fw_req));

> 

> -		data_len = qat_sym_dp_parse_data_vec(qp, req, vec->sgl[i].vec,

> -			vec->sgl[i].num);

> +		data_len = qat_sym_dp_parse_data_vec(qp, req,

> +			vec->src_sgl[i].vec,

> +			vec->src_sgl[i].num);

>  		if (unlikely(data_len < 0))

>  			break;

>  		req->comn_mid.opaque_data = (uint64_t)(uintptr_t)user_data[i];

> -		if (unlikely(enqueue_one_chain_job(ctx, req, vec->sgl[i].vec,

> -			vec->sgl[i].num, &vec->iv[i], &vec->digest[i],

> -				&vec->auth_iv[i], ofs, (uint32_t)data_len)))

> +		if (unlikely(enqueue_one_chain_job(ctx, req,

> +			vec->src_sgl[i].vec, vec->src_sgl[i].num,

> +			&vec->iv[i], &vec->digest[i],

> +			&vec->auth_iv[i], ofs, (uint32_t)data_len)))

>  			break;

> 

>  		tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask;

> @@ -688,8 +692,9 @@ qat_sym_dp_enqueue_aead_jobs(void *qp_data, uint8_t *drv_ctx,

>  			(uint8_t *)tx_queue->base_addr + tail);

>  		rte_mov128((uint8_t *)req, (const uint8_t *)&(ctx->fw_req));

> 

> -		data_len = qat_sym_dp_parse_data_vec(qp, req, vec->sgl[i].vec,

> -			vec->sgl[i].num);

> +		data_len = qat_sym_dp_parse_data_vec(qp, req,

> +			vec->src_sgl[i].vec,

> +			vec->src_sgl[i].num);

>  		if (unlikely(data_len < 0))

>  			break;

>  		req->comn_mid.opaque_data = (uint64_t)(uintptr_t)user_data[i];

> diff --git a/lib/cryptodev/rte_crypto_sym.h b/lib/cryptodev/rte_crypto_sym.h

> index 58c0724743..dcc0bd5933 100644

> --- a/lib/cryptodev/rte_crypto_sym.h

> +++ b/lib/cryptodev/rte_crypto_sym.h

> @@ -69,7 +69,7 @@ struct rte_crypto_sym_vec {

>  	/** number of operations to perform */

>  	uint32_t num;

>  	/** array of SGL vectors */

> -	struct rte_crypto_sgl *sgl;

> +	struct rte_crypto_sgl *src_sgl;

>  	/** array of pointers to cipher IV */

>  	struct rte_crypto_va_iova_ptr *iv;

>  	/** array of pointers to digest */

> diff --git a/lib/ipsec/misc.h b/lib/ipsec/misc.h

> index 79b9a20762..58ff538141 100644

> --- a/lib/ipsec/misc.h

> +++ b/lib/ipsec/misc.h

> @@ -136,7 +136,7 @@ cpu_crypto_bulk(const struct rte_ipsec_session *ss,

>  		/* not enough space in vec[] to hold all segments */

>  		if (vcnt < 0) {

>  			/* fill the request structure */

> -			symvec.sgl = &vecpkt[j];

> +			symvec.src_sgl = &vecpkt[j];

>  			symvec.iv = &iv[j];

>  			symvec.digest = &dgst[j];

>  			symvec.aad = &aad[j];

> @@ -160,7 +160,7 @@ cpu_crypto_bulk(const struct rte_ipsec_session *ss,

>  	}

> 

>  	/* fill the request structure */

> -	symvec.sgl = &vecpkt[j];

> +	symvec.src_sgl = &vecpkt[j];

>  	symvec.iv = &iv[j];

>  	symvec.aad = &aad[j];

>  	symvec.digest = &dgst[j];

> --


Acked-by: Konstantin Ananyev <konstantin.ananyev@intel.com>


> 2.17.1
diff mbox series

Patch

diff --git a/app/test/test_cryptodev.c b/app/test/test_cryptodev.c
index 843d07ba37..ed63524edc 100644
--- a/app/test/test_cryptodev.c
+++ b/app/test/test_cryptodev.c
@@ -221,7 +221,7 @@  process_sym_raw_dp_op(uint8_t dev_id, uint16_t qp_id,
 	digest.va = NULL;
 	sgl.vec = data_vec;
 	vec.num = 1;
-	vec.sgl = &sgl;
+	vec.src_sgl = &sgl;
 	vec.iv = &cipher_iv;
 	vec.digest = &digest;
 	vec.aad = &aad_auth_iv;
@@ -385,7 +385,7 @@  process_cpu_aead_op(uint8_t dev_id, struct rte_crypto_op *op)
 
 	sgl.vec = vec;
 	sgl.num = n;
-	symvec.sgl = &sgl;
+	symvec.src_sgl = &sgl;
 	symvec.iv = &iv_ptr;
 	symvec.digest = &digest_ptr;
 	symvec.aad = &aad_ptr;
@@ -431,7 +431,7 @@  process_cpu_crypt_auth_op(uint8_t dev_id, struct rte_crypto_op *op)
 
 	sgl.vec = vec;
 	sgl.num = n;
-	symvec.sgl = &sgl;
+	symvec.src_sgl = &sgl;
 	symvec.iv = &iv_ptr;
 	symvec.digest = &digest_ptr;
 	symvec.status = &st;
diff --git a/drivers/crypto/aesni_gcm/aesni_gcm_pmd.c b/drivers/crypto/aesni_gcm/aesni_gcm_pmd.c
index 886e2a5aaa..5fbb9b79f8 100644
--- a/drivers/crypto/aesni_gcm/aesni_gcm_pmd.c
+++ b/drivers/crypto/aesni_gcm/aesni_gcm_pmd.c
@@ -535,7 +535,7 @@  aesni_gcm_sgl_encrypt(struct aesni_gcm_session *s,
 	processed = 0;
 	for (i = 0; i < vec->num; ++i) {
 		aesni_gcm_process_gcm_sgl_op(s, gdata_ctx,
-			&vec->sgl[i], vec->iv[i].va,
+			&vec->src_sgl[i], vec->iv[i].va,
 			vec->aad[i].va);
 		vec->status[i] = aesni_gcm_sgl_op_finalize_encryption(s,
 			gdata_ctx, vec->digest[i].va);
@@ -554,7 +554,7 @@  aesni_gcm_sgl_decrypt(struct aesni_gcm_session *s,
 	processed = 0;
 	for (i = 0; i < vec->num; ++i) {
 		aesni_gcm_process_gcm_sgl_op(s, gdata_ctx,
-			&vec->sgl[i], vec->iv[i].va,
+			&vec->src_sgl[i], vec->iv[i].va,
 			vec->aad[i].va);
 		 vec->status[i] = aesni_gcm_sgl_op_finalize_decryption(s,
 			gdata_ctx, vec->digest[i].va);
@@ -572,13 +572,13 @@  aesni_gmac_sgl_generate(struct aesni_gcm_session *s,
 
 	processed = 0;
 	for (i = 0; i < vec->num; ++i) {
-		if (vec->sgl[i].num != 1) {
+		if (vec->src_sgl[i].num != 1) {
 			vec->status[i] = ENOTSUP;
 			continue;
 		}
 
 		aesni_gcm_process_gmac_sgl_op(s, gdata_ctx,
-			&vec->sgl[i], vec->iv[i].va);
+			&vec->src_sgl[i], vec->iv[i].va);
 		vec->status[i] = aesni_gcm_sgl_op_finalize_encryption(s,
 			gdata_ctx, vec->digest[i].va);
 		processed += (vec->status[i] == 0);
@@ -595,13 +595,13 @@  aesni_gmac_sgl_verify(struct aesni_gcm_session *s,
 
 	processed = 0;
 	for (i = 0; i < vec->num; ++i) {
-		if (vec->sgl[i].num != 1) {
+		if (vec->src_sgl[i].num != 1) {
 			vec->status[i] = ENOTSUP;
 			continue;
 		}
 
 		aesni_gcm_process_gmac_sgl_op(s, gdata_ctx,
-			&vec->sgl[i], vec->iv[i].va);
+			&vec->src_sgl[i], vec->iv[i].va);
 		vec->status[i] = aesni_gcm_sgl_op_finalize_decryption(s,
 			gdata_ctx, vec->digest[i].va);
 		processed += (vec->status[i] == 0);
diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
index a01c826a3c..1b05099446 100644
--- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
+++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
@@ -2002,14 +2002,14 @@  aesni_mb_cpu_crypto_process_bulk(struct rte_cryptodev *dev,
 	for (i = 0, j = 0, k = 0; i != vec->num; i++) {
 
 
-		ret = check_crypto_sgl(sofs, vec->sgl + i);
+		ret = check_crypto_sgl(sofs, vec->src_sgl + i);
 		if (ret != 0) {
 			vec->status[i] = ret;
 			continue;
 		}
 
-		buf = vec->sgl[i].vec[0].base;
-		len = vec->sgl[i].vec[0].len;
+		buf = vec->src_sgl[i].vec[0].base;
+		len = vec->src_sgl[i].vec[0].len;
 
 		job = IMB_GET_NEXT_JOB(mb_mgr);
 		if (job == NULL) {
diff --git a/drivers/crypto/qat/qat_sym_hw_dp.c b/drivers/crypto/qat/qat_sym_hw_dp.c
index ac9ac05363..4870ebf66a 100644
--- a/drivers/crypto/qat/qat_sym_hw_dp.c
+++ b/drivers/crypto/qat/qat_sym_hw_dp.c
@@ -181,8 +181,9 @@  qat_sym_dp_enqueue_cipher_jobs(void *qp_data, uint8_t *drv_ctx,
 			(uint8_t *)tx_queue->base_addr + tail);
 		rte_mov128((uint8_t *)req, (const uint8_t *)&(ctx->fw_req));
 
-		data_len = qat_sym_dp_parse_data_vec(qp, req, vec->sgl[i].vec,
-			vec->sgl[i].num);
+		data_len = qat_sym_dp_parse_data_vec(qp, req,
+			vec->src_sgl[i].vec,
+			vec->src_sgl[i].num);
 		if (unlikely(data_len < 0))
 			break;
 		req->comn_mid.opaque_data = (uint64_t)(uintptr_t)user_data[i];
@@ -302,8 +303,9 @@  qat_sym_dp_enqueue_auth_jobs(void *qp_data, uint8_t *drv_ctx,
 			(uint8_t *)tx_queue->base_addr + tail);
 		rte_mov128((uint8_t *)req, (const uint8_t *)&(ctx->fw_req));
 
-		data_len = qat_sym_dp_parse_data_vec(qp, req, vec->sgl[i].vec,
-			vec->sgl[i].num);
+		data_len = qat_sym_dp_parse_data_vec(qp, req,
+			vec->src_sgl[i].vec,
+			vec->src_sgl[i].num);
 		if (unlikely(data_len < 0))
 			break;
 		req->comn_mid.opaque_data = (uint64_t)(uintptr_t)user_data[i];
@@ -484,14 +486,16 @@  qat_sym_dp_enqueue_chain_jobs(void *qp_data, uint8_t *drv_ctx,
 			(uint8_t *)tx_queue->base_addr + tail);
 		rte_mov128((uint8_t *)req, (const uint8_t *)&(ctx->fw_req));
 
-		data_len = qat_sym_dp_parse_data_vec(qp, req, vec->sgl[i].vec,
-			vec->sgl[i].num);
+		data_len = qat_sym_dp_parse_data_vec(qp, req,
+			vec->src_sgl[i].vec,
+			vec->src_sgl[i].num);
 		if (unlikely(data_len < 0))
 			break;
 		req->comn_mid.opaque_data = (uint64_t)(uintptr_t)user_data[i];
-		if (unlikely(enqueue_one_chain_job(ctx, req, vec->sgl[i].vec,
-			vec->sgl[i].num, &vec->iv[i], &vec->digest[i],
-				&vec->auth_iv[i], ofs, (uint32_t)data_len)))
+		if (unlikely(enqueue_one_chain_job(ctx, req,
+			vec->src_sgl[i].vec, vec->src_sgl[i].num,
+			&vec->iv[i], &vec->digest[i],
+			&vec->auth_iv[i], ofs, (uint32_t)data_len)))
 			break;
 
 		tail = (tail + tx_queue->msg_size) & tx_queue->modulo_mask;
@@ -688,8 +692,9 @@  qat_sym_dp_enqueue_aead_jobs(void *qp_data, uint8_t *drv_ctx,
 			(uint8_t *)tx_queue->base_addr + tail);
 		rte_mov128((uint8_t *)req, (const uint8_t *)&(ctx->fw_req));
 
-		data_len = qat_sym_dp_parse_data_vec(qp, req, vec->sgl[i].vec,
-			vec->sgl[i].num);
+		data_len = qat_sym_dp_parse_data_vec(qp, req,
+			vec->src_sgl[i].vec,
+			vec->src_sgl[i].num);
 		if (unlikely(data_len < 0))
 			break;
 		req->comn_mid.opaque_data = (uint64_t)(uintptr_t)user_data[i];
diff --git a/lib/cryptodev/rte_crypto_sym.h b/lib/cryptodev/rte_crypto_sym.h
index 58c0724743..dcc0bd5933 100644
--- a/lib/cryptodev/rte_crypto_sym.h
+++ b/lib/cryptodev/rte_crypto_sym.h
@@ -69,7 +69,7 @@  struct rte_crypto_sym_vec {
 	/** number of operations to perform */
 	uint32_t num;
 	/** array of SGL vectors */
-	struct rte_crypto_sgl *sgl;
+	struct rte_crypto_sgl *src_sgl;
 	/** array of pointers to cipher IV */
 	struct rte_crypto_va_iova_ptr *iv;
 	/** array of pointers to digest */
diff --git a/lib/ipsec/misc.h b/lib/ipsec/misc.h
index 79b9a20762..58ff538141 100644
--- a/lib/ipsec/misc.h
+++ b/lib/ipsec/misc.h
@@ -136,7 +136,7 @@  cpu_crypto_bulk(const struct rte_ipsec_session *ss,
 		/* not enough space in vec[] to hold all segments */
 		if (vcnt < 0) {
 			/* fill the request structure */
-			symvec.sgl = &vecpkt[j];
+			symvec.src_sgl = &vecpkt[j];
 			symvec.iv = &iv[j];
 			symvec.digest = &dgst[j];
 			symvec.aad = &aad[j];
@@ -160,7 +160,7 @@  cpu_crypto_bulk(const struct rte_ipsec_session *ss,
 	}
 
 	/* fill the request structure */
-	symvec.sgl = &vecpkt[j];
+	symvec.src_sgl = &vecpkt[j];
 	symvec.iv = &iv[j];
 	symvec.aad = &aad[j];
 	symvec.digest = &dgst[j];