From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 65CF9A0506; Mon, 4 Apr 2022 15:06:54 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 99E9C42826; Mon, 4 Apr 2022 15:06:50 +0200 (CEST) Received: from mga18.intel.com (mga18.intel.com [134.134.136.126]) by mails.dpdk.org (Postfix) with ESMTP id 8DBE04068C for ; Mon, 4 Apr 2022 15:06:46 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1649077606; x=1680613606; h=from:to:cc:subject:date:message-id; bh=YkS/G38bYipa26IllYE0iuyXhaTeLy64GNmMPzr3AJA=; b=eIa4KhxTz7cNl2oZ3w7uN6g/Hx/gT2CF5akF2oMXOpR4XOLLLYygSqCV nG5EPdoNx3YsZoZZj5S8wrBvWgjTBP1L4Y/Mkatp8S67nyZwduR1Ds2WE 1c9HwTR9Wx4DPYxZZQNyTJ6H9nmhSg5rBxmRwIL5iHY0Dc43GwDScrTr3 V1KP3WbSwULLzthFFYh5e1NKxwiUoKXcipN8mNWT/EHqjeDDdhEpeciIa V1BzmuAUk4Rc1E11f/RuDGaUTuQvHBWwp37QoNRR9YW54dpvvx3XKCdDr bjcqzCK2bvriwP8hiRBfJU2AbQ2U2NLfDijNf330jqMn/GZ3YbSeUiW3G A==; X-IronPort-AV: E=McAfee;i="6200,9189,10306"; a="242650352" X-IronPort-AV: E=Sophos;i="5.90,234,1643702400"; d="scan'208";a="242650352" Received: from orsmga006.jf.intel.com ([10.7.209.51]) by orsmga106.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 04 Apr 2022 06:06:25 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.90,234,1643702400"; d="scan'208";a="523573768" Received: from silpixa00400308.ir.intel.com ([10.237.214.95]) by orsmga006.jf.intel.com with ESMTP; 04 Apr 2022 06:06:23 -0700 From: Arek Kusztal To: dev@dpdk.org Cc: gakhil@marvell.com, roy.fan.zhang@intel.com, Arek Kusztal Subject: [PATCH 1/2] crypto/qat: refactor asym algorithm macros and logs Date: Mon, 4 Apr 2022 14:06:16 +0100 Message-Id: <20220404130617.5225-1-arkadiuszx.kusztal@intel.com> X-Mailer: git-send-email 2.17.1 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org This commit unifies macros for asymmetric parameters, therefore making code easier to maintain. It additionally changes some of PMD output logs that right now can only be seen in debug mode. Signed-off-by: Arek Kusztal --- drivers/crypto/qat/qat_asym.c | 230 ++++++++++++++++++------------------------ drivers/crypto/qat/qat_asym.h | 3 +- drivers/crypto/qat/qat_ec.h | 1 - 3 files changed, 101 insertions(+), 133 deletions(-) diff --git a/drivers/crypto/qat/qat_asym.c b/drivers/crypto/qat/qat_asym.c index 479d5308cf..d2041b2efa 100644 --- a/drivers/crypto/qat/qat_asym.c +++ b/drivers/crypto/qat/qat_asym.c @@ -34,7 +34,7 @@ static const struct rte_driver cryptodev_qat_asym_driver = { /* * Macros with suffix _F are used with some of predefinded identifiers: * - cookie->input_buffer - * - qat_alg_bytesize + * - qat_func_alignsize */ #if RTE_LOG_DP_LEVEL >= RTE_LOG_DEBUG #define HEXDUMP(name, where, size) QAT_DP_HEXDUMP_LOG(DEBUG, name, \ @@ -43,8 +43,8 @@ static const struct rte_driver cryptodev_qat_asym_driver = { &where[idx * size], size) #define HEXDUMP_OFF_F(name, idx) QAT_DP_HEXDUMP_LOG(DEBUG, name, \ - &cookie->input_buffer[idx * qat_alg_bytesize], \ - qat_alg_bytesize) + &cookie->input_buffer[idx * qat_func_alignsize], \ + qat_func_alignsize) #else #define HEXDUMP(name, where, size) #define HEXDUMP_OFF(name, where, size, idx) @@ -69,36 +69,28 @@ static const struct rte_driver cryptodev_qat_asym_driver = { } \ } while (0) -#define SET_PKE_LN(where, what, how, idx) \ - rte_memcpy(where[idx] + how - \ - what.length, \ - what.data, \ - what.length) - -#define SET_PKE_LN_9A(where, what, how, idx) \ - rte_memcpy(&where[idx * RTE_ALIGN_CEIL(how, 8)] + \ - RTE_ALIGN_CEIL(how, 8) - \ +#define SET_PKE_LN(what, how, idx) \ + rte_memcpy(cookie->input_array[idx] + how - \ what.length, \ what.data, \ what.length) -#define SET_PKE_LN_EC(where, what, how, idx) \ - rte_memcpy(where[idx] + \ - RTE_ALIGN_CEIL(how, 8) - \ - how, \ - what.data, \ - how) +#define SET_PKE_LN_EC(curve, p, idx) \ + rte_memcpy(cookie->input_array[idx] + \ + qat_func_alignsize - curve.bytesize, \ + curve.p.data, curve.bytesize) -#define SET_PKE_LN_9A_F(what, idx) \ - rte_memcpy(&cookie->input_buffer[idx * qat_alg_bytesize] + \ - qat_alg_bytesize - what.length, \ +#define SET_PKE_9A_IN(what, idx) \ + rte_memcpy(&cookie->input_buffer[idx * \ + qat_func_alignsize] + \ + qat_func_alignsize - what.length, \ what.data, what.length) -#define SET_PKE_LN_EC_F(what, how, idx) \ +#define SET_PKE_9A_EC(curve, p, idx) \ rte_memcpy(&cookie->input_buffer[idx * \ - RTE_ALIGN_CEIL(how, 8)] + \ - RTE_ALIGN_CEIL(how, 8) - how, \ - what.data, how) + qat_func_alignsize] + \ + qat_func_alignsize - curve.bytesize, \ + curve.p.data, curve.bytesize) static void request_init(struct icp_qat_fw_pke_request *qat_req) @@ -231,12 +223,9 @@ modexp_set_input(struct rte_crypto_asym_op *asym_op, } alg_bytesize = qat_function.bytesize; - SET_PKE_LN(cookie->input_array, asym_op->modex.base, - alg_bytesize, 0); - SET_PKE_LN(cookie->input_array, xform->modex.exponent, - alg_bytesize, 1); - SET_PKE_LN(cookie->input_array, xform->modex.modulus, - alg_bytesize, 2); + SET_PKE_LN(asym_op->modex.base, alg_bytesize, 0); + SET_PKE_LN(xform->modex.exponent, alg_bytesize, 1); + SET_PKE_LN(xform->modex.modulus, alg_bytesize, 2); cookie->alg_bytesize = alg_bytesize; qat_req->pke_hdr.cd_pars.func_id = func_id; @@ -290,10 +279,8 @@ modinv_set_input(struct rte_crypto_asym_op *asym_op, } alg_bytesize = qat_function.bytesize; - SET_PKE_LN(cookie->input_array, asym_op->modinv.base, - alg_bytesize, 0); - SET_PKE_LN(cookie->input_array, xform->modinv.modulus, - alg_bytesize, 1); + SET_PKE_LN(asym_op->modinv.base, alg_bytesize, 0); + SET_PKE_LN(xform->modinv.modulus, alg_bytesize, 1); cookie->alg_bytesize = alg_bytesize; qat_req->pke_hdr.cd_pars.func_id = func_id; @@ -347,8 +334,7 @@ rsa_set_pub_input(struct rte_crypto_asym_op *asym_op, if (asym_op->rsa.op_type == RTE_CRYPTO_ASYM_OP_ENCRYPT) { switch (asym_op->rsa.pad) { case RTE_CRYPTO_RSA_PADDING_NONE: - SET_PKE_LN(cookie->input_array, asym_op->rsa.message, - alg_bytesize, 0); + SET_PKE_LN(asym_op->rsa.message, alg_bytesize, 0); break; default: QAT_LOG(ERR, @@ -360,8 +346,7 @@ rsa_set_pub_input(struct rte_crypto_asym_op *asym_op, } else { switch (asym_op->rsa.pad) { case RTE_CRYPTO_RSA_PADDING_NONE: - SET_PKE_LN(cookie->input_array, asym_op->rsa.sign, - alg_bytesize, 0); + SET_PKE_LN(asym_op->rsa.sign, alg_bytesize, 0); break; default: QAT_LOG(ERR, @@ -372,10 +357,8 @@ rsa_set_pub_input(struct rte_crypto_asym_op *asym_op, alg_bytesize); } - SET_PKE_LN(cookie->input_array, xform->rsa.e, - alg_bytesize, 1); - SET_PKE_LN(cookie->input_array, xform->rsa.n, - alg_bytesize, 2); + SET_PKE_LN(xform->rsa.e, alg_bytesize, 1); + SET_PKE_LN(xform->rsa.n, alg_bytesize, 2); cookie->alg_bytesize = alg_bytesize; qat_req->pke_hdr.cd_pars.func_id = func_id; @@ -407,16 +390,11 @@ rsa_set_priv_input(struct rte_crypto_asym_op *asym_op, qat_req->input_param_count = QAT_ASYM_RSA_QT_NUM_IN_PARAMS; - SET_PKE_LN(cookie->input_array, xform->rsa.qt.p, - (alg_bytesize >> 1), 1); - SET_PKE_LN(cookie->input_array, xform->rsa.qt.q, - (alg_bytesize >> 1), 2); - SET_PKE_LN(cookie->input_array, xform->rsa.qt.dP, - (alg_bytesize >> 1), 3); - SET_PKE_LN(cookie->input_array, xform->rsa.qt.dQ, - (alg_bytesize >> 1), 4); - SET_PKE_LN(cookie->input_array, xform->rsa.qt.qInv, - (alg_bytesize >> 1), 5); + SET_PKE_LN(xform->rsa.qt.p, (alg_bytesize >> 1), 1); + SET_PKE_LN(xform->rsa.qt.q, (alg_bytesize >> 1), 2); + SET_PKE_LN(xform->rsa.qt.dP, (alg_bytesize >> 1), 3); + SET_PKE_LN(xform->rsa.qt.dQ, (alg_bytesize >> 1), 4); + SET_PKE_LN(xform->rsa.qt.qInv, (alg_bytesize >> 1), 5); HEXDUMP("RSA p", cookie->input_array[1], alg_bytesize); @@ -438,10 +416,8 @@ rsa_set_priv_input(struct rte_crypto_asym_op *asym_op, } alg_bytesize = qat_function.bytesize; - SET_PKE_LN(cookie->input_array, xform->rsa.d, - alg_bytesize, 1); - SET_PKE_LN(cookie->input_array, xform->rsa.n, - alg_bytesize, 2); + SET_PKE_LN(xform->rsa.d, alg_bytesize, 1); + SET_PKE_LN(xform->rsa.n, alg_bytesize, 2); HEXDUMP("RSA d", cookie->input_array[1], alg_bytesize); @@ -456,8 +432,7 @@ rsa_set_priv_input(struct rte_crypto_asym_op *asym_op, RTE_CRYPTO_ASYM_OP_DECRYPT) { switch (asym_op->rsa.pad) { case RTE_CRYPTO_RSA_PADDING_NONE: - SET_PKE_LN(cookie->input_array, asym_op->rsa.cipher, - alg_bytesize, 0); + SET_PKE_LN(asym_op->rsa.cipher, alg_bytesize, 0); HEXDUMP("RSA ciphertext", cookie->input_array[0], alg_bytesize); break; @@ -471,8 +446,7 @@ rsa_set_priv_input(struct rte_crypto_asym_op *asym_op, RTE_CRYPTO_ASYM_OP_SIGN) { switch (asym_op->rsa.pad) { case RTE_CRYPTO_RSA_PADDING_NONE: - SET_PKE_LN(cookie->input_array, asym_op->rsa.message, - alg_bytesize, 0); + SET_PKE_LN(asym_op->rsa.message, alg_bytesize, 0); HEXDUMP("RSA text to be signed", cookie->input_array[0], alg_bytesize); break; @@ -580,12 +554,12 @@ ecdsa_set_input(struct rte_crypto_asym_op *asym_op, struct rte_crypto_asym_xform *xform) { struct qat_asym_function qat_function; - uint32_t alg_bytesize, qat_alg_bytesize, func_id; + uint32_t qat_func_alignsize, func_id; int curve_id; curve_id = pick_curve(xform); if (curve_id < 0) { - QAT_LOG(ERR, "Incorrect elliptic curve"); + QAT_LOG(DEBUG, "Incorrect elliptic curve"); return -EINVAL; } @@ -597,20 +571,21 @@ ecdsa_set_input(struct rte_crypto_asym_op *asym_op, QAT_LOG(ERR, "Cannot obtain functionality id"); return -EINVAL; } - alg_bytesize = qat_function.bytesize; - qat_alg_bytesize = RTE_ALIGN_CEIL(alg_bytesize, 8); - - SET_PKE_LN_9A_F(asym_op->ecdsa.pkey, 0); - SET_PKE_LN_9A_F(asym_op->ecdsa.message, 1); - SET_PKE_LN_9A_F(asym_op->ecdsa.k, 2); - SET_PKE_LN_EC_F(curve[curve_id].b, alg_bytesize, 3); - SET_PKE_LN_EC_F(curve[curve_id].a, alg_bytesize, 4); - SET_PKE_LN_EC_F(curve[curve_id].p, alg_bytesize, 5); - SET_PKE_LN_EC_F(curve[curve_id].n, alg_bytesize, 6); - SET_PKE_LN_EC_F(curve[curve_id].y, alg_bytesize, 7); - SET_PKE_LN_EC_F(curve[curve_id].x, alg_bytesize, 8); - - cookie->alg_bytesize = alg_bytesize; + qat_func_alignsize = + RTE_ALIGN_CEIL(qat_function.bytesize, 8); + + SET_PKE_9A_IN(asym_op->ecdsa.pkey, 0); + SET_PKE_9A_IN(asym_op->ecdsa.message, 1); + SET_PKE_9A_IN(asym_op->ecdsa.k, 2); + SET_PKE_9A_EC(curve[curve_id], b, 3); + SET_PKE_9A_EC(curve[curve_id], a, 4); + SET_PKE_9A_EC(curve[curve_id], p, 5); + SET_PKE_9A_EC(curve[curve_id], n, 6); + SET_PKE_9A_EC(curve[curve_id], y, 7); + SET_PKE_9A_EC(curve[curve_id], x, 8); + + cookie->alg_bytesize = curve[curve_id].bytesize; + cookie->qat_func_alignsize = qat_func_alignsize; qat_req->pke_hdr.cd_pars.func_id = func_id; qat_req->input_param_count = QAT_ASYM_ECDSA_RS_SIGN_IN_PARAMS; @@ -633,22 +608,22 @@ ecdsa_set_input(struct rte_crypto_asym_op *asym_op, QAT_LOG(ERR, "Cannot obtain functionality id"); return -EINVAL; } - alg_bytesize = qat_function.bytesize; - qat_alg_bytesize = RTE_ALIGN_CEIL(alg_bytesize, 8); - - SET_PKE_LN_9A_F(asym_op->ecdsa.message, 10); - SET_PKE_LN_9A_F(asym_op->ecdsa.s, 9); - SET_PKE_LN_9A_F(asym_op->ecdsa.r, 8); - SET_PKE_LN_EC_F(curve[curve_id].n, alg_bytesize, 7); - SET_PKE_LN_EC_F(curve[curve_id].x, alg_bytesize, 6); - SET_PKE_LN_EC_F(curve[curve_id].y, alg_bytesize, 5); - SET_PKE_LN_9A_F(asym_op->ecdsa.q.x, 4); - SET_PKE_LN_9A_F(asym_op->ecdsa.q.y, 3); - SET_PKE_LN_EC_F(curve[curve_id].a, alg_bytesize, 2); - SET_PKE_LN_EC_F(curve[curve_id].b, alg_bytesize, 1); - SET_PKE_LN_EC_F(curve[curve_id].p, alg_bytesize, 0); - - cookie->alg_bytesize = alg_bytesize; + qat_func_alignsize = RTE_ALIGN_CEIL(qat_function.bytesize, 8); + + SET_PKE_9A_IN(asym_op->ecdsa.message, 10); + SET_PKE_9A_IN(asym_op->ecdsa.s, 9); + SET_PKE_9A_IN(asym_op->ecdsa.r, 8); + SET_PKE_9A_EC(curve[curve_id], n, 7); + SET_PKE_9A_EC(curve[curve_id], x, 6); + SET_PKE_9A_EC(curve[curve_id], y, 5); + SET_PKE_9A_IN(asym_op->ecdsa.q.x, 4); + SET_PKE_9A_IN(asym_op->ecdsa.q.y, 3); + SET_PKE_9A_EC(curve[curve_id], a, 2); + SET_PKE_9A_EC(curve[curve_id], b, 1); + SET_PKE_9A_EC(curve[curve_id], p, 0); + + cookie->alg_bytesize = curve[curve_id].bytesize; + cookie->qat_func_alignsize = qat_func_alignsize; qat_req->pke_hdr.cd_pars.func_id = func_id; qat_req->input_param_count = QAT_ASYM_ECDSA_RS_VERIFY_IN_PARAMS; @@ -679,8 +654,8 @@ ecdsa_collect(struct rte_crypto_asym_op *asym_op, struct qat_asym_op_cookie *cookie) { uint32_t alg_bytesize = cookie->alg_bytesize; - uint32_t qat_alg_bytesize = RTE_ALIGN_CEIL(cookie->alg_bytesize, 8); - uint32_t ltrim = qat_alg_bytesize - alg_bytesize; + uint32_t qat_func_alignsize = cookie->qat_func_alignsize; + uint32_t ltrim = qat_func_alignsize - alg_bytesize; if (asym_op->rsa.op_type == RTE_CRYPTO_ASYM_OP_SIGN) { uint8_t *r = asym_op->ecdsa.r.data; @@ -692,9 +667,9 @@ ecdsa_collect(struct rte_crypto_asym_op *asym_op, rte_memcpy(s, &cookie->output_array[1][ltrim], alg_bytesize); HEXDUMP("R", cookie->output_array[0], - alg_bytesize); + qat_func_alignsize); HEXDUMP("S", cookie->output_array[1], - alg_bytesize); + qat_func_alignsize); } return RTE_CRYPTO_OP_STATUS_SUCCESS; } @@ -706,12 +681,12 @@ ecpm_set_input(struct rte_crypto_asym_op *asym_op, struct rte_crypto_asym_xform *xform) { struct qat_asym_function qat_function; - uint32_t alg_bytesize, __rte_unused qat_alg_bytesize, func_id; + uint32_t qat_func_alignsize, func_id; int curve_id; curve_id = pick_curve(xform); if (curve_id < 0) { - QAT_LOG(ERR, "Incorrect elliptic curve"); + QAT_LOG(DEBUG, "Incorrect elliptic curve"); return -EINVAL; } @@ -721,38 +696,31 @@ ecpm_set_input(struct rte_crypto_asym_op *asym_op, QAT_LOG(ERR, "Cannot obtain functionality id"); return -EINVAL; } - alg_bytesize = qat_function.bytesize; - qat_alg_bytesize = RTE_ALIGN_CEIL(alg_bytesize, 8); - - SET_PKE_LN_EC(cookie->input_array, asym_op->ecpm.scalar, - asym_op->ecpm.scalar.length, 0); - SET_PKE_LN_EC(cookie->input_array, asym_op->ecpm.p.x, - asym_op->ecpm.p.x.length, 1); - SET_PKE_LN_EC(cookie->input_array, asym_op->ecpm.p.y, - asym_op->ecpm.p.y.length, 2); - SET_PKE_LN_EC(cookie->input_array, curve[curve_id].a, - alg_bytesize, 3); - SET_PKE_LN_EC(cookie->input_array, curve[curve_id].b, - alg_bytesize, 4); - SET_PKE_LN_EC(cookie->input_array, curve[curve_id].p, - alg_bytesize, 5); - SET_PKE_LN_EC(cookie->input_array, curve[curve_id].h, - alg_bytesize, 6); - - cookie->alg_bytesize = alg_bytesize; + qat_func_alignsize = RTE_ALIGN_CEIL(qat_function.bytesize, 8); + + SET_PKE_LN(asym_op->ecpm.scalar, qat_func_alignsize, 0); + SET_PKE_LN(asym_op->ecpm.p.x, qat_func_alignsize, 1); + SET_PKE_LN(asym_op->ecpm.p.y, qat_func_alignsize, 2); + SET_PKE_LN_EC(curve[curve_id], a, 3); + SET_PKE_LN_EC(curve[curve_id], b, 4); + SET_PKE_LN_EC(curve[curve_id], p, 5); + SET_PKE_LN_EC(curve[curve_id], h, 6); + + cookie->alg_bytesize = curve[curve_id].bytesize; + cookie->qat_func_alignsize = qat_func_alignsize; qat_req->pke_hdr.cd_pars.func_id = func_id; qat_req->input_param_count = QAT_ASYM_ECPM_IN_PARAMS; qat_req->output_param_count = QAT_ASYM_ECPM_OUT_PARAMS; - HEXDUMP("k", cookie->input_array[0], qat_alg_bytesize); - HEXDUMP("xG", cookie->input_array[1], qat_alg_bytesize); - HEXDUMP("yG", cookie->input_array[2], qat_alg_bytesize); - HEXDUMP("a", cookie->input_array[3], qat_alg_bytesize); - HEXDUMP("b", cookie->input_array[4], qat_alg_bytesize); - HEXDUMP("q", cookie->input_array[5], qat_alg_bytesize); - HEXDUMP("h", cookie->input_array[6], qat_alg_bytesize); + HEXDUMP("k", cookie->input_array[0], qat_func_alignsize); + HEXDUMP("xG", cookie->input_array[1], qat_func_alignsize); + HEXDUMP("yG", cookie->input_array[2], qat_func_alignsize); + HEXDUMP("a", cookie->input_array[3], qat_func_alignsize); + HEXDUMP("b", cookie->input_array[4], qat_func_alignsize); + HEXDUMP("q", cookie->input_array[5], qat_func_alignsize); + HEXDUMP("h", cookie->input_array[6], qat_func_alignsize); return 0; } @@ -764,8 +732,8 @@ ecpm_collect(struct rte_crypto_asym_op *asym_op, uint8_t *x = asym_op->ecpm.r.x.data; uint8_t *y = asym_op->ecpm.r.y.data; uint32_t alg_bytesize = cookie->alg_bytesize; - uint32_t qat_alg_bytesize = RTE_ALIGN_CEIL(cookie->alg_bytesize, 8); - uint32_t ltrim = qat_alg_bytesize - alg_bytesize; + uint32_t qat_func_alignsize = cookie->qat_func_alignsize; + uint32_t ltrim = qat_func_alignsize - alg_bytesize; asym_op->ecpm.r.x.length = alg_bytesize; asym_op->ecpm.r.y.length = alg_bytesize; @@ -773,9 +741,9 @@ ecpm_collect(struct rte_crypto_asym_op *asym_op, rte_memcpy(y, &cookie->output_array[1][ltrim], alg_bytesize); HEXDUMP("rX", cookie->output_array[0], - alg_bytesize); + qat_func_alignsize); HEXDUMP("rY", cookie->output_array[1], - alg_bytesize); + qat_func_alignsize); return RTE_CRYPTO_OP_STATUS_SUCCESS; } @@ -910,13 +878,13 @@ qat_asym_process_response(void **out_op, uint8_t *resp, cookie->error = 0; if (op->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED) op->status = RTE_CRYPTO_OP_STATUS_ERROR; - QAT_DP_LOG(ERR, "Cookie status returned error"); + QAT_DP_LOG(DEBUG, "Cookie status returned error"); } else { if (ICP_QAT_FW_PKE_RESP_PKE_STAT_GET( resp_msg->pke_resp_hdr.resp_status.pke_resp_flags)) { if (op->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED) op->status = RTE_CRYPTO_OP_STATUS_ERROR; - QAT_DP_LOG(ERR, "Asymmetric response status" + QAT_DP_LOG(DEBUG, "Asymmetric response status" " returned error"); } if (resp_msg->pke_resp_hdr.resp_status.comn_err_code) { diff --git a/drivers/crypto/qat/qat_asym.h b/drivers/crypto/qat/qat_asym.h index b1d403486f..6628656b42 100644 --- a/drivers/crypto/qat/qat_asym.h +++ b/drivers/crypto/qat/qat_asym.h @@ -58,8 +58,9 @@ typedef uint64_t large_int_ptr; } struct qat_asym_op_cookie { - size_t alg_bytesize; uint64_t error; + uint32_t alg_bytesize; /*< Bytesize of algorithm */ + uint32_t qat_func_alignsize; /*< Aligned bytesize of qat function */ rte_iova_t input_addr; rte_iova_t output_addr; large_int_ptr input_params_ptrs[MAX_PKE_PARAMS] _PKE_ALIGN_; diff --git a/drivers/crypto/qat/qat_ec.h b/drivers/crypto/qat/qat_ec.h index a310e3f4d3..1bcd7d1408 100644 --- a/drivers/crypto/qat/qat_ec.h +++ b/drivers/crypto/qat/qat_ec.h @@ -92,7 +92,6 @@ static struct elliptic_curve curve[] = { }, }, }, - [SECP521R1] = { .name = "secp521r1", .bytesize = 66, -- 2.13.6