From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id E458F43C0B; Wed, 28 Feb 2024 12:33:13 +0100 (CET) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 67F9640295; Wed, 28 Feb 2024 12:33:13 +0100 (CET) Received: from mgamail.intel.com (mgamail.intel.com [192.198.163.14]) by mails.dpdk.org (Postfix) with ESMTP id B5AAC4003C for ; Wed, 28 Feb 2024 12:33:10 +0100 (CET) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1709119991; x=1740655991; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=9HiF50nSRX0ZQM2VjToXjQB8/4zU3Q2/es3jsIPhRJc=; b=XzZ/kIhuboJG0XI7qiwKncymhU7nYWvVu/HUxHnwzvgeiezJWyzyjDLA ChEaWSwcRufb4Wh2qa5awlFSwjwaqZnbAvGzfBwDUpfcV+41vt7W0rpfl 776eZYxg8uEWYjU6/z6BlW+C9dp8Bk0KLHNavNDCP5rinkq2mAs4ksqWf MSjxzQcV4FpXwC8NFHZFDUpJH6T4oGW862y65+7pw4annraAoyRj2rero sT32wzjM++oWniuXJ63pqdBso1IQBtxLWellZmt9VEVlD+ACYcsoLxh4b e7k00/xAowIlgO5othX6reMImk4vF/fOd8EP0/k4RdUn1UEXjkqM9gSXj w==; X-IronPort-AV: E=McAfee;i="6600,9927,10996"; a="3676156" X-IronPort-AV: E=Sophos;i="6.06,190,1705392000"; d="scan'208";a="3676156" Received: from fmviesa003.fm.intel.com ([10.60.135.143]) by fmvoesa108.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 28 Feb 2024 03:33:09 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="6.06,190,1705392000"; d="scan'208";a="11984994" Received: from unknown (HELO silpixa00400886.ir.intel.com) ([10.243.23.150]) by fmviesa003.fm.intel.com with ESMTP; 28 Feb 2024 03:33:07 -0800 From: Brian Dooley To: Kai Ji , Pablo de Lara Cc: dev@dpdk.org, gakhil@marvell.com, Brian Dooley Subject: [PATCH v4] crypto/ipsec_mb: unified IPsec MB interface Date: Wed, 28 Feb 2024 11:33:01 +0000 Message-Id: <20240228113301.934291-1-brian.dooley@intel.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20231212153640.1561504-1-brian.dooley@intel.com> References: <20231212153640.1561504-1-brian.dooley@intel.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Currently IPsec MB provides both the JOB API and direct API. AESNI_MB PMD is using the JOB API codepath while ZUC, KASUMI, SNOW3G and CHACHA20_POLY1305 are using the direct API. Instead of using the direct API for these PMDs, they should now make use of the JOB API codepath. This would remove all use of the IPsec MB direct API for these PMDs. Signed-off-by: Brian Dooley --- v2: - Fix compilation failure v3: - Remove session configure pointer for each PMD v4: - Keep AES GCM PMD and fix extern issue --- doc/guides/rel_notes/release_24_03.rst | 6 + drivers/crypto/ipsec_mb/pmd_aesni_mb.c | 10 +- drivers/crypto/ipsec_mb/pmd_aesni_mb_priv.h | 15 +- drivers/crypto/ipsec_mb/pmd_chacha_poly.c | 338 +---------- .../crypto/ipsec_mb/pmd_chacha_poly_priv.h | 28 - drivers/crypto/ipsec_mb/pmd_kasumi.c | 410 +------------ drivers/crypto/ipsec_mb/pmd_kasumi_priv.h | 20 - drivers/crypto/ipsec_mb/pmd_snow3g.c | 543 +----------------- drivers/crypto/ipsec_mb/pmd_snow3g_priv.h | 21 - drivers/crypto/ipsec_mb/pmd_zuc.c | 347 +---------- drivers/crypto/ipsec_mb/pmd_zuc_priv.h | 20 - 11 files changed, 48 insertions(+), 1710 deletions(-) diff --git a/doc/guides/rel_notes/release_24_03.rst b/doc/guides/rel_notes/release_24_03.rst index 879bb4944c..6c5b76cef5 100644 --- a/doc/guides/rel_notes/release_24_03.rst +++ b/doc/guides/rel_notes/release_24_03.rst @@ -138,6 +138,12 @@ New Features to support TLS v1.2, TLS v1.3 and DTLS v1.2. * Added PMD API to allow raw submission of instructions to CPT. +* **Updated ipsec_mb crypto driver.** + + * Kasumi, Snow3G, ChaChaPoly and ZUC PMDs now share the job API codepath + with AESNI_MB PMD. Depending on the architecture, the performance of ZUC + crypto PMD is approximately 10% less for small fixed packet sizes. + Removed Items ------------- diff --git a/drivers/crypto/ipsec_mb/pmd_aesni_mb.c b/drivers/crypto/ipsec_mb/pmd_aesni_mb.c index 4de4866cf3..7d4dbc91ef 100644 --- a/drivers/crypto/ipsec_mb/pmd_aesni_mb.c +++ b/drivers/crypto/ipsec_mb/pmd_aesni_mb.c @@ -8,6 +8,8 @@ RTE_DEFINE_PER_LCORE(pid_t, pid); +uint8_t pmd_driver_id_aesni_mb; + struct aesni_mb_op_buf_data { struct rte_mbuf *m; uint32_t offset; @@ -761,7 +763,7 @@ aesni_mb_set_session_aead_parameters(const IMB_MGR *mb_mgr, } /** Configure a aesni multi-buffer session from a crypto xform chain */ -static int +int aesni_mb_session_configure(IMB_MGR *mb_mgr, void *priv_sess, const struct rte_crypto_sym_xform *xform) @@ -2131,7 +2133,7 @@ set_job_null_op(IMB_JOB *job, struct rte_crypto_op *op) } #if IMB_VERSION(1, 2, 0) < IMB_VERSION_NUM -static uint16_t +uint16_t aesni_mb_dequeue_burst(void *queue_pair, struct rte_crypto_op **ops, uint16_t nb_ops) { @@ -2321,7 +2323,7 @@ flush_mb_mgr(struct ipsec_mb_qp *qp, IMB_MGR *mb_mgr, return processed_ops; } -static uint16_t +uint16_t aesni_mb_dequeue_burst(void *queue_pair, struct rte_crypto_op **ops, uint16_t nb_ops) { @@ -2456,7 +2458,7 @@ verify_sync_dgst(struct rte_crypto_sym_vec *vec, return k; } -static uint32_t +uint32_t aesni_mb_process_bulk(struct rte_cryptodev *dev __rte_unused, struct rte_cryptodev_sym_session *sess, union rte_crypto_sym_ofs sofs, struct rte_crypto_sym_vec *vec) diff --git a/drivers/crypto/ipsec_mb/pmd_aesni_mb_priv.h b/drivers/crypto/ipsec_mb/pmd_aesni_mb_priv.h index 85994fe5a1..2d462a7f68 100644 --- a/drivers/crypto/ipsec_mb/pmd_aesni_mb_priv.h +++ b/drivers/crypto/ipsec_mb/pmd_aesni_mb_priv.h @@ -21,6 +21,19 @@ #define MAX_NUM_SEGS 16 #endif +int +aesni_mb_session_configure(IMB_MGR * m __rte_unused, void *priv_sess, + const struct rte_crypto_sym_xform *xform); + +uint16_t +aesni_mb_dequeue_burst(void *queue_pair, struct rte_crypto_op **ops, + uint16_t nb_ops); + +uint32_t +aesni_mb_process_bulk(struct rte_cryptodev *dev __rte_unused, + struct rte_cryptodev_sym_session *sess, union rte_crypto_sym_ofs sofs, + struct rte_crypto_sym_vec *vec); + static const struct rte_cryptodev_capabilities aesni_mb_capabilities[] = { { /* MD5 HMAC */ .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, @@ -722,8 +735,6 @@ static const struct rte_cryptodev_capabilities aesni_mb_capabilities[] = { RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST() }; -uint8_t pmd_driver_id_aesni_mb; - struct aesni_mb_qp_data { uint8_t temp_digests[IMB_MAX_JOBS][DIGEST_LENGTH_MAX]; /* *< Buffers used to store the digest generated diff --git a/drivers/crypto/ipsec_mb/pmd_chacha_poly.c b/drivers/crypto/ipsec_mb/pmd_chacha_poly.c index 97e7cef233..7436353fc2 100644 --- a/drivers/crypto/ipsec_mb/pmd_chacha_poly.c +++ b/drivers/crypto/ipsec_mb/pmd_chacha_poly.c @@ -3,334 +3,7 @@ */ #include "pmd_chacha_poly_priv.h" - -/** Parse crypto xform chain and set private session parameters. */ -static int -chacha20_poly1305_session_configure(IMB_MGR * mb_mgr __rte_unused, - void *priv_sess, const struct rte_crypto_sym_xform *xform) -{ - struct chacha20_poly1305_session *sess = priv_sess; - const struct rte_crypto_sym_xform *auth_xform; - const struct rte_crypto_sym_xform *cipher_xform; - const struct rte_crypto_sym_xform *aead_xform; - - uint8_t key_length; - const uint8_t *key; - enum ipsec_mb_operation mode; - int ret = 0; - - ret = ipsec_mb_parse_xform(xform, &mode, &auth_xform, - &cipher_xform, &aead_xform); - if (ret) - return ret; - - sess->op = mode; - - switch (sess->op) { - case IPSEC_MB_OP_AEAD_AUTHENTICATED_ENCRYPT: - case IPSEC_MB_OP_AEAD_AUTHENTICATED_DECRYPT: - if (aead_xform->aead.algo != - RTE_CRYPTO_AEAD_CHACHA20_POLY1305) { - IPSEC_MB_LOG(ERR, - "The only combined operation supported is CHACHA20 POLY1305"); - ret = -ENOTSUP; - goto error_exit; - } - /* Set IV parameters */ - sess->iv.offset = aead_xform->aead.iv.offset; - sess->iv.length = aead_xform->aead.iv.length; - key_length = aead_xform->aead.key.length; - key = aead_xform->aead.key.data; - sess->aad_length = aead_xform->aead.aad_length; - sess->req_digest_length = aead_xform->aead.digest_length; - break; - default: - IPSEC_MB_LOG( - ERR, "Wrong xform type, has to be AEAD or authentication"); - ret = -ENOTSUP; - goto error_exit; - } - - /* IV check */ - if (sess->iv.length != CHACHA20_POLY1305_IV_LENGTH && - sess->iv.length != 0) { - IPSEC_MB_LOG(ERR, "Wrong IV length"); - ret = -EINVAL; - goto error_exit; - } - - /* Check key length */ - if (key_length != CHACHA20_POLY1305_KEY_SIZE) { - IPSEC_MB_LOG(ERR, "Invalid key length"); - ret = -EINVAL; - goto error_exit; - } else { - memcpy(sess->key, key, CHACHA20_POLY1305_KEY_SIZE); - } - - /* Digest check */ - if (sess->req_digest_length != CHACHA20_POLY1305_DIGEST_LENGTH) { - IPSEC_MB_LOG(ERR, "Invalid digest length"); - ret = -EINVAL; - goto error_exit; - } else { - sess->gen_digest_length = CHACHA20_POLY1305_DIGEST_LENGTH; - } - -error_exit: - return ret; -} - -/** - * Process a crypto operation, calling - * the direct chacha poly API from the multi buffer library. - * - * @param qp queue pair - * @param op symmetric crypto operation - * @param session chacha poly session - * - * @return - * - Return 0 if success - */ -static int -chacha20_poly1305_crypto_op(struct ipsec_mb_qp *qp, struct rte_crypto_op *op, - struct chacha20_poly1305_session *session) -{ - struct chacha20_poly1305_qp_data *qp_data = - ipsec_mb_get_qp_private_data(qp); - uint8_t *src, *dst; - uint8_t *iv_ptr; - struct rte_crypto_sym_op *sym_op = op->sym; - struct rte_mbuf *m_src = sym_op->m_src; - uint32_t offset, data_offset, data_length; - uint32_t part_len, data_len; - int total_len; - uint8_t *tag; - unsigned int oop = 0; - - offset = sym_op->aead.data.offset; - data_offset = offset; - data_length = sym_op->aead.data.length; - RTE_ASSERT(m_src != NULL); - - while (offset >= m_src->data_len && data_length != 0) { - offset -= m_src->data_len; - m_src = m_src->next; - - RTE_ASSERT(m_src != NULL); - } - - src = rte_pktmbuf_mtod_offset(m_src, uint8_t *, offset); - - data_len = m_src->data_len - offset; - part_len = (data_len < data_length) ? data_len : - data_length; - - /* In-place */ - if (sym_op->m_dst == NULL || (sym_op->m_dst == sym_op->m_src)) - dst = src; - /* Out-of-place */ - else { - oop = 1; - /* Segmented destination buffer is not supported - * if operation is Out-of-place - */ - RTE_ASSERT(rte_pktmbuf_is_contiguous(sym_op->m_dst)); - dst = rte_pktmbuf_mtod_offset(sym_op->m_dst, uint8_t *, - data_offset); - } - - iv_ptr = rte_crypto_op_ctod_offset(op, uint8_t *, - session->iv.offset); - - IMB_CHACHA20_POLY1305_INIT(qp->mb_mgr, session->key, - &qp_data->chacha20_poly1305_ctx_data, - iv_ptr, sym_op->aead.aad.data, - (uint64_t)session->aad_length); - - if (session->op == IPSEC_MB_OP_AEAD_AUTHENTICATED_ENCRYPT) { - IMB_CHACHA20_POLY1305_ENC_UPDATE(qp->mb_mgr, - session->key, - &qp_data->chacha20_poly1305_ctx_data, - dst, src, (uint64_t)part_len); - total_len = data_length - part_len; - - while (total_len) { - m_src = m_src->next; - RTE_ASSERT(m_src != NULL); - - src = rte_pktmbuf_mtod(m_src, uint8_t *); - if (oop) - dst += part_len; - else - dst = src; - part_len = (m_src->data_len < total_len) ? - m_src->data_len : total_len; - - if (dst == NULL || src == NULL) { - IPSEC_MB_LOG(ERR, "Invalid src or dst input"); - return -EINVAL; - } - IMB_CHACHA20_POLY1305_ENC_UPDATE(qp->mb_mgr, - session->key, - &qp_data->chacha20_poly1305_ctx_data, - dst, src, (uint64_t)part_len); - total_len -= part_len; - if (total_len < 0) { - IPSEC_MB_LOG(ERR, "Invalid part len"); - return -EINVAL; - } - } - - tag = sym_op->aead.digest.data; - IMB_CHACHA20_POLY1305_ENC_FINALIZE(qp->mb_mgr, - &qp_data->chacha20_poly1305_ctx_data, - tag, session->gen_digest_length); - - } else { - IMB_CHACHA20_POLY1305_DEC_UPDATE(qp->mb_mgr, - session->key, - &qp_data->chacha20_poly1305_ctx_data, - dst, src, (uint64_t)part_len); - - total_len = data_length - part_len; - - while (total_len) { - m_src = m_src->next; - - RTE_ASSERT(m_src != NULL); - - src = rte_pktmbuf_mtod(m_src, uint8_t *); - if (oop) - dst += part_len; - else - dst = src; - part_len = (m_src->data_len < total_len) ? - m_src->data_len : total_len; - - if (dst == NULL || src == NULL) { - IPSEC_MB_LOG(ERR, "Invalid src or dst input"); - return -EINVAL; - } - IMB_CHACHA20_POLY1305_DEC_UPDATE(qp->mb_mgr, - session->key, - &qp_data->chacha20_poly1305_ctx_data, - dst, src, (uint64_t)part_len); - total_len -= part_len; - if (total_len < 0) { - IPSEC_MB_LOG(ERR, "Invalid part len"); - return -EINVAL; - } - } - - tag = qp_data->temp_digest; - IMB_CHACHA20_POLY1305_DEC_FINALIZE(qp->mb_mgr, - &qp_data->chacha20_poly1305_ctx_data, - tag, session->gen_digest_length); - } - - return 0; -} - -/** - * Process a completed chacha poly op - * - * @param qp Queue Pair to process - * @param op Crypto operation - * @param sess Crypto session - * - * @return - * - void - */ -static void -post_process_chacha20_poly1305_crypto_op(struct ipsec_mb_qp *qp, - struct rte_crypto_op *op, - struct chacha20_poly1305_session *session) -{ - struct chacha20_poly1305_qp_data *qp_data = - ipsec_mb_get_qp_private_data(qp); - - op->status = RTE_CRYPTO_OP_STATUS_SUCCESS; - /* Verify digest if required */ - if (session->op == IPSEC_MB_OP_AEAD_AUTHENTICATED_DECRYPT || - session->op == IPSEC_MB_OP_HASH_VERIFY_ONLY) { - uint8_t *digest = op->sym->aead.digest.data; - uint8_t *tag = qp_data->temp_digest; - -#ifdef RTE_LIBRTE_PMD_CHACHA20_POLY1305_DEBUG - rte_hexdump(stdout, "auth tag (orig):", - digest, session->req_digest_length); - rte_hexdump(stdout, "auth tag (calc):", - tag, session->req_digest_length); -#endif - if (memcmp(tag, digest, session->req_digest_length) != 0) - op->status = RTE_CRYPTO_OP_STATUS_AUTH_FAILED; - - } - -} - -/** - * Process a completed Chacha20_poly1305 request - * - * @param qp Queue Pair to process - * @param op Crypto operation - * @param sess Crypto session - * - * @return - * - void - */ -static void -handle_completed_chacha20_poly1305_crypto_op(struct ipsec_mb_qp *qp, - struct rte_crypto_op *op, - struct chacha20_poly1305_session *sess) -{ - post_process_chacha20_poly1305_crypto_op(qp, op, sess); - - /* Free session if a session-less crypto op */ - if (op->sess_type == RTE_CRYPTO_OP_SESSIONLESS) { - memset(sess, 0, sizeof(struct chacha20_poly1305_session)); - rte_mempool_put(qp->sess_mp, op->sym->session); - op->sym->session = NULL; - } -} - -static uint16_t -chacha20_poly1305_pmd_dequeue_burst(void *queue_pair, - struct rte_crypto_op **ops, uint16_t nb_ops) -{ - struct chacha20_poly1305_session *sess; - struct ipsec_mb_qp *qp = queue_pair; - - int retval = 0; - unsigned int i = 0, nb_dequeued; - - nb_dequeued = rte_ring_dequeue_burst(qp->ingress_queue, - (void **)ops, nb_ops, NULL); - - for (i = 0; i < nb_dequeued; i++) { - - sess = ipsec_mb_get_session_private(qp, ops[i]); - if (unlikely(sess == NULL)) { - ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - qp->stats.dequeue_err_count++; - break; - } - - retval = chacha20_poly1305_crypto_op(qp, ops[i], sess); - if (retval < 0) { - ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - qp->stats.dequeue_err_count++; - break; - } - - handle_completed_chacha20_poly1305_crypto_op(qp, ops[i], sess); - } - - qp->stats.dequeued_count += i; - - return i; -} +#include "pmd_aesni_mb_priv.h" struct rte_cryptodev_ops chacha20_poly1305_pmd_ops = { .dev_configure = ipsec_mb_config, @@ -384,7 +57,7 @@ RTE_INIT(ipsec_mb_register_chacha20_poly1305) = &ipsec_mb_pmds[IPSEC_MB_PMD_TYPE_CHACHA20_POLY1305]; chacha_poly_data->caps = chacha20_poly1305_capabilities; - chacha_poly_data->dequeue_burst = chacha20_poly1305_pmd_dequeue_burst; + chacha_poly_data->dequeue_burst = aesni_mb_dequeue_burst; chacha_poly_data->feature_flags = RTE_CRYPTODEV_FF_SYMMETRIC_CRYPTO | RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING | @@ -395,10 +68,9 @@ RTE_INIT(ipsec_mb_register_chacha20_poly1305) RTE_CRYPTODEV_FF_SYM_SESSIONLESS; chacha_poly_data->internals_priv_size = 0; chacha_poly_data->ops = &chacha20_poly1305_pmd_ops; - chacha_poly_data->qp_priv_size = - sizeof(struct chacha20_poly1305_qp_data); + chacha_poly_data->qp_priv_size = sizeof(struct aesni_mb_qp_data); chacha_poly_data->session_configure = - chacha20_poly1305_session_configure; + aesni_mb_session_configure; chacha_poly_data->session_priv_size = - sizeof(struct chacha20_poly1305_session); + sizeof(struct aesni_mb_session); } diff --git a/drivers/crypto/ipsec_mb/pmd_chacha_poly_priv.h b/drivers/crypto/ipsec_mb/pmd_chacha_poly_priv.h index 842f62f5d1..e668bfe07f 100644 --- a/drivers/crypto/ipsec_mb/pmd_chacha_poly_priv.h +++ b/drivers/crypto/ipsec_mb/pmd_chacha_poly_priv.h @@ -7,9 +7,7 @@ #include "ipsec_mb_private.h" -#define CHACHA20_POLY1305_IV_LENGTH 12 #define CHACHA20_POLY1305_DIGEST_LENGTH 16 -#define CHACHA20_POLY1305_KEY_SIZE 32 static const struct rte_cryptodev_capabilities chacha20_poly1305_capabilities[] = { @@ -45,30 +43,4 @@ struct rte_cryptodev_capabilities chacha20_poly1305_capabilities[] = { uint8_t pmd_driver_id_chacha20_poly1305; -/** CHACHA20 POLY1305 private session structure */ -struct chacha20_poly1305_session { - struct { - uint16_t length; - uint16_t offset; - } iv; - /**< IV parameters */ - uint16_t aad_length; - /**< AAD length */ - uint16_t req_digest_length; - /**< Requested digest length */ - uint16_t gen_digest_length; - /**< Generated digest length */ - uint8_t key[CHACHA20_POLY1305_KEY_SIZE]; - enum ipsec_mb_operation op; -} __rte_cache_aligned; - -struct chacha20_poly1305_qp_data { - struct chacha20_poly1305_context_data chacha20_poly1305_ctx_data; - uint8_t temp_digest[CHACHA20_POLY1305_DIGEST_LENGTH]; - /**< Buffer used to store the digest generated - * by the driver when verifying a digest provided - * by the user (using authentication verify operation) - */ -}; - #endif /* _PMD_CHACHA_POLY_PRIV_H_ */ diff --git a/drivers/crypto/ipsec_mb/pmd_kasumi.c b/drivers/crypto/ipsec_mb/pmd_kasumi.c index 70536ec3dc..c3571ec81b 100644 --- a/drivers/crypto/ipsec_mb/pmd_kasumi.c +++ b/drivers/crypto/ipsec_mb/pmd_kasumi.c @@ -10,406 +10,7 @@ #include #include "pmd_kasumi_priv.h" - -/** Parse crypto xform chain and set private session parameters. */ -static int -kasumi_session_configure(IMB_MGR *mgr, void *priv_sess, - const struct rte_crypto_sym_xform *xform) -{ - const struct rte_crypto_sym_xform *auth_xform = NULL; - const struct rte_crypto_sym_xform *cipher_xform = NULL; - enum ipsec_mb_operation mode; - struct kasumi_session *sess = (struct kasumi_session *)priv_sess; - /* Select Crypto operation - hash then cipher / cipher then hash */ - int ret = ipsec_mb_parse_xform(xform, &mode, &auth_xform, - &cipher_xform, NULL); - - if (ret) - return ret; - - if (cipher_xform) { - /* Only KASUMI F8 supported */ - if (cipher_xform->cipher.algo != RTE_CRYPTO_CIPHER_KASUMI_F8) { - IPSEC_MB_LOG(ERR, "Unsupported cipher algorithm "); - return -ENOTSUP; - } - - sess->cipher_iv_offset = cipher_xform->cipher.iv.offset; - if (cipher_xform->cipher.iv.length != KASUMI_IV_LENGTH) { - IPSEC_MB_LOG(ERR, "Wrong IV length"); - return -EINVAL; - } - - /* Initialize key */ - IMB_KASUMI_INIT_F8_KEY_SCHED(mgr, - cipher_xform->cipher.key.data, - &sess->pKeySched_cipher); - } - - if (auth_xform) { - /* Only KASUMI F9 supported */ - if (auth_xform->auth.algo != RTE_CRYPTO_AUTH_KASUMI_F9) { - IPSEC_MB_LOG(ERR, "Unsupported authentication"); - return -ENOTSUP; - } - - if (auth_xform->auth.digest_length != KASUMI_DIGEST_LENGTH) { - IPSEC_MB_LOG(ERR, "Wrong digest length"); - return -EINVAL; - } - - sess->auth_op = auth_xform->auth.op; - - /* Initialize key */ - IMB_KASUMI_INIT_F9_KEY_SCHED(mgr, auth_xform->auth.key.data, - &sess->pKeySched_hash); - } - - sess->op = mode; - return ret; -} - -/** Encrypt/decrypt mbufs with same cipher key. */ -static uint8_t -process_kasumi_cipher_op(struct ipsec_mb_qp *qp, struct rte_crypto_op **ops, - struct kasumi_session *session, uint8_t num_ops) -{ - unsigned int i; - uint8_t processed_ops = 0; - const void *src[num_ops]; - void *dst[num_ops]; - uint8_t *iv_ptr; - uint64_t iv[num_ops]; - uint32_t num_bytes[num_ops]; - - for (i = 0; i < num_ops; i++) { - src[i] = rte_pktmbuf_mtod_offset(ops[i]->sym->m_src, - uint8_t *, - (ops[i]->sym->cipher.data.offset >> 3)); - dst[i] = ops[i]->sym->m_dst - ? rte_pktmbuf_mtod_offset(ops[i]->sym->m_dst, - uint8_t *, - (ops[i]->sym->cipher.data.offset >> 3)) - : rte_pktmbuf_mtod_offset(ops[i]->sym->m_src, - uint8_t *, - (ops[i]->sym->cipher.data.offset >> 3)); - iv_ptr = rte_crypto_op_ctod_offset(ops[i], uint8_t *, - session->cipher_iv_offset); - iv[i] = *((uint64_t *)(iv_ptr)); - num_bytes[i] = ops[i]->sym->cipher.data.length >> 3; - - processed_ops++; - } - - if (processed_ops != 0) - IMB_KASUMI_F8_N_BUFFER(qp->mb_mgr, &session->pKeySched_cipher, - iv, src, dst, num_bytes, - processed_ops); - - return processed_ops; -} - -/** Encrypt/decrypt mbuf (bit level function). */ -static uint8_t -process_kasumi_cipher_op_bit(struct ipsec_mb_qp *qp, struct rte_crypto_op *op, - struct kasumi_session *session) -{ - uint8_t *src, *dst; - uint8_t *iv_ptr; - uint64_t iv; - uint32_t length_in_bits, offset_in_bits; - - offset_in_bits = op->sym->cipher.data.offset; - src = rte_pktmbuf_mtod(op->sym->m_src, uint8_t *); - if (op->sym->m_dst == NULL) - dst = src; - else - dst = rte_pktmbuf_mtod(op->sym->m_dst, uint8_t *); - iv_ptr = rte_crypto_op_ctod_offset(op, uint8_t *, - session->cipher_iv_offset); - iv = *((uint64_t *)(iv_ptr)); - length_in_bits = op->sym->cipher.data.length; - - IMB_KASUMI_F8_1_BUFFER_BIT(qp->mb_mgr, &session->pKeySched_cipher, iv, - src, dst, length_in_bits, offset_in_bits); - - return 1; -} - -/** Generate/verify hash from mbufs with same hash key. */ -static int -process_kasumi_hash_op(struct ipsec_mb_qp *qp, struct rte_crypto_op **ops, - struct kasumi_session *session, uint8_t num_ops) -{ - unsigned int i; - uint8_t processed_ops = 0; - uint8_t *src, *dst; - uint32_t length_in_bits; - uint32_t num_bytes; - struct kasumi_qp_data *qp_data = ipsec_mb_get_qp_private_data(qp); - - for (i = 0; i < num_ops; i++) { - /* Data must be byte aligned */ - if ((ops[i]->sym->auth.data.offset % BYTE_LEN) != 0) { - ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - IPSEC_MB_LOG(ERR, "Invalid Offset"); - break; - } - - length_in_bits = ops[i]->sym->auth.data.length; - - src = rte_pktmbuf_mtod_offset(ops[i]->sym->m_src, uint8_t *, - (ops[i]->sym->auth.data.offset >> 3)); - /* Direction from next bit after end of message */ - num_bytes = length_in_bits >> 3; - - if (session->auth_op == RTE_CRYPTO_AUTH_OP_VERIFY) { - dst = qp_data->temp_digest; - IMB_KASUMI_F9_1_BUFFER(qp->mb_mgr, - &session->pKeySched_hash, src, - num_bytes, dst); - - /* Verify digest. */ - if (memcmp(dst, ops[i]->sym->auth.digest.data, - KASUMI_DIGEST_LENGTH) - != 0) - ops[i]->status - = RTE_CRYPTO_OP_STATUS_AUTH_FAILED; - } else { - dst = ops[i]->sym->auth.digest.data; - - IMB_KASUMI_F9_1_BUFFER(qp->mb_mgr, - &session->pKeySched_hash, src, - num_bytes, dst); - } - processed_ops++; - } - - return processed_ops; -} - -/** Process a batch of crypto ops which shares the same session. */ -static int -process_ops(struct rte_crypto_op **ops, struct kasumi_session *session, - struct ipsec_mb_qp *qp, uint8_t num_ops) -{ - unsigned int i; - unsigned int processed_ops; - - switch (session->op) { - case IPSEC_MB_OP_ENCRYPT_ONLY: - case IPSEC_MB_OP_DECRYPT_ONLY: - processed_ops - = process_kasumi_cipher_op(qp, ops, session, num_ops); - break; - case IPSEC_MB_OP_HASH_GEN_ONLY: - case IPSEC_MB_OP_HASH_VERIFY_ONLY: - processed_ops - = process_kasumi_hash_op(qp, ops, session, num_ops); - break; - case IPSEC_MB_OP_ENCRYPT_THEN_HASH_GEN: - case IPSEC_MB_OP_DECRYPT_THEN_HASH_VERIFY: - processed_ops - = process_kasumi_cipher_op(qp, ops, session, num_ops); - process_kasumi_hash_op(qp, ops, session, processed_ops); - break; - case IPSEC_MB_OP_HASH_VERIFY_THEN_DECRYPT: - case IPSEC_MB_OP_HASH_GEN_THEN_ENCRYPT: - processed_ops - = process_kasumi_hash_op(qp, ops, session, num_ops); - process_kasumi_cipher_op(qp, ops, session, processed_ops); - break; - default: - /* Operation not supported. */ - processed_ops = 0; - } - - for (i = 0; i < num_ops; i++) { - /* - * If there was no error/authentication failure, - * change status to successful. - */ - if (ops[i]->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED) - ops[i]->status = RTE_CRYPTO_OP_STATUS_SUCCESS; - /* Free session if a session-less crypto op. */ - if (ops[i]->sess_type == RTE_CRYPTO_OP_SESSIONLESS) { - memset(session, 0, sizeof(struct kasumi_session)); - rte_mempool_put(qp->sess_mp, ops[i]->sym->session); - ops[i]->sym->session = NULL; - } - } - return processed_ops; -} - -/** Process a crypto op with length/offset in bits. */ -static int -process_op_bit(struct rte_crypto_op *op, struct kasumi_session *session, - struct ipsec_mb_qp *qp) -{ - unsigned int processed_op; - - switch (session->op) { - /* case KASUMI_OP_ONLY_CIPHER: */ - case IPSEC_MB_OP_ENCRYPT_ONLY: - case IPSEC_MB_OP_DECRYPT_ONLY: - processed_op = process_kasumi_cipher_op_bit(qp, op, session); - break; - /* case KASUMI_OP_ONLY_AUTH: */ - case IPSEC_MB_OP_HASH_GEN_ONLY: - case IPSEC_MB_OP_HASH_VERIFY_ONLY: - processed_op = process_kasumi_hash_op(qp, &op, session, 1); - break; - /* case KASUMI_OP_CIPHER_AUTH: */ - case IPSEC_MB_OP_ENCRYPT_THEN_HASH_GEN: - processed_op = process_kasumi_cipher_op_bit(qp, op, session); - if (processed_op == 1) - process_kasumi_hash_op(qp, &op, session, 1); - break; - /* case KASUMI_OP_AUTH_CIPHER: */ - case IPSEC_MB_OP_HASH_VERIFY_THEN_DECRYPT: - processed_op = process_kasumi_hash_op(qp, &op, session, 1); - if (processed_op == 1) - process_kasumi_cipher_op_bit(qp, op, session); - break; - default: - /* Operation not supported. */ - processed_op = 0; - } - - /* - * If there was no error/authentication failure, - * change status to successful. - */ - if (op->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED) - op->status = RTE_CRYPTO_OP_STATUS_SUCCESS; - - /* Free session if a session-less crypto op. */ - if (op->sess_type == RTE_CRYPTO_OP_SESSIONLESS) { - memset(CRYPTODEV_GET_SYM_SESS_PRIV(op->sym->session), 0, - sizeof(struct kasumi_session)); - rte_mempool_put(qp->sess_mp, (void *)op->sym->session); - op->sym->session = NULL; - } - return processed_op; -} - -static uint16_t -kasumi_pmd_dequeue_burst(void *queue_pair, struct rte_crypto_op **ops, - uint16_t nb_ops) -{ - struct rte_crypto_op *c_ops[nb_ops]; - struct rte_crypto_op *curr_c_op = NULL; - - struct kasumi_session *prev_sess = NULL, *curr_sess = NULL; - struct ipsec_mb_qp *qp = queue_pair; - unsigned int i; - uint8_t burst_size = 0; - uint8_t processed_ops; - unsigned int nb_dequeued; - - nb_dequeued = rte_ring_dequeue_burst(qp->ingress_queue, - (void **)ops, nb_ops, NULL); - for (i = 0; i < nb_dequeued; i++) { - curr_c_op = ops[i]; - -#ifdef RTE_LIBRTE_PMD_KASUMI_DEBUG - if (!rte_pktmbuf_is_contiguous(curr_c_op->sym->m_src) - || (curr_c_op->sym->m_dst != NULL - && !rte_pktmbuf_is_contiguous( - curr_c_op->sym->m_dst))) { - IPSEC_MB_LOG(ERR, - "PMD supports only contiguous mbufs, op (%p) provides noncontiguous mbuf as source/destination buffer.", - curr_c_op); - curr_c_op->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - break; - } -#endif - - /* Set status as enqueued (not processed yet) by default. */ - curr_c_op->status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED; - - curr_sess = (struct kasumi_session *) - ipsec_mb_get_session_private(qp, curr_c_op); - if (unlikely(curr_sess == NULL - || curr_sess->op == IPSEC_MB_OP_NOT_SUPPORTED)) { - curr_c_op->status - = RTE_CRYPTO_OP_STATUS_INVALID_SESSION; - break; - } - - /* If length/offset is at bit-level, process this buffer alone. - */ - if (((curr_c_op->sym->cipher.data.length % BYTE_LEN) != 0) - || ((ops[i]->sym->cipher.data.offset % BYTE_LEN) != 0)) { - /* Process the ops of the previous session. */ - if (prev_sess != NULL) { - processed_ops = process_ops(c_ops, prev_sess, - qp, burst_size); - if (processed_ops < burst_size) { - burst_size = 0; - break; - } - - burst_size = 0; - prev_sess = NULL; - } - - processed_ops = process_op_bit(curr_c_op, - curr_sess, qp); - if (processed_ops != 1) - break; - - continue; - } - - /* Batch ops that share the same session. */ - if (prev_sess == NULL) { - prev_sess = curr_sess; - c_ops[burst_size++] = curr_c_op; - } else if (curr_sess == prev_sess) { - c_ops[burst_size++] = curr_c_op; - /* - * When there are enough ops to process in a batch, - * process them, and start a new batch. - */ - if (burst_size == KASUMI_MAX_BURST) { - processed_ops = process_ops(c_ops, prev_sess, - qp, burst_size); - if (processed_ops < burst_size) { - burst_size = 0; - break; - } - - burst_size = 0; - prev_sess = NULL; - } - } else { - /* - * Different session, process the ops - * of the previous session. - */ - processed_ops = process_ops(c_ops, prev_sess, qp, - burst_size); - if (processed_ops < burst_size) { - burst_size = 0; - break; - } - - burst_size = 0; - prev_sess = curr_sess; - - c_ops[burst_size++] = curr_c_op; - } - } - - if (burst_size != 0) { - /* Process the crypto ops of the last session. */ - processed_ops = process_ops(c_ops, prev_sess, qp, burst_size); - } - - qp->stats.dequeued_count += i; - return i; -} +#include "pmd_aesni_mb_priv.h" struct rte_cryptodev_ops kasumi_pmd_ops = { .dev_configure = ipsec_mb_config, @@ -460,7 +61,7 @@ RTE_INIT(ipsec_mb_register_kasumi) = &ipsec_mb_pmds[IPSEC_MB_PMD_TYPE_KASUMI]; kasumi_data->caps = kasumi_capabilities; - kasumi_data->dequeue_burst = kasumi_pmd_dequeue_burst; + kasumi_data->dequeue_burst = aesni_mb_dequeue_burst; kasumi_data->feature_flags = RTE_CRYPTODEV_FF_SYMMETRIC_CRYPTO | RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING | RTE_CRYPTODEV_FF_NON_BYTE_ALIGNED_DATA @@ -469,7 +70,8 @@ RTE_INIT(ipsec_mb_register_kasumi) | RTE_CRYPTODEV_FF_OOP_LB_IN_LB_OUT; kasumi_data->internals_priv_size = 0; kasumi_data->ops = &kasumi_pmd_ops; - kasumi_data->qp_priv_size = sizeof(struct kasumi_qp_data); - kasumi_data->session_configure = kasumi_session_configure; - kasumi_data->session_priv_size = sizeof(struct kasumi_session); + kasumi_data->qp_priv_size = sizeof(struct aesni_mb_qp_data); + kasumi_data->session_configure = aesni_mb_session_configure; + kasumi_data->session_priv_size = + sizeof(struct aesni_mb_session); } diff --git a/drivers/crypto/ipsec_mb/pmd_kasumi_priv.h b/drivers/crypto/ipsec_mb/pmd_kasumi_priv.h index 8db1d1cc5b..3223cf1a14 100644 --- a/drivers/crypto/ipsec_mb/pmd_kasumi_priv.h +++ b/drivers/crypto/ipsec_mb/pmd_kasumi_priv.h @@ -9,8 +9,6 @@ #define KASUMI_KEY_LENGTH 16 #define KASUMI_IV_LENGTH 8 -#define KASUMI_MAX_BURST 4 -#define BYTE_LEN 8 #define KASUMI_DIGEST_LENGTH 4 uint8_t pmd_driver_id_kasumi; @@ -60,22 +58,4 @@ static const struct rte_cryptodev_capabilities kasumi_capabilities[] = { RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST() }; -/** KASUMI private session structure */ -struct kasumi_session { - /* Keys have to be 16-byte aligned */ - kasumi_key_sched_t pKeySched_cipher; - kasumi_key_sched_t pKeySched_hash; - enum ipsec_mb_operation op; - enum rte_crypto_auth_operation auth_op; - uint16_t cipher_iv_offset; -} __rte_cache_aligned; - -struct kasumi_qp_data { - uint8_t temp_digest[KASUMI_DIGEST_LENGTH]; - /* *< Buffers used to store the digest generated - * by the driver when verifying a digest provided - * by the user (using authentication verify operation) - */ -}; - #endif /* _PMD_KASUMI_PRIV_H_ */ diff --git a/drivers/crypto/ipsec_mb/pmd_snow3g.c b/drivers/crypto/ipsec_mb/pmd_snow3g.c index a96779f059..957f6aade8 100644 --- a/drivers/crypto/ipsec_mb/pmd_snow3g.c +++ b/drivers/crypto/ipsec_mb/pmd_snow3g.c @@ -3,539 +3,7 @@ */ #include "pmd_snow3g_priv.h" - -/** Parse crypto xform chain and set private session parameters. */ -static int -snow3g_session_configure(IMB_MGR *mgr, void *priv_sess, - const struct rte_crypto_sym_xform *xform) -{ - struct snow3g_session *sess = (struct snow3g_session *)priv_sess; - const struct rte_crypto_sym_xform *auth_xform = NULL; - const struct rte_crypto_sym_xform *cipher_xform = NULL; - enum ipsec_mb_operation mode; - - /* Select Crypto operation - hash then cipher / cipher then hash */ - int ret = ipsec_mb_parse_xform(xform, &mode, &auth_xform, - &cipher_xform, NULL); - if (ret) - return ret; - - if (cipher_xform) { - /* Only SNOW 3G UEA2 supported */ - if (cipher_xform->cipher.algo != RTE_CRYPTO_CIPHER_SNOW3G_UEA2) - return -ENOTSUP; - - if (cipher_xform->cipher.iv.length != SNOW3G_IV_LENGTH) { - IPSEC_MB_LOG(ERR, "Wrong IV length"); - return -EINVAL; - } - if (cipher_xform->cipher.key.length > SNOW3G_MAX_KEY_SIZE) { - IPSEC_MB_LOG(ERR, "Not enough memory to store the key"); - return -ENOMEM; - } - - sess->cipher_iv_offset = cipher_xform->cipher.iv.offset; - - /* Initialize key */ - IMB_SNOW3G_INIT_KEY_SCHED(mgr, cipher_xform->cipher.key.data, - &sess->pKeySched_cipher); - } - - if (auth_xform) { - /* Only SNOW 3G UIA2 supported */ - if (auth_xform->auth.algo != RTE_CRYPTO_AUTH_SNOW3G_UIA2) - return -ENOTSUP; - - if (auth_xform->auth.digest_length != SNOW3G_DIGEST_LENGTH) { - IPSEC_MB_LOG(ERR, "Wrong digest length"); - return -EINVAL; - } - if (auth_xform->auth.key.length > SNOW3G_MAX_KEY_SIZE) { - IPSEC_MB_LOG(ERR, "Not enough memory to store the key"); - return -ENOMEM; - } - - sess->auth_op = auth_xform->auth.op; - - if (auth_xform->auth.iv.length != SNOW3G_IV_LENGTH) { - IPSEC_MB_LOG(ERR, "Wrong IV length"); - return -EINVAL; - } - sess->auth_iv_offset = auth_xform->auth.iv.offset; - - /* Initialize key */ - IMB_SNOW3G_INIT_KEY_SCHED(mgr, auth_xform->auth.key.data, - &sess->pKeySched_hash); - } - - sess->op = mode; - - return 0; -} - -/** Check if conditions are met for digest-appended operations */ -static uint8_t * -snow3g_digest_appended_in_src(struct rte_crypto_op *op) -{ - unsigned int auth_size, cipher_size; - - auth_size = (op->sym->auth.data.offset >> 3) + - (op->sym->auth.data.length >> 3); - cipher_size = (op->sym->cipher.data.offset >> 3) + - (op->sym->cipher.data.length >> 3); - - if (auth_size < cipher_size) - return rte_pktmbuf_mtod_offset(op->sym->m_src, - uint8_t *, auth_size); - - return NULL; -} - -/** Encrypt/decrypt mbufs with same cipher key. */ -static uint8_t -process_snow3g_cipher_op(struct ipsec_mb_qp *qp, struct rte_crypto_op **ops, - struct snow3g_session *session, - uint8_t num_ops) -{ - uint32_t i; - uint8_t processed_ops = 0; - const void *src[SNOW3G_MAX_BURST] = {NULL}; - void *dst[SNOW3G_MAX_BURST] = {NULL}; - uint8_t *digest_appended[SNOW3G_MAX_BURST] = {NULL}; - const void *iv[SNOW3G_MAX_BURST] = {NULL}; - uint32_t num_bytes[SNOW3G_MAX_BURST] = {0}; - uint32_t cipher_off, cipher_len; - int unencrypted_bytes = 0; - - for (i = 0; i < num_ops; i++) { - - cipher_off = ops[i]->sym->cipher.data.offset >> 3; - cipher_len = ops[i]->sym->cipher.data.length >> 3; - src[i] = rte_pktmbuf_mtod_offset( - ops[i]->sym->m_src, uint8_t *, cipher_off); - - /* If out-of-place operation */ - if (ops[i]->sym->m_dst && - ops[i]->sym->m_src != ops[i]->sym->m_dst) { - dst[i] = rte_pktmbuf_mtod_offset( - ops[i]->sym->m_dst, uint8_t *, cipher_off); - - /* In case of out-of-place, auth-cipher operation - * with partial encryption of the digest, copy - * the remaining, unencrypted part. - */ - if (session->op == IPSEC_MB_OP_HASH_VERIFY_THEN_DECRYPT - || session->op == IPSEC_MB_OP_HASH_GEN_THEN_ENCRYPT) - unencrypted_bytes = - (ops[i]->sym->auth.data.offset >> 3) + - (ops[i]->sym->auth.data.length >> 3) + - (SNOW3G_DIGEST_LENGTH) - - cipher_off - cipher_len; - if (unencrypted_bytes > 0) - rte_memcpy( - rte_pktmbuf_mtod_offset( - ops[i]->sym->m_dst, uint8_t *, - cipher_off + cipher_len), - rte_pktmbuf_mtod_offset( - ops[i]->sym->m_src, uint8_t *, - cipher_off + cipher_len), - unencrypted_bytes); - } else - dst[i] = rte_pktmbuf_mtod_offset(ops[i]->sym->m_src, - uint8_t *, cipher_off); - - iv[i] = rte_crypto_op_ctod_offset(ops[i], uint8_t *, - session->cipher_iv_offset); - num_bytes[i] = cipher_len; - processed_ops++; - } - - IMB_SNOW3G_F8_N_BUFFER(qp->mb_mgr, &session->pKeySched_cipher, iv, - src, dst, num_bytes, processed_ops); - - /* Take care of the raw digest data in src buffer */ - for (i = 0; i < num_ops; i++) { - if ((session->op == IPSEC_MB_OP_HASH_VERIFY_THEN_DECRYPT || - session->op == IPSEC_MB_OP_HASH_GEN_THEN_ENCRYPT) && - ops[i]->sym->m_dst != NULL) { - digest_appended[i] = - snow3g_digest_appended_in_src(ops[i]); - /* Clear unencrypted digest from - * the src buffer - */ - if (digest_appended[i] != NULL) - memset(digest_appended[i], - 0, SNOW3G_DIGEST_LENGTH); - } - } - return processed_ops; -} - -/** Encrypt/decrypt mbuf (bit level function). */ -static uint8_t -process_snow3g_cipher_op_bit(struct ipsec_mb_qp *qp, - struct rte_crypto_op *op, - struct snow3g_session *session) -{ - uint8_t *src, *dst; - uint8_t *iv; - uint32_t length_in_bits, offset_in_bits; - int unencrypted_bytes = 0; - - offset_in_bits = op->sym->cipher.data.offset; - src = rte_pktmbuf_mtod(op->sym->m_src, uint8_t *); - if (op->sym->m_dst == NULL) { - op->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - IPSEC_MB_LOG(ERR, "bit-level in-place not supported\n"); - return 0; - } - length_in_bits = op->sym->cipher.data.length; - dst = rte_pktmbuf_mtod(op->sym->m_dst, uint8_t *); - /* In case of out-of-place, auth-cipher operation - * with partial encryption of the digest, copy - * the remaining, unencrypted part. - */ - if (session->op == IPSEC_MB_OP_HASH_VERIFY_THEN_DECRYPT || - session->op == IPSEC_MB_OP_HASH_GEN_THEN_ENCRYPT) - unencrypted_bytes = - (op->sym->auth.data.offset >> 3) + - (op->sym->auth.data.length >> 3) + - (SNOW3G_DIGEST_LENGTH) - - (offset_in_bits >> 3) - - (length_in_bits >> 3); - if (unencrypted_bytes > 0) - rte_memcpy( - rte_pktmbuf_mtod_offset( - op->sym->m_dst, uint8_t *, - (length_in_bits >> 3)), - rte_pktmbuf_mtod_offset( - op->sym->m_src, uint8_t *, - (length_in_bits >> 3)), - unencrypted_bytes); - - iv = rte_crypto_op_ctod_offset(op, uint8_t *, - session->cipher_iv_offset); - - IMB_SNOW3G_F8_1_BUFFER_BIT(qp->mb_mgr, &session->pKeySched_cipher, iv, - src, dst, length_in_bits, offset_in_bits); - - return 1; -} - -/** Generate/verify hash from mbufs with same hash key. */ -static int -process_snow3g_hash_op(struct ipsec_mb_qp *qp, struct rte_crypto_op **ops, - struct snow3g_session *session, - uint8_t num_ops) -{ - uint32_t i; - uint8_t processed_ops = 0; - uint8_t *src, *dst; - uint32_t length_in_bits; - uint8_t *iv; - uint8_t digest_appended = 0; - struct snow3g_qp_data *qp_data = ipsec_mb_get_qp_private_data(qp); - - for (i = 0; i < num_ops; i++) { - /* Data must be byte aligned */ - if ((ops[i]->sym->auth.data.offset % BYTE_LEN) != 0) { - ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - IPSEC_MB_LOG(ERR, "Offset"); - break; - } - - dst = NULL; - - length_in_bits = ops[i]->sym->auth.data.length; - - src = rte_pktmbuf_mtod_offset(ops[i]->sym->m_src, uint8_t *, - (ops[i]->sym->auth.data.offset >> 3)); - iv = rte_crypto_op_ctod_offset(ops[i], uint8_t *, - session->auth_iv_offset); - - if (session->auth_op == RTE_CRYPTO_AUTH_OP_VERIFY) { - dst = qp_data->temp_digest; - /* Handle auth cipher verify oop case*/ - if ((session->op == - IPSEC_MB_OP_ENCRYPT_THEN_HASH_GEN || - session->op == - IPSEC_MB_OP_DECRYPT_THEN_HASH_VERIFY) && - ops[i]->sym->m_dst != NULL) - src = rte_pktmbuf_mtod_offset( - ops[i]->sym->m_dst, uint8_t *, - ops[i]->sym->auth.data.offset >> 3); - - IMB_SNOW3G_F9_1_BUFFER(qp->mb_mgr, - &session->pKeySched_hash, - iv, src, length_in_bits, dst); - /* Verify digest. */ - if (memcmp(dst, ops[i]->sym->auth.digest.data, - SNOW3G_DIGEST_LENGTH) != 0) - ops[i]->status = - RTE_CRYPTO_OP_STATUS_AUTH_FAILED; - } else { - if (session->op == - IPSEC_MB_OP_HASH_VERIFY_THEN_DECRYPT || - session->op == - IPSEC_MB_OP_HASH_GEN_THEN_ENCRYPT) - dst = snow3g_digest_appended_in_src(ops[i]); - - if (dst != NULL) - digest_appended = 1; - else - dst = ops[i]->sym->auth.digest.data; - - IMB_SNOW3G_F9_1_BUFFER(qp->mb_mgr, - &session->pKeySched_hash, - iv, src, length_in_bits, dst); - - /* Copy back digest from src to auth.digest.data */ - if (digest_appended) - rte_memcpy(ops[i]->sym->auth.digest.data, - dst, SNOW3G_DIGEST_LENGTH); - } - processed_ops++; - } - - return processed_ops; -} - -/** Process a batch of crypto ops which shares the same session. */ -static int -process_ops(struct rte_crypto_op **ops, struct snow3g_session *session, - struct ipsec_mb_qp *qp, uint8_t num_ops) -{ - uint32_t i; - uint32_t processed_ops; - -#ifdef RTE_LIBRTE_PMD_SNOW3G_DEBUG - for (i = 0; i < num_ops; i++) { - if (!rte_pktmbuf_is_contiguous(ops[i]->sym->m_src) || - (ops[i]->sym->m_dst != NULL && - !rte_pktmbuf_is_contiguous( - ops[i]->sym->m_dst))) { - IPSEC_MB_LOG(ERR, - "PMD supports only contiguous mbufs, " - "op (%p) provides noncontiguous mbuf as " - "source/destination buffer.\n", ops[i]); - ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - return 0; - } - } -#endif - - switch (session->op) { - case IPSEC_MB_OP_ENCRYPT_ONLY: - case IPSEC_MB_OP_DECRYPT_ONLY: - processed_ops = process_snow3g_cipher_op(qp, ops, - session, num_ops); - break; - case IPSEC_MB_OP_HASH_GEN_ONLY: - case IPSEC_MB_OP_HASH_VERIFY_ONLY: - processed_ops = process_snow3g_hash_op(qp, ops, session, - num_ops); - break; - case IPSEC_MB_OP_ENCRYPT_THEN_HASH_GEN: - case IPSEC_MB_OP_DECRYPT_THEN_HASH_VERIFY: - processed_ops = process_snow3g_cipher_op(qp, ops, session, - num_ops); - process_snow3g_hash_op(qp, ops, session, processed_ops); - break; - case IPSEC_MB_OP_HASH_VERIFY_THEN_DECRYPT: - case IPSEC_MB_OP_HASH_GEN_THEN_ENCRYPT: - processed_ops = process_snow3g_hash_op(qp, ops, session, - num_ops); - process_snow3g_cipher_op(qp, ops, session, processed_ops); - break; - default: - /* Operation not supported. */ - processed_ops = 0; - } - - for (i = 0; i < num_ops; i++) { - /* - * If there was no error/authentication failure, - * change status to successful. - */ - if (ops[i]->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED) - ops[i]->status = RTE_CRYPTO_OP_STATUS_SUCCESS; - /* Free session if a session-less crypto op. */ - if (ops[i]->sess_type == RTE_CRYPTO_OP_SESSIONLESS) { - memset(session, 0, sizeof(struct snow3g_session)); - rte_mempool_put(qp->sess_mp, ops[i]->sym->session); - ops[i]->sym->session = NULL; - } - } - return processed_ops; -} - -/** Process a crypto op with length/offset in bits. */ -static int -process_op_bit(struct rte_crypto_op *op, struct snow3g_session *session, - struct ipsec_mb_qp *qp) -{ - unsigned int processed_op; - int ret; - - switch (session->op) { - case IPSEC_MB_OP_ENCRYPT_ONLY: - case IPSEC_MB_OP_DECRYPT_ONLY: - - processed_op = process_snow3g_cipher_op_bit(qp, op, - session); - break; - case IPSEC_MB_OP_HASH_GEN_ONLY: - case IPSEC_MB_OP_HASH_VERIFY_ONLY: - processed_op = process_snow3g_hash_op(qp, &op, session, 1); - break; - case IPSEC_MB_OP_ENCRYPT_THEN_HASH_GEN: - case IPSEC_MB_OP_DECRYPT_THEN_HASH_VERIFY: - processed_op = process_snow3g_cipher_op_bit(qp, op, session); - if (processed_op == 1) - process_snow3g_hash_op(qp, &op, session, 1); - break; - case IPSEC_MB_OP_HASH_VERIFY_THEN_DECRYPT: - case IPSEC_MB_OP_HASH_GEN_THEN_ENCRYPT: - processed_op = process_snow3g_hash_op(qp, &op, session, 1); - if (processed_op == 1) - process_snow3g_cipher_op_bit(qp, op, session); - break; - default: - /* Operation not supported. */ - processed_op = 0; - } - - /* - * If there was no error/authentication failure, - * change status to successful. - */ - if (op->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED) - op->status = RTE_CRYPTO_OP_STATUS_SUCCESS; - - /* Free session if a session-less crypto op. */ - if (op->sess_type == RTE_CRYPTO_OP_SESSIONLESS) { - memset(CRYPTODEV_GET_SYM_SESS_PRIV(op->sym->session), 0, - sizeof(struct snow3g_session)); - rte_mempool_put(qp->sess_mp, (void *)op->sym->session); - op->sym->session = NULL; - } - - if (unlikely(processed_op != 1)) - return 0; - - ret = rte_ring_enqueue(qp->ingress_queue, op); - if (ret != 0) - return ret; - - return 1; -} - -static uint16_t -snow3g_pmd_dequeue_burst(void *queue_pair, - struct rte_crypto_op **ops, uint16_t nb_ops) -{ - struct ipsec_mb_qp *qp = queue_pair; - struct rte_crypto_op *c_ops[SNOW3G_MAX_BURST]; - struct rte_crypto_op *curr_c_op; - - struct snow3g_session *prev_sess = NULL, *curr_sess = NULL; - uint32_t i; - uint8_t burst_size = 0; - uint8_t processed_ops; - uint32_t nb_dequeued; - - nb_dequeued = rte_ring_dequeue_burst(qp->ingress_queue, - (void **)ops, nb_ops, NULL); - - for (i = 0; i < nb_dequeued; i++) { - curr_c_op = ops[i]; - - /* Set status as enqueued (not processed yet) by default. */ - curr_c_op->status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED; - - curr_sess = ipsec_mb_get_session_private(qp, curr_c_op); - if (unlikely(curr_sess == NULL || - curr_sess->op == IPSEC_MB_OP_NOT_SUPPORTED)) { - curr_c_op->status = - RTE_CRYPTO_OP_STATUS_INVALID_SESSION; - break; - } - - /* If length/offset is at bit-level, - * process this buffer alone. - */ - if (((curr_c_op->sym->cipher.data.length % BYTE_LEN) != 0) - || ((curr_c_op->sym->cipher.data.offset - % BYTE_LEN) != 0)) { - /* Process the ops of the previous session. */ - if (prev_sess != NULL) { - processed_ops = process_ops(c_ops, prev_sess, - qp, burst_size); - if (processed_ops < burst_size) { - burst_size = 0; - break; - } - - burst_size = 0; - prev_sess = NULL; - } - - processed_ops = process_op_bit(curr_c_op, curr_sess, qp); - if (processed_ops != 1) - break; - - continue; - } - - /* Batch ops that share the same session. */ - if (prev_sess == NULL) { - prev_sess = curr_sess; - c_ops[burst_size++] = curr_c_op; - } else if (curr_sess == prev_sess) { - c_ops[burst_size++] = curr_c_op; - /* - * When there are enough ops to process in a batch, - * process them, and start a new batch. - */ - if (burst_size == SNOW3G_MAX_BURST) { - processed_ops = process_ops(c_ops, prev_sess, - qp, burst_size); - if (processed_ops < burst_size) { - burst_size = 0; - break; - } - - burst_size = 0; - prev_sess = NULL; - } - } else { - /* - * Different session, process the ops - * of the previous session. - */ - processed_ops = process_ops(c_ops, prev_sess, - qp, burst_size); - if (processed_ops < burst_size) { - burst_size = 0; - break; - } - - burst_size = 0; - prev_sess = curr_sess; - - c_ops[burst_size++] = curr_c_op; - } - } - - if (burst_size != 0) { - /* Process the crypto ops of the last session. */ - processed_ops = process_ops(c_ops, prev_sess, - qp, burst_size); - } - - qp->stats.dequeued_count += i; - return i; -} +#include "pmd_aesni_mb_priv.h" struct rte_cryptodev_ops snow3g_pmd_ops = { .dev_configure = ipsec_mb_config, @@ -586,7 +54,7 @@ RTE_INIT(ipsec_mb_register_snow3g) = &ipsec_mb_pmds[IPSEC_MB_PMD_TYPE_SNOW3G]; snow3g_data->caps = snow3g_capabilities; - snow3g_data->dequeue_burst = snow3g_pmd_dequeue_burst; + snow3g_data->dequeue_burst = aesni_mb_dequeue_burst; snow3g_data->feature_flags = RTE_CRYPTODEV_FF_SYMMETRIC_CRYPTO | RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING | RTE_CRYPTODEV_FF_NON_BYTE_ALIGNED_DATA | @@ -595,7 +63,8 @@ RTE_INIT(ipsec_mb_register_snow3g) RTE_CRYPTODEV_FF_DIGEST_ENCRYPTED; snow3g_data->internals_priv_size = 0; snow3g_data->ops = &snow3g_pmd_ops; - snow3g_data->qp_priv_size = sizeof(struct snow3g_qp_data); - snow3g_data->session_configure = snow3g_session_configure; - snow3g_data->session_priv_size = sizeof(struct snow3g_session); + snow3g_data->qp_priv_size = sizeof(struct aesni_mb_qp_data); + snow3g_data->session_configure = aesni_mb_session_configure; + snow3g_data->session_priv_size = + sizeof(struct aesni_mb_session); } diff --git a/drivers/crypto/ipsec_mb/pmd_snow3g_priv.h b/drivers/crypto/ipsec_mb/pmd_snow3g_priv.h index ca1ce7f9d6..3ceb33b602 100644 --- a/drivers/crypto/ipsec_mb/pmd_snow3g_priv.h +++ b/drivers/crypto/ipsec_mb/pmd_snow3g_priv.h @@ -8,10 +8,7 @@ #include "ipsec_mb_private.h" #define SNOW3G_IV_LENGTH 16 -#define SNOW3G_MAX_BURST 8 -#define BYTE_LEN 8 #define SNOW3G_DIGEST_LENGTH 4 -#define SNOW3G_MAX_KEY_SIZE 128 uint8_t pmd_driver_id_snow3g; @@ -64,22 +61,4 @@ static const struct rte_cryptodev_capabilities snow3g_capabilities[] = { RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST() }; -/** SNOW 3G private session structure */ -struct snow3g_session { - enum ipsec_mb_operation op; - enum rte_crypto_auth_operation auth_op; - snow3g_key_schedule_t pKeySched_cipher; - snow3g_key_schedule_t pKeySched_hash; - uint16_t cipher_iv_offset; - uint16_t auth_iv_offset; -} __rte_cache_aligned; - -struct snow3g_qp_data { - uint8_t temp_digest[SNOW3G_DIGEST_LENGTH]; - /**< Buffer used to store the digest generated - * by the driver when verifying a digest provided - * by the user (using authentication verify operation) - */ -}; - #endif /* _PMD_SNOW3G_PRIV_H_ */ diff --git a/drivers/crypto/ipsec_mb/pmd_zuc.c b/drivers/crypto/ipsec_mb/pmd_zuc.c index 44781be1d1..b72191c7a7 100644 --- a/drivers/crypto/ipsec_mb/pmd_zuc.c +++ b/drivers/crypto/ipsec_mb/pmd_zuc.c @@ -3,343 +3,7 @@ */ #include "pmd_zuc_priv.h" - -/** Parse crypto xform chain and set private session parameters. */ -static int -zuc_session_configure(__rte_unused IMB_MGR * mgr, void *zuc_sess, - const struct rte_crypto_sym_xform *xform) -{ - struct zuc_session *sess = (struct zuc_session *) zuc_sess; - const struct rte_crypto_sym_xform *auth_xform = NULL; - const struct rte_crypto_sym_xform *cipher_xform = NULL; - enum ipsec_mb_operation mode; - /* Select Crypto operation - hash then cipher / cipher then hash */ - int ret = ipsec_mb_parse_xform(xform, &mode, &auth_xform, - &cipher_xform, NULL); - - if (ret) - return ret; - - if (cipher_xform) { - /* Only ZUC EEA3 supported */ - if (cipher_xform->cipher.algo != RTE_CRYPTO_CIPHER_ZUC_EEA3) - return -ENOTSUP; - - if (cipher_xform->cipher.iv.length != ZUC_IV_KEY_LENGTH) { - IPSEC_MB_LOG(ERR, "Wrong IV length"); - return -EINVAL; - } - sess->cipher_iv_offset = cipher_xform->cipher.iv.offset; - - /* Copy the key */ - memcpy(sess->pKey_cipher, cipher_xform->cipher.key.data, - ZUC_IV_KEY_LENGTH); - } - - if (auth_xform) { - /* Only ZUC EIA3 supported */ - if (auth_xform->auth.algo != RTE_CRYPTO_AUTH_ZUC_EIA3) - return -ENOTSUP; - - if (auth_xform->auth.digest_length != ZUC_DIGEST_LENGTH) { - IPSEC_MB_LOG(ERR, "Wrong digest length"); - return -EINVAL; - } - - sess->auth_op = auth_xform->auth.op; - - if (auth_xform->auth.iv.length != ZUC_IV_KEY_LENGTH) { - IPSEC_MB_LOG(ERR, "Wrong IV length"); - return -EINVAL; - } - sess->auth_iv_offset = auth_xform->auth.iv.offset; - - /* Copy the key */ - memcpy(sess->pKey_hash, auth_xform->auth.key.data, - ZUC_IV_KEY_LENGTH); - } - - sess->op = mode; - return 0; -} - -/** Encrypt/decrypt mbufs. */ -static uint8_t -process_zuc_cipher_op(struct ipsec_mb_qp *qp, struct rte_crypto_op **ops, - struct zuc_session **sessions, - uint8_t num_ops) -{ - unsigned int i; - uint8_t processed_ops = 0; - const void *src[ZUC_MAX_BURST]; - void *dst[ZUC_MAX_BURST]; - const void *iv[ZUC_MAX_BURST]; - uint32_t num_bytes[ZUC_MAX_BURST]; - const void *cipher_keys[ZUC_MAX_BURST]; - struct zuc_session *sess; - - for (i = 0; i < num_ops; i++) { - if (((ops[i]->sym->cipher.data.length % BYTE_LEN) != 0) - || ((ops[i]->sym->cipher.data.offset - % BYTE_LEN) != 0)) { - ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - IPSEC_MB_LOG(ERR, "Data Length or offset"); - break; - } - - sess = sessions[i]; - -#ifdef RTE_LIBRTE_PMD_ZUC_DEBUG - if (!rte_pktmbuf_is_contiguous(ops[i]->sym->m_src) || - (ops[i]->sym->m_dst != NULL && - !rte_pktmbuf_is_contiguous( - ops[i]->sym->m_dst))) { - IPSEC_MB_LOG(ERR, "PMD supports only " - " contiguous mbufs, op (%p) " - "provides noncontiguous mbuf " - "as source/destination buffer.\n", - "PMD supports only contiguous mbufs, " - "op (%p) provides noncontiguous mbuf " - "as source/destination buffer.\n", - ops[i]); - ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - break; - } -#endif - - src[i] = rte_pktmbuf_mtod_offset(ops[i]->sym->m_src, - uint8_t *, - (ops[i]->sym->cipher.data.offset >> 3)); - dst[i] = ops[i]->sym->m_dst ? - rte_pktmbuf_mtod_offset(ops[i]->sym->m_dst, uint8_t *, - (ops[i]->sym->cipher.data.offset >> 3)) : - rte_pktmbuf_mtod_offset(ops[i]->sym->m_src, uint8_t *, - (ops[i]->sym->cipher.data.offset >> 3)); - iv[i] = rte_crypto_op_ctod_offset(ops[i], uint8_t *, - sess->cipher_iv_offset); - num_bytes[i] = ops[i]->sym->cipher.data.length >> 3; - - cipher_keys[i] = sess->pKey_cipher; - - processed_ops++; - } - - IMB_ZUC_EEA3_N_BUFFER(qp->mb_mgr, (const void **)cipher_keys, - (const void **)iv, (const void **)src, (void **)dst, - num_bytes, processed_ops); - - return processed_ops; -} - -/** Generate/verify hash from mbufs. */ -static int -process_zuc_hash_op(struct ipsec_mb_qp *qp, struct rte_crypto_op **ops, - struct zuc_session **sessions, - uint8_t num_ops) -{ - unsigned int i; - uint8_t processed_ops = 0; - uint8_t *src[ZUC_MAX_BURST] = { 0 }; - uint32_t *dst[ZUC_MAX_BURST]; - uint32_t length_in_bits[ZUC_MAX_BURST] = { 0 }; - uint8_t *iv[ZUC_MAX_BURST] = { 0 }; - const void *hash_keys[ZUC_MAX_BURST] = { 0 }; - struct zuc_session *sess; - struct zuc_qp_data *qp_data = ipsec_mb_get_qp_private_data(qp); - - - for (i = 0; i < num_ops; i++) { - /* Data must be byte aligned */ - if ((ops[i]->sym->auth.data.offset % BYTE_LEN) != 0) { - ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS; - IPSEC_MB_LOG(ERR, "Offset"); - break; - } - - sess = sessions[i]; - - length_in_bits[i] = ops[i]->sym->auth.data.length; - - src[i] = rte_pktmbuf_mtod_offset(ops[i]->sym->m_src, - uint8_t *, - (ops[i]->sym->auth.data.offset >> 3)); - iv[i] = rte_crypto_op_ctod_offset(ops[i], uint8_t *, - sess->auth_iv_offset); - - hash_keys[i] = sess->pKey_hash; - if (sess->auth_op == RTE_CRYPTO_AUTH_OP_VERIFY) - dst[i] = (uint32_t *)qp_data->temp_digest[i]; - else - dst[i] = (uint32_t *)ops[i]->sym->auth.digest.data; - - processed_ops++; - } - - IMB_ZUC_EIA3_N_BUFFER(qp->mb_mgr, (const void **)hash_keys, - (const void * const *)iv, (const void * const *)src, - length_in_bits, dst, processed_ops); - - /* - * If tag needs to be verified, compare generated tag - * with attached tag - */ - for (i = 0; i < processed_ops; i++) - if (sessions[i]->auth_op == RTE_CRYPTO_AUTH_OP_VERIFY) - if (memcmp(dst[i], ops[i]->sym->auth.digest.data, - ZUC_DIGEST_LENGTH) != 0) - ops[i]->status = - RTE_CRYPTO_OP_STATUS_AUTH_FAILED; - - return processed_ops; -} - -/** Process a batch of crypto ops which shares the same operation type. */ -static int -process_ops(struct rte_crypto_op **ops, enum ipsec_mb_operation op_type, - struct zuc_session **sessions, - struct ipsec_mb_qp *qp, uint8_t num_ops) -{ - unsigned int i; - unsigned int processed_ops = 0; - - switch (op_type) { - case IPSEC_MB_OP_ENCRYPT_ONLY: - case IPSEC_MB_OP_DECRYPT_ONLY: - processed_ops = process_zuc_cipher_op(qp, ops, - sessions, num_ops); - break; - case IPSEC_MB_OP_HASH_GEN_ONLY: - case IPSEC_MB_OP_HASH_VERIFY_ONLY: - processed_ops = process_zuc_hash_op(qp, ops, sessions, - num_ops); - break; - case IPSEC_MB_OP_ENCRYPT_THEN_HASH_GEN: - case IPSEC_MB_OP_DECRYPT_THEN_HASH_VERIFY: - processed_ops = process_zuc_cipher_op(qp, ops, sessions, - num_ops); - process_zuc_hash_op(qp, ops, sessions, processed_ops); - break; - case IPSEC_MB_OP_HASH_VERIFY_THEN_DECRYPT: - case IPSEC_MB_OP_HASH_GEN_THEN_ENCRYPT: - processed_ops = process_zuc_hash_op(qp, ops, sessions, - num_ops); - process_zuc_cipher_op(qp, ops, sessions, processed_ops); - break; - default: - /* Operation not supported. */ - for (i = 0; i < num_ops; i++) - ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_SESSION; - } - - for (i = 0; i < num_ops; i++) { - /* - * If there was no error/authentication failure, - * change status to successful. - */ - if (ops[i]->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED) - ops[i]->status = RTE_CRYPTO_OP_STATUS_SUCCESS; - /* Free session if a session-less crypto op. */ - if (ops[i]->sess_type == RTE_CRYPTO_OP_SESSIONLESS) { - memset(sessions[i], 0, sizeof(struct zuc_session)); - rte_mempool_put(qp->sess_mp, ops[i]->sym->session); - ops[i]->sym->session = NULL; - } - } - return processed_ops; -} - -static uint16_t -zuc_pmd_dequeue_burst(void *queue_pair, - struct rte_crypto_op **c_ops, uint16_t nb_ops) -{ - - struct rte_crypto_op *curr_c_op; - - struct zuc_session *curr_sess; - struct zuc_session *sessions[ZUC_MAX_BURST]; - struct rte_crypto_op *int_c_ops[ZUC_MAX_BURST]; - enum ipsec_mb_operation prev_zuc_op = IPSEC_MB_OP_NOT_SUPPORTED; - enum ipsec_mb_operation curr_zuc_op; - struct ipsec_mb_qp *qp = queue_pair; - unsigned int nb_dequeued; - unsigned int i; - uint8_t burst_size = 0; - uint8_t processed_ops; - - nb_dequeued = rte_ring_dequeue_burst(qp->ingress_queue, - (void **)c_ops, nb_ops, NULL); - - - for (i = 0; i < nb_dequeued; i++) { - curr_c_op = c_ops[i]; - - curr_sess = (struct zuc_session *) - ipsec_mb_get_session_private(qp, curr_c_op); - if (unlikely(curr_sess == NULL)) { - curr_c_op->status = - RTE_CRYPTO_OP_STATUS_INVALID_SESSION; - break; - } - - curr_zuc_op = curr_sess->op; - - /* - * Batch ops that share the same operation type - * (cipher only, auth only...). - */ - if (burst_size == 0) { - prev_zuc_op = curr_zuc_op; - int_c_ops[0] = curr_c_op; - sessions[0] = curr_sess; - burst_size++; - } else if (curr_zuc_op == prev_zuc_op) { - int_c_ops[burst_size] = curr_c_op; - sessions[burst_size] = curr_sess; - burst_size++; - /* - * When there are enough ops to process in a batch, - * process them, and start a new batch. - */ - if (burst_size == ZUC_MAX_BURST) { - processed_ops = process_ops(int_c_ops, curr_zuc_op, - sessions, qp, burst_size); - if (processed_ops < burst_size) { - burst_size = 0; - break; - } - - burst_size = 0; - } - } else { - /* - * Different operation type, process the ops - * of the previous type. - */ - processed_ops = process_ops(int_c_ops, prev_zuc_op, - sessions, qp, burst_size); - if (processed_ops < burst_size) { - burst_size = 0; - break; - } - - burst_size = 0; - prev_zuc_op = curr_zuc_op; - - int_c_ops[0] = curr_c_op; - sessions[0] = curr_sess; - burst_size++; - } - } - - if (burst_size != 0) { - /* Process the crypto ops of the last operation type. */ - processed_ops = process_ops(int_c_ops, prev_zuc_op, - sessions, qp, burst_size); - } - - qp->stats.dequeued_count += i; - return i; -} +#include "pmd_aesni_mb_priv.h" struct rte_cryptodev_ops zuc_pmd_ops = { .dev_configure = ipsec_mb_config, @@ -390,7 +54,7 @@ RTE_INIT(ipsec_mb_register_zuc) = &ipsec_mb_pmds[IPSEC_MB_PMD_TYPE_ZUC]; zuc_data->caps = zuc_capabilities; - zuc_data->dequeue_burst = zuc_pmd_dequeue_burst; + zuc_data->dequeue_burst = aesni_mb_dequeue_burst; zuc_data->feature_flags = RTE_CRYPTODEV_FF_SYMMETRIC_CRYPTO | RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING | RTE_CRYPTODEV_FF_NON_BYTE_ALIGNED_DATA @@ -399,7 +63,8 @@ RTE_INIT(ipsec_mb_register_zuc) | RTE_CRYPTODEV_FF_OOP_LB_IN_LB_OUT; zuc_data->internals_priv_size = 0; zuc_data->ops = &zuc_pmd_ops; - zuc_data->qp_priv_size = sizeof(struct zuc_qp_data); - zuc_data->session_configure = zuc_session_configure; - zuc_data->session_priv_size = sizeof(struct zuc_session); + zuc_data->qp_priv_size = sizeof(struct aesni_mb_qp_data); + zuc_data->session_configure = aesni_mb_session_configure; + zuc_data->session_priv_size = + sizeof(struct aesni_mb_session); } diff --git a/drivers/crypto/ipsec_mb/pmd_zuc_priv.h b/drivers/crypto/ipsec_mb/pmd_zuc_priv.h index 76fd6758c2..a1e8e3aade 100644 --- a/drivers/crypto/ipsec_mb/pmd_zuc_priv.h +++ b/drivers/crypto/ipsec_mb/pmd_zuc_priv.h @@ -10,7 +10,6 @@ #define ZUC_IV_KEY_LENGTH 16 #define ZUC_DIGEST_LENGTH 4 #define ZUC_MAX_BURST 16 -#define BYTE_LEN 8 uint8_t pmd_driver_id_zuc; @@ -63,23 +62,4 @@ static const struct rte_cryptodev_capabilities zuc_capabilities[] = { RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST() }; -/** ZUC private session structure */ -struct zuc_session { - enum ipsec_mb_operation op; - enum rte_crypto_auth_operation auth_op; - uint8_t pKey_cipher[ZUC_IV_KEY_LENGTH]; - uint8_t pKey_hash[ZUC_IV_KEY_LENGTH]; - uint16_t cipher_iv_offset; - uint16_t auth_iv_offset; -} __rte_cache_aligned; - -struct zuc_qp_data { - - uint8_t temp_digest[ZUC_MAX_BURST][ZUC_DIGEST_LENGTH]; - /* *< Buffers used to store the digest generated - * by the driver when verifying a digest provided - * by the user (using authentication verify operation) - */ -}; - #endif /* _PMD_ZUC_PRIV_H_ */ -- 2.25.1