From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mga18.intel.com (mga18.intel.com [134.134.136.126]) by dpdk.org (Postfix) with ESMTP id C38624C88 for ; Mon, 4 Mar 2019 12:48:51 +0100 (CET) X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga005.fm.intel.com ([10.253.24.32]) by orsmga106.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 04 Mar 2019 03:48:49 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.58,439,1544515200"; d="scan'208";a="325227740" Received: from silpixa00398673.ir.intel.com (HELO silpixa00398673.ger.corp.intel.com) ([10.237.223.136]) by fmsmga005.fm.intel.com with ESMTP; 04 Mar 2019 03:48:47 -0800 From: Fan Zhang To: dev@dpdk.org Cc: akhil.goyal@nxp.com, roy.fan.zhang@intel.com, arkadiuszx.kusztal@intel.com, fiona.trahe@intel.com Date: Mon, 4 Mar 2019 11:47:17 +0000 Message-Id: <20190304114717.18152-1-roy.fan.zhang@intel.com> X-Mailer: git-send-email 2.14.5 Subject: [dpdk-dev] [PATCH] crypto/aesni_mb: support newer version library only X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 04 Mar 2019 11:48:52 -0000 As stated in 19.02 deprecation notice, this patch updates the aesni_mb PMD to remove the support of older Intel-ipsec-mb library version eariler than 0.52. Signed-off-by: Fan Zhang --- Although the PMD dependency to the library is changed, the ABI/API are not updated as the APIs to access the PMD remains intact. When the user compile the PMD a error message "IPSec MB version >= 0.52" will be displayed to inform the user to update the library. doc/guides/cryptodevs/aesni_mb.rst | 8 +- doc/guides/rel_notes/deprecation.rst | 3 - drivers/crypto/aesni_mb/Makefile | 21 +- drivers/crypto/aesni_mb/aesni_mb_ops.h | 302 ----- drivers/crypto/aesni_mb/meson.build | 13 +- drivers/crypto/aesni_mb/rte_aesni_mb_pmd_compat.c | 1313 -------------------- .../crypto/aesni_mb/rte_aesni_mb_pmd_ops_compat.c | 745 ----------- drivers/crypto/aesni_mb/rte_aesni_mb_pmd_private.h | 37 - 8 files changed, 16 insertions(+), 2426 deletions(-) delete mode 100644 drivers/crypto/aesni_mb/aesni_mb_ops.h delete mode 100644 drivers/crypto/aesni_mb/rte_aesni_mb_pmd_compat.c delete mode 100644 drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops_compat.c diff --git a/doc/guides/cryptodevs/aesni_mb.rst b/doc/guides/cryptodevs/aesni_mb.rst index 47f2ecc2f..d3567d279 100644 --- a/doc/guides/cryptodevs/aesni_mb.rst +++ b/doc/guides/cryptodevs/aesni_mb.rst @@ -57,10 +57,7 @@ Limitations * Chained mbufs are not supported. * Only in-place is currently supported (destination address is the same as source address). -* RTE_CRYPTO_AEAD_AES_GCM only works properly when the multi-buffer library is - 0.51.0 or newer. -* RTE_CRYPTO_HASH_AES_GMAC is supported by library version v0.51 or later. -* RTE_CRYPTO_HASH_SHA* is supported by library version v0.52 or later. +* Only support Intel multi buffer library version 0.52 or later. Installation @@ -92,7 +89,8 @@ and the Multi-Buffer library version supported by them: 17.05 - 17.08 0.45 - 0.48 17.11 0.47 - 0.48 18.02 0.48 - 18.05+ 0.49+ + 18.05 0.49 + 19.02+ 0.52+ ============== ============================ diff --git a/doc/guides/rel_notes/deprecation.rst b/doc/guides/rel_notes/deprecation.rst index 1b4fcb7e6..8adeaa552 100644 --- a/doc/guides/rel_notes/deprecation.rst +++ b/doc/guides/rel_notes/deprecation.rst @@ -72,6 +72,3 @@ Deprecation Notices replace ``enum rte_meter_color`` in meter library in 19.05. This will help to consolidate color definition, which is currently replicated in many places, such as: rte_meter.h, rte_mtr.h, rte_tm.h. - -* crypto/aesni_mb: the minimum supported intel-ipsec-mb library version will be - changed from 0.49.0 to 0.52.0. diff --git a/drivers/crypto/aesni_mb/Makefile b/drivers/crypto/aesni_mb/Makefile index 8d2024c9e..f3035340a 100644 --- a/drivers/crypto/aesni_mb/Makefile +++ b/drivers/crypto/aesni_mb/Makefile @@ -32,19 +32,14 @@ IMB_VERSION = $(shell grep -e "IMB_VERSION_STR" $(IMB_HDR) | cut -d'"' -f2) IMB_VERSION_NUM = $(shell grep -e "IMB_VERSION_NUM" $(IMB_HDR) | cut -d' ' -f3) ifeq ($(IMB_VERSION),) - # files for older version of IMB - SRCS-$(CONFIG_RTE_LIBRTE_PMD_AESNI_MB) += rte_aesni_mb_pmd_compat.c - SRCS-$(CONFIG_RTE_LIBRTE_PMD_AESNI_MB) += rte_aesni_mb_pmd_ops_compat.c -else - ifeq ($(shell expr $(IMB_VERSION_NUM) \>= 0x3400), 1) - # files for a new version of IMB - SRCS-$(CONFIG_RTE_LIBRTE_PMD_AESNI_MB) += rte_aesni_mb_pmd.c - SRCS-$(CONFIG_RTE_LIBRTE_PMD_AESNI_MB) += rte_aesni_mb_pmd_ops.c - else - # files for older version of IMB - SRCS-$(CONFIG_RTE_LIBRTE_PMD_AESNI_MB) += rte_aesni_mb_pmd_compat.c - SRCS-$(CONFIG_RTE_LIBRTE_PMD_AESNI_MB) += rte_aesni_mb_pmd_ops_compat.c - endif +$(error "IPSec_MB version >= 0.52 is required") endif +ifeq ($(shell expr $(IMB_VERSION_NUM) \< 0x3400), 1) +$(error "IPSec_MB version >= 0.52 is required") +endif + +SRCS-$(CONFIG_RTE_LIBRTE_PMD_AESNI_MB) += rte_aesni_mb_pmd.c +SRCS-$(CONFIG_RTE_LIBRTE_PMD_AESNI_MB) += rte_aesni_mb_pmd_ops.c + include $(RTE_SDK)/mk/rte.lib.mk diff --git a/drivers/crypto/aesni_mb/aesni_mb_ops.h b/drivers/crypto/aesni_mb/aesni_mb_ops.h deleted file mode 100644 index 575d6a5b8..000000000 --- a/drivers/crypto/aesni_mb/aesni_mb_ops.h +++ /dev/null @@ -1,302 +0,0 @@ -/* SPDX-License-Identifier: BSD-3-Clause - * Copyright(c) 2015 Intel Corporation - */ - -#ifndef _AESNI_MB_OPS_H_ -#define _AESNI_MB_OPS_H_ - -#ifndef LINUX -#define LINUX -#endif - -#include - -/* - * IMB_VERSION_NUM macro was introduced in version Multi-buffer 0.50, - * so if macro is not defined, it means that the version is 0.49. - */ -#if !defined(IMB_VERSION_NUM) -#define IMB_VERSION(a, b, c) (((a) << 16) + ((b) << 8) + (c)) -#define IMB_VERSION_NUM IMB_VERSION(0, 49, 0) -#endif - -enum aesni_mb_vector_mode { - RTE_AESNI_MB_NOT_SUPPORTED = 0, - RTE_AESNI_MB_SSE, - RTE_AESNI_MB_AVX, - RTE_AESNI_MB_AVX2, - RTE_AESNI_MB_AVX512 -}; - -typedef void (*md5_one_block_t)(const void *data, void *digest); - -typedef void (*sha1_one_block_t)(const void *data, void *digest); -typedef void (*sha224_one_block_t)(const void *data, void *digest); -typedef void (*sha256_one_block_t)(const void *data, void *digest); -typedef void (*sha384_one_block_t)(const void *data, void *digest); -typedef void (*sha512_one_block_t)(const void *data, void *digest); - -typedef void (*aes_keyexp_128_t) - (const void *key, void *enc_exp_keys, void *dec_exp_keys); -typedef void (*aes_keyexp_192_t) - (const void *key, void *enc_exp_keys, void *dec_exp_keys); -typedef void (*aes_keyexp_256_t) - (const void *key, void *enc_exp_keys, void *dec_exp_keys); -typedef void (*aes_xcbc_expand_key_t) - (const void *key, void *exp_k1, void *k2, void *k3); -typedef void (*aes_cmac_sub_key_gen_t) - (const void *exp_key, void *k2, void *k3); -typedef void (*aes_cmac_keyexp_t) - (const void *key, void *keyexp); -typedef void (*aes_gcm_keyexp_t) - (const void *key, struct gcm_key_data *keyexp); - -/** Multi-buffer library function pointer table */ -struct aesni_mb_op_fns { - struct { - init_mb_mgr_t init_mgr; - /**< Initialise scheduler */ - get_next_job_t get_next; - /**< Get next free job structure */ - submit_job_t submit; - /**< Submit job to scheduler */ - get_completed_job_t get_completed_job; - /**< Get completed job */ - flush_job_t flush_job; - /**< flush jobs from manager */ - } job; - /**< multi buffer manager functions */ - - struct { - struct { - md5_one_block_t md5; - /**< MD5 one block hash */ - sha1_one_block_t sha1; - /**< SHA1 one block hash */ - sha224_one_block_t sha224; - /**< SHA224 one block hash */ - sha256_one_block_t sha256; - /**< SHA256 one block hash */ - sha384_one_block_t sha384; - /**< SHA384 one block hash */ - sha512_one_block_t sha512; - /**< SHA512 one block hash */ - } one_block; - /**< one block hash functions */ - - struct { - aes_keyexp_128_t aes128; - /**< AES128 key expansions */ - aes_keyexp_192_t aes192; - /**< AES192 key expansions */ - aes_keyexp_256_t aes256; - /**< AES256 key expansions */ - aes_xcbc_expand_key_t aes_xcbc; - /**< AES XCBC key epansions */ - aes_cmac_sub_key_gen_t aes_cmac_subkey; - /**< AES CMAC subkey expansions */ - aes_cmac_keyexp_t aes_cmac_expkey; - /**< AES CMAC key expansions */ - aes_gcm_keyexp_t aes_gcm_128; - /**< AES GCM 128 key expansions */ - aes_gcm_keyexp_t aes_gcm_192; - /**< AES GCM 192 key expansions */ - aes_gcm_keyexp_t aes_gcm_256; - /**< AES GCM 256 key expansions */ - } keyexp; - /**< Key expansion functions */ -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - struct { - hash_fn_t sha1; - hash_fn_t sha224; - hash_fn_t sha256; - hash_fn_t sha384; - hash_fn_t sha512; - } multi_block; - /** multi block hash functions */ -#endif - } aux; - /**< Auxiliary functions */ -}; - - -static const struct aesni_mb_op_fns job_ops[] = { - [RTE_AESNI_MB_NOT_SUPPORTED] = { - .job = { - NULL - }, - .aux = { - .one_block = { - NULL - }, - .keyexp = { - NULL - }, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .multi_block = { - NULL - } -#endif - - } - }, - [RTE_AESNI_MB_SSE] = { - .job = { - init_mb_mgr_sse, - get_next_job_sse, - submit_job_sse, - get_completed_job_sse, - flush_job_sse - }, - .aux = { - .one_block = { - md5_one_block_sse, - sha1_one_block_sse, - sha224_one_block_sse, - sha256_one_block_sse, - sha384_one_block_sse, - sha512_one_block_sse - }, - .keyexp = { - aes_keyexp_128_sse, - aes_keyexp_192_sse, - aes_keyexp_256_sse, - aes_xcbc_expand_key_sse, - aes_cmac_subkey_gen_sse, - aes_keyexp_128_enc_sse, - aes_gcm_pre_128_sse, - aes_gcm_pre_192_sse, - aes_gcm_pre_256_sse - }, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .multi_block = { - sha1_sse, - sha224_sse, - sha256_sse, - sha384_sse, - sha512_sse - } -#endif - } - }, - [RTE_AESNI_MB_AVX] = { - .job = { - init_mb_mgr_avx, - get_next_job_avx, - submit_job_avx, - get_completed_job_avx, - flush_job_avx - }, - .aux = { - .one_block = { - md5_one_block_avx, - sha1_one_block_avx, - sha224_one_block_avx, - sha256_one_block_avx, - sha384_one_block_avx, - sha512_one_block_avx - }, - .keyexp = { - aes_keyexp_128_avx, - aes_keyexp_192_avx, - aes_keyexp_256_avx, - aes_xcbc_expand_key_avx, - aes_cmac_subkey_gen_avx, - aes_keyexp_128_enc_avx, - aes_gcm_pre_128_avx_gen2, - aes_gcm_pre_192_avx_gen2, - aes_gcm_pre_256_avx_gen2 - }, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .multi_block = { - sha1_avx, - sha224_avx, - sha256_avx, - sha384_avx, - sha512_avx - } -#endif - } - }, - [RTE_AESNI_MB_AVX2] = { - .job = { - init_mb_mgr_avx2, - get_next_job_avx2, - submit_job_avx2, - get_completed_job_avx2, - flush_job_avx2 - }, - .aux = { - .one_block = { - md5_one_block_avx2, - sha1_one_block_avx2, - sha224_one_block_avx2, - sha256_one_block_avx2, - sha384_one_block_avx2, - sha512_one_block_avx2 - }, - .keyexp = { - aes_keyexp_128_avx2, - aes_keyexp_192_avx2, - aes_keyexp_256_avx2, - aes_xcbc_expand_key_avx2, - aes_cmac_subkey_gen_avx2, - aes_keyexp_128_enc_avx2, - aes_gcm_pre_128_avx_gen4, - aes_gcm_pre_192_avx_gen4, - aes_gcm_pre_256_avx_gen4 - }, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .multi_block = { - sha1_avx2, - sha224_avx2, - sha256_avx2, - sha384_avx2, - sha512_avx2 - } -#endif - } - }, - [RTE_AESNI_MB_AVX512] = { - .job = { - init_mb_mgr_avx512, - get_next_job_avx512, - submit_job_avx512, - get_completed_job_avx512, - flush_job_avx512 - }, - .aux = { - .one_block = { - md5_one_block_avx512, - sha1_one_block_avx512, - sha224_one_block_avx512, - sha256_one_block_avx512, - sha384_one_block_avx512, - sha512_one_block_avx512 - }, - .keyexp = { - aes_keyexp_128_avx512, - aes_keyexp_192_avx512, - aes_keyexp_256_avx512, - aes_xcbc_expand_key_avx512, - aes_cmac_subkey_gen_avx512, - aes_keyexp_128_enc_avx512, - aes_gcm_pre_128_avx_gen4, - aes_gcm_pre_192_avx_gen4, - aes_gcm_pre_256_avx_gen4 - }, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .multi_block = { - sha1_avx512, - sha224_avx512, - sha256_avx512, - sha384_avx512, - sha512_avx512 - } -#endif - } - } -}; - - -#endif /* _AESNI_MB_OPS_H_ */ diff --git a/drivers/crypto/aesni_mb/meson.build b/drivers/crypto/aesni_mb/meson.build index 6313c4bd0..fbc4878af 100644 --- a/drivers/crypto/aesni_mb/meson.build +++ b/drivers/crypto/aesni_mb/meson.build @@ -10,16 +10,13 @@ else imb_arr = cc.get_define('IMB_VERSION_STR', prefix : '#include').split('"') - imb_ver =''.join(imb_arr) + imb_ver = ''.join(imb_arr) - if imb_ver.version_compare('>=' + IPSec_MB_ver_0_52) - message('Build for a new version of library IPSec_MB[' + imb_ver + ']') - sources = files('rte_aesni_mb_pmd.c', - 'rte_aesni_mb_pmd_ops.c') + if (imb_ver == '') or (imb_ver.version_compare('<' + IPSec_MB_ver_0_52)) + message('IPSec_MB version >= 0.52 is required') + build = false else - sources = files('rte_aesni_mb_pmd_compat.c', - 'rte_aesni_mb_pmd_ops_compat.c') - message('Build for older version of library IPSec_MB[' + imb_ver + ']') + sources = files('rte_aesni_mb_pmd.c', 'rte_aesni_mb_pmd_ops.c') endif endif diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_compat.c b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_compat.c deleted file mode 100644 index 8020f68e3..000000000 --- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_compat.c +++ /dev/null @@ -1,1313 +0,0 @@ -/* SPDX-License-Identifier: BSD-3-Clause - * Copyright(c) 2015-2017 Intel Corporation - */ - -#include - -#include -#include -#include -#include -#include -#include -#include - -#include "rte_aesni_mb_pmd_private.h" - -#define AES_CCM_DIGEST_MIN_LEN 4 -#define AES_CCM_DIGEST_MAX_LEN 16 -#define HMAC_MAX_BLOCK_SIZE 128 -static uint8_t cryptodev_driver_id; - -typedef void (*hash_one_block_t)(const void *data, void *digest); -typedef void (*aes_keyexp_t)(const void *key, void *enc_exp_keys, void *dec_exp_keys); - -/** - * Calculate the authentication pre-computes - * - * @param one_block_hash Function pointer to calculate digest on ipad/opad - * @param ipad Inner pad output byte array - * @param opad Outer pad output byte array - * @param hkey Authentication key - * @param hkey_len Authentication key length - * @param blocksize Block size of selected hash algo - */ -static void -calculate_auth_precomputes(hash_one_block_t one_block_hash, - uint8_t *ipad, uint8_t *opad, - uint8_t *hkey, uint16_t hkey_len, - uint16_t blocksize) -{ - unsigned i, length; - - uint8_t ipad_buf[blocksize] __rte_aligned(16); - uint8_t opad_buf[blocksize] __rte_aligned(16); - - /* Setup inner and outer pads */ - memset(ipad_buf, HMAC_IPAD_VALUE, blocksize); - memset(opad_buf, HMAC_OPAD_VALUE, blocksize); - - /* XOR hash key with inner and outer pads */ - length = hkey_len > blocksize ? blocksize : hkey_len; - - for (i = 0; i < length; i++) { - ipad_buf[i] ^= hkey[i]; - opad_buf[i] ^= hkey[i]; - } - - /* Compute partial hashes */ - (*one_block_hash)(ipad_buf, ipad); - (*one_block_hash)(opad_buf, opad); - - /* Clean up stack */ - memset(ipad_buf, 0, blocksize); - memset(opad_buf, 0, blocksize); -} - -/** Get xform chain order */ -static enum aesni_mb_operation -aesni_mb_get_chain_order(const struct rte_crypto_sym_xform *xform) -{ - if (xform == NULL) - return AESNI_MB_OP_NOT_SUPPORTED; - - if (xform->type == RTE_CRYPTO_SYM_XFORM_CIPHER) { - if (xform->next == NULL) - return AESNI_MB_OP_CIPHER_ONLY; - if (xform->next->type == RTE_CRYPTO_SYM_XFORM_AUTH) - return AESNI_MB_OP_CIPHER_HASH; - } - - if (xform->type == RTE_CRYPTO_SYM_XFORM_AUTH) { - if (xform->next == NULL) - return AESNI_MB_OP_HASH_ONLY; - if (xform->next->type == RTE_CRYPTO_SYM_XFORM_CIPHER) - return AESNI_MB_OP_HASH_CIPHER; - } - - if (xform->type == RTE_CRYPTO_SYM_XFORM_AEAD) { - if (xform->aead.algo == RTE_CRYPTO_AEAD_AES_CCM || - xform->aead.algo == RTE_CRYPTO_AEAD_AES_GCM) { - if (xform->aead.op == RTE_CRYPTO_AEAD_OP_ENCRYPT) - return AESNI_MB_OP_AEAD_CIPHER_HASH; - else - return AESNI_MB_OP_AEAD_HASH_CIPHER; - } - } - - return AESNI_MB_OP_NOT_SUPPORTED; -} - -/** Set session authentication parameters */ -static int -aesni_mb_set_session_auth_parameters(const struct aesni_mb_op_fns *mb_ops, - struct aesni_mb_session *sess, - const struct rte_crypto_sym_xform *xform) -{ - hash_one_block_t hash_oneblock_fn; - unsigned int key_larger_block_size = 0; - uint8_t hashed_key[HMAC_MAX_BLOCK_SIZE] = { 0 }; - - if (xform == NULL) { - sess->auth.algo = NULL_HASH; - return 0; - } - - if (xform->type != RTE_CRYPTO_SYM_XFORM_AUTH) { - AESNI_MB_LOG(ERR, "Crypto xform struct not of type auth"); - return -1; - } - - /* Set the request digest size */ - sess->auth.req_digest_len = xform->auth.digest_length; - - /* Select auth generate/verify */ - sess->auth.operation = xform->auth.op; - - /* Set Authentication Parameters */ - if (xform->auth.algo == RTE_CRYPTO_AUTH_AES_XCBC_MAC) { - sess->auth.algo = AES_XCBC; - - uint16_t xcbc_mac_digest_len = - get_truncated_digest_byte_length(AES_XCBC); - if (sess->auth.req_digest_len != xcbc_mac_digest_len) { - AESNI_MB_LOG(ERR, "Invalid digest size\n"); - return -EINVAL; - } - sess->auth.gen_digest_len = sess->auth.req_digest_len; - (*mb_ops->aux.keyexp.aes_xcbc)(xform->auth.key.data, - sess->auth.xcbc.k1_expanded, - sess->auth.xcbc.k2, sess->auth.xcbc.k3); - return 0; - } - - if (xform->auth.algo == RTE_CRYPTO_AUTH_AES_CMAC) { - sess->auth.algo = AES_CMAC; - - uint16_t cmac_digest_len = get_digest_byte_length(AES_CMAC); - - if (sess->auth.req_digest_len > cmac_digest_len) { - AESNI_MB_LOG(ERR, "Invalid digest size\n"); - return -EINVAL; - } - /* - * Multi-buffer lib supports digest sizes from 4 to 16 bytes - * in version 0.50 and sizes of 12 and 16 bytes, - * in version 0.49. - * If size requested is different, generate the full digest - * (16 bytes) in a temporary location and then memcpy - * the requested number of bytes. - */ -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - if (sess->auth.req_digest_len < 4) -#else - uint16_t cmac_trunc_digest_len = - get_truncated_digest_byte_length(AES_CMAC); - if (sess->auth.req_digest_len != cmac_digest_len && - sess->auth.req_digest_len != cmac_trunc_digest_len) -#endif - sess->auth.gen_digest_len = cmac_digest_len; - else - sess->auth.gen_digest_len = sess->auth.req_digest_len; - (*mb_ops->aux.keyexp.aes_cmac_expkey)(xform->auth.key.data, - sess->auth.cmac.expkey); - - (*mb_ops->aux.keyexp.aes_cmac_subkey)(sess->auth.cmac.expkey, - sess->auth.cmac.skey1, sess->auth.cmac.skey2); - return 0; - } - - if (xform->auth.algo == RTE_CRYPTO_AUTH_AES_GMAC) { - if (xform->auth.op == RTE_CRYPTO_AUTH_OP_GENERATE) { - sess->cipher.direction = ENCRYPT; - sess->chain_order = CIPHER_HASH; - } else - sess->cipher.direction = DECRYPT; - - sess->auth.algo = AES_GMAC; - /* - * Multi-buffer lib supports 8, 12 and 16 bytes of digest. - * If size requested is different, generate the full digest - * (16 bytes) in a temporary location and then memcpy - * the requested number of bytes. - */ - if (sess->auth.req_digest_len != 16 && - sess->auth.req_digest_len != 12 && - sess->auth.req_digest_len != 8) { - sess->auth.gen_digest_len = 16; - } else { - sess->auth.gen_digest_len = sess->auth.req_digest_len; - } - sess->iv.length = xform->auth.iv.length; - sess->iv.offset = xform->auth.iv.offset; - - switch (xform->auth.key.length) { - case AES_128_BYTES: - sess->cipher.key_length_in_bytes = AES_128_BYTES; - (mb_ops->aux.keyexp.aes_gcm_128)(xform->auth.key.data, - &sess->cipher.gcm_key); - break; - case AES_192_BYTES: - sess->cipher.key_length_in_bytes = AES_192_BYTES; - (mb_ops->aux.keyexp.aes_gcm_192)(xform->auth.key.data, - &sess->cipher.gcm_key); - break; - case AES_256_BYTES: - sess->cipher.key_length_in_bytes = AES_256_BYTES; - (mb_ops->aux.keyexp.aes_gcm_256)(xform->auth.key.data, - &sess->cipher.gcm_key); - break; - default: - RTE_LOG(ERR, PMD, "failed to parse test type\n"); - return -EINVAL; - } - - return 0; - } - - switch (xform->auth.algo) { - case RTE_CRYPTO_AUTH_MD5_HMAC: - sess->auth.algo = MD5; - hash_oneblock_fn = mb_ops->aux.one_block.md5; - break; - case RTE_CRYPTO_AUTH_SHA1_HMAC: - sess->auth.algo = SHA1; - hash_oneblock_fn = mb_ops->aux.one_block.sha1; -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - if (xform->auth.key.length > get_auth_algo_blocksize(SHA1)) { - mb_ops->aux.multi_block.sha1( - xform->auth.key.data, - xform->auth.key.length, - hashed_key); - key_larger_block_size = 1; - } -#endif - break; - case RTE_CRYPTO_AUTH_SHA224_HMAC: - sess->auth.algo = SHA_224; - hash_oneblock_fn = mb_ops->aux.one_block.sha224; -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - if (xform->auth.key.length > get_auth_algo_blocksize(SHA_224)) { - mb_ops->aux.multi_block.sha224( - xform->auth.key.data, - xform->auth.key.length, - hashed_key); - key_larger_block_size = 1; - } -#endif - break; - case RTE_CRYPTO_AUTH_SHA256_HMAC: - sess->auth.algo = SHA_256; - hash_oneblock_fn = mb_ops->aux.one_block.sha256; -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - if (xform->auth.key.length > get_auth_algo_blocksize(SHA_256)) { - mb_ops->aux.multi_block.sha256( - xform->auth.key.data, - xform->auth.key.length, - hashed_key); - key_larger_block_size = 1; - } -#endif - break; - case RTE_CRYPTO_AUTH_SHA384_HMAC: - sess->auth.algo = SHA_384; - hash_oneblock_fn = mb_ops->aux.one_block.sha384; -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - if (xform->auth.key.length > get_auth_algo_blocksize(SHA_384)) { - mb_ops->aux.multi_block.sha384( - xform->auth.key.data, - xform->auth.key.length, - hashed_key); - key_larger_block_size = 1; - } -#endif - break; - case RTE_CRYPTO_AUTH_SHA512_HMAC: - sess->auth.algo = SHA_512; - hash_oneblock_fn = mb_ops->aux.one_block.sha512; -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - if (xform->auth.key.length > get_auth_algo_blocksize(SHA_512)) { - mb_ops->aux.multi_block.sha512( - xform->auth.key.data, - xform->auth.key.length, - hashed_key); - key_larger_block_size = 1; - } -#endif - break; - default: - AESNI_MB_LOG(ERR, "Unsupported authentication algorithm selection"); - return -ENOTSUP; - } - uint16_t trunc_digest_size = - get_truncated_digest_byte_length(sess->auth.algo); - uint16_t full_digest_size = - get_digest_byte_length(sess->auth.algo); - -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - if (sess->auth.req_digest_len > full_digest_size || - sess->auth.req_digest_len == 0) { -#else - if (sess->auth.req_digest_len != trunc_digest_size) { -#endif - AESNI_MB_LOG(ERR, "Invalid digest size\n"); - return -EINVAL; - } - - if (sess->auth.req_digest_len != trunc_digest_size && - sess->auth.req_digest_len != full_digest_size) - sess->auth.gen_digest_len = full_digest_size; - else - sess->auth.gen_digest_len = sess->auth.req_digest_len; - - /* Calculate Authentication precomputes */ - if (key_larger_block_size) { - calculate_auth_precomputes(hash_oneblock_fn, - sess->auth.pads.inner, sess->auth.pads.outer, - hashed_key, - xform->auth.key.length, - get_auth_algo_blocksize(sess->auth.algo)); - } else { - calculate_auth_precomputes(hash_oneblock_fn, - sess->auth.pads.inner, sess->auth.pads.outer, - xform->auth.key.data, - xform->auth.key.length, - get_auth_algo_blocksize(sess->auth.algo)); - } - - return 0; -} - -/** Set session cipher parameters */ -static int -aesni_mb_set_session_cipher_parameters(const struct aesni_mb_op_fns *mb_ops, - struct aesni_mb_session *sess, - const struct rte_crypto_sym_xform *xform) -{ - uint8_t is_aes = 0; - uint8_t is_3DES = 0; - aes_keyexp_t aes_keyexp_fn; - - if (xform == NULL) { - sess->cipher.mode = NULL_CIPHER; - return 0; - } - - if (xform->type != RTE_CRYPTO_SYM_XFORM_CIPHER) { - AESNI_MB_LOG(ERR, "Crypto xform struct not of type cipher"); - return -EINVAL; - } - - /* Select cipher direction */ - switch (xform->cipher.op) { - case RTE_CRYPTO_CIPHER_OP_ENCRYPT: - sess->cipher.direction = ENCRYPT; - break; - case RTE_CRYPTO_CIPHER_OP_DECRYPT: - sess->cipher.direction = DECRYPT; - break; - default: - AESNI_MB_LOG(ERR, "Invalid cipher operation parameter"); - return -EINVAL; - } - - /* Select cipher mode */ - switch (xform->cipher.algo) { - case RTE_CRYPTO_CIPHER_AES_CBC: - sess->cipher.mode = CBC; - is_aes = 1; - break; - case RTE_CRYPTO_CIPHER_AES_CTR: - sess->cipher.mode = CNTR; - is_aes = 1; - break; - case RTE_CRYPTO_CIPHER_AES_DOCSISBPI: - sess->cipher.mode = DOCSIS_SEC_BPI; - is_aes = 1; - break; - case RTE_CRYPTO_CIPHER_DES_CBC: - sess->cipher.mode = DES; - break; - case RTE_CRYPTO_CIPHER_DES_DOCSISBPI: - sess->cipher.mode = DOCSIS_DES; - break; - case RTE_CRYPTO_CIPHER_3DES_CBC: - sess->cipher.mode = DES3; - is_3DES = 1; - break; - default: - AESNI_MB_LOG(ERR, "Unsupported cipher mode parameter"); - return -ENOTSUP; - } - - /* Set IV parameters */ - sess->iv.offset = xform->cipher.iv.offset; - sess->iv.length = xform->cipher.iv.length; - - /* Check key length and choose key expansion function for AES */ - if (is_aes) { - switch (xform->cipher.key.length) { - case AES_128_BYTES: - sess->cipher.key_length_in_bytes = AES_128_BYTES; - aes_keyexp_fn = mb_ops->aux.keyexp.aes128; - break; - case AES_192_BYTES: - sess->cipher.key_length_in_bytes = AES_192_BYTES; - aes_keyexp_fn = mb_ops->aux.keyexp.aes192; - break; - case AES_256_BYTES: - sess->cipher.key_length_in_bytes = AES_256_BYTES; - aes_keyexp_fn = mb_ops->aux.keyexp.aes256; - break; - default: - AESNI_MB_LOG(ERR, "Invalid cipher key length"); - return -EINVAL; - } - - /* Expanded cipher keys */ - (*aes_keyexp_fn)(xform->cipher.key.data, - sess->cipher.expanded_aes_keys.encode, - sess->cipher.expanded_aes_keys.decode); - - } else if (is_3DES) { - uint64_t *keys[3] = {sess->cipher.exp_3des_keys.key[0], - sess->cipher.exp_3des_keys.key[1], - sess->cipher.exp_3des_keys.key[2]}; - - switch (xform->cipher.key.length) { - case 24: - des_key_schedule(keys[0], xform->cipher.key.data); - des_key_schedule(keys[1], xform->cipher.key.data+8); - des_key_schedule(keys[2], xform->cipher.key.data+16); - - /* Initialize keys - 24 bytes: [K1-K2-K3] */ - sess->cipher.exp_3des_keys.ks_ptr[0] = keys[0]; - sess->cipher.exp_3des_keys.ks_ptr[1] = keys[1]; - sess->cipher.exp_3des_keys.ks_ptr[2] = keys[2]; - break; - case 16: - des_key_schedule(keys[0], xform->cipher.key.data); - des_key_schedule(keys[1], xform->cipher.key.data+8); - - /* Initialize keys - 16 bytes: [K1=K1,K2=K2,K3=K1] */ - sess->cipher.exp_3des_keys.ks_ptr[0] = keys[0]; - sess->cipher.exp_3des_keys.ks_ptr[1] = keys[1]; - sess->cipher.exp_3des_keys.ks_ptr[2] = keys[0]; - break; - case 8: - des_key_schedule(keys[0], xform->cipher.key.data); - - /* Initialize keys - 8 bytes: [K1 = K2 = K3] */ - sess->cipher.exp_3des_keys.ks_ptr[0] = keys[0]; - sess->cipher.exp_3des_keys.ks_ptr[1] = keys[0]; - sess->cipher.exp_3des_keys.ks_ptr[2] = keys[0]; - break; - default: - AESNI_MB_LOG(ERR, "Invalid cipher key length"); - return -EINVAL; - } - -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - sess->cipher.key_length_in_bytes = 24; -#else - sess->cipher.key_length_in_bytes = 8; -#endif - } else { - if (xform->cipher.key.length != 8) { - AESNI_MB_LOG(ERR, "Invalid cipher key length"); - return -EINVAL; - } - sess->cipher.key_length_in_bytes = 8; - - des_key_schedule((uint64_t *)sess->cipher.expanded_aes_keys.encode, - xform->cipher.key.data); - des_key_schedule((uint64_t *)sess->cipher.expanded_aes_keys.decode, - xform->cipher.key.data); - } - - return 0; -} - -static int -aesni_mb_set_session_aead_parameters(const struct aesni_mb_op_fns *mb_ops, - struct aesni_mb_session *sess, - const struct rte_crypto_sym_xform *xform) -{ - union { - aes_keyexp_t aes_keyexp_fn; - aes_gcm_keyexp_t aes_gcm_keyexp_fn; - } keyexp; - - switch (xform->aead.op) { - case RTE_CRYPTO_AEAD_OP_ENCRYPT: - sess->cipher.direction = ENCRYPT; - sess->auth.operation = RTE_CRYPTO_AUTH_OP_GENERATE; - break; - case RTE_CRYPTO_AEAD_OP_DECRYPT: - sess->cipher.direction = DECRYPT; - sess->auth.operation = RTE_CRYPTO_AUTH_OP_VERIFY; - break; - default: - AESNI_MB_LOG(ERR, "Invalid aead operation parameter"); - return -EINVAL; - } - - switch (xform->aead.algo) { - case RTE_CRYPTO_AEAD_AES_CCM: - sess->cipher.mode = CCM; - sess->auth.algo = AES_CCM; - - /* Check key length and choose key expansion function for AES */ - switch (xform->aead.key.length) { - case AES_128_BYTES: - sess->cipher.key_length_in_bytes = AES_128_BYTES; - keyexp.aes_keyexp_fn = mb_ops->aux.keyexp.aes128; - break; - default: - AESNI_MB_LOG(ERR, "Invalid cipher key length"); - return -EINVAL; - } - - /* Expanded cipher keys */ - (*keyexp.aes_keyexp_fn)(xform->aead.key.data, - sess->cipher.expanded_aes_keys.encode, - sess->cipher.expanded_aes_keys.decode); - break; - - case RTE_CRYPTO_AEAD_AES_GCM: - sess->cipher.mode = GCM; - sess->auth.algo = AES_GMAC; - - switch (xform->aead.key.length) { - case AES_128_BYTES: - sess->cipher.key_length_in_bytes = AES_128_BYTES; - keyexp.aes_gcm_keyexp_fn = - mb_ops->aux.keyexp.aes_gcm_128; - break; - case AES_192_BYTES: - sess->cipher.key_length_in_bytes = AES_192_BYTES; - keyexp.aes_gcm_keyexp_fn = - mb_ops->aux.keyexp.aes_gcm_192; - break; - case AES_256_BYTES: - sess->cipher.key_length_in_bytes = AES_256_BYTES; - keyexp.aes_gcm_keyexp_fn = - mb_ops->aux.keyexp.aes_gcm_256; - break; - default: - AESNI_MB_LOG(ERR, "Invalid cipher key length"); - return -EINVAL; - } - - (keyexp.aes_gcm_keyexp_fn)(xform->aead.key.data, - &sess->cipher.gcm_key); - break; - - default: - AESNI_MB_LOG(ERR, "Unsupported aead mode parameter"); - return -ENOTSUP; - } - - /* Set IV parameters */ - sess->iv.offset = xform->aead.iv.offset; - sess->iv.length = xform->aead.iv.length; - - sess->auth.req_digest_len = xform->aead.digest_length; - /* CCM digests must be between 4 and 16 and an even number */ - if (sess->auth.req_digest_len < AES_CCM_DIGEST_MIN_LEN || - sess->auth.req_digest_len > AES_CCM_DIGEST_MAX_LEN || - (sess->auth.req_digest_len & 1) == 1) { - AESNI_MB_LOG(ERR, "Invalid digest size\n"); - return -EINVAL; - } - sess->auth.gen_digest_len = sess->auth.req_digest_len; - - return 0; -} - -/** Parse crypto xform chain and set private session parameters */ -int -aesni_mb_set_session_parameters(const struct aesni_mb_op_fns *mb_ops, - struct aesni_mb_session *sess, - const struct rte_crypto_sym_xform *xform) -{ - const struct rte_crypto_sym_xform *auth_xform = NULL; - const struct rte_crypto_sym_xform *cipher_xform = NULL; - const struct rte_crypto_sym_xform *aead_xform = NULL; - int ret; - - /* Select Crypto operation - hash then cipher / cipher then hash */ - switch (aesni_mb_get_chain_order(xform)) { - case AESNI_MB_OP_HASH_CIPHER: - sess->chain_order = HASH_CIPHER; - auth_xform = xform; - cipher_xform = xform->next; - break; - case AESNI_MB_OP_CIPHER_HASH: - sess->chain_order = CIPHER_HASH; - auth_xform = xform->next; - cipher_xform = xform; - break; - case AESNI_MB_OP_HASH_ONLY: - sess->chain_order = HASH_CIPHER; - auth_xform = xform; - cipher_xform = NULL; - break; - case AESNI_MB_OP_CIPHER_ONLY: - /* - * Multi buffer library operates only at two modes, - * CIPHER_HASH and HASH_CIPHER. When doing ciphering only, - * chain order depends on cipher operation: encryption is always - * the first operation and decryption the last one. - */ - if (xform->cipher.op == RTE_CRYPTO_CIPHER_OP_ENCRYPT) - sess->chain_order = CIPHER_HASH; - else - sess->chain_order = HASH_CIPHER; - auth_xform = NULL; - cipher_xform = xform; - break; - case AESNI_MB_OP_AEAD_CIPHER_HASH: - sess->chain_order = CIPHER_HASH; - sess->aead.aad_len = xform->aead.aad_length; - aead_xform = xform; - break; - case AESNI_MB_OP_AEAD_HASH_CIPHER: - sess->chain_order = HASH_CIPHER; - sess->aead.aad_len = xform->aead.aad_length; - aead_xform = xform; - break; - case AESNI_MB_OP_NOT_SUPPORTED: - default: - AESNI_MB_LOG(ERR, "Unsupported operation chain order parameter"); - return -ENOTSUP; - } - - /* Default IV length = 0 */ - sess->iv.length = 0; - - ret = aesni_mb_set_session_auth_parameters(mb_ops, sess, auth_xform); - if (ret != 0) { - AESNI_MB_LOG(ERR, "Invalid/unsupported authentication parameters"); - return ret; - } - - ret = aesni_mb_set_session_cipher_parameters(mb_ops, sess, - cipher_xform); - if (ret != 0) { - AESNI_MB_LOG(ERR, "Invalid/unsupported cipher parameters"); - return ret; - } - - if (aead_xform) { - ret = aesni_mb_set_session_aead_parameters(mb_ops, sess, - aead_xform); - if (ret != 0) { - AESNI_MB_LOG(ERR, "Invalid/unsupported aead parameters"); - return ret; - } - } - - return 0; -} - -/** - * burst enqueue, place crypto operations on ingress queue for processing. - * - * @param __qp Queue Pair to process - * @param ops Crypto operations for processing - * @param nb_ops Number of crypto operations for processing - * - * @return - * - Number of crypto operations enqueued - */ -static uint16_t -aesni_mb_pmd_enqueue_burst(void *__qp, struct rte_crypto_op **ops, - uint16_t nb_ops) -{ - struct aesni_mb_qp *qp = __qp; - - unsigned int nb_enqueued; - - nb_enqueued = rte_ring_enqueue_burst(qp->ingress_queue, - (void **)ops, nb_ops, NULL); - - qp->stats.enqueued_count += nb_enqueued; - - return nb_enqueued; -} - -/** Get multi buffer session */ -static inline struct aesni_mb_session * -get_session(struct aesni_mb_qp *qp, struct rte_crypto_op *op) -{ - struct aesni_mb_session *sess = NULL; - - if (op->sess_type == RTE_CRYPTO_OP_WITH_SESSION) { - if (likely(op->sym->session != NULL)) - sess = (struct aesni_mb_session *) - get_sym_session_private_data( - op->sym->session, - cryptodev_driver_id); - } else { - void *_sess = NULL; - void *_sess_private_data = NULL; - - if (rte_mempool_get(qp->sess_mp, (void **)&_sess)) - return NULL; - - if (rte_mempool_get(qp->sess_mp_priv, (void **)&_sess_private_data)) - return NULL; - - sess = (struct aesni_mb_session *)_sess_private_data; - - if (unlikely(aesni_mb_set_session_parameters(qp->op_fns, - sess, op->sym->xform) != 0)) { - rte_mempool_put(qp->sess_mp, _sess); - rte_mempool_put(qp->sess_mp_priv, _sess_private_data); - sess = NULL; - } - op->sym->session = (struct rte_cryptodev_sym_session *)_sess; - set_sym_session_private_data(op->sym->session, - cryptodev_driver_id, _sess_private_data); - } - - if (unlikely(sess == NULL)) - op->status = RTE_CRYPTO_OP_STATUS_INVALID_SESSION; - - return sess; -} - -/** - * Process a crypto operation and complete a JOB_AES_HMAC job structure for - * submission to the multi buffer library for processing. - * - * @param qp queue pair - * @param job JOB_AES_HMAC structure to fill - * @param m mbuf to process - * - * @return - * - Completed JOB_AES_HMAC structure pointer on success - * - NULL pointer if completion of JOB_AES_HMAC structure isn't possible - */ -static inline int -set_mb_job_params(JOB_AES_HMAC *job, struct aesni_mb_qp *qp, - struct rte_crypto_op *op, uint8_t *digest_idx) -{ - struct rte_mbuf *m_src = op->sym->m_src, *m_dst; - struct aesni_mb_session *session; - uint16_t m_offset = 0; - - session = get_session(qp, op); - if (session == NULL) { - op->status = RTE_CRYPTO_OP_STATUS_INVALID_SESSION; - return -1; - } - - /* Set crypto operation */ - job->chain_order = session->chain_order; - - /* Set cipher parameters */ - job->cipher_direction = session->cipher.direction; - job->cipher_mode = session->cipher.mode; - - job->aes_key_len_in_bytes = session->cipher.key_length_in_bytes; - - /* Set authentication parameters */ - job->hash_alg = session->auth.algo; - - switch (job->hash_alg) { - case AES_XCBC: - job->u.XCBC._k1_expanded = session->auth.xcbc.k1_expanded; - job->u.XCBC._k2 = session->auth.xcbc.k2; - job->u.XCBC._k3 = session->auth.xcbc.k3; - - job->aes_enc_key_expanded = - session->cipher.expanded_aes_keys.encode; - job->aes_dec_key_expanded = - session->cipher.expanded_aes_keys.decode; - break; - - case AES_CCM: - job->u.CCM.aad = op->sym->aead.aad.data + 18; - job->u.CCM.aad_len_in_bytes = session->aead.aad_len; - job->aes_enc_key_expanded = - session->cipher.expanded_aes_keys.encode; - job->aes_dec_key_expanded = - session->cipher.expanded_aes_keys.decode; - break; - - case AES_CMAC: - job->u.CMAC._key_expanded = session->auth.cmac.expkey; - job->u.CMAC._skey1 = session->auth.cmac.skey1; - job->u.CMAC._skey2 = session->auth.cmac.skey2; - job->aes_enc_key_expanded = - session->cipher.expanded_aes_keys.encode; - job->aes_dec_key_expanded = - session->cipher.expanded_aes_keys.decode; - break; - - case AES_GMAC: - if (session->cipher.mode == GCM) { - job->u.GCM.aad = op->sym->aead.aad.data; - job->u.GCM.aad_len_in_bytes = session->aead.aad_len; - } else { - /* For GMAC */ - job->u.GCM.aad = rte_pktmbuf_mtod_offset(m_src, - uint8_t *, op->sym->auth.data.offset); - job->u.GCM.aad_len_in_bytes = op->sym->auth.data.length; - job->cipher_mode = GCM; - } - job->aes_enc_key_expanded = &session->cipher.gcm_key; - job->aes_dec_key_expanded = &session->cipher.gcm_key; - break; - - default: - job->u.HMAC._hashed_auth_key_xor_ipad = session->auth.pads.inner; - job->u.HMAC._hashed_auth_key_xor_opad = session->auth.pads.outer; - - if (job->cipher_mode == DES3) { - job->aes_enc_key_expanded = - session->cipher.exp_3des_keys.ks_ptr; - job->aes_dec_key_expanded = - session->cipher.exp_3des_keys.ks_ptr; - } else { - job->aes_enc_key_expanded = - session->cipher.expanded_aes_keys.encode; - job->aes_dec_key_expanded = - session->cipher.expanded_aes_keys.decode; - } - } - - /* Mutable crypto operation parameters */ - if (op->sym->m_dst) { - m_src = m_dst = op->sym->m_dst; - - /* append space for output data to mbuf */ - char *odata = rte_pktmbuf_append(m_dst, - rte_pktmbuf_data_len(op->sym->m_src)); - if (odata == NULL) { - AESNI_MB_LOG(ERR, "failed to allocate space in destination " - "mbuf for source data"); - op->status = RTE_CRYPTO_OP_STATUS_ERROR; - return -1; - } - - memcpy(odata, rte_pktmbuf_mtod(op->sym->m_src, void*), - rte_pktmbuf_data_len(op->sym->m_src)); - } else { - m_dst = m_src; - if (job->hash_alg == AES_CCM || (job->hash_alg == AES_GMAC && - session->cipher.mode == GCM)) - m_offset = op->sym->aead.data.offset; - else - m_offset = op->sym->cipher.data.offset; - } - - /* Set digest output location */ - if (job->hash_alg != NULL_HASH && - session->auth.operation == RTE_CRYPTO_AUTH_OP_VERIFY) { - job->auth_tag_output = qp->temp_digests[*digest_idx]; - *digest_idx = (*digest_idx + 1) % MAX_JOBS; - } else { - if (job->hash_alg == AES_CCM || (job->hash_alg == AES_GMAC && - session->cipher.mode == GCM)) - job->auth_tag_output = op->sym->aead.digest.data; - else - job->auth_tag_output = op->sym->auth.digest.data; - - if (session->auth.req_digest_len != session->auth.gen_digest_len) { - job->auth_tag_output = qp->temp_digests[*digest_idx]; - *digest_idx = (*digest_idx + 1) % MAX_JOBS; - } - } - /* - * Multi-buffer library current only support returning a truncated - * digest length as specified in the relevant IPsec RFCs - */ - - /* Set digest length */ - job->auth_tag_output_len_in_bytes = session->auth.gen_digest_len; - - /* Set IV parameters */ - job->iv_len_in_bytes = session->iv.length; - - /* Data Parameter */ - job->src = rte_pktmbuf_mtod(m_src, uint8_t *); - job->dst = rte_pktmbuf_mtod_offset(m_dst, uint8_t *, m_offset); - - switch (job->hash_alg) { - case AES_CCM: - job->cipher_start_src_offset_in_bytes = - op->sym->aead.data.offset; - job->msg_len_to_cipher_in_bytes = op->sym->aead.data.length; - job->hash_start_src_offset_in_bytes = op->sym->aead.data.offset; - job->msg_len_to_hash_in_bytes = op->sym->aead.data.length; - - job->iv = rte_crypto_op_ctod_offset(op, uint8_t *, - session->iv.offset + 1); - break; - - case AES_GMAC: - if (session->cipher.mode == GCM) { - job->cipher_start_src_offset_in_bytes = - op->sym->aead.data.offset; - job->hash_start_src_offset_in_bytes = - op->sym->aead.data.offset; - job->msg_len_to_cipher_in_bytes = - op->sym->aead.data.length; - job->msg_len_to_hash_in_bytes = - op->sym->aead.data.length; - } else { - job->cipher_start_src_offset_in_bytes = - op->sym->auth.data.offset; - job->hash_start_src_offset_in_bytes = - op->sym->auth.data.offset; - job->msg_len_to_cipher_in_bytes = 0; - job->msg_len_to_hash_in_bytes = 0; - } - job->iv = rte_crypto_op_ctod_offset(op, uint8_t *, - session->iv.offset); - - break; - - default: - job->cipher_start_src_offset_in_bytes = - op->sym->cipher.data.offset; - job->msg_len_to_cipher_in_bytes = op->sym->cipher.data.length; - - job->hash_start_src_offset_in_bytes = op->sym->auth.data.offset; - job->msg_len_to_hash_in_bytes = op->sym->auth.data.length; - - job->iv = rte_crypto_op_ctod_offset(op, uint8_t *, - session->iv.offset); - } - - /* Set user data to be crypto operation data struct */ - job->user_data = op; - - return 0; -} - -static inline void -verify_digest(JOB_AES_HMAC *job, void *digest, uint16_t len, uint8_t *status) -{ - if (memcmp(job->auth_tag_output, digest, len) != 0) - *status = RTE_CRYPTO_OP_STATUS_AUTH_FAILED; -} - -static inline void -generate_digest(JOB_AES_HMAC *job, struct rte_crypto_op *op, - struct aesni_mb_session *sess) -{ - /* No extra copy neeed */ - if (likely(sess->auth.req_digest_len == sess->auth.gen_digest_len)) - return; - - /* - * This can only happen for HMAC, so only digest - * for authentication algos is required - */ - memcpy(op->sym->auth.digest.data, job->auth_tag_output, - sess->auth.req_digest_len); -} - -/** - * Process a completed job and return rte_mbuf which job processed - * - * @param qp Queue Pair to process - * @param job JOB_AES_HMAC job to process - * - * @return - * - Returns processed crypto operation. - * - Returns NULL on invalid job - */ -static inline struct rte_crypto_op * -post_process_mb_job(struct aesni_mb_qp *qp, JOB_AES_HMAC *job) -{ - struct rte_crypto_op *op = (struct rte_crypto_op *)job->user_data; - struct aesni_mb_session *sess = get_sym_session_private_data( - op->sym->session, - cryptodev_driver_id); - - if (likely(op->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED)) { - switch (job->status) { - case STS_COMPLETED: - op->status = RTE_CRYPTO_OP_STATUS_SUCCESS; - - if (job->hash_alg == NULL_HASH) - break; - - if (sess->auth.operation == RTE_CRYPTO_AUTH_OP_VERIFY) { - if (job->hash_alg == AES_CCM || - (job->hash_alg == AES_GMAC && - sess->cipher.mode == GCM)) - verify_digest(job, - op->sym->aead.digest.data, - sess->auth.req_digest_len, - &op->status); - else - verify_digest(job, - op->sym->auth.digest.data, - sess->auth.req_digest_len, - &op->status); - } else - generate_digest(job, op, sess); - break; - default: - op->status = RTE_CRYPTO_OP_STATUS_ERROR; - } - } - - /* Free session if a session-less crypto op */ - if (op->sess_type == RTE_CRYPTO_OP_SESSIONLESS) { - memset(sess, 0, sizeof(struct aesni_mb_session)); - memset(op->sym->session, 0, - rte_cryptodev_sym_get_existing_header_session_size( - op->sym->session)); - rte_mempool_put(qp->sess_mp_priv, sess); - rte_mempool_put(qp->sess_mp, op->sym->session); - op->sym->session = NULL; - } - - return op; -} - -/** - * Process a completed JOB_AES_HMAC job and keep processing jobs until - * get_completed_job return NULL - * - * @param qp Queue Pair to process - * @param job JOB_AES_HMAC job - * - * @return - * - Number of processed jobs - */ -static unsigned -handle_completed_jobs(struct aesni_mb_qp *qp, JOB_AES_HMAC *job, - struct rte_crypto_op **ops, uint16_t nb_ops) -{ - struct rte_crypto_op *op = NULL; - unsigned processed_jobs = 0; - - while (job != NULL) { - op = post_process_mb_job(qp, job); - - if (op) { - ops[processed_jobs++] = op; - qp->stats.dequeued_count++; - } else { - qp->stats.dequeue_err_count++; - break; - } - if (processed_jobs == nb_ops) - break; - - job = (*qp->op_fns->job.get_completed_job)(qp->mb_mgr); - } - - return processed_jobs; -} - -static inline uint16_t -flush_mb_mgr(struct aesni_mb_qp *qp, struct rte_crypto_op **ops, - uint16_t nb_ops) -{ - int processed_ops = 0; - - /* Flush the remaining jobs */ - JOB_AES_HMAC *job = (*qp->op_fns->job.flush_job)(qp->mb_mgr); - - if (job) - processed_ops += handle_completed_jobs(qp, job, - &ops[processed_ops], nb_ops - processed_ops); - - return processed_ops; -} - -static inline JOB_AES_HMAC * -set_job_null_op(JOB_AES_HMAC *job, struct rte_crypto_op *op) -{ - job->chain_order = HASH_CIPHER; - job->cipher_mode = NULL_CIPHER; - job->hash_alg = NULL_HASH; - job->cipher_direction = DECRYPT; - - /* Set user data to be crypto operation data struct */ - job->user_data = op; - - return job; -} - -static uint16_t -aesni_mb_pmd_dequeue_burst(void *queue_pair, struct rte_crypto_op **ops, - uint16_t nb_ops) -{ - struct aesni_mb_qp *qp = queue_pair; - - struct rte_crypto_op *op; - JOB_AES_HMAC *job; - - int retval, processed_jobs = 0; - - if (unlikely(nb_ops == 0)) - return 0; - - uint8_t digest_idx = qp->digest_idx; - do { - /* Get next free mb job struct from mb manager */ - job = (*qp->op_fns->job.get_next)(qp->mb_mgr); - if (unlikely(job == NULL)) { - /* if no free mb job structs we need to flush mb_mgr */ - processed_jobs += flush_mb_mgr(qp, - &ops[processed_jobs], - nb_ops - processed_jobs); - - if (nb_ops == processed_jobs) - break; - - job = (*qp->op_fns->job.get_next)(qp->mb_mgr); - } - - /* - * Get next operation to process from ingress queue. - * There is no need to return the job to the MB_MGR - * if there are no more operations to process, since the MB_MGR - * can use that pointer again in next get_next calls. - */ - retval = rte_ring_dequeue(qp->ingress_queue, (void **)&op); - if (retval < 0) - break; - - retval = set_mb_job_params(job, qp, op, &digest_idx); - if (unlikely(retval != 0)) { - qp->stats.dequeue_err_count++; - set_job_null_op(job, op); - } - - /* Submit job to multi-buffer for processing */ - job = (*qp->op_fns->job.submit)(qp->mb_mgr); - - /* - * If submit returns a processed job then handle it, - * before submitting subsequent jobs - */ - if (job) - processed_jobs += handle_completed_jobs(qp, job, - &ops[processed_jobs], - nb_ops - processed_jobs); - - } while (processed_jobs < nb_ops); - - qp->digest_idx = digest_idx; - - if (processed_jobs < 1) - processed_jobs += flush_mb_mgr(qp, - &ops[processed_jobs], - nb_ops - processed_jobs); - - return processed_jobs; -} - -static int cryptodev_aesni_mb_remove(struct rte_vdev_device *vdev); - -static int -cryptodev_aesni_mb_create(const char *name, - struct rte_vdev_device *vdev, - struct rte_cryptodev_pmd_init_params *init_params) -{ - struct rte_cryptodev *dev; - struct aesni_mb_private *internals; - enum aesni_mb_vector_mode vector_mode; - - /* Check CPU for support for AES instruction set */ - if (!rte_cpu_get_flag_enabled(RTE_CPUFLAG_AES)) { - AESNI_MB_LOG(ERR, "AES instructions not supported by CPU"); - return -EFAULT; - } - - dev = rte_cryptodev_pmd_create(name, &vdev->device, init_params); - if (dev == NULL) { - AESNI_MB_LOG(ERR, "failed to create cryptodev vdev"); - return -ENODEV; - } - - /* Check CPU for supported vector instruction set */ - if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_AVX512F)) - vector_mode = RTE_AESNI_MB_AVX512; - else if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_AVX2)) - vector_mode = RTE_AESNI_MB_AVX2; - else if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_AVX)) - vector_mode = RTE_AESNI_MB_AVX; - else - vector_mode = RTE_AESNI_MB_SSE; - - dev->driver_id = cryptodev_driver_id; - dev->dev_ops = rte_aesni_mb_pmd_ops; - - /* register rx/tx burst functions for data path */ - dev->dequeue_burst = aesni_mb_pmd_dequeue_burst; - dev->enqueue_burst = aesni_mb_pmd_enqueue_burst; - - dev->feature_flags = RTE_CRYPTODEV_FF_SYMMETRIC_CRYPTO | - RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING | - RTE_CRYPTODEV_FF_CPU_AESNI; - - switch (vector_mode) { - case RTE_AESNI_MB_SSE: - dev->feature_flags |= RTE_CRYPTODEV_FF_CPU_SSE; - break; - case RTE_AESNI_MB_AVX: - dev->feature_flags |= RTE_CRYPTODEV_FF_CPU_AVX; - break; - case RTE_AESNI_MB_AVX2: - dev->feature_flags |= RTE_CRYPTODEV_FF_CPU_AVX2; - break; - case RTE_AESNI_MB_AVX512: - dev->feature_flags |= RTE_CRYPTODEV_FF_CPU_AVX512; - break; - default: - break; - } - - /* Set vector instructions mode supported */ - internals = dev->data->dev_private; - - internals->vector_mode = vector_mode; - internals->max_nb_queue_pairs = init_params->max_nb_queue_pairs; - -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - AESNI_MB_LOG(INFO, "IPSec Multi-buffer library version used: %s\n", - imb_get_version_str()); -#else - AESNI_MB_LOG(INFO, "IPSec Multi-buffer library version used: 0.49.0\n"); -#endif - - return 0; -} - -static int -cryptodev_aesni_mb_probe(struct rte_vdev_device *vdev) -{ - struct rte_cryptodev_pmd_init_params init_params = { - "", - sizeof(struct aesni_mb_private), - rte_socket_id(), - RTE_CRYPTODEV_PMD_DEFAULT_MAX_NB_QUEUE_PAIRS - }; - const char *name, *args; - int retval; - - name = rte_vdev_device_name(vdev); - if (name == NULL) - return -EINVAL; - - args = rte_vdev_device_args(vdev); - - retval = rte_cryptodev_pmd_parse_input_args(&init_params, args); - if (retval) { - AESNI_MB_LOG(ERR, "Failed to parse initialisation arguments[%s]", - args); - return -EINVAL; - } - - return cryptodev_aesni_mb_create(name, vdev, &init_params); -} - -static int -cryptodev_aesni_mb_remove(struct rte_vdev_device *vdev) -{ - struct rte_cryptodev *cryptodev; - const char *name; - - name = rte_vdev_device_name(vdev); - if (name == NULL) - return -EINVAL; - - cryptodev = rte_cryptodev_pmd_get_named_dev(name); - if (cryptodev == NULL) - return -ENODEV; - - return rte_cryptodev_pmd_destroy(cryptodev); -} - -static struct rte_vdev_driver cryptodev_aesni_mb_pmd_drv = { - .probe = cryptodev_aesni_mb_probe, - .remove = cryptodev_aesni_mb_remove -}; - -static struct cryptodev_driver aesni_mb_crypto_drv; - -RTE_PMD_REGISTER_VDEV(CRYPTODEV_NAME_AESNI_MB_PMD, cryptodev_aesni_mb_pmd_drv); -RTE_PMD_REGISTER_ALIAS(CRYPTODEV_NAME_AESNI_MB_PMD, cryptodev_aesni_mb_pmd); -RTE_PMD_REGISTER_PARAM_STRING(CRYPTODEV_NAME_AESNI_MB_PMD, - "max_nb_queue_pairs= " - "socket_id="); -RTE_PMD_REGISTER_CRYPTO_DRIVER(aesni_mb_crypto_drv, - cryptodev_aesni_mb_pmd_drv.driver, - cryptodev_driver_id); - -RTE_INIT(aesni_mb_init_log) -{ - aesni_mb_logtype_driver = rte_log_register("pmd.crypto.aesni_mb"); -} diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops_compat.c b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops_compat.c deleted file mode 100644 index 79a38b25e..000000000 --- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops_compat.c +++ /dev/null @@ -1,745 +0,0 @@ -/* SPDX-License-Identifier: BSD-3-Clause - * Copyright(c) 2015-2017 Intel Corporation - */ - -#include - -#include -#include -#include - -#include "rte_aesni_mb_pmd_private.h" - - -static const struct rte_cryptodev_capabilities aesni_mb_pmd_capabilities[] = { - { /* MD5 HMAC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH, - {.auth = { - .algo = RTE_CRYPTO_AUTH_MD5_HMAC, - .block_size = 64, - .key_size = { - .min = 1, - .max = 64, - .increment = 1 - }, - .digest_size = { -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .min = 1, - .max = 16, - .increment = 1 -#else - .min = 12, - .max = 12, - .increment = 0 -#endif - }, - .iv_size = { 0 } - }, } - }, } - }, - { /* SHA1 HMAC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH, - {.auth = { - .algo = RTE_CRYPTO_AUTH_SHA1_HMAC, - .block_size = 64, - .key_size = { - .min = 1, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .max = 65535, -#else - .max = 64, -#endif - .increment = 1 - }, - .digest_size = { -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .min = 1, - .max = 20, - .increment = 1 -#else - .min = 12, - .max = 12, - .increment = 0 -#endif - }, - .iv_size = { 0 } - }, } - }, } - }, - { /* SHA224 HMAC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH, - {.auth = { - .algo = RTE_CRYPTO_AUTH_SHA224_HMAC, - .block_size = 64, - .key_size = { - .min = 1, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .max = 65535, -#else - .max = 64, -#endif - .increment = 1 - }, - .digest_size = { -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .min = 1, - .max = 28, - .increment = 1 -#else - .min = 14, - .max = 14, - .increment = 0 -#endif - }, - .iv_size = { 0 } - }, } - }, } - }, - { /* SHA256 HMAC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH, - {.auth = { - .algo = RTE_CRYPTO_AUTH_SHA256_HMAC, - .block_size = 64, - .key_size = { - .min = 1, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .max = 65535, -#else - .max = 64, -#endif - .increment = 1 - }, - .digest_size = { -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .min = 1, - .max = 32, - .increment = 1 -#else - .min = 16, - .max = 16, - .increment = 0 -#endif - }, - .iv_size = { 0 } - }, } - }, } - }, - { /* SHA384 HMAC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH, - {.auth = { - .algo = RTE_CRYPTO_AUTH_SHA384_HMAC, - .block_size = 128, - .key_size = { - .min = 1, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .max = 65535, -#else - .max = 128, -#endif - .increment = 1 - }, - .digest_size = { -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .min = 1, - .max = 48, - .increment = 1 -#else - .min = 24, - .max = 24, - .increment = 0 -#endif - }, - .iv_size = { 0 } - }, } - }, } - }, - { /* SHA512 HMAC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH, - {.auth = { - .algo = RTE_CRYPTO_AUTH_SHA512_HMAC, - .block_size = 128, - .key_size = { - .min = 1, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .max = 65535, -#else - .max = 128, -#endif - .increment = 1 - }, - .digest_size = { -#if IMB_VERSION_NUM >= IMB_VERSION(0, 50, 0) - .min = 1, - .max = 64, - .increment = 1 -#else - .min = 32, - .max = 32, - .increment = 0 -#endif - }, - .iv_size = { 0 } - }, } - }, } - }, - { /* AES XCBC HMAC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH, - {.auth = { - .algo = RTE_CRYPTO_AUTH_AES_XCBC_MAC, - .block_size = 16, - .key_size = { - .min = 16, - .max = 16, - .increment = 0 - }, - .digest_size = { - .min = 12, - .max = 12, - .increment = 0 - }, - .iv_size = { 0 } - }, } - }, } - }, - { /* AES CBC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER, - {.cipher = { - .algo = RTE_CRYPTO_CIPHER_AES_CBC, - .block_size = 16, - .key_size = { - .min = 16, - .max = 32, - .increment = 8 - }, - .iv_size = { - .min = 16, - .max = 16, - .increment = 0 - } - }, } - }, } - }, - { /* AES CTR */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER, - {.cipher = { - .algo = RTE_CRYPTO_CIPHER_AES_CTR, - .block_size = 16, - .key_size = { - .min = 16, - .max = 32, - .increment = 8 - }, - .iv_size = { - .min = 12, - .max = 16, - .increment = 4 - } - }, } - }, } - }, - { /* AES DOCSIS BPI */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER, - {.cipher = { - .algo = RTE_CRYPTO_CIPHER_AES_DOCSISBPI, - .block_size = 16, - .key_size = { - .min = 16, - .max = 16, - .increment = 0 - }, - .iv_size = { - .min = 16, - .max = 16, - .increment = 0 - } - }, } - }, } - }, - { /* DES CBC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER, - {.cipher = { - .algo = RTE_CRYPTO_CIPHER_DES_CBC, - .block_size = 8, - .key_size = { - .min = 8, - .max = 8, - .increment = 0 - }, - .iv_size = { - .min = 8, - .max = 8, - .increment = 0 - } - }, } - }, } - }, - { /* 3DES CBC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER, - {.cipher = { - .algo = RTE_CRYPTO_CIPHER_3DES_CBC, - .block_size = 8, - .key_size = { - .min = 8, - .max = 24, - .increment = 8 - }, - .iv_size = { - .min = 8, - .max = 8, - .increment = 0 - } - }, } - }, } - }, - { /* DES DOCSIS BPI */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER, - {.cipher = { - .algo = RTE_CRYPTO_CIPHER_DES_DOCSISBPI, - .block_size = 8, - .key_size = { - .min = 8, - .max = 8, - .increment = 0 - }, - .iv_size = { - .min = 8, - .max = 8, - .increment = 0 - } - }, } - }, } - }, - { /* AES CCM */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AEAD, - {.aead = { - .algo = RTE_CRYPTO_AEAD_AES_CCM, - .block_size = 16, - .key_size = { - .min = 16, - .max = 16, - .increment = 0 - }, - .digest_size = { - .min = 4, - .max = 16, - .increment = 2 - }, - .aad_size = { - .min = 0, - .max = 46, - .increment = 1 - }, - .iv_size = { - .min = 7, - .max = 13, - .increment = 1 - }, - }, } - }, } - }, - { /* AES CMAC */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH, - {.auth = { - .algo = RTE_CRYPTO_AUTH_AES_CMAC, - .block_size = 16, - .key_size = { - .min = 16, - .max = 16, - .increment = 0 - }, - .digest_size = { - .min = 1, - .max = 16, - .increment = 1 - }, - .iv_size = { 0 } - }, } - }, } - }, - { /* AES GCM */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AEAD, - {.aead = { - .algo = RTE_CRYPTO_AEAD_AES_GCM, - .block_size = 16, - .key_size = { - .min = 16, - .max = 32, - .increment = 8 - }, - .digest_size = { - .min = 8, - .max = 16, - .increment = 4 - }, - .aad_size = { - .min = 0, - .max = 65535, - .increment = 1 - }, - .iv_size = { - .min = 12, - .max = 12, - .increment = 0 - } - }, } - }, } - }, - { /* AES GMAC (AUTH) */ - .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, - {.sym = { - .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH, - {.auth = { - .algo = RTE_CRYPTO_AUTH_AES_GMAC, - .block_size = 16, - .key_size = { - .min = 16, - .max = 32, - .increment = 8 - }, - .digest_size = { - .min = 8, - .max = 16, - .increment = 4 - }, - .iv_size = { - .min = 12, - .max = 12, - .increment = 0 - } - }, } - }, } - }, - RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST() -}; - - -/** Configure device */ -static int -aesni_mb_pmd_config(__rte_unused struct rte_cryptodev *dev, - __rte_unused struct rte_cryptodev_config *config) -{ - return 0; -} - -/** Start device */ -static int -aesni_mb_pmd_start(__rte_unused struct rte_cryptodev *dev) -{ - return 0; -} - -/** Stop device */ -static void -aesni_mb_pmd_stop(__rte_unused struct rte_cryptodev *dev) -{ -} - -/** Close device */ -static int -aesni_mb_pmd_close(__rte_unused struct rte_cryptodev *dev) -{ - return 0; -} - - -/** Get device statistics */ -static void -aesni_mb_pmd_stats_get(struct rte_cryptodev *dev, - struct rte_cryptodev_stats *stats) -{ - int qp_id; - - for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { - struct aesni_mb_qp *qp = dev->data->queue_pairs[qp_id]; - - stats->enqueued_count += qp->stats.enqueued_count; - stats->dequeued_count += qp->stats.dequeued_count; - - stats->enqueue_err_count += qp->stats.enqueue_err_count; - stats->dequeue_err_count += qp->stats.dequeue_err_count; - } -} - -/** Reset device statistics */ -static void -aesni_mb_pmd_stats_reset(struct rte_cryptodev *dev) -{ - int qp_id; - - for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { - struct aesni_mb_qp *qp = dev->data->queue_pairs[qp_id]; - - memset(&qp->stats, 0, sizeof(qp->stats)); - } -} - - -/** Get device info */ -static void -aesni_mb_pmd_info_get(struct rte_cryptodev *dev, - struct rte_cryptodev_info *dev_info) -{ - struct aesni_mb_private *internals = dev->data->dev_private; - - if (dev_info != NULL) { - dev_info->driver_id = dev->driver_id; - dev_info->feature_flags = dev->feature_flags; - dev_info->capabilities = aesni_mb_pmd_capabilities; - dev_info->max_nb_queue_pairs = internals->max_nb_queue_pairs; - /* No limit of number of sessions */ - dev_info->sym.max_nb_sessions = 0; - } -} - -/** Release queue pair */ -static int -aesni_mb_pmd_qp_release(struct rte_cryptodev *dev, uint16_t qp_id) -{ - struct aesni_mb_qp *qp = dev->data->queue_pairs[qp_id]; - struct rte_ring *r = NULL; - - if (qp != NULL) { - r = rte_ring_lookup(qp->name); - if (r) - rte_ring_free(r); - if (qp->mb_mgr) - free_mb_mgr(qp->mb_mgr); - rte_free(qp); - dev->data->queue_pairs[qp_id] = NULL; - } - return 0; -} - -/** set a unique name for the queue pair based on it's name, dev_id and qp_id */ -static int -aesni_mb_pmd_qp_set_unique_name(struct rte_cryptodev *dev, - struct aesni_mb_qp *qp) -{ - unsigned n = snprintf(qp->name, sizeof(qp->name), - "aesni_mb_pmd_%u_qp_%u", - dev->data->dev_id, qp->id); - - if (n >= sizeof(qp->name)) - return -1; - - return 0; -} - -/** Create a ring to place processed operations on */ -static struct rte_ring * -aesni_mb_pmd_qp_create_processed_ops_ring(struct aesni_mb_qp *qp, - unsigned int ring_size, int socket_id) -{ - struct rte_ring *r; - char ring_name[RTE_CRYPTODEV_NAME_MAX_LEN]; - - unsigned int n = snprintf(ring_name, sizeof(ring_name), "%s", qp->name); - - if (n >= sizeof(ring_name)) - return NULL; - - r = rte_ring_lookup(ring_name); - if (r) { - if (rte_ring_get_size(r) >= ring_size) { - AESNI_MB_LOG(INFO, "Reusing existing ring %s for processed ops", - ring_name); - return r; - } - - AESNI_MB_LOG(ERR, "Unable to reuse existing ring %s for processed ops", - ring_name); - return NULL; - } - - return rte_ring_create(ring_name, ring_size, socket_id, - RING_F_SP_ENQ | RING_F_SC_DEQ); -} - -/** Setup a queue pair */ -static int -aesni_mb_pmd_qp_setup(struct rte_cryptodev *dev, uint16_t qp_id, - const struct rte_cryptodev_qp_conf *qp_conf, - int socket_id) -{ - struct aesni_mb_qp *qp = NULL; - struct aesni_mb_private *internals = dev->data->dev_private; - int ret = -1; - - /* Free memory prior to re-allocation if needed. */ - if (dev->data->queue_pairs[qp_id] != NULL) - aesni_mb_pmd_qp_release(dev, qp_id); - - /* Allocate the queue pair data structure. */ - qp = rte_zmalloc_socket("AES-NI PMD Queue Pair", sizeof(*qp), - RTE_CACHE_LINE_SIZE, socket_id); - if (qp == NULL) - return -ENOMEM; - - qp->id = qp_id; - dev->data->queue_pairs[qp_id] = qp; - - if (aesni_mb_pmd_qp_set_unique_name(dev, qp)) - goto qp_setup_cleanup; - - - qp->mb_mgr = alloc_mb_mgr(0); - if (qp->mb_mgr == NULL) { - ret = -ENOMEM; - goto qp_setup_cleanup; - } - - qp->op_fns = &job_ops[internals->vector_mode]; - - qp->ingress_queue = aesni_mb_pmd_qp_create_processed_ops_ring(qp, - qp_conf->nb_descriptors, socket_id); - if (qp->ingress_queue == NULL) { - ret = -1; - goto qp_setup_cleanup; - } - - qp->sess_mp = qp_conf->mp_session; - qp->sess_mp_priv = qp_conf->mp_session_private; - - memset(&qp->stats, 0, sizeof(qp->stats)); - - char mp_name[RTE_MEMPOOL_NAMESIZE]; - - snprintf(mp_name, RTE_MEMPOOL_NAMESIZE, - "digest_mp_%u_%u", dev->data->dev_id, qp_id); - - /* Initialise multi-buffer manager */ - (*qp->op_fns->job.init_mgr)(qp->mb_mgr); - return 0; - -qp_setup_cleanup: - if (qp) { - if (qp->mb_mgr == NULL) - free_mb_mgr(qp->mb_mgr); - rte_free(qp); - } - - return ret; -} - -/** Return the number of allocated queue pairs */ -static uint32_t -aesni_mb_pmd_qp_count(struct rte_cryptodev *dev) -{ - return dev->data->nb_queue_pairs; -} - -/** Returns the size of the aesni multi-buffer session structure */ -static unsigned -aesni_mb_pmd_sym_session_get_size(struct rte_cryptodev *dev __rte_unused) -{ - return sizeof(struct aesni_mb_session); -} - -/** Configure a aesni multi-buffer session from a crypto xform chain */ -static int -aesni_mb_pmd_sym_session_configure(struct rte_cryptodev *dev, - struct rte_crypto_sym_xform *xform, - struct rte_cryptodev_sym_session *sess, - struct rte_mempool *mempool) -{ - void *sess_private_data; - struct aesni_mb_private *internals = dev->data->dev_private; - int ret; - - if (unlikely(sess == NULL)) { - AESNI_MB_LOG(ERR, "invalid session struct"); - return -EINVAL; - } - - if (rte_mempool_get(mempool, &sess_private_data)) { - AESNI_MB_LOG(ERR, - "Couldn't get object from session mempool"); - return -ENOMEM; - } - - ret = aesni_mb_set_session_parameters(&job_ops[internals->vector_mode], - sess_private_data, xform); - if (ret != 0) { - AESNI_MB_LOG(ERR, "failed configure session parameters"); - - /* Return session to mempool */ - rte_mempool_put(mempool, sess_private_data); - return ret; - } - - set_sym_session_private_data(sess, dev->driver_id, - sess_private_data); - - return 0; -} - -/** Clear the memory of session so it doesn't leave key material behind */ -static void -aesni_mb_pmd_sym_session_clear(struct rte_cryptodev *dev, - struct rte_cryptodev_sym_session *sess) -{ - uint8_t index = dev->driver_id; - void *sess_priv = get_sym_session_private_data(sess, index); - - /* Zero out the whole structure */ - if (sess_priv) { - memset(sess_priv, 0, sizeof(struct aesni_mb_session)); - struct rte_mempool *sess_mp = rte_mempool_from_obj(sess_priv); - set_sym_session_private_data(sess, index, NULL); - rte_mempool_put(sess_mp, sess_priv); - } -} - -struct rte_cryptodev_ops aesni_mb_pmd_ops = { - .dev_configure = aesni_mb_pmd_config, - .dev_start = aesni_mb_pmd_start, - .dev_stop = aesni_mb_pmd_stop, - .dev_close = aesni_mb_pmd_close, - - .stats_get = aesni_mb_pmd_stats_get, - .stats_reset = aesni_mb_pmd_stats_reset, - - .dev_infos_get = aesni_mb_pmd_info_get, - - .queue_pair_setup = aesni_mb_pmd_qp_setup, - .queue_pair_release = aesni_mb_pmd_qp_release, - .queue_pair_count = aesni_mb_pmd_qp_count, - - .sym_session_get_size = aesni_mb_pmd_sym_session_get_size, - .sym_session_configure = aesni_mb_pmd_sym_session_configure, - .sym_session_clear = aesni_mb_pmd_sym_session_clear -}; - -struct rte_cryptodev_ops *rte_aesni_mb_pmd_ops = &aesni_mb_pmd_ops; diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_private.h b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_private.h index 61f419dda..4d439360f 100644 --- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_private.h +++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_private.h @@ -7,21 +7,6 @@ #include - -/* - * IMB_VERSION_NUM macro was introduced in version Multi-buffer 0.50, - * so if macro is not defined, it means that the version is 0.49. - */ -#if !defined(IMB_VERSION_NUM) -#define IMB_VERSION(a, b, c) (((a) << 16) + ((b) << 8) + (c)) -#define IMB_VERSION_NUM IMB_VERSION(0, 49, 0) -#endif - -#if IMB_VERSION_NUM < IMB_VERSION(0, 52, 0) -#include "aesni_mb_ops.h" -#endif - -#if IMB_VERSION_NUM >= IMB_VERSION(0, 52, 0) enum aesni_mb_vector_mode { RTE_AESNI_MB_NOT_SUPPORTED = 0, RTE_AESNI_MB_SSE, @@ -29,8 +14,6 @@ enum aesni_mb_vector_mode { RTE_AESNI_MB_AVX2, RTE_AESNI_MB_AVX512 }; -#endif - #define CRYPTODEV_NAME_AESNI_MB_PMD crypto_aesni_mb /**< AES-NI Multi buffer PMD device name */ @@ -109,13 +92,11 @@ static const unsigned auth_digest_byte_lengths[] = { [AES_CMAC] = 16, [AES_GMAC] = 12, [NULL_HASH] = 0, -#if IMB_VERSION_NUM >= IMB_VERSION(0, 52, 0) [PLAIN_SHA1] = 20, [PLAIN_SHA_224] = 28, [PLAIN_SHA_256] = 32, [PLAIN_SHA_384] = 48, [PLAIN_SHA_512] = 64 -#endif /**< Vector mode dependent pointer table of the multi-buffer APIs */ }; @@ -149,10 +130,8 @@ struct aesni_mb_private { /**< CPU vector instruction set mode */ unsigned max_nb_queue_pairs; /**< Max number of queue pairs supported by device */ -#if IMB_VERSION_NUM >= IMB_VERSION(0, 52, 0) MB_MGR *mb_mgr; /**< Multi-buffer instance */ -#endif }; /** AESNI Multi buffer queue pair */ @@ -160,10 +139,6 @@ struct aesni_mb_qp { uint16_t id; /**< Queue Pair Identifier */ char name[RTE_CRYPTODEV_NAME_MAX_LEN]; -#if IMB_VERSION_NUM < IMB_VERSION(0, 52, 0) - /**< Unique Queue Pair Name */ - const struct aesni_mb_op_fns *op_fns; -#endif /**< Unique Queue Pair Name */ MB_MGR *mb_mgr; /**< Multi-buffer instance */ @@ -277,22 +252,10 @@ struct aesni_mb_session { } aead; } __rte_cache_aligned; - - -#if IMB_VERSION_NUM >= IMB_VERSION(0, 52, 0) -/** - * - */ extern int aesni_mb_set_session_parameters(const MB_MGR *mb_mgr, struct aesni_mb_session *sess, const struct rte_crypto_sym_xform *xform); -#else -extern int -aesni_mb_set_session_parameters(const struct aesni_mb_op_fns *mb_ops, - struct aesni_mb_session *sess, - const struct rte_crypto_sym_xform *xform); -#endif /** device specific operations function pointer structure */ extern struct rte_cryptodev_ops *rte_aesni_mb_pmd_ops; -- 2.14.5