From: Shijith Thotton <sthotton@marvell.com>
To: <dev@dpdk.org>
Cc: <sthotton@marvell.com>, <pbhagavatula@marvell.com>,
<anoobj@marvell.com>, <jerinj@marvell.com>,
<abhinandan.gujjar@intel.com>, <adwivedi@marvell.com>,
<gakhil@marvell.com>
Subject: [dpdk-dev] [PATCH 2/2] drivers: add octeontx crypto adapter data path
Date: Tue, 22 Jun 2021 22:18:09 +0530 [thread overview]
Message-ID: <185977c5fa60dad0ea0b89e7edbb369bc58726b8.1624379833.git.sthotton@marvell.com> (raw)
In-Reply-To: <cover.1624379833.git.sthotton@marvell.com>
Added support for crypto adapter OP_FORWARD mode.
As OcteonTx CPT crypto completions could be out of order, each crypto op
is enqueued to CPT, dequeued from CPT and enqueued to SSO one-by-one.
Signed-off-by: Shijith Thotton <sthotton@marvell.com>
---
doc/guides/rel_notes/release_21_08.rst | 4 +
drivers/common/cpt/cpt_common.h | 2 +-
drivers/crypto/octeontx/meson.build | 5 +
drivers/crypto/octeontx/otx_cryptodev_ops.c | 272 +++++++++++++-----
drivers/crypto/octeontx/otx_cryptodev_ops.h | 7 +
drivers/crypto/octeontx/version.map | 9 +
drivers/event/octeontx/ssovf_evdev.c | 3 +-
drivers/event/octeontx/ssovf_worker.c | 11 +
drivers/event/octeontx/ssovf_worker.h | 25 +-
.../octeontx2/otx2_evdev_crypto_adptr_rx.h | 6 +-
10 files changed, 259 insertions(+), 85 deletions(-)
diff --git a/doc/guides/rel_notes/release_21_08.rst b/doc/guides/rel_notes/release_21_08.rst
index a6ecfdf3ce..70a91cc654 100644
--- a/doc/guides/rel_notes/release_21_08.rst
+++ b/doc/guides/rel_notes/release_21_08.rst
@@ -55,6 +55,10 @@ New Features
Also, make sure to start the actual text at the margin.
=======================================================
+* **Updated Marvell OCTEON TX PMD.**
+
+ Added support for crypto adapter OP_FORWARD mode.
+
Removed Items
-------------
diff --git a/drivers/common/cpt/cpt_common.h b/drivers/common/cpt/cpt_common.h
index 7fea0ca879..724e5ec736 100644
--- a/drivers/common/cpt/cpt_common.h
+++ b/drivers/common/cpt/cpt_common.h
@@ -54,7 +54,7 @@ struct cpt_request_info {
uint64_t ei2;
} ist;
uint8_t *rptr;
- const struct otx2_cpt_qp *qp;
+ const void *qp;
/** Control path fields */
uint64_t time_out;
diff --git a/drivers/crypto/octeontx/meson.build b/drivers/crypto/octeontx/meson.build
index 37603c5c89..3ae6729e8f 100644
--- a/drivers/crypto/octeontx/meson.build
+++ b/drivers/crypto/octeontx/meson.build
@@ -6,6 +6,7 @@ if not is_linux
endif
deps += ['bus_pci']
+deps += ['bus_vdev']
deps += ['common_cpt']
deps += ['eventdev']
@@ -18,3 +19,7 @@ sources = files(
)
includes += include_directories('../../common/cpt')
+includes += include_directories('../../common/octeontx')
+includes += include_directories('../../event/octeontx')
+includes += include_directories('../../mempool/octeontx')
+includes += include_directories('../../net/octeontx')
diff --git a/drivers/crypto/octeontx/otx_cryptodev_ops.c b/drivers/crypto/octeontx/otx_cryptodev_ops.c
index d75f4b5f81..2ec95bbca4 100644
--- a/drivers/crypto/octeontx/otx_cryptodev_ops.c
+++ b/drivers/crypto/octeontx/otx_cryptodev_ops.c
@@ -6,6 +6,8 @@
#include <rte_bus_pci.h>
#include <rte_cryptodev.h>
#include <rte_cryptodev_pmd.h>
+#include <rte_eventdev.h>
+#include <rte_event_crypto_adapter.h>
#include <rte_errno.h>
#include <rte_malloc.h>
#include <rte_mempool.h>
@@ -21,6 +23,8 @@
#include "cpt_ucode.h"
#include "cpt_ucode_asym.h"
+#include "ssovf_worker.h"
+
static uint64_t otx_fpm_iova[CPT_EC_ID_PMAX];
/* Forward declarations */
@@ -412,15 +416,17 @@ otx_cpt_asym_session_clear(struct rte_cryptodev *dev,
rte_mempool_put(sess_mp, priv);
}
-static __rte_always_inline int32_t __rte_hot
+static __rte_always_inline void * __rte_hot
otx_cpt_request_enqueue(struct cpt_instance *instance,
struct pending_queue *pqueue,
void *req, uint64_t cpt_inst_w7)
{
struct cpt_request_info *user_req = (struct cpt_request_info *)req;
- if (unlikely(pqueue->pending_count >= DEFAULT_CMD_QLEN))
- return -EAGAIN;
+ if (unlikely(pqueue->pending_count >= DEFAULT_CMD_QLEN)) {
+ rte_errno = EAGAIN;
+ return NULL;
+ }
fill_cpt_inst(instance, req, cpt_inst_w7);
@@ -434,18 +440,12 @@ otx_cpt_request_enqueue(struct cpt_instance *instance,
/* Default mode of software queue */
mark_cpt_inst(instance);
- pqueue->req_queue[pqueue->enq_tail] = (uintptr_t)user_req;
-
- /* We will use soft queue length here to limit requests */
- MOD_INC(pqueue->enq_tail, DEFAULT_CMD_QLEN);
- pqueue->pending_count += 1;
-
CPT_LOG_DP_DEBUG("Submitted NB cmd with request: %p "
"op: %p", user_req, user_req->op);
- return 0;
+ return req;
}
-static __rte_always_inline int __rte_hot
+static __rte_always_inline void * __rte_hot
otx_cpt_enq_single_asym(struct cpt_instance *instance,
struct rte_crypto_op *op,
struct pending_queue *pqueue)
@@ -456,11 +456,13 @@ otx_cpt_enq_single_asym(struct cpt_instance *instance,
struct cpt_asym_sess_misc *sess;
uintptr_t *cop;
void *mdata;
+ void *req;
int ret;
if (unlikely(rte_mempool_get(minfo->pool, &mdata) < 0)) {
CPT_LOG_DP_ERR("Could not allocate meta buffer for request");
- return -ENOMEM;
+ rte_errno = ENOMEM;
+ return NULL;
}
sess = get_asym_session_private_data(asym_op->session,
@@ -506,27 +508,26 @@ otx_cpt_enq_single_asym(struct cpt_instance *instance,
default:
op->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS;
- ret = -EINVAL;
+ rte_errno = EINVAL;
goto req_fail;
}
- ret = otx_cpt_request_enqueue(instance, pqueue, params.req,
+ req = otx_cpt_request_enqueue(instance, pqueue, params.req,
sess->cpt_inst_w7);
-
- if (unlikely(ret)) {
+ if (unlikely(!req)) {
CPT_LOG_DP_ERR("Could not enqueue crypto req");
goto req_fail;
}
- return 0;
+ return req;
req_fail:
free_op_meta(mdata, minfo->pool);
- return ret;
+ return NULL;
}
-static __rte_always_inline int __rte_hot
+static __rte_always_inline void * __rte_hot
otx_cpt_enq_single_sym(struct cpt_instance *instance,
struct rte_crypto_op *op,
struct pending_queue *pqueue)
@@ -536,6 +537,7 @@ otx_cpt_enq_single_sym(struct cpt_instance *instance,
struct cpt_request_info *prep_req;
void *mdata = NULL;
int ret = 0;
+ void *req;
uint64_t cpt_op;
sess = (struct cpt_sess_misc *)
@@ -554,23 +556,22 @@ otx_cpt_enq_single_sym(struct cpt_instance *instance,
if (unlikely(ret)) {
CPT_LOG_DP_ERR("prep cryto req : op %p, cpt_op 0x%x "
"ret 0x%x", op, (unsigned int)cpt_op, ret);
- return ret;
+ return NULL;
}
/* Enqueue prepared instruction to h/w */
- ret = otx_cpt_request_enqueue(instance, pqueue, prep_req,
+ req = otx_cpt_request_enqueue(instance, pqueue, prep_req,
sess->cpt_inst_w7);
-
- if (unlikely(ret)) {
+ if (unlikely(!req)) {
/* Buffer allocated for request preparation need to be freed */
free_op_meta(mdata, instance->meta_info.pool);
- return ret;
+ return NULL;
}
- return 0;
+ return req;
}
-static __rte_always_inline int __rte_hot
+static __rte_always_inline void * __rte_hot
otx_cpt_enq_single_sym_sessless(struct cpt_instance *instance,
struct rte_crypto_op *op,
struct pending_queue *pend_q)
@@ -578,12 +579,15 @@ otx_cpt_enq_single_sym_sessless(struct cpt_instance *instance,
const int driver_id = otx_cryptodev_driver_id;
struct rte_crypto_sym_op *sym_op = op->sym;
struct rte_cryptodev_sym_session *sess;
+ void *req;
int ret;
/* Create temporary session */
sess = rte_cryptodev_sym_session_create(instance->sess_mp);
- if (sess == NULL)
- return -ENOMEM;
+ if (sess == NULL) {
+ rte_errno = ENOMEM;
+ return NULL;
+ }
ret = sym_session_configure(driver_id, sym_op->xform, sess,
instance->sess_mp_priv);
@@ -592,24 +596,24 @@ otx_cpt_enq_single_sym_sessless(struct cpt_instance *instance,
sym_op->session = sess;
- ret = otx_cpt_enq_single_sym(instance, op, pend_q);
+ req = otx_cpt_enq_single_sym(instance, op, pend_q);
- if (unlikely(ret))
+ if (unlikely(!req))
goto priv_put;
- return 0;
+ return req;
priv_put:
sym_session_clear(driver_id, sess);
sess_put:
rte_mempool_put(instance->sess_mp, sess);
- return ret;
+ return NULL;
}
#define OP_TYPE_SYM 0
#define OP_TYPE_ASYM 1
-static __rte_always_inline int __rte_hot
+static __rte_always_inline void *__rte_hot
otx_cpt_enq_single(struct cpt_instance *inst,
struct rte_crypto_op *op,
struct pending_queue *pqueue,
@@ -631,7 +635,8 @@ otx_cpt_enq_single(struct cpt_instance *inst,
}
/* Should not reach here */
- return -ENOTSUP;
+ rte_errno = ENOTSUP;
+ return NULL;
}
static __rte_always_inline uint16_t __rte_hot
@@ -640,7 +645,7 @@ otx_cpt_pkt_enqueue(void *qptr, struct rte_crypto_op **ops, uint16_t nb_ops,
{
struct cpt_instance *instance = (struct cpt_instance *)qptr;
uint16_t count;
- int ret;
+ void *req;
struct cpt_vf *cptvf = (struct cpt_vf *)instance;
struct pending_queue *pqueue = &cptvf->pqueue;
@@ -652,10 +657,15 @@ otx_cpt_pkt_enqueue(void *qptr, struct rte_crypto_op **ops, uint16_t nb_ops,
while (likely(count < nb_ops)) {
/* Enqueue single op */
- ret = otx_cpt_enq_single(instance, ops[count], pqueue, op_type);
+ req = otx_cpt_enq_single(instance, ops[count], pqueue, op_type);
- if (unlikely(ret))
+ if (unlikely(!req))
break;
+
+ pqueue->req_queue[pqueue->enq_tail] = (uintptr_t)req;
+ /* We will use soft queue length here to limit requests */
+ MOD_INC(pqueue->enq_tail, DEFAULT_CMD_QLEN);
+ pqueue->pending_count += 1;
count++;
}
otx_cpt_ring_dbell(instance, count);
@@ -674,6 +684,83 @@ otx_cpt_enqueue_sym(void *qptr, struct rte_crypto_op **ops, uint16_t nb_ops)
return otx_cpt_pkt_enqueue(qptr, ops, nb_ops, OP_TYPE_SYM);
}
+static __rte_always_inline void
+submit_request_to_sso(struct ssows *ws, struct cpt_request_info *req,
+ struct rte_event *rsp_info)
+{
+ uint64_t add_work;
+
+ add_work = rsp_info->flow_id | (RTE_EVENT_TYPE_CRYPTODEV << 28) |
+ ((uint64_t)(rsp_info->sched_type) << 32);
+
+ if (!rsp_info->sched_type)
+ ssows_head_wait(ws);
+
+ rte_atomic_thread_fence(__ATOMIC_RELEASE);
+ ssovf_store_pair(add_work, (uint64_t)req, ws->grps[rsp_info->queue_id]);
+}
+
+static inline union rte_event_crypto_metadata *
+get_event_crypto_mdata(struct rte_crypto_op *op)
+{
+ union rte_event_crypto_metadata *ec_mdata;
+
+ if (unlikely(!op))
+ return NULL;
+
+ if (op->sess_type == RTE_CRYPTO_OP_WITH_SESSION)
+ ec_mdata = rte_cryptodev_sym_session_get_user_data(
+ op->sym->session);
+ else if (op->sess_type == RTE_CRYPTO_OP_SESSIONLESS &&
+ op->private_data_offset)
+ ec_mdata = (union rte_event_crypto_metadata *)
+ ((uint8_t *)op + op->private_data_offset);
+ else
+ return NULL;
+
+ return ec_mdata;
+}
+
+uint16_t __rte_hot
+otx_crypto_adapter_enqueue(void *port, struct rte_crypto_op *op)
+{
+ union rte_event_crypto_metadata *ec_mdata;
+ struct cpt_instance *instance;
+ struct cpt_request_info *req;
+ struct rte_event *rsp_info;
+ uint8_t op_type, cdev_id;
+ uint16_t qp_id;
+
+ ec_mdata = get_event_crypto_mdata(op);
+ if (unlikely(!ec_mdata)) {
+ rte_errno = EINVAL;
+ return 0;
+ }
+
+ cdev_id = ec_mdata->request_info.cdev_id;
+ qp_id = ec_mdata->request_info.queue_pair_id;
+ rsp_info = &ec_mdata->response_info;
+ instance = rte_cryptodevs[cdev_id].data->queue_pairs[qp_id];
+
+ if (unlikely(!instance->ca_enabled)) {
+ rte_errno = EINVAL;
+ return 0;
+ }
+
+ op_type = op->type == RTE_CRYPTO_OP_TYPE_SYMMETRIC ? OP_TYPE_SYM :
+ OP_TYPE_ASYM;
+ req = otx_cpt_enq_single(instance, op,
+ &((struct cpt_vf *)instance)->pqueue, op_type);
+ if (unlikely(!req))
+ return 0;
+
+ otx_cpt_ring_dbell(instance, 1);
+ req->qp = instance;
+ submit_request_to_sso(port, req, rsp_info);
+
+ return 1;
+}
+
static inline void
otx_cpt_asym_rsa_op(struct rte_crypto_op *cop, struct cpt_request_info *req,
struct rte_crypto_rsa_xform *rsa_ctx)
@@ -820,6 +907,50 @@ otx_cpt_dequeue_post_process(struct rte_crypto_op *cop, uintptr_t *rsp,
return;
}
+static inline void
+free_sym_session_data(const struct cpt_instance *instance,
+ struct rte_crypto_op *cop)
+{
+ void *sess_private_data_t = get_sym_session_private_data(
+ cop->sym->session, otx_cryptodev_driver_id);
+ memset(sess_private_data_t, 0, cpt_get_session_size());
+ memset(cop->sym->session, 0,
+ rte_cryptodev_sym_get_existing_header_session_size(
+ cop->sym->session));
+ rte_mempool_put(instance->sess_mp_priv, sess_private_data_t);
+ rte_mempool_put(instance->sess_mp, cop->sym->session);
+ cop->sym->session = NULL;
+}
+
+static __rte_always_inline struct rte_crypto_op *
+otx_cpt_process_response(const struct cpt_instance *instance, uintptr_t *rsp,
+ uint8_t cc, const uint8_t op_type)
+{
+ struct rte_crypto_op *cop;
+ void *metabuf;
+
+ metabuf = (void *)rsp[0];
+ cop = (void *)rsp[1];
+
+ /* Check completion code */
+ if (likely(cc == 0)) {
+ /* H/w success pkt. Post process */
+ otx_cpt_dequeue_post_process(cop, rsp, op_type);
+ } else if (cc == ERR_GC_ICV_MISCOMPARE) {
+ /* auth data mismatch */
+ cop->status = RTE_CRYPTO_OP_STATUS_AUTH_FAILED;
+ } else {
+ /* Error */
+ cop->status = RTE_CRYPTO_OP_STATUS_ERROR;
+ }
+
+ if (unlikely(cop->sess_type == RTE_CRYPTO_OP_SESSIONLESS))
+ free_sym_session_data(instance, cop);
+ free_op_meta(metabuf, instance->meta_info.pool);
+
+ return cop;
+}
+
static __rte_always_inline uint16_t __rte_hot
otx_cpt_pkt_dequeue(void *qptr, struct rte_crypto_op **ops, uint16_t nb_ops,
const uint8_t op_type)
@@ -832,9 +963,6 @@ otx_cpt_pkt_dequeue(void *qptr, struct rte_crypto_op **ops, uint16_t nb_ops,
uint8_t ret;
int nb_completed;
struct pending_queue *pqueue = &cptvf->pqueue;
- struct rte_crypto_op *cop;
- void *metabuf;
- uintptr_t *rsp;
pcount = pqueue->pending_count;
count = (nb_ops > pcount) ? pcount : nb_ops;
@@ -869,45 +997,11 @@ otx_cpt_pkt_dequeue(void *qptr, struct rte_crypto_op **ops, uint16_t nb_ops,
nb_completed = i;
for (i = 0; i < nb_completed; i++) {
-
- rsp = (void *)ops[i];
-
if (likely((i + 1) < nb_completed))
rte_prefetch0(ops[i+1]);
- metabuf = (void *)rsp[0];
- cop = (void *)rsp[1];
-
- ops[i] = cop;
-
- /* Check completion code */
-
- if (likely(cc[i] == 0)) {
- /* H/w success pkt. Post process */
- otx_cpt_dequeue_post_process(cop, rsp, op_type);
- } else if (cc[i] == ERR_GC_ICV_MISCOMPARE) {
- /* auth data mismatch */
- cop->status = RTE_CRYPTO_OP_STATUS_AUTH_FAILED;
- } else {
- /* Error */
- cop->status = RTE_CRYPTO_OP_STATUS_ERROR;
- }
-
- if (unlikely(cop->sess_type == RTE_CRYPTO_OP_SESSIONLESS)) {
- void *sess_private_data_t =
- get_sym_session_private_data(cop->sym->session,
- otx_cryptodev_driver_id);
- memset(sess_private_data_t, 0,
- cpt_get_session_size());
- memset(cop->sym->session, 0,
- rte_cryptodev_sym_get_existing_header_session_size(
- cop->sym->session));
- rte_mempool_put(instance->sess_mp_priv,
- sess_private_data_t);
- rte_mempool_put(instance->sess_mp, cop->sym->session);
- cop->sym->session = NULL;
- }
- free_op_meta(metabuf, instance->meta_info.pool);
+ ops[i] = otx_cpt_process_response(instance, (void *)ops[i],
+ cc[i], op_type);
}
return nb_completed;
@@ -925,6 +1019,32 @@ otx_cpt_dequeue_sym(void *qptr, struct rte_crypto_op **ops, uint16_t nb_ops)
return otx_cpt_pkt_dequeue(qptr, ops, nb_ops, OP_TYPE_SYM);
}
+uint64_t __rte_hot
+otx_crypto_adapter_dequeue(uint64_t get_work1)
+{
+ const struct cpt_instance *instance;
+ struct cpt_request_info *req;
+ struct rte_crypto_op *cop;
+ uint8_t cc, op_type;
+ uintptr_t *rsp;
+
+ req = (struct cpt_request_info *)(get_work1);
+ instance = req->qp;
+ rsp = req->op;
+ cop = (void *)rsp[1];
+ op_type = cop->type == RTE_CRYPTO_OP_TYPE_SYMMETRIC ? OP_TYPE_SYM :
+ OP_TYPE_ASYM;
+
+ do {
+ cc = check_nb_command_id(
+ req, (struct cpt_instance *)(uintptr_t)instance);
+ } while (cc == ERR_REQ_PENDING);
+
+ cop = otx_cpt_process_response(instance, (void *)req->op, cc, op_type);
+
+ return (uint64_t)(cop);
+}
+
static struct rte_cryptodev_ops cptvf_ops = {
/* Device related operations */
.dev_configure = otx_cpt_dev_config,
diff --git a/drivers/crypto/octeontx/otx_cryptodev_ops.h b/drivers/crypto/octeontx/otx_cryptodev_ops.h
index fac8a3c006..46f474c1ba 100644
--- a/drivers/crypto/octeontx/otx_cryptodev_ops.h
+++ b/drivers/crypto/octeontx/otx_cryptodev_ops.h
@@ -14,4 +14,11 @@
int
otx_cpt_dev_create(struct rte_cryptodev *c_dev);
+__rte_internal
+uint16_t __rte_hot
+otx_crypto_adapter_enqueue(void *port, struct rte_crypto_op *op);
+__rte_internal
+uint64_t __rte_hot
+otx_crypto_adapter_dequeue(uint64_t get_work1);
+
#endif /* _OTX_CRYPTODEV_OPS_H_ */
diff --git a/drivers/crypto/octeontx/version.map b/drivers/crypto/octeontx/version.map
index 4a76d1d52d..41f33a4ecf 100644
--- a/drivers/crypto/octeontx/version.map
+++ b/drivers/crypto/octeontx/version.map
@@ -1,3 +1,12 @@
DPDK_21 {
local: *;
};
+
+INTERNAL {
+ global:
+
+ otx_crypto_adapter_enqueue;
+ otx_crypto_adapter_dequeue;
+
+ local: *;
+};
diff --git a/drivers/event/octeontx/ssovf_evdev.c b/drivers/event/octeontx/ssovf_evdev.c
index 25bf207db6..b93f6ec8c6 100644
--- a/drivers/event/octeontx/ssovf_evdev.c
+++ b/drivers/event/octeontx/ssovf_evdev.c
@@ -734,7 +734,8 @@ ssovf_crypto_adapter_caps_get(const struct rte_eventdev *dev,
RTE_SET_USED(dev);
RTE_SET_USED(cdev);
- *caps = 0;
+ *caps = RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD |
+ RTE_EVENT_CRYPTO_ADAPTER_CAP_SESSION_PRIVATE_DATA;
return 0;
}
diff --git a/drivers/event/octeontx/ssovf_worker.c b/drivers/event/octeontx/ssovf_worker.c
index a9149fb934..8b056ddc5a 100644
--- a/drivers/event/octeontx/ssovf_worker.c
+++ b/drivers/event/octeontx/ssovf_worker.c
@@ -322,6 +322,15 @@ sso_event_tx_adapter_enqueue_ ## name(void *port, struct rte_event ev[], \
SSO_TX_ADPTR_ENQ_FASTPATH_FUNC
#undef T
+static uint16_t __rte_hot
+ssow_crypto_adapter_enqueue(void *port, struct rte_event ev[],
+ uint16_t nb_events)
+{
+ RTE_SET_USED(nb_events);
+
+ return otx_crypto_adapter_enqueue(port, ev->event_ptr);
+}
+
void
ssovf_fastpath_fns_set(struct rte_eventdev *dev)
{
@@ -332,6 +341,8 @@ ssovf_fastpath_fns_set(struct rte_eventdev *dev)
dev->enqueue_new_burst = ssows_enq_new_burst;
dev->enqueue_forward_burst = ssows_enq_fwd_burst;
+ dev->ca_enqueue = ssow_crypto_adapter_enqueue;
+
const event_tx_adapter_enqueue ssow_txa_enqueue[2][2][2][2] = {
#define T(name, f3, f2, f1, f0, sz, flags) \
[f3][f2][f1][f0] = sso_event_tx_adapter_enqueue_ ##name,
diff --git a/drivers/event/octeontx/ssovf_worker.h b/drivers/event/octeontx/ssovf_worker.h
index 4354f007d7..29b5f48761 100644
--- a/drivers/event/octeontx/ssovf_worker.h
+++ b/drivers/event/octeontx/ssovf_worker.h
@@ -4,6 +4,9 @@
#include <arpa/inet.h>
+#ifndef _SSOVF_WORKER_H_
+#define _SSOVF_WORKER_H_
+
#include <rte_common.h>
#include <rte_branch_prediction.h>
@@ -11,6 +14,7 @@
#include "ssovf_evdev.h"
#include "octeontx_rxtx.h"
+#include "otx_cryptodev_ops.h"
/* Alignment */
#define OCCTX_ALIGN 128
@@ -174,14 +178,17 @@ ssows_get_work(struct ssows *ws, struct rte_event *ev, const uint16_t flag)
sched_type_queue = sched_type_queue << 38;
ev->event = sched_type_queue | (get_work0 & 0xffffffff);
- if (get_work1 && ev->event_type == RTE_EVENT_TYPE_ETHDEV) {
- ev->mbuf = ssovf_octeontx_wqe_to_pkt(get_work1,
- (ev->event >> 20) & 0x7F, flag, ws->lookup_mem);
+ if (get_work1) {
+ if (ev->event_type == RTE_EVENT_TYPE_ETHDEV)
+ get_work1 = (uint64_t)ssovf_octeontx_wqe_to_pkt(
+ get_work1, (ev->event >> 20) & 0x7F, flag,
+ ws->lookup_mem);
+ else if (ev->event_type == RTE_EVENT_TYPE_CRYPTODEV)
+ get_work1 = otx_crypto_adapter_dequeue(get_work1);
+ ev->u64 = get_work1;
} else if (unlikely((get_work0 & 0xFFFFFFFF) == 0xFFFFFFFF)) {
ssovf_octeontx_wqe_free(get_work1);
return 0;
- } else {
- ev->u64 = get_work1;
}
return !!get_work1;
@@ -254,3 +261,11 @@ ssows_swtag_wait(struct ssows *ws)
while (ssovf_read64(ws->base + SSOW_VHWS_SWTP))
;
}
+
+static __rte_always_inline void
+ssows_head_wait(struct ssows *ws)
+{
+ while (!(ssovf_read64(ws->base + SSOW_VHWS_TAG) & (1UL << 35)))
+ ;
+}
+#endif /* _SSOVF_WORKER_H_ */
diff --git a/drivers/event/octeontx2/otx2_evdev_crypto_adptr_rx.h b/drivers/event/octeontx2/otx2_evdev_crypto_adptr_rx.h
index 9e331fdd75..a543225376 100644
--- a/drivers/event/octeontx2/otx2_evdev_crypto_adptr_rx.h
+++ b/drivers/event/octeontx2/otx2_evdev_crypto_adptr_rx.h
@@ -54,6 +54,7 @@ static inline uint64_t
otx2_handle_crypto_event(uint64_t get_work1)
{
struct cpt_request_info *req;
+ const struct otx2_cpt_qp *qp;
struct rte_crypto_op *cop;
uintptr_t *rsp;
void *metabuf;
@@ -61,14 +62,15 @@ otx2_handle_crypto_event(uint64_t get_work1)
req = (struct cpt_request_info *)(get_work1);
cc = otx2_cpt_compcode_get(req);
+ qp = req->qp;
rsp = req->op;
metabuf = (void *)rsp[0];
cop = (void *)rsp[1];
- otx2_ca_deq_post_process(req->qp, cop, rsp, cc);
+ otx2_ca_deq_post_process(qp, cop, rsp, cc);
- rte_mempool_put(req->qp->meta_info.pool, metabuf);
+ rte_mempool_put(qp->meta_info.pool, metabuf);
return (uint64_t)(cop);
}
--
2.25.1
next prev parent reply other threads:[~2021-06-22 16:49 UTC|newest]
Thread overview: 24+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-06-22 16:48 [dpdk-dev] [PATCH 0/2] OCTEONTX crypto adapter support Shijith Thotton
2021-06-22 16:48 ` [dpdk-dev] [PATCH 1/2] drivers: add octeontx crypto adapter framework Shijith Thotton
2021-06-22 16:48 ` Shijith Thotton [this message]
2021-07-06 20:20 ` [dpdk-dev] [PATCH 2/2] drivers: add octeontx crypto adapter data path Akhil Goyal
2021-06-23 20:53 ` [dpdk-dev] [PATCH v2 0/2] OCTEONTX crypto adapter support Shijith Thotton
2021-06-23 20:53 ` [dpdk-dev] [PATCH v2 1/2] drivers: add octeontx crypto adapter framework Shijith Thotton
2021-07-06 20:14 ` Akhil Goyal
2021-07-15 14:21 ` David Marchand
2021-07-16 8:39 ` [dpdk-dev] [EXT] " Akhil Goyal
2021-07-20 11:58 ` Akhil Goyal
2021-07-20 12:14 ` David Marchand
2021-07-21 9:44 ` Thomas Monjalon
2021-07-21 15:11 ` Brandon Lo
2021-07-22 7:45 ` Akhil Goyal
2021-07-22 9:06 ` [dpdk-dev] [PATCH] crypto/octeontx: enable build on non Linux OS Shijith Thotton
2021-07-22 9:17 ` Akhil Goyal
2021-07-22 19:06 ` Thomas Monjalon
2021-07-22 19:08 ` Thomas Monjalon
2021-07-22 20:20 ` Brandon Lo
2021-07-22 20:32 ` Thomas Monjalon
2021-06-23 20:53 ` [dpdk-dev] [PATCH v2 2/2] drivers: add octeontx crypto adapter data path Shijith Thotton
2021-06-30 8:54 ` Akhil Goyal
2021-06-30 16:23 ` [dpdk-dev] [dpdk-ci] " Brandon Lo
2021-07-07 9:28 ` [dpdk-dev] [PATCH v2 0/2] OCTEONTX crypto adapter support Akhil Goyal
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=185977c5fa60dad0ea0b89e7edbb369bc58726b8.1624379833.git.sthotton@marvell.com \
--to=sthotton@marvell.com \
--cc=abhinandan.gujjar@intel.com \
--cc=adwivedi@marvell.com \
--cc=anoobj@marvell.com \
--cc=dev@dpdk.org \
--cc=gakhil@marvell.com \
--cc=jerinj@marvell.com \
--cc=pbhagavatula@marvell.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).