* [PATCH v2 13/19] net/xsc: add ethdev start and stop ops
@ 2024-09-11 2:07 WanRenyong
0 siblings, 0 replies; only message in thread
From: WanRenyong @ 2024-09-11 2:07 UTC (permalink / raw)
To: dev; +Cc: ferruh.yigit, thomas, WanRenyong, Rong Qian
Implement xsc ethdev start and stop function.
Signed-off-by: WanRenyong <wanry@yunsilicon.com>
Signed-off-by: Rong Qian <qianr@yunsilicon.com>
---
drivers/net/xsc/meson.build | 1 +
drivers/net/xsc/xsc_ctrl.h | 152 ++++++-
drivers/net/xsc/xsc_defs.h | 2 +
drivers/net/xsc/xsc_dev.h | 3 +
drivers/net/xsc/xsc_ethdev.c | 740 ++++++++++++++++++++++++++++++++++-
drivers/net/xsc/xsc_ethdev.h | 10 +
drivers/net/xsc/xsc_rxtx.c | 22 ++
drivers/net/xsc/xsc_rxtx.h | 68 +++-
8 files changed, 994 insertions(+), 4 deletions(-)
create mode 100644 drivers/net/xsc/xsc_rxtx.c
diff --git a/drivers/net/xsc/meson.build b/drivers/net/xsc/meson.build
index f38ebdfe0f..5f15c8def5 100644
--- a/drivers/net/xsc/meson.build
+++ b/drivers/net/xsc/meson.build
@@ -11,6 +11,7 @@ sources = files(
'xsc_dev.c',
'xsc_utils.c',
'xsc_ctrl.c',
+ 'xsc_rxtx.c',
)
libnames = ['ibverbs']
diff --git a/drivers/net/xsc/xsc_ctrl.h b/drivers/net/xsc/xsc_ctrl.h
index c33e625097..e51847d68f 100644
--- a/drivers/net/xsc/xsc_ctrl.h
+++ b/drivers/net/xsc/xsc_ctrl.h
@@ -5,7 +5,17 @@
#ifndef _XSC_CTRL_H_
#define _XSC_CTRL_H_
+#include <sys/types.h>
+#include <unistd.h>
+#include <string.h>
+#include <dirent.h>
+#include <net/if.h>
#include <sys/ioctl.h>
+#include <infiniband/verbs.h>
+
+#ifndef PAGE_SIZE
+#define PAGE_SIZE 4096
+#endif
#define XSC_IOCTL_CHECK_FIELD 0x01234567
@@ -25,6 +35,17 @@ enum xsc_ioctl_opmod {
XSC_IOCTL_OP_GET_LOCAL,
};
+#define XSC_DIV_ROUND_UP(n, d) ({ \
+ typeof(d) _d = (d); \
+ typeof(n) _n = (n); \
+ ((_n) + (_d) - 1) / (_d); \
+})
+
+enum {
+ XSC_IOCTL_SET_QP_STATUS = 0x200,
+ XSC_IOCTL_SET_MAX
+};
+
struct xsc_ioctl_attr {
uint16_t opcode; /* ioctl cmd */
uint16_t length; /* data length */
@@ -40,7 +61,18 @@ struct xsc_ioctl_hdr {
struct xsc_ioctl_attr attr;
};
-/* ioctl */
+enum {
+ XSC_QUEUE_TYPE_RDMA_RC = 0,
+ XSC_QUEUE_TYPE_RDMA_MAD = 1,
+ XSC_QUEUE_TYPE_RAW = 2,
+ XSC_QUEUE_TYPE_VIRTIO_NET = 3,
+ XSC_QUEUE_TYPE_VIRTIO_BLK = 4,
+ XSC_QUEUE_TYPE_RAW_TPE = 5,
+ XSC_QUEUE_TYPE_RAW_TSO = 6,
+ XSC_QUEUE_TYPE_RAW_TX = 7,
+ XSC_QUEUE_TYPE_INVALID = 0xFF,
+};
+
struct xsc_inbox_hdr {
__be16 opcode;
uint8_t rsvd[4];
@@ -53,7 +85,6 @@ struct xsc_outbox_hdr {
__be32 syndrome;
};
-/* ioctl mbox */
struct xsc_ioctl_mbox_in {
struct xsc_inbox_hdr hdr;
__be16 len;
@@ -96,6 +127,54 @@ struct xsc_cmd_modify_nic_hca_mbox_out {
uint8_t rsvd0[4];
};
+struct xsc_create_qp_request {
+ __be16 input_qpn;
+ __be16 pa_num;
+ uint8_t qp_type;
+ uint8_t log_sq_sz;
+ uint8_t log_rq_sz;
+ uint8_t dma_direct;
+ __be32 pdn;
+ __be16 cqn_send;
+ __be16 cqn_recv;
+ __be16 glb_funcid;
+ uint8_t page_shift;
+ uint8_t rsvd;
+ __be64 pas[];
+};
+
+struct xsc_create_multiqp_mbox_in {
+ struct xsc_inbox_hdr hdr;
+ __be16 qp_num;
+ uint8_t qp_type;
+ uint8_t rsvd;
+ __be32 req_len;
+ uint8_t data[];
+};
+
+struct xsc_create_multiqp_mbox_out {
+ struct xsc_outbox_hdr hdr;
+ __be32 qpn_base;
+};
+
+
+struct xsc_destroy_qp_mbox_in {
+ struct xsc_inbox_hdr hdr;
+ __be32 qpn;
+ uint8_t rsvd[4];
+};
+
+struct xsc_destroy_qp_mbox_out {
+ struct xsc_outbox_hdr hdr;
+ uint8_t rsvd[8];
+};
+
+struct xsc_ioctl_qp_range {
+ uint16_t opcode;
+ int num;
+ uint32_t qpn;
+};
+
struct xsc_ioctl_data_tl {
uint16_t table;
uint16_t opmod;
@@ -136,6 +215,75 @@ struct xsc_ioctl_get_hwinfo {
uint8_t esw_mode;
};
+/* for xscdv providers */
+#if !HAVE_XSC_DV_PROVIDER
+enum xscdv_obj_type {
+ XSCDV_OBJ_QP = 1 << 0,
+ XSCDV_OBJ_CQ = 1 << 1,
+ XSCDV_OBJ_SRQ = 1 << 2,
+ XSCDV_OBJ_RWQ = 1 << 3,
+ XSCDV_OBJ_DM = 1 << 4,
+ XSCDV_OBJ_AH = 1 << 5,
+ XSCDV_OBJ_PD = 1 << 6,
+};
+
+enum xsc_qp_create_flags {
+ XSC_QP_CREATE_RAWPACKE_TSO = 1 << 0,
+ XSC_QP_CREATE_RAWPACKET_TSO = 1 << 0,
+ XSC_QP_CREATE_RAWPACKET_TX = 1 << 1,
+};
+
+struct xscdv_cq_init_attr {
+ uint64_t comp_mask; /* Use enum xscdv_cq_init_attr_mask */
+ uint8_t cqe_comp_res_format; /* Use enum xscdv_cqe_comp_res_format */
+ uint32_t flags;
+ uint16_t cqe_size; /* when XSCDV_CQ_INIT_ATTR_MASK_CQE_SIZE set */
+};
+
+struct xscdv_obj {
+ struct {
+ struct ibv_qp *in;
+ struct xscdv_qp *out;
+ } qp;
+ struct {
+ struct ibv_cq *in;
+ struct xscdv_cq *out;
+ } cq;
+};
+
+struct xscdv_qp {
+ __le32 *dbrec;
+ struct {
+ void *buf;
+ uint32_t wqe_cnt;
+ uint32_t stride;
+ __le32 *db;
+ } sq;
+ struct {
+ void *buf;
+ uint32_t wqe_cnt;
+ uint32_t stride;
+ __le32 *db;
+ } rq;
+ uint64_t comp_mask;
+ uint32_t tirn;
+ uint32_t tisn;
+ uint32_t rqn;
+ uint32_t sqn;
+};
+
+struct xscdv_cq {
+ void *buf;
+ __le32 *dbrec;
+ __le32 *db;
+ uint32_t cqe_cnt;
+ uint32_t cqe_size;
+ uint32_t cqn;
+ uint64_t comp_mask;
+};
+
+#endif
+
int xsc_ioctl(struct xsc_dev *dev, int cmd, int opcode,
void *data_in, int in_len, void *data_out, int out_len);
int xsc_mailbox_exec(struct xsc_dev *dev, void *data_in,
diff --git a/drivers/net/xsc/xsc_defs.h b/drivers/net/xsc/xsc_defs.h
index 7dc57e5717..769cbad812 100644
--- a/drivers/net/xsc/xsc_defs.h
+++ b/drivers/net/xsc/xsc_defs.h
@@ -11,6 +11,8 @@
#define XSC_VFREP_BASE_LOGICAL_PORT 1081
#define XSC_MAX_MAC_ADDRESSES 3
+#define XSC_SEND_WQE_DS 3
+#define XSC_ESEG_EXTRA_DATA_SIZE 48u
enum xsc_nic_mode {
XSC_NIC_MODE_LEGACY,
diff --git a/drivers/net/xsc/xsc_dev.h b/drivers/net/xsc/xsc_dev.h
index f77551f1c5..a24ae582f3 100644
--- a/drivers/net/xsc/xsc_dev.h
+++ b/drivers/net/xsc/xsc_dev.h
@@ -5,6 +5,9 @@
#ifndef _XSC_DEV_H_
#define _XSC_DEV_H_
+#if HAVE_XSC_DV_PROVIDER
+#include <infiniband/xscdv.h>
+#endif
#include <infiniband/verbs.h>
#include "xsc_defs.h"
diff --git a/drivers/net/xsc/xsc_ethdev.c b/drivers/net/xsc/xsc_ethdev.c
index 25b2a4c44d..35daa16528 100644
--- a/drivers/net/xsc/xsc_ethdev.c
+++ b/drivers/net/xsc/xsc_ethdev.c
@@ -2,6 +2,8 @@
* Copyright 2024 Yunsilicon Technology Co., Ltd.
*/
+#include <net/if.h>
+
#include <ethdev_pci.h>
#include "xsc_log.h"
@@ -9,10 +11,29 @@
#include "xsc_dev.h"
#include "xsc_ethdev.h"
#include "xsc_utils.h"
-
#include "xsc_ctrl.h"
#include "xsc_rxtx.h"
+static __rte_always_inline struct xsc_rxq_data *
+xsc_rxq_get(struct rte_eth_dev *dev, uint16_t idx)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+
+ if (priv->rxqs != NULL && (*priv->rxqs)[idx] != NULL)
+ return (*priv->rxqs)[idx];
+ return NULL;
+}
+
+static __rte_always_inline struct xsc_txq_data *
+xsc_txq_get(struct rte_eth_dev *dev, uint16_t idx)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+
+ if (priv->txqs != NULL && (*priv->txqs)[idx] != NULL)
+ return (*priv->txqs)[idx];
+ return NULL;
+}
+
static int
xsc_rss_modify_cmd(struct xsc_ethdev_priv *priv, uint8_t *rss_key,
uint8_t rss_key_len)
@@ -84,6 +105,31 @@ xsc_ethdev_rss_hash_update(struct rte_eth_dev *dev,
return ret;
}
+static int
+xsc_modify_qpn_status(uint32_t qpn, int num, int opcode, struct xsc_dev *xdev)
+{
+ struct {
+ struct xsc_ioctl_data_tl tl;
+ struct xsc_ioctl_qp_range info;
+ } data_info;
+
+ int ret;
+
+ data_info.tl.opmod = XSC_IOCTL_SET_QP_STATUS;
+ data_info.info.opcode = opcode;
+ data_info.info.qpn = qpn;
+ data_info.info.num = num;
+
+ ret = xsc_ioctl(xdev, XSC_IOCTL_DRV_GET, XSC_IOCTL_SET_QP_STATUS,
+ &data_info, sizeof(data_info), NULL, 0);
+ if (ret != 0) {
+ rte_errno = ret;
+ PMD_DRV_LOG(ERR, "modify qp status fail, ret = %d\n", ret);
+ }
+
+ return ret;
+}
+
static int
xsc_ethdev_configure(struct rte_eth_dev *dev)
{
@@ -140,6 +186,685 @@ xsc_ethdev_configure(struct rte_eth_dev *dev)
return -rte_errno;
}
+static int
+xsc_init_obj(struct xscdv_obj *obj, uint64_t obj_type)
+{
+#if HAVE_XSC_DV_PROVIDER
+ return xscdv_init_obj(obj, obj_type);
+#else
+ (void)obj;
+ (void)obj_type;
+ return 0;
+#endif
+}
+
+static void
+xsc_txq_elts_alloc(struct xsc_txq_data *txq_data)
+{
+ const uint32_t elts_s = 1 << txq_data->elts_n;
+ uint32_t i;
+
+ for (i = 0; i < elts_s; ++i)
+ txq_data->elts[i] = NULL;
+ txq_data->elts_head = 0;
+ txq_data->elts_tail = 0;
+ txq_data->elts_comp = 0;
+}
+
+static void
+xsc_txq_elts_free(struct xsc_txq_data *txq_data)
+{
+ const uint16_t elts_n = 1 << txq_data->elts_n;
+ const uint16_t elts_m = elts_n - 1;
+ uint16_t elts_head = txq_data->elts_head;
+ uint16_t elts_tail = txq_data->elts_tail;
+ struct rte_mbuf *(*elts)[elts_n] = &txq_data->elts;
+
+ txq_data->elts_head = 0;
+ txq_data->elts_tail = 0;
+ txq_data->elts_comp = 0;
+
+ while (elts_tail != elts_head) {
+ struct rte_mbuf *elt = (*elts)[elts_tail & elts_m];
+ rte_pktmbuf_free_seg(elt);
+ ++elts_tail;
+ }
+ PMD_DRV_LOG(DEBUG, "Port %u txq %u free elts", txq_data->port_id, txq_data->idx);
+}
+
+static struct ibv_qp *
+xsc_txq_ibv_qp_create(struct rte_eth_dev *dev, uint16_t idx)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ struct xsc_txq_data *txq_data = xsc_txq_get(dev, idx);
+ struct ibv_qp *qp_obj = NULL;
+ struct ibv_qp_init_attr_ex qp_attr = { 0 };
+ const int desc = 1 << txq_data->elts_n;
+ uint64_t tx_offloads = dev->data->dev_conf.txmode.offloads;
+
+ qp_attr.send_cq = txq_data->cq;
+ qp_attr.recv_cq = txq_data->cq;
+ qp_attr.cap.max_send_wr = desc;
+ qp_attr.cap.max_recv_wr = 0;
+ qp_attr.cap.max_send_sge = 1;
+ qp_attr.qp_type = IBV_QPT_RAW_PACKET;
+ qp_attr.pd = priv->xdev->ibv_pd;
+ qp_attr.sq_sig_all = 0;
+
+ if (tx_offloads & (RTE_ETH_TX_OFFLOAD_TCP_TSO)) {
+ qp_attr.create_flags = XSC_QP_CREATE_RAWPACKET_TSO;
+ qp_attr.comp_mask = IBV_QP_INIT_ATTR_PD | IBV_QP_INIT_ATTR_CREATE_FLAGS;
+ txq_data->tso_en = 1;
+ PMD_DRV_LOG(DEBUG, "Port %u txq %u, create tso qp",
+ dev->data->port_id, idx);
+ } else {
+ qp_attr.create_flags = XSC_QP_CREATE_RAWPACKET_TX;
+ qp_attr.comp_mask = IBV_QP_INIT_ATTR_PD | IBV_QP_INIT_ATTR_CREATE_FLAGS;
+ PMD_DRV_LOG(DEBUG, "Port %u txq %u, create non-tso qp",
+ dev->data->port_id, idx);
+ }
+
+ qp_obj = ibv_create_qp_ex(priv->xdev->ibv_ctx, &qp_attr);
+ if (qp_obj == NULL) {
+ PMD_DRV_LOG(ERR, "Port %u txq %u, create %s qp fail, errno=%d",
+ dev->data->port_id, idx,
+ qp_attr.create_flags & XSC_QP_CREATE_RAWPACKET_TSO ?
+ "tso" : "non-tso", errno);
+
+ if (!(tx_offloads & (RTE_ETH_TX_OFFLOAD_TCP_TSO))) {
+ qp_attr.create_flags = XSC_QP_CREATE_RAWPACKET_TSO;
+ qp_attr.comp_mask = IBV_QP_INIT_ATTR_PD |
+ IBV_QP_INIT_ATTR_CREATE_FLAGS;
+ PMD_DRV_LOG(DEBUG, "Port %u txq %u, recreate tso qp",
+ dev->data->port_id, idx);
+
+ qp_obj = ibv_create_qp_ex(priv->xdev->ibv_ctx, &qp_attr);
+ if (qp_obj == NULL)
+ PMD_DRV_LOG(ERR, "Port %u txq %u, recreate tso qp fail, errno=%d",
+ dev->data->port_id, idx, errno);
+ else
+ txq_data->tso_en = 1;
+ }
+ }
+
+ return qp_obj;
+}
+
+static int
+xsc_txq_ibv_obj_new(struct rte_eth_dev *dev, uint16_t idx)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ struct xsc_txq_data *txq_data = xsc_txq_get(dev, idx);
+ struct xsc_hwinfo *hwinfo = &priv->xdev->hwinfo;
+ struct xscdv_qp qp_info = { 0 };
+ struct xscdv_cq cq_info = { 0 };
+ struct xscdv_obj obj;
+ const int desc = 1 << txq_data->elts_n;
+ uint32_t cqe_n;
+ int ret = 0;
+
+ cqe_n = desc;
+ txq_data->cq = ibv_create_cq(priv->xdev->ibv_ctx, cqe_n, NULL, NULL, 0);
+ if (txq_data->cq == NULL) {
+ PMD_DRV_LOG(ERR, "Port %u txq %u, create cq fail",
+ dev->data->port_id, idx);
+ rte_errno = errno;
+ goto error;
+ }
+
+ txq_data->qp = xsc_txq_ibv_qp_create(dev, idx);
+ if (txq_data->qp == NULL) {
+ rte_errno = errno;
+ goto error;
+ }
+
+ obj.cq.in = txq_data->cq;
+ obj.cq.out = &cq_info;
+ obj.qp.in = txq_data->qp;
+ obj.qp.out = &qp_info;
+ ret = xsc_init_obj(&obj, XSCDV_OBJ_CQ | XSCDV_OBJ_QP);
+ if (ret != 0) {
+ rte_errno = errno;
+ goto error;
+ }
+
+ txq_data->cqe_n = rte_log2_u32(cq_info.cqe_cnt);
+ txq_data->cqe_s = 1 << txq_data->cqe_n;
+ txq_data->cqe_m = txq_data->cqe_s - 1;
+ txq_data->qpn = ((struct ibv_qp *)txq_data->qp)->qp_num;
+ txq_data->wqes = qp_info.sq.buf;
+ txq_data->wqe_n = rte_log2_u32(qp_info.sq.wqe_cnt);
+ txq_data->wqe_s = 1 << txq_data->wqe_n;
+ txq_data->wqe_m = txq_data->wqe_s - 1;
+ txq_data->wqe_ds_n = rte_log2_u32(hwinfo->send_seg_num);
+
+ /* txq doobell */
+ txq_data->qp_db = qp_info.sq.db;
+ /* cqe doobell */
+ txq_data->cq_db = cq_info.db;
+ txq_data->cqn = cq_info.cqn;
+
+ txq_data->cqes = (volatile struct xsc_cqe *)cq_info.buf;
+ txq_data->cq_ci = 0;
+ txq_data->cq_pi = 0;
+ txq_data->wqe_ci = 0;
+ txq_data->wqe_pi = 0;
+ txq_data->wqe_comp = 0;
+ xsc_modify_qpn_status(txq_data->qpn, 1, XSC_CMD_OP_RTR2RTS_QP, priv->xdev);
+ dev->data->tx_queue_state[idx] = RTE_ETH_QUEUE_STATE_STARTED;
+
+ PMD_DRV_LOG(INFO, "Port %u create tx cq, cqe_s:%d, cqe_n:%d, cq_db=%p, cqn:%d",
+ dev->data->port_id,
+ txq_data->cqe_s, txq_data->cqe_n,
+ txq_data->cq_db, txq_data->cqn);
+
+ PMD_DRV_LOG(INFO, "Port %u create tx qp, wqe_s:%d, wqe_n:%d, qp_db=%p, qpn:%d",
+ dev->data->port_id,
+ txq_data->wqe_s, txq_data->wqe_n,
+ txq_data->qp_db, txq_data->qpn);
+
+ return 0;
+
+error:
+ return -rte_errno;
+}
+
+static void
+xsc_txq_ibv_obj_release(struct xsc_txq_data *txq_data)
+{
+ PMD_DRV_LOG(DEBUG, "destroy tx queue %u, portid %u\n",
+ txq_data->idx, txq_data->port_id);
+ if (txq_data->qp != NULL)
+ ibv_destroy_qp(txq_data->qp);
+ if (txq_data->cq != NULL)
+ ibv_destroy_cq(txq_data->cq);
+}
+
+static void
+xsc_ethdev_txq_release(struct rte_eth_dev *dev, uint16_t idx)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ struct xsc_txq_data *txq_data = xsc_txq_get(dev, idx);
+
+ if (txq_data == NULL)
+ return;
+ xsc_txq_ibv_obj_release(txq_data);
+ if (txq_data->fcqs != NULL)
+ rte_free(txq_data->fcqs);
+ txq_data->fcqs = NULL;
+ xsc_txq_elts_free(txq_data);
+ rte_free(txq_data);
+ (*priv->txqs)[idx] = NULL;
+
+ dev->data->tx_queues[idx] = NULL;
+ dev->data->tx_queue_state[idx] = RTE_ETH_QUEUE_STATE_STOPPED;
+}
+
+static int
+xsc_txq_start(struct rte_eth_dev *dev)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ struct xsc_txq_data *txq_data;
+ uint16_t i;
+ int ret;
+ size_t size;
+
+ for (i = 0; i != priv->num_sq; ++i) {
+ txq_data = xsc_txq_get(dev, i);
+ xsc_txq_elts_alloc(txq_data);
+ ret = xsc_txq_ibv_obj_new(dev, i);
+ if (ret < 0)
+ goto error;
+
+ size = txq_data->cqe_s * sizeof(*txq_data->fcqs);
+ txq_data->fcqs = rte_zmalloc(NULL, size, RTE_CACHE_LINE_SIZE);
+ if (!txq_data->fcqs) {
+ PMD_DRV_LOG(ERR, "Port %u txq %u alloc fcqs memory failed",
+ dev->data->port_id, i);
+ rte_errno = ENOMEM;
+ goto error;
+ }
+ }
+
+ return 0;
+
+error:
+ /* Queue resources are released by xsc_ethdev_start calling the stop interface */
+ return -rte_errno;
+}
+
+static int
+xsc_rxq_elts_alloc(struct xsc_rxq_data *rxq_data)
+{
+ uint32_t elts_s = rxq_data->wqe_s;
+ struct rte_mbuf *mbuf;
+ uint32_t i;
+
+ for (i = 0; (i != elts_s); ++i) {
+ mbuf = rte_pktmbuf_alloc(rxq_data->mp);
+ if (mbuf == NULL) {
+ PMD_DRV_LOG(ERR, "Port %u rxq %u empty mbuf pool",
+ rxq_data->port_id, rxq_data->idx);
+ rte_errno = ENOMEM;
+ goto error;
+ }
+
+ mbuf->port = rxq_data->port_id;
+ mbuf->nb_segs = 1;
+ rte_pktmbuf_data_len(mbuf) = rte_pktmbuf_data_room_size(rxq_data->mp);
+ rte_pktmbuf_pkt_len(mbuf) = rte_pktmbuf_data_room_size(rxq_data->mp);
+ (*rxq_data->elts)[i] = mbuf;
+ }
+
+ return 0;
+error:
+ elts_s = i;
+ for (i = 0; (i != elts_s); ++i) {
+ if ((*rxq_data->elts)[i] != NULL)
+ rte_pktmbuf_free_seg((*rxq_data->elts)[i]);
+ (*rxq_data->elts)[i] = NULL;
+ }
+
+ PMD_DRV_LOG(ERR, "Port %u rxq %u start failed, free elts",
+ rxq_data->port_id, rxq_data->idx);
+
+ return -rte_errno;
+}
+
+static void
+xsc_rxq_elts_free(struct xsc_rxq_data *rxq_data)
+{
+ uint16_t i;
+
+ if (rxq_data->elts == NULL)
+ return;
+ for (i = 0; i != rxq_data->wqe_s; ++i) {
+ if ((*rxq_data->elts)[i] != NULL)
+ rte_pktmbuf_free_seg((*rxq_data->elts)[i]);
+ (*rxq_data->elts)[i] = NULL;
+ }
+
+ PMD_DRV_LOG(DEBUG, "Port %u rxq %u free elts", rxq_data->port_id, rxq_data->idx);
+}
+
+static void
+xsc_rxq_rss_obj_release(struct xsc_ethdev_priv *priv, struct xsc_rxq_data *rxq_data)
+{
+ struct xsc_destroy_qp_mbox_in in = { .hdr = { 0 } };
+ struct xsc_destroy_qp_mbox_out out = { .hdr = { 0 } };
+ int ret, in_len, out_len;
+ uint32_t qpn = rxq_data->qpn;
+
+ xsc_modify_qpn_status(qpn, 1, XSC_CMD_OP_QP_2RST, priv->xdev);
+
+ in_len = sizeof(struct xsc_destroy_qp_mbox_in);
+ out_len = sizeof(struct xsc_destroy_qp_mbox_out);
+ in.hdr.opcode = rte_cpu_to_be_16(XSC_CMD_OP_DESTROY_QP);
+ in.qpn = rte_cpu_to_be_32(rxq_data->qpn);
+
+ ret = xsc_mailbox_exec(priv->xdev, &in, in_len, &out, out_len);
+ if (ret != 0 || out.hdr.status != 0) {
+ PMD_DRV_LOG(ERR, "release rss rq failed, port id=%d, "
+ "qid=%d, err=%d, out.status=%u\n",
+ rxq_data->port_id, rxq_data->idx, ret, out.hdr.status);
+ rte_errno = ENOEXEC;
+ return;
+ }
+
+ if (rxq_data->rq_pas != NULL)
+ rte_memzone_free(rxq_data->rq_pas);
+
+ if (rxq_data->cq != NULL)
+ ibv_destroy_cq(rxq_data->cq);
+ rxq_data->cq = NULL;
+}
+
+static void
+xsc_ethdev_rxq_release(struct rte_eth_dev *dev, uint16_t idx)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ struct xsc_rxq_data *rxq_data = xsc_rxq_get(dev, idx);
+
+ if (rxq_data == NULL)
+ return;
+ xsc_rxq_rss_obj_release(priv, rxq_data);
+ xsc_rxq_elts_free(rxq_data);
+ rte_free(rxq_data);
+ (*priv->rxqs)[idx] = NULL;
+
+ dev->data->rx_queues[idx] = NULL;
+ dev->data->rx_queue_state[idx] = RTE_ETH_QUEUE_STATE_STOPPED;
+}
+
+static void
+xsc_rxq_initialize(struct xsc_ethdev_priv *priv, struct xsc_rxq_data *rxq_data)
+{
+ const uint32_t wqe_n = rxq_data->wqe_s;
+ uint32_t i;
+ uint32_t seg_len = 0;
+ struct xsc_hwinfo *hwinfo = &priv->xdev->hwinfo;
+ uint32_t rx_ds_num = hwinfo->recv_seg_num;
+ uint32_t log2ds = rte_log2_u32(rx_ds_num);
+ uintptr_t addr;
+ struct rte_mbuf *mbuf;
+ volatile struct xsc_wqe_data_seg *seg;
+
+ for (i = 0; (i != wqe_n); ++i) {
+ mbuf = (*rxq_data->elts)[i];
+ seg = &((volatile struct xsc_wqe_data_seg *)rxq_data->wqes)[i * rx_ds_num];
+ addr = (uintptr_t)rte_pktmbuf_iova(mbuf);
+ seg_len = rte_pktmbuf_data_len(mbuf);
+ *seg = (struct xsc_wqe_data_seg){
+ .va = rte_cpu_to_le_64(addr),
+ .seg_len = rte_cpu_to_le_32(seg_len),
+ .lkey = 0,
+ };
+ }
+
+ rxq_data->rq_ci = wqe_n;
+ rxq_data->sge_n = rte_log2_u32(rx_ds_num);
+
+ rte_io_wmb();
+ union xsc_recv_doorbell recv_db = {
+ .recv_data = 0
+ };
+
+ recv_db.next_pid = wqe_n << log2ds;
+ recv_db.qp_num = rxq_data->qpn;
+ *rxq_data->rq_db = rte_cpu_to_le_32(recv_db.recv_data);
+}
+
+static int
+xsc_rxq_rss_qp_create(struct rte_eth_dev *dev)
+{
+ struct xsc_create_multiqp_mbox_in *in;
+ struct xsc_create_qp_request *req;
+ struct xsc_create_multiqp_mbox_out *out;
+ uint8_t log_ele;
+ uint64_t iova;
+ int wqe_n;
+ int in_len, out_len, cmd_len;
+ int entry_total_len, entry_len;
+ uint8_t log_rq_sz, log_sq_sz = 0;
+ int j, ret;
+ uint16_t i, pa_num;
+ int rqn_base;
+ size_t page_size = PAGE_SIZE;
+ struct xsc_rxq_data *rxq_data;
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ struct xsc_hwinfo *hwinfo = &priv->xdev->hwinfo;
+ int port_id = dev->data->port_id;
+ char name[64] = { 0 };
+
+ /* paslen */
+ rxq_data = xsc_rxq_get(dev, 0);
+ log_ele = rte_log2_u32(sizeof(struct xsc_wqe_data_seg));
+ wqe_n = rxq_data->wqe_s;
+ log_rq_sz = rte_log2_u32(wqe_n * hwinfo->recv_seg_num);
+
+ pa_num = XSC_DIV_ROUND_UP((1 << (log_rq_sz + log_sq_sz + log_ele)), page_size);
+ entry_len = sizeof(struct xsc_create_qp_request) +
+ sizeof(uint64_t) * pa_num;
+ entry_total_len = entry_len * priv->num_rq;
+
+ in_len = sizeof(struct xsc_create_multiqp_mbox_in) + entry_total_len;
+ out_len = sizeof(struct xsc_create_multiqp_mbox_out) + entry_total_len;
+ cmd_len = RTE_MAX(in_len, out_len);
+ in = rte_zmalloc(NULL, cmd_len, RTE_CACHE_LINE_SIZE);
+ if (in == NULL) {
+ rte_errno = ENOMEM;
+ PMD_DRV_LOG(ERR, "Alloc rss qp create cmd memory failed\n");
+ goto error;
+ }
+
+ in->qp_num = rte_cpu_to_be_16((uint16_t)priv->num_rq);
+ in->qp_type = XSC_QUEUE_TYPE_RAW;
+ in->req_len = rte_cpu_to_be_32(cmd_len);
+
+ for (i = 0; i < priv->num_rq; i++) {
+ rxq_data = (*priv->rxqs)[i];
+ req = (struct xsc_create_qp_request *)(&in->data[0] + entry_len * i);
+ req->input_qpn = rte_cpu_to_be_16(0); /* useless for eth */
+ req->pa_num = rte_cpu_to_be_16(pa_num);
+ req->qp_type = XSC_QUEUE_TYPE_RAW;
+ req->log_rq_sz = log_rq_sz;
+ req->cqn_recv = rte_cpu_to_be_16((uint16_t)rxq_data->cqn);
+ req->cqn_send = req->cqn_recv;
+ req->glb_funcid = rte_cpu_to_be_16((uint16_t)hwinfo->func_id);
+ /* alloc pas addr */
+ sprintf(name, "wqe_mem_rx_%d_%d", port_id, i);
+ rxq_data->rq_pas = rte_memzone_reserve_aligned(name,
+ (page_size * pa_num), SOCKET_ID_ANY, 0, page_size);
+ if (rxq_data->rq_pas == NULL) {
+ rte_errno = ENOMEM;
+ PMD_DRV_LOG(ERR, "Alloc rxq pas memory failed\n");
+ goto error;
+ }
+
+ iova = rxq_data->rq_pas->iova;
+ for (j = 0; j < pa_num; j++)
+ req->pas[j] = rte_cpu_to_be_64(iova + j * page_size);
+ }
+
+ in->hdr.opcode = rte_cpu_to_be_16(XSC_CMD_OP_CREATE_MULTI_QP);
+ out = (struct xsc_create_multiqp_mbox_out *)in;
+ ret = xsc_mailbox_exec(priv->xdev, in, in_len, out, out_len);
+ if (ret != 0 || out->hdr.status != 0) {
+ PMD_DRV_LOG(ERR, "Create rss rq failed, port id=%d, "
+ "qp_num=%d, type=%d, err=%d, out.status=%u\n",
+ port_id, priv->num_rq, XSC_QUEUE_TYPE_RAW, ret, out->hdr.status);
+ rte_errno = ENOEXEC;
+ goto error;
+ }
+ rqn_base = rte_be_to_cpu_32(out->qpn_base) & 0xffffff;
+
+ for (i = 0; i < priv->num_rq; i++) {
+ rxq_data = xsc_rxq_get(dev, i);
+ rxq_data->wqes = rxq_data->rq_pas->addr;
+ rxq_data->rq_db = (uint32_t *)((uint8_t *)priv->xdev->bar_addr +
+ XSC_RXQ_DB_PF_OFFSET);
+ rxq_data->qpn = rqn_base + i;
+ xsc_modify_qpn_status(rxq_data->qpn, 1, XSC_CMD_OP_RTR2RTS_QP,
+ priv->xdev);
+ xsc_rxq_initialize(priv, rxq_data);
+ rxq_data->cq_ci = 0;
+ priv->dev_data->rx_queue_state[i] = RTE_ETH_QUEUE_STATE_STARTED;
+ PMD_DRV_LOG(INFO, "Port %u create rx qp, wqe_s:%d, wqe_n:%d, qp_db=%p, qpn:%d",
+ dev->data->port_id,
+ rxq_data->wqe_s, rxq_data->wqe_n,
+ rxq_data->rq_db, rxq_data->qpn);
+ }
+
+ if (in != NULL)
+ rte_free(in);
+ return 0;
+
+error:
+ if (in != NULL)
+ rte_free(in);
+ return -rte_errno;
+}
+
+static int
+xsc_rxq_rss_obj_new(struct rte_eth_dev *dev)
+{
+ int ret;
+ uint32_t i;
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ uint16_t port_id = priv->dev_data->port_id;
+ struct xsc_rxq_data *rxq_data;
+ struct xscdv_cq cq_info = { 0 };
+ struct xscdv_obj obj;
+ uint32_t cqe_s;
+
+ /* Create CQ */
+ for (i = 0; i < priv->num_rq; ++i) {
+ rxq_data = xsc_rxq_get(dev, i);
+ cqe_s = rxq_data->wqe_s;
+ rxq_data->cq = ibv_create_cq(priv->xdev->ibv_ctx, cqe_s, NULL, NULL, 0);
+ if (rxq_data->cq == NULL) {
+ PMD_DRV_LOG(ERR, "Port %u rxq %u create cq fail", port_id, i);
+ rte_errno = errno;
+ goto error;
+ }
+ obj.cq.in = rxq_data->cq;
+ obj.cq.out = &cq_info;
+ ret = xsc_init_obj(&obj, XSCDV_OBJ_CQ);
+ if (ret) {
+ rte_errno = errno;
+ goto error;
+ }
+
+ rxq_data->cqe_n = rte_log2_u32(cq_info.cqe_cnt);
+ rxq_data->cqe_s = 1 << rxq_data->cqe_n;
+ rxq_data->cqe_m = rxq_data->cqe_s - 1;
+ rxq_data->cqes = (volatile struct xsc_cqe (*)[])(uintptr_t)cq_info.buf;
+ rxq_data->cq_db = cq_info.db;
+ rxq_data->cqn = cq_info.cqn;
+ PMD_DRV_LOG(INFO, "Port %u create rx cq, cqe_s:%d, cqe_n:%d, cq_db=%p, cqn:%d",
+ dev->data->port_id,
+ rxq_data->cqe_s, rxq_data->cqe_n,
+ rxq_data->cq_db, rxq_data->cqn);
+ }
+
+ ret = xsc_rxq_rss_qp_create(dev);
+ if (ret != 0) {
+ PMD_DRV_LOG(ERR, "Port %u rss rxq create fail", port_id);
+ goto error;
+ }
+ return 0;
+
+error:
+ return -rte_errno;
+}
+
+static void
+xsc_txq_stop(struct rte_eth_dev *dev)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ uint16_t i;
+
+ for (i = 0; i != priv->num_sq; ++i)
+ xsc_ethdev_txq_release(dev, i);
+ priv->txqs = NULL;
+}
+
+static void
+xsc_rxq_stop(struct rte_eth_dev *dev)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ uint16_t i;
+
+ for (i = 0; i != priv->num_rq; ++i)
+ xsc_ethdev_rxq_release(dev, i);
+ priv->rxqs = NULL;
+}
+
+static int
+xsc_rxq_start(struct rte_eth_dev *dev)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ struct xsc_rxq_data *rxq_data;
+ uint16_t i;
+ int ret;
+
+ for (i = 0; i != priv->num_rq; ++i) {
+ rxq_data = xsc_rxq_get(dev, i);
+ if (dev->data->rx_queue_state[i] != RTE_ETH_QUEUE_STATE_STARTED) {
+ ret = xsc_rxq_elts_alloc(rxq_data);
+ if (ret != 0)
+ goto error;
+ }
+ }
+
+ ret = xsc_rxq_rss_obj_new(dev);
+ if (ret != 0)
+ goto error;
+
+ return 0;
+error:
+ /* Queue resources are released by xsc_ethdev_start calling the stop interface */
+ return -rte_errno;
+}
+
+static int
+xsc_ethdev_start(struct rte_eth_dev *dev)
+{
+ int ret;
+
+ ret = xsc_txq_start(dev);
+ if (ret) {
+ PMD_DRV_LOG(ERR, "Port %u txq start failed: %s",
+ dev->data->port_id, strerror(rte_errno));
+ goto error;
+ }
+
+ ret = xsc_rxq_start(dev);
+ if (ret) {
+ PMD_DRV_LOG(ERR, "Port %u Rx queue start failed: %s",
+ dev->data->port_id, strerror(rte_errno));
+ goto error;
+ }
+
+ dev->data->dev_started = 1;
+
+ rte_wmb();
+ dev->rx_pkt_burst = xsc_rx_burst;
+ dev->tx_pkt_burst = xsc_tx_burst;
+
+ return 0;
+
+error:
+ dev->data->dev_started = 0;
+ xsc_txq_stop(dev);
+ xsc_rxq_stop(dev);
+ return -rte_errno;
+}
+
+static int
+xsc_ethdev_stop(struct rte_eth_dev *dev)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+ uint16_t i;
+
+ PMD_DRV_LOG(DEBUG, "Port %u stopping", dev->data->port_id);
+ dev->data->dev_started = 0;
+ dev->rx_pkt_burst = rte_eth_pkt_burst_dummy;
+ dev->tx_pkt_burst = rte_eth_pkt_burst_dummy;
+ rte_wmb();
+
+ rte_delay_us_sleep(1000 * priv->num_rq);
+ for (i = 0; i < priv->num_rq; ++i)
+ dev->data->rx_queue_state[i] = RTE_ETH_QUEUE_STATE_STOPPED;
+ for (i = 0; i < priv->num_sq; ++i)
+ dev->data->tx_queue_state[i] = RTE_ETH_QUEUE_STATE_STOPPED;
+ return 0;
+}
+
+static int
+xsc_ethdev_close(struct rte_eth_dev *dev)
+{
+ struct xsc_ethdev_priv *priv = TO_XSC_ETHDEV_PRIV(dev);
+
+ PMD_DRV_LOG(DEBUG, "Port %u closing", dev->data->port_id);
+ dev->rx_pkt_burst = rte_eth_pkt_burst_dummy;
+ dev->tx_pkt_burst = rte_eth_pkt_burst_dummy;
+ rte_wmb();
+
+ rte_delay_us_sleep(1000);
+ xsc_txq_stop(dev);
+ rte_delay_us_sleep(1000);
+ xsc_rxq_stop(dev);
+
+ if (priv->rss_conf.rss_key != NULL)
+ rte_free(priv->rss_conf.rss_key);
+
+ /* priv is released in rte_eth_dev_release_port */
+
+ dev->data->mac_addrs = NULL;
+ return 0;
+}
+
static int
xsc_ethdev_rx_queue_setup(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc,
uint32_t socket, const struct rte_eth_rxconf *conf,
@@ -219,10 +944,23 @@ xsc_ethdev_tx_queue_setup(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc,
return 0;
}
+static int
+xsc_ethdev_link_update(__rte_unused struct rte_eth_dev *dev,
+ __rte_unused int wait_to_complete)
+{
+ return 0;
+}
+
const struct eth_dev_ops xsc_dev_ops = {
.dev_configure = xsc_ethdev_configure,
+ .dev_start = xsc_ethdev_start,
+ .dev_stop = xsc_ethdev_stop,
+ .dev_close = xsc_ethdev_close,
+ .link_update = xsc_ethdev_link_update,
.rx_queue_setup = xsc_ethdev_rx_queue_setup,
.tx_queue_setup = xsc_ethdev_tx_queue_setup,
+ .rx_queue_release = xsc_ethdev_rxq_release,
+ .tx_queue_release = xsc_ethdev_txq_release,
.rss_hash_update = xsc_ethdev_rss_hash_update,
.rss_hash_conf_get = xsc_ethdev_rss_hash_conf_get,
};
diff --git a/drivers/net/xsc/xsc_ethdev.h b/drivers/net/xsc/xsc_ethdev.h
index fb92d47dd0..00d3671c31 100644
--- a/drivers/net/xsc/xsc_ethdev.h
+++ b/drivers/net/xsc/xsc_ethdev.h
@@ -5,7 +5,17 @@
#ifndef _XSC_ETHDEV_H_
#define _XSC_ETHDEV_H_
+#include "xsc_dev.h"
+#include "xsc_utils.h"
+#include "xsc_defs.h"
+
#define XSC_RSS_HASH_KEY_LEN 52
+#define XSC_RXQ_DB_PF_OFFSET 0x4804000
+
+#define XSC_CMD_OP_DESTROY_QP 0x501
+#define XSC_CMD_OP_RTR2RTS_QP 0x504
+#define XSC_CMD_OP_QP_2RST 0x50A
+#define XSC_CMD_OP_CREATE_MULTI_QP 0x515
#define XSC_MAX_DESC_NUMBER 1024
#define XSC_RX_FREE_THRESH 32
diff --git a/drivers/net/xsc/xsc_rxtx.c b/drivers/net/xsc/xsc_rxtx.c
new file mode 100644
index 0000000000..66b1511c6a
--- /dev/null
+++ b/drivers/net/xsc/xsc_rxtx.c
@@ -0,0 +1,22 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright 2024 Yunsilicon Technology Co., Ltd.
+ */
+
+#include "xsc_log.h"
+#include "xsc_defs.h"
+#include "xsc_dev.h"
+#include "xsc_ethdev.h"
+#include "xsc_rxtx.h"
+
+uint16_t
+xsc_rx_burst(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n)
+{
+ return 0;
+}
+
+uint16_t
+xsc_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n)
+{
+ return 0;
+}
+
diff --git a/drivers/net/xsc/xsc_rxtx.h b/drivers/net/xsc/xsc_rxtx.h
index 9b072e06e0..0b2fded42b 100644
--- a/drivers/net/xsc/xsc_rxtx.h
+++ b/drivers/net/xsc/xsc_rxtx.h
@@ -5,6 +5,61 @@
#ifndef _XSC_RXTX_H_
#define _XSC_RXTX_H_
+struct xsc_send_wqe_ctrl_seg {
+ __le32 msg_opcode:8;
+ __le32 with_immdt:1;
+ __le32 csum_en:2;
+ __le32 ds_data_num:5;
+ __le32 wqe_id:16;
+ __le32 msg_len;
+ union {
+ __le32 opcode_data;
+ struct {
+ uint8_t has_pph:1;
+ uint8_t so_type:1;
+ __le16 so_data_size:14;
+ uint8_t rsv1:8;
+ uint8_t so_hdr_len:8;
+ };
+ struct {
+ __le16 desc_id;
+ __le16 is_last_wqe:1;
+ __le16 dst_qp_id:15;
+ };
+ };
+ uint8_t se:1;
+ uint8_t ce:1;
+ __le32 rsv2:30;
+};
+
+struct xsc_wqe_data_seg {
+ union {
+ uint32_t in_line : 1;
+ struct {
+ uint32_t rsv1 : 1;
+ __le32 seg_len : 31;
+ __le32 lkey;
+ __le64 va;
+ };
+ struct {
+ uint32_t rsv2 : 1;
+ uint32_t len : 7;
+ uint8_t in_line_data[15];
+ };
+ };
+} __rte_packed;
+
+struct xsc_wqe {
+ union {
+ struct xsc_send_wqe_ctrl_seg cseg;
+ uint32_t ctrl[4];
+ };
+ union {
+ struct xsc_wqe_data_seg dseg[XSC_SEND_WQE_DS];
+ uint8_t data[XSC_ESEG_EXTRA_DATA_SIZE];
+ };
+} __rte_packed;
+
struct xsc_cqe {
union {
uint8_t msg_opcode;
@@ -111,5 +166,16 @@ struct __rte_cache_aligned xsc_rxq_data {
uint32_t rss_hash:1; /* RSS hash enabled */
};
-#endif /* _XSC_RXTX_H_ */
+union xsc_recv_doorbell {
+ struct {
+ uint32_t next_pid : 13;
+ uint32_t qp_num : 15;
+ uint32_t rsv : 4;
+ };
+ uint32_t recv_data;
+};
+
+uint16_t xsc_rx_burst(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n);
+uint16_t xsc_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n);
+#endif /* _XSC_RXTX_H_ */
--
2.25.1
^ permalink raw reply [flat|nested] only message in thread
only message in thread, other threads:[~2024-09-11 2:09 UTC | newest]
Thread overview: (only message) (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2024-09-11 2:07 [PATCH v2 13/19] net/xsc: add ethdev start and stop ops WanRenyong
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).