DPDK patches and discussions
 help / color / mirror / Atom feed
* [PATCH] net/octeon_ep: add octeon CN10K support
@ 2022-08-16 10:34 Sathesh Edara
  2022-08-27 13:07 ` Jerin Jacob
  2022-08-29 15:45 ` [PATCH v2] net/octeon_ep: support CN10K SoC Sathesh Edara
  0 siblings, 2 replies; 5+ messages in thread
From: Sathesh Edara @ 2022-08-16 10:34 UTC (permalink / raw)
  To: sburla, jerinj, sedara, Radha Mohan Chintakuntla, Veerasenareddy Burru
  Cc: dev

Added Octeon CN10K EP PMD functionality.

Signed-off-by: Sathesh Edara <sedara@marvell.com>
---

This patch depends on https://patches.dpdk.org/project/dpdk/list/?series=24103

 drivers/net/octeon_ep/cnxk_ep_vf.c    | 370 ++++++++++++++++++++++++++
 drivers/net/octeon_ep/cnxk_ep_vf.h    | 161 +++++++++++
 drivers/net/octeon_ep/meson.build     |   1 +
 drivers/net/octeon_ep/otx2_ep_vf.c    |  52 ++--
 drivers/net/octeon_ep/otx_ep_common.h |   4 +
 drivers/net/octeon_ep/otx_ep_ethdev.c |   9 +
 6 files changed, 571 insertions(+), 26 deletions(-)
 create mode 100644 drivers/net/octeon_ep/cnxk_ep_vf.c
 create mode 100644 drivers/net/octeon_ep/cnxk_ep_vf.h

diff --git a/drivers/net/octeon_ep/cnxk_ep_vf.c b/drivers/net/octeon_ep/cnxk_ep_vf.c
new file mode 100644
index 0000000000..04c516d037
--- /dev/null
+++ b/drivers/net/octeon_ep/cnxk_ep_vf.c
@@ -0,0 +1,370 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+
+#include <rte_common.h>
+#include <rte_cycles.h>
+#include "otx_ep_common.h"
+#include "cnxk_ep_vf.h"
+
+static void
+cnxk_ep_vf_setup_global_iq_reg(struct otx_ep_device *otx_ep, int q_no)
+{
+	volatile uint64_t reg_val = 0ull;
+
+	/* Select ES, RO, NS, RDSIZE,DPTR Format#0 for IQs
+	 * IS_64B is by default enabled.
+	 */
+	reg_val = oct_ep_read64(otx_ep->hw_addr +
+				CNXK_EP_R_IN_CONTROL(q_no));
+
+	reg_val |= CNXK_EP_R_IN_CTL_RDSIZE;
+	reg_val |= CNXK_EP_R_IN_CTL_IS_64B;
+	reg_val |= CNXK_EP_R_IN_CTL_ESR;
+
+	oct_ep_write64(reg_val, otx_ep->hw_addr +
+		       CNXK_EP_R_IN_CONTROL(q_no));
+}
+
+static void
+cnxk_ep_vf_setup_global_oq_reg(struct otx_ep_device *otx_ep, int q_no)
+{
+	volatile uint64_t reg_val = 0ull;
+
+	reg_val = oct_ep_read64(otx_ep->hw_addr +
+				CNXK_EP_R_OUT_CONTROL(q_no));
+
+	reg_val &= ~(CNXK_EP_R_OUT_CTL_IMODE);
+	reg_val &= ~(CNXK_EP_R_OUT_CTL_ROR_P);
+	reg_val &= ~(CNXK_EP_R_OUT_CTL_NSR_P);
+	reg_val &= ~(CNXK_EP_R_OUT_CTL_ROR_I);
+	reg_val &= ~(CNXK_EP_R_OUT_CTL_NSR_I);
+	reg_val &= ~(CNXK_EP_R_OUT_CTL_ROR_D);
+	reg_val &= ~(CNXK_EP_R_OUT_CTL_NSR_D);
+	reg_val &= ~(CNXK_EP_R_OUT_CTL_ES_I | CNXK_EP_R_OUT_CTL_ES_D);
+
+	/* INFO/DATA ptr swap is required  */
+	reg_val |= (CNXK_EP_R_OUT_CTL_ES_P);
+	oct_ep_write64(reg_val, otx_ep->hw_addr +
+		       CNXK_EP_R_OUT_CONTROL(q_no));
+}
+
+static void
+cnxk_ep_vf_setup_global_input_regs(struct otx_ep_device *otx_ep)
+{
+	uint64_t q_no = 0ull;
+
+	for (q_no = 0; q_no < (otx_ep->sriov_info.rings_per_vf); q_no++)
+		cnxk_ep_vf_setup_global_iq_reg(otx_ep, q_no);
+}
+
+static void
+cnxk_ep_vf_setup_global_output_regs(struct otx_ep_device *otx_ep)
+{
+	uint32_t q_no;
+
+	for (q_no = 0; q_no < (otx_ep->sriov_info.rings_per_vf); q_no++)
+		cnxk_ep_vf_setup_global_oq_reg(otx_ep, q_no);
+}
+
+static void
+cnxk_ep_vf_setup_device_regs(struct otx_ep_device *otx_ep)
+{
+	cnxk_ep_vf_setup_global_input_regs(otx_ep);
+	cnxk_ep_vf_setup_global_output_regs(otx_ep);
+}
+
+static void
+cnxk_ep_vf_setup_iq_regs(struct otx_ep_device *otx_ep, uint32_t iq_no)
+{
+	struct otx_ep_instr_queue *iq = otx_ep->instr_queue[iq_no];
+	volatile uint64_t reg_val = 0ull;
+
+	reg_val = oct_ep_read64(otx_ep->hw_addr +
+				CNXK_EP_R_IN_CONTROL(iq_no));
+
+	/* Wait till IDLE to set to 1, not supposed to configure BADDR
+	 * as long as IDLE is 0
+	 */
+	if (!(reg_val & CNXK_EP_R_IN_CTL_IDLE)) {
+		do {
+			reg_val = oct_ep_read64(otx_ep->hw_addr +
+					      CNXK_EP_R_IN_CONTROL(iq_no));
+		} while (!(reg_val & CNXK_EP_R_IN_CTL_IDLE));
+	}
+
+	/* Write the start of the input queue's ring and its size  */
+	oct_ep_write64(iq->base_addr_dma, otx_ep->hw_addr +
+		     CNXK_EP_R_IN_INSTR_BADDR(iq_no));
+	oct_ep_write64(iq->nb_desc, otx_ep->hw_addr +
+		     CNXK_EP_R_IN_INSTR_RSIZE(iq_no));
+
+	/* Remember the doorbell & instruction count register addr
+	 * for this queue
+	 */
+	iq->doorbell_reg = (uint8_t *)otx_ep->hw_addr +
+			   CNXK_EP_R_IN_INSTR_DBELL(iq_no);
+	iq->inst_cnt_reg = (uint8_t *)otx_ep->hw_addr +
+			   CNXK_EP_R_IN_CNTS(iq_no);
+
+	otx_ep_dbg("InstQ[%d]:dbell reg @ 0x%p instcnt_reg @ 0x%p",
+		   iq_no, iq->doorbell_reg, iq->inst_cnt_reg);
+
+	do {
+		reg_val = rte_read32(iq->inst_cnt_reg);
+		rte_write32(reg_val, iq->inst_cnt_reg);
+	} while (reg_val != 0);
+
+	/* IN INTR_THRESHOLD is set to max(FFFFFFFF) which disable the IN INTR
+	 * to raise
+	 */
+	oct_ep_write64(OTX_EP_CLEAR_SDP_IN_INT_LVLS,
+		     otx_ep->hw_addr + CNXK_EP_R_IN_INT_LEVELS(iq_no));
+}
+
+static void
+cnxk_ep_vf_setup_oq_regs(struct otx_ep_device *otx_ep, uint32_t oq_no)
+{
+	volatile uint64_t reg_val = 0ull;
+	uint64_t oq_ctl = 0ull;
+	struct otx_ep_droq *droq = otx_ep->droq[oq_no];
+
+	/* Wait on IDLE to set to 1, supposed to configure BADDR
+	 * as log as IDLE is 0
+	 */
+	reg_val = oct_ep_read64(otx_ep->hw_addr +
+				CNXK_EP_R_OUT_CONTROL(oq_no));
+
+	while (!(reg_val & CNXK_EP_R_OUT_CTL_IDLE)) {
+		reg_val = oct_ep_read64(otx_ep->hw_addr +
+				      CNXK_EP_R_OUT_CONTROL(oq_no));
+	}
+
+	oct_ep_write64(droq->desc_ring_dma, otx_ep->hw_addr +
+		     CNXK_EP_R_OUT_SLIST_BADDR(oq_no));
+	oct_ep_write64(droq->nb_desc, otx_ep->hw_addr +
+		     CNXK_EP_R_OUT_SLIST_RSIZE(oq_no));
+
+	oq_ctl = oct_ep_read64(otx_ep->hw_addr +
+			       CNXK_EP_R_OUT_CONTROL(oq_no));
+
+	/* Clear the ISIZE and BSIZE (22-0) */
+	oq_ctl &= ~(OTX_EP_CLEAR_ISIZE_BSIZE);
+
+	/* Populate the BSIZE (15-0) */
+	oq_ctl |= (droq->buffer_size & OTX_EP_DROQ_BUFSZ_MASK);
+
+	oct_ep_write64(oq_ctl, otx_ep->hw_addr +
+		       CNXK_EP_R_OUT_CONTROL(oq_no));
+
+	/* Mapped address of the pkt_sent and pkts_credit regs */
+	droq->pkts_sent_reg = (uint8_t *)otx_ep->hw_addr +
+			      CNXK_EP_R_OUT_CNTS(oq_no);
+	droq->pkts_credit_reg = (uint8_t *)otx_ep->hw_addr +
+				CNXK_EP_R_OUT_SLIST_DBELL(oq_no);
+
+	rte_write64(OTX_EP_CLEAR_OUT_INT_LVLS,
+		    otx_ep->hw_addr + CNXK_EP_R_OUT_INT_LEVELS(oq_no));
+
+	/* Clear PKT_CNT register */
+	rte_write64(OTX_EP_CLEAR_SDP_OUT_PKT_CNT, (uint8_t *)otx_ep->hw_addr +
+		    CNXK_EP_R_OUT_PKT_CNT(oq_no));
+
+	/* Clear the OQ doorbell  */
+	rte_write32(OTX_EP_CLEAR_SLIST_DBELL, droq->pkts_credit_reg);
+	while ((rte_read32(droq->pkts_credit_reg) != 0ull)) {
+		rte_write32(OTX_EP_CLEAR_SLIST_DBELL, droq->pkts_credit_reg);
+		rte_delay_ms(1);
+	}
+	otx_ep_dbg("SDP_R[%d]_credit:%x", oq_no,
+		   rte_read32(droq->pkts_credit_reg));
+
+	/* Clear the OQ_OUT_CNTS doorbell  */
+	reg_val = rte_read32(droq->pkts_sent_reg);
+	rte_write32((uint32_t)reg_val, droq->pkts_sent_reg);
+
+	otx_ep_dbg("SDP_R[%d]_sent: %x", oq_no,
+		   rte_read32(droq->pkts_sent_reg));
+
+	while (((rte_read32(droq->pkts_sent_reg)) != 0ull)) {
+		reg_val = rte_read32(droq->pkts_sent_reg);
+		rte_write32((uint32_t)reg_val, droq->pkts_sent_reg);
+		rte_delay_ms(1);
+	}
+	otx_ep_dbg("SDP_R[%d]_sent: %x", oq_no,
+		   rte_read32(droq->pkts_sent_reg));
+}
+
+static int
+cnxk_ep_vf_enable_iq(struct otx_ep_device *otx_ep, uint32_t q_no)
+{
+	uint64_t loop = OTX_EP_BUSY_LOOP_COUNT;
+	uint64_t reg_val = 0ull;
+
+	/* Resetting doorbells during IQ enabling also to handle abrupt
+	 * guest reboot. IQ reset does not clear the doorbells.
+	 */
+	oct_ep_write64(0xFFFFFFFF, otx_ep->hw_addr +
+		     CNXK_EP_R_IN_INSTR_DBELL(q_no));
+
+	while (((oct_ep_read64(otx_ep->hw_addr +
+		 CNXK_EP_R_IN_INSTR_DBELL(q_no))) != 0ull) && loop--) {
+		rte_delay_ms(1);
+	}
+
+	if (!loop) {
+		otx_ep_err("INSTR DBELL not coming back to 0\n");
+		return -EIO;
+	}
+
+	reg_val = oct_ep_read64(otx_ep->hw_addr + CNXK_EP_R_IN_ENABLE(q_no));
+	reg_val |= 0x1ull;
+
+	oct_ep_write64(reg_val, otx_ep->hw_addr + CNXK_EP_R_IN_ENABLE(q_no));
+
+	otx_ep_info("IQ[%d] enable done", q_no);
+
+	return 0;
+}
+
+static int
+cnxk_ep_vf_enable_oq(struct otx_ep_device *otx_ep, uint32_t q_no)
+{
+	uint64_t reg_val = 0ull;
+
+	reg_val = oct_ep_read64(otx_ep->hw_addr +
+				CNXK_EP_R_OUT_ENABLE(q_no));
+	reg_val |= 0x1ull;
+	oct_ep_write64(reg_val, otx_ep->hw_addr +
+		       CNXK_EP_R_OUT_ENABLE(q_no));
+
+	otx_ep_info("OQ[%d] enable done", q_no);
+
+	return 0;
+}
+
+static int
+cnxk_ep_vf_enable_io_queues(struct otx_ep_device *otx_ep)
+{
+	uint32_t q_no = 0;
+	int ret;
+
+	for (q_no = 0; q_no < otx_ep->nb_tx_queues; q_no++) {
+		ret = cnxk_ep_vf_enable_iq(otx_ep, q_no);
+		if (ret)
+			return ret;
+	}
+
+	for (q_no = 0; q_no < otx_ep->nb_rx_queues; q_no++)
+		cnxk_ep_vf_enable_oq(otx_ep, q_no);
+
+	return 0;
+}
+
+static void
+cnxk_ep_vf_disable_iq(struct otx_ep_device *otx_ep, uint32_t q_no)
+{
+	uint64_t reg_val = 0ull;
+
+	/* Reset the doorbell register for this Input Queue. */
+	reg_val = oct_ep_read64(otx_ep->hw_addr + CNXK_EP_R_IN_ENABLE(q_no));
+	reg_val &= ~0x1ull;
+
+	oct_ep_write64(reg_val, otx_ep->hw_addr + CNXK_EP_R_IN_ENABLE(q_no));
+}
+
+static void
+cnxk_ep_vf_disable_oq(struct otx_ep_device *otx_ep, uint32_t q_no)
+{
+	volatile uint64_t reg_val = 0ull;
+
+	reg_val = oct_ep_read64(otx_ep->hw_addr +
+				CNXK_EP_R_OUT_ENABLE(q_no));
+	reg_val &= ~0x1ull;
+
+	oct_ep_write64(reg_val, otx_ep->hw_addr +
+		       CNXK_EP_R_OUT_ENABLE(q_no));
+}
+
+static void
+cnxk_ep_vf_disable_io_queues(struct otx_ep_device *otx_ep)
+{
+	uint32_t q_no = 0;
+
+	for (q_no = 0; q_no < otx_ep->sriov_info.rings_per_vf; q_no++) {
+		cnxk_ep_vf_disable_iq(otx_ep, q_no);
+		cnxk_ep_vf_disable_oq(otx_ep, q_no);
+	}
+}
+
+static const struct otx_ep_config default_cnxk_ep_conf = {
+	/* IQ attributes */
+	.iq                        = {
+		.max_iqs           = OTX_EP_CFG_IO_QUEUES,
+		.instr_type        = OTX_EP_64BYTE_INSTR,
+		.pending_list_size = (OTX_EP_MAX_IQ_DESCRIPTORS *
+				      OTX_EP_CFG_IO_QUEUES),
+	},
+
+	/* OQ attributes */
+	.oq                        = {
+		.max_oqs           = OTX_EP_CFG_IO_QUEUES,
+		.info_ptr          = OTX_EP_OQ_INFOPTR_MODE,
+		.refill_threshold  = OTX_EP_OQ_REFIL_THRESHOLD,
+	},
+
+	.num_iqdef_descs           = OTX_EP_MAX_IQ_DESCRIPTORS,
+	.num_oqdef_descs           = OTX_EP_MAX_OQ_DESCRIPTORS,
+	.oqdef_buf_size            = OTX_EP_OQ_BUF_SIZE,
+};
+
+static const struct otx_ep_config*
+cnxk_ep_get_defconf(struct otx_ep_device *otx_ep_dev __rte_unused)
+{
+	const struct otx_ep_config *default_conf = NULL;
+
+	default_conf = &default_cnxk_ep_conf;
+
+	return default_conf;
+}
+
+int
+cnxk_ep_vf_setup_device(struct otx_ep_device *otx_ep)
+{
+	uint64_t reg_val = 0ull;
+
+	/* If application doesn't provide its conf, use driver default conf */
+	if (otx_ep->conf == NULL) {
+		otx_ep->conf = cnxk_ep_get_defconf(otx_ep);
+		if (otx_ep->conf == NULL) {
+			otx_ep_err("SDP VF default config not found");
+			return -ENOENT;
+		}
+		otx_ep_info("Default config is used");
+	}
+
+	/* Get IOQs (RPVF] count */
+	reg_val = oct_ep_read64(otx_ep->hw_addr + CNXK_EP_R_IN_CONTROL(0));
+
+	otx_ep->sriov_info.rings_per_vf =
+		((reg_val >> CNXK_EP_R_IN_CTL_RPVF_POS) &
+		 CNXK_EP_R_IN_CTL_RPVF_MASK);
+
+	otx_ep_info("SDP RPVF: %d", otx_ep->sriov_info.rings_per_vf);
+
+	otx_ep->fn_list.setup_iq_regs       = cnxk_ep_vf_setup_iq_regs;
+	otx_ep->fn_list.setup_oq_regs       = cnxk_ep_vf_setup_oq_regs;
+
+	otx_ep->fn_list.setup_device_regs   = cnxk_ep_vf_setup_device_regs;
+
+	otx_ep->fn_list.enable_io_queues    = cnxk_ep_vf_enable_io_queues;
+	otx_ep->fn_list.disable_io_queues   = cnxk_ep_vf_disable_io_queues;
+
+	otx_ep->fn_list.enable_iq           = cnxk_ep_vf_enable_iq;
+	otx_ep->fn_list.disable_iq          = cnxk_ep_vf_disable_iq;
+
+	otx_ep->fn_list.enable_oq           = cnxk_ep_vf_enable_oq;
+	otx_ep->fn_list.disable_oq          = cnxk_ep_vf_disable_oq;
+
+	return 0;
+}
diff --git a/drivers/net/octeon_ep/cnxk_ep_vf.h b/drivers/net/octeon_ep/cnxk_ep_vf.h
new file mode 100644
index 0000000000..4af20d5be9
--- /dev/null
+++ b/drivers/net/octeon_ep/cnxk_ep_vf.h
@@ -0,0 +1,161 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+#ifndef _CNXK_EP_VF_H_
+#define _CNXK_EP_VF_H_
+
+#include <rte_io.h>
+
+#define CNXK_CONFIG_XPANSION_BAR             0x38
+#define CNXK_CONFIG_PCIE_CAP                 0x70
+#define CNXK_CONFIG_PCIE_DEVCAP              0x74
+#define CNXK_CONFIG_PCIE_DEVCTL              0x78
+#define CNXK_CONFIG_PCIE_LINKCAP             0x7C
+#define CNXK_CONFIG_PCIE_LINKCTL             0x80
+#define CNXK_CONFIG_PCIE_SLOTCAP             0x84
+#define CNXK_CONFIG_PCIE_SLOTCTL             0x88
+#define CNXK_CONFIG_PCIE_FLTMSK              0x720
+
+#define CNXK_EP_RING_OFFSET                    (0x1ULL << 17)
+
+#define CNXK_EP_R_IN_CONTROL_START          0x10000
+#define CNXK_EP_R_IN_ENABLE_START           0x10010
+#define CNXK_EP_R_IN_INSTR_BADDR_START      0x10020
+#define CNXK_EP_R_IN_INSTR_RSIZE_START      0x10030
+#define CNXK_EP_R_IN_INSTR_DBELL_START      0x10040
+#define CNXK_EP_R_IN_CNTS_START             0x10050
+#define CNXK_EP_R_IN_INT_LEVELS_START       0x10060
+#define CNXK_EP_R_IN_PKT_CNT_START          0x10080
+#define CNXK_EP_R_IN_BYTE_CNT_START         0x10090
+
+#define CNXK_EP_R_IN_CONTROL(ring)             \
+	(CNXK_EP_R_IN_CONTROL_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_IN_ENABLE(ring)              \
+	(CNXK_EP_R_IN_ENABLE_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_IN_INSTR_BADDR(ring)          \
+	(CNXK_EP_R_IN_INSTR_BADDR_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_IN_INSTR_RSIZE(ring)          \
+	(CNXK_EP_R_IN_INSTR_RSIZE_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_IN_INSTR_DBELL(ring)          \
+	(CNXK_EP_R_IN_INSTR_DBELL_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_IN_CNTS(ring)                \
+	(CNXK_EP_R_IN_CNTS_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_IN_INT_LEVELS(ring)          \
+	(CNXK_EP_R_IN_INT_LEVELS_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_IN_PKT_CNT(ring)             \
+	(CNXK_EP_R_IN_PKT_CNT_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_IN_BYTE_CNT(ring)            \
+	(CNXK_EP_R_IN_BYTE_CNT_START +  ((ring) * CNXK_EP_RING_OFFSET))
+
+
+/** Rings per Virtual Function **/
+#define CNXK_EP_R_IN_CTL_RPVF_MASK	(0xF)
+#define	CNXK_EP_R_IN_CTL_RPVF_POS	(48)
+
+/* Number of instructions to be read in one MAC read request.
+ * setting to Max value(4)
+ */
+#define CNXK_EP_R_IN_CTL_IDLE		(0x1ULL << 28)
+#define CNXK_EP_R_IN_CTL_RDSIZE		(0x3ULL << 25)
+#define CNXK_EP_R_IN_CTL_IS_64B		(0x1ULL << 24)
+#define CNXK_EP_R_IN_CTL_D_NSR		(0x1ULL << 8)
+#define CNXK_EP_R_IN_CTL_D_ROR		(0x1ULL << 5)
+#define CNXK_EP_R_IN_CTL_NSR		(0x1ULL << 3)
+#define CNXK_EP_R_IN_CTL_ROR		(0x1ULL << 0)
+#define CNXK_EP_R_IN_CTL_ESR             (0x1ull << 1)
+
+#define CNXK_EP_R_IN_CTL_MASK                    \
+	(CNXK_EP_R_IN_CTL_RDSIZE                 \
+	| CNXK_EP_R_IN_CTL_IS_64B)
+
+#define CNXK_EP_R_OUT_CNTS_START           0x10100
+#define CNXK_EP_R_OUT_INT_LEVELS_START     0x10110
+#define CNXK_EP_R_OUT_SLIST_BADDR_START    0x10120
+#define CNXK_EP_R_OUT_SLIST_RSIZE_START    0x10130
+#define CNXK_EP_R_OUT_SLIST_DBELL_START    0x10140
+#define CNXK_EP_R_OUT_CONTROL_START        0x10150
+/* WMARK need to be set; New in CN10K */
+#define CNXK_EP_R_OUT_WMARK_START          0x10160
+#define CNXK_EP_R_OUT_ENABLE_START         0x10170
+#define CNXK_EP_R_OUT_PKT_CNT_START        0x10180
+#define CNXK_EP_R_OUT_BYTE_CNT_START       0x10190
+
+#define CNXK_EP_R_OUT_CNTS(ring)                \
+	(CNXK_EP_R_OUT_CNTS_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_OUT_INT_LEVELS(ring)          \
+	(CNXK_EP_R_OUT_INT_LEVELS_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_OUT_SLIST_BADDR(ring)          \
+	(CNXK_EP_R_OUT_SLIST_BADDR_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_OUT_SLIST_RSIZE(ring)          \
+	(CNXK_EP_R_OUT_SLIST_RSIZE_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_OUT_SLIST_DBELL(ring)          \
+	(CNXK_EP_R_OUT_SLIST_DBELL_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_OUT_CONTROL(ring)              \
+	(CNXK_EP_R_OUT_CONTROL_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_OUT_ENABLE(ring)               \
+	(CNXK_EP_R_OUT_ENABLE_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_OUT_WMARK(ring)                \
+	(CNXK_EP_R_OUT_WMARK_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_OUT_PKT_CNT(ring)              \
+	(CNXK_EP_R_OUT_PKT_CNT_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+#define CNXK_EP_R_OUT_BYTE_CNT(ring)             \
+	(CNXK_EP_R_OUT_BYTE_CNT_START + ((ring) * CNXK_EP_RING_OFFSET))
+
+/*------------------ R_OUT Masks ----------------*/
+#define CNXK_EP_R_OUT_INT_LEVELS_BMODE       (1ULL << 63)
+#define CNXK_EP_R_OUT_INT_LEVELS_TIMET       (32)
+
+#define CNXK_EP_R_OUT_CTL_IDLE               (1ULL << 40)
+#define CNXK_EP_R_OUT_CTL_ES_I         (1ull << 34)
+#define CNXK_EP_R_OUT_CTL_NSR_I              (1ULL << 33)
+#define CNXK_EP_R_OUT_CTL_ROR_I              (1ULL << 32)
+#define CNXK_EP_R_OUT_CTL_ES_D         (1ull << 30)
+#define CNXK_EP_R_OUT_CTL_NSR_D              (1ULL << 29)
+#define CNXK_EP_R_OUT_CTL_ROR_D              (1ULL << 28)
+#define CNXK_EP_R_OUT_CTL_ES_P         (1ull << 26)
+#define CNXK_EP_R_OUT_CTL_NSR_P              (1ULL << 25)
+#define CNXK_EP_R_OUT_CTL_ROR_P              (1ULL << 24)
+#define CNXK_EP_R_OUT_CTL_IMODE              (1ULL << 23)
+
+#define PCI_DEVID_CNXK_EP_NET_VF		0xB903
+
+int
+cnxk_ep_vf_setup_device(struct otx_ep_device *sdpvf);
+
+struct cnxk_ep_instr_64B {
+	/* Pointer where the input data is available. */
+	uint64_t dptr;
+
+	/* OTX_EP Instruction Header. */
+	union otx_ep_instr_ih ih;
+
+	/** Pointer where the response for a RAW mode packet
+	 * will be written by OCTEON TX.
+	 */
+	uint64_t rptr;
+
+	/* Input Request Header. */
+	union otx_ep_instr_irh irh;
+
+	/* Additional headers available in a 64-byte instruction. */
+	uint64_t exhdr[4];
+};
+
+#endif /*_CNXK_EP_VF_H_ */
diff --git a/drivers/net/octeon_ep/meson.build b/drivers/net/octeon_ep/meson.build
index 957e8fc48a..a267b60290 100644
--- a/drivers/net/octeon_ep/meson.build
+++ b/drivers/net/octeon_ep/meson.build
@@ -7,4 +7,5 @@ sources = files(
         'otx_ep_rxtx.c',
         'otx_ep_vf.c',
         'otx2_ep_vf.c',
+        'cnxk_ep_vf.c',
 )
diff --git a/drivers/net/octeon_ep/otx2_ep_vf.c b/drivers/net/octeon_ep/otx2_ep_vf.c
index 85e14a998f..09fb621643 100644
--- a/drivers/net/octeon_ep/otx2_ep_vf.c
+++ b/drivers/net/octeon_ep/otx2_ep_vf.c
@@ -15,13 +15,13 @@ otx2_vf_setup_global_iq_reg(struct otx_ep_device *otx_ep, int q_no)
 	/* Select ES, RO, NS, RDSIZE,DPTR Format#0 for IQs
 	 * IS_64B is by default enabled.
 	 */
-	reg_val = otx2_read64(otx_ep->hw_addr + SDP_VF_R_IN_CONTROL(q_no));
+	reg_val = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_IN_CONTROL(q_no));
 
 	reg_val |= SDP_VF_R_IN_CTL_RDSIZE;
 	reg_val |= SDP_VF_R_IN_CTL_IS_64B;
 	reg_val |= SDP_VF_R_IN_CTL_ESR;
 
-	otx2_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_IN_CONTROL(q_no));
+	oct_ep_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_IN_CONTROL(q_no));
 }
 
 static void
@@ -29,7 +29,7 @@ otx2_vf_setup_global_oq_reg(struct otx_ep_device *otx_ep, int q_no)
 {
 	volatile uint64_t reg_val = 0ull;
 
-	reg_val = otx2_read64(otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(q_no));
+	reg_val = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(q_no));
 
 	reg_val &= ~(SDP_VF_R_OUT_CTL_IMODE);
 	reg_val &= ~(SDP_VF_R_OUT_CTL_ROR_P);
@@ -44,7 +44,7 @@ otx2_vf_setup_global_oq_reg(struct otx_ep_device *otx_ep, int q_no)
 	/* INFO/DATA ptr swap is required  */
 	reg_val |= (SDP_VF_R_OUT_CTL_ES_P);
 
-	otx2_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(q_no));
+	oct_ep_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(q_no));
 }
 
 static void
@@ -78,22 +78,22 @@ otx2_vf_setup_iq_regs(struct otx_ep_device *otx_ep, uint32_t iq_no)
 	struct otx_ep_instr_queue *iq = otx_ep->instr_queue[iq_no];
 	volatile uint64_t reg_val = 0ull;
 
-	reg_val = otx2_read64(otx_ep->hw_addr + SDP_VF_R_IN_CONTROL(iq_no));
+	reg_val = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_IN_CONTROL(iq_no));
 
 	/* Wait till IDLE to set to 1, not supposed to configure BADDR
 	 * as long as IDLE is 0
 	 */
 	if (!(reg_val & SDP_VF_R_IN_CTL_IDLE)) {
 		do {
-			reg_val = otx2_read64(otx_ep->hw_addr +
+			reg_val = oct_ep_read64(otx_ep->hw_addr +
 					      SDP_VF_R_IN_CONTROL(iq_no));
 		} while (!(reg_val & SDP_VF_R_IN_CTL_IDLE));
 	}
 
 	/* Write the start of the input queue's ring and its size  */
-	otx2_write64(iq->base_addr_dma, otx_ep->hw_addr +
+	oct_ep_write64(iq->base_addr_dma, otx_ep->hw_addr +
 		     SDP_VF_R_IN_INSTR_BADDR(iq_no));
-	otx2_write64(iq->nb_desc, otx_ep->hw_addr +
+	oct_ep_write64(iq->nb_desc, otx_ep->hw_addr +
 		     SDP_VF_R_IN_INSTR_RSIZE(iq_no));
 
 	/* Remember the doorbell & instruction count register addr
@@ -115,7 +115,7 @@ otx2_vf_setup_iq_regs(struct otx_ep_device *otx_ep, uint32_t iq_no)
 	/* IN INTR_THRESHOLD is set to max(FFFFFFFF) which disable the IN INTR
 	 * to raise
 	 */
-	otx2_write64(OTX_EP_CLEAR_SDP_IN_INT_LVLS,
+	oct_ep_write64(OTX_EP_CLEAR_SDP_IN_INT_LVLS,
 		     otx_ep->hw_addr + SDP_VF_R_IN_INT_LEVELS(iq_no));
 }
 
@@ -129,19 +129,19 @@ otx2_vf_setup_oq_regs(struct otx_ep_device *otx_ep, uint32_t oq_no)
 	/* Wait on IDLE to set to 1, supposed to configure BADDR
 	 * as log as IDLE is 0
 	 */
-	reg_val = otx2_read64(otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(oq_no));
+	reg_val = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(oq_no));
 
 	while (!(reg_val & SDP_VF_R_OUT_CTL_IDLE)) {
-		reg_val = otx2_read64(otx_ep->hw_addr +
+		reg_val = oct_ep_read64(otx_ep->hw_addr +
 				      SDP_VF_R_OUT_CONTROL(oq_no));
 	}
 
-	otx2_write64(droq->desc_ring_dma, otx_ep->hw_addr +
+	oct_ep_write64(droq->desc_ring_dma, otx_ep->hw_addr +
 		     SDP_VF_R_OUT_SLIST_BADDR(oq_no));
-	otx2_write64(droq->nb_desc, otx_ep->hw_addr +
+	oct_ep_write64(droq->nb_desc, otx_ep->hw_addr +
 		     SDP_VF_R_OUT_SLIST_RSIZE(oq_no));
 
-	oq_ctl = otx2_read64(otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(oq_no));
+	oq_ctl = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(oq_no));
 
 	/* Clear the ISIZE and BSIZE (22-0) */
 	oq_ctl &= ~(OTX_EP_CLEAR_ISIZE_BSIZE);
@@ -149,7 +149,7 @@ otx2_vf_setup_oq_regs(struct otx_ep_device *otx_ep, uint32_t oq_no)
 	/* Populate the BSIZE (15-0) */
 	oq_ctl |= (droq->buffer_size & OTX_EP_DROQ_BUFSZ_MASK);
 
-	otx2_write64(oq_ctl, otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(oq_no));
+	oct_ep_write64(oq_ctl, otx_ep->hw_addr + SDP_VF_R_OUT_CONTROL(oq_no));
 
 	/* Mapped address of the pkt_sent and pkts_credit regs */
 	droq->pkts_sent_reg = (uint8_t *)otx_ep->hw_addr +
@@ -198,10 +198,10 @@ otx2_vf_enable_iq(struct otx_ep_device *otx_ep, uint32_t q_no)
 	/* Resetting doorbells during IQ enabling also to handle abrupt
 	 * guest reboot. IQ reset does not clear the doorbells.
 	 */
-	otx2_write64(0xFFFFFFFF, otx_ep->hw_addr +
+	oct_ep_write64(0xFFFFFFFF, otx_ep->hw_addr +
 		     SDP_VF_R_IN_INSTR_DBELL(q_no));
 
-	while (((otx2_read64(otx_ep->hw_addr +
+	while (((oct_ep_read64(otx_ep->hw_addr +
 		 SDP_VF_R_IN_INSTR_DBELL(q_no))) != 0ull) && loop--) {
 		rte_delay_ms(1);
 	}
@@ -211,10 +211,10 @@ otx2_vf_enable_iq(struct otx_ep_device *otx_ep, uint32_t q_no)
 		return -EIO;
 	}
 
-	reg_val = otx2_read64(otx_ep->hw_addr + SDP_VF_R_IN_ENABLE(q_no));
+	reg_val = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_IN_ENABLE(q_no));
 	reg_val |= 0x1ull;
 
-	otx2_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_IN_ENABLE(q_no));
+	oct_ep_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_IN_ENABLE(q_no));
 
 	otx_ep_info("IQ[%d] enable done", q_no);
 
@@ -226,9 +226,9 @@ otx2_vf_enable_oq(struct otx_ep_device *otx_ep, uint32_t q_no)
 {
 	uint64_t reg_val = 0ull;
 
-	reg_val = otx2_read64(otx_ep->hw_addr + SDP_VF_R_OUT_ENABLE(q_no));
+	reg_val = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_OUT_ENABLE(q_no));
 	reg_val |= 0x1ull;
-	otx2_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_OUT_ENABLE(q_no));
+	oct_ep_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_OUT_ENABLE(q_no));
 
 	otx_ep_info("OQ[%d] enable done", q_no);
 
@@ -259,10 +259,10 @@ otx2_vf_disable_iq(struct otx_ep_device *otx_ep, uint32_t q_no)
 	uint64_t reg_val = 0ull;
 
 	/* Reset the doorbell register for this Input Queue. */
-	reg_val = otx2_read64(otx_ep->hw_addr + SDP_VF_R_IN_ENABLE(q_no));
+	reg_val = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_IN_ENABLE(q_no));
 	reg_val &= ~0x1ull;
 
-	otx2_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_IN_ENABLE(q_no));
+	oct_ep_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_IN_ENABLE(q_no));
 }
 
 static void
@@ -270,10 +270,10 @@ otx2_vf_disable_oq(struct otx_ep_device *otx_ep, uint32_t q_no)
 {
 	volatile uint64_t reg_val = 0ull;
 
-	reg_val = otx2_read64(otx_ep->hw_addr + SDP_VF_R_OUT_ENABLE(q_no));
+	reg_val = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_OUT_ENABLE(q_no));
 	reg_val &= ~0x1ull;
 
-	otx2_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_OUT_ENABLE(q_no));
+	oct_ep_write64(reg_val, otx_ep->hw_addr + SDP_VF_R_OUT_ENABLE(q_no));
 }
 
 static void
@@ -334,7 +334,7 @@ otx2_ep_vf_setup_device(struct otx_ep_device *otx_ep)
 	}
 
 	/* Get IOQs (RPVF] count */
-	reg_val = otx2_read64(otx_ep->hw_addr + SDP_VF_R_IN_CONTROL(0));
+	reg_val = oct_ep_read64(otx_ep->hw_addr + SDP_VF_R_IN_CONTROL(0));
 
 	otx_ep->sriov_info.rings_per_vf = ((reg_val >> SDP_VF_R_IN_CTL_RPVF_POS)
 					  & SDP_VF_R_IN_CTL_RPVF_MASK);
diff --git a/drivers/net/octeon_ep/otx_ep_common.h b/drivers/net/octeon_ep/otx_ep_common.h
index 8a59a1a194..7eb50af75a 100644
--- a/drivers/net/octeon_ep/otx_ep_common.h
+++ b/drivers/net/octeon_ep/otx_ep_common.h
@@ -57,6 +57,10 @@
 		"%s():%u " fmt "\n",				\
 		__func__, __LINE__, ##args)
 
+/* IO Access */
+#define oct_ep_read64(addr) rte_read64_relaxed((void *)(addr))
+#define oct_ep_write64(val, addr) rte_write64_relaxed((val), (void *)(addr))
+
 /* Input Request Header format */
 union otx_ep_instr_irh {
 	uint64_t u64;
diff --git a/drivers/net/octeon_ep/otx_ep_ethdev.c b/drivers/net/octeon_ep/otx_ep_ethdev.c
index 77def6daa1..c8f4abe4ca 100644
--- a/drivers/net/octeon_ep/otx_ep_ethdev.c
+++ b/drivers/net/octeon_ep/otx_ep_ethdev.c
@@ -7,6 +7,7 @@
 #include "otx_ep_common.h"
 #include "otx_ep_vf.h"
 #include "otx2_ep_vf.h"
+#include "cnxk_ep_vf.h"
 #include "otx_ep_rxtx.h"
 
 #define OTX_EP_DEV(_eth_dev) \
@@ -108,6 +109,11 @@ otx_ep_chip_specific_setup(struct otx_ep_device *otx_epvf)
 		ret = otx2_ep_vf_setup_device(otx_epvf);
 		otx_epvf->fn_list.disable_io_queues(otx_epvf);
 		break;
+	case PCI_DEVID_CNXK_EP_NET_VF:
+		otx_epvf->chip_id = dev_id;
+		ret = cnxk_ep_vf_setup_device(otx_epvf);
+		otx_epvf->fn_list.disable_io_queues(otx_epvf);
+		break;
 	default:
 		otx_ep_err("Unsupported device\n");
 		ret = -EINVAL;
@@ -140,6 +146,8 @@ otx_epdev_init(struct otx_ep_device *otx_epvf)
 	else if (otx_epvf->chip_id == PCI_DEVID_CN9K_EP_NET_VF ||
 		 otx_epvf->chip_id == PCI_DEVID_CN98XX_EP_NET_VF)
 		otx_epvf->eth_dev->tx_pkt_burst = &otx2_ep_xmit_pkts;
+	else if (otx_epvf->chip_id == PCI_DEVID_CNXK_EP_NET_VF)
+		otx_epvf->eth_dev->tx_pkt_burst = &otx2_ep_xmit_pkts;
 	ethdev_queues = (uint32_t)(otx_epvf->sriov_info.rings_per_vf);
 	otx_epvf->max_rx_queues = ethdev_queues;
 	otx_epvf->max_tx_queues = ethdev_queues;
@@ -521,6 +529,7 @@ static const struct rte_pci_id pci_id_otx_ep_map[] = {
 	{ RTE_PCI_DEVICE(PCI_VENDOR_ID_CAVIUM, PCI_DEVID_OCTEONTX_EP_VF) },
 	{ RTE_PCI_DEVICE(PCI_VENDOR_ID_CAVIUM, PCI_DEVID_CN9K_EP_NET_VF) },
 	{ RTE_PCI_DEVICE(PCI_VENDOR_ID_CAVIUM, PCI_DEVID_CN98XX_EP_NET_VF) },
+	{ RTE_PCI_DEVICE(PCI_VENDOR_ID_CAVIUM, PCI_DEVID_CNXK_EP_NET_VF) },
 	{ .vendor_id = 0, /* sentinel */ }
 };
 
-- 
2.36.1


^ permalink raw reply	[flat|nested] 5+ messages in thread

end of thread, other threads:[~2022-08-30 10:25 UTC | newest]

Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-08-16 10:34 [PATCH] net/octeon_ep: add octeon CN10K support Sathesh Edara
2022-08-27 13:07 ` Jerin Jacob
2022-08-29 15:45 ` [PATCH v2] net/octeon_ep: support CN10K SoC Sathesh Edara
2022-08-30  7:39   ` Sathesh B Edara
2022-08-30 10:24   ` Jerin Jacob

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).