From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 74942A034F; Mon, 11 Oct 2021 16:51:47 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id DB8924111D; Mon, 11 Oct 2021 16:49:58 +0200 (CEST) Received: from shelob.oktetlabs.ru (shelob.oktetlabs.ru [91.220.146.113]) by mails.dpdk.org (Postfix) with ESMTP id 533894117C for ; Mon, 11 Oct 2021 16:49:55 +0200 (CEST) Received: by shelob.oktetlabs.ru (Postfix, from userid 122) id 2074A7F702; Mon, 11 Oct 2021 17:49:55 +0300 (MSK) X-Spam-Checker-Version: SpamAssassin 3.4.2 (2018-09-13) on shelob.oktetlabs.ru X-Spam-Level: X-Spam-Status: No, score=0.8 required=5.0 tests=ALL_TRUSTED, DKIM_ADSP_DISCARD, URIBL_BLOCKED autolearn=no autolearn_force=no version=3.4.2 Received: from aros.oktetlabs.ru (aros.oktetlabs.ru [192.168.38.17]) by shelob.oktetlabs.ru (Postfix) with ESMTP id 4D7987F70D; Mon, 11 Oct 2021 17:49:09 +0300 (MSK) DKIM-Filter: OpenDKIM Filter v2.11.0 shelob.oktetlabs.ru 4D7987F70D Authentication-Results: shelob.oktetlabs.ru/4D7987F70D; dkim=none; dkim-atps=neutral From: Andrew Rybchenko To: dev@dpdk.org Cc: Igor Romanov , Andy Moreton , Ivan Malov Date: Mon, 11 Oct 2021 17:48:41 +0300 Message-Id: <20211011144857.446802-23-andrew.rybchenko@oktetlabs.ru> X-Mailer: git-send-email 2.30.2 In-Reply-To: <20211011144857.446802-1-andrew.rybchenko@oktetlabs.ru> References: <20210827065717.1838258-1-andrew.rybchenko@oktetlabs.ru> <20211011144857.446802-1-andrew.rybchenko@oktetlabs.ru> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Subject: [dpdk-dev] [PATCH v2 22/38] net/sfc: implement representor Tx routine X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" From: Igor Romanov Forward traffic that is transmitted from a port representor to the corresponding virtual function using the dedicated TxQ. Signed-off-by: Igor Romanov Signed-off-by: Andrew Rybchenko Reviewed-by: Andy Moreton Reviewed-by: Ivan Malov --- drivers/net/sfc/sfc_repr.c | 45 ++++++++++++++++ drivers/net/sfc/sfc_repr_proxy.c | 88 +++++++++++++++++++++++++++++++- drivers/net/sfc/sfc_repr_proxy.h | 8 +++ 3 files changed, 140 insertions(+), 1 deletion(-) diff --git a/drivers/net/sfc/sfc_repr.c b/drivers/net/sfc/sfc_repr.c index 922f4da432..3e74313b12 100644 --- a/drivers/net/sfc/sfc_repr.c +++ b/drivers/net/sfc/sfc_repr.c @@ -160,6 +160,49 @@ sfc_repr_tx_queue_stop(void *queue) rte_ring_reset(txq->ring); } +static uint16_t +sfc_repr_tx_burst(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) +{ + struct sfc_repr_txq *txq = tx_queue; + unsigned int n_tx; + void **objs; + uint16_t i; + + /* + * mbuf is likely cache-hot. Set flag and egress m-port here instead of + * doing that in representors proxy. Also, it should help to avoid + * cache bounce. Moreover, potentially, it allows to use one + * multi-producer single-consumer ring for all representors. + * + * The only potential problem is doing so many times if enqueue + * fails and sender retries. + */ + for (i = 0; i < nb_pkts; ++i) { + struct rte_mbuf *m = tx_pkts[i]; + + m->ol_flags |= sfc_dp_mport_override; + *RTE_MBUF_DYNFIELD(m, sfc_dp_mport_offset, + efx_mport_id_t *) = txq->egress_mport; + } + + objs = (void *)&tx_pkts[0]; + n_tx = rte_ring_sp_enqueue_burst(txq->ring, objs, nb_pkts, NULL); + + /* + * Remove m-port override flag from packets that were not enqueued + * Setting the flag only for enqueued packets after the burst is + * not possible since the ownership of enqueued packets is + * transferred to representor proxy. + */ + for (i = n_tx; i < nb_pkts; ++i) { + struct rte_mbuf *m = tx_pkts[i]; + + m->ol_flags &= ~sfc_dp_mport_override; + } + + return n_tx; +} + static int sfc_repr_start(struct rte_eth_dev *dev) { @@ -760,6 +803,7 @@ sfc_repr_dev_close(struct rte_eth_dev *dev) (void)sfc_repr_proxy_del_port(srs->pf_port_id, srs->repr_id); + dev->tx_pkt_burst = NULL; dev->dev_ops = NULL; sfc_repr_unlock(sr); @@ -880,6 +924,7 @@ sfc_repr_eth_dev_init(struct rte_eth_dev *dev, void *init_params) goto fail_mac_addrs; } + dev->tx_pkt_burst = sfc_repr_tx_burst; dev->dev_ops = &sfc_repr_dev_ops; sr->state = SFC_ETHDEV_INITIALIZED; diff --git a/drivers/net/sfc/sfc_repr_proxy.c b/drivers/net/sfc/sfc_repr_proxy.c index ea03d5afdd..d8934bab65 100644 --- a/drivers/net/sfc/sfc_repr_proxy.c +++ b/drivers/net/sfc/sfc_repr_proxy.c @@ -25,6 +25,12 @@ */ #define SFC_REPR_PROXY_MBOX_POLL_TIMEOUT_MS 1000 +/** + * Amount of time to wait for the representor proxy routine (which is + * running on a service core) to terminate after service core is stopped. + */ +#define SFC_REPR_PROXY_ROUTINE_TERMINATE_TIMEOUT_MS 10000 + static struct sfc_repr_proxy * sfc_repr_proxy_by_adapter(struct sfc_adapter *sa) { @@ -148,16 +154,71 @@ sfc_repr_proxy_mbox_handle(struct sfc_repr_proxy *rp) __atomic_store_n(&mbox->ack, true, __ATOMIC_RELEASE); } +static void +sfc_repr_proxy_handle_tx(struct sfc_repr_proxy_dp_txq *rp_txq, + struct sfc_repr_proxy_txq *repr_txq) +{ + /* + * With multiple representor proxy queues configured it is + * possible that not all of the corresponding representor + * queues were created. Skip the queues that do not exist. + */ + if (repr_txq->ring == NULL) + return; + + if (rp_txq->available < RTE_DIM(rp_txq->tx_pkts)) { + rp_txq->available += + rte_ring_sc_dequeue_burst(repr_txq->ring, + (void **)(&rp_txq->tx_pkts[rp_txq->available]), + RTE_DIM(rp_txq->tx_pkts) - rp_txq->available, + NULL); + + if (rp_txq->available == rp_txq->transmitted) + return; + } + + rp_txq->transmitted += rp_txq->pkt_burst(rp_txq->dp, + &rp_txq->tx_pkts[rp_txq->transmitted], + rp_txq->available - rp_txq->transmitted); + + if (rp_txq->available == rp_txq->transmitted) { + rp_txq->available = 0; + rp_txq->transmitted = 0; + } +} + static int32_t sfc_repr_proxy_routine(void *arg) { + struct sfc_repr_proxy_port *port; struct sfc_repr_proxy *rp = arg; + unsigned int i; sfc_repr_proxy_mbox_handle(rp); + TAILQ_FOREACH(port, &rp->ports, entries) { + if (!port->started) + continue; + + for (i = 0; i < rp->nb_txq; i++) + sfc_repr_proxy_handle_tx(&rp->dp_txq[i], &port->txq[i]); + } + return 0; } +static struct sfc_txq_info * +sfc_repr_proxy_txq_info_get(struct sfc_adapter *sa, unsigned int repr_queue_id) +{ + struct sfc_adapter_shared *sas = sfc_sa2shared(sa); + struct sfc_repr_proxy_dp_txq *dp_txq; + + SFC_ASSERT(repr_queue_id < sfc_repr_nb_txq(sas)); + dp_txq = &sa->repr_proxy.dp_txq[repr_queue_id]; + + return &sas->txq_info[dp_txq->sw_index]; +} + static int sfc_repr_proxy_txq_attach(struct sfc_adapter *sa) { @@ -289,11 +350,20 @@ sfc_repr_proxy_txq_fini(struct sfc_adapter *sa) static int sfc_repr_proxy_txq_start(struct sfc_adapter *sa) { + struct sfc_adapter_shared * const sas = sfc_sa2shared(sa); struct sfc_repr_proxy *rp = &sa->repr_proxy; + unsigned int i; sfc_log_init(sa, "entry"); - RTE_SET_USED(rp); + for (i = 0; i < sfc_repr_nb_txq(sas); i++) { + struct sfc_repr_proxy_dp_txq *txq = &rp->dp_txq[i]; + + txq->dp = sfc_repr_proxy_txq_info_get(sa, i)->dp; + txq->pkt_burst = sa->eth_dev->tx_pkt_burst; + txq->available = 0; + txq->transmitted = 0; + } sfc_log_init(sa, "done"); @@ -922,6 +992,8 @@ sfc_repr_proxy_start(struct sfc_adapter *sa) if (rc != 0) goto fail_txq_start; + rp->nb_txq = sfc_repr_nb_txq(sas); + /* Service core may be in "stopped" state, start it */ rc = rte_service_lcore_start(rp->service_core_id); if (rc != 0 && rc != -EALREADY) { @@ -1007,6 +1079,9 @@ sfc_repr_proxy_stop(struct sfc_adapter *sa) struct sfc_adapter_shared * const sas = sfc_sa2shared(sa); struct sfc_repr_proxy *rp = &sa->repr_proxy; struct sfc_repr_proxy_port *port; + const unsigned int wait_ms_total = + SFC_REPR_PROXY_ROUTINE_TERMINATE_TIMEOUT_MS; + unsigned int i; int rc; sfc_log_init(sa, "entry"); @@ -1050,6 +1125,17 @@ sfc_repr_proxy_stop(struct sfc_adapter *sa) /* Service lcore may be shared and we never stop it */ + /* + * Wait for the representor proxy routine to finish the last iteration. + * Give up on timeout. + */ + for (i = 0; i < wait_ms_total; i++) { + if (rte_service_may_be_active(rp->service_id) == 0) + break; + + rte_delay_ms(1); + } + sfc_repr_proxy_rxq_stop(sa); sfc_repr_proxy_txq_stop(sa); diff --git a/drivers/net/sfc/sfc_repr_proxy.h b/drivers/net/sfc/sfc_repr_proxy.h index c350713a55..d47e0a431a 100644 --- a/drivers/net/sfc/sfc_repr_proxy.h +++ b/drivers/net/sfc/sfc_repr_proxy.h @@ -79,6 +79,13 @@ struct sfc_repr_proxy_dp_rxq { }; struct sfc_repr_proxy_dp_txq { + eth_tx_burst_t pkt_burst; + struct sfc_dp_txq *dp; + + unsigned int available; + unsigned int transmitted; + struct rte_mbuf *tx_pkts[SFC_REPR_PROXY_TX_BURST]; + sfc_sw_index_t sw_index; }; @@ -110,6 +117,7 @@ struct sfc_repr_proxy { struct sfc_repr_proxy_filter mport_filter; struct sfc_repr_proxy_mbox mbox; + unsigned int nb_txq; }; struct sfc_adapter; -- 2.30.2