From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-wi0-f177.google.com (mail-wi0-f177.google.com [209.85.212.177]) by dpdk.org (Postfix) with ESMTP id B6365C4A8 for ; Tue, 30 Jun 2015 11:29:05 +0200 (CEST) Received: by widjy10 with SMTP id jy10so24869491wid.1 for ; Tue, 30 Jun 2015 02:29:05 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=d4Jf6V2InQ5KlEa/bCFP91UXQ4BhA14moqrh0WdvRuQ=; b=mHG/jrCcXdnQCyfTXaQWhrmTfrL21UDeHYCyoFj8X0LExTRO02KDXfEsbTp9spdIJR x4La1CoOdDuKMmACdvF0mhIKFL12cVV+YQsG+TZFG6vjfAWD6W8K0aT88Z97m2RBUr7B ZMtdy1RdQgKj1AG36IRTfsEp71rkdvnvb3A8itun/p+9m9DHaE/CSV2ol4AnM8rxkRqS yuVnHYdZAt25UI0R/VDjb8qup/u6ABtFYjW+1Jy5wjF6VK4VMqkfTbdUR24y4x7VHaP5 i9YIZr64FmayC6oCOsOepBKSYG/FB5GoeSztVoKj6GhnCFnqqys8qfI+N/cvL7e6HEfC 0D1w== X-Gm-Message-State: ALoCoQksTxHctCWR7eNQ0G9skXFYdKtx5eccnPlLOX97OXvKlbObFrLZkpYI3haq6Uz7VED2sa2E X-Received: by 10.194.22.105 with SMTP id c9mr39506202wjf.120.1435656545524; Tue, 30 Jun 2015 02:29:05 -0700 (PDT) Received: from 6wind.com (6wind.net2.nerim.net. [213.41.151.210]) by mx.google.com with ESMTPSA id c11sm16132903wib.1.2015.06.30.02.29.04 (version=TLSv1.2 cipher=RC4-SHA bits=128/128); Tue, 30 Jun 2015 02:29:05 -0700 (PDT) From: Adrien Mazarguil To: dev@dpdk.org Date: Tue, 30 Jun 2015 11:28:02 +0200 Message-Id: <1435656489-27986-17-git-send-email-adrien.mazarguil@6wind.com> X-Mailer: git-send-email 2.1.0 In-Reply-To: <1435656489-27986-1-git-send-email-adrien.mazarguil@6wind.com> References: <1433546120-2254-1-git-send-email-adrien.mazarguil@6wind.com> <1435656489-27986-1-git-send-email-adrien.mazarguil@6wind.com> Cc: Alex Rosenbaum Subject: [dpdk-dev] [PATCH v2 16/23] mlx4: move scattered TX processing to helper function X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: patches and discussions about DPDK List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 30 Jun 2015 09:29:06 -0000 This commit makes scattered TX support entirely optional by moving it to a separate function that is only available when MLX4_PMD_SGE_WR_N > 1. Improves performance when scattered support is not needed. Signed-off-by: Alex Rosenbaum Signed-off-by: Adrien Mazarguil --- drivers/net/mlx4/mlx4.c | 248 +++++++++++++++++++++++++++++++++--------------- 1 file changed, 170 insertions(+), 78 deletions(-) diff --git a/drivers/net/mlx4/mlx4.c b/drivers/net/mlx4/mlx4.c index 3dff64d..acf1290 100644 --- a/drivers/net/mlx4/mlx4.c +++ b/drivers/net/mlx4/mlx4.c @@ -1025,6 +1025,8 @@ txq_mp2mr(struct txq *txq, struct rte_mempool *mp) return txq->mp2mr[i].lkey; } +#if MLX4_PMD_SGE_WR_N > 1 + /** * Copy scattered mbuf contents to a single linear buffer. * @@ -1058,6 +1060,146 @@ linearize_mbuf(linear_t *linear, struct rte_mbuf *buf) } /** + * Handle scattered buffers for mlx4_tx_burst(). + * + * @param txq + * TX queue structure. + * @param segs + * Number of segments in buf. + * @param elt + * TX queue element to fill. + * @param[in] buf + * Buffer to process. + * @param elts_head + * Index of the linear buffer to use if necessary (normally txq->elts_head). + * + * @return + * Processed packet size in bytes or (unsigned int)-1 in case of failure. + */ +static unsigned int +tx_burst_sg(struct txq *txq, unsigned int segs, struct txq_elt *elt, + struct rte_mbuf *buf, unsigned int elts_head) +{ + struct ibv_send_wr *wr = &elt->wr; + unsigned int sent_size = 0; + unsigned int j; + int linearize = 0; + + /* When there are too many segments, extra segments are + * linearized in the last SGE. */ + if (unlikely(segs > elemof(elt->sges))) { + segs = (elemof(elt->sges) - 1); + linearize = 1; + } + /* Set WR fields. */ + assert((rte_pktmbuf_mtod(buf, uintptr_t) - + (uintptr_t)buf) <= 0xffff); + WR_ID(wr->wr_id).offset = + (rte_pktmbuf_mtod(buf, uintptr_t) - + (uintptr_t)buf); + wr->num_sge = segs; + /* Register segments as SGEs. */ + for (j = 0; (j != segs); ++j) { + struct ibv_sge *sge = &elt->sges[j]; + uint32_t lkey; + + /* Retrieve Memory Region key for this memory pool. */ + lkey = txq_mp2mr(txq, buf->pool); + if (unlikely(lkey == (uint32_t)-1)) { + /* MR does not exist. */ + DEBUG("%p: unable to get MP <-> MR association", + (void *)txq); + /* Clean up TX element. */ + WR_ID(elt->wr.wr_id).offset = 0; +#ifndef NDEBUG + /* For assert(). */ + while (j) { + --j; + --sge; + sge->addr = 0; + sge->length = 0; + sge->lkey = 0; + } + wr->num_sge = 0; +#endif + goto stop; + } + /* Sanity checks, only relevant with debugging enabled. */ + assert(sge->addr == 0); + assert(sge->length == 0); + assert(sge->lkey == 0); + /* Update SGE. */ + sge->addr = rte_pktmbuf_mtod(buf, uintptr_t); + if (txq->priv->vf) + rte_prefetch0((volatile void *) + (uintptr_t)sge->addr); + sge->length = DATA_LEN(buf); + sge->lkey = lkey; + sent_size += sge->length; + buf = NEXT(buf); + } + /* If buf is not NULL here and is not going to be linearized, + * nb_segs is not valid. */ + assert(j == segs); + assert((buf == NULL) || (linearize)); + /* Linearize extra segments. */ + if (linearize) { + struct ibv_sge *sge = &elt->sges[segs]; + linear_t *linear = &(*txq->elts_linear)[elts_head]; + unsigned int size = linearize_mbuf(linear, buf); + + assert(segs == (elemof(elt->sges) - 1)); + if (size == 0) { + /* Invalid packet. */ + DEBUG("%p: packet too large to be linearized.", + (void *)txq); + /* Clean up TX element. */ + WR_ID(elt->wr.wr_id).offset = 0; +#ifndef NDEBUG + /* For assert(). */ + while (j) { + --j; + --sge; + sge->addr = 0; + sge->length = 0; + sge->lkey = 0; + } + wr->num_sge = 0; +#endif + goto stop; + } + /* If MLX4_PMD_SGE_WR_N is 1, free mbuf immediately + * and clear offset from WR ID. */ + if (elemof(elt->sges) == 1) { + do { + struct rte_mbuf *next = NEXT(buf); + + rte_pktmbuf_free_seg(buf); + buf = next; + } while (buf != NULL); + WR_ID(wr->wr_id).offset = 0; + } + /* Set WR fields and fill SGE with linear buffer. */ + ++wr->num_sge; + /* Sanity checks, only relevant with debugging + * enabled. */ + assert(sge->addr == 0); + assert(sge->length == 0); + assert(sge->lkey == 0); + /* Update SGE. */ + sge->addr = (uintptr_t)&(*linear)[0]; + sge->length = size; + sge->lkey = txq->mr_linear->lkey; + sent_size += size; + } + return sent_size; +stop: + return -1; +} + +#endif /* MLX4_PMD_SGE_WR_N > 1 */ + +/** * DPDK callback for TX. * * @param dpdk_txq @@ -1106,8 +1248,9 @@ mlx4_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n) #ifdef MLX4_PMD_SOFT_COUNTERS unsigned int sent_size = 0; #endif +#ifndef NDEBUG unsigned int j; - int linearize = 0; +#endif uint32_t send_flags = 0; /* Clean up old buffer. */ @@ -1143,24 +1286,19 @@ mlx4_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n) assert(wr->sg_list == &elt->sges[0]); assert(wr->num_sge == 0); assert(wr->opcode == IBV_WR_SEND); - /* When there are too many segments, extra segments are - * linearized in the last SGE. */ - if (unlikely(segs > elemof(elt->sges))) { - segs = (elemof(elt->sges) - 1); - linearize = 1; - } - /* Set WR fields. */ - assert((rte_pktmbuf_mtod(buf, uintptr_t) - - (uintptr_t)buf) <= 0xffff); - WR_ID(wr->wr_id).offset = - (rte_pktmbuf_mtod(buf, uintptr_t) - - (uintptr_t)buf); - wr->num_sge = segs; - /* Register segments as SGEs. */ - for (j = 0; (j != segs); ++j) { - struct ibv_sge *sge = &elt->sges[j]; + if (likely(segs == 1)) { + struct ibv_sge *sge = &elt->sges[0]; uint32_t lkey; + /* Set WR fields. */ + assert((rte_pktmbuf_mtod(buf, uintptr_t) - + (uintptr_t)buf) <= 0xffff); + WR_ID(wr->wr_id).offset = + (rte_pktmbuf_mtod(buf, uintptr_t) - + (uintptr_t)buf); + wr->num_sge = segs; + /* Register segment as SGE. */ + sge = &elt->sges[0]; /* Retrieve Memory Region key for this memory pool. */ lkey = txq_mp2mr(txq, buf->pool); if (unlikely(lkey == (uint32_t)-1)) { @@ -1171,13 +1309,9 @@ mlx4_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n) WR_ID(elt->wr.wr_id).offset = 0; #ifndef NDEBUG /* For assert(). */ - while (j) { - --j; - --sge; - sge->addr = 0; - sge->length = 0; - sge->lkey = 0; - } + sge->addr = 0; + sge->length = 0; + sge->lkey = 0; wr->num_sge = 0; #endif goto stop; @@ -1197,63 +1331,21 @@ mlx4_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n) #ifdef MLX4_PMD_SOFT_COUNTERS sent_size += sge->length; #endif - buf = NEXT(buf); - } - /* If buf is not NULL here and is not going to be linearized, - * nb_segs is not valid. */ - assert(j == segs); - assert((buf == NULL) || (linearize)); - /* Linearize extra segments. */ - if (linearize) { - struct ibv_sge *sge = &elt->sges[segs]; - linear_t *linear = &(*txq->elts_linear)[elts_head]; - unsigned int size = linearize_mbuf(linear, buf); - - assert(segs == (elemof(elt->sges) - 1)); - if (size == 0) { - /* Invalid packet. */ - DEBUG("%p: packet too large to be linearized.", - (void *)txq); - /* Clean up TX element. */ - WR_ID(elt->wr.wr_id).offset = 0; -#ifndef NDEBUG - /* For assert(). */ - while (j) { - --j; - --sge; - sge->addr = 0; - sge->length = 0; - sge->lkey = 0; - } - wr->num_sge = 0; -#endif - goto stop; - } - /* If MLX4_PMD_SGE_WR_N is 1, free mbuf immediately - * and clear offset from WR ID. */ - if (elemof(elt->sges) == 1) { - do { - struct rte_mbuf *next = NEXT(buf); + } else { +#if MLX4_PMD_SGE_WR_N > 1 + unsigned int ret; - rte_pktmbuf_free_seg(buf); - buf = next; - } while (buf != NULL); - WR_ID(wr->wr_id).offset = 0; - } - /* Set WR fields and fill SGE with linear buffer. */ - ++wr->num_sge; - /* Sanity checks, only relevant with debugging - * enabled. */ - assert(sge->addr == 0); - assert(sge->length == 0); - assert(sge->lkey == 0); - /* Update SGE. */ - sge->addr = (uintptr_t)&(*linear)[0]; - sge->length = size; - sge->lkey = txq->mr_linear->lkey; + ret = tx_burst_sg(txq, segs, elt, buf, elts_head); + if (ret == (unsigned int)-1) + goto stop; #ifdef MLX4_PMD_SOFT_COUNTERS - sent_size += size; + sent_size += ret; #endif +#else /* MLX4_PMD_SGE_WR_N > 1 */ + DEBUG("%p: TX scattered buffers support not" + " compiled in", (void *)txq); + goto stop; +#endif /* MLX4_PMD_SGE_WR_N > 1 */ } /* Link WRs together for ibv_post_send(). */ *wr_next = wr; -- 2.1.0