From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-wg0-f41.google.com (mail-wg0-f41.google.com [74.125.82.41]) by dpdk.org (Postfix) with ESMTP id 362C05A99 for ; Thu, 5 Mar 2015 12:28:13 +0100 (CET) Received: by wggx12 with SMTP id x12so52760739wgg.6 for ; Thu, 05 Mar 2015 03:28:13 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=dHTlc75VkpkDjgUxTzDJVbFrlFpSbHTtSVaMTHELogM=; b=a4awkjrp/5fT1JqTe+fMeD5RVZIsRQvr6dpAGE5IdLkZxEKENDjGgw+jZdJHaD+qqx 0jE2uovqdYYWDiavaDkoh162o/jIPQNAtRpQlmpNHM9v/Rhp4tV+KMG17CqT3VEFLooR fWUrOQReXz5crbT6FAFH2cxje4jbsfG+M+k7slbV6Jq5BkNO1twioPkq/zSJvYaLF1bv u57SESEbFGFmS86MwukR2hLyoDZGxMIUiu3PYaqApkKiBTclSO16+Co7MrHazU5L1KDF 12rxeq38OFCC45tF2ivZFQUd6TD7gORODcOUSXfF52/rWm1eu3oVIjcbSMPKjSNSEReO +jcw== X-Gm-Message-State: ALoCoQmU2w5y5GdFCSsmpfL6d58kFeMXY5sesJ+hYwHK/GUhyqYoj0M3sI3oGUdTJlt4I32WuYDG X-Received: by 10.181.12.5 with SMTP id em5mr22086678wid.84.1425554893095; Thu, 05 Mar 2015 03:28:13 -0800 (PST) Received: from vladz-laptop.cloudius-systems.com. ([212.143.139.214]) by mx.google.com with ESMTPSA id e2sm10069621wjy.46.2015.03.05.03.28.12 (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Thu, 05 Mar 2015 03:28:12 -0800 (PST) From: Vlad Zolotarov To: dev@dpdk.org Date: Thu, 5 Mar 2015 13:28:02 +0200 Message-Id: <1425554885-16901-4-git-send-email-vladz@cloudius-systems.com> X-Mailer: git-send-email 2.1.0 In-Reply-To: <1425554885-16901-1-git-send-email-vladz@cloudius-systems.com> References: <1425554885-16901-1-git-send-email-vladz@cloudius-systems.com> Subject: [dpdk-dev] [PATCH v2 3/6] ixgbe: Code refactoring X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: patches and discussions about DPDK List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Thu, 05 Mar 2015 11:28:13 -0000 - ixgbe_rx_alloc_bufs(): - Reset the rte_mbuf fields only when requested. - Take the RDT update out of the function. - Add the stub when RTE_LIBRTE_IXGBE_RX_ALLOW_BULK_ALLOC is not defined. - ixgbe_recv_scattered_pkts(): - Take the code that updates the fields of the cluster's HEAD buffer into the inline function. Signed-off-by: Vlad Zolotarov --- lib/librte_pmd_ixgbe/ixgbe_rxtx.c | 116 +++++++++++++++++++++++--------------- 1 file changed, 70 insertions(+), 46 deletions(-) diff --git a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c index 35a88d8..4c67a9e 100644 --- a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c +++ b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c @@ -1022,7 +1022,7 @@ ixgbe_rx_scan_hw_ring(struct igb_rx_queue *rxq) } static inline int -ixgbe_rx_alloc_bufs(struct igb_rx_queue *rxq) +ixgbe_rx_alloc_bufs(struct igb_rx_queue *rxq, bool reset_mbuf) { volatile union ixgbe_adv_rx_desc *rxdp; struct igb_rx_entry *rxep; @@ -1043,11 +1043,14 @@ ixgbe_rx_alloc_bufs(struct igb_rx_queue *rxq) for (i = 0; i < rxq->rx_free_thresh; ++i) { /* populate the static rte mbuf fields */ mb = rxep[i].mbuf; - rte_mbuf_refcnt_set(mb, 1); - mb->next = NULL; + if (reset_mbuf) { + rte_mbuf_refcnt_set(mb, 1); + mb->next = NULL; + mb->nb_segs = 1; + mb->port = rxq->port_id; + } + mb->data_off = RTE_PKTMBUF_HEADROOM; - mb->nb_segs = 1; - mb->port = rxq->port_id; /* populate the descriptors */ dma_addr = rte_cpu_to_le_64(RTE_MBUF_DATA_DMA_ADDR_DEFAULT(mb)); @@ -1055,10 +1058,6 @@ ixgbe_rx_alloc_bufs(struct igb_rx_queue *rxq) rxdp[i].read.pkt_addr = dma_addr; } - /* update tail pointer */ - rte_wmb(); - IXGBE_PCI_REG_WRITE(rxq->rdt_reg_addr, rxq->rx_free_trigger); - /* update state of internal queue structure */ rxq->rx_free_trigger = rxq->rx_free_trigger + rxq->rx_free_thresh; if (rxq->rx_free_trigger >= rxq->nb_rx_desc) @@ -1110,7 +1109,9 @@ rx_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, /* if required, allocate new buffers to replenish descriptors */ if (rxq->rx_tail > rxq->rx_free_trigger) { - if (ixgbe_rx_alloc_bufs(rxq) != 0) { + uint16_t cur_free_trigger = rxq->rx_free_trigger; + + if (ixgbe_rx_alloc_bufs(rxq, true) != 0) { int i, j; PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u " "queue_id=%u", (unsigned) rxq->port_id, @@ -1130,6 +1131,10 @@ rx_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, return 0; } + + /* update tail pointer */ + rte_wmb(); + IXGBE_PCI_REG_WRITE(rxq->rdt_reg_addr, cur_free_trigger); } if (rxq->rx_tail >= rxq->nb_rx_desc) @@ -1169,6 +1174,13 @@ ixgbe_recv_pkts_bulk_alloc(void *rx_queue, struct rte_mbuf **rx_pkts, return nb_rx; } +#else +static inline int +ixgbe_rx_alloc_bufs(__rte_unused struct igb_rx_queue *rxq, + __rte_unused bool reset_mbuf) +{ + return -ENOMEM; +} #endif /* RTE_LIBRTE_IXGBE_RX_ALLOW_BULK_ALLOC */ uint16_t @@ -1353,6 +1365,51 @@ ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, return (nb_rx); } +/** + * Initialize the first mbuf of the returned packet: + * - RX port identifier, + * - hardware offload data, if any: + * - RSS flag & hash, + * - IP checksum flag, + * - VLAN TCI, if any, + * - error flags. + * @head HEAD of the packet cluster + * @desc HW descriptor to get data from + * @port_id Port ID of the Rx queue + */ +static inline void ixgbe_fill_cluster_head_buf( + struct rte_mbuf *head, + union ixgbe_adv_rx_desc *desc, + uint8_t port_id, + uint32_t staterr) +{ + uint32_t hlen_type_rss; + uint64_t pkt_flags; + + head->port = port_id; + + /* + * The vlan_tci field is only valid when PKT_RX_VLAN_PKT is + * set in the pkt_flags field. + */ + head->vlan_tci = rte_le_to_cpu_16(desc->wb.upper.vlan); + hlen_type_rss = rte_le_to_cpu_32(desc->wb.lower.lo_dword.data); + pkt_flags = rx_desc_hlen_type_rss_to_pkt_flags(hlen_type_rss); + pkt_flags |= rx_desc_status_to_pkt_flags(staterr); + pkt_flags |= rx_desc_error_to_pkt_flags(staterr); + head->ol_flags = pkt_flags; + + if (likely(pkt_flags & PKT_RX_RSS_HASH)) + head->hash.rss = rte_le_to_cpu_32(desc->wb.lower.hi_dword.rss); + else if (pkt_flags & PKT_RX_FDIR) { + head->hash.fdir.hash = + rte_le_to_cpu_16(desc->wb.lower.hi_dword.csum_ip.csum) + & IXGBE_ATR_HASH_MASK; + head->hash.fdir.id = + rte_le_to_cpu_16(desc->wb.lower.hi_dword.csum_ip.ip_id); + } +} + uint16_t ixgbe_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t nb_pkts) @@ -1369,12 +1426,10 @@ ixgbe_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, union ixgbe_adv_rx_desc rxd; uint64_t dma; /* Physical address of mbuf data buffer */ uint32_t staterr; - uint32_t hlen_type_rss; uint16_t rx_id; uint16_t nb_rx; uint16_t nb_hold; uint16_t data_len; - uint64_t pkt_flags; nb_rx = 0; nb_hold = 0; @@ -1532,40 +1587,9 @@ ixgbe_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, (uint16_t) (data_len - ETHER_CRC_LEN); } - /* - * Initialize the first mbuf of the returned packet: - * - RX port identifier, - * - hardware offload data, if any: - * - RSS flag & hash, - * - IP checksum flag, - * - VLAN TCI, if any, - * - error flags. - */ - first_seg->port = rxq->port_id; - - /* - * The vlan_tci field is only valid when PKT_RX_VLAN_PKT is - * set in the pkt_flags field. - */ - first_seg->vlan_tci = rte_le_to_cpu_16(rxd.wb.upper.vlan); - hlen_type_rss = rte_le_to_cpu_32(rxd.wb.lower.lo_dword.data); - pkt_flags = rx_desc_hlen_type_rss_to_pkt_flags(hlen_type_rss); - pkt_flags = (pkt_flags | - rx_desc_status_to_pkt_flags(staterr)); - pkt_flags = (pkt_flags | - rx_desc_error_to_pkt_flags(staterr)); - first_seg->ol_flags = pkt_flags; - - if (likely(pkt_flags & PKT_RX_RSS_HASH)) - first_seg->hash.rss = - rte_le_to_cpu_32(rxd.wb.lower.hi_dword.rss); - else if (pkt_flags & PKT_RX_FDIR) { - first_seg->hash.fdir.hash = - rte_le_to_cpu_16(rxd.wb.lower.hi_dword.csum_ip.csum) - & IXGBE_ATR_HASH_MASK; - first_seg->hash.fdir.id = - rte_le_to_cpu_16(rxd.wb.lower.hi_dword.csum_ip.ip_id); - } + /* Initialize the first mbuf of the returned packet */ + ixgbe_fill_cluster_head_buf(first_seg, &rxd, rxq->port_id, + staterr); /* Prefetch data of first segment, if configured to do so. */ rte_packet_prefetch((char *)first_seg->buf_addr + -- 2.1.0