* [PATCH] net/tap: fix L4 checksum
@ 2023-08-22 7:32 David Marchand
2023-08-22 8:55 ` Olivier Matz
` (2 more replies)
0 siblings, 3 replies; 10+ messages in thread
From: David Marchand @ 2023-08-22 7:32 UTC (permalink / raw)
To: dev
Cc: stable, Ales Musil, Thomas Monjalon, Ophir Munk, Keith Wiles,
Raslan Darawsheh
The L4 checksum offloading API does not require l4_len to be set.
Make the driver discover the L4 headers size by itself.
Fixes: 6546e76056e3 ("net/tap: calculate checksums of multi segs packets")
Cc: stable@dpdk.org
Signed-off-by: David Marchand <david.marchand@redhat.com>
Tested-by: Ales Musil <amusil@redhat.com>
---
.mailmap | 1 +
drivers/net/tap/rte_eth_tap.c | 13 +++++++++++--
2 files changed, 12 insertions(+), 2 deletions(-)
diff --git a/.mailmap b/.mailmap
index 864d33ee46..b6a21b35cb 100644
--- a/.mailmap
+++ b/.mailmap
@@ -40,6 +40,7 @@ Aleksandr Loktionov <aleksandr.loktionov@intel.com>
Aleksandr Miloshenko <a.miloshenko@f5.com>
Aleksey Baulin <aleksey.baulin@gmail.com>
Aleksey Katargin <gureedo@gmail.com>
+Ales Musil <amusil@redhat.com>
Alexander Bechikov <asb.tyum@gmail.com>
Alexander Belyakov <abelyako@gmail.com>
Alexander Chernavin <achernavin@netgate.com>
diff --git a/drivers/net/tap/rte_eth_tap.c b/drivers/net/tap/rte_eth_tap.c
index bf98f75559..0ab214847a 100644
--- a/drivers/net/tap/rte_eth_tap.c
+++ b/drivers/net/tap/rte_eth_tap.c
@@ -645,13 +645,22 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
((mbuf->ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4) ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM))) {
+ unsigned int l4_len = 0;
+
is_cksum = 1;
+ if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
+ RTE_MBUF_F_TX_UDP_CKSUM)
+ l4_len = sizeof(struct rte_udp_hdr);
+ else if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
+ RTE_MBUF_F_TX_TCP_CKSUM)
+ l4_len = sizeof(struct rte_tcp_hdr);
+
/* Support only packets with at least layer 4
* header included in the first segment
*/
seg_len = rte_pktmbuf_data_len(mbuf);
- l234_hlen = mbuf->l2_len + mbuf->l3_len + mbuf->l4_len;
+ l234_hlen = mbuf->l2_len + mbuf->l3_len + l4_len;
if (seg_len < l234_hlen)
return -1;
@@ -661,7 +670,7 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
rte_memcpy(m_copy, rte_pktmbuf_mtod(mbuf, void *),
l234_hlen);
tap_tx_l3_cksum(m_copy, mbuf->ol_flags,
- mbuf->l2_len, mbuf->l3_len, mbuf->l4_len,
+ mbuf->l2_len, mbuf->l3_len, l4_len,
&l4_cksum, &l4_phdr_cksum,
&l4_raw_cksum);
iovecs[k].iov_base = m_copy;
--
2.41.0
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [PATCH] net/tap: fix L4 checksum
2023-08-22 7:32 [PATCH] net/tap: fix L4 checksum David Marchand
@ 2023-08-22 8:55 ` Olivier Matz
2023-08-22 15:44 ` David Marchand
2023-08-23 16:01 ` [PATCH v2 1/3] net/tap: fix L4 checksum offloading David Marchand
2023-08-24 7:18 ` [PATCH v3 1/3] net/tap: fix L4 " David Marchand
2 siblings, 1 reply; 10+ messages in thread
From: Olivier Matz @ 2023-08-22 8:55 UTC (permalink / raw)
To: David Marchand
Cc: dev, stable, Ales Musil, Thomas Monjalon, Ophir Munk,
Keith Wiles, Raslan Darawsheh
Hi David,
On Tue, Aug 22, 2023 at 09:32:44AM +0200, David Marchand wrote:
> The L4 checksum offloading API does not require l4_len to be set.
> Make the driver discover the L4 headers size by itself.
>
> Fixes: 6546e76056e3 ("net/tap: calculate checksums of multi segs packets")
> Cc: stable@dpdk.org
>
> Signed-off-by: David Marchand <david.marchand@redhat.com>
> Tested-by: Ales Musil <amusil@redhat.com>
> ---
> .mailmap | 1 +
> drivers/net/tap/rte_eth_tap.c | 13 +++++++++++--
> 2 files changed, 12 insertions(+), 2 deletions(-)
>
> diff --git a/.mailmap b/.mailmap
> index 864d33ee46..b6a21b35cb 100644
> --- a/.mailmap
> +++ b/.mailmap
> @@ -40,6 +40,7 @@ Aleksandr Loktionov <aleksandr.loktionov@intel.com>
> Aleksandr Miloshenko <a.miloshenko@f5.com>
> Aleksey Baulin <aleksey.baulin@gmail.com>
> Aleksey Katargin <gureedo@gmail.com>
> +Ales Musil <amusil@redhat.com>
> Alexander Bechikov <asb.tyum@gmail.com>
> Alexander Belyakov <abelyako@gmail.com>
> Alexander Chernavin <achernavin@netgate.com>
> diff --git a/drivers/net/tap/rte_eth_tap.c b/drivers/net/tap/rte_eth_tap.c
> index bf98f75559..0ab214847a 100644
> --- a/drivers/net/tap/rte_eth_tap.c
> +++ b/drivers/net/tap/rte_eth_tap.c
> @@ -645,13 +645,22 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
> ((mbuf->ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4) ||
> (mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM ||
> (mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM))) {
While looking at the patch, I noticed this line:
mbuf->ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4)
I think only RTE_MBUF_F_TX_IP_CKSUM should be checked.
> + unsigned int l4_len = 0;
> +
> is_cksum = 1;
>
> + if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
> + RTE_MBUF_F_TX_UDP_CKSUM)
> + l4_len = sizeof(struct rte_udp_hdr);
> + else if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
> + RTE_MBUF_F_TX_TCP_CKSUM)
> + l4_len = sizeof(struct rte_tcp_hdr);
> +
> /* Support only packets with at least layer 4
> * header included in the first segment
> */
> seg_len = rte_pktmbuf_data_len(mbuf);
> - l234_hlen = mbuf->l2_len + mbuf->l3_len + mbuf->l4_len;
> + l234_hlen = mbuf->l2_len + mbuf->l3_len + l4_len;
> if (seg_len < l234_hlen)
> return -1;
>
> @@ -661,7 +670,7 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
> rte_memcpy(m_copy, rte_pktmbuf_mtod(mbuf, void *),
> l234_hlen);
> tap_tx_l3_cksum(m_copy, mbuf->ol_flags,
> - mbuf->l2_len, mbuf->l3_len, mbuf->l4_len,
> + mbuf->l2_len, mbuf->l3_len, l4_len,
> &l4_cksum, &l4_phdr_cksum,
> &l4_raw_cksum);
> iovecs[k].iov_base = m_copy;
> --
> 2.41.0
>
Using rte_ipv4_udptcp_cksum() in this code would probably simplify it, and may
solve other issues (for instance the 0 checksum for UDP which has a special
meaning).
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [PATCH] net/tap: fix L4 checksum
2023-08-22 8:55 ` Olivier Matz
@ 2023-08-22 15:44 ` David Marchand
0 siblings, 0 replies; 10+ messages in thread
From: David Marchand @ 2023-08-22 15:44 UTC (permalink / raw)
To: Olivier Matz
Cc: dev, stable, Ales Musil, Thomas Monjalon, Ophir Munk,
Keith Wiles, Raslan Darawsheh
On Tue, Aug 22, 2023 at 10:55 AM Olivier Matz <olivier.matz@6wind.com> wrote:
>
> Hi David,
>
> On Tue, Aug 22, 2023 at 09:32:44AM +0200, David Marchand wrote:
> > The L4 checksum offloading API does not require l4_len to be set.
> > Make the driver discover the L4 headers size by itself.
> >
> > Fixes: 6546e76056e3 ("net/tap: calculate checksums of multi segs packets")
> > Cc: stable@dpdk.org
> >
> > Signed-off-by: David Marchand <david.marchand@redhat.com>
> > Tested-by: Ales Musil <amusil@redhat.com>
> > ---
> > .mailmap | 1 +
> > drivers/net/tap/rte_eth_tap.c | 13 +++++++++++--
> > 2 files changed, 12 insertions(+), 2 deletions(-)
> >
> > diff --git a/.mailmap b/.mailmap
> > index 864d33ee46..b6a21b35cb 100644
> > --- a/.mailmap
> > +++ b/.mailmap
> > @@ -40,6 +40,7 @@ Aleksandr Loktionov <aleksandr.loktionov@intel.com>
> > Aleksandr Miloshenko <a.miloshenko@f5.com>
> > Aleksey Baulin <aleksey.baulin@gmail.com>
> > Aleksey Katargin <gureedo@gmail.com>
> > +Ales Musil <amusil@redhat.com>
> > Alexander Bechikov <asb.tyum@gmail.com>
> > Alexander Belyakov <abelyako@gmail.com>
> > Alexander Chernavin <achernavin@netgate.com>
> > diff --git a/drivers/net/tap/rte_eth_tap.c b/drivers/net/tap/rte_eth_tap.c
> > index bf98f75559..0ab214847a 100644
> > --- a/drivers/net/tap/rte_eth_tap.c
> > +++ b/drivers/net/tap/rte_eth_tap.c
> > @@ -645,13 +645,22 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
> > ((mbuf->ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4) ||
> > (mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM ||
> > (mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM))) {
>
> While looking at the patch, I noticed this line:
>
> mbuf->ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4)
>
> I think only RTE_MBUF_F_TX_IP_CKSUM should be checked.
And tap_tx_l3_cksum is wrong too:
if (ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4)) {
This is a separate issue, I'll send another patch.
>
> > + unsigned int l4_len = 0;
> > +
> > is_cksum = 1;
> >
> > + if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
> > + RTE_MBUF_F_TX_UDP_CKSUM)
> > + l4_len = sizeof(struct rte_udp_hdr);
> > + else if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
> > + RTE_MBUF_F_TX_TCP_CKSUM)
> > + l4_len = sizeof(struct rte_tcp_hdr);
> > +
> > /* Support only packets with at least layer 4
> > * header included in the first segment
> > */
> > seg_len = rte_pktmbuf_data_len(mbuf);
> > - l234_hlen = mbuf->l2_len + mbuf->l3_len + mbuf->l4_len;
> > + l234_hlen = mbuf->l2_len + mbuf->l3_len + l4_len;
> > if (seg_len < l234_hlen)
> > return -1;
> >
> > @@ -661,7 +670,7 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
> > rte_memcpy(m_copy, rte_pktmbuf_mtod(mbuf, void *),
> > l234_hlen);
> > tap_tx_l3_cksum(m_copy, mbuf->ol_flags,
> > - mbuf->l2_len, mbuf->l3_len, mbuf->l4_len,
> > + mbuf->l2_len, mbuf->l3_len, l4_len,
> > &l4_cksum, &l4_phdr_cksum,
> > &l4_raw_cksum);
> > iovecs[k].iov_base = m_copy;
> > --
> > 2.41.0
> >
>
> Using rte_ipv4_udptcp_cksum() in this code would probably simplify it, and may
> solve other issues (for instance the 0 checksum for UDP which has a special
> meaning).
I agree such a rework would make the code easier to read, and may
solve other issues.
But I prefer to keep my original fix as is, and do what you propose as
a followup patch.
--
David Marchand
^ permalink raw reply [flat|nested] 10+ messages in thread
* [PATCH v2 1/3] net/tap: fix L4 checksum offloading
2023-08-22 7:32 [PATCH] net/tap: fix L4 checksum David Marchand
2023-08-22 8:55 ` Olivier Matz
@ 2023-08-23 16:01 ` David Marchand
2023-08-23 16:01 ` [PATCH v2 2/3] net/tap: fix IPv4 " David Marchand
2023-08-23 16:01 ` [PATCH v2 3/3] net/tap: rework " David Marchand
2023-08-24 7:18 ` [PATCH v3 1/3] net/tap: fix L4 " David Marchand
2 siblings, 2 replies; 10+ messages in thread
From: David Marchand @ 2023-08-23 16:01 UTC (permalink / raw)
To: dev
Cc: olivier.matz, stable, Ales Musil, Thomas Monjalon, Ophir Munk,
Keith Wiles, Raslan Darawsheh
The L4 checksum offloading API does not require l4_len to be set.
Make the driver discover the L4 headers size by itself.
Fixes: 6546e76056e3 ("net/tap: calculate checksums of multi segs packets")
Cc: stable@dpdk.org
Signed-off-by: David Marchand <david.marchand@redhat.com>
Tested-by: Ales Musil <amusil@redhat.com>
---
.mailmap | 1 +
drivers/net/tap/rte_eth_tap.c | 13 +++++++++++--
2 files changed, 12 insertions(+), 2 deletions(-)
diff --git a/.mailmap b/.mailmap
index 864d33ee46..b6a21b35cb 100644
--- a/.mailmap
+++ b/.mailmap
@@ -40,6 +40,7 @@ Aleksandr Loktionov <aleksandr.loktionov@intel.com>
Aleksandr Miloshenko <a.miloshenko@f5.com>
Aleksey Baulin <aleksey.baulin@gmail.com>
Aleksey Katargin <gureedo@gmail.com>
+Ales Musil <amusil@redhat.com>
Alexander Bechikov <asb.tyum@gmail.com>
Alexander Belyakov <abelyako@gmail.com>
Alexander Chernavin <achernavin@netgate.com>
diff --git a/drivers/net/tap/rte_eth_tap.c b/drivers/net/tap/rte_eth_tap.c
index bf98f75559..0ab214847a 100644
--- a/drivers/net/tap/rte_eth_tap.c
+++ b/drivers/net/tap/rte_eth_tap.c
@@ -645,13 +645,22 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
((mbuf->ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4) ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM))) {
+ unsigned int l4_len = 0;
+
is_cksum = 1;
+ if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
+ RTE_MBUF_F_TX_UDP_CKSUM)
+ l4_len = sizeof(struct rte_udp_hdr);
+ else if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
+ RTE_MBUF_F_TX_TCP_CKSUM)
+ l4_len = sizeof(struct rte_tcp_hdr);
+
/* Support only packets with at least layer 4
* header included in the first segment
*/
seg_len = rte_pktmbuf_data_len(mbuf);
- l234_hlen = mbuf->l2_len + mbuf->l3_len + mbuf->l4_len;
+ l234_hlen = mbuf->l2_len + mbuf->l3_len + l4_len;
if (seg_len < l234_hlen)
return -1;
@@ -661,7 +670,7 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
rte_memcpy(m_copy, rte_pktmbuf_mtod(mbuf, void *),
l234_hlen);
tap_tx_l3_cksum(m_copy, mbuf->ol_flags,
- mbuf->l2_len, mbuf->l3_len, mbuf->l4_len,
+ mbuf->l2_len, mbuf->l3_len, l4_len,
&l4_cksum, &l4_phdr_cksum,
&l4_raw_cksum);
iovecs[k].iov_base = m_copy;
--
2.41.0
^ permalink raw reply [flat|nested] 10+ messages in thread
* [PATCH v2 2/3] net/tap: fix IPv4 checksum offloading
2023-08-23 16:01 ` [PATCH v2 1/3] net/tap: fix L4 checksum offloading David Marchand
@ 2023-08-23 16:01 ` David Marchand
2023-08-23 16:01 ` [PATCH v2 3/3] net/tap: rework " David Marchand
1 sibling, 0 replies; 10+ messages in thread
From: David Marchand @ 2023-08-23 16:01 UTC (permalink / raw)
To: dev; +Cc: olivier.matz
Checking that one of RTE_MBUF_F_TX_IPV4 or RTE_MBUF_F_TX_IP_CKSUM is
present is not compliant with the offloading API which specifies that IP
checksum requires RTE_MBUF_F_TX_IP_CKSUM.
On the other hand, RTE_MBUF_F_TX_IP_CKSUM is invalid for IPv6 packets,
so we can simply check for RTE_MBUF_F_TX_IP_CKSUM and assume this is an
IPv4 packet.
Signed-off-by: David Marchand <david.marchand@redhat.com>
---
drivers/net/tap/rte_eth_tap.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git a/drivers/net/tap/rte_eth_tap.c b/drivers/net/tap/rte_eth_tap.c
index 0ab214847a..30b45ddc67 100644
--- a/drivers/net/tap/rte_eth_tap.c
+++ b/drivers/net/tap/rte_eth_tap.c
@@ -559,7 +559,7 @@ tap_tx_l3_cksum(char *packet, uint64_t ol_flags, unsigned int l2_len,
{
void *l3_hdr = packet + l2_len;
- if (ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4)) {
+ if (ol_flags & RTE_MBUF_F_TX_IP_CKSUM) {
struct rte_ipv4_hdr *iph = l3_hdr;
uint16_t cksum;
@@ -642,7 +642,7 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
nb_segs = mbuf->nb_segs;
if (txq->csum &&
- ((mbuf->ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4) ||
+ ((mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM))) {
unsigned int l4_len = 0;
--
2.41.0
^ permalink raw reply [flat|nested] 10+ messages in thread
* [PATCH v2 3/3] net/tap: rework checksum offloading
2023-08-23 16:01 ` [PATCH v2 1/3] net/tap: fix L4 checksum offloading David Marchand
2023-08-23 16:01 ` [PATCH v2 2/3] net/tap: fix IPv4 " David Marchand
@ 2023-08-23 16:01 ` David Marchand
1 sibling, 0 replies; 10+ messages in thread
From: David Marchand @ 2023-08-23 16:01 UTC (permalink / raw)
To: dev; +Cc: olivier.matz
Get rid of all the complicated code which copies data on the stack:
- allocate a new segment from the same mempool than the original mbuf,
- copy headers data in this segment,
- chain the new segment in place of headers of the original mbuf,
- use existing helpers for computing IP and TCP/UDP checksums,
- simplify the iovecs array filling,
With this rework, special care is needed for releasing mbufs in
pmd_tx_burst().
Signed-off-by: David Marchand <david.marchand@redhat.com>
---
drivers/net/tap/rte_eth_tap.c | 205 ++++++++++++----------------------
1 file changed, 73 insertions(+), 132 deletions(-)
diff --git a/drivers/net/tap/rte_eth_tap.c b/drivers/net/tap/rte_eth_tap.c
index 30b45ddc67..57d1126ce3 100644
--- a/drivers/net/tap/rte_eth_tap.c
+++ b/drivers/net/tap/rte_eth_tap.c
@@ -521,79 +521,13 @@ pmd_rx_burst(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
return num_rx;
}
-/* Finalize l4 checksum calculation */
-static void
-tap_tx_l4_cksum(uint16_t *l4_cksum, uint16_t l4_phdr_cksum,
- uint32_t l4_raw_cksum)
-{
- if (l4_cksum) {
- uint32_t cksum;
-
- cksum = __rte_raw_cksum_reduce(l4_raw_cksum);
- cksum += l4_phdr_cksum;
-
- cksum = ((cksum & 0xffff0000) >> 16) + (cksum & 0xffff);
- cksum = (~cksum) & 0xffff;
- if (cksum == 0)
- cksum = 0xffff;
- *l4_cksum = cksum;
- }
-}
-
-/* Accumulate L4 raw checksums */
-static void
-tap_tx_l4_add_rcksum(char *l4_data, unsigned int l4_len, uint16_t *l4_cksum,
- uint32_t *l4_raw_cksum)
-{
- if (l4_cksum == NULL)
- return;
-
- *l4_raw_cksum = __rte_raw_cksum(l4_data, l4_len, *l4_raw_cksum);
-}
-
-/* L3 and L4 pseudo headers checksum offloads */
-static void
-tap_tx_l3_cksum(char *packet, uint64_t ol_flags, unsigned int l2_len,
- unsigned int l3_len, unsigned int l4_len, uint16_t **l4_cksum,
- uint16_t *l4_phdr_cksum, uint32_t *l4_raw_cksum)
-{
- void *l3_hdr = packet + l2_len;
-
- if (ol_flags & RTE_MBUF_F_TX_IP_CKSUM) {
- struct rte_ipv4_hdr *iph = l3_hdr;
- uint16_t cksum;
-
- iph->hdr_checksum = 0;
- cksum = rte_raw_cksum(iph, l3_len);
- iph->hdr_checksum = (cksum == 0xffff) ? cksum : ~cksum;
- }
- if (ol_flags & RTE_MBUF_F_TX_L4_MASK) {
- void *l4_hdr;
-
- l4_hdr = packet + l2_len + l3_len;
- if ((ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM)
- *l4_cksum = &((struct rte_udp_hdr *)l4_hdr)->dgram_cksum;
- else if ((ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM)
- *l4_cksum = &((struct rte_tcp_hdr *)l4_hdr)->cksum;
- else
- return;
- **l4_cksum = 0;
- if (ol_flags & RTE_MBUF_F_TX_IPV4)
- *l4_phdr_cksum = rte_ipv4_phdr_cksum(l3_hdr, 0);
- else
- *l4_phdr_cksum = rte_ipv6_phdr_cksum(l3_hdr, 0);
- *l4_raw_cksum = __rte_raw_cksum(l4_hdr, l4_len, 0);
- }
-}
-
static inline int
tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
struct rte_mbuf **pmbufs,
uint16_t *num_packets, unsigned long *num_tx_bytes)
{
- int i;
- uint16_t l234_hlen;
struct pmd_process_private *process_private;
+ int i;
process_private = rte_eth_devices[txq->out_port].process_private;
@@ -602,19 +536,12 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
struct iovec iovecs[mbuf->nb_segs + 2];
struct tun_pi pi = { .flags = 0, .proto = 0x00 };
struct rte_mbuf *seg = mbuf;
- char m_copy[mbuf->data_len];
+ uint64_t l4_ol_flags;
int proto;
int n;
int j;
int k; /* current index in iovecs for copying segments */
- uint16_t seg_len; /* length of first segment */
- uint16_t nb_segs;
- uint16_t *l4_cksum; /* l4 checksum (pseudo header + payload) */
- uint32_t l4_raw_cksum = 0; /* TCP/UDP payload raw checksum */
- uint16_t l4_phdr_cksum = 0; /* TCP/UDP pseudo header checksum */
- uint16_t is_cksum = 0; /* in case cksum should be offloaded */
-
- l4_cksum = NULL;
+
if (txq->type == ETH_TUNTAP_TYPE_TUN) {
/*
* TUN and TAP are created with IFF_NO_PI disabled.
@@ -640,73 +567,83 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
iovecs[k].iov_len = sizeof(pi);
k++;
- nb_segs = mbuf->nb_segs;
- if (txq->csum &&
- ((mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM ||
- (mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM ||
- (mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM))) {
- unsigned int l4_len = 0;
-
- is_cksum = 1;
-
- if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
- RTE_MBUF_F_TX_UDP_CKSUM)
- l4_len = sizeof(struct rte_udp_hdr);
- else if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
- RTE_MBUF_F_TX_TCP_CKSUM)
- l4_len = sizeof(struct rte_tcp_hdr);
+ l4_ol_flags = mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK;
+ if (txq->csum && (mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM ||
+ l4_ol_flags == RTE_MBUF_F_TX_UDP_CKSUM ||
+ l4_ol_flags == RTE_MBUF_F_TX_TCP_CKSUM)) {
+ unsigned hdrlens = mbuf->l2_len + mbuf->l3_len;
+ uint16_t *l4_cksum;
+ void *l3_hdr;
+
+ if (l4_ol_flags == RTE_MBUF_F_TX_UDP_CKSUM)
+ hdrlens += sizeof(struct rte_udp_hdr);
+ else if (l4_ol_flags == RTE_MBUF_F_TX_TCP_CKSUM)
+ hdrlens += sizeof(struct rte_tcp_hdr);
+ else if (l4_ol_flags != RTE_MBUF_F_TX_L4_NO_CKSUM)
+ return -1;
/* Support only packets with at least layer 4
* header included in the first segment
*/
- seg_len = rte_pktmbuf_data_len(mbuf);
- l234_hlen = mbuf->l2_len + mbuf->l3_len + l4_len;
- if (seg_len < l234_hlen)
+ if (rte_pktmbuf_data_len(mbuf) < hdrlens)
return -1;
- /* To change checksums, work on a * copy of l2, l3
- * headers + l4 pseudo header
+ /* To change checksums (considering that a mbuf can be
+ * indirect, for example), copy l2, l3 and l4 headers
+ * in a new segment and chain it to existing data
*/
- rte_memcpy(m_copy, rte_pktmbuf_mtod(mbuf, void *),
- l234_hlen);
- tap_tx_l3_cksum(m_copy, mbuf->ol_flags,
- mbuf->l2_len, mbuf->l3_len, l4_len,
- &l4_cksum, &l4_phdr_cksum,
- &l4_raw_cksum);
- iovecs[k].iov_base = m_copy;
- iovecs[k].iov_len = l234_hlen;
- k++;
+ seg = rte_pktmbuf_copy(mbuf, mbuf->pool, 0, hdrlens);
+ if (seg == NULL)
+ return -1;
+ rte_pktmbuf_adj(mbuf, hdrlens);
+ rte_pktmbuf_chain(seg, mbuf);
+ pmbufs[i] = mbuf = seg;
+
+ l3_hdr = rte_pktmbuf_mtod_offset(mbuf, void *, mbuf->l2_len);
+ if (mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) {
+ struct rte_ipv4_hdr *iph = l3_hdr;
- /* Update next iovecs[] beyond l2, l3, l4 headers */
- if (seg_len > l234_hlen) {
- iovecs[k].iov_len = seg_len - l234_hlen;
- iovecs[k].iov_base =
- rte_pktmbuf_mtod(seg, char *) +
- l234_hlen;
- tap_tx_l4_add_rcksum(iovecs[k].iov_base,
- iovecs[k].iov_len, l4_cksum,
- &l4_raw_cksum);
- k++;
- nb_segs++;
+ iph->hdr_checksum = 0;
+ iph->hdr_checksum = rte_ipv4_cksum(iph);
}
- seg = seg->next;
+
+ if (l4_ol_flags == RTE_MBUF_F_TX_L4_NO_CKSUM)
+ goto skip_l4_cksum;
+
+ if (l4_ol_flags == RTE_MBUF_F_TX_UDP_CKSUM) {
+ struct rte_udp_hdr *udp_hdr;
+
+ udp_hdr = rte_pktmbuf_mtod_offset(mbuf, struct rte_udp_hdr *,
+ mbuf->l2_len + mbuf->l3_len);
+ l4_cksum = &udp_hdr->dgram_cksum;
+ } else {
+ struct rte_tcp_hdr *tcp_hdr;
+
+ tcp_hdr = rte_pktmbuf_mtod_offset(mbuf, struct rte_tcp_hdr *,
+ mbuf->l2_len + mbuf->l3_len);
+ l4_cksum = &tcp_hdr->cksum;
+ }
+
+ *l4_cksum = 0;
+ if (mbuf->ol_flags & RTE_MBUF_F_TX_IPV4) {
+ *l4_cksum = rte_ipv4_udptcp_cksum_mbuf(mbuf, l3_hdr,
+ mbuf->l2_len + mbuf->l3_len);
+ } else {
+ *l4_cksum = rte_ipv6_udptcp_cksum_mbuf(mbuf, l3_hdr,
+ mbuf->l2_len + mbuf->l3_len);
+ }
+skip_l4_cksum:
}
- for (j = k; j <= nb_segs; j++) {
- iovecs[j].iov_len = rte_pktmbuf_data_len(seg);
- iovecs[j].iov_base = rte_pktmbuf_mtod(seg, void *);
- if (is_cksum)
- tap_tx_l4_add_rcksum(iovecs[j].iov_base,
- iovecs[j].iov_len, l4_cksum,
- &l4_raw_cksum);
+ for (j = 0; j < mbuf->nb_segs; j++) {
+ iovecs[k].iov_len = rte_pktmbuf_data_len(seg);
+ iovecs[k].iov_base = rte_pktmbuf_mtod(seg, void *);
+ k++;
seg = seg->next;
}
- if (is_cksum)
- tap_tx_l4_cksum(l4_cksum, l4_phdr_cksum, l4_raw_cksum);
-
/* copy the tx frame data */
- n = writev(process_private->txq_fds[txq->queue_id], iovecs, j);
+ n = writev(process_private->txq_fds[txq->queue_id], iovecs, k);
if (n <= 0)
return -1;
@@ -801,11 +738,15 @@ pmd_tx_burst(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
break;
}
num_tx++;
- /* free original mbuf */
- rte_pktmbuf_free(mbuf_in);
- /* free tso mbufs */
- if (num_tso_mbufs > 0)
+ if (num_tso_mbufs == 0) {
+ /* tap_write_mbufs may prepend a segment to mbuf_in */
+ rte_pktmbuf_free(mbuf[0]);
+ } else {
+ /* free original mbuf */
+ rte_pktmbuf_free(mbuf_in);
+ /* free tso mbufs */
rte_pktmbuf_free_bulk(mbuf, num_tso_mbufs);
+ }
}
txq->stats.opackets += num_packets;
--
2.41.0
^ permalink raw reply [flat|nested] 10+ messages in thread
* [PATCH v3 1/3] net/tap: fix L4 checksum offloading
2023-08-22 7:32 [PATCH] net/tap: fix L4 checksum David Marchand
2023-08-22 8:55 ` Olivier Matz
2023-08-23 16:01 ` [PATCH v2 1/3] net/tap: fix L4 checksum offloading David Marchand
@ 2023-08-24 7:18 ` David Marchand
2023-08-24 7:18 ` [PATCH v3 2/3] net/tap: fix IPv4 " David Marchand
` (2 more replies)
2 siblings, 3 replies; 10+ messages in thread
From: David Marchand @ 2023-08-24 7:18 UTC (permalink / raw)
To: dev
Cc: olivier.matz, stable, Ales Musil, Thomas Monjalon, Keith Wiles,
Raslan Darawsheh, Ophir Munk
The L4 checksum offloading API does not require l4_len to be set.
Make the driver discover the L4 headers size by itself.
Fixes: 6546e76056e3 ("net/tap: calculate checksums of multi segs packets")
Cc: stable@dpdk.org
Signed-off-by: David Marchand <david.marchand@redhat.com>
Tested-by: Ales Musil <amusil@redhat.com>
---
.mailmap | 1 +
drivers/net/tap/rte_eth_tap.c | 13 +++++++++++--
2 files changed, 12 insertions(+), 2 deletions(-)
diff --git a/.mailmap b/.mailmap
index 864d33ee46..b6a21b35cb 100644
--- a/.mailmap
+++ b/.mailmap
@@ -40,6 +40,7 @@ Aleksandr Loktionov <aleksandr.loktionov@intel.com>
Aleksandr Miloshenko <a.miloshenko@f5.com>
Aleksey Baulin <aleksey.baulin@gmail.com>
Aleksey Katargin <gureedo@gmail.com>
+Ales Musil <amusil@redhat.com>
Alexander Bechikov <asb.tyum@gmail.com>
Alexander Belyakov <abelyako@gmail.com>
Alexander Chernavin <achernavin@netgate.com>
diff --git a/drivers/net/tap/rte_eth_tap.c b/drivers/net/tap/rte_eth_tap.c
index bf98f75559..0ab214847a 100644
--- a/drivers/net/tap/rte_eth_tap.c
+++ b/drivers/net/tap/rte_eth_tap.c
@@ -645,13 +645,22 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
((mbuf->ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4) ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM))) {
+ unsigned int l4_len = 0;
+
is_cksum = 1;
+ if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
+ RTE_MBUF_F_TX_UDP_CKSUM)
+ l4_len = sizeof(struct rte_udp_hdr);
+ else if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
+ RTE_MBUF_F_TX_TCP_CKSUM)
+ l4_len = sizeof(struct rte_tcp_hdr);
+
/* Support only packets with at least layer 4
* header included in the first segment
*/
seg_len = rte_pktmbuf_data_len(mbuf);
- l234_hlen = mbuf->l2_len + mbuf->l3_len + mbuf->l4_len;
+ l234_hlen = mbuf->l2_len + mbuf->l3_len + l4_len;
if (seg_len < l234_hlen)
return -1;
@@ -661,7 +670,7 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
rte_memcpy(m_copy, rte_pktmbuf_mtod(mbuf, void *),
l234_hlen);
tap_tx_l3_cksum(m_copy, mbuf->ol_flags,
- mbuf->l2_len, mbuf->l3_len, mbuf->l4_len,
+ mbuf->l2_len, mbuf->l3_len, l4_len,
&l4_cksum, &l4_phdr_cksum,
&l4_raw_cksum);
iovecs[k].iov_base = m_copy;
--
2.41.0
^ permalink raw reply [flat|nested] 10+ messages in thread
* [PATCH v3 2/3] net/tap: fix IPv4 checksum offloading
2023-08-24 7:18 ` [PATCH v3 1/3] net/tap: fix L4 " David Marchand
@ 2023-08-24 7:18 ` David Marchand
2023-08-24 7:18 ` [PATCH v3 3/3] net/tap: rework " David Marchand
2023-11-02 1:21 ` [PATCH v3 1/3] net/tap: fix L4 " Ferruh Yigit
2 siblings, 0 replies; 10+ messages in thread
From: David Marchand @ 2023-08-24 7:18 UTC (permalink / raw)
To: dev; +Cc: olivier.matz
Checking that one of RTE_MBUF_F_TX_IPV4 or RTE_MBUF_F_TX_IP_CKSUM is
present is not compliant with the offloading API which specifies that IP
checksum requires RTE_MBUF_F_TX_IP_CKSUM.
On the other hand, RTE_MBUF_F_TX_IP_CKSUM is invalid for IPv6 packets,
so we can simply check for RTE_MBUF_F_TX_IP_CKSUM and assume this is an
IPv4 packet.
Signed-off-by: David Marchand <david.marchand@redhat.com>
---
drivers/net/tap/rte_eth_tap.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git a/drivers/net/tap/rte_eth_tap.c b/drivers/net/tap/rte_eth_tap.c
index 0ab214847a..30b45ddc67 100644
--- a/drivers/net/tap/rte_eth_tap.c
+++ b/drivers/net/tap/rte_eth_tap.c
@@ -559,7 +559,7 @@ tap_tx_l3_cksum(char *packet, uint64_t ol_flags, unsigned int l2_len,
{
void *l3_hdr = packet + l2_len;
- if (ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4)) {
+ if (ol_flags & RTE_MBUF_F_TX_IP_CKSUM) {
struct rte_ipv4_hdr *iph = l3_hdr;
uint16_t cksum;
@@ -642,7 +642,7 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
nb_segs = mbuf->nb_segs;
if (txq->csum &&
- ((mbuf->ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_IPV4) ||
+ ((mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM ||
(mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM))) {
unsigned int l4_len = 0;
--
2.41.0
^ permalink raw reply [flat|nested] 10+ messages in thread
* [PATCH v3 3/3] net/tap: rework checksum offloading
2023-08-24 7:18 ` [PATCH v3 1/3] net/tap: fix L4 " David Marchand
2023-08-24 7:18 ` [PATCH v3 2/3] net/tap: fix IPv4 " David Marchand
@ 2023-08-24 7:18 ` David Marchand
2023-11-02 1:21 ` [PATCH v3 1/3] net/tap: fix L4 " Ferruh Yigit
2 siblings, 0 replies; 10+ messages in thread
From: David Marchand @ 2023-08-24 7:18 UTC (permalink / raw)
To: dev; +Cc: olivier.matz
Get rid of all the complicated code which copies data on the stack:
- allocate a new segment from the same mempool than the original mbuf,
- copy headers data in this segment,
- chain the new segment in place of headers of the original mbuf,
- use existing helpers for computing IP and TCP/UDP checksums,
- simplify the iovecs array filling,
With this rework, special care is needed for releasing mbufs in
pmd_tx_burst().
Signed-off-by: David Marchand <david.marchand@redhat.com>
---
Changes since v2:
- fixed build issue (strangely Fedora gcc seems lenient with goto
statement location...),
---
drivers/net/tap/rte_eth_tap.c | 205 ++++++++++++----------------------
1 file changed, 73 insertions(+), 132 deletions(-)
diff --git a/drivers/net/tap/rte_eth_tap.c b/drivers/net/tap/rte_eth_tap.c
index 30b45ddc67..90e95dba44 100644
--- a/drivers/net/tap/rte_eth_tap.c
+++ b/drivers/net/tap/rte_eth_tap.c
@@ -521,79 +521,13 @@ pmd_rx_burst(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
return num_rx;
}
-/* Finalize l4 checksum calculation */
-static void
-tap_tx_l4_cksum(uint16_t *l4_cksum, uint16_t l4_phdr_cksum,
- uint32_t l4_raw_cksum)
-{
- if (l4_cksum) {
- uint32_t cksum;
-
- cksum = __rte_raw_cksum_reduce(l4_raw_cksum);
- cksum += l4_phdr_cksum;
-
- cksum = ((cksum & 0xffff0000) >> 16) + (cksum & 0xffff);
- cksum = (~cksum) & 0xffff;
- if (cksum == 0)
- cksum = 0xffff;
- *l4_cksum = cksum;
- }
-}
-
-/* Accumulate L4 raw checksums */
-static void
-tap_tx_l4_add_rcksum(char *l4_data, unsigned int l4_len, uint16_t *l4_cksum,
- uint32_t *l4_raw_cksum)
-{
- if (l4_cksum == NULL)
- return;
-
- *l4_raw_cksum = __rte_raw_cksum(l4_data, l4_len, *l4_raw_cksum);
-}
-
-/* L3 and L4 pseudo headers checksum offloads */
-static void
-tap_tx_l3_cksum(char *packet, uint64_t ol_flags, unsigned int l2_len,
- unsigned int l3_len, unsigned int l4_len, uint16_t **l4_cksum,
- uint16_t *l4_phdr_cksum, uint32_t *l4_raw_cksum)
-{
- void *l3_hdr = packet + l2_len;
-
- if (ol_flags & RTE_MBUF_F_TX_IP_CKSUM) {
- struct rte_ipv4_hdr *iph = l3_hdr;
- uint16_t cksum;
-
- iph->hdr_checksum = 0;
- cksum = rte_raw_cksum(iph, l3_len);
- iph->hdr_checksum = (cksum == 0xffff) ? cksum : ~cksum;
- }
- if (ol_flags & RTE_MBUF_F_TX_L4_MASK) {
- void *l4_hdr;
-
- l4_hdr = packet + l2_len + l3_len;
- if ((ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM)
- *l4_cksum = &((struct rte_udp_hdr *)l4_hdr)->dgram_cksum;
- else if ((ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM)
- *l4_cksum = &((struct rte_tcp_hdr *)l4_hdr)->cksum;
- else
- return;
- **l4_cksum = 0;
- if (ol_flags & RTE_MBUF_F_TX_IPV4)
- *l4_phdr_cksum = rte_ipv4_phdr_cksum(l3_hdr, 0);
- else
- *l4_phdr_cksum = rte_ipv6_phdr_cksum(l3_hdr, 0);
- *l4_raw_cksum = __rte_raw_cksum(l4_hdr, l4_len, 0);
- }
-}
-
static inline int
tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
struct rte_mbuf **pmbufs,
uint16_t *num_packets, unsigned long *num_tx_bytes)
{
- int i;
- uint16_t l234_hlen;
struct pmd_process_private *process_private;
+ int i;
process_private = rte_eth_devices[txq->out_port].process_private;
@@ -602,19 +536,12 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
struct iovec iovecs[mbuf->nb_segs + 2];
struct tun_pi pi = { .flags = 0, .proto = 0x00 };
struct rte_mbuf *seg = mbuf;
- char m_copy[mbuf->data_len];
+ uint64_t l4_ol_flags;
int proto;
int n;
int j;
int k; /* current index in iovecs for copying segments */
- uint16_t seg_len; /* length of first segment */
- uint16_t nb_segs;
- uint16_t *l4_cksum; /* l4 checksum (pseudo header + payload) */
- uint32_t l4_raw_cksum = 0; /* TCP/UDP payload raw checksum */
- uint16_t l4_phdr_cksum = 0; /* TCP/UDP pseudo header checksum */
- uint16_t is_cksum = 0; /* in case cksum should be offloaded */
-
- l4_cksum = NULL;
+
if (txq->type == ETH_TUNTAP_TYPE_TUN) {
/*
* TUN and TAP are created with IFF_NO_PI disabled.
@@ -640,73 +567,83 @@ tap_write_mbufs(struct tx_queue *txq, uint16_t num_mbufs,
iovecs[k].iov_len = sizeof(pi);
k++;
- nb_segs = mbuf->nb_segs;
- if (txq->csum &&
- ((mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM ||
- (mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_UDP_CKSUM ||
- (mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) == RTE_MBUF_F_TX_TCP_CKSUM))) {
- unsigned int l4_len = 0;
-
- is_cksum = 1;
-
- if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
- RTE_MBUF_F_TX_UDP_CKSUM)
- l4_len = sizeof(struct rte_udp_hdr);
- else if ((mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK) ==
- RTE_MBUF_F_TX_TCP_CKSUM)
- l4_len = sizeof(struct rte_tcp_hdr);
+ l4_ol_flags = mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK;
+ if (txq->csum && (mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM ||
+ l4_ol_flags == RTE_MBUF_F_TX_UDP_CKSUM ||
+ l4_ol_flags == RTE_MBUF_F_TX_TCP_CKSUM)) {
+ unsigned hdrlens = mbuf->l2_len + mbuf->l3_len;
+ uint16_t *l4_cksum;
+ void *l3_hdr;
+
+ if (l4_ol_flags == RTE_MBUF_F_TX_UDP_CKSUM)
+ hdrlens += sizeof(struct rte_udp_hdr);
+ else if (l4_ol_flags == RTE_MBUF_F_TX_TCP_CKSUM)
+ hdrlens += sizeof(struct rte_tcp_hdr);
+ else if (l4_ol_flags != RTE_MBUF_F_TX_L4_NO_CKSUM)
+ return -1;
/* Support only packets with at least layer 4
* header included in the first segment
*/
- seg_len = rte_pktmbuf_data_len(mbuf);
- l234_hlen = mbuf->l2_len + mbuf->l3_len + l4_len;
- if (seg_len < l234_hlen)
+ if (rte_pktmbuf_data_len(mbuf) < hdrlens)
return -1;
- /* To change checksums, work on a * copy of l2, l3
- * headers + l4 pseudo header
+ /* To change checksums (considering that a mbuf can be
+ * indirect, for example), copy l2, l3 and l4 headers
+ * in a new segment and chain it to existing data
*/
- rte_memcpy(m_copy, rte_pktmbuf_mtod(mbuf, void *),
- l234_hlen);
- tap_tx_l3_cksum(m_copy, mbuf->ol_flags,
- mbuf->l2_len, mbuf->l3_len, l4_len,
- &l4_cksum, &l4_phdr_cksum,
- &l4_raw_cksum);
- iovecs[k].iov_base = m_copy;
- iovecs[k].iov_len = l234_hlen;
- k++;
+ seg = rte_pktmbuf_copy(mbuf, mbuf->pool, 0, hdrlens);
+ if (seg == NULL)
+ return -1;
+ rte_pktmbuf_adj(mbuf, hdrlens);
+ rte_pktmbuf_chain(seg, mbuf);
+ pmbufs[i] = mbuf = seg;
+
+ l3_hdr = rte_pktmbuf_mtod_offset(mbuf, void *, mbuf->l2_len);
+ if (mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) {
+ struct rte_ipv4_hdr *iph = l3_hdr;
- /* Update next iovecs[] beyond l2, l3, l4 headers */
- if (seg_len > l234_hlen) {
- iovecs[k].iov_len = seg_len - l234_hlen;
- iovecs[k].iov_base =
- rte_pktmbuf_mtod(seg, char *) +
- l234_hlen;
- tap_tx_l4_add_rcksum(iovecs[k].iov_base,
- iovecs[k].iov_len, l4_cksum,
- &l4_raw_cksum);
- k++;
- nb_segs++;
+ iph->hdr_checksum = 0;
+ iph->hdr_checksum = rte_ipv4_cksum(iph);
+ }
+
+ if (l4_ol_flags == RTE_MBUF_F_TX_L4_NO_CKSUM)
+ goto skip_l4_cksum;
+
+ if (l4_ol_flags == RTE_MBUF_F_TX_UDP_CKSUM) {
+ struct rte_udp_hdr *udp_hdr;
+
+ udp_hdr = rte_pktmbuf_mtod_offset(mbuf, struct rte_udp_hdr *,
+ mbuf->l2_len + mbuf->l3_len);
+ l4_cksum = &udp_hdr->dgram_cksum;
+ } else {
+ struct rte_tcp_hdr *tcp_hdr;
+
+ tcp_hdr = rte_pktmbuf_mtod_offset(mbuf, struct rte_tcp_hdr *,
+ mbuf->l2_len + mbuf->l3_len);
+ l4_cksum = &tcp_hdr->cksum;
+ }
+
+ *l4_cksum = 0;
+ if (mbuf->ol_flags & RTE_MBUF_F_TX_IPV4) {
+ *l4_cksum = rte_ipv4_udptcp_cksum_mbuf(mbuf, l3_hdr,
+ mbuf->l2_len + mbuf->l3_len);
+ } else {
+ *l4_cksum = rte_ipv6_udptcp_cksum_mbuf(mbuf, l3_hdr,
+ mbuf->l2_len + mbuf->l3_len);
}
- seg = seg->next;
}
- for (j = k; j <= nb_segs; j++) {
- iovecs[j].iov_len = rte_pktmbuf_data_len(seg);
- iovecs[j].iov_base = rte_pktmbuf_mtod(seg, void *);
- if (is_cksum)
- tap_tx_l4_add_rcksum(iovecs[j].iov_base,
- iovecs[j].iov_len, l4_cksum,
- &l4_raw_cksum);
+skip_l4_cksum:
+ for (j = 0; j < mbuf->nb_segs; j++) {
+ iovecs[k].iov_len = rte_pktmbuf_data_len(seg);
+ iovecs[k].iov_base = rte_pktmbuf_mtod(seg, void *);
+ k++;
seg = seg->next;
}
- if (is_cksum)
- tap_tx_l4_cksum(l4_cksum, l4_phdr_cksum, l4_raw_cksum);
-
/* copy the tx frame data */
- n = writev(process_private->txq_fds[txq->queue_id], iovecs, j);
+ n = writev(process_private->txq_fds[txq->queue_id], iovecs, k);
if (n <= 0)
return -1;
@@ -801,11 +738,15 @@ pmd_tx_burst(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
break;
}
num_tx++;
- /* free original mbuf */
- rte_pktmbuf_free(mbuf_in);
- /* free tso mbufs */
- if (num_tso_mbufs > 0)
+ if (num_tso_mbufs == 0) {
+ /* tap_write_mbufs may prepend a segment to mbuf_in */
+ rte_pktmbuf_free(mbuf[0]);
+ } else {
+ /* free original mbuf */
+ rte_pktmbuf_free(mbuf_in);
+ /* free tso mbufs */
rte_pktmbuf_free_bulk(mbuf, num_tso_mbufs);
+ }
}
txq->stats.opackets += num_packets;
--
2.41.0
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [PATCH v3 1/3] net/tap: fix L4 checksum offloading
2023-08-24 7:18 ` [PATCH v3 1/3] net/tap: fix L4 " David Marchand
2023-08-24 7:18 ` [PATCH v3 2/3] net/tap: fix IPv4 " David Marchand
2023-08-24 7:18 ` [PATCH v3 3/3] net/tap: rework " David Marchand
@ 2023-11-02 1:21 ` Ferruh Yigit
2 siblings, 0 replies; 10+ messages in thread
From: Ferruh Yigit @ 2023-11-02 1:21 UTC (permalink / raw)
To: David Marchand, dev
Cc: olivier.matz, stable, Ales Musil, Thomas Monjalon, Keith Wiles,
Raslan Darawsheh, Ophir Munk
On 8/24/2023 8:18 AM, David Marchand wrote:
> The L4 checksum offloading API does not require l4_len to be set.
> Make the driver discover the L4 headers size by itself.
>
> Fixes: 6546e76056e3 ("net/tap: calculate checksums of multi segs packets")
> Cc: stable@dpdk.org
>
> Signed-off-by: David Marchand <david.marchand@redhat.com>
> Tested-by: Ales Musil <amusil@redhat.com>
>
For series,
Acked-by: Ferruh Yigit <ferruh.yigit@amd.com>
Series applied to dpdk-next-net/main, thanks.
^ permalink raw reply [flat|nested] 10+ messages in thread
end of thread, other threads:[~2023-11-02 1:21 UTC | newest]
Thread overview: 10+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2023-08-22 7:32 [PATCH] net/tap: fix L4 checksum David Marchand
2023-08-22 8:55 ` Olivier Matz
2023-08-22 15:44 ` David Marchand
2023-08-23 16:01 ` [PATCH v2 1/3] net/tap: fix L4 checksum offloading David Marchand
2023-08-23 16:01 ` [PATCH v2 2/3] net/tap: fix IPv4 " David Marchand
2023-08-23 16:01 ` [PATCH v2 3/3] net/tap: rework " David Marchand
2023-08-24 7:18 ` [PATCH v3 1/3] net/tap: fix L4 " David Marchand
2023-08-24 7:18 ` [PATCH v3 2/3] net/tap: fix IPv4 " David Marchand
2023-08-24 7:18 ` [PATCH v3 3/3] net/tap: rework " David Marchand
2023-11-02 1:21 ` [PATCH v3 1/3] net/tap: fix L4 " Ferruh Yigit
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).