* [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 @ 2017-09-25 17:19 Damien Clabaut 2017-09-26 0:46 ` Yongseok Koh 2017-09-26 1:46 ` Wiles, Keith 0 siblings, 2 replies; 8+ messages in thread From: Damien Clabaut @ 2017-09-25 17:19 UTC (permalink / raw) To: dev Hello DPDK devs, I am sending this message here as I did not find a bugtracker on the website. If this is the wrong place, I would kindly apologize and ask you to redirect me to the proper place, Thank you. Description of the issue: I am using pktgen-dpdk to replay a pcap file containing exactly 1 packet. The packet in question is generated using this Scapy command: pkt=(Ether(src="ec:0d:9a:37:d1:ab",dst="7c:fe:90:31:0d:52")/Dot1Q(vlan=2)/IP(dst="192.168.0.254")/UDP(sport=1020,dport=1021)/Raw(RandBin(size=8500))) The pcap is then replayed in pktgen-dpdk: ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m [1-7].0 -s 0:pcap/8500Bpp.pcap When I run this on a machine with Mellanox ConnectX-4 NIC (MCX4), the switch towards which I generate traffic gets a strange behaviour #sh int et29/1 | i rate 5 seconds input rate 39.4 Gbps (98.4% with framing overhead), 0 packets/sec A capture of this traffic (I used a monitor session to redirect all to a different port, connected to a machine on which I ran tcpdump) gives me this: 19:04:50.210792 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 19:04:50.210795 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 19:04:50.210796 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 19:04:50.210797 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 19:04:50.210799 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 The issue cannot be reproduced if any of the following conditions is met: - Set the size in the Raw(RandBin()) to a value lower than 1500 - Send the packet from a Mellanox ConnectX-3 (MCX3) NIC (both machines are indentical in terms of software). Is this a known problem ? I remain available for any question you may have. Regards, -- Damien Clabaut R&D vRouter ovh.qc.ca ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 2017-09-25 17:19 [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 Damien Clabaut @ 2017-09-26 0:46 ` Yongseok Koh 2017-09-26 12:43 ` Damien Clabaut 2017-09-26 1:46 ` Wiles, Keith 1 sibling, 1 reply; 8+ messages in thread From: Yongseok Koh @ 2017-09-26 0:46 UTC (permalink / raw) To: Damien Clabaut; +Cc: dev Hi, Damien Can you please let me know the versions of your SW - pktgen-dpdk, DPDK and MLNX_OFED? Also Firmware version if available. Thanks, Yongseok > On Sep 25, 2017, at 10:19 AM, Damien Clabaut <damien.clabaut@corp.ovh.com> wrote: > > Hello DPDK devs, > > I am sending this message here as I did not find a bugtracker on the website. > > If this is the wrong place, I would kindly apologize and ask you to redirect me to the proper place, > > Thank you. > > Description of the issue: > > I am using pktgen-dpdk to replay a pcap file containing exactly 1 packet. > > The packet in question is generated using this Scapy command: > > pkt=(Ether(src="ec:0d:9a:37:d1:ab",dst="7c:fe:90:31:0d:52")/Dot1Q(vlan=2)/IP(dst="192.168.0.254")/UDP(sport=1020,dport=1021)/Raw(RandBin(size=8500))) > > The pcap is then replayed in pktgen-dpdk: > > ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m [1-7].0 -s 0:pcap/8500Bpp.pcap > > When I run this on a machine with Mellanox ConnectX-4 NIC (MCX4), the switch towards which I generate traffic gets a strange behaviour > > #sh int et29/1 | i rate > 5 seconds input rate 39.4 Gbps (98.4% with framing overhead), 0 packets/sec > > A capture of this traffic (I used a monitor session to redirect all to a different port, connected to a machine on which I ran tcpdump) gives me this: > > 19:04:50.210792 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > 19:04:50.210795 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > 19:04:50.210796 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > 19:04:50.210797 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > 19:04:50.210799 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > > The issue cannot be reproduced if any of the following conditions is met: > > - Set the size in the Raw(RandBin()) to a value lower than 1500 > > - Send the packet from a Mellanox ConnectX-3 (MCX3) NIC (both machines are indentical in terms of software). > > Is this a known problem ? > > I remain available for any question you may have. > > Regards, > > -- > Damien Clabaut > R&D vRouter > ovh.qc.ca > ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 2017-09-26 0:46 ` Yongseok Koh @ 2017-09-26 12:43 ` Damien Clabaut 0 siblings, 0 replies; 8+ messages in thread From: Damien Clabaut @ 2017-09-26 12:43 UTC (permalink / raw) To: dev Hello and thank you for your answer, We use the following versions: MLNX DPDK version 16.11, revision 3.0 pktgen-dpdk version 3.3.8 mlnx_ofed version 4.1-1.0.2.0 fw revision: 12.18.1000 Regards, On 2017-09-25 08:46 PM, Yongseok Koh wrote: > Hi, Damien > > Can you please let me know the versions of your SW - pktgen-dpdk, DPDK and MLNX_OFED? > Also Firmware version if available. > > Thanks, > Yongseok > >> On Sep 25, 2017, at 10:19 AM, Damien Clabaut <damien.clabaut@corp.ovh.com> wrote: >> >> Hello DPDK devs, >> >> I am sending this message here as I did not find a bugtracker on the website. >> >> If this is the wrong place, I would kindly apologize and ask you to redirect me to the proper place, >> >> Thank you. >> >> Description of the issue: >> >> I am using pktgen-dpdk to replay a pcap file containing exactly 1 packet. >> >> The packet in question is generated using this Scapy command: >> >> pkt=(Ether(src="ec:0d:9a:37:d1:ab",dst="7c:fe:90:31:0d:52")/Dot1Q(vlan=2)/IP(dst="192.168.0.254")/UDP(sport=1020,dport=1021)/Raw(RandBin(size=8500))) >> >> The pcap is then replayed in pktgen-dpdk: >> >> ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m [1-7].0 -s 0:pcap/8500Bpp.pcap >> >> When I run this on a machine with Mellanox ConnectX-4 NIC (MCX4), the switch towards which I generate traffic gets a strange behaviour >> >> #sh int et29/1 | i rate >> 5 seconds input rate 39.4 Gbps (98.4% with framing overhead), 0 packets/sec >> >> A capture of this traffic (I used a monitor session to redirect all to a different port, connected to a machine on which I ran tcpdump) gives me this: >> >> 19:04:50.210792 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> 19:04:50.210795 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> 19:04:50.210796 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> 19:04:50.210797 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> 19:04:50.210799 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> >> The issue cannot be reproduced if any of the following conditions is met: >> >> - Set the size in the Raw(RandBin()) to a value lower than 1500 >> >> - Send the packet from a Mellanox ConnectX-3 (MCX3) NIC (both machines are indentical in terms of software). >> >> Is this a known problem ? >> >> I remain available for any question you may have. >> >> Regards, >> >> -- >> Damien Clabaut >> R&D vRouter >> ovh.qc.ca >> -- Damien Clabaut R&D vRouter ovh.qc.ca ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 2017-09-25 17:19 [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 Damien Clabaut 2017-09-26 0:46 ` Yongseok Koh @ 2017-09-26 1:46 ` Wiles, Keith 2017-09-26 13:09 ` Damien Clabaut 1 sibling, 1 reply; 8+ messages in thread From: Wiles, Keith @ 2017-09-26 1:46 UTC (permalink / raw) To: Damien Clabaut; +Cc: dev > On Sep 25, 2017, at 6:19 PM, Damien Clabaut <damien.clabaut@corp.ovh.com> wrote: > > Hello DPDK devs, > > I am sending this message here as I did not find a bugtracker on the website. > > If this is the wrong place, I would kindly apologize and ask you to redirect me to the proper place, > > Thank you. > > Description of the issue: > > I am using pktgen-dpdk to replay a pcap file containing exactly 1 packet. > > The packet in question is generated using this Scapy command: > > pkt=(Ether(src="ec:0d:9a:37:d1:ab",dst="7c:fe:90:31:0d:52")/Dot1Q(vlan=2)/IP(dst="192.168.0.254")/UDP(sport=1020,dport=1021)/Raw(RandBin(size=8500))) > > The pcap is then replayed in pktgen-dpdk: > > ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m [1-7].0 -s 0:pcap/8500Bpp.pcap This could be the issue as I can not setup your system (no cards). The pktgen command line is using 1-7 cores for TX/RX of packets. This means to pktgen to send the pcap from each core and this means the packet will be sent from each core. If you set the number of TX/RX cores to 1.0 then you should only see one. I assume you are using 1-7 cores to increase the bit rate closer to the performance of the card. > > When I run this on a machine with Mellanox ConnectX-4 NIC (MCX4), the switch towards which I generate traffic gets a strange behaviour > > #sh int et29/1 | i rate > 5 seconds input rate 39.4 Gbps (98.4% with framing overhead), 0 packets/sec > > A capture of this traffic (I used a monitor session to redirect all to a different port, connected to a machine on which I ran tcpdump) gives me this: > > 19:04:50.210792 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > 19:04:50.210795 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > 19:04:50.210796 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > 19:04:50.210797 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > 19:04:50.210799 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 > > The issue cannot be reproduced if any of the following conditions is met: > > - Set the size in the Raw(RandBin()) to a value lower than 1500 > > - Send the packet from a Mellanox ConnectX-3 (MCX3) NIC (both machines are indentical in terms of software). > > Is this a known problem ? > > I remain available for any question you may have. > > Regards, > > -- > Damien Clabaut > R&D vRouter > ovh.qc.ca > Regards, Keith ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 2017-09-26 1:46 ` Wiles, Keith @ 2017-09-26 13:09 ` Damien Clabaut 2017-09-28 20:44 ` Wiles, Keith 0 siblings, 1 reply; 8+ messages in thread From: Damien Clabaut @ 2017-09-26 13:09 UTC (permalink / raw) To: dev Hello Keith and thank you for your answer, The goal is indeed to generate as much traffic per machine as possible (we use pktgen-dpdk to benchmark datacenter routers before putting them on production). For this we use all available CPU power to send packets. Following your suggestion, I modified my command to: ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m 1.0 -s 0:pcap/8500Bpp.pcap The issue is still reproduced, though will slightly lower performance (reaching linerate at 8500 Bpp does not require much processing power) #sh int et 29/1 | i rate 5 seconds input rate 36.2 Gbps (90.5% with framing overhead), 0 packets/sec 5 seconds output rate 56 bps (0.0% with framing overhead), 0 packets/sec Regards, PS: Sorry for answering you directly, sending this message a second time on ML On 2017-09-25 09:46 PM, Wiles, Keith wrote: >> On Sep 25, 2017, at 6:19 PM, Damien Clabaut <damien.clabaut@corp.ovh.com> wrote: >> >> Hello DPDK devs, >> >> I am sending this message here as I did not find a bugtracker on the website. >> >> If this is the wrong place, I would kindly apologize and ask you to redirect me to the proper place, >> >> Thank you. >> >> Description of the issue: >> >> I am using pktgen-dpdk to replay a pcap file containing exactly 1 packet. >> >> The packet in question is generated using this Scapy command: >> >> pkt=(Ether(src="ec:0d:9a:37:d1:ab",dst="7c:fe:90:31:0d:52")/Dot1Q(vlan=2)/IP(dst="192.168.0.254")/UDP(sport=1020,dport=1021)/Raw(RandBin(size=8500))) >> >> The pcap is then replayed in pktgen-dpdk: >> >> ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m [1-7].0 -s 0:pcap/8500Bpp.pcap > This could be the issue as I can not setup your system (no cards). The pktgen command line is using 1-7 cores for TX/RX of packets. This means to pktgen to send the pcap from each core and this means the packet will be sent from each core. If you set the number of TX/RX cores to 1.0 then you should only see one. I assume you are using 1-7 cores to increase the bit rate closer to the performance of the card. > >> When I run this on a machine with Mellanox ConnectX-4 NIC (MCX4), the switch towards which I generate traffic gets a strange behaviour >> >> #sh int et29/1 | i rate >> 5 seconds input rate 39.4 Gbps (98.4% with framing overhead), 0 packets/sec >> >> A capture of this traffic (I used a monitor session to redirect all to a different port, connected to a machine on which I ran tcpdump) gives me this: >> >> 19:04:50.210792 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> 19:04:50.210795 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> 19:04:50.210796 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> 19:04:50.210797 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> 19:04:50.210799 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >> >> The issue cannot be reproduced if any of the following conditions is met: >> >> - Set the size in the Raw(RandBin()) to a value lower than 1500 >> >> - Send the packet from a Mellanox ConnectX-3 (MCX3) NIC (both machines are indentical in terms of software). >> >> Is this a known problem ? >> >> I remain available for any question you may have. >> >> Regards, >> >> -- >> Damien Clabaut >> R&D vRouter >> ovh.qc.ca >> > Regards, > Keith > -- Damien Clabaut R&D vRouter ovh.qc.ca ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 2017-09-26 13:09 ` Damien Clabaut @ 2017-09-28 20:44 ` Wiles, Keith 2017-10-11 18:59 ` Yongseok Koh 0 siblings, 1 reply; 8+ messages in thread From: Wiles, Keith @ 2017-09-28 20:44 UTC (permalink / raw) To: Damien Clabaut; +Cc: dev > On Sep 26, 2017, at 8:09 AM, Damien Clabaut <damien.clabaut@corp.ovh.com> wrote: > > Hello Keith and thank you for your answer, > > The goal is indeed to generate as much traffic per machine as possible (we use pktgen-dpdk to benchmark datacenter routers before putting them on production). > > For this we use all available CPU power to send packets. > > Following your suggestion, I modified my command to: > > ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m 1.0 -s 0:pcap/8500Bpp.pcap I just noticed you were sending 8500 byte frames and you have to modify Pktgen to increase the size of the mbufs in the mempool. I only configure the mbufs to 1518 byte buffers or really 2048 byte, but I only deal with 1518 max size. The size can be changed, but I am not next to a machine right now. > > The issue is still reproduced, though will slightly lower performance (reaching linerate at 8500 Bpp does not require much processing power) > > #sh int et 29/1 | i rate > 5 seconds input rate 36.2 Gbps (90.5% with framing overhead), 0 packets/sec > 5 seconds output rate 56 bps (0.0% with framing overhead), 0 packets/sec > > Regards, > > PS: Sorry for answering you directly, sending this message a second time on ML > > > On 2017-09-25 09:46 PM, Wiles, Keith wrote: >>> On Sep 25, 2017, at 6:19 PM, Damien Clabaut <damien.clabaut@corp.ovh.com> wrote: >>> >>> Hello DPDK devs, >>> >>> I am sending this message here as I did not find a bugtracker on the website. >>> >>> If this is the wrong place, I would kindly apologize and ask you to redirect me to the proper place, >>> >>> Thank you. >>> >>> Description of the issue: >>> >>> I am using pktgen-dpdk to replay a pcap file containing exactly 1 packet. >>> >>> The packet in question is generated using this Scapy command: >>> >>> pkt=(Ether(src="ec:0d:9a:37:d1:ab",dst="7c:fe:90:31:0d:52")/Dot1Q(vlan=2)/IP(dst="192.168.0.254")/UDP(sport=1020,dport=1021)/Raw(RandBin(size=8500))) >>> >>> The pcap is then replayed in pktgen-dpdk: >>> >>> ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m [1-7].0 -s 0:pcap/8500Bpp.pcap >> This could be the issue as I can not setup your system (no cards). The pktgen command line is using 1-7 cores for TX/RX of packets. This means to pktgen to send the pcap from each core and this means the packet will be sent from each core. If you set the number of TX/RX cores to 1.0 then you should only see one. I assume you are using 1-7 cores to increase the bit rate closer to the performance of the card. >> >>> When I run this on a machine with Mellanox ConnectX-4 NIC (MCX4), the switch towards which I generate traffic gets a strange behaviour >>> >>> #sh int et29/1 | i rate >>> 5 seconds input rate 39.4 Gbps (98.4% with framing overhead), 0 packets/sec >>> >>> A capture of this traffic (I used a monitor session to redirect all to a different port, connected to a machine on which I ran tcpdump) gives me this: >>> >>> 19:04:50.210792 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >>> 19:04:50.210795 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >>> 19:04:50.210796 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >>> 19:04:50.210797 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >>> 19:04:50.210799 00:00:00:00:00:00 (oui Ethernet) > 00:00:00:00:00:00 (oui Ethernet) Null Unnumbered, ef, Flags [Poll], length 1500 >>> >>> The issue cannot be reproduced if any of the following conditions is met: >>> >>> - Set the size in the Raw(RandBin()) to a value lower than 1500 >>> >>> - Send the packet from a Mellanox ConnectX-3 (MCX3) NIC (both machines are indentical in terms of software). >>> >>> Is this a known problem ? >>> >>> I remain available for any question you may have. >>> >>> Regards, >>> >>> -- >>> Damien Clabaut >>> R&D vRouter >>> ovh.qc.ca >>> >> Regards, >> Keith >> > > -- > Damien Clabaut > R&D vRouter > ovh.qc.ca > Regards, Keith ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 2017-09-28 20:44 ` Wiles, Keith @ 2017-10-11 18:59 ` Yongseok Koh 2017-10-12 12:34 ` Damien Clabaut 0 siblings, 1 reply; 8+ messages in thread From: Yongseok Koh @ 2017-10-11 18:59 UTC (permalink / raw) To: Damien Clabaut; +Cc: Wiles, Keith, dev On Thu, Sep 28, 2017 at 08:44:26PM +0000, Wiles, Keith wrote: > > > On Sep 26, 2017, at 8:09 AM, Damien Clabaut <damien.clabaut@corp.ovh.com> wrote: > > > > Hello Keith and thank you for your answer, > > > > The goal is indeed to generate as much traffic per machine as possible (we use pktgen-dpdk to benchmark datacenter routers before putting them on production). > > > > For this we use all available CPU power to send packets. > > > > Following your suggestion, I modified my command to: > > > > ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m 1.0 -s 0:pcap/8500Bpp.pcap > > > I just noticed you were sending 8500 byte frames and you have to modify Pktgen > to increase the size of the mbufs in the mempool. I only configure the mbufs > to 1518 byte buffers or really 2048 byte, but I only deal with 1518 max size. > The size can be changed, but I am not next to a machine right now. Hi Damien, Could you manage to resolve this issue? Keith mentioned pktgen doesn't support jumbo frames w/o modifying code. Do you still have an issue with Mellanox NIC and its PMDs? Please let me know. Thanks Yongseok ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 2017-10-11 18:59 ` Yongseok Koh @ 2017-10-12 12:34 ` Damien Clabaut 0 siblings, 0 replies; 8+ messages in thread From: Damien Clabaut @ 2017-10-12 12:34 UTC (permalink / raw) Cc: dev Hello, I opened a ticket with Mellanox in parallel. We are trying to figure out why it doesn't work on MCX4 but it works on MCX3, even though MCX3 is not officially supported and neither are jumbo frames. In case you want to check, the case ID is 00392710. Regards, On 2017-10-11 02:59 PM, Yongseok Koh wrote: > On Thu, Sep 28, 2017 at 08:44:26PM +0000, Wiles, Keith wrote: >>> On Sep 26, 2017, at 8:09 AM, Damien Clabaut <damien.clabaut@corp.ovh.com> wrote: >>> >>> Hello Keith and thank you for your answer, >>> >>> The goal is indeed to generate as much traffic per machine as possible (we use pktgen-dpdk to benchmark datacenter routers before putting them on production). >>> >>> For this we use all available CPU power to send packets. >>> >>> Following your suggestion, I modified my command to: >>> >>> ./app/app/x86_64-native-linuxapp-gcc/pktgen -l 0-7 -- -m 1.0 -s 0:pcap/8500Bpp.pcap >> >> I just noticed you were sending 8500 byte frames and you have to modify Pktgen >> to increase the size of the mbufs in the mempool. I only configure the mbufs >> to 1518 byte buffers or really 2048 byte, but I only deal with 1518 max size. >> The size can be changed, but I am not next to a machine right now. > Hi Damien, > > Could you manage to resolve this issue? Keith mentioned pktgen doesn't support > jumbo frames w/o modifying code. Do you still have an issue with Mellanox NIC > and its PMDs? Please let me know. > > Thanks > Yongseok -- Damien Clabaut R&D vRouter ovh.qc.ca ^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2017-10-12 12:34 UTC | newest] Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2017-09-25 17:19 [dpdk-dev] Issue with pktgen-dpdk replaying >1500bytes pcap on MCX4 Damien Clabaut 2017-09-26 0:46 ` Yongseok Koh 2017-09-26 12:43 ` Damien Clabaut 2017-09-26 1:46 ` Wiles, Keith 2017-09-26 13:09 ` Damien Clabaut 2017-09-28 20:44 ` Wiles, Keith 2017-10-11 18:59 ` Yongseok Koh 2017-10-12 12:34 ` Damien Clabaut
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).