From: "Ferriter, Cian" <cian.ferriter@intel.com>
To: "Richardson, Bruce" <bruce.richardson@intel.com>,
"Yigit, Ferruh" <ferruh.yigit@intel.com>,
"Mcnamara, John" <john.mcnamara@intel.com>,
"Kovacevic, Marko" <marko.kovacevic@intel.com>
Cc: "dev@dpdk.org" <dev@dpdk.org>
Subject: Re: [dpdk-dev] [PATCH 19.08 v2] net/pcap: enable infinitely rxing a pcap file
Date: Wed, 5 Jun 2019 12:46:02 +0000 [thread overview]
Message-ID: <579B86492DFB364BBA627A48FB30C90E75D518F7@IRSMSX102.ger.corp.intel.com> (raw)
In-Reply-To: <20190605115620.32047-1-cian.ferriter@intel.com>
Adding in my changelog at the top of this, since I forgot to add it in the original mail:
v2:
* Rework the method of filling the ring to infinitely rx from
* Avoids potential huge allocation of mbufs
* Removes double allocation of mbufs used during queue setup
* rename count_packets_in_pcaps to count_packets_in_pcap
* initialize pcap_pkt_count in count_packets_in_pcap
* use RTE_PMD_REGISTER_PARAM_STRING <0|1> rather than <int>
* replace calls to rte_panic with proper error returning
* count rx and tx stat bytes in pcap_rx_infinite and tx_drop
* make internals->infinite_rx = infinite_rx assignment unconditional
* add cleanup for infinite_rx in eth_dev_close and pmd_pcap_remove
* add cleanup when multi seg mbufs are found
* add some clarifications to the documentation update
> -----Original Message-----
> From: Ferriter, Cian
> Sent: 05 June 2019 12:56
> To: Richardson, Bruce <bruce.richardson@intel.com>; Yigit, Ferruh
> <ferruh.yigit@intel.com>; Mcnamara, John <john.mcnamara@intel.com>;
> Kovacevic, Marko <marko.kovacevic@intel.com>
> Cc: dev@dpdk.org; Ferriter, Cian <cian.ferriter@intel.com>
> Subject: [PATCH 19.08 v2] net/pcap: enable infinitely rxing a pcap file
>
> It can be useful to use pcap files for some rudimental performance testing.
> This patch enables this functionality in the pcap driver.
>
> At a high level, this works by creaing a ring of sufficient size to store the
> packets in the pcap file passed to the application. When the rx function for
> this mode is called, packets are dequeued from the ring for use by the
> application and also enqueued back on to the ring to be "received" again.
>
> A tx_drop mode is also added since transmitting to a tx_pcap file isn't
> desirable at a high traffic rate.
>
> Jumbo frames are not supported in this mode. When filling the ring at rx
> queue setup time, the presence of multi segment mbufs is checked for.
> The PMD will exit on detection of these multi segment mbufs.
>
> Signed-off-by: Cian Ferriter <cian.ferriter@intel.com>
> ---
> doc/guides/nics/pcap_ring.rst | 19 +++
> drivers/net/pcap/rte_eth_pcap.c | 268
> ++++++++++++++++++++++++++++++--
> 2 files changed, 277 insertions(+), 10 deletions(-)
>
> diff --git a/doc/guides/nics/pcap_ring.rst b/doc/guides/nics/pcap_ring.rst
> index c1ef9196b..b272e6fe3 100644
> --- a/doc/guides/nics/pcap_ring.rst
> +++ b/doc/guides/nics/pcap_ring.rst
> @@ -106,6 +106,25 @@ Runtime Config Options
>
> --vdev 'net_pcap0,iface=eth0,phy_mac=1'
>
> +- Use the RX PCAP file to infinitely receive packets
> +
> + In case ``rx_pcap=`` configuration is set, user may want to use the
> + selected PCAP file for rudimental performance testing. This can be done
> with a ``devarg`` ``infinite_rx``, for example::
> +
> + --vdev 'net_pcap0,rx_pcap=file_rx.pcap,infinite_rx=1,tx_drop=1'
> +
> + When this mode is used, it is recommended to use the ``tx_drop``
> ``devarg``.
> +
> + This option is device wide, so all queues on a device will either have this
> enabled or disabled.
> +
> +- Drop all packets on transmit
> +
> + The user may want to drop all packets on tx for a device. This can be done
> with the ``tx_drop`` ``devarg``, for example::
> +
> + --vdev 'net_pcap0,rx_pcap=file_rx.pcap,tx_drop=1'
> +
> + One tx drop queue is created for each rxq on that device.
> +
> Examples of Usage
> ^^^^^^^^^^^^^^^^^
>
> diff --git a/drivers/net/pcap/rte_eth_pcap.c
> b/drivers/net/pcap/rte_eth_pcap.c index 10277b9b6..9d239d171 100644
> --- a/drivers/net/pcap/rte_eth_pcap.c
> +++ b/drivers/net/pcap/rte_eth_pcap.c
> @@ -39,6 +39,8 @@
> #define ETH_PCAP_TX_IFACE_ARG "tx_iface"
> #define ETH_PCAP_IFACE_ARG "iface"
> #define ETH_PCAP_PHY_MAC_ARG "phy_mac"
> +#define ETH_PCAP_INFINITE_RX_ARG "infinite_rx"
> +#define ETH_PCAP_TX_DROP_ARG "tx_drop"
>
> #define ETH_PCAP_ARG_MAXLEN 64
>
> @@ -64,6 +66,9 @@ struct pcap_rx_queue {
> struct queue_stat rx_stat;
> char name[PATH_MAX];
> char type[ETH_PCAP_ARG_MAXLEN];
> +
> + /* Contains pre-generated packets to be looped through */
> + struct rte_ring *pkts;
> };
>
> struct pcap_tx_queue {
> @@ -82,6 +87,7 @@ struct pmd_internals {
> int if_index;
> int single_iface;
> int phy_mac;
> + int infinite_rx;
> };
>
> struct pmd_process_private {
> @@ -109,6 +115,8 @@ static const char *valid_arguments[] = {
> ETH_PCAP_TX_IFACE_ARG,
> ETH_PCAP_IFACE_ARG,
> ETH_PCAP_PHY_MAC_ARG,
> + ETH_PCAP_INFINITE_RX_ARG,
> + ETH_PCAP_TX_DROP_ARG,
> NULL
> };
>
> @@ -178,6 +186,43 @@ eth_pcap_gather_data(unsigned char *data, struct
> rte_mbuf *mbuf)
> }
> }
>
> +static uint16_t
> +eth_pcap_rx_infinite(void *queue, struct rte_mbuf **bufs, uint16_t
> +nb_pkts) {
> + int i;
> + struct pcap_rx_queue *pcap_q = queue;
> + uint32_t rx_bytes = 0;
> +
> + if (unlikely(nb_pkts == 0))
> + return 0;
> +
> + if (rte_pktmbuf_alloc_bulk(pcap_q->mb_pool, bufs, nb_pkts) != 0)
> + return 0;
> +
> + for (i = 0; i < nb_pkts; i++) {
> + struct rte_mbuf *pcap_buf;
> + int err = rte_ring_dequeue(pcap_q->pkts, (void
> **)&pcap_buf);
> + if (err)
> + return i;
> +
> + rte_memcpy(rte_pktmbuf_mtod(bufs[i], void *),
> + rte_pktmbuf_mtod(pcap_buf, void *),
> + pcap_buf->data_len);
> + bufs[i]->data_len = pcap_buf->data_len;
> + bufs[i]->pkt_len = pcap_buf->pkt_len;
> + bufs[i]->port = pcap_q->port_id;
> + rx_bytes += pcap_buf->data_len;
> +
> + /* Enqueue packet back on ring to allow infinite rx. */
> + rte_ring_enqueue(pcap_q->pkts, pcap_buf);
> + }
> +
> + pcap_q->rx_stat.pkts += i;
> + pcap_q->rx_stat.bytes += rx_bytes;
> +
> + return i;
> +}
> +
> static uint16_t
> eth_pcap_rx(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts) { @@
> -320,6 +365,30 @@ eth_pcap_tx_dumper(void *queue, struct rte_mbuf
> **bufs, uint16_t nb_pkts)
> return num_tx;
> }
>
> +/*
> + * Callback to handle dropping packets in the infinite rx case.
> + */
> +static uint16_t
> +eth_tx_drop(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts) {
> + unsigned int i;
> + uint32_t tx_bytes = 0;
> + struct pcap_tx_queue *tx_queue = queue;
> +
> + if (unlikely(nb_pkts == 0))
> + return 0;
> +
> + for (i = 0; i < nb_pkts; i++) {
> + tx_bytes += bufs[i]->data_len;
> + rte_pktmbuf_free(bufs[i]);
> + }
> +
> + tx_queue->tx_stat.pkts += nb_pkts;
> + tx_queue->tx_stat.bytes += tx_bytes;
> +
> + return i;
> +}
> +
> /*
> * Callback to handle sending packets through a real NIC.
> */
> @@ -447,6 +516,24 @@ open_single_rx_pcap(const char *pcap_filename,
> pcap_t **pcap)
> return 0;
> }
>
> +static uint64_t
> +count_packets_in_pcap(pcap_t **pcap, struct pcap_rx_queue *pcap_q) {
> + const u_char *packet;
> + struct pcap_pkthdr header;
> + uint64_t pcap_pkt_count = 0;
> +
> + while ((packet = pcap_next(*pcap, &header)))
> + pcap_pkt_count++;
> +
> + /* The pcap is reopened so it can be used as normal later. */
> + pcap_close(*pcap);
> + *pcap = NULL;
> + open_single_rx_pcap(pcap_q->name, pcap);
> +
> + return pcap_pkt_count;
> +}
> +
> static int
> eth_dev_start(struct rte_eth_dev *dev)
> {
> @@ -639,8 +726,25 @@ eth_stats_reset(struct rte_eth_dev *dev) }
>
> static void
> -eth_dev_close(struct rte_eth_dev *dev __rte_unused)
> +eth_dev_close(struct rte_eth_dev *dev)
> {
> + unsigned int i;
> + struct pmd_internals *internals = dev->data->dev_private;
> +
> + /* Device wide flag, but cleanup must be performed per queue. */
> + if (internals->infinite_rx) {
> + for (i = 0; i < dev->data->nb_rx_queues; i++) {
> + struct pcap_rx_queue *pcap_q = &internals-
> >rx_queue[i];
> + struct rte_mbuf *pcap_buf;
> +
> + while (!rte_ring_dequeue(pcap_q->pkts,
> + (void **)&pcap_buf))
> + rte_pktmbuf_free(pcap_buf);
> +
> + rte_ring_free(pcap_q->pkts);
> + }
> + }
> +
> }
>
> static void
> @@ -671,6 +775,59 @@ eth_rx_queue_setup(struct rte_eth_dev *dev,
> pcap_q->queue_id = rx_queue_id;
> dev->data->rx_queues[rx_queue_id] = pcap_q;
>
> + if (internals->infinite_rx) {
> + struct pmd_process_private *pp;
> + char ring_name[NAME_MAX];
> + static uint32_t ring_number;
> + uint64_t pcap_pkt_count = 0;
> + struct rte_mbuf *bufs[1];
> + pcap_t **pcap;
> +
> + pp = rte_eth_devices[pcap_q->port_id].process_private;
> + pcap = &pp->rx_pcap[pcap_q->queue_id];
> +
> + if (unlikely(*pcap == NULL))
> + return -ENOENT;
> +
> + pcap_pkt_count = count_packets_in_pcap(pcap, pcap_q);
> +
> + snprintf(ring_name, sizeof(ring_name), "PCAP_RING%"
> PRIu16,
> + ring_number);
> +
> + pcap_q->pkts = rte_ring_create(ring_name,
> + rte_align64pow2(pcap_pkt_count + 1), 0,
> + RING_F_SP_ENQ | RING_F_SC_DEQ);
> + ring_number++;
> + if (!pcap_q->pkts)
> + return -ENOENT;
> +
> + /* Fill ring with packets from PCAP file one by one. */
> + while (eth_pcap_rx(pcap_q, bufs, 1)) {
> + /* Check for multiseg mbufs. */
> + if (bufs[0]->nb_segs != 1) {
> + rte_pktmbuf_free(*bufs);
> +
> + while (!rte_ring_dequeue(pcap_q->pkts,
> + (void **)bufs))
> + rte_pktmbuf_free(*bufs);
> +
> + rte_ring_free(pcap_q->pkts);
> + PMD_LOG(ERR, "Multiseg mbufs are not
> supported in infinite_rx "
> + "mode.");
> + return -EINVAL;
> + }
> +
> + rte_ring_enqueue_bulk(pcap_q->pkts,
> + (void * const *)bufs, 1, NULL);
> + }
> + /*
> + * Reset the stats for this queue since eth_pcap_rx calls
> above
> + * didn't result in the application receiving packets.
> + */
> + pcap_q->rx_stat.pkts = 0;
> + pcap_q->rx_stat.bytes = 0;
> + }
> +
> return 0;
> }
>
> @@ -908,6 +1065,34 @@ select_phy_mac(const char *key __rte_unused,
> const char *value,
> return 0;
> }
>
> +static int
> +get_infinite_rx_arg(const char *key __rte_unused,
> + const char *value, void *extra_args)
> +{
> + if (extra_args) {
> + const int infinite_rx = atoi(value);
> + int *enable_infinite_rx = extra_args;
> +
> + if (infinite_rx)
> + *enable_infinite_rx = 1;
> + }
> + return 0;
> +}
> +
> +static int
> +get_tx_drop_arg(const char *key __rte_unused,
> + const char *value, void *extra_args)
> +{
> + if (extra_args) {
> + const int tx_drop = atoi(value);
> + int *enable_tx_drop = extra_args;
> +
> + if (tx_drop)
> + *enable_tx_drop = 1;
> + }
> + return 0;
> +}
> +
> static struct rte_vdev_driver pmd_pcap_drv;
>
> static int
> @@ -1105,7 +1290,8 @@ static int
> eth_from_pcaps(struct rte_vdev_device *vdev,
> struct pmd_devargs *rx_queues, const unsigned int
> nb_rx_queues,
> struct pmd_devargs *tx_queues, const unsigned int
> nb_tx_queues,
> - int single_iface, unsigned int using_dumpers)
> + int single_iface, unsigned int using_dumpers,
> + unsigned int infinite_rx, unsigned int tx_drop)
> {
> struct pmd_internals *internals = NULL;
> struct rte_eth_dev *eth_dev = NULL;
> @@ -1132,9 +1318,17 @@ eth_from_pcaps(struct rte_vdev_device *vdev,
> }
> }
>
> - eth_dev->rx_pkt_burst = eth_pcap_rx;
> + internals->infinite_rx = infinite_rx;
> + /* Assign rx ops. */
> + if (infinite_rx)
> + eth_dev->rx_pkt_burst = eth_pcap_rx_infinite;
> + else
> + eth_dev->rx_pkt_burst = eth_pcap_rx;
>
> - if (using_dumpers)
> + /* Assign tx ops. */
> + if (tx_drop)
> + eth_dev->tx_pkt_burst = eth_tx_drop;
> + else if (using_dumpers)
> eth_dev->tx_pkt_burst = eth_pcap_tx_dumper;
> else
> eth_dev->tx_pkt_burst = eth_pcap_tx;
> @@ -1148,6 +1342,7 @@ pmd_pcap_probe(struct rte_vdev_device *dev) {
> const char *name;
> unsigned int is_rx_pcap = 0, is_tx_pcap = 0;
> + unsigned int infinite_rx = 0, infinite_rx_arg_cnt = 0, tx_drop = 0;
> struct rte_kvargs *kvlist;
> struct pmd_devargs pcaps = {0};
> struct pmd_devargs dumpers = {0};
> @@ -1216,7 +1411,25 @@ pmd_pcap_probe(struct rte_vdev_device *dev)
> is_rx_pcap = rte_kvargs_count(kvlist, ETH_PCAP_RX_PCAP_ARG) ? 1
> : 0;
> pcaps.num_of_queue = 0;
>
> + infinite_rx_arg_cnt = rte_kvargs_count(kvlist,
> + ETH_PCAP_INFINITE_RX_ARG);
> +
> if (is_rx_pcap) {
> + /*
> + * We check whether we want to infinitely rx the pcap file.
> + */
> + if (infinite_rx_arg_cnt == 1) {
> + ret = rte_kvargs_process(kvlist,
> + ETH_PCAP_INFINITE_RX_ARG,
> + &get_infinite_rx_arg, &infinite_rx);
> + if (ret < 0)
> + goto free_kvlist;
> + } else if (infinite_rx_arg_cnt >= 1) {
> + PMD_LOG(WARNING, "infinite_rx has not been
> enabled since the "
> + "argument has been provided more
> than once "
> + "for %s", name);
> + }
> +
> ret = rte_kvargs_process(kvlist, ETH_PCAP_RX_PCAP_ARG,
> &open_rx_pcap, &pcaps);
> } else {
> @@ -1228,18 +1441,32 @@ pmd_pcap_probe(struct rte_vdev_device *dev)
> goto free_kvlist;
>
> /*
> - * We check whether we want to open a TX stream to a real NIC or a
> - * pcap file
> + * We check whether we want to open a TX stream to a real NIC,
> + * a pcap file, or drop packets on tx
> */
> is_tx_pcap = rte_kvargs_count(kvlist, ETH_PCAP_TX_PCAP_ARG) ? 1
> : 0;
> dumpers.num_of_queue = 0;
>
> - if (is_tx_pcap)
> + if (rte_kvargs_count(kvlist, ETH_PCAP_TX_DROP_ARG) == 1) {
> + ret = rte_kvargs_process(kvlist, ETH_PCAP_TX_DROP_ARG,
> + &get_tx_drop_arg, &tx_drop);
> + if (ret < 0)
> + goto free_kvlist;
> + }
> +
> + if (tx_drop) {
> + unsigned int i;
> + /* Add 1 dummy queue per rxq which counts and drops
> packets. */
> + for (i = 0; i < pcaps.num_of_queue; i++)
> + ret = add_queue(&dumpers, "dummy", "tx_drop",
> NULL,
> + NULL);
> + } else if (is_tx_pcap) {
> ret = rte_kvargs_process(kvlist, ETH_PCAP_TX_PCAP_ARG,
> &open_tx_pcap, &dumpers);
> - else
> + } else {
> ret = rte_kvargs_process(kvlist, ETH_PCAP_TX_IFACE_ARG,
> &open_tx_iface, &dumpers);
> + }
>
> if (ret < 0)
> goto free_kvlist;
> @@ -1285,8 +1512,12 @@ pmd_pcap_probe(struct rte_vdev_device *dev)
> goto free_kvlist;
> }
>
> + PMD_LOG(INFO, "Configure pmd_pcap: infinite_rx is %s",
> + infinite_rx ? "enabled" : "disabled");
> +
> ret = eth_from_pcaps(dev, &pcaps, pcaps.num_of_queue,
> &dumpers,
> - dumpers.num_of_queue, single_iface, is_tx_pcap);
> + dumpers.num_of_queue, single_iface, is_tx_pcap,
> infinite_rx,
> + tx_drop);
>
> free_kvlist:
> rte_kvargs_free(kvlist);
> @@ -1297,6 +1528,7 @@ pmd_pcap_probe(struct rte_vdev_device *dev)
> static int pmd_pcap_remove(struct rte_vdev_device *dev) {
> + unsigned int i;
> struct pmd_internals *internals = NULL;
> struct rte_eth_dev *eth_dev = NULL;
>
> @@ -1318,6 +1550,20 @@ pmd_pcap_remove(struct rte_vdev_device *dev)
> eth_dev->data->mac_addrs = NULL;
> }
>
> + /* Device wide flag, but cleanup must be performed per queue. */
> + if (internals->infinite_rx) {
> + for (i = 0; i < eth_dev->data->nb_rx_queues; i++) {
> + struct pcap_rx_queue *pcap_q = &internals-
> >rx_queue[i];
> + struct rte_mbuf *pcap_buf;
> +
> + while (!rte_ring_dequeue(pcap_q->pkts,
> + (void **)&pcap_buf))
> + rte_pktmbuf_free(pcap_buf);
> +
> + rte_ring_free(pcap_q->pkts);
> + }
> + }
> +
> rte_free(eth_dev->process_private);
> rte_eth_dev_release_port(eth_dev);
>
> @@ -1338,7 +1584,9 @@ RTE_PMD_REGISTER_PARAM_STRING(net_pcap,
> ETH_PCAP_RX_IFACE_IN_ARG "=<ifc> "
> ETH_PCAP_TX_IFACE_ARG "=<ifc> "
> ETH_PCAP_IFACE_ARG "=<ifc> "
> - ETH_PCAP_PHY_MAC_ARG "=<int>");
> + ETH_PCAP_PHY_MAC_ARG "=<int>"
> + ETH_PCAP_INFINITE_RX_ARG "=<0|1>"
> + ETH_PCAP_TX_DROP_ARG "=<0|1>");
>
> RTE_INIT(eth_pcap_init_log)
> {
> --
> 2.17.1
next prev parent reply other threads:[~2019-06-05 12:46 UTC|newest]
Thread overview: 4+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-06-05 11:56 Cian Ferriter
2019-06-05 12:46 ` Ferriter, Cian [this message]
2019-06-12 14:10 ` Ferruh Yigit
2019-06-14 14:48 ` Ferriter, Cian
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=579B86492DFB364BBA627A48FB30C90E75D518F7@IRSMSX102.ger.corp.intel.com \
--to=cian.ferriter@intel.com \
--cc=bruce.richardson@intel.com \
--cc=dev@dpdk.org \
--cc=ferruh.yigit@intel.com \
--cc=john.mcnamara@intel.com \
--cc=marko.kovacevic@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).