DPDK patches and discussions
 help / color / mirror / Atom feed
From: "Zhang, Qi Z" <qi.z.zhang@intel.com>
To: "Liu, Mingxia" <mingxia.liu@intel.com>,
	"dev@dpdk.org" <dev@dpdk.org>,
	"Wu, Jingjing" <jingjing.wu@intel.com>,
	"Xing, Beilei" <beilei.xing@intel.com>
Subject: RE: [PATCH v6 1/6] common/idpf: add hw statistics
Date: Wed, 8 Feb 2023 02:00:00 +0000	[thread overview]
Message-ID: <DM4PR11MB59946DF079C4AC6ED6FDD8F9D7D89@DM4PR11MB5994.namprd11.prod.outlook.com> (raw)
In-Reply-To: <20230207101650.2402452-2-mingxia.liu@intel.com>



> -----Original Message-----
> From: Liu, Mingxia <mingxia.liu@intel.com>
> Sent: Tuesday, February 7, 2023 6:17 PM
> To: dev@dpdk.org; Zhang, Qi Z <qi.z.zhang@intel.com>; Wu, Jingjing
> <jingjing.wu@intel.com>; Xing, Beilei <beilei.xing@intel.com>
> Cc: Liu, Mingxia <mingxia.liu@intel.com>
> Subject: [PATCH v6 1/6] common/idpf: add hw statistics

Suggest to use ./devtools/check-git-log.sh to fix any title warning if possible
Also the main purpose of this patch is to support stats_get /stats_reset API,
the prefix is more reasonable to be "net/idpf" but not "common/idpf.

Please fix other patches if any similar issue. 

> 
> This patch add hardware packets/bytes statistics.
> 
> Signed-off-by: Mingxia Liu <mingxia.liu@intel.com>
> ---
>  drivers/common/idpf/idpf_common_device.c   | 17 +++++
>  drivers/common/idpf/idpf_common_device.h   |  4 +
>  drivers/common/idpf/idpf_common_virtchnl.c | 27 +++++++
> drivers/common/idpf/idpf_common_virtchnl.h |  3 +
>  drivers/common/idpf/version.map            |  2 +
>  drivers/net/idpf/idpf_ethdev.c             | 86 ++++++++++++++++++++++
>  6 files changed, 139 insertions(+)
> 
> diff --git a/drivers/common/idpf/idpf_common_device.c
> b/drivers/common/idpf/idpf_common_device.c
> index 48b3e3c0dd..5475a3e52c 100644
> --- a/drivers/common/idpf/idpf_common_device.c
> +++ b/drivers/common/idpf/idpf_common_device.c
> @@ -652,4 +652,21 @@ idpf_vport_info_init(struct idpf_vport *vport,
>  	return 0;
>  }
> 
> +void
> +idpf_vport_stats_update(struct virtchnl2_vport_stats *oes, struct
> +virtchnl2_vport_stats *nes) {
> +	nes->rx_bytes = nes->rx_bytes - oes->rx_bytes;
> +	nes->rx_unicast = nes->rx_unicast - oes->rx_unicast;
> +	nes->rx_multicast = nes->rx_multicast - oes->rx_multicast;
> +	nes->rx_broadcast = nes->rx_broadcast - oes->rx_broadcast;
> +	nes->rx_errors = nes->rx_errors - oes->rx_errors;
> +	nes->rx_discards = nes->rx_discards - oes->rx_discards;
> +	nes->tx_bytes = nes->tx_bytes - oes->tx_bytes;
> +	nes->tx_unicast = nes->tx_unicast - oes->tx_unicast;
> +	nes->tx_multicast = nes->tx_multicast - oes->tx_multicast;
> +	nes->tx_broadcast = nes->tx_broadcast - oes->tx_broadcast;
> +	nes->tx_errors = nes->tx_errors - oes->tx_errors;
> +	nes->tx_discards = nes->tx_discards - oes->tx_discards; }
> +
>  RTE_LOG_REGISTER_SUFFIX(idpf_common_logtype, common, NOTICE); diff
> --git a/drivers/common/idpf/idpf_common_device.h
> b/drivers/common/idpf/idpf_common_device.h
> index 545117df79..1d8e7d405a 100644
> --- a/drivers/common/idpf/idpf_common_device.h
> +++ b/drivers/common/idpf/idpf_common_device.h
> @@ -115,6 +115,8 @@ struct idpf_vport {
>  	bool tx_vec_allowed;
>  	bool rx_use_avx512;
>  	bool tx_use_avx512;
> +
> +	struct virtchnl2_vport_stats eth_stats_offset;
>  };
> 
>  /* Message type read in virtual channel from PF */ @@ -191,5 +193,7 @@
> int idpf_vport_irq_unmap_config(struct idpf_vport *vport, uint16_t
> nb_rx_queues)  __rte_internal  int idpf_vport_info_init(struct idpf_vport
> *vport,
>  			 struct virtchnl2_create_vport *vport_info);
> +__rte_internal
> +void idpf_vport_stats_update(struct virtchnl2_vport_stats *oes, struct
> +virtchnl2_vport_stats *nes);
> 
>  #endif /* _IDPF_COMMON_DEVICE_H_ */
> diff --git a/drivers/common/idpf/idpf_common_virtchnl.c
> b/drivers/common/idpf/idpf_common_virtchnl.c
> index 31fadefbd3..40cff34c09 100644
> --- a/drivers/common/idpf/idpf_common_virtchnl.c
> +++ b/drivers/common/idpf/idpf_common_virtchnl.c
> @@ -217,6 +217,7 @@ idpf_vc_cmd_execute(struct idpf_adapter *adapter,
> struct idpf_cmd_info *args)
>  	case VIRTCHNL2_OP_UNMAP_QUEUE_VECTOR:
>  	case VIRTCHNL2_OP_ALLOC_VECTORS:
>  	case VIRTCHNL2_OP_DEALLOC_VECTORS:
> +	case VIRTCHNL2_OP_GET_STATS:
>  		/* for init virtchnl ops, need to poll the response */
>  		err = idpf_vc_one_msg_read(adapter, args->ops, args-
> >out_size, args->out_buffer);
>  		clear_cmd(adapter);
> @@ -806,6 +807,32 @@ idpf_vc_ptype_info_query(struct idpf_adapter
> *adapter)
>  	return err;
>  }
> 
> +int
> +idpf_vc_stats_query(struct idpf_vport *vport,
> +		struct virtchnl2_vport_stats **pstats) {
> +	struct idpf_adapter *adapter = vport->adapter;
> +	struct virtchnl2_vport_stats vport_stats;
> +	struct idpf_cmd_info args;
> +	int err;
> +
> +	vport_stats.vport_id = vport->vport_id;
> +	args.ops = VIRTCHNL2_OP_GET_STATS;
> +	args.in_args = (u8 *)&vport_stats;
> +	args.in_args_size = sizeof(vport_stats);
> +	args.out_buffer = adapter->mbx_resp;
> +	args.out_size = IDPF_DFLT_MBX_BUF_SIZE;
> +
> +	err = idpf_vc_cmd_execute(adapter, &args);
> +	if (err) {
> +		DRV_LOG(ERR, "Failed to execute command of
> VIRTCHNL2_OP_GET_STATS");
> +		*pstats = NULL;
> +		return err;
> +	}
> +	*pstats = (struct virtchnl2_vport_stats *)args.out_buffer;
> +	return 0;
> +}
> +
>  #define IDPF_RX_BUF_STRIDE		64
>  int
>  idpf_vc_rxq_config(struct idpf_vport *vport, struct idpf_rx_queue *rxq) diff -
> -git a/drivers/common/idpf/idpf_common_virtchnl.h
> b/drivers/common/idpf/idpf_common_virtchnl.h
> index c105f02836..6b94fd5b8f 100644
> --- a/drivers/common/idpf/idpf_common_virtchnl.h
> +++ b/drivers/common/idpf/idpf_common_virtchnl.h
> @@ -49,4 +49,7 @@ __rte_internal
>  int idpf_vc_rxq_config(struct idpf_vport *vport, struct idpf_rx_queue *rxq);
> __rte_internal  int idpf_vc_txq_config(struct idpf_vport *vport, struct
> idpf_tx_queue *txq);
> +__rte_internal
> +int idpf_vc_stats_query(struct idpf_vport *vport,
> +			struct virtchnl2_vport_stats **pstats);
>  #endif /* _IDPF_COMMON_VIRTCHNL_H_ */
> diff --git a/drivers/common/idpf/version.map
> b/drivers/common/idpf/version.map index 8b33130bd6..e6a02828ba
> 100644
> --- a/drivers/common/idpf/version.map
> +++ b/drivers/common/idpf/version.map
> @@ -46,6 +46,7 @@ INTERNAL {
>  	idpf_vc_rss_key_set;
>  	idpf_vc_rss_lut_set;
>  	idpf_vc_rxq_config;
> +	idpf_vc_stats_query;
>  	idpf_vc_txq_config;
>  	idpf_vc_vectors_alloc;
>  	idpf_vc_vectors_dealloc;
> @@ -59,6 +60,7 @@ INTERNAL {
>  	idpf_vport_irq_map_config;
>  	idpf_vport_irq_unmap_config;
>  	idpf_vport_rss_config;
> +	idpf_vport_stats_update;
> 
>  	local: *;
>  };
> diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
> index 33f5e90743..02ddb0330a 100644
> --- a/drivers/net/idpf/idpf_ethdev.c
> +++ b/drivers/net/idpf/idpf_ethdev.c
> @@ -140,6 +140,87 @@ idpf_dev_supported_ptypes_get(struct rte_eth_dev
> *dev __rte_unused)
>  	return ptypes;
>  }
> 
> +static uint64_t
> +idpf_get_mbuf_alloc_failed_stats(struct rte_eth_dev *dev) {
> +	uint64_t mbuf_alloc_failed = 0;
> +	struct idpf_rx_queue *rxq;
> +	int i = 0;
> +
> +	for (i = 0; i < dev->data->nb_rx_queues; i++) {
> +		rxq = dev->data->rx_queues[i];
> +		mbuf_alloc_failed += __atomic_load_n(&rxq-
> >rx_stats.mbuf_alloc_failed,
> +						     __ATOMIC_RELAXED);
> +	}
> +
> +	return mbuf_alloc_failed;
> +}
> +
> +static int
> +idpf_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats
> +*stats) {
> +	struct idpf_vport *vport =
> +		(struct idpf_vport *)dev->data->dev_private;
> +	struct virtchnl2_vport_stats *pstats = NULL;
> +	int ret;
> +
> +	ret = idpf_vc_stats_query(vport, &pstats);
> +	if (ret == 0) {
> +		uint8_t crc_stats_len = (dev->data-
> >dev_conf.rxmode.offloads &
> +					 RTE_ETH_RX_OFFLOAD_KEEP_CRC) ?
> 0 :
> +					 RTE_ETHER_CRC_LEN;
> +
> +		idpf_vport_stats_update(&vport->eth_stats_offset, pstats);
> +		stats->ipackets = pstats->rx_unicast + pstats->rx_multicast +
> +				pstats->rx_broadcast - pstats->rx_discards;
> +		stats->opackets = pstats->tx_broadcast + pstats-
> >tx_multicast +
> +						pstats->tx_unicast;
> +		stats->imissed = pstats->rx_discards;
> +		stats->oerrors = pstats->tx_errors + pstats->tx_discards;
> +		stats->ibytes = pstats->rx_bytes;
> +		stats->ibytes -= stats->ipackets * crc_stats_len;
> +		stats->obytes = pstats->tx_bytes;
> +
> +		dev->data->rx_mbuf_alloc_failed =
> idpf_get_mbuf_alloc_failed_stats(dev);
> +		stats->rx_nombuf = dev->data->rx_mbuf_alloc_failed;
> +	} else {
> +		PMD_DRV_LOG(ERR, "Get statistics failed");
> +	}
> +	return ret;
> +}
> +
> +static void
> +idpf_reset_mbuf_alloc_failed_stats(struct rte_eth_dev *dev) {
> +	struct idpf_rx_queue *rxq;
> +	int i;
> +
> +	for (i = 0; i < dev->data->nb_rx_queues; i++) {
> +		rxq = dev->data->rx_queues[i];
> +		__atomic_store_n(&rxq->rx_stats.mbuf_alloc_failed, 0,
> __ATOMIC_RELAXED);
> +	}
> +}
> +
> +static int
> +idpf_dev_stats_reset(struct rte_eth_dev *dev) {
> +	struct idpf_vport *vport =
> +		(struct idpf_vport *)dev->data->dev_private;
> +	struct virtchnl2_vport_stats *pstats = NULL;
> +	int ret;
> +
> +	ret = idpf_vc_stats_query(vport, &pstats);
> +	if (ret != 0)
> +		return ret;
> +
> +	/* set stats offset base on current values */
> +	vport->eth_stats_offset = *pstats;
> +
> +	idpf_reset_mbuf_alloc_failed_stats(dev);
> +
> +	return 0;
> +}
> +
>  static int
>  idpf_init_rss(struct idpf_vport *vport)  { @@ -327,6 +408,9 @@
> idpf_dev_start(struct rte_eth_dev *dev)
>  		goto err_vport;
>  	}
> 
> +	if (idpf_dev_stats_reset(dev))
> +		PMD_DRV_LOG(ERR, "Failed to reset stats");
> +
>  	vport->stopped = 0;
> 
>  	return 0;
> @@ -606,6 +690,8 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
>  	.tx_queue_release		= idpf_dev_tx_queue_release,
>  	.mtu_set			= idpf_dev_mtu_set,
>  	.dev_supported_ptypes_get	= idpf_dev_supported_ptypes_get,
> +	.stats_get			= idpf_dev_stats_get,
> +	.stats_reset			= idpf_dev_stats_reset,
>  };
> 
>  static uint16_t
> --
> 2.25.1


  reply	other threads:[~2023-02-08  2:00 UTC|newest]

Thread overview: 63+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-12-16  9:36 [PATCH 0/7] add idpf pmd enhancement features Mingxia Liu
2022-12-16  9:37 ` [PATCH 1/7] common/idpf: add hw statistics Mingxia Liu
2022-12-16  9:37 ` [PATCH 2/7] common/idpf: add RSS set/get ops Mingxia Liu
2022-12-16  9:37 ` [PATCH 3/7] common/idpf: support single q scatter RX datapath Mingxia Liu
2022-12-16  9:37 ` [PATCH 4/7] common/idpf: add rss_offload hash in singleq rx Mingxia Liu
2022-12-16  9:37 ` [PATCH 5/7] common/idpf: add alarm to support handle vchnl message Mingxia Liu
2022-12-16  9:37 ` [PATCH 6/7] common/idpf: add xstats ops Mingxia Liu
2022-12-16  9:37 ` [PATCH 7/7] common/idpf: update mbuf_alloc_failed multi-thread process Mingxia Liu
2023-01-11  7:15 ` [PATCH 0/6] add idpf pmd enhancement features Mingxia Liu
2023-01-11  7:15   ` [PATCH v2 1/6] common/idpf: add hw statistics Mingxia Liu
2023-01-11  7:15   ` [PATCH v2 2/6] common/idpf: add RSS set/get ops Mingxia Liu
2023-01-11  7:15   ` [PATCH v2 3/6] common/idpf: support single q scatter RX datapath Mingxia Liu
2023-01-11  7:15   ` [PATCH v2 4/6] common/idpf: add rss_offload hash in singleq rx Mingxia Liu
2023-01-11  7:15   ` [PATCH v2 5/6] common/idpf: add alarm to support handle vchnl message Mingxia Liu
2023-01-11  7:15   ` [PATCH v2 6/6] common/idpf: add xstats ops Mingxia Liu
2023-01-18  7:14   ` [PATCH v3 0/6] add idpf pmd enhancement features Mingxia Liu
2023-01-18  7:14     ` [PATCH v3 1/6] common/idpf: add hw statistics Mingxia Liu
2023-02-01  8:48       ` Wu, Jingjing
2023-02-01 12:34         ` Liu, Mingxia
2023-01-18  7:14     ` [PATCH v3 2/6] common/idpf: add RSS set/get ops Mingxia Liu
2023-02-02  3:28       ` Wu, Jingjing
2023-02-07  3:10         ` Liu, Mingxia
2023-01-18  7:14     ` [PATCH v3 3/6] common/idpf: support single q scatter RX datapath Mingxia Liu
2023-02-02  3:45       ` Wu, Jingjing
2023-02-02  7:19         ` Liu, Mingxia
2023-01-18  7:14     ` [PATCH v3 4/6] common/idpf: add rss_offload hash in singleq rx Mingxia Liu
2023-01-18  7:14     ` [PATCH v3 5/6] common/idpf: add alarm to support handle vchnl message Mingxia Liu
2023-02-02  4:23       ` Wu, Jingjing
2023-02-02  7:39         ` Liu, Mingxia
2023-02-02  8:46           ` Wu, Jingjing
2023-01-18  7:14     ` [PATCH v3 6/6] common/idpf: add xstats ops Mingxia Liu
2023-02-07  9:56     ` [PATCH v4 0/6] add idpf pmd enhancement features Mingxia Liu
2023-02-07  9:56       ` [PATCH v4 1/6] common/idpf: add hw statistics Mingxia Liu
2023-02-07  9:56       ` [PATCH v4 2/6] common/idpf: add RSS set/get ops Mingxia Liu
2023-02-07  9:56       ` [PATCH v4 3/6] common/idpf: support single q scatter RX datapath Mingxia Liu
2023-02-07  9:56       ` [PATCH v4 4/6] common/idpf: add rss_offload hash in singleq rx Mingxia Liu
2023-02-07  9:57       ` [PATCH v4 5/6] common/idpf: add alarm to support handle vchnl message Mingxia Liu
2023-02-07  9:57       ` [PATCH v4 6/6] common/idpf: add xstats ops Mingxia Liu
2023-02-07 10:08       ` [PATCH v4 0/6] add idpf pmd enhancement features Mingxia Liu
2023-02-07 10:08         ` [PATCH v5 1/6] common/idpf: add hw statistics Mingxia Liu
2023-02-07 10:16           ` [PATCH v6 0/6] add idpf pmd enhancement features Mingxia Liu
2023-02-07 10:16             ` [PATCH v6 1/6] common/idpf: add hw statistics Mingxia Liu
2023-02-08  2:00               ` Zhang, Qi Z [this message]
2023-02-08  8:28                 ` Liu, Mingxia
2023-02-07 10:16             ` [PATCH v6 2/6] common/idpf: add RSS set/get ops Mingxia Liu
2023-02-07 10:16             ` [PATCH v6 3/6] common/idpf: support single q scatter RX datapath Mingxia Liu
2023-02-07 10:16             ` [PATCH v6 4/6] common/idpf: add rss_offload hash in singleq rx Mingxia Liu
2023-02-07 10:16             ` [PATCH v6 5/6] common/idpf: add alarm to support handle vchnl message Mingxia Liu
2023-02-07 10:16             ` [PATCH v6 6/6] common/idpf: add xstats ops Mingxia Liu
2023-02-08  0:28             ` [PATCH v6 0/6] add idpf pmd enhancement features Wu, Jingjing
2023-02-08  7:33             ` [PATCH v7 " Mingxia Liu
2023-02-08  7:33               ` [PATCH v7 1/6] net/idpf: add hw statistics Mingxia Liu
2023-02-08  7:33               ` [PATCH v7 2/6] net/idpf: add RSS set/get ops Mingxia Liu
2023-02-08  7:33               ` [PATCH v7 3/6] net/idpf: support single q scatter RX datapath Mingxia Liu
2023-02-08  7:33               ` [PATCH v7 4/6] net/idpf: add rss_offload hash in singleq rx Mingxia Liu
2023-02-08  7:34               ` [PATCH v7 5/6] net/idpf: add alarm to support handle vchnl message Mingxia Liu
2023-02-08  7:34               ` [PATCH v7 6/6] net/idpf: add xstats ops Mingxia Liu
2023-02-08  9:32               ` [PATCH v7 0/6] add idpf pmd enhancement features Zhang, Qi Z
2023-02-07 10:08         ` [PATCH v5 2/6] common/idpf: add RSS set/get ops Mingxia Liu
2023-02-07 10:08         ` [PATCH v5 3/6] common/idpf: support single q scatter RX datapath Mingxia Liu
2023-02-07 10:08         ` [PATCH v5 4/6] common/idpf: add rss_offload hash in singleq rx Mingxia Liu
2023-02-07 10:08         ` [PATCH v5 5/6] common/idpf: add alarm to support handle vchnl message Mingxia Liu
2023-02-07 10:08         ` [PATCH v5 6/6] common/idpf: add xstats ops Mingxia Liu

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=DM4PR11MB59946DF079C4AC6ED6FDD8F9D7D89@DM4PR11MB5994.namprd11.prod.outlook.com \
    --to=qi.z.zhang@intel.com \
    --cc=beilei.xing@intel.com \
    --cc=dev@dpdk.org \
    --cc=jingjing.wu@intel.com \
    --cc=mingxia.liu@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).