patches for DPDK stable branches
 help / color / mirror / Atom feed
From: Raslan Darawsheh <rasland@nvidia.com>
To: Alexander Kozyrev <akozyrev@nvidia.com>, "dev@dpdk.org" <dev@dpdk.org>
Cc: "stable@dpdk.org" <stable@dpdk.org>,
	Slava Ovsiienko <viacheslavo@nvidia.com>
Subject: Re: [dpdk-stable] [PATCH] net/mlx5: fix Rx queue count calculation
Date: Thu, 1 Oct 2020 13:28:34 +0000
Message-ID: <DM6PR12MB274898FEF4FE6AE25E44DB81CF300@DM6PR12MB2748.namprd12.prod.outlook.com> (raw)
In-Reply-To: <20200929183623.22685-1-akozyrev@nvidia.com>

Hi,

> -----Original Message-----
> From: Alexander Kozyrev <akozyrev@nvidia.com>
> Sent: Tuesday, September 29, 2020 9:36 PM
> To: dev@dpdk.org
> Cc: stable@dpdk.org; Raslan Darawsheh <rasland@nvidia.com>; Slava
> Ovsiienko <viacheslavo@nvidia.com>
> Subject: [PATCH] net/mlx5: fix Rx queue count calculation
> 
> There are a few discrepancies in the Rx queue count calculation.
> 
> The wrong index is used to calculate the number of used descriptors
> in an Rx queue in case of the compressed CQE processing. The global
> CQ index is used while we really need an internal index in a single
> compressed session to get the right number of elements processed.
> 
> The total number of CQs should be used instead of the number of mbufs
> to find out about the maximum number of Rx descriptors. These numbers
> are not equal for the Multi-Packet Rx queue.
> 
> Allow the Rx queue count calculation for all possible Rx bursts since
> CQ handling is the same for regular, vectorized, and multi-packet Rx
> queues.
> 
> Fixes: 26f0488344 ("net/mlx5: support Rx queue count API")
> Cc: stable@dpdk.org
> 
> Signed-off-by: Alexander Kozyrev <akozyrev@nvidia.com>
> Acked-by: Viacheslav Ovsiienko <viacheslavo@nvidia.com>
> ---
>  drivers/net/mlx5/mlx5_rxtx.c | 27 ++++++++++++---------------
>  1 file changed, 12 insertions(+), 15 deletions(-)
> 
> diff --git a/drivers/net/mlx5/mlx5_rxtx.c b/drivers/net/mlx5/mlx5_rxtx.c
> index 101555ed03..4755980d5b 100644
> --- a/drivers/net/mlx5/mlx5_rxtx.c
> +++ b/drivers/net/mlx5/mlx5_rxtx.c
> @@ -465,19 +465,11 @@ rx_queue_count(struct mlx5_rxq_data *rxq)
>  {
>  	struct rxq_zip *zip = &rxq->zip;
>  	volatile struct mlx5_cqe *cqe;
> +	unsigned int cq_ci = rxq->cq_ci;
>  	const unsigned int cqe_n = (1 << rxq->cqe_n);
>  	const unsigned int cqe_cnt = cqe_n - 1;
> -	unsigned int cq_ci;
> -	unsigned int used;
> +	unsigned int used = 0;
> 
> -	/* if we are processing a compressed cqe */
> -	if (zip->ai) {
> -		used = zip->cqe_cnt - zip->ca;
> -		cq_ci = zip->cq_ci;
> -	} else {
> -		used = 0;
> -		cq_ci = rxq->cq_ci;
> -	}
>  	cqe = &(*rxq->cqes)[cq_ci & cqe_cnt];
>  	while (check_cqe(cqe, cqe_n, cq_ci) !=
> MLX5_CQE_STATUS_HW_OWN) {
>  		int8_t op_own;
> @@ -485,14 +477,17 @@ rx_queue_count(struct mlx5_rxq_data *rxq)
> 
>  		op_own = cqe->op_own;
>  		if (MLX5_CQE_FORMAT(op_own) == MLX5_COMPRESSED)
> -			n = rte_be_to_cpu_32(cqe->byte_cnt);
> +			if (unlikely(zip->ai))
> +				n = zip->cqe_cnt - zip->ai;
> +			else
> +				n = rte_be_to_cpu_32(cqe->byte_cnt);
>  		else
>  			n = 1;
>  		cq_ci += n;
>  		used += n;
>  		cqe = &(*rxq->cqes)[cq_ci & cqe_cnt];
>  	}
> -	used = RTE_MIN(used, (1U << rxq->elts_n) - 1);
> +	used = RTE_MIN(used, cqe_n);
>  	return used;
>  }
> 
> @@ -515,11 +510,12 @@ mlx5_rx_descriptor_status(void *rx_queue,
> uint16_t offset)
>  			container_of(rxq, struct mlx5_rxq_ctrl, rxq);
>  	struct rte_eth_dev *dev = ETH_DEV(rxq_ctrl->priv);
> 
> -	if (dev->rx_pkt_burst != mlx5_rx_burst) {
> +	if (dev->rx_pkt_burst == NULL ||
> +	    dev->rx_pkt_burst == removed_rx_burst) {
>  		rte_errno = ENOTSUP;
>  		return -rte_errno;
>  	}
> -	if (offset >= (1 << rxq->elts_n)) {
> +	if (offset >= (1 << rxq->cqe_n)) {
>  		rte_errno = EINVAL;
>  		return -rte_errno;
>  	}
> @@ -630,7 +626,8 @@ mlx5_rx_queue_count(struct rte_eth_dev *dev,
> uint16_t rx_queue_id)
>  	struct mlx5_priv *priv = dev->data->dev_private;
>  	struct mlx5_rxq_data *rxq;
> 
> -	if (dev->rx_pkt_burst != mlx5_rx_burst) {
> +	if (dev->rx_pkt_burst == NULL ||
> +	    dev->rx_pkt_burst == removed_rx_burst) {
>  		rte_errno = ENOTSUP;
>  		return -rte_errno;
>  	}
> --
> 2.24.1

Patch applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh

      reply	other threads:[~2020-10-01 13:28 UTC|newest]

Thread overview: 2+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-09-29 18:36 Alexander Kozyrev
2020-10-01 13:28 ` Raslan Darawsheh [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=DM6PR12MB274898FEF4FE6AE25E44DB81CF300@DM6PR12MB2748.namprd12.prod.outlook.com \
    --to=rasland@nvidia.com \
    --cc=akozyrev@nvidia.com \
    --cc=dev@dpdk.org \
    --cc=stable@dpdk.org \
    --cc=viacheslavo@nvidia.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

patches for DPDK stable branches

This inbox may be cloned and mirrored by anyone:

	git clone --mirror https://inbox.dpdk.org/stable/0 stable/git/0.git

	# If you have public-inbox 1.1+ installed, you may
	# initialize and index your mirror using the following commands:
	public-inbox-init -V2 stable stable/ https://inbox.dpdk.org/stable \
		stable@dpdk.org
	public-inbox-index stable

Example config snippet for mirrors.
Newsgroup available over NNTP:
	nntp://inbox.dpdk.org/inbox.dpdk.stable


AGPL code for this site: git clone https://public-inbox.org/public-inbox.git