DPDK patches and discussions
 help / color / mirror / Atom feed
From: "Deng, KaiwenX" <kaiwenx.deng@intel.com>
To: Ferruh Yigit <ferruh.yigit@amd.com>, "dev@dpdk.org" <dev@dpdk.org>
Cc: "stable@dpdk.org" <stable@dpdk.org>,
	"Yang, Qiming" <qiming.yang@intel.com>,
	"Zhou, YidingX" <yidingx.zhou@intel.com>,
	"Chas Williams" <chas3@att.com>,
	"Min Hu (Connor)" <humin29@huawei.com>,
	"Wu, Jingjing" <jingjing.wu@intel.com>,
	"Xing, Beilei" <beilei.xing@intel.com>,
	Mike Pattrick <mkp@redhat.com>,
	"Zhang, Qi Z" <qi.z.zhang@intel.com>,
	"Doherty, Declan" <declan.doherty@intel.com>,
	"Mrzyglod, Daniel T" <daniel.t.mrzyglod@intel.com>,
	Dapeng Yu <dapengx.yu@intel.com>
Subject: RE: [PATCH v3] net/iavf: fix iavf query stats in intr thread
Date: Wed, 29 Mar 2023 06:41:52 +0000	[thread overview]
Message-ID: <BYAPR11MB2903549816D468A6A836A98B8E899@BYAPR11MB2903.namprd11.prod.outlook.com> (raw)
In-Reply-To: <abcddece-b2b7-c090-010d-56e2d94b6a91@amd.com>



> -----Original Message-----
> From: Ferruh Yigit <ferruh.yigit@amd.com>
> Sent: Monday, March 27, 2023 8:32 PM
> To: Deng, KaiwenX <kaiwenx.deng@intel.com>; dev@dpdk.org
> Cc: stable@dpdk.org; Yang, Qiming <qiming.yang@intel.com>; Zhou, YidingX
> <yidingx.zhou@intel.com>; Chas Williams <chas3@att.com>; Min Hu (Connor)
> <humin29@huawei.com>; Wu, Jingjing <jingjing.wu@intel.com>; Xing, Beilei
> <beilei.xing@intel.com>; Mike Pattrick <mkp@redhat.com>; Zhang, Qi Z
> <qi.z.zhang@intel.com>; Doherty, Declan <declan.doherty@intel.com>;
> Mrzyglod, Daniel T <daniel.t.mrzyglod@intel.com>; Dapeng Yu
> <dapengx.yu@intel.com>
> Subject: Re: [PATCH v3] net/iavf: fix iavf query stats in intr thread
> 
> On 3/27/2023 6:31 AM, Deng, KaiwenX wrote:
> >
> >
> >> -----Original Message-----
> >> From: Ferruh Yigit <ferruh.yigit@amd.com>
> >> Sent: Thursday, March 23, 2023 11:39 PM
> >> To: Deng, KaiwenX <kaiwenx.deng@intel.com>; dev@dpdk.org
> >> Cc: stable@dpdk.org; Yang, Qiming <qiming.yang@intel.com>; Zhou,
> >> YidingX <yidingx.zhou@intel.com>; Chas Williams <chas3@att.com>; Min
> >> Hu (Connor) <humin29@huawei.com>; Wu, Jingjing
> >> <jingjing.wu@intel.com>; Xing, Beilei <beilei.xing@intel.com>; Mike
> >> Pattrick <mkp@redhat.com>; Zhang, Qi Z <qi.z.zhang@intel.com>;
> >> Doherty, Declan <declan.doherty@intel.com>; Mrzyglod, Daniel T
> >> <daniel.t.mrzyglod@intel.com>; Dapeng Yu <dapengx.yu@intel.com>
> >> Subject: Re: [PATCH v3] net/iavf: fix iavf query stats in intr thread
> >>
> >> On 3/22/2023 7:26 AM, Kaiwen Deng wrote:
> >>> When iavf send query-stats command in eal-intr-thread through
> >>> virtual channel, there will be no response received from
> >>> iavf_dev_virtchnl_handler for this command during block and wait.
> >>> Because iavf_dev_virtchnl_handler is also registered in eal-intr-thread.
> >>>
> >>> When vf device is bonded as BONDING_MODE_TLB mode, the slave
> device
> >>> update callback will registered in alarm and called by
> >>> eal-intr-thread, it would also raise the above issue.
> >>>
> >>> This commit add to poll the response for VIRTCHNL_OP_GET_STATS
> when
> >> it
> >>> is called by eal-intr-thread to fix this issue.
> >>>
> >>> Fixes: 91bf37d250aa ("net/iavf: add lock for VF commands")
> >>> Fixes: 22b123a36d07 ("net/avf: initialize PMD")
> >>> Fixes: 7c76a747e68c ("bond: add mode 5")
> >>> Fixes: 435d523112cc ("net/iavf: fix multi-process shared data")
> >>> Fixes: cb5c1b91f76f ("net/iavf: add thread for event callbacks")
> >>
> >>
> >> Hi Kaiwen,
> >>
> >> Above commit already seems trying to address same issue, it creates
> >> "iavf- event-thread" control thread to asyncroniously handle the
> >> interrupts, in non- interrupt context, why it is not working?
> >>
> >> Instead of adding 'rte_thread_is_intr()' checks, can't you make sure
> >> all interrupts handled in control tread?
> >>
> >> And can you please provide a stack trace in commit log, to describe
> >> the issue better?
> > Hi Ferru,
> > Sorry for my late reply, And thanks for your review.
> >
> > The above commit does not fix this issue when we need to get the
> returned data.
> > If we call iavf_query_stats and wait for response statistics in the intr-thread.
> > iavf_handle_virtchnl_msg is also registered in the intr_thread and
> > will not be executed while waiting.
> >
> 
> Got it, since return value is required, API can't be called asyncroniously.
> 
> 
> 
> I think 'rte_thread_is_intr()' checks may cause more trouble for you in long
> term,
> 
> - why 'iavf_query_stats()' is called in the iterrupt thread, can it be prevented?
> 
> - does it make sense to allways poll messages from PF (for simplification)?
> 
Virtual channel commands sometimes need to be registered to alarm 
so that they can be called periodically, and alarm is registered  to be 
called in interrupt threads.

For some commands that do not require a return value, It cannot 
support asynchronous if allways poll messages from PF.
> 
> If answer to both are 'No', I am OK to continue with current proposal if you
> are happy with it.
> 
> 
> > This commit I changed it to polling for replies to commands executed in the
> interrupt thread.
> >
> > main thread                                                                                                interrupt
> thread
> >      |                                                                                                                               |
> >      |                                                                                                                               |
> > iavf_query_stats                                                                                                        |
> > iavf_execute_vf_cmd                                                                                               |
> > iavf_aq_send_msg_to_pf  and wait handle complete                                       |
> >      |                                                                                                                              |
> >      |---------------------------------------------------------------------------------------
> ----->|
> >      |                                                                                                                              |
> >      |
> iavf_handle_virtchnl_msg
> >      |                                                                                                                             |
> >      |<-------------------------------------------------------------------------------------
> -------|
> >      |                                                                                                                             |
> > iavf_execute_vf_cmd get response                                                                      |
> >      |                                                                                                                             |
> >
> > The above is the stack trace for the normal execution of
> > iavf_query_stats in the main thread.
> >
> >  interrupt thread
> >      |
> >      |
> > iavf_query_stats
> > iavf_execute_vf_cmd
> > iavf_aq_send_msg_to_pf wait handle complete(1 sec)
> iavf_execute_vf_cmd
> > timeout
> >      |
> >      |
> > iavf_handle_virtchnl_msg
> >      |
> >
> > The above is the stack trace for the abnormal execution of
> > iavf_query_stats in the interrupt thread.
> >>
> >>> Cc: stable@dpdk.org
> >>>
> >>> Signed-off-by: Kaiwen Deng <kaiwenx.deng@intel.com>
> >>> ---
> >>> Changes since v2:
> >>> - Add to poll the response for VIRTCHNL_OP_GET_STATS.
> >>>
> >>> Changes since v1:
> >>> - Add lock to avoid race condition.
> >>> ---
> >>> ---
> >>>  drivers/net/bonding/rte_eth_bond_pmd.c |  7 ++-
> >>>  drivers/net/iavf/iavf_ethdev.c         |  5 +-
> >>>  drivers/net/iavf/iavf_vchnl.c          | 71 ++++++++++++++++++--------
> >>>  3 files changed, 58 insertions(+), 25 deletions(-)
> >>>
> >>> diff --git a/drivers/net/bonding/rte_eth_bond_pmd.c
> >>> b/drivers/net/bonding/rte_eth_bond_pmd.c
> >>> index f0c4f7d26b..edce621496 100644
> >>> --- a/drivers/net/bonding/rte_eth_bond_pmd.c
> >>> +++ b/drivers/net/bonding/rte_eth_bond_pmd.c
> >>> @@ -894,6 +894,7 @@ bond_ethdev_update_tlb_slave_cb(void *arg)
> >>>  	uint8_t update_stats = 0;
> >>>  	uint16_t slave_id;
> >>>  	uint16_t i;
> >>> +	int ret;
> >>>
> >>>  	internals->slave_update_idx++;
> >>>
> >>> @@ -903,7 +904,10 @@ bond_ethdev_update_tlb_slave_cb(void *arg)
> >>>
> >>>  	for (i = 0; i < internals->active_slave_count; i++) {
> >>>  		slave_id = internals->active_slaves[i];
> >>> -		rte_eth_stats_get(slave_id, &slave_stats);
> >>> +		ret = rte_eth_stats_get(slave_id, &slave_stats);
> >>> +		if (ret)
> >>> +			goto OUT;
> >>> +
> >>>  		tx_bytes = slave_stats.obytes - tlb_last_obytets[slave_id];
> >>>  		bandwidth_left(slave_id, tx_bytes,
> >>>  				internals->slave_update_idx, &bwg_array[i]);
> >> @@ -922,6 +926,7 @@
> >>> bond_ethdev_update_tlb_slave_cb(void *arg)
> >>>  	for (i = 0; i < slave_count; i++)
> >>>  		internals->tlb_slaves_order[i] = bwg_array[i].slave;
> >>>
> >>> +OUT:
> >>>  	rte_eal_alarm_set(REORDER_PERIOD_MS * 1000,
> >> bond_ethdev_update_tlb_slave_cb,
> >>>  			(struct bond_dev_private *)internals);  } diff --git
> >>> a/drivers/net/iavf/iavf_ethdev.c b/drivers/net/iavf/iavf_ethdev.c
> >>> index 3196210f2c..d6e1f1a7f4 100644
> >>> --- a/drivers/net/iavf/iavf_ethdev.c
> >>> +++ b/drivers/net/iavf/iavf_ethdev.c
> >>> @@ -2607,6 +2607,9 @@ iavf_dev_init(struct rte_eth_dev *eth_dev)
> >>>  	adapter->dev_data = eth_dev->data;
> >>>  	adapter->stopped = 1;
> >>>
> >>> +	if (iavf_dev_event_handler_init())
> >>> +		goto init_vf_err;
> >>> +
> >>>  	if (iavf_init_vf(eth_dev) != 0) {
> >>>  		PMD_INIT_LOG(ERR, "Init vf failed");
> >>>  		return -1;
> >>> @@ -2634,8 +2637,6 @@ iavf_dev_init(struct rte_eth_dev *eth_dev)
> >>>  	rte_ether_addr_copy((struct rte_ether_addr *)hw->mac.addr,
> >>>  			&eth_dev->data->mac_addrs[0]);
> >>>
> >>> -	if (iavf_dev_event_handler_init())
> >>> -		goto init_vf_err;
> >>>
> >>>  	if (vf->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_WB_ON_ITR)
> >> {
> >>>  		/* register callback func to eal lib */ diff --git
> >>> a/drivers/net/iavf/iavf_vchnl.c b/drivers/net/iavf/iavf_vchnl.c
> >>> index 9adaadb173..aeffb07cca 100644
> >>> --- a/drivers/net/iavf/iavf_vchnl.c
> >>> +++ b/drivers/net/iavf/iavf_vchnl.c
> >>> @@ -256,6 +256,7 @@ iavf_read_msg_from_pf(struct iavf_adapter
> >> *adapter, uint16_t buf_len,
> >>>  				vf->link_speed =
> >> iavf_convert_link_speed(speed);
> >>>  			}
> >>>  			iavf_dev_link_update(vf->eth_dev, 0);
> >>> +			iavf_dev_event_post(vf->eth_dev,
> >> RTE_ETH_EVENT_INTR_LSC, NULL, 0);
> >>>  			PMD_DRV_LOG(INFO, "Link status update:%s",
> >>>  					vf->link_up ? "up" : "down");
> >>>  			break;
> >>> @@ -368,28 +369,48 @@ iavf_execute_vf_cmd(struct iavf_adapter
> >> *adapter, struct iavf_cmd_info *args,
> >>>  		_clear_cmd(vf);
> >>>  		break;
> >>>  	default:
> >>> -		/* For other virtchnl ops in running time,
> >>> -		 * wait for the cmd done flag.
> >>> -		 */
> >>> -		do {
> >>> -			if (vf->pend_cmd == VIRTCHNL_OP_UNKNOWN)
> >>> -				break;
> >>> -			iavf_msec_delay(ASQ_DELAY_MS);
> >>> -			/* If don't read msg or read sys event, continue */
> >>> -		} while (i++ < MAX_TRY_TIMES);
> >>> -
> >>> -		if (i >= MAX_TRY_TIMES) {
> >>> -			PMD_DRV_LOG(ERR, "No response for cmd %d",
> >> args->ops);
> >>> +		if (rte_thread_is_intr()) {
> >>> +			/* For virtchnl ops were executed in eal_intr_thread,
> >>> +			 * need to poll the response.
> >>> +			 */
> >>> +			do {
> >>> +				result = iavf_read_msg_from_pf(adapter,
> >> args->out_size,
> >>> +							args->out_buffer);
> >>> +				if (result == IAVF_MSG_CMD)
> >>> +					break;
> >>> +				iavf_msec_delay(ASQ_DELAY_MS);
> >>> +			} while (i++ < MAX_TRY_TIMES);
> >>> +			if (i >= MAX_TRY_TIMES ||
> >>> +				vf->cmd_retval !=
> >> VIRTCHNL_STATUS_SUCCESS) {
> >>> +				err = -1;
> >>> +				PMD_DRV_LOG(ERR, "No response or return
> >> failure (%d)"
> >>> +						" for cmd %d", vf-
> >>> cmd_retval, args->ops);
> >>> +			}
> >>>  			_clear_cmd(vf);
> >>> -			err = -EIO;
> >>> -		} else if (vf->cmd_retval ==
> >>> -			   VIRTCHNL_STATUS_ERR_NOT_SUPPORTED) {
> >>> -			PMD_DRV_LOG(ERR, "Cmd %d not supported", args-
> >>> ops);
> >>> -			err = -ENOTSUP;
> >>> -		} else if (vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
> >>> -			PMD_DRV_LOG(ERR, "Return failure %d for cmd %d",
> >>> -				    vf->cmd_retval, args->ops);
> >>> -			err = -EINVAL;
> >>> +		} else {
> >>> +			/* For other virtchnl ops in running time,
> >>> +			 * wait for the cmd done flag.
> >>> +			 */
> >>> +			do {
> >>> +				if (vf->pend_cmd ==
> >> VIRTCHNL_OP_UNKNOWN)
> >>> +					break;
> >>> +				iavf_msec_delay(ASQ_DELAY_MS);
> >>> +				/* If don't read msg or read sys event,
> >> continue */
> >>> +			} while (i++ < MAX_TRY_TIMES);
> >>> +
> >>> +			if (i >= MAX_TRY_TIMES) {
> >>> +				PMD_DRV_LOG(ERR, "No response for
> >> cmd %d", args->ops);
> >>> +				_clear_cmd(vf);
> >>> +				err = -EIO;
> >>> +			} else if (vf->cmd_retval ==
> >>> +				VIRTCHNL_STATUS_ERR_NOT_SUPPORTED) {
> >>> +				PMD_DRV_LOG(ERR, "Cmd %d not
> >> supported", args->ops);
> >>> +				err = -ENOTSUP;
> >>> +			} else if (vf->cmd_retval !=
> >> VIRTCHNL_STATUS_SUCCESS) {
> >>> +				PMD_DRV_LOG(ERR, "Return failure %d for
> >> cmd %d",
> >>> +						vf->cmd_retval, args->ops);
> >>> +				err = -EINVAL;
> >>> +			}
> >>>  		}
> >>>  		break;
> >>>  	}
> >>> @@ -403,8 +424,14 @@ iavf_execute_vf_cmd_safe(struct iavf_adapter
> >>> *adapter,  {
> >>>  	struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
> >>>  	int ret;
> >>> +	int is_intr_thread = rte_thread_is_intr();
> >>>
> >>> -	rte_spinlock_lock(&vf->aq_lock);
> >>> +	if (is_intr_thread) {
> >>> +		if (!rte_spinlock_trylock(&vf->aq_lock))
> >>> +			return -EIO;
> >>> +	} else {
> >>> +		rte_spinlock_lock(&vf->aq_lock);
> >>> +	}
> >>>  	ret = iavf_execute_vf_cmd(adapter, args, async);
> >>>  	rte_spinlock_unlock(&vf->aq_lock);
> >>>
> >


  parent reply	other threads:[~2023-03-29  6:42 UTC|newest]

Thread overview: 19+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-02-22  4:40 [PATCH] " Kaiwen Deng
2023-02-27  0:56 ` Zhang, Qi Z
2023-02-27  6:02   ` Deng, KaiwenX
2023-03-07  3:27   ` Deng, KaiwenX
2023-03-07  2:55 ` [PATCH v2] " Kaiwen Deng
2023-03-15 13:40   ` Zhang, Qi Z
2023-03-22  7:26   ` [PATCH v3] " Kaiwen Deng
2023-03-23 15:39     ` Ferruh Yigit
2023-03-27  5:31       ` Deng, KaiwenX
2023-03-27 12:31         ` Ferruh Yigit
2023-03-27 12:37           ` Ferruh Yigit
2023-03-29  7:53             ` Deng, KaiwenX
2023-05-05  2:31             ` Deng, KaiwenX
2023-05-23  1:45               ` Deng, KaiwenX
2023-05-23 10:26                 ` Ferruh Yigit
2023-03-29  6:41           ` Deng, KaiwenX [this message]
2023-06-06  5:41     ` Jiale, SongX
2023-06-07  2:03     ` [PATCH v4] " Kaiwen Deng
2023-06-07  4:01       ` Zhang, Qi Z

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=BYAPR11MB2903549816D468A6A836A98B8E899@BYAPR11MB2903.namprd11.prod.outlook.com \
    --to=kaiwenx.deng@intel.com \
    --cc=beilei.xing@intel.com \
    --cc=chas3@att.com \
    --cc=daniel.t.mrzyglod@intel.com \
    --cc=dapengx.yu@intel.com \
    --cc=declan.doherty@intel.com \
    --cc=dev@dpdk.org \
    --cc=ferruh.yigit@amd.com \
    --cc=humin29@huawei.com \
    --cc=jingjing.wu@intel.com \
    --cc=mkp@redhat.com \
    --cc=qi.z.zhang@intel.com \
    --cc=qiming.yang@intel.com \
    --cc=stable@dpdk.org \
    --cc=yidingx.zhou@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).