From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 11C3AA04A5; Wed, 17 Jun 2020 14:29:46 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id E64FC1150; Wed, 17 Jun 2020 14:29:45 +0200 (CEST) Received: from us-smtp-1.mimecast.com (us-smtp-delivery-1.mimecast.com [205.139.110.120]) by dpdk.org (Postfix) with ESMTP id 7F78B100C for ; Wed, 17 Jun 2020 14:29:44 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1592396984; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:autocrypt:autocrypt; bh=irUb1cZAqEnbd0xOPYxfKBLsFE/6GhwOPbSy5ppS3h4=; b=J7qTDiZhrdZNwBM5B3UQAwAGrOsFsPL3n4FTls8bjgHimyPt6su1I0hRAJiQkOyczoq+gr 14ACTsmPFAdC2XDUADiYJpIxfjSq6eIiSy/RPVLt+0uhNMmXcKwvy5//7eKyXVIDrSWeuC rpDIka/y18WVk5NKDomHPDUpDslrYEs= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-508-pwt3VO18OW69JpOAnPFHeg-1; Wed, 17 Jun 2020 08:29:40 -0400 X-MC-Unique: pwt3VO18OW69JpOAnPFHeg-1 Received: from smtp.corp.redhat.com (int-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.12]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 59399134CF; Wed, 17 Jun 2020 12:29:39 +0000 (UTC) Received: from [10.36.110.7] (unknown [10.36.110.7]) by smtp.corp.redhat.com (Postfix) with ESMTPS id D4F4860BE2; Wed, 17 Jun 2020 12:29:36 +0000 (UTC) To: Matan Azrad , "xiaolong.ye@intel.com" , Shahaf Shuler , "amorenoz@redhat.com" , "xiao.w.wang@intel.com" , Slava Ovsiienko , "dev@dpdk.org" Cc: "jasowang@redhat.com" , "lulu@redhat.com" References: <20200514080218.1435344-1-maxime.coquelin@redhat.com> <20200514080218.1435344-10-maxime.coquelin@redhat.com> <0216165f-aedd-06c7-5a90-2cd0d238b143@redhat.com> <97528794-aa14-37da-a5bb-1e1d46e9127a@redhat.com> <64c59e91-79ef-bc55-bc64-1995cfb03c84@redhat.com> <095af180-bb29-e890-5197-8bfbb8ea4c41@redhat.com> From: Maxime Coquelin Autocrypt: addr=maxime.coquelin@redhat.com; keydata= mQINBFOEQQIBEADjNLYZZqghYuWv1nlLisptPJp+TSxE/KuP7x47e1Gr5/oMDJ1OKNG8rlNg kLgBQUki3voWhUbMb69ybqdMUHOl21DGCj0BTU3lXwapYXOAnsh8q6RRM+deUpasyT+Jvf3a gU35dgZcomRh5HPmKMU4KfeA38cVUebsFec1HuJAWzOb/UdtQkYyZR4rbzw8SbsOemtMtwOx YdXodneQD7KuRU9IhJKiEfipwqk2pufm2VSGl570l5ANyWMA/XADNhcEXhpkZ1Iwj3TWO7XR uH4xfvPl8nBsLo/EbEI7fbuUULcAnHfowQslPUm6/yaGv6cT5160SPXT1t8U9QDO6aTSo59N jH519JS8oeKZB1n1eLDslCfBpIpWkW8ZElGkOGWAN0vmpLfdyiqBNNyS3eGAfMkJ6b1A24un /TKc6j2QxM0QK4yZGfAxDxtvDv9LFXec8ENJYsbiR6WHRHq7wXl/n8guyh5AuBNQ3LIK44x0 KjGXP1FJkUhUuruGyZsMrDLBRHYi+hhDAgRjqHgoXi5XGETA1PAiNBNnQwMf5aubt+mE2Q5r qLNTgwSo2dpTU3+mJ3y3KlsIfoaxYI7XNsPRXGnZi4hbxmeb2NSXgdCXhX3nELUNYm4ArKBP LugOIT/zRwk0H0+RVwL2zHdMO1Tht1UOFGfOZpvuBF60jhMzbQARAQABtCxNYXhpbWUgQ29x dWVsaW4gPG1heGltZS5jb3F1ZWxpbkByZWRoYXQuY29tPokCOAQTAQIAIgUCV3u/5QIbAwYL CQgHAwIGFQgCCQoLBBYCAwECHgECF4AACgkQyjiNKEaHD4ma2g/+P+Hg9WkONPaY1J4AR7Uf kBneosS4NO3CRy0x4WYmUSLYMLx1I3VH6SVjqZ6uBoYy6Fs6TbF6SHNc7QbB6Qjo3neqnQR1 71Ua1MFvIob8vUEl3jAR/+oaE1UJKrxjWztpppQTukIk4oJOmXbL0nj3d8dA2QgHdTyttZ1H xzZJWWz6vqxCrUqHU7RSH9iWg9R2iuTzii4/vk1oi4Qz7y/q8ONOq6ffOy/t5xSZOMtZCspu Mll2Szzpc/trFO0pLH4LZZfz/nXh2uuUbk8qRIJBIjZH3ZQfACffgfNefLe2PxMqJZ8mFJXc RQO0ONZvwoOoHL6CcnFZp2i0P5ddduzwPdGsPq1bnIXnZqJSl3dUfh3xG5ArkliZ/++zGF1O wvpGvpIuOgLqjyCNNRoR7cP7y8F24gWE/HqJBXs1qzdj/5Hr68NVPV1Tu/l2D1KMOcL5sOrz 2jLXauqDWn1Okk9hkXAP7+0Cmi6QwAPuBT3i6t2e8UdtMtCE4sLesWS/XohnSFFscZR6Vaf3 gKdWiJ/fW64L6b9gjkWtHd4jAJBAIAx1JM6xcA1xMbAFsD8gA2oDBWogHGYcScY/4riDNKXi lw92d6IEHnSf6y7KJCKq8F+Jrj2BwRJiFKTJ6ChbOpyyR6nGTckzsLgday2KxBIyuh4w+hMq TGDSp2rmWGJjASq5Ag0EVPSbkwEQAMkaNc084Qvql+XW+wcUIY+Dn9A2D1gMr2BVwdSfVDN7 0ZYxo9PvSkzh6eQmnZNQtl8WSHl3VG3IEDQzsMQ2ftZn2sxjcCadexrQQv3Lu60Tgj7YVYRM H+fLYt9W5YuWduJ+FPLbjIKynBf6JCRMWr75QAOhhhaI0tsie3eDsKQBA0w7WCuPiZiheJaL 4MDe9hcH4rM3ybnRW7K2dLszWNhHVoYSFlZGYh+MGpuODeQKDS035+4H2rEWgg+iaOwqD7bg CQXwTZ1kSrm8NxIRVD3MBtzp9SZdUHLfmBl/tLVwDSZvHZhhvJHC6Lj6VL4jPXF5K2+Nn/Su CQmEBisOmwnXZhhu8ulAZ7S2tcl94DCo60ReheDoPBU8PR2TLg8rS5f9w6mLYarvQWL7cDtT d2eX3Z6TggfNINr/RTFrrAd7NHl5h3OnlXj7PQ1f0kfufduOeCQddJN4gsQfxo/qvWVB7PaE 1WTIggPmWS+Xxijk7xG6x9McTdmGhYaPZBpAxewK8ypl5+yubVsE9yOOhKMVo9DoVCjh5To5 aph7CQWfQsV7cd9PfSJjI2lXI0dhEXhQ7lRCFpf3V3mD6CyrhpcJpV6XVGjxJvGUale7+IOp sQIbPKUHpB2F+ZUPWds9yyVxGwDxD8WLqKKy0WLIjkkSsOb9UBNzgRyzrEC9lgQ/ABEBAAGJ Ah8EGAECAAkFAlT0m5MCGwwACgkQyjiNKEaHD4nU8hAAtt0xFJAy0sOWqSmyxTc7FUcX+pbD KVyPlpl6urKKMk1XtVMUPuae/+UwvIt0urk1mXi6DnrAN50TmQqvdjcPTQ6uoZ8zjgGeASZg jj0/bJGhgUr9U7oG7Hh2F8vzpOqZrdd65MRkxmc7bWj1k81tOU2woR/Gy8xLzi0k0KUa8ueB iYOcZcIGTcs9CssVwQjYaXRoeT65LJnTxYZif2pfNxfINFzCGw42s3EtZFteczClKcVSJ1+L +QUY/J24x0/ocQX/M1PwtZbB4c/2Pg/t5FS+s6UB1Ce08xsJDcwyOPIH6O3tccZuriHgvqKP yKz/Ble76+NFlTK1mpUlfM7PVhD5XzrDUEHWRTeTJSvJ8TIPL4uyfzhjHhlkCU0mw7Pscyxn DE8G0UYMEaNgaZap8dcGMYH/96EfE5s/nTX0M6MXV0yots7U2BDb4soLCxLOJz4tAFDtNFtA wLBhXRSvWhdBJZiig/9CG3dXmKfi2H+wdUCSvEFHRpgo7GK8/Kh3vGhgKmnnxhl8ACBaGy9n fxjSxjSO6rj4/MeenmlJw1yebzkX8ZmaSi8BHe+n6jTGEFNrbiOdWpJgc5yHIZZnwXaW54QT UhhSjDL1rV2B4F28w30jYmlRmm2RdN7iCZfbyP3dvFQTzQ4ySquuPkIGcOOHrvZzxbRjzMx1 Mwqu3GQ= Message-ID: <7cdc8505-a667-81a3-b8ed-22dd34b958f7@redhat.com> Date: Wed, 17 Jun 2020 14:29:34 +0200 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.8.0 MIME-Version: 1.0 In-Reply-To: Content-Language: en-US X-Scanned-By: MIMEDefang 2.79 on 10.5.11.12 X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 8bit Subject: Re: [dpdk-dev] [PATCH 9/9] vhost: only use vDPA config workaround if needed X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" On 6/17/20 1:04 PM, Matan Azrad wrote: >>> Don’t you think that only enabled queues must be fully initialized when >> their status is changed from disabled to enabled? >>> So, you can assume that disabled queues can stay "not fully initialized"... >> >> That may work but might not be following the Virtio spec as with 1.0 we >> shouldn't process the rings before DRIVER_OK is set (but we cannot be sure >> we follow it anyway without SET_STATUS support). >> >> I propose to cook a patch doing the following: >> 1. virtio_is_ready() will only ensure the first queue pair is ready (i.e. enabled >> and configured). Meaning that app's new_device callback and vDPA drivers >> dev_conf callback will be called with only the first queue pair configured and >> enabled. >> >> 2. Before handling a new vhost-user request, it saves the ready status for >> every queue pair. >> >> 3. Same handling of the requests, except that we won't notify the vdpa >> driver and the application of vring state changes in the >> VHOST_USER_SET_VRING_ENABLE handler. >> >> 4. Once the Vhost-user request is handled, it compares the new ready status >> foe every queues with the old one and send queue state event changes >> accordingly. > > Looks very nice to me. Cool! > More points: > By this method some queues may be configured by the set_vring_state operation so the next calls are expected to be called for each queue by the driver from the set_vring_state callback : > 1. rte_vhost_enable_guest_notification > This one takes datapath lock so we need to be sure that datapath lock is not locked on this queue from the same caller thread (maybe to not takes datapath locks when vdpa is configured at all). Good point, I agree we shouldn't need to use the access lock when vdpa is configured. We may want to document that all the control path is assumed to be single thread though. > 2. rte_vhost_host_notifier_ctrl > This function API is per device and not per queue, maybe we need to change this function to be per queue (add new for now and deprecate the old one in 20.11). This one is still experimental, so no issue in reworking the API to make it per queue without deprecation notice. > 3. Need to be sure that if ready queue configuration is changed after dev_conf, we should notify it to the driver. (maybe by set_vring_state(disabl) and set_vring_state(enable)). Agree, I'm not sure yet if we should just toggle set_vring_state as you proposes, or if we should have a new callback for this. >> It is likely to need changes in the .dev_conf and .set_vring_state >> implementations by the drivers. > > Yes, for Mellanox it is very easy change. > Intel? > > >>> >>>> With VHOST_USER_SET_STATUS, we will be able to handle this properly, >>>> as the backend can be sure the guest won't initialize more queues as >>>> soon as DRIVER_OK Virtio status bit is set. In my v2, I can add one >>>> patch to handle this case properly, by "destorying" queues metadata >>>> as soon as DRIVER_OK is received. >>>> >>>> Note that it was the exact reason why I first tried to add support >>>> for VHOST_USER_SET_STATUS more than two years ago...: >>>> https://eur03.safelinks.protection.outlook.com/?url=https%3A%2F%2Flis >>>> ts.g >>>> nu.org%2Farchive%2Fhtml%2Fqemu-devel%2F2018- >>>> >> 02%2Fmsg04560.html&data=02%7C01%7Cmatan%40mellanox.com%7C >>>> >> bed5d361fbff47ab766008d80c99cc53%7Ca652971c7d2e4d9ba6a4d149256f461 >>>> >> b%7C0%7C0%7C637273201984684513&sdata=KGJjdEtEN54duNu41rhBIw >>>> o4tmdWn6QD4yvdR3zeLI8%3D&reserved=0 >>>> >>>> What do you think? >>> >>> Yes, I agree it may be solved by VHOST_USER_SET_STATUS (and probably a >>> lot of other issues), But I think we need support also legacy QEMU versions >> if we can... >> >> I think the SET_STATUS support is important to be compliant with the Virtio >> specifictation. >> >>> Don't you think so? > > Yes, I agree. > >> >> We can try that. >> I will try to cook something this week, but it will require validation with OVS >> to be sure we don't break multiqueue. I will send it as RFC, and count on you >> to try it with your mlx5 vDPA driver. >> >> Does it work for you? (note I'll be on vacation from July 1st to 17th) > > Sure, > Do you have capacity to do it this week? > I can help..... That would be welcome, as I initially planned to spend time on reviewing & merging patches this week. Thanks, Maxime > Matan > > >> >> Thanks, >> Maxime >> >>>> Regards, >>>> Maxime >>> >