From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 0896AA00BE; Thu, 16 Jun 2022 11:02:18 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id ABB0642BCA; Thu, 16 Jun 2022 11:02:17 +0200 (CEST) Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by mails.dpdk.org (Postfix) with ESMTP id EE89E4281E for ; Thu, 16 Jun 2022 11:02:16 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1655370136; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=4oKYX/tlyXRLpjMWXpMZ1/o9T0J2QPZU3r32ICdRrMU=; b=RbsFChEqE9zimBZhB3HDTx8s58p+IXnjMC2Tvus8EtERLB9fkE/KYf1JETJLOspzB20RCJ pv/iEIsgYkkJCw2+Kl+g2SbgcEwMGpovYRkoSjlb3uQHV9dlpFTmIWFgvMjxm/SJWyq5Ch wZqi03WvTTaQ3EDQexoMA1dYpmbU3U0= Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-348-Bfc9FplYNoaGe1jCRS2BnA-1; Thu, 16 Jun 2022 05:02:10 -0400 X-MC-Unique: Bfc9FplYNoaGe1jCRS2BnA-1 Received: from smtp.corp.redhat.com (int-mx05.intmail.prod.int.rdu2.redhat.com [10.11.54.5]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id 4D745811E75; Thu, 16 Jun 2022 09:02:10 +0000 (UTC) Received: from [10.39.208.34] (unknown [10.39.208.34]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 42CE110725; Thu, 16 Jun 2022 09:02:08 +0000 (UTC) Message-ID: Date: Thu, 16 Jun 2022 11:02:06 +0200 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:91.0) Gecko/20100101 Thunderbird/91.9.0 Subject: Re: [PATCH v2 00/15] mlx5/vdpa: optimize live migration time From: Maxime Coquelin To: Li Zhang , orika@nvidia.com, viacheslavo@nvidia.com, matan@nvidia.com, shahafs@nvidia.com Cc: dev@dpdk.org, thomas@monjalon.net, rasland@nvidia.com, roniba@nvidia.com References: <20220408075606.33056-1-lizh@nvidia.com> <20220616023012.16013-1-lizh@nvidia.com> In-Reply-To: X-Scanned-By: MIMEDefang 2.79 on 10.11.54.5 Authentication-Results: relay.mimecast.com; auth=pass smtp.auth=CUSA124A263 smtp.mailfrom=maxime.coquelin@redhat.com X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Language: en-US Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org On 6/16/22 09:24, Maxime Coquelin wrote: > Hi Li, > > On 6/16/22 04:29, Li Zhang wrote: >> Allow the driver to use internal threads to >> obtain fast configuration. >> All the threads will be open on the same core of >> the event completion queue scheduling thread. >> >> Add max_conf_threads parameter to configure >> the maximum number of internal threads in addition to >> the caller thread (8 is suggested). >> These internal threads to pipeline handle VDPA tasks >> in system and shared with all VDPA devices. >> Default is 0, don't use internal threads for configuration. >> >> Depends-on: series=21868 ("vdpa/mlx5: improve device shutdown time") >> http://patchwork.dpdk.org/project/dpdk/list/?series=21868 >> >> RFC ("Add vDPA multi-threads optiomization") >> https://patchwork.dpdk.org/project/dpdk/cover/20220408075606.33056-1-lizh@nvidia.com/ >> > > I just notice there was a RFC that was sent on time because I was not > cc'ed. I thought V1, which arrived on June 6th was targetting v22.11. Ok, so checking with Thomas, get_maintainer.pl script does not return me for vDPA drivers patches, so that 'explain why I'm not cc'ed automatically. Also, the auto-delegation script in patchwork seems to assign it to Andrew, that's why I did not see it. I'll try to review it tomorrow. > Given how late we are in the schedule for v22.07, this series will be > postponed to v22.11. > > Regards, > Maxime > >> V2: >> * Drop eal device removal patch in series. >> * Add release note in release_22_07.rst. >> >> Li Zhang (12): >>    vdpa/mlx5: fix usage of capability for max number of virtqs >>    common/mlx5: extend virtq modifiable fields >>    vdpa/mlx5: pre-create virtq in the prob >>    vdpa/mlx5: optimize datapath-control synchronization >>    vdpa/mlx5: add multi-thread management for configuration >>    vdpa/mlx5: add task ring for MT management >>    vdpa/mlx5: add MT task for VM memory registration >>    vdpa/mlx5: add virtq creation task for MT management >>    vdpa/mlx5: add virtq LM log task >>    vdpa/mlx5: add device close task >>    vdpa/mlx5: add virtq sub-resources creation >>    vdpa/mlx5: prepare virtqueue resource creation >> >> Yajun Wu (3): >>    vdpa/mlx5: support pre create virtq resource >>    common/mlx5: add DevX API to move QP to reset state >>    vdpa/mlx5: support event qp reuse >> >>   doc/guides/rel_notes/release_22_07.rst |   5 + >>   doc/guides/vdpadevs/mlx5.rst           |  25 + >>   drivers/common/mlx5/mlx5_devx_cmds.c   |  77 ++- >>   drivers/common/mlx5/mlx5_devx_cmds.h   |   6 +- >>   drivers/common/mlx5/mlx5_prm.h         |  30 +- >>   drivers/vdpa/mlx5/meson.build          |   1 + >>   drivers/vdpa/mlx5/mlx5_vdpa.c          | 270 ++++++++-- >>   drivers/vdpa/mlx5/mlx5_vdpa.h          | 152 +++++- >>   drivers/vdpa/mlx5/mlx5_vdpa_cthread.c  | 360 ++++++++++++++ >>   drivers/vdpa/mlx5/mlx5_vdpa_event.c    | 160 ++++-- >>   drivers/vdpa/mlx5/mlx5_vdpa_lm.c       | 128 ++++- >>   drivers/vdpa/mlx5/mlx5_vdpa_mem.c      | 270 ++++++---- >>   drivers/vdpa/mlx5/mlx5_vdpa_steer.c    |  22 +- >>   drivers/vdpa/mlx5/mlx5_vdpa_virtq.c    | 654 ++++++++++++++++++------- >>   14 files changed, 1776 insertions(+), 384 deletions(-) >>   create mode 100644 drivers/vdpa/mlx5/mlx5_vdpa_cthread.c >>