From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id A300EA0545; Mon, 20 Jun 2022 11:25:56 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 894F240F19; Mon, 20 Jun 2022 11:25:56 +0200 (CEST) Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) by mails.dpdk.org (Postfix) with ESMTP id 219F740150 for ; Mon, 20 Jun 2022 11:25:55 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1655717154; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=g4YJbXe3NAHJggv7TcdodEM3EhJpaN/2DeLrAgGoJYM=; b=QRXfFmm32COcS6nBWVHCfjAXSaAhVjd6xOF1Wvgq1nm4gRh5+IC6LotSZH+uL5cI/Nw0SR SbMHLQoWqb272vCuXlMabBoE/kpkF1NcyZ6WWFT2KeVrUt5nLEENvoAFp9aPBw6EiO+dr5 p5ziiYyHZz4u8NdQXZivfgBXajPcKKA= Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-459-dhLmmYpfN2K3FLW7crJgTQ-1; Mon, 20 Jun 2022 05:25:48 -0400 X-MC-Unique: dhLmmYpfN2K3FLW7crJgTQ-1 Received: from smtp.corp.redhat.com (int-mx09.intmail.prod.int.rdu2.redhat.com [10.11.54.9]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id BE7BC85A582; Mon, 20 Jun 2022 09:25:47 +0000 (UTC) Received: from [10.39.208.29] (unknown [10.39.208.29]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 05A8D492CA4; Mon, 20 Jun 2022 09:25:45 +0000 (UTC) Message-ID: <8e8177f6-2a65-9512-f483-7732789fcd7a@redhat.com> Date: Mon, 20 Jun 2022 11:25:44 +0200 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:91.0) Gecko/20100101 Thunderbird/91.9.0 Subject: Re: [PATCH v4 07/15] vdpa/mlx5: optimize datapath-control synchronization To: Li Zhang , orika@nvidia.com, viacheslavo@nvidia.com, matan@nvidia.com, shahafs@nvidia.com Cc: dev@dpdk.org, thomas@monjalon.net, rasland@nvidia.com, roniba@nvidia.com References: <20220408075606.33056-1-lizh@nvidia.com> <20220618090258.91157-1-lizh@nvidia.com> <20220618090258.91157-8-lizh@nvidia.com> From: Maxime Coquelin In-Reply-To: <20220618090258.91157-8-lizh@nvidia.com> X-Scanned-By: MIMEDefang 2.85 on 10.11.54.9 Authentication-Results: relay.mimecast.com; auth=pass smtp.auth=CUSA124A263 smtp.mailfrom=maxime.coquelin@redhat.com X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Language: en-US Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org On 6/18/22 11:02, Li Zhang wrote: > The driver used a single global lock for any synchronization > needed for the datapath and control path. > It is better to group the critical sections with > the other ones that should be synchronized. > > Replace the global lock with the following locks: > > 1.virtq locks(per virtq) synchronize datapath polling and > parallel configurations on the same virtq. > 2.A doorbell lock synchronizes doorbell update, > which is shared for all the virtqs in the device. > 3.A steering lock for the shared steering objects updates. > > Signed-off-by: Li Zhang > Acked-by: Matan Azrad > --- > drivers/vdpa/mlx5/mlx5_vdpa.c | 24 ++++--- > drivers/vdpa/mlx5/mlx5_vdpa.h | 13 ++-- > drivers/vdpa/mlx5/mlx5_vdpa_event.c | 97 ++++++++++++++++++----------- > drivers/vdpa/mlx5/mlx5_vdpa_lm.c | 36 ++++++++--- > drivers/vdpa/mlx5/mlx5_vdpa_steer.c | 7 ++- > drivers/vdpa/mlx5/mlx5_vdpa_virtq.c | 88 +++++++++++++++++++------- > 6 files changed, 186 insertions(+), 79 deletions(-) > Reviewed-by: Maxime Coquelin Thanks, Maxime