From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mx1.redhat.com (mx1.redhat.com [209.132.183.28]) by dpdk.org (Postfix) with ESMTP id 9BEF93979 for ; Tue, 17 Nov 2015 10:24:23 +0100 (CET) Received: from int-mx09.intmail.prod.int.phx2.redhat.com (int-mx09.intmail.prod.int.phx2.redhat.com [10.5.11.22]) by mx1.redhat.com (Postfix) with ESMTPS id EE8AE8E50E; Tue, 17 Nov 2015 09:24:22 +0000 (UTC) Received: from [10.72.7.81] (vpn1-7-81.pek2.redhat.com [10.72.7.81]) by int-mx09.intmail.prod.int.phx2.redhat.com (8.14.4/8.14.4) with ESMTP id tAH9OEJ8031740; Tue, 17 Nov 2015 04:24:16 -0500 To: "Michael S. Tsirkin" , Flavio Leitner References: <20151021133043-mutt-send-email-mst@redhat.com> <20151021124815.GG3115@yliu-dev.sh.intel.com> <20151021172336-mutt-send-email-mst@redhat.com> <20151022094955.GR3115@yliu-dev.sh.intel.com> <20151022142141-mutt-send-email-mst@redhat.com> <20151024023408.GA7182@x240.home> <20151024204401-mutt-send-email-mst@redhat.com> <20151028203041.GB22208@x240.home> <20151028230324-mutt-send-email-mst@redhat.com> <20151116222057.GA2340@x240.home> <20151117094416-mutt-send-email-mst@redhat.com> From: Jason Wang Message-ID: <564AF23E.4000006@redhat.com> Date: Tue, 17 Nov 2015 17:24:14 +0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:38.0) Gecko/20100101 Thunderbird/38.3.0 MIME-Version: 1.0 In-Reply-To: <20151117094416-mutt-send-email-mst@redhat.com> Content-Type: text/plain; charset=windows-1252 Content-Transfer-Encoding: 7bit X-Scanned-By: MIMEDefang 2.68 on 10.5.11.22 Cc: dev@dpdk.org, marcel@redhat.com Subject: Re: [dpdk-dev] [PATCH v7 4/8] vhost: rxtx: use queue id instead of constant ring index X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: patches and discussions about DPDK List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 17 Nov 2015 09:24:24 -0000 On 11/17/2015 04:23 PM, Michael S. Tsirkin wrote: > On Mon, Nov 16, 2015 at 02:20:57PM -0800, Flavio Leitner wrote: >> > On Wed, Oct 28, 2015 at 11:12:25PM +0200, Michael S. Tsirkin wrote: >>> > > On Wed, Oct 28, 2015 at 06:30:41PM -0200, Flavio Leitner wrote: >>>> > > > On Sat, Oct 24, 2015 at 08:47:10PM +0300, Michael S. Tsirkin wrote: >>>>> > > > > On Sat, Oct 24, 2015 at 12:34:08AM -0200, Flavio Leitner wrote: >>>>>> > > > > > On Thu, Oct 22, 2015 at 02:32:31PM +0300, Michael S. Tsirkin wrote: >>>>>>> > > > > > > On Thu, Oct 22, 2015 at 05:49:55PM +0800, Yuanhan Liu wrote: >>>>>>>> > > > > > > > On Wed, Oct 21, 2015 at 05:26:18PM +0300, Michael S. Tsirkin wrote: >>>>>>>>> > > > > > > > > On Wed, Oct 21, 2015 at 08:48:15PM +0800, Yuanhan Liu wrote: >>>>>>>>>>> > > > > > > > > > > Please note that for virtio devices, guest is supposed to >>>>>>>>>>> > > > > > > > > > > control the placement of incoming packets in RX queues. >>>>>>>>>> > > > > > > > > > >>>>>>>>>> > > > > > > > > > I may not follow you. >>>>>>>>>> > > > > > > > > > >>>>>>>>>> > > > > > > > > > Enqueuing packets to a RX queue is done at vhost lib, outside the >>>>>>>>>> > > > > > > > > > guest, how could the guest take the control here? >>>>>>>>>> > > > > > > > > > >>>>>>>>>> > > > > > > > > > --yliu >>>>>>>>> > > > > > > > > >>>>>>>>> > > > > > > > > vhost should do what guest told it to. >>>>>>>>> > > > > > > > > >>>>>>>>> > > > > > > > > See virtio spec: >>>>>>>>> > > > > > > > > 5.1.6.5.5 Automatic receive steering in multiqueue mode >>>>>>>> > > > > > > > >>>>>>>> > > > > > > > Spec says: >>>>>>>> > > > > > > > >>>>>>>> > > > > > > > After the driver transmitted a packet of a flow on transmitqX, >>>>>>>> > > > > > > > the device SHOULD cause incoming packets for that flow to be >>>>>>>> > > > > > > > steered to receiveqX. >>>>>>>> > > > > > > > >>>>>>>> > > > > > > > >>>>>>>> > > > > > > > Michael, I still have no idea how vhost could know the flow even >>>>>>>> > > > > > > > after discussion with Huawei. Could you be more specific about >>>>>>>> > > > > > > > this? Say, how could guest know that? And how could guest tell >>>>>>>> > > > > > > > vhost which RX is gonna to use? >>>>>>>> > > > > > > > >>>>>>>> > > > > > > > Thanks. >>>>>>>> > > > > > > > >>>>>>>> > > > > > > > --yliu >>>>>>> > > > > > > >>>>>>> > > > > > > I don't really understand the question. >>>>>>> > > > > > > >>>>>>> > > > > > > When guests transmits a packet, it makes a decision >>>>>>> > > > > > > about the flow to use, and maps that to a tx/rx pair of queues. >>>>>>> > > > > > > >>>>>>> > > > > > > It sends packets out on the tx queue and expects device to >>>>>>> > > > > > > return packets from the same flow on the rx queue. >>>>>> > > > > > >>>>>> > > > > > Why? I can understand that there should be a mapping between >>>>>> > > > > > flows and queues in a way that there is no re-ordering, but >>>>>> > > > > > I can't see the relation of receiving a flow with a TX queue. >>>>>> > > > > > >>>>>> > > > > > fbl >>>>> > > > > >>>>> > > > > That's the way virtio chose to program the rx steering logic. >>>>> > > > > >>>>> > > > > It's low overhead (no special commands), and >>>>> > > > > works well for TCP when user is an endpoint since rx and tx >>>>> > > > > for tcp are generally tied (because of ack handling). >> > >> > It is low overhead for the control plane, but not for the data plane. > Well, there's zero data plane overhead within the guest. > You can't go lower :) > >>>>> > > > > We can discuss other ways, e.g. special commands for guests to >>>>> > > > > program steering. >>>>> > > > > We'd have to first see some data showing the current scheme >>>>> > > > > is problematic somehow. >> > >> > The issue is that the spec assumes the packets are coming in >> > a serialized way and the distribution will be made by vhost-user >> > but that isn't necessarily true. >> > > Making the distribution guest controlled is obviously the right > thing to do if guest is the endpoint: we need guest scheduler to > make the decisions, it's the only entity that knows > how are tasks distributed across VCPUs. > > It's possible that this is not the right thing for when guest > is just doing bridging between two VNICs: > are you saying packets should just go from RX queue N > on eth0 to TX queue N on eth1, making host make all > the queue selection decisions? The problem looks like current automatic steering policy is not flexible for all kinds of workload in guest. So we can implement the feature of ntuple filters and export the interfaces to let guest/drivers to decide. > > This sounds reasonable. Since there's a mix of local and > bridged traffic normally, does this mean we need > a per-packet flag that tells host to > ignore the packet for classification purposes? This may not work well for all workloads. E.g shot live connections.