From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mga03.intel.com (mga03.intel.com [134.134.136.65]) by dpdk.org (Postfix) with ESMTP id 5E7E3F72 for ; Wed, 13 Jan 2016 17:40:12 +0100 (CET) Received: from orsmga002.jf.intel.com ([10.7.209.21]) by orsmga103.jf.intel.com with ESMTP; 13 Jan 2016 08:40:11 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.22,290,1449561600"; d="scan'208";a="889773605" Received: from bricha3-mobl3.ger.corp.intel.com ([10.237.208.64]) by orsmga002.jf.intel.com with SMTP; 13 Jan 2016 08:40:09 -0800 Received: by (sSMTP sendmail emulation); Wed, 13 Jan 2016 16:40:08 +0025 Date: Wed, 13 Jan 2016 16:40:08 +0000 From: Bruce Richardson To: "Hanoch Haim (hhaim)" Message-ID: <20160113164008.GA16736@bricha3-MOBL3> References: <568A7959.7030506@6wind.com> <7f5255b98dcb4f2396ada16d5eb43e5a@XCH-RTP-017.cisco.com> <568BA1A1.2070300@6wind.com> <521c3b8c6a014c24824bb1c5f17dca03@XCH-RTP-017.cisco.com> <20160113114820.GB7216@bricha3-MOBL3> <6d8a8e02397b4a01940064eaa1ceb581@XCH-RTP-017.cisco.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <6d8a8e02397b4a01940064eaa1ceb581@XCH-RTP-017.cisco.com> Organization: Intel Shannon Ltd. User-Agent: Mutt/1.5.23 (2014-03-12) Cc: "dev@dpdk.org" , "Itay Marom \(imarom\)" , "Ido Barnea \(ibarnea\)" Subject: Re: [dpdk-dev] [PATCH v2] mbuf: optimize rte_mbuf_refcnt_update X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: patches and discussions about DPDK List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Wed, 13 Jan 2016 16:40:12 -0000 On Wed, Jan 13, 2016 at 04:28:34PM +0000, Hanoch Haim (hhaim) wrote: > Hi Bruce. > This is exactly what was possible before the patch and does *not* work after this patch. > I think it is more related to the semantic of the API. My understanding of the API was that I could allocate a share mbuf (ref=1) and then attach it from many concurrent threads (as each thread attach opt inc in 1 and driver reduces, atomic by 1) > After the patch there is a need to make the ref of the share mbuf to be 2 (why?) and free it twice at the end( ?), this makes it cumbersome to use it and for sure there is a need to describe this semantic change. > There is no such thing as allocating a shared mbuf. Mbufs are always allocated with a reference count of one. To share an mbuf among threads, then the ref count must be increased. NOTE: this is a different operation to that of "attaching" another mbuf to it - it just happens that when attaching one mbuf to another you have to increase the reference count as the attaching mbuf acts like another thread holding a pointer to the original mbuf. Once finished with the mbuf, each thread (or pointing mbuf) needs to decrement the reference counter, and the last one holding the reference is then responsible for freeing that buffer. The easiest way to do this is to use the free call and let it handle the reference counts, but if it makes more logical sense in an app to do the decrement and checking itself (so that inc's and dec's match in the app code), then that can work too, obviously. > However, this change is good for the common case, and I think it is better to keep it. > The options I see are: > 1) Document the current behavior > 2) Create two sets of API with two types of semantic Given the confusion I think documenting the way things are "meant to work" is a good idea. /Bruce > > thanks, > Hanoh > > -----Original Message----- > From: Bruce Richardson [mailto:bruce.richardson@intel.com] > Sent: Wednesday, January 13, 2016 1:48 PM > To: Hanoch Haim (hhaim) > Cc: Olivier MATZ; dev@dpdk.org; Ido Barnea (ibarnea); Itay Marom (imarom) > Subject: Re: [dpdk-dev] [PATCH v2] mbuf: optimize rte_mbuf_refcnt_update > > On Tue, Jan 05, 2016 at 11:11:24AM +0000, Hanoch Haim (hhaim) wrote: > > Hi Oliver, > > Thank you for the fast response and it would be great to open a discussion on that. > > In general our project can leverage your optimization and I think it is great (we should have thought about it) . We can use it using the workaround I described. > > However, for me it seems odd that rte_pktmbuf_attach () that does not *change* anything in m_const, except of the *atomic* ref counter does not work in parallel. > > The example I gave is a classic use case of rte_pktmbuf_attach (multicast ) and I don't see why it wouldn't work after your optimization. > > > > Do you have a pointer to the documentation that state that that you can't call the atomic ref counter from more than one thread? > > > Hi, > > actually, the issue is not that you can't work with the reference counter field from multiple threads, or that you can't use an mbuf from multiple threads, it's that if you are working with the same mbuf in multiple threads you have multiple references to the mbuf and your application must increase the reference counter appropriately. For example, if thread A is going to pass an mbuf to thread B and keep using it itself, you must increment the reference counter in thread A before enqueuing it to B. > > Regards, > /Bruce >