From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id C5D19A04F0; Tue, 10 Dec 2019 19:07:25 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 077931B13C; Tue, 10 Dec 2019 19:07:25 +0100 (CET) Received: from mga02.intel.com (mga02.intel.com [134.134.136.20]) by dpdk.org (Postfix) with ESMTP id 2B04D23D for ; Tue, 10 Dec 2019 19:07:22 +0100 (CET) X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga101.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 10 Dec 2019 10:07:21 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.69,300,1571727600"; d="scan'208";a="244937248" Received: from fyigit-mobl.ger.corp.intel.com (HELO [10.237.221.96]) ([10.237.221.96]) by fmsmga002.fm.intel.com with ESMTP; 10 Dec 2019 10:07:19 -0800 To: Shahaf Shuler , Olivier Matz , "dev@dpdk.org" , "Ananyev, Konstantin" , Thomas Monjalon , Bruce Richardson , Andrew Rybchenko References: <20180123135308.tr7nmuqsdeogm7bl@glumotte.dev.6wind.com> From: Ferruh Yigit Autocrypt: addr=ferruh.yigit@intel.com; keydata= mQINBFXZCFABEADCujshBOAaqPZpwShdkzkyGpJ15lmxiSr3jVMqOtQS/sB3FYLT0/d3+bvy qbL9YnlbPyRvZfnP3pXiKwkRoR1RJwEo2BOf6hxdzTmLRtGtwWzI9MwrUPj6n/ldiD58VAGQ +iR1I/z9UBUN/ZMksElA2D7Jgg7vZ78iKwNnd+vLBD6I61kVrZ45Vjo3r+pPOByUBXOUlxp9 GWEKKIrJ4eogqkVNSixN16VYK7xR+5OUkBYUO+sE6etSxCr7BahMPKxH+XPlZZjKrxciaWQb +dElz3Ab4Opl+ZT/bK2huX+W+NJBEBVzjTkhjSTjcyRdxvS1gwWRuXqAml/sh+KQjPV1PPHF YK5LcqLkle+OKTCa82OvUb7cr+ALxATIZXQkgmn+zFT8UzSS3aiBBohg3BtbTIWy51jNlYdy ezUZ4UxKSsFuUTPt+JjHQBvF7WKbmNGS3fCid5Iag4tWOfZoqiCNzxApkVugltxoc6rG2TyX CmI2rP0mQ0GOsGXA3+3c1MCdQFzdIn/5tLBZyKy4F54UFo35eOX8/g7OaE+xrgY/4bZjpxC1 1pd66AAtKb3aNXpHvIfkVV6NYloo52H+FUE5ZDPNCGD0/btFGPWmWRmkPybzColTy7fmPaGz cBcEEqHK4T0aY4UJmE7Ylvg255Kz7s6wGZe6IR3N0cKNv++O7QARAQABtCVGZXJydWggWWln aXQgPGZlcnJ1aC55aWdpdEBpbnRlbC5jb20+iQJUBBMBCgA+AhsDAh4BAheABQsJCAcDBRUK CQgLBRYCAwEAFiEE0jZTh0IuwoTjmYHH+TPrQ98TYR8FAl1meboFCQlupOoACgkQ+TPrQ98T YR9ACBAAv2tomhyxY0Tp9Up7mNGLfEdBu/7joB/vIdqMRv63ojkwr9orQq5V16V/25+JEAD0 60cKodBDM6HdUvqLHatS8fooWRueSXHKYwJ3vxyB2tWDyZrLzLI1jxEvunGodoIzUOtum0Ce gPynnfQCelXBja0BwLXJMplM6TY1wXX22ap0ZViC0m714U5U4LQpzjabtFtjT8qOUR6L7hfy YQ72PBuktGb00UR/N5UrR6GqB0x4W41aZBHXfUQnvWIMmmCrRUJX36hOTYBzh+x86ULgg7H2 1499tA4o6rvE13FiGccplBNWCAIroAe/G11rdoN5NBgYVXu++38gTa/MBmIt6zRi6ch15oLA Ln2vHOdqhrgDuxjhMpG2bpNE36DG/V9WWyWdIRlz3NYPCDM/S3anbHlhjStXHOz1uHOnerXM 1jEjcsvmj1vSyYoQMyRcRJmBZLrekvgZeh7nJzbPHxtth8M7AoqiZ/o/BpYU+0xZ+J5/szWZ aYxxmIRu5ejFf+Wn9s5eXNHmyqxBidpCWvcbKYDBnkw2+Y9E5YTpL0mS0dCCOlrO7gca27ux ybtbj84aaW1g0CfIlUnOtHgMCmz6zPXThb+A8H8j3O6qmPoVqT3qnq3Uhy6GOoH8Fdu2Vchh TWiF5yo+pvUagQP6LpslffufSnu+RKAagkj7/RSuZV25Ag0EV9ZMvgEQAKc0Db17xNqtSwEv mfp4tkddwW9XA0tWWKtY4KUdd/jijYqc3fDD54ESYpV8QWj0xK4YM0dLxnDU2IYxjEshSB1T qAatVWz9WtBYvzalsyTqMKP3w34FciuL7orXP4AibPtrHuIXWQOBECcVZTTOdZYGAzaYzxiA ONzF9eTiwIqe9/oaOjTwTLnOarHt16QApTYQSnxDUQljeNvKYt1lZE/gAUUxNLWsYyTT+22/ vU0GDUahsJxs1+f1yEr+OGrFiEAmqrzpF0lCS3f/3HVTU6rS9cK3glVUeaTF4+1SK5ZNO35p iVQCwphmxa+dwTG/DvvHYCtgOZorTJ+OHfvCnSVjsM4kcXGjJPy3JZmUtyL9UxEbYlrffGPQ I3gLXIGD5AN5XdAXFCjjaID/KR1c9RHd7Oaw0Pdcq9UtMLgM1vdX8RlDuMGPrj5sQrRVbgYH fVU/TQCk1C9KhzOwg4Ap2T3tE1umY/DqrXQgsgH71PXFucVjOyHMYXXugLT8YQ0gcBPHy9mZ qw5mgOI5lCl6d4uCcUT0l/OEtPG/rA1lxz8ctdFBVOQOxCvwRG2QCgcJ/UTn5vlivul+cThi 6ERPvjqjblLncQtRg8izj2qgmwQkvfj+h7Ex88bI8iWtu5+I3K3LmNz/UxHBSWEmUnkg4fJl Rr7oItHsZ0ia6wWQ8lQnABEBAAGJAjwEGAEKACYCGwwWIQTSNlOHQi7ChOOZgcf5M+tD3xNh HwUCXWZ5wAUJB3FgggAKCRD5M+tD3xNhH2O+D/9OEz62YuJQLuIuOfL67eFTIB5/1+0j8Tsu o2psca1PUQ61SZJZOMl6VwNxpdvEaolVdrpnSxUF31kPEvR0Igy8HysQ11pj8AcgH0a9FrvU /8k2Roccd2ZIdpNLkirGFZR7LtRw41Kt1Jg+lafI0efkiHKMT/6D/P1EUp1RxOBNtWGV2hrd 0Yg9ds+VMphHHU69fDH02SwgpvXwG8Qm14Zi5WQ66R4CtTkHuYtA63sS17vMl8fDuTCtvfPF HzvdJLIhDYN3Mm1oMjKLlq4PUdYh68Fiwm+boJoBUFGuregJFlO3hM7uHBDhSEnXQr5mqpPM 6R/7Q5BjAxrwVBisH0yQGjsWlnysRWNfExAE2sRePSl0or9q19ddkRYltl6X4FDUXy2DTXa9 a+Fw4e1EvmcF3PjmTYs9IE3Vc64CRQXkhujcN4ZZh5lvOpU8WgyDxFq7bavFnSS6kx7Tk29/ wNJBp+cf9qsQxLbqhW5kfORuZGecus0TLcmpZEFKKjTJBK9gELRBB/zoN3j41hlEl7uTUXTI JQFLhpsFlEdKLujyvT/aCwP3XWT+B2uZDKrMAElF6ltpTxI53JYi22WO7NH7MR16Fhi4R6vh FHNBOkiAhUpoXRZXaCR6+X4qwA8CwHGqHRBfYFSU/Ulq1ZLR+S3hNj2mbnSx0lBs1eEqe2vh cA== Cc: Matan Azrad , Jerin Jacob Kollanukkaran Message-ID: <65f5f247-15e7-ac0a-183e-8a66193f426f@intel.com> Date: Tue, 10 Dec 2019 18:07:19 +0000 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 8bit Subject: Re: [dpdk-dev] questions about new offload ethdev api X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" On 1/23/2018 2:34 PM, Shahaf Shuler wrote: > Tuesday, January 23, 2018 3:53 PM, Olivier Matz: <...> >> >> 2/ meaning of rxmode.jumbo_frame, rxmode.enable_scatter, >> rxmode.max_rx_pkt_len >> >> While it's not related to the new API, it is probably a good opportunity >> to clarify the meaning of these flags. I'm not able to find a good >> documentation about them. >> >> Here is my understanding, the configuration only depends on: - the maximum >> rx frame length - the amount of data available in a mbuf (minus headroom) >> >> Flags to set in rxmode (example): >> +---------------+----------------+----------------+-----------------+ | >> |mbuf_data_len=1K|mbuf_data_len=2K|mbuf_data_len=16K| >> +---------------+----------------+----------------+-----------------+ >> |max_rx_len=1500|enable_scatter | | | >> +---------------+----------------+----------------+-----------------+ >> |max_rx_len=9000|enable_scatter, |enable_scatter, |jumbo_frame | | >> |jumbo_frame |jumbo_frame | | >> +---------------+----------------+----------------+-----------------+ >> >> If this table is correct, the flag jumbo_frame would be equivalent to check >> if max_rx_pkt_len is above a threshold. >> >> And enable_scatter could be deduced from the mbuf size of the given rxq >> (which is a bit harder but maybe doable). > > I glad you raised this subject. We had a lot of discussion on it internally > in Mellanox. > > I fully agree. All application needs is to specify the maximum packet size it > wants to receive. > > I think also the lack of documentation is causing PMDs to use those flags > wrongly. For example - some PMDs set the jumbo_frame flag internally without > it being set by the application. > > I would like to add one more item : MTU. What is the relation (if any) > between setting MTU and the max_rx_len ? I know MTU stands for Max Transmit > Unit, however at least in Linux it is the same for the Send and the receive. > > (Resurrecting the thread after two years, I will reply again with latest understanding.) Thanks Olivier for above summary and table, and unfortunately usage still not consistent between PMDs. According my understanding: 'max_rx_pkt_len' is user configuration value, to limit the size packet that is shared with host, but this doesn't limit the size of packet that NIC receives. Like if the mbuf size of the mempool used by a queue is 1024 bytes, we don't want packets bigger than buffer size, but if NIC supports it is possible receive 6000 bytes packet and split data into multiple buffers, and we can use multi segment packets to represent it. So what we need is NIC ability to limit the size of data to share to host and scattered Rx support (device + driver). But MTU limits the size of the packet that NIC receives. Assuming above are correct J, Using mbuf data size as 'max_rx_pkt_len' without asking from user is an option, but perhaps user has different reason to limit packet size, so I think better to keep as different config option. I think PMD itself enabling "jumbo frame" offload is not too bad, and acceptable, since providing a large MTU already implies it. But not sure about PMD enabling scattered Rx, application may want to force to receive single segment mbufs, for that case PMD enabling this config on its own looks like a problem. But user really needs this when a packet doesn't fit to the mbuf, so providing a MTU larger than 'max_rx_pkt_len' _may_ imply enabling scattered Rx, I assume this is the logic in some PMDs which looks acceptable. And PMD behavior should be according for mentioned configs: 1) Don't change user provided 'max_rx_pkt_len' value 2) If jumbo frame is not enabled, don't limit the size of packets to the host (I think this is based on assumption that mbuf size always will be > 1514) 3) When user request to set the MTU bigger than ETH_MAX, PMD enable jumbo frame support (if it is not enabled by user already and supported by HW). If HW doesn't support if of course it should fail. 4) When user request to set MTU bigger than 'max_rx_pkt_len' 4a) if "scattered Rx" is enabled, configure the MTU and limit packet size to host to 'max_rx_pkt_len' 4b) if "scattered Rx" is not enabled but HW supports it, enable "scattered Rx" by PMD, configure the MTU and limit packet size to host to 'max_rx_pkt_len' 4c) if "scattered Rx" is not enabled and not supported by HW, fail MTU set. 4d) if HW doesn't support to limit the packet size to host, but requested MTU bigger than 'max_rx_pkt_len' it should fail. Btw, I am aware of that some PMDs have a larger MTU by default and can't limit the packet size to host to 'max_rx_pkt_len' value, I don't know what to do in that case, fail in configure? Or at least be sure configured mempool's mbuf size is big enough? Thanks, ferruh