From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from EUR01-DB5-obe.outbound.protection.outlook.com (mail-db5eur01on0082.outbound.protection.outlook.com [104.47.2.82]) by dpdk.org (Postfix) with ESMTP id C6AA55398 for ; Sun, 10 Sep 2017 14:08:00 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=Mellanox.com; s=selector1; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version; bh=mBCItwksIEAtAxTxEVPQR0j7Iu+DjOUGv+RFUez8kfY=; b=q0rp5K1HEHwbEFDzKPQLaW27psOVqpZKuQ2PHGuUXwd17EmDDtgmlh9NwwSux3w0mPtqLVACdHHO2DNVYhfO3aKnVMBQVveSWsIBLhMhN7I8iAfNSqTOuCVLiR6lemgyTlGB7tSM1WKNfwpgs0zao5E7H45UMqRY5VzIfPKTg0U= Authentication-Results: spf=none (sender IP is ) smtp.mailfrom=shahafs@mellanox.com; Received: from mellanox.com (82.166.227.17) by AM4PR05MB3140.eurprd05.prod.outlook.com (2603:10a6:205:3::17) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384_P256) id 15.20.35.12; Sun, 10 Sep 2017 12:07:59 +0000 From: Shahaf Shuler To: thomas@monjalon.net Cc: dev@dpdk.org Date: Sun, 10 Sep 2017 15:07:48 +0300 Message-Id: <725cf2c5c2f8c163081958320bc1dbeeeeb1d1ad.1505044395.git.shahafs@mellanox.com> X-Mailer: git-send-email 2.12.0 In-Reply-To: References: MIME-Version: 1.0 Content-Type: text/plain X-Originating-IP: [82.166.227.17] X-ClientProxiedBy: VI1PR0502CA0025.eurprd05.prod.outlook.com (2603:10a6:803:1::38) To AM4PR05MB3140.eurprd05.prod.outlook.com (2603:10a6:205:3::17) X-MS-PublicTrafficType: Email X-MS-Office365-Filtering-Correlation-Id: 5eaf448f-7d7e-4e0c-b055-08d4f8449451 X-MS-Office365-Filtering-HT: Tenant X-Microsoft-Antispam: UriScan:; BCL:0; PCL:0; RULEID:(300000500095)(300135000095)(300000501095)(300135300095)(22001)(300000502095)(300135100095)(2017030254152)(48565401081)(300000503095)(300135400095)(201703131423075)(201703031133081)(201702281549075)(300000504095)(300135200095)(300000505095)(300135600095)(300000506095)(300135500095); SRVR:AM4PR05MB3140; X-Microsoft-Exchange-Diagnostics: 1; AM4PR05MB3140; 3:3jzkXJrbOGLekxUkNSk29Ef+jppo/DB/02qBCQy0X/OPGVp6dgR3OzqmvW3fuCOpD1fvfjVty6UMSAIoY1LpyTQYHi0KPpvJLPhkJgcerH8fKm2ks4iEAZVspUwWUiq0Yq/atoXkOWtnqX6NzpZe51mV3Y79YrAjXs+Oixx2AVxZEJx4xdhekx35DfpikmaA9HOFtHpBYNuz8DaoEE0n9IT5p/TJv/tiJALGe5sKgR90+dOuZ7uinxVA64pnJfjL; 25:sSs0FQS66CqxMdy5BbK/5ylCeuFkxXQmqivSIDTN5pptvqYEwwgWyHkAVxe4XwDgVnRd73ejgw2Mi+ryK1qGHruF0Xjoz4Ah+9ZT64rmhMIBodWwK8F2Dtw47QYdx6W7xPNQWC+0WRwb+eq3cdcFU6IYSffAxjieXO2wfHtdBYFY7bSkScMOR8dSqJOKj+6LUjCtiQMofFb0G+Ol3daKWIcR+AetcUinyYOfG2eUUKpqEjd6wsljvggKG1mZXzCqgX/KBiWboJfXFqcKwy/661CuAM22yfzp5g9tx0RfuuYXOLcPbTkme15Eo7oqg63HfJFA2gIYmjk35C6aDXjHUg==; 31:OpaB6xokzsk4G5le5ahrB7huyUW4xEF/hbRwXtZVpx+fb2dF4KYJXamlFwbsKQi/Fcs+IatCuUS7P07ORsoAtQUYXCk7nCsV/65C22q4rvRQjQFr4hvSeRv8IMXtVhB1U/oPB0hV8g0iHWIsnvSxSdHnEcZLP4mh9P17be3tycNt+5C0V8gYJj0De+SxO913sziP3SE2abmVf4CuyTUd0kotwCVDKd2do/asGqk1+rk= X-MS-TrafficTypeDiagnostic: AM4PR05MB3140: X-LD-Processed: a652971c-7d2e-4d9b-a6a4-d149256f461b,ExtAddr X-Microsoft-Exchange-Diagnostics: 1; AM4PR05MB3140; 20:7qOcfnBsRZ/uBzRCvVwvNMFQdj/hGU8f3KcbgjM87gJwwhg2r1+njtuHGNzhbCEaZYdDf9ts9e/fuq/8sYHjU6cbX5I9xKJcpXAsamJAVkGwPE/84EbUO0dEzrYGe5rHq1y7Iaqx7ixg+E00rhVCAXgQmEee0eKbhkmBgFyAq5wACPA80mpJsdczZRuxM7A4rQTk3v8U6Yo05/gOdpSLzuDXOg4qPDy+xUV1t4vKeYiEnoPnOafD3vIGP3oYZX5+Ep2DAYfkWilMRjTmelUrkKbi8BUVnLOmyPsjCBP0mBsg9BVpf9/6DapC2qKB3uQFcCpRc//MTp4h4qaRvgc7X+/STVswZnwwTbs8X0Tuyi9q0reRTZFRi8dI+e7WGjJ2jo5YJg0UDdZtRf6oeMETMDnx+4vIOEuQd5xLcTKKRLUKjK7BF71rnAZ8mJfhcHL8dWaI2XNd7oh/ZqQnkVBjTd7tc2AyRqaJHklDO/W0CPMBp8snJHbccSB7YNNZxXph; 4:+Bz9aZQ4HDdQ/X+Y93T7smMdFjvhQ7GYRqWT/ALLYYKGDGQ7cLXr5qxJ+Rl/OFPeyMzypOAk/jGBP9rfVdquGgc+3V+byzdASyUZZmceZpm34i1D0RbbpuhdZQAsCMA/LeSJAByeASMjOomhGbrmVzwIxElSD0kI6VookXw56oi9C9a7Csu6phT/KBqviWohilWrrfsESHuM2QAusymeKpl25NFKOywPTzHbpb5Mqu7Fp7M0IwlwvpoiQf2AhWNoaL/1nkvCamgngLa+3N0bBwPw7pUu88lX3mCG6vTVM3dRZh/bLIG+rzKaXnGyv8VOAL3uU9eWbTk2faPruXxUSA== X-Exchange-Antispam-Report-Test: UriScan:(20558992708506)(278428928389397); X-Microsoft-Antispam-PRVS: X-Exchange-Antispam-Report-CFA-Test: BCL:0; PCL:0; RULEID:(100000700101)(100105000095)(100000701101)(100105300095)(100000702101)(100105100095)(6040450)(2401047)(8121501046)(5005006)(93006095)(93001095)(3002001)(10201501046)(100000703101)(100105400095)(6055026)(6041248)(20161123560025)(20161123555025)(20161123564025)(20161123558100)(20161123562025)(201703131423075)(201702281528075)(201703061421075)(201703061406153)(6072148)(201708071742011)(100000704101)(100105200095)(100000705101)(100105500095); SRVR:AM4PR05MB3140; BCL:0; PCL:0; RULEID:(100000800101)(100110000095)(100000801101)(100110300095)(100000802101)(100110100095)(100000803101)(100110400095)(100000804101)(100110200095)(100000805101)(100110500095); SRVR:AM4PR05MB3140; X-Forefront-PRVS: 04267075BD X-Forefront-Antispam-Report: SFV:NSPM; SFS:(10009020)(4630300001)(7370300001)(6009001)(39860400002)(189002)(199003)(305945005)(50466002)(48376002)(50226002)(189998001)(2906002)(8676002)(21086003)(5660300001)(25786009)(66066001)(3846002)(69596002)(42186005)(47776003)(6116002)(7736002)(478600001)(7350300001)(81156014)(81166006)(4326008)(50986999)(55016002)(101416001)(53936002)(6666003)(86362001)(68736007)(6916009)(2950100002)(76176999)(105586002)(2351001)(2361001)(118296001)(106356001)(36756003)(110136004)(5003940100001)(97736004)(33646002)(33026002); DIR:OUT; SFP:1101; SCL:1; SRVR:AM4PR05MB3140; H:mellanox.com; FPR:; SPF:None; PTR:InfoNoRecords; MX:1; A:1; LANG:en; Received-SPF: None (protection.outlook.com: mellanox.com does not designate permitted sender hosts) X-Microsoft-Exchange-Diagnostics: =?us-ascii?Q?1; AM4PR05MB3140; 23:z/ZjdtzOL8fK1IPH8UYdS3HCX1VZZybp8r4MR65to?= =?us-ascii?Q?udk3JT3Xub1LT2DQmWYvqUr2L6NpCpAMbBzYJk3jMi1prWD3e1TKKMvNVdZa?= =?us-ascii?Q?5EiHVADHayIQ9Asp8HvXNR4SJd5qaPgaipmNBYaX6ofds08r6ZFrql/fGuXe?= =?us-ascii?Q?0JT0kVsWsh5r0378B+uNufhXr9nFwJasX47oUyxD3HMGTtFj8H2Snc9xM+ts?= =?us-ascii?Q?Xw/k5w3XZ0cku8W5CoeXk/jZ1duPWpskx4ooYKGT7cAbcwNcs5OvO5OpCdT9?= =?us-ascii?Q?v1D/PsZWP+Ju7DFmB6Pdqp9Fzewz9oOhZr/TbQykvQmIhkwWu6f38vIZm2FS?= =?us-ascii?Q?BfxZyc28vdVPdGklOtDAk5TVZrgUbWwB+KBWS7NyBX6nQ079hGfrXGEgQo8W?= =?us-ascii?Q?PUraJI7TbVZPsI6Rsk/Dry659QPMod55ZpC5S4/jg3XduI2jBeTBKIwmxVhQ?= =?us-ascii?Q?UJOx5Wh/I6HXD9fH8q/o7f0KDL7ZcLjsKcyLnTfP9A0ntkQTuK6QaN4okZ4D?= =?us-ascii?Q?FheXjR4Jo+HW7QLQfsFQFPHtipNASuqZpBKf9ptmizkDGYbbdWcaqp1k4s9S?= =?us-ascii?Q?UolSYhMx/Q6bW7txq+KSwgwqPdDRvpIYj5l7aE9o+W7uqEixuC+rAPo2Px0U?= =?us-ascii?Q?qGguXyI4mvex03OYjS9pEyH7xZejMfiBYwIDFlrTvI2dcf4V0NGCn7Su6YPh?= =?us-ascii?Q?ldv07eOKgJ75YSN1NxvmWkWjL/IDc+FUS+jORoAADUQKQ84rotOpsJsIGtpo?= =?us-ascii?Q?n0Jg8moHCQecHKiuBZQG/N5HJ+giuUCvdT4hMFAUrTGTYC+gNsTBKGEvdXcy?= =?us-ascii?Q?Jl1cEcKMV1w/wkR3Ak/eUau+2Dls7uDRj4Hh5vN/u8kX1XQ3NoF6xkiYlqMw?= =?us-ascii?Q?lQx5+zgJqFZRG1WnXpfoPgnJ0NrAasx5vBloYOBJBnbdeOyR1eZ0GCIkDqoP?= =?us-ascii?Q?i6dl1PkvVBSQn4feomvihGUgAJEUFjfzkqERnQiRAym42G4G/ahaMB/MWwUZ?= =?us-ascii?Q?leOEXphSF5tSKE5JWOxGsFiuT51b3JCQVKopkaIH6eMw4n9Es4jkEWrsZUhU?= =?us-ascii?Q?sKabKeFBA2UtM29Jm9k19s3wBt3OXDCaWAXx+u/x+/HXILm0isyzo7+eX1PE?= =?us-ascii?Q?7tTiq2TQWeOa0qapAjUvPAOPzobV2ntVBqv6g5PHZubxQ7btwlnDpAf39YIr?= =?us-ascii?Q?+G1bHRCArC/L+c1jnLWsrhfIJdsyNHfWfxD?= X-Microsoft-Exchange-Diagnostics: 1; AM4PR05MB3140; 6:tksTe/yTkxkc97VPmpTDILjpc60dxwcU1sIPVHcAHdlME1Mg4S6DenUcyO/EPJl4jQvlgOqJhUk9StsWI5c6DycIwIrJyRxCHgS61wIvIyFhFY9MbZZBpnlmwU82H5pGijjt2eq/Mbv+FbuVDe6oXqROee0v+aQprfZBFP+sD32DYhoFp2UrBdM7ObJoHMQcFgnOKc2YlgQ/RXrynWHNwxhvXPKqFZhQ3/HovGv3rYy+fd3kDbEEHmmyMPAPNTFA0YD0OuEHH5LSjZU0skikAfWfgyWjYkUZLxUy0D7FHUJBw4ky+LmDlc7jgUcua7FQaTyfLzpojYqFrd7sJ/Ch7w==; 5:fvQdxERoNErzu2JwazwbswZOYdhwZXpPfK4SZ8Fo7eZazoEyoxyBvA7SQMm7lR9ttxs4MzlKgVHbXwXRSWffpYHDM3sVsknQlP5HD5Gc98nBCuzr9n6Gsyfey3t11UHYnrozlOxPYOcOhvwhzTEe1A==; 24:0z+h1PgQ615k7zQbKuAgOiYgvaVQEShSqkEdFNJFFOi/JOBRRkW52hrRGp4zC8cVLMRaZCvpflYZ0ifL7hrJDtxsXEkuyjMeXjx2v08O1uY=; 7:O5bqE6Y4ZZFETNZQPDZjKRdaWjcm0U79GsfmhBOvRG6mYpVjN7Nj1UAtj0olx7IukgGzfHPrSfNm+tc6epx6QmE4tMPbJdub3rNNJGp0qj98TEb02BngKE+olYciGD49x7cQoz8o3DD+OGgVbXdyTMIucTtY7O+UmX8vGGGY8DPiBbS42DsIWMvvwko5NfsIOhoPob8HbfVigdg43PmBU4s+f5DHGZ3LPJiyf8Y5V00= SpamDiagnosticOutput: 1:99 SpamDiagnosticMetadata: NSPM X-OriginatorOrg: Mellanox.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 10 Sep 2017 12:07:59.1397 (UTC) X-MS-Exchange-CrossTenant-FromEntityHeader: Hosted X-MS-Exchange-Transport-CrossTenantHeadersStamped: AM4PR05MB3140 Subject: [dpdk-dev] [PATCH v2 1/2] ethdev: introduce Rx queue offloads API X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sun, 10 Sep 2017 12:08:01 -0000 Introduce a new API to configure Rx offloads. In the new API, offloads are divided into per-port and per-queue offloads. The PMD reports capability for each of them. Offloads are enabled using the existing DEV_RX_OFFLOAD_* flags. To enable per-port offload, the offload should be set on both device configuration and queue configuration. To enable per-queue offload, the offloads can be set only on queue configuration. Applications should set the ignore_offload_bitfield bit on rxmode structure in order to move to the new API. The old Rx offloads API is kept for the meanwhile, in order to enable a smooth transition for PMDs and application to the new API. Signed-off-by: Shahaf Shuler --- doc/guides/nics/features.rst | 19 +++-- lib/librte_ether/rte_ethdev.c | 156 +++++++++++++++++++++++++++++++++---- lib/librte_ether/rte_ethdev.h | 52 ++++++++++++- 3 files changed, 204 insertions(+), 23 deletions(-) diff --git a/doc/guides/nics/features.rst b/doc/guides/nics/features.rst index 37ffbc68c..f2c8497c2 100644 --- a/doc/guides/nics/features.rst +++ b/doc/guides/nics/features.rst @@ -179,7 +179,7 @@ Jumbo frame Supports Rx jumbo frames. -* **[uses] user config**: ``dev_conf.rxmode.jumbo_frame``, +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_JUMBO_FRAME``. ``dev_conf.rxmode.max_rx_pkt_len``. * **[related] rte_eth_dev_info**: ``max_rx_pktlen``. * **[related] API**: ``rte_eth_dev_set_mtu()``. @@ -192,7 +192,7 @@ Scattered Rx Supports receiving segmented mbufs. -* **[uses] user config**: ``dev_conf.rxmode.enable_scatter``. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_SCATTER``. * **[implements] datapath**: ``Scattered Rx function``. * **[implements] rte_eth_dev_data**: ``scattered_rx``. * **[provides] eth_dev_ops**: ``rxq_info_get:scattered_rx``. @@ -206,7 +206,7 @@ LRO Supports Large Receive Offload. -* **[uses] user config**: ``dev_conf.rxmode.enable_lro``. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_TCP_LRO``. * **[implements] datapath**: ``LRO functionality``. * **[implements] rte_eth_dev_data**: ``lro``. * **[provides] mbuf**: ``mbuf.ol_flags:PKT_RX_LRO``, ``mbuf.tso_segsz``. @@ -363,7 +363,7 @@ VLAN filter Supports filtering of a VLAN Tag identifier. -* **[uses] user config**: ``dev_conf.rxmode.hw_vlan_filter``. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_VLAN_FILTER``. * **[implements] eth_dev_ops**: ``vlan_filter_set``. * **[related] API**: ``rte_eth_dev_vlan_filter()``. @@ -499,7 +499,7 @@ CRC offload Supports CRC stripping by hardware. -* **[uses] user config**: ``dev_conf.rxmode.hw_strip_crc``. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_CRC_STRIP``. .. _nic_features_vlan_offload: @@ -509,8 +509,7 @@ VLAN offload Supports VLAN offload to hardware. -* **[uses] user config**: ``dev_conf.rxmode.hw_vlan_strip``, - ``dev_conf.rxmode.hw_vlan_filter``, ``dev_conf.rxmode.hw_vlan_extend``. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_VLAN_STRIP,DEV_RX_OFFLOAD_VLAN_FILTER,DEV_RX_OFFLOAD_VLAN_EXTEND``. * **[implements] eth_dev_ops**: ``vlan_offload_set``. * **[provides] mbuf**: ``mbuf.ol_flags:PKT_RX_VLAN_STRIPPED``, ``mbuf.vlan_tci``. * **[provides] rte_eth_dev_info**: ``rx_offload_capa:DEV_RX_OFFLOAD_VLAN_STRIP``, @@ -526,6 +525,7 @@ QinQ offload Supports QinQ (queue in queue) offload. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_QINQ_STRIP``. * **[uses] mbuf**: ``mbuf.ol_flags:PKT_TX_QINQ_PKT``. * **[provides] mbuf**: ``mbuf.ol_flags:PKT_RX_QINQ_STRIPPED``, ``mbuf.vlan_tci``, ``mbuf.vlan_tci_outer``. @@ -540,7 +540,7 @@ L3 checksum offload Supports L3 checksum offload. -* **[uses] user config**: ``dev_conf.rxmode.hw_ip_checksum``. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_IPV4_CKSUM``. * **[uses] mbuf**: ``mbuf.ol_flags:PKT_TX_IP_CKSUM``, ``mbuf.ol_flags:PKT_TX_IPV4`` | ``PKT_TX_IPV6``. * **[provides] mbuf**: ``mbuf.ol_flags:PKT_RX_IP_CKSUM_UNKNOWN`` | @@ -557,6 +557,7 @@ L4 checksum offload Supports L4 checksum offload. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_UDP_CKSUM,DEV_RX_OFFLOAD_TCP_CKSUM``. * **[uses] mbuf**: ``mbuf.ol_flags:PKT_TX_IPV4`` | ``PKT_TX_IPV6``, ``mbuf.ol_flags:PKT_TX_L4_NO_CKSUM`` | ``PKT_TX_TCP_CKSUM`` | ``PKT_TX_SCTP_CKSUM`` | ``PKT_TX_UDP_CKSUM``. @@ -574,6 +575,7 @@ MACsec offload Supports MACsec. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_MACSEC_STRIP``. * **[uses] mbuf**: ``mbuf.ol_flags:PKT_TX_MACSEC``. * **[provides] rte_eth_dev_info**: ``rx_offload_capa:DEV_RX_OFFLOAD_MACSEC_STRIP``, ``tx_offload_capa:DEV_TX_OFFLOAD_MACSEC_INSERT``. @@ -586,6 +588,7 @@ Inner L3 checksum Supports inner packet L3 checksum. +* **[uses] rte_eth_rxq_conf**: ``offloads:DEV_RX_OFFLOAD_OUTER_IPV4_CKSUM``. * **[uses] mbuf**: ``mbuf.ol_flags:PKT_TX_IP_CKSUM``, ``mbuf.ol_flags:PKT_TX_IPV4`` | ``PKT_TX_IPV6``, ``mbuf.ol_flags:PKT_TX_OUTER_IP_CKSUM``, diff --git a/lib/librte_ether/rte_ethdev.c b/lib/librte_ether/rte_ethdev.c index 0597641ee..b3c10701e 100644 --- a/lib/librte_ether/rte_ethdev.c +++ b/lib/librte_ether/rte_ethdev.c @@ -687,12 +687,90 @@ rte_eth_speed_bitflag(uint32_t speed, int duplex) } } +/** + * A conversion function from rxmode bitfield API. + */ +static void +rte_eth_convert_rx_offload_bitfield(const struct rte_eth_rxmode *rxmode, + uint64_t *rx_offloads) +{ + uint64_t offloads = 0; + + if (rxmode->header_split == 1) + offloads |= DEV_RX_OFFLOAD_HEADER_SPLIT; + if (rxmode->hw_ip_checksum == 1) + offloads |= DEV_RX_OFFLOAD_CHECKSUM; + if (rxmode->hw_vlan_filter == 1) + offloads |= DEV_RX_OFFLOAD_VLAN_FILTER; + if (rxmode->hw_vlan_strip == 1) + offloads |= DEV_RX_OFFLOAD_VLAN_STRIP; + if (rxmode->hw_vlan_extend == 1) + offloads |= DEV_RX_OFFLOAD_VLAN_EXTEND; + if (rxmode->jumbo_frame == 1) + offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; + if (rxmode->hw_strip_crc == 1) + offloads |= DEV_RX_OFFLOAD_CRC_STRIP; + if (rxmode->enable_scatter == 1) + offloads |= DEV_RX_OFFLOAD_SCATTER; + if (rxmode->enable_lro == 1) + offloads |= DEV_RX_OFFLOAD_TCP_LRO; + + *rx_offloads = offloads; +} + +/** + * A conversion function from rxmode offloads API. + */ +static void +rte_eth_convert_rx_offloads(const uint64_t rx_offloads, + struct rte_eth_rxmode *rxmode) +{ + + if (rx_offloads & DEV_RX_OFFLOAD_HEADER_SPLIT) + rxmode->header_split = 1; + else + rxmode->header_split = 0; + if (rx_offloads & DEV_RX_OFFLOAD_CHECKSUM) + rxmode->hw_ip_checksum = 1; + else + rxmode->hw_ip_checksum = 0; + if (rx_offloads & DEV_RX_OFFLOAD_VLAN_FILTER) + rxmode->hw_vlan_filter = 1; + else + rxmode->hw_vlan_filter = 0; + if (rx_offloads & DEV_RX_OFFLOAD_VLAN_STRIP) + rxmode->hw_vlan_strip = 1; + else + rxmode->hw_vlan_strip = 0; + if (rx_offloads & DEV_RX_OFFLOAD_VLAN_EXTEND) + rxmode->hw_vlan_extend = 1; + else + rxmode->hw_vlan_extend = 0; + if (rx_offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) + rxmode->jumbo_frame = 1; + else + rxmode->jumbo_frame = 0; + if (rx_offloads & DEV_RX_OFFLOAD_CRC_STRIP) + rxmode->hw_strip_crc = 1; + else + rxmode->hw_strip_crc = 0; + if (rx_offloads & DEV_RX_OFFLOAD_SCATTER) + rxmode->enable_scatter = 1; + else + rxmode->enable_scatter = 0; + if (rx_offloads & DEV_RX_OFFLOAD_TCP_LRO) + rxmode->enable_lro = 1; + else + rxmode->enable_lro = 0; +} + int rte_eth_dev_configure(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q, const struct rte_eth_conf *dev_conf) { struct rte_eth_dev *dev; struct rte_eth_dev_info dev_info; + struct rte_eth_conf local_conf = *dev_conf; int diag; RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -EINVAL); @@ -722,8 +800,20 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q, return -EBUSY; } + /* + * Convert between the offloads API to enable PMDs to support + * only one of them. + */ + if ((dev_conf->rxmode.ignore_offload_bitfield == 0)) { + rte_eth_convert_rx_offload_bitfield( + &dev_conf->rxmode, &local_conf.rxmode.offloads); + } else { + rte_eth_convert_rx_offloads(dev_conf->rxmode.offloads, + &local_conf.rxmode); + } + /* Copy the dev_conf parameter into the dev structure */ - memcpy(&dev->data->dev_conf, dev_conf, sizeof(dev->data->dev_conf)); + memcpy(&dev->data->dev_conf, &local_conf, sizeof(dev->data->dev_conf)); /* * Check that the numbers of RX and TX queues are not greater @@ -767,7 +857,7 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q, * If jumbo frames are enabled, check that the maximum RX packet * length is supported by the configured device. */ - if (dev_conf->rxmode.jumbo_frame == 1) { + if (local_conf.rxmode.offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) { if (dev_conf->rxmode.max_rx_pkt_len > dev_info.max_rx_pktlen) { RTE_PMD_DEBUG_TRACE("ethdev port_id=%d max_rx_pkt_len %u" @@ -1004,6 +1094,7 @@ rte_eth_rx_queue_setup(uint8_t port_id, uint16_t rx_queue_id, uint32_t mbp_buf_size; struct rte_eth_dev *dev; struct rte_eth_dev_info dev_info; + struct rte_eth_rxconf local_conf; void **rxq; RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -EINVAL); @@ -1074,8 +1165,18 @@ rte_eth_rx_queue_setup(uint8_t port_id, uint16_t rx_queue_id, if (rx_conf == NULL) rx_conf = &dev_info.default_rxconf; + local_conf = *rx_conf; + if (dev->data->dev_conf.rxmode.ignore_offload_bitfield == 0) { + /** + * Reflect port offloads to queue offloads in order for + * offloads to not be discarded. + */ + rte_eth_convert_rx_offload_bitfield(&dev->data->dev_conf.rxmode, + &local_conf.offloads); + } + ret = (*dev->dev_ops->rx_queue_setup)(dev, rx_queue_id, nb_rx_desc, - socket_id, rx_conf, mp); + socket_id, &local_conf, mp); if (!ret) { if (!dev->data->min_rx_buf_size || dev->data->min_rx_buf_size > mbp_buf_size) @@ -1979,7 +2080,8 @@ rte_eth_dev_vlan_filter(uint8_t port_id, uint16_t vlan_id, int on) RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV); dev = &rte_eth_devices[port_id]; - if (!(dev->data->dev_conf.rxmode.hw_vlan_filter)) { + if (!(dev->data->dev_conf.rxmode.offloads & + DEV_RX_OFFLOAD_VLAN_FILTER)) { RTE_PMD_DEBUG_TRACE("port %d: vlan-filtering disabled\n", port_id); return -ENOSYS; } @@ -2055,23 +2157,41 @@ rte_eth_dev_set_vlan_offload(uint8_t port_id, int offload_mask) /*check which option changed by application*/ cur = !!(offload_mask & ETH_VLAN_STRIP_OFFLOAD); - org = !!(dev->data->dev_conf.rxmode.hw_vlan_strip); + org = !!(dev->data->dev_conf.rxmode.offloads & + DEV_RX_OFFLOAD_VLAN_STRIP); if (cur != org) { - dev->data->dev_conf.rxmode.hw_vlan_strip = (uint8_t)cur; + if (cur) + dev->data->dev_conf.rxmode.offloads |= + DEV_RX_OFFLOAD_VLAN_STRIP; + else + dev->data->dev_conf.rxmode.offloads &= + ~DEV_RX_OFFLOAD_VLAN_STRIP; mask |= ETH_VLAN_STRIP_MASK; } cur = !!(offload_mask & ETH_VLAN_FILTER_OFFLOAD); - org = !!(dev->data->dev_conf.rxmode.hw_vlan_filter); + org = !!(dev->data->dev_conf.rxmode.offloads & + DEV_RX_OFFLOAD_VLAN_FILTER); if (cur != org) { - dev->data->dev_conf.rxmode.hw_vlan_filter = (uint8_t)cur; + if (cur) + dev->data->dev_conf.rxmode.offloads |= + DEV_RX_OFFLOAD_VLAN_FILTER; + else + dev->data->dev_conf.rxmode.offloads &= + ~DEV_RX_OFFLOAD_VLAN_FILTER; mask |= ETH_VLAN_FILTER_MASK; } cur = !!(offload_mask & ETH_VLAN_EXTEND_OFFLOAD); - org = !!(dev->data->dev_conf.rxmode.hw_vlan_extend); + org = !!(dev->data->dev_conf.rxmode.offloads & + DEV_RX_OFFLOAD_VLAN_EXTEND); if (cur != org) { - dev->data->dev_conf.rxmode.hw_vlan_extend = (uint8_t)cur; + if (cur) + dev->data->dev_conf.rxmode.offloads |= + DEV_RX_OFFLOAD_VLAN_EXTEND; + else + dev->data->dev_conf.rxmode.offloads &= + ~DEV_RX_OFFLOAD_VLAN_EXTEND; mask |= ETH_VLAN_EXTEND_MASK; } @@ -2080,6 +2200,13 @@ rte_eth_dev_set_vlan_offload(uint8_t port_id, int offload_mask) return ret; RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->vlan_offload_set, -ENOTSUP); + + /* + * Convert to the offload bitfield API just in case the underlying PMD + * still supporting it. + */ + rte_eth_convert_rx_offloads(dev->data->dev_conf.rxmode.offloads, + &dev->data->dev_conf.rxmode); (*dev->dev_ops->vlan_offload_set)(dev, mask); return ret; @@ -2094,13 +2221,16 @@ rte_eth_dev_get_vlan_offload(uint8_t port_id) RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV); dev = &rte_eth_devices[port_id]; - if (dev->data->dev_conf.rxmode.hw_vlan_strip) + if (dev->data->dev_conf.rxmode.offloads & + DEV_RX_OFFLOAD_VLAN_STRIP) ret |= ETH_VLAN_STRIP_OFFLOAD; - if (dev->data->dev_conf.rxmode.hw_vlan_filter) + if (dev->data->dev_conf.rxmode.offloads & + DEV_RX_OFFLOAD_VLAN_FILTER) ret |= ETH_VLAN_FILTER_OFFLOAD; - if (dev->data->dev_conf.rxmode.hw_vlan_extend) + if (dev->data->dev_conf.rxmode.offloads & + DEV_RX_OFFLOAD_VLAN_EXTEND) ret |= ETH_VLAN_EXTEND_OFFLOAD; return ret; diff --git a/lib/librte_ether/rte_ethdev.h b/lib/librte_ether/rte_ethdev.h index 0adf3274a..f424cba04 100644 --- a/lib/librte_ether/rte_ethdev.h +++ b/lib/librte_ether/rte_ethdev.h @@ -348,7 +348,18 @@ struct rte_eth_rxmode { enum rte_eth_rx_mq_mode mq_mode; uint32_t max_rx_pkt_len; /**< Only used if jumbo_frame enabled. */ uint16_t split_hdr_size; /**< hdr buf size (header_split enabled).*/ + uint64_t offloads; + /** + * Per-port Rx offloads to be set using DEV_RX_OFFLOAD_* flags. + * Only offloads set on rx_offload_capa field on rte_eth_dev_info + * structure are allowed to be set. + */ __extension__ + /** + * Below bitfield API is obsolete. Application should + * enable per-port offloads using the offload field + * above. + */ uint16_t header_split : 1, /**< Header Split enable. */ hw_ip_checksum : 1, /**< IP/UDP/TCP checksum offload enable. */ hw_vlan_filter : 1, /**< VLAN filter enable. */ @@ -357,7 +368,17 @@ struct rte_eth_rxmode { jumbo_frame : 1, /**< Jumbo Frame Receipt enable. */ hw_strip_crc : 1, /**< Enable CRC stripping by hardware. */ enable_scatter : 1, /**< Enable scatter packets rx handler */ - enable_lro : 1; /**< Enable LRO */ + enable_lro : 1, /**< Enable LRO */ + ignore_offload_bitfield : 1; + /** + * When set the offload bitfield should be ignored. + * Instead per-port Rx offloads should be set on offloads + * field above. + * Per-queue offloads shuold be set on rte_eth_rxq_conf + * structure. + * This bit is temporary till rxmode bitfield offloads API will + * be deprecated. + */ }; /** @@ -691,6 +712,12 @@ struct rte_eth_rxconf { uint16_t rx_free_thresh; /**< Drives the freeing of RX descriptors. */ uint8_t rx_drop_en; /**< Drop packets if no descriptors are available. */ uint8_t rx_deferred_start; /**< Do not start queue with rte_eth_dev_start(). */ + uint64_t offloads; + /** + * Per-queue Rx offloads to be set using DEV_RX_OFFLOAD_* flags. + * Only offloads set on rx_queue_offload_capa field on rte_eth_dev_info + * structure are allowed to be set. + */ }; #define ETH_TXQ_FLAGS_NOMULTSEGS 0x0001 /**< nb_segs=1 for all mbufs */ @@ -706,6 +733,7 @@ struct rte_eth_rxconf { #define ETH_TXQ_FLAGS_NOXSUMS \ (ETH_TXQ_FLAGS_NOXSUMSCTP | ETH_TXQ_FLAGS_NOXSUMUDP | \ ETH_TXQ_FLAGS_NOXSUMTCP) + /** * A structure used to configure a TX ring of an Ethernet port. */ @@ -907,6 +935,18 @@ struct rte_eth_conf { #define DEV_RX_OFFLOAD_QINQ_STRIP 0x00000020 #define DEV_RX_OFFLOAD_OUTER_IPV4_CKSUM 0x00000040 #define DEV_RX_OFFLOAD_MACSEC_STRIP 0x00000080 +#define DEV_RX_OFFLOAD_HEADER_SPLIT 0x00000100 +#define DEV_RX_OFFLOAD_VLAN_FILTER 0x00000200 +#define DEV_RX_OFFLOAD_VLAN_EXTEND 0x00000400 +#define DEV_RX_OFFLOAD_JUMBO_FRAME 0x00000800 +#define DEV_RX_OFFLOAD_CRC_STRIP 0x00001000 +#define DEV_RX_OFFLOAD_SCATTER 0x00002000 +#define DEV_RX_OFFLOAD_CHECKSUM (DEV_RX_OFFLOAD_IPV4_CKSUM | \ + DEV_RX_OFFLOAD_UDP_CKSUM | \ + DEV_RX_OFFLOAD_TCP_CKSUM) +#define DEV_RX_OFFLOAD_VLAN (DEV_RX_OFFLOAD_VLAN_STRIP | \ + DEV_RX_OFFLOAD_VLAN_FILTER | \ + DEV_RX_OFFLOAD_VLAN_EXTEND) /** * TX offload capabilities of a device. @@ -949,8 +989,11 @@ struct rte_eth_dev_info { /** Maximum number of hash MAC addresses for MTA and UTA. */ uint16_t max_vfs; /**< Maximum number of VFs. */ uint16_t max_vmdq_pools; /**< Maximum number of VMDq pools. */ - uint32_t rx_offload_capa; /**< Device RX offload capabilities. */ + uint64_t rx_offload_capa; + /**< Device per port RX offload capabilities. */ uint32_t tx_offload_capa; /**< Device TX offload capabilities. */ + uint64_t rx_queue_offload_capa; + /**< Device per queue RX offload capabilities. */ uint16_t reta_size; /**< Device redirection table size, the total number of entries. */ uint8_t hash_key_size; /**< Hash key size in bytes */ @@ -1870,6 +1913,9 @@ uint32_t rte_eth_speed_bitflag(uint32_t speed, int duplex); * each statically configurable offload hardware feature provided by * Ethernet devices, such as IP checksum or VLAN tag stripping for * example. + * The Rx offload bitfield API is obsolete and will be deprecated. + * Applications should set the ignore_bitfield_offloads bit on *rxmode* + * structure and use offloads field to set per-port offloads instead. * - the Receive Side Scaling (RSS) configuration when using multiple RX * queues per port. * @@ -1923,6 +1969,8 @@ void _rte_eth_dev_reset(struct rte_eth_dev *dev); * The *rx_conf* structure contains an *rx_thresh* structure with the values * of the Prefetch, Host, and Write-Back threshold registers of the receive * ring. + * In addition it contains the hardware offloads features to activate using + * the DEV_RX_OFFLOAD_* flags. * @param mb_pool * The pointer to the memory pool from which to allocate *rte_mbuf* network * memory buffers to populate each descriptor of the receive ring. -- 2.12.0