From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-we0-f177.google.com (mail-we0-f177.google.com [74.125.82.177]) by dpdk.org (Postfix) with ESMTP id BFB3B5A71 for ; Tue, 13 Jan 2015 11:02:10 +0100 (CET) Received: by mail-we0-f177.google.com with SMTP id q59so1861555wes.8 for ; Tue, 13 Jan 2015 02:02:10 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:message-id:date:from:user-agent:mime-version:to :subject:references:in-reply-to:content-type :content-transfer-encoding; bh=WjFBbxgfmi9bkAoC75gceuK4AxByuEGfCgU7iaozGEA=; b=ANHjT29lZ42GZ7JM17sLs5/gCVq0+QU2NAAIDx8TQHWCyLofRaSK+ldfms1V33QGK7 C7v/tHXkkWznFFqCSAqNUM1rra9I6Fz35cDwdwUf8wmYaan5EP6Hg0N40U7fVKOeWxfH 61U0a80oI0ubApbskrYHSSY1IYZvOuiE28s8CRvch0C8TV6z/YwwipwYIaML4lnBw7AX xZV/AMysStv2HcM8EcCUROmakVdB77Vxoj140gFQKehRV6X6LGuIMWXC53zOXPTsXdyJ SqO2nAKVP4JunjBAuiBF1J7xyrI/9V8mrWGDyP1gFKeiXvJu7566OKwu03L0BrToC2St 6oug== X-Gm-Message-State: ALoCoQlmIyjMQrgziYktBttINCcff8XB2uoraAI3U4I4lorp5uea5HHiw5ganwA3HAnXQzWD9rfz X-Received: by 10.194.20.67 with SMTP id l3mr18894194wje.94.1421143330525; Tue, 13 Jan 2015 02:02:10 -0800 (PST) Received: from [10.0.0.4] ([109.66.137.113]) by mx.google.com with ESMTPSA id bt2sm7306964wib.9.2015.01.13.02.02.09 (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Tue, 13 Jan 2015 02:02:10 -0800 (PST) Message-ID: <54B4ED20.4010905@cloudius-systems.com> Date: Tue, 13 Jan 2015 12:02:08 +0200 From: Vlad Zolotarov User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:31.0) Gecko/20100101 Thunderbird/31.3.0 MIME-Version: 1.0 To: "Jastrzebski, MichalX K" , "dev@dpdk.org" References: <6c3329$jtfs2e@orsmga002.jf.intel.com> <60ABE07DBB3A454EB7FAD707B4BB1582138D2553@IRSMSX109.ger.corp.intel.com> In-Reply-To: <60ABE07DBB3A454EB7FAD707B4BB1582138D2553@IRSMSX109.ger.corp.intel.com> Content-Type: text/plain; charset=windows-1252; format=flowed Content-Transfer-Encoding: 7bit Subject: Re: [dpdk-dev] [PATCH 1/2] pmd: add DCB for VF for ixgbe X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: patches and discussions about DPDK List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 13 Jan 2015 10:02:11 -0000 On 01/12/15 17:46, Jastrzebski, MichalX K wrote: >> -----Original Message----- >> From: dev [mailto:dev-bounces@dpdk.org] On Behalf Of Michal Jastrzebski >> Sent: Monday, January 12, 2015 3:43 PM >> To: dev@dpdk.org >> Subject: [dpdk-dev] [PATCH 1/2] pmd: add DCB for VF for ixgbe >> >> Date: Mon, 12 Jan 2015 15:39:40 +0100 >> Message-Id: <1421073581-6644-2-git-send-email- >> michalx.k.jastrzebski@intel.com> >> X-Mailer: git-send-email 2.1.1 >> In-Reply-To: <1421073581-6644-1-git-send-email- >> michalx.k.jastrzebski@intel.com> >> References: <1421073581-6644-1-git-send-email- >> michalx.k.jastrzebski@intel.com> >> >> From: Pawel Wodkowski >> >> >> This patch add support for DCB in SRIOV mode. When no PFC >> >> is enabled this feature might be used as multiple queues >> >> (up to 8 or 4) for VF. >> >> >> >> It incorporate following modifications: >> >> - Allow zero rx/tx queues to be passed to rte_eth_dev_configure(). >> >> Rationale: >> >> in SRIOV mode PF use first free VF to RX/TX. If VF count >> >> is 16 or 32 all recources are assigned to VFs so PF can >> >> be used only for configuration. >> >> - split nb_q_per_pool to nb_rx_q_per_pool and nb_tx_q_per_pool >> >> Rationale: >> >> rx and tx number of queue might be different if RX and TX are >> >> configured in different mode. This allow to inform VF about >> >> proper number of queues. >> >> - extern mailbox API for DCB mode >> >> >> >> Signed-off-by: Pawel Wodkowski >> >> --- >> >> lib/librte_ether/rte_ethdev.c | 84 +++++++++++++++++++++--------- >> >> lib/librte_ether/rte_ethdev.h | 5 +- >> >> lib/librte_pmd_e1000/igb_pf.c | 3 +- >> >> lib/librte_pmd_ixgbe/ixgbe_ethdev.c | 10 ++-- >> >> lib/librte_pmd_ixgbe/ixgbe_ethdev.h | 1 + >> >> lib/librte_pmd_ixgbe/ixgbe_pf.c | 98 ++++++++++++++++++++++++++++++-- >> --- >> >> lib/librte_pmd_ixgbe/ixgbe_rxtx.c | 7 ++- >> >> 7 files changed, 159 insertions(+), 49 deletions(-) >> >> >> >> diff --git a/lib/librte_ether/rte_ethdev.c b/lib/librte_ether/rte_ethdev.c >> >> index 95f2ceb..4c1a494 100644 >> >> --- a/lib/librte_ether/rte_ethdev.c >> >> +++ b/lib/librte_ether/rte_ethdev.c >> >> @@ -333,7 +333,7 @@ rte_eth_dev_rx_queue_config(struct rte_eth_dev >> *dev, uint16_t nb_queues) >> >> dev->data->rx_queues = rte_zmalloc("ethdev->rx_queues", >> >> sizeof(dev->data->rx_queues[0]) * nb_queues, >> >> RTE_CACHE_LINE_SIZE); >> >> - if (dev->data->rx_queues == NULL) { >> >> + if (dev->data->rx_queues == NULL && nb_queues > 0) { >> >> dev->data->nb_rx_queues = 0; >> >> return -(ENOMEM); >> >> } >> >> @@ -475,7 +475,7 @@ rte_eth_dev_tx_queue_config(struct rte_eth_dev >> *dev, uint16_t nb_queues) >> >> dev->data->tx_queues = rte_zmalloc("ethdev->tx_queues", >> >> sizeof(dev->data->tx_queues[0]) * nb_queues, >> >> RTE_CACHE_LINE_SIZE); >> >> - if (dev->data->tx_queues == NULL) { >> >> + if (dev->data->tx_queues == NULL && nb_queues > 0) { >> >> dev->data->nb_tx_queues = 0; >> >> return -(ENOMEM); >> >> } >> >> @@ -507,6 +507,7 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, >> uint16_t nb_rx_q, uint16_t nb_tx_q, >> >> const struct rte_eth_conf *dev_conf) >> >> { >> >> struct rte_eth_dev *dev = &rte_eth_devices[port_id]; >> >> + struct rte_eth_dev_info dev_info; >> >> >> >> if (RTE_ETH_DEV_SRIOV(dev).active != 0) { >> >> /* check multi-queue mode */ >> >> @@ -524,11 +525,33 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, >> uint16_t nb_rx_q, uint16_t nb_tx_q, >> >> return (-EINVAL); >> >> } >> >> >> >> + if ((dev_conf->rxmode.mq_mode == >> ETH_MQ_RX_VMDQ_DCB) && >> >> + (dev_conf->txmode.mq_mode == >> ETH_MQ_TX_VMDQ_DCB)) { >> >> + enum rte_eth_nb_pools rx_pools = >> >> + dev_conf- >>> rx_adv_conf.vmdq_dcb_conf.nb_queue_pools; >> + enum rte_eth_nb_pools tx_pools = >> >> + dev_conf- >>> tx_adv_conf.vmdq_dcb_tx_conf.nb_queue_pools; >> + >> >> + if (rx_pools != tx_pools) { >> >> + /* Only equal number of pools is supported >> when >> >> + * DCB+VMDq in SRIOV */ >> >> + PMD_DEBUG_TRACE("ethdev port_id=%" >> PRIu8 >> >> + " SRIOV active, DCB+VMDQ >> mode, " >> >> + "number of rx and tx pools is >> not eqaul\n", >> >> + port_id); >> >> + return (-EINVAL); >> >> + } >> >> + } >> >> + >> >> + uint16_t nb_rx_q_per_pool = >> RTE_ETH_DEV_SRIOV(dev).nb_rx_q_per_pool; >> >> + uint16_t nb_tx_q_per_pool = >> RTE_ETH_DEV_SRIOV(dev).nb_tx_q_per_pool; >> >> + >> >> switch (dev_conf->rxmode.mq_mode) { >> >> - case ETH_MQ_RX_VMDQ_RSS: >> >> case ETH_MQ_RX_VMDQ_DCB: >> >> + break; >> >> + case ETH_MQ_RX_VMDQ_RSS: >> >> case ETH_MQ_RX_VMDQ_DCB_RSS: >> >> - /* DCB/RSS VMDQ in SRIOV mode, not implement yet >> */ >> >> + /* RSS, DCB+RSS VMDQ in SRIOV mode, not >> implement yet */ >> >> PMD_DEBUG_TRACE("ethdev port_id=%" PRIu8 >> >> " SRIOV active, " >> >> "unsupported VMDQ mq_mode rx >> %u\n", >> >> @@ -537,37 +560,32 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, >> uint16_t nb_rx_q, uint16_t nb_tx_q, >> >> default: /* ETH_MQ_RX_VMDQ_ONLY or ETH_MQ_RX_NONE >> */ >> >> /* if nothing mq mode configure, use default scheme >> */ >> >> dev->data->dev_conf.rxmode.mq_mode = >> ETH_MQ_RX_VMDQ_ONLY; >> >> - if (RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool > 1) >> >> - RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool = >> 1; >> >> + if (nb_rx_q_per_pool > 1) >> >> + nb_rx_q_per_pool = 1; >> >> break; >> >> } >> >> >> >> switch (dev_conf->txmode.mq_mode) { >> >> - case ETH_MQ_TX_VMDQ_DCB: >> >> - /* DCB VMDQ in SRIOV mode, not implement yet */ >> >> - PMD_DEBUG_TRACE("ethdev port_id=%" PRIu8 >> >> - " SRIOV active, " >> >> - "unsupported VMDQ mq_mode tx >> %u\n", >> >> - port_id, dev_conf- >>> txmode.mq_mode); >> - return (-EINVAL); >> >> + case ETH_MQ_TX_VMDQ_DCB: /* DCB VMDQ in SRIOV >> mode*/ >> >> + break; >> >> default: /* ETH_MQ_TX_VMDQ_ONLY or ETH_MQ_TX_NONE >> */ >> >> /* if nothing mq mode configure, use default scheme >> */ >> >> dev->data->dev_conf.txmode.mq_mode = >> ETH_MQ_TX_VMDQ_ONLY; >> >> - if (RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool > 1) >> >> - RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool = >> 1; >> >> + if (nb_tx_q_per_pool > 1) >> >> + nb_tx_q_per_pool = 1; >> >> break; >> >> } >> >> >> >> /* check valid queue number */ >> >> - if ((nb_rx_q > RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool) || >> >> - (nb_tx_q > RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool)) { >> >> + if (nb_rx_q > nb_rx_q_per_pool || nb_tx_q > >> nb_tx_q_per_pool) { >> >> PMD_DEBUG_TRACE("ethdev port_id=%d SRIOV >> active, " >> >> - "queue number must less equal to %d\n", >> >> - port_id, >> RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool); >> >> + "rx/tx queue number must less equal to >> %d/%d\n", >> >> + port_id, >> RTE_ETH_DEV_SRIOV(dev).nb_rx_q_per_pool, >> >> + >> RTE_ETH_DEV_SRIOV(dev).nb_tx_q_per_pool); >> >> return (-EINVAL); >> >> } >> >> } else { >> >> - /* For vmdb+dcb mode check our configuration before we go >> further */ >> >> + /* For vmdq+dcb mode check our configuration before we go >> further */ >> >> if (dev_conf->rxmode.mq_mode == ETH_MQ_RX_VMDQ_DCB) >> { >> >> const struct rte_eth_vmdq_dcb_conf *conf; >> >> >> >> @@ -606,11 +624,20 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, >> uint16_t nb_rx_q, uint16_t nb_tx_q, >> >> } >> >> } >> >> >> >> + /* For DCB we need to obtain maximum number of queues >> dinamically, >> >> + * as this depends on max VF exported in PF */ >> >> + if ((dev_conf->rxmode.mq_mode == ETH_MQ_RX_DCB) || >> >> + (dev_conf->txmode.mq_mode == ETH_MQ_TX_DCB)) >> { >> >> + >> >> + FUNC_PTR_OR_ERR_RET(*dev->dev_ops- >>> dev_infos_get, -ENOTSUP); >> + (*dev->dev_ops->dev_infos_get)(dev, >> &dev_info); >> >> + } >> >> + >> >> /* For DCB mode check our configuration before we go further >> */ >> >> if (dev_conf->rxmode.mq_mode == ETH_MQ_RX_DCB) { >> >> const struct rte_eth_dcb_rx_conf *conf; >> >> >> >> - if (nb_rx_q != ETH_DCB_NUM_QUEUES) { >> >> + if (nb_rx_q != dev_info.max_rx_queues) { >> >> PMD_DEBUG_TRACE("ethdev port_id=%d >> DCB, nb_rx_q " >> >> "!= %d\n", >> >> port_id, >> ETH_DCB_NUM_QUEUES); >> >> @@ -630,7 +657,7 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, >> uint16_t nb_rx_q, uint16_t nb_tx_q, >> >> if (dev_conf->txmode.mq_mode == ETH_MQ_TX_DCB) { >> >> const struct rte_eth_dcb_tx_conf *conf; >> >> >> >> - if (nb_tx_q != ETH_DCB_NUM_QUEUES) { >> >> + if (nb_tx_q != dev_info.max_tx_queues) { >> >> PMD_DEBUG_TRACE("ethdev port_id=%d >> DCB, nb_tx_q " >> >> "!= %d\n", >> >> port_id, >> ETH_DCB_NUM_QUEUES); >> >> @@ -690,7 +717,10 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t >> nb_rx_q, uint16_t nb_tx_q, >> >> } >> >> if (nb_rx_q == 0) { >> >> PMD_DEBUG_TRACE("ethdev port_id=%d nb_rx_q == 0\n", >> port_id); >> >> - return (-EINVAL); >> >> + /* In SRIOV there can be no free resource for PF. So permit use >> only >> >> + * for configuration. */ >> >> + if (RTE_ETH_DEV_SRIOV(dev).active == 0) >> >> + return (-EINVAL); >> >> } >> >> >> >> if (nb_tx_q > dev_info.max_tx_queues) { >> >> @@ -698,9 +728,13 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t >> nb_rx_q, uint16_t nb_tx_q, >> >> port_id, nb_tx_q, dev_info.max_tx_queues); >> >> return (-EINVAL); >> >> } >> >> + >> >> if (nb_tx_q == 0) { >> >> PMD_DEBUG_TRACE("ethdev port_id=%d nb_tx_q == 0\n", >> port_id); >> >> - return (-EINVAL); >> >> + /* In SRIOV there can be no free resource for PF. So permit use >> only >> >> + * for configuration. */ >> >> + if (RTE_ETH_DEV_SRIOV(dev).active == 0) >> >> + return (-EINVAL); >> >> } >> >> >> >> /* Copy the dev_conf parameter into the dev structure */ >> >> @@ -750,7 +784,7 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t >> nb_rx_q, uint16_t nb_tx_q, >> >> ETHER_MAX_LEN; >> >> } >> >> >> >> - /* multipe queue mode checking */ >> >> + /* multiple queue mode checking */ >> >> diag = rte_eth_dev_check_mq_mode(port_id, nb_rx_q, nb_tx_q, >> dev_conf); >> >> if (diag != 0) { >> >> PMD_DEBUG_TRACE("port%d rte_eth_dev_check_mq_mode = >> %d\n", >> >> diff --git a/lib/librte_ether/rte_ethdev.h b/lib/librte_ether/rte_ethdev.h >> >> index ce0528f..04fda83 100644 >> >> --- a/lib/librte_ether/rte_ethdev.h >> >> +++ b/lib/librte_ether/rte_ethdev.h >> >> @@ -299,7 +299,7 @@ enum rte_eth_rx_mq_mode { >> >> enum rte_eth_tx_mq_mode { >> >> ETH_MQ_TX_NONE = 0, /**< It is in neither DCB nor VT mode. */ >> >> ETH_MQ_TX_DCB, /**< For TX side,only DCB is on. */ >> >> - ETH_MQ_TX_VMDQ_DCB, /**< For TX side,both DCB and VT is >> on. */ >> >> + ETH_MQ_TX_VMDQ_DCB, /**< For TX side,both DCB and VT is on. >> */ >> >> ETH_MQ_TX_VMDQ_ONLY, /**< Only VT on, no DCB */ >> >> }; >> >> >> >> @@ -1569,7 +1569,8 @@ struct rte_eth_dev { >> >> >> >> struct rte_eth_dev_sriov { >> >> uint8_t active; /**< SRIOV is active with 16, 32 or 64 pools */ >> >> - uint8_t nb_q_per_pool; /**< rx queue number per pool */ >> >> + uint8_t nb_rx_q_per_pool; /**< rx queue number per pool */ >> >> + uint8_t nb_tx_q_per_pool; /**< tx queue number per pool */ >> >> uint16_t def_vmdq_idx; /**< Default pool num used for PF */ >> >> uint16_t def_pool_q_idx; /**< Default pool queue start reg index */ >> >> }; >> >> diff --git a/lib/librte_pmd_e1000/igb_pf.c b/lib/librte_pmd_e1000/igb_pf.c >> >> index bc3816a..9d2f858 100644 >> >> --- a/lib/librte_pmd_e1000/igb_pf.c >> >> +++ b/lib/librte_pmd_e1000/igb_pf.c >> >> @@ -115,7 +115,8 @@ void igb_pf_host_init(struct rte_eth_dev *eth_dev) >> >> rte_panic("Cannot allocate memory for private VF data\n"); >> >> >> >> RTE_ETH_DEV_SRIOV(eth_dev).active = ETH_8_POOLS; >> >> - RTE_ETH_DEV_SRIOV(eth_dev).nb_q_per_pool = nb_queue; >> >> + RTE_ETH_DEV_SRIOV(eth_dev).nb_rx_q_per_pool = nb_queue; >> >> + RTE_ETH_DEV_SRIOV(eth_dev).nb_tx_q_per_pool = nb_queue; >> >> RTE_ETH_DEV_SRIOV(eth_dev).def_vmdq_idx = vf_num; >> >> RTE_ETH_DEV_SRIOV(eth_dev).def_pool_q_idx = (uint16_t)(vf_num * >> nb_queue); >> >> >> >> diff --git a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c >> b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c >> >> index 3fc3738..347f03c 100644 >> >> --- a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c >> >> +++ b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c >> >> @@ -3555,14 +3555,14 @@ static int ixgbe_set_vf_rate_limit(struct >> rte_eth_dev *dev, uint16_t vf, >> >> struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data- >>> dev_private); >> struct ixgbe_vf_info *vfinfo = >> >> *(IXGBE_DEV_PRIVATE_TO_P_VFDATA(dev->data- >>> dev_private)); >> - uint8_t nb_q_per_pool = RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool; >> >> + uint8_t nb_tx_q_per_pool = >> RTE_ETH_DEV_SRIOV(dev).nb_tx_q_per_pool; >> >> uint32_t queue_stride = >> >> IXGBE_MAX_RX_QUEUE_NUM / >> RTE_ETH_DEV_SRIOV(dev).active; >> >> uint32_t queue_idx = vf * queue_stride, idx = 0, vf_idx; >> >> - uint32_t queue_end = queue_idx + nb_q_per_pool - 1; >> >> + uint32_t tx_queue_end = queue_idx + nb_tx_q_per_pool - 1; >> >> uint16_t total_rate = 0; >> >> >> >> - if (queue_end >= hw->mac.max_tx_queues) >> >> + if (tx_queue_end >= hw->mac.max_tx_queues) >> >> return -EINVAL; >> >> >> >> if (vfinfo != NULL) { >> >> @@ -3577,7 +3577,7 @@ static int ixgbe_set_vf_rate_limit(struct rte_eth_dev >> *dev, uint16_t vf, >> >> return -EINVAL; >> >> >> >> /* Store tx_rate for this vf. */ >> >> - for (idx = 0; idx < nb_q_per_pool; idx++) { >> >> + for (idx = 0; idx < nb_tx_q_per_pool; idx++) { >> >> if (((uint64_t)0x1 << idx) & q_msk) { >> >> if (vfinfo[vf].tx_rate[idx] != tx_rate) >> >> vfinfo[vf].tx_rate[idx] = tx_rate; >> >> @@ -3595,7 +3595,7 @@ static int ixgbe_set_vf_rate_limit(struct rte_eth_dev >> *dev, uint16_t vf, >> >> } >> >> >> >> /* Set RTTBCNRC of each queue/pool for vf X */ >> >> - for (; queue_idx <= queue_end; queue_idx++) { >> >> + for (; queue_idx <= tx_queue_end; queue_idx++) { >> >> if (0x1 & q_msk) >> >> ixgbe_set_queue_rate_limit(dev, queue_idx, tx_rate); >> >> q_msk = q_msk >> 1; >> >> diff --git a/lib/librte_pmd_ixgbe/ixgbe_ethdev.h >> b/lib/librte_pmd_ixgbe/ixgbe_ethdev.h >> >> index ca99170..ebf16e9 100644 >> >> --- a/lib/librte_pmd_ixgbe/ixgbe_ethdev.h >> >> +++ b/lib/librte_pmd_ixgbe/ixgbe_ethdev.h >> >> @@ -159,6 +159,7 @@ struct ixgbe_vf_info { >> >> uint16_t tx_rate[IXGBE_MAX_QUEUE_NUM_PER_VF]; >> >> uint16_t vlan_count; >> >> uint8_t spoofchk_enabled; >> >> + unsigned int vf_api; >> >> }; >> >> >> >> /* >> >> diff --git a/lib/librte_pmd_ixgbe/ixgbe_pf.c b/lib/librte_pmd_ixgbe/ixgbe_pf.c >> >> index 51da1fd..4d30bcf 100644 >> >> --- a/lib/librte_pmd_ixgbe/ixgbe_pf.c >> >> +++ b/lib/librte_pmd_ixgbe/ixgbe_pf.c >> >> @@ -127,7 +127,8 @@ void ixgbe_pf_host_init(struct rte_eth_dev *eth_dev) >> >> RTE_ETH_DEV_SRIOV(eth_dev).active = ETH_16_POOLS; >> >> } >> >> >> >> - RTE_ETH_DEV_SRIOV(eth_dev).nb_q_per_pool = nb_queue; >> >> + RTE_ETH_DEV_SRIOV(eth_dev).nb_rx_q_per_pool = nb_queue; >> >> + RTE_ETH_DEV_SRIOV(eth_dev).nb_tx_q_per_pool = nb_queue; >> >> RTE_ETH_DEV_SRIOV(eth_dev).def_vmdq_idx = vf_num; >> >> RTE_ETH_DEV_SRIOV(eth_dev).def_pool_q_idx = (uint16_t)(vf_num * >> nb_queue); >> >> >> >> @@ -189,7 +190,7 @@ int ixgbe_pf_host_configure(struct rte_eth_dev >> *eth_dev) >> >> hw->mac.ops.set_vmdq(hw, 0, >> RTE_ETH_DEV_SRIOV(eth_dev).def_vmdq_idx); >> >> >> >> /* >> >> - * SW msut set GCR_EXT.VT_Mode the same as GPIE.VT_Mode >> >> + * SW must set GCR_EXT.VT_Mode the same as GPIE.VT_Mode >> >> */ >> >> gcr_ext = IXGBE_READ_REG(hw, IXGBE_GCR_EXT); >> >> gcr_ext &= ~IXGBE_GCR_EXT_VT_MODE_MASK; >> >> @@ -214,19 +215,19 @@ int ixgbe_pf_host_configure(struct rte_eth_dev >> *eth_dev) >> >> } >> >> >> >> IXGBE_WRITE_REG(hw, IXGBE_GCR_EXT, gcr_ext); >> >> - IXGBE_WRITE_REG(hw, IXGBE_GPIE, gpie); >> >> + IXGBE_WRITE_REG(hw, IXGBE_GPIE, gpie); >> >> >> >> - /* >> >> + /* >> >> * enable vlan filtering and allow all vlan tags through >> >> */ >> >> - vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); >> >> - vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */ >> >> - IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl); >> >> + vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); >> >> + vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */ >> >> + IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl); >> >> >> >> - /* VFTA - enable all vlan filters */ >> >> - for (i = 0; i < IXGBE_MAX_VFTA; i++) { >> >> - IXGBE_WRITE_REG(hw, IXGBE_VFTA(i), 0xFFFFFFFF); >> >> - } >> >> + /* VFTA - enable all vlan filters */ >> >> + for (i = 0; i < IXGBE_MAX_VFTA; i++) { >> >> + IXGBE_WRITE_REG(hw, IXGBE_VFTA(i), 0xFFFFFFFF); >> >> + } >> >> >> >> /* Enable MAC Anti-Spoofing */ >> >> hw->mac.ops.set_mac_anti_spoofing(hw, FALSE, vf_num); >> >> @@ -369,6 +370,73 @@ ixgbe_vf_reset(struct rte_eth_dev *dev, uint16_t vf, >> uint32_t *msgbuf) >> >> } >> >> >> >> static int >> >> +ixgbe_negotiate_vf_api(struct rte_eth_dev *dev, uint32_t vf, uint32_t >> *msgbuf) >> >> +{ >> >> + struct ixgbe_vf_info *vfinfo = >> >> + *(IXGBE_DEV_PRIVATE_TO_P_VFDATA(dev->data- >>> dev_private)); >> + int api = msgbuf[1]; >> >> + >> >> + switch (api) { >> >> + case ixgbe_mbox_api_10: >> >> + case ixgbe_mbox_api_11: >> >> + vfinfo[vf].vf_api = api; >> >> + return 0; >> >> + default: >> >> + break; >> >> + } >> >> + >> >> + RTE_LOG(DEBUG, PMD, "VF %d requested invalid api version %u\n", vf, >> api); >> >> + return -1; >> >> +} >> >> + >> >> +static int >> >> +ixgbe_get_vf_queues(struct rte_eth_dev *dev, uint32_t vf, uint32_t *msgbuf) >> >> +{ >> >> + struct ixgbe_vf_info *vfinfo = >> >> + *(IXGBE_DEV_PRIVATE_TO_P_VFDATA(dev->data- >>> dev_private)); >> + struct ixgbe_dcb_config *dcb_cfg = >> >> + IXGBE_DEV_PRIVATE_TO_DCB_CFG(dev->data- >>> dev_private); >> + >> >> + uint8_t num_tcs = dcb_cfg->num_tcs.pg_tcs; >> >> + >> >> + /* verify the PF is supporting the correct APIs */ >> >> + switch (vfinfo[vf].vf_api) { >> >> + case ixgbe_mbox_api_10: >> >> + case ixgbe_mbox_api_11: >> >> + break; >> >> + default: >> >> + return -1; >> >> + } >> >> + >> >> + if (RTE_ETH_DEV_SRIOV(dev).active) { >> >> + if (dev->data->dev_conf.rxmode.mq_mode == >> ETH_MQ_RX_VMDQ_DCB) >> >> + msgbuf[IXGBE_VF_TX_QUEUES] = num_tcs; >> >> + else >> >> + msgbuf[IXGBE_VF_TX_QUEUES] = 1; >> >> + >> >> + if (dev->data->dev_conf.txmode.mq_mode == >> ETH_MQ_TX_VMDQ_DCB) >> >> + msgbuf[IXGBE_VF_RX_QUEUES] = num_tcs; >> >> + else >> >> + msgbuf[IXGBE_VF_RX_QUEUES] = 1; >> >> + } else { >> >> + /* only allow 1 Tx queue for bandwidth limiting */ >> >> + msgbuf[IXGBE_VF_TX_QUEUES] = 1; >> >> + msgbuf[IXGBE_VF_RX_QUEUES] = 1; >> >> + } >> >> + >> >> + /* notify VF of need for VLAN tag stripping, and correct queue */ >> >> + if (num_tcs) >> >> + msgbuf[IXGBE_VF_TRANS_VLAN] = num_tcs; >> >> + else >> >> + msgbuf[IXGBE_VF_TRANS_VLAN] = 0; >> >> + >> >> + /* notify VF of default queue */ >> >> + msgbuf[IXGBE_VF_DEF_QUEUE] = 0; >> >> + >> >> + return 0; >> >> +} >> >> + >> >> +static int >> >> ixgbe_vf_set_mac_addr(struct rte_eth_dev *dev, uint32_t vf, uint32_t >> *msgbuf) >> >> { >> >> struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data- >>> dev_private); >> @@ -512,6 +580,12 @@ ixgbe_rcv_msg_from_vf(struct rte_eth_dev *dev, >> uint16_t vf) >> >> case IXGBE_VF_SET_VLAN: >> >> retval = ixgbe_vf_set_vlan(dev, vf, msgbuf); >> >> break; >> >> + case IXGBE_VF_API_NEGOTIATE: >> >> + retval = ixgbe_negotiate_vf_api(dev, vf, msgbuf); >> >> + break; >> >> + case IXGBE_VF_GET_QUEUES: >> >> + retval = ixgbe_get_vf_queues(dev, vf, msgbuf); >> >> + break; >> >> default: >> >> PMD_DRV_LOG(DEBUG, "Unhandled Msg %8.8x", >> (unsigned)msgbuf[0]); >> >> retval = IXGBE_ERR_MBX; >> >> @@ -526,7 +600,7 @@ ixgbe_rcv_msg_from_vf(struct rte_eth_dev *dev, >> uint16_t vf) >> >> >> >> msgbuf[0] |= IXGBE_VT_MSGTYPE_CTS; >> >> >> >> - ixgbe_write_mbx(hw, msgbuf, 1, vf); >> >> + ixgbe_write_mbx(hw, msgbuf, mbx_size, vf); >> >> >> >> return retval; >> >> } >> >> diff --git a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c >> b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c >> >> index e10d6a2..49b44fe 100644 >> >> --- a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c >> >> +++ b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c >> >> @@ -3166,10 +3166,9 @@ void ixgbe_configure_dcb(struct rte_eth_dev *dev) >> >> >> >> /* check support mq_mode for DCB */ >> >> if ((dev_conf->rxmode.mq_mode != ETH_MQ_RX_VMDQ_DCB) && >> >> - (dev_conf->rxmode.mq_mode != ETH_MQ_RX_DCB)) >> >> - return; >> >> - >> >> - if (dev->data->nb_rx_queues != ETH_DCB_NUM_QUEUES) >> >> + (dev_conf->rxmode.mq_mode != ETH_MQ_RX_DCB) && >> >> + (dev_conf->txmode.mq_mode != ETH_MQ_TX_VMDQ_DCB) && >> >> + (dev_conf->txmode.mq_mode != ETH_MQ_TX_DCB)) >> >> return; >> >> >> >> /** Configure DCB hardware **/ >> >> -- >> >> 1.7.9.5 >> >> > Self nacked - because of wrong message format. Yeah, there is something really wrong with this email formatting.... ;) Note that since u (i guess) haven't used 'git send-email' for this series - it doesn't look like a series (at least in my thunderbird).