From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-yh0-f49.google.com (mail-yh0-f49.google.com [209.85.213.49]) by dpdk.org (Postfix) with ESMTP id C74935941 for ; Tue, 2 Sep 2014 17:15:25 +0200 (CEST) Received: by mail-yh0-f49.google.com with SMTP id z6so4381709yhz.36 for ; Tue, 02 Sep 2014 08:19:57 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:date :message-id:subject:from:to:cc:content-type; bh=xfTCO9s53Z11GxcexhlrZNIMqngVvxabaAXWaD+PhDA=; b=MtwHelD4zT/2ahQfTuCjpi2bkGyTU+g4NDZkkfJld4NjxbR9BGQ1sFVedpxxHIVn7o dflFvLL6/O1CI9bktB32lb0TdeQcTz4imh/cQWTfFaCiQ5TEZPpa2evEY9Ek25cTlwW6 sBHBHz1ylFl2YFOF3Z1YgCJxxcuu5QVfz2/77XA708CLdsn2Sz5kCcvQdg8XU0cleKFP RDJadmFSMURFUQW3+5LNzP4DzFFCCEmWETMGBID8juM2MnjIpS1iX8OU35x6odmZc01p ZBL6j7N2yXsYeeMLCB4qSZ8s7j3xm+iyzTa1aAGpROo1Wr8Y5awPax+AylLikgnjejLN rv3g== X-Gm-Message-State: ALoCoQn0bdlOk7KeHh1r0B+PGs1Y7LzXZPhc2QmBD846PClY0e2NWCLsbDi6EKL9xiyhAPLCMkPb MIME-Version: 1.0 X-Received: by 10.236.50.198 with SMTP id z46mr4256964yhb.115.1409671197624; Tue, 02 Sep 2014 08:19:57 -0700 (PDT) Received: by 10.170.96.213 with HTTP; Tue, 2 Sep 2014 08:19:57 -0700 (PDT) In-Reply-To: <1409567080-27083-4-git-send-email-david.marchand@6wind.com> References: <1409567080-27083-1-git-send-email-david.marchand@6wind.com> <1409567080-27083-4-git-send-email-david.marchand@6wind.com> Date: Tue, 2 Sep 2014 10:19:57 -0500 Message-ID: From: Jay Rolette To: David Marchand Content-Type: text/plain; charset=UTF-8 X-Content-Filtered-By: Mailman/MimeDel 2.1.15 Cc: "dev@dpdk.org" Subject: Re: [dpdk-dev] [PATCH v2 03/17] ixgbe: clean log messages X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: patches and discussions about DPDK List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 02 Sep 2014 15:15:26 -0000 Looks good On Mon, Sep 1, 2014 at 5:24 AM, David Marchand wrote: > Clean log messages: > - remove leading \n in some messages, > - remove trailing \n in some messages, > - split multi lines messages, > - replace some PMD_INIT_LOG(DEBUG, "some_func") with PMD_INIT_FUNC_TRACE(). > > Signed-off-by: David Marchand > --- > lib/librte_pmd_ixgbe/ixgbe_ethdev.c | 150 > +++++++++++++++++------------------ > lib/librte_pmd_ixgbe/ixgbe_fdir.c | 6 +- > lib/librte_pmd_ixgbe/ixgbe_rxtx.c | 93 +++++++++++----------- > 3 files changed, 124 insertions(+), 125 deletions(-) > > diff --git a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c > b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c > index a8a7ed6..1419494 100644 > --- a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c > +++ b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c > @@ -547,12 +547,12 @@ ixgbe_dev_queue_stats_mapping_set(struct rte_eth_dev > *eth_dev, > if ((hw->mac.type != ixgbe_mac_82599EB) && (hw->mac.type != > ixgbe_mac_X540)) > return -ENOSYS; > > - PMD_INIT_LOG(INFO, "Setting port %d, %s queue_id %d to stat index > %d\n", > + PMD_INIT_LOG(INFO, "Setting port %d, %s queue_id %d to stat index > %d", > (int)(eth_dev->data->port_id), is_rx ? "RX" : "TX", > queue_id, stat_idx); > > n = (uint8_t)(queue_id / NB_QMAP_FIELDS_PER_QSM_REG); > if (n >= IXGBE_NB_STAT_MAPPING_REGS) { > - PMD_INIT_LOG(ERR, "Nb of stat mapping registers > exceeded\n"); > + PMD_INIT_LOG(ERR, "Nb of stat mapping registers exceeded"); > return -EIO; > } > offset = (uint8_t)(queue_id % NB_QMAP_FIELDS_PER_QSM_REG); > @@ -572,19 +572,20 @@ ixgbe_dev_queue_stats_mapping_set(struct rte_eth_dev > *eth_dev, > else > stat_mappings->rqsmr[n] |= qsmr_mask; > > - PMD_INIT_LOG(INFO, "Set port %d, %s queue_id %d to stat index %d\n" > - "%s[%d] = 0x%08x\n", > - (int)(eth_dev->data->port_id), is_rx ? "RX" : "TX", > queue_id, stat_idx, > - is_rx ? "RQSMR" : "TQSM",n, is_rx ? > stat_mappings->rqsmr[n] : stat_mappings->tqsm[n]); > + PMD_INIT_LOG(INFO, "Set port %d, %s queue_id %d to stat index %d", > + (int)(eth_dev->data->port_id), is_rx ? "RX" : "TX", > + queue_id, stat_idx); > + PMD_INIT_LOG(INFO, "%s[%d] = 0x%08x", is_rx ? "RQSMR" : "TQSM", n, > + is_rx ? stat_mappings->rqsmr[n] : > stat_mappings->tqsm[n]); > > /* Now write the mapping in the appropriate register */ > if (is_rx) { > - PMD_INIT_LOG(INFO, "Write 0x%x to RX IXGBE stat mapping > reg:%d\n", > + PMD_INIT_LOG(INFO, "Write 0x%x to RX IXGBE stat mapping > reg:%d", > stat_mappings->rqsmr[n], n); > IXGBE_WRITE_REG(hw, IXGBE_RQSMR(n), > stat_mappings->rqsmr[n]); > } > else { > - PMD_INIT_LOG(INFO, "Write 0x%x to TX IXGBE stat mapping > reg:%d\n", > + PMD_INIT_LOG(INFO, "Write 0x%x to TX IXGBE stat mapping > reg:%d", > stat_mappings->tqsm[n], n); > IXGBE_WRITE_REG(hw, IXGBE_TQSM(n), stat_mappings->tqsm[n]); > } > @@ -790,12 +791,13 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct > eth_driver *eth_drv, > > if (diag == IXGBE_ERR_EEPROM_VERSION) { > PMD_INIT_LOG(ERR, "This device is a pre-production > adapter/" > - "LOM. Please be aware there may be issues associated " > - "with your hardware.\n If you are experiencing > problems " > + "LOM. Please be aware there may be issues associated " > + "with your hardware."); > + PMD_INIT_LOG(ERR, "If you are experiencing problems " > "please contact your Intel or hardware representative " > - "who provided you with this hardware.\n"); > + "who provided you with this hardware."); > } else if (diag == IXGBE_ERR_SFP_NOT_SUPPORTED) > - PMD_INIT_LOG(ERR, "Unsupported SFP+ Module\n"); > + PMD_INIT_LOG(ERR, "Unsupported SFP+ Module"); > if (diag) { > PMD_INIT_LOG(ERR, "Hardware Initialization Failure: %d", > diag); > return -EIO; > @@ -811,10 +813,9 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct > eth_driver *eth_drv, > eth_dev->data->mac_addrs = rte_zmalloc("ixgbe", ETHER_ADDR_LEN * > hw->mac.num_rar_entries, 0); > if (eth_dev->data->mac_addrs == NULL) { > - PMD_INIT_LOG(ERR, > - "Failed to allocate %u bytes needed to store " > - "MAC addresses", > - ETHER_ADDR_LEN * hw->mac.num_rar_entries); > + PMD_INIT_LOG(ERR, "Failed to allocate %u bytes needed to > store " > + "MAC addresses", > + ETHER_ADDR_LEN * hw->mac.num_rar_entries); > return -ENOMEM; > } > /* Copy the permanent MAC address */ > @@ -825,9 +826,9 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct > eth_driver *eth_drv, > eth_dev->data->hash_mac_addrs = rte_zmalloc("ixgbe", > ETHER_ADDR_LEN * > IXGBE_VMDQ_NUM_UC_MAC, 0); > if (eth_dev->data->hash_mac_addrs == NULL) { > - PMD_INIT_LOG(ERR, > - "Failed to allocate %d bytes needed to store MAC > addresses", > - ETHER_ADDR_LEN * IXGBE_VMDQ_NUM_UC_MAC); > + PMD_INIT_LOG(ERR, "Failed to allocate %d bytes needed to > store " > + "MAC addresses", > + ETHER_ADDR_LEN * IXGBE_VMDQ_NUM_UC_MAC); > return -ENOMEM; > } > > @@ -849,12 +850,11 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct > eth_driver *eth_drv, > IXGBE_WRITE_FLUSH(hw); > > if (ixgbe_is_sfp(hw) && hw->phy.sfp_type != > ixgbe_sfp_type_not_present) > - PMD_INIT_LOG(DEBUG, > - "MAC: %d, PHY: %d, SFP+: %d + PMD_INIT_LOG(DEBUG, "MAC: %d, PHY: %d, SFP+: %d", > (int) hw->mac.type, (int) hw->phy.type, > (int) hw->phy.sfp_type); > else > - PMD_INIT_LOG(DEBUG, "MAC: %d, PHY: %d\n", > + PMD_INIT_LOG(DEBUG, "MAC: %d, PHY: %d", > (int) hw->mac.type, (int) hw->phy.type); > > PMD_INIT_LOG(DEBUG, "port %d vendorID=0x%x deviceID=0x%x", > @@ -933,7 +933,7 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct > eth_driver *eth_drv, > > IXGBE_DEV_PRIVATE_TO_HWSTRIP_BITMAP(eth_dev->data->dev_private); > struct ether_addr *perm_addr = (struct ether_addr *) > hw->mac.perm_addr; > > - PMD_INIT_LOG(DEBUG, "eth_ixgbevf_dev_init"); > + PMD_INIT_FUNC_TRACE(); > > eth_dev->dev_ops = &ixgbevf_eth_dev_ops; > eth_dev->rx_pkt_burst = &ixgbe_recv_pkts; > @@ -963,7 +963,8 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct > eth_driver *eth_drv, > /* Initialize the shared code (base driver) */ > diag = ixgbe_init_shared_code(hw); > if (diag != IXGBE_SUCCESS) { > - PMD_INIT_LOG(ERR, "Shared code init failed for ixgbevf: > %d", diag); > + PMD_INIT_LOG(ERR, "Shared code init failed for ixgbevf: > %d", > + diag); > return -EIO; > } > > @@ -996,10 +997,9 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct > eth_driver *eth_drv, > eth_dev->data->mac_addrs = rte_zmalloc("ixgbevf", ETHER_ADDR_LEN * > hw->mac.num_rar_entries, 0); > if (eth_dev->data->mac_addrs == NULL) { > - PMD_INIT_LOG(ERR, > - "Failed to allocate %u bytes needed to store " > - "MAC addresses", > - ETHER_ADDR_LEN * hw->mac.num_rar_entries); > + PMD_INIT_LOG(ERR, "Failed to allocate %u bytes needed to > store " > + "MAC addresses", > + ETHER_ADDR_LEN * hw->mac.num_rar_entries); > return -ENOMEM; > } > > @@ -1033,13 +1033,14 @@ eth_ixgbevf_dev_init(__attribute__((unused)) > struct eth_driver *eth_drv, > break; > > default: > - PMD_INIT_LOG(ERR, "VF Initialization Failure: %d", > diag); > + PMD_INIT_LOG(ERR, "VF Initialization Failure: %d", > + diag); > return (-EIO); > } > > - PMD_INIT_LOG(DEBUG, "\nport %d vendorID=0x%x deviceID=0x%x > mac.type=%s\n", > - eth_dev->data->port_id, pci_dev->id.vendor_id, > pci_dev->id.device_id, > - "ixgbe_mac_82599_vf"); > + PMD_INIT_LOG(DEBUG, "port %d vendorID=0x%x deviceID=0x%x > mac.type=%s", > + eth_dev->data->port_id, pci_dev->id.vendor_id, > + pci_dev->id.device_id, "ixgbe_mac_82599_vf"); > > return 0; > } > @@ -1416,8 +1417,8 @@ ixgbe_dev_start(struct rte_eth_dev *dev) > > /* IXGBE devices don't support half duplex */ > if ((dev->data->dev_conf.link_duplex != ETH_LINK_AUTONEG_DUPLEX) && > - (dev->data->dev_conf.link_duplex != > ETH_LINK_FULL_DUPLEX)) { > - PMD_INIT_LOG(ERR, "Invalid link_duplex (%hu) for port > %hhu\n", > + (dev->data->dev_conf.link_duplex != ETH_LINK_FULL_DUPLEX)) { > + PMD_INIT_LOG(ERR, "Invalid link_duplex (%hu) for port > %hhu", > dev->data->dev_conf.link_duplex, > dev->data->port_id); > return -EINVAL; > @@ -1443,7 +1444,7 @@ ixgbe_dev_start(struct rte_eth_dev *dev) > /* This can fail when allocating mbufs for descriptor rings */ > err = ixgbe_dev_rx_init(dev); > if (err) { > - PMD_INIT_LOG(ERR, "Unable to initialize RX hardware\n"); > + PMD_INIT_LOG(ERR, "Unable to initialize RX hardware"); > goto error; > } > > @@ -1490,9 +1491,9 @@ ixgbe_dev_start(struct rte_eth_dev *dev) > speed = IXGBE_LINK_SPEED_10GB_FULL; > break; > default: > - PMD_INIT_LOG(ERR, "Invalid link_speed (%hu) for port > %hhu\n", > - dev->data->dev_conf.link_speed, > - dev->data->port_id); > + PMD_INIT_LOG(ERR, "Invalid link_speed (%hu) for port %hhu", > + dev->data->dev_conf.link_speed, > + dev->data->port_id); > goto error; > } > > @@ -1598,10 +1599,8 @@ ixgbe_dev_set_link_up(struct rte_eth_dev *dev) > #ifdef RTE_NIC_BYPASS > if (hw->device_id == IXGBE_DEV_ID_82599_BYPASS) { > /* Not suported in bypass mode */ > - PMD_INIT_LOG(ERR, > - "\nSet link up is not supported " > - "by device id 0x%x\n", > - hw->device_id); > + PMD_INIT_LOG(ERR, "Set link up is not supported " > + "by device id 0x%x", hw->device_id); > return -ENOTSUP; > } > #endif > @@ -1610,8 +1609,8 @@ ixgbe_dev_set_link_up(struct rte_eth_dev *dev) > return 0; > } > > - PMD_INIT_LOG(ERR, "\nSet link up is not supported by device id > 0x%x\n", > - hw->device_id); > + PMD_INIT_LOG(ERR, "Set link up is not supported by device id 0x%x", > + hw->device_id); > return -ENOTSUP; > } > > @@ -1627,10 +1626,8 @@ ixgbe_dev_set_link_down(struct rte_eth_dev *dev) > #ifdef RTE_NIC_BYPASS > if (hw->device_id == IXGBE_DEV_ID_82599_BYPASS) { > /* Not suported in bypass mode */ > - PMD_INIT_LOG(ERR, > - "\nSet link down is not supported " > - "by device id 0x%x\n", > - hw->device_id); > + PMD_INIT_LOG(ERR, "Set link down is not supported " > + "by device id 0x%x", hw->device_id); > return -ENOTSUP; > } > #endif > @@ -1639,9 +1636,8 @@ ixgbe_dev_set_link_down(struct rte_eth_dev *dev) > return 0; > } > > - PMD_INIT_LOG(ERR, > - "\nSet link down is not supported by device id 0x%x\n", > - hw->device_id); > + PMD_INIT_LOG(ERR, "Set link down is not supported by device id > 0x%x", > + hw->device_id); > return -ENOTSUP; > } > > @@ -2179,7 +2175,7 @@ ixgbe_dev_interrupt_action(struct rte_eth_dev *dev) > struct rte_eth_link link; > int intr_enable_delay = false; > > - PMD_DRV_LOG(DEBUG, "intr action type %d\n", intr->flags); > + PMD_DRV_LOG(DEBUG, "intr action type %d", intr->flags); > > if (intr->flags & IXGBE_FLAG_MAILBOX) { > ixgbe_pf_mbx_process(dev); > @@ -2209,7 +2205,8 @@ ixgbe_dev_interrupt_action(struct rte_eth_dev *dev) > > if (intr_enable_delay) { > if (rte_eal_alarm_set(timeout * 1000, > - ixgbe_dev_interrupt_delayed_handler, > (void*)dev) < 0) > + ixgbe_dev_interrupt_delayed_handler, > + (void *)dev) < 0) > PMD_DRV_LOG(ERR, "Error setting alarm"); > } else { > PMD_DRV_LOG(DEBUG, "enable intr immediately"); > @@ -2256,7 +2253,7 @@ ixgbe_dev_interrupt_delayed_handler(void *param) > _rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC); > } > > - PMD_DRV_LOG(DEBUG, "enable intr in delayed handler S[%08x]\n", > eicr); > + PMD_DRV_LOG(DEBUG, "enable intr in delayed handler S[%08x]", eicr); > ixgbe_enable_intr(dev); > rte_intr_enable(&(dev->pci_dev->intr_handle)); > } > @@ -2370,7 +2367,7 @@ ixgbe_flow_ctrl_set(struct rte_eth_dev *dev, struct > rte_eth_fc_conf *fc_conf) > if (fc_conf->autoneg != !hw->fc.disable_fc_autoneg) > return -ENOTSUP; > rx_buf_size = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(0)); > - PMD_INIT_LOG(DEBUG, "Rx packet buffer size = 0x%x \n", > rx_buf_size); > + PMD_INIT_LOG(DEBUG, "Rx packet buffer size = 0x%x", rx_buf_size); > > /* > * At least reserve one Ethernet frame for watermark > @@ -2379,8 +2376,8 @@ ixgbe_flow_ctrl_set(struct rte_eth_dev *dev, struct > rte_eth_fc_conf *fc_conf) > max_high_water = (rx_buf_size - ETHER_MAX_LEN) >> > IXGBE_RXPBSIZE_SHIFT; > if ((fc_conf->high_water > max_high_water) || > (fc_conf->high_water < fc_conf->low_water)) { > - PMD_INIT_LOG(ERR, "Invalid high/low water setup value in > KB\n"); > - PMD_INIT_LOG(ERR, "High_water must <= 0x%x\n", > max_high_water); > + PMD_INIT_LOG(ERR, "Invalid high/low water setup value in > KB"); > + PMD_INIT_LOG(ERR, "High_water must <= 0x%x", > max_high_water); > return (-EINVAL); > } > > @@ -2412,7 +2409,7 @@ ixgbe_flow_ctrl_set(struct rte_eth_dev *dev, struct > rte_eth_fc_conf *fc_conf) > return 0; > } > > - PMD_INIT_LOG(ERR, "ixgbe_fc_enable = 0x%x \n", err); > + PMD_INIT_LOG(ERR, "ixgbe_fc_enable = 0x%x", err); > return -EIO; > } > > @@ -2442,13 +2439,13 @@ ixgbe_dcb_pfc_enable_generic(struct ixgbe_hw > *hw,uint8_t tc_num) > if (hw->fc.current_mode & ixgbe_fc_tx_pause) { > /* High/Low water can not be 0 */ > if( (!hw->fc.high_water[tc_num])|| > (!hw->fc.low_water[tc_num])) { > - PMD_INIT_LOG(ERR,"Invalid water mark > configuration\n"); > + PMD_INIT_LOG(ERR, "Invalid water mark > configuration"); > ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS; > goto out; > } > > if(hw->fc.low_water[tc_num] >= hw->fc.high_water[tc_num]) { > - PMD_INIT_LOG(ERR,"Invalid water mark > configuration\n"); > + PMD_INIT_LOG(ERR, "Invalid water mark > configuration"); > ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS; > goto out; > } > @@ -2592,7 +2589,7 @@ ixgbe_priority_flow_ctrl_set(struct rte_eth_dev > *dev, struct rte_eth_pfc_conf *p > ixgbe_dcb_unpack_map_cee(dcb_config, IXGBE_DCB_RX_CONFIG, map); > tc_num = map[pfc_conf->priority]; > rx_buf_size = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(tc_num)); > - PMD_INIT_LOG(DEBUG, "Rx packet buffer size = 0x%x \n", > rx_buf_size); > + PMD_INIT_LOG(DEBUG, "Rx packet buffer size = 0x%x", rx_buf_size); > /* > * At least reserve one Ethernet frame for watermark > * high_water/low_water in kilo bytes for ixgbe > @@ -2600,8 +2597,8 @@ ixgbe_priority_flow_ctrl_set(struct rte_eth_dev > *dev, struct rte_eth_pfc_conf *p > max_high_water = (rx_buf_size - ETHER_MAX_LEN) >> > IXGBE_RXPBSIZE_SHIFT; > if ((pfc_conf->fc.high_water > max_high_water) || > (pfc_conf->fc.high_water <= pfc_conf->fc.low_water)) { > - PMD_INIT_LOG(ERR, "Invalid high/low water setup value in > KB\n"); > - PMD_INIT_LOG(ERR, "High_water must <= 0x%x\n", > max_high_water); > + PMD_INIT_LOG(ERR, "Invalid high/low water setup value in > KB"); > + PMD_INIT_LOG(ERR, "High_water must <= 0x%x", > max_high_water); > return (-EINVAL); > } > > @@ -2617,7 +2614,7 @@ ixgbe_priority_flow_ctrl_set(struct rte_eth_dev > *dev, struct rte_eth_pfc_conf *p > if ((err == IXGBE_SUCCESS) || (err == IXGBE_ERR_FC_NOT_NEGOTIATED)) > return 0; > > - PMD_INIT_LOG(ERR, "ixgbe_dcb_pfc_enable = 0x%x \n", err); > + PMD_INIT_LOG(ERR, "ixgbe_dcb_pfc_enable = 0x%x", err); > return -EIO; > } > > @@ -2764,7 +2761,7 @@ ixgbe_dev_mtu_set(struct rte_eth_dev *dev, uint16_t > mtu) > static void > ixgbevf_intr_disable(struct ixgbe_hw *hw) > { > - PMD_INIT_LOG(DEBUG, "ixgbevf_intr_disable"); > + PMD_INIT_FUNC_TRACE(); > > /* Clear interrupt mask to stop from interrupts being generated */ > IXGBE_WRITE_REG(hw, IXGBE_VTEIMC, IXGBE_VF_IRQ_CLEAR_MASK); > @@ -2777,8 +2774,8 @@ ixgbevf_dev_configure(struct rte_eth_dev *dev) > { > struct rte_eth_conf* conf = &dev->data->dev_conf; > > - PMD_INIT_LOG(DEBUG, "\nConfigured Virtual Function port id: %d\n", > - dev->data->port_id); > + PMD_INIT_LOG(DEBUG, "Configured Virtual Function port id: %d", > + dev->data->port_id); > > /* > * VF has no ability to enable/disable HW CRC > @@ -2786,12 +2783,12 @@ ixgbevf_dev_configure(struct rte_eth_dev *dev) > */ > #ifndef RTE_LIBRTE_IXGBE_PF_DISABLE_STRIP_CRC > if (!conf->rxmode.hw_strip_crc) { > - PMD_INIT_LOG(INFO, "VF can't disable HW CRC Strip\n"); > + PMD_INIT_LOG(INFO, "VF can't disable HW CRC Strip"); > conf->rxmode.hw_strip_crc = 1; > } > #else > if (conf->rxmode.hw_strip_crc) { > - PMD_INIT_LOG(INFO, "VF can't enable HW CRC Strip\n"); > + PMD_INIT_LOG(INFO, "VF can't enable HW CRC Strip"); > conf->rxmode.hw_strip_crc = 0; > } > #endif > @@ -2806,7 +2803,7 @@ ixgbevf_dev_start(struct rte_eth_dev *dev) > IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private); > int err, mask = 0; > > - PMD_INIT_LOG(DEBUG, "ixgbevf_dev_start"); > + PMD_INIT_FUNC_TRACE(); > > hw->mac.ops.reset_hw(hw); > > @@ -2818,7 +2815,7 @@ ixgbevf_dev_start(struct rte_eth_dev *dev) > /* This can fail when allocating mbufs for descriptor rings */ > err = ixgbevf_dev_rx_init(dev); > if (err) { > - PMD_INIT_LOG(ERR, "Unable to initialize RX hardware > (%d)\n", err); > + PMD_INIT_LOG(ERR, "Unable to initialize RX hardware (%d)", > err); > ixgbe_dev_clear_queues(dev); > return err; > } > @@ -2841,7 +2838,7 @@ ixgbevf_dev_stop(struct rte_eth_dev *dev) > { > struct ixgbe_hw *hw = > IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private); > > - PMD_INIT_LOG(DEBUG, "ixgbevf_dev_stop"); > + PMD_INIT_FUNC_TRACE(); > > hw->adapter_stopped = TRUE; > ixgbe_stop_adapter(hw); > @@ -2860,7 +2857,7 @@ ixgbevf_dev_close(struct rte_eth_dev *dev) > { > struct ixgbe_hw *hw = > IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private); > > - PMD_INIT_LOG(DEBUG, "ixgbevf_dev_close"); > + PMD_INIT_FUNC_TRACE(); > > ixgbe_reset_hw(hw); > > @@ -2969,7 +2966,7 @@ ixgbe_vmdq_mode_check(struct ixgbe_hw *hw) > /* we only need to do this if VMDq is enabled */ > reg_val = IXGBE_READ_REG(hw, IXGBE_VT_CTL); > if (!(reg_val & IXGBE_VT_CTL_VT_ENABLE)) { > - PMD_INIT_LOG(ERR, "VMDq must be enabled for this > setting\n"); > + PMD_INIT_LOG(ERR, "VMDq must be enabled for this setting"); > return (-1); > } > > @@ -3098,7 +3095,7 @@ ixgbe_set_pool_rx_mode(struct rte_eth_dev *dev, > uint16_t pool, > > if (hw->mac.type == ixgbe_mac_82598EB) { > PMD_INIT_LOG(ERR, "setting VF receive mode set should be > done" > - " on 82599 hardware and newer\n"); > + " on 82599 hardware and newer"); > return (-ENOTSUP); > } > if (ixgbe_vmdq_mode_check(hw) < 0) > @@ -3513,8 +3510,7 @@ ixgbevf_remove_mac_addr(struct rte_eth_dev *dev, > uint32_t index) > continue; > diag = ixgbevf_set_uc_addr_vf(hw, 2, mac_addr->addr_bytes); > if (diag != 0) > - PMD_DRV_LOG(ERR, > - "Adding again MAC address " > + PMD_DRV_LOG(ERR, "Adding again MAC address " > "%02x:%02x:%02x:%02x:%02x:%02x failed " > "diag=%d", > mac_addr->addr_bytes[0], > diff --git a/lib/librte_pmd_ixgbe/ixgbe_fdir.c > b/lib/librte_pmd_ixgbe/ixgbe_fdir.c > index 6c0a530..8819aac 100644 > --- a/lib/librte_pmd_ixgbe/ixgbe_fdir.c > +++ b/lib/librte_pmd_ixgbe/ixgbe_fdir.c > @@ -112,7 +112,7 @@ static void fdir_enable_82599(struct ixgbe_hw *hw, u32 > fdirctrl) > } > > if (i >= IXGBE_FDIR_INIT_DONE_POLL) > - PMD_INIT_LOG(WARNING, "Flow Director poll time > exceeded!\n"); > + PMD_INIT_LOG(WARNING, "Flow Director poll time exceeded!"); > } > > /* > @@ -381,7 +381,7 @@ fdir_add_signature_filter_82599(struct ixgbe_hw *hw, > fdirhashcmd |= fdirhash; > IXGBE_WRITE_REG64(hw, IXGBE_FDIRHASH, fdirhashcmd); > > - PMD_INIT_LOG(DEBUG, "Tx Queue=%x hash=%x\n", queue, > (u32)fdirhashcmd); > + PMD_INIT_LOG(DEBUG, "Tx Queue=%x hash=%x", queue, > (u32)fdirhashcmd); > } > > /* > @@ -614,7 +614,7 @@ fdir_set_input_mask_82599(struct ixgbe_hw *hw, > /* use the L4 protocol mask for raw IPv4/IPv6 traffic */ > fdirm |= IXGBE_FDIRM_L4P; > if (input_mask->dst_port_mask || > input_mask->src_port_mask) { > - PMD_INIT_LOG(ERR, " Error on src/dst port mask\n"); > + PMD_INIT_LOG(ERR, " Error on src/dst port mask"); > return -EINVAL; > } > } > diff --git a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c > b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c > index 46962bc..981df60 100644 > --- a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c > +++ b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c > @@ -490,8 +490,7 @@ ixgbe_xmit_cleanup(struct igb_tx_queue *txq) > desc_to_clean_to = sw_ring[desc_to_clean_to].last_id; > if (! (txr[desc_to_clean_to].wb.status & IXGBE_TXD_STAT_DD)) > { > - PMD_TX_FREE_LOG(DEBUG, > - "TX descriptor %4u is not done" > + PMD_TX_FREE_LOG(DEBUG, "TX descriptor %4u is not done" > "(port=%d queue=%d)", > desc_to_clean_to, > txq->port_id, txq->queue_id); > @@ -507,8 +506,7 @@ ixgbe_xmit_cleanup(struct igb_tx_queue *txq) > nb_tx_to_clean = (uint16_t)(desc_to_clean_to - > last_desc_cleaned); > > - PMD_TX_FREE_LOG(DEBUG, > - "Cleaning %4u TX descriptors: %4u to %4u " > + PMD_TX_FREE_LOG(DEBUG, "Cleaning %4u TX descriptors: %4u to %4u " > "(port=%d queue=%d)", > nb_tx_to_clean, last_desc_cleaned, > desc_to_clean_to, > txq->port_id, txq->queue_id); > @@ -614,7 +612,7 @@ ixgbe_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > tx_last = (uint16_t) (tx_last - txq->nb_tx_desc); > > PMD_TX_LOG(DEBUG, "port_id=%u queue_id=%u pktlen=%u" > - " tx_first=%u tx_last=%u\n", > + " tx_first=%u tx_last=%u", > (unsigned) txq->port_id, > (unsigned) txq->queue_id, > (unsigned) pkt_len, > @@ -627,8 +625,7 @@ ixgbe_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > * nb_used better be less than or equal to > txq->tx_rs_thresh > */ > if (nb_used > txq->nb_tx_free) { > - PMD_TX_FREE_LOG(DEBUG, > - "Not enough free TX descriptors " > + PMD_TX_FREE_LOG(DEBUG, "Not enough free TX > descriptors " > "nb_used=%4u nb_free=%4u " > "(port=%d queue=%d)", > nb_used, txq->nb_tx_free, > @@ -1066,7 +1063,7 @@ rx_recv_pkts(void *rx_queue, struct rte_mbuf > **rx_pkts, > if (ixgbe_rx_alloc_bufs(rxq) != 0) { > int i, j; > PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u > " > - "queue_id=%u\n", (unsigned) > rxq->port_id, > + "queue_id=%u", (unsigned) rxq->port_id, > (unsigned) rxq->queue_id); > > > rte_eth_devices[rxq->port_id].data->rx_mbuf_alloc_failed += > @@ -1193,7 +1190,7 @@ ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf > **rx_pkts, > * frames to its peer(s). > */ > PMD_RX_LOG(DEBUG, "port_id=%u queue_id=%u rx_id=%u " > - "ext_err_stat=0x%08x pkt_len=%u\n", > + "ext_err_stat=0x%08x pkt_len=%u", > (unsigned) rxq->port_id, (unsigned) > rxq->queue_id, > (unsigned) rx_id, (unsigned) staterr, > (unsigned) > rte_le_to_cpu_16(rxd.wb.upper.length)); > @@ -1201,7 +1198,7 @@ ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf > **rx_pkts, > nmb = rte_rxmbuf_alloc(rxq->mb_pool); > if (nmb == NULL) { > PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u > " > - "queue_id=%u\n", (unsigned) > rxq->port_id, > + "queue_id=%u", (unsigned) rxq->port_id, > (unsigned) rxq->queue_id); > > rte_eth_devices[rxq->port_id].data->rx_mbuf_alloc_failed++; > break; > @@ -1296,7 +1293,7 @@ ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf > **rx_pkts, > nb_hold = (uint16_t) (nb_hold + rxq->nb_rx_hold); > if (nb_hold > rxq->rx_free_thresh) { > PMD_RX_LOG(DEBUG, "port_id=%u queue_id=%u rx_tail=%u " > - "nb_hold=%u nb_rx=%u\n", > + "nb_hold=%u nb_rx=%u", > (unsigned) rxq->port_id, (unsigned) > rxq->queue_id, > (unsigned) rx_id, (unsigned) nb_hold, > (unsigned) nb_rx); > @@ -1383,8 +1380,8 @@ ixgbe_recv_scattered_pkts(void *rx_queue, struct > rte_mbuf **rx_pkts, > * to happen by sending specific "back-pressure" flow > control > * frames to its peer(s). > */ > - PMD_RX_LOG(DEBUG, "\nport_id=%u queue_id=%u rx_id=%u " > - "staterr=0x%x data_len=%u\n", > + PMD_RX_LOG(DEBUG, "port_id=%u queue_id=%u rx_id=%u " > + "staterr=0x%x data_len=%u", > (unsigned) rxq->port_id, (unsigned) > rxq->queue_id, > (unsigned) rx_id, (unsigned) staterr, > (unsigned) > rte_le_to_cpu_16(rxd.wb.upper.length)); > @@ -1392,7 +1389,7 @@ ixgbe_recv_scattered_pkts(void *rx_queue, struct > rte_mbuf **rx_pkts, > nmb = rte_rxmbuf_alloc(rxq->mb_pool); > if (nmb == NULL) { > PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u > " > - "queue_id=%u\n", (unsigned) > rxq->port_id, > + "queue_id=%u", (unsigned) rxq->port_id, > (unsigned) rxq->queue_id); > > rte_eth_devices[rxq->port_id].data->rx_mbuf_alloc_failed++; > break; > @@ -1561,7 +1558,7 @@ ixgbe_recv_scattered_pkts(void *rx_queue, struct > rte_mbuf **rx_pkts, > nb_hold = (uint16_t) (nb_hold + rxq->nb_rx_hold); > if (nb_hold > rxq->rx_free_thresh) { > PMD_RX_LOG(DEBUG, "port_id=%u queue_id=%u rx_tail=%u " > - "nb_hold=%u nb_rx=%u\n", > + "nb_hold=%u nb_rx=%u", > (unsigned) rxq->port_id, (unsigned) > rxq->queue_id, > (unsigned) rx_id, (unsigned) nb_hold, > (unsigned) nb_rx); > @@ -1767,7 +1764,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev, > if (tx_rs_thresh >= (nb_desc - 2)) { > PMD_INIT_LOG(ERR, "tx_rs_thresh must be less than the > number " > "of TX descriptors minus 2. (tx_rs_thresh=%u " > - "port=%d queue=%d)\n", (unsigned > int)tx_rs_thresh, > + "port=%d queue=%d)", (unsigned > int)tx_rs_thresh, > (int)dev->data->port_id, (int)queue_idx); > return -(EINVAL); > } > @@ -1775,7 +1772,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev, > PMD_INIT_LOG(ERR, "tx_rs_thresh must be less than the " > "tx_free_thresh must be less than the number > of " > "TX descriptors minus 3. (tx_free_thresh=%u " > - "port=%d queue=%d)\n", > + "port=%d queue=%d)", > (unsigned int)tx_free_thresh, > (int)dev->data->port_id, (int)queue_idx); > return -(EINVAL); > @@ -1783,7 +1780,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev, > if (tx_rs_thresh > tx_free_thresh) { > PMD_INIT_LOG(ERR, "tx_rs_thresh must be less than or equal > to " > "tx_free_thresh. (tx_free_thresh=%u " > - "tx_rs_thresh=%u port=%d queue=%d)\n", > + "tx_rs_thresh=%u port=%d queue=%d)", > (unsigned int)tx_free_thresh, > (unsigned int)tx_rs_thresh, > (int)dev->data->port_id, > @@ -1793,7 +1790,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev, > if ((nb_desc % tx_rs_thresh) != 0) { > PMD_INIT_LOG(ERR, "tx_rs_thresh must be a divisor of the " > "number of TX descriptors. (tx_rs_thresh=%u " > - "port=%d queue=%d)\n", (unsigned > int)tx_rs_thresh, > + "port=%d queue=%d)", (unsigned > int)tx_rs_thresh, > (int)dev->data->port_id, (int)queue_idx); > return -(EINVAL); > } > @@ -1807,7 +1804,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev, > if ((tx_rs_thresh > 1) && (tx_conf->tx_thresh.wthresh != 0)) { > PMD_INIT_LOG(ERR, "TX WTHRESH must be set to 0 if " > "tx_rs_thresh is greater than 1. > (tx_rs_thresh=%u " > - "port=%d queue=%d)\n", (unsigned > int)tx_rs_thresh, > + "port=%d queue=%d)", (unsigned > int)tx_rs_thresh, > (int)dev->data->port_id, (int)queue_idx); > return -(EINVAL); > } > @@ -1873,26 +1870,32 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev, > ixgbe_tx_queue_release(txq); > return (-ENOMEM); > } > - PMD_INIT_LOG(DEBUG, "sw_ring=%p hw_ring=%p dma_addr=0x%"PRIx64"\n", > + PMD_INIT_LOG(DEBUG, "sw_ring=%p hw_ring=%p dma_addr=0x%"PRIx64, > txq->sw_ring, txq->tx_ring, txq->tx_ring_phys_addr); > > /* Use a simple Tx queue (no offloads, no multi segs) if possible > */ > if (((txq->txq_flags & IXGBE_SIMPLE_FLAGS) == IXGBE_SIMPLE_FLAGS) > && > (txq->tx_rs_thresh >= RTE_PMD_IXGBE_TX_MAX_BURST)) { > - PMD_INIT_LOG(INFO, "Using simple tx code path\n"); > + PMD_INIT_LOG(INFO, "Using simple tx code path"); > #ifdef RTE_IXGBE_INC_VECTOR > if (txq->tx_rs_thresh <= RTE_IXGBE_TX_MAX_FREE_BUF_SZ && > ixgbe_txq_vec_setup(txq, socket_id) == 0) { > - PMD_INIT_LOG(INFO, "Vector tx enabled.\n"); > + PMD_INIT_LOG(INFO, "Vector tx enabled."); > dev->tx_pkt_burst = ixgbe_xmit_pkts_vec; > } > else > #endif > dev->tx_pkt_burst = ixgbe_xmit_pkts_simple; > } else { > - PMD_INIT_LOG(INFO, "Using full-featured tx code path\n"); > - PMD_INIT_LOG(INFO, " - txq_flags = %lx > [IXGBE_SIMPLE_FLAGS=%lx]\n", (long unsigned)txq->txq_flags, (long > unsigned)IXGBE_SIMPLE_FLAGS); > - PMD_INIT_LOG(INFO, " - tx_rs_thresh = %lu > [RTE_PMD_IXGBE_TX_MAX_BURST=%lu]\n", (long unsigned)txq->tx_rs_thresh, > (long unsigned)RTE_PMD_IXGBE_TX_MAX_BURST); > + PMD_INIT_LOG(INFO, "Using full-featured tx code path"); > + PMD_INIT_LOG(INFO, " - txq_flags = %lx " > + "[IXGBE_SIMPLE_FLAGS=%lx]", > + (long unsigned)txq->txq_flags, > + (long unsigned)IXGBE_SIMPLE_FLAGS); > + PMD_INIT_LOG(INFO, " - tx_rs_thresh = %lu " > + "[RTE_PMD_IXGBE_TX_MAX_BURST=%lu]", > + (long unsigned)txq->tx_rs_thresh, > + (long unsigned)RTE_PMD_IXGBE_TX_MAX_BURST); > dev->tx_pkt_burst = ixgbe_xmit_pkts; > } > > @@ -2152,7 +2155,7 @@ ixgbe_dev_rx_queue_setup(struct rte_eth_dev *dev, > ixgbe_rx_queue_release(rxq); > return (-ENOMEM); > } > - PMD_INIT_LOG(DEBUG, "sw_ring=%p hw_ring=%p dma_addr=0x%"PRIx64"\n", > + PMD_INIT_LOG(DEBUG, "sw_ring=%p hw_ring=%p dma_addr=0x%"PRIx64, > rxq->sw_ring, rxq->rx_ring, rxq->rx_ring_phys_addr); > > /* > @@ -2166,13 +2169,13 @@ ixgbe_dev_rx_queue_setup(struct rte_eth_dev *dev, > #ifdef RTE_LIBRTE_IXGBE_RX_ALLOW_BULK_ALLOC > PMD_INIT_LOG(DEBUG, "Rx Burst Bulk Alloc Preconditions are > " > "satisfied. Rx Burst Bulk Alloc function will > be " > - "used on port=%d, queue=%d.\n", > + "used on port=%d, queue=%d.", > rxq->port_id, rxq->queue_id); > dev->rx_pkt_burst = ixgbe_recv_pkts_bulk_alloc; > #ifdef RTE_IXGBE_INC_VECTOR > if (!ixgbe_rx_vec_condition_check(dev)) { > PMD_INIT_LOG(INFO, "Vector rx enabled, please make > " > - "sure RX burst size no less than > 32.\n"); > + "sure RX burst size no less than > 32."); > ixgbe_rxq_vec_setup(rxq, socket_id); > dev->rx_pkt_burst = ixgbe_recv_pkts_vec; > } > @@ -2182,7 +2185,7 @@ ixgbe_dev_rx_queue_setup(struct rte_eth_dev *dev, > PMD_INIT_LOG(DEBUG, "Rx Burst Bulk Alloc Preconditions " > "are not satisfied, Scattered Rx is > requested, " > "or RTE_LIBRTE_IXGBE_RX_ALLOW_BULK_ALLOC is > not " > - "enabled (port=%d, queue=%d).\n", > + "enabled (port=%d, queue=%d).", > rxq->port_id, rxq->queue_id); > } > dev->data->rx_queues[queue_idx] = rxq; > @@ -2201,7 +2204,7 @@ ixgbe_dev_rx_queue_count(struct rte_eth_dev *dev, > uint16_t rx_queue_id) > uint32_t desc = 0; > > if (rx_queue_id >= dev->data->nb_rx_queues) { > - PMD_RX_LOG(ERR, "Invalid RX queue id=%d\n", rx_queue_id); > + PMD_RX_LOG(ERR, "Invalid RX queue id=%d", rx_queue_id); > return 0; > } > > @@ -2917,7 +2920,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev, > ixgbe_dcb_rx_hw_config(hw, dcb_config); > break; > default: > - PMD_INIT_LOG(ERR, "Incorrect DCB RX mode configuration\n"); > + PMD_INIT_LOG(ERR, "Incorrect DCB RX mode configuration"); > break; > } > switch (dev->data->dev_conf.txmode.mq_mode) { > @@ -2939,7 +2942,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev, > ixgbe_dcb_tx_hw_config(hw, dcb_config); > break; > default: > - PMD_INIT_LOG(ERR, "Incorrect DCB TX mode configuration\n"); > + PMD_INIT_LOG(ERR, "Incorrect DCB TX mode configuration"); > break; > } > > @@ -3210,7 +3213,7 @@ ixgbe_alloc_rx_queue_mbufs(struct igb_rx_queue *rxq) > volatile union ixgbe_adv_rx_desc *rxd; > struct rte_mbuf *mbuf = rte_rxmbuf_alloc(rxq->mb_pool); > if (mbuf == NULL) { > - PMD_INIT_LOG(ERR, "RX mbuf alloc failed > queue_id=%u\n", > + PMD_INIT_LOG(ERR, "RX mbuf alloc failed > queue_id=%u", > (unsigned) rxq->queue_id); > return (-ENOMEM); > } > @@ -3282,7 +3285,7 @@ ixgbe_dev_mq_rx_configure(struct rte_eth_dev *dev) > IXGBE_WRITE_REG(hw, IXGBE_MRQC, > IXGBE_MRQC_VMDQRT8TCEN); > break; > default: > - PMD_INIT_LOG(ERR, "invalid pool number in IOV > mode\n"); > + PMD_INIT_LOG(ERR, "invalid pool number in IOV > mode"); > } > } > > @@ -3335,7 +3338,7 @@ ixgbe_dev_mq_tx_configure(struct rte_eth_dev *dev) > break; > default: > mtqc = IXGBE_MTQC_64Q_1PB; > - PMD_INIT_LOG(ERR, "invalid pool number in IOV > mode\n"); > + PMD_INIT_LOG(ERR, "invalid pool number in IOV > mode"); > } > IXGBE_WRITE_REG(hw, IXGBE_MTQC, mtqc); > } > @@ -3603,7 +3606,7 @@ ixgbe_setup_loopback_link_82599(struct ixgbe_hw *hw) > if (ixgbe_verify_lesm_fw_enabled_82599(hw)) { > if (hw->mac.ops.acquire_swfw_sync(hw, > IXGBE_GSSR_MAC_CSR_SM) != > IXGBE_SUCCESS) { > - PMD_INIT_LOG(ERR, "Could not enable loopback > mode\n"); > + PMD_INIT_LOG(ERR, "Could not enable loopback > mode"); > /* ignore error */ > return; > } > @@ -3699,7 +3702,7 @@ ixgbe_dev_rx_queue_start(struct rte_eth_dev *dev, > uint16_t rx_queue_id) > /* Allocate buffers for descriptor rings */ > if (ixgbe_alloc_rx_queue_mbufs(rxq) != 0) { > PMD_INIT_LOG(ERR, > - "Could not alloc mbuf for queue:%d\n", > + "Could not alloc mbuf for queue:%d", > rx_queue_id); > return -1; > } > @@ -3715,7 +3718,7 @@ ixgbe_dev_rx_queue_start(struct rte_eth_dev *dev, > uint16_t rx_queue_id) > } while (--poll_ms && !(rxdctl & IXGBE_RXDCTL_ENABLE)); > if (!poll_ms) > PMD_INIT_LOG(ERR, "Could not enable " > - "Rx Queue %d\n", rx_queue_id); > + "Rx Queue %d", rx_queue_id); > rte_wmb(); > IXGBE_WRITE_REG(hw, IXGBE_RDH(rxq->reg_idx), 0); > IXGBE_WRITE_REG(hw, IXGBE_RDT(rxq->reg_idx), > rxq->nb_rx_desc - 1); > @@ -3754,7 +3757,7 @@ ixgbe_dev_rx_queue_stop(struct rte_eth_dev *dev, > uint16_t rx_queue_id) > } while (--poll_ms && (rxdctl | IXGBE_RXDCTL_ENABLE)); > if (!poll_ms) > PMD_INIT_LOG(ERR, "Could not disable " > - "Rx Queue %d\n", rx_queue_id); > + "Rx Queue %d", rx_queue_id); > > rte_delay_us(RTE_IXGBE_WAIT_100_US); > > @@ -3797,7 +3800,7 @@ ixgbe_dev_tx_queue_start(struct rte_eth_dev *dev, > uint16_t tx_queue_id) > } while (--poll_ms && !(txdctl & > IXGBE_TXDCTL_ENABLE)); > if (!poll_ms) > PMD_INIT_LOG(ERR, "Could not enable " > - "Tx Queue %d\n", tx_queue_id); > + "Tx Queue %d", tx_queue_id); > } > rte_wmb(); > IXGBE_WRITE_REG(hw, IXGBE_TDH(txq->reg_idx), 0); > @@ -3838,7 +3841,7 @@ ixgbe_dev_tx_queue_stop(struct rte_eth_dev *dev, > uint16_t tx_queue_id) > } while (--poll_ms && (txtdh != txtdt)); > if (!poll_ms) > PMD_INIT_LOG(ERR, > - "Tx Queue %d is not empty when > stopping.\n", > + "Tx Queue %d is not empty when stopping.", > tx_queue_id); > } > > @@ -3856,7 +3859,7 @@ ixgbe_dev_tx_queue_stop(struct rte_eth_dev *dev, > uint16_t tx_queue_id) > } while (--poll_ms && (txdctl | > IXGBE_TXDCTL_ENABLE)); > if (!poll_ms) > PMD_INIT_LOG(ERR, "Could not disable " > - "Tx Queue %d\n", tx_queue_id); > + "Tx Queue %d", tx_queue_id); > } > > if (txq->ops != NULL) { > @@ -4073,7 +4076,7 @@ ixgbevf_dev_rxtx_start(struct rte_eth_dev *dev) > } while (--poll_ms && !(txdctl & IXGBE_TXDCTL_ENABLE)); > if (!poll_ms) > PMD_INIT_LOG(ERR, "Could not enable " > - "Tx Queue %d\n", i); > + "Tx Queue %d", i); > } > for (i = 0; i < dev->data->nb_rx_queues; i++) { > > @@ -4091,7 +4094,7 @@ ixgbevf_dev_rxtx_start(struct rte_eth_dev *dev) > } while (--poll_ms && !(rxdctl & IXGBE_RXDCTL_ENABLE)); > if (!poll_ms) > PMD_INIT_LOG(ERR, "Could not enable " > - "Rx Queue %d\n", i); > + "Rx Queue %d", i); > rte_wmb(); > IXGBE_WRITE_REG(hw, IXGBE_VFRDT(i), rxq->nb_rx_desc - 1); > > -- > 1.7.10.4 > > Reviewed-By: Jay Rolette