From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 9730F424D8; Wed, 12 Jun 2024 06:38:07 +0200 (CEST) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 2CEAA40E48; Wed, 12 Jun 2024 06:37:07 +0200 (CEST) Received: from mgamail.intel.com (mgamail.intel.com [198.175.65.14]) by mails.dpdk.org (Postfix) with ESMTP id B3E2740E42 for ; Wed, 12 Jun 2024 06:37:05 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1718167026; x=1749703026; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=0cQ4xCg1wziDktKuV2mR82bPebYi9rO09FUUNHjuNdM=; b=OWlKGDL0CVzU8M2gl4ieN02k8wNVhZ8M2c2D244DjlPcnNKpJC5HaAK2 DEPEfSZmAtNp7+rU6c1xzEElSCDAYIPJ88E3UuszYCt8mUJOnkrtbFr2Q aFY+rjX9PDxUnlc6jLpNoCXOvTFotb58id3u3UGVZBb6kANc2QaBOGenQ pnlBTIRWsbhiNEqTVPUgN4pV29xrKZ7jWe4qXk08tNvKHMWvYu4B5yA6G U8vb+B1gSo5HSvQLoc0dg1x6sSOOCWlbs0LLoQ3CtENiapE3HEX12r76k S5aIVSy9/NumlLW9sR4Cbyz/rGtRqupLCkgX9HOLzYZ2/jM3xnrjkg+0I g==; X-CSE-ConnectionGUID: MBHYx6j7Sb+3+myXzuUDQg== X-CSE-MsgGUID: zXnWl7rMRhS1NvFatwZINA== X-IronPort-AV: E=McAfee;i="6600,9927,11100"; a="18742204" X-IronPort-AV: E=Sophos;i="6.08,232,1712646000"; d="scan'208";a="18742204" Received: from orviesa008.jf.intel.com ([10.64.159.148]) by orvoesa106.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 11 Jun 2024 21:37:06 -0700 X-CSE-ConnectionGUID: YAkGyOokRwOv+6YNEuqeAg== X-CSE-MsgGUID: WSIj4Z46RYuRa0setGJeWg== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="6.08,232,1712646000"; d="scan'208";a="40281965" Received: from unknown (HELO npf-hyd-clx-03..) ([10.145.170.182]) by orviesa008.jf.intel.com with ESMTP; 11 Jun 2024 21:37:01 -0700 From: Soumyadeep Hore To: bruce.richardson@intel.com, anatoly.burakov@intel.com Cc: dev@dpdk.org Subject: [PATCH v3 13/22] common/idpf: avoid variable 0-init Date: Wed, 12 Jun 2024 03:52:48 +0000 Message-ID: <20240612035257.2245824-14-soumyadeep.hore@intel.com> X-Mailer: git-send-email 2.43.0 In-Reply-To: <20240612035257.2245824-1-soumyadeep.hore@intel.com> References: <20240604080611.2197835-1-soumyadeep.hore@intel.com> <20240612035257.2245824-1-soumyadeep.hore@intel.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Don't initialize the variables if not needed. Also use 'err' instead of 'status', 'ret_code', 'ret' etc. for consistency and change the return label 'sq_send_command_out' to 'err_unlock'. Signed-off-by: Soumyadeep Hore --- drivers/common/idpf/base/idpf_controlq.c | 63 +++++++++---------- .../common/idpf/base/idpf_controlq_setup.c | 18 +++--- 2 files changed, 39 insertions(+), 42 deletions(-) diff --git a/drivers/common/idpf/base/idpf_controlq.c b/drivers/common/idpf/base/idpf_controlq.c index b5ba9c3bd0..bd23e54421 100644 --- a/drivers/common/idpf/base/idpf_controlq.c +++ b/drivers/common/idpf/base/idpf_controlq.c @@ -61,7 +61,7 @@ static void idpf_ctlq_init_regs(struct idpf_hw *hw, struct idpf_ctlq_info *cq, */ static void idpf_ctlq_init_rxq_bufs(struct idpf_ctlq_info *cq) { - int i = 0; + int i; for (i = 0; i < cq->ring_size; i++) { struct idpf_ctlq_desc *desc = IDPF_CTLQ_DESC(cq, i); @@ -134,7 +134,7 @@ int idpf_ctlq_add(struct idpf_hw *hw, { struct idpf_ctlq_info *cq; bool is_rxq = false; - int status = 0; + int err; if (!qinfo->len || !qinfo->buf_size || qinfo->len > IDPF_CTLQ_MAX_RING_SIZE || @@ -164,16 +164,16 @@ int idpf_ctlq_add(struct idpf_hw *hw, is_rxq = true; /* fallthrough */ case IDPF_CTLQ_TYPE_MAILBOX_TX: - status = idpf_ctlq_alloc_ring_res(hw, cq); + err = idpf_ctlq_alloc_ring_res(hw, cq); break; default: - status = -EINVAL; + err = -EINVAL; break; } - if (status) + if (err) #ifdef NVME_CPF - return status; + return err; #else goto init_free_q; #endif @@ -187,7 +187,7 @@ int idpf_ctlq_add(struct idpf_hw *hw, idpf_calloc(hw, qinfo->len, sizeof(struct idpf_ctlq_msg *)); if (!cq->bi.tx_msg) { - status = -ENOMEM; + err = -ENOMEM; goto init_dealloc_q_mem; } #endif @@ -203,17 +203,16 @@ int idpf_ctlq_add(struct idpf_hw *hw, #ifndef NVME_CPF *cq_out = cq; - return status; + return 0; init_dealloc_q_mem: /* free ring buffers and the ring itself */ idpf_ctlq_dealloc_ring_res(hw, cq); init_free_q: idpf_free(hw, cq); - cq = NULL; #endif - return status; + return err; } /** @@ -249,8 +248,8 @@ int idpf_ctlq_init(struct idpf_hw *hw, u8 num_q, #endif { struct idpf_ctlq_info *cq = NULL, *tmp = NULL; - int ret_code = 0; - int i = 0; + int err; + int i; LIST_INIT(&hw->cq_list_head); @@ -261,19 +260,19 @@ int idpf_ctlq_init(struct idpf_hw *hw, u8 num_q, cq = *(ctlq + i); #endif - ret_code = idpf_ctlq_add(hw, qinfo, &cq); - if (ret_code) + err = idpf_ctlq_add(hw, qinfo, &cq); + if (err) goto init_destroy_qs; } - return ret_code; + return 0; init_destroy_qs: LIST_FOR_EACH_ENTRY_SAFE(cq, tmp, &hw->cq_list_head, idpf_ctlq_info, cq_list) idpf_ctlq_remove(hw, cq); - return ret_code; + return err; } /** @@ -307,9 +306,9 @@ int idpf_ctlq_send(struct idpf_hw *hw, struct idpf_ctlq_info *cq, u16 num_q_msg, struct idpf_ctlq_msg q_msg[]) { struct idpf_ctlq_desc *desc; - int num_desc_avail = 0; - int status = 0; - int i = 0; + int num_desc_avail; + int err = 0; + int i; if (!cq || !cq->ring_size) return -ENOBUFS; @@ -319,8 +318,8 @@ int idpf_ctlq_send(struct idpf_hw *hw, struct idpf_ctlq_info *cq, /* Ensure there are enough descriptors to send all messages */ num_desc_avail = IDPF_CTLQ_DESC_UNUSED(cq); if (num_desc_avail == 0 || num_desc_avail < num_q_msg) { - status = -ENOSPC; - goto sq_send_command_out; + err = -ENOSPC; + goto err_unlock; } for (i = 0; i < num_q_msg; i++) { @@ -391,10 +390,10 @@ int idpf_ctlq_send(struct idpf_hw *hw, struct idpf_ctlq_info *cq, wr32(hw, cq->reg.tail, cq->next_to_use); -sq_send_command_out: +err_unlock: idpf_release_lock(&cq->cq_lock); - return status; + return err; } /** @@ -418,9 +417,8 @@ static int __idpf_ctlq_clean_sq(struct idpf_ctlq_info *cq, u16 *clean_count, struct idpf_ctlq_msg *msg_status[], bool force) { struct idpf_ctlq_desc *desc; - u16 i = 0, num_to_clean; + u16 i, num_to_clean; u16 ntc, desc_err; - int ret = 0; if (!cq || !cq->ring_size) return -ENOBUFS; @@ -467,7 +465,7 @@ static int __idpf_ctlq_clean_sq(struct idpf_ctlq_info *cq, u16 *clean_count, /* Return number of descriptors actually cleaned */ *clean_count = i; - return ret; + return 0; } /** @@ -534,7 +532,6 @@ int idpf_ctlq_post_rx_buffs(struct idpf_hw *hw, struct idpf_ctlq_info *cq, u16 ntp = cq->next_to_post; bool buffs_avail = false; u16 tbp = ntp + 1; - int status = 0; int i = 0; if (*buff_count > cq->ring_size) @@ -635,7 +632,7 @@ int idpf_ctlq_post_rx_buffs(struct idpf_hw *hw, struct idpf_ctlq_info *cq, /* return the number of buffers that were not posted */ *buff_count = *buff_count - i; - return status; + return 0; } /** @@ -654,8 +651,8 @@ int idpf_ctlq_recv(struct idpf_ctlq_info *cq, u16 *num_q_msg, { u16 num_to_clean, ntc, ret_val, flags; struct idpf_ctlq_desc *desc; - int ret_code = 0; - u16 i = 0; + int err = 0; + u16 i; if (!cq || !cq->ring_size) return -ENOBUFS; @@ -688,7 +685,7 @@ int idpf_ctlq_recv(struct idpf_ctlq_info *cq, u16 *num_q_msg, IDPF_CTLQ_FLAG_FTYPE_S; if (flags & IDPF_CTLQ_FLAG_ERR) - ret_code = -EBADMSG; + err = -EBADMSG; q_msg[i].cookie.mbx.chnl_opcode = LE32_TO_CPU(desc->cookie_high); q_msg[i].cookie.mbx.chnl_retval = LE32_TO_CPU(desc->cookie_low); @@ -734,7 +731,7 @@ int idpf_ctlq_recv(struct idpf_ctlq_info *cq, u16 *num_q_msg, *num_q_msg = i; if (*num_q_msg == 0) - ret_code = -ENOMSG; + err = -ENOMSG; - return ret_code; + return err; } diff --git a/drivers/common/idpf/base/idpf_controlq_setup.c b/drivers/common/idpf/base/idpf_controlq_setup.c index 21f43c74f5..cd6bcb1cf0 100644 --- a/drivers/common/idpf/base/idpf_controlq_setup.c +++ b/drivers/common/idpf/base/idpf_controlq_setup.c @@ -1,5 +1,5 @@ /* SPDX-License-Identifier: BSD-3-Clause - * Copyright(c) 2001-2023 Intel Corporation + * Copyright(c) 2001-2024 Intel Corporation */ @@ -34,7 +34,7 @@ static int idpf_ctlq_alloc_desc_ring(struct idpf_hw *hw, static int idpf_ctlq_alloc_bufs(struct idpf_hw *hw, struct idpf_ctlq_info *cq) { - int i = 0; + int i; /* Do not allocate DMA buffers for transmit queues */ if (cq->cq_type == IDPF_CTLQ_TYPE_MAILBOX_TX) @@ -153,20 +153,20 @@ void idpf_ctlq_dealloc_ring_res(struct idpf_hw *hw, struct idpf_ctlq_info *cq) */ int idpf_ctlq_alloc_ring_res(struct idpf_hw *hw, struct idpf_ctlq_info *cq) { - int ret_code; + int err; /* verify input for valid configuration */ if (!cq->ring_size || !cq->buf_size) return -EINVAL; /* allocate the ring memory */ - ret_code = idpf_ctlq_alloc_desc_ring(hw, cq); - if (ret_code) - return ret_code; + err = idpf_ctlq_alloc_desc_ring(hw, cq); + if (err) + return err; /* allocate buffers in the rings */ - ret_code = idpf_ctlq_alloc_bufs(hw, cq); - if (ret_code) + err = idpf_ctlq_alloc_bufs(hw, cq); + if (err) goto idpf_init_cq_free_ring; /* success! */ @@ -174,5 +174,5 @@ int idpf_ctlq_alloc_ring_res(struct idpf_hw *hw, struct idpf_ctlq_info *cq) idpf_init_cq_free_ring: idpf_free_dma_mem(hw, &cq->desc_ring); - return ret_code; + return err; } -- 2.43.0