From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 3AFC0A034F; Tue, 30 Mar 2021 21:37:33 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 5D917140ED1; Tue, 30 Mar 2021 21:36:50 +0200 (CEST) Received: from mga04.intel.com (mga04.intel.com [192.55.52.120]) by mails.dpdk.org (Postfix) with ESMTP id 27190140E80 for ; Tue, 30 Mar 2021 21:36:41 +0200 (CEST) IronPort-SDR: MQJ+AxIlUi6PnYiwvvYYuv54NYlgQUTrBS1Sz/ortmLAxPu0FVCzE+RLHbM0hmEnsEWJxyVzwJ xBPr1jCAf/4A== X-IronPort-AV: E=McAfee;i="6000,8403,9939"; a="189601135" X-IronPort-AV: E=Sophos;i="5.81,291,1610438400"; d="scan'208";a="189601135" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 30 Mar 2021 12:36:39 -0700 IronPort-SDR: QyZgpAv5UmCnWck3R95hFsmGGAMEsLRrukeqqrZN5hpex3WUB0JqmCrrf6q5JXlWeoYpT+Btkl WOvV33sxL/Hg== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.81,291,1610438400"; d="scan'208";a="418309696" Received: from txasoft-yocto.an.intel.com ([10.123.72.192]) by orsmga008.jf.intel.com with ESMTP; 30 Mar 2021 12:36:39 -0700 From: Timothy McDaniel To: Cc: dev@dpdk.org, erik.g.carrillo@intel.com, gage.eads@intel.com, harry.van.haaren@intel.com, jerinj@marvell.com, thomas@monjalon.net Date: Tue, 30 Mar 2021 14:35:19 -0500 Message-Id: <1617132940-24800-7-git-send-email-timothy.mcdaniel@intel.com> X-Mailer: git-send-email 1.7.10 In-Reply-To: <1617132940-24800-1-git-send-email-timothy.mcdaniel@intel.com> References: <20210316221857.2254-2-timothy.mcdaniel@intel.com> <1617132940-24800-1-git-send-email-timothy.mcdaniel@intel.com> Subject: [dpdk-dev] [PATCH v2 06/27] event/dlb2: add V2.5 create ldb queue X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Updated low level hardware functions to add DLB 2.5 support for creating load balanced queues. Signed-off-by: Timothy McDaniel --- drivers/event/dlb2/pf/base/dlb2_resource.c | 397 ------------------ .../event/dlb2/pf/base/dlb2_resource_new.c | 391 +++++++++++++++++ 2 files changed, 391 insertions(+), 397 deletions(-) diff --git a/drivers/event/dlb2/pf/base/dlb2_resource.c b/drivers/event/dlb2/pf/base/dlb2_resource.c index 041aeaeee..f8b85bc57 100644 --- a/drivers/event/dlb2/pf/base/dlb2_resource.c +++ b/drivers/event/dlb2/pf/base/dlb2_resource.c @@ -1149,403 +1149,6 @@ unsigned int dlb2_finish_map_qid_procedures(struct dlb2_hw *hw) return num; } - -static void dlb2_configure_ldb_queue(struct dlb2_hw *hw, - struct dlb2_hw_domain *domain, - struct dlb2_ldb_queue *queue, - struct dlb2_create_ldb_queue_args *args, - bool vdev_req, - unsigned int vdev_id) -{ - union dlb2_sys_vf_ldb_vqid_v r0 = { {0} }; - union dlb2_sys_vf_ldb_vqid2qid r1 = { {0} }; - union dlb2_sys_ldb_qid2vqid r2 = { {0} }; - union dlb2_sys_ldb_vasqid_v r3 = { {0} }; - union dlb2_lsp_qid_ldb_infl_lim r4 = { {0} }; - union dlb2_lsp_qid_aqed_active_lim r5 = { {0} }; - union dlb2_aqed_pipe_qid_hid_width r6 = { {0} }; - union dlb2_sys_ldb_qid_its r7 = { {0} }; - union dlb2_lsp_qid_atm_depth_thrsh r8 = { {0} }; - union dlb2_lsp_qid_naldb_depth_thrsh r9 = { {0} }; - union dlb2_aqed_pipe_qid_fid_lim r10 = { {0} }; - union dlb2_chp_ord_qid_sn_map r11 = { {0} }; - union dlb2_sys_ldb_qid_cfg_v r12 = { {0} }; - union dlb2_sys_ldb_qid_v r13 = { {0} }; - - struct dlb2_sn_group *sn_group; - unsigned int offs; - - /* QID write permissions are turned on when the domain is started */ - r3.field.vasqid_v = 0; - - offs = domain->id.phys_id * DLB2_MAX_NUM_LDB_QUEUES + - queue->id.phys_id; - - DLB2_CSR_WR(hw, DLB2_SYS_LDB_VASQID_V(offs), r3.val); - - /* - * Unordered QIDs get 4K inflights, ordered get as many as the number - * of sequence numbers. - */ - r4.field.limit = args->num_qid_inflights; - - DLB2_CSR_WR(hw, DLB2_LSP_QID_LDB_INFL_LIM(queue->id.phys_id), r4.val); - - r5.field.limit = queue->aqed_limit; - - if (r5.field.limit > DLB2_MAX_NUM_AQED_ENTRIES) - r5.field.limit = DLB2_MAX_NUM_AQED_ENTRIES; - - DLB2_CSR_WR(hw, - DLB2_LSP_QID_AQED_ACTIVE_LIM(queue->id.phys_id), - r5.val); - - switch (args->lock_id_comp_level) { - case 64: - r6.field.compress_code = 1; - break; - case 128: - r6.field.compress_code = 2; - break; - case 256: - r6.field.compress_code = 3; - break; - case 512: - r6.field.compress_code = 4; - break; - case 1024: - r6.field.compress_code = 5; - break; - case 2048: - r6.field.compress_code = 6; - break; - case 4096: - r6.field.compress_code = 7; - break; - case 0: - case 65536: - r6.field.compress_code = 0; - } - - DLB2_CSR_WR(hw, - DLB2_AQED_PIPE_QID_HID_WIDTH(queue->id.phys_id), - r6.val); - - /* Don't timestamp QEs that pass through this queue */ - r7.field.qid_its = 0; - - DLB2_CSR_WR(hw, - DLB2_SYS_LDB_QID_ITS(queue->id.phys_id), - r7.val); - - r8.field.thresh = args->depth_threshold; - - DLB2_CSR_WR(hw, - DLB2_LSP_QID_ATM_DEPTH_THRSH(queue->id.phys_id), - r8.val); - - r9.field.thresh = args->depth_threshold; - - DLB2_CSR_WR(hw, - DLB2_LSP_QID_NALDB_DEPTH_THRSH(queue->id.phys_id), - r9.val); - - /* - * This register limits the number of inflight flows a queue can have - * at one time. It has an upper bound of 2048, but can be - * over-subscribed. 512 is chosen so that a single queue doesn't use - * the entire atomic storage, but can use a substantial portion if - * needed. - */ - r10.field.qid_fid_limit = 512; - - DLB2_CSR_WR(hw, - DLB2_AQED_PIPE_QID_FID_LIM(queue->id.phys_id), - r10.val); - - /* Configure SNs */ - sn_group = &hw->rsrcs.sn_groups[queue->sn_group]; - r11.field.mode = sn_group->mode; - r11.field.slot = queue->sn_slot; - r11.field.grp = sn_group->id; - - DLB2_CSR_WR(hw, DLB2_CHP_ORD_QID_SN_MAP(queue->id.phys_id), r11.val); - - r12.field.sn_cfg_v = (args->num_sequence_numbers != 0); - r12.field.fid_cfg_v = (args->num_atomic_inflights != 0); - - DLB2_CSR_WR(hw, DLB2_SYS_LDB_QID_CFG_V(queue->id.phys_id), r12.val); - - if (vdev_req) { - offs = vdev_id * DLB2_MAX_NUM_LDB_QUEUES + queue->id.virt_id; - - r0.field.vqid_v = 1; - - DLB2_CSR_WR(hw, DLB2_SYS_VF_LDB_VQID_V(offs), r0.val); - - r1.field.qid = queue->id.phys_id; - - DLB2_CSR_WR(hw, DLB2_SYS_VF_LDB_VQID2QID(offs), r1.val); - - r2.field.vqid = queue->id.virt_id; - - DLB2_CSR_WR(hw, - DLB2_SYS_LDB_QID2VQID(queue->id.phys_id), - r2.val); - } - - r13.field.qid_v = 1; - - DLB2_CSR_WR(hw, DLB2_SYS_LDB_QID_V(queue->id.phys_id), r13.val); -} - -static int -dlb2_ldb_queue_attach_to_sn_group(struct dlb2_hw *hw, - struct dlb2_ldb_queue *queue, - struct dlb2_create_ldb_queue_args *args) -{ - int slot = -1; - int i; - - queue->sn_cfg_valid = false; - - if (args->num_sequence_numbers == 0) - return 0; - - for (i = 0; i < DLB2_MAX_NUM_SEQUENCE_NUMBER_GROUPS; i++) { - struct dlb2_sn_group *group = &hw->rsrcs.sn_groups[i]; - - if (group->sequence_numbers_per_queue == - args->num_sequence_numbers && - !dlb2_sn_group_full(group)) { - slot = dlb2_sn_group_alloc_slot(group); - if (slot >= 0) - break; - } - } - - if (slot == -1) { - DLB2_HW_ERR(hw, - "[%s():%d] Internal error: no sequence number slots available\n", - __func__, __LINE__); - return -EFAULT; - } - - queue->sn_cfg_valid = true; - queue->sn_group = i; - queue->sn_slot = slot; - return 0; -} - -static int -dlb2_ldb_queue_attach_resources(struct dlb2_hw *hw, - struct dlb2_hw_domain *domain, - struct dlb2_ldb_queue *queue, - struct dlb2_create_ldb_queue_args *args) -{ - int ret; - - ret = dlb2_ldb_queue_attach_to_sn_group(hw, queue, args); - if (ret) - return ret; - - /* Attach QID inflights */ - queue->num_qid_inflights = args->num_qid_inflights; - - /* Attach atomic inflights */ - queue->aqed_limit = args->num_atomic_inflights; - - domain->num_avail_aqed_entries -= args->num_atomic_inflights; - domain->num_used_aqed_entries += args->num_atomic_inflights; - - return 0; -} - -static int -dlb2_verify_create_ldb_queue_args(struct dlb2_hw *hw, - u32 domain_id, - struct dlb2_create_ldb_queue_args *args, - struct dlb2_cmd_response *resp, - bool vdev_req, - unsigned int vdev_id) -{ - struct dlb2_hw_domain *domain; - int i; - - domain = dlb2_get_domain_from_id(hw, domain_id, vdev_req, vdev_id); - - if (domain == NULL) { - resp->status = DLB2_ST_INVALID_DOMAIN_ID; - return -EINVAL; - } - - if (!domain->configured) { - resp->status = DLB2_ST_DOMAIN_NOT_CONFIGURED; - return -EINVAL; - } - - if (domain->started) { - resp->status = DLB2_ST_DOMAIN_STARTED; - return -EINVAL; - } - - if (dlb2_list_empty(&domain->avail_ldb_queues)) { - resp->status = DLB2_ST_LDB_QUEUES_UNAVAILABLE; - return -EINVAL; - } - - if (args->num_sequence_numbers) { - for (i = 0; i < DLB2_MAX_NUM_SEQUENCE_NUMBER_GROUPS; i++) { - struct dlb2_sn_group *group = &hw->rsrcs.sn_groups[i]; - - if (group->sequence_numbers_per_queue == - args->num_sequence_numbers && - !dlb2_sn_group_full(group)) - break; - } - - if (i == DLB2_MAX_NUM_SEQUENCE_NUMBER_GROUPS) { - resp->status = DLB2_ST_SEQUENCE_NUMBERS_UNAVAILABLE; - return -EINVAL; - } - } - - if (args->num_qid_inflights > 4096) { - resp->status = DLB2_ST_INVALID_QID_INFLIGHT_ALLOCATION; - return -EINVAL; - } - - /* Inflights must be <= number of sequence numbers if ordered */ - if (args->num_sequence_numbers != 0 && - args->num_qid_inflights > args->num_sequence_numbers) { - resp->status = DLB2_ST_INVALID_QID_INFLIGHT_ALLOCATION; - return -EINVAL; - } - - if (domain->num_avail_aqed_entries < args->num_atomic_inflights) { - resp->status = DLB2_ST_ATOMIC_INFLIGHTS_UNAVAILABLE; - return -EINVAL; - } - - if (args->num_atomic_inflights && - args->lock_id_comp_level != 0 && - args->lock_id_comp_level != 64 && - args->lock_id_comp_level != 128 && - args->lock_id_comp_level != 256 && - args->lock_id_comp_level != 512 && - args->lock_id_comp_level != 1024 && - args->lock_id_comp_level != 2048 && - args->lock_id_comp_level != 4096 && - args->lock_id_comp_level != 65536) { - resp->status = DLB2_ST_INVALID_LOCK_ID_COMP_LEVEL; - return -EINVAL; - } - - return 0; -} - -static void -dlb2_log_create_ldb_queue_args(struct dlb2_hw *hw, - u32 domain_id, - struct dlb2_create_ldb_queue_args *args, - bool vdev_req, - unsigned int vdev_id) -{ - DLB2_HW_DBG(hw, "DLB2 create load-balanced queue arguments:\n"); - if (vdev_req) - DLB2_HW_DBG(hw, "(Request from vdev %d)\n", vdev_id); - DLB2_HW_DBG(hw, "\tDomain ID: %d\n", - domain_id); - DLB2_HW_DBG(hw, "\tNumber of sequence numbers: %d\n", - args->num_sequence_numbers); - DLB2_HW_DBG(hw, "\tNumber of QID inflights: %d\n", - args->num_qid_inflights); - DLB2_HW_DBG(hw, "\tNumber of ATM inflights: %d\n", - args->num_atomic_inflights); -} - -/** - * dlb2_hw_create_ldb_queue() - Allocate and initialize a DLB LDB queue. - * @hw: Contains the current state of the DLB2 hardware. - * @domain_id: Domain ID - * @args: User-provided arguments. - * @resp: Response to user. - * @vdev_req: Request came from a virtual device. - * @vdev_id: If vdev_req is true, this contains the virtual device's ID. - * - * Return: returns < 0 on error, 0 otherwise. If the driver is unable to - * satisfy a request, resp->status will be set accordingly. - */ -int dlb2_hw_create_ldb_queue(struct dlb2_hw *hw, - u32 domain_id, - struct dlb2_create_ldb_queue_args *args, - struct dlb2_cmd_response *resp, - bool vdev_req, - unsigned int vdev_id) -{ - struct dlb2_hw_domain *domain; - struct dlb2_ldb_queue *queue; - int ret; - - dlb2_log_create_ldb_queue_args(hw, domain_id, args, vdev_req, vdev_id); - - /* - * Verify that hardware resources are available before attempting to - * satisfy the request. This simplifies the error unwinding code. - */ - ret = dlb2_verify_create_ldb_queue_args(hw, - domain_id, - args, - resp, - vdev_req, - vdev_id); - if (ret) - return ret; - - domain = dlb2_get_domain_from_id(hw, domain_id, vdev_req, vdev_id); - if (domain == NULL) { - DLB2_HW_ERR(hw, - "[%s():%d] Internal error: domain not found\n", - __func__, __LINE__); - return -EFAULT; - } - - queue = DLB2_DOM_LIST_HEAD(domain->avail_ldb_queues, typeof(*queue)); - if (queue == NULL) { - DLB2_HW_ERR(hw, - "[%s():%d] Internal error: no available ldb queues\n", - __func__, __LINE__); - return -EFAULT; - } - - ret = dlb2_ldb_queue_attach_resources(hw, domain, queue, args); - if (ret < 0) { - DLB2_HW_ERR(hw, - "[%s():%d] Internal error: failed to attach the ldb queue resources\n", - __func__, __LINE__); - return ret; - } - - dlb2_configure_ldb_queue(hw, domain, queue, args, vdev_req, vdev_id); - - queue->num_mappings = 0; - - queue->configured = true; - - /* - * Configuration succeeded, so move the resource from the 'avail' to - * the 'used' list. - */ - dlb2_list_del(&domain->avail_ldb_queues, &queue->domain_list); - - dlb2_list_add(&domain->used_ldb_queues, &queue->domain_list); - - resp->status = 0; - resp->id = (vdev_req) ? queue->id.virt_id : queue->id.phys_id; - - return 0; -} - int dlb2_get_group_sequence_numbers(struct dlb2_hw *hw, unsigned int group_id) { if (group_id >= DLB2_MAX_NUM_SEQUENCE_NUMBER_GROUPS) diff --git a/drivers/event/dlb2/pf/base/dlb2_resource_new.c b/drivers/event/dlb2/pf/base/dlb2_resource_new.c index 641812412..b52d2becd 100644 --- a/drivers/event/dlb2/pf/base/dlb2_resource_new.c +++ b/drivers/event/dlb2/pf/base/dlb2_resource_new.c @@ -3581,3 +3581,394 @@ int dlb2_reset_domain(struct dlb2_hw *hw, /* Hardware reset complete. Reset the domain's software state */ return dlb2_domain_reset_software_state(hw, domain); } + +static void +dlb2_log_create_ldb_queue_args(struct dlb2_hw *hw, + u32 domain_id, + struct dlb2_create_ldb_queue_args *args, + bool vdev_req, + unsigned int vdev_id) +{ + DLB2_HW_DBG(hw, "DLB2 create load-balanced queue arguments:\n"); + if (vdev_req) + DLB2_HW_DBG(hw, "(Request from vdev %d)\n", vdev_id); + DLB2_HW_DBG(hw, "\tDomain ID: %d\n", + domain_id); + DLB2_HW_DBG(hw, "\tNumber of sequence numbers: %d\n", + args->num_sequence_numbers); + DLB2_HW_DBG(hw, "\tNumber of QID inflights: %d\n", + args->num_qid_inflights); + DLB2_HW_DBG(hw, "\tNumber of ATM inflights: %d\n", + args->num_atomic_inflights); +} + +static int +dlb2_ldb_queue_attach_to_sn_group(struct dlb2_hw *hw, + struct dlb2_ldb_queue *queue, + struct dlb2_create_ldb_queue_args *args) +{ + int slot = -1; + int i; + + queue->sn_cfg_valid = false; + + if (args->num_sequence_numbers == 0) + return 0; + + for (i = 0; i < DLB2_MAX_NUM_SEQUENCE_NUMBER_GROUPS; i++) { + struct dlb2_sn_group *group = &hw->rsrcs.sn_groups[i]; + + if (group->sequence_numbers_per_queue == + args->num_sequence_numbers && + !dlb2_sn_group_full(group)) { + slot = dlb2_sn_group_alloc_slot(group); + if (slot >= 0) + break; + } + } + + if (slot == -1) { + DLB2_HW_ERR(hw, + "[%s():%d] Internal error: no sequence number slots available\n", + __func__, __LINE__); + return -EFAULT; + } + + queue->sn_cfg_valid = true; + queue->sn_group = i; + queue->sn_slot = slot; + return 0; +} + +static int +dlb2_verify_create_ldb_queue_args(struct dlb2_hw *hw, + u32 domain_id, + struct dlb2_create_ldb_queue_args *args, + struct dlb2_cmd_response *resp, + bool vdev_req, + unsigned int vdev_id, + struct dlb2_hw_domain **out_domain, + struct dlb2_ldb_queue **out_queue) +{ + struct dlb2_hw_domain *domain; + struct dlb2_ldb_queue *queue; + int i; + + domain = dlb2_get_domain_from_id(hw, domain_id, vdev_req, vdev_id); + + if (!domain) { + resp->status = DLB2_ST_INVALID_DOMAIN_ID; + return -EINVAL; + } + + if (!domain->configured) { + resp->status = DLB2_ST_DOMAIN_NOT_CONFIGURED; + return -EINVAL; + } + + if (domain->started) { + resp->status = DLB2_ST_DOMAIN_STARTED; + return -EINVAL; + } + + queue = DLB2_DOM_LIST_HEAD(domain->avail_ldb_queues, typeof(*queue)); + if (!queue) { + resp->status = DLB2_ST_LDB_QUEUES_UNAVAILABLE; + return -EINVAL; + } + + if (args->num_sequence_numbers) { + for (i = 0; i < DLB2_MAX_NUM_SEQUENCE_NUMBER_GROUPS; i++) { + struct dlb2_sn_group *group = &hw->rsrcs.sn_groups[i]; + + if (group->sequence_numbers_per_queue == + args->num_sequence_numbers && + !dlb2_sn_group_full(group)) + break; + } + + if (i == DLB2_MAX_NUM_SEQUENCE_NUMBER_GROUPS) { + resp->status = DLB2_ST_SEQUENCE_NUMBERS_UNAVAILABLE; + return -EINVAL; + } + } + + if (args->num_qid_inflights > 4096) { + resp->status = DLB2_ST_INVALID_QID_INFLIGHT_ALLOCATION; + return -EINVAL; + } + + /* Inflights must be <= number of sequence numbers if ordered */ + if (args->num_sequence_numbers != 0 && + args->num_qid_inflights > args->num_sequence_numbers) { + resp->status = DLB2_ST_INVALID_QID_INFLIGHT_ALLOCATION; + return -EINVAL; + } + + if (domain->num_avail_aqed_entries < args->num_atomic_inflights) { + resp->status = DLB2_ST_ATOMIC_INFLIGHTS_UNAVAILABLE; + return -EINVAL; + } + + if (args->num_atomic_inflights && + args->lock_id_comp_level != 0 && + args->lock_id_comp_level != 64 && + args->lock_id_comp_level != 128 && + args->lock_id_comp_level != 256 && + args->lock_id_comp_level != 512 && + args->lock_id_comp_level != 1024 && + args->lock_id_comp_level != 2048 && + args->lock_id_comp_level != 4096 && + args->lock_id_comp_level != 65536) { + resp->status = DLB2_ST_INVALID_LOCK_ID_COMP_LEVEL; + return -EINVAL; + } + + *out_domain = domain; + *out_queue = queue; + + return 0; +} + +static int +dlb2_ldb_queue_attach_resources(struct dlb2_hw *hw, + struct dlb2_hw_domain *domain, + struct dlb2_ldb_queue *queue, + struct dlb2_create_ldb_queue_args *args) +{ + int ret; + ret = dlb2_ldb_queue_attach_to_sn_group(hw, queue, args); + if (ret) + return ret; + + /* Attach QID inflights */ + queue->num_qid_inflights = args->num_qid_inflights; + + /* Attach atomic inflights */ + queue->aqed_limit = args->num_atomic_inflights; + + domain->num_avail_aqed_entries -= args->num_atomic_inflights; + domain->num_used_aqed_entries += args->num_atomic_inflights; + + return 0; +} + +static void dlb2_configure_ldb_queue(struct dlb2_hw *hw, + struct dlb2_hw_domain *domain, + struct dlb2_ldb_queue *queue, + struct dlb2_create_ldb_queue_args *args, + bool vdev_req, + unsigned int vdev_id) +{ + struct dlb2_sn_group *sn_group; + unsigned int offs; + u32 reg = 0; + u32 alimit; + + /* QID write permissions are turned on when the domain is started */ + offs = domain->id.phys_id * DLB2_MAX_NUM_LDB_QUEUES + queue->id.phys_id; + + DLB2_CSR_WR(hw, DLB2_SYS_LDB_VASQID_V(offs), reg); + + /* + * Unordered QIDs get 4K inflights, ordered get as many as the number + * of sequence numbers. + */ + DLB2_BITS_SET(reg, args->num_qid_inflights, + DLB2_LSP_QID_LDB_INFL_LIM_LIMIT); + DLB2_CSR_WR(hw, DLB2_LSP_QID_LDB_INFL_LIM(hw->ver, + queue->id.phys_id), reg); + + alimit = queue->aqed_limit; + + if (alimit > DLB2_MAX_NUM_AQED_ENTRIES) + alimit = DLB2_MAX_NUM_AQED_ENTRIES; + + reg = 0; + DLB2_BITS_SET(reg, alimit, DLB2_LSP_QID_AQED_ACTIVE_LIM_LIMIT); + DLB2_CSR_WR(hw, + DLB2_LSP_QID_AQED_ACTIVE_LIM(hw->ver, + queue->id.phys_id), reg); + + reg = 0; + switch (args->lock_id_comp_level) { + case 64: + DLB2_BITS_SET(reg, 1, DLB2_AQED_QID_HID_WIDTH_COMPRESS_CODE); + break; + case 128: + DLB2_BITS_SET(reg, 2, DLB2_AQED_QID_HID_WIDTH_COMPRESS_CODE); + break; + case 256: + DLB2_BITS_SET(reg, 3, DLB2_AQED_QID_HID_WIDTH_COMPRESS_CODE); + break; + case 512: + DLB2_BITS_SET(reg, 4, DLB2_AQED_QID_HID_WIDTH_COMPRESS_CODE); + break; + case 1024: + DLB2_BITS_SET(reg, 5, DLB2_AQED_QID_HID_WIDTH_COMPRESS_CODE); + break; + case 2048: + DLB2_BITS_SET(reg, 6, DLB2_AQED_QID_HID_WIDTH_COMPRESS_CODE); + break; + case 4096: + DLB2_BITS_SET(reg, 7, DLB2_AQED_QID_HID_WIDTH_COMPRESS_CODE); + break; + default: + /* No compression by default */ + break; + } + + DLB2_CSR_WR(hw, DLB2_AQED_QID_HID_WIDTH(queue->id.phys_id), reg); + + reg = 0; + /* Don't timestamp QEs that pass through this queue */ + DLB2_CSR_WR(hw, DLB2_SYS_LDB_QID_ITS(queue->id.phys_id), reg); + + DLB2_BITS_SET(reg, args->depth_threshold, + DLB2_LSP_QID_ATM_DEPTH_THRSH_THRESH); + DLB2_CSR_WR(hw, + DLB2_LSP_QID_ATM_DEPTH_THRSH(hw->ver, + queue->id.phys_id), reg); + + reg = 0; + DLB2_BITS_SET(reg, args->depth_threshold, + DLB2_LSP_QID_NALDB_DEPTH_THRSH_THRESH); + DLB2_CSR_WR(hw, + DLB2_LSP_QID_NALDB_DEPTH_THRSH(hw->ver, queue->id.phys_id), + reg); + + /* + * This register limits the number of inflight flows a queue can have + * at one time. It has an upper bound of 2048, but can be + * over-subscribed. 512 is chosen so that a single queue does not use + * the entire atomic storage, but can use a substantial portion if + * needed. + */ + reg = 0; + DLB2_BITS_SET(reg, 512, DLB2_AQED_QID_FID_LIM_QID_FID_LIMIT); + DLB2_CSR_WR(hw, DLB2_AQED_QID_FID_LIM(queue->id.phys_id), reg); + + /* Configure SNs */ + reg = 0; + sn_group = &hw->rsrcs.sn_groups[queue->sn_group]; + DLB2_BITS_SET(reg, sn_group->mode, DLB2_CHP_ORD_QID_SN_MAP_MODE); + DLB2_BITS_SET(reg, queue->sn_slot, DLB2_CHP_ORD_QID_SN_MAP_SLOT); + DLB2_BITS_SET(reg, sn_group->id, DLB2_CHP_ORD_QID_SN_MAP_GRP); + + DLB2_CSR_WR(hw, + DLB2_CHP_ORD_QID_SN_MAP(hw->ver, queue->id.phys_id), reg); + + reg = 0; + DLB2_BITS_SET(reg, (args->num_sequence_numbers != 0), + DLB2_SYS_LDB_QID_CFG_V_SN_CFG_V); + DLB2_BITS_SET(reg, (args->num_atomic_inflights != 0), + DLB2_SYS_LDB_QID_CFG_V_FID_CFG_V); + + DLB2_CSR_WR(hw, DLB2_SYS_LDB_QID_CFG_V(queue->id.phys_id), reg); + + if (vdev_req) { + offs = vdev_id * DLB2_MAX_NUM_LDB_QUEUES + queue->id.virt_id; + + reg = 0; + DLB2_BIT_SET(reg, DLB2_SYS_VF_LDB_VQID_V_VQID_V); + DLB2_CSR_WR(hw, DLB2_SYS_VF_LDB_VQID_V(offs), reg); + + reg = 0; + DLB2_BITS_SET(reg, queue->id.phys_id, + DLB2_SYS_VF_LDB_VQID2QID_QID); + DLB2_CSR_WR(hw, DLB2_SYS_VF_LDB_VQID2QID(offs), reg); + + reg = 0; + DLB2_BITS_SET(reg, queue->id.virt_id, + DLB2_SYS_LDB_QID2VQID_VQID); + DLB2_CSR_WR(hw, DLB2_SYS_LDB_QID2VQID(queue->id.phys_id), reg); + } + + reg = 0; + DLB2_BIT_SET(reg, DLB2_SYS_LDB_QID_V_QID_V); + DLB2_CSR_WR(hw, DLB2_SYS_LDB_QID_V(queue->id.phys_id), reg); +} + +/** + * dlb2_hw_create_ldb_queue() - create a load-balanced queue + * @hw: dlb2_hw handle for a particular device. + * @domain_id: domain ID. + * @args: queue creation arguments. + * @resp: response structure. + * @vdev_req: indicates whether this request came from a vdev. + * @vdev_id: If vdev_req is true, this contains the vdev's ID. + * + * This function creates a load-balanced queue. + * + * A vdev can be either an SR-IOV virtual function or a Scalable IOV virtual + * device. + * + * Return: + * Returns 0 upon success, < 0 otherwise. If an error occurs, resp->status is + * assigned a detailed error code from enum dlb2_error. If successful, resp->id + * contains the queue ID. + * + * resp->id contains a virtual ID if vdev_req is true. + * + * Errors: + * EINVAL - A requested resource is unavailable, the domain is not configured, + * the domain has already been started, or the requested queue name is + * already in use. + * EFAULT - Internal error (resp->status not set). + */ +int dlb2_hw_create_ldb_queue(struct dlb2_hw *hw, + u32 domain_id, + struct dlb2_create_ldb_queue_args *args, + struct dlb2_cmd_response *resp, + bool vdev_req, + unsigned int vdev_id) +{ + struct dlb2_hw_domain *domain; + struct dlb2_ldb_queue *queue; + int ret; + + dlb2_log_create_ldb_queue_args(hw, domain_id, args, vdev_req, vdev_id); + + /* + * Verify that hardware resources are available before attempting to + * satisfy the request. This simplifies the error unwinding code. + */ + ret = dlb2_verify_create_ldb_queue_args(hw, + domain_id, + args, + resp, + vdev_req, + vdev_id, + &domain, + &queue); + if (ret) + return ret; + + ret = dlb2_ldb_queue_attach_resources(hw, domain, queue, args); + + if (ret) { + DLB2_HW_ERR(hw, + "[%s():%d] Internal error: failed to attach the ldb queue resources\n", + __func__, __LINE__); + return ret; + } + + dlb2_configure_ldb_queue(hw, domain, queue, args, vdev_req, vdev_id); + + queue->num_mappings = 0; + + queue->configured = true; + + /* + * Configuration succeeded, so move the resource from the 'avail' to + * the 'used' list. + */ + dlb2_list_del(&domain->avail_ldb_queues, &queue->domain_list); + + dlb2_list_add(&domain->used_ldb_queues, &queue->domain_list); + + resp->status = 0; + resp->id = (vdev_req) ? queue->id.virt_id : queue->id.phys_id; + + return 0; +} -- 2.23.0