From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id E66C8A04E7; Mon, 2 Nov 2020 00:32:31 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 25768592C; Mon, 2 Nov 2020 00:28:54 +0100 (CET) Received: from mga12.intel.com (mga12.intel.com [192.55.52.136]) by dpdk.org (Postfix) with ESMTP id 0CFD52BFF for ; Mon, 2 Nov 2020 00:28:39 +0100 (CET) IronPort-SDR: 2VnqGi++ihVMxWRMN0z45dy6O+csqh3Xpcr6vsQNvtPzpXC27G+xmOm4RijEf/uQwX2z0YlwmT ckSCp4eEbhxg== X-IronPort-AV: E=McAfee;i="6000,8403,9792"; a="148099230" X-IronPort-AV: E=Sophos;i="5.77,443,1596524400"; d="scan'208";a="148099230" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga007.fm.intel.com ([10.253.24.52]) by fmsmga106.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 01 Nov 2020 15:28:39 -0800 IronPort-SDR: JY0UmQEEXxbIKj5uJ33qNdvEsGtMvrAfdMV+XQ1wL++TnXe+xpWpPxTxTW/uf/nx2bfK9zUKhv XlLZhFPi9KlA== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.77,443,1596524400"; d="scan'208";a="305521526" Received: from txasoft-yocto.an.intel.com ([10.123.72.192]) by fmsmga007.fm.intel.com with ESMTP; 01 Nov 2020 15:28:38 -0800 From: Timothy McDaniel To: Cc: dev@dpdk.org, erik.g.carrillo@intel.com, gage.eads@intel.com, harry.van.haaren@intel.com, jerinj@marvell.com, thomas@monjalon.net Date: Sun, 1 Nov 2020 17:30:02 -0600 Message-Id: <1604273415-13912-11-git-send-email-timothy.mcdaniel@intel.com> X-Mailer: git-send-email 1.7.10 In-Reply-To: <1604273415-13912-1-git-send-email-timothy.mcdaniel@intel.com> References: <20200612212434.6852-2-timothy.mcdaniel@intel.com> <1604273415-13912-1-git-send-email-timothy.mcdaniel@intel.com> Subject: [dpdk-dev] [PATCH v16 10/23] event/dlb: add infos get and configure X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Add support for configuring the DLB hardware. In particular, this patch configures the DLB hardware's scheduling domain, such that it is provisioned with the requested number of ports and queues, provided sufficient resources are available. Individual queues and ports are configured later in port setup and eventdev start. Signed-off-by: Timothy McDaniel Reviewed-by: Gage Eads --- doc/guides/eventdevs/dlb.rst | 48 + drivers/event/dlb/dlb.c | 397 +++ drivers/event/dlb/dlb_iface.c | 11 + drivers/event/dlb/dlb_iface.h | 11 + drivers/event/dlb/pf/base/dlb_resource.c | 4100 +++++++++++++++++++++++++++++- drivers/event/dlb/pf/dlb_pf.c | 88 + 6 files changed, 4562 insertions(+), 93 deletions(-) diff --git a/doc/guides/eventdevs/dlb.rst b/doc/guides/eventdevs/dlb.rst index 92341c0..2d7999b 100644 --- a/doc/guides/eventdevs/dlb.rst +++ b/doc/guides/eventdevs/dlb.rst @@ -34,3 +34,51 @@ detailed understanding of the hardware, but these details are important when writing high-performance code. This section describes the places where the eventdev API and DLB misalign. +Scheduling Domain Configuration +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + +There are 32 scheduling domainis the DLB. +When one is configured, it allocates load-balanced and +directed queues, ports, credits, and other hardware resources. Some +resource allocations are user-controlled -- the number of queues, for example +-- and others, like credit pools (one directed and one load-balanced pool per +scheduling domain), are not. + +The DLB is a closed system eventdev, and as such the ``nb_events_limit`` device +setup argument and the per-port ``new_event_threshold`` argument apply as +defined in the eventdev header file. The limit is applied to all enqueues, +regardless of whether it will consume a directed or load-balanced credit. + +Reconfiguration +~~~~~~~~~~~~~~~ + +The Eventdev API allows one to reconfigure a device, its ports, and its queues +by first stopping the device, calling the configuration function(s), then +restarting the device. The DLB does not support configuring an individual queue +or port without first reconfiguring the entire device, however, so there are +certain reconfiguration sequences that are valid in the eventdev API but not +supported by the PMD. + +Specifically, the PMD supports the following configuration sequence: +1. Configure and start the device +2. Stop the device +3. (Optional) Reconfigure the device +4. (Optional) If step 3 is run: + + a. Setup queue(s). The reconfigured queue(s) lose their previous port links. + b. The reconfigured port(s) lose their previous queue links. + +5. (Optional, only if steps 4a and 4b are run) Link port(s) to queue(s) +6. Restart the device. If the device is reconfigured in step 3 but one or more + of its ports or queues are not, the PMD will apply their previous + configuration (including port->queue links) at this time. + +The PMD does not support the following configuration sequences: +1. Configure and start the device +2. Stop the device +3. Setup queue or setup port +4. Start the device + +This sequence is not supported because the event device must be reconfigured +before its ports or queues can be. + diff --git a/drivers/event/dlb/dlb.c b/drivers/event/dlb/dlb.c index 62b9695..c038794 100644 --- a/drivers/event/dlb/dlb.c +++ b/drivers/event/dlb/dlb.c @@ -139,6 +139,19 @@ dlb_hw_query_resources(struct dlb_eventdev *dlb) return 0; } +static void +dlb_free_qe_mem(struct dlb_port *qm_port) +{ + if (qm_port == NULL) + return; + + rte_free(qm_port->qe4); + qm_port->qe4 = NULL; + + rte_free(qm_port->consume_qe); + qm_port->consume_qe = NULL; +} + /* Wrapper for string to int conversion. Substituted for atoi(...), which is * unsafe. */ @@ -231,6 +244,388 @@ set_num_dir_credits(const char *key __rte_unused, DLB_MAX_NUM_DIR_CREDITS); return -EINVAL; } + return 0; +} + +/* VDEV-only notes: + * This function first unmaps all memory mappings and closes the + * domain's file descriptor, which causes the driver to reset the + * scheduling domain. Once that completes (when close() returns), we + * can safely free the dynamically allocated memory used by the + * scheduling domain. + * + * PF-only notes: + * We will maintain a use count and use that to determine when + * a reset is required. In PF mode, we never mmap, or munmap + * device memory, and we own the entire physical PCI device. + */ + +static void +dlb_hw_reset_sched_domain(const struct rte_eventdev *dev, bool reconfig) +{ + struct dlb_eventdev *dlb = dlb_pmd_priv(dev); + enum dlb_configuration_state config_state; + int i, j; + + /* Close and reset the domain */ + dlb_iface_domain_close(dlb); + + /* Free all dynamically allocated port memory */ + for (i = 0; i < dlb->num_ports; i++) + dlb_free_qe_mem(&dlb->ev_ports[i].qm_port); + + /* If reconfiguring, mark the device's queues and ports as "previously + * configured." If the user does not reconfigure them, the PMD will + * reapply their previous configuration when the device is started. + */ + config_state = (reconfig) ? DLB_PREV_CONFIGURED : DLB_NOT_CONFIGURED; + + for (i = 0; i < dlb->num_ports; i++) { + dlb->ev_ports[i].qm_port.config_state = config_state; + /* Reset setup_done so ports can be reconfigured */ + dlb->ev_ports[i].setup_done = false; + for (j = 0; j < DLB_MAX_NUM_QIDS_PER_LDB_CQ; j++) + dlb->ev_ports[i].link[j].mapped = false; + } + + for (i = 0; i < dlb->num_queues; i++) + dlb->ev_queues[i].qm_queue.config_state = config_state; + + for (i = 0; i < DLB_MAX_NUM_QUEUES; i++) + dlb->ev_queues[i].setup_done = false; + + dlb->num_ports = 0; + dlb->num_ldb_ports = 0; + dlb->num_dir_ports = 0; + dlb->num_queues = 0; + dlb->num_ldb_queues = 0; + dlb->num_dir_queues = 0; + dlb->configured = false; +} + +static int +dlb_ldb_credit_pool_create(struct dlb_hw_dev *handle) +{ + struct dlb_create_ldb_pool_args cfg; + struct dlb_cmd_response response; + int ret; + + if (handle == NULL) + return -EINVAL; + + if (!handle->cfg.resources.num_ldb_credits) { + handle->cfg.ldb_credit_pool_id = 0; + handle->cfg.num_ldb_credits = 0; + return 0; + } + + cfg.response = (uintptr_t)&response; + cfg.num_ldb_credits = handle->cfg.resources.num_ldb_credits; + + ret = dlb_iface_ldb_credit_pool_create(handle, + &cfg); + if (ret < 0) { + DLB_LOG_ERR("dlb: ldb_credit_pool_create ret=%d (driver status: %s)\n", + ret, dlb_error_strings[response.status]); + } + + handle->cfg.ldb_credit_pool_id = response.id; + handle->cfg.num_ldb_credits = cfg.num_ldb_credits; + + return ret; +} + +static int +dlb_dir_credit_pool_create(struct dlb_hw_dev *handle) +{ + struct dlb_create_dir_pool_args cfg; + struct dlb_cmd_response response; + int ret; + + if (handle == NULL) + return -EINVAL; + + if (!handle->cfg.resources.num_dir_credits) { + handle->cfg.dir_credit_pool_id = 0; + handle->cfg.num_dir_credits = 0; + return 0; + } + + cfg.response = (uintptr_t)&response; + cfg.num_dir_credits = handle->cfg.resources.num_dir_credits; + + ret = dlb_iface_dir_credit_pool_create(handle, &cfg); + if (ret < 0) + DLB_LOG_ERR("dlb: dir_credit_pool_create ret=%d (driver status: %s)\n", + ret, dlb_error_strings[response.status]); + + handle->cfg.dir_credit_pool_id = response.id; + handle->cfg.num_dir_credits = cfg.num_dir_credits; + + return ret; +} + +static int +dlb_hw_create_sched_domain(struct dlb_hw_dev *handle, + struct dlb_eventdev *dlb, + const struct dlb_hw_rsrcs *resources_asked) +{ + int ret = 0; + struct dlb_create_sched_domain_args *config_params; + struct dlb_cmd_response response; + + if (resources_asked == NULL) { + DLB_LOG_ERR("dlb: dlb_create NULL parameter\n"); + ret = EINVAL; + goto error_exit; + } + + /* Map generic qm resources to dlb resources */ + config_params = &handle->cfg.resources; + + config_params->response = (uintptr_t)&response; + + /* DIR ports and queues */ + + config_params->num_dir_ports = + resources_asked->num_dir_ports; + + config_params->num_dir_credits = + resources_asked->num_dir_credits; + + /* LDB ports and queues */ + + config_params->num_ldb_queues = + resources_asked->num_ldb_queues; + + config_params->num_ldb_ports = + resources_asked->num_ldb_ports; + + config_params->num_ldb_credits = + resources_asked->num_ldb_credits; + + config_params->num_atomic_inflights = + dlb->num_atm_inflights_per_queue * + config_params->num_ldb_queues; + + config_params->num_hist_list_entries = config_params->num_ldb_ports * + DLB_NUM_HIST_LIST_ENTRIES_PER_LDB_PORT; + + /* dlb limited to 1 credit pool per queue type */ + config_params->num_ldb_credit_pools = 1; + config_params->num_dir_credit_pools = 1; + + DLB_LOG_DBG("sched domain create - ldb_qs=%d, ldb_ports=%d, dir_ports=%d, atomic_inflights=%d, hist_list_entries=%d, ldb_credits=%d, dir_credits=%d, ldb_cred_pools=%d, dir-credit_pools=%d\n", + config_params->num_ldb_queues, + config_params->num_ldb_ports, + config_params->num_dir_ports, + config_params->num_atomic_inflights, + config_params->num_hist_list_entries, + config_params->num_ldb_credits, + config_params->num_dir_credits, + config_params->num_ldb_credit_pools, + config_params->num_dir_credit_pools); + + /* Configure the QM */ + + ret = dlb_iface_sched_domain_create(handle, config_params); + if (ret < 0) { + DLB_LOG_ERR("dlb: domain create failed, device_id = %d, (driver ret = %d, extra status: %s)\n", + handle->device_id, + ret, + dlb_error_strings[response.status]); + goto error_exit; + } + + handle->domain_id = response.id; + handle->domain_id_valid = 1; + + config_params->response = 0; + + ret = dlb_ldb_credit_pool_create(handle); + if (ret < 0) { + DLB_LOG_ERR("dlb: create ldb credit pool failed\n"); + goto error_exit2; + } + + ret = dlb_dir_credit_pool_create(handle); + if (ret < 0) { + DLB_LOG_ERR("dlb: create dir credit pool failed\n"); + goto error_exit2; + } + + handle->cfg.configured = true; + + return 0; + +error_exit2: + dlb_iface_domain_close(dlb); + +error_exit: + return ret; +} + +/* End HW specific */ +static void +dlb_eventdev_info_get(struct rte_eventdev *dev, + struct rte_event_dev_info *dev_info) +{ + struct dlb_eventdev *dlb = dlb_pmd_priv(dev); + int ret; + + ret = dlb_hw_query_resources(dlb); + if (ret) { + const struct rte_eventdev_data *data = dev->data; + + DLB_LOG_ERR("get resources err=%d, devid=%d\n", + ret, data->dev_id); + /* fn is void, so fall through and return values set up in + * probe + */ + } + + /* Add num resources currently owned by this domain. + * These would become available if the scheduling domain were reset due + * to the application recalling eventdev_configure to *reconfigure* the + * domain. + */ + evdev_dlb_default_info.max_event_ports += dlb->num_ldb_ports; + evdev_dlb_default_info.max_event_queues += dlb->num_ldb_queues; + evdev_dlb_default_info.max_num_events += dlb->num_ldb_credits; + + /* In DLB A-stepping hardware, applications are limited to 128 + * configured ports (load-balanced or directed). The reported number of + * available ports must reflect this. + */ + if (dlb->revision < DLB_REV_B0) { + int used_ports; + + used_ports = DLB_MAX_NUM_LDB_PORTS + DLB_MAX_NUM_DIR_PORTS - + dlb->hw_rsrc_query_results.num_ldb_ports - + dlb->hw_rsrc_query_results.num_dir_ports; + + evdev_dlb_default_info.max_event_ports = + RTE_MIN(evdev_dlb_default_info.max_event_ports, + 128 - used_ports); + } + + evdev_dlb_default_info.max_event_queues = + RTE_MIN(evdev_dlb_default_info.max_event_queues, + RTE_EVENT_MAX_QUEUES_PER_DEV); + + evdev_dlb_default_info.max_num_events = + RTE_MIN(evdev_dlb_default_info.max_num_events, + dlb->max_num_events_override); + + *dev_info = evdev_dlb_default_info; +} + +/* Note: 1 QM instance per QM device, QM instance/device == event device */ +static int +dlb_eventdev_configure(const struct rte_eventdev *dev) +{ + struct dlb_eventdev *dlb = dlb_pmd_priv(dev); + struct dlb_hw_dev *handle = &dlb->qm_instance; + struct dlb_hw_rsrcs *rsrcs = &handle->info.hw_rsrc_max; + const struct rte_eventdev_data *data = dev->data; + const struct rte_event_dev_config *config = &data->dev_conf; + int ret; + + /* If this eventdev is already configured, we must release the current + * scheduling domain before attempting to configure a new one. + */ + if (dlb->configured) { + dlb_hw_reset_sched_domain(dev, true); + + ret = dlb_hw_query_resources(dlb); + if (ret) { + DLB_LOG_ERR("get resources err=%d, devid=%d\n", + ret, data->dev_id); + return ret; + } + } + + if (config->nb_event_queues > rsrcs->num_queues) { + DLB_LOG_ERR("nb_event_queues parameter (%d) exceeds the QM device's capabilities (%d).\n", + config->nb_event_queues, + rsrcs->num_queues); + return -EINVAL; + } + if (config->nb_event_ports > (rsrcs->num_ldb_ports + + rsrcs->num_dir_ports)) { + DLB_LOG_ERR("nb_event_ports parameter (%d) exceeds the QM device's capabilities (%d).\n", + config->nb_event_ports, + (rsrcs->num_ldb_ports + rsrcs->num_dir_ports)); + return -EINVAL; + } + if (config->nb_events_limit > rsrcs->nb_events_limit) { + DLB_LOG_ERR("nb_events_limit parameter (%d) exceeds the QM device's capabilities (%d).\n", + config->nb_events_limit, + rsrcs->nb_events_limit); + return -EINVAL; + } + + if (config->event_dev_cfg & RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT) + dlb->global_dequeue_wait = false; + else { + uint32_t timeout32; + + dlb->global_dequeue_wait = true; + + timeout32 = config->dequeue_timeout_ns; + + dlb->global_dequeue_wait_ticks = + timeout32 * (rte_get_timer_hz() / 1E9); + } + + /* Does this platform support umonitor/umwait? */ + if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_WAITPKG)) { + if (RTE_LIBRTE_PMD_DLB_UMWAIT_CTL_STATE != 0 && + RTE_LIBRTE_PMD_DLB_UMWAIT_CTL_STATE != 1) { + DLB_LOG_ERR("invalid value (%d) for RTE_LIBRTE_PMD_DLB_UMWAIT_CTL_STATE must be 0 or 1.\n", + RTE_LIBRTE_PMD_DLB_UMWAIT_CTL_STATE); + return -EINVAL; + } + dlb->umwait_allowed = true; + } + + rsrcs->num_dir_ports = config->nb_single_link_event_port_queues; + rsrcs->num_ldb_ports = config->nb_event_ports - rsrcs->num_dir_ports; + /* 1 dir queue per dir port */ + rsrcs->num_ldb_queues = config->nb_event_queues - rsrcs->num_dir_ports; + + /* Scale down nb_events_limit by 4 for directed credits, since there + * are 4x as many load-balanced credits. + */ + rsrcs->num_ldb_credits = 0; + rsrcs->num_dir_credits = 0; + + if (rsrcs->num_ldb_queues) + rsrcs->num_ldb_credits = config->nb_events_limit; + if (rsrcs->num_dir_ports) + rsrcs->num_dir_credits = config->nb_events_limit / 4; + if (dlb->num_dir_credits_override != -1) + rsrcs->num_dir_credits = dlb->num_dir_credits_override; + + if (dlb_hw_create_sched_domain(handle, dlb, rsrcs) < 0) { + DLB_LOG_ERR("dlb_hw_create_sched_domain failed\n"); + return -ENODEV; + } + + dlb->new_event_limit = config->nb_events_limit; + __atomic_store_n(&dlb->inflights, 0, __ATOMIC_SEQ_CST); + + /* Save number of ports/queues for this event dev */ + dlb->num_ports = config->nb_event_ports; + dlb->num_queues = config->nb_event_queues; + dlb->num_dir_ports = rsrcs->num_dir_ports; + dlb->num_ldb_ports = dlb->num_ports - dlb->num_dir_ports; + dlb->num_ldb_queues = dlb->num_queues - dlb->num_dir_ports; + dlb->num_dir_queues = dlb->num_dir_ports; + dlb->num_ldb_credits = rsrcs->num_ldb_credits; + dlb->num_dir_credits = rsrcs->num_dir_credits; + + dlb->configured = true; return 0; } @@ -309,6 +704,8 @@ void dlb_entry_points_init(struct rte_eventdev *dev) { static struct rte_eventdev_ops dlb_eventdev_entry_ops = { + .dev_infos_get = dlb_eventdev_info_get, + .dev_configure = dlb_eventdev_configure, .dump = dlb_eventdev_dump, .xstats_get = dlb_eventdev_xstats_get, .xstats_get_names = dlb_eventdev_xstats_get_names, diff --git a/drivers/event/dlb/dlb_iface.c b/drivers/event/dlb/dlb_iface.c index dd72120..f3e82f2 100644 --- a/drivers/event/dlb/dlb_iface.c +++ b/drivers/event/dlb/dlb_iface.c @@ -16,12 +16,23 @@ void (*dlb_iface_low_level_io_init)(struct dlb_eventdev *dlb); int (*dlb_iface_open)(struct dlb_hw_dev *handle, const char *name); +void (*dlb_iface_domain_close)(struct dlb_eventdev *dlb); + int (*dlb_iface_get_device_version)(struct dlb_hw_dev *handle, uint8_t *revision); int (*dlb_iface_get_num_resources)(struct dlb_hw_dev *handle, struct dlb_get_num_resources_args *rsrcs); +int (*dlb_iface_sched_domain_create)(struct dlb_hw_dev *handle, + struct dlb_create_sched_domain_args *args); + +int (*dlb_iface_ldb_credit_pool_create)(struct dlb_hw_dev *handle, + struct dlb_create_ldb_pool_args *cfg); + +int (*dlb_iface_dir_credit_pool_create)(struct dlb_hw_dev *handle, + struct dlb_create_dir_pool_args *cfg); + int (*dlb_iface_get_cq_poll_mode)(struct dlb_hw_dev *handle, enum dlb_cq_poll_modes *mode); diff --git a/drivers/event/dlb/dlb_iface.h b/drivers/event/dlb/dlb_iface.h index 416d1b3..d576232 100644 --- a/drivers/event/dlb/dlb_iface.h +++ b/drivers/event/dlb/dlb_iface.h @@ -15,12 +15,23 @@ extern void (*dlb_iface_low_level_io_init)(struct dlb_eventdev *dlb); extern int (*dlb_iface_open)(struct dlb_hw_dev *handle, const char *name); +extern void (*dlb_iface_domain_close)(struct dlb_eventdev *dlb); + extern int (*dlb_iface_get_device_version)(struct dlb_hw_dev *handle, uint8_t *revision); extern int (*dlb_iface_get_num_resources)(struct dlb_hw_dev *handle, struct dlb_get_num_resources_args *rsrcs); +extern int (*dlb_iface_sched_domain_create)(struct dlb_hw_dev *handle, + struct dlb_create_sched_domain_args *args); + +extern int (*dlb_iface_ldb_credit_pool_create)(struct dlb_hw_dev *handle, + struct dlb_create_ldb_pool_args *cfg); + +extern int (*dlb_iface_dir_credit_pool_create)(struct dlb_hw_dev *handle, + struct dlb_create_dir_pool_args *cfg); + extern int (*dlb_iface_get_cq_poll_mode)(struct dlb_hw_dev *handle, enum dlb_cq_poll_modes *mode); diff --git a/drivers/event/dlb/pf/base/dlb_resource.c b/drivers/event/dlb/pf/base/dlb_resource.c index 9c4267b..2f8ffec 100644 --- a/drivers/event/dlb/pf/base/dlb_resource.c +++ b/drivers/event/dlb/pf/base/dlb_resource.c @@ -9,107 +9,30 @@ #include "dlb_osdep_bitmap.h" #include "dlb_osdep_types.h" #include "dlb_regs.h" +#include "../../dlb_priv.h" +#include "../../dlb_inline_fns.h" -void dlb_disable_dp_vasr_feature(struct dlb_hw *hw) -{ - union dlb_dp_dir_csr_ctrl r0; - - r0.val = DLB_CSR_RD(hw, DLB_DP_DIR_CSR_CTRL); - - r0.field.cfg_vasr_dis = 1; - - DLB_CSR_WR(hw, DLB_DP_DIR_CSR_CTRL, r0.val); -} - -void dlb_enable_excess_tokens_alarm(struct dlb_hw *hw) -{ - union dlb_chp_cfg_chp_csr_ctrl r0; - - r0.val = DLB_CSR_RD(hw, DLB_CHP_CFG_CHP_CSR_CTRL); - - r0.val |= 1 << DLB_CHP_CFG_EXCESS_TOKENS_SHIFT; - - DLB_CSR_WR(hw, DLB_CHP_CFG_CHP_CSR_CTRL, r0.val); -} - -void dlb_hw_enable_sparse_ldb_cq_mode(struct dlb_hw *hw) -{ - union dlb_sys_cq_mode r0; - - r0.val = DLB_CSR_RD(hw, DLB_SYS_CQ_MODE); - - r0.field.ldb_cq64 = 1; - - DLB_CSR_WR(hw, DLB_SYS_CQ_MODE, r0.val); -} +#define DLB_DOM_LIST_HEAD(head, type) \ + DLB_LIST_HEAD((head), type, domain_list) -void dlb_hw_enable_sparse_dir_cq_mode(struct dlb_hw *hw) -{ - union dlb_sys_cq_mode r0; - - r0.val = DLB_CSR_RD(hw, DLB_SYS_CQ_MODE); - - r0.field.dir_cq64 = 1; - - DLB_CSR_WR(hw, DLB_SYS_CQ_MODE, r0.val); -} +#define DLB_FUNC_LIST_HEAD(head, type) \ + DLB_LIST_HEAD((head), type, func_list) -void dlb_hw_disable_pf_to_vf_isr_pend_err(struct dlb_hw *hw) -{ - union dlb_sys_sys_alarm_int_enable r0; +#define DLB_DOM_LIST_FOR(head, ptr, iter) \ + DLB_LIST_FOR_EACH(head, ptr, domain_list, iter) - r0.val = DLB_CSR_RD(hw, DLB_SYS_SYS_ALARM_INT_ENABLE); +#define DLB_FUNC_LIST_FOR(head, ptr, iter) \ + DLB_LIST_FOR_EACH(head, ptr, func_list, iter) - r0.field.pf_to_vf_isr_pend_error = 0; +#define DLB_DOM_LIST_FOR_SAFE(head, ptr, ptr_tmp, it, it_tmp) \ + DLB_LIST_FOR_EACH_SAFE((head), ptr, ptr_tmp, domain_list, it, it_tmp) - DLB_CSR_WR(hw, DLB_SYS_SYS_ALARM_INT_ENABLE, r0.val); -} +#define DLB_FUNC_LIST_FOR_SAFE(head, ptr, ptr_tmp, it, it_tmp) \ + DLB_LIST_FOR_EACH_SAFE((head), ptr, ptr_tmp, func_list, it, it_tmp) -void dlb_hw_get_num_resources(struct dlb_hw *hw, - struct dlb_get_num_resources_args *arg) +static inline void dlb_flush_csr(struct dlb_hw *hw) { - struct dlb_function_resources *rsrcs; - struct dlb_bitmap *map; - - rsrcs = &hw->pf; - - arg->num_sched_domains = rsrcs->num_avail_domains; - - arg->num_ldb_queues = rsrcs->num_avail_ldb_queues; - - arg->num_ldb_ports = rsrcs->num_avail_ldb_ports; - - arg->num_dir_ports = rsrcs->num_avail_dir_pq_pairs; - - map = rsrcs->avail_aqed_freelist_entries; - - arg->num_atomic_inflights = dlb_bitmap_count(map); - - arg->max_contiguous_atomic_inflights = - dlb_bitmap_longest_set_range(map); - - map = rsrcs->avail_hist_list_entries; - - arg->num_hist_list_entries = dlb_bitmap_count(map); - - arg->max_contiguous_hist_list_entries = - dlb_bitmap_longest_set_range(map); - - map = rsrcs->avail_qed_freelist_entries; - - arg->num_ldb_credits = dlb_bitmap_count(map); - - arg->max_contiguous_ldb_credits = dlb_bitmap_longest_set_range(map); - - map = rsrcs->avail_dqed_freelist_entries; - - arg->num_dir_credits = dlb_bitmap_count(map); - - arg->max_contiguous_dir_credits = dlb_bitmap_longest_set_range(map); - - arg->num_ldb_credit_pools = rsrcs->num_avail_ldb_credit_pools; - - arg->num_dir_credit_pools = rsrcs->num_avail_dir_credit_pools; + DLB_CSR_RD(hw, DLB_SYS_TOTAL_VAS); } static void dlb_init_fn_rsrc_lists(struct dlb_function_resources *rsrc) @@ -290,6 +213,3997 @@ void dlb_resource_free(struct dlb_hw *hw) dlb_bitmap_free(hw->pf.avail_aqed_freelist_entries); } +static struct dlb_domain *dlb_get_domain_from_id(struct dlb_hw *hw, u32 id) +{ + if (id >= DLB_MAX_NUM_DOMAINS) + return NULL; + + return &hw->domains[id]; +} + +static int dlb_attach_ldb_queues(struct dlb_hw *hw, + struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + u32 num_queues, + struct dlb_cmd_response *resp) +{ + unsigned int i, j; + + if (rsrcs->num_avail_ldb_queues < num_queues) { + resp->status = DLB_ST_LDB_QUEUES_UNAVAILABLE; + return -1; + } + + for (i = 0; i < num_queues; i++) { + struct dlb_ldb_queue *queue; + + queue = DLB_FUNC_LIST_HEAD(rsrcs->avail_ldb_queues, + typeof(*queue)); + if (queue == NULL) { + DLB_HW_ERR(hw, + "[%s()] Internal error: domain validation failed\n", + __func__); + goto cleanup; + } + + dlb_list_del(&rsrcs->avail_ldb_queues, &queue->func_list); + + queue->domain_id = domain->id; + queue->owned = true; + + dlb_list_add(&domain->avail_ldb_queues, &queue->domain_list); + } + + rsrcs->num_avail_ldb_queues -= num_queues; + + return 0; + +cleanup: + + /* Return the assigned queues */ + for (j = 0; j < i; j++) { + struct dlb_ldb_queue *queue; + + queue = DLB_FUNC_LIST_HEAD(domain->avail_ldb_queues, + typeof(*queue)); + /* Unrecoverable internal error */ + if (queue == NULL) + break; + + queue->owned = false; + + dlb_list_del(&domain->avail_ldb_queues, &queue->domain_list); + + dlb_list_add(&rsrcs->avail_ldb_queues, &queue->func_list); + } + + return -EFAULT; +} + +static struct dlb_ldb_port * +dlb_get_next_ldb_port(struct dlb_hw *hw, + struct dlb_function_resources *rsrcs, + u32 domain_id) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + + /* To reduce the odds of consecutive load-balanced ports mapping to the + * same queue(s), the driver attempts to allocate ports whose neighbors + * are owned by a different domain. + */ + DLB_FUNC_LIST_FOR(rsrcs->avail_ldb_ports, port, iter) { + u32 next, prev; + u32 phys_id; + + phys_id = port->id; + next = phys_id + 1; + prev = phys_id - 1; + + if (phys_id == DLB_MAX_NUM_LDB_PORTS - 1) + next = 0; + if (phys_id == 0) + prev = DLB_MAX_NUM_LDB_PORTS - 1; + + if (!hw->rsrcs.ldb_ports[next].owned || + hw->rsrcs.ldb_ports[next].domain_id == domain_id) + continue; + + if (!hw->rsrcs.ldb_ports[prev].owned || + hw->rsrcs.ldb_ports[prev].domain_id == domain_id) + continue; + + return port; + } + + /* Failing that, the driver looks for a port with one neighbor owned by + * a different domain and the other unallocated. + */ + DLB_FUNC_LIST_FOR(rsrcs->avail_ldb_ports, port, iter) { + u32 next, prev; + u32 phys_id; + + phys_id = port->id; + next = phys_id + 1; + prev = phys_id - 1; + + if (phys_id == DLB_MAX_NUM_LDB_PORTS - 1) + next = 0; + if (phys_id == 0) + prev = DLB_MAX_NUM_LDB_PORTS - 1; + + if (!hw->rsrcs.ldb_ports[prev].owned && + hw->rsrcs.ldb_ports[next].owned && + hw->rsrcs.ldb_ports[next].domain_id != domain_id) + return port; + + if (!hw->rsrcs.ldb_ports[next].owned && + hw->rsrcs.ldb_ports[prev].owned && + hw->rsrcs.ldb_ports[prev].domain_id != domain_id) + return port; + } + + /* Failing that, the driver looks for a port with both neighbors + * unallocated. + */ + DLB_FUNC_LIST_FOR(rsrcs->avail_ldb_ports, port, iter) { + u32 next, prev; + u32 phys_id; + + phys_id = port->id; + next = phys_id + 1; + prev = phys_id - 1; + + if (phys_id == DLB_MAX_NUM_LDB_PORTS - 1) + next = 0; + if (phys_id == 0) + prev = DLB_MAX_NUM_LDB_PORTS - 1; + + if (!hw->rsrcs.ldb_ports[prev].owned && + !hw->rsrcs.ldb_ports[next].owned) + return port; + } + + /* If all else fails, the driver returns the next available port. */ + return DLB_FUNC_LIST_HEAD(rsrcs->avail_ldb_ports, typeof(*port)); +} + +static int dlb_attach_ldb_ports(struct dlb_hw *hw, + struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + u32 num_ports, + struct dlb_cmd_response *resp) +{ + unsigned int i, j; + + if (rsrcs->num_avail_ldb_ports < num_ports) { + resp->status = DLB_ST_LDB_PORTS_UNAVAILABLE; + return -1; + } + + for (i = 0; i < num_ports; i++) { + struct dlb_ldb_port *port; + + port = dlb_get_next_ldb_port(hw, rsrcs, domain->id); + + if (port == NULL) { + DLB_HW_ERR(hw, + "[%s()] Internal error: domain validation failed\n", + __func__); + goto cleanup; + } + + dlb_list_del(&rsrcs->avail_ldb_ports, &port->func_list); + + port->domain_id = domain->id; + port->owned = true; + + dlb_list_add(&domain->avail_ldb_ports, &port->domain_list); + } + + rsrcs->num_avail_ldb_ports -= num_ports; + + return 0; + +cleanup: + + /* Return the assigned ports */ + for (j = 0; j < i; j++) { + struct dlb_ldb_port *port; + + port = DLB_FUNC_LIST_HEAD(domain->avail_ldb_ports, + typeof(*port)); + /* Unrecoverable internal error */ + if (port == NULL) + break; + + port->owned = false; + + dlb_list_del(&domain->avail_ldb_ports, &port->domain_list); + + dlb_list_add(&rsrcs->avail_ldb_ports, &port->func_list); + } + + return -EFAULT; +} + +static int dlb_attach_dir_ports(struct dlb_hw *hw, + struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + u32 num_ports, + struct dlb_cmd_response *resp) +{ + unsigned int i, j; + + if (rsrcs->num_avail_dir_pq_pairs < num_ports) { + resp->status = DLB_ST_DIR_PORTS_UNAVAILABLE; + return -1; + } + + for (i = 0; i < num_ports; i++) { + struct dlb_dir_pq_pair *port; + + port = DLB_FUNC_LIST_HEAD(rsrcs->avail_dir_pq_pairs, + typeof(*port)); + if (port == NULL) { + DLB_HW_ERR(hw, + "[%s()] Internal error: domain validation failed\n", + __func__); + goto cleanup; + } + + dlb_list_del(&rsrcs->avail_dir_pq_pairs, &port->func_list); + + port->domain_id = domain->id; + port->owned = true; + + dlb_list_add(&domain->avail_dir_pq_pairs, &port->domain_list); + } + + rsrcs->num_avail_dir_pq_pairs -= num_ports; + + return 0; + +cleanup: + + /* Return the assigned ports */ + for (j = 0; j < i; j++) { + struct dlb_dir_pq_pair *port; + + port = DLB_FUNC_LIST_HEAD(domain->avail_dir_pq_pairs, + typeof(*port)); + /* Unrecoverable internal error */ + if (port == NULL) + break; + + port->owned = false; + + dlb_list_del(&domain->avail_dir_pq_pairs, &port->domain_list); + + dlb_list_add(&rsrcs->avail_dir_pq_pairs, &port->func_list); + } + + return -EFAULT; +} + +static int dlb_attach_ldb_credits(struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + u32 num_credits, + struct dlb_cmd_response *resp) +{ + struct dlb_bitmap *bitmap = rsrcs->avail_qed_freelist_entries; + + if (dlb_bitmap_count(bitmap) < (int)num_credits) { + resp->status = DLB_ST_LDB_CREDITS_UNAVAILABLE; + return -1; + } + + if (num_credits) { + int base; + + base = dlb_bitmap_find_set_bit_range(bitmap, num_credits); + if (base < 0) + goto error; + + domain->qed_freelist.base = base; + domain->qed_freelist.bound = base + num_credits; + domain->qed_freelist.offset = 0; + + dlb_bitmap_clear_range(bitmap, base, num_credits); + } + + return 0; + +error: + resp->status = DLB_ST_QED_FREELIST_ENTRIES_UNAVAILABLE; + return -1; +} + +static int dlb_attach_dir_credits(struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + u32 num_credits, + struct dlb_cmd_response *resp) +{ + struct dlb_bitmap *bitmap = rsrcs->avail_dqed_freelist_entries; + + if (dlb_bitmap_count(bitmap) < (int)num_credits) { + resp->status = DLB_ST_DIR_CREDITS_UNAVAILABLE; + return -1; + } + + if (num_credits) { + int base; + + base = dlb_bitmap_find_set_bit_range(bitmap, num_credits); + if (base < 0) + goto error; + + domain->dqed_freelist.base = base; + domain->dqed_freelist.bound = base + num_credits; + domain->dqed_freelist.offset = 0; + + dlb_bitmap_clear_range(bitmap, base, num_credits); + } + + return 0; + +error: + resp->status = DLB_ST_DQED_FREELIST_ENTRIES_UNAVAILABLE; + return -1; +} + +static int dlb_attach_ldb_credit_pools(struct dlb_hw *hw, + struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + u32 num_credit_pools, + struct dlb_cmd_response *resp) +{ + unsigned int i, j; + + if (rsrcs->num_avail_ldb_credit_pools < num_credit_pools) { + resp->status = DLB_ST_LDB_CREDIT_POOLS_UNAVAILABLE; + return -1; + } + + for (i = 0; i < num_credit_pools; i++) { + struct dlb_credit_pool *pool; + + pool = DLB_FUNC_LIST_HEAD(rsrcs->avail_ldb_credit_pools, + typeof(*pool)); + if (pool == NULL) { + DLB_HW_ERR(hw, + "[%s()] Internal error: domain validation failed\n", + __func__); + goto cleanup; + } + + dlb_list_del(&rsrcs->avail_ldb_credit_pools, + &pool->func_list); + + pool->domain_id = domain->id; + pool->owned = true; + + dlb_list_add(&domain->avail_ldb_credit_pools, + &pool->domain_list); + } + + rsrcs->num_avail_ldb_credit_pools -= num_credit_pools; + + return 0; + +cleanup: + + /* Return the assigned credit pools */ + for (j = 0; j < i; j++) { + struct dlb_credit_pool *pool; + + pool = DLB_FUNC_LIST_HEAD(domain->avail_ldb_credit_pools, + typeof(*pool)); + /* Unrecoverable internal error */ + if (pool == NULL) + break; + + pool->owned = false; + + dlb_list_del(&domain->avail_ldb_credit_pools, + &pool->domain_list); + + dlb_list_add(&rsrcs->avail_ldb_credit_pools, + &pool->func_list); + } + + return -EFAULT; +} + +static int dlb_attach_dir_credit_pools(struct dlb_hw *hw, + struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + u32 num_credit_pools, + struct dlb_cmd_response *resp) +{ + unsigned int i, j; + + if (rsrcs->num_avail_dir_credit_pools < num_credit_pools) { + resp->status = DLB_ST_DIR_CREDIT_POOLS_UNAVAILABLE; + return -1; + } + + for (i = 0; i < num_credit_pools; i++) { + struct dlb_credit_pool *pool; + + pool = DLB_FUNC_LIST_HEAD(rsrcs->avail_dir_credit_pools, + typeof(*pool)); + if (pool == NULL) { + DLB_HW_ERR(hw, + "[%s()] Internal error: domain validation failed\n", + __func__); + goto cleanup; + } + + dlb_list_del(&rsrcs->avail_dir_credit_pools, + &pool->func_list); + + pool->domain_id = domain->id; + pool->owned = true; + + dlb_list_add(&domain->avail_dir_credit_pools, + &pool->domain_list); + } + + rsrcs->num_avail_dir_credit_pools -= num_credit_pools; + + return 0; + +cleanup: + + /* Return the assigned credit pools */ + for (j = 0; j < i; j++) { + struct dlb_credit_pool *pool; + + pool = DLB_FUNC_LIST_HEAD(domain->avail_dir_credit_pools, + typeof(*pool)); + /* Unrecoverable internal error */ + if (pool == NULL) + break; + + pool->owned = false; + + dlb_list_del(&domain->avail_dir_credit_pools, + &pool->domain_list); + + dlb_list_add(&rsrcs->avail_dir_credit_pools, + &pool->func_list); + } + + return -EFAULT; +} + +static int +dlb_attach_domain_hist_list_entries(struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + u32 num_hist_list_entries, + struct dlb_cmd_response *resp) +{ + struct dlb_bitmap *bitmap; + int base; + + if (num_hist_list_entries) { + bitmap = rsrcs->avail_hist_list_entries; + + base = dlb_bitmap_find_set_bit_range(bitmap, + num_hist_list_entries); + if (base < 0) + goto error; + + domain->total_hist_list_entries = num_hist_list_entries; + domain->avail_hist_list_entries = num_hist_list_entries; + domain->hist_list_entry_base = base; + domain->hist_list_entry_offset = 0; + + dlb_bitmap_clear_range(bitmap, base, num_hist_list_entries); + } + return 0; + +error: + resp->status = DLB_ST_HIST_LIST_ENTRIES_UNAVAILABLE; + return -1; +} + +static int dlb_attach_atomic_inflights(struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + u32 num_atomic_inflights, + struct dlb_cmd_response *resp) +{ + if (num_atomic_inflights) { + struct dlb_bitmap *bitmap = + rsrcs->avail_aqed_freelist_entries; + int base; + + base = dlb_bitmap_find_set_bit_range(bitmap, + num_atomic_inflights); + if (base < 0) + goto error; + + domain->aqed_freelist.base = base; + domain->aqed_freelist.bound = base + num_atomic_inflights; + domain->aqed_freelist.offset = 0; + + dlb_bitmap_clear_range(bitmap, base, num_atomic_inflights); + } + + return 0; + +error: + resp->status = DLB_ST_ATOMIC_INFLIGHTS_UNAVAILABLE; + return -1; +} + + +static int +dlb_domain_attach_resources(struct dlb_hw *hw, + struct dlb_function_resources *rsrcs, + struct dlb_domain *domain, + struct dlb_create_sched_domain_args *args, + struct dlb_cmd_response *resp) +{ + int ret; + + ret = dlb_attach_ldb_queues(hw, + rsrcs, + domain, + args->num_ldb_queues, + resp); + if (ret < 0) + return ret; + + ret = dlb_attach_ldb_ports(hw, + rsrcs, + domain, + args->num_ldb_ports, + resp); + if (ret < 0) + return ret; + + ret = dlb_attach_dir_ports(hw, + rsrcs, + domain, + args->num_dir_ports, + resp); + if (ret < 0) + return ret; + + ret = dlb_attach_ldb_credits(rsrcs, + domain, + args->num_ldb_credits, + resp); + if (ret < 0) + return ret; + + ret = dlb_attach_dir_credits(rsrcs, + domain, + args->num_dir_credits, + resp); + if (ret < 0) + return ret; + + ret = dlb_attach_ldb_credit_pools(hw, + rsrcs, + domain, + args->num_ldb_credit_pools, + resp); + if (ret < 0) + return ret; + + ret = dlb_attach_dir_credit_pools(hw, + rsrcs, + domain, + args->num_dir_credit_pools, + resp); + if (ret < 0) + return ret; + + ret = dlb_attach_domain_hist_list_entries(rsrcs, + domain, + args->num_hist_list_entries, + resp); + if (ret < 0) + return ret; + + ret = dlb_attach_atomic_inflights(rsrcs, + domain, + args->num_atomic_inflights, + resp); + if (ret < 0) + return ret; + + domain->configured = true; + + domain->started = false; + + rsrcs->num_avail_domains--; + + return 0; +} + +static void dlb_ldb_port_cq_enable(struct dlb_hw *hw, + struct dlb_ldb_port *port) +{ + union dlb_lsp_cq_ldb_dsbl reg; + + /* Don't re-enable the port if a removal is pending. The caller should + * mark this port as enabled (if it isn't already), and when the + * removal completes the port will be enabled. + */ + if (port->num_pending_removals) + return; + + reg.field.disabled = 0; + + DLB_CSR_WR(hw, DLB_LSP_CQ_LDB_DSBL(port->id), reg.val); + + dlb_flush_csr(hw); +} + +static void dlb_dir_port_cq_enable(struct dlb_hw *hw, + struct dlb_dir_pq_pair *port) +{ + union dlb_lsp_cq_dir_dsbl reg; + + reg.field.disabled = 0; + + DLB_CSR_WR(hw, DLB_LSP_CQ_DIR_DSBL(port->id), reg.val); + + dlb_flush_csr(hw); +} + + +static void dlb_ldb_port_cq_disable(struct dlb_hw *hw, + struct dlb_ldb_port *port) +{ + union dlb_lsp_cq_ldb_dsbl reg; + + reg.field.disabled = 1; + + DLB_CSR_WR(hw, DLB_LSP_CQ_LDB_DSBL(port->id), reg.val); + + dlb_flush_csr(hw); +} + +static void dlb_dir_port_cq_disable(struct dlb_hw *hw, + struct dlb_dir_pq_pair *port) +{ + union dlb_lsp_cq_dir_dsbl reg; + + reg.field.disabled = 1; + + DLB_CSR_WR(hw, DLB_LSP_CQ_DIR_DSBL(port->id), reg.val); + + dlb_flush_csr(hw); +} + + + +void dlb_disable_dp_vasr_feature(struct dlb_hw *hw) +{ + union dlb_dp_dir_csr_ctrl r0; + + r0.val = DLB_CSR_RD(hw, DLB_DP_DIR_CSR_CTRL); + + r0.field.cfg_vasr_dis = 1; + + DLB_CSR_WR(hw, DLB_DP_DIR_CSR_CTRL, r0.val); +} + +void dlb_enable_excess_tokens_alarm(struct dlb_hw *hw) +{ + union dlb_chp_cfg_chp_csr_ctrl r0; + + r0.val = DLB_CSR_RD(hw, DLB_CHP_CFG_CHP_CSR_CTRL); + + r0.val |= 1 << DLB_CHP_CFG_EXCESS_TOKENS_SHIFT; + + DLB_CSR_WR(hw, DLB_CHP_CFG_CHP_CSR_CTRL, r0.val); +} + +void dlb_hw_enable_sparse_ldb_cq_mode(struct dlb_hw *hw) +{ + union dlb_sys_cq_mode r0; + + r0.val = DLB_CSR_RD(hw, DLB_SYS_CQ_MODE); + + r0.field.ldb_cq64 = 1; + + DLB_CSR_WR(hw, DLB_SYS_CQ_MODE, r0.val); +} + +void dlb_hw_enable_sparse_dir_cq_mode(struct dlb_hw *hw) +{ + union dlb_sys_cq_mode r0; + + r0.val = DLB_CSR_RD(hw, DLB_SYS_CQ_MODE); + + r0.field.dir_cq64 = 1; + + DLB_CSR_WR(hw, DLB_SYS_CQ_MODE, r0.val); +} + +void dlb_hw_disable_pf_to_vf_isr_pend_err(struct dlb_hw *hw) +{ + union dlb_sys_sys_alarm_int_enable r0; + + r0.val = DLB_CSR_RD(hw, DLB_SYS_SYS_ALARM_INT_ENABLE); + + r0.field.pf_to_vf_isr_pend_error = 0; + + DLB_CSR_WR(hw, DLB_SYS_SYS_ALARM_INT_ENABLE, r0.val); +} + +static unsigned int +dlb_get_num_ports_in_use(struct dlb_hw *hw) +{ + unsigned int i, n = 0; + + for (i = 0; i < DLB_MAX_NUM_LDB_PORTS; i++) + if (hw->rsrcs.ldb_ports[i].owned) + n++; + + for (i = 0; i < DLB_MAX_NUM_DIR_PORTS; i++) + if (hw->rsrcs.dir_pq_pairs[i].owned) + n++; + + return n; +} + +static bool dlb_port_find_slot(struct dlb_ldb_port *port, + enum dlb_qid_map_state state, + int *slot) +{ + int i; + + for (i = 0; i < DLB_MAX_NUM_QIDS_PER_LDB_CQ; i++) { + if (port->qid_map[i].state == state) + break; + } + + *slot = i; + + return (i < DLB_MAX_NUM_QIDS_PER_LDB_CQ); +} + +static bool dlb_port_find_slot_queue(struct dlb_ldb_port *port, + enum dlb_qid_map_state state, + struct dlb_ldb_queue *queue, + int *slot) +{ + int i; + + for (i = 0; i < DLB_MAX_NUM_QIDS_PER_LDB_CQ; i++) { + if (port->qid_map[i].state == state && + port->qid_map[i].qid == queue->id) + break; + } + + *slot = i; + + return (i < DLB_MAX_NUM_QIDS_PER_LDB_CQ); +} + +static int dlb_port_slot_state_transition(struct dlb_hw *hw, + struct dlb_ldb_port *port, + struct dlb_ldb_queue *queue, + int slot, + enum dlb_qid_map_state new_state) +{ + enum dlb_qid_map_state curr_state = port->qid_map[slot].state; + struct dlb_domain *domain; + + domain = dlb_get_domain_from_id(hw, port->domain_id); + if (domain == NULL) { + DLB_HW_ERR(hw, + "[%s()] Internal error: unable to find domain %d\n", + __func__, port->domain_id); + return -EFAULT; + } + + switch (curr_state) { + case DLB_QUEUE_UNMAPPED: + switch (new_state) { + case DLB_QUEUE_MAPPED: + queue->num_mappings++; + port->num_mappings++; + break; + case DLB_QUEUE_MAP_IN_PROGRESS: + queue->num_pending_additions++; + domain->num_pending_additions++; + break; + default: + goto error; + } + break; + case DLB_QUEUE_MAPPED: + switch (new_state) { + case DLB_QUEUE_UNMAPPED: + queue->num_mappings--; + port->num_mappings--; + break; + case DLB_QUEUE_UNMAP_IN_PROGRESS: + port->num_pending_removals++; + domain->num_pending_removals++; + break; + case DLB_QUEUE_MAPPED: + /* Priority change, nothing to update */ + break; + default: + goto error; + } + break; + case DLB_QUEUE_MAP_IN_PROGRESS: + switch (new_state) { + case DLB_QUEUE_UNMAPPED: + queue->num_pending_additions--; + domain->num_pending_additions--; + break; + case DLB_QUEUE_MAPPED: + queue->num_mappings++; + port->num_mappings++; + queue->num_pending_additions--; + domain->num_pending_additions--; + break; + default: + goto error; + } + break; + case DLB_QUEUE_UNMAP_IN_PROGRESS: + switch (new_state) { + case DLB_QUEUE_UNMAPPED: + port->num_pending_removals--; + domain->num_pending_removals--; + queue->num_mappings--; + port->num_mappings--; + break; + case DLB_QUEUE_MAPPED: + port->num_pending_removals--; + domain->num_pending_removals--; + break; + case DLB_QUEUE_UNMAP_IN_PROGRESS_PENDING_MAP: + /* Nothing to update */ + break; + default: + goto error; + } + break; + case DLB_QUEUE_UNMAP_IN_PROGRESS_PENDING_MAP: + switch (new_state) { + case DLB_QUEUE_UNMAP_IN_PROGRESS: + /* Nothing to update */ + break; + case DLB_QUEUE_UNMAPPED: + /* An UNMAP_IN_PROGRESS_PENDING_MAP slot briefly + * becomes UNMAPPED before it transitions to + * MAP_IN_PROGRESS. + */ + queue->num_mappings--; + port->num_mappings--; + port->num_pending_removals--; + domain->num_pending_removals--; + break; + default: + goto error; + } + break; + default: + goto error; + } + + port->qid_map[slot].state = new_state; + + DLB_HW_INFO(hw, + "[%s()] queue %d -> port %d state transition (%d -> %d)\n", + __func__, queue->id, port->id, curr_state, + new_state); + return 0; + +error: + DLB_HW_ERR(hw, + "[%s()] Internal error: invalid queue %d -> port %d state transition (%d -> %d)\n", + __func__, queue->id, port->id, curr_state, + new_state); + return -EFAULT; +} + +/* dlb_ldb_queue_{enable, disable}_mapped_cqs() don't operate exactly as their + * function names imply, and should only be called by the dynamic CQ mapping + * code. + */ +static void dlb_ldb_queue_disable_mapped_cqs(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_ldb_queue *queue) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + int slot; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) { + enum dlb_qid_map_state state = DLB_QUEUE_MAPPED; + + if (!dlb_port_find_slot_queue(port, state, queue, &slot)) + continue; + + if (port->enabled) + dlb_ldb_port_cq_disable(hw, port); + } +} + +static void dlb_ldb_queue_enable_mapped_cqs(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_ldb_queue *queue) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + int slot; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) { + enum dlb_qid_map_state state = DLB_QUEUE_MAPPED; + + if (!dlb_port_find_slot_queue(port, state, queue, &slot)) + continue; + + if (port->enabled) + dlb_ldb_port_cq_enable(hw, port); + } +} + +static int dlb_ldb_port_map_qid_static(struct dlb_hw *hw, + struct dlb_ldb_port *p, + struct dlb_ldb_queue *q, + u8 priority) +{ + union dlb_lsp_cq2priov r0; + union dlb_lsp_cq2qid r1; + union dlb_atm_pipe_qid_ldb_qid2cqidx r2; + union dlb_lsp_qid_ldb_qid2cqidx r3; + union dlb_lsp_qid_ldb_qid2cqidx2 r4; + enum dlb_qid_map_state state; + int i; + + /* Look for a pending or already mapped slot, else an unused slot */ + if (!dlb_port_find_slot_queue(p, DLB_QUEUE_MAP_IN_PROGRESS, q, &i) && + !dlb_port_find_slot_queue(p, DLB_QUEUE_MAPPED, q, &i) && + !dlb_port_find_slot(p, DLB_QUEUE_UNMAPPED, &i)) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: CQ has no available QID mapping slots\n", + __func__, __LINE__); + return -EFAULT; + } + + if (i >= DLB_MAX_NUM_QIDS_PER_LDB_CQ) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: port slot tracking failed\n", + __func__, __LINE__); + return -EFAULT; + } + + /* Read-modify-write the priority and valid bit register */ + r0.val = DLB_CSR_RD(hw, DLB_LSP_CQ2PRIOV(p->id)); + + r0.field.v |= 1 << i; + r0.field.prio |= (priority & 0x7) << i * 3; + + DLB_CSR_WR(hw, DLB_LSP_CQ2PRIOV(p->id), r0.val); + + /* Read-modify-write the QID map register */ + r1.val = DLB_CSR_RD(hw, DLB_LSP_CQ2QID(p->id, i / 4)); + + if (i == 0 || i == 4) + r1.field.qid_p0 = q->id; + if (i == 1 || i == 5) + r1.field.qid_p1 = q->id; + if (i == 2 || i == 6) + r1.field.qid_p2 = q->id; + if (i == 3 || i == 7) + r1.field.qid_p3 = q->id; + + DLB_CSR_WR(hw, DLB_LSP_CQ2QID(p->id, i / 4), r1.val); + + r2.val = DLB_CSR_RD(hw, + DLB_ATM_PIPE_QID_LDB_QID2CQIDX(q->id, + p->id / 4)); + + r3.val = DLB_CSR_RD(hw, + DLB_LSP_QID_LDB_QID2CQIDX(q->id, + p->id / 4)); + + r4.val = DLB_CSR_RD(hw, + DLB_LSP_QID_LDB_QID2CQIDX2(q->id, + p->id / 4)); + + switch (p->id % 4) { + case 0: + r2.field.cq_p0 |= 1 << i; + r3.field.cq_p0 |= 1 << i; + r4.field.cq_p0 |= 1 << i; + break; + + case 1: + r2.field.cq_p1 |= 1 << i; + r3.field.cq_p1 |= 1 << i; + r4.field.cq_p1 |= 1 << i; + break; + + case 2: + r2.field.cq_p2 |= 1 << i; + r3.field.cq_p2 |= 1 << i; + r4.field.cq_p2 |= 1 << i; + break; + + case 3: + r2.field.cq_p3 |= 1 << i; + r3.field.cq_p3 |= 1 << i; + r4.field.cq_p3 |= 1 << i; + break; + } + + DLB_CSR_WR(hw, + DLB_ATM_PIPE_QID_LDB_QID2CQIDX(q->id, + p->id / 4), + r2.val); + + DLB_CSR_WR(hw, + DLB_LSP_QID_LDB_QID2CQIDX(q->id, + p->id / 4), + r3.val); + + DLB_CSR_WR(hw, + DLB_LSP_QID_LDB_QID2CQIDX2(q->id, + p->id / 4), + r4.val); + + dlb_flush_csr(hw); + + p->qid_map[i].qid = q->id; + p->qid_map[i].priority = priority; + + state = DLB_QUEUE_MAPPED; + + return dlb_port_slot_state_transition(hw, p, q, i, state); +} + +static int dlb_ldb_port_set_has_work_bits(struct dlb_hw *hw, + struct dlb_ldb_port *port, + struct dlb_ldb_queue *queue, + int slot) +{ + union dlb_lsp_qid_aqed_active_cnt r0; + union dlb_lsp_qid_ldb_enqueue_cnt r1; + union dlb_lsp_ldb_sched_ctrl r2 = { {0} }; + + /* Set the atomic scheduling haswork bit */ + r0.val = DLB_CSR_RD(hw, DLB_LSP_QID_AQED_ACTIVE_CNT(queue->id)); + + r2.field.cq = port->id; + r2.field.qidix = slot; + r2.field.value = 1; + r2.field.rlist_haswork_v = r0.field.count > 0; + + /* Set the non-atomic scheduling haswork bit */ + DLB_CSR_WR(hw, DLB_LSP_LDB_SCHED_CTRL, r2.val); + + r1.val = DLB_CSR_RD(hw, DLB_LSP_QID_LDB_ENQUEUE_CNT(queue->id)); + + memset(&r2, 0, sizeof(r2)); + + r2.field.cq = port->id; + r2.field.qidix = slot; + r2.field.value = 1; + r2.field.nalb_haswork_v = (r1.field.count > 0); + + DLB_CSR_WR(hw, DLB_LSP_LDB_SCHED_CTRL, r2.val); + + dlb_flush_csr(hw); + + return 0; +} + +static void dlb_ldb_port_clear_queue_if_status(struct dlb_hw *hw, + struct dlb_ldb_port *port, + int slot) +{ + union dlb_lsp_ldb_sched_ctrl r0 = { {0} }; + + r0.field.cq = port->id; + r0.field.qidix = slot; + r0.field.value = 0; + r0.field.inflight_ok_v = 1; + + DLB_CSR_WR(hw, DLB_LSP_LDB_SCHED_CTRL, r0.val); + + dlb_flush_csr(hw); +} + +static void dlb_ldb_port_set_queue_if_status(struct dlb_hw *hw, + struct dlb_ldb_port *port, + int slot) +{ + union dlb_lsp_ldb_sched_ctrl r0 = { {0} }; + + r0.field.cq = port->id; + r0.field.qidix = slot; + r0.field.value = 1; + r0.field.inflight_ok_v = 1; + + DLB_CSR_WR(hw, DLB_LSP_LDB_SCHED_CTRL, r0.val); + + dlb_flush_csr(hw); +} + +static void dlb_ldb_queue_set_inflight_limit(struct dlb_hw *hw, + struct dlb_ldb_queue *queue) +{ + union dlb_lsp_qid_ldb_infl_lim r0 = { {0} }; + + r0.field.limit = queue->num_qid_inflights; + + DLB_CSR_WR(hw, DLB_LSP_QID_LDB_INFL_LIM(queue->id), r0.val); +} + +static void dlb_ldb_queue_clear_inflight_limit(struct dlb_hw *hw, + struct dlb_ldb_queue *queue) +{ + DLB_CSR_WR(hw, + DLB_LSP_QID_LDB_INFL_LIM(queue->id), + DLB_LSP_QID_LDB_INFL_LIM_RST); +} + +static int dlb_ldb_port_finish_map_qid_dynamic(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_ldb_port *port, + struct dlb_ldb_queue *queue) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_lsp_qid_ldb_infl_cnt r0; + enum dlb_qid_map_state state; + int slot, ret; + u8 prio; + + r0.val = DLB_CSR_RD(hw, DLB_LSP_QID_LDB_INFL_CNT(queue->id)); + + if (r0.field.count) { + DLB_HW_ERR(hw, + "[%s()] Internal error: non-zero QID inflight count\n", + __func__); + return -EFAULT; + } + + /* For each port with a pending mapping to this queue, perform the + * static mapping and set the corresponding has_work bits. + */ + state = DLB_QUEUE_MAP_IN_PROGRESS; + if (!dlb_port_find_slot_queue(port, state, queue, &slot)) + return -EINVAL; + + if (slot >= DLB_MAX_NUM_QIDS_PER_LDB_CQ) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: port slot tracking failed\n", + __func__, __LINE__); + return -EFAULT; + } + + prio = port->qid_map[slot].priority; + + /* Update the CQ2QID, CQ2PRIOV, and QID2CQIDX registers, and + * the port's qid_map state. + */ + ret = dlb_ldb_port_map_qid_static(hw, port, queue, prio); + if (ret) + return ret; + + ret = dlb_ldb_port_set_has_work_bits(hw, port, queue, slot); + if (ret) + return ret; + + /* Ensure IF_status(cq,qid) is 0 before enabling the port to + * prevent spurious schedules to cause the queue's inflight + * count to increase. + */ + dlb_ldb_port_clear_queue_if_status(hw, port, slot); + + /* Reset the queue's inflight status */ + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) { + state = DLB_QUEUE_MAPPED; + if (!dlb_port_find_slot_queue(port, state, queue, &slot)) + continue; + + dlb_ldb_port_set_queue_if_status(hw, port, slot); + } + + dlb_ldb_queue_set_inflight_limit(hw, queue); + + /* Re-enable CQs mapped to this queue */ + dlb_ldb_queue_enable_mapped_cqs(hw, domain, queue); + + /* If this queue has other mappings pending, clear its inflight limit */ + if (queue->num_pending_additions > 0) + dlb_ldb_queue_clear_inflight_limit(hw, queue); + + return 0; +} + +/** + * dlb_ldb_port_map_qid_dynamic() - perform a "dynamic" QID->CQ mapping + * @hw: dlb_hw handle for a particular device. + * @port: load-balanced port + * @queue: load-balanced queue + * @priority: queue servicing priority + * + * Returns 0 if the queue was mapped, 1 if the mapping is scheduled to occur + * at a later point, and <0 if an error occurred. + */ +static int dlb_ldb_port_map_qid_dynamic(struct dlb_hw *hw, + struct dlb_ldb_port *port, + struct dlb_ldb_queue *queue, + u8 priority) +{ + union dlb_lsp_qid_ldb_infl_cnt r0 = { {0} }; + enum dlb_qid_map_state state; + struct dlb_domain *domain; + int slot, ret; + + domain = dlb_get_domain_from_id(hw, port->domain_id); + if (domain == NULL) { + DLB_HW_ERR(hw, + "[%s()] Internal error: unable to find domain %d\n", + __func__, port->domain_id); + return -EFAULT; + } + + /* Set the QID inflight limit to 0 to prevent further scheduling of the + * queue. + */ + DLB_CSR_WR(hw, DLB_LSP_QID_LDB_INFL_LIM(queue->id), 0); + + if (!dlb_port_find_slot(port, DLB_QUEUE_UNMAPPED, &slot)) { + DLB_HW_ERR(hw, + "Internal error: No available unmapped slots\n"); + return -EFAULT; + } + + if (slot >= DLB_MAX_NUM_QIDS_PER_LDB_CQ) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: port slot tracking failed\n", + __func__, __LINE__); + return -EFAULT; + } + + port->qid_map[slot].qid = queue->id; + port->qid_map[slot].priority = priority; + + state = DLB_QUEUE_MAP_IN_PROGRESS; + ret = dlb_port_slot_state_transition(hw, port, queue, slot, state); + if (ret) + return ret; + + r0.val = DLB_CSR_RD(hw, DLB_LSP_QID_LDB_INFL_CNT(queue->id)); + + if (r0.field.count) { + /* The queue is owed completions so it's not safe to map it + * yet. Schedule a kernel thread to complete the mapping later, + * once software has completed all the queue's inflight events. + */ + if (!os_worker_active(hw)) + os_schedule_work(hw); + + return 1; + } + + /* Disable the affected CQ, and the CQs already mapped to the QID, + * before reading the QID's inflight count a second time. There is an + * unlikely race in which the QID may schedule one more QE after we + * read an inflight count of 0, and disabling the CQs guarantees that + * the race will not occur after a re-read of the inflight count + * register. + */ + if (port->enabled) + dlb_ldb_port_cq_disable(hw, port); + + dlb_ldb_queue_disable_mapped_cqs(hw, domain, queue); + + r0.val = DLB_CSR_RD(hw, DLB_LSP_QID_LDB_INFL_CNT(queue->id)); + + if (r0.field.count) { + if (port->enabled) + dlb_ldb_port_cq_enable(hw, port); + + dlb_ldb_queue_enable_mapped_cqs(hw, domain, queue); + + /* The queue is owed completions so it's not safe to map it + * yet. Schedule a kernel thread to complete the mapping later, + * once software has completed all the queue's inflight events. + */ + if (!os_worker_active(hw)) + os_schedule_work(hw); + + return 1; + } + + return dlb_ldb_port_finish_map_qid_dynamic(hw, domain, port, queue); +} + + +static int dlb_ldb_port_map_qid(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_ldb_port *port, + struct dlb_ldb_queue *queue, + u8 prio) +{ + if (domain->started) + return dlb_ldb_port_map_qid_dynamic(hw, port, queue, prio); + else + return dlb_ldb_port_map_qid_static(hw, port, queue, prio); +} + +static int dlb_ldb_port_unmap_qid(struct dlb_hw *hw, + struct dlb_ldb_port *port, + struct dlb_ldb_queue *queue) +{ + enum dlb_qid_map_state mapped, in_progress, pending_map, unmapped; + union dlb_lsp_cq2priov r0; + union dlb_atm_pipe_qid_ldb_qid2cqidx r1; + union dlb_lsp_qid_ldb_qid2cqidx r2; + union dlb_lsp_qid_ldb_qid2cqidx2 r3; + u32 queue_id; + u32 port_id; + int i; + + /* Find the queue's slot */ + mapped = DLB_QUEUE_MAPPED; + in_progress = DLB_QUEUE_UNMAP_IN_PROGRESS; + pending_map = DLB_QUEUE_UNMAP_IN_PROGRESS_PENDING_MAP; + + if (!dlb_port_find_slot_queue(port, mapped, queue, &i) && + !dlb_port_find_slot_queue(port, in_progress, queue, &i) && + !dlb_port_find_slot_queue(port, pending_map, queue, &i)) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: QID %d isn't mapped\n", + __func__, __LINE__, queue->id); + return -EFAULT; + } + + if (i >= DLB_MAX_NUM_QIDS_PER_LDB_CQ) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: port slot tracking failed\n", + __func__, __LINE__); + return -EFAULT; + } + + port_id = port->id; + queue_id = queue->id; + + /* Read-modify-write the priority and valid bit register */ + r0.val = DLB_CSR_RD(hw, DLB_LSP_CQ2PRIOV(port_id)); + + r0.field.v &= ~(1 << i); + + DLB_CSR_WR(hw, DLB_LSP_CQ2PRIOV(port_id), r0.val); + + r1.val = DLB_CSR_RD(hw, + DLB_ATM_PIPE_QID_LDB_QID2CQIDX(queue_id, + port_id / 4)); + + r2.val = DLB_CSR_RD(hw, + DLB_LSP_QID_LDB_QID2CQIDX(queue_id, + port_id / 4)); + + r3.val = DLB_CSR_RD(hw, + DLB_LSP_QID_LDB_QID2CQIDX2(queue_id, + port_id / 4)); + + switch (port_id % 4) { + case 0: + r1.field.cq_p0 &= ~(1 << i); + r2.field.cq_p0 &= ~(1 << i); + r3.field.cq_p0 &= ~(1 << i); + break; + + case 1: + r1.field.cq_p1 &= ~(1 << i); + r2.field.cq_p1 &= ~(1 << i); + r3.field.cq_p1 &= ~(1 << i); + break; + + case 2: + r1.field.cq_p2 &= ~(1 << i); + r2.field.cq_p2 &= ~(1 << i); + r3.field.cq_p2 &= ~(1 << i); + break; + + case 3: + r1.field.cq_p3 &= ~(1 << i); + r2.field.cq_p3 &= ~(1 << i); + r3.field.cq_p3 &= ~(1 << i); + break; + } + + DLB_CSR_WR(hw, + DLB_ATM_PIPE_QID_LDB_QID2CQIDX(queue_id, port_id / 4), + r1.val); + + DLB_CSR_WR(hw, + DLB_LSP_QID_LDB_QID2CQIDX(queue_id, port_id / 4), + r2.val); + + DLB_CSR_WR(hw, + DLB_LSP_QID_LDB_QID2CQIDX2(queue_id, port_id / 4), + r3.val); + + dlb_flush_csr(hw); + + unmapped = DLB_QUEUE_UNMAPPED; + + return dlb_port_slot_state_transition(hw, port, queue, i, unmapped); +} + +static int +dlb_verify_create_sched_domain_args(struct dlb_hw *hw, + struct dlb_function_resources *rsrcs, + struct dlb_create_sched_domain_args *args, + struct dlb_cmd_response *resp) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_bitmap *ldb_credit_freelist; + struct dlb_bitmap *dir_credit_freelist; + unsigned int ldb_credit_freelist_count; + unsigned int dir_credit_freelist_count; + unsigned int max_contig_aqed_entries; + unsigned int max_contig_dqed_entries; + unsigned int max_contig_qed_entries; + unsigned int max_contig_hl_entries; + struct dlb_bitmap *aqed_freelist; + enum dlb_dev_revision revision; + + ldb_credit_freelist = rsrcs->avail_qed_freelist_entries; + dir_credit_freelist = rsrcs->avail_dqed_freelist_entries; + aqed_freelist = rsrcs->avail_aqed_freelist_entries; + + ldb_credit_freelist_count = dlb_bitmap_count(ldb_credit_freelist); + dir_credit_freelist_count = dlb_bitmap_count(dir_credit_freelist); + + max_contig_hl_entries = + dlb_bitmap_longest_set_range(rsrcs->avail_hist_list_entries); + max_contig_aqed_entries = + dlb_bitmap_longest_set_range(aqed_freelist); + max_contig_qed_entries = + dlb_bitmap_longest_set_range(ldb_credit_freelist); + max_contig_dqed_entries = + dlb_bitmap_longest_set_range(dir_credit_freelist); + + if (rsrcs->num_avail_domains < 1) + resp->status = DLB_ST_DOMAIN_UNAVAILABLE; + else if (rsrcs->num_avail_ldb_queues < args->num_ldb_queues) + resp->status = DLB_ST_LDB_QUEUES_UNAVAILABLE; + else if (rsrcs->num_avail_ldb_ports < args->num_ldb_ports) + resp->status = DLB_ST_LDB_PORTS_UNAVAILABLE; + else if (args->num_ldb_queues > 0 && args->num_ldb_ports == 0) + resp->status = DLB_ST_LDB_PORT_REQUIRED_FOR_LDB_QUEUES; + else if (rsrcs->num_avail_dir_pq_pairs < args->num_dir_ports) + resp->status = DLB_ST_DIR_PORTS_UNAVAILABLE; + else if (ldb_credit_freelist_count < args->num_ldb_credits) + resp->status = DLB_ST_LDB_CREDITS_UNAVAILABLE; + else if (dir_credit_freelist_count < args->num_dir_credits) + resp->status = DLB_ST_DIR_CREDITS_UNAVAILABLE; + else if (rsrcs->num_avail_ldb_credit_pools < args->num_ldb_credit_pools) + resp->status = DLB_ST_LDB_CREDIT_POOLS_UNAVAILABLE; + else if (rsrcs->num_avail_dir_credit_pools < args->num_dir_credit_pools) + resp->status = DLB_ST_DIR_CREDIT_POOLS_UNAVAILABLE; + else if (max_contig_hl_entries < args->num_hist_list_entries) + resp->status = DLB_ST_HIST_LIST_ENTRIES_UNAVAILABLE; + else if (max_contig_aqed_entries < args->num_atomic_inflights) + resp->status = DLB_ST_ATOMIC_INFLIGHTS_UNAVAILABLE; + else if (max_contig_qed_entries < args->num_ldb_credits) + resp->status = DLB_ST_QED_FREELIST_ENTRIES_UNAVAILABLE; + else if (max_contig_dqed_entries < args->num_dir_credits) + resp->status = DLB_ST_DQED_FREELIST_ENTRIES_UNAVAILABLE; + + /* DLB A-stepping workaround for hardware write buffer lock up issue: + * limit the maximum configured ports to less than 128 and disable CQ + * occupancy interrupts. + */ + revision = os_get_dev_revision(hw); + + if (revision < DLB_B0) { + u32 n = dlb_get_num_ports_in_use(hw); + + n += args->num_ldb_ports + args->num_dir_ports; + + if (n >= DLB_A_STEP_MAX_PORTS) + resp->status = args->num_ldb_ports ? + DLB_ST_LDB_PORTS_UNAVAILABLE : + DLB_ST_DIR_PORTS_UNAVAILABLE; + } + + if (resp->status) + return -1; + + return 0; +} + + +static void +dlb_log_create_sched_domain_args(struct dlb_hw *hw, + struct dlb_create_sched_domain_args *args) +{ + DLB_HW_INFO(hw, "DLB create sched domain arguments:\n"); + DLB_HW_INFO(hw, "\tNumber of LDB queues: %d\n", + args->num_ldb_queues); + DLB_HW_INFO(hw, "\tNumber of LDB ports: %d\n", + args->num_ldb_ports); + DLB_HW_INFO(hw, "\tNumber of DIR ports: %d\n", + args->num_dir_ports); + DLB_HW_INFO(hw, "\tNumber of ATM inflights: %d\n", + args->num_atomic_inflights); + DLB_HW_INFO(hw, "\tNumber of hist list entries: %d\n", + args->num_hist_list_entries); + DLB_HW_INFO(hw, "\tNumber of LDB credits: %d\n", + args->num_ldb_credits); + DLB_HW_INFO(hw, "\tNumber of DIR credits: %d\n", + args->num_dir_credits); + DLB_HW_INFO(hw, "\tNumber of LDB credit pools: %d\n", + args->num_ldb_credit_pools); + DLB_HW_INFO(hw, "\tNumber of DIR credit pools: %d\n", + args->num_dir_credit_pools); +} + +/** + * dlb_hw_create_sched_domain() - Allocate and initialize a DLB scheduling + * domain and its resources. + * @hw: Contains the current state of the DLB hardware. + * @args: User-provided arguments. + * @resp: Response to user. + * + * Return: returns < 0 on error, 0 otherwise. If the driver is unable to + * satisfy a request, resp->status will be set accordingly. + */ +int dlb_hw_create_sched_domain(struct dlb_hw *hw, + struct dlb_create_sched_domain_args *args, + struct dlb_cmd_response *resp) +{ + struct dlb_domain *domain; + struct dlb_function_resources *rsrcs; + int ret; + + rsrcs = &hw->pf; + + dlb_log_create_sched_domain_args(hw, args); + + /* Verify that hardware resources are available before attempting to + * satisfy the request. This simplifies the error unwinding code. + */ + if (dlb_verify_create_sched_domain_args(hw, rsrcs, args, resp)) + return -EINVAL; + + domain = DLB_FUNC_LIST_HEAD(rsrcs->avail_domains, typeof(*domain)); + + /* Verification should catch this. */ + if (domain == NULL) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: no available domains\n", + __func__, __LINE__); + return -EFAULT; + } + + if (domain->configured) { + DLB_HW_ERR(hw, + "[%s()] Internal error: avail_domains contains configured domains.\n", + __func__); + return -EFAULT; + } + + dlb_init_domain_rsrc_lists(domain); + + /* Verification should catch this too. */ + ret = dlb_domain_attach_resources(hw, rsrcs, domain, args, resp); + if (ret < 0) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to verify args.\n", + __func__); + + return -EFAULT; + } + + dlb_list_del(&rsrcs->avail_domains, &domain->func_list); + + dlb_list_add(&rsrcs->used_domains, &domain->func_list); + + resp->id = domain->id; + resp->status = 0; + + return 0; +} + +static void +dlb_configure_ldb_credit_pool(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_create_ldb_pool_args *args, + struct dlb_credit_pool *pool) +{ + union dlb_sys_ldb_pool_enbld r0 = { {0} }; + union dlb_chp_ldb_pool_crd_lim r1 = { {0} }; + union dlb_chp_ldb_pool_crd_cnt r2 = { {0} }; + union dlb_chp_qed_fl_base r3 = { {0} }; + union dlb_chp_qed_fl_lim r4 = { {0} }; + union dlb_chp_qed_fl_push_ptr r5 = { {0} }; + union dlb_chp_qed_fl_pop_ptr r6 = { {0} }; + + r1.field.limit = args->num_ldb_credits; + + DLB_CSR_WR(hw, DLB_CHP_LDB_POOL_CRD_LIM(pool->id), r1.val); + + r2.field.count = args->num_ldb_credits; + + DLB_CSR_WR(hw, DLB_CHP_LDB_POOL_CRD_CNT(pool->id), r2.val); + + r3.field.base = domain->qed_freelist.base + domain->qed_freelist.offset; + + DLB_CSR_WR(hw, DLB_CHP_QED_FL_BASE(pool->id), r3.val); + + r4.field.freelist_disable = 0; + r4.field.limit = r3.field.base + args->num_ldb_credits - 1; + + DLB_CSR_WR(hw, DLB_CHP_QED_FL_LIM(pool->id), r4.val); + + r5.field.push_ptr = r3.field.base; + r5.field.generation = 1; + + DLB_CSR_WR(hw, DLB_CHP_QED_FL_PUSH_PTR(pool->id), r5.val); + + r6.field.pop_ptr = r3.field.base; + r6.field.generation = 0; + + DLB_CSR_WR(hw, DLB_CHP_QED_FL_POP_PTR(pool->id), r6.val); + + r0.field.pool_enabled = 1; + + DLB_CSR_WR(hw, DLB_SYS_LDB_POOL_ENBLD(pool->id), r0.val); + + pool->avail_credits = args->num_ldb_credits; + pool->total_credits = args->num_ldb_credits; + domain->qed_freelist.offset += args->num_ldb_credits; + + pool->configured = true; +} + +static int +dlb_verify_create_ldb_pool_args(struct dlb_hw *hw, + u32 domain_id, + struct dlb_create_ldb_pool_args *args, + struct dlb_cmd_response *resp) +{ + struct dlb_freelist *qed_freelist; + struct dlb_domain *domain; + + domain = dlb_get_domain_from_id(hw, domain_id); + + if (domain == NULL) { + resp->status = DLB_ST_INVALID_DOMAIN_ID; + return -1; + } + + if (!domain->configured) { + resp->status = DLB_ST_DOMAIN_NOT_CONFIGURED; + return -1; + } + + qed_freelist = &domain->qed_freelist; + + if (dlb_freelist_count(qed_freelist) < args->num_ldb_credits) { + resp->status = DLB_ST_LDB_CREDITS_UNAVAILABLE; + return -1; + } + + if (dlb_list_empty(&domain->avail_ldb_credit_pools)) { + resp->status = DLB_ST_LDB_CREDIT_POOLS_UNAVAILABLE; + return -1; + } + + if (domain->started) { + resp->status = DLB_ST_DOMAIN_STARTED; + return -1; + } + + return 0; +} + +static void +dlb_log_create_ldb_pool_args(struct dlb_hw *hw, + u32 domain_id, + struct dlb_create_ldb_pool_args *args) +{ + DLB_HW_INFO(hw, "DLB create load-balanced credit pool arguments:\n"); + DLB_HW_INFO(hw, "\tDomain ID: %d\n", domain_id); + DLB_HW_INFO(hw, "\tNumber of LDB credits: %d\n", + args->num_ldb_credits); +} + +/** + * dlb_hw_create_ldb_pool() - Allocate and initialize a DLB credit pool. + * @hw: Contains the current state of the DLB hardware. + * @args: User-provided arguments. + * @resp: Response to user. + * + * Return: returns < 0 on error, 0 otherwise. If the driver is unable to + * satisfy a request, resp->status will be set accordingly. + */ +int dlb_hw_create_ldb_pool(struct dlb_hw *hw, + u32 domain_id, + struct dlb_create_ldb_pool_args *args, + struct dlb_cmd_response *resp) +{ + struct dlb_credit_pool *pool; + struct dlb_domain *domain; + + dlb_log_create_ldb_pool_args(hw, domain_id, args); + + /* Verify that hardware resources are available before attempting to + * satisfy the request. This simplifies the error unwinding code. + */ + if (dlb_verify_create_ldb_pool_args(hw, domain_id, args, resp)) + return -EINVAL; + + domain = dlb_get_domain_from_id(hw, domain_id); + if (domain == NULL) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: domain not found\n", + __func__, __LINE__); + return -EFAULT; + } + + pool = DLB_DOM_LIST_HEAD(domain->avail_ldb_credit_pools, typeof(*pool)); + + /* Verification should catch this. */ + if (pool == NULL) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: no available ldb credit pools\n", + __func__, __LINE__); + return -EFAULT; + } + + dlb_configure_ldb_credit_pool(hw, domain, args, pool); + + /* Configuration succeeded, so move the resource from the 'avail' to + * the 'used' list. + */ + dlb_list_del(&domain->avail_ldb_credit_pools, &pool->domain_list); + + dlb_list_add(&domain->used_ldb_credit_pools, &pool->domain_list); + + resp->status = 0; + resp->id = pool->id; + + return 0; +} + +static void +dlb_configure_dir_credit_pool(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_create_dir_pool_args *args, + struct dlb_credit_pool *pool) +{ + union dlb_sys_dir_pool_enbld r0 = { {0} }; + union dlb_chp_dir_pool_crd_lim r1 = { {0} }; + union dlb_chp_dir_pool_crd_cnt r2 = { {0} }; + union dlb_chp_dqed_fl_base r3 = { {0} }; + union dlb_chp_dqed_fl_lim r4 = { {0} }; + union dlb_chp_dqed_fl_push_ptr r5 = { {0} }; + union dlb_chp_dqed_fl_pop_ptr r6 = { {0} }; + + r1.field.limit = args->num_dir_credits; + + DLB_CSR_WR(hw, DLB_CHP_DIR_POOL_CRD_LIM(pool->id), r1.val); + + r2.field.count = args->num_dir_credits; + + DLB_CSR_WR(hw, DLB_CHP_DIR_POOL_CRD_CNT(pool->id), r2.val); + + r3.field.base = domain->dqed_freelist.base + + domain->dqed_freelist.offset; + + DLB_CSR_WR(hw, DLB_CHP_DQED_FL_BASE(pool->id), r3.val); + + r4.field.freelist_disable = 0; + r4.field.limit = r3.field.base + args->num_dir_credits - 1; + + DLB_CSR_WR(hw, DLB_CHP_DQED_FL_LIM(pool->id), r4.val); + + r5.field.push_ptr = r3.field.base; + r5.field.generation = 1; + + DLB_CSR_WR(hw, DLB_CHP_DQED_FL_PUSH_PTR(pool->id), r5.val); + + r6.field.pop_ptr = r3.field.base; + r6.field.generation = 0; + + DLB_CSR_WR(hw, DLB_CHP_DQED_FL_POP_PTR(pool->id), r6.val); + + r0.field.pool_enabled = 1; + + DLB_CSR_WR(hw, DLB_SYS_DIR_POOL_ENBLD(pool->id), r0.val); + + pool->avail_credits = args->num_dir_credits; + pool->total_credits = args->num_dir_credits; + domain->dqed_freelist.offset += args->num_dir_credits; + + pool->configured = true; +} + +static int +dlb_verify_create_dir_pool_args(struct dlb_hw *hw, + u32 domain_id, + struct dlb_create_dir_pool_args *args, + struct dlb_cmd_response *resp) +{ + struct dlb_freelist *dqed_freelist; + struct dlb_domain *domain; + + domain = dlb_get_domain_from_id(hw, domain_id); + + if (domain == NULL) { + resp->status = DLB_ST_INVALID_DOMAIN_ID; + return -1; + } + + if (!domain->configured) { + resp->status = DLB_ST_DOMAIN_NOT_CONFIGURED; + return -1; + } + + dqed_freelist = &domain->dqed_freelist; + + if (dlb_freelist_count(dqed_freelist) < args->num_dir_credits) { + resp->status = DLB_ST_DIR_CREDITS_UNAVAILABLE; + return -1; + } + + if (dlb_list_empty(&domain->avail_dir_credit_pools)) { + resp->status = DLB_ST_DIR_CREDIT_POOLS_UNAVAILABLE; + return -1; + } + + if (domain->started) { + resp->status = DLB_ST_DOMAIN_STARTED; + return -1; + } + + return 0; +} + +static void +dlb_log_create_dir_pool_args(struct dlb_hw *hw, + u32 domain_id, + struct dlb_create_dir_pool_args *args) +{ + DLB_HW_INFO(hw, "DLB create directed credit pool arguments:\n"); + DLB_HW_INFO(hw, "\tDomain ID: %d\n", domain_id); + DLB_HW_INFO(hw, "\tNumber of DIR credits: %d\n", + args->num_dir_credits); +} + +/** + * dlb_hw_create_dir_pool() - Allocate and initialize a DLB credit pool. + * @hw: Contains the current state of the DLB hardware. + * @args: User-provided arguments. + * @resp: Response to user. + * + * Return: returns < 0 on error, 0 otherwise. If the driver is unable to + * satisfy a request, resp->status will be set accordingly. + */ +int dlb_hw_create_dir_pool(struct dlb_hw *hw, + u32 domain_id, + struct dlb_create_dir_pool_args *args, + struct dlb_cmd_response *resp) +{ + struct dlb_credit_pool *pool; + struct dlb_domain *domain; + + dlb_log_create_dir_pool_args(hw, domain_id, args); + + /* Verify that hardware resources are available before attempting to + * satisfy the request. This simplifies the error unwinding code. + */ + /* At least one available pool */ + if (dlb_verify_create_dir_pool_args(hw, domain_id, args, resp)) + return -EINVAL; + + domain = dlb_get_domain_from_id(hw, domain_id); + if (domain == NULL) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: domain not found\n", + __func__, __LINE__); + return -EFAULT; + } + + pool = DLB_DOM_LIST_HEAD(domain->avail_dir_credit_pools, typeof(*pool)); + + /* Verification should catch this. */ + if (pool == NULL) { + DLB_HW_ERR(hw, + "[%s():%d] Internal error: no available dir credit pools\n", + __func__, __LINE__); + return -EFAULT; + } + + dlb_configure_dir_credit_pool(hw, domain, args, pool); + + /* Configuration succeeded, so move the resource from the 'avail' to + * the 'used' list. + */ + dlb_list_del(&domain->avail_dir_credit_pools, &pool->domain_list); + + dlb_list_add(&domain->used_dir_credit_pools, &pool->domain_list); + + resp->status = 0; + resp->id = pool->id; + + return 0; +} + +static u32 dlb_ldb_cq_inflight_count(struct dlb_hw *hw, + struct dlb_ldb_port *port) +{ + union dlb_lsp_cq_ldb_infl_cnt r0; + + r0.val = DLB_CSR_RD(hw, DLB_LSP_CQ_LDB_INFL_CNT(port->id)); + + return r0.field.count; +} + +static u32 dlb_ldb_cq_token_count(struct dlb_hw *hw, + struct dlb_ldb_port *port) +{ + union dlb_lsp_cq_ldb_tkn_cnt r0; + + r0.val = DLB_CSR_RD(hw, DLB_LSP_CQ_LDB_TKN_CNT(port->id)); + + return r0.field.token_count; +} + +static int dlb_drain_ldb_cq(struct dlb_hw *hw, struct dlb_ldb_port *port) +{ + u32 infl_cnt, tkn_cnt; + unsigned int i; + + infl_cnt = dlb_ldb_cq_inflight_count(hw, port); + + /* Account for the initial token count, which is used in order to + * provide a CQ with depth less than 8. + */ + tkn_cnt = dlb_ldb_cq_token_count(hw, port) - port->init_tkn_cnt; + + if (infl_cnt || tkn_cnt) { + struct dlb_hcw hcw_mem[8], *hcw; + void *pp_addr; + + pp_addr = os_map_producer_port(hw, port->id, true); + + /* Point hcw to a 64B-aligned location */ + hcw = (struct dlb_hcw *)((uintptr_t)&hcw_mem[4] & ~0x3F); + + /* Program the first HCW for a completion and token return and + * the other HCWs as NOOPS + */ + + memset(hcw, 0, 4 * sizeof(*hcw)); + hcw->qe_comp = (infl_cnt > 0); + hcw->cq_token = (tkn_cnt > 0); + hcw->lock_id = tkn_cnt - 1; + + /* Return tokens in the first HCW */ + dlb_movdir64b(pp_addr, hcw); + + hcw->cq_token = 0; + + /* Issue remaining completions (if any) */ + for (i = 1; i < infl_cnt; i++) + dlb_movdir64b(pp_addr, hcw); + + os_fence_hcw(hw, pp_addr); + + os_unmap_producer_port(hw, pp_addr); + } + + return 0; +} + +static int dlb_domain_drain_ldb_cqs(struct dlb_hw *hw, + struct dlb_domain *domain, + bool toggle_port) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + int ret; + + /* If the domain hasn't been started, there's no traffic to drain */ + if (!domain->started) + return 0; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) { + if (toggle_port) + dlb_ldb_port_cq_disable(hw, port); + + ret = dlb_drain_ldb_cq(hw, port); + if (ret < 0) + return ret; + + if (toggle_port) + dlb_ldb_port_cq_enable(hw, port); + } + + return 0; +} + +static void dlb_domain_disable_ldb_queue_write_perms(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + int domain_offset = domain->id * DLB_MAX_NUM_LDB_QUEUES; + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_sys_ldb_vasqid_v r0; + struct dlb_ldb_queue *queue; + + r0.field.vasqid_v = 0; + + DLB_DOM_LIST_FOR(domain->used_ldb_queues, queue, iter) { + int idx = domain_offset + queue->id; + + DLB_CSR_WR(hw, DLB_SYS_LDB_VASQID_V(idx), r0.val); + } +} + +static void dlb_domain_disable_ldb_seq_checks(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_chp_sn_chk_enbl r1; + struct dlb_ldb_port *port; + + r1.field.en = 0; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) + DLB_CSR_WR(hw, + DLB_CHP_SN_CHK_ENBL(port->id), + r1.val); +} + +static void dlb_domain_disable_ldb_port_crd_updates(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_chp_ldb_pp_crd_req_state r0; + struct dlb_ldb_port *port; + + r0.field.no_pp_credit_update = 1; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_CRD_REQ_STATE(port->id), + r0.val); +} + +static void dlb_domain_disable_ldb_port_interrupts(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_chp_ldb_cq_int_enb r0 = { {0} }; + union dlb_chp_ldb_cq_wd_enb r1 = { {0} }; + struct dlb_ldb_port *port; + + r0.field.en_tim = 0; + r0.field.en_depth = 0; + + r1.field.wd_enable = 0; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) { + DLB_CSR_WR(hw, + DLB_CHP_LDB_CQ_INT_ENB(port->id), + r0.val); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_CQ_WD_ENB(port->id), + r1.val); + } +} + +static void dlb_domain_disable_dir_queue_write_perms(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + int domain_offset = domain->id * DLB_MAX_NUM_DIR_PORTS; + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_sys_dir_vasqid_v r0; + struct dlb_dir_pq_pair *port; + + r0.field.vasqid_v = 0; + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, port, iter) { + int idx = domain_offset + port->id; + + DLB_CSR_WR(hw, DLB_SYS_DIR_VASQID_V(idx), r0.val); + } +} + +static void dlb_domain_disable_dir_port_interrupts(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_chp_dir_cq_int_enb r0 = { {0} }; + union dlb_chp_dir_cq_wd_enb r1 = { {0} }; + struct dlb_dir_pq_pair *port; + + r0.field.en_tim = 0; + r0.field.en_depth = 0; + + r1.field.wd_enable = 0; + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, port, iter) { + DLB_CSR_WR(hw, + DLB_CHP_DIR_CQ_INT_ENB(port->id), + r0.val); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_CQ_WD_ENB(port->id), + r1.val); + } +} + +static void dlb_domain_disable_dir_port_crd_updates(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_chp_dir_pp_crd_req_state r0; + struct dlb_dir_pq_pair *port; + + r0.field.no_pp_credit_update = 1; + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, port, iter) + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_CRD_REQ_STATE(port->id), + r0.val); +} + +static void dlb_domain_disable_dir_cqs(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_dir_pq_pair *port; + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, port, iter) { + port->enabled = false; + + dlb_dir_port_cq_disable(hw, port); + } +} + +static void dlb_domain_disable_ldb_cqs(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) { + port->enabled = false; + + dlb_ldb_port_cq_disable(hw, port); + } +} + +static void dlb_domain_enable_ldb_cqs(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) { + port->enabled = true; + + dlb_ldb_port_cq_enable(hw, port); + } +} + +static struct dlb_ldb_queue *dlb_get_ldb_queue_from_id(struct dlb_hw *hw, + u32 id) +{ + if (id >= DLB_MAX_NUM_LDB_QUEUES) + return NULL; + + return &hw->rsrcs.ldb_queues[id]; +} + +static void dlb_ldb_port_clear_has_work_bits(struct dlb_hw *hw, + struct dlb_ldb_port *port, + u8 slot) +{ + union dlb_lsp_ldb_sched_ctrl r2 = { {0} }; + + r2.field.cq = port->id; + r2.field.qidix = slot; + r2.field.value = 0; + r2.field.rlist_haswork_v = 1; + + DLB_CSR_WR(hw, DLB_LSP_LDB_SCHED_CTRL, r2.val); + + memset(&r2, 0, sizeof(r2)); + + r2.field.cq = port->id; + r2.field.qidix = slot; + r2.field.value = 0; + r2.field.nalb_haswork_v = 1; + + DLB_CSR_WR(hw, DLB_LSP_LDB_SCHED_CTRL, r2.val); + + dlb_flush_csr(hw); +} + +static void dlb_domain_finish_map_port(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_ldb_port *port) +{ + int i; + + for (i = 0; i < DLB_MAX_NUM_QIDS_PER_LDB_CQ; i++) { + union dlb_lsp_qid_ldb_infl_cnt r0; + struct dlb_ldb_queue *queue; + int qid; + + if (port->qid_map[i].state != DLB_QUEUE_MAP_IN_PROGRESS) + continue; + + qid = port->qid_map[i].qid; + + queue = dlb_get_ldb_queue_from_id(hw, qid); + + if (queue == NULL) { + DLB_HW_ERR(hw, + "[%s()] Internal error: unable to find queue %d\n", + __func__, qid); + continue; + } + + r0.val = DLB_CSR_RD(hw, DLB_LSP_QID_LDB_INFL_CNT(qid)); + + if (r0.field.count) + continue; + + /* Disable the affected CQ, and the CQs already mapped to the + * QID, before reading the QID's inflight count a second time. + * There is an unlikely race in which the QID may schedule one + * more QE after we read an inflight count of 0, and disabling + * the CQs guarantees that the race will not occur after a + * re-read of the inflight count register. + */ + if (port->enabled) + dlb_ldb_port_cq_disable(hw, port); + + dlb_ldb_queue_disable_mapped_cqs(hw, domain, queue); + + r0.val = DLB_CSR_RD(hw, DLB_LSP_QID_LDB_INFL_CNT(qid)); + + if (r0.field.count) { + if (port->enabled) + dlb_ldb_port_cq_enable(hw, port); + + dlb_ldb_queue_enable_mapped_cqs(hw, domain, queue); + + continue; + } + + dlb_ldb_port_finish_map_qid_dynamic(hw, domain, port, queue); + } +} + +static unsigned int +dlb_domain_finish_map_qid_procedures(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + + if (!domain->configured || domain->num_pending_additions == 0) + return 0; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) + dlb_domain_finish_map_port(hw, domain, port); + + return domain->num_pending_additions; +} + +unsigned int dlb_finish_map_qid_procedures(struct dlb_hw *hw) +{ + int i, num = 0; + + /* Finish queue map jobs for any domain that needs it */ + for (i = 0; i < DLB_MAX_NUM_DOMAINS; i++) { + struct dlb_domain *domain = &hw->domains[i]; + + num += dlb_domain_finish_map_qid_procedures(hw, domain); + } + + return num; +} + + +static int dlb_domain_wait_for_ldb_cqs_to_empty(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) { + int i; + + for (i = 0; i < DLB_MAX_CQ_COMP_CHECK_LOOPS; i++) { + if (dlb_ldb_cq_inflight_count(hw, port) == 0) + break; + } + + if (i == DLB_MAX_CQ_COMP_CHECK_LOOPS) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to flush load-balanced port %d's completions.\n", + __func__, port->id); + return -EFAULT; + } + } + + return 0; +} + + +static void dlb_domain_finish_unmap_port_slot(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_ldb_port *port, + int slot) +{ + enum dlb_qid_map_state state; + struct dlb_ldb_queue *queue; + + queue = &hw->rsrcs.ldb_queues[port->qid_map[slot].qid]; + + state = port->qid_map[slot].state; + + /* Update the QID2CQIDX and CQ2QID vectors */ + dlb_ldb_port_unmap_qid(hw, port, queue); + + /* Ensure the QID will not be serviced by this {CQ, slot} by clearing + * the has_work bits + */ + dlb_ldb_port_clear_has_work_bits(hw, port, slot); + + /* Reset the {CQ, slot} to its default state */ + dlb_ldb_port_set_queue_if_status(hw, port, slot); + + /* Re-enable the CQ if it was not manually disabled by the user */ + if (port->enabled) + dlb_ldb_port_cq_enable(hw, port); + + /* If there is a mapping that is pending this slot's removal, perform + * the mapping now. + */ + if (state == DLB_QUEUE_UNMAP_IN_PROGRESS_PENDING_MAP) { + struct dlb_ldb_port_qid_map *map; + struct dlb_ldb_queue *map_queue; + u8 prio; + + map = &port->qid_map[slot]; + + map->qid = map->pending_qid; + map->priority = map->pending_priority; + + map_queue = &hw->rsrcs.ldb_queues[map->qid]; + prio = map->priority; + + dlb_ldb_port_map_qid(hw, domain, port, map_queue, prio); + } +} + +static bool dlb_domain_finish_unmap_port(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_ldb_port *port) +{ + union dlb_lsp_cq_ldb_infl_cnt r0; + int i; + + if (port->num_pending_removals == 0) + return false; + + /* The unmap requires all the CQ's outstanding inflights to be + * completed. + */ + r0.val = DLB_CSR_RD(hw, DLB_LSP_CQ_LDB_INFL_CNT(port->id)); + if (r0.field.count > 0) + return false; + + for (i = 0; i < DLB_MAX_NUM_QIDS_PER_LDB_CQ; i++) { + struct dlb_ldb_port_qid_map *map; + + map = &port->qid_map[i]; + + if (map->state != DLB_QUEUE_UNMAP_IN_PROGRESS && + map->state != DLB_QUEUE_UNMAP_IN_PROGRESS_PENDING_MAP) + continue; + + dlb_domain_finish_unmap_port_slot(hw, domain, port, i); + } + + return true; +} + +static unsigned int +dlb_domain_finish_unmap_qid_procedures(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + + if (!domain->configured || domain->num_pending_removals == 0) + return 0; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) + dlb_domain_finish_unmap_port(hw, domain, port); + + return domain->num_pending_removals; +} + +unsigned int dlb_finish_unmap_qid_procedures(struct dlb_hw *hw) +{ + int i, num = 0; + + /* Finish queue unmap jobs for any domain that needs it */ + for (i = 0; i < DLB_MAX_NUM_DOMAINS; i++) { + struct dlb_domain *domain = &hw->domains[i]; + + num += dlb_domain_finish_unmap_qid_procedures(hw, domain); + } + + return num; +} + +/* Returns whether the queue is empty, including its inflight and replay + * counts. + */ +static bool dlb_ldb_queue_is_empty(struct dlb_hw *hw, + struct dlb_ldb_queue *queue) +{ + union dlb_lsp_qid_ldb_replay_cnt r0; + union dlb_lsp_qid_aqed_active_cnt r1; + union dlb_lsp_qid_atq_enqueue_cnt r2; + union dlb_lsp_qid_ldb_enqueue_cnt r3; + union dlb_lsp_qid_ldb_infl_cnt r4; + + r0.val = DLB_CSR_RD(hw, + DLB_LSP_QID_LDB_REPLAY_CNT(queue->id)); + if (r0.val) + return false; + + r1.val = DLB_CSR_RD(hw, + DLB_LSP_QID_AQED_ACTIVE_CNT(queue->id)); + if (r1.val) + return false; + + r2.val = DLB_CSR_RD(hw, + DLB_LSP_QID_ATQ_ENQUEUE_CNT(queue->id)); + if (r2.val) + return false; + + r3.val = DLB_CSR_RD(hw, + DLB_LSP_QID_LDB_ENQUEUE_CNT(queue->id)); + if (r3.val) + return false; + + r4.val = DLB_CSR_RD(hw, + DLB_LSP_QID_LDB_INFL_CNT(queue->id)); + if (r4.val) + return false; + + return true; +} + +static bool dlb_domain_mapped_queues_empty(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_queue *queue; + + DLB_DOM_LIST_FOR(domain->used_ldb_queues, queue, iter) { + if (queue->num_mappings == 0) + continue; + + if (!dlb_ldb_queue_is_empty(hw, queue)) + return false; + } + + return true; +} + +static int dlb_domain_drain_mapped_queues(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + int i, ret; + + /* If the domain hasn't been started, there's no traffic to drain */ + if (!domain->started) + return 0; + + if (domain->num_pending_removals > 0) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to unmap domain queues\n", + __func__); + return -EFAULT; + } + + for (i = 0; i < DLB_MAX_QID_EMPTY_CHECK_LOOPS; i++) { + ret = dlb_domain_drain_ldb_cqs(hw, domain, true); + if (ret < 0) + return ret; + + if (dlb_domain_mapped_queues_empty(hw, domain)) + break; + } + + if (i == DLB_MAX_QID_EMPTY_CHECK_LOOPS) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to empty queues\n", + __func__); + return -EFAULT; + } + + /* Drain the CQs one more time. For the queues to go empty, they would + * have scheduled one or more QEs. + */ + ret = dlb_domain_drain_ldb_cqs(hw, domain, true); + if (ret < 0) + return ret; + + return 0; +} + +static int dlb_domain_drain_unmapped_queue(struct dlb_hw *hw, + struct dlb_domain *domain, + struct dlb_ldb_queue *queue) +{ + struct dlb_ldb_port *port; + int ret; + + /* If a domain has LDB queues, it must have LDB ports */ + if (dlb_list_empty(&domain->used_ldb_ports)) { + DLB_HW_ERR(hw, + "[%s()] Internal error: No configured LDB ports\n", + __func__); + return -EFAULT; + } + + port = DLB_DOM_LIST_HEAD(domain->used_ldb_ports, typeof(*port)); + + /* If necessary, free up a QID slot in this CQ */ + if (port->num_mappings == DLB_MAX_NUM_QIDS_PER_LDB_CQ) { + struct dlb_ldb_queue *mapped_queue; + + mapped_queue = &hw->rsrcs.ldb_queues[port->qid_map[0].qid]; + + ret = dlb_ldb_port_unmap_qid(hw, port, mapped_queue); + if (ret) + return ret; + } + + ret = dlb_ldb_port_map_qid_dynamic(hw, port, queue, 0); + if (ret) + return ret; + + return dlb_domain_drain_mapped_queues(hw, domain); +} + +static int dlb_domain_drain_unmapped_queues(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_queue *queue; + int ret; + + /* If the domain hasn't been started, there's no traffic to drain */ + if (!domain->started) + return 0; + + DLB_DOM_LIST_FOR(domain->used_ldb_queues, queue, iter) { + if (queue->num_mappings != 0 || + dlb_ldb_queue_is_empty(hw, queue)) + continue; + + ret = dlb_domain_drain_unmapped_queue(hw, domain, queue); + if (ret) + return ret; + } + + return 0; +} + +static int dlb_domain_wait_for_ldb_pool_refill(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_credit_pool *pool; + + /* Confirm that all credits are returned to the domain's credit pools */ + DLB_DOM_LIST_FOR(domain->used_ldb_credit_pools, pool, iter) { + union dlb_chp_qed_fl_push_ptr r0; + union dlb_chp_qed_fl_pop_ptr r1; + unsigned long pop_offs, push_offs; + int i; + + push_offs = DLB_CHP_QED_FL_PUSH_PTR(pool->id); + pop_offs = DLB_CHP_QED_FL_POP_PTR(pool->id); + + for (i = 0; i < DLB_MAX_QID_EMPTY_CHECK_LOOPS; i++) { + r0.val = DLB_CSR_RD(hw, push_offs); + + r1.val = DLB_CSR_RD(hw, pop_offs); + + /* Break early if the freelist is replenished */ + if (r1.field.pop_ptr == r0.field.push_ptr && + r1.field.generation != r0.field.generation) { + break; + } + } + + /* Error if the freelist is not full */ + if (r1.field.pop_ptr != r0.field.push_ptr || + r1.field.generation == r0.field.generation) { + return -EFAULT; + } + } + + return 0; +} + +static int dlb_domain_wait_for_dir_pool_refill(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_credit_pool *pool; + + /* Confirm that all credits are returned to the domain's credit pools */ + DLB_DOM_LIST_FOR(domain->used_dir_credit_pools, pool, iter) { + union dlb_chp_dqed_fl_push_ptr r0; + union dlb_chp_dqed_fl_pop_ptr r1; + unsigned long pop_offs, push_offs; + int i; + + push_offs = DLB_CHP_DQED_FL_PUSH_PTR(pool->id); + pop_offs = DLB_CHP_DQED_FL_POP_PTR(pool->id); + + for (i = 0; i < DLB_MAX_QID_EMPTY_CHECK_LOOPS; i++) { + r0.val = DLB_CSR_RD(hw, push_offs); + + r1.val = DLB_CSR_RD(hw, pop_offs); + + /* Break early if the freelist is replenished */ + if (r1.field.pop_ptr == r0.field.push_ptr && + r1.field.generation != r0.field.generation) { + break; + } + } + + /* Error if the freelist is not full */ + if (r1.field.pop_ptr != r0.field.push_ptr || + r1.field.generation == r0.field.generation) { + return -EFAULT; + } + } + + return 0; +} + +static u32 dlb_dir_queue_depth(struct dlb_hw *hw, + struct dlb_dir_pq_pair *queue) +{ + union dlb_lsp_qid_dir_enqueue_cnt r0; + + r0.val = DLB_CSR_RD(hw, DLB_LSP_QID_DIR_ENQUEUE_CNT(queue->id)); + + return r0.field.count; +} + +static bool dlb_dir_queue_is_empty(struct dlb_hw *hw, + struct dlb_dir_pq_pair *queue) +{ + return dlb_dir_queue_depth(hw, queue) == 0; +} + +static bool dlb_domain_dir_queues_empty(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_dir_pq_pair *queue; + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, queue, iter) { + if (!dlb_dir_queue_is_empty(hw, queue)) + return false; + } + + return true; +} + +static u32 dlb_dir_cq_token_count(struct dlb_hw *hw, + struct dlb_dir_pq_pair *port) +{ + union dlb_lsp_cq_dir_tkn_cnt r0; + + r0.val = DLB_CSR_RD(hw, DLB_LSP_CQ_DIR_TKN_CNT(port->id)); + + return r0.field.count; +} + +static void dlb_drain_dir_cq(struct dlb_hw *hw, struct dlb_dir_pq_pair *port) +{ + unsigned int port_id = port->id; + u32 cnt; + + /* Return any outstanding tokens */ + cnt = dlb_dir_cq_token_count(hw, port); + + if (cnt != 0) { + struct dlb_hcw hcw_mem[8], *hcw; + void *pp_addr; + + pp_addr = os_map_producer_port(hw, port_id, false); + + /* Point hcw to a 64B-aligned location */ + hcw = (struct dlb_hcw *)((uintptr_t)&hcw_mem[4] & ~0x3F); + + /* Program the first HCW for a batch token return and + * the rest as NOOPS + */ + memset(hcw, 0, 4 * sizeof(*hcw)); + hcw->cq_token = 1; + hcw->lock_id = cnt - 1; + + dlb_movdir64b(pp_addr, hcw); + + os_fence_hcw(hw, pp_addr); + + os_unmap_producer_port(hw, pp_addr); + } +} + +static int dlb_domain_drain_dir_cqs(struct dlb_hw *hw, + struct dlb_domain *domain, + bool toggle_port) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_dir_pq_pair *port; + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, port, iter) { + /* Can't drain a port if it's not configured, and there's + * nothing to drain if its queue is unconfigured. + */ + if (!port->port_configured || !port->queue_configured) + continue; + + if (toggle_port) + dlb_dir_port_cq_disable(hw, port); + + dlb_drain_dir_cq(hw, port); + + if (toggle_port) + dlb_dir_port_cq_enable(hw, port); + } + + return 0; +} + +static int dlb_domain_drain_dir_queues(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + int i; + + /* If the domain hasn't been started, there's no traffic to drain */ + if (!domain->started) + return 0; + + for (i = 0; i < DLB_MAX_QID_EMPTY_CHECK_LOOPS; i++) { + dlb_domain_drain_dir_cqs(hw, domain, true); + + if (dlb_domain_dir_queues_empty(hw, domain)) + break; + } + + if (i == DLB_MAX_QID_EMPTY_CHECK_LOOPS) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to empty queues\n", + __func__); + return -EFAULT; + } + + /* Drain the CQs one more time. For the queues to go empty, they would + * have scheduled one or more QEs. + */ + dlb_domain_drain_dir_cqs(hw, domain, true); + + return 0; +} + +static void dlb_domain_disable_dir_producer_ports(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_dir_pq_pair *port; + union dlb_sys_dir_pp_v r1; + + r1.field.pp_v = 0; + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, port, iter) + DLB_CSR_WR(hw, + DLB_SYS_DIR_PP_V(port->id), + r1.val); +} + +static void dlb_domain_disable_ldb_producer_ports(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_sys_ldb_pp_v r1; + struct dlb_ldb_port *port; + + r1.field.pp_v = 0; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) { + DLB_CSR_WR(hw, + DLB_SYS_LDB_PP_V(port->id), + r1.val); + + hw->pf.num_enabled_ldb_ports--; + } +} + +static void dlb_domain_disable_dir_pools(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_sys_dir_pool_enbld r0 = { {0} }; + struct dlb_credit_pool *pool; + + DLB_DOM_LIST_FOR(domain->used_dir_credit_pools, pool, iter) + DLB_CSR_WR(hw, + DLB_SYS_DIR_POOL_ENBLD(pool->id), + r0.val); +} + +static void dlb_domain_disable_ldb_pools(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + union dlb_sys_ldb_pool_enbld r0 = { {0} }; + struct dlb_credit_pool *pool; + + DLB_DOM_LIST_FOR(domain->used_ldb_credit_pools, pool, iter) + DLB_CSR_WR(hw, + DLB_SYS_LDB_POOL_ENBLD(pool->id), + r0.val); +} + +static int dlb_reset_hw_resource(struct dlb_hw *hw, int type, int id) +{ + union dlb_cfg_mstr_diag_reset_sts r0 = { {0} }; + union dlb_cfg_mstr_bcast_reset_vf_start r1 = { {0} }; + int i; + + r1.field.vf_reset_start = 1; + + r1.field.vf_reset_type = type; + r1.field.vf_reset_id = id; + + DLB_CSR_WR(hw, DLB_CFG_MSTR_BCAST_RESET_VF_START, r1.val); + + /* Wait for hardware to complete. This is a finite time operation, + * but wait set a loop bound just in case. + */ + for (i = 0; i < 1024 * 1024; i++) { + r0.val = DLB_CSR_RD(hw, DLB_CFG_MSTR_DIAG_RESET_STS); + + if (r0.field.chp_vf_reset_done && + r0.field.rop_vf_reset_done && + r0.field.lsp_vf_reset_done && + r0.field.nalb_vf_reset_done && + r0.field.ap_vf_reset_done && + r0.field.dp_vf_reset_done && + r0.field.qed_vf_reset_done && + r0.field.dqed_vf_reset_done && + r0.field.aqed_vf_reset_done) + return 0; + + os_udelay(1); + } + + return -ETIMEDOUT; +} + +static int dlb_domain_reset_hw_resources(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_dir_pq_pair *dir_port; + struct dlb_ldb_queue *ldb_queue; + struct dlb_ldb_port *ldb_port; + struct dlb_credit_pool *pool; + int ret; + + DLB_DOM_LIST_FOR(domain->used_ldb_credit_pools, pool, iter) { + ret = dlb_reset_hw_resource(hw, + VF_RST_TYPE_POOL_LDB, + pool->id); + if (ret) + return ret; + } + + DLB_DOM_LIST_FOR(domain->used_dir_credit_pools, pool, iter) { + ret = dlb_reset_hw_resource(hw, + VF_RST_TYPE_POOL_DIR, + pool->id); + if (ret) + return ret; + } + + DLB_DOM_LIST_FOR(domain->used_ldb_queues, ldb_queue, iter) { + ret = dlb_reset_hw_resource(hw, + VF_RST_TYPE_QID_LDB, + ldb_queue->id); + if (ret) + return ret; + } + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, dir_port, iter) { + ret = dlb_reset_hw_resource(hw, + VF_RST_TYPE_QID_DIR, + dir_port->id); + if (ret) + return ret; + } + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, ldb_port, iter) { + ret = dlb_reset_hw_resource(hw, + VF_RST_TYPE_CQ_LDB, + ldb_port->id); + if (ret) + return ret; + } + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, dir_port, iter) { + ret = dlb_reset_hw_resource(hw, + VF_RST_TYPE_CQ_DIR, + dir_port->id); + if (ret) + return ret; + } + + return 0; +} + +static int dlb_domain_verify_reset_success(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_dir_pq_pair *dir_port; + struct dlb_ldb_port *ldb_port; + struct dlb_credit_pool *pool; + struct dlb_ldb_queue *queue; + + /* Confirm that all credits are returned to the domain's credit pools */ + DLB_DOM_LIST_FOR(domain->used_dir_credit_pools, pool, iter) { + union dlb_chp_dqed_fl_pop_ptr r0; + union dlb_chp_dqed_fl_push_ptr r1; + + r0.val = DLB_CSR_RD(hw, + DLB_CHP_DQED_FL_POP_PTR(pool->id)); + + r1.val = DLB_CSR_RD(hw, + DLB_CHP_DQED_FL_PUSH_PTR(pool->id)); + + if (r0.field.pop_ptr != r1.field.push_ptr || + r0.field.generation == r1.field.generation) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to refill directed pool %d's credits.\n", + __func__, pool->id); + return -EFAULT; + } + } + + /* Confirm that all the domain's queue's inflight counts and AQED + * active counts are 0. + */ + DLB_DOM_LIST_FOR(domain->used_ldb_queues, queue, iter) { + if (!dlb_ldb_queue_is_empty(hw, queue)) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to empty ldb queue %d\n", + __func__, queue->id); + return -EFAULT; + } + } + + /* Confirm that all the domain's CQs inflight and token counts are 0. */ + DLB_DOM_LIST_FOR(domain->used_ldb_ports, ldb_port, iter) { + if (dlb_ldb_cq_inflight_count(hw, ldb_port) || + dlb_ldb_cq_token_count(hw, ldb_port)) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to empty ldb port %d\n", + __func__, ldb_port->id); + return -EFAULT; + } + } + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, dir_port, iter) { + if (!dlb_dir_queue_is_empty(hw, dir_port)) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to empty dir queue %d\n", + __func__, dir_port->id); + return -EFAULT; + } + + if (dlb_dir_cq_token_count(hw, dir_port)) { + DLB_HW_ERR(hw, + "[%s()] Internal error: failed to empty dir port %d\n", + __func__, dir_port->id); + return -EFAULT; + } + } + + return 0; +} + +static void __dlb_domain_reset_ldb_port_registers(struct dlb_hw *hw, + struct dlb_ldb_port *port) +{ + union dlb_chp_ldb_pp_state_reset r0 = { {0} }; + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_CRD_REQ_STATE(port->id), + DLB_CHP_LDB_PP_CRD_REQ_STATE_RST); + + /* Reset the port's load-balanced and directed credit state */ + r0.field.dir_type = 0; + r0.field.reset_pp_state = 1; + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_STATE_RESET(port->id), + r0.val); + + r0.field.dir_type = 1; + r0.field.reset_pp_state = 1; + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_STATE_RESET(port->id), + r0.val); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_DIR_PUSH_PTR(port->id), + DLB_CHP_LDB_PP_DIR_PUSH_PTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_LDB_PUSH_PTR(port->id), + DLB_CHP_LDB_PP_LDB_PUSH_PTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_LDB_MIN_CRD_QNT(port->id), + DLB_CHP_LDB_PP_LDB_MIN_CRD_QNT_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_LDB_CRD_LWM(port->id), + DLB_CHP_LDB_PP_LDB_CRD_LWM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_LDB_CRD_HWM(port->id), + DLB_CHP_LDB_PP_LDB_CRD_HWM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_LDB_PP2POOL(port->id), + DLB_CHP_LDB_LDB_PP2POOL_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_DIR_MIN_CRD_QNT(port->id), + DLB_CHP_LDB_PP_DIR_MIN_CRD_QNT_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_DIR_CRD_LWM(port->id), + DLB_CHP_LDB_PP_DIR_CRD_LWM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_PP_DIR_CRD_HWM(port->id), + DLB_CHP_LDB_PP_DIR_CRD_HWM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_DIR_PP2POOL(port->id), + DLB_CHP_LDB_DIR_PP2POOL_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_PP2LDBPOOL(port->id), + DLB_SYS_LDB_PP2LDBPOOL_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_PP2DIRPOOL(port->id), + DLB_SYS_LDB_PP2DIRPOOL_RST); + + DLB_CSR_WR(hw, + DLB_CHP_HIST_LIST_LIM(port->id), + DLB_CHP_HIST_LIST_LIM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_HIST_LIST_BASE(port->id), + DLB_CHP_HIST_LIST_BASE_RST); + + DLB_CSR_WR(hw, + DLB_CHP_HIST_LIST_POP_PTR(port->id), + DLB_CHP_HIST_LIST_POP_PTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_HIST_LIST_PUSH_PTR(port->id), + DLB_CHP_HIST_LIST_PUSH_PTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_CQ_WPTR(port->id), + DLB_CHP_LDB_CQ_WPTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_CQ_INT_DEPTH_THRSH(port->id), + DLB_CHP_LDB_CQ_INT_DEPTH_THRSH_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_CQ_TMR_THRESHOLD(port->id), + DLB_CHP_LDB_CQ_TMR_THRESHOLD_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_CQ_INT_ENB(port->id), + DLB_CHP_LDB_CQ_INT_ENB_RST); + + DLB_CSR_WR(hw, + DLB_LSP_CQ_LDB_INFL_LIM(port->id), + DLB_LSP_CQ_LDB_INFL_LIM_RST); + + DLB_CSR_WR(hw, + DLB_LSP_CQ2PRIOV(port->id), + DLB_LSP_CQ2PRIOV_RST); + + DLB_CSR_WR(hw, + DLB_LSP_CQ_LDB_TOT_SCH_CNT_CTRL(port->id), + DLB_LSP_CQ_LDB_TOT_SCH_CNT_CTRL_RST); + + DLB_CSR_WR(hw, + DLB_LSP_CQ_LDB_TKN_DEPTH_SEL(port->id), + DLB_LSP_CQ_LDB_TKN_DEPTH_SEL_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_CQ_TKN_DEPTH_SEL(port->id), + DLB_CHP_LDB_CQ_TKN_DEPTH_SEL_RST); + + DLB_CSR_WR(hw, + DLB_LSP_CQ_LDB_DSBL(port->id), + DLB_LSP_CQ_LDB_DSBL_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_CQ2VF_PF(port->id), + DLB_SYS_LDB_CQ2VF_PF_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_PP2VF_PF(port->id), + DLB_SYS_LDB_PP2VF_PF_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_CQ_ADDR_L(port->id), + DLB_SYS_LDB_CQ_ADDR_L_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_CQ_ADDR_U(port->id), + DLB_SYS_LDB_CQ_ADDR_U_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_PP_ADDR_L(port->id), + DLB_SYS_LDB_PP_ADDR_L_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_PP_ADDR_U(port->id), + DLB_SYS_LDB_PP_ADDR_U_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_PP_V(port->id), + DLB_SYS_LDB_PP_V_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_PP2VAS(port->id), + DLB_SYS_LDB_PP2VAS_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_CQ_ISR(port->id), + DLB_SYS_LDB_CQ_ISR_RST); + + DLB_CSR_WR(hw, + DLB_SYS_WBUF_LDB_FLAGS(port->id), + DLB_SYS_WBUF_LDB_FLAGS_RST); +} + +static void __dlb_domain_reset_dir_port_registers(struct dlb_hw *hw, + struct dlb_dir_pq_pair *port) +{ + union dlb_chp_dir_pp_state_reset r0 = { {0} }; + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_CRD_REQ_STATE(port->id), + DLB_CHP_DIR_PP_CRD_REQ_STATE_RST); + + /* Reset the port's load-balanced and directed credit state */ + r0.field.dir_type = 0; + r0.field.reset_pp_state = 1; + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_STATE_RESET(port->id), + r0.val); + + r0.field.dir_type = 1; + r0.field.reset_pp_state = 1; + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_STATE_RESET(port->id), + r0.val); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_DIR_PUSH_PTR(port->id), + DLB_CHP_DIR_PP_DIR_PUSH_PTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_LDB_PUSH_PTR(port->id), + DLB_CHP_DIR_PP_LDB_PUSH_PTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_LDB_MIN_CRD_QNT(port->id), + DLB_CHP_DIR_PP_LDB_MIN_CRD_QNT_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_LDB_CRD_LWM(port->id), + DLB_CHP_DIR_PP_LDB_CRD_LWM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_LDB_CRD_HWM(port->id), + DLB_CHP_DIR_PP_LDB_CRD_HWM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_LDB_PP2POOL(port->id), + DLB_CHP_DIR_LDB_PP2POOL_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_DIR_MIN_CRD_QNT(port->id), + DLB_CHP_DIR_PP_DIR_MIN_CRD_QNT_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_DIR_CRD_LWM(port->id), + DLB_CHP_DIR_PP_DIR_CRD_LWM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_PP_DIR_CRD_HWM(port->id), + DLB_CHP_DIR_PP_DIR_CRD_HWM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_DIR_PP2POOL(port->id), + DLB_CHP_DIR_DIR_PP2POOL_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_PP2LDBPOOL(port->id), + DLB_SYS_DIR_PP2LDBPOOL_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_PP2DIRPOOL(port->id), + DLB_SYS_DIR_PP2DIRPOOL_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_CQ_WPTR(port->id), + DLB_CHP_DIR_CQ_WPTR_RST); + + DLB_CSR_WR(hw, + DLB_LSP_CQ_DIR_TKN_DEPTH_SEL_DSI(port->id), + DLB_LSP_CQ_DIR_TKN_DEPTH_SEL_DSI_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_CQ_TKN_DEPTH_SEL(port->id), + DLB_CHP_DIR_CQ_TKN_DEPTH_SEL_RST); + + DLB_CSR_WR(hw, + DLB_LSP_CQ_DIR_DSBL(port->id), + DLB_LSP_CQ_DIR_DSBL_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_CQ_WPTR(port->id), + DLB_CHP_DIR_CQ_WPTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_CQ_INT_DEPTH_THRSH(port->id), + DLB_CHP_DIR_CQ_INT_DEPTH_THRSH_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_CQ_TMR_THRESHOLD(port->id), + DLB_CHP_DIR_CQ_TMR_THRESHOLD_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_CQ_INT_ENB(port->id), + DLB_CHP_DIR_CQ_INT_ENB_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_CQ2VF_PF(port->id), + DLB_SYS_DIR_CQ2VF_PF_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_PP2VF_PF(port->id), + DLB_SYS_DIR_PP2VF_PF_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_CQ_ADDR_L(port->id), + DLB_SYS_DIR_CQ_ADDR_L_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_CQ_ADDR_U(port->id), + DLB_SYS_DIR_CQ_ADDR_U_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_PP_ADDR_L(port->id), + DLB_SYS_DIR_PP_ADDR_L_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_PP_ADDR_U(port->id), + DLB_SYS_DIR_PP_ADDR_U_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_PP_V(port->id), + DLB_SYS_DIR_PP_V_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_PP2VAS(port->id), + DLB_SYS_DIR_PP2VAS_RST); + + DLB_CSR_WR(hw, + DLB_SYS_DIR_CQ_ISR(port->id), + DLB_SYS_DIR_CQ_ISR_RST); + + DLB_CSR_WR(hw, + DLB_SYS_WBUF_DIR_FLAGS(port->id), + DLB_SYS_WBUF_DIR_FLAGS_RST); +} + +static void dlb_domain_reset_dir_port_registers(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_dir_pq_pair *port; + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, port, iter) + __dlb_domain_reset_dir_port_registers(hw, port); +} + +static void dlb_domain_reset_ldb_queue_registers(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_queue *queue; + + DLB_DOM_LIST_FOR(domain->used_ldb_queues, queue, iter) { + DLB_CSR_WR(hw, + DLB_AQED_PIPE_FL_LIM(queue->id), + DLB_AQED_PIPE_FL_LIM_RST); + + DLB_CSR_WR(hw, + DLB_AQED_PIPE_FL_BASE(queue->id), + DLB_AQED_PIPE_FL_BASE_RST); + + DLB_CSR_WR(hw, + DLB_AQED_PIPE_FL_POP_PTR(queue->id), + DLB_AQED_PIPE_FL_POP_PTR_RST); + + DLB_CSR_WR(hw, + DLB_AQED_PIPE_FL_PUSH_PTR(queue->id), + DLB_AQED_PIPE_FL_PUSH_PTR_RST); + + DLB_CSR_WR(hw, + DLB_AQED_PIPE_QID_FID_LIM(queue->id), + DLB_AQED_PIPE_QID_FID_LIM_RST); + + DLB_CSR_WR(hw, + DLB_LSP_QID_AQED_ACTIVE_LIM(queue->id), + DLB_LSP_QID_AQED_ACTIVE_LIM_RST); + + DLB_CSR_WR(hw, + DLB_LSP_QID_LDB_INFL_LIM(queue->id), + DLB_LSP_QID_LDB_INFL_LIM_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_QID_V(queue->id), + DLB_SYS_LDB_QID_V_RST); + + DLB_CSR_WR(hw, + DLB_SYS_LDB_QID_V(queue->id), + DLB_SYS_LDB_QID_V_RST); + + DLB_CSR_WR(hw, + DLB_CHP_ORD_QID_SN(queue->id), + DLB_CHP_ORD_QID_SN_RST); + + DLB_CSR_WR(hw, + DLB_CHP_ORD_QID_SN_MAP(queue->id), + DLB_CHP_ORD_QID_SN_MAP_RST); + + DLB_CSR_WR(hw, + DLB_RO_PIPE_QID2GRPSLT(queue->id), + DLB_RO_PIPE_QID2GRPSLT_RST); + } +} + +static void dlb_domain_reset_dir_queue_registers(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_dir_pq_pair *queue; + + DLB_DOM_LIST_FOR(domain->used_dir_pq_pairs, queue, iter) { + DLB_CSR_WR(hw, + DLB_SYS_DIR_QID_V(queue->id), + DLB_SYS_DIR_QID_V_RST); + } +} + +static void dlb_domain_reset_ldb_pool_registers(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_credit_pool *pool; + + DLB_DOM_LIST_FOR(domain->used_ldb_credit_pools, pool, iter) { + DLB_CSR_WR(hw, + DLB_CHP_LDB_POOL_CRD_LIM(pool->id), + DLB_CHP_LDB_POOL_CRD_LIM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_LDB_POOL_CRD_CNT(pool->id), + DLB_CHP_LDB_POOL_CRD_CNT_RST); + + DLB_CSR_WR(hw, + DLB_CHP_QED_FL_BASE(pool->id), + DLB_CHP_QED_FL_BASE_RST); + + DLB_CSR_WR(hw, + DLB_CHP_QED_FL_LIM(pool->id), + DLB_CHP_QED_FL_LIM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_QED_FL_PUSH_PTR(pool->id), + DLB_CHP_QED_FL_PUSH_PTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_QED_FL_POP_PTR(pool->id), + DLB_CHP_QED_FL_POP_PTR_RST); + } +} + +static void dlb_domain_reset_dir_pool_registers(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_credit_pool *pool; + + DLB_DOM_LIST_FOR(domain->used_dir_credit_pools, pool, iter) { + DLB_CSR_WR(hw, + DLB_CHP_DIR_POOL_CRD_LIM(pool->id), + DLB_CHP_DIR_POOL_CRD_LIM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DIR_POOL_CRD_CNT(pool->id), + DLB_CHP_DIR_POOL_CRD_CNT_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DQED_FL_BASE(pool->id), + DLB_CHP_DQED_FL_BASE_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DQED_FL_LIM(pool->id), + DLB_CHP_DQED_FL_LIM_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DQED_FL_PUSH_PTR(pool->id), + DLB_CHP_DQED_FL_PUSH_PTR_RST); + + DLB_CSR_WR(hw, + DLB_CHP_DQED_FL_POP_PTR(pool->id), + DLB_CHP_DQED_FL_POP_PTR_RST); + } +} + +static void dlb_domain_reset_ldb_port_registers(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_list_entry *iter; + RTE_SET_USED(iter); + struct dlb_ldb_port *port; + + DLB_DOM_LIST_FOR(domain->used_ldb_ports, port, iter) + __dlb_domain_reset_ldb_port_registers(hw, port); +} + +static void dlb_domain_reset_registers(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + dlb_domain_reset_ldb_port_registers(hw, domain); + + dlb_domain_reset_dir_port_registers(hw, domain); + + dlb_domain_reset_ldb_queue_registers(hw, domain); + + dlb_domain_reset_dir_queue_registers(hw, domain); + + dlb_domain_reset_ldb_pool_registers(hw, domain); + + dlb_domain_reset_dir_pool_registers(hw, domain); +} + +static int dlb_domain_reset_software_state(struct dlb_hw *hw, + struct dlb_domain *domain) +{ + struct dlb_ldb_queue *tmp_ldb_queue; + RTE_SET_USED(tmp_ldb_queue); + struct dlb_dir_pq_pair *tmp_dir_port; + RTE_SET_USED(tmp_dir_port); + struct dlb_ldb_port *tmp_ldb_port; + RTE_SET_USED(tmp_ldb_port); + struct dlb_credit_pool *tmp_pool; + RTE_SET_USED(tmp_pool); + struct dlb_list_entry *iter1; + RTE_SET_USED(iter1); + struct dlb_list_entry *iter2; + RTE_SET_USED(iter2); + struct dlb_ldb_queue *ldb_queue; + struct dlb_dir_pq_pair *dir_port; + struct dlb_ldb_port *ldb_port; + struct dlb_credit_pool *pool; + + struct dlb_function_resources *rsrcs; + struct dlb_list_head *list; + int ret; + + rsrcs = domain->parent_func; + + /* Move the domain's ldb queues to the function's avail list */ + list = &domain->used_ldb_queues; + DLB_DOM_LIST_FOR_SAFE(*list, ldb_queue, tmp_ldb_queue, iter1, iter2) { + if (ldb_queue->sn_cfg_valid) { + struct dlb_sn_group *grp; + + grp = &hw->rsrcs.sn_groups[ldb_queue->sn_group]; + + dlb_sn_group_free_slot(grp, ldb_queue->sn_slot); + ldb_queue->sn_cfg_valid = false; + } + + ldb_queue->owned = false; + ldb_queue->num_mappings = 0; + ldb_queue->num_pending_additions = 0; + + dlb_list_del(&domain->used_ldb_queues, &ldb_queue->domain_list); + dlb_list_add(&rsrcs->avail_ldb_queues, &ldb_queue->func_list); + rsrcs->num_avail_ldb_queues++; + } + + list = &domain->avail_ldb_queues; + DLB_DOM_LIST_FOR_SAFE(*list, ldb_queue, tmp_ldb_queue, iter1, iter2) { + ldb_queue->owned = false; + + dlb_list_del(&domain->avail_ldb_queues, + &ldb_queue->domain_list); + dlb_list_add(&rsrcs->avail_ldb_queues, + &ldb_queue->func_list); + rsrcs->num_avail_ldb_queues++; + } + + /* Move the domain's ldb ports to the function's avail list */ + list = &domain->used_ldb_ports; + DLB_DOM_LIST_FOR_SAFE(*list, ldb_port, tmp_ldb_port, iter1, iter2) { + int i; + + ldb_port->owned = false; + ldb_port->configured = false; + ldb_port->num_pending_removals = 0; + ldb_port->num_mappings = 0; + for (i = 0; i < DLB_MAX_NUM_QIDS_PER_LDB_CQ; i++) + ldb_port->qid_map[i].state = DLB_QUEUE_UNMAPPED; + + dlb_list_del(&domain->used_ldb_ports, &ldb_port->domain_list); + dlb_list_add(&rsrcs->avail_ldb_ports, &ldb_port->func_list); + rsrcs->num_avail_ldb_ports++; + } + + list = &domain->avail_ldb_ports; + DLB_DOM_LIST_FOR_SAFE(*list, ldb_port, tmp_ldb_port, iter1, iter2) { + ldb_port->owned = false; + + dlb_list_del(&domain->avail_ldb_ports, &ldb_port->domain_list); + dlb_list_add(&rsrcs->avail_ldb_ports, &ldb_port->func_list); + rsrcs->num_avail_ldb_ports++; + } + + /* Move the domain's dir ports to the function's avail list */ + list = &domain->used_dir_pq_pairs; + DLB_DOM_LIST_FOR_SAFE(*list, dir_port, tmp_dir_port, iter1, iter2) { + dir_port->owned = false; + dir_port->port_configured = false; + + dlb_list_del(&domain->used_dir_pq_pairs, + &dir_port->domain_list); + + dlb_list_add(&rsrcs->avail_dir_pq_pairs, + &dir_port->func_list); + rsrcs->num_avail_dir_pq_pairs++; + } + + list = &domain->avail_dir_pq_pairs; + DLB_DOM_LIST_FOR_SAFE(*list, dir_port, tmp_dir_port, iter1, iter2) { + dir_port->owned = false; + + dlb_list_del(&domain->avail_dir_pq_pairs, + &dir_port->domain_list); + + dlb_list_add(&rsrcs->avail_dir_pq_pairs, + &dir_port->func_list); + rsrcs->num_avail_dir_pq_pairs++; + } + + /* Return hist list entries to the function */ + ret = dlb_bitmap_set_range(rsrcs->avail_hist_list_entries, + domain->hist_list_entry_base, + domain->total_hist_list_entries); + if (ret) { + DLB_HW_ERR(hw, + "[%s()] Internal error: domain hist list base does not match the function's bitmap.\n", + __func__); + return -EFAULT; + } + + domain->total_hist_list_entries = 0; + domain->avail_hist_list_entries = 0; + domain->hist_list_entry_base = 0; + domain->hist_list_entry_offset = 0; + + /* Return QED entries to the function */ + ret = dlb_bitmap_set_range(rsrcs->avail_qed_freelist_entries, + domain->qed_freelist.base, + (domain->qed_freelist.bound - + domain->qed_freelist.base)); + if (ret) { + DLB_HW_ERR(hw, + "[%s()] Internal error: domain QED base does not match the function's bitmap.\n", + __func__); + return -EFAULT; + } + + domain->qed_freelist.base = 0; + domain->qed_freelist.bound = 0; + domain->qed_freelist.offset = 0; + + /* Return DQED entries back to the function */ + ret = dlb_bitmap_set_range(rsrcs->avail_dqed_freelist_entries, + domain->dqed_freelist.base, + (domain->dqed_freelist.bound - + domain->dqed_freelist.base)); + if (ret) { + DLB_HW_ERR(hw, + "[%s()] Internal error: domain DQED base does not match the function's bitmap.\n", + __func__); + return -EFAULT; + } + + domain->dqed_freelist.base = 0; + domain->dqed_freelist.bound = 0; + domain->dqed_freelist.offset = 0; + + /* Return AQED entries back to the function */ + ret = dlb_bitmap_set_range(rsrcs->avail_aqed_freelist_entries, + domain->aqed_freelist.base, + (domain->aqed_freelist.bound - + domain->aqed_freelist.base)); + if (ret) { + DLB_HW_ERR(hw, + "[%s()] Internal error: domain AQED base does not match the function's bitmap.\n", + __func__); + return -EFAULT; + } + + domain->aqed_freelist.base = 0; + domain->aqed_freelist.bound = 0; + domain->aqed_freelist.offset = 0; + + /* Return ldb credit pools back to the function's avail list */ + list = &domain->used_ldb_credit_pools; + DLB_DOM_LIST_FOR_SAFE(*list, pool, tmp_pool, iter1, iter2) { + pool->owned = false; + pool->configured = false; + + dlb_list_del(&domain->used_ldb_credit_pools, + &pool->domain_list); + dlb_list_add(&rsrcs->avail_ldb_credit_pools, + &pool->func_list); + rsrcs->num_avail_ldb_credit_pools++; + } + + list = &domain->avail_ldb_credit_pools; + DLB_DOM_LIST_FOR_SAFE(*list, pool, tmp_pool, iter1, iter2) { + pool->owned = false; + + dlb_list_del(&domain->avail_ldb_credit_pools, + &pool->domain_list); + dlb_list_add(&rsrcs->avail_ldb_credit_pools, + &pool->func_list); + rsrcs->num_avail_ldb_credit_pools++; + } + + /* Move dir credit pools back to the function */ + list = &domain->used_dir_credit_pools; + DLB_DOM_LIST_FOR_SAFE(*list, pool, tmp_pool, iter1, iter2) { + pool->owned = false; + pool->configured = false; + + dlb_list_del(&domain->used_dir_credit_pools, + &pool->domain_list); + dlb_list_add(&rsrcs->avail_dir_credit_pools, + &pool->func_list); + rsrcs->num_avail_dir_credit_pools++; + } + + list = &domain->avail_dir_credit_pools; + DLB_DOM_LIST_FOR_SAFE(*list, pool, tmp_pool, iter1, iter2) { + pool->owned = false; + + dlb_list_del(&domain->avail_dir_credit_pools, + &pool->domain_list); + dlb_list_add(&rsrcs->avail_dir_credit_pools, + &pool->func_list); + rsrcs->num_avail_dir_credit_pools++; + } + + domain->num_pending_removals = 0; + domain->num_pending_additions = 0; + domain->configured = false; + domain->started = false; + + /* Move the domain out of the used_domains list and back to the + * function's avail_domains list. + */ + dlb_list_del(&rsrcs->used_domains, &domain->func_list); + dlb_list_add(&rsrcs->avail_domains, &domain->func_list); + rsrcs->num_avail_domains++; + + return 0; +} + +static void dlb_log_reset_domain(struct dlb_hw *hw, u32 domain_id) +{ + DLB_HW_INFO(hw, "DLB reset domain:\n"); + DLB_HW_INFO(hw, "\tDomain ID: %d\n", domain_id); +} + +/** + * dlb_reset_domain() - Reset a DLB scheduling domain and its associated + * hardware resources. + * @hw: Contains the current state of the DLB hardware. + * @args: User-provided arguments. + * @resp: Response to user. + * + * Note: User software *must* stop sending to this domain's producer ports + * before invoking this function, otherwise undefined behavior will result. + * + * Return: returns < 0 on error, 0 otherwise. + */ +int dlb_reset_domain(struct dlb_hw *hw, u32 domain_id) +{ + struct dlb_domain *domain; + int ret; + + dlb_log_reset_domain(hw, domain_id); + + domain = dlb_get_domain_from_id(hw, domain_id); + + if (domain == NULL || !domain->configured) + return -EINVAL; + + /* For each queue owned by this domain, disable its write permissions to + * cause any traffic sent to it to be dropped. Well-behaved software + * should not be sending QEs at this point. + */ + dlb_domain_disable_dir_queue_write_perms(hw, domain); + + dlb_domain_disable_ldb_queue_write_perms(hw, domain); + + /* Disable credit updates and turn off completion tracking on all the + * domain's PPs. + */ + dlb_domain_disable_dir_port_crd_updates(hw, domain); + + dlb_domain_disable_ldb_port_crd_updates(hw, domain); + + dlb_domain_disable_dir_port_interrupts(hw, domain); + + dlb_domain_disable_ldb_port_interrupts(hw, domain); + + dlb_domain_disable_ldb_seq_checks(hw, domain); + + /* Disable the LDB CQs and drain them in order to complete the map and + * unmap procedures, which require zero CQ inflights and zero QID + * inflights respectively. + */ + dlb_domain_disable_ldb_cqs(hw, domain); + + ret = dlb_domain_drain_ldb_cqs(hw, domain, false); + if (ret < 0) + return ret; + + ret = dlb_domain_wait_for_ldb_cqs_to_empty(hw, domain); + if (ret < 0) + return ret; + + ret = dlb_domain_finish_unmap_qid_procedures(hw, domain); + if (ret < 0) + return ret; + + ret = dlb_domain_finish_map_qid_procedures(hw, domain); + if (ret < 0) + return ret; + + /* Re-enable the CQs in order to drain the mapped queues. */ + dlb_domain_enable_ldb_cqs(hw, domain); + + ret = dlb_domain_drain_mapped_queues(hw, domain); + if (ret < 0) + return ret; + + ret = dlb_domain_drain_unmapped_queues(hw, domain); + if (ret < 0) + return ret; + + ret = dlb_domain_wait_for_ldb_pool_refill(hw, domain); + if (ret) { + DLB_HW_ERR(hw, + "[%s()] Internal error: LDB credits failed to refill\n", + __func__); + return ret; + } + + /* Done draining LDB QEs, so disable the CQs. */ + dlb_domain_disable_ldb_cqs(hw, domain); + + /* Directed queues are reset in dlb_domain_reset_hw_resources(), but + * that process does not decrement the directed queue size counters used + * by SMON for its average DQED depth measurement. So, we manually drain + * the directed queues here. + */ + dlb_domain_drain_dir_queues(hw, domain); + + ret = dlb_domain_wait_for_dir_pool_refill(hw, domain); + if (ret) { + DLB_HW_ERR(hw, + "[%s()] Internal error: DIR credits failed to refill\n", + __func__); + return ret; + } + + /* Done draining DIR QEs, so disable the CQs. */ + dlb_domain_disable_dir_cqs(hw, domain); + + dlb_domain_disable_dir_producer_ports(hw, domain); + + dlb_domain_disable_ldb_producer_ports(hw, domain); + + dlb_domain_disable_dir_pools(hw, domain); + + dlb_domain_disable_ldb_pools(hw, domain); + + /* Reset the QID, credit pool, and CQ hardware. + * + * Note: DLB 1.0 A0 h/w does not disarm CQ interrupts during sched + * domain reset. + * A spurious interrupt can occur on subsequent use of a reset CQ. + */ + ret = dlb_domain_reset_hw_resources(hw, domain); + if (ret) + return ret; + + ret = dlb_domain_verify_reset_success(hw, domain); + if (ret) + return ret; + + dlb_domain_reset_registers(hw, domain); + + /* Hardware reset complete. Reset the domain's software state */ + ret = dlb_domain_reset_software_state(hw, domain); + if (ret) + return ret; + + return 0; +} + +void dlb_hw_get_num_resources(struct dlb_hw *hw, + struct dlb_get_num_resources_args *arg) +{ + struct dlb_function_resources *rsrcs; + struct dlb_bitmap *map; + + rsrcs = &hw->pf; + + arg->num_sched_domains = rsrcs->num_avail_domains; + + arg->num_ldb_queues = rsrcs->num_avail_ldb_queues; + + arg->num_ldb_ports = rsrcs->num_avail_ldb_ports; + + arg->num_dir_ports = rsrcs->num_avail_dir_pq_pairs; + + map = rsrcs->avail_aqed_freelist_entries; + + arg->num_atomic_inflights = dlb_bitmap_count(map); + + arg->max_contiguous_atomic_inflights = + dlb_bitmap_longest_set_range(map); + + map = rsrcs->avail_hist_list_entries; + + arg->num_hist_list_entries = dlb_bitmap_count(map); + + arg->max_contiguous_hist_list_entries = + dlb_bitmap_longest_set_range(map); + + map = rsrcs->avail_qed_freelist_entries; + + arg->num_ldb_credits = dlb_bitmap_count(map); + + arg->max_contiguous_ldb_credits = dlb_bitmap_longest_set_range(map); + + map = rsrcs->avail_dqed_freelist_entries; + + arg->num_dir_credits = dlb_bitmap_count(map); + + arg->max_contiguous_dir_credits = dlb_bitmap_longest_set_range(map); + + arg->num_ldb_credit_pools = rsrcs->num_avail_ldb_credit_pools; + + arg->num_dir_credit_pools = rsrcs->num_avail_dir_credit_pools; +} + void dlb_hw_disable_vf_to_pf_isr_pend_err(struct dlb_hw *hw) { union dlb_sys_sys_alarm_int_enable r0; diff --git a/drivers/event/dlb/pf/dlb_pf.c b/drivers/event/dlb/pf/dlb_pf.c index 7fc85e9..57a150c 100644 --- a/drivers/event/dlb/pf/dlb_pf.c +++ b/drivers/event/dlb/pf/dlb_pf.c @@ -78,6 +78,17 @@ dlb_pf_open(struct dlb_hw_dev *handle, const char *name) return 0; } +static void +dlb_pf_domain_close(struct dlb_eventdev *dlb) +{ + struct dlb_dev *dlb_dev = (struct dlb_dev *)dlb->qm_instance.pf_dev; + int ret; + + ret = dlb_reset_domain(&dlb_dev->hw, dlb->qm_instance.domain_id); + if (ret) + DLB_LOG_ERR("dlb_pf_reset_domain err %d", ret); +} + static int dlb_pf_get_device_version(struct dlb_hw_dev *handle, uint8_t *revision) @@ -101,6 +112,79 @@ dlb_pf_get_num_resources(struct dlb_hw_dev *handle, } static int +dlb_pf_sched_domain_create(struct dlb_hw_dev *handle, + struct dlb_create_sched_domain_args *arg) +{ + struct dlb_dev *dlb_dev = (struct dlb_dev *)handle->pf_dev; + struct dlb_cmd_response response = {0}; + int ret; + + DLB_INFO(dev->dlb_device, "Entering %s()\n", __func__); + + if (dlb_dev->domain_reset_failed) { + response.status = DLB_ST_DOMAIN_RESET_FAILED; + ret = -EINVAL; + goto done; + } + + ret = dlb_hw_create_sched_domain(&dlb_dev->hw, arg, &response); + if (ret) + goto done; + +done: + + *(struct dlb_cmd_response *)arg->response = response; + + DLB_INFO(dev->dlb_device, "Exiting %s() with ret=%d\n", __func__, ret); + + return ret; +} + +static int +dlb_pf_ldb_credit_pool_create(struct dlb_hw_dev *handle, + struct dlb_create_ldb_pool_args *cfg) +{ + struct dlb_dev *dlb_dev = (struct dlb_dev *)handle->pf_dev; + struct dlb_cmd_response response = {0}; + int ret; + + DLB_INFO(dev->dlb_device, "Entering %s()\n", __func__); + + ret = dlb_hw_create_ldb_pool(&dlb_dev->hw, + handle->domain_id, + cfg, + &response); + + *(struct dlb_cmd_response *)cfg->response = response; + + DLB_INFO(dev->dlb_device, "Exiting %s() with ret=%d\n", __func__, ret); + + return ret; +} + +static int +dlb_pf_dir_credit_pool_create(struct dlb_hw_dev *handle, + struct dlb_create_dir_pool_args *cfg) +{ + struct dlb_dev *dlb_dev = (struct dlb_dev *)handle->pf_dev; + struct dlb_cmd_response response = {0}; + int ret; + + DLB_INFO(dev->dlb_device, "Entering %s()\n", __func__); + + ret = dlb_hw_create_dir_pool(&dlb_dev->hw, + handle->domain_id, + cfg, + &response); + + *(struct dlb_cmd_response *)cfg->response = response; + + DLB_INFO(dev->dlb_device, "Exiting %s() with ret=%d\n", __func__, ret); + + return ret; +} + +static int dlb_pf_get_cq_poll_mode(struct dlb_hw_dev *handle, enum dlb_cq_poll_modes *mode) { @@ -119,8 +203,12 @@ dlb_pf_iface_fn_ptrs_init(void) { dlb_iface_low_level_io_init = dlb_pf_low_level_io_init; dlb_iface_open = dlb_pf_open; + dlb_iface_domain_close = dlb_pf_domain_close; dlb_iface_get_device_version = dlb_pf_get_device_version; dlb_iface_get_num_resources = dlb_pf_get_num_resources; + dlb_iface_sched_domain_create = dlb_pf_sched_domain_create; + dlb_iface_ldb_credit_pool_create = dlb_pf_ldb_credit_pool_create; + dlb_iface_dir_credit_pool_create = dlb_pf_dir_credit_pool_create; dlb_iface_get_cq_poll_mode = dlb_pf_get_cq_poll_mode; } -- 2.6.4