DPDK patches and discussions
 help / color / mirror / Atom feed
From: Jerin Jacob <jerinjacobk@gmail.com>
To: pbhagavatula@marvell.com
Cc: jerinj@marvell.com, sthotton@marvell.com,
	timothy.mcdaniel@intel.com,  hemant.agrawal@nxp.com,
	sachin.saxena@nxp.com, mattias.ronnblom@ericsson.com,
	 liangma@liangbit.com, peter.mccarthy@intel.com,
	harry.van.haaren@intel.com,  erik.g.carrillo@intel.com,
	abhinandan.gujjar@intel.com,  s.v.naga.harish.k@intel.com,
	anatoly.burakov@intel.com, dev@dpdk.org
Subject: Re: [PATCH v4 1/3] eventdev: introduce link profiles
Date: Tue, 3 Oct 2023 12:25:43 +0530	[thread overview]
Message-ID: <CALBAE1OCQaki7mGJKWQQvL3UJA5MZtnsXbMp5mu=pxOh_L6zXA@mail.gmail.com> (raw)
In-Reply-To: <20230928101205.4352-2-pbhagavatula@marvell.com>

On Thu, Sep 28, 2023 at 9:41 PM <pbhagavatula@marvell.com> wrote:
>
> From: Pavan Nikhilesh <pbhagavatula@marvell.com>

DMA adapter merge created conflicts, Please rebase to next-event tree.

[for-main]dell[dpdk-next-eventdev] $ git pw series  apply 29675
Failed to apply patch:
Applying: eventdev: introduce link profiles
Using index info to reconstruct a base tree...
M       config/rte_config.h
M       doc/guides/eventdevs/features/default.ini
M       doc/guides/prog_guide/eventdev.rst
M       doc/guides/rel_notes/release_23_11.rst
M       lib/eventdev/eventdev_pmd.h
M       lib/eventdev/eventdev_private.c
M       lib/eventdev/rte_eventdev.c
M       lib/eventdev/rte_eventdev.h
M       lib/eventdev/rte_eventdev_core.h
M       lib/eventdev/version.map
Falling back to patching base and 3-way merge...
Auto-merging lib/eventdev/version.map
Auto-merging lib/eventdev/rte_eventdev_core.h
CONFLICT (content): Merge conflict in lib/eventdev/rte_eventdev_core.h
Auto-merging lib/eventdev/rte_eventdev.h
Auto-merging lib/eventdev/rte_eventdev.c
Auto-merging lib/eventdev/eventdev_private.c
CONFLICT (content): Merge conflict in lib/eventdev/eventdev_private.c
Auto-merging lib/eventdev/eventdev_pmd.h
CONFLICT (content): Merge conflict in lib/eventdev/eventdev_pmd.h
Auto-merging doc/guides/rel_notes/release_23_11.rst
Auto-merging doc/guides/prog_guide/eventdev.rst
Auto-merging doc/guides/eventdevs/features/default.ini
Auto-merging config/rte_config.h
error: Failed to merge in the changes.
hint: Use 'git am --show-current-patch=diff' to see the failed patch
Patch failed at 0001 eventdev: introduce link profiles
When you have resolved this problem, run "git am --continue".
If you prefer to skip this patch, run "git am --skip" instead.
To restore the original branch and stop patching, run "git am --abort".
>
> A collection of event queues linked to an event port can be
> associated with a unique identifier called as a link profile, multiple
> such profiles can be created based on the event device capability
> using the function `rte_event_port_profile_links_set` which takes
> arguments similar to `rte_event_port_link` in addition to the profile
> identifier.
>
> The maximum link profiles that are supported by an event device
> is advertised through the structure member
> `rte_event_dev_info::max_profiles_per_port`.
> By default, event ports are configured to use the link profile 0
> on initialization.
>
> Once multiple link profiles are set up and the event device is started,
> the application can use the function `rte_event_port_profile_switch`
> to change the currently active profile on an event port. This effects
> the next `rte_event_dequeue_burst` call, where the event queues
> associated with the newly active link profile will participate in
> scheduling.
>
> An unlink function `rte_event_port_profile_unlink` is provided
> to modify the links associated to a profile, and
> `rte_event_port_profile_links_get` can be used to retrieve the
> links associated with a profile.
>
> Using Link profiles can reduce the overhead of linking/unlinking and
> waiting for unlinks in progress in fast-path and gives applications
> the ability to switch between preset profiles on the fly.
>
> Signed-off-by: Pavan Nikhilesh <pbhagavatula@marvell.com>
> Acked-by: Jerin Jacob <jerinj@marvell.com>
> ---
>  config/rte_config.h                       |   1 +
>  doc/guides/eventdevs/features/default.ini |   1 +
>  doc/guides/prog_guide/eventdev.rst        |  40 ++++
>  doc/guides/rel_notes/release_23_11.rst    |  10 +
>  lib/eventdev/eventdev_pmd.h               |  59 +++++-
>  lib/eventdev/eventdev_private.c           |   9 +
>  lib/eventdev/eventdev_trace.h             |  32 +++
>  lib/eventdev/eventdev_trace_points.c      |  12 ++
>  lib/eventdev/rte_eventdev.c               | 150 +++++++++++---
>  lib/eventdev/rte_eventdev.h               | 231 ++++++++++++++++++++++
>  lib/eventdev/rte_eventdev_core.h          |   6 +-
>  lib/eventdev/rte_eventdev_trace_fp.h      |   8 +
>  lib/eventdev/version.map                  |   4 +
>  13 files changed, 535 insertions(+), 28 deletions(-)
>
> diff --git a/config/rte_config.h b/config/rte_config.h
> index 400e44e3cf..d43b3eecb8 100644
> --- a/config/rte_config.h
> +++ b/config/rte_config.h
> @@ -73,6 +73,7 @@
>  #define RTE_EVENT_MAX_DEVS 16
>  #define RTE_EVENT_MAX_PORTS_PER_DEV 255
>  #define RTE_EVENT_MAX_QUEUES_PER_DEV 255
> +#define RTE_EVENT_MAX_PROFILES_PER_PORT 8
>  #define RTE_EVENT_TIMER_ADAPTER_NUM_MAX 32
>  #define RTE_EVENT_ETH_INTR_RING_SIZE 1024
>  #define RTE_EVENT_CRYPTO_ADAPTER_MAX_INSTANCE 32
> diff --git a/doc/guides/eventdevs/features/default.ini b/doc/guides/eventdevs/features/default.ini
> index 00360f60c6..1c0082352b 100644
> --- a/doc/guides/eventdevs/features/default.ini
> +++ b/doc/guides/eventdevs/features/default.ini
> @@ -18,6 +18,7 @@ multiple_queue_port        =
>  carry_flow_id              =
>  maintenance_free           =
>  runtime_queue_attr         =
> +profile_links              =
>
>  ;
>  ; Features of a default Ethernet Rx adapter.
> diff --git a/doc/guides/prog_guide/eventdev.rst b/doc/guides/prog_guide/eventdev.rst
> index 2c83176846..4bc0de4cdc 100644
> --- a/doc/guides/prog_guide/eventdev.rst
> +++ b/doc/guides/prog_guide/eventdev.rst
> @@ -317,6 +317,46 @@ can be achieved like this:
>          }
>          int links_made = rte_event_port_link(dev_id, tx_port_id, &single_link_q, &priority, 1);
>
> +Linking Queues to Ports with link profiles
> +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> +
> +An application can use link profiles if supported by the underlying event device to setup up
> +multiple link profile per port and change them run time depending up on heuristic data.
> +Using Link profiles can reduce the overhead of linking/unlinking and wait for unlinks in progress
> +in fast-path and gives applications the ability to switch between preset profiles on the fly.
> +
> +An Example use case could be as follows.
> +
> +Config path:
> +
> +.. code-block:: c
> +
> +        uint8_t lq[4] = {4, 5, 6, 7};
> +        uint8_t hq[4] = {0, 1, 2, 3};
> +
> +        if (rte_event_dev_info.max_profiles_per_port < 2)
> +            return -ENOTSUP;
> +
> +        rte_event_port_profile_links_set(0, 0, hq, NULL, 4, 0);
> +        rte_event_port_profile_links_set(0, 0, lq, NULL, 4, 1);
> +
> +Worker path:
> +
> +.. code-block:: c
> +
> +        uint8_t profile_id_to_switch;
> +
> +        while (1) {
> +            deq = rte_event_dequeue_burst(0, 0, &ev, 1, 0);
> +            if (deq == 0) {
> +                profile_id_to_switch = app_find_profile_id_to_switch();
> +                rte_event_port_profile_switch(0, 0, profile_id_to_switch);
> +                continue;
> +            }
> +
> +            // Process the event received.
> +        }
> +
>  Starting the EventDev
>  ~~~~~~~~~~~~~~~~~~~~~
>
> diff --git a/doc/guides/rel_notes/release_23_11.rst b/doc/guides/rel_notes/release_23_11.rst
> index b34ddc0860..e08e2eadce 100644
> --- a/doc/guides/rel_notes/release_23_11.rst
> +++ b/doc/guides/rel_notes/release_23_11.rst
> @@ -89,6 +89,16 @@ New Features
>    * Added support for ``remaining_ticks_get`` timer adapter PMD callback
>      to get the remaining ticks to expire for a given event timer.
>
> +* **Added eventdev support to link queues to port with link profile.**
> +
> +  Introduced event link profiles that can be used to associated links between
> +  event queues and an event port with a unique identifier termed as link profile.
> +  The profile can be used to switch between the associated links in fast-path
> +  without the additional overhead of linking/unlinking and waiting for unlinking.
> +
> +  * Added ``rte_event_port_profile_links_set``, ``rte_event_port_profile_unlink``
> +    ``rte_event_port_profile_links_get`` and ``rte_event_port_profile_switch``
> +    APIs to enable this feature.
>
>  Removed Items
>  -------------
> diff --git a/lib/eventdev/eventdev_pmd.h b/lib/eventdev/eventdev_pmd.h
> index f62f42e140..9585c0ca24 100644
> --- a/lib/eventdev/eventdev_pmd.h
> +++ b/lib/eventdev/eventdev_pmd.h
> @@ -119,8 +119,8 @@ struct rte_eventdev_data {
>         /**< Array of port configuration structures. */
>         struct rte_event_queue_conf queues_cfg[RTE_EVENT_MAX_QUEUES_PER_DEV];
>         /**< Array of queue configuration structures. */
> -       uint16_t links_map[RTE_EVENT_MAX_PORTS_PER_DEV *
> -                          RTE_EVENT_MAX_QUEUES_PER_DEV];
> +       uint16_t links_map[RTE_EVENT_MAX_PROFILES_PER_PORT]
> +                         [RTE_EVENT_MAX_PORTS_PER_DEV * RTE_EVENT_MAX_QUEUES_PER_DEV];
>         /**< Memory to store queues to port connections. */
>         void *dev_private;
>         /**< PMD-specific private data */
> @@ -178,6 +178,9 @@ struct rte_eventdev {
>         event_tx_adapter_enqueue_t txa_enqueue;
>         /**< Pointer to PMD eth Tx adapter enqueue function. */
>         event_crypto_adapter_enqueue_t ca_enqueue;
> +       /**< PMD Crypto adapter enqueue function. */
> +       event_profile_switch_t profile_switch;
> +       /**< PMD Event switch profile function. */
>
>         uint64_t reserved_64s[4]; /**< Reserved for future fields */
>         void *reserved_ptrs[3];   /**< Reserved for future fields */
> @@ -437,6 +440,32 @@ typedef int (*eventdev_port_link_t)(struct rte_eventdev *dev, void *port,
>                 const uint8_t queues[], const uint8_t priorities[],
>                 uint16_t nb_links);
>
> +/**
> + * Link multiple source event queues associated with a link profile to a
> + * destination event port.
> + *
> + * @param dev
> + *   Event device pointer
> + * @param port
> + *   Event port pointer
> + * @param queues
> + *   Points to an array of *nb_links* event queues to be linked
> + *   to the event port.
> + * @param priorities
> + *   Points to an array of *nb_links* service priorities associated with each
> + *   event queue link to event port.
> + * @param nb_links
> + *   The number of links to establish.
> + * @param profile_id
> + *   The profile ID to associate the links.
> + *
> + * @return
> + *   Returns 0 on success.
> + */
> +typedef int (*eventdev_port_link_profile_t)(struct rte_eventdev *dev, void *port,
> +                                           const uint8_t queues[], const uint8_t priorities[],
> +                                           uint16_t nb_links, uint8_t profile_id);
> +
>  /**
>   * Unlink multiple source event queues from destination event port.
>   *
> @@ -455,6 +484,28 @@ typedef int (*eventdev_port_link_t)(struct rte_eventdev *dev, void *port,
>  typedef int (*eventdev_port_unlink_t)(struct rte_eventdev *dev, void *port,
>                 uint8_t queues[], uint16_t nb_unlinks);
>
> +/**
> + * Unlink multiple source event queues associated with a link profile from
> + * destination event port.
> + *
> + * @param dev
> + *   Event device pointer
> + * @param port
> + *   Event port pointer
> + * @param queues
> + *   An array of *nb_unlinks* event queues to be unlinked from the event port.
> + * @param nb_unlinks
> + *   The number of unlinks to establish
> + * @param profile_id
> + *   The profile ID of the associated links.
> + *
> + * @return
> + *   Returns 0 on success.
> + */
> +typedef int (*eventdev_port_unlink_profile_t)(struct rte_eventdev *dev, void *port,
> +                                             uint8_t queues[], uint16_t nb_unlinks,
> +                                             uint8_t profile_id);
> +
>  /**
>   * Unlinks in progress. Returns number of unlinks that the PMD is currently
>   * performing, but have not yet been completed.
> @@ -1348,8 +1399,12 @@ struct eventdev_ops {
>
>         eventdev_port_link_t port_link;
>         /**< Link event queues to an event port. */
> +       eventdev_port_link_profile_t port_link_profile;
> +       /**< Link event queues associated with a profile to an event port. */
>         eventdev_port_unlink_t port_unlink;
>         /**< Unlink event queues from an event port. */
> +       eventdev_port_unlink_profile_t port_unlink_profile;
> +       /**< Unlink event queues associated with a profile from an event port. */
>         eventdev_port_unlinks_in_progress_t port_unlinks_in_progress;
>         /**< Unlinks in progress on an event port. */
>         eventdev_dequeue_timeout_ticks_t timeout_ticks;
> diff --git a/lib/eventdev/eventdev_private.c b/lib/eventdev/eventdev_private.c
> index 1d3d9d357e..b90a3a3833 100644
> --- a/lib/eventdev/eventdev_private.c
> +++ b/lib/eventdev/eventdev_private.c
> @@ -81,6 +81,13 @@ dummy_event_crypto_adapter_enqueue(__rte_unused void *port,
>         return 0;
>  }
>
> +static int
> +dummy_event_port_profile_switch(__rte_unused void *port, __rte_unused uint8_t profile_id)
> +{
> +       RTE_EDEV_LOG_ERR("change profile requested for unconfigured event device");
> +       return -EINVAL;
> +}
> +
>  void
>  event_dev_fp_ops_reset(struct rte_event_fp_ops *fp_op)
>  {
> @@ -97,6 +104,7 @@ event_dev_fp_ops_reset(struct rte_event_fp_ops *fp_op)
>                 .txa_enqueue_same_dest =
>                         dummy_event_tx_adapter_enqueue_same_dest,
>                 .ca_enqueue = dummy_event_crypto_adapter_enqueue,
> +               .profile_switch = dummy_event_port_profile_switch,
>                 .data = dummy_data,
>         };
>
> @@ -117,5 +125,6 @@ event_dev_fp_ops_set(struct rte_event_fp_ops *fp_op,
>         fp_op->txa_enqueue = dev->txa_enqueue;
>         fp_op->txa_enqueue_same_dest = dev->txa_enqueue_same_dest;
>         fp_op->ca_enqueue = dev->ca_enqueue;
> +       fp_op->profile_switch = dev->profile_switch;
>         fp_op->data = dev->data->ports;
>  }
> diff --git a/lib/eventdev/eventdev_trace.h b/lib/eventdev/eventdev_trace.h
> index f008ef0091..9c2b261c06 100644
> --- a/lib/eventdev/eventdev_trace.h
> +++ b/lib/eventdev/eventdev_trace.h
> @@ -76,6 +76,17 @@ RTE_TRACE_POINT(
>         rte_trace_point_emit_int(rc);
>  )
>
> +RTE_TRACE_POINT(
> +       rte_eventdev_trace_port_profile_links_set,
> +       RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id,
> +               uint16_t nb_links, uint8_t profile_id, int rc),
> +       rte_trace_point_emit_u8(dev_id);
> +       rte_trace_point_emit_u8(port_id);
> +       rte_trace_point_emit_u16(nb_links);
> +       rte_trace_point_emit_u8(profile_id);
> +       rte_trace_point_emit_int(rc);
> +)
> +
>  RTE_TRACE_POINT(
>         rte_eventdev_trace_port_unlink,
>         RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id,
> @@ -86,6 +97,17 @@ RTE_TRACE_POINT(
>         rte_trace_point_emit_int(rc);
>  )
>
> +RTE_TRACE_POINT(
> +       rte_eventdev_trace_port_profile_unlink,
> +       RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id,
> +               uint16_t nb_unlinks, uint8_t profile_id, int rc),
> +       rte_trace_point_emit_u8(dev_id);
> +       rte_trace_point_emit_u8(port_id);
> +       rte_trace_point_emit_u16(nb_unlinks);
> +       rte_trace_point_emit_u8(profile_id);
> +       rte_trace_point_emit_int(rc);
> +)
> +
>  RTE_TRACE_POINT(
>         rte_eventdev_trace_start,
>         RTE_TRACE_POINT_ARGS(uint8_t dev_id, int rc),
> @@ -487,6 +509,16 @@ RTE_TRACE_POINT(
>         rte_trace_point_emit_int(count);
>  )
>
> +RTE_TRACE_POINT(
> +       rte_eventdev_trace_port_profile_links_get,
> +       RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id, uint8_t profile_id,
> +               int count),
> +       rte_trace_point_emit_u8(dev_id);
> +       rte_trace_point_emit_u8(port_id);
> +       rte_trace_point_emit_u8(profile_id);
> +       rte_trace_point_emit_int(count);
> +)
> +
>  RTE_TRACE_POINT(
>         rte_eventdev_trace_port_unlinks_in_progress,
>         RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id),
> diff --git a/lib/eventdev/eventdev_trace_points.c b/lib/eventdev/eventdev_trace_points.c
> index 76144cfe75..8024e07531 100644
> --- a/lib/eventdev/eventdev_trace_points.c
> +++ b/lib/eventdev/eventdev_trace_points.c
> @@ -19,9 +19,15 @@ RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_setup,
>  RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_link,
>         lib.eventdev.port.link)
>
> +RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_profile_links_set,
> +       lib.eventdev.port.profile.links.set)
> +
>  RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_unlink,
>         lib.eventdev.port.unlink)
>
> +RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_profile_unlink,
> +       lib.eventdev.port.profile.unlink)
> +
>  RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_start,
>         lib.eventdev.start)
>
> @@ -40,6 +46,9 @@ RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_deq_burst,
>  RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_maintain,
>         lib.eventdev.maintain)
>
> +RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_profile_switch,
> +       lib.eventdev.port.profile.switch)
> +
>  /* Eventdev Rx adapter trace points */
>  RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_eth_rx_adapter_create,
>         lib.eventdev.rx.adapter.create)
> @@ -206,6 +215,9 @@ RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_default_conf_get,
>  RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_links_get,
>         lib.eventdev.port.links.get)
>
> +RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_profile_links_get,
> +       lib.eventdev.port.profile.links.get)
> +
>  RTE_TRACE_POINT_REGISTER(rte_eventdev_trace_port_unlinks_in_progress,
>         lib.eventdev.port.unlinks.in.progress)
>
> diff --git a/lib/eventdev/rte_eventdev.c b/lib/eventdev/rte_eventdev.c
> index 6ab4524332..33a3154d5d 100644
> --- a/lib/eventdev/rte_eventdev.c
> +++ b/lib/eventdev/rte_eventdev.c
> @@ -95,6 +95,7 @@ rte_event_dev_info_get(uint8_t dev_id, struct rte_event_dev_info *dev_info)
>                 return -EINVAL;
>
>         memset(dev_info, 0, sizeof(struct rte_event_dev_info));
> +       dev_info->max_profiles_per_port = 1;
>
>         if (*dev->dev_ops->dev_infos_get == NULL)
>                 return -ENOTSUP;
> @@ -270,7 +271,7 @@ event_dev_port_config(struct rte_eventdev *dev, uint8_t nb_ports)
>         void **ports;
>         uint16_t *links_map;
>         struct rte_event_port_conf *ports_cfg;
> -       unsigned int i;
> +       unsigned int i, j;
>
>         RTE_EDEV_LOG_DEBUG("Setup %d ports on device %u", nb_ports,
>                          dev->data->dev_id);
> @@ -281,7 +282,6 @@ event_dev_port_config(struct rte_eventdev *dev, uint8_t nb_ports)
>
>                 ports = dev->data->ports;
>                 ports_cfg = dev->data->ports_cfg;
> -               links_map = dev->data->links_map;
>
>                 for (i = nb_ports; i < old_nb_ports; i++)
>                         (*dev->dev_ops->port_release)(ports[i]);
> @@ -297,9 +297,11 @@ event_dev_port_config(struct rte_eventdev *dev, uint8_t nb_ports)
>                                 sizeof(ports[0]) * new_ps);
>                         memset(ports_cfg + old_nb_ports, 0,
>                                 sizeof(ports_cfg[0]) * new_ps);
> -                       for (i = old_links_map_end; i < links_map_end; i++)
> -                               links_map[i] =
> -                                       EVENT_QUEUE_SERVICE_PRIORITY_INVALID;
> +                       for (i = 0; i < RTE_EVENT_MAX_PROFILES_PER_PORT; i++) {
> +                               links_map = dev->data->links_map[i];
> +                               for (j = old_links_map_end; j < links_map_end; j++)
> +                                       links_map[j] = EVENT_QUEUE_SERVICE_PRIORITY_INVALID;
> +                       }
>                 }
>         } else {
>                 if (*dev->dev_ops->port_release == NULL)
> @@ -953,21 +955,45 @@ rte_event_port_link(uint8_t dev_id, uint8_t port_id,
>                     const uint8_t queues[], const uint8_t priorities[],
>                     uint16_t nb_links)
>  {
> -       struct rte_eventdev *dev;
> -       uint8_t queues_list[RTE_EVENT_MAX_QUEUES_PER_DEV];
> +       return rte_event_port_profile_links_set(dev_id, port_id, queues, priorities, nb_links, 0);
> +}
> +
> +int
> +rte_event_port_profile_links_set(uint8_t dev_id, uint8_t port_id, const uint8_t queues[],
> +                                const uint8_t priorities[], uint16_t nb_links, uint8_t profile_id)
> +{
>         uint8_t priorities_list[RTE_EVENT_MAX_QUEUES_PER_DEV];
> +       uint8_t queues_list[RTE_EVENT_MAX_QUEUES_PER_DEV];
> +       struct rte_event_dev_info info;
> +       struct rte_eventdev *dev;
>         uint16_t *links_map;
>         int i, diag;
>
>         RTE_EVENTDEV_VALID_DEVID_OR_ERRNO_RET(dev_id, EINVAL, 0);
>         dev = &rte_eventdevs[dev_id];
>
> +       if (*dev->dev_ops->dev_infos_get == NULL)
> +               return -ENOTSUP;
> +
> +       (*dev->dev_ops->dev_infos_get)(dev, &info);
> +       if (profile_id >= RTE_EVENT_MAX_PROFILES_PER_PORT ||
> +           profile_id >= info.max_profiles_per_port) {
> +               RTE_EDEV_LOG_ERR("Invalid profile_id=%" PRIu8, profile_id);
> +               return -EINVAL;
> +       }
> +
>         if (*dev->dev_ops->port_link == NULL) {
>                 RTE_EDEV_LOG_ERR("Function not supported\n");
>                 rte_errno = ENOTSUP;
>                 return 0;
>         }
>
> +       if (profile_id && *dev->dev_ops->port_link_profile == NULL) {
> +               RTE_EDEV_LOG_ERR("Function not supported\n");
> +               rte_errno = ENOTSUP;
> +               return 0;
> +       }
> +
>         if (!is_valid_port(dev, port_id)) {
>                 RTE_EDEV_LOG_ERR("Invalid port_id=%" PRIu8, port_id);
>                 rte_errno = EINVAL;
> @@ -995,18 +1021,22 @@ rte_event_port_link(uint8_t dev_id, uint8_t port_id,
>                         return 0;
>                 }
>
> -       diag = (*dev->dev_ops->port_link)(dev, dev->data->ports[port_id],
> -                                               queues, priorities, nb_links);
> +       if (profile_id)
> +               diag = (*dev->dev_ops->port_link_profile)(dev, dev->data->ports[port_id], queues,
> +                                                         priorities, nb_links, profile_id);
> +       else
> +               diag = (*dev->dev_ops->port_link)(dev, dev->data->ports[port_id], queues,
> +                                                 priorities, nb_links);
>         if (diag < 0)
>                 return diag;
>
> -       links_map = dev->data->links_map;
> +       links_map = dev->data->links_map[profile_id];
>         /* Point links_map to this port specific area */
>         links_map += (port_id * RTE_EVENT_MAX_QUEUES_PER_DEV);
>         for (i = 0; i < diag; i++)
>                 links_map[queues[i]] = (uint8_t)priorities[i];
>
> -       rte_eventdev_trace_port_link(dev_id, port_id, nb_links, diag);
> +       rte_eventdev_trace_port_profile_links_set(dev_id, port_id, nb_links, profile_id, diag);
>         return diag;
>  }
>
> @@ -1014,27 +1044,51 @@ int
>  rte_event_port_unlink(uint8_t dev_id, uint8_t port_id,
>                       uint8_t queues[], uint16_t nb_unlinks)
>  {
> -       struct rte_eventdev *dev;
> +       return rte_event_port_profile_unlink(dev_id, port_id, queues, nb_unlinks, 0);
> +}
> +
> +int
> +rte_event_port_profile_unlink(uint8_t dev_id, uint8_t port_id, uint8_t queues[],
> +                             uint16_t nb_unlinks, uint8_t profile_id)
> +{
>         uint8_t all_queues[RTE_EVENT_MAX_QUEUES_PER_DEV];
> -       int i, diag, j;
> +       struct rte_event_dev_info info;
> +       struct rte_eventdev *dev;
>         uint16_t *links_map;
> +       int i, diag, j;
>
>         RTE_EVENTDEV_VALID_DEVID_OR_ERRNO_RET(dev_id, EINVAL, 0);
>         dev = &rte_eventdevs[dev_id];
>
> +       if (*dev->dev_ops->dev_infos_get == NULL)
> +               return -ENOTSUP;
> +
> +       (*dev->dev_ops->dev_infos_get)(dev, &info);
> +       if (profile_id >= RTE_EVENT_MAX_PROFILES_PER_PORT ||
> +           profile_id >= info.max_profiles_per_port) {
> +               RTE_EDEV_LOG_ERR("Invalid profile_id=%" PRIu8, profile_id);
> +               return -EINVAL;
> +       }
> +
>         if (*dev->dev_ops->port_unlink == NULL) {
>                 RTE_EDEV_LOG_ERR("Function not supported");
>                 rte_errno = ENOTSUP;
>                 return 0;
>         }
>
> +       if (profile_id && *dev->dev_ops->port_unlink_profile == NULL) {
> +               RTE_EDEV_LOG_ERR("Function not supported");
> +               rte_errno = ENOTSUP;
> +               return 0;
> +       }
> +
>         if (!is_valid_port(dev, port_id)) {
>                 RTE_EDEV_LOG_ERR("Invalid port_id=%" PRIu8, port_id);
>                 rte_errno = EINVAL;
>                 return 0;
>         }
>
> -       links_map = dev->data->links_map;
> +       links_map = dev->data->links_map[profile_id];
>         /* Point links_map to this port specific area */
>         links_map += (port_id * RTE_EVENT_MAX_QUEUES_PER_DEV);
>
> @@ -1063,16 +1117,19 @@ rte_event_port_unlink(uint8_t dev_id, uint8_t port_id,
>                         return 0;
>                 }
>
> -       diag = (*dev->dev_ops->port_unlink)(dev, dev->data->ports[port_id],
> -                                       queues, nb_unlinks);
> -
> +       if (profile_id)
> +               diag = (*dev->dev_ops->port_unlink_profile)(dev, dev->data->ports[port_id], queues,
> +                                                           nb_unlinks, profile_id);
> +       else
> +               diag = (*dev->dev_ops->port_unlink)(dev, dev->data->ports[port_id], queues,
> +                                                   nb_unlinks);
>         if (diag < 0)
>                 return diag;
>
>         for (i = 0; i < diag; i++)
>                 links_map[queues[i]] = EVENT_QUEUE_SERVICE_PRIORITY_INVALID;
>
> -       rte_eventdev_trace_port_unlink(dev_id, port_id, nb_unlinks, diag);
> +       rte_eventdev_trace_port_profile_unlink(dev_id, port_id, nb_unlinks, profile_id, diag);
>         return diag;
>  }
>
> @@ -1116,7 +1173,8 @@ rte_event_port_links_get(uint8_t dev_id, uint8_t port_id,
>                 return -EINVAL;
>         }
>
> -       links_map = dev->data->links_map;
> +       /* Use the default profile_id. */
> +       links_map = dev->data->links_map[0];
>         /* Point links_map to this port specific area */
>         links_map += (port_id * RTE_EVENT_MAX_QUEUES_PER_DEV);
>         for (i = 0; i < dev->data->nb_queues; i++) {
> @@ -1132,6 +1190,49 @@ rte_event_port_links_get(uint8_t dev_id, uint8_t port_id,
>         return count;
>  }
>
> +int
> +rte_event_port_profile_links_get(uint8_t dev_id, uint8_t port_id, uint8_t queues[],
> +                                uint8_t priorities[], uint8_t profile_id)
> +{
> +       struct rte_event_dev_info info;
> +       struct rte_eventdev *dev;
> +       uint16_t *links_map;
> +       int i, count = 0;
> +
> +       RTE_EVENTDEV_VALID_DEVID_OR_ERR_RET(dev_id, -EINVAL);
> +
> +       dev = &rte_eventdevs[dev_id];
> +       if (*dev->dev_ops->dev_infos_get == NULL)
> +               return -ENOTSUP;
> +
> +       (*dev->dev_ops->dev_infos_get)(dev, &info);
> +       if (profile_id >= RTE_EVENT_MAX_PROFILES_PER_PORT ||
> +           profile_id >= info.max_profiles_per_port) {
> +               RTE_EDEV_LOG_ERR("Invalid profile_id=%" PRIu8, profile_id);
> +               return -EINVAL;
> +       }
> +
> +       if (!is_valid_port(dev, port_id)) {
> +               RTE_EDEV_LOG_ERR("Invalid port_id=%" PRIu8, port_id);
> +               return -EINVAL;
> +       }
> +
> +       links_map = dev->data->links_map[profile_id];
> +       /* Point links_map to this port specific area */
> +       links_map += (port_id * RTE_EVENT_MAX_QUEUES_PER_DEV);
> +       for (i = 0; i < dev->data->nb_queues; i++) {
> +               if (links_map[i] != EVENT_QUEUE_SERVICE_PRIORITY_INVALID) {
> +                       queues[count] = i;
> +                       priorities[count] = (uint8_t)links_map[i];
> +                       ++count;
> +               }
> +       }
> +
> +       rte_eventdev_trace_port_profile_links_get(dev_id, port_id, profile_id, count);
> +
> +       return count;
> +}
> +
>  int
>  rte_event_dequeue_timeout_ticks(uint8_t dev_id, uint64_t ns,
>                                  uint64_t *timeout_ticks)
> @@ -1440,7 +1541,7 @@ eventdev_data_alloc(uint8_t dev_id, struct rte_eventdev_data **data,
>  {
>         char mz_name[RTE_EVENTDEV_NAME_MAX_LEN];
>         const struct rte_memzone *mz;
> -       int n;
> +       int i, n;
>
>         /* Generate memzone name */
>         n = snprintf(mz_name, sizeof(mz_name), "rte_eventdev_data_%u", dev_id);
> @@ -1460,11 +1561,10 @@ eventdev_data_alloc(uint8_t dev_id, struct rte_eventdev_data **data,
>         *data = mz->addr;
>         if (rte_eal_process_type() == RTE_PROC_PRIMARY) {
>                 memset(*data, 0, sizeof(struct rte_eventdev_data));
> -               for (n = 0; n < RTE_EVENT_MAX_PORTS_PER_DEV *
> -                                       RTE_EVENT_MAX_QUEUES_PER_DEV;
> -                    n++)
> -                       (*data)->links_map[n] =
> -                               EVENT_QUEUE_SERVICE_PRIORITY_INVALID;
> +               for (i = 0; i < RTE_EVENT_MAX_PROFILES_PER_PORT; i++)
> +                       for (n = 0; n < RTE_EVENT_MAX_PORTS_PER_DEV * RTE_EVENT_MAX_QUEUES_PER_DEV;
> +                            n++)
> +                               (*data)->links_map[i][n] = EVENT_QUEUE_SERVICE_PRIORITY_INVALID;
>         }
>
>         return 0;
> diff --git a/lib/eventdev/rte_eventdev.h b/lib/eventdev/rte_eventdev.h
> index 2ba8a7b090..23cbff939f 100644
> --- a/lib/eventdev/rte_eventdev.h
> +++ b/lib/eventdev/rte_eventdev.h
> @@ -320,6 +320,12 @@ struct rte_event;
>   * rte_event_queue_setup().
>   */
>
> +#define RTE_EVENT_DEV_CAP_PROFILE_LINK (1ULL << 12)
> +/**< Event device is capable of supporting multiple link profiles per event port
> + * i.e., the value of `rte_event_dev_info::max_profiles_per_port` is greater
> + * than one.
> + */
> +
>  /* Event device priority levels */
>  #define RTE_EVENT_DEV_PRIORITY_HIGHEST   0
>  /**< Highest priority expressed across eventdev subsystem
> @@ -446,6 +452,10 @@ struct rte_event_dev_info {
>          * device. These ports and queues are not accounted for in
>          * max_event_ports or max_event_queues.
>          */
> +       uint8_t max_profiles_per_port;
> +       /**< Maximum number of event queue profiles per event port.
> +        * A device that doesn't support multiple profiles will set this as 1.
> +        */
>  };
>
>  /**
> @@ -1536,6 +1546,10 @@ rte_event_dequeue_timeout_ticks(uint8_t dev_id, uint64_t ns,
>   * latency of critical work by establishing the link with more event ports
>   * at runtime.
>   *
> + * When the value of ``rte_event_dev_info::max_profiles_per_port`` is greater
> + * than or equal to one, this function links the event queues to the default
> + * profile_id i.e. profile_id 0 of the event port.
> + *
>   * @param dev_id
>   *   The identifier of the device.
>   *
> @@ -1593,6 +1607,10 @@ rte_event_port_link(uint8_t dev_id, uint8_t port_id,
>   * Event queue(s) to event port unlink establishment can be changed at runtime
>   * without re-configuring the device.
>   *
> + * When the value of ``rte_event_dev_info::max_profiles_per_port`` is greater
> + * than or equal to one, this function unlinks the event queues from the default
> + * profile identifier i.e. profile 0 of the event port.
> + *
>   * @see rte_event_port_unlinks_in_progress() to poll for completed unlinks.
>   *
>   * @param dev_id
> @@ -1626,6 +1644,136 @@ int
>  rte_event_port_unlink(uint8_t dev_id, uint8_t port_id,
>                       uint8_t queues[], uint16_t nb_unlinks);
>
> +/**
> + * Link multiple source event queues supplied in *queues* to the destination
> + * event port designated by its *port_id* with associated profile identifier
> + * supplied in *profile_id* with service priorities supplied in *priorities*
> + * on the event device designated by its *dev_id*.
> + *
> + * If *profile_id* is set to 0 then, the links created by the call `rte_event_port_link`
> + * will be overwritten.
> + *
> + * Event ports by default use profile_id 0 unless it is changed using the
> + * call ``rte_event_port_profile_switch()``.
> + *
> + * The link establishment shall enable the event port *port_id* from
> + * receiving events from the specified event queue(s) supplied in *queues*
> + *
> + * An event queue may link to one or more event ports.
> + * The number of links can be established from an event queue to event port is
> + * implementation defined.
> + *
> + * Event queue(s) to event port link establishment can be changed at runtime
> + * without re-configuring the device to support scaling and to reduce the
> + * latency of critical work by establishing the link with more event ports
> + * at runtime.
> + *
> + * @param dev_id
> + *   The identifier of the device.
> + *
> + * @param port_id
> + *   Event port identifier to select the destination port to link.
> + *
> + * @param queues
> + *   Points to an array of *nb_links* event queues to be linked
> + *   to the event port.
> + *   NULL value is allowed, in which case this function links all the configured
> + *   event queues *nb_event_queues* which previously supplied to
> + *   rte_event_dev_configure() to the event port *port_id*
> + *
> + * @param priorities
> + *   Points to an array of *nb_links* service priorities associated with each
> + *   event queue link to event port.
> + *   The priority defines the event port's servicing priority for
> + *   event queue, which may be ignored by an implementation.
> + *   The requested priority should in the range of
> + *   [RTE_EVENT_DEV_PRIORITY_HIGHEST, RTE_EVENT_DEV_PRIORITY_LOWEST].
> + *   The implementation shall normalize the requested priority to
> + *   implementation supported priority value.
> + *   NULL value is allowed, in which case this function links the event queues
> + *   with RTE_EVENT_DEV_PRIORITY_NORMAL servicing priority
> + *
> + * @param nb_links
> + *   The number of links to establish. This parameter is ignored if queues is
> + *   NULL.
> + *
> + * @param profile_id
> + *   The profile identifier associated with the links between event queues and
> + *   event port. Should be less than the max capability reported by
> + *   ``rte_event_dev_info::max_profiles_per_port``
> + *
> + * @return
> + * The number of links actually established. The return value can be less than
> + * the value of the *nb_links* parameter when the implementation has the
> + * limitation on specific queue to port link establishment or if invalid
> + * parameters are specified in *queues*
> + * If the return value is less than *nb_links*, the remaining links at the end
> + * of link[] are not established, and the caller has to take care of them.
> + * If return value is less than *nb_links* then implementation shall update the
> + * rte_errno accordingly, Possible rte_errno values are
> + * (EDQUOT) Quota exceeded(Application tried to link the queue configured with
> + *  RTE_EVENT_QUEUE_CFG_SINGLE_LINK to more than one event ports)
> + * (EINVAL) Invalid parameter
> + *
> + */
> +__rte_experimental
> +int
> +rte_event_port_profile_links_set(uint8_t dev_id, uint8_t port_id, const uint8_t queues[],
> +                                const uint8_t priorities[], uint16_t nb_links, uint8_t profile_id);
> +
> +/**
> + * Unlink multiple source event queues supplied in *queues* that belong to profile
> + * designated by *profile_id* from the destination event port designated by its
> + * *port_id* on the event device designated by its *dev_id*.
> + *
> + * If *profile_id* is set to 0 i.e., the default profile then, then this function
> + * will act as ``rte_event_port_unlink``.
> + *
> + * The unlink call issues an async request to disable the event port *port_id*
> + * from receiving events from the specified event queue *queue_id*.
> + * Event queue(s) to event port unlink establishment can be changed at runtime
> + * without re-configuring the device.
> + *
> + * @see rte_event_port_unlinks_in_progress() to poll for completed unlinks.
> + *
> + * @param dev_id
> + *   The identifier of the device.
> + *
> + * @param port_id
> + *   Event port identifier to select the destination port to unlink.
> + *
> + * @param queues
> + *   Points to an array of *nb_unlinks* event queues to be unlinked
> + *   from the event port.
> + *   NULL value is allowed, in which case this function unlinks all the
> + *   event queue(s) from the event port *port_id*.
> + *
> + * @param nb_unlinks
> + *   The number of unlinks to establish. This parameter is ignored if queues is
> + *   NULL.
> + *
> + * @param profile_id
> + *   The profile identifier associated with the links between event queues and
> + *   event port. Should be less than the max capability reported by
> + *   ``rte_event_dev_info::max_profiles_per_port``
> + *
> + * @return
> + * The number of unlinks successfully requested. The return value can be less
> + * than the value of the *nb_unlinks* parameter when the implementation has the
> + * limitation on specific queue to port unlink establishment or
> + * if invalid parameters are specified.
> + * If the return value is less than *nb_unlinks*, the remaining queues at the
> + * end of queues[] are not unlinked, and the caller has to take care of them.
> + * If return value is less than *nb_unlinks* then implementation shall update
> + * the rte_errno accordingly, Possible rte_errno values are
> + * (EINVAL) Invalid parameter
> + *
> + */
> +__rte_experimental
> +int
> +rte_event_port_profile_unlink(uint8_t dev_id, uint8_t port_id, uint8_t queues[],
> +                             uint16_t nb_unlinks, uint8_t profile_id);
> +
>  /**
>   * Returns the number of unlinks in progress.
>   *
> @@ -1680,6 +1828,42 @@ int
>  rte_event_port_links_get(uint8_t dev_id, uint8_t port_id,
>                          uint8_t queues[], uint8_t priorities[]);
>
> +/**
> + * Retrieve the list of source event queues and its service priority
> + * associated to a *profile_id* and linked to the destination event port
> + * designated by its *port_id* on the event device designated by its *dev_id*.
> + *
> + * @param dev_id
> + *   The identifier of the device.
> + *
> + * @param port_id
> + *   Event port identifier.
> + *
> + * @param[out] queues
> + *   Points to an array of *queues* for output.
> + *   The caller has to allocate *RTE_EVENT_MAX_QUEUES_PER_DEV* bytes to
> + *   store the event queue(s) linked with event port *port_id*
> + *
> + * @param[out] priorities
> + *   Points to an array of *priorities* for output.
> + *   The caller has to allocate *RTE_EVENT_MAX_QUEUES_PER_DEV* bytes to
> + *   store the service priority associated with each event queue linked
> + *
> + * @param profile_id
> + *   The profile identifier associated with the links between event queues and
> + *   event port. Should be less than the max capability reported by
> + *   ``rte_event_dev_info::max_profiles_per_port``
> + *
> + * @return
> + * The number of links established on the event port designated by its
> + *  *port_id*.
> + * - <0 on failure.
> + */
> +__rte_experimental
> +int
> +rte_event_port_profile_links_get(uint8_t dev_id, uint8_t port_id, uint8_t queues[],
> +                                uint8_t priorities[], uint8_t profile_id);
> +
>  /**
>   * Retrieve the service ID of the event dev. If the adapter doesn't use
>   * a rte_service function, this function returns -ESRCH.
> @@ -2265,6 +2449,53 @@ rte_event_maintain(uint8_t dev_id, uint8_t port_id, int op)
>         return 0;
>  }
>
> +/**
> + * Change the active profile on an event port.
> + *
> + * This function is used to change the current active profile on an event port
> + * when multiple link profiles are configured on an event port through the
> + * function call ``rte_event_port_profile_links_set``.
> + *
> + * On the subsequent ``rte_event_dequeue_burst`` call, only the event queues
> + * that were associated with the newly active profile will participate in
> + * scheduling.
> + *
> + * @param dev_id
> + *   The identifier of the device.
> + * @param port_id
> + *   The identifier of the event port.
> + * @param profile_id
> + *   The identifier of the profile.
> + * @return
> + *  - 0 on success.
> + *  - -EINVAL if *dev_id*,  *port_id*, or *profile_id* is invalid.
> + */
> +__rte_experimental
> +static inline uint8_t
> +rte_event_port_profile_switch(uint8_t dev_id, uint8_t port_id, uint8_t profile_id)
> +{
> +       const struct rte_event_fp_ops *fp_ops;
> +       void *port;
> +
> +       fp_ops = &rte_event_fp_ops[dev_id];
> +       port = fp_ops->data[port_id];
> +
> +#ifdef RTE_LIBRTE_EVENTDEV_DEBUG
> +       if (dev_id >= RTE_EVENT_MAX_DEVS ||
> +           port_id >= RTE_EVENT_MAX_PORTS_PER_DEV)
> +               return -EINVAL;
> +
> +       if (port == NULL)
> +               return -EINVAL;
> +
> +       if (profile_id >= RTE_EVENT_MAX_PROFILES_PER_PORT)
> +               return -EINVAL;
> +#endif
> +       rte_eventdev_trace_port_profile_switch(dev_id, port_id, profile_id);
> +
> +       return fp_ops->profile_switch(port, profile_id);
> +}
> +
>  #ifdef __cplusplus
>  }
>  #endif
> diff --git a/lib/eventdev/rte_eventdev_core.h b/lib/eventdev/rte_eventdev_core.h
> index c27a52ccc0..5af646ed5c 100644
> --- a/lib/eventdev/rte_eventdev_core.h
> +++ b/lib/eventdev/rte_eventdev_core.h
> @@ -42,6 +42,8 @@ typedef uint16_t (*event_crypto_adapter_enqueue_t)(void *port,
>                                                    uint16_t nb_events);
>  /**< @internal Enqueue burst of events on crypto adapter */
>
> +typedef int (*event_profile_switch_t)(void *port, uint8_t profile);
> +
>  struct rte_event_fp_ops {
>         void **data;
>         /**< points to array of internal port data pointers */
> @@ -65,7 +67,9 @@ struct rte_event_fp_ops {
>         /**< PMD Tx adapter enqueue same destination function. */
>         event_crypto_adapter_enqueue_t ca_enqueue;
>         /**< PMD Crypto adapter enqueue function. */
> -       uintptr_t reserved[5];
> +       event_profile_switch_t profile_switch;
> +       /**< PMD Event switch profile function. */
> +       uintptr_t reserved[4];
>  } __rte_cache_aligned;
>
>  extern struct rte_event_fp_ops rte_event_fp_ops[RTE_EVENT_MAX_DEVS];
> diff --git a/lib/eventdev/rte_eventdev_trace_fp.h b/lib/eventdev/rte_eventdev_trace_fp.h
> index af2172d2a5..04d510ad00 100644
> --- a/lib/eventdev/rte_eventdev_trace_fp.h
> +++ b/lib/eventdev/rte_eventdev_trace_fp.h
> @@ -46,6 +46,14 @@ RTE_TRACE_POINT_FP(
>         rte_trace_point_emit_int(op);
>  )
>
> +RTE_TRACE_POINT_FP(
> +       rte_eventdev_trace_port_profile_switch,
> +       RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id, uint8_t profile),
> +       rte_trace_point_emit_u8(dev_id);
> +       rte_trace_point_emit_u8(port_id);
> +       rte_trace_point_emit_u8(profile);
> +)
> +
>  RTE_TRACE_POINT_FP(
>         rte_eventdev_trace_eth_tx_adapter_enqueue,
>         RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id, void *ev_table,
> diff --git a/lib/eventdev/version.map b/lib/eventdev/version.map
> index 7ce09a87bb..f88decee39 100644
> --- a/lib/eventdev/version.map
> +++ b/lib/eventdev/version.map
> @@ -134,6 +134,10 @@ EXPERIMENTAL {
>
>         # added in 23.11
>         rte_event_eth_rx_adapter_create_ext_with_params;
> +       rte_event_port_profile_links_set;
> +       rte_event_port_profile_unlink;
> +       rte_event_port_profile_links_get;
> +       __rte_eventdev_trace_port_profile_switch;
>  };
>
>  INTERNAL {
> --
> 2.25.1
>

  reply	other threads:[~2023-10-03  6:56 UTC|newest]

Thread overview: 44+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-08-09 14:26 [RFC 0/3] Introduce event " pbhagavatula
2023-08-09 14:26 ` [RFC 1/3] eventdev: introduce " pbhagavatula
2023-08-18 10:27   ` Jerin Jacob
2023-08-09 14:26 ` [RFC 2/3] event/cnxk: implement event " pbhagavatula
2023-08-09 14:26 ` [RFC 3/3] test/event: add event link profile test pbhagavatula
2023-08-09 19:45 ` [RFC 0/3] Introduce event link profiles Mattias Rönnblom
2023-08-10  5:17   ` [EXT] " Pavan Nikhilesh Bhagavatula
2023-08-12  5:52     ` Mattias Rönnblom
2023-08-14 11:29       ` Pavan Nikhilesh Bhagavatula
2023-08-25 18:44 ` [PATCH " pbhagavatula
2023-08-25 18:44   ` [PATCH 1/3] eventdev: introduce " pbhagavatula
2023-08-25 18:44   ` [PATCH 2/3] event/cnxk: implement event " pbhagavatula
2023-08-25 18:44   ` [PATCH 3/3] test/event: add event link profile test pbhagavatula
2023-08-31 20:44   ` [PATCH v2 0/3] Introduce event link profiles pbhagavatula
2023-08-31 20:44     ` [PATCH v2 1/3] eventdev: introduce " pbhagavatula
2023-09-20  4:22       ` Jerin Jacob
2023-08-31 20:44     ` [PATCH v2 2/3] event/cnxk: implement event " pbhagavatula
2023-08-31 20:44     ` [PATCH v2 3/3] test/event: add event link profile test pbhagavatula
2023-09-21 10:28     ` [PATCH v3 0/3] Introduce event link profiles pbhagavatula
2023-09-21 10:28       ` [PATCH v3 1/3] eventdev: introduce " pbhagavatula
2023-09-27 15:23         ` Jerin Jacob
2023-09-21 10:28       ` [PATCH v3 2/3] event/cnxk: implement event " pbhagavatula
2023-09-27 15:29         ` Jerin Jacob
2023-09-21 10:28       ` [PATCH v3 3/3] test/event: add event link profile test pbhagavatula
2023-09-27 14:56       ` [PATCH v3 0/3] Introduce event link profiles Jerin Jacob
2023-09-28 10:12       ` [PATCH v4 " pbhagavatula
2023-09-28 10:12         ` [PATCH v4 1/3] eventdev: introduce " pbhagavatula
2023-10-03  6:55           ` Jerin Jacob [this message]
2023-09-28 10:12         ` [PATCH v4 2/3] event/cnxk: implement event " pbhagavatula
2023-09-28 10:12         ` [PATCH v4 3/3] test/event: add event link profile test pbhagavatula
2023-09-28 14:45         ` [PATCH v4 0/3] Introduce event link profiles Jerin Jacob
2023-09-29  9:27           ` [EXT] " Pavan Nikhilesh Bhagavatula
2023-10-03  7:51         ` [PATCH v5 " pbhagavatula
2023-10-03  7:51           ` [PATCH v5 1/3] eventdev: introduce " pbhagavatula
2023-10-03  7:51           ` [PATCH v5 2/3] event/cnxk: implement event " pbhagavatula
2023-10-03  7:51           ` [PATCH v5 3/3] test/event: add event link profile test pbhagavatula
2023-10-03  9:47           ` [PATCH v6 0/3] Introduce event link profiles pbhagavatula
2023-10-03  9:47             ` [PATCH v6 1/3] eventdev: introduce " pbhagavatula
2023-10-03  9:47             ` [PATCH v6 2/3] event/cnxk: implement event " pbhagavatula
2023-10-03  9:47             ` [PATCH v6 3/3] test/event: add event link profile test pbhagavatula
2023-10-03 10:36             ` [PATCH v6 0/3] Introduce event link profiles Jerin Jacob
2023-10-03 14:12               ` Bruce Richardson
2023-10-03 15:17                 ` Jerin Jacob
2023-10-03 15:32                   ` [EXT] " Pavan Nikhilesh Bhagavatula

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='CALBAE1OCQaki7mGJKWQQvL3UJA5MZtnsXbMp5mu=pxOh_L6zXA@mail.gmail.com' \
    --to=jerinjacobk@gmail.com \
    --cc=abhinandan.gujjar@intel.com \
    --cc=anatoly.burakov@intel.com \
    --cc=dev@dpdk.org \
    --cc=erik.g.carrillo@intel.com \
    --cc=harry.van.haaren@intel.com \
    --cc=hemant.agrawal@nxp.com \
    --cc=jerinj@marvell.com \
    --cc=liangma@liangbit.com \
    --cc=mattias.ronnblom@ericsson.com \
    --cc=pbhagavatula@marvell.com \
    --cc=peter.mccarthy@intel.com \
    --cc=s.v.naga.harish.k@intel.com \
    --cc=sachin.saxena@nxp.com \
    --cc=sthotton@marvell.com \
    --cc=timothy.mcdaniel@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).