DPDK patches and discussions
 help / color / mirror / Atom feed
From: Jerin Jacob <jerinjacobk@gmail.com>
To: Shijith Thotton <sthotton@marvell.com>
Cc: dpdk-dev <dev@dpdk.org>,
	Pavan Nikhilesh <pbhagavatula@marvell.com>,
	 Jerin Jacob <jerinj@marvell.com>
Subject: Re: [PATCH] event/cnxk: add timer adapter periodic mode support
Date: Mon, 24 Jan 2022 14:37:42 +0530	[thread overview]
Message-ID: <CALBAE1N7xG9dD4wTn4ZKs3pq2vR4g_d5rZFe5h28x=f_N7+B+Q@mail.gmail.com> (raw)
In-Reply-To: <5919b346b947e8a9968e974dfdede7d3ae8c0eff.1640273981.git.sthotton@marvell.com>

On Thu, Dec 23, 2021 at 9:36 PM Shijith Thotton <sthotton@marvell.com> wrote:
>
> Add support for event timer adapter periodic mode capability.
>
> Signed-off-by: Shijith Thotton <sthotton@marvell.com>

Acked-by: Jerin Jacob <jerinj@marvell.com>
Applied to dpdk-next-net-eventdev/for-main. Thanks

> ---
> Depends-on: series-20928 (event/cnxk: update min interval calculation)
>
>  drivers/event/cnxk/cnxk_tim_evdev.c | 21 +++++++++++++++++----
>  drivers/event/cnxk/cnxk_tim_evdev.h |  1 +
>  2 files changed, 18 insertions(+), 4 deletions(-)
>
> diff --git a/drivers/event/cnxk/cnxk_tim_evdev.c b/drivers/event/cnxk/cnxk_tim_evdev.c
> index 5d52a39752..4d22ac6ac3 100644
> --- a/drivers/event/cnxk/cnxk_tim_evdev.c
> +++ b/drivers/event/cnxk/cnxk_tim_evdev.c
> @@ -58,7 +58,7 @@ cnxk_tim_chnk_pool_create(struct cnxk_tim_ring *tim_ring,
>                 }
>                 tim_ring->aura = roc_npa_aura_handle_to_aura(
>                         tim_ring->chunk_pool->pool_id);
> -               tim_ring->ena_dfb = 0;
> +               tim_ring->ena_dfb = tim_ring->ena_periodic ? 1 : 0;
>         } else {
>                 tim_ring->chunk_pool = rte_mempool_create(
>                         pool_name, tim_ring->nb_chunks, tim_ring->chunk_sz,
> @@ -112,7 +112,9 @@ cnxk_tim_ring_info_get(const struct rte_event_timer_adapter *adptr,
>         struct cnxk_tim_ring *tim_ring = adptr->data->adapter_priv;
>
>         adptr_info->max_tmo_ns = tim_ring->max_tout;
> -       adptr_info->min_resolution_ns = tim_ring->tck_nsec;
> +       adptr_info->min_resolution_ns = tim_ring->ena_periodic ?
> +                                               tim_ring->max_tout :
> +                                               tim_ring->tck_nsec;
>         rte_memcpy(&adptr_info->conf, &adptr->data->conf,
>                    sizeof(struct rte_event_timer_adapter_conf));
>  }
> @@ -237,6 +239,12 @@ cnxk_tim_ring_create(struct rte_event_timer_adapter *adptr)
>                 goto tim_hw_free;
>         }
>
> +       if (rcfg->flags & RTE_EVENT_TIMER_ADAPTER_F_PERIODIC) {
> +               /* Use 2 buckets to avoid contention */
> +               rcfg->timer_tick_ns /= 2;
> +               tim_ring->ena_periodic = 1;
> +       }
> +
>         if (rcfg->timer_tick_ns < min_intvl_ns) {
>                 if (rcfg->flags & RTE_EVENT_TIMER_ADAPTER_F_ADJUST_RES) {
>                         rcfg->timer_tick_ns = min_intvl_ns;
> @@ -246,6 +254,9 @@ cnxk_tim_ring_create(struct rte_event_timer_adapter *adptr)
>                 }
>         }
>
> +       if (tim_ring->ena_periodic)
> +               rcfg->max_tmo_ns = rcfg->timer_tick_ns * 2;
> +
>         if (rcfg->timer_tick_ns > rcfg->max_tmo_ns) {
>                 plt_err("Max timeout to too high");
>                 rc = -ERANGE;
> @@ -322,7 +333,8 @@ cnxk_tim_ring_create(struct rte_event_timer_adapter *adptr)
>         if (rc < 0)
>                 goto tim_bkt_free;
>
> -       rc = roc_tim_lf_config(&dev->tim, tim_ring->ring_id, clk_src, 0, 0,
> +       rc = roc_tim_lf_config(&dev->tim, tim_ring->ring_id, clk_src,
> +                              tim_ring->ena_periodic, tim_ring->ena_dfb,
>                                tim_ring->nb_bkts, tim_ring->chunk_sz,
>                                tim_ring->tck_int, tim_ring->tck_nsec, clk_freq);
>         if (rc < 0) {
> @@ -493,7 +505,8 @@ cnxk_tim_caps_get(const struct rte_eventdev *evdev, uint64_t flags,
>
>         /* Store evdev pointer for later use. */
>         dev->event_dev = (struct rte_eventdev *)(uintptr_t)evdev;
> -       *caps = RTE_EVENT_TIMER_ADAPTER_CAP_INTERNAL_PORT;
> +       *caps = RTE_EVENT_TIMER_ADAPTER_CAP_INTERNAL_PORT |
> +               RTE_EVENT_TIMER_ADAPTER_CAP_PERIODIC;
>         *ops = &cnxk_tim_ops;
>
>         return 0;
> diff --git a/drivers/event/cnxk/cnxk_tim_evdev.h b/drivers/event/cnxk/cnxk_tim_evdev.h
> index 6b5342cc34..91a90ee2ce 100644
> --- a/drivers/event/cnxk/cnxk_tim_evdev.h
> +++ b/drivers/event/cnxk/cnxk_tim_evdev.h
> @@ -135,6 +135,7 @@ struct cnxk_tim_ring {
>         uint8_t enable_stats;
>         uint8_t disable_npa;
>         uint8_t ena_dfb;
> +       uint8_t ena_periodic;
>         uint16_t ring_id;
>         uint32_t aura;
>         uint64_t nb_timers;
> --
> 2.25.1
>

      reply	other threads:[~2022-01-24  9:08 UTC|newest]

Thread overview: 2+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-12-23 16:06 Shijith Thotton
2022-01-24  9:07 ` Jerin Jacob [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='CALBAE1N7xG9dD4wTn4ZKs3pq2vR4g_d5rZFe5h28x=f_N7+B+Q@mail.gmail.com' \
    --to=jerinjacobk@gmail.com \
    --cc=dev@dpdk.org \
    --cc=jerinj@marvell.com \
    --cc=pbhagavatula@marvell.com \
    --cc=sthotton@marvell.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).