DPDK patches and discussions
 help / color / mirror / Atom feed
From: Chengwen Feng <fengchengwen@huawei.com>
To: <thomas@monjalon.net>, <dev@dpdk.org>
Cc: <gmuthukrishn@marvell.com>, <tangkunshan@huawei.com>
Subject: [PATCH 2/2] dma/skeleton: support fill ops
Date: Fri, 26 Jan 2024 08:57:26 +0000	[thread overview]
Message-ID: <20240126085726.54581-3-fengchengwen@huawei.com> (raw)
In-Reply-To: <20240126085726.54581-1-fengchengwen@huawei.com>

Add support for fill operation.

Signed-off-by: Chengwen Feng <fengchengwen@huawei.com>
---
 drivers/dma/skeleton/skeleton_dmadev.c | 53 +++++++++++++++++++++++---
 drivers/dma/skeleton/skeleton_dmadev.h | 16 +++++---
 2 files changed, 59 insertions(+), 10 deletions(-)

diff --git a/drivers/dma/skeleton/skeleton_dmadev.c b/drivers/dma/skeleton/skeleton_dmadev.c
index d1d257a064..48f88f9fc1 100644
--- a/drivers/dma/skeleton/skeleton_dmadev.c
+++ b/drivers/dma/skeleton/skeleton_dmadev.c
@@ -38,7 +38,8 @@ skeldma_info_get(const struct rte_dma_dev *dev, struct rte_dma_info *dev_info,
 	dev_info->dev_capa = RTE_DMA_CAPA_MEM_TO_MEM |
 			     RTE_DMA_CAPA_SVA |
 			     RTE_DMA_CAPA_OPS_COPY |
-			     RTE_DMA_CAPA_OPS_COPY_SG;
+			     RTE_DMA_CAPA_OPS_COPY_SG |
+			     RTE_DMA_CAPA_OPS_FILL;
 	dev_info->max_vchans = 1;
 	dev_info->max_desc = SKELDMA_MAX_DESC;
 	dev_info->min_desc = SKELDMA_MIN_DESC;
@@ -100,8 +101,19 @@ do_copy_sg(struct skeldma_desc *desc)
 	}
 }
 
+static inline void
+do_fill(struct skeldma_desc *desc)
+{
+	uint8_t *fills = (uint8_t *)&desc->fill.pattern;
+	uint8_t *dst = (uint8_t *)desc->fill.dst;
+	uint32_t i;
+
+	for (i = 0; i < desc->fill.len; i++)
+		dst[i] = fills[i % 8];
+}
+
 static uint32_t
-cpucopy_thread(void *param)
+cpuwork_thread(void *param)
 {
 #define SLEEP_THRESHOLD		10000
 #define SLEEP_US_VAL		10
@@ -127,6 +139,8 @@ cpucopy_thread(void *param)
 			rte_memcpy(desc->copy.dst, desc->copy.src, desc->copy.len);
 		else if (desc->op == SKELDMA_OP_COPY_SG)
 			do_copy_sg(desc);
+		else if (desc->op == SKELDMA_OP_FILL)
+			do_fill(desc);
 
 		__atomic_fetch_add(&hw->completed_count, 1, __ATOMIC_RELEASE);
 		(void)rte_ring_enqueue(hw->desc_completed, (void *)desc);
@@ -162,7 +176,7 @@ skeldma_start(struct rte_dma_dev *dev)
 	 * 1) fflush pending/running/completed ring to empty ring.
 	 * 2) init ring idx to zero.
 	 * 3) init running statistics.
-	 * 4) mark cpucopy task exit_flag to false.
+	 * 4) mark cpuwork task exit_flag to false.
 	 */
 	fflush_ring(hw, hw->desc_pending);
 	fflush_ring(hw, hw->desc_running);
@@ -178,9 +192,9 @@ skeldma_start(struct rte_dma_dev *dev)
 
 	snprintf(name, sizeof(name), "dma-skel%d", dev->data->dev_id);
 	ret = rte_thread_create_internal_control(&hw->thread, name,
-			cpucopy_thread, dev);
+			cpuwork_thread, dev);
 	if (ret) {
-		SKELDMA_LOG(ERR, "Start cpucopy thread fail!");
+		SKELDMA_LOG(ERR, "Start cpuwork thread fail!");
 		return -EINVAL;
 	}
 
@@ -462,6 +476,34 @@ skeldma_copy_sg(void *dev_private, uint16_t vchan,
 	return hw->ridx++;
 }
 
+static int
+skeldma_fill(void *dev_private, uint16_t vchan,
+	     uint64_t pattern, rte_iova_t dst,
+	     uint32_t length, uint64_t flags)
+{
+	struct skeldma_hw *hw = dev_private;
+	struct skeldma_desc *desc;
+	int ret;
+
+	RTE_SET_USED(vchan);
+
+	ret = rte_ring_dequeue(hw->desc_empty, (void **)&desc);
+	if (ret)
+		return -ENOSPC;
+	desc->op = SKELDMA_OP_FILL;
+	desc->ridx = hw->ridx;
+	desc->fill.dst = (void *)(uintptr_t)dst;
+	desc->fill.len = length;
+	desc->fill.pattern = pattern;
+	if (flags & RTE_DMA_OP_FLAG_SUBMIT)
+		submit(hw, desc);
+	else
+		(void)rte_ring_enqueue(hw->desc_pending, (void *)desc);
+	hw->submitted_count++;
+
+	return hw->ridx++;
+}
+
 static int
 skeldma_submit(void *dev_private, uint16_t vchan)
 {
@@ -573,6 +615,7 @@ skeldma_create(const char *name, struct rte_vdev_device *vdev, int lcore_id)
 	dev->fp_obj->dev_private = dev->data->dev_private;
 	dev->fp_obj->copy = skeldma_copy;
 	dev->fp_obj->copy_sg = skeldma_copy_sg;
+	dev->fp_obj->fill = skeldma_fill;
 	dev->fp_obj->submit = skeldma_submit;
 	dev->fp_obj->completed = skeldma_completed;
 	dev->fp_obj->completed_status = skeldma_completed_status;
diff --git a/drivers/dma/skeleton/skeleton_dmadev.h b/drivers/dma/skeleton/skeleton_dmadev.h
index 7d32dd5095..c9bf3153ba 100644
--- a/drivers/dma/skeleton/skeleton_dmadev.h
+++ b/drivers/dma/skeleton/skeleton_dmadev.h
@@ -16,6 +16,7 @@
 enum skeldma_op {
 	SKELDMA_OP_COPY,
 	SKELDMA_OP_COPY_SG,
+	SKELDMA_OP_FILL,
 };
 
 struct skeldma_desc {
@@ -34,14 +35,19 @@ struct skeldma_desc {
 			uint16_t nb_src;
 			uint16_t nb_dst;
 		} copy_sg;
+		struct {
+			void *dst;
+			uint32_t len;
+			uint64_t pattern;
+		} fill;
 	};
 };
 
 struct skeldma_hw {
-	int lcore_id; /* cpucopy task affinity core */
+	int lcore_id; /* cpuwork task affinity core */
 	int socket_id;
-	rte_thread_t thread; /* cpucopy task thread */
-	volatile int exit_flag; /* cpucopy task exit flag */
+	rte_thread_t thread; /* cpuwork task thread */
+	volatile int exit_flag; /* cpuwork task exit flag */
 
 	struct skeldma_desc *desc_mem;
 
@@ -57,7 +63,7 @@ struct skeldma_hw {
 	 *       |get completed     |------------------|    |
 	 *       |                                     |    |
 	 *       |                                     v    v
-	 *  -----------     cpucopy thread working     -----------
+	 *  -----------     cpuwork thread working     -----------
 	 *  |completed|<-------------------------------| running |
 	 *  -----------                                -----------
 	 */
@@ -72,7 +78,7 @@ struct skeldma_hw {
 	uint16_t last_ridx;
 	uint64_t submitted_count;
 
-	/* Cache delimiter for cpucopy thread's operation data */
+	/* Cache delimiter for cpuwork thread's operation data */
 	char cache2 __rte_cache_aligned;
 	volatile uint32_t zero_req_count;
 	uint64_t completed_count;
-- 
2.17.1


  parent reply	other threads:[~2024-01-26  9:01 UTC|newest]

Thread overview: 8+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-01-26  8:57 [PATCH 0/2] dma/skeleton: add support for SG copy and " Chengwen Feng
2024-01-26  8:57 ` [PATCH 1/2] dma/skeleton: support SG copy ops Chengwen Feng
2024-03-06 20:48   ` Thomas Monjalon
2024-03-07 10:44     ` Ferruh Yigit
2024-03-07 13:12       ` Thomas Monjalon
2024-03-07 13:15         ` Morten Brørup
2024-01-26  8:57 ` Chengwen Feng [this message]
2024-03-06 20:49 ` [PATCH 0/2] dma/skeleton: add support for SG copy and fill ops Thomas Monjalon

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20240126085726.54581-3-fengchengwen@huawei.com \
    --to=fengchengwen@huawei.com \
    --cc=dev@dpdk.org \
    --cc=gmuthukrishn@marvell.com \
    --cc=tangkunshan@huawei.com \
    --cc=thomas@monjalon.net \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).