From: Piotr Azarewicz <piotrx.t.azarewicz@intel.com>
To: dev@dpdk.org
Subject: [dpdk-dev] [PATCH v2 1/3] port: add mp/mc ring ports
Date: Thu, 24 Sep 2015 11:55:37 +0200 [thread overview]
Message-ID: <1443088539-28194-2-git-send-email-piotrx.t.azarewicz@intel.com> (raw)
In-Reply-To: <1443088539-28194-1-git-send-email-piotrx.t.azarewicz@intel.com>
ring_multi_reader input port (on top of multi consumer rte_ring)
ring_multi_writer output port (on top of multi producer rte_ring)
Signed-off-by: Piotr Azarewicz <piotrx.t.azarewicz@intel.com>
---
lib/librte_port/rte_port_ring.c | 311 +++++++++++++++++++++++++++++++++++----
lib/librte_port/rte_port_ring.h | 35 ++++-
2 files changed, 316 insertions(+), 30 deletions(-)
diff --git a/lib/librte_port/rte_port_ring.c b/lib/librte_port/rte_port_ring.c
index 9461c05..755dfc1 100644
--- a/lib/librte_port/rte_port_ring.c
+++ b/lib/librte_port/rte_port_ring.c
@@ -63,15 +63,19 @@ struct rte_port_ring_reader {
};
static void *
-rte_port_ring_reader_create(void *params, int socket_id)
+rte_port_ring_reader_create_internal(void *params, int socket_id,
+ uint32_t is_multi)
{
struct rte_port_ring_reader_params *conf =
(struct rte_port_ring_reader_params *) params;
struct rte_port_ring_reader *port;
/* Check input parameters */
- if (conf == NULL) {
- RTE_LOG(ERR, PORT, "%s: params is NULL\n", __func__);
+ if ((conf == NULL) ||
+ (conf->ring == NULL) ||
+ (conf->ring->cons.sc_dequeue && is_multi) ||
+ (!(conf->ring->cons.sc_dequeue) && !is_multi)) {
+ RTE_LOG(ERR, PORT, "%s: Invalid Parameters\n", __func__);
return NULL;
}
@@ -89,6 +93,18 @@ rte_port_ring_reader_create(void *params, int socket_id)
return port;
}
+static void *
+rte_port_ring_reader_create(void *params, int socket_id)
+{
+ return rte_port_ring_reader_create_internal(params, socket_id, 0);
+}
+
+static void *
+rte_port_ring_multi_reader_create(void *params, int socket_id)
+{
+ return rte_port_ring_reader_create_internal(params, socket_id, 1);
+}
+
static int
rte_port_ring_reader_rx(void *port, struct rte_mbuf **pkts, uint32_t n_pkts)
{
@@ -102,6 +118,19 @@ rte_port_ring_reader_rx(void *port, struct rte_mbuf **pkts, uint32_t n_pkts)
}
static int
+rte_port_ring_multi_reader_rx(void *port, struct rte_mbuf **pkts,
+ uint32_t n_pkts)
+{
+ struct rte_port_ring_reader *p = (struct rte_port_ring_reader *) port;
+ uint32_t nb_rx;
+
+ nb_rx = rte_ring_mc_dequeue_burst(p->ring, (void **) pkts, n_pkts);
+ RTE_PORT_RING_READER_STATS_PKTS_IN_ADD(p, nb_rx);
+
+ return nb_rx;
+}
+
+static int
rte_port_ring_reader_free(void *port)
{
if (port == NULL) {
@@ -155,10 +184,12 @@ struct rte_port_ring_writer {
uint32_t tx_burst_sz;
uint32_t tx_buf_count;
uint64_t bsz_mask;
+ uint32_t is_multi;
};
static void *
-rte_port_ring_writer_create(void *params, int socket_id)
+rte_port_ring_writer_create_internal(void *params, int socket_id,
+ uint32_t is_multi)
{
struct rte_port_ring_writer_params *conf =
(struct rte_port_ring_writer_params *) params;
@@ -166,7 +197,9 @@ rte_port_ring_writer_create(void *params, int socket_id)
/* Check input parameters */
if ((conf == NULL) ||
- (conf->ring == NULL) ||
+ (conf->ring == NULL) ||
+ (conf->ring->prod.sp_enqueue && is_multi) ||
+ (!(conf->ring->prod.sp_enqueue) && !is_multi) ||
(conf->tx_burst_sz > RTE_PORT_IN_BURST_SIZE_MAX)) {
RTE_LOG(ERR, PORT, "%s: Invalid Parameters\n", __func__);
return NULL;
@@ -185,10 +218,23 @@ rte_port_ring_writer_create(void *params, int socket_id)
port->tx_burst_sz = conf->tx_burst_sz;
port->tx_buf_count = 0;
port->bsz_mask = 1LLU << (conf->tx_burst_sz - 1);
+ port->is_multi = is_multi;
return port;
}
+static void *
+rte_port_ring_writer_create(void *params, int socket_id)
+{
+ return rte_port_ring_writer_create_internal(params, socket_id, 0);
+}
+
+static void *
+rte_port_ring_multi_writer_create(void *params, int socket_id)
+{
+ return rte_port_ring_writer_create_internal(params, socket_id, 1);
+}
+
static inline void
send_burst(struct rte_port_ring_writer *p)
{
@@ -204,6 +250,21 @@ send_burst(struct rte_port_ring_writer *p)
p->tx_buf_count = 0;
}
+static inline void
+send_burst_mp(struct rte_port_ring_writer *p)
+{
+ uint32_t nb_tx;
+
+ nb_tx = rte_ring_mp_enqueue_burst(p->ring, (void **)p->tx_buf,
+ p->tx_buf_count);
+
+ RTE_PORT_RING_WRITER_STATS_PKTS_DROP_ADD(p, p->tx_buf_count - nb_tx);
+ for ( ; nb_tx < p->tx_buf_count; nb_tx++)
+ rte_pktmbuf_free(p->tx_buf[nb_tx]);
+
+ p->tx_buf_count = 0;
+}
+
static int
rte_port_ring_writer_tx(void *port, struct rte_mbuf *pkt)
{
@@ -218,9 +279,23 @@ rte_port_ring_writer_tx(void *port, struct rte_mbuf *pkt)
}
static int
-rte_port_ring_writer_tx_bulk(void *port,
+rte_port_ring_multi_writer_tx(void *port, struct rte_mbuf *pkt)
+{
+ struct rte_port_ring_writer *p = (struct rte_port_ring_writer *) port;
+
+ p->tx_buf[p->tx_buf_count++] = pkt;
+ RTE_PORT_RING_WRITER_STATS_PKTS_IN_ADD(p, 1);
+ if (p->tx_buf_count >= p->tx_burst_sz)
+ send_burst_mp(p);
+
+ return 0;
+}
+
+static inline int __attribute__((always_inline))
+rte_port_ring_writer_tx_bulk_internal(void *port,
struct rte_mbuf **pkts,
- uint64_t pkts_mask)
+ uint64_t pkts_mask,
+ uint32_t is_multi)
{
struct rte_port_ring_writer *p =
(struct rte_port_ring_writer *) port;
@@ -234,11 +309,20 @@ rte_port_ring_writer_tx_bulk(void *port,
uint64_t n_pkts = __builtin_popcountll(pkts_mask);
uint32_t n_pkts_ok;
- if (tx_buf_count)
- send_burst(p);
+ if (tx_buf_count) {
+ if (is_multi)
+ send_burst_mp(p);
+ else
+ send_burst(p);
+ }
RTE_PORT_RING_WRITER_STATS_PKTS_IN_ADD(p, n_pkts);
- n_pkts_ok = rte_ring_sp_enqueue_burst(p->ring, (void **)pkts, n_pkts);
+ if (is_multi)
+ n_pkts_ok = rte_ring_mp_enqueue_burst(p->ring, (void **)pkts,
+ n_pkts);
+ else
+ n_pkts_ok = rte_ring_sp_enqueue_burst(p->ring, (void **)pkts,
+ n_pkts);
RTE_PORT_RING_WRITER_STATS_PKTS_DROP_ADD(p, n_pkts - n_pkts_ok);
for ( ; n_pkts_ok < n_pkts; n_pkts_ok++) {
@@ -258,14 +342,34 @@ rte_port_ring_writer_tx_bulk(void *port,
}
p->tx_buf_count = tx_buf_count;
- if (tx_buf_count >= p->tx_burst_sz)
- send_burst(p);
+ if (tx_buf_count >= p->tx_burst_sz) {
+ if (is_multi)
+ send_burst_mp(p);
+ else
+ send_burst(p);
+ }
}
return 0;
}
static int
+rte_port_ring_writer_tx_bulk(void *port,
+ struct rte_mbuf **pkts,
+ uint64_t pkts_mask)
+{
+ return rte_port_ring_writer_tx_bulk_internal(port, pkts, pkts_mask, 0);
+}
+
+static int
+rte_port_ring_multi_writer_tx_bulk(void *port,
+ struct rte_mbuf **pkts,
+ uint64_t pkts_mask)
+{
+ return rte_port_ring_writer_tx_bulk_internal(port, pkts, pkts_mask, 1);
+}
+
+static int
rte_port_ring_writer_flush(void *port)
{
struct rte_port_ring_writer *p = (struct rte_port_ring_writer *) port;
@@ -277,14 +381,31 @@ rte_port_ring_writer_flush(void *port)
}
static int
+rte_port_ring_multi_writer_flush(void *port)
+{
+ struct rte_port_ring_writer *p = (struct rte_port_ring_writer *) port;
+
+ if (p->tx_buf_count > 0)
+ send_burst_mp(p);
+
+ return 0;
+}
+
+static int
rte_port_ring_writer_free(void *port)
{
+ struct rte_port_ring_writer *p = (struct rte_port_ring_writer *) port;
+
if (port == NULL) {
RTE_LOG(ERR, PORT, "%s: Port is NULL\n", __func__);
return -EINVAL;
}
- rte_port_ring_writer_flush(port);
+ if (p->is_multi)
+ rte_port_ring_multi_writer_flush(port);
+ else
+ rte_port_ring_writer_flush(port);
+
rte_free(port);
return 0;
@@ -332,10 +453,12 @@ struct rte_port_ring_writer_nodrop {
uint32_t tx_buf_count;
uint64_t bsz_mask;
uint64_t n_retries;
+ uint32_t is_multi;
};
static void *
-rte_port_ring_writer_nodrop_create(void *params, int socket_id)
+rte_port_ring_writer_nodrop_create_internal(void *params, int socket_id,
+ uint32_t is_multi)
{
struct rte_port_ring_writer_nodrop_params *conf =
(struct rte_port_ring_writer_nodrop_params *) params;
@@ -343,7 +466,9 @@ rte_port_ring_writer_nodrop_create(void *params, int socket_id)
/* Check input parameters */
if ((conf == NULL) ||
- (conf->ring == NULL) ||
+ (conf->ring == NULL) ||
+ (conf->ring->prod.sp_enqueue && is_multi) ||
+ (!(conf->ring->prod.sp_enqueue) && !is_multi) ||
(conf->tx_burst_sz > RTE_PORT_IN_BURST_SIZE_MAX)) {
RTE_LOG(ERR, PORT, "%s: Invalid Parameters\n", __func__);
return NULL;
@@ -362,6 +487,7 @@ rte_port_ring_writer_nodrop_create(void *params, int socket_id)
port->tx_burst_sz = conf->tx_burst_sz;
port->tx_buf_count = 0;
port->bsz_mask = 1LLU << (conf->tx_burst_sz - 1);
+ port->is_multi = is_multi;
/*
* When n_retries is 0 it means that we should wait for every packet to
@@ -373,6 +499,18 @@ rte_port_ring_writer_nodrop_create(void *params, int socket_id)
return port;
}
+static void *
+rte_port_ring_writer_nodrop_create(void *params, int socket_id)
+{
+ return rte_port_ring_writer_nodrop_create_internal(params, socket_id, 0);
+}
+
+static void *
+rte_port_ring_multi_writer_nodrop_create(void *params, int socket_id)
+{
+ return rte_port_ring_writer_nodrop_create_internal(params, socket_id, 1);
+}
+
static inline void
send_burst_nodrop(struct rte_port_ring_writer_nodrop *p)
{
@@ -402,6 +540,35 @@ send_burst_nodrop(struct rte_port_ring_writer_nodrop *p)
p->tx_buf_count = 0;
}
+static inline void
+send_burst_mp_nodrop(struct rte_port_ring_writer_nodrop *p)
+{
+ uint32_t nb_tx = 0, i;
+
+ nb_tx = rte_ring_mp_enqueue_burst(p->ring, (void **)p->tx_buf,
+ p->tx_buf_count);
+
+ /* We sent all the packets in a first try */
+ if (nb_tx >= p->tx_buf_count)
+ return;
+
+ for (i = 0; i < p->n_retries; i++) {
+ nb_tx += rte_ring_mp_enqueue_burst(p->ring,
+ (void **) (p->tx_buf + nb_tx), p->tx_buf_count - nb_tx);
+
+ /* We sent all the packets in more than one try */
+ if (nb_tx >= p->tx_buf_count)
+ return;
+ }
+
+ /* We didn't send the packets in maximum allowed attempts */
+ RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_DROP_ADD(p, p->tx_buf_count - nb_tx);
+ for ( ; nb_tx < p->tx_buf_count; nb_tx++)
+ rte_pktmbuf_free(p->tx_buf[nb_tx]);
+
+ p->tx_buf_count = 0;
+}
+
static int
rte_port_ring_writer_nodrop_tx(void *port, struct rte_mbuf *pkt)
{
@@ -417,9 +584,24 @@ rte_port_ring_writer_nodrop_tx(void *port, struct rte_mbuf *pkt)
}
static int
-rte_port_ring_writer_nodrop_tx_bulk(void *port,
+rte_port_ring_multi_writer_nodrop_tx(void *port, struct rte_mbuf *pkt)
+{
+ struct rte_port_ring_writer_nodrop *p =
+ (struct rte_port_ring_writer_nodrop *) port;
+
+ p->tx_buf[p->tx_buf_count++] = pkt;
+ RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_IN_ADD(p, 1);
+ if (p->tx_buf_count >= p->tx_burst_sz)
+ send_burst_mp_nodrop(p);
+
+ return 0;
+}
+
+static inline int __attribute__((always_inline))
+rte_port_ring_writer_nodrop_tx_bulk_internal(void *port,
struct rte_mbuf **pkts,
- uint64_t pkts_mask)
+ uint64_t pkts_mask,
+ uint32_t is_multi)
{
struct rte_port_ring_writer_nodrop *p =
(struct rte_port_ring_writer_nodrop *) port;
@@ -433,24 +615,37 @@ rte_port_ring_writer_nodrop_tx_bulk(void *port,
uint64_t n_pkts = __builtin_popcountll(pkts_mask);
uint32_t n_pkts_ok;
- if (tx_buf_count)
- send_burst_nodrop(p);
+ if (tx_buf_count) {
+ if (is_multi)
+ send_burst_mp_nodrop(p);
+ else
+ send_burst_nodrop(p);
+ }
RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_IN_ADD(p, n_pkts);
- n_pkts_ok = rte_ring_sp_enqueue_burst(p->ring, (void **)pkts, n_pkts);
+ if (is_multi)
+ n_pkts_ok =
+ rte_ring_mp_enqueue_burst(p->ring, (void **)pkts, n_pkts);
+ else
+ n_pkts_ok =
+ rte_ring_sp_enqueue_burst(p->ring, (void **)pkts, n_pkts);
if (n_pkts_ok >= n_pkts)
return 0;
/*
- * If we didnt manage to send all packets in single burst, move
+ * If we didn't manage to send all packets in single burst, move
* remaining packets to the buffer and call send burst.
*/
for (; n_pkts_ok < n_pkts; n_pkts_ok++) {
struct rte_mbuf *pkt = pkts[n_pkts_ok];
+
p->tx_buf[p->tx_buf_count++] = pkt;
}
- send_burst_nodrop(p);
+ if (is_multi)
+ send_burst_mp_nodrop(p);
+ else
+ send_burst_nodrop(p);
} else {
for ( ; pkts_mask; ) {
uint32_t pkt_index = __builtin_ctzll(pkts_mask);
@@ -463,14 +658,36 @@ rte_port_ring_writer_nodrop_tx_bulk(void *port,
}
p->tx_buf_count = tx_buf_count;
- if (tx_buf_count >= p->tx_burst_sz)
- send_burst_nodrop(p);
+ if (tx_buf_count >= p->tx_burst_sz) {
+ if (is_multi)
+ send_burst_mp_nodrop(p);
+ else
+ send_burst_nodrop(p);
+ }
}
return 0;
}
static int
+rte_port_ring_writer_nodrop_tx_bulk(void *port,
+ struct rte_mbuf **pkts,
+ uint64_t pkts_mask)
+{
+ return
+ rte_port_ring_writer_nodrop_tx_bulk_internal(port, pkts, pkts_mask, 0);
+}
+
+static int
+rte_port_ring_multi_writer_nodrop_tx_bulk(void *port,
+ struct rte_mbuf **pkts,
+ uint64_t pkts_mask)
+{
+ return
+ rte_port_ring_writer_nodrop_tx_bulk_internal(port, pkts, pkts_mask, 1);
+}
+
+static int
rte_port_ring_writer_nodrop_flush(void *port)
{
struct rte_port_ring_writer_nodrop *p =
@@ -483,14 +700,33 @@ rte_port_ring_writer_nodrop_flush(void *port)
}
static int
+rte_port_ring_multi_writer_nodrop_flush(void *port)
+{
+ struct rte_port_ring_writer_nodrop *p =
+ (struct rte_port_ring_writer_nodrop *) port;
+
+ if (p->tx_buf_count > 0)
+ send_burst_mp_nodrop(p);
+
+ return 0;
+}
+
+static int
rte_port_ring_writer_nodrop_free(void *port)
{
+ struct rte_port_ring_writer_nodrop *p =
+ (struct rte_port_ring_writer_nodrop *) port;
+
if (port == NULL) {
RTE_LOG(ERR, PORT, "%s: Port is NULL\n", __func__);
return -EINVAL;
}
- rte_port_ring_writer_nodrop_flush(port);
+ if (p->is_multi)
+ rte_port_ring_multi_writer_nodrop_flush(port);
+ else
+ rte_port_ring_writer_nodrop_flush(port);
+
rte_free(port);
return 0;
@@ -539,3 +775,28 @@ struct rte_port_out_ops rte_port_ring_writer_nodrop_ops = {
.f_flush = rte_port_ring_writer_nodrop_flush,
.f_stats = rte_port_ring_writer_nodrop_stats_read,
};
+
+struct rte_port_in_ops rte_port_ring_multi_reader_ops = {
+ .f_create = rte_port_ring_multi_reader_create,
+ .f_free = rte_port_ring_reader_free,
+ .f_rx = rte_port_ring_multi_reader_rx,
+ .f_stats = rte_port_ring_reader_stats_read,
+};
+
+struct rte_port_out_ops rte_port_ring_multi_writer_ops = {
+ .f_create = rte_port_ring_multi_writer_create,
+ .f_free = rte_port_ring_writer_free,
+ .f_tx = rte_port_ring_multi_writer_tx,
+ .f_tx_bulk = rte_port_ring_multi_writer_tx_bulk,
+ .f_flush = rte_port_ring_multi_writer_flush,
+ .f_stats = rte_port_ring_writer_stats_read,
+};
+
+struct rte_port_out_ops rte_port_ring_multi_writer_nodrop_ops = {
+ .f_create = rte_port_ring_multi_writer_nodrop_create,
+ .f_free = rte_port_ring_writer_nodrop_free,
+ .f_tx = rte_port_ring_multi_writer_nodrop_tx,
+ .f_tx_bulk = rte_port_ring_multi_writer_nodrop_tx_bulk,
+ .f_flush = rte_port_ring_multi_writer_nodrop_flush,
+ .f_stats = rte_port_ring_writer_nodrop_stats_read,
+};
diff --git a/lib/librte_port/rte_port_ring.h b/lib/librte_port/rte_port_ring.h
index 89a219b..de377d2 100644
--- a/lib/librte_port/rte_port_ring.h
+++ b/lib/librte_port/rte_port_ring.h
@@ -42,8 +42,14 @@ extern "C" {
* @file
* RTE Port Ring
*
- * ring_reader: input port built on top of pre-initialized single consumer ring
- * ring_writer: output port built on top of pre-initialized single producer ring
+ * ring_reader:
+ * input port built on top of pre-initialized single consumer ring
+ * ring_writer:
+ * output port built on top of pre-initialized single producer ring
+ * ring_multi_reader:
+ * input port built on top of pre-initialized multi consumers ring
+ * ring_multi_writer:
+ * output port built on top of pre-initialized multi producers ring
*
***/
@@ -55,7 +61,7 @@ extern "C" {
/** ring_reader port parameters */
struct rte_port_ring_reader_params {
- /** Underlying single consumer ring that has to be pre-initialized */
+ /** Underlying consumer ring that has to be pre-initialized */
struct rte_ring *ring;
};
@@ -64,7 +70,7 @@ extern struct rte_port_in_ops rte_port_ring_reader_ops;
/** ring_writer port parameters */
struct rte_port_ring_writer_params {
- /** Underlying single producer ring that has to be pre-initialized */
+ /** Underlying producer ring that has to be pre-initialized */
struct rte_ring *ring;
/** Recommended burst size to ring. The actual burst size can be
@@ -77,7 +83,7 @@ extern struct rte_port_out_ops rte_port_ring_writer_ops;
/** ring_writer_nodrop port parameters */
struct rte_port_ring_writer_nodrop_params {
- /** Underlying single producer ring that has to be pre-initialized */
+ /** Underlying producer ring that has to be pre-initialized */
struct rte_ring *ring;
/** Recommended burst size to ring. The actual burst size can be
@@ -91,6 +97,25 @@ struct rte_port_ring_writer_nodrop_params {
/** ring_writer_nodrop port operations */
extern struct rte_port_out_ops rte_port_ring_writer_nodrop_ops;
+/** ring_multi_reader port parameters */
+#define rte_port_ring_multi_reader_params rte_port_ring_reader_params
+
+/** ring_multi_reader port operations */
+extern struct rte_port_in_ops rte_port_ring_multi_reader_ops;
+
+/** ring_multi_writer port parameters */
+#define rte_port_ring_multi_writer_params rte_port_ring_writer_params
+
+/** ring_multi_writer port operations */
+extern struct rte_port_out_ops rte_port_ring_multi_writer_ops;
+
+/** ring_multi_writer_nodrop port parameters */
+#define rte_port_ring_multi_writer_nodrop_params \
+ rte_port_ring_writer_nodrop_params
+
+/** ring_multi_writer_nodrop port operations */
+extern struct rte_port_out_ops rte_port_ring_multi_writer_nodrop_ops;
+
#ifdef __cplusplus
}
#endif
--
1.7.9.5
next prev parent reply other threads:[~2015-09-24 9:55 UTC|newest]
Thread overview: 30+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-09-15 13:06 [dpdk-dev] [PATCH v1 0/3] ip_pipeline: add MP/MC and frag/ras support to SWQs Piotr Azarewicz
2015-09-15 13:06 ` [dpdk-dev] [PATCH v1 1/3] port: add mp/mc ring ports Piotr Azarewicz
2015-09-21 22:34 ` Stephen Hemminger
2015-09-22 11:35 ` Dumitrescu, Cristian
2015-09-21 22:35 ` Stephen Hemminger
2015-09-22 11:34 ` Dumitrescu, Cristian
2015-09-22 14:23 ` Thomas Monjalon
2015-09-22 16:34 ` Stephen Hemminger
2015-09-23 13:07 ` Dumitrescu, Cristian
2015-09-21 22:37 ` Stephen Hemminger
2015-09-22 11:34 ` Dumitrescu, Cristian
2015-09-15 13:06 ` [dpdk-dev] [PATCH v1 2/3] port: fix ras " Piotr Azarewicz
2015-09-15 13:06 ` [dpdk-dev] [PATCH v1 3/3] examples/ip_pipeline: add mp/mc and frag/ras swq Piotr Azarewicz
2015-09-15 13:49 ` [dpdk-dev] [PATCH v1 0/3] ip_pipeline: add MP/MC and frag/ras support to SWQs Dumitrescu, Cristian
2015-09-24 9:55 ` [dpdk-dev] [PATCH v2 " Piotr Azarewicz
2015-09-24 9:55 ` Piotr Azarewicz [this message]
2015-10-19 15:18 ` [dpdk-dev] [PATCH v2 1/3] port: add mp/mc ring ports Thomas Monjalon
2015-09-24 9:55 ` [dpdk-dev] [PATCH v2 2/3] port: fix ras " Piotr Azarewicz
2015-10-19 15:15 ` Thomas Monjalon
2015-09-24 9:55 ` [dpdk-dev] [PATCH v2 3/3] examples/ip_pipeline: add mp/mc and frag/ras swq Piotr Azarewicz
2015-09-24 10:28 ` [dpdk-dev] [PATCH v2 0/3] ip_pipeline: add MP/MC and frag/ras support to SWQs Dumitrescu, Cristian
2015-10-20 14:36 ` [dpdk-dev] [PATCH v3 " Piotr Azarewicz
2015-10-20 14:36 ` [dpdk-dev] [PATCH v3 1/3] port: add mp/mc ring ports Piotr Azarewicz
2015-10-20 14:36 ` [dpdk-dev] [PATCH v3 2/3] port: fix ras " Piotr Azarewicz
2015-10-20 14:36 ` [dpdk-dev] [PATCH v3 3/3] examples/ip_pipeline: add mp/mc and frag/ras swq Piotr Azarewicz
2015-10-28 13:30 ` [dpdk-dev] [PATCH v4 0/3] ip_pipeline: add MP/MC and frag/ras support to SWQs Piotr Azarewicz
2015-10-28 13:30 ` [dpdk-dev] [PATCH v4 1/3] port: add mp/mc ring ports Piotr Azarewicz
2015-10-28 13:30 ` [dpdk-dev] [PATCH v4 2/3] port: fix ras/frag " Piotr Azarewicz
2015-10-28 13:30 ` [dpdk-dev] [PATCH v4 3/3] examples/ip_pipeline: add mp/mc and frag/ras swq Piotr Azarewicz
2015-11-25 22:08 ` [dpdk-dev] [PATCH v4 0/3] ip_pipeline: add MP/MC and frag/ras support to SWQs Thomas Monjalon
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1443088539-28194-2-git-send-email-piotrx.t.azarewicz@intel.com \
--to=piotrx.t.azarewicz@intel.com \
--cc=dev@dpdk.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).