From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mga01.intel.com (mga01.intel.com [192.55.52.88]) by dpdk.org (Postfix) with ESMTP id 7ED494CA6 for ; Mon, 1 Apr 2019 14:57:24 +0200 (CEST) X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga008.fm.intel.com ([10.253.24.58]) by fmsmga101.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 01 Apr 2019 05:57:24 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.60,296,1549958400"; d="scan'208";a="136564326" Received: from sivswdev08.ir.intel.com ([10.237.217.47]) by fmsmga008.fm.intel.com with ESMTP; 01 Apr 2019 05:57:22 -0700 From: Konstantin Ananyev To: dev@dpdk.org Cc: akhil.goyal@nxp.com, olivier.matz@6wind.com, Konstantin Ananyev Date: Mon, 1 Apr 2019 13:56:51 +0100 Message-Id: <20190401125656.7636-5-konstantin.ananyev@intel.com> X-Mailer: git-send-email 2.18.0 In-Reply-To: <20190401125656.7636-1-konstantin.ananyev@intel.com> References: <20190329102726.27716-1-konstantin.ananyev@intel.com> <20190401125656.7636-1-konstantin.ananyev@intel.com> Subject: [dpdk-dev] [PATCH v5 4/9] ipsec: change the way unprocessed mbufs are accounted X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 01 Apr 2019 12:57:25 -0000 As was pointed in one of previous reviews - we can avoid updating contents of mbuf array for successfully processed packets. Instead store indexes of failed packets, to move them beyond the good ones later. Signed-off-by: Konstantin Ananyev Acked-by: Akhil Goyal --- lib/librte_ipsec/sa.c | 164 +++++++++++++++++++++++------------------- 1 file changed, 92 insertions(+), 72 deletions(-) diff --git a/lib/librte_ipsec/sa.c b/lib/librte_ipsec/sa.c index 97c0f8c61..e728c6a28 100644 --- a/lib/librte_ipsec/sa.c +++ b/lib/librte_ipsec/sa.c @@ -450,14 +450,31 @@ rte_ipsec_sa_init(struct rte_ipsec_sa *sa, const struct rte_ipsec_sa_prm *prm, return sz; } +/* + * Move bad (unprocessed) mbufs beyond the good (processed) ones. + * bad_idx[] contains the indexes of bad mbufs inside the mb[]. + */ static inline void -mbuf_bulk_copy(struct rte_mbuf *dst[], struct rte_mbuf * const src[], - uint32_t num) +move_bad_mbufs(struct rte_mbuf *mb[], const uint32_t bad_idx[], uint32_t nb_mb, + uint32_t nb_bad) { - uint32_t i; + uint32_t i, j, k; + struct rte_mbuf *drb[nb_bad]; - for (i = 0; i != num; i++) - dst[i] = src[i]; + j = 0; + k = 0; + + /* copy bad ones into a temp place */ + for (i = 0; i != nb_mb; i++) { + if (j != nb_bad && i == bad_idx[j]) + drb[j++] = mb[i]; + else + mb[k++] = mb[i]; + } + + /* copy bad ones after the good ones */ + for (i = 0; i != nb_bad; i++) + mb[k + i] = drb[i]; } /* @@ -667,7 +684,7 @@ outb_tun_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -691,17 +708,17 @@ outb_tun_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], outb_pkt_xprepare(sa, sqc, &icv); lksd_none_cop_prepare(cop[k], cs, mb[i]); esp_outb_cop_prepare(cop[k], sa, iv, &icv, 0, rc); - mb[k++] = mb[i]; + k++; /* failure, put packet into the death-row */ } else { - dr[i - k] = mb[i]; + dr[i - k] = i; rte_errno = -rc; } } /* copy not prepared mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); return k; } @@ -802,7 +819,7 @@ outb_trs_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -830,17 +847,17 @@ outb_trs_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], outb_pkt_xprepare(sa, sqc, &icv); lksd_none_cop_prepare(cop[k], cs, mb[i]); esp_outb_cop_prepare(cop[k], sa, iv, &icv, l2 + l3, rc); - mb[k++] = mb[i]; + k++; /* failure, put packet into the death-row */ } else { - dr[i - k] = mb[i]; + dr[i - k] = i; rte_errno = -rc; } } /* copy not prepared mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); return k; } @@ -1019,7 +1036,7 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; struct replay_sqn *rsn; union sym_op_data icv; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -1036,10 +1053,9 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], hl, rc); } - if (rc == 0) - mb[k++] = mb[i]; - else { - dr[i - k] = mb[i]; + k += (rc == 0); + if (rc != 0) { + dr[i - k] = i; rte_errno = -rc; } } @@ -1048,7 +1064,7 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], /* copy not prepared mbufs beyond good ones */ if (k != num && k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + move_bad_mbufs(mb, dr, num, num - k); return k; } @@ -1200,7 +1216,7 @@ esp_inb_trs_single_pkt_process(struct rte_ipsec_sa *sa, struct rte_mbuf *mb, */ static inline uint16_t esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[], - struct rte_mbuf *mb[], struct rte_mbuf *dr[], uint16_t num) + uint32_t dr[], uint16_t num) { uint32_t i, k; struct replay_sqn *rsn; @@ -1210,9 +1226,9 @@ esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[], k = 0; for (i = 0; i != num; i++) { if (esn_inb_update_sqn(rsn, sa, sqn[i]) == 0) - mb[k++] = mb[i]; + k++; else - dr[i - k] = mb[i]; + dr[i - k] = i; } rsn_update_finish(sa, rsn); @@ -1226,10 +1242,10 @@ static uint16_t inb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { - uint32_t i, k; + uint32_t i, k, n; struct rte_ipsec_sa *sa; uint32_t sqn[num]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1239,23 +1255,27 @@ inb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], for (i = 0; i != num; i++) { /* good packet */ if (esp_inb_tun_single_pkt_process(sa, mb[i], sqn + k) == 0) - mb[k++] = mb[i]; + k++; /* bad packet, will drop from furhter processing */ else - dr[i - k] = mb[i]; + dr[i - k] = i; } - /* update seq # and replay winow */ - k = esp_inb_rsn_update(sa, sqn, mb, dr + i - k, k); - /* handle unprocessed mbufs */ - if (k != num) { + if (k != num && k != 0) + move_bad_mbufs(mb, dr, num, num - k); + + /* update SQN and replay winow */ + n = esp_inb_rsn_update(sa, sqn, dr, k); + + /* handle mbufs with wrong SQN */ + if (n != k && n != 0) + move_bad_mbufs(mb, dr, k, k - n); + + if (n != num) rte_errno = EBADMSG; - if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); - } - return k; + return n; } /* @@ -1265,10 +1285,10 @@ static uint16_t inb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { - uint32_t i, k; + uint32_t i, k, n; uint32_t sqn[num]; struct rte_ipsec_sa *sa; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1278,23 +1298,27 @@ inb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], for (i = 0; i != num; i++) { /* good packet */ if (esp_inb_trs_single_pkt_process(sa, mb[i], sqn + k) == 0) - mb[k++] = mb[i]; + k++; /* bad packet, will drop from furhter processing */ else - dr[i - k] = mb[i]; + dr[i - k] = i; } - /* update seq # and replay winow */ - k = esp_inb_rsn_update(sa, sqn, mb, dr + i - k, k); - /* handle unprocessed mbufs */ - if (k != num) { + if (k != num && k != 0) + move_bad_mbufs(mb, dr, num, num - k); + + /* update SQN and replay winow */ + n = esp_inb_rsn_update(sa, sqn, dr, k); + + /* handle mbufs with wrong SQN */ + if (n != k && n != 0) + move_bad_mbufs(mb, dr, k, k - n); + + if (n != num) rte_errno = EBADMSG; - if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); - } - return k; + return n; } /* @@ -1310,7 +1334,7 @@ outb_sqh_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint32_t i, k, icv_len, *icv; struct rte_mbuf *ml; struct rte_ipsec_sa *sa; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1323,16 +1347,16 @@ outb_sqh_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], icv = rte_pktmbuf_mtod_offset(ml, void *, ml->data_len - icv_len); remove_sqh(icv, icv_len); - mb[k++] = mb[i]; + k++; } else - dr[i - k] = mb[i]; + dr[i - k] = i; } /* handle unprocessed mbufs */ if (k != num) { rte_errno = EBADMSG; if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + move_bad_mbufs(mb, dr, num, num - k); } return k; @@ -1352,23 +1376,23 @@ pkt_flag_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { uint32_t i, k; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; RTE_SET_USED(ss); k = 0; for (i = 0; i != num; i++) { if ((mb[i]->ol_flags & PKT_RX_SEC_OFFLOAD_FAILED) == 0) - mb[k++] = mb[i]; + k++; else - dr[i - k] = mb[i]; + dr[i - k] = i; } /* handle unprocessed mbufs */ if (k != num) { rte_errno = EBADMSG; if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + move_bad_mbufs(mb, dr, num, num - k); } return k; @@ -1409,7 +1433,7 @@ inline_outb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_ipsec_sa *sa; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1427,22 +1451,20 @@ inline_outb_tun_pkt_process(const struct rte_ipsec_session *ss, /* try to update the packet itself */ rc = esp_outb_tun_pkt_prepare(sa, sqc, iv, mb[i], &icv); - /* success, update mbuf fields */ - if (rc >= 0) - mb[k++] = mb[i]; + k += (rc >= 0); + /* failure, put packet into the death-row */ - else { - dr[i - k] = mb[i]; + if (rc < 0) { + dr[i - k] = i; rte_errno = -rc; } } - inline_outb_mbuf_prepare(ss, mb, k); - /* copy not processed mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); + inline_outb_mbuf_prepare(ss, mb, k); return k; } @@ -1461,7 +1483,7 @@ inline_outb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_ipsec_sa *sa; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1483,22 +1505,20 @@ inline_outb_trs_pkt_process(const struct rte_ipsec_session *ss, rc = esp_outb_trs_pkt_prepare(sa, sqc, iv, mb[i], l2, l3, &icv); - /* success, update mbuf fields */ - if (rc >= 0) - mb[k++] = mb[i]; + k += (rc >= 0); + /* failure, put packet into the death-row */ - else { - dr[i - k] = mb[i]; + if (rc < 0) { + dr[i - k] = i; rte_errno = -rc; } } - inline_outb_mbuf_prepare(ss, mb, k); - /* copy not processed mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); + inline_outb_mbuf_prepare(ss, mb, k); return k; } -- 2.17.1 From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from dpdk.org (dpdk.org [92.243.14.124]) by dpdk.space (Postfix) with ESMTP id 594F5A0679 for ; Mon, 1 Apr 2019 14:57:54 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id F31344CC5; Mon, 1 Apr 2019 14:57:27 +0200 (CEST) Received: from mga01.intel.com (mga01.intel.com [192.55.52.88]) by dpdk.org (Postfix) with ESMTP id 7ED494CA6 for ; Mon, 1 Apr 2019 14:57:24 +0200 (CEST) X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga008.fm.intel.com ([10.253.24.58]) by fmsmga101.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 01 Apr 2019 05:57:24 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.60,296,1549958400"; d="scan'208";a="136564326" Received: from sivswdev08.ir.intel.com ([10.237.217.47]) by fmsmga008.fm.intel.com with ESMTP; 01 Apr 2019 05:57:22 -0700 From: Konstantin Ananyev To: dev@dpdk.org Cc: akhil.goyal@nxp.com, olivier.matz@6wind.com, Konstantin Ananyev Date: Mon, 1 Apr 2019 13:56:51 +0100 Message-Id: <20190401125656.7636-5-konstantin.ananyev@intel.com> X-Mailer: git-send-email 2.18.0 In-Reply-To: <20190401125656.7636-1-konstantin.ananyev@intel.com> References: <20190329102726.27716-1-konstantin.ananyev@intel.com> <20190401125656.7636-1-konstantin.ananyev@intel.com> Subject: [dpdk-dev] [PATCH v5 4/9] ipsec: change the way unprocessed mbufs are accounted X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Content-Type: text/plain; charset="UTF-8" Message-ID: <20190401125651.c5693d3VlPlNmv5ZaYC4HhpM4BgftuLPz0lCPWGyKT8@z> As was pointed in one of previous reviews - we can avoid updating contents of mbuf array for successfully processed packets. Instead store indexes of failed packets, to move them beyond the good ones later. Signed-off-by: Konstantin Ananyev Acked-by: Akhil Goyal --- lib/librte_ipsec/sa.c | 164 +++++++++++++++++++++++------------------- 1 file changed, 92 insertions(+), 72 deletions(-) diff --git a/lib/librte_ipsec/sa.c b/lib/librte_ipsec/sa.c index 97c0f8c61..e728c6a28 100644 --- a/lib/librte_ipsec/sa.c +++ b/lib/librte_ipsec/sa.c @@ -450,14 +450,31 @@ rte_ipsec_sa_init(struct rte_ipsec_sa *sa, const struct rte_ipsec_sa_prm *prm, return sz; } +/* + * Move bad (unprocessed) mbufs beyond the good (processed) ones. + * bad_idx[] contains the indexes of bad mbufs inside the mb[]. + */ static inline void -mbuf_bulk_copy(struct rte_mbuf *dst[], struct rte_mbuf * const src[], - uint32_t num) +move_bad_mbufs(struct rte_mbuf *mb[], const uint32_t bad_idx[], uint32_t nb_mb, + uint32_t nb_bad) { - uint32_t i; + uint32_t i, j, k; + struct rte_mbuf *drb[nb_bad]; - for (i = 0; i != num; i++) - dst[i] = src[i]; + j = 0; + k = 0; + + /* copy bad ones into a temp place */ + for (i = 0; i != nb_mb; i++) { + if (j != nb_bad && i == bad_idx[j]) + drb[j++] = mb[i]; + else + mb[k++] = mb[i]; + } + + /* copy bad ones after the good ones */ + for (i = 0; i != nb_bad; i++) + mb[k + i] = drb[i]; } /* @@ -667,7 +684,7 @@ outb_tun_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -691,17 +708,17 @@ outb_tun_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], outb_pkt_xprepare(sa, sqc, &icv); lksd_none_cop_prepare(cop[k], cs, mb[i]); esp_outb_cop_prepare(cop[k], sa, iv, &icv, 0, rc); - mb[k++] = mb[i]; + k++; /* failure, put packet into the death-row */ } else { - dr[i - k] = mb[i]; + dr[i - k] = i; rte_errno = -rc; } } /* copy not prepared mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); return k; } @@ -802,7 +819,7 @@ outb_trs_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -830,17 +847,17 @@ outb_trs_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], outb_pkt_xprepare(sa, sqc, &icv); lksd_none_cop_prepare(cop[k], cs, mb[i]); esp_outb_cop_prepare(cop[k], sa, iv, &icv, l2 + l3, rc); - mb[k++] = mb[i]; + k++; /* failure, put packet into the death-row */ } else { - dr[i - k] = mb[i]; + dr[i - k] = i; rte_errno = -rc; } } /* copy not prepared mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); return k; } @@ -1019,7 +1036,7 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; struct replay_sqn *rsn; union sym_op_data icv; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -1036,10 +1053,9 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], hl, rc); } - if (rc == 0) - mb[k++] = mb[i]; - else { - dr[i - k] = mb[i]; + k += (rc == 0); + if (rc != 0) { + dr[i - k] = i; rte_errno = -rc; } } @@ -1048,7 +1064,7 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], /* copy not prepared mbufs beyond good ones */ if (k != num && k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + move_bad_mbufs(mb, dr, num, num - k); return k; } @@ -1200,7 +1216,7 @@ esp_inb_trs_single_pkt_process(struct rte_ipsec_sa *sa, struct rte_mbuf *mb, */ static inline uint16_t esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[], - struct rte_mbuf *mb[], struct rte_mbuf *dr[], uint16_t num) + uint32_t dr[], uint16_t num) { uint32_t i, k; struct replay_sqn *rsn; @@ -1210,9 +1226,9 @@ esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[], k = 0; for (i = 0; i != num; i++) { if (esn_inb_update_sqn(rsn, sa, sqn[i]) == 0) - mb[k++] = mb[i]; + k++; else - dr[i - k] = mb[i]; + dr[i - k] = i; } rsn_update_finish(sa, rsn); @@ -1226,10 +1242,10 @@ static uint16_t inb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { - uint32_t i, k; + uint32_t i, k, n; struct rte_ipsec_sa *sa; uint32_t sqn[num]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1239,23 +1255,27 @@ inb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], for (i = 0; i != num; i++) { /* good packet */ if (esp_inb_tun_single_pkt_process(sa, mb[i], sqn + k) == 0) - mb[k++] = mb[i]; + k++; /* bad packet, will drop from furhter processing */ else - dr[i - k] = mb[i]; + dr[i - k] = i; } - /* update seq # and replay winow */ - k = esp_inb_rsn_update(sa, sqn, mb, dr + i - k, k); - /* handle unprocessed mbufs */ - if (k != num) { + if (k != num && k != 0) + move_bad_mbufs(mb, dr, num, num - k); + + /* update SQN and replay winow */ + n = esp_inb_rsn_update(sa, sqn, dr, k); + + /* handle mbufs with wrong SQN */ + if (n != k && n != 0) + move_bad_mbufs(mb, dr, k, k - n); + + if (n != num) rte_errno = EBADMSG; - if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); - } - return k; + return n; } /* @@ -1265,10 +1285,10 @@ static uint16_t inb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { - uint32_t i, k; + uint32_t i, k, n; uint32_t sqn[num]; struct rte_ipsec_sa *sa; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1278,23 +1298,27 @@ inb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], for (i = 0; i != num; i++) { /* good packet */ if (esp_inb_trs_single_pkt_process(sa, mb[i], sqn + k) == 0) - mb[k++] = mb[i]; + k++; /* bad packet, will drop from furhter processing */ else - dr[i - k] = mb[i]; + dr[i - k] = i; } - /* update seq # and replay winow */ - k = esp_inb_rsn_update(sa, sqn, mb, dr + i - k, k); - /* handle unprocessed mbufs */ - if (k != num) { + if (k != num && k != 0) + move_bad_mbufs(mb, dr, num, num - k); + + /* update SQN and replay winow */ + n = esp_inb_rsn_update(sa, sqn, dr, k); + + /* handle mbufs with wrong SQN */ + if (n != k && n != 0) + move_bad_mbufs(mb, dr, k, k - n); + + if (n != num) rte_errno = EBADMSG; - if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); - } - return k; + return n; } /* @@ -1310,7 +1334,7 @@ outb_sqh_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint32_t i, k, icv_len, *icv; struct rte_mbuf *ml; struct rte_ipsec_sa *sa; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1323,16 +1347,16 @@ outb_sqh_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], icv = rte_pktmbuf_mtod_offset(ml, void *, ml->data_len - icv_len); remove_sqh(icv, icv_len); - mb[k++] = mb[i]; + k++; } else - dr[i - k] = mb[i]; + dr[i - k] = i; } /* handle unprocessed mbufs */ if (k != num) { rte_errno = EBADMSG; if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + move_bad_mbufs(mb, dr, num, num - k); } return k; @@ -1352,23 +1376,23 @@ pkt_flag_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { uint32_t i, k; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; RTE_SET_USED(ss); k = 0; for (i = 0; i != num; i++) { if ((mb[i]->ol_flags & PKT_RX_SEC_OFFLOAD_FAILED) == 0) - mb[k++] = mb[i]; + k++; else - dr[i - k] = mb[i]; + dr[i - k] = i; } /* handle unprocessed mbufs */ if (k != num) { rte_errno = EBADMSG; if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + move_bad_mbufs(mb, dr, num, num - k); } return k; @@ -1409,7 +1433,7 @@ inline_outb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_ipsec_sa *sa; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1427,22 +1451,20 @@ inline_outb_tun_pkt_process(const struct rte_ipsec_session *ss, /* try to update the packet itself */ rc = esp_outb_tun_pkt_prepare(sa, sqc, iv, mb[i], &icv); - /* success, update mbuf fields */ - if (rc >= 0) - mb[k++] = mb[i]; + k += (rc >= 0); + /* failure, put packet into the death-row */ - else { - dr[i - k] = mb[i]; + if (rc < 0) { + dr[i - k] = i; rte_errno = -rc; } } - inline_outb_mbuf_prepare(ss, mb, k); - /* copy not processed mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); + inline_outb_mbuf_prepare(ss, mb, k); return k; } @@ -1461,7 +1483,7 @@ inline_outb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_ipsec_sa *sa; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1483,22 +1505,20 @@ inline_outb_trs_pkt_process(const struct rte_ipsec_session *ss, rc = esp_outb_trs_pkt_prepare(sa, sqc, iv, mb[i], l2, l3, &icv); - /* success, update mbuf fields */ - if (rc >= 0) - mb[k++] = mb[i]; + k += (rc >= 0); + /* failure, put packet into the death-row */ - else { - dr[i - k] = mb[i]; + if (rc < 0) { + dr[i - k] = i; rte_errno = -rc; } } - inline_outb_mbuf_prepare(ss, mb, k); - /* copy not processed mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); + inline_outb_mbuf_prepare(ss, mb, k); return k; } -- 2.17.1