From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from dpdk.org (dpdk.org [92.243.14.124]) by dpdk.space (Postfix) with ESMTP id 2555BA05D3 for ; Fri, 29 Mar 2019 11:28:17 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id E8F664CBD; Fri, 29 Mar 2019 11:28:08 +0100 (CET) Received: from mga02.intel.com (mga02.intel.com [134.134.136.20]) by dpdk.org (Postfix) with ESMTP id 47FDB4CA6 for ; Fri, 29 Mar 2019 11:28:04 +0100 (CET) X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga006.jf.intel.com ([10.7.209.51]) by orsmga101.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 29 Mar 2019 03:28:03 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.60,284,1549958400"; d="scan'208";a="131234678" Received: from sivswdev08.ir.intel.com ([10.237.217.47]) by orsmga006.jf.intel.com with ESMTP; 29 Mar 2019 03:28:01 -0700 From: Konstantin Ananyev To: dev@dpdk.org Cc: akhil.goyal@nxp.com, olivier.matz@6wind.com, Konstantin Ananyev Date: Fri, 29 Mar 2019 10:27:21 +0000 Message-Id: <20190329102726.27716-5-konstantin.ananyev@intel.com> X-Mailer: git-send-email 2.18.0 In-Reply-To: <20190329102726.27716-1-konstantin.ananyev@intel.com> References: <20190326154320.29913-1-konstantin.ananyev@intel.com> <20190329102726.27716-1-konstantin.ananyev@intel.com> Subject: [dpdk-dev] [PATCH v4 4/9] ipsec: change the way unprocessed mbufs are accounted X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Content-Type: text/plain; charset="UTF-8" Message-ID: <20190329102721.sHlyGraBwfM-GAUxOqrm0hqot3m0EK-GJW597vfCjtc@z> As was pointed in one of previous reviews - we can avoid updating contents of mbuf array for successfully processed packets. Instead store indexes of failed packets, to move them beyond the good ones later. Signed-off-by: Konstantin Ananyev Acked-by: Akhil Goyal --- lib/librte_ipsec/sa.c | 164 +++++++++++++++++++++++------------------- 1 file changed, 92 insertions(+), 72 deletions(-) diff --git a/lib/librte_ipsec/sa.c b/lib/librte_ipsec/sa.c index 97c0f8c61..e728c6a28 100644 --- a/lib/librte_ipsec/sa.c +++ b/lib/librte_ipsec/sa.c @@ -450,14 +450,31 @@ rte_ipsec_sa_init(struct rte_ipsec_sa *sa, const struct rte_ipsec_sa_prm *prm, return sz; } +/* + * Move bad (unprocessed) mbufs beyond the good (processed) ones. + * bad_idx[] contains the indexes of bad mbufs inside the mb[]. + */ static inline void -mbuf_bulk_copy(struct rte_mbuf *dst[], struct rte_mbuf * const src[], - uint32_t num) +move_bad_mbufs(struct rte_mbuf *mb[], const uint32_t bad_idx[], uint32_t nb_mb, + uint32_t nb_bad) { - uint32_t i; + uint32_t i, j, k; + struct rte_mbuf *drb[nb_bad]; - for (i = 0; i != num; i++) - dst[i] = src[i]; + j = 0; + k = 0; + + /* copy bad ones into a temp place */ + for (i = 0; i != nb_mb; i++) { + if (j != nb_bad && i == bad_idx[j]) + drb[j++] = mb[i]; + else + mb[k++] = mb[i]; + } + + /* copy bad ones after the good ones */ + for (i = 0; i != nb_bad; i++) + mb[k + i] = drb[i]; } /* @@ -667,7 +684,7 @@ outb_tun_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -691,17 +708,17 @@ outb_tun_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], outb_pkt_xprepare(sa, sqc, &icv); lksd_none_cop_prepare(cop[k], cs, mb[i]); esp_outb_cop_prepare(cop[k], sa, iv, &icv, 0, rc); - mb[k++] = mb[i]; + k++; /* failure, put packet into the death-row */ } else { - dr[i - k] = mb[i]; + dr[i - k] = i; rte_errno = -rc; } } /* copy not prepared mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); return k; } @@ -802,7 +819,7 @@ outb_trs_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -830,17 +847,17 @@ outb_trs_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], outb_pkt_xprepare(sa, sqc, &icv); lksd_none_cop_prepare(cop[k], cs, mb[i]); esp_outb_cop_prepare(cop[k], sa, iv, &icv, l2 + l3, rc); - mb[k++] = mb[i]; + k++; /* failure, put packet into the death-row */ } else { - dr[i - k] = mb[i]; + dr[i - k] = i; rte_errno = -rc; } } /* copy not prepared mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); return k; } @@ -1019,7 +1036,7 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; struct replay_sqn *rsn; union sym_op_data icv; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -1036,10 +1053,9 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], hl, rc); } - if (rc == 0) - mb[k++] = mb[i]; - else { - dr[i - k] = mb[i]; + k += (rc == 0); + if (rc != 0) { + dr[i - k] = i; rte_errno = -rc; } } @@ -1048,7 +1064,7 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], /* copy not prepared mbufs beyond good ones */ if (k != num && k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + move_bad_mbufs(mb, dr, num, num - k); return k; } @@ -1200,7 +1216,7 @@ esp_inb_trs_single_pkt_process(struct rte_ipsec_sa *sa, struct rte_mbuf *mb, */ static inline uint16_t esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[], - struct rte_mbuf *mb[], struct rte_mbuf *dr[], uint16_t num) + uint32_t dr[], uint16_t num) { uint32_t i, k; struct replay_sqn *rsn; @@ -1210,9 +1226,9 @@ esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[], k = 0; for (i = 0; i != num; i++) { if (esn_inb_update_sqn(rsn, sa, sqn[i]) == 0) - mb[k++] = mb[i]; + k++; else - dr[i - k] = mb[i]; + dr[i - k] = i; } rsn_update_finish(sa, rsn); @@ -1226,10 +1242,10 @@ static uint16_t inb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { - uint32_t i, k; + uint32_t i, k, n; struct rte_ipsec_sa *sa; uint32_t sqn[num]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1239,23 +1255,27 @@ inb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], for (i = 0; i != num; i++) { /* good packet */ if (esp_inb_tun_single_pkt_process(sa, mb[i], sqn + k) == 0) - mb[k++] = mb[i]; + k++; /* bad packet, will drop from furhter processing */ else - dr[i - k] = mb[i]; + dr[i - k] = i; } - /* update seq # and replay winow */ - k = esp_inb_rsn_update(sa, sqn, mb, dr + i - k, k); - /* handle unprocessed mbufs */ - if (k != num) { + if (k != num && k != 0) + move_bad_mbufs(mb, dr, num, num - k); + + /* update SQN and replay winow */ + n = esp_inb_rsn_update(sa, sqn, dr, k); + + /* handle mbufs with wrong SQN */ + if (n != k && n != 0) + move_bad_mbufs(mb, dr, k, k - n); + + if (n != num) rte_errno = EBADMSG; - if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); - } - return k; + return n; } /* @@ -1265,10 +1285,10 @@ static uint16_t inb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { - uint32_t i, k; + uint32_t i, k, n; uint32_t sqn[num]; struct rte_ipsec_sa *sa; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1278,23 +1298,27 @@ inb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], for (i = 0; i != num; i++) { /* good packet */ if (esp_inb_trs_single_pkt_process(sa, mb[i], sqn + k) == 0) - mb[k++] = mb[i]; + k++; /* bad packet, will drop from furhter processing */ else - dr[i - k] = mb[i]; + dr[i - k] = i; } - /* update seq # and replay winow */ - k = esp_inb_rsn_update(sa, sqn, mb, dr + i - k, k); - /* handle unprocessed mbufs */ - if (k != num) { + if (k != num && k != 0) + move_bad_mbufs(mb, dr, num, num - k); + + /* update SQN and replay winow */ + n = esp_inb_rsn_update(sa, sqn, dr, k); + + /* handle mbufs with wrong SQN */ + if (n != k && n != 0) + move_bad_mbufs(mb, dr, k, k - n); + + if (n != num) rte_errno = EBADMSG; - if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); - } - return k; + return n; } /* @@ -1310,7 +1334,7 @@ outb_sqh_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint32_t i, k, icv_len, *icv; struct rte_mbuf *ml; struct rte_ipsec_sa *sa; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1323,16 +1347,16 @@ outb_sqh_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], icv = rte_pktmbuf_mtod_offset(ml, void *, ml->data_len - icv_len); remove_sqh(icv, icv_len); - mb[k++] = mb[i]; + k++; } else - dr[i - k] = mb[i]; + dr[i - k] = i; } /* handle unprocessed mbufs */ if (k != num) { rte_errno = EBADMSG; if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + move_bad_mbufs(mb, dr, num, num - k); } return k; @@ -1352,23 +1376,23 @@ pkt_flag_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { uint32_t i, k; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; RTE_SET_USED(ss); k = 0; for (i = 0; i != num; i++) { if ((mb[i]->ol_flags & PKT_RX_SEC_OFFLOAD_FAILED) == 0) - mb[k++] = mb[i]; + k++; else - dr[i - k] = mb[i]; + dr[i - k] = i; } /* handle unprocessed mbufs */ if (k != num) { rte_errno = EBADMSG; if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + move_bad_mbufs(mb, dr, num, num - k); } return k; @@ -1409,7 +1433,7 @@ inline_outb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_ipsec_sa *sa; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1427,22 +1451,20 @@ inline_outb_tun_pkt_process(const struct rte_ipsec_session *ss, /* try to update the packet itself */ rc = esp_outb_tun_pkt_prepare(sa, sqc, iv, mb[i], &icv); - /* success, update mbuf fields */ - if (rc >= 0) - mb[k++] = mb[i]; + k += (rc >= 0); + /* failure, put packet into the death-row */ - else { - dr[i - k] = mb[i]; + if (rc < 0) { + dr[i - k] = i; rte_errno = -rc; } } - inline_outb_mbuf_prepare(ss, mb, k); - /* copy not processed mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); + inline_outb_mbuf_prepare(ss, mb, k); return k; } @@ -1461,7 +1483,7 @@ inline_outb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_ipsec_sa *sa; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1483,22 +1505,20 @@ inline_outb_trs_pkt_process(const struct rte_ipsec_session *ss, rc = esp_outb_trs_pkt_prepare(sa, sqc, iv, mb[i], l2, l3, &icv); - /* success, update mbuf fields */ - if (rc >= 0) - mb[k++] = mb[i]; + k += (rc >= 0); + /* failure, put packet into the death-row */ - else { - dr[i - k] = mb[i]; + if (rc < 0) { + dr[i - k] = i; rte_errno = -rc; } } - inline_outb_mbuf_prepare(ss, mb, k); - /* copy not processed mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + move_bad_mbufs(mb, dr, n, n - k); + inline_outb_mbuf_prepare(ss, mb, k); return k; } -- 2.17.1