From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id EF8D2A034C; Thu, 28 Apr 2022 17:05:12 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 8ACA742819; Thu, 28 Apr 2022 17:05:12 +0200 (CEST) Received: from mx0b-0016f401.pphosted.com (mx0b-0016f401.pphosted.com [67.231.156.173]) by mails.dpdk.org (Postfix) with ESMTP id 1003F40E50 for ; Thu, 28 Apr 2022 17:05:10 +0200 (CEST) Received: from pps.filterd (m0045851.ppops.net [127.0.0.1]) by mx0b-0016f401.pphosted.com (8.16.1.2/8.16.1.2) with ESMTP id 23SA6flY016804; Thu, 28 Apr 2022 08:05:09 -0700 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=marvell.com; h=from : to : cc : subject : date : message-id : in-reply-to : references : mime-version : content-type; s=pfpt0220; bh=TzHyO3dfrJNYfRaow9EVoXpIKlLd8cATrHnFZWmu4v0=; b=c7Sem3yhZD5GAsyhXDD0/QMNtLk7hRHW8ro3arFGvoUEDZqL2s7ekL2l8R7ZtqEGYbN0 jzwpIsX12gbnogqWeThwpf+ipwte0grBu839huWIxOmALF6Wk/oNxIQ3MB0+qLIDHStC FxJe82+QC/UCnXlEJ89k/BsyfCWstQbw52nynPEyYqBMhxRzdOHFy6lP+mKIbjuOUeIE R7saTCKvqC5wox4AhoE4RxaMhEqY10r8uzBAl5AtTsD0HVDEwa2cP/QztNEiA1KOIRl9 XmUojBku6FTuVUFeqan/3xzYDIrb9BL4+cTWi+fXM8bC4V6R2+qbdBn3wuMe1ovWgyhU jA== Received: from dc5-exch02.marvell.com ([199.233.59.182]) by mx0b-0016f401.pphosted.com (PPS) with ESMTPS id 3fqpvy1j0r-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-SHA384 bits=256 verify=NOT); Thu, 28 Apr 2022 08:05:09 -0700 Received: from DC5-EXCH01.marvell.com (10.69.176.38) by DC5-EXCH02.marvell.com (10.69.176.39) with Microsoft SMTP Server (TLS) id 15.0.1497.18; Thu, 28 Apr 2022 08:05:07 -0700 Received: from maili.marvell.com (10.69.176.80) by DC5-EXCH01.marvell.com (10.69.176.38) with Microsoft SMTP Server id 15.0.1497.2 via Frontend Transport; Thu, 28 Apr 2022 08:05:06 -0700 Received: from hyd1588t430.marvell.com (unknown [10.29.52.204]) by maili.marvell.com (Postfix) with ESMTP id EB07D5B6936; Thu, 28 Apr 2022 08:05:03 -0700 (PDT) From: Nithin Dabilpuram To: , , Radu Nicolau , Akhil Goyal CC: , , Nithin Dabilpuram Subject: [PATCH v3 1/7] examples/ipsec-secgw: move fast path helper functions Date: Thu, 28 Apr 2022 20:34:53 +0530 Message-ID: <20220428150459.23950-1-ndabilpuram@marvell.com> X-Mailer: git-send-email 2.8.4 In-Reply-To: <20220322175902.363520-1-ndabilpuram@marvell.com> References: <20220322175902.363520-1-ndabilpuram@marvell.com> MIME-Version: 1.0 Content-Type: text/plain X-Proofpoint-GUID: xd8EE0k20vQAoIZPhzi5BMatWV2qn44R X-Proofpoint-ORIG-GUID: xd8EE0k20vQAoIZPhzi5BMatWV2qn44R X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.205,Aquarius:18.0.858,Hydra:6.0.486,FMLib:17.11.64.514 definitions=2022-04-28_02,2022-04-28_01,2022-02-23_01 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Move fast path helper functions to header file for easy access. Signed-off-by: Nithin Dabilpuram --- v3: - In patch 7/7, in Inline Protocol single sa mode's worker thread, further reduce processing by getting the proto from the Single outbound SA flags. Keep processing as minimal as possible as single-sa mode is only for benchmarking drivers. v2: - Moved this patch from 4/7 to 1/7 to keep all moving as first patch without any change in function. - In patch 1/7, handled comments from Konstantin to check for capabilities before using Tx offload in case of LA and also to enable Checksum offload in case of TSO+Inline Protocol - In patch 2/7, handled comments from Konstantin to use RTE_ETH_IS* macros and - In patch 2/7, used tx_offload field and RTE_MBUF_L2_LEN_BITS shift to write to mbuf->tx_offload instead of bitfield access so that it is cleared and there is only stores and no loads. - In patch 5/7, made few fast path functions always_inline examples/ipsec-secgw/ipsec-secgw.c | 575 +--------------------------------- examples/ipsec-secgw/ipsec-secgw.h | 4 + examples/ipsec-secgw/ipsec.h | 34 ++ examples/ipsec-secgw/ipsec_process.c | 49 +-- examples/ipsec-secgw/ipsec_worker.h | 588 +++++++++++++++++++++++++++++++++++ 5 files changed, 630 insertions(+), 620 deletions(-) diff --git a/examples/ipsec-secgw/ipsec-secgw.c b/examples/ipsec-secgw/ipsec-secgw.c index 42b5081..959a20b 100644 --- a/examples/ipsec-secgw/ipsec-secgw.c +++ b/examples/ipsec-secgw/ipsec-secgw.c @@ -70,11 +70,6 @@ volatile bool force_quit; #define BURST_TX_DRAIN_US 100 /* TX drain every ~100us */ -/* Configure how many packets ahead to prefetch, when reading packets */ -#define PREFETCH_OFFSET 3 - -#define MAX_RX_QUEUE_PER_LCORE 16 - #define MAX_LCORE_PARAMS 1024 /* @@ -191,9 +186,9 @@ static uint64_t dev_tx_offload = UINT64_MAX; /* * global values that determine multi-seg policy */ -static uint32_t frag_tbl_sz; +uint32_t frag_tbl_sz; static uint32_t frame_buf_size = RTE_MBUF_DEFAULT_BUF_SIZE; -static uint32_t mtu_size = RTE_ETHER_MTU; +uint32_t mtu_size = RTE_ETHER_MTU; static uint64_t frag_ttl_ns = MAX_FRAG_TTL_NS; static uint32_t stats_interval; @@ -205,11 +200,6 @@ struct app_sa_prm app_sa_prm = { }; static const char *cfgfile; -struct lcore_rx_queue { - uint16_t port_id; - uint8_t queue_id; -} __rte_cache_aligned; - struct lcore_params { uint16_t port_id; uint8_t queue_id; @@ -224,28 +214,7 @@ static uint16_t nb_lcore_params; static struct rte_hash *cdev_map_in; static struct rte_hash *cdev_map_out; -struct buffer { - uint16_t len; - struct rte_mbuf *m_table[MAX_PKT_BURST] __rte_aligned(sizeof(void *)); -}; - -struct lcore_conf { - uint16_t nb_rx_queue; - struct lcore_rx_queue rx_queue_list[MAX_RX_QUEUE_PER_LCORE]; - uint16_t tx_queue_id[RTE_MAX_ETHPORTS]; - struct buffer tx_mbufs[RTE_MAX_ETHPORTS]; - struct ipsec_ctx inbound; - struct ipsec_ctx outbound; - struct rt_ctx *rt4_ctx; - struct rt_ctx *rt6_ctx; - struct { - struct rte_ip_frag_tbl *tbl; - struct rte_mempool *pool_indir; - struct rte_ip_frag_death_row dr; - } frag; -} __rte_cache_aligned; - -static struct lcore_conf lcore_conf[RTE_MAX_LCORE]; +struct lcore_conf lcore_conf[RTE_MAX_LCORE]; static struct rte_eth_conf port_conf = { .rxmode = { @@ -281,32 +250,6 @@ multi_seg_required(void) frame_buf_size || frag_tbl_sz != 0); } -static inline void -adjust_ipv4_pktlen(struct rte_mbuf *m, const struct rte_ipv4_hdr *iph, - uint32_t l2_len) -{ - uint32_t plen, trim; - - plen = rte_be_to_cpu_16(iph->total_length) + l2_len; - if (plen < m->pkt_len) { - trim = m->pkt_len - plen; - rte_pktmbuf_trim(m, trim); - } -} - -static inline void -adjust_ipv6_pktlen(struct rte_mbuf *m, const struct rte_ipv6_hdr *iph, - uint32_t l2_len) -{ - uint32_t plen, trim; - - plen = rte_be_to_cpu_16(iph->payload_len) + sizeof(*iph) + l2_len; - if (plen < m->pkt_len) { - trim = m->pkt_len - plen; - rte_pktmbuf_trim(m, trim); - } -} - struct ipsec_core_statistics core_statistics[RTE_MAX_LCORE]; @@ -371,369 +314,6 @@ print_stats_cb(__rte_unused void *param) rte_eal_alarm_set(stats_interval * US_PER_S, print_stats_cb, NULL); } -static inline void -prepare_one_packet(struct rte_mbuf *pkt, struct ipsec_traffic *t) -{ - const struct rte_ether_hdr *eth; - const struct rte_ipv4_hdr *iph4; - const struct rte_ipv6_hdr *iph6; - const struct rte_udp_hdr *udp; - uint16_t ip4_hdr_len; - uint16_t nat_port; - - eth = rte_pktmbuf_mtod(pkt, const struct rte_ether_hdr *); - if (eth->ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4)) { - - iph4 = (const struct rte_ipv4_hdr *)rte_pktmbuf_adj(pkt, - RTE_ETHER_HDR_LEN); - adjust_ipv4_pktlen(pkt, iph4, 0); - - switch (iph4->next_proto_id) { - case IPPROTO_ESP: - t->ipsec.pkts[(t->ipsec.num)++] = pkt; - break; - case IPPROTO_UDP: - if (app_sa_prm.udp_encap == 1) { - ip4_hdr_len = ((iph4->version_ihl & - RTE_IPV4_HDR_IHL_MASK) * - RTE_IPV4_IHL_MULTIPLIER); - udp = rte_pktmbuf_mtod_offset(pkt, - struct rte_udp_hdr *, ip4_hdr_len); - nat_port = rte_cpu_to_be_16(IPSEC_NAT_T_PORT); - if (udp->src_port == nat_port || - udp->dst_port == nat_port){ - t->ipsec.pkts[(t->ipsec.num)++] = pkt; - pkt->packet_type |= - MBUF_PTYPE_TUNNEL_ESP_IN_UDP; - break; - } - } - /* Fall through */ - default: - t->ip4.data[t->ip4.num] = &iph4->next_proto_id; - t->ip4.pkts[(t->ip4.num)++] = pkt; - } - pkt->l2_len = 0; - pkt->l3_len = sizeof(*iph4); - pkt->packet_type |= RTE_PTYPE_L3_IPV4; - if (pkt->packet_type & RTE_PTYPE_L4_TCP) - pkt->l4_len = sizeof(struct rte_tcp_hdr); - else if (pkt->packet_type & RTE_PTYPE_L4_UDP) - pkt->l4_len = sizeof(struct rte_udp_hdr); - } else if (eth->ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6)) { - int next_proto; - size_t l3len, ext_len; - uint8_t *p; - - /* get protocol type */ - iph6 = (const struct rte_ipv6_hdr *)rte_pktmbuf_adj(pkt, - RTE_ETHER_HDR_LEN); - adjust_ipv6_pktlen(pkt, iph6, 0); - - next_proto = iph6->proto; - - /* determine l3 header size up to ESP extension */ - l3len = sizeof(struct ip6_hdr); - p = rte_pktmbuf_mtod(pkt, uint8_t *); - while (next_proto != IPPROTO_ESP && l3len < pkt->data_len && - (next_proto = rte_ipv6_get_next_ext(p + l3len, - next_proto, &ext_len)) >= 0) - l3len += ext_len; - - /* drop packet when IPv6 header exceeds first segment length */ - if (unlikely(l3len > pkt->data_len)) { - free_pkts(&pkt, 1); - return; - } - - switch (next_proto) { - case IPPROTO_ESP: - t->ipsec.pkts[(t->ipsec.num)++] = pkt; - break; - case IPPROTO_UDP: - if (app_sa_prm.udp_encap == 1) { - udp = rte_pktmbuf_mtod_offset(pkt, - struct rte_udp_hdr *, l3len); - nat_port = rte_cpu_to_be_16(IPSEC_NAT_T_PORT); - if (udp->src_port == nat_port || - udp->dst_port == nat_port){ - t->ipsec.pkts[(t->ipsec.num)++] = pkt; - pkt->packet_type |= - MBUF_PTYPE_TUNNEL_ESP_IN_UDP; - break; - } - } - /* Fall through */ - default: - t->ip6.data[t->ip6.num] = &iph6->proto; - t->ip6.pkts[(t->ip6.num)++] = pkt; - } - pkt->l2_len = 0; - pkt->l3_len = l3len; - pkt->packet_type |= RTE_PTYPE_L3_IPV6; - } else { - /* Unknown/Unsupported type, drop the packet */ - RTE_LOG(ERR, IPSEC, "Unsupported packet type 0x%x\n", - rte_be_to_cpu_16(eth->ether_type)); - free_pkts(&pkt, 1); - return; - } - - /* Check if the packet has been processed inline. For inline protocol - * processed packets, the metadata in the mbuf can be used to identify - * the security processing done on the packet. The metadata will be - * used to retrieve the application registered userdata associated - * with the security session. - */ - - if (pkt->ol_flags & RTE_MBUF_F_RX_SEC_OFFLOAD && - rte_security_dynfield_is_registered()) { - struct ipsec_sa *sa; - struct ipsec_mbuf_metadata *priv; - struct rte_security_ctx *ctx = (struct rte_security_ctx *) - rte_eth_dev_get_sec_ctx( - pkt->port); - - /* Retrieve the userdata registered. Here, the userdata - * registered is the SA pointer. - */ - sa = (struct ipsec_sa *)rte_security_get_userdata(ctx, - *rte_security_dynfield(pkt)); - if (sa == NULL) { - /* userdata could not be retrieved */ - return; - } - - /* Save SA as priv member in mbuf. This will be used in the - * IPsec selector(SP-SA) check. - */ - - priv = get_priv(pkt); - priv->sa = sa; - } -} - -static inline void -prepare_traffic(struct rte_mbuf **pkts, struct ipsec_traffic *t, - uint16_t nb_pkts) -{ - int32_t i; - - t->ipsec.num = 0; - t->ip4.num = 0; - t->ip6.num = 0; - - for (i = 0; i < (nb_pkts - PREFETCH_OFFSET); i++) { - rte_prefetch0(rte_pktmbuf_mtod(pkts[i + PREFETCH_OFFSET], - void *)); - prepare_one_packet(pkts[i], t); - } - /* Process left packets */ - for (; i < nb_pkts; i++) - prepare_one_packet(pkts[i], t); -} - -static inline void -prepare_tx_pkt(struct rte_mbuf *pkt, uint16_t port, - const struct lcore_conf *qconf) -{ - struct ip *ip; - struct rte_ether_hdr *ethhdr; - - ip = rte_pktmbuf_mtod(pkt, struct ip *); - - ethhdr = (struct rte_ether_hdr *) - rte_pktmbuf_prepend(pkt, RTE_ETHER_HDR_LEN); - - if (ip->ip_v == IPVERSION) { - pkt->ol_flags |= qconf->outbound.ipv4_offloads; - pkt->l3_len = sizeof(struct ip); - pkt->l2_len = RTE_ETHER_HDR_LEN; - - ip->ip_sum = 0; - - /* calculate IPv4 cksum in SW */ - if ((pkt->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) == 0) - ip->ip_sum = rte_ipv4_cksum((struct rte_ipv4_hdr *)ip); - - ethhdr->ether_type = rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4); - } else { - pkt->ol_flags |= qconf->outbound.ipv6_offloads; - pkt->l3_len = sizeof(struct ip6_hdr); - pkt->l2_len = RTE_ETHER_HDR_LEN; - - ethhdr->ether_type = rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6); - } - - memcpy(ðhdr->src_addr, ðaddr_tbl[port].src, - sizeof(struct rte_ether_addr)); - memcpy(ðhdr->dst_addr, ðaddr_tbl[port].dst, - sizeof(struct rte_ether_addr)); -} - -static inline void -prepare_tx_burst(struct rte_mbuf *pkts[], uint16_t nb_pkts, uint16_t port, - const struct lcore_conf *qconf) -{ - int32_t i; - const int32_t prefetch_offset = 2; - - for (i = 0; i < (nb_pkts - prefetch_offset); i++) { - rte_mbuf_prefetch_part2(pkts[i + prefetch_offset]); - prepare_tx_pkt(pkts[i], port, qconf); - } - /* Process left packets */ - for (; i < nb_pkts; i++) - prepare_tx_pkt(pkts[i], port, qconf); -} - -/* Send burst of packets on an output interface */ -static inline int32_t -send_burst(struct lcore_conf *qconf, uint16_t n, uint16_t port) -{ - struct rte_mbuf **m_table; - int32_t ret; - uint16_t queueid; - - queueid = qconf->tx_queue_id[port]; - m_table = (struct rte_mbuf **)qconf->tx_mbufs[port].m_table; - - prepare_tx_burst(m_table, n, port, qconf); - - ret = rte_eth_tx_burst(port, queueid, m_table, n); - - core_stats_update_tx(ret); - - if (unlikely(ret < n)) { - do { - free_pkts(&m_table[ret], 1); - } while (++ret < n); - } - - return 0; -} - -/* - * Helper function to fragment and queue for TX one packet. - */ -static inline uint32_t -send_fragment_packet(struct lcore_conf *qconf, struct rte_mbuf *m, - uint16_t port, uint8_t proto) -{ - struct buffer *tbl; - uint32_t len, n; - int32_t rc; - - tbl = qconf->tx_mbufs + port; - len = tbl->len; - - /* free space for new fragments */ - if (len + RTE_LIBRTE_IP_FRAG_MAX_FRAG >= RTE_DIM(tbl->m_table)) { - send_burst(qconf, len, port); - len = 0; - } - - n = RTE_DIM(tbl->m_table) - len; - - if (proto == IPPROTO_IP) - rc = rte_ipv4_fragment_packet(m, tbl->m_table + len, - n, mtu_size, m->pool, qconf->frag.pool_indir); - else - rc = rte_ipv6_fragment_packet(m, tbl->m_table + len, - n, mtu_size, m->pool, qconf->frag.pool_indir); - - if (rc >= 0) - len += rc; - else - RTE_LOG(ERR, IPSEC, - "%s: failed to fragment packet with size %u, " - "error code: %d\n", - __func__, m->pkt_len, rte_errno); - - free_pkts(&m, 1); - return len; -} - -/* Enqueue a single packet, and send burst if queue is filled */ -static inline int32_t -send_single_packet(struct rte_mbuf *m, uint16_t port, uint8_t proto) -{ - uint32_t lcore_id; - uint16_t len; - struct lcore_conf *qconf; - - lcore_id = rte_lcore_id(); - - qconf = &lcore_conf[lcore_id]; - len = qconf->tx_mbufs[port].len; - - if (m->pkt_len <= mtu_size) { - qconf->tx_mbufs[port].m_table[len] = m; - len++; - - /* need to fragment the packet */ - } else if (frag_tbl_sz > 0) - len = send_fragment_packet(qconf, m, port, proto); - else - free_pkts(&m, 1); - - /* enough pkts to be sent */ - if (unlikely(len == MAX_PKT_BURST)) { - send_burst(qconf, MAX_PKT_BURST, port); - len = 0; - } - - qconf->tx_mbufs[port].len = len; - return 0; -} - -static inline void -inbound_sp_sa(struct sp_ctx *sp, struct sa_ctx *sa, struct traffic_type *ip, - uint16_t lim, struct ipsec_spd_stats *stats) -{ - struct rte_mbuf *m; - uint32_t i, j, res, sa_idx; - - if (ip->num == 0 || sp == NULL) - return; - - rte_acl_classify((struct rte_acl_ctx *)sp, ip->data, ip->res, - ip->num, DEFAULT_MAX_CATEGORIES); - - j = 0; - for (i = 0; i < ip->num; i++) { - m = ip->pkts[i]; - res = ip->res[i]; - if (res == BYPASS) { - ip->pkts[j++] = m; - stats->bypass++; - continue; - } - if (res == DISCARD) { - free_pkts(&m, 1); - stats->discard++; - continue; - } - - /* Only check SPI match for processed IPSec packets */ - if (i < lim && ((m->ol_flags & RTE_MBUF_F_RX_SEC_OFFLOAD) == 0)) { - stats->discard++; - free_pkts(&m, 1); - continue; - } - - sa_idx = res - 1; - if (!inbound_sa_check(sa, m, sa_idx)) { - stats->discard++; - free_pkts(&m, 1); - continue; - } - ip->pkts[j++] = m; - stats->protect++; - } - ip->num = j; -} - static void split46_traffic(struct ipsec_traffic *trf, struct rte_mbuf *mb[], uint32_t num) { @@ -962,140 +542,6 @@ process_pkts_outbound_nosp(struct ipsec_ctx *ipsec_ctx, } } -static inline int32_t -get_hop_for_offload_pkt(struct rte_mbuf *pkt, int is_ipv6) -{ - struct ipsec_mbuf_metadata *priv; - struct ipsec_sa *sa; - - priv = get_priv(pkt); - - sa = priv->sa; - if (unlikely(sa == NULL)) { - RTE_LOG(ERR, IPSEC, "SA not saved in private data\n"); - goto fail; - } - - if (is_ipv6) - return sa->portid; - - /* else */ - return (sa->portid | RTE_LPM_LOOKUP_SUCCESS); - -fail: - if (is_ipv6) - return -1; - - /* else */ - return 0; -} - -static inline void -route4_pkts(struct rt_ctx *rt_ctx, struct rte_mbuf *pkts[], uint8_t nb_pkts) -{ - uint32_t hop[MAX_PKT_BURST * 2]; - uint32_t dst_ip[MAX_PKT_BURST * 2]; - int32_t pkt_hop = 0; - uint16_t i, offset; - uint16_t lpm_pkts = 0; - unsigned int lcoreid = rte_lcore_id(); - - if (nb_pkts == 0) - return; - - /* Need to do an LPM lookup for non-inline packets. Inline packets will - * have port ID in the SA - */ - - for (i = 0; i < nb_pkts; i++) { - if (!(pkts[i]->ol_flags & RTE_MBUF_F_TX_SEC_OFFLOAD)) { - /* Security offload not enabled. So an LPM lookup is - * required to get the hop - */ - offset = offsetof(struct ip, ip_dst); - dst_ip[lpm_pkts] = *rte_pktmbuf_mtod_offset(pkts[i], - uint32_t *, offset); - dst_ip[lpm_pkts] = rte_be_to_cpu_32(dst_ip[lpm_pkts]); - lpm_pkts++; - } - } - - rte_lpm_lookup_bulk((struct rte_lpm *)rt_ctx, dst_ip, hop, lpm_pkts); - - lpm_pkts = 0; - - for (i = 0; i < nb_pkts; i++) { - if (pkts[i]->ol_flags & RTE_MBUF_F_TX_SEC_OFFLOAD) { - /* Read hop from the SA */ - pkt_hop = get_hop_for_offload_pkt(pkts[i], 0); - } else { - /* Need to use hop returned by lookup */ - pkt_hop = hop[lpm_pkts++]; - } - - if ((pkt_hop & RTE_LPM_LOOKUP_SUCCESS) == 0) { - core_statistics[lcoreid].lpm4.miss++; - free_pkts(&pkts[i], 1); - continue; - } - send_single_packet(pkts[i], pkt_hop & 0xff, IPPROTO_IP); - } -} - -static inline void -route6_pkts(struct rt_ctx *rt_ctx, struct rte_mbuf *pkts[], uint8_t nb_pkts) -{ - int32_t hop[MAX_PKT_BURST * 2]; - uint8_t dst_ip[MAX_PKT_BURST * 2][16]; - uint8_t *ip6_dst; - int32_t pkt_hop = 0; - uint16_t i, offset; - uint16_t lpm_pkts = 0; - unsigned int lcoreid = rte_lcore_id(); - - if (nb_pkts == 0) - return; - - /* Need to do an LPM lookup for non-inline packets. Inline packets will - * have port ID in the SA - */ - - for (i = 0; i < nb_pkts; i++) { - if (!(pkts[i]->ol_flags & RTE_MBUF_F_TX_SEC_OFFLOAD)) { - /* Security offload not enabled. So an LPM lookup is - * required to get the hop - */ - offset = offsetof(struct ip6_hdr, ip6_dst); - ip6_dst = rte_pktmbuf_mtod_offset(pkts[i], uint8_t *, - offset); - memcpy(&dst_ip[lpm_pkts][0], ip6_dst, 16); - lpm_pkts++; - } - } - - rte_lpm6_lookup_bulk_func((struct rte_lpm6 *)rt_ctx, dst_ip, hop, - lpm_pkts); - - lpm_pkts = 0; - - for (i = 0; i < nb_pkts; i++) { - if (pkts[i]->ol_flags & RTE_MBUF_F_TX_SEC_OFFLOAD) { - /* Read hop from the SA */ - pkt_hop = get_hop_for_offload_pkt(pkts[i], 1); - } else { - /* Need to use hop returned by lookup */ - pkt_hop = hop[lpm_pkts++]; - } - - if (pkt_hop == -1) { - core_statistics[lcoreid].lpm6.miss++; - free_pkts(&pkts[i], 1); - continue; - } - send_single_packet(pkts[i], pkt_hop & 0xff, IPPROTO_IPV6); - } -} - static inline void process_pkts(struct lcore_conf *qconf, struct rte_mbuf **pkts, uint8_t nb_pkts, uint16_t portid) @@ -1121,21 +567,6 @@ process_pkts(struct lcore_conf *qconf, struct rte_mbuf **pkts, } static inline void -drain_tx_buffers(struct lcore_conf *qconf) -{ - struct buffer *buf; - uint32_t portid; - - for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++) { - buf = &qconf->tx_mbufs[portid]; - if (buf->len == 0) - continue; - send_burst(qconf, buf->len, portid); - buf->len = 0; - } -} - -static inline void drain_crypto_buffers(struct lcore_conf *qconf) { uint32_t i; diff --git a/examples/ipsec-secgw/ipsec-secgw.h b/examples/ipsec-secgw/ipsec-secgw.h index 24f11ad..fceb835 100644 --- a/examples/ipsec-secgw/ipsec-secgw.h +++ b/examples/ipsec-secgw/ipsec-secgw.h @@ -6,6 +6,7 @@ #include +#define MAX_RX_QUEUE_PER_LCORE 16 #define NB_SOCKETS 4 @@ -136,6 +137,9 @@ extern uint32_t nb_bufs_in_pool; extern bool per_port_pool; +extern uint32_t mtu_size; +extern uint32_t frag_tbl_sz; + static inline uint8_t is_unprotected_port(uint16_t port_id) { diff --git a/examples/ipsec-secgw/ipsec.h b/examples/ipsec-secgw/ipsec.h index ccfde8e..9a4e7ea 100644 --- a/examples/ipsec-secgw/ipsec.h +++ b/examples/ipsec-secgw/ipsec.h @@ -9,6 +9,7 @@ #include #include +#include #include #include #include @@ -37,6 +38,11 @@ #define IP6_VERSION (6) +#define SATP_OUT_IPV4(t) \ + ((((t) & RTE_IPSEC_SATP_MODE_MASK) == RTE_IPSEC_SATP_MODE_TRANS && \ + (((t) & RTE_IPSEC_SATP_IPV_MASK) == RTE_IPSEC_SATP_IPV4)) || \ + ((t) & RTE_IPSEC_SATP_MODE_MASK) == RTE_IPSEC_SATP_MODE_TUNLV4) + struct rte_crypto_xform; struct ipsec_xform; struct rte_mbuf; @@ -260,6 +266,34 @@ struct cnt_blk { uint32_t cnt; } __rte_packed; +struct lcore_rx_queue { + uint16_t port_id; + uint8_t queue_id; +} __rte_cache_aligned; + +struct buffer { + uint16_t len; + struct rte_mbuf *m_table[MAX_PKT_BURST] __rte_aligned(sizeof(void *)); +}; + +struct lcore_conf { + uint16_t nb_rx_queue; + struct lcore_rx_queue rx_queue_list[MAX_RX_QUEUE_PER_LCORE]; + uint16_t tx_queue_id[RTE_MAX_ETHPORTS]; + struct buffer tx_mbufs[RTE_MAX_ETHPORTS]; + struct ipsec_ctx inbound; + struct ipsec_ctx outbound; + struct rt_ctx *rt4_ctx; + struct rt_ctx *rt6_ctx; + struct { + struct rte_ip_frag_tbl *tbl; + struct rte_mempool *pool_indir; + struct rte_ip_frag_death_row dr; + } frag; +} __rte_cache_aligned; + +extern struct lcore_conf lcore_conf[RTE_MAX_LCORE]; + /* Socket ctx */ extern struct socket_ctx socket_ctx[NB_SOCKETS]; diff --git a/examples/ipsec-secgw/ipsec_process.c b/examples/ipsec-secgw/ipsec_process.c index 285e9c7..089d89f 100644 --- a/examples/ipsec-secgw/ipsec_process.c +++ b/examples/ipsec-secgw/ipsec_process.c @@ -13,11 +13,7 @@ #include "ipsec.h" #include "ipsec-secgw.h" - -#define SATP_OUT_IPV4(t) \ - ((((t) & RTE_IPSEC_SATP_MODE_MASK) == RTE_IPSEC_SATP_MODE_TRANS && \ - (((t) & RTE_IPSEC_SATP_IPV_MASK) == RTE_IPSEC_SATP_IPV4)) || \ - ((t) & RTE_IPSEC_SATP_MODE_MASK) == RTE_IPSEC_SATP_MODE_TUNLV4) +#include "ipsec_worker.h" /* helper routine to free bulk of crypto-ops and related packets */ static inline void @@ -209,49 +205,6 @@ ipsec_prepare_crypto_group(struct ipsec_ctx *ctx, struct ipsec_sa *sa, } /* - * helper routine for inline and cpu(synchronous) processing - * this is just to satisfy inbound_sa_check() and get_hop_for_offload_pkt(). - * Should be removed in future. - */ -static inline void -prep_process_group(void *sa, struct rte_mbuf *mb[], uint32_t cnt) -{ - uint32_t j; - struct ipsec_mbuf_metadata *priv; - - for (j = 0; j != cnt; j++) { - priv = get_priv(mb[j]); - priv->sa = sa; - /* setup TSO related fields if TSO enabled*/ - if (priv->sa->mss) { - uint32_t ptype = mb[j]->packet_type; - /* only TCP is supported */ - if ((ptype & RTE_PTYPE_L4_MASK) == RTE_PTYPE_L4_TCP) { - mb[j]->tso_segsz = priv->sa->mss; - if ((IS_TUNNEL(priv->sa->flags))) { - mb[j]->outer_l3_len = mb[j]->l3_len; - mb[j]->outer_l2_len = mb[j]->l2_len; - mb[j]->ol_flags |= - RTE_MBUF_F_TX_TUNNEL_ESP; - if (RTE_ETH_IS_IPV4_HDR(ptype)) - mb[j]->ol_flags |= - RTE_MBUF_F_TX_OUTER_IP_CKSUM; - } - mb[j]->l4_len = sizeof(struct rte_tcp_hdr); - mb[j]->ol_flags |= (RTE_MBUF_F_TX_TCP_SEG | - RTE_MBUF_F_TX_TCP_CKSUM); - if (RTE_ETH_IS_IPV4_HDR(ptype)) - mb[j]->ol_flags |= - RTE_MBUF_F_TX_OUTER_IPV4; - else - mb[j]->ol_flags |= - RTE_MBUF_F_TX_OUTER_IPV6; - } - } - } -} - -/* * finish processing of packets successfully decrypted by an inline processor */ static uint32_t diff --git a/examples/ipsec-secgw/ipsec_worker.h b/examples/ipsec-secgw/ipsec_worker.h index 5d85cf1..e0b0a82 100644 --- a/examples/ipsec-secgw/ipsec_worker.h +++ b/examples/ipsec-secgw/ipsec_worker.h @@ -4,8 +4,15 @@ #ifndef _IPSEC_WORKER_H_ #define _IPSEC_WORKER_H_ +#include +#include +#include +#include + #include "ipsec.h" +/* Configure how many packets ahead to prefetch, when reading packets */ +#define PREFETCH_OFFSET 3 enum pkt_type { PKT_TYPE_PLAIN_IPV4 = 1, PKT_TYPE_IPSEC_IPV4, @@ -38,4 +45,585 @@ void ipsec_poll_mode_worker(void); int ipsec_launch_one_lcore(void *args); +/* + * helper routine for inline and cpu(synchronous) processing + * this is just to satisfy inbound_sa_check() and get_hop_for_offload_pkt(). + * Should be removed in future. + */ +static inline void +prep_process_group(void *sa, struct rte_mbuf *mb[], uint32_t cnt) +{ + uint32_t j; + struct ipsec_mbuf_metadata *priv; + + for (j = 0; j != cnt; j++) { + priv = get_priv(mb[j]); + priv->sa = sa; + /* setup TSO related fields if TSO enabled*/ + if (priv->sa->mss) { + uint32_t ptype = mb[j]->packet_type; + /* only TCP is supported */ + if ((ptype & RTE_PTYPE_L4_MASK) == RTE_PTYPE_L4_TCP) { + mb[j]->tso_segsz = priv->sa->mss; + if ((IS_TUNNEL(priv->sa->flags))) { + mb[j]->outer_l3_len = mb[j]->l3_len; + mb[j]->outer_l2_len = mb[j]->l2_len; + mb[j]->ol_flags |= + RTE_MBUF_F_TX_TUNNEL_ESP; + if (RTE_ETH_IS_IPV4_HDR(ptype)) + mb[j]->ol_flags |= + RTE_MBUF_F_TX_OUTER_IP_CKSUM; + } + mb[j]->l4_len = sizeof(struct rte_tcp_hdr); + mb[j]->ol_flags |= (RTE_MBUF_F_TX_TCP_SEG | + RTE_MBUF_F_TX_TCP_CKSUM); + if (RTE_ETH_IS_IPV4_HDR(ptype)) + mb[j]->ol_flags |= + RTE_MBUF_F_TX_OUTER_IPV4; + else + mb[j]->ol_flags |= + RTE_MBUF_F_TX_OUTER_IPV6; + } + } + } +} + +static inline void +adjust_ipv4_pktlen(struct rte_mbuf *m, const struct rte_ipv4_hdr *iph, + uint32_t l2_len) +{ + uint32_t plen, trim; + + plen = rte_be_to_cpu_16(iph->total_length) + l2_len; + if (plen < m->pkt_len) { + trim = m->pkt_len - plen; + rte_pktmbuf_trim(m, trim); + } +} + +static inline void +adjust_ipv6_pktlen(struct rte_mbuf *m, const struct rte_ipv6_hdr *iph, + uint32_t l2_len) +{ + uint32_t plen, trim; + + plen = rte_be_to_cpu_16(iph->payload_len) + sizeof(*iph) + l2_len; + if (plen < m->pkt_len) { + trim = m->pkt_len - plen; + rte_pktmbuf_trim(m, trim); + } +} + +static inline void +prepare_one_packet(struct rte_mbuf *pkt, struct ipsec_traffic *t) +{ + const struct rte_ether_hdr *eth; + const struct rte_ipv4_hdr *iph4; + const struct rte_ipv6_hdr *iph6; + const struct rte_udp_hdr *udp; + uint16_t ip4_hdr_len; + uint16_t nat_port; + + eth = rte_pktmbuf_mtod(pkt, const struct rte_ether_hdr *); + if (eth->ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4)) { + + iph4 = (const struct rte_ipv4_hdr *)rte_pktmbuf_adj(pkt, + RTE_ETHER_HDR_LEN); + adjust_ipv4_pktlen(pkt, iph4, 0); + + switch (iph4->next_proto_id) { + case IPPROTO_ESP: + t->ipsec.pkts[(t->ipsec.num)++] = pkt; + break; + case IPPROTO_UDP: + if (app_sa_prm.udp_encap == 1) { + ip4_hdr_len = ((iph4->version_ihl & + RTE_IPV4_HDR_IHL_MASK) * + RTE_IPV4_IHL_MULTIPLIER); + udp = rte_pktmbuf_mtod_offset(pkt, + struct rte_udp_hdr *, ip4_hdr_len); + nat_port = rte_cpu_to_be_16(IPSEC_NAT_T_PORT); + if (udp->src_port == nat_port || + udp->dst_port == nat_port){ + t->ipsec.pkts[(t->ipsec.num)++] = pkt; + pkt->packet_type |= + MBUF_PTYPE_TUNNEL_ESP_IN_UDP; + break; + } + } + /* Fall through */ + default: + t->ip4.data[t->ip4.num] = &iph4->next_proto_id; + t->ip4.pkts[(t->ip4.num)++] = pkt; + } + pkt->l2_len = 0; + pkt->l3_len = sizeof(*iph4); + pkt->packet_type |= RTE_PTYPE_L3_IPV4; + if (pkt->packet_type & RTE_PTYPE_L4_TCP) + pkt->l4_len = sizeof(struct rte_tcp_hdr); + else if (pkt->packet_type & RTE_PTYPE_L4_UDP) + pkt->l4_len = sizeof(struct rte_udp_hdr); + } else if (eth->ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6)) { + int next_proto; + size_t l3len, ext_len; + uint8_t *p; + + /* get protocol type */ + iph6 = (const struct rte_ipv6_hdr *)rte_pktmbuf_adj(pkt, + RTE_ETHER_HDR_LEN); + adjust_ipv6_pktlen(pkt, iph6, 0); + + next_proto = iph6->proto; + + /* determine l3 header size up to ESP extension */ + l3len = sizeof(struct ip6_hdr); + p = rte_pktmbuf_mtod(pkt, uint8_t *); + while (next_proto != IPPROTO_ESP && l3len < pkt->data_len && + (next_proto = rte_ipv6_get_next_ext(p + l3len, + next_proto, &ext_len)) >= 0) + l3len += ext_len; + + /* drop packet when IPv6 header exceeds first segment length */ + if (unlikely(l3len > pkt->data_len)) { + free_pkts(&pkt, 1); + return; + } + + switch (next_proto) { + case IPPROTO_ESP: + t->ipsec.pkts[(t->ipsec.num)++] = pkt; + break; + case IPPROTO_UDP: + if (app_sa_prm.udp_encap == 1) { + udp = rte_pktmbuf_mtod_offset(pkt, + struct rte_udp_hdr *, l3len); + nat_port = rte_cpu_to_be_16(IPSEC_NAT_T_PORT); + if (udp->src_port == nat_port || + udp->dst_port == nat_port){ + t->ipsec.pkts[(t->ipsec.num)++] = pkt; + pkt->packet_type |= + MBUF_PTYPE_TUNNEL_ESP_IN_UDP; + break; + } + } + /* Fall through */ + default: + t->ip6.data[t->ip6.num] = &iph6->proto; + t->ip6.pkts[(t->ip6.num)++] = pkt; + } + pkt->l2_len = 0; + pkt->l3_len = l3len; + pkt->packet_type |= RTE_PTYPE_L3_IPV6; + } else { + /* Unknown/Unsupported type, drop the packet */ + RTE_LOG(ERR, IPSEC, "Unsupported packet type 0x%x\n", + rte_be_to_cpu_16(eth->ether_type)); + free_pkts(&pkt, 1); + return; + } + + /* Check if the packet has been processed inline. For inline protocol + * processed packets, the metadata in the mbuf can be used to identify + * the security processing done on the packet. The metadata will be + * used to retrieve the application registered userdata associated + * with the security session. + */ + + if (pkt->ol_flags & RTE_MBUF_F_RX_SEC_OFFLOAD && + rte_security_dynfield_is_registered()) { + struct ipsec_sa *sa; + struct ipsec_mbuf_metadata *priv; + struct rte_security_ctx *ctx = (struct rte_security_ctx *) + rte_eth_dev_get_sec_ctx( + pkt->port); + + /* Retrieve the userdata registered. Here, the userdata + * registered is the SA pointer. + */ + sa = (struct ipsec_sa *)rte_security_get_userdata(ctx, + *rte_security_dynfield(pkt)); + if (sa == NULL) { + /* userdata could not be retrieved */ + return; + } + + /* Save SA as priv member in mbuf. This will be used in the + * IPsec selector(SP-SA) check. + */ + + priv = get_priv(pkt); + priv->sa = sa; + } +} + +static inline void +prepare_traffic(struct rte_mbuf **pkts, struct ipsec_traffic *t, + uint16_t nb_pkts) +{ + int32_t i; + + t->ipsec.num = 0; + t->ip4.num = 0; + t->ip6.num = 0; + + for (i = 0; i < (nb_pkts - PREFETCH_OFFSET); i++) { + rte_prefetch0(rte_pktmbuf_mtod(pkts[i + PREFETCH_OFFSET], + void *)); + prepare_one_packet(pkts[i], t); + } + /* Process left packets */ + for (; i < nb_pkts; i++) + prepare_one_packet(pkts[i], t); +} + +static inline void +prepare_tx_pkt(struct rte_mbuf *pkt, uint16_t port, + const struct lcore_conf *qconf) +{ + struct ip *ip; + struct rte_ether_hdr *ethhdr; + + ip = rte_pktmbuf_mtod(pkt, struct ip *); + + ethhdr = (struct rte_ether_hdr *) + rte_pktmbuf_prepend(pkt, RTE_ETHER_HDR_LEN); + + if (ip->ip_v == IPVERSION) { + pkt->ol_flags |= qconf->outbound.ipv4_offloads; + pkt->l3_len = sizeof(struct ip); + pkt->l2_len = RTE_ETHER_HDR_LEN; + + ip->ip_sum = 0; + + /* calculate IPv4 cksum in SW */ + if ((pkt->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) == 0) + ip->ip_sum = rte_ipv4_cksum((struct rte_ipv4_hdr *)ip); + + ethhdr->ether_type = rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4); + } else { + pkt->ol_flags |= qconf->outbound.ipv6_offloads; + pkt->l3_len = sizeof(struct ip6_hdr); + pkt->l2_len = RTE_ETHER_HDR_LEN; + + ethhdr->ether_type = rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6); + } + + memcpy(ðhdr->src_addr, ðaddr_tbl[port].src, + sizeof(struct rte_ether_addr)); + memcpy(ðhdr->dst_addr, ðaddr_tbl[port].dst, + sizeof(struct rte_ether_addr)); +} + +static inline void +prepare_tx_burst(struct rte_mbuf *pkts[], uint16_t nb_pkts, uint16_t port, + const struct lcore_conf *qconf) +{ + int32_t i; + const int32_t prefetch_offset = 2; + + for (i = 0; i < (nb_pkts - prefetch_offset); i++) { + rte_mbuf_prefetch_part2(pkts[i + prefetch_offset]); + prepare_tx_pkt(pkts[i], port, qconf); + } + /* Process left packets */ + for (; i < nb_pkts; i++) + prepare_tx_pkt(pkts[i], port, qconf); +} + +/* Send burst of packets on an output interface */ +static inline int32_t +send_burst(struct lcore_conf *qconf, uint16_t n, uint16_t port) +{ + struct rte_mbuf **m_table; + int32_t ret; + uint16_t queueid; + + queueid = qconf->tx_queue_id[port]; + m_table = (struct rte_mbuf **)qconf->tx_mbufs[port].m_table; + + prepare_tx_burst(m_table, n, port, qconf); + + ret = rte_eth_tx_burst(port, queueid, m_table, n); + + core_stats_update_tx(ret); + + if (unlikely(ret < n)) { + do { + free_pkts(&m_table[ret], 1); + } while (++ret < n); + } + + return 0; +} + +/* + * Helper function to fragment and queue for TX one packet. + */ +static inline uint32_t +send_fragment_packet(struct lcore_conf *qconf, struct rte_mbuf *m, + uint16_t port, uint8_t proto) +{ + struct buffer *tbl; + uint32_t len, n; + int32_t rc; + + tbl = qconf->tx_mbufs + port; + len = tbl->len; + + /* free space for new fragments */ + if (len + RTE_LIBRTE_IP_FRAG_MAX_FRAG >= RTE_DIM(tbl->m_table)) { + send_burst(qconf, len, port); + len = 0; + } + + n = RTE_DIM(tbl->m_table) - len; + + if (proto == IPPROTO_IP) + rc = rte_ipv4_fragment_packet(m, tbl->m_table + len, + n, mtu_size, m->pool, qconf->frag.pool_indir); + else + rc = rte_ipv6_fragment_packet(m, tbl->m_table + len, + n, mtu_size, m->pool, qconf->frag.pool_indir); + + if (rc >= 0) + len += rc; + else + RTE_LOG(ERR, IPSEC, + "%s: failed to fragment packet with size %u, " + "error code: %d\n", + __func__, m->pkt_len, rte_errno); + + free_pkts(&m, 1); + return len; +} + +/* Enqueue a single packet, and send burst if queue is filled */ +static inline int32_t +send_single_packet(struct rte_mbuf *m, uint16_t port, uint8_t proto) +{ + uint32_t lcore_id; + uint16_t len; + struct lcore_conf *qconf; + + lcore_id = rte_lcore_id(); + + qconf = &lcore_conf[lcore_id]; + len = qconf->tx_mbufs[port].len; + + if (m->pkt_len <= mtu_size) { + qconf->tx_mbufs[port].m_table[len] = m; + len++; + + /* need to fragment the packet */ + } else if (frag_tbl_sz > 0) + len = send_fragment_packet(qconf, m, port, proto); + else + free_pkts(&m, 1); + + /* enough pkts to be sent */ + if (unlikely(len == MAX_PKT_BURST)) { + send_burst(qconf, MAX_PKT_BURST, port); + len = 0; + } + + qconf->tx_mbufs[port].len = len; + return 0; +} + +static inline void +inbound_sp_sa(struct sp_ctx *sp, struct sa_ctx *sa, struct traffic_type *ip, + uint16_t lim, struct ipsec_spd_stats *stats) +{ + struct rte_mbuf *m; + uint32_t i, j, res, sa_idx; + + if (ip->num == 0 || sp == NULL) + return; + + rte_acl_classify((struct rte_acl_ctx *)sp, ip->data, ip->res, + ip->num, DEFAULT_MAX_CATEGORIES); + + j = 0; + for (i = 0; i < ip->num; i++) { + m = ip->pkts[i]; + res = ip->res[i]; + if (res == BYPASS) { + ip->pkts[j++] = m; + stats->bypass++; + continue; + } + if (res == DISCARD) { + free_pkts(&m, 1); + stats->discard++; + continue; + } + + /* Only check SPI match for processed IPSec packets */ + if (i < lim && ((m->ol_flags & RTE_MBUF_F_RX_SEC_OFFLOAD) == 0)) { + stats->discard++; + free_pkts(&m, 1); + continue; + } + + sa_idx = res - 1; + if (!inbound_sa_check(sa, m, sa_idx)) { + stats->discard++; + free_pkts(&m, 1); + continue; + } + ip->pkts[j++] = m; + stats->protect++; + } + ip->num = j; +} + +static inline int32_t +get_hop_for_offload_pkt(struct rte_mbuf *pkt, int is_ipv6) +{ + struct ipsec_mbuf_metadata *priv; + struct ipsec_sa *sa; + + priv = get_priv(pkt); + + sa = priv->sa; + if (unlikely(sa == NULL)) { + RTE_LOG(ERR, IPSEC, "SA not saved in private data\n"); + goto fail; + } + + if (is_ipv6) + return sa->portid; + + /* else */ + return (sa->portid | RTE_LPM_LOOKUP_SUCCESS); + +fail: + if (is_ipv6) + return -1; + + /* else */ + return 0; +} + +static inline void +route4_pkts(struct rt_ctx *rt_ctx, struct rte_mbuf *pkts[], uint8_t nb_pkts) +{ + uint32_t hop[MAX_PKT_BURST * 2]; + uint32_t dst_ip[MAX_PKT_BURST * 2]; + int32_t pkt_hop = 0; + uint16_t i, offset; + uint16_t lpm_pkts = 0; + unsigned int lcoreid = rte_lcore_id(); + + if (nb_pkts == 0) + return; + + /* Need to do an LPM lookup for non-inline packets. Inline packets will + * have port ID in the SA + */ + + for (i = 0; i < nb_pkts; i++) { + if (!(pkts[i]->ol_flags & RTE_MBUF_F_TX_SEC_OFFLOAD)) { + /* Security offload not enabled. So an LPM lookup is + * required to get the hop + */ + offset = offsetof(struct ip, ip_dst); + dst_ip[lpm_pkts] = *rte_pktmbuf_mtod_offset(pkts[i], + uint32_t *, offset); + dst_ip[lpm_pkts] = rte_be_to_cpu_32(dst_ip[lpm_pkts]); + lpm_pkts++; + } + } + + rte_lpm_lookup_bulk((struct rte_lpm *)rt_ctx, dst_ip, hop, lpm_pkts); + + lpm_pkts = 0; + + for (i = 0; i < nb_pkts; i++) { + if (pkts[i]->ol_flags & RTE_MBUF_F_TX_SEC_OFFLOAD) { + /* Read hop from the SA */ + pkt_hop = get_hop_for_offload_pkt(pkts[i], 0); + } else { + /* Need to use hop returned by lookup */ + pkt_hop = hop[lpm_pkts++]; + } + + if ((pkt_hop & RTE_LPM_LOOKUP_SUCCESS) == 0) { + core_statistics[lcoreid].lpm4.miss++; + free_pkts(&pkts[i], 1); + continue; + } + send_single_packet(pkts[i], pkt_hop & 0xff, IPPROTO_IP); + } +} + +static inline void +route6_pkts(struct rt_ctx *rt_ctx, struct rte_mbuf *pkts[], uint8_t nb_pkts) +{ + int32_t hop[MAX_PKT_BURST * 2]; + uint8_t dst_ip[MAX_PKT_BURST * 2][16]; + uint8_t *ip6_dst; + int32_t pkt_hop = 0; + uint16_t i, offset; + uint16_t lpm_pkts = 0; + unsigned int lcoreid = rte_lcore_id(); + + if (nb_pkts == 0) + return; + + /* Need to do an LPM lookup for non-inline packets. Inline packets will + * have port ID in the SA + */ + + for (i = 0; i < nb_pkts; i++) { + if (!(pkts[i]->ol_flags & RTE_MBUF_F_TX_SEC_OFFLOAD)) { + /* Security offload not enabled. So an LPM lookup is + * required to get the hop + */ + offset = offsetof(struct ip6_hdr, ip6_dst); + ip6_dst = rte_pktmbuf_mtod_offset(pkts[i], uint8_t *, + offset); + memcpy(&dst_ip[lpm_pkts][0], ip6_dst, 16); + lpm_pkts++; + } + } + + rte_lpm6_lookup_bulk_func((struct rte_lpm6 *)rt_ctx, dst_ip, hop, + lpm_pkts); + + lpm_pkts = 0; + + for (i = 0; i < nb_pkts; i++) { + if (pkts[i]->ol_flags & RTE_MBUF_F_TX_SEC_OFFLOAD) { + /* Read hop from the SA */ + pkt_hop = get_hop_for_offload_pkt(pkts[i], 1); + } else { + /* Need to use hop returned by lookup */ + pkt_hop = hop[lpm_pkts++]; + } + + if (pkt_hop == -1) { + core_statistics[lcoreid].lpm6.miss++; + free_pkts(&pkts[i], 1); + continue; + } + send_single_packet(pkts[i], pkt_hop & 0xff, IPPROTO_IPV6); + } +} + +static inline void +drain_tx_buffers(struct lcore_conf *qconf) +{ + struct buffer *buf; + uint32_t portid; + + for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++) { + buf = &qconf->tx_mbufs[portid]; + if (buf->len == 0) + continue; + send_burst(qconf, buf->len, portid); + buf->len = 0; + } +} + #endif /* _IPSEC_WORKER_H_ */ -- 2.8.4