DPDK patches and discussions
 help / color / mirror / Atom feed
From: Bingbin Chen <chen.bingbin@zte.com.cn>
To: stephen@networkplumber.org, wang.junlong1@zte.com.cn,
	yang.yonggang@zte.com.cn
Cc: dev@dpdk.org, Bingbin Chen <chen.bingbin@zte.com.cn>
Subject: [PATCH v5 14/14] net/zxdh: fix debugging errors
Date: Wed, 19 Mar 2025 16:58:08 +0800	[thread overview]
Message-ID: <20250319085808.1912523-15-chen.bingbin@zte.com.cn> (raw)
In-Reply-To: <20250319085808.1912523-1-chen.bingbin@zte.com.cn>


[-- Attachment #1.1.1: Type: text/plain, Size: 97798 bytes --]

Fix zxdh driver packet sending and receiving errors.

Signed-off-by: Bingbin Chen <chen.bingbin@zte.com.cn>
---
 drivers/net/zxdh/zxdh_common.h     |  99 ++++++++
 drivers/net/zxdh/zxdh_ethdev.c     |  19 +-
 drivers/net/zxdh/zxdh_ethdev_ops.c | 109 +++++----
 drivers/net/zxdh/zxdh_ethdev_ops.h |  35 +++
 drivers/net/zxdh/zxdh_msg.c        | 381 +++++++++++++++++------------
 drivers/net/zxdh/zxdh_msg.h        |  99 +++++---
 drivers/net/zxdh/zxdh_mtr.c        |  50 ++--
 drivers/net/zxdh/zxdh_mtr.h        |   5 +
 drivers/net/zxdh/zxdh_np.c         |  70 +++++-
 drivers/net/zxdh/zxdh_np.h         |   8 +-
 drivers/net/zxdh/zxdh_pci.c        |  24 +-
 drivers/net/zxdh/zxdh_pci.h        |   2 +-
 drivers/net/zxdh/zxdh_rxtx.c       |  12 +-
 drivers/net/zxdh/zxdh_tables.c     |  42 ++--
 drivers/net/zxdh/zxdh_tables.h     |  68 +++--
 15 files changed, 708 insertions(+), 315 deletions(-)

diff --git a/drivers/net/zxdh/zxdh_common.h b/drivers/net/zxdh/zxdh_common.h
index d78a822ebf..b34dcddd4d 100644
--- a/drivers/net/zxdh/zxdh_common.h
+++ b/drivers/net/zxdh/zxdh_common.h
@@ -20,6 +20,105 @@ struct zxdh_res_para {
 	uint16_t src_type; /* refer to BAR_DRIVER_TYPE */
 };
 
+static inline size_t
+zxdh_get_value(uint32_t fld_sz, uint8_t *addr) {
+	size_t result = 0;
+	switch (fld_sz) {
+	case 1:
+		result = *((uint8_t *)addr);
+		break;
+	case 2:
+		result = *((uint16_t *)addr);
+		break;
+	case 4:
+		result = *((uint32_t *)addr);
+		break;
+	case 8:
+		result = *((uint64_t *)addr);
+		break;
+	default:
+		printf("Error: unreachable field size %u\n", fld_sz);
+		break;
+	}
+	return result;
+}
+
+static inline void
+zxdh_set_value(uint32_t fld_sz, uint8_t *addr, size_t value) {
+	switch (fld_sz) {
+	case 1:
+		*(uint8_t *)addr = (uint8_t)value;
+		break;
+	case 2:
+		*(uint16_t *)addr = (uint16_t)value;
+		break;
+	case 4:
+		*(uint32_t *)addr = (uint32_t)value;
+		break;
+	case 8:
+		*(uint64_t *)addr = (uint64_t)value;
+		break;
+	default:
+		printf("Error: unreachable field size %u\n", fld_sz);
+		break;
+	}
+}
+
+#define __zxdh_nullp(typ) ((struct zxdh_ifc_##typ##_bits *)0)
+#define __zxdh_bit_sz(typ, fld) sizeof(__zxdh_nullp(typ)->fld)
+#define __zxdh_bit_off(typ, fld) ((unsigned int)(uintptr_t) \
+				  (&(__zxdh_nullp(typ)->fld)))
+#define __zxdh_dw_bit_off(typ, fld) (32 - __zxdh_bit_sz(typ, fld) - \
+				    (__zxdh_bit_off(typ, fld) & 0x1f))
+#define __zxdh_dw_off(typ, fld) (__zxdh_bit_off(typ, fld) / 32)
+#define __zxdh_64_off(typ, fld) (__zxdh_bit_off(typ, fld) / 64)
+#define __zxdh_dw_mask(typ, fld) (__zxdh_mask(typ, fld) << \
+				  __zxdh_dw_bit_off(typ, fld))
+#define __zxdh_mask(typ, fld) ((uint32_t)((1ull << __zxdh_bit_sz(typ, fld)) - 1))
+#define __zxdh_16_off(typ, fld) (__zxdh_bit_off(typ, fld) / 16)
+#define __zxdh_16_bit_off(typ, fld) (16 - __zxdh_bit_sz(typ, fld) - \
+				    (__zxdh_bit_off(typ, fld) & 0xf))
+#define __zxdh_mask16(typ, fld) ((uint16_t)((1ull << __zxdh_bit_sz(typ, fld)) - 1))
+#define __zxdh_16_mask(typ, fld) (__zxdh_mask16(typ, fld) << \
+				  __zxdh_16_bit_off(typ, fld))
+#define ZXDH_ST_SZ_BYTES(typ) (sizeof(struct zxdh_ifc_##typ##_bits) / 8)
+#define ZXDH_ST_SZ_DW(typ) (sizeof(struct zxdh_ifc_##typ##_bits) / 32)
+#define ZXDH_BYTE_OFF(typ, fld) (__zxdh_bit_off(typ, fld) / 8)
+#define ZXDH_ADDR_OF(typ, p, fld) ((uint8_t *)(p) + ZXDH_BYTE_OFF(typ, fld))
+
+#define BUILD_BUG_ON(condition) do { \
+	if (condition) \
+		__builtin_unreachable(); \
+	} while (0)
+
+#define ZXDH_SET(typ, p, fld, v) do { \
+	BUILD_BUG_ON(__zxdh_bit_sz(typ, fld) % 8); \
+	uint32_t fld_sz = __zxdh_bit_sz(typ, fld) / 8; \
+	uint8_t *addr = ZXDH_ADDR_OF(typ, p, fld); \
+	zxdh_set_value(fld_sz, addr, v); \
+} while (0)
+
+#define ZXDH_GET(typ, p, fld) ({ \
+	BUILD_BUG_ON(__zxdh_bit_sz(typ, fld) % 8); \
+	uint32_t fld_sz = __zxdh_bit_sz(typ, fld) / 8; \
+	uint8_t *addr = ZXDH_ADDR_OF(typ, p, fld); \
+	zxdh_get_value(fld_sz, addr); \
+})
+
+#define ZXDH_SET_ARRAY(typ, p, fld, index, value, type) \
+	do { \
+		type *addr = (type *)((uint8_t *)ZXDH_ADDR_OF(typ, p, fld) + \
+		(index) * sizeof(type)); \
+		*addr = (type)(value); \
+	} while (0)
+
+#define ZXDH_GET_ARRAY(typ, p, fld, index, type) \
+	({ \
+		type *addr = (type *)((uint8_t *)ZXDH_ADDR_OF(typ, p, fld) + \
+		(index) * sizeof(type)); \
+		*addr; \
+	})
+
 int32_t zxdh_phyport_get(struct rte_eth_dev *dev, uint8_t *phyport);
 int32_t zxdh_panelid_get(struct rte_eth_dev *dev, uint8_t *pannelid);
 int32_t zxdh_hashidx_get(struct rte_eth_dev *dev, uint8_t *hash_idx);
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index ea8b18e5e1..ba7ea52d20 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -1250,10 +1250,6 @@ zxdh_set_rxtx_funcs(struct rte_eth_dev *eth_dev)
 {
 	struct zxdh_hw *hw = eth_dev->data->dev_private;
 
-	if (!zxdh_pci_packed_queue(hw)) {
-		PMD_DRV_LOG(ERR, "port %u not support packed queue", eth_dev->data->port_id);
-		return -1;
-	}
 	if (!zxdh_pci_with_feature(hw, ZXDH_NET_F_MRG_RXBUF)) {
 		PMD_DRV_LOG(ERR, "port %u not support rx mergeable", eth_dev->data->port_id);
 		return -1;
@@ -1498,6 +1494,8 @@ zxdh_dtb_dump_res_init(struct zxdh_hw *hw, ZXDH_DEV_INIT_CTRL_T *dpp_ctrl)
 
 	struct zxdh_dtb_bulk_dump_info dtb_dump_baseres[] = {
 		{"sdt_vport_att_table", 4 * 1024 * 1024, ZXDH_SDT_VPORT_ATT_TABLE, NULL},
+		{"sdt_vlan_att_table", 4 * 1024 * 1024, ZXDH_SDT_VLAN_ATT_TABLE, NULL},
+		{"sdt_rss_table", 4 * 1024 * 1024, ZXDH_SDT_RSS_ATT_TABLE, NULL},
 		{"sdt_l2_entry_table0", 5 * 1024 * 1024, ZXDH_SDT_L2_ENTRY_TABLE0, NULL},
 		{"sdt_l2_entry_table1", 5 * 1024 * 1024, ZXDH_SDT_L2_ENTRY_TABLE1, NULL},
 		{"sdt_l2_entry_table2", 5 * 1024 * 1024, ZXDH_SDT_L2_ENTRY_TABLE2, NULL},
@@ -1514,7 +1512,8 @@ zxdh_dtb_dump_res_init(struct zxdh_hw *hw, ZXDH_DEV_INIT_CTRL_T *dpp_ctrl)
 	for (i = 0; i < (int)RTE_DIM(dtb_dump_baseres); i++) {
 		struct zxdh_dtb_bulk_dump_info *p = dtb_dump_baseres + i;
 		char buf[ZXDH_MAX_NAME_LEN] = {0};
-
+		memset(buf, '\0', sizeof(buf));
+		sprintf(buf, "%s_%x", p->mz_name, hw->dev_id);
 		p->mz_name = buf;
 
 		const struct rte_memzone *generic_dump_mz =
@@ -1544,6 +1543,7 @@ zxdh_np_dtb_res_init(struct rte_eth_dev *dev)
 	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_bar_offset_params param = {0};
 	struct zxdh_bar_offset_res res = {0};
+	char buf[ZXDH_MAX_NAME_LEN] = {0};
 	struct zxdh_dtb_shared_data *dtb_data = &hw->dev_sd->dtb_sd;
 	int ret = 0;
 
@@ -1569,7 +1569,7 @@ zxdh_np_dtb_res_init(struct rte_eth_dev *dev)
 	dpp_ctrl->vport = hw->vport.vport;
 	dpp_ctrl->vector = ZXDH_MSIX_INTR_DTB_VEC;
 	strlcpy(dpp_ctrl->port_name, dev->device->name, sizeof(dpp_ctrl->port_name));
-	dpp_ctrl->pcie_vir_addr = (uint32_t)hw->bar_addr[0];
+	dpp_ctrl->pcie_vir_addr = (uint64_t)hw->bar_addr[0];
 
 	param.pcie_id = hw->pcie_id;
 	param.virt_addr = hw->bar_addr[0] + ZXDH_CTRLCH_OFFSET;
@@ -1584,7 +1584,8 @@ zxdh_np_dtb_res_init(struct rte_eth_dev *dev)
 	dpp_ctrl->np_bar_offset = res.bar_offset;
 
 	if (!dtb_data->dtb_table_conf_mz) {
-		const struct rte_memzone *conf_mz = rte_memzone_reserve_aligned("zxdh_dtb_table_conf_mz",
+		sprintf(buf, "%s_%x", "zxdh_dtb_table_conf_mz", hw->dev_id);
+		const struct rte_memzone *conf_mz = rte_memzone_reserve_aligned(buf,
 				ZXDH_DTB_TABLE_CONF_SIZE, SOCKET_ID_ANY, 0, RTE_CACHE_LINE_SIZE);
 
 		if (conf_mz == NULL) {
@@ -1600,7 +1601,9 @@ zxdh_np_dtb_res_init(struct rte_eth_dev *dev)
 	}
 
 	if (!dtb_data->dtb_table_dump_mz) {
-		const struct rte_memzone *dump_mz = rte_memzone_reserve_aligned("zxdh_dtb_table_dump_mz",
+		memset(buf, '\0', sizeof(buf));
+		sprintf(buf, "%s_%x", "zxdh_dtb_table_dump_mz", hw->dev_id);
+		const struct rte_memzone *dump_mz = rte_memzone_reserve_aligned(buf,
 				ZXDH_DTB_TABLE_DUMP_SIZE, SOCKET_ID_ANY, 0, RTE_CACHE_LINE_SIZE);
 
 		if (dump_mz == NULL) {
diff --git a/drivers/net/zxdh/zxdh_ethdev_ops.c b/drivers/net/zxdh/zxdh_ethdev_ops.c
index 2b02734c62..f8e8d26c50 100644
--- a/drivers/net/zxdh/zxdh_ethdev_ops.c
+++ b/drivers/net/zxdh/zxdh_ethdev_ops.c
@@ -15,6 +15,7 @@
 #include "zxdh_np.h"
 #include "zxdh_queue.h"
 #include "zxdh_mtr.h"
+#include "zxdh_common.h"
 
 #define ZXDH_VLAN_FILTER_GROUPS       64
 #define ZXDH_INVALID_LOGIC_QID        0xFFFFU
@@ -278,9 +279,11 @@ zxdh_link_info_get(struct rte_eth_dev *dev, struct rte_eth_link *link)
 {
 	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_msg_info msg_info = {0};
-	struct zxdh_msg_reply_info reply_info = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 	uint16_t status = 0;
 	int32_t ret = 0;
+	void *reply_body_addr = ZXDH_ADDR_OF(msg_reply_info, zxdh_msg_reply_info, reply_body);
+	void *link_msg_addr = ZXDH_ADDR_OF(msg_reply_body, reply_body_addr, link_msg);
 
 	if (zxdh_pci_with_feature(hw, ZXDH_NET_F_STATUS))
 		zxdh_pci_read_dev_config(hw, offsetof(struct zxdh_net_config, status),
@@ -295,17 +298,18 @@ zxdh_link_info_get(struct rte_eth_dev *dev, struct rte_eth_link *link)
 		zxdh_agent_msg_build(hw, ZXDH_MAC_LINK_GET, &msg_info);
 
 		ret = zxdh_send_msg_to_riscv(dev, &msg_info, sizeof(struct zxdh_msg_info),
-				&reply_info, sizeof(struct zxdh_msg_reply_info),
+				zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info),
 				ZXDH_BAR_MODULE_MAC);
 		if (ret) {
 			PMD_DRV_LOG(ERR, "Failed to send msg: port 0x%x msg type %d",
 					hw->vport.vport, ZXDH_MAC_LINK_GET);
 			return -1;
 		}
-		link->link_speed = reply_info.reply_body.link_msg.speed;
-		link->link_autoneg = reply_info.reply_body.link_msg.autoneg;
-		hw->speed_mode = reply_info.reply_body.link_msg.speed_modes;
-		if ((reply_info.reply_body.link_msg.duplex & RTE_ETH_LINK_FULL_DUPLEX) ==
+
+		link->link_speed = ZXDH_GET(link_info_msg, link_msg_addr, speed);
+		link->link_autoneg = ZXDH_GET(link_info_msg, link_msg_addr, autoneg);
+		hw->speed_mode = ZXDH_GET(link_info_msg, link_msg_addr, speed_modes);
+		if ((ZXDH_GET(link_info_msg, link_msg_addr, duplex) & RTE_ETH_LINK_FULL_DUPLEX) ==
 				RTE_ETH_LINK_FULL_DUPLEX)
 			link->link_duplex = RTE_ETH_LINK_FULL_DUPLEX;
 		else
@@ -433,7 +437,7 @@ zxdh_dev_mac_addr_set(struct rte_eth_dev *dev, struct rte_ether_addr *addr)
 		ret = zxdh_del_mac_table(hw, hw->vport.vport, old_addr,
 			hw->hash_search_index, 0, 0);
 		if (ret) {
-			PMD_DRV_LOG(ERR, "mac_addr_add failed, code:%d", ret);
+			PMD_DRV_LOG(ERR, "mac_addr_del failed, code:%d", ret);
 			return ret;
 		}
 		hw->uc_num--;
@@ -467,6 +471,8 @@ zxdh_dev_mac_addr_set(struct rte_eth_dev *dev, struct rte_ether_addr *addr)
 		hw->uc_num--;
 	}
 	rte_ether_addr_copy(addr, (struct rte_ether_addr *)hw->mac_addr);
+	zxdh_pci_write_dev_config(hw, offsetof(struct zxdh_net_config, mac),
+								&hw->mac_addr, RTE_ETHER_ADDR_LEN);
 	return ret;
 }
 
@@ -566,7 +572,7 @@ zxdh_dev_mac_addr_add(struct rte_eth_dev *dev, struct rte_ether_addr *mac_addr,
 
 void zxdh_dev_mac_addr_remove(struct rte_eth_dev *dev, uint32_t index)
 {
-	struct zxdh_hw *hw	= dev->data->dev_private;
+	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_msg_info msg_info = {0};
 	struct rte_ether_addr *mac_addr = &dev->data->mac_addrs[index];
 	uint16_t ret = 0;
@@ -1072,7 +1078,9 @@ zxdh_dev_rss_reta_query(struct rte_eth_dev *dev,
 {
 	struct zxdh_hw *hw = (struct zxdh_hw *)dev->data->dev_private;
 	struct zxdh_msg_info msg = {0};
-	struct zxdh_msg_reply_info reply_msg = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
+	void *reply_body_addr = ZXDH_ADDR_OF(msg_reply_info, zxdh_msg_reply_info, reply_body);
+	void *rss_reta_msg_addr = ZXDH_ADDR_OF(msg_reply_body, reply_body_addr, rss_reta_msg);
 	uint16_t idx;
 	uint16_t i;
 	int ret = 0;
@@ -1094,21 +1102,21 @@ zxdh_dev_rss_reta_query(struct rte_eth_dev *dev,
 	zxdh_msg_head_build(hw, ZXDH_RSS_RETA_GET, &msg);
 
 	if (hw->is_pf) {
-		ret = zxdh_rss_table_get(hw, hw->vport.vport, &reply_msg.reply_body.rss_reta);
+		ret = zxdh_rss_table_get(hw, hw->vport.vport, rss_reta_msg_addr);
 		if (ret) {
 			PMD_DRV_LOG(ERR, "rss reta table set failed");
 			return -EINVAL;
 		}
 	} else {
 		ret = zxdh_vf_send_msg_to_pf(dev, &msg, sizeof(struct zxdh_msg_info),
-					&reply_msg, sizeof(struct zxdh_msg_reply_info));
+					zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info));
 		if (ret) {
 			PMD_DRV_LOG(ERR, "vf rss reta table get failed");
 			return -EINVAL;
 		}
 	}
 
-	struct zxdh_rss_reta *reta_table = &reply_msg.reply_body.rss_reta;
+	struct zxdh_rss_reta *reta_table = rss_reta_msg_addr;
 
 	for (idx = 0, i = 0; i < reta_size; ++i) {
 		idx = i / RTE_ETH_RETA_GROUP_SIZE;
@@ -1232,10 +1240,13 @@ zxdh_rss_hash_conf_get(struct rte_eth_dev *dev, struct rte_eth_rss_conf *rss_con
 	struct zxdh_hw *hw = (struct zxdh_hw *)dev->data->dev_private;
 	struct rte_eth_rss_conf *old_rss_conf = &dev->data->dev_conf.rx_adv_conf.rss_conf;
 	struct zxdh_msg_info msg = {0};
-	struct zxdh_msg_reply_info reply_msg = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
+	void *reply_body_addr = ZXDH_ADDR_OF(msg_reply_info, zxdh_msg_reply_info, reply_body);
+	void *rss_hf_msg_addr = ZXDH_ADDR_OF(msg_reply_body, reply_body_addr, rss_hf_msg);
 	struct zxdh_port_attr_table port_attr = {0};
-	int ret;
+	uint32_t rss_hf;
 	uint32_t hw_hf;
+	int ret;
 
 	if (rss_conf == NULL) {
 		PMD_DRV_LOG(ERR, "rss conf is NULL");
@@ -1252,16 +1263,17 @@ zxdh_rss_hash_conf_get(struct rte_eth_dev *dev, struct rte_eth_rss_conf *rss_con
 			PMD_DRV_LOG(ERR, "rss hash factor set failed");
 			return -EINVAL;
 		}
-		reply_msg.reply_body.rss_hf.rss_hf = port_attr.rss_hash_factor;
+		ZXDH_SET(rss_hf, rss_hf_msg_addr, rss_hf, port_attr.rss_hash_factor);
 	} else {
 		ret = zxdh_vf_send_msg_to_pf(dev, &msg, sizeof(struct zxdh_msg_info),
-				&reply_msg, sizeof(struct zxdh_msg_reply_info));
+				zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info));
 		if (ret) {
 			PMD_DRV_LOG(ERR, "rss hash factor set failed");
 			return -EINVAL;
 		}
 	}
-	rss_conf->rss_hf = zxdh_rss_hf_to_eth(reply_msg.reply_body.rss_hf.rss_hf);
+	rss_hf = ZXDH_GET(rss_hf, rss_hf_msg_addr, rss_hf);
+	rss_conf->rss_hf = zxdh_rss_hf_to_eth(rss_hf);
 
 	return 0;
 }
@@ -1382,7 +1394,8 @@ zxdh_hw_vqm_stats_get(struct rte_eth_dev *dev, enum zxdh_agent_msg_type opcode,
 {
 	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_msg_info msg_info = {0};
-	struct zxdh_msg_reply_info reply_info = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
+	void *reply_body_addr = ZXDH_ADDR_OF(msg_reply_info, zxdh_msg_reply_info, reply_body);
 	enum ZXDH_BAR_MODULE_ID module_id;
 	int ret = 0;
 
@@ -1404,14 +1417,15 @@ zxdh_hw_vqm_stats_get(struct rte_eth_dev *dev, enum zxdh_agent_msg_type opcode,
 	zxdh_agent_msg_build(hw, opcode, &msg_info);
 
 	ret = zxdh_send_msg_to_riscv(dev, &msg_info, sizeof(struct zxdh_msg_info),
-				&reply_info, sizeof(struct zxdh_msg_reply_info), module_id);
+				zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info), module_id);
 	if (ret) {
 		PMD_DRV_LOG(ERR, "Failed to get hw stats");
 		return -1;
 	}
-	struct zxdh_msg_reply_body *reply_body = &reply_info.reply_body;
 
-	memcpy(hw_stats, &reply_body->vqm_stats, sizeof(struct zxdh_hw_vqm_stats));
+	void *hw_vqm_stats = ZXDH_ADDR_OF(msg_reply_body, reply_body_addr, vqm_stats);
+	memcpy(hw_stats, hw_vqm_stats, sizeof(struct zxdh_hw_vqm_stats));
+
 	return 0;
 }
 
@@ -1578,7 +1592,9 @@ zxdh_hw_np_stats_get(struct rte_eth_dev *dev, struct zxdh_hw_np_stats *np_stats)
 {
 	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_msg_info msg_info = {0};
-	struct zxdh_msg_reply_info reply_info = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
+	void *reply_body_addr = ZXDH_ADDR_OF(msg_reply_info, zxdh_msg_reply_info, reply_body);
+	void *hw_stas_addr = ZXDH_ADDR_OF(msg_reply_body, reply_body_addr, hw_stats);
 	int ret = 0;
 
 	if (hw->is_pf) {
@@ -1590,13 +1606,13 @@ zxdh_hw_np_stats_get(struct rte_eth_dev *dev, struct zxdh_hw_np_stats *np_stats)
 	} else {
 		zxdh_msg_head_build(hw, ZXDH_GET_NP_STATS, &msg_info);
 		ret = zxdh_vf_send_msg_to_pf(dev, &msg_info, sizeof(struct zxdh_msg_info),
-					&reply_info, sizeof(struct zxdh_msg_reply_info));
+					zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info));
 		if (ret) {
 			PMD_DRV_LOG(ERR,
 				"Failed to send msg: port 0x%x msg type", hw->vport.vport);
 			return -1;
 		}
-		memcpy(np_stats, &reply_info.reply_body.np_stats, sizeof(struct zxdh_hw_np_stats));
+		memcpy(np_stats, hw_stas_addr, sizeof(struct zxdh_hw_np_stats));
 	}
 	return ret;
 }
@@ -1666,7 +1682,7 @@ zxdh_hw_stats_reset(struct rte_eth_dev *dev, enum zxdh_agent_msg_type opcode)
 {
 	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_msg_info msg_info = {0};
-	struct zxdh_msg_reply_info reply_info = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 	enum ZXDH_BAR_MODULE_ID module_id;
 	int ret = 0;
 
@@ -1685,7 +1701,7 @@ zxdh_hw_stats_reset(struct rte_eth_dev *dev, enum zxdh_agent_msg_type opcode)
 	zxdh_agent_msg_build(hw, opcode, &msg_info);
 
 	ret = zxdh_send_msg_to_riscv(dev, &msg_info, sizeof(struct zxdh_msg_info),
-				&reply_info, sizeof(struct zxdh_msg_reply_info), module_id);
+				zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info), module_id);
 	if (ret) {
 		PMD_DRV_LOG(ERR, "Failed to reset hw stats");
 		return -1;
@@ -1767,13 +1783,13 @@ zxdh_hw_np_stats_vf_reset(struct rte_eth_dev *dev)
 {
 	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_msg_info msg_info = {0};
-	struct zxdh_msg_reply_info reply_info = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 	int ret = 0;
 
 	msg_info.data.np_stats_query.clear_mode = 1;
 	zxdh_msg_head_build(hw, ZXDH_GET_NP_STATS, &msg_info);
 	ret = zxdh_vf_send_msg_to_pf(dev, &msg_info, sizeof(struct zxdh_msg_info),
-			&reply_info, sizeof(reply_info));
+			zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info));
 	if (ret)
 		PMD_DRV_LOG(ERR, "Failed to send ZXDH_PORT_METER_STAT_GET msg. code:%d", ret);
 
@@ -2034,29 +2050,24 @@ zxdh_dev_fw_version_get(struct rte_eth_dev *dev,
 {
 	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_msg_info msg_info = {0};
-	struct zxdh_msg_reply_info reply_info = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
+	void *reply_body_addr = ZXDH_ADDR_OF(msg_reply_info, zxdh_msg_reply_info, reply_body);
+	void *flash_msg_addr = ZXDH_ADDR_OF(msg_reply_body, reply_body_addr, flash_msg);
 	char fw_ver[ZXDH_FWVERS_LEN] = {0};
 	uint32_t ret = 0;
 
 	zxdh_agent_msg_build(hw, ZXDH_FLASH_FIR_VERSION_GET, &msg_info);
 
-	struct zxdh_msg_recviver_mem rsp_data = {
-			.recv_buffer = (void *)&reply_info,
-			.buffer_len = sizeof(struct zxdh_msg_reply_info),
-	};
-
 	ret = zxdh_send_msg_to_riscv(dev, &msg_info, sizeof(struct zxdh_msg_info),
-				&reply_info, sizeof(struct zxdh_msg_reply_info),
+				zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info),
 				ZXDH_MODULE_FLASH);
 	if (ret) {
 		PMD_DRV_LOG(ERR, "Failed to send msg: port 0x%x msg type %d",
 				hw->vport.vport, ZXDH_FLASH_FIR_VERSION_GET);
 		return -1;
 	}
-	struct zxdh_msg_reply_body *ack_msg =
-			 &(((struct zxdh_msg_reply_info *)rsp_data.recv_buffer)->reply_body);
 
-	memcpy(fw_ver, ack_msg->flash_msg.firmware_version, ZXDH_FWVERS_LEN);
+	memcpy(fw_ver, flash_msg_addr, ZXDH_FWVERS_LEN);
 	snprintf(fw_version, ZXDH_FWVERS_LEN - 1, "%s", fw_ver);
 
 	return 0;
@@ -2068,7 +2079,7 @@ zxdh_en_module_eeprom_read(struct rte_eth_dev *dev,
 {
 	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_msg_info msg_info = {0};
-	struct zxdh_msg_reply_info reply_info = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 	uint8_t ret = 0;
 
 	zxdh_agent_msg_build(hw, ZXDH_MAC_MODULE_EEPROM_READ, &msg_info);
@@ -2079,26 +2090,24 @@ zxdh_en_module_eeprom_read(struct rte_eth_dev *dev,
 	msg_info.data.module_eeprom_msg.offset = query->offset;
 	msg_info.data.module_eeprom_msg.length = query->length;
 
-	struct zxdh_msg_recviver_mem rsp_data = {
-			.recv_buffer = (void *)&reply_info,
-			.buffer_len = sizeof(struct zxdh_msg_reply_info),
-	};
-
 	ret = zxdh_send_msg_to_riscv(dev, &msg_info, sizeof(struct zxdh_msg_info),
-				&reply_info, sizeof(struct zxdh_msg_reply_info),
+				zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info),
 				ZXDH_BAR_MODULE_MAC);
 	if (ret) {
 		PMD_DRV_LOG(ERR, "Failed to send msg: port 0x%x msg type %d",
 				hw->vport.vport, ZXDH_MAC_MODULE_EEPROM_READ);
 		return -1;
 	}
-	struct zxdh_msg_reply_body *ack_msg =
-			 &(((struct zxdh_msg_reply_info *)rsp_data.recv_buffer)->reply_body);
-
+	void *reply_body_addr = ZXDH_ADDR_OF(msg_reply_info, zxdh_msg_reply_info, reply_body);
+	void *module_eeprom_msg_addr =
+			ZXDH_ADDR_OF(msg_reply_body, reply_body_addr, module_eeprom_msg);
+	void *agent_mac_module_eeprom_msg_data_addr =
+			ZXDH_ADDR_OF(agent_mac_module_eeprom_msg, module_eeprom_msg_addr, data);
+	uint8_t length = ZXDH_GET(agent_mac_module_eeprom_msg, module_eeprom_msg_addr, length);
 	if (data)
-		memcpy(data, ack_msg->module_eeprom_msg.data, ack_msg->module_eeprom_msg.length);
+		memcpy(data, agent_mac_module_eeprom_msg_data_addr, length);
 
-	return ack_msg->module_eeprom_msg.length;
+	return length;
 }
 
 int
diff --git a/drivers/net/zxdh/zxdh_ethdev_ops.h b/drivers/net/zxdh/zxdh_ethdev_ops.h
index a5162a6d6b..97a1eb4532 100644
--- a/drivers/net/zxdh/zxdh_ethdev_ops.h
+++ b/drivers/net/zxdh/zxdh_ethdev_ops.h
@@ -62,6 +62,30 @@ struct zxdh_hw_np_stats {
 	uint64_t tx_ssvpc_pkts;
 };
 
+struct zxdh_ifc_hw_np_stats_bits {
+	uint8_t rx_unicast_pkts[0x40];
+	uint8_t tx_unicast_pkts[0x40];
+	uint8_t rx_unicast_bytes[0x40];
+	uint8_t tx_unicast_bytes[0x40];
+	uint8_t rx_multicast_pkts[0x40];
+	uint8_t tx_multicast_pkts[0x40];
+	uint8_t rx_multicast_bytes[0x40];
+	uint8_t tx_multicast_bytes[0x40];
+	uint8_t rx_broadcast_pkts[0x40];
+	uint8_t tx_broadcast_pkts[0x40];
+	uint8_t rx_broadcast_bytes[0x40];
+	uint8_t tx_broadcast_bytes[0x40];
+	uint8_t rx_mtu_drop_pkts[0x40];
+	uint8_t tx_mtu_drop_pkts[0x40];
+	uint8_t rx_mtu_drop_bytes[0x40];
+	uint8_t tx_mtu_drop_bytes[0x40];
+	uint8_t rx_mtr_drop_pkts[0x40];
+	uint8_t tx_mtr_drop_pkts[0x40];
+	uint8_t rx_mtr_drop_bytes[0x40];
+	uint8_t tx_mtr_drop_bytes[0x40];
+	uint8_t tx_ssvpc_pkts[0x40];
+};
+
 struct zxdh_hw_vqm_stats {
 	uint64_t rx_total;
 	uint64_t tx_total;
@@ -72,6 +96,17 @@ struct zxdh_hw_vqm_stats {
 	uint64_t rx_drop;
 };
 
+struct zxdh_ifc_hw_vqm_stats_bits {
+	uint8_t rx_total[0x40];
+	uint8_t tx_total[0x40];
+	uint8_t rx_bytes[0x40];
+	uint8_t tx_bytes[0x40];
+	uint8_t rx_error[0x40];
+	uint8_t tx_error[0x40];
+	uint8_t rx_drop[0x40];
+};
+
+
 int zxdh_dev_set_link_up(struct rte_eth_dev *dev);
 int zxdh_dev_set_link_down(struct rte_eth_dev *dev);
 int32_t zxdh_dev_link_update(struct rte_eth_dev *dev, int32_t wait_to_complete);
diff --git a/drivers/net/zxdh/zxdh_msg.c b/drivers/net/zxdh/zxdh_msg.c
index 96ad638e83..02ecd93b12 100644
--- a/drivers/net/zxdh/zxdh_msg.c
+++ b/drivers/net/zxdh/zxdh_msg.c
@@ -18,6 +18,7 @@
 #include "zxdh_pci.h"
 #include "zxdh_tables.h"
 #include "zxdh_np.h"
+#include "zxdh_common.h"
 
 #define ZXDH_REPS_INFO_FLAG_USABLE  0x00
 #define ZXDH_BAR_SEQID_NUM_MAX      256
@@ -695,7 +696,7 @@ static uint16_t
 zxdh_bar_chan_sync_msg_reps_get(uint64_t subchan_addr,
 		uint64_t recv_buffer, uint16_t buffer_len)
 {
-	struct zxdh_bar_msg_header msg_header = {0};
+	struct zxdh_bar_msg_header msg_header;
 	uint16_t msg_id = 0;
 	uint16_t msg_len = 0;
 
@@ -987,7 +988,7 @@ zxdh_bar_chan_msg_header_check(struct zxdh_bar_msg_header *msg_header)
 int
 zxdh_bar_irq_recv(uint8_t src, uint8_t dst, uint64_t virt_addr, void *dev)
 {
-	struct zxdh_bar_msg_header msg_header = {0};
+	struct zxdh_bar_msg_header msg_header;
 	uint64_t recv_addr = 0;
 	uint64_t reps_addr = 0;
 	uint16_t ret = 0;
@@ -1083,22 +1084,20 @@ zxdh_vf_send_msg_to_pf(struct rte_eth_dev *dev,  void *msg_req,
 {
 	struct zxdh_hw *hw  = dev->data->dev_private;
 	struct zxdh_msg_recviver_mem result = {0};
-	struct zxdh_msg_reply_info reply_info = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 	int ret = 0;
 
 	if (reply) {
-		RTE_ASSERT(reply_len < sizeof(struct zxdh_msg_reply_info));
+		RTE_ASSERT(reply_len < ZXDH_ST_SZ_BYTES(msg_reply_info));
 		result.recv_buffer  = reply;
 		result.buffer_len = reply_len;
 	} else {
-		result.recv_buffer = &reply_info;
-		result.buffer_len = sizeof(reply_info);
+		result.recv_buffer = zxdh_msg_reply_info;
+		result.buffer_len = ZXDH_ST_SZ_BYTES(msg_reply_info);
 	}
 
-	struct zxdh_msg_reply_head *reply_head =
-				&(((struct zxdh_msg_reply_info *)result.recv_buffer)->reply_head);
-	struct zxdh_msg_reply_body *reply_body =
-				&(((struct zxdh_msg_reply_info *)result.recv_buffer)->reply_body);
+	void *reply_head_addr = ZXDH_ADDR_OF(msg_reply_info, result.recv_buffer, reply_head);
+	void *reply_body_addr = ZXDH_ADDR_OF(msg_reply_info, result.recv_buffer, reply_body);
 
 	struct zxdh_pci_bar_msg in = {
 		.virt_addr = (uint64_t)(hw->bar_addr[ZXDH_BAR0_INDEX] +
@@ -1118,12 +1117,16 @@ zxdh_vf_send_msg_to_pf(struct rte_eth_dev *dev,  void *msg_req,
 			"vf[%d] send bar msg to pf failed.ret %d", hw->vport.vfid, ret);
 		return -1;
 	}
-	if (reply_head->flag != ZXDH_MSG_REPS_OK) {
+
+	uint8_t flag = ZXDH_GET(msg_reply_head, reply_head_addr, flag);
+	uint16_t reps_len = ZXDH_GET(msg_reply_head, reply_head_addr, reps_len);
+	if (flag != ZXDH_MSG_REPS_OK) {
 		PMD_MSG_LOG(ERR, "vf[%d] get pf reply failed: reply_head flag : 0x%x(0xff is OK).replylen %d",
-				hw->vport.vfid, reply_head->flag, reply_head->reps_len);
+				hw->vport.vfid, flag, reps_len);
 		return -1;
 	}
-	if (reply_body->flag != ZXDH_REPS_SUCC) {
+	uint8_t reply_body_flag = ZXDH_GET(msg_reply_body, reply_body_addr, flag);
+	if (reply_body_flag != ZXDH_REPS_SUCC) {
 		PMD_MSG_LOG(ERR, "vf[%d] msg processing failed", hw->vfid);
 		return -1;
 	}
@@ -1137,23 +1140,19 @@ zxdh_send_msg_to_riscv(struct rte_eth_dev *dev, void *msg_req,
 {
 	struct zxdh_hw *hw = dev->data->dev_private;
 	struct zxdh_msg_recviver_mem result = {0};
-	struct zxdh_msg_reply_info reply_info = {0};
+	uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 
 	if (reply) {
-		RTE_ASSERT(reply_len < sizeof(struct zxdh_msg_reply_info));
+		RTE_ASSERT(reply_len < ZXDH_ST_SZ_BYTES(msg_reply_info));
 		result.recv_buffer  = reply;
 		result.buffer_len = reply_len;
 	} else {
-		result.recv_buffer = &reply_info;
-		result.buffer_len = sizeof(reply_info);
+		result.recv_buffer = zxdh_msg_reply_info;
+		result.buffer_len = ZXDH_ST_SZ_BYTES(msg_reply_info);
 	}
-	struct zxdh_msg_reply_head *reply_head =
-				&(((struct zxdh_msg_reply_info *)result.recv_buffer)->reply_head);
-	struct zxdh_msg_reply_body *reply_body =
-				&(((struct zxdh_msg_reply_info *)result.recv_buffer)->reply_body);
 
 	struct zxdh_pci_bar_msg in = {
-		.payload_addr = &msg_req,
+		.payload_addr = msg_req,
 		.payload_len = msg_req_len,
 		.virt_addr = (uint64_t)(hw->bar_addr[ZXDH_BAR0_INDEX] + ZXDH_CTRLCH_OFFSET),
 		.src = hw->is_pf ? ZXDH_MSG_CHAN_END_PF : ZXDH_MSG_CHAN_END_VF,
@@ -1166,15 +1165,6 @@ zxdh_send_msg_to_riscv(struct rte_eth_dev *dev, void *msg_req,
 		PMD_MSG_LOG(ERR, "Failed to send sync messages or receive response");
 		return -1;
 	}
-	if (reply_head->flag != ZXDH_MSG_REPS_OK) {
-		PMD_MSG_LOG(ERR, "vf[%d] get pf reply failed: reply_head flag : 0x%x(0xff is OK).replylen %d",
-				hw->vport.vfid, reply_head->flag, reply_head->reps_len);
-		return -1;
-	}
-	if (reply_body->flag != ZXDH_REPS_SUCC) {
-		PMD_MSG_LOG(ERR, "vf[%d] msg processing failed", hw->vfid);
-		return -1;
-	}
 
 	return 0;
 }
@@ -1245,7 +1235,7 @@ zxdh_vf_promisc_uninit(struct zxdh_hw *hw, union zxdh_virport_num vport)
 
 static int
 zxdh_vf_port_init(struct zxdh_hw *pf_hw, uint16_t vport, void *cfg_data,
-		struct zxdh_msg_reply_body *res_info, uint16_t *res_len)
+		void *res_info, uint16_t *res_len)
 {
 	struct zxdh_port_attr_table port_attr = {0};
 	union zxdh_virport_num port = {.vport = vport};
@@ -1275,14 +1265,13 @@ zxdh_vf_port_init(struct zxdh_hw *pf_hw, uint16_t vport, void *cfg_data,
 		PMD_DRV_LOG(ERR, "vf_promisc_table_init failed, code:%d", ret);
 		goto proc_end;
 	}
-
-	res_info->flag = ZXDH_REPS_SUCC;
-	*res_len = sizeof(res_info->flag);
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_SUCC);
+	*res_len = sizeof(uint8_t);
 
 	return ret;
 proc_end:
-	*res_len = sizeof(res_info->flag);
-	res_info->flag = ZXDH_REPS_FAIL;
+	*res_len = sizeof(uint8_t);
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_FAIL);
 	return ret;
 }
 
@@ -1311,11 +1300,12 @@ zxdh_mac_clear(struct zxdh_hw *hw, union zxdh_virport_num vport)
 static int
 zxdh_vf_port_uninit(struct zxdh_hw *pf_hw,
 		uint16_t vport, void *cfg_data __rte_unused,
-		struct zxdh_msg_reply_body *res_info, uint16_t *res_len)
+		void *res_info, uint16_t *res_len)
 {
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "uninit";
 	struct zxdh_port_attr_table port_attr = {0};
 	union zxdh_virport_num vport_num = {.vport = vport};
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, res_info, reply_data);
 	int ret = 0;
 
 	*res_len =  ZXDH_MSG_REPLYBODY_HEAD;
@@ -1340,20 +1330,20 @@ zxdh_vf_port_uninit(struct zxdh_hw *pf_hw,
 	}
 
 	*res_len += strlen(str);
-	rte_memcpy(&res_info->reply_data, str, strlen(str) + 1);
-	res_info->flag = ZXDH_REPS_SUCC;
+	rte_memcpy(reply_data_addr, str, strlen(str) + 1);
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_SUCC);
 	return ret;
 
 proc_end:
 	*res_len += strlen(str);
-	rte_memcpy(&res_info->reply_data, str, strlen(str) + 1);
-	res_info->flag = ZXDH_REPS_FAIL;
+	rte_memcpy(reply_data_addr, str, strlen(str) + 1);
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_FAIL);
 	return ret;
 }
 
 static int
 zxdh_add_vf_mac_table(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-		struct zxdh_msg_reply_body *reply_body, uint16_t *reply_len)
+		void *reply_body, uint16_t *reply_len)
 {
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "add mac";
 	union zxdh_virport_num port = {0};
@@ -1362,6 +1352,8 @@ zxdh_add_vf_mac_table(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
 	int i = 0, ret = 0;
 	uint16_t vf_id = port.vfid;
 	port.vport = vport;
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, reply_body, reply_data);
+	void *mac_reply_msg_addr = ZXDH_ADDR_OF(msg_reply_body, reply_body, mac_reply_msg);
 
 	for (i = 0; i < ZXDH_MAX_MAC_ADDRS; i++)
 		if (rte_is_same_ether_addr(&hw->vfinfo[vf_id].vf_mac[i], addr))
@@ -1369,7 +1361,7 @@ zxdh_add_vf_mac_table(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
 
 	ret = zxdh_add_mac_table(hw, vport, addr, hw->hash_search_index, 0, 0);
 	if (ret == -EADDRINUSE) {
-		reply_body->mac_reply_msg.mac_flag = ZXDH_EEXIST_MAC_FLAG;
+		ZXDH_SET(mac_reply_msg, mac_reply_msg_addr, mac_flag, ZXDH_EEXIST_MAC_FLAG);
 		PMD_DRV_LOG(ERR, "vf vport 0x%x set mac ret 0x%x failed. mac is in used.",
 				port.vport, ret);
 		goto failure;
@@ -1389,26 +1381,27 @@ zxdh_add_vf_mac_table(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
 success:
 	sprintf(str, " vport 0x%x set mac ret 0x%x\n", port.vport, ret);
 	*reply_len =  strlen(str) + ZXDH_MSG_REPLYBODY_HEAD;
-	rte_memcpy(&reply_body->reply_data, str, strlen(str) + 1);
-	reply_body->flag = ZXDH_REPS_SUCC;
+	rte_memcpy(reply_data_addr, str, strlen(str) + 1);
+	ZXDH_SET(msg_reply_body, reply_body, flag, ZXDH_REPS_SUCC);
 	PMD_DRV_LOG(DEBUG, " reply len %d", *reply_len);
 	return ret;
 
 failure:
 	*reply_len = strlen(str) + ZXDH_MSG_REPLYBODY_HEAD;
-	reply_body->flag = ZXDH_REPS_FAIL;
+	ZXDH_SET(msg_reply_body, reply_body, flag, ZXDH_REPS_FAIL);
 	return ret;
 }
 
 static int
 zxdh_del_vf_mac_table(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-	struct zxdh_msg_reply_body *res_info, uint16_t *res_len)
+	void *res_info, uint16_t *res_len)
 {
 	int ret, i = 0;
 	struct zxdh_mac_filter *mac_filter = (struct zxdh_mac_filter *)cfg_data;
 	union zxdh_virport_num  port = (union zxdh_virport_num)vport;
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "del mac";
 	uint16_t  vf_id = port.vfid;
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, res_info, reply_data);
 
 	PMD_DRV_LOG(DEBUG, "[PF GET MSG FROM VF]--vf mac to del.");
 	ret = zxdh_del_mac_table(hw, vport, &mac_filter->mac, hw->hash_search_index, 0, 0);
@@ -1428,19 +1421,19 @@ zxdh_del_vf_mac_table(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
 
 	sprintf(str, "vport 0x%x del mac ret 0x%x\n", port.vport, ret);
 	*res_len =  strlen(str) + ZXDH_MSG_REPLYBODY_HEAD;
-	rte_memcpy(&res_info->reply_data, str, strlen(str) + 1);
-	res_info->flag = ZXDH_REPS_SUCC;
+	rte_memcpy(reply_data_addr, str, strlen(str) + 1);
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_SUCC);
 	return ret;
 
 proc_end:
 	*res_len = strlen(str) + ZXDH_MSG_REPLYBODY_HEAD;
-	res_info->flag = ZXDH_REPS_FAIL;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_FAIL);
 	return ret;
 }
 
 static int
 zxdh_vf_promisc_set(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-		struct zxdh_msg_reply_body *reply, uint16_t *res_len)
+		void *reply, uint16_t *res_len)
 {
 	struct zxdh_port_promisc_msg *promisc_msg = (struct zxdh_port_promisc_msg *)cfg_data;
 	int ret = 0;
@@ -1458,24 +1451,25 @@ zxdh_vf_promisc_set(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
 		goto proc_end;
 	}
 
-	*res_len = sizeof(struct zxdh_port_attr_set_msg) + sizeof(enum zxdh_reps_flag);
-	reply->flag = ZXDH_REPS_SUCC;
+	*res_len = sizeof(struct zxdh_port_attr_set_msg) + sizeof(uint8_t);
+	ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_SUCC);
 
 	return ret;
 
 proc_end:
-	*res_len = sizeof(struct zxdh_port_attr_set_msg) + sizeof(enum zxdh_reps_flag);
-	reply->flag = ZXDH_REPS_FAIL;
+	*res_len = sizeof(struct zxdh_port_attr_set_msg) + sizeof(uint8_t);
+	ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_FAIL);
 	return ret;
 }
 
 static int
 zxdh_vf_vlan_filter_table_process(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-		struct zxdh_msg_reply_body *res_info, uint16_t *res_len, uint8_t enable)
+		void *res_info, uint16_t *res_len, uint8_t enable)
 {
 	struct zxdh_vlan_filter *vlan_filter = cfg_data;
 	uint16_t vlan_id =  vlan_filter->vlan_id;
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "vlan filter table";
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, res_info, reply_data);
 	int ret = 0;
 
 	ret = zxdh_vlan_filter_table_set(hw, vport, vlan_id, enable);
@@ -1483,33 +1477,38 @@ zxdh_vf_vlan_filter_table_process(struct zxdh_hw *hw, uint16_t vport, void *cfg_
 		sprintf(str, "vlan filter op-code[%d] vlan id:%d failed, code:%d\n",
 			enable, vlan_id, ret);
 
-	*res_len = strlen(str) + sizeof(enum zxdh_reps_flag);
-	memcpy(&res_info->reply_data, str, strlen(str) + 1);
-	res_info->flag = (ret == 0) ? ZXDH_REPS_SUCC : ZXDH_REPS_FAIL;
+	*res_len = strlen(str) + sizeof(uint8_t);
+
+	memcpy(reply_data_addr, str, strlen(str) + 1);
+	if (ret == 0)
+		ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_SUCC);
+	else
+		ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_FAIL);
 	return ret;
 }
 
 static int
 zxdh_vf_vlan_filter_table_add(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-		struct zxdh_msg_reply_body *res_info, uint16_t *res_len)
+		void *res_info, uint16_t *res_len)
 {
 	return zxdh_vf_vlan_filter_table_process(hw, vport, cfg_data, res_info, res_len, 1);
 }
 
 static int
 zxdh_vf_vlan_filter_table_del(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-		struct zxdh_msg_reply_body *res_info, uint16_t *res_len)
+		void *res_info, uint16_t *res_len)
 {
 	return zxdh_vf_vlan_filter_table_process(hw, vport, cfg_data, res_info, res_len, 0);
 }
 
 static int
 zxdh_vf_set_vlan_filter(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-		struct zxdh_msg_reply_body *reply, uint16_t *res_len)
+		void *reply, uint16_t *res_len)
 {
 	struct zxdh_vlan_filter_set *vlan_filter = cfg_data;
 	union zxdh_virport_num port = (union zxdh_virport_num)vport;
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "vlan filter";
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, reply, reply_data);
 	int ret = 0;
 	uint16_t vfid = port.vfid;
 
@@ -1517,19 +1516,23 @@ zxdh_vf_set_vlan_filter(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
 	if (ret)
 		sprintf(str, "[vfid:%d] vlan filter. set failed, ret:%d\n", vfid, ret);
 
-	*res_len = strlen(str) + sizeof(enum zxdh_reps_flag);
-	reply->flag = (ret == 0) ? ZXDH_REPS_SUCC : ZXDH_REPS_FAIL;
-	memcpy(&reply->reply_data, str, strlen(str) + 1);
+	*res_len = strlen(str) + sizeof(uint8_t);
+	if (ret == 0)
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_SUCC);
+	else
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_FAIL);
+	memcpy(reply_data_addr, str, strlen(str) + 1);
 	return ret;
 }
 
 static int
 zxdh_vf_set_vlan_offload(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-		struct zxdh_msg_reply_body *reply, uint16_t *res_len)
+		void *reply, uint16_t *res_len)
 {
 	struct zxdh_vlan_offload *vlan_offload = cfg_data;
 	union zxdh_virport_num port = (union zxdh_virport_num)vport;
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "vlan offload";
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, reply, reply_data);
 	int ret = 0;
 	uint16_t vfid = port.vfid;
 
@@ -1540,18 +1543,23 @@ zxdh_vf_set_vlan_offload(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
 	if (ret)
 		sprintf(str, "[vfid:%d] vlan offload set failed, ret:%d\n", vfid, ret);
 
-	*res_len = strlen(str) + sizeof(enum zxdh_reps_flag);
-	reply->flag = (ret == 0) ? ZXDH_REPS_SUCC : ZXDH_REPS_FAIL;
-	memcpy(&reply->reply_data, str, strlen(str) + 1);
+	*res_len = strlen(str) + sizeof(uint8_t);
+	if (ret == 0)
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_SUCC);
+	else
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_FAIL);
+	memcpy(reply_data_addr, str, strlen(str) + 1);
 	return ret;
 }
 
 static int
 zxdh_vf_rss_hf_get(struct zxdh_hw *hw, uint16_t vport, void *cfg_data __rte_unused,
-			struct zxdh_msg_reply_body *reply, uint16_t *res_len)
+			void *reply, uint16_t *res_len)
 {
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "rss_hf";
 	struct zxdh_port_attr_table vport_att = {0};
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, reply, reply_data);
+	void *rss_hf_msg_addr = ZXDH_ADDR_OF(msg_reply_body, reply, rss_hf_msg);
 	int ret = 0;
 
 	ret = zxdh_get_port_attr(hw, vport, &vport_att);
@@ -1561,22 +1569,26 @@ zxdh_vf_rss_hf_get(struct zxdh_hw *hw, uint16_t vport, void *cfg_data __rte_unus
 		goto proc_end;
 	}
 
-	reply->rss_hf.rss_hf = vport_att.rss_hash_factor;
+	ZXDH_SET(rss_hf, rss_hf_msg_addr, rss_hf, vport_att.rss_hash_factor);
 
 proc_end:
-	*res_len = strlen(str) + sizeof(enum zxdh_reps_flag);
-	reply->flag = (ret == 0) ? ZXDH_REPS_SUCC : ZXDH_REPS_FAIL;
-	memcpy(&reply->reply_data, str, strlen(str) + 1);
+	*res_len = strlen(str) + sizeof(uint8_t);
+	if (ret == 0)
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_SUCC);
+	else
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_FAIL);
+	memcpy(reply_data_addr, str, strlen(str) + 1);
 	return ret;
 }
 
 static int
 zxdh_vf_rss_hf_set(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-			struct zxdh_msg_reply_body *reply, uint16_t *res_len)
+			void *reply, uint16_t *res_len)
 {
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "rss_hf";
 	struct zxdh_rss_hf *rss_hf = cfg_data;
 	struct zxdh_port_attr_table vport_att = {0};
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, reply, reply_data);
 	int ret = 0;
 
 	ret = zxdh_get_port_attr(hw, vport, &vport_att);
@@ -1596,19 +1608,23 @@ zxdh_vf_rss_hf_set(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
 	}
 
 proc_end:
-	*res_len = strlen(str) + sizeof(enum zxdh_reps_flag);
-	reply->flag = (ret == 0) ? ZXDH_REPS_SUCC : ZXDH_REPS_FAIL;
-	memcpy(&reply->reply_data, str, strlen(str) + 1);
+	*res_len = strlen(str) + sizeof(uint8_t);
+	if (ret == 0)
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_SUCC);
+	else
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_FAIL);
+	memcpy(reply_data_addr, str, strlen(str) + 1);
 	return ret;
 }
 
 static int
 zxdh_vf_rss_enable(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-			struct zxdh_msg_reply_body *reply, uint16_t *res_len)
+			void *reply, uint16_t *res_len)
 {
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "rss_enable";
 	struct zxdh_rss_enable *rss_enable = cfg_data;
 	struct zxdh_port_attr_table vport_att = {0};
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, reply, reply_data);
 	int ret = 0;
 
 	ret = zxdh_get_port_attr(hw, vport, &vport_att);
@@ -1628,51 +1644,63 @@ zxdh_vf_rss_enable(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
 	}
 
 proc_end:
-	*res_len = strlen(str) + sizeof(enum zxdh_reps_flag);
-	reply->flag = (ret == 0) ? ZXDH_REPS_SUCC : ZXDH_REPS_FAIL;
-	memcpy(&reply->reply_data, str, strlen(str) + 1);
+	*res_len = strlen(str) + sizeof(uint8_t);
+	if (ret == 0)
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_SUCC);
+	else
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_FAIL);
+	memcpy(reply_data_addr, str, strlen(str) + 1);
 	return ret;
 }
 
 static int
 zxdh_vf_rss_table_set(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-		struct zxdh_msg_reply_body *reply, uint16_t *res_len)
+		void *reply, uint16_t *res_len)
 {
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "rss_table";
 	struct zxdh_rss_reta *rss_reta = cfg_data;
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, reply, reply_data);
 	int32_t ret = 0;
 
 	ret = zxdh_rss_table_set(hw, vport, rss_reta);
 	if (ret)
 		sprintf(str, "set rss reta tbl failed, code:%d", ret);
 
-	*res_len = strlen(str) + sizeof(enum zxdh_reps_flag);
-	reply->flag = (ret == 0) ? ZXDH_REPS_SUCC : ZXDH_REPS_FAIL;
-	memcpy(&reply->reply_data, str, strlen(str) + 1);
+	*res_len = strlen(str) + sizeof(uint8_t);
+	if (ret == 0)
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_SUCC);
+	else
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_FAIL);
+	memcpy(reply_data_addr, str, strlen(str) + 1);
 	return ret;
 }
 
 static int
 zxdh_vf_rss_table_get(struct zxdh_hw *hw, uint16_t vport, void *cfg_data __rte_unused,
-		struct zxdh_msg_reply_body *reply, uint16_t *res_len)
+		void *reply, uint16_t *res_len)
 {
 	char str[ZXDH_MSG_REPLY_BODY_MAX_LEN] = "rss_table";
-	struct zxdh_rss_reta *rss_reta = &reply->rss_reta;
+	void *rss_reta_msg_addr = ZXDH_ADDR_OF(msg_reply_body, reply, rss_reta_msg);
+	struct zxdh_rss_reta *rss_reta = (struct zxdh_rss_reta *)rss_reta_msg_addr;
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, reply, reply_data);
 	int ret = 0;
 
 	ret = zxdh_rss_table_get(hw, vport, rss_reta);
 	if (ret)
 		sprintf(str, "set rss reta tbl failed, code:%d", ret);
 
-	*res_len = strlen(str) + sizeof(enum zxdh_reps_flag);
-	reply->flag = (ret == 0) ? ZXDH_REPS_SUCC : ZXDH_REPS_FAIL;
-	memcpy(&reply->reply_data, str, strlen(str) + 1);
+	*res_len = strlen(str) + sizeof(uint8_t);
+	if (ret == 0)
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_SUCC);
+	else
+		ZXDH_SET(msg_reply_body, reply, flag, ZXDH_REPS_FAIL);
+	memcpy(reply_data_addr, str, strlen(str) + 1);
 	return ret;
 }
 
 static int
 zxdh_vf_port_attr_set(struct zxdh_hw *pf_hw, uint16_t vport, void *cfg_data,
-	struct zxdh_msg_reply_body *res_info, uint16_t *res_len)
+	void *res_info, uint16_t *res_len)
 {
 	RTE_ASSERT(!cfg_data || !pf_hw);
 	if (res_info)
@@ -1734,7 +1762,7 @@ zxdh_vf_port_attr_set(struct zxdh_hw *pf_hw, uint16_t vport, void *cfg_data,
 
 static int
 zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
-		void *cfg_data, struct zxdh_msg_reply_body *res_info,
+		void *cfg_data, void *res_info,
 		uint16_t *res_len)
 {
 	struct zxdh_np_stats_updata_msg *np_stats_query =
@@ -1745,6 +1773,47 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 	uint32_t idx = 0;
 	int ret = 0;
 
+	void *hw_stats_addr = ZXDH_ADDR_OF(msg_reply_body, res_info, hw_stats);
+	void *tx_unicast_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_unicast_pkts);
+	void *rx_unicast_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_unicast_pkts);
+	void *tx_unicast_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_unicast_bytes);
+	void *rx_unicast_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_unicast_bytes);
+	void *tx_multicast_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_multicast_pkts);
+	void *rx_multicast_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_multicast_pkts);
+	void *tx_multicast_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_multicast_bytes);
+	void *rx_multicast_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_multicast_bytes);
+	void *tx_broadcast_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_broadcast_pkts);
+	void *tx_broadcast_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_broadcast_bytes);
+	void *rx_broadcast_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_broadcast_pkts);
+	void *rx_broadcast_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_broadcast_bytes);
+	void *tx_mtu_drop_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_mtu_drop_pkts);
+	void *tx_mtu_drop_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_mtu_drop_bytes);
+	void *rx_mtu_drop_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_mtu_drop_pkts);
+	void *rx_mtu_drop_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_mtu_drop_bytes);
+	void *tx_mtr_drop_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_mtr_drop_pkts);
+	void *tx_mtr_drop_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, tx_mtr_drop_bytes);
+	void *rx_mtr_drop_pkts_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_mtr_drop_pkts);
+	void *rx_mtr_drop_bytes_addr =
+		ZXDH_ADDR_OF(hw_np_stats, hw_stats_addr, rx_mtr_drop_bytes);
 	if (!res_len || !res_info) {
 		PMD_DRV_LOG(ERR, "get stat invalid inparams");
 		return -1;
@@ -1760,8 +1829,8 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_unicast_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_unicast_bytes);
+	zxdh_data_hi_to_lo(tx_unicast_pkts_addr);
+	zxdh_data_hi_to_lo(tx_unicast_bytes_addr);
 
 	idx = zxdh_vport_to_vfid(vport_num) + ZXDH_UNICAST_STATS_INGRESS_BASE;
 	memset(&stats_data, 0, sizeof(stats_data));
@@ -1771,8 +1840,8 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_unicast_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_unicast_bytes);
+	zxdh_data_hi_to_lo(rx_unicast_pkts_addr);
+	zxdh_data_hi_to_lo(rx_unicast_bytes_addr);
 
 	idx = zxdh_vport_to_vfid(vport_num) + ZXDH_MULTICAST_STATS_EGRESS_BASE;
 	ret = zxdh_np_dtb_stats_get(pf_hw->dev_id, pf_hw->dev_sd->dtb_sd.queueid,
@@ -1781,8 +1850,8 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_multicast_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_multicast_bytes);
+	zxdh_data_hi_to_lo(tx_multicast_pkts_addr);
+	zxdh_data_hi_to_lo(tx_multicast_bytes_addr);
 
 	idx = zxdh_vport_to_vfid(vport_num) + ZXDH_MULTICAST_STATS_INGRESS_BASE;
 	memset(&stats_data, 0, sizeof(stats_data));
@@ -1792,8 +1861,8 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_multicast_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_multicast_bytes);
+	zxdh_data_hi_to_lo(rx_multicast_pkts_addr);
+	zxdh_data_hi_to_lo(rx_multicast_bytes_addr);
 
 	idx = zxdh_vport_to_vfid(vport_num) + ZXDH_BROAD_STATS_EGRESS_BASE;
 	ret = zxdh_np_dtb_stats_get(pf_hw->dev_id, pf_hw->dev_sd->dtb_sd.queueid,
@@ -1802,8 +1871,8 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_broadcast_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_broadcast_bytes);
+	zxdh_data_hi_to_lo(tx_broadcast_pkts_addr);
+	zxdh_data_hi_to_lo(tx_broadcast_bytes_addr);
 
 	idx = zxdh_vport_to_vfid(vport_num) + ZXDH_BROAD_STATS_INGRESS_BASE;
 	memset(&stats_data, 0, sizeof(stats_data));
@@ -1813,8 +1882,8 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_broadcast_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_broadcast_bytes);
+	zxdh_data_hi_to_lo(rx_broadcast_pkts_addr);
+	zxdh_data_hi_to_lo(rx_broadcast_bytes_addr);
 
 	idx = zxdh_vport_to_vfid(vport_num) + ZXDH_MTU_STATS_EGRESS_BASE;
 	memset(&stats_data, 0, sizeof(stats_data));
@@ -1824,10 +1893,10 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	res_info->np_stats.tx_mtu_drop_pkts = stats_data.n_pkts_dropped;
-	res_info->np_stats.tx_mtu_drop_bytes = stats_data.n_bytes_dropped;
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_mtu_drop_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_mtu_drop_bytes);
+	ZXDH_SET(hw_np_stats, hw_stats_addr, tx_mtu_drop_pkts, stats_data.n_pkts_dropped);
+	ZXDH_SET(hw_np_stats, hw_stats_addr, tx_mtu_drop_bytes, stats_data.n_bytes_dropped);
+	zxdh_data_hi_to_lo(tx_mtu_drop_pkts_addr);
+	zxdh_data_hi_to_lo(tx_mtu_drop_bytes_addr);
 
 	idx = zxdh_vport_to_vfid(vport_num) + ZXDH_MTU_STATS_INGRESS_BASE;
 	memset(&stats_data, 0, sizeof(stats_data));
@@ -1837,10 +1906,10 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	res_info->np_stats.rx_mtu_drop_pkts = stats_data.n_pkts_dropped;
-	res_info->np_stats.rx_mtu_drop_bytes = stats_data.n_bytes_dropped;
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_mtu_drop_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_mtu_drop_bytes);
+	ZXDH_SET(hw_np_stats, hw_stats_addr, rx_mtu_drop_pkts, stats_data.n_pkts_dropped);
+	ZXDH_SET(hw_np_stats, hw_stats_addr, rx_mtu_drop_bytes, stats_data.n_bytes_dropped);
+	zxdh_data_hi_to_lo(rx_mtu_drop_pkts_addr);
+	zxdh_data_hi_to_lo(rx_mtu_drop_bytes_addr);
 
 	idx = zxdh_vport_to_vfid(vport_num) + ZXDH_MTR_STATS_EGRESS_BASE;
 	memset(&stats_data, 0, sizeof(stats_data));
@@ -1850,10 +1919,11 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	res_info->np_stats.tx_mtr_drop_pkts = stats_data.n_pkts_dropped;
-	res_info->np_stats.tx_mtr_drop_bytes = stats_data.n_bytes_dropped;
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_mtr_drop_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.tx_mtr_drop_bytes);
+	ZXDH_SET(hw_np_stats, hw_stats_addr, tx_mtr_drop_pkts, stats_data.n_pkts_dropped);
+	ZXDH_SET(hw_np_stats, hw_stats_addr, tx_mtr_drop_bytes, stats_data.n_bytes_dropped);
+
+	zxdh_data_hi_to_lo(tx_mtr_drop_pkts_addr);
+	zxdh_data_hi_to_lo(tx_mtr_drop_bytes_addr);
 
 	idx = zxdh_vport_to_vfid(vport_num) + ZXDH_MTR_STATS_INGRESS_BASE;
 	memset(&stats_data, 0, sizeof(stats_data));
@@ -1863,10 +1933,11 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 		PMD_DRV_LOG(ERR, "get stats failed. code:%d", ret);
 		return ret;
 	}
-	res_info->np_stats.rx_mtr_drop_pkts = stats_data.n_pkts_dropped;
-	res_info->np_stats.rx_mtr_drop_bytes = stats_data.n_bytes_dropped;
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_mtr_drop_pkts);
-	zxdh_data_hi_to_lo(&res_info->np_stats.rx_mtr_drop_bytes);
+	ZXDH_SET(hw_np_stats, hw_stats_addr, rx_mtr_drop_pkts, stats_data.n_pkts_dropped);
+	ZXDH_SET(hw_np_stats, hw_stats_addr, rx_mtr_drop_bytes, stats_data.n_bytes_dropped);
+
+	zxdh_data_hi_to_lo(rx_mtr_drop_pkts_addr);
+	zxdh_data_hi_to_lo(rx_mtr_drop_bytes_addr);
 	*res_len = sizeof(struct zxdh_hw_np_stats);
 
 	return 0;
@@ -1875,12 +1946,13 @@ zxdh_vf_np_stats_update(struct zxdh_hw *pf_hw, uint16_t vport,
 static int
 zxdh_vf_mtr_hw_stats_get(struct zxdh_hw *pf_hw,
 	uint16_t vport, void *cfg_data,
-	struct zxdh_msg_reply_body *res_info,
+	void *res_info,
 	uint16_t *res_len)
 {
 	struct zxdh_mtr_stats_query  *zxdh_mtr_stats_query =
 			(struct zxdh_mtr_stats_query  *)cfg_data;
 	union zxdh_virport_num v_port = {.vport = vport};
+	uint8_t *hw_mtr_stats_addr = ZXDH_ADDR_OF(msg_reply_body, res_info, hw_mtr_stats);
 	int ret = 0;
 
 	uint32_t stat_baseaddr = zxdh_mtr_stats_query->direction ==
@@ -1892,14 +1964,14 @@ zxdh_vf_mtr_hw_stats_get(struct zxdh_hw *pf_hw,
 		PMD_DRV_LOG(ERR, "get stat invalid in params");
 		return -1;
 	}
-	res_info->flag = ZXDH_REPS_FAIL;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_FAIL);
 	ret = zxdh_np_dtb_stats_get(pf_hw->dev_id, pf_hw->dev_sd->dtb_sd.queueid,
-				1, idx, (uint32_t *)&res_info->hw_mtr_stats);
+				1, idx, (uint32_t *)hw_mtr_stats_addr);
 	if (ret) {
 		PMD_DRV_LOG(ERR, "get dir %d stats  failed", zxdh_mtr_stats_query->direction);
 		return ret;
 	}
-	res_info->flag = ZXDH_REPS_SUCC;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_SUCC);
 	*res_len = sizeof(struct zxdh_hw_mtr_stats);
 	return 0;
 }
@@ -1908,7 +1980,7 @@ static int
 zxdh_vf_mtr_hw_profile_add(struct zxdh_hw *pf_hw,
 	uint16_t vport,
 	void *cfg_data,
-	struct zxdh_msg_reply_body *res_info,
+	void *res_info,
 	uint16_t *res_len)
 {
 	if (!cfg_data || !res_len || !res_info) {
@@ -1917,16 +1989,18 @@ zxdh_vf_mtr_hw_profile_add(struct zxdh_hw *pf_hw,
 	}
 	struct rte_mtr_error error = {0};
 	int ret = 0;
-	uint64_t profile_id = HW_PROFILE_MAX;
+	uint64_t hw_profile_id = HW_PROFILE_MAX;
+	void *mtr_profile_info_addr = ZXDH_ADDR_OF(msg_reply_body, res_info, mtr_profile_info);
 
 	struct zxdh_plcr_profile_add  *zxdh_plcr_profile_add =
 		(struct zxdh_plcr_profile_add *)cfg_data;
 
-	res_info->flag = ZXDH_REPS_FAIL;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_FAIL);
+
 	*res_len = sizeof(struct zxdh_mtr_profile_info);
 	ret = zxdh_hw_profile_alloc_direct(pf_hw->eth_dev,
 		zxdh_plcr_profile_add->car_type,
-		&profile_id, &error);
+		&hw_profile_id, &error);
 
 	if (ret) {
 		PMD_DRV_LOG(ERR, "pf 0x%x for vf 0x%x alloc hw profile failed",
@@ -1935,9 +2009,9 @@ zxdh_vf_mtr_hw_profile_add(struct zxdh_hw *pf_hw,
 		);
 		return -1;
 	}
-	zxdh_hw_profile_ref(profile_id);
-	res_info->mtr_profile_info.profile_id = profile_id;
-	res_info->flag = ZXDH_REPS_SUCC;
+	zxdh_hw_profile_ref(hw_profile_id);
+	ZXDH_SET(mtr_profile_info, mtr_profile_info_addr, profile_id, hw_profile_id);
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_SUCC);
 
 	return 0;
 }
@@ -1946,7 +2020,7 @@ static int
 zxdh_vf_mtr_hw_profile_del(struct zxdh_hw *pf_hw,
 	uint16_t vport,
 	void *cfg_data,
-	struct zxdh_msg_reply_body *res_info,
+	void *res_info,
 	uint16_t *res_len)
 {
 	if (!cfg_data || !res_len || !res_info) {
@@ -1954,7 +2028,7 @@ zxdh_vf_mtr_hw_profile_del(struct zxdh_hw *pf_hw,
 		return -1;
 	}
 
-	res_info->flag = ZXDH_REPS_FAIL;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_FAIL);
 	*res_len = 0;
 	struct zxdh_plcr_profile_free *mtr_profile_free = (struct zxdh_plcr_profile_free *)cfg_data;
 	uint64_t profile_id = mtr_profile_free->profile_id;
@@ -1980,7 +2054,7 @@ zxdh_vf_mtr_hw_profile_del(struct zxdh_hw *pf_hw,
 				RTE_MTR_ERROR_TYPE_METER_PROFILE_ID, NULL,
 				"Meter offload del profile failed ");
 	}
-	res_info->flag = ZXDH_REPS_SUCC;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_SUCC);
 	return 0;
 }
 
@@ -1988,7 +2062,7 @@ static int
 zxdh_vf_mtr_hw_plcrflow_cfg(struct zxdh_hw *pf_hw,
 	uint16_t vport,
 	void *cfg_data,
-	struct zxdh_msg_reply_body *res_info,
+	void *res_info,
 	uint16_t *res_len)
 {
 	int ret = 0;
@@ -2000,7 +2074,7 @@ zxdh_vf_mtr_hw_plcrflow_cfg(struct zxdh_hw *pf_hw,
 	struct rte_mtr_error error = {0};
 	struct zxdh_plcr_flow_cfg *zxdh_plcr_flow_cfg = (struct zxdh_plcr_flow_cfg *)cfg_data;
 
-	res_info->flag = ZXDH_REPS_FAIL;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_FAIL);
 	*res_len = 0;
 	ret = zxdh_np_stat_car_queue_cfg_set(pf_hw->dev_id,
 		zxdh_plcr_flow_cfg->car_type,
@@ -2019,7 +2093,7 @@ zxdh_vf_mtr_hw_plcrflow_cfg(struct zxdh_hw *pf_hw,
 				RTE_MTR_ERROR_TYPE_MTR_PARAMS,
 				NULL, "Failed to bind plcr flow.");
 	}
-	res_info->flag = ZXDH_REPS_SUCC;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_SUCC);
 	return 0;
 }
 
@@ -2027,7 +2101,7 @@ static int
 zxdh_vf_mtr_hw_profile_cfg(struct zxdh_hw *pf_hw __rte_unused,
 	uint16_t vport,
 	void *cfg_data,
-	struct zxdh_msg_reply_body *res_info,
+	void *res_info,
 	uint16_t *res_len)
 {
 	int ret = 0;
@@ -2036,14 +2110,15 @@ zxdh_vf_mtr_hw_profile_cfg(struct zxdh_hw *pf_hw __rte_unused,
 		PMD_DRV_LOG(ERR, " cfg profile invalid inparams");
 		return -1;
 	}
-	res_info->flag = ZXDH_REPS_FAIL;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_FAIL);
 	*res_len = 0;
 	struct rte_mtr_error error = {0};
 	struct zxdh_plcr_profile_cfg *zxdh_plcr_profile_cfg =
 		(struct zxdh_plcr_profile_cfg *)cfg_data;
 	union zxdh_offload_profile_cfg *plcr_param = &zxdh_plcr_profile_cfg->plcr_param;
 
-	ret = zxdh_np_car_profile_cfg_set(vport,
+	ret = zxdh_np_car_profile_cfg_set(pf_hw->dev_id,
+		vport,
 		zxdh_plcr_profile_cfg->car_type,
 		zxdh_plcr_profile_cfg->packet_mode,
 		zxdh_plcr_profile_cfg->hw_profile_id,
@@ -2052,7 +2127,7 @@ zxdh_vf_mtr_hw_profile_cfg(struct zxdh_hw *pf_hw __rte_unused,
 		PMD_DRV_LOG(ERR, "(vport %d)config hw profilefailed", vport);
 		return -rte_mtr_error_set(&error, ENOTSUP, RTE_MTR_ERROR_TYPE_METER_PROFILE, NULL, "Meter offload cfg profile failed");
 	}
-	res_info->flag = ZXDH_REPS_SUCC;
+	ZXDH_SET(msg_reply_body, res_info, flag, ZXDH_REPS_SUCC);
 	return 0;
 }
 
@@ -2083,7 +2158,7 @@ static const zxdh_msg_process_callback zxdh_proc_cb[] = {
 
 static inline int
 zxdh_config_process_callback(struct zxdh_hw *hw, struct zxdh_msg_info *msg_info,
-	struct zxdh_msg_reply_body *res, uint16_t *res_len)
+	void *res, uint16_t *res_len)
 {
 	struct zxdh_msg_head *msghead = &msg_info->msg_head;
 	int ret = -1;
@@ -2096,13 +2171,13 @@ zxdh_config_process_callback(struct zxdh_hw *hw, struct zxdh_msg_info *msg_info,
 		ret = zxdh_proc_cb[msghead->msg_type](hw, msghead->vport,
 					(void *)&msg_info->data, res, res_len);
 		if (!ret)
-			res->flag = ZXDH_REPS_SUCC;
+			ZXDH_SET(msg_reply_body, res, flag, ZXDH_REPS_SUCC);
 		else
-			res->flag = ZXDH_REPS_FAIL;
+			ZXDH_SET(msg_reply_body, res, flag, ZXDH_REPS_FAIL);
 	} else {
-		res->flag = ZXDH_REPS_INVALID;
+		ZXDH_SET(msg_reply_body, res, flag, ZXDH_REPS_INVALID);
 	}
-	*res_len += sizeof(res->flag);
+	*res_len += sizeof(uint8_t);
 	return ret;
 }
 
@@ -2111,7 +2186,7 @@ pf_recv_bar_msg(void *pay_load, uint16_t len, void *reps_buffer,
 	uint16_t *reps_len, void *eth_dev)
 {
 	struct zxdh_msg_info *msg_info = (struct zxdh_msg_info *)pay_load;
-	struct zxdh_msg_reply_body *reply_body = reps_buffer;
+	void *reply_data_addr = ZXDH_ADDR_OF(msg_reply_body, reps_buffer, reply_data);
 	struct rte_eth_dev *dev = (struct rte_eth_dev *)eth_dev;
 	int32_t ret = 0;
 	struct zxdh_hw *hw;
@@ -2131,14 +2206,14 @@ pf_recv_bar_msg(void *pay_load, uint16_t len, void *reps_buffer,
 		goto msg_proc_end;
 	}
 
-	ret = zxdh_config_process_callback(hw, msg_info, reply_body, &reply_len);
-	*reps_len = reply_len + sizeof(struct zxdh_msg_reply_head);
+	ret = zxdh_config_process_callback(hw, msg_info, reps_buffer, &reply_len);
+	*reps_len = reply_len + ZXDH_ST_SZ_BYTES(msg_reply_head);
 	return ret;
 
 msg_proc_end:
-	memcpy(reply_body->reply_data, &ret, sizeof(ret));
+	memcpy(reply_data_addr, &ret, sizeof(ret));
 	reply_len = sizeof(ret);
-	*reps_len = sizeof(struct zxdh_msg_reply_head) + reply_len;
+	*reps_len = ZXDH_ST_SZ_BYTES(msg_reply_head) + reply_len;
 	return ret;
 }
 
diff --git a/drivers/net/zxdh/zxdh_msg.h b/drivers/net/zxdh/zxdh_msg.h
index 58836bb4b7..7dad6f7335 100644
--- a/drivers/net/zxdh/zxdh_msg.h
+++ b/drivers/net/zxdh/zxdh_msg.h
@@ -11,6 +11,7 @@
 
 #include "zxdh_ethdev_ops.h"
 #include "zxdh_mtr.h"
+#include "zxdh_common.h"
 
 #define ZXDH_BAR0_INDEX                 0
 #define ZXDH_CTRLCH_OFFSET              (0x2000)
@@ -46,8 +47,10 @@
 
 #define ZXDH_MSG_REPLYBODY_HEAD    sizeof(enum zxdh_reps_flag)
 #define ZXDH_MSG_HEADER_SIZE       4
-#define ZXDH_MSG_REPLY_BODY_MAX_LEN  \
-		(ZXDH_MSG_PAYLOAD_MAX_LEN - sizeof(struct zxdh_msg_reply_head))
+#define ZXDH_MSG_REPLY_BODY_MAX_LEN \
+		(ZXDH_MSG_PAYLOAD_MAX_LEN - ZXDH_MSG_HEADER_SIZE)
+#define ZXDH_MSG_REPLY_DATA \
+		(ZXDH_MSG_REPLY_BODY_MAX_LEN - ZXDH_MSG_REPLYBODY_HEAD)
 
 #define ZXDH_MSG_HEAD_LEN            8
 #define ZXDH_MSG_REQ_BODY_MAX_LEN  \
@@ -329,12 +332,6 @@ struct zxdh_offset_get_msg {
 	uint16_t type;
 };
 
-struct zxdh_msg_reply_head {
-	uint8_t flag;
-	uint16_t reps_len;
-	uint8_t resvd;
-};
-
 enum zxdh_reps_flag {
 	ZXDH_REPS_FAIL,
 	ZXDH_REPS_SUCC = 0xaa,
@@ -354,18 +351,39 @@ struct zxdh_link_info_msg {
 	uint32_t speed;
 };
 
+struct zxdh_ifc_link_info_msg_bits {
+	uint8_t autoneg[0x8];
+	uint8_t link_state[0x8];
+	uint8_t blink_enable[0x8];
+	uint8_t duplex[0x8];
+	uint8_t speed_modes[0x20];
+	uint8_t speed[0x20];
+};
+
 struct zxdh_rss_reta {
 	uint32_t reta[RTE_ETH_RSS_RETA_SIZE_256];
 };
 
+struct zxdh_ifc_rss_reta_bits {
+	uint32_t reta[RTE_ETH_RSS_RETA_SIZE_256 * 8];
+};
+
 struct zxdh_rss_hf {
 	uint32_t rss_hf;
 };
 
+struct zxdh_ifc_rss_hf_bits {
+	uint8_t rss_hf[0x20];
+};
+
 struct zxdh_mac_reply_msg {
 	uint8_t mac_flag;
 };
 
+struct zxdh_ifc_mac_reply_msg_bits {
+	uint8_t mac_flag[0x8];
+};
+
 struct zxdh_mac_module_eeprom_msg {
 	uint8_t i2c_addr;
 	uint8_t bank;
@@ -375,34 +393,57 @@ struct zxdh_mac_module_eeprom_msg {
 	uint8_t data[ZXDH_MODULE_EEPROM_DATA_LEN];
 };
 
+struct zxdh_ifc_agent_mac_module_eeprom_msg_bits {
+	uint8_t i2c_addr[0x8];
+	uint8_t bank[0x8];
+	uint8_t page[0x8];
+	uint8_t offset[0x8];
+	uint8_t length[0x8];
+	uint8_t data[ZXDH_MODULE_EEPROM_DATA_LEN * 8];
+};
+
 struct zxdh_flash_msg {
 	uint8_t firmware_version[ZXDH_FWVERS_LEN];
 };
 
+struct zxdh_ifc_agent_flash_msg_bits {
+	uint8_t firmware_version[0x100];
+};
+
 struct zxdh_mtr_profile_info {
 	uint64_t profile_id;
 };
 
-struct zxdh_msg_reply_body {
-	enum zxdh_reps_flag flag;
+struct zxdh_ifc_mtr_profile_info_bits {
+	uint8_t profile_id[0x40];
+};
+
+struct zxdh_ifc_msg_reply_body_bits {
+	uint8_t flag[0x8];
 	union {
-		uint8_t reply_data[ZXDH_MSG_REPLY_BODY_MAX_LEN - sizeof(enum zxdh_reps_flag)];
-		struct zxdh_hw_np_stats np_stats;
-		struct zxdh_link_info_msg link_msg;
-		struct zxdh_rss_reta rss_reta;
-		struct zxdh_rss_hf rss_hf;
-		struct zxdh_hw_vqm_stats vqm_stats;
-		struct zxdh_mac_reply_msg mac_reply_msg;
-		struct zxdh_flash_msg flash_msg;
-		struct zxdh_mac_module_eeprom_msg module_eeprom_msg;
-		struct zxdh_mtr_profile_info mtr_profile_info;
-		struct zxdh_mtr_stats hw_mtr_stats;
+		uint8_t reply_data[ZXDH_MSG_REPLY_DATA * 8];
+		struct zxdh_ifc_hw_np_stats_bits hw_stats;
+		struct zxdh_ifc_link_info_msg_bits link_msg;
+		struct zxdh_ifc_rss_reta_bits rss_reta_msg;
+		struct zxdh_ifc_rss_hf_bits rss_hf_msg;
+		struct zxdh_ifc_hw_vqm_stats_bits vqm_stats;
+		struct zxdh_ifc_mac_reply_msg_bits mac_reply_msg;
+		struct zxdh_ifc_agent_flash_msg_bits flash_msg;
+		struct zxdh_ifc_agent_mac_module_eeprom_msg_bits module_eeprom_msg;
+		struct zxdh_ifc_mtr_profile_info_bits  mtr_profile_info;
+		struct zxdh_ifc_mtr_stats_bits hw_mtr_stats;
 	};
 };
 
-struct zxdh_msg_reply_info {
-	struct zxdh_msg_reply_head reply_head;
-	struct zxdh_msg_reply_body reply_body;
+struct zxdh_ifc_msg_reply_head_bits {
+	uint8_t flag[0x8];
+	uint8_t reps_len[0x10];
+	uint8_t resvd[0x8];
+};
+
+struct zxdh_ifc_msg_reply_info_bits {
+	struct zxdh_ifc_msg_reply_head_bits reply_head;
+	struct zxdh_ifc_msg_reply_body_bits reply_body;
 };
 
 struct zxdh_vf_init_msg {
@@ -412,12 +453,12 @@ struct zxdh_vf_init_msg {
 	uint8_t rss_enable;
 };
 
-struct zxdh_msg_head {
-	enum zxdh_msg_type msg_type;
+struct __rte_packed_begin zxdh_msg_head {
+	uint8_t msg_type;
 	uint16_t  vport;
 	uint16_t  vf_id;
 	uint16_t pcieid;
-};
+} __rte_packed_end;
 
 struct zxdh_port_attr_set_msg {
 	uint32_t mode;
@@ -455,7 +496,7 @@ struct zxdh_rss_enable {
 };
 
 struct zxdh_agent_msg_head {
-	enum zxdh_agent_msg_type msg_type;
+	uint8_t msg_type;
 	uint8_t panel_id;
 	uint8_t phyport;
 	uint8_t rsv;
@@ -526,7 +567,7 @@ struct zxdh_msg_info {
 typedef int (*zxdh_bar_chan_msg_recv_callback)(void *pay_load, uint16_t len,
 		void *reps_buffer, uint16_t *reps_len, void *dev);
 typedef int (*zxdh_msg_process_callback)(struct zxdh_hw *hw, uint16_t vport, void *cfg_data,
-	struct zxdh_msg_reply_body *res_info, uint16_t *res_len);
+	void *res_info, uint16_t *res_len);
 
 typedef int (*zxdh_bar_chan_msg_recv_callback)(void *pay_load, uint16_t len,
 			void *reps_buffer, uint16_t *reps_len, void *dev);
diff --git a/drivers/net/zxdh/zxdh_mtr.c b/drivers/net/zxdh/zxdh_mtr.c
index 3797a5b29b..b23f6535a8 100644
--- a/drivers/net/zxdh/zxdh_mtr.c
+++ b/drivers/net/zxdh/zxdh_mtr.c
@@ -281,7 +281,7 @@ zxdh_hw_profile_free_direct(struct rte_eth_dev *dev, ZXDH_PROFILE_TYPE car_type,
 {
 	struct zxdh_hw *hw = dev->data->dev_private;
 	uint16_t vport = hw->vport.vport;
-	int ret = zxdh_np_car_profile_id_delete(vport, car_type,
+	int ret = zxdh_np_car_profile_id_delete(hw->dev_id, vport, car_type,
 			(uint64_t)hw_profile_id);
 	if (ret) {
 		PMD_DRV_LOG(ERR, "port %u free hw profile %u failed", vport, hw_profile_id);
@@ -299,7 +299,7 @@ zxdh_hw_profile_alloc_direct(struct rte_eth_dev *dev, ZXDH_PROFILE_TYPE car_type
 	uint64_t profile_id = HW_PROFILE_MAX;
 	struct zxdh_hw *hw = dev->data->dev_private;
 	uint16_t vport = hw->vport.vport;
-	int ret = zxdh_np_car_profile_id_add(vport, car_type, &profile_id);
+	int ret = zxdh_np_car_profile_id_add(hw->dev_id, vport, car_type, &profile_id);
 
 	if (ret) {
 		PMD_DRV_LOG(ERR, "port %u alloc hw profile failed", vport);
@@ -326,7 +326,7 @@ zxdh_hw_profile_free(struct rte_eth_dev *dev, uint8_t car_type,
 		ret = zxdh_hw_profile_free_direct(dev, car_type, (uint64_t)hw_profile_id, error);
 	} else {
 		struct zxdh_msg_info msg_info = {0};
-		struct zxdh_msg_reply_info reply_info = {0};
+		uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 		struct zxdh_plcr_profile_free *zxdh_plcr_profile_free =
 			&msg_info.data.zxdh_plcr_profile_free;
 
@@ -335,7 +335,7 @@ zxdh_hw_profile_free(struct rte_eth_dev *dev, uint8_t car_type,
 		zxdh_msg_head_build(hw, ZXDH_PLCR_CAR_PROFILE_ID_DELETE, &msg_info);
 		ret = zxdh_vf_send_msg_to_pf(dev, &msg_info,
 			ZXDH_MSG_HEAD_LEN + sizeof(struct zxdh_plcr_profile_free),
-			&reply_info, sizeof(struct zxdh_msg_reply_info));
+			zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info));
 
 		if (ret)
 			return -rte_mtr_error_set(error, ENOTSUP,
@@ -357,15 +357,19 @@ zxdh_hw_profile_alloc(struct rte_eth_dev *dev, uint64_t *hw_profile_id,
 		ret = zxdh_hw_profile_alloc_direct(dev, CAR_A, hw_profile_id, error);
 	} else {
 		struct zxdh_msg_info msg_info = {0};
-		struct zxdh_msg_reply_info reply_info = {0};
+		uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 		struct zxdh_plcr_profile_add  *zxdh_plcr_profile_add =
 			&msg_info.data.zxdh_plcr_profile_add;
+		void *reply_body_addr =
+			ZXDH_ADDR_OF(msg_reply_info, zxdh_msg_reply_info, reply_body);
+		void *mtr_profile_info_addr =
+			ZXDH_ADDR_OF(msg_reply_body, reply_body_addr, mtr_profile_info);
 
 		zxdh_plcr_profile_add->car_type = CAR_A;
 		zxdh_msg_head_build(hw, ZXDH_PLCR_CAR_PROFILE_ID_ADD, &msg_info);
 		ret = zxdh_vf_send_msg_to_pf(dev, &msg_info,
 			ZXDH_MSG_HEAD_LEN + sizeof(struct zxdh_plcr_profile_add),
-			&reply_info, sizeof(struct zxdh_msg_reply_info));
+			zxdh_msg_reply_info, ZXDH_ST_SZ_BYTES(msg_reply_info));
 
 		if (ret) {
 			PMD_DRV_LOG(ERR,
@@ -376,7 +380,7 @@ zxdh_hw_profile_alloc(struct rte_eth_dev *dev, uint64_t *hw_profile_id,
 					RTE_MTR_ERROR_TYPE_METER_PROFILE_ID, NULL,
 					"Meter offload alloc profile  id msg failed ");
 		}
-		*hw_profile_id = reply_info.reply_body.mtr_profile_info.profile_id;
+		*hw_profile_id = ZXDH_GET(mtr_profile_info, mtr_profile_info_addr, profile_id);
 		if (*hw_profile_id == ZXDH_HW_PROFILE_MAX) {
 			return -rte_mtr_error_set(error, ENOTSUP,
 					RTE_MTR_ERROR_TYPE_METER_PROFILE_ID, NULL,
@@ -432,7 +436,7 @@ zxdh_mtr_hw_counter_query(struct rte_eth_dev *dev,
 		}
 	} else { /* send msg to pf */
 		struct zxdh_msg_info msg_info = {0};
-		struct zxdh_msg_reply_info reply_info = {0};
+		uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 		struct zxdh_mtr_stats_query *zxdh_mtr_stats_query =
 				&msg_info.data.zxdh_mtr_stats_query;
 
@@ -442,8 +446,8 @@ zxdh_mtr_hw_counter_query(struct rte_eth_dev *dev,
 		ret = zxdh_vf_send_msg_to_pf(dev,
 			&msg_info,
 			sizeof(msg_info),
-			&reply_info,
-			sizeof(struct zxdh_msg_reply_info));
+			zxdh_msg_reply_info,
+			ZXDH_ST_SZ_BYTES(msg_reply_info));
 
 		if (ret) {
 			PMD_DRV_LOG(ERR,
@@ -451,7 +455,11 @@ zxdh_mtr_hw_counter_query(struct rte_eth_dev *dev,
 				hw->vport.vport);
 			return -rte_mtr_error_set(error, ENOTSUP, RTE_MTR_ERROR_TYPE_STATS, NULL, "Meter offload alloc profile failed");
 		}
-		struct zxdh_mtr_stats *hw_mtr_stats = &reply_info.reply_body.hw_mtr_stats;
+		void *reply_body_addr =
+			ZXDH_ADDR_OF(msg_reply_info, zxdh_msg_reply_info, reply_body);
+		void *hw_mtr_stats_addr =
+			ZXDH_ADDR_OF(msg_reply_body, reply_body_addr, hw_mtr_stats);
+		struct zxdh_mtr_stats *hw_mtr_stats = (struct zxdh_mtr_stats *)hw_mtr_stats_addr;
 
 		mtr_stats->n_bytes_dropped = hw_mtr_stats->n_bytes_dropped;
 		mtr_stats->n_pkts_dropped = hw_mtr_stats->n_pkts_dropped;
@@ -551,7 +559,9 @@ zxdh_hw_profile_config_direct(struct rte_eth_dev *dev __rte_unused,
 	struct zxdh_meter_profile *mp,
 	struct rte_mtr_error *error)
 {
-	int ret = zxdh_np_car_profile_cfg_set(mp->hw_profile_owner_vport,
+	struct zxdh_hw *hw = dev->data->dev_private;
+	int ret = zxdh_np_car_profile_cfg_set(hw->dev_id,
+		mp->hw_profile_owner_vport,
 		car_type, mp->profile.packet_mode,
 		(uint32_t)hw_profile_id, &mp->plcr_param);
 	if (ret) {
@@ -573,7 +583,7 @@ static int zxdh_hw_profile_config(struct rte_eth_dev *dev, uint16_t hw_profile_i
 		ret = zxdh_hw_profile_config_direct(dev, CAR_A, hw_profile_id, mp, error);
 	} else {
 		struct zxdh_msg_info msg_info = {0};
-		struct zxdh_msg_reply_info reply_info = {0};
+		uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 		struct zxdh_plcr_profile_cfg *zxdh_plcr_profile_cfg =
 			&msg_info.data.zxdh_plcr_profile_cfg;
 
@@ -588,9 +598,9 @@ static int zxdh_hw_profile_config(struct rte_eth_dev *dev, uint16_t hw_profile_i
 		ret = zxdh_vf_send_msg_to_pf(dev,
 			&msg_info,
 			ZXDH_MSG_HEAD_LEN + sizeof(struct zxdh_plcr_profile_cfg),
-			&reply_info,
-			sizeof(struct zxdh_msg_reply_info));
-
+			zxdh_msg_reply_info,
+			ZXDH_ST_SZ_BYTES(msg_reply_info)
+		);
 		if (ret) {
 			PMD_DRV_LOG(ERR,
 				"Failed msg: port 0x%x msg type ZXDH_PLCR_CAR_PROFILE_CFG_SET ",
@@ -874,7 +884,7 @@ zxdh_set_mtr_enable(struct rte_eth_dev *dev, uint8_t dir, bool enable, struct rt
 
 	if (priv->is_pf) {
 		ret = zxdh_get_port_attr(priv, priv->vport.vport, &port_attr);
-		port_attr.ingress_meter_enable = enable;
+		port_attr.egress_meter_enable = enable;
 		ret = zxdh_set_port_attr(priv, priv->vport.vport, &port_attr);
 		if (ret) {
 			PMD_DRV_LOG(ERR, "%s set port attr failed", __func__);
@@ -936,7 +946,7 @@ zxdh_hw_plcrflow_config(struct rte_eth_dev *dev, uint16_t hw_flow_id,
 		}
 	} else {
 		struct zxdh_msg_info msg_info = {0};
-		struct zxdh_msg_reply_info reply_info = {0};
+		uint8_t zxdh_msg_reply_info[ZXDH_ST_SZ_BYTES(msg_reply_info)] = {0};
 		struct zxdh_plcr_flow_cfg *zxdh_plcr_flow_cfg = &msg_info.data.zxdh_plcr_flow_cfg;
 
 		zxdh_plcr_flow_cfg->car_type = CAR_A;
@@ -947,8 +957,8 @@ zxdh_hw_plcrflow_config(struct rte_eth_dev *dev, uint16_t hw_flow_id,
 		zxdh_msg_head_build(hw, ZXDH_PLCR_CAR_QUEUE_CFG_SET, &msg_info);
 		ret = zxdh_vf_send_msg_to_pf(dev, &msg_info,
 			ZXDH_MSG_HEAD_LEN + sizeof(struct zxdh_plcr_flow_cfg),
-			&reply_info,
-			sizeof(struct zxdh_msg_reply_info));
+			zxdh_msg_reply_info,
+			ZXDH_ST_SZ_BYTES(msg_reply_info));
 		if (ret) {
 			PMD_DRV_LOG(ERR,
 				"Failed msg: port 0x%x msg type ZXDH_PLCR_CAR_QUEUE_CFG_SET ",
diff --git a/drivers/net/zxdh/zxdh_mtr.h b/drivers/net/zxdh/zxdh_mtr.h
index 749ceb7479..3efcb6b591 100644
--- a/drivers/net/zxdh/zxdh_mtr.h
+++ b/drivers/net/zxdh/zxdh_mtr.h
@@ -93,6 +93,11 @@ struct zxdh_mtr_stats {
 	uint64_t n_bytes_dropped;
 };
 
+struct zxdh_ifc_mtr_stats_bits {
+	uint8_t n_pkts_dropped[0x40];
+	uint8_t n_bytes_dropped[0x40];
+};
+
 struct zxdh_hw_mtr_stats {
 	uint32_t n_pkts_dropped_hi;
 	uint32_t n_pkts_dropped_lo;
diff --git a/drivers/net/zxdh/zxdh_np.c b/drivers/net/zxdh/zxdh_np.c
index ab8b3ae688..66902e7e92 100644
--- a/drivers/net/zxdh/zxdh_np.c
+++ b/drivers/net/zxdh/zxdh_np.c
@@ -2350,6 +2350,8 @@ zxdh_np_dev_add(uint32_t  dev_id, ZXDH_DEV_TYPE_E dev_type,
 
 	rte_spinlock_init(&p_dev_info->dtb_spinlock.spinlock);
 
+	rte_spinlock_init(&p_dev_info->smmu0_spinlock.spinlock);
+
 	for (i = 0; i < ZXDH_DTB_QUEUE_NUM_MAX; i++)
 		rte_spinlock_init(&p_dev_info->dtb_queue_spinlock[i].spinlock);
 
@@ -3391,6 +3393,32 @@ zxdh_np_reg_read(uint32_t dev_id, uint32_t reg_no,
 	return rc;
 }
 
+static uint32_t
+zxdh_np_reg_read32(uint32_t dev_id, uint32_t reg_no,
+	uint32_t m_offset, uint32_t n_offset, uint32_t *p_data)
+{
+	uint32_t rc = 0;
+	uint32_t addr = 0;
+	ZXDH_REG_T *p_reg_info = &g_dpp_reg_info[reg_no];
+	uint32_t p_buff[ZXDH_REG_DATA_MAX] = {0};
+	uint32_t reg_real_no = p_reg_info->reg_no;
+	uint32_t reg_type = p_reg_info->flags;
+	uint32_t reg_module = p_reg_info->module_no;
+
+	addr = zxdh_np_reg_get_reg_addr(reg_no, m_offset, n_offset);
+
+	if (reg_module == DTB4K) {
+		rc = p_reg_info->p_read_fun(dev_id, addr, p_data);
+		ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, "p_reg_info->p_read_fun");
+	} else {
+		rc = zxdh_np_agent_channel_reg_read(dev_id, reg_type, reg_real_no, 4, addr, p_buff);
+		ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, "zxdh_np_agent_channel_reg_read");
+		*p_data = p_buff[0];
+	}
+
+	return rc;
+}
+
 static uint32_t
 zxdh_np_dtb_queue_vm_info_get(uint32_t dev_id,
 		uint32_t queue_id,
@@ -10542,9 +10570,9 @@ zxdh_np_se_done_status_check(uint32_t dev_id, uint32_t reg_no, uint32_t pos)
 	uint32_t done_flag = 0;
 
 	while (!done_flag) {
-		rc = zxdh_np_reg_read(dev_id, reg_no, 0, 0, &data);
+		rc = zxdh_np_reg_read32(dev_id, reg_no, 0, 0, &data);
 		if (rc != 0) {
-			PMD_DRV_LOG(ERR, "reg_read fail!");
+			PMD_DRV_LOG(ERR, "reg_read32 fail!");
 			return rc;
 		}
 
@@ -10577,10 +10605,17 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 	uint32_t temp_data[4] = {0};
 	uint32_t *p_temp_data = NULL;
 	ZXDH_SMMU0_SMMU0_CPU_IND_CMD_T cpu_ind_cmd = {0};
+	ZXDH_SPINLOCK_T *p_ind_spinlock = NULL;
+
+	rc = zxdh_np_dev_opr_spinlock_get(dev_id, ZXDH_DEV_SPINLOCK_T_SMMU0, &p_ind_spinlock);
+	ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, "zxdh_np_dev_opr_spinlock_get");
+
+	rte_spinlock_lock(&p_ind_spinlock->spinlock);
 
 	rc = zxdh_np_se_done_status_check(dev_id, ZXDH_SMMU0_SMMU0_WR_ARB_CPU_RDYR, 0);
 	if (rc != ZXDH_OK) {
 		PMD_DRV_LOG(ERR, "se done status check failed, rc=0x%x.", rc);
+		rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 		return ZXDH_ERR;
 	}
 
@@ -10592,11 +10627,13 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 		switch (rd_mode) {
 		case ZXDH_ERAM128_OPR_128b:
 			if ((0xFFFFFFFF - (base_addr)) < (index)) {
+				rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 				PMD_DRV_LOG(ERR, "index 0x%x is invalid!", index);
 				return ZXDH_PAR_CHK_INVALID_INDEX;
 			}
 			if (base_addr + index > ZXDH_SE_SMMU0_ERAM_ADDR_NUM_TOTAL - 1) {
 				PMD_DRV_LOG(ERR, "index out of range!");
+				rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 				return ZXDH_ERR;
 			}
 			row_index = (index << 7) & ZXDH_ERAM128_BADDR_MASK;
@@ -10604,6 +10641,7 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 		case ZXDH_ERAM128_OPR_64b:
 			if ((base_addr + (index >> 1)) > ZXDH_SE_SMMU0_ERAM_ADDR_NUM_TOTAL - 1) {
 				PMD_DRV_LOG(ERR, "index out of range!");
+				rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 				return ZXDH_ERR;
 			}
 			row_index = (index << 6) & ZXDH_ERAM128_BADDR_MASK;
@@ -10612,6 +10650,7 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 		case ZXDH_ERAM128_OPR_32b:
 			if ((base_addr + (index >> 2)) > ZXDH_SE_SMMU0_ERAM_ADDR_NUM_TOTAL - 1) {
 				PMD_DRV_LOG(ERR, "index out of range!");
+				rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 				return ZXDH_ERR;
 			}
 			row_index = (index << 5) & ZXDH_ERAM128_BADDR_MASK;
@@ -10620,6 +10659,7 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 		case ZXDH_ERAM128_OPR_1b:
 			if ((base_addr + (index >> 7)) > ZXDH_SE_SMMU0_ERAM_ADDR_NUM_TOTAL - 1) {
 				PMD_DRV_LOG(ERR, "index out of range!");
+				rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 				return ZXDH_ERR;
 			}
 			row_index = index & ZXDH_ERAM128_BADDR_MASK;
@@ -10638,10 +10678,12 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 		case ZXDH_ERAM128_OPR_128b:
 			if ((0xFFFFFFFF - (base_addr)) < (index)) {
 				PMD_DRV_LOG(ERR, "index 0x%x is invalid!", index);
+				rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 				return ZXDH_PAR_CHK_INVALID_INDEX;
 			}
 			if (base_addr + index > ZXDH_SE_SMMU0_ERAM_ADDR_NUM_TOTAL - 1) {
 				PMD_DRV_LOG(ERR, "index out of range!");
+				rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 				return ZXDH_ERR;
 			}
 			row_index = (index << 7);
@@ -10650,6 +10692,7 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 		case ZXDH_ERAM128_OPR_64b:
 			if ((base_addr + (index >> 1)) > ZXDH_SE_SMMU0_ERAM_ADDR_NUM_TOTAL - 1) {
 				PMD_DRV_LOG(ERR, "index out of range!");
+				rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 				return ZXDH_ERR;
 			}
 			row_index = (index << 6);
@@ -10658,6 +10701,7 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 		case ZXDH_ERAM128_OPR_32b:
 			if ((base_addr + (index >> 2)) > ZXDH_SE_SMMU0_ERAM_ADDR_NUM_TOTAL - 1) {
 				PMD_DRV_LOG(ERR, "index out of range!");
+				rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 				return ZXDH_ERR;
 			}
 			row_index = (index << 5);
@@ -10665,7 +10709,8 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 			break;
 		case ZXDH_ERAM128_OPR_1b:
 			PMD_DRV_LOG(ERR, "rd_clr_mode[%u] or rd_mode[%u] error!",
-			rd_clr_mode, rd_mode);
+				rd_clr_mode, rd_mode);
+			rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 			return ZXDH_ERR;
 		default:
 			break;
@@ -10680,12 +10725,14 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 			&cpu_ind_cmd);
 	if (rc != ZXDH_OK) {
 		PMD_DRV_LOG(ERR, "zxdh_np_reg_write failed, rc=0x%x.", rc);
+		rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 		return ZXDH_ERR;
 	}
 
 	rc = zxdh_np_se_done_status_check(dev_id, ZXDH_SMMU0_SMMU0_CPU_IND_RD_DONER, 0);
 	if (rc != ZXDH_OK) {
 		PMD_DRV_LOG(ERR, "se done status check failed, rc=0x%x.", rc);
+		rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 		return ZXDH_ERR;
 	}
 
@@ -10698,6 +10745,7 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 			p_temp_data + 3 - i);
 		if (rc != ZXDH_OK) {
 			PMD_DRV_LOG(ERR, "zxdh_np_reg_write failed, rc=0x%x.", rc);
+			rte_spinlock_unlock(&p_ind_spinlock->spinlock);
 			return ZXDH_ERR;
 		}
 	}
@@ -10736,6 +10784,8 @@ zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
 		}
 	}
 
+	rte_spinlock_unlock(&p_ind_spinlock->spinlock);
+
 	return rc;
 }
 
@@ -10789,7 +10839,7 @@ zxdh_np_agent_channel_plcr_sync_send(uint32_t dev_id, ZXDH_AGENT_CHANNEL_PLCR_MS
 {
 	uint32_t ret = 0;
 	ZXDH_AGENT_CHANNEL_MSG_T agent_msg = {
-		.msg = (void *)&p_msg,
+		.msg = (void *)p_msg,
 		.msg_len = sizeof(ZXDH_AGENT_CHANNEL_PLCR_MSG_T),
 	};
 
@@ -11012,7 +11062,8 @@ zxdh_np_stat_carc_queue_cfg_set(uint32_t dev_id,
 }
 
 uint32_t
-zxdh_np_car_profile_id_add(uint32_t vport_id,
+zxdh_np_car_profile_id_add(uint32_t dev_id,
+		uint32_t vport_id,
 		uint32_t flags,
 		uint64_t *p_profile_id)
 {
@@ -11027,7 +11078,7 @@ zxdh_np_car_profile_id_add(uint32_t vport_id,
 		PMD_DRV_LOG(ERR, "profile_id point null!");
 		return ZXDH_PAR_CHK_POINT_NULL;
 	}
-	ret = zxdh_np_agent_channel_plcr_profileid_request(0, vport_id, flags, profile_id);
+	ret = zxdh_np_agent_channel_plcr_profileid_request(dev_id, vport_id, flags, profile_id);
 
 	profile_id_h = *(profile_id + 1);
 	profile_id_l = *profile_id;
@@ -11045,14 +11096,14 @@ zxdh_np_car_profile_id_add(uint32_t vport_id,
 }
 
 uint32_t
-zxdh_np_car_profile_cfg_set(uint32_t vport_id __rte_unused,
+zxdh_np_car_profile_cfg_set(uint32_t dev_id,
+		uint32_t vport_id __rte_unused,
 		uint32_t car_type,
 		uint32_t pkt_sign,
 		uint32_t profile_id,
 		void *p_car_profile_cfg)
 {
 	uint32_t ret = 0;
-	uint32_t dev_id = 0;
 
 	ret = zxdh_np_agent_channel_plcr_car_rate(dev_id, car_type,
 		pkt_sign, profile_id, p_car_profile_cfg);
@@ -11065,11 +11116,10 @@ zxdh_np_car_profile_cfg_set(uint32_t vport_id __rte_unused,
 }
 
 uint32_t
-zxdh_np_car_profile_id_delete(uint32_t vport_id,
+zxdh_np_car_profile_id_delete(uint32_t dev_id, uint32_t vport_id,
 	uint32_t flags, uint64_t profile_id)
 {
 	uint32_t ret = 0;
-	uint32_t dev_id = 0;
 	uint32_t profileid = profile_id & 0xFFFF;
 
 	ret = zxdh_np_agent_channel_plcr_profileid_release(dev_id, vport_id, flags, profileid);
diff --git a/drivers/net/zxdh/zxdh_np.h b/drivers/net/zxdh/zxdh_np.h
index b1d8b1aef8..1b8f17474d 100644
--- a/drivers/net/zxdh/zxdh_np.h
+++ b/drivers/net/zxdh/zxdh_np.h
@@ -1934,15 +1934,17 @@ uint32_t zxdh_np_stat_ppu_cnt_get_ex(uint32_t dev_id,
 			uint32_t clr_mode,
 			uint32_t *p_data);
 uint32_t
-zxdh_np_car_profile_id_add(uint32_t vport_id,
+zxdh_np_car_profile_id_add(uint32_t dev_id,
+			uint32_t vport_id,
 			uint32_t flags,
 			uint64_t *p_profile_id);
-uint32_t zxdh_np_car_profile_cfg_set(uint32_t vport_id,
+uint32_t zxdh_np_car_profile_cfg_set(uint32_t dev_id,
+			uint32_t vport_id,
 			uint32_t car_type,
 			uint32_t pkt_sign,
 			uint32_t profile_id,
 			void *p_car_profile_cfg);
-uint32_t zxdh_np_car_profile_id_delete(uint32_t vport_id,
+uint32_t zxdh_np_car_profile_id_delete(uint32_t dev_id, uint32_t vport_id,
 			uint32_t flags, uint64_t profile_id);
 uint32_t zxdh_np_stat_car_queue_cfg_set(uint32_t dev_id,
 			uint32_t car_type,
diff --git a/drivers/net/zxdh/zxdh_pci.c b/drivers/net/zxdh/zxdh_pci.c
index 3d1a3ff0dd..4ff0f065df 100644
--- a/drivers/net/zxdh/zxdh_pci.c
+++ b/drivers/net/zxdh/zxdh_pci.c
@@ -27,6 +27,23 @@
 		1ULL << ZXDH_F_NOTIFICATION_DATA | \
 		1ULL << ZXDH_NET_F_MAC)
 
+#define ZXDH_PMD_DEFAULT_HOST_FEATURES   \
+	(1ULL << ZXDH_NET_F_MRG_RXBUF | \
+	 1ULL << ZXDH_NET_F_STATUS    | \
+	 1ULL << ZXDH_NET_F_MQ        | \
+	 1ULL << ZXDH_F_ANY_LAYOUT    | \
+	 1ULL << ZXDH_F_VERSION_1   | \
+	 1ULL << ZXDH_F_RING_PACKED | \
+	 1ULL << ZXDH_F_IN_ORDER    | \
+	 1ULL << ZXDH_F_NOTIFICATION_DATA |\
+	 1ULL << ZXDH_NET_F_MAC | \
+	 1ULL << ZXDH_NET_F_CSUM |\
+	 1ULL << ZXDH_NET_F_GUEST_CSUM |\
+	 1ULL << ZXDH_NET_F_GUEST_TSO4 |\
+	 1ULL << ZXDH_NET_F_GUEST_TSO6 |\
+	 1ULL << ZXDH_NET_F_HOST_TSO4 |\
+	 1ULL << ZXDH_NET_F_HOST_TSO6)
+
 static void
 zxdh_read_dev_config(struct zxdh_hw *hw, size_t offset,
 		void *dst, int32_t length)
@@ -391,13 +408,18 @@ zxdh_pci_read_dev_config(struct zxdh_hw *hw, size_t offset, void *dst, int32_t l
 	ZXDH_VTPCI_OPS(hw)->read_dev_cfg(hw, offset, dst, length);
 }
 
+void zxdh_pci_write_dev_config(struct zxdh_hw *hw, size_t offset, const void *src, int32_t length)
+{
+	ZXDH_VTPCI_OPS(hw)->write_dev_cfg(hw, offset, src, length);
+}
+
 void
 zxdh_get_pci_dev_config(struct zxdh_hw *hw)
 {
 	uint64_t guest_features = 0;
 	uint64_t nego_features = 0;
 
-	hw->host_features = zxdh_pci_get_features(hw);
+	hw->host_features = ZXDH_PMD_DEFAULT_HOST_FEATURES;
 
 	guest_features = (uint64_t)ZXDH_PMD_DEFAULT_GUEST_FEATURES;
 	nego_features = guest_features & hw->host_features;
diff --git a/drivers/net/zxdh/zxdh_pci.h b/drivers/net/zxdh/zxdh_pci.h
index 9b8bef6c09..a1834f6615 100644
--- a/drivers/net/zxdh/zxdh_pci.h
+++ b/drivers/net/zxdh/zxdh_pci.h
@@ -162,7 +162,7 @@ void zxdh_pci_read_dev_config(struct zxdh_hw *hw, size_t offset,
 
 int32_t zxdh_read_pci_caps(struct rte_pci_device *dev, struct zxdh_hw *hw);
 void zxdh_get_pci_dev_config(struct zxdh_hw *hw);
-
+void zxdh_pci_write_dev_config(struct zxdh_hw *hw, size_t offset, const void *src, int32_t length);
 uint16_t zxdh_pci_get_features(struct zxdh_hw *hw);
 enum zxdh_msix_status zxdh_pci_msix_detect(struct rte_pci_device *dev);
 uint8_t zxdh_pci_isr(struct zxdh_hw *hw);
diff --git a/drivers/net/zxdh/zxdh_rxtx.c b/drivers/net/zxdh/zxdh_rxtx.c
index 1921a23f25..9bb57d4f84 100644
--- a/drivers/net/zxdh/zxdh_rxtx.c
+++ b/drivers/net/zxdh/zxdh_rxtx.c
@@ -760,7 +760,6 @@ zxdh_rx_update_mbuf(struct rte_mbuf *m, struct zxdh_net_hdr_ul *hdr)
 		idx = (pkt_type_inner >> 4)  & 0xF;
 		m->packet_type |= zxdh_inner_l4_type[idx];
 	}
-
 }
 
 static void zxdh_discard_rxbuf(struct zxdh_virtqueue *vq, struct rte_mbuf *m)
@@ -818,9 +817,14 @@ zxdh_recv_pkts_packed(void *rx_queue, struct rte_mbuf **rx_pkts,
 		seg_num  = header->type_hdr.num_buffers;
 
 		/* Private queue only handle type hdr */
-		hdr_size = ZXDH_TYPE_HDR_SIZE;
-		rxm->pkt_len = ((header->type_hdr.port & 0x7f) << 8) +
-							header->type_hdr.pd_len;
+		hdr_size = header->type_hdr.pd_len << 1;
+		if (unlikely(hdr_size > lens[i] || hdr_size < ZXDH_TYPE_HDR_SIZE)) {
+			PMD_RX_LOG(ERR, "hdr_size:%u is invalid", hdr_size);
+			rte_pktmbuf_free(rxm);
+			rxvq->stats.errors++;
+			rxvq->stats.invalid_hdr_len_err++;
+			continue;
+		}
 		rxm->data_off += hdr_size;
 		rxm->nb_segs = seg_num;
 		rxm->ol_flags = 0;
diff --git a/drivers/net/zxdh/zxdh_tables.c b/drivers/net/zxdh/zxdh_tables.c
index 253d9ce438..ab91d51948 100644
--- a/drivers/net/zxdh/zxdh_tables.c
+++ b/drivers/net/zxdh/zxdh_tables.c
@@ -8,14 +8,7 @@
 #include "zxdh_tables.h"
 #include "zxdh_logs.h"
 
-#define ZXDH_SDT_VPORT_ATT_TABLE          1
-#define ZXDH_SDT_PANEL_ATT_TABLE          2
-#define ZXDH_SDT_RSS_ATT_TABLE            3
-#define ZXDH_SDT_VLAN_ATT_TABLE           4
-#define ZXDH_SDT_BROCAST_ATT_TABLE        6
-#define ZXDH_SDT_UNICAST_ATT_TABLE        10
-#define ZXDH_SDT_MULTICAST_ATT_TABLE      11
-#define ZXDH_SDT_PORT_VLAN_ATT_TABLE      16
+
 
 #define ZXDH_MAC_HASH_INDEX_BASE          64
 #define ZXDH_MAC_HASH_INDEX(index)        (ZXDH_MAC_HASH_INDEX_BASE + (index))
@@ -40,15 +33,16 @@ zxdh_set_port_attr(struct zxdh_hw *hw, uint16_t vport, struct zxdh_port_attr_tab
 {
 	struct zxdh_dtb_shared_data *dtb_data = &hw->dev_sd->dtb_sd;
 	union zxdh_virport_num vport_num = (union zxdh_virport_num)vport;
+	uint16_t vfid = zxdh_vport_to_vfid(vport_num);
 	int ret = 0;
 
-	ZXDH_DTB_ERAM_ENTRY_INFO_T entry = {vport_num.vfid, (uint32_t *)port_attr};
+	ZXDH_DTB_ERAM_ENTRY_INFO_T entry = {vfid, (uint32_t *)port_attr};
 	ZXDH_DTB_USER_ENTRY_T user_entry_write = {ZXDH_SDT_VPORT_ATT_TABLE, (void *)&entry};
 
 	ret = zxdh_np_dtb_table_entry_write(hw->slot_id,
 				dtb_data->queueid, 1, &user_entry_write);
 	if (ret != 0)
-		PMD_DRV_LOG(ERR, "write vport_att failed vfid:%d failed", vport_num.vfid);
+		PMD_DRV_LOG(ERR, "write vport_att failed vfid:%d failed", vfid);
 
 	return ret;
 }
@@ -72,6 +66,7 @@ zxdh_port_attr_init(struct rte_eth_dev *dev)
 		port_attr.mtu = dev->data->mtu;
 		port_attr.mtu_enable = 1;
 		port_attr.is_up = 0;
+		port_attr.hash_search_index = hw->hash_search_index;
 		if (!port_attr.rss_enable)
 			port_attr.port_base_qid = 0;
 
@@ -144,6 +139,7 @@ int zxdh_panel_table_init(struct rte_eth_dev *dev)
 	panel.pf_vfid = zxdh_vport_to_vfid(hw->vport);
 	panel.mtu_enable = 1;
 	panel.mtu = dev->data->mtu;
+	panel.port_vfid_1588 = panel.pf_vfid;
 
 	ZXDH_DTB_ERAM_ENTRY_INFO_T panel_entry = {
 		.index = hw->phyport,
@@ -212,13 +208,14 @@ zxdh_get_port_attr(struct zxdh_hw *hw, uint16_t vport, struct zxdh_port_attr_tab
 {
 	struct zxdh_dtb_shared_data *dtb_data = &hw->dev_sd->dtb_sd;
 	union zxdh_virport_num vport_num = (union zxdh_virport_num)vport;
-	ZXDH_DTB_ERAM_ENTRY_INFO_T entry = {vport_num.vfid, (uint32_t *)port_attr};
+	uint16_t vfid = zxdh_vport_to_vfid(vport_num);
+	ZXDH_DTB_ERAM_ENTRY_INFO_T entry = {vfid, (uint32_t *)port_attr};
 	ZXDH_DTB_USER_ENTRY_T user_entry_get = {ZXDH_SDT_VPORT_ATT_TABLE, &entry};
 	int ret;
 
 	ret = zxdh_np_dtb_table_entry_get(hw->slot_id, dtb_data->queueid, &user_entry_get, 1);
 	if (ret != 0)
-		PMD_DRV_LOG(ERR, "get port_attr vfid:%d failed, ret:%d", vport_num.vfid, ret);
+		PMD_DRV_LOG(ERR, "get port_attr vfid:%d failed, ret:%d", vfid, ret);
 
 	return ret;
 }
@@ -229,7 +226,8 @@ zxdh_delete_port_attr(struct zxdh_hw *hw, uint16_t vport,
 {
 	struct zxdh_dtb_shared_data *dtb_data = &hw->dev_sd->dtb_sd;
 	union zxdh_virport_num vport_num = (union zxdh_virport_num)vport;
-	ZXDH_DTB_ERAM_ENTRY_INFO_T entry = {vport_num.vfid, (uint32_t *)port_attr};
+	uint16_t vfid = zxdh_vport_to_vfid(vport_num);
+	ZXDH_DTB_ERAM_ENTRY_INFO_T entry = {vfid, (uint32_t *)port_attr};
 	ZXDH_DTB_USER_ENTRY_T user_entry = {
 		.sdt_no = ZXDH_SDT_VPORT_ATT_TABLE,
 		.p_entry_data = (void *)&entry
@@ -247,9 +245,9 @@ zxdh_add_mac_table(struct zxdh_hw *hw, uint16_t vport, struct rte_ether_addr *ad
 	struct zxdh_mac_unicast_table unicast_table = {0};
 	struct zxdh_mac_multicast_table multicast_table = {0};
 	union zxdh_virport_num vport_num = (union zxdh_virport_num)vport;
+	uint16_t vfid = zxdh_vport_to_vfid(vport_num);
 	uint32_t ret;
 	uint16_t group_id = 0;
-	uint16_t vfid = vport_num.vfid;
 
 	if (rte_is_unicast_ether_addr(addr)) {
 		rte_memcpy(unicast_table.key.dmac_addr, addr, sizeof(struct rte_ether_addr));
@@ -351,15 +349,17 @@ zxdh_del_mac_table(struct zxdh_hw *hw, uint16_t vport, struct rte_ether_addr *ad
 	struct zxdh_mac_unicast_table unicast_table = {0};
 	struct zxdh_mac_multicast_table multicast_table = {0};
 	union zxdh_virport_num vport_num = (union zxdh_virport_num)vport;
-	uint32_t ret, del_flag = 0;
-	uint16_t group_id = 0;
 	union zxdh_virport_num port = (union zxdh_virport_num)vport;
 	uint16_t vfid = zxdh_vport_to_vfid(port);
+	uint32_t ret, del_flag = 0;
+	uint16_t group_id = 0;
 
 	if (rte_is_unicast_ether_addr(addr)) {
 		rte_memcpy(unicast_table.key.dmac_addr, addr, sizeof(struct rte_ether_addr));
 		unicast_table.key.sriov_vlan_id = srv_vlanid;
 		unicast_table.key.sriov_vlan_tpid = srv_tpid;
+		unicast_table.entry.hit_flag = 0;
+		unicast_table.entry.vfid = rte_cpu_to_be_16(vfid & 0x7ff);
 
 		ZXDH_DTB_HASH_ENTRY_INFO_T dtb_hash_entry = {
 			.p_actu_key = (uint8_t *)&unicast_table.key,
@@ -800,6 +800,7 @@ zxdh_rss_table_set(struct zxdh_hw *hw, uint16_t vport, struct zxdh_rss_reta *rss
 	struct zxdh_dtb_shared_data *dtb_data = &hw->dev_sd->dtb_sd;
 	struct zxdh_rss_to_vqid_table rss_vqid = {0};
 	union zxdh_virport_num vport_num = (union zxdh_virport_num)vport;
+	uint16_t vfid = zxdh_vport_to_vfid(vport_num);
 	int ret = 0;
 
 	for (uint16_t i = 0; i < RTE_ETH_RSS_RETA_SIZE_256 / 8; i++) {
@@ -820,7 +821,7 @@ zxdh_rss_table_set(struct zxdh_hw *hw, uint16_t vport, struct zxdh_rss_reta *rss
 			rss_vqid.vqm_qid[0] |= 0x8000;
 #endif
 		ZXDH_DTB_ERAM_ENTRY_INFO_T entry = {
-			.index = vport_num.vfid * 32 + i,
+			.index = vfid * 32 + i,
 			.p_data = (uint32_t *)&rss_vqid
 		};
 		ZXDH_DTB_USER_ENTRY_T user_entry_write = {
@@ -830,7 +831,7 @@ zxdh_rss_table_set(struct zxdh_hw *hw, uint16_t vport, struct zxdh_rss_reta *rss
 		ret = zxdh_np_dtb_table_entry_write(hw->slot_id,
 					dtb_data->queueid, 1, &user_entry_write);
 		if (ret != 0) {
-			PMD_DRV_LOG(ERR, "write rss base qid failed vfid:%d", vport_num.vfid);
+			PMD_DRV_LOG(ERR, "write rss base qid failed vfid:%d", vfid);
 			return ret;
 		}
 	}
@@ -843,16 +844,17 @@ zxdh_rss_table_get(struct zxdh_hw *hw, uint16_t vport, struct zxdh_rss_reta *rss
 	struct zxdh_dtb_shared_data *dtb_data = &hw->dev_sd->dtb_sd;
 	struct zxdh_rss_to_vqid_table rss_vqid = {0};
 	union zxdh_virport_num vport_num = (union zxdh_virport_num)vport;
+	uint16_t vfid = zxdh_vport_to_vfid(vport_num);
 	int ret = 0;
 
 	for (uint16_t i = 0; i < RTE_ETH_RSS_RETA_SIZE_256 / 8; i++) {
-		ZXDH_DTB_ERAM_ENTRY_INFO_T entry = {vport_num.vfid * 32 + i, (uint32_t *)&rss_vqid};
+		ZXDH_DTB_ERAM_ENTRY_INFO_T entry = {vfid * 32 + i, (uint32_t *)&rss_vqid};
 		ZXDH_DTB_USER_ENTRY_T user_entry = {ZXDH_SDT_RSS_ATT_TABLE, &entry};
 
 		ret = zxdh_np_dtb_table_entry_get(hw->slot_id,
 					dtb_data->queueid, &user_entry, 1);
 		if (ret != 0) {
-			PMD_DRV_LOG(ERR, "get rss tbl failed, vfid:%d", vport_num.vfid);
+			PMD_DRV_LOG(ERR, "get rss tbl failed, vfid:%d", vfid);
 			return -1;
 		}
 
diff --git a/drivers/net/zxdh/zxdh_tables.h b/drivers/net/zxdh/zxdh_tables.h
index 2f2ada3a9f..cb34e38be8 100644
--- a/drivers/net/zxdh/zxdh_tables.h
+++ b/drivers/net/zxdh/zxdh_tables.h
@@ -9,7 +9,13 @@
 
 /* eram */
 #define ZXDH_SDT_VPORT_ATT_TABLE          1
-
+#define ZXDH_SDT_PANEL_ATT_TABLE          2
+#define ZXDH_SDT_RSS_ATT_TABLE            3
+#define ZXDH_SDT_VLAN_ATT_TABLE           4
+#define ZXDH_SDT_BROCAST_ATT_TABLE        6
+#define ZXDH_SDT_UNICAST_ATT_TABLE        10
+#define ZXDH_SDT_MULTICAST_ATT_TABLE      11
+#define ZXDH_SDT_PORT_VLAN_ATT_TABLE      16
 /* hash */
 #define ZXDH_SDT_L2_ENTRY_TABLE0          64
 #define ZXDH_SDT_L2_ENTRY_TABLE1          65
@@ -80,8 +86,6 @@
 #define ZXDH_MTR_STATS_EGRESS_BASE           0x7481
 #define ZXDH_MTR_STATS_INGRESS_BASE          0x7C81
 
-extern struct zxdh_dtb_shared_data g_dtb_data;
-
 struct zxdh_port_vlan_table {
 #if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN
 	uint16_t business_vlan_tpid:16;
@@ -233,19 +237,51 @@ struct zxdh_port_attr_table {
 };
 
 struct zxdh_panel_table {
-	uint16_t port_vfid_1588 : 11,
-			 rsv2           : 5;
-	uint16_t pf_vfid        : 11,
-			 rsv1           : 1,
-			 enable_1588_tc : 2,
-			 trust_mode     : 1,
-			 hit_flag       : 1;
-	uint32_t mtu            : 16,
-			 mtu_enable     : 1,
-			 rsv            : 3,
-			 tm_base_queue  : 12;
-	uint32_t rsv_1;
-	uint32_t rsv_2;
+#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN
+	uint16_t port_vfid_1588   : 11,
+				rsv2             : 5;
+	uint16_t rsv1             : 11,
+			 tm_shape_enable  : 1,
+			 enable_1588_tc   : 2,
+			 trust_mode       : 1,
+			 hit_flag         : 1;
+	uint16_t mtu              : 16;
+	uint16_t mtu_enable       : 1,
+			 rsv              : 3,
+			 tm_base_queue    : 12;
+	uint16_t lacp_pf_qid      : 12,
+				rsv5             : 4;
+	uint16_t lacp_pf_vfid     : 11,
+				rsv6             : 2,
+				member_port_up   : 1,
+				bond_link_up     : 1,
+				hw_bond_enable   : 1;
+	uint16_t rsv3             : 16;
+	uint16_t pf_vfid          : 11,
+				rsv4             : 5;
+#else
+	uint16_t rsv1             : 11,
+				tm_shape_enable  : 1,
+				enable_1588_tc   : 2,
+				trust_mode       : 1,
+				hit_flag         : 1;
+	uint16_t port_vfid_1588   : 11,
+				rsv2             : 5;
+	uint16_t mtu_enable       : 1,
+				rsv              : 3,
+				tm_base_queue    : 12;
+	uint16_t mtu              : 16;
+	uint16_t lacp_pf_vfid     : 11,
+				rsv6             : 2,
+				member_port_up   : 1,
+				bond_link_up     : 1,
+				hw_bond_enable   : 1;
+	uint16_t lacp_pf_qid      : 12,
+				rsv5             : 4;
+	uint16_t pf_vfid          : 11,
+				rsv4             : 5;
+	uint16_t rsv3             : 16;
+#endif
 }; /* 16B */
 
 struct zxdh_mac_unicast_key {
-- 
2.27.0

[-- Attachment #1.1.2: Type: text/html , Size: 217264 bytes --]

  parent reply	other threads:[~2025-03-19  9:16 UTC|newest]

Thread overview: 95+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2025-02-10  1:44 [PATCH v1 01/14] net/zxdh: add network processor registers ops Bingbin Chen
2025-02-10  1:46 ` [PATCH v1 02/14] net/zxdh: support compatibility check Bingbin Chen
2025-02-10 17:25   ` Stephen Hemminger
2025-02-10  1:47 ` [PATCH v1 03/14] net/zxdh: add agent channel Bingbin Chen
2025-02-10 17:28   ` Stephen Hemminger
2025-02-10 17:30   ` Stephen Hemminger
2025-02-10 17:31   ` Stephen Hemminger
2025-02-10 18:23   ` Stephen Hemminger
2025-02-10  1:47 ` [PATCH v1 04/14] net/zxdh: modify dtb queue ops Bingbin Chen
2025-02-10 17:31   ` Stephen Hemminger
2025-02-10  1:48 ` [PATCH v1 05/14] net/zxdh: add tables dump address ops Bingbin Chen
2025-02-10 17:33   ` Stephen Hemminger
2025-02-10  1:50 ` [PATCH v1 06/14] net/zxdh: add eram tables ops Bingbin Chen
2025-02-10  1:50   ` [PATCH v1 07/14] net/zxdh: get flow tables resources Bingbin Chen
2025-02-10 17:35     ` Stephen Hemminger
2025-02-10 17:35     ` Stephen Hemminger
2025-02-10  1:50   ` [PATCH v1 08/14] net/zxdh: support hash resources configuration Bingbin Chen
2025-02-10 17:36     ` Stephen Hemminger
2025-02-10  1:50   ` [PATCH v1 09/14] net/zxdh: implement tables initialization Bingbin Chen
2025-02-10 17:40     ` Stephen Hemminger
2025-02-10 17:43     ` Stephen Hemminger
2025-02-10  1:50   ` [PATCH v1 10/14] net/zxdh: support hash tables write and delete ops Bingbin Chen
2025-02-10 17:45     ` Stephen Hemminger
2025-02-10  1:50   ` [PATCH v1 11/14] net/zxdh: get hash table entry result Bingbin Chen
2025-02-10 17:46     ` Stephen Hemminger
2025-02-10  1:50   ` [PATCH v1 12/14] net/zxdh: delete all hash entries Bingbin Chen
2025-02-10 17:47     ` Stephen Hemminger
2025-02-10  1:50   ` [PATCH v1 13/14] net/zxdh: add acl tables ops Bingbin Chen
2025-02-10  1:50   ` [PATCH v1 14/14] net/zxdh: clean stat values Bingbin Chen
2025-02-10 17:50     ` Stephen Hemminger
2025-02-10 17:50     ` Stephen Hemminger
2025-02-10 18:19     ` Stephen Hemminger
2025-02-22  7:22 ` [PATCH v2 00/14] add network processor ops Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 01/14] net/zxdh: add network processor registers ops Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 02/14] net/zxdh: support compatibility check Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 03/14] net/zxdh: add agent channel Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 04/14] net/zxdh: modify dtb queue ops Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 05/14] net/zxdh: add tables dump address ops Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 06/14] net/zxdh: add eram tables ops Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 07/14] net/zxdh: get flow tables resources Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 08/14] net/zxdh: support hash resources configuration Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 09/14] net/zxdh: implement tables initialization Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 10/14] net/zxdh: support hash tables write and delete ops Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 11/14] net/zxdh: get hash table entry result Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 12/14] net/zxdh: delete all hash entries Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 13/14] net/zxdh: add acl tables ops Bingbin Chen
2025-02-22  7:22   ` [PATCH v2 14/14] net/zxdh: clean stat values Bingbin Chen
2025-02-22 17:34     ` Stephen Hemminger
2025-03-05  8:13 ` [PATCH v3 00/14] net/zxdh: add network processor ops Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 01/14] net/zxdh: add network processor registers ops Bingbin Chen
2025-03-17 14:57     ` [PATCH v4 00/14] net/zxdh: add network processor ops Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 01/14] net/zxdh: add network processor registers ops Bingbin Chen
2025-03-19  8:57         ` [PATCH v5 00/14] net/zxdh: add network processor ops Bingbin Chen
2025-03-19  8:57           ` [PATCH v5 01/14] net/zxdh: add network processor registers ops Bingbin Chen
2025-03-19  8:57           ` [PATCH v5 02/14] net/zxdh: support compatibility check Bingbin Chen
2025-03-19  8:57           ` [PATCH v5 03/14] net/zxdh: add agent channel Bingbin Chen
2025-03-19  8:57           ` [PATCH v5 04/14] net/zxdh: modify dtb queue ops Bingbin Chen
2025-03-19  8:57           ` [PATCH v5 05/14] net/zxdh: add tables dump address ops Bingbin Chen
2025-03-19  8:58           ` [PATCH v5 06/14] net/zxdh: add eram tables ops Bingbin Chen
2025-03-19  8:58           ` [PATCH v5 07/14] net/zxdh: get flow tables resources Bingbin Chen
2025-03-19  8:58           ` [PATCH v5 08/14] net/zxdh: support hash resources configuration Bingbin Chen
2025-03-19  8:58           ` [PATCH v5 09/14] net/zxdh: implement tables initialization Bingbin Chen
2025-03-19  8:58           ` [PATCH v5 10/14] net/zxdh: support hash tables write and delete ops Bingbin Chen
2025-03-19  8:58           ` [PATCH v5 11/14] net/zxdh: get hash table entry result Bingbin Chen
2025-03-19  8:58           ` [PATCH v5 12/14] net/zxdh: delete all hash entries Bingbin Chen
2025-03-19  8:58           ` [PATCH v5 13/14] net/zxdh: add acl tables ops Bingbin Chen
2025-03-19  8:58           ` Bingbin Chen [this message]
2025-03-19  9:31           ` [v5,00/14] net/zxdh: add network processor ops Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 02/14] net/zxdh: support compatibility check Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 03/14] net/zxdh: add agent channel Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 04/14] net/zxdh: modify dtb queue ops Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 05/14] net/zxdh: add tables dump address ops Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 06/14] net/zxdh: add eram tables ops Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 07/14] net/zxdh: get flow tables resources Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 08/14] net/zxdh: support hash resources configuration Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 09/14] net/zxdh: implement tables initialization Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 10/14] net/zxdh: support hash tables write and delete ops Bingbin Chen
2025-03-17 14:57       ` [PATCH v4 11/14] net/zxdh: get hash table entry result Bingbin Chen
2025-03-17 14:58       ` [PATCH v4 12/14] net/zxdh: delete all hash entries Bingbin Chen
2025-03-17 14:58       ` [PATCH v4 13/14] net/zxdh: add acl tables ops Bingbin Chen
2025-03-17 14:58       ` [PATCH v4 14/14] net/zxdh: fix debugging errors Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 02/14] net/zxdh: support compatibility check Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 03/14] net/zxdh: add agent channel Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 04/14] net/zxdh: modify dtb queue ops Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 05/14] net/zxdh: add tables dump address ops Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 06/14] net/zxdh: add eram tables ops Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 07/14] net/zxdh: get flow tables resources Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 08/14] net/zxdh: support hash resources configuration Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 09/14] net/zxdh: implement tables initialization Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 10/14] net/zxdh: support hash tables write and delete ops Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 11/14] net/zxdh: get hash table entry result Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 12/14] net/zxdh: delete all hash entries Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 13/14] net/zxdh: add acl tables ops Bingbin Chen
2025-03-05  8:13   ` [PATCH v3 14/14] net/zxdh: modify parameters of the plcr function Bingbin Chen
2025-03-10 23:19 ` [PATCH v1 01/14] net/zxdh: add network processor registers ops Stephen Hemminger

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20250319085808.1912523-15-chen.bingbin@zte.com.cn \
    --to=chen.bingbin@zte.com.cn \
    --cc=dev@dpdk.org \
    --cc=stephen@networkplumber.org \
    --cc=wang.junlong1@zte.com.cn \
    --cc=yang.yonggang@zte.com.cn \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).