DPDK patches and discussions
 help / color / mirror / Atom feed
* [PATCH v2 0/2] vhost: add VDUSE reconnection support
@ 2024-09-20 15:09 Maxime Coquelin
  2024-09-20 15:09 ` [PATCH v2 1/2] vhost: add logging mechanism for reconnection Maxime Coquelin
  2024-09-20 15:09 ` [PATCH v2 2/2] vhost: add reconnection support to VDUSE Maxime Coquelin
  0 siblings, 2 replies; 8+ messages in thread
From: Maxime Coquelin @ 2024-09-20 15:09 UTC (permalink / raw)
  To: dev, david.marchand, chenbox; +Cc: Maxime Coquelin

This series adds support for VDUSE reconnection.

First patch introduces the reconnection file layout and
track the virtqueues available index updates in the
datapath and control queue.

Second patch adds VDUSE reconnect intialization and some
sanity checks to prevent incompatible reconnections.

Changes in v2:
==============
- Added more sanity checks at reconnection
- Improve versionning
- Fix error loggin (Chenbo)
- Clarify why offloading device start is required (Chenbo)
- Change runtime path to <RUNPATH>/vduse instead of <RUNPATH>/dpdk/vduse

Maxime Coquelin (2):
  vhost: add logging mechanism for reconnection
  vhost: add reconnection support to VDUSE

 lib/vhost/vduse.c           | 308 +++++++++++++++++++++++++++++++-----
 lib/vhost/vhost.h           |  41 ++++-
 lib/vhost/virtio_net.c      |   8 +
 lib/vhost/virtio_net_ctrl.c |   2 +
 4 files changed, 316 insertions(+), 43 deletions(-)

-- 
2.46.0


^ permalink raw reply	[flat|nested] 8+ messages in thread

* [PATCH v2 1/2] vhost: add logging mechanism for reconnection
  2024-09-20 15:09 [PATCH v2 0/2] vhost: add VDUSE reconnection support Maxime Coquelin
@ 2024-09-20 15:09 ` Maxime Coquelin
  2024-09-23  7:12   ` Chenbo Xia
  2024-09-23 15:42   ` David Marchand
  2024-09-20 15:09 ` [PATCH v2 2/2] vhost: add reconnection support to VDUSE Maxime Coquelin
  1 sibling, 2 replies; 8+ messages in thread
From: Maxime Coquelin @ 2024-09-20 15:09 UTC (permalink / raw)
  To: dev, david.marchand, chenbox; +Cc: Maxime Coquelin

This patch introduces a way for backend to keep track
of the needed information to be able to reconnect without
frontend cooperation.

It will be used for VDUSE, which does not provide interface
for the backend to save and later recover local virtqueues
metadata needed to reconnect.

Vhost-user support could also be added for improved packed
ring reconnection support.

Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
---
 lib/vhost/vhost.h           | 41 ++++++++++++++++++++++++++++++++++---
 lib/vhost/virtio_net.c      |  8 ++++++++
 lib/vhost/virtio_net_ctrl.c |  2 ++
 3 files changed, 48 insertions(+), 3 deletions(-)

diff --git a/lib/vhost/vhost.h b/lib/vhost/vhost.h
index cd3fa55f1b..1f4192f5d1 100644
--- a/lib/vhost/vhost.h
+++ b/lib/vhost/vhost.h
@@ -269,6 +269,24 @@ struct vhost_async {
 	};
 };
 
+#define VHOST_RECONNECT_VERSION		0x0
+#define VHOST_MAX_VRING			0x100
+#define VHOST_MAX_QUEUE_PAIRS		0x80
+
+struct __rte_cache_aligned vhost_reconnect_vring {
+	uint16_t last_avail_idx;
+	bool avail_wrap_counter;
+};
+
+struct vhost_reconnect_data {
+	uint32_t version;
+	uint64_t features;
+	uint8_t status;
+	struct virtio_net_config config;
+	uint32_t nr_vrings;
+	struct vhost_reconnect_vring vring[VHOST_MAX_VRING];
+};
+
 /**
  * Structure contains variables relevant to RX/TX virtqueues.
  */
@@ -351,6 +369,7 @@ struct __rte_cache_aligned vhost_virtqueue {
 	struct virtqueue_stats	stats;
 
 	RTE_ATOMIC(bool) irq_pending;
+	struct vhost_reconnect_vring *reconnect_log;
 };
 
 /* Virtio device status as per Virtio specification */
@@ -362,9 +381,6 @@ struct __rte_cache_aligned vhost_virtqueue {
 #define VIRTIO_DEVICE_STATUS_DEV_NEED_RESET	0x40
 #define VIRTIO_DEVICE_STATUS_FAILED		0x80
 
-#define VHOST_MAX_VRING			0x100
-#define VHOST_MAX_QUEUE_PAIRS		0x80
-
 /* Declare IOMMU related bits for older kernels */
 #ifndef VIRTIO_F_IOMMU_PLATFORM
 
@@ -538,8 +554,26 @@ struct __rte_cache_aligned virtio_net {
 	struct rte_vhost_user_extern_ops extern_ops;
 
 	struct vhost_backend_ops *backend_ops;
+
+	struct vhost_reconnect_data *reconnect_log;
 };
 
+static __rte_always_inline void
+vhost_virtqueue_reconnect_log_split(struct vhost_virtqueue *vq)
+{
+	if (vq->reconnect_log != NULL)
+		vq->reconnect_log->last_avail_idx = vq->last_avail_idx;
+}
+
+static __rte_always_inline void
+vhost_virtqueue_reconnect_log_packed(struct vhost_virtqueue *vq)
+{
+	if (vq->reconnect_log != NULL) {
+		vq->reconnect_log->last_avail_idx = vq->last_avail_idx;
+		vq->reconnect_log->avail_wrap_counter = vq->avail_wrap_counter;
+	}
+}
+
 static inline void
 vq_assert_lock__(struct virtio_net *dev, struct vhost_virtqueue *vq, const char *func)
 	__rte_assert_exclusive_lock(&vq->access_lock)
@@ -584,6 +618,7 @@ vq_inc_last_avail_packed(struct vhost_virtqueue *vq, uint16_t num)
 		vq->avail_wrap_counter ^= 1;
 		vq->last_avail_idx -= vq->size;
 	}
+	vhost_virtqueue_reconnect_log_packed(vq);
 }
 
 void __vhost_log_cache_write(struct virtio_net *dev,
diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index 370402d849..f66a0c82f8 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -1445,6 +1445,7 @@ virtio_dev_rx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
 		}
 
 		vq->last_avail_idx += num_buffers;
+		vhost_virtqueue_reconnect_log_split(vq);
 	}
 
 	do_data_copy_enqueue(dev, vq);
@@ -1857,6 +1858,7 @@ virtio_dev_rx_async_submit_split(struct virtio_net *dev, struct vhost_virtqueue
 		pkts_info[slot_idx].mbuf = pkts[pkt_idx];
 
 		vq->last_avail_idx += num_buffers;
+		vhost_virtqueue_reconnect_log_split(vq);
 	}
 
 	if (unlikely(pkt_idx == 0))
@@ -1885,6 +1887,7 @@ virtio_dev_rx_async_submit_split(struct virtio_net *dev, struct vhost_virtqueue
 		/* recover shadow used ring and available ring */
 		vq->shadow_used_idx -= num_descs;
 		vq->last_avail_idx -= num_descs;
+		vhost_virtqueue_reconnect_log_split(vq);
 	}
 
 	/* keep used descriptors */
@@ -2100,6 +2103,7 @@ dma_error_handler_packed(struct vhost_virtqueue *vq, uint16_t slot_idx,
 		vq->last_avail_idx = vq->last_avail_idx + vq->size - descs_err;
 		vq->avail_wrap_counter ^= 1;
 	}
+	vhost_virtqueue_reconnect_log_packed(vq);
 
 	if (async->buffer_idx_packed >= buffers_err)
 		async->buffer_idx_packed -= buffers_err;
@@ -3182,6 +3186,7 @@ virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
 
 	if (likely(vq->shadow_used_idx)) {
 		vq->last_avail_idx += vq->shadow_used_idx;
+		vhost_virtqueue_reconnect_log_split(vq);
 		do_data_copy_dequeue(vq);
 		flush_shadow_used_ring_split(dev, vq);
 		vhost_vring_call_split(dev, vq);
@@ -3854,6 +3859,7 @@ virtio_dev_tx_async_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
 		async->desc_idx_split++;
 
 		vq->last_avail_idx++;
+		vhost_virtqueue_reconnect_log_split(vq);
 	}
 
 	if (unlikely(dropped))
@@ -3872,6 +3878,7 @@ virtio_dev_tx_async_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
 		pkt_idx = n_xfer;
 		/* recover available ring */
 		vq->last_avail_idx -= pkt_err;
+		vhost_virtqueue_reconnect_log_split(vq);
 
 		/**
 		 * recover async channel copy related structures and free pktmbufs
@@ -4153,6 +4160,7 @@ virtio_dev_tx_async_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,
 			vq->last_avail_idx += vq->size - descs_err;
 			vq->avail_wrap_counter ^= 1;
 		}
+		vhost_virtqueue_reconnect_log_packed(vq);
 	}
 
 	async->pkts_idx += pkt_idx;
diff --git a/lib/vhost/virtio_net_ctrl.c b/lib/vhost/virtio_net_ctrl.c
index 8f78122361..b8ee94018e 100644
--- a/lib/vhost/virtio_net_ctrl.c
+++ b/lib/vhost/virtio_net_ctrl.c
@@ -169,6 +169,7 @@ virtio_net_ctrl_pop(struct virtio_net *dev, struct vhost_virtqueue *cvq,
 	cvq->last_avail_idx++;
 	if (cvq->last_avail_idx >= cvq->size)
 		cvq->last_avail_idx -= cvq->size;
+	vhost_virtqueue_reconnect_log_split(cvq);
 
 	if (dev->features & (1ULL << VIRTIO_RING_F_EVENT_IDX))
 		vhost_avail_event(cvq) = cvq->last_avail_idx;
@@ -181,6 +182,7 @@ virtio_net_ctrl_pop(struct virtio_net *dev, struct vhost_virtqueue *cvq,
 	cvq->last_avail_idx++;
 	if (cvq->last_avail_idx >= cvq->size)
 		cvq->last_avail_idx -= cvq->size;
+	vhost_virtqueue_reconnect_log_split(cvq);
 
 	if (dev->features & (1ULL << VIRTIO_RING_F_EVENT_IDX))
 		vhost_avail_event(cvq) = cvq->last_avail_idx;
-- 
2.46.0


^ permalink raw reply	[flat|nested] 8+ messages in thread

* [PATCH v2 2/2] vhost: add reconnection support to VDUSE
  2024-09-20 15:09 [PATCH v2 0/2] vhost: add VDUSE reconnection support Maxime Coquelin
  2024-09-20 15:09 ` [PATCH v2 1/2] vhost: add logging mechanism for reconnection Maxime Coquelin
@ 2024-09-20 15:09 ` Maxime Coquelin
  2024-09-23  7:13   ` Chenbo Xia
  2024-09-23 16:43   ` David Marchand
  1 sibling, 2 replies; 8+ messages in thread
From: Maxime Coquelin @ 2024-09-20 15:09 UTC (permalink / raw)
  To: dev, david.marchand, chenbox; +Cc: Maxime Coquelin

This patch enables VDUSE reconnection support making use of
the newly introduced reconnection mechanism in Vhost
library.

At DPDK VDUSE device creation time, there are two
possibilities:
 1. The Kernel VDUSE device does not exist:
  a. A reconnection file named after the VUDSE device name
     is created in VDUSE tmpfs.
  b. The file is truncated to 'struct vhost_reconnect_data'
     size, and mmapped.
  c. Negotiated features, Virtio status... are saved for
     sanity checks at reconnect time.
 2. The Kernel VDUSE device already exists:
  a. Exit with failure if no reconnect file exists for
     this device.
  b. Open and mmap the reconnect file.
  c. Perform sanity check to ensure features are compatible.
  d. Restore virtqueues' available indexes at startup time.

Then at runtime, the virtqueues' available index are logged by
the Vhost reconnection mechanism.

At DPDK VDUSE device destruction time, there are two
possibilities:
 1. The Kernel VDUSE device destruction succeed, which
    means it is no more attached to the vDPA bus. The
    reconnection file is unmapped and then removed.
 2. The Kernel VDUSE device destruction failed, meaning it
    is no more attached to the vDPA bus. The reconnection
    file is unmapped but not removed to make possible later
    reconnection.

Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
---
 lib/vhost/vduse.c | 308 ++++++++++++++++++++++++++++++++++++++++------
 1 file changed, 268 insertions(+), 40 deletions(-)

diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
index c66602905c..f9ac317438 100644
--- a/lib/vhost/vduse.c
+++ b/lib/vhost/vduse.c
@@ -136,7 +136,7 @@ vduse_control_queue_event(int fd, void *arg, int *remove __rte_unused)
 }
 
 static void
-vduse_vring_setup(struct virtio_net *dev, unsigned int index)
+vduse_vring_setup(struct virtio_net *dev, unsigned int index, bool reconnect)
 {
 	struct vhost_virtqueue *vq = dev->virtqueue[index];
 	struct vhost_vring_addr *ra = &vq->ring_addrs;
@@ -152,6 +152,19 @@ vduse_vring_setup(struct virtio_net *dev, unsigned int index)
 		return;
 	}
 
+	if (reconnect) {
+		vq->last_avail_idx = vq->reconnect_log->last_avail_idx;
+		vq->last_used_idx = vq->reconnect_log->last_avail_idx;
+	} else {
+		vq->last_avail_idx = vq_info.split.avail_index;
+		vq->last_used_idx = vq_info.split.avail_index;
+	}
+	vq->size = vq_info.num;
+	vq->ready = true;
+	vq->enabled = vq_info.ready;
+	ra->desc_user_addr = vq_info.desc_addr;
+	ra->avail_user_addr = vq_info.driver_addr;
+	ra->used_user_addr = vq_info.device_addr;
 	VHOST_CONFIG_LOG(dev->ifname, INFO, "VQ %u info:", index);
 	VHOST_CONFIG_LOG(dev->ifname, INFO, "\tnum: %u", vq_info.num);
 	VHOST_CONFIG_LOG(dev->ifname, INFO, "\tdesc_addr: %llx",
@@ -160,17 +173,9 @@ vduse_vring_setup(struct virtio_net *dev, unsigned int index)
 			(unsigned long long)vq_info.driver_addr);
 	VHOST_CONFIG_LOG(dev->ifname, INFO, "\tdevice_addr: %llx",
 			(unsigned long long)vq_info.device_addr);
-	VHOST_CONFIG_LOG(dev->ifname, INFO, "\tavail_idx: %u", vq_info.split.avail_index);
+	VHOST_CONFIG_LOG(dev->ifname, INFO, "\tavail_idx: %u", vq->last_avail_idx);
+	VHOST_CONFIG_LOG(dev->ifname, INFO, "\tused_idx: %u", vq->last_used_idx);
 	VHOST_CONFIG_LOG(dev->ifname, INFO, "\tready: %u", vq_info.ready);
-
-	vq->last_avail_idx = vq_info.split.avail_index;
-	vq->size = vq_info.num;
-	vq->ready = true;
-	vq->enabled = vq_info.ready;
-	ra->desc_user_addr = vq_info.desc_addr;
-	ra->avail_user_addr = vq_info.driver_addr;
-	ra->used_user_addr = vq_info.device_addr;
-
 	vq->kickfd = eventfd(0, EFD_NONBLOCK | EFD_CLOEXEC);
 	if (vq->kickfd < 0) {
 		VHOST_CONFIG_LOG(dev->ifname, ERR, "Failed to init kickfd for VQ %u: %s",
@@ -267,7 +272,7 @@ vduse_vring_cleanup(struct virtio_net *dev, unsigned int index)
 }
 
 static void
-vduse_device_start(struct virtio_net *dev)
+vduse_device_start(struct virtio_net *dev, bool reconnect)
 {
 	unsigned int i, ret;
 
@@ -287,6 +292,15 @@ vduse_device_start(struct virtio_net *dev)
 		return;
 	}
 
+	if (reconnect && dev->features != dev->reconnect_log->features) {
+		VHOST_CONFIG_LOG(dev->ifname, ERR,
+				"Mismatch between reconnect file features 0x%" PRIx64 " & device features 0x%" PRIx64,
+				dev->reconnect_log->features, dev->features);
+		return;
+	}
+
+	dev->reconnect_log->features = dev->features;
+
 	VHOST_CONFIG_LOG(dev->ifname, INFO, "Negotiated Virtio features: 0x%" PRIx64,
 		dev->features);
 
@@ -300,7 +314,7 @@ vduse_device_start(struct virtio_net *dev)
 	}
 
 	for (i = 0; i < dev->nr_vring; i++)
-		vduse_vring_setup(dev, i);
+		vduse_vring_setup(dev, i, reconnect);
 
 	dev->flags |= VIRTIO_DEV_READY;
 
@@ -373,6 +387,7 @@ vduse_events_handler(int fd, void *arg, int *remove __rte_unused)
 				req.s.status);
 		old_status = dev->status;
 		dev->status = req.s.status;
+		dev->reconnect_log->status = dev->status;
 		resp.result = VDUSE_REQ_RESULT_OK;
 		break;
 	case VDUSE_UPDATE_IOTLB:
@@ -398,7 +413,7 @@ vduse_events_handler(int fd, void *arg, int *remove __rte_unused)
 
 	if ((old_status ^ dev->status) & VIRTIO_DEVICE_STATUS_DRIVER_OK) {
 		if (dev->status & VIRTIO_DEVICE_STATUS_DRIVER_OK)
-			vduse_device_start(dev);
+			vduse_device_start(dev, false);
 		else
 			vduse_device_stop(dev);
 	}
@@ -407,10 +422,64 @@ vduse_events_handler(int fd, void *arg, int *remove __rte_unused)
 			vduse_req_id_to_str(req.type), req.type);
 }
 
+static char vduse_reconnect_dir[PATH_MAX];
+static bool vduse_reconnect_path_set;
+
+static int
+vduse_reconnect_path_init(void)
+{
+	const char *directory;
+	int ret;
+
+	/* from RuntimeDirectory= see systemd.exec */
+	directory = getenv("RUNTIME_DIRECTORY");
+	if (directory == NULL) {
+		/*
+		 * Used standard convention defined in
+		 * XDG Base Directory Specification and
+		 * Filesystem Hierarchy Standard.
+		 */
+		if (getuid() == 0)
+			directory = "/var/run";
+		else
+			directory = getenv("XDG_RUNTIME_DIR") ? : "/tmp";
+	}
+
+	ret = snprintf(vduse_reconnect_dir, sizeof(vduse_reconnect_dir), "%s/vduse",
+			directory);
+	if (ret < 0 || ret == sizeof(vduse_reconnect_dir)) {
+		VHOST_CONFIG_LOG("vduse", ERR, "Error creating VDUSE reconnect path name");
+		return -1;
+	}
+
+	ret = mkdir(vduse_reconnect_dir, 0700);
+	if (ret < 0 && errno != EEXIST) {
+		VHOST_CONFIG_LOG("vduse", ERR, "Error creating '%s': %s",
+				vduse_reconnect_dir, strerror(errno));
+		return -1;
+	}
+
+	VHOST_CONFIG_LOG("vduse", INFO, "Created VDUSE reconnect directory in %s",
+			vduse_reconnect_dir);
+
+	return 0;
+}
+
+static void
+vduse_reconnect_handler(int fd, void *arg, int *remove)
+{
+	struct virtio_net *dev = arg;
+
+	vduse_device_start(dev, true);
+
+	close(fd);
+	*remove = 1;
+}
+
 int
 vduse_device_create(const char *path, bool compliant_ol_flags)
 {
-	int control_fd, dev_fd, vid, ret;
+	int control_fd, dev_fd, vid, ret, reco_fd;
 	uint32_t i, max_queue_pairs, total_queues;
 	struct virtio_net *dev;
 	struct virtio_net_config vnet_config = {{ 0 }};
@@ -418,6 +487,9 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
 	uint64_t features;
 	struct vduse_dev_config *dev_config = NULL;
 	const char *name = path + strlen("/dev/vduse/");
+	char reconnect_file[PATH_MAX];
+	struct vhost_reconnect_data *reconnect_log = NULL;
+	bool reconnect = false;
 
 	if (vduse.fdset == NULL) {
 		vduse.fdset = fdset_init("vduse-evt");
@@ -427,6 +499,20 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
 		}
 	}
 
+	if (vduse_reconnect_path_set == false) {
+		if (vduse_reconnect_path_init() < 0) {
+			VHOST_CONFIG_LOG(path, ERR, "failed to initialize reconnect path");
+			return -1;
+		}
+		vduse_reconnect_path_set = true;
+	}
+
+	ret = snprintf(reconnect_file, sizeof(reconnect_file), "%s/%s", vduse_reconnect_dir, name);
+	if (ret < 0 || ret == sizeof(reconnect_file)) {
+		VHOST_CONFIG_LOG(name, ERR, "Failed to create vduse reconnect path name");
+		return -1;
+	}
+
 	control_fd = open(VDUSE_CTRL_PATH, O_RDWR);
 	if (control_fd < 0) {
 		VHOST_CONFIG_LOG(name, ERR, "Failed to open %s: %s",
@@ -441,14 +527,6 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
 		goto out_ctrl_close;
 	}
 
-	dev_config = malloc(offsetof(struct vduse_dev_config, config) +
-			sizeof(vnet_config));
-	if (!dev_config) {
-		VHOST_CONFIG_LOG(name, ERR, "Failed to allocate VDUSE config");
-		ret = -1;
-		goto out_ctrl_close;
-	}
-
 	ret = rte_vhost_driver_get_features(path, &features);
 	if (ret < 0) {
 		VHOST_CONFIG_LOG(name, ERR, "Failed to get backend features");
@@ -469,23 +547,118 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
 	else
 		total_queues += 1; /* Includes ctrl queue */
 
-	vnet_config.max_virtqueue_pairs = max_queue_pairs;
-	memset(dev_config, 0, sizeof(struct vduse_dev_config));
+	if (access(path, F_OK) == 0) {
+		VHOST_CONFIG_LOG(name, INFO, "Device already exists, reconnecting...");
+		reconnect = true;
+
+		reco_fd = open(reconnect_file, O_RDWR, 0600);
+		if (reco_fd < 0) {
+			if (errno == ENOENT)
+				VHOST_CONFIG_LOG(name, ERR, "Missing reconnect file (%s)",
+						reconnect_file);
+			else
+				VHOST_CONFIG_LOG(name, ERR, "Failed to open reconnect file %s (%s)",
+						reconnect_file, strerror(errno));
+			ret = -1;
+			goto out_ctrl_close;
+		}
 
-	strncpy(dev_config->name, name, VDUSE_NAME_MAX - 1);
-	dev_config->device_id = VIRTIO_ID_NET;
-	dev_config->vendor_id = 0;
-	dev_config->features = features;
-	dev_config->vq_num = total_queues;
-	dev_config->vq_align = sysconf(_SC_PAGE_SIZE);
-	dev_config->config_size = sizeof(struct virtio_net_config);
-	memcpy(dev_config->config, &vnet_config, sizeof(vnet_config));
+		reconnect_log = mmap(NULL, sizeof(*reconnect_log), PROT_READ | PROT_WRITE,
+				MAP_SHARED, reco_fd, 0);
+		close(reco_fd);
+		if (reconnect_log == MAP_FAILED) {
+			VHOST_CONFIG_LOG(name, ERR, "Failed to mmap reconnect file %s (%s)",
+					reconnect_file, strerror(errno));
+			ret = -1;
+			goto out_ctrl_close;
+		}
 
-	ret = ioctl(control_fd, VDUSE_CREATE_DEV, dev_config);
-	if (ret < 0) {
-		VHOST_CONFIG_LOG(name, ERR, "Failed to create VDUSE device: %s",
-				strerror(errno));
-		goto out_free;
+		if (reconnect_log->version != VHOST_RECONNECT_VERSION) {
+			VHOST_CONFIG_LOG(name, ERR,
+					"Version mismatch between backend (0x%x) & reconnection file (0x%x)",
+					VHOST_RECONNECT_VERSION, reconnect_log->version);
+		}
+
+		if ((reconnect_log->features & features) != reconnect_log->features) {
+			VHOST_CONFIG_LOG(name, ERR,
+					"Features mismatch between backend (0x%" PRIx64 ") & reconnection file (0x%" PRIx64 ")",
+					features, reconnect_log->features);
+			ret = -1;
+			goto out_ctrl_close;
+		}
+
+		if (reconnect_log->nr_vrings != total_queues) {
+			VHOST_CONFIG_LOG(name, ERR,
+					"Queues number mismatch between backend (%u) and reconnection file (%u)",
+					total_queues, reconnect_log->nr_vrings);
+			ret = -1;
+			goto out_ctrl_close;
+		}
+	} else {
+		reco_fd = open(reconnect_file, O_CREAT | O_EXCL | O_RDWR, 0600);
+		if (reco_fd < 0) {
+			if (errno == EEXIST) {
+				VHOST_CONFIG_LOG(name, ERR, "Reconnect file %s exists but not the device",
+						reconnect_file);
+			} else {
+				VHOST_CONFIG_LOG(name, ERR, "Failed to open reconnect file %s (%s)",
+						reconnect_file, strerror(errno));
+			}
+			ret = -1;
+			goto out_ctrl_close;
+		}
+
+		ret = ftruncate(reco_fd, sizeof(*reconnect_log));
+		if (ret < 0) {
+			VHOST_CONFIG_LOG(name, ERR, "Failed to truncate reconnect file %s (%s)",
+					reconnect_file, strerror(errno));
+			close(reco_fd);
+			goto out_ctrl_close;
+		}
+
+		reconnect_log = mmap(NULL, sizeof(*reconnect_log), PROT_READ | PROT_WRITE,
+					MAP_SHARED, reco_fd, 0);
+		close(reco_fd);
+		if (reconnect_log == MAP_FAILED) {
+			VHOST_CONFIG_LOG(name, ERR, "Failed to mmap reconnect file %s (%s)",
+					reconnect_file, strerror(errno));
+			ret = -1;
+			goto out_ctrl_close;
+		}
+
+		reconnect_log->version = VHOST_RECONNECT_VERSION;
+
+		dev_config = malloc(offsetof(struct vduse_dev_config, config) +
+				sizeof(vnet_config));
+		if (!dev_config) {
+			VHOST_CONFIG_LOG(name, ERR, "Failed to allocate VDUSE config");
+			ret = -1;
+			goto out_ctrl_close;
+		}
+
+		vnet_config.max_virtqueue_pairs = max_queue_pairs;
+		memset(dev_config, 0, sizeof(struct vduse_dev_config));
+
+		rte_strscpy(dev_config->name, name, VDUSE_NAME_MAX - 1);
+		dev_config->device_id = VIRTIO_ID_NET;
+		dev_config->vendor_id = 0;
+		dev_config->features = features;
+		dev_config->vq_num = total_queues;
+		dev_config->vq_align = sysconf(_SC_PAGE_SIZE);
+		dev_config->config_size = sizeof(struct virtio_net_config);
+		memcpy(dev_config->config, &vnet_config, sizeof(vnet_config));
+
+		ret = ioctl(control_fd, VDUSE_CREATE_DEV, dev_config);
+		if (ret < 0) {
+			VHOST_CONFIG_LOG(name, ERR, "Failed to create VDUSE device: %s",
+					strerror(errno));
+			goto out_free;
+		}
+
+		memcpy(&reconnect_log->config, &vnet_config, sizeof(vnet_config));
+		reconnect_log->nr_vrings = total_queues;
+		free(dev_config);
+		dev_config = NULL;
 	}
 
 	dev_fd = open(path, O_RDWR);
@@ -519,10 +692,15 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
 	strncpy(dev->ifname, path, IF_NAME_SZ - 1);
 	dev->vduse_ctrl_fd = control_fd;
 	dev->vduse_dev_fd = dev_fd;
+	dev->reconnect_log = reconnect_log;
+	if (reconnect)
+		dev->status = dev->reconnect_log->status;
+
 	vhost_setup_virtio_net(dev->vid, true, compliant_ol_flags, true, true);
 
 	for (i = 0; i < total_queues; i++) {
 		struct vduse_vq_config vq_cfg = { 0 };
+		struct vhost_virtqueue *vq;
 
 		ret = alloc_vring_queue(dev, i);
 		if (ret) {
@@ -530,6 +708,12 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
 			goto out_dev_destroy;
 		}
 
+		vq = dev->virtqueue[i];
+		vq->reconnect_log = &reconnect_log->vring[i];
+
+		if (reconnect)
+			continue;
+
 		vq_cfg.index = i;
 		vq_cfg.max_size = 1024;
 
@@ -549,7 +733,32 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
 		goto out_dev_destroy;
 	}
 
-	free(dev_config);
+	if (reconnect && dev->status & VIRTIO_DEVICE_STATUS_DRIVER_OK)  {
+		/*
+		 * Make vduse_device_start() being executed in the same
+		 * context for both reconnection and fresh startup.
+		 */
+		reco_fd = eventfd(0, EFD_NONBLOCK | EFD_CLOEXEC);
+		if (reco_fd < 0) {
+			VHOST_CONFIG_LOG(name, ERR, "Failed to create reco_fd: %s",
+					strerror(errno));
+			ret = -1;
+			goto out_dev_destroy;
+		}
+
+		ret = fdset_add(vduse.fdset, reco_fd, vduse_reconnect_handler, NULL, dev);
+		if (ret) {
+			VHOST_CONFIG_LOG(name, ERR, "Failed to add reconnect fd %d to vduse fdset",
+					reco_fd);
+			goto out_dev_destroy;
+		}
+
+		ret = eventfd_write(reco_fd, (eventfd_t)1);
+		if (ret < 0) {
+			VHOST_CONFIG_LOG(name, ERR, "Failed to write to reconnect eventfd");
+			goto out_dev_destroy;
+		}
+	}
 
 	return 0;
 
@@ -587,6 +796,9 @@ vduse_device_destroy(const char *path)
 	if (vid == RTE_MAX_VHOST_DEVICE)
 		return -1;
 
+	if (dev->reconnect_log)
+		munmap(dev->reconnect_log, sizeof(*dev->reconnect_log));
+
 	vduse_device_stop(dev);
 
 	fdset_del(vduse.fdset, dev->vduse_dev_fd);
@@ -597,10 +809,26 @@ vduse_device_destroy(const char *path)
 	}
 
 	if (dev->vduse_ctrl_fd >= 0) {
+		char reconnect_file[PATH_MAX];
+
 		ret = ioctl(dev->vduse_ctrl_fd, VDUSE_DESTROY_DEV, name);
-		if (ret)
+		if (ret) {
 			VHOST_CONFIG_LOG(name, ERR, "Failed to destroy VDUSE device: %s",
 					strerror(errno));
+		} else {
+			/*
+			 * VDUSE device was no more attached to the vDPA bus,
+			 * so we can remove the reconnect file.
+			 */
+			ret = snprintf(reconnect_file, sizeof(reconnect_file), "%s/%s",
+					vduse_reconnect_dir, name);
+			if (ret < 0 || ret == sizeof(reconnect_file))
+				VHOST_CONFIG_LOG(name, ERR,
+						"Failed to create vduse reconnect path name");
+			else
+				unlink(reconnect_file);
+		}
+
 		close(dev->vduse_ctrl_fd);
 		dev->vduse_ctrl_fd = -1;
 	}
-- 
2.46.0


^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH v2 1/2] vhost: add logging mechanism for reconnection
  2024-09-20 15:09 ` [PATCH v2 1/2] vhost: add logging mechanism for reconnection Maxime Coquelin
@ 2024-09-23  7:12   ` Chenbo Xia
  2024-09-23 15:42   ` David Marchand
  1 sibling, 0 replies; 8+ messages in thread
From: Chenbo Xia @ 2024-09-23  7:12 UTC (permalink / raw)
  To: Maxime Coquelin; +Cc: dev, David Marchand


> On Sep 20, 2024, at 23:09, Maxime Coquelin <maxime.coquelin@redhat.com> wrote:
> 
> External email: Use caution opening links or attachments
> 
> 
> This patch introduces a way for backend to keep track
> of the needed information to be able to reconnect without
> frontend cooperation.
> 
> It will be used for VDUSE, which does not provide interface
> for the backend to save and later recover local virtqueues
> metadata needed to reconnect.
> 
> Vhost-user support could also be added for improved packed
> ring reconnection support.
> 
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
> ---
> lib/vhost/vhost.h           | 41 ++++++++++++++++++++++++++++++++++---
> lib/vhost/virtio_net.c      |  8 ++++++++
> lib/vhost/virtio_net_ctrl.c |  2 ++
> 3 files changed, 48 insertions(+), 3 deletions(-)
> 
> diff --git a/lib/vhost/vhost.h b/lib/vhost/vhost.h
> index cd3fa55f1b..1f4192f5d1 100644
> --- a/lib/vhost/vhost.h
> +++ b/lib/vhost/vhost.h
> @@ -269,6 +269,24 @@ struct vhost_async {
>        };
> };
> 
> +#define VHOST_RECONNECT_VERSION                0x0
> +#define VHOST_MAX_VRING                        0x100
> +#define VHOST_MAX_QUEUE_PAIRS          0x80
> +
> +struct __rte_cache_aligned vhost_reconnect_vring {
> +       uint16_t last_avail_idx;
> +       bool avail_wrap_counter;
> +};
> +
> +struct vhost_reconnect_data {
> +       uint32_t version;
> +       uint64_t features;
> +       uint8_t status;
> +       struct virtio_net_config config;
> +       uint32_t nr_vrings;
> +       struct vhost_reconnect_vring vring[VHOST_MAX_VRING];
> +};
> +
> /**
>  * Structure contains variables relevant to RX/TX virtqueues.
>  */
> @@ -351,6 +369,7 @@ struct __rte_cache_aligned vhost_virtqueue {
>        struct virtqueue_stats  stats;
> 
>        RTE_ATOMIC(bool) irq_pending;
> +       struct vhost_reconnect_vring *reconnect_log;
> };
> 
> /* Virtio device status as per Virtio specification */
> @@ -362,9 +381,6 @@ struct __rte_cache_aligned vhost_virtqueue {
> #define VIRTIO_DEVICE_STATUS_DEV_NEED_RESET    0x40
> #define VIRTIO_DEVICE_STATUS_FAILED            0x80
> 
> -#define VHOST_MAX_VRING                        0x100
> -#define VHOST_MAX_QUEUE_PAIRS          0x80
> -
> /* Declare IOMMU related bits for older kernels */
> #ifndef VIRTIO_F_IOMMU_PLATFORM
> 
> @@ -538,8 +554,26 @@ struct __rte_cache_aligned virtio_net {
>        struct rte_vhost_user_extern_ops extern_ops;
> 
>        struct vhost_backend_ops *backend_ops;
> +
> +       struct vhost_reconnect_data *reconnect_log;
> };
> 
> +static __rte_always_inline void
> +vhost_virtqueue_reconnect_log_split(struct vhost_virtqueue *vq)
> +{
> +       if (vq->reconnect_log != NULL)
> +               vq->reconnect_log->last_avail_idx = vq->last_avail_idx;
> +}
> +
> +static __rte_always_inline void
> +vhost_virtqueue_reconnect_log_packed(struct vhost_virtqueue *vq)
> +{
> +       if (vq->reconnect_log != NULL) {
> +               vq->reconnect_log->last_avail_idx = vq->last_avail_idx;
> +               vq->reconnect_log->avail_wrap_counter = vq->avail_wrap_counter;
> +       }
> +}
> +
> static inline void
> vq_assert_lock__(struct virtio_net *dev, struct vhost_virtqueue *vq, const char *func)
>        __rte_assert_exclusive_lock(&vq->access_lock)
> @@ -584,6 +618,7 @@ vq_inc_last_avail_packed(struct vhost_virtqueue *vq, uint16_t num)
>                vq->avail_wrap_counter ^= 1;
>                vq->last_avail_idx -= vq->size;
>        }
> +       vhost_virtqueue_reconnect_log_packed(vq);
> }
> 
> void __vhost_log_cache_write(struct virtio_net *dev,
> diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
> index 370402d849..f66a0c82f8 100644
> --- a/lib/vhost/virtio_net.c
> +++ b/lib/vhost/virtio_net.c
> @@ -1445,6 +1445,7 @@ virtio_dev_rx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
>                }
> 
>                vq->last_avail_idx += num_buffers;
> +               vhost_virtqueue_reconnect_log_split(vq);
>        }
> 
>        do_data_copy_enqueue(dev, vq);
> @@ -1857,6 +1858,7 @@ virtio_dev_rx_async_submit_split(struct virtio_net *dev, struct vhost_virtqueue
>                pkts_info[slot_idx].mbuf = pkts[pkt_idx];
> 
>                vq->last_avail_idx += num_buffers;
> +               vhost_virtqueue_reconnect_log_split(vq);
>        }
> 
>        if (unlikely(pkt_idx == 0))
> @@ -1885,6 +1887,7 @@ virtio_dev_rx_async_submit_split(struct virtio_net *dev, struct vhost_virtqueue
>                /* recover shadow used ring and available ring */
>                vq->shadow_used_idx -= num_descs;
>                vq->last_avail_idx -= num_descs;
> +               vhost_virtqueue_reconnect_log_split(vq);
>        }
> 
>        /* keep used descriptors */
> @@ -2100,6 +2103,7 @@ dma_error_handler_packed(struct vhost_virtqueue *vq, uint16_t slot_idx,
>                vq->last_avail_idx = vq->last_avail_idx + vq->size - descs_err;
>                vq->avail_wrap_counter ^= 1;
>        }
> +       vhost_virtqueue_reconnect_log_packed(vq);
> 
>        if (async->buffer_idx_packed >= buffers_err)
>                async->buffer_idx_packed -= buffers_err;
> @@ -3182,6 +3186,7 @@ virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
> 
>        if (likely(vq->shadow_used_idx)) {
>                vq->last_avail_idx += vq->shadow_used_idx;
> +               vhost_virtqueue_reconnect_log_split(vq);
>                do_data_copy_dequeue(vq);
>                flush_shadow_used_ring_split(dev, vq);
>                vhost_vring_call_split(dev, vq);
> @@ -3854,6 +3859,7 @@ virtio_dev_tx_async_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
>                async->desc_idx_split++;
> 
>                vq->last_avail_idx++;
> +               vhost_virtqueue_reconnect_log_split(vq);
>        }
> 
>        if (unlikely(dropped))
> @@ -3872,6 +3878,7 @@ virtio_dev_tx_async_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
>                pkt_idx = n_xfer;
>                /* recover available ring */
>                vq->last_avail_idx -= pkt_err;
> +               vhost_virtqueue_reconnect_log_split(vq);
> 
>                /**
>                 * recover async channel copy related structures and free pktmbufs
> @@ -4153,6 +4160,7 @@ virtio_dev_tx_async_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,
>                        vq->last_avail_idx += vq->size - descs_err;
>                        vq->avail_wrap_counter ^= 1;
>                }
> +               vhost_virtqueue_reconnect_log_packed(vq);
>        }
> 
>        async->pkts_idx += pkt_idx;
> diff --git a/lib/vhost/virtio_net_ctrl.c b/lib/vhost/virtio_net_ctrl.c
> index 8f78122361..b8ee94018e 100644
> --- a/lib/vhost/virtio_net_ctrl.c
> +++ b/lib/vhost/virtio_net_ctrl.c
> @@ -169,6 +169,7 @@ virtio_net_ctrl_pop(struct virtio_net *dev, struct vhost_virtqueue *cvq,
>        cvq->last_avail_idx++;
>        if (cvq->last_avail_idx >= cvq->size)
>                cvq->last_avail_idx -= cvq->size;
> +       vhost_virtqueue_reconnect_log_split(cvq);
> 
>        if (dev->features & (1ULL << VIRTIO_RING_F_EVENT_IDX))
>                vhost_avail_event(cvq) = cvq->last_avail_idx;
> @@ -181,6 +182,7 @@ virtio_net_ctrl_pop(struct virtio_net *dev, struct vhost_virtqueue *cvq,
>        cvq->last_avail_idx++;
>        if (cvq->last_avail_idx >= cvq->size)
>                cvq->last_avail_idx -= cvq->size;
> +       vhost_virtqueue_reconnect_log_split(cvq);
> 
>        if (dev->features & (1ULL << VIRTIO_RING_F_EVENT_IDX))
>                vhost_avail_event(cvq) = cvq->last_avail_idx;
> --
> 2.46.0
> 

Reviewed-by: Chenbo Xia <chenbox@nvidia.com>


^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH v2 2/2] vhost: add reconnection support to VDUSE
  2024-09-20 15:09 ` [PATCH v2 2/2] vhost: add reconnection support to VDUSE Maxime Coquelin
@ 2024-09-23  7:13   ` Chenbo Xia
  2024-09-23 16:43   ` David Marchand
  1 sibling, 0 replies; 8+ messages in thread
From: Chenbo Xia @ 2024-09-23  7:13 UTC (permalink / raw)
  To: Maxime Coquelin; +Cc: dev, David Marchand


> On Sep 20, 2024, at 23:09, Maxime Coquelin <maxime.coquelin@redhat.com> wrote:
> 
> External email: Use caution opening links or attachments
> 
> 
> This patch enables VDUSE reconnection support making use of
> the newly introduced reconnection mechanism in Vhost
> library.
> 
> At DPDK VDUSE device creation time, there are two
> possibilities:
> 1. The Kernel VDUSE device does not exist:
>  a. A reconnection file named after the VUDSE device name
>     is created in VDUSE tmpfs.
>  b. The file is truncated to 'struct vhost_reconnect_data'
>     size, and mmapped.
>  c. Negotiated features, Virtio status... are saved for
>     sanity checks at reconnect time.
> 2. The Kernel VDUSE device already exists:
>  a. Exit with failure if no reconnect file exists for
>     this device.
>  b. Open and mmap the reconnect file.
>  c. Perform sanity check to ensure features are compatible.
>  d. Restore virtqueues' available indexes at startup time.
> 
> Then at runtime, the virtqueues' available index are logged by
> the Vhost reconnection mechanism.
> 
> At DPDK VDUSE device destruction time, there are two
> possibilities:
> 1. The Kernel VDUSE device destruction succeed, which
>    means it is no more attached to the vDPA bus. The
>    reconnection file is unmapped and then removed.
> 2. The Kernel VDUSE device destruction failed, meaning it
>    is no more attached to the vDPA bus. The reconnection
>    file is unmapped but not removed to make possible later
>    reconnection.
> 
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
> ---
> lib/vhost/vduse.c | 308 ++++++++++++++++++++++++++++++++++++++++------
> 1 file changed, 268 insertions(+), 40 deletions(-)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index c66602905c..f9ac317438 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -136,7 +136,7 @@ vduse_control_queue_event(int fd, void *arg, int *remove __rte_unused)
> }
> 
> static void
> -vduse_vring_setup(struct virtio_net *dev, unsigned int index)
> +vduse_vring_setup(struct virtio_net *dev, unsigned int index, bool reconnect)
> {
>        struct vhost_virtqueue *vq = dev->virtqueue[index];
>        struct vhost_vring_addr *ra = &vq->ring_addrs;
> @@ -152,6 +152,19 @@ vduse_vring_setup(struct virtio_net *dev, unsigned int index)
>                return;
>        }
> 
> +       if (reconnect) {
> +               vq->last_avail_idx = vq->reconnect_log->last_avail_idx;
> +               vq->last_used_idx = vq->reconnect_log->last_avail_idx;
> +       } else {
> +               vq->last_avail_idx = vq_info.split.avail_index;
> +               vq->last_used_idx = vq_info.split.avail_index;
> +       }
> +       vq->size = vq_info.num;
> +       vq->ready = true;
> +       vq->enabled = vq_info.ready;
> +       ra->desc_user_addr = vq_info.desc_addr;
> +       ra->avail_user_addr = vq_info.driver_addr;
> +       ra->used_user_addr = vq_info.device_addr;
>        VHOST_CONFIG_LOG(dev->ifname, INFO, "VQ %u info:", index);
>        VHOST_CONFIG_LOG(dev->ifname, INFO, "\tnum: %u", vq_info.num);
>        VHOST_CONFIG_LOG(dev->ifname, INFO, "\tdesc_addr: %llx",
> @@ -160,17 +173,9 @@ vduse_vring_setup(struct virtio_net *dev, unsigned int index)
>                        (unsigned long long)vq_info.driver_addr);
>        VHOST_CONFIG_LOG(dev->ifname, INFO, "\tdevice_addr: %llx",
>                        (unsigned long long)vq_info.device_addr);
> -       VHOST_CONFIG_LOG(dev->ifname, INFO, "\tavail_idx: %u", vq_info.split.avail_index);
> +       VHOST_CONFIG_LOG(dev->ifname, INFO, "\tavail_idx: %u", vq->last_avail_idx);
> +       VHOST_CONFIG_LOG(dev->ifname, INFO, "\tused_idx: %u", vq->last_used_idx);
>        VHOST_CONFIG_LOG(dev->ifname, INFO, "\tready: %u", vq_info.ready);
> -
> -       vq->last_avail_idx = vq_info.split.avail_index;
> -       vq->size = vq_info.num;
> -       vq->ready = true;
> -       vq->enabled = vq_info.ready;
> -       ra->desc_user_addr = vq_info.desc_addr;
> -       ra->avail_user_addr = vq_info.driver_addr;
> -       ra->used_user_addr = vq_info.device_addr;
> -
>        vq->kickfd = eventfd(0, EFD_NONBLOCK | EFD_CLOEXEC);
>        if (vq->kickfd < 0) {
>                VHOST_CONFIG_LOG(dev->ifname, ERR, "Failed to init kickfd for VQ %u: %s",
> @@ -267,7 +272,7 @@ vduse_vring_cleanup(struct virtio_net *dev, unsigned int index)
> }
> 
> static void
> -vduse_device_start(struct virtio_net *dev)
> +vduse_device_start(struct virtio_net *dev, bool reconnect)
> {
>        unsigned int i, ret;
> 
> @@ -287,6 +292,15 @@ vduse_device_start(struct virtio_net *dev)
>                return;
>        }
> 
> +       if (reconnect && dev->features != dev->reconnect_log->features) {
> +               VHOST_CONFIG_LOG(dev->ifname, ERR,
> +                               "Mismatch between reconnect file features 0x%" PRIx64 " & device features 0x%" PRIx64,
> +                               dev->reconnect_log->features, dev->features);
> +               return;
> +       }
> +
> +       dev->reconnect_log->features = dev->features;
> +
>        VHOST_CONFIG_LOG(dev->ifname, INFO, "Negotiated Virtio features: 0x%" PRIx64,
>                dev->features);
> 
> @@ -300,7 +314,7 @@ vduse_device_start(struct virtio_net *dev)
>        }
> 
>        for (i = 0; i < dev->nr_vring; i++)
> -               vduse_vring_setup(dev, i);
> +               vduse_vring_setup(dev, i, reconnect);
> 
>        dev->flags |= VIRTIO_DEV_READY;
> 
> @@ -373,6 +387,7 @@ vduse_events_handler(int fd, void *arg, int *remove __rte_unused)
>                                req.s.status);
>                old_status = dev->status;
>                dev->status = req.s.status;
> +               dev->reconnect_log->status = dev->status;
>                resp.result = VDUSE_REQ_RESULT_OK;
>                break;
>        case VDUSE_UPDATE_IOTLB:
> @@ -398,7 +413,7 @@ vduse_events_handler(int fd, void *arg, int *remove __rte_unused)
> 
>        if ((old_status ^ dev->status) & VIRTIO_DEVICE_STATUS_DRIVER_OK) {
>                if (dev->status & VIRTIO_DEVICE_STATUS_DRIVER_OK)
> -                       vduse_device_start(dev);
> +                       vduse_device_start(dev, false);
>                else
>                        vduse_device_stop(dev);
>        }
> @@ -407,10 +422,64 @@ vduse_events_handler(int fd, void *arg, int *remove __rte_unused)
>                        vduse_req_id_to_str(req.type), req.type);
> }
> 
> +static char vduse_reconnect_dir[PATH_MAX];
> +static bool vduse_reconnect_path_set;
> +
> +static int
> +vduse_reconnect_path_init(void)
> +{
> +       const char *directory;
> +       int ret;
> +
> +       /* from RuntimeDirectory= see systemd.exec */
> +       directory = getenv("RUNTIME_DIRECTORY");
> +       if (directory == NULL) {
> +               /*
> +                * Used standard convention defined in
> +                * XDG Base Directory Specification and
> +                * Filesystem Hierarchy Standard.
> +                */
> +               if (getuid() == 0)
> +                       directory = "/var/run";
> +               else
> +                       directory = getenv("XDG_RUNTIME_DIR") ? : "/tmp";
> +       }
> +
> +       ret = snprintf(vduse_reconnect_dir, sizeof(vduse_reconnect_dir), "%s/vduse",
> +                       directory);
> +       if (ret < 0 || ret == sizeof(vduse_reconnect_dir)) {
> +               VHOST_CONFIG_LOG("vduse", ERR, "Error creating VDUSE reconnect path name");
> +               return -1;
> +       }
> +
> +       ret = mkdir(vduse_reconnect_dir, 0700);
> +       if (ret < 0 && errno != EEXIST) {
> +               VHOST_CONFIG_LOG("vduse", ERR, "Error creating '%s': %s",
> +                               vduse_reconnect_dir, strerror(errno));
> +               return -1;
> +       }
> +
> +       VHOST_CONFIG_LOG("vduse", INFO, "Created VDUSE reconnect directory in %s",
> +                       vduse_reconnect_dir);
> +
> +       return 0;
> +}
> +
> +static void
> +vduse_reconnect_handler(int fd, void *arg, int *remove)
> +{
> +       struct virtio_net *dev = arg;
> +
> +       vduse_device_start(dev, true);
> +
> +       close(fd);
> +       *remove = 1;
> +}
> +
> int
> vduse_device_create(const char *path, bool compliant_ol_flags)
> {
> -       int control_fd, dev_fd, vid, ret;
> +       int control_fd, dev_fd, vid, ret, reco_fd;
>        uint32_t i, max_queue_pairs, total_queues;
>        struct virtio_net *dev;
>        struct virtio_net_config vnet_config = {{ 0 }};
> @@ -418,6 +487,9 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
>        uint64_t features;
>        struct vduse_dev_config *dev_config = NULL;
>        const char *name = path + strlen("/dev/vduse/");
> +       char reconnect_file[PATH_MAX];
> +       struct vhost_reconnect_data *reconnect_log = NULL;
> +       bool reconnect = false;
> 
>        if (vduse.fdset == NULL) {
>                vduse.fdset = fdset_init("vduse-evt");
> @@ -427,6 +499,20 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
>                }
>        }
> 
> +       if (vduse_reconnect_path_set == false) {
> +               if (vduse_reconnect_path_init() < 0) {
> +                       VHOST_CONFIG_LOG(path, ERR, "failed to initialize reconnect path");
> +                       return -1;
> +               }
> +               vduse_reconnect_path_set = true;
> +       }
> +
> +       ret = snprintf(reconnect_file, sizeof(reconnect_file), "%s/%s", vduse_reconnect_dir, name);
> +       if (ret < 0 || ret == sizeof(reconnect_file)) {
> +               VHOST_CONFIG_LOG(name, ERR, "Failed to create vduse reconnect path name");
> +               return -1;
> +       }
> +
>        control_fd = open(VDUSE_CTRL_PATH, O_RDWR);
>        if (control_fd < 0) {
>                VHOST_CONFIG_LOG(name, ERR, "Failed to open %s: %s",
> @@ -441,14 +527,6 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
>                goto out_ctrl_close;
>        }
> 
> -       dev_config = malloc(offsetof(struct vduse_dev_config, config) +
> -                       sizeof(vnet_config));
> -       if (!dev_config) {
> -               VHOST_CONFIG_LOG(name, ERR, "Failed to allocate VDUSE config");
> -               ret = -1;
> -               goto out_ctrl_close;
> -       }
> -
>        ret = rte_vhost_driver_get_features(path, &features);
>        if (ret < 0) {
>                VHOST_CONFIG_LOG(name, ERR, "Failed to get backend features");
> @@ -469,23 +547,118 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
>        else
>                total_queues += 1; /* Includes ctrl queue */
> 
> -       vnet_config.max_virtqueue_pairs = max_queue_pairs;
> -       memset(dev_config, 0, sizeof(struct vduse_dev_config));
> +       if (access(path, F_OK) == 0) {
> +               VHOST_CONFIG_LOG(name, INFO, "Device already exists, reconnecting...");
> +               reconnect = true;
> +
> +               reco_fd = open(reconnect_file, O_RDWR, 0600);
> +               if (reco_fd < 0) {
> +                       if (errno == ENOENT)
> +                               VHOST_CONFIG_LOG(name, ERR, "Missing reconnect file (%s)",
> +                                               reconnect_file);
> +                       else
> +                               VHOST_CONFIG_LOG(name, ERR, "Failed to open reconnect file %s (%s)",
> +                                               reconnect_file, strerror(errno));
> +                       ret = -1;
> +                       goto out_ctrl_close;
> +               }
> 
> -       strncpy(dev_config->name, name, VDUSE_NAME_MAX - 1);
> -       dev_config->device_id = VIRTIO_ID_NET;
> -       dev_config->vendor_id = 0;
> -       dev_config->features = features;
> -       dev_config->vq_num = total_queues;
> -       dev_config->vq_align = sysconf(_SC_PAGE_SIZE);
> -       dev_config->config_size = sizeof(struct virtio_net_config);
> -       memcpy(dev_config->config, &vnet_config, sizeof(vnet_config));
> +               reconnect_log = mmap(NULL, sizeof(*reconnect_log), PROT_READ | PROT_WRITE,
> +                               MAP_SHARED, reco_fd, 0);
> +               close(reco_fd);
> +               if (reconnect_log == MAP_FAILED) {
> +                       VHOST_CONFIG_LOG(name, ERR, "Failed to mmap reconnect file %s (%s)",
> +                                       reconnect_file, strerror(errno));
> +                       ret = -1;
> +                       goto out_ctrl_close;
> +               }
> 
> -       ret = ioctl(control_fd, VDUSE_CREATE_DEV, dev_config);
> -       if (ret < 0) {
> -               VHOST_CONFIG_LOG(name, ERR, "Failed to create VDUSE device: %s",
> -                               strerror(errno));
> -               goto out_free;
> +               if (reconnect_log->version != VHOST_RECONNECT_VERSION) {
> +                       VHOST_CONFIG_LOG(name, ERR,
> +                                       "Version mismatch between backend (0x%x) & reconnection file (0x%x)",
> +                                       VHOST_RECONNECT_VERSION, reconnect_log->version);
> +               }
> +
> +               if ((reconnect_log->features & features) != reconnect_log->features) {
> +                       VHOST_CONFIG_LOG(name, ERR,
> +                                       "Features mismatch between backend (0x%" PRIx64 ") & reconnection file (0x%" PRIx64 ")",
> +                                       features, reconnect_log->features);
> +                       ret = -1;
> +                       goto out_ctrl_close;
> +               }
> +
> +               if (reconnect_log->nr_vrings != total_queues) {
> +                       VHOST_CONFIG_LOG(name, ERR,
> +                                       "Queues number mismatch between backend (%u) and reconnection file (%u)",
> +                                       total_queues, reconnect_log->nr_vrings);
> +                       ret = -1;
> +                       goto out_ctrl_close;
> +               }
> +       } else {
> +               reco_fd = open(reconnect_file, O_CREAT | O_EXCL | O_RDWR, 0600);
> +               if (reco_fd < 0) {
> +                       if (errno == EEXIST) {
> +                               VHOST_CONFIG_LOG(name, ERR, "Reconnect file %s exists but not the device",
> +                                               reconnect_file);
> +                       } else {
> +                               VHOST_CONFIG_LOG(name, ERR, "Failed to open reconnect file %s (%s)",
> +                                               reconnect_file, strerror(errno));
> +                       }
> +                       ret = -1;
> +                       goto out_ctrl_close;
> +               }
> +
> +               ret = ftruncate(reco_fd, sizeof(*reconnect_log));
> +               if (ret < 0) {
> +                       VHOST_CONFIG_LOG(name, ERR, "Failed to truncate reconnect file %s (%s)",
> +                                       reconnect_file, strerror(errno));
> +                       close(reco_fd);
> +                       goto out_ctrl_close;
> +               }
> +
> +               reconnect_log = mmap(NULL, sizeof(*reconnect_log), PROT_READ | PROT_WRITE,
> +                                       MAP_SHARED, reco_fd, 0);
> +               close(reco_fd);
> +               if (reconnect_log == MAP_FAILED) {
> +                       VHOST_CONFIG_LOG(name, ERR, "Failed to mmap reconnect file %s (%s)",
> +                                       reconnect_file, strerror(errno));
> +                       ret = -1;
> +                       goto out_ctrl_close;
> +               }
> +
> +               reconnect_log->version = VHOST_RECONNECT_VERSION;
> +
> +               dev_config = malloc(offsetof(struct vduse_dev_config, config) +
> +                               sizeof(vnet_config));
> +               if (!dev_config) {
> +                       VHOST_CONFIG_LOG(name, ERR, "Failed to allocate VDUSE config");
> +                       ret = -1;
> +                       goto out_ctrl_close;
> +               }
> +
> +               vnet_config.max_virtqueue_pairs = max_queue_pairs;
> +               memset(dev_config, 0, sizeof(struct vduse_dev_config));
> +
> +               rte_strscpy(dev_config->name, name, VDUSE_NAME_MAX - 1);
> +               dev_config->device_id = VIRTIO_ID_NET;
> +               dev_config->vendor_id = 0;
> +               dev_config->features = features;
> +               dev_config->vq_num = total_queues;
> +               dev_config->vq_align = sysconf(_SC_PAGE_SIZE);
> +               dev_config->config_size = sizeof(struct virtio_net_config);
> +               memcpy(dev_config->config, &vnet_config, sizeof(vnet_config));
> +
> +               ret = ioctl(control_fd, VDUSE_CREATE_DEV, dev_config);
> +               if (ret < 0) {
> +                       VHOST_CONFIG_LOG(name, ERR, "Failed to create VDUSE device: %s",
> +                                       strerror(errno));
> +                       goto out_free;
> +               }
> +
> +               memcpy(&reconnect_log->config, &vnet_config, sizeof(vnet_config));
> +               reconnect_log->nr_vrings = total_queues;
> +               free(dev_config);
> +               dev_config = NULL;
>        }
> 
>        dev_fd = open(path, O_RDWR);
> @@ -519,10 +692,15 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
>        strncpy(dev->ifname, path, IF_NAME_SZ - 1);
>        dev->vduse_ctrl_fd = control_fd;
>        dev->vduse_dev_fd = dev_fd;
> +       dev->reconnect_log = reconnect_log;
> +       if (reconnect)
> +               dev->status = dev->reconnect_log->status;
> +
>        vhost_setup_virtio_net(dev->vid, true, compliant_ol_flags, true, true);
> 
>        for (i = 0; i < total_queues; i++) {
>                struct vduse_vq_config vq_cfg = { 0 };
> +               struct vhost_virtqueue *vq;
> 
>                ret = alloc_vring_queue(dev, i);
>                if (ret) {
> @@ -530,6 +708,12 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
>                        goto out_dev_destroy;
>                }
> 
> +               vq = dev->virtqueue[i];
> +               vq->reconnect_log = &reconnect_log->vring[i];
> +
> +               if (reconnect)
> +                       continue;
> +
>                vq_cfg.index = i;
>                vq_cfg.max_size = 1024;
> 
> @@ -549,7 +733,32 @@ vduse_device_create(const char *path, bool compliant_ol_flags)
>                goto out_dev_destroy;
>        }
> 
> -       free(dev_config);
> +       if (reconnect && dev->status & VIRTIO_DEVICE_STATUS_DRIVER_OK)  {
> +               /*
> +                * Make vduse_device_start() being executed in the same
> +                * context for both reconnection and fresh startup.
> +                */
> +               reco_fd = eventfd(0, EFD_NONBLOCK | EFD_CLOEXEC);
> +               if (reco_fd < 0) {
> +                       VHOST_CONFIG_LOG(name, ERR, "Failed to create reco_fd: %s",
> +                                       strerror(errno));
> +                       ret = -1;
> +                       goto out_dev_destroy;
> +               }
> +
> +               ret = fdset_add(vduse.fdset, reco_fd, vduse_reconnect_handler, NULL, dev);
> +               if (ret) {
> +                       VHOST_CONFIG_LOG(name, ERR, "Failed to add reconnect fd %d to vduse fdset",
> +                                       reco_fd);
> +                       goto out_dev_destroy;
> +               }
> +
> +               ret = eventfd_write(reco_fd, (eventfd_t)1);
> +               if (ret < 0) {
> +                       VHOST_CONFIG_LOG(name, ERR, "Failed to write to reconnect eventfd");
> +                       goto out_dev_destroy;
> +               }
> +       }
> 
>        return 0;
> 
> @@ -587,6 +796,9 @@ vduse_device_destroy(const char *path)
>        if (vid == RTE_MAX_VHOST_DEVICE)
>                return -1;
> 
> +       if (dev->reconnect_log)
> +               munmap(dev->reconnect_log, sizeof(*dev->reconnect_log));
> +
>        vduse_device_stop(dev);
> 
>        fdset_del(vduse.fdset, dev->vduse_dev_fd);
> @@ -597,10 +809,26 @@ vduse_device_destroy(const char *path)
>        }
> 
>        if (dev->vduse_ctrl_fd >= 0) {
> +               char reconnect_file[PATH_MAX];
> +
>                ret = ioctl(dev->vduse_ctrl_fd, VDUSE_DESTROY_DEV, name);
> -               if (ret)
> +               if (ret) {
>                        VHOST_CONFIG_LOG(name, ERR, "Failed to destroy VDUSE device: %s",
>                                        strerror(errno));
> +               } else {
> +                       /*
> +                        * VDUSE device was no more attached to the vDPA bus,
> +                        * so we can remove the reconnect file.
> +                        */
> +                       ret = snprintf(reconnect_file, sizeof(reconnect_file), "%s/%s",
> +                                       vduse_reconnect_dir, name);
> +                       if (ret < 0 || ret == sizeof(reconnect_file))
> +                               VHOST_CONFIG_LOG(name, ERR,
> +                                               "Failed to create vduse reconnect path name");
> +                       else
> +                               unlink(reconnect_file);
> +               }
> +
>                close(dev->vduse_ctrl_fd);
>                dev->vduse_ctrl_fd = -1;
>        }
> --
> 2.46.0
> 

Reviewed-by: Chenbo Xia <chenbox@nvidia.com>


^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH v2 1/2] vhost: add logging mechanism for reconnection
  2024-09-20 15:09 ` [PATCH v2 1/2] vhost: add logging mechanism for reconnection Maxime Coquelin
  2024-09-23  7:12   ` Chenbo Xia
@ 2024-09-23 15:42   ` David Marchand
  2024-09-23 19:46     ` Maxime Coquelin
  1 sibling, 1 reply; 8+ messages in thread
From: David Marchand @ 2024-09-23 15:42 UTC (permalink / raw)
  To: Maxime Coquelin; +Cc: dev, chenbox

On Fri, Sep 20, 2024 at 11:09 AM Maxime Coquelin
<maxime.coquelin@redhat.com> wrote:
>
> This patch introduces a way for backend to keep track
> of the needed information to be able to reconnect without
> frontend cooperation.
>
> It will be used for VDUSE, which does not provide interface
> for the backend to save and later recover local virtqueues
> metadata needed to reconnect.
>
> Vhost-user support could also be added for improved packed
> ring reconnection support.
>
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>

vq->last_avail_idx gets updated in other places and I suspect we are
missing some calls to vhost_virtqueue_reconnect_log_split/packed.
I spotted:
- lib/vhost/vhost.c: rte_vhost_set_vring_base()
- lib/vhost/vhost_user.c:  translate_ring_addresses(),
vhost_user_set_vring_base().

The rest lgtm.


-- 
David Marchand


^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH v2 2/2] vhost: add reconnection support to VDUSE
  2024-09-20 15:09 ` [PATCH v2 2/2] vhost: add reconnection support to VDUSE Maxime Coquelin
  2024-09-23  7:13   ` Chenbo Xia
@ 2024-09-23 16:43   ` David Marchand
  1 sibling, 0 replies; 8+ messages in thread
From: David Marchand @ 2024-09-23 16:43 UTC (permalink / raw)
  To: Maxime Coquelin; +Cc: dev, chenbox

On Fri, Sep 20, 2024 at 11:09 AM Maxime Coquelin
<maxime.coquelin@redhat.com> wrote:
>
> This patch enables VDUSE reconnection support making use of
> the newly introduced reconnection mechanism in Vhost
> library.
>
> At DPDK VDUSE device creation time, there are two
> possibilities:
>  1. The Kernel VDUSE device does not exist:
>   a. A reconnection file named after the VUDSE device name

VDUSE*

>      is created in VDUSE tmpfs.
>   b. The file is truncated to 'struct vhost_reconnect_data'
>      size, and mmapped.
>   c. Negotiated features, Virtio status... are saved for
>      sanity checks at reconnect time.
>  2. The Kernel VDUSE device already exists:
>   a. Exit with failure if no reconnect file exists for
>      this device.
>   b. Open and mmap the reconnect file.
>   c. Perform sanity check to ensure features are compatible.
>   d. Restore virtqueues' available indexes at startup time.
>
> Then at runtime, the virtqueues' available index are logged by
> the Vhost reconnection mechanism.
>
> At DPDK VDUSE device destruction time, there are two
> possibilities:
>  1. The Kernel VDUSE device destruction succeed, which

succeeded*


>     means it is no more attached to the vDPA bus. The
>     reconnection file is unmapped and then removed.
>  2. The Kernel VDUSE device destruction failed, meaning it
>     is no more attached to the vDPA bus. The reconnection
>     file is unmapped but not removed to make possible later
>     reconnection.
>
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>

Reviewed-by: David Marchand <david.marchand@redhat.com>


-- 
David Marchand


^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH v2 1/2] vhost: add logging mechanism for reconnection
  2024-09-23 15:42   ` David Marchand
@ 2024-09-23 19:46     ` Maxime Coquelin
  0 siblings, 0 replies; 8+ messages in thread
From: Maxime Coquelin @ 2024-09-23 19:46 UTC (permalink / raw)
  To: David Marchand; +Cc: dev, chenbox



On 9/23/24 17:42, David Marchand wrote:
> On Fri, Sep 20, 2024 at 11:09 AM Maxime Coquelin
> <maxime.coquelin@redhat.com> wrote:
>>
>> This patch introduces a way for backend to keep track
>> of the needed information to be able to reconnect without
>> frontend cooperation.
>>
>> It will be used for VDUSE, which does not provide interface
>> for the backend to save and later recover local virtqueues
>> metadata needed to reconnect.
>>
>> Vhost-user support could also be added for improved packed
>> ring reconnection support.
>>
>> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
> 
> vq->last_avail_idx gets updated in other places and I suspect we are
> missing some calls to vhost_virtqueue_reconnect_log_split/packed.
> I spotted:
> - lib/vhost/vhost.c: rte_vhost_set_vring_base()
> - lib/vhost/vhost_user.c:  translate_ring_addresses(),
> vhost_user_set_vring_base().
> 
> The rest lgtm.
> 
> 

Thanks, I patched the missing avail index updates you identified.

And also added your R-by as asked off-list.

Maxime


^ permalink raw reply	[flat|nested] 8+ messages in thread

end of thread, other threads:[~2024-09-23 19:47 UTC | newest]

Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2024-09-20 15:09 [PATCH v2 0/2] vhost: add VDUSE reconnection support Maxime Coquelin
2024-09-20 15:09 ` [PATCH v2 1/2] vhost: add logging mechanism for reconnection Maxime Coquelin
2024-09-23  7:12   ` Chenbo Xia
2024-09-23 15:42   ` David Marchand
2024-09-23 19:46     ` Maxime Coquelin
2024-09-20 15:09 ` [PATCH v2 2/2] vhost: add reconnection support to VDUSE Maxime Coquelin
2024-09-23  7:13   ` Chenbo Xia
2024-09-23 16:43   ` David Marchand

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).