From: Nikos Dragazis <ndragazis@arrikto.com>
To: dev@dpdk.org
Cc: Maxime Coquelin <maxime.coquelin@redhat.com>,
Tiwei Bie <tiwei.bie@intel.com>,
Zhihong Wang <zhihong.wang@intel.com>,
Stefan Hajnoczi <stefanha@redhat.com>,
Wei Wang <wei.w.wang@intel.com>,
Stojaczyk Dariusz <dariusz.stojaczyk@intel.com>,
Vangelis Koukis <vkoukis@arrikto.com>
Subject: [dpdk-dev] [PATCH 04/28] vhost: move socket fd and un sockaddr
Date: Wed, 19 Jun 2019 18:14:29 +0300 [thread overview]
Message-ID: <1560957293-17294-5-git-send-email-ndragazis@arrikto.com> (raw)
In-Reply-To: <1560957293-17294-1-git-send-email-ndragazis@arrikto.com>
The socket file descriptor and AF_UNIX sockaddr are specific to the
AF_UNIX transport, so move them into trans_af_unix.c.
In order to do this, we need to begin defining the vhost_transport_ops
interface that will allow librte_vhost to support multiple transports.
This patch adds socket_init() and socket_cleanup() to
vhost_transport_ops.
Signed-off-by: Nikos Dragazis <ndragazis@arrikto.com>
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
---
lib/librte_vhost/socket.c | 11 ++------
lib/librte_vhost/trans_af_unix.c | 55 ++++++++++++++++++++++++++++++++--------
lib/librte_vhost/vhost.h | 30 ++++++++++++++++++----
3 files changed, 72 insertions(+), 24 deletions(-)
diff --git a/lib/librte_vhost/socket.c b/lib/librte_vhost/socket.c
index 60d3546..3b5608c 100644
--- a/lib/librte_vhost/socket.c
+++ b/lib/librte_vhost/socket.c
@@ -408,7 +408,7 @@ rte_vhost_driver_register(const char *path, uint64_t flags)
} else {
vsocket->is_server = true;
}
- ret = create_unix_socket(vsocket);
+ ret = trans_ops->socket_init(vsocket, flags);
if (ret < 0) {
goto out_mutex;
}
@@ -480,14 +480,7 @@ rte_vhost_driver_unregister(const char *path)
}
pthread_mutex_unlock(&vsocket->conn_mutex);
- if (vsocket->is_server) {
- fdset_del(&vhost_user.fdset,
- vsocket->socket_fd);
- close(vsocket->socket_fd);
- unlink(path);
- } else if (vsocket->reconnect) {
- vhost_user_remove_reconnect(vsocket);
- }
+ vsocket->trans_ops->socket_cleanup(vsocket);
pthread_mutex_destroy(&vsocket->conn_mutex);
vhost_user_socket_mem_free(vsocket);
diff --git a/lib/librte_vhost/trans_af_unix.c b/lib/librte_vhost/trans_af_unix.c
index 4de2579..f23bb9c 100644
--- a/lib/librte_vhost/trans_af_unix.c
+++ b/lib/librte_vhost/trans_af_unix.c
@@ -4,6 +4,8 @@
* Copyright(c) 2019 Arrikto Inc.
*/
+#include <sys/socket.h>
+#include <sys/un.h>
#include <fcntl.h>
#include <rte_log.h>
@@ -15,8 +17,11 @@
struct af_unix_socket {
struct vhost_user_socket socket; /* must be the first field! */
+ int socket_fd;
+ struct sockaddr_un un;
};
+static int create_unix_socket(struct vhost_user_socket *vsocket);
static void vhost_user_read_cb(int connfd, void *dat, int *remove);
/*
@@ -244,11 +249,13 @@ vhost_user_read_cb(int connfd, void *dat, int *remove)
}
}
-int
+static int
create_unix_socket(struct vhost_user_socket *vsocket)
{
+ struct af_unix_socket *af_vsocket =
+ container_of(vsocket, struct af_unix_socket, socket);
int fd;
- struct sockaddr_un *un = &vsocket->un;
+ struct sockaddr_un *un = &af_vsocket->un;
fd = socket(AF_UNIX, SOCK_STREAM, 0);
if (fd < 0)
@@ -269,15 +276,17 @@ create_unix_socket(struct vhost_user_socket *vsocket)
strncpy(un->sun_path, vsocket->path, sizeof(un->sun_path));
un->sun_path[sizeof(un->sun_path) - 1] = '\0';
- vsocket->socket_fd = fd;
+ af_vsocket->socket_fd = fd;
return 0;
}
int
vhost_user_start_server(struct vhost_user_socket *vsocket)
{
+ struct af_unix_socket *af_vsocket =
+ container_of(vsocket, struct af_unix_socket, socket);
int ret;
- int fd = vsocket->socket_fd;
+ int fd = af_vsocket->socket_fd;
const char *path = vsocket->path;
/*
@@ -290,7 +299,7 @@ vhost_user_start_server(struct vhost_user_socket *vsocket)
* The user must ensure that the socket does not exist before
* registering the vhost driver in server mode.
*/
- ret = bind(fd, (struct sockaddr *)&vsocket->un, sizeof(vsocket->un));
+ ret = bind(fd, (struct sockaddr *)&af_vsocket->un, sizeof(af_vsocket->un));
if (ret < 0) {
RTE_LOG(ERR, VHOST_CONFIG,
"failed to bind to %s: %s; remove it and try again\n",
@@ -432,13 +441,15 @@ vhost_user_reconnect_init(void)
int
vhost_user_start_client(struct vhost_user_socket *vsocket)
{
+ struct af_unix_socket *af_vsocket =
+ container_of(vsocket, struct af_unix_socket, socket);
int ret;
- int fd = vsocket->socket_fd;
+ int fd = af_vsocket->socket_fd;
const char *path = vsocket->path;
struct vhost_user_reconnect *reconn;
- ret = vhost_user_connect_nonblock(fd, (struct sockaddr *)&vsocket->un,
- sizeof(vsocket->un));
+ ret = vhost_user_connect_nonblock(fd, (struct sockaddr *)&af_vsocket->un,
+ sizeof(af_vsocket->un));
if (ret == 0) {
vhost_user_add_connection(fd, vsocket);
return 0;
@@ -461,7 +472,7 @@ vhost_user_start_client(struct vhost_user_socket *vsocket)
close(fd);
return -1;
}
- reconn->un = vsocket->un;
+ reconn->un = af_vsocket->un;
reconn->fd = fd;
reconn->vsocket = vsocket;
pthread_mutex_lock(&reconn_list.mutex);
@@ -471,7 +482,7 @@ vhost_user_start_client(struct vhost_user_socket *vsocket)
return 0;
}
-bool
+static bool
vhost_user_remove_reconnect(struct vhost_user_socket *vsocket)
{
int found = false;
@@ -496,6 +507,28 @@ vhost_user_remove_reconnect(struct vhost_user_socket *vsocket)
}
static int
+af_unix_socket_init(struct vhost_user_socket *vsocket,
+ uint64_t flags __rte_unused)
+{
+ return create_unix_socket(vsocket);
+}
+
+static void
+af_unix_socket_cleanup(struct vhost_user_socket *vsocket)
+{
+ struct af_unix_socket *af_vsocket =
+ container_of(vsocket, struct af_unix_socket, socket);
+
+ if (vsocket->is_server) {
+ fdset_del(&vhost_user.fdset, af_vsocket->socket_fd);
+ close(af_vsocket->socket_fd);
+ unlink(vsocket->path);
+ } else if (vsocket->reconnect) {
+ vhost_user_remove_reconnect(vsocket);
+ }
+}
+
+static int
af_unix_vring_call(struct virtio_net *dev __rte_unused,
struct vhost_virtqueue *vq)
{
@@ -506,5 +539,7 @@ af_unix_vring_call(struct virtio_net *dev __rte_unused,
const struct vhost_transport_ops af_unix_trans_ops = {
.socket_size = sizeof(struct af_unix_socket),
+ .socket_init = af_unix_socket_init,
+ .socket_cleanup = af_unix_socket_cleanup,
.vring_call = af_unix_vring_call,
};
diff --git a/lib/librte_vhost/vhost.h b/lib/librte_vhost/vhost.h
index 9615392..40b5c25 100644
--- a/lib/librte_vhost/vhost.h
+++ b/lib/librte_vhost/vhost.h
@@ -14,7 +14,6 @@
#include <linux/vhost.h>
#include <linux/virtio_net.h>
#include <sys/socket.h>
-#include <sys/un.h> /* TODO remove when trans_af_unix.c refactoring is done */
#include <linux/if.h>
#include <pthread.h>
@@ -291,6 +290,7 @@ struct guest_page {
};
struct virtio_net;
+struct vhost_user_socket;
/**
* A structure containing function pointers for transport-specific operations.
@@ -300,6 +300,30 @@ struct vhost_transport_ops {
size_t socket_size;
/**
+ * Initialize a vhost-user socket that is being created by
+ * rte_vhost_driver_register(). This function checks that the flags
+ * are valid but does not establish a vhost-user connection.
+ *
+ * @param vsocket
+ * new socket
+ * @param flags
+ * flags argument from rte_vhost_driver_register()
+ * @return
+ * 0 on success, -1 on failure
+ */
+ int (*socket_init)(struct vhost_user_socket *vsocket, uint64_t flags);
+
+ /**
+ * Free resources associated with a socket, including any established
+ * connections. This function calls vhost_destroy_device() to destroy
+ * established connections for this socket.
+ *
+ * @param vsocket
+ * vhost socket
+ */
+ void (*socket_cleanup)(struct vhost_user_socket *vsocket);
+
+ /**
* Notify the guest that used descriptors have been added to the vring.
* The VRING_AVAIL_F_NO_INTERRUPT flag and event idx have already been checked
* so this function just needs to perform the notification.
@@ -387,8 +411,6 @@ struct vhost_user_socket {
struct vhost_user_connection_list conn_list;
pthread_mutex_t conn_mutex;
char *path;
- int socket_fd;
- struct sockaddr_un un;
bool is_server;
bool reconnect;
bool dequeue_zero_copy;
@@ -436,14 +458,12 @@ struct vhost_user {
extern struct vhost_user vhost_user;
-int create_unix_socket(struct vhost_user_socket *vsocket);
int vhost_user_start_server(struct vhost_user_socket *vsocket);
int vhost_user_start_client(struct vhost_user_socket *vsocket);
extern pthread_t reconn_tid;
int vhost_user_reconnect_init(void);
-bool vhost_user_remove_reconnect(struct vhost_user_socket *vsocket);
static __rte_always_inline bool
vq_is_packed(struct virtio_net *dev)
--
2.7.4
next prev parent reply other threads:[~2019-06-19 15:16 UTC|newest]
Thread overview: 40+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-06-19 15:14 [dpdk-dev] [PATCH 00/28] vhost: add virtio-vhost-user transport Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 01/28] vhost: introduce vhost transport operations structure Nikos Dragazis
2019-06-19 20:14 ` Aaron Conole
2019-06-20 10:30 ` Bruce Richardson
2019-06-20 18:24 ` Nikos Dragazis
2019-06-20 18:19 ` Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 02/28] vhost: move socket management code Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 03/28] vhost: allocate per-socket transport state Nikos Dragazis
2019-06-19 15:14 ` Nikos Dragazis [this message]
2019-06-19 15:14 ` [dpdk-dev] [PATCH 05/28] vhost: move start server/client calls Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 06/28] vhost: move vhost-user connection Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 07/28] vhost: move vhost-user reconnection Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 08/28] vhost: move vhost-user fdset Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 09/28] vhost: propagate vhost transport operations Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 10/28] vhost: use a single structure for the device state Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 11/28] vhost: extract socket I/O into transport Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 12/28] vhost: move slave request fd and lock Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 13/28] vhost: move mmap/munmap Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 14/28] vhost: move setup of the log memory region Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 15/28] vhost: remove main fd parameter from msg handlers Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 16/28] vhost: move postcopy live migration code Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 17/28] vhost: support registering additional vhost-user transports Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 18/28] drivers/virtio_vhost_user: add virtio PCI framework Nikos Dragazis
2019-09-05 16:34 ` Maxime Coquelin
2019-09-09 8:42 ` Nikos Dragazis
2019-09-09 8:44 ` Maxime Coquelin
2019-06-19 15:14 ` [dpdk-dev] [PATCH 19/28] vhost: add index field in vhost virtqueues Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 20/28] drivers: add virtio-vhost-user transport Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 21/28] drivers/virtio_vhost_user: use additional device resources Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 22/28] vhost: add flag for choosing vhost-user transport Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 23/28] net/vhost: add virtio-vhost-user support Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 24/28] examples/vhost_scsi: add --socket-file argument Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 25/28] examples/vhost_scsi: add virtio-vhost-user support Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 26/28] mk: link apps with virtio-vhost-user driver Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 27/28] config: add option for the virtio-vhost-user transport Nikos Dragazis
2019-06-19 15:14 ` [dpdk-dev] [PATCH 28/28] usertools: add virtio-vhost-user devices to dpdk-devbind.py Nikos Dragazis
[not found] ` <CGME20190620113240eucas1p22ca4faa64a36bbb7aec38a81298ade56@eucas1p2.samsung.com>
2019-06-20 11:32 ` [dpdk-dev] [PATCH 00/28] vhost: add virtio-vhost-user transport Ilya Maximets
2019-06-20 23:44 ` Nikos Dragazis
2019-06-20 11:35 ` Maxime Coquelin
2019-06-22 20:26 ` Nikos Dragazis
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1560957293-17294-5-git-send-email-ndragazis@arrikto.com \
--to=ndragazis@arrikto.com \
--cc=dariusz.stojaczyk@intel.com \
--cc=dev@dpdk.org \
--cc=maxime.coquelin@redhat.com \
--cc=stefanha@redhat.com \
--cc=tiwei.bie@intel.com \
--cc=vkoukis@arrikto.com \
--cc=wei.w.wang@intel.com \
--cc=zhihong.wang@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).