From: Tetsuya Mukawa <mukawa@igel.co.jp>
To: dev@dpdk.org
Cc: nakajima.yoshihiro@lab.ntt.co.jp, masutani.hitoshi@lab.ntt.co.jp
Subject: [dpdk-dev] [RFC PATCH 7/7] lib/librte_vhost: Add vhost-user implementation
Date: Thu, 6 Nov 2014 20:14:31 +0900 [thread overview]
Message-ID: <1415272471-3299-8-git-send-email-mukawa@igel.co.jp> (raw)
In-Reply-To: <1415272471-3299-1-git-send-email-mukawa@igel.co.jp>
This patch adds vhost-user implementation to librte_vhost.
To communicate with vhost-user of QEMU, speficy VHOST_DRV_USER as
a vhost_driver_type_t variable in rte_vhost_driver_register().
Signed-off-by: Tetsuya Mukawa <mukawa@igel.co.jp>
---
lib/librte_vhost/rte_virtio_net.h | 19 +-
lib/librte_vhost/vhost-net-user.c | 541 +++++++++++++++++++++++++++++++++++++
lib/librte_vhost/vhost-net.c | 39 ++-
lib/librte_vhost/vhost-net.h | 7 +
lib/librte_vhost/virtio-net-user.c | 410 ++++++++++++++++++++++++++++
lib/librte_vhost/virtio-net.c | 64 ++++-
6 files changed, 1073 insertions(+), 7 deletions(-)
create mode 100644 lib/librte_vhost/vhost-net-user.c
create mode 100644 lib/librte_vhost/virtio-net-user.c
diff --git a/lib/librte_vhost/rte_virtio_net.h b/lib/librte_vhost/rte_virtio_net.h
index a9e20ea..af07900 100644
--- a/lib/librte_vhost/rte_virtio_net.h
+++ b/lib/librte_vhost/rte_virtio_net.h
@@ -75,17 +75,32 @@ struct buf_vector {
*/
typedef enum {
VHOST_DRV_CUSE, /* cuse driver */
+ VHOST_DRV_USER, /* vhost-user driver */
VHOST_DRV_NUM /* the number of vhost driver types */
} vhost_driver_type_t;
+
+/**
+ * Structure contains vhost-user session specific information
+ */
+struct vhost_user_session {
+ int fh; /**< session identifier */
+ pthread_t tid; /**< thread id of session handler */
+ int socketfd; /**< fd of socket */
+ int interval; /**< reconnection interval of session */
+};
+
/**
* Structure contains information relating vhost driver.
*/
struct vhost_driver {
vhost_driver_type_t type; /**< driver type. */
const char *dev_name; /**< accessing device name. */
+ void *priv; /**< private data. */
union {
struct fuse_session *cuse_session; /**< fuse session. */
+ struct vhost_user_session *user_session;
+ /**< vhost-user session. */
};
};
@@ -199,9 +214,11 @@ struct vhost_driver *rte_vhost_driver_register(
const char *dev_name, vhost_driver_type_t type);
/* Register callbacks. */
-int rte_vhost_driver_callback_register(struct virtio_net_device_ops const * const);
+int rte_vhost_driver_callback_register(struct vhost_driver *drv,
+ struct virtio_net_device_ops const * const, void *priv);
/* Start vhost driver session blocking loop. */
int rte_vhost_driver_session_start(struct vhost_driver *drv);
+void rte_vhost_driver_session_stop(struct vhost_driver *drv);
/**
* This function adds buffers to the virtio devices RX virtqueue. Buffers can
diff --git a/lib/librte_vhost/vhost-net-user.c b/lib/librte_vhost/vhost-net-user.c
new file mode 100644
index 0000000..434f20f
--- /dev/null
+++ b/lib/librte_vhost/vhost-net-user.c
@@ -0,0 +1,541 @@
+/*-
+ * BSD LICENSE
+ *
+ * Copyright (c) 2014 IGEL Co/.Ltd.
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ *
+ * * Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * * Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in
+ * the documentation and/or other materials provided with the
+ * distribution.
+ * * Neither the name of IGEL nor the names of its
+ * contributors may be used to endorse or promote products derived
+ * from this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+ * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+ * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+ * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+ * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+ * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+ * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+ * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+ */
+
+#include <sys/types.h>
+#include <sys/socket.h>
+#include <linux/un.h>
+
+#define VHOST_USER_MAX_DEVICE (32)
+#define VHOST_USER_MAX_FD_NUM (3)
+
+/* start id of vhost user device */
+rte_atomic16_t vhost_user_device_id;
+
+static struct vhost_net_device_ops const *ops;
+
+typedef enum VhostUserRequest {
+ VHOST_USER_NONE = 0,
+ VHOST_USER_GET_FEATURES = 1,
+ VHOST_USER_SET_FEATURES = 2,
+ VHOST_USER_SET_OWNER = 3,
+ VHOST_USER_RESET_OWNER = 4,
+ VHOST_USER_SET_MEM_TABLE = 5,
+ VHOST_USER_SET_LOG_BASE = 6,
+ VHOST_USER_SET_LOG_FD = 7,
+ VHOST_USER_SET_VRING_NUM = 8,
+ VHOST_USER_SET_VRING_ADDR = 9,
+ VHOST_USER_SET_VRING_BASE = 10,
+ VHOST_USER_GET_VRING_BASE = 11,
+ VHOST_USER_SET_VRING_KICK = 12,
+ VHOST_USER_SET_VRING_CALL = 13,
+ VHOST_USER_SET_VRING_ERR = 14,
+ VHOST_USER_MAX
+} VhostUserRequest;
+
+#define VHOST_MEMORY_MAX_NREGIONS 8
+
+typedef struct VhostUserMemoryRegion {
+ uint64_t guest_phys_addr;
+ uint64_t memory_size;
+ uint64_t userspace_addr;
+ uint64_t mmap_offset;
+} VhostUserMemoryRegion;
+
+typedef struct VhostUserMemory {
+ uint32_t nregions;
+ uint32_t padding;
+ VhostUserMemoryRegion regions[VHOST_MEMORY_MAX_NREGIONS];
+} VhostUserMemory;
+
+typedef struct VhostUserMsg {
+ VhostUserRequest request;
+
+#define VHOST_USER_VERSION_MASK (0x3)
+#define VHOST_USER_REPLY_MASK (0x1<<2)
+ uint32_t flags;
+ uint32_t size; /* the following payload size */
+ union {
+#define VHOST_USER_VRING_IDX_MASK (0xff)
+#define VHOST_USER_VRING_NOFD_MASK (0x1<<8)
+ uint64_t u64;
+ struct vhost_vring_state state;
+ struct vhost_vring_addr addr;
+ VhostUserMemory memory;
+ };
+} __attribute__((packed)) VhostUserMsg;
+
+static VhostUserMsg m __attribute__ ((unused));
+#define VHOST_USER_HDR_SIZE (sizeof(m.request) \
+ + sizeof(m.flags) + sizeof(m.size))
+
+/* The version of the protocol we support */
+#define VHOST_USER_VERSION (0x1)
+
+static unsigned long int ioctl_to_vhost_user_request[VHOST_USER_MAX] = {
+ -1, /* VHOST_USER_NONE */
+ VHOST_GET_FEATURES, /* VHOST_USER_GET_FEATURES */
+ VHOST_SET_FEATURES, /* VHOST_USER_SET_FEATURES */
+ VHOST_SET_OWNER, /* VHOST_USER_SET_OWNER */
+ VHOST_RESET_OWNER, /* VHOST_USER_RESET_OWNER */
+ VHOST_SET_MEM_TABLE, /* VHOST_USER_SET_MEM_TABLE */
+ VHOST_SET_LOG_BASE, /* VHOST_USER_SET_LOG_BASE */
+ VHOST_SET_LOG_FD, /* VHOST_USER_SET_LOG_FD */
+ VHOST_SET_VRING_NUM, /* VHOST_USER_SET_VRING_NUM */
+ VHOST_SET_VRING_ADDR, /* VHOST_USER_SET_VRING_ADDR */
+ VHOST_SET_VRING_BASE, /* VHOST_USER_SET_VRING_BASE */
+ VHOST_GET_VRING_BASE, /* VHOST_USER_GET_VRING_BASE */
+ VHOST_SET_VRING_KICK, /* VHOST_USER_SET_VRING_KICK */
+ VHOST_SET_VRING_CALL, /* VHOST_USER_SET_VRING_CALL */
+ VHOST_SET_VRING_ERR /* VHOST_USER_SET_VRING_ERR */
+};
+
+/**
+ * Returns vhost_device_ctx from given fuse_req_t. The index is populated later when
+ * the device is added to the device linked list.
+ */
+static struct vhost_device_ctx
+vhost_driver_to_vhost_ctx(struct vhost_driver *drv)
+{
+ struct vhost_device_ctx ctx;
+ int device_id = drv->user_session->fh;
+
+ ctx.type = VHOST_DRV_USER;
+ ctx.fh = device_id;
+ ctx.user.drv = drv;
+
+ return ctx;
+}
+
+/**
+ * When the device is created in QEMU it gets initialised here and added to the device linked list.
+ */
+static int
+vhost_user_open(struct vhost_driver *drv)
+{
+ struct vhost_device_ctx ctx = vhost_driver_to_vhost_ctx(drv);
+
+ int ret;
+
+ ret = ops->new_device(ctx);
+ if (ret == -1)
+ return -1;
+
+ RTE_LOG(INFO, VHOST_CONFIG, "(%"PRIu64") Device configuration started\n", ctx.fh);
+
+ return 0;
+}
+
+/**
+ * When QEMU is shutdown or killed the device gets released.
+ */
+static void
+vhost_user_release(struct vhost_driver *drv)
+{
+ struct vhost_device_ctx ctx = vhost_driver_to_vhost_ctx(drv);
+
+ ops->destroy_device(ctx);
+ RTE_LOG(INFO, VHOST_CONFIG, "(%"PRIu64") Device released\n", ctx.fh);
+}
+
+/**
+ * Send data to vhost-user device on a QEMU.
+ */
+static int
+vhost_user_write(struct vhost_driver *drv, VhostUserMsg *msg,
+ int *fds, size_t fd_num)
+{
+ int fd, len;
+ size_t fd_size = fd_num * sizeof(int);
+ char control[CMSG_SPACE(fd_size)];
+ struct msghdr msg_header;
+ struct iovec iov[1];
+ struct cmsghdr *cmsg_header;
+ struct vhost_device_ctx ctx = vhost_driver_to_vhost_ctx(drv);
+
+ if ((drv == NULL) || (msg == NULL))
+ return -EINVAL;
+
+ fd = drv->user_session->socketfd;
+
+ memset(&msg_header, 0, sizeof(msg_header));
+ memset(control, 0, sizeof(control));
+
+ /* set the payload */
+ iov[0].iov_base = (void *)msg;
+ iov[0].iov_len = VHOST_USER_HDR_SIZE + msg->size;
+
+ msg_header.msg_iov = iov;
+ msg_header.msg_iovlen = 1;
+
+ if (fd_num) {
+ msg_header.msg_control = control;
+ msg_header.msg_controllen = sizeof(control);
+ cmsg_header = CMSG_FIRSTHDR(&msg_header);
+ cmsg_header->cmsg_len = CMSG_LEN(fd_size);
+ cmsg_header->cmsg_level = SOL_SOCKET;
+ cmsg_header->cmsg_type = SCM_RIGHTS;
+ memcpy(CMSG_DATA(cmsg_header), fds, fd_size);
+ } else {
+ msg_header.msg_control = 0;
+ msg_header.msg_controllen = 0;
+ }
+
+ do {
+ len = sendmsg(fd, &msg_header, 0);
+ } while (len < 0 && errno == EINTR);
+
+ if (len < 0)
+ goto error;
+
+ return 0;
+
+error:
+ RTE_LOG(INFO, VHOST_CONFIG, "(%"PRIu64") Device cannot send message\n", ctx.fh);
+ return -EFAULT;
+}
+
+/**
+ * Receive data from vhost-user device on a QEMU.
+ */
+static int
+vhost_user_read(struct vhost_driver *drv, VhostUserMsg *msg,
+ int *fds, size_t *fd_num)
+{
+ int fd, len;
+ size_t fd_size = (*fd_num) * sizeof(int);
+ char control[CMSG_SPACE(fd_size)];
+ struct msghdr msg_header;
+ struct iovec iov[1];
+ struct cmsghdr *cmsg_header;
+ struct vhost_device_ctx ctx = vhost_driver_to_vhost_ctx(drv);
+
+ if ((drv == NULL) || (msg == NULL))
+ return -EINVAL;
+
+ fd = drv->user_session->socketfd;
+
+ memset(&msg_header, 0, sizeof(msg_header));
+ memset(control, 0, sizeof(control));
+ *fd_num = 0;
+
+ /* set the payload */
+ iov[0].iov_base = (void *)msg;
+ iov[0].iov_len = VHOST_USER_HDR_SIZE;
+
+ msg_header.msg_iov = iov;
+ msg_header.msg_iovlen = 1;
+ msg_header.msg_control = control;
+ msg_header.msg_controllen = sizeof(control);
+
+ if ((len = recvmsg(fd, &msg_header, 0)) <= 0)
+ goto error;
+
+ if (msg_header.msg_flags & (MSG_TRUNC | MSG_CTRUNC))
+ goto error;
+
+ cmsg_header = CMSG_FIRSTHDR(&msg_header);
+ if (cmsg_header && cmsg_header->cmsg_len > 0 &&
+ cmsg_header->cmsg_level == SOL_SOCKET &&
+ cmsg_header->cmsg_type == SCM_RIGHTS) {
+ if (fd_size >= cmsg_header->cmsg_len - CMSG_LEN(0)) {
+ fd_size = cmsg_header->cmsg_len - CMSG_LEN(0);
+ memcpy(fds, CMSG_DATA(cmsg_header), fd_size);
+ *fd_num = fd_size / sizeof(int);
+ }
+ }
+
+ if (read(fd, ((char *)msg) + len, msg->size) < 0)
+ goto error;
+
+ return 0;
+
+error:
+ RTE_LOG(INFO, VHOST_CONFIG, "(%"PRIu64") Device cannot receive message\n", ctx.fh);
+ return -EFAULT;
+}
+
+/*
+ * Boilerplate code for vhost-user IOCTL
+ * Implicit arguments: ctx, req, result.
+ */
+#define VHOST_USER_IOCTL(func) do { \
+ result = (func)(ctx); \
+} while (0)
+
+/*
+ * Boilerplate code for vhost-user Read IOCTL
+ * Implicit arguments: ctx, req, result, in_bufsz, in_buf.
+ */
+#define VHOST_USER_IOCTL_R(type, var, func) do {\
+ result = func(ctx, &(var)); \
+} while (0)
+
+/*
+ * Boilerplate code for vhost-user Write IOCTL
+ * Implicit arguments: ctx, req, result, out_bufsz.
+ */
+#define VHOST_USER_IOCTL_W(type, var, func) do {\
+ result = (func)(ctx, &(var)); \
+ msg->flags |= VHOST_USER_REPLY_MASK; \
+ msg->size = sizeof(type); \
+ vhost_user_write(drv, msg, NULL, 0); \
+} while (0)
+
+/*
+ * Boilerplate code for vhost-user Read/Write IOCTL
+ * Implicit arguments: ctx, req, result, in_bufsz, in_buf.
+ */
+#define VHOST_USER_IOCTL_RW(type1, var1, type2, var2, func) do {\
+ result = (func)(ctx, (var1), &(var2)); \
+ msg->flags |= VHOST_USER_REPLY_MASK; \
+ msg->size = sizeof(type2); \
+ vhost_user_write(drv, msg, NULL, 0); \
+} while (0)
+
+/**
+ * The IOCTLs are handled using unix domain socket in userspace.
+ */
+ static int
+vhost_user_ioctl(struct vhost_driver *drv, VhostUserMsg *msg,
+ int *fds, int fd_num)
+{
+ struct vhost_device_ctx ctx = vhost_driver_to_vhost_ctx(drv);
+ struct vhost_vring_file file;
+ int result = 0;
+
+ switch (ioctl_to_vhost_user_request[msg->request]) {
+ case VHOST_GET_FEATURES:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_GET_FEATURES\n", ctx.fh);
+ VHOST_USER_IOCTL_W(uint64_t, msg->u64, ops->get_features);
+ break;
+
+ case VHOST_SET_FEATURES:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_SET_FEATURES\n", ctx.fh);
+ VHOST_USER_IOCTL_R(uint64_t, msg->u64, ops->set_features);
+ break;
+
+ case VHOST_RESET_OWNER:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_RESET_OWNER\n", ctx.fh);
+ VHOST_USER_IOCTL(ops->reset_owner);
+ break;
+
+ case VHOST_SET_OWNER:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_SET_OWNER\n", ctx.fh);
+ VHOST_USER_IOCTL(ops->set_owner);
+ break;
+
+ case VHOST_SET_MEM_TABLE:
+ /*TODO fix race condition.*/
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_SET_MEM_TABLE\n", ctx.fh);
+ /* all fds should be same, because physical memory consist of an one file */
+ ctx.user.fds = fds;
+ ctx.user.fd_num = fd_num;
+ result = ops->set_mem_table(ctx, &msg->memory, msg->memory.nregions);
+ break;
+
+ case VHOST_SET_VRING_NUM:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_SET_VRING_NUM\n", ctx.fh);
+ VHOST_USER_IOCTL_R(struct vhost_vring_state, msg->state, ops->set_vring_num);
+ break;
+
+ case VHOST_SET_VRING_BASE:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_SET_VRING_BASE\n", ctx.fh);
+ VHOST_USER_IOCTL_R(struct vhost_vring_state, msg->state, ops->set_vring_base);
+ break;
+
+ case VHOST_GET_VRING_BASE:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_GET_VRING_BASE\n", ctx.fh);
+ VHOST_USER_IOCTL_RW(uint32_t, msg->addr.index, struct vhost_vring_state, msg->state, ops->get_vring_base);
+ break;
+
+ case VHOST_SET_VRING_ADDR:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_SET_VRING_ADDR\n", ctx.fh);
+ VHOST_USER_IOCTL_R(struct vhost_vring_addr, msg->addr, ops->set_vring_addr);
+ break;
+
+ case VHOST_SET_VRING_KICK:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_SET_VRING_KICK\n", ctx.fh);
+ ctx.user.fds = fds;
+ ctx.user.fd_num = fd_num;
+ file.index = msg->u64;
+ VHOST_USER_IOCTL_R(struct vhost_vring_file, file, ops->set_vring_kick);
+ break;
+
+ case VHOST_SET_VRING_CALL:
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: VHOST_SET_VRING_CALL\n", ctx.fh);
+ ctx.user.fds = fds;
+ ctx.user.fd_num = fd_num;
+ file.index = msg->u64;
+ VHOST_USER_IOCTL_R(struct vhost_vring_file, file, ops->set_vring_call);
+ break;
+
+ case VHOST_NET_SET_BACKEND:
+ default:
+ RTE_LOG(ERR, VHOST_CONFIG, "(%"PRIu64") IOCTL: DOESN NOT EXIST\n", ctx.fh);
+ result = -1;
+ }
+
+ if (result < 0)
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: FAIL\n", ctx.fh);
+ else
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") IOCTL: SUCCESS\n", ctx.fh);
+
+ return result;
+}
+
+/**
+ * vhost-user specific registration.
+ */
+static int
+vhost_user_driver_register(struct vhost_driver *drv)
+{
+ if ((drv == NULL) || (drv->dev_name == NULL) ||
+ (strlen(drv->dev_name) > UNIX_PATH_MAX - 1))
+ return -1;
+
+ ops = get_virtio_net_callbacks(drv->type);
+
+ drv->user_session = rte_malloc(NULL, sizeof(struct vhost_user_session), CACHE_LINE_SIZE);
+ if (drv->user_session == NULL)
+ return -1;
+
+ drv->user_session->fh =
+ rte_atomic16_add_return(&vhost_user_device_id, 1) - 1; /* fh of first device is zero */
+ drv->user_session->interval = 1;
+
+ return 0;
+}
+
+/**
+ * When vhost-user driver starts, the session handler communicates with vhost-user
+ * device on a QEMU using a unix domain sokcet.
+ */
+static void *
+vhost_user_session_handler(void *data)
+{
+ struct vhost_driver *drv = data;
+ int ret;
+ struct sockaddr_un caddr;
+ VhostUserMsg msg;
+ int fds[VHOST_USER_MAX_FD_NUM];
+ int socketfd;
+ int interval;
+ size_t fd_num;
+
+ if ((drv == NULL) || (drv->dev_name == NULL))
+ return NULL;
+
+ bzero(&caddr, sizeof(caddr));
+ caddr.sun_family = AF_LOCAL;
+ strncpy((char *)&caddr.sun_path, drv->dev_name, strlen(drv->dev_name));
+
+reconnect:
+ drv->user_session->socketfd = socket(AF_UNIX, SOCK_STREAM, 0);
+ if (drv->user_session->socketfd < 0)
+ return NULL;
+
+ socketfd = drv->user_session->socketfd;
+ interval = drv->user_session->interval;
+ while (1) {
+ ret = connect(socketfd, (struct sockaddr *)&caddr, sizeof(caddr));
+ if (ret == 0)
+ break; /* success */
+ sleep(interval);
+ }
+
+ ret = vhost_user_open(drv);
+ if (ret != 0) {
+ RTE_LOG(ERR, VHOST_CONFIG, "(Socket %s) open failuer\n", drv->dev_name);
+ return NULL;
+ }
+
+ for (;;) {
+ fd_num = VHOST_USER_MAX_FD_NUM;
+ bzero(&msg, sizeof(VhostUserMsg));
+ ret = vhost_user_read(drv, &msg, fds, &fd_num);
+ if (ret != 0) {
+ RTE_LOG(ERR, VHOST_CONFIG, "(Socket %s) read failuer\n", drv->dev_name);
+ vhost_user_release(drv);
+ goto reconnect;
+ }
+
+ ret = vhost_user_ioctl(drv, &msg, fds, fd_num);
+ if (ret != 0) {
+ RTE_LOG(ERR, VHOST_CONFIG, "(Socket %s) request failuer\n", drv->dev_name);
+ vhost_user_release(drv);
+ goto reconnect;
+ }
+ }
+
+ return NULL;
+}
+
+/**
+ * Create session handler
+ */
+static int
+vhost_user_driver_start(struct vhost_driver *drv)
+{
+ if (pthread_create(&drv->user_session->tid, NULL, vhost_user_session_handler, drv)) {
+ RTE_LOG(ERR, VHOST_CONFIG,
+ "(Socket %s) starting event handler failuer\n", drv->dev_name);
+ return -1;
+ }
+
+ /* TODO: The event handler thread may need to run on a core user speficied. */
+
+ return 0;
+}
+
+/**
+ * Destroy session handler
+ */
+static void
+vhost_user_driver_stop(struct vhost_driver *drv)
+{
+ pthread_t *tid = &drv->user_session->tid;
+
+ if (pthread_create(tid, NULL, vhost_user_session_handler, drv)) {
+ RTE_LOG(ERR, VHOST_CONFIG,
+ "(Socket %s) starting event handler failuer\n", drv->dev_name);
+ return;
+ }
+
+ /* stop event thread and wait until connection is closed */
+ if (*tid) {
+ pthread_cancel(*tid);
+ pthread_join(*tid, NULL);
+ }
+
+ vhost_user_release(drv);
+}
diff --git a/lib/librte_vhost/vhost-net.c b/lib/librte_vhost/vhost-net.c
index b0de5fd..10f41e9 100644
--- a/lib/librte_vhost/vhost-net.c
+++ b/lib/librte_vhost/vhost-net.c
@@ -42,6 +42,11 @@
*/
#include "vhost-net-cdev.c"
+/*
+ * Include vhost-user depend functions and definitions
+ */
+#include "vhost-net-user.c"
+
/**
* This function abstracts cuse and vhost-user driver registration.
*/
@@ -65,10 +70,17 @@ rte_vhost_driver_register(const char *dev_name, vhost_driver_type_t type)
if (ret != 0)
goto err;
break;
+ case VHOST_DRV_USER:
+ ret = vhost_user_driver_register(drv);
+ break;
default:
+ ret = -EINVAL;
break;
}
+ if (ret != 0)
+ goto err;
+
return drv;
err:
free(drv);
@@ -81,17 +93,40 @@ err:
int
rte_vhost_driver_session_start(struct vhost_driver *drv)
{
+ int ret;
+
if (drv == NULL)
return -ENODEV;
switch (drv->type) {
case VHOST_DRV_CUSE:
- vhost_cuse_driver_session_start(drv);
+ ret = vhost_cuse_driver_session_start(drv);
+ break;
+ case VHOST_DRV_USER:
+ ret = vhost_user_driver_start(drv);
break;
default:
+ ret = -EINVAL;
break;
}
- return 0;
+ return ret;
}
+/**
+ * The vhost session is closed, only allow for vhost-user.
+ */
+void
+rte_vhost_driver_session_stop(struct vhost_driver *drv)
+{
+ if (drv == NULL)
+ return;
+
+ switch (drv->type) {
+ case VHOST_DRV_USER:
+ vhost_user_driver_stop(drv);
+ break;
+ default:
+ break;
+ }
+}
diff --git a/lib/librte_vhost/vhost-net.h b/lib/librte_vhost/vhost-net.h
index ef04832..0e36ba0 100644
--- a/lib/librte_vhost/vhost-net.h
+++ b/lib/librte_vhost/vhost-net.h
@@ -76,6 +76,12 @@ struct vhost_device_cuse_ctx {
pid_t pid; /* PID of process calling the IOCTL. */
};
+struct vhost_device_user_ctx {
+ int *fds;
+ int fd_num;
+ struct vhost_driver *drv;
+};
+
/*
* Structure used to identify device context.
*/
@@ -83,6 +89,7 @@ struct vhost_device_ctx {
vhost_driver_type_t type; /* driver type. */
uint64_t fh; /* Populated with fi->fh to track the device index. */
union {
+ struct vhost_device_user_ctx user;
struct vhost_device_cuse_ctx cdev;
};
};
diff --git a/lib/librte_vhost/virtio-net-user.c b/lib/librte_vhost/virtio-net-user.c
new file mode 100644
index 0000000..1e78f98
--- /dev/null
+++ b/lib/librte_vhost/virtio-net-user.c
@@ -0,0 +1,410 @@
+/*-
+ * BSD LICENSE
+ *
+ * Copyright (c) 2014 IGEL Co.,Ltd.
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ *
+ * * Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * * Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in
+ * the documentation and/or other materials provided with the
+ * distribution.
+ * * Neither the name of IGEL nor the names of its
+ * contributors may be used to endorse or promote products derived
+ * from this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+ * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+ * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+ * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+ * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+ * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+ * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+ * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+ */
+
+#include <sys/types.h>
+#include <sys/stat.h>
+#include <fcntl.h>
+#include <unistd.h>
+#include <sys/mman.h>
+#include <stdlib.h>
+
+/* Functions defined in virtio_net.c */
+static void init_device(struct vhost_device_ctx ctx, struct virtio_net *dev);
+static void cleanup_device(struct virtio_net *dev);
+static void free_device(struct virtio_net_config_ll *ll_dev);
+static int new_device(struct vhost_device_ctx ctx);
+static void destroy_device(struct vhost_device_ctx ctx);
+static int set_owner(struct vhost_device_ctx ctx);
+static int reset_owner(struct vhost_device_ctx ctx);
+static int get_features(struct vhost_device_ctx ctx, uint64_t *pu);
+static int set_features(struct vhost_device_ctx ctx, uint64_t *pu);
+static int set_vring_num(struct vhost_device_ctx ctx, struct vhost_vring_state *state);
+static int set_vring_addr(struct vhost_device_ctx ctx, struct vhost_vring_addr *addr);
+static int set_vring_base(struct vhost_device_ctx ctx, struct vhost_vring_state *state);
+static int set_backend(struct vhost_device_ctx ctx, struct vhost_vring_file *file);
+
+/* Root address of the linked list in the configuration core. */
+static struct virtio_net_config_ll *user_ll_root;
+
+/**
+ * Retrieves an entry from the devices configuration linked list.
+ */
+static struct virtio_net_config_ll *
+user_get_config_ll_entry(struct vhost_device_ctx ctx)
+{
+ struct virtio_net_config_ll *ll_dev = user_ll_root;
+
+ /* Loop through linked list until the device_fh is found. */
+ while (ll_dev != NULL) {
+ if (ll_dev->dev.device_fh == ctx.fh)
+ return ll_dev;
+ ll_dev = ll_dev->next;
+ }
+
+ return NULL;
+}
+
+/**
+ * Searches the configuration core linked list and retrieves the device if it exists.
+ */
+static struct virtio_net *
+user_get_device(struct vhost_device_ctx ctx)
+{
+ struct virtio_net_config_ll *ll_dev;
+
+ ll_dev = user_get_config_ll_entry(ctx);
+
+ /* If a matching entry is found in the linked list, return the device in that entry. */
+ if (ll_dev)
+ return &ll_dev->dev;
+
+ RTE_LOG(ERR, VHOST_CONFIG, "(%"PRIu64") Device not found in linked list.\n", ctx.fh);
+ return NULL;
+}
+
+/**
+ * Add entry containing a device to the device configuration linked list.
+ */
+static void
+user_add_config_ll_entry(struct virtio_net_config_ll *new_ll_dev)
+{
+ struct virtio_net_config_ll *ll_dev = user_ll_root;
+
+ /* If ll_dev == NULL then this is the first device so go to else */
+ if (ll_dev) {
+ /* If the 1st device_fh != 0 then we insert our device here. */
+ if (ll_dev->dev.device_fh != 0) {
+ new_ll_dev->dev.device_fh = 0;
+ new_ll_dev->next = ll_dev;
+ user_ll_root = new_ll_dev;
+ } else {
+ /* Increment through the ll until we find un unused device_fh. Insert the device at that entry*/
+ while ((ll_dev->next != NULL) && (ll_dev->dev.device_fh == (ll_dev->next->dev.device_fh - 1)))
+ ll_dev = ll_dev->next;
+
+ new_ll_dev->dev.device_fh = ll_dev->dev.device_fh + 1;
+ new_ll_dev->next = ll_dev->next;
+ ll_dev->next = new_ll_dev;
+ }
+ } else {
+ user_ll_root = new_ll_dev;
+ user_ll_root->dev.device_fh = 0;
+ }
+
+}
+
+/**
+ * Remove an entry from the device configuration linked list.
+ */
+static struct virtio_net_config_ll *
+user_rm_config_ll_entry(struct virtio_net_config_ll *ll_dev, struct virtio_net_config_ll *ll_dev_last)
+{
+ /* First remove the device and then clean it up. */
+ if (ll_dev == user_ll_root) {
+ user_ll_root = ll_dev->next;
+ cleanup_device(&ll_dev->dev);
+ free_device(ll_dev);
+ return user_ll_root;
+ } else {
+ if (likely(ll_dev_last != NULL)) {
+ ll_dev_last->next = ll_dev->next;
+ cleanup_device(&ll_dev->dev);
+ free_device(ll_dev);
+ return ll_dev_last->next;
+ } else {
+ cleanup_device(&ll_dev->dev);
+ free_device(ll_dev);
+ RTE_LOG(ERR, VHOST_CONFIG, "Remove entry from config_ll failed\n");
+ return NULL;
+ }
+ }
+}
+
+/**
+ * Returns the root entry of linked list
+ */
+static struct virtio_net_config_ll *
+user_get_config_ll_root(void)
+{
+ return user_ll_root;
+}
+
+/**
+ * vhost-user specific device initialization.
+ */
+static void
+user_init_device(struct vhost_device_ctx ctx, struct virtio_net *dev)
+{
+ dev->priv = ctx.user.drv->priv;
+}
+
+/**
+ * Locate the file containing QEMU's memory space and map it to our address space.
+ */
+static int
+user_host_memory_map(struct virtio_net *dev, struct virtio_memory *mem, int fd, size_t size)
+{
+ void *map;
+
+ map = mmap(0, size, PROT_READ|PROT_WRITE , MAP_POPULATE|MAP_SHARED, fd, 0);
+ close(fd);
+
+ if (map == MAP_FAILED) {
+ RTE_LOG(ERR, VHOST_CONFIG, "(%"PRIu64") Error mapping the file fd %d\n", dev->device_fh, fd);
+ return -1;
+ }
+
+ /* Store the memory address and size in the device data structure */
+ mem->mapped_address = (uint64_t)(uintptr_t)map;
+ mem->mapped_size = size;
+
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") Mem File: fd: %d - Size: %llu - VA: %p\n", dev->device_fh,
+ fd, (long long unsigned)mem->mapped_size, map);
+
+ return 0;
+}
+
+/*
+ * Called from IOCTL: VHOST_SET_MEM_TABLE
+ * This function creates and populates the memory structure for the device. This includes
+ * storing offsets used to translate buffer addresses.
+ */
+static int
+user_set_mem_table(struct vhost_device_ctx ctx, const void *mem_regions_addr, uint32_t nregions)
+{
+ struct virtio_net *dev;
+ struct vhost_memory_region *mem_regions;
+ struct virtio_memory *mem;
+ uint64_t size = offsetof(struct vhost_memory, regions);
+ uint32_t regionidx, valid_regions;
+ size_t guest_memory_size = 0;
+
+ dev = user_get_device(ctx);
+ if (dev == NULL)
+ return -1;
+
+ if (dev->mem) {
+ munmap((void *)(uintptr_t)dev->mem->mapped_address, (size_t)dev->mem->mapped_size);
+ free(dev->mem);
+ }
+
+ /* Malloc the memory structure depending on the number of regions. */
+ mem = calloc(1, sizeof(struct virtio_memory) + (sizeof(struct virtio_memory_regions) * nregions));
+ if (mem == NULL) {
+ RTE_LOG(ERR, VHOST_CONFIG, "(%"PRIu64") Failed to allocate memory for dev->mem.\n", dev->device_fh);
+ return -1;
+ }
+
+ mem->nregions = nregions;
+
+ mem_regions = (void*)(uintptr_t)((uint64_t)(uintptr_t)mem_regions_addr + size);
+
+ for (regionidx = 0; regionidx < mem->nregions; regionidx++) {
+ /* Populate the region structure for each region. */
+ mem->regions[regionidx].guest_phys_address = mem_regions[regionidx].guest_phys_addr;
+ mem->regions[regionidx].guest_phys_address_end = mem->regions[regionidx].guest_phys_address +
+ mem_regions[regionidx].memory_size;
+ mem->regions[regionidx].memory_size = mem_regions[regionidx].memory_size;
+ mem->regions[regionidx].userspace_address = mem_regions[regionidx].userspace_addr;
+ guest_memory_size += mem_regions[regionidx].memory_size;
+
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") REGION: %u - GPA: %p - QEMU VA: %p - SIZE (%"PRIu64")\n", dev->device_fh,
+ regionidx, (void*)(uintptr_t)mem->regions[regionidx].guest_phys_address,
+ (void*)(uintptr_t)mem->regions[regionidx].userspace_address,
+ mem->regions[regionidx].memory_size);
+ }
+
+ for (regionidx = 0; regionidx < mem->nregions; regionidx++) {
+ /*set the base address mapping*/
+ if (mem->regions[regionidx].guest_phys_address == 0x0) {
+ mem->base_address = mem->regions[regionidx].userspace_address;
+ /* Map VM memory file */
+ if (user_host_memory_map(dev, mem, ctx.user.fds[regionidx], guest_memory_size) != 0) {
+ free(mem);
+ return -1;
+ }
+ } else
+ close(ctx.user.fds[regionidx]);
+ }
+
+ /* Check that we have a valid base address. */
+ if (mem->base_address == 0) {
+ RTE_LOG(ERR, VHOST_CONFIG, "(%"PRIu64") Failed to find base address of qemu memory file.\n", dev->device_fh);
+ free(mem);
+ return -1;
+ }
+
+ /* Check if all of our regions have valid mappings. Usually one does not exist in the QEMU memory file. */
+ valid_regions = mem->nregions;
+ for (regionidx = 0; regionidx < mem->nregions; regionidx++) {
+ if ((mem->regions[regionidx].userspace_address < mem->base_address) ||
+ (mem->regions[regionidx].userspace_address > (mem->base_address + mem->mapped_size)))
+ valid_regions--;
+ }
+
+ /* If a region does not have a valid mapping we rebuild our memory struct to contain only valid entries. */
+ if (valid_regions != mem->nregions) {
+ LOG_DEBUG(VHOST_CONFIG, "(%"PRIu64") Not all memory regions exist in the QEMU mem file. Re-populating mem structure\n",
+ dev->device_fh);
+
+ /* Re-populate the memory structure with only valid regions. Invalid regions are over-written with memmove. */
+ valid_regions = 0;
+
+ for (regionidx = mem->nregions; 0 != regionidx--;) {
+ if ((mem->regions[regionidx].userspace_address < mem->base_address) ||
+ (mem->regions[regionidx].userspace_address > (mem->base_address + mem->mapped_size))) {
+ memmove(&mem->regions[regionidx], &mem->regions[regionidx + 1],
+ sizeof(struct virtio_memory_regions) * valid_regions);
+ } else {
+ valid_regions++;
+ }
+ }
+ }
+ mem->nregions = valid_regions;
+ dev->mem = mem;
+
+ /*
+ * Calculate the address offset for each region. This offset is used to identify the vhost virtual address
+ * corresponding to a QEMU guest physical address.
+ */
+ for (regionidx = 0; regionidx < dev->mem->nregions; regionidx++)
+ dev->mem->regions[regionidx].address_offset = dev->mem->regions[regionidx].userspace_address - dev->mem->base_address
+ + dev->mem->mapped_address - dev->mem->regions[regionidx].guest_phys_address;
+
+ return 0;
+}
+
+/**
+ * Called from IOCTL: VHOST_GET_VRING_BASE
+ * We send the virtio device our available ring last used index.
+ */
+static int
+user_get_vring_base(struct vhost_device_ctx ctx, uint32_t index, struct vhost_vring_state *state)
+{
+ struct virtio_net *dev;
+
+ dev = user_get_device(ctx);
+ if (dev == NULL)
+ return -1;
+
+ state->index = index;
+ /* State->index refers to the queue index. The TX queue is 1, RX queue is 0. */
+ state->num = dev->virtqueue[state->index]->last_used_idx;
+
+ return 0;
+}
+
+/**
+ * Called from IOCTL: VHOST_SET_VRING_CALL
+ * The virtio device sends an eventfd to interrupt the guest. This fd gets copied in
+ * to our process space.
+ * Also this message is sent when virtio-net device is reset by device driver on QEMU.
+ */
+static int
+user_set_vring_call(struct vhost_device_ctx ctx, struct vhost_vring_file *file)
+{
+ struct virtio_net *dev;
+ struct vhost_virtqueue *vq;
+
+ dev = user_get_device(ctx);
+ if (dev == NULL)
+ return -1;
+
+ /* file->index refers to the queue index. The TX queue is 1, RX queue is 0. */
+ vq = dev->virtqueue[file->index];
+
+ if (vq->kickfd)
+ close((int)vq->kickfd);
+
+ /* Populate the eventfd_copy structure and call eventfd_copy. */
+ vq->kickfd = ctx.user.fds[0];
+
+ return 0;
+}
+
+/**
+ * Called from IOCTL: VHOST_SET_VRING_KICK
+ * The virtio device sends an eventfd that it can use to notify us. This fd gets copied in
+ * to our process space.
+ */
+static int
+user_set_vring_kick(struct vhost_device_ctx ctx, struct vhost_vring_file *file)
+{
+ struct virtio_net *dev;
+ struct vhost_virtqueue *vq;
+
+ dev = user_get_device(ctx);
+ if (dev == NULL)
+ return -1;
+
+ /* file->index refers to the queue index. The TX queue is 1, RX queue is 0. */
+ vq = dev->virtqueue[file->index];
+
+ if (vq->callfd)
+ close((int)vq->callfd);
+
+ /* Populate the eventfd_copy structure and call eventfd_copy. */
+ vq->callfd = ctx.user.fds[0];
+
+ if ((dev->virtqueue[VIRTIO_RXQ] != NULL) && (dev->virtqueue[VIRTIO_TXQ]) != NULL)
+ return set_backend(ctx, file);
+
+ return 0;
+}
+
+/**
+ * Function pointers are set for the device operations to allow to call functions
+ * when an IOCTL, device_add or device_release is received.
+ */
+static const struct vhost_net_device_ops vhost_user_device_ops = {
+ .new_device = new_device,
+ .destroy_device = destroy_device,
+
+ .get_features = get_features,
+ .set_features = set_features,
+
+ .set_mem_table = user_set_mem_table,
+
+ .set_vring_num = set_vring_num,
+ .set_vring_addr = set_vring_addr,
+ .set_vring_base = set_vring_base,
+ .get_vring_base = user_get_vring_base,
+
+ .set_vring_kick = user_set_vring_kick,
+ .set_vring_call = user_set_vring_call,
+
+ .set_backend = set_backend,
+
+ .set_owner = set_owner,
+ .reset_owner = reset_owner,
+};
diff --git a/lib/librte_vhost/virtio-net.c b/lib/librte_vhost/virtio-net.c
index 13fbb6f..db810e7 100644
--- a/lib/librte_vhost/virtio-net.c
+++ b/lib/librte_vhost/virtio-net.c
@@ -96,6 +96,12 @@ qva_to_vva(struct virtio_net *dev, uint64_t qemu_va)
*/
#include "virtio-net-cdev.c"
+/**
+ * Include vhost-user depend functions and definitions.
+ */
+#include "virtio-net-user.c"
+
+
/*
* Retrieves an entry from the devices configuration linked list.
*/
@@ -105,6 +111,8 @@ get_config_ll_entry(struct vhost_device_ctx ctx)
switch (ctx.type) {
case VHOST_DRV_CUSE:
return cdev_get_config_ll_entry(ctx);
+ case VHOST_DRV_USER:
+ return user_get_config_ll_entry(ctx);
default:
break;
}
@@ -120,6 +128,8 @@ get_device(struct vhost_device_ctx ctx)
switch (ctx.type) {
case VHOST_DRV_CUSE:
return cdev_get_device(ctx);
+ case VHOST_DRV_USER:
+ return user_get_device(ctx);
default:
break;
}
@@ -136,6 +146,8 @@ add_config_ll_entry(vhost_driver_type_t type,
switch (type) {
case VHOST_DRV_CUSE:
return cdev_add_config_ll_entry(new_ll_dev);
+ case VHOST_DRV_USER:
+ return user_add_config_ll_entry(new_ll_dev);
default:
break;
}
@@ -149,8 +161,39 @@ cleanup_device(struct virtio_net *dev)
{
/* Unmap QEMU memory file if mapped. */
if (dev->mem) {
- munmap((void *)(uintptr_t)dev->mem->mapped_address,
- (size_t)dev->mem->mapped_size);
+ {
+ /*
+ * 'munmap()' will be failed when mapped_size isn't
+ * aligned with hugepage size.
+ * Usually a file size of QEMU physical memory is
+ * aligned by hugepage size. So In a case of CUSE,
+ * there is no problem. But with vhost-user, there is
+ * no way to get physical memory size.
+ *
+ * Let's assume hugepage size is 2MB or 1GB here.
+ * BTW, 'mmap()' automatically fixed size parameter
+ * to be aligned. Why does 'munmap()' do like so?
+ */
+ int ret = 0;
+ size_t hugepagesize, size = dev->mem->mapped_size;
+
+ /* assume hugepage size is 2MB */
+ hugepagesize = 2 * 1024 * 1024;
+ size = (size + hugepagesize - 1) /
+ hugepagesize * hugepagesize;
+ ret = munmap((void *)(uintptr_t)
+ dev->mem->mapped_address,
+ size);
+ if (ret) {
+ /* assume hugepage size is 1GB, try again */
+ hugepagesize = 1024 * 1024 * 1024;
+ size = (size + hugepagesize - 1) /
+ hugepagesize * hugepagesize;
+ munmap((void *)(uintptr_t)
+ dev->mem->mapped_address,
+ size);
+ }
+ }
free(dev->mem);
}
@@ -187,6 +230,8 @@ rm_config_ll_entry(vhost_driver_type_t type,
switch (type) {
case VHOST_DRV_CUSE:
return cdev_rm_config_ll_entry(ll_dev, ll_dev_last);
+ case VHOST_DRV_USER:
+ return user_rm_config_ll_entry(ll_dev, ll_dev_last);
default:
break;
}
@@ -201,7 +246,9 @@ get_config_ll_root(struct vhost_device_ctx ctx)
{
switch (ctx.type) {
case VHOST_DRV_CUSE:
- return cdev_get_config_ll_root(ctx);
+ return cdev_get_config_ll_root();
+ case VHOST_DRV_USER:
+ return user_get_config_ll_root();
default:
break;
}
@@ -232,6 +279,8 @@ init_device(struct vhost_device_ctx ctx, struct virtio_net *dev)
switch (ctx.type) {
case VHOST_DRV_CUSE:
return cdev_init_device(ctx, dev);
+ case VHOST_DRV_USER:
+ return user_init_device(ctx, dev);
default:
break;
}
@@ -527,6 +576,8 @@ get_virtio_net_callbacks(vhost_driver_type_t type)
switch (type) {
case VHOST_DRV_CUSE:
return &vhost_cuse_device_ops;
+ case VHOST_DRV_USER:
+ return &vhost_user_device_ops;
default:
break;
}
@@ -570,9 +621,14 @@ int rte_vhost_feature_enable(uint64_t feature_mask)
* Register ops so that we can add/remove device to data core.
*/
int
-rte_vhost_driver_callback_register(struct virtio_net_device_ops const * const ops)
+rte_vhost_driver_callback_register(struct vhost_driver *drv,
+ struct virtio_net_device_ops const * const ops, void *priv)
{
+ if (drv == NULL || ops == NULL)
+ return -1;
+
notify_ops = ops;
+ drv->priv = priv;
return 0;
}
--
1.9.1
next prev parent reply other threads:[~2014-11-06 11:05 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-11-06 11:14 [dpdk-dev] [RFC PATCH 0/7] lib/librte_vhost: Add vhost-user extension Tetsuya Mukawa
2014-11-06 11:14 ` [dpdk-dev] [RFC PATCH 1/7] lib/librte_vhost: Fix host_memory_map() to handle various memory regions Tetsuya Mukawa
2014-11-06 11:14 ` [dpdk-dev] [RFC PATCH 2/7] lib/librte_vhost: Add an abstraction layer for vhost backends Tetsuya Mukawa
2014-11-06 11:14 ` [dpdk-dev] [RFC PATCH 3/7] lib/librte_vhost: Add an abstraction layer tointerpret messages Tetsuya Mukawa
2014-11-07 20:43 ` Xie, Huawei
2014-11-10 5:12 ` Tetsuya Mukawa
2014-11-10 8:07 ` Xie, Huawei
2014-11-10 8:44 ` Tetsuya Mukawa
2014-11-06 11:14 ` [dpdk-dev] [RFC PATCH 4/7] lib/librte_vhost: Move vhost vhost-cuse device list and accessor functions Tetsuya Mukawa
2014-11-06 11:14 ` [dpdk-dev] [RFC PATCH 5/7] lib/librte_vhost: Add a vhost session abstraction Tetsuya Mukawa
2014-11-06 11:14 ` [dpdk-dev] [RFC PATCH 6/7] lib/librte_vhost: Add vhost-cuse/user specific initialization Tetsuya Mukawa
2014-11-06 11:14 ` Tetsuya Mukawa [this message]
2014-11-07 21:25 ` [dpdk-dev] [RFC PATCH 7/7] lib/librte_vhost: Add vhost-user implementation Xie, Huawei
2014-11-10 5:11 ` Tetsuya Mukawa
2014-11-10 8:18 ` Xie, Huawei
2014-11-10 8:55 ` Tetsuya Mukawa
2014-11-14 0:07 ` Xie, Huawei
2014-11-14 4:41 ` Tetsuya Mukawa
2014-11-07 3:33 ` [dpdk-dev] [RFC PATCH 0/7] lib/librte_vhost: Add vhost-user extension Xie, Huawei
2014-11-07 5:09 ` Tetsuya Mukawa
[not found] ` <C37D651A908B024F974696C65296B57B0F2E3C93@SHSMSX101.ccr.corp.intel.com>
2014-11-07 6:16 ` Tetsuya Mukawa
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1415272471-3299-8-git-send-email-mukawa@igel.co.jp \
--to=mukawa@igel.co.jp \
--cc=dev@dpdk.org \
--cc=masutani.hitoshi@lab.ntt.co.jp \
--cc=nakajima.yoshihiro@lab.ntt.co.jp \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).