* Re: [dpdk-dev] [PATCH v4] eal: add asynchronous request API to DPDK IPC
2018-03-13 17:42 ` [dpdk-dev] [PATCH v4] " Anatoly Burakov
@ 2018-03-23 15:38 ` Tan, Jianfeng
2018-03-23 18:21 ` Burakov, Anatoly
2018-03-24 12:46 ` [dpdk-dev] [PATCH v5 1/2] eal: rename IPC sync request to pending request Anatoly Burakov
2018-03-24 12:46 ` [dpdk-dev] [PATCH v5 " Anatoly Burakov
2 siblings, 1 reply; 40+ messages in thread
From: Tan, Jianfeng @ 2018-03-23 15:38 UTC (permalink / raw)
To: Anatoly Burakov, dev; +Cc: konstantin.ananyev
Hi Anatoly,
Two general comments firstly.
Q1. As I understand, malloc usage as an example, when a primary process
receives a request in rte_mp_handle thread, it will allocate memory, and
broadcast an async request to all secondary processes, and it will
return immediately; then responses are replied from each secondary
process, which are recorded at rte_mp_async_handle thread firstly;
either timeout or responses are all collected, rte_mp_async_handle will
trigger an async action.
I agree the necessity of the async request API; without it, each caller
who has similar requirement needs lots of code to implement it.
But can we achieve that without creating another pthread by leveraging
the alarm set? For example, to send an async request,
step 1. set an alarm;
step 2. send the request;
step 3. receive and record responses in mp thread;
step 4. if alarm timeout, trigger the async action with timeout result;
step 5. if all responses are collected, cancel the alarm, and trigger
the async action with success result.
I don't have strong objection for adding another thread, and actually
this is an internal thing in ipc system, we can optimize it later.
Q2. Do we really need to register actions for result handler of async
request? Instead, we can put it as a parameter into
rte_mp_request_async(), whenever the request fails or succeeds, we just
invoke the function.
Please see some other comments inline.
On 3/14/2018 1:42 AM, Anatoly Burakov wrote:
> This API is similar to the blocking API that is already present,
> but reply will be received in a separate callback by the caller.
>
> Under the hood, we create a separate thread to deal with replies to
> asynchronous requests, that will just wait to be notified by the
> main thread, or woken up on a timer (it'll wake itself up every
> minute regardless of whether it was called, but if there are no
> requests in the queue, nothing will be done and it'll go to sleep
> for another minute).
Wait for 1min seems a little strange to me; it shall wake up for a
latest timeout of pending async requests?
> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
> ---
>
> Notes:
> v4:
> - rebase on top of latest IPC Improvements patchset [2]
> v3:
> - added support for MP_IGN messages introduced in
> IPC improvements v5 patchset
> v2:
> - fixed deadlocks and race conditions by not calling
> callbacks while iterating over sync request list
> - fixed use-after-free by making a copy of request
> - changed API to also give user a copy of original
> request, so that they know to which message the
> callback is a reply to
> - fixed missing .map file entries
>
> This patch is dependent upon previously published patchsets
> for IPC fixes [1] and improvements [2].
>
> rte_mp_action_unregister and rte_mp_async_reply_unregister
> do the same thing - should we perhaps make it one function?
>
> [1] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Fixes/
> [2] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Improvements/
>
> lib/librte_eal/common/eal_common_proc.c | 563 ++++++++++++++++++++++++++++++--
> lib/librte_eal/common/include/rte_eal.h | 72 ++++
> lib/librte_eal/rte_eal_version.map | 3 +
> 3 files changed, 607 insertions(+), 31 deletions(-)
>
> diff --git a/lib/librte_eal/common/eal_common_proc.c b/lib/librte_eal/common/eal_common_proc.c
> index 4131b67..50d6506 100644
> --- a/lib/librte_eal/common/eal_common_proc.c
> +++ b/lib/librte_eal/common/eal_common_proc.c
> @@ -26,6 +26,7 @@
> #include <rte_errno.h>
> #include <rte_lcore.h>
> #include <rte_log.h>
> +#include <rte_tailq.h>
>
> #include "eal_private.h"
> #include "eal_filesystem.h"
> @@ -39,7 +40,11 @@ static pthread_mutex_t mp_mutex_action = PTHREAD_MUTEX_INITIALIZER;
> struct action_entry {
> TAILQ_ENTRY(action_entry) next;
> char action_name[RTE_MP_MAX_NAME_LEN];
> - rte_mp_t action;
> + RTE_STD_C11
> + union {
> + rte_mp_t action;
> + rte_mp_async_reply_t reply;
> + };
> };
>
> /** Double linked list of actions. */
> @@ -60,13 +65,37 @@ struct mp_msg_internal {
> struct rte_mp_msg msg;
> };
>
> +enum mp_request_type {
> + REQUEST_TYPE_SYNC,
> + REQUEST_TYPE_ASYNC
> +};
> +
> +struct async_request_shared_param {
> + struct rte_mp_reply *user_reply;
> + struct timespec *end;
Why we have to make these two as pointers? Operating on pointers
introduce unnecessary complexity.
> + int n_requests_processed;
It sounds like recording how many requests are sent, but it means how
many responses are received, right? n_responses sounds better?
> +};
> +
> +struct async_request_param {
> + struct async_request_shared_param *param;
> +};
> +
> +struct sync_request_param {
> + pthread_cond_t cond;
> +};
> +
> struct sync_request {
I know "sync_" in the original version was for synchronization between
the blocked thread (who sends the request) and the mp thread.
But it indeed makes the code difficult to understand. We may change the
name to "pending_request"?
> TAILQ_ENTRY(sync_request) next;
> - int reply_received;
> + enum mp_request_type type;
> char dst[PATH_MAX];
> struct rte_mp_msg *request;
> - struct rte_mp_msg *reply;
> - pthread_cond_t cond;
> + struct rte_mp_msg *reply_msg;
> + int reply_received;
> + RTE_STD_C11
> + union {
> + struct sync_request_param sync;
> + struct async_request_param async;
> + };
> };
Too many structs are defined? How about just putting it like this:
struct pending_request {
TAILQ_ENTRY(sync_request) next;
enum mp_request_type type;
char dst[PATH_MAX];
struct rte_mp_msg *request;
struct rte_mp_msg *reply_msg;
int reply_received;
RTE_STD_C11
union {
/* for sync request */
struct {
pthread_cond_t cond; /* used for mp thread to
wake up requesting thread */
};
/* for async request */
struct {
struct rte_mp_reply user_reply;
struct timespec end;
int n_requests_processed; /* store how requests */
};
};
};
>
> TAILQ_HEAD(sync_request_list, sync_request);
> @@ -74,9 +103,12 @@ TAILQ_HEAD(sync_request_list, sync_request);
> static struct {
> struct sync_request_list requests;
> pthread_mutex_t lock;
> + pthread_cond_t async_cond;
> } sync_requests = {
> .requests = TAILQ_HEAD_INITIALIZER(sync_requests.requests),
> - .lock = PTHREAD_MUTEX_INITIALIZER
> + .lock = PTHREAD_MUTEX_INITIALIZER,
> + .async_cond = PTHREAD_COND_INITIALIZER
> + /**< used in async requests only */
> };
>
> /* forward declarations */
> @@ -164,53 +196,97 @@ validate_action_name(const char *name)
> return 0;
> }
>
> -int __rte_experimental
> -rte_mp_action_register(const char *name, rte_mp_t action)
> +static struct action_entry *
> +action_register(const char *name)
> {
> struct action_entry *entry;
>
> if (validate_action_name(name))
> - return -1;
> + return NULL;
>
> entry = malloc(sizeof(struct action_entry));
> if (entry == NULL) {
> rte_errno = ENOMEM;
> - return -1;
> + return NULL;
> }
> strcpy(entry->action_name, name);
> - entry->action = action;
>
> - pthread_mutex_lock(&mp_mutex_action);
> if (find_action_entry_by_name(name) != NULL) {
> pthread_mutex_unlock(&mp_mutex_action);
> rte_errno = EEXIST;
> free(entry);
> - return -1;
> + return NULL;
> }
> TAILQ_INSERT_TAIL(&action_entry_list, entry, next);
> - pthread_mutex_unlock(&mp_mutex_action);
> - return 0;
> +
> + /* async and sync replies are handled by different threads, so even
> + * though they a share pointer in a union, one will never trigger in
> + * place of the other.
> + */
> +
> + return entry;
> }
>
> -void __rte_experimental
> -rte_mp_action_unregister(const char *name)
> +static void
> +action_unregister(const char *name)
> {
> struct action_entry *entry;
>
> if (validate_action_name(name))
> return;
>
> - pthread_mutex_lock(&mp_mutex_action);
> entry = find_action_entry_by_name(name);
> if (entry == NULL) {
> - pthread_mutex_unlock(&mp_mutex_action);
> return;
> }
> TAILQ_REMOVE(&action_entry_list, entry, next);
> - pthread_mutex_unlock(&mp_mutex_action);
> free(entry);
> }
>
> +int __rte_experimental
> +rte_mp_action_register(const char *name, rte_mp_t action)
> +{
> + struct action_entry *entry;
> +
> + pthread_mutex_lock(&mp_mutex_action);
> +
> + entry = action_register(name);
> + if (entry != NULL)
> + entry->action = action;
> + pthread_mutex_unlock(&mp_mutex_action);
> +
> + return entry == NULL ? -1 : 0;
> +}
> +
> +void __rte_experimental
> +rte_mp_action_unregister(const char *name)
> +{
> + pthread_mutex_lock(&mp_mutex_action);
> + action_unregister(name);
> + pthread_mutex_unlock(&mp_mutex_action);
> +}
> +
> +int __rte_experimental
> +rte_mp_async_reply_register(const char *name, rte_mp_async_reply_t reply)
> +{
> + struct action_entry *entry;
> +
> + pthread_mutex_lock(&mp_mutex_action);
> +
> + entry = action_register(name);
> + if (entry != NULL)
> + entry->reply = reply;
> + pthread_mutex_unlock(&mp_mutex_action);
> +
> + return entry == NULL ? -1 : 0;
> +}
> +
> +void __rte_experimental
> +rte_mp_async_reply_unregister(const char *name)
> +{
> + rte_mp_action_unregister(name);
> +}
> +
> static int
> read_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
> {
> @@ -270,10 +346,14 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
> pthread_mutex_lock(&sync_requests.lock);
> sync_req = find_sync_request(s->sun_path, msg->name);
> if (sync_req) {
> - memcpy(sync_req->reply, msg, sizeof(*msg));
> + memcpy(sync_req->reply_msg, msg, sizeof(*msg));
> /* -1 indicates that we've been asked to ignore */
> sync_req->reply_received = m->type == MP_REP ? 1 : -1;
> - pthread_cond_signal(&sync_req->cond);
> +
> + if (sync_req->type == REQUEST_TYPE_SYNC)
> + pthread_cond_signal(&sync_req->sync.cond);
> + else if (sync_req->type == REQUEST_TYPE_ASYNC)
> + pthread_cond_signal(&sync_requests.async_cond);
> } else
> RTE_LOG(ERR, EAL, "Drop mp reply: %s\n", msg->name);
> pthread_mutex_unlock(&sync_requests.lock);
> @@ -320,6 +400,204 @@ mp_handle(void *arg __rte_unused)
> }
>
> static int
> +timespec_cmp(const struct timespec *a, const struct timespec *b)
> +{
> + if (a->tv_sec < b->tv_sec)
> + return -1;
> + if (a->tv_sec > b->tv_sec)
> + return 1;
> + if (a->tv_nsec < b->tv_nsec)
> + return -1;
> + if (a->tv_nsec > b->tv_nsec)
> + return 1;
> + return 0;
> +}
> +
> +enum async_action {
> + ACTION_NONE, /**< don't do anything */
> + ACTION_FREE, /**< free the action entry, but don't trigger callback */
> + ACTION_TRIGGER /**< trigger callback, then free action entry */
> +};
> +
> +static enum async_action
> +process_async_request(struct sync_request *sr, const struct timespec *now)
> +{
> + struct async_request_shared_param *param;
> + struct rte_mp_reply *reply;
> + bool timeout, received, last_msg;
> +
> + param = sr->async.param;
> + reply = param->user_reply;
> +
> + /* did we timeout? */
> + timeout = timespec_cmp(param->end, now) <= 0;
> +
> + /* did we receive a response? */
> + received = sr->reply_received != 0;
> +
> + /* if we didn't time out, and we didn't receive a response, ignore */
> + if (!timeout && !received)
> + return ACTION_NONE;
> +
> + /* if we received a response, adjust relevant data and copy mesasge. */
> + if (sr->reply_received == 1 && sr->reply_msg) {
> + struct rte_mp_msg *msg, *user_msgs, *tmp;
> +
> + msg = sr->reply_msg;
> + user_msgs = reply->msgs;
> +
> + tmp = realloc(user_msgs, sizeof(*msg) *
> + (reply->nb_received + 1));
> + if (!tmp) {
> + RTE_LOG(ERR, EAL, "Fail to alloc reply for request %s:%s\n",
> + sr->dst, sr->request->name);
> + /* this entry is going to be removed and its message
> + * dropped, but we don't want to leak memory, so
> + * continue.
> + */
> + } else {
> + user_msgs = tmp;
> + reply->msgs = user_msgs;
> + memcpy(&user_msgs[reply->nb_received],
> + msg, sizeof(*msg));
> + reply->nb_received++;
> + }
> + } else if (sr->reply_received == -1) {
> + /* we were asked to ignore this process */
> + reply->nb_sent--;
> + }
> + free(sr->reply_msg);
> +
> + /* mark this request as processed */
> + param->n_requests_processed++;
> +
> + /* if number of sent messages is zero, we're short-circuiting */
> + last_msg = param->n_requests_processed == reply->nb_sent ||
> + reply->nb_sent == 0;
> +
> + return last_msg ? ACTION_TRIGGER : ACTION_FREE;
> +}
> +
> +static void
> +trigger_async_action(struct sync_request *sr)
> +{
> + struct async_request_shared_param *param;
> + struct rte_mp_reply *reply;
> +
> + param = sr->async.param;
> + reply = param->user_reply;
> +
> + pthread_mutex_lock(&mp_mutex_action);
> + struct action_entry *entry =
> + find_action_entry_by_name(sr->request->name);
> + pthread_mutex_unlock(&mp_mutex_action);
> + if (!entry)
> + RTE_LOG(ERR, EAL, "Cannot find async request callback for %s\n",
> + sr->request->name);
> + else
> + entry->reply(sr->request, reply);
> + /* clean up */
> + free(sr->async.param->user_reply->msgs);
> + free(sr->async.param->user_reply);
> + free(sr->async.param->end);
> + free(sr->async.param);
> + free(sr->request);
> +}
> +
> +static void *
> +async_reply_handle(void *arg __rte_unused)
> +{
> + struct sync_request *sr;
> + struct timeval now;
> + struct timespec timeout, ts_now;
> + do {
Put while(1) here so that it's clear to be an infinite loop.
> + struct sync_request *trigger = NULL;
> + int ret;
> + bool dontwait = false;
> +
> + pthread_mutex_lock(&sync_requests.lock);
> +
> + if (gettimeofday(&now, NULL) < 0) {
> + RTE_LOG(ERR, EAL, "Cannot get current time\n");
> + pthread_mutex_unlock(&sync_requests.lock);
> + break;
> + }
> +
> + /* set a 60 second timeout by default */
> + timeout.tv_nsec = (now.tv_usec * 1000 + 60) % 1000000000;
> + timeout.tv_sec = now.tv_sec + 60 +
> + (now.tv_usec * 1000 + 60) / 1000000000;
> +
> + /* scan through the list and see if there are any timeouts that
> + * are earlier than our current timeout.
> + */
> + TAILQ_FOREACH(sr, &sync_requests.requests, next) {
> + if (sr->type != REQUEST_TYPE_ASYNC)
> + continue;
> + if (timespec_cmp(sr->async.param->end, &timeout) < 0)
> + memcpy(&timeout, sr->async.param->end,
> + sizeof(timeout));
> +
> + /* sometimes, we don't even wait */
> + if (sr->reply_received) {
> + dontwait = true;
> + break;
> + }
> + }
> +
> + /* now, wait until we either time out or get woken up */
> + if (!dontwait)
> + ret = pthread_cond_timedwait(&sync_requests.async_cond,
> + &sync_requests.lock, &timeout);
> + else
> + ret = 0;
> +
> + if (gettimeofday(&now, NULL) < 0) {
> + RTE_LOG(ERR, EAL, "Cannot get current time\n");
> + break;
> + }
> + ts_now.tv_nsec = now.tv_usec * 1000;
> + ts_now.tv_sec = now.tv_sec;
> +
> + if (ret == 0 || ret == ETIMEDOUT) {
> + struct sync_request *next;
> + /* we've either been woken up, or we timed out */
> +
> + /* we have still the lock, check if anything needs
> + * processing.
> + */
> + TAILQ_FOREACH_SAFE(sr, &sync_requests.requests, next,
> + next) {
> + enum async_action action;
> + if (sr->type != REQUEST_TYPE_ASYNC)
> + continue;
> +
> + action = process_async_request(sr, &ts_now);
> + if (action == ACTION_FREE) {
> + TAILQ_REMOVE(&sync_requests.requests,
> + sr, next);
> + free(sr);
> + } else if (action == ACTION_TRIGGER &&
> + trigger == NULL) {
> + TAILQ_REMOVE(&sync_requests.requests,
> + sr, next);
> + trigger = sr;
> + }
> + }
> + }
> + pthread_mutex_unlock(&sync_requests.lock);
> + if (trigger) {
> + trigger_async_action(trigger);
> + free(trigger);
> + }
> + } while (1);
> +
> + RTE_LOG(ERR, EAL, "ERROR: asynchronous requests disabled\n");
> +
> + return NULL;
> +}
> +
> +static int
> open_socket_fd(void)
> {
> char peer_name[PATH_MAX] = {0};
> @@ -382,7 +660,7 @@ rte_mp_channel_init(void)
> char thread_name[RTE_MAX_THREAD_NAME_LEN];
> char path[PATH_MAX];
> int dir_fd;
> - pthread_t tid;
> + pthread_t mp_handle_tid, async_reply_handle_tid;
>
> /* create filter path */
> create_socket_path("*", path, sizeof(path));
> @@ -419,7 +697,16 @@ rte_mp_channel_init(void)
> return -1;
> }
>
> - if (pthread_create(&tid, NULL, mp_handle, NULL) < 0) {
> + if (pthread_create(&mp_handle_tid, NULL, mp_handle, NULL) < 0) {
> + RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
> + strerror(errno));
> + close(mp_fd);
> + mp_fd = -1;
> + return -1;
> + }
> +
> + if (pthread_create(&async_reply_handle_tid, NULL,
> + async_reply_handle, NULL) < 0) {
> RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
> strerror(errno));
> close(mp_fd);
> @@ -430,7 +717,11 @@ rte_mp_channel_init(void)
>
> /* try best to set thread name */
> snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_handle");
> - rte_thread_setname(tid, thread_name);
> + rte_thread_setname(mp_handle_tid, thread_name);
> +
> + /* try best to set thread name */
> + snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_async_handle");
> + rte_thread_setname(async_reply_handle_tid, thread_name);
>
> /* unlock the directory */
> flock(dir_fd, LOCK_UN);
> @@ -602,18 +893,77 @@ rte_mp_sendmsg(struct rte_mp_msg *msg)
> }
>
> static int
> -mp_request_one(const char *dst, struct rte_mp_msg *req,
> +mp_request_async(const char *dst, struct rte_mp_msg *req,
> + struct async_request_shared_param *param)
> +{
> + struct rte_mp_msg *reply_msg;
> + struct sync_request *sync_req, *exist;
> + int ret;
> +
> + sync_req = malloc(sizeof(*sync_req));
> + reply_msg = malloc(sizeof(*reply_msg));
> + if (sync_req == NULL || reply_msg == NULL) {
> + RTE_LOG(ERR, EAL, "Could not allocate space for sync request\n");
> + rte_errno = ENOMEM;
> + ret = -1;
> + goto fail;
> + }
> +
> + memset(sync_req, 0, sizeof(*sync_req));
> + memset(reply_msg, 0, sizeof(*reply_msg));
> +
> + sync_req->type = REQUEST_TYPE_ASYNC;
> + strcpy(sync_req->dst, dst);
> + sync_req->request = req;
> + sync_req->reply_msg = reply_msg;
> + sync_req->async.param = param;
> +
> + /* queue already locked by caller */
> +
> + exist = find_sync_request(dst, req->name);
> + if (!exist)
> + TAILQ_INSERT_TAIL(&sync_requests.requests, sync_req, next);
> + if (exist) {
> + RTE_LOG(ERR, EAL, "A pending request %s:%s\n", dst, req->name);
> + rte_errno = EEXIST;
> + ret = -1;
> + goto fail;
> + }
> +
> + ret = send_msg(dst, req, MP_REQ);
> + if (ret < 0) {
> + RTE_LOG(ERR, EAL, "Fail to send request %s:%s\n",
> + dst, req->name);
> + ret = -1;
> + goto fail;
> + } else if (ret == 0) {
> + ret = 0;
> + goto fail;
> + }
> +
> + param->user_reply->nb_sent++;
> +
> + return 0;
> +fail:
> + free(sync_req);
> + free(reply_msg);
> + return ret;
> +}
> +
> +static int
> +mp_request_sync(const char *dst, struct rte_mp_msg *req,
> struct rte_mp_reply *reply, const struct timespec *ts)
> {
> int ret;
> struct rte_mp_msg msg, *tmp;
> struct sync_request sync_req, *exist;
>
> + sync_req.type = REQUEST_TYPE_SYNC;
> sync_req.reply_received = 0;
> strcpy(sync_req.dst, dst);
> sync_req.request = req;
> - sync_req.reply = &msg;
> - pthread_cond_init(&sync_req.cond, NULL);
> + sync_req.reply_msg = &msg;
> + pthread_cond_init(&sync_req.sync.cond, NULL);
>
> pthread_mutex_lock(&sync_requests.lock);
> exist = find_sync_request(dst, req->name);
> @@ -637,7 +987,7 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
> reply->nb_sent++;
>
> do {
> - ret = pthread_cond_timedwait(&sync_req.cond,
> + ret = pthread_cond_timedwait(&sync_req.sync.cond,
> &sync_requests.lock, ts);
> } while (ret != 0 && ret != ETIMEDOUT);
>
> @@ -703,7 +1053,7 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
>
> /* for secondary process, send request to the primary process only */
> if (rte_eal_process_type() == RTE_PROC_SECONDARY)
> - return mp_request_one(eal_mp_socket_path(), req, reply, &end);
> + return mp_request_sync(eal_mp_socket_path(), req, reply, &end);
>
> /* for primary process, broadcast request, and collect reply 1 by 1 */
> mp_dir = opendir(mp_dir_path);
> @@ -732,7 +1082,7 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
> snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
> ent->d_name);
>
> - if (mp_request_one(path, req, reply, &end))
> + if (mp_request_sync(path, req, reply, &end))
> ret = -1;
> }
> /* unlock the directory */
> @@ -744,9 +1094,160 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
> }
>
> int __rte_experimental
> -rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
> +rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts)
> {
> + struct rte_mp_msg *copy;
> + struct sync_request *dummy;
> + struct async_request_shared_param *param = NULL;
> + struct rte_mp_reply *reply = NULL;
> + int dir_fd, ret = 0;
> + DIR *mp_dir;
> + struct dirent *ent;
> + struct timeval now;
> + struct timespec *end = NULL;
> +
> + RTE_LOG(DEBUG, EAL, "request: %s\n", req->name);
> +
> + if (check_input(req) == false)
> + return -1;
> + if (gettimeofday(&now, NULL) < 0) {
> + RTE_LOG(ERR, EAL, "Faile to get current time\n");
> + rte_errno = errno;
> + return -1;
> + }
> + copy = malloc(sizeof(*copy));
> + dummy = malloc(sizeof(*dummy));
> + param = malloc(sizeof(*param));
> + reply = malloc(sizeof(*reply));
> + end = malloc(sizeof(*end));
> + if (copy == NULL || dummy == NULL || param == NULL || reply == NULL ||
> + end == NULL) {
> + RTE_LOG(ERR, EAL, "Failed to allocate memory for async reply\n");
> + rte_errno = ENOMEM;
> + goto fail;
> + }
> +
> + memset(copy, 0, sizeof(*copy));
> + memset(dummy, 0, sizeof(*dummy));
> + memset(param, 0, sizeof(*param));
> + memset(reply, 0, sizeof(*reply));
> + memset(end, 0, sizeof(*end));
> +
> + /* copy message */
> + memcpy(copy, req, sizeof(*copy));
> +
> + param->n_requests_processed = 0;
> + param->end = end;
> + param->user_reply = reply;
>
> + end->tv_nsec = (now.tv_usec * 1000 + ts->tv_nsec) % 1000000000;
> + end->tv_sec = now.tv_sec + ts->tv_sec +
> + (now.tv_usec * 1000 + ts->tv_nsec) / 1000000000;
> + reply->nb_sent = 0;
> + reply->nb_received = 0;
> + reply->msgs = NULL;
> +
> + /* we have to lock the request queue here, as we will be adding a bunch
> + * of requests to the queue at once, and some of the replies may arrive
> + * before we add all of the requests to the queue.
> + */
> + pthread_mutex_lock(&sync_requests.lock);
Why do we share this lock for both sync and async requests?
> +
> + /* we have to ensure that callback gets triggered even if we don't send
> + * anything, therefore earlier we have allocated a dummy request. put it
> + * on the queue and fill it. we will remove it once we know we sent
> + * something.
> + */
> + dummy->type = REQUEST_TYPE_ASYNC;
> + dummy->request = copy;
> + dummy->reply_msg = NULL;
> + dummy->async.param = param;
> + dummy->reply_received = 1; /* short-circuit the timeout */
> +
> + TAILQ_INSERT_TAIL(&sync_requests.requests, dummy, next);
> +
> + /* for secondary process, send request to the primary process only */
> + if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
> + ret = mp_request_async(eal_mp_socket_path(), copy, param);
> +
> + /* if we sent something, remove dummy request from the queue */
> + if (reply->nb_sent != 0) {
> + TAILQ_REMOVE(&sync_requests.requests, dummy, next);
> + free(dummy);
> + dummy = NULL;
> + }
> +
> + pthread_mutex_unlock(&sync_requests.lock);
> +
> + /* if we couldn't send anything, clean up */
> + if (ret != 0)
> + goto fail;
> + return 0;
> + }
> +
> + /* for primary process, broadcast request */
> + mp_dir = opendir(mp_dir_path);
> + if (!mp_dir) {
> + RTE_LOG(ERR, EAL, "Unable to open directory %s\n", mp_dir_path);
> + rte_errno = errno;
> + goto unlock_fail;
> + }
> + dir_fd = dirfd(mp_dir);
> +
> + /* lock the directory to prevent processes spinning up while we send */
> + if (flock(dir_fd, LOCK_EX)) {
> + RTE_LOG(ERR, EAL, "Unable to lock directory %s\n",
> + mp_dir_path);
> + rte_errno = errno;
> + goto closedir_fail;
> + }
> +
> + while ((ent = readdir(mp_dir))) {
> + char path[PATH_MAX];
> +
> + if (fnmatch(mp_filter, ent->d_name, 0) != 0)
> + continue;
> +
> + snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
> + ent->d_name);
> +
> + if (mp_request_async(path, copy, param))
> + ret = -1;
> + }
> + /* if we sent something, remove dummy request from the queue */
> + if (reply->nb_sent != 0) {
> + TAILQ_REMOVE(&sync_requests.requests, dummy, next);
> + free(dummy);
> + dummy = NULL;
> + }
> + /* trigger async request thread wake up */
> + pthread_cond_signal(&sync_requests.async_cond);
> +
> + /* finally, unlock the queue */
> + pthread_mutex_unlock(&sync_requests.lock);
> +
> + /* unlock the directory */
> + flock(dir_fd, LOCK_UN);
> +
> + /* dir_fd automatically closed on closedir */
> + closedir(mp_dir);
> + return ret;
> +closedir_fail:
> + closedir(mp_dir);
> +unlock_fail:
> + pthread_mutex_unlock(&sync_requests.lock);
> +fail:
> + free(dummy);
> + free(param);
> + free(reply);
> + free(end);
> + free(copy);
> + return -1;
> +}
> +
> +int __rte_experimental
> +rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
> +{
> RTE_LOG(DEBUG, EAL, "reply: %s\n", msg->name);
>
> if (check_input(msg) == false)
> diff --git a/lib/librte_eal/common/include/rte_eal.h b/lib/librte_eal/common/include/rte_eal.h
> index 044474e..93ca4cc 100644
> --- a/lib/librte_eal/common/include/rte_eal.h
> +++ b/lib/librte_eal/common/include/rte_eal.h
> @@ -230,6 +230,16 @@ struct rte_mp_reply {
> typedef int (*rte_mp_t)(const struct rte_mp_msg *msg, const void *peer);
>
> /**
> + * Asynchronous reply function typedef used by other components.
> + *
> + * As we create socket channel for primary/secondary communication, use
Here are two spaces.
> + * this function typedef to register action for coming responses to asynchronous
> + * requests.
> + */
> +typedef int (*rte_mp_async_reply_t)(const struct rte_mp_msg *request,
> + const struct rte_mp_reply *reply);
> +
> +/**
> * @warning
> * @b EXPERIMENTAL: this API may change without prior notice
> *
> @@ -273,6 +283,46 @@ rte_mp_action_unregister(const char *name);
> * @warning
> * @b EXPERIMENTAL: this API may change without prior notice
> *
> + * Register an asynchronous reply callback for primary/secondary communication.
> + *
> + * Call this function to register a callback for asynchronous requests, if the
> + * calling component wants to receive responses to its own asynchronous requests
> + * from the corresponding component in its primary or secondary processes.
> + *
> + * @param name
> + * The name argument plays as a unique key to find the action.
> + *
> + * @param reply
> + * The reply argument is the function pointer to the reply callback.
> + *
> + * @return
> + * - 0 on success.
> + * - (<0) on failure.
> + */
> +int __rte_experimental
> +rte_mp_async_reply_register(const char *name, rte_mp_async_reply_t reply);
> +
> +/**
> + * @warning
> + * @b EXPERIMENTAL: this API may change without prior notice
> + *
> + * Unregister an asynchronous reply callback.
> + *
> + * Call this function to unregister a callback if the calling component does
> + * not want responses the messages from the corresponding component in its
> + * primary process or secondary processes.
> + *
> + * @param name
> + * The name argument plays as a unique key to find the action.
> + *
> + */
> +void __rte_experimental
> +rte_mp_async_reply_unregister(const char *name);
> +
> +/**
> + * @warning
> + * @b EXPERIMENTAL: this API may change without prior notice
> + *
> * Send a message to the peer process.
> *
> * This function will send a message which will be responsed by the action
> @@ -321,6 +371,28 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
> * @warning
> * @b EXPERIMENTAL: this API may change without prior notice
> *
> + * Send a request to the peer process and expect a reply in a separate callback.
> + *
> + * This function sends a request message to the peer process, and will not
> + * block. Instead, reply will be received in a separate callback.
> + *
> + * @param req
> + * The req argument contains the customized request message.
> + *
> + * @param ts
> + * The ts argument specifies how long we can wait for the peer(s) to reply.
> + *
> + * @return
> + * - On success, return 0.
> + * - On failure, return -1, and the reason will be stored in rte_errno.
> + */
> +int __rte_experimental
> +rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts);
> +
> +/**
> + * @warning
> + * @b EXPERIMENTAL: this API may change without prior notice
> + *
> * Send a reply to the peer process.
> *
> * This function will send a reply message in response to a request message
> diff --git a/lib/librte_eal/rte_eal_version.map b/lib/librte_eal/rte_eal_version.map
> index d123602..1d88437 100644
> --- a/lib/librte_eal/rte_eal_version.map
> +++ b/lib/librte_eal/rte_eal_version.map
> @@ -223,8 +223,11 @@ EXPERIMENTAL {
> rte_eal_mbuf_user_pool_ops;
> rte_mp_action_register;
> rte_mp_action_unregister;
> + rte_mp_async_reply_register;
> + rte_mp_async_reply_unregister;
> rte_mp_sendmsg;
> rte_mp_request;
> + rte_mp_request_async;
> rte_mp_reply;
> rte_service_attr_get;
> rte_service_attr_reset_all;
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v4] eal: add asynchronous request API to DPDK IPC
2018-03-23 15:38 ` Tan, Jianfeng
@ 2018-03-23 18:21 ` Burakov, Anatoly
2018-03-24 13:22 ` Burakov, Anatoly
0 siblings, 1 reply; 40+ messages in thread
From: Burakov, Anatoly @ 2018-03-23 18:21 UTC (permalink / raw)
To: Tan, Jianfeng, dev; +Cc: konstantin.ananyev
On 23-Mar-18 3:38 PM, Tan, Jianfeng wrote:
> Hi Anatoly,
>
> Two general comments firstly.
>
> Q1. As I understand, malloc usage as an example, when a primary process
> receives a request in rte_mp_handle thread, it will allocate memory, and
> broadcast an async request to all secondary processes, and it will
> return immediately; then responses are replied from each secondary
> process, which are recorded at rte_mp_async_handle thread firstly;
> either timeout or responses are all collected, rte_mp_async_handle will
> trigger an async action.
>
> I agree the necessity of the async request API; without it, each caller
> who has similar requirement needs lots of code to implement it.
> But can we achieve that without creating another pthread by leveraging
> the alarm set? For example, to send an async request,
> step 1. set an alarm;
> step 2. send the request;
> step 3. receive and record responses in mp thread;
> step 4. if alarm timeout, trigger the async action with timeout result;
> step 5. if all responses are collected, cancel the alarm, and trigger
> the async action with success result.
That would work, but this "alarm" sounds like another thread. The main
thread is always blocked in recvmsg(), and interrupting that would
involve a signal, which is a recipe for disaster (you know, global
overwritable signal handlers, undefined behavior as to who actually gets
the signal, stuff like that).
That is, unless you were referring to DPDK's own "alarm" API, in which
case it's a chicken and egg problem - we want to use alarm for IPC, but
can't because rte_malloc relies on IPC, which relies on alarm, which
relies on rte_malloc. So unless we rewrite (or add an option for) alarm
API to not use rte_malloc'd memory, this isn't going to work.
We of course can do it, but really, i would be inclined to leave this as
is in the interests of merging this earlier, unless there are strong
objections to it.
>
> I don't have strong objection for adding another thread, and actually
> this is an internal thing in ipc system, we can optimize it later.
>
> Q2. Do we really need to register actions for result handler of async
> request? Instead, we can put it as a parameter into
> rte_mp_request_async(), whenever the request fails or succeeds, we just
> invoke the function.
That can be done, sure. This would look cleaner on the backend too, so
i'll probably do that for v5.
>
> Please see some other comments inline.
>
> On 3/14/2018 1:42 AM, Anatoly Burakov wrote:
>> This API is similar to the blocking API that is already present,
>> but reply will be received in a separate callback by the caller.
>>
>> Under the hood, we create a separate thread to deal with replies to
>> asynchronous requests, that will just wait to be notified by the
>> main thread, or woken up on a timer (it'll wake itself up every
>> minute regardless of whether it was called, but if there are no
>> requests in the queue, nothing will be done and it'll go to sleep
>> for another minute).
>
> Wait for 1min seems a little strange to me; it shall wake up for a
> latest timeout of pending async requests?
We do. However, we have to wait for *something* if there aren't any
asynchronous requests pending. There isn't a way to put "wait infinite
amount" as a time value, so i opted for next best thing - large enough
to not cause any performance issues. The timeout is arbitrary.
>> /** Double linked list of actions. */
>> @@ -60,13 +65,37 @@ struct mp_msg_internal {
>> struct rte_mp_msg msg;
>> };
>> +enum mp_request_type {
>> + REQUEST_TYPE_SYNC,
>> + REQUEST_TYPE_ASYNC
>> +};
>> +
>> +struct async_request_shared_param {
>> + struct rte_mp_reply *user_reply;
>> + struct timespec *end;
>
> Why we have to make these two as pointers? Operating on pointers
> introduce unnecessary complexity.
Because those are shared between different pending requests. Each
pending request gets its own entry in the queue (because it expects
answer from a particular process), but the request data (callback,
number of requests processed, etc.) is shared between all requests for
this sync operation. We don't have the luxury of storing all of that in
a local variable like we do with synchronous requests :)
>
>> + int n_requests_processed;
>
> It sounds like recording how many requests are sent, but it means how
> many responses are received, right? n_responses sounds better?
No, it doesn't. Well, perhaps "n_requests_processed" should be
"n_responses_processed", but not "n_responses", because this is a value
that is indicating how many responses we've already seen (to know when
to trigger the callback).
>
>> +};
>> +
>> +struct async_request_param {
>> + struct async_request_shared_param *param;
>> +};
>> +
>> +struct sync_request_param {
>> + pthread_cond_t cond;
>> +};
>> +
>> struct sync_request {
>
> I know "sync_" in the original version was for synchronization between
> the blocked thread (who sends the request) and the mp thread.
>
> But it indeed makes the code difficult to understand. We may change the
> name to "pending_request"?
This can be done, sure.
>
>
>> TAILQ_ENTRY(sync_request) next;
>> - int reply_received;
>> + enum mp_request_type type;
>> char dst[PATH_MAX];
>> struct rte_mp_msg *request;
>> - struct rte_mp_msg *reply;
>> - pthread_cond_t cond;
>> + struct rte_mp_msg *reply_msg;
>> + int reply_received;
>> + RTE_STD_C11
>> + union {
>> + struct sync_request_param sync;
>> + struct async_request_param async;
>> + };
>> };
>
> Too many structs are defined? How about just putting it like this:
>
> struct pending_request {
> TAILQ_ENTRY(sync_request) next;
> enum mp_request_type type;
> char dst[PATH_MAX];
> struct rte_mp_msg *request;
> struct rte_mp_msg *reply_msg;
> int reply_received;
> RTE_STD_C11
> union {
> /* for sync request */
> struct {
> pthread_cond_t cond; /* used for mp thread to
> wake up requesting thread */
> };
>
> /* for async request */
> struct {
> struct rte_mp_reply user_reply;
> struct timespec end;
> int n_requests_processed; /* store how requests */
> };
> };
> };
That can work, sure. However, i actually think that my approach is
clearer, because when you're working with autocomplete and a proper IDE,
it's clear which values are for which case (and i would argue it makes
the code more readable as well).
>> +static void *
>> +async_reply_handle(void *arg __rte_unused)
>> +{
>> + struct sync_request *sr;
>> + struct timeval now;
>> + struct timespec timeout, ts_now;
>> + do {
>
> Put while(1) here so that it's clear to be an infinite loop.
Will do.
>> + /* we have to lock the request queue here, as we will be adding a
>> bunch
>> + * of requests to the queue at once, and some of the replies may
>> arrive
>> + * before we add all of the requests to the queue.
>> + */
>> + pthread_mutex_lock(&sync_requests.lock);
>
> Why do we share this lock for both sync and async requests?
Because sync and async requests share the queue.
>> + * Asynchronous reply function typedef used by other components.
>> + *
>> + * As we create socket channel for primary/secondary communication, use
>
> Here are two spaces.
Will fix.
--
Thanks,
Anatoly
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v4] eal: add asynchronous request API to DPDK IPC
2018-03-23 18:21 ` Burakov, Anatoly
@ 2018-03-24 13:22 ` Burakov, Anatoly
0 siblings, 0 replies; 40+ messages in thread
From: Burakov, Anatoly @ 2018-03-24 13:22 UTC (permalink / raw)
To: Tan, Jianfeng, dev; +Cc: konstantin.ananyev
A few of my yesterday's replies made no sense... Lesson learned: don't
reply to code review comments on a late friday evening :)
On 23-Mar-18 6:21 PM, Burakov, Anatoly wrote:
> On 23-Mar-18 3:38 PM, Tan, Jianfeng wrote:
>
> We do. However, we have to wait for *something* if there aren't any
> asynchronous requests pending. There isn't a way to put "wait infinite
> amount" as a time value, so i opted for next best thing - large enough
> to not cause any performance issues. The timeout is arbitrary.
>
Didn't realize we were holding the lock, so we could choose between wait
and timed wait. Fixed in v5.
>>> /** Double linked list of actions. */
>>> @@ -60,13 +65,37 @@ struct mp_msg_internal {
>>> struct rte_mp_msg msg;
>>> };
>>> +enum mp_request_type {
>>> + REQUEST_TYPE_SYNC,
>>> + REQUEST_TYPE_ASYNC
>>> +};
>>> +
>>> +struct async_request_shared_param {
>>> + struct rte_mp_reply *user_reply;
>>> + struct timespec *end;
>>
>> Why we have to make these two as pointers? Operating on pointers
>> introduce unnecessary complexity.
>
> Because those are shared between different pending requests. Each
> pending request gets its own entry in the queue (because it expects
> answer from a particular process), but the request data (callback,
> number of requests processed, etc.) is shared between all requests for
> this sync operation. We don't have the luxury of storing all of that in
> a local variable like we do with synchronous requests :)
Missed the fact that you weren't referring to the need of storing these
in shared_param but rather to the fact that i was storing
malloc-allocated values that are shared, as pointers in shared param
structure, when i could just as easily store actual structs there. Fixed
in v5.
>> Too many structs are defined? How about just putting it like this:
>>
>> struct pending_request {
>> TAILQ_ENTRY(sync_request) next;
>> enum mp_request_type type;
>> char dst[PATH_MAX];
>> struct rte_mp_msg *request;
>> struct rte_mp_msg *reply_msg;
>> int reply_received;
>> RTE_STD_C11
>> union {
>> /* for sync request */
>> struct {
>> pthread_cond_t cond; /* used for mp thread to
>> wake up requesting thread */
>> };
>>
>> /* for async request */
>> struct {
>> struct rte_mp_reply user_reply;
>> struct timespec end;
>> int n_requests_processed; /* store how
>> requests */
>> };
>> };
>> };
>
> That can work, sure. However, i actually think that my approach is
> clearer, because when you're working with autocomplete and a proper IDE,
> it's clear which values are for which case (and i would argue it makes
> the code more readable as well).
Again, didn't realize that you were referring to defining all of the
structs and values outside of pending request, rather than storing them
as anonymous structs (which would indeed cause problems with
autocomplete and readability). Fixed in v5.
Thanks again for your review!
--
Thanks,
Anatoly
^ permalink raw reply [flat|nested] 40+ messages in thread
* [dpdk-dev] [PATCH v5 1/2] eal: rename IPC sync request to pending request
2018-03-13 17:42 ` [dpdk-dev] [PATCH v4] " Anatoly Burakov
2018-03-23 15:38 ` Tan, Jianfeng
@ 2018-03-24 12:46 ` Anatoly Burakov
2018-03-26 7:31 ` Tan, Jianfeng
` (2 more replies)
2018-03-24 12:46 ` [dpdk-dev] [PATCH v5 " Anatoly Burakov
2 siblings, 3 replies; 40+ messages in thread
From: Anatoly Burakov @ 2018-03-24 12:46 UTC (permalink / raw)
To: dev; +Cc: failed, to, remove, 'kernel'
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Suggested-by: Jianfeng Tan <jianfeng.tan@intel.com>
---
lib/librte_eal/common/eal_common_proc.c | 38 ++++++++++++++++-----------------
1 file changed, 19 insertions(+), 19 deletions(-)
diff --git a/lib/librte_eal/common/eal_common_proc.c b/lib/librte_eal/common/eal_common_proc.c
index 4131b67..52b6ab2 100644
--- a/lib/librte_eal/common/eal_common_proc.c
+++ b/lib/librte_eal/common/eal_common_proc.c
@@ -60,8 +60,8 @@ struct mp_msg_internal {
struct rte_mp_msg msg;
};
-struct sync_request {
- TAILQ_ENTRY(sync_request) next;
+struct pending_request {
+ TAILQ_ENTRY(pending_request) next;
int reply_received;
char dst[PATH_MAX];
struct rte_mp_msg *request;
@@ -69,13 +69,13 @@ struct sync_request {
pthread_cond_t cond;
};
-TAILQ_HEAD(sync_request_list, sync_request);
+TAILQ_HEAD(pending_request_list, pending_request);
static struct {
- struct sync_request_list requests;
+ struct pending_request_list requests;
pthread_mutex_t lock;
-} sync_requests = {
- .requests = TAILQ_HEAD_INITIALIZER(sync_requests.requests),
+} pending_requests = {
+ .requests = TAILQ_HEAD_INITIALIZER(pending_requests.requests),
.lock = PTHREAD_MUTEX_INITIALIZER
};
@@ -84,12 +84,12 @@ static int
mp_send(struct rte_mp_msg *msg, const char *peer, int type);
-static struct sync_request *
+static struct pending_request *
find_sync_request(const char *dst, const char *act_name)
{
- struct sync_request *r;
+ struct pending_request *r;
- TAILQ_FOREACH(r, &sync_requests.requests, next) {
+ TAILQ_FOREACH(r, &pending_requests.requests, next) {
if (!strcmp(r->dst, dst) &&
!strcmp(r->request->name, act_name))
break;
@@ -259,7 +259,7 @@ read_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
static void
process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
{
- struct sync_request *sync_req;
+ struct pending_request *sync_req;
struct action_entry *entry;
struct rte_mp_msg *msg = &m->msg;
rte_mp_t action = NULL;
@@ -267,7 +267,7 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
RTE_LOG(DEBUG, EAL, "msg: %s\n", msg->name);
if (m->type == MP_REP || m->type == MP_IGN) {
- pthread_mutex_lock(&sync_requests.lock);
+ pthread_mutex_lock(&pending_requests.lock);
sync_req = find_sync_request(s->sun_path, msg->name);
if (sync_req) {
memcpy(sync_req->reply, msg, sizeof(*msg));
@@ -276,7 +276,7 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
pthread_cond_signal(&sync_req->cond);
} else
RTE_LOG(ERR, EAL, "Drop mp reply: %s\n", msg->name);
- pthread_mutex_unlock(&sync_requests.lock);
+ pthread_mutex_unlock(&pending_requests.lock);
return;
}
@@ -607,7 +607,7 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
{
int ret;
struct rte_mp_msg msg, *tmp;
- struct sync_request sync_req, *exist;
+ struct pending_request sync_req, *exist;
sync_req.reply_received = 0;
strcpy(sync_req.dst, dst);
@@ -615,14 +615,14 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
sync_req.reply = &msg;
pthread_cond_init(&sync_req.cond, NULL);
- pthread_mutex_lock(&sync_requests.lock);
+ pthread_mutex_lock(&pending_requests.lock);
exist = find_sync_request(dst, req->name);
if (!exist)
- TAILQ_INSERT_TAIL(&sync_requests.requests, &sync_req, next);
+ TAILQ_INSERT_TAIL(&pending_requests.requests, &sync_req, next);
if (exist) {
RTE_LOG(ERR, EAL, "A pending request %s:%s\n", dst, req->name);
rte_errno = EEXIST;
- pthread_mutex_unlock(&sync_requests.lock);
+ pthread_mutex_unlock(&pending_requests.lock);
return -1;
}
@@ -638,12 +638,12 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
do {
ret = pthread_cond_timedwait(&sync_req.cond,
- &sync_requests.lock, ts);
+ &pending_requests.lock, ts);
} while (ret != 0 && ret != ETIMEDOUT);
/* We got the lock now */
- TAILQ_REMOVE(&sync_requests.requests, &sync_req, next);
- pthread_mutex_unlock(&sync_requests.lock);
+ TAILQ_REMOVE(&pending_requests.requests, &sync_req, next);
+ pthread_mutex_unlock(&pending_requests.lock);
if (sync_req.reply_received == 0) {
RTE_LOG(ERR, EAL, "Fail to recv reply for request %s:%s\n",
--
2.7.4
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v5 1/2] eal: rename IPC sync request to pending request
2018-03-24 12:46 ` [dpdk-dev] [PATCH v5 1/2] eal: rename IPC sync request to pending request Anatoly Burakov
@ 2018-03-26 7:31 ` Tan, Jianfeng
2018-03-27 13:59 ` [dpdk-dev] [PATCH v6 " Anatoly Burakov
2018-03-27 13:59 ` [dpdk-dev] [PATCH v6 2/2] " Anatoly Burakov
2 siblings, 0 replies; 40+ messages in thread
From: Tan, Jianfeng @ 2018-03-26 7:31 UTC (permalink / raw)
To: Anatoly Burakov, dev; +Cc: konstantin.ananyev
On 3/24/2018 8:46 PM, Anatoly Burakov wrote:
> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
> Suggested-by: Jianfeng Tan <jianfeng.tan@intel.com>
Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
^ permalink raw reply [flat|nested] 40+ messages in thread
* [dpdk-dev] [PATCH v6 1/2] eal: rename IPC sync request to pending request
2018-03-24 12:46 ` [dpdk-dev] [PATCH v5 1/2] eal: rename IPC sync request to pending request Anatoly Burakov
2018-03-26 7:31 ` Tan, Jianfeng
@ 2018-03-27 13:59 ` Anatoly Burakov
2018-03-27 16:27 ` Thomas Monjalon
` (3 more replies)
2018-03-27 13:59 ` [dpdk-dev] [PATCH v6 2/2] " Anatoly Burakov
2 siblings, 4 replies; 40+ messages in thread
From: Anatoly Burakov @ 2018-03-27 13:59 UTC (permalink / raw)
To: dev; +Cc: failed, to, remove, 'kernel'
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Suggested-by: Jianfeng Tan <jianfeng.tan@intel.com>
Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
---
lib/librte_eal/common/eal_common_proc.c | 38 ++++++++++++++++-----------------
1 file changed, 19 insertions(+), 19 deletions(-)
diff --git a/lib/librte_eal/common/eal_common_proc.c b/lib/librte_eal/common/eal_common_proc.c
index 4131b67..52b6ab2 100644
--- a/lib/librte_eal/common/eal_common_proc.c
+++ b/lib/librte_eal/common/eal_common_proc.c
@@ -60,8 +60,8 @@ struct mp_msg_internal {
struct rte_mp_msg msg;
};
-struct sync_request {
- TAILQ_ENTRY(sync_request) next;
+struct pending_request {
+ TAILQ_ENTRY(pending_request) next;
int reply_received;
char dst[PATH_MAX];
struct rte_mp_msg *request;
@@ -69,13 +69,13 @@ struct sync_request {
pthread_cond_t cond;
};
-TAILQ_HEAD(sync_request_list, sync_request);
+TAILQ_HEAD(pending_request_list, pending_request);
static struct {
- struct sync_request_list requests;
+ struct pending_request_list requests;
pthread_mutex_t lock;
-} sync_requests = {
- .requests = TAILQ_HEAD_INITIALIZER(sync_requests.requests),
+} pending_requests = {
+ .requests = TAILQ_HEAD_INITIALIZER(pending_requests.requests),
.lock = PTHREAD_MUTEX_INITIALIZER
};
@@ -84,12 +84,12 @@ static int
mp_send(struct rte_mp_msg *msg, const char *peer, int type);
-static struct sync_request *
+static struct pending_request *
find_sync_request(const char *dst, const char *act_name)
{
- struct sync_request *r;
+ struct pending_request *r;
- TAILQ_FOREACH(r, &sync_requests.requests, next) {
+ TAILQ_FOREACH(r, &pending_requests.requests, next) {
if (!strcmp(r->dst, dst) &&
!strcmp(r->request->name, act_name))
break;
@@ -259,7 +259,7 @@ read_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
static void
process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
{
- struct sync_request *sync_req;
+ struct pending_request *sync_req;
struct action_entry *entry;
struct rte_mp_msg *msg = &m->msg;
rte_mp_t action = NULL;
@@ -267,7 +267,7 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
RTE_LOG(DEBUG, EAL, "msg: %s\n", msg->name);
if (m->type == MP_REP || m->type == MP_IGN) {
- pthread_mutex_lock(&sync_requests.lock);
+ pthread_mutex_lock(&pending_requests.lock);
sync_req = find_sync_request(s->sun_path, msg->name);
if (sync_req) {
memcpy(sync_req->reply, msg, sizeof(*msg));
@@ -276,7 +276,7 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
pthread_cond_signal(&sync_req->cond);
} else
RTE_LOG(ERR, EAL, "Drop mp reply: %s\n", msg->name);
- pthread_mutex_unlock(&sync_requests.lock);
+ pthread_mutex_unlock(&pending_requests.lock);
return;
}
@@ -607,7 +607,7 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
{
int ret;
struct rte_mp_msg msg, *tmp;
- struct sync_request sync_req, *exist;
+ struct pending_request sync_req, *exist;
sync_req.reply_received = 0;
strcpy(sync_req.dst, dst);
@@ -615,14 +615,14 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
sync_req.reply = &msg;
pthread_cond_init(&sync_req.cond, NULL);
- pthread_mutex_lock(&sync_requests.lock);
+ pthread_mutex_lock(&pending_requests.lock);
exist = find_sync_request(dst, req->name);
if (!exist)
- TAILQ_INSERT_TAIL(&sync_requests.requests, &sync_req, next);
+ TAILQ_INSERT_TAIL(&pending_requests.requests, &sync_req, next);
if (exist) {
RTE_LOG(ERR, EAL, "A pending request %s:%s\n", dst, req->name);
rte_errno = EEXIST;
- pthread_mutex_unlock(&sync_requests.lock);
+ pthread_mutex_unlock(&pending_requests.lock);
return -1;
}
@@ -638,12 +638,12 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
do {
ret = pthread_cond_timedwait(&sync_req.cond,
- &sync_requests.lock, ts);
+ &pending_requests.lock, ts);
} while (ret != 0 && ret != ETIMEDOUT);
/* We got the lock now */
- TAILQ_REMOVE(&sync_requests.requests, &sync_req, next);
- pthread_mutex_unlock(&sync_requests.lock);
+ TAILQ_REMOVE(&pending_requests.requests, &sync_req, next);
+ pthread_mutex_unlock(&pending_requests.lock);
if (sync_req.reply_received == 0) {
RTE_LOG(ERR, EAL, "Fail to recv reply for request %s:%s\n",
--
2.7.4
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 1/2] eal: rename IPC sync request to pending request
2018-03-27 13:59 ` [dpdk-dev] [PATCH v6 " Anatoly Burakov
@ 2018-03-27 16:27 ` Thomas Monjalon
2018-03-28 9:15 ` Burakov, Anatoly
2018-03-31 17:06 ` [dpdk-dev] [PATCH v7 1/3] " Anatoly Burakov
` (2 subsequent siblings)
3 siblings, 1 reply; 40+ messages in thread
From: Thomas Monjalon @ 2018-03-27 16:27 UTC (permalink / raw)
To: Anatoly Burakov; +Cc: dev, jianfeng.tan, konstantin.ananyev
27/03/2018 15:59, Anatoly Burakov:
> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
> Suggested-by: Jianfeng Tan <jianfeng.tan@intel.com>
> Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
You probably have a good explanation for this change.
Please explain in the commit message.
Generally speaking, every commits should have an explanation,
it is a basic quality rule.
Thanks
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 1/2] eal: rename IPC sync request to pending request
2018-03-27 16:27 ` Thomas Monjalon
@ 2018-03-28 9:15 ` Burakov, Anatoly
2018-03-28 10:08 ` Thomas Monjalon
0 siblings, 1 reply; 40+ messages in thread
From: Burakov, Anatoly @ 2018-03-28 9:15 UTC (permalink / raw)
To: Thomas Monjalon; +Cc: dev, jianfeng.tan, konstantin.ananyev
On 27-Mar-18 5:27 PM, Thomas Monjalon wrote:
> 27/03/2018 15:59, Anatoly Burakov:
>> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
>> Suggested-by: Jianfeng Tan <jianfeng.tan@intel.com>
>> Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
>
> You probably have a good explanation for this change.
> Please explain in the commit message.
>
> Generally speaking, every commits should have an explanation,
> it is a basic quality rule.
> Thanks
>
This is just to reduce noise in the following patch.
--
Thanks,
Anatoly
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 1/2] eal: rename IPC sync request to pending request
2018-03-28 9:15 ` Burakov, Anatoly
@ 2018-03-28 10:08 ` Thomas Monjalon
2018-03-28 10:57 ` Burakov, Anatoly
0 siblings, 1 reply; 40+ messages in thread
From: Thomas Monjalon @ 2018-03-28 10:08 UTC (permalink / raw)
To: Burakov, Anatoly; +Cc: dev, jianfeng.tan, konstantin.ananyev
28/03/2018 11:15, Burakov, Anatoly:
> On 27-Mar-18 5:27 PM, Thomas Monjalon wrote:
> > 27/03/2018 15:59, Anatoly Burakov:
> >> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
> >> Suggested-by: Jianfeng Tan <jianfeng.tan@intel.com>
> >> Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
> >
> > You probably have a good explanation for this change.
> > Please explain in the commit message.
> >
> > Generally speaking, every commits should have an explanation,
> > it is a basic quality rule.
> > Thanks
> >
>
> This is just to reduce noise in the following patch.
No Anatoly, you should explain why a rename is needed.
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 1/2] eal: rename IPC sync request to pending request
2018-03-28 10:08 ` Thomas Monjalon
@ 2018-03-28 10:57 ` Burakov, Anatoly
0 siblings, 0 replies; 40+ messages in thread
From: Burakov, Anatoly @ 2018-03-28 10:57 UTC (permalink / raw)
To: Thomas Monjalon; +Cc: dev, jianfeng.tan, konstantin.ananyev
On 28-Mar-18 11:08 AM, Thomas Monjalon wrote:
> 28/03/2018 11:15, Burakov, Anatoly:
>> On 27-Mar-18 5:27 PM, Thomas Monjalon wrote:
>>> 27/03/2018 15:59, Anatoly Burakov:
>>>> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
>>>> Suggested-by: Jianfeng Tan <jianfeng.tan@intel.com>
>>>> Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
>>>
>>> You probably have a good explanation for this change.
>>> Please explain in the commit message.
>>>
>>> Generally speaking, every commits should have an explanation,
>>> it is a basic quality rule.
>>> Thanks
>>>
>>
>> This is just to reduce noise in the following patch.
>
> No Anatoly, you should explain why a rename is needed.
>
I know :) I'm going to be submitting a v7 with function renames anyway,
so i'll fix this as well.
--
Thanks,
Anatoly
^ permalink raw reply [flat|nested] 40+ messages in thread
* [dpdk-dev] [PATCH v7 1/3] eal: rename IPC sync request to pending request
2018-03-27 13:59 ` [dpdk-dev] [PATCH v6 " Anatoly Burakov
2018-03-27 16:27 ` Thomas Monjalon
@ 2018-03-31 17:06 ` Anatoly Burakov
2018-03-31 17:06 ` [dpdk-dev] [PATCH v7 2/3] eal: rename mp_request to mp_request_sync Anatoly Burakov
2018-03-31 17:06 ` [dpdk-dev] [PATCH v7 3/3] eal: add asynchronous request API to DPDK IPC Anatoly Burakov
3 siblings, 0 replies; 40+ messages in thread
From: Anatoly Burakov @ 2018-03-31 17:06 UTC (permalink / raw)
To: dev; +Cc: jianfeng.tan, konstantin.ananyev, thomas
Originally, there was only one type of request which was used
for multiprocess synchronization (hence the name - sync request).
However, now that we are going to have two types of requests,
synchronous and asynchronous, having it named "sync request" is
very confusing, so we will rename it to "pending request". This
is internal-only, so no externally visible API changes.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Suggested-by: Jianfeng Tan <jianfeng.tan@intel.com>
Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
---
Notes:
v7:
- Provide explanation as to why this change is being made
lib/librte_eal/common/eal_common_proc.c | 38 ++++++++++++++++-----------------
1 file changed, 19 insertions(+), 19 deletions(-)
diff --git a/lib/librte_eal/common/eal_common_proc.c b/lib/librte_eal/common/eal_common_proc.c
index 4131b67..52b6ab2 100644
--- a/lib/librte_eal/common/eal_common_proc.c
+++ b/lib/librte_eal/common/eal_common_proc.c
@@ -60,8 +60,8 @@ struct mp_msg_internal {
struct rte_mp_msg msg;
};
-struct sync_request {
- TAILQ_ENTRY(sync_request) next;
+struct pending_request {
+ TAILQ_ENTRY(pending_request) next;
int reply_received;
char dst[PATH_MAX];
struct rte_mp_msg *request;
@@ -69,13 +69,13 @@ struct sync_request {
pthread_cond_t cond;
};
-TAILQ_HEAD(sync_request_list, sync_request);
+TAILQ_HEAD(pending_request_list, pending_request);
static struct {
- struct sync_request_list requests;
+ struct pending_request_list requests;
pthread_mutex_t lock;
-} sync_requests = {
- .requests = TAILQ_HEAD_INITIALIZER(sync_requests.requests),
+} pending_requests = {
+ .requests = TAILQ_HEAD_INITIALIZER(pending_requests.requests),
.lock = PTHREAD_MUTEX_INITIALIZER
};
@@ -84,12 +84,12 @@ static int
mp_send(struct rte_mp_msg *msg, const char *peer, int type);
-static struct sync_request *
+static struct pending_request *
find_sync_request(const char *dst, const char *act_name)
{
- struct sync_request *r;
+ struct pending_request *r;
- TAILQ_FOREACH(r, &sync_requests.requests, next) {
+ TAILQ_FOREACH(r, &pending_requests.requests, next) {
if (!strcmp(r->dst, dst) &&
!strcmp(r->request->name, act_name))
break;
@@ -259,7 +259,7 @@ read_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
static void
process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
{
- struct sync_request *sync_req;
+ struct pending_request *sync_req;
struct action_entry *entry;
struct rte_mp_msg *msg = &m->msg;
rte_mp_t action = NULL;
@@ -267,7 +267,7 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
RTE_LOG(DEBUG, EAL, "msg: %s\n", msg->name);
if (m->type == MP_REP || m->type == MP_IGN) {
- pthread_mutex_lock(&sync_requests.lock);
+ pthread_mutex_lock(&pending_requests.lock);
sync_req = find_sync_request(s->sun_path, msg->name);
if (sync_req) {
memcpy(sync_req->reply, msg, sizeof(*msg));
@@ -276,7 +276,7 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
pthread_cond_signal(&sync_req->cond);
} else
RTE_LOG(ERR, EAL, "Drop mp reply: %s\n", msg->name);
- pthread_mutex_unlock(&sync_requests.lock);
+ pthread_mutex_unlock(&pending_requests.lock);
return;
}
@@ -607,7 +607,7 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
{
int ret;
struct rte_mp_msg msg, *tmp;
- struct sync_request sync_req, *exist;
+ struct pending_request sync_req, *exist;
sync_req.reply_received = 0;
strcpy(sync_req.dst, dst);
@@ -615,14 +615,14 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
sync_req.reply = &msg;
pthread_cond_init(&sync_req.cond, NULL);
- pthread_mutex_lock(&sync_requests.lock);
+ pthread_mutex_lock(&pending_requests.lock);
exist = find_sync_request(dst, req->name);
if (!exist)
- TAILQ_INSERT_TAIL(&sync_requests.requests, &sync_req, next);
+ TAILQ_INSERT_TAIL(&pending_requests.requests, &sync_req, next);
if (exist) {
RTE_LOG(ERR, EAL, "A pending request %s:%s\n", dst, req->name);
rte_errno = EEXIST;
- pthread_mutex_unlock(&sync_requests.lock);
+ pthread_mutex_unlock(&pending_requests.lock);
return -1;
}
@@ -638,12 +638,12 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
do {
ret = pthread_cond_timedwait(&sync_req.cond,
- &sync_requests.lock, ts);
+ &pending_requests.lock, ts);
} while (ret != 0 && ret != ETIMEDOUT);
/* We got the lock now */
- TAILQ_REMOVE(&sync_requests.requests, &sync_req, next);
- pthread_mutex_unlock(&sync_requests.lock);
+ TAILQ_REMOVE(&pending_requests.requests, &sync_req, next);
+ pthread_mutex_unlock(&pending_requests.lock);
if (sync_req.reply_received == 0) {
RTE_LOG(ERR, EAL, "Fail to recv reply for request %s:%s\n",
--
2.7.4
^ permalink raw reply [flat|nested] 40+ messages in thread
* [dpdk-dev] [PATCH v7 2/3] eal: rename mp_request to mp_request_sync
2018-03-27 13:59 ` [dpdk-dev] [PATCH v6 " Anatoly Burakov
2018-03-27 16:27 ` Thomas Monjalon
2018-03-31 17:06 ` [dpdk-dev] [PATCH v7 1/3] " Anatoly Burakov
@ 2018-03-31 17:06 ` Anatoly Burakov
2018-04-02 5:09 ` Tan, Jianfeng
2018-03-31 17:06 ` [dpdk-dev] [PATCH v7 3/3] eal: add asynchronous request API to DPDK IPC Anatoly Burakov
3 siblings, 1 reply; 40+ messages in thread
From: Anatoly Burakov @ 2018-03-31 17:06 UTC (permalink / raw)
To: dev; +Cc: jianfeng.tan, konstantin.ananyev, thomas
Rename rte_mp_request to rte_mp_request_sync to indicate
that this request will be done synchronously (as opposed to
asynchronous request, which comes in next patch).
Also, fix alphabetical ordering for .map file.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Suggested-by: Thomas Monjalon <thomas@monjalon.net>
---
Notes:
v7:
- Added this patch
lib/librte_eal/common/eal_common_proc.c | 2 +-
lib/librte_eal/common/include/rte_eal.h | 2 +-
lib/librte_eal/rte_eal_version.map | 4 ++--
3 files changed, 4 insertions(+), 4 deletions(-)
diff --git a/lib/librte_eal/common/eal_common_proc.c b/lib/librte_eal/common/eal_common_proc.c
index 52b6ab2..b704f5a 100644
--- a/lib/librte_eal/common/eal_common_proc.c
+++ b/lib/librte_eal/common/eal_common_proc.c
@@ -674,7 +674,7 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
}
int __rte_experimental
-rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
+rte_mp_request_sync(struct rte_mp_msg *req, struct rte_mp_reply *reply,
const struct timespec *ts)
{
int dir_fd, ret = 0;
diff --git a/lib/librte_eal/common/include/rte_eal.h b/lib/librte_eal/common/include/rte_eal.h
index 044474e..d1cc89e 100644
--- a/lib/librte_eal/common/include/rte_eal.h
+++ b/lib/librte_eal/common/include/rte_eal.h
@@ -314,7 +314,7 @@ rte_mp_sendmsg(struct rte_mp_msg *msg);
* - On failure, return -1, and the reason will be stored in rte_errno.
*/
int __rte_experimental
-rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
+rte_mp_request_sync(struct rte_mp_msg *req, struct rte_mp_reply *reply,
const struct timespec *ts);
/**
diff --git a/lib/librte_eal/rte_eal_version.map b/lib/librte_eal/rte_eal_version.map
index d123602..7b66c73 100644
--- a/lib/librte_eal/rte_eal_version.map
+++ b/lib/librte_eal/rte_eal_version.map
@@ -223,9 +223,9 @@ EXPERIMENTAL {
rte_eal_mbuf_user_pool_ops;
rte_mp_action_register;
rte_mp_action_unregister;
- rte_mp_sendmsg;
- rte_mp_request;
rte_mp_reply;
+ rte_mp_request_sync;
+ rte_mp_sendmsg;
rte_service_attr_get;
rte_service_attr_reset_all;
rte_service_component_register;
--
2.7.4
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v7 2/3] eal: rename mp_request to mp_request_sync
2018-03-31 17:06 ` [dpdk-dev] [PATCH v7 2/3] eal: rename mp_request to mp_request_sync Anatoly Burakov
@ 2018-04-02 5:09 ` Tan, Jianfeng
0 siblings, 0 replies; 40+ messages in thread
From: Tan, Jianfeng @ 2018-04-02 5:09 UTC (permalink / raw)
To: Burakov, Anatoly, dev; +Cc: Ananyev, Konstantin, thomas
> -----Original Message-----
> From: Burakov, Anatoly
> Sent: Sunday, April 1, 2018 1:06 AM
> To: dev@dpdk.org
> Cc: Tan, Jianfeng; Ananyev, Konstantin; thomas@monjalon.net
> Subject: [PATCH v7 2/3] eal: rename mp_request to mp_request_sync
>
> Rename rte_mp_request to rte_mp_request_sync to indicate
> that this request will be done synchronously (as opposed to
> asynchronous request, which comes in next patch).
>
> Also, fix alphabetical ordering for .map file.
>
> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
> Suggested-by: Thomas Monjalon <thomas@monjalon.net>
Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
Thanks!
> ---
>
> Notes:
> v7:
> - Added this patch
>
> lib/librte_eal/common/eal_common_proc.c | 2 +-
> lib/librte_eal/common/include/rte_eal.h | 2 +-
> lib/librte_eal/rte_eal_version.map | 4 ++--
> 3 files changed, 4 insertions(+), 4 deletions(-)
>
> diff --git a/lib/librte_eal/common/eal_common_proc.c
> b/lib/librte_eal/common/eal_common_proc.c
> index 52b6ab2..b704f5a 100644
> --- a/lib/librte_eal/common/eal_common_proc.c
> +++ b/lib/librte_eal/common/eal_common_proc.c
> @@ -674,7 +674,7 @@ mp_request_one(const char *dst, struct rte_mp_msg
> *req,
> }
>
> int __rte_experimental
> -rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
> +rte_mp_request_sync(struct rte_mp_msg *req, struct rte_mp_reply
> *reply,
> const struct timespec *ts)
> {
> int dir_fd, ret = 0;
> diff --git a/lib/librte_eal/common/include/rte_eal.h
> b/lib/librte_eal/common/include/rte_eal.h
> index 044474e..d1cc89e 100644
> --- a/lib/librte_eal/common/include/rte_eal.h
> +++ b/lib/librte_eal/common/include/rte_eal.h
> @@ -314,7 +314,7 @@ rte_mp_sendmsg(struct rte_mp_msg *msg);
> * - On failure, return -1, and the reason will be stored in rte_errno.
> */
> int __rte_experimental
> -rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
> +rte_mp_request_sync(struct rte_mp_msg *req, struct rte_mp_reply
> *reply,
> const struct timespec *ts);
>
> /**
> diff --git a/lib/librte_eal/rte_eal_version.map
> b/lib/librte_eal/rte_eal_version.map
> index d123602..7b66c73 100644
> --- a/lib/librte_eal/rte_eal_version.map
> +++ b/lib/librte_eal/rte_eal_version.map
> @@ -223,9 +223,9 @@ EXPERIMENTAL {
> rte_eal_mbuf_user_pool_ops;
> rte_mp_action_register;
> rte_mp_action_unregister;
> - rte_mp_sendmsg;
> - rte_mp_request;
> rte_mp_reply;
> + rte_mp_request_sync;
> + rte_mp_sendmsg;
> rte_service_attr_get;
> rte_service_attr_reset_all;
> rte_service_component_register;
> --
> 2.7.4
^ permalink raw reply [flat|nested] 40+ messages in thread
* [dpdk-dev] [PATCH v7 3/3] eal: add asynchronous request API to DPDK IPC
2018-03-27 13:59 ` [dpdk-dev] [PATCH v6 " Anatoly Burakov
` (2 preceding siblings ...)
2018-03-31 17:06 ` [dpdk-dev] [PATCH v7 2/3] eal: rename mp_request to mp_request_sync Anatoly Burakov
@ 2018-03-31 17:06 ` Anatoly Burakov
2018-04-04 22:15 ` Thomas Monjalon
3 siblings, 1 reply; 40+ messages in thread
From: Anatoly Burakov @ 2018-03-31 17:06 UTC (permalink / raw)
To: dev; +Cc: jianfeng.tan, konstantin.ananyev, thomas
This API is similar to the blocking API that is already present,
but reply will be received in a separate callback by the caller
(callback specified at the time of request, rather than registering
for it in advance).
Under the hood, we create a separate thread to deal with replies to
asynchronous requests, that will just wait to be notified by the
main thread, or woken up on a timer.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
---
Notes:
v6:
- address review comments from Jianfeng
- do not add dummy item to queue by default
v5:
- addressed review comments from Jianfeng
- split into two patches to avoid rename noise
- do not mark ignored message as processed
v4:
- rebase on top of latest IPC Improvements patchset [2]
v3:
- added support for MP_IGN messages introduced in
IPC improvements v5 patchset
v2:
- fixed deadlocks and race conditions by not calling
callbacks while iterating over sync request list
- fixed use-after-free by making a copy of request
- changed API to also give user a copy of original
request, so that they know to which message the
callback is a reply to
- fixed missing .map file entries
This patch is dependent upon previously published patchsets
for IPC fixes [1] and improvements [2].
rte_mp_action_unregister and rte_mp_async_reply_unregister
do the same thing - should we perhaps make it one function?
[1] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Fixes/
[2] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Improvements/
lib/librte_eal/common/eal_common_proc.c | 458 +++++++++++++++++++++++++++++++-
lib/librte_eal/common/include/rte_eal.h | 36 +++
lib/librte_eal/rte_eal_version.map | 1 +
3 files changed, 482 insertions(+), 13 deletions(-)
diff --git a/lib/librte_eal/common/eal_common_proc.c b/lib/librte_eal/common/eal_common_proc.c
index b704f5a..f98622f 100644
--- a/lib/librte_eal/common/eal_common_proc.c
+++ b/lib/librte_eal/common/eal_common_proc.c
@@ -26,6 +26,7 @@
#include <rte_errno.h>
#include <rte_lcore.h>
#include <rte_log.h>
+#include <rte_tailq.h>
#include "eal_private.h"
#include "eal_filesystem.h"
@@ -60,13 +61,32 @@ struct mp_msg_internal {
struct rte_mp_msg msg;
};
+struct async_request_param {
+ rte_mp_async_reply_t clb;
+ struct rte_mp_reply user_reply;
+ struct timespec end;
+ int n_responses_processed;
+};
+
struct pending_request {
TAILQ_ENTRY(pending_request) next;
- int reply_received;
+ enum {
+ REQUEST_TYPE_SYNC,
+ REQUEST_TYPE_ASYNC
+ } type;
char dst[PATH_MAX];
struct rte_mp_msg *request;
struct rte_mp_msg *reply;
- pthread_cond_t cond;
+ int reply_received;
+ RTE_STD_C11
+ union {
+ struct {
+ struct async_request_param *param;
+ } async;
+ struct {
+ pthread_cond_t cond;
+ } sync;
+ };
};
TAILQ_HEAD(pending_request_list, pending_request);
@@ -74,9 +94,12 @@ TAILQ_HEAD(pending_request_list, pending_request);
static struct {
struct pending_request_list requests;
pthread_mutex_t lock;
+ pthread_cond_t async_cond;
} pending_requests = {
.requests = TAILQ_HEAD_INITIALIZER(pending_requests.requests),
- .lock = PTHREAD_MUTEX_INITIALIZER
+ .lock = PTHREAD_MUTEX_INITIALIZER,
+ .async_cond = PTHREAD_COND_INITIALIZER
+ /**< used in async requests only */
};
/* forward declarations */
@@ -273,7 +296,12 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
memcpy(sync_req->reply, msg, sizeof(*msg));
/* -1 indicates that we've been asked to ignore */
sync_req->reply_received = m->type == MP_REP ? 1 : -1;
- pthread_cond_signal(&sync_req->cond);
+
+ if (sync_req->type == REQUEST_TYPE_SYNC)
+ pthread_cond_signal(&sync_req->sync.cond);
+ else if (sync_req->type == REQUEST_TYPE_ASYNC)
+ pthread_cond_signal(
+ &pending_requests.async_cond);
} else
RTE_LOG(ERR, EAL, "Drop mp reply: %s\n", msg->name);
pthread_mutex_unlock(&pending_requests.lock);
@@ -320,6 +348,189 @@ mp_handle(void *arg __rte_unused)
}
static int
+timespec_cmp(const struct timespec *a, const struct timespec *b)
+{
+ if (a->tv_sec < b->tv_sec)
+ return -1;
+ if (a->tv_sec > b->tv_sec)
+ return 1;
+ if (a->tv_nsec < b->tv_nsec)
+ return -1;
+ if (a->tv_nsec > b->tv_nsec)
+ return 1;
+ return 0;
+}
+
+enum async_action {
+ ACTION_NONE, /**< don't do anything */
+ ACTION_FREE, /**< free the action entry, but don't trigger callback */
+ ACTION_TRIGGER /**< trigger callback, then free action entry */
+};
+
+static enum async_action
+process_async_request(struct pending_request *sr, const struct timespec *now)
+{
+ struct async_request_param *param;
+ struct rte_mp_reply *reply;
+ bool timeout, received, last_msg;
+
+ param = sr->async.param;
+ reply = ¶m->user_reply;
+
+ /* did we timeout? */
+ timeout = timespec_cmp(¶m->end, now) <= 0;
+
+ /* did we receive a response? */
+ received = sr->reply_received != 0;
+
+ /* if we didn't time out, and we didn't receive a response, ignore */
+ if (!timeout && !received)
+ return ACTION_NONE;
+
+ /* if we received a response, adjust relevant data and copy mesasge. */
+ if (sr->reply_received == 1 && sr->reply) {
+ struct rte_mp_msg *msg, *user_msgs, *tmp;
+
+ msg = sr->reply;
+ user_msgs = reply->msgs;
+
+ tmp = realloc(user_msgs, sizeof(*msg) *
+ (reply->nb_received + 1));
+ if (!tmp) {
+ RTE_LOG(ERR, EAL, "Fail to alloc reply for request %s:%s\n",
+ sr->dst, sr->request->name);
+ /* this entry is going to be removed and its message
+ * dropped, but we don't want to leak memory, so
+ * continue.
+ */
+ } else {
+ user_msgs = tmp;
+ reply->msgs = user_msgs;
+ memcpy(&user_msgs[reply->nb_received],
+ msg, sizeof(*msg));
+ reply->nb_received++;
+ }
+
+ /* mark this request as processed */
+ param->n_responses_processed++;
+ } else if (sr->reply_received == -1) {
+ /* we were asked to ignore this process */
+ reply->nb_sent--;
+ }
+ free(sr->reply);
+
+ last_msg = param->n_responses_processed == reply->nb_sent;
+
+ return last_msg ? ACTION_TRIGGER : ACTION_FREE;
+}
+
+static void
+trigger_async_action(struct pending_request *sr)
+{
+ struct async_request_param *param;
+ struct rte_mp_reply *reply;
+
+ param = sr->async.param;
+ reply = ¶m->user_reply;
+
+ param->clb(sr->request, reply);
+
+ /* clean up */
+ free(sr->async.param->user_reply.msgs);
+ free(sr->async.param);
+ free(sr->request);
+}
+
+static void *
+async_reply_handle(void *arg __rte_unused)
+{
+ struct pending_request *sr;
+ struct timeval now;
+ struct timespec timeout, ts_now;
+ while (1) {
+ struct pending_request *trigger = NULL;
+ int ret;
+ bool nowait = false;
+ bool timedwait = false;
+
+ pthread_mutex_lock(&pending_requests.lock);
+
+ /* scan through the list and see if there are any timeouts that
+ * are earlier than our current timeout.
+ */
+ TAILQ_FOREACH(sr, &pending_requests.requests, next) {
+ if (sr->type != REQUEST_TYPE_ASYNC)
+ continue;
+ if (!timedwait || timespec_cmp(&sr->async.param->end,
+ &timeout) < 0) {
+ memcpy(&timeout, &sr->async.param->end,
+ sizeof(timeout));
+ timedwait = true;
+ }
+
+ /* sometimes, we don't even wait */
+ if (sr->reply_received) {
+ nowait = true;
+ break;
+ }
+ }
+
+ if (nowait)
+ ret = 0;
+ else if (timedwait)
+ ret = pthread_cond_timedwait(
+ &pending_requests.async_cond,
+ &pending_requests.lock, &timeout);
+ else
+ ret = pthread_cond_wait(&pending_requests.async_cond,
+ &pending_requests.lock);
+
+ if (gettimeofday(&now, NULL) < 0) {
+ RTE_LOG(ERR, EAL, "Cannot get current time\n");
+ break;
+ }
+ ts_now.tv_nsec = now.tv_usec * 1000;
+ ts_now.tv_sec = now.tv_sec;
+
+ if (ret == 0 || ret == ETIMEDOUT) {
+ struct pending_request *next;
+ /* we've either been woken up, or we timed out */
+
+ /* we have still the lock, check if anything needs
+ * processing.
+ */
+ TAILQ_FOREACH_SAFE(sr, &pending_requests.requests, next,
+ next) {
+ enum async_action action;
+ if (sr->type != REQUEST_TYPE_ASYNC)
+ continue;
+
+ action = process_async_request(sr, &ts_now);
+ if (action == ACTION_FREE) {
+ TAILQ_REMOVE(&pending_requests.requests,
+ sr, next);
+ free(sr);
+ } else if (action == ACTION_TRIGGER &&
+ trigger == NULL) {
+ TAILQ_REMOVE(&pending_requests.requests,
+ sr, next);
+ trigger = sr;
+ }
+ }
+ }
+ pthread_mutex_unlock(&pending_requests.lock);
+ if (trigger) {
+ trigger_async_action(trigger);
+ free(trigger);
+ }
+ };
+
+ RTE_LOG(ERR, EAL, "ERROR: asynchronous requests disabled\n");
+
+ return NULL;
+}
+
+static int
open_socket_fd(void)
{
char peer_name[PATH_MAX] = {0};
@@ -382,7 +593,7 @@ rte_mp_channel_init(void)
char thread_name[RTE_MAX_THREAD_NAME_LEN];
char path[PATH_MAX];
int dir_fd;
- pthread_t tid;
+ pthread_t mp_handle_tid, async_reply_handle_tid;
/* create filter path */
create_socket_path("*", path, sizeof(path));
@@ -419,7 +630,16 @@ rte_mp_channel_init(void)
return -1;
}
- if (pthread_create(&tid, NULL, mp_handle, NULL) < 0) {
+ if (pthread_create(&mp_handle_tid, NULL, mp_handle, NULL) < 0) {
+ RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
+ strerror(errno));
+ close(mp_fd);
+ mp_fd = -1;
+ return -1;
+ }
+
+ if (pthread_create(&async_reply_handle_tid, NULL,
+ async_reply_handle, NULL) < 0) {
RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
strerror(errno));
close(mp_fd);
@@ -430,7 +650,11 @@ rte_mp_channel_init(void)
/* try best to set thread name */
snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_handle");
- rte_thread_setname(tid, thread_name);
+ rte_thread_setname(mp_handle_tid, thread_name);
+
+ /* try best to set thread name */
+ snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_async_handle");
+ rte_thread_setname(async_reply_handle_tid, thread_name);
/* unlock the directory */
flock(dir_fd, LOCK_UN);
@@ -602,18 +826,77 @@ rte_mp_sendmsg(struct rte_mp_msg *msg)
}
static int
-mp_request_one(const char *dst, struct rte_mp_msg *req,
+mp_request_async(const char *dst, struct rte_mp_msg *req,
+ struct async_request_param *param)
+{
+ struct rte_mp_msg *reply_msg;
+ struct pending_request *sync_req, *exist;
+ int ret;
+
+ sync_req = malloc(sizeof(*sync_req));
+ reply_msg = malloc(sizeof(*reply_msg));
+ if (sync_req == NULL || reply_msg == NULL) {
+ RTE_LOG(ERR, EAL, "Could not allocate space for sync request\n");
+ rte_errno = ENOMEM;
+ ret = -1;
+ goto fail;
+ }
+
+ memset(sync_req, 0, sizeof(*sync_req));
+ memset(reply_msg, 0, sizeof(*reply_msg));
+
+ sync_req->type = REQUEST_TYPE_ASYNC;
+ strcpy(sync_req->dst, dst);
+ sync_req->request = req;
+ sync_req->reply = reply_msg;
+ sync_req->async.param = param;
+
+ /* queue already locked by caller */
+
+ exist = find_sync_request(dst, req->name);
+ if (!exist) {
+ TAILQ_INSERT_TAIL(&pending_requests.requests, sync_req, next);
+ } else {
+ RTE_LOG(ERR, EAL, "A pending request %s:%s\n", dst, req->name);
+ rte_errno = EEXIST;
+ ret = -1;
+ goto fail;
+ }
+
+ ret = send_msg(dst, req, MP_REQ);
+ if (ret < 0) {
+ RTE_LOG(ERR, EAL, "Fail to send request %s:%s\n",
+ dst, req->name);
+ ret = -1;
+ goto fail;
+ } else if (ret == 0) {
+ ret = 0;
+ goto fail;
+ }
+
+ param->user_reply.nb_sent++;
+
+ return 0;
+fail:
+ free(sync_req);
+ free(reply_msg);
+ return ret;
+}
+
+static int
+mp_request_sync(const char *dst, struct rte_mp_msg *req,
struct rte_mp_reply *reply, const struct timespec *ts)
{
int ret;
struct rte_mp_msg msg, *tmp;
struct pending_request sync_req, *exist;
+ sync_req.type = REQUEST_TYPE_SYNC;
sync_req.reply_received = 0;
strcpy(sync_req.dst, dst);
sync_req.request = req;
sync_req.reply = &msg;
- pthread_cond_init(&sync_req.cond, NULL);
+ pthread_cond_init(&sync_req.sync.cond, NULL);
pthread_mutex_lock(&pending_requests.lock);
exist = find_sync_request(dst, req->name);
@@ -637,7 +920,7 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
reply->nb_sent++;
do {
- ret = pthread_cond_timedwait(&sync_req.cond,
+ ret = pthread_cond_timedwait(&sync_req.sync.cond,
&pending_requests.lock, ts);
} while (ret != 0 && ret != ETIMEDOUT);
@@ -703,7 +986,7 @@ rte_mp_request_sync(struct rte_mp_msg *req, struct rte_mp_reply *reply,
/* for secondary process, send request to the primary process only */
if (rte_eal_process_type() == RTE_PROC_SECONDARY)
- return mp_request_one(eal_mp_socket_path(), req, reply, &end);
+ return mp_request_sync(eal_mp_socket_path(), req, reply, &end);
/* for primary process, broadcast request, and collect reply 1 by 1 */
mp_dir = opendir(mp_dir_path);
@@ -732,7 +1015,7 @@ rte_mp_request_sync(struct rte_mp_msg *req, struct rte_mp_reply *reply,
snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
ent->d_name);
- if (mp_request_one(path, req, reply, &end))
+ if (mp_request_sync(path, req, reply, &end))
ret = -1;
}
/* unlock the directory */
@@ -744,9 +1027,158 @@ rte_mp_request_sync(struct rte_mp_msg *req, struct rte_mp_reply *reply,
}
int __rte_experimental
-rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
+rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts,
+ rte_mp_async_reply_t clb)
{
+ struct rte_mp_msg *copy;
+ struct pending_request *dummy;
+ struct async_request_param *param;
+ struct rte_mp_reply *reply;
+ int dir_fd, ret = 0;
+ DIR *mp_dir;
+ struct dirent *ent;
+ struct timeval now;
+ struct timespec *end;
+ bool dummy_used = false;
+
+ RTE_LOG(DEBUG, EAL, "request: %s\n", req->name);
+
+ if (check_input(req) == false)
+ return -1;
+ if (gettimeofday(&now, NULL) < 0) {
+ RTE_LOG(ERR, EAL, "Faile to get current time\n");
+ rte_errno = errno;
+ return -1;
+ }
+ copy = malloc(sizeof(*copy));
+ dummy = malloc(sizeof(*dummy));
+ param = malloc(sizeof(*param));
+ if (copy == NULL || dummy == NULL || param == NULL) {
+ RTE_LOG(ERR, EAL, "Failed to allocate memory for async reply\n");
+ rte_errno = ENOMEM;
+ goto fail;
+ }
+
+ memset(copy, 0, sizeof(*copy));
+ memset(dummy, 0, sizeof(*dummy));
+ memset(param, 0, sizeof(*param));
+
+ /* copy message */
+ memcpy(copy, req, sizeof(*copy));
+
+ param->n_responses_processed = 0;
+ param->clb = clb;
+ end = ¶m->end;
+ reply = ¶m->user_reply;
+
+ end->tv_nsec = (now.tv_usec * 1000 + ts->tv_nsec) % 1000000000;
+ end->tv_sec = now.tv_sec + ts->tv_sec +
+ (now.tv_usec * 1000 + ts->tv_nsec) / 1000000000;
+ reply->nb_sent = 0;
+ reply->nb_received = 0;
+ reply->msgs = NULL;
+
+ /* we have to lock the request queue here, as we will be adding a bunch
+ * of requests to the queue at once, and some of the replies may arrive
+ * before we add all of the requests to the queue.
+ */
+ pthread_mutex_lock(&pending_requests.lock);
+ /* we have to ensure that callback gets triggered even if we don't send
+ * anything, therefore earlier we have allocated a dummy request. fill
+ * it, and put it on the queue if we don't send any requests.
+ */
+ dummy->type = REQUEST_TYPE_ASYNC;
+ dummy->request = copy;
+ dummy->reply = NULL;
+ dummy->async.param = param;
+ dummy->reply_received = 1; /* short-circuit the timeout */
+
+ /* for secondary process, send request to the primary process only */
+ if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
+ ret = mp_request_async(eal_mp_socket_path(), copy, param);
+
+ /* if we didn't send anything, put dummy request on the queue */
+ if (ret == 0 && reply->nb_sent == 0) {
+ TAILQ_INSERT_TAIL(&pending_requests.requests, dummy,
+ next);
+ dummy_used = true;
+ }
+
+ pthread_mutex_unlock(&pending_requests.lock);
+
+ /* if we couldn't send anything, clean up */
+ if (ret != 0)
+ goto fail;
+ return 0;
+ }
+
+ /* for primary process, broadcast request */
+ mp_dir = opendir(mp_dir_path);
+ if (!mp_dir) {
+ RTE_LOG(ERR, EAL, "Unable to open directory %s\n", mp_dir_path);
+ rte_errno = errno;
+ goto unlock_fail;
+ }
+ dir_fd = dirfd(mp_dir);
+
+ /* lock the directory to prevent processes spinning up while we send */
+ if (flock(dir_fd, LOCK_EX)) {
+ RTE_LOG(ERR, EAL, "Unable to lock directory %s\n",
+ mp_dir_path);
+ rte_errno = errno;
+ goto closedir_fail;
+ }
+
+ while ((ent = readdir(mp_dir))) {
+ char path[PATH_MAX];
+
+ if (fnmatch(mp_filter, ent->d_name, 0) != 0)
+ continue;
+
+ snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
+ ent->d_name);
+
+ if (mp_request_async(path, copy, param))
+ ret = -1;
+ }
+ /* if we didn't send anything, put dummy request on the queue */
+ if (ret == 0 && reply->nb_sent == 0) {
+ TAILQ_INSERT_HEAD(&pending_requests.requests, dummy, next);
+ dummy_used = true;
+ }
+
+ /* trigger async request thread wake up */
+ pthread_cond_signal(&pending_requests.async_cond);
+
+ /* finally, unlock the queue */
+ pthread_mutex_unlock(&pending_requests.lock);
+
+ /* unlock the directory */
+ flock(dir_fd, LOCK_UN);
+
+ /* dir_fd automatically closed on closedir */
+ closedir(mp_dir);
+
+ /* if dummy was unused, free it */
+ if (!dummy_used)
+ free(dummy);
+
+ return ret;
+closedir_fail:
+ closedir(mp_dir);
+unlock_fail:
+ pthread_mutex_unlock(&pending_requests.lock);
+fail:
+ free(dummy);
+ free(param);
+ free(copy);
+ return -1;
+}
+
+int __rte_experimental
+rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
+{
RTE_LOG(DEBUG, EAL, "reply: %s\n", msg->name);
if (check_input(msg) == false)
diff --git a/lib/librte_eal/common/include/rte_eal.h b/lib/librte_eal/common/include/rte_eal.h
index d1cc89e..b804ff5 100644
--- a/lib/librte_eal/common/include/rte_eal.h
+++ b/lib/librte_eal/common/include/rte_eal.h
@@ -230,6 +230,16 @@ struct rte_mp_reply {
typedef int (*rte_mp_t)(const struct rte_mp_msg *msg, const void *peer);
/**
+ * Asynchronous reply function typedef used by other components.
+ *
+ * As we create socket channel for primary/secondary communication, use
+ * this function typedef to register action for coming responses to asynchronous
+ * requests.
+ */
+typedef int (*rte_mp_async_reply_t)(const struct rte_mp_msg *request,
+ const struct rte_mp_reply *reply);
+
+/**
* @warning
* @b EXPERIMENTAL: this API may change without prior notice
*
@@ -321,6 +331,32 @@ rte_mp_request_sync(struct rte_mp_msg *req, struct rte_mp_reply *reply,
* @warning
* @b EXPERIMENTAL: this API may change without prior notice
*
+ * Send a request to the peer process and expect a reply in a separate callback.
+ *
+ * This function sends a request message to the peer process, and will not
+ * block. Instead, reply will be received in a separate callback.
+ *
+ * @param req
+ * The req argument contains the customized request message.
+ *
+ * @param ts
+ * The ts argument specifies how long we can wait for the peer(s) to reply.
+ *
+ * @param clb
+ * The callback to trigger when all responses for this request have arrived.
+ *
+ * @return
+ * - On success, return 0.
+ * - On failure, return -1, and the reason will be stored in rte_errno.
+ */
+int __rte_experimental
+rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts,
+ rte_mp_async_reply_t clb);
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice
+ *
* Send a reply to the peer process.
*
* This function will send a reply message in response to a request message
diff --git a/lib/librte_eal/rte_eal_version.map b/lib/librte_eal/rte_eal_version.map
index 7b66c73..cb44b0a 100644
--- a/lib/librte_eal/rte_eal_version.map
+++ b/lib/librte_eal/rte_eal_version.map
@@ -225,6 +225,7 @@ EXPERIMENTAL {
rte_mp_action_unregister;
rte_mp_reply;
rte_mp_request_sync;
+ rte_mp_request_async;
rte_mp_sendmsg;
rte_service_attr_get;
rte_service_attr_reset_all;
--
2.7.4
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v7 3/3] eal: add asynchronous request API to DPDK IPC
2018-03-31 17:06 ` [dpdk-dev] [PATCH v7 3/3] eal: add asynchronous request API to DPDK IPC Anatoly Burakov
@ 2018-04-04 22:15 ` Thomas Monjalon
0 siblings, 0 replies; 40+ messages in thread
From: Thomas Monjalon @ 2018-04-04 22:15 UTC (permalink / raw)
To: Anatoly Burakov; +Cc: dev, jianfeng.tan, konstantin.ananyev
31/03/2018 19:06, Anatoly Burakov:
> This API is similar to the blocking API that is already present,
> but reply will be received in a separate callback by the caller
> (callback specified at the time of request, rather than registering
> for it in advance).
>
> Under the hood, we create a separate thread to deal with replies to
> asynchronous requests, that will just wait to be notified by the
> main thread, or woken up on a timer.
>
> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
> Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
Applied, thanks
You are expected to work on removing IPC threads during the next release cycle.
Thanks in advance :)
^ permalink raw reply [flat|nested] 40+ messages in thread
* [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-24 12:46 ` [dpdk-dev] [PATCH v5 1/2] eal: rename IPC sync request to pending request Anatoly Burakov
2018-03-26 7:31 ` Tan, Jianfeng
2018-03-27 13:59 ` [dpdk-dev] [PATCH v6 " Anatoly Burakov
@ 2018-03-27 13:59 ` Anatoly Burakov
2018-03-27 16:33 ` Thomas Monjalon
2 siblings, 1 reply; 40+ messages in thread
From: Anatoly Burakov @ 2018-03-27 13:59 UTC (permalink / raw)
To: dev; +Cc: failed, to, remove, 'kernel'
This API is similar to the blocking API that is already present,
but reply will be received in a separate callback by the caller
(callback specified at the time of request, rather than registering
for it in advance).
Under the hood, we create a separate thread to deal with replies to
asynchronous requests, that will just wait to be notified by the
main thread, or woken up on a timer.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
---
Notes:
v6:
- address review comments from Jianfeng
- do not add dummy item to queue by default
v5:
- addressed review comments from Jianfeng
- split into two patches to avoid rename noise
- do not mark ignored message as processed
v4:
- rebase on top of latest IPC Improvements patchset [2]
v3:
- added support for MP_IGN messages introduced in
IPC improvements v5 patchset
v2:
- fixed deadlocks and race conditions by not calling
callbacks while iterating over sync request list
- fixed use-after-free by making a copy of request
- changed API to also give user a copy of original
request, so that they know to which message the
callback is a reply to
- fixed missing .map file entries
This patch is dependent upon previously published patchsets
for IPC fixes [1] and improvements [2].
rte_mp_action_unregister and rte_mp_async_reply_unregister
do the same thing - should we perhaps make it one function?
[1] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Fixes/
[2] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Improvements/
lib/librte_eal/common/eal_common_proc.c | 458 +++++++++++++++++++++++++++++++-
lib/librte_eal/common/include/rte_eal.h | 36 +++
lib/librte_eal/rte_eal_version.map | 1 +
3 files changed, 482 insertions(+), 13 deletions(-)
diff --git a/lib/librte_eal/common/eal_common_proc.c b/lib/librte_eal/common/eal_common_proc.c
index 52b6ab2..139b653 100644
--- a/lib/librte_eal/common/eal_common_proc.c
+++ b/lib/librte_eal/common/eal_common_proc.c
@@ -26,6 +26,7 @@
#include <rte_errno.h>
#include <rte_lcore.h>
#include <rte_log.h>
+#include <rte_tailq.h>
#include "eal_private.h"
#include "eal_filesystem.h"
@@ -60,13 +61,32 @@ struct mp_msg_internal {
struct rte_mp_msg msg;
};
+struct async_request_param {
+ rte_mp_async_reply_t clb;
+ struct rte_mp_reply user_reply;
+ struct timespec end;
+ int n_responses_processed;
+};
+
struct pending_request {
TAILQ_ENTRY(pending_request) next;
- int reply_received;
+ enum {
+ REQUEST_TYPE_SYNC,
+ REQUEST_TYPE_ASYNC
+ } type;
char dst[PATH_MAX];
struct rte_mp_msg *request;
struct rte_mp_msg *reply;
- pthread_cond_t cond;
+ int reply_received;
+ RTE_STD_C11
+ union {
+ struct {
+ struct async_request_param *param;
+ } async;
+ struct {
+ pthread_cond_t cond;
+ } sync;
+ };
};
TAILQ_HEAD(pending_request_list, pending_request);
@@ -74,9 +94,12 @@ TAILQ_HEAD(pending_request_list, pending_request);
static struct {
struct pending_request_list requests;
pthread_mutex_t lock;
+ pthread_cond_t async_cond;
} pending_requests = {
.requests = TAILQ_HEAD_INITIALIZER(pending_requests.requests),
- .lock = PTHREAD_MUTEX_INITIALIZER
+ .lock = PTHREAD_MUTEX_INITIALIZER,
+ .async_cond = PTHREAD_COND_INITIALIZER
+ /**< used in async requests only */
};
/* forward declarations */
@@ -273,7 +296,12 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
memcpy(sync_req->reply, msg, sizeof(*msg));
/* -1 indicates that we've been asked to ignore */
sync_req->reply_received = m->type == MP_REP ? 1 : -1;
- pthread_cond_signal(&sync_req->cond);
+
+ if (sync_req->type == REQUEST_TYPE_SYNC)
+ pthread_cond_signal(&sync_req->sync.cond);
+ else if (sync_req->type == REQUEST_TYPE_ASYNC)
+ pthread_cond_signal(
+ &pending_requests.async_cond);
} else
RTE_LOG(ERR, EAL, "Drop mp reply: %s\n", msg->name);
pthread_mutex_unlock(&pending_requests.lock);
@@ -320,6 +348,189 @@ mp_handle(void *arg __rte_unused)
}
static int
+timespec_cmp(const struct timespec *a, const struct timespec *b)
+{
+ if (a->tv_sec < b->tv_sec)
+ return -1;
+ if (a->tv_sec > b->tv_sec)
+ return 1;
+ if (a->tv_nsec < b->tv_nsec)
+ return -1;
+ if (a->tv_nsec > b->tv_nsec)
+ return 1;
+ return 0;
+}
+
+enum async_action {
+ ACTION_NONE, /**< don't do anything */
+ ACTION_FREE, /**< free the action entry, but don't trigger callback */
+ ACTION_TRIGGER /**< trigger callback, then free action entry */
+};
+
+static enum async_action
+process_async_request(struct pending_request *sr, const struct timespec *now)
+{
+ struct async_request_param *param;
+ struct rte_mp_reply *reply;
+ bool timeout, received, last_msg;
+
+ param = sr->async.param;
+ reply = ¶m->user_reply;
+
+ /* did we timeout? */
+ timeout = timespec_cmp(¶m->end, now) <= 0;
+
+ /* did we receive a response? */
+ received = sr->reply_received != 0;
+
+ /* if we didn't time out, and we didn't receive a response, ignore */
+ if (!timeout && !received)
+ return ACTION_NONE;
+
+ /* if we received a response, adjust relevant data and copy mesasge. */
+ if (sr->reply_received == 1 && sr->reply) {
+ struct rte_mp_msg *msg, *user_msgs, *tmp;
+
+ msg = sr->reply;
+ user_msgs = reply->msgs;
+
+ tmp = realloc(user_msgs, sizeof(*msg) *
+ (reply->nb_received + 1));
+ if (!tmp) {
+ RTE_LOG(ERR, EAL, "Fail to alloc reply for request %s:%s\n",
+ sr->dst, sr->request->name);
+ /* this entry is going to be removed and its message
+ * dropped, but we don't want to leak memory, so
+ * continue.
+ */
+ } else {
+ user_msgs = tmp;
+ reply->msgs = user_msgs;
+ memcpy(&user_msgs[reply->nb_received],
+ msg, sizeof(*msg));
+ reply->nb_received++;
+ }
+
+ /* mark this request as processed */
+ param->n_responses_processed++;
+ } else if (sr->reply_received == -1) {
+ /* we were asked to ignore this process */
+ reply->nb_sent--;
+ }
+ free(sr->reply);
+
+ last_msg = param->n_responses_processed == reply->nb_sent;
+
+ return last_msg ? ACTION_TRIGGER : ACTION_FREE;
+}
+
+static void
+trigger_async_action(struct pending_request *sr)
+{
+ struct async_request_param *param;
+ struct rte_mp_reply *reply;
+
+ param = sr->async.param;
+ reply = ¶m->user_reply;
+
+ param->clb(sr->request, reply);
+
+ /* clean up */
+ free(sr->async.param->user_reply.msgs);
+ free(sr->async.param);
+ free(sr->request);
+}
+
+static void *
+async_reply_handle(void *arg __rte_unused)
+{
+ struct pending_request *sr;
+ struct timeval now;
+ struct timespec timeout, ts_now;
+ while (1) {
+ struct pending_request *trigger = NULL;
+ int ret;
+ bool nowait = false;
+ bool timedwait = false;
+
+ pthread_mutex_lock(&pending_requests.lock);
+
+ /* scan through the list and see if there are any timeouts that
+ * are earlier than our current timeout.
+ */
+ TAILQ_FOREACH(sr, &pending_requests.requests, next) {
+ if (sr->type != REQUEST_TYPE_ASYNC)
+ continue;
+ if (!timedwait || timespec_cmp(&sr->async.param->end,
+ &timeout) < 0) {
+ memcpy(&timeout, &sr->async.param->end,
+ sizeof(timeout));
+ timedwait = true;
+ }
+
+ /* sometimes, we don't even wait */
+ if (sr->reply_received) {
+ nowait = true;
+ break;
+ }
+ }
+
+ if (nowait)
+ ret = 0;
+ else if (timedwait)
+ ret = pthread_cond_timedwait(
+ &pending_requests.async_cond,
+ &pending_requests.lock, &timeout);
+ else
+ ret = pthread_cond_wait(&pending_requests.async_cond,
+ &pending_requests.lock);
+
+ if (gettimeofday(&now, NULL) < 0) {
+ RTE_LOG(ERR, EAL, "Cannot get current time\n");
+ break;
+ }
+ ts_now.tv_nsec = now.tv_usec * 1000;
+ ts_now.tv_sec = now.tv_sec;
+
+ if (ret == 0 || ret == ETIMEDOUT) {
+ struct pending_request *next;
+ /* we've either been woken up, or we timed out */
+
+ /* we have still the lock, check if anything needs
+ * processing.
+ */
+ TAILQ_FOREACH_SAFE(sr, &pending_requests.requests, next,
+ next) {
+ enum async_action action;
+ if (sr->type != REQUEST_TYPE_ASYNC)
+ continue;
+
+ action = process_async_request(sr, &ts_now);
+ if (action == ACTION_FREE) {
+ TAILQ_REMOVE(&pending_requests.requests,
+ sr, next);
+ free(sr);
+ } else if (action == ACTION_TRIGGER &&
+ trigger == NULL) {
+ TAILQ_REMOVE(&pending_requests.requests,
+ sr, next);
+ trigger = sr;
+ }
+ }
+ }
+ pthread_mutex_unlock(&pending_requests.lock);
+ if (trigger) {
+ trigger_async_action(trigger);
+ free(trigger);
+ }
+ };
+
+ RTE_LOG(ERR, EAL, "ERROR: asynchronous requests disabled\n");
+
+ return NULL;
+}
+
+static int
open_socket_fd(void)
{
char peer_name[PATH_MAX] = {0};
@@ -382,7 +593,7 @@ rte_mp_channel_init(void)
char thread_name[RTE_MAX_THREAD_NAME_LEN];
char path[PATH_MAX];
int dir_fd;
- pthread_t tid;
+ pthread_t mp_handle_tid, async_reply_handle_tid;
/* create filter path */
create_socket_path("*", path, sizeof(path));
@@ -419,7 +630,16 @@ rte_mp_channel_init(void)
return -1;
}
- if (pthread_create(&tid, NULL, mp_handle, NULL) < 0) {
+ if (pthread_create(&mp_handle_tid, NULL, mp_handle, NULL) < 0) {
+ RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
+ strerror(errno));
+ close(mp_fd);
+ mp_fd = -1;
+ return -1;
+ }
+
+ if (pthread_create(&async_reply_handle_tid, NULL,
+ async_reply_handle, NULL) < 0) {
RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
strerror(errno));
close(mp_fd);
@@ -430,7 +650,11 @@ rte_mp_channel_init(void)
/* try best to set thread name */
snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_handle");
- rte_thread_setname(tid, thread_name);
+ rte_thread_setname(mp_handle_tid, thread_name);
+
+ /* try best to set thread name */
+ snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_async_handle");
+ rte_thread_setname(async_reply_handle_tid, thread_name);
/* unlock the directory */
flock(dir_fd, LOCK_UN);
@@ -602,18 +826,77 @@ rte_mp_sendmsg(struct rte_mp_msg *msg)
}
static int
-mp_request_one(const char *dst, struct rte_mp_msg *req,
+mp_request_async(const char *dst, struct rte_mp_msg *req,
+ struct async_request_param *param)
+{
+ struct rte_mp_msg *reply_msg;
+ struct pending_request *sync_req, *exist;
+ int ret;
+
+ sync_req = malloc(sizeof(*sync_req));
+ reply_msg = malloc(sizeof(*reply_msg));
+ if (sync_req == NULL || reply_msg == NULL) {
+ RTE_LOG(ERR, EAL, "Could not allocate space for sync request\n");
+ rte_errno = ENOMEM;
+ ret = -1;
+ goto fail;
+ }
+
+ memset(sync_req, 0, sizeof(*sync_req));
+ memset(reply_msg, 0, sizeof(*reply_msg));
+
+ sync_req->type = REQUEST_TYPE_ASYNC;
+ strcpy(sync_req->dst, dst);
+ sync_req->request = req;
+ sync_req->reply = reply_msg;
+ sync_req->async.param = param;
+
+ /* queue already locked by caller */
+
+ exist = find_sync_request(dst, req->name);
+ if (!exist) {
+ TAILQ_INSERT_TAIL(&pending_requests.requests, sync_req, next);
+ } else {
+ RTE_LOG(ERR, EAL, "A pending request %s:%s\n", dst, req->name);
+ rte_errno = EEXIST;
+ ret = -1;
+ goto fail;
+ }
+
+ ret = send_msg(dst, req, MP_REQ);
+ if (ret < 0) {
+ RTE_LOG(ERR, EAL, "Fail to send request %s:%s\n",
+ dst, req->name);
+ ret = -1;
+ goto fail;
+ } else if (ret == 0) {
+ ret = 0;
+ goto fail;
+ }
+
+ param->user_reply.nb_sent++;
+
+ return 0;
+fail:
+ free(sync_req);
+ free(reply_msg);
+ return ret;
+}
+
+static int
+mp_request_sync(const char *dst, struct rte_mp_msg *req,
struct rte_mp_reply *reply, const struct timespec *ts)
{
int ret;
struct rte_mp_msg msg, *tmp;
struct pending_request sync_req, *exist;
+ sync_req.type = REQUEST_TYPE_SYNC;
sync_req.reply_received = 0;
strcpy(sync_req.dst, dst);
sync_req.request = req;
sync_req.reply = &msg;
- pthread_cond_init(&sync_req.cond, NULL);
+ pthread_cond_init(&sync_req.sync.cond, NULL);
pthread_mutex_lock(&pending_requests.lock);
exist = find_sync_request(dst, req->name);
@@ -637,7 +920,7 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
reply->nb_sent++;
do {
- ret = pthread_cond_timedwait(&sync_req.cond,
+ ret = pthread_cond_timedwait(&sync_req.sync.cond,
&pending_requests.lock, ts);
} while (ret != 0 && ret != ETIMEDOUT);
@@ -703,7 +986,7 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
/* for secondary process, send request to the primary process only */
if (rte_eal_process_type() == RTE_PROC_SECONDARY)
- return mp_request_one(eal_mp_socket_path(), req, reply, &end);
+ return mp_request_sync(eal_mp_socket_path(), req, reply, &end);
/* for primary process, broadcast request, and collect reply 1 by 1 */
mp_dir = opendir(mp_dir_path);
@@ -732,7 +1015,7 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
ent->d_name);
- if (mp_request_one(path, req, reply, &end))
+ if (mp_request_sync(path, req, reply, &end))
ret = -1;
}
/* unlock the directory */
@@ -744,9 +1027,158 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
}
int __rte_experimental
-rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
+rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts,
+ rte_mp_async_reply_t clb)
{
+ struct rte_mp_msg *copy;
+ struct pending_request *dummy;
+ struct async_request_param *param;
+ struct rte_mp_reply *reply;
+ int dir_fd, ret = 0;
+ DIR *mp_dir;
+ struct dirent *ent;
+ struct timeval now;
+ struct timespec *end;
+ bool dummy_used = false;
+
+ RTE_LOG(DEBUG, EAL, "request: %s\n", req->name);
+
+ if (check_input(req) == false)
+ return -1;
+ if (gettimeofday(&now, NULL) < 0) {
+ RTE_LOG(ERR, EAL, "Faile to get current time\n");
+ rte_errno = errno;
+ return -1;
+ }
+ copy = malloc(sizeof(*copy));
+ dummy = malloc(sizeof(*dummy));
+ param = malloc(sizeof(*param));
+ if (copy == NULL || dummy == NULL || param == NULL) {
+ RTE_LOG(ERR, EAL, "Failed to allocate memory for async reply\n");
+ rte_errno = ENOMEM;
+ goto fail;
+ }
+
+ memset(copy, 0, sizeof(*copy));
+ memset(dummy, 0, sizeof(*dummy));
+ memset(param, 0, sizeof(*param));
+
+ /* copy message */
+ memcpy(copy, req, sizeof(*copy));
+
+ param->n_responses_processed = 0;
+ param->clb = clb;
+ end = ¶m->end;
+ reply = ¶m->user_reply;
+
+ end->tv_nsec = (now.tv_usec * 1000 + ts->tv_nsec) % 1000000000;
+ end->tv_sec = now.tv_sec + ts->tv_sec +
+ (now.tv_usec * 1000 + ts->tv_nsec) / 1000000000;
+ reply->nb_sent = 0;
+ reply->nb_received = 0;
+ reply->msgs = NULL;
+
+ /* we have to lock the request queue here, as we will be adding a bunch
+ * of requests to the queue at once, and some of the replies may arrive
+ * before we add all of the requests to the queue.
+ */
+ pthread_mutex_lock(&pending_requests.lock);
+ /* we have to ensure that callback gets triggered even if we don't send
+ * anything, therefore earlier we have allocated a dummy request. fill
+ * it, and put it on the queue if we don't send any requests.
+ */
+ dummy->type = REQUEST_TYPE_ASYNC;
+ dummy->request = copy;
+ dummy->reply = NULL;
+ dummy->async.param = param;
+ dummy->reply_received = 1; /* short-circuit the timeout */
+
+ /* for secondary process, send request to the primary process only */
+ if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
+ ret = mp_request_async(eal_mp_socket_path(), copy, param);
+
+ /* if we didn't send anything, put dummy request on the queue */
+ if (ret == 0 && reply->nb_sent == 0) {
+ TAILQ_INSERT_TAIL(&pending_requests.requests, dummy,
+ next);
+ dummy_used = true;
+ }
+
+ pthread_mutex_unlock(&pending_requests.lock);
+
+ /* if we couldn't send anything, clean up */
+ if (ret != 0)
+ goto fail;
+ return 0;
+ }
+
+ /* for primary process, broadcast request */
+ mp_dir = opendir(mp_dir_path);
+ if (!mp_dir) {
+ RTE_LOG(ERR, EAL, "Unable to open directory %s\n", mp_dir_path);
+ rte_errno = errno;
+ goto unlock_fail;
+ }
+ dir_fd = dirfd(mp_dir);
+
+ /* lock the directory to prevent processes spinning up while we send */
+ if (flock(dir_fd, LOCK_EX)) {
+ RTE_LOG(ERR, EAL, "Unable to lock directory %s\n",
+ mp_dir_path);
+ rte_errno = errno;
+ goto closedir_fail;
+ }
+
+ while ((ent = readdir(mp_dir))) {
+ char path[PATH_MAX];
+
+ if (fnmatch(mp_filter, ent->d_name, 0) != 0)
+ continue;
+
+ snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
+ ent->d_name);
+
+ if (mp_request_async(path, copy, param))
+ ret = -1;
+ }
+ /* if we didn't send anything, put dummy request on the queue */
+ if (ret == 0 && reply->nb_sent == 0) {
+ TAILQ_INSERT_HEAD(&pending_requests.requests, dummy, next);
+ dummy_used = true;
+ }
+
+ /* trigger async request thread wake up */
+ pthread_cond_signal(&pending_requests.async_cond);
+
+ /* finally, unlock the queue */
+ pthread_mutex_unlock(&pending_requests.lock);
+
+ /* unlock the directory */
+ flock(dir_fd, LOCK_UN);
+
+ /* dir_fd automatically closed on closedir */
+ closedir(mp_dir);
+
+ /* if dummy was unused, free it */
+ if (!dummy_used)
+ free(dummy);
+
+ return ret;
+closedir_fail:
+ closedir(mp_dir);
+unlock_fail:
+ pthread_mutex_unlock(&pending_requests.lock);
+fail:
+ free(dummy);
+ free(param);
+ free(copy);
+ return -1;
+}
+
+int __rte_experimental
+rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
+{
RTE_LOG(DEBUG, EAL, "reply: %s\n", msg->name);
if (check_input(msg) == false)
diff --git a/lib/librte_eal/common/include/rte_eal.h b/lib/librte_eal/common/include/rte_eal.h
index 044474e..87ebfd0 100644
--- a/lib/librte_eal/common/include/rte_eal.h
+++ b/lib/librte_eal/common/include/rte_eal.h
@@ -230,6 +230,16 @@ struct rte_mp_reply {
typedef int (*rte_mp_t)(const struct rte_mp_msg *msg, const void *peer);
/**
+ * Asynchronous reply function typedef used by other components.
+ *
+ * As we create socket channel for primary/secondary communication, use
+ * this function typedef to register action for coming responses to asynchronous
+ * requests.
+ */
+typedef int (*rte_mp_async_reply_t)(const struct rte_mp_msg *request,
+ const struct rte_mp_reply *reply);
+
+/**
* @warning
* @b EXPERIMENTAL: this API may change without prior notice
*
@@ -321,6 +331,32 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
* @warning
* @b EXPERIMENTAL: this API may change without prior notice
*
+ * Send a request to the peer process and expect a reply in a separate callback.
+ *
+ * This function sends a request message to the peer process, and will not
+ * block. Instead, reply will be received in a separate callback.
+ *
+ * @param req
+ * The req argument contains the customized request message.
+ *
+ * @param ts
+ * The ts argument specifies how long we can wait for the peer(s) to reply.
+ *
+ * @param clb
+ * The callback to trigger when all responses for this request have arrived.
+ *
+ * @return
+ * - On success, return 0.
+ * - On failure, return -1, and the reason will be stored in rte_errno.
+ */
+int __rte_experimental
+rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts,
+ rte_mp_async_reply_t clb);
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice
+ *
* Send a reply to the peer process.
*
* This function will send a reply message in response to a request message
diff --git a/lib/librte_eal/rte_eal_version.map b/lib/librte_eal/rte_eal_version.map
index d123602..328a0be 100644
--- a/lib/librte_eal/rte_eal_version.map
+++ b/lib/librte_eal/rte_eal_version.map
@@ -225,6 +225,7 @@ EXPERIMENTAL {
rte_mp_action_unregister;
rte_mp_sendmsg;
rte_mp_request;
+ rte_mp_request_async;
rte_mp_reply;
rte_service_attr_get;
rte_service_attr_reset_all;
--
2.7.4
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-27 13:59 ` [dpdk-dev] [PATCH v6 2/2] " Anatoly Burakov
@ 2018-03-27 16:33 ` Thomas Monjalon
2018-03-28 2:08 ` Tan, Jianfeng
0 siblings, 1 reply; 40+ messages in thread
From: Thomas Monjalon @ 2018-03-27 16:33 UTC (permalink / raw)
To: Anatoly Burakov; +Cc: dev, jianfeng.tan, konstantin.ananyev, harry.van.haaren
27/03/2018 15:59, Anatoly Burakov:
> Under the hood, we create a separate thread to deal with replies to
> asynchronous requests, that will just wait to be notified by the
> main thread, or woken up on a timer.
I really don't like that a library is creating a thread.
We don't even know where the thread is created (which core).
Can it be a rte_service? or in the interrupt thread?
> --- a/lib/librte_eal/rte_eal_version.map
> +++ b/lib/librte_eal/rte_eal_version.map
> @@ -225,6 +225,7 @@ EXPERIMENTAL {
> rte_mp_action_unregister;
> rte_mp_sendmsg;
> rte_mp_request;
> + rte_mp_request_async;
So there is rte_mp_request and rte_mp_request_async?
You should rename rte_mp_request, I guess.
> rte_mp_reply;
> rte_service_attr_get;
> rte_service_attr_reset_all;
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-27 16:33 ` Thomas Monjalon
@ 2018-03-28 2:08 ` Tan, Jianfeng
2018-03-28 7:29 ` Thomas Monjalon
0 siblings, 1 reply; 40+ messages in thread
From: Tan, Jianfeng @ 2018-03-28 2:08 UTC (permalink / raw)
To: Thomas Monjalon, Burakov, Anatoly
Cc: dev, Ananyev, Konstantin, Van Haaren, Harry
Hi Thomas ,
> -----Original Message-----
> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> Sent: Wednesday, March 28, 2018 12:34 AM
> To: Burakov, Anatoly
> Cc: dev@dpdk.org; Tan, Jianfeng; Ananyev, Konstantin; Van Haaren, Harry
> Subject: Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to
> DPDK IPC
>
> 27/03/2018 15:59, Anatoly Burakov:
> > Under the hood, we create a separate thread to deal with replies to
> > asynchronous requests, that will just wait to be notified by the
> > main thread, or woken up on a timer.
>
> I really don't like that a library is creating a thread.
> We don't even know where the thread is created (which core).
> Can it be a rte_service? or in the interrupt thread?
Agree that we'd better not adding so many threads in a library.
I was considering to merge all the threads into the interrupt thread, however, we don't have an interrupt thread in freebsd. Further, we don't implement alarm API in freebsd. That's why I tend to current implementation, and optimize it later.
For rte_service, it may be not a good idea to reply on it as it needs explicit API calls to setup.
>
>
> > --- a/lib/librte_eal/rte_eal_version.map
> > +++ b/lib/librte_eal/rte_eal_version.map
> > @@ -225,6 +225,7 @@ EXPERIMENTAL {
> > rte_mp_action_unregister;
> > rte_mp_sendmsg;
> > rte_mp_request;
> > + rte_mp_request_async;
>
> So there is rte_mp_request and rte_mp_request_async?
> You should rename rte_mp_request, I guess.
+1.
Thanks,
Jianfeng
>
> > rte_mp_reply;
> > rte_service_attr_get;
> > rte_service_attr_reset_all;
>
>
>
>
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 2:08 ` Tan, Jianfeng
@ 2018-03-28 7:29 ` Thomas Monjalon
2018-03-28 8:22 ` Van Haaren, Harry
2018-03-28 9:11 ` Bruce Richardson
0 siblings, 2 replies; 40+ messages in thread
From: Thomas Monjalon @ 2018-03-28 7:29 UTC (permalink / raw)
To: Tan, Jianfeng
Cc: Burakov, Anatoly, dev, Ananyev, Konstantin, Van Haaren, Harry
28/03/2018 04:08, Tan, Jianfeng:
> Hi Thomas ,
>
> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> > 27/03/2018 15:59, Anatoly Burakov:
> > > Under the hood, we create a separate thread to deal with replies to
> > > asynchronous requests, that will just wait to be notified by the
> > > main thread, or woken up on a timer.
> >
> > I really don't like that a library is creating a thread.
> > We don't even know where the thread is created (which core).
> > Can it be a rte_service? or in the interrupt thread?
>
> Agree that we'd better not adding so many threads in a library.
>
> I was considering to merge all the threads into the interrupt thread, however, we don't have an interrupt thread in freebsd. Further, we don't implement alarm API in freebsd. That's why I tend to current implementation, and optimize it later.
I would prefer we improve the current code now instead of polluting more
with more uncontrolled threads.
> For rte_service, it may be not a good idea to reply on it as it needs explicit API calls to setup.
I don't see the issue of the explicit API.
The IPC is a new service.
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 7:29 ` Thomas Monjalon
@ 2018-03-28 8:22 ` Van Haaren, Harry
2018-03-28 8:55 ` Tan, Jianfeng
2018-03-28 9:11 ` Bruce Richardson
1 sibling, 1 reply; 40+ messages in thread
From: Van Haaren, Harry @ 2018-03-28 8:22 UTC (permalink / raw)
To: Thomas Monjalon, Tan, Jianfeng; +Cc: Burakov, Anatoly, dev, Ananyev, Konstantin
> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> Sent: Wednesday, March 28, 2018 8:30 AM
> To: Tan, Jianfeng <jianfeng.tan@intel.com>
> Cc: Burakov, Anatoly <anatoly.burakov@intel.com>; dev@dpdk.org; Ananyev,
> Konstantin <konstantin.ananyev@intel.com>; Van Haaren, Harry
> <harry.van.haaren@intel.com>
> Subject: Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to
> DPDK IPC
>
> 28/03/2018 04:08, Tan, Jianfeng:
> > Hi Thomas ,
> >
> > From: Thomas Monjalon [mailto:thomas@monjalon.net]
> > > 27/03/2018 15:59, Anatoly Burakov:
> > > > Under the hood, we create a separate thread to deal with replies to
> > > > asynchronous requests, that will just wait to be notified by the
> > > > main thread, or woken up on a timer.
> > >
> > > I really don't like that a library is creating a thread.
> > > We don't even know where the thread is created (which core).
> > > Can it be a rte_service? or in the interrupt thread?
> >
> > Agree that we'd better not adding so many threads in a library.
> >
> > I was considering to merge all the threads into the interrupt thread,
> however, we don't have an interrupt thread in freebsd. Further, we don't
> implement alarm API in freebsd. That's why I tend to current implementation,
> and optimize it later.
>
> I would prefer we improve the current code now instead of polluting more
> with more uncontrolled threads.
>
> > For rte_service, it may be not a good idea to reply on it as it needs
> explicit API calls to setup.
>
> I don't see the issue of the explicit API.
> The IPC is a new service.
Although I do like to see new services, if we want to enable "core" dpdk functionality with Services, we need a proper designed solution for that. Service cores is not intended for "occasional" work - there is no method to block and sleep on a specific service until work becomes available, so this would imply a busy-polling. Using a service (hence busy polling) for rte_malloc()-based memory mapping requests is inefficient, and total overkill :)
For this patch I suggest to use some blocking-read capable mechanism.
The above said, in the longer term it would be good to have a design that allows new file-descriptors to be added to a "dpdk core" thread, which performs occasional lengthy work if the FD has data available.
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 8:22 ` Van Haaren, Harry
@ 2018-03-28 8:55 ` Tan, Jianfeng
2018-03-28 9:10 ` Van Haaren, Harry
2018-03-28 9:21 ` Burakov, Anatoly
0 siblings, 2 replies; 40+ messages in thread
From: Tan, Jianfeng @ 2018-03-28 8:55 UTC (permalink / raw)
To: Van Haaren, Harry, Thomas Monjalon
Cc: Burakov, Anatoly, dev, Ananyev, Konstantin
Hi Thomas and Harry,
On 3/28/2018 4:22 PM, Van Haaren, Harry wrote:
>> From: Thomas Monjalon [mailto:thomas@monjalon.net]
>> Sent: Wednesday, March 28, 2018 8:30 AM
>> To: Tan, Jianfeng <jianfeng.tan@intel.com>
>> Cc: Burakov, Anatoly <anatoly.burakov@intel.com>; dev@dpdk.org; Ananyev,
>> Konstantin <konstantin.ananyev@intel.com>; Van Haaren, Harry
>> <harry.van.haaren@intel.com>
>> Subject: Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to
>> DPDK IPC
>>
>> 28/03/2018 04:08, Tan, Jianfeng:
>>> Hi Thomas ,
>>>
>>> From: Thomas Monjalon [mailto:thomas@monjalon.net]
>>>> 27/03/2018 15:59, Anatoly Burakov:
>>>>> Under the hood, we create a separate thread to deal with replies to
>>>>> asynchronous requests, that will just wait to be notified by the
>>>>> main thread, or woken up on a timer.
>>>> I really don't like that a library is creating a thread.
>>>> We don't even know where the thread is created (which core).
>>>> Can it be a rte_service? or in the interrupt thread?
>>> Agree that we'd better not adding so many threads in a library.
>>>
>>> I was considering to merge all the threads into the interrupt thread,
>> however, we don't have an interrupt thread in freebsd. Further, we don't
>> implement alarm API in freebsd. That's why I tend to current implementation,
>> and optimize it later.
>>
>> I would prefer we improve the current code now instead of polluting more
>> with more uncontrolled threads.
>>
>>> For rte_service, it may be not a good idea to reply on it as it needs
>> explicit API calls to setup.
>>
>> I don't see the issue of the explicit API.
>> The IPC is a new service.
My concern is that not every DPDK application sets up rte_service, but
IPC will be used for very fundamental functions, like memory allocation.
We could not possibly ask all DPDK applications to add rte_service now.
And also take Harry's comments below into consideration, most likely, we
will move these threads into interrupt thread now by adding
> Although I do like to see new services, if we want to enable "core" dpdk functionality with Services, we need a proper designed solution for that. Service cores is not intended for "occasional" work - there is no method to block and sleep on a specific service until work becomes available, so this would imply a busy-polling. Using a service (hence busy polling) for rte_malloc()-based memory mapping requests is inefficient, and total overkill :)
>
> For this patch I suggest to use some blocking-read capable mechanism.
The problem here is that we add too many threads; blocking-read does not
decrease # of threads.
>
> The above said, in the longer term it would be good to have a design that allows new file-descriptors to be added to a "dpdk core" thread, which performs occasional lengthy work if the FD has data available.
Interrupt thread vs rte_service, which is the direction to go? We
actually have some others threads, in vhost and even virtio-user; we can
also avoid those threads if we have a clear direction.
Thanks,
Jianfeng
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 8:55 ` Tan, Jianfeng
@ 2018-03-28 9:10 ` Van Haaren, Harry
2018-03-28 9:21 ` Burakov, Anatoly
1 sibling, 0 replies; 40+ messages in thread
From: Van Haaren, Harry @ 2018-03-28 9:10 UTC (permalink / raw)
To: Tan, Jianfeng, Thomas Monjalon; +Cc: Burakov, Anatoly, dev, Ananyev, Konstantin
> -----Original Message-----
> From: Tan, Jianfeng
> Sent: Wednesday, March 28, 2018 9:55 AM
> To: Van Haaren, Harry <harry.van.haaren@intel.com>; Thomas Monjalon
> <thomas@monjalon.net>
> Cc: Burakov, Anatoly <anatoly.burakov@intel.com>; dev@dpdk.org; Ananyev,
> Konstantin <konstantin.ananyev@intel.com>
> Subject: Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to
> DPDK IPC
>
> Hi Thomas and Harry,
Hey,
> On 3/28/2018 4:22 PM, Van Haaren, Harry wrote:
> >> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> >> Sent: Wednesday, March 28, 2018 8:30 AM
> >> To: Tan, Jianfeng <jianfeng.tan@intel.com>
> >> Cc: Burakov, Anatoly <anatoly.burakov@intel.com>; dev@dpdk.org; Ananyev,
> >> Konstantin <konstantin.ananyev@intel.com>; Van Haaren, Harry
> >> <harry.van.haaren@intel.com>
> >> Subject: Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API
> to
> >> DPDK IPC
> >>
> >> 28/03/2018 04:08, Tan, Jianfeng:
> >>> Hi Thomas ,
> >>>
> >>> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> >>>> 27/03/2018 15:59, Anatoly Burakov:
> >>>>> Under the hood, we create a separate thread to deal with replies to
> >>>>> asynchronous requests, that will just wait to be notified by the
> >>>>> main thread, or woken up on a timer.
> >>>> I really don't like that a library is creating a thread.
> >>>> We don't even know where the thread is created (which core).
> >>>> Can it be a rte_service? or in the interrupt thread?
> >>> Agree that we'd better not adding so many threads in a library.
> >>>
> >>> I was considering to merge all the threads into the interrupt thread,
> >> however, we don't have an interrupt thread in freebsd. Further, we don't
> >> implement alarm API in freebsd. That's why I tend to current
> implementation,
> >> and optimize it later.
> >>
> >> I would prefer we improve the current code now instead of polluting more
> >> with more uncontrolled threads.
> >>
> >>> For rte_service, it may be not a good idea to reply on it as it needs
> >> explicit API calls to setup.
> >>
> >> I don't see the issue of the explicit API.
> >> The IPC is a new service.
>
> My concern is that not every DPDK application sets up rte_service, but
> IPC will be used for very fundamental functions, like memory allocation.
> We could not possibly ask all DPDK applications to add rte_service now.
>
> And also take Harry's comments below into consideration, most likely, we
> will move these threads into interrupt thread now by adding
I don't suggest moving everything into interrupt thread.
We need to ensure the interrupt thread (aka, link status interrupt thread) is not
busy for too long. Certain work may cause un-acceptable delay in the interrupt
thread, hence perhaps we should have 2 core DPDK threads:
1) EAL interrupt thread: handles only interrupts, so is ~always blocked on some FDs
2) EAL "core" thread: handles the IPC work here, and any other irregular work that
can take some time to complete. Adding all core DPDK work to this thread enables
us to refactor and create a scalable / coherent design.
> > Although I do like to see new services, if we want to enable "core" dpdk
> functionality with Services, we need a proper designed solution for that.
> Service cores is not intended for "occasional" work - there is no method to
> block and sleep on a specific service until work becomes available, so this
> would imply a busy-polling. Using a service (hence busy polling) for
> rte_malloc()-based memory mapping requests is inefficient, and total
> overkill :)
> >
> > For this patch I suggest to use some blocking-read capable mechanism.
>
> The problem here is that we add too many threads; blocking-read does not
> decrease # of threads.
To correct my statement just above - some method capable of blocking reads (as opposed to
busy polling). In my mind, this method *must* allow waiting on multiple FDs,
as to *require* only 1 thread. We could use more if it makes sense to do so.
> > The above said, in the longer term it would be good to have a design that
> allows new file-descriptors to be added to a "dpdk core" thread, which
> performs occasional lengthy work if the FD has data available.
>
> Interrupt thread vs rte_service, which is the direction to go? We
> actually have some others threads, in vhost and even virtio-user; we can
> also avoid those threads if we have a clear direction.
I don't think that service is the correct hammer for this problem.
As to interrupt thread or creating a new thread, what makes more sense given
the current codebase? Is the current implementation an acceptable short-term
solution, that gets reworked to be more generic in future?
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 8:55 ` Tan, Jianfeng
2018-03-28 9:10 ` Van Haaren, Harry
@ 2018-03-28 9:21 ` Burakov, Anatoly
2018-03-28 9:53 ` Thomas Monjalon
1 sibling, 1 reply; 40+ messages in thread
From: Burakov, Anatoly @ 2018-03-28 9:21 UTC (permalink / raw)
To: Tan, Jianfeng, Van Haaren, Harry, Thomas Monjalon
Cc: dev, Ananyev, Konstantin
On 28-Mar-18 9:55 AM, Tan, Jianfeng wrote:
> Hi Thomas and Harry,
>
>
> On 3/28/2018 4:22 PM, Van Haaren, Harry wrote:
>>> From: Thomas Monjalon [mailto:thomas@monjalon.net]
>>> Sent: Wednesday, March 28, 2018 8:30 AM
>>> To: Tan, Jianfeng <jianfeng.tan@intel.com>
>>> Cc: Burakov, Anatoly <anatoly.burakov@intel.com>; dev@dpdk.org; Ananyev,
>>> Konstantin <konstantin.ananyev@intel.com>; Van Haaren, Harry
>>> <harry.van.haaren@intel.com>
>>> Subject: Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request
>>> API to
>>> DPDK IPC
>>>
>>> 28/03/2018 04:08, Tan, Jianfeng:
>>>> Hi Thomas ,
>>>>
>>>> From: Thomas Monjalon [mailto:thomas@monjalon.net]
>>>>> 27/03/2018 15:59, Anatoly Burakov:
>>>>>> Under the hood, we create a separate thread to deal with replies to
>>>>>> asynchronous requests, that will just wait to be notified by the
>>>>>> main thread, or woken up on a timer.
>>>>> I really don't like that a library is creating a thread.
>>>>> We don't even know where the thread is created (which core).
>>>>> Can it be a rte_service? or in the interrupt thread?
>>>> Agree that we'd better not adding so many threads in a library.
>>>>
>>>> I was considering to merge all the threads into the interrupt thread,
>>> however, we don't have an interrupt thread in freebsd. Further, we don't
>>> implement alarm API in freebsd. That's why I tend to current
>>> implementation,
>>> and optimize it later.
>>>
>>> I would prefer we improve the current code now instead of polluting more
>>> with more uncontrolled threads.
>>>
>>>> For rte_service, it may be not a good idea to reply on it as it needs
>>> explicit API calls to setup.
>>>
>>> I don't see the issue of the explicit API.
>>> The IPC is a new service.
>
> My concern is that not every DPDK application sets up rte_service, but
> IPC will be used for very fundamental functions, like memory allocation.
> We could not possibly ask all DPDK applications to add rte_service now.
>
> And also take Harry's comments below into consideration, most likely, we
> will move these threads into interrupt thread now by adding
>
>> Although I do like to see new services, if we want to enable "core"
>> dpdk functionality with Services, we need a proper designed solution
>> for that. Service cores is not intended for "occasional" work - there
>> is no method to block and sleep on a specific service until work
>> becomes available, so this would imply a busy-polling. Using a service
>> (hence busy polling) for rte_malloc()-based memory mapping requests is
>> inefficient, and total overkill :)
>>
>> For this patch I suggest to use some blocking-read capable mechanism.
>
> The problem here is that we add too many threads; blocking-read does not
> decrease # of threads.
>
>>
>> The above said, in the longer term it would be good to have a design
>> that allows new file-descriptors to be added to a "dpdk core" thread,
>> which performs occasional lengthy work if the FD has data available.
>
> Interrupt thread vs rte_service, which is the direction to go? We
> actually have some others threads, in vhost and even virtio-user; we can
> also avoid those threads if we have a clear direction.
>
> Thanks,
> Jianfeng
>
Hi all,
First of all, @Thomas, this is not a "new library" - it's part of EAL.
We're going to be removing a few threads from EAL as it is because of
IPC (Jianfeng has already submitted patches for those), so i don't think
it's such a big deal to have two IPC threads instead of one. I'm open to
suggestions on how to make this work without a second thread, but i
don't see it.
We've discussed possibility of using rte_service internally, but decided
against it for reasons already outlined by Harry - it's not a suitable
mechanism for this kind of thing, not as it is.
Using interrupt thread for this _will_ work, however this will require a
a lot more changes, as currently alarm API allocates everything through
rte_malloc, while we want to use IPC for rte_malloc (which would make it
a circular dependency). So it'll probably be more API and more
complexity for dealing with malloc vs rte_malloc allocations. Hence the
least-bad approach taken here: a new thread.
--
Thanks,
Anatoly
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 9:21 ` Burakov, Anatoly
@ 2018-03-28 9:53 ` Thomas Monjalon
2018-03-28 10:42 ` Burakov, Anatoly
0 siblings, 1 reply; 40+ messages in thread
From: Thomas Monjalon @ 2018-03-28 9:53 UTC (permalink / raw)
To: Burakov, Anatoly, Tan, Jianfeng
Cc: Van Haaren, Harry, dev, Ananyev, Konstantin
28/03/2018 11:21, Burakov, Anatoly:
> On 28-Mar-18 9:55 AM, Tan, Jianfeng wrote:
> > On 3/28/2018 4:22 PM, Van Haaren, Harry wrote:
> >> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> >>> 28/03/2018 04:08, Tan, Jianfeng:
> >>>> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> >>>>> 27/03/2018 15:59, Anatoly Burakov:
> >>>>>> Under the hood, we create a separate thread to deal with replies to
> >>>>>> asynchronous requests, that will just wait to be notified by the
> >>>>>> main thread, or woken up on a timer.
> >>>>>
> >>>>> I really don't like that a library is creating a thread.
> >>>>> We don't even know where the thread is created (which core).
> >>>>> Can it be a rte_service? or in the interrupt thread?
> >>>>
> >>>> Agree that we'd better not adding so many threads in a library.
> >>>>
> >>>> I was considering to merge all the threads into the interrupt thread,
> >>> however, we don't have an interrupt thread in freebsd. Further, we don't
> >>> implement alarm API in freebsd. That's why I tend to current
> >>> implementation,
> >>> and optimize it later.
> >>>
> >>> I would prefer we improve the current code now instead of polluting more
> >>> with more uncontrolled threads.
> >>>
> >>>> For rte_service, it may be not a good idea to reply on it as it needs
> >>> explicit API calls to setup.
> >>>
> >>> I don't see the issue of the explicit API.
> >>> The IPC is a new service.
> >
> > My concern is that not every DPDK application sets up rte_service, but
> > IPC will be used for very fundamental functions, like memory allocation.
> > We could not possibly ask all DPDK applications to add rte_service now.
> >
> > And also take Harry's comments below into consideration, most likely, we
> > will move these threads into interrupt thread now by adding
> >
> >> Although I do like to see new services, if we want to enable "core"
> >> dpdk functionality with Services, we need a proper designed solution
> >> for that. Service cores is not intended for "occasional" work - there
> >> is no method to block and sleep on a specific service until work
> >> becomes available, so this would imply a busy-polling. Using a service
> >> (hence busy polling) for rte_malloc()-based memory mapping requests is
> >> inefficient, and total overkill :)
> >>
> >> For this patch I suggest to use some blocking-read capable mechanism.
> >
> > The problem here is that we add too many threads; blocking-read does not
> > decrease # of threads.
> >
> >>
> >> The above said, in the longer term it would be good to have a design
> >> that allows new file-descriptors to be added to a "dpdk core" thread,
> >> which performs occasional lengthy work if the FD has data available.
> >
> > Interrupt thread vs rte_service, which is the direction to go? We
> > actually have some others threads, in vhost and even virtio-user; we can
> > also avoid those threads if we have a clear direction.
> >
> > Thanks,
> > Jianfeng
> >
>
> Hi all,
>
> First of all, @Thomas, this is not a "new library" - it's part of EAL.
I did not say it is a new library.
> We're going to be removing a few threads from EAL as it is because of
> IPC (Jianfeng has already submitted patches for those),
I don't understand.
Which threads are you going to remove? Which patch?
> so i don't think
> it's such a big deal to have two IPC threads instead of one. I'm open to
> suggestions on how to make this work without a second thread, but i
> don't see it.
I am not against the second thread.
I am against both threads :)
> We've discussed possibility of using rte_service internally, but decided
> against it for reasons already outlined by Harry - it's not a suitable
> mechanism for this kind of thing, not as it is.
>
> Using interrupt thread for this _will_ work, however this will require a
> a lot more changes, as currently alarm API allocates everything through
> rte_malloc, while we want to use IPC for rte_malloc (which would make it
> a circular dependency). So it'll probably be more API and more
> complexity for dealing with malloc vs rte_malloc allocations. Hence the
> least-bad approach taken here: a new thread.
If everybody is happy enough with "least bad" design and not trying
to improve the core design, what can I say?
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 9:53 ` Thomas Monjalon
@ 2018-03-28 10:42 ` Burakov, Anatoly
2018-03-28 11:26 ` Thomas Monjalon
0 siblings, 1 reply; 40+ messages in thread
From: Burakov, Anatoly @ 2018-03-28 10:42 UTC (permalink / raw)
To: Thomas Monjalon, Tan, Jianfeng
Cc: Van Haaren, Harry, dev, Ananyev, Konstantin
On 28-Mar-18 10:53 AM, Thomas Monjalon wrote:
> 28/03/2018 11:21, Burakov, Anatoly:
>> On 28-Mar-18 9:55 AM, Tan, Jianfeng wrote:
>>> On 3/28/2018 4:22 PM, Van Haaren, Harry wrote:
>>>> From: Thomas Monjalon [mailto:thomas@monjalon.net]
>>>>> 28/03/2018 04:08, Tan, Jianfeng:
>>>>>> From: Thomas Monjalon [mailto:thomas@monjalon.net]
>>>>>>> 27/03/2018 15:59, Anatoly Burakov:
>>>>>>>> Under the hood, we create a separate thread to deal with replies to
>>>>>>>> asynchronous requests, that will just wait to be notified by the
>>>>>>>> main thread, or woken up on a timer.
>>>>>>>
>>>>>>> I really don't like that a library is creating a thread.
>>>>>>> We don't even know where the thread is created (which core).
>>>>>>> Can it be a rte_service? or in the interrupt thread?
>>>>>>
>>>>>> Agree that we'd better not adding so many threads in a library.
>>>>>>
>>>>>> I was considering to merge all the threads into the interrupt thread,
>>>>> however, we don't have an interrupt thread in freebsd. Further, we don't
>>>>> implement alarm API in freebsd. That's why I tend to current
>>>>> implementation,
>>>>> and optimize it later.
>>>>>
>>>>> I would prefer we improve the current code now instead of polluting more
>>>>> with more uncontrolled threads.
>>>>>
>>>>>> For rte_service, it may be not a good idea to reply on it as it needs
>>>>> explicit API calls to setup.
>>>>>
>>>>> I don't see the issue of the explicit API.
>>>>> The IPC is a new service.
>>>
>>> My concern is that not every DPDK application sets up rte_service, but
>>> IPC will be used for very fundamental functions, like memory allocation.
>>> We could not possibly ask all DPDK applications to add rte_service now.
>>>
>>> And also take Harry's comments below into consideration, most likely, we
>>> will move these threads into interrupt thread now by adding
>>>
>>>> Although I do like to see new services, if we want to enable "core"
>>>> dpdk functionality with Services, we need a proper designed solution
>>>> for that. Service cores is not intended for "occasional" work - there
>>>> is no method to block and sleep on a specific service until work
>>>> becomes available, so this would imply a busy-polling. Using a service
>>>> (hence busy polling) for rte_malloc()-based memory mapping requests is
>>>> inefficient, and total overkill :)
>>>>
>>>> For this patch I suggest to use some blocking-read capable mechanism.
>>>
>>> The problem here is that we add too many threads; blocking-read does not
>>> decrease # of threads.
>>>
>>>>
>>>> The above said, in the longer term it would be good to have a design
>>>> that allows new file-descriptors to be added to a "dpdk core" thread,
>>>> which performs occasional lengthy work if the FD has data available.
>>>
>>> Interrupt thread vs rte_service, which is the direction to go? We
>>> actually have some others threads, in vhost and even virtio-user; we can
>>> also avoid those threads if we have a clear direction.
>>>
>>> Thanks,
>>> Jianfeng
>>>
>>
>> Hi all,
>>
>> First of all, @Thomas, this is not a "new library" - it's part of EAL.
>
> I did not say it is a new library.
>
>> We're going to be removing a few threads from EAL as it is because of
>> IPC (Jianfeng has already submitted patches for those),
>
> I don't understand.
> Which threads are you going to remove? Which patch?
There are separate patches that get rid of some EAL threads we have from
before (e.g. VFIO thread) - this is the reason IPC was created in the
first place. These aren't relevant to this patchset per se, i'm just
saying that it's not like we're adding to the pile of already existing
threads without taking anything away :)
>
>> so i don't think
>> it's such a big deal to have two IPC threads instead of one. I'm open to
>> suggestions on how to make this work without a second thread, but i
>> don't see it.
>
> I am not against the second thread.
> I am against both threads :)
Well, the first thread is already in DPDK. To provide some context,
first implementation for DPDK IPC was suggested for 17.11, and (without
many conceptual changes) was merged in 18.02. I think it's a bit late to
be against both threads :)
>
>> We've discussed possibility of using rte_service internally, but decided
>> against it for reasons already outlined by Harry - it's not a suitable
>> mechanism for this kind of thing, not as it is.
>>
>> Using interrupt thread for this _will_ work, however this will require a
>> a lot more changes, as currently alarm API allocates everything through
>> rte_malloc, while we want to use IPC for rte_malloc (which would make it
>> a circular dependency). So it'll probably be more API and more
>> complexity for dealing with malloc vs rte_malloc allocations. Hence the
>> least-bad approach taken here: a new thread.
>
> If everybody is happy enough with "least bad" design and not trying
> to improve the core design, what can I say?
I'm not against trying to improve the core design. I'm just saying that,
had this kind of feedback been provided just a bit earlier, I would've
had time to fix it in time for deadlines. However, because memory rework
patchset depends on this API, i would suggest merging it in now, as is,
and commit to a roadmap of improvements for next release(s).
For starters, we could plan on removing alarm thread's dependency on
rte_malloc and just use regular malloc API's in there, and rework
asynchronous IPC API to use that instead. This shouldn't be much work,
and will presumably make you halfway happy, as one of the threads will
be gone :)
We can then look into removing the second thread and moving the entirety
of DPDK IPC into the interrupt thread. I'm not too sure how would that
work, but i haven't looked at it in any detail, so maybe it is feasible.
Can we agree on this? It would be great to do everything perfectly from
the first try, but having a goal in sight and working towards it is fine
too, even if not all of the steps we take are perfect.
--
Thanks,
Anatoly
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 10:42 ` Burakov, Anatoly
@ 2018-03-28 11:26 ` Thomas Monjalon
2018-03-28 12:21 ` Burakov, Anatoly
0 siblings, 1 reply; 40+ messages in thread
From: Thomas Monjalon @ 2018-03-28 11:26 UTC (permalink / raw)
To: Burakov, Anatoly
Cc: Tan, Jianfeng, Van Haaren, Harry, dev, Ananyev, Konstantin,
bruce.richardson, olivier.matz
28/03/2018 12:42, Burakov, Anatoly:
> On 28-Mar-18 10:53 AM, Thomas Monjalon wrote:
> > 28/03/2018 11:21, Burakov, Anatoly:
> >> so i don't think
> >> it's such a big deal to have two IPC threads instead of one. I'm open to
> >> suggestions on how to make this work without a second thread, but i
> >> don't see it.
> >
> > I am not against the second thread.
> > I am against both threads :)
>
> Well, the first thread is already in DPDK. To provide some context,
> first implementation for DPDK IPC was suggested for 17.11, and (without
> many conceptual changes) was merged in 18.02. I think it's a bit late to
> be against both threads :)
No, it's never too late to discuss.
Merging a patch does not prevent discussing it :)
> >> We've discussed possibility of using rte_service internally, but decided
> >> against it for reasons already outlined by Harry - it's not a suitable
> >> mechanism for this kind of thing, not as it is.
> >>
> >> Using interrupt thread for this _will_ work, however this will require a
> >> a lot more changes, as currently alarm API allocates everything through
> >> rte_malloc, while we want to use IPC for rte_malloc (which would make it
> >> a circular dependency). So it'll probably be more API and more
> >> complexity for dealing with malloc vs rte_malloc allocations. Hence the
> >> least-bad approach taken here: a new thread.
> >
> > If everybody is happy enough with "least bad" design and not trying
> > to improve the core design, what can I say?
>
> I'm not against trying to improve the core design. I'm just saying that,
> had this kind of feedback been provided just a bit earlier, I would've
> had time to fix it in time for deadlines. However, because memory rework
> patchset depends on this API, i would suggest merging it in now, as is,
> and commit to a roadmap of improvements for next release(s).
Actually, you had the feedback yourself from the beginning.
You decided to gave up with interrupt thread because its implementation
is not complete (and maybe far from perfect).
There are some communities where it is not acceptable to workaround
core issues because of timing issues. I think we accept it in DPDK,
but I continue to question it, in order to be sure that everybody is OK
with this kind of tradeoff.
> For starters, we could plan on removing alarm thread's dependency on
> rte_malloc and just use regular malloc API's in there, and rework
> asynchronous IPC API to use that instead. This shouldn't be much work,
> and will presumably make you halfway happy, as one of the threads will
> be gone :)
>
> We can then look into removing the second thread and moving the entirety
> of DPDK IPC into the interrupt thread. I'm not too sure how would that
> work, but i haven't looked at it in any detail, so maybe it is feasible.
>
> Can we agree on this? It would be great to do everything perfectly from
> the first try, but having a goal in sight and working towards it is fine
> too, even if not all of the steps we take are perfect.
The main concern is API.
If all these changes are internal only, and does not involve any major
API change, then I guess it is OK to pospone them in next release.
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 11:26 ` Thomas Monjalon
@ 2018-03-28 12:21 ` Burakov, Anatoly
0 siblings, 0 replies; 40+ messages in thread
From: Burakov, Anatoly @ 2018-03-28 12:21 UTC (permalink / raw)
To: Thomas Monjalon
Cc: Tan, Jianfeng, Van Haaren, Harry, dev, Ananyev, Konstantin,
bruce.richardson, olivier.matz
On 28-Mar-18 12:26 PM, Thomas Monjalon wrote:
> 28/03/2018 12:42, Burakov, Anatoly:
>> On 28-Mar-18 10:53 AM, Thomas Monjalon wrote:
>>> 28/03/2018 11:21, Burakov, Anatoly:
>> I'm not against trying to improve the core design. I'm just saying that,
>> had this kind of feedback been provided just a bit earlier, I would've
>> had time to fix it in time for deadlines. However, because memory rework
>> patchset depends on this API, i would suggest merging it in now, as is,
>> and commit to a roadmap of improvements for next release(s).
>
> Actually, you had the feedback yourself from the beginning.
> You decided to gave up with interrupt thread because its implementation
> is not complete (and maybe far from perfect).
That's not quite how i see it, but OK, suppose so.
> There are some communities where it is not acceptable to workaround
> core issues because of timing issues. I think we accept it in DPDK,
> but I continue to question it, in order to be sure that everybody is OK
> with this kind of tradeoff.
The way i see it, not all API's are equal; some are more important than
others. This is a new, experimental API that is not core to any DPDK
function - it's not used on any hotpaths nor is it even that demanding
(the two threads will be sleeping 99.999% of the time anyway). I think
we're allowed to experiment on it before settling on an implementation
that satisfies everyone :)
>> For starters, we could plan on removing alarm thread's dependency on
>> rte_malloc and just use regular malloc API's in there, and rework
>> asynchronous IPC API to use that instead. This shouldn't be much work,
>> and will presumably make you halfway happy, as one of the threads will
>> be gone :)
>>
>> We can then look into removing the second thread and moving the entirety
>> of DPDK IPC into the interrupt thread. I'm not too sure how would that
>> work, but i haven't looked at it in any detail, so maybe it is feasible.
>>
>> Can we agree on this? It would be great to do everything perfectly from
>> the first try, but having a goal in sight and working towards it is fine
>> too, even if not all of the steps we take are perfect.
>
> The main concern is API.
> If all these changes are internal only, and does not involve any major
> API change, then I guess it is OK to pospone them in next release.
>
Yes, all of this is/will be internal to DPDK IPC - no externally visible
changes whatsoever.
--
Thanks,
Anatoly
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v6 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-28 7:29 ` Thomas Monjalon
2018-03-28 8:22 ` Van Haaren, Harry
@ 2018-03-28 9:11 ` Bruce Richardson
1 sibling, 0 replies; 40+ messages in thread
From: Bruce Richardson @ 2018-03-28 9:11 UTC (permalink / raw)
To: Thomas Monjalon
Cc: Tan, Jianfeng, Burakov, Anatoly, dev, Ananyev, Konstantin,
Van Haaren, Harry
On Wed, Mar 28, 2018 at 09:29:35AM +0200, Thomas Monjalon wrote:
> 28/03/2018 04:08, Tan, Jianfeng:
> > Hi Thomas ,
> >
> > From: Thomas Monjalon [mailto:thomas@monjalon.net]
> > > 27/03/2018 15:59, Anatoly Burakov:
> > > > Under the hood, we create a separate thread to deal with replies to
> > > > asynchronous requests, that will just wait to be notified by the
> > > > main thread, or woken up on a timer.
> > >
> > > I really don't like that a library is creating a thread.
> > > We don't even know where the thread is created (which core).
> > > Can it be a rte_service? or in the interrupt thread?
> >
> > Agree that we'd better not adding so many threads in a library.
> >
> > I was considering to merge all the threads into the interrupt thread, however, we don't have an interrupt thread in freebsd. Further, we don't implement alarm API in freebsd. That's why I tend to current implementation, and optimize it later.
>
> I would prefer we improve the current code now instead of polluting more
> with more uncontrolled threads.
>
+1
I think it would be worthwhile adding an interrupt thread to BSD, and it
should not be a massive amount of work. Having a single interrupt thread
has a lot of benefits, I think.
/Bruce
^ permalink raw reply [flat|nested] 40+ messages in thread
* [dpdk-dev] [PATCH v5 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-13 17:42 ` [dpdk-dev] [PATCH v4] " Anatoly Burakov
2018-03-23 15:38 ` Tan, Jianfeng
2018-03-24 12:46 ` [dpdk-dev] [PATCH v5 1/2] eal: rename IPC sync request to pending request Anatoly Burakov
@ 2018-03-24 12:46 ` Anatoly Burakov
2018-03-26 14:15 ` Tan, Jianfeng
2 siblings, 1 reply; 40+ messages in thread
From: Anatoly Burakov @ 2018-03-24 12:46 UTC (permalink / raw)
To: dev; +Cc: failed, to, remove, 'kernel'
This API is similar to the blocking API that is already present,
but reply will be received in a separate callback by the caller
(callback specified at the time of request, rather than registering
for it in advance).
Under the hood, we create a separate thread to deal with replies to
asynchronous requests, that will just wait to be notified by the
main thread, or woken up on a timer.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
---
Notes:
v5:
- addressed review comments from Jianfeng
- split into two patches to avoid rename noise
- do not mark ignored message as processed
v4:
- rebase on top of latest IPC Improvements patchset [2]
v3:
- added support for MP_IGN messages introduced in
IPC improvements v5 patchset
v2:
- fixed deadlocks and race conditions by not calling
callbacks while iterating over sync request list
- fixed use-after-free by making a copy of request
- changed API to also give user a copy of original
request, so that they know to which message the
callback is a reply to
- fixed missing .map file entries
This patch is dependent upon previously published patchsets
for IPC fixes [1] and improvements [2].
rte_mp_action_unregister and rte_mp_async_reply_unregister
do the same thing - should we perhaps make it one function?
[1] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Fixes/
[2] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Improvements/
lib/librte_eal/common/eal_common_proc.c | 455 +++++++++++++++++++++++++++++++-
lib/librte_eal/common/include/rte_eal.h | 36 +++
lib/librte_eal/rte_eal_version.map | 1 +
3 files changed, 479 insertions(+), 13 deletions(-)
diff --git a/lib/librte_eal/common/eal_common_proc.c b/lib/librte_eal/common/eal_common_proc.c
index 52b6ab2..c86252c 100644
--- a/lib/librte_eal/common/eal_common_proc.c
+++ b/lib/librte_eal/common/eal_common_proc.c
@@ -26,6 +26,7 @@
#include <rte_errno.h>
#include <rte_lcore.h>
#include <rte_log.h>
+#include <rte_tailq.h>
#include "eal_private.h"
#include "eal_filesystem.h"
@@ -60,13 +61,32 @@ struct mp_msg_internal {
struct rte_mp_msg msg;
};
+struct async_request_param {
+ rte_mp_async_reply_t clb;
+ struct rte_mp_reply user_reply;
+ struct timespec end;
+ int n_responses_processed;
+};
+
struct pending_request {
TAILQ_ENTRY(pending_request) next;
- int reply_received;
+ enum {
+ REQUEST_TYPE_SYNC,
+ REQUEST_TYPE_ASYNC
+ } type;
char dst[PATH_MAX];
struct rte_mp_msg *request;
struct rte_mp_msg *reply;
- pthread_cond_t cond;
+ int reply_received;
+ RTE_STD_C11
+ union {
+ struct {
+ struct async_request_param *param;
+ } async;
+ struct {
+ pthread_cond_t cond;
+ } sync;
+ };
};
TAILQ_HEAD(pending_request_list, pending_request);
@@ -74,9 +94,12 @@ TAILQ_HEAD(pending_request_list, pending_request);
static struct {
struct pending_request_list requests;
pthread_mutex_t lock;
+ pthread_cond_t async_cond;
} pending_requests = {
.requests = TAILQ_HEAD_INITIALIZER(pending_requests.requests),
- .lock = PTHREAD_MUTEX_INITIALIZER
+ .lock = PTHREAD_MUTEX_INITIALIZER,
+ .async_cond = PTHREAD_COND_INITIALIZER
+ /**< used in async requests only */
};
/* forward declarations */
@@ -273,7 +296,12 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
memcpy(sync_req->reply, msg, sizeof(*msg));
/* -1 indicates that we've been asked to ignore */
sync_req->reply_received = m->type == MP_REP ? 1 : -1;
- pthread_cond_signal(&sync_req->cond);
+
+ if (sync_req->type == REQUEST_TYPE_SYNC)
+ pthread_cond_signal(&sync_req->sync.cond);
+ else if (sync_req->type == REQUEST_TYPE_ASYNC)
+ pthread_cond_signal(
+ &pending_requests.async_cond);
} else
RTE_LOG(ERR, EAL, "Drop mp reply: %s\n", msg->name);
pthread_mutex_unlock(&pending_requests.lock);
@@ -320,6 +348,189 @@ mp_handle(void *arg __rte_unused)
}
static int
+timespec_cmp(const struct timespec *a, const struct timespec *b)
+{
+ if (a->tv_sec < b->tv_sec)
+ return -1;
+ if (a->tv_sec > b->tv_sec)
+ return 1;
+ if (a->tv_nsec < b->tv_nsec)
+ return -1;
+ if (a->tv_nsec > b->tv_nsec)
+ return 1;
+ return 0;
+}
+
+enum async_action {
+ ACTION_NONE, /**< don't do anything */
+ ACTION_FREE, /**< free the action entry, but don't trigger callback */
+ ACTION_TRIGGER /**< trigger callback, then free action entry */
+};
+
+static enum async_action
+process_async_request(struct pending_request *sr, const struct timespec *now)
+{
+ struct async_request_param *param;
+ struct rte_mp_reply *reply;
+ bool timeout, received, last_msg;
+
+ param = sr->async.param;
+ reply = ¶m->user_reply;
+
+ /* did we timeout? */
+ timeout = timespec_cmp(¶m->end, now) <= 0;
+
+ /* did we receive a response? */
+ received = sr->reply_received != 0;
+
+ /* if we didn't time out, and we didn't receive a response, ignore */
+ if (!timeout && !received)
+ return ACTION_NONE;
+
+ /* if we received a response, adjust relevant data and copy mesasge. */
+ if (sr->reply_received == 1 && sr->reply) {
+ struct rte_mp_msg *msg, *user_msgs, *tmp;
+
+ msg = sr->reply;
+ user_msgs = reply->msgs;
+
+ tmp = realloc(user_msgs, sizeof(*msg) *
+ (reply->nb_received + 1));
+ if (!tmp) {
+ RTE_LOG(ERR, EAL, "Fail to alloc reply for request %s:%s\n",
+ sr->dst, sr->request->name);
+ /* this entry is going to be removed and its message
+ * dropped, but we don't want to leak memory, so
+ * continue.
+ */
+ } else {
+ user_msgs = tmp;
+ reply->msgs = user_msgs;
+ memcpy(&user_msgs[reply->nb_received],
+ msg, sizeof(*msg));
+ reply->nb_received++;
+ }
+
+ /* mark this request as processed */
+ param->n_responses_processed++;
+ } else if (sr->reply_received == -1) {
+ /* we were asked to ignore this process */
+ reply->nb_sent--;
+ }
+ free(sr->reply);
+
+ last_msg = param->n_responses_processed == reply->nb_sent;
+
+ return last_msg ? ACTION_TRIGGER : ACTION_FREE;
+}
+
+static void
+trigger_async_action(struct pending_request *sr)
+{
+ struct async_request_param *param;
+ struct rte_mp_reply *reply;
+
+ param = sr->async.param;
+ reply = ¶m->user_reply;
+
+ param->clb(sr->request, reply);
+
+ /* clean up */
+ free(sr->async.param->user_reply.msgs);
+ free(sr->async.param);
+ free(sr->request);
+}
+
+static void *
+async_reply_handle(void *arg __rte_unused)
+{
+ struct pending_request *sr;
+ struct timeval now;
+ struct timespec timeout, ts_now;
+ while (1) {
+ struct pending_request *trigger = NULL;
+ int ret;
+ bool nowait = false;
+ bool timedwait = false;
+
+ pthread_mutex_lock(&pending_requests.lock);
+
+ /* scan through the list and see if there are any timeouts that
+ * are earlier than our current timeout.
+ */
+ TAILQ_FOREACH(sr, &pending_requests.requests, next) {
+ if (sr->type != REQUEST_TYPE_ASYNC)
+ continue;
+ if (!timedwait || timespec_cmp(&sr->async.param->end,
+ &timeout) < 0) {
+ memcpy(&timeout, &sr->async.param->end,
+ sizeof(timeout));
+ timedwait = true;
+ }
+
+ /* sometimes, we don't even wait */
+ if (sr->reply_received) {
+ nowait = true;
+ break;
+ }
+ }
+
+ if (nowait)
+ ret = 0;
+ else if (timedwait)
+ ret = pthread_cond_timedwait(
+ &pending_requests.async_cond,
+ &pending_requests.lock, &timeout);
+ else
+ ret = pthread_cond_wait(&pending_requests.async_cond,
+ &pending_requests.lock);
+
+ if (gettimeofday(&now, NULL) < 0) {
+ RTE_LOG(ERR, EAL, "Cannot get current time\n");
+ break;
+ }
+ ts_now.tv_nsec = now.tv_usec * 1000;
+ ts_now.tv_sec = now.tv_sec;
+
+ if (ret == 0 || ret == ETIMEDOUT) {
+ struct pending_request *next;
+ /* we've either been woken up, or we timed out */
+
+ /* we have still the lock, check if anything needs
+ * processing.
+ */
+ TAILQ_FOREACH_SAFE(sr, &pending_requests.requests, next,
+ next) {
+ enum async_action action;
+ if (sr->type != REQUEST_TYPE_ASYNC)
+ continue;
+
+ action = process_async_request(sr, &ts_now);
+ if (action == ACTION_FREE) {
+ TAILQ_REMOVE(&pending_requests.requests,
+ sr, next);
+ free(sr);
+ } else if (action == ACTION_TRIGGER &&
+ trigger == NULL) {
+ TAILQ_REMOVE(&pending_requests.requests,
+ sr, next);
+ trigger = sr;
+ }
+ }
+ }
+ pthread_mutex_unlock(&pending_requests.lock);
+ if (trigger) {
+ trigger_async_action(trigger);
+ free(trigger);
+ }
+ };
+
+ RTE_LOG(ERR, EAL, "ERROR: asynchronous requests disabled\n");
+
+ return NULL;
+}
+
+static int
open_socket_fd(void)
{
char peer_name[PATH_MAX] = {0};
@@ -382,7 +593,7 @@ rte_mp_channel_init(void)
char thread_name[RTE_MAX_THREAD_NAME_LEN];
char path[PATH_MAX];
int dir_fd;
- pthread_t tid;
+ pthread_t mp_handle_tid, async_reply_handle_tid;
/* create filter path */
create_socket_path("*", path, sizeof(path));
@@ -419,7 +630,16 @@ rte_mp_channel_init(void)
return -1;
}
- if (pthread_create(&tid, NULL, mp_handle, NULL) < 0) {
+ if (pthread_create(&mp_handle_tid, NULL, mp_handle, NULL) < 0) {
+ RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
+ strerror(errno));
+ close(mp_fd);
+ mp_fd = -1;
+ return -1;
+ }
+
+ if (pthread_create(&async_reply_handle_tid, NULL,
+ async_reply_handle, NULL) < 0) {
RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
strerror(errno));
close(mp_fd);
@@ -430,7 +650,11 @@ rte_mp_channel_init(void)
/* try best to set thread name */
snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_handle");
- rte_thread_setname(tid, thread_name);
+ rte_thread_setname(mp_handle_tid, thread_name);
+
+ /* try best to set thread name */
+ snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_async_handle");
+ rte_thread_setname(async_reply_handle_tid, thread_name);
/* unlock the directory */
flock(dir_fd, LOCK_UN);
@@ -602,18 +826,77 @@ rte_mp_sendmsg(struct rte_mp_msg *msg)
}
static int
-mp_request_one(const char *dst, struct rte_mp_msg *req,
+mp_request_async(const char *dst, struct rte_mp_msg *req,
+ struct async_request_param *param)
+{
+ struct rte_mp_msg *reply_msg;
+ struct pending_request *sync_req, *exist;
+ int ret;
+
+ sync_req = malloc(sizeof(*sync_req));
+ reply_msg = malloc(sizeof(*reply_msg));
+ if (sync_req == NULL || reply_msg == NULL) {
+ RTE_LOG(ERR, EAL, "Could not allocate space for sync request\n");
+ rte_errno = ENOMEM;
+ ret = -1;
+ goto fail;
+ }
+
+ memset(sync_req, 0, sizeof(*sync_req));
+ memset(reply_msg, 0, sizeof(*reply_msg));
+
+ sync_req->type = REQUEST_TYPE_ASYNC;
+ strcpy(sync_req->dst, dst);
+ sync_req->request = req;
+ sync_req->reply = reply_msg;
+ sync_req->async.param = param;
+
+ /* queue already locked by caller */
+
+ exist = find_sync_request(dst, req->name);
+ if (!exist)
+ TAILQ_INSERT_TAIL(&pending_requests.requests, sync_req, next);
+ if (exist) {
+ RTE_LOG(ERR, EAL, "A pending request %s:%s\n", dst, req->name);
+ rte_errno = EEXIST;
+ ret = -1;
+ goto fail;
+ }
+
+ ret = send_msg(dst, req, MP_REQ);
+ if (ret < 0) {
+ RTE_LOG(ERR, EAL, "Fail to send request %s:%s\n",
+ dst, req->name);
+ ret = -1;
+ goto fail;
+ } else if (ret == 0) {
+ ret = 0;
+ goto fail;
+ }
+
+ param->user_reply.nb_sent++;
+
+ return 0;
+fail:
+ free(sync_req);
+ free(reply_msg);
+ return ret;
+}
+
+static int
+mp_request_sync(const char *dst, struct rte_mp_msg *req,
struct rte_mp_reply *reply, const struct timespec *ts)
{
int ret;
struct rte_mp_msg msg, *tmp;
struct pending_request sync_req, *exist;
+ sync_req.type = REQUEST_TYPE_SYNC;
sync_req.reply_received = 0;
strcpy(sync_req.dst, dst);
sync_req.request = req;
sync_req.reply = &msg;
- pthread_cond_init(&sync_req.cond, NULL);
+ pthread_cond_init(&sync_req.sync.cond, NULL);
pthread_mutex_lock(&pending_requests.lock);
exist = find_sync_request(dst, req->name);
@@ -637,7 +920,7 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
reply->nb_sent++;
do {
- ret = pthread_cond_timedwait(&sync_req.cond,
+ ret = pthread_cond_timedwait(&sync_req.sync.cond,
&pending_requests.lock, ts);
} while (ret != 0 && ret != ETIMEDOUT);
@@ -703,7 +986,7 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
/* for secondary process, send request to the primary process only */
if (rte_eal_process_type() == RTE_PROC_SECONDARY)
- return mp_request_one(eal_mp_socket_path(), req, reply, &end);
+ return mp_request_sync(eal_mp_socket_path(), req, reply, &end);
/* for primary process, broadcast request, and collect reply 1 by 1 */
mp_dir = opendir(mp_dir_path);
@@ -732,7 +1015,7 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
ent->d_name);
- if (mp_request_one(path, req, reply, &end))
+ if (mp_request_sync(path, req, reply, &end))
ret = -1;
}
/* unlock the directory */
@@ -744,9 +1027,155 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
}
int __rte_experimental
-rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
+rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts,
+ rte_mp_async_reply_t clb)
{
+ struct rte_mp_msg *copy;
+ struct pending_request *dummy;
+ struct async_request_param *param = NULL;
+ struct rte_mp_reply *reply;
+ int dir_fd, ret = 0;
+ DIR *mp_dir;
+ struct dirent *ent;
+ struct timeval now;
+ struct timespec *end;
+
+ RTE_LOG(DEBUG, EAL, "request: %s\n", req->name);
+
+ if (check_input(req) == false)
+ return -1;
+ if (gettimeofday(&now, NULL) < 0) {
+ RTE_LOG(ERR, EAL, "Faile to get current time\n");
+ rte_errno = errno;
+ return -1;
+ }
+ copy = malloc(sizeof(*copy));
+ dummy = malloc(sizeof(*dummy));
+ param = malloc(sizeof(*param));
+ if (copy == NULL || dummy == NULL || param == NULL) {
+ RTE_LOG(ERR, EAL, "Failed to allocate memory for async reply\n");
+ rte_errno = ENOMEM;
+ goto fail;
+ }
+
+ memset(copy, 0, sizeof(*copy));
+ memset(dummy, 0, sizeof(*dummy));
+ memset(param, 0, sizeof(*param));
+
+ /* copy message */
+ memcpy(copy, req, sizeof(*copy));
+
+ param->n_responses_processed = 0;
+ param->clb = clb;
+ end = ¶m->end;
+ reply = ¶m->user_reply;
+
+ end->tv_nsec = (now.tv_usec * 1000 + ts->tv_nsec) % 1000000000;
+ end->tv_sec = now.tv_sec + ts->tv_sec +
+ (now.tv_usec * 1000 + ts->tv_nsec) / 1000000000;
+ reply->nb_sent = 0;
+ reply->nb_received = 0;
+ reply->msgs = NULL;
+ /* we have to lock the request queue here, as we will be adding a bunch
+ * of requests to the queue at once, and some of the replies may arrive
+ * before we add all of the requests to the queue.
+ */
+ pthread_mutex_lock(&pending_requests.lock);
+
+ /* we have to ensure that callback gets triggered even if we don't send
+ * anything, therefore earlier we have allocated a dummy request. put it
+ * on the queue and fill it. we will remove it once we know we sent
+ * something.
+ */
+ dummy->type = REQUEST_TYPE_ASYNC;
+ dummy->request = copy;
+ dummy->reply = NULL;
+ dummy->async.param = param;
+ dummy->reply_received = 1; /* short-circuit the timeout */
+
+ TAILQ_INSERT_TAIL(&pending_requests.requests, dummy, next);
+
+ /* for secondary process, send request to the primary process only */
+ if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
+ ret = mp_request_async(eal_mp_socket_path(), copy, param);
+
+ /* if we sent something, remove dummy request from the queue */
+ if (reply->nb_sent != 0) {
+ TAILQ_REMOVE(&pending_requests.requests, dummy, next);
+ free(dummy);
+ dummy = NULL;
+ }
+
+ pthread_mutex_unlock(&pending_requests.lock);
+
+ /* if we couldn't send anything, clean up */
+ if (ret != 0)
+ goto fail;
+ return 0;
+ }
+
+ /* for primary process, broadcast request */
+ mp_dir = opendir(mp_dir_path);
+ if (!mp_dir) {
+ RTE_LOG(ERR, EAL, "Unable to open directory %s\n", mp_dir_path);
+ rte_errno = errno;
+ goto unlock_fail;
+ }
+ dir_fd = dirfd(mp_dir);
+
+ /* lock the directory to prevent processes spinning up while we send */
+ if (flock(dir_fd, LOCK_EX)) {
+ RTE_LOG(ERR, EAL, "Unable to lock directory %s\n",
+ mp_dir_path);
+ rte_errno = errno;
+ goto closedir_fail;
+ }
+
+ while ((ent = readdir(mp_dir))) {
+ char path[PATH_MAX];
+
+ if (fnmatch(mp_filter, ent->d_name, 0) != 0)
+ continue;
+
+ snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
+ ent->d_name);
+
+ if (mp_request_async(path, copy, param))
+ ret = -1;
+ }
+ /* if we sent something, remove dummy request from the queue */
+ if (reply->nb_sent != 0) {
+ TAILQ_REMOVE(&pending_requests.requests, dummy, next);
+ free(dummy);
+ dummy = NULL;
+ }
+ /* trigger async request thread wake up */
+ pthread_cond_signal(&pending_requests.async_cond);
+
+ /* finally, unlock the queue */
+ pthread_mutex_unlock(&pending_requests.lock);
+
+ /* unlock the directory */
+ flock(dir_fd, LOCK_UN);
+
+ /* dir_fd automatically closed on closedir */
+ closedir(mp_dir);
+ return ret;
+closedir_fail:
+ closedir(mp_dir);
+unlock_fail:
+ pthread_mutex_unlock(&pending_requests.lock);
+fail:
+ free(dummy);
+ free(param);
+ free(copy);
+ return -1;
+}
+
+int __rte_experimental
+rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
+{
RTE_LOG(DEBUG, EAL, "reply: %s\n", msg->name);
if (check_input(msg) == false)
diff --git a/lib/librte_eal/common/include/rte_eal.h b/lib/librte_eal/common/include/rte_eal.h
index 044474e..87ebfd0 100644
--- a/lib/librte_eal/common/include/rte_eal.h
+++ b/lib/librte_eal/common/include/rte_eal.h
@@ -230,6 +230,16 @@ struct rte_mp_reply {
typedef int (*rte_mp_t)(const struct rte_mp_msg *msg, const void *peer);
/**
+ * Asynchronous reply function typedef used by other components.
+ *
+ * As we create socket channel for primary/secondary communication, use
+ * this function typedef to register action for coming responses to asynchronous
+ * requests.
+ */
+typedef int (*rte_mp_async_reply_t)(const struct rte_mp_msg *request,
+ const struct rte_mp_reply *reply);
+
+/**
* @warning
* @b EXPERIMENTAL: this API may change without prior notice
*
@@ -321,6 +331,32 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
* @warning
* @b EXPERIMENTAL: this API may change without prior notice
*
+ * Send a request to the peer process and expect a reply in a separate callback.
+ *
+ * This function sends a request message to the peer process, and will not
+ * block. Instead, reply will be received in a separate callback.
+ *
+ * @param req
+ * The req argument contains the customized request message.
+ *
+ * @param ts
+ * The ts argument specifies how long we can wait for the peer(s) to reply.
+ *
+ * @param clb
+ * The callback to trigger when all responses for this request have arrived.
+ *
+ * @return
+ * - On success, return 0.
+ * - On failure, return -1, and the reason will be stored in rte_errno.
+ */
+int __rte_experimental
+rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts,
+ rte_mp_async_reply_t clb);
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice
+ *
* Send a reply to the peer process.
*
* This function will send a reply message in response to a request message
diff --git a/lib/librte_eal/rte_eal_version.map b/lib/librte_eal/rte_eal_version.map
index d123602..328a0be 100644
--- a/lib/librte_eal/rte_eal_version.map
+++ b/lib/librte_eal/rte_eal_version.map
@@ -225,6 +225,7 @@ EXPERIMENTAL {
rte_mp_action_unregister;
rte_mp_sendmsg;
rte_mp_request;
+ rte_mp_request_async;
rte_mp_reply;
rte_service_attr_get;
rte_service_attr_reset_all;
--
2.7.4
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v5 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-24 12:46 ` [dpdk-dev] [PATCH v5 " Anatoly Burakov
@ 2018-03-26 14:15 ` Tan, Jianfeng
2018-03-26 14:28 ` Burakov, Anatoly
0 siblings, 1 reply; 40+ messages in thread
From: Tan, Jianfeng @ 2018-03-26 14:15 UTC (permalink / raw)
To: Anatoly Burakov, dev
Cc: failed, to, remove, Directory, not, empty, konstantin.ananyev
On 3/24/2018 8:46 PM, Anatoly Burakov wrote:
> This API is similar to the blocking API that is already present,
> but reply will be received in a separate callback by the caller
> (callback specified at the time of request, rather than registering
> for it in advance).
>
> Under the hood, we create a separate thread to deal with replies to
> asynchronous requests, that will just wait to be notified by the
> main thread, or woken up on a timer.
>
> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Generally, it looks great to me except some trivial nits, so
Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
> ---
>
> Notes:
> v5:
> - addressed review comments from Jianfeng
> - split into two patches to avoid rename noise
> - do not mark ignored message as processed
> v4:
> - rebase on top of latest IPC Improvements patchset [2]
>
> v3:
> - added support for MP_IGN messages introduced in
> IPC improvements v5 patchset
> v2:
> - fixed deadlocks and race conditions by not calling
> callbacks while iterating over sync request list
> - fixed use-after-free by making a copy of request
> - changed API to also give user a copy of original
> request, so that they know to which message the
> callback is a reply to
> - fixed missing .map file entries
>
> This patch is dependent upon previously published patchsets
> for IPC fixes [1] and improvements [2].
>
> rte_mp_action_unregister and rte_mp_async_reply_unregister
> do the same thing - should we perhaps make it one function?
>
> [1] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Fixes/
> [2] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Improvements/
>
> lib/librte_eal/common/eal_common_proc.c | 455 +++++++++++++++++++++++++++++++-
> lib/librte_eal/common/include/rte_eal.h | 36 +++
> lib/librte_eal/rte_eal_version.map | 1 +
> 3 files changed, 479 insertions(+), 13 deletions(-)
>
> diff --git a/lib/librte_eal/common/eal_common_proc.c b/lib/librte_eal/common/eal_common_proc.c
> index 52b6ab2..c86252c 100644
> --- a/lib/librte_eal/common/eal_common_proc.c
> +++ b/lib/librte_eal/common/eal_common_proc.c
> @@ -26,6 +26,7 @@
> #include <rte_errno.h>
> #include <rte_lcore.h>
> #include <rte_log.h>
> +#include <rte_tailq.h>
>
> #include "eal_private.h"
> #include "eal_filesystem.h"
> @@ -60,13 +61,32 @@ struct mp_msg_internal {
> struct rte_mp_msg msg;
> };
>
> +struct async_request_param {
> + rte_mp_async_reply_t clb;
> + struct rte_mp_reply user_reply;
> + struct timespec end;
> + int n_responses_processed;
> +};
> +
> struct pending_request {
> TAILQ_ENTRY(pending_request) next;
> - int reply_received;
> + enum {
> + REQUEST_TYPE_SYNC,
> + REQUEST_TYPE_ASYNC
> + } type;
> char dst[PATH_MAX];
> struct rte_mp_msg *request;
> struct rte_mp_msg *reply;
> - pthread_cond_t cond;
> + int reply_received;
> + RTE_STD_C11
> + union {
> + struct {
> + struct async_request_param *param;
> + } async;
> + struct {
> + pthread_cond_t cond;
> + } sync;
> + };
> };
>
> TAILQ_HEAD(pending_request_list, pending_request);
> @@ -74,9 +94,12 @@ TAILQ_HEAD(pending_request_list, pending_request);
> static struct {
> struct pending_request_list requests;
> pthread_mutex_t lock;
> + pthread_cond_t async_cond;
> } pending_requests = {
> .requests = TAILQ_HEAD_INITIALIZER(pending_requests.requests),
> - .lock = PTHREAD_MUTEX_INITIALIZER
> + .lock = PTHREAD_MUTEX_INITIALIZER,
> + .async_cond = PTHREAD_COND_INITIALIZER
> + /**< used in async requests only */
> };
>
> /* forward declarations */
> @@ -273,7 +296,12 @@ process_msg(struct mp_msg_internal *m, struct sockaddr_un *s)
> memcpy(sync_req->reply, msg, sizeof(*msg));
> /* -1 indicates that we've been asked to ignore */
> sync_req->reply_received = m->type == MP_REP ? 1 : -1;
> - pthread_cond_signal(&sync_req->cond);
> +
> + if (sync_req->type == REQUEST_TYPE_SYNC)
> + pthread_cond_signal(&sync_req->sync.cond);
> + else if (sync_req->type == REQUEST_TYPE_ASYNC)
> + pthread_cond_signal(
> + &pending_requests.async_cond);
> } else
> RTE_LOG(ERR, EAL, "Drop mp reply: %s\n", msg->name);
> pthread_mutex_unlock(&pending_requests.lock);
> @@ -320,6 +348,189 @@ mp_handle(void *arg __rte_unused)
> }
>
> static int
> +timespec_cmp(const struct timespec *a, const struct timespec *b)
> +{
> + if (a->tv_sec < b->tv_sec)
> + return -1;
> + if (a->tv_sec > b->tv_sec)
> + return 1;
> + if (a->tv_nsec < b->tv_nsec)
> + return -1;
> + if (a->tv_nsec > b->tv_nsec)
> + return 1;
> + return 0;
> +}
> +
> +enum async_action {
> + ACTION_NONE, /**< don't do anything */
> + ACTION_FREE, /**< free the action entry, but don't trigger callback */
> + ACTION_TRIGGER /**< trigger callback, then free action entry */
> +};
> +
> +static enum async_action
> +process_async_request(struct pending_request *sr, const struct timespec *now)
> +{
> + struct async_request_param *param;
> + struct rte_mp_reply *reply;
> + bool timeout, received, last_msg;
> +
> + param = sr->async.param;
> + reply = ¶m->user_reply;
> +
> + /* did we timeout? */
> + timeout = timespec_cmp(¶m->end, now) <= 0;
> +
> + /* did we receive a response? */
> + received = sr->reply_received != 0;
> +
> + /* if we didn't time out, and we didn't receive a response, ignore */
> + if (!timeout && !received)
> + return ACTION_NONE;
> +
> + /* if we received a response, adjust relevant data and copy mesasge. */
> + if (sr->reply_received == 1 && sr->reply) {
> + struct rte_mp_msg *msg, *user_msgs, *tmp;
> +
> + msg = sr->reply;
> + user_msgs = reply->msgs;
> +
> + tmp = realloc(user_msgs, sizeof(*msg) *
> + (reply->nb_received + 1));
> + if (!tmp) {
> + RTE_LOG(ERR, EAL, "Fail to alloc reply for request %s:%s\n",
> + sr->dst, sr->request->name);
> + /* this entry is going to be removed and its message
> + * dropped, but we don't want to leak memory, so
> + * continue.
> + */
> + } else {
> + user_msgs = tmp;
> + reply->msgs = user_msgs;
> + memcpy(&user_msgs[reply->nb_received],
> + msg, sizeof(*msg));
> + reply->nb_received++;
> + }
> +
> + /* mark this request as processed */
> + param->n_responses_processed++;
> + } else if (sr->reply_received == -1) {
> + /* we were asked to ignore this process */
> + reply->nb_sent--;
> + }
> + free(sr->reply);
> +
> + last_msg = param->n_responses_processed == reply->nb_sent;
> +
> + return last_msg ? ACTION_TRIGGER : ACTION_FREE;
> +}
> +
> +static void
> +trigger_async_action(struct pending_request *sr)
> +{
> + struct async_request_param *param;
> + struct rte_mp_reply *reply;
> +
> + param = sr->async.param;
> + reply = ¶m->user_reply;
> +
> + param->clb(sr->request, reply);
> +
> + /* clean up */
> + free(sr->async.param->user_reply.msgs);
How about simple "free(reply->msgs);"?
> + free(sr->async.param);
> + free(sr->request);
> +}
> +
> +static void *
> +async_reply_handle(void *arg __rte_unused)
> +{
> + struct pending_request *sr;
> + struct timeval now;
> + struct timespec timeout, ts_now;
> + while (1) {
> + struct pending_request *trigger = NULL;
> + int ret;
> + bool nowait = false;
> + bool timedwait = false;
> +
> + pthread_mutex_lock(&pending_requests.lock);
> +
> + /* scan through the list and see if there are any timeouts that
> + * are earlier than our current timeout.
> + */
> + TAILQ_FOREACH(sr, &pending_requests.requests, next) {
> + if (sr->type != REQUEST_TYPE_ASYNC)
> + continue;
> + if (!timedwait || timespec_cmp(&sr->async.param->end,
> + &timeout) < 0) {
> + memcpy(&timeout, &sr->async.param->end,
> + sizeof(timeout));
> + timedwait = true;
> + }
> +
> + /* sometimes, we don't even wait */
> + if (sr->reply_received) {
> + nowait = true;
> + break;
> + }
> + }
> +
> + if (nowait)
> + ret = 0;
> + else if (timedwait)
> + ret = pthread_cond_timedwait(
> + &pending_requests.async_cond,
> + &pending_requests.lock, &timeout);
> + else
> + ret = pthread_cond_wait(&pending_requests.async_cond,
> + &pending_requests.lock);
> +
> + if (gettimeofday(&now, NULL) < 0) {
> + RTE_LOG(ERR, EAL, "Cannot get current time\n");
> + break;
> + }
> + ts_now.tv_nsec = now.tv_usec * 1000;
> + ts_now.tv_sec = now.tv_sec;
> +
> + if (ret == 0 || ret == ETIMEDOUT) {
> + struct pending_request *next;
> + /* we've either been woken up, or we timed out */
> +
> + /* we have still the lock, check if anything needs
> + * processing.
> + */
> + TAILQ_FOREACH_SAFE(sr, &pending_requests.requests, next,
> + next) {
> + enum async_action action;
> + if (sr->type != REQUEST_TYPE_ASYNC)
> + continue;
> +
> + action = process_async_request(sr, &ts_now);
> + if (action == ACTION_FREE) {
> + TAILQ_REMOVE(&pending_requests.requests,
> + sr, next);
> + free(sr);
> + } else if (action == ACTION_TRIGGER &&
> + trigger == NULL) {
> + TAILQ_REMOVE(&pending_requests.requests,
> + sr, next);
> + trigger = sr;
> + }
> + }
> + }
> + pthread_mutex_unlock(&pending_requests.lock);
> + if (trigger) {
> + trigger_async_action(trigger);
> + free(trigger);
> + }
> + };
> +
> + RTE_LOG(ERR, EAL, "ERROR: asynchronous requests disabled\n");
> +
> + return NULL;
> +}
> +
> +static int
> open_socket_fd(void)
> {
> char peer_name[PATH_MAX] = {0};
> @@ -382,7 +593,7 @@ rte_mp_channel_init(void)
> char thread_name[RTE_MAX_THREAD_NAME_LEN];
> char path[PATH_MAX];
> int dir_fd;
> - pthread_t tid;
> + pthread_t mp_handle_tid, async_reply_handle_tid;
>
> /* create filter path */
> create_socket_path("*", path, sizeof(path));
> @@ -419,7 +630,16 @@ rte_mp_channel_init(void)
> return -1;
> }
>
> - if (pthread_create(&tid, NULL, mp_handle, NULL) < 0) {
> + if (pthread_create(&mp_handle_tid, NULL, mp_handle, NULL) < 0) {
> + RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
> + strerror(errno));
> + close(mp_fd);
> + mp_fd = -1;
> + return -1;
> + }
> +
> + if (pthread_create(&async_reply_handle_tid, NULL,
> + async_reply_handle, NULL) < 0) {
> RTE_LOG(ERR, EAL, "failed to create mp thead: %s\n",
> strerror(errno));
> close(mp_fd);
> @@ -430,7 +650,11 @@ rte_mp_channel_init(void)
>
> /* try best to set thread name */
> snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_handle");
> - rte_thread_setname(tid, thread_name);
> + rte_thread_setname(mp_handle_tid, thread_name);
> +
> + /* try best to set thread name */
> + snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN, "rte_mp_async_handle");
> + rte_thread_setname(async_reply_handle_tid, thread_name);
>
> /* unlock the directory */
> flock(dir_fd, LOCK_UN);
> @@ -602,18 +826,77 @@ rte_mp_sendmsg(struct rte_mp_msg *msg)
> }
>
> static int
> -mp_request_one(const char *dst, struct rte_mp_msg *req,
> +mp_request_async(const char *dst, struct rte_mp_msg *req,
> + struct async_request_param *param)
> +{
> + struct rte_mp_msg *reply_msg;
> + struct pending_request *sync_req, *exist;
> + int ret;
> +
> + sync_req = malloc(sizeof(*sync_req));
> + reply_msg = malloc(sizeof(*reply_msg));
> + if (sync_req == NULL || reply_msg == NULL) {
> + RTE_LOG(ERR, EAL, "Could not allocate space for sync request\n");
> + rte_errno = ENOMEM;
> + ret = -1;
> + goto fail;
> + }
> +
> + memset(sync_req, 0, sizeof(*sync_req));
> + memset(reply_msg, 0, sizeof(*reply_msg));
> +
> + sync_req->type = REQUEST_TYPE_ASYNC;
> + strcpy(sync_req->dst, dst);
> + sync_req->request = req;
> + sync_req->reply = reply_msg;
> + sync_req->async.param = param;
> +
> + /* queue already locked by caller */
> +
> + exist = find_sync_request(dst, req->name);
> + if (!exist)
> + TAILQ_INSERT_TAIL(&pending_requests.requests, sync_req, next);
> + if (exist) {
else?
> + RTE_LOG(ERR, EAL, "A pending request %s:%s\n", dst, req->name);
> + rte_errno = EEXIST;
> + ret = -1;
> + goto fail;
> + }
> +
> + ret = send_msg(dst, req, MP_REQ);
> + if (ret < 0) {
> + RTE_LOG(ERR, EAL, "Fail to send request %s:%s\n",
> + dst, req->name);
> + ret = -1;
> + goto fail;
> + } else if (ret == 0) {
> + ret = 0;
> + goto fail;
> + }
> +
> + param->user_reply.nb_sent++;
> +
> + return 0;
> +fail:
> + free(sync_req);
> + free(reply_msg);
> + return ret;
> +}
> +
> +static int
> +mp_request_sync(const char *dst, struct rte_mp_msg *req,
> struct rte_mp_reply *reply, const struct timespec *ts)
> {
> int ret;
> struct rte_mp_msg msg, *tmp;
> struct pending_request sync_req, *exist;
>
> + sync_req.type = REQUEST_TYPE_SYNC;
> sync_req.reply_received = 0;
> strcpy(sync_req.dst, dst);
> sync_req.request = req;
> sync_req.reply = &msg;
> - pthread_cond_init(&sync_req.cond, NULL);
> + pthread_cond_init(&sync_req.sync.cond, NULL);
>
> pthread_mutex_lock(&pending_requests.lock);
> exist = find_sync_request(dst, req->name);
> @@ -637,7 +920,7 @@ mp_request_one(const char *dst, struct rte_mp_msg *req,
> reply->nb_sent++;
>
> do {
> - ret = pthread_cond_timedwait(&sync_req.cond,
> + ret = pthread_cond_timedwait(&sync_req.sync.cond,
> &pending_requests.lock, ts);
> } while (ret != 0 && ret != ETIMEDOUT);
>
> @@ -703,7 +986,7 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
>
> /* for secondary process, send request to the primary process only */
> if (rte_eal_process_type() == RTE_PROC_SECONDARY)
> - return mp_request_one(eal_mp_socket_path(), req, reply, &end);
> + return mp_request_sync(eal_mp_socket_path(), req, reply, &end);
>
> /* for primary process, broadcast request, and collect reply 1 by 1 */
> mp_dir = opendir(mp_dir_path);
> @@ -732,7 +1015,7 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
> snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
> ent->d_name);
>
> - if (mp_request_one(path, req, reply, &end))
> + if (mp_request_sync(path, req, reply, &end))
> ret = -1;
> }
> /* unlock the directory */
> @@ -744,9 +1027,155 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
> }
>
> int __rte_experimental
> -rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
> +rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts,
> + rte_mp_async_reply_t clb)
> {
> + struct rte_mp_msg *copy;
> + struct pending_request *dummy;
> + struct async_request_param *param = NULL;
No need to assign it to NULL.
> + struct rte_mp_reply *reply;
> + int dir_fd, ret = 0;
> + DIR *mp_dir;
> + struct dirent *ent;
> + struct timeval now;
> + struct timespec *end;
> +
> + RTE_LOG(DEBUG, EAL, "request: %s\n", req->name);
> +
> + if (check_input(req) == false)
> + return -1;
> + if (gettimeofday(&now, NULL) < 0) {
> + RTE_LOG(ERR, EAL, "Faile to get current time\n");
> + rte_errno = errno;
> + return -1;
> + }
> + copy = malloc(sizeof(*copy));
> + dummy = malloc(sizeof(*dummy));
> + param = malloc(sizeof(*param));
> + if (copy == NULL || dummy == NULL || param == NULL) {
> + RTE_LOG(ERR, EAL, "Failed to allocate memory for async reply\n");
> + rte_errno = ENOMEM;
> + goto fail;
> + }
> +
> + memset(copy, 0, sizeof(*copy));
> + memset(dummy, 0, sizeof(*dummy));
> + memset(param, 0, sizeof(*param));
> +
> + /* copy message */
> + memcpy(copy, req, sizeof(*copy));
> +
> + param->n_responses_processed = 0;
> + param->clb = clb;
> + end = ¶m->end;
> + reply = ¶m->user_reply;
> +
> + end->tv_nsec = (now.tv_usec * 1000 + ts->tv_nsec) % 1000000000;
> + end->tv_sec = now.tv_sec + ts->tv_sec +
> + (now.tv_usec * 1000 + ts->tv_nsec) / 1000000000;
> + reply->nb_sent = 0;
> + reply->nb_received = 0;
> + reply->msgs = NULL;
>
> + /* we have to lock the request queue here, as we will be adding a bunch
> + * of requests to the queue at once, and some of the replies may arrive
> + * before we add all of the requests to the queue.
> + */
> + pthread_mutex_lock(&pending_requests.lock);
> +
> + /* we have to ensure that callback gets triggered even if we don't send
> + * anything, therefore earlier we have allocated a dummy request. put it
> + * on the queue and fill it. we will remove it once we know we sent
> + * something.
> + */
Or we can add this dummy at last if it's necessary, instead of adding
firstly and remove if not necessary? No strong option here.
> + dummy->type = REQUEST_TYPE_ASYNC;
> + dummy->request = copy;
> + dummy->reply = NULL;
> + dummy->async.param = param;
> + dummy->reply_received = 1; /* short-circuit the timeout */
> +
> + TAILQ_INSERT_TAIL(&pending_requests.requests, dummy, next);
> +
> + /* for secondary process, send request to the primary process only */
> + if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
> + ret = mp_request_async(eal_mp_socket_path(), copy, param);
> +
> + /* if we sent something, remove dummy request from the queue */
> + if (reply->nb_sent != 0) {
> + TAILQ_REMOVE(&pending_requests.requests, dummy, next);
> + free(dummy);
> + dummy = NULL;
> + }
> +
> + pthread_mutex_unlock(&pending_requests.lock);
> +
> + /* if we couldn't send anything, clean up */
> + if (ret != 0)
> + goto fail;
> + return 0;
> + }
> +
> + /* for primary process, broadcast request */
> + mp_dir = opendir(mp_dir_path);
> + if (!mp_dir) {
> + RTE_LOG(ERR, EAL, "Unable to open directory %s\n", mp_dir_path);
> + rte_errno = errno;
> + goto unlock_fail;
> + }
> + dir_fd = dirfd(mp_dir);
> +
> + /* lock the directory to prevent processes spinning up while we send */
> + if (flock(dir_fd, LOCK_EX)) {
> + RTE_LOG(ERR, EAL, "Unable to lock directory %s\n",
> + mp_dir_path);
> + rte_errno = errno;
> + goto closedir_fail;
> + }
> +
> + while ((ent = readdir(mp_dir))) {
> + char path[PATH_MAX];
> +
> + if (fnmatch(mp_filter, ent->d_name, 0) != 0)
> + continue;
> +
> + snprintf(path, sizeof(path), "%s/%s", mp_dir_path,
> + ent->d_name);
> +
> + if (mp_request_async(path, copy, param))
> + ret = -1;
> + }
> + /* if we sent something, remove dummy request from the queue */
> + if (reply->nb_sent != 0) {
> + TAILQ_REMOVE(&pending_requests.requests, dummy, next);
> + free(dummy);
> + dummy = NULL;
> + }
> + /* trigger async request thread wake up */
> + pthread_cond_signal(&pending_requests.async_cond);
> +
> + /* finally, unlock the queue */
> + pthread_mutex_unlock(&pending_requests.lock);
> +
> + /* unlock the directory */
> + flock(dir_fd, LOCK_UN);
> +
> + /* dir_fd automatically closed on closedir */
> + closedir(mp_dir);
> + return ret;
> +closedir_fail:
> + closedir(mp_dir);
> +unlock_fail:
> + pthread_mutex_unlock(&pending_requests.lock);
> +fail:
> + free(dummy);
> + free(param);
> + free(copy);
> + return -1;
> +}
> +
> +int __rte_experimental
> +rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
> +{
> RTE_LOG(DEBUG, EAL, "reply: %s\n", msg->name);
>
> if (check_input(msg) == false)
> diff --git a/lib/librte_eal/common/include/rte_eal.h b/lib/librte_eal/common/include/rte_eal.h
> index 044474e..87ebfd0 100644
> --- a/lib/librte_eal/common/include/rte_eal.h
> +++ b/lib/librte_eal/common/include/rte_eal.h
> @@ -230,6 +230,16 @@ struct rte_mp_reply {
> typedef int (*rte_mp_t)(const struct rte_mp_msg *msg, const void *peer);
>
> /**
> + * Asynchronous reply function typedef used by other components.
> + *
> + * As we create socket channel for primary/secondary communication, use
> + * this function typedef to register action for coming responses to asynchronous
> + * requests.
> + */
> +typedef int (*rte_mp_async_reply_t)(const struct rte_mp_msg *request,
> + const struct rte_mp_reply *reply);
> +
> +/**
> * @warning
> * @b EXPERIMENTAL: this API may change without prior notice
> *
> @@ -321,6 +331,32 @@ rte_mp_request(struct rte_mp_msg *req, struct rte_mp_reply *reply,
> * @warning
> * @b EXPERIMENTAL: this API may change without prior notice
> *
> + * Send a request to the peer process and expect a reply in a separate callback.
> + *
> + * This function sends a request message to the peer process, and will not
> + * block. Instead, reply will be received in a separate callback.
> + *
> + * @param req
> + * The req argument contains the customized request message.
> + *
> + * @param ts
> + * The ts argument specifies how long we can wait for the peer(s) to reply.
> + *
> + * @param clb
> + * The callback to trigger when all responses for this request have arrived.
> + *
> + * @return
> + * - On success, return 0.
> + * - On failure, return -1, and the reason will be stored in rte_errno.
> + */
> +int __rte_experimental
> +rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts,
> + rte_mp_async_reply_t clb);
> +
> +/**
> + * @warning
> + * @b EXPERIMENTAL: this API may change without prior notice
> + *
> * Send a reply to the peer process.
> *
> * This function will send a reply message in response to a request message
> diff --git a/lib/librte_eal/rte_eal_version.map b/lib/librte_eal/rte_eal_version.map
> index d123602..328a0be 100644
> --- a/lib/librte_eal/rte_eal_version.map
> +++ b/lib/librte_eal/rte_eal_version.map
> @@ -225,6 +225,7 @@ EXPERIMENTAL {
> rte_mp_action_unregister;
> rte_mp_sendmsg;
> rte_mp_request;
> + rte_mp_request_async;
> rte_mp_reply;
> rte_service_attr_get;
> rte_service_attr_reset_all;
^ permalink raw reply [flat|nested] 40+ messages in thread
* Re: [dpdk-dev] [PATCH v5 2/2] eal: add asynchronous request API to DPDK IPC
2018-03-26 14:15 ` Tan, Jianfeng
@ 2018-03-26 14:28 ` Burakov, Anatoly
0 siblings, 0 replies; 40+ messages in thread
From: Burakov, Anatoly @ 2018-03-26 14:28 UTC (permalink / raw)
To: Tan, Jianfeng, dev
Cc: failed, to, remove, Directory, not, empty, konstantin.ananyev
On 26-Mar-18 3:15 PM, Tan, Jianfeng wrote:
>
>
> On 3/24/2018 8:46 PM, Anatoly Burakov wrote:
>> This API is similar to the blocking API that is already present,
>> but reply will be received in a separate callback by the caller
>> (callback specified at the time of request, rather than registering
>> for it in advance).
>>
>> Under the hood, we create a separate thread to deal with replies to
>> asynchronous requests, that will just wait to be notified by the
>> main thread, or woken up on a timer.
>>
>> Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
>
> Generally, it looks great to me except some trivial nits, so
>
> Acked-by: Jianfeng Tan <jianfeng.tan@intel.com>
Thanks!
>> +static void
>> +trigger_async_action(struct pending_request *sr)
>> +{
>> + struct async_request_param *param;
>> + struct rte_mp_reply *reply;
>> +
>> + param = sr->async.param;
>> + reply = ¶m->user_reply;
>> +
>> + param->clb(sr->request, reply);
>> +
>> + /* clean up */
>> + free(sr->async.param->user_reply.msgs);
>
> How about simple "free(reply->msgs);"?
>
I would prefer leaving it as is, as it makes it clear that i'm freeing
everything to do with sync request.
>> +
>> + sync_req->type = REQUEST_TYPE_ASYNC;
>> + strcpy(sync_req->dst, dst);
>> + sync_req->request = req;
>> + sync_req->reply = reply_msg;
>> + sync_req->async.param = param;
>> +
>> + /* queue already locked by caller */
>> +
>> + exist = find_sync_request(dst, req->name);
>> + if (!exist)
>> + TAILQ_INSERT_TAIL(&pending_requests.requests, sync_req, next);
>> + if (exist) {
>
> else?
>
Will fix in v6
>> @@ -744,9 +1027,155 @@ rte_mp_request(struct rte_mp_msg *req, struct
>> rte_mp_reply *reply,
>> }
>> int __rte_experimental
>> -rte_mp_reply(struct rte_mp_msg *msg, const char *peer)
>> +rte_mp_request_async(struct rte_mp_msg *req, const struct timespec *ts,
>> + rte_mp_async_reply_t clb)
>> {
>> + struct rte_mp_msg *copy;
>> + struct pending_request *dummy;
>> + struct async_request_param *param = NULL;
>
> No need to assign it to NULL.
>
Will fix in v6.
>> + /* we have to lock the request queue here, as we will be adding a
>> bunch
>> + * of requests to the queue at once, and some of the replies may
>> arrive
>> + * before we add all of the requests to the queue.
>> + */
>> + pthread_mutex_lock(&pending_requests.lock);
>> +
>> + /* we have to ensure that callback gets triggered even if we
>> don't send
>> + * anything, therefore earlier we have allocated a dummy request.
>> put it
>> + * on the queue and fill it. we will remove it once we know we sent
>> + * something.
>> + */
>
> Or we can add this dummy at last if it's necessary, instead of adding
> firstly and remove if not necessary? No strong option here.
>
Yep, sure, will fix in v6.
--
Thanks,
Anatoly
^ permalink raw reply [flat|nested] 40+ messages in thread