2013-03-25 01:00:48 +00:00
|
|
|
/*
|
|
|
|
* async_nif: An async thread-pool layer for Erlang's NIF API
|
|
|
|
*
|
|
|
|
* Copyright (c) 2012 Basho Technologies, Inc. All Rights Reserved.
|
|
|
|
* Author: Gregory Burd <greg@basho.com> <greg@burd.me>
|
|
|
|
*
|
|
|
|
* This file is provided to you under the Apache License,
|
|
|
|
* Version 2.0 (the "License"); you may not use this file
|
|
|
|
* except in compliance with the License. You may obtain
|
|
|
|
* a copy of the License at
|
|
|
|
*
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
*
|
|
|
|
* Unless required by applicable law or agreed to in writing,
|
|
|
|
* software distributed under the License is distributed on an
|
|
|
|
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
|
|
* KIND, either express or implied. See the License for the
|
|
|
|
* specific language governing permissions and limitations
|
|
|
|
* under the License.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef __ASYNC_NIF_H__
|
|
|
|
#define __ASYNC_NIF_H__
|
|
|
|
|
|
|
|
#if defined(__cplusplus)
|
|
|
|
extern "C" {
|
|
|
|
#endif
|
|
|
|
|
2013-04-17 20:48:23 +00:00
|
|
|
#include <assert.h>
|
2013-04-18 14:32:29 +00:00
|
|
|
#include "fifo_q.h"
|
2013-04-14 23:33:14 +00:00
|
|
|
#ifdef ASYNC_NIF_STATS
|
2013-04-15 19:22:12 +00:00
|
|
|
#include "stats.h" // TODO: measure, measure... measure again
|
2013-04-14 23:33:14 +00:00
|
|
|
#endif
|
|
|
|
|
2013-04-14 12:44:54 +00:00
|
|
|
#define ASYNC_NIF_MAX_WORKERS 128
|
2013-04-15 19:22:12 +00:00
|
|
|
#define ASYNC_NIF_WORKER_QUEUE_SIZE 500
|
|
|
|
|
|
|
|
DECL_FIFO_QUEUE(reqs, struct async_nif_req_entry);
|
|
|
|
|
2013-04-14 12:44:54 +00:00
|
|
|
struct async_nif_work_queue {
|
|
|
|
ErlNifMutex *reqs_mutex;
|
|
|
|
ErlNifCond *reqs_cnd;
|
2013-04-15 19:22:12 +00:00
|
|
|
FIFO_QUEUE_TYPE(reqs) reqs;
|
2013-04-14 12:44:54 +00:00
|
|
|
};
|
|
|
|
|
2013-03-25 01:00:48 +00:00
|
|
|
struct async_nif_worker_entry {
|
|
|
|
ErlNifTid tid;
|
2013-04-14 12:44:54 +00:00
|
|
|
unsigned int worker_id;
|
|
|
|
struct async_nif_state *async_nif;
|
|
|
|
struct async_nif_work_queue *q;
|
2013-03-25 01:00:48 +00:00
|
|
|
};
|
|
|
|
|
2013-04-05 22:09:54 +00:00
|
|
|
struct async_nif_state {
|
2013-04-14 12:44:54 +00:00
|
|
|
unsigned int shutdown;
|
2013-04-05 22:09:54 +00:00
|
|
|
unsigned int num_workers;
|
|
|
|
struct async_nif_worker_entry worker_entries[ASYNC_NIF_MAX_WORKERS];
|
2013-04-14 12:44:54 +00:00
|
|
|
unsigned int num_queues;
|
|
|
|
unsigned int next_q;
|
2013-04-17 15:17:13 +00:00
|
|
|
struct async_nif_work_queue queues[];
|
2013-04-06 15:05:41 +00:00
|
|
|
};
|
|
|
|
|
2013-03-25 01:00:48 +00:00
|
|
|
#define ASYNC_NIF_DECL(decl, frame, pre_block, work_block, post_block) \
|
|
|
|
struct decl ## _args frame; \
|
2013-04-05 22:09:54 +00:00
|
|
|
static void fn_work_ ## decl (ErlNifEnv *env, ERL_NIF_TERM ref, ErlNifPid *pid, unsigned int worker_id, struct decl ## _args *args) work_block \
|
2013-03-25 01:00:48 +00:00
|
|
|
static void fn_post_ ## decl (struct decl ## _args *args) { \
|
2013-04-16 21:09:34 +00:00
|
|
|
do post_block while(0); \
|
2013-03-25 01:00:48 +00:00
|
|
|
} \
|
|
|
|
static ERL_NIF_TERM decl(ErlNifEnv* env, int argc, const ERL_NIF_TERM argv_in[]) { \
|
|
|
|
struct decl ## _args on_stack_args; \
|
|
|
|
struct decl ## _args *args = &on_stack_args; \
|
|
|
|
struct decl ## _args *copy_of_args; \
|
2013-04-16 21:09:34 +00:00
|
|
|
struct async_nif_req_entry *req = NULL; \
|
2013-04-15 22:46:06 +00:00
|
|
|
const char *affinity = NULL; \
|
2013-03-25 01:00:48 +00:00
|
|
|
ErlNifEnv *new_env = NULL; \
|
|
|
|
/* argv[0] is a ref used for selective recv */ \
|
2013-04-15 04:08:01 +00:00
|
|
|
const ERL_NIF_TERM *argv = argv_in + 1; \
|
|
|
|
argc -= 1; \
|
2013-04-18 14:32:29 +00:00
|
|
|
/* Note: !!! this assumes that the first element of priv_data is ours */ \
|
|
|
|
struct async_nif_state *async_nif = *(struct async_nif_state**)enif_priv_data(env); \
|
2013-04-16 21:09:34 +00:00
|
|
|
if (async_nif->shutdown) \
|
2013-03-25 01:00:48 +00:00
|
|
|
return enif_make_tuple2(env, enif_make_atom(env, "error"), \
|
|
|
|
enif_make_atom(env, "shutdown")); \
|
|
|
|
if (!(new_env = enif_alloc_env())) { \
|
|
|
|
return enif_make_tuple2(env, enif_make_atom(env, "error"), \
|
|
|
|
enif_make_atom(env, "enomem")); \
|
|
|
|
} \
|
|
|
|
do pre_block while(0); \
|
2013-04-15 21:37:14 +00:00
|
|
|
req = (struct async_nif_req_entry*)enif_alloc(sizeof(struct async_nif_req_entry)); \
|
|
|
|
if (!req) { \
|
|
|
|
fn_post_ ## decl (args); \
|
|
|
|
enif_free_env(new_env); \
|
|
|
|
return enif_make_tuple2(env, enif_make_atom(env, "error"), \
|
|
|
|
enif_make_atom(env, "enomem")); \
|
|
|
|
} \
|
2013-04-16 21:09:34 +00:00
|
|
|
memset(req, 0, sizeof(struct async_nif_req_entry)); \
|
2013-03-25 01:00:48 +00:00
|
|
|
copy_of_args = (struct decl ## _args *)enif_alloc(sizeof(struct decl ## _args)); \
|
|
|
|
if (!copy_of_args) { \
|
|
|
|
fn_post_ ## decl (args); \
|
2013-04-17 15:17:13 +00:00
|
|
|
enif_free(req); \
|
2013-03-25 01:00:48 +00:00
|
|
|
enif_free_env(new_env); \
|
|
|
|
return enif_make_tuple2(env, enif_make_atom(env, "error"), \
|
|
|
|
enif_make_atom(env, "enomem")); \
|
|
|
|
} \
|
|
|
|
memcpy(copy_of_args, args, sizeof(struct decl ## _args)); \
|
2013-04-15 21:37:14 +00:00
|
|
|
req->env = new_env; \
|
|
|
|
req->ref = enif_make_copy(new_env, argv_in[0]); \
|
|
|
|
enif_self(env, &req->pid); \
|
|
|
|
req->args = (void*)copy_of_args; \
|
|
|
|
req->fn_work = (void (*)(ErlNifEnv *, ERL_NIF_TERM, ErlNifPid*, unsigned int, void *))fn_work_ ## decl ; \
|
2013-04-17 20:48:23 +00:00
|
|
|
req->fn_post = (void (*)(void *))fn_post_ ## decl; \
|
2013-04-17 15:17:13 +00:00
|
|
|
int h = -1; \
|
2013-04-15 22:46:06 +00:00
|
|
|
if (affinity) \
|
|
|
|
h = async_nif_str_hash_func(affinity) % async_nif->num_queues; \
|
2013-04-16 21:09:34 +00:00
|
|
|
ERL_NIF_TERM reply = async_nif_enqueue_req(async_nif, req, h); \
|
2013-04-17 15:17:13 +00:00
|
|
|
if (!reply) { \
|
2013-04-17 20:48:23 +00:00
|
|
|
fn_post_ ## decl (args); \
|
2013-04-17 15:17:13 +00:00
|
|
|
enif_free(req); \
|
|
|
|
enif_free_env(new_env); \
|
|
|
|
enif_free(copy_of_args); \
|
|
|
|
return enif_make_tuple2(env, enif_make_atom(env, "error"), \
|
|
|
|
enif_make_atom(env, "shutdown")); \
|
|
|
|
} \
|
2013-04-16 21:09:34 +00:00
|
|
|
return reply; \
|
2013-03-25 01:00:48 +00:00
|
|
|
}
|
|
|
|
|
2013-04-15 19:22:12 +00:00
|
|
|
#define ASYNC_NIF_INIT(name) \
|
2013-04-12 19:25:56 +00:00
|
|
|
static ErlNifMutex *name##_async_nif_coord = NULL;
|
|
|
|
|
|
|
|
#define ASYNC_NIF_LOAD(name, priv) do { \
|
|
|
|
if (!name##_async_nif_coord) \
|
|
|
|
name##_async_nif_coord = enif_mutex_create(NULL); \
|
|
|
|
enif_mutex_lock(name##_async_nif_coord); \
|
|
|
|
priv = async_nif_load(); \
|
|
|
|
enif_mutex_unlock(name##_async_nif_coord); \
|
|
|
|
} while(0);
|
2013-04-18 14:32:29 +00:00
|
|
|
#define ASYNC_NIF_UNLOAD(name, env, priv) do { \
|
2013-04-12 19:25:56 +00:00
|
|
|
if (!name##_async_nif_coord) \
|
|
|
|
name##_async_nif_coord = enif_mutex_create(NULL); \
|
|
|
|
enif_mutex_lock(name##_async_nif_coord); \
|
2013-04-18 14:32:29 +00:00
|
|
|
async_nif_unload(env, priv); \
|
2013-04-12 19:25:56 +00:00
|
|
|
enif_mutex_unlock(name##_async_nif_coord); \
|
|
|
|
enif_mutex_destroy(name##_async_nif_coord); \
|
|
|
|
name##_async_nif_coord = NULL; \
|
|
|
|
} while(0);
|
|
|
|
#define ASYNC_NIF_UPGRADE(name, env) do { \
|
|
|
|
if (!name##_async_nif_coord) \
|
|
|
|
name##_async_nif_coord = enif_mutex_create(NULL); \
|
|
|
|
enif_mutex_lock(name##_async_nif_coord); \
|
|
|
|
async_nif_upgrade(env); \
|
|
|
|
enif_mutex_unlock(name##_async_nif_coord); \
|
|
|
|
} while(0);
|
2013-03-25 01:00:48 +00:00
|
|
|
|
|
|
|
#define ASYNC_NIF_RETURN_BADARG() return enif_make_badarg(env);
|
|
|
|
#define ASYNC_NIF_WORK_ENV new_env
|
|
|
|
|
|
|
|
#define ASYNC_NIF_REPLY(msg) enif_send(NULL, pid, env, enif_make_tuple2(env, ref, msg))
|
2013-04-17 15:17:13 +00:00
|
|
|
// TODO: fix, currently NOREPLY() will block cause the recieve in async_nif.hrl wait forever
|
|
|
|
#define ASYNC_NIF_NOREPLY() enif_free_env(env)
|
2013-03-25 01:00:48 +00:00
|
|
|
|
2013-04-15 19:22:12 +00:00
|
|
|
/**
|
2013-04-15 22:46:06 +00:00
|
|
|
* TODO:
|
|
|
|
*/
|
|
|
|
static inline unsigned int async_nif_str_hash_func(const char *s)
|
|
|
|
{
|
|
|
|
unsigned int h = (unsigned int)*s;
|
|
|
|
if (h) for (++s ; *s; ++s) h = (h << 5) - h + (unsigned int)*s;
|
|
|
|
return h;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* TODO:
|
2013-04-15 19:22:12 +00:00
|
|
|
*/
|
2013-04-05 22:09:54 +00:00
|
|
|
static ERL_NIF_TERM
|
2013-04-17 15:17:13 +00:00
|
|
|
async_nif_enqueue_req(struct async_nif_state* async_nif, struct async_nif_req_entry *req, int hint)
|
2013-03-25 01:00:48 +00:00
|
|
|
{
|
2013-04-17 15:17:13 +00:00
|
|
|
/* Identify the most appropriate worker for this request. */
|
|
|
|
unsigned int qid = (hint != -1) ? hint : async_nif->next_q;
|
|
|
|
struct async_nif_work_queue *q = NULL;
|
|
|
|
do {
|
2013-04-15 04:08:01 +00:00
|
|
|
q = &async_nif->queues[qid];
|
2013-04-17 15:17:13 +00:00
|
|
|
enif_mutex_lock(q->reqs_mutex);
|
|
|
|
|
|
|
|
/* Now that we hold the lock, check for shutdown. As long as we
|
|
|
|
hold this lock either a) we're shutting down so exit now or
|
|
|
|
b) this queue will be valid until we release the lock. */
|
|
|
|
if (async_nif->shutdown)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (fifo_q_full(reqs, q->reqs)) { // TODO: || (q->avg_latency > median_latency)
|
|
|
|
enif_mutex_unlock(q->reqs_mutex);
|
|
|
|
qid = (qid + 1) % async_nif->num_queues;
|
|
|
|
q = &async_nif->queues[qid];
|
|
|
|
} else {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
} while(1);
|
|
|
|
|
|
|
|
/* And add the request to their work queue. */
|
2013-04-15 19:22:12 +00:00
|
|
|
fifo_q_put(reqs, q->reqs, req);
|
2013-04-15 21:37:14 +00:00
|
|
|
|
2013-04-12 19:25:56 +00:00
|
|
|
/* Build the term before releasing the lock so as not to race on the use of
|
2013-04-17 15:17:13 +00:00
|
|
|
the req pointer (which will soon become invalid in another thread
|
|
|
|
performing the request). */
|
2013-04-12 19:25:56 +00:00
|
|
|
ERL_NIF_TERM reply = enif_make_tuple2(req->env, enif_make_atom(req->env, "ok"),
|
2013-04-17 15:17:13 +00:00
|
|
|
enif_make_atom(req->env, "enqueued"));
|
2013-04-14 12:44:54 +00:00
|
|
|
enif_mutex_unlock(q->reqs_mutex);
|
2013-04-14 17:54:45 +00:00
|
|
|
enif_cond_signal(q->reqs_cnd);
|
2013-04-12 19:25:56 +00:00
|
|
|
return reply;
|
2013-03-25 01:00:48 +00:00
|
|
|
}
|
|
|
|
|
2013-04-06 15:05:41 +00:00
|
|
|
static void *
|
|
|
|
async_nif_worker_fn(void *arg)
|
2013-03-25 01:00:48 +00:00
|
|
|
{
|
2013-04-14 12:44:54 +00:00
|
|
|
struct async_nif_worker_entry *we = (struct async_nif_worker_entry *)arg;
|
|
|
|
unsigned int worker_id = we->worker_id;
|
|
|
|
struct async_nif_state *async_nif = we->async_nif;
|
|
|
|
struct async_nif_work_queue *q = we->q;
|
2013-04-17 17:19:06 +00:00
|
|
|
struct async_nif_req_entry *req = NULL;
|
2013-03-25 01:00:48 +00:00
|
|
|
|
|
|
|
for(;;) {
|
|
|
|
/* Examine the request queue, are there things to be done? */
|
2013-04-14 12:44:54 +00:00
|
|
|
enif_mutex_lock(q->reqs_mutex);
|
2013-03-25 01:00:48 +00:00
|
|
|
check_again_for_work:
|
2013-04-12 19:25:56 +00:00
|
|
|
if (async_nif->shutdown) {
|
2013-04-14 12:44:54 +00:00
|
|
|
enif_mutex_unlock(q->reqs_mutex);
|
2013-04-12 19:25:56 +00:00
|
|
|
break;
|
|
|
|
}
|
2013-04-15 19:22:12 +00:00
|
|
|
if (fifo_q_empty(reqs, q->reqs)) {
|
2013-04-12 19:25:56 +00:00
|
|
|
/* Queue is empty, wait for work */
|
2013-04-14 12:44:54 +00:00
|
|
|
enif_cond_wait(q->reqs_cnd, q->reqs_mutex);
|
2013-03-25 01:00:48 +00:00
|
|
|
goto check_again_for_work;
|
|
|
|
} else {
|
2013-04-17 20:48:23 +00:00
|
|
|
assert(fifo_q_size(reqs, q->reqs) > 0);
|
|
|
|
assert(fifo_q_size(reqs, q->reqs) < fifo_q_capacity(reqs, q->reqs));
|
2013-04-15 04:08:01 +00:00
|
|
|
/* At this point the next req is ours to process and we hold the
|
2013-04-17 17:19:06 +00:00
|
|
|
reqs_mutex lock. Take the request off the queue. */
|
|
|
|
req = fifo_q_get(reqs, q->reqs);
|
|
|
|
enif_mutex_unlock(q->reqs_mutex);
|
|
|
|
|
2013-04-17 20:48:23 +00:00
|
|
|
/* Ensure that there is at least one other worker watching this queue. */
|
2013-04-17 17:19:06 +00:00
|
|
|
enif_cond_signal(q->reqs_cnd);
|
|
|
|
|
2013-04-17 20:48:23 +00:00
|
|
|
/* Finally, do the work, */
|
2013-04-17 17:19:06 +00:00
|
|
|
req->fn_work(req->env, req->ref, &req->pid, worker_id, req->args);
|
2013-04-17 20:48:23 +00:00
|
|
|
|
|
|
|
/* and then call the post-work cleanup function. */
|
2013-04-17 17:19:06 +00:00
|
|
|
req->fn_post(req->args);
|
2013-04-17 20:48:23 +00:00
|
|
|
|
|
|
|
/* Note: we don't call enif_free_env(req->env) because somewhere in the
|
|
|
|
work block there has been a call to enif_send() which invalidates the
|
|
|
|
environment we created which in theory will cause the next GC to free
|
|
|
|
it for us). If a work block doesn't call ASYNC_NIF_REPLY() at some
|
|
|
|
point then it must call ASYNC_NIF_NOREPLY() to free this env. */
|
2013-04-17 17:19:06 +00:00
|
|
|
enif_free(req->args);
|
|
|
|
enif_free(req);
|
2013-04-17 20:48:23 +00:00
|
|
|
req = NULL;
|
2013-03-25 01:00:48 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
enif_thread_exit(0);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-04-12 19:25:56 +00:00
|
|
|
static void
|
2013-04-18 14:32:29 +00:00
|
|
|
async_nif_unload(ErlNifEnv *env, struct async_nif_state *async_nif)
|
2013-03-25 01:00:48 +00:00
|
|
|
{
|
|
|
|
unsigned int i;
|
2013-04-17 20:48:23 +00:00
|
|
|
unsigned int num_queues = async_nif->num_queues;
|
2013-04-17 22:26:59 +00:00
|
|
|
struct async_nif_work_queue *q = NULL;
|
2013-03-25 01:00:48 +00:00
|
|
|
|
2013-04-17 15:17:13 +00:00
|
|
|
/* Signal the worker threads, stop what you're doing and exit. To
|
|
|
|
ensure that we don't race with the enqueue() process we first
|
|
|
|
lock all the worker queues, then set shutdown to true, then
|
|
|
|
unlock. The enqueue function will take the queue mutex, then
|
|
|
|
test for shutdown condition, then enqueue only if not shutting
|
|
|
|
down. */
|
2013-04-17 22:26:59 +00:00
|
|
|
for (i = 0; i < num_queues; i++) {
|
|
|
|
q = &async_nif->queues[i];
|
|
|
|
enif_mutex_lock(q->reqs_mutex);
|
|
|
|
}
|
2013-04-05 22:09:54 +00:00
|
|
|
async_nif->shutdown = 1;
|
2013-04-17 20:48:23 +00:00
|
|
|
for (i = 0; i < num_queues; i++) {
|
2013-04-17 22:26:59 +00:00
|
|
|
q = &async_nif->queues[i];
|
2013-04-14 12:44:54 +00:00
|
|
|
enif_cond_broadcast(q->reqs_cnd);
|
2013-04-17 22:26:59 +00:00
|
|
|
enif_mutex_unlock(q->reqs_mutex);
|
2013-04-14 12:44:54 +00:00
|
|
|
}
|
2013-03-25 01:00:48 +00:00
|
|
|
|
|
|
|
/* Join for the now exiting worker threads. */
|
2013-04-07 13:21:47 +00:00
|
|
|
for (i = 0; i < async_nif->num_workers; ++i) {
|
2013-04-17 15:17:13 +00:00
|
|
|
void *exit_value = 0; /* We ignore the thread_join's exit value. */
|
2013-04-05 22:09:54 +00:00
|
|
|
enif_thread_join(async_nif->worker_entries[i].tid, &exit_value);
|
2013-03-25 01:00:48 +00:00
|
|
|
}
|
|
|
|
|
2013-04-14 12:44:54 +00:00
|
|
|
/* Cleanup requests, mutexes and conditions in each work queue. */
|
2013-04-17 15:17:13 +00:00
|
|
|
for (i = 0; i < num_queues; i++) {
|
2013-04-17 22:26:59 +00:00
|
|
|
q = &async_nif->queues[i];
|
2013-04-14 12:44:54 +00:00
|
|
|
|
|
|
|
/* Worker threads are stopped, now toss anything left in the queue. */
|
|
|
|
struct async_nif_req_entry *req = NULL;
|
2013-04-15 19:22:12 +00:00
|
|
|
fifo_q_foreach(reqs, q->reqs, req, {
|
2013-04-14 12:44:54 +00:00
|
|
|
enif_send(NULL, &req->pid, req->env,
|
|
|
|
enif_make_tuple2(req->env, enif_make_atom(req->env, "error"),
|
|
|
|
enif_make_atom(req->env, "shutdown")));
|
|
|
|
req->fn_post(req->args);
|
|
|
|
enif_free(req->args);
|
2013-04-15 21:37:14 +00:00
|
|
|
enif_free(req);
|
2013-04-15 19:22:12 +00:00
|
|
|
});
|
|
|
|
fifo_q_free(reqs, q->reqs);
|
2013-04-17 22:26:59 +00:00
|
|
|
enif_mutex_destroy(q->reqs_mutex);
|
|
|
|
enif_cond_destroy(q->reqs_cnd);
|
2013-03-25 01:00:48 +00:00
|
|
|
}
|
2013-04-17 22:26:59 +00:00
|
|
|
memset(async_nif, 0, sizeof(struct async_nif_state) + (sizeof(struct async_nif_work_queue) * async_nif->num_queues));
|
2013-04-12 19:25:56 +00:00
|
|
|
enif_free(async_nif);
|
2013-03-25 01:00:48 +00:00
|
|
|
}
|
|
|
|
|
2013-04-05 22:09:54 +00:00
|
|
|
static void *
|
2013-04-18 14:32:29 +00:00
|
|
|
async_nif_load()
|
2013-03-25 01:00:48 +00:00
|
|
|
{
|
2013-04-06 15:05:41 +00:00
|
|
|
static int has_init = 0;
|
2013-04-17 22:26:59 +00:00
|
|
|
unsigned int i, j, num_queues;
|
2013-04-06 15:05:41 +00:00
|
|
|
ErlNifSysInfo info;
|
2013-04-05 22:09:54 +00:00
|
|
|
struct async_nif_state *async_nif;
|
2013-03-25 01:00:48 +00:00
|
|
|
|
|
|
|
/* Don't init more than once. */
|
2013-04-06 15:05:41 +00:00
|
|
|
if (has_init) return 0;
|
|
|
|
else has_init = 1;
|
2013-03-25 01:00:48 +00:00
|
|
|
|
2013-04-05 22:09:54 +00:00
|
|
|
/* Find out how many schedulers there are. */
|
2013-04-06 15:05:41 +00:00
|
|
|
enif_system_info(&info, sizeof(ErlNifSysInfo));
|
2013-04-05 22:09:54 +00:00
|
|
|
|
2013-04-17 22:26:59 +00:00
|
|
|
/* Size the number of work queues according to schedulers. */
|
|
|
|
if (info.scheduler_threads > ASYNC_NIF_MAX_WORKERS / 2) {
|
|
|
|
num_queues = ASYNC_NIF_MAX_WORKERS / 2;
|
|
|
|
} else {
|
|
|
|
int remainder = ASYNC_NIF_MAX_WORKERS % info.scheduler_threads;
|
|
|
|
if (remainder != 0)
|
|
|
|
num_queues = info.scheduler_threads - remainder;
|
|
|
|
else
|
|
|
|
num_queues = info.scheduler_threads;
|
|
|
|
if (num_queues < 2)
|
|
|
|
num_queues = 2;
|
|
|
|
}
|
|
|
|
|
2013-04-05 22:09:54 +00:00
|
|
|
/* Init our portion of priv_data's module-specific state. */
|
2013-04-17 15:17:13 +00:00
|
|
|
async_nif = enif_alloc(sizeof(struct async_nif_state) +
|
2013-04-17 22:26:59 +00:00
|
|
|
sizeof(struct async_nif_work_queue) * num_queues);
|
2013-04-05 22:09:54 +00:00
|
|
|
if (!async_nif)
|
|
|
|
return NULL;
|
2013-04-17 15:17:13 +00:00
|
|
|
memset(async_nif, 0, sizeof(struct async_nif_state) +
|
2013-04-17 22:26:59 +00:00
|
|
|
sizeof(struct async_nif_work_queue) * num_queues);
|
2013-03-25 01:00:48 +00:00
|
|
|
|
2013-04-17 22:26:59 +00:00
|
|
|
async_nif->num_queues = num_queues;
|
2013-04-17 20:48:23 +00:00
|
|
|
async_nif->num_workers = ASYNC_NIF_MAX_WORKERS; // TODO: start with 2 per queue, then grow if needed
|
2013-04-14 12:44:54 +00:00
|
|
|
async_nif->next_q = 0;
|
|
|
|
async_nif->shutdown = 0;
|
|
|
|
|
|
|
|
for (i = 0; i < async_nif->num_queues; i++) {
|
|
|
|
struct async_nif_work_queue *q = &async_nif->queues[i];
|
2013-04-15 19:22:12 +00:00
|
|
|
q->reqs = fifo_q_new(reqs, ASYNC_NIF_WORKER_QUEUE_SIZE);
|
2013-04-14 12:44:54 +00:00
|
|
|
q->reqs_mutex = enif_mutex_create(NULL);
|
|
|
|
q->reqs_cnd = enif_cond_create(NULL);
|
|
|
|
}
|
2013-03-25 01:00:48 +00:00
|
|
|
|
|
|
|
/* Setup the thread pool management. */
|
2013-04-14 21:23:57 +00:00
|
|
|
memset(async_nif->worker_entries, 0, sizeof(struct async_nif_worker_entry) * ASYNC_NIF_MAX_WORKERS);
|
2013-04-05 22:09:54 +00:00
|
|
|
|
2013-04-14 12:44:54 +00:00
|
|
|
/* Start the worker threads. */
|
2013-04-17 20:48:23 +00:00
|
|
|
for (i = 0; i < async_nif->num_workers; i++) {
|
2013-04-14 12:44:54 +00:00
|
|
|
struct async_nif_worker_entry *we = &async_nif->worker_entries[i];
|
|
|
|
we->async_nif = async_nif;
|
|
|
|
we->worker_id = i;
|
|
|
|
we->q = &async_nif->queues[i % async_nif->num_queues];
|
2013-04-05 22:09:54 +00:00
|
|
|
if (enif_thread_create(NULL, &async_nif->worker_entries[i].tid,
|
2013-04-14 12:44:54 +00:00
|
|
|
&async_nif_worker_fn, (void*)we, NULL) != 0) {
|
2013-04-05 22:09:54 +00:00
|
|
|
async_nif->shutdown = 1;
|
2013-04-14 12:44:54 +00:00
|
|
|
|
|
|
|
for (j = 0; j < async_nif->num_queues; j++) {
|
|
|
|
struct async_nif_work_queue *q = &async_nif->queues[j];
|
|
|
|
enif_cond_broadcast(q->reqs_cnd);
|
|
|
|
}
|
|
|
|
|
2013-03-25 01:00:48 +00:00
|
|
|
while(i-- > 0) {
|
|
|
|
void *exit_value = 0; /* Ignore this. */
|
2013-04-05 22:09:54 +00:00
|
|
|
enif_thread_join(async_nif->worker_entries[i].tid, &exit_value);
|
2013-03-25 01:00:48 +00:00
|
|
|
}
|
2013-04-14 12:44:54 +00:00
|
|
|
|
2013-04-17 20:48:23 +00:00
|
|
|
for (j = 0; j < async_nif->num_queues; j++) {
|
|
|
|
struct async_nif_work_queue *q = &async_nif->queues[j];
|
|
|
|
enif_mutex_destroy(q->reqs_mutex);
|
|
|
|
enif_cond_destroy(q->reqs_cnd);
|
|
|
|
}
|
|
|
|
|
2013-04-14 21:23:57 +00:00
|
|
|
memset(async_nif->worker_entries, 0, sizeof(struct async_nif_worker_entry) * ASYNC_NIF_MAX_WORKERS);
|
2013-04-17 20:48:23 +00:00
|
|
|
enif_free(async_nif);
|
2013-04-05 22:09:54 +00:00
|
|
|
return NULL;
|
2013-03-25 01:00:48 +00:00
|
|
|
}
|
|
|
|
}
|
2013-04-05 22:09:54 +00:00
|
|
|
return async_nif;
|
2013-03-25 01:00:48 +00:00
|
|
|
}
|
|
|
|
|
2013-04-12 19:25:56 +00:00
|
|
|
static void
|
|
|
|
async_nif_upgrade(ErlNifEnv *env)
|
|
|
|
{
|
|
|
|
// TODO:
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2013-03-25 01:00:48 +00:00
|
|
|
#if defined(__cplusplus)
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#endif // __ASYNC_NIF_H__
|