Update workqueue implementation to use a single queue for the work

Previously I used one queue per worker; now I use one queue for
everyone.  The "broadcast" code is gone, replaced with an idempotent
'update' operation.
This commit is contained in:
Nick Mathewson 2015-01-14 13:29:58 -05:00
parent 051ad788e0
commit a52e549124
5 changed files with 140 additions and 113 deletions

View file

@ -5,6 +5,7 @@
#include "orconfig.h" #include "orconfig.h"
#include <pthread.h> #include <pthread.h>
#include <signal.h>
#include "compat.h" #include "compat.h"
#include "torlog.h" #include "torlog.h"

View file

@ -1,4 +1,4 @@
/* Copyright (c) 2013, The Tor Project, Inc. */ /* Copyright (c) 2013-2015, The Tor Project, Inc. */
/* See LICENSE for licensing information */ /* See LICENSE for licensing information */
#include "orconfig.h" #include "orconfig.h"
@ -13,8 +13,23 @@ struct threadpool_s {
/** An array of pointers to workerthread_t: one for each running worker /** An array of pointers to workerthread_t: one for each running worker
* thread. */ * thread. */
struct workerthread_s **threads; struct workerthread_s **threads;
/** Index of the next thread that we'll give work to.*/
int next_for_work; /** Condition variable that we wait on when we have no work, and which
* gets signaled when our queue becomes nonempty. */
tor_cond_t condition;
/** Queue of pending work that we have to do. */
TOR_TAILQ_HEAD(, workqueue_entry_s) work;
/** The current 'update generation' of the threadpool. Any thread that is
* at an earlier generation needs to run the update function. */
unsigned generation;
/** Function that should be run for updates on each thread. */
int (*update_fn)(void *, void *);
/** Function to free update arguments if they can't be run. */
void (*free_update_arg_fn)(void *);
/** Array of n_threads update arguments. */
void **update_args;
/** Number of elements in threads. */ /** Number of elements in threads. */
int n_threads; int n_threads;
@ -34,10 +49,10 @@ struct workqueue_entry_s {
/** The next workqueue_entry_t that's pending on the same thread or /** The next workqueue_entry_t that's pending on the same thread or
* reply queue. */ * reply queue. */
TOR_TAILQ_ENTRY(workqueue_entry_s) next_work; TOR_TAILQ_ENTRY(workqueue_entry_s) next_work;
/** The thread to which this workqueue_entry_t was assigned. This field /** The threadpool to which this workqueue_entry_t was assigned. This field
* is set when the workqueue_entry_t is created, and won't be cleared until * is set when the workqueue_entry_t is created, and won't be cleared until
* after it's handled in the main thread. */ * after it's handled in the main thread. */
struct workerthread_s *on_thread; struct threadpool_s *on_pool;
/** True iff this entry is waiting for a worker to start processing it. */ /** True iff this entry is waiting for a worker to start processing it. */
uint8_t pending; uint8_t pending;
/** Function to run in the worker thread. */ /** Function to run in the worker thread. */
@ -62,13 +77,10 @@ struct replyqueue_s {
* contention, each gets its own queue. This breaks the guarantee that that * contention, each gets its own queue. This breaks the guarantee that that
* queued work will get executed strictly in order. */ * queued work will get executed strictly in order. */
typedef struct workerthread_s { typedef struct workerthread_s {
/** Lock to protect all fields of this thread and its queue. */ /** Which thread it this? In range 0..in_pool->n_threads-1 */
tor_mutex_t lock; int index;
/** Condition variable that we wait on when we have no work, and which /** The pool this thread is a part of. */
* gets signaled when our queue becomes nonempty. */ struct threadpool_s *in_pool;
tor_cond_t condition;
/** Queue of pending work that we have to do. */
TOR_TAILQ_HEAD(, workqueue_entry_s) work;
/** True iff this thread is currently in its loop. (Not currently used.) */ /** True iff this thread is currently in its loop. (Not currently used.) */
unsigned is_running; unsigned is_running;
/** True iff this thread has crashed or is shut down for some reason. (Not /** True iff this thread has crashed or is shut down for some reason. (Not
@ -81,6 +93,8 @@ typedef struct workerthread_s {
void *state; void *state;
/** Reply queue to which we pass our results. */ /** Reply queue to which we pass our results. */
replyqueue_t *reply_queue; replyqueue_t *reply_queue;
/** The current update generation of this thread */
unsigned generation;
} workerthread_t; } workerthread_t;
static void queue_reply(replyqueue_t *queue, workqueue_entry_t *work); static void queue_reply(replyqueue_t *queue, workqueue_entry_t *work);
@ -132,13 +146,13 @@ workqueue_entry_cancel(workqueue_entry_t *ent)
{ {
int cancelled = 0; int cancelled = 0;
void *result = NULL; void *result = NULL;
tor_mutex_acquire(&ent->on_thread->lock); tor_mutex_acquire(&ent->on_pool->lock);
if (ent->pending) { if (ent->pending) {
TOR_TAILQ_REMOVE(&ent->on_thread->work, ent, next_work); TOR_TAILQ_REMOVE(&ent->on_pool->work, ent, next_work);
cancelled = 1; cancelled = 1;
result = ent->arg; result = ent->arg;
} }
tor_mutex_release(&ent->on_thread->lock); tor_mutex_release(&ent->on_pool->lock);
if (cancelled) { if (cancelled) {
tor_free(ent); tor_free(ent);
@ -146,6 +160,16 @@ workqueue_entry_cancel(workqueue_entry_t *ent)
return result; return result;
} }
/**DOCDOC
must hold lock */
static int
worker_thread_has_work(workerthread_t *thread)
{
return !TOR_TAILQ_EMPTY(&thread->in_pool->work) ||
thread->generation != thread->in_pool->generation;
}
/** /**
* Main function for the worker thread. * Main function for the worker thread.
*/ */
@ -153,20 +177,39 @@ static void
worker_thread_main(void *thread_) worker_thread_main(void *thread_)
{ {
workerthread_t *thread = thread_; workerthread_t *thread = thread_;
threadpool_t *pool = thread->in_pool;
workqueue_entry_t *work; workqueue_entry_t *work;
int result; int result;
tor_mutex_acquire(&thread->lock);
thread->is_running = 1; thread->is_running = 1;
tor_mutex_acquire(&pool->lock);
while (1) { while (1) {
/* lock must be held at this point. */ /* lock must be held at this point. */
while (!TOR_TAILQ_EMPTY(&thread->work)) { while (worker_thread_has_work(thread)) {
/* lock must be held at this point. */ /* lock must be held at this point. */
if (thread->in_pool->generation != thread->generation) {
void *arg = thread->in_pool->update_args[thread->index];
thread->in_pool->update_args[thread->index] = NULL;
int (*update_fn)(void*,void*) = thread->in_pool->update_fn;
thread->generation = thread->in_pool->generation;
tor_mutex_release(&pool->lock);
work = TOR_TAILQ_FIRST(&thread->work); int r = update_fn(thread->state, arg);
TOR_TAILQ_REMOVE(&thread->work, work, next_work);
if (r < 0) {
thread->is_running = 0;
thread->is_shut_down = 1;
return;
}
tor_mutex_acquire(&pool->lock);
continue;
}
work = TOR_TAILQ_FIRST(&pool->work);
TOR_TAILQ_REMOVE(&pool->work, work, next_work);
work->pending = 0; work->pending = 0;
tor_mutex_release(&thread->lock); tor_mutex_release(&pool->lock);
/* We run the work function without holding the thread lock. This /* We run the work function without holding the thread lock. This
* is the main thread's first opportunity to give us more work. */ * is the main thread's first opportunity to give us more work. */
@ -175,25 +218,23 @@ worker_thread_main(void *thread_)
/* Queue the reply for the main thread. */ /* Queue the reply for the main thread. */
queue_reply(thread->reply_queue, work); queue_reply(thread->reply_queue, work);
tor_mutex_acquire(&thread->lock);
/* We may need to exit the thread. */ /* We may need to exit the thread. */
if (result >= WQ_RPL_ERROR) { if (result >= WQ_RPL_ERROR) {
thread->is_running = 0; thread->is_running = 0;
thread->is_shut_down = 1; thread->is_shut_down = 1;
tor_mutex_release(&thread->lock);
return; return;
} }
tor_mutex_acquire(&pool->lock);
} }
/* At this point the lock is held, and there is no work in this thread's /* At this point the lock is held, and there is no work in this thread's
* queue. */ * queue. */
/* TODO: Try work-stealing. */
/* TODO: support an idle-function */ /* TODO: support an idle-function */
/* Okay. Now, wait till somebody has work for us. */ /* Okay. Now, wait till somebody has work for us. */
/* XXXX we could just omit waiting and instead */ /* XXXX we could just omit waiting and instead */
thread->waiting = 1; thread->waiting = 1;
if (tor_cond_wait(&thread->condition, &thread->lock, NULL) < 0) { if (tor_cond_wait(&pool->condition, &pool->lock, NULL) < 0) {
/* XXXX ERROR */ /* XXXX ERROR */
} }
thread->waiting = 0; thread->waiting = 0;
@ -221,14 +262,12 @@ queue_reply(replyqueue_t *queue, workqueue_entry_t *work)
/** Allocate and start a new worker thread to use state object <b>state</b>, /** Allocate and start a new worker thread to use state object <b>state</b>,
* and send responses to <b>replyqueue</b>. */ * and send responses to <b>replyqueue</b>. */
static workerthread_t * static workerthread_t *
workerthread_new(void *state, replyqueue_t *replyqueue) workerthread_new(void *state, threadpool_t *pool, replyqueue_t *replyqueue)
{ {
workerthread_t *thr = tor_malloc_zero(sizeof(workerthread_t)); workerthread_t *thr = tor_malloc_zero(sizeof(workerthread_t));
tor_mutex_init_for_cond(&thr->lock);
tor_cond_init(&thr->condition);
TOR_TAILQ_INIT(&thr->work);
thr->state = state; thr->state = state;
thr->reply_queue = replyqueue; thr->reply_queue = replyqueue;
thr->in_pool = pool;
if (spawn_func(worker_thread_main, thr) < 0) { if (spawn_func(worker_thread_main, thr) < 0) {
log_err(LD_GENERAL, "Can't launch worker thread."); log_err(LD_GENERAL, "Can't launch worker thread.");
@ -238,30 +277,6 @@ workerthread_new(void *state, replyqueue_t *replyqueue)
return thr; return thr;
} }
/**
* Add an item of work to a single worker thread. See threadpool_queue_work(*)
* for arguments.
*/
static workqueue_entry_t *
workerthread_queue_work(workerthread_t *worker,
int (*fn)(void *, void *),
void (*reply_fn)(void *),
void *arg)
{
workqueue_entry_t *ent = workqueue_entry_new(fn, reply_fn, arg);
tor_mutex_acquire(&worker->lock);
ent->on_thread = worker;
ent->pending = 1;
TOR_TAILQ_INSERT_TAIL(&worker->work, ent, next_work);
if (worker->waiting) /* XXXX inside or outside of lock?? */
tor_cond_signal_one(&worker->condition);
tor_mutex_release(&worker->lock);
return ent;
}
/** /**
* Queue an item of work for a thread in a thread pool. The function * Queue an item of work for a thread in a thread pool. The function
* <b>fn</b> will be run in a worker thread, and will receive as arguments the * <b>fn</b> will be run in a worker thread, and will receive as arguments the
@ -285,20 +300,19 @@ threadpool_queue_work(threadpool_t *pool,
void (*reply_fn)(void *), void (*reply_fn)(void *),
void *arg) void *arg)
{ {
workerthread_t *worker; workqueue_entry_t *ent = workqueue_entry_new(fn, reply_fn, arg);
ent->on_pool = pool;
ent->pending = 1;
tor_mutex_acquire(&pool->lock); tor_mutex_acquire(&pool->lock);
/* Pick the next thread in random-access order. */
worker = pool->threads[pool->next_for_work++]; TOR_TAILQ_INSERT_TAIL(&pool->work, ent, next_work);
if (!worker) {
tor_mutex_release(&pool->lock);
return NULL;
}
if (pool->next_for_work >= pool->n_threads)
pool->next_for_work = 0;
tor_mutex_release(&pool->lock); tor_mutex_release(&pool->lock);
return workerthread_queue_work(worker, fn, reply_fn, arg); tor_cond_signal_one(&pool->condition);
return ent;
} }
/** /**
@ -309,30 +323,56 @@ threadpool_queue_work(threadpool_t *pool,
* <b>arg</b> value is passed to <b>dup_fn</b> once per each thread to * <b>arg</b> value is passed to <b>dup_fn</b> once per each thread to
* make a copy of it. * make a copy of it.
* *
* UPDATE FUNCTIONS MUST BE IDEMPOTENT. We do not guarantee that every update
* will be run. If a new update is scheduled before the old update finishes
* running, then the new will replace the old in any threads that haven't run
* it yet.
*
* Return 0 on success, -1 on failure. * Return 0 on success, -1 on failure.
*/ */
int int
threadpool_queue_for_all(threadpool_t *pool, threadpool_queue_update(threadpool_t *pool,
void *(*dup_fn)(void *), void *(*dup_fn)(void *),
int (*fn)(void *, void *), int (*fn)(void *, void *),
void (*reply_fn)(void *), void (*free_fn)(void *),
void *arg) void *arg)
{ {
int i = 0; int i, n_threads;
workerthread_t *worker; void (*old_args_free_fn)(void *arg);
void *arg_copy; void **old_args;
while (1) { void **new_args;
tor_mutex_acquire(&pool->lock);
if (i >= pool->n_threads) {
tor_mutex_release(&pool->lock);
return 0;
}
worker = pool->threads[i++];
tor_mutex_release(&pool->lock);
arg_copy = dup_fn ? dup_fn(arg) : arg; tor_mutex_acquire(&pool->lock);
/* CHECK*/ workerthread_queue_work(worker, fn, reply_fn, arg_copy); n_threads = pool->n_threads;
old_args = pool->update_args;
old_args_free_fn = pool->free_update_arg_fn;
new_args = tor_calloc(n_threads, sizeof(void*));
for (i = 0; i < n_threads; ++i) {
if (dup_fn)
new_args[i] = dup_fn(arg);
else
new_args[i] = arg;
} }
pool->update_args = new_args;
pool->free_update_arg_fn = free_fn;
pool->update_fn = fn;
++pool->generation;
tor_mutex_release(&pool->lock);
tor_cond_signal_all(&pool->condition);
if (old_args) {
for (i = 0; i < n_threads; ++i) {
if (old_args[i] && old_args_free_fn)
old_args_free_fn(old_args[i]);
}
tor_free(old_args);
}
return 0;
} }
/** Launch threads until we have <b>n</b>. */ /** Launch threads until we have <b>n</b>. */
@ -346,7 +386,8 @@ threadpool_start_threads(threadpool_t *pool, int n)
while (pool->n_threads < n) { while (pool->n_threads < n) {
void *state = pool->new_thread_state_fn(pool->new_thread_state_arg); void *state = pool->new_thread_state_fn(pool->new_thread_state_arg);
workerthread_t *thr = workerthread_new(state, pool->reply_queue); workerthread_t *thr = workerthread_new(state, pool, pool->reply_queue);
thr->index = pool->n_threads;
if (!thr) { if (!thr) {
tor_mutex_release(&pool->lock); tor_mutex_release(&pool->lock);
@ -375,7 +416,10 @@ threadpool_new(int n_threads,
{ {
threadpool_t *pool; threadpool_t *pool;
pool = tor_malloc_zero(sizeof(threadpool_t)); pool = tor_malloc_zero(sizeof(threadpool_t));
tor_mutex_init(&pool->lock); tor_mutex_init_nonrecursive(&pool->lock);
tor_cond_init(&pool->condition);
TOR_TAILQ_INIT(&pool->work);
pool->new_thread_state_fn = new_thread_state_fn; pool->new_thread_state_fn = new_thread_state_fn;
pool->new_thread_state_arg = arg; pool->new_thread_state_arg = arg;
pool->free_thread_state_fn = free_thread_state_fn; pool->free_thread_state_fn = free_thread_state_fn;
@ -447,7 +491,7 @@ replyqueue_process(replyqueue_t *queue)
workqueue_entry_t *work = TOR_TAILQ_FIRST(&queue->answers); workqueue_entry_t *work = TOR_TAILQ_FIRST(&queue->answers);
TOR_TAILQ_REMOVE(&queue->answers, work, next_work); TOR_TAILQ_REMOVE(&queue->answers, work, next_work);
tor_mutex_release(&queue->lock); tor_mutex_release(&queue->lock);
work->on_thread = NULL; work->on_pool = NULL;
work->reply_fn(work->arg); work->reply_fn(work->arg);
workqueue_entry_free(work); workqueue_entry_free(work);

View file

@ -27,11 +27,11 @@ workqueue_entry_t *threadpool_queue_work(threadpool_t *pool,
int (*fn)(void *, void *), int (*fn)(void *, void *),
void (*reply_fn)(void *), void (*reply_fn)(void *),
void *arg); void *arg);
int threadpool_queue_for_all(threadpool_t *pool, int threadpool_queue_update(threadpool_t *pool,
void *(*dup_fn)(void *), void *(*dup_fn)(void *),
int (*fn)(void *, void *), int (*fn)(void *, void *),
void (*reply_fn)(void *), void (*free_fn)(void *),
void *arg); void *arg);
void *workqueue_entry_cancel(workqueue_entry_t *pending_work); void *workqueue_entry_cancel(workqueue_entry_t *pending_work);
threadpool_t *threadpool_new(int n_threads, threadpool_t *threadpool_new(int n_threads,
replyqueue_t *replyqueue, replyqueue_t *replyqueue,

View file

@ -170,11 +170,6 @@ update_state_threadfn(void *state_, void *work_)
++state->generation; ++state->generation;
return WQ_RPL_REPLY; return WQ_RPL_REPLY;
} }
static void
update_state_replyfn(void *work_)
{
tor_free(work_);
}
/** Called when the onion key has changed and we need to spawn new /** Called when the onion key has changed and we need to spawn new
* cpuworkers. Close all currently idle cpuworkers, and mark the last * cpuworkers. Close all currently idle cpuworkers, and mark the last
@ -183,11 +178,11 @@ update_state_replyfn(void *work_)
void void
cpuworkers_rotate_keyinfo(void) cpuworkers_rotate_keyinfo(void)
{ {
if (threadpool_queue_for_all(threadpool, if (threadpool_queue_update(threadpool,
worker_state_new, worker_state_new,
update_state_threadfn, update_state_threadfn,
update_state_replyfn, worker_state_free,
NULL)) { NULL)) {
log_warn(LD_OR, "Failed to queue key update for worker threads."); log_warn(LD_OR, "Failed to queue key update for worker threads.");
} }
} }

View file

@ -132,7 +132,6 @@ new_state(void *arg)
/* Every thread gets its own keys. not a problem for benchmarking */ /* Every thread gets its own keys. not a problem for benchmarking */
st->rsa = crypto_pk_new(); st->rsa = crypto_pk_new();
if (crypto_pk_generate_key_with_bits(st->rsa, 1024) < 0) { if (crypto_pk_generate_key_with_bits(st->rsa, 1024) < 0) {
puts("keygen failed");
crypto_pk_free(st->rsa); crypto_pk_free(st->rsa);
tor_free(st); tor_free(st);
return NULL; return NULL;
@ -213,7 +212,6 @@ add_n_work_items(threadpool_t *tp, int n)
while (n_queued++ < n) { while (n_queued++ < n) {
ent = add_work(tp); ent = add_work(tp);
if (! ent) { if (! ent) {
puts("Couldn't add work.");
tor_event_base_loopexit(tor_libevent_get_base(), NULL); tor_event_base_loopexit(tor_libevent_get_base(), NULL);
return -1; return -1;
} }
@ -238,18 +236,6 @@ add_n_work_items(threadpool_t *tp, int n)
} }
static int shutting_down = 0; static int shutting_down = 0;
static int n_shutdowns_done = 0;
static void
shutdown_reply(void *arg)
{
(void)arg;
tor_assert(shutting_down);
++n_shutdowns_done;
if (n_shutdowns_done == opt_n_threads) {
tor_event_base_loopexit(tor_libevent_get_base(), NULL);
}
}
static void static void
replysock_readable_cb(tor_socket_t sock, short what, void *arg) replysock_readable_cb(tor_socket_t sock, short what, void *arg)
@ -297,8 +283,8 @@ replysock_readable_cb(tor_socket_t sock, short what, void *arg)
n_received+n_successful_cancel == n_sent && n_received+n_successful_cancel == n_sent &&
n_sent >= opt_n_items) { n_sent >= opt_n_items) {
shutting_down = 1; shutting_down = 1;
threadpool_queue_for_all(tp, NULL, threadpool_queue_update(tp, NULL,
workqueue_do_shutdown, shutdown_reply, NULL); workqueue_do_shutdown, NULL, NULL);
} }
} }
@ -410,8 +396,9 @@ main(int argc, char **argv)
event_base_loop(tor_libevent_get_base(), 0); event_base_loop(tor_libevent_get_base(), 0);
if (n_sent != opt_n_items || n_received+n_successful_cancel != n_sent || if (n_sent != opt_n_items || n_received+n_successful_cancel != n_sent) {
n_shutdowns_done != opt_n_threads) { printf("%d vs %d\n", n_sent, opt_n_items);
printf("%d+%d vs %d\n", n_received, n_successful_cancel, n_sent);
puts("FAIL"); puts("FAIL");
return 1; return 1;
} else { } else {