mirror of
				https://github.com/asterisk/asterisk.git
				synced 2025-10-26 22:30:28 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			1437 lines
		
	
	
		
			39 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1437 lines
		
	
	
		
			39 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Asterisk -- An open source telephony toolkit.
 | |
|  *
 | |
|  * Copyright (C) 2012-2013, Digium, Inc.
 | |
|  *
 | |
|  * Mark Michelson <mmmichelson@digium.com>
 | |
|  *
 | |
|  * See http://www.asterisk.org for more information about
 | |
|  * the Asterisk project. Please do not directly contact
 | |
|  * any of the maintainers of this project for assistance;
 | |
|  * the project provides a web site, mailing lists and IRC
 | |
|  * channels for your use.
 | |
|  *
 | |
|  * This program is free software, distributed under the terms of
 | |
|  * the GNU General Public License Version 2. See the LICENSE file
 | |
|  * at the top of the source tree.
 | |
|  */
 | |
| 
 | |
| 
 | |
| #include "asterisk.h"
 | |
| 
 | |
| #include "asterisk/threadpool.h"
 | |
| #include "asterisk/taskprocessor.h"
 | |
| #include "asterisk/astobj2.h"
 | |
| #include "asterisk/utils.h"
 | |
| 
 | |
| /* Needs to stay prime if increased */
 | |
| #define THREAD_BUCKETS 89
 | |
| 
 | |
| /*!
 | |
|  * \brief An opaque threadpool structure
 | |
|  *
 | |
|  * A threadpool is a collection of threads that execute
 | |
|  * tasks from a common queue.
 | |
|  */
 | |
| struct ast_threadpool {
 | |
| 	/*! Threadpool listener */
 | |
| 	struct ast_threadpool_listener *listener;
 | |
| 	/*!
 | |
| 	 * \brief The container of active threads.
 | |
| 	 * Active threads are those that are currently running tasks
 | |
| 	 */
 | |
| 	struct ao2_container *active_threads;
 | |
| 	/*!
 | |
| 	 * \brief The container of idle threads.
 | |
| 	 * Idle threads are those that are currently waiting to run tasks
 | |
| 	 */
 | |
| 	struct ao2_container *idle_threads;
 | |
| 	/*!
 | |
| 	 * \brief The container of zombie threads.
 | |
| 	 * Zombie threads may be running tasks, but they are scheduled to die soon
 | |
| 	 */
 | |
| 	struct ao2_container *zombie_threads;
 | |
| 	/*!
 | |
| 	 * \brief The main taskprocessor
 | |
| 	 *
 | |
| 	 * Tasks that are queued in this taskprocessor are
 | |
| 	 * doled out to the worker threads. Worker threads that
 | |
| 	 * execute tasks from the threadpool are executing tasks
 | |
| 	 * in this taskprocessor.
 | |
| 	 *
 | |
| 	 * The threadpool itself is actually the private data for
 | |
| 	 * this taskprocessor's listener. This way, as taskprocessor
 | |
| 	 * changes occur, the threadpool can alert its listeners
 | |
| 	 * appropriately.
 | |
| 	 */
 | |
| 	struct ast_taskprocessor *tps;
 | |
| 	/*!
 | |
| 	 * \brief The control taskprocessor
 | |
| 	 *
 | |
| 	 * This is a standard taskprocessor that uses the default
 | |
| 	 * taskprocessor listener. In other words, all tasks queued to
 | |
| 	 * this taskprocessor have a single thread that executes the
 | |
| 	 * tasks.
 | |
| 	 *
 | |
| 	 * All tasks that modify the state of the threadpool and all tasks
 | |
| 	 * that call out to threadpool listeners are pushed to this
 | |
| 	 * taskprocessor.
 | |
| 	 *
 | |
| 	 * For instance, when the threadpool changes sizes, a task is put
 | |
| 	 * into this taskprocessor to do so. When it comes time to tell the
 | |
| 	 * threadpool listener that worker threads have changed state,
 | |
| 	 * the task is placed in this taskprocessor.
 | |
| 	 *
 | |
| 	 * This is done for three main reasons
 | |
| 	 * 1) It ensures that listeners are given an accurate portrayal
 | |
| 	 * of the threadpool's current state. In other words, when a listener
 | |
| 	 * gets told a count of active, idle and zombie threads, it does not
 | |
| 	 * need to worry that internal state of the threadpool might be different
 | |
| 	 * from what it has been told.
 | |
| 	 * 2) It minimizes the locking required in both the threadpool and in
 | |
| 	 * threadpool listener's callbacks.
 | |
| 	 * 3) It ensures that listener callbacks are called in the same order
 | |
| 	 * that the threadpool had its state change.
 | |
| 	 */
 | |
| 	struct ast_taskprocessor *control_tps;
 | |
| 	/*! True if the threadpool is in the process of shutting down */
 | |
| 	int shutting_down;
 | |
| 	/*! Threadpool-specific options */
 | |
| 	struct ast_threadpool_options options;
 | |
| };
 | |
| 
 | |
| /*!
 | |
|  * \brief listener for a threadpool
 | |
|  *
 | |
|  * The listener is notified of changes in a threadpool. It can
 | |
|  * react by doing things like increasing the number of threads
 | |
|  * in the pool
 | |
|  */
 | |
| struct ast_threadpool_listener {
 | |
| 	/*! Callbacks called by the threadpool */
 | |
| 	const struct ast_threadpool_listener_callbacks *callbacks;
 | |
| 	/*! User data for the listener */
 | |
| 	void *user_data;
 | |
| };
 | |
| 
 | |
| /*!
 | |
|  * \brief states for worker threads
 | |
|  */
 | |
| enum worker_state {
 | |
| 	/*! The worker is either active or idle */
 | |
| 	ALIVE,
 | |
| 	/*!
 | |
| 	 * The worker has been asked to shut down but
 | |
| 	 * may still be in the process of executing tasks.
 | |
| 	 * This transition happens when the threadpool needs
 | |
| 	 * to shrink and needs to kill active threads in order
 | |
| 	 * to do so.
 | |
| 	 */
 | |
| 	ZOMBIE,
 | |
| 	/*!
 | |
| 	 * The worker has been asked to shut down. Typically
 | |
| 	 * only idle threads go to this state directly, but
 | |
| 	 * active threads may go straight to this state when
 | |
| 	 * the threadpool is shut down.
 | |
| 	 */
 | |
| 	DEAD,
 | |
| };
 | |
| 
 | |
| /*!
 | |
|  * A thread that executes threadpool tasks
 | |
|  */
 | |
| struct worker_thread {
 | |
| 	/*! A unique (within a run of Asterisk) ID for the thread. Used for hashing and searching */
 | |
| 	int id;
 | |
| 	/*! Condition used in conjunction with state changes */
 | |
| 	ast_cond_t cond;
 | |
| 	/*! Lock used alongside the condition for state changes */
 | |
| 	ast_mutex_t lock;
 | |
| 	/*! The actual thread that is executing tasks */
 | |
| 	pthread_t thread;
 | |
| 	/*! A pointer to the threadpool. Needed to be able to execute tasks */
 | |
| 	struct ast_threadpool *pool;
 | |
| 	/*! The current state of the worker thread */
 | |
| 	enum worker_state state;
 | |
| 	/*! A boolean used to determine if an idle thread should become active */
 | |
| 	int wake_up;
 | |
| 	/*! Options for this threadpool */
 | |
| 	struct ast_threadpool_options options;
 | |
| };
 | |
| 
 | |
| /* Worker thread forward declarations. See definitions for documentation */
 | |
| static int worker_thread_hash(const void *obj, int flags);
 | |
| static int worker_thread_cmp(void *obj, void *arg, int flags);
 | |
| static void worker_thread_destroy(void *obj);
 | |
| static void worker_active(struct worker_thread *worker);
 | |
| static void *worker_start(void *arg);
 | |
| static struct worker_thread *worker_thread_alloc(struct ast_threadpool *pool);
 | |
| static int worker_thread_start(struct worker_thread *worker);
 | |
| static int worker_idle(struct worker_thread *worker);
 | |
| static int worker_set_state(struct worker_thread *worker, enum worker_state state);
 | |
| static void worker_shutdown(struct worker_thread *worker);
 | |
| 
 | |
| /*!
 | |
|  * \brief Notify the threadpool listener that the state has changed.
 | |
|  *
 | |
|  * This notifies the threadpool listener via its state_changed callback.
 | |
|  * \param pool The threadpool whose state has changed
 | |
|  */
 | |
| static void threadpool_send_state_changed(struct ast_threadpool *pool)
 | |
| {
 | |
| 	int active_size = ao2_container_count(pool->active_threads);
 | |
| 	int idle_size = ao2_container_count(pool->idle_threads);
 | |
| 
 | |
| 	if (pool->listener && pool->listener->callbacks->state_changed) {
 | |
| 		pool->listener->callbacks->state_changed(pool, pool->listener, active_size, idle_size);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Struct used for queued operations involving worker state changes
 | |
|  */
 | |
| struct thread_worker_pair {
 | |
| 	/*! Threadpool that contains the worker whose state has changed */
 | |
| 	struct ast_threadpool *pool;
 | |
| 	/*! Worker whose state has changed */
 | |
| 	struct worker_thread *worker;
 | |
| };
 | |
| 
 | |
| /*!
 | |
|  * \brief Destructor for thread_worker_pair
 | |
|  */
 | |
| static void thread_worker_pair_free(struct thread_worker_pair *pair)
 | |
| {
 | |
| 	ao2_ref(pair->worker, -1);
 | |
| 	ast_free(pair);
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Allocate and initialize a thread_worker_pair
 | |
|  * \param pool Threadpool to assign to the thread_worker_pair
 | |
|  * \param worker Worker thread to assign to the thread_worker_pair
 | |
|  */
 | |
| static struct thread_worker_pair *thread_worker_pair_alloc(struct ast_threadpool *pool,
 | |
| 		struct worker_thread *worker)
 | |
| {
 | |
| 	struct thread_worker_pair *pair = ast_malloc(sizeof(*pair));
 | |
| 	if (!pair) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	pair->pool = pool;
 | |
| 	ao2_ref(worker, +1);
 | |
| 	pair->worker = worker;
 | |
| 
 | |
| 	return pair;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Move a worker thread from the active container to the idle container.
 | |
|  *
 | |
|  * This function is called from the threadpool's control taskprocessor thread.
 | |
|  * \param data A thread_worker_pair containing the threadpool and the worker to move.
 | |
|  * \return 0
 | |
|  */
 | |
| static int queued_active_thread_idle(void *data)
 | |
| {
 | |
| 	struct thread_worker_pair *pair = data;
 | |
| 
 | |
| 	ao2_link(pair->pool->idle_threads, pair->worker);
 | |
| 	ao2_unlink(pair->pool->active_threads, pair->worker);
 | |
| 
 | |
| 	threadpool_send_state_changed(pair->pool);
 | |
| 
 | |
| 	thread_worker_pair_free(pair);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Queue a task to move a thread from the active list to the idle list
 | |
|  *
 | |
|  * This is called by a worker thread when it runs out of tasks to perform and
 | |
|  * goes idle.
 | |
|  * \param pool The threadpool to which the worker belongs
 | |
|  * \param worker The worker thread that has gone idle
 | |
|  */
 | |
| static void threadpool_active_thread_idle(struct ast_threadpool *pool,
 | |
| 		struct worker_thread *worker)
 | |
| {
 | |
| 	struct thread_worker_pair *pair;
 | |
| 	SCOPED_AO2LOCK(lock, pool);
 | |
| 
 | |
| 	if (pool->shutting_down) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	pair = thread_worker_pair_alloc(pool, worker);
 | |
| 	if (!pair) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if (ast_taskprocessor_push(pool->control_tps, queued_active_thread_idle, pair)) {
 | |
| 		thread_worker_pair_free(pair);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Kill a zombie thread
 | |
|  *
 | |
|  * This runs from the threadpool's control taskprocessor thread.
 | |
|  *
 | |
|  * \param data A thread_worker_pair containing the threadpool and the zombie thread
 | |
|  * \return 0
 | |
|  */
 | |
| static int queued_zombie_thread_dead(void *data)
 | |
| {
 | |
| 	struct thread_worker_pair *pair = data;
 | |
| 
 | |
| 	ao2_unlink(pair->pool->zombie_threads, pair->worker);
 | |
| 	threadpool_send_state_changed(pair->pool);
 | |
| 
 | |
| 	thread_worker_pair_free(pair);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Queue a task to kill a zombie thread
 | |
|  *
 | |
|  * This is called by a worker thread when it acknowledges that it is time for
 | |
|  * it to die.
 | |
|  */
 | |
| static void threadpool_zombie_thread_dead(struct ast_threadpool *pool,
 | |
| 		struct worker_thread *worker)
 | |
| {
 | |
| 	struct thread_worker_pair *pair;
 | |
| 	SCOPED_AO2LOCK(lock, pool);
 | |
| 
 | |
| 	if (pool->shutting_down) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	pair = thread_worker_pair_alloc(pool, worker);
 | |
| 	if (!pair) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if (ast_taskprocessor_push(pool->control_tps, queued_zombie_thread_dead, pair)) {
 | |
| 		thread_worker_pair_free(pair);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int queued_idle_thread_dead(void *data)
 | |
| {
 | |
| 	struct thread_worker_pair *pair = data;
 | |
| 
 | |
| 	ao2_unlink(pair->pool->idle_threads, pair->worker);
 | |
| 	threadpool_send_state_changed(pair->pool);
 | |
| 
 | |
| 	thread_worker_pair_free(pair);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void threadpool_idle_thread_dead(struct ast_threadpool *pool,
 | |
| 		struct worker_thread *worker)
 | |
| {
 | |
| 	struct thread_worker_pair *pair;
 | |
| 	SCOPED_AO2LOCK(lock, pool);
 | |
| 
 | |
| 	if (pool->shutting_down) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	pair = thread_worker_pair_alloc(pool, worker);
 | |
| 	if (!pair) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if (ast_taskprocessor_push(pool->control_tps, queued_idle_thread_dead, pair)) {
 | |
| 		thread_worker_pair_free(pair);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Execute a task in the threadpool
 | |
|  *
 | |
|  * This is the function that worker threads call in order to execute tasks
 | |
|  * in the threadpool
 | |
|  *
 | |
|  * \param pool The pool to which the tasks belong.
 | |
|  * \retval 0 Either the pool has been shut down or there are no tasks.
 | |
|  * \retval 1 There are still tasks remaining in the pool.
 | |
|  */
 | |
| static int threadpool_execute(struct ast_threadpool *pool)
 | |
| {
 | |
| 	ao2_lock(pool);
 | |
| 	if (!pool->shutting_down) {
 | |
| 		ao2_unlock(pool);
 | |
| 		return ast_taskprocessor_execute(pool->tps);
 | |
| 	}
 | |
| 	ao2_unlock(pool);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Destroy a threadpool's components.
 | |
|  *
 | |
|  * This is the destructor called automatically when the threadpool's
 | |
|  * reference count reaches zero. This is not to be confused with
 | |
|  * threadpool_destroy.
 | |
|  *
 | |
|  * By the time this actually gets called, most of the cleanup has already
 | |
|  * been done in the pool. The only thing left to do is to release the
 | |
|  * final reference to the threadpool listener.
 | |
|  *
 | |
|  * \param obj The pool to destroy
 | |
|  */
 | |
| static void threadpool_destructor(void *obj)
 | |
| {
 | |
| 	struct ast_threadpool *pool = obj;
 | |
| 	ao2_cleanup(pool->listener);
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Allocate a threadpool
 | |
|  *
 | |
|  * This is implemented as a taskprocessor listener's alloc callback. This
 | |
|  * is because the threadpool exists as the private data on a taskprocessor
 | |
|  * listener.
 | |
|  *
 | |
|  * \param name The name of the threadpool.
 | |
|  * \param options The options the threadpool uses.
 | |
|  * \retval NULL Could not initialize threadpool properly
 | |
|  * \retval non-NULL The newly-allocated threadpool
 | |
|  */
 | |
| static struct ast_threadpool *threadpool_alloc(const char *name, const struct ast_threadpool_options *options)
 | |
| {
 | |
| 	RAII_VAR(struct ast_threadpool *, pool, NULL, ao2_cleanup);
 | |
| 	struct ast_str *control_tps_name;
 | |
| 
 | |
| 	pool = ao2_alloc(sizeof(*pool), threadpool_destructor);
 | |
| 	control_tps_name = ast_str_create(64);
 | |
| 	if (!pool || !control_tps_name) {
 | |
| 		ast_free(control_tps_name);
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	ast_str_set(&control_tps_name, 0, "%s/pool-control", name);
 | |
| 
 | |
| 	pool->control_tps = ast_taskprocessor_get(ast_str_buffer(control_tps_name), TPS_REF_DEFAULT);
 | |
| 	ast_free(control_tps_name);
 | |
| 	if (!pool->control_tps) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	pool->active_threads = ao2_container_alloc_hash(AO2_ALLOC_OPT_LOCK_MUTEX, 0,
 | |
| 		THREAD_BUCKETS, worker_thread_hash, NULL, worker_thread_cmp);
 | |
| 	if (!pool->active_threads) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	pool->idle_threads = ao2_container_alloc_hash(AO2_ALLOC_OPT_LOCK_MUTEX, 0,
 | |
| 		THREAD_BUCKETS, worker_thread_hash, NULL, worker_thread_cmp);
 | |
| 	if (!pool->idle_threads) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	pool->zombie_threads = ao2_container_alloc_hash(AO2_ALLOC_OPT_LOCK_MUTEX, 0,
 | |
| 		THREAD_BUCKETS, worker_thread_hash, NULL, worker_thread_cmp);
 | |
| 	if (!pool->zombie_threads) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	pool->options = *options;
 | |
| 
 | |
| 	ao2_ref(pool, +1);
 | |
| 	return pool;
 | |
| }
 | |
| 
 | |
| static int threadpool_tps_start(struct ast_taskprocessor_listener *listener)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief helper used for queued task when tasks are pushed
 | |
|  */
 | |
| struct task_pushed_data {
 | |
| 	/*! Pool into which a task was pushed */
 | |
| 	struct ast_threadpool *pool;
 | |
| 	/*! Indicator of whether the pool had no tasks prior to the new task being added */
 | |
| 	int was_empty;
 | |
| };
 | |
| 
 | |
| /*!
 | |
|  * \brief Allocate and initialize a task_pushed_data
 | |
|  * \param pool The threadpool to set in the task_pushed_data
 | |
|  * \param was_empty The was_empty value to set in the task_pushed_data
 | |
|  * \retval NULL Unable to allocate task_pushed_data
 | |
|  * \retval non-NULL The newly-allocated task_pushed_data
 | |
|  */
 | |
| static struct task_pushed_data *task_pushed_data_alloc(struct ast_threadpool *pool,
 | |
| 		int was_empty)
 | |
| {
 | |
| 	struct task_pushed_data *tpd = ast_malloc(sizeof(*tpd));
 | |
| 
 | |
| 	if (!tpd) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	tpd->pool = pool;
 | |
| 	tpd->was_empty = was_empty;
 | |
| 	return tpd;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Activate idle threads
 | |
|  *
 | |
|  * This function always returns CMP_MATCH because all workers that this
 | |
|  * function acts on need to be seen as matches so they are unlinked from the
 | |
|  * list of idle threads.
 | |
|  *
 | |
|  * Called as an ao2_callback in the threadpool's control taskprocessor thread.
 | |
|  * \param obj The worker to activate
 | |
|  * \param arg The pool where the worker belongs
 | |
|  * \param flags
 | |
|  * \retval CMP_MATCH
 | |
|  */
 | |
| static int activate_thread(void *obj, void *arg, int flags)
 | |
| {
 | |
| 	struct worker_thread *worker = obj;
 | |
| 	struct ast_threadpool *pool = arg;
 | |
| 
 | |
| 	if (!ao2_link(pool->active_threads, worker)) {
 | |
| 		/* If we can't link the idle thread into the active container, then
 | |
| 		 * we'll just leave the thread idle and not wake it up.
 | |
| 		 */
 | |
| 		ast_log(LOG_WARNING, "Failed to activate thread %d. Remaining idle\n",
 | |
| 				worker->id);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	if (worker_set_state(worker, ALIVE)) {
 | |
| 		ast_debug(1, "Failed to activate thread %d. It is dead\n",
 | |
| 				worker->id);
 | |
| 		/* The worker thread will no longer exist in the active threads or
 | |
| 		 * idle threads container after this.
 | |
| 		 */
 | |
| 		ao2_unlink(pool->active_threads, worker);
 | |
| 	}
 | |
| 
 | |
| 	return CMP_MATCH;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Add threads to the threadpool
 | |
|  *
 | |
|  * This function is called from the threadpool's control taskprocessor thread.
 | |
|  * \param pool The pool that is expanding
 | |
|  * \param delta The number of threads to add to the pool
 | |
|  */
 | |
| static void grow(struct ast_threadpool *pool, int delta)
 | |
| {
 | |
| 	int i;
 | |
| 
 | |
| 	int current_size = ao2_container_count(pool->active_threads) +
 | |
| 		ao2_container_count(pool->idle_threads);
 | |
| 
 | |
| 	if (pool->options.max_size && current_size + delta > pool->options.max_size) {
 | |
| 		delta = pool->options.max_size - current_size;
 | |
| 	}
 | |
| 
 | |
| 	ast_debug(3, "Increasing threadpool %s's size by %d\n",
 | |
| 			ast_taskprocessor_name(pool->tps), delta);
 | |
| 
 | |
| 	for (i = 0; i < delta; ++i) {
 | |
| 		struct worker_thread *worker = worker_thread_alloc(pool);
 | |
| 		if (!worker) {
 | |
| 			return;
 | |
| 		}
 | |
| 		if (ao2_link(pool->idle_threads, worker)) {
 | |
| 			if (worker_thread_start(worker)) {
 | |
| 				ast_log(LOG_ERROR, "Unable to start worker thread %d. Destroying.\n", worker->id);
 | |
| 				ao2_unlink(pool->active_threads, worker);
 | |
| 			}
 | |
| 		} else {
 | |
| 			ast_log(LOG_WARNING, "Failed to activate worker thread %d. Destroying.\n", worker->id);
 | |
| 		}
 | |
| 		ao2_ref(worker, -1);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Queued task called when tasks are pushed into the threadpool
 | |
|  *
 | |
|  * This function first calls into the threadpool's listener to let it know
 | |
|  * that a task has been pushed. It then wakes up all idle threads and moves
 | |
|  * them into the active thread container.
 | |
|  * \param data A task_pushed_data
 | |
|  * \return 0
 | |
|  */
 | |
| static int queued_task_pushed(void *data)
 | |
| {
 | |
| 	struct task_pushed_data *tpd = data;
 | |
| 	struct ast_threadpool *pool = tpd->pool;
 | |
| 	int was_empty = tpd->was_empty;
 | |
| 	unsigned int existing_active;
 | |
| 
 | |
| 	ast_free(tpd);
 | |
| 
 | |
| 	if (pool->listener && pool->listener->callbacks->task_pushed) {
 | |
| 		pool->listener->callbacks->task_pushed(pool, pool->listener, was_empty);
 | |
| 	}
 | |
| 
 | |
| 	existing_active = ao2_container_count(pool->active_threads);
 | |
| 
 | |
| 	/* The first pass transitions any existing idle threads to be active, and
 | |
| 	 * will also remove any worker threads that have recently entered the dead
 | |
| 	 * state.
 | |
| 	 */
 | |
| 	ao2_callback(pool->idle_threads, OBJ_UNLINK | OBJ_NOLOCK | OBJ_NODATA,
 | |
| 			activate_thread, pool);
 | |
| 
 | |
| 	/* If no idle threads could be transitioned to active grow the pool as permitted. */
 | |
| 	if (ao2_container_count(pool->active_threads) == existing_active) {
 | |
| 		if (!pool->options.auto_increment) {
 | |
| 			return 0;
 | |
| 		}
 | |
| 		grow(pool, pool->options.auto_increment);
 | |
| 		/* An optional second pass transitions any newly added threads. */
 | |
| 		ao2_callback(pool->idle_threads, OBJ_UNLINK | OBJ_NOLOCK | OBJ_NODATA,
 | |
| 				activate_thread, pool);
 | |
| 	}
 | |
| 
 | |
| 	threadpool_send_state_changed(pool);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Taskprocessor listener callback called when a task is added
 | |
|  *
 | |
|  * The threadpool uses this opportunity to queue a task on its control taskprocessor
 | |
|  * in order to activate idle threads and notify the threadpool listener that the
 | |
|  * task has been pushed.
 | |
|  * \param listener The taskprocessor listener. The threadpool is the listener's private data
 | |
|  * \param was_empty True if the taskprocessor was empty prior to the task being pushed
 | |
|  */
 | |
| static void threadpool_tps_task_pushed(struct ast_taskprocessor_listener *listener,
 | |
| 		int was_empty)
 | |
| {
 | |
| 	struct ast_threadpool *pool = ast_taskprocessor_listener_get_user_data(listener);
 | |
| 	struct task_pushed_data *tpd;
 | |
| 	SCOPED_AO2LOCK(lock, pool);
 | |
| 
 | |
| 	if (pool->shutting_down) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	tpd = task_pushed_data_alloc(pool, was_empty);
 | |
| 	if (!tpd) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if (ast_taskprocessor_push(pool->control_tps, queued_task_pushed, tpd)) {
 | |
| 		ast_free(tpd);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Queued task that handles the case where the threadpool's taskprocessor is emptied
 | |
|  *
 | |
|  * This simply lets the threadpool's listener know that the threadpool is devoid of tasks
 | |
|  * \param data The pool that has become empty
 | |
|  * \return 0
 | |
|  */
 | |
| static int queued_emptied(void *data)
 | |
| {
 | |
| 	struct ast_threadpool *pool = data;
 | |
| 
 | |
| 	/* We already checked for existence of this callback when this was queued */
 | |
| 	pool->listener->callbacks->emptied(pool, pool->listener);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Taskprocessor listener emptied callback
 | |
|  *
 | |
|  * The threadpool queues a task to let the threadpool listener know that
 | |
|  * the threadpool no longer contains any tasks.
 | |
|  * \param listener The taskprocessor listener. The threadpool is the listener's private data.
 | |
|  */
 | |
| static void threadpool_tps_emptied(struct ast_taskprocessor_listener *listener)
 | |
| {
 | |
| 	struct ast_threadpool *pool = ast_taskprocessor_listener_get_user_data(listener);
 | |
| 	SCOPED_AO2LOCK(lock, pool);
 | |
| 
 | |
| 	if (pool->shutting_down) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if (pool->listener && pool->listener->callbacks->emptied) {
 | |
| 		if (ast_taskprocessor_push(pool->control_tps, queued_emptied, pool)) {
 | |
| 			/* Nothing to do here but we need the check to keep the compiler happy. */
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Taskprocessor listener shutdown callback
 | |
|  *
 | |
|  * The threadpool will shut down and destroy all of its worker threads when
 | |
|  * this is called back. By the time this gets called, the taskprocessor's
 | |
|  * control taskprocessor has already been destroyed. Therefore there is no risk
 | |
|  * in outright destroying the worker threads here.
 | |
|  * \param listener The taskprocessor listener. The threadpool is the listener's private data.
 | |
|  */
 | |
| static void threadpool_tps_shutdown(struct ast_taskprocessor_listener *listener)
 | |
| {
 | |
| 	struct ast_threadpool *pool = ast_taskprocessor_listener_get_user_data(listener);
 | |
| 
 | |
| 	if (pool->listener && pool->listener->callbacks->shutdown) {
 | |
| 		pool->listener->callbacks->shutdown(pool->listener);
 | |
| 	}
 | |
| 	ao2_cleanup(pool->active_threads);
 | |
| 	ao2_cleanup(pool->idle_threads);
 | |
| 	ao2_cleanup(pool->zombie_threads);
 | |
| 	ao2_cleanup(pool);
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Table of taskprocessor listener callbacks for threadpool's main taskprocessor
 | |
|  */
 | |
| static struct ast_taskprocessor_listener_callbacks threadpool_tps_listener_callbacks = {
 | |
| 	.start = threadpool_tps_start,
 | |
| 	.task_pushed = threadpool_tps_task_pushed,
 | |
| 	.emptied = threadpool_tps_emptied,
 | |
| 	.shutdown = threadpool_tps_shutdown,
 | |
| };
 | |
| 
 | |
| /*!
 | |
|  * \brief ao2 callback to kill a set number of threads.
 | |
|  *
 | |
|  * Threads will be unlinked from the container as long as the
 | |
|  * counter has not reached zero. The counter is decremented with
 | |
|  * each thread that is removed.
 | |
|  * \param obj The worker thread up for possible destruction
 | |
|  * \param arg The counter
 | |
|  * \param flags Unused
 | |
|  * \retval CMP_MATCH The counter has not reached zero, so this flag should be removed.
 | |
|  * \retval CMP_STOP The counter has reached zero so no more threads should be removed.
 | |
|  */
 | |
| static int kill_threads(void *obj, void *arg, int flags)
 | |
| {
 | |
| 	int *num_to_kill = arg;
 | |
| 
 | |
| 	if (*num_to_kill > 0) {
 | |
| 		--(*num_to_kill);
 | |
| 		return CMP_MATCH;
 | |
| 	} else {
 | |
| 		return CMP_STOP;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief ao2 callback to zombify a set number of threads.
 | |
|  *
 | |
|  * Threads will be zombified as long as the counter has not reached
 | |
|  * zero. The counter is decremented with each thread that is zombified.
 | |
|  *
 | |
|  * Zombifying a thread involves removing it from its current container,
 | |
|  * adding it to the zombie container, and changing the state of the
 | |
|  * worker to a zombie
 | |
|  *
 | |
|  * This callback is called from the threadpool control taskprocessor thread.
 | |
|  *
 | |
|  * \param obj The worker thread that may be zombified
 | |
|  * \param arg The pool to which the worker belongs
 | |
|  * \param data The counter
 | |
|  * \param flags Unused
 | |
|  * \retval CMP_MATCH The zombified thread should be removed from its current container
 | |
|  * \retval CMP_STOP Stop attempting to zombify threads
 | |
|  */
 | |
| static int zombify_threads(void *obj, void *arg, void *data, int flags)
 | |
| {
 | |
| 	struct worker_thread *worker = obj;
 | |
| 	struct ast_threadpool *pool = arg;
 | |
| 	int *num_to_zombify = data;
 | |
| 
 | |
| 	if ((*num_to_zombify)-- > 0) {
 | |
| 		if (!ao2_link(pool->zombie_threads, worker)) {
 | |
| 			ast_log(LOG_WARNING, "Failed to zombify active thread %d. Thread will remain active\n", worker->id);
 | |
| 			return 0;
 | |
| 		}
 | |
| 		worker_set_state(worker, ZOMBIE);
 | |
| 		return CMP_MATCH;
 | |
| 	} else {
 | |
| 		return CMP_STOP;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Remove threads from the threadpool
 | |
|  *
 | |
|  * The preference is to kill idle threads. However, if there are
 | |
|  * more threads to remove than there are idle threads, then active
 | |
|  * threads will be zombified instead.
 | |
|  *
 | |
|  * This function is called from the threadpool control taskprocessor thread.
 | |
|  *
 | |
|  * \param pool The threadpool to remove threads from
 | |
|  * \param delta The number of threads to remove
 | |
|  */
 | |
| static void shrink(struct ast_threadpool *pool, int delta)
 | |
| {
 | |
| 	/*
 | |
| 	 * Preference is to kill idle threads, but
 | |
| 	 * we'll move on to deactivating active threads
 | |
| 	 * if we have to
 | |
| 	 */
 | |
| 	int idle_threads = ao2_container_count(pool->idle_threads);
 | |
| 	int idle_threads_to_kill = MIN(delta, idle_threads);
 | |
| 	int active_threads_to_zombify = delta - idle_threads_to_kill;
 | |
| 
 | |
| 	ast_debug(3, "Destroying %d idle threads in threadpool %s\n", idle_threads_to_kill,
 | |
| 			ast_taskprocessor_name(pool->tps));
 | |
| 
 | |
| 	ao2_callback(pool->idle_threads, OBJ_UNLINK | OBJ_NOLOCK | OBJ_NODATA | OBJ_MULTIPLE,
 | |
| 			kill_threads, &idle_threads_to_kill);
 | |
| 
 | |
| 	ast_debug(3, "Destroying %d active threads in threadpool %s\n", active_threads_to_zombify,
 | |
| 			ast_taskprocessor_name(pool->tps));
 | |
| 
 | |
| 	ao2_callback_data(pool->active_threads, OBJ_UNLINK | OBJ_NOLOCK | OBJ_NODATA | OBJ_MULTIPLE,
 | |
| 			zombify_threads, pool, &active_threads_to_zombify);
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Helper struct used for queued operations that change the size of the threadpool
 | |
|  */
 | |
| struct set_size_data {
 | |
| 	/*! The pool whose size is to change */
 | |
| 	struct ast_threadpool *pool;
 | |
| 	/*! The requested new size of the pool */
 | |
| 	unsigned int size;
 | |
| };
 | |
| 
 | |
| /*!
 | |
|  * \brief Allocate and initialize a set_size_data
 | |
|  * \param pool The pool for the set_size_data
 | |
|  * \param size The size to store in the set_size_data
 | |
|  */
 | |
| static struct set_size_data *set_size_data_alloc(struct ast_threadpool *pool,
 | |
| 		unsigned int size)
 | |
| {
 | |
| 	struct set_size_data *ssd = ast_malloc(sizeof(*ssd));
 | |
| 	if (!ssd) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	ssd->pool = pool;
 | |
| 	ssd->size = size;
 | |
| 	return ssd;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Change the size of the threadpool
 | |
|  *
 | |
|  * This can either result in shrinking or growing the threadpool depending
 | |
|  * on the new desired size and the current size.
 | |
|  *
 | |
|  * This function is run from the threadpool control taskprocessor thread
 | |
|  *
 | |
|  * \param data A set_size_data used for determining how to act
 | |
|  * \return 0
 | |
|  */
 | |
| static int queued_set_size(void *data)
 | |
| {
 | |
| 	struct set_size_data *ssd = data;
 | |
| 	struct ast_threadpool *pool = ssd->pool;
 | |
| 	unsigned int num_threads = ssd->size;
 | |
| 
 | |
| 	/* We don't count zombie threads as being "live" when potentially resizing */
 | |
| 	unsigned int current_size = ao2_container_count(pool->active_threads) +
 | |
| 			ao2_container_count(pool->idle_threads);
 | |
| 
 | |
| 	ast_free(ssd);
 | |
| 
 | |
| 	if (current_size == num_threads) {
 | |
| 		ast_debug(3, "Not changing threadpool size since new size %u is the same as current %u\n",
 | |
| 			  num_threads, current_size);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	if (current_size < num_threads) {
 | |
| 		ao2_callback(pool->idle_threads, OBJ_UNLINK | OBJ_NOLOCK | OBJ_NODATA | OBJ_MULTIPLE,
 | |
| 				activate_thread, pool);
 | |
| 
 | |
| 		/* As the above may have altered the number of current threads update it */
 | |
| 		current_size = ao2_container_count(pool->active_threads) +
 | |
| 				ao2_container_count(pool->idle_threads);
 | |
| 		grow(pool, num_threads - current_size);
 | |
| 		ao2_callback(pool->idle_threads, OBJ_UNLINK | OBJ_NOLOCK | OBJ_NODATA | OBJ_MULTIPLE,
 | |
| 				activate_thread, pool);
 | |
| 	} else {
 | |
| 		shrink(pool, current_size - num_threads);
 | |
| 	}
 | |
| 
 | |
| 	threadpool_send_state_changed(pool);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void ast_threadpool_set_size(struct ast_threadpool *pool, unsigned int size)
 | |
| {
 | |
| 	struct set_size_data *ssd;
 | |
| 	SCOPED_AO2LOCK(lock, pool);
 | |
| 
 | |
| 	if (pool->shutting_down) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	ssd = set_size_data_alloc(pool, size);
 | |
| 	if (!ssd) {
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if (ast_taskprocessor_push(pool->control_tps, queued_set_size, ssd)) {
 | |
| 		ast_free(ssd);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| struct ast_threadpool_listener *ast_threadpool_listener_alloc(
 | |
| 		const struct ast_threadpool_listener_callbacks *callbacks, void *user_data)
 | |
| {
 | |
| 	struct ast_threadpool_listener *listener = ao2_alloc(sizeof(*listener), NULL);
 | |
| 	if (!listener) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	listener->callbacks = callbacks;
 | |
| 	listener->user_data = user_data;
 | |
| 	return listener;
 | |
| }
 | |
| 
 | |
| void *ast_threadpool_listener_get_user_data(const struct ast_threadpool_listener *listener)
 | |
| {
 | |
| 	return listener->user_data;
 | |
| }
 | |
| 
 | |
| struct pool_options_pair {
 | |
| 	struct ast_threadpool *pool;
 | |
| 	struct ast_threadpool_options options;
 | |
| };
 | |
| 
 | |
| struct ast_threadpool *ast_threadpool_create(const char *name,
 | |
| 		struct ast_threadpool_listener *listener,
 | |
| 		const struct ast_threadpool_options *options)
 | |
| {
 | |
| 	struct ast_taskprocessor *tps;
 | |
| 	RAII_VAR(struct ast_taskprocessor_listener *, tps_listener, NULL, ao2_cleanup);
 | |
| 	RAII_VAR(struct ast_threadpool *, pool, NULL, ao2_cleanup);
 | |
| 	char *fullname;
 | |
| 
 | |
| 	pool = threadpool_alloc(name, options);
 | |
| 	if (!pool) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	tps_listener = ast_taskprocessor_listener_alloc(&threadpool_tps_listener_callbacks, pool);
 | |
| 	if (!tps_listener) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	if (options->version != AST_THREADPOOL_OPTIONS_VERSION) {
 | |
| 		ast_log(LOG_WARNING, "Incompatible version of threadpool options in use.\n");
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	fullname = ast_alloca(strlen(name) + strlen("/pool") + 1);
 | |
| 	sprintf(fullname, "%s/pool", name); /* Safe */
 | |
| 	tps = ast_taskprocessor_create_with_listener(fullname, tps_listener);
 | |
| 	if (!tps) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	pool->tps = tps;
 | |
| 	if (listener) {
 | |
| 		ao2_ref(listener, +1);
 | |
| 		pool->listener = listener;
 | |
| 	}
 | |
| 	ast_threadpool_set_size(pool, pool->options.initial_size);
 | |
| 	ao2_ref(pool, +1);
 | |
| 	return pool;
 | |
| }
 | |
| 
 | |
| int ast_threadpool_push(struct ast_threadpool *pool, int (*task)(void *data), void *data)
 | |
| {
 | |
| 	SCOPED_AO2LOCK(lock, pool);
 | |
| 	if (!pool->shutting_down) {
 | |
| 		return ast_taskprocessor_push(pool->tps, task, data);
 | |
| 	}
 | |
| 	return -1;
 | |
| }
 | |
| 
 | |
| void ast_threadpool_shutdown(struct ast_threadpool *pool)
 | |
| {
 | |
| 	if (!pool) {
 | |
| 		return;
 | |
| 	}
 | |
| 	/* Shut down the taskprocessors and everything else just
 | |
| 	 * takes care of itself via the taskprocessor callbacks
 | |
| 	 */
 | |
| 	ao2_lock(pool);
 | |
| 	pool->shutting_down = 1;
 | |
| 	ao2_unlock(pool);
 | |
| 	ast_taskprocessor_unreference(pool->control_tps);
 | |
| 	ast_taskprocessor_unreference(pool->tps);
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * A monotonically increasing integer used for worker
 | |
|  * thread identification.
 | |
|  */
 | |
| static int worker_id_counter;
 | |
| 
 | |
| static int worker_thread_hash(const void *obj, int flags)
 | |
| {
 | |
| 	const struct worker_thread *worker = obj;
 | |
| 
 | |
| 	return worker->id;
 | |
| }
 | |
| 
 | |
| static int worker_thread_cmp(void *obj, void *arg, int flags)
 | |
| {
 | |
| 	struct worker_thread *worker1 = obj;
 | |
| 	struct worker_thread *worker2 = arg;
 | |
| 
 | |
| 	return worker1->id == worker2->id ? CMP_MATCH : 0;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief shut a worker thread down
 | |
|  *
 | |
|  * Set the worker dead and then wait for its thread
 | |
|  * to finish executing.
 | |
|  *
 | |
|  * \param worker The worker thread to shut down
 | |
|  */
 | |
| static void worker_shutdown(struct worker_thread *worker)
 | |
| {
 | |
| 	worker_set_state(worker, DEAD);
 | |
| 	if (worker->thread != AST_PTHREADT_NULL) {
 | |
| 		pthread_join(worker->thread, NULL);
 | |
| 		worker->thread = AST_PTHREADT_NULL;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Worker thread destructor
 | |
|  *
 | |
|  * Called automatically when refcount reaches 0. Shuts
 | |
|  * down the worker thread and destroys its component
 | |
|  * parts
 | |
|  */
 | |
| static void worker_thread_destroy(void *obj)
 | |
| {
 | |
| 	struct worker_thread *worker = obj;
 | |
| 	ast_debug(3, "Destroying worker thread %d\n", worker->id);
 | |
| 	worker_shutdown(worker);
 | |
| 	ast_mutex_destroy(&worker->lock);
 | |
| 	ast_cond_destroy(&worker->cond);
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief start point for worker threads
 | |
|  *
 | |
|  * Worker threads start in the active state but may
 | |
|  * immediately go idle if there is no work to be
 | |
|  * done
 | |
|  *
 | |
|  * \param arg The worker thread
 | |
|  */
 | |
| static void *worker_start(void *arg)
 | |
| {
 | |
| 	struct worker_thread *worker = arg;
 | |
| 	enum worker_state saved_state;
 | |
| 
 | |
| 	if (worker->options.thread_start) {
 | |
| 		worker->options.thread_start();
 | |
| 	}
 | |
| 
 | |
| 	ast_mutex_lock(&worker->lock);
 | |
| 	while (worker_idle(worker)) {
 | |
| 		ast_mutex_unlock(&worker->lock);
 | |
| 		worker_active(worker);
 | |
| 		ast_mutex_lock(&worker->lock);
 | |
| 		if (worker->state != ALIVE) {
 | |
| 			break;
 | |
| 		}
 | |
| 		threadpool_active_thread_idle(worker->pool, worker);
 | |
| 	}
 | |
| 	saved_state = worker->state;
 | |
| 	ast_mutex_unlock(&worker->lock);
 | |
| 
 | |
| 	/* Reaching this portion means the thread is
 | |
| 	 * on death's door. It may have been killed while
 | |
| 	 * it was idle, in which case it can just die
 | |
| 	 * peacefully. If it's a zombie, though, then
 | |
| 	 * it needs to let the pool know so
 | |
| 	 * that the thread can be removed from the
 | |
| 	 * list of zombie threads.
 | |
| 	 */
 | |
| 	if (saved_state == ZOMBIE) {
 | |
| 		threadpool_zombie_thread_dead(worker->pool, worker);
 | |
| 	}
 | |
| 
 | |
| 	if (worker->options.thread_end) {
 | |
| 		worker->options.thread_end();
 | |
| 	}
 | |
| 	return NULL;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Allocate and initialize a new worker thread
 | |
|  *
 | |
|  * This will create, initialize, and start the thread.
 | |
|  *
 | |
|  * \param pool The threadpool to which the worker will be added
 | |
|  * \retval NULL Failed to allocate or start the worker thread
 | |
|  * \retval non-NULL The newly-created worker thread
 | |
|  */
 | |
| static struct worker_thread *worker_thread_alloc(struct ast_threadpool *pool)
 | |
| {
 | |
| 	struct worker_thread *worker = ao2_alloc(sizeof(*worker), worker_thread_destroy);
 | |
| 	if (!worker) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	worker->id = ast_atomic_fetchadd_int(&worker_id_counter, 1);
 | |
| 	ast_mutex_init(&worker->lock);
 | |
| 	ast_cond_init(&worker->cond, NULL);
 | |
| 	worker->pool = pool;
 | |
| 	worker->thread = AST_PTHREADT_NULL;
 | |
| 	worker->state = ALIVE;
 | |
| 	worker->options = pool->options;
 | |
| 	return worker;
 | |
| }
 | |
| 
 | |
| static int worker_thread_start(struct worker_thread *worker)
 | |
| {
 | |
| 	return ast_pthread_create(&worker->thread, NULL, worker_start, worker);
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Active loop for worker threads
 | |
|  *
 | |
|  * The worker will stay in this loop for its lifetime,
 | |
|  * executing tasks as they become available. If there
 | |
|  * are no tasks currently available, then the thread
 | |
|  * will go idle.
 | |
|  *
 | |
|  * \param worker The worker thread executing tasks.
 | |
|  */
 | |
| static void worker_active(struct worker_thread *worker)
 | |
| {
 | |
| 	int alive;
 | |
| 
 | |
| 	/* The following is equivalent to
 | |
| 	 *
 | |
| 	 * while (threadpool_execute(worker->pool));
 | |
| 	 *
 | |
| 	 * However, reviewers have suggested in the past
 | |
| 	 * doing that can cause optimizers to (wrongly)
 | |
| 	 * optimize the code away.
 | |
| 	 */
 | |
| 	do {
 | |
| 		alive = threadpool_execute(worker->pool);
 | |
| 	} while (alive);
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Idle function for worker threads
 | |
|  *
 | |
|  * The worker waits here until it gets told by the threadpool
 | |
|  * to wake up.
 | |
|  *
 | |
|  * worker is locked before entering this function.
 | |
|  *
 | |
|  * \param worker The idle worker
 | |
|  * \retval 0 The thread is being woken up so that it can conclude.
 | |
|  * \retval non-zero The thread is being woken up to do more work.
 | |
|  */
 | |
| static int worker_idle(struct worker_thread *worker)
 | |
| {
 | |
| 	struct timeval start = ast_tvnow();
 | |
| 	struct timespec end = {
 | |
| 		.tv_sec = start.tv_sec + worker->options.idle_timeout,
 | |
| 		.tv_nsec = start.tv_usec * 1000,
 | |
| 	};
 | |
| 	while (!worker->wake_up) {
 | |
| 		if (worker->options.idle_timeout <= 0) {
 | |
| 			ast_cond_wait(&worker->cond, &worker->lock);
 | |
| 		} else if (ast_cond_timedwait(&worker->cond, &worker->lock, &end) == ETIMEDOUT) {
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if (!worker->wake_up) {
 | |
| 		ast_debug(1, "Worker thread idle timeout reached. Dying.\n");
 | |
| 		threadpool_idle_thread_dead(worker->pool, worker);
 | |
| 		worker->state = DEAD;
 | |
| 	}
 | |
| 	worker->wake_up = 0;
 | |
| 	return worker->state == ALIVE;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \brief Change a worker's state
 | |
|  *
 | |
|  * The threadpool calls into this function in order to let a worker know
 | |
|  * how it should proceed.
 | |
|  *
 | |
|  * \retval -1 failure (state transition not permitted)
 | |
|  * \retval 0 success
 | |
|  */
 | |
| static int worker_set_state(struct worker_thread *worker, enum worker_state state)
 | |
| {
 | |
| 	SCOPED_MUTEX(lock, &worker->lock);
 | |
| 
 | |
| 	switch (state) {
 | |
| 	case ALIVE:
 | |
| 		/* This can occur due to a race condition between being told to go active
 | |
| 		 * and an idle timeout happening.
 | |
| 		 */
 | |
| 		if (worker->state == DEAD) {
 | |
| 			return -1;
 | |
| 		}
 | |
| 		ast_assert(worker->state != ZOMBIE);
 | |
| 		break;
 | |
| 	case DEAD:
 | |
| 		break;
 | |
| 	case ZOMBIE:
 | |
| 		ast_assert(worker->state != DEAD);
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	worker->state = state;
 | |
| 	worker->wake_up = 1;
 | |
| 	ast_cond_signal(&worker->cond);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*! Serializer group shutdown control object. */
 | |
| struct ast_serializer_shutdown_group {
 | |
| 	/*! Shutdown thread waits on this conditional. */
 | |
| 	ast_cond_t cond;
 | |
| 	/*! Count of serializers needing to shutdown. */
 | |
| 	int count;
 | |
| };
 | |
| 
 | |
| static void serializer_shutdown_group_dtor(void *vdoomed)
 | |
| {
 | |
| 	struct ast_serializer_shutdown_group *doomed = vdoomed;
 | |
| 
 | |
| 	ast_cond_destroy(&doomed->cond);
 | |
| }
 | |
| 
 | |
| struct ast_serializer_shutdown_group *ast_serializer_shutdown_group_alloc(void)
 | |
| {
 | |
| 	struct ast_serializer_shutdown_group *shutdown_group;
 | |
| 
 | |
| 	shutdown_group = ao2_alloc(sizeof(*shutdown_group), serializer_shutdown_group_dtor);
 | |
| 	if (!shutdown_group) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	ast_cond_init(&shutdown_group->cond, NULL);
 | |
| 	return shutdown_group;
 | |
| }
 | |
| 
 | |
| int ast_serializer_shutdown_group_join(struct ast_serializer_shutdown_group *shutdown_group, int timeout)
 | |
| {
 | |
| 	int remaining;
 | |
| 	ast_mutex_t *lock;
 | |
| 
 | |
| 	if (!shutdown_group) {
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	lock = ao2_object_get_lockaddr(shutdown_group);
 | |
| 	ast_assert(lock != NULL);
 | |
| 
 | |
| 	ao2_lock(shutdown_group);
 | |
| 	if (timeout) {
 | |
| 		struct timeval start;
 | |
| 		struct timespec end;
 | |
| 
 | |
| 		start = ast_tvnow();
 | |
| 		end.tv_sec = start.tv_sec + timeout;
 | |
| 		end.tv_nsec = start.tv_usec * 1000;
 | |
| 		while (shutdown_group->count) {
 | |
| 			if (ast_cond_timedwait(&shutdown_group->cond, lock, &end)) {
 | |
| 				/* Error or timed out waiting for the count to reach zero. */
 | |
| 				break;
 | |
| 			}
 | |
| 		}
 | |
| 	} else {
 | |
| 		while (shutdown_group->count) {
 | |
| 			if (ast_cond_wait(&shutdown_group->cond, lock)) {
 | |
| 				/* Error */
 | |
| 				break;
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	remaining = shutdown_group->count;
 | |
| 	ao2_unlock(shutdown_group);
 | |
| 	return remaining;
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \internal
 | |
|  * \brief Increment the number of serializer members in the group.
 | |
|  * \since 13.5.0
 | |
|  *
 | |
|  * \param shutdown_group Group shutdown controller.
 | |
|  */
 | |
| static void serializer_shutdown_group_inc(struct ast_serializer_shutdown_group *shutdown_group)
 | |
| {
 | |
| 	ao2_lock(shutdown_group);
 | |
| 	++shutdown_group->count;
 | |
| 	ao2_unlock(shutdown_group);
 | |
| }
 | |
| 
 | |
| /*!
 | |
|  * \internal
 | |
|  * \brief Decrement the number of serializer members in the group.
 | |
|  * \since 13.5.0
 | |
|  *
 | |
|  * \param shutdown_group Group shutdown controller.
 | |
|  */
 | |
| static void serializer_shutdown_group_dec(struct ast_serializer_shutdown_group *shutdown_group)
 | |
| {
 | |
| 	ao2_lock(shutdown_group);
 | |
| 	--shutdown_group->count;
 | |
| 	if (!shutdown_group->count) {
 | |
| 		ast_cond_signal(&shutdown_group->cond);
 | |
| 	}
 | |
| 	ao2_unlock(shutdown_group);
 | |
| }
 | |
| 
 | |
| struct serializer {
 | |
| 	/*! Threadpool the serializer will use to process the jobs. */
 | |
| 	struct ast_threadpool *pool;
 | |
| 	/*! Which group will wait for this serializer to shutdown. */
 | |
| 	struct ast_serializer_shutdown_group *shutdown_group;
 | |
| };
 | |
| 
 | |
| static void serializer_dtor(void *obj)
 | |
| {
 | |
| 	struct serializer *ser = obj;
 | |
| 
 | |
| 	ao2_cleanup(ser->pool);
 | |
| 	ser->pool = NULL;
 | |
| 	ao2_cleanup(ser->shutdown_group);
 | |
| 	ser->shutdown_group = NULL;
 | |
| }
 | |
| 
 | |
| static struct serializer *serializer_create(struct ast_threadpool *pool,
 | |
| 	struct ast_serializer_shutdown_group *shutdown_group)
 | |
| {
 | |
| 	struct serializer *ser;
 | |
| 
 | |
| 	ser = ao2_alloc_options(sizeof(*ser), serializer_dtor, AO2_ALLOC_OPT_LOCK_NOLOCK);
 | |
| 	if (!ser) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	ao2_ref(pool, +1);
 | |
| 	ser->pool = pool;
 | |
| 	ser->shutdown_group = ao2_bump(shutdown_group);
 | |
| 	return ser;
 | |
| }
 | |
| 
 | |
| AST_THREADSTORAGE_RAW(current_serializer);
 | |
| 
 | |
| static int execute_tasks(void *data)
 | |
| {
 | |
| 	struct ast_taskprocessor *tps = data;
 | |
| 
 | |
| 	ast_threadstorage_set_ptr(¤t_serializer, tps);
 | |
| 	while (ast_taskprocessor_execute(tps)) {
 | |
| 		/* No-op */
 | |
| 	}
 | |
| 	ast_threadstorage_set_ptr(¤t_serializer, NULL);
 | |
| 
 | |
| 	ast_taskprocessor_unreference(tps);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void serializer_task_pushed(struct ast_taskprocessor_listener *listener, int was_empty)
 | |
| {
 | |
| 	if (was_empty) {
 | |
| 		struct serializer *ser = ast_taskprocessor_listener_get_user_data(listener);
 | |
| 		struct ast_taskprocessor *tps = ast_taskprocessor_listener_get_tps(listener);
 | |
| 
 | |
| 		if (ast_threadpool_push(ser->pool, execute_tasks, tps)) {
 | |
| 			ast_taskprocessor_unreference(tps);
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int serializer_start(struct ast_taskprocessor_listener *listener)
 | |
| {
 | |
| 	/* No-op */
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void serializer_shutdown(struct ast_taskprocessor_listener *listener)
 | |
| {
 | |
| 	struct serializer *ser = ast_taskprocessor_listener_get_user_data(listener);
 | |
| 
 | |
| 	if (ser->shutdown_group) {
 | |
| 		serializer_shutdown_group_dec(ser->shutdown_group);
 | |
| 	}
 | |
| 	ao2_cleanup(ser);
 | |
| }
 | |
| 
 | |
| static struct ast_taskprocessor_listener_callbacks serializer_tps_listener_callbacks = {
 | |
| 	.task_pushed = serializer_task_pushed,
 | |
| 	.start = serializer_start,
 | |
| 	.shutdown = serializer_shutdown,
 | |
| };
 | |
| 
 | |
| struct ast_taskprocessor *ast_threadpool_serializer_get_current(void)
 | |
| {
 | |
| 	return ast_threadstorage_get_ptr(¤t_serializer);
 | |
| }
 | |
| 
 | |
| struct ast_taskprocessor *ast_threadpool_serializer_group(const char *name,
 | |
| 	struct ast_threadpool *pool, struct ast_serializer_shutdown_group *shutdown_group)
 | |
| {
 | |
| 	struct serializer *ser;
 | |
| 	struct ast_taskprocessor_listener *listener;
 | |
| 	struct ast_taskprocessor *tps;
 | |
| 
 | |
| 	ser = serializer_create(pool, shutdown_group);
 | |
| 	if (!ser) {
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	listener = ast_taskprocessor_listener_alloc(&serializer_tps_listener_callbacks, ser);
 | |
| 	if (!listener) {
 | |
| 		ao2_ref(ser, -1);
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	tps = ast_taskprocessor_create_with_listener(name, listener);
 | |
| 	if (!tps) {
 | |
| 		/* ser ref transferred to listener but not cleaned without tps */
 | |
| 		ao2_ref(ser, -1);
 | |
| 	} else if (shutdown_group) {
 | |
| 		serializer_shutdown_group_inc(shutdown_group);
 | |
| 	}
 | |
| 
 | |
| 	ao2_ref(listener, -1);
 | |
| 	return tps;
 | |
| }
 | |
| 
 | |
| struct ast_taskprocessor *ast_threadpool_serializer(const char *name, struct ast_threadpool *pool)
 | |
| {
 | |
| 	return ast_threadpool_serializer_group(name, pool, NULL);
 | |
| }
 | |
| 
 | |
| long ast_threadpool_queue_size(struct ast_threadpool *pool)
 | |
| {
 | |
| 	return ast_taskprocessor_size(pool->tps);
 | |
| }
 |