asterisk/main/sched.c

785 lines
19 KiB
C
Raw Normal View History

/*
* Asterisk -- An open source telephony toolkit.
*
* Copyright (C) 1999 - 2010, Digium, Inc.
*
* Mark Spencer <markster@digium.com>
* Russell Bryant <russell@digium.com>
*
* See http://www.asterisk.org for more information about
* the Asterisk project. Please do not directly contact
* any of the maintainers of this project for assistance;
* the project provides a web site, mailing lists and IRC
* channels for your use.
*
* This program is free software, distributed under the terms of
* the GNU General Public License Version 2. See the LICENSE file
* at the top of the source tree.
*/
/*! \file
*
* \brief Scheduler Routines (from cheops-NG)
*
* \author Mark Spencer <markster@digium.com>
*/
/*** MODULEINFO
<support_level>core</support_level>
***/
#include "asterisk.h"
git migration: Refactor the ASTERISK_FILE_VERSION macro Git does not support the ability to replace a token with a version string during check-in. While it does have support for replacing a token on clone, this is somewhat sub-optimal: the token is replaced with the object hash, which is not particularly easy for human consumption. What's more, in practice, the source file version was often not terribly useful. Generally, when triaging bugs, the overall version of Asterisk is far more useful than an individual SVN version of a file. As a result, this patch removes Asterisk's support for showing source file versions. Specifically, it does the following: * Rename ASTERISK_FILE_VERSION macro to ASTERISK_REGISTER_FILE, and remove passing the version in with the macro. Other facilities than 'core show file version' make use of the file names, such as setting a debug level only on a specific file. As such, the act of registering source files with the Asterisk core still has use. The macro rename now reflects the new macro purpose. * main/asterisk: - Refactor the file_version structure to reflect that it no longer tracks a version field. - Remove the "core show file version" CLI command. Without the file version, it is no longer useful. - Remove the ast_file_version_find function. The file version is no longer tracked. - Rename ast_register_file_version/ast_unregister_file_version to ast_register_file/ast_unregister_file, respectively. * main/manager: Remove value from the Version key of the ModuleCheck Action. The actual key itself has not been removed, as doing so would absolutely constitute a backwards incompatible change. However, since the file version is no longer tracked, there is no need to attempt to include it in the Version key. * UPGRADE: Add notes for: - Modification to the ModuleCheck AMI Action - Removal of the "core show file version" CLI command Change-Id: I6cf0ff280e1668bf4957dc21f32a5ff43444a40e
2015-04-12 02:38:22 +00:00
ASTERISK_REGISTER_FILE()
#ifdef DEBUG_SCHEDULER
#define DEBUG(a) do { \
if (option_debug) \
DEBUG_M(a) \
} while (0)
#else
#define DEBUG(a)
#endif
#include <sys/time.h>
#include "asterisk/sched.h"
#include "asterisk/channel.h"
#include "asterisk/lock.h"
#include "asterisk/utils.h"
#include "asterisk/heap.h"
#include "asterisk/threadstorage.h"
/*!
* \brief Max num of schedule structs
*
* \note The max number of schedule structs to keep around
* for use. Undefine to disable schedule structure
* caching. (Only disable this on very low memory
* machines)
*/
#define SCHED_MAX_CACHE 128
AST_THREADSTORAGE(last_del_id);
/*!
* \brief Scheduler ID holder
*
* These form a queue on a scheduler context. When a new
* scheduled item is created, a sched_id is popped off the
* queue and its id is assigned to the new scheduled item.
* When the scheduled task is complete, the sched_id on that
* task is then pushed to the back of the queue to be re-used
* on some future scheduled item.
*/
struct sched_id {
/*! Immutable ID number that is copied onto the scheduled task */
int id;
AST_LIST_ENTRY(sched_id) list;
};
struct sched {
AST_LIST_ENTRY(sched) list;
/*! The ID that has been popped off the scheduler context's queue */
struct sched_id *sched_id;
struct timeval when; /*!< Absolute time event should take place */
int resched; /*!< When to reschedule */
int variable; /*!< Use return value from callback to reschedule */
const void *data; /*!< Data */
ast_sched_cb callback; /*!< Callback */
ssize_t __heap_index;
/*!
* Used to synchronize between thread running a task and thread
* attempting to delete a task
*/
ast_cond_t cond;
/*! Indication that a running task was deleted. */
unsigned int deleted:1;
};
struct sched_thread {
pthread_t thread;
ast_cond_t cond;
unsigned int stop:1;
};
struct ast_sched_context {
ast_mutex_t lock;
unsigned int eventcnt; /*!< Number of events processed */
unsigned int highwater; /*!< highest count so far */
struct ast_heap *sched_heap;
struct sched_thread *sched_thread;
/*! The scheduled task that is currently executing */
struct sched *currently_executing;
#ifdef SCHED_MAX_CACHE
AST_LIST_HEAD_NOLOCK(, sched) schedc; /*!< Cache of unused schedule structures and how many */
unsigned int schedccnt;
#endif
/*! Queue of scheduler task IDs to assign */
AST_LIST_HEAD_NOLOCK(, sched_id) id_queue;
/*! The number of IDs in the id_queue */
int id_queue_size;
};
static void *sched_run(void *data)
{
struct ast_sched_context *con = data;
while (!con->sched_thread->stop) {
int ms;
struct timespec ts = {
.tv_sec = 0,
};
ast_mutex_lock(&con->lock);
if (con->sched_thread->stop) {
ast_mutex_unlock(&con->lock);
return NULL;
}
ms = ast_sched_wait(con);
if (ms == -1) {
ast_cond_wait(&con->sched_thread->cond, &con->lock);
} else {
struct timeval tv;
tv = ast_tvadd(ast_tvnow(), ast_samp2tv(ms, 1000));
ts.tv_sec = tv.tv_sec;
ts.tv_nsec = tv.tv_usec * 1000;
ast_cond_timedwait(&con->sched_thread->cond, &con->lock, &ts);
}
ast_mutex_unlock(&con->lock);
if (con->sched_thread->stop) {
return NULL;
}
ast_sched_runq(con);
}
return NULL;
}
static void sched_thread_destroy(struct ast_sched_context *con)
{
if (!con->sched_thread) {
return;
}
if (con->sched_thread->thread != AST_PTHREADT_NULL) {
ast_mutex_lock(&con->lock);
con->sched_thread->stop = 1;
ast_cond_signal(&con->sched_thread->cond);
ast_mutex_unlock(&con->lock);
pthread_join(con->sched_thread->thread, NULL);
con->sched_thread->thread = AST_PTHREADT_NULL;
}
ast_cond_destroy(&con->sched_thread->cond);
ast_free(con->sched_thread);
con->sched_thread = NULL;
}
int ast_sched_start_thread(struct ast_sched_context *con)
{
struct sched_thread *st;
if (con->sched_thread) {
ast_log(LOG_ERROR, "Thread already started on this scheduler context\n");
return -1;
}
if (!(st = ast_calloc(1, sizeof(*st)))) {
return -1;
}
ast_cond_init(&st->cond, NULL);
st->thread = AST_PTHREADT_NULL;
con->sched_thread = st;
if (ast_pthread_create_background(&st->thread, NULL, sched_run, con)) {
ast_log(LOG_ERROR, "Failed to create scheduler thread\n");
sched_thread_destroy(con);
return -1;
}
return 0;
}
static int sched_time_cmp(void *a, void *b)
{
return ast_tvcmp(((struct sched *) b)->when, ((struct sched *) a)->when);
}
struct ast_sched_context *ast_sched_context_create(void)
{
struct ast_sched_context *tmp;
if (!(tmp = ast_calloc(1, sizeof(*tmp)))) {
return NULL;
}
ast_mutex_init(&tmp->lock);
tmp->eventcnt = 1;
AST_LIST_HEAD_INIT_NOLOCK(&tmp->id_queue);
if (!(tmp->sched_heap = ast_heap_create(8, sched_time_cmp,
offsetof(struct sched, __heap_index)))) {
ast_sched_context_destroy(tmp);
return NULL;
}
return tmp;
}
static void sched_free(struct sched *task)
{
/* task->sched_id will be NULL most of the time, but when the
* scheduler context shuts down, it will free all scheduled
* tasks, and in that case, the task->sched_id will be non-NULL
*/
ast_free(task->sched_id);
ast_cond_destroy(&task->cond);
ast_free(task);
}
void ast_sched_context_destroy(struct ast_sched_context *con)
{
struct sched *s;
struct sched_id *sid;
sched_thread_destroy(con);
con->sched_thread = NULL;
ast_mutex_lock(&con->lock);
#ifdef SCHED_MAX_CACHE
while ((s = AST_LIST_REMOVE_HEAD(&con->schedc, list))) {
sched_free(s);
}
#endif
if (con->sched_heap) {
while ((s = ast_heap_pop(con->sched_heap))) {
sched_free(s);
}
ast_heap_destroy(con->sched_heap);
con->sched_heap = NULL;
}
while ((sid = AST_LIST_REMOVE_HEAD(&con->id_queue, list))) {
ast_free(sid);
}
ast_mutex_unlock(&con->lock);
ast_mutex_destroy(&con->lock);
ast_free(con);
}
#define ID_QUEUE_INCREMENT 16
/*!
* \brief Add new scheduler IDs to the queue.
*
* \retval The number of IDs added to the queue
*/
static int add_ids(struct ast_sched_context *con)
{
int new_size;
int original_size;
int i;
original_size = con->id_queue_size;
/* So we don't go overboard with the mallocs here, we'll just up
* the size of the list by a fixed amount each time instead of
* multiplying the size by any particular factor
*/
new_size = original_size + ID_QUEUE_INCREMENT;
if (new_size < 0) {
/* Overflow. Cap it at INT_MAX. */
new_size = INT_MAX;
}
for (i = original_size; i < new_size; ++i) {
struct sched_id *new_id;
new_id = ast_calloc(1, sizeof(*new_id));
if (!new_id) {
break;
}
new_id->id = i;
AST_LIST_INSERT_TAIL(&con->id_queue, new_id, list);
++con->id_queue_size;
}
return con->id_queue_size - original_size;
}
static int set_sched_id(struct ast_sched_context *con, struct sched *new_sched)
{
if (AST_LIST_EMPTY(&con->id_queue) && (add_ids(con) == 0)) {
return -1;
}
new_sched->sched_id = AST_LIST_REMOVE_HEAD(&con->id_queue, list);
return 0;
}
static void sched_release(struct ast_sched_context *con, struct sched *tmp)
{
if (tmp->sched_id) {
AST_LIST_INSERT_TAIL(&con->id_queue, tmp->sched_id, list);
tmp->sched_id = NULL;
}
/*
* Add to the cache, or just free() if we
* already have too many cache entries
*/
#ifdef SCHED_MAX_CACHE
if (con->schedccnt < SCHED_MAX_CACHE) {
AST_LIST_INSERT_HEAD(&con->schedc, tmp, list);
con->schedccnt++;
} else
#endif
sched_free(tmp);
}
static struct sched *sched_alloc(struct ast_sched_context *con)
{
struct sched *tmp;
/*
* We keep a small cache of schedule entries
* to minimize the number of necessary malloc()'s
*/
#ifdef SCHED_MAX_CACHE
if ((tmp = AST_LIST_REMOVE_HEAD(&con->schedc, list))) {
con->schedccnt--;
} else
#endif
{
tmp = ast_calloc(1, sizeof(*tmp));
if (!tmp) {
return NULL;
}
ast_cond_init(&tmp->cond, NULL);
}
if (set_sched_id(con, tmp)) {
sched_release(con, tmp);
return NULL;
}
return tmp;
}
void ast_sched_clean_by_callback(struct ast_sched_context *con, ast_sched_cb match, ast_sched_cb cleanup_cb)
{
int i = 1;
struct sched *current;
ast_mutex_lock(&con->lock);
while ((current = ast_heap_peek(con->sched_heap, i))) {
if (current->callback != match) {
i++;
continue;
}
ast_heap_remove(con->sched_heap, current);
cleanup_cb(current->data);
sched_release(con, current);
}
ast_mutex_unlock(&con->lock);
}
/*! \brief
* Return the number of milliseconds
* until the next scheduled event
*/
int ast_sched_wait(struct ast_sched_context *con)
{
int ms;
struct sched *s;
DEBUG(ast_debug(1, "ast_sched_wait()\n"));
ast_mutex_lock(&con->lock);
if ((s = ast_heap_peek(con->sched_heap, 1))) {
ms = ast_tvdiff_ms(s->when, ast_tvnow());
if (ms < 0) {
ms = 0;
}
} else {
ms = -1;
}
ast_mutex_unlock(&con->lock);
return ms;
}
/*! \brief
* Take a sched structure and put it in the
* queue, such that the soonest event is
* first in the list.
*/
static void schedule(struct ast_sched_context *con, struct sched *s)
{
ast_heap_push(con->sched_heap, s);
if (ast_heap_size(con->sched_heap) > con->highwater) {
con->highwater = ast_heap_size(con->sched_heap);
}
}
/*! \brief
* given the last event *tv and the offset in milliseconds 'when',
* computes the next value,
*/
static int sched_settime(struct timeval *t, int when)
{
struct timeval now = ast_tvnow();
if (when < 0) {
/*
* A negative when value is likely a bug as it
* represents a VERY large timeout time.
*/
ast_log(LOG_WARNING,
"Bug likely: Negative time interval %d (interpreted as %u ms) requested!\n",
when, (unsigned int) when);
ast_assert(0);
}
/*ast_debug(1, "TV -> %lu,%lu\n", tv->tv_sec, tv->tv_usec);*/
if (ast_tvzero(*t)) /* not supplied, default to now */
*t = now;
*t = ast_tvadd(*t, ast_samp2tv(when, 1000));
if (ast_tvcmp(*t, now) < 0) {
*t = now;
}
return 0;
}
int ast_sched_replace_variable(int old_id, struct ast_sched_context *con, int when, ast_sched_cb callback, const void *data, int variable)
{
/* 0 means the schedule item is new; do not delete */
if (old_id > 0) {
AST_SCHED_DEL(con, old_id);
}
return ast_sched_add_variable(con, when, callback, data, variable);
}
/*! \brief
* Schedule callback(data) to happen when ms into the future
*/
int ast_sched_add_variable(struct ast_sched_context *con, int when, ast_sched_cb callback, const void *data, int variable)
{
struct sched *tmp;
int res = -1;
DEBUG(ast_debug(1, "ast_sched_add()\n"));
ast_mutex_lock(&con->lock);
if ((tmp = sched_alloc(con))) {
con->eventcnt++;
tmp->callback = callback;
tmp->data = data;
tmp->resched = when;
tmp->variable = variable;
tmp->when = ast_tv(0, 0);
tmp->deleted = 0;
if (sched_settime(&tmp->when, when)) {
sched_release(con, tmp);
} else {
schedule(con, tmp);
res = tmp->sched_id->id;
}
}
#ifdef DUMP_SCHEDULER
/* Dump contents of the context while we have the lock so nothing gets screwed up by accident. */
if (option_debug)
ast_sched_dump(con);
#endif
if (con->sched_thread) {
ast_cond_signal(&con->sched_thread->cond);
}
ast_mutex_unlock(&con->lock);
return res;
}
int ast_sched_replace(int old_id, struct ast_sched_context *con, int when, ast_sched_cb callback, const void *data)
{
if (old_id > -1) {
AST_SCHED_DEL(con, old_id);
}
return ast_sched_add(con, when, callback, data);
}
int ast_sched_add(struct ast_sched_context *con, int when, ast_sched_cb callback, const void *data)
{
return ast_sched_add_variable(con, when, callback, data, 0);
}
static struct sched *sched_find(struct ast_sched_context *con, int id)
{
int x;
size_t heap_size;
heap_size = ast_heap_size(con->sched_heap);
for (x = 1; x <= heap_size; x++) {
struct sched *cur = ast_heap_peek(con->sched_heap, x);
if (cur->sched_id->id == id) {
return cur;
}
}
return NULL;
}
const void *ast_sched_find_data(struct ast_sched_context *con, int id)
{
struct sched *s;
const void *data = NULL;
ast_mutex_lock(&con->lock);
s = sched_find(con, id);
if (s) {
data = s->data;
}
ast_mutex_unlock(&con->lock);
return data;
}
/*! \brief
* Delete the schedule entry with number
* "id". It's nearly impossible that there
* would be two or more in the list with that
* id.
*/
#ifndef AST_DEVMODE
int ast_sched_del(struct ast_sched_context *con, int id)
#else
int _ast_sched_del(struct ast_sched_context *con, int id, const char *file, int line, const char *function)
#endif
{
struct sched *s = NULL;
int *last_id = ast_threadstorage_get(&last_del_id, sizeof(int));
DEBUG(ast_debug(1, "ast_sched_del(%d)\n", id));
if (id < 0) {
return 0;
}
ast_mutex_lock(&con->lock);
s = sched_find(con, id);
if (s) {
if (!ast_heap_remove(con->sched_heap, s)) {
ast_log(LOG_WARNING,"sched entry %d not in the sched heap?\n", s->sched_id->id);
}
sched_release(con, s);
} else if (con->currently_executing && (id == con->currently_executing->sched_id->id)) {
s = con->currently_executing;
s->deleted = 1;
/* Wait for executing task to complete so that caller of ast_sched_del() does not
* free memory out from under the task.
*/
while (con->currently_executing && (id == con->currently_executing->sched_id->id)) {
ast_cond_wait(&s->cond, &con->lock);
}
/* Do not sched_release() here because ast_sched_runq() will do it */
}
#ifdef DUMP_SCHEDULER
/* Dump contents of the context while we have the lock so nothing gets screwed up by accident. */
if (option_debug)
ast_sched_dump(con);
#endif
if (con->sched_thread) {
ast_cond_signal(&con->sched_thread->cond);
}
ast_mutex_unlock(&con->lock);
if (!s && *last_id != id) {
ast_debug(1, "Attempted to delete nonexistent schedule entry %d!\n", id);
/* Removing nonexistent schedule entry shouldn't trigger assert (it was enabled in DEV_MODE);
* because in many places entries is deleted without having valid id. */
*last_id = id;
return -1;
} else if (!s) {
return -1;
}
return 0;
}
void ast_sched_report(struct ast_sched_context *con, struct ast_str **buf, struct ast_cb_names *cbnames)
{
int i, x;
struct sched *cur;
int countlist[cbnames->numassocs + 1];
size_t heap_size;
memset(countlist, 0, sizeof(countlist));
ast_str_set(buf, 0, " Highwater = %u\n schedcnt = %zu\n", con->highwater, ast_heap_size(con->sched_heap));
ast_mutex_lock(&con->lock);
heap_size = ast_heap_size(con->sched_heap);
for (x = 1; x <= heap_size; x++) {
cur = ast_heap_peek(con->sched_heap, x);
/* match the callback to the cblist */
for (i = 0; i < cbnames->numassocs; i++) {
if (cur->callback == cbnames->cblist[i]) {
break;
}
}
if (i < cbnames->numassocs) {
countlist[i]++;
} else {
countlist[cbnames->numassocs]++;
}
}
ast_mutex_unlock(&con->lock);
for (i = 0; i < cbnames->numassocs; i++) {
ast_str_append(buf, 0, " %s : %d\n", cbnames->list[i], countlist[i]);
}
ast_str_append(buf, 0, " <unknown> : %d\n", countlist[cbnames->numassocs]);
}
/*! \brief Dump the contents of the scheduler to LOG_DEBUG */
void ast_sched_dump(struct ast_sched_context *con)
{
struct sched *q;
struct timeval when = ast_tvnow();
int x;
size_t heap_size;
#ifdef SCHED_MAX_CACHE
ast_debug(1, "Asterisk Schedule Dump (%zu in Q, %u Total, %u Cache, %u high-water)\n", ast_heap_size(con->sched_heap), con->eventcnt - 1, con->schedccnt, con->highwater);
#else
ast_debug(1, "Asterisk Schedule Dump (%zu in Q, %u Total, %u high-water)\n", ast_heap_size(con->sched_heap), con->eventcnt - 1, con->highwater);
#endif
ast_debug(1, "=============================================================\n");
ast_debug(1, "|ID Callback Data Time (sec:ms) |\n");
ast_debug(1, "+-----+-----------------+-----------------+-----------------+\n");
ast_mutex_lock(&con->lock);
heap_size = ast_heap_size(con->sched_heap);
for (x = 1; x <= heap_size; x++) {
struct timeval delta;
q = ast_heap_peek(con->sched_heap, x);
delta = ast_tvsub(q->when, when);
ast_debug(1, "|%.4d | %-15p | %-15p | %.6ld : %.6ld |\n",
q->sched_id->id,
q->callback,
q->data,
(long)delta.tv_sec,
(long int)delta.tv_usec);
}
ast_mutex_unlock(&con->lock);
ast_debug(1, "=============================================================\n");
}
/*! \brief
* Launch all events which need to be run at this time.
*/
int ast_sched_runq(struct ast_sched_context *con)
{
struct sched *current;
struct timeval when;
int numevents;
int res;
DEBUG(ast_debug(1, "ast_sched_runq()\n"));
ast_mutex_lock(&con->lock);
when = ast_tvadd(ast_tvnow(), ast_tv(0, 1000));
for (numevents = 0; (current = ast_heap_peek(con->sched_heap, 1)); numevents++) {
/* schedule all events which are going to expire within 1ms.
* We only care about millisecond accuracy anyway, so this will
* help us get more than one event at one time if they are very
* close together.
*/
if (ast_tvcmp(current->when, when) != -1) {
break;
}
current = ast_heap_pop(con->sched_heap);
/*
* At this point, the schedule queue is still intact. We
* have removed the first event and the rest is still there,
* so it's permissible for the callback to add new events, but
* trying to delete itself won't work because it isn't in
* the schedule queue. If that's what it wants to do, it
* should return 0.
*/
con->currently_executing = current;
ast_mutex_unlock(&con->lock);
res = current->callback(current->data);
ast_mutex_lock(&con->lock);
con->currently_executing = NULL;
ast_cond_signal(&current->cond);
if (res && !current->deleted) {
/*
* If they return non-zero, we should schedule them to be
* run again.
*/
if (sched_settime(&current->when, current->variable? res : current->resched)) {
sched_release(con, current);
} else {
schedule(con, current);
}
} else {
/* No longer needed, so release it */
sched_release(con, current);
}
}
ast_mutex_unlock(&con->lock);
return numevents;
}
long ast_sched_when(struct ast_sched_context *con,int id)
{
struct sched *s;
long secs = -1;
DEBUG(ast_debug(1, "ast_sched_when()\n"));
ast_mutex_lock(&con->lock);
s = sched_find(con, id);
if (s) {
struct timeval now = ast_tvnow();
secs = s->when.tv_sec - now.tv_sec;
}
ast_mutex_unlock(&con->lock);
return secs;
}