mirror of
https://git.proxmox.com/git/mirror_zfs.git
synced 2024-12-28 11:59:34 +03:00
46364cb2f3
OpenZFS provides a library called tpool which implements thread pools for user space applications. Porting this library means the zpool utility no longer needs to borrow the kernel mutex and taskq interfaces from libzpool. This code was updated to use the tpool library which behaves in a very similar fashion. Porting libtpool was relatively straight forward and minimal modifications were needed. The core changes were: * Fully convert the library to use pthreads. * Updated signal handling. * lmalloc/lfree converted to calloc/free * Implemented portable pthread_attr_clone() function. Finally, update the build system such that libzpool.so is no longer linked in to zfs(8), zpool(8), etc. All that is required is libzfs to which the zcommon soures were added (which is the way it always should have been). Removing the libzpool dependency resulted in several build issues which needed to be resolved. * Moved zfeature support to module/zcommon/zfeature_common.c * Moved ratelimiting to to module/zfs/zfs_ratelimit.c * Moved get_system_hostid() to lib/libspl/gethostid.c * Removed use of cmn_err() in zcommon source * Removed dprintf_setup() call from zpool_main.c and zfs_main.c * Removed highbit() and lowbit() * Removed unnecessary library dependencies from Makefiles * Removed fletcher-4 kstat in user space * Added sha2 support explicitly to libzfs * Added highbit64() and lowbit64() to zpool_util.c Reviewed-by: Tony Hutter <hutter2@llnl.gov> Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov> Closes #6442
651 lines
16 KiB
C
651 lines
16 KiB
C
/*
|
|
* CDDL HEADER START
|
|
*
|
|
* The contents of this file are subject to the terms of the
|
|
* Common Development and Distribution License (the "License").
|
|
* You may not use this file except in compliance with the License.
|
|
*
|
|
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
|
|
* or http://www.opensolaris.org/os/licensing.
|
|
* See the License for the specific language governing permissions
|
|
* and limitations under the License.
|
|
*
|
|
* When distributing Covered Code, include this CDDL HEADER in each
|
|
* file and include the License file at usr/src/OPENSOLARIS.LICENSE.
|
|
* If applicable, add the following below this CDDL HEADER, with the
|
|
* fields enclosed by brackets "[]" replaced with your own identifying
|
|
* information: Portions Copyright [yyyy] [name of copyright owner]
|
|
*
|
|
* CDDL HEADER END
|
|
*/
|
|
|
|
/*
|
|
* Copyright 2008 Sun Microsystems, Inc. All rights reserved.
|
|
* Use is subject to license terms.
|
|
*/
|
|
|
|
#include <stdlib.h>
|
|
#include <signal.h>
|
|
#include <errno.h>
|
|
#include <assert.h>
|
|
#include "thread_pool_impl.h"
|
|
|
|
static pthread_mutex_t thread_pool_lock = PTHREAD_MUTEX_INITIALIZER;
|
|
static tpool_t *thread_pools = NULL;
|
|
|
|
static void
|
|
delete_pool(tpool_t *tpool)
|
|
{
|
|
tpool_job_t *job;
|
|
|
|
ASSERT(tpool->tp_current == 0 && tpool->tp_active == NULL);
|
|
|
|
/*
|
|
* Unlink the pool from the global list of all pools.
|
|
*/
|
|
(void) pthread_mutex_lock(&thread_pool_lock);
|
|
if (thread_pools == tpool)
|
|
thread_pools = tpool->tp_forw;
|
|
if (thread_pools == tpool)
|
|
thread_pools = NULL;
|
|
else {
|
|
tpool->tp_back->tp_forw = tpool->tp_forw;
|
|
tpool->tp_forw->tp_back = tpool->tp_back;
|
|
}
|
|
pthread_mutex_unlock(&thread_pool_lock);
|
|
|
|
/*
|
|
* There should be no pending jobs, but just in case...
|
|
*/
|
|
for (job = tpool->tp_head; job != NULL; job = tpool->tp_head) {
|
|
tpool->tp_head = job->tpj_next;
|
|
free(job);
|
|
}
|
|
(void) pthread_attr_destroy(&tpool->tp_attr);
|
|
free(tpool);
|
|
}
|
|
|
|
/*
|
|
* Worker thread is terminating.
|
|
*/
|
|
static void
|
|
worker_cleanup(void *arg)
|
|
{
|
|
tpool_t *tpool = (tpool_t *)arg;
|
|
|
|
if (--tpool->tp_current == 0 &&
|
|
(tpool->tp_flags & (TP_DESTROY | TP_ABANDON))) {
|
|
if (tpool->tp_flags & TP_ABANDON) {
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
delete_pool(tpool);
|
|
return;
|
|
}
|
|
if (tpool->tp_flags & TP_DESTROY)
|
|
(void) pthread_cond_broadcast(&tpool->tp_busycv);
|
|
}
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
}
|
|
|
|
static void
|
|
notify_waiters(tpool_t *tpool)
|
|
{
|
|
if (tpool->tp_head == NULL && tpool->tp_active == NULL) {
|
|
tpool->tp_flags &= ~TP_WAIT;
|
|
(void) pthread_cond_broadcast(&tpool->tp_waitcv);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Called by a worker thread on return from a tpool_dispatch()d job.
|
|
*/
|
|
static void
|
|
job_cleanup(void *arg)
|
|
{
|
|
tpool_t *tpool = (tpool_t *)arg;
|
|
|
|
pthread_t my_tid = pthread_self();
|
|
tpool_active_t *activep;
|
|
tpool_active_t **activepp;
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
/* CSTYLED */
|
|
for (activepp = &tpool->tp_active;; activepp = &activep->tpa_next) {
|
|
activep = *activepp;
|
|
if (activep->tpa_tid == my_tid) {
|
|
*activepp = activep->tpa_next;
|
|
break;
|
|
}
|
|
}
|
|
if (tpool->tp_flags & TP_WAIT)
|
|
notify_waiters(tpool);
|
|
}
|
|
|
|
static void *
|
|
tpool_worker(void *arg)
|
|
{
|
|
tpool_t *tpool = (tpool_t *)arg;
|
|
int elapsed;
|
|
tpool_job_t *job;
|
|
void (*func)(void *);
|
|
tpool_active_t active;
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
pthread_cleanup_push(worker_cleanup, tpool);
|
|
|
|
/*
|
|
* This is the worker's main loop.
|
|
* It will only be left if a timeout or an error has occured.
|
|
*/
|
|
active.tpa_tid = pthread_self();
|
|
for (;;) {
|
|
elapsed = 0;
|
|
tpool->tp_idle++;
|
|
if (tpool->tp_flags & TP_WAIT)
|
|
notify_waiters(tpool);
|
|
while ((tpool->tp_head == NULL ||
|
|
(tpool->tp_flags & TP_SUSPEND)) &&
|
|
!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON))) {
|
|
if (tpool->tp_current <= tpool->tp_minimum ||
|
|
tpool->tp_linger == 0) {
|
|
(void) pthread_cond_wait(&tpool->tp_workcv,
|
|
&tpool->tp_mutex);
|
|
} else {
|
|
struct timespec ts;
|
|
|
|
clock_gettime(CLOCK_REALTIME, &ts);
|
|
ts.tv_sec += tpool->tp_linger;
|
|
|
|
if (pthread_cond_timedwait(&tpool->tp_workcv,
|
|
&tpool->tp_mutex, &ts) != 0) {
|
|
elapsed = 1;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
tpool->tp_idle--;
|
|
if (tpool->tp_flags & TP_DESTROY)
|
|
break;
|
|
if (tpool->tp_flags & TP_ABANDON) {
|
|
/* can't abandon a suspended pool */
|
|
if (tpool->tp_flags & TP_SUSPEND) {
|
|
tpool->tp_flags &= ~TP_SUSPEND;
|
|
(void) pthread_cond_broadcast(
|
|
&tpool->tp_workcv);
|
|
}
|
|
if (tpool->tp_head == NULL)
|
|
break;
|
|
}
|
|
if ((job = tpool->tp_head) != NULL &&
|
|
!(tpool->tp_flags & TP_SUSPEND)) {
|
|
elapsed = 0;
|
|
func = job->tpj_func;
|
|
arg = job->tpj_arg;
|
|
tpool->tp_head = job->tpj_next;
|
|
if (job == tpool->tp_tail)
|
|
tpool->tp_tail = NULL;
|
|
tpool->tp_njobs--;
|
|
active.tpa_next = tpool->tp_active;
|
|
tpool->tp_active = &active;
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
pthread_cleanup_push(job_cleanup, tpool);
|
|
free(job);
|
|
|
|
sigset_t maskset;
|
|
(void) pthread_sigmask(SIG_SETMASK, NULL, &maskset);
|
|
|
|
/*
|
|
* Call the specified function.
|
|
*/
|
|
func(arg);
|
|
/*
|
|
* We don't know what this thread has been doing,
|
|
* so we reset its signal mask and cancellation
|
|
* state back to the values prior to calling func().
|
|
*/
|
|
(void) pthread_sigmask(SIG_SETMASK, &maskset, NULL);
|
|
(void) pthread_setcanceltype(PTHREAD_CANCEL_DEFERRED,
|
|
NULL);
|
|
(void) pthread_setcancelstate(PTHREAD_CANCEL_ENABLE,
|
|
NULL);
|
|
pthread_cleanup_pop(1);
|
|
}
|
|
if (elapsed && tpool->tp_current > tpool->tp_minimum) {
|
|
/*
|
|
* We timed out and there is no work to be done
|
|
* and the number of workers exceeds the minimum.
|
|
* Exit now to reduce the size of the pool.
|
|
*/
|
|
break;
|
|
}
|
|
}
|
|
pthread_cleanup_pop(1);
|
|
return (arg);
|
|
}
|
|
|
|
/*
|
|
* Create a worker thread, with default signals blocked.
|
|
*/
|
|
static int
|
|
create_worker(tpool_t *tpool)
|
|
{
|
|
pthread_t thread;
|
|
sigset_t oset;
|
|
int error;
|
|
|
|
(void) pthread_sigmask(SIG_SETMASK, NULL, &oset);
|
|
error = pthread_create(&thread, &tpool->tp_attr, tpool_worker, tpool);
|
|
(void) pthread_sigmask(SIG_SETMASK, &oset, NULL);
|
|
return (error);
|
|
}
|
|
|
|
|
|
/*
|
|
* pthread_attr_clone: make a copy of a pthread_attr_t. When old_attr
|
|
* is NULL initialize the cloned attr using default values.
|
|
*/
|
|
static int
|
|
pthread_attr_clone(pthread_attr_t *attr, const pthread_attr_t *old_attr)
|
|
{
|
|
int error;
|
|
|
|
error = pthread_attr_init(attr);
|
|
if (error || (old_attr == NULL))
|
|
return (error);
|
|
|
|
cpu_set_t cpuset;
|
|
size_t cpusetsize = sizeof (cpuset);
|
|
error = pthread_attr_getaffinity_np(old_attr, cpusetsize, &cpuset);
|
|
if (error == 0)
|
|
error = pthread_attr_setaffinity_np(attr, cpusetsize, &cpuset);
|
|
if (error)
|
|
goto error;
|
|
|
|
int detachstate;
|
|
error = pthread_attr_getdetachstate(old_attr, &detachstate);
|
|
if (error == 0)
|
|
error = pthread_attr_setdetachstate(attr, detachstate);
|
|
if (error)
|
|
goto error;
|
|
|
|
size_t guardsize;
|
|
error = pthread_attr_getguardsize(old_attr, &guardsize);
|
|
if (error == 0)
|
|
error = pthread_attr_setguardsize(attr, guardsize);
|
|
if (error)
|
|
goto error;
|
|
|
|
int inheritsched;
|
|
error = pthread_attr_getinheritsched(old_attr, &inheritsched);
|
|
if (error == 0)
|
|
error = pthread_attr_setinheritsched(attr, inheritsched);
|
|
if (error)
|
|
goto error;
|
|
|
|
struct sched_param param;
|
|
error = pthread_attr_getschedparam(old_attr, ¶m);
|
|
if (error == 0)
|
|
error = pthread_attr_setschedparam(attr, ¶m);
|
|
if (error)
|
|
goto error;
|
|
|
|
int policy;
|
|
error = pthread_attr_getschedpolicy(old_attr, &policy);
|
|
if (error == 0)
|
|
error = pthread_attr_setschedpolicy(attr, policy);
|
|
if (error)
|
|
goto error;
|
|
|
|
int scope;
|
|
error = pthread_attr_getscope(old_attr, &scope);
|
|
if (error == 0)
|
|
error = pthread_attr_setscope(attr, scope);
|
|
if (error)
|
|
goto error;
|
|
|
|
void *stackaddr;
|
|
size_t stacksize;
|
|
error = pthread_attr_getstack(old_attr, &stackaddr, &stacksize);
|
|
if (error == 0)
|
|
error = pthread_attr_setstack(attr, stackaddr, stacksize);
|
|
if (error)
|
|
goto error;
|
|
|
|
return (0);
|
|
error:
|
|
pthread_attr_destroy(attr);
|
|
return (error);
|
|
}
|
|
|
|
tpool_t *
|
|
tpool_create(uint_t min_threads, uint_t max_threads, uint_t linger,
|
|
pthread_attr_t *attr)
|
|
{
|
|
tpool_t *tpool;
|
|
void *stackaddr;
|
|
size_t stacksize;
|
|
size_t minstack;
|
|
int error;
|
|
|
|
if (min_threads > max_threads || max_threads < 1) {
|
|
errno = EINVAL;
|
|
return (NULL);
|
|
}
|
|
if (attr != NULL) {
|
|
if (pthread_attr_getstack(attr, &stackaddr, &stacksize) != 0) {
|
|
errno = EINVAL;
|
|
return (NULL);
|
|
}
|
|
/*
|
|
* Allow only one thread in the pool with a specified stack.
|
|
* Require threads to have at least the minimum stack size.
|
|
*/
|
|
minstack = PTHREAD_STACK_MIN;
|
|
if (stackaddr != NULL) {
|
|
if (stacksize < minstack || max_threads != 1) {
|
|
errno = EINVAL;
|
|
return (NULL);
|
|
}
|
|
} else if (stacksize != 0 && stacksize < minstack) {
|
|
errno = EINVAL;
|
|
return (NULL);
|
|
}
|
|
}
|
|
|
|
tpool = calloc(1, sizeof (*tpool));
|
|
if (tpool == NULL) {
|
|
errno = ENOMEM;
|
|
return (NULL);
|
|
}
|
|
(void) pthread_mutex_init(&tpool->tp_mutex, NULL);
|
|
(void) pthread_cond_init(&tpool->tp_busycv, NULL);
|
|
(void) pthread_cond_init(&tpool->tp_workcv, NULL);
|
|
(void) pthread_cond_init(&tpool->tp_waitcv, NULL);
|
|
tpool->tp_minimum = min_threads;
|
|
tpool->tp_maximum = max_threads;
|
|
tpool->tp_linger = linger;
|
|
|
|
/*
|
|
* We cannot just copy the attribute pointer.
|
|
* We need to initialize a new pthread_attr_t structure
|
|
* with the values from the user-supplied pthread_attr_t.
|
|
* If the attribute pointer is NULL, we need to initialize
|
|
* the new pthread_attr_t structure with default values.
|
|
*/
|
|
error = pthread_attr_clone(&tpool->tp_attr, attr);
|
|
if (error) {
|
|
free(tpool);
|
|
errno = error;
|
|
return (NULL);
|
|
}
|
|
|
|
/* make all pool threads be detached daemon threads */
|
|
(void) pthread_attr_setdetachstate(&tpool->tp_attr,
|
|
PTHREAD_CREATE_DETACHED);
|
|
|
|
/* insert into the global list of all thread pools */
|
|
pthread_mutex_lock(&thread_pool_lock);
|
|
if (thread_pools == NULL) {
|
|
tpool->tp_forw = tpool;
|
|
tpool->tp_back = tpool;
|
|
thread_pools = tpool;
|
|
} else {
|
|
thread_pools->tp_back->tp_forw = tpool;
|
|
tpool->tp_forw = thread_pools;
|
|
tpool->tp_back = thread_pools->tp_back;
|
|
thread_pools->tp_back = tpool;
|
|
}
|
|
pthread_mutex_unlock(&thread_pool_lock);
|
|
|
|
return (tpool);
|
|
}
|
|
|
|
/*
|
|
* Dispatch a work request to the thread pool.
|
|
* If there are idle workers, awaken one.
|
|
* Else, if the maximum number of workers has
|
|
* not been reached, spawn a new worker thread.
|
|
* Else just return with the job added to the queue.
|
|
*/
|
|
int
|
|
tpool_dispatch(tpool_t *tpool, void (*func)(void *), void *arg)
|
|
{
|
|
tpool_job_t *job;
|
|
|
|
ASSERT(!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON)));
|
|
|
|
if ((job = calloc(1, sizeof (*job))) == NULL)
|
|
return (-1);
|
|
job->tpj_next = NULL;
|
|
job->tpj_func = func;
|
|
job->tpj_arg = arg;
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
|
|
if (tpool->tp_head == NULL)
|
|
tpool->tp_head = job;
|
|
else
|
|
tpool->tp_tail->tpj_next = job;
|
|
tpool->tp_tail = job;
|
|
tpool->tp_njobs++;
|
|
|
|
if (!(tpool->tp_flags & TP_SUSPEND)) {
|
|
if (tpool->tp_idle > 0)
|
|
(void) pthread_cond_signal(&tpool->tp_workcv);
|
|
else if (tpool->tp_current < tpool->tp_maximum &&
|
|
create_worker(tpool) == 0)
|
|
tpool->tp_current++;
|
|
}
|
|
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
return (0);
|
|
}
|
|
|
|
static void
|
|
tpool_cleanup(void *arg)
|
|
{
|
|
tpool_t *tpool = (tpool_t *)arg;
|
|
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
}
|
|
|
|
/*
|
|
* Assumes: by the time tpool_destroy() is called no one will use this
|
|
* thread pool in any way and no one will try to dispatch entries to it.
|
|
* Calling tpool_destroy() from a job in the pool will cause deadlock.
|
|
*/
|
|
void
|
|
tpool_destroy(tpool_t *tpool)
|
|
{
|
|
tpool_active_t *activep;
|
|
|
|
ASSERT(!tpool_member(tpool));
|
|
ASSERT(!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON)));
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
pthread_cleanup_push(tpool_cleanup, tpool);
|
|
|
|
/* mark the pool as being destroyed; wakeup idle workers */
|
|
tpool->tp_flags |= TP_DESTROY;
|
|
tpool->tp_flags &= ~TP_SUSPEND;
|
|
(void) pthread_cond_broadcast(&tpool->tp_workcv);
|
|
|
|
/* cancel all active workers */
|
|
for (activep = tpool->tp_active; activep; activep = activep->tpa_next)
|
|
(void) pthread_cancel(activep->tpa_tid);
|
|
|
|
/* wait for all active workers to finish */
|
|
while (tpool->tp_active != NULL) {
|
|
tpool->tp_flags |= TP_WAIT;
|
|
(void) pthread_cond_wait(&tpool->tp_waitcv, &tpool->tp_mutex);
|
|
}
|
|
|
|
/* the last worker to terminate will wake us up */
|
|
while (tpool->tp_current != 0)
|
|
(void) pthread_cond_wait(&tpool->tp_busycv, &tpool->tp_mutex);
|
|
|
|
pthread_cleanup_pop(1); /* pthread_mutex_unlock(&tpool->tp_mutex); */
|
|
delete_pool(tpool);
|
|
}
|
|
|
|
/*
|
|
* Like tpool_destroy(), but don't cancel workers or wait for them to finish.
|
|
* The last worker to terminate will delete the pool.
|
|
*/
|
|
void
|
|
tpool_abandon(tpool_t *tpool)
|
|
{
|
|
ASSERT(!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON)));
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
if (tpool->tp_current == 0) {
|
|
/* no workers, just delete the pool */
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
delete_pool(tpool);
|
|
} else {
|
|
/* wake up all workers, last one will delete the pool */
|
|
tpool->tp_flags |= TP_ABANDON;
|
|
tpool->tp_flags &= ~TP_SUSPEND;
|
|
(void) pthread_cond_broadcast(&tpool->tp_workcv);
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Wait for all jobs to complete.
|
|
* Calling tpool_wait() from a job in the pool will cause deadlock.
|
|
*/
|
|
void
|
|
tpool_wait(tpool_t *tpool)
|
|
{
|
|
ASSERT(!tpool_member(tpool));
|
|
ASSERT(!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON)));
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
pthread_cleanup_push(tpool_cleanup, tpool);
|
|
while (tpool->tp_head != NULL || tpool->tp_active != NULL) {
|
|
tpool->tp_flags |= TP_WAIT;
|
|
(void) pthread_cond_wait(&tpool->tp_waitcv, &tpool->tp_mutex);
|
|
ASSERT(!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON)));
|
|
}
|
|
pthread_cleanup_pop(1); /* pthread_mutex_unlock(&tpool->tp_mutex); */
|
|
}
|
|
|
|
void
|
|
tpool_suspend(tpool_t *tpool)
|
|
{
|
|
ASSERT(!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON)));
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
tpool->tp_flags |= TP_SUSPEND;
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
}
|
|
|
|
int
|
|
tpool_suspended(tpool_t *tpool)
|
|
{
|
|
int suspended;
|
|
|
|
ASSERT(!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON)));
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
suspended = (tpool->tp_flags & TP_SUSPEND) != 0;
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
|
|
return (suspended);
|
|
}
|
|
|
|
void
|
|
tpool_resume(tpool_t *tpool)
|
|
{
|
|
int excess;
|
|
|
|
ASSERT(!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON)));
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
if (!(tpool->tp_flags & TP_SUSPEND)) {
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
return;
|
|
}
|
|
tpool->tp_flags &= ~TP_SUSPEND;
|
|
(void) pthread_cond_broadcast(&tpool->tp_workcv);
|
|
excess = tpool->tp_njobs - tpool->tp_idle;
|
|
while (excess-- > 0 && tpool->tp_current < tpool->tp_maximum) {
|
|
if (create_worker(tpool) != 0)
|
|
break; /* pthread_create() failed */
|
|
tpool->tp_current++;
|
|
}
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
}
|
|
|
|
int
|
|
tpool_member(tpool_t *tpool)
|
|
{
|
|
pthread_t my_tid = pthread_self();
|
|
tpool_active_t *activep;
|
|
|
|
ASSERT(!(tpool->tp_flags & (TP_DESTROY | TP_ABANDON)));
|
|
|
|
pthread_mutex_lock(&tpool->tp_mutex);
|
|
for (activep = tpool->tp_active; activep; activep = activep->tpa_next) {
|
|
if (activep->tpa_tid == my_tid) {
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
return (1);
|
|
}
|
|
}
|
|
pthread_mutex_unlock(&tpool->tp_mutex);
|
|
return (0);
|
|
}
|
|
|
|
void
|
|
postfork1_child_tpool(void)
|
|
{
|
|
pthread_t my_tid = pthread_self();
|
|
tpool_t *tpool;
|
|
tpool_job_t *job;
|
|
|
|
/*
|
|
* All of the thread pool workers are gone, except possibly
|
|
* for the current thread, if it is a thread pool worker thread.
|
|
* Retain the thread pools, but make them all empty. Whatever
|
|
* jobs were queued or running belong to the parent process.
|
|
*/
|
|
top:
|
|
if ((tpool = thread_pools) == NULL)
|
|
return;
|
|
|
|
do {
|
|
tpool_active_t *activep;
|
|
|
|
(void) pthread_mutex_init(&tpool->tp_mutex, NULL);
|
|
(void) pthread_cond_init(&tpool->tp_busycv, NULL);
|
|
(void) pthread_cond_init(&tpool->tp_workcv, NULL);
|
|
(void) pthread_cond_init(&tpool->tp_waitcv, NULL);
|
|
for (job = tpool->tp_head; job; job = tpool->tp_head) {
|
|
tpool->tp_head = job->tpj_next;
|
|
free(job);
|
|
}
|
|
tpool->tp_tail = NULL;
|
|
tpool->tp_njobs = 0;
|
|
for (activep = tpool->tp_active; activep;
|
|
activep = activep->tpa_next) {
|
|
if (activep->tpa_tid == my_tid) {
|
|
activep->tpa_next = NULL;
|
|
break;
|
|
}
|
|
}
|
|
tpool->tp_idle = 0;
|
|
tpool->tp_current = 0;
|
|
if ((tpool->tp_active = activep) != NULL)
|
|
tpool->tp_current = 1;
|
|
tpool->tp_flags &= ~TP_WAIT;
|
|
if (tpool->tp_flags & (TP_DESTROY | TP_ABANDON)) {
|
|
tpool->tp_flags &= ~TP_DESTROY;
|
|
tpool->tp_flags |= TP_ABANDON;
|
|
if (tpool->tp_current == 0) {
|
|
delete_pool(tpool);
|
|
goto top; /* start over */
|
|
}
|
|
}
|
|
} while ((tpool = tpool->tp_forw) != thread_pools);
|
|
}
|