#include "pthread_impl.h" void __pthread_testcancel(void); int __pthread_mutex_lock(pthread_mutex_t *); int __pthread_mutex_unlock(pthread_mutex_t *); int __pthread_setcancelstate(int, int *); /* * struct waiter * * Waiter objects have automatic storage on the waiting thread, and * are used in building a linked list representing waiters currently * waiting on the condition variable or a group of waiters woken * together by a broadcast or signal; in the case of signal, this is a * degenerate list of one member. * * Waiter lists attached to the condition variable itself are * protected by the lock on the cv. Detached waiter lists are never * modified again, but can only be traversed in reverse order, and are * protected by the "barrier" locks in each node, which are unlocked * in turn to control wake order. * * Since process-shared cond var semantics do not necessarily allow * one thread to see another's automatic storage (they may be in * different processes), the waiter list is not used for the * process-shared case, but the structure is still used to store data * needed by the cancellation cleanup handler. */ struct waiter { struct waiter *prev, *next; volatile int state, barrier; volatile int *notify; }; /* Self-synchronized-destruction-safe lock functions */ static inline void lock(volatile int *l) { if (a_cas(l, 0, 1)) { a_cas(l, 1, 2); do __wait(l, 0, 2, 1); while (a_cas(l, 0, 2)); } } static inline void unlock(volatile int *l) { if (a_swap(l, 0)==2) __wake(l, 1, 1); } static inline void unlock_requeue(volatile int *l, volatile int *r, int w) { a_store(l, 0); if (w) __wake(l, 1, 1); else __syscall(SYS_futex, l, FUTEX_REQUEUE|128, 0, 1, r) != -ENOSYS || __syscall(SYS_futex, l, FUTEX_REQUEUE, 0, 1, r); } enum { WAITING, SIGNALED, LEAVING, }; int __pthread_cond_timedwait(pthread_cond_t *restrict c, pthread_mutex_t *restrict m, const struct timespec *restrict ts) { struct waiter node = { 0 }; int e, seq, clock = c->_c_clock, cs, shared=0, oldstate, tmp; volatile int *fut; if ((m->_m_type&15) && (m->_m_lock&INT_MAX) != __pthread_self()->tid) return EPERM; if (ts && ts->tv_nsec >= 1000000000UL) return EINVAL; __pthread_testcancel(); if (c->_c_shared) { shared = 1; fut = &c->_c_seq; seq = c->_c_seq; a_inc(&c->_c_waiters); } else { lock(&c->_c_lock); seq = node.barrier = 2; fut = &node.barrier; node.state = WAITING; node.next = c->_c_head; c->_c_head = &node; if (!c->_c_tail) c->_c_tail = &node; else node.next->prev = &node; unlock(&c->_c_lock); } __pthread_mutex_unlock(m); __pthread_setcancelstate(PTHREAD_CANCEL_MASKED, &cs); if (cs == PTHREAD_CANCEL_DISABLE) __pthread_setcancelstate(cs, 0); do e = __timedwait_cp(fut, seq, clock, ts, !shared); while (*fut==seq && (!e || e==EINTR)); if (e == EINTR) e = 0; if (shared) { /* Suppress cancellation if a signal was potentially * consumed; this is a legitimate form of spurious * wake even if not. */ if (e == ECANCELED && c->_c_seq != seq) e = 0; if (a_fetch_add(&c->_c_waiters, -1) == -0x7fffffff) __wake(&c->_c_waiters, 1, 0); oldstate = WAITING; goto relock; } oldstate = a_cas(&node.state, WAITING, LEAVING); if (oldstate == WAITING) { /* Access to cv object is valid because this waiter was not * yet signaled and a new signal/broadcast cannot return * after seeing a LEAVING waiter without getting notified * via the futex notify below. */ lock(&c->_c_lock); if (c->_c_head == &node) c->_c_head = node.next; else if (node.prev) node.prev->next = node.next; if (c->_c_tail == &node) c->_c_tail = node.prev; else if (node.next) node.next->prev = node.prev; unlock(&c->_c_lock); if (node.notify) { if (a_fetch_add(node.notify, -1)==1) __wake(node.notify, 1, 1); } } else { /* Lock barrier first to control wake order. */ lock(&node.barrier); } relock: /* Errors locking the mutex override any existing error or * cancellation, since the caller must see them to know the * state of the mutex. */ if ((tmp = pthread_mutex_lock(m))) e = tmp; if (oldstate == WAITING) goto done; if (!node.next) a_inc(&m->_m_waiters); /* Unlock the barrier that's holding back the next waiter, and * either wake it or requeue it to the mutex. */ if (node.prev) unlock_requeue(&node.prev->barrier, &m->_m_lock, m->_m_type & 128); else a_dec(&m->_m_waiters); /* Since a signal was consumed, cancellation is not permitted. */ if (e == ECANCELED) e = 0; done: __pthread_setcancelstate(cs, 0); if (e == ECANCELED) { __pthread_testcancel(); __pthread_setcancelstate(PTHREAD_CANCEL_DISABLE, 0); } return e; } int __private_cond_signal(pthread_cond_t *c, int n) { struct waiter *p, *first=0; volatile int ref = 0; int cur; lock(&c->_c_lock); for (p=c->_c_tail; n && p; p=p->prev) { if (a_cas(&p->state, WAITING, SIGNALED) != WAITING) { ref++; p->notify = &ref; } else { n--; if (!first) first=p; } } /* Split the list, leaving any remainder on the cv. */ if (p) { if (p->next) p->next->prev = 0; p->next = 0; } else { c->_c_head = 0; } c->_c_tail = p; unlock(&c->_c_lock); /* Wait for any waiters in the LEAVING state to remove * themselves from the list before returning or allowing * signaled threads to proceed. */ while ((cur = ref)) __wait(&ref, 0, cur, 1); /* Allow first signaled waiter, if any, to proceed. */ if (first) unlock(&first->barrier); return 0; } weak_alias(__pthread_cond_timedwait, pthread_cond_timedwait);