/* thread.cc: Locking and threading module functions This file is part of Cygwin. This software is a copyrighted work licensed under the terms of the Cygwin license. Please consult the file "CYGWIN_LICENSE" for details. */ /* Implementation overview and caveats: Win32 puts some contraints on what can and cannot be implemented. Where possible we work around those contrainsts. Where we cannot work around the constraints we either pretend to be conformant, or return an error code. Some caveats: PROCESS_SHARED objects, while they pretend to be process shared, may not actually work. Some test cases are needed to determine win32's behaviour. My suspicion is that the win32 handle needs to be opened with different flags for proper operation. R.Collins, April 2001. */ #include "winsup.h" #include "create_posix_thread.h" #include "path.h" #include #include #include "sigproc.h" #include "fhandler.h" #include "dtable.h" #include "cygheap.h" #include "ntdll.h" #include "cygwait.h" #include "exception.h" /* For Linux compatibility, the length of a thread name is 16 characters. */ #define THRNAMELEN 16 extern "C" void __fp_lock_all (); extern "C" void __fp_unlock_all (); extern "C" bool valid_sched_parameters(const struct sched_param *); extern "C" int sched_get_thread_priority(HANDLE thread); extern "C" int sched_set_thread_priority(HANDLE thread, int priority); extern int threadsafe; const pthread_t pthread_mutex::_new_mutex = (pthread_t) 1; const pthread_t pthread_mutex::_unlocked_mutex = (pthread_t) 2; const pthread_t pthread_mutex::_destroyed_mutex = (pthread_t) 3; template static inline void delete_and_clear (T * * const ptr) { delete *ptr; *ptr = 0; } inline bool pthread_mutex::no_owner() { int res; if (!owner) { debug_printf ("NULL owner value"); res = 1; } else if (owner == _destroyed_mutex) { paranoid_printf ("attempt to use destroyed mutex"); res = 1; } else if (owner == _new_mutex || owner == _unlocked_mutex) res = 1; else res = 0; return res; } #undef __getreent extern "C" struct _reent * __getreent () { return &_my_tls.local_clib; } extern "C" void __cygwin_lock_init (_LOCK_T *lock) { *lock = _LOCK_T_INITIALIZER; } extern "C" void __cygwin_lock_init_recursive (_LOCK_T *lock) { *lock = _LOCK_T_RECURSIVE_INITIALIZER; } extern "C" void __cygwin_lock_fini (_LOCK_T *lock) { pthread_mutex_destroy ((pthread_mutex_t*) lock); } extern "C" void __cygwin_lock_lock (_LOCK_T *lock) { paranoid_printf ("threadcount %d. locking", MT_INTERFACE->threadcount); pthread_mutex_lock ((pthread_mutex_t*) lock); } extern "C" int __cygwin_lock_trylock (_LOCK_T *lock) { return pthread_mutex_trylock ((pthread_mutex_t*) lock); } extern "C" void __cygwin_lock_unlock (_LOCK_T *lock) { pthread_mutex_unlock ((pthread_mutex_t*) lock); paranoid_printf ("threadcount %d. unlocked", MT_INTERFACE->threadcount); } static inline verifyable_object_state verifyable_object_isvalid (void const *objectptr, thread_magic_t magic, void *static_ptr1 = NULL, void *static_ptr2 = NULL, void *static_ptr3 = NULL) { verifyable_object_state state = INVALID_OBJECT; __try { if (!objectptr || !(*(const char **) objectptr)) __leave; verifyable_object **object = (verifyable_object **) objectptr; if ((static_ptr1 && *object == static_ptr1) || (static_ptr2 && *object == static_ptr2) || (static_ptr3 && *object == static_ptr3)) state = VALID_STATIC_OBJECT; else if ((*object)->magic == magic) state = VALID_OBJECT; } __except (NO_ERROR) {} __endtry return state; } /* static members */ inline bool pthread_attr::is_good_object (pthread_attr_t const *attr) { if (verifyable_object_isvalid (attr, PTHREAD_ATTR_MAGIC) != VALID_OBJECT) return false; return true; } inline bool pthread_condattr::is_good_object (pthread_condattr_t const *attr) { if (verifyable_object_isvalid (attr, PTHREAD_CONDATTR_MAGIC) != VALID_OBJECT) return false; return true; } inline bool pthread_rwlockattr::is_good_object (pthread_rwlockattr_t const *attr) { if (verifyable_object_isvalid (attr, PTHREAD_RWLOCKATTR_MAGIC) != VALID_OBJECT) return false; return true; } inline bool pthread_key::is_good_object (pthread_key_t const *key) { if (verifyable_object_isvalid (key, PTHREAD_KEY_MAGIC) != VALID_OBJECT) return false; return true; } inline bool pthread_spinlock::is_good_object (pthread_spinlock_t const *mutex) { if (verifyable_object_isvalid (mutex, PTHREAD_SPINLOCK_MAGIC) != VALID_OBJECT) return false; return true; } inline bool pthread_mutex::is_good_object (pthread_mutex_t const *mutex) { if (verifyable_object_isvalid (mutex, PTHREAD_MUTEX_MAGIC) != VALID_OBJECT) return false; return true; } inline bool pthread_mutex::is_initializer (pthread_mutex_t const *mutex) { if (verifyable_object_isvalid (mutex, PTHREAD_MUTEX_MAGIC, PTHREAD_RECURSIVE_MUTEX_INITIALIZER_NP, PTHREAD_NORMAL_MUTEX_INITIALIZER_NP, PTHREAD_ERRORCHECK_MUTEX_INITIALIZER_NP) != VALID_STATIC_OBJECT) return false; return true; } inline bool pthread_mutex::is_initializer_or_object (pthread_mutex_t const *mutex) { if (verifyable_object_isvalid (mutex, PTHREAD_MUTEX_MAGIC, PTHREAD_RECURSIVE_MUTEX_INITIALIZER_NP, PTHREAD_NORMAL_MUTEX_INITIALIZER_NP, PTHREAD_ERRORCHECK_MUTEX_INITIALIZER_NP) == INVALID_OBJECT) return false; return true; } /* FIXME: Accommodate PTHREAD_MUTEX_ERRORCHECK */ inline bool pthread_mutex::can_be_unlocked () { pthread_t self = pthread::self (); /* Check if the mutex is owned by the current thread and can be unlocked. * Also check for the ANONYMOUS owner to cover NORMAL mutexes as well. */ bool res = type == PTHREAD_MUTEX_NORMAL || no_owner () || (recursion_counter == 1 && pthread::equal (owner, self)); pthread_printf ("recursion_counter %u res %d", recursion_counter, res); return res; } inline bool pthread_mutexattr::is_good_object (pthread_mutexattr_t const * attr) { if (verifyable_object_isvalid (attr, PTHREAD_MUTEXATTR_MAGIC) != VALID_OBJECT) return false; return true; } inline bool __attribute__ ((used)) pthread::is_good_object (pthread_t const *thread) { if (verifyable_object_isvalid (thread, PTHREAD_MAGIC) != VALID_OBJECT) return false; return true; } /* Thread synchronisation */ inline bool pthread_cond::is_good_object (pthread_cond_t const *cond) { if (verifyable_object_isvalid (cond, PTHREAD_COND_MAGIC) != VALID_OBJECT) return false; return true; } inline bool pthread_cond::is_initializer (pthread_cond_t const *cond) { if (verifyable_object_isvalid (cond, PTHREAD_COND_MAGIC, PTHREAD_COND_INITIALIZER) != VALID_STATIC_OBJECT) return false; return true; } inline bool pthread_cond::is_initializer_or_object (pthread_cond_t const *cond) { if (verifyable_object_isvalid (cond, PTHREAD_COND_MAGIC, PTHREAD_COND_INITIALIZER) == INVALID_OBJECT) return false; return true; } inline bool pthread_barrierattr::is_good_object (pthread_barrierattr_t const *cond) { if (verifyable_object_isvalid (cond, PTHREAD_BARRIERATTR_MAGIC) != VALID_OBJECT) return false; return true; } inline bool pthread_barrier::is_good_object (pthread_barrier_t const *cond) { if (verifyable_object_isvalid (cond, PTHREAD_BARRIER_MAGIC) != VALID_OBJECT) return false; return true; } /* RW locks */ inline bool pthread_rwlock::is_good_object (pthread_rwlock_t const *rwlock) { if (verifyable_object_isvalid (rwlock, PTHREAD_RWLOCK_MAGIC) != VALID_OBJECT) return false; return true; } inline bool pthread_rwlock::is_initializer (pthread_rwlock_t const *rwlock) { if (verifyable_object_isvalid (rwlock, PTHREAD_RWLOCK_MAGIC, PTHREAD_RWLOCK_INITIALIZER) != VALID_STATIC_OBJECT) return false; return true; } inline bool pthread_rwlock::is_initializer_or_object (pthread_rwlock_t const *rwlock) { if (verifyable_object_isvalid (rwlock, PTHREAD_RWLOCK_MAGIC, PTHREAD_RWLOCK_INITIALIZER) == INVALID_OBJECT) return false; return true; } inline bool semaphore::is_good_object (sem_t const * sem) { if (verifyable_object_isvalid (sem, SEM_MAGIC) != VALID_OBJECT) return false; return true; } void MTinterface::Init () { pthread_mutex::init_mutex (); pthread_cond::init_mutex (); pthread_rwlock::init_mutex (); } void MTinterface::fixup_before_fork () { pthread_key::fixup_before_fork (); semaphore::fixup_before_fork (); } /* This function is called from a single threaded process */ void MTinterface::fixup_after_fork () { pthread_key::fixup_after_fork (); threadcount = 0; pthread::init_mainthread (); pthread::fixup_after_fork (); pthread_mutex::fixup_after_fork (); pthread_cond::fixup_after_fork (); pthread_rwlock::fixup_after_fork (); semaphore::fixup_after_fork (); } /* pthread calls */ /* static methods */ void pthread::init_mainthread () { pthread *thread = _my_tls.tid; if (!thread || thread == pthread_null::get_null_pthread ()) { thread = new pthread (); if (!thread) api_fatal ("failed to create mainthread object"); } thread->set_tls_self_pointer (); thread->thread_id = GetCurrentThreadId (); if (!DuplicateHandle (GetCurrentProcess (), GetCurrentThread (), GetCurrentProcess (), &thread->win32_obj_id, 0, FALSE, DUPLICATE_SAME_ACCESS)) api_fatal ("failed to create mainthread handle"); if (!thread->create_cancel_event ()) api_fatal ("couldn't create cancel event for main thread"); VerifyHandle (thread->win32_obj_id); /* Make sure the pthread mutex is recursive. See comment in pthread::precreate (called only for subsequent pthreads) for a description. */ thread->mutex.set_type (PTHREAD_MUTEX_RECURSIVE); thread->postcreate (); } pthread * pthread::self () { pthread *thread = _my_tls.tid; if (!thread) { thread = pthread_null::get_null_pthread (); thread->set_tls_self_pointer (); } return thread; } void pthread::set_tls_self_pointer () { cygtls = &_my_tls; _my_tls.tid = this; } List pthread::threads; /* member methods */ pthread::pthread ():verifyable_object (PTHREAD_MAGIC), win32_obj_id (0), valid (false), suspended (false), canceled (false), cancelstate (0), canceltype (0), cancel_event (0), joiner (NULL), next (NULL), cleanup_stack (NULL) { if (this != pthread_null::get_null_pthread ()) threads.insert (this); sigprocmask (SIG_SETMASK, NULL, &parent_sigmask); } pthread::~pthread () { if (win32_obj_id) CloseHandle (win32_obj_id); if (cancel_event) CloseHandle (cancel_event); if (this != pthread_null::get_null_pthread ()) threads.remove (this); } bool pthread::create_cancel_event () { cancel_event = ::CreateEvent (&sec_none_nih, true, false, NULL); if (!cancel_event) { system_printf ("couldn't create cancel event, %E"); /* we need the event for correct behaviour */ return false; } return true; } void pthread::precreate (pthread_attr *newattr) { pthread_mutex *verifyable_mutex_obj = &mutex; /* already running ? */ if (win32_obj_id) return; if (newattr) { attr.joinable = newattr->joinable; attr.contentionscope = newattr->contentionscope; attr.inheritsched = newattr->inheritsched; attr.schedparam = newattr->schedparam; attr.stackaddr = newattr->stackaddr; attr.stacksize = newattr->stacksize; attr.guardsize = newattr->guardsize; } if (!pthread_mutex::is_good_object (&verifyable_mutex_obj)) { thread_printf ("New thread object access mutex is not valid. this %p", this); magic = 0; return; } /* This mutex MUST be recursive. Consider the following scenario: - The thread installs a cleanup handler. - The cleanup handler calls a function which itself installs a cleanup handler. - pthread_cancel is called for this thread. - The thread's cleanup handler is called under mutex lock condition. - The cleanup handler calls the subsequent function with cleanup handler. - The function runs to completion, so it calls pthread_cleanup_pop. - pthread_cleanup_pop calls pthread::pop_cleanup_handler which will again try to lock the mutex. - Deadlock. */ mutex.set_type (PTHREAD_MUTEX_RECURSIVE); if (!create_cancel_event ()) magic = 0; } bool pthread::create (void *(*func) (void *), pthread_attr *newattr, void *threadarg) { bool retval; precreate (newattr); if (!magic) return false; function = func; arg = threadarg; mutex.lock (); /* stackaddr holds the uppermost stack address. See the comments in pthread_attr_setstack and pthread_attr_setstackaddr for a description. */ ULONG stacksize = attr.stacksize ?: get_rlimit_stack (); PVOID stackaddr = attr.stackaddr ? ((caddr_t) attr.stackaddr - stacksize) : NULL; win32_obj_id = create_posix_thread (thread_init_wrapper, this, stackaddr, stacksize, attr.guardsize, 0, &thread_id); if (!win32_obj_id) { thread_printf ("CreateThread failed: this %p, %E", this); magic = 0; } else { postcreate (); while (!cygtls) yield (); } retval = magic; mutex.unlock (); return retval; } void pthread::postcreate () { valid = true; InterlockedIncrement (&MT_INTERFACE->threadcount); /* Per POSIX the new thread inherits the sched priority from its caller thread if PTHREAD_INHERIT_SCHED is set. FIXME: set the priority appropriately for system contention scope */ if (attr.inheritsched == PTHREAD_INHERIT_SCHED) attr.schedparam.sched_priority = sched_get_thread_priority (GetCurrentThread ()); if (attr.schedparam.sched_priority) sched_set_thread_priority (win32_obj_id, attr.schedparam.sched_priority); } void pthread::exit (void *value_ptr) { class pthread *thread = this; _cygtls *tls = cygtls; /* Save cygtls before deleting this. */ // run cleanup handlers pop_all_cleanup_handlers (); pthread_key::run_all_destructors (); mutex.lock (); // cleanup if thread is in detached state and not joined if (equal (joiner, thread)) delete this; else { valid = false; return_ptr = value_ptr; mutex.unlock (); } if (_my_tls.local_clib.__cleanup == _cygtls::cleanup_early) _my_tls.local_clib.__cleanup = NULL; _reclaim_reent (_REENT); if (InterlockedDecrement (&MT_INTERFACE->threadcount) == 0) ::exit (0); else { if (tls == _main_tls) { cygheap->find_tls (tls); /* Lock _main_tls mutex. */ _cygtls *dummy = (_cygtls *) malloc (sizeof (_cygtls)); *dummy = *_main_tls; _main_tls = dummy; _main_tls->initialized = 0; } /* This also unlocks and closes the _main_tls mutex. */ tls->remove (INFINITE); ExitThread (0); } } int pthread::cancel () { class pthread *thread = this; class pthread *self = pthread::self (); mutex.lock (); if (!valid) { mutex.unlock (); return 0; } if (canceltype == PTHREAD_CANCEL_DEFERRED || cancelstate == PTHREAD_CANCEL_DISABLE) { // cancel deferred mutex.unlock (); canceled = true; SetEvent (cancel_event); return 0; } else if (equal (thread, self)) { mutex.unlock (); cancel_self (); return 0; // Never reached } // cancel asynchronous SuspendThread (win32_obj_id); if (WaitForSingleObject (win32_obj_id, 0) == WAIT_TIMEOUT) { CONTEXT context; context.ContextFlags = CONTEXT_CONTROL; GetThreadContext (win32_obj_id, &context); /* The OS is not foolproof in terms of asynchronous thread cancellation and tends to hang infinitely if we change the instruction pointer. So just don't cancel asynchronously if the thread is currently executing Windows code. Rely on deferred cancellation in this case. */ threadlist_t *tl_entry = cygheap->find_tls (cygtls); if (!cygtls->inside_kernel (&context)) { context.Rip = (ULONG_PTR) pthread::static_cancel_self; SetThreadContext (win32_obj_id, &context); } cygheap->unlock_tls (tl_entry); } mutex.unlock (); /* See above. For instance, a thread which waits for a semaphore in sem_wait will call cygwait which in turn calls WFMO. While this WFMO call is cancelable by setting the thread's cancel_event object, the OS apparently refuses to set the thread's context and continues to wait for the WFMO conditions. This is *not* reflected in the return value of SetThreadContext or ResumeThread, btw. So, what we do here is to set the cancel_event as well to allow at least a deferred cancel. */ canceled = true; SetEvent (cancel_event); ResumeThread (win32_obj_id); return 0; } /* TODO: Insert pthread_testcancel into the required functions. Here are the lists of required and optional functions per POSIX.1-2001 and POSIX.1-2008. A star (*) indicates that the Cygwin function already is a cancellation point (aka "calls pthread_testcancel"), an o (o) indicates that the function is not implemented in Cygwin. Required cancellation points: * accept () * aio_suspend () * clock_nanosleep () * close () * connect () * creat () * fcntl () F_SETLKW * fdatasync () * fsync () o getmsg () o getpmsg () * lockf () F_LOCK * mq_receive () * mq_send () * mq_timedreceive () * mq_timedsend () msgrcv () msgsnd () * msync () * nanosleep () * open () * openat () * pause () * poll () * pread () * pselect () * pthread_cond_timedwait () * pthread_cond_wait () * pthread_join () * pthread_testcancel () o putmsg () o putpmsg () * pwrite () * read () * readv () * recv () * recvfrom () * recvmsg () * select () * sem_timedwait () * sem_wait () * send () * sendmsg () * sendto () * sigpause () * sigsuspend () * sigtimedwait () * sigwait () * sigwaitinfo () * sleep () * system () * tcdrain () * usleep () * wait () * wait3() o waitid () * waitpid () * write () * writev () Optional cancellation points: access () asctime () asctime_r () catclose () Implemented externally: libcatgets catgets () Implemented externally: libcatgets catopen () Implemented externally: libcatgets chmod () chown () closedir () closelog () ctermid () ctime () ctime_r () dbm_close () Implemented externally: libgdbm dbm_delete () Implemented externally: libgdbm dbm_fetch () Implemented externally: libgdbm dbm_nextkey () Implemented externally: libgdbm dbm_open () Implemented externally: libgdbm dbm_store () Implemented externally: libgdbm dlclose () dlopen () dprintf () endgrent () endhostent () o endnetent () endprotoent () endpwent () endservent () endutxent () faccessat () fchmod () fchmodat () fchown () fchownat () * fclose () * fcntl () (any value) fflush () fgetc () fgetpos () fgets () fgetwc () fgetws () o fmtmsg () fopen () fpathconf () fprintf () fputc () fputs () fputwc () fputws () fread () freopen () fscanf () fseek () fseeko () fsetpos () fstat () fstatat () ftell () ftello () ftw () futimens () fwprintf () fwrite () fwscanf () getaddrinfo () getc () getc_unlocked () getchar () getchar_unlocked () getcwd () o getdate () getdelim () getgrent () getgrgid () getgrgid_r () getgrnam () getgrnam_r () gethostbyaddr () gethostbyname () gethostent () gethostid () gethostname () getline () getlogin () getlogin_r () getnameinfo () o getnetbyaddr () o getnetbyname () o getnetent () getopt () (if opterr is nonzero) getprotobyname () getprotobynumber () getprotoent () getpwent () * getpwnam () * getpwnam_r () * getpwuid () * getpwuid_r () gets () getservbyname () getservbyport () getservent () getutxent () getutxid () getutxline () getwc () getwchar () getwd () glob () iconv_close () Implemented externally: libiconv iconv_open () Implemented externally: libiconv ioctl () link () linkat () * lio_listio () localtime () localtime_r () * lockf () lseek () lstat () mkdir () mkdirat () mkdtemp () mkfifo () mkfifoat () mknod () mknodat () mkstemp () mktime () nftw () opendir () openlog () pathconf () pclose () perror () popen () posix_fadvise () posix_fallocate () posix_madvise () posix_openpt () posix_spawn () posix_spawnp () o posix_trace_clear () o posix_trace_close () o posix_trace_create () o posix_trace_create_withlog () o posix_trace_eventtypelist_getnext_id () o posix_trace_eventtypelist_rewind () o posix_trace_flush () o posix_trace_get_attr () o posix_trace_get_filter () o posix_trace_get_status () o posix_trace_getnext_event () o posix_trace_open () o posix_trace_rewind () o posix_trace_set_filter () o posix_trace_shutdown () o posix_trace_timedgetnext_event () o posix_typed_mem_open () printf () psiginfo () psignal () pthread_rwlock_rdlock () o pthread_rwlock_timedrdlock () o pthread_rwlock_timedwrlock () pthread_rwlock_wrlock () putc () putc_unlocked () putchar () putchar_unlocked () puts () pututxline () putwc () putwchar () readdir () readdir_r () readlink () readlinkat () remove () rename () renameat () rewind () rewinddir () scandir () scanf () seekdir () semop () setgrent () sethostent () o setnetent () setprotoent () setpwent () setservent () setutxent () sigpause () stat () strerror () strerror_r () strftime () symlink () symlinkat () sync () syslog () tmpfile () tmpnam () ttyname () ttyname_r () tzset () ungetc () ungetwc () unlink () unlinkat () utime () utimensat () utimes () vdprintf () vfprintf () vfwprintf () vprintf () vwprintf () wcsftime () wordexp () wprintf () wscanf () An implementation may also mark other functions not specified in the standard as cancellation points. In particular, an implementation is likely to mark any nonstandard function that may block as a cancellation point. */ void pthread::testcancel () { if (cancelstate == PTHREAD_CANCEL_DISABLE) return; /* We check for the canceled flag first. This allows to use the pthread_testcancel function a lot without adding the overhead of an OS call. Only if the thread is marked as canceled, we wait for cancel_event being really set, on the off-chance that pthread_cancel gets interrupted or terminated before calling SetEvent. */ if (canceled && IsEventSignalled (cancel_event)) cancel_self (); } /* Return cancel event handle if it exists *and* cancel is not disabled. This function is supposed to be used from other functions which are cancelable and need the cancel event in a WFMO call. */ HANDLE pthread::get_cancel_event () { pthread_t thread = pthread::self (); return (thread && thread->cancel_event && thread->cancelstate != PTHREAD_CANCEL_DISABLE) ? thread->cancel_event : NULL; } void pthread::static_cancel_self () { pthread::self ()->cancel_self (); } int pthread::setcancelstate (int state, int *oldstate) { if (state != PTHREAD_CANCEL_ENABLE && state != PTHREAD_CANCEL_DISABLE) return EINVAL; if (oldstate) *oldstate = cancelstate; cancelstate = state; return 0; } int pthread::setcanceltype (int type, int *oldtype) { if (type != PTHREAD_CANCEL_DEFERRED && type != PTHREAD_CANCEL_ASYNCHRONOUS) return EINVAL; if (oldtype) *oldtype = canceltype; canceltype = type; return 0; } void pthread::push_cleanup_handler (__pthread_cleanup_handler *handler) { if (this != self ()) // TODO: do it? api_fatal ("Attempt to push a cleanup handler across threads"); handler->next = cleanup_stack; cleanup_stack = handler; } void pthread::pop_cleanup_handler (int const execute) { if (this != self ()) // TODO: send a signal or something to the thread ? api_fatal ("Attempt to execute a cleanup handler across threads"); mutex.lock (); if (cleanup_stack != NULL) { __pthread_cleanup_handler *handler = cleanup_stack; if (execute) (*handler->function) (handler->arg); cleanup_stack = handler->next; } mutex.unlock (); } void pthread::pop_all_cleanup_handlers () { /* We will no honor cancels since the thread is exiting. */ cancelstate = PTHREAD_CANCEL_DISABLE; while (cleanup_stack != NULL) pop_cleanup_handler (1); } void pthread::cancel_self () { /* Can someone explain why the pthread:: is needed here? g++ complains without it. */ pthread::exit (PTHREAD_CANCELED); } DWORD pthread::get_thread_id () { return thread_id; } void pthread::_fixup_after_fork () { /* set thread to not running if it is not the forking thread */ if (this != pthread::self ()) { magic = 0; valid = false; win32_obj_id = NULL; canceled = false; cancel_event = NULL; } } void pthread::suspend_except_self () { if (valid && this != pthread::self ()) SuspendThread (win32_obj_id); } void pthread::resume () { if (valid) ResumeThread (win32_obj_id); } /* instance members */ pthread_attr::pthread_attr ():verifyable_object (PTHREAD_ATTR_MAGIC), joinable (PTHREAD_CREATE_JOINABLE), contentionscope (PTHREAD_SCOPE_PROCESS), inheritsched (PTHREAD_INHERIT_SCHED), stackaddr (NULL), stacksize (0), guardsize (wincap.def_guard_page_size ()), name (NULL) { schedparam.sched_priority = 0; } pthread_attr::~pthread_attr () { } pthread_condattr::pthread_condattr ():verifyable_object (PTHREAD_CONDATTR_MAGIC), shared (PTHREAD_PROCESS_PRIVATE), clock_id (CLOCK_REALTIME) { } pthread_condattr::~pthread_condattr () { } List pthread_cond::conds; /* This is used for cond creation protection within a single process only */ fast_mutex NO_COPY pthread_cond::cond_initialization_lock; /* We can only be called once. TODO: (no rush) use a non copied memory section to hold an initialization flag. */ void pthread_cond::init_mutex () { if (!cond_initialization_lock.init ()) api_fatal ("Could not create win32 Mutex for pthread cond static initializer support."); } pthread_cond::pthread_cond (pthread_condattr *attr) : verifyable_object (PTHREAD_COND_MAGIC), shared (0), clock_id (CLOCK_REALTIME), waiting (0), pending (0), sem_wait (NULL), mtx_cond(NULL), next (NULL) { pthread_mutex *verifyable_mutex_obj; if (attr) { clock_id = attr->clock_id; if (attr->shared != PTHREAD_PROCESS_PRIVATE) { magic = 0; return; } } verifyable_mutex_obj = &mtx_in; if (!pthread_mutex::is_good_object (&verifyable_mutex_obj)) { thread_printf ("Internal cond mutex is not valid. this %p", this); magic = 0; return; } /* * Change the mutex type to NORMAL. * This mutex MUST be of type normal */ mtx_in.set_type (PTHREAD_MUTEX_NORMAL); verifyable_mutex_obj = &mtx_out; if (!pthread_mutex::is_good_object (&verifyable_mutex_obj)) { thread_printf ("Internal cond mutex is not valid. this %p", this); magic = 0; return; } /* Change the mutex type to NORMAL to speed up mutex operations */ mtx_out.set_type (PTHREAD_MUTEX_NORMAL); sem_wait = ::CreateSemaphore (&sec_none_nih, 0, INT32_MAX, NULL); if (!sem_wait) { pthread_printf ("CreateSemaphore failed. %E"); magic = 0; return; } conds.insert (this); } pthread_cond::~pthread_cond () { if (sem_wait) CloseHandle (sem_wait); conds.remove (this); } void pthread_cond::unblock (const bool all) { LONG releaseable; /* * Block outgoing threads (and avoid simultanous unblocks) */ mtx_out.lock (); releaseable = waiting - pending; if (releaseable) { LONG released; if (!pending) { /* * Block incoming threads until all waiting threads are released. */ mtx_in.lock (); /* * Calculate releaseable again because threads can enter until * the semaphore has been taken, but they can not leave, therefore pending * is unchanged and releaseable can only get higher */ releaseable = waiting - pending; } released = all ? releaseable : 1; pending += released; /* * Signal threads */ ::ReleaseSemaphore (sem_wait, released, NULL); } /* * And let the threads release. */ mtx_out.unlock (); } int pthread_cond::wait (pthread_mutex_t mutex, PLARGE_INTEGER timeout) { DWORD rv; mtx_in.lock (); if (InterlockedIncrement (&waiting) == 1) mtx_cond = mutex; else if (mtx_cond != mutex) { InterlockedDecrement (&waiting); mtx_in.unlock (); return EINVAL; } mtx_in.unlock (); /* * Release the mutex and wait on semaphore */ ++mutex->condwaits; mutex->unlock (); rv = cygwait (sem_wait, timeout, cw_cancel | cw_sig_restart); mtx_out.lock (); if (rv != WAIT_OBJECT_0 && WaitForSingleObject (sem_wait, 0) == WAIT_OBJECT_0) /* Thread got cancelled ot timed out while a signalling is in progress. Set wait result back to signaled */ rv = WAIT_OBJECT_0; InterlockedDecrement (&waiting); if (rv == WAIT_OBJECT_0 && --pending == 0) /* * All signaled threads are released, * new threads can enter Wait */ mtx_in.unlock (); mtx_out.unlock (); mutex->lock (); --mutex->condwaits; if (rv == WAIT_CANCELED) pthread::static_cancel_self (); else if (rv == WAIT_TIMEOUT) return ETIMEDOUT; return 0; } void pthread_cond::_fixup_after_fork () { waiting = pending = 0; mtx_cond = NULL; /* Unlock eventually locked mutexes */ mtx_in.unlock (); mtx_out.unlock (); sem_wait = ::CreateSemaphore (&sec_none_nih, 0, INT32_MAX, NULL); if (!sem_wait) api_fatal ("pthread_cond::_fixup_after_fork () failed to recreate win32 semaphore"); } pthread_barrierattr::pthread_barrierattr () : verifyable_object (PTHREAD_BARRIERATTR_MAGIC) , shared (PTHREAD_PROCESS_PRIVATE) { } pthread_barrierattr::~pthread_barrierattr () { } pthread_barrier::pthread_barrier () : verifyable_object (PTHREAD_BARRIER_MAGIC) { } pthread_barrier::~pthread_barrier () { } pthread_rwlockattr::pthread_rwlockattr ():verifyable_object (PTHREAD_RWLOCKATTR_MAGIC), shared (PTHREAD_PROCESS_PRIVATE) { } pthread_rwlockattr::~pthread_rwlockattr () { } List pthread_rwlock::rwlocks; /* This is used for rwlock creation protection within a single process only */ fast_mutex NO_COPY pthread_rwlock::rwlock_initialization_lock; /* We can only be called once. TODO: (no rush) use a non copied memory section to hold an initialization flag. */ void pthread_rwlock::init_mutex () { if (!rwlock_initialization_lock.init ()) api_fatal ("Could not create win32 Mutex for pthread rwlock static initializer support."); } pthread_rwlock::pthread_rwlock (pthread_rwlockattr *attr) : verifyable_object (PTHREAD_RWLOCK_MAGIC), shared (0), waiting_readers (0), waiting_writers (0), writer (NULL), readers (NULL), readers_mx (), mtx (NULL), cond_readers (NULL), cond_writers (NULL), next (NULL) { pthread_mutex *verifyable_mutex_obj = &mtx; pthread_cond *verifyable_cond_obj; if (!readers_mx.init ()) { thread_printf ("Internal rwlock synchronisation mutex is not valid. this %p", this); magic = 0; return; } if (attr) if (attr->shared != PTHREAD_PROCESS_PRIVATE) { magic = 0; return; } if (!pthread_mutex::is_good_object (&verifyable_mutex_obj)) { thread_printf ("Internal rwlock mutex is not valid. this %p", this); magic = 0; return; } /* Change the mutex type to NORMAL to speed up mutex operations */ mtx.set_type (PTHREAD_MUTEX_NORMAL); verifyable_cond_obj = &cond_readers; if (!pthread_cond::is_good_object (&verifyable_cond_obj)) { thread_printf ("Internal rwlock readers cond is not valid. this %p", this); magic = 0; return; } verifyable_cond_obj = &cond_writers; if (!pthread_cond::is_good_object (&verifyable_cond_obj)) { thread_printf ("Internal rwlock writers cond is not valid. this %p", this); magic = 0; return; } rwlocks.insert (this); } pthread_rwlock::~pthread_rwlock () { rwlocks.remove (this); } int pthread_rwlock::rdlock (PLARGE_INTEGER timeout) { int result = 0; struct RWLOCK_READER *reader; mtx.lock (); reader = lookup_reader (); if (reader) { if (reader->n < UINT32_MAX) ++reader->n; else result = EAGAIN; goto DONE; } while (writer || waiting_writers) { int ret; pthread_cleanup_push (pthread_rwlock::rdlock_cleanup, this); ++waiting_readers; ret = cond_readers.wait (&mtx, timeout); --waiting_readers; pthread_cleanup_pop (0); if (ret == ETIMEDOUT) { result = ETIMEDOUT; goto DONE; } } if ((reader = add_reader ())) ++reader->n; else { result = EAGAIN; goto DONE; } DONE: mtx.unlock (); return result; } int pthread_rwlock::tryrdlock () { int result = 0; mtx.lock (); if (writer || waiting_writers) result = EBUSY; else { RWLOCK_READER *reader = lookup_reader (); if (!reader) reader = add_reader (); if (reader && reader->n < UINT32_MAX) ++reader->n; else result = EAGAIN; } mtx.unlock (); return result; } int pthread_rwlock::wrlock (PLARGE_INTEGER timeout) { int result = 0; pthread_t self = pthread::self (); mtx.lock (); if (writer == self || lookup_reader ()) { result = EDEADLK; goto DONE; } while (writer || readers) { int ret; pthread_cleanup_push (pthread_rwlock::wrlock_cleanup, this); ++waiting_writers; ret = cond_writers.wait (&mtx, timeout); --waiting_writers; pthread_cleanup_pop (0); if (ret == ETIMEDOUT) { result = ETIMEDOUT; goto DONE; } } writer = self; DONE: mtx.unlock (); return result; } int pthread_rwlock::trywrlock () { int result = 0; pthread_t self = pthread::self (); mtx.lock (); if (writer || readers) result = EBUSY; else writer = self; mtx.unlock (); return result; } int pthread_rwlock::unlock () { int result = 0; mtx.lock (); if (writer) { if (writer != pthread::self ()) { result = EPERM; goto DONE; } writer = NULL; } else { struct RWLOCK_READER *reader = lookup_reader (); if (!reader) { result = EPERM; goto DONE; } if (--reader->n > 0) goto DONE; remove_reader (reader); delete reader; } release (); DONE: mtx.unlock (); return result; } pthread_rwlock::RWLOCK_READER * pthread_rwlock::add_reader () { RWLOCK_READER *rd = new RWLOCK_READER; if (rd) List_insert_nolock (readers, rd); return rd; } void pthread_rwlock::remove_reader (struct RWLOCK_READER *rd) { List_remove (readers_mx, readers, rd); } struct pthread_rwlock::RWLOCK_READER * pthread_rwlock::lookup_reader () { readers_mx.lock (); pthread_t thread = pthread::self (); struct RWLOCK_READER *cur = readers; while (cur && cur->thread != thread) cur = cur->next; readers_mx.unlock (); return cur; } void pthread_rwlock::rdlock_cleanup (void *arg) { pthread_rwlock *rwlock = (pthread_rwlock *) arg; --(rwlock->waiting_readers); rwlock->release (); rwlock->mtx.unlock (); } void pthread_rwlock::wrlock_cleanup (void *arg) { pthread_rwlock *rwlock = (pthread_rwlock *) arg; --(rwlock->waiting_writers); rwlock->release (); rwlock->mtx.unlock (); } void pthread_rwlock::_fixup_after_fork () { pthread_t self = pthread::self (); struct RWLOCK_READER **temp = &readers; waiting_readers = 0; waiting_writers = 0; if (!readers_mx.init ()) api_fatal ("pthread_rwlock::_fixup_after_fork () failed to recreate mutex"); /* Unlock eventually locked mutex */ mtx.unlock (); /* * Remove all readers except self */ while (*temp) { if ((*temp)->thread == self) temp = &((*temp)->next); else { struct RWLOCK_READER *cur = *temp; *temp = (*temp)->next; delete cur; } } } /* pthread_key */ /* static members */ /* This stores pthread_key information across fork() boundaries */ List pthread_key::keys; /* non-static members */ pthread_key::pthread_key (void (*aDestructor) (void *)):verifyable_object (PTHREAD_KEY_MAGIC), destructor (aDestructor) { tls_index = TlsAlloc (); if (tls_index == TLS_OUT_OF_INDEXES) magic = 0; else keys.insert (this); } pthread_key::~pthread_key () { /* We may need to make the list code lock the list during operations */ if (magic != 0) { keys.remove (this); TlsFree (tls_index); } } void pthread_key::_fixup_before_fork () { fork_buf = get (); } void pthread_key::_fixup_after_fork () { tls_index = TlsAlloc (); if (tls_index == TLS_OUT_OF_INDEXES) api_fatal ("pthread_key::recreate_key_from_buffer () failed to reallocate Tls storage"); set (fork_buf); } bool pthread_key::iterate_dtors_once_more; void pthread_key::run_destructor () { if (destructor) { void *oldValue = get (); if (oldValue) { set (NULL); destructor (oldValue); if (get ()) iterate_dtors_once_more = true; } } } /* pshared mutexs */ /* static members */ List pthread_mutex::mutexes; /* This is used for mutex creation protection within a single process only */ fast_mutex NO_COPY pthread_mutex::mutex_initialization_lock; void pthread_mutex::init_mutex () { if (!mutex_initialization_lock.init ()) api_fatal ("Could not create win32 Mutex for pthread mutex static initializer support."); } pthread_mutex::pthread_mutex (pthread_mutexattr *attr) : verifyable_object (0), /* set magic to zero initially */ lock_counter (0), win32_obj_id (NULL), owner (_new_mutex), #ifdef DEBUGGING tid (0), #endif recursion_counter (0), condwaits (0), type (PTHREAD_MUTEX_NORMAL), pshared (PTHREAD_PROCESS_PRIVATE) { win32_obj_id = ::CreateEvent (&sec_none_nih, false, false, NULL); if (!win32_obj_id) return; /*attr checked in the C call */ if (!attr) /* handled in the caller */; else if (attr->pshared != PTHREAD_PROCESS_SHARED) type = attr->mutextype; else return; /* Not implemented */ magic = PTHREAD_MUTEX_MAGIC; mutexes.insert (this); } pthread_mutex::~pthread_mutex () { if (win32_obj_id) { CloseHandle (win32_obj_id); win32_obj_id = NULL; } mutexes.remove (this); owner = _destroyed_mutex; magic = 0; } int pthread_mutex::lock (PLARGE_INTEGER timeout) { pthread_t self = ::pthread_self (); int result = 0; if (InterlockedIncrement (&lock_counter) == 1) set_owner (self); else if (type == PTHREAD_MUTEX_NORMAL /* potentially causes deadlock */ || !pthread::equal (owner, self)) { if (cygwait (win32_obj_id, timeout, cw_sig | cw_sig_restart) != WAIT_TIMEOUT) set_owner (self); else { InterlockedDecrement (&lock_counter); result = ETIMEDOUT; } } else { InterlockedDecrement (&lock_counter); if (type == PTHREAD_MUTEX_RECURSIVE) result = lock_recursive (); else result = EDEADLK; } pthread_printf ("mutex %p, self %p, owner %p, lock_counter %d, recursion_counter %u", this, self, owner, lock_counter, recursion_counter); return result; } int pthread_mutex::unlock () { int res = 0; pthread_t self = ::pthread_self (); if (type == PTHREAD_MUTEX_NORMAL) /* no error checking */; else if (no_owner ()) res = type == PTHREAD_MUTEX_ERRORCHECK ? EPERM : 0; else if (!pthread::equal (owner, self)) res = EPERM; if (!res && recursion_counter > 0 && --recursion_counter == 0) /* Don't try to unlock anything if recursion_counter == 0. This means the mutex was never locked or that we've forked. */ { owner = (pthread_t) _unlocked_mutex; #ifdef DEBUGGING tid = 0; // thread-id #endif if (InterlockedDecrement (&lock_counter)) ::SetEvent (win32_obj_id); // Another thread is waiting res = 0; } pthread_printf ("mutex %p, owner %p, self %p, lock_counter %d, recursion_counter %u, type %d, res %d", this, owner, self, lock_counter, recursion_counter, type, res); return res; } int pthread_mutex::trylock () { pthread_t self = ::pthread_self (); int result = 0; if (InterlockedCompareExchange (&lock_counter, 1, 0) == 0) set_owner (self); else if (type == PTHREAD_MUTEX_RECURSIVE && pthread::equal (owner, self)) result = lock_recursive (); else result = EBUSY; return result; } int pthread_mutex::destroy () { if (condwaits || trylock ()) // Do not destroy a condwaited or locked mutex return EBUSY; else if (recursion_counter > 1) { // Do not destroy a recursive locked mutex recursion_counter--; return EBUSY; } delete this; return 0; } void pthread_mutex::_fixup_after_fork () { pthread_printf ("mutex %p", this); if (pshared != PTHREAD_PROCESS_PRIVATE) api_fatal ("pthread_mutex::_fixup_after_fork () doesn't understand PROCESS_SHARED mutex's"); /* All waiting threads are gone after a fork */ recursion_counter = 0; lock_counter = 0; condwaits = 0; #ifdef DEBUGGING tid = 0xffffffff; /* Don't know the tid after a fork */ #endif win32_obj_id = ::CreateEvent (&sec_none_nih, false, false, NULL); if (!win32_obj_id) api_fatal ("pthread_mutex::_fixup_after_fork () failed to recreate win32 event for mutex"); } pthread_mutexattr::pthread_mutexattr ():verifyable_object (PTHREAD_MUTEXATTR_MAGIC), pshared (PTHREAD_PROCESS_PRIVATE), mutextype (PTHREAD_MUTEX_NORMAL) { } pthread_mutexattr::~pthread_mutexattr () { } /* pshared spinlocks The infrastructure is provided by the underlying pthread_mutex class. The rest is a simplification implementing spin locking. */ pthread_spinlock::pthread_spinlock (int pshared) : pthread_mutex (NULL) { magic = PTHREAD_SPINLOCK_MAGIC; set_type (PTHREAD_MUTEX_NORMAL); set_shared (pshared); } int pthread_spinlock::lock () { pthread_t self = ::pthread_self (); int result = -1; unsigned spins = 0; /* We want to spin using 'pause' instruction on multi-core system but we want to avoid this on single-core systems. The limit of 1000 spins is semi-arbitrary. Microsoft suggests (in their InitializeCriticalSectionAndSpinCount documentation on MSDN) they are using spin count limit 4000 for their heap manager critical sections. Other source suggest spin count as small as 200 for fast path of mutex locking. */ unsigned const FAST_SPINS_LIMIT = wincap.cpu_count () != 1 ? 1000 : 0; do { if (InterlockedExchange (&lock_counter, 1) == 0) { set_owner (self); result = 0; } else if (unlikely(pthread::equal (owner, self))) result = EDEADLK; else if (spins < FAST_SPINS_LIMIT) { ++spins; __asm__ volatile ("pause":::); } else { /* Minimal timeout to minimize CPU usage while still spinning. */ LARGE_INTEGER timeout; timeout.QuadPart = -10000LL; /* FIXME: no cancel? */ cygwait (win32_obj_id, &timeout, cw_sig); } } while (result == -1); pthread_printf ("spinlock %p, self %p, owner %p", this, self, owner); return result; } int pthread_spinlock::unlock () { pthread_t self = ::pthread_self (); int result = 0; if (!pthread::equal (owner, self)) result = EPERM; else { owner = (pthread_t) _unlocked_mutex; #ifdef DEBUGGING tid = 0; // thread-id #endif InterlockedExchange (&lock_counter, 0); ::SetEvent (win32_obj_id); result = 0; } pthread_printf ("spinlock %p, owner %p, self %p, res %d", this, owner, self, result); return result; } DWORD pthread::thread_init_wrapper (void *arg) { exception protect; pthread *thread = (pthread *) arg; /* This *must* be set prior to calling set_tls_self_pointer or there is a race with the signal processing code which may miss the signal mask settings. */ _my_tls.sigmask = thread->parent_sigmask; thread->set_tls_self_pointer (); // Give thread default name SetThreadName (GetCurrentThreadId (), program_invocation_short_name); thread->mutex.lock (); // if thread is detached force cleanup on exit if (thread->attr.joinable == PTHREAD_CREATE_DETACHED && thread->joiner == NULL) thread->joiner = thread; thread->mutex.unlock (); debug_printf ("tid %p", &_my_tls); thread_printf ("started thread %p %p %p %p %p %p", arg, &_my_tls.local_clib, _impure_ptr, thread, thread->function, thread->arg); // call the user's thread void *ret = thread->function (thread->arg); thread->exit (ret); return 0; // just for show. Never returns. } unsigned long pthread::getsequence_np () { return get_thread_id (); } int pthread::create (pthread_t *thread, const pthread_attr_t *attr, void *(*start_routine) (void *), void *arg) { if (attr && !pthread_attr::is_good_object (attr)) return EINVAL; *thread = new pthread (); if (!(*thread)->create (start_routine, attr ? *attr : NULL, arg)) { delete (*thread); *thread = NULL; return EAGAIN; } return 0; } int pthread::once (pthread_once_t *once_control, void (*init_routine) (void)) { /* Sign bit of once_control->state is used as done flag. Similarly, the next significant bit is used as destroyed flag. */ const int32_t DONE = 0x80000000; const int32_t DESTROYED = 0x40000000; static_assert (sizeof (once_control->state) >= sizeof (int32_t)); static_assert (sizeof (once_control->state) == sizeof (LONG)); if (once_control->state & DONE) return 0; /* The type of &once_control->state is int *, which is compatible with LONG * (that is the type of the pointer argument of InterlockedXxx()). */ if ((InterlockedIncrement (&once_control->state) & DONE) == 0) { pthread_mutex_lock (&once_control->mutex); if (!(once_control->state & DONE)) { init_routine (); InterlockedOr (&once_control->state, DONE); } pthread_mutex_unlock (&once_control->mutex); } InterlockedDecrement (&once_control->state); if (InterlockedCompareExchange (&once_control->state, DONE | DESTROYED, DONE) == DONE) pthread_mutex_destroy (&once_control->mutex); return 0; } int pthread::cancel (pthread_t thread) { if (!is_good_object (&thread)) return ESRCH; return thread->cancel (); } void pthread::atforkprepare () { callback *cb = MT_INTERFACE->pthread_prepare; while (cb) { cb->cb (); cb = cb->next; } __fp_lock_all (); MT_INTERFACE->fixup_before_fork (); } void pthread::atforkparent () { __fp_unlock_all (); callback *cb = MT_INTERFACE->pthread_parent; while (cb) { cb->cb (); cb = cb->next; } } void pthread::atforkchild () { MT_INTERFACE->fixup_after_fork (); __fp_unlock_all (); callback *cb = MT_INTERFACE->pthread_child; while (cb) { cb->cb (); cb = cb->next; } } /* Register a set of functions to run before and after fork. prepare calls are called in LI-FC order. parent and child calls are called in FI-FC order. */ int pthread::atfork (void (*prepare)(void), void (*parent)(void), void (*child)(void)) { callback *prepcb = NULL, *parentcb = NULL, *childcb = NULL; if (prepare) { prepcb = new callback; if (!prepcb) return ENOMEM; } if (parent) { parentcb = new callback; if (!parentcb) { if (prepcb) delete prepcb; return ENOMEM; } } if (child) { childcb = new callback; if (!childcb) { if (prepcb) delete prepcb; if (parentcb) delete parentcb; return ENOMEM; } } if (prepcb) { prepcb->cb = prepare; List_insert_nolock (MT_INTERFACE->pthread_prepare, prepcb); } if (parentcb) { parentcb->cb = parent; callback **t = &MT_INTERFACE->pthread_parent; while (*t) t = &(*t)->next; /* t = pointer to last next in the list */ List_insert_nolock (*t, parentcb); } if (childcb) { childcb->cb = child; callback **t = &MT_INTERFACE->pthread_child; while (*t) t = &(*t)->next; /* t = pointer to last next in the list */ List_insert_nolock (*t, childcb); } return 0; } int pthread::join (pthread_t *thread, void **return_val, PLARGE_INTEGER timeout) { pthread_t joiner = self (); joiner->testcancel (); // Initialize return val with NULL if (return_val) *return_val = NULL; if (!is_good_object (&joiner)) return EINVAL; if (!is_good_object (thread)) return ESRCH; if (equal (*thread,joiner)) return EDEADLK; (*thread)->mutex.lock (); if ((*thread)->attr.joinable == PTHREAD_CREATE_DETACHED) { (*thread)->mutex.unlock (); return EINVAL; } else { (*thread)->joiner = joiner; (*thread)->attr.joinable = PTHREAD_CREATE_DETACHED; (*thread)->mutex.unlock (); switch (cygwait ((*thread)->win32_obj_id, timeout, cw_sig | cw_sig_restart | cw_cancel)) { case WAIT_OBJECT_0: if (return_val) *return_val = (*thread)->return_ptr; delete (*thread); break; case WAIT_CANCELED: // set joined thread back to joinable since we got canceled (*thread)->joiner = NULL; (*thread)->attr.joinable = PTHREAD_CREATE_JOINABLE; joiner->cancel_self (); // never reached break; case WAIT_TIMEOUT: // set joined thread back to joinable since we got canceled (*thread)->joiner = NULL; (*thread)->attr.joinable = PTHREAD_CREATE_JOINABLE; return (timeout && timeout->QuadPart == 0LL) ? EBUSY : ETIMEDOUT; default: // should never happen return EINVAL; } } return 0; } int pthread::detach (pthread_t *thread) { if (!is_good_object (thread)) return ESRCH; (*thread)->mutex.lock (); if ((*thread)->attr.joinable == PTHREAD_CREATE_DETACHED) { (*thread)->mutex.unlock (); return EINVAL; } // check if thread is still alive if ((*thread)->valid && WaitForSingleObject ((*thread)->win32_obj_id, 0) == WAIT_TIMEOUT) { // force cleanup on exit (*thread)->joiner = *thread; (*thread)->attr.joinable = PTHREAD_CREATE_DETACHED; (*thread)->mutex.unlock (); } else { // thread has already terminated. (*thread)->mutex.unlock (); delete (*thread); } return 0; } int pthread::suspend (pthread_t *thread) { if (!is_good_object (thread)) return ESRCH; if ((*thread)->suspended == false) { (*thread)->suspended = true; SuspendThread ((*thread)->win32_obj_id); } return 0; } int pthread::resume (pthread_t *thread) { if (!is_good_object (thread)) return ESRCH; if ((*thread)->suspended == true) ResumeThread ((*thread)->win32_obj_id); (*thread)->suspended = false; return 0; } static inline int pthread_convert_abstime (clockid_t clock_id, const struct timespec *abstime, PLARGE_INTEGER timeout) { struct timespec tp; /* According to SUSv3, the abstime value must be checked for validity. */ if (!valid_timespec (*abstime)) return EINVAL; /* Check for immediate timeout before converting */ clock_gettime (clock_id, &tp); if (tp.tv_sec > abstime->tv_sec || (tp.tv_sec == abstime->tv_sec && tp.tv_nsec > abstime->tv_nsec)) return ETIMEDOUT; timeout->QuadPart = abstime->tv_sec * NS100PERSEC + (abstime->tv_nsec + (NSPERSEC/NS100PERSEC) - 1) / (NSPERSEC/NS100PERSEC); switch (clock_id) { case CLOCK_REALTIME_COARSE: case CLOCK_REALTIME: timeout->QuadPart += FACTOR; break; default: /* other clocks must be handled as relative timeout */ timeout->QuadPart -= tp.tv_sec * NS100PERSEC + tp.tv_nsec / (NSPERSEC/NS100PERSEC); timeout->QuadPart *= -1LL; break; } return 0; } int pthread_cond::init (pthread_cond_t *cond, const pthread_condattr_t *attr) { pthread_cond_t new_cond; if (attr && !pthread_condattr::is_good_object (attr)) return EINVAL; cond_initialization_lock.lock (); new_cond = new pthread_cond (attr ? (*attr) : NULL); if (!is_good_object (&new_cond)) { delete new_cond; cond_initialization_lock.unlock (); return EAGAIN; } int ret = 0; __try { *cond = new_cond; } __except (NO_ERROR) { delete new_cond; ret = EINVAL; } __endtry cond_initialization_lock.unlock (); return ret; } int pthread_rwlock::init (pthread_rwlock_t *rwlock, const pthread_rwlockattr_t *attr) { pthread_rwlock_t new_rwlock; if (attr && !pthread_rwlockattr::is_good_object (attr)) return EINVAL; rwlock_initialization_lock.lock (); new_rwlock = new pthread_rwlock (attr ? (*attr) : NULL); if (!is_good_object (&new_rwlock)) { delete new_rwlock; rwlock_initialization_lock.unlock (); return EAGAIN; } int ret = 0; __try { *rwlock = new_rwlock; } __except (NO_ERROR) { delete new_rwlock; ret = EINVAL; } __endtry rwlock_initialization_lock.unlock (); return ret; } /* Mutexes */ int pthread_mutex::init (pthread_mutex_t *mutex, const pthread_mutexattr_t *attr, const pthread_mutex_t initializer) { if (attr && !pthread_mutexattr::is_good_object (attr)) return EINVAL; mutex_initialization_lock.lock (); if (initializer == NULL || pthread_mutex::is_initializer (mutex)) { pthread_mutex_t new_mutex = new pthread_mutex (attr ? (*attr) : NULL); if (!is_good_object (&new_mutex)) { delete new_mutex; mutex_initialization_lock.unlock (); return EAGAIN; } if (!attr && initializer) { if (initializer == PTHREAD_RECURSIVE_MUTEX_INITIALIZER_NP) new_mutex->type = PTHREAD_MUTEX_RECURSIVE; else if (initializer == PTHREAD_NORMAL_MUTEX_INITIALIZER_NP) new_mutex->type = PTHREAD_MUTEX_NORMAL; else if (initializer == PTHREAD_ERRORCHECK_MUTEX_INITIALIZER_NP) new_mutex->type = PTHREAD_MUTEX_ERRORCHECK; } __try { *mutex = new_mutex; } __except (NO_ERROR) { delete new_mutex; mutex_initialization_lock.unlock (); return EINVAL; } __endtry } mutex_initialization_lock.unlock (); pthread_printf ("*mutex %p, attr %p, initializer %p", *mutex, attr, initializer); return 0; } /* Spinlocks */ int pthread_spinlock::init (pthread_spinlock_t *spinlock, int pshared) { pthread_spinlock_t new_spinlock = new pthread_spinlock (pshared); if (!is_good_object (&new_spinlock)) { delete new_spinlock; return EAGAIN; } __try { *spinlock = new_spinlock; } __except (NO_ERROR) { delete new_spinlock; return EINVAL; } __endtry pthread_printf ("*spinlock %p, pshared %d", *spinlock, pshared); return 0; } /* Semaphores */ List semaphore::semaphores; semaphore::semaphore (int pshared, unsigned int value) : verifyable_object (SEM_MAGIC), shared (pshared), currentvalue (-1), startvalue (value), fd (-1), hash (0ULL), sem (NULL) { SECURITY_ATTRIBUTES sa = (pshared != PTHREAD_PROCESS_PRIVATE) ? sec_all : sec_none_nih; this->win32_obj_id = ::CreateSemaphore (&sa, value, INT32_MAX, NULL); if (!this->win32_obj_id) magic = 0; semaphores.insert (this); } semaphore::semaphore (unsigned long long shash, LUID sluid, int sfd, sem_t *ssem, int oflag, mode_t mode, unsigned int value) : verifyable_object (SEM_MAGIC), shared (PTHREAD_PROCESS_SHARED), currentvalue (-1), /* Unused for named semaphores. */ startvalue (value), fd (sfd), hash (shash), luid (sluid), sem (ssem) { char name[MAX_PATH]; __small_sprintf (name, "semaphore/%016X%08x%08x", hash, luid.HighPart, luid.LowPart); this->win32_obj_id = ::CreateSemaphore (&sec_all, value, INT32_MAX, name); if (!this->win32_obj_id) magic = 0; if (GetLastError () == ERROR_ALREADY_EXISTS && (oflag & O_EXCL)) { __seterrno (); CloseHandle (this->win32_obj_id); magic = 0; } semaphores.insert (this); } semaphore::~semaphore () { if (win32_obj_id) CloseHandle (win32_obj_id); semaphores.remove (this); } void semaphore::_post () { LONG dummy; ReleaseSemaphore (win32_obj_id, 1, &dummy); } int semaphore::_getvalue (int *sval) { NTSTATUS status; SEMAPHORE_BASIC_INFORMATION sbi; status = NtQuerySemaphore (win32_obj_id, SemaphoreBasicInformation, &sbi, sizeof sbi, NULL); int res; if (NT_SUCCESS (status)) { *sval = sbi.CurrentCount; res = 0; } else { *sval = startvalue; __seterrno_from_nt_status (status); res = -1; } return res; } int semaphore::_trywait () { /* FIXME: signals should be able to interrupt semaphores... We probably need WaitForMultipleObjects here. */ if (WaitForSingleObject (win32_obj_id, 0) == WAIT_TIMEOUT) { set_errno (EAGAIN); return -1; } return 0; } int semaphore::_wait (PLARGE_INTEGER timeout) { __try { switch (cygwait (win32_obj_id, timeout, cw_cancel | cw_cancel_self | cw_sig_eintr)) { case WAIT_OBJECT_0: break; case WAIT_SIGNALED: set_errno (EINTR); return -1; case WAIT_TIMEOUT: set_errno (ETIMEDOUT); return -1; default: pthread_printf ("cygwait failed. %E"); __seterrno (); return -1; } } __except (NO_ERROR) {} __endtry return 0; } void semaphore::_fixup_before_fork () { NTSTATUS status; SEMAPHORE_BASIC_INFORMATION sbi; status = NtQuerySemaphore (win32_obj_id, SemaphoreBasicInformation, &sbi, sizeof sbi, NULL); if (NT_SUCCESS (status)) currentvalue = sbi.CurrentCount; else currentvalue = startvalue; } void semaphore::_fixup_after_fork () { if (shared == PTHREAD_PROCESS_PRIVATE) { pthread_printf ("sem %p", this); win32_obj_id = ::CreateSemaphore (&sec_none_nih, currentvalue, INT32_MAX, NULL); if (!win32_obj_id) api_fatal ("failed to create new win32 semaphore, " "currentvalue %ld, %E", currentvalue); } } void semaphore::_terminate () { int _sem_close (sem_t *, bool); if (sem) _sem_close (sem, false); } /* static members */ int semaphore::init (sem_t *sem, int pshared, unsigned int value) { /* We can't tell the difference between reinitialising an existing semaphore and initialising a semaphore who's contents happen to be a valid pointer */ if (is_good_object (sem)) paranoid_printf ("potential attempt to reinitialise a semaphore"); if (value > SEM_VALUE_MAX) { set_errno(EINVAL); return -1; } *sem = new semaphore (pshared, value); if (!is_good_object (sem)) { delete (*sem); *sem = NULL; set_errno(EAGAIN); return -1; } return 0; } int semaphore::destroy (sem_t *sem) { if (!is_good_object (sem)) { set_errno(EINVAL); return -1; } /* It's invalid to destroy a semaphore not opened with sem_init. */ if ((*sem)->fd != -1) { set_errno(EINVAL); return -1; } /* FIXME - new feature - test for busy against threads... */ delete (*sem); *sem = NULL; return 0; } int semaphore::close (sem_t *sem) { if (!is_good_object (sem)) { set_errno(EINVAL); return -1; } /* It's invalid to close a semaphore not opened with sem_open. */ if ((*sem)->fd == -1) { set_errno(EINVAL); return -1; } delete (*sem); delete sem; return 0; } sem_t * semaphore::open (unsigned long long hash, LUID luid, int fd, int oflag, mode_t mode, unsigned int value, bool &wasopen) { if (value > SEM_VALUE_MAX) { set_errno (EINVAL); return NULL; } /* sem_open is supposed to return the same pointer, if the same named semaphore is opened multiple times in the same process, as long as the semaphore hasn't been closed or unlinked in the meantime. */ semaphores.mx.lock (); for (semaphore *sema = semaphores.head; sema; sema = sema->next) if (sema->fd >= 0 && sema->hash == hash && sema->luid.HighPart == luid.HighPart && sema->luid.LowPart == luid.LowPart) { wasopen = true; semaphores.mx.unlock (); return sema->sem; } semaphores.mx.unlock (); wasopen = false; sem_t *sem = new sem_t; if (!sem) { set_errno (ENOMEM); return NULL; } *sem = new semaphore (hash, luid, fd, sem, oflag, mode, value); if (!is_good_object (sem)) { delete *sem; delete sem; return NULL; } return sem; } int semaphore::wait (sem_t *sem) { pthread_testcancel (); if (!is_good_object (sem)) { set_errno (EINVAL); return -1; } return (*sem)->_wait (); } int semaphore::trywait (sem_t *sem) { if (!is_good_object (sem)) { set_errno (EINVAL); return -1; } return (*sem)->_trywait (); } int semaphore::clockwait (sem_t *sem, clockid_t clock_id, const struct timespec *abstime) { LARGE_INTEGER timeout; if (!is_good_object (sem)) { set_errno (EINVAL); return -1; } /* According to SUSv3, abstime need not be checked for validity, if the semaphore can be locked immediately. */ if (!(*sem)->_trywait ()) return 0; __try { int err = pthread_convert_abstime (clock_id, abstime, &timeout); if (err) return err; return (*sem)->_wait (&timeout); } __except (NO_ERROR) {} __endtry return EINVAL; } int semaphore::post (sem_t *sem) { if (!is_good_object (sem)) { set_errno (EINVAL); return -1; } (*sem)->_post (); return 0; } int semaphore::getvalue (sem_t *sem, int *sval) { __try { if (is_good_object (sem)) return (*sem)->_getvalue (sval); } __except (NO_ERROR) {} __endtry set_errno (EINVAL); return -1; } int semaphore::getinternal (sem_t *sem, int *sfd, unsigned long long *shash, LUID *sluid, unsigned int *sval) { __try { if (!is_good_object (sem)) __leave; if ((*sfd = (*sem)->fd) < 0) __leave; *shash = (*sem)->hash; *sluid = (*sem)->luid; /* POSIX defines the value in calls to sem_init/sem_open as unsigned, but the sem_getvalue gets a pointer to int to return the value. Go figure! */ return (*sem)->_getvalue ((int *)sval); } __except (NO_ERROR) {} __endtry set_errno (EINVAL); return -1; } /* pthread_null */ pthread * pthread_null::get_null_pthread () { /* because of weird entry points */ _instance.magic = 0; return &_instance; } pthread_null::pthread_null () { attr.joinable = PTHREAD_CREATE_DETACHED; /* Mark ourselves as invalid */ magic = 0; } pthread_null::~pthread_null () { } bool pthread_null::create (void *(*)(void *), pthread_attr *, void *) { return true; } void pthread_null::exit (void *value_ptr) { _my_tls.remove (INFINITE); ExitThread (0); } int pthread_null::cancel () { return 0; } void pthread_null::testcancel () { } int pthread_null::setcancelstate (int state, int *oldstate) { return EINVAL; } int pthread_null::setcanceltype (int type, int *oldtype) { return EINVAL; } void pthread_null::push_cleanup_handler (__pthread_cleanup_handler *handler) { } void pthread_null::pop_cleanup_handler (int const execute) { } unsigned long pthread_null::getsequence_np () { return 0; } pthread_null pthread_null::_instance; int pthread_barrier::init (const pthread_barrierattr_t * attr, unsigned count) { pthread_mutex_t * mutex = NULL; if (unlikely ((attr != NULL && (! pthread_barrierattr::is_good_object (attr) || (*attr)->shared == PTHREAD_PROCESS_SHARED)) || count == 0)) return EINVAL; int retval = pthread_mutex_init (&mtx, NULL); if (unlikely (retval != 0)) return retval; retval = pthread_cond_init (&cond, NULL); if (unlikely (retval != 0)) { int ret = pthread_mutex_destroy (mutex); if (ret != 0) api_fatal ("pthread_mutex_destroy (%p) = %d", mutex, ret); mtx = NULL; return retval; } cnt = count; cyc = 0; wt = 0; return 0; } int pthread_barrier::destroy () { if (unlikely (wt != 0)) return EBUSY; int retval = pthread_cond_destroy (&cond); if (unlikely (retval != 0)) return retval; else cond = NULL; retval = pthread_mutex_destroy (&mtx); if (unlikely (retval != 0)) return retval; else mtx = NULL; cnt = 0; cyc = 0; wt = 0; return 0; } int pthread_barrier::wait () { int retval = pthread_mutex_lock (&mtx); if (unlikely (retval != 0)) return retval; if (unlikely (wt >= cnt)) { api_fatal ("wt >= cnt (%u >= %u)", wt, cnt); return EINVAL; } if (unlikely (++wt == cnt)) { ++cyc; /* This is the last thread to reach the barrier. Signal the waiting threads to wake up and continue. */ retval = pthread_cond_broadcast (&cond); if (unlikely (retval != 0)) goto cond_error; wt = 0; retval = pthread_mutex_unlock (&mtx); if (unlikely (retval != 0)) abort (); return PTHREAD_BARRIER_SERIAL_THREAD; } else { uint64_t cycle = cyc; do { retval = pthread_cond_wait (&cond, &mtx); if (unlikely (retval != 0)) goto cond_error; } while (unlikely (cycle == cyc)); retval = pthread_mutex_unlock (&mtx); if (unlikely (retval != 0)) api_fatal ("pthread_mutex_unlock (%p) = %d", &mtx, retval); return 0; } cond_error: { --wt; int ret = pthread_mutex_unlock (&mtx); if (unlikely (ret != 0)) api_fatal ("pthread_mutex_unlock (%p) = %d", &mtx, ret); return retval; } } /* Returns running thread's name; works for both cygthreads and pthreads */ char * mythreadname (void) { char *result = (char *) cygthread::name (); if (result == _my_tls.locals.unknown_thread_name) { result[0] = '\0'; pthread_getname_np (pthread_self (), result, (size_t) THRNAMELEN); } return result; } extern "C" { /* Thread creation */ int pthread_create (pthread_t *thread, const pthread_attr_t *attr, void *(*start_routine) (void *), void *arg) { return pthread::create (thread, attr, start_routine, arg); } int pthread_once (pthread_once_t * once_control, void (*init_routine) (void)) { return pthread::once (once_control, init_routine); } int pthread_atfork (void (*prepare)(void), void (*parent)(void), void (*child)(void)) { return pthread::atfork (prepare, parent, child); } /* ID */ pthread_t pthread_self () { return pthread::self (); } int pthread_equal (pthread_t t1, pthread_t t2) { return pthread::equal (t1, t2); } unsigned long pthread_getsequence_np (pthread_t * thread) { if (!pthread::is_good_object (thread)) return EINVAL; return (*thread)->getsequence_np (); } /* Thread name */ int pthread_getname_np (pthread_t thread, char *buf, size_t buflen) { char *name; if (!pthread::is_good_object (&thread)) return ESRCH; if (!thread->attr.name) name = program_invocation_short_name; else name = thread->attr.name; /* Return ERANGE if the provided buffer is less than THRNAMELEN. Truncate and zero-terminate the name to fit in buf. This means we always return something if the buffer is THRNAMELEN or larger, but there is no way to tell if we have the whole name. */ if (buflen < THRNAMELEN) return ERANGE; int ret = 0; __try { strlcpy (buf, name, buflen); } __except (NO_ERROR) { ret = EFAULT; } __endtry return ret; } int pthread_setname_np (pthread_t thread, const char *name) { char *oldname, *cp; if (!pthread::is_good_object (&thread)) return ESRCH; if (strlen (name) > THRNAMELEN) return ERANGE; cp = strdup (name); if (!cp) return ENOMEM; oldname = thread->attr.name; thread->attr.name = cp; SetThreadName (GetThreadId (thread->win32_obj_id), thread->attr.name); if (oldname) free (oldname); return 0; } /* Thread exit */ void pthread_exit (void *value_ptr) { pthread::self ()->exit (value_ptr); __builtin_unreachable (); /* FIXME: don't know why this is necessary */ } int pthread_detach (pthread_t thread) { return pthread::detach (&thread); } int pthread_join (pthread_t thread, void **return_val) { return pthread::join (&thread, (void **) return_val, NULL); } int pthread_tryjoin_np (pthread_t thread, void **return_val) { LARGE_INTEGER timeout = { QuadPart:0LL }; return pthread::join (&thread, (void **) return_val, &timeout); } int pthread_timedjoin_np (pthread_t thread, void **return_val, const struct timespec *abstime) { LARGE_INTEGER timeout; int err = pthread_convert_abstime (CLOCK_REALTIME, abstime, &timeout); if (err) return err; return pthread::join (&thread, (void **) return_val, &timeout); } /* Thread suspend/resume */ /* This isn't a posix call... should we keep it? */ int pthread_suspend (pthread_t thread) { return pthread::suspend (&thread); } /* same */ int pthread_continue (pthread_t thread) { return pthread::resume (&thread); } /* Thread signal */ int pthread_kill (pthread_t thread, int sig) { // lock myself, for the use of thread2signal // two different kills might clash: FIXME if (!pthread::is_good_object (&thread)) return EINVAL; /* check that sig is in right range */ if (sig < 0 || sig >= _NSIG) return EINVAL; siginfo_t si = {0}; si.si_signo = sig; si.si_code = SI_USER; si.si_pid = myself->pid; si.si_uid = myself->uid; int rval; if (!thread->valid) rval = ESRCH; else if (sig) { rval = (int) sig_send (NULL, si, thread->cygtls); if (rval == -1) rval = get_errno (); } else switch (WaitForSingleObject (thread->win32_obj_id, 0)) { case WAIT_TIMEOUT: rval = 0; break; default: rval = ESRCH; break; } // unlock myself return rval; } int pthread_sigmask (int operation, const sigset_t *set, sigset_t *old_set) { int res = handle_sigprocmask (operation, set, old_set, _my_tls.sigmask); syscall_printf ("%d = pthread_sigmask(%d, %p, %p)", res, operation, set, old_set); return res; } int pthread_sigqueue (pthread_t thread, int sig, const union sigval value) { siginfo_t si = {0}; if (!pthread::is_good_object (&thread)) return EINVAL; if (!thread->valid) return ESRCH; si.si_signo = sig; si.si_code = SI_QUEUE; si.si_value = value; si.si_pid = myself->pid; si.si_uid = myself->uid; return (int) sig_send (NULL, si, thread->cygtls); } /* Cancelability */ int pthread_cancel (pthread_t thread) { return pthread::cancel (thread); } int pthread_setcancelstate (int state, int *oldstate) { return pthread::self ()->setcancelstate (state, oldstate); } int pthread_setcanceltype (int type, int *oldtype) { return pthread::self ()->setcanceltype (type, oldtype); } void pthread_testcancel () { pthread::self ()->testcancel (); } void _pthread_cleanup_push (__pthread_cleanup_handler *handler) { pthread::self ()->push_cleanup_handler (handler); } void _pthread_cleanup_pop (int execute) { pthread::self ()->pop_cleanup_handler (execute); } /* provided for source level compatability. See http://www.opengroup.org/onlinepubs/007908799/xsh/pthread_getconcurrency.html */ int pthread_getconcurrency () { return MT_INTERFACE->concurrency; } /* provided for source level compatability. See http://www.opengroup.org/onlinepubs/007908799/xsh/pthread_getconcurrency.html */ int pthread_setconcurrency (int new_level) { if (new_level < 0) return EINVAL; MT_INTERFACE->concurrency = new_level; return 0; } /* Thread scheduling */ /* keep this in sync with sched.cc */ int pthread_getschedparam (pthread_t thread, int *policy, struct sched_param *param) { if (!pthread::is_good_object (&thread)) return ESRCH; *policy = SCHED_FIFO; param->sched_priority = sched_get_thread_priority (thread->win32_obj_id); return 0; } /* keep this in sync with sched.cc */ int pthread_setschedparam (pthread_t thread, int policy, const struct sched_param *param) { if (!pthread::is_good_object (&thread)) return ESRCH; if (policy != SCHED_FIFO) return ENOTSUP; if (!param) return EINVAL; int rv = sched_set_thread_priority (thread->win32_obj_id, param->sched_priority); if (!rv) thread->attr.schedparam.sched_priority = param->sched_priority; return rv; } int pthread_setschedprio (pthread_t thread, int priority) { if (!pthread::is_good_object (&thread)) return ESRCH; int rv = sched_set_thread_priority (thread->win32_obj_id, priority); if (!rv) thread->attr.schedparam.sched_priority = priority; return rv; } /* Thread affinity */ int pthread_getaffinity_np (pthread_t thread, size_t sizeof_set, cpu_set_t *set) { if (!pthread::is_good_object (&thread)) return ESRCH; return sched_get_thread_affinity (thread->win32_obj_id, sizeof_set, set); } int pthread_setaffinity_np (pthread_t thread, size_t sizeof_set, const cpu_set_t *set) { if (!pthread::is_good_object (&thread)) return ESRCH; return sched_set_thread_affinity (thread->win32_obj_id, sizeof_set, set); } /* pthread_attr */ int pthread_attr_init (pthread_attr_t *attr) { *attr = new pthread_attr; if (!pthread_attr::is_good_object (attr)) { delete (*attr); *attr = NULL; return ENOMEM; } return 0; } int pthread_attr_getinheritsched (const pthread_attr_t *attr, int *inheritsched) { if (!pthread_attr::is_good_object (attr)) return EINVAL; *inheritsched = (*attr)->inheritsched; return 0; } int pthread_attr_getschedparam (const pthread_attr_t *attr, struct sched_param *param) { if (!pthread_attr::is_good_object (attr)) return EINVAL; *param = (*attr)->schedparam; return 0; } /* From a pure code point of view, this should call a helper in sched.cc, to allow for someone adding scheduler policy changes to win32 in the future. However that's extremely unlikely, so short and sweet will do us */ int pthread_attr_getschedpolicy (const pthread_attr_t *attr, int *policy) { if (!pthread_attr::is_good_object (attr)) return EINVAL; *policy = SCHED_FIFO; return 0; } int pthread_attr_getscope (const pthread_attr_t *attr, int *contentionscope) { if (!pthread_attr::is_good_object (attr)) return EINVAL; *contentionscope = (*attr)->contentionscope; return 0; } int pthread_attr_setdetachstate (pthread_attr_t *attr, int detachstate) { if (!pthread_attr::is_good_object (attr)) return EINVAL; if (detachstate < 0 || detachstate > 1) return EINVAL; (*attr)->joinable = detachstate; return 0; } int pthread_attr_getdetachstate (const pthread_attr_t *attr, int *detachstate) { if (!pthread_attr::is_good_object (attr)) return EINVAL; *detachstate = (*attr)->joinable; return 0; } int pthread_attr_setinheritsched (pthread_attr_t *attr, int inheritsched) { if (!pthread_attr::is_good_object (attr)) return EINVAL; if (inheritsched != PTHREAD_INHERIT_SCHED && inheritsched != PTHREAD_EXPLICIT_SCHED) return ENOTSUP; (*attr)->inheritsched = inheritsched; return 0; } int pthread_attr_setschedparam (pthread_attr_t *attr, const struct sched_param *param) { if (!pthread_attr::is_good_object (attr)) return EINVAL; if (!valid_sched_parameters (param)) return ENOTSUP; (*attr)->schedparam = *param; return 0; } /* See __pthread_attr_getschedpolicy for some notes */ int pthread_attr_setschedpolicy (pthread_attr_t *attr, int policy) { if (!pthread_attr::is_good_object (attr)) return EINVAL; if (policy != SCHED_FIFO) return ENOTSUP; return 0; } int pthread_attr_setscope (pthread_attr_t *attr, int contentionscope) { if (!pthread_attr::is_good_object (attr)) return EINVAL; if (contentionscope != PTHREAD_SCOPE_SYSTEM && contentionscope != PTHREAD_SCOPE_PROCESS) return EINVAL; /* In future, we may be able to support system scope by escalating the thread priority to exceed the priority class. For now we only support PROCESS scope. */ if (contentionscope != PTHREAD_SCOPE_PROCESS) return ENOTSUP; (*attr)->contentionscope = contentionscope; return 0; } int pthread_attr_setstack (pthread_attr_t *attr, void *addr, size_t size) { if (!pthread_attr::is_good_object (attr)) return EINVAL; if (addr == NULL) return EINVAL; if (size < PTHREAD_STACK_MIN) return EINVAL; /* The incoming address addr points to the lowest addressable byte of a buffer of size bytes. Due to the way pthread_attr_setstackaddr is defined on Linux, the lowest address ot the stack can't be reliably computed when using pthread_attr_setstackaddr/pthread_attr_setstacksize. Therefore we store the uppermost address of the stack in stackaddr. See also the comment in pthread_attr_setstackaddr. */ (*attr)->stackaddr = (caddr_t) addr + size; (*attr)->stacksize = size; return 0; } int pthread_attr_getstack (const pthread_attr_t *attr, void **addr, size_t *size) { if (!pthread_attr::is_good_object (attr)) return EINVAL; /* stackaddr holds the uppermost stack address. See the comment in pthread_attr_setstack. */ *addr = (caddr_t) (*attr)->stackaddr - (*attr)->stacksize; *size = (*attr)->stacksize; return 0; } int pthread_attr_setstackaddr (pthread_attr_t *attr, void *addr) { if (!pthread_attr::is_good_object (attr)) return EINVAL; if (addr == NULL) return EINVAL; /* This function is deprecated in SUSv4, but SUSv3 didn't define if the incoming stack address is the lowest address of the memory area defined as stack, or if it's the start address of the stack at which it begins its growth. On Linux it's the latter which means the uppermost stack address on x86 based systems. See comment in pthread_attr_setstack as well. */ (*attr)->stackaddr = addr; return 0; } int pthread_attr_getstackaddr (const pthread_attr_t *attr, void **addr) { if (!pthread_attr::is_good_object (attr)) return EINVAL; /* See comment in pthread_attr_setstackaddr. */ *addr = (*attr)->stackaddr; return 0; } int pthread_attr_setstacksize (pthread_attr_t *attr, size_t size) { if (!pthread_attr::is_good_object (attr)) return EINVAL; if (size < PTHREAD_STACK_MIN) return EINVAL; (*attr)->stacksize = size; return 0; } int pthread_attr_getstacksize (const pthread_attr_t *attr, size_t *size) { if (!pthread_attr::is_good_object (attr)) return EINVAL; /* If the stacksize has not been set by the application, return the default stacksize. Note that this is different from what pthread_attr_getstack returns. */ *size = (*attr)->stacksize ?: get_rlimit_stack (); return 0; } int pthread_attr_setguardsize (pthread_attr_t *attr, size_t size) { if (!pthread_attr::is_good_object (attr)) return EINVAL; /* We don't support a guardsize of more than 1 Meg. */ if (size > 1024 * 1024) return EINVAL; (*attr)->guardsize = size; return 0; } int pthread_attr_getguardsize (const pthread_attr_t *attr, size_t *size) { if (!pthread_attr::is_good_object (attr)) return EINVAL; *size = (*attr)->guardsize; return 0; } int pthread_attr_destroy (pthread_attr_t *attr) { if (!pthread_attr::is_good_object (attr)) return EINVAL; delete (*attr); *attr = NULL; return 0; } int pthread_getattr_np (pthread_t thread, pthread_attr_t *attr) { THREAD_BASIC_INFORMATION tbi; NTSTATUS status; if (!pthread::is_good_object (&thread)) return ESRCH; /* attr may not be pre-initialized */ if (!pthread_attr::is_good_object (attr)) { int rv = pthread_attr_init (attr); if (rv != 0) return rv; } (*attr)->joinable = thread->attr.joinable; (*attr)->contentionscope = thread->attr.contentionscope; (*attr)->inheritsched = thread->attr.inheritsched; (*attr)->schedparam = thread->attr.schedparam; (*attr)->guardsize = thread->attr.guardsize; status = NtQueryInformationThread (thread->win32_obj_id, ThreadBasicInformation, &tbi, sizeof (tbi), NULL); if (NT_SUCCESS (status)) { PTEB teb = (PTEB) tbi.TebBaseAddress; /* stackaddr holds the uppermost stack address. See the comments in pthread_attr_setstack and pthread_attr_setstackaddr for a description. */ (*attr)->stackaddr = teb->Tib.StackBase; (*attr)->stacksize = (uintptr_t) teb->Tib.StackBase - (uintptr_t) (teb->DeallocationStack ?: teb->Tib.StackLimit); } else { debug_printf ("NtQueryInformationThread(ThreadBasicInformation), " "status %y", status); (*attr)->stackaddr = thread->attr.stackaddr; (*attr)->stacksize = thread->attr.stacksize; } return 0; } /* Thread Specific Data */ int pthread_key_create (pthread_key_t *key, void (*destructor) (void *)) { *key = new pthread_key (destructor); if (!pthread_key::is_good_object (key)) { delete (*key); *key = NULL; return EAGAIN; } return 0; } int pthread_key_delete (pthread_key_t key) { if (!pthread_key::is_good_object (&key)) return EINVAL; delete (key); return 0; } void * pthread_getspecific (pthread_key_t key) { if (!pthread_key::is_good_object (&key)) return NULL; return (key)->get (); } int pthread_setspecific (pthread_key_t key, const void *value) { if (!pthread_key::is_good_object (&key)) return EINVAL; (key)->set (value); return 0; } /* Mutexes */ int pthread_mutex_init (pthread_mutex_t * mutex, const pthread_mutexattr_t * attr) { return pthread_mutex::init (mutex, attr, NULL); } int pthread_mutex_getprioceiling (const pthread_mutex_t *mutex, int *prioceiling) { /* We don't define _POSIX_THREAD_PRIO_PROTECT because we do't currently support mutex priorities. We can support mutex priorities in the future though: Store a priority with each mutex. When the mutex is optained, set the thread priority as appropriate When the mutex is released, reset the thread priority. */ return ENOSYS; } int pthread_mutex_lock (pthread_mutex_t *mutex) { if (pthread_mutex::is_initializer (mutex)) pthread_mutex::init (mutex, NULL, *mutex); if (!pthread_mutex::is_good_object (mutex)) return EINVAL; return (*mutex)->lock (); } int pthread_mutex_clocklock (pthread_mutex_t *mutex, clockid_t clock_id, const struct timespec *abstime) { LARGE_INTEGER timeout; if (pthread_mutex::is_initializer (mutex)) pthread_mutex::init (mutex, NULL, *mutex); if (!pthread_mutex::is_good_object (mutex)) return EINVAL; /* According to SUSv3, abstime need not be checked for validity, if the mutex can be locked immediately. */ if (!(*mutex)->trylock ()) return 0; __try { int err = pthread_convert_abstime (clock_id, abstime, &timeout); if (err) return err; return (*mutex)->lock (&timeout); } __except (NO_ERROR) {} __endtry return EINVAL; } int pthread_mutex_timedlock (pthread_mutex_t *mutex, const struct timespec *abstime) { return pthread_mutex_clocklock (mutex, CLOCK_REALTIME, abstime); } int pthread_mutex_trylock (pthread_mutex_t *mutex) { if (pthread_mutex::is_initializer (mutex)) pthread_mutex::init (mutex, NULL, *mutex); if (!pthread_mutex::is_good_object (mutex)) return EINVAL; return (*mutex)->trylock (); } int pthread_mutex_unlock (pthread_mutex_t *mutex) { if (pthread_mutex::is_initializer (mutex)) pthread_mutex::init (mutex, NULL, *mutex); if (!pthread_mutex::is_good_object (mutex)) return EINVAL; return (*mutex)->unlock (); } int pthread_mutex_destroy (pthread_mutex_t *mutex) { int rv; if (pthread_mutex::is_initializer (mutex)) return 0; if (!pthread_mutex::is_good_object (mutex)) return EINVAL; rv = (*mutex)->destroy (); if (rv) return rv; *mutex = NULL; return 0; } int pthread_mutex_setprioceiling (pthread_mutex_t *mutex, int prioceiling, int *old_ceiling) { return ENOSYS; } /* Mutex attributes */ /* Win32 doesn't support mutex priorities - see __pthread_mutex_getprioceiling for more detail */ int pthread_mutexattr_getprotocol (const pthread_mutexattr_t *attr, int *protocol) { if (!pthread_mutexattr::is_good_object (attr)) return EINVAL; return ENOSYS; } int pthread_mutexattr_getpshared (const pthread_mutexattr_t *attr, int *pshared) { if (!pthread_mutexattr::is_good_object (attr)) return EINVAL; *pshared = (*attr)->pshared; return 0; } int pthread_mutexattr_gettype (const pthread_mutexattr_t *attr, int *type) { if (!pthread_mutexattr::is_good_object (attr)) return EINVAL; *type = (*attr)->mutextype; return 0; } /* FIXME: write and test process shared mutex's. */ int pthread_mutexattr_init (pthread_mutexattr_t *attr) { *attr = new pthread_mutexattr (); if (!pthread_mutexattr::is_good_object (attr)) { delete (*attr); *attr = NULL; return ENOMEM; } return 0; } int pthread_mutexattr_destroy (pthread_mutexattr_t *attr) { if (!pthread_mutexattr::is_good_object (attr)) return EINVAL; delete (*attr); *attr = NULL; return 0; } /* Win32 doesn't support mutex priorities */ int pthread_mutexattr_setprotocol (pthread_mutexattr_t *attr, int protocol) { if (!pthread_mutexattr::is_good_object (attr)) return EINVAL; return ENOSYS; } /* Win32 doesn't support mutex priorities */ int pthread_mutexattr_setprioceiling (pthread_mutexattr_t *attr, int prioceiling) { if (!pthread_mutexattr::is_good_object (attr)) return EINVAL; return ENOSYS; } int pthread_mutexattr_getprioceiling (const pthread_mutexattr_t *attr, int *prioceiling) { if (!pthread_mutexattr::is_good_object (attr)) return EINVAL; return ENOSYS; } int pthread_mutexattr_setpshared (pthread_mutexattr_t *attr, int pshared) { if (!pthread_mutexattr::is_good_object (attr)) return EINVAL; /* we don't use pshared for anything as yet. We need to test PROCESS_SHARED *functionality */ if (pshared != PTHREAD_PROCESS_PRIVATE) return EINVAL; (*attr)->pshared = pshared; return 0; } /* see pthread_mutex_gettype */ int pthread_mutexattr_settype (pthread_mutexattr_t *attr, int type) { if (!pthread_mutexattr::is_good_object (attr)) return EINVAL; switch (type) { case PTHREAD_MUTEX_ERRORCHECK: case PTHREAD_MUTEX_RECURSIVE: case PTHREAD_MUTEX_NORMAL: (*attr)->mutextype = type; break; default: return EINVAL; } return 0; } /* Spinlocks */ int pthread_spin_init (pthread_spinlock_t *spinlock, int pshared) { return pthread_spinlock::init (spinlock, pshared); } int pthread_spin_lock (pthread_spinlock_t *spinlock) { if (!pthread_spinlock::is_good_object (spinlock)) return EINVAL; return (*spinlock)->lock (); } int pthread_spin_trylock (pthread_spinlock_t *spinlock) { if (!pthread_spinlock::is_good_object (spinlock)) return EINVAL; return (*spinlock)->trylock (); } int pthread_spin_unlock (pthread_spinlock_t *spinlock) { if (!pthread_spinlock::is_good_object (spinlock)) return EINVAL; return (*spinlock)->unlock (); } int pthread_spin_destroy (pthread_spinlock_t *spinlock) { if (!pthread_spinlock::is_good_object (spinlock)) return EINVAL; return (*spinlock)->destroy (); } /* Synchronisation */ int pthread_cond_init (pthread_cond_t * cond, const pthread_condattr_t * attr) { return pthread_cond::init (cond, attr); } int pthread_cond_destroy (pthread_cond_t *cond) { if (pthread_cond::is_initializer (cond)) return 0; if (!pthread_cond::is_good_object (cond)) return EINVAL; /* reads are atomic */ if ((*cond)->waiting) return EBUSY; delete (*cond); *cond = NULL; return 0; } int pthread_cond_broadcast (pthread_cond_t *cond) { if (pthread_cond::is_initializer (cond)) return 0; if (!pthread_cond::is_good_object (cond)) return EINVAL; (*cond)->unblock (true); return 0; } int pthread_cond_signal (pthread_cond_t *cond) { if (pthread_cond::is_initializer (cond)) return 0; if (!pthread_cond::is_good_object (cond)) return EINVAL; (*cond)->unblock (false); return 0; } static int __pthread_cond_wait_init (pthread_cond_t *cond, pthread_mutex_t *mutex) { if (!pthread_mutex::is_good_object (mutex)) return EINVAL; if (!(*mutex)->can_be_unlocked ()) return EPERM; if (pthread_cond::is_initializer (cond)) pthread_cond::init (cond, NULL); if (!pthread_cond::is_good_object (cond)) return EINVAL; return 0; } static int __pthread_cond_clockwait (pthread_cond_t *cond, pthread_mutex_t *mutex, clockid_t clock_id, const struct timespec *abstime) { int err = 0; LARGE_INTEGER timeout; do { err = pthread_convert_abstime (clock_id, abstime, &timeout); if (err) break; err = (*cond)->wait (*mutex, &timeout); } while (err == ETIMEDOUT); return err; } int pthread_cond_clockwait (pthread_cond_t *cond, pthread_mutex_t *mutex, clockid_t clock_id, const struct timespec *abstime) { int err = 0; pthread_testcancel (); __try { err = __pthread_cond_wait_init (cond, mutex); if (err) __leave; err = __pthread_cond_clockwait (cond, mutex, clock_id, abstime); } __except (NO_ERROR) { return EINVAL; } __endtry return err; } int pthread_cond_timedwait (pthread_cond_t *cond, pthread_mutex_t *mutex, const struct timespec *abstime) { int err = 0; pthread_testcancel (); __try { err = __pthread_cond_wait_init (cond, mutex); if (err) __leave; err = __pthread_cond_clockwait (cond, mutex, (*cond)->clock_id, abstime); } __except (NO_ERROR) { return EINVAL; } __endtry return err; } int pthread_cond_wait (pthread_cond_t *cond, pthread_mutex_t *mutex) { pthread_testcancel (); int err = __pthread_cond_wait_init (cond, mutex); if (err) return err; return (*cond)->wait (*mutex, NULL); } /* Thread cond attributes */ int pthread_condattr_init (pthread_condattr_t *condattr) { *condattr = new pthread_condattr; if (!pthread_condattr::is_good_object (condattr)) { delete (*condattr); *condattr = NULL; return ENOMEM; } return 0; } int pthread_condattr_getpshared (const pthread_condattr_t *attr, int *pshared) { if (!pthread_condattr::is_good_object (attr)) return EINVAL; *pshared = (*attr)->shared; return 0; } int pthread_condattr_setpshared (pthread_condattr_t *attr, int pshared) { if (!pthread_condattr::is_good_object (attr)) return EINVAL; if ((pshared < 0) || (pshared > 1)) return EINVAL; /* shared cond vars not currently supported */ if (pshared != PTHREAD_PROCESS_PRIVATE) return EINVAL; (*attr)->shared = pshared; return 0; } int pthread_condattr_getclock (const pthread_condattr_t *attr, clockid_t *clock_id) { if (!pthread_condattr::is_good_object (attr)) return EINVAL; *clock_id = (*attr)->clock_id; return 0; } int pthread_condattr_setclock (pthread_condattr_t *attr, clockid_t clock_id) { if (!pthread_condattr::is_good_object (attr)) return EINVAL; if (CLOCKID_IS_PROCESS (clock_id) || CLOCKID_IS_THREAD (clock_id) || clock_id >= MAX_CLOCKS) return EINVAL; (*attr)->clock_id = clock_id; return 0; } int pthread_condattr_destroy (pthread_condattr_t *condattr) { if (!pthread_condattr::is_good_object (condattr)) return EINVAL; delete (*condattr); *condattr = NULL; return 0; } /* RW Locks */ int pthread_rwlock_init (pthread_rwlock_t *rwlock, const pthread_rwlockattr_t *attr) { return pthread_rwlock::init (rwlock, attr); } int pthread_rwlock_destroy (pthread_rwlock_t *rwlock) { if (pthread_rwlock::is_initializer (rwlock)) return 0; if (!pthread_rwlock::is_good_object (rwlock)) return EINVAL; if ((*rwlock)->writer || (*rwlock)->readers || (*rwlock)->waiting_readers || (*rwlock)->waiting_writers) return EBUSY; delete (*rwlock); *rwlock = NULL; return 0; } int pthread_rwlock_rdlock (pthread_rwlock_t *rwlock) { pthread_testcancel (); if (pthread_rwlock::is_initializer (rwlock)) pthread_rwlock::init (rwlock, NULL); if (!pthread_rwlock::is_good_object (rwlock)) return EINVAL; return (*rwlock)->rdlock (); } int pthread_rwlock_clockrdlock (pthread_rwlock_t *rwlock, clockid_t clock_id, const struct timespec *abstime) { LARGE_INTEGER timeout; pthread_testcancel (); if (pthread_rwlock::is_initializer (rwlock)) pthread_rwlock::init (rwlock, NULL); if (!pthread_rwlock::is_good_object (rwlock)) return EINVAL; /* According to SUSv3, abstime need not be checked for validity, if the rwlock can be locked immediately. */ if (!(*rwlock)->tryrdlock ()) return 0; __try { int err = pthread_convert_abstime (clock_id, abstime, &timeout); if (err) return err; return (*rwlock)->rdlock (&timeout); } __except (NO_ERROR) {} __endtry return EINVAL; } int pthread_rwlock_timedrdlock (pthread_rwlock_t *rwlock, const struct timespec *abstime) { return pthread_rwlock_clockrdlock (rwlock, CLOCK_REALTIME, abstime); } int pthread_rwlock_tryrdlock (pthread_rwlock_t *rwlock) { if (pthread_rwlock::is_initializer (rwlock)) pthread_rwlock::init (rwlock, NULL); if (!pthread_rwlock::is_good_object (rwlock)) return EINVAL; return (*rwlock)->tryrdlock (); } int pthread_rwlock_wrlock (pthread_rwlock_t *rwlock) { pthread_testcancel (); if (pthread_rwlock::is_initializer (rwlock)) pthread_rwlock::init (rwlock, NULL); if (!pthread_rwlock::is_good_object (rwlock)) return EINVAL; return (*rwlock)->wrlock (); } int pthread_rwlock_clockwrlock (pthread_rwlock_t *rwlock, clockid_t clock_id, const struct timespec *abstime) { LARGE_INTEGER timeout; pthread_testcancel (); if (pthread_rwlock::is_initializer (rwlock)) pthread_rwlock::init (rwlock, NULL); if (!pthread_rwlock::is_good_object (rwlock)) return EINVAL; /* According to SUSv3, abstime need not be checked for validity, if the rwlock can be locked immediately. */ if (!(*rwlock)->trywrlock ()) return 0; __try { int err = pthread_convert_abstime (clock_id, abstime, &timeout); if (err) return err; return (*rwlock)->wrlock (&timeout); } __except (NO_ERROR) {} __endtry return EINVAL; } int pthread_rwlock_timedwrlock (pthread_rwlock_t *rwlock, const struct timespec *abstime) { return pthread_rwlock_clockwrlock (rwlock, CLOCK_REALTIME, abstime); } int pthread_rwlock_trywrlock (pthread_rwlock_t *rwlock) { if (pthread_rwlock::is_initializer (rwlock)) pthread_rwlock::init (rwlock, NULL); if (!pthread_rwlock::is_good_object (rwlock)) return EINVAL; return (*rwlock)->trywrlock (); } int pthread_rwlock_unlock (pthread_rwlock_t *rwlock) { if (pthread_rwlock::is_initializer (rwlock)) return 0; if (!pthread_rwlock::is_good_object (rwlock)) return EINVAL; return (*rwlock)->unlock (); } /* RW Lock attributes */ int pthread_rwlockattr_init (pthread_rwlockattr_t *rwlockattr) { *rwlockattr = new pthread_rwlockattr; if (!pthread_rwlockattr::is_good_object (rwlockattr)) { delete (*rwlockattr); *rwlockattr = NULL; return ENOMEM; } return 0; } int pthread_rwlockattr_getpshared (const pthread_rwlockattr_t *attr, int *pshared) { if (!pthread_rwlockattr::is_good_object (attr)) return EINVAL; *pshared = (*attr)->shared; return 0; } int pthread_rwlockattr_setpshared (pthread_rwlockattr_t *attr, int pshared) { if (!pthread_rwlockattr::is_good_object (attr)) return EINVAL; if ((pshared < 0) || (pshared > 1)) return EINVAL; /* shared rwlock vars not currently supported */ if (pshared != PTHREAD_PROCESS_PRIVATE) return EINVAL; (*attr)->shared = pshared; return 0; } int pthread_rwlockattr_destroy (pthread_rwlockattr_t *rwlockattr) { if (!pthread_rwlockattr::is_good_object (rwlockattr)) return EINVAL; delete (*rwlockattr); *rwlockattr = NULL; return 0; } /* Barriers */ int pthread_barrier_init (pthread_barrier_t * bar, const pthread_barrierattr_t * attr, unsigned count) { if (unlikely (bar == NULL)) return EINVAL; *bar = new pthread_barrier; return (*bar)->init (attr, count); } int pthread_barrier_destroy (pthread_barrier_t * bar) { if (unlikely (! pthread_barrier::is_good_object (bar))) return EINVAL; int ret; ret = (*bar)->destroy (); if (ret == 0) delete_and_clear (bar); return ret; } int pthread_barrier_wait (pthread_barrier_t * bar) { if (unlikely (! pthread_barrier::is_good_object (bar))) return EINVAL; return (*bar)->wait (); } /* Barrier attributes */ int pthread_barrierattr_init (pthread_barrierattr_t * battr) { if (unlikely (battr == NULL)) return EINVAL; *battr = new pthread_barrierattr; (*battr)->shared = PTHREAD_PROCESS_PRIVATE; return 0; } int pthread_barrierattr_setpshared (pthread_barrierattr_t * battr, int shared) { if (unlikely (! pthread_barrierattr::is_good_object (battr))) return EINVAL; if (unlikely (shared != PTHREAD_PROCESS_SHARED && shared != PTHREAD_PROCESS_PRIVATE)) return EINVAL; (*battr)->shared = shared; return 0; } int pthread_barrierattr_getpshared (const pthread_barrierattr_t * battr, int * shared) { if (unlikely (! pthread_barrierattr::is_good_object (battr) || shared == NULL)) return EINVAL; *shared = (*battr)->shared; return 0; } int pthread_barrierattr_destroy (pthread_barrierattr_t * battr) { if (unlikely (! pthread_barrierattr::is_good_object (battr))) return EINVAL; delete_and_clear (battr); return 0; } /* Thread clock ID */ int pthread_getcpuclockid (pthread_t thread, clockid_t *clk_id) { if (!pthread::is_good_object (&thread)) return (ESRCH); *clk_id = (clockid_t) THREADID_TO_CLOCKID (thread->getsequence_np ()); return 0; } }