thr_private.h revision 330897
1/*- 2 * SPDX-License-Identifier: BSD-2-Clause-FreeBSD 3 * 4 * Copyright (C) 2005 Daniel M. Eischen <deischen@freebsd.org> 5 * Copyright (c) 2005 David Xu <davidxu@freebsd.org> 6 * Copyright (c) 1995-1998 John Birrell <jb@cimlogic.com.au>. 7 * 8 * All rights reserved. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice unmodified, this list of conditions, and the following 15 * disclaimer. 16 * 2. Redistributions in binary form must reproduce the above copyright 17 * notice, this list of conditions and the following disclaimer in the 18 * documentation and/or other materials provided with the distribution. 19 * 20 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 21 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 22 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 23 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 24 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 25 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 26 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 27 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 28 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 29 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 30 * 31 * $FreeBSD: stable/11/lib/libthr/thread/thr_private.h 330897 2018-03-14 03:19:51Z eadler $ 32 */ 33 34#ifndef _THR_PRIVATE_H 35#define _THR_PRIVATE_H 36 37/* 38 * Include files. 39 */ 40#include <sys/types.h> 41#include <sys/time.h> 42#include <sys/cdefs.h> 43#include <sys/queue.h> 44#include <sys/param.h> 45#include <sys/cpuset.h> 46#include <machine/atomic.h> 47#include <errno.h> 48#include <limits.h> 49#include <signal.h> 50#include <stdbool.h> 51#include <stddef.h> 52#include <stdio.h> 53#include <unistd.h> 54#include <ucontext.h> 55#include <sys/thr.h> 56#include <pthread.h> 57 58__NULLABILITY_PRAGMA_PUSH 59 60#define SYM_FB10(sym) __CONCAT(sym, _fb10) 61#define SYM_FBP10(sym) __CONCAT(sym, _fbp10) 62#define WEAK_REF(sym, alias) __weak_reference(sym, alias) 63#define SYM_COMPAT(sym, impl, ver) __sym_compat(sym, impl, ver) 64#define SYM_DEFAULT(sym, impl, ver) __sym_default(sym, impl, ver) 65 66#define FB10_COMPAT(func, sym) \ 67 WEAK_REF(func, SYM_FB10(sym)); \ 68 SYM_COMPAT(sym, SYM_FB10(sym), FBSD_1.0) 69 70#define FB10_COMPAT_PRIVATE(func, sym) \ 71 WEAK_REF(func, SYM_FBP10(sym)); \ 72 SYM_DEFAULT(sym, SYM_FBP10(sym), FBSDprivate_1.0) 73 74struct pthread; 75extern struct pthread *_thr_initial __hidden; 76 77#include "pthread_md.h" 78#include "thr_umtx.h" 79#include "thread_db.h" 80 81#ifdef _PTHREAD_FORCED_UNWIND 82#define _BSD_SOURCE 83#include <unwind.h> 84#endif 85 86typedef TAILQ_HEAD(pthreadlist, pthread) pthreadlist; 87typedef TAILQ_HEAD(atfork_head, pthread_atfork) atfork_head; 88TAILQ_HEAD(mutex_queue, pthread_mutex); 89 90/* Signal to do cancellation */ 91#define SIGCANCEL SIGTHR 92 93/* 94 * Kernel fatal error handler macro. 95 */ 96#define PANIC(args...) _thread_exitf(__FILE__, __LINE__, ##args) 97 98/* Output debug messages like this: */ 99#define stdout_debug(args...) _thread_printf(STDOUT_FILENO, ##args) 100#define stderr_debug(args...) _thread_printf(STDERR_FILENO, ##args) 101 102#ifdef _PTHREADS_INVARIANTS 103#define THR_ASSERT(cond, msg) do { \ 104 if (__predict_false(!(cond))) \ 105 PANIC(msg); \ 106} while (0) 107#else 108#define THR_ASSERT(cond, msg) 109#endif 110 111#ifdef PIC 112# define STATIC_LIB_REQUIRE(name) 113#else 114# define STATIC_LIB_REQUIRE(name) __asm (".globl " #name) 115#endif 116 117#define TIMESPEC_ADD(dst, src, val) \ 118 do { \ 119 (dst)->tv_sec = (src)->tv_sec + (val)->tv_sec; \ 120 (dst)->tv_nsec = (src)->tv_nsec + (val)->tv_nsec; \ 121 if ((dst)->tv_nsec >= 1000000000) { \ 122 (dst)->tv_sec++; \ 123 (dst)->tv_nsec -= 1000000000; \ 124 } \ 125 } while (0) 126 127#define TIMESPEC_SUB(dst, src, val) \ 128 do { \ 129 (dst)->tv_sec = (src)->tv_sec - (val)->tv_sec; \ 130 (dst)->tv_nsec = (src)->tv_nsec - (val)->tv_nsec; \ 131 if ((dst)->tv_nsec < 0) { \ 132 (dst)->tv_sec--; \ 133 (dst)->tv_nsec += 1000000000; \ 134 } \ 135 } while (0) 136 137/* Magic cookie set for shared pthread locks and cv's pointers */ 138#define THR_PSHARED_PTR \ 139 ((void *)(uintptr_t)((1ULL << (NBBY * sizeof(long) - 1)) | 1)) 140 141/* XXX These values should be same as those defined in pthread.h */ 142#define THR_MUTEX_INITIALIZER ((struct pthread_mutex *)NULL) 143#define THR_ADAPTIVE_MUTEX_INITIALIZER ((struct pthread_mutex *)1) 144#define THR_MUTEX_DESTROYED ((struct pthread_mutex *)2) 145#define THR_COND_INITIALIZER ((struct pthread_cond *)NULL) 146#define THR_COND_DESTROYED ((struct pthread_cond *)1) 147#define THR_RWLOCK_INITIALIZER ((struct pthread_rwlock *)NULL) 148#define THR_RWLOCK_DESTROYED ((struct pthread_rwlock *)1) 149 150#define PMUTEX_FLAG_TYPE_MASK 0x0ff 151#define PMUTEX_FLAG_PRIVATE 0x100 152#define PMUTEX_FLAG_DEFERRED 0x200 153#define PMUTEX_TYPE(mtxflags) ((mtxflags) & PMUTEX_FLAG_TYPE_MASK) 154 155#define PMUTEX_OWNER_ID(m) ((m)->m_lock.m_owner & ~UMUTEX_CONTESTED) 156 157#define MAX_DEFER_WAITERS 50 158 159/* 160 * Values for pthread_mutex m_ps indicator. 161 */ 162#define PMUTEX_INITSTAGE_ALLOC 0 163#define PMUTEX_INITSTAGE_BUSY 1 164#define PMUTEX_INITSTAGE_DONE 2 165 166struct pthread_mutex { 167 /* 168 * Lock for accesses to this structure. 169 */ 170 struct umutex m_lock; 171 int m_flags; 172 int m_count; 173 int m_spinloops; 174 int m_yieldloops; 175 int m_ps; /* pshared init stage */ 176 /* 177 * Link for all mutexes a thread currently owns, of the same 178 * prio type. 179 */ 180 TAILQ_ENTRY(pthread_mutex) m_qe; 181 /* Link for all private mutexes a thread currently owns. */ 182 TAILQ_ENTRY(pthread_mutex) m_pqe; 183 struct pthread_mutex *m_rb_prev; 184}; 185 186struct pthread_mutex_attr { 187 enum pthread_mutextype m_type; 188 int m_protocol; 189 int m_ceiling; 190 int m_pshared; 191 int m_robust; 192}; 193 194#define PTHREAD_MUTEXATTR_STATIC_INITIALIZER \ 195 { PTHREAD_MUTEX_DEFAULT, PTHREAD_PRIO_NONE, 0, MUTEX_FLAGS_PRIVATE, \ 196 PTHREAD_MUTEX_STALLED } 197 198struct pthread_cond { 199 __uint32_t __has_user_waiters; 200 struct ucond kcond; 201}; 202 203struct pthread_cond_attr { 204 int c_pshared; 205 int c_clockid; 206}; 207 208struct pthread_barrier { 209 struct umutex b_lock; 210 struct ucond b_cv; 211 int64_t b_cycle; 212 int b_count; 213 int b_waiters; 214 int b_refcount; 215 int b_destroying; 216}; 217 218struct pthread_barrierattr { 219 int pshared; 220}; 221 222struct pthread_spinlock { 223 struct umutex s_lock; 224}; 225 226/* 227 * Flags for condition variables. 228 */ 229#define COND_FLAGS_PRIVATE 0x01 230#define COND_FLAGS_INITED 0x02 231#define COND_FLAGS_BUSY 0x04 232 233/* 234 * Cleanup definitions. 235 */ 236struct pthread_cleanup { 237 struct pthread_cleanup *prev; 238 void (*routine)(void *); 239 void *routine_arg; 240 int onheap; 241}; 242 243#define THR_CLEANUP_PUSH(td, func, arg) { \ 244 struct pthread_cleanup __cup; \ 245 \ 246 __cup.routine = func; \ 247 __cup.routine_arg = arg; \ 248 __cup.onheap = 0; \ 249 __cup.prev = (td)->cleanup; \ 250 (td)->cleanup = &__cup; 251 252#define THR_CLEANUP_POP(td, exec) \ 253 (td)->cleanup = __cup.prev; \ 254 if ((exec) != 0) \ 255 __cup.routine(__cup.routine_arg); \ 256} 257 258struct pthread_atfork { 259 TAILQ_ENTRY(pthread_atfork) qe; 260 void (*prepare)(void); 261 void (*parent)(void); 262 void (*child)(void); 263}; 264 265struct pthread_attr { 266#define pthread_attr_start_copy sched_policy 267 int sched_policy; 268 int sched_inherit; 269 int prio; 270 int suspend; 271#define THR_STACK_USER 0x100 /* 0xFF reserved for <pthread.h> */ 272 int flags; 273 void *stackaddr_attr; 274 size_t stacksize_attr; 275 size_t guardsize_attr; 276#define pthread_attr_end_copy cpuset 277 cpuset_t *cpuset; 278 size_t cpusetsize; 279}; 280 281struct wake_addr { 282 struct wake_addr *link; 283 unsigned int value; 284 char pad[12]; 285}; 286 287struct sleepqueue { 288 TAILQ_HEAD(, pthread) sq_blocked; 289 SLIST_HEAD(, sleepqueue) sq_freeq; 290 LIST_ENTRY(sleepqueue) sq_hash; 291 SLIST_ENTRY(sleepqueue) sq_flink; 292 void *sq_wchan; 293 int sq_type; 294}; 295 296/* 297 * Thread creation state attributes. 298 */ 299#define THR_CREATE_RUNNING 0 300#define THR_CREATE_SUSPENDED 1 301 302/* 303 * Miscellaneous definitions. 304 */ 305#define THR_STACK_DEFAULT (sizeof(void *) / 4 * 1024 * 1024) 306 307/* 308 * Maximum size of initial thread's stack. This perhaps deserves to be larger 309 * than the stacks of other threads, since many applications are likely to run 310 * almost entirely on this stack. 311 */ 312#define THR_STACK_INITIAL (THR_STACK_DEFAULT * 2) 313 314/* 315 * Define priorities returned by kernel. 316 */ 317#define THR_MIN_PRIORITY (_thr_priorities[SCHED_OTHER-1].pri_min) 318#define THR_MAX_PRIORITY (_thr_priorities[SCHED_OTHER-1].pri_max) 319#define THR_DEF_PRIORITY (_thr_priorities[SCHED_OTHER-1].pri_default) 320 321#define THR_MIN_RR_PRIORITY (_thr_priorities[SCHED_RR-1].pri_min) 322#define THR_MAX_RR_PRIORITY (_thr_priorities[SCHED_RR-1].pri_max) 323#define THR_DEF_RR_PRIORITY (_thr_priorities[SCHED_RR-1].pri_default) 324 325/* XXX The SCHED_FIFO should have same priority range as SCHED_RR */ 326#define THR_MIN_FIFO_PRIORITY (_thr_priorities[SCHED_FIFO_1].pri_min) 327#define THR_MAX_FIFO_PRIORITY (_thr_priorities[SCHED_FIFO-1].pri_max) 328#define THR_DEF_FIFO_PRIORITY (_thr_priorities[SCHED_FIFO-1].pri_default) 329 330struct pthread_prio { 331 int pri_min; 332 int pri_max; 333 int pri_default; 334}; 335 336struct pthread_rwlockattr { 337 int pshared; 338}; 339 340struct pthread_rwlock { 341 struct urwlock lock; 342 uint32_t owner; 343}; 344 345/* 346 * Thread states. 347 */ 348enum pthread_state { 349 PS_RUNNING, 350 PS_DEAD 351}; 352 353struct pthread_specific_elem { 354 const void *data; 355 int seqno; 356}; 357 358struct pthread_key { 359 volatile int allocated; 360 int seqno; 361 void (*destructor)(void *); 362}; 363 364/* 365 * lwpid_t is 32bit but kernel thr API exports tid as long type 366 * to preserve the ABI for M:N model in very early date (r131431). 367 */ 368#define TID(thread) ((uint32_t) ((thread)->tid)) 369 370/* 371 * Thread structure. 372 */ 373struct pthread { 374#define _pthread_startzero tid 375 /* Kernel thread id. */ 376 long tid; 377#define TID_TERMINATED 1 378 379 /* 380 * Lock for accesses to this thread structure. 381 */ 382 struct umutex lock; 383 384 /* Internal condition variable cycle number. */ 385 uint32_t cycle; 386 387 /* How many low level locks the thread held. */ 388 int locklevel; 389 390 /* 391 * Set to non-zero when this thread has entered a critical 392 * region. We allow for recursive entries into critical regions. 393 */ 394 int critical_count; 395 396 /* Signal blocked counter. */ 397 int sigblock; 398 399 /* Queue entry for list of all threads. */ 400 TAILQ_ENTRY(pthread) tle; /* link for all threads in process */ 401 402 /* Queue entry for GC lists. */ 403 TAILQ_ENTRY(pthread) gcle; 404 405 /* Hash queue entry. */ 406 LIST_ENTRY(pthread) hle; 407 408 /* Sleep queue entry */ 409 TAILQ_ENTRY(pthread) wle; 410 411 /* Threads reference count. */ 412 int refcount; 413 414 /* 415 * Thread start routine, argument, stack pointer and thread 416 * attributes. 417 */ 418 void *(*start_routine)(void *); 419 void *arg; 420 struct pthread_attr attr; 421 422#define SHOULD_CANCEL(thr) \ 423 ((thr)->cancel_pending && (thr)->cancel_enable && \ 424 (thr)->no_cancel == 0) 425 426 /* Cancellation is enabled */ 427 int cancel_enable; 428 429 /* Cancellation request is pending */ 430 int cancel_pending; 431 432 /* Thread is at cancellation point */ 433 int cancel_point; 434 435 /* Cancellation is temporarily disabled */ 436 int no_cancel; 437 438 /* Asynchronouse cancellation is enabled */ 439 int cancel_async; 440 441 /* Cancellation is in progress */ 442 int cancelling; 443 444 /* Thread temporary signal mask. */ 445 sigset_t sigmask; 446 447 /* Thread should unblock SIGCANCEL. */ 448 int unblock_sigcancel; 449 450 /* In sigsuspend state */ 451 int in_sigsuspend; 452 453 /* deferred signal info */ 454 siginfo_t deferred_siginfo; 455 456 /* signal mask to restore. */ 457 sigset_t deferred_sigmask; 458 459 /* the sigaction should be used for deferred signal. */ 460 struct sigaction deferred_sigact; 461 462 /* deferred signal delivery is performed, do not reenter. */ 463 int deferred_run; 464 465 /* Force new thread to exit. */ 466 int force_exit; 467 468 /* Thread state: */ 469 enum pthread_state state; 470 471 /* 472 * Error variable used instead of errno. The function __error() 473 * returns a pointer to this. 474 */ 475 int error; 476 477 /* 478 * The joiner is the thread that is joining to this thread. The 479 * join status keeps track of a join operation to another thread. 480 */ 481 struct pthread *joiner; 482 483 /* Miscellaneous flags; only set with scheduling lock held. */ 484 int flags; 485#define THR_FLAGS_PRIVATE 0x0001 486#define THR_FLAGS_NEED_SUSPEND 0x0002 /* thread should be suspended */ 487#define THR_FLAGS_SUSPENDED 0x0004 /* thread is suspended */ 488#define THR_FLAGS_DETACHED 0x0008 /* thread is detached */ 489 490 /* Thread list flags; only set with thread list lock held. */ 491 int tlflags; 492#define TLFLAGS_GC_SAFE 0x0001 /* thread safe for cleaning */ 493#define TLFLAGS_IN_TDLIST 0x0002 /* thread in all thread list */ 494#define TLFLAGS_IN_GCLIST 0x0004 /* thread in gc list */ 495 496 /* 497 * Queues of the owned mutexes. Private queue must have index 498 * + 1 of the corresponding full queue. 499 */ 500#define TMQ_NORM 0 /* NORMAL or PRIO_INHERIT normal */ 501#define TMQ_NORM_PRIV 1 /* NORMAL or PRIO_INHERIT normal priv */ 502#define TMQ_NORM_PP 2 /* PRIO_PROTECT normal mutexes */ 503#define TMQ_NORM_PP_PRIV 3 /* PRIO_PROTECT normal priv */ 504#define TMQ_ROBUST_PP 4 /* PRIO_PROTECT robust mutexes */ 505#define TMQ_ROBUST_PP_PRIV 5 /* PRIO_PROTECT robust priv */ 506#define TMQ_NITEMS 6 507 struct mutex_queue mq[TMQ_NITEMS]; 508 509 void *ret; 510 struct pthread_specific_elem *specific; 511 int specific_data_count; 512 513 /* Number rwlocks rdlocks held. */ 514 int rdlock_count; 515 516 /* 517 * Current locks bitmap for rtld. */ 518 int rtld_bits; 519 520 /* Thread control block */ 521 struct tcb *tcb; 522 523 /* Cleanup handlers Link List */ 524 struct pthread_cleanup *cleanup; 525 526#ifdef _PTHREAD_FORCED_UNWIND 527 struct _Unwind_Exception ex; 528 void *unwind_stackend; 529 int unwind_disabled; 530#endif 531 532 /* 533 * Magic value to help recognize a valid thread structure 534 * from an invalid one: 535 */ 536#define THR_MAGIC ((u_int32_t) 0xd09ba115) 537 u_int32_t magic; 538 539 /* Enable event reporting */ 540 int report_events; 541 542 /* Event mask */ 543 int event_mask; 544 545 /* Event */ 546 td_event_msg_t event_buf; 547 548 /* Wait channel */ 549 void *wchan; 550 551 /* Referenced mutex. */ 552 struct pthread_mutex *mutex_obj; 553 554 /* Thread will sleep. */ 555 int will_sleep; 556 557 /* Number of threads deferred. */ 558 int nwaiter_defer; 559 560 int robust_inited; 561 uintptr_t robust_list; 562 uintptr_t priv_robust_list; 563 uintptr_t inact_mtx; 564 565 /* Deferred threads from pthread_cond_signal. */ 566 unsigned int *defer_waiters[MAX_DEFER_WAITERS]; 567#define _pthread_endzero wake_addr 568 569 struct wake_addr *wake_addr; 570#define WAKE_ADDR(td) ((td)->wake_addr) 571 572 /* Sleep queue */ 573 struct sleepqueue *sleepqueue; 574 575}; 576 577#define THR_SHOULD_GC(thrd) \ 578 ((thrd)->refcount == 0 && (thrd)->state == PS_DEAD && \ 579 ((thrd)->flags & THR_FLAGS_DETACHED) != 0) 580 581#define THR_IN_CRITICAL(thrd) \ 582 (((thrd)->locklevel > 0) || \ 583 ((thrd)->critical_count > 0)) 584 585#define THR_CRITICAL_ENTER(thrd) \ 586 (thrd)->critical_count++ 587 588#define THR_CRITICAL_LEAVE(thrd) \ 589 do { \ 590 (thrd)->critical_count--; \ 591 _thr_ast(thrd); \ 592 } while (0) 593 594#define THR_UMUTEX_TRYLOCK(thrd, lck) \ 595 _thr_umutex_trylock((lck), TID(thrd)) 596 597#define THR_UMUTEX_LOCK(thrd, lck) \ 598 _thr_umutex_lock((lck), TID(thrd)) 599 600#define THR_UMUTEX_TIMEDLOCK(thrd, lck, timo) \ 601 _thr_umutex_timedlock((lck), TID(thrd), (timo)) 602 603#define THR_UMUTEX_UNLOCK(thrd, lck) \ 604 _thr_umutex_unlock((lck), TID(thrd)) 605 606#define THR_LOCK_ACQUIRE(thrd, lck) \ 607do { \ 608 (thrd)->locklevel++; \ 609 _thr_umutex_lock(lck, TID(thrd)); \ 610} while (0) 611 612#define THR_LOCK_ACQUIRE_SPIN(thrd, lck) \ 613do { \ 614 (thrd)->locklevel++; \ 615 _thr_umutex_lock_spin(lck, TID(thrd)); \ 616} while (0) 617 618#ifdef _PTHREADS_INVARIANTS 619#define THR_ASSERT_LOCKLEVEL(thrd) \ 620do { \ 621 if (__predict_false((thrd)->locklevel <= 0)) \ 622 _thr_assert_lock_level(); \ 623} while (0) 624#else 625#define THR_ASSERT_LOCKLEVEL(thrd) 626#endif 627 628#define THR_LOCK_RELEASE(thrd, lck) \ 629do { \ 630 THR_ASSERT_LOCKLEVEL(thrd); \ 631 _thr_umutex_unlock((lck), TID(thrd)); \ 632 (thrd)->locklevel--; \ 633 _thr_ast(thrd); \ 634} while (0) 635 636#define THR_LOCK(curthrd) THR_LOCK_ACQUIRE(curthrd, &(curthrd)->lock) 637#define THR_UNLOCK(curthrd) THR_LOCK_RELEASE(curthrd, &(curthrd)->lock) 638#define THR_THREAD_LOCK(curthrd, thr) THR_LOCK_ACQUIRE(curthrd, &(thr)->lock) 639#define THR_THREAD_UNLOCK(curthrd, thr) THR_LOCK_RELEASE(curthrd, &(thr)->lock) 640 641#define THREAD_LIST_RDLOCK(curthrd) \ 642do { \ 643 (curthrd)->locklevel++; \ 644 _thr_rwl_rdlock(&_thr_list_lock); \ 645} while (0) 646 647#define THREAD_LIST_WRLOCK(curthrd) \ 648do { \ 649 (curthrd)->locklevel++; \ 650 _thr_rwl_wrlock(&_thr_list_lock); \ 651} while (0) 652 653#define THREAD_LIST_UNLOCK(curthrd) \ 654do { \ 655 _thr_rwl_unlock(&_thr_list_lock); \ 656 (curthrd)->locklevel--; \ 657 _thr_ast(curthrd); \ 658} while (0) 659 660/* 661 * Macros to insert/remove threads to the all thread list and 662 * the gc list. 663 */ 664#define THR_LIST_ADD(thrd) do { \ 665 if (((thrd)->tlflags & TLFLAGS_IN_TDLIST) == 0) { \ 666 TAILQ_INSERT_HEAD(&_thread_list, thrd, tle); \ 667 _thr_hash_add(thrd); \ 668 (thrd)->tlflags |= TLFLAGS_IN_TDLIST; \ 669 } \ 670} while (0) 671#define THR_LIST_REMOVE(thrd) do { \ 672 if (((thrd)->tlflags & TLFLAGS_IN_TDLIST) != 0) { \ 673 TAILQ_REMOVE(&_thread_list, thrd, tle); \ 674 _thr_hash_remove(thrd); \ 675 (thrd)->tlflags &= ~TLFLAGS_IN_TDLIST; \ 676 } \ 677} while (0) 678#define THR_GCLIST_ADD(thrd) do { \ 679 if (((thrd)->tlflags & TLFLAGS_IN_GCLIST) == 0) { \ 680 TAILQ_INSERT_HEAD(&_thread_gc_list, thrd, gcle);\ 681 (thrd)->tlflags |= TLFLAGS_IN_GCLIST; \ 682 _gc_count++; \ 683 } \ 684} while (0) 685#define THR_GCLIST_REMOVE(thrd) do { \ 686 if (((thrd)->tlflags & TLFLAGS_IN_GCLIST) != 0) { \ 687 TAILQ_REMOVE(&_thread_gc_list, thrd, gcle); \ 688 (thrd)->tlflags &= ~TLFLAGS_IN_GCLIST; \ 689 _gc_count--; \ 690 } \ 691} while (0) 692 693#define THR_REF_ADD(curthread, pthread) { \ 694 THR_CRITICAL_ENTER(curthread); \ 695 pthread->refcount++; \ 696} while (0) 697 698#define THR_REF_DEL(curthread, pthread) { \ 699 pthread->refcount--; \ 700 THR_CRITICAL_LEAVE(curthread); \ 701} while (0) 702 703#define GC_NEEDED() (_gc_count >= 5) 704 705#define SHOULD_REPORT_EVENT(curthr, e) \ 706 (curthr->report_events && \ 707 (((curthr)->event_mask | _thread_event_mask ) & e) != 0) 708 709#ifndef __LIBC_ISTHREADED_DECLARED 710#define __LIBC_ISTHREADED_DECLARED 711extern int __isthreaded; 712#endif 713 714/* 715 * Global variables for the pthread kernel. 716 */ 717 718extern char *_usrstack __hidden; 719 720/* For debugger */ 721extern int _libthr_debug; 722extern int _thread_event_mask; 723extern struct pthread *_thread_last_event; 724 725/* List of all threads: */ 726extern pthreadlist _thread_list; 727 728/* List of threads needing GC: */ 729extern pthreadlist _thread_gc_list __hidden; 730 731extern int _thread_active_threads; 732extern atfork_head _thr_atfork_list __hidden; 733extern struct urwlock _thr_atfork_lock __hidden; 734 735/* Default thread attributes: */ 736extern struct pthread_attr _pthread_attr_default __hidden; 737 738/* Default mutex attributes: */ 739extern struct pthread_mutex_attr _pthread_mutexattr_default __hidden; 740extern struct pthread_mutex_attr _pthread_mutexattr_adaptive_default __hidden; 741 742/* Default condition variable attributes: */ 743extern struct pthread_cond_attr _pthread_condattr_default __hidden; 744 745extern struct pthread_prio _thr_priorities[] __hidden; 746 747extern int _thr_is_smp __hidden; 748 749extern size_t _thr_guard_default __hidden; 750extern size_t _thr_stack_default __hidden; 751extern size_t _thr_stack_initial __hidden; 752extern int _thr_page_size __hidden; 753extern int _thr_spinloops __hidden; 754extern int _thr_yieldloops __hidden; 755extern int _thr_queuefifo __hidden; 756 757/* Garbage thread count. */ 758extern int _gc_count __hidden; 759 760extern struct umutex _mutex_static_lock __hidden; 761extern struct umutex _cond_static_lock __hidden; 762extern struct umutex _rwlock_static_lock __hidden; 763extern struct umutex _keytable_lock __hidden; 764extern struct urwlock _thr_list_lock __hidden; 765extern struct umutex _thr_event_lock __hidden; 766extern struct umutex _suspend_all_lock __hidden; 767extern int _suspend_all_waiters __hidden; 768extern int _suspend_all_cycle __hidden; 769extern struct pthread *_single_thread __hidden; 770 771/* 772 * Function prototype definitions. 773 */ 774__BEGIN_DECLS 775int _thr_setthreaded(int) __hidden; 776int _mutex_cv_lock(struct pthread_mutex *, int, bool) __hidden; 777int _mutex_cv_unlock(struct pthread_mutex *, int *, int *) __hidden; 778int _mutex_cv_attach(struct pthread_mutex *, int) __hidden; 779int _mutex_cv_detach(struct pthread_mutex *, int *) __hidden; 780int _mutex_owned(struct pthread *, const struct pthread_mutex *) __hidden; 781int _mutex_reinit(pthread_mutex_t *) __hidden; 782void _mutex_fork(struct pthread *curthread) __hidden; 783int _mutex_enter_robust(struct pthread *curthread, struct pthread_mutex *m) 784 __hidden; 785void _mutex_leave_robust(struct pthread *curthread, struct pthread_mutex *m) 786 __hidden; 787void _libpthread_init(struct pthread *) __hidden; 788struct pthread *_thr_alloc(struct pthread *) __hidden; 789void _thread_exit(const char *, int, const char *) __hidden __dead2; 790void _thread_exitf(const char *, int, const char *, ...) __hidden __dead2 791 __printflike(3, 4); 792int _thr_ref_add(struct pthread *, struct pthread *, int) __hidden; 793void _thr_ref_delete(struct pthread *, struct pthread *) __hidden; 794void _thr_ref_delete_unlocked(struct pthread *, struct pthread *) __hidden; 795int _thr_find_thread(struct pthread *, struct pthread *, int) __hidden; 796void _thr_rtld_init(void) __hidden; 797void _thr_rtld_postfork_child(void) __hidden; 798int _thr_stack_alloc(struct pthread_attr *) __hidden; 799void _thr_stack_free(struct pthread_attr *) __hidden; 800void _thr_free(struct pthread *, struct pthread *) __hidden; 801void _thr_gc(struct pthread *) __hidden; 802void _thread_cleanupspecific(void) __hidden; 803void _thread_printf(int, const char *, ...) __hidden __printflike(2, 3); 804void _thread_vprintf(int, const char *, va_list) __hidden; 805void _thr_spinlock_init(void) __hidden; 806void _thr_cancel_enter(struct pthread *) __hidden; 807void _thr_cancel_enter2(struct pthread *, int) __hidden; 808void _thr_cancel_leave(struct pthread *, int) __hidden; 809void _thr_testcancel(struct pthread *) __hidden; 810void _thr_signal_block(struct pthread *) __hidden; 811void _thr_signal_unblock(struct pthread *) __hidden; 812void _thr_signal_init(int) __hidden; 813void _thr_signal_deinit(void) __hidden; 814int _thr_send_sig(struct pthread *, int sig) __hidden; 815void _thr_list_init(void) __hidden; 816void _thr_hash_add(struct pthread *) __hidden; 817void _thr_hash_remove(struct pthread *) __hidden; 818struct pthread *_thr_hash_find(struct pthread *) __hidden; 819void _thr_link(struct pthread *, struct pthread *) __hidden; 820void _thr_unlink(struct pthread *, struct pthread *) __hidden; 821void _thr_assert_lock_level(void) __hidden __dead2; 822void _thr_ast(struct pthread *) __hidden; 823void _thr_report_creation(struct pthread *curthread, 824 struct pthread *newthread) __hidden; 825void _thr_report_death(struct pthread *curthread) __hidden; 826int _thr_getscheduler(lwpid_t, int *, struct sched_param *) __hidden; 827int _thr_setscheduler(lwpid_t, int, const struct sched_param *) __hidden; 828void _thr_signal_prefork(void) __hidden; 829void _thr_signal_postfork(void) __hidden; 830void _thr_signal_postfork_child(void) __hidden; 831void _thr_suspend_all_lock(struct pthread *) __hidden; 832void _thr_suspend_all_unlock(struct pthread *) __hidden; 833void _thr_try_gc(struct pthread *, struct pthread *) __hidden; 834int _rtp_to_schedparam(const struct rtprio *rtp, int *policy, 835 struct sched_param *param) __hidden; 836int _schedparam_to_rtp(int policy, const struct sched_param *param, 837 struct rtprio *rtp) __hidden; 838void _thread_bp_create(void); 839void _thread_bp_death(void); 840int _sched_yield(void); 841 842void _pthread_cleanup_push(void (*)(void *), void *); 843void _pthread_cleanup_pop(int); 844void _pthread_exit_mask(void *status, sigset_t *mask) __dead2 __hidden; 845#ifndef _LIBC_PRIVATE_H_ 846void _pthread_cancel_enter(int maycancel); 847void _pthread_cancel_leave(int maycancel); 848#endif 849int _pthread_mutex_consistent(pthread_mutex_t * _Nonnull); 850int _pthread_mutexattr_getrobust(pthread_mutexattr_t * _Nonnull __restrict, 851 int * _Nonnull __restrict); 852int _pthread_mutexattr_setrobust(pthread_mutexattr_t * _Nonnull, int); 853 854/* #include <fcntl.h> */ 855#ifdef _SYS_FCNTL_H_ 856#ifndef _LIBC_PRIVATE_H_ 857int __sys_fcntl(int, int, ...); 858int __sys_openat(int, const char *, int, ...); 859#endif /* _LIBC_PRIVATE_H_ */ 860#endif /* _SYS_FCNTL_H_ */ 861 862/* #include <signal.h> */ 863#ifdef _SIGNAL_H_ 864int __sys_kill(pid_t, int); 865int __sys_sigaltstack(const struct sigaltstack *, struct sigaltstack *); 866int __sys_sigpending(sigset_t *); 867int __sys_sigreturn(const ucontext_t *); 868#ifndef _LIBC_PRIVATE_H_ 869int __sys_sigaction(int, const struct sigaction *, struct sigaction *); 870int __sys_sigprocmask(int, const sigset_t *, sigset_t *); 871int __sys_sigsuspend(const sigset_t *); 872int __sys_sigtimedwait(const sigset_t *, siginfo_t *, 873 const struct timespec *); 874int __sys_sigwait(const sigset_t *, int *); 875int __sys_sigwaitinfo(const sigset_t *set, siginfo_t *info); 876#endif /* _LIBC_PRIVATE_H_ */ 877#endif /* _SYS_FCNTL_H_ */ 878 879/* #include <time.h> */ 880#ifdef _TIME_H_ 881#ifndef _LIBC_PRIVATE_H_ 882int __sys_clock_nanosleep(clockid_t, int, const struct timespec *, 883 struct timespec *); 884int __sys_nanosleep(const struct timespec *, struct timespec *); 885#endif /* _LIBC_PRIVATE_H_ */ 886#endif /* _SYS_FCNTL_H_ */ 887 888/* #include <sys/ucontext.h> */ 889#ifdef _SYS_UCONTEXT_H_ 890#ifndef _LIBC_PRIVATE_H_ 891int __sys_setcontext(const ucontext_t *ucp); 892int __sys_swapcontext(ucontext_t *oucp, const ucontext_t *ucp); 893#endif /* _LIBC_PRIVATE_H_ */ 894#endif /* _SYS_FCNTL_H_ */ 895 896/* #include <unistd.h> */ 897#ifdef _UNISTD_H_ 898void __sys_exit(int); 899pid_t __sys_getpid(void); 900#ifndef _LIBC_PRIVATE_H_ 901int __sys_close(int); 902int __sys_fork(void); 903ssize_t __sys_read(int, void *, size_t); 904#endif /* _LIBC_PRIVATE_H_ */ 905#endif /* _SYS_FCNTL_H_ */ 906 907static inline int 908_thr_isthreaded(void) 909{ 910 return (__isthreaded != 0); 911} 912 913static inline int 914_thr_is_inited(void) 915{ 916 return (_thr_initial != NULL); 917} 918 919static inline void 920_thr_check_init(void) 921{ 922 if (_thr_initial == NULL) 923 _libpthread_init(NULL); 924} 925 926struct wake_addr *_thr_alloc_wake_addr(void); 927void _thr_release_wake_addr(struct wake_addr *); 928int _thr_sleep(struct pthread *, int, const struct timespec *); 929 930void _thr_wake_addr_init(void) __hidden; 931 932static inline void 933_thr_clear_wake(struct pthread *td) 934{ 935 td->wake_addr->value = 0; 936} 937 938static inline int 939_thr_is_woken(struct pthread *td) 940{ 941 return td->wake_addr->value != 0; 942} 943 944static inline void 945_thr_set_wake(unsigned int *waddr) 946{ 947 *waddr = 1; 948 _thr_umtx_wake(waddr, INT_MAX, 0); 949} 950 951void _thr_wake_all(unsigned int *waddrs[], int) __hidden; 952 953static inline struct pthread * 954_sleepq_first(struct sleepqueue *sq) 955{ 956 return TAILQ_FIRST(&sq->sq_blocked); 957} 958 959void _sleepq_init(void) __hidden; 960struct sleepqueue *_sleepq_alloc(void) __hidden; 961void _sleepq_free(struct sleepqueue *) __hidden; 962void _sleepq_lock(void *) __hidden; 963void _sleepq_unlock(void *) __hidden; 964struct sleepqueue *_sleepq_lookup(void *) __hidden; 965void _sleepq_add(void *, struct pthread *) __hidden; 966int _sleepq_remove(struct sleepqueue *, struct pthread *) __hidden; 967void _sleepq_drop(struct sleepqueue *, 968 void (*cb)(struct pthread *, void *arg), void *) __hidden; 969 970int _pthread_mutex_init_calloc_cb(pthread_mutex_t *mutex, 971 void *(calloc_cb)(size_t, size_t)); 972 973struct dl_phdr_info; 974void __pthread_cxa_finalize(struct dl_phdr_info *phdr_info); 975void _thr_tsd_unload(struct dl_phdr_info *phdr_info) __hidden; 976void _thr_sigact_unload(struct dl_phdr_info *phdr_info) __hidden; 977void _thr_stack_fix_protection(struct pthread *thrd); 978 979int *__error_threaded(void) __hidden; 980void __thr_interpose_libc(void) __hidden; 981pid_t __thr_fork(void); 982int __thr_setcontext(const ucontext_t *ucp); 983int __thr_sigaction(int sig, const struct sigaction *act, 984 struct sigaction *oact) __hidden; 985int __thr_sigprocmask(int how, const sigset_t *set, sigset_t *oset); 986int __thr_sigsuspend(const sigset_t * set); 987int __thr_sigtimedwait(const sigset_t *set, siginfo_t *info, 988 const struct timespec * timeout); 989int __thr_sigwait(const sigset_t *set, int *sig); 990int __thr_sigwaitinfo(const sigset_t *set, siginfo_t *info); 991int __thr_swapcontext(ucontext_t *oucp, const ucontext_t *ucp); 992 993void __thr_map_stacks_exec(void); 994 995struct _spinlock; 996void __thr_spinunlock(struct _spinlock *lck); 997void __thr_spinlock(struct _spinlock *lck); 998 999struct tcb *_tcb_ctor(struct pthread *, int); 1000void _tcb_dtor(struct tcb *); 1001 1002void __thr_pshared_init(void) __hidden; 1003void *__thr_pshared_offpage(void *key, int doalloc) __hidden; 1004void __thr_pshared_destroy(void *key) __hidden; 1005void __thr_pshared_atfork_pre(void) __hidden; 1006void __thr_pshared_atfork_post(void) __hidden; 1007 1008__END_DECLS 1009__NULLABILITY_PRAGMA_POP 1010 1011#endif /* !_THR_PRIVATE_H */ 1012