24#if !KMP_OS_DRAGONFLY && !KMP_OS_FREEBSD && !KMP_OS_NETBSD && !KMP_OS_OPENBSD
31#include <sys/resource.h>
34#include <libperfstat.h>
36#include <sys/syscall.h>
43#include <sys/sysinfo.h>
59#include <sys/sysctl.h>
60#elif KMP_OS_DRAGONFLY || KMP_OS_FREEBSD
62#include <sys/sysctl.h>
64#include <pthread_np.h>
68#elif KMP_OS_NETBSD || KMP_OS_OPENBSD
70#include <sys/sysctl.h>
78#include <sys/loadavg.h>
89#ifndef TIMEVAL_TO_TIMESPEC
91#define TIMEVAL_TO_TIMESPEC(tv, ts) \
93 (ts)->tv_sec = (tv)->tv_sec; \
94 (ts)->tv_nsec = (tv)->tv_usec * 1000; \
99#define TS2NS(timespec) \
100 (((timespec).tv_sec * (long int)1e9) + (timespec).tv_nsec)
104#if KMP_HANDLE_SIGNALS
105typedef void (*sig_func_t)(
int);
107static sigset_t __kmp_sigset;
124static void __kmp_print_cond(
char *buffer, kmp_cond_align_t *cond) {
125 KMP_SNPRINTF(buffer, 128,
"(cond (lock (%ld, %d)), (descr (%p)))",
126 cond->c_cond.__c_lock.__status, cond->c_cond.__c_lock.__spinlock,
127 cond->c_cond.__c_waiting);
131#if ((KMP_OS_LINUX || KMP_OS_FREEBSD || KMP_OS_NETBSD || KMP_OS_DRAGONFLY || \
133 KMP_AFFINITY_SUPPORTED)
139 "Illegal set affinity operation when not capable");
141 kmp_affin_mask_t *
mask;
142 KMP_CPU_ALLOC_ON_STACK(
mask);
144 KMP_CPU_SET(which,
mask);
145 __kmp_set_system_affinity(
mask,
TRUE);
146 KMP_CPU_FREE_FROM_STACK(
mask);
159 if (mask_size %
sizeof(__kmp_affin_mask_size))
160 mask_size +=
sizeof(__kmp_affin_mask_size) -
161 mask_size %
sizeof(__kmp_affin_mask_size);
162 KMP_AFFINITY_ENABLE(mask_size);
164 (
"__kmp_affinity_determine_capable: "
165 "AIX OS affinity interface bindprocessor functional (mask size = "
166 "%" KMP_SIZE_T_SPEC
").\n",
167 __kmp_affin_mask_size));
179#define KMP_CPU_SET_SIZE_LIMIT (1024 * 1024)
180#define KMP_CPU_SET_TRY_SIZE CACHE_LINE
181#elif KMP_OS_FREEBSD || KMP_OS_DRAGONFLY
182#define KMP_CPU_SET_SIZE_LIMIT (sizeof(cpuset_t))
184#define KMP_CPU_SET_SIZE_LIMIT (256)
187 int verbose = __kmp_affinity.flags.verbose;
188 int warnings = __kmp_affinity.flags.warnings;
189 enum affinity_type
type = __kmp_affinity.type;
198 gCode = syscall(__NR_sched_getaffinity, 0, KMP_CPU_SET_TRY_SIZE,
buf);
199 KA_TRACE(30, (
"__kmp_affinity_determine_capable: "
200 "initial getaffinity call returned %ld errno = %d\n",
203 if (gCode < 0 && errno != EINVAL) {
206 (warnings && (
type != affinity_none) && (
type != affinity_default) &&
207 (
type != affinity_disabled))) {
216 KMP_AFFINITY_DISABLE();
219 }
else if (gCode > 0) {
221 KMP_AFFINITY_ENABLE(gCode);
222 KA_TRACE(10, (
"__kmp_affinity_determine_capable: "
223 "affinity supported (mask size %d)\n",
224 (
int)__kmp_affin_mask_size));
231 KA_TRACE(30, (
"__kmp_affinity_determine_capable: "
232 "searching for proper set size\n"));
234 for (
size = 1;
size <= KMP_CPU_SET_SIZE_LIMIT;
size *= 2) {
235 gCode = syscall(__NR_sched_getaffinity, 0,
size,
buf);
236 KA_TRACE(30, (
"__kmp_affinity_determine_capable: "
237 "getaffinity for mask size %ld returned %ld errno = %d\n",
238 size, gCode, errno));
241 if (errno == ENOSYS) {
243 KA_TRACE(30, (
"__kmp_affinity_determine_capable: "
244 "inconsistent OS call behavior: errno == ENOSYS for mask "
248 (warnings && (
type != affinity_none) &&
249 (
type != affinity_default) && (
type != affinity_disabled))) {
258 KMP_AFFINITY_DISABLE();
265 KMP_AFFINITY_ENABLE(gCode);
266 KA_TRACE(10, (
"__kmp_affinity_determine_capable: "
267 "affinity supported (mask size %d)\n",
268 (
int)__kmp_affin_mask_size));
272#elif KMP_OS_FREEBSD || KMP_OS_NETBSD || KMP_OS_DRAGONFLY
276 gCode = pthread_getaffinity_np(pthread_self(), KMP_CPU_SET_SIZE_LIMIT,
277 reinterpret_cast<cpuset_t *
>(
buf));
278 KA_TRACE(30, (
"__kmp_affinity_determine_capable: "
279 "initial getaffinity call returned %d errno = %d\n",
282 KMP_AFFINITY_ENABLE(KMP_CPU_SET_SIZE_LIMIT);
283 KA_TRACE(10, (
"__kmp_affinity_determine_capable: "
284 "affinity supported (mask size %d)\n",
285 (
int)__kmp_affin_mask_size));
293 KMP_AFFINITY_DISABLE();
294 KA_TRACE(10, (
"__kmp_affinity_determine_capable: "
295 "cannot determine mask size - affinity not supported\n"));
296 if (verbose || (warnings && (
type != affinity_none) &&
297 (
type != affinity_default) && (
type != affinity_disabled))) {
307int __kmp_futex_determine_capable() {
309 long rc = syscall(__NR_futex, &
loc, FUTEX_WAKE, 1, NULL, NULL, 0);
310 int retval = (rc == 0) || (errno != ENOSYS);
313 (
"__kmp_futex_determine_capable: rc = %d errno = %d\n", rc, errno));
314 KA_TRACE(10, (
"__kmp_futex_determine_capable: futex syscall%s supported\n",
315 retval ?
"" :
" not"));
322#if (KMP_ARCH_X86 || KMP_ARCH_X86_64 || KMP_ARCH_WASM) && (!KMP_ASM_INTRINS)
330 new_value = old_value |
d;
335 new_value = old_value |
d;
344 new_value = old_value &
d;
349 new_value = old_value &
d;
358 new_value = old_value |
d;
363 new_value = old_value |
d;
372 new_value = old_value &
d;
377 new_value = old_value &
d;
382#if KMP_ARCH_X86 || KMP_ARCH_WASM
387 new_value = old_value +
d;
392 new_value = old_value +
d;
401 new_value = old_value +
d;
406 new_value = old_value +
d;
416 new_value = old_value |
d;
420 new_value = old_value |
d;
429 new_value = old_value &
d;
433 new_value = old_value &
d;
447#ifdef KMP_CANCEL_THREADS
448 KA_TRACE(10, (
"__kmp_terminate_thread: kill (%d)\n", gtid));
449 status = pthread_cancel(th->th.th_info.ds.ds_thread);
464#if KMP_OS_LINUX || KMP_OS_DRAGONFLY || KMP_OS_FREEBSD || KMP_OS_NETBSD || \
465 KMP_OS_HURD || KMP_OS_SOLARIS || KMP_OS_AIX
477 if ((
status = thr_stksegment(&
s)) < 0) {
483 KA_TRACE(60, (
"__kmp_set_stack_info: T#%d thr_stksegment returned size:"
484 " %lu, low addr: %p\n",
489 status = pthread_attr_init(&attr);
491#if KMP_OS_DRAGONFLY || KMP_OS_FREEBSD || KMP_OS_NETBSD
492 status = pthread_attr_get_np(pthread_self(), &attr);
495 status = pthread_getattr_np(pthread_self(), &attr);
501 (
"__kmp_set_stack_info: T#%d pthread_attr_getstack returned size:"
502 " %lu, low addr: %p\n",
504 status = pthread_attr_destroy(&attr);
519 TCW_PTR(th->th.th_info.ds.ds_stacksize, 0);
520 TCW_PTR(th->th.th_info.ds.ds_stackbase, &stack_data);
521 TCW_4(th->th.th_info.ds.ds_stackgrow,
TRUE);
526 int status, old_type, old_state;
527#ifdef KMP_BLOCK_SIGNALS
528 sigset_t new_set, old_set;
531#if KMP_OS_LINUX || KMP_OS_DRAGONFLY || KMP_OS_FREEBSD || KMP_OS_NETBSD || \
532 KMP_OS_OPENBSD || KMP_OS_HURD || KMP_OS_SOLARIS || KMP_OS_AIX
533 void *
volatile padding = 0;
537 gtid = ((
kmp_info_t *)thr)->th.th_info.ds.ds_gtid;
544 __kmp_stats_thread_ptr = ((
kmp_info_t *)thr)->th.th_stats;
545 __kmp_stats_thread_ptr->startLife();
551 __kmp_itt_thread_name(gtid);
554#if KMP_AFFINITY_SUPPORTED
555 __kmp_affinity_bind_init_mask(gtid);
558#ifdef KMP_CANCEL_THREADS
559 status = pthread_setcanceltype(PTHREAD_CANCEL_ASYNCHRONOUS, &old_type);
562 status = pthread_setcancelstate(PTHREAD_CANCEL_ENABLE, &old_state);
566#if KMP_ARCH_X86 || KMP_ARCH_X86_64
568 __kmp_clear_x87_fpu_status_word();
569 __kmp_load_x87_fpu_control_word(&__kmp_init_x87_fpu_control_word);
570 __kmp_load_mxcsr(&__kmp_init_mxcsr);
573#ifdef KMP_BLOCK_SIGNALS
574 status = sigfillset(&new_set);
576 status = pthread_sigmask(SIG_BLOCK, &new_set, &old_set);
580#if KMP_OS_LINUX || KMP_OS_DRAGONFLY || KMP_OS_FREEBSD || KMP_OS_NETBSD || \
581 KMP_OS_OPENBSD || KMP_OS_HURD || KMP_OS_SOLARIS || KMP_OS_AIX
595#ifdef KMP_BLOCK_SIGNALS
596 status = pthread_sigmask(SIG_SETMASK, &old_set, NULL);
606static void *__kmp_launch_monitor(
void *thr) {
607 int status, old_type, old_state;
608#ifdef KMP_BLOCK_SIGNALS
611 struct timespec interval;
615 KA_TRACE(10, (
"__kmp_launch_monitor: #1 launched\n"));
627 __kmp_itt_thread_ignore();
635#ifdef KMP_CANCEL_THREADS
636 status = pthread_setcanceltype(PTHREAD_CANCEL_ASYNCHRONOUS, &old_type);
639 status = pthread_setcancelstate(PTHREAD_CANCEL_ENABLE, &old_state);
648 int sched = sched_getscheduler(0);
649 if (
sched == SCHED_FIFO ||
sched == SCHED_RR) {
652 struct sched_param param;
653 int max_priority = sched_get_priority_max(
sched);
656 sched_getparam(0, ¶m);
657 if (param.sched_priority < max_priority) {
658 param.sched_priority += 1;
659 rc = sched_setscheduler(0,
sched, ¶m);
685 if (__kmp_monitor_wakeups == 1) {
687 interval.tv_nsec = 0;
693 KA_TRACE(10, (
"__kmp_launch_monitor: #2 monitor\n"));
701 KA_TRACE(15, (
"__kmp_launch_monitor: update\n"));
703 status = gettimeofday(&tval, NULL);
707 now.tv_sec += interval.tv_sec;
708 now.tv_nsec += interval.tv_nsec;
736 KA_TRACE(10, (
"__kmp_launch_monitor: #3 cleanup\n"));
738#ifdef KMP_BLOCK_SIGNALS
739 status = sigfillset(&new_set);
741 status = pthread_sigmask(SIG_UNBLOCK, &new_set, NULL);
745 KA_TRACE(10, (
"__kmp_launch_monitor: #4 finished\n"));
753 KA_TRACE(10, (
"__kmp_launch_monitor: #5 terminate sig=%d\n",
764 KA_TRACE(10, (
"__kmp_launch_monitor: #6 raise sig=%d\n",
771 KA_TRACE(10, (
"__kmp_launch_monitor: #7 exit\n"));
779 pthread_attr_t thread_attr;
782 th->th.th_info.ds.ds_gtid = gtid;
793 th->th.th_stats = __kmp_stats_list->push_back(gtid);
797 th->th.th_stats = __kmp_stats_thread_ptr;
804 KA_TRACE(10, (
"__kmp_create_worker: uber thread (%d)\n", gtid));
805 th->th.th_info.ds.ds_thread = pthread_self();
811 KA_TRACE(10, (
"__kmp_create_worker: try to create thread (%d)\n", gtid));
815#ifdef KMP_THREAD_ATTR
816 status = pthread_attr_init(&thread_attr);
820 status = pthread_attr_setdetachstate(&thread_attr, PTHREAD_CREATE_JOINABLE);
834 KA_TRACE(10, (
"__kmp_create_worker: T#%d, default stacksize = %lu bytes, "
835 "__kmp_stksize = %lu bytes, final stacksize = %lu bytes\n",
838#ifdef _POSIX_THREAD_ATTR_STACKSIZE
839 status = pthread_attr_setstacksize(&thread_attr, stack_size);
840#ifdef KMP_BACKUP_STKSIZE
845 KA_TRACE(10, (
"__kmp_create_worker: T#%d, default stacksize = %lu bytes, "
846 "__kmp_stksize = %lu bytes, (backup) final stacksize = %lu "
849 status = pthread_attr_setstacksize(&thread_attr, stack_size);
863 if (
status != 0 || !handle) {
864#ifdef _POSIX_THREAD_ATTR_STACKSIZE
881 th->th.th_info.ds.ds_thread = handle;
883#ifdef KMP_THREAD_ATTR
884 status = pthread_attr_destroy(&thread_attr);
897 KA_TRACE(10, (
"__kmp_create_worker: done creating thread (%d)\n", gtid));
904 pthread_attr_t thread_attr;
907 int auto_adj_size =
FALSE;
911 KA_TRACE(10, (
"__kmp_create_monitor: skipping monitor thread because of "
913 th->th.th_info.ds.ds_tid = 0;
914 th->th.th_info.ds.ds_gtid = 0;
917 KA_TRACE(10, (
"__kmp_create_monitor: try to create monitor\n"));
930#ifdef KMP_THREAD_ATTR
931 if (__kmp_monitor_stksize == 0) {
932 __kmp_monitor_stksize = KMP_DEFAULT_MONITOR_STKSIZE;
933 auto_adj_size =
TRUE;
935 status = pthread_attr_init(&thread_attr);
939 status = pthread_attr_setdetachstate(&thread_attr, PTHREAD_CREATE_JOINABLE);
944#ifdef _POSIX_THREAD_ATTR_STACKSIZE
945 status = pthread_attr_getstacksize(&thread_attr, &
size);
952 if (__kmp_monitor_stksize == 0) {
953 __kmp_monitor_stksize = KMP_DEFAULT_MONITOR_STKSIZE;
959 KA_TRACE(10, (
"__kmp_create_monitor: default stacksize = %lu bytes,"
960 "requested stacksize = %lu bytes\n",
961 size, __kmp_monitor_stksize));
966#ifdef _POSIX_THREAD_ATTR_STACKSIZE
967 KA_TRACE(10, (
"__kmp_create_monitor: setting stacksize = %lu bytes,",
968 __kmp_monitor_stksize));
969 status = pthread_attr_setstacksize(&thread_attr, __kmp_monitor_stksize);
972 __kmp_monitor_stksize *= 2;
977 KMP_MSG(CantSetMonitorStackSize, (
long int)__kmp_monitor_stksize),
986 pthread_create(&handle, &thread_attr, __kmp_launch_monitor, (
void *)th);
989#ifdef _POSIX_THREAD_ATTR_STACKSIZE
991 if (auto_adj_size && (__kmp_monitor_stksize < (
size_t)0x40000000)) {
992 __kmp_monitor_stksize *= 2;
1012 th->th.th_info.ds.ds_thread = handle;
1014#if KMP_REAL_TIME_FIX
1022#ifdef KMP_THREAD_ATTR
1023 status = pthread_attr_destroy(&thread_attr);
1036 KA_TRACE(10, (
"__kmp_create_monitor: monitor created %#.8lx\n",
1037 th->th.th_info.ds.ds_thread));
1046 pthread_exit((
void *)(intptr_t)exit_status);
1051void __kmp_resume_monitor();
1057 KA_TRACE(10, (
"__kmp_reap_monitor: try to reap monitor thread with handle"
1059 th->th.th_info.ds.ds_thread));
1066 KA_TRACE(10, (
"__kmp_reap_monitor: monitor did not start, returning\n"));
1076 status = pthread_kill(th->th.th_info.ds.ds_thread, 0);
1078 __kmp_resume_monitor();
1080 KA_TRACE(10, (
"__kmp_reap_monitor: try to join with monitor\n"));
1081 status = pthread_join(th->th.th_info.ds.ds_thread, &exit_val);
1082 if (exit_val != th) {
1089 KA_TRACE(10, (
"__kmp_reap_monitor: done reaping monitor thread with handle"
1091 th->th.th_info.ds.ds_thread));
1108 10, (
"__kmp_reap_worker: try to reap T#%d\n", th->th.th_info.ds.ds_gtid));
1110 status = pthread_join(th->th.th_info.ds.ds_thread, &exit_val);
1116 if (exit_val != th) {
1117 KA_TRACE(10, (
"__kmp_reap_worker: worker T#%d did not reap properly, "
1119 th->th.th_info.ds.ds_gtid, exit_val));
1125 KA_TRACE(10, (
"__kmp_reap_worker: done reaping T#%d\n",
1126 th->th.th_info.ds.ds_gtid));
1131#if KMP_HANDLE_SIGNALS
1133static void __kmp_null_handler(
int signo) {
1137static void __kmp_team_handler(
int signo) {
1175static void __kmp_sigaction(
int signum,
const struct sigaction *act,
1176 struct sigaction *oldact) {
1177 int rc = sigaction(signum, act, oldact);
1181static void __kmp_install_one_handler(
int sig, sig_func_t handler_func,
1182 int parallel_init) {
1185 (
"__kmp_install_one_handler( %d, ..., %d )\n", sig, parallel_init));
1186 if (parallel_init) {
1187 struct sigaction new_action;
1188 struct sigaction old_action;
1189 new_action.sa_handler = handler_func;
1190 new_action.sa_flags = 0;
1191 sigfillset(&new_action.sa_mask);
1192 __kmp_sigaction(sig, &new_action, &old_action);
1193 if (old_action.sa_handler == __kmp_sighldrs[sig].sa_handler) {
1194 sigaddset(&__kmp_sigset, sig);
1197 __kmp_sigaction(sig, &old_action, NULL);
1201 __kmp_sigaction(sig, NULL, &__kmp_sighldrs[sig]);
1206static void __kmp_remove_one_handler(
int sig) {
1207 KB_TRACE(60, (
"__kmp_remove_one_handler( %d )\n", sig));
1208 if (sigismember(&__kmp_sigset, sig)) {
1209 struct sigaction old;
1211 __kmp_sigaction(sig, &__kmp_sighldrs[sig], &old);
1212 if ((old.sa_handler != __kmp_team_handler) &&
1213 (old.sa_handler != __kmp_null_handler)) {
1215 KB_TRACE(10, (
"__kmp_remove_one_handler: oops, not our handler, "
1216 "restoring: sig=%d\n",
1218 __kmp_sigaction(sig, &old, NULL);
1220 sigdelset(&__kmp_sigset, sig);
1225void __kmp_install_signals(
int parallel_init) {
1226 KB_TRACE(10, (
"__kmp_install_signals( %d )\n", parallel_init));
1227 if (__kmp_handle_signals || !parallel_init) {
1230 sigemptyset(&__kmp_sigset);
1231 __kmp_install_one_handler(SIGHUP, __kmp_team_handler, parallel_init);
1232 __kmp_install_one_handler(SIGINT, __kmp_team_handler, parallel_init);
1233 __kmp_install_one_handler(SIGQUIT, __kmp_team_handler, parallel_init);
1234 __kmp_install_one_handler(SIGILL, __kmp_team_handler, parallel_init);
1235 __kmp_install_one_handler(SIGABRT, __kmp_team_handler, parallel_init);
1236 __kmp_install_one_handler(SIGFPE, __kmp_team_handler, parallel_init);
1237 __kmp_install_one_handler(SIGBUS, __kmp_team_handler, parallel_init);
1238 __kmp_install_one_handler(SIGSEGV, __kmp_team_handler, parallel_init);
1240 __kmp_install_one_handler(SIGSYS, __kmp_team_handler, parallel_init);
1242 __kmp_install_one_handler(SIGTERM, __kmp_team_handler, parallel_init);
1244 __kmp_install_one_handler(SIGPIPE, __kmp_team_handler, parallel_init);
1249void __kmp_remove_signals(
void) {
1251 KB_TRACE(10, (
"__kmp_remove_signals()\n"));
1252 for (sig = 1; sig < NSIG; ++sig) {
1253 __kmp_remove_one_handler(sig);
1260#ifdef KMP_CANCEL_THREADS
1262 status = pthread_setcancelstate(new_state, &old_state);
1269#ifdef KMP_CANCEL_THREADS
1271 status = pthread_setcancelstate(PTHREAD_CANCEL_DISABLE, old_state);
1298#if KMP_AFFINITY_SUPPORTED
1299#if KMP_OS_LINUX || KMP_OS_FREEBSD || KMP_OS_NETBSD || KMP_OS_DRAGONFLY || \
1303 kmp_set_thread_affinity_mask_initial();
1311 for (kmp_affinity_t *affinity : __kmp_affinities)
1312 *affinity = KMP_AFFINITY_INIT(affinity->env_var);
1313 __kmp_affin_fullMask =
nullptr;
1314 __kmp_affin_origMask =
nullptr;
1319 __kmp_init_monitor = 0;
1328#if !KMP_USE_DYNAMIC_LOCK
1344 KA_TRACE(10, (
"__kmp_atfork_child: checking cache address list %p\n",
1350 KC_TRACE(50, (
"__kmp_atfork_child: zeroing cache at address %p\n",
1412 if (old_value == new_value)
1416 &th->th.th_suspend_init_count, old_value, -1)) {
1423 status = pthread_cond_init(&th->th.th_suspend_cv.c_cond,
1426 status = pthread_mutex_init(&th->th.th_suspend_mx.m_mutex,
1439 status = pthread_cond_destroy(&th->th.th_suspend_cv.c_cond);
1443 status = pthread_mutex_destroy(&th->th.th_suspend_mx.m_mutex);
1447 --th->th.th_suspend_init_count;
1455 return (pthread_mutex_trylock(&th->th.th_suspend_mx.m_mutex) == 0);
1459 int status = pthread_mutex_lock(&th->th.th_suspend_mx.m_mutex);
1464 int status = pthread_mutex_unlock(&th->th.th_suspend_mx.m_mutex);
1475 typename C::flag_t old_spin;
1477 KF_TRACE(30, (
"__kmp_suspend_template: T#%d enter for flag = %p\n", th_gtid,
1484 KF_TRACE(10, (
"__kmp_suspend_template: T#%d setting sleep bit for spin(%p)\n",
1485 th_gtid,
flag->get()));
1489 old_spin =
flag->set_sleeping();
1491 th->th.th_sleep_loc_type =
flag->get_type();
1494 flag->unset_sleeping();
1495 TCW_PTR(th->th.th_sleep_loc, NULL);
1500 KF_TRACE(5, (
"__kmp_suspend_template: T#%d set sleep bit for spin(%p)==%x,"
1502 th_gtid,
flag->get(),
flag->load(), old_spin));
1504 if (
flag->done_check_val(old_spin) ||
flag->done_check()) {
1505 flag->unset_sleeping();
1506 TCW_PTR(th->th.th_sleep_loc, NULL);
1508 KF_TRACE(5, (
"__kmp_suspend_template: T#%d false alarm, reset sleep bit "
1510 th_gtid,
flag->get()));
1515 int deactivated =
FALSE;
1517 while (
flag->is_sleeping()) {
1520 __kmp_suspend_count++;
1521 __kmp_print_cond(buffer, &th->th.th_suspend_cv);
1522 __kmp_printf(
"__kmp_suspend_template: suspending T#%d: %s\n", th_gtid,
1528 th->th.th_active =
FALSE;
1529 if (th->th.th_active_in_pool) {
1530 th->th.th_active_in_pool =
FALSE;
1540#if USE_SUSPEND_TIMEOUT
1541 struct timespec now;
1542 struct timeval tval;
1545 status = gettimeofday(&tval, NULL);
1550 now.tv_sec += msecs / 1000;
1551 now.tv_nsec += (msecs % 1000) * 1000;
1553 KF_TRACE(15, (
"__kmp_suspend_template: T#%d about to perform "
1554 "pthread_cond_timedwait\n",
1556 status = pthread_cond_timedwait(&th->th.th_suspend_cv.c_cond,
1557 &th->th.th_suspend_mx.m_mutex, &now);
1559 KF_TRACE(15, (
"__kmp_suspend_template: T#%d about to perform"
1560 " pthread_cond_wait\n",
1562 status = pthread_cond_wait(&th->th.th_suspend_cv.c_cond,
1563 &th->th.th_suspend_mx.m_mutex);
1572 if (!
flag->is_sleeping() &&
1577 flag->unset_sleeping();
1578 TCW_PTR(th->th.th_sleep_loc, NULL);
1582 if (
status == ETIMEDOUT) {
1583 if (
flag->is_sleeping()) {
1585 (
"__kmp_suspend_template: T#%d timeout wakeup\n", th_gtid));
1587 KF_TRACE(2, (
"__kmp_suspend_template: T#%d timeout wakeup, sleep bit "
1590 TCW_PTR(th->th.th_sleep_loc, NULL);
1593 }
else if (
flag->is_sleeping()) {
1595 (
"__kmp_suspend_template: T#%d spurious wakeup\n", th_gtid));
1602 th->th.th_active =
TRUE;
1603 if (
TCR_4(th->th.th_in_pool)) {
1605 th->th.th_active_in_pool =
TRUE;
1611 TCW_PTR(th->th.th_sleep_loc, NULL);
1619 __kmp_print_cond(buffer, &th->th.th_suspend_cv);
1620 __kmp_printf(
"__kmp_suspend_template: T#%d has awakened: %s\n", th_gtid,
1626 KF_TRACE(30, (
"__kmp_suspend_template: T#%d exit\n", th_gtid));
1629template <
bool C,
bool S>
1633template <
bool C,
bool S>
1637template <
bool C,
bool S>
1666 KF_TRACE(30, (
"__kmp_resume_template: T#%d wants to wakeup T#%d enter\n",
1667 gtid, target_gtid));
1674 if (!
flag ||
flag != th->th.th_sleep_loc) {
1677 flag = (
C *)
CCAST(
void *, th->th.th_sleep_loc);
1683 KF_TRACE(5, (
"__kmp_resume_template: T#%d exiting, thread T#%d already "
1684 "awake: flag(%p)\n",
1685 gtid, target_gtid, (
void *)NULL));
1688 }
else if (
flag->get_type() != th->th.th_sleep_loc_type) {
1693 (
"__kmp_resume_template: T#%d retrying, thread T#%d Mismatch flag(%p), "
1694 "spin(%p) type=%d ptr_type=%d\n",
1695 gtid, target_gtid,
flag,
flag->get(),
flag->get_type(),
1696 th->th.th_sleep_loc_type));
1702 if (!
flag->is_sleeping()) {
1703 KF_TRACE(5, (
"__kmp_resume_template: T#%d exiting, thread T#%d already "
1704 "awake: flag(%p): %u\n",
1705 gtid, target_gtid,
flag->get(), (
unsigned int)
flag->load()));
1711 flag->unset_sleeping();
1712 TCW_PTR(th->th.th_sleep_loc, NULL);
1715 KF_TRACE(5, (
"__kmp_resume_template: T#%d about to wakeup T#%d, reset "
1716 "sleep bit for flag's loc(%p): %u\n",
1717 gtid, target_gtid,
flag->get(), (
unsigned int)
flag->load()));
1722 __kmp_print_cond(buffer, &th->th.th_suspend_cv);
1723 __kmp_printf(
"__kmp_resume_template: T#%d resuming T#%d: %s\n", gtid,
1724 target_gtid, buffer);
1727 status = pthread_cond_signal(&th->th.th_suspend_cv.c_cond);
1730 KF_TRACE(30, (
"__kmp_resume_template: T#%d exiting after signaling wake up"
1732 gtid, target_gtid));
1735template <
bool C,
bool S>
1739template <
bool C,
bool S>
1743template <
bool C,
bool S>
1758void __kmp_resume_monitor() {
1763 KF_TRACE(30, (
"__kmp_resume_monitor: T#%d wants to wakeup T#%d enter\n", gtid,
1773 __kmp_printf(
"__kmp_resume_monitor: T#%d resuming T#%d: %s\n", gtid,
1781 KF_TRACE(30, (
"__kmp_resume_monitor: T#%d exiting after signaling wake up"
1793 (
void *)(intptr_t)(gtid + 1));
1796 KA_TRACE(50, (
"__kmp_gtid_set_specific: runtime shutdown, returning\n"));
1803 KA_TRACE(50, (
"__kmp_gtid_get_specific: runtime shutdown, returning "
1804 "KMP_GTID_SHUTDOWN\n"));
1813 KA_TRACE(50, (
"__kmp_gtid_get_specific: key:%d gtid:%d\n",
1824 return (
double)(buffer.tms_utime + buffer.tms_cutime) /
1825 (
double)CLOCKS_PER_SEC;
1830 struct rusage r_usage;
1832 memset(info, 0,
sizeof(*info));
1834 status = getrusage(RUSAGE_SELF, &r_usage);
1839 info->
maxrss = r_usage.ru_maxrss;
1841 info->
minflt = r_usage.ru_minflt;
1843 info->
majflt = r_usage.ru_majflt;
1845 info->
nswap = r_usage.ru_nswap;
1847 info->
inblock = r_usage.ru_inblock;
1849 info->
oublock = r_usage.ru_oublock;
1851 info->
nvcsw = r_usage.ru_nvcsw;
1853 info->
nivcsw = r_usage.ru_nivcsw;
1861 struct timeval tval;
1862 struct timespec
stop;
1865 status = gettimeofday(&tval, NULL);
1869 *delta = (t_ns * 1e-9);
1873 struct timeval tval;
1875 status = gettimeofday(&tval, NULL);
1888#elif KMP_OS_DRAGONFLY || KMP_OS_FREEBSD || KMP_OS_NETBSD || KMP_OS_OPENBSD || \
1889 KMP_OS_HURD || KMP_OS_SOLARIS || KMP_OS_WASI || KMP_OS_AIX
1895 size_t len =
sizeof(
r);
1896 sysctlbyname(
"hw.logicalcpu", &
r, &len, NULL, 0);
1900#error "Unknown or unsupported OS."
1904 return r > 0 ?
r : 2;
1912 va_start(
args, format);
1913 FILE *
f = fopen(
path,
"rb");
1927 pthread_mutexattr_t mutex_attr;
1928 pthread_condattr_t cond_attr;
1934#if (KMP_ARCH_X86 || KMP_ARCH_X86_64)
1935 if (!__kmp_cpuinfo.initialized) {
1936 __kmp_query_cpuid(&__kmp_cpuinfo);
1946 status = getrlimit(RLIMIT_STACK, &rlim);
1953 if (sysconf(_SC_THREADS)) {
1986 status = pthread_mutexattr_init(&mutex_attr);
1990 status = pthread_mutexattr_destroy(&mutex_attr);
1992 status = pthread_condattr_init(&cond_attr);
1996 status = pthread_condattr_destroy(&cond_attr);
1999 __kmp_itt_initialize();
2013 __kmp_itt_destroy();
2027#if KMP_AFFINITY_SUPPORTED
2028 __kmp_affinity_uninitialize();
2044 status = clock_gettime(CLOCK_PROCESS_CPUTIME_ID, &ts);
2051 status = gettimeofday(&tv, NULL);
2064 gettimeofday(&t, NULL);
2070#if KMP_ARCH_X86 || KMP_ARCH_X86_64
2080 diff = nsec2 - nsec;
2082 double tpus = 1000.0 * (
double)(
delay + (now - goal)) / (
double)diff;
2099#if KMP_OS_LINUX || KMP_OS_HURD
2107 file = fopen(
name,
"r");
2112 void *beginning = NULL;
2113 void *ending = NULL;
2116 rc = fscanf(file,
"%p-%p %4s %*[^\n]\n", &beginning, &ending, perms);
2124 if ((
addr >= beginning) && (
addr < ending)) {
2126 if (strcmp(perms,
"rw") == 0) {
2140 int mib[] = {CTL_KERN, KERN_PROC, KERN_PROC_VMMAP, getpid()};
2141 rc = sysctl(mib, 4, NULL, &lstsz, NULL, 0);
2146 lstsz = lstsz * 4 / 3;
2148 rc = sysctl(mib, 4,
buf, &lstsz, NULL, 0);
2155 char *up =
buf + lstsz;
2158 struct kinfo_vmentry *cur =
reinterpret_cast<struct kinfo_vmentry *
>(lw);
2159 size_t cursz = cur->kve_structsize;
2162 void *start =
reinterpret_cast<void *
>(cur->kve_start);
2163 void *
end =
reinterpret_cast<void *
>(cur->kve_end);
2166 if ((cur->kve_protection & KVME_PROT_READ) != 0 &&
2167 (cur->kve_protection & KVME_PROT_WRITE) != 0) {
2175#elif KMP_OS_DRAGONFLY
2176 char err[_POSIX2_LINE_MAX];
2180 vm_map_entry entry, *c;
2186 fd = kvm_openfiles(
nullptr,
nullptr,
nullptr, O_RDONLY,
err);
2191 proc = kvm_getprocs(fd, KERN_PROC_PID, getpid(), &num);
2193 if (kvm_read(fd,
static_cast<uintptr_t
>(proc->kp_paddr), &
p,
sizeof(
p)) !=
2195 kvm_read(fd,
reinterpret_cast<uintptr_t
>(
p.p_vmspace), &sp,
sizeof(sp)) !=
2203 uaddr =
reinterpret_cast<uintptr_t
>(
addr);
2204 for (c = kvm_vm_map_entry_first(fd, cur, &entry); c;
2205 c = kvm_vm_map_entry_next(fd, c, &entry)) {
2206 if ((uaddr >= entry.ba.start) && (uaddr <= entry.ba.end)) {
2207 if ((entry.protection & VM_PROT_READ) != 0 &&
2208 (entry.protection & VM_PROT_WRITE) != 0) {
2224 pid_t pid = getpid();
2225 struct ps_prochandle *fd = Pgrab(pid, PGRAB_RDONLY, &
err);
2233 size_t sz = (1 << 20);
2234 file = open(
name, O_RDONLY);
2242 while (sz > 0 && (
rd = pread(file,
buf, sz, 0)) == sz) {
2249 map =
reinterpret_cast<prmap_t *
>(
buf);
2250 uaddr =
reinterpret_cast<uintptr_t
>(
addr);
2252 for (cur = map;
rd > 0; cur++,
rd = -
sizeof(*map)) {
2253 if ((uaddr >= cur->pr_vaddr) && (uaddr < cur->pr_vaddr)) {
2254 if ((cur->pr_mflags & MA_READ) != 0 && (cur->pr_mflags & MA_WRITE) != 0) {
2271 rc = vm_read_overwrite(
2273 (vm_address_t)(
addr),
2275 (vm_address_t)(&buffer),
2288 mib[2] = VM_PROC_MAP;
2290 mib[4] =
sizeof(
struct kinfo_vmentry);
2293 rc = sysctl(mib, __arraycount(mib), NULL, &
size, NULL, 0);
2301 rc = sysctl(mib, __arraycount(mib), kiv, &
size, NULL, 0);
2305 for (
size_t i = 0;
i <
size;
i++) {
2306 if (kiv[
i].kve_start >= (uint64_t)
addr &&
2307 kiv[
i].kve_end <= (uint64_t)
addr) {
2317 mib[1] = KERN_PROC_VMMAP;
2322 rc = sysctl(mib, 3, NULL, &
size, NULL, 0);
2327 struct kinfo_vmentry kiv = {.kve_start = 0};
2329 while ((rc = sysctl(mib, 3, &kiv, &
size, NULL, 0)) == 0) {
2331 if (kiv.kve_end ==
end)
2334 if (kiv.kve_start >= (uint64_t)
addr && kiv.kve_end <= (uint64_t)
addr) {
2341 found = (
int)
addr < (__builtin_wasm_memory_size(0) * PAGESIZE);
2344 uint32_t loadQueryBufSize = 4096u;
2349 if (loadQueryBuf == NULL) {
2353 rc = loadquery(L_GETXINFO | L_IGNOREUNLOAD, loadQueryBuf, loadQueryBufSize);
2356 if (errno != ENOMEM) {
2360 loadQueryBufSize <<= 1;
2367 struct ld_xinfo *curLdInfo = (
struct ld_xinfo *)loadQueryBuf;
2371 uintptr_t curDataStart = (uintptr_t)curLdInfo->ldinfo_dataorg;
2372 uintptr_t curDataEnd = curDataStart + curLdInfo->ldinfo_datasize;
2375 if (curDataStart <= (uintptr_t)
addr && (uintptr_t)
addr < curDataEnd) {
2379 if (curLdInfo->ldinfo_next == 0u) {
2383 curLdInfo = (
struct ld_xinfo *)((
char *)curLdInfo + curLdInfo->ldinfo_next);
2389#error "Unknown or unsupported OS"
2397#ifdef USE_LOAD_BALANCE
2399#if KMP_OS_DARWIN || KMP_OS_DRAGONFLY || KMP_OS_FREEBSD || KMP_OS_NETBSD || \
2400 KMP_OS_OPENBSD || KMP_OS_SOLARIS
2411 int res = getloadavg(averages, 3);
2416 if (__kmp_load_balance_interval < 180 && (
res >= 1)) {
2417 ret_avg = (
int)averages[0];
2418 }
else if ((__kmp_load_balance_interval >= 180 &&
2419 __kmp_load_balance_interval < 600) &&
2421 ret_avg = (
int)averages[1];
2422 }
else if ((__kmp_load_balance_interval >= 600) && (
res == 3)) {
2423 ret_avg = (
int)averages[2];
2437 static int glb_running_threads = 0;
2439 static double glb_call_time = 0;
2440 int running_threads = 0;
2442 double call_time = 0.0;
2446 if (glb_call_time &&
2447 (call_time - glb_call_time < __kmp_load_balance_interval))
2448 return glb_running_threads;
2450 glb_call_time = call_time;
2457 int logical_cpus = perfstat_cpu(NULL, NULL,
sizeof(perfstat_cpu_t), 0);
2458 if (logical_cpus <= 0) {
2464 logical_cpus *
sizeof(perfstat_cpu_t));
2465 if (cpu_stat == NULL) {
2472 perfstat_id_t first_cpu_name;
2473 strcpy(first_cpu_name.name, FIRST_CPU);
2476 int rc = perfstat_cpu(&first_cpu_name, cpu_stat,
sizeof(perfstat_cpu_t),
2484 for (
int i = 0;
i < logical_cpus; ++
i) {
2485 running_threads += cpu_stat[
i].runque;
2486 if (running_threads >= max)
2494 if (running_threads <= 0)
2495 running_threads = 1;
2499 glb_running_threads = running_threads;
2501 return running_threads;
2511 static int permanent_error = 0;
2512 static int glb_running_threads = 0;
2514 static double glb_call_time = 0;
2516 int running_threads = 0;
2518 DIR *proc_dir = NULL;
2519 struct dirent *proc_entry = NULL;
2522 DIR *task_dir = NULL;
2524 int task_path_fixed_len;
2528 int stat_path_fixed_len;
2531 int total_processes = 0;
2534 double call_time = 0.0;
2541 if (glb_call_time &&
2542 (call_time - glb_call_time < __kmp_load_balance_interval)) {
2543 running_threads = glb_running_threads;
2547 glb_call_time = call_time;
2550 if (permanent_error) {
2551 running_threads = -1;
2560 proc_dir = opendir(
"/proc");
2561 if (proc_dir == NULL) {
2564 running_threads = -1;
2565 permanent_error = 1;
2571 task_path_fixed_len = task_path.
used;
2573 proc_entry = readdir(proc_dir);
2574 while (proc_entry != NULL) {
2577 if (proc_entry->d_type == DT_DIR && isdigit(proc_entry->d_name[0])) {
2588 strcmp(proc_entry->d_name,
"1") == 0);
2591 task_path.
used = task_path_fixed_len;
2596 task_dir = opendir(task_path.
str);
2597 if (task_dir == NULL) {
2606 if (strcmp(proc_entry->d_name,
"1") == 0) {
2607 running_threads = -1;
2608 permanent_error = 1;
2616 stat_path_fixed_len = stat_path.
used;
2621 if (proc_entry->d_type == DT_DIR && isdigit(
task_entry->d_name[0])) {
2628 stat_path_fixed_len;
2635 stat_file = open(stat_path.
str, O_RDONLY);
2636 if (stat_file == -1) {
2666 len = read(stat_file, buffer,
sizeof(buffer) - 1);
2673 char *close_parent = strstr(buffer,
") ");
2674 if (close_parent != NULL) {
2675 char state = *(close_parent + 2);
2678 if (running_threads >= max) {
2694 proc_entry = readdir(proc_dir);
2701 if (running_threads <= 0) {
2702 running_threads = 1;
2706 if (proc_dir != NULL) {
2710 if (task_dir != NULL) {
2714 if (stat_file != -1) {
2718 glb_running_threads = running_threads;
2720 return running_threads;
2728#if !(KMP_ARCH_X86 || KMP_ARCH_X86_64 || KMP_MIC || \
2729 ((KMP_OS_LINUX || KMP_OS_DARWIN) && KMP_ARCH_AARCH64) || \
2730 KMP_ARCH_PPC64 || KMP_ARCH_RISCV64 || KMP_ARCH_LOONGARCH64 || \
2731 KMP_ARCH_ARM || KMP_ARCH_VE || KMP_ARCH_S390X || KMP_ARCH_PPC_XCOFF || \
2732 KMP_ARCH_AARCH64_32)
2748 void *,
void *,
void *);
2750 void *,
void *,
void *,
void *);
2752 void *,
void *,
void *,
void *,
void *);
2754 void *,
void *,
void *,
void *,
void *,
void *);
2756 void *,
void *,
void *,
void *,
void *,
void *,
2759 void *,
void *,
void *,
void *,
void *,
void *,
2762 void *,
void *,
void *,
void *,
void *,
void *,
2763 void *,
void *,
void *);
2765 void *,
void *,
void *,
void *,
void *,
void *,
2766 void *,
void *,
void *,
void *);
2768 void *,
void *,
void *,
void *,
void *,
void *,
2769 void *,
void *,
void *,
void *,
void *);
2777 void **exit_frame_ptr
2786 fprintf(stderr,
"Too many args to microtask: %d!\n", argc);
2796 (*(
microtask_t2)pkfn)(>id, &tid, p_argv[0], p_argv[1]);
2799 (*(
microtask_t3)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2]);
2802 (*(
microtask_t4)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2806 (*(
microtask_t5)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2807 p_argv[3], p_argv[4]);
2810 (*(
microtask_t6)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2811 p_argv[3], p_argv[4], p_argv[5]);
2814 (*(
microtask_t7)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2815 p_argv[3], p_argv[4], p_argv[5], p_argv[6]);
2818 (*(
microtask_t8)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2819 p_argv[3], p_argv[4], p_argv[5], p_argv[6],
2823 (*(
microtask_t9)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2824 p_argv[3], p_argv[4], p_argv[5], p_argv[6], p_argv[7],
2828 (*(
microtask_t10)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2829 p_argv[3], p_argv[4], p_argv[5], p_argv[6],
2830 p_argv[7], p_argv[8], p_argv[9]);
2833 (*(
microtask_t11)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2834 p_argv[3], p_argv[4], p_argv[5], p_argv[6],
2835 p_argv[7], p_argv[8], p_argv[9], p_argv[10]);
2838 (*(
microtask_t12)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2839 p_argv[3], p_argv[4], p_argv[5], p_argv[6],
2840 p_argv[7], p_argv[8], p_argv[9], p_argv[10],
2844 (*(
microtask_t13)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2845 p_argv[3], p_argv[4], p_argv[5], p_argv[6],
2846 p_argv[7], p_argv[8], p_argv[9], p_argv[10],
2847 p_argv[11], p_argv[12]);
2850 (*(
microtask_t14)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2851 p_argv[3], p_argv[4], p_argv[5], p_argv[6],
2852 p_argv[7], p_argv[8], p_argv[9], p_argv[10],
2853 p_argv[11], p_argv[12], p_argv[13]);
2856 (*(
microtask_t15)pkfn)(>id, &tid, p_argv[0], p_argv[1], p_argv[2],
2857 p_argv[3], p_argv[4], p_argv[5], p_argv[6],
2858 p_argv[7], p_argv[8], p_argv[9], p_argv[10],
2859 p_argv[11], p_argv[12], p_argv[13], p_argv[14]);
2872pthread_cond_t hidden_helper_threads_initz_cond_var;
2873pthread_mutex_t hidden_helper_threads_initz_lock;
2874volatile int hidden_helper_initz_signaled =
FALSE;
2877pthread_cond_t hidden_helper_threads_deinitz_cond_var;
2878pthread_mutex_t hidden_helper_threads_deinitz_lock;
2879volatile int hidden_helper_deinitz_signaled =
FALSE;
2882pthread_cond_t hidden_helper_main_thread_cond_var;
2883pthread_mutex_t hidden_helper_main_thread_lock;
2884volatile int hidden_helper_main_thread_signaled =
FALSE;
2889sem_t hidden_helper_task_sem;
2893 int status = sem_wait(&hidden_helper_task_sem);
2900 pthread_cond_init(&hidden_helper_threads_initz_cond_var,
nullptr);
2903 status = pthread_cond_init(&hidden_helper_threads_deinitz_cond_var,
nullptr);
2906 status = pthread_cond_init(&hidden_helper_main_thread_cond_var,
nullptr);
2909 status = pthread_mutex_init(&hidden_helper_threads_initz_lock,
nullptr);
2912 status = pthread_mutex_init(&hidden_helper_threads_deinitz_lock,
nullptr);
2915 status = pthread_mutex_init(&hidden_helper_main_thread_lock,
nullptr);
2919 status = sem_init(&hidden_helper_task_sem, 0, 0);
2926 [](
void *) ->
void * {
2937 int status = pthread_mutex_lock(&hidden_helper_threads_initz_lock);
2940 if (!
TCR_4(hidden_helper_initz_signaled)) {
2941 status = pthread_cond_wait(&hidden_helper_threads_initz_cond_var,
2942 &hidden_helper_threads_initz_lock);
2946 status = pthread_mutex_unlock(&hidden_helper_threads_initz_lock);
2952 int status = pthread_mutex_lock(&hidden_helper_threads_initz_lock);
2955 status = pthread_cond_signal(&hidden_helper_threads_initz_cond_var);
2960 status = pthread_mutex_unlock(&hidden_helper_threads_initz_lock);
2967 int status = pthread_mutex_lock(&hidden_helper_main_thread_lock);
2970 if (!
TCR_4(hidden_helper_main_thread_signaled)) {
2971 status = pthread_cond_wait(&hidden_helper_main_thread_cond_var,
2972 &hidden_helper_main_thread_lock);
2976 status = pthread_mutex_unlock(&hidden_helper_main_thread_lock);
2983 int status = pthread_mutex_lock(&hidden_helper_main_thread_lock);
2986 status = pthread_cond_signal(&hidden_helper_main_thread_cond_var);
2992 status = pthread_mutex_unlock(&hidden_helper_main_thread_lock);
2997 int status = sem_post(&hidden_helper_task_sem);
3004 int status = pthread_mutex_lock(&hidden_helper_threads_deinitz_lock);
3007 if (!
TCR_4(hidden_helper_deinitz_signaled)) {
3008 status = pthread_cond_wait(&hidden_helper_threads_deinitz_cond_var,
3009 &hidden_helper_threads_deinitz_lock);
3013 status = pthread_mutex_unlock(&hidden_helper_threads_deinitz_lock);
3018 int status = pthread_mutex_lock(&hidden_helper_threads_deinitz_lock);
3021 status = pthread_cond_signal(&hidden_helper_threads_deinitz_cond_var);
3026 status = pthread_mutex_unlock(&hidden_helper_threads_deinitz_lock);
3031 KMP_ASSERT(0 &&
"Hidden helper task is not supported on this OS");
3035 KMP_ASSERT(0 &&
"Hidden helper task is not supported on this OS");
3039 KMP_ASSERT(0 &&
"Hidden helper task is not supported on this OS");
3043 KMP_ASSERT(0 &&
"Hidden helper task is not supported on this OS");
3047 KMP_ASSERT(0 &&
"Hidden helper task is not supported on this OS");
3051 KMP_ASSERT(0 &&
"Hidden helper task is not supported on this OS");
3055 KMP_ASSERT(0 &&
"Hidden helper task is not supported on this OS");
3059 KMP_ASSERT(0 &&
"Hidden helper task is not supported on this OS");
3063 KMP_ASSERT(0 &&
"Hidden helper task is not supported on this OS");
3068 DIR *dir = opendir(
"/dev/shm");
3072 }
else if (ENOENT == errno) {
3080 DIR *dir = opendir(
"/tmp");
3084 }
else if (ENOENT == errno) {
int task_entry(kmp_int32 gtid, kmp_task_t *task)
__itt_string_handle * name
void const char const char int ITT_FORMAT __itt_group_sync s
void const char const char int ITT_FORMAT __itt_group_sync x void const char ITT_FORMAT __itt_group_sync s void ITT_FORMAT __itt_group_sync p void ITT_FORMAT p void ITT_FORMAT p no args __itt_suppress_mode_t unsigned int mask
void const char const char int ITT_FORMAT __itt_group_sync x void const char ITT_FORMAT __itt_group_sync s void ITT_FORMAT __itt_group_sync p void ITT_FORMAT p void ITT_FORMAT p no args __itt_suppress_mode_t unsigned int void size_t ITT_FORMAT d void ITT_FORMAT p void ITT_FORMAT p __itt_model_site __itt_model_site_instance ITT_FORMAT p __itt_model_task __itt_model_task_instance ITT_FORMAT p void ITT_FORMAT p void ITT_FORMAT p void size_t ITT_FORMAT d void ITT_FORMAT p const wchar_t ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s no args void ITT_FORMAT p size_t ITT_FORMAT d no args const wchar_t const wchar_t ITT_FORMAT s __itt_heap_function void size_t int ITT_FORMAT d __itt_heap_function void ITT_FORMAT p __itt_heap_function void void size_t int ITT_FORMAT d no args no args unsigned int ITT_FORMAT u const __itt_domain __itt_id ITT_FORMAT lu const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain __itt_id ITT_FORMAT p const __itt_domain __itt_id __itt_timestamp __itt_timestamp end
void const char const char int ITT_FORMAT __itt_group_sync x void const char ITT_FORMAT __itt_group_sync s void ITT_FORMAT __itt_group_sync p void ITT_FORMAT p void ITT_FORMAT p no args __itt_suppress_mode_t unsigned int void size_t ITT_FORMAT d
void const char const char int ITT_FORMAT __itt_group_sync x void const char ITT_FORMAT __itt_group_sync s void ITT_FORMAT __itt_group_sync p void ITT_FORMAT p void ITT_FORMAT p no args __itt_suppress_mode_t unsigned int void size_t ITT_FORMAT d void ITT_FORMAT p void ITT_FORMAT p __itt_model_site __itt_model_site_instance ITT_FORMAT p __itt_model_task __itt_model_task_instance ITT_FORMAT p void ITT_FORMAT p void ITT_FORMAT p void size_t ITT_FORMAT d void ITT_FORMAT p const wchar_t ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s no args void ITT_FORMAT p size_t ITT_FORMAT d no args const wchar_t const wchar_t ITT_FORMAT s __itt_heap_function void size_t int ITT_FORMAT d __itt_heap_function void ITT_FORMAT p __itt_heap_function void void size_t int ITT_FORMAT d no args no args unsigned int ITT_FORMAT u const __itt_domain __itt_id ITT_FORMAT lu const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain __itt_id ITT_FORMAT p const __itt_domain __itt_id __itt_timestamp __itt_timestamp ITT_FORMAT lu const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain ITT_FORMAT p const __itt_domain __itt_string_handle unsigned long long ITT_FORMAT lu const __itt_domain __itt_string_handle unsigned long long ITT_FORMAT lu const __itt_domain __itt_id __itt_string_handle __itt_metadata_type size_t void ITT_FORMAT p const __itt_domain __itt_id __itt_string_handle const wchar_t size_t ITT_FORMAT lu const __itt_domain __itt_id __itt_relation __itt_id ITT_FORMAT p const wchar_t int ITT_FORMAT __itt_group_mark d __itt_event ITT_FORMAT __itt_group_mark d void const wchar_t const wchar_t int ITT_FORMAT __itt_group_sync __itt_group_fsync x void const wchar_t int const wchar_t int int ITT_FORMAT __itt_group_sync __itt_group_fsync x void ITT_FORMAT __itt_group_sync __itt_group_fsync p void ITT_FORMAT __itt_group_sync __itt_group_fsync p void size_t ITT_FORMAT lu no args __itt_obj_prop_t __itt_obj_state_t ITT_FORMAT d const char ITT_FORMAT s const char ITT_FORMAT s __itt_frame ITT_FORMAT p __itt_counter ITT_FORMAT p __itt_counter unsigned long long ITT_FORMAT lu __itt_counter unsigned long long ITT_FORMAT lu __itt_counter __itt_clock_domain unsigned long long void ITT_FORMAT p const wchar_t ITT_FORMAT S __itt_mark_type const wchar_t ITT_FORMAT S __itt_mark_type const char ITT_FORMAT s __itt_mark_type ITT_FORMAT d __itt_caller ITT_FORMAT p __itt_caller ITT_FORMAT p no args const __itt_domain __itt_clock_domain unsigned long long __itt_id ITT_FORMAT lu const __itt_domain __itt_clock_domain unsigned long long __itt_id __itt_id void ITT_FORMAT p const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain __itt_id ITT_FORMAT lu const __itt_domain __itt_clock_domain unsigned long long __itt_id __itt_string_handle __itt_scope ITT_FORMAT d const __itt_domain __itt_scope __itt_string_handle const char size_t ITT_FORMAT lu const __itt_domain __itt_clock_domain unsigned long long __itt_relation __itt_id ITT_FORMAT lu __itt_track_group __itt_string_handle __itt_track_group_type ITT_FORMAT d __itt_track ITT_FORMAT p void int const int int const char int ITT_FORMAT d void void const char * path
void const char const char int ITT_FORMAT __itt_group_sync x void const char ITT_FORMAT __itt_group_sync s void ITT_FORMAT __itt_group_sync p void ITT_FORMAT p void ITT_FORMAT p no args __itt_suppress_mode_t unsigned int void size_t ITT_FORMAT d void ITT_FORMAT p void ITT_FORMAT p __itt_model_site __itt_model_site_instance ITT_FORMAT p __itt_model_task __itt_model_task_instance ITT_FORMAT p void ITT_FORMAT p void ITT_FORMAT p void size_t ITT_FORMAT d void ITT_FORMAT p const wchar_t ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s no args void ITT_FORMAT p size_t count
void const char const char int ITT_FORMAT __itt_group_sync x void const char ITT_FORMAT __itt_group_sync s void ITT_FORMAT __itt_group_sync p void ITT_FORMAT p void ITT_FORMAT p no args __itt_suppress_mode_t unsigned int void size_t size
void const char const char int ITT_FORMAT __itt_group_sync p
void const char const char int ITT_FORMAT __itt_group_sync x void const char ITT_FORMAT __itt_group_sync s void ITT_FORMAT __itt_group_sync p void ITT_FORMAT p void ITT_FORMAT p no args __itt_suppress_mode_t unsigned int void size_t ITT_FORMAT d void ITT_FORMAT p void ITT_FORMAT p __itt_model_site __itt_model_site_instance ITT_FORMAT p __itt_model_task __itt_model_task_instance ITT_FORMAT p void ITT_FORMAT p void ITT_FORMAT p void size_t ITT_FORMAT d void ITT_FORMAT p const wchar_t ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s no args void ITT_FORMAT p size_t ITT_FORMAT d no args const wchar_t const wchar_t ITT_FORMAT s __itt_heap_function void size_t int ITT_FORMAT d __itt_heap_function void ITT_FORMAT p __itt_heap_function void void size_t int ITT_FORMAT d no args no args unsigned int ITT_FORMAT u const __itt_domain __itt_id ITT_FORMAT lu const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain __itt_id ITT_FORMAT p const __itt_domain __itt_id __itt_timestamp __itt_timestamp ITT_FORMAT lu const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain ITT_FORMAT p const __itt_domain __itt_string_handle unsigned long long ITT_FORMAT lu const __itt_domain __itt_string_handle unsigned long long ITT_FORMAT lu const __itt_domain __itt_id __itt_string_handle __itt_metadata_type size_t void ITT_FORMAT p const __itt_domain __itt_id __itt_string_handle const wchar_t size_t ITT_FORMAT lu const __itt_domain __itt_id __itt_relation __itt_id ITT_FORMAT p const wchar_t int ITT_FORMAT __itt_group_mark d int
void const char const char int ITT_FORMAT __itt_group_sync x void const char ITT_FORMAT __itt_group_sync s void ITT_FORMAT __itt_group_sync p void ITT_FORMAT p void ITT_FORMAT p no args __itt_suppress_mode_t unsigned int void size_t ITT_FORMAT d void ITT_FORMAT p void ITT_FORMAT p __itt_model_site __itt_model_site_instance ITT_FORMAT p __itt_model_task __itt_model_task_instance ITT_FORMAT p void ITT_FORMAT p void ITT_FORMAT p void size_t ITT_FORMAT d void ITT_FORMAT p const wchar_t ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s no args void ITT_FORMAT p size_t ITT_FORMAT d no args const wchar_t const wchar_t ITT_FORMAT s __itt_heap_function void size_t int ITT_FORMAT d __itt_heap_function void ITT_FORMAT p __itt_heap_function void void size_t int ITT_FORMAT d no args no args unsigned int ITT_FORMAT u const __itt_domain __itt_id ITT_FORMAT lu const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain __itt_id ITT_FORMAT p const __itt_domain __itt_id __itt_timestamp __itt_timestamp ITT_FORMAT lu const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain ITT_FORMAT p const __itt_domain __itt_string_handle unsigned long long ITT_FORMAT lu const __itt_domain __itt_string_handle unsigned long long ITT_FORMAT lu const __itt_domain __itt_id __itt_string_handle __itt_metadata_type type
volatile kmp_team_t * __kmp_team_pool
int __kmp_generate_warnings
volatile int __kmp_init_user_locks
#define KMP_INTERNAL_MALLOC(sz)
kmp_bootstrap_lock_t __kmp_initz_lock
kmp_global_t __kmp_global
kmp_pause_status_t __kmp_pause_status
#define KMP_MAX_BLOCKTIME
void __kmp_check_stack_overlap(kmp_info_t *thr)
#define KMP_INTERNAL_REALLOC(p, sz)
volatile kmp_info_t * __kmp_thread_pool
volatile int __kmp_init_gtid
kmp_bootstrap_lock_t __kmp_task_team_lock
kmp_nested_proc_bind_t __kmp_nested_proc_bind
#define KMP_GTID_SHUTDOWN
kmp_cached_addr_t * __kmp_threadpriv_cache_list
volatile int __kmp_all_nth
void __kmp_check_stksize(size_t *val)
volatile int __kmp_init_common
static bool KMP_UBER_GTID(int gtid)
#define KMP_DEFAULT_STKSIZE
volatile int __kmp_init_middle
std::atomic< int > __kmp_thread_pool_active_nth
void __kmp_hidden_helper_threads_initz_routine()
volatile int __kmp_need_register_serial
kmp_bootstrap_lock_t __kmp_forkjoin_lock
void * __kmp_launch_thread(kmp_info_t *thr)
kmp_info_t ** __kmp_threads
int __kmp_need_register_atfork
void __kmp_internal_end_dest(void *)
void __kmp_unregister_library(void)
volatile int __kmp_init_parallel
kmp_key_t __kmp_gtid_threadprivate_key
kmp_info_t * __kmp_thread_pool_insert_pt
kmp_uint32 __kmp_neq_4(kmp_uint32 value, kmp_uint32 checker)
size_t __kmp_sys_min_stksize
#define KMP_INTERNAL_FREE(p)
int __kmp_threads_capacity
void __kmp_serial_initialize(void)
kmp_uint32 __kmp_wait_4(kmp_uint32 volatile *spinner, kmp_uint32 checker, kmp_uint32(*pred)(kmp_uint32, kmp_uint32), void *obj)
volatile int __kmp_init_serial
static void __kmp_type_convert(T1 src, T2 *dest)
union KMP_ALIGN_CACHE kmp_info kmp_info_t
kmp_uint64 __kmp_hardware_timestamp(void)
kmp_topology_t * __kmp_topology
KMP_ARCH_X86 KMP_ARCH_X86 long double
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 kmp_int8
void __kmp_dump_debug_buffer(void)
void __kmp_debug_printf(char const *format,...)
#define KMP_DEBUG_ASSERT(cond)
#define KMP_ASSERT2(cond, msg)
unsigned long long kmp_uint64
static volatile kmp_i18n_cat_status_t status
static void __kmp_msg(kmp_msg_severity_t severity, kmp_msg_t message, va_list ap)
void __kmp_fatal(kmp_msg_t message,...)
#define KMP_CHECK_SYSFAIL(func, error)
#define KMP_CHECK_SYSFAIL_ERRNO(func, status)
#define KMP_SYSFAIL(func, error)
kmp_bootstrap_lock_t __kmp_console_lock
kmp_bootstrap_lock_t __kmp_stdio_lock
void __kmp_printf(char const *format,...)
int __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid)
kmp_block_of_locks * __kmp_lock_blocks
int __kmp_release_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid)
kmp_lock_table_t __kmp_user_lock_table
static void __kmp_release_bootstrap_lock(kmp_bootstrap_lock_t *lck)
static int __kmp_acquire_bootstrap_lock(kmp_bootstrap_lock_t *lck)
static void __kmp_init_bootstrap_lock(kmp_bootstrap_lock_t *lck)
void(* microtask_t)(int *gtid, int *npr,...)
#define KMP_COMPARE_AND_STORE_REL64(p, cv, sv)
#define KMP_COMPARE_AND_STORE_REL32(p, cv, sv)
#define KMP_ATOMIC_ST_REL(p, v)
bool __kmp_atomic_compare_store(std::atomic< T > *p, T expected, T desired)
#define KMP_ATOMIC_LD_ACQ(p)
#define STATIC_EFI2_WORKAROUND
#define KMP_COMPARE_AND_STORE_REL8(p, cv, sv)
#define KMP_ATOMIC_DEC(p)
#define KMP_ATOMIC_LD_RLX(p)
#define KMP_ATOMIC_INC(p)
Functions for collecting statistics.
#define KMP_INIT_PARTITIONED_TIMERS(name)
#define KMP_TIME_DEVELOPER_PARTITIONED_BLOCK(n)
#define KMP_SET_THREAD_STATE(state_name)
void __kmp_str_buf_clear(kmp_str_buf_t *buffer)
void __kmp_str_buf_free(kmp_str_buf_t *buffer)
char * __kmp_str_format(char const *format,...)
void __kmp_str_buf_cat(kmp_str_buf_t *buffer, char const *str, size_t len)
void __kmp_str_free(char **str)
#define __kmp_str_buf_init(b)
static void __kmp_null_resume_wrapper(kmp_info_t *thr)
#define OMPT_GET_FRAME_ADDRESS(level)
struct kmp_cached_addr * next
kmp_lock_index_t allocated
kmp_proc_bind_t * bind_types
void(* microtask_t13)(int *, int *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *)
static void __kmp_atfork_child(void)
void __kmp_hidden_helper_worker_thread_signal()
template void __kmp_atomic_resume_64< false, true >(int, kmp_atomic_flag_64< false, true > *)
template void __kmp_suspend_64< true, false >(int, kmp_flag_64< true, false > *)
void(* microtask_t5)(int *, int *, void *, void *, void *, void *, void *)
static kmp_mutex_align_t __kmp_wait_mx
void __kmp_read_system_time(double *delta)
void(* microtask_t0)(int *, int *)
void __kmp_hidden_helper_threads_initz_wait()
static void __kmp_suspend_template(int th_gtid, C *flag)
void __kmp_enable(int new_state)
void __kmp_unlock_suspend_mx(kmp_info_t *th)
void(* microtask_t7)(int *, int *, void *, void *, void *, void *, void *, void *, void *)
void __kmp_do_initialize_hidden_helper_threads()
static pthread_mutexattr_t __kmp_suspend_mutex_attr
void(* microtask_t1)(int *, int *, void *)
void __kmp_thread_sleep(int millis)
static kmp_int32 __kmp_set_stack_info(int gtid, kmp_info_t *th)
void __kmp_suspend_64(int th_gtid, kmp_flag_64< C, S > *flag)
static int __kmp_init_runtime
static void __kmp_atfork_parent(void)
void __kmp_hidden_helper_main_thread_release()
template void __kmp_resume_32< false, false >(int, kmp_flag_32< false, false > *)
void __kmp_suspend_initialize(void)
kmp_uint64 __kmp_now_nsec()
kmp_uint64 __kmp_ticks_per_usec
int __kmp_is_address_mapped(void *addr)
#define TIMEVAL_TO_TIMESPEC(tv, ts)
double __kmp_read_cpu_time(void)
void(* microtask_t12)(int *, int *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *)
void __kmp_suspend_32(int th_gtid, kmp_flag_32< C, S > *flag)
void __kmp_hidden_helper_worker_thread_wait()
template void __kmp_resume_32< false, true >(int, kmp_flag_32< false, true > *)
void __kmp_disable(int *old_state)
static pthread_condattr_t __kmp_suspend_cond_attr
void __kmp_reap_monitor(kmp_info_t *th)
template void __kmp_atomic_suspend_64< true, false >(int, kmp_atomic_flag_64< true, false > *)
void __kmp_hidden_helper_threads_deinitz_wait()
kmp_uint64 __kmp_ticks_per_msec
template void __kmp_atomic_suspend_64< false, true >(int, kmp_atomic_flag_64< false, true > *)
int __kmp_gtid_get_specific()
void __kmp_hidden_helper_main_thread_wait()
void(* microtask_t4)(int *, int *, void *, void *, void *, void *)
static struct kmp_sys_timer __kmp_sys_timer_data
void(* microtask_t6)(int *, int *, void *, void *, void *, void *, void *, void *)
void(* microtask_t14)(int *, int *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *)
void(* microtask_t11)(int *, int *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *)
int __kmp_invoke_microtask(microtask_t pkfn, int gtid, int tid, int argc, void *p_argv[])
void __kmp_hidden_helper_initz_release()
void __kmp_suspend_uninitialize_thread(kmp_info_t *th)
static int __kmp_fork_count
void __kmp_suspend_oncore(int th_gtid, kmp_flag_oncore *flag)
void __kmp_lock_suspend_mx(kmp_info_t *th)
void __kmp_terminate_thread(int gtid)
void(* microtask_t2)(int *, int *, void *, void *)
void __kmp_create_worker(int gtid, kmp_info_t *th, size_t stack_size)
template void __kmp_resume_64< false, true >(int, kmp_flag_64< false, true > *)
static void __kmp_atfork_prepare(void)
int __kmp_read_from_file(char const *path, char const *format,...)
void __kmp_elapsed(double *t)
void(* microtask_t9)(int *, int *, void *, void *, void *, void *, void *, void *, void *, void *, void *)
void __kmp_clear_system_time(void)
template void __kmp_suspend_64< false, true >(int, kmp_flag_64< false, true > *)
static void __kmp_resume_template(int target_gtid, C *flag)
void __kmp_runtime_destroy(void)
int __kmp_read_system_info(struct kmp_sys_info *info)
static void * __kmp_launch_worker(void *thr)
template void __kmp_suspend_32< false, false >(int, kmp_flag_32< false, false > *)
void(* microtask_t15)(int *, int *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *)
void __kmp_resume_oncore(int target_gtid, kmp_flag_oncore *flag)
void(* microtask_t10)(int *, int *, void *, void *, void *, void *, void *, void *, void *, void *, void *, void *)
void __kmp_runtime_initialize(void)
int __kmp_try_suspend_mx(kmp_info_t *th)
void __kmp_resume_64(int target_gtid, kmp_flag_64< C, S > *flag)
void __kmp_resume_32(int target_gtid, kmp_flag_32< C, S > *flag)
void __kmp_atomic_suspend_64(int th_gtid, kmp_atomic_flag_64< C, S > *flag)
static kmp_cond_align_t __kmp_wait_cv
void __kmp_elapsed_tick(double *t)
void __kmp_gtid_set_specific(int gtid)
static int __kmp_get_xproc(void)
void(* microtask_t3)(int *, int *, void *, void *, void *)
void __kmp_atomic_resume_64(int target_gtid, kmp_atomic_flag_64< C, S > *flag)
void(* microtask_t8)(int *, int *, void *, void *, void *, void *, void *, void *, void *, void *)
void __kmp_suspend_initialize_thread(kmp_info_t *th)
void __kmp_exit_thread(int exit_status)
void __kmp_register_atfork(void)
void __kmp_reap_worker(kmp_info_t *th)
void __kmp_hidden_helper_threads_deinitz_release()
void __kmp_affinity_determine_capable(const char *env_var)
void __kmp_affinity_bind_thread(int proc)
int __kmp_get_load_balance(int max)
void __kmp_initialize_system_tick(void)