15 #include "kmp_affinity.h" 16 #if KMP_USE_HIER_SCHED 17 #include "kmp_dispatch_hier.h" 20 kmp_key_t __kmp_gtid_threadprivate_key;
22 #if KMP_ARCH_X86 || KMP_ARCH_X86_64 23 kmp_cpuinfo_t __kmp_cpuinfo = {0};
27 #include "kmp_stats.h" 29 kmp_tas_lock_t __kmp_stats_lock;
33 kmp_stats_list *__kmp_stats_list;
36 KMP_THREAD_LOCAL kmp_stats_list *__kmp_stats_thread_ptr = NULL;
39 tsc_tick_count __kmp_stats_start_time;
45 volatile int __kmp_init_serial = FALSE;
46 volatile int __kmp_init_gtid = FALSE;
47 volatile int __kmp_init_common = FALSE;
48 volatile int __kmp_init_middle = FALSE;
49 volatile int __kmp_init_parallel = FALSE;
51 volatile int __kmp_init_monitor =
54 volatile int __kmp_init_user_locks = FALSE;
57 kmp_cached_addr_t *__kmp_threadpriv_cache_list = NULL;
59 int __kmp_init_counter = 0;
60 int __kmp_root_counter = 0;
61 int __kmp_version = 0;
63 std::atomic<kmp_int32> __kmp_team_counter = ATOMIC_VAR_INIT(0);
64 std::atomic<kmp_int32> __kmp_task_counter = ATOMIC_VAR_INIT(0);
66 unsigned int __kmp_init_wait =
67 KMP_DEFAULT_INIT_WAIT;
68 unsigned int __kmp_next_wait =
69 KMP_DEFAULT_NEXT_WAIT;
71 size_t __kmp_stksize = KMP_DEFAULT_STKSIZE;
73 size_t __kmp_monitor_stksize = 0;
75 size_t __kmp_stkoffset = KMP_DEFAULT_STKOFFSET;
76 int __kmp_stkpadding = KMP_MIN_STKPADDING;
78 size_t __kmp_malloc_pool_incr = KMP_DEFAULT_MALLOC_POOL_INCR;
82 kmp_uint32 __kmp_barrier_gather_bb_dflt = 2;
84 kmp_uint32 __kmp_barrier_release_bb_dflt = 2;
87 kmp_bar_pat_e __kmp_barrier_gather_pat_dflt = bp_hyper_bar;
89 kmp_bar_pat_e __kmp_barrier_release_pat_dflt = bp_hyper_bar;
92 kmp_uint32 __kmp_barrier_gather_branch_bits[bs_last_barrier] = {0};
93 kmp_uint32 __kmp_barrier_release_branch_bits[bs_last_barrier] = {0};
94 kmp_bar_pat_e __kmp_barrier_gather_pattern[bs_last_barrier] = {bp_linear_bar};
95 kmp_bar_pat_e __kmp_barrier_release_pattern[bs_last_barrier] = {bp_linear_bar};
96 char const *__kmp_barrier_branch_bit_env_name[bs_last_barrier] = {
97 "KMP_PLAIN_BARRIER",
"KMP_FORKJOIN_BARRIER" 98 #if KMP_FAST_REDUCTION_BARRIER 100 "KMP_REDUCTION_BARRIER" 101 #endif // KMP_FAST_REDUCTION_BARRIER 103 char const *__kmp_barrier_pattern_env_name[bs_last_barrier] = {
104 "KMP_PLAIN_BARRIER_PATTERN",
"KMP_FORKJOIN_BARRIER_PATTERN" 105 #if KMP_FAST_REDUCTION_BARRIER 107 "KMP_REDUCTION_BARRIER_PATTERN" 108 #endif // KMP_FAST_REDUCTION_BARRIER 110 char const *__kmp_barrier_type_name[bs_last_barrier] = {
"plain",
"forkjoin" 111 #if KMP_FAST_REDUCTION_BARRIER 114 #endif // KMP_FAST_REDUCTION_BARRIER 116 char const *__kmp_barrier_pattern_name[bp_last_bar] = {
"linear",
"tree",
117 "hyper",
"hierarchical"};
119 int __kmp_allThreadsSpecified = 0;
120 size_t __kmp_align_alloc = CACHE_LINE;
122 int __kmp_generate_warnings = kmp_warnings_low;
123 int __kmp_reserve_warn = 0;
125 int __kmp_avail_proc = 0;
126 size_t __kmp_sys_min_stksize = KMP_MIN_STKSIZE;
127 int __kmp_sys_max_nth = KMP_MAX_NTH;
128 int __kmp_max_nth = 0;
129 int __kmp_cg_max_nth = 0;
130 int __kmp_teams_max_nth = 0;
131 int __kmp_threads_capacity = 0;
132 int __kmp_dflt_team_nth = 0;
133 int __kmp_dflt_team_nth_ub = 0;
134 int __kmp_tp_capacity = 0;
135 int __kmp_tp_cached = 0;
136 int __kmp_dflt_nested = FALSE;
137 int __kmp_dispatch_num_buffers = KMP_DFLT_DISP_NUM_BUFF;
138 int __kmp_dflt_max_active_levels =
139 KMP_MAX_ACTIVE_LEVELS_LIMIT;
140 #if KMP_NESTED_HOT_TEAMS 141 int __kmp_hot_teams_mode = 0;
143 int __kmp_hot_teams_max_level = 1;
145 enum library_type __kmp_library = library_none;
149 kmp_sch_static_greedy;
151 kmp_sch_guided_iterative_chunked;
153 kmp_sch_guided_analytical_chunked;
154 #if KMP_USE_HIER_SCHED 155 int __kmp_dispatch_hand_threading = 0;
156 int __kmp_hier_max_units[kmp_hier_layer_e::LAYER_LAST + 1];
157 int __kmp_hier_threads_per[kmp_hier_layer_e::LAYER_LAST + 1];
158 kmp_hier_sched_env_t __kmp_hier_scheds = {0, 0, NULL, NULL, NULL};
160 int __kmp_dflt_blocktime = KMP_DEFAULT_BLOCKTIME;
162 int __kmp_monitor_wakeups = KMP_MIN_MONITOR_WAKEUPS;
163 int __kmp_bt_intervals = KMP_INTERVALS_FROM_BLOCKTIME(KMP_DEFAULT_BLOCKTIME,
164 KMP_MIN_MONITOR_WAKEUPS);
166 #ifdef KMP_ADJUST_BLOCKTIME 167 int __kmp_zero_bt = FALSE;
169 #ifdef KMP_DFLT_NTH_CORES 170 int __kmp_ncores = 0;
173 int __kmp_abort_delay = 0;
174 #if KMP_OS_LINUX && defined(KMP_TDATA_GTID) 175 int __kmp_gtid_mode = 3;
176 int __kmp_adjust_gtid_mode = FALSE;
178 int __kmp_gtid_mode = 2;
179 int __kmp_adjust_gtid_mode = FALSE;
181 int __kmp_gtid_mode = 0;
182 int __kmp_adjust_gtid_mode = TRUE;
184 #ifdef KMP_TDATA_GTID 185 KMP_THREAD_LOCAL
int __kmp_gtid = KMP_GTID_DNE;
187 int __kmp_tls_gtid_min = INT_MAX;
188 int __kmp_foreign_tp = TRUE;
189 #if KMP_ARCH_X86 || KMP_ARCH_X86_64 190 int __kmp_inherit_fp_control = TRUE;
191 kmp_int16 __kmp_init_x87_fpu_control_word = 0;
192 kmp_uint32 __kmp_init_mxcsr = 0;
195 #ifdef USE_LOAD_BALANCE 196 double __kmp_load_balance_interval = 1.0;
199 kmp_nested_nthreads_t __kmp_nested_nth = {NULL, 0, 0};
201 #if KMP_USE_ADAPTIVE_LOCKS 203 kmp_adaptive_backoff_params_t __kmp_adaptive_backoff_params = {
206 #if KMP_DEBUG_ADAPTIVE_LOCKS 207 const char *__kmp_speculative_statsfile =
"-";
210 #endif // KMP_USE_ADAPTIVE_LOCKS 213 int __kmp_display_env = FALSE;
214 int __kmp_display_env_verbose = FALSE;
215 int __kmp_omp_cancellation = FALSE;
219 enum sched_type __kmp_sch_map[kmp_sched_upper - kmp_sched_lower_ext +
220 kmp_sched_upper_std - kmp_sched_lower - 2] = {
221 kmp_sch_static_chunked,
222 kmp_sch_dynamic_chunked,
231 enum clock_function_type __kmp_clock_function;
232 int __kmp_clock_function_param;
235 #if KMP_MIC_SUPPORTED 236 enum mic_type __kmp_mic_type = non_mic;
239 #if KMP_AFFINITY_SUPPORTED 241 KMPAffinity *__kmp_affinity_dispatch = NULL;
244 int __kmp_hwloc_error = FALSE;
245 hwloc_topology_t __kmp_hwloc_topology = NULL;
246 int __kmp_numa_detected = FALSE;
247 int __kmp_tile_depth = 0;
251 #if KMP_GROUP_AFFINITY 252 int __kmp_num_proc_groups = 1;
254 kmp_GetActiveProcessorCount_t __kmp_GetActiveProcessorCount = NULL;
255 kmp_GetActiveProcessorGroupCount_t __kmp_GetActiveProcessorGroupCount = NULL;
256 kmp_GetThreadGroupAffinity_t __kmp_GetThreadGroupAffinity = NULL;
257 kmp_SetThreadGroupAffinity_t __kmp_SetThreadGroupAffinity = NULL;
260 size_t __kmp_affin_mask_size = 0;
261 enum affinity_type __kmp_affinity_type = affinity_default;
262 enum affinity_gran __kmp_affinity_gran = affinity_gran_default;
263 int __kmp_affinity_gran_levels = -1;
264 int __kmp_affinity_dups = TRUE;
265 enum affinity_top_method __kmp_affinity_top_method =
266 affinity_top_method_default;
267 int __kmp_affinity_compact = 0;
268 int __kmp_affinity_offset = 0;
269 int __kmp_affinity_verbose = FALSE;
270 int __kmp_affinity_warnings = TRUE;
271 int __kmp_affinity_respect_mask = affinity_respect_mask_default;
272 char *__kmp_affinity_proclist = NULL;
273 kmp_affin_mask_t *__kmp_affinity_masks = NULL;
274 unsigned __kmp_affinity_num_masks = 0;
276 char *__kmp_cpuinfo_file = NULL;
281 kmp_nested_proc_bind_t __kmp_nested_proc_bind = {NULL, 0, 0};
282 int __kmp_affinity_num_places = 0;
286 int __kmp_display_affinity = FALSE;
287 char *__kmp_affinity_format = NULL;
288 #endif // OMP_50_ENABLED 290 kmp_hws_item_t __kmp_hws_socket = {0, 0};
291 kmp_hws_item_t __kmp_hws_node = {0, 0};
292 kmp_hws_item_t __kmp_hws_tile = {0, 0};
293 kmp_hws_item_t __kmp_hws_core = {0, 0};
294 kmp_hws_item_t __kmp_hws_proc = {0, 0};
295 int __kmp_hws_requested = 0;
296 int __kmp_hws_abs_flag = 0;
299 kmp_int32 __kmp_default_device = 0;
302 kmp_tasking_mode_t __kmp_tasking_mode = tskm_task_teams;
304 kmp_int32 __kmp_max_task_priority = 0;
305 kmp_uint64 __kmp_taskloop_min_tasks = 0;
309 int __kmp_memkind_available = 0;
310 int __kmp_hbw_mem_available = 0;
311 const omp_allocator_t *OMP_NULL_ALLOCATOR = NULL;
312 const omp_allocator_t *omp_default_mem_alloc = (
const omp_allocator_t *)1;
313 const omp_allocator_t *omp_large_cap_mem_alloc = (
const omp_allocator_t *)2;
314 const omp_allocator_t *omp_const_mem_alloc = (
const omp_allocator_t *)3;
315 const omp_allocator_t *omp_high_bw_mem_alloc = (
const omp_allocator_t *)4;
316 const omp_allocator_t *omp_low_lat_mem_alloc = (
const omp_allocator_t *)5;
317 const omp_allocator_t *omp_cgroup_mem_alloc = (
const omp_allocator_t *)6;
318 const omp_allocator_t *omp_pteam_mem_alloc = (
const omp_allocator_t *)7;
319 const omp_allocator_t *omp_thread_mem_alloc = (
const omp_allocator_t *)8;
320 void *
const *__kmp_def_allocator = omp_default_mem_alloc;
328 KMP_BUILD_ASSERT(
sizeof(kmp_tasking_flags_t) == 4);
330 int __kmp_task_stealing_constraint = 1;
333 int __kmp_suspend_count = 0;
336 int __kmp_settings = FALSE;
337 int __kmp_duplicate_library_ok = 0;
339 int __kmp_forkjoin_frames = 1;
340 int __kmp_forkjoin_frames_mode = 3;
342 PACKED_REDUCTION_METHOD_T __kmp_force_reduction_method =
343 reduction_method_not_defined;
344 int __kmp_determ_red = FALSE;
357 int __kmp_debug_buf =
359 int __kmp_debug_buf_lines =
360 KMP_DEBUG_BUF_LINES_INIT;
361 int __kmp_debug_buf_chars =
362 KMP_DEBUG_BUF_CHARS_INIT;
363 int __kmp_debug_buf_atomic =
366 char *__kmp_debug_buffer = NULL;
367 std::atomic<int> __kmp_debug_count =
369 int __kmp_debug_buf_warn_chars =
376 char __kmp_par_range_routine[KMP_PAR_RANGE_ROUTINE_LEN] = {
'\0'};
377 char __kmp_par_range_filename[KMP_PAR_RANGE_FILENAME_LEN] = {
'\0'};
378 int __kmp_par_range_lb = 0;
379 int __kmp_par_range_ub = INT_MAX;
383 int __kmp_storage_map =
385 int __kmp_storage_map_verbose =
387 int __kmp_storage_map_verbose_specified = FALSE;
390 int __kmp_need_register_atfork =
392 int __kmp_need_register_atfork_specified = TRUE;
394 int __kmp_env_stksize = FALSE;
395 int __kmp_env_blocktime = FALSE;
396 int __kmp_env_checks = FALSE;
397 int __kmp_env_consistency_check = FALSE;
399 kmp_uint32 __kmp_yield_init = KMP_INIT_WAIT;
400 kmp_uint32 __kmp_yield_next = KMP_NEXT_WAIT;
403 kmp_uint32 __kmp_yielding_on = 1;
406 kmp_uint32 __kmp_yield_cycle = 0;
408 kmp_uint32 __kmp_yield_cycle = 1;
410 kmp_int32 __kmp_yield_on_count =
412 kmp_int32 __kmp_yield_off_count =
422 kmp_info_t **__kmp_threads = NULL;
423 kmp_root_t **__kmp_root = NULL;
427 volatile int __kmp_nth = 0;
428 volatile int __kmp_all_nth = 0;
429 int __kmp_thread_pool_nth = 0;
430 volatile kmp_info_t *__kmp_thread_pool = NULL;
431 volatile kmp_team_t *__kmp_team_pool = NULL;
434 std::atomic<int> __kmp_thread_pool_active_nth = ATOMIC_VAR_INIT(0);
439 kmp_global_t __kmp_global = {{0}};
445 #if KMP_USE_INTERNODE_ALIGNMENT 448 KMP_ALIGN_CACHE_INTERNODE
450 KMP_BOOTSTRAP_LOCK_INIT(__kmp_initz_lock);
451 KMP_ALIGN_CACHE_INTERNODE
452 KMP_BOOTSTRAP_LOCK_INIT(__kmp_forkjoin_lock);
453 KMP_ALIGN_CACHE_INTERNODE
454 KMP_BOOTSTRAP_LOCK_INIT(__kmp_exit_lock);
457 KMP_ALIGN_CACHE_INTERNODE
458 KMP_BOOTSTRAP_LOCK_INIT(__kmp_monitor_lock);
462 KMP_ALIGN_CACHE_INTERNODE
463 KMP_BOOTSTRAP_LOCK_INIT(__kmp_tp_cached_lock);
465 KMP_ALIGN_CACHE_INTERNODE
466 KMP_LOCK_INIT(__kmp_global_lock);
467 KMP_ALIGN_CACHE_INTERNODE
468 kmp_queuing_lock_t __kmp_dispatch_lock;
469 KMP_ALIGN_CACHE_INTERNODE
470 KMP_LOCK_INIT(__kmp_debug_lock);
474 KMP_BOOTSTRAP_LOCK_INIT(__kmp_initz_lock);
475 KMP_BOOTSTRAP_LOCK_INIT(__kmp_forkjoin_lock);
476 KMP_BOOTSTRAP_LOCK_INIT(__kmp_exit_lock);
479 KMP_BOOTSTRAP_LOCK_INIT(__kmp_monitor_lock);
483 KMP_BOOTSTRAP_LOCK_INIT(__kmp_tp_cached_lock);
486 KMP_LOCK_INIT(__kmp_global_lock);
488 kmp_queuing_lock_t __kmp_dispatch_lock;
490 KMP_LOCK_INIT(__kmp_debug_lock);
495 #if KMP_HANDLE_SIGNALS 515 int __kmp_handle_signals = FALSE;
519 int get_suspend_count_(
void) {
520 int count = __kmp_suspend_count;
521 __kmp_suspend_count = 0;
524 void set_suspend_count_(
int *value) { __kmp_suspend_count = *value; }
528 int _You_must_link_with_exactly_one_OpenMP_library = 1;
529 int _You_must_link_with_Intel_OpenMP_library = 1;
530 #if KMP_OS_WINDOWS && (KMP_VERSION_MAJOR > 4) 531 int _You_must_link_with_Microsoft_OpenMP_library = 1;
535 kmp_target_offload_kind_t __kmp_target_offload = tgt_default;