Lines Matching defs:thread

69 static kmp_dephash_t *__kmp_dephash_extend(kmp_info_t *thread,
82 h = (kmp_dephash_t *)__kmp_fast_allocate(thread, size_to_allocate);
84 h = (kmp_dephash_t *)__kmp_thread_malloc(thread, size_to_allocate);
117 __kmp_fast_free(thread, current_dephash);
119 __kmp_thread_free(thread, current_dephash);
125 static kmp_dephash_t *__kmp_dephash_create(kmp_info_t *thread,
139 h = (kmp_dephash_t *)__kmp_fast_allocate(thread, size);
141 h = (kmp_dephash_t *)__kmp_thread_malloc(thread, size);
157 static kmp_dephash_entry *__kmp_dephash_find(kmp_info_t *thread,
162 *hash = __kmp_dephash_extend(thread, h);
173 // create entry. This is only done by one thread so no locking required
176 thread, sizeof(kmp_dephash_entry_t));
179 thread, sizeof(kmp_dephash_entry_t));
199 static kmp_depnode_list_t *__kmp_add_node(kmp_info_t *thread,
206 thread, sizeof(kmp_depnode_list_t));
209 thread, sizeof(kmp_depnode_list_t));
298 __kmp_depnode_link_successor(kmp_int32 gtid, kmp_info_t *thread,
325 dep->dn.successors = __kmp_add_node(thread, dep->dn.successors, node);
341 kmp_info_t *thread,
367 sink->dn.successors = __kmp_add_node(thread, sink->dn.successors, source);
399 kmp_info_t *thread = __kmp_threads[gtid];
404 __kmp_depnode_link_successor(gtid, thread, task, node, h->last_all);
405 __kmp_node_deref(thread, h->last_all);
427 __kmp_depnode_link_successor(gtid, thread, task, node, last_set);
428 __kmp_depnode_list_free(thread, last_set);
429 __kmp_depnode_list_free(thread, prev_set);
435 __kmp_depnode_link_successor(gtid, thread, task, node, last_out);
437 __kmp_node_deref(thread, last_out);
459 kmp_info_t *thread = __kmp_threads[gtid];
468 __kmp_dephash_find(thread, hash, dep->base_addr);
476 __kmp_depnode_link_successor(gtid, thread, task, node, last_set);
477 __kmp_depnode_list_free(thread, last_set);
478 __kmp_depnode_list_free(thread, prev_set);
484 __kmp_depnode_link_successor(gtid, thread, task, node, last_out);
486 __kmp_node_deref(thread, last_out);
500 __kmp_depnode_link_successor(gtid, thread, task, node, last_out);
503 __kmp_depnode_link_successor(gtid, thread, task, node, prev_set);
506 __kmp_node_deref(thread, last_out);
508 __kmp_depnode_list_free(thread, prev_set);
514 __kmp_depnode_link_successor(gtid, thread, task, node, last_set);
516 __kmp_node_deref(thread, last_out);
519 __kmp_depnode_list_free(thread, prev_set);
533 info->last_set = __kmp_add_node(thread, info->last_set, node);
622 // doesn't need to be atomic as no other thread is going to be accessing this
664 @param gtid Global Thread ID of encountering thread
675 Schedule a non-thread-switchable task with dependences for execution
687 kmp_info_t *thread = __kmp_threads[gtid];
688 kmp_taskdata_t *current_task = thread->th.th_current_task;
757 thread, (ndeps + ndeps_noalias) * sizeof(ompt_dependence_t));
802 KMP_OMPT_DEPS_FREE(thread, ompt_deps);
810 kmp_task_team_t *task_team = thread->th.th_task_team;
818 current_task->td_dephash = __kmp_dephash_create(thread, current_task);
822 (kmp_depnode_t *)__kmp_fast_allocate(thread, sizeof(kmp_depnode_t));
825 (kmp_depnode_t *)__kmp_thread_malloc(thread, sizeof(kmp_depnode_t));
880 @param gtid Global Thread ID of encountering thread
914 kmp_info_t *thread = __kmp_threads[gtid];
915 kmp_taskdata_t *current_task = thread->th.th_current_task;
923 ompt_data_t *taskwait_task_data = &thread->th.ompt_thread_info.task_data;
945 thread, (ndeps + ndeps_noalias) * sizeof(ompt_dependence_t));
981 KMP_OMPT_DEPS_FREE(thread, ompt_deps);
994 ignore && thread->th.th_task_team != NULL &&
995 thread->th.th_task_team->tt.tt_found_proxy_tasks == FALSE &&
996 thread->th.th_task_team->tt.tt_hidden_helper_task_encountered == FALSE;
1028 flag.execute_tasks(thread, gtid, FALSE,