Lines Matching full:bt

210 void __kmp_dist_barrier_wakeup(enum barrier_type bt, kmp_team_t *team,
214 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
227 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
247 gtid, team->t.t_id, tid, bt));
372 gtid, team->t.t_id, tid, bt));
376 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
387 gtid, tid, bt));
389 thr_bar = &this_thr->th.th_bar[bt].bb;
414 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
423 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
429 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
460 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
471 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
502 __kmp_dist_barrier_wakeup(bt, team, tid + 1, group_end, 1, tid);
524 __kmp_dist_barrier_wakeup(bt, team, tid + b->threads_per_group, nproc,
542 __kmp_dist_barrier_wakeup(bt, team, tid + 1, group_end, 1, tid);
551 gtid, team->t.t_id, tid, bt));
557 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
561 kmp_bstate_t *thr_bar = &this_thr->th.th_bar[bt].bb;
567 gtid, team->t.t_id, tid, bt));
593 kmp_balign_team_t *team_bar = &team->t.t_bar[bt];
604 KMP_CACHE_PREFETCH(&other_threads[i + 1]->th.th_bar[bt].bb.b_arrived);
610 &other_threads[i]->th.th_bar[bt].bb.b_arrived, new_state));
615 &other_threads[i]->th.th_bar[bt].bb.b_arrived, new_state);
619 kmp_flag_64<> flag(&other_threads[i]->th.th_bar[bt].bb.b_arrived,
653 gtid, team->t.t_id, tid, bt));
659 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
662 kmp_bstate_t *thr_bar = &this_thr->th.th_bar[bt].bb;
676 gtid, team->t.t_id, tid, bt));
700 KMP_CACHE_PREFETCH(&other_threads[i + 1]->th.th_bar[bt].bb.b_go);
707 team->t.t_id, i, &other_threads[i]->th.th_bar[bt].bb.b_go,
708 other_threads[i]->th.th_bar[bt].bb.b_go,
709 other_threads[i]->th.th_bar[bt].bb.b_go + KMP_BARRIER_STATE_BUMP));
710 kmp_flag_64<> flag(&other_threads[i]->th.th_bar[bt].bb.b_go,
734 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
744 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
761 gtid, team->t.t_id, tid, bt));
766 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
769 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj));
773 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
776 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj));
780 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
783 bt, this_thr, gtid, tid, propagate_icvs USE_ITT_BUILD_ARG(itt_sync_obj));
787 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
790 bt, this_thr, gtid, tid, propagate_icvs USE_ITT_BUILD_ARG(itt_sync_obj));
795 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
799 kmp_bstate_t *thr_bar = &this_thr->th.th_bar[bt].bb;
802 kmp_uint32 branch_bits = __kmp_barrier_gather_branch_bits[bt];
810 gtid, team->t.t_id, tid, bt));
825 new_state = team->t.t_bar[bt].b_arrived + KMP_BARRIER_STATE_BUMP;
829 kmp_bstate_t *child_bar = &child_thr->th.th_bar[bt].bb;
834 &other_threads[child_tid + 1]->th.th_bar[bt].bb.b_arrived);
887 team->t.t_bar[bt].b_arrived = new_state;
889 team->t.t_bar[bt].b_arrived += KMP_BARRIER_STATE_BUMP;
893 &team->t.t_bar[bt].b_arrived, team->t.t_bar[bt].b_arrived));
897 gtid, team->t.t_id, tid, bt));
901 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
905 kmp_bstate_t *thr_bar = &this_thr->th.th_bar[bt].bb;
907 kmp_uint32 branch_bits = __kmp_barrier_release_branch_bits[bt];
928 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
938 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
956 gtid, team->t.t_id, tid, bt));
967 kmp_bstate_t *child_bar = &child_thr->th.th_bar[bt].bb;
972 &other_threads[child_tid + 1]->th.th_bar[bt].bb.b_go);
1002 gtid, team->t.t_id, tid, bt));
1007 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
1011 kmp_bstate_t *thr_bar = &this_thr->th.th_bar[bt].bb;
1015 kmp_uint32 branch_bits = __kmp_barrier_gather_branch_bits[bt];
1023 gtid, team->t.t_id, tid, bt));
1063 new_state = team->t.t_bar[bt].b_arrived + KMP_BARRIER_STATE_BUMP;
1068 kmp_bstate_t *child_bar = &child_thr->th.th_bar[bt].bb;
1074 &other_threads[next_child_tid]->th.th_bar[bt].bb.b_arrived);
1110 team->t.t_bar[bt].b_arrived += KMP_BARRIER_STATE_BUMP;
1112 team->t.t_bar[bt].b_arrived = new_state;
1116 &team->t.t_bar[bt].b_arrived, team->t.t_bar[bt].b_arrived));
1120 gtid, team->t.t_id, tid, bt));
1126 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
1130 kmp_bstate_t *thr_bar = &this_thr->th.th_bar[bt].bb;
1133 kmp_uint32 branch_bits = __kmp_barrier_release_branch_bits[bt];
1149 gtid, team->t.t_id, tid, bt));
1169 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
1179 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
1234 kmp_bstate_t *child_bar = &child_thr->th.th_bar[bt].bb;
1244 &other_threads[next_child_tid]->th.th_bar[bt].bb.b_go);
1277 gtid, team->t.t_id, tid, bt));
1290 static bool __kmp_init_hierarchical_barrier_thread(enum barrier_type bt,
1335 &team->t.t_threads[thr_bar->parent_tid]->th.th_bar[bt].bb;
1340 &team->t.t_threads[thr_bar->parent_tid]->th.th_bar[bt].bb;
1358 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
1362 kmp_bstate_t *thr_bar = &this_thr->th.th_bar[bt].bb;
1379 gtid, team->t.t_id, tid, bt));
1389 (void)__kmp_init_hierarchical_barrier_thread(bt, thr_bar, nproc, gtid, tid,
1395 (kmp_uint64)team->t.t_bar[bt].b_arrived + KMP_BARRIER_STATE_BUMP;
1403 : team->t.t_bar[bt].b_arrived | thr_bar->leaf_state;
1436 kmp_bstate_t *child_bar = &child_thr->th.th_bar[bt].bb;
1465 kmp_bstate_t *child_bar = &child_thr->th.th_bar[bt].bb;
1507 thr_bar->b_arrived = team->t.t_bar[bt].b_arrived + KMP_BARRIER_STATE_BUMP;
1514 team->t.t_bar[bt].b_arrived = new_state;
1518 &team->t.t_bar[bt].b_arrived, team->t.t_bar[bt].b_arrived));
1523 gtid, team->t.t_id, tid, bt));
1527 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
1531 kmp_bstate_t *thr_bar = &this_thr->th.th_bar[bt].bb;
1540 gtid, team->t.t_id, tid, bt));
1557 thr_bar->offset + 1, bt,
1571 if (bt == bs_forkjoin_barrier && TCR_4(__kmp_global.g.g_done))
1604 team_change = __kmp_init_hierarchical_barrier_thread(bt, thr_bar, nproc, gtid,
1655 &team->t.t_threads[child_tid]->th.th_bar[bt].bb;
1686 kmp_bstate_t *child_bar = &child_thr->th.th_bar[bt].bb;
1712 kmp_bstate_t *child_bar = &child_thr->th.th_bar[bt].bb;
1734 gtid, team->t.t_id, tid, bt));
1767 static int __kmp_barrier_template(enum barrier_type bt, int gtid, int is_split,
1793 barrier_kind = __ompt_get_barrier_kind(bt, this_thr);
1836 itt_sync_obj = __kmp_itt_barrier_object(gtid, bt, 1);
1868 team->t.t_bar[bt].b_master_arrived += 1;
1870 this_thr->th.th_bar[bt].bb.b_worker_arrived += 1;
1883 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj));
1885 switch (__kmp_barrier_gather_pattern[bt]) {
1887 __kmp_dist_barrier_gather(bt, this_thr, gtid, tid,
1893 KMP_ASSERT(__kmp_barrier_gather_branch_bits[bt]);
1894 __kmp_hyper_barrier_gather(bt, this_thr, gtid, tid,
1900 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj));
1905 KMP_ASSERT(__kmp_barrier_gather_branch_bits[bt]);
1906 __kmp_tree_barrier_gather(bt, this_thr, gtid, tid,
1911 __kmp_linear_barrier_gather(bt, this_thr, gtid, tid,
1927 team->t.t_bar[bt].b_team_arrived += 1;
2001 bt, this_thr, gtid, tid, FALSE USE_ITT_BUILD_ARG(itt_sync_obj));
2003 switch (__kmp_barrier_release_pattern[bt]) {
2005 KMP_ASSERT(__kmp_barrier_release_branch_bits[bt]);
2006 __kmp_dist_barrier_release(bt, this_thr, gtid, tid,
2011 KMP_ASSERT(__kmp_barrier_release_branch_bits[bt]);
2012 __kmp_hyper_barrier_release(bt, this_thr, gtid, tid,
2018 bt, this_thr, gtid, tid, FALSE USE_ITT_BUILD_ARG(itt_sync_obj));
2022 KMP_ASSERT(__kmp_barrier_release_branch_bits[bt]);
2023 __kmp_tree_barrier_release(bt, this_thr, gtid, tid,
2028 __kmp_linear_barrier_release(bt, this_thr, gtid, tid,
2052 itt_sync_obj = __kmp_itt_barrier_object(gtid, bt, 1);
2099 int __kmp_barrier(enum barrier_type bt, int gtid, int is_split,
2102 return __kmp_barrier_template<>(bt, gtid, is_split, reduce_size, reduce_data,
2130 void __kmp_end_split_barrier(enum barrier_type bt, int gtid) {
2133 KMP_DEBUG_ASSERT(bt < bs_last_barrier);
2140 switch (__kmp_barrier_release_pattern[bt]) {
2142 __kmp_dist_barrier_release(bt, this_thr, gtid, tid,
2147 KMP_ASSERT(__kmp_barrier_release_branch_bits[bt]);
2148 __kmp_hyper_barrier_release(bt, this_thr, gtid, tid,
2153 __kmp_hierarchical_barrier_release(bt, this_thr, gtid, tid,
2158 KMP_ASSERT(__kmp_barrier_release_branch_bits[bt]);
2159 __kmp_tree_barrier_release(bt, this_thr, gtid, tid,
2164 __kmp_linear_barrier_release(bt, this_thr, gtid, tid,