Lines Matching defs:lck

1057   ompt_wait_id_t lck;
1062 lck = (ompt_wait_id_t)(uintptr_t)&team->t.t_ordered.dt.t_value;
1064 th->th.ompt_thread_info.wait_id = lck;
1071 ompt_mutex_ordered, omp_lock_hint_none, kmp_mutex_impl_spin, lck,
1091 ompt_mutex_ordered, (ompt_wait_id_t)(uintptr_t)lck, codeptr_ra);
1147 kmp_indirect_lock_t **lck;
1148 lck = (kmp_indirect_lock_t **)crit;
1158 int status = KMP_COMPARE_AND_STORE_PTR(lck, nullptr, ilk);
1167 KMP_DEBUG_ASSERT(*lck != NULL);
1291 kmp_user_lock_p lck = (kmp_user_lock_p)TCR_PTR(*lck_pp);
1293 if (lck == NULL) {
1298 lck = __kmp_user_lock_allocate(&idx, gtid, kmp_lf_critical_section);
1299 __kmp_init_user_lock_with_checks(lck);
1300 __kmp_set_user_lock_location(lck, loc);
1302 __kmp_itt_critical_creating(lck);
1313 int status = KMP_COMPARE_AND_STORE_PTR(lck_pp, 0, lck);
1318 __kmp_itt_critical_destroyed(lck);
1322 __kmp_destroy_user_lock_with_checks(lck);
1323 __kmp_user_lock_free(&idx, gtid, lck);
1324 lck = (kmp_user_lock_p)TCR_PTR(*lck_pp);
1325 KMP_DEBUG_ASSERT(lck != NULL);
1328 return lck;
1356 kmp_user_lock_p lck;
1367 (sizeof(lck->tas.lk.poll) <= OMP_CRITICAL_SIZE)) {
1368 lck = (kmp_user_lock_p)crit;
1372 (sizeof(lck->futex.lk.poll) <= OMP_CRITICAL_SIZE)) {
1373 lck = (kmp_user_lock_p)crit;
1377 lck = __kmp_get_critical_section_ptr(crit, loc, global_tid);
1381 __kmp_push_sync(global_tid, ct_critical, loc, lck);
1389 __kmp_itt_critical_acquiring(lck);
1398 ti.wait_id = (ompt_wait_id_t)(uintptr_t)lck;
1406 (ompt_wait_id_t)(uintptr_t)lck, codeptr_ra);
1412 __kmp_acquire_user_lock_with_checks(lck, global_tid);
1415 __kmp_itt_critical_acquired(lck);
1426 ompt_mutex_critical, (ompt_wait_id_t)(uintptr_t)lck, codeptr_ra);
1573 kmp_user_lock_p lck;
1603 lck = (kmp_user_lock_p)lk;
1605 __kmp_push_sync(global_tid, ct_critical, loc, lck,
1609 __kmp_itt_critical_acquiring(lck);
1616 ti.wait_id = (ompt_wait_id_t)(uintptr_t)lck;
1623 __ompt_get_mutex_impl_type(crit), (ompt_wait_id_t)(uintptr_t)lck,
1630 KMP_ACQUIRE_TAS_LOCK(lck, global_tid);
1634 KMP_ACQUIRE_FUTEX_LOCK(lck, global_tid);
1642 lck = ilk->lock;
1644 __kmp_push_sync(global_tid, ct_critical, loc, lck,
1648 __kmp_itt_critical_acquiring(lck);
1655 ti.wait_id = (ompt_wait_id_t)(uintptr_t)lck;
1662 __ompt_get_mutex_impl_type(0, ilk), (ompt_wait_id_t)(uintptr_t)lck,
1667 KMP_I_LOCK_FUNC(ilk, set)(lck, global_tid);
1672 __kmp_itt_critical_acquired(lck);
1683 ompt_mutex_critical, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
1705 kmp_user_lock_p lck;
1712 lck = (kmp_user_lock_p)crit;
1713 KMP_ASSERT(lck != NULL);
1718 __kmp_itt_critical_releasing(lck);
1722 KMP_RELEASE_TAS_LOCK(lck, global_tid);
1726 KMP_RELEASE_FUTEX_LOCK(lck, global_tid);
1730 KMP_D_LOCK_FUNC(lck, unset)((kmp_dyna_lock_t *)lck, global_tid);
1736 lck = ilk->lock;
1741 __kmp_itt_critical_releasing(lck);
1743 KMP_I_LOCK_FUNC(ilk, unset)(lck, global_tid);
1749 (sizeof(lck->tas.lk.poll) <= OMP_CRITICAL_SIZE)) {
1750 lck = (kmp_user_lock_p)crit;
1754 (sizeof(lck->futex.lk.poll) <= OMP_CRITICAL_SIZE)) {
1755 lck = (kmp_user_lock_p)crit;
1759 lck = (kmp_user_lock_p)TCR_PTR(*((kmp_user_lock_p *)crit));
1762 KMP_ASSERT(lck != NULL);
1768 __kmp_itt_critical_releasing(lck);
1772 __kmp_release_user_lock_with_checks(lck, global_tid);
1782 ompt_mutex_critical, (ompt_wait_id_t)(uintptr_t)lck,
2538 kmp_user_lock_p lck;
2550 (sizeof(lck->tas.lk.poll) <= OMP_LOCK_T_SIZE)) {
2551 lck = (kmp_user_lock_p)user_lock;
2555 (sizeof(lck->futex.lk.poll) <= OMP_LOCK_T_SIZE)) {
2556 lck = (kmp_user_lock_p)user_lock;
2560 lck = __kmp_user_lock_allocate(user_lock, gtid, 0);
2562 INIT_LOCK(lck);
2563 __kmp_set_user_lock_location(lck, loc);
2578 __kmp_itt_lock_creating(lck);
2610 kmp_user_lock_p lck;
2622 (sizeof(lck->tas.lk.poll) + sizeof(lck->tas.lk.depth_locked) <=
2624 lck = (kmp_user_lock_p)user_lock;
2628 (sizeof(lck->futex.lk.poll) + sizeof(lck->futex.lk.depth_locked) <=
2630 lck = (kmp_user_lock_p)user_lock;
2634 lck = __kmp_user_lock_allocate(user_lock, gtid, 0);
2637 INIT_NESTED_LOCK(lck);
2638 __kmp_set_user_lock_location(lck, loc);
2653 __kmp_itt_lock_creating(lck);
2663 kmp_user_lock_p lck;
2665 lck = ((kmp_indirect_lock_t *)KMP_LOOKUP_I_LOCK(user_lock))->lock;
2667 lck = (kmp_user_lock_p)user_lock;
2669 __kmp_itt_lock_destroyed(lck);
2683 kmp_user_lock_p lck;
2686 (sizeof(lck->tas.lk.poll) <= OMP_LOCK_T_SIZE)) {
2687 lck = (kmp_user_lock_p)user_lock;
2691 (sizeof(lck->futex.lk.poll) <= OMP_LOCK_T_SIZE)) {
2692 lck = (kmp_user_lock_p)user_lock;
2696 lck = __kmp_lookup_user_lock(user_lock, "omp_destroy_lock");
2711 __kmp_itt_lock_destroyed(lck);
2713 DESTROY_LOCK(lck);
2716 (sizeof(lck->tas.lk.poll) <= OMP_LOCK_T_SIZE)) {
2721 (sizeof(lck->futex.lk.poll) <= OMP_LOCK_T_SIZE)) {
2726 __kmp_user_lock_free(user_lock, gtid, lck);
2753 kmp_user_lock_p lck;
2756 (sizeof(lck->tas.lk.poll) + sizeof(lck->tas.lk.depth_locked) <=
2758 lck = (kmp_user_lock_p)user_lock;
2762 (sizeof(lck->futex.lk.poll) + sizeof(lck->futex.lk.depth_locked) <=
2764 lck = (kmp_user_lock_p)user_lock;
2768 lck = __kmp_lookup_user_lock(user_lock, "omp_destroy_nest_lock");
2783 __kmp_itt_lock_destroyed(lck);
2786 DESTROY_NESTED_LOCK(lck);
2789 (sizeof(lck->tas.lk.poll) + sizeof(lck->tas.lk.depth_locked) <=
2795 (sizeof(lck->futex.lk.poll) + sizeof(lck->futex.lk.depth_locked) <=
2801 __kmp_user_lock_free(user_lock, gtid, lck);
2851 kmp_user_lock_p lck;
2854 (sizeof(lck->tas.lk.poll) <= OMP_LOCK_T_SIZE)) {
2855 lck = (kmp_user_lock_p)user_lock;
2859 (sizeof(lck->futex.lk.poll) <= OMP_LOCK_T_SIZE)) {
2860 lck = (kmp_user_lock_p)user_lock;
2864 lck = __kmp_lookup_user_lock(user_lock, "omp_set_lock");
2868 __kmp_itt_lock_acquiring(lck);
2878 (ompt_wait_id_t)(uintptr_t)lck, codeptr);
2882 ACQUIRE_LOCK(lck, gtid);
2885 __kmp_itt_lock_acquired(lck);
2891 ompt_mutex_lock, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
2946 kmp_user_lock_p lck;
2949 (sizeof(lck->tas.lk.poll) + sizeof(lck->tas.lk.depth_locked) <=
2951 lck = (kmp_user_lock_p)user_lock;
2955 (sizeof(lck->futex.lk.poll) + sizeof(lck->futex.lk.depth_locked) <=
2957 lck = (kmp_user_lock_p)user_lock;
2961 lck = __kmp_lookup_user_lock(user_lock, "omp_set_nest_lock");
2965 __kmp_itt_lock_acquiring(lck);
2976 __ompt_get_mutex_impl_type(), (ompt_wait_id_t)(uintptr_t)lck,
2982 ACQUIRE_NESTED_LOCK(lck, gtid, &acquire_status);
2985 __kmp_itt_lock_acquired(lck);
2994 ompt_mutex_nest_lock, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3000 ompt_scope_begin, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3042 kmp_user_lock_p lck;
3048 (sizeof(lck->tas.lk.poll) <= OMP_LOCK_T_SIZE)) {
3065 ompt_mutex_lock, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3071 lck = (kmp_user_lock_p)user_lock;
3076 (sizeof(lck->futex.lk.poll) <= OMP_LOCK_T_SIZE)) {
3077 lck = (kmp_user_lock_p)user_lock;
3081 lck = __kmp_lookup_user_lock(user_lock, "omp_unset_lock");
3085 __kmp_itt_lock_releasing(lck);
3088 RELEASE_LOCK(lck, gtid);
3097 ompt_mutex_lock, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3138 kmp_user_lock_p lck;
3143 (sizeof(lck->tas.lk.poll) + sizeof(lck->tas.lk.depth_locked) <=
3175 ompt_mutex_nest_lock, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3180 ompt_mutex_scope_end, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3187 lck = (kmp_user_lock_p)user_lock;
3192 (sizeof(lck->futex.lk.poll) + sizeof(lck->futex.lk.depth_locked) <=
3194 lck = (kmp_user_lock_p)user_lock;
3198 lck = __kmp_lookup_user_lock(user_lock, "omp_unset_nest_lock");
3202 __kmp_itt_lock_releasing(lck);
3206 release_status = RELEASE_NESTED_LOCK(lck, gtid);
3217 ompt_mutex_nest_lock, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3222 ompt_mutex_scope_end, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3284 kmp_user_lock_p lck;
3288 (sizeof(lck->tas.lk.poll) <= OMP_LOCK_T_SIZE)) {
3289 lck = (kmp_user_lock_p)user_lock;
3293 (sizeof(lck->futex.lk.poll) <= OMP_LOCK_T_SIZE)) {
3294 lck = (kmp_user_lock_p)user_lock;
3298 lck = __kmp_lookup_user_lock(user_lock, "omp_test_lock");
3302 __kmp_itt_lock_acquiring(lck);
3312 (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3316 rc = TEST_LOCK(lck, gtid);
3319 __kmp_itt_lock_acquired(lck);
3321 __kmp_itt_lock_cancelled(lck);
3327 ompt_mutex_test_lock, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3387 kmp_user_lock_p lck;
3391 (sizeof(lck->tas.lk.poll) + sizeof(lck->tas.lk.depth_locked) <=
3393 lck = (kmp_user_lock_p)user_lock;
3397 (sizeof(lck->futex.lk.poll) + sizeof(lck->futex.lk.depth_locked) <=
3399 lck = (kmp_user_lock_p)user_lock;
3403 lck = __kmp_lookup_user_lock(user_lock, "omp_test_nest_lock");
3407 __kmp_itt_lock_acquiring(lck);
3419 __ompt_get_mutex_impl_type(), (ompt_wait_id_t)(uintptr_t)lck,
3424 rc = TEST_NESTED_LOCK(lck, gtid);
3427 __kmp_itt_lock_acquired(lck);
3429 __kmp_itt_lock_cancelled(lck);
3438 ompt_mutex_test_nest_lock, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3444 ompt_mutex_scope_begin, (ompt_wait_id_t)(uintptr_t)lck, codeptr);
3481 kmp_user_lock_p lck;
3500 lck = (kmp_user_lock_p)lk;
3501 KMP_DEBUG_ASSERT(lck != NULL);
3503 __kmp_push_sync(global_tid, ct_critical, loc, lck, __kmp_user_lock_seq);
3508 lck = ilk->lock;
3509 KMP_DEBUG_ASSERT(lck != NULL);
3511 __kmp_push_sync(global_tid, ct_critical, loc, lck, __kmp_user_lock_seq);
3513 KMP_I_LOCK_FUNC(ilk, set)(lck, global_tid);
3522 lck = (kmp_user_lock_p)crit;
3524 lck = __kmp_get_critical_section_ptr(crit, loc, global_tid);
3526 KMP_DEBUG_ASSERT(lck != NULL);
3529 __kmp_push_sync(global_tid, ct_critical, loc, lck);
3531 __kmp_acquire_user_lock_with_checks(lck, global_tid);
3541 kmp_user_lock_p lck;
3546 lck = (kmp_user_lock_p)crit;
3549 KMP_D_LOCK_FUNC(lck, unset)((kmp_dyna_lock_t *)lck, global_tid);
3564 lck = *((kmp_user_lock_p *)crit);
3565 KMP_ASSERT(lck != NULL);
3567 lck = (kmp_user_lock_p)crit;
3573 __kmp_release_user_lock_with_checks(lck, global_tid);
3623 @param lck pointer to the unique lock data structure
3633 kmp_critical_name *lck) {
3684 loc, global_tid, num_vars, reduce_size, reduce_data, reduce_func, lck);
3692 __kmp_enter_critical_section_reduce_block(loc, global_tid, lck);
3786 @param lck pointer to the unique lock data structure
3791 kmp_critical_name *lck) {
3804 __kmp_end_critical_section_reduce_block(loc, global_tid, lck);
3853 @param lck pointer to the unique lock data structure
3862 kmp_critical_name *lck) {
3897 loc, global_tid, num_vars, reduce_size, reduce_data, reduce_func, lck);
3905 __kmp_enter_critical_section_reduce_block(loc, global_tid, lck);
3975 @param lck pointer to the unique lock data structure
3978 The <tt>lck</tt> pointer must be the same as that used in the corresponding
3982 kmp_critical_name *lck) {
4002 __kmp_end_critical_section_reduce_block(loc, global_tid, lck);