/netbsd-src/external/mpl/bind/dist/tests/isc/ |
H A D | uv_wrap.h | 103 if (atomic_load(&__state_uv_udp_open) == 0) { 106 return atomic_load(&__state_uv_udp_open); in __wrap_uv_udp_open() 114 if (atomic_load(&__state_uv_udp_bind) == 0) { in __wrap_uv_udp_bind() 117 return atomic_load(&__state_uv_udp_bind); in __wrap_uv_udp_bind() 124 if (atomic_load(&__state_uv_udp_connect) == 0) { 127 return atomic_load(&__state_uv_udp_connect); in __wrap_uv_udp_connect() 135 if (atomic_load(&__state_uv_udp_getpeername) == 0) { 138 return atomic_load(&__state_uv_udp_getpeername); in __wrap_uv_udp_getpeername() 145 if (atomic_load(&__state_uv_udp_getsockname) == 0) { 148 return atomic_load( [all...] |
H A D | task_test.c |
|
H A D | doh_test.c | 122 int __v = atomic_load(&v); \ 123 assert_true(__v >= atomic_load(&total_sends)); \ 128 int __v = atomic_load(&v); \ 129 assert_true(__v >= atomic_load(&total_sends) / 2); \ 135 #define X(v) fprintf(stderr, #v " = %" PRIu64 "\n", atomic_load(&v)) 142 if (!atomic_load(&use_PROXY)) { in connect_send_cb() 144 } else if (atomic_load(&use_TLS) && atomic_load(&use_PROXY_over_TLS)) { in connect_send_cb() 320 atomic_store(&nsends, atomic_load(&total_sends)); in setup_test() 416 if (atomic_load( in doh_reply_sent_cb() [all...] |
H A D | quota_test.c | 196 assert_int_equal(atomic_load(&cb_calls), 0); in ISC_RUN_TEST_IMPL() 201 assert_int_equal(atomic_load(&cb_calls), i + 1); in ISC_RUN_TEST_IMPL() 205 assert_int_equal(atomic_load(&cb_calls), 10); in ISC_RUN_TEST_IMPL() 216 assert_int_equal(atomic_load(&cb_calls), 10); in ISC_RUN_TEST_IMPL() 287 for (i = 0; i < (int)atomic_load(&g_tnum); i++) { in quota_thread() 293 direct += atomic_load(&qtis[i].direct); in ISC_RUN_TEST_IMPL() 294 ncallback += atomic_load(&qtis[i].callback); in ISC_RUN_TEST_IMPL()
|
/netbsd-src/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/ |
H A D | sanitizer_addrhashmap.h | 121 uptr addr1 = atomic_load(&c->addr, memory_order_acquire); in ForEach() 128 (AddBucket *)atomic_load(&bucket->add, memory_order_acquire)) { in ForEach() 131 uptr addr1 = atomic_load(&c->addr, memory_order_acquire); in ForEach() 225 uptr addr1 = atomic_load(&c->addr, memory_order_acquire); in acquire() 233 if (atomic_load(&b->add, memory_order_relaxed)) { in acquire() 235 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); in acquire() 238 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 254 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 266 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); in acquire() 270 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() [all …]
|
H A D | sanitizer_mutex.h | 42 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), 1); in CheckLocked() 184 state = atomic_load(&state_, memory_order_relaxed); in Lock() 204 state = atomic_load(&state_, memory_order_relaxed); in Lock() 254 state = atomic_load(&state_, memory_order_relaxed); in ReadLock() 270 state = atomic_load(&state_, memory_order_relaxed); in ReadLock() 301 CHECK(atomic_load(&state_, memory_order_relaxed) & kWriterLock); in CheckWriteLocked() 307 CHECK(atomic_load(&state_, memory_order_relaxed) & kReaderLockMask); in CheckReadLocked()
|
H A D | sanitizer_lfstack.h | 32 return (atomic_load(&head_, memory_order_relaxed) & kPtrMask) == 0; in Empty() 36 u64 cmp = atomic_load(&head_, memory_order_relaxed); in Push() 48 u64 cmp = atomic_load(&head_, memory_order_acquire); in Pop()
|
/netbsd-src/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/ |
H A D | sanitizer_addrhashmap.h | 186 uptr addr1 = atomic_load(&c->addr, memory_order_acquire); in acquire() 194 if (atomic_load(&b->add, memory_order_relaxed)) { in acquire() 196 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); in acquire() 199 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 215 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 227 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); in acquire() 231 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 255 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 288 CHECK_EQ(atomic_load(&c->addr, memory_order_relaxed), 0); in acquire() 299 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in release() [all …]
|
H A D | sanitizer_mutex.h | 42 CHECK_NE(atomic_load(&state_, memory_order_relaxed), 0); in CheckLocked() 54 if (atomic_load(&state_, memory_order_relaxed) == 0 in LockSlow() 103 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked); in ~RWMutex() 135 CHECK_NE(atomic_load(&state_, memory_order_relaxed), kUnlocked); in CheckLocked() 153 u32 cmp = atomic_load(&state_, memory_order_relaxed); in LockSlow() 167 u32 prev = atomic_load(&state_, memory_order_acquire); in ReadLockSlow()
|
H A D | sanitizer_lfstack.h | 31 return (atomic_load(&head_, memory_order_relaxed) & kPtrMask) == 0; in Empty() 35 u64 cmp = atomic_load(&head_, memory_order_relaxed); in Push() 47 u64 cmp = atomic_load(&head_, memory_order_acquire); in Pop()
|
/netbsd-src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/ |
H A D | sanitizer_addrhashmap.h | 188 uptr addr1 = atomic_load(&c->addr, memory_order_acquire); in acquire() 196 if (atomic_load(&b->add, memory_order_relaxed)) { in acquire() 198 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); in acquire() 201 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 217 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 229 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); in acquire() 233 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 257 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire() 290 CHECK_EQ(atomic_load(&c->addr, memory_order_relaxed), 0); in acquire() 301 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in release() [all …]
|
H A D | sanitizer_mutex.h | 44 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), 1); in CheckLocked() 56 if (atomic_load(&state_, memory_order_relaxed) == 0 in LockSlow() 105 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked); in ~RWMutex() 137 CHECK_NE(atomic_load(&state_, memory_order_relaxed), kUnlocked); in CheckLocked() 155 u32 cmp = atomic_load(&state_, memory_order_relaxed); in LockSlow() 169 u32 prev = atomic_load(&state_, memory_order_acquire); in ReadLockSlow()
|
H A D | sanitizer_lfstack.h | 33 return (atomic_load(&head_, memory_order_relaxed) & kPtrMask) == 0; in Empty() 37 u64 cmp = atomic_load(&head_, memory_order_relaxed); in Push() 49 u64 cmp = atomic_load(&head_, memory_order_acquire); in Pop()
|
/netbsd-src/external/mpl/dhcp/bind/dist/lib/isc/ |
H A D | tls.c | 65 REQUIRE(!atomic_load(&init_done)); in tls_initialize() 123 REQUIRE(atomic_load(&init_done)); in isc__tls_initialize() 128 REQUIRE(atomic_load(&init_done)); in tls_shutdown() 129 REQUIRE(!atomic_load(&shut_done)); in tls_shutdown() 162 REQUIRE(atomic_load(&shut_done)); in isc__tls_shutdown()
|
/netbsd-src/external/gpl3/gcc/dist/libsanitizer/asan/ |
H A D | asan_allocator.cpp | 64 u64 context = atomic_load(atomic_context, memory_order_relaxed); in AtomicContextLoad() 173 return atomic_load(&magic, memory_order_acquire) == kAllocBegMagic in Get() 349 return atomic_load(&rss_limit_exceeded, memory_order_relaxed); in RssLimitExceeded() 361 if (ac && atomic_load(&ac->chunk_state, memory_order_acquire) == in RePoisonChunk() 402 options->min_redzone = atomic_load(&min_redzone, memory_order_acquire); in GetOptions() 403 options->max_redzone = atomic_load(&max_redzone, memory_order_acquire); in GetOptions() 406 atomic_load(&alloc_dealloc_mismatch, memory_order_acquire); in GetOptions() 421 u32 min_log = RZSize2Log(atomic_load(&min_redzone, memory_order_acquire)); in ComputeRZLog() 422 u32 max_log = RZSize2Log(atomic_load(&max_redzone, memory_order_acquire)); in ComputeRZLog() 449 u8 left_state = atomic_load(&left_chunk->chunk_state, memory_order_relaxed); in ChooseChunk() [all …]
|
/netbsd-src/external/mpl/bind/dist/lib/isc/netmgr/ |
H A D | tcp.c |
|
H A D | tcpdns.c |
|
H A D | netmgr.c |
|
/netbsd-src/external/mpl/bind/dist/tests/dns/ |
H A D | zt_test.c | 268 assert_false(atomic_load(&done)); in ISC_RUN_TEST_IMPL()
|
/netbsd-src/external/mpl/dhcp/bind/dist/lib/isc/netmgr/ |
H A D | netmgr.c | 457 REQUIRE(!atomic_load(&mgr->paused)); in isc_nm_pause() 480 while (atomic_load(&mgr->workers_paused) != mgr->workers_running) { in isc_nm_pause() 507 REQUIRE(atomic_load(&mgr->paused)); in isc_nm_resume() 531 while (atomic_load(&mgr->workers_paused) != 0) { in isc_nm_resume() 670 *initial = atomic_load(&mgr->init); in isc_nm_gettimeouts() 674 *idle = atomic_load(&mgr->idle); in isc_nm_gettimeouts() 678 *keepalive = atomic_load(&mgr->keepalive); in isc_nm_gettimeouts() 682 *advertised = atomic_load(&mgr->advertised); in isc_nm_gettimeouts() 731 INSIST(atomic_load(&mgr->interlocked) != isc_nm_tid()); in nm_thread() 1156 return (atomic_load(&sock->parent->active)); in isc__nmsocket_active() [all …]
|
H A D | tcp.c | 184 if (!atomic_load(&sock->active)) { in tcp_connect_direct() 187 INSIST(atomic_load(&sock->active)); in tcp_connect_direct() 242 if (atomic_load(&sock->timedout)) { in tcp_connect_cb() 247 if (!atomic_load(&sock->connecting)) { in tcp_connect_cb() 493 while (atomic_load(&sock->rchildren) != sock->nchildren) { in isc_nm_listentcp() 503 REQUIRE(atomic_load(&sock->rchildren) == sock->nchildren); in isc_nm_listentcp() 750 (atomic_load(&sock->keepalive) in isc__nm_tcp_read() 751 ? atomic_load(&sock->mgr->keepalive) in isc__nm_tcp_read() 752 : atomic_load(&sock->mgr->idle)); in isc__nm_tcp_read() 903 if (!atomic_load(&sock->client)) { in isc__nm_tcp_read_cb() [all …]
|
H A D | tcpdns.c | 155 if (!atomic_load(&sock->active)) { in tcpdns_connect_direct() 158 INSIST(atomic_load(&sock->active)); in tcpdns_connect_direct() 212 if (atomic_load(&sock->timedout)) { in tcpdns_connect_cb() 459 while (atomic_load(&sock->rchildren) != sock->nchildren) { in isc_nm_listentcpdns() 469 REQUIRE(atomic_load(&sock->rchildren) == sock->nchildren); in isc_nm_listentcpdns() 717 (atomic_load(&sock->keepalive) in isc__nm_tcpdns_read() 718 ? atomic_load(&sock->mgr->keepalive) in isc__nm_tcpdns_read() 719 : atomic_load(&sock->mgr->idle)); in isc__nm_tcpdns_read() 889 if (!atomic_load(&sock->client)) { in isc__nm_tcpdns_read_cb() 890 sock->read_timeout = atomic_load(&sock->mgr->idle); in isc__nm_tcpdns_read_cb() [all …]
|
/netbsd-src/external/gpl3/gcc/dist/libsanitizer/tsan/ |
H A D | tsan_external.cpp | 35 if (tag >= atomic_load(&used_tags, memory_order_relaxed)) return nullptr; in GetTagData() 54 uptr tag_count = atomic_load(&used_tags, memory_order_relaxed); in TagFromShadowStackFrame() 64 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); in ExternalAccess() 101 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); in __tsan_external_assign_tag()
|
/netbsd-src/external/gpl3/gcc.old/dist/libsanitizer/tsan/ |
H A D | tsan_external.cc | 30 if (tag >= atomic_load(&used_tags, memory_order_relaxed)) return nullptr; in GetTagData() 49 uptr tag_count = atomic_load(&used_tags, memory_order_relaxed); in TagFromShadowStackFrame() 60 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); in ExternalAccess() 98 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); in __tsan_external_assign_tag()
|
/netbsd-src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/ |
H A D | tsan_external.cc | 32 if (tag >= atomic_load(&used_tags, memory_order_relaxed)) return nullptr; in GetTagData() 51 uptr tag_count = atomic_load(&used_tags, memory_order_relaxed); in TagFromShadowStackFrame() 62 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); in ExternalAccess() 100 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); in __tsan_external_assign_tag()
|