| /openbsd-src/sys/dev/pci/drm/i915/selftests/ |
| H A D | i915_syncmap.c | 146 static int check_seqno(struct i915_syncmap *leaf, unsigned int idx, u32 seqno) in check_seqno() argument 154 if (__sync_seqno(leaf)[idx] != seqno) { in check_seqno() 156 __func__, idx, __sync_seqno(leaf)[idx], seqno); in check_seqno() 163 static int check_one(struct i915_syncmap **sync, u64 context, u32 seqno) in check_one() argument 167 err = i915_syncmap_set(sync, context, seqno); in check_one() 189 err = check_seqno((*sync), ilog2((*sync)->bitmap), seqno); in check_one() 193 if (!i915_syncmap_is_later(sync, context, seqno)) { in check_one() 195 context, seqno); in check_one() 238 static int check_leaf(struct i915_syncmap **sync, u64 context, u32 seqno) in check_leaf() argument 242 err = i915_syncmap_set(sync, context, seqno); in check_leaf() [all …]
|
| H A D | igt_spinner.c | 89 if (!spin->seqno) { in igt_spinner_pin() 94 spin->seqno = memset(vaddr, 0xff, PAGE_SIZE); in igt_spinner_pin() 178 *batch++ = rq->fence.seqno; in igt_spinner_create_request() 219 u32 *seqno = spin->seqno + seqno_offset(rq->fence.context); in hws_seqno() local 221 return READ_ONCE(*seqno); in hws_seqno() 243 if (spin->seqno) { in igt_spinner_fini() 256 rq->fence.seqno), in igt_wait_for_spinner() 259 rq->fence.seqno), in igt_wait_for_spinner()
|
| /openbsd-src/sys/dev/pci/drm/radeon/ |
| H A D | radeon_trace.h | 127 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), 129 TP_ARGS(dev, ring, seqno), 134 __field(u32, seqno) 140 __entry->seqno = seqno; 144 __entry->dev, __entry->ring, __entry->seqno) 149 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), 151 TP_ARGS(dev, ring, seqno) 156 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), 158 TP_ARGS(dev, ring, seqno) 163 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), [all …]
|
| /openbsd-src/sys/dev/pci/drm/i915/gt/ |
| H A D | selftest_timeline.c | 199 u32 seqno; member 211 if (__intel_timeline_sync_is_later(tl, ctx, p->seqno) != p->expected) { in __igt_sync() 213 name, p->name, ctx, p->seqno, str_yes_no(p->expected)); in __igt_sync() 218 ret = __intel_timeline_sync_set(tl, ctx, p->seqno); in __igt_sync() 396 u32 seqno = prandom_u32_state(&prng); in bench_sync() local 398 if (!__intel_timeline_sync_is_later(&tl, id, seqno)) in bench_sync() 399 __intel_timeline_sync_set(&tl, id, seqno); in bench_sync() 496 if (READ_ONCE(*tl->hwsp_seqno) != tl->seqno) { in checked_tl_write() 498 *tl->hwsp_seqno, tl->seqno); in checked_tl_write() 697 u32 seqno[2]; in live_hwsp_wrap() local [all …]
|
| H A D | intel_tlb.c | 112 static bool tlb_seqno_passed(const struct intel_gt *gt, u32 seqno) in tlb_seqno_passed() argument 117 return (s32)(cur - ALIGN(seqno, 2)) > 0; in tlb_seqno_passed() 120 void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno) in intel_gt_invalidate_tlb_full() argument 130 if (tlb_seqno_passed(gt, seqno)) in intel_gt_invalidate_tlb_full() 135 if (tlb_seqno_passed(gt, seqno)) in intel_gt_invalidate_tlb_full() 141 write_seqcount_invalidate(>->tlb.seqno); in intel_gt_invalidate_tlb_full() 144 gt->tlb.seqno.seq.sequence += 2; in intel_gt_invalidate_tlb_full() 154 seqcount_mutex_init(>->tlb.seqno, >->tlb.invalidate_lock); in intel_gt_init_tlb()
|
| H A D | intel_timeline.h | 46 u64 context, u32 seqno) in __intel_timeline_sync_set() argument 48 return i915_syncmap_set(&tl->sync, context, seqno); in __intel_timeline_sync_set() 54 return __intel_timeline_sync_set(tl, fence->context, fence->seqno); in intel_timeline_sync_set() 58 u64 context, u32 seqno) in __intel_timeline_sync_is_later() argument 60 return i915_syncmap_is_later(&tl->sync, context, seqno); in __intel_timeline_sync_is_later() 66 return __intel_timeline_sync_is_later(tl, fence->context, fence->seqno); in intel_timeline_sync_is_later() 74 u32 *seqno);
|
| H A D | intel_timeline.c | 230 WRITE_ONCE(*hwsp_seqno, tl->seqno); in intel_timeline_reset_seqno() 302 GEM_BUG_ON(tl->seqno & tl->has_initial_breadcrumb); in timeline_advance() 304 return tl->seqno += 1 + tl->has_initial_breadcrumb; in timeline_advance() 309 u32 *seqno) in __intel_timeline_get_seqno() argument 321 *seqno = timeline_advance(tl); in __intel_timeline_get_seqno() 322 GEM_BUG_ON(i915_seqno_passed(*tl->hwsp_seqno, *seqno)); in __intel_timeline_get_seqno() 328 u32 *seqno) in intel_timeline_get_seqno() argument 330 *seqno = timeline_advance(tl); in intel_timeline_get_seqno() 333 if (unlikely(!*seqno && tl->has_initial_breadcrumb)) in intel_timeline_get_seqno() 334 return __intel_timeline_get_seqno(tl, seqno); in intel_timeline_get_seqno() [all …]
|
| H A D | intel_tlb.h | 14 void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno); 21 return seqprop_sequence(>->tlb.seqno); in intel_gt_tlb_seqno()
|
| H A D | intel_breadcrumbs.c | 114 i915_seqno_passed(rq->fence.seqno, in check_signal_order() 115 list_next_entry(rq, signal_link)->fence.seqno)) in check_signal_order() 119 i915_seqno_passed(list_prev_entry(rq, signal_link)->fence.seqno, in check_signal_order() 120 rq->fence.seqno)) in check_signal_order() 387 if (i915_seqno_passed(rq->fence.seqno, it->fence.seqno)) in insert_breadcrumb() 500 rq->fence.context, rq->fence.seqno, in print_signals()
|
| H A D | selftest_hangcheck.c | 37 u32 *seqno; member 73 h->seqno = memset(vaddr, 0xff, PAGE_SIZE); in hang_init() 176 *batch++ = rq->fence.seqno; in hang_create_request() 190 *batch++ = rq->fence.seqno; in hang_create_request() 203 *batch++ = rq->fence.seqno; in hang_create_request() 215 *batch++ = rq->fence.seqno; in hang_create_request() 255 return READ_ONCE(h->seqno[rq->fence.context % (PAGE_SIZE/sizeof(u32))]); in hws_seqno() 277 rq->fence.seqno), in wait_until_running() 280 rq->fence.seqno), in wait_until_running() 757 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in __igt_reset_engine() [all …]
|
| H A D | gen6_engine_cs.c | 167 *cs++ = rq->fence.seqno; in gen6_emit_breadcrumb_rcs() 364 *cs++ = rq->fence.seqno; in gen7_emit_breadcrumb_rcs() 382 *cs++ = rq->fence.seqno; in gen6_emit_breadcrumb_xcs() 403 *cs++ = rq->fence.seqno; in gen7_emit_breadcrumb_xcs() 408 *cs++ = rq->fence.seqno; in gen7_emit_breadcrumb_xcs()
|
| /openbsd-src/sys/dev/pci/drm/i915/ |
| H A D | i915_syncmap.c | 154 bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno) in i915_syncmap_is_later() argument 195 return seqno_later(__sync_seqno(p)[idx], seqno); in i915_syncmap_is_later() 214 static inline void __sync_set_seqno(struct i915_syncmap *p, u64 id, u32 seqno) in __sync_set_seqno() argument 219 __sync_seqno(p)[idx] = seqno; in __sync_set_seqno() 230 static noinline int __sync_set(struct i915_syncmap **root, u64 id, u32 seqno) in __sync_set() argument 335 __sync_set_seqno(p, id, seqno); in __sync_set() 353 int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno) in i915_syncmap_set() argument 362 __sync_set_seqno(p, id, seqno); in i915_syncmap_set() 366 return __sync_set(root, id, seqno); in i915_syncmap_set()
|
| H A D | i915_request.h | 67 rq__->fence.context, rq__->fence.seqno, \ 510 u32 seqno; in hwsp_seqno() local 513 seqno = __hwsp_seqno(rq); in hwsp_seqno() 516 return seqno; in hwsp_seqno() 521 return i915_seqno_passed(__hwsp_seqno(rq), rq->fence.seqno - 1); in __i915_request_has_started() 612 return i915_seqno_passed(__hwsp_seqno(rq), rq->fence.seqno); in __i915_request_is_complete() 634 (u32 *)&rq->fence.seqno); in i915_request_mark_complete()
|
| H A D | i915_syncmap.h | 34 int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno); 35 bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno);
|
| H A D | i915_request.c | 535 inflight = i915_seqno_passed(rq->fence.seqno, in __request_in_flight() 536 signal->fence.seqno); in __request_in_flight() 969 u32 seqno; in __i915_request_create() local 1029 ret = intel_timeline_get_seqno(tl, rq, &seqno); in __i915_request_create() 1034 tl->fence_context, seqno); in __i915_request_create() 1227 u32 seqno) in __emit_semaphore_wait() argument 1263 *cs++ = seqno; in __emit_semaphore_wait() 1318 if (__emit_semaphore_wait(to, from, from->fence.seqno)) in emit_semaphore_wait() 1335 fence->seqno - 1); in intel_timeline_sync_has_start() 1341 return __intel_timeline_sync_set(tl, fence->context, fence->seqno - 1); in intel_timeline_sync_set_start() [all …]
|
| /openbsd-src/usr.sbin/tcpdump/ |
| H A D | print-gtp.c | 507 ntohs(gh->length), ntohs(gh->seqno), ntohs(gh->flow), in gtp_v0_print() 551 printf("GTPv0' (len %u, seq %u) ", len, ntohs(gph->seqno)); in gtp_v0_print_prime() 759 u_int16_t *lenp, *seqno, len; in gtp_v0_print_tlv() local 841 seqno = (u_int16_t *)cp; in gtp_v0_print_tlv() 843 printf(" %u", ntohs(*seqno)); in gtp_v0_print_tlv() 844 seqno++; in gtp_v0_print_tlv() 845 len -= sizeof(*seqno); in gtp_v0_print_tlv() 917 printf(" [seq %u]", ntohs(ghe->seqno)); in gtp_v1_print() 1044 printf(" (len %u, seq %u) ", len, ntohs(gph->seqno)); in gtp_v1_print_prime() 1353 u_int16_t *lenp, *seqno, len; in gtp_v1_print_tlv() local [all …]
|
| H A D | gtp.h | 375 u_int16_t seqno; member 388 u_int16_t seqno; member 400 u_int16_t seqno; member 409 u_int16_t seqno; member
|
| /openbsd-src/regress/sys/crypto/aesxts/ |
| H A D | aes_xts.c | 51 u_int64_t seqno; member 1749 do_aes_xts(u_int8_t *key, int klen, u_int64_t seqno, in do_aes_xts() argument 1759 aes_xts_reinit((caddr_t)&ctx, (u_int8_t *)&seqno); in do_aes_xts() 1787 if (do_aes_xts(tv->key, tv->key_len, tv->seqno, tv->plaintext, in main() 1800 if (do_aes_xts(tv->key, tv->key_len, tv->seqno, tv->ciphertext, in main()
|
| /openbsd-src/usr.bin/ssh/ |
| H A D | mac.c | 160 mac_compute(struct sshmac *mac, u_int32_t seqno, in mac_compute() argument 176 put_u32(b, seqno); in mac_compute() 185 POKE_U64(nonce, seqno); in mac_compute() 190 put_u64(nonce, seqno); in mac_compute() 206 mac_check(struct sshmac *mac, u_int32_t seqno, in mac_check() argument 215 if ((r = mac_compute(mac, seqno, data, dlen, in mac_check()
|
| /openbsd-src/sys/dev/pci/drm/scheduler/ |
| H A D | gpu_scheduler_trace.h | 93 __field(unsigned, seqno) 101 __entry->seqno = fence->seqno; 106 __entry->seqno)
|
| /openbsd-src/sys/net/ |
| H A D | ppp-deflate.c | 56 int seqno; member 201 state->seqno = 0; in z_comp_init() 216 state->seqno = 0; in z_comp_reset() 265 wptr[0] = state->seqno >> 8; in z_compress() 266 wptr[1] = state->seqno; in z_compress() 276 ++state->seqno; in z_compress() 421 state->seqno = 0; in z_decomp_init() 437 state->seqno = 0; in z_decomp_reset() 484 if (seq != state->seqno) { in z_decompress() 487 state->unit, seq, state->seqno); in z_decompress() [all …]
|
| /openbsd-src/sys/dev/pci/drm/amd/amdgpu/ |
| H A D | amdgpu_trace.h | 173 __field(unsigned int, seqno) 183 __entry->seqno = job->base.s_fence->finished.seqno; 189 __entry->seqno, __get_str(ring), __entry->num_ibs) 199 __field(unsigned int, seqno) 208 __entry->seqno = job->base.s_fence->finished.seqno; 214 __entry->seqno, __get_str(ring), __entry->num_ibs) 525 __field(unsigned, seqno) 533 __entry->seqno = fence->seqno; 538 __entry->seqno)
|
| /openbsd-src/gnu/usr.bin/perl/pod/ |
| H A D | splitman | 4 if ($seqno = 1 .. /^\.TH/) { 5 unless ($seqno =~ /e/i) {
|
| /openbsd-src/sys/dev/pci/drm/include/linux/ |
| H A D | dma-fence.h | 20 uint64_t seqno; member 115 return __dma_fence_is_later(a->seqno, b->seqno, a->ops); in dma_fence_is_later()
|
| /openbsd-src/sys/dev/pci/ |
| H A D | qlereg.h | 479 u_int8_t seqno; member 520 u_int8_t seqno; member 536 u_int8_t seqno; member 545 u_int8_t seqno; member 585 u_int8_t seqno; member 624 u_int8_t seqno; member
|