| /netbsd-src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
| H A D | vmwgfx_irq.c | 156 static bool vmw_fifo_idle(struct vmw_private *dev_priv, uint32_t seqno) in vmw_fifo_idle() argument 166 uint32_t seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE); in vmw_update_seqno() local 170 if (dev_priv->last_read_seqno != seqno) { in vmw_update_seqno() 171 dev_priv->last_read_seqno = seqno; in vmw_update_seqno() 172 vmw_marker_pull(&fifo_state->marker_queue, seqno); in vmw_update_seqno() 178 uint32_t seqno) in vmw_seqno_passed() argument 185 if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) in vmw_seqno_passed() 190 if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) in vmw_seqno_passed() 194 vmw_fifo_idle(dev_priv, seqno)) in vmw_seqno_passed() 202 ret = ((atomic_read(&dev_priv->marker_seq) - seqno) in vmw_seqno_passed() [all …]
|
| H A D | vmwgfx_marker.c | 38 uint32_t seqno; member 61 uint32_t seqno) in vmw_marker_push() argument 68 marker->seqno = seqno; in vmw_marker_push() 95 if (signaled_seqno - marker->seqno > (1 << 30)) in vmw_marker_pull() 136 uint32_t seqno; in vmw_wait_lag() local 142 seqno = atomic_read(&dev_priv->marker_seq); in vmw_wait_lag() 146 seqno = marker->seqno; in vmw_wait_lag() 150 ret = vmw_wait_seqno(dev_priv, false, seqno, true, in vmw_wait_lag() 156 (void) vmw_marker_pull(queue, seqno); in vmw_wait_lag()
|
| H A D | vmwgfx_fence.c | 152 u32 seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE); in vmw_fence_enable_signaling() local 153 if (seqno - fence->base.seqno < VMW_FENCE_WRAP) in vmw_fence_enable_signaling() 378 struct vmw_fence_obj *fence, u32 seqno, in vmw_fence_obj_init() argument 384 fman->ctx, seqno); in vmw_fence_obj_init() 457 vmw_mmio_write(fence->base.seqno, in vmw_fence_goal_new_locked() 494 goal_seqno - fence->base.seqno < VMW_FENCE_WRAP)) in vmw_fence_goal_check_locked() 497 vmw_mmio_write(fence->base.seqno, fifo_mem + SVGA_FIFO_FENCE_GOAL); in vmw_fence_goal_check_locked() 508 uint32_t seqno, new_seqno; in __vmw_fences_update() local 511 seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE); in __vmw_fences_update() 514 if (seqno - fence->base.seqno < VMW_FENCE_WRAP) { in __vmw_fences_update() [all …]
|
| /netbsd-src/sys/external/bsd/drm2/linux/ |
| H A D | linux_dma_fence_chain.c | 51 struct dma_fence *fence, uint64_t seqno) in dma_fence_chain_init() argument 62 !__dma_fence_is_later(seqno, prev->seqno, prev->ops)) { in dma_fence_chain_init() 65 seqno = MAX(prev->seqno, seqno); in dma_fence_chain_init() 69 chain->prev_seqno = prev->seqno; in dma_fence_chain_init() 73 context, seqno); in dma_fence_chain_init() 289 dma_fence_chain_find_seqno(struct dma_fence **fencep, uint64_t seqno) in dma_fence_chain_find_seqno() argument 293 if (seqno == 0) in dma_fence_chain_find_seqno() 297 if (chain == NULL || chain->base.seqno < seqno) in dma_fence_chain_find_seqno() 302 to_dma_fence_chain(*fencep)->prev_seqno < seqno) in dma_fence_chain_find_seqno()
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
| H A D | i915_syncmap.c | 151 static int check_seqno(struct i915_syncmap *leaf, unsigned int idx, u32 seqno) in check_seqno() argument 159 if (__sync_seqno(leaf)[idx] != seqno) { in check_seqno() 161 __func__, idx, __sync_seqno(leaf)[idx], seqno); in check_seqno() 168 static int check_one(struct i915_syncmap **sync, u64 context, u32 seqno) in check_one() argument 172 err = i915_syncmap_set(sync, context, seqno); in check_one() 194 err = check_seqno((*sync), ilog2((*sync)->bitmap), seqno); in check_one() 198 if (!i915_syncmap_is_later(sync, context, seqno)) { in check_one() 200 context, seqno); in check_one() 243 static int check_leaf(struct i915_syncmap **sync, u64 context, u32 seqno) in check_leaf() argument 247 err = i915_syncmap_set(sync, context, seqno); in check_leaf() [all …]
|
| H A D | igt_spinner.c | 44 spin->seqno = memset(vaddr, 0xff, PAGE_SIZE); in igt_spinner_init() 158 *batch++ = rq->fence.seqno; in igt_spinner_create_request() 204 u32 *seqno = spin->seqno + seqno_offset(rq->fence.context); in hws_seqno() local 206 return READ_ONCE(*seqno); in hws_seqno() 229 rq->fence.seqno), in igt_wait_for_spinner() 232 rq->fence.seqno), in igt_wait_for_spinner()
|
| /netbsd-src/crypto/external/bsd/heimdal/dist/lib/krb5/ |
| H A D | generate_seq_number.c | 41 uint32_t *seqno) in krb5_generate_seq_number() argument 43 if (RAND_bytes((void *)seqno, sizeof(*seqno)) <= 0) in krb5_generate_seq_number() 46 *seqno &= 0x3fffffff; in krb5_generate_seq_number() 47 if (*seqno == 0) in krb5_generate_seq_number() 48 *seqno = 1; in krb5_generate_seq_number()
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/radeon/ |
| H A D | radeon_trace.h | 129 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), 131 TP_ARGS(dev, ring, seqno), 136 __field(u32, seqno) 142 __entry->seqno = seqno; 146 __entry->dev, __entry->ring, __entry->seqno) 151 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), 153 TP_ARGS(dev, ring, seqno) 158 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), 160 TP_ARGS(dev, ring, seqno) 165 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), [all …]
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
| H A D | intel_timeline.h | 58 u64 context, u32 seqno) in __intel_timeline_sync_set() argument 60 return i915_syncmap_set(&tl->sync, context, seqno); in __intel_timeline_sync_set() 66 return __intel_timeline_sync_set(tl, fence->context, fence->seqno); in intel_timeline_sync_set() 70 u64 context, u32 seqno) in __intel_timeline_sync_is_later() argument 72 return i915_syncmap_is_later(&tl->sync, context, seqno); in __intel_timeline_sync_is_later() 78 return __intel_timeline_sync_is_later(tl, fence->context, fence->seqno); in intel_timeline_sync_is_later() 85 u32 *seqno);
|
| H A D | selftest_timeline.c | 169 u32 seqno; member 181 if (__intel_timeline_sync_is_later(tl, ctx, p->seqno) != p->expected) { in __igt_sync() 183 name, p->name, ctx, p->seqno, yesno(p->expected)); in __igt_sync() 188 ret = __intel_timeline_sync_set(tl, ctx, p->seqno); in __igt_sync() 366 u32 seqno = prandom_u32_state(&prng); in bench_sync() local 368 if (!__intel_timeline_sync_is_later(&tl, id, seqno)) in bench_sync() 369 __intel_timeline_sync_set(&tl, id, seqno); in bench_sync() 496 if (*tl->hwsp_seqno != tl->seqno) { in checked_intel_timeline_create() 498 *tl->hwsp_seqno, tl->seqno); in checked_intel_timeline_create() 678 u32 seqno[2]; in live_hwsp_wrap() local [all …]
|
| H A D | intel_breadcrumbs.c | 94 return i915_seqno_passed(__hwsp_seqno(rq), rq->fence.seqno); in __request_completed() 101 i915_seqno_passed(rq->fence.seqno, in check_signal_order() 102 list_next_entry(rq, signal_link)->fence.seqno)) in check_signal_order() 106 i915_seqno_passed(list_prev_entry(rq, signal_link)->fence.seqno, in check_signal_order() 107 rq->fence.seqno)) in check_signal_order() 332 if (i915_seqno_passed(rq->fence.seqno, it->fence.seqno)) in i915_request_enable_breadcrumb() 390 (uint64_t)rq->fence.seqno, in intel_engine_print_breadcrumbs()
|
| H A D | intel_timeline.c | 397 GEM_BUG_ON(tl->seqno & tl->has_initial_breadcrumb); in timeline_advance() 399 return tl->seqno += 1 + tl->has_initial_breadcrumb; in timeline_advance() 404 tl->seqno -= 1 + tl->has_initial_breadcrumb; in timeline_rollback() 410 u32 *seqno) in __intel_timeline_get_seqno() argument 484 *seqno = timeline_advance(tl); in __intel_timeline_get_seqno() 485 GEM_BUG_ON(i915_seqno_passed(*tl->hwsp_seqno, *seqno)); in __intel_timeline_get_seqno() 499 u32 *seqno) in intel_timeline_get_seqno() argument 501 *seqno = timeline_advance(tl); in intel_timeline_get_seqno() 504 if (unlikely(!*seqno && tl->hwsp_cacheline)) in intel_timeline_get_seqno() 505 return __intel_timeline_get_seqno(tl, rq, seqno); in intel_timeline_get_seqno()
|
| H A D | selftest_hangcheck.c | 56 u32 *seqno; member 92 h->seqno = memset(vaddr, 0xff, PAGE_SIZE); in hang_init() 210 *batch++ = rq->fence.seqno; in hang_create_request() 224 *batch++ = rq->fence.seqno; in hang_create_request() 237 *batch++ = rq->fence.seqno; in hang_create_request() 249 *batch++ = rq->fence.seqno; in hang_create_request() 289 return READ_ONCE(h->seqno[rq->fence.context % (PAGE_SIZE/sizeof(u32))]); in hws_seqno() 311 rq->fence.seqno), in wait_until_running() 314 rq->fence.seqno), in wait_until_running() 607 __func__, rq->fence.seqno, hws_seqno(&h, rq)); in __igt_reset_engine() [all …]
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/i915/ |
| H A D | i915_syncmap.c | 159 bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno) in i915_syncmap_is_later() argument 200 return seqno_later(__sync_seqno(p)[idx], seqno); in i915_syncmap_is_later() 219 static inline void __sync_set_seqno(struct i915_syncmap *p, u64 id, u32 seqno) in __sync_set_seqno() argument 224 __sync_seqno(p)[idx] = seqno; in __sync_set_seqno() 235 static noinline int __sync_set(struct i915_syncmap **root, u64 id, u32 seqno) in __sync_set() argument 340 __sync_set_seqno(p, id, seqno); in __sync_set() 358 int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno) in i915_syncmap_set() argument 367 __sync_set_seqno(p, id, seqno); in i915_syncmap_set() 371 return __sync_set(root, id, seqno); in i915_syncmap_set()
|
| H A D | i915_request.h | 57 rq__->fence.context, rq__->fence.seqno, \ 422 u32 seqno; in hwsp_seqno() local 425 seqno = __hwsp_seqno(rq); in hwsp_seqno() 428 return seqno; in hwsp_seqno() 433 return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno - 1); in __i915_request_has_started() 513 return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno); in i915_request_completed() 518 rq->hwsp_seqno = (u32 *)&rq->fence.seqno; /* decouple from HWSP */ in i915_request_mark_complete()
|
| H A D | i915_syncmap.h | 36 int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno); 37 bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno);
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/virtio/ |
| H A D | virtgpu_fence.c | 52 if (WARN_ON_ONCE(fence->f.seqno == 0)) in virtio_fence_signaled() 56 if (atomic64_read(&fence->drv->last_seq) >= fence->f.seqno) in virtio_fence_signaled() 63 snprintf(str, size, "%llu", f->seqno); in virtio_fence_value_str() 108 fence->f.seqno = ++drv->sync_seq; in virtio_gpu_fence_emit() 116 cmd_hdr->fence_id = cpu_to_le64(fence->f.seqno); in virtio_gpu_fence_emit() 129 if (last_seq < fence->f.seqno) in virtio_gpu_fence_event_process()
|
| /netbsd-src/sys/net/ |
| H A D | ppp-deflate.c | 64 int seqno; member 208 state->seqno = 0; in z_comp_init() 223 state->seqno = 0; in z_comp_reset() 272 wptr[0] = state->seqno >> 8; in z_compress() 273 wptr[1] = state->seqno; in z_compress() 283 ++state->seqno; in z_compress() 425 state->seqno = 0; in z_decomp_init() 441 state->seqno = 0; in z_decomp_reset() 488 if (seq != state->seqno) { in z_decompress() 491 state->unit, seq, state->seqno); in z_decompress() [all …]
|
| H A D | bsd-comp.c | 93 uint16_t seqno; /* sequence # of next packet */ member 294 db->seqno = 0; in bsd_reset() 541 *wptr++ = db->seqno >> 8; in bsd_compress() 542 *wptr++ = db->seqno; in bsd_compress() 544 ++db->seqno; in bsd_compress() 691 db->seqno++; in bsd_incomp() 837 if (seq != db->seqno) { in bsd_decompress() 840 db->unit, seq, db->seqno - 1); in bsd_decompress() 843 ++db->seqno; in bsd_decompress() 924 max_ent, explen, db->seqno); in bsd_decompress()
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
| H A D | amdgpu_trace.h | 173 __field(unsigned int, seqno) 183 __entry->seqno = job->base.s_fence->finished.seqno; 189 __entry->seqno, __get_str(ring), __entry->num_ibs) 199 __field(unsigned int, seqno) 208 __entry->seqno = job->base.s_fence->finished.seqno; 214 __entry->seqno, __get_str(ring), __entry->num_ibs) 482 __field(unsigned, seqno) 490 __entry->seqno = fence->seqno; 495 __entry->seqno)
|
| /netbsd-src/crypto/external/bsd/openssh/dist/ |
| H A D | mac.c | 163 mac_compute(struct sshmac *mac, u_int32_t seqno, in mac_compute() argument 179 put_u32(b, seqno); in mac_compute() 188 POKE_U64(nonce, seqno); in mac_compute() 193 put_u64(nonce, seqno); in mac_compute() 209 mac_check(struct sshmac *mac, u_int32_t seqno, in mac_check() argument 218 if ((r = mac_compute(mac, seqno, data, dlen, in mac_check()
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/scheduler/ |
| H A D | gpu_scheduler_trace.h | 85 __field(unsigned, seqno) 93 __entry->seqno = fence->seqno; 98 __entry->seqno)
|
| /netbsd-src/sys/external/bsd/drm2/include/ |
| H A D | radeon_trace.h | 56 uint32_t seqno __unused) in trace_radeon_fence_emit() 62 int ring __unused, uint32_t seqno __unused) in trace_radeon_fence_wait_begin() 68 uint32_t seqno __unused) in trace_radeon_fence_wait_end()
|
| /netbsd-src/external/ibm-public/postfix/dist/src/fsstone/ |
| H A D | fsstone.c | 80 static void make_file(int seqno, int size) in make_file() argument 87 sprintf(path, "%06d", seqno); in make_file() 108 static void use_file(int seqno) in use_file() argument 114 sprintf(path, "%06d", seqno); in use_file()
|
| /netbsd-src/usr.sbin/timed/timed/ |
| H A D | measure.c | 63 static n_short seqno = 0; variable 135 oicp.icmp_seq = seqno; in measure() 212 || icp.icmp_seq < seqno in measure() 272 seqno += TRIALS; /* allocate our sequence numbers */ in measure()
|