| /netbsd-src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
| H A D | igt_spinner.c | 52 spin->batch = vaddr; in igt_spinner_init() 102 u32 *batch; in igt_spinner_create_request() local 140 batch = spin->batch; in igt_spinner_create_request() 143 *batch++ = MI_STORE_DWORD_IMM_GEN4; in igt_spinner_create_request() 144 *batch++ = lower_32_bits(hws_address(hws, rq)); in igt_spinner_create_request() 145 *batch++ = upper_32_bits(hws_address(hws, rq)); in igt_spinner_create_request() 147 *batch++ = MI_STORE_DWORD_IMM_GEN4; in igt_spinner_create_request() 148 *batch++ = 0; in igt_spinner_create_request() 149 *batch++ = hws_address(hws, rq); in igt_spinner_create_request() 151 *batch++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT; in igt_spinner_create_request() [all …]
|
| H A D | i915_request.c | 654 struct i915_vma *batch) in empty_request() argument 664 batch->node.start, in empty_request() 665 batch->node.size, in empty_request() 681 struct i915_vma *batch; in live_empty_request() local 690 batch = empty_batch(i915); in live_empty_request() 691 if (IS_ERR(batch)) in live_empty_request() 692 return PTR_ERR(batch); in live_empty_request() 707 request = empty_request(engine, batch); in live_empty_request() 720 request = empty_request(engine, batch); in live_empty_request() 750 i915_vma_unpin(batch); in live_empty_request() [all …]
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
| H A D | i915_gem_object_blt.c | 27 struct i915_vma *batch; in intel_emit_vma_fill_blt() local 90 batch = i915_vma_instance(pool->obj, ce->vm, NULL); in intel_emit_vma_fill_blt() 91 if (IS_ERR(batch)) { in intel_emit_vma_fill_blt() 92 err = PTR_ERR(batch); in intel_emit_vma_fill_blt() 96 err = i915_vma_pin(batch, 0, 0, PIN_USER); in intel_emit_vma_fill_blt() 100 batch->private = pool; in intel_emit_vma_fill_blt() 101 return batch; in intel_emit_vma_fill_blt() 137 struct i915_vma *batch; in i915_gem_object_fill_blt() local 155 batch = intel_emit_vma_fill_blt(ce, vma, value); in i915_gem_object_fill_blt() 156 if (IS_ERR(batch)) { in i915_gem_object_fill_blt() [all …]
|
| H A D | i915_gem_execbuffer.c | 240 struct i915_vma *batch; /** identity of the batch obj/vma */ member 534 eb->batch = vma; in eb_add_vma() 552 eb->batch = NULL; in eb_add_vma() 743 unsigned int i, batch; in eb_lookup_vmas() local 749 batch = eb_batch_index(eb); in eb_lookup_vmas() 801 err = eb_add_vma(eb, i, batch, vma); in eb_lookup_vmas() 929 GEM_BUG_ON(cache->rq_size >= cache->rq->batch->obj->base.size / sizeof(u32)); in reloc_gpu_flush() 932 __i915_gem_object_flush_map(cache->rq->batch->obj, 0, cache->rq_size); in reloc_gpu_flush() 933 i915_gem_object_unpin_map(cache->rq->batch->obj); in reloc_gpu_flush() 1155 struct i915_vma *batch; in __reloc_gpu_alloc() local [all …]
|
| H A D | i915_gem_client_blt.c | 166 struct i915_vma *batch; in clear_pages_worker() local 184 batch = intel_emit_vma_fill_blt(w->ce, vma, w->value); in clear_pages_worker() 185 if (IS_ERR(batch)) { in clear_pages_worker() 186 err = PTR_ERR(batch); in clear_pages_worker() 201 err = intel_emit_vma_mark_active(batch, rq); in clear_pages_worker() 221 batch->node.start, batch->node.size, in clear_pages_worker() 231 intel_emit_vma_release(w->ce, batch); in clear_pages_worker()
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
| H A D | igt_gem_utils.c | 117 struct i915_vma *batch; in igt_gpu_fill_dw() local 124 batch = igt_emit_store_dw(vma, offset, count, val); in igt_gpu_fill_dw() 125 if (IS_ERR(batch)) in igt_gpu_fill_dw() 126 return PTR_ERR(batch); in igt_gpu_fill_dw() 139 batch->node.start, batch->node.size, in igt_gpu_fill_dw() 144 i915_vma_lock(batch); in igt_gpu_fill_dw() 145 err = i915_request_await_object(rq, batch->obj, false); in igt_gpu_fill_dw() 147 err = i915_vma_move_to_active(batch, rq, 0); in igt_gpu_fill_dw() 148 i915_vma_unlock(batch); in igt_gpu_fill_dw() 162 i915_vma_unpin_and_release(&batch, 0); in igt_gpu_fill_dw() [all …]
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
| H A D | vmwgfx_mob.c | 241 struct vmw_otable_batch *batch) in vmw_otable_batch_setup() argument 245 struct vmw_otable *otables = batch->otables; in vmw_otable_batch_setup() 254 for (i = 0; i < batch->num_otables; ++i) { in vmw_otable_batch_setup() 266 0, false, &batch->otable_bo); in vmw_otable_batch_setup() 271 ret = ttm_bo_reserve(batch->otable_bo, false, true, NULL); in vmw_otable_batch_setup() 273 ret = vmw_bo_driver.ttm_tt_populate(batch->otable_bo->ttm, &ctx); in vmw_otable_batch_setup() 276 ret = vmw_bo_map_dma(batch->otable_bo); in vmw_otable_batch_setup() 280 ttm_bo_unreserve(batch->otable_bo); in vmw_otable_batch_setup() 283 for (i = 0; i < batch->num_otables; ++i) { in vmw_otable_batch_setup() 284 if (!batch->otables[i].enabled) in vmw_otable_batch_setup() [all …]
|
| /netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
| H A D | selftest_hangcheck.c | 57 u32 *batch; member 100 h->batch = vaddr; in hang_init() 147 u32 *batch; in hang_create_request() local 167 h->batch = vaddr; in hang_create_request() 205 batch = h->batch; in hang_create_request() 207 *batch++ = MI_STORE_DWORD_IMM_GEN4; in hang_create_request() 208 *batch++ = lower_32_bits(hws_address(hws, rq)); in hang_create_request() 209 *batch++ = upper_32_bits(hws_address(hws, rq)); in hang_create_request() 210 *batch++ = rq->fence.seqno; in hang_create_request() 211 *batch++ = MI_ARB_CHECK; in hang_create_request() [all …]
|
| H A D | selftest_workarounds.c | 479 struct i915_vma *batch; in check_dirty_whitelist() local 487 batch = create_batch(ce->vm); in check_dirty_whitelist() 488 if (IS_ERR(batch)) { in check_dirty_whitelist() 489 err = PTR_ERR(batch); in check_dirty_whitelist() 518 cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC); in check_dirty_whitelist() 567 i915_gem_object_flush_map(batch->obj); in check_dirty_whitelist() 568 i915_gem_object_unpin_map(batch->obj); in check_dirty_whitelist() 583 i915_vma_lock(batch); in check_dirty_whitelist() 584 err = i915_request_await_object(rq, batch->obj, false); in check_dirty_whitelist() 586 err = i915_vma_move_to_active(batch, rq, 0); in check_dirty_whitelist() [all …]
|
| H A D | intel_lrc.c | 3109 gen8_emit_flush_coherentl3_wa(struct intel_engine_cs *engine, u32 *batch) in gen8_emit_flush_coherentl3_wa() argument 3112 *batch++ = MI_STORE_REGISTER_MEM_GEN8 | MI_SRM_LRM_GLOBAL_GTT; in gen8_emit_flush_coherentl3_wa() 3113 *batch++ = i915_mmio_reg_offset(GEN8_L3SQCREG4); in gen8_emit_flush_coherentl3_wa() 3114 *batch++ = intel_gt_scratch_offset(engine->gt, in gen8_emit_flush_coherentl3_wa() 3116 *batch++ = 0; in gen8_emit_flush_coherentl3_wa() 3118 *batch++ = MI_LOAD_REGISTER_IMM(1); in gen8_emit_flush_coherentl3_wa() 3119 *batch++ = i915_mmio_reg_offset(GEN8_L3SQCREG4); in gen8_emit_flush_coherentl3_wa() 3120 *batch++ = 0x40400000 | GEN8_LQSC_FLUSH_COHERENT_LINES; in gen8_emit_flush_coherentl3_wa() 3122 batch = gen8_emit_pipe_control(batch, in gen8_emit_flush_coherentl3_wa() 3127 *batch++ = MI_LOAD_REGISTER_MEM_GEN8 | MI_SRM_LRM_GLOBAL_GTT; in gen8_emit_flush_coherentl3_wa() [all …]
|
| H A D | intel_engine.h | 246 static inline u32 *gen8_emit_pipe_control(u32 *batch, u32 flags, u32 offset) in gen8_emit_pipe_control() argument 248 memset(batch, 0, 6 * sizeof(u32)); in gen8_emit_pipe_control() 250 batch[0] = GFX_OP_PIPE_CONTROL(6); in gen8_emit_pipe_control() 251 batch[1] = flags; in gen8_emit_pipe_control() 252 batch[2] = offset; in gen8_emit_pipe_control() 254 return batch + 6; in gen8_emit_pipe_control()
|
| /netbsd-src/sys/external/bsd/drm/dist/shared-core/ |
| H A D | i915_dma.c | 486 drm_i915_batchbuffer_t * batch) in i915_dispatch_batchbuffer() argument 489 struct drm_clip_rect __user *boxes = batch->cliprects; in i915_dispatch_batchbuffer() 490 int nbox = batch->num_cliprects; in i915_dispatch_batchbuffer() 494 if ((batch->start | batch->used) & 0x7) { in i915_dispatch_batchbuffer() 506 batch->DR1, batch->DR4); in i915_dispatch_batchbuffer() 515 OUT_RING(batch->start); in i915_dispatch_batchbuffer() 518 OUT_RING(batch->start | MI_BATCH_NON_SECURE); in i915_dispatch_batchbuffer() 524 OUT_RING(batch->start | MI_BATCH_NON_SECURE); in i915_dispatch_batchbuffer() 525 OUT_RING(batch->start + batch->used - 4); in i915_dispatch_batchbuffer() 613 drm_i915_batchbuffer_t *batch = data; in i915_batchbuffer() local [all …]
|
| /netbsd-src/usr.bin/mail/ |
| H A D | mime_detach.c | 62 int batch; member 71 detach_ctl.batch = value(ENAME_MIME_DETACH_BATCH) != NULL; in mime_detach_control() 72 detach_ctl.ask = detach_ctl.batch ? 0 : 1; in mime_detach_control() 97 if (!detach_ctl.batch) { in detach_get_fname() 146 detach_ctl.batch = 1; in detach_open_core() 155 detach_ctl.batch = 1; in detach_open_core() 237 detach_ctl.batch = 0; in detach_open_target() 242 } while (!detach_ctl.batch); in detach_open_target()
|
| /netbsd-src/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/ |
| H A D | sanitizer_quarantine.h | 32 void *batch[kSize]; member 36 batch[0] = ptr; in init() 47 batch[count++] = ptr; in push_back() 60 batch[count + i] = from->batch[i]; in merge() 187 CHECK(kPrefetch <= ARRAY_SIZE(b->batch)); in DoRecycle() 189 PREFETCH(b->batch[i]); in DoRecycle() 192 PREFETCH(b->batch[i + kPrefetch]); in DoRecycle() 193 cb.Recycle((Node*)b->batch[i]); in DoRecycle()
|
| H A D | sanitizer_allocator_local_cache.h | 168 void *res = c->batch[--c->count]; in Allocate() 169 PREFETCH(c->batch[c->count - 1]); in Allocate() 183 c->batch[c->count++] = p; in Deallocate() 211 void *batch[2 * TransferBatch::kMaxNumCached]; member 248 b->CopyToArray(c->batch); in Refill() 259 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]); in Drain() 267 b->SetFromArray(&c->batch[first_idx_to_drain], count); in Drain()
|
| /netbsd-src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/ |
| H A D | sanitizer_quarantine.h | 33 void *batch[kSize]; member 37 batch[0] = ptr; in init() 48 batch[count++] = ptr; in push_back() 61 batch[count + i] = from->batch[i]; in merge() 187 CHECK(kPrefetch <= ARRAY_SIZE(b->batch)); in DoRecycle() 189 PREFETCH(b->batch[i]); in DoRecycle() 192 PREFETCH(b->batch[i + kPrefetch]); in DoRecycle() 193 cb.Recycle((Node*)b->batch[i]); in DoRecycle()
|
| H A D | sanitizer_allocator_local_cache.h | 169 void *res = c->batch[--c->count]; in Allocate() 170 PREFETCH(c->batch[c->count - 1]); in Allocate() 184 c->batch[c->count++] = p; in Deallocate() 212 void *batch[2 * TransferBatch::kMaxNumCached]; member 249 b->CopyToArray(c->batch); in Refill() 260 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]); in Drain() 268 b->SetFromArray(&c->batch[first_idx_to_drain], count); in Drain()
|
| /netbsd-src/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/ |
| H A D | sanitizer_quarantine.h | 31 void *batch[kSize]; member 35 batch[0] = ptr; in init() 46 batch[count++] = ptr; in push_back() 59 batch[count + i] = from->batch[i]; in merge() 185 CHECK(kPrefetch <= ARRAY_SIZE(b->batch)); in DoRecycle() 187 PREFETCH(b->batch[i]); in DoRecycle() 190 PREFETCH(b->batch[i + kPrefetch]); in DoRecycle() 191 cb.Recycle((Node*)b->batch[i]); in DoRecycle()
|
| H A D | sanitizer_allocator_local_cache.h | 167 void *res = c->batch[--c->count]; in Allocate() 168 PREFETCH(c->batch[c->count - 1]); in Allocate() 182 c->batch[c->count++] = p; in Deallocate() 210 void *batch[2 * TransferBatch::kMaxNumCached]; member 247 b->CopyToArray(c->batch); in Refill() 258 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]); in Drain() 266 b->SetFromArray(&c->batch[first_idx_to_drain], count); in Drain()
|
| /netbsd-src/external/gpl3/gcc.old/dist/libsanitizer/tsan/ |
| H A D | tsan_dense_alloc.h | 110 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); in Refill() local 114 new(batch + i) T; in Refill() 115 *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size; in Refill() 117 *(IndexT*)(batch + kL2Size - 1) = 0; in Refill() 119 map_[fillpos_++] = batch; in Refill()
|
| /netbsd-src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/ |
| H A D | tsan_dense_alloc.h | 112 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); in Refill() local 116 new(batch + i) T; in Refill() 117 *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size; in Refill() 119 *(IndexT*)(batch + kL2Size - 1) = 0; in Refill() 121 map_[fillpos_++] = batch; in Refill()
|
| /netbsd-src/external/gpl3/gcc/dist/libsanitizer/tsan/ |
| H A D | tsan_dense_alloc.h | 125 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); in Refill() local 129 new(batch + i) T; in Refill() 130 *(IndexT *)(batch + i) = i + 1 + fillpos * kL2Size; in Refill() 132 *(IndexT*)(batch + kL2Size - 1) = 0; in Refill() 134 map_[fillpos] = batch; in Refill()
|
| /netbsd-src/external/bsd/openldap/dist/tests/data/tls/ |
| H A D | create-crt.sh | 59 -batch > /dev/null 2>&1 63 -batch >/dev/null 2>&1 74 -batch >/dev/null 2>&1 78 -cert ca/certs/testsuiteCA.crt -batch >/dev/null 2>&1
|
| /netbsd-src/tests/usr.bin/gdb/ |
| H A D | t_regress.sh | 48 gdb --batch -x test.gdb dig >gdb.out 68 gdb --batch -x test.gdb ./test >gdb.out 2>&1 87 gdb --batch -x test.gdb >gdb.out 2>&1
|
| /netbsd-src/external/mpl/bind/dist/bin/tests/system/ |
| H A D | get_core_dumps.sh | 33 …binary=$(gdb --batch --core="$coredump" 2>/dev/null | sed -ne "s|Core was generated by \`\([^' ]*\… 38 -batch \ 47 -batch \
|