/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gvt/ |
H A D | dmabuf.c | 460 vgpu->id, kref_read(&dmabuf_obj->kref), in intel_vgpu_query_plane() 517 __func__, kref_read(&dmabuf_obj->kref), ret); in intel_vgpu_query_plane() 584 kref_read(&dmabuf_obj->kref), in intel_vgpu_get_dmabuf() 587 kref_read(&obj->base.refcount)); in intel_vgpu_get_dmabuf()
|
/netbsd-src/sys/external/bsd/drm2/dist/include/drm/ttm/ |
H A D | ttm_bo_driver.h | 752 WARN_ON(!kref_read(&bo->kref)); in ttm_bo_reserve() 773 WARN_ON(!kref_read(&bo->kref)); in ttm_bo_reserve_slowpath()
|
/netbsd-src/sys/external/bsd/drm2/dist/include/drm/ |
H A D | drm_framebuffer.h | 261 return kref_read(&fb->base.refcount); in drm_framebuffer_read_refcount()
|
/netbsd-src/sys/external/bsd/drm2/include/linux/ |
H A D | kref.h | 163 kref_read(const struct kref *kref) in kref_read() function
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/ |
H A D | drm_mode_object.c | 202 DRM_DEBUG("OBJ ID: %d (%d)\n", obj->id, kref_read(&obj->refcount)); in drm_mode_object_put() 219 DRM_DEBUG("OBJ ID: %d (%d)\n", obj->id, kref_read(&obj->refcount)); in drm_mode_object_get()
|
H A D | drm_dp_mst_topology.c | 1366 DRM_DEBUG("mstb %p (%d)\n", mstb, kref_read(&mstb->malloc_kref)); in drm_dp_mst_get_mstb_malloc() 1383 DRM_DEBUG("mstb %p (%d)\n", mstb, kref_read(&mstb->malloc_kref) - 1); in drm_dp_mst_put_mstb_malloc() 1417 DRM_DEBUG("port %p (%d)\n", port, kref_read(&port->malloc_kref)); in drm_dp_mst_get_port_malloc() 1434 DRM_DEBUG("port %p (%d)\n", port, kref_read(&port->malloc_kref) - 1); in drm_dp_mst_put_port_malloc() 1660 mstb, kref_read(&mstb->topology_kref)); in drm_dp_mst_topology_try_get_mstb() 1688 WARN_ON(kref_read(&mstb->topology_kref) == 0); in drm_dp_mst_topology_get_mstb() 1690 DRM_DEBUG("mstb %p (%d)\n", mstb, kref_read(&mstb->topology_kref)); in drm_dp_mst_topology_get_mstb() 1713 mstb, kref_read(&mstb->topology_kref) - 1); in drm_dp_mst_topology_put_mstb() 1777 port, kref_read(&port->topology_kref)); in drm_dp_mst_topology_try_get_port() 1802 WARN_ON(kref_read(&port->topology_kref) == 0); in drm_dp_mst_topology_get_port() [all …]
|
H A D | drm_debugfs.c | 125 kref_read(&obj->refcount)); in drm_gem_one_name_info()
|
H A D | drm_gem.c | 1343 kref_read(&obj->refcount)); in drm_gem_print_info()
|
/netbsd-src/sys/external/bsd/drm2/linux/ |
H A D | linux_dma_fence_chain.c | 173 if (kref_read(&prev->refcount) > 1) in dma_fence_chain_release()
|
H A D | linux_dma_fence.c | 211 KASSERT(kref_read(&fence->refcount) == 0 || in dma_fence_reset() 212 kref_read(&fence->refcount) == 1); in dma_fence_reset()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
H A D | i915_gem_shrinker.c | 494 GEM_BUG_ON(!kref_read(&obj->base.refcount)); in __i915_gem_object_make_shrinkable()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/nouveau/ |
H A D | nouveau_fence.c | 516 WARN_ON(kref_read(&fence->base.refcount) <= 1); in nouveau_fence_no_signaling()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
H A D | ttm_object.c | 340 if (unlikely(kref_read(&ref->kref) == 0)) in ttm_ref_object_exists()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
H A D | amdgpu_ctx.c | 604 if (kref_read(&ctx->refcount) != 1) { in amdgpu_ctx_mgr_entity_fini()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/ttm/ |
H A D | ttm_bo.c | 180 BUG_ON(kref_read(&bo->list_kref)); in ttm_bo_release_list() 181 BUG_ON(kref_read(&bo->kref)); in ttm_bo_release_list()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/amd/display/amdgpu_dm/ |
H A D | amdgpu_dm.c | 1697 WARN_ON(kref_read(&dm_new_crtc_state->stream->refcount) > 1); in dm_resume() 1706 WARN_ON(kref_read(&dm_new_plane_state->dc_state->refcount) > 1); in dm_resume()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
H A D | intel_lrc.c | 1480 GEM_BUG_ON(!kref_read(&rq->fence.refcount)); in assert_pending_valid()
|