Lines Matching refs:tbo

35 static void qxl_ttm_bo_destroy(struct ttm_buffer_object *tbo)  in qxl_ttm_bo_destroy()  argument
40 bo = to_qxl_bo(tbo); in qxl_ttm_bo_destroy()
41 qdev = (struct qxl_device *)bo->tbo.base.dev->dev_private; in qxl_ttm_bo_destroy()
48 drm_gem_object_release(&bo->tbo.base); in qxl_ttm_bo_destroy()
67 if (qbo->tbo.base.size <= PAGE_SIZE) in qxl_ttm_placement_from_domain()
121 r = drm_gem_object_init(&qdev->ddev, &bo->tbo.base, size); in qxl_bo_create()
126 bo->tbo.base.funcs = &qxl_object_funcs; in qxl_bo_create()
137 r = ttm_bo_init(&qdev->mman.bdev, &bo->tbo, size, type, in qxl_bo_create()
162 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); in qxl_bo_kmap()
179 if (bo->tbo.mem.mem_type == TTM_PL_VRAM) in qxl_bo_kmap_atomic_page()
181 else if (bo->tbo.mem.mem_type == TTM_PL_PRIV) in qxl_bo_kmap_atomic_page()
186 ret = qxl_ttm_io_mem_reserve(bo->tbo.bdev, &bo->tbo.mem); in qxl_bo_kmap_atomic_page()
188 return io_mapping_map_atomic_wc(map, bo->tbo.mem.bus.offset + page_offset); in qxl_bo_kmap_atomic_page()
217 if ((bo->tbo.mem.mem_type != TTM_PL_VRAM) && in qxl_bo_kunmap_atomic_page()
218 (bo->tbo.mem.mem_type != TTM_PL_PRIV)) in qxl_bo_kunmap_atomic_page()
232 drm_gem_object_put_unlocked(&(*bo)->tbo.base); in qxl_bo_unref()
238 drm_gem_object_get(&bo->tbo.base); in qxl_bo_ref()
245 struct drm_device *ddev = bo->tbo.base.dev; in __qxl_bo_pin()
253 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in __qxl_bo_pin()
265 struct drm_device *ddev = bo->tbo.base.dev; in __qxl_bo_unpin()
277 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in __qxl_bo_unpin()
328 &bo->tbo.base, bo, (unsigned long)bo->tbo.base.size, in qxl_bo_force_delete()
329 *((unsigned long *)&bo->tbo.base.refcount)); in qxl_bo_force_delete()
334 drm_gem_object_put_unlocked(&bo->tbo.base); in qxl_bo_force_delete()