Lines Matching refs:ttm

71 	if (bo->ttm)  in ttm_tt_create()
98 bo->ttm = bdev->funcs->ttm_tt_create(bo, page_flags); in ttm_tt_create()
99 if (unlikely(bo->ttm == NULL)) in ttm_tt_create()
102 WARN_ON(bo->ttm->page_flags & TTM_TT_FLAG_EXTERNAL_MAPPABLE && in ttm_tt_create()
103 !(bo->ttm->page_flags & TTM_TT_FLAG_EXTERNAL)); in ttm_tt_create()
111 static int ttm_tt_alloc_page_directory(struct ttm_tt *ttm) in ttm_tt_alloc_page_directory() argument
113 ttm->pages = kvcalloc(ttm->num_pages, sizeof(void*), GFP_KERNEL); in ttm_tt_alloc_page_directory()
114 if (!ttm->pages) in ttm_tt_alloc_page_directory()
116 ttm->orders = kvmalloc_array(ttm->num_pages, in ttm_tt_alloc_page_directory()
118 if (!ttm->orders) in ttm_tt_alloc_page_directory()
123 static int ttm_dma_tt_alloc_page_directory(struct ttm_tt *ttm) in ttm_dma_tt_alloc_page_directory() argument
125 ttm->pages = kvcalloc(ttm->num_pages, sizeof(*ttm->pages) + in ttm_dma_tt_alloc_page_directory()
126 sizeof(*ttm->dma_address), GFP_KERNEL); in ttm_dma_tt_alloc_page_directory()
127 if (!ttm->pages) in ttm_dma_tt_alloc_page_directory()
130 ttm->dma_address = (void *)(ttm->pages + ttm->num_pages); in ttm_dma_tt_alloc_page_directory()
132 ttm->orders = kvmalloc_array(ttm->num_pages, in ttm_dma_tt_alloc_page_directory()
135 if (!ttm->orders) in ttm_dma_tt_alloc_page_directory()
140 static int ttm_sg_tt_alloc_page_directory(struct ttm_tt *ttm) in ttm_sg_tt_alloc_page_directory() argument
142 ttm->dma_address = kvcalloc(ttm->num_pages, sizeof(*ttm->dma_address), in ttm_sg_tt_alloc_page_directory()
144 if (!ttm->dma_address) in ttm_sg_tt_alloc_page_directory()
150 void ttm_tt_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) in ttm_tt_destroy() argument
152 bdev->funcs->ttm_tt_destroy(bdev, ttm); in ttm_tt_destroy()
155 static void ttm_tt_init_fields(struct ttm_tt *ttm, in ttm_tt_init_fields() argument
161 ttm->num_pages = (PAGE_ALIGN(bo->base.size) >> PAGE_SHIFT) + extra_pages; in ttm_tt_init_fields()
162 ttm->page_flags = page_flags; in ttm_tt_init_fields()
163 ttm->dma_address = NULL; in ttm_tt_init_fields()
164 ttm->swap_storage = NULL; in ttm_tt_init_fields()
165 ttm->sg = bo->sg; in ttm_tt_init_fields()
166 ttm->caching = caching; in ttm_tt_init_fields()
167 ttm->dmat = bo->bdev->dmat; in ttm_tt_init_fields()
168 ttm->map = NULL; in ttm_tt_init_fields()
169 ttm->segs = NULL; in ttm_tt_init_fields()
172 int ttm_tt_init(struct ttm_tt *ttm, struct ttm_buffer_object *bo, in ttm_tt_init() argument
176 ttm_tt_init_fields(ttm, bo, page_flags, caching, extra_pages); in ttm_tt_init()
178 if (ttm_tt_alloc_page_directory(ttm)) { in ttm_tt_init()
186 void ttm_tt_fini(struct ttm_tt *ttm) in ttm_tt_fini() argument
188 WARN_ON(ttm->page_flags & TTM_TT_FLAG_PRIV_POPULATED); in ttm_tt_fini()
190 if (ttm->swap_storage) in ttm_tt_fini()
191 uao_detach(ttm->swap_storage); in ttm_tt_fini()
192 ttm->swap_storage = NULL; in ttm_tt_fini()
194 if (ttm->pages) in ttm_tt_fini()
195 kvfree(ttm->pages); in ttm_tt_fini()
197 kvfree(ttm->dma_address); in ttm_tt_fini()
198 kvfree(ttm->orders); in ttm_tt_fini()
199 ttm->pages = NULL; in ttm_tt_fini()
200 ttm->dma_address = NULL; in ttm_tt_fini()
201 ttm->orders = NULL; in ttm_tt_fini()
203 if (ttm->map) in ttm_tt_fini()
204 bus_dmamap_destroy(ttm->dmat, ttm->map); in ttm_tt_fini()
205 if (ttm->segs) in ttm_tt_fini()
206 km_free(ttm->segs, round_page(ttm->num_pages * in ttm_tt_fini()
211 int ttm_sg_tt_init(struct ttm_tt *ttm, struct ttm_buffer_object *bo, in ttm_sg_tt_init() argument
217 ttm_tt_init_fields(ttm, bo, page_flags, caching, 0); in ttm_sg_tt_init()
220 ret = ttm_sg_tt_alloc_page_directory(ttm); in ttm_sg_tt_init()
222 ret = ttm_dma_tt_alloc_page_directory(ttm); in ttm_sg_tt_init()
228 ttm->segs = km_alloc(round_page(ttm->num_pages * in ttm_sg_tt_init()
233 if (bus_dmamap_create(ttm->dmat, ttm->num_pages << PAGE_SHIFT, in ttm_sg_tt_init()
234 ttm->num_pages, ttm->num_pages << PAGE_SHIFT, 0, flags, in ttm_sg_tt_init()
235 &ttm->map)) { in ttm_sg_tt_init()
236 km_free(ttm->segs, round_page(ttm->num_pages * in ttm_sg_tt_init()
238 if (ttm->pages) { in ttm_sg_tt_init()
239 kvfree(ttm->pages); in ttm_sg_tt_init()
240 kvfree(ttm->orders); in ttm_sg_tt_init()
242 kvfree(ttm->dma_address); in ttm_sg_tt_init()
243 ttm->pages = NULL; in ttm_sg_tt_init()
244 ttm->orders = NULL; in ttm_sg_tt_init()
245 ttm->dma_address = NULL; in ttm_sg_tt_init()
254 int ttm_tt_swapin(struct ttm_tt *ttm) in ttm_tt_swapin() argument
262 swap_storage = ttm->swap_storage; in ttm_tt_swapin()
266 if (uvm_obj_wire(swap_storage, 0, ttm->num_pages << PAGE_SHIFT, in ttm_tt_swapin()
273 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapin()
274 to_page = ttm->pages[i]; in ttm_tt_swapin()
284 uvm_obj_unwire(swap_storage, 0, ttm->num_pages << PAGE_SHIFT); in ttm_tt_swapin()
287 ttm->swap_storage = NULL; in ttm_tt_swapin()
288 ttm->page_flags &= ~TTM_TT_FLAG_SWAPPED; in ttm_tt_swapin()
306 int ttm_tt_swapout(struct ttm_device *bdev, struct ttm_tt *ttm, in ttm_tt_swapout() argument
312 loff_t size = (loff_t)ttm->num_pages << PAGE_SHIFT; in ttm_tt_swapout()
326 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapout()
327 from_page = ttm->pages[i]; in ttm_tt_swapout()
342 ttm_tt_unpopulate(bdev, ttm); in ttm_tt_swapout()
343 ttm->swap_storage = swap_storage; in ttm_tt_swapout()
344 ttm->page_flags |= TTM_TT_FLAG_SWAPPED; in ttm_tt_swapout()
346 return ttm->num_pages; in ttm_tt_swapout()
356 struct ttm_tt *ttm, struct ttm_operation_ctx *ctx) in ttm_tt_populate() argument
360 if (!ttm) in ttm_tt_populate()
363 if (ttm_tt_is_populated(ttm)) in ttm_tt_populate()
366 if (!(ttm->page_flags & TTM_TT_FLAG_EXTERNAL)) { in ttm_tt_populate()
367 atomic_long_add(ttm->num_pages, &ttm_pages_allocated); in ttm_tt_populate()
369 atomic_long_add(ttm->num_pages, in ttm_tt_populate()
385 ret = bdev->funcs->ttm_tt_populate(bdev, ttm, ctx); in ttm_tt_populate()
387 ret = ttm_pool_alloc(&bdev->pool, ttm, ctx); in ttm_tt_populate()
391 ttm->page_flags |= TTM_TT_FLAG_PRIV_POPULATED; in ttm_tt_populate()
392 if (unlikely(ttm->page_flags & TTM_TT_FLAG_SWAPPED)) { in ttm_tt_populate()
393 ret = ttm_tt_swapin(ttm); in ttm_tt_populate()
395 ttm_tt_unpopulate(bdev, ttm); in ttm_tt_populate()
403 if (!(ttm->page_flags & TTM_TT_FLAG_EXTERNAL)) { in ttm_tt_populate()
404 atomic_long_sub(ttm->num_pages, &ttm_pages_allocated); in ttm_tt_populate()
406 atomic_long_sub(ttm->num_pages, in ttm_tt_populate()
413 void ttm_tt_unpopulate(struct ttm_device *bdev, struct ttm_tt *ttm) in ttm_tt_unpopulate() argument
415 if (!ttm_tt_is_populated(ttm)) in ttm_tt_unpopulate()
419 bdev->funcs->ttm_tt_unpopulate(bdev, ttm); in ttm_tt_unpopulate()
421 ttm_pool_free(&bdev->pool, ttm); in ttm_tt_unpopulate()
423 if (!(ttm->page_flags & TTM_TT_FLAG_EXTERNAL)) { in ttm_tt_unpopulate()
424 atomic_long_sub(ttm->num_pages, &ttm_pages_allocated); in ttm_tt_unpopulate()
426 atomic_long_sub(ttm->num_pages, in ttm_tt_unpopulate()
430 ttm->page_flags &= ~TTM_TT_FLAG_PRIV_POPULATED; in ttm_tt_unpopulate()