Lines Matching refs:vm
151 struct amdgpu_vm *vm; member
209 struct amdgpu_vm *vm; member
409 int amdgpu_vm_set_pasid(struct amdgpu_device *adev, struct amdgpu_vm *vm,
412 long amdgpu_vm_wait_idle(struct amdgpu_vm *vm, long timeout);
413 int amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm, int32_t xcp_id);
414 int amdgpu_vm_make_compute(struct amdgpu_device *adev, struct amdgpu_vm *vm);
415 void amdgpu_vm_release_compute(struct amdgpu_device *adev, struct amdgpu_vm *vm);
416 void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm);
417 int amdgpu_vm_lock_pd(struct amdgpu_vm *vm, struct drm_exec *exec,
419 bool amdgpu_vm_ready(struct amdgpu_vm *vm);
420 uint64_t amdgpu_vm_generation(struct amdgpu_device *adev, struct amdgpu_vm *vm);
421 int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
426 struct amdgpu_vm *vm, bool immediate);
428 struct amdgpu_vm *vm,
431 struct amdgpu_vm *vm);
433 struct amdgpu_vm *vm, struct amdgpu_bo *bo);
434 int amdgpu_vm_update_range(struct amdgpu_device *adev, struct amdgpu_vm *vm,
447 struct amdgpu_bo_va *amdgpu_vm_bo_find(struct amdgpu_vm *vm,
450 struct amdgpu_vm *vm,
464 struct amdgpu_vm *vm,
466 struct amdgpu_bo_va_mapping *amdgpu_vm_bo_lookup_mapping(struct amdgpu_vm *vm,
468 void amdgpu_vm_bo_trace_cs(struct amdgpu_vm *vm, struct ww_acquire_ctx *ticket);
485 void amdgpu_vm_set_task_info(struct amdgpu_vm *vm);
488 struct amdgpu_vm *vm);
489 void amdgpu_vm_get_memory(struct amdgpu_vm *vm,
492 int amdgpu_vm_pt_clear(struct amdgpu_device *adev, struct amdgpu_vm *vm,
494 int amdgpu_vm_pt_create(struct amdgpu_device *adev, struct amdgpu_vm *vm,
497 void amdgpu_vm_pt_free_root(struct amdgpu_device *adev, struct amdgpu_vm *vm);
499 struct amdgpu_vm *vm);
509 void amdgpu_debugfs_vm_bo_info(struct amdgpu_vm *vm, struct seq_file *m);
512 int amdgpu_vm_pt_map_tables(struct amdgpu_device *adev, struct amdgpu_vm *vm);
521 static inline uint64_t amdgpu_vm_tlb_seq(struct amdgpu_vm *vm) in amdgpu_vm_tlb_seq() argument
532 lock = vm->last_tlb_flush->lock; in amdgpu_vm_tlb_seq()
538 return atomic64_read(&vm->tlb_seq); in amdgpu_vm_tlb_seq()
546 static inline void amdgpu_vm_eviction_lock(struct amdgpu_vm *vm) in amdgpu_vm_eviction_lock() argument
548 mutex_lock(&vm->eviction_lock); in amdgpu_vm_eviction_lock()
550 vm->saved_flags = memalloc_noreclaim_save(); in amdgpu_vm_eviction_lock()
554 static inline bool amdgpu_vm_eviction_trylock(struct amdgpu_vm *vm) in amdgpu_vm_eviction_trylock() argument
556 if (mutex_trylock(&vm->eviction_lock)) { in amdgpu_vm_eviction_trylock()
558 vm->saved_flags = memalloc_noreclaim_save(); in amdgpu_vm_eviction_trylock()
565 static inline void amdgpu_vm_eviction_unlock(struct amdgpu_vm *vm) in amdgpu_vm_eviction_unlock() argument
568 memalloc_noreclaim_restore(vm->saved_flags); in amdgpu_vm_eviction_unlock()
570 mutex_unlock(&vm->eviction_lock); in amdgpu_vm_eviction_unlock()