Lines Matching refs:sync

455 static int vm_update_pds(struct amdgpu_vm *vm, struct amdgpu_sync *sync)
465 return amdgpu_sync_fence(sync, vm->last_update);
1068 * all the reservation info in a context structure. Optionally a sync
1076 /* Pointer to sync object */
1077 struct amdgpu_sync *sync;
1102 ctx->sync = &mem->sync;
1141 ctx->sync = &mem->sync;
1176 * @wait: Optionally wait for a sync object representing pending VM updates
1189 ret = amdgpu_sync_wait(ctx->sync, intr);
1192 ctx->sync = NULL;
1198 struct amdgpu_sync *sync)
1208 amdgpu_sync_fence(sync, bo_va->last_pt_update);
1213 struct amdgpu_sync *sync)
1230 return amdgpu_sync_fence(sync, bo_va->last_pt_update);
1235 struct amdgpu_sync *sync,
1253 ret = update_gpuvm_pte(mem, entry, sync);
1262 unmap_bo_from_gpuvm(mem, entry, sync);
1283 struct amdgpu_sync *sync)
1292 ret = amdgpu_sync_resv(NULL, sync, pd->tbo.base.resv,
1303 struct amdgpu_sync *sync)
1310 ret = vm_update_pds(peer_vm, sync);
1719 amdgpu_sync_create(&(*mem)->sync);
1792 amdgpu_sync_free(&(*mem)->sync);
1875 /* Free the sync object */
1876 amdgpu_sync_free(&mem->sync);
2010 ret = map_bo_to_gpuvm(mem, entry, ctx.sync,
2017 ret = vm_update_pds(avm, ctx.sync);
2113 unmap_bo_from_gpuvm(mem, entry, ctx.sync);
2140 struct amdgpu_sync sync;
2143 amdgpu_sync_create(&sync);
2146 amdgpu_sync_clone(&mem->sync, &sync);
2149 ret = amdgpu_sync_wait(&sync, intr);
2150 amdgpu_sync_free(&sync);
2354 amdgpu_sync_create(&(*mem)->sync);
2531 struct amdgpu_sync sync;
2539 amdgpu_sync_create(&sync);
2599 ret = update_gpuvm_pte(mem, attachment, &sync);
2612 ret = process_update_pds(process_info, &sync);
2616 amdgpu_sync_wait(&sync, false);
2617 amdgpu_sync_free(&sync);
2816 pr_debug("Memory eviction: Failed to sync to PD BO moving fence. Try again\n");
2940 amdgpu_sync_create(&(*mem)->sync);
2976 amdgpu_sync_free(&(*mem)->sync);
3006 amdgpu_sync_free(&kgd_mem->sync);