static int amdgpu_ttm_backend_bind(struct ttm_bo_device *bdev,
struct ttm_tt *ttm,
struct ttm_resource *bo_mem);
+static void amdgpu_ttm_backend_unbind(struct ttm_bo_device *bdev,
+ struct ttm_tt *ttm);
static int amdgpu_ttm_init_on_chip(struct amdgpu_device *adev,
unsigned int type,
goto out_cleanup;
}
- /* move BO (in tmp_mem) to new_mem */
- r = ttm_bo_move_to_system(bo, ctx);
+ r = ttm_bo_wait_ctx(bo, ctx);
if (unlikely(r))
goto out_cleanup;
+ amdgpu_ttm_backend_unbind(bo->bdev, bo->ttm);
+ ttm_resource_free(bo, &bo->mem);
ttm_bo_assign_mem(bo, new_mem);
out_cleanup:
ttm_resource_free(bo, &tmp_mem);
if (old_mem->mem_type == TTM_PL_TT &&
new_mem->mem_type == TTM_PL_SYSTEM) {
- r = ttm_bo_move_to_system(bo, ctx);
+ r = ttm_bo_wait_ctx(bo, ctx);
if (r)
return r;
+
+ amdgpu_ttm_backend_unbind(bo->bdev, bo->ttm);
+ ttm_resource_free(bo, &bo->mem);
ttm_bo_assign_mem(bo, new_mem);
return 0;
}
static int nouveau_ttm_tt_bind(struct ttm_bo_device *bdev, struct ttm_tt *ttm,
struct ttm_resource *reg);
+static void nouveau_ttm_tt_unbind(struct ttm_bo_device *bdev, struct ttm_tt *ttm);
/*
* NV10-NV40 tiling helpers
if (ret)
goto out;
- ret = ttm_bo_move_to_system(bo, ctx);
+ ret = ttm_bo_wait_ctx(bo, ctx);
if (ret)
goto out;
+ nouveau_ttm_tt_unbind(bo->bdev, bo->ttm);
+ ttm_resource_free(bo, &bo->mem);
ttm_bo_assign_mem(bo, &tmp_reg);
out:
ttm_resource_free(bo, &tmp_reg);
if (old_reg->mem_type == TTM_PL_TT &&
new_reg->mem_type == TTM_PL_SYSTEM) {
- ret = ttm_bo_move_to_system(bo, ctx);
- if (ret)
- return ret;
+ nouveau_ttm_tt_unbind(bo->bdev, bo->ttm);
+ ttm_resource_free(bo, &bo->mem);
ttm_bo_assign_mem(bo, new_reg);
goto out;
}
static int radeon_ttm_tt_bind(struct ttm_bo_device *bdev,
struct ttm_tt *ttm,
struct ttm_resource *bo_mem);
+static void radeon_ttm_tt_unbind(struct ttm_bo_device *bdev,
+ struct ttm_tt *ttm);
struct radeon_device *radeon_get_rdev(struct ttm_bo_device *bdev)
{
if (unlikely(r)) {
goto out_cleanup;
}
- r = ttm_bo_move_to_system(bo, ctx);
+ r = ttm_bo_wait_ctx(bo, ctx);
if (unlikely(r))
goto out_cleanup;
+ radeon_ttm_tt_unbind(bo->bdev, bo->ttm);
+ ttm_resource_free(bo, &bo->mem);
ttm_bo_assign_mem(bo, new_mem);
out_cleanup:
ttm_resource_free(bo, &tmp_mem);
if (old_mem->mem_type == TTM_PL_TT &&
new_mem->mem_type == TTM_PL_SYSTEM) {
- r = ttm_bo_move_to_system(bo, ctx);
- if (r)
- return r;
-
+ radeon_ttm_tt_unbind(bo->bdev, bo->ttm);
+ ttm_resource_free(bo, &bo->mem);
ttm_bo_assign_mem(bo, new_mem);
return 0;
}
return 0;
}
-int ttm_bo_move_to_system(struct ttm_buffer_object *bo,
- struct ttm_operation_ctx *ctx)
-{
- int ret;
-
- if (bo->mem.mem_type == TTM_PL_SYSTEM)
- return 0;
-
- ret = ttm_bo_wait_ctx(bo, ctx);
- if (unlikely(ret != 0)) {
- if (ret != -ERESTARTSYS)
- pr_err("Failed to expire sync object before unbinding TTM\n");
- return ret;
- }
-
- ttm_bo_tt_unbind(bo);
- ttm_resource_free(bo, &bo->mem);
- return 0;
-}
-EXPORT_SYMBOL(ttm_bo_move_to_system);
-
int ttm_mem_io_reserve(struct ttm_bo_device *bdev,
struct ttm_resource *mem)
{
ttm_bo_assign_mem(bo, new_mem);
return 0;
}
- ret = ttm_bo_move_to_system(bo, ctx);
+ ret = ttm_bo_wait_ctx(bo, ctx);
if (ret)
return ret;
+
+ vmw_ttm_unbind(bo->bdev, bo->ttm);
+ ttm_resource_free(bo, &bo->mem);
ttm_bo_assign_mem(bo, new_mem);
return 0;
} else {
int ttm_bo_move_to_new_tt_mem(struct ttm_buffer_object *bo,
struct ttm_operation_ctx *ctx,
struct ttm_resource *new_mem);
-int ttm_bo_move_to_system(struct ttm_buffer_object *bo,
- struct ttm_operation_ctx *ctx);
/**
* ttm_bo_move_memcpy