drm/i915/gvt: Use common mapping routines for indirect_ctx object
authorChris Wilson <chris@chris-wilson.co.uk>
Wed, 19 Oct 2016 10:11:45 +0000 (11:11 +0100)
committerZhenyu Wang <zhenyuw@linux.intel.com>
Thu, 20 Oct 2016 09:31:34 +0000 (17:31 +0800)
We have the ability to map an object, so use it rather than opencode it
badly.

Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Zhenyu Wang <zhenyuw@linux.intel.com>
drivers/gpu/drm/i915/gvt/cmd_parser.c
drivers/gpu/drm/i915/gvt/execlist.c

index d942da9a0c8a7e296506e3acb274acf703dbb058..153943a9411e4d948c3ef2d8a91c455cd20aea44 100644 (file)
@@ -2717,7 +2717,7 @@ static int shadow_indirect_ctx(struct intel_shadow_wa_ctx *wa_ctx)
        unsigned long guest_gma = wa_ctx->indirect_ctx.guest_gma;
        struct drm_i915_gem_object *obj;
        int ret = 0;
-       void *dest = NULL;
+       void *map;
 
        obj = i915_gem_object_create(dev,
                                     roundup(ctx_size + CACHELINE_BYTES,
@@ -2725,18 +2725,12 @@ static int shadow_indirect_ctx(struct intel_shadow_wa_ctx *wa_ctx)
        if (IS_ERR(obj))
                return PTR_ERR(obj);
 
-       ret = i915_gem_object_get_pages(obj);
-       if (ret)
-               goto put_obj;
-
-       i915_gem_object_pin_pages(obj);
-
        /* get the va of the shadow batch buffer */
-       dest = (void *)vmap_batch(obj, 0, ctx_size + CACHELINE_BYTES);
-       if (!dest) {
+       map = i915_gem_object_pin_map(obj, I915_MAP_WB);
+       if (IS_ERR(map)) {
                gvt_err("failed to vmap shadow indirect ctx\n");
-               ret = -ENOMEM;
-               goto unpin_src;
+               ret = PTR_ERR(map);
+               goto put_obj;
        }
 
        ret = i915_gem_object_set_to_cpu_domain(obj, false);
@@ -2745,25 +2739,21 @@ static int shadow_indirect_ctx(struct intel_shadow_wa_ctx *wa_ctx)
                goto unmap_src;
        }
 
-       wa_ctx->indirect_ctx.shadow_va = dest;
-
-       memset(dest, 0, round_up(ctx_size + CACHELINE_BYTES, PAGE_SIZE));
-
        ret = copy_gma_to_hva(wa_ctx->workload->vgpu,
                                wa_ctx->workload->vgpu->gtt.ggtt_mm,
-                               guest_gma, guest_gma + ctx_size, dest);
+                               guest_gma, guest_gma + ctx_size,
+                               map);
        if (ret) {
                gvt_err("fail to copy guest indirect ctx\n");
                goto unmap_src;
        }
 
        wa_ctx->indirect_ctx.obj = obj;
+       wa_ctx->indirect_ctx.shadow_va = map;
        return 0;
 
 unmap_src:
-       vunmap(dest);
-unpin_src:
-       i915_gem_object_unpin_pages(wa_ctx->indirect_ctx.obj);
+       i915_gem_object_unpin_map(obj);
 put_obj:
        i915_gem_object_put(wa_ctx->indirect_ctx.obj);
        return ret;
index 5534336814f00ce6a1f8682eab187d28e10b9325..88430ca23504b484e9a4dc21143a4518fb00a94e 100644 (file)
@@ -518,8 +518,8 @@ static void release_shadow_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx)
        if (wa_ctx->indirect_ctx.size == 0)
                return;
 
+       i915_gem_object_unpin_map(wa_ctx->indirect_ctx.obj);
        i915_gem_object_put(wa_ctx->indirect_ctx.obj);
-       kvfree(wa_ctx->indirect_ctx.shadow_va);
 }
 
 static int complete_execlist_workload(struct intel_vgpu_workload *workload)