drm/i915: Fix workarounds selftest, part 1
pin_map needs the ww lock, so ensure we pin both before submission. Signed-off-by: Maarten Lankhorst <maarten.lankhorst@linux.intel.com> Reviewed-by: Thomas Hellström <thomas.hellstrom@linux.intel.com> [danvet: Again pick older version just to side-step conflicts.] Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch> Link: https://patchwork.freedesktop.org/patch/msgid/20210128162612.927917-32-maarten.lankhorst@linux.intel.com Link: https://patchwork.freedesktop.org/patch/msgid/20210323155059.628690-32-maarten.lankhorst@linux.intel.com
This commit is contained in:
parent
f1ac8a0292
commit
74827b539c
@ -411,6 +411,9 @@ void i915_gem_object_writeback(struct drm_i915_gem_object *obj);
|
||||
void *__must_check i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
|
||||
enum i915_map_type type);
|
||||
|
||||
void *__must_check i915_gem_object_pin_map_unlocked(struct drm_i915_gem_object *obj,
|
||||
enum i915_map_type type);
|
||||
|
||||
void __i915_gem_object_flush_map(struct drm_i915_gem_object *obj,
|
||||
unsigned long offset,
|
||||
unsigned long size);
|
||||
|
@ -400,6 +400,18 @@ err_unpin:
|
||||
goto out_unlock;
|
||||
}
|
||||
|
||||
void *i915_gem_object_pin_map_unlocked(struct drm_i915_gem_object *obj,
|
||||
enum i915_map_type type)
|
||||
{
|
||||
void *ret;
|
||||
|
||||
i915_gem_object_lock(obj, NULL);
|
||||
ret = i915_gem_object_pin_map(obj, type);
|
||||
i915_gem_object_unlock(obj);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
void __i915_gem_object_flush_map(struct drm_i915_gem_object *obj,
|
||||
unsigned long offset,
|
||||
unsigned long size)
|
||||
|
@ -112,7 +112,7 @@ read_nonprivs(struct intel_context *ce)
|
||||
|
||||
i915_gem_object_set_cache_coherency(result, I915_CACHE_LLC);
|
||||
|
||||
cs = i915_gem_object_pin_map(result, I915_MAP_WB);
|
||||
cs = i915_gem_object_pin_map_unlocked(result, I915_MAP_WB);
|
||||
if (IS_ERR(cs)) {
|
||||
err = PTR_ERR(cs);
|
||||
goto err_obj;
|
||||
@ -218,7 +218,7 @@ static int check_whitelist(struct intel_context *ce)
|
||||
i915_gem_object_lock(results, NULL);
|
||||
intel_wedge_on_timeout(&wedge, engine->gt, HZ / 5) /* safety net! */
|
||||
err = i915_gem_object_set_to_cpu_domain(results, false);
|
||||
i915_gem_object_unlock(results);
|
||||
|
||||
if (intel_gt_is_wedged(engine->gt))
|
||||
err = -EIO;
|
||||
if (err)
|
||||
@ -246,6 +246,7 @@ static int check_whitelist(struct intel_context *ce)
|
||||
|
||||
i915_gem_object_unpin_map(results);
|
||||
out_put:
|
||||
i915_gem_object_unlock(results);
|
||||
i915_gem_object_put(results);
|
||||
return err;
|
||||
}
|
||||
@ -502,6 +503,7 @@ static int check_dirty_whitelist(struct intel_context *ce)
|
||||
|
||||
for (i = 0; i < engine->whitelist.count; i++) {
|
||||
u32 reg = i915_mmio_reg_offset(engine->whitelist.list[i].reg);
|
||||
struct i915_gem_ww_ctx ww;
|
||||
u64 addr = scratch->node.start;
|
||||
struct i915_request *rq;
|
||||
u32 srm, lrm, rsvd;
|
||||
@ -517,6 +519,29 @@ static int check_dirty_whitelist(struct intel_context *ce)
|
||||
|
||||
ro_reg = ro_register(reg);
|
||||
|
||||
i915_gem_ww_ctx_init(&ww, false);
|
||||
retry:
|
||||
cs = NULL;
|
||||
err = i915_gem_object_lock(scratch->obj, &ww);
|
||||
if (!err)
|
||||
err = i915_gem_object_lock(batch->obj, &ww);
|
||||
if (!err)
|
||||
err = intel_context_pin_ww(ce, &ww);
|
||||
if (err)
|
||||
goto out;
|
||||
|
||||
cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
|
||||
if (IS_ERR(cs)) {
|
||||
err = PTR_ERR(cs);
|
||||
goto out_ctx;
|
||||
}
|
||||
|
||||
results = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB);
|
||||
if (IS_ERR(results)) {
|
||||
err = PTR_ERR(results);
|
||||
goto out_unmap_batch;
|
||||
}
|
||||
|
||||
/* Clear non priv flags */
|
||||
reg &= RING_FORCE_TO_NONPRIV_ADDRESS_MASK;
|
||||
|
||||
@ -528,12 +553,6 @@ static int check_dirty_whitelist(struct intel_context *ce)
|
||||
pr_debug("%s: Writing garbage to %x\n",
|
||||
engine->name, reg);
|
||||
|
||||
cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
|
||||
if (IS_ERR(cs)) {
|
||||
err = PTR_ERR(cs);
|
||||
goto out_batch;
|
||||
}
|
||||
|
||||
/* SRM original */
|
||||
*cs++ = srm;
|
||||
*cs++ = reg;
|
||||
@ -580,11 +599,12 @@ static int check_dirty_whitelist(struct intel_context *ce)
|
||||
i915_gem_object_flush_map(batch->obj);
|
||||
i915_gem_object_unpin_map(batch->obj);
|
||||
intel_gt_chipset_flush(engine->gt);
|
||||
cs = NULL;
|
||||
|
||||
rq = intel_context_create_request(ce);
|
||||
rq = i915_request_create(ce);
|
||||
if (IS_ERR(rq)) {
|
||||
err = PTR_ERR(rq);
|
||||
goto out_batch;
|
||||
goto out_unmap_scratch;
|
||||
}
|
||||
|
||||
if (engine->emit_init_breadcrumb) { /* Be nice if we hang */
|
||||
@ -593,20 +613,16 @@ static int check_dirty_whitelist(struct intel_context *ce)
|
||||
goto err_request;
|
||||
}
|
||||
|
||||
i915_vma_lock(batch);
|
||||
err = i915_request_await_object(rq, batch->obj, false);
|
||||
if (err == 0)
|
||||
err = i915_vma_move_to_active(batch, rq, 0);
|
||||
i915_vma_unlock(batch);
|
||||
if (err)
|
||||
goto err_request;
|
||||
|
||||
i915_vma_lock(scratch);
|
||||
err = i915_request_await_object(rq, scratch->obj, true);
|
||||
if (err == 0)
|
||||
err = i915_vma_move_to_active(scratch, rq,
|
||||
EXEC_OBJECT_WRITE);
|
||||
i915_vma_unlock(scratch);
|
||||
if (err)
|
||||
goto err_request;
|
||||
|
||||
@ -622,13 +638,7 @@ err_request:
|
||||
pr_err("%s: Futzing %x timedout; cancelling test\n",
|
||||
engine->name, reg);
|
||||
intel_gt_set_wedged(engine->gt);
|
||||
goto out_batch;
|
||||
}
|
||||
|
||||
results = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB);
|
||||
if (IS_ERR(results)) {
|
||||
err = PTR_ERR(results);
|
||||
goto out_batch;
|
||||
goto out_unmap_scratch;
|
||||
}
|
||||
|
||||
GEM_BUG_ON(values[ARRAY_SIZE(values) - 1] != 0xffffffff);
|
||||
@ -639,7 +649,7 @@ err_request:
|
||||
pr_err("%s: Unable to write to whitelisted register %x\n",
|
||||
engine->name, reg);
|
||||
err = -EINVAL;
|
||||
goto out_unpin;
|
||||
goto out_unmap_scratch;
|
||||
}
|
||||
} else {
|
||||
rsvd = 0;
|
||||
@ -705,15 +715,27 @@ err_request:
|
||||
|
||||
err = -EINVAL;
|
||||
}
|
||||
out_unpin:
|
||||
out_unmap_scratch:
|
||||
i915_gem_object_unpin_map(scratch->obj);
|
||||
out_unmap_batch:
|
||||
if (cs)
|
||||
i915_gem_object_unpin_map(batch->obj);
|
||||
out_ctx:
|
||||
intel_context_unpin(ce);
|
||||
out:
|
||||
if (err == -EDEADLK) {
|
||||
err = i915_gem_ww_ctx_backoff(&ww);
|
||||
if (!err)
|
||||
goto retry;
|
||||
}
|
||||
i915_gem_ww_ctx_fini(&ww);
|
||||
if (err)
|
||||
break;
|
||||
}
|
||||
|
||||
if (igt_flush_test(engine->i915))
|
||||
err = -EIO;
|
||||
out_batch:
|
||||
|
||||
i915_vma_unpin_and_release(&batch, 0);
|
||||
out_scratch:
|
||||
i915_vma_unpin_and_release(&scratch, 0);
|
||||
@ -847,7 +869,7 @@ static int scrub_whitelisted_registers(struct intel_context *ce)
|
||||
if (IS_ERR(batch))
|
||||
return PTR_ERR(batch);
|
||||
|
||||
cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
|
||||
cs = i915_gem_object_pin_map_unlocked(batch->obj, I915_MAP_WC);
|
||||
if (IS_ERR(cs)) {
|
||||
err = PTR_ERR(cs);
|
||||
goto err_batch;
|
||||
@ -982,11 +1004,11 @@ check_whitelisted_registers(struct intel_engine_cs *engine,
|
||||
u32 *a, *b;
|
||||
int i, err;
|
||||
|
||||
a = i915_gem_object_pin_map(A->obj, I915_MAP_WB);
|
||||
a = i915_gem_object_pin_map_unlocked(A->obj, I915_MAP_WB);
|
||||
if (IS_ERR(a))
|
||||
return PTR_ERR(a);
|
||||
|
||||
b = i915_gem_object_pin_map(B->obj, I915_MAP_WB);
|
||||
b = i915_gem_object_pin_map_unlocked(B->obj, I915_MAP_WB);
|
||||
if (IS_ERR(b)) {
|
||||
err = PTR_ERR(b);
|
||||
goto err_a;
|
||||
|
Loading…
Reference in New Issue
Block a user