1 1.1 riastrad /* $NetBSD: selftest_lrc.c,v 1.2 2021/12/18 23:45:30 riastradh Exp $ */ 2 1.1 riastrad 3 1.1 riastrad /* 4 1.1 riastrad * SPDX-License-Identifier: MIT 5 1.1 riastrad * 6 1.1 riastrad * Copyright 2018 Intel Corporation 7 1.1 riastrad */ 8 1.1 riastrad 9 1.1 riastrad #include <sys/cdefs.h> 10 1.1 riastrad __KERNEL_RCSID(0, "$NetBSD: selftest_lrc.c,v 1.2 2021/12/18 23:45:30 riastradh Exp $"); 11 1.1 riastrad 12 1.1 riastrad #include <linux/prime_numbers.h> 13 1.1 riastrad 14 1.1 riastrad #include "gem/i915_gem_pm.h" 15 1.1 riastrad #include "gt/intel_engine_heartbeat.h" 16 1.1 riastrad #include "gt/intel_reset.h" 17 1.1 riastrad 18 1.1 riastrad #include "i915_selftest.h" 19 1.1 riastrad #include "selftests/i915_random.h" 20 1.1 riastrad #include "selftests/igt_flush_test.h" 21 1.1 riastrad #include "selftests/igt_live_test.h" 22 1.1 riastrad #include "selftests/igt_spinner.h" 23 1.1 riastrad #include "selftests/lib_sw_fence.h" 24 1.1 riastrad 25 1.1 riastrad #include "gem/selftests/igt_gem_utils.h" 26 1.1 riastrad #include "gem/selftests/mock_context.h" 27 1.1 riastrad 28 1.1 riastrad #define CS_GPR(engine, n) ((engine)->mmio_base + 0x600 + (n) * 4) 29 1.1 riastrad #define NUM_GPR_DW (16 * 2) /* each GPR is 2 dwords */ 30 1.1 riastrad 31 1.1 riastrad static struct i915_vma *create_scratch(struct intel_gt *gt) 32 1.1 riastrad { 33 1.1 riastrad struct drm_i915_gem_object *obj; 34 1.1 riastrad struct i915_vma *vma; 35 1.1 riastrad int err; 36 1.1 riastrad 37 1.1 riastrad obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE); 38 1.1 riastrad if (IS_ERR(obj)) 39 1.1 riastrad return ERR_CAST(obj); 40 1.1 riastrad 41 1.1 riastrad i915_gem_object_set_cache_coherency(obj, I915_CACHING_CACHED); 42 1.1 riastrad 43 1.1 riastrad vma = i915_vma_instance(obj, >->ggtt->vm, NULL); 44 1.1 riastrad if (IS_ERR(vma)) { 45 1.1 riastrad i915_gem_object_put(obj); 46 1.1 riastrad return vma; 47 1.1 riastrad } 48 1.1 riastrad 49 1.1 riastrad err = i915_vma_pin(vma, 0, 0, PIN_GLOBAL); 50 1.1 riastrad if (err) { 51 1.1 riastrad i915_gem_object_put(obj); 52 1.1 riastrad return ERR_PTR(err); 53 1.1 riastrad } 54 1.1 riastrad 55 1.1 riastrad return vma; 56 1.1 riastrad } 57 1.1 riastrad 58 1.1 riastrad static void engine_heartbeat_disable(struct intel_engine_cs *engine, 59 1.1 riastrad unsigned long *saved) 60 1.1 riastrad { 61 1.1 riastrad *saved = engine->props.heartbeat_interval_ms; 62 1.1 riastrad engine->props.heartbeat_interval_ms = 0; 63 1.1 riastrad 64 1.1 riastrad intel_engine_pm_get(engine); 65 1.1 riastrad intel_engine_park_heartbeat(engine); 66 1.1 riastrad } 67 1.1 riastrad 68 1.1 riastrad static void engine_heartbeat_enable(struct intel_engine_cs *engine, 69 1.1 riastrad unsigned long saved) 70 1.1 riastrad { 71 1.1 riastrad intel_engine_pm_put(engine); 72 1.1 riastrad 73 1.1 riastrad engine->props.heartbeat_interval_ms = saved; 74 1.1 riastrad } 75 1.1 riastrad 76 1.1 riastrad static int live_sanitycheck(void *arg) 77 1.1 riastrad { 78 1.1 riastrad struct intel_gt *gt = arg; 79 1.1 riastrad struct intel_engine_cs *engine; 80 1.1 riastrad enum intel_engine_id id; 81 1.1 riastrad struct igt_spinner spin; 82 1.1 riastrad int err = 0; 83 1.1 riastrad 84 1.1 riastrad if (!HAS_LOGICAL_RING_CONTEXTS(gt->i915)) 85 1.1 riastrad return 0; 86 1.1 riastrad 87 1.1 riastrad if (igt_spinner_init(&spin, gt)) 88 1.1 riastrad return -ENOMEM; 89 1.1 riastrad 90 1.1 riastrad for_each_engine(engine, gt, id) { 91 1.1 riastrad struct intel_context *ce; 92 1.1 riastrad struct i915_request *rq; 93 1.1 riastrad 94 1.1 riastrad ce = intel_context_create(engine); 95 1.1 riastrad if (IS_ERR(ce)) { 96 1.1 riastrad err = PTR_ERR(ce); 97 1.1 riastrad break; 98 1.1 riastrad } 99 1.1 riastrad 100 1.1 riastrad rq = igt_spinner_create_request(&spin, ce, MI_NOOP); 101 1.1 riastrad if (IS_ERR(rq)) { 102 1.1 riastrad err = PTR_ERR(rq); 103 1.1 riastrad goto out_ctx; 104 1.1 riastrad } 105 1.1 riastrad 106 1.1 riastrad i915_request_add(rq); 107 1.1 riastrad if (!igt_wait_for_spinner(&spin, rq)) { 108 1.1 riastrad GEM_TRACE("spinner failed to start\n"); 109 1.1 riastrad GEM_TRACE_DUMP(); 110 1.1 riastrad intel_gt_set_wedged(gt); 111 1.1 riastrad err = -EIO; 112 1.1 riastrad goto out_ctx; 113 1.1 riastrad } 114 1.1 riastrad 115 1.1 riastrad igt_spinner_end(&spin); 116 1.1 riastrad if (igt_flush_test(gt->i915)) { 117 1.1 riastrad err = -EIO; 118 1.1 riastrad goto out_ctx; 119 1.1 riastrad } 120 1.1 riastrad 121 1.1 riastrad out_ctx: 122 1.1 riastrad intel_context_put(ce); 123 1.1 riastrad if (err) 124 1.1 riastrad break; 125 1.1 riastrad } 126 1.1 riastrad 127 1.1 riastrad igt_spinner_fini(&spin); 128 1.1 riastrad return err; 129 1.1 riastrad } 130 1.1 riastrad 131 1.1 riastrad static int live_unlite_restore(struct intel_gt *gt, int prio) 132 1.1 riastrad { 133 1.1 riastrad struct intel_engine_cs *engine; 134 1.1 riastrad enum intel_engine_id id; 135 1.1 riastrad struct igt_spinner spin; 136 1.1 riastrad int err = -ENOMEM; 137 1.1 riastrad 138 1.1 riastrad /* 139 1.1 riastrad * Check that we can correctly context switch between 2 instances 140 1.1 riastrad * on the same engine from the same parent context. 141 1.1 riastrad */ 142 1.1 riastrad 143 1.1 riastrad if (igt_spinner_init(&spin, gt)) 144 1.1 riastrad return err; 145 1.1 riastrad 146 1.1 riastrad err = 0; 147 1.1 riastrad for_each_engine(engine, gt, id) { 148 1.1 riastrad struct intel_context *ce[2] = {}; 149 1.1 riastrad struct i915_request *rq[2]; 150 1.1 riastrad struct igt_live_test t; 151 1.1 riastrad unsigned long saved; 152 1.1 riastrad int n; 153 1.1 riastrad 154 1.1 riastrad if (prio && !intel_engine_has_preemption(engine)) 155 1.1 riastrad continue; 156 1.1 riastrad 157 1.1 riastrad if (!intel_engine_can_store_dword(engine)) 158 1.1 riastrad continue; 159 1.1 riastrad 160 1.1 riastrad if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) { 161 1.1 riastrad err = -EIO; 162 1.1 riastrad break; 163 1.1 riastrad } 164 1.1 riastrad engine_heartbeat_disable(engine, &saved); 165 1.1 riastrad 166 1.1 riastrad for (n = 0; n < ARRAY_SIZE(ce); n++) { 167 1.1 riastrad struct intel_context *tmp; 168 1.1 riastrad 169 1.1 riastrad tmp = intel_context_create(engine); 170 1.1 riastrad if (IS_ERR(tmp)) { 171 1.1 riastrad err = PTR_ERR(tmp); 172 1.1 riastrad goto err_ce; 173 1.1 riastrad } 174 1.1 riastrad 175 1.1 riastrad err = intel_context_pin(tmp); 176 1.1 riastrad if (err) { 177 1.1 riastrad intel_context_put(tmp); 178 1.1 riastrad goto err_ce; 179 1.1 riastrad } 180 1.1 riastrad 181 1.1 riastrad /* 182 1.1 riastrad * Setup the pair of contexts such that if we 183 1.1 riastrad * lite-restore using the RING_TAIL from ce[1] it 184 1.1 riastrad * will execute garbage from ce[0]->ring. 185 1.1 riastrad */ 186 1.1 riastrad memset(tmp->ring->vaddr, 187 1.1 riastrad POISON_INUSE, /* IPEHR: 0x5a5a5a5a [hung!] */ 188 1.1 riastrad tmp->ring->vma->size); 189 1.1 riastrad 190 1.1 riastrad ce[n] = tmp; 191 1.1 riastrad } 192 1.1 riastrad GEM_BUG_ON(!ce[1]->ring->size); 193 1.1 riastrad intel_ring_reset(ce[1]->ring, ce[1]->ring->size / 2); 194 1.1 riastrad __execlists_update_reg_state(ce[1], engine, ce[1]->ring->head); 195 1.1 riastrad 196 1.1 riastrad rq[0] = igt_spinner_create_request(&spin, ce[0], MI_ARB_CHECK); 197 1.1 riastrad if (IS_ERR(rq[0])) { 198 1.1 riastrad err = PTR_ERR(rq[0]); 199 1.1 riastrad goto err_ce; 200 1.1 riastrad } 201 1.1 riastrad 202 1.1 riastrad i915_request_get(rq[0]); 203 1.1 riastrad i915_request_add(rq[0]); 204 1.1 riastrad GEM_BUG_ON(rq[0]->postfix > ce[1]->ring->emit); 205 1.1 riastrad 206 1.1 riastrad if (!igt_wait_for_spinner(&spin, rq[0])) { 207 1.1 riastrad i915_request_put(rq[0]); 208 1.1 riastrad goto err_ce; 209 1.1 riastrad } 210 1.1 riastrad 211 1.1 riastrad rq[1] = i915_request_create(ce[1]); 212 1.1 riastrad if (IS_ERR(rq[1])) { 213 1.1 riastrad err = PTR_ERR(rq[1]); 214 1.1 riastrad i915_request_put(rq[0]); 215 1.1 riastrad goto err_ce; 216 1.1 riastrad } 217 1.1 riastrad 218 1.1 riastrad if (!prio) { 219 1.1 riastrad /* 220 1.1 riastrad * Ensure we do the switch to ce[1] on completion. 221 1.1 riastrad * 222 1.1 riastrad * rq[0] is already submitted, so this should reduce 223 1.1 riastrad * to a no-op (a wait on a request on the same engine 224 1.1 riastrad * uses the submit fence, not the completion fence), 225 1.1 riastrad * but it will install a dependency on rq[1] for rq[0] 226 1.1 riastrad * that will prevent the pair being reordered by 227 1.1 riastrad * timeslicing. 228 1.1 riastrad */ 229 1.1 riastrad i915_request_await_dma_fence(rq[1], &rq[0]->fence); 230 1.1 riastrad } 231 1.1 riastrad 232 1.1 riastrad i915_request_get(rq[1]); 233 1.1 riastrad i915_request_add(rq[1]); 234 1.1 riastrad GEM_BUG_ON(rq[1]->postfix <= rq[0]->postfix); 235 1.1 riastrad i915_request_put(rq[0]); 236 1.1 riastrad 237 1.1 riastrad if (prio) { 238 1.1 riastrad struct i915_sched_attr attr = { 239 1.1 riastrad .priority = prio, 240 1.1 riastrad }; 241 1.1 riastrad 242 1.1 riastrad /* Alternatively preempt the spinner with ce[1] */ 243 1.1 riastrad engine->schedule(rq[1], &attr); 244 1.1 riastrad } 245 1.1 riastrad 246 1.1 riastrad /* And switch back to ce[0] for good measure */ 247 1.1 riastrad rq[0] = i915_request_create(ce[0]); 248 1.1 riastrad if (IS_ERR(rq[0])) { 249 1.1 riastrad err = PTR_ERR(rq[0]); 250 1.1 riastrad i915_request_put(rq[1]); 251 1.1 riastrad goto err_ce; 252 1.1 riastrad } 253 1.1 riastrad 254 1.1 riastrad i915_request_await_dma_fence(rq[0], &rq[1]->fence); 255 1.1 riastrad i915_request_get(rq[0]); 256 1.1 riastrad i915_request_add(rq[0]); 257 1.1 riastrad GEM_BUG_ON(rq[0]->postfix > rq[1]->postfix); 258 1.1 riastrad i915_request_put(rq[1]); 259 1.1 riastrad i915_request_put(rq[0]); 260 1.1 riastrad 261 1.1 riastrad err_ce: 262 1.1 riastrad tasklet_kill(&engine->execlists.tasklet); /* flush submission */ 263 1.1 riastrad igt_spinner_end(&spin); 264 1.1 riastrad for (n = 0; n < ARRAY_SIZE(ce); n++) { 265 1.1 riastrad if (IS_ERR_OR_NULL(ce[n])) 266 1.1 riastrad break; 267 1.1 riastrad 268 1.1 riastrad intel_context_unpin(ce[n]); 269 1.1 riastrad intel_context_put(ce[n]); 270 1.1 riastrad } 271 1.1 riastrad 272 1.1 riastrad engine_heartbeat_enable(engine, saved); 273 1.1 riastrad if (igt_live_test_end(&t)) 274 1.1 riastrad err = -EIO; 275 1.1 riastrad if (err) 276 1.1 riastrad break; 277 1.1 riastrad } 278 1.1 riastrad 279 1.1 riastrad igt_spinner_fini(&spin); 280 1.1 riastrad return err; 281 1.1 riastrad } 282 1.1 riastrad 283 1.1 riastrad static int live_unlite_switch(void *arg) 284 1.1 riastrad { 285 1.1 riastrad return live_unlite_restore(arg, 0); 286 1.1 riastrad } 287 1.1 riastrad 288 1.1 riastrad static int live_unlite_preempt(void *arg) 289 1.1 riastrad { 290 1.1 riastrad return live_unlite_restore(arg, I915_USER_PRIORITY(I915_PRIORITY_MAX)); 291 1.1 riastrad } 292 1.1 riastrad 293 1.1 riastrad static int live_hold_reset(void *arg) 294 1.1 riastrad { 295 1.1 riastrad struct intel_gt *gt = arg; 296 1.1 riastrad struct intel_engine_cs *engine; 297 1.1 riastrad enum intel_engine_id id; 298 1.1 riastrad struct igt_spinner spin; 299 1.1 riastrad int err = 0; 300 1.1 riastrad 301 1.1 riastrad /* 302 1.1 riastrad * In order to support offline error capture for fast preempt reset, 303 1.1 riastrad * we need to decouple the guilty request and ensure that it and its 304 1.1 riastrad * descendents are not executed while the capture is in progress. 305 1.1 riastrad */ 306 1.1 riastrad 307 1.1 riastrad if (!intel_has_reset_engine(gt)) 308 1.1 riastrad return 0; 309 1.1 riastrad 310 1.1 riastrad if (igt_spinner_init(&spin, gt)) 311 1.1 riastrad return -ENOMEM; 312 1.1 riastrad 313 1.1 riastrad for_each_engine(engine, gt, id) { 314 1.1 riastrad struct intel_context *ce; 315 1.1 riastrad unsigned long heartbeat; 316 1.1 riastrad struct i915_request *rq; 317 1.1 riastrad 318 1.1 riastrad ce = intel_context_create(engine); 319 1.1 riastrad if (IS_ERR(ce)) { 320 1.1 riastrad err = PTR_ERR(ce); 321 1.1 riastrad break; 322 1.1 riastrad } 323 1.1 riastrad 324 1.1 riastrad engine_heartbeat_disable(engine, &heartbeat); 325 1.1 riastrad 326 1.1 riastrad rq = igt_spinner_create_request(&spin, ce, MI_ARB_CHECK); 327 1.1 riastrad if (IS_ERR(rq)) { 328 1.1 riastrad err = PTR_ERR(rq); 329 1.1 riastrad goto out; 330 1.1 riastrad } 331 1.1 riastrad i915_request_add(rq); 332 1.1 riastrad 333 1.1 riastrad if (!igt_wait_for_spinner(&spin, rq)) { 334 1.1 riastrad intel_gt_set_wedged(gt); 335 1.1 riastrad err = -ETIME; 336 1.1 riastrad goto out; 337 1.1 riastrad } 338 1.1 riastrad 339 1.1 riastrad /* We have our request executing, now remove it and reset */ 340 1.1 riastrad 341 1.1 riastrad if (test_and_set_bit(I915_RESET_ENGINE + id, 342 1.1 riastrad >->reset.flags)) { 343 1.1 riastrad intel_gt_set_wedged(gt); 344 1.1 riastrad err = -EBUSY; 345 1.1 riastrad goto out; 346 1.1 riastrad } 347 1.1 riastrad tasklet_disable(&engine->execlists.tasklet); 348 1.1 riastrad 349 1.1 riastrad engine->execlists.tasklet.func(engine->execlists.tasklet.data); 350 1.1 riastrad GEM_BUG_ON(execlists_active(&engine->execlists) != rq); 351 1.1 riastrad 352 1.1 riastrad i915_request_get(rq); 353 1.1 riastrad execlists_hold(engine, rq); 354 1.1 riastrad GEM_BUG_ON(!i915_request_on_hold(rq)); 355 1.1 riastrad 356 1.1 riastrad intel_engine_reset(engine, NULL); 357 1.1 riastrad GEM_BUG_ON(rq->fence.error != -EIO); 358 1.1 riastrad 359 1.1 riastrad tasklet_enable(&engine->execlists.tasklet); 360 1.1 riastrad clear_and_wake_up_bit(I915_RESET_ENGINE + id, 361 1.1 riastrad >->reset.flags); 362 1.1 riastrad 363 1.1 riastrad /* Check that we do not resubmit the held request */ 364 1.1 riastrad if (!i915_request_wait(rq, 0, HZ / 5)) { 365 1.1 riastrad pr_err("%s: on hold request completed!\n", 366 1.1 riastrad engine->name); 367 1.1 riastrad i915_request_put(rq); 368 1.1 riastrad err = -EIO; 369 1.1 riastrad goto out; 370 1.1 riastrad } 371 1.1 riastrad GEM_BUG_ON(!i915_request_on_hold(rq)); 372 1.1 riastrad 373 1.1 riastrad /* But is resubmitted on release */ 374 1.1 riastrad execlists_unhold(engine, rq); 375 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 5) < 0) { 376 1.1 riastrad pr_err("%s: held request did not complete!\n", 377 1.1 riastrad engine->name); 378 1.1 riastrad intel_gt_set_wedged(gt); 379 1.1 riastrad err = -ETIME; 380 1.1 riastrad } 381 1.1 riastrad i915_request_put(rq); 382 1.1 riastrad 383 1.1 riastrad out: 384 1.1 riastrad engine_heartbeat_enable(engine, heartbeat); 385 1.1 riastrad intel_context_put(ce); 386 1.1 riastrad if (err) 387 1.1 riastrad break; 388 1.1 riastrad } 389 1.1 riastrad 390 1.1 riastrad igt_spinner_fini(&spin); 391 1.1 riastrad return err; 392 1.1 riastrad } 393 1.1 riastrad 394 1.1 riastrad static int 395 1.1 riastrad emit_semaphore_chain(struct i915_request *rq, struct i915_vma *vma, int idx) 396 1.1 riastrad { 397 1.1 riastrad u32 *cs; 398 1.1 riastrad 399 1.1 riastrad cs = intel_ring_begin(rq, 10); 400 1.1 riastrad if (IS_ERR(cs)) 401 1.1 riastrad return PTR_ERR(cs); 402 1.1 riastrad 403 1.1 riastrad *cs++ = MI_ARB_ON_OFF | MI_ARB_ENABLE; 404 1.1 riastrad 405 1.1 riastrad *cs++ = MI_SEMAPHORE_WAIT | 406 1.1 riastrad MI_SEMAPHORE_GLOBAL_GTT | 407 1.1 riastrad MI_SEMAPHORE_POLL | 408 1.1 riastrad MI_SEMAPHORE_SAD_NEQ_SDD; 409 1.1 riastrad *cs++ = 0; 410 1.1 riastrad *cs++ = i915_ggtt_offset(vma) + 4 * idx; 411 1.1 riastrad *cs++ = 0; 412 1.1 riastrad 413 1.1 riastrad if (idx > 0) { 414 1.1 riastrad *cs++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT; 415 1.1 riastrad *cs++ = i915_ggtt_offset(vma) + 4 * (idx - 1); 416 1.1 riastrad *cs++ = 0; 417 1.1 riastrad *cs++ = 1; 418 1.1 riastrad } else { 419 1.1 riastrad *cs++ = MI_NOOP; 420 1.1 riastrad *cs++ = MI_NOOP; 421 1.1 riastrad *cs++ = MI_NOOP; 422 1.1 riastrad *cs++ = MI_NOOP; 423 1.1 riastrad } 424 1.1 riastrad 425 1.1 riastrad *cs++ = MI_ARB_ON_OFF | MI_ARB_DISABLE; 426 1.1 riastrad 427 1.1 riastrad intel_ring_advance(rq, cs); 428 1.1 riastrad return 0; 429 1.1 riastrad } 430 1.1 riastrad 431 1.1 riastrad static struct i915_request * 432 1.1 riastrad semaphore_queue(struct intel_engine_cs *engine, struct i915_vma *vma, int idx) 433 1.1 riastrad { 434 1.1 riastrad struct intel_context *ce; 435 1.1 riastrad struct i915_request *rq; 436 1.1 riastrad int err; 437 1.1 riastrad 438 1.1 riastrad ce = intel_context_create(engine); 439 1.1 riastrad if (IS_ERR(ce)) 440 1.1 riastrad return ERR_CAST(ce); 441 1.1 riastrad 442 1.1 riastrad rq = intel_context_create_request(ce); 443 1.1 riastrad if (IS_ERR(rq)) 444 1.1 riastrad goto out_ce; 445 1.1 riastrad 446 1.1 riastrad err = 0; 447 1.1 riastrad if (rq->engine->emit_init_breadcrumb) 448 1.1 riastrad err = rq->engine->emit_init_breadcrumb(rq); 449 1.1 riastrad if (err == 0) 450 1.1 riastrad err = emit_semaphore_chain(rq, vma, idx); 451 1.1 riastrad if (err == 0) 452 1.1 riastrad i915_request_get(rq); 453 1.1 riastrad i915_request_add(rq); 454 1.1 riastrad if (err) 455 1.1 riastrad rq = ERR_PTR(err); 456 1.1 riastrad 457 1.1 riastrad out_ce: 458 1.1 riastrad intel_context_put(ce); 459 1.1 riastrad return rq; 460 1.1 riastrad } 461 1.1 riastrad 462 1.1 riastrad static int 463 1.1 riastrad release_queue(struct intel_engine_cs *engine, 464 1.1 riastrad struct i915_vma *vma, 465 1.1 riastrad int idx, int prio) 466 1.1 riastrad { 467 1.1 riastrad struct i915_sched_attr attr = { 468 1.1 riastrad .priority = prio, 469 1.1 riastrad }; 470 1.1 riastrad struct i915_request *rq; 471 1.1 riastrad u32 *cs; 472 1.1 riastrad 473 1.1 riastrad rq = intel_engine_create_kernel_request(engine); 474 1.1 riastrad if (IS_ERR(rq)) 475 1.1 riastrad return PTR_ERR(rq); 476 1.1 riastrad 477 1.1 riastrad cs = intel_ring_begin(rq, 4); 478 1.1 riastrad if (IS_ERR(cs)) { 479 1.1 riastrad i915_request_add(rq); 480 1.1 riastrad return PTR_ERR(cs); 481 1.1 riastrad } 482 1.1 riastrad 483 1.1 riastrad *cs++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT; 484 1.1 riastrad *cs++ = i915_ggtt_offset(vma) + 4 * (idx - 1); 485 1.1 riastrad *cs++ = 0; 486 1.1 riastrad *cs++ = 1; 487 1.1 riastrad 488 1.1 riastrad intel_ring_advance(rq, cs); 489 1.1 riastrad 490 1.1 riastrad i915_request_get(rq); 491 1.1 riastrad i915_request_add(rq); 492 1.1 riastrad 493 1.1 riastrad local_bh_disable(); 494 1.1 riastrad engine->schedule(rq, &attr); 495 1.1 riastrad local_bh_enable(); /* kick tasklet */ 496 1.1 riastrad 497 1.1 riastrad i915_request_put(rq); 498 1.1 riastrad 499 1.1 riastrad return 0; 500 1.1 riastrad } 501 1.1 riastrad 502 1.1 riastrad static int 503 1.1 riastrad slice_semaphore_queue(struct intel_engine_cs *outer, 504 1.1 riastrad struct i915_vma *vma, 505 1.1 riastrad int count) 506 1.1 riastrad { 507 1.1 riastrad struct intel_engine_cs *engine; 508 1.1 riastrad struct i915_request *head; 509 1.1 riastrad enum intel_engine_id id; 510 1.1 riastrad int err, i, n = 0; 511 1.1 riastrad 512 1.1 riastrad head = semaphore_queue(outer, vma, n++); 513 1.1 riastrad if (IS_ERR(head)) 514 1.1 riastrad return PTR_ERR(head); 515 1.1 riastrad 516 1.1 riastrad for_each_engine(engine, outer->gt, id) { 517 1.1 riastrad for (i = 0; i < count; i++) { 518 1.1 riastrad struct i915_request *rq; 519 1.1 riastrad 520 1.1 riastrad rq = semaphore_queue(engine, vma, n++); 521 1.1 riastrad if (IS_ERR(rq)) { 522 1.1 riastrad err = PTR_ERR(rq); 523 1.1 riastrad goto out; 524 1.1 riastrad } 525 1.1 riastrad 526 1.1 riastrad i915_request_put(rq); 527 1.1 riastrad } 528 1.1 riastrad } 529 1.1 riastrad 530 1.1 riastrad err = release_queue(outer, vma, n, INT_MAX); 531 1.1 riastrad if (err) 532 1.1 riastrad goto out; 533 1.1 riastrad 534 1.1 riastrad if (i915_request_wait(head, 0, 535 1.1 riastrad 2 * RUNTIME_INFO(outer->i915)->num_engines * (count + 2) * (count + 3)) < 0) { 536 1.1 riastrad pr_err("Failed to slice along semaphore chain of length (%d, %d)!\n", 537 1.1 riastrad count, n); 538 1.1 riastrad GEM_TRACE_DUMP(); 539 1.1 riastrad intel_gt_set_wedged(outer->gt); 540 1.1 riastrad err = -EIO; 541 1.1 riastrad } 542 1.1 riastrad 543 1.1 riastrad out: 544 1.1 riastrad i915_request_put(head); 545 1.1 riastrad return err; 546 1.1 riastrad } 547 1.1 riastrad 548 1.1 riastrad static int live_timeslice_preempt(void *arg) 549 1.1 riastrad { 550 1.1 riastrad struct intel_gt *gt = arg; 551 1.1 riastrad struct drm_i915_gem_object *obj; 552 1.1 riastrad struct i915_vma *vma; 553 1.1 riastrad void *vaddr; 554 1.1 riastrad int err = 0; 555 1.1 riastrad int count; 556 1.1 riastrad 557 1.1 riastrad /* 558 1.1 riastrad * If a request takes too long, we would like to give other users 559 1.1 riastrad * a fair go on the GPU. In particular, users may create batches 560 1.1 riastrad * that wait upon external input, where that input may even be 561 1.1 riastrad * supplied by another GPU job. To avoid blocking forever, we 562 1.1 riastrad * need to preempt the current task and replace it with another 563 1.1 riastrad * ready task. 564 1.1 riastrad */ 565 1.1 riastrad if (!IS_ACTIVE(CONFIG_DRM_I915_TIMESLICE_DURATION)) 566 1.1 riastrad return 0; 567 1.1 riastrad 568 1.1 riastrad obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE); 569 1.1 riastrad if (IS_ERR(obj)) 570 1.1 riastrad return PTR_ERR(obj); 571 1.1 riastrad 572 1.1 riastrad vma = i915_vma_instance(obj, >->ggtt->vm, NULL); 573 1.1 riastrad if (IS_ERR(vma)) { 574 1.1 riastrad err = PTR_ERR(vma); 575 1.1 riastrad goto err_obj; 576 1.1 riastrad } 577 1.1 riastrad 578 1.1 riastrad vaddr = i915_gem_object_pin_map(obj, I915_MAP_WC); 579 1.1 riastrad if (IS_ERR(vaddr)) { 580 1.1 riastrad err = PTR_ERR(vaddr); 581 1.1 riastrad goto err_obj; 582 1.1 riastrad } 583 1.1 riastrad 584 1.1 riastrad err = i915_vma_pin(vma, 0, 0, PIN_GLOBAL); 585 1.1 riastrad if (err) 586 1.1 riastrad goto err_map; 587 1.1 riastrad 588 1.1 riastrad for_each_prime_number_from(count, 1, 16) { 589 1.1 riastrad struct intel_engine_cs *engine; 590 1.1 riastrad enum intel_engine_id id; 591 1.1 riastrad 592 1.1 riastrad for_each_engine(engine, gt, id) { 593 1.1 riastrad unsigned long saved; 594 1.1 riastrad 595 1.1 riastrad if (!intel_engine_has_preemption(engine)) 596 1.1 riastrad continue; 597 1.1 riastrad 598 1.1 riastrad memset(vaddr, 0, PAGE_SIZE); 599 1.1 riastrad 600 1.1 riastrad engine_heartbeat_disable(engine, &saved); 601 1.1 riastrad err = slice_semaphore_queue(engine, vma, count); 602 1.1 riastrad engine_heartbeat_enable(engine, saved); 603 1.1 riastrad if (err) 604 1.1 riastrad goto err_pin; 605 1.1 riastrad 606 1.1 riastrad if (igt_flush_test(gt->i915)) { 607 1.1 riastrad err = -EIO; 608 1.1 riastrad goto err_pin; 609 1.1 riastrad } 610 1.1 riastrad } 611 1.1 riastrad } 612 1.1 riastrad 613 1.1 riastrad err_pin: 614 1.1 riastrad i915_vma_unpin(vma); 615 1.1 riastrad err_map: 616 1.1 riastrad i915_gem_object_unpin_map(obj); 617 1.1 riastrad err_obj: 618 1.1 riastrad i915_gem_object_put(obj); 619 1.1 riastrad return err; 620 1.1 riastrad } 621 1.1 riastrad 622 1.1 riastrad static struct i915_request *nop_request(struct intel_engine_cs *engine) 623 1.1 riastrad { 624 1.1 riastrad struct i915_request *rq; 625 1.1 riastrad 626 1.1 riastrad rq = intel_engine_create_kernel_request(engine); 627 1.1 riastrad if (IS_ERR(rq)) 628 1.1 riastrad return rq; 629 1.1 riastrad 630 1.1 riastrad i915_request_get(rq); 631 1.1 riastrad i915_request_add(rq); 632 1.1 riastrad 633 1.1 riastrad return rq; 634 1.1 riastrad } 635 1.1 riastrad 636 1.1 riastrad static int wait_for_submit(struct intel_engine_cs *engine, 637 1.1 riastrad struct i915_request *rq, 638 1.1 riastrad unsigned long timeout) 639 1.1 riastrad { 640 1.1 riastrad timeout += jiffies; 641 1.1 riastrad do { 642 1.1 riastrad cond_resched(); 643 1.1 riastrad intel_engine_flush_submission(engine); 644 1.1 riastrad if (i915_request_is_active(rq)) 645 1.1 riastrad return 0; 646 1.1 riastrad } while (time_before(jiffies, timeout)); 647 1.1 riastrad 648 1.1 riastrad return -ETIME; 649 1.1 riastrad } 650 1.1 riastrad 651 1.1 riastrad static long timeslice_threshold(const struct intel_engine_cs *engine) 652 1.1 riastrad { 653 1.1 riastrad return 2 * msecs_to_jiffies_timeout(timeslice(engine)) + 1; 654 1.1 riastrad } 655 1.1 riastrad 656 1.1 riastrad static int live_timeslice_queue(void *arg) 657 1.1 riastrad { 658 1.1 riastrad struct intel_gt *gt = arg; 659 1.1 riastrad struct drm_i915_gem_object *obj; 660 1.1 riastrad struct intel_engine_cs *engine; 661 1.1 riastrad enum intel_engine_id id; 662 1.1 riastrad struct i915_vma *vma; 663 1.1 riastrad void *vaddr; 664 1.1 riastrad int err = 0; 665 1.1 riastrad 666 1.1 riastrad /* 667 1.1 riastrad * Make sure that even if ELSP[0] and ELSP[1] are filled with 668 1.1 riastrad * timeslicing between them disabled, we *do* enable timeslicing 669 1.1 riastrad * if the queue demands it. (Normally, we do not submit if 670 1.1 riastrad * ELSP[1] is already occupied, so must rely on timeslicing to 671 1.1 riastrad * eject ELSP[0] in favour of the queue.) 672 1.1 riastrad */ 673 1.1 riastrad if (!IS_ACTIVE(CONFIG_DRM_I915_TIMESLICE_DURATION)) 674 1.1 riastrad return 0; 675 1.1 riastrad 676 1.1 riastrad obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE); 677 1.1 riastrad if (IS_ERR(obj)) 678 1.1 riastrad return PTR_ERR(obj); 679 1.1 riastrad 680 1.1 riastrad vma = i915_vma_instance(obj, >->ggtt->vm, NULL); 681 1.1 riastrad if (IS_ERR(vma)) { 682 1.1 riastrad err = PTR_ERR(vma); 683 1.1 riastrad goto err_obj; 684 1.1 riastrad } 685 1.1 riastrad 686 1.1 riastrad vaddr = i915_gem_object_pin_map(obj, I915_MAP_WC); 687 1.1 riastrad if (IS_ERR(vaddr)) { 688 1.1 riastrad err = PTR_ERR(vaddr); 689 1.1 riastrad goto err_obj; 690 1.1 riastrad } 691 1.1 riastrad 692 1.1 riastrad err = i915_vma_pin(vma, 0, 0, PIN_GLOBAL); 693 1.1 riastrad if (err) 694 1.1 riastrad goto err_map; 695 1.1 riastrad 696 1.1 riastrad for_each_engine(engine, gt, id) { 697 1.1 riastrad struct i915_sched_attr attr = { 698 1.1 riastrad .priority = I915_USER_PRIORITY(I915_PRIORITY_MAX), 699 1.1 riastrad }; 700 1.1 riastrad struct i915_request *rq, *nop; 701 1.1 riastrad unsigned long saved; 702 1.1 riastrad 703 1.1 riastrad if (!intel_engine_has_preemption(engine)) 704 1.1 riastrad continue; 705 1.1 riastrad 706 1.1 riastrad engine_heartbeat_disable(engine, &saved); 707 1.1 riastrad memset(vaddr, 0, PAGE_SIZE); 708 1.1 riastrad 709 1.1 riastrad /* ELSP[0]: semaphore wait */ 710 1.1 riastrad rq = semaphore_queue(engine, vma, 0); 711 1.1 riastrad if (IS_ERR(rq)) { 712 1.1 riastrad err = PTR_ERR(rq); 713 1.1 riastrad goto err_heartbeat; 714 1.1 riastrad } 715 1.1 riastrad engine->schedule(rq, &attr); 716 1.1 riastrad err = wait_for_submit(engine, rq, HZ / 2); 717 1.1 riastrad if (err) { 718 1.1 riastrad pr_err("%s: Timed out trying to submit semaphores\n", 719 1.1 riastrad engine->name); 720 1.1 riastrad goto err_rq; 721 1.1 riastrad } 722 1.1 riastrad 723 1.1 riastrad /* ELSP[1]: nop request */ 724 1.1 riastrad nop = nop_request(engine); 725 1.1 riastrad if (IS_ERR(nop)) { 726 1.1 riastrad err = PTR_ERR(nop); 727 1.1 riastrad goto err_rq; 728 1.1 riastrad } 729 1.1 riastrad err = wait_for_submit(engine, nop, HZ / 2); 730 1.1 riastrad i915_request_put(nop); 731 1.1 riastrad if (err) { 732 1.1 riastrad pr_err("%s: Timed out trying to submit nop\n", 733 1.1 riastrad engine->name); 734 1.1 riastrad goto err_rq; 735 1.1 riastrad } 736 1.1 riastrad 737 1.1 riastrad GEM_BUG_ON(i915_request_completed(rq)); 738 1.1 riastrad GEM_BUG_ON(execlists_active(&engine->execlists) != rq); 739 1.1 riastrad 740 1.1 riastrad /* Queue: semaphore signal, matching priority as semaphore */ 741 1.1 riastrad err = release_queue(engine, vma, 1, effective_prio(rq)); 742 1.1 riastrad if (err) 743 1.1 riastrad goto err_rq; 744 1.1 riastrad 745 1.1 riastrad intel_engine_flush_submission(engine); 746 1.1 riastrad if (!READ_ONCE(engine->execlists.timer.expires) && 747 1.1 riastrad !i915_request_completed(rq)) { 748 1.1 riastrad struct drm_printer p = 749 1.1 riastrad drm_info_printer(gt->i915->drm.dev); 750 1.1 riastrad 751 1.1 riastrad GEM_TRACE_ERR("%s: Failed to enable timeslicing!\n", 752 1.1 riastrad engine->name); 753 1.1 riastrad intel_engine_dump(engine, &p, 754 1.1 riastrad "%s\n", engine->name); 755 1.1 riastrad GEM_TRACE_DUMP(); 756 1.1 riastrad 757 1.1 riastrad memset(vaddr, 0xff, PAGE_SIZE); 758 1.1 riastrad err = -EINVAL; 759 1.1 riastrad } 760 1.1 riastrad 761 1.1 riastrad /* Timeslice every jiffy, so within 2 we should signal */ 762 1.1 riastrad if (i915_request_wait(rq, 0, timeslice_threshold(engine)) < 0) { 763 1.1 riastrad struct drm_printer p = 764 1.1 riastrad drm_info_printer(gt->i915->drm.dev); 765 1.1 riastrad 766 1.1 riastrad pr_err("%s: Failed to timeslice into queue\n", 767 1.1 riastrad engine->name); 768 1.1 riastrad intel_engine_dump(engine, &p, 769 1.1 riastrad "%s\n", engine->name); 770 1.1 riastrad 771 1.1 riastrad memset(vaddr, 0xff, PAGE_SIZE); 772 1.1 riastrad err = -EIO; 773 1.1 riastrad } 774 1.1 riastrad err_rq: 775 1.1 riastrad i915_request_put(rq); 776 1.1 riastrad err_heartbeat: 777 1.1 riastrad engine_heartbeat_enable(engine, saved); 778 1.1 riastrad if (err) 779 1.1 riastrad break; 780 1.1 riastrad } 781 1.1 riastrad 782 1.1 riastrad i915_vma_unpin(vma); 783 1.1 riastrad err_map: 784 1.1 riastrad i915_gem_object_unpin_map(obj); 785 1.1 riastrad err_obj: 786 1.1 riastrad i915_gem_object_put(obj); 787 1.1 riastrad return err; 788 1.1 riastrad } 789 1.1 riastrad 790 1.1 riastrad static int live_busywait_preempt(void *arg) 791 1.1 riastrad { 792 1.1 riastrad struct intel_gt *gt = arg; 793 1.1 riastrad struct i915_gem_context *ctx_hi, *ctx_lo; 794 1.1 riastrad struct intel_engine_cs *engine; 795 1.1 riastrad struct drm_i915_gem_object *obj; 796 1.1 riastrad struct i915_vma *vma; 797 1.1 riastrad enum intel_engine_id id; 798 1.1 riastrad int err = -ENOMEM; 799 1.1 riastrad u32 *map; 800 1.1 riastrad 801 1.1 riastrad /* 802 1.1 riastrad * Verify that even without HAS_LOGICAL_RING_PREEMPTION, we can 803 1.1 riastrad * preempt the busywaits used to synchronise between rings. 804 1.1 riastrad */ 805 1.1 riastrad 806 1.1 riastrad ctx_hi = kernel_context(gt->i915); 807 1.1 riastrad if (!ctx_hi) 808 1.1 riastrad return -ENOMEM; 809 1.1 riastrad ctx_hi->sched.priority = 810 1.1 riastrad I915_USER_PRIORITY(I915_CONTEXT_MAX_USER_PRIORITY); 811 1.1 riastrad 812 1.1 riastrad ctx_lo = kernel_context(gt->i915); 813 1.1 riastrad if (!ctx_lo) 814 1.1 riastrad goto err_ctx_hi; 815 1.1 riastrad ctx_lo->sched.priority = 816 1.1 riastrad I915_USER_PRIORITY(I915_CONTEXT_MIN_USER_PRIORITY); 817 1.1 riastrad 818 1.1 riastrad obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE); 819 1.1 riastrad if (IS_ERR(obj)) { 820 1.1 riastrad err = PTR_ERR(obj); 821 1.1 riastrad goto err_ctx_lo; 822 1.1 riastrad } 823 1.1 riastrad 824 1.1 riastrad map = i915_gem_object_pin_map(obj, I915_MAP_WC); 825 1.1 riastrad if (IS_ERR(map)) { 826 1.1 riastrad err = PTR_ERR(map); 827 1.1 riastrad goto err_obj; 828 1.1 riastrad } 829 1.1 riastrad 830 1.1 riastrad vma = i915_vma_instance(obj, >->ggtt->vm, NULL); 831 1.1 riastrad if (IS_ERR(vma)) { 832 1.1 riastrad err = PTR_ERR(vma); 833 1.1 riastrad goto err_map; 834 1.1 riastrad } 835 1.1 riastrad 836 1.1 riastrad err = i915_vma_pin(vma, 0, 0, PIN_GLOBAL); 837 1.1 riastrad if (err) 838 1.1 riastrad goto err_map; 839 1.1 riastrad 840 1.1 riastrad for_each_engine(engine, gt, id) { 841 1.1 riastrad struct i915_request *lo, *hi; 842 1.1 riastrad struct igt_live_test t; 843 1.1 riastrad u32 *cs; 844 1.1 riastrad 845 1.1 riastrad if (!intel_engine_has_preemption(engine)) 846 1.1 riastrad continue; 847 1.1 riastrad 848 1.1 riastrad if (!intel_engine_can_store_dword(engine)) 849 1.1 riastrad continue; 850 1.1 riastrad 851 1.1 riastrad if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) { 852 1.1 riastrad err = -EIO; 853 1.1 riastrad goto err_vma; 854 1.1 riastrad } 855 1.1 riastrad 856 1.1 riastrad /* 857 1.1 riastrad * We create two requests. The low priority request 858 1.1 riastrad * busywaits on a semaphore (inside the ringbuffer where 859 1.1 riastrad * is should be preemptible) and the high priority requests 860 1.1 riastrad * uses a MI_STORE_DWORD_IMM to update the semaphore value 861 1.1 riastrad * allowing the first request to complete. If preemption 862 1.1 riastrad * fails, we hang instead. 863 1.1 riastrad */ 864 1.1 riastrad 865 1.1 riastrad lo = igt_request_alloc(ctx_lo, engine); 866 1.1 riastrad if (IS_ERR(lo)) { 867 1.1 riastrad err = PTR_ERR(lo); 868 1.1 riastrad goto err_vma; 869 1.1 riastrad } 870 1.1 riastrad 871 1.1 riastrad cs = intel_ring_begin(lo, 8); 872 1.1 riastrad if (IS_ERR(cs)) { 873 1.1 riastrad err = PTR_ERR(cs); 874 1.1 riastrad i915_request_add(lo); 875 1.1 riastrad goto err_vma; 876 1.1 riastrad } 877 1.1 riastrad 878 1.1 riastrad *cs++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT; 879 1.1 riastrad *cs++ = i915_ggtt_offset(vma); 880 1.1 riastrad *cs++ = 0; 881 1.1 riastrad *cs++ = 1; 882 1.1 riastrad 883 1.1 riastrad /* XXX Do we need a flush + invalidate here? */ 884 1.1 riastrad 885 1.1 riastrad *cs++ = MI_SEMAPHORE_WAIT | 886 1.1 riastrad MI_SEMAPHORE_GLOBAL_GTT | 887 1.1 riastrad MI_SEMAPHORE_POLL | 888 1.1 riastrad MI_SEMAPHORE_SAD_EQ_SDD; 889 1.1 riastrad *cs++ = 0; 890 1.1 riastrad *cs++ = i915_ggtt_offset(vma); 891 1.1 riastrad *cs++ = 0; 892 1.1 riastrad 893 1.1 riastrad intel_ring_advance(lo, cs); 894 1.1 riastrad 895 1.1 riastrad i915_request_get(lo); 896 1.1 riastrad i915_request_add(lo); 897 1.1 riastrad 898 1.1 riastrad if (wait_for(READ_ONCE(*map), 10)) { 899 1.1 riastrad i915_request_put(lo); 900 1.1 riastrad err = -ETIMEDOUT; 901 1.1 riastrad goto err_vma; 902 1.1 riastrad } 903 1.1 riastrad 904 1.1 riastrad /* Low priority request should be busywaiting now */ 905 1.1 riastrad if (i915_request_wait(lo, 0, 1) != -ETIME) { 906 1.1 riastrad i915_request_put(lo); 907 1.1 riastrad pr_err("%s: Busywaiting request did not!\n", 908 1.1 riastrad engine->name); 909 1.1 riastrad err = -EIO; 910 1.1 riastrad goto err_vma; 911 1.1 riastrad } 912 1.1 riastrad 913 1.1 riastrad hi = igt_request_alloc(ctx_hi, engine); 914 1.1 riastrad if (IS_ERR(hi)) { 915 1.1 riastrad err = PTR_ERR(hi); 916 1.1 riastrad i915_request_put(lo); 917 1.1 riastrad goto err_vma; 918 1.1 riastrad } 919 1.1 riastrad 920 1.1 riastrad cs = intel_ring_begin(hi, 4); 921 1.1 riastrad if (IS_ERR(cs)) { 922 1.1 riastrad err = PTR_ERR(cs); 923 1.1 riastrad i915_request_add(hi); 924 1.1 riastrad i915_request_put(lo); 925 1.1 riastrad goto err_vma; 926 1.1 riastrad } 927 1.1 riastrad 928 1.1 riastrad *cs++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT; 929 1.1 riastrad *cs++ = i915_ggtt_offset(vma); 930 1.1 riastrad *cs++ = 0; 931 1.1 riastrad *cs++ = 0; 932 1.1 riastrad 933 1.1 riastrad intel_ring_advance(hi, cs); 934 1.1 riastrad i915_request_add(hi); 935 1.1 riastrad 936 1.1 riastrad if (i915_request_wait(lo, 0, HZ / 5) < 0) { 937 1.1 riastrad struct drm_printer p = drm_info_printer(gt->i915->drm.dev); 938 1.1 riastrad 939 1.1 riastrad pr_err("%s: Failed to preempt semaphore busywait!\n", 940 1.1 riastrad engine->name); 941 1.1 riastrad 942 1.1 riastrad intel_engine_dump(engine, &p, "%s\n", engine->name); 943 1.1 riastrad GEM_TRACE_DUMP(); 944 1.1 riastrad 945 1.1 riastrad i915_request_put(lo); 946 1.1 riastrad intel_gt_set_wedged(gt); 947 1.1 riastrad err = -EIO; 948 1.1 riastrad goto err_vma; 949 1.1 riastrad } 950 1.1 riastrad GEM_BUG_ON(READ_ONCE(*map)); 951 1.1 riastrad i915_request_put(lo); 952 1.1 riastrad 953 1.1 riastrad if (igt_live_test_end(&t)) { 954 1.1 riastrad err = -EIO; 955 1.1 riastrad goto err_vma; 956 1.1 riastrad } 957 1.1 riastrad } 958 1.1 riastrad 959 1.1 riastrad err = 0; 960 1.1 riastrad err_vma: 961 1.1 riastrad i915_vma_unpin(vma); 962 1.1 riastrad err_map: 963 1.1 riastrad i915_gem_object_unpin_map(obj); 964 1.1 riastrad err_obj: 965 1.1 riastrad i915_gem_object_put(obj); 966 1.1 riastrad err_ctx_lo: 967 1.1 riastrad kernel_context_close(ctx_lo); 968 1.1 riastrad err_ctx_hi: 969 1.1 riastrad kernel_context_close(ctx_hi); 970 1.1 riastrad return err; 971 1.1 riastrad } 972 1.1 riastrad 973 1.1 riastrad static struct i915_request * 974 1.1 riastrad spinner_create_request(struct igt_spinner *spin, 975 1.1 riastrad struct i915_gem_context *ctx, 976 1.1 riastrad struct intel_engine_cs *engine, 977 1.1 riastrad u32 arb) 978 1.1 riastrad { 979 1.1 riastrad struct intel_context *ce; 980 1.1 riastrad struct i915_request *rq; 981 1.1 riastrad 982 1.1 riastrad ce = i915_gem_context_get_engine(ctx, engine->legacy_idx); 983 1.1 riastrad if (IS_ERR(ce)) 984 1.1 riastrad return ERR_CAST(ce); 985 1.1 riastrad 986 1.1 riastrad rq = igt_spinner_create_request(spin, ce, arb); 987 1.1 riastrad intel_context_put(ce); 988 1.1 riastrad return rq; 989 1.1 riastrad } 990 1.1 riastrad 991 1.1 riastrad static int live_preempt(void *arg) 992 1.1 riastrad { 993 1.1 riastrad struct intel_gt *gt = arg; 994 1.1 riastrad struct i915_gem_context *ctx_hi, *ctx_lo; 995 1.1 riastrad struct igt_spinner spin_hi, spin_lo; 996 1.1 riastrad struct intel_engine_cs *engine; 997 1.1 riastrad enum intel_engine_id id; 998 1.1 riastrad int err = -ENOMEM; 999 1.1 riastrad 1000 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 1001 1.1 riastrad return 0; 1002 1.1 riastrad 1003 1.1 riastrad if (!(gt->i915->caps.scheduler & I915_SCHEDULER_CAP_PREEMPTION)) 1004 1.1 riastrad pr_err("Logical preemption supported, but not exposed\n"); 1005 1.1 riastrad 1006 1.1 riastrad if (igt_spinner_init(&spin_hi, gt)) 1007 1.1 riastrad return -ENOMEM; 1008 1.1 riastrad 1009 1.1 riastrad if (igt_spinner_init(&spin_lo, gt)) 1010 1.1 riastrad goto err_spin_hi; 1011 1.1 riastrad 1012 1.1 riastrad ctx_hi = kernel_context(gt->i915); 1013 1.1 riastrad if (!ctx_hi) 1014 1.1 riastrad goto err_spin_lo; 1015 1.1 riastrad ctx_hi->sched.priority = 1016 1.1 riastrad I915_USER_PRIORITY(I915_CONTEXT_MAX_USER_PRIORITY); 1017 1.1 riastrad 1018 1.1 riastrad ctx_lo = kernel_context(gt->i915); 1019 1.1 riastrad if (!ctx_lo) 1020 1.1 riastrad goto err_ctx_hi; 1021 1.1 riastrad ctx_lo->sched.priority = 1022 1.1 riastrad I915_USER_PRIORITY(I915_CONTEXT_MIN_USER_PRIORITY); 1023 1.1 riastrad 1024 1.1 riastrad for_each_engine(engine, gt, id) { 1025 1.1 riastrad struct igt_live_test t; 1026 1.1 riastrad struct i915_request *rq; 1027 1.1 riastrad 1028 1.1 riastrad if (!intel_engine_has_preemption(engine)) 1029 1.1 riastrad continue; 1030 1.1 riastrad 1031 1.1 riastrad if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) { 1032 1.1 riastrad err = -EIO; 1033 1.1 riastrad goto err_ctx_lo; 1034 1.1 riastrad } 1035 1.1 riastrad 1036 1.1 riastrad rq = spinner_create_request(&spin_lo, ctx_lo, engine, 1037 1.1 riastrad MI_ARB_CHECK); 1038 1.1 riastrad if (IS_ERR(rq)) { 1039 1.1 riastrad err = PTR_ERR(rq); 1040 1.1 riastrad goto err_ctx_lo; 1041 1.1 riastrad } 1042 1.1 riastrad 1043 1.1 riastrad i915_request_add(rq); 1044 1.1 riastrad if (!igt_wait_for_spinner(&spin_lo, rq)) { 1045 1.1 riastrad GEM_TRACE("lo spinner failed to start\n"); 1046 1.1 riastrad GEM_TRACE_DUMP(); 1047 1.1 riastrad intel_gt_set_wedged(gt); 1048 1.1 riastrad err = -EIO; 1049 1.1 riastrad goto err_ctx_lo; 1050 1.1 riastrad } 1051 1.1 riastrad 1052 1.1 riastrad rq = spinner_create_request(&spin_hi, ctx_hi, engine, 1053 1.1 riastrad MI_ARB_CHECK); 1054 1.1 riastrad if (IS_ERR(rq)) { 1055 1.1 riastrad igt_spinner_end(&spin_lo); 1056 1.1 riastrad err = PTR_ERR(rq); 1057 1.1 riastrad goto err_ctx_lo; 1058 1.1 riastrad } 1059 1.1 riastrad 1060 1.1 riastrad i915_request_add(rq); 1061 1.1 riastrad if (!igt_wait_for_spinner(&spin_hi, rq)) { 1062 1.1 riastrad GEM_TRACE("hi spinner failed to start\n"); 1063 1.1 riastrad GEM_TRACE_DUMP(); 1064 1.1 riastrad intel_gt_set_wedged(gt); 1065 1.1 riastrad err = -EIO; 1066 1.1 riastrad goto err_ctx_lo; 1067 1.1 riastrad } 1068 1.1 riastrad 1069 1.1 riastrad igt_spinner_end(&spin_hi); 1070 1.1 riastrad igt_spinner_end(&spin_lo); 1071 1.1 riastrad 1072 1.1 riastrad if (igt_live_test_end(&t)) { 1073 1.1 riastrad err = -EIO; 1074 1.1 riastrad goto err_ctx_lo; 1075 1.1 riastrad } 1076 1.1 riastrad } 1077 1.1 riastrad 1078 1.1 riastrad err = 0; 1079 1.1 riastrad err_ctx_lo: 1080 1.1 riastrad kernel_context_close(ctx_lo); 1081 1.1 riastrad err_ctx_hi: 1082 1.1 riastrad kernel_context_close(ctx_hi); 1083 1.1 riastrad err_spin_lo: 1084 1.1 riastrad igt_spinner_fini(&spin_lo); 1085 1.1 riastrad err_spin_hi: 1086 1.1 riastrad igt_spinner_fini(&spin_hi); 1087 1.1 riastrad return err; 1088 1.1 riastrad } 1089 1.1 riastrad 1090 1.1 riastrad static int live_late_preempt(void *arg) 1091 1.1 riastrad { 1092 1.1 riastrad struct intel_gt *gt = arg; 1093 1.1 riastrad struct i915_gem_context *ctx_hi, *ctx_lo; 1094 1.1 riastrad struct igt_spinner spin_hi, spin_lo; 1095 1.1 riastrad struct intel_engine_cs *engine; 1096 1.1 riastrad struct i915_sched_attr attr = {}; 1097 1.1 riastrad enum intel_engine_id id; 1098 1.1 riastrad int err = -ENOMEM; 1099 1.1 riastrad 1100 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 1101 1.1 riastrad return 0; 1102 1.1 riastrad 1103 1.1 riastrad if (igt_spinner_init(&spin_hi, gt)) 1104 1.1 riastrad return -ENOMEM; 1105 1.1 riastrad 1106 1.1 riastrad if (igt_spinner_init(&spin_lo, gt)) 1107 1.1 riastrad goto err_spin_hi; 1108 1.1 riastrad 1109 1.1 riastrad ctx_hi = kernel_context(gt->i915); 1110 1.1 riastrad if (!ctx_hi) 1111 1.1 riastrad goto err_spin_lo; 1112 1.1 riastrad 1113 1.1 riastrad ctx_lo = kernel_context(gt->i915); 1114 1.1 riastrad if (!ctx_lo) 1115 1.1 riastrad goto err_ctx_hi; 1116 1.1 riastrad 1117 1.1 riastrad /* Make sure ctx_lo stays before ctx_hi until we trigger preemption. */ 1118 1.1 riastrad ctx_lo->sched.priority = I915_USER_PRIORITY(1); 1119 1.1 riastrad 1120 1.1 riastrad for_each_engine(engine, gt, id) { 1121 1.1 riastrad struct igt_live_test t; 1122 1.1 riastrad struct i915_request *rq; 1123 1.1 riastrad 1124 1.1 riastrad if (!intel_engine_has_preemption(engine)) 1125 1.1 riastrad continue; 1126 1.1 riastrad 1127 1.1 riastrad if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) { 1128 1.1 riastrad err = -EIO; 1129 1.1 riastrad goto err_ctx_lo; 1130 1.1 riastrad } 1131 1.1 riastrad 1132 1.1 riastrad rq = spinner_create_request(&spin_lo, ctx_lo, engine, 1133 1.1 riastrad MI_ARB_CHECK); 1134 1.1 riastrad if (IS_ERR(rq)) { 1135 1.1 riastrad err = PTR_ERR(rq); 1136 1.1 riastrad goto err_ctx_lo; 1137 1.1 riastrad } 1138 1.1 riastrad 1139 1.1 riastrad i915_request_add(rq); 1140 1.1 riastrad if (!igt_wait_for_spinner(&spin_lo, rq)) { 1141 1.1 riastrad pr_err("First context failed to start\n"); 1142 1.1 riastrad goto err_wedged; 1143 1.1 riastrad } 1144 1.1 riastrad 1145 1.1 riastrad rq = spinner_create_request(&spin_hi, ctx_hi, engine, 1146 1.1 riastrad MI_NOOP); 1147 1.1 riastrad if (IS_ERR(rq)) { 1148 1.1 riastrad igt_spinner_end(&spin_lo); 1149 1.1 riastrad err = PTR_ERR(rq); 1150 1.1 riastrad goto err_ctx_lo; 1151 1.1 riastrad } 1152 1.1 riastrad 1153 1.1 riastrad i915_request_add(rq); 1154 1.1 riastrad if (igt_wait_for_spinner(&spin_hi, rq)) { 1155 1.1 riastrad pr_err("Second context overtook first?\n"); 1156 1.1 riastrad goto err_wedged; 1157 1.1 riastrad } 1158 1.1 riastrad 1159 1.1 riastrad attr.priority = I915_USER_PRIORITY(I915_PRIORITY_MAX); 1160 1.1 riastrad engine->schedule(rq, &attr); 1161 1.1 riastrad 1162 1.1 riastrad if (!igt_wait_for_spinner(&spin_hi, rq)) { 1163 1.1 riastrad pr_err("High priority context failed to preempt the low priority context\n"); 1164 1.1 riastrad GEM_TRACE_DUMP(); 1165 1.1 riastrad goto err_wedged; 1166 1.1 riastrad } 1167 1.1 riastrad 1168 1.1 riastrad igt_spinner_end(&spin_hi); 1169 1.1 riastrad igt_spinner_end(&spin_lo); 1170 1.1 riastrad 1171 1.1 riastrad if (igt_live_test_end(&t)) { 1172 1.1 riastrad err = -EIO; 1173 1.1 riastrad goto err_ctx_lo; 1174 1.1 riastrad } 1175 1.1 riastrad } 1176 1.1 riastrad 1177 1.1 riastrad err = 0; 1178 1.1 riastrad err_ctx_lo: 1179 1.1 riastrad kernel_context_close(ctx_lo); 1180 1.1 riastrad err_ctx_hi: 1181 1.1 riastrad kernel_context_close(ctx_hi); 1182 1.1 riastrad err_spin_lo: 1183 1.1 riastrad igt_spinner_fini(&spin_lo); 1184 1.1 riastrad err_spin_hi: 1185 1.1 riastrad igt_spinner_fini(&spin_hi); 1186 1.1 riastrad return err; 1187 1.1 riastrad 1188 1.1 riastrad err_wedged: 1189 1.1 riastrad igt_spinner_end(&spin_hi); 1190 1.1 riastrad igt_spinner_end(&spin_lo); 1191 1.1 riastrad intel_gt_set_wedged(gt); 1192 1.1 riastrad err = -EIO; 1193 1.1 riastrad goto err_ctx_lo; 1194 1.1 riastrad } 1195 1.1 riastrad 1196 1.1 riastrad struct preempt_client { 1197 1.1 riastrad struct igt_spinner spin; 1198 1.1 riastrad struct i915_gem_context *ctx; 1199 1.1 riastrad }; 1200 1.1 riastrad 1201 1.1 riastrad static int preempt_client_init(struct intel_gt *gt, struct preempt_client *c) 1202 1.1 riastrad { 1203 1.1 riastrad c->ctx = kernel_context(gt->i915); 1204 1.1 riastrad if (!c->ctx) 1205 1.1 riastrad return -ENOMEM; 1206 1.1 riastrad 1207 1.1 riastrad if (igt_spinner_init(&c->spin, gt)) 1208 1.1 riastrad goto err_ctx; 1209 1.1 riastrad 1210 1.1 riastrad return 0; 1211 1.1 riastrad 1212 1.1 riastrad err_ctx: 1213 1.1 riastrad kernel_context_close(c->ctx); 1214 1.1 riastrad return -ENOMEM; 1215 1.1 riastrad } 1216 1.1 riastrad 1217 1.1 riastrad static void preempt_client_fini(struct preempt_client *c) 1218 1.1 riastrad { 1219 1.1 riastrad igt_spinner_fini(&c->spin); 1220 1.1 riastrad kernel_context_close(c->ctx); 1221 1.1 riastrad } 1222 1.1 riastrad 1223 1.1 riastrad static int live_nopreempt(void *arg) 1224 1.1 riastrad { 1225 1.1 riastrad struct intel_gt *gt = arg; 1226 1.1 riastrad struct intel_engine_cs *engine; 1227 1.1 riastrad struct preempt_client a, b; 1228 1.1 riastrad enum intel_engine_id id; 1229 1.1 riastrad int err = -ENOMEM; 1230 1.1 riastrad 1231 1.1 riastrad /* 1232 1.1 riastrad * Verify that we can disable preemption for an individual request 1233 1.1 riastrad * that may be being observed and not want to be interrupted. 1234 1.1 riastrad */ 1235 1.1 riastrad 1236 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 1237 1.1 riastrad return 0; 1238 1.1 riastrad 1239 1.1 riastrad if (preempt_client_init(gt, &a)) 1240 1.1 riastrad return -ENOMEM; 1241 1.1 riastrad if (preempt_client_init(gt, &b)) 1242 1.1 riastrad goto err_client_a; 1243 1.1 riastrad b.ctx->sched.priority = I915_USER_PRIORITY(I915_PRIORITY_MAX); 1244 1.1 riastrad 1245 1.1 riastrad for_each_engine(engine, gt, id) { 1246 1.1 riastrad struct i915_request *rq_a, *rq_b; 1247 1.1 riastrad 1248 1.1 riastrad if (!intel_engine_has_preemption(engine)) 1249 1.1 riastrad continue; 1250 1.1 riastrad 1251 1.1 riastrad engine->execlists.preempt_hang.count = 0; 1252 1.1 riastrad 1253 1.1 riastrad rq_a = spinner_create_request(&a.spin, 1254 1.1 riastrad a.ctx, engine, 1255 1.1 riastrad MI_ARB_CHECK); 1256 1.1 riastrad if (IS_ERR(rq_a)) { 1257 1.1 riastrad err = PTR_ERR(rq_a); 1258 1.1 riastrad goto err_client_b; 1259 1.1 riastrad } 1260 1.1 riastrad 1261 1.1 riastrad /* Low priority client, but unpreemptable! */ 1262 1.1 riastrad __set_bit(I915_FENCE_FLAG_NOPREEMPT, &rq_a->fence.flags); 1263 1.1 riastrad 1264 1.1 riastrad i915_request_add(rq_a); 1265 1.1 riastrad if (!igt_wait_for_spinner(&a.spin, rq_a)) { 1266 1.1 riastrad pr_err("First client failed to start\n"); 1267 1.1 riastrad goto err_wedged; 1268 1.1 riastrad } 1269 1.1 riastrad 1270 1.1 riastrad rq_b = spinner_create_request(&b.spin, 1271 1.1 riastrad b.ctx, engine, 1272 1.1 riastrad MI_ARB_CHECK); 1273 1.1 riastrad if (IS_ERR(rq_b)) { 1274 1.1 riastrad err = PTR_ERR(rq_b); 1275 1.1 riastrad goto err_client_b; 1276 1.1 riastrad } 1277 1.1 riastrad 1278 1.1 riastrad i915_request_add(rq_b); 1279 1.1 riastrad 1280 1.1 riastrad /* B is much more important than A! (But A is unpreemptable.) */ 1281 1.1 riastrad GEM_BUG_ON(rq_prio(rq_b) <= rq_prio(rq_a)); 1282 1.1 riastrad 1283 1.1 riastrad /* Wait long enough for preemption and timeslicing */ 1284 1.1 riastrad if (igt_wait_for_spinner(&b.spin, rq_b)) { 1285 1.1 riastrad pr_err("Second client started too early!\n"); 1286 1.1 riastrad goto err_wedged; 1287 1.1 riastrad } 1288 1.1 riastrad 1289 1.1 riastrad igt_spinner_end(&a.spin); 1290 1.1 riastrad 1291 1.1 riastrad if (!igt_wait_for_spinner(&b.spin, rq_b)) { 1292 1.1 riastrad pr_err("Second client failed to start\n"); 1293 1.1 riastrad goto err_wedged; 1294 1.1 riastrad } 1295 1.1 riastrad 1296 1.1 riastrad igt_spinner_end(&b.spin); 1297 1.1 riastrad 1298 1.1 riastrad if (engine->execlists.preempt_hang.count) { 1299 1.1 riastrad pr_err("Preemption recorded x%d; should have been suppressed!\n", 1300 1.1 riastrad engine->execlists.preempt_hang.count); 1301 1.1 riastrad err = -EINVAL; 1302 1.1 riastrad goto err_wedged; 1303 1.1 riastrad } 1304 1.1 riastrad 1305 1.1 riastrad if (igt_flush_test(gt->i915)) 1306 1.1 riastrad goto err_wedged; 1307 1.1 riastrad } 1308 1.1 riastrad 1309 1.1 riastrad err = 0; 1310 1.1 riastrad err_client_b: 1311 1.1 riastrad preempt_client_fini(&b); 1312 1.1 riastrad err_client_a: 1313 1.1 riastrad preempt_client_fini(&a); 1314 1.1 riastrad return err; 1315 1.1 riastrad 1316 1.1 riastrad err_wedged: 1317 1.1 riastrad igt_spinner_end(&b.spin); 1318 1.1 riastrad igt_spinner_end(&a.spin); 1319 1.1 riastrad intel_gt_set_wedged(gt); 1320 1.1 riastrad err = -EIO; 1321 1.1 riastrad goto err_client_b; 1322 1.1 riastrad } 1323 1.1 riastrad 1324 1.1 riastrad struct live_preempt_cancel { 1325 1.1 riastrad struct intel_engine_cs *engine; 1326 1.1 riastrad struct preempt_client a, b; 1327 1.1 riastrad }; 1328 1.1 riastrad 1329 1.1 riastrad static int __cancel_active0(struct live_preempt_cancel *arg) 1330 1.1 riastrad { 1331 1.1 riastrad struct i915_request *rq; 1332 1.1 riastrad struct igt_live_test t; 1333 1.1 riastrad int err; 1334 1.1 riastrad 1335 1.1 riastrad /* Preempt cancel of ELSP0 */ 1336 1.1 riastrad GEM_TRACE("%s(%s)\n", __func__, arg->engine->name); 1337 1.1 riastrad if (igt_live_test_begin(&t, arg->engine->i915, 1338 1.1 riastrad __func__, arg->engine->name)) 1339 1.1 riastrad return -EIO; 1340 1.1 riastrad 1341 1.1 riastrad rq = spinner_create_request(&arg->a.spin, 1342 1.1 riastrad arg->a.ctx, arg->engine, 1343 1.1 riastrad MI_ARB_CHECK); 1344 1.1 riastrad if (IS_ERR(rq)) 1345 1.1 riastrad return PTR_ERR(rq); 1346 1.1 riastrad 1347 1.1 riastrad clear_bit(CONTEXT_BANNED, &rq->context->flags); 1348 1.1 riastrad i915_request_get(rq); 1349 1.1 riastrad i915_request_add(rq); 1350 1.1 riastrad if (!igt_wait_for_spinner(&arg->a.spin, rq)) { 1351 1.1 riastrad err = -EIO; 1352 1.1 riastrad goto out; 1353 1.1 riastrad } 1354 1.1 riastrad 1355 1.1 riastrad intel_context_set_banned(rq->context); 1356 1.1 riastrad err = intel_engine_pulse(arg->engine); 1357 1.1 riastrad if (err) 1358 1.1 riastrad goto out; 1359 1.1 riastrad 1360 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 5) < 0) { 1361 1.1 riastrad err = -EIO; 1362 1.1 riastrad goto out; 1363 1.1 riastrad } 1364 1.1 riastrad 1365 1.1 riastrad if (rq->fence.error != -EIO) { 1366 1.1 riastrad pr_err("Cancelled inflight0 request did not report -EIO\n"); 1367 1.1 riastrad err = -EINVAL; 1368 1.1 riastrad goto out; 1369 1.1 riastrad } 1370 1.1 riastrad 1371 1.1 riastrad out: 1372 1.1 riastrad i915_request_put(rq); 1373 1.1 riastrad if (igt_live_test_end(&t)) 1374 1.1 riastrad err = -EIO; 1375 1.1 riastrad return err; 1376 1.1 riastrad } 1377 1.1 riastrad 1378 1.1 riastrad static int __cancel_active1(struct live_preempt_cancel *arg) 1379 1.1 riastrad { 1380 1.1 riastrad struct i915_request *rq[2] = {}; 1381 1.1 riastrad struct igt_live_test t; 1382 1.1 riastrad int err; 1383 1.1 riastrad 1384 1.1 riastrad /* Preempt cancel of ELSP1 */ 1385 1.1 riastrad GEM_TRACE("%s(%s)\n", __func__, arg->engine->name); 1386 1.1 riastrad if (igt_live_test_begin(&t, arg->engine->i915, 1387 1.1 riastrad __func__, arg->engine->name)) 1388 1.1 riastrad return -EIO; 1389 1.1 riastrad 1390 1.1 riastrad rq[0] = spinner_create_request(&arg->a.spin, 1391 1.1 riastrad arg->a.ctx, arg->engine, 1392 1.1 riastrad MI_NOOP); /* no preemption */ 1393 1.1 riastrad if (IS_ERR(rq[0])) 1394 1.1 riastrad return PTR_ERR(rq[0]); 1395 1.1 riastrad 1396 1.1 riastrad clear_bit(CONTEXT_BANNED, &rq[0]->context->flags); 1397 1.1 riastrad i915_request_get(rq[0]); 1398 1.1 riastrad i915_request_add(rq[0]); 1399 1.1 riastrad if (!igt_wait_for_spinner(&arg->a.spin, rq[0])) { 1400 1.1 riastrad err = -EIO; 1401 1.1 riastrad goto out; 1402 1.1 riastrad } 1403 1.1 riastrad 1404 1.1 riastrad rq[1] = spinner_create_request(&arg->b.spin, 1405 1.1 riastrad arg->b.ctx, arg->engine, 1406 1.1 riastrad MI_ARB_CHECK); 1407 1.1 riastrad if (IS_ERR(rq[1])) { 1408 1.1 riastrad err = PTR_ERR(rq[1]); 1409 1.1 riastrad goto out; 1410 1.1 riastrad } 1411 1.1 riastrad 1412 1.1 riastrad clear_bit(CONTEXT_BANNED, &rq[1]->context->flags); 1413 1.1 riastrad i915_request_get(rq[1]); 1414 1.1 riastrad err = i915_request_await_dma_fence(rq[1], &rq[0]->fence); 1415 1.1 riastrad i915_request_add(rq[1]); 1416 1.1 riastrad if (err) 1417 1.1 riastrad goto out; 1418 1.1 riastrad 1419 1.1 riastrad intel_context_set_banned(rq[1]->context); 1420 1.1 riastrad err = intel_engine_pulse(arg->engine); 1421 1.1 riastrad if (err) 1422 1.1 riastrad goto out; 1423 1.1 riastrad 1424 1.1 riastrad igt_spinner_end(&arg->a.spin); 1425 1.1 riastrad if (i915_request_wait(rq[1], 0, HZ / 5) < 0) { 1426 1.1 riastrad err = -EIO; 1427 1.1 riastrad goto out; 1428 1.1 riastrad } 1429 1.1 riastrad 1430 1.1 riastrad if (rq[0]->fence.error != 0) { 1431 1.1 riastrad pr_err("Normal inflight0 request did not complete\n"); 1432 1.1 riastrad err = -EINVAL; 1433 1.1 riastrad goto out; 1434 1.1 riastrad } 1435 1.1 riastrad 1436 1.1 riastrad if (rq[1]->fence.error != -EIO) { 1437 1.1 riastrad pr_err("Cancelled inflight1 request did not report -EIO\n"); 1438 1.1 riastrad err = -EINVAL; 1439 1.1 riastrad goto out; 1440 1.1 riastrad } 1441 1.1 riastrad 1442 1.1 riastrad out: 1443 1.1 riastrad i915_request_put(rq[1]); 1444 1.1 riastrad i915_request_put(rq[0]); 1445 1.1 riastrad if (igt_live_test_end(&t)) 1446 1.1 riastrad err = -EIO; 1447 1.1 riastrad return err; 1448 1.1 riastrad } 1449 1.1 riastrad 1450 1.1 riastrad static int __cancel_queued(struct live_preempt_cancel *arg) 1451 1.1 riastrad { 1452 1.1 riastrad struct i915_request *rq[3] = {}; 1453 1.1 riastrad struct igt_live_test t; 1454 1.1 riastrad int err; 1455 1.1 riastrad 1456 1.1 riastrad /* Full ELSP and one in the wings */ 1457 1.1 riastrad GEM_TRACE("%s(%s)\n", __func__, arg->engine->name); 1458 1.1 riastrad if (igt_live_test_begin(&t, arg->engine->i915, 1459 1.1 riastrad __func__, arg->engine->name)) 1460 1.1 riastrad return -EIO; 1461 1.1 riastrad 1462 1.1 riastrad rq[0] = spinner_create_request(&arg->a.spin, 1463 1.1 riastrad arg->a.ctx, arg->engine, 1464 1.1 riastrad MI_ARB_CHECK); 1465 1.1 riastrad if (IS_ERR(rq[0])) 1466 1.1 riastrad return PTR_ERR(rq[0]); 1467 1.1 riastrad 1468 1.1 riastrad clear_bit(CONTEXT_BANNED, &rq[0]->context->flags); 1469 1.1 riastrad i915_request_get(rq[0]); 1470 1.1 riastrad i915_request_add(rq[0]); 1471 1.1 riastrad if (!igt_wait_for_spinner(&arg->a.spin, rq[0])) { 1472 1.1 riastrad err = -EIO; 1473 1.1 riastrad goto out; 1474 1.1 riastrad } 1475 1.1 riastrad 1476 1.1 riastrad rq[1] = igt_request_alloc(arg->b.ctx, arg->engine); 1477 1.1 riastrad if (IS_ERR(rq[1])) { 1478 1.1 riastrad err = PTR_ERR(rq[1]); 1479 1.1 riastrad goto out; 1480 1.1 riastrad } 1481 1.1 riastrad 1482 1.1 riastrad clear_bit(CONTEXT_BANNED, &rq[1]->context->flags); 1483 1.1 riastrad i915_request_get(rq[1]); 1484 1.1 riastrad err = i915_request_await_dma_fence(rq[1], &rq[0]->fence); 1485 1.1 riastrad i915_request_add(rq[1]); 1486 1.1 riastrad if (err) 1487 1.1 riastrad goto out; 1488 1.1 riastrad 1489 1.1 riastrad rq[2] = spinner_create_request(&arg->b.spin, 1490 1.1 riastrad arg->a.ctx, arg->engine, 1491 1.1 riastrad MI_ARB_CHECK); 1492 1.1 riastrad if (IS_ERR(rq[2])) { 1493 1.1 riastrad err = PTR_ERR(rq[2]); 1494 1.1 riastrad goto out; 1495 1.1 riastrad } 1496 1.1 riastrad 1497 1.1 riastrad i915_request_get(rq[2]); 1498 1.1 riastrad err = i915_request_await_dma_fence(rq[2], &rq[1]->fence); 1499 1.1 riastrad i915_request_add(rq[2]); 1500 1.1 riastrad if (err) 1501 1.1 riastrad goto out; 1502 1.1 riastrad 1503 1.1 riastrad intel_context_set_banned(rq[2]->context); 1504 1.1 riastrad err = intel_engine_pulse(arg->engine); 1505 1.1 riastrad if (err) 1506 1.1 riastrad goto out; 1507 1.1 riastrad 1508 1.1 riastrad if (i915_request_wait(rq[2], 0, HZ / 5) < 0) { 1509 1.1 riastrad err = -EIO; 1510 1.1 riastrad goto out; 1511 1.1 riastrad } 1512 1.1 riastrad 1513 1.1 riastrad if (rq[0]->fence.error != -EIO) { 1514 1.1 riastrad pr_err("Cancelled inflight0 request did not report -EIO\n"); 1515 1.1 riastrad err = -EINVAL; 1516 1.1 riastrad goto out; 1517 1.1 riastrad } 1518 1.1 riastrad 1519 1.1 riastrad if (rq[1]->fence.error != 0) { 1520 1.1 riastrad pr_err("Normal inflight1 request did not complete\n"); 1521 1.1 riastrad err = -EINVAL; 1522 1.1 riastrad goto out; 1523 1.1 riastrad } 1524 1.1 riastrad 1525 1.1 riastrad if (rq[2]->fence.error != -EIO) { 1526 1.1 riastrad pr_err("Cancelled queued request did not report -EIO\n"); 1527 1.1 riastrad err = -EINVAL; 1528 1.1 riastrad goto out; 1529 1.1 riastrad } 1530 1.1 riastrad 1531 1.1 riastrad out: 1532 1.1 riastrad i915_request_put(rq[2]); 1533 1.1 riastrad i915_request_put(rq[1]); 1534 1.1 riastrad i915_request_put(rq[0]); 1535 1.1 riastrad if (igt_live_test_end(&t)) 1536 1.1 riastrad err = -EIO; 1537 1.1 riastrad return err; 1538 1.1 riastrad } 1539 1.1 riastrad 1540 1.1 riastrad static int __cancel_hostile(struct live_preempt_cancel *arg) 1541 1.1 riastrad { 1542 1.1 riastrad struct i915_request *rq; 1543 1.1 riastrad int err; 1544 1.1 riastrad 1545 1.1 riastrad /* Preempt cancel non-preemptible spinner in ELSP0 */ 1546 1.1 riastrad if (!IS_ACTIVE(CONFIG_DRM_I915_PREEMPT_TIMEOUT)) 1547 1.1 riastrad return 0; 1548 1.1 riastrad 1549 1.1 riastrad GEM_TRACE("%s(%s)\n", __func__, arg->engine->name); 1550 1.1 riastrad rq = spinner_create_request(&arg->a.spin, 1551 1.1 riastrad arg->a.ctx, arg->engine, 1552 1.1 riastrad MI_NOOP); /* preemption disabled */ 1553 1.1 riastrad if (IS_ERR(rq)) 1554 1.1 riastrad return PTR_ERR(rq); 1555 1.1 riastrad 1556 1.1 riastrad clear_bit(CONTEXT_BANNED, &rq->context->flags); 1557 1.1 riastrad i915_request_get(rq); 1558 1.1 riastrad i915_request_add(rq); 1559 1.1 riastrad if (!igt_wait_for_spinner(&arg->a.spin, rq)) { 1560 1.1 riastrad err = -EIO; 1561 1.1 riastrad goto out; 1562 1.1 riastrad } 1563 1.1 riastrad 1564 1.1 riastrad intel_context_set_banned(rq->context); 1565 1.1 riastrad err = intel_engine_pulse(arg->engine); /* force reset */ 1566 1.1 riastrad if (err) 1567 1.1 riastrad goto out; 1568 1.1 riastrad 1569 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 5) < 0) { 1570 1.1 riastrad err = -EIO; 1571 1.1 riastrad goto out; 1572 1.1 riastrad } 1573 1.1 riastrad 1574 1.1 riastrad if (rq->fence.error != -EIO) { 1575 1.1 riastrad pr_err("Cancelled inflight0 request did not report -EIO\n"); 1576 1.1 riastrad err = -EINVAL; 1577 1.1 riastrad goto out; 1578 1.1 riastrad } 1579 1.1 riastrad 1580 1.1 riastrad out: 1581 1.1 riastrad i915_request_put(rq); 1582 1.1 riastrad if (igt_flush_test(arg->engine->i915)) 1583 1.1 riastrad err = -EIO; 1584 1.1 riastrad return err; 1585 1.1 riastrad } 1586 1.1 riastrad 1587 1.1 riastrad static int live_preempt_cancel(void *arg) 1588 1.1 riastrad { 1589 1.1 riastrad struct intel_gt *gt = arg; 1590 1.1 riastrad struct live_preempt_cancel data; 1591 1.1 riastrad enum intel_engine_id id; 1592 1.1 riastrad int err = -ENOMEM; 1593 1.1 riastrad 1594 1.1 riastrad /* 1595 1.1 riastrad * To cancel an inflight context, we need to first remove it from the 1596 1.1 riastrad * GPU. That sounds like preemption! Plus a little bit of bookkeeping. 1597 1.1 riastrad */ 1598 1.1 riastrad 1599 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 1600 1.1 riastrad return 0; 1601 1.1 riastrad 1602 1.1 riastrad if (preempt_client_init(gt, &data.a)) 1603 1.1 riastrad return -ENOMEM; 1604 1.1 riastrad if (preempt_client_init(gt, &data.b)) 1605 1.1 riastrad goto err_client_a; 1606 1.1 riastrad 1607 1.1 riastrad for_each_engine(data.engine, gt, id) { 1608 1.1 riastrad if (!intel_engine_has_preemption(data.engine)) 1609 1.1 riastrad continue; 1610 1.1 riastrad 1611 1.1 riastrad err = __cancel_active0(&data); 1612 1.1 riastrad if (err) 1613 1.1 riastrad goto err_wedged; 1614 1.1 riastrad 1615 1.1 riastrad err = __cancel_active1(&data); 1616 1.1 riastrad if (err) 1617 1.1 riastrad goto err_wedged; 1618 1.1 riastrad 1619 1.1 riastrad err = __cancel_queued(&data); 1620 1.1 riastrad if (err) 1621 1.1 riastrad goto err_wedged; 1622 1.1 riastrad 1623 1.1 riastrad err = __cancel_hostile(&data); 1624 1.1 riastrad if (err) 1625 1.1 riastrad goto err_wedged; 1626 1.1 riastrad } 1627 1.1 riastrad 1628 1.1 riastrad err = 0; 1629 1.1 riastrad err_client_b: 1630 1.1 riastrad preempt_client_fini(&data.b); 1631 1.1 riastrad err_client_a: 1632 1.1 riastrad preempt_client_fini(&data.a); 1633 1.1 riastrad return err; 1634 1.1 riastrad 1635 1.1 riastrad err_wedged: 1636 1.1 riastrad GEM_TRACE_DUMP(); 1637 1.1 riastrad igt_spinner_end(&data.b.spin); 1638 1.1 riastrad igt_spinner_end(&data.a.spin); 1639 1.1 riastrad intel_gt_set_wedged(gt); 1640 1.1 riastrad goto err_client_b; 1641 1.1 riastrad } 1642 1.1 riastrad 1643 1.1 riastrad static int live_suppress_self_preempt(void *arg) 1644 1.1 riastrad { 1645 1.1 riastrad struct intel_gt *gt = arg; 1646 1.1 riastrad struct intel_engine_cs *engine; 1647 1.1 riastrad struct i915_sched_attr attr = { 1648 1.1 riastrad .priority = I915_USER_PRIORITY(I915_PRIORITY_MAX) 1649 1.1 riastrad }; 1650 1.1 riastrad struct preempt_client a, b; 1651 1.1 riastrad enum intel_engine_id id; 1652 1.1 riastrad int err = -ENOMEM; 1653 1.1 riastrad 1654 1.1 riastrad /* 1655 1.1 riastrad * Verify that if a preemption request does not cause a change in 1656 1.1 riastrad * the current execution order, the preempt-to-idle injection is 1657 1.1 riastrad * skipped and that we do not accidentally apply it after the CS 1658 1.1 riastrad * completion event. 1659 1.1 riastrad */ 1660 1.1 riastrad 1661 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 1662 1.1 riastrad return 0; 1663 1.1 riastrad 1664 1.1 riastrad if (USES_GUC_SUBMISSION(gt->i915)) 1665 1.1 riastrad return 0; /* presume black blox */ 1666 1.1 riastrad 1667 1.1 riastrad if (intel_vgpu_active(gt->i915)) 1668 1.1 riastrad return 0; /* GVT forces single port & request submission */ 1669 1.1 riastrad 1670 1.1 riastrad if (preempt_client_init(gt, &a)) 1671 1.1 riastrad return -ENOMEM; 1672 1.1 riastrad if (preempt_client_init(gt, &b)) 1673 1.1 riastrad goto err_client_a; 1674 1.1 riastrad 1675 1.1 riastrad for_each_engine(engine, gt, id) { 1676 1.1 riastrad struct i915_request *rq_a, *rq_b; 1677 1.1 riastrad int depth; 1678 1.1 riastrad 1679 1.1 riastrad if (!intel_engine_has_preemption(engine)) 1680 1.1 riastrad continue; 1681 1.1 riastrad 1682 1.1 riastrad if (igt_flush_test(gt->i915)) 1683 1.1 riastrad goto err_wedged; 1684 1.1 riastrad 1685 1.1 riastrad intel_engine_pm_get(engine); 1686 1.1 riastrad engine->execlists.preempt_hang.count = 0; 1687 1.1 riastrad 1688 1.1 riastrad rq_a = spinner_create_request(&a.spin, 1689 1.1 riastrad a.ctx, engine, 1690 1.1 riastrad MI_NOOP); 1691 1.1 riastrad if (IS_ERR(rq_a)) { 1692 1.1 riastrad err = PTR_ERR(rq_a); 1693 1.1 riastrad intel_engine_pm_put(engine); 1694 1.1 riastrad goto err_client_b; 1695 1.1 riastrad } 1696 1.1 riastrad 1697 1.1 riastrad i915_request_add(rq_a); 1698 1.1 riastrad if (!igt_wait_for_spinner(&a.spin, rq_a)) { 1699 1.1 riastrad pr_err("First client failed to start\n"); 1700 1.1 riastrad intel_engine_pm_put(engine); 1701 1.1 riastrad goto err_wedged; 1702 1.1 riastrad } 1703 1.1 riastrad 1704 1.1 riastrad /* Keep postponing the timer to avoid premature slicing */ 1705 1.1 riastrad mod_timer(&engine->execlists.timer, jiffies + HZ); 1706 1.1 riastrad for (depth = 0; depth < 8; depth++) { 1707 1.1 riastrad rq_b = spinner_create_request(&b.spin, 1708 1.1 riastrad b.ctx, engine, 1709 1.1 riastrad MI_NOOP); 1710 1.1 riastrad if (IS_ERR(rq_b)) { 1711 1.1 riastrad err = PTR_ERR(rq_b); 1712 1.1 riastrad intel_engine_pm_put(engine); 1713 1.1 riastrad goto err_client_b; 1714 1.1 riastrad } 1715 1.1 riastrad i915_request_add(rq_b); 1716 1.1 riastrad 1717 1.1 riastrad GEM_BUG_ON(i915_request_completed(rq_a)); 1718 1.1 riastrad engine->schedule(rq_a, &attr); 1719 1.1 riastrad igt_spinner_end(&a.spin); 1720 1.1 riastrad 1721 1.1 riastrad if (!igt_wait_for_spinner(&b.spin, rq_b)) { 1722 1.1 riastrad pr_err("Second client failed to start\n"); 1723 1.1 riastrad intel_engine_pm_put(engine); 1724 1.1 riastrad goto err_wedged; 1725 1.1 riastrad } 1726 1.1 riastrad 1727 1.1 riastrad swap(a, b); 1728 1.1 riastrad rq_a = rq_b; 1729 1.1 riastrad } 1730 1.1 riastrad igt_spinner_end(&a.spin); 1731 1.1 riastrad 1732 1.1 riastrad if (engine->execlists.preempt_hang.count) { 1733 1.1 riastrad pr_err("Preemption on %s recorded x%d, depth %d; should have been suppressed!\n", 1734 1.1 riastrad engine->name, 1735 1.1 riastrad engine->execlists.preempt_hang.count, 1736 1.1 riastrad depth); 1737 1.1 riastrad intel_engine_pm_put(engine); 1738 1.1 riastrad err = -EINVAL; 1739 1.1 riastrad goto err_client_b; 1740 1.1 riastrad } 1741 1.1 riastrad 1742 1.1 riastrad intel_engine_pm_put(engine); 1743 1.1 riastrad if (igt_flush_test(gt->i915)) 1744 1.1 riastrad goto err_wedged; 1745 1.1 riastrad } 1746 1.1 riastrad 1747 1.1 riastrad err = 0; 1748 1.1 riastrad err_client_b: 1749 1.1 riastrad preempt_client_fini(&b); 1750 1.1 riastrad err_client_a: 1751 1.1 riastrad preempt_client_fini(&a); 1752 1.1 riastrad return err; 1753 1.1 riastrad 1754 1.1 riastrad err_wedged: 1755 1.1 riastrad igt_spinner_end(&b.spin); 1756 1.1 riastrad igt_spinner_end(&a.spin); 1757 1.1 riastrad intel_gt_set_wedged(gt); 1758 1.1 riastrad err = -EIO; 1759 1.1 riastrad goto err_client_b; 1760 1.1 riastrad } 1761 1.1 riastrad 1762 1.1 riastrad static int __i915_sw_fence_call 1763 1.1 riastrad dummy_notify(struct i915_sw_fence *fence, enum i915_sw_fence_notify state) 1764 1.1 riastrad { 1765 1.1 riastrad return NOTIFY_DONE; 1766 1.1 riastrad } 1767 1.1 riastrad 1768 1.1 riastrad static struct i915_request *dummy_request(struct intel_engine_cs *engine) 1769 1.1 riastrad { 1770 1.1 riastrad struct i915_request *rq; 1771 1.1 riastrad 1772 1.1 riastrad rq = kzalloc(sizeof(*rq), GFP_KERNEL); 1773 1.1 riastrad if (!rq) 1774 1.1 riastrad return NULL; 1775 1.1 riastrad 1776 1.1 riastrad rq->engine = engine; 1777 1.1 riastrad 1778 1.1 riastrad spin_lock_init(&rq->lock); 1779 1.1 riastrad INIT_LIST_HEAD(&rq->fence.cb_list); 1780 1.1 riastrad rq->fence.lock = &rq->lock; 1781 1.1 riastrad rq->fence.ops = &i915_fence_ops; 1782 1.1 riastrad 1783 1.1 riastrad i915_sched_node_init(&rq->sched); 1784 1.1 riastrad 1785 1.1 riastrad /* mark this request as permanently incomplete */ 1786 1.1 riastrad rq->fence.seqno = 1; 1787 1.1 riastrad BUILD_BUG_ON(sizeof(rq->fence.seqno) != 8); /* upper 32b == 0 */ 1788 1.1 riastrad rq->hwsp_seqno = (u32 *)&rq->fence.seqno + 1; 1789 1.1 riastrad GEM_BUG_ON(i915_request_completed(rq)); 1790 1.1 riastrad 1791 1.1 riastrad i915_sw_fence_init(&rq->submit, dummy_notify); 1792 1.1 riastrad set_bit(I915_FENCE_FLAG_ACTIVE, &rq->fence.flags); 1793 1.1 riastrad 1794 1.1 riastrad spin_lock_init(&rq->lock); 1795 1.1 riastrad rq->fence.lock = &rq->lock; 1796 1.1 riastrad INIT_LIST_HEAD(&rq->fence.cb_list); 1797 1.1 riastrad 1798 1.1 riastrad return rq; 1799 1.1 riastrad } 1800 1.1 riastrad 1801 1.1 riastrad static void dummy_request_free(struct i915_request *dummy) 1802 1.1 riastrad { 1803 1.1 riastrad /* We have to fake the CS interrupt to kick the next request */ 1804 1.1 riastrad i915_sw_fence_commit(&dummy->submit); 1805 1.1 riastrad 1806 1.1 riastrad i915_request_mark_complete(dummy); 1807 1.1 riastrad dma_fence_signal(&dummy->fence); 1808 1.1 riastrad 1809 1.1 riastrad i915_sched_node_fini(&dummy->sched); 1810 1.1 riastrad i915_sw_fence_fini(&dummy->submit); 1811 1.1 riastrad 1812 1.1 riastrad dma_fence_free(&dummy->fence); 1813 1.1 riastrad } 1814 1.1 riastrad 1815 1.1 riastrad static int live_suppress_wait_preempt(void *arg) 1816 1.1 riastrad { 1817 1.1 riastrad struct intel_gt *gt = arg; 1818 1.1 riastrad struct preempt_client client[4]; 1819 1.1 riastrad struct i915_request *rq[ARRAY_SIZE(client)] = {}; 1820 1.1 riastrad struct intel_engine_cs *engine; 1821 1.1 riastrad enum intel_engine_id id; 1822 1.1 riastrad int err = -ENOMEM; 1823 1.1 riastrad int i; 1824 1.1 riastrad 1825 1.1 riastrad /* 1826 1.1 riastrad * Waiters are given a little priority nudge, but not enough 1827 1.1 riastrad * to actually cause any preemption. Double check that we do 1828 1.1 riastrad * not needlessly generate preempt-to-idle cycles. 1829 1.1 riastrad */ 1830 1.1 riastrad 1831 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 1832 1.1 riastrad return 0; 1833 1.1 riastrad 1834 1.1 riastrad if (preempt_client_init(gt, &client[0])) /* ELSP[0] */ 1835 1.1 riastrad return -ENOMEM; 1836 1.1 riastrad if (preempt_client_init(gt, &client[1])) /* ELSP[1] */ 1837 1.1 riastrad goto err_client_0; 1838 1.1 riastrad if (preempt_client_init(gt, &client[2])) /* head of queue */ 1839 1.1 riastrad goto err_client_1; 1840 1.1 riastrad if (preempt_client_init(gt, &client[3])) /* bystander */ 1841 1.1 riastrad goto err_client_2; 1842 1.1 riastrad 1843 1.1 riastrad for_each_engine(engine, gt, id) { 1844 1.1 riastrad int depth; 1845 1.1 riastrad 1846 1.1 riastrad if (!intel_engine_has_preemption(engine)) 1847 1.1 riastrad continue; 1848 1.1 riastrad 1849 1.1 riastrad if (!engine->emit_init_breadcrumb) 1850 1.1 riastrad continue; 1851 1.1 riastrad 1852 1.1 riastrad for (depth = 0; depth < ARRAY_SIZE(client); depth++) { 1853 1.1 riastrad struct i915_request *dummy; 1854 1.1 riastrad 1855 1.1 riastrad engine->execlists.preempt_hang.count = 0; 1856 1.1 riastrad 1857 1.1 riastrad dummy = dummy_request(engine); 1858 1.1 riastrad if (!dummy) 1859 1.1 riastrad goto err_client_3; 1860 1.1 riastrad 1861 1.1 riastrad for (i = 0; i < ARRAY_SIZE(client); i++) { 1862 1.1 riastrad struct i915_request *this; 1863 1.1 riastrad 1864 1.1 riastrad this = spinner_create_request(&client[i].spin, 1865 1.1 riastrad client[i].ctx, engine, 1866 1.1 riastrad MI_NOOP); 1867 1.1 riastrad if (IS_ERR(this)) { 1868 1.1 riastrad err = PTR_ERR(this); 1869 1.1 riastrad goto err_wedged; 1870 1.1 riastrad } 1871 1.1 riastrad 1872 1.1 riastrad /* Disable NEWCLIENT promotion */ 1873 1.1 riastrad __i915_active_fence_set(&i915_request_timeline(this)->last_request, 1874 1.1 riastrad &dummy->fence); 1875 1.1 riastrad 1876 1.1 riastrad rq[i] = i915_request_get(this); 1877 1.1 riastrad i915_request_add(this); 1878 1.1 riastrad } 1879 1.1 riastrad 1880 1.1 riastrad dummy_request_free(dummy); 1881 1.1 riastrad 1882 1.1 riastrad GEM_BUG_ON(i915_request_completed(rq[0])); 1883 1.1 riastrad if (!igt_wait_for_spinner(&client[0].spin, rq[0])) { 1884 1.1 riastrad pr_err("%s: First client failed to start\n", 1885 1.1 riastrad engine->name); 1886 1.1 riastrad goto err_wedged; 1887 1.1 riastrad } 1888 1.1 riastrad GEM_BUG_ON(!i915_request_started(rq[0])); 1889 1.1 riastrad 1890 1.1 riastrad if (i915_request_wait(rq[depth], 1891 1.1 riastrad I915_WAIT_PRIORITY, 1892 1.1 riastrad 1) != -ETIME) { 1893 1.1 riastrad pr_err("%s: Waiter depth:%d completed!\n", 1894 1.1 riastrad engine->name, depth); 1895 1.1 riastrad goto err_wedged; 1896 1.1 riastrad } 1897 1.1 riastrad 1898 1.1 riastrad for (i = 0; i < ARRAY_SIZE(client); i++) { 1899 1.1 riastrad igt_spinner_end(&client[i].spin); 1900 1.1 riastrad i915_request_put(rq[i]); 1901 1.1 riastrad rq[i] = NULL; 1902 1.1 riastrad } 1903 1.1 riastrad 1904 1.1 riastrad if (igt_flush_test(gt->i915)) 1905 1.1 riastrad goto err_wedged; 1906 1.1 riastrad 1907 1.1 riastrad if (engine->execlists.preempt_hang.count) { 1908 1.1 riastrad pr_err("%s: Preemption recorded x%d, depth %d; should have been suppressed!\n", 1909 1.1 riastrad engine->name, 1910 1.1 riastrad engine->execlists.preempt_hang.count, 1911 1.1 riastrad depth); 1912 1.1 riastrad err = -EINVAL; 1913 1.1 riastrad goto err_client_3; 1914 1.1 riastrad } 1915 1.1 riastrad } 1916 1.1 riastrad } 1917 1.1 riastrad 1918 1.1 riastrad err = 0; 1919 1.1 riastrad err_client_3: 1920 1.1 riastrad preempt_client_fini(&client[3]); 1921 1.1 riastrad err_client_2: 1922 1.1 riastrad preempt_client_fini(&client[2]); 1923 1.1 riastrad err_client_1: 1924 1.1 riastrad preempt_client_fini(&client[1]); 1925 1.1 riastrad err_client_0: 1926 1.1 riastrad preempt_client_fini(&client[0]); 1927 1.1 riastrad return err; 1928 1.1 riastrad 1929 1.1 riastrad err_wedged: 1930 1.1 riastrad for (i = 0; i < ARRAY_SIZE(client); i++) { 1931 1.1 riastrad igt_spinner_end(&client[i].spin); 1932 1.1 riastrad i915_request_put(rq[i]); 1933 1.1 riastrad } 1934 1.1 riastrad intel_gt_set_wedged(gt); 1935 1.1 riastrad err = -EIO; 1936 1.1 riastrad goto err_client_3; 1937 1.1 riastrad } 1938 1.1 riastrad 1939 1.1 riastrad static int live_chain_preempt(void *arg) 1940 1.1 riastrad { 1941 1.1 riastrad struct intel_gt *gt = arg; 1942 1.1 riastrad struct intel_engine_cs *engine; 1943 1.1 riastrad struct preempt_client hi, lo; 1944 1.1 riastrad enum intel_engine_id id; 1945 1.1 riastrad int err = -ENOMEM; 1946 1.1 riastrad 1947 1.1 riastrad /* 1948 1.1 riastrad * Build a chain AB...BA between two contexts (A, B) and request 1949 1.1 riastrad * preemption of the last request. It should then complete before 1950 1.1 riastrad * the previously submitted spinner in B. 1951 1.1 riastrad */ 1952 1.1 riastrad 1953 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 1954 1.1 riastrad return 0; 1955 1.1 riastrad 1956 1.1 riastrad if (preempt_client_init(gt, &hi)) 1957 1.1 riastrad return -ENOMEM; 1958 1.1 riastrad 1959 1.1 riastrad if (preempt_client_init(gt, &lo)) 1960 1.1 riastrad goto err_client_hi; 1961 1.1 riastrad 1962 1.1 riastrad for_each_engine(engine, gt, id) { 1963 1.1 riastrad struct i915_sched_attr attr = { 1964 1.1 riastrad .priority = I915_USER_PRIORITY(I915_PRIORITY_MAX), 1965 1.1 riastrad }; 1966 1.1 riastrad struct igt_live_test t; 1967 1.1 riastrad struct i915_request *rq; 1968 1.1 riastrad int ring_size, count, i; 1969 1.1 riastrad 1970 1.1 riastrad if (!intel_engine_has_preemption(engine)) 1971 1.1 riastrad continue; 1972 1.1 riastrad 1973 1.1 riastrad rq = spinner_create_request(&lo.spin, 1974 1.1 riastrad lo.ctx, engine, 1975 1.1 riastrad MI_ARB_CHECK); 1976 1.1 riastrad if (IS_ERR(rq)) 1977 1.1 riastrad goto err_wedged; 1978 1.1 riastrad 1979 1.1 riastrad i915_request_get(rq); 1980 1.1 riastrad i915_request_add(rq); 1981 1.1 riastrad 1982 1.1 riastrad ring_size = rq->wa_tail - rq->head; 1983 1.1 riastrad if (ring_size < 0) 1984 1.1 riastrad ring_size += rq->ring->size; 1985 1.1 riastrad ring_size = rq->ring->size / ring_size; 1986 1.1 riastrad pr_debug("%s(%s): Using maximum of %d requests\n", 1987 1.1 riastrad __func__, engine->name, ring_size); 1988 1.1 riastrad 1989 1.1 riastrad igt_spinner_end(&lo.spin); 1990 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 2) < 0) { 1991 1.1 riastrad pr_err("Timed out waiting to flush %s\n", engine->name); 1992 1.1 riastrad i915_request_put(rq); 1993 1.1 riastrad goto err_wedged; 1994 1.1 riastrad } 1995 1.1 riastrad i915_request_put(rq); 1996 1.1 riastrad 1997 1.1 riastrad if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) { 1998 1.1 riastrad err = -EIO; 1999 1.1 riastrad goto err_wedged; 2000 1.1 riastrad } 2001 1.1 riastrad 2002 1.1 riastrad for_each_prime_number_from(count, 1, ring_size) { 2003 1.1 riastrad rq = spinner_create_request(&hi.spin, 2004 1.1 riastrad hi.ctx, engine, 2005 1.1 riastrad MI_ARB_CHECK); 2006 1.1 riastrad if (IS_ERR(rq)) 2007 1.1 riastrad goto err_wedged; 2008 1.1 riastrad i915_request_add(rq); 2009 1.1 riastrad if (!igt_wait_for_spinner(&hi.spin, rq)) 2010 1.1 riastrad goto err_wedged; 2011 1.1 riastrad 2012 1.1 riastrad rq = spinner_create_request(&lo.spin, 2013 1.1 riastrad lo.ctx, engine, 2014 1.1 riastrad MI_ARB_CHECK); 2015 1.1 riastrad if (IS_ERR(rq)) 2016 1.1 riastrad goto err_wedged; 2017 1.1 riastrad i915_request_add(rq); 2018 1.1 riastrad 2019 1.1 riastrad for (i = 0; i < count; i++) { 2020 1.1 riastrad rq = igt_request_alloc(lo.ctx, engine); 2021 1.1 riastrad if (IS_ERR(rq)) 2022 1.1 riastrad goto err_wedged; 2023 1.1 riastrad i915_request_add(rq); 2024 1.1 riastrad } 2025 1.1 riastrad 2026 1.1 riastrad rq = igt_request_alloc(hi.ctx, engine); 2027 1.1 riastrad if (IS_ERR(rq)) 2028 1.1 riastrad goto err_wedged; 2029 1.1 riastrad 2030 1.1 riastrad i915_request_get(rq); 2031 1.1 riastrad i915_request_add(rq); 2032 1.1 riastrad engine->schedule(rq, &attr); 2033 1.1 riastrad 2034 1.1 riastrad igt_spinner_end(&hi.spin); 2035 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 5) < 0) { 2036 1.1 riastrad struct drm_printer p = 2037 1.1 riastrad drm_info_printer(gt->i915->drm.dev); 2038 1.1 riastrad 2039 1.1 riastrad pr_err("Failed to preempt over chain of %d\n", 2040 1.1 riastrad count); 2041 1.1 riastrad intel_engine_dump(engine, &p, 2042 1.1 riastrad "%s\n", engine->name); 2043 1.1 riastrad i915_request_put(rq); 2044 1.1 riastrad goto err_wedged; 2045 1.1 riastrad } 2046 1.1 riastrad igt_spinner_end(&lo.spin); 2047 1.1 riastrad i915_request_put(rq); 2048 1.1 riastrad 2049 1.1 riastrad rq = igt_request_alloc(lo.ctx, engine); 2050 1.1 riastrad if (IS_ERR(rq)) 2051 1.1 riastrad goto err_wedged; 2052 1.1 riastrad 2053 1.1 riastrad i915_request_get(rq); 2054 1.1 riastrad i915_request_add(rq); 2055 1.1 riastrad 2056 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 5) < 0) { 2057 1.1 riastrad struct drm_printer p = 2058 1.1 riastrad drm_info_printer(gt->i915->drm.dev); 2059 1.1 riastrad 2060 1.1 riastrad pr_err("Failed to flush low priority chain of %d requests\n", 2061 1.1 riastrad count); 2062 1.1 riastrad intel_engine_dump(engine, &p, 2063 1.1 riastrad "%s\n", engine->name); 2064 1.1 riastrad 2065 1.1 riastrad i915_request_put(rq); 2066 1.1 riastrad goto err_wedged; 2067 1.1 riastrad } 2068 1.1 riastrad i915_request_put(rq); 2069 1.1 riastrad } 2070 1.1 riastrad 2071 1.1 riastrad if (igt_live_test_end(&t)) { 2072 1.1 riastrad err = -EIO; 2073 1.1 riastrad goto err_wedged; 2074 1.1 riastrad } 2075 1.1 riastrad } 2076 1.1 riastrad 2077 1.1 riastrad err = 0; 2078 1.1 riastrad err_client_lo: 2079 1.1 riastrad preempt_client_fini(&lo); 2080 1.1 riastrad err_client_hi: 2081 1.1 riastrad preempt_client_fini(&hi); 2082 1.1 riastrad return err; 2083 1.1 riastrad 2084 1.1 riastrad err_wedged: 2085 1.1 riastrad igt_spinner_end(&hi.spin); 2086 1.1 riastrad igt_spinner_end(&lo.spin); 2087 1.1 riastrad intel_gt_set_wedged(gt); 2088 1.1 riastrad err = -EIO; 2089 1.1 riastrad goto err_client_lo; 2090 1.1 riastrad } 2091 1.1 riastrad 2092 1.1 riastrad static int create_gang(struct intel_engine_cs *engine, 2093 1.1 riastrad struct i915_request **prev) 2094 1.1 riastrad { 2095 1.1 riastrad struct drm_i915_gem_object *obj; 2096 1.1 riastrad struct intel_context *ce; 2097 1.1 riastrad struct i915_request *rq; 2098 1.1 riastrad struct i915_vma *vma; 2099 1.1 riastrad u32 *cs; 2100 1.1 riastrad int err; 2101 1.1 riastrad 2102 1.1 riastrad ce = intel_context_create(engine); 2103 1.1 riastrad if (IS_ERR(ce)) 2104 1.1 riastrad return PTR_ERR(ce); 2105 1.1 riastrad 2106 1.1 riastrad obj = i915_gem_object_create_internal(engine->i915, 4096); 2107 1.1 riastrad if (IS_ERR(obj)) { 2108 1.1 riastrad err = PTR_ERR(obj); 2109 1.1 riastrad goto err_ce; 2110 1.1 riastrad } 2111 1.1 riastrad 2112 1.1 riastrad vma = i915_vma_instance(obj, ce->vm, NULL); 2113 1.1 riastrad if (IS_ERR(vma)) { 2114 1.1 riastrad err = PTR_ERR(vma); 2115 1.1 riastrad goto err_obj; 2116 1.1 riastrad } 2117 1.1 riastrad 2118 1.1 riastrad err = i915_vma_pin(vma, 0, 0, PIN_USER); 2119 1.1 riastrad if (err) 2120 1.1 riastrad goto err_obj; 2121 1.1 riastrad 2122 1.1 riastrad cs = i915_gem_object_pin_map(obj, I915_MAP_WC); 2123 1.1 riastrad if (IS_ERR(cs)) 2124 1.1 riastrad goto err_obj; 2125 1.1 riastrad 2126 1.1 riastrad /* Semaphore target: spin until zero */ 2127 1.1 riastrad *cs++ = MI_ARB_ON_OFF | MI_ARB_ENABLE; 2128 1.1 riastrad 2129 1.1 riastrad *cs++ = MI_SEMAPHORE_WAIT | 2130 1.1 riastrad MI_SEMAPHORE_POLL | 2131 1.1 riastrad MI_SEMAPHORE_SAD_EQ_SDD; 2132 1.1 riastrad *cs++ = 0; 2133 1.1 riastrad *cs++ = lower_32_bits(vma->node.start); 2134 1.1 riastrad *cs++ = upper_32_bits(vma->node.start); 2135 1.1 riastrad 2136 1.1 riastrad if (*prev) { 2137 1.1 riastrad u64 offset = (*prev)->batch->node.start; 2138 1.1 riastrad 2139 1.1 riastrad /* Terminate the spinner in the next lower priority batch. */ 2140 1.1 riastrad *cs++ = MI_STORE_DWORD_IMM_GEN4; 2141 1.1 riastrad *cs++ = lower_32_bits(offset); 2142 1.1 riastrad *cs++ = upper_32_bits(offset); 2143 1.1 riastrad *cs++ = 0; 2144 1.1 riastrad } 2145 1.1 riastrad 2146 1.1 riastrad *cs++ = MI_BATCH_BUFFER_END; 2147 1.1 riastrad i915_gem_object_flush_map(obj); 2148 1.1 riastrad i915_gem_object_unpin_map(obj); 2149 1.1 riastrad 2150 1.1 riastrad rq = intel_context_create_request(ce); 2151 1.1 riastrad if (IS_ERR(rq)) 2152 1.1 riastrad goto err_obj; 2153 1.1 riastrad 2154 1.1 riastrad rq->batch = vma; 2155 1.1 riastrad i915_request_get(rq); 2156 1.1 riastrad 2157 1.1 riastrad i915_vma_lock(vma); 2158 1.1 riastrad err = i915_request_await_object(rq, vma->obj, false); 2159 1.1 riastrad if (!err) 2160 1.1 riastrad err = i915_vma_move_to_active(vma, rq, 0); 2161 1.1 riastrad if (!err) 2162 1.1 riastrad err = rq->engine->emit_bb_start(rq, 2163 1.1 riastrad vma->node.start, 2164 1.1 riastrad PAGE_SIZE, 0); 2165 1.1 riastrad i915_vma_unlock(vma); 2166 1.1 riastrad i915_request_add(rq); 2167 1.1 riastrad if (err) 2168 1.1 riastrad goto err_rq; 2169 1.1 riastrad 2170 1.1 riastrad i915_gem_object_put(obj); 2171 1.1 riastrad intel_context_put(ce); 2172 1.1 riastrad 2173 1.1 riastrad rq->client_link.next = &(*prev)->client_link; 2174 1.1 riastrad *prev = rq; 2175 1.1 riastrad return 0; 2176 1.1 riastrad 2177 1.1 riastrad err_rq: 2178 1.1 riastrad i915_request_put(rq); 2179 1.1 riastrad err_obj: 2180 1.1 riastrad i915_gem_object_put(obj); 2181 1.1 riastrad err_ce: 2182 1.1 riastrad intel_context_put(ce); 2183 1.1 riastrad return err; 2184 1.1 riastrad } 2185 1.1 riastrad 2186 1.1 riastrad static int live_preempt_gang(void *arg) 2187 1.1 riastrad { 2188 1.1 riastrad struct intel_gt *gt = arg; 2189 1.1 riastrad struct intel_engine_cs *engine; 2190 1.1 riastrad enum intel_engine_id id; 2191 1.1 riastrad 2192 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 2193 1.1 riastrad return 0; 2194 1.1 riastrad 2195 1.1 riastrad /* 2196 1.1 riastrad * Build as long a chain of preempters as we can, with each 2197 1.1 riastrad * request higher priority than the last. Once we are ready, we release 2198 1.1 riastrad * the last batch which then precolates down the chain, each releasing 2199 1.1 riastrad * the next oldest in turn. The intent is to simply push as hard as we 2200 1.1 riastrad * can with the number of preemptions, trying to exceed narrow HW 2201 1.1 riastrad * limits. At a minimum, we insist that we can sort all the user 2202 1.1 riastrad * high priority levels into execution order. 2203 1.1 riastrad */ 2204 1.1 riastrad 2205 1.1 riastrad for_each_engine(engine, gt, id) { 2206 1.1 riastrad struct i915_request *rq = NULL; 2207 1.1 riastrad struct igt_live_test t; 2208 1.1 riastrad IGT_TIMEOUT(end_time); 2209 1.1 riastrad int prio = 0; 2210 1.1 riastrad int err = 0; 2211 1.1 riastrad u32 *cs; 2212 1.1 riastrad 2213 1.1 riastrad if (!intel_engine_has_preemption(engine)) 2214 1.1 riastrad continue; 2215 1.1 riastrad 2216 1.1 riastrad if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) 2217 1.1 riastrad return -EIO; 2218 1.1 riastrad 2219 1.1 riastrad do { 2220 1.1 riastrad struct i915_sched_attr attr = { 2221 1.1 riastrad .priority = I915_USER_PRIORITY(prio++), 2222 1.1 riastrad }; 2223 1.1 riastrad 2224 1.1 riastrad err = create_gang(engine, &rq); 2225 1.1 riastrad if (err) 2226 1.1 riastrad break; 2227 1.1 riastrad 2228 1.1 riastrad /* Submit each spinner at increasing priority */ 2229 1.1 riastrad engine->schedule(rq, &attr); 2230 1.1 riastrad 2231 1.1 riastrad if (prio <= I915_PRIORITY_MAX) 2232 1.1 riastrad continue; 2233 1.1 riastrad 2234 1.1 riastrad if (prio > (INT_MAX >> I915_USER_PRIORITY_SHIFT)) 2235 1.1 riastrad break; 2236 1.1 riastrad 2237 1.1 riastrad if (__igt_timeout(end_time, NULL)) 2238 1.1 riastrad break; 2239 1.1 riastrad } while (1); 2240 1.1 riastrad pr_debug("%s: Preempt chain of %d requests\n", 2241 1.1 riastrad engine->name, prio); 2242 1.1 riastrad 2243 1.1 riastrad /* 2244 1.1 riastrad * Such that the last spinner is the highest priority and 2245 1.1 riastrad * should execute first. When that spinner completes, 2246 1.1 riastrad * it will terminate the next lowest spinner until there 2247 1.1 riastrad * are no more spinners and the gang is complete. 2248 1.1 riastrad */ 2249 1.1 riastrad cs = i915_gem_object_pin_map(rq->batch->obj, I915_MAP_WC); 2250 1.1 riastrad if (!IS_ERR(cs)) { 2251 1.1 riastrad *cs = 0; 2252 1.1 riastrad i915_gem_object_unpin_map(rq->batch->obj); 2253 1.1 riastrad } else { 2254 1.1 riastrad err = PTR_ERR(cs); 2255 1.1 riastrad intel_gt_set_wedged(gt); 2256 1.1 riastrad } 2257 1.1 riastrad 2258 1.1 riastrad while (rq) { /* wait for each rq from highest to lowest prio */ 2259 1.1 riastrad struct i915_request *n = 2260 1.1 riastrad list_next_entry(rq, client_link); 2261 1.1 riastrad 2262 1.1 riastrad if (err == 0 && i915_request_wait(rq, 0, HZ / 5) < 0) { 2263 1.1 riastrad struct drm_printer p = 2264 1.1 riastrad drm_info_printer(engine->i915->drm.dev); 2265 1.1 riastrad 2266 1.1 riastrad pr_err("Failed to flush chain of %d requests, at %d\n", 2267 1.1 riastrad prio, rq_prio(rq) >> I915_USER_PRIORITY_SHIFT); 2268 1.1 riastrad intel_engine_dump(engine, &p, 2269 1.1 riastrad "%s\n", engine->name); 2270 1.1 riastrad 2271 1.1 riastrad err = -ETIME; 2272 1.1 riastrad } 2273 1.1 riastrad 2274 1.1 riastrad i915_request_put(rq); 2275 1.1 riastrad rq = n; 2276 1.1 riastrad } 2277 1.1 riastrad 2278 1.1 riastrad if (igt_live_test_end(&t)) 2279 1.1 riastrad err = -EIO; 2280 1.1 riastrad if (err) 2281 1.1 riastrad return err; 2282 1.1 riastrad } 2283 1.1 riastrad 2284 1.1 riastrad return 0; 2285 1.1 riastrad } 2286 1.1 riastrad 2287 1.1 riastrad static int live_preempt_hang(void *arg) 2288 1.1 riastrad { 2289 1.1 riastrad struct intel_gt *gt = arg; 2290 1.1 riastrad struct i915_gem_context *ctx_hi, *ctx_lo; 2291 1.1 riastrad struct igt_spinner spin_hi, spin_lo; 2292 1.1 riastrad struct intel_engine_cs *engine; 2293 1.1 riastrad enum intel_engine_id id; 2294 1.1 riastrad int err = -ENOMEM; 2295 1.1 riastrad 2296 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 2297 1.1 riastrad return 0; 2298 1.1 riastrad 2299 1.1 riastrad if (!intel_has_reset_engine(gt)) 2300 1.1 riastrad return 0; 2301 1.1 riastrad 2302 1.1 riastrad if (igt_spinner_init(&spin_hi, gt)) 2303 1.1 riastrad return -ENOMEM; 2304 1.1 riastrad 2305 1.1 riastrad if (igt_spinner_init(&spin_lo, gt)) 2306 1.1 riastrad goto err_spin_hi; 2307 1.1 riastrad 2308 1.1 riastrad ctx_hi = kernel_context(gt->i915); 2309 1.1 riastrad if (!ctx_hi) 2310 1.1 riastrad goto err_spin_lo; 2311 1.1 riastrad ctx_hi->sched.priority = 2312 1.1 riastrad I915_USER_PRIORITY(I915_CONTEXT_MAX_USER_PRIORITY); 2313 1.1 riastrad 2314 1.1 riastrad ctx_lo = kernel_context(gt->i915); 2315 1.1 riastrad if (!ctx_lo) 2316 1.1 riastrad goto err_ctx_hi; 2317 1.1 riastrad ctx_lo->sched.priority = 2318 1.1 riastrad I915_USER_PRIORITY(I915_CONTEXT_MIN_USER_PRIORITY); 2319 1.1 riastrad 2320 1.1 riastrad for_each_engine(engine, gt, id) { 2321 1.1 riastrad struct i915_request *rq; 2322 1.1 riastrad 2323 1.1 riastrad if (!intel_engine_has_preemption(engine)) 2324 1.1 riastrad continue; 2325 1.1 riastrad 2326 1.1 riastrad rq = spinner_create_request(&spin_lo, ctx_lo, engine, 2327 1.1 riastrad MI_ARB_CHECK); 2328 1.1 riastrad if (IS_ERR(rq)) { 2329 1.1 riastrad err = PTR_ERR(rq); 2330 1.1 riastrad goto err_ctx_lo; 2331 1.1 riastrad } 2332 1.1 riastrad 2333 1.1 riastrad i915_request_add(rq); 2334 1.1 riastrad if (!igt_wait_for_spinner(&spin_lo, rq)) { 2335 1.1 riastrad GEM_TRACE("lo spinner failed to start\n"); 2336 1.1 riastrad GEM_TRACE_DUMP(); 2337 1.1 riastrad intel_gt_set_wedged(gt); 2338 1.1 riastrad err = -EIO; 2339 1.1 riastrad goto err_ctx_lo; 2340 1.1 riastrad } 2341 1.1 riastrad 2342 1.1 riastrad rq = spinner_create_request(&spin_hi, ctx_hi, engine, 2343 1.1 riastrad MI_ARB_CHECK); 2344 1.1 riastrad if (IS_ERR(rq)) { 2345 1.1 riastrad igt_spinner_end(&spin_lo); 2346 1.1 riastrad err = PTR_ERR(rq); 2347 1.1 riastrad goto err_ctx_lo; 2348 1.1 riastrad } 2349 1.1 riastrad 2350 1.1 riastrad init_completion(&engine->execlists.preempt_hang.completion); 2351 1.1 riastrad engine->execlists.preempt_hang.inject_hang = true; 2352 1.1 riastrad 2353 1.1 riastrad i915_request_add(rq); 2354 1.1 riastrad 2355 1.1 riastrad if (!wait_for_completion_timeout(&engine->execlists.preempt_hang.completion, 2356 1.1 riastrad HZ / 10)) { 2357 1.1 riastrad pr_err("Preemption did not occur within timeout!"); 2358 1.1 riastrad GEM_TRACE_DUMP(); 2359 1.1 riastrad intel_gt_set_wedged(gt); 2360 1.1 riastrad err = -EIO; 2361 1.1 riastrad goto err_ctx_lo; 2362 1.1 riastrad } 2363 1.1 riastrad 2364 1.1 riastrad set_bit(I915_RESET_ENGINE + id, >->reset.flags); 2365 1.1 riastrad intel_engine_reset(engine, NULL); 2366 1.1 riastrad clear_bit(I915_RESET_ENGINE + id, >->reset.flags); 2367 1.1 riastrad 2368 1.1 riastrad engine->execlists.preempt_hang.inject_hang = false; 2369 1.1 riastrad 2370 1.1 riastrad if (!igt_wait_for_spinner(&spin_hi, rq)) { 2371 1.1 riastrad GEM_TRACE("hi spinner failed to start\n"); 2372 1.1 riastrad GEM_TRACE_DUMP(); 2373 1.1 riastrad intel_gt_set_wedged(gt); 2374 1.1 riastrad err = -EIO; 2375 1.1 riastrad goto err_ctx_lo; 2376 1.1 riastrad } 2377 1.1 riastrad 2378 1.1 riastrad igt_spinner_end(&spin_hi); 2379 1.1 riastrad igt_spinner_end(&spin_lo); 2380 1.1 riastrad if (igt_flush_test(gt->i915)) { 2381 1.1 riastrad err = -EIO; 2382 1.1 riastrad goto err_ctx_lo; 2383 1.1 riastrad } 2384 1.1 riastrad } 2385 1.1 riastrad 2386 1.1 riastrad err = 0; 2387 1.1 riastrad err_ctx_lo: 2388 1.1 riastrad kernel_context_close(ctx_lo); 2389 1.1 riastrad err_ctx_hi: 2390 1.1 riastrad kernel_context_close(ctx_hi); 2391 1.1 riastrad err_spin_lo: 2392 1.1 riastrad igt_spinner_fini(&spin_lo); 2393 1.1 riastrad err_spin_hi: 2394 1.1 riastrad igt_spinner_fini(&spin_hi); 2395 1.1 riastrad return err; 2396 1.1 riastrad } 2397 1.1 riastrad 2398 1.1 riastrad static int live_preempt_timeout(void *arg) 2399 1.1 riastrad { 2400 1.1 riastrad struct intel_gt *gt = arg; 2401 1.1 riastrad struct i915_gem_context *ctx_hi, *ctx_lo; 2402 1.1 riastrad struct igt_spinner spin_lo; 2403 1.1 riastrad struct intel_engine_cs *engine; 2404 1.1 riastrad enum intel_engine_id id; 2405 1.1 riastrad int err = -ENOMEM; 2406 1.1 riastrad 2407 1.1 riastrad /* 2408 1.1 riastrad * Check that we force preemption to occur by cancelling the previous 2409 1.1 riastrad * context if it refuses to yield the GPU. 2410 1.1 riastrad */ 2411 1.1 riastrad if (!IS_ACTIVE(CONFIG_DRM_I915_PREEMPT_TIMEOUT)) 2412 1.1 riastrad return 0; 2413 1.1 riastrad 2414 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(gt->i915)) 2415 1.1 riastrad return 0; 2416 1.1 riastrad 2417 1.1 riastrad if (!intel_has_reset_engine(gt)) 2418 1.1 riastrad return 0; 2419 1.1 riastrad 2420 1.1 riastrad if (igt_spinner_init(&spin_lo, gt)) 2421 1.1 riastrad return -ENOMEM; 2422 1.1 riastrad 2423 1.1 riastrad ctx_hi = kernel_context(gt->i915); 2424 1.1 riastrad if (!ctx_hi) 2425 1.1 riastrad goto err_spin_lo; 2426 1.1 riastrad ctx_hi->sched.priority = 2427 1.1 riastrad I915_USER_PRIORITY(I915_CONTEXT_MAX_USER_PRIORITY); 2428 1.1 riastrad 2429 1.1 riastrad ctx_lo = kernel_context(gt->i915); 2430 1.1 riastrad if (!ctx_lo) 2431 1.1 riastrad goto err_ctx_hi; 2432 1.1 riastrad ctx_lo->sched.priority = 2433 1.1 riastrad I915_USER_PRIORITY(I915_CONTEXT_MIN_USER_PRIORITY); 2434 1.1 riastrad 2435 1.1 riastrad for_each_engine(engine, gt, id) { 2436 1.1 riastrad unsigned long saved_timeout; 2437 1.1 riastrad struct i915_request *rq; 2438 1.1 riastrad 2439 1.1 riastrad if (!intel_engine_has_preemption(engine)) 2440 1.1 riastrad continue; 2441 1.1 riastrad 2442 1.1 riastrad rq = spinner_create_request(&spin_lo, ctx_lo, engine, 2443 1.1 riastrad MI_NOOP); /* preemption disabled */ 2444 1.1 riastrad if (IS_ERR(rq)) { 2445 1.1 riastrad err = PTR_ERR(rq); 2446 1.1 riastrad goto err_ctx_lo; 2447 1.1 riastrad } 2448 1.1 riastrad 2449 1.1 riastrad i915_request_add(rq); 2450 1.1 riastrad if (!igt_wait_for_spinner(&spin_lo, rq)) { 2451 1.1 riastrad intel_gt_set_wedged(gt); 2452 1.1 riastrad err = -EIO; 2453 1.1 riastrad goto err_ctx_lo; 2454 1.1 riastrad } 2455 1.1 riastrad 2456 1.1 riastrad rq = igt_request_alloc(ctx_hi, engine); 2457 1.1 riastrad if (IS_ERR(rq)) { 2458 1.1 riastrad igt_spinner_end(&spin_lo); 2459 1.1 riastrad err = PTR_ERR(rq); 2460 1.1 riastrad goto err_ctx_lo; 2461 1.1 riastrad } 2462 1.1 riastrad 2463 1.1 riastrad /* Flush the previous CS ack before changing timeouts */ 2464 1.1 riastrad while (READ_ONCE(engine->execlists.pending[0])) 2465 1.1 riastrad cpu_relax(); 2466 1.1 riastrad 2467 1.1 riastrad saved_timeout = engine->props.preempt_timeout_ms; 2468 1.1 riastrad engine->props.preempt_timeout_ms = 1; /* in ms, -> 1 jiffie */ 2469 1.1 riastrad 2470 1.1 riastrad i915_request_get(rq); 2471 1.1 riastrad i915_request_add(rq); 2472 1.1 riastrad 2473 1.1 riastrad intel_engine_flush_submission(engine); 2474 1.1 riastrad engine->props.preempt_timeout_ms = saved_timeout; 2475 1.1 riastrad 2476 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 10) < 0) { 2477 1.1 riastrad intel_gt_set_wedged(gt); 2478 1.1 riastrad i915_request_put(rq); 2479 1.1 riastrad err = -ETIME; 2480 1.1 riastrad goto err_ctx_lo; 2481 1.1 riastrad } 2482 1.1 riastrad 2483 1.1 riastrad igt_spinner_end(&spin_lo); 2484 1.1 riastrad i915_request_put(rq); 2485 1.1 riastrad } 2486 1.1 riastrad 2487 1.1 riastrad err = 0; 2488 1.1 riastrad err_ctx_lo: 2489 1.1 riastrad kernel_context_close(ctx_lo); 2490 1.1 riastrad err_ctx_hi: 2491 1.1 riastrad kernel_context_close(ctx_hi); 2492 1.1 riastrad err_spin_lo: 2493 1.1 riastrad igt_spinner_fini(&spin_lo); 2494 1.1 riastrad return err; 2495 1.1 riastrad } 2496 1.1 riastrad 2497 1.1 riastrad static int random_range(struct rnd_state *rnd, int min, int max) 2498 1.1 riastrad { 2499 1.1 riastrad return i915_prandom_u32_max_state(max - min, rnd) + min; 2500 1.1 riastrad } 2501 1.1 riastrad 2502 1.1 riastrad static int random_priority(struct rnd_state *rnd) 2503 1.1 riastrad { 2504 1.1 riastrad return random_range(rnd, I915_PRIORITY_MIN, I915_PRIORITY_MAX); 2505 1.1 riastrad } 2506 1.1 riastrad 2507 1.1 riastrad struct preempt_smoke { 2508 1.1 riastrad struct intel_gt *gt; 2509 1.1 riastrad struct i915_gem_context **contexts; 2510 1.1 riastrad struct intel_engine_cs *engine; 2511 1.1 riastrad struct drm_i915_gem_object *batch; 2512 1.1 riastrad unsigned int ncontext; 2513 1.1 riastrad struct rnd_state prng; 2514 1.1 riastrad unsigned long count; 2515 1.1 riastrad }; 2516 1.1 riastrad 2517 1.1 riastrad static struct i915_gem_context *smoke_context(struct preempt_smoke *smoke) 2518 1.1 riastrad { 2519 1.1 riastrad return smoke->contexts[i915_prandom_u32_max_state(smoke->ncontext, 2520 1.1 riastrad &smoke->prng)]; 2521 1.1 riastrad } 2522 1.1 riastrad 2523 1.1 riastrad static int smoke_submit(struct preempt_smoke *smoke, 2524 1.1 riastrad struct i915_gem_context *ctx, int prio, 2525 1.1 riastrad struct drm_i915_gem_object *batch) 2526 1.1 riastrad { 2527 1.1 riastrad struct i915_request *rq; 2528 1.1 riastrad struct i915_vma *vma = NULL; 2529 1.1 riastrad int err = 0; 2530 1.1 riastrad 2531 1.1 riastrad if (batch) { 2532 1.1 riastrad struct i915_address_space *vm; 2533 1.1 riastrad 2534 1.1 riastrad vm = i915_gem_context_get_vm_rcu(ctx); 2535 1.1 riastrad vma = i915_vma_instance(batch, vm, NULL); 2536 1.1 riastrad i915_vm_put(vm); 2537 1.1 riastrad if (IS_ERR(vma)) 2538 1.1 riastrad return PTR_ERR(vma); 2539 1.1 riastrad 2540 1.1 riastrad err = i915_vma_pin(vma, 0, 0, PIN_USER); 2541 1.1 riastrad if (err) 2542 1.1 riastrad return err; 2543 1.1 riastrad } 2544 1.1 riastrad 2545 1.1 riastrad ctx->sched.priority = prio; 2546 1.1 riastrad 2547 1.1 riastrad rq = igt_request_alloc(ctx, smoke->engine); 2548 1.1 riastrad if (IS_ERR(rq)) { 2549 1.1 riastrad err = PTR_ERR(rq); 2550 1.1 riastrad goto unpin; 2551 1.1 riastrad } 2552 1.1 riastrad 2553 1.1 riastrad if (vma) { 2554 1.1 riastrad i915_vma_lock(vma); 2555 1.1 riastrad err = i915_request_await_object(rq, vma->obj, false); 2556 1.1 riastrad if (!err) 2557 1.1 riastrad err = i915_vma_move_to_active(vma, rq, 0); 2558 1.1 riastrad if (!err) 2559 1.1 riastrad err = rq->engine->emit_bb_start(rq, 2560 1.1 riastrad vma->node.start, 2561 1.1 riastrad PAGE_SIZE, 0); 2562 1.1 riastrad i915_vma_unlock(vma); 2563 1.1 riastrad } 2564 1.1 riastrad 2565 1.1 riastrad i915_request_add(rq); 2566 1.1 riastrad 2567 1.1 riastrad unpin: 2568 1.1 riastrad if (vma) 2569 1.1 riastrad i915_vma_unpin(vma); 2570 1.1 riastrad 2571 1.1 riastrad return err; 2572 1.1 riastrad } 2573 1.1 riastrad 2574 1.1 riastrad static int smoke_crescendo_thread(void *arg) 2575 1.1 riastrad { 2576 1.1 riastrad struct preempt_smoke *smoke = arg; 2577 1.1 riastrad IGT_TIMEOUT(end_time); 2578 1.1 riastrad unsigned long count; 2579 1.1 riastrad 2580 1.1 riastrad count = 0; 2581 1.1 riastrad do { 2582 1.1 riastrad struct i915_gem_context *ctx = smoke_context(smoke); 2583 1.1 riastrad int err; 2584 1.1 riastrad 2585 1.1 riastrad err = smoke_submit(smoke, 2586 1.1 riastrad ctx, count % I915_PRIORITY_MAX, 2587 1.1 riastrad smoke->batch); 2588 1.1 riastrad if (err) 2589 1.1 riastrad return err; 2590 1.1 riastrad 2591 1.1 riastrad count++; 2592 1.1 riastrad } while (!__igt_timeout(end_time, NULL)); 2593 1.1 riastrad 2594 1.1 riastrad smoke->count = count; 2595 1.1 riastrad return 0; 2596 1.1 riastrad } 2597 1.1 riastrad 2598 1.1 riastrad static int smoke_crescendo(struct preempt_smoke *smoke, unsigned int flags) 2599 1.1 riastrad #define BATCH BIT(0) 2600 1.1 riastrad { 2601 1.1 riastrad struct task_struct *tsk[I915_NUM_ENGINES] = {}; 2602 1.1 riastrad struct preempt_smoke arg[I915_NUM_ENGINES]; 2603 1.1 riastrad struct intel_engine_cs *engine; 2604 1.1 riastrad enum intel_engine_id id; 2605 1.1 riastrad unsigned long count; 2606 1.1 riastrad int err = 0; 2607 1.1 riastrad 2608 1.1 riastrad for_each_engine(engine, smoke->gt, id) { 2609 1.1 riastrad arg[id] = *smoke; 2610 1.1 riastrad arg[id].engine = engine; 2611 1.1 riastrad if (!(flags & BATCH)) 2612 1.1 riastrad arg[id].batch = NULL; 2613 1.1 riastrad arg[id].count = 0; 2614 1.1 riastrad 2615 1.1 riastrad tsk[id] = kthread_run(smoke_crescendo_thread, &arg, 2616 1.1 riastrad "igt/smoke:%d", id); 2617 1.1 riastrad if (IS_ERR(tsk[id])) { 2618 1.1 riastrad err = PTR_ERR(tsk[id]); 2619 1.1 riastrad break; 2620 1.1 riastrad } 2621 1.1 riastrad get_task_struct(tsk[id]); 2622 1.1 riastrad } 2623 1.1 riastrad 2624 1.1 riastrad yield(); /* start all threads before we kthread_stop() */ 2625 1.1 riastrad 2626 1.1 riastrad count = 0; 2627 1.1 riastrad for_each_engine(engine, smoke->gt, id) { 2628 1.1 riastrad int status; 2629 1.1 riastrad 2630 1.1 riastrad if (IS_ERR_OR_NULL(tsk[id])) 2631 1.1 riastrad continue; 2632 1.1 riastrad 2633 1.1 riastrad status = kthread_stop(tsk[id]); 2634 1.1 riastrad if (status && !err) 2635 1.1 riastrad err = status; 2636 1.1 riastrad 2637 1.1 riastrad count += arg[id].count; 2638 1.1 riastrad 2639 1.1 riastrad put_task_struct(tsk[id]); 2640 1.1 riastrad } 2641 1.1 riastrad 2642 1.1 riastrad pr_info("Submitted %lu crescendo:%x requests across %d engines and %d contexts\n", 2643 1.1 riastrad count, flags, 2644 1.1 riastrad RUNTIME_INFO(smoke->gt->i915)->num_engines, smoke->ncontext); 2645 1.1 riastrad return 0; 2646 1.1 riastrad } 2647 1.1 riastrad 2648 1.1 riastrad static int smoke_random(struct preempt_smoke *smoke, unsigned int flags) 2649 1.1 riastrad { 2650 1.1 riastrad enum intel_engine_id id; 2651 1.1 riastrad IGT_TIMEOUT(end_time); 2652 1.1 riastrad unsigned long count; 2653 1.1 riastrad 2654 1.1 riastrad count = 0; 2655 1.1 riastrad do { 2656 1.1 riastrad for_each_engine(smoke->engine, smoke->gt, id) { 2657 1.1 riastrad struct i915_gem_context *ctx = smoke_context(smoke); 2658 1.1 riastrad int err; 2659 1.1 riastrad 2660 1.1 riastrad err = smoke_submit(smoke, 2661 1.1 riastrad ctx, random_priority(&smoke->prng), 2662 1.1 riastrad flags & BATCH ? smoke->batch : NULL); 2663 1.1 riastrad if (err) 2664 1.1 riastrad return err; 2665 1.1 riastrad 2666 1.1 riastrad count++; 2667 1.1 riastrad } 2668 1.1 riastrad } while (!__igt_timeout(end_time, NULL)); 2669 1.1 riastrad 2670 1.1 riastrad pr_info("Submitted %lu random:%x requests across %d engines and %d contexts\n", 2671 1.1 riastrad count, flags, 2672 1.1 riastrad RUNTIME_INFO(smoke->gt->i915)->num_engines, smoke->ncontext); 2673 1.1 riastrad return 0; 2674 1.1 riastrad } 2675 1.1 riastrad 2676 1.1 riastrad static int live_preempt_smoke(void *arg) 2677 1.1 riastrad { 2678 1.1 riastrad struct preempt_smoke smoke = { 2679 1.1 riastrad .gt = arg, 2680 1.1 riastrad .prng = I915_RND_STATE_INITIALIZER(i915_selftest.random_seed), 2681 1.1 riastrad .ncontext = 1024, 2682 1.1 riastrad }; 2683 1.1 riastrad const unsigned int phase[] = { 0, BATCH }; 2684 1.1 riastrad struct igt_live_test t; 2685 1.1 riastrad int err = -ENOMEM; 2686 1.1 riastrad u32 *cs; 2687 1.1 riastrad int n; 2688 1.1 riastrad 2689 1.1 riastrad if (!HAS_LOGICAL_RING_PREEMPTION(smoke.gt->i915)) 2690 1.1 riastrad return 0; 2691 1.1 riastrad 2692 1.1 riastrad smoke.contexts = kmalloc_array(smoke.ncontext, 2693 1.1 riastrad sizeof(*smoke.contexts), 2694 1.1 riastrad GFP_KERNEL); 2695 1.1 riastrad if (!smoke.contexts) 2696 1.1 riastrad return -ENOMEM; 2697 1.1 riastrad 2698 1.1 riastrad smoke.batch = 2699 1.1 riastrad i915_gem_object_create_internal(smoke.gt->i915, PAGE_SIZE); 2700 1.1 riastrad if (IS_ERR(smoke.batch)) { 2701 1.1 riastrad err = PTR_ERR(smoke.batch); 2702 1.1 riastrad goto err_free; 2703 1.1 riastrad } 2704 1.1 riastrad 2705 1.1 riastrad cs = i915_gem_object_pin_map(smoke.batch, I915_MAP_WB); 2706 1.1 riastrad if (IS_ERR(cs)) { 2707 1.1 riastrad err = PTR_ERR(cs); 2708 1.1 riastrad goto err_batch; 2709 1.1 riastrad } 2710 1.1 riastrad for (n = 0; n < PAGE_SIZE / sizeof(*cs) - 1; n++) 2711 1.1 riastrad cs[n] = MI_ARB_CHECK; 2712 1.1 riastrad cs[n] = MI_BATCH_BUFFER_END; 2713 1.1 riastrad i915_gem_object_flush_map(smoke.batch); 2714 1.1 riastrad i915_gem_object_unpin_map(smoke.batch); 2715 1.1 riastrad 2716 1.1 riastrad if (igt_live_test_begin(&t, smoke.gt->i915, __func__, "all")) { 2717 1.1 riastrad err = -EIO; 2718 1.1 riastrad goto err_batch; 2719 1.1 riastrad } 2720 1.1 riastrad 2721 1.1 riastrad for (n = 0; n < smoke.ncontext; n++) { 2722 1.1 riastrad smoke.contexts[n] = kernel_context(smoke.gt->i915); 2723 1.1 riastrad if (!smoke.contexts[n]) 2724 1.1 riastrad goto err_ctx; 2725 1.1 riastrad } 2726 1.1 riastrad 2727 1.1 riastrad for (n = 0; n < ARRAY_SIZE(phase); n++) { 2728 1.1 riastrad err = smoke_crescendo(&smoke, phase[n]); 2729 1.1 riastrad if (err) 2730 1.1 riastrad goto err_ctx; 2731 1.1 riastrad 2732 1.1 riastrad err = smoke_random(&smoke, phase[n]); 2733 1.1 riastrad if (err) 2734 1.1 riastrad goto err_ctx; 2735 1.1 riastrad } 2736 1.1 riastrad 2737 1.1 riastrad err_ctx: 2738 1.1 riastrad if (igt_live_test_end(&t)) 2739 1.1 riastrad err = -EIO; 2740 1.1 riastrad 2741 1.1 riastrad for (n = 0; n < smoke.ncontext; n++) { 2742 1.1 riastrad if (!smoke.contexts[n]) 2743 1.1 riastrad break; 2744 1.1 riastrad kernel_context_close(smoke.contexts[n]); 2745 1.1 riastrad } 2746 1.1 riastrad 2747 1.1 riastrad err_batch: 2748 1.1 riastrad i915_gem_object_put(smoke.batch); 2749 1.1 riastrad err_free: 2750 1.1 riastrad kfree(smoke.contexts); 2751 1.1 riastrad 2752 1.1 riastrad return err; 2753 1.1 riastrad } 2754 1.1 riastrad 2755 1.1 riastrad static int nop_virtual_engine(struct intel_gt *gt, 2756 1.1 riastrad struct intel_engine_cs **siblings, 2757 1.1 riastrad unsigned int nsibling, 2758 1.1 riastrad unsigned int nctx, 2759 1.1 riastrad unsigned int flags) 2760 1.1 riastrad #define CHAIN BIT(0) 2761 1.1 riastrad { 2762 1.1 riastrad IGT_TIMEOUT(end_time); 2763 1.1 riastrad struct i915_request *request[16] = {}; 2764 1.1 riastrad struct intel_context *ve[16]; 2765 1.1 riastrad unsigned long n, prime, nc; 2766 1.1 riastrad struct igt_live_test t; 2767 1.1 riastrad ktime_t times[2] = {}; 2768 1.1 riastrad int err; 2769 1.1 riastrad 2770 1.1 riastrad GEM_BUG_ON(!nctx || nctx > ARRAY_SIZE(ve)); 2771 1.1 riastrad 2772 1.1 riastrad for (n = 0; n < nctx; n++) { 2773 1.1 riastrad ve[n] = intel_execlists_create_virtual(siblings, nsibling); 2774 1.1 riastrad if (IS_ERR(ve[n])) { 2775 1.1 riastrad err = PTR_ERR(ve[n]); 2776 1.1 riastrad nctx = n; 2777 1.1 riastrad goto out; 2778 1.1 riastrad } 2779 1.1 riastrad 2780 1.1 riastrad err = intel_context_pin(ve[n]); 2781 1.1 riastrad if (err) { 2782 1.1 riastrad intel_context_put(ve[n]); 2783 1.1 riastrad nctx = n; 2784 1.1 riastrad goto out; 2785 1.1 riastrad } 2786 1.1 riastrad } 2787 1.1 riastrad 2788 1.1 riastrad err = igt_live_test_begin(&t, gt->i915, __func__, ve[0]->engine->name); 2789 1.1 riastrad if (err) 2790 1.1 riastrad goto out; 2791 1.1 riastrad 2792 1.1 riastrad for_each_prime_number_from(prime, 1, 8192) { 2793 1.1 riastrad times[1] = ktime_get_raw(); 2794 1.1 riastrad 2795 1.1 riastrad if (flags & CHAIN) { 2796 1.1 riastrad for (nc = 0; nc < nctx; nc++) { 2797 1.1 riastrad for (n = 0; n < prime; n++) { 2798 1.1 riastrad struct i915_request *rq; 2799 1.1 riastrad 2800 1.1 riastrad rq = i915_request_create(ve[nc]); 2801 1.1 riastrad if (IS_ERR(rq)) { 2802 1.1 riastrad err = PTR_ERR(rq); 2803 1.1 riastrad goto out; 2804 1.1 riastrad } 2805 1.1 riastrad 2806 1.1 riastrad if (request[nc]) 2807 1.1 riastrad i915_request_put(request[nc]); 2808 1.1 riastrad request[nc] = i915_request_get(rq); 2809 1.1 riastrad i915_request_add(rq); 2810 1.1 riastrad } 2811 1.1 riastrad } 2812 1.1 riastrad } else { 2813 1.1 riastrad for (n = 0; n < prime; n++) { 2814 1.1 riastrad for (nc = 0; nc < nctx; nc++) { 2815 1.1 riastrad struct i915_request *rq; 2816 1.1 riastrad 2817 1.1 riastrad rq = i915_request_create(ve[nc]); 2818 1.1 riastrad if (IS_ERR(rq)) { 2819 1.1 riastrad err = PTR_ERR(rq); 2820 1.1 riastrad goto out; 2821 1.1 riastrad } 2822 1.1 riastrad 2823 1.1 riastrad if (request[nc]) 2824 1.1 riastrad i915_request_put(request[nc]); 2825 1.1 riastrad request[nc] = i915_request_get(rq); 2826 1.1 riastrad i915_request_add(rq); 2827 1.1 riastrad } 2828 1.1 riastrad } 2829 1.1 riastrad } 2830 1.1 riastrad 2831 1.1 riastrad for (nc = 0; nc < nctx; nc++) { 2832 1.1 riastrad if (i915_request_wait(request[nc], 0, HZ / 10) < 0) { 2833 1.1 riastrad pr_err("%s(%s): wait for %llx:%lld timed out\n", 2834 1.1 riastrad __func__, ve[0]->engine->name, 2835 1.1 riastrad request[nc]->fence.context, 2836 1.1 riastrad request[nc]->fence.seqno); 2837 1.1 riastrad 2838 1.1 riastrad GEM_TRACE("%s(%s) failed at request %llx:%lld\n", 2839 1.1 riastrad __func__, ve[0]->engine->name, 2840 1.1 riastrad request[nc]->fence.context, 2841 1.1 riastrad request[nc]->fence.seqno); 2842 1.1 riastrad GEM_TRACE_DUMP(); 2843 1.1 riastrad intel_gt_set_wedged(gt); 2844 1.1 riastrad break; 2845 1.1 riastrad } 2846 1.1 riastrad } 2847 1.1 riastrad 2848 1.1 riastrad times[1] = ktime_sub(ktime_get_raw(), times[1]); 2849 1.1 riastrad if (prime == 1) 2850 1.1 riastrad times[0] = times[1]; 2851 1.1 riastrad 2852 1.1 riastrad for (nc = 0; nc < nctx; nc++) { 2853 1.1 riastrad i915_request_put(request[nc]); 2854 1.1 riastrad request[nc] = NULL; 2855 1.1 riastrad } 2856 1.1 riastrad 2857 1.1 riastrad if (__igt_timeout(end_time, NULL)) 2858 1.1 riastrad break; 2859 1.1 riastrad } 2860 1.1 riastrad 2861 1.1 riastrad err = igt_live_test_end(&t); 2862 1.1 riastrad if (err) 2863 1.1 riastrad goto out; 2864 1.1 riastrad 2865 1.1 riastrad pr_info("Requestx%d latencies on %s: 1 = %lluns, %lu = %lluns\n", 2866 1.1 riastrad nctx, ve[0]->engine->name, ktime_to_ns(times[0]), 2867 1.1 riastrad prime, div64_u64(ktime_to_ns(times[1]), prime)); 2868 1.1 riastrad 2869 1.1 riastrad out: 2870 1.1 riastrad if (igt_flush_test(gt->i915)) 2871 1.1 riastrad err = -EIO; 2872 1.1 riastrad 2873 1.1 riastrad for (nc = 0; nc < nctx; nc++) { 2874 1.1 riastrad i915_request_put(request[nc]); 2875 1.1 riastrad intel_context_unpin(ve[nc]); 2876 1.1 riastrad intel_context_put(ve[nc]); 2877 1.1 riastrad } 2878 1.1 riastrad return err; 2879 1.1 riastrad } 2880 1.1 riastrad 2881 1.1 riastrad static int live_virtual_engine(void *arg) 2882 1.1 riastrad { 2883 1.1 riastrad struct intel_gt *gt = arg; 2884 1.1 riastrad struct intel_engine_cs *siblings[MAX_ENGINE_INSTANCE + 1]; 2885 1.1 riastrad struct intel_engine_cs *engine; 2886 1.1 riastrad enum intel_engine_id id; 2887 1.1 riastrad unsigned int class, inst; 2888 1.1 riastrad int err; 2889 1.1 riastrad 2890 1.1 riastrad if (USES_GUC_SUBMISSION(gt->i915)) 2891 1.1 riastrad return 0; 2892 1.1 riastrad 2893 1.1 riastrad for_each_engine(engine, gt, id) { 2894 1.1 riastrad err = nop_virtual_engine(gt, &engine, 1, 1, 0); 2895 1.1 riastrad if (err) { 2896 1.1 riastrad pr_err("Failed to wrap engine %s: err=%d\n", 2897 1.1 riastrad engine->name, err); 2898 1.1 riastrad return err; 2899 1.1 riastrad } 2900 1.1 riastrad } 2901 1.1 riastrad 2902 1.1 riastrad for (class = 0; class <= MAX_ENGINE_CLASS; class++) { 2903 1.1 riastrad int nsibling, n; 2904 1.1 riastrad 2905 1.1 riastrad nsibling = 0; 2906 1.1 riastrad for (inst = 0; inst <= MAX_ENGINE_INSTANCE; inst++) { 2907 1.1 riastrad if (!gt->engine_class[class][inst]) 2908 1.1 riastrad continue; 2909 1.1 riastrad 2910 1.1 riastrad siblings[nsibling++] = gt->engine_class[class][inst]; 2911 1.1 riastrad } 2912 1.1 riastrad if (nsibling < 2) 2913 1.1 riastrad continue; 2914 1.1 riastrad 2915 1.1 riastrad for (n = 1; n <= nsibling + 1; n++) { 2916 1.1 riastrad err = nop_virtual_engine(gt, siblings, nsibling, 2917 1.1 riastrad n, 0); 2918 1.1 riastrad if (err) 2919 1.1 riastrad return err; 2920 1.1 riastrad } 2921 1.1 riastrad 2922 1.1 riastrad err = nop_virtual_engine(gt, siblings, nsibling, n, CHAIN); 2923 1.1 riastrad if (err) 2924 1.1 riastrad return err; 2925 1.1 riastrad } 2926 1.1 riastrad 2927 1.1 riastrad return 0; 2928 1.1 riastrad } 2929 1.1 riastrad 2930 1.1 riastrad static int mask_virtual_engine(struct intel_gt *gt, 2931 1.1 riastrad struct intel_engine_cs **siblings, 2932 1.1 riastrad unsigned int nsibling) 2933 1.1 riastrad { 2934 1.1 riastrad struct i915_request *request[MAX_ENGINE_INSTANCE + 1]; 2935 1.1 riastrad struct intel_context *ve; 2936 1.1 riastrad struct igt_live_test t; 2937 1.1 riastrad unsigned int n; 2938 1.1 riastrad int err; 2939 1.1 riastrad 2940 1.1 riastrad /* 2941 1.1 riastrad * Check that by setting the execution mask on a request, we can 2942 1.1 riastrad * restrict it to our desired engine within the virtual engine. 2943 1.1 riastrad */ 2944 1.1 riastrad 2945 1.1 riastrad ve = intel_execlists_create_virtual(siblings, nsibling); 2946 1.1 riastrad if (IS_ERR(ve)) { 2947 1.1 riastrad err = PTR_ERR(ve); 2948 1.1 riastrad goto out_close; 2949 1.1 riastrad } 2950 1.1 riastrad 2951 1.1 riastrad err = intel_context_pin(ve); 2952 1.1 riastrad if (err) 2953 1.1 riastrad goto out_put; 2954 1.1 riastrad 2955 1.1 riastrad err = igt_live_test_begin(&t, gt->i915, __func__, ve->engine->name); 2956 1.1 riastrad if (err) 2957 1.1 riastrad goto out_unpin; 2958 1.1 riastrad 2959 1.1 riastrad for (n = 0; n < nsibling; n++) { 2960 1.1 riastrad request[n] = i915_request_create(ve); 2961 1.1 riastrad if (IS_ERR(request[n])) { 2962 1.1 riastrad err = PTR_ERR(request[n]); 2963 1.1 riastrad nsibling = n; 2964 1.1 riastrad goto out; 2965 1.1 riastrad } 2966 1.1 riastrad 2967 1.1 riastrad /* Reverse order as it's more likely to be unnatural */ 2968 1.1 riastrad request[n]->execution_mask = siblings[nsibling - n - 1]->mask; 2969 1.1 riastrad 2970 1.1 riastrad i915_request_get(request[n]); 2971 1.1 riastrad i915_request_add(request[n]); 2972 1.1 riastrad } 2973 1.1 riastrad 2974 1.1 riastrad for (n = 0; n < nsibling; n++) { 2975 1.1 riastrad if (i915_request_wait(request[n], 0, HZ / 10) < 0) { 2976 1.1 riastrad pr_err("%s(%s): wait for %llx:%lld timed out\n", 2977 1.1 riastrad __func__, ve->engine->name, 2978 1.1 riastrad request[n]->fence.context, 2979 1.1 riastrad request[n]->fence.seqno); 2980 1.1 riastrad 2981 1.1 riastrad GEM_TRACE("%s(%s) failed at request %llx:%lld\n", 2982 1.1 riastrad __func__, ve->engine->name, 2983 1.1 riastrad request[n]->fence.context, 2984 1.1 riastrad request[n]->fence.seqno); 2985 1.1 riastrad GEM_TRACE_DUMP(); 2986 1.1 riastrad intel_gt_set_wedged(gt); 2987 1.1 riastrad err = -EIO; 2988 1.1 riastrad goto out; 2989 1.1 riastrad } 2990 1.1 riastrad 2991 1.1 riastrad if (request[n]->engine != siblings[nsibling - n - 1]) { 2992 1.1 riastrad pr_err("Executed on wrong sibling '%s', expected '%s'\n", 2993 1.1 riastrad request[n]->engine->name, 2994 1.1 riastrad siblings[nsibling - n - 1]->name); 2995 1.1 riastrad err = -EINVAL; 2996 1.1 riastrad goto out; 2997 1.1 riastrad } 2998 1.1 riastrad } 2999 1.1 riastrad 3000 1.1 riastrad err = igt_live_test_end(&t); 3001 1.1 riastrad out: 3002 1.1 riastrad if (igt_flush_test(gt->i915)) 3003 1.1 riastrad err = -EIO; 3004 1.1 riastrad 3005 1.1 riastrad for (n = 0; n < nsibling; n++) 3006 1.1 riastrad i915_request_put(request[n]); 3007 1.1 riastrad 3008 1.1 riastrad out_unpin: 3009 1.1 riastrad intel_context_unpin(ve); 3010 1.1 riastrad out_put: 3011 1.1 riastrad intel_context_put(ve); 3012 1.1 riastrad out_close: 3013 1.1 riastrad return err; 3014 1.1 riastrad } 3015 1.1 riastrad 3016 1.1 riastrad static int live_virtual_mask(void *arg) 3017 1.1 riastrad { 3018 1.1 riastrad struct intel_gt *gt = arg; 3019 1.1 riastrad struct intel_engine_cs *siblings[MAX_ENGINE_INSTANCE + 1]; 3020 1.1 riastrad unsigned int class, inst; 3021 1.1 riastrad int err; 3022 1.1 riastrad 3023 1.1 riastrad if (USES_GUC_SUBMISSION(gt->i915)) 3024 1.1 riastrad return 0; 3025 1.1 riastrad 3026 1.1 riastrad for (class = 0; class <= MAX_ENGINE_CLASS; class++) { 3027 1.1 riastrad unsigned int nsibling; 3028 1.1 riastrad 3029 1.1 riastrad nsibling = 0; 3030 1.1 riastrad for (inst = 0; inst <= MAX_ENGINE_INSTANCE; inst++) { 3031 1.1 riastrad if (!gt->engine_class[class][inst]) 3032 1.1 riastrad break; 3033 1.1 riastrad 3034 1.1 riastrad siblings[nsibling++] = gt->engine_class[class][inst]; 3035 1.1 riastrad } 3036 1.1 riastrad if (nsibling < 2) 3037 1.1 riastrad continue; 3038 1.1 riastrad 3039 1.1 riastrad err = mask_virtual_engine(gt, siblings, nsibling); 3040 1.1 riastrad if (err) 3041 1.1 riastrad return err; 3042 1.1 riastrad } 3043 1.1 riastrad 3044 1.1 riastrad return 0; 3045 1.1 riastrad } 3046 1.1 riastrad 3047 1.1 riastrad static int preserved_virtual_engine(struct intel_gt *gt, 3048 1.1 riastrad struct intel_engine_cs **siblings, 3049 1.1 riastrad unsigned int nsibling) 3050 1.1 riastrad { 3051 1.1 riastrad struct i915_request *last = NULL; 3052 1.1 riastrad struct intel_context *ve; 3053 1.1 riastrad struct i915_vma *scratch; 3054 1.1 riastrad struct igt_live_test t; 3055 1.1 riastrad unsigned int n; 3056 1.1 riastrad int err = 0; 3057 1.1 riastrad u32 *cs; 3058 1.1 riastrad 3059 1.1 riastrad scratch = create_scratch(siblings[0]->gt); 3060 1.1 riastrad if (IS_ERR(scratch)) 3061 1.1 riastrad return PTR_ERR(scratch); 3062 1.1 riastrad 3063 1.1 riastrad ve = intel_execlists_create_virtual(siblings, nsibling); 3064 1.1 riastrad if (IS_ERR(ve)) { 3065 1.1 riastrad err = PTR_ERR(ve); 3066 1.1 riastrad goto out_scratch; 3067 1.1 riastrad } 3068 1.1 riastrad 3069 1.1 riastrad err = intel_context_pin(ve); 3070 1.1 riastrad if (err) 3071 1.1 riastrad goto out_put; 3072 1.1 riastrad 3073 1.1 riastrad err = igt_live_test_begin(&t, gt->i915, __func__, ve->engine->name); 3074 1.1 riastrad if (err) 3075 1.1 riastrad goto out_unpin; 3076 1.1 riastrad 3077 1.1 riastrad for (n = 0; n < NUM_GPR_DW; n++) { 3078 1.1 riastrad struct intel_engine_cs *engine = siblings[n % nsibling]; 3079 1.1 riastrad struct i915_request *rq; 3080 1.1 riastrad 3081 1.1 riastrad rq = i915_request_create(ve); 3082 1.1 riastrad if (IS_ERR(rq)) { 3083 1.1 riastrad err = PTR_ERR(rq); 3084 1.1 riastrad goto out_end; 3085 1.1 riastrad } 3086 1.1 riastrad 3087 1.1 riastrad i915_request_put(last); 3088 1.1 riastrad last = i915_request_get(rq); 3089 1.1 riastrad 3090 1.1 riastrad cs = intel_ring_begin(rq, 8); 3091 1.1 riastrad if (IS_ERR(cs)) { 3092 1.1 riastrad i915_request_add(rq); 3093 1.1 riastrad err = PTR_ERR(cs); 3094 1.1 riastrad goto out_end; 3095 1.1 riastrad } 3096 1.1 riastrad 3097 1.1 riastrad *cs++ = MI_STORE_REGISTER_MEM_GEN8 | MI_USE_GGTT; 3098 1.1 riastrad *cs++ = CS_GPR(engine, n); 3099 1.1 riastrad *cs++ = i915_ggtt_offset(scratch) + n * sizeof(u32); 3100 1.1 riastrad *cs++ = 0; 3101 1.1 riastrad 3102 1.1 riastrad *cs++ = MI_LOAD_REGISTER_IMM(1); 3103 1.1 riastrad *cs++ = CS_GPR(engine, (n + 1) % NUM_GPR_DW); 3104 1.1 riastrad *cs++ = n + 1; 3105 1.1 riastrad 3106 1.1 riastrad *cs++ = MI_NOOP; 3107 1.1 riastrad intel_ring_advance(rq, cs); 3108 1.1 riastrad 3109 1.1 riastrad /* Restrict this request to run on a particular engine */ 3110 1.1 riastrad rq->execution_mask = engine->mask; 3111 1.1 riastrad i915_request_add(rq); 3112 1.1 riastrad } 3113 1.1 riastrad 3114 1.1 riastrad if (i915_request_wait(last, 0, HZ / 5) < 0) { 3115 1.1 riastrad err = -ETIME; 3116 1.1 riastrad goto out_end; 3117 1.1 riastrad } 3118 1.1 riastrad 3119 1.1 riastrad cs = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB); 3120 1.1 riastrad if (IS_ERR(cs)) { 3121 1.1 riastrad err = PTR_ERR(cs); 3122 1.1 riastrad goto out_end; 3123 1.1 riastrad } 3124 1.1 riastrad 3125 1.1 riastrad for (n = 0; n < NUM_GPR_DW; n++) { 3126 1.1 riastrad if (cs[n] != n) { 3127 1.1 riastrad pr_err("Incorrect value[%d] found for GPR[%d]\n", 3128 1.1 riastrad cs[n], n); 3129 1.1 riastrad err = -EINVAL; 3130 1.1 riastrad break; 3131 1.1 riastrad } 3132 1.1 riastrad } 3133 1.1 riastrad 3134 1.1 riastrad i915_gem_object_unpin_map(scratch->obj); 3135 1.1 riastrad 3136 1.1 riastrad out_end: 3137 1.1 riastrad if (igt_live_test_end(&t)) 3138 1.1 riastrad err = -EIO; 3139 1.1 riastrad i915_request_put(last); 3140 1.1 riastrad out_unpin: 3141 1.1 riastrad intel_context_unpin(ve); 3142 1.1 riastrad out_put: 3143 1.1 riastrad intel_context_put(ve); 3144 1.1 riastrad out_scratch: 3145 1.1 riastrad i915_vma_unpin_and_release(&scratch, 0); 3146 1.1 riastrad return err; 3147 1.1 riastrad } 3148 1.1 riastrad 3149 1.1 riastrad static int live_virtual_preserved(void *arg) 3150 1.1 riastrad { 3151 1.1 riastrad struct intel_gt *gt = arg; 3152 1.1 riastrad struct intel_engine_cs *siblings[MAX_ENGINE_INSTANCE + 1]; 3153 1.1 riastrad unsigned int class, inst; 3154 1.1 riastrad 3155 1.1 riastrad /* 3156 1.1 riastrad * Check that the context image retains non-privileged (user) registers 3157 1.1 riastrad * from one engine to the next. For this we check that the CS_GPR 3158 1.1 riastrad * are preserved. 3159 1.1 riastrad */ 3160 1.1 riastrad 3161 1.1 riastrad if (USES_GUC_SUBMISSION(gt->i915)) 3162 1.1 riastrad return 0; 3163 1.1 riastrad 3164 1.1 riastrad /* As we use CS_GPR we cannot run before they existed on all engines. */ 3165 1.1 riastrad if (INTEL_GEN(gt->i915) < 9) 3166 1.1 riastrad return 0; 3167 1.1 riastrad 3168 1.1 riastrad for (class = 0; class <= MAX_ENGINE_CLASS; class++) { 3169 1.1 riastrad int nsibling, err; 3170 1.1 riastrad 3171 1.1 riastrad nsibling = 0; 3172 1.1 riastrad for (inst = 0; inst <= MAX_ENGINE_INSTANCE; inst++) { 3173 1.1 riastrad if (!gt->engine_class[class][inst]) 3174 1.1 riastrad continue; 3175 1.1 riastrad 3176 1.1 riastrad siblings[nsibling++] = gt->engine_class[class][inst]; 3177 1.1 riastrad } 3178 1.1 riastrad if (nsibling < 2) 3179 1.1 riastrad continue; 3180 1.1 riastrad 3181 1.1 riastrad err = preserved_virtual_engine(gt, siblings, nsibling); 3182 1.1 riastrad if (err) 3183 1.1 riastrad return err; 3184 1.1 riastrad } 3185 1.1 riastrad 3186 1.1 riastrad return 0; 3187 1.1 riastrad } 3188 1.1 riastrad 3189 1.1 riastrad static int bond_virtual_engine(struct intel_gt *gt, 3190 1.1 riastrad unsigned int class, 3191 1.1 riastrad struct intel_engine_cs **siblings, 3192 1.1 riastrad unsigned int nsibling, 3193 1.1 riastrad unsigned int flags) 3194 1.1 riastrad #define BOND_SCHEDULE BIT(0) 3195 1.1 riastrad { 3196 1.1 riastrad struct intel_engine_cs *master; 3197 1.1 riastrad struct i915_request *rq[16]; 3198 1.1 riastrad enum intel_engine_id id; 3199 1.1 riastrad struct igt_spinner spin; 3200 1.1 riastrad unsigned long n; 3201 1.1 riastrad int err; 3202 1.1 riastrad 3203 1.1 riastrad /* 3204 1.1 riastrad * A set of bonded requests is intended to be run concurrently 3205 1.1 riastrad * across a number of engines. We use one request per-engine 3206 1.1 riastrad * and a magic fence to schedule each of the bonded requests 3207 1.1 riastrad * at the same time. A consequence of our current scheduler is that 3208 1.1 riastrad * we only move requests to the HW ready queue when the request 3209 1.1 riastrad * becomes ready, that is when all of its prerequisite fences have 3210 1.1 riastrad * been signaled. As one of those fences is the master submit fence, 3211 1.1 riastrad * there is a delay on all secondary fences as the HW may be 3212 1.1 riastrad * currently busy. Equally, as all the requests are independent, 3213 1.1 riastrad * they may have other fences that delay individual request 3214 1.1 riastrad * submission to HW. Ergo, we do not guarantee that all requests are 3215 1.1 riastrad * immediately submitted to HW at the same time, just that if the 3216 1.1 riastrad * rules are abided by, they are ready at the same time as the 3217 1.1 riastrad * first is submitted. Userspace can embed semaphores in its batch 3218 1.1 riastrad * to ensure parallel execution of its phases as it requires. 3219 1.1 riastrad * Though naturally it gets requested that perhaps the scheduler should 3220 1.1 riastrad * take care of parallel execution, even across preemption events on 3221 1.1 riastrad * different HW. (The proper answer is of course "lalalala".) 3222 1.1 riastrad * 3223 1.1 riastrad * With the submit-fence, we have identified three possible phases 3224 1.1 riastrad * of synchronisation depending on the master fence: queued (not 3225 1.1 riastrad * ready), executing, and signaled. The first two are quite simple 3226 1.1 riastrad * and checked below. However, the signaled master fence handling is 3227 1.1 riastrad * contentious. Currently we do not distinguish between a signaled 3228 1.1 riastrad * fence and an expired fence, as once signaled it does not convey 3229 1.1 riastrad * any information about the previous execution. It may even be freed 3230 1.1 riastrad * and hence checking later it may not exist at all. Ergo we currently 3231 1.1 riastrad * do not apply the bonding constraint for an already signaled fence, 3232 1.1 riastrad * as our expectation is that it should not constrain the secondaries 3233 1.1 riastrad * and is outside of the scope of the bonded request API (i.e. all 3234 1.1 riastrad * userspace requests are meant to be running in parallel). As 3235 1.1 riastrad * it imposes no constraint, and is effectively a no-op, we do not 3236 1.1 riastrad * check below as normal execution flows are checked extensively above. 3237 1.1 riastrad * 3238 1.1 riastrad * XXX Is the degenerate handling of signaled submit fences the 3239 1.1 riastrad * expected behaviour for userpace? 3240 1.1 riastrad */ 3241 1.1 riastrad 3242 1.1 riastrad GEM_BUG_ON(nsibling >= ARRAY_SIZE(rq) - 1); 3243 1.1 riastrad 3244 1.1 riastrad if (igt_spinner_init(&spin, gt)) 3245 1.1 riastrad return -ENOMEM; 3246 1.1 riastrad 3247 1.1 riastrad err = 0; 3248 1.1 riastrad rq[0] = ERR_PTR(-ENOMEM); 3249 1.1 riastrad for_each_engine(master, gt, id) { 3250 1.1 riastrad struct i915_sw_fence fence = {}; 3251 1.1 riastrad 3252 1.1 riastrad if (master->class == class) 3253 1.1 riastrad continue; 3254 1.1 riastrad 3255 1.1 riastrad memset_p((void *)rq, ERR_PTR(-EINVAL), ARRAY_SIZE(rq)); 3256 1.1 riastrad 3257 1.1 riastrad rq[0] = igt_spinner_create_request(&spin, 3258 1.1 riastrad master->kernel_context, 3259 1.1 riastrad MI_NOOP); 3260 1.1 riastrad if (IS_ERR(rq[0])) { 3261 1.1 riastrad err = PTR_ERR(rq[0]); 3262 1.1 riastrad goto out; 3263 1.1 riastrad } 3264 1.1 riastrad i915_request_get(rq[0]); 3265 1.1 riastrad 3266 1.1 riastrad if (flags & BOND_SCHEDULE) { 3267 1.1 riastrad onstack_fence_init(&fence); 3268 1.1 riastrad err = i915_sw_fence_await_sw_fence_gfp(&rq[0]->submit, 3269 1.1 riastrad &fence, 3270 1.1 riastrad GFP_KERNEL); 3271 1.1 riastrad } 3272 1.1 riastrad 3273 1.1 riastrad i915_request_add(rq[0]); 3274 1.1 riastrad if (err < 0) 3275 1.1 riastrad goto out; 3276 1.1 riastrad 3277 1.1 riastrad if (!(flags & BOND_SCHEDULE) && 3278 1.1 riastrad !igt_wait_for_spinner(&spin, rq[0])) { 3279 1.1 riastrad err = -EIO; 3280 1.1 riastrad goto out; 3281 1.1 riastrad } 3282 1.1 riastrad 3283 1.1 riastrad for (n = 0; n < nsibling; n++) { 3284 1.1 riastrad struct intel_context *ve; 3285 1.1 riastrad 3286 1.1 riastrad ve = intel_execlists_create_virtual(siblings, nsibling); 3287 1.1 riastrad if (IS_ERR(ve)) { 3288 1.1 riastrad err = PTR_ERR(ve); 3289 1.1 riastrad onstack_fence_fini(&fence); 3290 1.1 riastrad goto out; 3291 1.1 riastrad } 3292 1.1 riastrad 3293 1.1 riastrad err = intel_virtual_engine_attach_bond(ve->engine, 3294 1.1 riastrad master, 3295 1.1 riastrad siblings[n]); 3296 1.1 riastrad if (err) { 3297 1.1 riastrad intel_context_put(ve); 3298 1.1 riastrad onstack_fence_fini(&fence); 3299 1.1 riastrad goto out; 3300 1.1 riastrad } 3301 1.1 riastrad 3302 1.1 riastrad err = intel_context_pin(ve); 3303 1.1 riastrad intel_context_put(ve); 3304 1.1 riastrad if (err) { 3305 1.1 riastrad onstack_fence_fini(&fence); 3306 1.1 riastrad goto out; 3307 1.1 riastrad } 3308 1.1 riastrad 3309 1.1 riastrad rq[n + 1] = i915_request_create(ve); 3310 1.1 riastrad intel_context_unpin(ve); 3311 1.1 riastrad if (IS_ERR(rq[n + 1])) { 3312 1.1 riastrad err = PTR_ERR(rq[n + 1]); 3313 1.1 riastrad onstack_fence_fini(&fence); 3314 1.1 riastrad goto out; 3315 1.1 riastrad } 3316 1.1 riastrad i915_request_get(rq[n + 1]); 3317 1.1 riastrad 3318 1.1 riastrad err = i915_request_await_execution(rq[n + 1], 3319 1.1 riastrad &rq[0]->fence, 3320 1.1 riastrad ve->engine->bond_execute); 3321 1.1 riastrad i915_request_add(rq[n + 1]); 3322 1.1 riastrad if (err < 0) { 3323 1.1 riastrad onstack_fence_fini(&fence); 3324 1.1 riastrad goto out; 3325 1.1 riastrad } 3326 1.1 riastrad } 3327 1.1 riastrad onstack_fence_fini(&fence); 3328 1.1 riastrad intel_engine_flush_submission(master); 3329 1.1 riastrad igt_spinner_end(&spin); 3330 1.1 riastrad 3331 1.1 riastrad if (i915_request_wait(rq[0], 0, HZ / 10) < 0) { 3332 1.1 riastrad pr_err("Master request did not execute (on %s)!\n", 3333 1.1 riastrad rq[0]->engine->name); 3334 1.1 riastrad err = -EIO; 3335 1.1 riastrad goto out; 3336 1.1 riastrad } 3337 1.1 riastrad 3338 1.1 riastrad for (n = 0; n < nsibling; n++) { 3339 1.1 riastrad if (i915_request_wait(rq[n + 1], 0, 3340 1.1 riastrad MAX_SCHEDULE_TIMEOUT) < 0) { 3341 1.1 riastrad err = -EIO; 3342 1.1 riastrad goto out; 3343 1.1 riastrad } 3344 1.1 riastrad 3345 1.1 riastrad if (rq[n + 1]->engine != siblings[n]) { 3346 1.1 riastrad pr_err("Bonded request did not execute on target engine: expected %s, used %s; master was %s\n", 3347 1.1 riastrad siblings[n]->name, 3348 1.1 riastrad rq[n + 1]->engine->name, 3349 1.1 riastrad rq[0]->engine->name); 3350 1.1 riastrad err = -EINVAL; 3351 1.1 riastrad goto out; 3352 1.1 riastrad } 3353 1.1 riastrad } 3354 1.1 riastrad 3355 1.1 riastrad for (n = 0; !IS_ERR(rq[n]); n++) 3356 1.1 riastrad i915_request_put(rq[n]); 3357 1.1 riastrad rq[0] = ERR_PTR(-ENOMEM); 3358 1.1 riastrad } 3359 1.1 riastrad 3360 1.1 riastrad out: 3361 1.1 riastrad for (n = 0; !IS_ERR(rq[n]); n++) 3362 1.1 riastrad i915_request_put(rq[n]); 3363 1.1 riastrad if (igt_flush_test(gt->i915)) 3364 1.1 riastrad err = -EIO; 3365 1.1 riastrad 3366 1.1 riastrad igt_spinner_fini(&spin); 3367 1.1 riastrad return err; 3368 1.1 riastrad } 3369 1.1 riastrad 3370 1.1 riastrad static int live_virtual_bond(void *arg) 3371 1.1 riastrad { 3372 1.1 riastrad static const struct phase { 3373 1.1 riastrad const char *name; 3374 1.1 riastrad unsigned int flags; 3375 1.1 riastrad } phases[] = { 3376 1.1 riastrad { "", 0 }, 3377 1.1 riastrad { "schedule", BOND_SCHEDULE }, 3378 1.1 riastrad { }, 3379 1.1 riastrad }; 3380 1.1 riastrad struct intel_gt *gt = arg; 3381 1.1 riastrad struct intel_engine_cs *siblings[MAX_ENGINE_INSTANCE + 1]; 3382 1.1 riastrad unsigned int class, inst; 3383 1.1 riastrad int err; 3384 1.1 riastrad 3385 1.1 riastrad if (USES_GUC_SUBMISSION(gt->i915)) 3386 1.1 riastrad return 0; 3387 1.1 riastrad 3388 1.1 riastrad for (class = 0; class <= MAX_ENGINE_CLASS; class++) { 3389 1.1 riastrad const struct phase *p; 3390 1.1 riastrad int nsibling; 3391 1.1 riastrad 3392 1.1 riastrad nsibling = 0; 3393 1.1 riastrad for (inst = 0; inst <= MAX_ENGINE_INSTANCE; inst++) { 3394 1.1 riastrad if (!gt->engine_class[class][inst]) 3395 1.1 riastrad break; 3396 1.1 riastrad 3397 1.1 riastrad GEM_BUG_ON(nsibling == ARRAY_SIZE(siblings)); 3398 1.1 riastrad siblings[nsibling++] = gt->engine_class[class][inst]; 3399 1.1 riastrad } 3400 1.1 riastrad if (nsibling < 2) 3401 1.1 riastrad continue; 3402 1.1 riastrad 3403 1.1 riastrad for (p = phases; p->name; p++) { 3404 1.1 riastrad err = bond_virtual_engine(gt, 3405 1.1 riastrad class, siblings, nsibling, 3406 1.1 riastrad p->flags); 3407 1.1 riastrad if (err) { 3408 1.1 riastrad pr_err("%s(%s): failed class=%d, nsibling=%d, err=%d\n", 3409 1.1 riastrad __func__, p->name, class, nsibling, err); 3410 1.1 riastrad return err; 3411 1.1 riastrad } 3412 1.1 riastrad } 3413 1.1 riastrad } 3414 1.1 riastrad 3415 1.1 riastrad return 0; 3416 1.1 riastrad } 3417 1.1 riastrad 3418 1.1 riastrad static int reset_virtual_engine(struct intel_gt *gt, 3419 1.1 riastrad struct intel_engine_cs **siblings, 3420 1.1 riastrad unsigned int nsibling) 3421 1.1 riastrad { 3422 1.1 riastrad struct intel_engine_cs *engine; 3423 1.1 riastrad struct intel_context *ve; 3424 1.1 riastrad unsigned long *heartbeat; 3425 1.1 riastrad struct igt_spinner spin; 3426 1.1 riastrad struct i915_request *rq; 3427 1.1 riastrad unsigned int n; 3428 1.1 riastrad int err = 0; 3429 1.1 riastrad 3430 1.1 riastrad /* 3431 1.1 riastrad * In order to support offline error capture for fast preempt reset, 3432 1.1 riastrad * we need to decouple the guilty request and ensure that it and its 3433 1.1 riastrad * descendents are not executed while the capture is in progress. 3434 1.1 riastrad */ 3435 1.1 riastrad 3436 1.1 riastrad heartbeat = kmalloc_array(nsibling, sizeof(*heartbeat), GFP_KERNEL); 3437 1.1 riastrad if (!heartbeat) 3438 1.1 riastrad return -ENOMEM; 3439 1.1 riastrad 3440 1.1 riastrad if (igt_spinner_init(&spin, gt)) { 3441 1.1 riastrad err = -ENOMEM; 3442 1.1 riastrad goto out_free; 3443 1.1 riastrad } 3444 1.1 riastrad 3445 1.1 riastrad ve = intel_execlists_create_virtual(siblings, nsibling); 3446 1.1 riastrad if (IS_ERR(ve)) { 3447 1.1 riastrad err = PTR_ERR(ve); 3448 1.1 riastrad goto out_spin; 3449 1.1 riastrad } 3450 1.1 riastrad 3451 1.1 riastrad for (n = 0; n < nsibling; n++) 3452 1.1 riastrad engine_heartbeat_disable(siblings[n], &heartbeat[n]); 3453 1.1 riastrad 3454 1.1 riastrad rq = igt_spinner_create_request(&spin, ve, MI_ARB_CHECK); 3455 1.1 riastrad if (IS_ERR(rq)) { 3456 1.1 riastrad err = PTR_ERR(rq); 3457 1.1 riastrad goto out_heartbeat; 3458 1.1 riastrad } 3459 1.1 riastrad i915_request_add(rq); 3460 1.1 riastrad 3461 1.1 riastrad if (!igt_wait_for_spinner(&spin, rq)) { 3462 1.1 riastrad intel_gt_set_wedged(gt); 3463 1.1 riastrad err = -ETIME; 3464 1.1 riastrad goto out_heartbeat; 3465 1.1 riastrad } 3466 1.1 riastrad 3467 1.1 riastrad engine = rq->engine; 3468 1.1 riastrad GEM_BUG_ON(engine == ve->engine); 3469 1.1 riastrad 3470 1.1 riastrad /* Take ownership of the reset and tasklet */ 3471 1.1 riastrad if (test_and_set_bit(I915_RESET_ENGINE + engine->id, 3472 1.1 riastrad >->reset.flags)) { 3473 1.1 riastrad intel_gt_set_wedged(gt); 3474 1.1 riastrad err = -EBUSY; 3475 1.1 riastrad goto out_heartbeat; 3476 1.1 riastrad } 3477 1.1 riastrad tasklet_disable(&engine->execlists.tasklet); 3478 1.1 riastrad 3479 1.1 riastrad engine->execlists.tasklet.func(engine->execlists.tasklet.data); 3480 1.1 riastrad GEM_BUG_ON(execlists_active(&engine->execlists) != rq); 3481 1.1 riastrad 3482 1.1 riastrad /* Fake a preemption event; failed of course */ 3483 1.1 riastrad spin_lock_irq(&engine->active.lock); 3484 1.1 riastrad __unwind_incomplete_requests(engine); 3485 1.1 riastrad spin_unlock_irq(&engine->active.lock); 3486 1.1 riastrad GEM_BUG_ON(rq->engine != ve->engine); 3487 1.1 riastrad 3488 1.1 riastrad /* Reset the engine while keeping our active request on hold */ 3489 1.1 riastrad execlists_hold(engine, rq); 3490 1.1 riastrad GEM_BUG_ON(!i915_request_on_hold(rq)); 3491 1.1 riastrad 3492 1.1 riastrad intel_engine_reset(engine, NULL); 3493 1.1 riastrad GEM_BUG_ON(rq->fence.error != -EIO); 3494 1.1 riastrad 3495 1.1 riastrad /* Release our grasp on the engine, letting CS flow again */ 3496 1.1 riastrad tasklet_enable(&engine->execlists.tasklet); 3497 1.1 riastrad clear_and_wake_up_bit(I915_RESET_ENGINE + engine->id, >->reset.flags); 3498 1.1 riastrad 3499 1.1 riastrad /* Check that we do not resubmit the held request */ 3500 1.1 riastrad i915_request_get(rq); 3501 1.1 riastrad if (!i915_request_wait(rq, 0, HZ / 5)) { 3502 1.1 riastrad pr_err("%s: on hold request completed!\n", 3503 1.1 riastrad engine->name); 3504 1.1 riastrad intel_gt_set_wedged(gt); 3505 1.1 riastrad err = -EIO; 3506 1.1 riastrad goto out_rq; 3507 1.1 riastrad } 3508 1.1 riastrad GEM_BUG_ON(!i915_request_on_hold(rq)); 3509 1.1 riastrad 3510 1.1 riastrad /* But is resubmitted on release */ 3511 1.1 riastrad execlists_unhold(engine, rq); 3512 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 5) < 0) { 3513 1.1 riastrad pr_err("%s: held request did not complete!\n", 3514 1.1 riastrad engine->name); 3515 1.1 riastrad intel_gt_set_wedged(gt); 3516 1.1 riastrad err = -ETIME; 3517 1.1 riastrad } 3518 1.1 riastrad 3519 1.1 riastrad out_rq: 3520 1.1 riastrad i915_request_put(rq); 3521 1.1 riastrad out_heartbeat: 3522 1.1 riastrad for (n = 0; n < nsibling; n++) 3523 1.1 riastrad engine_heartbeat_enable(siblings[n], heartbeat[n]); 3524 1.1 riastrad 3525 1.1 riastrad intel_context_put(ve); 3526 1.1 riastrad out_spin: 3527 1.1 riastrad igt_spinner_fini(&spin); 3528 1.1 riastrad out_free: 3529 1.1 riastrad kfree(heartbeat); 3530 1.1 riastrad return err; 3531 1.1 riastrad } 3532 1.1 riastrad 3533 1.1 riastrad static int live_virtual_reset(void *arg) 3534 1.1 riastrad { 3535 1.1 riastrad struct intel_gt *gt = arg; 3536 1.1 riastrad struct intel_engine_cs *siblings[MAX_ENGINE_INSTANCE + 1]; 3537 1.1 riastrad unsigned int class, inst; 3538 1.1 riastrad 3539 1.1 riastrad /* 3540 1.1 riastrad * Check that we handle a reset event within a virtual engine. 3541 1.1 riastrad * Only the physical engine is reset, but we have to check the flow 3542 1.1 riastrad * of the virtual requests around the reset, and make sure it is not 3543 1.1 riastrad * forgotten. 3544 1.1 riastrad */ 3545 1.1 riastrad 3546 1.1 riastrad if (USES_GUC_SUBMISSION(gt->i915)) 3547 1.1 riastrad return 0; 3548 1.1 riastrad 3549 1.1 riastrad if (!intel_has_reset_engine(gt)) 3550 1.1 riastrad return 0; 3551 1.1 riastrad 3552 1.1 riastrad for (class = 0; class <= MAX_ENGINE_CLASS; class++) { 3553 1.1 riastrad int nsibling, err; 3554 1.1 riastrad 3555 1.1 riastrad nsibling = 0; 3556 1.1 riastrad for (inst = 0; inst <= MAX_ENGINE_INSTANCE; inst++) { 3557 1.1 riastrad if (!gt->engine_class[class][inst]) 3558 1.1 riastrad continue; 3559 1.1 riastrad 3560 1.1 riastrad siblings[nsibling++] = gt->engine_class[class][inst]; 3561 1.1 riastrad } 3562 1.1 riastrad if (nsibling < 2) 3563 1.1 riastrad continue; 3564 1.1 riastrad 3565 1.1 riastrad err = reset_virtual_engine(gt, siblings, nsibling); 3566 1.1 riastrad if (err) 3567 1.1 riastrad return err; 3568 1.1 riastrad } 3569 1.1 riastrad 3570 1.1 riastrad return 0; 3571 1.1 riastrad } 3572 1.1 riastrad 3573 1.1 riastrad int intel_execlists_live_selftests(struct drm_i915_private *i915) 3574 1.1 riastrad { 3575 1.1 riastrad static const struct i915_subtest tests[] = { 3576 1.1 riastrad SUBTEST(live_sanitycheck), 3577 1.1 riastrad SUBTEST(live_unlite_switch), 3578 1.1 riastrad SUBTEST(live_unlite_preempt), 3579 1.1 riastrad SUBTEST(live_hold_reset), 3580 1.1 riastrad SUBTEST(live_timeslice_preempt), 3581 1.1 riastrad SUBTEST(live_timeslice_queue), 3582 1.1 riastrad SUBTEST(live_busywait_preempt), 3583 1.1 riastrad SUBTEST(live_preempt), 3584 1.1 riastrad SUBTEST(live_late_preempt), 3585 1.1 riastrad SUBTEST(live_nopreempt), 3586 1.1 riastrad SUBTEST(live_preempt_cancel), 3587 1.1 riastrad SUBTEST(live_suppress_self_preempt), 3588 1.1 riastrad SUBTEST(live_suppress_wait_preempt), 3589 1.1 riastrad SUBTEST(live_chain_preempt), 3590 1.1 riastrad SUBTEST(live_preempt_gang), 3591 1.1 riastrad SUBTEST(live_preempt_hang), 3592 1.1 riastrad SUBTEST(live_preempt_timeout), 3593 1.1 riastrad SUBTEST(live_preempt_smoke), 3594 1.1 riastrad SUBTEST(live_virtual_engine), 3595 1.1 riastrad SUBTEST(live_virtual_mask), 3596 1.1 riastrad SUBTEST(live_virtual_preserved), 3597 1.1 riastrad SUBTEST(live_virtual_bond), 3598 1.1 riastrad SUBTEST(live_virtual_reset), 3599 1.1 riastrad }; 3600 1.1 riastrad 3601 1.1 riastrad if (!HAS_EXECLISTS(i915)) 3602 1.1 riastrad return 0; 3603 1.1 riastrad 3604 1.1 riastrad if (intel_gt_is_wedged(&i915->gt)) 3605 1.1 riastrad return 0; 3606 1.1 riastrad 3607 1.1 riastrad return intel_gt_live_subtests(tests, &i915->gt); 3608 1.1 riastrad } 3609 1.1 riastrad 3610 1.1 riastrad static void hexdump(const void *buf, size_t len) 3611 1.1 riastrad { 3612 1.1 riastrad const size_t rowsize = 8 * sizeof(u32); 3613 1.1 riastrad const void *prev = NULL; 3614 1.1 riastrad bool skip = false; 3615 1.1 riastrad size_t pos; 3616 1.1 riastrad 3617 1.1 riastrad for (pos = 0; pos < len; pos += rowsize) { 3618 1.1 riastrad char line[128]; 3619 1.1 riastrad 3620 1.1 riastrad if (prev && !memcmp(prev, buf + pos, rowsize)) { 3621 1.1 riastrad if (!skip) { 3622 1.1 riastrad pr_info("*\n"); 3623 1.1 riastrad skip = true; 3624 1.1 riastrad } 3625 1.1 riastrad continue; 3626 1.1 riastrad } 3627 1.1 riastrad 3628 1.1 riastrad WARN_ON_ONCE(hex_dump_to_buffer(buf + pos, len - pos, 3629 1.1 riastrad rowsize, sizeof(u32), 3630 1.1 riastrad line, sizeof(line), 3631 1.1 riastrad false) >= sizeof(line)); 3632 1.1 riastrad pr_info("[%04zx] %s\n", pos, line); 3633 1.1 riastrad 3634 1.1 riastrad prev = buf + pos; 3635 1.1 riastrad skip = false; 3636 1.1 riastrad } 3637 1.1 riastrad } 3638 1.1 riastrad 3639 1.1 riastrad static int live_lrc_layout(void *arg) 3640 1.1 riastrad { 3641 1.1 riastrad struct intel_gt *gt = arg; 3642 1.1 riastrad struct intel_engine_cs *engine; 3643 1.1 riastrad enum intel_engine_id id; 3644 1.1 riastrad u32 *lrc; 3645 1.1 riastrad int err; 3646 1.1 riastrad 3647 1.1 riastrad /* 3648 1.1 riastrad * Check the registers offsets we use to create the initial reg state 3649 1.1 riastrad * match the layout saved by HW. 3650 1.1 riastrad */ 3651 1.1 riastrad 3652 1.1 riastrad lrc = kmalloc(PAGE_SIZE, GFP_KERNEL); 3653 1.1 riastrad if (!lrc) 3654 1.1 riastrad return -ENOMEM; 3655 1.1 riastrad 3656 1.1 riastrad err = 0; 3657 1.1 riastrad for_each_engine(engine, gt, id) { 3658 1.1 riastrad u32 *hw; 3659 1.1 riastrad int dw; 3660 1.1 riastrad 3661 1.1 riastrad if (!engine->default_state) 3662 1.1 riastrad continue; 3663 1.1 riastrad 3664 1.1 riastrad hw = i915_gem_object_pin_map(engine->default_state, 3665 1.1 riastrad I915_MAP_WB); 3666 1.1 riastrad if (IS_ERR(hw)) { 3667 1.1 riastrad err = PTR_ERR(hw); 3668 1.1 riastrad break; 3669 1.1 riastrad } 3670 1.1 riastrad hw += LRC_STATE_PN * PAGE_SIZE / sizeof(*hw); 3671 1.1 riastrad 3672 1.1 riastrad execlists_init_reg_state(memset(lrc, POISON_INUSE, PAGE_SIZE), 3673 1.1 riastrad engine->kernel_context, 3674 1.1 riastrad engine, 3675 1.1 riastrad engine->kernel_context->ring, 3676 1.1 riastrad true); 3677 1.1 riastrad 3678 1.1 riastrad dw = 0; 3679 1.1 riastrad do { 3680 1.1 riastrad u32 lri = hw[dw]; 3681 1.1 riastrad 3682 1.1 riastrad if (lri == 0) { 3683 1.1 riastrad dw++; 3684 1.1 riastrad continue; 3685 1.1 riastrad } 3686 1.1 riastrad 3687 1.1 riastrad if (lrc[dw] == 0) { 3688 1.1 riastrad pr_debug("%s: skipped instruction %x at dword %d\n", 3689 1.1 riastrad engine->name, lri, dw); 3690 1.1 riastrad dw++; 3691 1.1 riastrad continue; 3692 1.1 riastrad } 3693 1.1 riastrad 3694 1.1 riastrad if ((lri & GENMASK(31, 23)) != MI_INSTR(0x22, 0)) { 3695 1.1 riastrad pr_err("%s: Expected LRI command at dword %d, found %08x\n", 3696 1.1 riastrad engine->name, dw, lri); 3697 1.1 riastrad err = -EINVAL; 3698 1.1 riastrad break; 3699 1.1 riastrad } 3700 1.1 riastrad 3701 1.1 riastrad if (lrc[dw] != lri) { 3702 1.1 riastrad pr_err("%s: LRI command mismatch at dword %d, expected %08x found %08x\n", 3703 1.1 riastrad engine->name, dw, lri, lrc[dw]); 3704 1.1 riastrad err = -EINVAL; 3705 1.1 riastrad break; 3706 1.1 riastrad } 3707 1.1 riastrad 3708 1.1 riastrad lri &= 0x7f; 3709 1.1 riastrad lri++; 3710 1.1 riastrad dw++; 3711 1.1 riastrad 3712 1.1 riastrad while (lri) { 3713 1.1 riastrad if (hw[dw] != lrc[dw]) { 3714 1.1 riastrad pr_err("%s: Different registers found at dword %d, expected %x, found %x\n", 3715 1.1 riastrad engine->name, dw, hw[dw], lrc[dw]); 3716 1.1 riastrad err = -EINVAL; 3717 1.1 riastrad break; 3718 1.1 riastrad } 3719 1.1 riastrad 3720 1.1 riastrad /* 3721 1.1 riastrad * Skip over the actual register value as we 3722 1.1 riastrad * expect that to differ. 3723 1.1 riastrad */ 3724 1.1 riastrad dw += 2; 3725 1.1 riastrad lri -= 2; 3726 1.1 riastrad } 3727 1.1 riastrad } while ((lrc[dw] & ~BIT(0)) != MI_BATCH_BUFFER_END); 3728 1.1 riastrad 3729 1.1 riastrad if (err) { 3730 1.1 riastrad pr_info("%s: HW register image:\n", engine->name); 3731 1.1 riastrad hexdump(hw, PAGE_SIZE); 3732 1.1 riastrad 3733 1.1 riastrad pr_info("%s: SW register image:\n", engine->name); 3734 1.1 riastrad hexdump(lrc, PAGE_SIZE); 3735 1.1 riastrad } 3736 1.1 riastrad 3737 1.1 riastrad i915_gem_object_unpin_map(engine->default_state); 3738 1.1 riastrad if (err) 3739 1.1 riastrad break; 3740 1.1 riastrad } 3741 1.1 riastrad 3742 1.1 riastrad kfree(lrc); 3743 1.1 riastrad return err; 3744 1.1 riastrad } 3745 1.1 riastrad 3746 1.1 riastrad static int find_offset(const u32 *lri, u32 offset) 3747 1.1 riastrad { 3748 1.1 riastrad int i; 3749 1.1 riastrad 3750 1.1 riastrad for (i = 0; i < PAGE_SIZE / sizeof(u32); i++) 3751 1.1 riastrad if (lri[i] == offset) 3752 1.1 riastrad return i; 3753 1.1 riastrad 3754 1.1 riastrad return -1; 3755 1.1 riastrad } 3756 1.1 riastrad 3757 1.1 riastrad static int live_lrc_fixed(void *arg) 3758 1.1 riastrad { 3759 1.1 riastrad struct intel_gt *gt = arg; 3760 1.1 riastrad struct intel_engine_cs *engine; 3761 1.1 riastrad enum intel_engine_id id; 3762 1.1 riastrad int err = 0; 3763 1.1 riastrad 3764 1.1 riastrad /* 3765 1.1 riastrad * Check the assumed register offsets match the actual locations in 3766 1.1 riastrad * the context image. 3767 1.1 riastrad */ 3768 1.1 riastrad 3769 1.1 riastrad for_each_engine(engine, gt, id) { 3770 1.1 riastrad const struct { 3771 1.1 riastrad u32 reg; 3772 1.1 riastrad u32 offset; 3773 1.1 riastrad const char *name; 3774 1.1 riastrad } tbl[] = { 3775 1.1 riastrad { 3776 1.1 riastrad i915_mmio_reg_offset(RING_START(engine->mmio_base)), 3777 1.1 riastrad CTX_RING_START - 1, 3778 1.1 riastrad "RING_START" 3779 1.1 riastrad }, 3780 1.1 riastrad { 3781 1.1 riastrad i915_mmio_reg_offset(RING_CTL(engine->mmio_base)), 3782 1.1 riastrad CTX_RING_CTL - 1, 3783 1.1 riastrad "RING_CTL" 3784 1.1 riastrad }, 3785 1.1 riastrad { 3786 1.1 riastrad i915_mmio_reg_offset(RING_HEAD(engine->mmio_base)), 3787 1.1 riastrad CTX_RING_HEAD - 1, 3788 1.1 riastrad "RING_HEAD" 3789 1.1 riastrad }, 3790 1.1 riastrad { 3791 1.1 riastrad i915_mmio_reg_offset(RING_TAIL(engine->mmio_base)), 3792 1.1 riastrad CTX_RING_TAIL - 1, 3793 1.1 riastrad "RING_TAIL" 3794 1.1 riastrad }, 3795 1.1 riastrad { 3796 1.1 riastrad i915_mmio_reg_offset(RING_MI_MODE(engine->mmio_base)), 3797 1.1 riastrad lrc_ring_mi_mode(engine), 3798 1.1 riastrad "RING_MI_MODE" 3799 1.1 riastrad }, 3800 1.1 riastrad { 3801 1.1 riastrad i915_mmio_reg_offset(RING_BBSTATE(engine->mmio_base)), 3802 1.1 riastrad CTX_BB_STATE - 1, 3803 1.1 riastrad "BB_STATE" 3804 1.1 riastrad }, 3805 1.1 riastrad { }, 3806 1.1 riastrad }, *t; 3807 1.1 riastrad u32 *hw; 3808 1.1 riastrad 3809 1.1 riastrad if (!engine->default_state) 3810 1.1 riastrad continue; 3811 1.1 riastrad 3812 1.1 riastrad hw = i915_gem_object_pin_map(engine->default_state, 3813 1.1 riastrad I915_MAP_WB); 3814 1.1 riastrad if (IS_ERR(hw)) { 3815 1.1 riastrad err = PTR_ERR(hw); 3816 1.1 riastrad break; 3817 1.1 riastrad } 3818 1.1 riastrad hw += LRC_STATE_PN * PAGE_SIZE / sizeof(*hw); 3819 1.1 riastrad 3820 1.1 riastrad for (t = tbl; t->name; t++) { 3821 1.1 riastrad int dw = find_offset(hw, t->reg); 3822 1.1 riastrad 3823 1.1 riastrad if (dw != t->offset) { 3824 1.1 riastrad pr_err("%s: Offset for %s [0x%x] mismatch, found %x, expected %x\n", 3825 1.1 riastrad engine->name, 3826 1.1 riastrad t->name, 3827 1.1 riastrad t->reg, 3828 1.1 riastrad dw, 3829 1.1 riastrad t->offset); 3830 1.1 riastrad err = -EINVAL; 3831 1.1 riastrad } 3832 1.1 riastrad } 3833 1.1 riastrad 3834 1.1 riastrad i915_gem_object_unpin_map(engine->default_state); 3835 1.1 riastrad } 3836 1.1 riastrad 3837 1.1 riastrad return err; 3838 1.1 riastrad } 3839 1.1 riastrad 3840 1.1 riastrad static int __live_lrc_state(struct intel_engine_cs *engine, 3841 1.1 riastrad struct i915_vma *scratch) 3842 1.1 riastrad { 3843 1.1 riastrad struct intel_context *ce; 3844 1.1 riastrad struct i915_request *rq; 3845 1.1 riastrad enum { 3846 1.1 riastrad RING_START_IDX = 0, 3847 1.1 riastrad RING_TAIL_IDX, 3848 1.1 riastrad MAX_IDX 3849 1.1 riastrad }; 3850 1.1 riastrad u32 expected[MAX_IDX]; 3851 1.1 riastrad u32 *cs; 3852 1.1 riastrad int err; 3853 1.1 riastrad int n; 3854 1.1 riastrad 3855 1.1 riastrad ce = intel_context_create(engine); 3856 1.1 riastrad if (IS_ERR(ce)) 3857 1.1 riastrad return PTR_ERR(ce); 3858 1.1 riastrad 3859 1.1 riastrad err = intel_context_pin(ce); 3860 1.1 riastrad if (err) 3861 1.1 riastrad goto err_put; 3862 1.1 riastrad 3863 1.1 riastrad rq = i915_request_create(ce); 3864 1.1 riastrad if (IS_ERR(rq)) { 3865 1.1 riastrad err = PTR_ERR(rq); 3866 1.1 riastrad goto err_unpin; 3867 1.1 riastrad } 3868 1.1 riastrad 3869 1.1 riastrad cs = intel_ring_begin(rq, 4 * MAX_IDX); 3870 1.1 riastrad if (IS_ERR(cs)) { 3871 1.1 riastrad err = PTR_ERR(cs); 3872 1.1 riastrad i915_request_add(rq); 3873 1.1 riastrad goto err_unpin; 3874 1.1 riastrad } 3875 1.1 riastrad 3876 1.1 riastrad *cs++ = MI_STORE_REGISTER_MEM_GEN8 | MI_USE_GGTT; 3877 1.1 riastrad *cs++ = i915_mmio_reg_offset(RING_START(engine->mmio_base)); 3878 1.1 riastrad *cs++ = i915_ggtt_offset(scratch) + RING_START_IDX * sizeof(u32); 3879 1.1 riastrad *cs++ = 0; 3880 1.1 riastrad 3881 1.1 riastrad expected[RING_START_IDX] = i915_ggtt_offset(ce->ring->vma); 3882 1.1 riastrad 3883 1.1 riastrad *cs++ = MI_STORE_REGISTER_MEM_GEN8 | MI_USE_GGTT; 3884 1.1 riastrad *cs++ = i915_mmio_reg_offset(RING_TAIL(engine->mmio_base)); 3885 1.1 riastrad *cs++ = i915_ggtt_offset(scratch) + RING_TAIL_IDX * sizeof(u32); 3886 1.1 riastrad *cs++ = 0; 3887 1.1 riastrad 3888 1.1 riastrad i915_request_get(rq); 3889 1.1 riastrad i915_request_add(rq); 3890 1.1 riastrad 3891 1.1 riastrad intel_engine_flush_submission(engine); 3892 1.1 riastrad expected[RING_TAIL_IDX] = ce->ring->tail; 3893 1.1 riastrad 3894 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 5) < 0) { 3895 1.1 riastrad err = -ETIME; 3896 1.1 riastrad goto err_rq; 3897 1.1 riastrad } 3898 1.1 riastrad 3899 1.1 riastrad cs = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB); 3900 1.1 riastrad if (IS_ERR(cs)) { 3901 1.1 riastrad err = PTR_ERR(cs); 3902 1.1 riastrad goto err_rq; 3903 1.1 riastrad } 3904 1.1 riastrad 3905 1.1 riastrad for (n = 0; n < MAX_IDX; n++) { 3906 1.1 riastrad if (cs[n] != expected[n]) { 3907 1.1 riastrad pr_err("%s: Stored register[%d] value[0x%x] did not match expected[0x%x]\n", 3908 1.1 riastrad engine->name, n, cs[n], expected[n]); 3909 1.1 riastrad err = -EINVAL; 3910 1.1 riastrad break; 3911 1.1 riastrad } 3912 1.1 riastrad } 3913 1.1 riastrad 3914 1.1 riastrad i915_gem_object_unpin_map(scratch->obj); 3915 1.1 riastrad 3916 1.1 riastrad err_rq: 3917 1.1 riastrad i915_request_put(rq); 3918 1.1 riastrad err_unpin: 3919 1.1 riastrad intel_context_unpin(ce); 3920 1.1 riastrad err_put: 3921 1.1 riastrad intel_context_put(ce); 3922 1.1 riastrad return err; 3923 1.1 riastrad } 3924 1.1 riastrad 3925 1.1 riastrad static int live_lrc_state(void *arg) 3926 1.1 riastrad { 3927 1.1 riastrad struct intel_gt *gt = arg; 3928 1.1 riastrad struct intel_engine_cs *engine; 3929 1.1 riastrad struct i915_vma *scratch; 3930 1.1 riastrad enum intel_engine_id id; 3931 1.1 riastrad int err = 0; 3932 1.1 riastrad 3933 1.1 riastrad /* 3934 1.1 riastrad * Check the live register state matches what we expect for this 3935 1.1 riastrad * intel_context. 3936 1.1 riastrad */ 3937 1.1 riastrad 3938 1.1 riastrad scratch = create_scratch(gt); 3939 1.1 riastrad if (IS_ERR(scratch)) 3940 1.1 riastrad return PTR_ERR(scratch); 3941 1.1 riastrad 3942 1.1 riastrad for_each_engine(engine, gt, id) { 3943 1.1 riastrad err = __live_lrc_state(engine, scratch); 3944 1.1 riastrad if (err) 3945 1.1 riastrad break; 3946 1.1 riastrad } 3947 1.1 riastrad 3948 1.1 riastrad if (igt_flush_test(gt->i915)) 3949 1.1 riastrad err = -EIO; 3950 1.1 riastrad 3951 1.1 riastrad i915_vma_unpin_and_release(&scratch, 0); 3952 1.1 riastrad return err; 3953 1.1 riastrad } 3954 1.1 riastrad 3955 1.1 riastrad static int gpr_make_dirty(struct intel_engine_cs *engine) 3956 1.1 riastrad { 3957 1.1 riastrad struct i915_request *rq; 3958 1.1 riastrad u32 *cs; 3959 1.1 riastrad int n; 3960 1.1 riastrad 3961 1.1 riastrad rq = intel_engine_create_kernel_request(engine); 3962 1.1 riastrad if (IS_ERR(rq)) 3963 1.1 riastrad return PTR_ERR(rq); 3964 1.1 riastrad 3965 1.1 riastrad cs = intel_ring_begin(rq, 2 * NUM_GPR_DW + 2); 3966 1.1 riastrad if (IS_ERR(cs)) { 3967 1.1 riastrad i915_request_add(rq); 3968 1.1 riastrad return PTR_ERR(cs); 3969 1.1 riastrad } 3970 1.1 riastrad 3971 1.1 riastrad *cs++ = MI_LOAD_REGISTER_IMM(NUM_GPR_DW); 3972 1.1 riastrad for (n = 0; n < NUM_GPR_DW; n++) { 3973 1.1 riastrad *cs++ = CS_GPR(engine, n); 3974 1.1 riastrad *cs++ = STACK_MAGIC; 3975 1.1 riastrad } 3976 1.1 riastrad *cs++ = MI_NOOP; 3977 1.1 riastrad 3978 1.1 riastrad intel_ring_advance(rq, cs); 3979 1.1 riastrad i915_request_add(rq); 3980 1.1 riastrad 3981 1.1 riastrad return 0; 3982 1.1 riastrad } 3983 1.1 riastrad 3984 1.1 riastrad static int __live_gpr_clear(struct intel_engine_cs *engine, 3985 1.1 riastrad struct i915_vma *scratch) 3986 1.1 riastrad { 3987 1.1 riastrad struct intel_context *ce; 3988 1.1 riastrad struct i915_request *rq; 3989 1.1 riastrad u32 *cs; 3990 1.1 riastrad int err; 3991 1.1 riastrad int n; 3992 1.1 riastrad 3993 1.1 riastrad if (INTEL_GEN(engine->i915) < 9 && engine->class != RENDER_CLASS) 3994 1.1 riastrad return 0; /* GPR only on rcs0 for gen8 */ 3995 1.1 riastrad 3996 1.1 riastrad err = gpr_make_dirty(engine); 3997 1.1 riastrad if (err) 3998 1.1 riastrad return err; 3999 1.1 riastrad 4000 1.1 riastrad ce = intel_context_create(engine); 4001 1.1 riastrad if (IS_ERR(ce)) 4002 1.1 riastrad return PTR_ERR(ce); 4003 1.1 riastrad 4004 1.1 riastrad rq = intel_context_create_request(ce); 4005 1.1 riastrad if (IS_ERR(rq)) { 4006 1.1 riastrad err = PTR_ERR(rq); 4007 1.1 riastrad goto err_put; 4008 1.1 riastrad } 4009 1.1 riastrad 4010 1.1 riastrad cs = intel_ring_begin(rq, 4 * NUM_GPR_DW); 4011 1.1 riastrad if (IS_ERR(cs)) { 4012 1.1 riastrad err = PTR_ERR(cs); 4013 1.1 riastrad i915_request_add(rq); 4014 1.1 riastrad goto err_put; 4015 1.1 riastrad } 4016 1.1 riastrad 4017 1.1 riastrad for (n = 0; n < NUM_GPR_DW; n++) { 4018 1.1 riastrad *cs++ = MI_STORE_REGISTER_MEM_GEN8 | MI_USE_GGTT; 4019 1.1 riastrad *cs++ = CS_GPR(engine, n); 4020 1.1 riastrad *cs++ = i915_ggtt_offset(scratch) + n * sizeof(u32); 4021 1.1 riastrad *cs++ = 0; 4022 1.1 riastrad } 4023 1.1 riastrad 4024 1.1 riastrad i915_request_get(rq); 4025 1.1 riastrad i915_request_add(rq); 4026 1.1 riastrad 4027 1.1 riastrad if (i915_request_wait(rq, 0, HZ / 5) < 0) { 4028 1.1 riastrad err = -ETIME; 4029 1.1 riastrad goto err_rq; 4030 1.1 riastrad } 4031 1.1 riastrad 4032 1.1 riastrad cs = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB); 4033 1.1 riastrad if (IS_ERR(cs)) { 4034 1.1 riastrad err = PTR_ERR(cs); 4035 1.1 riastrad goto err_rq; 4036 1.1 riastrad } 4037 1.1 riastrad 4038 1.1 riastrad for (n = 0; n < NUM_GPR_DW; n++) { 4039 1.1 riastrad if (cs[n]) { 4040 1.1 riastrad pr_err("%s: GPR[%d].%s was not zero, found 0x%08x!\n", 4041 1.1 riastrad engine->name, 4042 1.1 riastrad n / 2, n & 1 ? "udw" : "ldw", 4043 1.1 riastrad cs[n]); 4044 1.1 riastrad err = -EINVAL; 4045 1.1 riastrad break; 4046 1.1 riastrad } 4047 1.1 riastrad } 4048 1.1 riastrad 4049 1.1 riastrad i915_gem_object_unpin_map(scratch->obj); 4050 1.1 riastrad 4051 1.1 riastrad err_rq: 4052 1.1 riastrad i915_request_put(rq); 4053 1.1 riastrad err_put: 4054 1.1 riastrad intel_context_put(ce); 4055 1.1 riastrad return err; 4056 1.1 riastrad } 4057 1.1 riastrad 4058 1.1 riastrad static int live_gpr_clear(void *arg) 4059 1.1 riastrad { 4060 1.1 riastrad struct intel_gt *gt = arg; 4061 1.1 riastrad struct intel_engine_cs *engine; 4062 1.1 riastrad struct i915_vma *scratch; 4063 1.1 riastrad enum intel_engine_id id; 4064 1.1 riastrad int err = 0; 4065 1.1 riastrad 4066 1.1 riastrad /* 4067 1.1 riastrad * Check that GPR registers are cleared in new contexts as we need 4068 1.1 riastrad * to avoid leaking any information from previous contexts. 4069 1.1 riastrad */ 4070 1.1 riastrad 4071 1.1 riastrad scratch = create_scratch(gt); 4072 1.1 riastrad if (IS_ERR(scratch)) 4073 1.1 riastrad return PTR_ERR(scratch); 4074 1.1 riastrad 4075 1.1 riastrad for_each_engine(engine, gt, id) { 4076 1.1 riastrad err = __live_gpr_clear(engine, scratch); 4077 1.1 riastrad if (err) 4078 1.1 riastrad break; 4079 1.1 riastrad } 4080 1.1 riastrad 4081 1.1 riastrad if (igt_flush_test(gt->i915)) 4082 1.1 riastrad err = -EIO; 4083 1.1 riastrad 4084 1.1 riastrad i915_vma_unpin_and_release(&scratch, 0); 4085 1.1 riastrad return err; 4086 1.1 riastrad } 4087 1.1 riastrad 4088 1.1 riastrad int intel_lrc_live_selftests(struct drm_i915_private *i915) 4089 1.1 riastrad { 4090 1.1 riastrad static const struct i915_subtest tests[] = { 4091 1.1 riastrad SUBTEST(live_lrc_layout), 4092 1.1 riastrad SUBTEST(live_lrc_fixed), 4093 1.1 riastrad SUBTEST(live_lrc_state), 4094 1.1 riastrad SUBTEST(live_gpr_clear), 4095 1.1 riastrad }; 4096 1.1 riastrad 4097 1.1 riastrad if (!HAS_LOGICAL_RING_CONTEXTS(i915)) 4098 1.1 riastrad return 0; 4099 1.1 riastrad 4100 1.1 riastrad return intel_gt_live_subtests(tests, &i915->gt); 4101 1.1 riastrad } 4102