xref: /openbmc/linux/drivers/gpu/drm/i915/gt/selftest_engine_pm.c (revision 0791faebfe750292a8a842b64795a390ca4a3b51)
124f90d66SChris Wilson // SPDX-License-Identifier: GPL-2.0
2c7302f20SChris Wilson /*
3c7302f20SChris Wilson  * Copyright © 2018 Intel Corporation
4c7302f20SChris Wilson  */
5c7302f20SChris Wilson 
68391c9b2SChris Wilson #include <linux/sort.h>
745233ab2SChris Wilson 
8*860cf3bdSTejas Upadhyay #include "gt/intel_gt_print.h"
9c7302f20SChris Wilson #include "i915_selftest.h"
10202b1f4cSMatt Roper #include "intel_engine_regs.h"
118391c9b2SChris Wilson #include "intel_gpu_commands.h"
128391c9b2SChris Wilson #include "intel_gt_clock_utils.h"
13c7302f20SChris Wilson #include "selftest_engine.h"
141b90e4a4SChris Wilson #include "selftest_engine_heartbeat.h"
15c7302f20SChris Wilson #include "selftests/igt_atomic.h"
161b90e4a4SChris Wilson #include "selftests/igt_flush_test.h"
171b90e4a4SChris Wilson #include "selftests/igt_spinner.h"
181b90e4a4SChris Wilson 
198391c9b2SChris Wilson #define COUNT 5
208391c9b2SChris Wilson 
cmp_u64(const void * A,const void * B)218391c9b2SChris Wilson static int cmp_u64(const void *A, const void *B)
228391c9b2SChris Wilson {
238391c9b2SChris Wilson 	const u64 *a = A, *b = B;
248391c9b2SChris Wilson 
258391c9b2SChris Wilson 	return *a - *b;
268391c9b2SChris Wilson }
278391c9b2SChris Wilson 
trifilter(u64 * a)288391c9b2SChris Wilson static u64 trifilter(u64 *a)
298391c9b2SChris Wilson {
308391c9b2SChris Wilson 	sort(a, COUNT, sizeof(*a), cmp_u64, NULL);
318391c9b2SChris Wilson 	return (a[1] + 2 * a[2] + a[3]) >> 2;
328391c9b2SChris Wilson }
338391c9b2SChris Wilson 
emit_wait(u32 * cs,u32 offset,int op,u32 value)348391c9b2SChris Wilson static u32 *emit_wait(u32 *cs, u32 offset, int op, u32 value)
358391c9b2SChris Wilson {
368391c9b2SChris Wilson 	*cs++ = MI_SEMAPHORE_WAIT |
378391c9b2SChris Wilson 		MI_SEMAPHORE_GLOBAL_GTT |
388391c9b2SChris Wilson 		MI_SEMAPHORE_POLL |
398391c9b2SChris Wilson 		op;
408391c9b2SChris Wilson 	*cs++ = value;
418391c9b2SChris Wilson 	*cs++ = offset;
428391c9b2SChris Wilson 	*cs++ = 0;
438391c9b2SChris Wilson 
448391c9b2SChris Wilson 	return cs;
458391c9b2SChris Wilson }
468391c9b2SChris Wilson 
emit_store(u32 * cs,u32 offset,u32 value)478391c9b2SChris Wilson static u32 *emit_store(u32 *cs, u32 offset, u32 value)
488391c9b2SChris Wilson {
498391c9b2SChris Wilson 	*cs++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT;
508391c9b2SChris Wilson 	*cs++ = offset;
518391c9b2SChris Wilson 	*cs++ = 0;
528391c9b2SChris Wilson 	*cs++ = value;
538391c9b2SChris Wilson 
548391c9b2SChris Wilson 	return cs;
558391c9b2SChris Wilson }
568391c9b2SChris Wilson 
emit_srm(u32 * cs,i915_reg_t reg,u32 offset)578391c9b2SChris Wilson static u32 *emit_srm(u32 *cs, i915_reg_t reg, u32 offset)
588391c9b2SChris Wilson {
598391c9b2SChris Wilson 	*cs++ = MI_STORE_REGISTER_MEM_GEN8 | MI_USE_GGTT;
608391c9b2SChris Wilson 	*cs++ = i915_mmio_reg_offset(reg);
618391c9b2SChris Wilson 	*cs++ = offset;
628391c9b2SChris Wilson 	*cs++ = 0;
638391c9b2SChris Wilson 
648391c9b2SChris Wilson 	return cs;
658391c9b2SChris Wilson }
668391c9b2SChris Wilson 
write_semaphore(u32 * x,u32 value)678391c9b2SChris Wilson static void write_semaphore(u32 *x, u32 value)
688391c9b2SChris Wilson {
698391c9b2SChris Wilson 	WRITE_ONCE(*x, value);
708391c9b2SChris Wilson 	wmb();
718391c9b2SChris Wilson }
728391c9b2SChris Wilson 
__measure_timestamps(struct intel_context * ce,u64 * dt,u64 * d_ring,u64 * d_ctx)738391c9b2SChris Wilson static int __measure_timestamps(struct intel_context *ce,
748391c9b2SChris Wilson 				u64 *dt, u64 *d_ring, u64 *d_ctx)
758391c9b2SChris Wilson {
768391c9b2SChris Wilson 	struct intel_engine_cs *engine = ce->engine;
778391c9b2SChris Wilson 	u32 *sema = memset32(engine->status_page.addr + 1000, 0, 5);
788391c9b2SChris Wilson 	u32 offset = i915_ggtt_offset(engine->status_page.vma);
798391c9b2SChris Wilson 	struct i915_request *rq;
808391c9b2SChris Wilson 	u32 *cs;
818391c9b2SChris Wilson 
828391c9b2SChris Wilson 	rq = intel_context_create_request(ce);
838391c9b2SChris Wilson 	if (IS_ERR(rq))
848391c9b2SChris Wilson 		return PTR_ERR(rq);
858391c9b2SChris Wilson 
868391c9b2SChris Wilson 	cs = intel_ring_begin(rq, 28);
878391c9b2SChris Wilson 	if (IS_ERR(cs)) {
888391c9b2SChris Wilson 		i915_request_add(rq);
898391c9b2SChris Wilson 		return PTR_ERR(cs);
908391c9b2SChris Wilson 	}
918391c9b2SChris Wilson 
928391c9b2SChris Wilson 	/* Signal & wait for start */
938391c9b2SChris Wilson 	cs = emit_store(cs, offset + 4008, 1);
948391c9b2SChris Wilson 	cs = emit_wait(cs, offset + 4008, MI_SEMAPHORE_SAD_NEQ_SDD, 1);
958391c9b2SChris Wilson 
968391c9b2SChris Wilson 	cs = emit_srm(cs, RING_TIMESTAMP(engine->mmio_base), offset + 4000);
978391c9b2SChris Wilson 	cs = emit_srm(cs, RING_CTX_TIMESTAMP(engine->mmio_base), offset + 4004);
988391c9b2SChris Wilson 
998391c9b2SChris Wilson 	/* Busy wait */
1008391c9b2SChris Wilson 	cs = emit_wait(cs, offset + 4008, MI_SEMAPHORE_SAD_EQ_SDD, 1);
1018391c9b2SChris Wilson 
1028391c9b2SChris Wilson 	cs = emit_srm(cs, RING_TIMESTAMP(engine->mmio_base), offset + 4016);
1038391c9b2SChris Wilson 	cs = emit_srm(cs, RING_CTX_TIMESTAMP(engine->mmio_base), offset + 4012);
1048391c9b2SChris Wilson 
1058391c9b2SChris Wilson 	intel_ring_advance(rq, cs);
1068391c9b2SChris Wilson 	i915_request_get(rq);
1078391c9b2SChris Wilson 	i915_request_add(rq);
1088391c9b2SChris Wilson 	intel_engine_flush_submission(engine);
1098391c9b2SChris Wilson 
1108391c9b2SChris Wilson 	/* Wait for the request to start executing, that then waits for us */
1118391c9b2SChris Wilson 	while (READ_ONCE(sema[2]) == 0)
1128391c9b2SChris Wilson 		cpu_relax();
1138391c9b2SChris Wilson 
1148391c9b2SChris Wilson 	/* Run the request for a 100us, sampling timestamps before/after */
115985458d7SChris Wilson 	local_irq_disable();
1168391c9b2SChris Wilson 	write_semaphore(&sema[2], 0);
117985458d7SChris Wilson 	while (READ_ONCE(sema[1]) == 0) /* wait for the gpu to catch up */
118985458d7SChris Wilson 		cpu_relax();
119985458d7SChris Wilson 	*dt = local_clock();
1208391c9b2SChris Wilson 	udelay(100);
1210399d0e3SChris Wilson 	*dt = local_clock() - *dt;
1228391c9b2SChris Wilson 	write_semaphore(&sema[2], 1);
123985458d7SChris Wilson 	local_irq_enable();
1248391c9b2SChris Wilson 
1258391c9b2SChris Wilson 	if (i915_request_wait(rq, 0, HZ / 2) < 0) {
1268391c9b2SChris Wilson 		i915_request_put(rq);
1278391c9b2SChris Wilson 		return -ETIME;
1288391c9b2SChris Wilson 	}
1298391c9b2SChris Wilson 	i915_request_put(rq);
1308391c9b2SChris Wilson 
1318391c9b2SChris Wilson 	pr_debug("%s CTX_TIMESTAMP: [%x, %x], RING_TIMESTAMP: [%x, %x]\n",
1328391c9b2SChris Wilson 		 engine->name, sema[1], sema[3], sema[0], sema[4]);
1338391c9b2SChris Wilson 
1348391c9b2SChris Wilson 	*d_ctx = sema[3] - sema[1];
1358391c9b2SChris Wilson 	*d_ring = sema[4] - sema[0];
1368391c9b2SChris Wilson 	return 0;
1378391c9b2SChris Wilson }
1388391c9b2SChris Wilson 
__live_engine_timestamps(struct intel_engine_cs * engine)1398391c9b2SChris Wilson static int __live_engine_timestamps(struct intel_engine_cs *engine)
1408391c9b2SChris Wilson {
1418391c9b2SChris Wilson 	u64 s_ring[COUNT], s_ctx[COUNT], st[COUNT], d_ring, d_ctx, dt;
1428391c9b2SChris Wilson 	struct intel_context *ce;
1438391c9b2SChris Wilson 	int i, err = 0;
1448391c9b2SChris Wilson 
1458391c9b2SChris Wilson 	ce = intel_context_create(engine);
1468391c9b2SChris Wilson 	if (IS_ERR(ce))
1478391c9b2SChris Wilson 		return PTR_ERR(ce);
1488391c9b2SChris Wilson 
1498391c9b2SChris Wilson 	for (i = 0; i < COUNT; i++) {
1508391c9b2SChris Wilson 		err = __measure_timestamps(ce, &st[i], &s_ring[i], &s_ctx[i]);
1518391c9b2SChris Wilson 		if (err)
1528391c9b2SChris Wilson 			break;
1538391c9b2SChris Wilson 	}
1548391c9b2SChris Wilson 	intel_context_put(ce);
1558391c9b2SChris Wilson 	if (err)
1568391c9b2SChris Wilson 		return err;
1578391c9b2SChris Wilson 
1588391c9b2SChris Wilson 	dt = trifilter(st);
1598391c9b2SChris Wilson 	d_ring = trifilter(s_ring);
1608391c9b2SChris Wilson 	d_ctx = trifilter(s_ctx);
1618391c9b2SChris Wilson 
162f170523aSChris Wilson 	pr_info("%s elapsed:%lldns, CTX_TIMESTAMP:%lldns, RING_TIMESTAMP:%lldns\n",
1638391c9b2SChris Wilson 		engine->name, dt,
1648391c9b2SChris Wilson 		intel_gt_clock_interval_to_ns(engine->gt, d_ctx),
1658391c9b2SChris Wilson 		intel_gt_clock_interval_to_ns(engine->gt, d_ring));
1668391c9b2SChris Wilson 
1678391c9b2SChris Wilson 	d_ring = intel_gt_clock_interval_to_ns(engine->gt, d_ring);
1688391c9b2SChris Wilson 	if (3 * dt > 4 * d_ring || 4 * dt < 3 * d_ring) {
1698391c9b2SChris Wilson 		pr_err("%s Mismatch between ring timestamp and walltime!\n",
1708391c9b2SChris Wilson 		       engine->name);
1718391c9b2SChris Wilson 		return -EINVAL;
1728391c9b2SChris Wilson 	}
1738391c9b2SChris Wilson 
1748391c9b2SChris Wilson 	d_ring = trifilter(s_ring);
1758391c9b2SChris Wilson 	d_ctx = trifilter(s_ctx);
1768391c9b2SChris Wilson 
177f170523aSChris Wilson 	d_ctx *= engine->gt->clock_frequency;
17853fe9cf2STejas Upadhyay 	if (GRAPHICS_VER(engine->i915) == 11)
17953fe9cf2STejas Upadhyay 		d_ring *= 12500000; /* Fixed 80ns for GEN11 ctx timestamp? */
1808391c9b2SChris Wilson 	else
181f170523aSChris Wilson 		d_ring *= engine->gt->clock_frequency;
1828391c9b2SChris Wilson 
1838391c9b2SChris Wilson 	if (3 * d_ctx > 4 * d_ring || 4 * d_ctx < 3 * d_ring) {
1848391c9b2SChris Wilson 		pr_err("%s Mismatch between ring and context timestamps!\n",
1858391c9b2SChris Wilson 		       engine->name);
1868391c9b2SChris Wilson 		return -EINVAL;
1878391c9b2SChris Wilson 	}
1888391c9b2SChris Wilson 
1898391c9b2SChris Wilson 	return 0;
1908391c9b2SChris Wilson }
1918391c9b2SChris Wilson 
live_engine_timestamps(void * arg)1928391c9b2SChris Wilson static int live_engine_timestamps(void *arg)
1938391c9b2SChris Wilson {
1948391c9b2SChris Wilson 	struct intel_gt *gt = arg;
1958391c9b2SChris Wilson 	struct intel_engine_cs *engine;
1968391c9b2SChris Wilson 	enum intel_engine_id id;
1978391c9b2SChris Wilson 
1988391c9b2SChris Wilson 	/*
1998391c9b2SChris Wilson 	 * Check that CS_TIMESTAMP / CTX_TIMESTAMP are in sync, i.e. share
2008391c9b2SChris Wilson 	 * the same CS clock.
2018391c9b2SChris Wilson 	 */
2028391c9b2SChris Wilson 
203c816723bSLucas De Marchi 	if (GRAPHICS_VER(gt->i915) < 8)
2048391c9b2SChris Wilson 		return 0;
2058391c9b2SChris Wilson 
2068391c9b2SChris Wilson 	for_each_engine(engine, gt, id) {
2078391c9b2SChris Wilson 		int err;
2088391c9b2SChris Wilson 
2098391c9b2SChris Wilson 		st_engine_heartbeat_disable(engine);
2108391c9b2SChris Wilson 		err = __live_engine_timestamps(engine);
2118391c9b2SChris Wilson 		st_engine_heartbeat_enable(engine);
2128391c9b2SChris Wilson 		if (err)
2138391c9b2SChris Wilson 			return err;
2148391c9b2SChris Wilson 	}
2158391c9b2SChris Wilson 
2168391c9b2SChris Wilson 	return 0;
2178391c9b2SChris Wilson }
2188391c9b2SChris Wilson 
__spin_until_busier(struct intel_engine_cs * engine,ktime_t busyness)21977cdd054SUmesh Nerlige Ramappa static int __spin_until_busier(struct intel_engine_cs *engine, ktime_t busyness)
22077cdd054SUmesh Nerlige Ramappa {
22177cdd054SUmesh Nerlige Ramappa 	ktime_t start, unused, dt;
22277cdd054SUmesh Nerlige Ramappa 
22377cdd054SUmesh Nerlige Ramappa 	if (!intel_engine_uses_guc(engine))
22477cdd054SUmesh Nerlige Ramappa 		return 0;
22577cdd054SUmesh Nerlige Ramappa 
22677cdd054SUmesh Nerlige Ramappa 	/*
22777cdd054SUmesh Nerlige Ramappa 	 * In GuC mode of submission, the busyness stats may get updated after
22877cdd054SUmesh Nerlige Ramappa 	 * the batch starts running. Poll for a change in busyness and timeout
22977cdd054SUmesh Nerlige Ramappa 	 * after 500 us.
23077cdd054SUmesh Nerlige Ramappa 	 */
23177cdd054SUmesh Nerlige Ramappa 	start = ktime_get();
23277cdd054SUmesh Nerlige Ramappa 	while (intel_engine_get_busy_time(engine, &unused) == busyness) {
23377cdd054SUmesh Nerlige Ramappa 		dt = ktime_get() - start;
2340b64e2e4SUmesh Nerlige Ramappa 		if (dt > 10000000) {
23577cdd054SUmesh Nerlige Ramappa 			pr_err("active wait timed out %lld\n", dt);
23677cdd054SUmesh Nerlige Ramappa 			ENGINE_TRACE(engine, "active wait time out %lld\n", dt);
23777cdd054SUmesh Nerlige Ramappa 			return -ETIME;
23877cdd054SUmesh Nerlige Ramappa 		}
23977cdd054SUmesh Nerlige Ramappa 	}
24077cdd054SUmesh Nerlige Ramappa 
24177cdd054SUmesh Nerlige Ramappa 	return 0;
24277cdd054SUmesh Nerlige Ramappa }
24377cdd054SUmesh Nerlige Ramappa 
live_engine_busy_stats(void * arg)2441b90e4a4SChris Wilson static int live_engine_busy_stats(void *arg)
2451b90e4a4SChris Wilson {
2461b90e4a4SChris Wilson 	struct intel_gt *gt = arg;
2471b90e4a4SChris Wilson 	struct intel_engine_cs *engine;
2481b90e4a4SChris Wilson 	enum intel_engine_id id;
2491b90e4a4SChris Wilson 	struct igt_spinner spin;
2501b90e4a4SChris Wilson 	int err = 0;
2511b90e4a4SChris Wilson 
2521b90e4a4SChris Wilson 	/*
2531b90e4a4SChris Wilson 	 * Check that if an engine supports busy-stats, they tell the truth.
2541b90e4a4SChris Wilson 	 */
2551b90e4a4SChris Wilson 
2561b90e4a4SChris Wilson 	if (igt_spinner_init(&spin, gt))
2571b90e4a4SChris Wilson 		return -ENOMEM;
2581b90e4a4SChris Wilson 
2591b90e4a4SChris Wilson 	GEM_BUG_ON(intel_gt_pm_is_awake(gt));
2601b90e4a4SChris Wilson 	for_each_engine(engine, gt, id) {
2611b90e4a4SChris Wilson 		struct i915_request *rq;
26277cdd054SUmesh Nerlige Ramappa 		ktime_t busyness, dummy;
263810b7ee3SChris Wilson 		ktime_t de, dt;
264810b7ee3SChris Wilson 		ktime_t t[2];
2651b90e4a4SChris Wilson 
2661b90e4a4SChris Wilson 		if (!intel_engine_supports_stats(engine))
2671b90e4a4SChris Wilson 			continue;
2681b90e4a4SChris Wilson 
2691b90e4a4SChris Wilson 		if (!intel_engine_can_store_dword(engine))
2701b90e4a4SChris Wilson 			continue;
2711b90e4a4SChris Wilson 
2721b90e4a4SChris Wilson 		if (intel_gt_pm_wait_for_idle(gt)) {
2731b90e4a4SChris Wilson 			err = -EBUSY;
2741b90e4a4SChris Wilson 			break;
2751b90e4a4SChris Wilson 		}
2761b90e4a4SChris Wilson 
2771b90e4a4SChris Wilson 		st_engine_heartbeat_disable(engine);
2781b90e4a4SChris Wilson 
2791b90e4a4SChris Wilson 		ENGINE_TRACE(engine, "measuring idle time\n");
2801b90e4a4SChris Wilson 		preempt_disable();
281810b7ee3SChris Wilson 		de = intel_engine_get_busy_time(engine, &t[0]);
2821b90e4a4SChris Wilson 		udelay(100);
283810b7ee3SChris Wilson 		de = ktime_sub(intel_engine_get_busy_time(engine, &t[1]), de);
2841b90e4a4SChris Wilson 		preempt_enable();
285810b7ee3SChris Wilson 		dt = ktime_sub(t[1], t[0]);
2861b90e4a4SChris Wilson 		if (de < 0 || de > 10) {
2871b90e4a4SChris Wilson 			pr_err("%s: reported %lldns [%d%%] busyness while sleeping [for %lldns]\n",
2881b90e4a4SChris Wilson 			       engine->name,
2891b90e4a4SChris Wilson 			       de, (int)div64_u64(100 * de, dt), dt);
2901b90e4a4SChris Wilson 			GEM_TRACE_DUMP();
2911b90e4a4SChris Wilson 			err = -EINVAL;
2921b90e4a4SChris Wilson 			goto end;
2931b90e4a4SChris Wilson 		}
2941b90e4a4SChris Wilson 
2951b90e4a4SChris Wilson 		/* 100% busy */
2961b90e4a4SChris Wilson 		rq = igt_spinner_create_request(&spin,
2971b90e4a4SChris Wilson 						engine->kernel_context,
2981b90e4a4SChris Wilson 						MI_NOOP);
2991b90e4a4SChris Wilson 		if (IS_ERR(rq)) {
3001b90e4a4SChris Wilson 			err = PTR_ERR(rq);
3011b90e4a4SChris Wilson 			goto end;
3021b90e4a4SChris Wilson 		}
3031b90e4a4SChris Wilson 		i915_request_add(rq);
3041b90e4a4SChris Wilson 
30577cdd054SUmesh Nerlige Ramappa 		busyness = intel_engine_get_busy_time(engine, &dummy);
3061b90e4a4SChris Wilson 		if (!igt_wait_for_spinner(&spin, rq)) {
3071b90e4a4SChris Wilson 			intel_gt_set_wedged(engine->gt);
3081b90e4a4SChris Wilson 			err = -ETIME;
3091b90e4a4SChris Wilson 			goto end;
3101b90e4a4SChris Wilson 		}
3111b90e4a4SChris Wilson 
31277cdd054SUmesh Nerlige Ramappa 		err = __spin_until_busier(engine, busyness);
31377cdd054SUmesh Nerlige Ramappa 		if (err) {
31477cdd054SUmesh Nerlige Ramappa 			GEM_TRACE_DUMP();
31577cdd054SUmesh Nerlige Ramappa 			goto end;
31677cdd054SUmesh Nerlige Ramappa 		}
31777cdd054SUmesh Nerlige Ramappa 
3181b90e4a4SChris Wilson 		ENGINE_TRACE(engine, "measuring busy time\n");
3191b90e4a4SChris Wilson 		preempt_disable();
320810b7ee3SChris Wilson 		de = intel_engine_get_busy_time(engine, &t[0]);
321529d95a6SUmesh Nerlige Ramappa 		mdelay(100);
322810b7ee3SChris Wilson 		de = ktime_sub(intel_engine_get_busy_time(engine, &t[1]), de);
3231b90e4a4SChris Wilson 		preempt_enable();
324810b7ee3SChris Wilson 		dt = ktime_sub(t[1], t[0]);
3251b90e4a4SChris Wilson 		if (100 * de < 95 * dt || 95 * de > 100 * dt) {
3261b90e4a4SChris Wilson 			pr_err("%s: reported %lldns [%d%%] busyness while spinning [for %lldns]\n",
3271b90e4a4SChris Wilson 			       engine->name,
3281b90e4a4SChris Wilson 			       de, (int)div64_u64(100 * de, dt), dt);
3291b90e4a4SChris Wilson 			GEM_TRACE_DUMP();
3301b90e4a4SChris Wilson 			err = -EINVAL;
3311b90e4a4SChris Wilson 			goto end;
3321b90e4a4SChris Wilson 		}
3331b90e4a4SChris Wilson 
3341b90e4a4SChris Wilson end:
3351b90e4a4SChris Wilson 		st_engine_heartbeat_enable(engine);
3361b90e4a4SChris Wilson 		igt_spinner_end(&spin);
3371b90e4a4SChris Wilson 		if (igt_flush_test(gt->i915))
3381b90e4a4SChris Wilson 			err = -EIO;
3391b90e4a4SChris Wilson 		if (err)
3401b90e4a4SChris Wilson 			break;
3411b90e4a4SChris Wilson 	}
3421b90e4a4SChris Wilson 
3431b90e4a4SChris Wilson 	igt_spinner_fini(&spin);
3441b90e4a4SChris Wilson 	if (igt_flush_test(gt->i915))
3451b90e4a4SChris Wilson 		err = -EIO;
3461b90e4a4SChris Wilson 	return err;
3471b90e4a4SChris Wilson }
348c7302f20SChris Wilson 
live_engine_pm(void * arg)349c7302f20SChris Wilson static int live_engine_pm(void *arg)
350c7302f20SChris Wilson {
351c7302f20SChris Wilson 	struct intel_gt *gt = arg;
352c7302f20SChris Wilson 	struct intel_engine_cs *engine;
353c7302f20SChris Wilson 	enum intel_engine_id id;
354c7302f20SChris Wilson 
355c7302f20SChris Wilson 	/*
356c7302f20SChris Wilson 	 * Check we can call intel_engine_pm_put from any context. No
357c7302f20SChris Wilson 	 * failures are reported directly, but if we mess up lockdep should
358c7302f20SChris Wilson 	 * tell us.
359c7302f20SChris Wilson 	 */
360c7302f20SChris Wilson 	if (intel_gt_pm_wait_for_idle(gt)) {
361c7302f20SChris Wilson 		pr_err("Unable to flush GT pm before test\n");
362c7302f20SChris Wilson 		return -EBUSY;
363c7302f20SChris Wilson 	}
364c7302f20SChris Wilson 
365c7302f20SChris Wilson 	GEM_BUG_ON(intel_gt_pm_is_awake(gt));
3665d904e3cSTvrtko Ursulin 	for_each_engine(engine, gt, id) {
367c7302f20SChris Wilson 		const typeof(*igt_atomic_phases) *p;
368c7302f20SChris Wilson 
369c7302f20SChris Wilson 		for (p = igt_atomic_phases; p->name; p++) {
370c7302f20SChris Wilson 			/*
371c7302f20SChris Wilson 			 * Acquisition is always synchronous, except if we
372c7302f20SChris Wilson 			 * know that the engine is already awake, in which
373c7302f20SChris Wilson 			 * case we should use intel_engine_pm_get_if_awake()
374c7302f20SChris Wilson 			 * to atomically grab the wakeref.
375c7302f20SChris Wilson 			 *
376c7302f20SChris Wilson 			 * In practice,
377c7302f20SChris Wilson 			 *    intel_engine_pm_get();
378c7302f20SChris Wilson 			 *    intel_engine_pm_put();
379c7302f20SChris Wilson 			 * occurs in one thread, while simultaneously
380c7302f20SChris Wilson 			 *    intel_engine_pm_get_if_awake();
381c7302f20SChris Wilson 			 *    intel_engine_pm_put();
382c7302f20SChris Wilson 			 * occurs from atomic context in another.
383c7302f20SChris Wilson 			 */
384c7302f20SChris Wilson 			GEM_BUG_ON(intel_engine_pm_is_awake(engine));
385c7302f20SChris Wilson 			intel_engine_pm_get(engine);
386c7302f20SChris Wilson 
387c7302f20SChris Wilson 			p->critical_section_begin();
388c7302f20SChris Wilson 			if (!intel_engine_pm_get_if_awake(engine))
389c7302f20SChris Wilson 				pr_err("intel_engine_pm_get_if_awake(%s) failed under %s\n",
390c7302f20SChris Wilson 				       engine->name, p->name);
391c7302f20SChris Wilson 			else
39207779a76SChris Wilson 				intel_engine_pm_put_async(engine);
39307779a76SChris Wilson 			intel_engine_pm_put_async(engine);
394c7302f20SChris Wilson 			p->critical_section_end();
395c7302f20SChris Wilson 
39607779a76SChris Wilson 			intel_engine_pm_flush(engine);
39707779a76SChris Wilson 
398c7302f20SChris Wilson 			if (intel_engine_pm_is_awake(engine)) {
399c7302f20SChris Wilson 				pr_err("%s is still awake after flushing pm\n",
400c7302f20SChris Wilson 				       engine->name);
401c7302f20SChris Wilson 				return -EINVAL;
402c7302f20SChris Wilson 			}
403c7302f20SChris Wilson 
404c7302f20SChris Wilson 			/* gt wakeref is async (deferred to workqueue) */
405c7302f20SChris Wilson 			if (intel_gt_pm_wait_for_idle(gt)) {
406*860cf3bdSTejas Upadhyay 				gt_err(gt, "GT failed to idle\n");
407c7302f20SChris Wilson 				return -EINVAL;
408c7302f20SChris Wilson 			}
409c7302f20SChris Wilson 		}
410c7302f20SChris Wilson 	}
411c7302f20SChris Wilson 
412c7302f20SChris Wilson 	return 0;
413c7302f20SChris Wilson }
414c7302f20SChris Wilson 
live_engine_pm_selftests(struct intel_gt * gt)415c7302f20SChris Wilson int live_engine_pm_selftests(struct intel_gt *gt)
416c7302f20SChris Wilson {
417c7302f20SChris Wilson 	static const struct i915_subtest tests[] = {
4188391c9b2SChris Wilson 		SUBTEST(live_engine_timestamps),
4191b90e4a4SChris Wilson 		SUBTEST(live_engine_busy_stats),
420c7302f20SChris Wilson 		SUBTEST(live_engine_pm),
421c7302f20SChris Wilson 	};
422c7302f20SChris Wilson 
423c7302f20SChris Wilson 	return intel_gt_live_subtests(tests, gt);
424c7302f20SChris Wilson }
425