1 /*
2  * SPDX-License-Identifier: MIT
3  *
4  * Copyright © 2017 Intel Corporation
5  */
6 
7 #include <linux/prime_numbers.h>
8 #include <linux/string_helpers.h>
9 #include <linux/swap.h>
10 
11 #include "i915_selftest.h"
12 
13 #include "gem/i915_gem_internal.h"
14 #include "gem/i915_gem_lmem.h"
15 #include "gem/i915_gem_pm.h"
16 #include "gem/i915_gem_region.h"
17 
18 #include "gt/intel_gt.h"
19 
20 #include "igt_gem_utils.h"
21 #include "mock_context.h"
22 
23 #include "selftests/mock_drm.h"
24 #include "selftests/mock_gem_device.h"
25 #include "selftests/mock_region.h"
26 #include "selftests/i915_random.h"
27 
28 static struct i915_gem_context *hugepage_ctx(struct drm_i915_private *i915,
29 					     struct file *file)
30 {
31 	struct i915_gem_context *ctx = live_context(i915, file);
32 	struct i915_address_space *vm;
33 
34 	if (IS_ERR(ctx))
35 		return ctx;
36 
37 	vm = ctx->vm;
38 	if (vm)
39 		WRITE_ONCE(vm->scrub_64K, true);
40 
41 	return ctx;
42 }
43 
44 static const unsigned int page_sizes[] = {
45 	I915_GTT_PAGE_SIZE_2M,
46 	I915_GTT_PAGE_SIZE_64K,
47 	I915_GTT_PAGE_SIZE_4K,
48 };
49 
50 static unsigned int get_largest_page_size(struct drm_i915_private *i915,
51 					  u64 rem)
52 {
53 	int i;
54 
55 	for (i = 0; i < ARRAY_SIZE(page_sizes); ++i) {
56 		unsigned int page_size = page_sizes[i];
57 
58 		if (HAS_PAGE_SIZES(i915, page_size) && rem >= page_size)
59 			return page_size;
60 	}
61 
62 	return 0;
63 }
64 
65 static void huge_pages_free_pages(struct sg_table *st)
66 {
67 	struct scatterlist *sg;
68 
69 	for (sg = st->sgl; sg; sg = __sg_next(sg)) {
70 		if (sg_page(sg))
71 			__free_pages(sg_page(sg), get_order(sg->length));
72 	}
73 
74 	sg_free_table(st);
75 	kfree(st);
76 }
77 
78 static int get_huge_pages(struct drm_i915_gem_object *obj)
79 {
80 #define GFP (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY)
81 	unsigned int page_mask = obj->mm.page_mask;
82 	struct sg_table *st;
83 	struct scatterlist *sg;
84 	unsigned int sg_page_sizes;
85 	u64 rem;
86 
87 	st = kmalloc(sizeof(*st), GFP);
88 	if (!st)
89 		return -ENOMEM;
90 
91 	if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) {
92 		kfree(st);
93 		return -ENOMEM;
94 	}
95 
96 	rem = obj->base.size;
97 	sg = st->sgl;
98 	st->nents = 0;
99 	sg_page_sizes = 0;
100 
101 	/*
102 	 * Our goal here is simple, we want to greedily fill the object from
103 	 * largest to smallest page-size, while ensuring that we use *every*
104 	 * page-size as per the given page-mask.
105 	 */
106 	do {
107 		unsigned int bit = ilog2(page_mask);
108 		unsigned int page_size = BIT(bit);
109 		int order = get_order(page_size);
110 
111 		do {
112 			struct page *page;
113 
114 			GEM_BUG_ON(order >= MAX_ORDER);
115 			page = alloc_pages(GFP | __GFP_ZERO, order);
116 			if (!page)
117 				goto err;
118 
119 			sg_set_page(sg, page, page_size, 0);
120 			sg_page_sizes |= page_size;
121 			st->nents++;
122 
123 			rem -= page_size;
124 			if (!rem) {
125 				sg_mark_end(sg);
126 				break;
127 			}
128 
129 			sg = __sg_next(sg);
130 		} while ((rem - ((page_size-1) & page_mask)) >= page_size);
131 
132 		page_mask &= (page_size-1);
133 	} while (page_mask);
134 
135 	if (i915_gem_gtt_prepare_pages(obj, st))
136 		goto err;
137 
138 	GEM_BUG_ON(sg_page_sizes != obj->mm.page_mask);
139 	__i915_gem_object_set_pages(obj, st, sg_page_sizes);
140 
141 	return 0;
142 
143 err:
144 	sg_set_page(sg, NULL, 0, 0);
145 	sg_mark_end(sg);
146 	huge_pages_free_pages(st);
147 
148 	return -ENOMEM;
149 }
150 
151 static void put_huge_pages(struct drm_i915_gem_object *obj,
152 			   struct sg_table *pages)
153 {
154 	i915_gem_gtt_finish_pages(obj, pages);
155 	huge_pages_free_pages(pages);
156 
157 	obj->mm.dirty = false;
158 
159 	__start_cpu_write(obj);
160 }
161 
162 static const struct drm_i915_gem_object_ops huge_page_ops = {
163 	.name = "huge-gem",
164 	.flags = I915_GEM_OBJECT_IS_SHRINKABLE,
165 	.get_pages = get_huge_pages,
166 	.put_pages = put_huge_pages,
167 };
168 
169 static struct drm_i915_gem_object *
170 huge_pages_object(struct drm_i915_private *i915,
171 		  u64 size,
172 		  unsigned int page_mask)
173 {
174 	static struct lock_class_key lock_class;
175 	struct drm_i915_gem_object *obj;
176 	unsigned int cache_level;
177 
178 	GEM_BUG_ON(!size);
179 	GEM_BUG_ON(!IS_ALIGNED(size, BIT(__ffs(page_mask))));
180 
181 	if (size >> PAGE_SHIFT > INT_MAX)
182 		return ERR_PTR(-E2BIG);
183 
184 	if (overflows_type(size, obj->base.size))
185 		return ERR_PTR(-E2BIG);
186 
187 	obj = i915_gem_object_alloc();
188 	if (!obj)
189 		return ERR_PTR(-ENOMEM);
190 
191 	drm_gem_private_object_init(&i915->drm, &obj->base, size);
192 	i915_gem_object_init(obj, &huge_page_ops, &lock_class, 0);
193 	obj->mem_flags |= I915_BO_FLAG_STRUCT_PAGE;
194 	i915_gem_object_set_volatile(obj);
195 
196 	obj->write_domain = I915_GEM_DOMAIN_CPU;
197 	obj->read_domains = I915_GEM_DOMAIN_CPU;
198 
199 	cache_level = HAS_LLC(i915) ? I915_CACHE_LLC : I915_CACHE_NONE;
200 	i915_gem_object_set_cache_coherency(obj, cache_level);
201 
202 	obj->mm.page_mask = page_mask;
203 
204 	return obj;
205 }
206 
207 static int fake_get_huge_pages(struct drm_i915_gem_object *obj)
208 {
209 	struct drm_i915_private *i915 = to_i915(obj->base.dev);
210 	const u64 max_len = rounddown_pow_of_two(UINT_MAX);
211 	struct sg_table *st;
212 	struct scatterlist *sg;
213 	unsigned int sg_page_sizes;
214 	u64 rem;
215 
216 	st = kmalloc(sizeof(*st), GFP);
217 	if (!st)
218 		return -ENOMEM;
219 
220 	if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) {
221 		kfree(st);
222 		return -ENOMEM;
223 	}
224 
225 	/* Use optimal page sized chunks to fill in the sg table */
226 	rem = obj->base.size;
227 	sg = st->sgl;
228 	st->nents = 0;
229 	sg_page_sizes = 0;
230 	do {
231 		unsigned int page_size = get_largest_page_size(i915, rem);
232 		unsigned int len = min(page_size * div_u64(rem, page_size),
233 				       max_len);
234 
235 		GEM_BUG_ON(!page_size);
236 
237 		sg->offset = 0;
238 		sg->length = len;
239 		sg_dma_len(sg) = len;
240 		sg_dma_address(sg) = page_size;
241 
242 		sg_page_sizes |= len;
243 
244 		st->nents++;
245 
246 		rem -= len;
247 		if (!rem) {
248 			sg_mark_end(sg);
249 			break;
250 		}
251 
252 		sg = sg_next(sg);
253 	} while (1);
254 
255 	i915_sg_trim(st);
256 
257 	__i915_gem_object_set_pages(obj, st, sg_page_sizes);
258 
259 	return 0;
260 }
261 
262 static int fake_get_huge_pages_single(struct drm_i915_gem_object *obj)
263 {
264 	struct drm_i915_private *i915 = to_i915(obj->base.dev);
265 	struct sg_table *st;
266 	struct scatterlist *sg;
267 	unsigned int page_size;
268 
269 	st = kmalloc(sizeof(*st), GFP);
270 	if (!st)
271 		return -ENOMEM;
272 
273 	if (sg_alloc_table(st, 1, GFP)) {
274 		kfree(st);
275 		return -ENOMEM;
276 	}
277 
278 	sg = st->sgl;
279 	st->nents = 1;
280 
281 	page_size = get_largest_page_size(i915, obj->base.size);
282 	GEM_BUG_ON(!page_size);
283 
284 	sg->offset = 0;
285 	sg->length = obj->base.size;
286 	sg_dma_len(sg) = obj->base.size;
287 	sg_dma_address(sg) = page_size;
288 
289 	__i915_gem_object_set_pages(obj, st, sg->length);
290 
291 	return 0;
292 #undef GFP
293 }
294 
295 static void fake_free_huge_pages(struct drm_i915_gem_object *obj,
296 				 struct sg_table *pages)
297 {
298 	sg_free_table(pages);
299 	kfree(pages);
300 }
301 
302 static void fake_put_huge_pages(struct drm_i915_gem_object *obj,
303 				struct sg_table *pages)
304 {
305 	fake_free_huge_pages(obj, pages);
306 	obj->mm.dirty = false;
307 }
308 
309 static const struct drm_i915_gem_object_ops fake_ops = {
310 	.name = "fake-gem",
311 	.flags = I915_GEM_OBJECT_IS_SHRINKABLE,
312 	.get_pages = fake_get_huge_pages,
313 	.put_pages = fake_put_huge_pages,
314 };
315 
316 static const struct drm_i915_gem_object_ops fake_ops_single = {
317 	.name = "fake-gem",
318 	.flags = I915_GEM_OBJECT_IS_SHRINKABLE,
319 	.get_pages = fake_get_huge_pages_single,
320 	.put_pages = fake_put_huge_pages,
321 };
322 
323 static struct drm_i915_gem_object *
324 fake_huge_pages_object(struct drm_i915_private *i915, u64 size, bool single)
325 {
326 	static struct lock_class_key lock_class;
327 	struct drm_i915_gem_object *obj;
328 
329 	GEM_BUG_ON(!size);
330 	GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE));
331 
332 	if (size >> PAGE_SHIFT > UINT_MAX)
333 		return ERR_PTR(-E2BIG);
334 
335 	if (overflows_type(size, obj->base.size))
336 		return ERR_PTR(-E2BIG);
337 
338 	obj = i915_gem_object_alloc();
339 	if (!obj)
340 		return ERR_PTR(-ENOMEM);
341 
342 	drm_gem_private_object_init(&i915->drm, &obj->base, size);
343 
344 	if (single)
345 		i915_gem_object_init(obj, &fake_ops_single, &lock_class, 0);
346 	else
347 		i915_gem_object_init(obj, &fake_ops, &lock_class, 0);
348 
349 	i915_gem_object_set_volatile(obj);
350 
351 	obj->write_domain = I915_GEM_DOMAIN_CPU;
352 	obj->read_domains = I915_GEM_DOMAIN_CPU;
353 	obj->cache_level = I915_CACHE_NONE;
354 
355 	return obj;
356 }
357 
358 static int igt_check_page_sizes(struct i915_vma *vma)
359 {
360 	struct drm_i915_private *i915 = vma->vm->i915;
361 	unsigned int supported = INTEL_INFO(i915)->page_sizes;
362 	struct drm_i915_gem_object *obj = vma->obj;
363 	int err;
364 
365 	/* We have to wait for the async bind to complete before our asserts */
366 	err = i915_vma_sync(vma);
367 	if (err)
368 		return err;
369 
370 	if (!HAS_PAGE_SIZES(i915, vma->page_sizes.sg)) {
371 		pr_err("unsupported page_sizes.sg=%u, supported=%u\n",
372 		       vma->page_sizes.sg & ~supported, supported);
373 		err = -EINVAL;
374 	}
375 
376 	if (!HAS_PAGE_SIZES(i915, vma->resource->page_sizes_gtt)) {
377 		pr_err("unsupported page_sizes.gtt=%u, supported=%u\n",
378 		       vma->resource->page_sizes_gtt & ~supported, supported);
379 		err = -EINVAL;
380 	}
381 
382 	if (vma->page_sizes.phys != obj->mm.page_sizes.phys) {
383 		pr_err("vma->page_sizes.phys(%u) != obj->mm.page_sizes.phys(%u)\n",
384 		       vma->page_sizes.phys, obj->mm.page_sizes.phys);
385 		err = -EINVAL;
386 	}
387 
388 	if (vma->page_sizes.sg != obj->mm.page_sizes.sg) {
389 		pr_err("vma->page_sizes.sg(%u) != obj->mm.page_sizes.sg(%u)\n",
390 		       vma->page_sizes.sg, obj->mm.page_sizes.sg);
391 		err = -EINVAL;
392 	}
393 
394 	/*
395 	 * The dma-api is like a box of chocolates when it comes to the
396 	 * alignment of dma addresses, however for LMEM we have total control
397 	 * and so can guarantee alignment, likewise when we allocate our blocks
398 	 * they should appear in descending order, and if we know that we align
399 	 * to the largest page size for the GTT address, we should be able to
400 	 * assert that if we see 2M physical pages then we should also get 2M
401 	 * GTT pages. If we don't then something might be wrong in our
402 	 * construction of the backing pages.
403 	 *
404 	 * Maintaining alignment is required to utilise huge pages in the ppGGT.
405 	 */
406 	if (i915_gem_object_is_lmem(obj) &&
407 	    IS_ALIGNED(vma->node.start, SZ_2M) &&
408 	    vma->page_sizes.sg & SZ_2M &&
409 	    vma->resource->page_sizes_gtt < SZ_2M) {
410 		pr_err("gtt pages mismatch for LMEM, expected 2M GTT pages, sg(%u), gtt(%u)\n",
411 		       vma->page_sizes.sg, vma->resource->page_sizes_gtt);
412 		err = -EINVAL;
413 	}
414 
415 	return err;
416 }
417 
418 static int igt_mock_exhaust_device_supported_pages(void *arg)
419 {
420 	struct i915_ppgtt *ppgtt = arg;
421 	struct drm_i915_private *i915 = ppgtt->vm.i915;
422 	unsigned int saved_mask = INTEL_INFO(i915)->page_sizes;
423 	struct drm_i915_gem_object *obj;
424 	struct i915_vma *vma;
425 	int i, j, single;
426 	int err;
427 
428 	/*
429 	 * Sanity check creating objects with every valid page support
430 	 * combination for our mock device.
431 	 */
432 
433 	for (i = 1; i < BIT(ARRAY_SIZE(page_sizes)); i++) {
434 		unsigned int combination = SZ_4K; /* Required for ppGTT */
435 
436 		for (j = 0; j < ARRAY_SIZE(page_sizes); j++) {
437 			if (i & BIT(j))
438 				combination |= page_sizes[j];
439 		}
440 
441 		mkwrite_device_info(i915)->page_sizes = combination;
442 
443 		for (single = 0; single <= 1; ++single) {
444 			obj = fake_huge_pages_object(i915, combination, !!single);
445 			if (IS_ERR(obj)) {
446 				err = PTR_ERR(obj);
447 				goto out_device;
448 			}
449 
450 			if (obj->base.size != combination) {
451 				pr_err("obj->base.size=%zu, expected=%u\n",
452 				       obj->base.size, combination);
453 				err = -EINVAL;
454 				goto out_put;
455 			}
456 
457 			vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
458 			if (IS_ERR(vma)) {
459 				err = PTR_ERR(vma);
460 				goto out_put;
461 			}
462 
463 			err = i915_vma_pin(vma, 0, 0, PIN_USER);
464 			if (err)
465 				goto out_put;
466 
467 			err = igt_check_page_sizes(vma);
468 
469 			if (vma->page_sizes.sg != combination) {
470 				pr_err("page_sizes.sg=%u, expected=%u\n",
471 				       vma->page_sizes.sg, combination);
472 				err = -EINVAL;
473 			}
474 
475 			i915_vma_unpin(vma);
476 			i915_gem_object_put(obj);
477 
478 			if (err)
479 				goto out_device;
480 		}
481 	}
482 
483 	goto out_device;
484 
485 out_put:
486 	i915_gem_object_put(obj);
487 out_device:
488 	mkwrite_device_info(i915)->page_sizes = saved_mask;
489 
490 	return err;
491 }
492 
493 static int igt_mock_memory_region_huge_pages(void *arg)
494 {
495 	const unsigned int flags[] = { 0, I915_BO_ALLOC_CONTIGUOUS };
496 	struct i915_ppgtt *ppgtt = arg;
497 	struct drm_i915_private *i915 = ppgtt->vm.i915;
498 	unsigned long supported = INTEL_INFO(i915)->page_sizes;
499 	struct intel_memory_region *mem;
500 	struct drm_i915_gem_object *obj;
501 	struct i915_vma *vma;
502 	int bit;
503 	int err = 0;
504 
505 	mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
506 	if (IS_ERR(mem)) {
507 		pr_err("%s failed to create memory region\n", __func__);
508 		return PTR_ERR(mem);
509 	}
510 
511 	for_each_set_bit(bit, &supported, ilog2(I915_GTT_MAX_PAGE_SIZE) + 1) {
512 		unsigned int page_size = BIT(bit);
513 		resource_size_t phys;
514 		int i;
515 
516 		for (i = 0; i < ARRAY_SIZE(flags); ++i) {
517 			obj = i915_gem_object_create_region(mem,
518 							    page_size, page_size,
519 							    flags[i]);
520 			if (IS_ERR(obj)) {
521 				err = PTR_ERR(obj);
522 				goto out_region;
523 			}
524 
525 			vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
526 			if (IS_ERR(vma)) {
527 				err = PTR_ERR(vma);
528 				goto out_put;
529 			}
530 
531 			err = i915_vma_pin(vma, 0, 0, PIN_USER);
532 			if (err)
533 				goto out_put;
534 
535 			err = igt_check_page_sizes(vma);
536 			if (err)
537 				goto out_unpin;
538 
539 			phys = i915_gem_object_get_dma_address(obj, 0);
540 			if (!IS_ALIGNED(phys, page_size)) {
541 				pr_err("%s addr misaligned(%pa) page_size=%u\n",
542 				       __func__, &phys, page_size);
543 				err = -EINVAL;
544 				goto out_unpin;
545 			}
546 
547 			if (vma->resource->page_sizes_gtt != page_size) {
548 				pr_err("%s page_sizes.gtt=%u, expected=%u\n",
549 				       __func__, vma->resource->page_sizes_gtt,
550 				       page_size);
551 				err = -EINVAL;
552 				goto out_unpin;
553 			}
554 
555 			i915_vma_unpin(vma);
556 			__i915_gem_object_put_pages(obj);
557 			i915_gem_object_put(obj);
558 		}
559 	}
560 
561 	goto out_region;
562 
563 out_unpin:
564 	i915_vma_unpin(vma);
565 out_put:
566 	i915_gem_object_put(obj);
567 out_region:
568 	intel_memory_region_destroy(mem);
569 	return err;
570 }
571 
572 static int igt_mock_ppgtt_misaligned_dma(void *arg)
573 {
574 	struct i915_ppgtt *ppgtt = arg;
575 	struct drm_i915_private *i915 = ppgtt->vm.i915;
576 	unsigned long supported = INTEL_INFO(i915)->page_sizes;
577 	struct drm_i915_gem_object *obj;
578 	int bit;
579 	int err;
580 
581 	/*
582 	 * Sanity check dma misalignment for huge pages -- the dma addresses we
583 	 * insert into the paging structures need to always respect the page
584 	 * size alignment.
585 	 */
586 
587 	bit = ilog2(I915_GTT_PAGE_SIZE_64K);
588 
589 	for_each_set_bit_from(bit, &supported,
590 			      ilog2(I915_GTT_MAX_PAGE_SIZE) + 1) {
591 		IGT_TIMEOUT(end_time);
592 		unsigned int page_size = BIT(bit);
593 		unsigned int flags = PIN_USER | PIN_OFFSET_FIXED;
594 		unsigned int offset;
595 		unsigned int size =
596 			round_up(page_size, I915_GTT_PAGE_SIZE_2M) << 1;
597 		struct i915_vma *vma;
598 
599 		obj = fake_huge_pages_object(i915, size, true);
600 		if (IS_ERR(obj))
601 			return PTR_ERR(obj);
602 
603 		if (obj->base.size != size) {
604 			pr_err("obj->base.size=%zu, expected=%u\n",
605 			       obj->base.size, size);
606 			err = -EINVAL;
607 			goto out_put;
608 		}
609 
610 		err = i915_gem_object_pin_pages_unlocked(obj);
611 		if (err)
612 			goto out_put;
613 
614 		/* Force the page size for this object */
615 		obj->mm.page_sizes.sg = page_size;
616 
617 		vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
618 		if (IS_ERR(vma)) {
619 			err = PTR_ERR(vma);
620 			goto out_unpin;
621 		}
622 
623 		err = i915_vma_pin(vma, 0, 0, flags);
624 		if (err)
625 			goto out_unpin;
626 
627 
628 		err = igt_check_page_sizes(vma);
629 
630 		if (vma->resource->page_sizes_gtt != page_size) {
631 			pr_err("page_sizes.gtt=%u, expected %u\n",
632 			       vma->resource->page_sizes_gtt, page_size);
633 			err = -EINVAL;
634 		}
635 
636 		i915_vma_unpin(vma);
637 
638 		if (err)
639 			goto out_unpin;
640 
641 		/*
642 		 * Try all the other valid offsets until the next
643 		 * boundary -- should always fall back to using 4K
644 		 * pages.
645 		 */
646 		for (offset = 4096; offset < page_size; offset += 4096) {
647 			err = i915_vma_unbind_unlocked(vma);
648 			if (err)
649 				goto out_unpin;
650 
651 			err = i915_vma_pin(vma, 0, 0, flags | offset);
652 			if (err)
653 				goto out_unpin;
654 
655 			err = igt_check_page_sizes(vma);
656 
657 			if (vma->resource->page_sizes_gtt != I915_GTT_PAGE_SIZE_4K) {
658 				pr_err("page_sizes.gtt=%u, expected %llu\n",
659 				       vma->resource->page_sizes_gtt,
660 				       I915_GTT_PAGE_SIZE_4K);
661 				err = -EINVAL;
662 			}
663 
664 			i915_vma_unpin(vma);
665 
666 			if (err)
667 				goto out_unpin;
668 
669 			if (igt_timeout(end_time,
670 					"%s timed out at offset %x with page-size %x\n",
671 					__func__, offset, page_size))
672 				break;
673 		}
674 
675 		i915_gem_object_lock(obj, NULL);
676 		i915_gem_object_unpin_pages(obj);
677 		__i915_gem_object_put_pages(obj);
678 		i915_gem_object_unlock(obj);
679 		i915_gem_object_put(obj);
680 	}
681 
682 	return 0;
683 
684 out_unpin:
685 	i915_gem_object_lock(obj, NULL);
686 	i915_gem_object_unpin_pages(obj);
687 	i915_gem_object_unlock(obj);
688 out_put:
689 	i915_gem_object_put(obj);
690 
691 	return err;
692 }
693 
694 static void close_object_list(struct list_head *objects,
695 			      struct i915_ppgtt *ppgtt)
696 {
697 	struct drm_i915_gem_object *obj, *on;
698 
699 	list_for_each_entry_safe(obj, on, objects, st_link) {
700 		list_del(&obj->st_link);
701 		i915_gem_object_lock(obj, NULL);
702 		i915_gem_object_unpin_pages(obj);
703 		__i915_gem_object_put_pages(obj);
704 		i915_gem_object_unlock(obj);
705 		i915_gem_object_put(obj);
706 	}
707 }
708 
709 static int igt_mock_ppgtt_huge_fill(void *arg)
710 {
711 	struct i915_ppgtt *ppgtt = arg;
712 	struct drm_i915_private *i915 = ppgtt->vm.i915;
713 	unsigned long max_pages = ppgtt->vm.total >> PAGE_SHIFT;
714 	unsigned long page_num;
715 	bool single = false;
716 	LIST_HEAD(objects);
717 	IGT_TIMEOUT(end_time);
718 	int err = -ENODEV;
719 
720 	for_each_prime_number_from(page_num, 1, max_pages) {
721 		struct drm_i915_gem_object *obj;
722 		u64 size = page_num << PAGE_SHIFT;
723 		struct i915_vma *vma;
724 		unsigned int expected_gtt = 0;
725 		int i;
726 
727 		obj = fake_huge_pages_object(i915, size, single);
728 		if (IS_ERR(obj)) {
729 			err = PTR_ERR(obj);
730 			break;
731 		}
732 
733 		if (obj->base.size != size) {
734 			pr_err("obj->base.size=%zd, expected=%llu\n",
735 			       obj->base.size, size);
736 			i915_gem_object_put(obj);
737 			err = -EINVAL;
738 			break;
739 		}
740 
741 		err = i915_gem_object_pin_pages_unlocked(obj);
742 		if (err) {
743 			i915_gem_object_put(obj);
744 			break;
745 		}
746 
747 		list_add(&obj->st_link, &objects);
748 
749 		vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
750 		if (IS_ERR(vma)) {
751 			err = PTR_ERR(vma);
752 			break;
753 		}
754 
755 		err = i915_vma_pin(vma, 0, 0, PIN_USER);
756 		if (err)
757 			break;
758 
759 		err = igt_check_page_sizes(vma);
760 		if (err) {
761 			i915_vma_unpin(vma);
762 			break;
763 		}
764 
765 		/*
766 		 * Figure out the expected gtt page size knowing that we go from
767 		 * largest to smallest page size sg chunks, and that we align to
768 		 * the largest page size.
769 		 */
770 		for (i = 0; i < ARRAY_SIZE(page_sizes); ++i) {
771 			unsigned int page_size = page_sizes[i];
772 
773 			if (HAS_PAGE_SIZES(i915, page_size) &&
774 			    size >= page_size) {
775 				expected_gtt |= page_size;
776 				size &= page_size-1;
777 			}
778 		}
779 
780 		GEM_BUG_ON(!expected_gtt);
781 		GEM_BUG_ON(size);
782 
783 		if (expected_gtt & I915_GTT_PAGE_SIZE_4K)
784 			expected_gtt &= ~I915_GTT_PAGE_SIZE_64K;
785 
786 		i915_vma_unpin(vma);
787 
788 		if (vma->page_sizes.sg & I915_GTT_PAGE_SIZE_64K) {
789 			if (!IS_ALIGNED(vma->node.start,
790 					I915_GTT_PAGE_SIZE_2M)) {
791 				pr_err("node.start(%llx) not aligned to 2M\n",
792 				       vma->node.start);
793 				err = -EINVAL;
794 				break;
795 			}
796 
797 			if (!IS_ALIGNED(vma->node.size,
798 					I915_GTT_PAGE_SIZE_2M)) {
799 				pr_err("node.size(%llx) not aligned to 2M\n",
800 				       vma->node.size);
801 				err = -EINVAL;
802 				break;
803 			}
804 		}
805 
806 		if (vma->resource->page_sizes_gtt != expected_gtt) {
807 			pr_err("gtt=%u, expected=%u, size=%zd, single=%s\n",
808 			       vma->resource->page_sizes_gtt, expected_gtt,
809 			       obj->base.size, str_yes_no(!!single));
810 			err = -EINVAL;
811 			break;
812 		}
813 
814 		if (igt_timeout(end_time,
815 				"%s timed out at size %zd\n",
816 				__func__, obj->base.size))
817 			break;
818 
819 		single = !single;
820 	}
821 
822 	close_object_list(&objects, ppgtt);
823 
824 	if (err == -ENOMEM || err == -ENOSPC)
825 		err = 0;
826 
827 	return err;
828 }
829 
830 static int igt_mock_ppgtt_64K(void *arg)
831 {
832 	struct i915_ppgtt *ppgtt = arg;
833 	struct drm_i915_private *i915 = ppgtt->vm.i915;
834 	struct drm_i915_gem_object *obj;
835 	const struct object_info {
836 		unsigned int size;
837 		unsigned int gtt;
838 		unsigned int offset;
839 	} objects[] = {
840 		/* Cases with forced padding/alignment */
841 		{
842 			.size = SZ_64K,
843 			.gtt = I915_GTT_PAGE_SIZE_64K,
844 			.offset = 0,
845 		},
846 		{
847 			.size = SZ_64K + SZ_4K,
848 			.gtt = I915_GTT_PAGE_SIZE_4K,
849 			.offset = 0,
850 		},
851 		{
852 			.size = SZ_64K - SZ_4K,
853 			.gtt = I915_GTT_PAGE_SIZE_4K,
854 			.offset = 0,
855 		},
856 		{
857 			.size = SZ_2M,
858 			.gtt = I915_GTT_PAGE_SIZE_64K,
859 			.offset = 0,
860 		},
861 		{
862 			.size = SZ_2M - SZ_4K,
863 			.gtt = I915_GTT_PAGE_SIZE_4K,
864 			.offset = 0,
865 		},
866 		{
867 			.size = SZ_2M + SZ_4K,
868 			.gtt = I915_GTT_PAGE_SIZE_64K | I915_GTT_PAGE_SIZE_4K,
869 			.offset = 0,
870 		},
871 		{
872 			.size = SZ_2M + SZ_64K,
873 			.gtt = I915_GTT_PAGE_SIZE_64K,
874 			.offset = 0,
875 		},
876 		{
877 			.size = SZ_2M - SZ_64K,
878 			.gtt = I915_GTT_PAGE_SIZE_64K,
879 			.offset = 0,
880 		},
881 		/* Try without any forced padding/alignment */
882 		{
883 			.size = SZ_64K,
884 			.offset = SZ_2M,
885 			.gtt = I915_GTT_PAGE_SIZE_4K,
886 		},
887 		{
888 			.size = SZ_128K,
889 			.offset = SZ_2M - SZ_64K,
890 			.gtt = I915_GTT_PAGE_SIZE_4K,
891 		},
892 	};
893 	struct i915_vma *vma;
894 	int i, single;
895 	int err;
896 
897 	/*
898 	 * Sanity check some of the trickiness with 64K pages -- either we can
899 	 * safely mark the whole page-table(2M block) as 64K, or we have to
900 	 * always fallback to 4K.
901 	 */
902 
903 	if (!HAS_PAGE_SIZES(i915, I915_GTT_PAGE_SIZE_64K))
904 		return 0;
905 
906 	for (i = 0; i < ARRAY_SIZE(objects); ++i) {
907 		unsigned int size = objects[i].size;
908 		unsigned int expected_gtt = objects[i].gtt;
909 		unsigned int offset = objects[i].offset;
910 		unsigned int flags = PIN_USER;
911 
912 		for (single = 0; single <= 1; single++) {
913 			obj = fake_huge_pages_object(i915, size, !!single);
914 			if (IS_ERR(obj))
915 				return PTR_ERR(obj);
916 
917 			err = i915_gem_object_pin_pages_unlocked(obj);
918 			if (err)
919 				goto out_object_put;
920 
921 			/*
922 			 * Disable 2M pages -- We only want to use 64K/4K pages
923 			 * for this test.
924 			 */
925 			obj->mm.page_sizes.sg &= ~I915_GTT_PAGE_SIZE_2M;
926 
927 			vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
928 			if (IS_ERR(vma)) {
929 				err = PTR_ERR(vma);
930 				goto out_object_unpin;
931 			}
932 
933 			if (offset)
934 				flags |= PIN_OFFSET_FIXED | offset;
935 
936 			err = i915_vma_pin(vma, 0, 0, flags);
937 			if (err)
938 				goto out_object_unpin;
939 
940 			err = igt_check_page_sizes(vma);
941 			if (err)
942 				goto out_vma_unpin;
943 
944 			if (!offset && vma->page_sizes.sg & I915_GTT_PAGE_SIZE_64K) {
945 				if (!IS_ALIGNED(vma->node.start,
946 						I915_GTT_PAGE_SIZE_2M)) {
947 					pr_err("node.start(%llx) not aligned to 2M\n",
948 					       vma->node.start);
949 					err = -EINVAL;
950 					goto out_vma_unpin;
951 				}
952 
953 				if (!IS_ALIGNED(vma->node.size,
954 						I915_GTT_PAGE_SIZE_2M)) {
955 					pr_err("node.size(%llx) not aligned to 2M\n",
956 					       vma->node.size);
957 					err = -EINVAL;
958 					goto out_vma_unpin;
959 				}
960 			}
961 
962 			if (vma->resource->page_sizes_gtt != expected_gtt) {
963 				pr_err("gtt=%u, expected=%u, i=%d, single=%s\n",
964 				       vma->resource->page_sizes_gtt,
965 				       expected_gtt, i, str_yes_no(!!single));
966 				err = -EINVAL;
967 				goto out_vma_unpin;
968 			}
969 
970 			i915_vma_unpin(vma);
971 			i915_gem_object_lock(obj, NULL);
972 			i915_gem_object_unpin_pages(obj);
973 			__i915_gem_object_put_pages(obj);
974 			i915_gem_object_unlock(obj);
975 			i915_gem_object_put(obj);
976 
977 			i915_gem_drain_freed_objects(i915);
978 		}
979 	}
980 
981 	return 0;
982 
983 out_vma_unpin:
984 	i915_vma_unpin(vma);
985 out_object_unpin:
986 	i915_gem_object_lock(obj, NULL);
987 	i915_gem_object_unpin_pages(obj);
988 	i915_gem_object_unlock(obj);
989 out_object_put:
990 	i915_gem_object_put(obj);
991 
992 	return err;
993 }
994 
995 static int gpu_write(struct intel_context *ce,
996 		     struct i915_vma *vma,
997 		     u32 dw,
998 		     u32 val)
999 {
1000 	int err;
1001 
1002 	i915_gem_object_lock(vma->obj, NULL);
1003 	err = i915_gem_object_set_to_gtt_domain(vma->obj, true);
1004 	i915_gem_object_unlock(vma->obj);
1005 	if (err)
1006 		return err;
1007 
1008 	return igt_gpu_fill_dw(ce, vma, dw * sizeof(u32),
1009 			       vma->size >> PAGE_SHIFT, val);
1010 }
1011 
1012 static int
1013 __cpu_check_shmem(struct drm_i915_gem_object *obj, u32 dword, u32 val)
1014 {
1015 	unsigned int needs_flush;
1016 	unsigned long n;
1017 	int err;
1018 
1019 	i915_gem_object_lock(obj, NULL);
1020 	err = i915_gem_object_prepare_read(obj, &needs_flush);
1021 	if (err)
1022 		goto err_unlock;
1023 
1024 	for (n = 0; n < obj->base.size >> PAGE_SHIFT; ++n) {
1025 		u32 *ptr = kmap_atomic(i915_gem_object_get_page(obj, n));
1026 
1027 		if (needs_flush & CLFLUSH_BEFORE)
1028 			drm_clflush_virt_range(ptr, PAGE_SIZE);
1029 
1030 		if (ptr[dword] != val) {
1031 			pr_err("n=%lu ptr[%u]=%u, val=%u\n",
1032 			       n, dword, ptr[dword], val);
1033 			kunmap_atomic(ptr);
1034 			err = -EINVAL;
1035 			break;
1036 		}
1037 
1038 		kunmap_atomic(ptr);
1039 	}
1040 
1041 	i915_gem_object_finish_access(obj);
1042 err_unlock:
1043 	i915_gem_object_unlock(obj);
1044 
1045 	return err;
1046 }
1047 
1048 static int __cpu_check_vmap(struct drm_i915_gem_object *obj, u32 dword, u32 val)
1049 {
1050 	unsigned long n = obj->base.size >> PAGE_SHIFT;
1051 	u32 *ptr;
1052 	int err;
1053 
1054 	err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
1055 	if (err)
1056 		return err;
1057 
1058 	ptr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
1059 	if (IS_ERR(ptr))
1060 		return PTR_ERR(ptr);
1061 
1062 	ptr += dword;
1063 	while (n--) {
1064 		if (*ptr != val) {
1065 			pr_err("base[%u]=%08x, val=%08x\n",
1066 			       dword, *ptr, val);
1067 			err = -EINVAL;
1068 			break;
1069 		}
1070 
1071 		ptr += PAGE_SIZE / sizeof(*ptr);
1072 	}
1073 
1074 	i915_gem_object_unpin_map(obj);
1075 	return err;
1076 }
1077 
1078 static int cpu_check(struct drm_i915_gem_object *obj, u32 dword, u32 val)
1079 {
1080 	if (i915_gem_object_has_struct_page(obj))
1081 		return __cpu_check_shmem(obj, dword, val);
1082 	else
1083 		return __cpu_check_vmap(obj, dword, val);
1084 }
1085 
1086 static int __igt_write_huge(struct intel_context *ce,
1087 			    struct drm_i915_gem_object *obj,
1088 			    u64 size, u64 offset,
1089 			    u32 dword, u32 val)
1090 {
1091 	unsigned int flags = PIN_USER | PIN_OFFSET_FIXED;
1092 	struct i915_vma *vma;
1093 	int err;
1094 
1095 	vma = i915_vma_instance(obj, ce->vm, NULL);
1096 	if (IS_ERR(vma))
1097 		return PTR_ERR(vma);
1098 
1099 	err = i915_vma_pin(vma, size, 0, flags | offset);
1100 	if (err) {
1101 		/*
1102 		 * The ggtt may have some pages reserved so
1103 		 * refrain from erroring out.
1104 		 */
1105 		if (err == -ENOSPC && i915_is_ggtt(ce->vm))
1106 			err = 0;
1107 
1108 		return err;
1109 	}
1110 
1111 	err = igt_check_page_sizes(vma);
1112 	if (err)
1113 		goto out_vma_unpin;
1114 
1115 	err = gpu_write(ce, vma, dword, val);
1116 	if (err) {
1117 		pr_err("gpu-write failed at offset=%llx\n", offset);
1118 		goto out_vma_unpin;
1119 	}
1120 
1121 	err = cpu_check(obj, dword, val);
1122 	if (err) {
1123 		pr_err("cpu-check failed at offset=%llx\n", offset);
1124 		goto out_vma_unpin;
1125 	}
1126 
1127 out_vma_unpin:
1128 	i915_vma_unpin(vma);
1129 	return err;
1130 }
1131 
1132 static int igt_write_huge(struct drm_i915_private *i915,
1133 			  struct drm_i915_gem_object *obj)
1134 {
1135 	struct i915_gem_engines *engines;
1136 	struct i915_gem_engines_iter it;
1137 	struct intel_context *ce;
1138 	I915_RND_STATE(prng);
1139 	IGT_TIMEOUT(end_time);
1140 	unsigned int max_page_size;
1141 	unsigned int count;
1142 	struct i915_gem_context *ctx;
1143 	struct file *file;
1144 	u64 max;
1145 	u64 num;
1146 	u64 size;
1147 	int *order;
1148 	int i, n;
1149 	int err = 0;
1150 
1151 	file = mock_file(i915);
1152 	if (IS_ERR(file))
1153 		return PTR_ERR(file);
1154 
1155 	ctx = hugepage_ctx(i915, file);
1156 	if (IS_ERR(ctx)) {
1157 		err = PTR_ERR(ctx);
1158 		goto out;
1159 	}
1160 
1161 	GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
1162 
1163 	size = obj->base.size;
1164 	if (obj->mm.page_sizes.sg & I915_GTT_PAGE_SIZE_64K)
1165 		size = round_up(size, I915_GTT_PAGE_SIZE_2M);
1166 
1167 	n = 0;
1168 	count = 0;
1169 	max = U64_MAX;
1170 	for_each_gem_engine(ce, i915_gem_context_lock_engines(ctx), it) {
1171 		count++;
1172 		if (!intel_engine_can_store_dword(ce->engine))
1173 			continue;
1174 
1175 		max = min(max, ce->vm->total);
1176 		n++;
1177 	}
1178 	i915_gem_context_unlock_engines(ctx);
1179 	if (!n)
1180 		goto out;
1181 
1182 	/*
1183 	 * To keep things interesting when alternating between engines in our
1184 	 * randomized order, lets also make feeding to the same engine a few
1185 	 * times in succession a possibility by enlarging the permutation array.
1186 	 */
1187 	order = i915_random_order(count * count, &prng);
1188 	if (!order)
1189 		return -ENOMEM;
1190 
1191 	max_page_size = rounddown_pow_of_two(obj->mm.page_sizes.sg);
1192 	max = div_u64(max - size, max_page_size);
1193 
1194 	/*
1195 	 * Try various offsets in an ascending/descending fashion until we
1196 	 * timeout -- we want to avoid issues hidden by effectively always using
1197 	 * offset = 0.
1198 	 */
1199 	i = 0;
1200 	engines = i915_gem_context_lock_engines(ctx);
1201 	for_each_prime_number_from(num, 0, max) {
1202 		u64 offset_low = num * max_page_size;
1203 		u64 offset_high = (max - num) * max_page_size;
1204 		u32 dword = offset_in_page(num) / 4;
1205 		struct intel_context *ce;
1206 
1207 		ce = engines->engines[order[i] % engines->num_engines];
1208 		i = (i + 1) % (count * count);
1209 		if (!ce || !intel_engine_can_store_dword(ce->engine))
1210 			continue;
1211 
1212 		/*
1213 		 * In order to utilize 64K pages we need to both pad the vma
1214 		 * size and ensure the vma offset is at the start of the pt
1215 		 * boundary, however to improve coverage we opt for testing both
1216 		 * aligned and unaligned offsets.
1217 		 */
1218 		if (obj->mm.page_sizes.sg & I915_GTT_PAGE_SIZE_64K)
1219 			offset_low = round_down(offset_low,
1220 						I915_GTT_PAGE_SIZE_2M);
1221 
1222 		err = __igt_write_huge(ce, obj, size, offset_low,
1223 				       dword, num + 1);
1224 		if (err)
1225 			break;
1226 
1227 		err = __igt_write_huge(ce, obj, size, offset_high,
1228 				       dword, num + 1);
1229 		if (err)
1230 			break;
1231 
1232 		if (igt_timeout(end_time,
1233 				"%s timed out on %s, offset_low=%llx offset_high=%llx, max_page_size=%x\n",
1234 				__func__, ce->engine->name, offset_low, offset_high,
1235 				max_page_size))
1236 			break;
1237 	}
1238 	i915_gem_context_unlock_engines(ctx);
1239 
1240 	kfree(order);
1241 
1242 out:
1243 	fput(file);
1244 	return err;
1245 }
1246 
1247 typedef struct drm_i915_gem_object *
1248 (*igt_create_fn)(struct drm_i915_private *i915, u32 size, u32 flags);
1249 
1250 static inline bool igt_can_allocate_thp(struct drm_i915_private *i915)
1251 {
1252 	return i915->mm.gemfs && has_transparent_hugepage();
1253 }
1254 
1255 static struct drm_i915_gem_object *
1256 igt_create_shmem(struct drm_i915_private *i915, u32 size, u32 flags)
1257 {
1258 	if (!igt_can_allocate_thp(i915)) {
1259 		pr_info("%s missing THP support, skipping\n", __func__);
1260 		return ERR_PTR(-ENODEV);
1261 	}
1262 
1263 	return i915_gem_object_create_shmem(i915, size);
1264 }
1265 
1266 static struct drm_i915_gem_object *
1267 igt_create_internal(struct drm_i915_private *i915, u32 size, u32 flags)
1268 {
1269 	return i915_gem_object_create_internal(i915, size);
1270 }
1271 
1272 static struct drm_i915_gem_object *
1273 igt_create_system(struct drm_i915_private *i915, u32 size, u32 flags)
1274 {
1275 	return huge_pages_object(i915, size, size);
1276 }
1277 
1278 static struct drm_i915_gem_object *
1279 igt_create_local(struct drm_i915_private *i915, u32 size, u32 flags)
1280 {
1281 	return i915_gem_object_create_lmem(i915, size, flags);
1282 }
1283 
1284 static u32 igt_random_size(struct rnd_state *prng,
1285 			   u32 min_page_size,
1286 			   u32 max_page_size)
1287 {
1288 	u64 mask;
1289 	u32 size;
1290 
1291 	GEM_BUG_ON(!is_power_of_2(min_page_size));
1292 	GEM_BUG_ON(!is_power_of_2(max_page_size));
1293 	GEM_BUG_ON(min_page_size < PAGE_SIZE);
1294 	GEM_BUG_ON(min_page_size > max_page_size);
1295 
1296 	mask = ((max_page_size << 1ULL) - 1) & PAGE_MASK;
1297 	size = prandom_u32_state(prng) & mask;
1298 	if (size < min_page_size)
1299 		size |= min_page_size;
1300 
1301 	return size;
1302 }
1303 
1304 static int igt_ppgtt_smoke_huge(void *arg)
1305 {
1306 	struct drm_i915_private *i915 = arg;
1307 	struct drm_i915_gem_object *obj;
1308 	I915_RND_STATE(prng);
1309 	struct {
1310 		igt_create_fn fn;
1311 		u32 min;
1312 		u32 max;
1313 	} backends[] = {
1314 		{ igt_create_internal, SZ_64K, SZ_2M,  },
1315 		{ igt_create_shmem,    SZ_64K, SZ_32M, },
1316 		{ igt_create_local,    SZ_64K, SZ_1G,  },
1317 	};
1318 	int err;
1319 	int i;
1320 
1321 	/*
1322 	 * Sanity check that the HW uses huge pages correctly through our
1323 	 * various backends -- ensure that our writes land in the right place.
1324 	 */
1325 
1326 	for (i = 0; i < ARRAY_SIZE(backends); ++i) {
1327 		u32 min = backends[i].min;
1328 		u32 max = backends[i].max;
1329 		u32 size = max;
1330 
1331 try_again:
1332 		size = igt_random_size(&prng, min, rounddown_pow_of_two(size));
1333 
1334 		obj = backends[i].fn(i915, size, 0);
1335 		if (IS_ERR(obj)) {
1336 			err = PTR_ERR(obj);
1337 			if (err == -E2BIG) {
1338 				size >>= 1;
1339 				goto try_again;
1340 			} else if (err == -ENODEV) {
1341 				err = 0;
1342 				continue;
1343 			}
1344 
1345 			return err;
1346 		}
1347 
1348 		err = i915_gem_object_pin_pages_unlocked(obj);
1349 		if (err) {
1350 			if (err == -ENXIO || err == -E2BIG || err == -ENOMEM) {
1351 				i915_gem_object_put(obj);
1352 				size >>= 1;
1353 				goto try_again;
1354 			}
1355 			goto out_put;
1356 		}
1357 
1358 		if (obj->mm.page_sizes.phys < min) {
1359 			pr_info("%s unable to allocate huge-page(s) with size=%u, i=%d\n",
1360 				__func__, size, i);
1361 			err = -ENOMEM;
1362 			goto out_unpin;
1363 		}
1364 
1365 		err = igt_write_huge(i915, obj);
1366 		if (err) {
1367 			pr_err("%s write-huge failed with size=%u, i=%d\n",
1368 			       __func__, size, i);
1369 		}
1370 out_unpin:
1371 		i915_gem_object_lock(obj, NULL);
1372 		i915_gem_object_unpin_pages(obj);
1373 		__i915_gem_object_put_pages(obj);
1374 		i915_gem_object_unlock(obj);
1375 out_put:
1376 		i915_gem_object_put(obj);
1377 
1378 		if (err == -ENOMEM || err == -ENXIO)
1379 			err = 0;
1380 
1381 		if (err)
1382 			break;
1383 
1384 		cond_resched();
1385 	}
1386 
1387 	return err;
1388 }
1389 
1390 static int igt_ppgtt_sanity_check(void *arg)
1391 {
1392 	struct drm_i915_private *i915 = arg;
1393 	unsigned int supported = INTEL_INFO(i915)->page_sizes;
1394 	struct {
1395 		igt_create_fn fn;
1396 		unsigned int flags;
1397 	} backends[] = {
1398 		{ igt_create_system, 0,                        },
1399 		{ igt_create_local,  0,                        },
1400 		{ igt_create_local,  I915_BO_ALLOC_CONTIGUOUS, },
1401 	};
1402 	struct {
1403 		u32 size;
1404 		u32 pages;
1405 	} combos[] = {
1406 		{ SZ_64K,		SZ_64K		},
1407 		{ SZ_2M,		SZ_2M		},
1408 		{ SZ_2M,		SZ_64K		},
1409 		{ SZ_2M - SZ_64K,	SZ_64K		},
1410 		{ SZ_2M - SZ_4K,	SZ_64K | SZ_4K	},
1411 		{ SZ_2M + SZ_4K,	SZ_64K | SZ_4K	},
1412 		{ SZ_2M + SZ_4K,	SZ_2M  | SZ_4K	},
1413 		{ SZ_2M + SZ_64K,	SZ_2M  | SZ_64K },
1414 	};
1415 	int i, j;
1416 	int err;
1417 
1418 	if (supported == I915_GTT_PAGE_SIZE_4K)
1419 		return 0;
1420 
1421 	/*
1422 	 * Sanity check that the HW behaves with a limited set of combinations.
1423 	 * We already have a bunch of randomised testing, which should give us
1424 	 * a decent amount of variation between runs, however we should keep
1425 	 * this to limit the chances of introducing a temporary regression, by
1426 	 * testing the most obvious cases that might make something blow up.
1427 	 */
1428 
1429 	for (i = 0; i < ARRAY_SIZE(backends); ++i) {
1430 		for (j = 0; j < ARRAY_SIZE(combos); ++j) {
1431 			struct drm_i915_gem_object *obj;
1432 			u32 size = combos[j].size;
1433 			u32 pages = combos[j].pages;
1434 
1435 			obj = backends[i].fn(i915, size, backends[i].flags);
1436 			if (IS_ERR(obj)) {
1437 				err = PTR_ERR(obj);
1438 				if (err == -ENODEV) {
1439 					pr_info("Device lacks local memory, skipping\n");
1440 					err = 0;
1441 					break;
1442 				}
1443 
1444 				return err;
1445 			}
1446 
1447 			err = i915_gem_object_pin_pages_unlocked(obj);
1448 			if (err) {
1449 				i915_gem_object_put(obj);
1450 				goto out;
1451 			}
1452 
1453 			GEM_BUG_ON(pages > obj->base.size);
1454 			pages = pages & supported;
1455 
1456 			if (pages)
1457 				obj->mm.page_sizes.sg = pages;
1458 
1459 			err = igt_write_huge(i915, obj);
1460 
1461 			i915_gem_object_lock(obj, NULL);
1462 			i915_gem_object_unpin_pages(obj);
1463 			__i915_gem_object_put_pages(obj);
1464 			i915_gem_object_unlock(obj);
1465 			i915_gem_object_put(obj);
1466 
1467 			if (err) {
1468 				pr_err("%s write-huge failed with size=%u pages=%u i=%d, j=%d\n",
1469 				       __func__, size, pages, i, j);
1470 				goto out;
1471 			}
1472 		}
1473 
1474 		cond_resched();
1475 	}
1476 
1477 out:
1478 	if (err == -ENOMEM)
1479 		err = 0;
1480 
1481 	return err;
1482 }
1483 
1484 static int igt_ppgtt_compact(void *arg)
1485 {
1486 	struct drm_i915_private *i915 = arg;
1487 	struct drm_i915_gem_object *obj;
1488 	int err;
1489 
1490 	/*
1491 	 * Simple test to catch issues with compact 64K pages -- since the pt is
1492 	 * compacted to 256B that gives us 32 entries per pt, however since the
1493 	 * backing page for the pt is 4K, any extra entries we might incorrectly
1494 	 * write out should be ignored by the HW. If ever hit such a case this
1495 	 * test should catch it since some of our writes would land in scratch.
1496 	 */
1497 
1498 	if (!HAS_64K_PAGES(i915)) {
1499 		pr_info("device lacks compact 64K page support, skipping\n");
1500 		return 0;
1501 	}
1502 
1503 	if (!HAS_LMEM(i915)) {
1504 		pr_info("device lacks LMEM support, skipping\n");
1505 		return 0;
1506 	}
1507 
1508 	/* We want the range to cover multiple page-table boundaries. */
1509 	obj = i915_gem_object_create_lmem(i915, SZ_4M, 0);
1510 	if (IS_ERR(obj))
1511 		return PTR_ERR(obj);
1512 
1513 	err = i915_gem_object_pin_pages_unlocked(obj);
1514 	if (err)
1515 		goto out_put;
1516 
1517 	if (obj->mm.page_sizes.phys < I915_GTT_PAGE_SIZE_64K) {
1518 		pr_info("LMEM compact unable to allocate huge-page(s)\n");
1519 		goto out_unpin;
1520 	}
1521 
1522 	/*
1523 	 * Disable 2M GTT pages by forcing the page-size to 64K for the GTT
1524 	 * insertion.
1525 	 */
1526 	obj->mm.page_sizes.sg = I915_GTT_PAGE_SIZE_64K;
1527 
1528 	err = igt_write_huge(i915, obj);
1529 	if (err)
1530 		pr_err("LMEM compact write-huge failed\n");
1531 
1532 out_unpin:
1533 	i915_gem_object_unpin_pages(obj);
1534 out_put:
1535 	i915_gem_object_put(obj);
1536 
1537 	if (err == -ENOMEM)
1538 		err = 0;
1539 
1540 	return err;
1541 }
1542 
1543 static int igt_tmpfs_fallback(void *arg)
1544 {
1545 	struct drm_i915_private *i915 = arg;
1546 	struct i915_address_space *vm;
1547 	struct i915_gem_context *ctx;
1548 	struct vfsmount *gemfs = i915->mm.gemfs;
1549 	struct drm_i915_gem_object *obj;
1550 	struct i915_vma *vma;
1551 	struct file *file;
1552 	u32 *vaddr;
1553 	int err = 0;
1554 
1555 	file = mock_file(i915);
1556 	if (IS_ERR(file))
1557 		return PTR_ERR(file);
1558 
1559 	ctx = hugepage_ctx(i915, file);
1560 	if (IS_ERR(ctx)) {
1561 		err = PTR_ERR(ctx);
1562 		goto out;
1563 	}
1564 	vm = i915_gem_context_get_eb_vm(ctx);
1565 
1566 	/*
1567 	 * Make sure that we don't burst into a ball of flames upon falling back
1568 	 * to tmpfs, which we rely on if on the off-chance we encouter a failure
1569 	 * when setting up gemfs.
1570 	 */
1571 
1572 	i915->mm.gemfs = NULL;
1573 
1574 	obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
1575 	if (IS_ERR(obj)) {
1576 		err = PTR_ERR(obj);
1577 		goto out_restore;
1578 	}
1579 
1580 	vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WB);
1581 	if (IS_ERR(vaddr)) {
1582 		err = PTR_ERR(vaddr);
1583 		goto out_put;
1584 	}
1585 	*vaddr = 0xdeadbeaf;
1586 
1587 	__i915_gem_object_flush_map(obj, 0, 64);
1588 	i915_gem_object_unpin_map(obj);
1589 
1590 	vma = i915_vma_instance(obj, vm, NULL);
1591 	if (IS_ERR(vma)) {
1592 		err = PTR_ERR(vma);
1593 		goto out_put;
1594 	}
1595 
1596 	err = i915_vma_pin(vma, 0, 0, PIN_USER);
1597 	if (err)
1598 		goto out_put;
1599 
1600 	err = igt_check_page_sizes(vma);
1601 
1602 	i915_vma_unpin(vma);
1603 out_put:
1604 	i915_gem_object_put(obj);
1605 out_restore:
1606 	i915->mm.gemfs = gemfs;
1607 
1608 	i915_vm_put(vm);
1609 out:
1610 	fput(file);
1611 	return err;
1612 }
1613 
1614 static int igt_shrink_thp(void *arg)
1615 {
1616 	struct drm_i915_private *i915 = arg;
1617 	struct i915_address_space *vm;
1618 	struct i915_gem_context *ctx;
1619 	struct drm_i915_gem_object *obj;
1620 	struct i915_gem_engines_iter it;
1621 	struct intel_context *ce;
1622 	struct i915_vma *vma;
1623 	struct file *file;
1624 	unsigned int flags = PIN_USER;
1625 	unsigned int n;
1626 	bool should_swap;
1627 	int err;
1628 
1629 	if (!igt_can_allocate_thp(i915)) {
1630 		pr_info("missing THP support, skipping\n");
1631 		return 0;
1632 	}
1633 
1634 	file = mock_file(i915);
1635 	if (IS_ERR(file))
1636 		return PTR_ERR(file);
1637 
1638 	ctx = hugepage_ctx(i915, file);
1639 	if (IS_ERR(ctx)) {
1640 		err = PTR_ERR(ctx);
1641 		goto out;
1642 	}
1643 	vm = i915_gem_context_get_eb_vm(ctx);
1644 
1645 	/*
1646 	 * Sanity check shrinking huge-paged object -- make sure nothing blows
1647 	 * up.
1648 	 */
1649 
1650 	obj = i915_gem_object_create_shmem(i915, SZ_2M);
1651 	if (IS_ERR(obj)) {
1652 		err = PTR_ERR(obj);
1653 		goto out_vm;
1654 	}
1655 
1656 	vma = i915_vma_instance(obj, vm, NULL);
1657 	if (IS_ERR(vma)) {
1658 		err = PTR_ERR(vma);
1659 		goto out_put;
1660 	}
1661 
1662 	err = i915_vma_pin(vma, 0, 0, flags);
1663 	if (err)
1664 		goto out_put;
1665 
1666 	if (obj->mm.page_sizes.phys < I915_GTT_PAGE_SIZE_2M) {
1667 		pr_info("failed to allocate THP, finishing test early\n");
1668 		goto out_unpin;
1669 	}
1670 
1671 	err = igt_check_page_sizes(vma);
1672 	if (err)
1673 		goto out_unpin;
1674 
1675 	n = 0;
1676 
1677 	for_each_gem_engine(ce, i915_gem_context_lock_engines(ctx), it) {
1678 		if (!intel_engine_can_store_dword(ce->engine))
1679 			continue;
1680 
1681 		err = gpu_write(ce, vma, n++, 0xdeadbeaf);
1682 		if (err)
1683 			break;
1684 	}
1685 	i915_gem_context_unlock_engines(ctx);
1686 	/*
1687 	 * Nuke everything *before* we unpin the pages so we can be reasonably
1688 	 * sure that when later checking get_nr_swap_pages() that some random
1689 	 * leftover object doesn't steal the remaining swap space.
1690 	 */
1691 	i915_gem_shrink(NULL, i915, -1UL, NULL,
1692 			I915_SHRINK_BOUND |
1693 			I915_SHRINK_UNBOUND |
1694 			I915_SHRINK_ACTIVE);
1695 	i915_vma_unpin(vma);
1696 	if (err)
1697 		goto out_put;
1698 
1699 	/*
1700 	 * Now that the pages are *unpinned* shrinking should invoke
1701 	 * shmem to truncate our pages, if we have available swap.
1702 	 */
1703 	should_swap = get_nr_swap_pages() > 0;
1704 	i915_gem_shrink(NULL, i915, -1UL, NULL,
1705 			I915_SHRINK_BOUND |
1706 			I915_SHRINK_UNBOUND |
1707 			I915_SHRINK_ACTIVE |
1708 			I915_SHRINK_WRITEBACK);
1709 	if (should_swap == i915_gem_object_has_pages(obj)) {
1710 		pr_err("unexpected pages mismatch, should_swap=%s\n",
1711 		       str_yes_no(should_swap));
1712 		err = -EINVAL;
1713 		goto out_put;
1714 	}
1715 
1716 	if (should_swap == (obj->mm.page_sizes.sg || obj->mm.page_sizes.phys)) {
1717 		pr_err("unexpected residual page-size bits, should_swap=%s\n",
1718 		       str_yes_no(should_swap));
1719 		err = -EINVAL;
1720 		goto out_put;
1721 	}
1722 
1723 	err = i915_vma_pin(vma, 0, 0, flags);
1724 	if (err)
1725 		goto out_put;
1726 
1727 	while (n--) {
1728 		err = cpu_check(obj, n, 0xdeadbeaf);
1729 		if (err)
1730 			break;
1731 	}
1732 
1733 out_unpin:
1734 	i915_vma_unpin(vma);
1735 out_put:
1736 	i915_gem_object_put(obj);
1737 out_vm:
1738 	i915_vm_put(vm);
1739 out:
1740 	fput(file);
1741 	return err;
1742 }
1743 
1744 int i915_gem_huge_page_mock_selftests(void)
1745 {
1746 	static const struct i915_subtest tests[] = {
1747 		SUBTEST(igt_mock_exhaust_device_supported_pages),
1748 		SUBTEST(igt_mock_memory_region_huge_pages),
1749 		SUBTEST(igt_mock_ppgtt_misaligned_dma),
1750 		SUBTEST(igt_mock_ppgtt_huge_fill),
1751 		SUBTEST(igt_mock_ppgtt_64K),
1752 	};
1753 	struct drm_i915_private *dev_priv;
1754 	struct i915_ppgtt *ppgtt;
1755 	int err;
1756 
1757 	dev_priv = mock_gem_device();
1758 	if (!dev_priv)
1759 		return -ENOMEM;
1760 
1761 	/* Pretend to be a device which supports the 48b PPGTT */
1762 	mkwrite_device_info(dev_priv)->ppgtt_type = INTEL_PPGTT_FULL;
1763 	mkwrite_device_info(dev_priv)->ppgtt_size = 48;
1764 
1765 	ppgtt = i915_ppgtt_create(to_gt(dev_priv), 0);
1766 	if (IS_ERR(ppgtt)) {
1767 		err = PTR_ERR(ppgtt);
1768 		goto out_unlock;
1769 	}
1770 
1771 	if (!i915_vm_is_4lvl(&ppgtt->vm)) {
1772 		pr_err("failed to create 48b PPGTT\n");
1773 		err = -EINVAL;
1774 		goto out_put;
1775 	}
1776 
1777 	/* If we were ever hit this then it's time to mock the 64K scratch */
1778 	if (!i915_vm_has_scratch_64K(&ppgtt->vm)) {
1779 		pr_err("PPGTT missing 64K scratch page\n");
1780 		err = -EINVAL;
1781 		goto out_put;
1782 	}
1783 
1784 	err = i915_subtests(tests, ppgtt);
1785 
1786 out_put:
1787 	i915_vm_put(&ppgtt->vm);
1788 out_unlock:
1789 	mock_destroy_device(dev_priv);
1790 	return err;
1791 }
1792 
1793 int i915_gem_huge_page_live_selftests(struct drm_i915_private *i915)
1794 {
1795 	static const struct i915_subtest tests[] = {
1796 		SUBTEST(igt_shrink_thp),
1797 		SUBTEST(igt_tmpfs_fallback),
1798 		SUBTEST(igt_ppgtt_smoke_huge),
1799 		SUBTEST(igt_ppgtt_sanity_check),
1800 		SUBTEST(igt_ppgtt_compact),
1801 	};
1802 
1803 	if (!HAS_PPGTT(i915)) {
1804 		pr_info("PPGTT not supported, skipping live-selftests\n");
1805 		return 0;
1806 	}
1807 
1808 	if (intel_gt_is_wedged(to_gt(i915)))
1809 		return 0;
1810 
1811 	return i915_live_subtests(tests, i915);
1812 }
1813