Merge branches 'pm-devfreq', 'pm-qos', 'pm-tools' and 'pm-docs'
[linux-2.6-block.git] / drivers / gpu / drm / i915 / selftests / intel_memory_region.c
CommitLineData
232a6eba
MA
1// SPDX-License-Identifier: MIT
2/*
3 * Copyright © 2019 Intel Corporation
4 */
5
6#include <linux/prime_numbers.h>
ba12993c 7#include <linux/sort.h>
232a6eba 8
6387a3c4
A
9#include <drm/drm_buddy.h>
10
232a6eba
MA
11#include "../i915_selftest.h"
12
13#include "mock_drm.h"
14#include "mock_gem_device.h"
15#include "mock_region.h"
16
340be48f 17#include "gem/i915_gem_context.h"
b908be54 18#include "gem/i915_gem_lmem.h"
232a6eba 19#include "gem/i915_gem_region.h"
2d45f668 20#include "gem/i915_gem_ttm.h"
340be48f 21#include "gem/selftests/igt_gem_utils.h"
232a6eba 22#include "gem/selftests/mock_context.h"
99919be7 23#include "gt/intel_engine_pm.h"
6804da20 24#include "gt/intel_engine_user.h"
b908be54 25#include "gt/intel_gt.h"
99919be7 26#include "gt/intel_migrate.h"
ba12993c 27#include "i915_memcpy.h"
d53ec322 28#include "i915_ttm_buddy_manager.h"
01377a0d 29#include "selftests/igt_flush_test.h"
2f0b97ca 30#include "selftests/i915_random.h"
232a6eba
MA
31
32static void close_objects(struct intel_memory_region *mem,
33 struct list_head *objects)
34{
35 struct drm_i915_private *i915 = mem->i915;
36 struct drm_i915_gem_object *obj, *on;
37
38 list_for_each_entry_safe(obj, on, objects, st_link) {
e70a27d7 39 i915_gem_object_lock(obj, NULL);
232a6eba
MA
40 if (i915_gem_object_has_pinned_pages(obj))
41 i915_gem_object_unpin_pages(obj);
42 /* No polluting the memory region between tests */
f86dbacb 43 __i915_gem_object_put_pages(obj);
e70a27d7 44 i915_gem_object_unlock(obj);
232a6eba
MA
45 list_del(&obj->st_link);
46 i915_gem_object_put(obj);
47 }
48
49 cond_resched();
50
51 i915_gem_drain_freed_objects(i915);
52}
53
54static int igt_mock_fill(void *arg)
55{
56 struct intel_memory_region *mem = arg;
57 resource_size_t total = resource_size(&mem->region);
58 resource_size_t page_size;
59 resource_size_t rem;
60 unsigned long max_pages;
61 unsigned long page_num;
62 LIST_HEAD(objects);
63 int err = 0;
64
d53ec322
MA
65 page_size = PAGE_SIZE;
66 max_pages = div64_u64(total, page_size);
232a6eba
MA
67 rem = total;
68
69 for_each_prime_number_from(page_num, 1, max_pages) {
70 resource_size_t size = page_num * page_size;
71 struct drm_i915_gem_object *obj;
72
d22632c8 73 obj = i915_gem_object_create_region(mem, size, 0, 0);
232a6eba
MA
74 if (IS_ERR(obj)) {
75 err = PTR_ERR(obj);
76 break;
77 }
78
e70a27d7 79 err = i915_gem_object_pin_pages_unlocked(obj);
232a6eba
MA
80 if (err) {
81 i915_gem_object_put(obj);
82 break;
83 }
84
85 list_add(&obj->st_link, &objects);
86 rem -= size;
87 }
88
89 if (err == -ENOMEM)
90 err = 0;
91 if (err == -ENXIO) {
92 if (page_num * page_size <= rem) {
93 pr_err("%s failed, space still left in region\n",
94 __func__);
95 err = -EINVAL;
96 } else {
97 err = 0;
98 }
99 }
100
101 close_objects(mem, &objects);
102
103 return err;
104}
105
2f0b97ca
MA
106static struct drm_i915_gem_object *
107igt_object_create(struct intel_memory_region *mem,
108 struct list_head *objects,
109 u64 size,
110 unsigned int flags)
111{
112 struct drm_i915_gem_object *obj;
113 int err;
114
d22632c8 115 obj = i915_gem_object_create_region(mem, size, 0, flags);
2f0b97ca
MA
116 if (IS_ERR(obj))
117 return obj;
118
e70a27d7 119 err = i915_gem_object_pin_pages_unlocked(obj);
2f0b97ca
MA
120 if (err)
121 goto put;
122
123 list_add(&obj->st_link, objects);
124 return obj;
125
126put:
127 i915_gem_object_put(obj);
128 return ERR_PTR(err);
129}
130
131static void igt_object_release(struct drm_i915_gem_object *obj)
132{
e70a27d7 133 i915_gem_object_lock(obj, NULL);
2f0b97ca 134 i915_gem_object_unpin_pages(obj);
f86dbacb 135 __i915_gem_object_put_pages(obj);
e70a27d7 136 i915_gem_object_unlock(obj);
2f0b97ca
MA
137 list_del(&obj->st_link);
138 i915_gem_object_put(obj);
139}
140
a2843b3b
CW
141static bool is_contiguous(struct drm_i915_gem_object *obj)
142{
143 struct scatterlist *sg;
144 dma_addr_t addr = -1;
145
146 for (sg = obj->mm.pages->sgl; sg; sg = sg_next(sg)) {
147 if (addr != -1 && sg_dma_address(sg) != addr)
148 return false;
149
150 addr = sg_dma_address(sg) + sg_dma_len(sg);
151 }
152
153 return true;
154}
155
adeca641
AJ
156static int igt_mock_reserve(void *arg)
157{
158 struct intel_memory_region *mem = arg;
d53ec322 159 struct drm_i915_private *i915 = mem->i915;
adeca641
AJ
160 resource_size_t avail = resource_size(&mem->region);
161 struct drm_i915_gem_object *obj;
162 const u32 chunk_size = SZ_32M;
163 u32 i, offset, count, *order;
164 u64 allocated, cur_avail;
165 I915_RND_STATE(prng);
166 LIST_HEAD(objects);
167 int err = 0;
168
adeca641
AJ
169 count = avail / chunk_size;
170 order = i915_random_order(count, &prng);
171 if (!order)
172 return 0;
173
235582ca 174 mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
d53ec322
MA
175 if (IS_ERR(mem)) {
176 pr_err("failed to create memory region\n");
177 err = PTR_ERR(mem);
b23228f2 178 goto out_free_order;
d53ec322
MA
179 }
180
adeca641
AJ
181 /* Reserve a bunch of ranges within the region */
182 for (i = 0; i < count; ++i) {
183 u64 start = order[i] * chunk_size;
184 u64 size = i915_prandom_u32_max_state(chunk_size, &prng);
185
186 /* Allow for some really big holes */
187 if (!size)
188 continue;
189
190 size = round_up(size, PAGE_SIZE);
191 offset = igt_random_offset(&prng, 0, chunk_size, size,
192 PAGE_SIZE);
193
194 err = intel_memory_region_reserve(mem, start + offset, size);
195 if (err) {
196 pr_err("%s failed to reserve range", __func__);
197 goto out_close;
198 }
199
200 /* XXX: maybe sanity check the block range here? */
201 avail -= size;
202 }
203
204 /* Try to see if we can allocate from the remaining space */
205 allocated = 0;
206 cur_avail = avail;
207 do {
208 u32 size = i915_prandom_u32_max_state(cur_avail, &prng);
209
210 size = max_t(u32, round_up(size, PAGE_SIZE), PAGE_SIZE);
211 obj = igt_object_create(mem, &objects, size, 0);
212 if (IS_ERR(obj)) {
d53ec322 213 if (PTR_ERR(obj) == -ENXIO)
adeca641 214 break;
d53ec322 215
adeca641
AJ
216 err = PTR_ERR(obj);
217 goto out_close;
218 }
219 cur_avail -= size;
220 allocated += size;
221 } while (1);
222
223 if (allocated != avail) {
224 pr_err("%s mismatch between allocation and free space", __func__);
225 err = -EINVAL;
226 }
227
228out_close:
adeca641 229 close_objects(mem, &objects);
8b1f7f92 230 intel_memory_region_destroy(mem);
b23228f2
MA
231out_free_order:
232 kfree(order);
adeca641
AJ
233 return err;
234}
235
2f0b97ca
MA
236static int igt_mock_contiguous(void *arg)
237{
238 struct intel_memory_region *mem = arg;
239 struct drm_i915_gem_object *obj;
240 unsigned long n_objects;
241 LIST_HEAD(objects);
242 LIST_HEAD(holes);
243 I915_RND_STATE(prng);
2f0b97ca
MA
244 resource_size_t total;
245 resource_size_t min;
280bc0ce 246 u64 target;
2f0b97ca
MA
247 int err = 0;
248
249 total = resource_size(&mem->region);
250
251 /* Min size */
d53ec322 252 obj = igt_object_create(mem, &objects, PAGE_SIZE,
2f0b97ca
MA
253 I915_BO_ALLOC_CONTIGUOUS);
254 if (IS_ERR(obj))
255 return PTR_ERR(obj);
256
a2843b3b
CW
257 if (!is_contiguous(obj)) {
258 pr_err("%s min object spans disjoint sg entries\n", __func__);
2f0b97ca
MA
259 err = -EINVAL;
260 goto err_close_objects;
261 }
262
263 igt_object_release(obj);
264
265 /* Max size */
266 obj = igt_object_create(mem, &objects, total, I915_BO_ALLOC_CONTIGUOUS);
267 if (IS_ERR(obj))
268 return PTR_ERR(obj);
269
a2843b3b
CW
270 if (!is_contiguous(obj)) {
271 pr_err("%s max object spans disjoint sg entries\n", __func__);
2f0b97ca
MA
272 err = -EINVAL;
273 goto err_close_objects;
274 }
275
276 igt_object_release(obj);
277
278 /* Internal fragmentation should not bleed into the object size */
280bc0ce
CW
279 target = i915_prandom_u64_state(&prng);
280 div64_u64_rem(target, total, &target);
281 target = round_up(target, PAGE_SIZE);
2f0b97ca
MA
282 target = max_t(u64, PAGE_SIZE, target);
283
284 obj = igt_object_create(mem, &objects, target,
285 I915_BO_ALLOC_CONTIGUOUS);
286 if (IS_ERR(obj))
287 return PTR_ERR(obj);
288
289 if (obj->base.size != target) {
280bc0ce
CW
290 pr_err("%s obj->base.size(%zx) != target(%llx)\n", __func__,
291 obj->base.size, target);
2f0b97ca
MA
292 err = -EINVAL;
293 goto err_close_objects;
294 }
295
a2843b3b
CW
296 if (!is_contiguous(obj)) {
297 pr_err("%s object spans disjoint sg entries\n", __func__);
2f0b97ca
MA
298 err = -EINVAL;
299 goto err_close_objects;
300 }
301
302 igt_object_release(obj);
303
304 /*
305 * Try to fragment the address space, such that half of it is free, but
306 * the max contiguous block size is SZ_64K.
307 */
308
309 target = SZ_64K;
310 n_objects = div64_u64(total, target);
311
312 while (n_objects--) {
313 struct list_head *list;
314
315 if (n_objects % 2)
316 list = &holes;
317 else
318 list = &objects;
319
320 obj = igt_object_create(mem, list, target,
321 I915_BO_ALLOC_CONTIGUOUS);
322 if (IS_ERR(obj)) {
323 err = PTR_ERR(obj);
324 goto err_close_objects;
325 }
326 }
327
328 close_objects(mem, &holes);
329
330 min = target;
331 target = total >> 1;
332
d53ec322
MA
333 /* Make sure we can still allocate all the fragmented space */
334 obj = igt_object_create(mem, &objects, target, 0);
335 if (IS_ERR(obj)) {
336 err = PTR_ERR(obj);
337 goto err_close_objects;
d1487389 338 }
2f0b97ca 339
d53ec322
MA
340 igt_object_release(obj);
341
2f0b97ca
MA
342 /*
343 * Even though we have enough free space, we don't have a big enough
344 * contiguous block. Make sure that holds true.
345 */
346
347 do {
348 bool should_fail = target > min;
349
350 obj = igt_object_create(mem, &objects, target,
351 I915_BO_ALLOC_CONTIGUOUS);
352 if (should_fail != IS_ERR(obj)) {
353 pr_err("%s target allocation(%llx) mismatch\n",
280bc0ce 354 __func__, target);
2f0b97ca
MA
355 err = -EINVAL;
356 goto err_close_objects;
357 }
358
359 target >>= 1;
d53ec322 360 } while (target >= PAGE_SIZE);
2f0b97ca
MA
361
362err_close_objects:
363 list_splice_tail(&holes, &objects);
364 close_objects(mem, &objects);
365 return err;
366}
367
09a729b1
MA
368static int igt_mock_splintered_region(void *arg)
369{
370 struct intel_memory_region *mem = arg;
371 struct drm_i915_private *i915 = mem->i915;
d53ec322 372 struct i915_ttm_buddy_resource *res;
09a729b1 373 struct drm_i915_gem_object *obj;
6387a3c4 374 struct drm_buddy *mm;
09a729b1
MA
375 unsigned int expected_order;
376 LIST_HEAD(objects);
377 u64 size;
378 int err = 0;
379
380 /*
381 * Sanity check we can still allocate everything even if the
d53ec322 382 * mm.max_order != mm.size. i.e our starting address space size is not a
09a729b1
MA
383 * power-of-two.
384 */
385
386 size = (SZ_4G - 1) & PAGE_MASK;
235582ca 387 mem = mock_region_create(i915, 0, size, PAGE_SIZE, 0, 0);
09a729b1
MA
388 if (IS_ERR(mem))
389 return PTR_ERR(mem);
390
09a729b1
MA
391 obj = igt_object_create(mem, &objects, size, 0);
392 if (IS_ERR(obj)) {
393 err = PTR_ERR(obj);
394 goto out_close;
395 }
396
d53ec322
MA
397 res = to_ttm_buddy_resource(obj->mm.res);
398 mm = res->mm;
399 if (mm->size != size) {
400 pr_err("%s size mismatch(%llu != %llu)\n",
401 __func__, mm->size, size);
402 err = -EINVAL;
403 goto out_put;
404 }
405
406 expected_order = get_order(rounddown_pow_of_two(size));
407 if (mm->max_order != expected_order) {
408 pr_err("%s order mismatch(%u != %u)\n",
409 __func__, mm->max_order, expected_order);
410 err = -EINVAL;
411 goto out_put;
412 }
413
09a729b1
MA
414 close_objects(mem, &objects);
415
416 /*
417 * While we should be able allocate everything without any flag
418 * restrictions, if we consider I915_BO_ALLOC_CONTIGUOUS then we are
419 * actually limited to the largest power-of-two for the region size i.e
420 * max_order, due to the inner workings of the buddy allocator. So make
421 * sure that does indeed hold true.
422 */
423
d53ec322
MA
424 obj = igt_object_create(mem, &objects, size, I915_BO_ALLOC_CONTIGUOUS);
425 if (!IS_ERR(obj)) {
426 pr_err("%s too large contiguous allocation was not rejected\n",
427 __func__);
428 err = -EINVAL;
429 goto out_close;
09a729b1
MA
430 }
431
432 obj = igt_object_create(mem, &objects, rounddown_pow_of_two(size),
433 I915_BO_ALLOC_CONTIGUOUS);
434 if (IS_ERR(obj)) {
435 pr_err("%s largest possible contiguous allocation failed\n",
436 __func__);
437 err = PTR_ERR(obj);
438 goto out_close;
439 }
440
441out_close:
442 close_objects(mem, &objects);
443out_put:
8b1f7f92 444 intel_memory_region_destroy(mem);
09a729b1
MA
445 return err;
446}
447
d53ec322
MA
448#ifndef SZ_8G
449#define SZ_8G BIT_ULL(33)
450#endif
451
452static int igt_mock_max_segment(void *arg)
453{
d53ec322
MA
454 struct intel_memory_region *mem = arg;
455 struct drm_i915_private *i915 = mem->i915;
456 struct i915_ttm_buddy_resource *res;
457 struct drm_i915_gem_object *obj;
6387a3c4
A
458 struct drm_buddy_block *block;
459 struct drm_buddy *mm;
d53ec322
MA
460 struct list_head *blocks;
461 struct scatterlist *sg;
aff1e0b0 462 I915_RND_STATE(prng);
d53ec322 463 LIST_HEAD(objects);
aff1e0b0
MA
464 unsigned int max_segment;
465 unsigned int ps;
d53ec322
MA
466 u64 size;
467 int err = 0;
468
469 /*
470 * While we may create very large contiguous blocks, we may need
471 * to break those down for consumption elsewhere. In particular,
472 * dma-mapping with scatterlist elements have an implicit limit of
473 * UINT_MAX on each element.
474 */
475
476 size = SZ_8G;
aff1e0b0
MA
477 ps = PAGE_SIZE;
478 if (i915_prandom_u64_state(&prng) & 1)
479 ps = SZ_64K; /* For something like DG2 */
480
481 max_segment = round_down(UINT_MAX, ps);
482
483 mem = mock_region_create(i915, 0, size, ps, 0, 0);
d53ec322
MA
484 if (IS_ERR(mem))
485 return PTR_ERR(mem);
486
487 obj = igt_object_create(mem, &objects, size, 0);
488 if (IS_ERR(obj)) {
489 err = PTR_ERR(obj);
490 goto out_put;
491 }
492
493 res = to_ttm_buddy_resource(obj->mm.res);
494 blocks = &res->blocks;
495 mm = res->mm;
496 size = 0;
497 list_for_each_entry(block, blocks, link) {
6387a3c4
A
498 if (drm_buddy_block_size(mm, block) > size)
499 size = drm_buddy_block_size(mm, block);
d53ec322
MA
500 }
501 if (size < max_segment) {
502 pr_err("%s: Failed to create a huge contiguous block [> %u], largest block %lld\n",
503 __func__, max_segment, size);
504 err = -EINVAL;
505 goto out_close;
506 }
507
508 for (sg = obj->mm.pages->sgl; sg; sg = sg_next(sg)) {
aff1e0b0
MA
509 dma_addr_t daddr = sg_dma_address(sg);
510
d53ec322
MA
511 if (sg->length > max_segment) {
512 pr_err("%s: Created an oversized scatterlist entry, %u > %u\n",
513 __func__, sg->length, max_segment);
514 err = -EINVAL;
515 goto out_close;
516 }
aff1e0b0
MA
517
518 if (!IS_ALIGNED(daddr, ps)) {
519 pr_err("%s: Created an unaligned scatterlist entry, addr=%pa, ps=%u\n",
520 __func__, &daddr, ps);
521 err = -EINVAL;
522 goto out_close;
523 }
d53ec322
MA
524 }
525
526out_close:
527 close_objects(mem, &objects);
528out_put:
8b1f7f92 529 intel_memory_region_destroy(mem);
d53ec322
MA
530 return err;
531}
532
2d45f668
MA
533static u64 igt_object_mappable_total(struct drm_i915_gem_object *obj)
534{
535 struct intel_memory_region *mr = obj->mm.region;
536 struct i915_ttm_buddy_resource *bman_res =
537 to_ttm_buddy_resource(obj->mm.res);
538 struct drm_buddy *mm = bman_res->mm;
539 struct drm_buddy_block *block;
540 u64 total;
541
542 total = 0;
543 list_for_each_entry(block, &bman_res->blocks, link) {
544 u64 start = drm_buddy_block_offset(block);
545 u64 end = start + drm_buddy_block_size(mm, block);
546
547 if (start < mr->io_size)
548 total += min_t(u64, end, mr->io_size) - start;
549 }
550
551 return total;
552}
553
554static int igt_mock_io_size(void *arg)
555{
556 struct intel_memory_region *mr = arg;
557 struct drm_i915_private *i915 = mr->i915;
558 struct drm_i915_gem_object *obj;
559 u64 mappable_theft_total;
560 u64 io_size;
561 u64 total;
562 u64 ps;
563 u64 rem;
564 u64 size;
565 I915_RND_STATE(prng);
566 LIST_HEAD(objects);
567 int err = 0;
568
569 ps = SZ_4K;
570 if (i915_prandom_u64_state(&prng) & 1)
571 ps = SZ_64K; /* For something like DG2 */
572
573 div64_u64_rem(i915_prandom_u64_state(&prng), SZ_8G, &total);
574 total = round_down(total, ps);
575 total = max_t(u64, total, SZ_1G);
576
577 div64_u64_rem(i915_prandom_u64_state(&prng), total - ps, &io_size);
578 io_size = round_down(io_size, ps);
579 io_size = max_t(u64, io_size, SZ_256M); /* 256M seems to be the common lower limit */
580
581 pr_info("%s with ps=%llx, io_size=%llx, total=%llx\n",
582 __func__, ps, io_size, total);
583
584 mr = mock_region_create(i915, 0, total, ps, 0, io_size);
585 if (IS_ERR(mr)) {
586 err = PTR_ERR(mr);
587 goto out_err;
588 }
589
590 mappable_theft_total = 0;
591 rem = total - io_size;
592 do {
593 div64_u64_rem(i915_prandom_u64_state(&prng), rem, &size);
594 size = round_down(size, ps);
595 size = max(size, ps);
596
597 obj = igt_object_create(mr, &objects, size,
598 I915_BO_ALLOC_GPU_ONLY);
599 if (IS_ERR(obj)) {
600 pr_err("%s TOPDOWN failed with rem=%llx, size=%llx\n",
601 __func__, rem, size);
602 err = PTR_ERR(obj);
603 goto out_close;
604 }
605
606 mappable_theft_total += igt_object_mappable_total(obj);
607 rem -= size;
608 } while (rem);
609
610 pr_info("%s mappable theft=(%lluMiB/%lluMiB), total=%lluMiB\n",
611 __func__,
612 (u64)mappable_theft_total >> 20,
613 (u64)io_size >> 20,
614 (u64)total >> 20);
615
616 /*
617 * Even if we allocate all of the non-mappable portion, we should still
618 * be able to dip into the mappable portion.
619 */
620 obj = igt_object_create(mr, &objects, io_size,
621 I915_BO_ALLOC_GPU_ONLY);
622 if (IS_ERR(obj)) {
623 pr_err("%s allocation unexpectedly failed\n", __func__);
624 err = PTR_ERR(obj);
625 goto out_close;
626 }
627
628 close_objects(mr, &objects);
629
630 rem = io_size;
631 do {
632 div64_u64_rem(i915_prandom_u64_state(&prng), rem, &size);
633 size = round_down(size, ps);
634 size = max(size, ps);
635
636 obj = igt_object_create(mr, &objects, size, 0);
637 if (IS_ERR(obj)) {
638 pr_err("%s MAPPABLE failed with rem=%llx, size=%llx\n",
639 __func__, rem, size);
640 err = PTR_ERR(obj);
641 goto out_close;
642 }
643
644 if (igt_object_mappable_total(obj) != size) {
645 pr_err("%s allocation is not mappable(size=%llx)\n",
646 __func__, size);
647 err = -EINVAL;
648 goto out_close;
649 }
650 rem -= size;
651 } while (rem);
652
653 /*
654 * We assume CPU access is required by default, which should result in a
655 * failure here, even though the non-mappable portion is free.
656 */
657 obj = igt_object_create(mr, &objects, ps, 0);
658 if (!IS_ERR(obj)) {
659 pr_err("%s allocation unexpectedly succeeded\n", __func__);
660 err = -EINVAL;
661 goto out_close;
662 }
663
664out_close:
665 close_objects(mr, &objects);
666 intel_memory_region_destroy(mr);
667out_err:
668 if (err == -ENOMEM)
669 err = 0;
670
671 return err;
672}
673
340be48f
MA
674static int igt_gpu_write_dw(struct intel_context *ce,
675 struct i915_vma *vma,
676 u32 dword,
677 u32 value)
678{
679 return igt_gpu_fill_dw(ce, vma, dword * sizeof(u32),
680 vma->size >> PAGE_SHIFT, value);
681}
682
683static int igt_cpu_check(struct drm_i915_gem_object *obj, u32 dword, u32 val)
684{
6056e500
CW
685 unsigned long n = obj->base.size >> PAGE_SHIFT;
686 u32 *ptr;
340be48f
MA
687 int err;
688
6056e500 689 err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
340be48f
MA
690 if (err)
691 return err;
692
4f869f1d 693 ptr = i915_gem_object_pin_map(obj, I915_MAP_WC);
6056e500
CW
694 if (IS_ERR(ptr))
695 return PTR_ERR(ptr);
340be48f 696
6056e500
CW
697 ptr += dword;
698 while (n--) {
699 if (*ptr != val) {
700 pr_err("base[%u]=%08x, val=%08x\n",
701 dword, *ptr, val);
340be48f
MA
702 err = -EINVAL;
703 break;
704 }
6056e500
CW
705
706 ptr += PAGE_SIZE / sizeof(*ptr);
340be48f
MA
707 }
708
6056e500 709 i915_gem_object_unpin_map(obj);
340be48f
MA
710 return err;
711}
712
713static int igt_gpu_write(struct i915_gem_context *ctx,
714 struct drm_i915_gem_object *obj)
715{
716 struct i915_gem_engines *engines;
717 struct i915_gem_engines_iter it;
718 struct i915_address_space *vm;
719 struct intel_context *ce;
720 I915_RND_STATE(prng);
721 IGT_TIMEOUT(end_time);
722 unsigned int count;
723 struct i915_vma *vma;
724 int *order;
725 int i, n;
726 int err = 0;
727
728 GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
729
730 n = 0;
731 count = 0;
732 for_each_gem_engine(ce, i915_gem_context_lock_engines(ctx), it) {
733 count++;
734 if (!intel_engine_can_store_dword(ce->engine))
735 continue;
736
737 vm = ce->vm;
738 n++;
739 }
740 i915_gem_context_unlock_engines(ctx);
741 if (!n)
742 return 0;
743
744 order = i915_random_order(count * count, &prng);
745 if (!order)
746 return -ENOMEM;
747
748 vma = i915_vma_instance(obj, vm, NULL);
749 if (IS_ERR(vma)) {
750 err = PTR_ERR(vma);
751 goto out_free;
752 }
753
754 err = i915_vma_pin(vma, 0, 0, PIN_USER);
755 if (err)
756 goto out_free;
757
758 i = 0;
759 engines = i915_gem_context_lock_engines(ctx);
760 do {
761 u32 rng = prandom_u32_state(&prng);
762 u32 dword = offset_in_page(rng) / 4;
763
764 ce = engines->engines[order[i] % engines->num_engines];
765 i = (i + 1) % (count * count);
766 if (!ce || !intel_engine_can_store_dword(ce->engine))
767 continue;
768
769 err = igt_gpu_write_dw(ce, vma, dword, rng);
770 if (err)
771 break;
772
4f869f1d 773 i915_gem_object_lock(obj, NULL);
340be48f 774 err = igt_cpu_check(obj, dword, rng);
4f869f1d 775 i915_gem_object_unlock(obj);
340be48f
MA
776 if (err)
777 break;
778 } while (!__igt_timeout(end_time, NULL));
779 i915_gem_context_unlock_engines(ctx);
780
781out_free:
782 kfree(order);
783
784 if (err == -ENOMEM)
785 err = 0;
786
787 return err;
788}
789
b908be54
MA
790static int igt_lmem_create(void *arg)
791{
792 struct drm_i915_private *i915 = arg;
793 struct drm_i915_gem_object *obj;
794 int err = 0;
795
796 obj = i915_gem_object_create_lmem(i915, PAGE_SIZE, 0);
797 if (IS_ERR(obj))
798 return PTR_ERR(obj);
799
e70a27d7 800 err = i915_gem_object_pin_pages_unlocked(obj);
b908be54
MA
801 if (err)
802 goto out_put;
803
804 i915_gem_object_unpin_pages(obj);
805out_put:
806 i915_gem_object_put(obj);
807
808 return err;
809}
810
d22632c8
MA
811static int igt_lmem_create_with_ps(void *arg)
812{
813 struct drm_i915_private *i915 = arg;
814 int err = 0;
815 u32 ps;
816
817 for (ps = PAGE_SIZE; ps <= SZ_1G; ps <<= 1) {
818 struct drm_i915_gem_object *obj;
819 dma_addr_t daddr;
820
821 obj = __i915_gem_object_create_lmem_with_ps(i915, ps, ps, 0);
822 if (IS_ERR(obj)) {
823 err = PTR_ERR(obj);
824 if (err == -ENXIO || err == -E2BIG) {
825 pr_info("%s not enough lmem for ps(%u) err=%d\n",
826 __func__, ps, err);
827 err = 0;
828 }
829
830 break;
831 }
832
833 if (obj->base.size != ps) {
834 pr_err("%s size(%zu) != ps(%u)\n",
835 __func__, obj->base.size, ps);
836 err = -EINVAL;
837 goto out_put;
838 }
839
840 i915_gem_object_lock(obj, NULL);
841 err = i915_gem_object_pin_pages(obj);
6e0c5bf0
MA
842 if (err) {
843 if (err == -ENXIO || err == -E2BIG || err == -ENOMEM) {
844 pr_info("%s not enough lmem for ps(%u) err=%d\n",
845 __func__, ps, err);
846 err = 0;
847 }
d22632c8 848 goto out_put;
6e0c5bf0 849 }
d22632c8
MA
850
851 daddr = i915_gem_object_get_dma_address(obj, 0);
852 if (!IS_ALIGNED(daddr, ps)) {
853 pr_err("%s daddr(%pa) not aligned with ps(%u)\n",
854 __func__, &daddr, ps);
855 err = -EINVAL;
856 goto out_unpin;
857 }
858
859out_unpin:
860 i915_gem_object_unpin_pages(obj);
861 __i915_gem_object_put_pages(obj);
862out_put:
863 i915_gem_object_unlock(obj);
864 i915_gem_object_put(obj);
865
866 if (err)
867 break;
868 }
869
870 return err;
871}
872
4f869f1d
MA
873static int igt_lmem_create_cleared_cpu(void *arg)
874{
875 struct drm_i915_private *i915 = arg;
876 I915_RND_STATE(prng);
877 IGT_TIMEOUT(end_time);
878 u32 size, i;
879 int err;
880
881 i915_gem_drain_freed_objects(i915);
882
883 size = max_t(u32, PAGE_SIZE, i915_prandom_u32_max_state(SZ_32M, &prng));
884 size = round_up(size, PAGE_SIZE);
885 i = 0;
886
887 do {
888 struct drm_i915_gem_object *obj;
889 unsigned int flags;
890 u32 dword, val;
891 void *vaddr;
892
893 /*
894 * Alternate between cleared and uncleared allocations, while
895 * also dirtying the pages each time to check that the pages are
896 * always cleared if requested, since we should get some overlap
897 * of the underlying pages, if not all, since we are the only
898 * user.
899 */
900
901 flags = I915_BO_ALLOC_CPU_CLEAR;
902 if (i & 1)
903 flags = 0;
904
905 obj = i915_gem_object_create_lmem(i915, size, flags);
906 if (IS_ERR(obj))
907 return PTR_ERR(obj);
908
909 i915_gem_object_lock(obj, NULL);
910 err = i915_gem_object_pin_pages(obj);
911 if (err)
912 goto out_put;
913
914 dword = i915_prandom_u32_max_state(PAGE_SIZE / sizeof(u32),
915 &prng);
916
917 if (flags & I915_BO_ALLOC_CPU_CLEAR) {
918 err = igt_cpu_check(obj, dword, 0);
919 if (err) {
920 pr_err("%s failed with size=%u, flags=%u\n",
921 __func__, size, flags);
922 goto out_unpin;
923 }
924 }
925
926 vaddr = i915_gem_object_pin_map(obj, I915_MAP_WC);
927 if (IS_ERR(vaddr)) {
928 err = PTR_ERR(vaddr);
929 goto out_unpin;
930 }
931
932 val = prandom_u32_state(&prng);
933
934 memset32(vaddr, val, obj->base.size / sizeof(u32));
935
936 i915_gem_object_flush_map(obj);
937 i915_gem_object_unpin_map(obj);
938out_unpin:
939 i915_gem_object_unpin_pages(obj);
940 __i915_gem_object_put_pages(obj);
941out_put:
942 i915_gem_object_unlock(obj);
943 i915_gem_object_put(obj);
944
945 if (err)
946 break;
947 ++i;
948 } while (!__igt_timeout(end_time, NULL));
949
950 pr_info("%s completed (%u) iterations\n", __func__, i);
951
952 return err;
953}
954
340be48f
MA
955static int igt_lmem_write_gpu(void *arg)
956{
957 struct drm_i915_private *i915 = arg;
958 struct drm_i915_gem_object *obj;
959 struct i915_gem_context *ctx;
a8c9a7f5 960 struct file *file;
340be48f
MA
961 I915_RND_STATE(prng);
962 u32 sz;
963 int err;
964
965 file = mock_file(i915);
966 if (IS_ERR(file))
967 return PTR_ERR(file);
968
969 ctx = live_context(i915, file);
970 if (IS_ERR(ctx)) {
971 err = PTR_ERR(ctx);
972 goto out_file;
973 }
974
975 sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
976
977 obj = i915_gem_object_create_lmem(i915, sz, 0);
978 if (IS_ERR(obj)) {
979 err = PTR_ERR(obj);
980 goto out_file;
981 }
982
e70a27d7 983 err = i915_gem_object_pin_pages_unlocked(obj);
340be48f
MA
984 if (err)
985 goto out_put;
986
987 err = igt_gpu_write(ctx, obj);
988 if (err)
989 pr_err("igt_gpu_write failed(%d)\n", err);
990
991 i915_gem_object_unpin_pages(obj);
992out_put:
993 i915_gem_object_put(obj);
994out_file:
a8c9a7f5 995 fput(file);
340be48f
MA
996 return err;
997}
998
6804da20
CW
999static struct intel_engine_cs *
1000random_engine_class(struct drm_i915_private *i915,
1001 unsigned int class,
1002 struct rnd_state *prng)
1003{
1004 struct intel_engine_cs *engine;
1005 unsigned int count;
1006
1007 count = 0;
1008 for (engine = intel_engine_lookup_user(i915, class, 0);
1009 engine && engine->uabi_class == class;
1010 engine = rb_entry_safe(rb_next(&engine->uabi_node),
1011 typeof(*engine), uabi_node))
1012 count++;
1013
1014 count = i915_prandom_u32_max_state(count, prng);
1015 return intel_engine_lookup_user(i915, class, count);
1016}
1017
01377a0d
AJ
1018static int igt_lmem_write_cpu(void *arg)
1019{
1020 struct drm_i915_private *i915 = arg;
1021 struct drm_i915_gem_object *obj;
1022 I915_RND_STATE(prng);
1023 IGT_TIMEOUT(end_time);
1024 u32 bytes[] = {
1025 0, /* rng placeholder */
1026 sizeof(u32),
1027 sizeof(u64),
1028 64, /* cl */
1029 PAGE_SIZE,
1030 PAGE_SIZE - sizeof(u32),
1031 PAGE_SIZE - sizeof(u64),
1032 PAGE_SIZE - 64,
1033 };
6804da20 1034 struct intel_engine_cs *engine;
99919be7 1035 struct i915_request *rq;
01377a0d
AJ
1036 u32 *vaddr;
1037 u32 sz;
1038 u32 i;
1039 int *order;
1040 int count;
1041 int err;
1042
6804da20
CW
1043 engine = random_engine_class(i915, I915_ENGINE_CLASS_COPY, &prng);
1044 if (!engine)
01377a0d
AJ
1045 return 0;
1046
6804da20
CW
1047 pr_info("%s: using %s\n", __func__, engine->name);
1048
01377a0d
AJ
1049 sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
1050 sz = max_t(u32, 2 * PAGE_SIZE, sz);
1051
1052 obj = i915_gem_object_create_lmem(i915, sz, I915_BO_ALLOC_CONTIGUOUS);
1053 if (IS_ERR(obj))
1054 return PTR_ERR(obj);
1055
e70a27d7 1056 vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
01377a0d
AJ
1057 if (IS_ERR(vaddr)) {
1058 err = PTR_ERR(vaddr);
1059 goto out_put;
1060 }
1061
99919be7 1062 i915_gem_object_lock(obj, NULL);
c8d4c18b
CK
1063
1064 err = dma_resv_reserve_fences(obj->base.resv, 1);
1065 if (err) {
1066 i915_gem_object_unlock(obj);
1067 goto out_put;
1068 }
1069
01377a0d 1070 /* Put the pages into a known state -- from the gpu for added fun */
de5825be 1071 intel_engine_pm_get(engine);
99919be7
TH
1072 err = intel_context_migrate_clear(engine->gt->migrate.context, NULL,
1073 obj->mm.pages->sgl, I915_CACHE_NONE,
1074 true, 0xdeadbeaf, &rq);
1075 if (rq) {
73511edf
CK
1076 dma_resv_add_fence(obj->base.resv, &rq->fence,
1077 DMA_RESV_USAGE_WRITE);
99919be7
TH
1078 i915_request_put(rq);
1079 }
01377a0d 1080
99919be7
TH
1081 intel_engine_pm_put(engine);
1082 if (!err)
1083 err = i915_gem_object_set_to_wc_domain(obj, true);
01377a0d
AJ
1084 i915_gem_object_unlock(obj);
1085 if (err)
1086 goto out_unpin;
1087
1088 count = ARRAY_SIZE(bytes);
1089 order = i915_random_order(count * count, &prng);
1090 if (!order) {
1091 err = -ENOMEM;
1092 goto out_unpin;
1093 }
1094
dd5e0249
CW
1095 /* A random multiple of u32, picked between [64, PAGE_SIZE - 64] */
1096 bytes[0] = igt_random_offset(&prng, 64, PAGE_SIZE - 64, 0, sizeof(u32));
1097 GEM_BUG_ON(!IS_ALIGNED(bytes[0], sizeof(u32)));
01377a0d
AJ
1098
1099 i = 0;
1100 do {
1101 u32 offset;
1102 u32 align;
1103 u32 dword;
1104 u32 size;
1105 u32 val;
1106
1107 size = bytes[order[i] % count];
1108 i = (i + 1) % (count * count);
1109
1110 align = bytes[order[i] % count];
1111 i = (i + 1) % (count * count);
1112
1113 align = max_t(u32, sizeof(u32), rounddown_pow_of_two(align));
1114
1115 offset = igt_random_offset(&prng, 0, obj->base.size,
1116 size, align);
1117
1118 val = prandom_u32_state(&prng);
1119 memset32(vaddr + offset / sizeof(u32), val ^ 0xdeadbeaf,
1120 size / sizeof(u32));
1121
1122 /*
1123 * Sample random dw -- don't waste precious time reading every
1124 * single dw.
1125 */
1126 dword = igt_random_offset(&prng, offset,
1127 offset + size,
1128 sizeof(u32), sizeof(u32));
1129 dword /= sizeof(u32);
1130 if (vaddr[dword] != (val ^ 0xdeadbeaf)) {
1131 pr_err("%s vaddr[%u]=%u, val=%u, size=%u, align=%u, offset=%u\n",
1132 __func__, dword, vaddr[dword], val ^ 0xdeadbeaf,
1133 size, align, offset);
1134 err = -EINVAL;
1135 break;
1136 }
1137 } while (!__igt_timeout(end_time, NULL));
1138
1139out_unpin:
1140 i915_gem_object_unpin_map(obj);
1141out_put:
1142 i915_gem_object_put(obj);
1143
1144 return err;
1145}
1146
ba12993c
MA
1147static const char *repr_type(u32 type)
1148{
1149 switch (type) {
1150 case I915_MAP_WB:
1151 return "WB";
1152 case I915_MAP_WC:
1153 return "WC";
1154 }
1155
1156 return "";
1157}
1158
1159static struct drm_i915_gem_object *
1160create_region_for_mapping(struct intel_memory_region *mr, u64 size, u32 type,
1161 void **out_addr)
1162{
1163 struct drm_i915_gem_object *obj;
1164 void *addr;
1165
d22632c8 1166 obj = i915_gem_object_create_region(mr, size, 0, 0);
45d41739
MA
1167 if (IS_ERR(obj)) {
1168 if (PTR_ERR(obj) == -ENOSPC) /* Stolen memory */
1169 return ERR_PTR(-ENODEV);
ba12993c 1170 return obj;
45d41739 1171 }
ba12993c 1172
e70a27d7 1173 addr = i915_gem_object_pin_map_unlocked(obj, type);
ba12993c
MA
1174 if (IS_ERR(addr)) {
1175 i915_gem_object_put(obj);
1176 if (PTR_ERR(addr) == -ENXIO)
1177 return ERR_PTR(-ENODEV);
1178 return addr;
1179 }
1180
1181 *out_addr = addr;
1182 return obj;
1183}
1184
1185static int wrap_ktime_compare(const void *A, const void *B)
1186{
1187 const ktime_t *a = A, *b = B;
1188
1189 return ktime_compare(*a, *b);
1190}
1191
1192static void igt_memcpy_long(void *dst, const void *src, size_t size)
1193{
1194 unsigned long *tmp = dst;
1195 const unsigned long *s = src;
1196
1197 size = size / sizeof(unsigned long);
1198 while (size--)
1199 *tmp++ = *s++;
1200}
1201
1202static inline void igt_memcpy(void *dst, const void *src, size_t size)
1203{
1204 memcpy(dst, src, size);
1205}
1206
1207static inline void igt_memcpy_from_wc(void *dst, const void *src, size_t size)
1208{
1209 i915_memcpy_from_wc(dst, src, size);
1210}
1211
1212static int _perf_memcpy(struct intel_memory_region *src_mr,
1213 struct intel_memory_region *dst_mr,
1214 u64 size, u32 src_type, u32 dst_type)
1215{
1216 struct drm_i915_private *i915 = src_mr->i915;
1217 const struct {
1218 const char *name;
1219 void (*copy)(void *dst, const void *src, size_t size);
1220 bool skip;
1221 } tests[] = {
1222 {
1223 "memcpy",
1224 igt_memcpy,
1225 },
1226 {
1227 "memcpy_long",
1228 igt_memcpy_long,
1229 },
1230 {
1231 "memcpy_from_wc",
1232 igt_memcpy_from_wc,
1233 !i915_has_memcpy_from_wc(),
1234 },
1235 };
1236 struct drm_i915_gem_object *src, *dst;
1237 void *src_addr, *dst_addr;
1238 int ret = 0;
1239 int i;
1240
1241 src = create_region_for_mapping(src_mr, size, src_type, &src_addr);
1242 if (IS_ERR(src)) {
1243 ret = PTR_ERR(src);
1244 goto out;
1245 }
1246
1247 dst = create_region_for_mapping(dst_mr, size, dst_type, &dst_addr);
1248 if (IS_ERR(dst)) {
1249 ret = PTR_ERR(dst);
1250 goto out_unpin_src;
1251 }
1252
1253 for (i = 0; i < ARRAY_SIZE(tests); ++i) {
1254 ktime_t t[5];
1255 int pass;
1256
1257 if (tests[i].skip)
1258 continue;
1259
1260 for (pass = 0; pass < ARRAY_SIZE(t); pass++) {
1261 ktime_t t0, t1;
1262
1263 t0 = ktime_get();
1264
1265 tests[i].copy(dst_addr, src_addr, size);
1266
1267 t1 = ktime_get();
1268 t[pass] = ktime_sub(t1, t0);
1269 }
1270
1271 sort(t, ARRAY_SIZE(t), sizeof(*t), wrap_ktime_compare, NULL);
c318a203
CW
1272 if (t[0] <= 0) {
1273 /* ignore the impossible to protect our sanity */
1274 pr_debug("Skipping %s src(%s, %s) -> dst(%s, %s) %14s %4lluKiB copy, unstable measurement [%lld, %lld]\n",
1275 __func__,
1276 src_mr->name, repr_type(src_type),
1277 dst_mr->name, repr_type(dst_type),
1278 tests[i].name, size >> 10,
1279 t[0], t[4]);
1280 continue;
1281 }
1282
ba12993c
MA
1283 pr_info("%s src(%s, %s) -> dst(%s, %s) %14s %4llu KiB copy: %5lld MiB/s\n",
1284 __func__,
c318a203
CW
1285 src_mr->name, repr_type(src_type),
1286 dst_mr->name, repr_type(dst_type),
1287 tests[i].name, size >> 10,
ba12993c
MA
1288 div64_u64(mul_u32_u32(4 * size,
1289 1000 * 1000 * 1000),
1290 t[1] + 2 * t[2] + t[3]) >> 20);
1291
1292 cond_resched();
1293 }
1294
1295 i915_gem_object_unpin_map(dst);
1296 i915_gem_object_put(dst);
1297out_unpin_src:
1298 i915_gem_object_unpin_map(src);
1299 i915_gem_object_put(src);
1300
1301 i915_gem_drain_freed_objects(i915);
1302out:
1303 if (ret == -ENODEV)
1304 ret = 0;
1305
1306 return ret;
1307}
1308
1309static int perf_memcpy(void *arg)
1310{
1311 struct drm_i915_private *i915 = arg;
1312 static const u32 types[] = {
1313 I915_MAP_WB,
1314 I915_MAP_WC,
1315 };
1316 static const u32 sizes[] = {
1317 SZ_4K,
1318 SZ_64K,
1319 SZ_4M,
1320 };
1321 struct intel_memory_region *src_mr, *dst_mr;
1322 int src_id, dst_id;
1323 int i, j, k;
1324 int ret;
1325
1326 for_each_memory_region(src_mr, i915, src_id) {
1327 for_each_memory_region(dst_mr, i915, dst_id) {
1328 for (i = 0; i < ARRAY_SIZE(sizes); ++i) {
1329 for (j = 0; j < ARRAY_SIZE(types); ++j) {
1330 for (k = 0; k < ARRAY_SIZE(types); ++k) {
1331 ret = _perf_memcpy(src_mr,
1332 dst_mr,
1333 sizes[i],
1334 types[j],
1335 types[k]);
1336 if (ret)
1337 return ret;
1338 }
1339 }
1340 }
1341 }
1342 }
1343
1344 return 0;
1345}
1346
232a6eba
MA
1347int intel_memory_region_mock_selftests(void)
1348{
1349 static const struct i915_subtest tests[] = {
adeca641 1350 SUBTEST(igt_mock_reserve),
232a6eba 1351 SUBTEST(igt_mock_fill),
2f0b97ca 1352 SUBTEST(igt_mock_contiguous),
09a729b1 1353 SUBTEST(igt_mock_splintered_region),
d53ec322 1354 SUBTEST(igt_mock_max_segment),
2d45f668 1355 SUBTEST(igt_mock_io_size),
232a6eba
MA
1356 };
1357 struct intel_memory_region *mem;
1358 struct drm_i915_private *i915;
1359 int err;
1360
1361 i915 = mock_gem_device();
1362 if (!i915)
1363 return -ENOMEM;
1364
235582ca 1365 mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
232a6eba
MA
1366 if (IS_ERR(mem)) {
1367 pr_err("failed to create memory region\n");
1368 err = PTR_ERR(mem);
1369 goto out_unref;
1370 }
1371
1372 err = i915_subtests(tests, mem);
1373
8b1f7f92 1374 intel_memory_region_destroy(mem);
232a6eba 1375out_unref:
82be0d75 1376 mock_destroy_device(i915);
232a6eba
MA
1377 return err;
1378}
b908be54
MA
1379
1380int intel_memory_region_live_selftests(struct drm_i915_private *i915)
1381{
1382 static const struct i915_subtest tests[] = {
1383 SUBTEST(igt_lmem_create),
d22632c8 1384 SUBTEST(igt_lmem_create_with_ps),
4f869f1d 1385 SUBTEST(igt_lmem_create_cleared_cpu),
01377a0d 1386 SUBTEST(igt_lmem_write_cpu),
340be48f 1387 SUBTEST(igt_lmem_write_gpu),
b908be54
MA
1388 };
1389
1390 if (!HAS_LMEM(i915)) {
1391 pr_info("device lacks LMEM support, skipping\n");
1392 return 0;
1393 }
1394
8c2699fa 1395 if (intel_gt_is_wedged(to_gt(i915)))
b908be54
MA
1396 return 0;
1397
1398 return i915_live_subtests(tests, i915);
1399}
ba12993c
MA
1400
1401int intel_memory_region_perf_selftests(struct drm_i915_private *i915)
1402{
1403 static const struct i915_subtest tests[] = {
1404 SUBTEST(perf_memcpy),
1405 };
1406
8c2699fa 1407 if (intel_gt_is_wedged(to_gt(i915)))
ba12993c
MA
1408 return 0;
1409
1410 return i915_live_subtests(tests, i915);
1411}