linux/drivers/gpu/drm/i915/selftests/intel_memory_region.c
<<
>>
Prefs
   1// SPDX-License-Identifier: MIT
   2/*
   3 * Copyright © 2019 Intel Corporation
   4 */
   5
   6#include <linux/prime_numbers.h>
   7#include <linux/sort.h>
   8
   9#include "../i915_selftest.h"
  10
  11#include "mock_drm.h"
  12#include "mock_gem_device.h"
  13#include "mock_region.h"
  14
  15#include "gem/i915_gem_context.h"
  16#include "gem/i915_gem_lmem.h"
  17#include "gem/i915_gem_region.h"
  18#include "gem/i915_gem_object_blt.h"
  19#include "gem/selftests/igt_gem_utils.h"
  20#include "gem/selftests/mock_context.h"
  21#include "gt/intel_engine_user.h"
  22#include "gt/intel_gt.h"
  23#include "i915_memcpy.h"
  24#include "selftests/igt_flush_test.h"
  25#include "selftests/i915_random.h"
  26
  27static void close_objects(struct intel_memory_region *mem,
  28                          struct list_head *objects)
  29{
  30        struct drm_i915_private *i915 = mem->i915;
  31        struct drm_i915_gem_object *obj, *on;
  32
  33        list_for_each_entry_safe(obj, on, objects, st_link) {
  34                if (i915_gem_object_has_pinned_pages(obj))
  35                        i915_gem_object_unpin_pages(obj);
  36                /* No polluting the memory region between tests */
  37                __i915_gem_object_put_pages(obj);
  38                list_del(&obj->st_link);
  39                i915_gem_object_put(obj);
  40        }
  41
  42        cond_resched();
  43
  44        i915_gem_drain_freed_objects(i915);
  45}
  46
  47static int igt_mock_fill(void *arg)
  48{
  49        struct intel_memory_region *mem = arg;
  50        resource_size_t total = resource_size(&mem->region);
  51        resource_size_t page_size;
  52        resource_size_t rem;
  53        unsigned long max_pages;
  54        unsigned long page_num;
  55        LIST_HEAD(objects);
  56        int err = 0;
  57
  58        page_size = mem->mm.chunk_size;
  59        max_pages = div64_u64(total, page_size);
  60        rem = total;
  61
  62        for_each_prime_number_from(page_num, 1, max_pages) {
  63                resource_size_t size = page_num * page_size;
  64                struct drm_i915_gem_object *obj;
  65
  66                obj = i915_gem_object_create_region(mem, size, 0);
  67                if (IS_ERR(obj)) {
  68                        err = PTR_ERR(obj);
  69                        break;
  70                }
  71
  72                err = i915_gem_object_pin_pages(obj);
  73                if (err) {
  74                        i915_gem_object_put(obj);
  75                        break;
  76                }
  77
  78                list_add(&obj->st_link, &objects);
  79                rem -= size;
  80        }
  81
  82        if (err == -ENOMEM)
  83                err = 0;
  84        if (err == -ENXIO) {
  85                if (page_num * page_size <= rem) {
  86                        pr_err("%s failed, space still left in region\n",
  87                               __func__);
  88                        err = -EINVAL;
  89                } else {
  90                        err = 0;
  91                }
  92        }
  93
  94        close_objects(mem, &objects);
  95
  96        return err;
  97}
  98
  99static struct drm_i915_gem_object *
 100igt_object_create(struct intel_memory_region *mem,
 101                  struct list_head *objects,
 102                  u64 size,
 103                  unsigned int flags)
 104{
 105        struct drm_i915_gem_object *obj;
 106        int err;
 107
 108        obj = i915_gem_object_create_region(mem, size, flags);
 109        if (IS_ERR(obj))
 110                return obj;
 111
 112        err = i915_gem_object_pin_pages(obj);
 113        if (err)
 114                goto put;
 115
 116        list_add(&obj->st_link, objects);
 117        return obj;
 118
 119put:
 120        i915_gem_object_put(obj);
 121        return ERR_PTR(err);
 122}
 123
 124static void igt_object_release(struct drm_i915_gem_object *obj)
 125{
 126        i915_gem_object_unpin_pages(obj);
 127        __i915_gem_object_put_pages(obj);
 128        list_del(&obj->st_link);
 129        i915_gem_object_put(obj);
 130}
 131
 132static int igt_mock_contiguous(void *arg)
 133{
 134        struct intel_memory_region *mem = arg;
 135        struct drm_i915_gem_object *obj;
 136        unsigned long n_objects;
 137        LIST_HEAD(objects);
 138        LIST_HEAD(holes);
 139        I915_RND_STATE(prng);
 140        resource_size_t total;
 141        resource_size_t min;
 142        u64 target;
 143        int err = 0;
 144
 145        total = resource_size(&mem->region);
 146
 147        /* Min size */
 148        obj = igt_object_create(mem, &objects, mem->mm.chunk_size,
 149                                I915_BO_ALLOC_CONTIGUOUS);
 150        if (IS_ERR(obj))
 151                return PTR_ERR(obj);
 152
 153        if (obj->mm.pages->nents != 1) {
 154                pr_err("%s min object spans multiple sg entries\n", __func__);
 155                err = -EINVAL;
 156                goto err_close_objects;
 157        }
 158
 159        igt_object_release(obj);
 160
 161        /* Max size */
 162        obj = igt_object_create(mem, &objects, total, I915_BO_ALLOC_CONTIGUOUS);
 163        if (IS_ERR(obj))
 164                return PTR_ERR(obj);
 165
 166        if (obj->mm.pages->nents != 1) {
 167                pr_err("%s max object spans multiple sg entries\n", __func__);
 168                err = -EINVAL;
 169                goto err_close_objects;
 170        }
 171
 172        igt_object_release(obj);
 173
 174        /* Internal fragmentation should not bleed into the object size */
 175        target = i915_prandom_u64_state(&prng);
 176        div64_u64_rem(target, total, &target);
 177        target = round_up(target, PAGE_SIZE);
 178        target = max_t(u64, PAGE_SIZE, target);
 179
 180        obj = igt_object_create(mem, &objects, target,
 181                                I915_BO_ALLOC_CONTIGUOUS);
 182        if (IS_ERR(obj))
 183                return PTR_ERR(obj);
 184
 185        if (obj->base.size != target) {
 186                pr_err("%s obj->base.size(%zx) != target(%llx)\n", __func__,
 187                       obj->base.size, target);
 188                err = -EINVAL;
 189                goto err_close_objects;
 190        }
 191
 192        if (obj->mm.pages->nents != 1) {
 193                pr_err("%s object spans multiple sg entries\n", __func__);
 194                err = -EINVAL;
 195                goto err_close_objects;
 196        }
 197
 198        igt_object_release(obj);
 199
 200        /*
 201         * Try to fragment the address space, such that half of it is free, but
 202         * the max contiguous block size is SZ_64K.
 203         */
 204
 205        target = SZ_64K;
 206        n_objects = div64_u64(total, target);
 207
 208        while (n_objects--) {
 209                struct list_head *list;
 210
 211                if (n_objects % 2)
 212                        list = &holes;
 213                else
 214                        list = &objects;
 215
 216                obj = igt_object_create(mem, list, target,
 217                                        I915_BO_ALLOC_CONTIGUOUS);
 218                if (IS_ERR(obj)) {
 219                        err = PTR_ERR(obj);
 220                        goto err_close_objects;
 221                }
 222        }
 223
 224        close_objects(mem, &holes);
 225
 226        min = target;
 227        target = total >> 1;
 228
 229        /* Make sure we can still allocate all the fragmented space */
 230        obj = igt_object_create(mem, &objects, target, 0);
 231        if (IS_ERR(obj)) {
 232                err = PTR_ERR(obj);
 233                goto err_close_objects;
 234        }
 235
 236        igt_object_release(obj);
 237
 238        /*
 239         * Even though we have enough free space, we don't have a big enough
 240         * contiguous block. Make sure that holds true.
 241         */
 242
 243        do {
 244                bool should_fail = target > min;
 245
 246                obj = igt_object_create(mem, &objects, target,
 247                                        I915_BO_ALLOC_CONTIGUOUS);
 248                if (should_fail != IS_ERR(obj)) {
 249                        pr_err("%s target allocation(%llx) mismatch\n",
 250                               __func__, target);
 251                        err = -EINVAL;
 252                        goto err_close_objects;
 253                }
 254
 255                target >>= 1;
 256        } while (target >= mem->mm.chunk_size);
 257
 258err_close_objects:
 259        list_splice_tail(&holes, &objects);
 260        close_objects(mem, &objects);
 261        return err;
 262}
 263
 264static int igt_gpu_write_dw(struct intel_context *ce,
 265                            struct i915_vma *vma,
 266                            u32 dword,
 267                            u32 value)
 268{
 269        return igt_gpu_fill_dw(ce, vma, dword * sizeof(u32),
 270                               vma->size >> PAGE_SHIFT, value);
 271}
 272
 273static int igt_cpu_check(struct drm_i915_gem_object *obj, u32 dword, u32 val)
 274{
 275        unsigned long n = obj->base.size >> PAGE_SHIFT;
 276        u32 *ptr;
 277        int err;
 278
 279        err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
 280        if (err)
 281                return err;
 282
 283        ptr = i915_gem_object_pin_map(obj, I915_MAP_WC);
 284        if (IS_ERR(ptr))
 285                return PTR_ERR(ptr);
 286
 287        ptr += dword;
 288        while (n--) {
 289                if (*ptr != val) {
 290                        pr_err("base[%u]=%08x, val=%08x\n",
 291                               dword, *ptr, val);
 292                        err = -EINVAL;
 293                        break;
 294                }
 295
 296                ptr += PAGE_SIZE / sizeof(*ptr);
 297        }
 298
 299        i915_gem_object_unpin_map(obj);
 300        return err;
 301}
 302
 303static int igt_gpu_write(struct i915_gem_context *ctx,
 304                         struct drm_i915_gem_object *obj)
 305{
 306        struct i915_gem_engines *engines;
 307        struct i915_gem_engines_iter it;
 308        struct i915_address_space *vm;
 309        struct intel_context *ce;
 310        I915_RND_STATE(prng);
 311        IGT_TIMEOUT(end_time);
 312        unsigned int count;
 313        struct i915_vma *vma;
 314        int *order;
 315        int i, n;
 316        int err = 0;
 317
 318        GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
 319
 320        n = 0;
 321        count = 0;
 322        for_each_gem_engine(ce, i915_gem_context_lock_engines(ctx), it) {
 323                count++;
 324                if (!intel_engine_can_store_dword(ce->engine))
 325                        continue;
 326
 327                vm = ce->vm;
 328                n++;
 329        }
 330        i915_gem_context_unlock_engines(ctx);
 331        if (!n)
 332                return 0;
 333
 334        order = i915_random_order(count * count, &prng);
 335        if (!order)
 336                return -ENOMEM;
 337
 338        vma = i915_vma_instance(obj, vm, NULL);
 339        if (IS_ERR(vma)) {
 340                err = PTR_ERR(vma);
 341                goto out_free;
 342        }
 343
 344        err = i915_vma_pin(vma, 0, 0, PIN_USER);
 345        if (err)
 346                goto out_free;
 347
 348        i = 0;
 349        engines = i915_gem_context_lock_engines(ctx);
 350        do {
 351                u32 rng = prandom_u32_state(&prng);
 352                u32 dword = offset_in_page(rng) / 4;
 353
 354                ce = engines->engines[order[i] % engines->num_engines];
 355                i = (i + 1) % (count * count);
 356                if (!ce || !intel_engine_can_store_dword(ce->engine))
 357                        continue;
 358
 359                err = igt_gpu_write_dw(ce, vma, dword, rng);
 360                if (err)
 361                        break;
 362
 363                err = igt_cpu_check(obj, dword, rng);
 364                if (err)
 365                        break;
 366        } while (!__igt_timeout(end_time, NULL));
 367        i915_gem_context_unlock_engines(ctx);
 368
 369out_free:
 370        kfree(order);
 371
 372        if (err == -ENOMEM)
 373                err = 0;
 374
 375        return err;
 376}
 377
 378static int igt_lmem_create(void *arg)
 379{
 380        struct drm_i915_private *i915 = arg;
 381        struct drm_i915_gem_object *obj;
 382        int err = 0;
 383
 384        obj = i915_gem_object_create_lmem(i915, PAGE_SIZE, 0);
 385        if (IS_ERR(obj))
 386                return PTR_ERR(obj);
 387
 388        err = i915_gem_object_pin_pages(obj);
 389        if (err)
 390                goto out_put;
 391
 392        i915_gem_object_unpin_pages(obj);
 393out_put:
 394        i915_gem_object_put(obj);
 395
 396        return err;
 397}
 398
 399static int igt_lmem_write_gpu(void *arg)
 400{
 401        struct drm_i915_private *i915 = arg;
 402        struct drm_i915_gem_object *obj;
 403        struct i915_gem_context *ctx;
 404        struct file *file;
 405        I915_RND_STATE(prng);
 406        u32 sz;
 407        int err;
 408
 409        file = mock_file(i915);
 410        if (IS_ERR(file))
 411                return PTR_ERR(file);
 412
 413        ctx = live_context(i915, file);
 414        if (IS_ERR(ctx)) {
 415                err = PTR_ERR(ctx);
 416                goto out_file;
 417        }
 418
 419        sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
 420
 421        obj = i915_gem_object_create_lmem(i915, sz, 0);
 422        if (IS_ERR(obj)) {
 423                err = PTR_ERR(obj);
 424                goto out_file;
 425        }
 426
 427        err = i915_gem_object_pin_pages(obj);
 428        if (err)
 429                goto out_put;
 430
 431        err = igt_gpu_write(ctx, obj);
 432        if (err)
 433                pr_err("igt_gpu_write failed(%d)\n", err);
 434
 435        i915_gem_object_unpin_pages(obj);
 436out_put:
 437        i915_gem_object_put(obj);
 438out_file:
 439        fput(file);
 440        return err;
 441}
 442
 443static struct intel_engine_cs *
 444random_engine_class(struct drm_i915_private *i915,
 445                    unsigned int class,
 446                    struct rnd_state *prng)
 447{
 448        struct intel_engine_cs *engine;
 449        unsigned int count;
 450
 451        count = 0;
 452        for (engine = intel_engine_lookup_user(i915, class, 0);
 453             engine && engine->uabi_class == class;
 454             engine = rb_entry_safe(rb_next(&engine->uabi_node),
 455                                    typeof(*engine), uabi_node))
 456                count++;
 457
 458        count = i915_prandom_u32_max_state(count, prng);
 459        return intel_engine_lookup_user(i915, class, count);
 460}
 461
 462static int igt_lmem_write_cpu(void *arg)
 463{
 464        struct drm_i915_private *i915 = arg;
 465        struct drm_i915_gem_object *obj;
 466        I915_RND_STATE(prng);
 467        IGT_TIMEOUT(end_time);
 468        u32 bytes[] = {
 469                0, /* rng placeholder */
 470                sizeof(u32),
 471                sizeof(u64),
 472                64, /* cl */
 473                PAGE_SIZE,
 474                PAGE_SIZE - sizeof(u32),
 475                PAGE_SIZE - sizeof(u64),
 476                PAGE_SIZE - 64,
 477        };
 478        struct intel_engine_cs *engine;
 479        u32 *vaddr;
 480        u32 sz;
 481        u32 i;
 482        int *order;
 483        int count;
 484        int err;
 485
 486        engine = random_engine_class(i915, I915_ENGINE_CLASS_COPY, &prng);
 487        if (!engine)
 488                return 0;
 489
 490        pr_info("%s: using %s\n", __func__, engine->name);
 491
 492        sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
 493        sz = max_t(u32, 2 * PAGE_SIZE, sz);
 494
 495        obj = i915_gem_object_create_lmem(i915, sz, I915_BO_ALLOC_CONTIGUOUS);
 496        if (IS_ERR(obj))
 497                return PTR_ERR(obj);
 498
 499        vaddr = i915_gem_object_pin_map(obj, I915_MAP_WC);
 500        if (IS_ERR(vaddr)) {
 501                err = PTR_ERR(vaddr);
 502                goto out_put;
 503        }
 504
 505        /* Put the pages into a known state -- from the gpu for added fun */
 506        intel_engine_pm_get(engine);
 507        err = i915_gem_object_fill_blt(obj, engine->kernel_context, 0xdeadbeaf);
 508        intel_engine_pm_put(engine);
 509        if (err)
 510                goto out_unpin;
 511
 512        i915_gem_object_lock(obj);
 513        err = i915_gem_object_set_to_wc_domain(obj, true);
 514        i915_gem_object_unlock(obj);
 515        if (err)
 516                goto out_unpin;
 517
 518        count = ARRAY_SIZE(bytes);
 519        order = i915_random_order(count * count, &prng);
 520        if (!order) {
 521                err = -ENOMEM;
 522                goto out_unpin;
 523        }
 524
 525        /* We want to throw in a random width/align */
 526        bytes[0] = igt_random_offset(&prng, 0, PAGE_SIZE, sizeof(u32),
 527                                     sizeof(u32));
 528
 529        i = 0;
 530        do {
 531                u32 offset;
 532                u32 align;
 533                u32 dword;
 534                u32 size;
 535                u32 val;
 536
 537                size = bytes[order[i] % count];
 538                i = (i + 1) % (count * count);
 539
 540                align = bytes[order[i] % count];
 541                i = (i + 1) % (count * count);
 542
 543                align = max_t(u32, sizeof(u32), rounddown_pow_of_two(align));
 544
 545                offset = igt_random_offset(&prng, 0, obj->base.size,
 546                                           size, align);
 547
 548                val = prandom_u32_state(&prng);
 549                memset32(vaddr + offset / sizeof(u32), val ^ 0xdeadbeaf,
 550                         size / sizeof(u32));
 551
 552                /*
 553                 * Sample random dw -- don't waste precious time reading every
 554                 * single dw.
 555                 */
 556                dword = igt_random_offset(&prng, offset,
 557                                          offset + size,
 558                                          sizeof(u32), sizeof(u32));
 559                dword /= sizeof(u32);
 560                if (vaddr[dword] != (val ^ 0xdeadbeaf)) {
 561                        pr_err("%s vaddr[%u]=%u, val=%u, size=%u, align=%u, offset=%u\n",
 562                               __func__, dword, vaddr[dword], val ^ 0xdeadbeaf,
 563                               size, align, offset);
 564                        err = -EINVAL;
 565                        break;
 566                }
 567        } while (!__igt_timeout(end_time, NULL));
 568
 569out_unpin:
 570        i915_gem_object_unpin_map(obj);
 571out_put:
 572        i915_gem_object_put(obj);
 573
 574        return err;
 575}
 576
 577static const char *repr_type(u32 type)
 578{
 579        switch (type) {
 580        case I915_MAP_WB:
 581                return "WB";
 582        case I915_MAP_WC:
 583                return "WC";
 584        }
 585
 586        return "";
 587}
 588
 589static struct drm_i915_gem_object *
 590create_region_for_mapping(struct intel_memory_region *mr, u64 size, u32 type,
 591                          void **out_addr)
 592{
 593        struct drm_i915_gem_object *obj;
 594        void *addr;
 595
 596        obj = i915_gem_object_create_region(mr, size, 0);
 597        if (IS_ERR(obj)) {
 598                if (PTR_ERR(obj) == -ENOSPC) /* Stolen memory */
 599                        return ERR_PTR(-ENODEV);
 600                return obj;
 601        }
 602
 603        addr = i915_gem_object_pin_map(obj, type);
 604        if (IS_ERR(addr)) {
 605                i915_gem_object_put(obj);
 606                if (PTR_ERR(addr) == -ENXIO)
 607                        return ERR_PTR(-ENODEV);
 608                return addr;
 609        }
 610
 611        *out_addr = addr;
 612        return obj;
 613}
 614
 615static int wrap_ktime_compare(const void *A, const void *B)
 616{
 617        const ktime_t *a = A, *b = B;
 618
 619        return ktime_compare(*a, *b);
 620}
 621
 622static void igt_memcpy_long(void *dst, const void *src, size_t size)
 623{
 624        unsigned long *tmp = dst;
 625        const unsigned long *s = src;
 626
 627        size = size / sizeof(unsigned long);
 628        while (size--)
 629                *tmp++ = *s++;
 630}
 631
 632static inline void igt_memcpy(void *dst, const void *src, size_t size)
 633{
 634        memcpy(dst, src, size);
 635}
 636
 637static inline void igt_memcpy_from_wc(void *dst, const void *src, size_t size)
 638{
 639        i915_memcpy_from_wc(dst, src, size);
 640}
 641
 642static int _perf_memcpy(struct intel_memory_region *src_mr,
 643                        struct intel_memory_region *dst_mr,
 644                        u64 size, u32 src_type, u32 dst_type)
 645{
 646        struct drm_i915_private *i915 = src_mr->i915;
 647        const struct {
 648                const char *name;
 649                void (*copy)(void *dst, const void *src, size_t size);
 650                bool skip;
 651        } tests[] = {
 652                {
 653                        "memcpy",
 654                        igt_memcpy,
 655                },
 656                {
 657                        "memcpy_long",
 658                        igt_memcpy_long,
 659                },
 660                {
 661                        "memcpy_from_wc",
 662                        igt_memcpy_from_wc,
 663                        !i915_has_memcpy_from_wc(),
 664                },
 665        };
 666        struct drm_i915_gem_object *src, *dst;
 667        void *src_addr, *dst_addr;
 668        int ret = 0;
 669        int i;
 670
 671        src = create_region_for_mapping(src_mr, size, src_type, &src_addr);
 672        if (IS_ERR(src)) {
 673                ret = PTR_ERR(src);
 674                goto out;
 675        }
 676
 677        dst = create_region_for_mapping(dst_mr, size, dst_type, &dst_addr);
 678        if (IS_ERR(dst)) {
 679                ret = PTR_ERR(dst);
 680                goto out_unpin_src;
 681        }
 682
 683        for (i = 0; i < ARRAY_SIZE(tests); ++i) {
 684                ktime_t t[5];
 685                int pass;
 686
 687                if (tests[i].skip)
 688                        continue;
 689
 690                for (pass = 0; pass < ARRAY_SIZE(t); pass++) {
 691                        ktime_t t0, t1;
 692
 693                        t0 = ktime_get();
 694
 695                        tests[i].copy(dst_addr, src_addr, size);
 696
 697                        t1 = ktime_get();
 698                        t[pass] = ktime_sub(t1, t0);
 699                }
 700
 701                sort(t, ARRAY_SIZE(t), sizeof(*t), wrap_ktime_compare, NULL);
 702                pr_info("%s src(%s, %s) -> dst(%s, %s) %14s %4llu KiB copy: %5lld MiB/s\n",
 703                        __func__,
 704                        src_mr->name,
 705                        repr_type(src_type),
 706                        dst_mr->name,
 707                        repr_type(dst_type),
 708                        tests[i].name,
 709                        size >> 10,
 710                        div64_u64(mul_u32_u32(4 * size,
 711                                              1000 * 1000 * 1000),
 712                                  t[1] + 2 * t[2] + t[3]) >> 20);
 713
 714                cond_resched();
 715        }
 716
 717        i915_gem_object_unpin_map(dst);
 718        i915_gem_object_put(dst);
 719out_unpin_src:
 720        i915_gem_object_unpin_map(src);
 721        i915_gem_object_put(src);
 722
 723        i915_gem_drain_freed_objects(i915);
 724out:
 725        if (ret == -ENODEV)
 726                ret = 0;
 727
 728        return ret;
 729}
 730
 731static int perf_memcpy(void *arg)
 732{
 733        struct drm_i915_private *i915 = arg;
 734        static const u32 types[] = {
 735                I915_MAP_WB,
 736                I915_MAP_WC,
 737        };
 738        static const u32 sizes[] = {
 739                SZ_4K,
 740                SZ_64K,
 741                SZ_4M,
 742        };
 743        struct intel_memory_region *src_mr, *dst_mr;
 744        int src_id, dst_id;
 745        int i, j, k;
 746        int ret;
 747
 748        for_each_memory_region(src_mr, i915, src_id) {
 749                for_each_memory_region(dst_mr, i915, dst_id) {
 750                        for (i = 0; i < ARRAY_SIZE(sizes); ++i) {
 751                                for (j = 0; j < ARRAY_SIZE(types); ++j) {
 752                                        for (k = 0; k < ARRAY_SIZE(types); ++k) {
 753                                                ret = _perf_memcpy(src_mr,
 754                                                                   dst_mr,
 755                                                                   sizes[i],
 756                                                                   types[j],
 757                                                                   types[k]);
 758                                                if (ret)
 759                                                        return ret;
 760                                        }
 761                                }
 762                        }
 763                }
 764        }
 765
 766        return 0;
 767}
 768
 769int intel_memory_region_mock_selftests(void)
 770{
 771        static const struct i915_subtest tests[] = {
 772                SUBTEST(igt_mock_fill),
 773                SUBTEST(igt_mock_contiguous),
 774        };
 775        struct intel_memory_region *mem;
 776        struct drm_i915_private *i915;
 777        int err;
 778
 779        i915 = mock_gem_device();
 780        if (!i915)
 781                return -ENOMEM;
 782
 783        mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0);
 784        if (IS_ERR(mem)) {
 785                pr_err("failed to create memory region\n");
 786                err = PTR_ERR(mem);
 787                goto out_unref;
 788        }
 789
 790        err = i915_subtests(tests, mem);
 791
 792        intel_memory_region_put(mem);
 793out_unref:
 794        drm_dev_put(&i915->drm);
 795        return err;
 796}
 797
 798int intel_memory_region_live_selftests(struct drm_i915_private *i915)
 799{
 800        static const struct i915_subtest tests[] = {
 801                SUBTEST(igt_lmem_create),
 802                SUBTEST(igt_lmem_write_cpu),
 803                SUBTEST(igt_lmem_write_gpu),
 804        };
 805
 806        if (!HAS_LMEM(i915)) {
 807                pr_info("device lacks LMEM support, skipping\n");
 808                return 0;
 809        }
 810
 811        if (intel_gt_is_wedged(&i915->gt))
 812                return 0;
 813
 814        return i915_live_subtests(tests, i915);
 815}
 816
 817int intel_memory_region_perf_selftests(struct drm_i915_private *i915)
 818{
 819        static const struct i915_subtest tests[] = {
 820                SUBTEST(perf_memcpy),
 821        };
 822
 823        if (intel_gt_is_wedged(&i915->gt))
 824                return 0;
 825
 826        return i915_live_subtests(tests, i915);
 827}
 828