Add a new kunit test gpu_test_buddy_alloc_range() that exercises the
__gpu_buddy_alloc_range() exact-range allocation path, triggered when
start + size == end with flags=0.

The test covers:
- Basic exact-range allocation of the full mm
- Exact-range allocation of equal sub-ranges (quarters)
- Minimum chunk-size exact ranges at start, middle, and end offsets
- Non power-of-two mm size with multiple roots, including cross-root
  exact-range allocation
- Randomized exact-range allocations of N contiguous page-aligned
  slices in random order
- Negative: partially allocated range must reject overlapping exact
  alloc
- Negative: checkerboard allocation pattern rejects exact range over
  partially occupied pairs
- Negative: misaligned start, unaligned size, and out-of-bounds end
- Free and re-allocate the same exact range across multiple iterations
- Various power-of-two exact ranges at natural alignment

Cc: Christian König <[email protected]>
Cc: Arunpravin Paneer Selvam <[email protected]>
Suggested-by: Matthew Auld <[email protected]>
Signed-off-by: Sanjay Yadav <[email protected]>
---
 drivers/gpu/tests/gpu_buddy_test.c | 327 +++++++++++++++++++++++++++++
 1 file changed, 327 insertions(+)

diff --git a/drivers/gpu/tests/gpu_buddy_test.c 
b/drivers/gpu/tests/gpu_buddy_test.c
index 5429010f34d3..9738fd09518f 100644
--- a/drivers/gpu/tests/gpu_buddy_test.c
+++ b/drivers/gpu/tests/gpu_buddy_test.c
@@ -362,6 +362,332 @@ static void gpu_test_buddy_alloc_range_bias(struct kunit 
*test)
        gpu_buddy_fini(&mm);
 }
 
+static void gpu_test_buddy_alloc_range(struct kunit *test)
+{
+       GPU_RND_STATE(prng, random_seed);
+       struct gpu_buddy_block *block;
+       struct gpu_buddy mm;
+       u32 mm_size, total;
+       LIST_HEAD(blocks);
+       LIST_HEAD(tmp);
+       u32 ps = SZ_4K;
+       int ret;
+
+       mm_size = SZ_16M;
+
+       KUNIT_ASSERT_FALSE_MSG(test, gpu_buddy_init(&mm, mm_size, ps),
+                              "buddy_init failed\n");
+
+       /*
+        * Basic exact-range allocation.
+        * Allocate the entire mm as one exact range (start + size == end).
+        * This is the simplest case exercising __gpu_buddy_alloc_range.
+        */
+       ret = gpu_buddy_alloc_blocks(&mm, 0, mm_size, mm_size, ps, &blocks, 0);
+       KUNIT_ASSERT_EQ_MSG(test, ret, 0,
+                           "exact-range alloc of full mm failed\n");
+
+       total = 0;
+       list_for_each_entry(block, &blocks, link) {
+               u64 offset = gpu_buddy_block_offset(block);
+               u64 bsize = gpu_buddy_block_size(&mm, block);
+
+               KUNIT_EXPECT_TRUE_MSG(test, offset + bsize <= (u64)mm_size,
+                                     "block [%llx, %llx) outside mm\n", 
offset, offset + bsize);
+               total += (u32)bsize;
+       }
+       KUNIT_EXPECT_EQ(test, total, mm_size);
+       KUNIT_EXPECT_EQ(test, mm.avail, 0ULL);
+
+       /* Full mm should be exhausted */
+       ret = gpu_buddy_alloc_blocks(&mm, 0, ps, ps, ps, &tmp, 0);
+       KUNIT_EXPECT_NE_MSG(test, ret, 0, "alloc should fail when mm is 
full\n");
+
+       gpu_buddy_free_list(&mm, &blocks, 0);
+       KUNIT_EXPECT_EQ(test, mm.avail, (u64)mm_size);
+       gpu_buddy_fini(&mm);
+
+       /*
+        * Exact-range allocation of sub-ranges.
+        * Split the mm into four equal quarters and allocate each as an exact
+        * range. Validates splitting and non-overlapping exact allocations.
+        */
+       KUNIT_ASSERT_FALSE(test, gpu_buddy_init(&mm, mm_size, ps));
+
+       {
+               u32 quarter = mm_size / 4;
+               int i;
+
+               for (i = 0; i < 4; i++) {
+                       u32 start = i * quarter;
+                       u32 end = start + quarter;
+
+                       ret = gpu_buddy_alloc_blocks(&mm, start, end, quarter, 
ps, &blocks, 0);
+                       KUNIT_ASSERT_EQ_MSG(test, ret, 0,
+                                           "exact-range alloc quarter %d [%x, 
%x) failed\n",
+                                           i, start, end);
+               }
+               KUNIT_EXPECT_EQ(test, mm.avail, 0ULL);
+               gpu_buddy_free_list(&mm, &blocks, 0);
+       }
+
+       gpu_buddy_fini(&mm);
+
+       /*
+        * Minimum chunk-size exact range at various offsets.
+        * Allocate single-page exact ranges at the start, middle and end.
+        */
+       KUNIT_ASSERT_FALSE(test, gpu_buddy_init(&mm, mm_size, ps));
+
+       ret = gpu_buddy_alloc_blocks(&mm, 0, ps, ps, ps, &blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       ret = gpu_buddy_alloc_blocks(&mm, mm_size / 2, mm_size / 2 + ps, ps, 
ps, &blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       ret = gpu_buddy_alloc_blocks(&mm, mm_size - ps, mm_size, ps, ps, 
&blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       total = 0;
+       list_for_each_entry(block, &blocks, link)
+               total += (u32)gpu_buddy_block_size(&mm, block);
+       KUNIT_EXPECT_EQ(test, total, 3 * ps);
+
+       gpu_buddy_free_list(&mm, &blocks, 0);
+       gpu_buddy_fini(&mm);
+
+       /*
+        * Non power-of-two mm size (multiple roots).
+        * Exact-range allocations that span root boundaries must still work.
+        */
+       mm_size = SZ_4M + SZ_2M + SZ_1M; /* 7 MiB, three roots */
+
+       KUNIT_ASSERT_FALSE(test, gpu_buddy_init(&mm, mm_size, ps));
+       KUNIT_EXPECT_GT(test, mm.n_roots, 1U);
+
+       /* Allocate first 4M root exactly */
+       ret = gpu_buddy_alloc_blocks(&mm, 0, SZ_4M, SZ_4M, ps, &blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       /* Allocate second root (4M-6M) exactly */
+       ret = gpu_buddy_alloc_blocks(&mm, SZ_4M, SZ_4M + SZ_2M, SZ_2M, ps, 
&blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       /* Allocate third root (6M-7M) exactly */
+       ret = gpu_buddy_alloc_blocks(&mm, SZ_4M + SZ_2M, mm_size, SZ_1M, ps, 
&blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       KUNIT_EXPECT_EQ(test, mm.avail, 0ULL);
+       gpu_buddy_free_list(&mm, &blocks, 0);
+
+       /* Cross-root exact-range: the entire non-pot mm */
+       ret = gpu_buddy_alloc_blocks(&mm, 0, mm_size, mm_size, ps, &blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+       KUNIT_EXPECT_EQ(test, mm.avail, 0ULL);
+
+       gpu_buddy_free_list(&mm, &blocks, 0);
+       gpu_buddy_fini(&mm);
+
+       /*
+        * Randomized exact-range allocations.
+        * Divide the mm into N random-sized, contiguous, page-aligned slices
+        * and allocate each as an exact range in random order.
+        */
+       mm_size = SZ_16M;
+       KUNIT_ASSERT_FALSE(test, gpu_buddy_init(&mm, mm_size, ps));
+
+       {
+#define N_RAND_RANGES 16
+               u32 ranges[N_RAND_RANGES + 1]; /* boundaries */
+               u32 order_arr[N_RAND_RANGES];
+               u32 remaining = mm_size;
+               int i;
+
+               ranges[0] = 0;
+               for (i = 0; i < N_RAND_RANGES - 1; i++) {
+                       u32 max_chunk = remaining - (N_RAND_RANGES - 1 - i) * 
ps;
+                       u32 sz = max(round_up(prandom_u32_state(&prng) % 
max_chunk, ps), ps);
+
+                       ranges[i + 1] = ranges[i] + sz;
+                       remaining -= sz;
+               }
+               ranges[N_RAND_RANGES] = mm_size;
+
+               /* Create a random order */
+               for (i = 0; i < N_RAND_RANGES; i++)
+                       order_arr[i] = i;
+               for (i = N_RAND_RANGES - 1; i > 0; i--) {
+                       u32 j = prandom_u32_state(&prng) % (i + 1);
+                       u32 tmp_val = order_arr[i];
+
+                       order_arr[i] = order_arr[j];
+                       order_arr[j] = tmp_val;
+               }
+
+               for (i = 0; i < N_RAND_RANGES; i++) {
+                       u32 idx = order_arr[i];
+                       u32 start = ranges[idx];
+                       u32 end = ranges[idx + 1];
+                       u32 sz = end - start;
+
+                       ret = gpu_buddy_alloc_blocks(&mm, start, end, sz, ps, 
&blocks, 0);
+                       KUNIT_ASSERT_EQ_MSG(test, ret, 0,
+                                           "random exact-range [%x, %x) sz=%x 
failed\n",
+                                           start, end, sz);
+               }
+
+               KUNIT_EXPECT_EQ(test, mm.avail, 0ULL);
+               gpu_buddy_free_list(&mm, &blocks, 0);
+#undef N_RAND_RANGES
+       }
+
+       gpu_buddy_fini(&mm);
+
+       /*
+        * Negative case - partially allocated range.
+        * Allocate the first half, then try to exact-range allocate the full
+        * mm. This must fail because the first half is already occupied.
+        */
+       mm_size = SZ_16M;
+       KUNIT_ASSERT_FALSE(test, gpu_buddy_init(&mm, mm_size, ps));
+
+       ret = gpu_buddy_alloc_blocks(&mm, 0, mm_size / 2, mm_size / 2, ps, 
&blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       ret = gpu_buddy_alloc_blocks(&mm, 0, mm_size, mm_size, ps, &tmp, 0);
+       KUNIT_EXPECT_NE_MSG(test, ret, 0,
+                           "exact-range alloc should fail when range is 
partially used\n");
+
+       /* Also try the already-occupied sub-range directly */
+       ret = gpu_buddy_alloc_blocks(&mm, 0, mm_size / 2, mm_size / 2, ps, 
&tmp, 0);
+       KUNIT_EXPECT_NE_MSG(test, ret, 0,
+                           "double alloc of same exact range should fail\n");
+
+       /* The free second half should still be allocatable */
+       ret = gpu_buddy_alloc_blocks(&mm, mm_size / 2, mm_size, mm_size / 2, 
ps, &blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       KUNIT_EXPECT_EQ(test, mm.avail, 0ULL);
+       gpu_buddy_free_list(&mm, &blocks, 0);
+       gpu_buddy_fini(&mm);
+
+       /*
+        * Negative case - checkerboard partial allocation.
+        * Allocate every other page-sized chunk in a small mm, then try to
+        * exact-range allocate a range covering two pages (one allocated, one
+        * free). This must fail.
+        */
+       mm_size = SZ_64K;
+       KUNIT_ASSERT_FALSE(test, gpu_buddy_init(&mm, mm_size, ps));
+
+       {
+               u32 off;
+
+               for (off = 0; off < mm_size; off += 2 * ps) {
+                       ret = gpu_buddy_alloc_blocks(&mm, off, off + ps, ps, 
ps, &blocks, 0);
+                       KUNIT_ASSERT_EQ(test, ret, 0);
+               }
+
+               /* Try exact range over a pair [allocated, free] */
+               ret = gpu_buddy_alloc_blocks(&mm, 0, 2 * ps, 2 * ps, ps, &tmp, 
0);
+               KUNIT_EXPECT_NE_MSG(test, ret, 0,
+                                   "exact-range over partially allocated pair 
should fail\n");
+
+               /* The free pages individually should still work */
+               ret = gpu_buddy_alloc_blocks(&mm, ps, 2 * ps, ps, ps, &blocks, 
0);
+               KUNIT_ASSERT_EQ(test, ret, 0);
+
+               gpu_buddy_free_list(&mm, &blocks, 0);
+       }
+
+       gpu_buddy_fini(&mm);
+
+       /* Negative case - misaligned start/end/size */
+       mm_size = SZ_16M;
+       KUNIT_ASSERT_FALSE(test, gpu_buddy_init(&mm, mm_size, ps));
+
+       /* start not aligned to chunk_size */
+       ret = gpu_buddy_alloc_blocks(&mm, ps / 2, ps / 2 + ps, ps, ps, &tmp, 0);
+       KUNIT_EXPECT_NE(test, ret, 0);
+
+       /* size not aligned */
+       ret = gpu_buddy_alloc_blocks(&mm, 0, ps + 1, ps + 1, ps, &tmp, 0);
+       KUNIT_EXPECT_NE(test, ret, 0);
+
+       /* end exceeds mm size */
+       ret = gpu_buddy_alloc_blocks(&mm, mm_size, mm_size + ps, ps, ps, &tmp, 
0);
+       KUNIT_EXPECT_NE(test, ret, 0);
+
+       gpu_buddy_fini(&mm);
+
+       /*
+        * Free and re-allocate the same exact range.
+        * This exercises merge-on-free followed by exact-range re-split.
+        */
+       mm_size = SZ_16M;
+       KUNIT_ASSERT_FALSE(test, gpu_buddy_init(&mm, mm_size, ps));
+
+       {
+               int i;
+
+               for (i = 0; i < 5; i++) {
+                       ret = gpu_buddy_alloc_blocks(&mm, SZ_4M, SZ_4M + SZ_2M,
+                                                    SZ_2M, ps, &blocks, 0);
+                       KUNIT_ASSERT_EQ_MSG(test, ret, 0,
+                                           "re-alloc iteration %d failed\n", 
i);
+
+                       total = 0;
+                       list_for_each_entry(block, &blocks, link) {
+                               u64 offset = gpu_buddy_block_offset(block);
+                               u64 bsize = gpu_buddy_block_size(&mm, block);
+
+                               KUNIT_EXPECT_GE(test, offset, (u64)SZ_4M);
+                               KUNIT_EXPECT_LE(test, offset + bsize, 
(u64)(SZ_4M + SZ_2M));
+                               total += (u32)bsize;
+                       }
+                       KUNIT_EXPECT_EQ(test, total, SZ_2M);
+
+                       gpu_buddy_free_list(&mm, &blocks, 0);
+               }
+
+               KUNIT_EXPECT_EQ(test, mm.avail, (u64)mm_size);
+       }
+
+       gpu_buddy_fini(&mm);
+
+       /*
+        * Various power-of-two exact ranges within a large mm.
+        * Allocate non-overlapping power-of-two exact ranges at their natural
+        * alignment, validating that the allocator handles different orders.
+        */
+       mm_size = SZ_16M;
+       KUNIT_ASSERT_FALSE(test, gpu_buddy_init(&mm, mm_size, ps));
+
+       /* Allocate 4K at offset 0 */
+       ret = gpu_buddy_alloc_blocks(&mm, 0, SZ_4K, SZ_4K, ps, &blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       /* Allocate 64K at offset 64K */
+       ret = gpu_buddy_alloc_blocks(&mm, SZ_64K, SZ_64K + SZ_64K, SZ_64K, ps, 
&blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       /* Allocate 1M at offset 1M */
+       ret = gpu_buddy_alloc_blocks(&mm, SZ_1M, SZ_1M + SZ_1M, SZ_1M, ps, 
&blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       /* Allocate 4M at offset 4M */
+       ret = gpu_buddy_alloc_blocks(&mm, SZ_4M, SZ_4M + SZ_4M, SZ_4M, ps, 
&blocks, 0);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       total = 0;
+       list_for_each_entry(block, &blocks, link)
+               total += (u32)gpu_buddy_block_size(&mm, block);
+       KUNIT_EXPECT_EQ(test, total, SZ_4K + SZ_64K + SZ_1M + SZ_4M);
+
+       gpu_buddy_free_list(&mm, &blocks, 0);
+       gpu_buddy_fini(&mm);
+}
+
 static void gpu_test_buddy_alloc_clear(struct kunit *test)
 {
        unsigned long n_pages, total, i = 0;
@@ -909,6 +1235,7 @@ static struct kunit_case gpu_buddy_tests[] = {
        KUNIT_CASE(gpu_test_buddy_alloc_pathological),
        KUNIT_CASE(gpu_test_buddy_alloc_contiguous),
        KUNIT_CASE(gpu_test_buddy_alloc_clear),
+       KUNIT_CASE(gpu_test_buddy_alloc_range),
        KUNIT_CASE(gpu_test_buddy_alloc_range_bias),
        KUNIT_CASE_SLOW(gpu_test_buddy_fragmentation_performance),
        KUNIT_CASE(gpu_test_buddy_alloc_exceeds_max_order),
-- 
2.52.0

Reply via email to