Hi Steve,

On Fri, 8 Dec 2023 14:28:05 +0000
Steven Price <steven.pr...@arm.com> wrote:

> > +/**
> > + * alloc_pt() - Custom page table allocator
> > + * @cookie: Cookie passed at page table allocation time.
> > + * @size: Size of the page table. This size should be fixed,
> > + * and determined at creation time based on the granule size.
> > + * @gfp: GFP flags.
> > + *
> > + * We want a custom allocator so we can use a cache for page table
> > + * allocations and amortize the cost of the over-reservation that's
> > + * done to allow asynchronous VM operations.
> > + *
> > + * Return: non-NULL on success, NULL if the allocation failed for any
> > + * reason.
> > + */
> > +static void *alloc_pt(void *cookie, size_t size, gfp_t gfp)
> > +{
> > +   struct panthor_vm *vm = cookie;
> > +   void *page;
> > +
> > +   /* Allocation of the root page table happening during init. */
> > +   if (unlikely(!vm->pgtbl_ops)) {  
> 
> I'm not that keen on using pgtbl_ops as the proxy for this. Can we use
> root_page_table instead?

Definitely, I actually intended to test ->root_page_table when I
introduced this field, but somehow forgot to update this part of the
code.

> 
> At the moment if the IOMMU code ever did multiple allocations during
> alloc_io_pgtable_ops() then we'd overwrite root_page_table and screw up
> on the free path.
> 
> If we use root_page_table == NULL as the check then things will
> 'cleanly' fail by falling through to the non-root case in that case.
> 
> Of course this really looks like we should have had a different
> allocator for the root table but I'm not (re)opening that can of worms! ;)
> 
> And of course it doesn't make any sense for the IOMMU code to do
> multiple allocations so this is all rather academic - but maybe one day
> there will be a different page table structure (16K pages maybe?).
> 
> > +           struct page *p;
> > +
> > +           drm_WARN_ON(&vm->ptdev->base, vm->op_ctx);
> > +           p = alloc_pages_node(dev_to_node(vm->ptdev->base.dev),
> > +                                gfp | __GFP_ZERO, get_order(size));
> > +           page = p ? page_address(p) : NULL;
> > +           vm->root_page_table = page;
> > +           return page;
> > +   }
> > +
> > +   /* We're not supposed to have anything bigger than 4k here, because we 
> > picked a
> > +    * 4k granule size at init time.
> > +    */
> > +   if (drm_WARN_ON(&vm->ptdev->base, size != SZ_4K))
> > +           return NULL;
> > +
> > +   /* We must have some op_ctx attached to the VM and it must have at 
> > least one
> > +    * free page.
> > +    */
> > +   if (drm_WARN_ON(&vm->ptdev->base, !vm->op_ctx) ||
> > +       drm_WARN_ON(&vm->ptdev->base,
> > +                   vm->op_ctx->rsvd_page_tables.ptr >= 
> > vm->op_ctx->rsvd_page_tables.count))
> > +           return NULL;
> > +
> > +   page = 
> > vm->op_ctx->rsvd_page_tables.pages[vm->op_ctx->rsvd_page_tables.ptr++];
> > +   memset(page, 0, SZ_4K);
> > +
> > +   /* Page table entries don't use virtual addresses, which trips out
> > +    * kmemleak. kmemleak_alloc_phys() might work, but physical addresses
> > +    * are mixed with other fields, and I fear kmemleak won't detect that
> > +    * either.
> > +    *
> > +    * Let's just ignore memory passed to the page-table driver for now.
> > +    */
> > +   kmemleak_ignore(page);
> > +   return page;
> > +}
> > +
> > +/**
> > + * @free_pt() - Custom page table free function
> > + * @cookie: Cookie passed at page table allocation time.
> > + * @data: Page table to free.
> > + * @size: Size of the page table. This size should be fixed,
> > + * and determined at creation time based on the granule size.
> > + */
> > +static void free_pt(void *cookie, void *data, size_t size)
> > +{
> > +   struct panthor_vm *vm = cookie;
> > +
> > +   if (unlikely(vm->root_page_table == data)) {
> > +           free_pages((unsigned long)data, get_order(size));  
> 
> Maybe add "vm->root_page_table = NULL;"?

Sure.

> 
> > +           return;
> > +   }

[...]

> > +/**
> > + * panthor_vm_alloc_va() - Allocate a region in the auto-va space
> > + * @VM: VM to allocate a region on.
> > + * @size: Size of the region.  
> 
> kerneldoc needs updating for the new arguments.

Will fix.

> 
> > + *
> > + * Some GPU objects, like heap chunks, are fully managed by the kernel and
> > + * need to be mapped to the userspace VM, in the region reserved for kernel
> > + * objects.
> > + *
> > + * This function takes care of allocating a region in this reserved space.
> > + *
> > + * Return: A valid pointer on success, and ERR_PTR() otherwise.  
> 
> Returns an error code not a pointer.

And that too.

> 
> > + */
> > +int
> > +panthor_vm_alloc_va(struct panthor_vm *vm, u64 va, u64 size,
> > +               struct drm_mm_node *va_node)
> > +{
> > +   int ret;
> > +
> > +   if (!size || (size & ~PAGE_MASK))
> > +           return -EINVAL;
> > +
> > +   if (va != PANTHOR_VM_KERNEL_AUTO_VA && (va & ~PAGE_MASK))
> > +           return -EINVAL;
> > +
> > +   mutex_lock(&vm->mm_lock);
> > +   if (va != PANTHOR_VM_KERNEL_AUTO_VA) {
> > +           memset(va_node, 0, sizeof(*va_node));  
> 
> This memset() seems redundant.

If we assume the va_node is initialized to zero, it's indeed redundant.
I'll update the doc to make this a caller requirement.

> I certainly can't see why it's only
> required on this path.

drm_mm_insert_node_in_range() seems to assign all fields explicitly,
while, according to the doc [1], drm_mm_reserve_node() wants the caller
to make sure the struct is zero-initialized, except for the start and
size fields.

> 
> > +           va_node->start = va;
> > +           va_node->size = size;
> > +           ret = drm_mm_reserve_node(&vm->mm, va_node);
> > +   } else {
> > +           ret = drm_mm_insert_node_in_range(&vm->mm, va_node, size,
> > +                                             size >= SZ_2M ? SZ_2M : SZ_4K,
> > +                                             0, vm->kernel_auto_va.start,
> > +                                             vm->kernel_auto_va.end,
> > +                                             DRM_MM_INSERT_BEST);
> > +   }
> > +   mutex_unlock(&vm->mm_lock);
> > +
> > +   return ret;
> > +}

[...]

> > +/*
> > + * Only 32 VMs per open file. If that becomes a limiting factor, we can
> > + * increase this number.
> > + */
> > +#define PANTHOR_MAX_VMS_PER_FILE   32
> > +
> > +/**
> > + * panthor_vm_pool_create_vm() - Create a VM
> > + * @pool: The VM to create this VM on.
> > + * @kernel_va_start: Start of the region reserved for kernel objects.
> > + * @kernel_va_range: Size of the region reserved for kernel objects.
> > + *
> > + * Return: 0 on success, a negative error code otherwise.  
> 
> Actually returns the (positive) id on success.

Will fix.

> 
> > + */
> > +int panthor_vm_pool_create_vm(struct panthor_device *ptdev,
> > +                         struct panthor_vm_pool *pool,
> > +                         struct drm_panthor_vm_create *args)
> > +{
> > +   u64 kernel_va_start, kernel_va_range;
> > +   struct panthor_vm *vm;
> > +   int ret;
> > +   u32 id;
> > +
> > +   ret = panthor_vm_create_check_args(ptdev, args, &kernel_va_start, 
> > &kernel_va_range);
> > +   if (ret)
> > +           return ret;
> > +
> > +   vm = panthor_vm_create(ptdev, false, kernel_va_start, kernel_va_range,
> > +                          kernel_va_start, kernel_va_range);
> > +   if (IS_ERR(vm))
> > +           return PTR_ERR(vm);
> > +
> > +   ret = xa_alloc(&pool->xa, &id, vm,
> > +                  XA_LIMIT(1, PANTHOR_MAX_VMS_PER_FILE), GFP_KERNEL);
> > +
> > +   if (ret) {
> > +           panthor_vm_put(vm);
> > +           return ret;
> > +   }
> > +
> > +   args->user_va_range = kernel_va_start;
> > +   return id;
> > +}

[...]

> > +/**
> > + * panthor_vm_put() - Release a reference on a VM
> > + * @vm: VM to release the reference on. Can be NULL.
> > + */
> > +void panthor_vm_put(struct panthor_vm *vm)
> > +{
> > +   static_assert(offsetof(struct panthor_vm, base) == 0);  
> 
> Yuk! ;)
> 
> I'd prefer:
> 
>   drm_gpuvm_put(vm ? &vm->base : NULL);
> 
> which my compiler turns into the same thing rather than relying on the
> type punning. You can keep the static_assert if you like, but I don't
> like relying on it for correct code generation. Although I'll admit I
> couldn't actually get the compiler to produce incorrect code when I tried.

Sure, I'll pick your suggestion here.

> 
> > +   drm_gpuvm_put(&vm->base);
> > +}

[...]

> > +
> > +/**
> > + * panthor_vm_prepare_mapped_bos_resvs() - Prepare resvs on VM BOs.
> > + * @exec: Locking/preparation context.
> > + * @vm: VM targeted by the GPU job.
> > + * @slot_count: Number of slots to reserve.
> > + *
> > + * GPU jobs assume all BOs bound to the VM at the time the job is submitted
> > + * are available when the job is executed. In order to guarantee that, we
> > + * need to reserve a slot on all BOs mapped to a VM and update this slot 
> > with
> > + * the job fence after its submission.
> > + *
> > + * Return: 0 on success, a negative error code otherwise.
> > + */
> > +int panthor_vm_prepare_mapped_bos_resvs(struct drm_exec *exec, struct 
> > panthor_vm *vm,
> > +                                   u32 slot_count)
> > +{
> > +   int ret;
> > +
> > +   /* Acquire the VM lock an reserve a slot for this GPU job. */
> > +   ret = drm_gpuvm_prepare_vm(&vm->base, exec, slot_count);
> > +   if (ret)
> > +           return ret;
> > +
> > +   /* VM operations are not protected by the VM resv-lock. We need to
> > +    * take the op_lock to make sure the shared_bos list is not updated
> > +    * while we're walking it.
> > +    */  
> 
> Is the above comment stale? AFAIK the shared_bos list doesn't exist
> anymore and this doesn't appear to relate to anything here.

Oops, indeed. That predates to transition to drm_gpuvm for the VM <-> BO
association.

Thanks for the review!

[1]https://elixir.bootlin.com/linux/latest/source/drivers/gpu/drm/drm_mm.c#L441

Reply via email to