Am 17.05.2017 um 10:46 schrieb zhoucm1:


On 2017年05月17日 16:40, Christian König wrote:
Am 17.05.2017 um 10:01 schrieb Michel Dänzer:
On 17/05/17 04:13 PM, zhoucm1 wrote:
On 2017年05月17日 14:57, Michel Dänzer wrote:
On 17/05/17 01:28 PM, zhoucm1 wrote:
On 2017年05月17日 11:15, Michel Dänzer wrote:
On 17/05/17 12:04 PM, zhoucm1 wrote:
On 2017年05月17日 09:18, Michel Dänzer wrote:
On 16/05/17 06:25 PM, Chunming Zhou wrote:
Change-Id: I8eb6d7f558da05510e429d3bf1d48c8cec6c1977
Signed-off-by: Chunming Zhou <david1.z...@amd.com>

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
index bca1fb5..f3e7525 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
@@ -2547,6 +2547,9 @@ int amdgpu_vm_ioctl(struct drm_device *dev,
void *data, struct drm_file *filp)
         case AMDGPU_VM_OP_UNRESERVE_VMID:
             amdgpu_vm_free_reserved_vmid(adev, &fpriv->vm,
AMDGPU_GFXHUB);
             break;
+    case AMDGPU_VM_OP_RESET:
+        fpriv->vram_lost_counter =
atomic_read(&adev->vram_lost_counter);
+        break;
How do you envision the UMDs using this? I can mostly think of them calling this ioctl when a context is created or destroyed. But that would also allow any other remaining contexts using the same DRM file descriptor to use all ioctls again. So, I think there needs to be a
vram_lost_counter in struct amdgpu_ctx instead of in struct
amdgpu_fpriv.
struct amdgpu_fpriv for vram_lost_counter is proper place, especially
for ioctl return value.
if you need to reset ctx one by one, we can mark all contexts of that
vm, and then reset by userspace.
I'm not following. With vram_lost_counter in amdgpu_fpriv, if any
context calls this ioctl, all other contexts using the same file
descriptor will also be considered safe again, right?
Yes, but it really depends on userspace requirement, if you need to
reset ctx one by one, we can mark all contexts of that vm to guilty, and
then reset one context by userspace.
Still not sure what you mean by that.

E.g. what do you mean by "guilty"? I thought that refers to the context which caused a hang. But it seems like you're using it to refer to any
context which hasn't reacted yet to VRAM contents being lost.
When vram is lost, we treat all contexts need to reset.
Essentially, your patches only track VRAM contents being lost per file
descriptor, not per context. I'm not sure (rather skeptical) that this
is suitable for OpenGL UMDs, since state is usually tracked per context.
Marek / Nicolai?

Oh, yeah that's a good point.

The problem with tracking it per context is that Vulkan also wants the ENODEV on the amdgpu_gem_va_ioct() and amdgpu_info_ioctl() which are context less.

But thinking more about this blocking those two doesn't make much sense. The VM content can be restored and why should be disallow reading GPU info?
I can re-paste the Vulkan APIs requiring ENODEV:
"

The Vulkan APIs listed below could return VK_ERROR_DEVICE_LOST according to the spec.

I tries to provide a list of u/k interfaces that could be called for each vk API.


Well those are the Vulkan requirements, but that doesn't necessary mean we must follow that on the kernel side. Keep in mind that Vulkan can't made any requirements towards the kernel driver.

IIRC we already have a query Vulkan can use to figure out if a GPU reset happened or not. So they could use that instead.

Regards,
Christian.

vkCreateDevice

-amdgpu_device_initialize.

-amdgpu_query_gpu_info

vkQueueSubmit

-amdgpu_cs_submit

vkWaitForFences

                amdgpu_cs_wait_fences

vkGetEventStatus

vkQueueWaitIdle

vkDeviceWaitIdle

vkGetQueryPoolResults**

                amdgpu_cs_query_Fence_status

vkQueueBindSparse**

                amdgpu_bo_va_op

                amdgpu_bo_va_op_raw

vkCreateSwapchainKHR**

vkAcquireNextImageKHR**

vkQueuePresentKHR

                Not related with u/k interface.**

**

Besides those listed above, I think amdgpu_cs_signal_Sem/amdgpu_cs_wait_sem should respond to gpu reset as well."

Christian.




_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx


_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

Reply via email to