On 2/20/19 12:28 AM, Mikhail Gavrilov wrote:
> On Tue, 19 Feb 2019 at 20:24, Grodzovsky, Andrey
> <andrey.grodzov...@amd.com> wrote:
>> Just pull in latest drm-next from here -
>> https://cgit.freedesktop.org/~agd5f/linux/log/?h=amd-staging-drm-next
>>
>> Andrey
> Tested this kernel and result not good for me.
> 1) "amdgpu 0000:0b:00.0: VM_L2_PROTECTION_FAULT_STATUS:0x0070113C"
> happens again. I thought this would fixed.

No, we only fixed the original deadlock with display driver during GPU 
reset. I still didn't have time to go over your captures for the GPU 
page fault.

The deadlock we see here is another deadlock, different from the one 
already fixed. I suggest you open a bugzilla ticket for this and add me 
there so we can track it and take care of it.

Andrey


>
> 2) After it "WARNING: possible circular locking dependency detected" happens.
>
> [  302.266337] ======================================================
> [  302.266338] WARNING: possible circular locking dependency detected
> [  302.266340] 5.0.0-rc1-drm-next-kernel+ #1 Tainted: G         C
> [  302.266341] ------------------------------------------------------
> [  302.266343] kworker/5:2/871 is trying to acquire lock:
> [  302.266345] 000000000abbb16a
> (&(&ring->fence_drv.lock)->rlock){-.-.}, at:
> dma_fence_remove_callback+0x1a/0x60
> [  302.266352]
>                 but task is already holding lock:
> [  302.266353] 000000006e32ba38
> (&(&sched->job_list_lock)->rlock){-.-.}, at: drm_sched_stop+0x34/0x140
> [gpu_sched]
> [  302.266358]
>                 which lock already depends on the new lock.
>
> [  302.266360]
>                 the existing dependency chain (in reverse order) is:
> [  302.266361]
>                 -> #1 (&(&sched->job_list_lock)->rlock){-.-.}:
> [  302.266366]        drm_sched_process_job+0x4d/0x180 [gpu_sched]
> [  302.266368]        dma_fence_signal+0x111/0x1a0
> [  302.266414]        amdgpu_fence_process+0xa3/0x100 [amdgpu]
> [  302.266470]        sdma_v4_0_process_trap_irq+0x6e/0xa0 [amdgpu]
> [  302.266523]        amdgpu_irq_dispatch+0xc0/0x250 [amdgpu]
> [  302.266576]        amdgpu_ih_process+0x84/0xf0 [amdgpu]
> [  302.266628]        amdgpu_irq_handler+0x1b/0x50 [amdgpu]
> [  302.266632]        __handle_irq_event_percpu+0x3f/0x290
> [  302.266635]        handle_irq_event_percpu+0x31/0x80
> [  302.266637]        handle_irq_event+0x34/0x51
> [  302.266639]        handle_edge_irq+0x7c/0x1a0
> [  302.266643]        handle_irq+0xbf/0x100
> [  302.266646]        do_IRQ+0x61/0x120
> [  302.266648]        ret_from_intr+0x0/0x22
> [  302.266651]        cpuidle_enter_state+0xbf/0x470
> [  302.266654]        do_idle+0x1ec/0x280
> [  302.266657]        cpu_startup_entry+0x19/0x20
> [  302.266660]        start_secondary+0x1b3/0x200
> [  302.266663]        secondary_startup_64+0xa4/0xb0
> [  302.266664]
>                 -> #0 (&(&ring->fence_drv.lock)->rlock){-.-.}:
> [  302.266668]        _raw_spin_lock_irqsave+0x49/0x83
> [  302.266670]        dma_fence_remove_callback+0x1a/0x60
> [  302.266673]        drm_sched_stop+0x59/0x140 [gpu_sched]
> [  302.266717]        amdgpu_device_pre_asic_reset+0x4f/0x240 [amdgpu]
> [  302.266761]        amdgpu_device_gpu_recover+0x88/0x7d0 [amdgpu]
> [  302.266822]        amdgpu_job_timedout+0x109/0x130 [amdgpu]
> [  302.266827]        drm_sched_job_timedout+0x40/0x70 [gpu_sched]
> [  302.266831]        process_one_work+0x272/0x5d0
> [  302.266834]        worker_thread+0x50/0x3b0
> [  302.266836]        kthread+0x108/0x140
> [  302.266839]        ret_from_fork+0x27/0x50
> [  302.266840]
>                 other info that might help us debug this:
>
> [  302.266841]  Possible unsafe locking scenario:
>
> [  302.266842]        CPU0                    CPU1
> [  302.266843]        ----                    ----
> [  302.266844]   lock(&(&sched->job_list_lock)->rlock);
> [  302.266846]
> lock(&(&ring->fence_drv.lock)->rlock);
> [  302.266847]
> lock(&(&sched->job_list_lock)->rlock);
> [  302.266849]   lock(&(&ring->fence_drv.lock)->rlock);
> [  302.266850]
>                  *** DEADLOCK ***
>
> [  302.266852] 5 locks held by kworker/5:2/871:
> [  302.266853]  #0: 00000000d133fb6e ((wq_completion)"events"){+.+.},
> at: process_one_work+0x1e9/0x5d0
> [  302.266857]  #1: 000000008a5c3f7e
> ((work_completion)(&(&sched->work_tdr)->work)){+.+.}, at:
> process_one_work+0x1e9/0x5d0
> [  302.266862]  #2: 00000000b9b2c76f (&adev->lock_reset){+.+.}, at:
> amdgpu_device_lock_adev+0x17/0x40 [amdgpu]
> [  302.266908]  #3: 00000000ac637728 (&dqm->lock_hidden){+.+.}, at:
> kgd2kfd_pre_reset+0x30/0x60 [amdgpu]
> [  302.266965]  #4: 000000006e32ba38
> (&(&sched->job_list_lock)->rlock){-.-.}, at: drm_sched_stop+0x34/0x140
> [gpu_sched]
> [  302.266971]
>                 stack backtrace:
> [  302.266975] CPU: 5 PID: 871 Comm: kworker/5:2 Tainted: G         C
>        5.0.0-rc1-drm-next-kernel+ #1
> [  302.266976] Hardware name: System manufacturer System Product
> Name/ROG STRIX X470-I GAMING, BIOS 1103 11/16/2018
> [  302.266980] Workqueue: events drm_sched_job_timedout [gpu_sched]
> [  302.266982] Call Trace:
> [  302.266987]  dump_stack+0x85/0xc0
> [  302.266991]  print_circular_bug.isra.0.cold+0x15c/0x195
> [  302.266994]  __lock_acquire+0x134c/0x1660
> [  302.266998]  ? add_lock_to_list.isra.0+0x67/0xb0
> [  302.267003]  lock_acquire+0xa2/0x1b0
> [  302.267006]  ? dma_fence_remove_callback+0x1a/0x60
> [  302.267011]  _raw_spin_lock_irqsave+0x49/0x83
> [  302.267013]  ? dma_fence_remove_callback+0x1a/0x60
> [  302.267016]  dma_fence_remove_callback+0x1a/0x60
> [  302.267020]  drm_sched_stop+0x59/0x140 [gpu_sched]
> [  302.267065]  amdgpu_device_pre_asic_reset+0x4f/0x240 [amdgpu]
> [  302.267110]  amdgpu_device_gpu_recover+0x88/0x7d0 [amdgpu]
> [  302.267173]  amdgpu_job_timedout+0x109/0x130 [amdgpu]
> [  302.267178]  drm_sched_job_timedout+0x40/0x70 [gpu_sched]
> [  302.267183]  process_one_work+0x272/0x5d0
> [  302.267188]  worker_thread+0x50/0x3b0
> [  302.267191]  kthread+0x108/0x140
> [  302.267194]  ? process_one_work+0x5d0/0x5d0
> [  302.267196]  ? kthread_park+0x90/0x90
> [  302.267199]  ret_from_fork+0x27/0x50
>
> 3) kernel logs flooded by message "[drm:amdgpu_cs_ioctl [amdgpu]]
> *ERROR* Failed to initialize parser -125!" until I killed
> gdm-wayland-session process.
> `# systemctl restart gdm` couldn't helps restart worked Wayland session.
>
> 4) finally after `# init 6` system actually hang.
>
> --
> Best Regards,
> Mike Gavrilov.
_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

Reply via email to