[RFC PATCH 0/1] Add AMDGPU_INFO_GUILTY_APP ioctl

Marek Olšák maraeo at gmail.com
Thu May 4 01:13:26 UTC 2023


On Wed, May 3, 2023, 14:53 André Almeida <andrealmeid at igalia.com> wrote:

> Em 03/05/2023 14:08, Marek Olšák escreveu:
> > GPU hangs are pretty common post-bringup. They are not common per user,
> > but if we gather all hangs from all users, we can have lots and lots of
> > them.
> >
> > GPU hangs are indeed not very debuggable. There are however some things
> > we can do:
> > - Identify the hanging IB by its VA (the kernel should know it)
>
> How can the kernel tell which VA range is being executed? I only found
> that information at mmCP_IB1_BASE_ regs, but as stated in this thread by
> Christian this is not reliable to be read.
>

The kernel receives the VA and the size via the CS ioctl. When user queues
are enabled, the kernel will no longer receive them.


> > - Read and parse the IB to detect memory corruption.
> > - Print active waves with shader disassembly if SQ isn't hung (often
> > it's not).
> >
> > Determining which packet the CP is stuck on is tricky. The CP has 2
> > engines (one frontend and one backend) that work on the same command
> > buffer. The frontend engine runs ahead, executes some packets and
> > forwards others to the backend engine. Only the frontend engine has the
> > command buffer VA somewhere. The backend engine only receives packets
> > from the frontend engine via a FIFO, so it might not be possible to tell
> > where it's stuck if it's stuck.
>
> Do they run at the same asynchronously or does the front end waits the
> back end to execute?
>

They run asynchronously and should run asynchronously for performance, but
they can be synchronized using a special packet (PFP_SYNC_ME).

Marek


> >
> > When the gfx pipeline hangs outside of shaders, making a scandump seems
> > to be the only way to have a chance at finding out what's going wrong,
> > and only AMD-internal versions of hw can be scanned.
> >
> > Marek
> >
> > On Wed, May 3, 2023 at 11:23 AM Christian König
> > <ckoenig.leichtzumerken at gmail.com
> > <mailto:ckoenig.leichtzumerken at gmail.com>> wrote:
> >
> >     Am 03.05.23 um 17:08 schrieb Felix Kuehling:
> >      > Am 2023-05-03 um 03:59 schrieb Christian König:
> >      >> Am 02.05.23 um 20:41 schrieb Alex Deucher:
> >      >>> On Tue, May 2, 2023 at 11:22 AM Timur Kristóf
> >      >>> <timur.kristof at gmail.com <mailto:timur.kristof at gmail.com>>
> wrote:
> >      >>>> [SNIP]
> >      >>>>>>>> In my opinion, the correct solution to those problems
> would be
> >      >>>>>>>> if
> >      >>>>>>>> the kernel could give userspace the necessary information
> >     about
> >      >>>>>>>> a
> >      >>>>>>>> GPU hang before a GPU reset.
> >      >>>>>>>>
> >      >>>>>>>   The fundamental problem here is that the kernel doesn't
> have
> >      >>>>>>> that
> >      >>>>>>> information either. We know which IB timed out and can
> >      >>>>>>> potentially do
> >      >>>>>>> a devcoredump when that happens, but that's it.
> >      >>>>>>
> >      >>>>>> Is it really not possible to know such a fundamental thing
> >     as what
> >      >>>>>> the
> >      >>>>>> GPU was doing when it hung? How are we supposed to do any
> >     kind of
> >      >>>>>> debugging without knowing that?
> >      >>
> >      >> Yes, that's indeed something at least I try to figure out for
> years
> >      >> as well.
> >      >>
> >      >> Basically there are two major problems:
> >      >> 1. When the ASIC is hung you can't talk to the firmware engines
> any
> >      >> more and most state is not exposed directly, but just through
> some
> >      >> fw/hw interface.
> >      >>     Just take a look at how umr reads the shader state from the
> SQ.
> >      >> When that block is hung you can't do that any more and basically
> >     have
> >      >> no chance at all to figure out why it's hung.
> >      >>
> >      >>     Same for other engines, I remember once spending a week
> >     figuring
> >      >> out why the UVD block is hung during suspend. Turned out to be a
> >      >> debugging nightmare because any time you touch any register of
> that
> >      >> block the whole system would hang.
> >      >>
> >      >> 2. There are tons of things going on in a pipeline fashion or
> even
> >      >> completely in parallel. For example the CP is just the beginning
> >     of a
> >      >> rather long pipeline which at the end produces a bunch of pixels.
> >      >>     In almost all cases I've seen you ran into a problem
> somewhere
> >      >> deep in the pipeline and only very rarely at the beginning.
> >      >>
> >      >>>>>>
> >      >>>>>> I wonder what AMD's Windows driver team is doing with this
> >     problem,
> >      >>>>>> surely they must have better tools to deal with GPU hangs?
> >      >>>>> For better or worse, most teams internally rely on scan dumps
> via
> >      >>>>> JTAG
> >      >>>>> which sort of limits the usefulness outside of AMD, but also
> >     gives
> >      >>>>> you
> >      >>>>> the exact state of the hardware when it's hung so the
> >     hardware teams
> >      >>>>> prefer it.
> >      >>>>>
> >      >>>> How does this approach scale? It's not something we can ask
> >     users to
> >      >>>> do, and even if all of us in the radv team had a JTAG device,
> we
> >      >>>> wouldn't be able to play every game that users experience
> >     random hangs
> >      >>>> with.
> >      >>> It doesn't scale or lend itself particularly well to external
> >      >>> development, but that's the current state of affairs.
> >      >>
> >      >> The usual approach seems to be to reproduce a problem in a lab
> and
> >      >> have a JTAG attached to give the hw guys a scan dump and they can
> >      >> then tell you why something didn't worked as expected.
> >      >
> >      > That's the worst-case scenario where you're debugging HW or FW
> >     issues.
> >      > Those should be pretty rare post-bringup. But are there hangs
> caused
> >      > by user mode driver or application bugs that are easier to debug
> and
> >      > probably don't even require a GPU reset? For example most VM
> faults
> >      > can be handled without hanging the GPU. Similarly, a shader in an
> >      > endless loop should not require a full GPU reset. In the KFD
> compute
> >      > case, that's still preemptible and the offending process can be
> >     killed
> >      > with Ctrl-C or debugged with rocm-gdb.
> >
> >     We also have infinite loop in shader abort for gfx and page faults
> are
> >     pretty rare with OpenGL (a bit more often with Vulkan) and can be
> >     handled gracefully on modern hw (they just spam the logs).
> >
> >     The majority of the problems is unfortunately that we really get hard
> >     hangs because of some hw issues. That can be caused by unlucky
> timing,
> >     power management or doing things in an order the hw doesn't expected.
> >
> >     Regards,
> >     Christian.
> >
> >      >
> >      > It's more complicated for graphics because of the more complex
> >      > pipeline and the lack of CWSR. But it should still be possible to
> do
> >      > some debugging without JTAG if the problem is in SW and not HW or
> >     FW.
> >      > It's probably worth improving that debugability without getting
> >      > hung-up on the worst case.
> >      >
> >      > Maybe user mode graphics queues will offer a better way of
> >     recovering
> >      > from these kinds of bugs, if the graphics pipeline can be unstuck
> >      > without a GPU reset, just by killing the offending user mode
> queue.
> >      >
> >      > Regards,
> >      >   Felix
> >      >
> >      >
> >      >>
> >      >> And yes that absolutely doesn't scale.
> >      >>
> >      >> Christian.
> >      >>
> >      >>>
> >      >>> Alex
> >      >>
> >
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.freedesktop.org/archives/amd-gfx/attachments/20230503/af05d1ac/attachment-0001.htm>


More information about the amd-gfx mailing list