Ping
Andrey
On 2022-05-13 11:41, Andrey Grodzovsky wrote:
Yes, exactly that's the idea.
Basically the reset domain knowns which amdgpu devices it needs to
reset together.
If you then represent that so that you always have a hive even when
you only have one device in it, or if you put an array of devices
which needs to be reset together into the reset domain doesn't matter.
Maybe go for the later approach, that is probably a bit cleaner and
less code to change.
Christian.
Unfortunately this approach raises also a few difficulties -
First - if holding array of devices in reset_domain then when you come
to GPU reset function you don't really know which adev is the one
triggered the reset and this is actually essential to some procedures
like emergency restart.
Second - in XGMI case we must take into account that one of the hive
members might go away in runtime (i could do echo 1 >
/sysfs/pci_id/remove on it for example at any moment) - so now we need
to maintain this array and mark such entry with NULL probably on XGMI
node removal , and then there might be hot insertion and all this adds
more complications.
I now tend to prefer your initial solution for it's simplicity and the
result will be what we need -
"E.g. in the reset code (either before or after the reset, that's
debatable) you do something like this:
for (i = 0; i < num_ring; ++i)
cancel_delayed_work(ring[i]->scheduler....)
cancel_work(adev->ras_work);
cancel_work(adev->iofault_work);
cancel_work(adev->debugfs_work);
"
Let me know what you think.
Andrey