Hi,
In pass-through mode it is essential to assign all GPUs connected through
NVLink to the same VM. If only a subset of these GPUs is assigned to a VM it
triggers the unrecoverable error XID 74 during boot corrupting the NVLink state
and rendering the NVLink bridge unusable. Therefore to avoid
I noticed this document
https://docs.nvidia.com/vgpu/16.0/grid-vgpu-release-notes-generic-linux-kvm/index.html#all-nvlink-gpus-must-be-passed-through-to-same-vm
has this to say
In pass through mode, all GPUs connected to each other through NVLink must be
assigned to the same VM. If a subset of
any progress in this gpu question?
in our setup we have supermicro boards with intel xeon gold 6146 + 2 T4
we add extra line in the /etc/default/grub
"rd.driver.blacklist=nouveau nouveau.modeset=0 pci-stub.ids=xxx:xxx
intel_iommu=on"
would be interesting if the nvlink was the showstopper.
Arman
same here ☺️, on Monday will check them.
Michael Jones schrieb am Fr., 4. Sept. 2020, 22:01:
> Yea pass through, I think vgpu you have to pay for driver upgrade with
> nvidia, I've not tried that and don't know the price, didn't find getting
> info on it easy last time I tried.
>
> Have used in
First things I'd check would be what driver is on host and that it's all
nvidia driver all the way make sure nouveau is blacklisted throughout
On Fri, 4 Sep 2020, 21:01 Michael Jones, wrote:
> Yea pass through, I think vgpu you have to pay for driver upgrade with
> nvidia, I've not tried that an
Yea pass through, I think vgpu you have to pay for driver upgrade with
nvidia, I've not tried that and don't know the price, didn't find getting
info on it easy last time I tried.
Have used in both legacy and uefi boot machines, don't know the chipsets
off the top of my head, will look on Monday.
Thanks Michael and Arman.
To make things clear, you guys are using Passthrough, right? It’s not vGPU. The
4x GPUs are added on the “Host Devices” tab of the VM.
What I’m trying to achieve is add the 4x V100 directly to one specific VM.
And finally can you guys confirm which BIOS type is being us
Also use multiple t4, also p4, titans, no issues but never used the nvlink
On Fri, 4 Sep 2020, 16:02 Arman Khalatyan, wrote:
> hi,
> with the 2xT4 we haven't seen any trouble. we have no nvlink there.
>
> did u try to disable the nvlink?
>
>
>
> Vinícius Ferrão via Users schrieb am Fr., 4. Sept
hi,
with the 2xT4 we haven't seen any trouble. we have no nvlink there.
did u try to disable the nvlink?
Vinícius Ferrão via Users schrieb am Fr., 4. Sept. 2020,
08:39:
> Hello, here we go again.
>
> I’m trying to passthrough 4x NVIDIA Tesla V100 GPUs (with NVLink) to a
> single VM; but thing
9 matches
Mail list logo