Running into a strange couple of issues.
First, running a system with the intent of using VT-d. Have 2 GPUS, integrated Intel and a Nvidia card. The intention is to use the integrated card for day to day usage. The nvidia for a kvm hosted win10 box solely for a couple of windows games that currently will not run under wine.
So here is the first issue I have run into - when clear boots, it boots under the integrated graphics, switches to the nVidea for login, then switches back to the integrated for the desktop. Works just fine, just is weird. When I tried to disable the device id:
echo "8086 10b9" > /sys/bus/pci/drivers/pci-stub/new_id echo "0000:01:00.0" > /sys/bus/pci/devices/0000:01:00.0/driver/unbind echo "0000:01:00.0" > /sys/bus/pci/drivers/pci-stub/bind
the graphic system just locked up on the second line (i used the device id of the nvidea card, not what I just pasted above).
so i figured i could live with the goofy login.
On to the next issue - so when configuring a kvm hosted machine, everything works great. can host machines with no issues until I configure a pass through PCI device for vt-d, pick the nvidea devices and just like before, the graphic subsystem on the host freezes.
in both freeze cases, it is just the graphic system that freezes, I can ssh into the box and reboot it.
So looking for suggestions.
(the bios is configured for both GPUs to be active, full vt-d support and so forth. was running windows 10 and nvidia before switching the system to clear)
Thanks in advance.