I am planning to build a multipurpose home server. It will be a NAS, virtualization host, and have the typical selfhosted services. I want all of these services to have high uptime and be protected from power surges/balckouts, so I will put my server on a UPS.
I also want to run an LLM server on this machine, so I plan to add one or more GPUs and pass them through to a VM. I do not care about high uptime on the LLM server. However, this of course means that I will need a more powerful UPS, which I do not have the space for.
My plan is to get a second power supply to power only the GPUs. I do not want to put this PSU on the UPS. I will turn on the second PSU via an Add2PSU.
In the event of a blackout, this means that the base system will get full power and the GPUs will get power via the PCIe slot, but they will lose the power from the dedicated power plug.
Obviously this will slow down or kill the LLM server, but will this have an effect on the rest of the system?
Nope. I actually did that unintentionally on a PC I built. I only used one power wire when the GPU needed 2 so it couldn’t use all the power it needed when running 100%. My understanding was PCI doesn’t support disconnecting devices so the system expects all components it starts up with to be available all the time. Lose one and the system goes down.
PCIe absolutely does support disconnecting devices. It is a hot swap bus, that’s how ExpressCard works. But it doesn’t mean that the board/uefi implements it correctly.
in other words: OP either needs to get a thunderbolt dock or straight up have 2 computers. The latter should not even consume that much more power if the PC gets shut down in the evening and woken up using wakeonlan in the morning.