Really glad that could help! Since I've got your attention, I couldn't get TensorRT to work on Windows. At least 50% chance I didn't install it properly, BUT at the same time your gui was showing my 1650 instead of the 3060. After looking for some setting about Cuda devices and finding none I gave up. Generation times and usage pointed clearly at a normal 3060 task, even if the gui had the temperature for the 1650.
But anyway! One thing I'd like to ask is (now that there's a viable way to use it on my Mini) an option to allow other computers to access it, and better yet the API like in Automatic1111. Like that I could do some kind of LLM on the 3060 (I like Pygmalion 6b) and stable diffusion on the Mac.
All that aside, thanks for making a viable alternative to Draw Things. As much as I like it and the interface, choice is always good... and yours has the potential to be usable in remote :D
3060 here, it might be the vram. SDXL eats a lot of it (and if you had say the vae in the wrong spot it would output very wrong images) so it might be that either 8gb aren't enough, or maybe they aren't enough with the resolution of your screen plus whatever you are running, like the browser.
Or, OR: the checkpoint is corrupted. I had that happen a couple of times in the past and the whole huge error with loading of another model was what happened.