Hey, maybe I'm a little late to the party but I would like to request that you benchmark using a real UI, rather than that txt2waifu generator Automatic1111.
As a professional organization you should have switched off of that godforsaken platform more than 2 months ago.
Also, please don't lie about their being no tensor rt support for SDXL models. Automatic1111 is NOT stable diffusion. Their individual lack of support and awfully slow pace at development is not representative of the Stable Diffusion scene as a whole.
The sooner people realize that it being called "stable diffusion webui" is something that the developers have chosen themselves rather than an official adoption by any of the creators of Stable Diffusion, the better.
Because the officially adopted webui is ComfyUI and frankly I'm already getting 1 image per second results without using tensor RT or anything specialized besides XFormers.
So, your benchmarking results are about as useful to real users as automatic1111 itself. Not very.
Okay, maybe take a chill pill and step it down a notch on the rhetoric.
First, Automatic1111 does have a "real" UI, which is ComfyUI as you point out. That is not really an important part of the testing. Prior testing used different front-ends, but the results weren't radically changed. Second, the last testing was done in early November, and
at the time, SDXL was not working with TensorRT — at least not without some extra effort and code that I didn't have access to. Is it working now? I don't know, as I haven't checked, but probably — Nvidia (and AMD and Intel) are routinely submitting source code updates.
It's clear you have an axe to grind with Automatic1111, and frankly I couldn't care less about that aspect. Automatic1111 has been forked by a lot of places, and it's generally considered the default choice for many people. More importantly, Nvidia, AMD, and even Intel have all offered up specific instructions and tuning to get it working on every modern GPU. That's a critical element of my testing: Being able to run the tests on non-Nvidia GPUs. I am not doing
any coding whatsoever for these tests, at least in regards to Stable Diffusion, but I've regularly touched base with the GPU companies regarding their "optimal test paths" and that's what was used here.
Is A1111 perfect? I'm sure it's not. But the "txt2waifu" insult you toss in there is meaningless invective. The models and prompts used are what ultimately determines the output, not the front-end and web interface. In this case, using SD1.5 means the results are dependent on the training of the model from HuggingFace, not A1111.
The rest of what you say is basically shouting into a vacuum. You got one image per second? Okay... on
what exactly? Because you didn't provide any details whatsoever about your test settings, hardware used, or even the software you used, and that makes your statements completely useless.
Which is ironic, because you assert that my benchmarking results are useless, even though I've provided detailed instructions on what was tested and how it was tested. So, if you really want to have a dialog with me about testing, drop the attitude and provide real suggestions and advice, not ad hominem attacks — against either me or the creator of Automatic1111.
This is the "nice response" from me. Again, tone it down, and if you have recommendations, give concrete examples and links rather than invective and thinly veiled insults.