This is the problem with university researchers and estimates. Is OpenAI actually using 45 GWh of electricity, purely for GPT-5 queries? It should be possible to determine appoximately how many data centers and such OpenAI uses, but without insider knowledge, it's impossible to say whether those GPUs and servers are being for:
- Running GPT-5
- Running GPT-4 and older models
- Running non-GPT models
- Training new models
- Running other infrastructure
- Basically anything else
45 GWh would mean that OpenAI is consistently using the equivalent of 1.875 GW of power, all day long, every day. That's easily the equivalent of a couple dozen large 75~100 megawatt data centers doing nothing other than running OpenAI. That figure might be viable, but even if OpenAI is using that much power, it's a safe bet that at present a large chunk of that power isn't currently being used to serve up GPT-5 responses.
Realistically? I'd guess no more than 10~20 percent is for GPT-5 inference. You could also argue that GPT-5 used probably thousands (tens of thousands) of GPUs for a couple of months to train the model. Again, that's just estimating, but obviously a ton of electricity gets consumed in the process. However, if that's factored into estimates at all, it would also mean the cost per query goes down over time, as the training power gets diffused across billions (trillions?) of queries.
Anyway, a quick estimate here. GPT-5 can respond to a typical request in about 17 seconds. That's based on me just running a query right now asking it to write me a short, funny story about training GPT-5. So, about 1000 tokens took 17 seconds. To get up to 18 Wh per query with that sort of napkin math, it would mean that the hardware used to respond to my query was consuming 3800 watts of power. Getting up to 3800 watts isn't hard. But Nvidia's GPUs are designed to run multiple concurrent workloads at the same time. If a single H100-based server (eight H100 GPUs) is running the query, it's probably also running a dozen other concurrent queries is my guess.
But I'll admit, I could be wrong. I'm not researching OpenAI power use or anything else. I just suspect that a lot of these estimates are more of a "worst-case" estimate than a real-world estimate. Probably both GPT-4 and GPT-5 average power use is a lot lower than these estimates on average, but proportionately GPT-5 likely does use 8X more power on average.