As far as efficiency, AMD leads the way. nVidia 5090 600-700Watt load consumption is double that of early AMD 9070XT power consumption leaked reports with the 9070XT showing performance both in raster and RT around the same as nVidia 5080.
In order to compare efficiency between hardware architectures, you need to look at not just one datapoint, but the whole perf/W curve. For Blackwell, the Ada (RTX 4000) data should mostly apply, since the process node and microarchitecture between the two are basically the same. Fortunately, Jarred did some good testing on that and found the RTX 4090 to be supremely efficient, especially if you just dial it back a little bit. However, note that at 100% power, it's not even using 400 W!
@JarredWaltonGPU , did you do similar testing for the RX 7900 XTX, by any chance?
You think AMD couldn't just double the power requirements and come up with a GPU that performs the same as a 5090?
If you're talking about running the same RX 9070XT die at double the power, then no. Power scaling is very nonlinear. If you're starting at the max non-OC power limit and you double that, it might be good for only like 20% more performance.
nVidia's stock price is free falling downwards, this is clearly a BIG mistake for nVidia on several fronts.
Like Icarus, fly too close to the sun and you're gonna get burnt. That said, Nvidia's stock won't crash, it'll just return back to some semblance of sanity. It's nothing to do with RTX 5000 and everything to do with their datacenter AI products, BTW. That's where they get most of their revenue, these days, and where investors were anticipating the stratospheric growth to continue.
Oh really? I found this datasheet which mentions only a 16 Gb chip.
Their product page also lists only 16Gb:
GDDR7 comes with a 70% increase in thermals ... these air cooled cards just aren't going to cut it and also why those few that actually got a 5090 are reporting system lockups after 1-3 hours of continuous use. "right?" yeah ok whatever.
First, the "thermals" must be a function of power, which is a function of data rate. So, specifically which data rates are you comparing? Also, the RTX 4090 used GDDR6X - are you comparing against that or regular GDDR6?
Second, the above datasheet claims it's 44.4% more efficient (i.e. in terms of pJ/b). In the comparison, they seem to imply they're comparing 18 Gbps GDDR6 vs. 32 Gbps GDDR7. When I apply that to the data rates given,
the GDDR7 should be using on 23.1% more power, not 70%!
Speaking of data rates, the RTX 5090 runs its memory at a modest 28 Gbps, whereas the RTX 5080 uses 30 Gbps. FWIW, the RXT 4090 ran its GDDR6X at 21 Gbps, but because it uses PAM4 (unlike regular GDDR6), the efficiency spec for GDDR6 isn't directly applicable. The info I could find for Micron's GDDR6X doesn't give such an efficiency stat.
So, I'd love to know more about this 70% stat, if you'd kindly enlighten us.
Workstation cards are an entirely different market and you're paying for specific application support/drivers, not so much the hardware and memory. Not to mention the memory bandwidth is actually lower on the W7900 as it uses ECC GDDR6 where as the 7900XTX does not use ECC.
Workstation cards typically have lower power limits and correspondingly reduced clocks. Performance is therefore a little lower, but not much. In the case of memory bandwidth, we're talking 864 GB/s for the Radeon Pro W7900 vs. 960 for the RX 7900 XTX.
As for memory capacity, that's absolutely a differentiator and probably one of the reasons for the lower memory clock. I'm pretty sure they do in-band ECC, and that's probably not being counted in the above data, because that discrepancy follows directly from the corresponding clocks of 18 Gbps vs. 20 Gbps for the gaming card.
Anyway, the whole workstation thing is a mere footnote. Like I said, both AMD and Nvidia do it, so I think it's not worth dwelling on.