Moore's law says 2x transistor density in 18months. I know that is no longer feasible. But transistor density is largely what determines performance and cost per transistor. With each generation the cost of each transistor should go down.
That said, yes wafer cost are going up. However this merely slows down the price decrease per transistor per generation.
What's funny about that is how Eben Upton said in Toms' launch coverage of the Pi 4 that they moved from 40 nm to 28 nm, because
that node had the lowest cost per transistor. And, at the time (though this was back in 2019 and a lot has changed), he expected the next Pi would also be made on 28 nm.
Now, perhaps he was accounting for not only the per-wafer costs, but also the various up-front and engineering costs, amortized over the lifetime (or at least the first year's worth) of production they expected to do. Otherwise, something seem very broken.
Looking at transistor count and cost per transistor based on wafer node, the margin Nvidia and amd are pulling are insane compared to what it used to be.
What figures are you using, for that? I think it's probably more telling to look at the gross margins in their quarterly financial reporting, since that captures all of the engineering, support, and marketing costs. Those bills all have to get paid somehow, and these companies' income is nearly 100% goods-based, not services (though Intel looks to be challenging that notion).
Basically the fps/dollar has stagnated.
That's not the fault of the hardware. If you look at GFLOPS/$, there's a strong generational rate of improvement. If fps/$ isn't going up, that says more about the games and what resolutions people use.
Year-Mo | Graphics Card | GFLOPS | MSRP | GFLOPS/$ |
---|
2013-11 | GTX 780 Ti | 5046 | $699 | 7.22 |
2015-06 | GTX 980 Ti | 6054 | $649 | 9.33 |
2017-03 | GXT 1080 Ti | 11340 | $699 | 16.22 |
2018-09 | RTX 2080 Ti | 13448 | $999 | 13.46 |
2020-09 | RTX 3090 | 35580 | $1,499 | 23.74 |
2022-10 | RTX 4090 | 82600 | $1,599 | 51.66 |
Okay, so there's a regression in the RTX 2080 Ti, but that's because they only went from 16 nm to 12 nm, and burned most of their additional silicon budget on ray tracing and tensor cores, while simultaneously increasing the price. A lot of people were snapping up GTX 1080 Ti's, at the time, in an open rejection of what a poor value it represented.
BTW, the small jump between the 780 Ti and 980 Ti is largely due to the fact that they were both made on 28 nm. Maybe, if Nvidia could've charged more for it, the 980 Ti would've been made on 20 nm (but, it's also possible there simply was no GPU-suited 20 nm node, at the time).
it was obvious before the release of the cards crypto upheld these high prices. How AMD and NVIDIA didn't see the collapse in demand at these prices is beyond me.
Ever heard of the game musical chairs? It's a little like that. You know the music is going to stop, but you don't know when. And, until it does, you have to keep pace with the others.
Six years back $750 would have gotten you a #1 card. $550 #2
Yeah, it's not fun to see prices increase faster than inflation.
It happened with CPUs, as well, once the core-count race got going. I still lament that you have to buy a lot more cores than you may want or need, just to get a desktop CPU with the highest turbo limits.