People don't call it PAM2 - its just a binary (i.e. NRZ) signal.I'm not familiar enough of the roadblocks which might exist between PAM3/4, but Ada supported PAM2/4 at the same time.
It was an increase, but nowhere near enough to justify buying a new card. The difference between the 10 and the 20 was minimal. And nothing compared to the 20 to 30, or the 30 to 40. It's the same thing for the 50 series. Yea, minimal increase in performance, but you're paying the extra for the "new and shiny" thing, which happens to be mostly based around "fake frames".im getting tired of people blaming gpu manufacturers for not having "enough" vram when the problem is that game devs are shoving bloatware rt features into games. ALSO vram isnt the only factor that makes a gpu "good"
that isnt exactly true...View: https://www.youtube.com/watch?v=iGUejCAupSg
20 series was still a massive leap in game performance + they upped the vram on their budget cards on that series.
No ... the kernel/user mode isn't relevant to my context. You seem to be focused on mapping of physical memory address space, I'm not. I saying you can direct allocation to VRAM with DX12, the OS will allow it.No it's you misunderstanding the documentation and not understanding kernel mode vs user mode.
I think the term is "supervised learning". The problems we ran into was more cycles actually reduced accuracy regardless of how good the data/source was ... we needed to establish a matrix to identify when patterns and relationship (training) was producing negative results.determining the meta-parameters of an AI model
Sounds like you're talking about overfitting. That happens when the size of your training set is too small for the model.The problems we ran into was more cycles actually reduced accuracy regardless of how good the data/source was ... we needed to establish a matrix to identify when patterns and relationship (training) was producing negative results.
Thanks for the chart, but you seem to have misread it. Micron clearly shows GDDR7 at 16-24Gb (32 Gbps) landing in 2024-2025. The 24GB+ that comes in 2026 is 36 Gbps.As for Micron it will be a couple of years:
I don't get this point. First, Snowden was complaining about the 5080 having only 16 GB. I've only heard a few AI folks complaining about the 5090's 32 GB, not gamers. Second, to increase either of these would've required doubling the GDDR7 chips (or widening the RTX 5080's memory interface), or else waiting for the 24 Gb chips to launch.With a 512 bit bus I know who I'm blaming for not putting more than 32GB on the card; but now we can have 48 (in our dreams).
Ok first off GDRR6X works out to about $5 per GB, where GDDR7 is estimated to be $7-$8 per GB....it's negligible price increase compared to the overall cost of the card.You presume that's even an option, but Micron is the sole supplier of GDDR6X (which is proprietary - not a JEDEC standard, I might add) and only offers it in 16 Gb capacity.
With GDDR7 now shipping, it's very unlikely Micron will introduce any higher-capacity GDDR6X dies, at this point.
Performance usually follows the improvement of whatever is the bottleneck. The 30% improvement suggests that the CUDA cores were the main bottleneck in the RTX 4090. So, that's why gaming only improved by the amount of CUDA cores they added (with an additional penalty for the clock speed decrease).
I think the RTX 5090's memory bandwidth & capacity went up so much purely for AI.
The article said a packaged die probably costs about $350:
You forgot the GDDR7 memory. More importantly, you're ignoring the engineering costs, which were estimated for Blackwell at $9 Billion, according to one estimate I read. There's also software and support. Drivers, technologies like DLSS, and Nvidia's extensive range of libraries for game developers cost a lot of money to develop and maintain.
After accounting for all their costs, Nvidia needs to make some profit, as do their AIB board partners, distributors, and finally retailers. So, even if the hardware costs like $550 or $600 to produce, I don't consider a MSRP of $2k too surprising. Could they sell it for less? Sure. But, in the absence of real competition at that level, why would they?
If we consider the situation of a large retail chain buying directly from an AIB partner, and each makes a 10% profit, then that puts their price at $1,652.89. If $250 of that is for memory, PCB, cooler, VRM, assembly, etc. then that puts Nvidia's price for the ASIC at about $1400. If Nvidia is making a 75% profit margin on these, which is the company's overall net profit, then it puts their cost at $800, after all engineering, software, R&D, and support are added in and gives them $600 of profit per 3rd party board that's sold at MSRP (which I know is a bit of a fiction, right now). So, they're taking just 30% of that overall $2k board price as profit. If their take went to zero, then the final cost of the board would only drop to $1270.50. So, don't look at that $550 or $600 figure and think It means Nvidia is making like $1400 or $1450 on each AIB card sold! Obviously, they make more on their Founders Edition, but I'm just talking about their MSRP, here.
Morally, do I think they should increase the price of their flagship gaming cards so much faster than inflation? No, I'm not happy with this situation, either. Were it not for the AI boom, I think these GPUs would neither be so expensive nor quite so powerful.
I don't really have much sympathy for people who expect to upgrade their graphics card, every generation. That's neither normal nor cost-efficient.
If someone had a 5-7 year old graphics card, my advice would be either to go for the RTX 5070 Ti or wait for at least the Supers to launch in a year or so. But, it depends a lot on their budget and what their old card is. The 5070 Ti is just what's most interesting to me, personally.
Given your account was registered in 2006, I have to ask if you simply weren't paying attention during the launch of the GTX 500, the GTX 700, or the GTX 900 series. All of those generations used the same node as their predecessor. And, in the case of RTX 2000, they were virtually the same. The fact of Nvidia reusing the microarchitecture of a prior generation also isn't new.
This is not a "sky is falling" moment, like you're trying to paint it. There are better nodes that Nvidia could use, but they're expensive and in high-demand right now. Given their relatively large dies, client GPUs aren't typically on leading-edge nodes. So, this ended up being a very incremental generation. That's how it goes, especially in the wake of RTX 4000 - which was above average in overall generational performance gains, if a bit on the pricey side.
BTW, I actually have a theory that Nvidia renamed some of the lower-end RTX 4000 models shortly before launch, in light of the weak GPU market at the time. That would explain the lackluster generational gains in some of the tiers.
Intel got through their quad-core era, finally. I'm not sure if there's a good analogy, here.
There's no way this degree of scarcity is working to Nvidia's advantage. They have an incentive to slightly undersupply the market, but if you do the math, there's no way they come out ahead with this ridiculous extent of a shortage.thanks to the manufactured scarcity, prices have now been raised to be between 60-80% over MSRP.
There's your answer. After two generations of above-average gains, they run out of gas. I've said it before: the seeds of dissatisfaction in the RTX 5000's were planted when Nvidia went so big in the RTX 4000 generation. Not only did it make RTX 4000 expensive, which made everyone angry, but it left them less room for the RTX 5000 to improve on it, without advancing to a newer node.Lets take a look back through history at Nvidia's 80 class cards.
780 2013 $649 = 24% uplift over previous gen
980 2014 $549 = 38% uplift
1080 2016 $649 =51% uplift
2080 2018 $699 =39% uplift
3080 2020 $699 =63% uplift
4080 2022 $1199=49% uplift
5080 2025 $ 999 =15% uplift
If you're not going to compare the launch models from one gen to the next, then you should look at all of the supers, not just the one that supports your argument. However, I think the better thing is comparing one generation's launch models to another's.Except the 4080 isn't really the comparison point for the 5080 is it? The $999 4080 super is, and that performance uplift is 8.3%
Again, that's not a fair comparison. Yield issues meant they couldn't sell fully functioning dies to everyone, so they reserved them for workstation cards. You have to compare what they actually sell, not imaginary products that would necessarily drive different economics.Also remember that the 4090, was still a cut down die....the full AD102 die had 18432 cuda cores.....meaning the 5090 is only a 15% increase in cuda cores over the previous generation.
No, but it's usually not cost-effective to upgrade every generation. So, I expect most normal people to upgrade every 2-3 generations. So, more like 3-5 years.You think people should hold onto their gpu's for 5-7 years?
You can say that about several of their previous generations. The product is the thing people are buying. The price, performance, and power are what count - not what's under the hood.Well the 5000 series is basically a slightly tweaked 4000 series, so it's already 3 years old out the gate,
I think the main use cases for DLSS and FG are using a higher resolution monitor, which is also going to demand more memory. Certainly, when you upgrade your monitor, it's reasonable to upgrade graphics cards.8 gigs hasn't been enough VRAM for several years, and 12 gigs is getting to be the minimum even for 1080p. Even more so as DLSS and FG increase VRAM usage....and both those technologies have been viewed as default requirements for the past three years.
This isn't about what I think Nvidia should have done. I never defended the decision to go with 16 GB in the RTX 5080. I just pointed out that they sort of brought this issue on themselves, by launching it without 24 Gb GDDR7 for whatever reason. I'm just explaining their decision space. It seems to me that you're so upset by their decisions that you're not looking at it from their perspective and trying to think about the constraints they're dealing with.You can't look at any single issue in a vacuum....it's the combination of everything, and there is no defense.
We can call it what it really is here: nvidia maintaining high margins. They didn't run out of gas so much as there's no market reason for them to either sell GPUs for less or make them bigger at the same price. There's no technical reason for them not putting another 10-15% cores at each tier below the 5090.There's your answer. After two generations of above-average gains, they run out of gas. I've said it before: the seeds of dissatisfaction in the RTX 5000's were planted when Nvidia went so big in the RTX 4000 generation. Not only did it make RTX 4000 expensive, which made everyone angry, but it left them less room for the RTX 5000 to improve on it, without advancing to a newer node.
My comment was targeted just at the generational gains, not pricing. Nvidia really can't make the RTX 5090 much bigger. That pretty much trickles down the lineup.We can call it what it really is here: nvidia maintaining high margins. They didn't run out of gas so much as there's no market reason for them to either sell GPUs for less or make them bigger at the same price.
One thing I've noticed about Nvidia is that they're pretty keen on maintaining similar perf/$, within a generation. The 5080's MSRP is exactly half of the 5090's and so is the number of its CUDA cores. In fact the die size of the GB103 is actually a bit more than half as big as the GB102.There's no technical reason for them not putting another 10-15% cores at each tier below the 5090.
Yes and no, because the scaling never used to be set by the halo tier product but rather whatever was below it. This price scaling off the halo tier product started with the 40 series and they're just carrying it forward because it means higher margins.One thing I've noticed about Nvidia is that they're pretty keen on maintaining similar perf/$, within a generation.
In the RTX 4000 series, the RTX 4070 Ti and RTX 4090 both had a better perf/$ ratio than the RTX 4080 (i.e. according to list prices). To remain in line with the other two, the RTX 4080 should've cost $1050, not $1200.Yes and no, because the scaling never used to be set by the halo tier product but rather whatever was below it This price scaling off the halo tier product started with the 40 series and they're just carrying it forward because it means higher margins.
Curious what data you're looking at which doesn't display close to linear performance scaling between those when using the 4090 as baseline. I think it was the 4070 which was slightly better than the rest of the 40 series (when compared to itself not necessarily 30 series). If you're just looking at FPS/$ that should always improve as you go down the stack.In the RTX 4000 series, the RTX 4070 Ti and RTX 4090 both had a better perf/$ ratio than the RTX 4080 (i.e. according to list prices). To remain in line with the other two, the RTX 4080 should've cost $1050, not $1200.
FLOPS/$. Your data is heavily skewed by games with low occupancy. Too much of the 4090's capacity is going to waste. Maybe try with RT and maxed out settings, if you're going to use averages. Either that, or use raster games that achieve high occupancy on the 4090.Curious what data you're looking at which doesn't display close to linear performance scaling between those when using the 4090 as baseline.
Then you're using a conditionally accurate metric which doesn't apply to general usage. That's certainly a valid way to measure, but not what I'd use.FLOPS/$. Your data is heavily skewed by games with low occupancy. Too much of the 4090's capacity is going to waste. Maybe try with RT and maxed out settings, if you're going to use averages. Either that, or use raster games that achieve high occupancy on the 4090.
It works for AI, which is largely driving their pricing and I think usually achieves good occupancy.Then you're using a conditionally accurate metric which doesn't apply to general usage. That's certainly a valid way to measure, but not what I'd use.
There's no way this degree of scarcity is working to Nvidia's advantage. They have an incentive to slightly undersupply the market, but if you do the math, there's no way they come out ahead with this ridiculous extent of a shortage.
Not sure about that, go back to 2014 and the Titan Z at $3000 ... nVidia were testing the consumer waters. That's when they tasted the lure of huge profit ... 4-way SLI with two Titan Z ... remember SLI, buying 2 GPUs at $1200 each for $2400. But, SLI had too many issues and people stopped buying two or more GPUs so nVidia needed to come up with ways to recover the drop in revenue by increasing single GPU prices.One thing I've noticed about Nvidia is that they're pretty keen on maintaining similar perf/$,
too bad both prices are out of reach for alot of people, specially considering those are probably US prices...When a partner card cost $2900 and nVidia FE card is $2000
Oh, the Titan cards were on their own planet. They never offered a comparable value to the cards below.Not sure about that, go back to 2014 and the Titan Z at $3000 ... nVidia were testing the consumer waters.
Except I doubt the partner cards would be selling for that much, if not for the scarcity. Their cost structure is the same as Nvidia's. The only difference is that they need to tack their own profit margins on the card and it's not nearly that much.When a partner card cost $2900 and nVidia FE card is $2000 and they both perform within 3% of each other ... you see the problem?
But I wasn't using ASUS as my example? I didn't even mention a specific partner? They all seem to be doing the price increase over MSRP, not just ASUS. But that's sorta my point, those partner cards would NOT be selling for that much over nVidia MSRP if it weren't for nVidia keeping supply short on purpose. And yes, supply will be held short for many months ... current suggestions by retailers like BestBuy is that 5090 supply may never be sufficient for the entire 2025 and into 2026.So, I'd say you're already biasing the argument, by using ASUS as your example.
AIBs rarely had more than one 4090 SKU at MSRP throughout the lifetime of the card (if that one was even available) so it really doesn't have much to do with supply. It has a lot more to do with the margins nvidia is making and since they effectively control the video card market AIBs don't have a choice but to go with it.But I wasn't using ASUS as my example? I didn't even mention a specific partner? They all seem to be doing the price increase over MSRP, not just ASUS. But that's sorta my point, those partner cards would NOT be selling for that much over nVidia MSRP if it weren't for nVidia keeping supply short on purpose. And yes, supply will be held short for many months ... current suggestions by retailers like BestBuy is that 5090 supply may never be sufficient for the entire 2025 and into 2026.