Micron expects to produce GDDR7 memory devices in the first half of 2024.
Micron to Introduce GDDR7 Memory in 1H 2024 : Read more
Micron to Introduce GDDR7 Memory in 1H 2024 : Read more
The scheduled forum maintenance has now been completed. If you spot any issues, please report them here in this thread. Thank you!
Can you also explain the PAM3 vs PAM4? Is it for power management? Why would ddr6x use pam4 and ddr7 use pam3?Usage of this next-gen mem type in 'consumer' products seems like a pipe dream for now.
Currently NVIDIA provides the fastest memory solution with it's RTX 40 series GPUs in the form of the GDDR6X which provides up to 22 Gbps pin speeds, and AMD's Radeon RX 7000 series cards on the other hand utilize the standard 20 Gbps GDDR6 tech.
So let's assume GDDR7 would deliver the following bandwidth figures, based on 36 Gbps pin speeds:
- 128-bit @ 36 Gbps: 576 GB/s
- 192-bit @ 36 Gbps: 846 GB/s
- 256-bit @ 36 Gbps: 1152 GB/s
- 320-bit @ 36 Gbps: 1440 GB/s
- 384-bit @ 36 Gbps: 1728 GB/s
The rollback from PAM4 to PAM3 was almost certainly because they analyzed results from GDDR6X and concluded that timing jitter and phase noises from pushing higher clocks with PAM3 would be easier to deal with than keeping signals clean enough for PAM4.Can you also explain the PAM3 vs PAM4? Is it for power management? Why would ddr6x use pam4 and ddr7 use pam3?
RTX 5000 set for spring 2025 release, so the timing is good for GDDR7 on the cards. As for individual cards:If Jeedeedeeareseven comes out in the first half of 2024, when does Jeedeedeeareseveneques come out? Any chance we will see that on the 5090Ti?
Why? We've seen how Nvidia is keen on cutting the number of chips and memory channels, on lower-end models. Faster RAM would let them do even more of that.As for individual cards:
...
RTX 5080 chance of sporting GDDR7: 50%
RTX 5070 Ti chance of sporting GDDR7: 50%
RTX 5070 chance of sporting GDDR7: 10%
RTX 5060 Ti chance of sporting GDDR7: 0%
RTX 5060 chance of sporting GDDR7: 0%
GDDR6 @ 20gbps is ~4.8W and GDDR6X @ 20gbps is ~4.65W (21gbps is ~4.87W)Normal memory is 2-3W per DIMM, GDDR6X is ~4W per chip, wonder how much power GDDR7 is going to guzzle at 32+Gbps.
At some point, HBM is going to become necessary just to keep the IO power in check.
Reviews bashing Nvidia and AMD to a lesser degree for insufficient VRAM and VRAM bandwidth might motivate them to rethink their bus shrinkage strategy. Current-gen stuff often runs into scaling issues previous-gen stuff doesn't have when dipping below 256bits/16GB. If next-gen bumps performance up ~50% for a given marketing tier like the 5090 rumors say it might, trimming bus some more with GDDR7 won't be an option. At best, it might make it possible to keep things the same as they are now.Why? We've seen how Nvidia is keen on cutting the number of chips and memory channels, on lower-end models. Faster RAM would let them do even more of that.
We already have 24Gbits DDR5, 24Gbits GDDR7 should be quite easily feasible. As mentioned in my bit reply above though, I don't think further memory interface trim will be possible as they'll need the extra bandwidth to feed the higher performance cores required to avoid having another "lowest sales in 20+ years" generation on their hands.I haven't seen in any documents or press releases yet. If they're able to launch 32Gb capacity that would mean client side they could drop bus width down fairly significantly without losing memory bandwidth over what is available today so long as they use higher end chips.
I was being a bit silly to suggest they would trim interfaces even further, but if they wish to scale performance while at least keeping interface widths the same, then they'll definitely need higher-performance memory (as you say).As mentioned in my bit reply above though, I don't think further memory interface trim will be possible as they'll need the extra bandwidth to feed the higher performance cores
Depends on how much of a premium GDDR7 is going to carry at least initially. Just like DDR5 and GDDR6X, the premium might be too steep for entry-level during the first year and change.So, it sounds like we agree that the lower-tier models could still be viable targets for GDDR7?
I sure wouldn't be that confident seeing as we still don't have over 16Gb GDDR6 despite enterprise needing massive VRAM capacity and the spec allowing for 32Gb. I certainly hope the GDDR7 spec has been designed around bigger capacity, but I wouldn't view it as a forgone conclusion.We already have 24Gbits DDR5, 24Gbits GDDR7 should be quite easily feasible.
I think it's going to depend on the silicon cost between cache and memory channels. If there's 32Gb capacity they could drop to a 96 bit bus and still have more memory bandwidth than now on 128 bit with 12GB capacity and the lowest end card could drop to 64 bit while being very close to the same bandwidth as 128 bit now and maintain 8GB.As mentioned in my bit reply above though, I don't think further memory interface trim will be possible as they'll need the extra bandwidth to feed the higher performance cores required to avoid having another "lowest sales in 20+ years" generation on their hands.
While the spec may technically allow larger sizes, the command word structure and the way the DRAM matrix operates may not allow exercising that extra hypothetical flexibility due to issues with making memory array rows wider or sense lines longer. DRAM manufacturers would likely prefer to have an extra bank address bit or two.I sure wouldn't be that confident seeing as we still don't have over 16Gb GDDR6 despite enterprise needing massive VRAM capacity and the spec allowing for 32Gb.
So where is the RTX 4060 with 21, 22.4 or 24 Gbps VRAM?Why? We've seen how Nvidia is keen on cutting the number of chips and memory channels, on lower-end models. Faster RAM would let them do even more of that.
Maybe they spent all the remaining budget on extra L2 cache?So where is the RTX 4060 with 21, 22.4 or 24 Gbps VRAM?
I think you're missing the point entirely. If GDDR7 has 32Gb chips that means nvidia could take a hypothetical 4060, cut it to a 64 bit bus thus using less silicon for the chip and only need to wire up 2 memory chips while maintaining the same capacity and bandwidth. This could be cheaper than using slower memory on a 128 bit bus and if it is you can be sure it's something nvidia would look at doing.So where is the RTX 4060 with 21, 22.4 or 24 Gbps VRAM?
I doubt GDDR7 will be anywhere as affordable as GDDR6 for the first year or two from launch. Much like DDR5, the 2nd or 3rd wave of GDDR7 chips will likely have much better timings and clock frequencies than the first-gen stuff too for a lower price. Also, if we get 32Gbits chips, they'll likely be at a steep premium for the first couple of years on top of that. I doubt 2x32Gbits GDDR7 chips will be viable at the entry-level for the first 2-3 years from such chips becoming available.This could be cheaper than using slower memory on a 128 bit bus and if it is you can be sure it's something nvidia would look at doing.
Oh the memory itself absolutely won't be cheaper at all. It would entirely come down to the silicon savings of getting more chips per wafer and lowering board complexity and whether those savings at scale make sense to pair with the more expensive memory. It is also dependent on 32Gb being widely available which I doubt it would be even if it launched with GDDR7 due to enterprise demands. Of course if the 2025 rumor for nvidia's next consumer architecture is accurate that could be far enough out since lower end tend to be the last in line.I doubt GDDR7 will be anywhere as affordable as GDDR6 for the first year or two from launch. Much like DDR5, the 2nd or 3rd wave of GDDR7 chips will likely have much better timings and clock frequencies than the first-gen stuff too for a lower price. Also, if we get 32Gbits chips, they'll likely be at a steep premium for the first couple of years on top of that. I doubt 2x32Gbits GDDR7 chips will be viable at the entry-level for the first 2-3 years from such chips becoming available.
RTX 5060 16GB GDDR6 24Gbps with 128 bit bus will have more memory bandwidth that RTX 5060 16GB GDDR7 36Gbps with 64 bit bus. But according to what you say there is a good chance for an RTX 5090 32GB GDDR7 36Gbps with 256 bit bus.I think you're missing the point entirely. If GDDR7 has 32Gb chips that means nvidia could take a hypothetical 4060, cut it to a 64 bit bus thus using less silicon for the chip and only need to wire up 2 memory chips while maintaining the same capacity and bandwidth. This could be cheaper than using slower memory on a 128 bit bus and if it is you can be sure it's something nvidia would look at doing.