Samsung unveils 32Gb GDDR6W chips with 22 GT/s data transfer rate.
Samsung's GDDR6W Doubles Performance and Capacity : Read more
Samsung's GDDR6W Doubles Performance and Capacity : Read more
The scheduled forum maintenance has now been completed. If you spot any issues, please report them here in this thread. Thank you!
If HBM was mass-manufactured on a scale comparable to GDDR6*, it would likely be cheaper than this fancy double-decked chip-and-PCB sandwich.Reads cool, but this sounds power hungry and hot right off the bat. I mean, probably worth the trade off vs HBM's implementation, but I would not say HBM is not the better option for true high performance computing as things stand now. Trade offs, for sure.
Likely the production cost of HBM could be cheaper, but I doubt the implementation cost would be. HBM requires a lot of traces.If HBM was mass-manufactured on a scale comparable to GDDR6*, it would likely be cheaper than this fancy double-decked chip-and-PCB sandwich.
As others have pointed out, it's not so much the cost of the HBM chip itself, but the whole "close to the CPU/SoC/GPU" restriction and interposer. I do believe you can do it outside of the interposer and use it as GDDR/DDR, but then the traces/routes on the PCB would still make the cost skyrocket compared to GDDR/DDR. The only comparable thing to HBM would be traditional eRAM in both performance and cost (as I understand it).If HBM was mass-manufactured on a scale comparable to GDDR6*, it would likely be cheaper than this fancy double-decked chip-and-PCB sandwich.
If HBM was mass-manufactured on a scale comparable to GDDR6*, it would likely be cheaper than this fancy double-decked chip-and-PCB sandwich.
The proximity isn't a limitation or restriction until you need more than 128GB and consumer GPUs aren't going to need that much VRAM within the next 20 years. We'll likely have 3D-stacked VRAM within the next 10 years.As others have pointed out, it's not so much the cost of the HBM chip itself, but the whole "close to the CPU/SoC/GPU" restriction and interposer.
I'm with you there. I rather GPUs stack HBM and charge me accordingly, but they won't, given the trade offs in cost.The proximity isn't a limitation or restriction until you need more than 128GB and consumer GPUs aren't going to need that much VRAM within the next 20 years. We'll likely have 3D-stacked VRAM within the next 10 years.
There won't be a "tradeoff in cost" when HBM or some future replacement becomes mainstream, it'll be cheaper than GDDR-whateveriscurrent.I'm with you there. I rather GPUs stack HBM and charge me accordingly, but they won't, given the trade offs in cost.
Hm... That's an interesting prediction, but I don't see it like you do. What makes you say that is a potential future?There won't be a "tradeoff in cost" when HBM or some future replacement becomes mainstream, it'll be cheaper than GDDR-whateveriscurrent.
All memory, regardless of whether it is as stand-alone packages or integrated on SoC interposer will be TSV stacked die eventually, at which point there will be effectively no manufacturing cost difference between HBM and everything else, apart from HBM or equivalent on-package memories foregoing individual packaging, saving costs.Hm... That's an interesting prediction, but I don't see it like you do. What makes you say that is a potential future?
So, in other words to simplify your argument, you see HBM and other "super fast" memory go the way of the Cache?All memory, regardless of whether it is as stand-alone packages or integrated on SoC interposer will be TSV stacked die eventually, at which point there will be effectively no manufacturing cost difference between HBM and everything else, apart from HBM or equivalent on-package memories foregoing individual packaging, saving costs.
AMD already uses chiplets for memory controllers and caches, they could easily integrate the HBM base die in those right now if they wanted to.So, in other words to simplify your argument, you see HBM and other "super fast" memory go the way of the Cache?
I can see that, kind of.
But tipping your argument over (or taking it from another angle), just making CPUs and GPUs have bigger caches via chiplets or just bigger dies, wouldn't that just make having a slower secondary memory subsystem a non-issue? Or less of an issue at least.AMD already uses chiplets for memory controllers and caches, they could easily integrate the HBM base die in those right now if they wanted to.
In he case of GPUs, stacking memory directly on top of the memory controller / cache dies would enable AMD to cram all the bandwidth they need in less than half as many MC&CDs. Eliminating the HBM base die would also allow higher speeds at the same or lower power.But tipping your argument over (or taking it from another angle), just making CPUs and GPUs have bigger caches via chiplets or just bigger dies, wouldn't that just make having a slower secondary memory subsystem a non-issue? Or less of an issue at least.