alceryes
Splendid
1080p will show the MOST CPU difference in FPS. Higher resolutions are usually more GPU bound.IF you read the article 15% was in gaming at 1080p
Last edited:
1080p will show the MOST CPU difference in FPS. Higher resolutions are usually more GPU bound.IF you read the article 15% was in gaming at 1080p
It's also a very early engineering sample, I wouldn't read too much into that.What's very curious and IMPORTANT to note here is they locked the processor at 4GHz. We all know the 5900X can run at considerably higher speeds here.
10:1 They are thermal throttling. Note the lack of thermals.
Call me crazy but eventually sockets will have cooling from below and above working in a compression fashion where each heat sink (front and back) pull against each other.
I think this wound only happen with very specialized PCBs (and CPUs) where the CPU is soldered to the board. Making a CPU removable necessitates (at least currently) having the CPU too far off the PCB to make cooling it from the underside practical.Call me crazy but eventually sockets will have cooling from below and above working in a compression fashion where each heat sink (front and back) pull against each other.
With one extra layer of silicon and interconnects between the CCD and IHS, thermals are almost certain to be a little more challenging.10:1 They are thermal throttling. Note the lack of thermals.
Call me crazy but eventually sockets will have cooling from below and above working in a compression fashion where each heat sink (front and back) pull against each other.
With one extra layer of silicon and interconnects between the CCD and IHS, thermals are almost certain to be a little more challenging.
As for cooling the bottom of the socket, I wouldn't expect too much out of that since heat has to go through the bed-of-pins down to the motherboard, through layers and whatever may be on the back. The thermal resistance from the die, through the CPU substrate and everything else to the back of the motherboard will be horrible. The most heatsinking I could imagine making sense there would be upgrading the mounting backplate to a small heatsink mainly to help cool the Vcore power and ground planes so they don't contribute to CPU temperature and maybe the socket just a little bit.
The SRAM itself uses almost no power. It is the tagRAM (the bit responsible for keeping track of what cache line is caching which memory chunk) that uses tons of power doing the lookups and if you are going to make the L3$ 6X as large, you can mitigate tagRAM power and latency by making each cache line 2-8X as big. With such a large cache, you should also be able to afford reducing the way-ness and associativeness of each L3$ block to simplify the tagRAM without hurting the hit rate much.SRAM cache tends to be expensive, power wise, when it's fired.
I thought that they implied the full 64MB cache stack on one CCD was about less than 1/2 the size of a 7nm core chiplet so if they were aiming for the full 128MB , there would be plenty of room (since it would be a 64MB stack on each chiplet). Lisa herself said the 64MB SRAM used in the demo was 36mm^2 which is just under 1/2 the die area of a full Zen 3 chiplet.Since AMD was speaking about gaming performance, this64MB128MB cache should be a mainstream thing at least on select gaming SKUs. Lower-end SKUs will likely use cache die defects for32-56MB64-112MB extra L3$.
(Edit: with the structural silicon, there may be smaller cache die or even models with none at all. On 7nm, 128MB would be pretty close to a whole CCD-sized die, not just a center sliver shown in illustrations.)
Update 6/1/2021 10am PT: AMD has confirmed to Tom's Hardware that Zen 3 Ryzen processors with 3D V-Cache will enter production later this year. The technology currently consists of a single layer of stacked L3 cache, but the underlying tech supports stacking multiple dies. The technology also doesn't require any specific software optimizations and should be transparent in terms of latency and thermals (no significant overhead in either). We also obtained further fine-grained details, stay tuned for additional coverage.
While I think AMD is doing very well, I am starting to wonder if this every increasing cache size is sustainable. AMD's solution on CPU and GPU seems to be quite similar, slap an oversized cache to improve performance. Cache takes up a lot of die space and even if you can stack it, there is still a limit to the height and width of the chip.
The die space here is non-existent. Since it is a different die it takes up no space of the processor.While I think AMD is doing very well, I am starting to wonder if this every increasing cache size is sustainable. AMD's solution on CPU and GPU seems to be quite similar, slap an oversized cache to improve performance. Cache takes up a lot of die space and even if you can stack it, there is still a limit to the height and width of the chip.
Intel's heterogeneous cores arrangement is because it cannot do 16 high-performance cores on 10nm within a reasonable TDP. AMD is planning to do heterogeneous cores too with Zen 4D as the power-efficient cores to go along Zen 5.To some extent Alder Lake with its big.Little enhancements strikes me as a play more for the mobile market than ultra performance users.
Everybody in discussion seems to be focused on Intel and Alder Lake, but is that ignoring theelephantApple in the living room? To some extent Alder Lake with its big.Little enhancements strikes me as a play more for the mobile market than ultra performance users.
How much of this new AMD/chiplets is a response to M1 and staying ahead of what Apple is up to?
Intel's heterogeneous cores arrangement is because it cannot do 16 high-performance cores on 10nm within a reasonable TDP. AMD is planning to do heterogeneous cores too with Zen 4D as the power-efficient cores to go along Zen 5.
Everybody in discussion seems to be focused on Intel and Alder Lake, but is that ignoring theelephantApple in the living room? To some extent Alder Lake with its big.Little enhancements strikes me as a play more for the mobile market than ultra performance users.
How much of this new AMD/chiplets is a response to M1 and staying ahead of what Apple is up to?
The instruction set is likely the same, just optimized for lower power and smaller die size: you can do AVX512 on a quarter-width ALU, you just need to break it down into four extra steps.I would have to see the Inst Set support for the low powered cored to figure out how it would affect things like gaming and encoding/decoding type task. I'm guessing AVX and other MIMD support on the low powered cores is non existent. A low powered core might be useful for something like handling a network stack, or system IO interrupts. But again, this is only relevant for low power users like laptops.
I was going to argue with you, but Intel's been trying to innovate for 3 years. If they'd have been trying to innovate for 6 years (before Ryzen) though, they probably wouldn't be in half the mess they're in right now.Intel has been innovating. The problem is they don't have anything to show for it because of the woes trying to get off 14nm.
I love the concept, but I have to say I'm wary of thermals in the way they described how they'll be doing the "ground leveling" of the surface.
Throwing more memory if the price increase is low... Sure, why not? Will it be cheap though? Hm... Doubt it, so this may only be a feature for TR-class or Ry-69x0-class. I'd love it if they released a Zen3 refresh with this just bolted on as a "preview" for enthusiasts. I'm sure that nieche market would pay whatever premium they ask for.
I seriously doubt this will make its way to lower SKUs. With the "G" and mobile APUs they made clear they'll be uplifting their monolithic dies and then the chiplets will cover the higher end of the performance spectrum, but will keep them separated. Or that's what I think.
Cheers!
While I think AMD is doing very well, I am starting to wonder if this every increasing cache size is sustainable. AMD's solution on CPU and GPU seems to be quite similar, slap an oversized cache to improve performance. Cache takes up a lot of die space and even if you can stack it, there is still a limit to the height and width of the chip.
Cache has diminishing returns after a certain point and also with larger cache sizes, latency tends to increase. We'll see how it ends up performing.
Cache has diminishing returns after a certain point and also with larger cache sizes, latency tends to increase. We'll see how it ends up performing.
I agree. But it's curious why there would be a focus low power cores unless it was purely for energy savings.
I would have to see the Inst Set support for the low powered cored to figure out how it would affect things like gaming and encoding/decoding type task. I'm guessing AVX and other MIMD support on the low powered cores is non existent. A low powered core might be useful for something like handling a network stack, or system IO interrupts. But again, this is only relevant for low power users like laptops.