Cascade Lake-X enters a segment now dominated by AMD.
Intel Core i9-10980XE Review: Intel Loses its Grip on HEDT : Read more
Intel Core i9-10980XE Review: Intel Loses its Grip on HEDT : Read more
Games benchmarks on a non gamer CPU. There is no sense. Please do compiling benchmarks and other stuff that make sense.
And please stop using Windows to do that.
Games benchmarks on a non gamer CPU. There is no sense. Please do compiling benchmarks and other stuff that make sense.
And please stop using Windows to do that.
Games benchmarks on a non gamer CPU. There is no sense. Please do compiling benchmarks and other stuff that make sense.
And please stop using Windows to do that.
<snip>
The AMD 3950X has 16 PCIe lanes, right? So for those of us who have multiple adapters such as RAID cards, USB or SATA port adapters, 10G NICs, etc, HEDT is the only way to go.
<snip>
The article says that "Intel still can't deal with heat".
<snip>
TDP is the wrong way to directly compare an Intel CPU with an AMD CPU. Neither vendor measures TDP in the same fashion so you should not compare them directly. On the most recent platforms, per watt consumed, you get more work done on the new AMD platform, plus most users don't have their chips running at max power 24/7, so why would you calculate your power usage against TDP even if it were comparable across brands?Disclaimer: I badly want to dump Intel and go AMD. But are the conditions right?
The AMD 3950X has 16 PCIe lanes, right? So for those of us who have multiple adapters such as RAID cards, USB or SATA port adapters, 10G NICs, etc, HEDT is the only way to go.
Someone once told me "No one in the world needs more than 16PCIe lanes, that's why mainstream CPUs have never gone over 16 lanes". If that were true the HEDT CPUs would not exist.
So we can say the 3950X destroys the Intel HEDT lineup, but only if you don't have anything other than ONE graphics card. As soon as you add other devices, you're blown.
The 3970X is $3199 where I am. That will drop by $100 by 2021.
The power consumption of 280w will cost me an extra $217 per year per PC. There are 3 HEDT PCs, so an extra $651 per year.
AMD: 1 PC @ 280w for 12 hours per day for 365 days at 43c per kilowatt hour = $527.74
Intel: 1 PC @ 165w for 12 hours per day for 365 days at 43c per kilowatt hour = $310.76
My 7900X is overclocked to 4.5GHZ all cores. Can I do that with any AMD HEDT CPU?
In summer the ambient temp here is 38 - 40 degrees Celsius. With a 280mm cooler and 11 case fans my system runs 10 degrees over ambient on idle, so 50c is not uncommon during the afternoons on idle. Put the system under load it easily sits at 80c and is very loud.
With a 280w CPU, how can I cool that? The article says that "Intel still can't deal with heat". Errr... Isn't 280w going to produce more heat than 165w. And isn't 165w much easier to cool? Am I missing something?
I'm going to have to replace motherboard and RAM too. That's another $2000 - $3000. With Intel my current memory will work and a new motherboard will set me back $900.
Like I said, I really want to go AMD, but I think the heat, energy and changeover costs are going to be prohibitive. PCIe4 is a big draw for AMD as it means I don't have to replace again when Intel finally gets with the program, but the other factors I fear are just too overwhelming to make AMD viable at this stage.
Darn it Intel is way cheaper when looked at from this perspective.
Fun times. The Tualatin was based on Coppermine and went to 1.4 GHz, outclassing Williamette at 1.8GHz by a wide margin. Northwood came out and beat it, but at the same time Intel was developing Pentium M based on...guess what? Tualatin.I'm surprised nobody caught this from the second paragraph of the article.
Intel's price cuts come as a byproduct of AMD's third-gen Ryzen and Threadripper processors, with the former bringing HEDT-class levels of performance to mainstream 400- and 500-series motherboards, while the latter lineup is so powerful that Intel, for once, doesn't even have a response.
For twice? This is a recall of the olden days of the first-gen slot-A Athlon processors. Now I'm not well-versed in TomsHardware articles circa 1999, but this was not hard to find at all:
Coppermine's architecture is still based on the architecture of Pentium Pro. This architecture won't be good enough to catch up with Athlon. It will be very hard for Intel to get Coppermine to clock frequencies of 700 and above and the P6-architecture may not benefit too much from even higher core clocks anymore. Athlon however is already faster than a Pentium III at the same clock speed, which will hardly change with Coppermine, and Athlon is designed to go way higher than 600 MHz. This design screams for higher clock speeds! AMD is probably for the first time in the very situation that Intel used to enjoy for such a long time. AMD might already be able to supply Athlons at even higher clock rates right now (650 MHz is currently the fastest Athlon), but there is no reason to do so.
https://www.tomshardware.com/reviews/athlon-processor,121-16.html
Intel didn't have a response back then either.
I agree with the first point here, which is why we point out that Intel has an advantage there for users that need the I/O.
On the second point, can you point me to where it says that in the article? I must've missed it. Taken in context, it says that Intel can't deal with the heat of adding more 14nm cores in the same physical package, which is accurate if it wants to maintain a decent clock rate.
How about idle and average power consumption, or completed work per watt, or even overall completed work in a given time-frame, which make a better case about AMD's current level of competitiveness.
I'm surprised nobody caught this from the second paragraph of the article.
Intel's price cuts come as a byproduct of AMD's third-gen Ryzen and Threadripper processors, with the former bringing HEDT-class levels of performance to mainstream 400- and 500-series motherboards, while the latter lineup is so powerful that Intel, for once, doesn't even have a response.
For twice? This is a recall of the olden days of the first-gen slot-A Athlon processors. Now I'm not well-versed in TomsHardware articles circa 1999, but this was not hard to find at all:
Coppermine's architecture is still based on the architecture of Pentium Pro. This architecture won't be good enough to catch up with Athlon. It will be very hard for Intel to get Coppermine to clock frequencies of 700 and above and the P6-architecture may not benefit too much from even higher core clocks anymore. Athlon however is already faster than a Pentium III at the same clock speed, which will hardly change with Coppermine, and Athlon is designed to go way higher than 600 MHz. This design screams for higher clock speeds! AMD is probably for the first time in the very situation that Intel used to enjoy for such a long time. AMD might already be able to supply Athlons at even higher clock rates right now (650 MHz is currently the fastest Athlon), but there is no reason to do so.
https://www.tomshardware.com/reviews/athlon-processor,121-16.html
Intel didn't have a response back then either.
No, it is 24 PCIe lanes from the CPU (four going to the chipset and four more usually going to an M2 NVMe slot) and depending on the motherboard you pick, you can have up to 20 more PCIe lanes from the x570 chipset for a total of up to 40x useable PCIe 4.0. Of course, most boards will configure some of the HSIO lanes for extra SATA, USB3 or M.2 ports instead and those won't be available for PCIe. Even if you are left with 10 spare PCIe lanes, that's enough for two PCIe4.0x4 and two 4.0x1 slots. May not sound like much but you need to keep in mind that 4.0x1 is 16Gbps, almost fast enough for dual-10G. Also, a growing number of higher-end boards have 10G on-board.The AMD 3950X has 16 PCIe lanes, right? So for those of us who have multiple adapters such as RAID cards, USB or SATA port adapters, 10G NICs, etc, HEDT is the only way to go.
I don't really get your reasoning. You're comparing a 32 core CPU to an 18 core. Of course it's going to cost more, and in return you get more performance. If you're actually running it full bore 12 hours a day, 365 days a year as you did in your power costs calculation, I can only assume you'd be using it for work in which case it'd likely end up paying for itself through increased productivity.All great points. I feel that it is going to come down the the tangibles for me. Initial outlay in $ and $ per year to run them.
Anyway, I think I've probably taken the conversation away from the original topic, so I'll leave it there.
Games benchmarks on a non gamer CPU. There is no sense. Please do compiling benchmarks and other stuff that make sense.
And please stop using Windows to do that.
Intel consciously neglected traditional CPUs back then to try and make itanium take off.Having an even stronger x86 competition than was already present would have been an even worse idea.I'm surprised nobody caught this from the second paragraph of the article.
Intel's price cuts come as a byproduct of AMD's third-gen Ryzen and Threadripper processors, with the former bringing HEDT-class levels of performance to mainstream 400- and 500-series motherboards, while the latter lineup is so powerful that Intel, for once, doesn't even have a response.
For twice? This is a recall of the olden days of the first-gen slot-A Athlon processors. Now I'm not well-versed in TomsHardware articles circa 1999, but this was not hard to find at all:
Coppermine's architecture is still based on the architecture of Pentium Pro. This architecture won't be good enough to catch up with Athlon. It will be very hard for Intel to get Coppermine to clock frequencies of 700 and above and the P6-architecture may not benefit too much from even higher core clocks anymore. Athlon however is already faster than a Pentium III at the same clock speed, which will hardly change with Coppermine, and Athlon is designed to go way higher than 600 MHz. This design screams for higher clock speeds! AMD is probably for the first time in the very situation that Intel used to enjoy for such a long time. AMD might already be able to supply Athlons at even higher clock rates right now (650 MHz is currently the fastest Athlon), but there is no reason to do so.
https://www.tomshardware.com/reviews/athlon-processor,121-16.html
Intel didn't have a response back then either.
Yes if you don't look at the new technologies then you won't see them...While the -10980XE is thoroughly unimpressive from a new technology standpoint,
Disclaimer: I badly want to dump Intel and go AMD. But are the conditions right?
The AMD 3950X has 16 PCIe lanes, right? So for those of us who have multiple adapters such as RAID cards, USB or SATA port adapters, 10G NICs, etc, HEDT is the only way to go.
Someone once told me "No one in the world needs more than 16PCIe lanes, that's why mainstream CPUs have never gone over 16 lanes". If that were true the HEDT CPUs would not exist.
So we can say the 3950X destroys the Intel HEDT lineup, but only if you don't have anything other than ONE graphics card. As soon as you add other devices, you're blown.
The 3970X is $3199 where I am. That will drop by $100 by 2021.
The power consumption of 280w will cost me an extra $217 per year per PC. There are 3 HEDT PCs, so an extra $651 per year.
AMD: 1 PC @ 280w for 12 hours per day for 365 days at 43c per kilowatt hour = $527.74
Intel: 1 PC @ 165w for 12 hours per day for 365 days at 43c per kilowatt hour = $310.76
My 7900X is overclocked to 4.5GHZ all cores. Can I do that with any AMD HEDT CPU?
In summer the ambient temp here is 38 - 40 degrees Celsius. With a 280mm cooler and 11 case fans my system runs 10 degrees over ambient on idle, so 50c is not uncommon during the afternoons on idle. Put the system under load it easily sits at 80c and is very loud.
With a 280w CPU, how can I cool that? The article says that "Intel still can't deal with heat". Errr... Isn't 280w going to produce more heat than 165w. And isn't 165w much easier to cool? Am I missing something?
I'm going to have to replace motherboard and RAM too. That's another $2000 - $3000. With Intel my current memory will work and a new motherboard will set me back $900.
Like I said, I really want to go AMD, but I think the heat, energy and changeover costs are going to be prohibitive. PCIe4 is a big draw for AMD as it means I don't have to replace again when Intel finally gets with the program, but the other factors I fear are just too overwhelming to make AMD viable at this stage.
Darn it Intel is way cheaper when looked at from this perspective.
Disclaimer: I badly want to dump Intel and go AMD. But are the conditions right?
The AMD 3950X has 16 PCIe lanes, right? So for those of us who have multiple adapters such as RAID cards, USB or SATA port adapters, 10G NICs, etc, HEDT is the only way to go.
Someone once told me "No one in the world needs more than 16PCIe lanes, that's why mainstream CPUs have never gone over 16 lanes". If that were true the HEDT CPUs would not exist.
So we can say the 3950X destroys the Intel HEDT lineup, but only if you don't have anything other than ONE graphics card. As soon as you add other devices, you're blown.
The 3970X is $3199 where I am. That will drop by $100 by 2021.
The power consumption of 280w will cost me an extra $217 per year per PC. There are 3 HEDT PCs, so an extra $651 per year.
AMD: 1 PC @ 280w for 12 hours per day for 365 days at 43c per kilowatt hour = $527.74
Intel: 1 PC @ 165w for 12 hours per day for 365 days at 43c per kilowatt hour = $310.76
My 7900X is overclocked to 4.5GHZ all cores. Can I do that with any AMD HEDT CPU?
In summer the ambient temp here is 38 - 40 degrees Celsius. With a 280mm cooler and 11 case fans my system runs 10 degrees over ambient on idle, so 50c is not uncommon during the afternoons on idle. Put the system under load it easily sits at 80c and is very loud.
With a 280w CPU, how can I cool that? The article says that "Intel still can't deal with heat". Errr... Isn't 280w going to produce more heat than 165w. And isn't 165w much easier to cool? Am I missing something?
I'm going to have to replace motherboard and RAM too. That's another $2000 - $3000. With Intel my current memory will work and a new motherboard will set me back $900.
Like I said, I really want to go AMD, but I think the heat, energy and changeover costs are going to be prohibitive. PCIe4 is a big draw for AMD as it means I don't have to replace again when Intel finally gets with the program, but the other factors I fear are just too overwhelming to make AMD viable at this stage.
Darn it Intel is way cheaper when looked at from this perspective.
There is more to the power usage than just the CPU. What you really need to know is total system power draw and guess what the AMD is better than Intel at that, even with 14 more cores. https://www.servethehome.com/amd-ryzen-threadripper-3970x-review-32-cores-of-madness/6/ The reason that happens is that with more cores each core doesn't have to run as hard to get more work done on parallel tasks.yes, sorry, my interpretation was not worded accurately.
Intel simply doesn't have room to add more cores, let alone deal with the increased heat, within the same package.
My point was that Intel is still going to be easier to cool producing only 165w vs AMD's 280w.
How do you calculate the watts, or heat for an overclocked CPU? I'm assuming the Intel is still more over-clockable than the AMD, so given the 10980XE's base clock of 3.00ghz, I wonder if I could still overclock it over 4.00ghz. How much heat would it produce then compared to the AMD?
Not that I can afford to spend $6000 to upgrade to the 3970X or $5000 to upgrade to the 3960X... And the 3950X is out because of PCIe lane limitations.
It looks like I'm stuck with Intel, unless I save my coins to go AMD. Makes me sick to the pit of my stomach
Yes let's be honest here,we can all bash intel all we want but if intel had done all that there would be no AMD left today...a 6c/12t sandybridge would have overshadowed anything AMD had by so much that nobody would have stayed with them.Let's be honest when sandy bridge launched, 6c/12t should have been the mainstream high end CPU, not the HEDT those should have gone up to 10c/20t. And take that thinking further Haswell's mainstream CPUs would have been 8c/16t threads, Broadwell 10c/20t, Skylake 12c/24t and Coffee lake should have been 14c/28t and then comet lake 16c/32t. With each new higher core count costing the same as lesser core count flagship it replaced. But Intel didn't see the value in treating it's consumer base well because they could get away with it as AMD had nothing to challenge them.
Sandy Bridge was already a 200+sqmm die and is getting pretty big by Intel mainstream CPU die size standards. Intel likes to keep mainstream CPUs closer to 150sqmm. While Intel could have axed the IGP to make room, it still isn't giving up the IGP for Comet Lake, so it clearly looks like IGPs are here to stay in mainstream CPUs regardless of how much it may hurt Intel's yield per wafer. Can't really blame them when most PCs ship to offices and other locations that don't require much more than something to connect one or more monitors to.Yes let's be honest here,we can all bash intel all we want but if intel had done all that there would be no AMD left today...a 6c/12t sandybridge would have overshadowed anything AMD had by so much that nobody would have stayed with them.
be able to supply Athlons at even higher clock rates right now (650 MHz is currently the fastest Athlon), but there is no reason to do so.
Intel consciously neglected traditional CPUs back then to try and make itanium take off.Having an even stronger x86 competition than was already present would have been an even worse idea.
Intel was consciously neglecting traditional cores in the last years to push out optane ram that they sell for ~$8000 a pop right now and to put laptop CPUs on m.2 cards as co-processors plus the new i/GPU and nervana.
Yes if you don't look at the new technologies then you won't see them...
- Intel® Deep Learning Boost (Intel® DL Boost)Yes
- Intel® Optane™ Memory Supported ‡Yes
- # of AVX-512 FMA Units2