haswell or piledriver

Page 6 - Seeking answers? Join the Tom's Hardware community: where nearly two million members share solutions and discuss the latest tech.
Solution
If you are planning to upgrade later on, Piledriver is in an architecture AMD is committed to until 2015 (AM3+ socket) and that socket will get Steamroller and eventually Excavator, if you go with something by intel, Haswell will be their new architecture so they will commit to that for at least 2 years.

If money is at all something you're concerned about, go with AMD, for the money you would spend on an intel system, a comparably built AMD system will have more processor and more GPU for the same money...

Also, PS4 and XBOX 720 are all running AMD hardware in their consoles, so future games will be optimized for AMD architecture anyway.


Look this is not going anywhere. I'd give no more reason to trust a manufacturer over someone else (other than hard claims such as 128 bit interfact). Claims of "up to 20% gain" are worthless. Look at bulldozer and what it was promised to be and what it delivered. There is a strong link between theoretical and actual performance in many cases. Its not an absolute but there is s strong relationship (ex the link between a synthetic such as cinebench and real world-adobe cs encoding is fairly linear).

The sandra sp5c results are one set of results that show a sandy bridge i3 (not ivy) still doing better than the a10. The sandy bridge cpu is using RAM that it is not rated for (only 1333 mhz for sandy bridge) and is a clear outlier among the other data. Its using 'overclocked' RAM (as you keep putting it) and so is not expected to show good scaling.

The e-350 will not be ram bottlenecked in any way by slower ram.

i7-975 (about much faster shows little or no difference in ram speed in that test). This is triple channel ram but 800 mhz triple channel is comparable to 1333 dual channel. If a much more powerful processor shows no difference then a puny e-350 is not going to care. (Pentium 4 beats the e-350 in that test and it has less bandwidth). None of the other cpu's have any faster ram there. e-350 needs bandwidth for the igp which is not used in the test.

032_premiere.png


025_photoshop.png
 


1) Yes, Bulldozer was a kind of disappointment for many people (Intel also makes mistakes). But I again wonder how you select a special case and take it as the general rule: as-this-happened-once-it-will-happen-now, as-this-game-run-bad-many-games-will-run-bad... You also miss that there are more than one company involved in the development of the PS4.

2) The Sandra sp5 results show that trinity A10 performs on the same league than the sandy i3 regarding the memory subsystem:

sisoft-mem.png


The difference of about 2.6 Gb/s is really minimal for everyone except you. It is specially minimal when one notices that the A10 was run at stock speeds whereas the i3 was run with memory overclocked by a factor of about 40% (on MHz) on a Z77 motherboard. The immense majority of owners of an i3 will be using stock speed memory and cheaper chipset motherboards and their bandwithds will be smaller. Next is an older test on an ivy i7 that gives an idea of the sensibility of this test to changes on the memory speed:

mem%20scaling%20sandra.png


The bandwidth increases about 7.5 Gb/s when memory is overcloked by a factor of 40% (MHz). Even if you take a half of that value for the sandy i3, this closes the above 2.6 Gb/s gap.

Therefore, it is fair to say that trinity at stock is on pair with an i3 (sandy) at stock. Your claims against AMD were a complete exaggeration.
 


AMD also claimed "up to 56% performance gain with trinity vs llano" for mobile. Anandtech article for a10-4600m. In reality trinity increased games performance by 19% average and 3d mark performance by 35%. 19 vs 56 seems a massive exxageration but all companies to this (nvidia is especially bad).

Yes the bandwidth increases by 7.5 GB/s that is however an increase in bandwidth of 7.5/17.1=44%. That is impressive scaling.

That cpu is sandy bridge not ivy bridge.

That same graph shows trinity with less bandwidth than llano, stepping back from your previous generation is kinda a fail.

If we assume that trinity is strongly bandwidth bound (which scaling seems to suggest) we would expect trinity to perform (17.8-15.2)/15.2=17% better with i3 level bandwidth. That is basically the improvement of richland on the high end level. (If you look at the gpu clocks the richland a10 is only about 6% faster than the a10 trinity).

I don't think you also got the part where I said the immense majority of owners of trinity will not be using 1866 mhz ram (I don't think any non-customizable storebought trinity system shipping in volume uses ram faster than 1600 mhz). I don't think systems with a i7 or 8350 are using 1866 mhz ram. Basically the only people who are using trinity with faster ram are those who are building it themself, which is a small section of the market.

Anyway here is the memory bandwidth of my system using the newest version of sandra (released march 24). i7-3630qm 8 GB CL 11 DDR3 1600. Approximately 20 GB/s. Which matches to the i7-3770k (its a desktop using lower latency ram + margin of error). This is an i7 but its a mobile version using mobile RAM at a higher latency.

memorybandwidth.png
 


Ok, we'll see what happens when they arrive.



Actually, in intel terms you're right...because all intel threads share resources with their cores.

In AMD terms, you're wrong...here's why:

AMD Modules have 4 times the L2 cache size being shared...so if you cut each modules cache in half, it's still twice the size of the intel chips cache. Plus, when you run threads on intel chips, you're sapping resources from a "real core". In an AMD processor, the architecture is designed so 8 cores can run up to 3 threads each, and the FPUs can run 2 threads each...so if you're running 8 or less threads in HSA architecture, you can actually not tap resources from one of your other cores. In intel, once you go over 4 threads, you're splitting up resources. So, the architecture differences, (as you pointed out above) mean that you cannot generically say that 8 cores are hyperthreading, because it's not the same thing.

Now, when steamroller comes (I realize this is slightly putting the cart before the horse, but bare with me...)...each core can run 4 threads and each FPU can 3 threads. The instructions will be more streamlined internally and a 30% single thread performance gain is expected...as well as a 15-20% performance gain in multi-threaded applications.
 
The good thing about hyperthreading is that it is very power efficient in heavily threaded workloads.

visual%20studio.png


Take a real world benchmark. The i7 is 34% faster than the i5 when it only uses a maximum of 13% more power for the system (idle usage is 67 watts so using just cpu performance power increase is 31%-- 34 vs 31 is an increase in efficiency). Considering that power vs performance does not increase linearly this represents a good gain.

AMD's method is good too in that they get much more performance than hyperthreading, unfortunately the module architecture only stretches its legs when at full load. Some tests show that there is a performance gain by deactivating every core of the bulldozer/piledriver module in single thread loads.

load%20power.png
 


1) In the first place I find truly amazing how you pick a 56% from Anandtech, then a 19% and a 35% from two arbitrary tests and you decide to compare the 56 against the 19, but not against the 35 in your new attempt to bash AMD. In any case, this is unimportant, because once again you and Anandtech are plain wrong.

Anandtech is omitting important info... What AMD said was that trinity A10-4600M would give up to up to a 56% increase over Llano A8-3500M in visual performance based on the figures of 3D Mark Vantage.

http://ir.amd.com/phoenix.zhtml?c=74093&p=irol-newsArticle&ID=1695793

Let us then see to what finally measured

3DMark%20V%20graphics.png


The measured increase in performance is of 55%, which is close to the promised 56%.

Moreover, trinity was found to be better than promised

http://news.softpedia.com/news/AMD-Trinity-A8-4500M-APU-Benchmarked-Better-Than-Expected-262701.shtml

2) In the past, you said that AMD trinity was a fail because i3 sandy was better on Sandra. I showed you this was not true and now you change the claim to AMD trinity is a fail because llano was better on Sandra. Well if you were to use your logic ("stepping back from your previous generation is kinda a fail") with ivy Bridge you would claim that ivy Bridge is a fail because an i7-3770k (ivy) gives less bandwidth than an i5-2550k (Sandy) in the same test

sandra%20mem%20bandwidth.png

 


Yes. Meanwhile, we can read what Sony said some days ago at Game Developers Conference:

Using DDR5 RAM inside the PS4 will mean that the system will be able to produce a higher quality of games than some high-end systems that are only equipped with DRR3 based memory.

or read enthusiasts

http://gamingbolt.com/ps4-will-will-out-power-most-pcs-for-years-to-come-avalanche-studios-cto
 


Some console developer says consoles are better than PCs... big deal. That's like a car salesman telling you a new SUV gets better gas mileage than a two-year old compact.

GPUs have been using GDDR5 for years. Everyone's trying to act like these consoles are going to show us tons of new hardware we haven't seen before. They're not. The chips themselves may be new models, but they're not new technology.

Anyone who spent $1000 in parts (spent wisely) on a gaming PC with the last 1-2 years will have something much faster than a next gen console. If you think a console with an APU is going to outperform a CPU and dedicated GPU using hundreds of more watts of power, you're drinking the marketing kool-aid.

If AMD had that technology, why aren't they putting it in their own parts? So they have this <150W APU they can manufacture for < $200 that outperforms an 8350 and 7970 put together... but they're saving it for Sony? Get out of here with that bull-ess. Are you guys even thinking about this garbage before you pass it off?
 


DDR5 is coming to desktops...and it's onboard on the consoles(in the form of shared DDR5), so they have access to RAM that you don't which means the CPU/GPU will be communicating data back and forth about 2-3x faster than your PC could ever dream of...so unless you're running something like a HD 7990, you won't be in the same ballpark.
 
So now consoles are going to be as powerful as a PC with a 7990? AMD themselves can't make a single GPU chip as powerful as their own 7990, but you think they're going to make a low power APU that's faster?

You are so full of it -- stop misleading people.
 


I am talking about data transfer rates...follow along...stop falling behind.

The 8 GB of DDR5 RAM on the console systems is MORE RAM than you have on a HD 7990, and data transfer rates are the same, so speed of data transfer is going to be even higher.

Plus, in the new architecture, CPU and GPU will be able to share memory addresses for specific things so they can work on the same threads/applications simultaneously.

THAT'S why Sony/Microsoft chose AMD...AMD already supports this technology in their GPUs and the Richland APUs are going to get it integrated. Plus, the FX series CPUs are going to be receiving a similar instruction update in the Steamroller iteration.
 


Regardless, the 56% was far less than the actual performance increase measured in games. Sorry I did not see that quote. AMD intentionally lead people to believe that trinity would be a massive gain in gpu performance when really in REAL WORLD tests the performance gain was about 1/3. The promised 20-40% gain of richland over trinity could be similar.

For an apu memory bandwidth is important. For strict cpu related tasks its significantly less so. I said this plenty of times.
I said many times that you must use a margin of error of around 5-10% when looking at something like memory bandwidth. 0.5 GB/s/20 is pretty insignificant.

Anyway here's another comparison of memory bandwidth between the a10 and the i3.

AIDAREADWRITECOPY_zps379b8194.png


Here is the other slide from toms hardware that has been posted before. Bandwidth is more than before. Either of way its margin of error.

mem%20scaling%20sandra.png


 


Benchmarks on PS4? That's a good one...

Maybe while I am benchmarking product that isn't out yet, I can also find a way to turn lead into gold?

You can find the specs...you can read about the system...google PS4 specs...I bet there are no less than 10-15 articles that come up right away talking about the hardware...

Do I have to explain the difference between DDR3 and DDR5 as well, or have you got a grasp on that?
 


1) It is not "Some console developer", it is the entire gaming developers ecosystem which is impressed by Sony next console. John Carmack (Id Software) already praised Sony for their choice of hardware; Eurogamer calls the graphics part of the console "impressive"; the console is being considered a "perfect gaming PC"; in fact, Mark Rein (Epic Games) has said yesterday about the PS4:

It's like giving you the world's best PC.

2) Believe it or not, but Sony, AMD, game developers... all they know that GDDR5 has been available on GPUs (yes the people who makes the cards and the games know this trivial stuff). The point is that most graphics cards for PCs have only 1 or 2 GB of GDDR5, with very high-end cards having 4 GB. This console has 8GB GDDR5 (not even an overly-expensive Titan has that!)

3) You also miss that this console is having a unified memory model and that its basic memory subsystem gives about 120 GB/s of bandwidth, whereas your high-end PC will be giving, at best, 25 GB/s.

4) You also miss that this console will be accessible close to bare metal level which is something you cannot even dream in your Windows 7/8 high-end PC.
 
8350, if you're going to make claims a PS4 is faster than a modern gaming PC, post your sources. It's a stupid claim and I'm not going to do research to prove you wrong -- I don't have to -- the burden of proof is on the one making the stupid claims. I've caught you lying about things multiple times.

juanrga, I understand a PS4 will use shared GDDR5. The CPU portion of the APU will have faster system memory to use -- that should help with some tasks -- but CPUs now aren't memory constrained. The benchmarks you guys are passing around are largely irrelevant because the effects essentially just show up in memory benchmarks. There's a reason there hasn't been a big push for DDR4 on desktops; it's not needed. Also, as a whole, even with faster memory, the CPU is still going to be a lot slower than modern quad core CPUs. You guys try and try to dispute this all you want, but they simply can't overcome physics. It's like expecting a 35W laptop part to keep up with a 130W server part. Or a 5W phone SoC to keep up with a 35W laptop CPU. In order to match a modern desktop CPU, they're going to need to match the transistor count and also the frequency -- that's not happening. To be clear I'm talking about chips manufactured at the same time. Don't bother going back in time to bring up 386s or something.

GPUs already use GDDR5 (and have been for years), so I don't know why anyone is expecting increased memory bandwidth for the component most responsible for graphics quality. Regardless, memory bandwidth is only responsible for part of the GPU's speed. The GPU in consoles is going to be entry-level gaming PC level, that's about how good the games will look on consoles. Again, they're not going to overcome physics. If AMD could make an APU that could compete with their enthusiast gaming cards for less money, they'd already be doing it.

Consoles are going with shared GDDR5 probably because it's faster for the GPU (edit: as opposed to using DDR3 for the whole system like Trinity, for example), and also cheaper to use one set of memory instead of two separate sets. They could have used DDR3 for the system memory and it wouldn't matter from a performance standpoint, but it would be more expensive to manufacture the unit.
 
Forgot to mention this. Console developers aren't programming in machine language for the same reason PC developers aren't. Not that it's not possible, but because it takes forever (as opposed to using more powerful high-level languages) and you don't get much benefit from it. Program inefficiency is largely due to lazy developing, not poor high-level compilers. Console developers may be more diligent at making their code efficient, but then again, they have to be.
 
Well at least you tried. This is better than you making wild, baseless claims.

"AMD has told us several times that Kaveri should be able to hit the 1.0 TFLOPs level of performance and if we compare to the current discrete GPUs would enable graphics performance similar to that of an under-clocked Radeon HD 7770."

So the guys from the site you linked (assuming AMD isn't embellishing like they did with Trinity) agree with me that the PS4 will be about as powerful as a current entry-level gaming PC. What have I been saying the entire time?... Next-gen consoles will be comparable to current entry-level gaming desktop PCs -- or relatively high-end gaming notebook PCs.

Development kits are just that. They allow developers to start making games before hardware is finalized. Otherwise there wouldn't be any games for 1-2 years after launch.
 


Sony has already stated onboard graphics will be 1.84 TFLOPS capable, which is between HD 7870 and HD 7950 performance for graphics. The speculated speed of the processor is ~2.0-2.4 GHz with 8 cores.

Furthermore, if the development kits shipped with the FX8350, ports to PC should go exceptionally well...

So, CPU wise, you'll be on par with a mid-high end 4 core because of data transfer rates that aren't available on desktops, and graphics wise, it's equivalent to roughly a HD 7890 (if such an animal existed)...with 4x the memory.
 
POST

YOUR

SOURCE

Your speculation means nothing to anyone, anywhere. What you say is contrary to current technology and the laws of thermodynamics. I've already proven you'll lie about facts.

I'm tired of talking to you. Is there a way I can unsubscribe to thread updates?