AMD CPU speculation... and expert conjecture

Page 117 - Seeking answers? Join the Tom's Hardware community: where nearly two million members share solutions and discuss the latest tech.
Status
Not open for further replies.

8350rocks

Distinguished


Titan and the GTX690 while quieter, also run quite a bit hotter...or did you read Tom's review? Under load the GTX 690 can reach temps of 90-100C and it would require additional cooling to be overclocked beyond the factory boost.

Additionally AMD's HD 7990 has 3 intake fans and vents alot of heat out the rear of the card.

Titan gets destroyed by the HD 7990 too btw...that's why nvidia is coming out with the Titan Ultra. Additionally, you don't get 7-8 games with Titan, so you're paying more for less.

Titan makes no sense, and the GTX 690 might be the quietest card, but it runs the hottest against the vendor cards that were factory overclocked, etc. Though the Powercolor card was the power hungriest, the HIS 7970X2 in the test was clearly the best performing card...Tom's, in unusual fashion, showed a preference for the nvidia card despite the fact that it's cooling was subpar and performance was 3rd best. They made a huge issue out of coil whine as well.
 

8350rocks

Distinguished


That's a supply issue...when they're 1/17th the size of intel, does it surprise you at all that they don't make as many chips?
 

kettu

Distinguished
May 28, 2009
243
0
18,710


I haven't heard that before. Got a link, I'm interested to read more about it.



That's a difficult call. It could be either way but I think it doesn't matter in the big picture. As far as I'm aware the front-end throughput difference clock-for-clock between BD and PD is minor at best. It's basically few tweaks here and there across the entire module that gave a nice performance boost for PD. A major contributor being the improved power efficiency allowing higher clockspeeds.

My understanding is that this 30% throughput increase is just refering to the improved fetch hardware and the addition of the second decode unit. (I can't remember off-hand if there were some other improvements there.) And that this increase is realised in situations where the module is fully loaded. How that translates to real world performance is another matter. In cases where the cores were starved I think the gains will be significant. I wouldn't be surprised if it was pretty close to 30%.

On top of that the actual core has it's own tweaks that will improve performance in other cases. But my guess is that BD/PD cores are being atleast somewhat starved in most multi-threaded situations. So in strictly single-threaded situations the performance improvement won't be that awesome. Is 7% - 10% the norm these days between CPU generations? But in most multi-threaded realworld cases it's going to be better than that.

This kind of a multi-threaded performance gains are a big deal for AMD. They're going to be much more competitive in the high-margin server market. It's also a pretty big deal in consumer markets with games becoming more and more threaded. I don't see how Intel can compete on gaming performance with their multiplier-locked dual-core line-up against Steamroller quad-cores. Now there will be lightly-threaded cases where i3 will be ahead of the steamroller but the problem I see is while SR's single-threaded performance is "good enough" I don't think same can be said for multi-threaded performance of i3. My guess is were going to see an unlocked i3 next-generation. That's the only way I can see them still being relevant. The i5 series will definately be a contender and with multi-threaded games I don't see unlocked i3's eating that much sales against i5 because it will be superior. Rattling sabres with the six-core SR.

As for the FX-6300 availability:
It can be had in Finland for ~130-140 euros. Roughly ~10-20 more than an i3 and ~30 less than FX-8320
 

8350rocks

Distinguished
Steamroller articles:

http://www.xbitlabs.com/news/cpu/display/20130331080217_AMD_We_Are_On_Track_With_Steamroller_Micro_Architecture_in_2013.html

http://www.bit-tech.net/news/hardware/2012/08/29/amd-steamroller/1

http://en.wikipedia.org/wiki/Bulldozer_(microarchitecture)

http://www.brightsideofnews.com/news/2013/3/6/analysis-amd-kaveri-apu-and-steamroller-core-architectural-enhancements-unveiled.aspx

Those articles should cover the amount of information available at this point in time...the last one discusses FPU enhancements in some detail...I couldn't find the link to the article that gave a thorough engineering level breakdown of the improvements, but those "layman terms" articles should give the gist of it.
 

kettu

Distinguished
May 28, 2009
243
0
18,710
Thanks. If you happen to stumble upon the more thorough article please share.

Edit:
What I got out of them is that AMD is dropping one MMX pipe from the FPU and reducing FP pipelinestages from 4 to 3 reducing instruction latency. Also various undisclosed tweaks.
 

one fan in titan and gtx690 vs 3 in 7990, with open air cooler design to boot. cooling the card would obviously be better than blower and central axial types. what's not obvious is that the cooler choice will affect heat buildup inside the case and case choice- which was my concern. to repeat - gtx690's cooler blows part of the hot air out of the case, titan's blows most of it out, 7990's dumps all of it inside the case.
just in case, casually trying to dismiss by 'like, most/all cases totally have side vents so what-everr..' - won't change anything. :D
from toms' review:
There is a price to be paid for this sort of design, though. Three fans, side-by-side, are only effective if they aren’t doing battle with each other. And that means channeling air vertically, rather than horizontally. At the end of the day, then, the slotted rear I/O panel that typically helps exhaust hot air from most graphics cards is almost non-functional. Instead, waste heat from both GPUs is jettisoned out the top of the card, right into your case.
now... i don't know what you consider 'rear' of the card.... may be you meant the 'other' rear...wherever that is..... :whistle:

:lol: no amount of 'destroying'(!) can take away the ease of using single, power gpu and the sheer possibility of combining with may be another two. then there's it's easy fitment in smaller cases worrying much less about heat buildup. since these are all $1k cards being discussed, these kinds of factors matter quite a bit. titan being outperformed by 7990 is not really anything earth shattering - titan's configured to underperform gtx690. titan's purpose is just different.
additionally, 7990 suffers badly from amd's driver support. amd's trying so hard to make 7990 look like it's worth the $1k yet they neglect to provide good drivers for it. i noticed 7990 getting hilariously 'destroyed' by [strike]titan[/strike] 7990(prototype drv) in crysis 3 (titan outperformed and kept up in some others), especially in practical fps. it should not have happened because 7990 has more powerful gpus, yet it did happen. 7990 woulda looked worse if amd hadn't provided optimizations for hitman, skyrim and others and the prototype driver.
may be in a few months or later this year, 7990 will seem worth it, but the driver delay is inexcusable.
titan 'ultra' is showing amd that nvidia can have multiple halo gfx cards(with better drivers) while amd is struggling to make one seem worthwhile. marketing.:p
tbh, i think you entirely missed the reason why nvidia is suddenly making so many high end cards available. the real reason is so simple- improved yields and/or stockpiling, and amd's clean sweep with the consoles, especially with ps4. imo from todays benches, seems like nvidia could care less about 7990 since it doesnt have that vital software component available yet.

heh. you can bundle all the games in the worlds for free, it won't count towards calculating price/performance of the card. if 7990 was $200 or more cheaper, may be woulda mattered. may be. in the end, you end up paying a lot more for quite less - with 7990.

titan makes no sense... to people who don't understand the purpose of halo products and simple marketing tactics; i'll just leave it at there. his' card never made it to the market, they withdrew. his is selling amd's reference 7990 instead of their own '7970x2'. coil whine is annoying and downright wrong for a highest end, halo card which should have top notch design. if it was 200-300 bucks cheaper, no one woulda cared.
.... most of these are addressed in toms' review at least, others touch on these in varying degrees. i sense lack of attention. :whistle:

size means nothing compared to execution. they have almost all other new cpus available but... just not the one that has the best chance for selling.

they don't make 'as many chips'? they flooded markets with llano apus and zambezi cpus that no one cared to buy soon after trinity and vishera came out. they make plenty of chips alright.
 

8350rocks

Distinguished


The card has a slotted design with exhaust vents on the rear of the card, where are you coming up with the design doesn't exhaust heat out of the rear of the card? The design schematics indicate that is not proper function. Perhaps the engineering sample was not 100% ready yet?


now... i don't know what you consider 'rear' of the card.... may be you meant the 'other' rear...wherever that is..... :whistle:

It's exactly where the design schematics indicated it should be...


:lol: no amount of 'destroying'(!) can take away the ease of using single, power gpu and the sheer possibility of combining with may be another two. then there's it's easy fitment in smaller cases worrying much less about heat buildup. since these are all $1k cards being discussed, these kinds of factors matter quite a bit. titan being outperformed by 7990 is not really anything earth shattering - titan's configured to underperform gtx690. titan's purpose is just different.
additionally, 7990 suffers badly from amd's driver support. amd's trying so hard to make 7990 look like it's worth the $1k yet they neglect to provide good drivers for it. i noticed 7990 getting hilariously 'destroyed' by [strike]titan[/strike] 7990(prototype drv) in crysis 3 (titan outperformed and kept up in some others), especially in practical fps. it should not have happened because 7990 has more powerful gpus, yet it did happen. 7990 woulda looked worse if amd hadn't provided optimizations for hitman, skyrim and others and the prototype driver.

HD 7990 drivers are in beta, how are they deficient? By the time the product is launched it should be fine...by the way...the Crossfire issues have been fixed now, and I have every reason to believe that the next set of drivers will be fantastic with Kudori being responsible for that.

may be in a few months or later this year, 7990 will seem worth it, but the driver delay is inexcusable.
titan 'ultra' is showing amd that nvidia can have multiple halo gfx cards(with better drivers) while amd is struggling to make one seem worthwhile. marketing.:p
tbh, i think you entirely missed the reason why nvidia is suddenly making so many high end cards available. the real reason is so simple- improved yields and/or stockpiling, and amd's clean sweep with the consoles, especially with ps4. imo from todays benches, seems like nvidia could care less about 7990 since it doesnt have that vital software component available yet.

The drivers issue will be fixed in short order.


heh. you can bundle all the games in the worlds for free, it won't count towards calculating price/performance of the card. if 7990 was $200 or more cheaper, may be woulda mattered. may be. in the end, you end up paying a lot more for quite less - with 7990.

No, what matters is it outperforms the GTX690 and GTX Titan, so you're getting more for the same money. If Titan was about $600 then it would make sense...especially considering the HD 7970 blows it away, and the GHz edition will nearly keep up with the GTX 690 in many applications.


titan makes no sense... to people who don't understand the purpose of halo products and simple marketing tactics; i'll just leave it at there. his' card never made it to the market, they withdrew. his is selling amd's reference 7990 instead of their own '7970x2'. coil whine is annoying and downright wrong for a highest end, halo card which should have top notch design. if it was 200-300 bucks cheaper, no one woulda cared.
.... most of these are addressed in toms' review at least, others touch on these in varying degrees. i sense lack of attention. :whistle:

Coil whine is a product of the massive cooling required to keep such a high end card cool. Would you rather have something inside your case running at 90C or 70C? Considering that most people will be running high end cooling systems that own these cards, and, will not be using an open benchmark setup like Tom's review did, the noise is acceptable. If you're running water cooling on your system, you won't even hear the fan in most cases over the massive push/pull fans on your radiator. So what's your point? This is not a product for a low end gaming machine...this card goes into a $2k computer build, and under those circumstances, you could use a high end sound dampening case to reduce any noise anyway.

they don't make 'as many chips'? they flooded markets with llano apus and zambezi cpus that no one cared to buy soon after trinity and vishera came out. they make plenty of chips alright.

Considering they possess 1/4 of the market share, I would say intel produces 3x more chips than AMD. Their biggest market presence is in the US, no doubt they have supply issues overseas, and that's something they could actively work toward remedying, but haven't addressed other issues first which are more important.

 




AMD doesn't make chips; they sold their fab unit, remember?
 


lame, baseless excuses. amd or anyone in the right mind won't give below-standard products to reviewers. they are well known for supplying reviewers with [strike]golden[/strike] cherry picked, best samples that deliver the best performance in order to look good in the benchmarks and tests.
i said the 7990's cooling design dumps heat inside the case. the vent in 7990 only dissipates minor part of the hot air from the closest fan, not from the other two further inside.
btw, fanboy linear mathematics (FL math(tm)) like '1 fan out of 3 is fiiine' does not apply here. :D


they're not just deficient, the sheer delay when this premium 7990 card is launching within weeks is bad. amd themselves have said that the prototype driver, the one that actually improves multi-gpu performance and experience, will come out in 2h 2013, could be as early as june or as late as october. either way it's way too long. per-game optimizations (in catalyst profiles) won't carry them for long, they only look passable in benchmark reviews, not so much in real life..

7990 trades blows with gtx 690 and only outperforms titan when the driver isn't making sure it can't. and performance/price does favor gtx690, titan only falls short in 1600p, if tpu's analysis is of any indication. even if i think tpu favors nvidia, 7990's perf/price against amd's own 2x 7970 is much worse. you really pay much [strike]less[/strike] more with 7990 and get less. the things that should justify 7990's extra price, don't deliver. worse, counting drivers.
titan shoulda been $600 or less.. even, only if there was a similarly competitive card from amd. your claim about 7970 beating/keeping up with titan and gtx690 is only right for niche gpgpu applications, that's nowhere near the general case. i noticed that you didn't use the word 'games'. :D

you need to read the section where toms' addresses this noise issue. seems you didn't understand it. i apologize for using 'coil whine' directly from your post due to negligence.
it's not because of 'massive cooling', the reason 'massive cooling' is required because the gpus produce more heat while trading blows with gtx690.
bringing up user-customized cooling, other cooling apparatus making more noise, 'high end sound dampening' and so on - make amd's cooling seem even worse. :lol: those are lame excuses anyway.
my point is simple - highest end products should be very, very good, especially in hardware. they should represent the best (of the best), the most innovation, the essence of a halo product. amd underdelivered (worse in some cases) in all of them. it's undoubtedly much better than previous 6990 and it does have the raw performance, but the whole package is sub-par for the price and for the competition. if it was cheaper, no question woulda be raised.

lawl, it's the opposite of the 'amd has significant share (>30%, iirc) in desktop cpus' pitch! just to make amd's poor execution look better.:no: there are no other issues more important than selling cpus - the right ones, as many as possible(regardless of yields), in potential markets. no one cares about intel when the fault is amd's own. :D
edit: fixed the brainfart and other errors.
 

8350rocks

Distinguished
So PS4 is heavily HSA oriented and the APU is really going to act like a CPU/GPGPU combination capable of running approximately 96 simultaneous threads based on the pipelines and capabilities of the 8 core jaguar.

It seems no PC will be able to keep up with that for some time to come. SR is only going to be able to run 44 simultaneous threads, and tier 1 GPUs (HD 7950+) allow running 4 compute threads at once...for a total of 48.

 
Hmm, I doubt that the PS4 will truly have the advantage, but I had my doubts with the PS3's RSX and its 128-bit bus since the 7900GTX it was based on had 2x the bandwidth, but of course does not run the games at anywhere near as good quality as the PS3's RSX+Cell (largely thanks to optimized code and not sheer power). However, I do think that PS4/Durango ports should optimize SR and PDs 2 /4 unused cores in modern titles, thus possibly gaining ground on Team Blue, but I would think that Haswell would still have better IPC.
 

juanrga

Distinguished
BANNED
Mar 19, 2013
5,278
0
17,790


As said the 35W TDP difference is considering that one is a K series. If you don't overclock the power consumptions are similar. The difference is due to slightly higher stock clocks.

19865_pwrl.png


About the HD 7990 vs GTX 690 vs GTX titan, a resume of what said:

the HD 7990 is faster than the GTX 690 in most tests that we ran, but not in all of them.

[...]

Even more surprising is the card's lack of noise and heat, [...] We'd say all in all it's probably just a bit quieter than the GTX 690, and gets about as hot.

http://www.maximumpc.com/article/videocards/amd_radeon_hd_7990_first_look

As the results on the preceding pages have shown, the Radeon HD 7990 puts up some impressive frame rates, outpacing the similarly priced GeForce GTX 690 and GTX Titan more often than not.

[...]

Power consumption under load was relatively high for the Radeon HD 7990, but we were expecting that to be the case considering the card is packing 8.6B transistors and 6GB of memory.

[...]

With that said, the Radeon HD 7990 is immensely powerful and it runs surprisingly cool and quiet, considering its performance.

[...]

The AMD Radeon HD 7990 is a beast.

http://hothardware.com/Reviews/AMD-Radeon-HD-7990-Review-The-Silent-Beast/?page=13
 

Cannot wait for some more P̶e̶n̶t̶i̶u̶m̶ ̶4̶ ̶P̶r̶e̶s̶s̶c̶o̶t̶ ̶v̶s̶ ̶A̶t̶h̶l̶o̶n̶ ̶6̶4̶ ̶W̶i̶n̶c̶h̶e̶s̶t̶e̶r̶ ̶ Intel i5 Haswell vs AMD FX Steamrolller action!

 

... you coulda at least posted the review link. anyway, interesting info. in the review, it says right at the top of the page that motherboard models affect power draw. the similar peak draw seems unusual because no one else seem to have measured the same numbers. if correct, it makes a10 5700 as unsuitable as 5800k for sff. i think 5700 and 5800k's similar consumption can be explained. xbitlabs noticed (and amd confirmed) that when certain artificial loads are put on the apus, protection kicks in and throttles clockrate under base rate as long as turbo core is enabled. 5800k coulda been throttling harder under load. i could be wrong about this though.
here's a hwmonitor screenshot from guru3d's a10 5700 review:
http://www.guru3d.com/articles_pages/amd_a10_5700k_review_apu,8.html
you can see the apu's power use @stock in the power section - not similar to a10 5800k in any way.
http://www.guru3d.com/articles_pages/amd_a10_5700k_review_apu,1.html
there's a significant difference between 5700 and 5800k's power use.
http://www.goldfries.com/images/hwreviews/2013/amd-a10/hwmonitor.jpg
this one is from a lesser known review site (my web filter doesn't rate it highly so you've been warned.), but it shows a10 5800k's power use @stock. also keep in mind that they use a gigabyte board while guru3d used asus iirc and vortez used asrock.

HH tested frame time and their conclusion is similar to others. mpc... just slapped things together.thanks for the synopses, changes nothing though. :)
 


Wut?

Explain how an 8 core processor can run 44 threads at a time. I REALLY want to hear this explanation.
 

8350rocks

Distinguished


Easily...

Each core is capable of 4 threads at once (4x8=32)
Each FPU is capable of running 3 threads at once (3x4=12)
Total:44

Those are engineered maximums...they cannot be exceeded. Now, they might not hit 44 at once without supremely good software utilization, but the capability is there.

By contrast, the i7-3770k can run 36 threads at once:
4 per real core and 2 per HT process: (6x4=24)
3 FP calculations per core (2 core/1 HT): (3x4=12)
Total: 36

The i7-4770k is rumored to be able to run 40 threads at once. though I am not sure of the breakdown.
 


Someone doesn't have a clue how CPU's work, apparently.

One set of registers = One thread on the core. 8 cores = 8 threads.

So yeah, I think you're talking out of your behind just a little bit...
 

8350rocks

Distinguished


It's CPU multi-tasking...this is EE 101 stuff guys... Back when you had a single core CPU...(some of you may not have been old enough to remember...not sure of ages here)...Your CPU is capable of multi-tasking. This means that the maximum number of threads a CPU can run at once are engineered maximums before the pipelines fill up. There is an acceptable limit to the amount of clocktime the engineers originally designed to allow to be split up under heavy load.

By concurrently, I don't mean 100% from each core to each thread. I am talking about one thread receiving closer to 60% clocktime and the other 3 splitting up approximately 40% or so...(not sure on the % split there...could be 50/50 or 70/30 etc. depending on design limits)...so while work is being done on background threads in each core, the foreground thread...(or highest prioritized thread)...is receiving the most resources until it is completed, then the CPU moves on to the next thread in order of priority, and continues working on the same threads it was while picking up one more in the open pipeline. If one thread is completed while running in the background, that pipeline hands the next thread to the scheduler, and the process is begun all over again.

This was the old school way they made your CPU able to run multiple programs at once...and why you noticed your single core CPU bogging down after more than 2-4 processes were open at once...do you guys think they lost 15 year old technology all of the sudden?

It won't execute them all at 100%, in fact, at those loads, no single thread would get 100% resources from any core/FPU...the % would be quite a bit less than that. Though rather than processing one single thread per core, you're doing more work in the same time frame...and enhancing the speed with which you can complete large voluminous tasks.

For example...on the i5-3570k...(without HT obviously)...you could have something like these programs running all at once:

Internet Explorer (1+ threads)
Windows Media Player (streaming) (1+ threads)
MS Word (1+ threads)
iTunes (downloading) (1+ threads)
Skyrim (1+ threads)

Now, by your logic, you could only run 4 of those or fewer...but your PC doesn't stop you and say..."Oh, no...you can't do that sir! We don't have 5 cores!" does it? No...clearly not. So what happens is, the programs scheduled with highest priority receive more resources than the background processes.

I am astonished that you guys don't already know this, or have forgotten about it. This is simple distribution of resources.
 
It's CPU multi-tasking...this is EE 101 stuff guys... Back when you had a single core CPU...(some of you may not have been old enough to remember...not sure of ages here)...Your CPU is capable of multi-tasking. This means that the maximum number of threads a CPU can run at once are engineered maximums before the pipelines fill up. There is an acceptable limit to the amount of clocktime the engineers originally designed to allow to be split up under heavy load.

Multithreading != Multiprocessing.

Multithreading: Using instruction level parallization to keep a single CPU pipeline moving at an optimal rate.
Multiprocessing: Using multiple CPU's or CPU cores to run multiple threads at the same time.

SMT is capable of running multiple threads on a single CPU, but aside from Intel Hyperthreading, isn't done on X86 CPU's. You are limited to run one thread context at a time per set of registers. Basically, the OS scheduler sends out any threads that are "ready to run" (non-blocked), and the CPU thread scheduler will scheduler appropriately. Under a Windows OS, the highest priority threads run, until they either finish, or block (cache miss, for example). Priorities get adjusted by the OS the next time the threads go out to the CPU.

For example...on the i5-3570k...(without HT obviously)...you could have something like these programs running all at once:

Internet Explorer (1+ threads)
Windows Media Player (streaming) (1+ threads)
MS Word (1+ threads)
iTunes (downloading) (1+ threads)
Skyrim (1+ threads)

They don't "run at once", they swap as needed, per the OS scheduler. Thats the entire point of preemptive multithreading. Using your example, at any one instant, only four of these applications could actually be running. The other one would be waiting until either its priority was greater then another application, or another application was blocked. But at any instant, you are limited to running FOUR threads at a time.

This also means that threads don't "share clocktime"; they are scheduled by the OS, given a priority, and the CPU will run the threads accordingly. Now, if a thread blocks (for any number of reasons), modern CPU's are smart enough to run some other thread instead, but there isn't any magical "Oh, this thread is getting 70% runtime, and this one gets 30%" that is determined beforehand.
 
Status
Not open for further replies.