AMD CPU speculation... and expert conjecture

Page 252 - Seeking answers? Join the Tom's Hardware community: where nearly two million members share solutions and discuss the latest tech.
Status
Not open for further replies.

juanrga

Distinguished
BANNED
Mar 19, 2013
5,278
0
17,790


Releasing a new kernel for supporting a new architecture is not the same than releasing fixes to patch the unfixable.

Therefore the unix/linux community had absolutely no problem with AMD, but when Microsoft has serious problems and released buggy hotfixes to fix that bad scheduler, the problem is... AMD.

According to you, the well-known fact that AMD chips work better under linux has nothing to do with Microsoft releasing mediocre products or only caring about Intel (Wintel).

DDR3 typical latencies 10-15 ns. GDDR5 typical latencies 10-12 ns. Myth: GDDR5 has very poor latencies...

This myth seems to have been initiated by Intel fanboys when Sony presented its AMD-powered console. You can find lots of posters in forums and blogs claiming that "PC is better", that GDDR5 is not used in gaming PCs (aka they did mean Wintel PCs) because drops CPU performance.

So many AMD hate around...
 

blackkstar

Honorable
Sep 30, 2012
468
0
10,780


It isn't very hard to figure out that this is Microsoft's fault. I don't know why you can't get people to see this.

AMD submitted changes for their CPUs to GCC and everything, and they showed up, just fine.

AMD then submits changes to Microsoft and it ends up a disaster.

Had this been AMD's fault, the Linux versions would be just as bad. AMD told MS what needed to be done, and MS half-assed it because they weren't going to make money off of an AMD patch. "Hey we improved performance for AMD FX CPUS!" isn't going to change someone's mind and make them go, "well, I was thinking about buying Windows but now that there's this patch, I'm definitely buying Windows!"

People need to open up to the idea of looking at things from a company's perspective. If you were MS and AMD showed up and said "please help our CPUs run faster by changing your kernel" you'd probably look at the folks you'd have to pay to make it happen, publishing the patch, and everything and then not caring at all about actually doing it.

And then compare to Linux where FX runs absolutely great (I'm on Gentoo right now, actually) where AMD pretty much posted patches and they got added, and it's not hard to see who is at fault.

The company who only cares about profits getting nagged to change their kernel for something that would not make the money and the patch turned out bad, while the open source community who doesn't care about profits had things turn out well.

 

jdwii

Splendid


Sorry to tell you but these hotfixes only improved performance by 3-5% not nearly enough to make bulldozer relevant in the high performance field(horrible design to begin with Latency is most important in a CPU and the module design raises latency)

That is why this design was originally shut down many engineers at Amd didn't want this to happen. Amd advertises Serial performance for their APU's but they don't offer the best of 2 worlds this design was a bad choice.
 

Only if someone doesn't have an in depth understanding of transistors and how processing is done. There is technically no difference between a "CPU" and a "GPU", they are both integrated electronics that use binary math to perform computations. The term "GPU" was coined by Nvidia when they integrated transition, lighting and geometry capabilities into their frame buffers. It was a marketing term used to differentiate their advanced processor from the competitions.

The concept of processing and processors needs to be fully understood before we can go any further. Your pocket calculator has a processor in it, its very basic with a limited range of instructions it can accept, yet it is still a processor. The processors used on modern graphics cards are just vector processors vs scalar processors that do integer work.

http://en.wikipedia.org/wiki/Vector_processor

The concept of using vector processors to do work has been around for decades. Using a combination of scalar and vector processors has been around just as long, early super computers used to do this.

What AMD is doing is just codifying a methodology for multiple non-uniform processors to communicate with each other. This is similar to what Intel did with the MMX and later SSE extensions, hey added it to the x86 ISA rather then make it an open extendable standard. Technically this is already happening as the "FPU" in modern CPU's is actually a separate co-processor that's operates in the same memory space. So now you will have three or more processors working together, the CPU, FPU (SIMD engine) and "other" (no official industry name has really risen up yet) and exchanging data / operations.

This is where I take issue
That is my point. Why would AMD waste die space, TDP, and money with extra SR cores for parallel workloads, when GCN CUs are much much faster at parallel workloads?

[strike]8 SR cores[/strike] --> 4 SR cores + 8 GCN-CU cores

That is incorrect. Code designed for / compiled on x86 will not auto-magically run on non-x86 processor. Those 8 SR "cores" would actually be faster with that workload. It would require a separate workload to be designed / compiled to get any advantage at all out of the additional co-processors.

This is one of the reasons I feel everyone should learn some ASM, it's really an eye opened to exactly how processors work and interact with everything else.

Some info about the above

http://en.wikipedia.org/wiki/Heterogeneous_computing
http://en.wikipedia.org/wiki/Asymmetric_multiprocessing

 

juanrga

Distinguished
BANNED
Mar 19, 2013
5,278
0
17,790


Agree. In fact, Microsoft had to fix its fix!



Bulldozer was not the brightest design made by AMD, but the hotfixes by Microsoft were very far from optimal. My position is well summarized in the next:

After their first attempt to improve Bulldozer performance through a Windows 7 scheduler was quickly pulled due to performance issues, Microsoft is yet again trying by introducing a new core scheduler update.

As we reported in our original Bulldozer review, AMD made it abundantly apparent that Bulldozer couldn’t realize its full potential under a Windows 7 environment. Their reasoning behind these claims was directed towards Windows 7′s core scheduling which just couldn’t take advantage of the new, advanced architecture. Windows 8 is supposed to do away with many of these issues through the introduction of a new scheduler but Microsoft has now rolled out another update (the first one caused performance issues and was pulled) which is supposed to grant the Bulldozer architecture additional performance improvements under the their current operating system.

[...]

Some of you may remember that AMD FX processors use a unique dual-core module architecture codenamed “Bulldozer”, which current versions of Windows® 7 were not specifically architected to utilize. In essence, for those with an AMD FX-8150 Processor, for example, Windows 7 sees the eight available cores and randomly assigns threads to them.

In initial testing of the upcoming Windows 8 operating system, we’ve seen performance improvements of up to 10% in some applications, when compared to Windows 7. This is because the system correctly recognizes the AMD FX processor architecture and cores. Thanks to close collaboration between Microsoft and AMD, Microsoft recently completed back-porting some of the Windows 8 scheduler code for AMD FX processors into a hotfix for Windows 7.

http://www.hardwarecanucks.com/news/cpu/microsoft-tries-again-second-win-7-bulldozer-hotfix-now-available/

Again: there was no problems in other operative systems.
 
I see the same stuff over and over and its giving me a headache. Multi-processing, Hetrogeneous computing, OpenCL while they all share commonality they are all by and large antiquated and the reason why they are still used is because hardware and software is most optimised for it. It is also naive to accept that this standard will be the same forever as we know it is antiquated and needs more efficient methodologies cue the issue here, yes the coders here know what they learned at universities but that is all they really know, then there are those that are accredited with changing the landscape of what is known. AMD would not commit to HSA like the others if it was going to end up in the same result. I won't comment on the intricacies as I leave that to those who are in the know of such but nothing is ever fixed or static and is there just a incling of the thought that maybe there are those that actually do know how to more streamline multi-processing, Heterogeneous compute methods to actually be notably faster than the standards of today? Is there any possibility that maybe someone has figured it out and probably not by mere coincidence but years of research? Or do we exist in this bubble that single thread legacy compute standards and all it entails is going to be around until man kills himself or some other vis major ends life as we know it?
 

juanrga

Distinguished
BANNED
Mar 19, 2013
5,278
0
17,790


Exactly. I don't have a deep understanding. AMD doesn't have a deep understanding. The people who made the Titan Supercomputer doesn't have a deep understanding. This is the reason why we continue to use the terms CPU and GPU. Moreover, I don't know the history behind VPU and GPU terms.

Now that you made your point, let me say that I will continue using the terms CPU and GPU, and others related such as APU, GPGPU... like I have been using up to now.



I am not saying that the GPU will run code compiled for the CPU. You already posted this kind of misunderstanding before. I already corrected you. I already mentioned you that the GPU in a HSA APU will be used for compute only with "HSA enabled software", still you insist on misunderstanding me...
 
The only thing really that is allowing AMD to keep it's head above water is OpenGL. OpenGL will run slightly better than Direct X, unless you get a higher end Nvidia card (770, 780, or Titan) which will boost the performance greatly of Direct X. The thing that I found the least realistic about games, is the lighting. In almost every game I play, the lighting is very bright compared to reality, no matter what settings you play at. DirectX does a better job at getting darker shades and colors, but at the same time getting better dynamic lighting results so the distortion between light and dark is better, as well as better texture results.

Then again, it is also the problem of how good the user uses Direct X. If it is executed correctly, it will look much more fantastic and beautiful than OpenGL, whilst keeping up with the frame rates of it.

AMD has the performance gain of OpenGL, however, DirectX does look better, and is easier to code for. If they make a more optimized or shorter code for DirectX, it is bound to be the same if not better quality, and run faster and better. CUDA is what helps boost this Direct X performance. Everyone says "it's a marketing gimmick" but its really the architecture of it that gives it advantages for PhysX and Direct X. CUDA is the name given for the architecture units that run the engines. This is literally no different than what AMD uses to power OpenGL better, except they didn't specifically name it.

Uh, no. Graphically, DX and OGL have the capabilities to output the same, exact, picture. OGL is a PITA to code with though. In my opinion, OGL as is currently designed is functionally obsolete, and the API needs a total re-write. But Direct X isn't graphically "superior", just easier to code with.
 
Releasing a new kernel for supporting a new architecture is not the same than releasing fixes to patch the unfixable.

Therefore the unix/linux community had absolutely no problem with AMD, but when Microsoft has serious problems and released buggy hotfixes to fix that bad scheduler, the problem is... AMD.

According to you, the well-known fact that AMD chips work better under linux has nothing to do with Microsoft releasing mediocre products or only caring about Intel (Wintel).

To be fair, its worth noting on of the major contributors to the Linux kernel is: AMD. I would hope their CPU had day 1 support, given AMD was the entity that released the patches to support BD. Also worth noting that GCC already had BD optimizations prior to the BD launch, where MSVC didn't. Finally, you have to note the frequent release of updates to the kernel (compared to Windows 2-4 year lifecycle) favors getting more out of architectures over time. I can say BD is probably 20%+ faster now then it was on release for Linux.

Also, funny how AMD did about the same in Linux/Windows at release:

http://www.phoronix.com/scan.php?page=article&item=amd_fx8150_bulldozer&num=6

Looses single threaded, competitive with Nahalem multithreaded, typically looses to SB. Same as Windows basically.

DDR3 typical latencies 10-15 ns. GDDR5 typical latencies 10-12 ns. Myth: GDDR5 has very poor latencies...

This myth seems to have been initiated by Intel fanboys when Sony presented its AMD-powered console. You can find lots of posters in forums and blogs claiming that "PC is better", that GDDR5 is not used in gaming PCs (aka they did mean Wintel PCs) because drops CPU performance.

Funny, I've seen much higher latencies for GDDR5. We'll see at release, but I'd expect a subset of workloads to show a decline in performance.

----------------

Also, just a little historical background, aside from being one of the 2-3 people here who nailed BD's performance prior to release, I'm also the guy who's long held X86 as an arch is simply non-competitive in the low-power (mobile) market, and that Intel's Larabee design (or whatever they are calling it theses days) had no chance of being competitive. I call it the way I see it. And right now, the way I see it, AMD followed the wrong CPU architecture, and that's the way I called it two years ago. Heck, given the move to Trinity/HSA, having lots of cores as the expense of per-core performance makes even LESS sense, given that type of work is best left offloaded to the APU anyways.

The best CPU company right now? Qualcomm. Aside from the fact they are now worth more then Intel, they have the best chance of having constant success in mobile, the one market that is growing at the moment. Sure, they don't have the best performance, but in terms of profit potential, Qualcomm is looking the best right now.
 


Yes, but its easier to talk "CPU" and "GPU" due to historical context, or in otherwords, the types of loads both are designed for. But yeah, language is going to be a problem going forward.

And yes, everyone should learn at least some ASM. You really can't talk about code optimization until you have an application to what the code actually DOES internally.
 

juanrga

Distinguished
BANNED
Mar 19, 2013
5,278
0
17,790


What is funny is that Microsoft is one of the top contributors to linux kernel

http://arstechnica.com/business/2012/04/linux-kernel-in-2011-15-million-total-lines-of-code-and-microsoft-is-a-top-contributor/



But that review was not showing the real Bulldozer performance, unlike what you believe:

Linux users will be able to take full advantage of the Bulldozer architecture sooner than Microsoft Windows customers, which will primarily see the real potential when Windows 8 is released. In the Linux world, there's still some Bulldozer kernel work that's not yet merged and presumably more compiler/kernel optimizations coming, but we will hopefully see all of that merged and ready in time for next spring when Ubuntu 12.04 LTS, Fedora 17, and other Linux distributions are pushing out their new versions. If you are so inclined, you can always pull the patches yourself, tune your compiler options, and make other tweaks today to take greater advantage of these new AMD processors. The upcoming FX-8150 Linux articles have more revealing information.

The poor performance of Buldozer in many popular reviews is a consequence of the bad scheduler in Windows 7, which does not understand CMT.



I know the myth is strong.



In a perfect world where only technological stuff matters, Bulldozer is a bad architecture. But when one puts Bulldozer in an economic, legal, and managing context, then it looks very much as the only move that AMD could do.

In any case this is the past; upcoming Steamroller is very much Bulldozer made right.
 

griptwister

Distinguished
Oct 7, 2012
1,437
0
19,460
Anyone else notice Newegg raised the prices on some of their Intel and Nvidia products today? The GTX 760s are priced about $50 more than normal, and a few of the z87s are about $20-30 more expensive... I'm just hoping their just trying to generate revenue before AMD releases their better cards.
 

8350rocks

Distinguished


They could be marking them up to mark them back down "on sale" as well...

EDIT: The downward spiral for M$ begins: http://www.tomshardware.com/news/surface-rt-lawsuit-class-action-shareholders-misleading,23875.html
-----------------------------------
AMD market share up 10% Q1 to Q2, Intel up 6%, Nvidia down 5%:
http://www.zdnet.com/pc-shipments-still-dismal-but-q2-gives-amd-and-intel-still-up-and-nvidia-down-analyst-7000019404/

It appears GF is already ramping up 28nm FD-SOI...power consumption drops and performance increases vs. 32nm HKMG PD-SOI.

slide8_no-NDA-short-Planar-28nm-FD-SOI-Technology-610x456.jpg


The interesting part of this graphic, is that it appears they expect to be on track for 20nm FD-SOI in 2014, and 14nm FD-SOI in 2015, though the next shrink looks to be 10nm and it's put out another 1.5-2 years after 14nm.

I think it's good that GF is licensing the ability to use the Silicon process from STMicro in France making all the FD-SOI advancements...

If this holds up, they will perhaps be able to hit the targets for their schedules. Which would mean we could easily see excavator sneak in at the end of next year as a paper launch on 20nm FD-SOI.

Given that Intel will not be shrinking to 14nm FinFET until 2015...AMD may actually gain a process advantage over Intel by that time, as 14nm FD-SOI > 14nm bulk FinFET in terms of both power consumption and performance. Additionally, based on the projections of GF, the 14nm bulk FinFET may actually mean the die is more costly for Intel.

If that's the case, then Intel may be considering retooling their complex process to move to FD-SOI at that point, however, while die costs would shrink some, the costs to change their fab around for SOI would likely offset much of the money they would save in the die costs.

I am curious to see how good Kaveri is, if it's strikingly improved over PD, then the shrink to 20nm may mean AMD could actually be right in the thick of things when excavator launches.

Interesting times to be sure...
-----------------------------------
Here is an interesting read on the SOI Consortium study...the link is to part 3 of the review, but the links to the first 2 parts are at the bottom:

http://www.advancedsubstratenews.com/2012/02/fd-soi-a-look-at-recent-consortium-resultspart-3-of-3-20nm-fd-soi-comes-out-way-ahead/
-----------------------------------
"Hawaii" GPU architecture to be unveiled in Hawaii in Sept:
http://semiaccurate.com/2013/08/07/amd-to-launch-hawaii-in-hawaii/

The most interesting bit in the read is how abysmal Nvidia is actually doing. Their roadmap has literally nothing to compete with AMDs new cards until late next year at the earliest. The even cancelled the Maxwell update that was supposed to come in the Spring.

They are literally floundering like a beached whale.
-----------------------------------
M$, Dell, Blackberry tied together? Maybe...

http://semiaccurate.com/2013/08/12/dell-needs-blackberry-microsoft-needs-dell/
 

jdwii

Splendid


I find this statement to be fallacious, data proves this was a bad choice over cost and performance. Maybe steamroller will finally make Amd some money instead of charging less for more transistors.
 

jdwii

Splendid
8350rocks, that chart basically says Amd can start production on steamroller using 28nm today and we could have them around 3 months or so, i hope Amd doesn't wait for 20nm we know GF will delay that design.
 


Err, the GTX 760 still seems to be around $260 You may be confusing the 4GBs and 2GBs. The 7900 series has done well, after all, it has been competing for 3 card generations now. History will repeat itself, 9970 to 780 as 7970 is to 580, which the cheaper 7950 defeated.
 

8350rocks

Distinguished


28nm production is ramping up, that's why Kaveri is taking just a tad longer...AMD couldn't resist the power saving features and increased performance...especially considering reduced costs per die.

20nm is going to be excavator. If they keep that on track, we could see excavator as early as Q4 next year.

I expect that once the 28nm FD-SOI process is mature and yields are good enough, we will see the return of the FX series or a rename to Phenom III with a SR product featuring 4/6/8 cores.

AMD likely is not committing to it publicly as they certainly don't want the press to be able to rip them if they cannot deliver on time. Especially considering the positive momentum that they have built as of late.
 


Perhaps AMD Phenom FX™ I agree that once the FD-SOI 28nm process matures, we will likely see *Insert Vishera successor's name here* being released.
 

8350rocks

Distinguished


Yeah...HD 7990 @ $699 + 3 free games worth $150 = HD 7990 for $549...who wouldn't buy that if they were looking at anything even remotely close...

EDIT: Sapphire Atomic HD 7990 inbound too: http://www.techpowerup.com/187162/sapphire-radeon-hd-7990-atomic-pcb-and-cooling-solution-pictured.html
 


7990 has\will continue to have 8 free games.
mother_of_god_in_hd_by_lemmino-d64ndao.png


http://www.anandtech.com/show/7218/amd-announces-never-settle-forever-bundle
www.newegg.com/Product/Product.aspx?Item=N82E16814202036
 

GOM3RPLY3R

Honorable
Mar 16, 2013
658
0
11,010


Excuse me as I flip my table:

(╯°□°)╯︵ ┻━┻

*puts it back* ┬──┬◡ノ(° -°ノ)
.
.
.
*throws Q6660* (╯°Д°)╯︵/(.□ . \)
 


OW! You mad that you don't have an excuse to go 780 SLI anymore :p?
 
Status
Not open for further replies.