Cazalan :
juanrga :
Huumm. This looks all suspicious. Then a simple search reveals that this technical paper is being cited in news sites that claim that Intel will compete with ARM on mobiles. Huuum more suspicion. Then we take a look to the conference website and we found Intel name among 'sponsors'
http://www.carch.ac.cn/~hpca19
Then we search the main author of the article and... surprise surprise he has an Intel folk as collaborator and several Intel students
http://research.cs.wisc.edu/vertical/wiki/index.php/Main/People
You're confusing an IEEE conference sponsor with the actual people who wrote that particular paper, which largely have nothing to do with the conference other than submitting their paper for review and presenting it. The conference doesn't pay people or provide funding to write papers.
There is a huge difference. A conference sponsor means they get their names on the SWAG (stuff we all get, t-shirts, coffee mugs, and crap like that) and signs and other hand outs that are given to the people who PAY to go to the conference.
There is an Intel person listed as a collaborator for the research group, but there is also a Broadcomm person. Broadcomm uses ARM cores in their products not x86 cores.
"Support for this research was provided by NSF grants CCF-0845751, CCF-0917238, and CNS-0917213,
and the Cisco Systems Distinguished Graduate Fellowship."
Cisco uses X86 and ARM products so your Intel bias assumption doesn't have much merit.
It's amazing how quick you are to cut down the paper but take marketing slides as gospel.
What is really amazing is how you have snipped relevant parts of my analysis of their paper, such as their scaling methodology and power consumption measurement favouring Intel.
I wrote 'sponsors' not sponsors.
As mentioned before, there are three (3) Intel persons in his research group (one collaborator plus two students), not only one (1).
No sure why you mention Broadcom and Cisco. Do they have a historical record of dishonest competition/cheating? Also, do you believe that because there is people from other companies the possibility of aggressive 'sponsoring' is eliminated? AMD, Nvidia, and VIA were members of SYSMARK consortium and still Intel cheated the SYSMARK benchmarks for artificially increasing the score of its products.
If that article that you mentioned was right and Intel (x86) is able to compete with ARM face to face, then Intel had not been caught using its ICC compiler to cheat the Antutu benchmarks this year.
The cheated benchmark gave ARM processors a mere 50% upgrade in performance whereas Intel chips were seeing increases of up to 292%.
Pay attention. The ICC biased CPU dispatcher gives Intel chips about a 15% advantage over the AMD chips. When fighting ARM, Intel had to give its chips up to ~10x more advantage.
http://www.androidauthority.com/analyst-says-intel-beating-arm-in-benchmarks-was-a-sham-243526/
http://www.eetimes.com/author.asp?section_id=36&doc_id=1318894
http://www.slashgear.com/intel-atom-z2580-antutu-benchmark-falls-20-following-revision-13290307/
http://seekingalpha.com/article/1550372-intel-and-antutu-mea-culpa
Cazalan :
juanrga :
Why did they choose Intel but not AMD chips? If they had tested Bulldozer FX instead of an SB i7, they had found poor efficiency and even poor performance in several tests. Why did they not test a modern ARM chip?
When it comes to conference papers they're often submitted a year in advance of the conference. And the work probably started a year or more before that, so you're looking at a 2+ year lag from start to presentation (Feb 2013).
So the question why they didn't test an A15 is because they simply weren't out yet. Apple was the first to ship a phone with one and that wasn't until September 2012, which is only 5 months before the paper was presented.
They're not doing speculation here. They have real world hardware they had to get in the lab, code to write, testing, getting results, taking measurements, getting feedback from advisers, peer reviews, etc.
They used what they had access to. The i7 Sandy is old as well. It came out in January 2011, which is around the time A9s were introduced to the market. I.e the first Tegra 2 phone in January 2011.
http://www.networkworld.com/news/2011/010511-nvidia-shows-off-lg-optimus.html
I have submitted conference and contest papers a pair of months in advance. However, there are other objections to your hypothesis. The first, that they cite references published so late as 2013. The second, that they mention the A15 and even give some numbers for it. The third, that they claim that they evaluated but rejected more modern SoCs such as Tegra 3 before starting the lab work.
Now explain us: how did they evaluate a 2012 product if their lab work was ready in early 2011 according to you?
The facts are:
- They rejected better ARM designs with improved performance and efficiency.
- They didn't test Bulldozer FX or similar ones but the best possible x86 chips.
- The scaling methodology and power measurement favoured Intel.
- Their compiler choices favoured Intel.
- «they believe the results would look much the same — in terms of the relationship between power and performance» when looking at newer ARM designs, including the 64-bit ARMv8 architecture.
- They have 3 Intel people in their small research group, but nobody from ARM or AMD.
All together implies bias pro Intel.
Cazalan :
juanrga :
Don't be fooled my friend. That is not for CPUs. I was discussing only CPUs.
That link is for Intel MIC project. The Xeon Phi coprocessor is somewhat based in x86. But it doesn't run the same x86 binaries that you can run in an i7-3770k for instance. For using the Phi you have to recompile the code at least and if you want to use all its potential you have to rewrite the code.
No one is making CPU only HPC from here on out. Even IBM systems are incorporating NVidia GPUs. BlueGene/Q is the only one that could be considered CPU only on the top500 list but it's floating point accelerator is essentially a GPU.
Your beloved project denver (Tegra 6) isn't getting it's massive performance gains from the ARM cores. The 100x performance increase is largely due to the Maxwell GPU cores that support CUDA, instead of their ancient GeForce ULP cores.
Recompiling for HPC environments is common place. They're all unique with various bandwidth, storage, cpu variations. Dig on the Phi all you want but it currently powers the #1 supercomputer in the world.
The same will have to be done for a supercomputer powered by Tegra 6. It will mostly be CUDA code with some ARM code for moving data around.
If you pay attention to the part of my message that _you snipped_. I already said you that CPU only HPCs are being abandoned. I was the first one in this forum who mentioned (it was many pages ago) how IBM joined to Nvidia to develop heterogeneous POWER8+CUDA supercomputers. You don't need to repeat what I already know.
I also said you that
ARM+CUDA or ARM+GCN will be the interesting combos.
What apparently you don't understand is that up to a 40% of the power consumption of heterogeneous supercomputers comes from the CPUs alone. Therefore it doesn't make many sense to improve the efficiency of the GPGPU by an order of magnitude without improving the efficiency of the CPUs first. That is why there are projects to replace x86 CPUs by ARM CPUs.
Nobody is saying you that Denver will be 100x faster than A9. But the A15 core is about a 70% faster than the previous A9 core. Nvidia claims that Denver will be faster than the last A57. Therefore, Denver being ~3x faster than A9 is a reasonable estimation.
My point was that once you recompile then the benefit of compatibility with legacy x86 binaries is lost. Sure that the Phi powers the #1 in top500, but not in green500, because that #1 is obtained by a brute force approach:
Looking at the latest data, what’s particularly astounding about Tianhe-2 is simply how large it is. Placing on the Top500 list requires both efficiency and brute force, and in the case of Tianhe-2 there’s an unprecedented amount of brute force in play. The official power consumption rating for Tianhe-2 is 17.8 megawatts, more than double Titan’s 8.2MW.
Take the old Titan, double it, and you recover the #1 (approx.). But as explained to you before, this pure brute approach doesn't work for exascale supercomputing. You cannot take a x86-based supercomputer and scale it 1000x times because you cannot generate the immense power needed for that beast. That is why HPC experts are developing an ARM supercomputer for exascale.
Also you seem to not understand that the performance of the Xeon Phi has very little to do with x86, but is directly related to an _extension_ of the x86 ISA which is exclusive for Phi. Follow my advice, don't be fooled by Intel marketing ;-)
I doubt any supercomputer will be powered by Tegra 6. Who said you that?
Cazalan :
juanrga :
Nvidia has already ported CUDA to ARM for instance. CUDA will be not used in servers.
Or course CUDA will be used in NVidia servers. That's how you program/execute Kepler/Maxwell cores.
https://developer.nvidia.com/content/cuda-arm-platforms-now-available
Read the link that you give about CUDA on ARM.
How many times the word "server" appears? Zero (0). How many times the word "supercomputer" and derivative (supercomputers, supercomputing, HPC) appears? Seven (7).