hafijur :
Low end to what what a 22nm cpu of ivy bridge is capable of. 45w is basically how much the i7 3770k consumes in electricity less then probably an e8400 core 2 duo system. Anyway its got high end performance but intel could easily release cpus 50% faster with 6 cores as the mainstream cpus if they wanted to.
To put this into perspective an amd a6 apu cpu test will take similar power consumption to an i7 3770k and we all know which cpu will win.
Intel really is probably near the top of what they can do with desktop performance with the i7-3770K and i7-4770K. The top Turbo speeds of those i7s are 3.9 GHz and overclockers are topping out somewhere around 500 MHz faster is all. That's a fuzz over 10% is all, roughly how much you could overclock a top-end chip before the early 2000s when chips' top speeds were determined by the clock speed potential of the process/arch rather than thermal reasons. I think that is
exactly what is happening today with Intel.
Power consumption and clock speeds are determined at the architecture, process ,and platform level today rather than formerly pretty much just the architecture level. Current manufacturing processes have gotten small enough that gate leakage can be a problem and really drives up power consumption. This problem just gets worse with each gate shrink and Intel is one shrink ahead of AMD. However, leaky gates have a low capacitance and clock to high heaven (witness the Phenom II X4 TWKR.) If you want lower power consumption, you give up clock speed. Intel obviously optimized the 22 nm process for low power consumption as they believe mobile, especially low-TDP mobile like tablets and "ultrabooks" are the future. As a result, maximum clock speeds are quite a bit lower than with the older, less-power-optimized 32 nm process.
Intel also has chosen to go for minimum power consumption with the chip and platform arch as well. All of us oldtime overclockers around before bus locks (hell, or
multipliers for that matter) know that the "non-CPU" parts of the system generally don't like to overclock nearly as much as the CPU. Intel has rolled the entire northbridge, most of the southbridge, and even the VRMs onboard Haswell in an effort to save power. There are bus locks to keep the "non-core" parts at normal speed (which is why you can only overclock with a K or X series unit). But these parts simply being on the chip and designed to run in a very narrow and much lower frequency range has an effect on the complexity and floorplan of the chip. AMD on the other hand uses the same platform architecture in the FXes that they used back in 2003 with the very first Athlon 64 FX. The only thing on-die besides the caches is the IMC. No VRMs, no GPU, no PCIe controller, no southbridge. The layout of the chip can be much more conducive for maximum performance.
So in short, Intel could very well make a 22 nm chip that burned 220 watts and it could be much faster than what they are currently offering. They just can't do it with anything resembling their current 22 nm process or platforms.
I can see in 2015 amd will be releasing 220w tdp chips to compete with intels 35w tdp chips thats how the two companies are going, amd crank the tdp up, intel can we go even lower and start focusing on arm cpu power envelope.
I don't really care about clock speed as it could be like the pentium m 1.3ghz that destroyed any p4 really out then. The funny thing is the intel pentium m cpu was the best out on the market then the athlon cpus then the crap p4 cpus. The pentium m cpus were like 6x better performance per watt then the p4 cpus.
The PIII Tualatin and and the Pentium Ms were good chips but they weren't faster than the competing desktop chips. First of all, the Athlon XP was roughly as fast per clock as the Tualatin, considerably better in FPU-heavy tasks (any K7/K8's FPU was a ton better than any P6/P6+ chip), and also clocked a lot higher. Ditto with the A64 vs. any Pentium M- roughly as fast per clock except a lot faster in FPU tasks, but clocks a lot higher. And that's only in 32 bit mode, the A64 was considerably faster running 64 bit code per clock than the 32-bit-only P-M with similar 32-bit code. The PIIIs and P-Ms were faster per clock than the P4 but the P4 Northwood and on were simply clocked so much higher that they ended up being considerably faster at the end of the day. The short-pipeline Tualatin and P-Ms didn't overclock that well either- you could get to around 1.8 GHz on a good 1.3-1.4 Tualatin and around 2.5 GHz on a P-M Dothan. That didn't hold a candle to a 4 GHz P4 or close to 3 GHz A64.
Anyway amd piledriver currently is like an intel p4 netburst cpu, very inefficient needs a lot more power to compete compared to a competitor cpu. i am prettyu sure the athlon 64 took like 70w less then a p4 single core cpu and basically outperformed it. Intel cpus currently are trying to outperform amd chips taking 90-100w less on peak load. Thats what innovation or amazing cpus are pushing the boundaries of moores law.
Thats why I hope steamroller gets at least 60% better performance per watt then the current piledriver cpu. I find it hard to believe how they can go really backwards on new 32nm process while intel made a gigantic leap on 32nm with sandy bridge.
Piledriver is not like a Netburst CPU. Bulldozer/Piledriver is actually much closer to something like the UltraSPARC T series in theory as it's aim was to maximize core count by sharing of some formerly unshared core resources. Netburst's goal was to try to increase performance by massively goosing clock speed and it did it by having a very long for the time period (and in Prescott, very long period) pipeline. The pipeline of BD/PD isn't that much longer than Haswell's. Clock speeds with the exception of the limited-quantity FX-9xxx series aren't that much higher than that of the Phenom IIs. Remember those got well into the high 3 GHz range and BD/PD are on a new, smaller process that should allow for better clocks. Bulldozer's problem was that it erred on the side of multithreaded > single-threaded performance being a priority and shared a little too much between cores. At least AMD realized this and Steamroller is "un-sharing" some of the resources to tip the balance back the other way some. If they were like Intel and Netburst, they'd double down and hang four int cores off of one decoder and FPU like Intel stretching Northwood's pipeline to a ridiculous 31 stages in Prescott.
If you are looking at maximum performance per watt, AMD's chip for that is not BD/PD/Steamroller but Jaguar. You don't have to design your top-line chip for minimum power consumption just to get a decently-performing low-power chip when you actually have a decent low-power chip design to begin with.
This is just hilarious. Intel main aim is innovation like low power cpu for ultrabook. Heck they even funded a lot of money into this. AMD just releasing junk on 32nm barely any better then the 45nm phenom cpu and worse then the old phenom 2 series on a lot of tasks like gflops on intel burn test I believe.
The "ultrabook" is a joke. It's essentially a non-Apple version of the MacBook Air and saddled with the same limitations- very high price, mediocre performance, and few expansion ports. The MacBook Air sells slowly and that should have given Intel a clue as to the "ultrabook" sales prospects. But I suppose the appeal of something that could try to push ASPs way up and make $300+ CPUs "make sense" was too much for Intel to resist. Meanwhile, people are buying boatloads of cheap madeinchina convertible tablets running non-x86 cell phone CPUs and costing 20% of what an "ultrabook" costs.
AMD's 32 nm is a better process than their 45 nm one. You seem to be very concerned with power consumption for mobile CPUs. The latest 32 nm APUs do a whole lot better for power consumption and performance than the 45 nm Champlain Phenom IIs ever did. Their peak GFLOPs are actually quite a bit higher as they have a built-in GPU which has a pretty high FP throughput, particularly in a purely theoretical benchmark like a GFLOPs benchmark. BTW, Intel Burn Test isn't a benchmark, it's a system validation tool to make sure the system has sufficient cooling to not fail when severely stressed.
AMD are a good 2-3 years behind. If it weren't for intel then we would not have efficient computers and haswell is the next big step. AMD bulldozer series cpu was so cap on power they had to delay it. It still is crap.
We would have efficient computers if Intel weren't around because the big driver of very low-power computing isn't Intel, it's TI, Samsung, Qualcomm, et. al with their ARM-based chips in tablets and such. Intel is a follower, not a leader, as evidenced by their laughably bad attempt at a low-power chip in the Atom and the creation of not one but two new categories of computers to try to shoehorn their chips into. The netbook is already dead and the "ultrabook" soon will be.
AMD had to delay Bulldozer as their 32 nm fab process was not ready, not because of power or the chip being "crap." The best example of a chip being delayed because it stank was the first Itanium. Boy did that ever suck and no, it did not get any better in the three years it sat around during the "we'll release it when we get our compiler working well, which will be Real Soon Now" phase.
Think of using amd as a server for a company or intel. Lets say amd systems for same performance total take 10000w. Intel for same performance take 4000-5000w thats a big difference isn't it."
First of all, the percentage difference in the amount of total power an Opteron vs. a Xeon draws is fairly small. The E7 Xeons actually are much worse than the Opterons in total power draw because the E7s are old Westmeres using FB-DIMM MTH type devices on their motherboards which just chew through the power. Most servers also have a lot of RAM and disks that draw a considerable amount of power and would be similar between the two. The cost difference between Xeon and Opteron servers are quite large, especially for 4P units. If you wanted to make a cost of ownership argument for Xeons, you won't make it on power or performance. Licensing with ridiculously expensive per-core fees like Oracle's would be a better case to make as you want to have as few cores as possible to avoid getting reamed as badly by using Xeons with identical performance using fewer cores. But then all that really says is that some vendors have stupid licensing schemes and massively overcharge.
8350rocks :
AMD may very well be ahead after Steamroller...it's that dramatic an improvement.
I'll believe it when I see it. Steamroller should improve single-thread performance somewhat but Bulldozer/Piledriver isn't bottlenecked when it's only running one thread in total. It's when it runs two threads on one module where it gets bottlenecked and this is where Steamroller will help a lot. But you won't see that unless you are using decently multithreaded tasks anyway (which BD/PD already do well at) or are using a completely brain-dead OS thread scheduler like Windows'.