The fact that AMD would follow up their dual-die, 12-core Magny-Cours with a dual-die, 16-core Opteron was a no-brainer.
Perhaps the biggest concern here is that AMD's targetted price-point competition, the Xeon 5670, may be beaten thoroughly on performance, but a lot of people look just to the superficial per-package TDP: and Sandy Bridge's low TDP DOES make it look pretty attractive to professionals, even if, at face value, the performance-per-watt may be lower; after all, that PPW only considers the cost of the CPU itself; the cost in money, power, and space for cooling may balance that out, and is harder to pin down, so many professionals may, in fact, just go with the lower TDP option in hopes that it will yield the best deal in the long run.
As for the debate of cores vs. modules... Each of the 16 "cores," in an Interlagos *IS* as capable, on paper, as a single core on a Sandy Bridge CPU. Each can execute its own instructions and operations, independent of the other. The issue here mostly stems from AMD's chosen implementation of AVX; like Sandy Bridge, Bulldozer cannot fully retire a AVX instruction in a single cycle with just a single core. However, while Intel's FPUs are more isolated, AMD's solution was to join the FPUs of both cores in a "module" at the hip; yes, the FPUs may be shared, but there's twice as many per module as a Phenom II has.
It's not the number that's the problem, but rather collisions; so far, with the architecture being alien to all operating systems out there, there's no system in place to ensure that, with both cores in heavy use, that both get smooth access to the FPU resources. Chances are that software won't ever fix this, and it's a hardware issue: then it's AMD's fault for not including better thread-management circuitry there. At the very least, the most likely short-term solution is an OS that adjusts its thread handling to act much like it does with an Intel Hyper-Threading CPU: fill all the even-numbered threads FIRST, then do the odd-numbered ones. That way, the most intensive threads won't have to clash with each other, sapping performance.
As for the desktop market... Barring a significant adjustment to compensate for this threading weakness, the per-clock performance of Bulldozer is not going to be able to match up to Intel's offerings. On the server side, with embarrassingly parallel workloads, this can be compensated for by brute force; simply upping the number of total cores, and damn the TDP. It may or may not work, though I can't help but recall if you replace "up the number of cores and damn the TDP" with "up the clock speed and damn the TDP," you have Intel's entire design philosophy from 2001-2003, and we know how well THAT turned out. Plus, of course, given the decidedly fewer-thread nature of home computing & gaming, this strategy won't impact outside the server market.
Ironically, though, if AMD either wants to grab more market share to displace the 2500K & 2600K et. al, or simply grab a higher slice of the "extreme enthusiast" crowd, they could opt to tap into the apparent clock speed headroom of Bulldozer, and get over their fear (a fear shared by Intel, admittedly) of actually binning a CPU at the dreaded 4.0 GHz mark or above. AMD claimed the 1 GHz mark, and Intel quickly beat them to the 2 GHz and 3 GHz ones... But right now AMD's best-equipped to hit 4.0 GHz first.
[citation][nom]otacon72[/nom]Um the Xeon X5670 has 6 cores. Intel should make an X5670 with 16 cores and blow the Opteron 6276 out of the water just to shut AMD up.[/citation]
Keep in mind that the X5670 MSRPs for $1440US, (which, given the history of the 12-core Magny Cours that preceded the Interlagos, is likely around the 6200's MSRP) so the price might've factored in here; remember that the "performance crown" is meaningless outside of the gaming sector, and "performance-per-price/watt" is king: after all, with the enterprise market, if one chip isn't powerful enough, your always-there solution is to just buy more.
[citation][nom]de5_roy[/nom]would love to see some server benches against xeons... and some real world general purpose benches e.g. gaming and transcoding on the side.[/citation]
Gaming isn't exactly a real-world benchmark for a server. LINPACK is, in fact, actually more realistic; after all, that's the benchmark used for measuring the world's most powerful supercomputers.
[citation][nom]shqtth[/nom]fpu wa snever designed to take orders from two cpus at once. two threads at once on an fpu is hard due to the stack structure of the fpu and the register structure of 3dnow/sse2.[/citation]
Actually, as I recall, Bulldozer actually removes support for 3DNow!... Ironically, perhaps retention of it would've allowed for a potential solution to FPU resource collisions... But requiring using an entirely extra extension just to compensate for that is just plain unfeasible anyway.
[citation][nom]hetneo[/nom]This is such a huge blunder on AMD's part. This is supposed to be their flagship server CPU but they stacked it against almost 2 years old[/citation]
Again, keep in mind that they're likely going for a price-point comparison. I think that, at this point, AMD's essentially decided to ditch the "go for the crown jewel" strategy; after all, it's mostly a marketing gimmick, and never is the part that rakes in the sales. And again, in the server marketplace, IT professionals don't need the highest-end chip to get the most FPS: if they need more power for their server farm, they just buy MORE CPUs; hence 2x 2.0 GHz CPUs for $200US each is EASILY a better deal than 1x 4.0 GHz CPU for $1,000US, just to give an example.
In this case, to be more direct, while Intel's current server flagship, the 10-core 8870, (12-core Sandy-Bridg-based Xeons aren't out 'till 2012) probably would best the Opteron 6200, it also MSRPs for $4,616US. And actually, given that the net theoretical throughput (instructions & ops) is only 36.5% greater than the 5670 given it's reduced clock speed, (which is slightly below the 36.8% increase in TDP) there's a chance that the 6200 would still come out on top given the margins of victory AMD's been boasting.
Of course, that just makes me wonder what AMD was thinking in picking the chip to compare it too; they could've retained a "it's more powerful" argument AND tacked on a "it's much lower in price" and a "it's lower in TDP" claim as well.