nVidia already supports DirectCompute and OpenCL on thier hardware. They have whole sections of thier website and documentation for that. You can already choose to use either of those, or the CUDA extensions for C++ or Fortran as well, on all nVidia CUDA enabled hardware. It's your choice which you prefer to use, and which is best suited for your application (and skills). Obviously a 4 or 6 core cpu will never give any significant performance compare to hundreds of cores and tons of dedicated memory on a specialized board, or even a consumer class GTX graphics board. But, as previous comments have said, cpu support gives access to developers that otherwise wouldn't be able to use the tech, and allows use in some hardware situations that, while not ideal, at least would enable it. Thier hardware is not exclusive to CUDA, and now CUDA won't be exclusive to thier hardware. A lot also depends on what type of application you're developing and what type of calculations are being performed. Like anything, different things are better suited. But, if you're working with the ideally suited types of software, there is no comparison in performance when designed well. There are many examples where the same type of application and calulations are compiled and run on different hardware and development languages and the difference in speeds is sometimes simply staggering. You just need to choose what best suits your needs.