SLI/Xfire Physics Vs. Ageia PhysX

Bluefinger

Distinguished
Mar 10, 2006
531
0
18,980
What are your personal thoughts on ATI's + Nvidia's solution to complicated in-game physics, compared to having a completely seperate processing unit devoted to real-time dynamic physics processing?

IMO, i think the "SLI-Physics" won't suffice in terms of being able to process physics. Since games are becoming more graphical and shader heavy, even SLI configurations would eventually struggle to render frames and process dynamic physics at the same time, where as having a seperate processing unit devoted for this purpose would obviously have its advantages. So, by keeping GPU's for rendering, and offloading physics from the CPU onto a PPU, in theory, the system would run more efficiently when presented with a physics-heavy game, since the only thing the CPU would have to do is process management, A.I., along with the rudimentary game engine processing.
 
Not sure how performance will play out SLI vs. PhysX, that remains to be seen after benchmarking. But I do appreciate NVidia's approach to this, their solution is basically a driver install with no additional hardware purchase for SLI users. By waiting this long after ATI announcing that were opening up the API, they neatly solve the chicken-and-egg dilemna; there's already an installed base of over 1 million SLI rigs out there. I predict that fact alone will attract developers to start using it, plus they're already familiar with NVidia coding. PhysX will be a new set of programming methods, and who's gonna speculatively invest in that when there's no boards out there in the wild? That's a big gamble and could mean a big hit to your R&D budget if those boards don't sell well. PhysX may perform better, but attracting delevopers to support it will be an uphill battle, and selling it will be hard since you'll need some big titles out there to support it and show it off in demos. Chicken-and-egg paradox.

I did notice today that the announced $10,000 monster rig from Dell is including the PhysX board:
http://www.tgdaily.com/2006/03/22/dell_xps600_renegade/
 
I think an independent processor will work better than a shared GPU. Also, a standalone processor will benefit people without SLI:

While the name "SLI Physics" indicates that two graphics processors are needed to run physics, Nvidia officials told TG Daily that owners of single-GPU systems will still see "some benefit," but couldn't provide more detail at this time. - article

Of course, the standalone processor will need to be fast enough to feed data to the GPU or at least synchronized with the GPU. I would expect lower latencies on the SLI version since the vid cards have their own interconnect and calculations are being done on the same processor/GPU.
 
I suspect that SLI Physics would act as the intermediate solution to complex physics in the mean time, since many games even today don't make full use of a dual-GPU setup. In terms of communication between devices, i would agree that a SLI-Physics config would have lower latencies, but in having lower latencies, it would become limited by the fact that it can't dedicate itself completely to processing dynamic physics, which is where the PhysX card would have the advantage. Also, the CPU would have to manage the communication between the PPU and GPU so heavy physics processing still puts pressure on CPU usuage, but not on the magnitude that it would have without either SLI Physics or a standalone PPU. I would reckon that the only way to really get the PPU solution going is by keeping the boards reasonably priced along with having a selection of games that would benefit loads by having the card. Personally, i can't wait to see some benchmarks or tech-demos to demonstrate its capabilities, because though the PPU would require a different code set than what nvidia are offering, the power of the card would certainly be enough to encourage game developers to at least experiement with the technology in their games.
 
I think we need to wait and see what ATI has going on. From an interview I read in cpu they're working on gpus as a co-processor for more then just physics, things like audio,cad and adding more to avivo. They also want to do it with any ATI gpu, so crossfire can use a x1600 and a x1900. The lower card does math and the higher renders. I read this in november so I may be off on this some but thats what I remember.
What I would like to see is the processes to work together and share the load or for developers to give the option like gl or dx. It seems nvidia rushed to be first on the block.
The ageia seems to be the better solution in the long run since it's a dedicated processor, if it can get noticed. With ATI and NVINIA trying to hold it down it will be a fight.
This is a link to a site about gpu's as coprocessors http://www.gpgpu.org/ it has some interesting info on it.
 
Well Ageia is getting some love theres already a few graphic engines that provide direct support for it. I.E. Unreal 3 engine being the most noted. Also the PS3 will be rocking some form of the Ageia PPU engine pre-loaded into one of the Cell's SPU's, not sure how thats all gonna work. But Ageia is definitely getting its product out there. In my opinion if BFG see's something in it to manufacture them, I'll trust them (even though there outragously overpriced) since there the former owners of the original SLi back in the 90's. But thats neither here nor there.

I'll fork up around $200 give or take $50 for an Ageia PPU over SLi Physics. One is essentially software based and the other is hardware, and hardware based processing is always better than software based.

On a side note: Doesnt it seem odd that nVidia released details about SLi Physics, right after launching Quad-SLi...
 
Concerning the PS3, it's getting the codebase loaded onto some of the threaded Cell CPU's, so from what i can imagine and speculate, the CPU's would be divided through threading on what core processes what task, so i guess some part of the processor will dedicated for physics processing... however, i am concerned whether a general cpu, even with multi threaded cpus, could cope with what Ageia reckons physics will become like.

As you can see, i'm really interested on in-game physics, partly because of what Half-life 2 managed to do with Havok. I also know Havok FX deals with using GPU's to process some physics, but specifically to do with visual physics effects, like explosions and the creation/movement of the debris from the explosion, so i reckon that software would take advantage of a SLI Physics configuration, or with whatever ATI offers up. Overall, i'll be keeping an eye on this, along with the whole Conroe/AM2 battle.
 
yes, interesting about the quadsli/sliphysics connection. things that make you go hmmm...

I also agree that ageia is getting plenty of exposure, and w/ native support in engines like unreal 3 "coding for it" is not an issue. Just a matter of how much hardware gets out there in how much time. w/ Dell offering their gaming rigs w/ them (and vodoo and others i assume) i think there is plenty of promise for ageia.

Personally I will wait and see what happens once the physX cards are @ retail. See what the industry does and what "real" benchmarks show for perf gains... then i will buy.
 
I think that the Ageia PhysX card will be the future, simply because of the psycology of enthusiasts. People are gonna look at the SLi and similar solutions and say, "Why is my video card trying to process physics? Wouldn't a dedicated physics card be better at that?" Whether it's true or not is besides the point. (Although I personally think that the physics card WILL end up being a better solution for high end gaming)

I have to wonder though, that with multicore CPU's coming out, what's to stop a developer from dedicating a core to computing physics? I mean, won't a dedicated physics card require a seperate thread to run correctly anyways? I'm probably getting a dual core on my next system (and soon enough, a quad core) Unless I'm convinced that these physics cards will have signifigantly different architecture that is optimized for physics, I'll probably just wait for games to be written for the CPU to handle the load.
 
I would reckon a multi-core CPU could handle some physics, or at least the management of a seperate PPU by concentrating a core or thread on the processing for that task. CPU's at the moment are powerful... but they lack the required architecture to help them cope with dynamic physics processing. However, what's stopping chip manufacturers from putting in maths/physics coprocessors inside their chip architectures in order to help them cope with such cpu-intensive tasks? Which makes me wonder... what will happen if someone manages to get a CPU with a maths/physics coprocessor and put that into a decent high-end system with a Ageia PhysX card.... :twisted:
 
NVidia's announcement was that they'll be opening up the GPU to doing alternate tasks, it just happened that the demo of it was a physics demo (I won't get into potential politics of that decision). Gee, you think it might be geared to help sales of SLI rigs? And to poke fun at ATI, where their announcement months ago of this is the usual ATI vaporware?

Using the vidcard wouldn't necessarily accelerate it beyond just having additional processor resources to work with, but with support for math functions that are present in 3D rendering, some of which are useful in physics calculations. almost like dedicating a CPU core to the task, but better since a GPU is far better at doing floating-point calculations.

I don't see the processor mfgrs wanting to put physics into an onboard coprocessor, 99% of their sales are corporate and non-gamer sales, and the R&D cost to modify an existing product line is staggering.

Dedicating a general-purpose processor core to doing physics wouldn't give much benefit, these specialized coprocessors have hardwired logic to accelerate some of the hairy math involved in calculating physics, like matrix transforms that take hundreds or thousands of CPU cycles to process on a general purpose CPU. (*Shudder* - I had to do matrix math in engineering college, ick) We're already beating the onboard math co-processors to death now and it's still not enough horsepower.
 
To me a Physics processor built from the ground up is better than a makeshift solution using a driver.

However, yes, Nvidia has an advantage because it has the developer support for its products, doesn't require a new code base, and theres already hardware out there.

But the Physx processor is gaining momentum. Personally I'll probably buy one and put it in my system.

With SLI-Physics I think it should be that one card renders the screen while the other is devoted to doing the physics calculations. That alone would speed things up because the rendering GPU wouldn't have to wait for the CPU to do the calculations and get the result back to the GPU (which might be what Nvidia does).

SLI definitely gained some more justification for its expense to me now. But if developers embrace a dedicated PPU then theres no need for SLI-Physics and one card will be fine still.
 
Dedicating a general-purpose processor core to doing physics wouldn't give much benefit, these specialized coprocessors have hardwired logic to accelerate some of the hairy math involved in calculating physics, like matrix transforms that take hundreds or thousands of CPU cycles to process on a general purpose CPU. (*Shudder* - I had to do matrix math in engineering college, ick) We're already beating the onboard math co-processors to death now and it's still not enough horsepower.

I see your point. Which is why i am slightly concerned about the PS3 in having the Ageia PhysX SDK codebase integrated into some of the cores in the Cell CPU. In theory, it should work, but when confronted with the "hairy maths" of complex dynamic physics, how will it cope in reality? The only way to find out is sit and watch....
 
Well, Ageia is an IP company, not a manufacturer, so I'm sure they'll figure out some way to maximize the potential of the Cell CPU.

I'm not sure of the capabilites of Cell's math copro, but since it's not saddled with legacy issues like I386 architecture, and it's designed to be part of a gaming system, I'm sure it's a lot more capable than I386, and that there's more hardwired support for advanced math functions.

Personally, I'd love to see PhysX take off and be supported. Hopefully they'll climb that steep slope.
 
Just a couple of thoughts...

1) a separate card is going to be cheaper than a full-blown graphics card;
2) a separate card will surely not consume as much power, nor generate as much heat, so will probably be either passively cooled or only require a small/slow fan. Quieter, in other words.

So, there's certainly no reason to get a second graphics card for processing physics, and if I already had an SLI config I'm obviously not too scared of spending money on bits of circuit board; I'm almost certainly going to want to get a separate card purely so my graphics cards can concentrate on what I bought them for.

Just what I came up with after 5 minute's thought, but I can't really see the argument for having your expensive graphics card process physics. As good as it may be for it, it's still not going to be as efficient as a PPU designed from the ground-up for the purpose. Obviously my thoughts about the power/heat are entirely conjecture, but I can't see the market for a dual-slot loud-as-hell physics card at all. :lol:
 
This looks like a good thing for the NB makers[ sis,via and the like] to license this and intergrate it in the NB[ really for AMD because there's no mem controller in it]. I could also see AMD adding something like a physic processor in the FX line since it's really only a gamers chip.
Is the new race in gpu's which one can do more nongraphics processing? From the GPGPU site it looks like this is something thats catching on.
 
Just a couple of thoughts...

1) a separate card is going to be cheaper than a full-blown graphics card;
2) a separate card will surely not consume as much power, nor generate as much heat, so will probably be either passively cooled or only require a small/slow fan. Quieter, in other words.

They're saying $100-$400 for these cards, so they're in the realm of a second vidcard.

The picture of the card at http://www.tgdaily.com/2006/03/22/ageia_physx/ shows a heatsink/fan on the board and a 4-pin Molex connector to the power supply, so it's going to be using a fair amount of power and generate some heat, but probably not as much as a high-end vidcard.

Going to a dedicated physics board will definitely be the better performer, but using unused vidcard power also has some attractiveness. I'll be watching for other areas of computing to invent interesting uses for this. Wouldn't accelerated DVD ripping/transcoding be nifty? Or HD/BluRay video ripping, which I'm sure will require a ton of horsepower once they crack the encryption within a week of its release :) Or SETI- or Folding@home processing...

Lots of possibilities.
 
Going to a dedicated physics board will definitely be the better performer, but using unused vidcard power also has some attractiveness. I'll be watching for other areas of computing to invent interesting uses for this. Wouldn't accelerated DVD ripping/transcoding be nifty? Or HD/BluRay video ripping, which I'm sure will require a ton of horsepower once they crack the encryption within a week of its release :) Or SETI- or Folding@home processing...

Lots of possibilities.

Using the PhysX chip for something other than physics processing... just as inventive as using the unused resources in a dual-GPU setup to process physics... oooo.... I'm starting to get lots of ideas on this one... :twisted:
 
Using the PhysX chip for something other than physics processing... just as inventive as using the unused resources in a dual-GPU setup to process physics... oooo.... I'm starting to get lots of ideas on this one... :twisted:

How about 3D rendering on the PhysX chip? 😀

I'll use it to run my Basic bubble sort faster......
 
Damn it! All these ideas!! I have decided I must get the card... along with the SDK to go with it... just so i can mess around with all that potential power...
 
How about an AOL accelerator? (sorry, I feel dirty just mentioning that cesspool)

Imagine if viruses or malware could get hold of this power!

Or real-time encryption of data for your hard drive, or better VPN connectivity.

Or making your Sims do stupid things even faster...

Oooh! Playing Solitaire with real physics applied to the cards!

But I digress.....
 
How about an AOL accelerator? (sorry, I feel dirty just mentioning that cesspool)

Shame on you! :evil: The mere mention of using specialised hardware for physics processing to try to improve the performance of such an abomination is blasphemy! You should scrub yourself thoroughly with a metal wire brush!
 
As for GPGPU and doing physics on GPUs, there are a lot of reasons why it will win out over AGEIA's hardware specific solution. There are two main issues to keep in mind on my points:

(1) Faster does not matter, just has to be fast enough. Physics is not an end-all be all task at this point. Devs are working within normal game models and have some limitations. While the fastest solution, in a perfect world, is best, the reality is each solution only has to be fast enough for the end product. PhysX does have limits (e.g. you are not going to get an ocean of complete fluid dynamics), so the question is can GPUs keep up with what devs are trying to do today?

(2) Game physics & Effects physics. HL2 had some game physics. Yet so far the games like UT2007 and City of Villans are talking about ONLY putting in effects physics. e.g. the UT example: Instead of 600 rocks falling off a cliff, instead 6,000. But it was noted this would NOT change gameplay. Ditto City of Villans. Blow up a news stand and paper goes everywhere. Here is the problem: GPUs already can do effects physics great. Go to ATI.com and look at the "ToyShop" demo -- there are a LOT of GPU based physics there. The 360 has an example in Kameo where it had 1M interactive particle bodies running in realtime. So until someone makes a game w/ gameplay physics (like dynamically destructable buildings) and it cannot be done on GPUs, devs are so far not doing anything that requires AGEIA. If the power goes untapped...

On to my points of why AGEIA's PhysX wont survive.

- PR and Marketing. No contest. ATI/NV have a ready made fan base.

- Established developer support. It is all about the games, namely getting a game that screams, "You MUST have AGEIA". But no developer can risk building around AGEIA due to install base issues, which leads to...

- Distribution Channels, board maker, and OEM support. This cannot be overlooked. NV/ATI have a decade or more of working supply lines and having great connection with board makers. AGEIA has an undersupported product that was delayed with 2 board makers getting on ship.

- Install base. SLI/Crossfire > PhysX. Developers will develop where the money is; developers will then code to the expectations of the least-common denominator.

- Open platform. There is competition on the GPUs (like MS unannounced Physics API and Havok) which means more developer choice. Thus far AGEIA has NOT gone out of their way to make their platform open (they say after release anyone can work on it but they are not making the runtimes and so forth open at this point)

- GPUs are more versatile. Every 3D game uses them. PhysX? Only games using Novodex will use it. GPUs are now also used in Vista and also can be used for GPGPU tasks. PhsyX? Nope.

- GPUs have 100% overlap in market. PPUs, on the other hand, will only work with a small percentage of games.

- Cost. $250-$300 (Alienware has it at $275) for a chip that only works in a few games is nuts. Return on investment is low. Would you rather have 1GPU+1PPU or 2GPUs, like 2x 7900GT? Which leads us to...

- SLI/Crossfire. Does a game use physics? Use 1 GPU for Physics, 1 GPU for 3D. No physics needed? Then use both GPUs for 3D! PPUs cannot do this, if the game does NOT use it, well, it sits there like a $250 paper weight.

- Old GPUs may in the future be left in your machine as a dedicated physics accelerator.

- ATI/NV may put smaller chips onboard as "dedicated" physics chips. Basically just an ALU engine.

- X1900XT(X). This chip has 300% the PS performance of an X1800XT, but in games it only gets a 0-20% lead in most cases. Why? All those ALUs are idle since most games currently don't use them. The X1900XTX has a TON of power just sitting there, unused... waiting to be used by something like physics!

- ALUs are CHEAP. It cost ATI less than 63M transistors to add 32 ALUs into R580, about 2M transistors an ALU. That is well over 200 billion flops (200GFLOPs). Basically ATI/NV could add ALUs just to have a hug surplus for physics. Since DX10 GPUs are more robust and are actually much better general processing oriented, a few tweaks and changes could make them VERY Physics friendly. The power is there. The X1900XT has a total flops of 500B. This will only increase. By fall 2007 we will break the 1 TFLOPs (Tera-flops, as in TRILLION) range in PROGRAMMABLE shaders. AGEIA does not have the R&D, product range, or the know how (they are still on the large, hot, and outdated 130nm node... I believe some of the NV35s were made on that!).

- Another level of latency. With PhysX you have a CPU, GPU, and PPU. ALL working over a bus fighting for traffic and coherancy. With a GPU you have only a CPU and GPU, limiting some of the traffic and coherancy.

- The chips are expensive for what they are. 125M transistors on the 130nm node process, 25W for the chip alone. To compare GPUs are on the 90nm processing, moving to the 80nm process and are 3x as big. What they lack in elegance they make up for in brute force.

GPUs are not perfect for physics. Advanced rigid body collision detection takes more work, and they may (or may not) be faster than AGEIAs solution. But pure speed is not always the determination of the winner. VALUE, Marketshare, and PR are. And in games, the most important factor is developer support.

Thus far AGEIA has over promised and never provided a product (well, until now). Delays have been common and the chip itself is waaaaay overpriced.

Yep, not a fan. More useless hardware. At least they woke up the GPU makers.

What is more interesting is how ATI and NV will fight it out. ATI is about 5x faster in GPGPU tasks than NV (that was comparing the 7800GTX with X1800XT), and ATI has done a lot of work on the 360 with MEMEXPORT to specifically accelerate this type of tasks. 6 months ago one of the "bullet points" of the X1000 series was it can be used to accelerate a number of tasks and ATI has had working physics demos up on hardware as old as the 9700. With all the extra ALUs on the R580 + the excellent dynamic branching, I get the feeling ATI could have a significant lead here.