I've been reading around the internet and I've noticed a lot of people mentioning that Cross-Firex setups can really improve performance yet most games don't use the full potential of 2GPU's and most of the time an insane amount of video ram is useless because no games are built to utilize that.
I personally have X2 HD 5770's CrossfireX and I've noticed almost doubled performance in certain games. And terrible performance in others including micro-stutter and shadow flickering issues. Dirt2 ran at 28-30 FPS on a single card everything on Ultra High with 16XAF and 8XAA. And with the 2nd card It runs solid 60FPS and never changes. For some odd reason if I disable v-sync it drops to like 5FPS weird bugg.
I've noticed with games like Assassins Creed II, Splinter Cell Conviction, Prototype, GTA IV, Just Cause 2 and others that the games run at 60FPS or higher if v-sync is disabled but randomly drop to the 20's and 30's even when the scenery doesn't change. I've heard with some games that companies port over from X-BOX 360 the performance sucks on PC's even with the best performance parts.
My question comes down to thinking games don't use all the potential of Cross-Fire/SLI setups.
Using EVEREST with my G15 keyboard I can monitor GPU utilization and even the games that run the slowest like GTA IV my graphics cards never even hit 50% they jump between 15%-30% on both cards. And the game runs terrible slow.
Now, I can run Crysis on Ultra High with no AA at high 50's to 70's FPS.
What is the deal with this?
Im thinking it's very similar to when Multi-Core Processors first came out. They had tons of power, but they had no applications using more then 1 thread and nobody saw a large performance increase.
Some feedback would be nice on this subject.
It's actually very aggravating considering I payed so much for both cards.
I personally have X2 HD 5770's CrossfireX and I've noticed almost doubled performance in certain games. And terrible performance in others including micro-stutter and shadow flickering issues. Dirt2 ran at 28-30 FPS on a single card everything on Ultra High with 16XAF and 8XAA. And with the 2nd card It runs solid 60FPS and never changes. For some odd reason if I disable v-sync it drops to like 5FPS weird bugg.
I've noticed with games like Assassins Creed II, Splinter Cell Conviction, Prototype, GTA IV, Just Cause 2 and others that the games run at 60FPS or higher if v-sync is disabled but randomly drop to the 20's and 30's even when the scenery doesn't change. I've heard with some games that companies port over from X-BOX 360 the performance sucks on PC's even with the best performance parts.
My question comes down to thinking games don't use all the potential of Cross-Fire/SLI setups.
Using EVEREST with my G15 keyboard I can monitor GPU utilization and even the games that run the slowest like GTA IV my graphics cards never even hit 50% they jump between 15%-30% on both cards. And the game runs terrible slow.
Now, I can run Crysis on Ultra High with no AA at high 50's to 70's FPS.
What is the deal with this?
Im thinking it's very similar to when Multi-Core Processors first came out. They had tons of power, but they had no applications using more then 1 thread and nobody saw a large performance increase.
Some feedback would be nice on this subject.
It's actually very aggravating considering I payed so much for both cards.