GF100 (Fermi) previews and discussion

Page 2 - Seeking answers? Join the Tom's Hardware community: where nearly two million members share solutions and discuss the latest tech.
Status
Not open for further replies.

i missed the 6000 cards news , are you guys joking or is there an actual press release?

http://forums.pureoverclock.com/showthread.php?threadid=7380

I dont see how taping out in Q4 is going to do much to bother nvidia.
 


We have absolutely nothing on the 6000 series other than it will be a new architecture, will probably be on 28nm at global found, and will probably not be here until late Q410 at the earliest.

As for this:

VIDIA will probably leave 5970 for at least 2nd quarter of 2010

At start they will give at least two cards of 5770 and 5850 range.
Then they will go for 5870
Then for lower market.

We only have info on the highest end cards. Nvidia will be top down like always. Probably we will see something faster than the 5870 first (360), then at the same time but in limited quantity we will see the 380 slipping in likely under the 5970. Around late spring they might have a working dual 360 card. We won't see 5850 level cards until the spring summer time frame, starting higher and moving lower.
 
http://www.hardocp.com/article/2010/01/17/nvidias_fermi_gf100_facts_opinions

And my favorite:

1263608214xxTstzDnsd_1_18_l.gif


Looks like Fermi blows the 5000 series out of the water...Redesigning the entire pipeline is paying of giant dividends for NVIDIA right now.

Also, I was right on the 512 cores while everyone else claimed "only" 448. 😀
 


Most of the major sites have stated that the tessellation implementation is too different to really compare using that benchmark. On Fermi it is tied to the shader cores, so it will hit performance more than on ATI. Though that may not matter if it is sufficiently faster.

As for the "only" 448.. didn't most people claim that the 512 will come around just not in great numbers right away? Well except for the trolls. I don't think anyone rational was claiming otherwise.


Also, until I am installing my fermi it is blowing nothing out of the water...

I'm excited about the possibility.. but we have nothing yet.. I can't say I'm too keen on slides from nvidia either.. as the ATI slides for the 5870 claimed some ridiculous things as well.
 

I agree but I do get a feeling fermi will be faster, waiting for the 3rd party benches to bring a realism to the table. I flip nvidia fans some pooh but NV does a great job ,and in areas ATI fails to compete. Some one showed me a gpu sales trend over the span of a few years and the entire market followed nvidias lead, ATI never rose above in discrete sales and when you think in terms of rendering farms and commercial gfx ability ATI is like nowheresville
 
Implemention != performance, true. But as others have pointed out, performance is what people buy, and the first comparison across a single benchmark gives a 20fps advantage to Fermi.

The only question I have now is actual gaming performance on non-DX11 SW and Price.
 
RAWR need more benchmarks preferably by 3rd party ppl not nvidia, as we all know by reading their driver change log apparently they know magic or something when it comes to performance increases and benchmarks.

Oh well their little **** tease with this *** they were showing off put me in a better mood about it although the price is still *** in the air.

If they can throw down a 5770 and 5850 competitor that out preforms it in DX11 then they can gain quite a bit of market back imo.
 

There is an ATX specified 300W limit (which is why the 5970 uses downclocked but higher binned 5870 Chips).

AMD circumvented the 300W limit by making their 5970 highly overclockable in the hands of the end user but the limit remains.

We know that a Fermi based card with 448SPs (and a 600-750MHz core clock) has a power envelope of >225W. How much of a cut down would nVIDIA need to get two Fermi's on a single card? I would think too much of a cut down, so much so that a Dual GPU card is no longer feasible as it cannot compete with the 5970.

I think nVIDIA will wait for a die shrink before they release a Dual GPU product.
 

good to know. Tsmc is at 28nm Q4 so likely not any earlier right? If so it seems ATI AMD has done a bang up strategy with 5970.
 

:lol: :lol: :lol: really , tell us how you truly feel.
 

Or they can just lop off parts(if i remember parts from the like 5-7 articles about the architecture right) until they get it down to the power requirements. They never said the dual gpu single card will be dual chips from their flagship card.
 


I am almost certain they will release a beast of a dual GPU card if only to ensure that the fastest thing on the planet is Nvidia for some unknown amount of time.

While the radeons are certainly more powerful per area of die.. I'm not sure how it will stack up performance/power. We might see a dual 8pin 'special' edition 5980 that is clocked slightly higher with better cooling (to overclock the 590 well requires special VRM cooling, they are not dealt with well on the stock cooler) but that is as far as ATI would bother going with this gen. While I would not at all be surprised to see a GTX390 that was effectively two 448 parts together with no regard for staying in spec. I'm sure there is some stupid way around the spec they will exploit.. They would only be after claiming the single card crown, if one fermi does not do it they will figure out a way to smash exactly 300W of computing power together as maintaining the crown seems important to their business model.
 

That benchmark is tessellation-heavy, more than most games will ever be. It doesn't represent the real world much more than 3D Mark.
 


It is still impressive to watch it chug it out..

Mind you, I'm more interested in how everything else light that benchmark is, as opposed to tessellation heavy. It doesn't look nice without tessellation giving me the impression that other than tessellation it is very easy on a GPU... I'd like to see how well fermi handles tessellation if it is also computing a texture fill heavy scene with AA. At what point does the dedicated hardware of ATI's tessellation engine overtake fermi, if ever, or is the computing power of fermi such that it won't make a difference.
 
Yes the unigine engine used was created very much for tesselation. Still I see not much worth discussing till they are out in the wild. So much speculation without any real numbers.

Odd that gamer posted a [H] link considering how much flak they are giving nvidia right now.

I thought it was odd as well.. though Kyle didn't write that article, just the bit at the end. He seemed rather impartial though, a few stern words here and there though he did seem impressed.
 
Speaking of Eyefinity, Nvidia will get it in their new cards through 3D Vision Surround. Ive allways thought that Nvidia would have to come out with there own version sooner or later, and well here it is!. The only realy incredibly stupid thing I see with Nvidias implementation is that they force you to have TWO cards in SLI. Youd think that they would learn form ATIs mistake, being that the 3rd monitor had to be Displayport, and just put 3 outputs on the back of the card...
 


Still, considering how badly the 5770 and below did in actual DX11 benchmarks, I'd say anything over 40FPS is outstanding at this point. Tesselation is the hardest part of DX11 computation, and if the first DX11 entry can handle that benchmark, which is overkill, I would expect to see much better numbers in DX11 titles (flawed as they may be at this point) compared to ATI.

If NVIDIA can keep the price of the top card under $450 and beat ATI in performance, we might yet have a repeat of the HD2000 vs G80...Everyone here knows I view the 5000 series as too weak for DX11/Tesselation (except maybe the 5870 and above; we'll see), so if these numbers hold, I don't think thats a bad comparison...

Having an actual DX11 game that wasn't a DX9 add-on would be interesting right about now...
 


Wait..lol?

Can we see half a g100 running tesselation before you claim the 5770 and below does dx11 'badly'? 😗

If NVIDIA can keep the price of the top card under $450 and beat ATI in performance, we might yet have a repeat of the HD2000 vs G80...

I really, really, really, really, really hope you are being sarcastic.
 

Yes... two cut down variants, but how much of a cut down would they need to fit the 300W power envelope? Probably something to the tune of a Dual 320SP part (to fit the power envelope) which only means a theoretical 640SP performance (not enough to compete with a Radeon HD 5970 (I would think).

I am thinking around 15-20% better performance for GF100 (Fermi) over a Radeon 5870 1GB (averaged out over various titles).

Radeon HD 5970 are two RV870 (Radeon HD 5870) cores (highly binned ones) on a single board (stock clock is lowered to fit the 300W envelope).

You would need around a 40% + win for GF100 over Radeon 5870 across the board to make a Dual GPU 320SP part feasible (and we have to remember that memory units in GF100 are tied to SP blocks so a 320SP card is lacking several memory blocks to boot).

I don't see a dual GPU card (could still happen I just don't see it) at launch. With the move to 32 or 28nm though, anything is possible in the second half of 2010.
 
The 3D nVidia Implementation is similar to ATIs 6 monitor. If its that much better, you shouldnt need SLI.
My take is this. Fermi is the first in a 3 gen (most likely) rehash down the road, and has the baby steps of whats going to happen down the road, much like the R600 was thus the 5xxx series, and next new gen, where we'll most likely see similar implementations, as early DX11 games wont be that hard, since most wont be ground up games using it, and top cards should handle it, its the second gen as usual thatll do better, as the DX model and games mature in its usage.
I would say AOHD is fairly demanding on DX11, and if used, its not nearly as scalable as tesselation is.
Theres other conversations going elsewheres saying some things, most arent forgetting where these "benches" are coming from.
The nVidia slide showing the 8xAA abilities at 35x16 is a great example of how these "benches" may be applied.
If the Fermi marks are indeed 2.33 times better than G200 at 25x16 using 8xAA, all this proves is what we do know.
We do know G200 8xAA sucks, and often fails at 25x16, so saying 2.33 is very subjective, and is what nVidia shows as triumph.

As for highend? I think we need to look at the partners here, as itll be them doing an end around on the PCI specs, and creating sone kind of frankensteinian "top" card, using more than 300 watts.
If this is so, not sure how nVidia will be able to claim victory here, as such a card will be outside their specs, even tho they can unofficially help in its desgn
 

A few more things we know
http://www.xtremesystems.org/forums/showpost.php?p=4204548&postcount=1484
Read the bottom of the post, and this is coming from Rys, whos had the best and closests understanding/ability of Fermi thus far.
I quote:
As for my clock estimates, I doubt 1700 MHz hot clock at launch 🙂sad, but the base clock should be usefully higher, up past 700 MHz. They still haven't talked about GeForce productisation or clocks, but at this point it looks unlikely the fastest launch GeForce will texture faster than a GTX 285.

Since its tied to the hot clock, going smaller will require large jumps in clocks to overcome in scaling, so, the lil Fermis may not show much promise at all, or worse.
 
Thats not a few more things we know, thats one persons supposition heres another on textures. Front page at Hocp.
http://www.hardocp.com/article/2010/01/17/nvidias_fermi_gf100_facts_opinions

Some GF100 Specifications

GF100 will have 512 CUDA cores, which more than doubles its cores compared to the GeForce GTX 285 GPU’s 240 core. There are 64 texture units, compared to the GTX 285’s 80, but the Texture Units have been moved inside the Third Generation Streaming Multiprocessors (SM)for improved efficiency and clock speed. In fact, the Texture Units will run at a higher clock speed than the core GPU clock. There are 48 ROP units, up from 32 on the GTX 285. The GF100 will use 384-bit GDDR5, so depending on clock speeds it actually operates at, there is potential for high memory bandwidth. These changes seem logical, and encouraging, but without knowing clock speeds actual shader performance is anyone’s guess.
 
Status
Not open for further replies.