GT300 NEWS

rangers

Distinguished
Nov 19, 2007
1,563
0
19,790
THERE'S A LOT of fake news going around about the upcoming GPUish chip called the GT300. Let's clear some air on this Larrabee-lite architecture.

First of all, almost everything you have heard about the two upcoming DX11 architectures is wrong. There is a single source making up news, and second rate sites are parroting it left and right. The R870 news is laughably inaccurate, and the GT300 info is quite curious too. Either ATI figured out a way to break the laws of physics with memory speed and Nvidia managed to almost double its transistor density - do the math on purported numbers, they aren't even in the ballpark - or someone is blatantly making up numbers.

That said, lets get on with what we know, and delve into the architectures a bit. The GT300 is going to lose, badly, in the GPU game, and we will go over why and how. First a little background science and math. There are three fabrication processes out there that ATI and Nvidia use, all from TSMC, 65nm, 55nm and 40nm.

They are each a 'half step' from the next, and 65nm to 40nm is a full step. If you do the math, the shrink from 65nm to 55nm ((55 * 55) / (65 *65) ~= 0.72) saves you about 1/4 the area, that is, 55nm is 0.72 of the area of 65nm for the same transistor count. 55nm shrunk to 40nm gives you 0.53 of the area, and 65nm shrunk to 40nm gives you 0.38 of the area. We will be using these later.

Second is the time it takes to do things. We will use the best case scenarios, with a hot lot from TSMC taking a mere six weeks, and the time from wafers in to boards out of an AIB being 12 weeks. Top it off with test and debug times of two weeks for first silicon and one week for each subsequent spin. To simplify rough calculations, all months will be assumed to have 4 weeks.

Okay, ATI stated that it will have DX11 GPUs on sale when Windows 7 launches, purportedly October 23, 2009. Since this was done in a financial conference call, SEC rules applying, you can be pretty sure ATI is serious about this. Nvidia on the other hand basically dodged the question, hard, in its conference call the other day.

At least you should know why Nvidia picked the farcical date of October 15 for its partners. Why farcical? Lets go over the numbers once again.

According to sources in Satan Clara, GT300 has not taped out yet, as of last week. It is still set for June, which means best case, June 1st. Add six weeks for first silicon, two more for initial debug, and you are at eight weeks, minimum. That means the go or no-go decision might be made as early as August 1st. If everything goes perfectly, and there is no second spin required, you would have to add 90 days to that, meaning November 1st, before you could see any boards.

So, if all the stars align, and everything goes perfectly, Nvidia could hit Q4 of 2009. But that won't happen.

Why not? There is a concept called risk when doing chips, and the GT300 is a high risk part. GT300 is the first chip of a new architecture, or so Nvidia claims. It is also going to be the first GDDR5 part, and moreover, it with be Nvidia's first 'big' chip on the 40nm process.

Nvidia chipmaking of late has been laughably bad. GT200 was slated for November of 2007 and came out in May or so in 2008, two quarters late. We are still waiting for the derivative parts. The shrink, GT206/GT200b is technically a no-brainer, but instead of arriving in August of 2008, it trickled out in January, 2009. The shrink of that to 40nm, the GT212/GT200c was flat out canceled, Nvidia couldn't do it.

The next largest 40nm part, the GT214 also failed, and it was redone as the GT215. The next smallest parts, the GT216 and GT218, very small chips, are hugely delayed, perhaps to finally show up in late June. Nvidia can't make a chip that is one-quarter of the purported size of the GT300 on the TSMC 40nm process. That is, make it at all, period - making it profitably is, well, a humorous concept for now.

GT300 is also the first DX11 part from the green team, and it didn't even have DX10.1 parts. Between the new process, larger size, bleeding-edge memory technology, dysfunctional design teams, new feature sets and fab partners trashed at every opportunity, you could hardly imagine ways to have more risk in a new chip design than Nvidia has with the GT300.

If everything goes perfectly and Nvidia puts out a GT300 with zero bugs, or easy fix minor bugs, then it could be out in November. Given that there is only one GPU that we have heard of that hit this milestone, a derivative part, not a new architecture, it is almost assuredly not going to happen. No OEM is going to bet their Windows 7 launch vehicles on Nvidia's track record. They remember the 9400, GT200, and well, everything else.

If there is only one respin, you are into 2010. If there is a second respin, then you might have a hard time hitting Q1 of 2010. Of late, we can't think of any Nvidia product that hasn't had at least two respins, be they simple optical shrinks or big chips.

Conversely, the ATI R870 is a low risk part. ATI has a functional 40nm part on the market with the RV740/HD4770, and has had GDDR5 on cards since last June. Heck, it basically developed GDDR5. The RV740 - again, a part already on the market - is rumored to be notably larger than either the GT216 or 218, and more or less the same size as the GT215 that Nvidia can't seem to make.

DX11 is a much funnier story. The DX10 feature list was quite long when it was first proposed. ATI dutifully worked with Microsoft to get it implemented, and did so with the HD2900. Nvidia stomped around like a petulant child and refused to support most of those features, and Microsoft stupidly capitulated and removed large tracts of of DX10 functionality.

This had several effects, the most notable being that the now castrated DX10 was a pretty sad API, barely moving anything forward. It also meant that ATI spent a lot of silicon area implementing things that would never be used. DX10.1 put some of those back, but not the big ones.

DX11 is basically what DX10 was meant to be with a few minor additions. That means ATI has had a mostly DX11 compliant part since the HD2900. The R870/HD5870 effectively will be the fourth generation DX11 GPU from the red team. Remember the tessellator? Been there, done that since 80nm parts.

This is not to say that is will be easy for either side, TSMC has basically come out and said that its 40nm process basically is horrid, an assertion backed up by everyone that uses it. That said, both the GT300 and R870 are designed for the process, so they are stuck with it. If yields can't be made economically viable, you will be in a situation of older 55nm parts going head to head for all of 2010. Given Nvidia's total lack of cost competitiveness on that node, it would be more a question of them surviving the year.

That brings us to the main point, what is GT300? If you recall Jen-Hsun's mocking jabs about Laughabee, you might find it ironic that GT300 is basically a Larrabee clone. Sadly though, it doesn't have the process tech, software support, or architecture behind it to make it work, but then again, this isn't the first time that Nvidia's grand prognostications have landed on its head.

s

The basic structure of GT300 is the same a Larrabee. Nvidia is going to use general purpose 'shaders' to do compute tasks, and the things that any sane company would put into dedicated hardware are going to be done in software. Basically DX11 will be shader code on top of a generic CPU-like structure. Just like Larrabee, but from the look of it, Larrabee got the underlying hardware right.

Before you jump up and down, and before all the Nvidiots start drooling, this is a massive problem for Nvidia. The chip was conceived at a time when Nvidia thought GPU compute was actually going to bring it some money, and it was an exit strategy for the company when GPUs went away.

It didn't happen that way, partially because of buggy hardware, partially because of over-promising and under-delivering, and then came the deathblows from Larrabee and Fusion. Nvidia's grand ambitions were stuffed into the dirt, and rightly so.

Nvidia Investor Relations tells people that between five to ten per cent of the GT200 die area is dedicated to GPU compute tasks. The GT300 goes way farther here, but let's be charitable and call it 10 per cent. This puts Nvidia at a 10 per cent areal disadvantage to ATI on the DX11 front, and that is before you talk about anything else. Out of the gate in second place.

On 55nm, the ATI RV790 basically ties the GT200b in performance, but does it in about 60 per cent of the area, and that means less than 60 per cent of the cost. Please note, we are not taking board costs into account, and if you look at yield too, things get very ugly for Nvidia. Suffice it to say that architecturally, GT200 is a dog, a fat, bloated dog.

Rather than go lean and mean for GT300, possibly with a multi-die strategy like ATI, Nvidia is going for bigger and less areally efficient. They are giving up GPU performance to chase a market that doesn't exist, but was a nice fantasy three years ago. Also, remember that part about ATI's DX10 being the vast majority of the current DX11? ATI is not going to have to bloat its die size to get to DX11, but Nvidia will be forced to, one way or another. Step 1) Collect Underpants. Step 2) ??? Step 3) Profit!

On the shrink from 55nm to 40nm, you about double your transistor count, but due to current leakage, doing so will hit a power wall. Let's assume that both sides can double their transistor counts and stay within their power budgets though, that is the best case for Nvidia.

If AMD doubles its transistor count, it could almost double performance. If it does, Nvidia will have to as well. But, because Nvidia has to add in all the DX11 features, or additional shaders to essentially dedicate to them, its chips' areal efficiency will likely go down. Meanwhile, ATI has those features already in place, and it will shrink its chip sizes to a quarter of what they were in the 2900, or half of what they were in the R770.

Nvidia will gain some area back when it goes to GDDR5. Then the open question will be how wide the memory interface will have to be to support a hugely inefficient GPGPU strategy. That code has to be loaded, stored and flushed, taking bandwidth and memory.

In the end, what you will end up with is ATI that can double performance if it choses to double shader count, while Nvidia can double shader count, but it will lose a lot of real world performance if it does.

In the R870, if you compare the time it takes to render 1 Million triangles from 250K using the tesselator, it will take a bit longer than running those same 1 Million triangles through without the tesselator. Tesselation takes no shader time, so other than latency and bandwidth, there is essentially zero cost. If ATI implemented things right, and remember, this is generation four of the technology, things should be almost transparent.

Contrast that with the GT300 approach. There is no dedicated tesselator, and if you use that DX11 feature, it will take large amounts of shader time, used inefficiently as is the case with general purpose hardware. You will then need the same shaders again to render the triangles. 250K to 1 Million triangles on the GT300 should be notably slower than straight 1 Million triangles.

The same should hold true for all DX11 features, ATI has dedicated hardware where applicable, Nvidia has general purpose shaders roped into doing things far less efficiently. When you turn on DX11 features, the GT300 will take a performance nosedive, the R870 won't.

Worse yet, when the derivatives come out, the proportion of shaders needed to run DX11 will go up for Nvidia, but the dedicated hardware won't change for ATI. It is currently selling parts on the low end of the market that have all the "almost DX11" features, and are doing so profitably. Nvidia will have a situation on its hands in the low end that will make the DX10 performance of the 8600 and 8400 class parts look like drag racers.

In the end, Nvidia architecturally did just about everything wrong with this part. It is chasing a market that doesn't exist, and skewing their parts away from their core purpose, graphics, to fulfill that pipe dream. Meanwhile, ATI will offer you an x86 hybrid Fusion part if that is what you want to do, and Intel will have Larrabee in the same time frame.

GT300 is basically Larrabee done wrong for the wrong reasons. Amusingly though, it misses both of the attempted targets. R870 should pummel it in DX10/DX11 performance, but if you buy a $400-600 GPU for ripping DVDs to your Ipod, Nvidia has a card for you. Maybe. Yield problems notwithstanding.

GT300 will be quarters late, and without a miracle, miss back to school, the Windows 7 launch, and Christmas. It won't come close to R870 in graphics performance, and it will cost much more to make. This is not an architecture that will dig Nvidia out of its hole, but instead will dig it deeper. It made a Laughabee

http://www.theinquirer.net/inquirer/news/1137331/a-look-nvidia-gt300-architecture
 

turboflame

Distinguished
Aug 6, 2006
1,046
0
19,290
On 55nm, the ATI RV790 basically ties the GT200b in performance, but does it in about 60 per cent of the area

When it's overclocked to the limit.

This article is biased as hell. ATI's tesselator isn't DX11 compatible IIRC. Plus I doubt Charlie was given the full schematics of a GT300 when the GPU isn't even taped out yet.

Nvidia hasn't released a new GPU on a new process since the FX series. If they do go 40nm they are going to have some issues with a huge die. Hopefully they learn from their mistakes of this generation, if they pull another G200 we might start seeing eVGA ATI cards.
 

The_Blood_Raven

Distinguished
Jan 2, 2008
2,567
0
20,790
He brings up a lot of good points that have been discussed before but does it in a very biased manor. Not to mention he also makes them up like his "actual" release date B.S.

No one will argue that nVidia has a good few things to overcome that are more serious for them than ATI, such as:
1) DX11 implementation
2) DDR5 implementation, tweaking, and the purchasing of the new modules
3) Reportedly another large chip = not very cost efficient
4) Physx and Cuda are going to fall on their face if demand doesn't increase for them

However, if the chip is as large as reported then I can not see how it will not perform well. DDR5 should solve all the bandwidths issues nVidia has been having and nVidia doesn't seem to have a problem with going over the top with the hardware without listening to the budget department.

Also ATI will not be doing great anyway. They have a headstart with most of the topics posted above, but they also have MUCH less in the way of funds for R&D like nVidia has. Also AMD still is suffering losses and isn't the most financially secure at the moment.

What will likely happen is a similar situation as we have now with the GTX 2xx vs 4xxx series, but ATI might have a slight edge with DX11.
 

Dekasav

Distinguished
Sep 2, 2008
1,243
0
19,310
He lost a lot of credibility for me when he was talking about die sizes and mentioned that " [GPUs are on] 65nm, 55nm and 40nm [processes].

They are each a 'half step' from the next, and 65nm to 40nm is a full step"

55nm to 40nm is a full step, 65nm to 45nm is a full step, 65nm to 40nm is a step and one half.
 
Yep, that wasn't biased. Its too bad, because he brought up good points, but they were lost in all the bashing. I'm fairly confident some things will work out for NVidia and some won't for ATI. And also, what makes him think his sources are so accurate? I also noticed he just took ATI's word for their release date and just threw NVidias out as lies. I guess we'll see.
 

Zerk

Distinguished
May 6, 2008
297
0
18,790
[quotemsg=6713806,3,186442]looks like someone got fired from Nvidia and decided to get back at them...[/quotemsg]


I Agree!


I don't think i would sit back for one second and say Nvidia doesn't have the knowledge or "Whatever" to make a GDDR5/40nm/DX11 Card.

With all the Great! Advances Nvidia has made in only the past year!

Well I'd say they might be on a slower but Much Smarter way to Advancements.

ATi has made some good stuff to "lets not let this be a one sided Battle "

Both have done Amazing Work!

ATi Slapping 500 or more Cores on one Card.

Nvidia with its Cuda,PhysX,Hybrid,Pure video HD.

Go to Nvidia's Web site you can find a lot of DX11 stuff.

I think they have a Nice surprise Coming up like they normally do.


You watch.

Nvidia is Still the King of Visual Entertainment.

Vary Goop post though.


Good job!

 

B-Unit

Distinguished
Oct 13, 2006
1,837
1
19,810
[quotemsg=6713838,7,186569]He lost a lot of credibility for me when he was talking about die sizes and mentioned that " [GPUs are on] 65nm, 55nm and 40nm [processes].

They are each a 'half step' from the next, and 65nm to 40nm is a full step"

55nm to 40nm is a full step, 65nm to 45nm is a full step, 65nm to 40nm is a step and one half.[/quotemsg]

Huh? If 65nm to 45nm is 1 full step (20nm difference) how is an additional 5nm a half?
 

daedalus685

Distinguished
Nov 11, 2008
1,558
1
19,810
[quotemsg=6713813,4,74232]
On 55nm, the ATI RV790 basically ties the GT200b in performance, but does it in about 60 per cent of the area

When it's overclocked to the limit.

[/quotemsg]

While I would usually agree, with almsot all ATI partners releasing 1ghz 4890's in the coming couple weeks I'd be really interested in a comparison of the best avaliable retail cards. Most OC reviews are self overclocked and what have you.

I'd have to imagine the 1ghz 4890's would be effectively on par with the highest OC 285's avaliable at retail. Though a factory OC of 150mhz will probalby push the 4890 into a higher price segment (I wonder though).

I hate comparing user overclocked cards to user over clocked cards and having people state "x is faster than y" as there are too many variables. But It would be nice to see a factoy overclocked roundup here, really put to rest what the fasted card you can "buy" is. Maybe next month.

I find it strange the way ATI has gone about the 4890. In previous generations they would have released an official GPU with official faster specs. (1900gt vs 1900xtx for example, even 4850 vs 4870) Yet this gen they are sticking with the 850mhz 4890 and just having the partners go as crazy as they want. I mean a 1ghz 4890 is as far removed from a 4890 as a 4890 is from the 4870... I question the point of calling it the same card at this point. It makes more sence to me to have the 4870, the 4890, and the 4895+ or something for the 1ghz cards (they msut have great yields if every partner is comgin out with a 1ghz)... With so many OC varients, even from release I wonder why they spec'd the stock so low in the first place.. at any rate..
 

daedalus685

Distinguished
Nov 11, 2008
1,558
1
19,810
For the last fcuking time, stop quoting whole articles, haven't you been warned enough about this dipshit!!!!.

But it's not like there is some easy way to make a clickable "link" to the source right? ;) Besides that, if he had linked it peolpe might have noticed it is from the inquirer.... lol
 

rangers

Distinguished
Nov 19, 2007
1,563
0
19,790
For the last fcuking time, stop quoting whole articles, haven't you been warned enough about this dipshit!!!!.


no i haven't, anyway couldn't give a flying *** what you think, and by warned, do you mean you opening your big mouth or being pulled up by one of the mods, yes to the first one and no to the second, piss off you NeoNatzi
 
[quotemsg=6713957,10,73097]Huh? If 65nm to 45nm is 1 full step (20nm difference) how is an additional 5nm a half?[/quotemsg]
Theoretically, the true half node would be 38.5, or half ways between 45 and 32nm, so 40nm is considered a half node.

Charlie does bring many truths, some stretched, some not at all in this, but even between his sometime stretches, hes biased. This all started awhile ago, when he and an nVidia "dude" got in a fight in HongKong. The words have gotten worse sine then. It was the mobile card failures at first, and he still hasnt let it go
 

turboflame

Distinguished
Aug 6, 2006
1,046
0
19,290
[quotemsg=6713868,9,175994]
I don't think i would sit back for one second and say Nvidia doesn't have the knowledge or "Whatever" to make a GDDR5/40nm/DX11 Card.
[/quotemsg]

Actually that's one area where Charlie has a point. Releasing a new architecture on a new process is very risky, Nvidia learned that the hard way when they released the FX series on a new 130nm process. Since then they've always played it safe and used a more mature process, that's why it took so long for them to transition to 55nm.
 
More to his point, if nVidia had actually made a DX 10.1 part, instead of pretending how useless it was, itd have a Tesselator, and the DX10.1 HW in their cards already. Now, theyll have to do that, plus a whole new arch, plus add in the requirements (die space), and all on a new node. Not the best scenario for any company
 

uncfan_2563

Distinguished
Apr 8, 2009
904
0
19,010
This is not the first time he has done this, to my knowledge you are not allowed to directly quote whole articles on the forums for multiple reasons including copywrite.

The author has not given him permission to use it, there was no indication where it was from, that inq. link could have been to another source saying the same thing, it is just not right.

There are certain ways of doing things on forums and that ain't one of them.

User agrees not to post any material that is protected by copyright, trademark or other proprietary right without the express permission of the owner(s) of said copyright, trademark or other proprietary right.
You aren't a mod and no one in their right mind will make you one just cuz u rip on people because of your interpretation of the rules. he brought the topic up to hear people's feedback and stuff okay? he could have just posted the link but this made it more convinient for all of us.

god...why do people just popup in a thread to start a fight?
 

tenaciousleydead

Distinguished
Dec 6, 2004
812
0
18,990
I bet you all your posts combined probably equal the length of his copied article, and to tell the truth i found his post a lot more interesting than yours stranger. If anybody enforced your fanatical rules this would be the shittiest forum
 

tenaciousleydead

Distinguished
Dec 6, 2004
812
0
18,990
some rules cause too much friction, some people cause too much friction. a bureaucratic system is far from efficient, if everybody had to site something just to start a thread around 90% of the populous wouldn't post, or would post on a forum that didn't enforce there rules. i just cant see how bending a minuscule rule to make a forum better causes you so much upheaval
 

rangers

Distinguished
Nov 19, 2007
1,563
0
19,790
the only person taking up space on here is you and BTW threats only work if you've got some kind of authority, and the last time i had a look you where still a little man with a limpdick
 

B-Unit

Distinguished
Oct 13, 2006
1,837
1
19,810
stranger is right, although I cant say that rangers specificly has been warned by mods, they have warned others about this kind of posting. It is a violation of copyright, and tom's is liable to be sued.
 

TRENDING THREADS