Yeah, like how cloud operators want Perf/W. And TSMC's 7 nm delivers the goods on that.the market is saying loud and clear they don't care about lithography used to make their computers.... They have 1 question - will it do what I want it to do?
The scheduled forum maintenance has now been completed. If you spot any issues, please report them here in this thread. Thank you!
Yeah, like how cloud operators want Perf/W. And TSMC's 7 nm delivers the goods on that.the market is saying loud and clear they don't care about lithography used to make their computers.... They have 1 question - will it do what I want it to do?
Uh, not really. Ray-triangle intersection tests are fairly compute intensive, but BVH traversal is pretty cheap, computationally. The problem is more bandwidth limited, as it lacks good coherency.Raytracing is ultra GPU calc inensive- there are no shortcuts
Except Toy Story wasn't ray traced! In fact, most movies didn't use ray tracing, until about 10-15 years ago.The goal of ray tracing is to be able to play in a Pixar movie level game in real time - all of those movies use ray tracing - some frames on the original Toy Story took 8 hours of CPU time minimum (per frame, 24fps or 30fps). Neither AMD nor Nvidia have a lock on ray tracing - and there are no shortcuts to full scene ray tracing - it is raw GPU power.
True full scene GI is basically just doing ray tracing on everything -- like what Minecraft RTX does, on some level. Theoretically, anything with full DXR support (or VulkanRT) can do full GI ... but most games using RTX are only focusing on one or a few subsets to keep performance up. Metro Exodus does RT GI, but it's actually indirect lighting and not the same as the full GI you're referring to.@JarredWaltonGPU
I'm probably reading more into this than you meant, but will AMD support Global Illumination? That would be a big deal, because it's one of the applications of Nvidia's tensor cores (i.e. for denoising - an essential part of GI).
We'll have to see. If AMD is saying RDNA2 has 50% higher perf/watt than RDNA1, I doubt it will be just from adding CUs. We know the CUs had to change (to add RT support at a minimum), and in the process I suspect AMD found some other items it could tweak to improve efficiency.I don't know about that. I suspect they just plowed that dividend into adding more CUs.
That still nets you more fps, but fps per GFLOPS would stay constant.
I really would love to know hard numbers on how many Radeon VII cards were produced and actually sold. I suspect the final number (not including Radeon Instinct) is extremely small -- less than 100K for sure, but probably less than 20K and maybe even less than 10K. Well, it might be higher because some imaging pros might have gone with Radeon VII instead of paying more for Radeon Instinct, but if we were able to look at GPUs actually used for playing games? I think the majority of Radeon VII parts used in gaming are only in the hands of reviewers and "Team Red" members. I don't know anyone who actually bought one, because it was overpriced and underperforming.Agreed on the RX 590, but I think AMD was just trying to cash in on cryptomining.
With regard to the Radeon VII, in fact it does basically match the RTX 2080! And it was virtually free to release, because it's a slightly-cripled datacenter GPU that they just sold to consumers as a bonus. Originally, they weren't planning to, but then they saw a market opportunity.
I don't know about that. I suspect they just plowed that dividend into adding more CUs.
That still nets you more fps, but fps per GFLOPS would stay constant.
My guess -- assuming Intel does multiple dies -- would be:Only time will tell. I'm curious if they gutted any int 8 or whatever to streamline for gaming. AMD are splitting GPU lines so they can sacrifice a bit of compute on the gaming cards like Nvidia does.
For that degree of partisan zeal, he'd either have to be a new employee (i.e. still in the "honeymoon" period) or an investor. I'm betting on #2. Probably a recent investor who's either under water or was betting it'll go up a lot more.I think we found an Intel employee.
Yeah, it doesn't even pass a basic sanity test. How could MS afford this or guarantee that the connection quality would support it, for all of the console's users?@Deicidium369 > the slide that said the full screen real time RT will be done IN THE CLOUD..
I didn't think so, but I wanted to check on it. Apparently, they did fix a couple bugs (Loop Stream Decoder) and add some in-silicon mitigations, but that's really it. And for models with higher core count, the total L3 cache increased, due to cache slices being part of the core tiles. But the cache-per-core ratio stayed constant.What has Intel changed about its micro-architecture since Skylake? Absolutely nothing.
Well, I guess whatever it is that Nvidia is calling GI. Related:Metro Exodus does RT GI, but it's actually indirect lighting and not the same as the full GI you're referring to.
Realistically, I think we need close to 10X the ray tracing hardware of today's GPUs to make modest full scene GI (or full ray tracing, if you prefer) viable. Maybe more than 10X.
Now that you mention it, that seems to be the case. When Turing launched, I'm sure I read something about how they were using the Tensor Cores for de-noising, though. I recall seeing demos of their AI denoising and how it was so good that they could even get away with just a couple lighting rays per image pixel.And right now, for games, Nvidia isn't even using the Tensor cores to do denoising AFAIK.
Well, it was priced comparably to RTX 2080 and GTX 1080 Ti, to which it's roughly equivalent.I don't know anyone who actually bought one, because it was overpriced and underperforming.
What's your deal with int8? You've mentioned it several times, but do you actually know what GCN had, in the way of int8, or what it's useful for?I'm curious if they gutted any int 8
Vega 7nm includes the additional instructions listed below:...* V_DOT4_I32_I8* V_DOT4_U32_U8...
This makes it sound like the improvement from GCN to RDNA is bigger than it is. GCN would dispatch an instruction from one wave every cycle, but it rotated between feeding 4 different SIMD16 pipes from 4 different waves. So, when you got the pipelines full, you'd retire 64 results every cycle. It's explained, here:GCN issued one instruction per wave every four cycles; RDNA issues an instruction every cycle.
AMD's terminology is to call these "shaders", which I feel is better, as it avoids promoting a misunderstanding of these cores as being comparable to CPU cores (which you've already said they're not).the PS4 clearly had the faster GPU. It had 18 CUs or 1152 GPU cores
No, Phoronix has reported that driver changes indicate Arcturus will lack any 3D graphics blocks.Arcturus could eventually make its way into consumer / prosumer graphics cards.
I'd wait simply because there is a fair chance of considerable price cuts across the 5000-series.Guys should I hold of buying 5600 xt and wait for 6600 xt with RT or not?
What are you using now, and how much are you willing to spend? My expectation is that Nvidia and AMD will both launch high-end ($500+) graphics cards in September/October. It will probably be several more months at least before the 'mid-range' cards come out. AMD launched the RX 5600 XT in January, so it's not really due for a refresh until at least next January. Plus, it sounds like there's a lot of demand for TSMC 7nm so AMD might not have a ton of chips ready to go. Don't be surprised if an RX 6600 XT (or whatever it's called) ends up being priced closer to $350-$400. $400 is certainly my expectation for Nvidia's RTX 3060, because I'm pessimistic. 😛Guys should I hold of buying 5600 xt and wait for 6600 xt with RT or not?
Thx for response. I was using gtx 970 until it broke. Now i have backup gpu in R5 240 with i7 4770, 16gb 1600 ddr3, 21:9 1080p 60Hz monitor. I play mostly Apex and thats not very demanding game. I wantto buy 144Hz monitor so therefore my interset for 5600 xt. Otherwise used rx 580 would be good replacement for my monitor.What are you using now, and how much are you willing to spend? My expectation is that Nvidia and AMD will both launch high-end ($500+) graphics cards in September/October. It will probably be several more months at least before the 'mid-range' cards come out. AMD launched the RX 5600 XT in January, so it's not really due for a refresh until at least next January. Plus, it sounds like there's a lot of demand for TSMC 7nm so AMD might not have a ton of chips ready to go. Don't be surprised if an RX 6600 XT (or whatever it's called) ends up being priced closer to $350-$400. $400 is certainly my expectation for Nvidia's RTX 3060, because I'm pessimistic. 😛
If you've got at least something at the RX 570 level (R9 390), I'd try to stretch that as long as you can until you get games that don't run fast enough at the settings you want. Once that happens, take the plunge on a new GPU. There are games where 1080p at high settings is going to fall below 60 fps on a 570, though, so it's at that point where I'd want to upgrade. Next month for Nvidia we should get Ampere, and then Navi 2x in October. Wait and see is the sage advice.
If we're low-balling, then so are most of the other rumors circulating. AMD hasn't had a chip bigger than 500mm2 since Fiji, which was not a great design overall. It did new stuff like HBM, but underperformed. If not for cryptocurrency mining, it wouldn't have sold many units at all -- and in fact I don't think the Fury cards ever showed up on the Steam Hardware Survey, which means they likely never even reached more than 0.15% of the gaming market.I believe that RDNA 2 is going to be bigger than 96 ROP. I think its going to have 128 at least & maybe even more, because I believe AMD is aiming to surpass the TITAN (Ampere).
So in my opinion I think Tom's Hardware estimates are low-balling AMD's supposed "HALO" product.
If we're low-balling, then so are most of the other rumors circulating. AMD hasn't had a chip bigger than 500mm2 since Fiji, which was not a great design overall. It did new stuff like HBM, but underperformed. If not for cryptocurrency mining, it wouldn't have sold many units at all -- and in fact I don't think the Fury cards ever showed up on the Steam Hardware Survey, which means they likely never even reached more than 0.15% of the gaming market.
Anyway, historically AMD has never gone quite as big as Nvidia and managed to pull it off. Fiji tried and failed, Vega tried and came up short. Hawaii by comparison was 'only' 438mm2. So now we have Navi 10 at 251mm2, and we're saying Navi 21 will double the CUs, add ray tracing hardware, and you think it will be even bigger than that? Navi 21 could very well be the largest GPU chip (not counting silicon interposer on Fiji and Vega) that AMD has ever created. I very much doubt it's going to be as large as GA100, however.
The reality is AMD doesn't sell as many cards into the professional space as Nvidia. It can't afford to make a huge chip that no one buys. That's the whole purpose of the chiplet approach for Zen 2. Maybe the Frontier supercomputer will change the dynamics and allow AMD to really go big, but it's a huge leap from Navi 10 to something that maxes out reticle size.
PS5 has 36 CUs clocked at up to 2.23 GHz. Xbox Series X has 52 CUs clocked at up to 1.825 GHz. Notice how more CUs ended up with a much lower clockspeed? That's my assumption for a 72 CU RDNA2 GPU -- it will not clock as high as the PS5, and possibly not even as high as the XBSX. But we don't know for sure -- a higher TDP could certainly allow for higher clocks. 300W and 2.0-2.1GHz? Maybe!There are three things AMD specifically put out going from RNDA 1x to RNDA 2x:
PS5 does 2.3ghz and that is with a CPU in the same core plus restricted by a game console configuration. We do not know the upper most end for clock speed with efficiency and power, clock power curve. AMD may well have an option to push 2.3ghz+ out of a 72CU-80CU GPU if they want to.
- IPC improvements
- Logic Enhancements to reduce complexity and switching power (better perf/w, 50% better)
- Physical Optimization -> INCREASE CLOCK SPEEDS
If AMD has the option to, they should blow up Nvidia Ampere line in Rasterization performance then it will become a battle of useful features after that.
That is most likely a power limiting factor, form factor, cooling and overall cost for a console. Still we won't know until real hardware is reviewed representing actual products. 1600mhz would be slower than the Vega LC as a note which clocked at 1700mhz. My 5700 XT does 2100mhz without too much effort. With a more mature process, experience and time (also less power needed for a given performance) it should clock rather high. Their goal was indeed faster clock speeds which 1600 would really be a fail in that goal.PS5 has 36 CUs clocked at up to 2.23 GHz. Xbox Series X has 52 CUs clocked at up to 1.825 GHz. Notice how more CUs ended up with a much lower clockspeed? That's my assumption for a 72 CU RDNA2 GPU -- it will not clock as high as the PS5, and possibly not even as high as the XBSX. But we don't know for sure -- a higher TDP could certainly allow for higher clocks. 300W and 2.0-2.1GHz? Maybe!
Unless my math is wrong; 80 Compute Units * 96 Raster Operations * 12.28 TFLOPS of single precision floating point (FP32).
Not bad AMD. Not bad. Let's see what that translates to in the real world