Guru3D forum user discovers a workaround to get RTX Voice to work with non-RTX graphics cards.
Nvidia RTX Voice Works Fine On Non-RTX GPUs : Read more
Nvidia RTX Voice Works Fine On Non-RTX GPUs : Read more
Clearly not.The latest discovery raises the question on whether RTX Voice really requires Tensor cores.
No. It's just audio processing. As I said in the comments on previous article about this, it might even be usable on CPUs.As things currently stand, the feature seems to run smoothly on just CUDA cores. Nevertheless, it remains to be seen if RTX Voice has any performance impact on GeForce graphics cards that lack proper Tensor cores.
I hate to say I really do ...but that dose not help you have to not buy the games that are more optimized to run on Nvida cards. The whole game market is pretty much focused on Nvida more so than AMD and it helps Nivda dose more to optimize their drivers more frequently.You know I would have respected NVIDIA more if they said, "This is a premium feature you need to pay for with a higher end RTX Card to receive"
Instead they just lie through their G-D teeth saying it requires "Tensor cores" to promote a marketing objective which is a blatant lie.
This is why I refuse to buy NVIDIA. They are just a bunch of a-clowns when it comes to be being straight forward and honest. They are anti competitive as well.
Cool. Thanks for the report!I tried RTX Voice on a GTX 750 Ti (Maxwell) card.
We don't actually know for sure how / what it uses RTX/tensor core features for. Part of me wonders if they are doing full sample analysis on the cores (which I think would be inefficient as technically you're working with a larger dataset) or using some other parts of the GPU to assist also, e.g. using NVENC block to perform real-time DCT of the audio stream then using the GPU matrix/tensor cores calculate and remove audio patterns which are programmed to be ignored, and shunting the end result back through to the user app (although that may also be more working overhead).Clearly not.
No. It's just audio processing. As I said in the comments on previous article about this, it might even be usable on CPUs.
Anyway, it can either keep up with realtime or not. That's really the only question. Without tensor cores, you're almost certainly using fp32, instead of fp16. So, precision shouldn't be an issue. It just comes down to performance.
Now, if someone can try it (either on a RTX or GTX card) and post their GPU utilization, that might shed some light on how much compute power it really requires.
Sadly, my Nvidia GPU is a 900-series...
Do we know that clock speeds on RTX GPUs don't rev up when using RTX voice?Nvidia didn't exactly lie about RTX Voice...
I installed it on my 1080 Ti, and it is just power hungry!!!! My GPU is always 139MHz, even when watching 4K YouTube. But when I activate RTX Voice, it clocks my GPU to almost 1800MHz........ ON STANDBY, with all apps closed!!!!! 😭
If it needs that much power on stand by, imagine how the games would run...................
And I'm only using the Input Device setting, not both.
So no, they didn't lie. It just works better on RTX GPUs. Same with Ray Tracing.
I have an RTX 2080 Ti and have been giving the RTX Voice program a try for a bit. Input works fine, with people saying they hear pretty much any and all background noise eliminated from my end.
Output is a different story however, since I cannot get it to work in any circumstances. Doesn't matter if I'm using speakers or headphones, or if I'm actually telling it to filter the background noise or not- I get absolutely nothing from the other end (tried with Skype, Discord, and Steam, though I think only the first two out of those three are actually officially supported).
Anyway, regarding GPU impact, my clock speeds for my video card stay constant, regardless if the program is up and running and being used during an active voice call, or if it's completely closed. 1350 mhz core clock, 7000 mhz memory clock (according to MSI Afterburner).
Looking at GPU usage during an actual call with someone (and only filtering input, since I can't get output to work) my usage seems to fluctuate between 4% and 6%. Only a marginal difference from the previous posters non-RTX card results.
Wow, so based on the 2080ti TDP (250W - I know TDP is not exactly equal to max power) that's very roughly about 10W power usage. Honestly... Can I say (donning flame proof coat) I don't think it's worth it... Not to say its worthless (and on a normal desktop PC it's ultimately negligible so on or off is irrelevant) I just mean cost vs reward for me is not worth it (efficiency wise its not great) - Certainly if I was working on a laptop on battery it is a significant draw.I have an RTX 2080 Ti and have been giving the RTX Voice program a try for a bit. Input works fine, with people saying they hear pretty much any and all background noise eliminated from my end.
Output is a different story however, since I cannot get it to work in any circumstances. Doesn't matter if I'm using speakers or headphones, or if I'm actually telling it to filter the background noise or not- I get absolutely nothing from the other end (tried with Skype, Discord, and Steam, though I think only the first two out of those three are actually officially supported).
Anyway, regarding GPU impact, my clock speeds for my video card stay constant, regardless if the program is up and running and being used during an active voice call, or if it's completely closed. 1350 mhz core clock, 7000 mhz memory clock (according to MSI Afterburner).
Looking at GPU usage during an actual call with someone (and only filtering input, since I can't get output to work) my usage seems to fluctuate between 4% and 6%. Only a marginal difference from the previous posters non-RTX card results.
You're confusing two different things, here.The more I read about the features that "utilizes" the tensor cores, the more I feel the tensor cores are a marketing gimmick in the retail market.
I carefully read everything they published about it, and there was never any suggestion that it was learning in realtime. They clearly said it was applying a pre-trained network to the data, which is realistically all you can really do.If you look at DLSS, when I first heard it, I thought its a great feature that the tensor cores can optimized the frames real time. Turns out that games first needs to be optimized, or more like the game developer needs to work with Nvidia to "teach" the AI how it should optimize. Nothing like a real time machine learning on our GPU to optimize performance.
I've said it before, but I'll at least use a different analogy: if you're going to use first-gen DLSS as an indictment of the whole idea, then you probably would have decided that automobiles could never replace horses, based on the first few machines that were built.Even though AMD did not feature any machine learning cores in the chip, their solution to DLSS is simple, just lower resolution based on how much performance you want back, and sharpen the details. When this solution first came about, most reviews did find that the supposedly less elegant AMD solution seem to produce better results both in terms of performance and IQ, as compared to the first gen DLSS.
Okay, so, this issue is truly unrelated to DLSS.Now if you look at this supposed tensor optimized RTX voice, again people can walk around the requirement for the tensor cores and make it work fine, or at least well enough. Perhaps there is some help from the tensor cores, but I don't think its tangible enough.
Wow, so based on the 2080ti TDP (250W - I know TDP is not exactly equal to max power) that's very roughly about 10W power usage. Honestly... Can I say (donning flame proof coat) I don't think it's worth it... Not to say its worthless (and on a normal desktop PC it's ultimately negligible so on or off is irrelevant) I just mean cost vs reward for me is not worth it (efficiency wise its not great) - Certainly if I was working on a laptop on battery it is a significant draw.
Let me put it another way - I expect phones will pick this up soon enough (many now having 'AI' processor cores) and they will need to perform the task using a hell of a lot less power and I expect they will probably pull it off using a much smaller power envelope and energy consumption cost.
Good point.I'd love to know what the trade off really is in terms of power usage / battery runtime on a mobile device
If it does even use the tensor cores, then you should know that their power-efficiency is much closer to that of hard-wired logic than CUDA cores.GPGPU assistance in apps tend to not be low power vs a dedicated IP/DSP block, e.g. NVENC vs CUDA.
GPU power usage doesn't scale lineraly with utilization. At low utilization, GPUs run at lower clocks, which is significantly more efficient.Yeah, the usage is <10% (from user accounts - it may not be that much and falsely reported but until we know better I'll take that as the number), but that could still be around 10 Watts of energy usage, so on a laptop with a mobile RTX GPU with say a 50WHr battery that's a fifth to a quarter of your capacity used up just for that feature.
Could be, but audio chips are cheap and tiny. Adding AI horsepower in the realm of 10% of a GPU (or even less) would add considerable cost to their solution.I suspect Realtek, etc., will eventually integrate such features into their audio codec companion chips relatively easily and probably more efficiently if less general purpose and more tweaked for audio use
GPUs, especially using tensor cores, are vastly more efficient at inferencing than CPUs. Even with specialized VNNI extensions, Intel's AVX-512 CPUs are no match for GPUs.Also it would be interesting (seeing as Discord / MS Teams are looking in to doing it already, not just nvidia) if the overhead on a normal CPU performing the calculation is low enough that it also doesn't actually impact power usage more so than a GPU based approach.
It's probably best to distinguish between AI and Deep Learning. Nvidia tech is legitimately deep learning (which usually requires on the order of a hundred or so iterations to converge - not thousands).'AI' cores - let me know when real 'AI' actually appears that doesn't need thousands of iterations of learning, etc., and can adapt to changes without needing additional 're-training' - until then this is just another 'programmed intelligence' example.
Well, now that same silicon might be used to improve your call quality.it pains me there is a large area of silicon reserved in my phone to essentially make a picture you've taken (where the imaging sensor and processor itself optimised to try and be as accurate and clear as possible despite their sizes) become less accurate, clear, or realistic
: )On a side note: Will be funny if someone uses it whilst doing a youtube or other video review of a keyboard and tries to capture the sound difference between a certain mechnical one vs another or membrane... doh
I would poke around in your mixer settings and make sure that there's not something muted or at low-volume. Sometimes, you have to click around to find mixer settings for different devices...Output is a different story however, since I cannot get it to work in any circumstances. Doesn't matter if I'm using speakers or headphones, or if I'm actually telling it to filter the background noise or not- I get absolutely nothing from the other end (tried with Skype, Discord, and Steam, though I think only the first two out of those three are actually officially supported).
If Nvidia can get the GPU to stay at base clocks, then it should use much less than that.Wow, so based on the 2080ti TDP (250W - I know TDP is not exactly equal to max power) that's very roughly about 10W power usage. Honestly... Can I say (donning flame proof coat) I don't think it's worth it...
Good point.
If it does even use the tensor cores, then you should know that their power-efficiency is much closer to that of hard-wired logic than CUDA cores.
GPU power usage doesn't scale lineraly with utilization. At low utilization, GPUs run at lower clocks, which is significantly more efficient.
Could be, but audio chips are cheap and tiny. Adding AI horsepower in the realm of 10% of a GPU (or even less) would add considerable cost to their solution.
Also, to the point about costs, they tend to be fabbed on much older, less power-efficient manufacturing nodes.
GPUs, especially using tensor cores, are vastly more efficient at inferencing than CPUs. Even with specialized VNNI extensions, Intel's AVX-512 CPUs are no match for GPUs.
So, if your concern is about power-efficiency, then it belongs in a GPU. Whether an iGPU or dGPU, a GPU-based solution will be preferable to a CPU-based one.
It's probably best to distinguish between AI and Deep Learning. Nvidia tech is legitimately deep learning (which usually requires on the order of a hundred or so iterations to converge - not thousands).
Programmed intelligence is something different, and shouldn't be confused with deep learning.
Well, now that same silicon might be used to improve your call quality.
: )
@GenericUser For me to make it to work in the output (to filter the sound of a youtube video by example). In windows I set as output "NVIDIA RTX voice". And in the RTX voice application I choose the output where I want to hear the output ("speaker").
I would poke around in your mixer settings and make sure that there's not something muted or at low-volume. Sometimes, you have to click around to find mixer settings for different devices...