News AMD released instructions for running DeepSeek on Ryzen AI CPUs and Radeon GPUs

I'm running on a desktop and a mini pc. The desktop has a 7700x, 64gb RAM, AND A7800XT. The mini pc has a 8845hs, 64gb RAM, and 780m internal gasoline graphics. Running Ollama in both dual boot. I get better a litlle inference performance on Ubuntu.
 
AMD taking advantage of Nvidia's moment of weakness.
Nvidia is in serious trouble when it comes to AI Model execution. Both Apple & AMD are offering compute platforms with up to 128GB of RAM that can execute VERY LARGE AI models. NVidia cannot touch the price/performance of these machines and apparently they have no plans to create a competing product anytime soon. It's for this reason that I bought my son a 48GB MacBook Pro M4Pro laptop - the ability to run larger AI models.

This weakness in NVidia hardware is also causing Mac Mini sales to skyrocket because you can put 64GB of RAM into an M4Pro model and run 64GB models that the 5090 will NEVER run for $2699.
 
Nvidia is in serious trouble when it comes to AI Model execution. Both Apple & AMD are offering compute platforms with up to 128GB of RAM that can execute VERY LARGE AI models. NVidia cannot touch the price/performance of these machines and apparently they have no plans to create a competing product anytime soon. It's for this reason that I bought my son a 48GB MacBook Pro M4Pro laptop - the ability to run larger AI models.

This weakness in NVidia hardware is also causing Mac Mini sales to skyrocket because you can put 64GB of RAM into an M4Pro model and run 64GB models that the 5090 will NEVER run for $2699.
I wanted to buy a 128 GB Strix Halo mini-PC in the coming months... but I'm afraid that with DeepSeek coming out, all of those Strix Halo will end up in hands of AI people.
 
You got it backwards or perhaps didn't really understand the article. The privacy issues apply to their apps, website and other products that links to the privacy policy. Honestly every AI company collects similar load of information, just not sending to China if that matters to you.

However, they distributed their code and weights with MIT license. What we have here is a local setup that can be run entirely offline, which truly eliminates the problem. If privacy is your concern, running open models locally is the only way to go and that's what this article is about.
 
Nvidia is in serious trouble when it comes to AI Model execution. Both Apple & AMD are offering compute platforms with up to 128GB of RAM that can execute VERY LARGE AI models. NVidia cannot touch the price/performance of these machines and apparently they have no plans to create a competing product anytime soon. It's for this reason that I bought my son a 48GB MacBook Pro M4Pro laptop - the ability to run larger AI models.

This weakness in NVidia hardware is also causing Mac Mini sales to skyrocket because you can put 64GB of RAM into an M4Pro model and run 64GB models that the 5090 will NEVER run for $2699.
https://www.tomshardware.com/pc-com...chip-reportedly-ready-for-tape-out-this-month