My use case is I have installed LLMs in my GPU using the method described in this Reddit Post. I want to run WizardLM-30B, which requires 27GB of RAM. (I also have a quantized version with lower RAM requirements, but I do not have a number on the RAM requirements. I just know that I can fit 40/61 layers on my 16GB GPU, implying I would need 25GB to run it entirely on GPU). I cannot currently run the LLM entirely on my ARC A770 16GB GPU, forcing the rest of it to use system ram and the CPU, slowing the model down considerably. I would like to have a GPU with enough memory to handle the model itself, however, all the GPUs that I have found with more than 24GB are extremely expensive. The 24GB Radeon RX 7900 XTX is a thousand dollars, but I haven't been able to find any GPUs with more memory that are even double that price.
Of course, this entire issue could be avoided if there is a way to split the LLM up among multiple GPUs, then I could just get a second A770. The A770 seems to have a very good value of VRAM per Dollar. Unfortunately, I have not been able to find any resources for this. This post I have found with my exact issue does not say that it is possible https://datascience.stackexchange.com/questions/121639/load-an-llm-in-multiple-gpus )
Here is a thread about 24GB GPUs,
https://forums.tomshardware.com/thr...n-gaming-graphics-card.3806569/#post-22998565
Here is my previous thread on this topic, before I refined my requirements
Does anyone know an affordable GPU with at least 27GB of VRAM?
Of course, this entire issue could be avoided if there is a way to split the LLM up among multiple GPUs, then I could just get a second A770. The A770 seems to have a very good value of VRAM per Dollar. Unfortunately, I have not been able to find any resources for this. This post I have found with my exact issue does not say that it is possible https://datascience.stackexchange.com/questions/121639/load-an-llm-in-multiple-gpus )
Here is a thread about 24GB GPUs,
https://forums.tomshardware.com/thr...n-gaming-graphics-card.3806569/#post-22998565
Here is my previous thread on this topic, before I refined my requirements
Question - How to shop for GPUs (or other hardware) for LLM Workloads ?
Hello, I would appreciate some guidance on what hardware (GPU or otherwise) I should purchase to enable me to run LLMs locally on my machine. Here are my system specs. CPU: AMD Ryzen 9 7950X Motherboard: ASRock X670E PG Lightning AM5 ATX Mainboard. RAM...
forums.tomshardware.com
Does anyone know an affordable GPU with at least 27GB of VRAM?