In general how much VRAM do I need for 14B and 24B models?
marauding_gibberish142
I didn't know that. I thought just one ROCM binary to install, run Ollama and that's it. Thanks for the explanation
Do you have any recommendations for running the Mistral small model? I'm very interested in it alongside CodeLlama, OogaBooga and others
Wait how does that work? How is 24GB enough for a 38B model?
The 7900XTX was $1000 when it launched, I wouldn't mind it used either.
I don't mind multiple GPUs but my motherboard doesn't have 2+ electrically connected X16 slots. I could build a new homeserver (I've been thinking about it) but consumer platforms simply don't have the PCIE lanes for 2 actual x16 slots. I'd have to go back to Broadwell Xeons for that, which are really power hungry. Oh well, I don't think it matters considering how power hungry GPUs are now.
I am OK with either Nvidia or AMD especially if Ollama supports it. With that said I have heard that AMD takes some manual effort whilst Nvidia is easier. Depends on how difficult ROCM is
Thank you. Are 14B models the biggest you can run comfortably?
Do you have 2 PCIE X16 slots on your motherboard (speaking in terms of electrical connections)?
Seedboxes go from €2 to €100+ a month depending on how much you will torrent and how much space you need on the box alongside other factors. My personal choices are Gigarapid and Ultra but there are others
What are you talking about mate? The VPS will be a wireguard server, your device will connect to it and use the VPS' IP address to connect to your game
I have an alternative for you if your power bills are cheap: X99 motherboard + CPU combos from China