Damus
Ivan · 7w
Good morning, Nostr. Who's running local LLMs? What are the best models that can run at home for coding on a beefy PC system? In 2026, I want to dig into local LLMs more and stop using Claude and Gemi...
John profile picture
24 gb VRAM (3090, 7900 cards): the latest mistral 24b, qwen3 32b and qwen3 30a3 (MoE)

48gb: 70b size models at decent quants, mistral dev large at lobotomized quants. Mistral dev large is the main one in this bracket. There might be other good 70b's released lately

96gb: gpt-oss 120b

This is to fit everything in vram. With MoE's (qwen3 30a3, GPT-oss) you can get by with VRAM+RAM without ruining your speed depending on the speed of your ram.

But it's usually a speed hit so I don't use anything that doesn't fit in VRAM
1
Kajoozie Maflingo · 7w
yup, and about 2tb storage just to fit all those models.
Ivan · 7w
Thank you ๐Ÿ™. Current cards I own 3090 and 7900xtx. Someone suggested dual 3090s. I'll check out the models thank you.