Thread

Zero-JS Hypermedia Browser

Relays: 5
Replies: 20
Generated: 02:20:12
Seriously considering putting two RTX 6000 in my upcoming build machine. Puts the price up an order of magnitude, but truly private AI might be a worthwhile investment. Never played with GPUs before, so informed thoughts welcome?
2025-11-30 07:25:36 from 1 relay(s) 8 replies ↓
Login to reply

Replies (20)

Yes, 96GB is quite ok, the problem is that you have the complexity of two cards, which is not so much fun to setup, you will need to handle this often. It's not a click and run solution, distributed inference even on the same machine adds complexity. Nvidia and Mac are not the only options. There's for example AMD Ryzen AI Max, easily with up to 128GB of shared memory. Framework for example sells these.
2025-11-30 08:01:44 from 1 relay(s) ↑ Parent 1 replies ↓ Reply
There's a pretty big gap between the light models and the heavy ones. Often theres a flagship MoE midel and a light version like GLM-4.5 and GLM-4.5 air 4.5 air fits easily on 96gb but 4.5 full needs 200gb+ just for the model (no context) (quantized to q4) CPU offloading makes them runnable but at like 10T/s which is pretty lame And then there's models like kimi k2 that are >500gb quantized I want a local rig but keep putting it off BC the reqs keep changing Stacking 5090s is nice BC they're 1/4 the price but stacking 6000s is just a nicer system (noise;/power etc)
2025-11-30 12:29:12 from 1 relay(s) ↑ Parent Reply
technically you could get rtx 6000 pro and get 96gb on one card it is true that having two cards is not the same as one big, theres penalty to it and added complexity but also comparing 128gb shared memory vs 96gb vram on gpus is not exactly the same, theres lots to consider in terms of worload and performance, sadly afaik desktop gpus don't have nvlink support anymore so you need a motherboard with good pcie 5.0 support so that you'll have max bw between cards also depends on what you want to do with them, what kinds of models you want to run generally the hardest workload to do locally is coding agents because they tend to like to have a lot of context so you're hitting vram limits very quickly with them and its hard to have comparable performance with small gpus overall - buying great gpu's is always a good idea since the models are getting better and better and the tooling is maturing so two rtx 6000 can give you a lot of bang for the buck. will it perform the same as using opus/gemini3/codex? gonna probably need to bump another order of magnitude for that 😅 you can get decent performance out of rtx 6000 with quantized models according to the internetz
2025-11-30 15:57:59 from 1 relay(s) ↑ Parent 1 replies ↓ Reply
I run a few models on 4090 locally. They are good enough for many tasks but there hallucination patterns are more obvious than for the larger paid models when it goes south. I mean to use langchain and vector databases on my extensive PDF library to see how much more I can get out of it.
2025-12-01 06:01:13 from 1 relay(s) ↑ Parent Reply