I used local LLM exclusively, mostly for coding. Two used 24g 3090's which provides 48g of vram. It runs models up to 70b with very fast performance.

When inferring or training,

1. It uses a lot of power, peaking around 800W

2. It spins up the fans pretty loudly

I don't think it's necessary to go local for open source coding though. Maple, mostly gpt-oss-120 is great for that. I think it is necessary to go local for uncensored models or training with your own data, and discussing things that don't fit mainstream bullshit narratives.

Reply to this note

Please Login to reply.

Discussion

I got a 3090 I

might pick another. Thank you!

Dual 3090s is (somehow) still the sweet spot for local