It's still not magical. I also think the reasoning part is wasted on goose since there's already a feedback loop

Reply to this note

Please Login to reply.

Discussion

Mistral small just dropped, and the 22B is supposed to support function calling

whats the best local model to use ?

Nothing yet. They're getting very close this month

I run 30 and 40B on my rtx 3090ti on ollama. Most of them run fine, as fast as chat gpt. Gonna need big improvements to go larger though even with 24gb vram. 72B is out of the question currently.

I started with a P40, then added a 3090. 48GB enough to run 70B models, but it might be time to add a 4090 as well.