does deepseek-r1 work with goose? I'm about to try.
Discussion
dang didn't work
It's still not magical. I also think the reasoning part is wasted on goose since there's already a feedback loop
Mistral small just dropped, and the 22B is supposed to support function calling
whats the best local model to use ?
Nothing yet. They're getting very close this month
I run 30 and 40B on my rtx 3090ti on ollama. Most of them run fine, as fast as chat gpt. Gonna need big improvements to go larger though even with 24gb vram. 72B is out of the question currently.
I started with a P40, then added a 3090. 48GB enough to run 70B models, but it might be time to add a 4090 as well.
Some said that this is a fine-tuned version of Llama, not the actual deepseek model.
goose manual seems to explain how it works with deepseek-r1.
https://block.github.io/goose/docs/getting-started/using-goose-free/#deepseek-r1