I like the idea of using a local 8B LLM on my laptop.

What’s the best way to add a web search layer on top of a model running f.e. in Ollama? Are there any standard good solutions?

I tried using RAG (Retrieval-Augmented Generation) myself, but it didn’t work as expected. Even using a separate embedded context LLM beforehand didn’t have the desired effect.

Reply to this note

Please Login to reply.

Discussion

Ollama? greatest name choice. in math: yama-scale (O(Alpaka)).

The Lama/Alpaka are there for wool and stuff. If they spit in one's face one is medium length as an adult.

For Germans this name fits.