Self host by copying the code and running it?? Via github?????

Reply to this note

Please Login to reply.

Discussion

Depending on your environment, you can get it up and running via Docker. Install Ollama and then setup Open WebUI which is a pretty good interface for managing models. Getting it running on windows was actually really easy when I first tested it there.

https://github.com/open-webui/open-webui

Without a GPU it is slow but still usable depending on the task, and you'll need at least 8-64GB of RAM depending on the model (llama 3 needs 48GB).

I have it running with a single RTX 3090 and the deepseek and mistral models are blazing fast, at least to me.