I am trying to run it headless in docker, and failing miserably. help would be nice.
Should i forget docker?
would LXC work?
I have a new micro-project.
It allows to use venice.ai lifetime Pro account with local apps that communicate over the ollama API (open-webui, or continue.dev your VS Code IDE / JetBrains IDE).
Check out the demo video, it's quite nice.
I can also provide a pro account if you don't want to fiddle with MOR tokens.
I started doing this because llama-3.1-405b is really a great model (I think better than ChatGPT for many coding tasks), but I can't run it locally on my laptop.
With this, I have everything set up to make it work locally, with the best open-source model available today.
https://pay.cypherpunk.today/apps/26zEBNn6FGAkzvVVuDMz3SXrKJLU/crowdfund

I am trying to run it headless in docker, and failing miserably. help would be nice.
Should i forget docker?
would LXC work?
I just run it in venv. What about just creating a new user and a python venv?
Headless is default, so that should not be a problem.