Other option is Ollama with many models available. No internet access. I use it on Linux and I find it useful.
Discussion
Ollama is extremely hostile to the llamacpp ecosystem so I don't recommend it. KoboldCpp has some ollama emulation on board. Ollama day one has been trying to create a walled garden by hijacking upstream model efforts, making a closed model repo and attempting to force a custom API trough.