Yes, but don't discount that value. I used to run llama.cpp on the command line because it's amazing that you can ./a-brain. These days I run ollama on the GPU computer and connect to it from my laptop or phone using Zed or Enchanted.
Discussion
No replies yet.