Replying to Avatar ChipTuner

I keep seeing people #vibing projects talking about running out of token, then I realized some of my friends have just been building their apps through the online chat interfaces, with Anthropic or OpenAI directly.

I have some suggestions. Not everyone likes the rate or bleeding edge quality of GitHub Copilot, but its $100/year for "unlimited" usage, plus access to like 10 models in chat, with about 5-6 models available in agent mode, where it basically writes code in your IDE for you. I've hit some rate limits when letting 3 Claude 4 agents cook in the background just rabidly clicking continue while testing this. That was only for a single model, I was able to switch to Gemini and continue for a while longer. On my regular daily usage, I never run into limits.

I've heard similar results with Cursor, who seems to be more on the bleeding edge but you have to use their IDE which is not for me. I also have no idea how the payment system is structured.

If you want to keep using your token based LLMs, you can use Continue.dev or Cline linked up to most of the big providers you already pay for in your IDE. VS Code is the most supported.

Finally you can use Cline or Continue with your own LLM server like ollama, or ollama + owui an the even recommend the models to use, but you'll need some serious hardware to get anywhere near the quality of the paid LLMs. My hardware is a little too dated to really use it full time. I love the privacy but it's not practical for me yet. Once LLM prices go up compared to hardware, I may invest. These two also have far more tuning abilities than Copilot or others IMO, just lack the inlegence.

Continue + Ollama with llama3.3 is definitely usable on a big rig. It is not as good as copilot with claud3.7 or Claude 4 but I like the fact that you can get 90% of the experience without leaving home.

Reply to this note

Please Login to reply.

Discussion

What processor do you have again? I might send some sats your way to test some C compilation things for me.

I9 13900k

Our last conversation had my thinking you already had a thread-ripper. I just remember we were daydreaming XD

Lol, I wish!! I'd love to shred some threads. I have some multicore core for playing with prime numbers. The 13900k is something like 70 times faster than my laptop, but I need more. I need to rewrite for cuda to use my 4090 but haven't got around to it yet.