Starting to run LLMs on my own hardware. These model files are often 4-40GB in size.
Where are the torrents? #asknostr
Starting to run LLMs on my own hardware. These model files are often 4-40GB in size.
Where are the torrents? #asknostr
Have you tried smaller models like Ollama for specific tasks?
Yeah, Iβve got llama 3 running via ollama. Downloading the bigger version now. Should be able to run it once my new RAM arrives.
But mostly, Iβm thinking about going forward. I plan to download and try out a whole bunch of different models all of which are extraordinarily large data files. This seems like a perfect job for BitTorrent.
Is that what iβd need to make a chat bot of myself?
I believe you could do this with ollama, yes. It supports saving and loading models, so you could feed it a bunch of info, then begin the bot serving from that point.
Which LLM are you running? Mine is ridiculously slow
Right now, Iβm running Mistral through ollama. Takes a while to get responses back. My machine is a 6-core Intel-based 2018 era Alienware gaming PC with 16GB RAM.
Iβve got 64GB RAM arriving today. I donβt expect it to help with performance, but it should enable me to run bigger models.
Youβre the one who is slow. π
ππ