Juraj-AI doing well.

https://nostrcheck.me/media/public/e6832c71599efd1469747b07f9e3770cf93c89caa1d0608389a10a973736d52d.webp

Reply to this note

Please Login to reply.

Discussion

Cool. How did you train the model? (tools, method, etc.)

How do I create my own?

I know how to run a model locally, using ollama.

Can you please share any good resources on how to create my own model based on my texts?

Still experimenting, I'm currently using llama.cpp including their fine tuning scripts. I have changed all the frameworks several times.

I'm programmatically creating question and answer data (it helps that I've written several books) by prompting a larger model (currently nous-hermes-llama-2-70B) and then some cleaning of the data, converting to alpaca format and finetuning (I currently fine tune the 13B version of the same model).

You need at least Apple Silicon hardware with lots of ram or a few nvidia gpus (vram more important than speed). Amd has shitty tooling.

I use both several nvidia cards and apple silicon. Nvidia and CUDA was probably forged in hell, Apple ecosystem is not so mature, but growing rapidly.

Thanks a lot for sharing.

Impressive.

Precious skill... I'd like to try and learn one day... But hardware requirements nontrivial...