local llama3.2 on your phone https://apps.apple.com/us/app/pocketpal-ai/id6502579498

Reply to this note

Please Login to reply.

Discussion

I'm curious how it is on battery life.. have you tried it out?

I just installed it now. I never used it before so I can't comment on battery.

No llama 3.2 1b though... my phone can't probably run 3b one

thanks for sharing πŸ€™

Use gpu instead of cpu? I would think you would want that on πŸ€”

that’s what I was thinking.. what would you recommend for layers on GPU, default 50 or 100

Depends what your device can handle. Think of it like this: the full model might be 4Gb, if your device has 8Gb it might fit in memory, so 100% of the layers can be loaded there (and still have some room for the system and apps and such). But if your device has only 6Gb or 4Gb, the whole model will not fit, so you will need to test if 50% can be loaded into memory, or maybe less. At some point it might not make sense to use the GPU if only too little layers are loaded there, since the overhead of combining CPU + GPU work can predominate. Also, you need free memory space for the context window, so bigger contexts will consume more space and leave less space for layers, while smaller context leaves space for more layers.

This is amazing. My iPhone needs more than 8GB of RAM for sure. Why so stingy Apple?!

Looks like I can cancel my chat GPT subscription now. I don't get why Microsoft building all these data centers when my phone can do the exact same thing..

seems like massive mal investment.

Meh. Get a gaming GPU with 10GB+ VRAM run Ollama + Open WebUI, then set up a wiregaurd VPN to the machine connect to it on your phone and run much more powerful models.

I seem to recall that Ollama has some shady code that is probably harvesting your data.

when android?

Same app is available for android

Anyone tried this? Anygood?