Pro@programming.dev to Technology@lemmy.worldEnglish · 1 month agoGoogle quietly released an app that lets you download and run AI models locallygithub.comexternal-linkmessage-square44linkfedilinkarrow-up11arrow-down10cross-posted to: localllama@sh.itjust.works
arrow-up11arrow-down1external-linkGoogle quietly released an app that lets you download and run AI models locallygithub.comPro@programming.dev to Technology@lemmy.worldEnglish · 1 month agomessage-square44linkfedilinkcross-posted to: localllama@sh.itjust.works
minus-squareAmbiguousProps@lemmy.todaylinkfedilinkEnglisharrow-up0·1 month agoThat’s fair, but I think I’d rather self host an Ollama server and connect to it with an Android client in that case. Much better performance.
minus-squareGreg Clarke@lemmy.calinkfedilinkEnglisharrow-up0·1 month agoYes, that’s my setup. But this will be useful for cases where internet connection is not reliable
minus-squareOhVenus_Baby@lemmy.mllinkfedilinkEnglisharrow-up0·1 month agoHow is Ollama compared to GPT models? I used the paid tier for work and I’m curious how this stacks up.
minus-squareGreg Clarke@lemmy.calinkfedilinkEnglisharrow-up0·1 month agoHas this actually been done? If so, I assume it would only be able to use the CPU
minus-squareEuphoma@lemmy.mllinkfedilinkEnglisharrow-up0·1 month agoYeah I have it in termux. Ollama is in the package repos for termux. The speed it generates does feel like cpu speed but idk
minus-squareDiplomjodler@lemmy.worldlinkfedilinkEnglisharrow-up0·1 month agoIs there any useful model you can run on a phone?
minus-squaregens@programming.devlinkfedilinkEnglisharrow-up0·1 month agoLlama.cpp (on which ollama runs on) can. And many chat programs for phones can use it.
Ollama can’t run on Android
That’s fair, but I think I’d rather self host an Ollama server and connect to it with an Android client in that case. Much better performance.
Yes, that’s my setup. But this will be useful for cases where internet connection is not reliable
How is Ollama compared to GPT models? I used the paid tier for work and I’m curious how this stacks up.
You can use it in termux
Has this actually been done? If so, I assume it would only be able to use the CPU
Yeah I have it in termux. Ollama is in the package repos for termux. The speed it generates does feel like cpu speed but idk
Try PocketPal instead
Is there any useful model you can run on a phone?
Llama.cpp (on which ollama runs on) can. And many chat programs for phones can use it.