3B q4_0 open llama model running on a 4gb Pixel #1667
BarfingLemurs
started this conversation in
Show and tell
Replies: 1 comment 1 reply
-
glorious. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I just wanted to share that via termux 4gb of ram is enough to run this model, taking 1.9 gb of ram.
My hope is that I can use an older phone as a chatbot, using the
talk-llama
example found in https://github.com/ggerganov/whisper.cpp. I could leave my the phone plugged 24/7 only using 1.5W of power. Currently termux wont detect any audio, but you can run the talk-llama using 3b model on a pc, it's fast.I also tried other 3b models running with an interface on kobold.cpp, it crashes due to the extra ram needed to load the interface.
Beta Was this translation helpful? Give feedback.
All reactions