Replies: 2 comments 2 replies
-
@idekterev most of the time, you need 6-8gb of ram to run 7b on android, there is a way to run 3b and 1b models using the ggml repo, that is pending, which may reduce ram requirements to 3-4gb ram for 4bit models. https://huggingface.co/replit/replit-code-v1-3b these are some models that could work in the near future |
Beta Was this translation helpful? Give feedback.
2 replies
-
check out my comment here #9318 (comment) |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
New to the space - apologies in advance for the possibly noob question.
Wondering at the possibility of running a quantized 7B model on a 32-bit arm (armeabi-v7a).
Is is possible to compile this for armeabi-v7a?
Beta Was this translation helpful? Give feedback.
All reactions