Running out of buffer for large models #361
Unanswered
MariusGalyan
asked this question in
Q&A
Replies: 1 comment 1 reply
-
I have the same problem and openend an Issue here: ggml-org/llama.cpp#1815 |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I am getting this error message running an 18GB large 30B parameter llm model:
buffer 'data' size 18300780544 is larger than buffer maximum of 17179869184
llama_init_from_file: failed to add buffer
Other llm models with smaller sizes run without any problems. I have a Mac Metal...
Beta Was this translation helpful? Give feedback.
All reactions