-
Notifications
You must be signed in to change notification settings - Fork 966
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GGML GPT-2 consistently dies at around 825 tokens with: ggml_new_object: not enough space in the context's memory pool #480
Comments
You can try increasing the buffer size here: Line 408 in 1a5d5f3
If you want a more reliable solution for different context sizes, you can use the allocator in |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Firstly, thanks to GG and contributors for a great library/utility.
When generating using gpt-2, ggml bombs out at around 824 or 825 tokens, reporting an error then dumping core.
I would expect there to be a problem (hopefully not involving fatal errors and core dumps) when the total tokens equal the context size, but 824 or 825 total seems an odd number?
The same error is referenced in the llama.cpp repo, but possibly for a different reason: ggerganov/llama.cpp#2404
REPRODUCE:
Clean build, CPU only, Ubuntu 22:
git pull && rm -Rf build && mkdir build && cd build && cmake .. && make
with ggml-model-f16.bin (gpt2-xl), eg
bin/gpt-2 -m ~/gpt-2/models/1558M/ggml-model-f32.bin -n ...
-n 823: ok (run completes without error)
-n 824: ggml_new_object: not enough space in the context's memory pool (needed 268457104, available 268435456)
-n 825: ggml_new_object: not enough space in the context's memory pool (needed 268457104, available 268435456)
with ggml-model-f32.bin (gpt2-xl):
-n 823: ok
-n 824: ok
-n 825: ggml_new_object: not enough space in the context's memory pool (needed 268457104, available 268435456)
Note: I had to repeat some runs several times as ggml will stop prematurely if an <|endoftext|> token is generated. Getting to 823+ tokens can take a few tries.
The text was updated successfully, but these errors were encountered: