-
Notifications
You must be signed in to change notification settings - Fork 959
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use custom GPT-J checkpoint #488
Comments
Hi, have you tried out on your end? What I did is download the pytorch file, which is approx 24G from the link you shared. I used one of the older convert-h5-to-ggml.py (https://github.com/ggerganov/ggml/blob/master/examples/gpt-j/convert-h5-to-ggml.py), the newer gguf convert tool produce int error. I was able to run without problems. However, because my machine is only 16G RAM, I run out of memory and the job is killed. I'm pretty positive that it will work if you have a machine with more than 24G RAM. Please keep me posted on the status of the conversion and the status of the MLPerf testing. I'm interested in the MLPerf too as I am going to run similar. Either post back here or DM me privately. See below
root@master:~/oldllama.cpp/models/gpt-j/checkpoint-final# python3 convert.py ./ 1 pytorch_model-00001-of-00003.bin |
I would like to run the
ggml/gpt-j
version on the MLPerf benchmark. Is it possible to use a fine-tuned GPT-J checkpoint listed here: https://github.com/mlcommons/inference/blob/master/language/gpt-j/README.md#download-gpt-j-model? The pre-trained version used in MLPerf isEleutherAI/gpt-j-6B
which is the same as what is used in ggml.The text was updated successfully, but these errors were encountered: