-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[QUESTION] How can I convert checkpoint tuned by EE-Tuning to Huggingface format? #15
Comments
Thanks for your information! I am also curious about how I can reproduce the results demonstrated in your paper and perform the downstream evaluation on the HELM benchmark. Could you please provide more details on this? |
We modified the MegatronClient, adding parameters related to EE-LLM. All other parts are directly inherited from HELM. |
Actually, I'm not familiar with HELM. Could you provide some demos or guidance on how to use the script MegatronClient? |
You can refer to the demo in data-juicer. |
Really appreciate for your help! |
Hello again! I have tried the evaluation framework proposed in data-juicer and get some benchmark results, such as ROUGE-2 in CNN/DM, F1 score in NarrativeQA, and EM in MMLU. However, I'm confused about how can I get efficiency results like inference time throughout the generation process. What should I modify in mymodel_example.yaml to parse the corresponding metric from HELM output? I would greatly appreciate your help and look forward to your prompt response. |
If you use the HELM provided by Data-Juicer, you can modify
The In addition, you also need to modify your megatron_client.py to return the new metric in your response. For example,
Helm will use the Note that the demo script provided by Data-Juicer is not for EE models, it only records some metrics for pretraining. |
Your client must return those metrics in response, and then HELM can summarize them, so you need to modify your client first, as shown in my previous comment. And in our paper's experiment, we set |
Thanks! |
Marking as stale. No activity in 60 days. |
I have fine-tuned the llama-7b model using EE-Tuning, and I now need to convert the checkpoint to the Hugging Face format to proceed with the evaluation process. How should I do this?
The text was updated successfully, but these errors were encountered: