-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
truthfulqa_mc2 is Nan, while truthfulqa_mc1 is 1.00 #714
Comments
I have same issue!But I have done some operation to change or move the lora weight in my code.
|
This issue should be solved in the |
@lintangsutawika I used the main branch and the issue is still there |
@lintangsutawika How to fix it? Can you share the PR? Thanks |
@choco9966 can you share a public model + sample command that reproduces this issue? |
I have finetued a model based on llama-2-hf, and run the evaluation with code and get truthfulqa_mc2 is Nan, while truthfulqa_mc1 is 1.00.
What does that means?
python main.py --model hf-causal-experimental --model_args pretrained=../mamba-gpt-7b-v2 --tasks anli_r1,anli_r2,anli_r3,arc_challenge,arc_easy,boolq,hellaswag,openbookqa,piqa,record,rte,truthfulqa_mc,wic,winogrande --device cuda:0
hf-causal-experimental (pretrained=../mamba-gpt-7b-v2), limit: None, provide_description: False, num_fewshot: 0, batch_size: None
The text was updated successfully, but these errors were encountered: