-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Long time testing Qwen2-72B #1984
Labels
bug
Something isn't working.
Comments
There have been a couple reports of odd scores or behavior when evaluating Qwen2 models. I hope to check this out soonish, if others aren't able to. |
Looks like |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
When testing Qwen/Qwen2-72B
parallelize=True,max_length=4096
ongenerate_until
task I am getting warnings:Number after
max_length
is getting lower in time.The problem is the test is taking few times more time than e.g. using llama3-70B. In the task there is
max_gen_toks: 50
.Is the long time related to the warning above?
I see that this model has
"max_new_tokens": 2048
in generation_config.json, but llama3 does not have.The text was updated successfully, but these errors were encountered: