Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Long time testing Qwen2-72B #1984

Open
djstrong opened this issue Jun 18, 2024 · 2 comments
Open

Long time testing Qwen2-72B #1984

djstrong opened this issue Jun 18, 2024 · 2 comments
Labels
bug Something isn't working.

Comments

@djstrong
Copy link
Contributor

djstrong commented Jun 18, 2024

When testing Qwen/Qwen2-72B parallelize=True,max_length=4096 on generate_until task I am getting warnings:

Both `max_new_tokens` (=2048) and `max_length`(=882) seem to have been set. `max_new_tokens` will take precedence. Please refer to the documentation for more information. (https://huggingface.co/docs/transformers/main/en/main_classes/text_generation)

Number after max_length is getting lower in time.

The problem is the test is taking few times more time than e.g. using llama3-70B. In the task there is max_gen_toks: 50.

Is the long time related to the warning above?

I see that this model has "max_new_tokens": 2048 in generation_config.json, but llama3 does not have.

@haileyschoelkopf
Copy link
Contributor

There have been a couple reports of odd scores or behavior when evaluating Qwen2 models. I hope to check this out soonish, if others aren't able to.

@haileyschoelkopf haileyschoelkopf added the bug Something isn't working. label Jun 19, 2024
@baberabb
Copy link
Contributor

Looks like max_new_tokens=2048 is taking precedence over max_tokens(len(inputs) + max_gen_toks)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working.
Projects
None yet
Development

No branches or pull requests

3 participants