Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Suboptimal Performance on Generation Tasks #1353

Closed
erenup opened this issue Jan 25, 2024 · 5 comments
Closed

Suboptimal Performance on Generation Tasks #1353

erenup opened this issue Jan 25, 2024 · 5 comments

Comments

@erenup
Copy link

erenup commented Jan 25, 2024

Hi @haileyschoelkopf

I find one general problem for generation tasks when I am developing the selfcheckgpt task.

The problem is that:

  1. For each task, we can set one prompt format in task.yaml. But different models have their own system prompts or starting token. So one task can not be compared across multiple models.
  2. For each task, we can set stopping tokens in the {"until": ["\n\n"}. But different models have their own ending tokens. So the model will generate something like this:
    image

I think this is a general problem and more people who are using this framework should be aware of it when they are trying to compare different models in on generation task.

Thank you very much.

@StellaAthena
Copy link
Member

We are currently working on supporting HF Chat Templating, see #1287

@haileyschoelkopf
Copy link
Contributor

Thanks @erenup for the feedback! With respect to the latter point--we currently support the passing of generation hyperparameters via the --gen_kwargs CLI flag. I can see about adding the ability to provide extra stop sequences via the CLI using this flag.

@erenup
Copy link
Author

erenup commented Jan 26, 2024

Thank you very much for the quick reply!

@erenup
Copy link
Author

erenup commented Feb 17, 2024

hi @haileyschoelkopf

I think this could be a potential way to fix the huggingface.py model generation performance with apply_chat_template:
add self.tokenizer.apply_chat_template(messages, tokenize=False) to tok_batch_encode function in huggingface.py
image

How do you think about it?

@LSinev
Copy link
Contributor

LSinev commented Jun 13, 2024

Is this still an issue after #1873?

@erenup erenup closed this as completed Jun 13, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants