Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix: Set model_max_length in the Tokenizer of DefaultPromptHandler #5596

Merged
merged 2 commits into from
Sep 1, 2023

Conversation

bogdankostic
Copy link
Contributor

Related Issues

Proposed Changes:

This PR sets the model_max_length parameter of the tokenizer of DefaultPromptHandler.

How did you test it?

I added a unit test + added a test case in an integration test.

Notes for the reviewer

Without this change, users are getting a warning message from the transformers library that the sequence length is too long for the model they are using in case they are using models supporting larger sequence length than 1024.

Checklist

@bogdankostic bogdankostic requested review from a team as code owners August 18, 2023 13:15
@bogdankostic bogdankostic requested review from dfokina and masci and removed request for a team August 18, 2023 13:15
@coveralls
Copy link
Collaborator

Pull Request Test Coverage Report for Build 5903109928

  • 0 of 0 changed or added relevant lines in 0 files are covered.
  • 1 unchanged line in 1 file lost coverage.
  • Overall coverage decreased (-0.002%) to 48.257%

Files with Coverage Reduction New Missed Lines %
utils/context_matching.py 1 95.7%
Totals Coverage Status
Change from base Build 5901803206: -0.002%
Covered Lines: 11460
Relevant Lines: 23748

💛 - Coveralls

@bogdankostic bogdankostic merged commit 1144039 into main Sep 1, 2023
55 checks passed
@bogdankostic bogdankostic deleted the model_max_length_prompthandler branch September 1, 2023 09:48
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
3 participants