Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support InternLM #362

Open
monatis opened this issue Jul 9, 2023 · 0 comments
Open

Support InternLM #362

monatis opened this issue Jul 9, 2023 · 0 comments
Labels
model Model specific

Comments

@monatis
Copy link
Contributor

monatis commented Jul 9, 2023

InternLM opensourced a new 7B-parameter model, and their published evaluation scores seem to be very promising.

According to their readme:

InternLM has open-sourced a 7 billion parameter base model and a chat model tailored for practical scenarios. The model has the following characteristics:

  • It leverages trillions of high-quality tokens for training to establish a powerful knowledge base.
  • It supports an 8k context window length, enabling longer input sequences and stronger reasoning capabilities.

A quick look at their source code reveals that it's based on GPTNeoX, saying the code is modified to accomodate minor architectural changes. It should be fairly easy to support InternLM with little to no modifications to the GPTNeoX example here. I'm mobile now, but I'll give it a try to it if anyone else has not yet.

@ggerganov ggerganov added the model Model specific label Jul 14, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
model Model specific
Projects
None yet
Development

No branches or pull requests

2 participants