Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About performance of LLama (llama2, llama3) model #215

Open
huazhenliu opened this issue Apr 25, 2024 · 1 comment
Open

About performance of LLama (llama2, llama3) model #215

huazhenliu opened this issue Apr 25, 2024 · 1 comment
Labels
question Further information is requested

Comments

@huazhenliu
Copy link

Thank you for your wonderful work!

Have you ever experimented with LLama2-7B as the model to do C-RLFT? How about the performance? Because OpenChat-3.5-0106 is based on Mistral, performance is really high, I have tried using LLama2-7B, the performance is not satisfied.

Another 2 questions: can chat model be used as the model to do C-RLFT? I think, some code needs to be done, e.g., chat template, etc.
How about LLama3-8B-instuct, how to easy train, any performance data?

Thanks in advance.

@imoneoi
Copy link
Owner

imoneoi commented Apr 26, 2024

Hi @huazhenliu We've tried Llama 2 13B, the performance is worse than Mistral 7B, so we've chosen Mistral 7B as the base model.

@imoneoi imoneoi added the question Further information is requested label Apr 26, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants