Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Google Gemma #214

Closed
surak opened this issue Feb 21, 2024 · 2 comments · Fixed by #256
Closed

Google Gemma #214

surak opened this issue Feb 21, 2024 · 2 comments · Fixed by #256

Comments

@surak
Copy link

surak commented Feb 21, 2024

Google released their new open model, Gemma. https://huggingface.co/google/gemma-7b-it

But we can't use it on fastchat with sglang yet:

File "/p/haicluster/llama/FastChat/sc_venv_sglang/sglang/python/sglang/srt/managers/router/model_runner.py", line 49, in get_model_cls_by_arch_name
2024-02-21 15:21:34 | INFO | stdout |     raise ValueError(
2024-02-21 15:21:34 | INFO | stdout | ValueError: Unsupported architectures: GemmaForCausalLM. Supported list: ['QWenLMHeadModel', 'Qwen2ForCausalLM', 'LlavaLlamaForCausalLM', 'MixtralForCausalLM', 'MistralForCausalLM', 'YiVLForCausalLM', 'LlamaForCausalLM']
@nopepper
Copy link

👍 I'd also be interested in this. I don't know how hard it would be to add support, though.

@Arcmoon-Hu
Copy link
Contributor

Apparently sglang doesn't support this model yet. May be you can create a branch support it.

@hnyls2002 hnyls2002 mentioned this issue Mar 4, 2024
@hnyls2002 hnyls2002 linked a pull request Mar 4, 2024 that will close this issue
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants