-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can't load Llama Model #1158
Comments
Hi @himanshud2611 , Could you share any relevant code snippets and/or stack traces? Seems like the error message you get from TGI is from not setting the model name correctly, while the Anyscale error message has a configured endpoint but potentially incorrect arguments passed in. |
Hi @arnavsinghvi11
At this point in time, I want to use Language Model API Client [lm = dspy.HFClientTGI(model="meta-llama/Llama-2-7b-hf", port=8080, url="https://localhost")] with same model name. Can you help me with setting ports and url here - how can I setup TGI server locally, help me with docker part. |
Can't you just run the 4 lines from the screenshot you shared? |
Hi @tom-doerr |
I don't think it makes sense for us to talk you through the fundamentals of docker, there's a lot to know. |
Thanks :) |
I've tried using Ollama to load Llama and issue is resolved |
I've worked with OpenAI and Mistral previously, the DSPy pipeline executed well. I've tried loading Llama-2-7b using HuggingFace and HFClient TGI but getting errors "No LM loaded" and Connection Error. I've also tried using Anyscale to load Llama-2-7b but got "Received invalid JSON response from server" error. Can anyone help the optimal way to Llama-2-7b to execute DSPy pipeline.
The text was updated successfully, but these errors were encountered: