Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Ollama support #18

Open
ThatOneCalculator opened this issue Jul 7, 2024 · 7 comments
Open

Ollama support #18

ThatOneCalculator opened this issue Jul 7, 2024 · 7 comments
Labels
enhancement New feature or request

Comments

@ThatOneCalculator
Copy link

Support for self-hosted Ollama servers would be great

@Taewan-P
Copy link
Owner

Taewan-P commented Jul 8, 2024

Do they have a unified output format? Any API references?

@ThatOneCalculator
Copy link
Author

@Taewan-P Taewan-P added the enhancement New feature or request label Jul 8, 2024
@randoentity
Copy link

Wouldn't it make more sense to just use the openai api spec? That adds support for many inference servers, including ollama (didn't they also just use the openai api spec?).

@Taewan-P
Copy link
Owner

@randoentity Wow, after seeing your comment I tried searching for that... and surprisingly it is (link)
This will make it more easy to implement the feature. Thanks! 😃

@ThatOneCalculator
Copy link
Author

I would recommend against using the OpenAI API layer, as it's still WIP and missing many features.

@Taewan-P
Copy link
Owner

There is not really much feature we need rn. If chat completion works, I think that's good for now.

@ThatOneCalculator
Copy link
Author

Ok, the chat completion works so as long as remote models can be listed with the regular Ollama API, I don't see an issue. It would be nice to have vision/etc in the future though

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants