Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batch Inference using Ray and vLLM #680

Open
ratnopamc opened this issue Oct 23, 2024 · 2 comments
Open

Batch Inference using Ray and vLLM #680

ratnopamc opened this issue Oct 23, 2024 · 2 comments
Assignees
Labels
enhancement New feature or request

Comments

@ratnopamc
Copy link
Collaborator

Community Note

  • Please vote on this issue by adding a 👍 reaction to the original issue to help the community and maintainers prioritize this request
  • Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
  • If you are interested in working on this issue or have submitted a pull request, please leave a comment

What is the outcome that you are trying to reach?

Current examples mostly showcase online inference.
We need to add example of batch inference on Ray with vLLM.

Describe the solution you would like

Add example of batch inference using RayJob under the JARK-stack blueprint (for GPUs).
Refer to Ray Documentation for an example.

Describe alternatives you have considered

Additional context

@askulkarni2 askulkarni2 added the enhancement New feature or request label Oct 23, 2024
@thangalv
Copy link

Interested in working on this issue

@ratnopamc
Copy link
Collaborator Author

Thanks @thangalv, assigned to you. Once you're done testing your changes, please raise a PR.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants