Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q4 2024
#9006 opened Oct 1, 2024 by simon-mo
Open 4
vLLM's V2 Engine Architecture
#8779 opened Sep 24, 2024 by simon-mo
Open 7
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[Bug]: vllm mistralai--Codestral-22B-v0.1 response is truncated bug Something isn't working
#9310 opened Oct 12, 2024 by Fly-Pluche
1 task done
[Bug]: latest docker build (0.6.2) got error due to VLLM_MAX_SIZE_MB bug Something isn't working
#9307 opened Oct 12, 2024 by ZJLi2013
1 task done
[Installation]: vllm installation error installation Installation problems
#9305 opened Oct 12, 2024 by leoneyar
1 task done
[New Model]: meta-llama/Llama-Guard-3-1B help wanted Extra attention is needed new model Requests to new models
#9294 opened Oct 11, 2024 by ayeganov
1 task done
[Bug]: Quantization example outdated (Ammo -> ModelOpt) bug Something isn't working
#9288 opened Oct 11, 2024 by kevalmorabia97
1 task done
[Bug]: Simultaneous mm calls lead to permanently degraded performance. bug Something isn't working
#9283 opened Oct 11, 2024 by SeanIsYoung
1 task done
[Bug]: MiniCPM3-4B is support lora by --enable-lora ? bug Something isn't working
#9282 opened Oct 11, 2024 by ML-GCN
1 task done
[Bug]: VLLM doesn't support LoRa with config modules_to_save bug Something isn't working
#9280 opened Oct 11, 2024 by fahadh4ilyas
1 task done
[Usage]: Manually Increasing inference time usage How to use vllm
#9274 opened Oct 11, 2024 by Playerrrrr
1 task done
[Usage]: blip2 inference code usage How to use vllm
#9270 opened Oct 11, 2024 by zhaoxueqi6666
1 task done
[Bug]: AsyncLLMEngine stuck on a single too long request bug Something isn't working
#9263 opened Oct 10, 2024 by rickyyx
1 task done
[Bug]: Streaming response fails after one token (0.5.3.post1) bug Something isn't working
#9260 opened Oct 10, 2024 by NeonDaniel
1 task done
[Usage]: running gated models offline usage How to use vllm
#9255 opened Oct 10, 2024 by SamuelBG13
[Bug]: new beam search implementation ignores stop conditions bug Something isn't working
#9253 opened Oct 10, 2024 by nFunctor
1 task done
[Installation]: build error from source on 4090 installation Installation problems
#9246 opened Oct 10, 2024 by ltm920716
1 task done
ProTip! Exclude everything labeled bug with -label:bug.