Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q4 2024
#9006 opened Oct 1, 2024 by simon-mo
Open 26
vLLM's V1 Engine Architecture
#8779 opened Sep 24, 2024 by simon-mo
Open 10
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[Bug]: Prefix caching doesn't work for LlavaOneVision bug Something isn't working
#11371 opened Dec 20, 2024 by sleepwalker2017
1 task done
[Bug]: The following fields were present in the request but ignored: {'schema'} bug Something isn't working
#11363 opened Dec 20, 2024 by Quang-elec44
1 task done
[Bug]: vllm 0.6.3.post1 crash when deploy qwen2vl 72b bug Something isn't working
#11356 opened Dec 20, 2024 by xxlight
1 task done
[New Model]: answerdotai/ModernBERT-large new model Requests to new models
#11347 opened Dec 19, 2024 by pooyadavoodi
1 task
[Performance]: 1P1D Disaggregation performance performance Performance-related issues
#11345 opened Dec 19, 2024 by Jeffwan
1 task done
[Bug]: Paligemma 2 model loading error bug Something isn't working
#11343 opened Dec 19, 2024 by mmderakhshani
1 task done
[Bug]: Multi-Node CPU Inference on MacOS calling intel_extension_for_pytorch bug Something isn't working
#11342 opened Dec 19, 2024 by MoSedkyy
1 task done
[Bug]: vllm crash when 20 concurrent test with long content (9k words) bug Something isn't working
#11335 opened Dec 19, 2024 by Flynn-Zh
1 task done
[Bug]: FP8 kvcache causes RuntimeError in v1 engine bug Something isn't working
#11329 opened Dec 19, 2024 by Nekofish-L
1 task done
[Usage]: how to use torch_compile usage How to use vllm
#11323 opened Dec 19, 2024 by chenglu66
1 task done
[Doc]: Update default max_num_batch_tokens for chunked prefill documentation Improvements or additions to documentation
#11319 opened Dec 19, 2024 by toslunar
1 task done
[Bug]: Chat with n>1 breaks xgrammar bug Something isn't working
#11312 opened Dec 18, 2024 by joerunde
1 task done
ProTip! What’s not been updated in a month: updated:<2024-11-20.