Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

[Build/CI] Fix libcuda.so linkage ci/build ready ONLY add when PR is ready to merge/full CI is needed
#12424 opened Jan 25, 2025 by tlrmchlsmth Loading…
[ROCm][AMD][Model] llama 3.2 support upstreaming
#12421 opened Jan 24, 2025 by maleksan85 Loading…
Fix the pydantic logging validator frontend
#12420 opened Jan 24, 2025 by maxdebayser Loading…
[Bugfix] Disable w16a16 2of4 sparse CompressedTensors24 ready ONLY add when PR is ready to merge/full CI is needed
#12417 opened Jan 24, 2025 by tlrmchlsmth Loading…
[V1][Metrics] Add initial Prometheus logger ready ONLY add when PR is ready to merge/full CI is needed
#12416 opened Jan 24, 2025 by markmc Draft
[V1] Revert uncache_blocks and support recaching full blocks ready ONLY add when PR is ready to merge/full CI is needed
#12415 opened Jan 24, 2025 by comaniac Loading…
[Usage] Add pipeline parallelism for usage stats
#12414 opened Jan 24, 2025 by simon-mo Loading…
[Frontend] Support override generation config in args ready ONLY add when PR is ready to merge/full CI is needed
#12409 opened Jan 24, 2025 by liuyanyi Loading…
[ci/build] detect and auto use cxx abi ci/build
#12403 opened Jan 24, 2025 by youkaichao Loading…
[MISC] add arg pad_for_invariant_seq_len
#12397 opened Jan 24, 2025 by MengqingCao Loading…
[Bugfix] Fix output_tokens is 0 if using tgi backend
#12394 opened Jan 24, 2025 by sywangyi Loading…
[Hardware][Intel GPU] add XPU bf16 support documentation Improvements or additions to documentation
#12392 opened Jan 24, 2025 by jikunshang Loading…
[Frontend] Rerank API (Jina- and Cohere-compatible API) documentation Improvements or additions to documentation frontend
#12376 opened Jan 24, 2025 by K-Mistele Loading…
[Core] add and implement VLLM_LOGITS_PROCESSOR_THREADS
#12368 opened Jan 23, 2025 by akeshet Loading…
FLOP counting for vLLM inference
#12341 opened Jan 23, 2025 by dianastea Draft
[Build] Only build 9.0a for scaled_mm and sparse kernels ci/build ready ONLY add when PR is ready to merge/full CI is needed
#12339 opened Jan 23, 2025 by LucasWilkinson Loading…
ProTip! Exclude everything labeled bug with -label:bug.