Skip to content

Pull requests: vllm-project/vllm-gaudi

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[WIP] Enable mamba
#138 opened Sep 4, 2025 by tianmu-li Draft
[SW-236089] UTs: multimodality correctness
#136 opened Sep 4, 2025 by kfojcik-intel Loading…
Update CODEOWNERS
#135 opened Sep 4, 2025 by afierka-intel Loading…
Reenabling llama4 models
#128 opened Sep 3, 2025 by afierka-intel Loading…
Add out-of-tree HPU schedulers
#119 opened Sep 1, 2025 by kzawora-intel Loading…
[WARMUP] fix update bucket
#118 opened Aug 29, 2025 by xuechendi Loading…
[Bucketing] WA for warmup big values - crash
#116 opened Aug 29, 2025 by adobrzyn Loading…
Re-quantize FP8 model with INC
#114 opened Aug 29, 2025 by yiliu30 Draft
[Bucketing] Read buckets from file
#101 opened Aug 23, 2025 by adobrzyn Draft
initial port for nixl
#100 opened Aug 22, 2025 by hsubramony Loading…
Add attention unit tests
#74 opened Aug 12, 2025 by tthaddey Loading…
Lookahead decoding
#72 opened Aug 11, 2025 by jkaniecki Loading…
Fixed Plugin Test
#70 opened Aug 8, 2025 by slokesha Loading…
[test] Add yaml files for fp8 tests
#53 opened Jul 29, 2025 by ulivne Loading…
Add sampler unit tests
#46 opened Jul 28, 2025 by kzawora-intel Loading…
Proper chunked prefill bucketing/warmup
#32 opened Jul 16, 2025 by kzawora-intel Loading…
ProTip! Updated in the last three days: updated:>2025-09-06.