Jump to content
vLLM
Pipelines
Test Suites
Log In
Sign Up
Learn More...
Pipelines
Test Suites
🐎
Performance Benchmark
Public
Builds
main
GitHub Icon
main
All users
All states
All dates
Fix erroneous "model doesn't support compile" warning (#16486)
#11869
16h
Tyler Michael Smith
·
main
GitHub Icon
70de35a88
·
Created
yesterday at 4:24 PM
Loading steps…
[Hardware][Intel-Gaudi] Multi-step scheduling implementation for HPU (#12779)
#11868
17h
Tomasz Zielinski
·
main
GitHub Icon
34b2cf3b3
·
Created
yesterday at 2:38 PM
Loading steps…
[Bugfix] Fix bugs of running Quark quantized models (#16236)
#11865
18h
Michael Goin
·
main
GitHub Icon
9e90c9f73
·
Created
yesterday at 2:18 PM
Loading steps…
[Kernel] support merge_attn_states CUDA kernel, 3x speedup (#16173)
#11864
19h
Michael Goin
·
main
GitHub Icon
e9528f6dc
·
Created
yesterday at 12:50 PM
Loading steps…
Don't install triton on `ppc64le` platform (#16470)
#11863
22h
Harry Mellor
·
main
GitHub Icon
51baa9c33
·
Created
yesterday at 10:11 AM
Loading steps…
[Misc] update api_client example (#16459)
#11862
22h
Reid
·
main
GitHub Icon
35e076b3a
·
Created
yesterday at 10:05 AM
Loading steps…
[Misc] Raise error for V1 not supporting Long LoRA. (#16415)
#11861
23h
Jee Jee Li
·
main
GitHub Icon
a26f59ccb
·
Created
yesterday at 8:51 AM
Loading steps…
Enforce valid max_num_batched_tokens when disable_chunked_mm_input=True (#16447)
#11860
1d
Michael Goin
·
main
GitHub Icon
aa3b3d76e
·
Created
yesterday at 8:09 AM
Loading steps…
[Core][LoRA][1/N] Add LoRA for EncoderDecoderModelRunner (#15990)
#11859
1d
Jee Jee Li
·
main
GitHub Icon
f7030df3b
·
Created
yesterday at 7:32 AM
Loading steps…
Revert "[Model] use AutoWeightsLoader for deepseek_v2, internlm2" (#16453)
#11858
1d
DefTruth
·
main
GitHub Icon
905e91e9a
·
Created
yesterday at 6:44 AM
Loading steps…
[Bugfix] Don't set an upper bound on repetition penalty (#16403)
#11857
1d
Alex Brooks
·
main
GitHub Icon
f8f9c0ba6
·
Created
yesterday at 6:19 AM
Loading steps…
[CPU][Bugfix] Fix CPU docker issues (#16454)
#11856
1d
Li, Jiang
·
main
GitHub Icon
dda811021
·
Created
yesterday at 6:19 AM
Loading steps…
[Bugfix][VLM] Fix failing Phi-4-MM multi-images tests and add vision-speech test (#16424)
#11855
1d
Isotr0py
·
main
GitHub Icon
93195146e
·
Created
yesterday at 4:57 AM
Loading steps…
Update supported_hardware.md for TPU INT8 (#16437)
#11854
1d
Michael Goin
·
main
GitHub Icon
ed3759954
·
Created
yesterday at 4:28 AM
Loading steps…
[Llama4] Enable attention temperature tuning by default for long context (>32k) (#16439)
#11853
1d
Yong Hoon Shin
·
main
GitHub Icon
99ef59cf7
·
Created
yesterday at 4:26 AM
Loading steps…
update benchmark_serving_structured_output to include auto backend (#16438)
#11852
1d
Chenyaaang
·
main
GitHub Icon
d544d141e
·
Created
yesterday at 4:25 AM
Loading steps…
check input length of sonnet samples (#16423)
#11851
1d
Jee Jee Li
·
main
GitHub Icon
3e397a948
·
Created
yesterday at 2:15 AM
Loading steps…
Fix range_ratio Bug in RandomDataset (#16126)
#11850
1d
Roger Wang
·
main
GitHub Icon
268c32507
·
Created
Thursday at 10:31 PM
Loading steps…
[TPU][V1] Disable per-request seed/Generator (#16172)
#11849
1d
Nicolò Lucchesi
·
main
GitHub Icon
3cc9af88f
·
Created
Thursday at 9:05 PM
Loading steps…
[Bugfix] Fix output token length check logic (#16419)
#11848
1d
Roger Wang
·
main
GitHub Icon
7cd0bd721
·
Created
Thursday at 8:16 PM
Loading steps…
‹ Prev
Next ›