This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
9206b3d7ecf5ac05ba22d45740609e570d3d171e
vllm
/
vllm
/
v1
/
core
History
Cody Yu
9206b3d7ec
[V1][PP] Run engine busy loop with batch queue (
#13064
)
2025-02-15 03:59:01 -08:00
..
__init__.py
[V1] Implement vLLM V1 [1/N] (
#9289
)
2024-10-22 01:24:07 -07:00
encoder_cache_manager.py
[V1][Bugfix] Copy encoder input ids to fix set iteration issue during VLM abort (
#13173
)
2025-02-12 12:58:11 -08:00
kv_cache_manager.py
[V1][Metrics] Add several request timing histograms (
#12644
)
2025-02-11 10:14:00 -05:00
kv_cache_utils.py
[V1][core] Implement pipeline parallel on Ray (
#12996
)
2025-02-13 08:02:46 +00:00
scheduler_output.py
[V1][Minor] Move scheduler outputs to a separate file (
#13062
)
2025-02-11 02:10:06 +00:00
scheduler.py
[V1][PP] Run engine busy loop with batch queue (
#13064
)
2025-02-15 03:59:01 -08:00