Logo
Explore Help
Register Sign In
biondizzle/vllm
1
0
Fork 0
You've already forked vllm
Code Issues Pull Requests Actions 2 Packages Projects Releases Wiki Activity
Files
b6be6f8d1e49d4aa884603e8675dc216be1cbd79
vllm/vllm/v1/worker
History
wwl2755 463bbb1835 [Bugfix][V1] Fix bug from putting llm_engine.model_executor in a background process (#15367)
Signed-off-by: wwl2755 <wangwenlong2755@gmail.com>
2025-04-03 07:32:10 +00:00
..
__init__.py
[V1] Implement vLLM V1 [1/N] (#9289)
2024-10-22 01:24:07 -07:00
block_table.py
Update deprecated Python 3.8 typing (#13971)
2025-03-02 17:34:51 -08:00
gpu_input_batch.py
[V1][Spec Decode] Implement Eagle Proposer [1/N] (#15729)
2025-04-01 12:33:16 -07:00
gpu_model_runner.py
[V1][Spec Decode] Implement Eagle Proposer [1/N] (#15729)
2025-04-01 12:33:16 -07:00
gpu_worker.py
[Bugfix][V1] Fix bug from putting llm_engine.model_executor in a background process (#15367)
2025-04-03 07:32:10 +00:00
lora_model_runner_mixin.py
[Kernels] LoRA - Retire SGMV and BGMV Kernels (#14685)
2025-03-18 09:47:53 +00:00
tpu_model_runner.py
[V1][TPU] Do not compile sampling more than needed (#15883)
2025-04-03 01:36:01 +00:00
tpu_worker.py
[TPU] optimize the all-reduce performance (#15903)
2025-04-03 00:25:14 +00:00
utils.py
[Bugfix] Check dimensions of multimodal embeddings in V1 (#15816)
2025-03-31 09:01:35 -07:00
worker_base.py
[v1] Refactor KVCacheConfig (#14079)
2025-03-21 04:56:27 -07:00
Powered by Gitea Version: 1.25.2 Page: 252ms Template: 3ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API