This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
83ca9ae47b4476c900738b6db0ff5fcdfce13a7b
vllm
/
tests
/
v1
/
tpu
History
Chengji Yao
a77aea59fd
[TPU] support attention head dim smaller than 128 (
#19620
)
...
Signed-off-by: Chengji Yao <
chengjiyao@google.com
> Co-authored-by: mgoin <
mgoin64@gmail.com
>
2025-06-16 06:40:53 +00:00
..
worker
[Core] Use tuple for kv cache group block ids (
#19175
)
2025-06-10 07:01:17 +02:00
__init__.py
[V1] TPU - Add tensor parallel support via Ray (
#13618
)
2025-03-08 08:19:38 -05:00
test_basic.py
[TPU] support attention head dim smaller than 128 (
#19620
)
2025-06-16 06:40:53 +00:00
test_mha_attn.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00
test_multimodal.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00
test_pallas.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00
test_perf.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00
test_sampler.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00
test_spmd_model_weight_loading.py
[TPU] Skip hanging tests (
#19115
)
2025-06-04 01:43:00 -07:00
test_topk_topp_sampler.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00
test_tpu_qkv_linear.py
[Hardware][TPU] Initial support of model parallelism with single worker using SPMD (
#18011
)
2025-06-03 00:06:20 +00:00