# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Standard attention benchmark runner - shared utilities for non-MLA benchmarks. This module provides helpers for running standard attention backends (FlashAttention, Triton, FlashInfer) with real vLLM integration. """ import logging import types from contextlib import contextmanager import numpy as np import torch from batch_spec import parse_batch_spec, reorder_for_flashinfer from common import BenchmarkConfig, BenchmarkResult, MockLayer, get_attention_scale from vllm.config import ( CacheConfig, CompilationConfig, DeviceConfig, LoadConfig, ModelConfig, ParallelConfig, SchedulerConfig, VllmConfig, set_current_vllm_config, ) from vllm.v1.attention.backends.utils import ( CommonAttentionMetadata, get_kv_cache_layout, set_kv_cache_layout, ) from vllm.v1.kv_cache_interface import FullAttentionSpec # ============================================================================ # Backend Configuration # ============================================================================ def _get_backend_config(backend: str) -> dict: """ Get backend configuration from AttentionBackendEnum. Args: backend: Backend name matching AttentionBackendEnum exactly (e.g., "FLASH_ATTN", "TRITON_ATTN", "FLASHINFER") Returns: Dict with backend_class """ from vllm.v1.attention.backends.registry import AttentionBackendEnum try: backend_enum = AttentionBackendEnum[backend] backend_class = backend_enum.get_class() except (KeyError, ValueError) as e: valid_backends = [b.name for b in AttentionBackendEnum if b.name != "CUSTOM"] raise ValueError( f"Unknown backend: {backend}. Valid backends: {valid_backends}" ) from e return {"backend_class": backend_class} @contextmanager def log_warnings_and_errors_only(): """Temporarily set vLLM logger to WARNING level.""" logger = logging.getLogger("vllm") old_level = logger.level logger.setLevel(logging.WARNING) try: yield finally: logger.setLevel(old_level) # ============================================================================ # Metadata Building Helpers # ============================================================================ def _build_common_attn_metadata( q_lens: list[int], kv_lens: list[int], block_size: int, device: torch.device, ) -> CommonAttentionMetadata: """Build CommonAttentionMetadata from query/kv lengths.""" batch_size = len(q_lens) total_tokens = sum(q_lens) query_start_loc = torch.zeros(batch_size + 1, dtype=torch.int32, device=device) query_start_loc[1:] = torch.tensor(q_lens, dtype=torch.int32, device=device).cumsum( 0 ) query_start_loc_cpu = query_start_loc.cpu() seq_lens = torch.tensor(kv_lens, dtype=torch.int32, device=device) max_seq_len = int(seq_lens.max().item()) max_blocks = (max(kv_lens) + block_size - 1) // block_size num_blocks = batch_size * max_blocks block_table_tensor = torch.arange( num_blocks, dtype=torch.int32, device=device ).view(batch_size, max_blocks) slot_mapping = torch.arange(total_tokens, dtype=torch.int64, device=device) max_query_len = max(q_lens) return CommonAttentionMetadata( query_start_loc=query_start_loc, query_start_loc_cpu=query_start_loc_cpu, seq_lens=seq_lens, num_reqs=batch_size, num_actual_tokens=total_tokens, max_query_len=max_query_len, max_seq_len=max_seq_len, block_table_tensor=block_table_tensor, slot_mapping=slot_mapping, causal=True, ) def _create_vllm_config( config: BenchmarkConfig, max_num_blocks: int, ) -> VllmConfig: """Create a VllmConfig for benchmarking with mock model methods.""" model_config = ModelConfig( model="meta-llama/Meta-Llama-3-8B", tokenizer="meta-llama/Meta-Llama-3-8B", trust_remote_code=False, dtype="auto", # Use model's native dtype seed=0, max_model_len=1024, ) cache_config = CacheConfig( block_size=config.block_size, cache_dtype="auto", swap_space=0, ) cache_config.num_gpu_blocks = max_num_blocks cache_config.num_cpu_blocks = 0 parallel_config = ParallelConfig(tensor_parallel_size=1) scheduler_config = SchedulerConfig( max_num_seqs=256, max_num_batched_tokens=8192, max_model_len=8192, is_encoder_decoder=False, enable_chunked_prefill=True, ) device_config = DeviceConfig() load_config = LoadConfig() compilation_config = CompilationConfig() # Add mock methods for benchmark config values model_config.get_num_layers = types.MethodType( lambda self: config.num_layers, model_config ) model_config.get_sliding_window_for_layer = types.MethodType( lambda self, i: None, model_config ) model_config.get_logits_soft_cap_for_layer = types.MethodType( lambda self, i: 0.0, model_config ) model_config.get_sm_scale_for_layer = types.MethodType( lambda self, i: 1.0 / config.head_dim**0.5, model_config ) model_config.get_num_attention_heads = types.MethodType( lambda self, parallel_config=None: config.num_q_heads, model_config ) model_config.get_num_kv_heads = types.MethodType( lambda self, parallel_config=None: config.num_kv_heads, model_config ) model_config.get_head_size = types.MethodType( lambda self: config.head_dim, model_config ) model_config.get_sliding_window = types.MethodType(lambda self: None, model_config) return VllmConfig( model_config=model_config, cache_config=cache_config, parallel_config=parallel_config, scheduler_config=scheduler_config, device_config=device_config, load_config=load_config, compilation_config=compilation_config, ) # ============================================================================ # Backend Initialization # ============================================================================ def _create_backend_impl( backend_cfg: dict, config: BenchmarkConfig, device: torch.device, dtype: torch.dtype, ): """Create backend implementation instance.""" backend_class = backend_cfg["backend_class"] scale = get_attention_scale(config.head_dim) impl = backend_class.get_impl_cls()( num_heads=config.num_q_heads, head_size=config.head_dim, scale=scale, num_kv_heads=config.num_kv_heads, alibi_slopes=None, sliding_window=None, kv_cache_dtype="auto", ) kv_cache_spec = FullAttentionSpec( block_size=config.block_size, num_kv_heads=config.num_kv_heads, head_size=config.head_dim, dtype=dtype, ) layer = MockLayer(device, kv_cache_spec=kv_cache_spec) return backend_class, impl, layer def _create_metadata_builder( backend_class, kv_cache_spec: FullAttentionSpec, vllm_config: VllmConfig, device: torch.device, backend_name: str = "", ): """Create metadata builder instance.""" layer_names = ["layer_0"] builder_cls = backend_class.get_builder_cls() # Flashinfer needs get_per_layer_parameters mocked since we don't have # real model layers registered if backend_name == "FLASHINFER": import unittest.mock from vllm.v1.attention.backends.utils import PerLayerParameters def mock_get_per_layer_parameters(vllm_config, layer_names, impl_cls): head_size = vllm_config.model_config.get_head_size() return { layer_name: PerLayerParameters( window_left=-1, # No sliding window logits_soft_cap=0.0, # No soft cap sm_scale=1.0 / (head_size**0.5), # Standard scale ) for layer_name in layer_names } with unittest.mock.patch( "vllm.v1.attention.backends.flashinfer.get_per_layer_parameters", mock_get_per_layer_parameters, ): return builder_cls( kv_cache_spec=kv_cache_spec, layer_names=layer_names, vllm_config=vllm_config, device=device, ) return builder_cls( kv_cache_spec=kv_cache_spec, layer_names=layer_names, vllm_config=vllm_config, device=device, ) # ============================================================================ # Tensor Creation Helpers # ============================================================================ def _create_input_tensors( config: BenchmarkConfig, total_q: int, device: torch.device, dtype: torch.dtype, ) -> tuple: """Create Q, K, V input tensors for all layers.""" q_list = [ torch.randn( total_q, config.num_q_heads, config.head_dim, device=device, dtype=dtype ) for _ in range(config.num_layers) ] k_list = [ torch.randn( total_q, config.num_kv_heads, config.head_dim, device=device, dtype=dtype ) for _ in range(config.num_layers) ] v_list = [ torch.randn( total_q, config.num_kv_heads, config.head_dim, device=device, dtype=dtype ) for _ in range(config.num_layers) ] return q_list, k_list, v_list def _create_kv_cache( config: BenchmarkConfig, max_num_blocks: int, backend_class, device: torch.device, dtype: torch.dtype, ) -> list: """Create KV cache tensors for all layers using the backend's methods. Uses the backend's get_kv_cache_shape() and get_kv_cache_stride_order() to create the cache with the correct shape and memory layout. """ # Get the logical shape from the backend cache_shape = backend_class.get_kv_cache_shape( num_blocks=max_num_blocks, block_size=config.block_size, num_kv_heads=config.num_kv_heads, head_size=config.head_dim, ) # Get the stride order for custom memory layout try: stride_order = backend_class.get_kv_cache_stride_order() assert len(stride_order) == len(cache_shape) except (AttributeError, NotImplementedError): stride_order = tuple(range(len(cache_shape))) # Permute shape to physical layout order physical_shape = tuple(cache_shape[i] for i in stride_order) # Compute inverse permutation to get back to logical view inv_order = [stride_order.index(i) for i in range(len(stride_order))] cache_list = [] for _ in range(config.num_layers): # Allocate in physical layout order (contiguous in memory) cache = torch.zeros(*physical_shape, device=device, dtype=dtype) # Permute to logical view cache = cache.permute(*inv_order) cache_list.append(cache) return cache_list # ============================================================================ # Benchmark Execution # ============================================================================ def _run_single_benchmark( config: BenchmarkConfig, impl, layer, q_list: list, k_list: list, v_list: list, cache_list: list, attn_metadata, device: torch.device, dtype: torch.dtype, ) -> tuple: """Run single benchmark iteration with warmup and timing loop.""" total_q = q_list[0].shape[0] out = torch.empty( total_q, config.num_q_heads, config.head_dim, device=device, dtype=dtype ) # Warmup for _ in range(config.warmup_iters): for i in range(config.num_layers): impl.forward( layer, q_list[i], k_list[i], v_list[i], cache_list[i], attn_metadata, output=out, ) torch.cuda.synchronize() # Benchmark times = [] for _ in range(config.repeats): start = torch.cuda.Event(enable_timing=True) end = torch.cuda.Event(enable_timing=True) start.record() for i in range(config.num_layers): impl.forward( layer, q_list[i], k_list[i], v_list[i], cache_list[i], attn_metadata, output=out, ) end.record() torch.cuda.synchronize() elapsed_ms = start.elapsed_time(end) times.append(elapsed_ms / 1000.0 / config.num_layers) # seconds per layer mem_stats = {} if config.profile_memory: mem_stats = { "allocated_mb": torch.cuda.memory_allocated(device) / 1024**2, "reserved_mb": torch.cuda.memory_reserved(device) / 1024**2, } return times, mem_stats # ============================================================================ # Public API # ============================================================================ def run_attention_benchmark(config: BenchmarkConfig) -> BenchmarkResult: """ Run standard attention benchmark with real kernels. Supports: FLASH_ATTN, TRITON_ATTN, FLASHINFER Args: config: Benchmark configuration Returns: BenchmarkResult with timing and memory statistics """ device = torch.device(config.device) torch.cuda.set_device(device) backend_cfg = _get_backend_config(config.backend) requests = parse_batch_spec(config.batch_spec) if config.backend == "FLASHINFER": requests = reorder_for_flashinfer(requests) q_lens = [r.q_len for r in requests] kv_lens = [r.kv_len for r in requests] total_q = sum(q_lens) max_kv = max(kv_lens) batch_size = len(q_lens) # Calculate total blocks needed: batch_size * max_blocks_per_request max_blocks_per_request = (max_kv + config.block_size - 1) // config.block_size max_num_blocks = batch_size * max_blocks_per_request # Suppress vLLM logs during setup to reduce spam with log_warnings_and_errors_only(): # Create vllm_config first - uses model's native dtype via "auto" vllm_config = _create_vllm_config(config, max_num_blocks) dtype = vllm_config.model_config.dtype # Wrap everything in set_current_vllm_config context # This is required for backends like flashinfer that need global config with set_current_vllm_config(vllm_config): backend_class, impl, layer = _create_backend_impl( backend_cfg, config, device, dtype ) # Set KV cache layout if the backend requires a specific one # (e.g., FlashInfer requires HND on SM100/Blackwell for TRTLLM attention) required_layout = backend_class.get_required_kv_cache_layout() if required_layout is not None: set_kv_cache_layout(required_layout) get_kv_cache_layout.cache_clear() common_metadata = _build_common_attn_metadata( q_lens, kv_lens, config.block_size, device ) kv_cache_spec = FullAttentionSpec( block_size=config.block_size, num_kv_heads=config.num_kv_heads, head_size=config.head_dim, dtype=dtype, ) builder = _create_metadata_builder( backend_class, kv_cache_spec, vllm_config, device, config.backend ) attn_metadata = builder.build( common_prefix_len=0, common_attn_metadata=common_metadata, ) q_list, k_list, v_list = _create_input_tensors( config, total_q, device, dtype ) cache_list = _create_kv_cache( config, max_num_blocks, backend_class, device, dtype ) times, mem_stats = _run_single_benchmark( config, impl, layer, q_list, k_list, v_list, cache_list, attn_metadata, device, dtype, ) mean_time = np.mean(times) throughput = total_q / mean_time if mean_time > 0 else 0 return BenchmarkResult( config=config, mean_time=mean_time, std_time=np.std(times), min_time=np.min(times), max_time=np.max(times), throughput_tokens_per_sec=throughput, memory_allocated_mb=mem_stats.get("allocated_mb"), memory_reserved_mb=mem_stats.get("reserved_mb"), )