vllm.attention.ops ΒΆ
 Modules:
| Name | Description | 
|---|---|
chunked_prefill_paged_decode |    |  
common |    |  
flashmla |    |  
merge_attn_states |    |  
paged_attn |    |  
pallas_kv_cache_update |    |  
prefix_prefill |    |  
rocm_aiter_mla |    |  
rocm_aiter_paged_attn |    |  
triton_decode_attention |    Memory-efficient attention for decoding.  |  
triton_flash_attention |    Fused Attention  |  
triton_merge_attn_states |    |  
triton_reshape_and_cache_flash |    |  
triton_unified_attention |    |  
vit_attn_wrappers |    This file contains ops for ViT attention to be compatible with torch.compile  |