Quantized Attention achieves speedup of 2-3x and 3-5x compared to FlashAttention and xformers, without lossing end-to-end metrics across language, image, and video models.
cuda triton attention vit quantization video-generation mlsys inference-acceleration efficient-attention llm llm-infra video-generate
-
Updated
Apr 21, 2025 - Cuda