Pinokio

SageAttention

https://github.com/deepbeepmeep/sageattentionupdated 3/4/2025, 4:50:49 PMindexed 5/12/2026, 5:43:11 PM

Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.

Posts

Sort
Loading…