SLA2: Sparse-Linear Attention with Learnable Routing and QAT Paper • 2602.12675 • Published about 1 month ago • 55 • 5
TurboDiffusion: Accelerating Video Diffusion Models by 100-200 Times Paper • 2512.16093 • Published Dec 18, 2025 • 95 • 7
SLA: Beyond Sparsity in Diffusion Transformers via Fine-Tunable Sparse-Linear Attention Paper • 2509.24006 • Published Sep 28, 2025 • 118 • 4
SLA: Beyond Sparsity in Diffusion Transformers via Fine-Tunable Sparse-Linear Attention Paper • 2509.24006 • Published Sep 28, 2025 • 118 • 4
SLA: Beyond Sparsity in Diffusion Transformers via Fine-Tunable Sparse-Linear Attention Paper • 2509.24006 • Published Sep 28, 2025 • 118 • 4
Insights into DeepSeek-V3: Scaling Challenges and Reflections on Hardware for AI Architectures Paper • 2505.09343 • Published May 14, 2025 • 76 • 5
SageAttention2++: A More Efficient Implementation of SageAttention2 Paper • 2505.21136 • Published May 27, 2025 • 45 • 3
SageAttention2++: A More Efficient Implementation of SageAttention2 Paper • 2505.21136 • Published May 27, 2025 • 45 • 3
SageAttention3: Microscaling FP4 Attention for Inference and An Exploration of 8-Bit Training Paper • 2505.11594 • Published May 16, 2025 • 75 • 8
SageAttention3: Microscaling FP4 Attention for Inference and An Exploration of 8-Bit Training Paper • 2505.11594 • Published May 16, 2025 • 75 • 8
Identifying Sensitive Weights via Post-quantization Integral Paper • 2503.01901 • Published Feb 28, 2025 • 8 • 2