view article Article KV Caching Explained: Optimizing Transformer Inference Efficiency not-lain • Jan 30, 2025 • 329
view article Article Ultra-Long Sequence Parallelism: Ulysses + Ring-Attention Technical Principles and Implementation exploding-gradients • Sep 16, 2025 • 20
view article Article You could have designed state of the art positional encoding FL33TW00D-HF • Nov 25, 2024 • 478
Running 3.84k The Ultra-Scale Playbook 🌌 3.84k The ultimate guide to training LLM on large GPU Clusters