Attention Sink in Transformers: A Survey on Utilization, Interpretation, and Mitigation Paper • 2604.10098 • Published 15 days ago • 75
MSA: Memory Sparse Attention for Efficient End-to-End Memory Model Scaling to 100M Tokens Paper • 2603.23516 • Published Mar 6 • 48
Recursive Language Models Meet Uncertainty: The Surprising Effectiveness of Self-Reflective Program Search for Long Context Paper • 2603.15653 • Published Mar 7 • 12
Qwen3.5-Abliterated-Opus-4.6-Distilled Collection Qwen3.5-Abliterated • 3 items • Updated Mar 8 • 1