Running 1 Estimated Performance Frontiers of Open LLM Leaderboard Tasks 🚀 1 Estimate LLM task performance from pretraining compute
Prescriptive Scaling Reveals the Evolution of Language Model Capabilities Paper • 2602.15327 • Published 12 days ago • 2
Prescriptive Scaling Reveals the Evolution of Language Model Capabilities Paper • 2602.15327 • Published 12 days ago • 2
Discovering Hierarchical Latent Capabilities of Language Models via Causal Representation Learning Paper • 2506.10378 • Published Jun 12, 2025 • 2 • 2
Running 3.71k The Ultra-Scale Playbook 🌌 3.71k The ultimate guide to training LLM on large GPU Clusters
Do the Rewards Justify the Means? Measuring Trade-Offs Between Rewards and Ethical Behavior in the MACHIAVELLI Benchmark Paper • 2304.03279 • Published Apr 6, 2023 • 2
CoLoR-Filter: Conditional Loss Reduction Filtering for Targeted Language Model Pre-training Paper • 2406.10670 • Published Jun 15, 2024 • 4
DataComp-LM: In search of the next generation of training sets for language models Paper • 2406.11794 • Published Jun 17, 2024 • 55
Eliminating Position Bias of Language Models: A Mechanistic Approach Paper • 2407.01100 • Published Jul 1, 2024 • 8
Mind the Gap: Examining the Self-Improvement Capabilities of Large Language Models Paper • 2412.02674 • Published Dec 3, 2024