Running on CPU Upgrade Featured 2.8k The Smol Training Playbook 📚 2.8k The secrets to building world-class LLMs
Running 3.62k The Ultra-Scale Playbook 🌌 3.62k The ultimate guide to training LLM on large GPU Clusters
view article Article KV Caching Explained: Optimizing Transformer Inference Efficiency Jan 30, 2025 • 210
Training Dynamics Impact Post-Training Quantization Robustness Paper • 2510.06213 • Published Oct 7, 2025 • 3
Training Dynamics Impact Post-Training Quantization Robustness Paper • 2510.06213 • Published Oct 7, 2025 • 3
view article Article Prefill and Decode for Concurrent Requests - Optimizing LLM Performance Apr 16, 2025 • 59
🧠SmolLM3 Collection Smol, multilingual, long-context reasoner • 14 items • Updated Oct 9, 2025 • 90
view article Article Unlocking Longer Generation with Key-Value Cache Quantization May 16, 2024 • 54