Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters Paper • 2408.03314 • Published Aug 6, 2024 • 66
Embarrassingly Simple Self-Distillation Improves Code Generation Paper • 2604.01193 • Published 24 days ago • 46
view article Article How we OCR'ed 30,000 papers using Codex, open OCR models and Jobs 18 days ago • 59
view article Article How I contributed a new model to the Transformers library using Codex 26 days ago • 49
view article Article Keep the Tokens Flowing: Lessons from 16 Open-Source RL Libraries +7 Mar 10 • 133
Data Repetition Beats Data Scaling in Long-CoT Supervised Fine-Tuning Paper • 2602.11149 • Published Feb 11 • 17
view article Article GGML and llama.cpp join HF to ensure the long-term progress of Local AI +4 Feb 20 • 503
view article Article I Let a Lobster Run My Jetson: What OpenClaw Taught Me About the Future of Computing Feb 19 • 16
Reasoning Cache: Continual Improvement Over Long Horizons via Short-Horizon RL Paper • 2602.03773 • Published Feb 3 • 13