How Do Large Language Monkeys Get Their Power (Laws)?
Abstract
We investigate the origins of power law scaling in large language model inference-time compute, explaining why repeated sampling yields predictable improvements in task performance.
Summary
Understanding the origins of power law scaling in large language model inference-time compute.
Overview
We investigate the origins of power law scaling in large language model inference-time compute, explaining why repeated sampling yields predictable improvements in task performance.
Oral presentation at ICML 2025.
