AlphaGo Shows How Search Can Turn RL Into Supervised Learning
Eric Jang rebuilds AlphaGo as a way to examine why its combination of search, value learning and self-play still matters for modern AI. His central claim is that AlphaGo’s Monte Carlo Tree Search turns each move into a better supervised-learning target, avoiding the long-horizon credit-assignment problem that makes much reinforcement learning for language models inefficient. Jang also argues that current LLM research assistants can already help execute and optimize experiments, but still struggle with the harder judgment of choosing which research paths are worth pursuing.
Dwarkesh Patel·May 15, 2026·28 min read