Ilya Sutskever – We're moving from the age of scaling to the age of research
🤖 AI Summary
Overview
This episode explores the transition from the age of scaling
to the age of research
in AI development, as discussed by Ilya Sutskever. Topics include the limitations of current AI training paradigms, the challenges of generalization, the future of superintelligence, and how SSI (Sutskever's company) plans to approach these challenges differently.
Notable Quotes
- The models are much more like the first student, but even more, because we train them so narrowly that they don’t generalize to other things.
– Ilya Sutskever, on why current AI struggles with generalization.
- If ideas are so cheap, how come no one's having any ideas?
– Ilya Sutskever, reflecting on the bottleneck of innovation in AI research.
- The whole problem of AI is the power. When the power is really big, what's going to happen?
– Ilya Sutskever, on the challenges of managing superintelligence.
🧠 The Transition from Scaling to Research
- Ilya Sutskever describes the age of scaling
(2012–2025) as a period where increasing compute, data, and model size drove progress. However, he argues that scaling alone is no longer sufficient.
- The age of research
focuses on understanding fundamental principles, such as improving generalization and addressing the disconnect between model performance on benchmarks and real-world utility.
- He emphasizes the need for new approaches, as scaling pre-training has diminishing returns and finite data limits.
📊 Generalization and Model Limitations
- Current AI models excel at specific tasks but fail to generalize effectively, often repeating errors or struggling with tasks outside their training distribution.
- Sutskever compares models to students: one who practices narrowly for competitive programming versus one who learns broadly. The latter generalizes better, akin to human learning.
- He highlights the importance of understanding why humans learn efficiently with limited data and how this could inspire better AI training paradigms.
🌍 The Path to Superintelligence
- SSI’s strategy involves focusing on research to develop AI systems capable of human-like continual learning.
- Sutskever discusses the concept of straight-shot superintelligence,
where AI is developed incrementally but with a focus on safety and alignment from the outset.
- He suggests that future AI systems should prioritize alignment with sentient life, arguing that empathy and care for sentience may emerge naturally in advanced systems.
🤝 Collaboration and Safety in AI Development
- Sutskever predicts increased collaboration between competing AI companies on safety measures as AI systems become more visibly powerful.
- He advocates for gradual deployment of AI to help society adapt and to identify potential risks early.
- He also emphasizes the importance of building systems that are robustly aligned to human values and capable of safe, incremental learning.
🎨 Research Taste and Inspiration
- Sutskever attributes his research success to a combination of aesthetic intuition, simplicity, and inspiration from human cognition.
- He stresses the importance of top-down beliefs in guiding research, especially when experimental results are ambiguous or contradictory.
- His approach involves seeking elegance and avoiding ugliness
in AI design, drawing heavily from how humans learn and process information.
AI-generated content may not be accurate or complete and should not be relied upon as a sole source of truth.
📋 Episode Description
Ilya & I discuss SSI’s strategy, the problems with pre-training, how to improve the generalization of AI models, and how to ensure AGI goes well.
Watch on YouTube; read the transcript.
Sponsors
* Gemini 3 is the first model I’ve used that can find connections I haven’t anticipated. I recently wrote a blog post on RL’s information efficiency, and Gemini 3 helped me think it all through. It also generated the relevant charts and ran toy ML experiments for me with zero bugs. Try Gemini 3 today at gemini.google
* Labelbox helped me create a tool to transcribe our episodes! I’ve struggled with transcription in the past because I don’t just want verbatim transcripts, I want transcripts reworded to read like essays. Labelbox helped me generate the exact data I needed for this. If you want to learn how Labelbox can help you (or if you want to try out the transcriber tool yourself), go to labelbox.com/dwarkesh
* Sardine is an AI risk management platform that brings together thousands of device, behavior, and identity signals to help you assess a user’s risk of fraud & abuse. Sardine also offers a suite of agents to automate investigations so that as fraudsters use AI to scale their attacks, you can use AI to scale your defenses. Learn more at sardine.ai/dwarkesh
To sponsor a future episode, visit dwarkesh.com/advertise.
Timestamps
(00:00:00) – Explaining model jaggedness
(00:09:39) - Emotions and value functions
(00:18:49) – What are we scaling?
(00:25:13) – Why humans generalize better than models
(00:35:45) – SSI’s plan to straight-shot superintelligence
(00:46:47) – SSI’s model will learn from deployment
(00:55:07) – How to think about powerful AGIs
(01:18:13) – “We are squarely an age of research company”
(01:30:26) – Self-play and multi-agent
(01:32:42) – Research taste
Get full access to Dwarkesh Podcast at www.dwarkesh.com/subscribe