Michael Nielsen – How science actually progresses

Michael Nielsen – How science actually progresses

April 07, 2026 2 hr 3 min
🎧 Listen Now

🤖 AI Summary

Overview

This episode explores the elusive nature of scientific progress, focusing on how breakthroughs are recognized and validated. Through historical examples like Einstein's special relativity, Darwin's theory of evolution, and the discovery of isotopes, the conversation examines the long and often ambiguous verification loops in science. Michael Nielsen also shares provocative ideas about the diversity of potential scientific and technological paths, including how alien civilizations might develop entirely different tech stacks, with implications for future inter-civilizational trade and cooperation.

Notable Quotes

- Great scientists can remain wrong for a very long time after the scientific community has broadly changed its opinion.Michael Nielsen, on the complexity of scientific progress.

- The tech tree of science is much larger than we realize. Most parts of it will never be explored.Michael Nielsen, on the vastness of unexplored scientific possibilities.

- If raw gradient descent were doing science, it wouldn’t have switched from Ptolemy to Copernicus—it would just keep adding epicycles.Dwarkesh Patel, on the limitations of AI in scientific discovery.

🧪 The Complexity of Scientific Verification

- The Michelson-Morley experiment, often credited with inspiring Einstein's special relativity, was not central to his thinking. Einstein later claimed he wasn’t even aware of it at the time.

- Verification loops in science can span centuries. For example, Aristarchus proposed heliocentrism in 200 BCE, but stellar parallax wasn’t measured until 1838—a 2,000-year gap.

- The discovery of isotopes illustrates how experimental results can actively mislead scientists for decades, requiring persistent defenders of seemingly discredited theories.

🌌 Alien Science and the Tech Tree Hypothesis

- Nielsen argues that alien civilizations might develop entirely different scientific and technological stacks due to the vastness of the tech tree.

- This challenges the assumption of a universal, linear progression of science and suggests that inter-civilizational trade could yield immense gains due to non-overlapping discoveries.

- The idea implies that friendliness and cooperation between civilizations could be highly rewarding, as each could offer unique insights and technologies.

📜 Historical Bottlenecks in Scientific Progress

- Darwin’s theory of evolution took centuries to emerge despite its conceptual simplicity. Factors like the lack of a deep-time framework (established by geologists like Charles Lyell) delayed its acceptance.

- Newtonian gravity succeeded in explaining Uranus’s orbit by predicting Neptune, but failed with Mercury’s orbit, which required Einstein’s general relativity. This highlights the difficulty of distinguishing between correct and incorrect theories based on anomalies.

- The history of quantum computing shows how breakthroughs often depend on the convergence of multiple factors, such as the rise of computation and advances in experimental physics in the 1980s.

🤖 AI and the Future of Science

- While AI excels in domains with tight verification loops (e.g., coding, protein folding), it struggles with the ambiguity and creativity required for paradigm-shifting discoveries.

- Nielsen suggests that AI might accelerate science by solving specific bottlenecks, but the broader process of scientific exploration still requires human-like intuition and creativity.

- The conversation explores whether AI could eventually develop new types of scientific explanations, distinct from human approaches, and how tools like AlphaFold might contain hidden insights waiting to be extracted.

📚 Learning and Intellectual Growth

- Dwarkesh reflects on the challenge of deeply learning from podcast interviews, noting the risk of superficial understanding without structured follow-up.

- Nielsen emphasizes the importance of creating high-stakes, demanding contexts for learning, such as writing a book or solving a specific problem.

- Both discuss the value of balancing routine work with high-variance, exploratory efforts to achieve deeper intellectual engagement.

AI-generated content may not be accurate or complete and should not be relied upon as a sole source of truth.

📋 Episode Description

The key question in this conversation is, how do we recognize scientific progress?It's especially relevant for closing the RL verification for scientific discovery. But it’s also a surprisingly mysterious and elusive question when you analyze the history of human science.

We approach this question through the stories of Einstein (who claimed that he hadn't even heard of the famous Michaelson Morely experiment which is supposed to have motivated special relativity until after he had come up with it), Darwin (why did it take till 1859 to lay out an idea whose essence every farmer since antiquity must have observed?, Prout (how do you recognize that isotopes exist if you cannot chemically separate them?), and many others.

The verification loop on scientific ideas is often extremely long and weirdly hostile. Ancient Athenians dismissed Aristarchus's heliocentrism in the 2nd century BC because it would imply that the stars should shift in the sky as the Earth orbits the sun. The first successful measurement of stellar parallax was in 1838. That's a 2,000-year verification loop.

But clearly human science is able to make progress faster than raw experimental falsification/verification would imply, and in cases where experiments are very ambiguous. How?

Michael has some very deep and provocative hypotheses about the nature of progress. One I found especially thought-provoking is that aliens will likely have a VERY different science + tech stack that us. Which contradicts the common sense picture of a linear tech tree that I was assuming. And has some interesting implications about how future civilizations might trade and cooperate with each other.

So many other interesting ideas. Really hope you enjoy this as much as I did.

Sponsors

* Labelbox researchers built a new safety benchmark. Why? Well, current safety benchmarks claim that attacks on top models are successful only a few percent of the time, but the prompts in those benchmarks don’t reflect how real bad actors actually write. You can read Labelbox’s research here: https://labelbox.com/blog/the-ai-safety-illusion-why-current-safety-datasets-fool-us-on-model-safety/. If this could be useful for your work, reach out at labelbox.com/dwarkesh

* Mercury has an MCP that lets you give an LLM access to your full transaction history, including things like attached receipts and internal notes. I just used it to categorize my 2025 transactions, and it worked shockingly well. Modern functionality like this is exactly why I use Mercury. Learn more at mercury.com

* Jane Street’s ML engineers presented some of their GPU optimization workflows at GTC, showing how