#490 – State of AI in 2026: LLMs, Coding, Scaling Laws, China, Agents, GPUs, AGI

#490 – State of AI in 2026: LLMs, Coding, Scaling Laws, China, Agents, GPUs, AGI

February 01, 2026 0 min
🎧 Listen Now

🤖 AI Summary

Overview

This episode dives deep into the current state of artificial intelligence in 2026, focusing on large language models (LLMs), coding advancements, scaling laws, open vs. closed source AI, and the global AI race between China and the United States. The discussion also explores the future of AI, including its economic impact, potential breakthroughs, and societal implications.

Notable Quotes

- Humans are in charge. AI is still a tool; it doesn't take agency from you.Sebastian Raschka, on the role of AI in human decision-making.

- The dream of one central model to rule everything is dying.Nathan Lambert, on the shift toward specialized AI systems.

- We might get to Mars because of the quiet force of AI making all human knowledge accessible.Lex Fridman, on the transformative potential of AI in education and innovation.

🌍 The Global AI Race: China vs. the US

- China's dominance in open-weight AI models: Companies like DeepSeek and Z.ai are leading with innovative open-weight models, challenging the US's position.

- US response: Initiatives like the Adam Project aim to bolster American open-source AI to compete with China's advancements.

- Cultural and economic differences: US companies focus on monetization, while Chinese companies leverage open models to gain influence globally.

- Open-source as a strategic advantage: Open models are seen as engines for innovation, but concerns about security and intellectual property persist.

🤖 The Evolution of LLMs and AI for Coding

- LLMs for coding: Tools like Claude Code and Cursor are revolutionizing software development, enabling faster debugging and code generation.

- Senior developers lead adoption: Experienced programmers are more likely to use AI-generated code, with 25% using it for over half of their shipped code.

- Challenges in full automation: While LLMs excel at specific coding tasks, achieving fully autonomous programming remains a distant goal due to the complexity of software systems.

📈 Scaling Laws and AI Training

- Scaling laws still hold: Increasing compute and data continues to improve model performance, but the low-hanging fruit in pretraining may be exhausted.

- Post-training innovations: Reinforcement Learning with Verifiable Rewards (RLVR) has emerged as a game-changer, enabling models to learn reasoning and tool use.

- Inference scaling: Models like OpenAI's O1 demonstrate how extended inference time can enhance reasoning capabilities.

🛠️ Open Source vs. Closed Source AI

- Open-source momentum: Open-weight models are driving innovation and accessibility, with Chinese companies leading the charge.

- Closed-source advantages: Companies like OpenAI and Anthropic focus on proprietary models with deep integrations, offering tailored solutions for enterprises.

- The future of open AI: US initiatives like the Adam Project aim to reclaim leadership in open-source AI, emphasizing its role in education and innovation.

🌌 The Future of AI and Society

- Timelines to AGI: Predictions vary, but full automation of programming and scientific discovery remains a long-term goal.

- Economic impact: While AI has yet to create a significant GDP leap, its potential to democratize knowledge and enhance productivity is immense.

- Human connection and creativity: As AI generates more slop (low-quality content), the value of human creativity, physical experiences, and community will likely increase.

- Ethical considerations: The societal implications of AI, from job displacement to misinformation, require careful navigation to ensure a positive future.

This episode offers a comprehensive look at the current and future landscape of AI, emphasizing both its transformative potential and the challenges it presents.

AI-generated content may not be accurate or complete and should not be relied upon as a sole source of truth.

📋 Episode Description

Nathan Lambert and Sebastian Raschka are machine learning researchers, engineers, and educators. Nathan is the post-training lead at the Allen Institute for AI (Ai2) and the author of The RLHF Book. Sebastian Raschka is the author of Build a Large Language Model (From Scratch) and Build a Reasoning Model (From Scratch).

Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep490-sc

See below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc.


Transcript:

https://lexfridman.com/ai-sota-2026-transcript


CONTACT LEX:

Feedback – give feedback to Lex: https://lexfridman.com/survey

AMA – submit questions, videos or call-in: https://lexfridman.com/ama

Hiring – join our team: https://lexfridman.com/hiring

Other – other ways to get in touch: https://lexfridman.com/contact


SPONSORS:

To support this podcast, check out our sponsors & get discounts:

Box: Intelligent content management platform.

Go to https://box.com/ai

Quo: Phone system (calls, texts, contacts) for businesses.

Go to https://quo.com/lex

UPLIFT Desk: Standing desks and office ergonomics.

Go to https://upliftdesk.com/lex

Fin: AI agent for customer service.

Go to https://fin.ai/lex

Shopify: Sell stuff online.

Go to https://shopify.com/lex

CodeRabbit: AI-powered code reviews.

Go to https://coderabbit.ai/lex

LMNT: Zero-sugar electrolyte drink mix.

Go to https://drinkLMNT.com/lex

Perplexity: AI-powered answer engine.

Go to https://perplexity.ai/


OUTLINE:

(00:00) – Introduction

(01:39) – Sponsors, Comments, and Reflections

(16:29) – China vs US: Who wins the AI race?

(25:11) – ChatGPT vs Claude vs Gemini vs Grok: Who is winning?

(36:11) – Best AI for coding

(43:02) – Open Source vs Closed Source LLMs

(54:41) – Transformers: Evolution of LLMs since 2019

(1:02:38) – AI Scaling Laws: Are they dead or still holding?

(1:18:45) – How AI is trained: Pre-training, Mid-training, and Post-training