Sergey Levine: Robotics and Machine Learning | Lex Fridman Podcast #108
Sergey Levine is a professor at Berkeley and a world-class researcher in deep learning, reinforcement learning, robotics, and computer vision, including the development of algorithms for end-to-end training of neural network policies that combine perception and control, scalable algorithms for inverse reinforcement learning, and deep RL algorithms.
Support this podcast by signing up with these sponsors:
– ExpressVPN at https://www.expressvpn.com/lexpod
– Cash App – use code “LexPodcast” and download:
– Cash App (App Store): https://apple.co/2sPrUHe
– Cash App (Google Play): https://bit.ly/2MlvP5w
EPISODE LINKS:
Sergey’s Twitter: https://twitter.com/svlevine
Sergey’s Website: http://rail.eecs.berkeley.edu/
Sergey’s Papers: https://scholar.google.com/citations?user=8R35rCwAAAAJ
PODCAST INFO:
Podcast website:
https://lexfridman.com/podcast
Apple Podcasts:
https://apple.co/2lwqZIr
Spotify:
https://spoti.fi/2nEwCF8
RSS:
https://lexfridman.com/feed/podcast/
Full episodes playlist:
https://www.youtube.com/playlist?list=PLrAXtmErZgOdP_8GztsuKi9nrraNbKKp4
Clips playlist:
https://www.youtube.com/playlist?list=PLrAXtmErZgOeciFP3CBCIEElOJeitOr41
OUTLINE:
0:00 – Introduction
3:05 – State-of-the-art robots vs humans
16:13 – Robotics may help us understand intelligence
22:49 – End-to-end learning in robotics
27:01 – Canonical problem in robotics
31:44 – Commonsense reasoning in robotics
34:41 – Can we solve robotics through learning?
44:55 – What is reinforcement learning?
1:06:36 – Tesla Autopilot
1:08:15 – Simulation in reinforcement learning
1:13:46 – Can we learn gravity from data?
1:16:03 – Self-play
1:17:39 – Reward functions
1:27:01 – Bitter lesson by Rich Sutton
1:32:13 – Advice for students interesting in AI
1:33:55 – Meaning of life
CONNECT:
– Subscribe to this YouTube channel
– Twitter: https://twitter.com/lexfridman
– LinkedIn: https://www.linkedin.com/in/lexfridman
– Facebook: https://www.facebook.com/LexFridmanPage
– Instagram: https://www.instagram.com/lexfridman
– Medium: https://medium.com/@lexfridman
– Support on Patreon: https://www.patreon.com/lexfridman

@lexfridman
January 27, 2026 at 4:40 am
I really enjoyed this conversation with Sergey. Here's the outline:
0:00 – Introduction
3:05 – State-of-the-art robots vs humans
16:13 – Robotics may help us understand intelligence
22:49 – End-to-end learning in robotics
27:01 – Canonical problem in robotics
31:44 – Commonsense reasoning in robotics
34:41 – Can we solve robotics through learning?
44:55 – What is reinforcement learning?
1:06:36 – Tesla Autopilot
1:08:15 – Simulation in reinforcement learning
1:13:46 – Can we learn gravity from data?
1:16:03 – Self-play
1:17:39 – Reward functions
1:27:01 – Bitter lesson by Rich Sutton
1:32:13 – Advice for students interesting in AI
1:33:55 – Meaning of life
@deepg2477
January 27, 2026 at 4:40 am
Lex always sound like he didnt get a good sleep
@erniea5843
January 27, 2026 at 4:40 am
Damn I miss the AI version of Lex’s podcast. The intellectual bar has come down since he’s gone mainstream 😢
@jewishgenes
January 27, 2026 at 4:40 am
He said fetch the coffee for me hahaha. Nigga android but I toured for 15 years & played at a high level.
It will be people like us helping robots understand androids like this guy 😂
@jewishgenes
January 27, 2026 at 4:40 am
This guy is knowledgeable yet so fucking frustrated he can”t relax to experience this experience of self.
@achunaryan3418
January 27, 2026 at 4:40 am
Sup vs unsup = On pol vs off pol? No its more than that.
@NeuroReview
January 27, 2026 at 4:40 am
Rating: 7.6/10
In Short: Robots are Key to Intelligence
Notes: Sergey is a very well spoken computer scientist 'nerd' (meant in a good way!), but can speak at a normal level and explain things very well. He made some great analogies, especially connecting to Lex's glacier idea, and also his own about broken plates, to describe some complex ideas. He was clearly very thoughtful about Lex's questions (which were well done and less added on then other pods at this time). Something that really came out from this was Sergeys excitement to the idea of 'understanding intelligence', and how that is the main goal of his work studying robotics and machine learning. He gave off a very humble vibe, and was a really good guest. Would have loved a longer podcast, and feel that this is an underrated conversation.
@miroslavdyer-wd1ei
January 27, 2026 at 4:40 am
I believe lex's podcasts form an historically important AIML archive about bleeding edge research
@miroslavdyer-wd1ei
January 27, 2026 at 4:40 am
lex always looks a little baked…
@TheArthurAbbott
January 27, 2026 at 4:40 am
Bro asks the meaning of life. Smirks.
Life is full of meaning. How dare we be so reductionist to pick just one and hold onto it like the gospel truth. Is it to live? To love? To make an (positive) impact? To be happy? It doesn't boil down.
But he smirks and asks the question, knowing the chance for a good response is well worth the ask.
@tablechairlamp
January 27, 2026 at 4:40 am
All the time, He looks downwards. Doesn't look at Lex in the eyes and answer. Classic textbook nerd and genius, Levine is.
@mj2068
January 27, 2026 at 4:40 am
the man is like the definition of a sharp man… in every way. i bet he's even tall😊
@skipperkongen
January 27, 2026 at 4:40 am
how legal would it be to build a robot, big or tiny, and let it roam freely outside where it could gather experience and have to survive on its own, e.g. keep power on its batteries? How legal would that be? Under what conditions would it be legal/illegal let alone ethical?
@skipperkongen
January 27, 2026 at 4:40 am
Is simulation learning a thing? We start by learning in a simulator, then we use the policy in the real world and learn that it behaves different there compared to the simulator. That means there is an error in the simulator that we then make a step to fix and repeat?
@skipperkongen
January 27, 2026 at 4:40 am
Do humans learn to wash the dishes by starting with an off-policy method of imitating our parents and then take it from there? Is it because breaking the dishes is associated with such a strong anti-reward, i.e. getting yelled at, that we burn that behaviour path to the ground and never use it again?
@hineko_
January 27, 2026 at 4:40 am
is it possible for jew to invite a non-jew, i wonder?
@burkebaby
January 27, 2026 at 4:40 am
Sergey is one of the few Deep Learning researchers I have a huge appreciation. He is brilliant, prolific and an incredible communicator in one of the hardest areas of AI.
@junweidong2448
January 27, 2026 at 4:40 am
The most SHARPE question of people from classical control theory is:"what can you ganrantee from pure learning?"— what is the answer.
@cogoid
January 27, 2026 at 4:40 am
Well put: 33:08 …the systems that we have now simply inhabit a different universe. [of pixels and sentences] And if we build AI systems that are forced to deal with … our universe maybe they will have to acquire our common sense…
@randpaul9863
January 27, 2026 at 4:40 am
I find that many of my favorite guests on this show are very humble
@TheAIEpiphany
January 27, 2026 at 4:40 am
I find it beautiful that Isaac Asimov had a profound impact on Sergey's passion for the field. It's funny how people have been influencing each other since the dawn of civilization.
@Eltopshottah
January 27, 2026 at 4:40 am
*intensely whistles *
@denniswigand8066
January 27, 2026 at 4:40 am
Really great! Sergey is a huge inspiration!
@TheRohr
January 27, 2026 at 4:40 am
My wild guess would be that it is indeed actually easier to built the universe than to build a brain (1:12:00).
@vast634
January 27, 2026 at 4:40 am
I think the idea of using a human operator to steer the robot – only using the robots sensors and actuator commands – makes sense, to get a sense of the available feedback and control that the robot offers. And potential problems if those are not sufficient. Also to determine the concrete problem space an AI has to operate it. if a human operator cannot solve the problem, its unlikely that the robot could (apart from problems that require very quick reactions).
@shoubhikdasguptadg9911
January 27, 2026 at 4:40 am
Lex "I will insert philosophy unnecessarily in to every thing " Fridman
@GeorgeHennegar
January 27, 2026 at 4:40 am
I love Lex's podcasts because if I feel desperate for meaning, I can pick any given podcast, fastforward to the end and find a potential answer.
@rogerab1792
January 27, 2026 at 4:40 am
1:21:40
@yuvalperry6688
January 27, 2026 at 4:40 am
Common sense is a property of experience
@margrietoregan828
January 27, 2026 at 4:40 am
With the greatest of respect but both Lex & Sergey are here unknowingly & sadly irredeemably, conflating apples & oranges, as ‘computation’ (‘counting with digits’) & ‘thinking’ (‘intelligent use of information’) are two entirely different phenomena. Correct answers & solutions to robotics & separately to AI, will not be forthcoming until the all-important distinctions between these two domains are fully recognised & in which domains investigators, engineers & technicians comfortably & competently work.
‘Computation’ is nothing more than ‘counting’, for which ‘digits’ make perfectly adequate stand-ins for whatever multiplicity of identical entities is being counted/calculated/computed.
Computers are nothing more than glorified abacuses – vastly accelerated, massively miniaturised, user-friendlied, now globally interconnected, electronically automated abacuses. ABACUSES. BEAN COUNTING MACHINES. A fact which George Gilder points out in ‘Life After Google’. Even IIT advocate, Christof Koch in ‘The Feeling of Life Itself’ says consciousness (that is real thinking) cannot be computed into being. As also among a growing number of others, does Sir Roger Penrose.
As such – as beads on strings – ‘digits’ cannot be utilised as the means with which to ‘think’, the latter requiring real, true, bona fide ‘information’. Indeed, ‘digits’’ inadequacy in this ‘intelligent thinking’ regard obtains no matter how many of them one has at one’s disposal, nor how cleverly arranged they are, nor how large, powerful & globally interconnected are the machines & devices operating on them. Let alone ‘intelligently’.
Starting with our own fingers & toes & moving on to scratch marks on prison walls, ‘digits’ (sic) are excellent instruments with which to count, calculate & compute, but, again, they are not units of information, & they cannot be utilised as aids to real, true bona fide ‘thinking’.
‘Thinking’ is ‘using information (not digits) to guide & direct action, such action as is executed in regard to whatever object/s &/or event/s initially gave rise to the information being so utilised’.
On the other hand & by this definition of ‘thinking’ (as opposed to merely ‘counting’/‘calculating’/‘computing’) anything that is equipped to first take in some certain kind & amount of real information from its surrounds – temperature, humidity & barometric pressure, say – & then use this information to guide & direct those of its actions which it then executes strictly in regard to those very objects & events which initially gave rise to the information it is so using, is ‘thinking’ – not just counting.
As several others have already noted, our own climate control systems think. As do our robovacs etc.
Even though these systems genuinely ‘think’, they do not do so ‘intelligently’. Only ‘wholes’ can think intelligently. All of our own otherwise perfectly properly ‘thinking’ servomechanisms – all of our own humanly-invented & manufactured cybernetic machines, tools, systems, & devices – are what Richard Dawkins correctly labelled ‘extended phenotypes’. They are components of our own intelligent use of the information they routinely gather & use within their many & various operations, but they do not do so in their own regard.
‘Intelligence’ is ‘thinking’ in a fitness maximising/optimising manner. That is to say, ‘intelligence’ is using information to guide and direct action/behaviour in a fitness maximising/optimising manner, such action as is executed by the thinking/acting agent in regard to whatever object/s &/or event’s having given rise to the information being so used in the first instance’.
Although it has been my most dubious fortune to have figured out ‘information’s’ correct ontological identity, I am not going to divulge it here in this YouTube comment (otherwise I’d have to kill you, as the saying goes), but I can most certainly assure you that once ‘information’s’ correct ontological identity is properly recognised, no great difficulty whatsoever attends the run-on tasks of further determining the ontological identities of all of the directly information-related phenomena such as ‘thought’, ‘mind’, ‘intelligence’ and ‘consciousness’ – & further, utilising information’s’ correct identity also enables us to clearly & easily distinguish between the two realms being conflated here – information & real thinking, and, digits & mere counting.
@edgarvolkov8631
January 27, 2026 at 4:40 am
Sorry for my romanticized comment
@henrikjohnsson7403
January 27, 2026 at 4:40 am
This was great! Thank you!
@sohamdats
January 27, 2026 at 4:40 am
This guy had the maximum number of papers in NIPS 2019 which is 12. That says something.
@rquevedo9264
January 27, 2026 at 4:40 am
1:14:12 "Do you think we can learn gravity from just data?":
well…yes? https://www.youtube.com/watch?v=LMb5tvW-UoQ&t=3s
@vivekmittal1454
January 27, 2026 at 4:40 am
I was waiting for this for soo long.
@dr.markbowler-smith1662
January 27, 2026 at 4:40 am
It strikes me that the robot wouldn't break so many dishes if it attributed value to physical objects. Maybe even the idea that dish one is worth x but dish two is worth x^2.
@Stwinky
January 27, 2026 at 4:40 am
Sergey is not only a great researcher, but an amazing instructor. He can take complex aspects of RL and explain them like I’m five.
@hesbonkiptoo1849
January 27, 2026 at 4:40 am
Is it just me or Sergey Levine looks like The Riddled
@Freedom4urlife
January 27, 2026 at 4:40 am
Your interviews are really good and much info,
But need some humor and comic to run
@Torterra_ghahhyhiHd
January 27, 2026 at 4:40 am
actually, i love to build an intelligence without any propose and ambition. because every filosofy stuff etc we thin is right may is really not. could be so much of prejudice inflated by ego or fear of wasted effort, o fear to die, etc. we have so limited real deep understanding. that we assume die i not good a life is good. and if it is wrong the go to the opposite directions. our mind is still mindless compared to real deep meaning in the universe. we say we are aware, we are not aware of the real values even the moral value, etc is full of bugs. we just intent to approach. so every wrong conception on filosofical stuff that we put in the ia it will be exponentially, reflected in the derived way and calculus an integrals problem magnify in the world. keep it pragmatic, keep it simple. do not create firearms. do not realize any complex task like save the world, eliminate poverty or hunger, suppose to get a "mind", be "kind", "don't be kind" (this base is not solid and part of this construction is prejudiced itself in our filosofical conception), etc. "love", "hate".
@frankiethefish73
January 27, 2026 at 4:40 am
Has anyone else noticed that Sergey speaks a little bit like Lorne Michaels, (aka Dr. Evil)?
Comments are closed.