TalkRL: The Reinforcement Learning Podcast

Nathan Lambert on Model-based RL, Trajectory-based models, Quadrotor control, Hyperparameter Optimization for MBRL, RL vs PID control, and more!

Show Notes

Nathan Lambert is a PhD Candidate at UC Berkeley. 


Featured References 

Learning Accurate Long-term Dynamics for Model-based Reinforcement Learning 
Nathan O. Lambert, Albert Wilcox, Howard Zhang, Kristofer S. J. Pister, Roberto Calandra 

Objective Mismatch in Model-based Reinforcement Learning 
Nathan Lambert, Brandon Amos, Omry Yadan, Roberto Calandra 

Low Level Control of a Quadrotor with Deep Model-Based Reinforcement Learning 
Nathan O. Lambert, Daniel S. Drew, Joseph Yaconelli, Roberto Calandra, Sergey Levine, Kristofer S.J. Pister 

On the Importance of Hyperparameter Optimization for Model-based Reinforcement Learning 
Baohe Zhang, Raghu Rajan, Luis Pineda, Nathan Lambert, André Biedenkapp, Kurtland Chua, Frank Hutter, Roberto Calandra 


Additional References 

Creators & Guests

Host
Robin Ranjit Singh Chauhan
🌱 Head of Eng @AgFunder 🧠 AI:Reinforcement Learning/ML/DL/NLP🎙️Host @TalkRLPodcast 💳 ex-@Microsoft ecomm PgmMgr 🤖 @UWaterloo CompEng 🇨🇦 🇮🇳

What is TalkRL: The Reinforcement Learning Podcast?

TalkRL podcast is All Reinforcement Learning, All the Time.
In-depth interviews with brilliant people at the forefront of RL research and practice.
Guests from places like MILA, OpenAI, MIT, DeepMind, Berkeley, Amii, Oxford, Google Research, Brown, Waymo, Caltech, and Vector Institute.
Hosted by Robin Ranjit Singh Chauhan.