DiscoverAI BreakdownLearning from Reward-Free Offline Data: A Case for Planning with Latent Dynamics Models
Learning from Reward-Free Offline Data: A Case for Planning with Latent Dynamics Models

Learning from Reward-Free Offline Data: A Case for Planning with Latent Dynamics Models

Update: 2025-08-15
Share

Description

In this episode, we discuss Learning from Reward-Free Offline Data: A Case for Planning with Latent Dynamics Models by Vlad Sobal, Wancong Zhang, Kyunghyun Cho, Randall Balestriero, Tim G. J. Rudner, Yann LeCun. The paper compares model-free reinforcement learning and model-based control methods for solving navigation tasks using offline, reward-free data. It finds that reinforcement learning performs best with large, high-quality datasets, while model-based planning with latent dynamics models generalizes better to new environments and handles suboptimal data more efficiently. Overall, latent model-based planning is highlighted as a robust approach for offline learning and adapting to diverse tasks.
Comments 
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

Learning from Reward-Free Offline Data: A Case for Planning with Latent Dynamics Models

Learning from Reward-Free Offline Data: A Case for Planning with Latent Dynamics Models

agibreakdown