Dinesh Jayaraman
Dinesh Jayaraman
Home
Research Group
Publications
Teaching
Imitation Learning
Keyframe-Focused Visual Imitation Learning
Identifying and upsampling important frames from demonstration data can significantly boost imitation learning from histories, and scales easily to complex settings such as autonomous driving from vision.
Chuan Wen
,
Jierui Lin
,
Jianing Qian
,
Yang Gao
,
Dinesh Jayaraman
SMIRL: Surprise Minimizing RL in Dynamic Environments
We formulate homeostasis as an intrinsic motivation objective and show interesting emergent behavior from minimizing Bayesian surprise with RL across many environments.
Glen Berseth
,
Daniel Geng
,
Coline Devin
,
Chelsea Finn
,
Dinesh Jayaraman
,
Sergey Levine
Fighting Copycat Agents in Behavioral Cloning from Multiple Observations.
Chuan Wen
,
Jierui Lin
,
Trevor Darrell
,
Dinesh Jayaraman
,
Yang Gao
Model-Based Inverse Reinforcement Learning from Visual Demonstrations
We learn reward functions in unsupervised object keypoint space, to allow us to follow third-person demonstrations with model-based RL.
Neha Das
,
Sarah Bechtle
,
Todor Davchev
,
Dinesh Jayaraman
,
Akshara Rai
,
Franziska Meier
Long-Horizon Visual Planning with Goal-Conditioned Hierarchical Predictors
To plan towards long-term goals through visual prediction, we propose a model based on two key ideas: (i) predict in a goal-conditioned way to restrict planning only to useful sequences, and (ii) recursively decompose the goal-conditioned prediction task into an increasingly fine series of subgoals.
Karl Pertsch
,
Oleg Rybkin*
,
Frederik Ebert
,
Chelsea Finn
,
Dinesh Jayaraman
,
Sergey Levine
Causal Confusion in Imitation Learning
“Causal confusion”, where spurious correlates are mistaken to be causes of expert actions, is commonly prevalent in imitation learning, leading to counterintuitive results where additional information can lead to worse task performance. How might one address this?
Pim de Haan
,
Dinesh Jayaraman
,
Sergey Levine
Time-Agnostic Prediction: Predicting Predictable Video Frames
In visual prediction tasks, letting your predictive model choose which times to predict does two things: (i) improves prediction quality, and (ii) leads to semantically coherent “bottleneck state” predictions, which are useful for planning.
Dinesh Jayaraman
,
Frederik Ebert
,
Alexei A Efros
,
Sergey Levine
Cite
×