November 08, 2021
Specifying tasks with videos is a powerful technique towards acquiring novel and general robot skills. However, reasoning over mechanics and dexterous interactions can make it challenging to scale learning contact-rich manipulation. In this work, we focus on the problem of visual non-prehensile planar manipulation: given a video of an object in planar motion, find contact-aware robot actions that reproduce the same object motion. We propose a novel architecture, Differentiable Learning for Manipulation (DLM), that combines video decoding neural models with priors from contact mechanics by leveraging differentiable optimization and finite difference based simulation. Through extensive simulated experiments, we investigate the interplay between traditional model-based techniques and modern deep learning approaches. We find that our modular and fully differentiable architecture performs better than learning-only methods on unseen objects and motions. https://github.com/baceituno/dlm.
Written by
Bernardo Aceituno
Alberto Rodriguez
Shubham Tulsiani
Abhinav Gupta
Mustafa Mukadam
Publisher
CoRL
June 04, 2023
Dahyun Kang, Peter Koniusz, Minsu Cho, Naila Murray
June 04, 2023
May 09, 2023
Rohit Girdhar, Alaa El-Nouby, Zhuang Liu, Mannat Singh, Kalyan Vasudev Alwala, Armand Joulin, Ishan Misra
May 09, 2023
May 04, 2023
Nicklas Hansen, Yixin Lin, Hao Su, Xiaolong Wang, Vikash Kumar, Aravind Rajeswaran
May 04, 2023
April 20, 2023
Xubo Liu, Egor Lakomkin, Dino Vougioukas, Pingchuan Ma, Honglie Chen, Ruiming Xie, Morrie Doulaty, Niko Moritz, Jachym Kolar, Stavros Petridis, Maja Pantic, Christian Fuegen
April 20, 2023
Latest Work
Our Actions
Newsletter