PhD Defense: From Demonstration to Dynamic Interaction: Enabling Long-Term Robotic Planning

Talk
Mara Levy
Time: 
05.09.2025 14:00 to 16:00

Robotic learning has seen rapid growth over the past decade, driven by advances in machine learning that have brought real-world deployment of robots closer to reality. Research in this area primarily falls into two categories: reinforcement learning and imitation learning. Despite their promise, both approaches face significant challenges, including limited data availability and the difficulty of obtaining accurate state representations. This thesis explores how we can advance these methods to enable robust performance in real-world, unstructured environments.
We begin by exploring how to redefine state representation, presenting two complementary approaches. The first focuses on human state representation but is easily extendable to robots. It significantly outperforms existing methods in generalizing to unseen states and varying camera viewpoints. The second approach introduces a more concise, keypoint-based representation. We show that this method enables training of robot policies with minimal demonstrations and generalizes effectively to new environments and objects of varying shapes and sizes.
Next, we turn to the problem of learning policies from a single demonstration, without relying on handcrafted reward functions. Remarkably, our method achieves comparable final performance to existing approaches while using 100× less data. Finally, we demonstrate how these methods can be deployed in dynamic environments, even when trained under static conditions. By layering a lightweight planner on top of a pretrained policy, we achieve substantial improvements over naïve replanning strategies, approaching oracle-level success rates.