Enhancing Reinforcement Learning with Error-Prone Language Models

GHC 6501

The correct specification of reward models is a well-known challenge in reinforcement learning. Hand-crafted reward functions, which are usually sparse, often lead to inefficient or suboptimal policies, misalignment with user values, or difficulties in attributing credit or blame within multi-agent systems. Reinforcement learning from human feedback is a successful technique that can mitigate such issues [...]

Efficient Multi-Agent Motion Planning using Local Policies

WEH 4625

Abstract: Teams of multiple robots working together can achieve challenging tasks like warehouse automation, search and rescue, and cooperative construction. However, finding efficient collision-free motions for all agents is extremely challenging as the complexity of the multi-agent motion planning (MAMP) problem grows exponentially with the number of agents. Multi-Agent Path Finding (MAPF) is a subset [...]

Foundation Control Model for General Embodied Intelligence

3305 Newell-Simon Hall

Abstract: With the growing accessibility of humanoid hardware and rapid advances in foundation models, we are entering an era where achieving general embodied intelligence is within reach—enabling humanoid robots to perform a wide range of tasks in human-centric environments. Despite significant progress in language and vision foundation models, controlling humanoids with high degrees of freedom [...]

Learning Humanoid Robots from Simulation to Real to Simulation

Gates Hillman Center 6115

Abstract: How do we teach humanoid robots to move like humans—and do so reliably in the real world? In this talk, I’ll share my journey in building a learning-based pipeline that closes the loop between simulation and reality for humanoid whole-body control. Starting from real-time teleoperation (H2O), to scalable data humanoid collection (OmniH2O), to learning [...]