Robots Doing Yoga

DrEureka blends large language models with reinforcement learning to simplify teaching robots complex skills, like balancing on a yoga ball.

nickbild
over 1 year ago Robotics
This robot learned to balance on a yoga ball with DrEureka (📷: J. Ma)

Training an autonomous robot to carry out any complex task is as much an art as it is a science. It involves trial and error, and lots of parameter tweaking by trained individuals that get a feel for the work through years of experience. Before autonomous robots can become commonplace helpers in our daily lives, we will need to develop better, more efficient, ways to teach them new skills.

Large language models (LLMs), such as GPT-4, have been explored recently as a better way to train robots. These algorithms excel at high-level semantic planning and generating context-aware responses, making them well-suited for guiding robots through complex tasks, in theory. However, using LLMs for low-level manipulation tasks still remains challenging, as it requires extensive domain-specific knowledge to create effective prompts and teach complex motor skills.

Reinforcement learning (RL), on the other hand, is a powerful method for teaching robots complex tasks through trial and error and learning from rewards. It has proven effective in achieving dexterous control in low-level manipulation tasks. RL can adapt to different environments and tasks, making it a very versatile option. Nonetheless, designing effective reward functions in RL can be a complicated process, often involving a lot of manual tweaking and the potential for unintended outcomes. Sparse rewards in real-world tasks can further hinder learning and slow down the process.

A tool called EUREKA was recently developed by researchers at NVIDIA in collaboration with their partners in academia that seeks to blend the best aspects of LLMs and RL to simplify robot learning. EUREKA leverages the knowledge of the world that is encoded in cutting-edge LLMs to produce an ideal reward function. That reward function can then be used by an RL algorithm to teach robots complex skills without needing intervention and fine-tuning by human experts.

Using EUREKA, a research group at the University of Pennsylvania has now developed an algorithm, called DrEureka, that can train robot dogs to do some very impressive things. In the initial demonstration, the robot was seen balancing on top of a yoga ball.

This feat was accomplished by training the robot dog entirely in a simulated environment. After training the DrEureka model with the help of EUREKA, it was deployed to the physical robot. Amazingly, the first time the robot was put on top of a yoga ball, it just worked. No fine-tuning was needed, no additional training data had to be collected, and there were no broken parts to repair. The system also proved itself to be robust as it maintained proper operation as different types of terrain, and other disturbances, were introduced.

Looking to the future, the team is planning to further enhance DrEureka so that it will be suitable for even more use cases. They note that while they trained their robot entirely in simulation, feedback from real-world experiments would further increase the algorithm’s accuracy. They also believe that bringing in additional sensing modalities could be beneficial.

nickbild

R&D, creativity, and building the next big thing you never knew you wanted are my specialties.

Latest Articles