AI teaches robot to balance on ball: It teaches the robot better than humans

May 8, 2024  12:19

A group of scientists from the University of Pennsylvania has developed the DrEureka system designed for training robots using large language models of artificial intelligence like OpenAI GPT-4. It turns out to be a more efficient method than a sequence of real-world tasks, but it requires special attention from humans due to the peculiarities of AI "thinking."

The DrEureka platform (Domain Randomization Eureka) has proven its effectiveness using the Unitree Go1 robot - a quadrupedal open-source machine. It involves training the robot in a simulated environment by randomizing key variables such as friction coefficients, mass, damping, center of gravity offset, and other parameters. Based on several user requests, the AI generates code describing the reward and penalty system for training the robot in a virtual environment. After each simulation, the AI analyzes how well the virtual robot performed the task and how its performance can be improved. It's important to note that the neural network can rapidly generate scenarios in large volumes and execute them simultaneously.

dreureka.jpg (288 KB)

The AI creates tasks with maximum and minimum parameter values at failure or mechanism breakdown points, achieving or exceeding which results in a reduction in the score for completing the training scenario. The research authors note that additional safety instructions are needed for the correct coding by AI; otherwise, the neural network starts to "cheat" in pursuit of maximum performance, which in the real world can lead to engine overheating or damage to the robot's limbs. In one of these unnatural scenarios, the virtual robot "discovered" that it could move faster by disabling one leg and starting to move on three.

Researchers instructed the AI to exercise special caution, considering that the trained robot would undergo tests in the real world. Therefore, the neural network created additional safety features for aspects such as smooth movements, horizontal orientation, torso height, and torque consideration for electric motors - it should not exceed the specified values. As a result, the DrEureka system succeeded in training the robot better than humans: the machine showed a 34% increase in movement speed and a 20% increase in the distance traversed over rough terrain. Researchers explained this difference in approaches by noting that when humans train, they break down the task into stages and find a solution for each of them, whereas GPT trains everything at once, a feat humans are clearly unable to replicate.

The DrEureka system has allowed a transition from simulation directly to real-world work. Project authors claim they could further enhance the platform's efficiency by providing AI with feedback from the real world - for this, neural networks would need to analyze video recordings of tests, not just error analysis in the robot's system logs. It takes an average person up to 1.5 years to learn to walk, and only a few can move on a yoga ball. The trained DrEureka robot effectively handles even this task.


 
 
 
 
  • Archive