AI researchers say they’ve created a framework for controlling four-legged robots that promises better energy efficiency and adaptability than more traditional model-based gait control of robotic legs. To demonstrate the robust nature of the framework that adjusts to conditions in real time, AI researchers made the system slip on frictionless surfaces to mimic a banana peel, ride a skateboard, and climb on a bridge while walking on a treadmill. An Nvidia company spokesperson told VentureBeat that only the frictionless surface test was conducted in real life due to limits placed on office staff size due to COVID-19 and that all other challenges took place in a simulation. Simulations are often used as training data for robotics systems before they’re used in robots in real life.
“Our framework learns a controller that can adapt to challenging environmental changes on the fly, including novel scenarios not seen during training. The learned controller is up to 85 percent more energy efficient and is more robust compared to baseline methods,” the paper reads. “At inference time, the high-level controller needs only evaluate a small multi-layer neural network, avoiding the use of an expensive model predictive control (MPC) strategy that might otherwise be required to optimize for long-term performance.”
The quadruped model is trained in simulation with a split-belt treadmill whose two tracks can change speed independently. That training in simulation is then transferred to a Laikago robot in the real world. Nvidia released video of simulations and laboratory work Monday. Also unveiled by Nvidia Monday: AI-powered videoconferencing service Maxine, and the Omniverse simulated environment for engineers in beta.
A paper detailing the framework for controlling quadruped legs was published one week ago on preprint repository arXiv. AI researchers from Nvidia; Caltech; University of Texas, Austin; and the Vector Institute at the University of Toronto contributed to the paper about the controller framework, which combines a high-level controller that uses reinforcement learning with a model-based lower level controller.
“By leveraging the advantages of both paradigms, we obtain a contact-adaptive controller that is more robust and energy efficient than those employing a fixed contact sequence,” the paper reads.
Researchers argue that a number of networks for controlling robotic legs are fixed and therefore unable to adapt to new circumstance, while adaptive networks are often energy intensive. Locomotion systems created with reinforcement learning are often less robust than model-based approaches, need a lot of training samples, or use a complicated approach to rewarding agents.
In similar news, earlier this year AI researchers from ETH Zurich detailed DeepGait, AI trained with reinforcement learning to do things like bridge unusually long gaps and walk on uneven terrain, at the International Conference on Robotics and Automation (CRA).