Robot Technology News
ROBO SPACE
Can robots learn from machine dreams?
MIT CSAIL researchers (left to right) Alan Yu, an undergraduate in electrical engineering and computer science (EECS); Phillip Isola, associate professor of EECS; and Ge Yang, a postdoctoral associate, developed an AI-powered simulator that generates unlimited, diverse, and realistic training data for robots. Robots trained in this virtual environment can seamlessly transfer their skills to the real world, performing at expert levels without additional fine-tuning. Credits:Photo: Michael Grimmett/MIT CSAIL
Can robots learn from machine dreams?
by Rachel Gordon | MIT CSAIL
Boston MA (SPX) Nov 20, 2024

For roboticists, one challenge towers above all others: generalization - the ability to create machines that can adapt to any environment or condition. Since the 1970s, the field has evolved from writing sophisticated programs to using deep learning, teaching robots to learn directly from human behavior. But a critical bottleneck remains: data quality. To improve, robots need to encounter scenarios that push the boundaries of their capabilities, operating at the edge of their mastery. This process traditionally requires human oversight, with operators carefully challenging robots to expand their abilities. As robots become more sophisticated, this hands-on approach hits a scaling problem: the demand for high-quality training data far outpaces humans' ability to provide it.

Now, a team of MIT Computer Science and Artificial Intelligence Laboratory (CSAIL) researchers has developed a novel approach to robot training that could significantly accelerate the deployment of adaptable, intelligent machines in real-world environments. The new system, called "LucidSim," uses recent advances in generative AI and physics simulators to create diverse and realistic virtual training environments, helping robots achieve expert-level performance in difficult tasks without any real-world data.

LucidSim combines physics simulation with generative AI models, addressing one of the most persistent challenges in robotics: transferring skills learned in simulation to the real world. "A fundamental challenge in robot learning has long been the 'sim-to-real gap' - the disparity between simulated training environments and the complex, unpredictable real world," says MIT CSAIL postdoc Ge Yang, a lead researcher on LucidSim. "Previous approaches often relied on depth sensors, which simplified the problem but missed crucial real-world complexities."

The multipronged system is a blend of different technologies. At its core, LucidSim uses large language models to generate various structured descriptions of environments. These descriptions are then transformed into images using generative models. To ensure that these images reflect real-world physics, an underlying physics simulator is used to guide the generation process.

The birth of an idea: From burritos to breakthroughs
The inspiration for LucidSim came from an unexpected place: a conversation outside Beantown Taqueria in Cambridge, Massachusetts. ??"We wanted to teach vision-equipped robots how to improve using human feedback. But then, we realized we didn't have a pure vision-based policy to begin with," says Alan Yu, an undergraduate student in electrical engineering and computer science (EECS) at MIT and co-lead author on LucidSim. "We kept talking about it as we walked down the street, and then we stopped outside the taqueria for about half-an-hour. That's where we had our moment."

To cook up their data, the team generated realistic images by extracting depth maps, which provide geometric information, and semantic masks, which label different parts of an image, from the simulated scene. They quickly realized, however, that with tight control on the composition of the image content, the model would produce similar images that weren't different from each other using the same prompt. So, they devised a way to source diverse text prompts from ChatGPT.

This approach, however, only resulted in a single image. To make short, coherent videos that serve as little "experiences" for the robot, the scientists hacked together some image magic into another novel technique the team created, called "Dreams In Motion." The system computes the movements of each pixel between frames, to warp a single generated image into a short, multi-frame video. Dreams In Motion does this by considering the 3D geometry of the scene and the relative changes in the robot's perspective.

"We outperform domain randomization, a method developed in 2017 that applies random colors and patterns to objects in the environment, which is still considered the go-to method these days," says Yu. "While this technique generates diverse data, it lacks realism. LucidSim addresses both diversity and realism problems. It's exciting that even without seeing the real world during training, the robot can recognize and navigate obstacles in real environments."

The team is particularly excited about the potential of applying LucidSim to domains outside quadruped locomotion and parkour, their main test bed. One example is mobile manipulation, where a mobile robot is tasked to handle objects in an open area; also, color perception is critical. "Today, these robots still learn from real-world demonstrations," says Yang. "Although collecting demonstrations is easy, scaling a real-world robot teleoperation setup to thousands of skills is challenging because a human has to physically set up each scene. We hope to make this easier, thus qualitatively more scalable, by moving data collection into a virtual environment."

Who's the real expert?
The team put LucidSim to the test against an alternative, where an expert teacher demonstrates the skill for the robot to learn from. The results were surprising: Robots trained by the expert struggled, succeeding only 15 percent of the time - and even quadrupling the amount of expert training data barely moved the needle. But when robots collected their own training data through LucidSim, the story changed dramatically. Just doubling the dataset size catapulted success rates to 88 percent. "And giving our robot more data monotonically improves its performance - eventually, the student becomes the expert," says Yang.

"One of the main challenges in sim-to-real transfer for robotics is achieving visual realism in simulated environments," says Stanford University assistant professor of electrical engineering Shuran Song, who wasn't involved in the research. "The LucidSim framework provides an elegant solution by using generative models to create diverse, highly realistic visual data for any simulation. This work could significantly accelerate the deployment of robots trained in virtual environments to real-world tasks."

From the streets of Cambridge to the cutting edge of robotics research, LucidSim is paving the way toward a new generation of intelligent, adaptable machines - ones that learn to navigate our complex world without ever setting foot in it.

Research Report:Learning Visual Parkour from Generated Images

Related Links
Computer Science and Artificial Intelligence Laboratory (CSAIL)
All about the robots on Earth and beyond!

Subscribe Free To Our Daily Newsletters
Tweet

RELATED CONTENT
The following news reports may link to other Space Media Network websites.
ROBO SPACE
Smart robots poised to boost green energy efficiency with wave prediction
Edinburgh, UK (SPX) Nov 15, 2024
Underwater robots capable of predicting real-time wave movements could significantly cut the costs of offshore renewable energy production, according to a new study. Engineers from the University of Edinburgh have developed innovative computational and experimental methods that allow autonomous robots to function stably in turbulent seas. This advancement could make the maintenance of offshore wind farms and tidal turbines more efficient and safer, eliminating current challenges posed by erratic w ... read more

ROBO SPACE
NASA Armstrong builds sensor pod for autonomous flight

Silent Arrow secures Air Force contract for new 200-mile logistics drone

ANELLO Photonics secures Series B funding to advance inertial navigation technology

N. Korean leader orders 'mass production' of attack drones

ROBO SPACE
Enormous potential for rare Earth elements found in US coal ash

Scientists show how a laser beam can cast a shadow

MIT engineers make converting CO2 into useful products more practical

British Museum secures record 1bn donation of Chinese ceramics

ROBO SPACE
MIT physicists predict exotic form of matter with potential for quantum computing

US finalizes up to $6.6 bn funding for chip giant TSMC

China's top chipmaker reports surge in profits

Nvidia surpasses Apple as world's biggest company

ROBO SPACE
Framatome's PROtect fuel achieves key milestone at Gosgen Nuclear Plant in Switzerland

Framatome and Nuclearelectrica partner to produce Lutetium-177 in Romania

Vietnam looks to restart nuclear power projects

US touts Trump-proof nuclear expansion plans at COP29

ROBO SPACE
Thai military accused of beating Myanmar man to death

Syrians, Iraqis archive IS jail crimes in virtual museum

U.S. military judge rules government plea deals with 9/11 suspects are 'enforceable contracts'

Poland alleges 'foreign agents' plotted using int'l parcel couriers to target transatlantic air traffic

ROBO SPACE
Climate finance can be hard sell, says aide to banks and PMs

'Vague' net zero rules threaten climate targets, scientists warn

US says its climate financing reached $11 bn this year

Trump picks Doug Burgum as energy czar in new administration

ROBO SPACE
Spinning fusion fuel for efficiency

Tackling the energy revolution, one sector at a time

NASA opens Power Systems essay contest for K12 students

In search of high-performance materials for fusion reactors

ROBO SPACE
China's commercial space sector expands as firms outline ambitious plans

China prepares Tianzhou 8 for upcoming launch to Tiangong station

Model of Haoloong Space Cargo Shuttle to Be Unveiled at Airshow China

Shenzhou 18 brings back samples for space habitability and materials research

Subscribe Free To Our Daily Newsletters




The content herein, unless otherwise known to be public domain, are Copyright 1995-2024 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us.