Robot Technology News  
ROBO SPACE
A simpler path to better computer vision
by Adam Zewe | MIT News
Boston MA (SPX) Nov 24, 2022

Researchers used a large collection of simple, un-curated synthetic image generation programs to pretrain a computer vision model for image classification. The researchers didn't curate or alter the programs, which each comprised just a few lines of code. In this image, the image sets in each row were produced using 3 different image generation programs.

Before a machine-learning model can complete a task, such as identifying cancer in medical images, the model must be trained. Training image classification models typically involves showing the model millions of example images gathered into a massive dataset.

However, using real image data can raise practical and ethical concerns: The images could run afoul of copyright laws, violate people's privacy, or be biased against a certain racial or ethnic group. To avoid these pitfalls, researchers can use image generation programs to create synthetic data for model training. But these techniques are limited because expert knowledge is often needed to hand-design an image generation program that can create effective training data.

Researchers from MIT, the MIT-IBM Watson AI Lab, and elsewhere took a different approach. Instead of designing customized image generation programs for a particular training task, they gathered a dataset of 21,000 publicly available programs from the internet. Then they used this large collection of basic image generation programs to train a computer vision model.

These programs produce diverse images that display simple colors and textures. The researchers didn't curate or alter the programs, which each comprised just a few lines of code.

The models they trained with this large dataset of programs classified images more accurately than other synthetically trained models. And, while their models underperformed those trained with real data, the researchers showed that increasing the number of image programs in the dataset also increased model performance, revealing a path to attaining higher accuracy.

"It turns out that using lots of programs that are uncurated is actually better than using a small set of programs that people need to manipulate. Data are important, but we have shown that you can go pretty far without real data," says Manel Baradad, an electrical engineering and computer science (EECS) graduate student working in the Computer Science and Artificial Intelligence Laboratory (CSAIL) and lead author of the paper describing this technique.

Co-authors include Tongzhou Wang, an EECS grad student in CSAIL; Rogerio Feris, principal scientist and manager at the MIT-IBM Watson AI Lab; Antonio Torralba, the Delta Electronics Professor of Electrical Engineering and Computer Science and a member of CSAIL; and senior author Phillip Isola, an associate professor in EECS and CSAIL; along with others at JPMorgan Chase Bank and Xyla, Inc. The research will be presented at the Conference on Neural Information Processing Systems.

Rethinking pretraining
Machine-learning models are typically pretrained, which means they are trained on one dataset first to help them build parameters that can be used to tackle a different task. A model for classifying X-rays might be pretrained using a huge dataset of synthetically generated images before it is trained for its actual task using a much smaller dataset of real X-rays.

These researchers previously showed that they could use a handful of image generation programs to create synthetic data for model pretraining, but the programs needed to be carefully designed so the synthetic images matched up with certain properties of real images. This made the technique difficult to scale up.

In the new work, they used an enormous dataset of uncurated image generation programs instead.

They began by gathering a collection of 21,000 images generation programs from the internet. All the programs are written in a simple programming language and comprise just a few snippets of code, so they generate images rapidly.

"These programs have been designed by developers all over the world to produce images that have some of the properties we are interested in. They produce images that look kind of like abstract art," Baradad explains.

These simple programs can run so quickly that the researchers didn't need to produce images in advance to train the model. The researchers found they could generate images and train the model simultaneously, which streamlines the process.

They used their massive dataset of image generation programs to pretrain computer vision models for both supervised and unsupervised image classification tasks. In supervised learning, the image data are labeled, while in unsupervised learning the model learns to categorize images without labels.

Improving accuracy
When they compared their pretrained models to state-of-the-art computer vision models that had been pretrained using synthetic data, their models were more accurate, meaning they put images into the correct categories more often. While the accuracy levels were still less than models trained on real data, their technique narrowed the performance gap between models trained on real data and those trained on synthetic data by 38 percent.

"Importantly, we show that for the number of programs you collect, performance scales logarithmically. We do not saturate performance, so if we collect more programs, the model would perform even better. So, there is a way to extend our approach," Manel says.

The researchers also used each individual image generation program for pretraining, in an effort to uncover factors that contribute to model accuracy. They found that when a program generates a more diverse set of images, the model performs better. They also found that colorful images with scenes that fill the entire canvas tend to improve model performance the most.

Now that they have demonstrated the success of this pretraining approach, the researchers want to extend their technique to other types of data, such as multimodal data that include text and images. They also want to continue exploring ways to improve image classification performance.

"There is still a gap to close with models trained on real data. This gives our research a direction that we hope others will follow," he says.

Research Report:"Procedural Image Programs for Representation Learning"


Related Links
Computer Science and Artificial Intelligence Laboratory
All about the robots on Earth and beyond!


Thanks for being here;
We need your help. The SpaceDaily news network continues to grow but revenues have never been harder to maintain.

With the rise of Ad Blockers, and Facebook - our traditional revenue sources via quality network advertising continues to decline. And unlike so many other news sites, we don't have a paywall - with those annoying usernames and passwords.

Our news coverage takes time and effort to publish 365 days a year.

If you find our news sites informative and useful then please consider becoming a regular supporter or for now make a one off contribution.
SpaceDaily Contributor
$5 Billed Once


credit card or paypal
SpaceDaily Monthly Supporter
$5 Billed Monthly


paypal only


ROBO SPACE
Robotic In-Space Mechanic aces tests, on track for launch
Washington DC (SPX) Nov 21, 2022
All component-level tests are complete on DARPA's Robotic Servicing of Geosynchronous Satellites (RSGS) program and the on-orbit demonstration mission is on schedule for launch in 2024. The RSGS goal is to enable inspection and servicing of satellites in geosynchronous Earth orbit (GEO), where hundreds of satellites provide communications, meteorological, national security, and other vital functions. Currently, no options exist for visual diagnosis, upgrades, or repairs of a malfunctioning satellite's c ... read more

Comment using your Disqus, Facebook, Google or Twitter login.



Share this article via these popular social media networks
del.icio.usdel.icio.us DiggDigg RedditReddit GoogleGoogle

ROBO SPACE
Northrop Grumman's C-UAS system of systems architecture excels during complex live fire tests

Taiwan's Chien Hsiang loitering drone was designed to destroy enemy radar and UAVs

China's MD-22 hypersonic UAV could be strategic asset in near- or outer-space

RDARS Eagle Nest Autonomous Drone-In-a-Box solution supports SpaceX Starlink Satellite Communications

ROBO SPACE
French-Lebanese architect seeks pro-climate construction transformation

Quandum Aerospace tested Zortrax resin 3D Printing Ecosystem

Talks kick off on global plastic trash treaty

Scientists demonstrate continuous-wave lasing of deep-ultraviolet laser diode at room temps

ROBO SPACE
NIST finds a sweet new way to print microchip patterns on curvy surfaces

US chip ban on China a dangerous game for all

A possible game changer for next generation microelectronics

NIST's grid of quantum islands could reveal secrets for powerful technologies

ROBO SPACE
Ukraine nuclear plants reconnect to grid after strikes: ministry

Ukraine nuclear plants cut from grid after strikes: operator

Kyrgyzstan mulls building nuclear plant with Russian help

Argonne awarded $6 million to develop technologies for recycling nuclear fuel

ROBO SPACE
Bangkok shrine bombing trial resumes after long delays

Sweden changes constitution to beef up anti-terror law

Swedish MPs to vote next week on beefing up anti-terror laws

Oldest Guantanamo Bay prisoner released to Pakistan

ROBO SPACE
Britain U-turns over energy-saving publicity drive; Legal challenge over Finnish climate inaction

Ottawa rolls out CAN$1.6 bn plan to adapt to climate change

Biden's ambitious climate plan stokes tension with EU allies

US to release emergency aid for Ukraine energy infrastructure

ROBO SPACE
Reversing the charge

China tests new Tianzhou fuel cell on route to Tiangong Station

POWER aims to create revolutionary power distribution network

Generating electricity from tacky tape

ROBO SPACE
China to launch Shenzhou XV on Tuesday

China aims to establish new global partnership in space exploration, innovation: CNSA

Space belongs to humanity, come share it with China

Shenzhou XV to begin mission to space soon









The content herein, unless otherwise known to be public domain, are Copyright 1995-2024 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us.