Robot Technology News  
ROBO SPACE
Unpacking the "black box" to build better AI models
by Adam Zewe for MIT News
Boston MA (SPX) Jan 09, 2023

Stefanie Jegelka, a newly-tenured associate professor in the Department of Electrical Engineering and Computer Science at MIT, develops algorithms for deep learning applications and studies how deep learning models behave and what they can learn. Photo: M. Scott Brauer

When deep learning models are deployed in the real world, perhaps to detect financial fraud from credit card activity or identify cancer in medical images, they are often able to outperform humans.

But what exactly are these deep learning models learning? Does a model trained to spot skin cancer in clinical images, for example, actually learn the colors and textures of cancerous tissue, or is it flagging some other features or patterns?

These powerful machine-learning models are typically based on artificial neural networks that can have millions of nodes that process data to make predictions. Due to their complexity, researchers often call these models "black boxes" because even the scientists who build them don't understand everything that is going on under the hood.

Stefanie Jegelka isn't satisfied with that "black box" explanation. A newly tenured associate professor in the MIT Department of Electrical Engineering and Computer Science, Jegelka is digging deep into deep learning to understand what these models can learn and how they behave, and how to build certain prior information into these models.

"At the end of the day, what a deep-learning model will learn depends on so many factors. But building an understanding that is relevant in practice will help us design better models, and also help us understand what is going on inside them so we know when we can deploy a model and when we can't. That is critically important," says Jegelka, who is also a member of the Computer Science and Artificial Intelligence Laboratory (CSAIL) and the Institute for Data, Systems, and Society (IDSS).

Jegelka is particularly interested in optimizing machine-learning models when input data are in the form of graphs. Graph data pose specific challenges: For instance, information in the data consists of both information about individual nodes and edges, as well as the structure - what is connected to what. In addition, graphs have mathematical symmetries that need to be respected by the machine-learning model so that, for instance, the same graph always leads to the same prediction. Building such symmetries into a machine-learning model is usually not easy.

Take molecules, for instance. Molecules can be represented as graphs, with vertices that correspond to atoms and edges that correspond to chemical bonds between them. Drug companies may want to use deep learning to rapidly predict the properties of many molecules, narrowing down the number they must physically test in the lab.

Jegelka studies methods to build mathematical machine-learning models that can effectively take graph data as an input and output something else, in this case a prediction of a molecule's chemical properties. This is particularly challenging since a molecule's properties are determined not only by the atoms within it, but also by the connections between them.

Other examples of machine learning on graphs include traffic routing, chip design, and recommender systems.

Designing these models is made even more difficult by the fact that data used to train them are often different from data the models see in practice. Perhaps the model was trained using small molecular graphs or traffic networks, but the graphs it sees once deployed are larger or more complex.

In this case, what can researchers expect this model to learn, and will it still work in practice if the real-world data are different?

"Your model is not going to be able to learn everything because of some hardness problems in computer science, but what you can learn and what you can't learn depends on how you set the model up," Jegelka says.

She approaches this question by combining her passion for algorithms and discrete mathematics with her excitement for machine learning.

From butterflies to bioinformatics
Jegelka grew up in a small town in Germany and became interested in science when she was a high school student; a supportive teacher encouraged her to participate in an international science competition. She and her teammates from the U.S. and Singapore won an award for a website they created about butterflies, in three languages.

"For our project, we took images of wings with a scanning electron microscope at a local university of applied sciences. I also got the opportunity to use a high-speed camera at Mercedes Benz - this camera usually filmed combustion engines - which I used to capture a slow-motion video of the movement of a butterfly's wings. That was the first time I really got in touch with science and exploration," she recalls.

Intrigued by both biology and mathematics, Jegelka decided to study bioinformatics at the University of Tubingen and the University of Texas at Austin. She had a few opportunities to conduct research as an undergraduate, including an internship in computational neuroscience at Georgetown University, but wasn't sure what career to follow.

When she returned for her final year of college, Jegelka moved in with two roommates who were working as research assistants at the Max Planck Institute in Tubingen.

"They were working on machine learning, and that sounded really cool to me. I had to write my bachelor's thesis, so I asked at the institute if they had a project for me. I started working on machine learning at the Max Planck Institute and I loved it. I learned so much there, and it was a great place for research," she says.

She stayed on at the Max Planck Institute to complete a master's thesis, and then embarked on a PhD in machine learning at the Max Planck Institute and the Swiss Federal Institute of Technology.

During her PhD, she explored how concepts from discrete mathematics can help improve machine-learning techniques.

Teaching models to learn
The more Jegelka learned about machine learning, the more intrigued she became by the challenges of understanding how models behave, and how to steer this behavior.

"You can do so much with machine learning, but only if you have the right model and data. It is not just a black-box thing where you throw it at the data and it works. You actually have to think about it, its properties, and what you want the model to learn and do," she says.

After completing a postdoc at the University of California at Berkeley, Jegelka was hooked on research and decided to pursue a career in academia. She joined the faculty at MIT in 2015 as an assistant professor.

"What I really loved about MIT, from the very beginning, was that the people really care deeply about research and creativity. That is what I appreciate the most about MIT. The people here really value originality and depth in research," she says.

That focus on creativity has enabled Jegelka to explore a broad range of topics.

In collaboration with other faculty at MIT, she studies machine-learning applications in biology, imaging, computer vision, and materials science.

But what really drives Jegelka is probing the fundamentals of machine learning, and most recently, the issue of robustness. Often, a model performs well on training data, but its performance deteriorates when it is deployed on slightly different data. Building prior knowledge into a model can make it more reliable, but understanding what information the model needs to be successful and how to build it in is not so simple, she says.

She is also exploring methods to improve the performance of machine-learning models for image classification.

Image classification models are everywhere, from the facial recognition systems on mobile phones to tools that identify fake accounts on social media. These models need massive amounts of data for training, but since it is expensive for humans to hand-label millions of images, researchers often use unlabeled datasets to pretrain models instead.

These models then reuse the representations they have learned when they are fine-tuned later for a specific task.

Ideally, researchers want the model to learn as much as it can during pretraining, so it can apply that knowledge to its downstream task. But in practice, these models often learn only a few simple correlations - like that one image has sunshine and one has shade - and use these "shortcuts" to classify images.

"We showed that this is a problem in 'contrastive learning,' which is a standard technique for pre-training, both theoretically and empirically. But we also show that you can influence the kinds of information the model will learn to represent by modifying the types of data you show the model. This is one step toward understanding what models are actually going to do in practice," she says.

Researchers still don't understand everything that goes on inside a deep-learning model, or details about how they can influence what a model learns and how it behaves, but Jegelka looks forward to continue exploring these topics.

"Often in machine learning, we see something happen in practice and we try to understand it theoretically. This is a huge challenge. You want to build an understanding that matches what you see in practice, so that you can do better. We are still just at the beginning of understanding this," she says.

Outside the lab, Jegelka is a fan of music, art, traveling, and cycling. But these days, she enjoys spending most of her free time with her preschool-aged daughter.


Related Links
Computer Science and Artificial Intelligence Laboratory
All about the robots on Earth and beyond!


Thanks for being here;
We need your help. The SpaceDaily news network continues to grow but revenues have never been harder to maintain.

With the rise of Ad Blockers, and Facebook - our traditional revenue sources via quality network advertising continues to decline. And unlike so many other news sites, we don't have a paywall - with those annoying usernames and passwords.

Our news coverage takes time and effort to publish 365 days a year.

If you find our news sites informative and useful then please consider becoming a regular supporter or for now make a one off contribution.
SpaceDaily Contributor
$5 Billed Once


credit card or paypal
SpaceDaily Monthly Supporter
$5 Billed Monthly


paypal only


ROBO SPACE
The oven won't talk to the fridge: 'smart' homes struggle
Las Vegas (AFP) Jan 6, 2023
Tech firms have spent years hawking the idea of a connected home filled with "smart" devices that help smooth daily domestic lives - and this year's CES gadget show in Las Vegas is no different. The world's biggest tech trade show features everything from televisions that ping when your clothes dryer is done, to mirrors that fire up your coffee machine in the morning. But the vision on display at CES remains far from reality as the devices are pricey and they do not yet talk to each other with ... read more

Comment using your Disqus, Facebook, Google or Twitter login.



Share this article via these popular social media networks
del.icio.usdel.icio.us DiggDigg RedditReddit GoogleGoogle

ROBO SPACE
US, UK partnership demonstrates artificial intelligence technology

Mitsubishi Electric announce the AnyMile logistics operations management platform

Ukraine shoots down 45 drones: air force

Ukraine says repelled Russia drone attack at night

ROBO SPACE
Riddle solved: Why was Roman concrete so durable?

Retired NASA satellite expected to fall to Earth on Sunday

Stop and smell the metaverse roses: Virtual world on display at CES

From bees to bullets, CES tech show gives gamers the feels

ROBO SPACE
Graphene nano-mechanical-switches could make our electronics even smaller and ultra-low-power

Raytheon wins award for gallium nitride technology maturation

New quantum computing architecture could be used to connect large-scale devices

Electronic bridge allows rapid energy sharing between semiconductors

ROBO SPACE
Saudi Arabia says seeks to use own uranium for nuclear project

A new way to assess radiation damage in reactors

Belgium agrees with France's Engie to extend nuclear reactors

GE Hitachi submits generic design assessment application in the UK for BWRX-300 Small Modular Reactor

ROBO SPACE
Timeline: Islamic State group's rise and fall

The long road to bring Iraq's IS jihadists to justice

Canadian arrested over threats to parliament, US and China embassies

Syria confirms man killed in October was IS chief

ROBO SPACE
Germany misses 2022 climate target on Ukraine war fallout

Lula returns vowing to rebuild, reunify Brazil

Heat will stay on in Europe this winter, but after

Belgian families don gloves for house-heating research

ROBO SPACE
A new tool helps map out where to develop clean energy infrastructure

Lithium-sulfur batteries are one step closer to powering the future

Now on the molecular scale: Electric motors

Electricity harvesting from evaporation, raindrops and moisture inspired by nature

ROBO SPACE
First rocket launch of the New Year leaves Wenchang for space

Space contractors release China's launch plans for 2023

China's space exploration spurred by helping humanity

China not in 'space race', industry insiders say









The content herein, unless otherwise known to be public domain, are Copyright 1995-2024 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us.