24/7 Space News
ROBO SPACE
Efficient technique improves machine-learning models' reliability
Researchers from MIT and the MIT-IBM Watson AI Lab have developed a new technique that can enable a machine-learning model to quantify how confident it is in its predictions, but does not require vast troves of new data and is much less computationally intensive than other techniques.
Efficient technique improves machine-learning models' reliability
by Adam Zewe for MIT News
Boston MA (SPX) Feb 14, 2023

Powerful machine-learning models are being used to help people tackle tough problems such as identifying disease in medical images or detecting road obstacles for autonomous vehicles. But machine-learning models can make mistakes, so in high-stakes settings it's critical that humans know when to trust a model's predictions.

Uncertainty quantification is one tool that improves a model's reliability; the model produces a score along with the prediction that expresses a confidence level that the prediction is correct. While uncertainty quantification can be useful, existing methods typically require retraining the entire model to give it that ability. Training involves showing a model millions of examples so it can learn a task. Retraining then requires millions of new data inputs, which can be expensive and difficult to obtain, and also uses huge amounts of computing resources.

Researchers at MIT and the MIT-IBM Watson AI Lab have now developed a technique that enables a model to perform more effective uncertainty quantification, while using far fewer computing resources than other methods, and no additional data. Their technique, which does not require a user to retrain or modify a model, is flexible enough for many applications.

The technique involves creating a simpler companion model that assists the original machine-learning model in estimating uncertainty. This smaller model is designed to identify different types of uncertainty, which can help researchers drill down on the root cause of inaccurate predictions.

"Uncertainty quantification is essential for both developers and users of machine-learning models. Developers can utilize uncertainty measurements to help develop more robust models, while for users, it can add another layer of trust and reliability when deploying models in the real world. Our work leads to a more flexible and practical solution for uncertainty quantification," says Maohao Shen, an electrical engineering and computer science graduate student and lead author of a paper on this technique.

Shen wrote the paper with Yuheng Bu, a former postdoc in the Research Laboratory of Electronics (RLE) who is now an assistant professor at the University of Florida; Prasanna Sattigeri, Soumya Ghosh, and Subhro Das, research staff members at the MIT-IBM Watson AI Lab; and senior author Gregory Wornell, the Sumitomo Professor in Engineering who leads the Signals, Information, and Algorithms Laboratory RLE and is a member of the MIT-IBM Watson AI Lab. The research will be presented at the AAAI Conference on Artificial Intelligence.

Quantifying uncertainty
In uncertainty quantification, a machine-learning model generates a numerical score with each output to reflect its confidence in that prediction's accuracy. Incorporating uncertainty quantification by building a new model from scratch or retraining an existing model typically requires a large amount of data and expensive computation, which is often impractical. What's more, existing methods sometimes have the unintended consequence of degrading the quality of the model's predictions.

The MIT and MIT-IBM Watson AI Lab researchers have thus zeroed in on the following problem: Given a pretrained model, how can they enable it to perform effective uncertainty quantification?

They solve this by creating a smaller and simpler model, known as a metamodel, that attaches to the larger, pretrained model and uses the features that larger model has already learned to help it make uncertainty quantification assessments.

"The metamodel can be applied to any pretrained model. It is better to have access to the internals of the model, because we can get much more information about the base model, but it will also work if you just have a final output. It can still predict a confidence score," Sattigeri says.

They design the metamodel to produce the uncertainty quantification output using a technique that includes both types of uncertainty: data uncertainty and model uncertainty. Data uncertainty is caused by corrupted data or inaccurate labels and can only be reduced by fixing the dataset or gathering new data. In model uncertainty, the model is not sure how to explain the newly observed data and might make incorrect predictions, most likely because it hasn't seen enough similar training examples. This issue is an especially challenging but common problem when models are deployed. In real-world settings, they often encounter data that are different from the training dataset.

"Has the reliability of your decisions changed when you use the model in a new setting? You want some way to have confidence in whether it is working in this new regime or whether you need to collect training data for this particular new setting," Wornell says.

Validating the quantification
Once a model produces an uncertainty quantification score, the user still needs some assurance that the score itself is accurate. Researchers often validate accuracy by creating a smaller dataset, held out from the original training data, and then testing the model on the held-out data. However, this technique does not work well in measuring uncertainty quantification because the model can achieve good prediction accuracy while still being over-confident, Shen says.

They created a new validation technique by adding noise to the data in the validation set - this noisy data is more like out-of-distribution data that can cause model uncertainty. The researchers use this noisy dataset to evaluate uncertainty quantifications.

They tested their approach by seeing how well a meta-model could capture different types of uncertainty for various downstream tasks, including out-of-distribution detection and misclassification detection. Their method not only outperformed all the baselines in each downstream task but also required less training time to achieve those results.

This technique could help researchers enable more machine-learning models to effectively perform uncertainty quantification, ultimately aiding users in making better decisions about when to trust predictions.

Moving forward, the researchers want to adapt their technique for newer classes of models, such as large language models that have a different structure than a traditional neural network, Shen says.

The work was funded, in part, by the MIT-IBM Watson AI Lab and the U.S. National Science Foundation.

Research Report:"Post-hoc Uncertainty Learning using a Dirichlet Meta-Mode"

Related Links
Signals, Information, and Algorithms Laboratory
All about the robots on Earth and beyond!

Subscribe Free To Our Daily Newsletters
Tweet

RELATED CONTENT
The following news reports may link to other Space Media Network websites.
ROBO SPACE
Killer AI? Dutch summit to focus on military use
The Hague (AFP) Feb 9, 2023
While chatbots have caught the world's imagination, should we be more worried about "slaughterbots"? The first international conference on responsible military uses of Artificial Intelligence (AI) is being held in the Netherlands next week. The United States and China are among around 50 countries that will attend, with hopes of producing a declaration at the end of the meeting in The Hague on February 15 and 16. Russia has not been invited over the invasion of Ukraine. "We truly see thi ... read more

ROBO SPACE
Roscosmos is delaying the launch of the replacement Soyuz for crew return

NASA, partners clear Axiom's second private astronaut mission crew

Spacecraft controllers aim for the heights

NASA launches new Framework for Procurement Ideas, Solutions

ROBO SPACE
Japan's H3 rocket fails to leave the launch pad

The roar and crackle of Artemis 1

Russian Progress cargo craft docks at space station suffers loss of coolant

NASA conducts first 2023 test of redesigned SLS rocket engine

ROBO SPACE
Study quantifies global impact of electricity in dust storms on Mars

Complex subsurface of Mars imaged by Chinese rover Zhurong

Perseverance shows off collection of Mars samples

Cloud gazing while we get ready to drill: Sols 3739-3741

ROBO SPACE
China's space station experiments pave way for new space technology

Large number of launches planned

Chinese astronauts complete first walk outside Tiangong space station

Shenzhou XV astronauts take their first spacewalk

ROBO SPACE
SpaceX launches 55 Starlink satellites early Sunday morning

FCC greenlights Amazon's Project Kuiper to deploy 3,236 satellites in LEO

Women and girls in science: the team helping to take us to Mars

AccelerComm, TTP set to soar with 5G NTN LEO Space Cell

ROBO SPACE
'Magic' solvent creates stronger thin films

Space Station research announcement for advanced materials and manufacturing open now

Astroscale wins Dstl funding for exploration of future Space-Based Space Domain Awareness missions

Sidius Space reaches an agreement with a Dutch organization to Deploy Lasercom Mission

ROBO SPACE
New models shed light on life's origin

Researchers focus AI on finding exoplanets

A nearby potentially habitable Earth-mass exoplanet

Two nearby exoplanets might be habitable

ROBO SPACE
JUICE's final take-off before lift-off

A new ring system discovered in our Solar System

SwRI models explain canyons on Pluto moon

NASA's Juno Team assessing camera after 48th flyby of Jupiter

Subscribe Free To Our Daily Newsletters




The content herein, unless otherwise known to be public domain, are Copyright 1995-2024 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us.