Close Menu
Versa AI hub
  • AI Ethics
  • AI Legislation
  • Business
  • Cybersecurity
  • Media and Entertainment
  • Content Creation
  • Art Generation
  • Research
  • Tools
  • Resources

Subscribe to Updates

Subscribe to our newsletter and stay updated with the latest news and exclusive offers.

What's Hot

How we achieved cutting-edge technology

December 8, 2025

New image verification feature added to Gemini app

December 7, 2025

Aluminum OS is the AI-powered successor to ChromeOS

December 7, 2025
Facebook X (Twitter) Instagram
Versa AI hubVersa AI hub
Monday, December 8
Facebook X (Twitter) Instagram
Login
  • AI Ethics
  • AI Legislation
  • Business
  • Cybersecurity
  • Media and Entertainment
  • Content Creation
  • Art Generation
  • Research
  • Tools
  • Resources
Versa AI hub
Home»Research»Teach AI to acknowledge uncertainty
Research

Teach AI to acknowledge uncertainty

versatileaiBy versatileaiJune 26, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
#image_title
Share
Facebook Twitter LinkedIn Pinterest Email

by
Jamie Patterson

/

Published

June 26, 2025

In high-stakes situations such as healthcare and weekday risks, it is safer to say “I don’t know” than to incorrectly answer. While doctors, game show contestants and standardized test takers understand this, most artificial intelligence applications prefer to give potentially wrong answers rather than admitting uncertainty.

Computer scientists at Johns Hopkins believe they have a solution. It is a new way that allows AI models to spend more time through problems, and uses confidence scores to determine that AI should say “I don’t know” rather than risking the wrong answer.

The research team will present their results at the 63rd Annual General Meeting of the 63rd Society of Computational Linguistics, which will be held in Vienna, Austria from July 27th to August 1st.

“If you want to be confident, thinking long means giving the system a more correct answer and a more incorrect answer.”

William Juray

PhD student, Whiting School of Engineering

“It all started when I saw that cutting-edge, large-scale language models spend more time solving more difficult problems. So will this additional thinking time help these models determine whether the problems have been resolved correctly and report them to users?” First author William Juray says he is a doctoral student studying computer science, affiliated with the Whiting School of Engineering’s Language and Speech Processing Center.

To investigate, the team generated length-length inference chains to answer difficult mathematical problems and measure how chain length affects both the final answer of the model and its confidence. Researchers only received model responses if confidence exceeded the given threshold.

They found that thinking more generally improves model accuracy and confidence. However, even if there is sufficient time to consider, the model can still make wild guesses and give the wrong answer. In fact, researchers have found that if a model sets a high standard for confidence and makes it think longer, the accuracy of the model actually decreases.

“This happens because the accuracy of the answer is just a part of the performance of the system,” explains Jurayj. “Increasing confidence means that the system thinks longer means that it provides a more correct answer and a more incorrect answer. In some settings, the extra correct answer is worth the risk. But in other, high-stakes environments, it may not be.”

Motivated by this discovery, the team proposed three different “odds” settings to punish the wrong answer. danger! Odds, if the correct answer is rewarded for the wrong one at the same rate, the odds are punished. And the odds of high stakes where the wrong answer is punished much more than the correct answer is rewarded.

They found that, under stricter odds, the model should refuse to answer the question if, after the model consumes a computational budget, it is not sufficiently confident in its answer. And at a higher confidence threshold, this means that more questions cannot be answered, but that’s not necessarily a bad thing.

“Students may find it a little frustrating to wait 10 minutes just to know that they need to solve a math problem because they are not sure about the AI ​​model,” Jurayj says. “However, in a high-stakes environment, this is much more preferred than waiting five minutes for an answer that appears to be correct but not.”

Now, the team is encouraging the larger AI research community to report performance that avoids model questions under exams and risk! The odds allow everyone to benefit from AI with better coordinated and confident.

“We hope that the research community will accept invitations to report performance in a setting at a non-zero cost for false answers, as this will naturally motivate the development of better ways to quantify uncertainty.”

Additional authors for this work include graduate students Jeffrey Chen and Benjamin van Durum.

author avatar
versatileai
See Full Bio
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleTop 3 AI Cryptocoins to watch in 2025 – Research Analysis – Prediction
Next Article Street Signs of the Street Signs DataVault AI (DVLT) and long-standing interviews, television commercials and iconic Billboard exposure featuring a 12-month national media series – filming will soon begin in collaboration with fintech.tv – Akron Beacon Journal
versatileai

Related Posts

Research

New AI research clarifies the origins of Papua New Guineans

July 22, 2025
Research

AI helps prevent medical errors in real clinics

July 22, 2025
Research

No one is surprised, and a new study says that AI overview causes a significant drop in search clicks

July 22, 2025
Add A Comment

Comments are closed.

Top Posts

UK and Germany plan to commercialize quantum supercomputing

December 5, 20255 Views

Tencent launches Hunyuan 3D AI asset generation engine

December 3, 20255 Views

Aluminum OS is the AI-powered successor to ChromeOS

December 7, 20254 Views
Stay In Touch
  • YouTube
  • TikTok
  • Twitter
  • Instagram
  • Threads
Latest Reviews

Subscribe to Updates

Subscribe to our newsletter and stay updated with the latest news and exclusive offers.

Most Popular

UK and Germany plan to commercialize quantum supercomputing

December 5, 20255 Views

Tencent launches Hunyuan 3D AI asset generation engine

December 3, 20255 Views

Aluminum OS is the AI-powered successor to ChromeOS

December 7, 20254 Views
Don't Miss

How we achieved cutting-edge technology

December 8, 2025

New image verification feature added to Gemini app

December 7, 2025

Aluminum OS is the AI-powered successor to ChromeOS

December 7, 2025
Service Area
X (Twitter) Instagram YouTube TikTok Threads RSS
  • About Us
  • Contact Us
  • Privacy Policy
  • Terms and Conditions
  • Disclaimer
© 2025 Versa AI Hub. All Rights Reserved.

Type above and press Enter to search. Press Esc to cancel.

Sign In or Register

Welcome Back!

Login to your account below.

Lost password?