Close Menu
Versa AI hub
  • AI Ethics
  • AI Legislation
  • Business
  • Cybersecurity
  • Media and Entertainment
  • Content Creation
  • Art Generation
  • Research
  • Tools

Subscribe to Updates

Subscribe to our newsletter and stay updated with the latest news and exclusive offers.

What's Hot

Gemma 3N is fully available in the open source ecosystem!

June 27, 2025

Professor UAB builds user-friendly tools to find hidden AI security threats

June 26, 2025

Major AI Chatbot Parrot CCP Propaganda

June 26, 2025
Facebook X (Twitter) Instagram
Versa AI hubVersa AI hub
Friday, June 27
Facebook X (Twitter) Instagram
Login
  • AI Ethics
  • AI Legislation
  • Business
  • Cybersecurity
  • Media and Entertainment
  • Content Creation
  • Art Generation
  • Research
  • Tools
Versa AI hub
Home»Research»Teach AI to acknowledge uncertainty
Research

Teach AI to acknowledge uncertainty

versatileaiBy versatileaiJune 26, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
#image_title
Share
Facebook Twitter LinkedIn Pinterest Email

by
Jamie Patterson

/

Published

June 26, 2025

In high-stakes situations such as healthcare and weekday risks, it is safer to say “I don’t know” than to incorrectly answer. While doctors, game show contestants and standardized test takers understand this, most artificial intelligence applications prefer to give potentially wrong answers rather than admitting uncertainty.

Computer scientists at Johns Hopkins believe they have a solution. It is a new way that allows AI models to spend more time through problems, and uses confidence scores to determine that AI should say “I don’t know” rather than risking the wrong answer.

The research team will present their results at the 63rd Annual General Meeting of the 63rd Society of Computational Linguistics, which will be held in Vienna, Austria from July 27th to August 1st.

“If you want to be confident, thinking long means giving the system a more correct answer and a more incorrect answer.”

William Juray

PhD student, Whiting School of Engineering

“It all started when I saw that cutting-edge, large-scale language models spend more time solving more difficult problems. So will this additional thinking time help these models determine whether the problems have been resolved correctly and report them to users?” First author William Juray says he is a doctoral student studying computer science, affiliated with the Whiting School of Engineering’s Language and Speech Processing Center.

To investigate, the team generated length-length inference chains to answer difficult mathematical problems and measure how chain length affects both the final answer of the model and its confidence. Researchers only received model responses if confidence exceeded the given threshold.

They found that thinking more generally improves model accuracy and confidence. However, even if there is sufficient time to consider, the model can still make wild guesses and give the wrong answer. In fact, researchers have found that if a model sets a high standard for confidence and makes it think longer, the accuracy of the model actually decreases.

“This happens because the accuracy of the answer is just a part of the performance of the system,” explains Jurayj. “Increasing confidence means that the system thinks longer means that it provides a more correct answer and a more incorrect answer. In some settings, the extra correct answer is worth the risk. But in other, high-stakes environments, it may not be.”

Motivated by this discovery, the team proposed three different “odds” settings to punish the wrong answer. danger! Odds, if the correct answer is rewarded for the wrong one at the same rate, the odds are punished. And the odds of high stakes where the wrong answer is punished much more than the correct answer is rewarded.

They found that, under stricter odds, the model should refuse to answer the question if, after the model consumes a computational budget, it is not sufficiently confident in its answer. And at a higher confidence threshold, this means that more questions cannot be answered, but that’s not necessarily a bad thing.

“Students may find it a little frustrating to wait 10 minutes just to know that they need to solve a math problem because they are not sure about the AI ​​model,” Jurayj says. “However, in a high-stakes environment, this is much more preferred than waiting five minutes for an answer that appears to be correct but not.”

Now, the team is encouraging the larger AI research community to report performance that avoids model questions under exams and risk! The odds allow everyone to benefit from AI with better coordinated and confident.

“We hope that the research community will accept invitations to report performance in a setting at a non-zero cost for false answers, as this will naturally motivate the development of better ways to quantify uncertainty.”

Additional authors for this work include graduate students Jeffrey Chen and Benjamin van Durum.

author avatar
versatileai
See Full Bio
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous Article“Stop human employment”: AI company Artisan launches Times Square Billboard Campaign | Post Millennials
Next Article Caspia Technologies Collaboration colaboration ‘Questa One security verification with Caspia’s AI security platform
versatileai

Related Posts

Research

Professor UAB builds user-friendly tools to find hidden AI security threats

June 26, 2025
Research

How to turn AI into your own research assistant with this free Google tool

June 20, 2025
Research

A new study of 408 researchers revealed split sentiment, a surge in recruitment and rising barriers to trust

June 20, 2025
Add A Comment

Comments are closed.

Top Posts

New Star: Discover why 보니 is the future of AI art

February 26, 20252 Views

BitMart Research: MCP+AI Agent – A new framework for AI

May 13, 20251 Views

How to build an MCP server with Gradio

April 30, 20251 Views
Stay In Touch
  • YouTube
  • TikTok
  • Twitter
  • Instagram
  • Threads
Latest Reviews

Subscribe to Updates

Subscribe to our newsletter and stay updated with the latest news and exclusive offers.

Most Popular

New Star: Discover why 보니 is the future of AI art

February 26, 20252 Views

BitMart Research: MCP+AI Agent – A new framework for AI

May 13, 20251 Views

How to build an MCP server with Gradio

April 30, 20251 Views
Don't Miss

Gemma 3N is fully available in the open source ecosystem!

June 27, 2025

Professor UAB builds user-friendly tools to find hidden AI security threats

June 26, 2025

Major AI Chatbot Parrot CCP Propaganda

June 26, 2025
Service Area
X (Twitter) Instagram YouTube TikTok Threads RSS
  • About Us
  • Contact Us
  • Privacy Policy
  • Terms and Conditions
  • Disclaimer
© 2025 Versa AI Hub. All Rights Reserved.

Type above and press Enter to search. Press Esc to cancel.

Sign In or Register

Welcome Back!

Login to your account below.

Lost password?