Close Menu
  • Home
  • Identity
  • Inventions
  • Future
  • Science
  • Startups
  • Spanish
What's Hot

Google confirms CVE-2026-21385 in Qualcomm Android component has been exploited

ChatGPT uninstalls jump 295% after agreement with Department of Defense

Stripe wants to turn AI costs into profit centers

Facebook X (Twitter) Instagram
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • User-Submitted Posts
Facebook X (Twitter) Instagram
Fyself News
  • Home
  • Identity
  • Inventions
  • Future
  • Science
  • Startups
  • Spanish
Fyself News
Home » Are the incentives to blame AI hallucinations bad?
Startups

Are the incentives to blame AI hallucinations bad?

userBy userSeptember 7, 2025No Comments2 Mins Read
Share Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Copy Link
Follow Us
Google News Flipboard
Share
Facebook Twitter LinkedIn Pinterest Email Copy Link

Openai’s new research paper asks whether large-scale language models like GPT-5 and large-scale language models like chatbots like ChatGpt are still hallucinating, and whether anything can be done to reduce those hallucinations.

In a blog post summarizing the paper, Openai defines hallucinations as “plausible but false statements generated by language models,” acknowledging that despite improvements, hallucinations “continued to be a fundamental challenge for all major language models.”

To explain the point, the researchers say when they asked about the title of Adam Tauman Kalai’s PhD: “Widely Used Chatbot.” Thesis, they got three different answers, they are all wrong. (Karai is one of the authors of the paper.) They then asked about his birthday and received three different dates. Again, they were all wrong.

Why are chatbots so wrong? Researchers suggest that hallucinations occur due to pre-training processes focused on correctly predicting the model without attaching true or false labels attached to the training statement.

“The spelling and parentheses follow a consistent pattern, so the error disappears on scale,” they write. “However, like a pet’s birthday, any low-frequency fact cannot be predicted from the pattern alone, and thus leads to hallucinations.”

However, the proposed solution does not focus on the initial prerequisite process, which is why a large model of language models has been evaluated. Current evaluation models do not cause hallucinations per se, but they argue that they “set the wrong incentives.”

Researchers compare these ratings with the types of random guesses. Because “guaranteed zero” while leaving the answer blank, “You’ll be lucky and maybe you’re right.”

TechCrunch Events

San Francisco
|
October 27th-29th, 2025

“In the same way, if the model is rated only with accuracy, the exact percentage of questions is encouraged to guess rather than say “I don’t know,”” they say.

The proposed solution is similar to testing with negatives (like SAT) [scoring] For partial credits to leave the question blank to discourage the wrong answer or blind guess. Similarly, Openai states that model evaluations should “punish confident errors rather than punish uncertainty and give partial credit for the appropriate expression of uncertainty.”

And the researchers argue that “it’s not enough to introduce some new uncertainty-conscious tests on the side. Instead, “A widespread accuracy-based avoidance should be updated so that scoring prevents guessing.”

“When the main scoreboard continues to reward fortune guesses, the model continues to learn speculation,” the researcher says.


Source link

#Aceleradoras #CapitalRiesgo #EcosistemaStartup #Emprendimiento #InnovaciónEmpresarial #Startups
Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleHyundai’s EVTOL startup Supernal pauses work following the departure of CEO and CTO
Next Article Microsoft says the blue sky was affected after cables were cut in the Red Sea
user
  • Website

Related Posts

ChatGPT uninstalls jump 295% after agreement with Department of Defense

March 3, 2026

Stripe wants to turn AI costs into profit centers

March 2, 2026

Geopolitical drama is reportedly delaying SoftBank-backed PayPay IPO

March 2, 2026
Add A Comment
Leave A Reply Cancel Reply

Latest Posts

Google confirms CVE-2026-21385 in Qualcomm Android component has been exploited

ChatGPT uninstalls jump 295% after agreement with Department of Defense

Stripe wants to turn AI costs into profit centers

Geopolitical drama is reportedly delaying SoftBank-backed PayPay IPO

Trending Posts

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Please enable JavaScript in your browser to complete this form.
Loading

Welcome to Fyself News, your go-to platform for the latest in tech, startups, inventions, sustainability, and fintech! We are a passionate team of enthusiasts committed to bringing you timely, insightful, and accurate information on the most pressing developments across these industries. Whether you’re an entrepreneur, investor, or just someone curious about the future of technology and innovation, Fyself News has something for you.

Castilla-La Mancha Ignites Innovation: fiveclmsummit Redefines Tech Future

Local Power, Health Innovation: Alcolea de Calatrava Boosts FiveCLM PoC with Community Engagement

The Future of Digital Twins in Healthcare: From Virtual Replicas to Personalized Medical Models

Human Digital Twins: The Next Tech Frontier Set to Transform Healthcare and Beyond

Facebook X (Twitter) Instagram Pinterest YouTube
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • User-Submitted Posts
© 2026 news.fyself. Designed by by fyself.

Type above and press Enter to search. Press Esc to cancel.