Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

Nvidia Halts H20 AI Chip Production as China Cracks Down on Purchases

A Survey on Large Language Model Benchmarks – Takara TLDR

ET Soonicorns Summit 2025: Can India build its own ChatGPT or DeepSeek?

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Business AI
    • Advanced AI News Features
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
MIT News

MIT Says 95% Of Enterprise AI Fail- Here’s What The 5% Are Doing Right

By Advanced AI EditorAugust 22, 2025No Comments5 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Young Happy Businesswoman Using Laptop Computer in Modern Office with Colleagues. Stylish Beautiful Financial Advisor Working. VFX Hologram Edit Visualizing Stock Exchange Interface, Opened Charts.

Young Happy Businesswoman Using Laptop Computer in Modern Office with Colleagues. Stylish Beautiful Financial Advisor Working. VFX Hologram Edit Visualizing Stock Exchange Interface, Opened Charts.

getty

A new report from MIT has sent shockwaves through the enterprise AI world. According to the State of AI in Business 2025 study, 95% of generative AI pilots deliver zero return on investment. The findings, based on 300 public deployments and more than 150 executive interviews, suggest that billions of dollars have been spent on AI experiments that never scale — and that most organizations are stuck on what MIT researchers call the “GenAI Divide.”

The numbers are stark. Forty percent of organizations say they’ve deployed AI tools, but only 5% have managed to integrate them into workflows at scale. Most projects die in pilot purgatory. Meanwhile, headlines are warning of an “AI bubble,” and investors are shorting AI stocks on the idea that generative AI’s big enterprise moment is already stalling out.

CEO PromptQL, TAnmai Gopal

PromptQL

But not everyone agrees with that reading.

“Confidently wrong is the problem,” says Tanmai Gopal, co-founder and CEO of PromptQL, a unicorn AI company that counts OpenAI, Airbus, Siemens, and NASA as customers. “If the system is not always accurate even the tiniest percent of the time, I need to know when it’s not. Otherwise, my minutes turn into hours; the ROI disappears.”

The Verification Tax

In his blog post, Being “Confidently Wrong” Is Holding AI Back, Gopal describes what he calls the “verification tax.”

“I don’t know when I might get an incorrect response from my AI. So I have to forensically check every response.”

This tax explains much of what MIT labeled as the GenAI Divide. Enterprises eagerly launch pilots, but employees end up spending so much time double-checking outputs that the promised efficiencies never materialize.

It’s not that generative AI lacks raw horsepower — the models can be dazzling. It’s that their confidence is uncalibrated. In regulated or high-stakes industries, one bad answer can outweigh ten good ones. As Gopal puts it: “For serious work, one high-confidence miss costs more credibility than ten successes earn.”

The Learning Gap

MIT’s researchers framed the same issue differently. They found that most enterprise AI tools don’t retain feedback, adapt to workflows, or improve over time. Without those qualities, they stall.

Gopal agrees. “Without high-quality uncertainty information, I don’t know whether a result is wrong because of ambiguity, missing context, stale data, or a model mistake. If I don’t know why it’s wrong, I’m not invested in making it successful.”

That insight matters because it reframes the entire conversation. If AI isn’t failing due to lack of capability, but because it hasn’t been designed to communicate its limits and learn from corrections, then the fix is less about building bigger models — and more about building humbler ones.

How PromptQL Solves It

PromptQL has built its entire platform around solving this exact problem — what Gopal calls the difference between being “confidently wrong” and “tentatively right.”

Instead of presenting outputs as gospel, PromptQL calibrates confidence at the response level:

Quantifies uncertainty. Every answer comes with a confidence score. If the system is unsure, it abstains — effectively saying “I don’t know.
Surfaces context gaps. Rather than hiding uncertainty, the system flags why an answer may be unreliable: missing data, ambiguity, or lack of context.
Builds an accuracy flywheel. Each abstention or correction becomes training fuel. PromptQL captures those signals, letting the system improve continuously — closing the “learning gap” MIT identified as the number one cause of pilot failure.
Integrates into workflows. Instead of sitting in a chatbox, PromptQL embeds directly into enterprise processes like contracts, engineering, or procurement, so uncertainty flags and corrections appear exactly where the work is happening.

“The starting point of this loop is if an AI system could tell the user when it’s not certain about its accuracy in a concrete and native way,” Gopal writes. That loop — abstain, get corrected, learn — is what he calls the accuracy flywheel. “We don’t need perfection; we need a loop that tightens.”

Tentatively Right Beats Confidently Wrong

This humility-first approach has led to adoption in some of the most skeptical corners of the enterprise market. While 95% of pilots stall, PromptQL is closing seven- and eight-figure contracts with Fortune 500s, government agencies, and regulated industries — the exact places MIT says AI has struggled to gain traction.

The company is living proof that enterprise AI is not failing. The wrong kind of enterprise AI is.

As Gopal puts it: “No amount of solving any other problem — integration, data readiness, organizational readiness — will change the fact that AI’s tendency to be confidently wrong keeps it out of real-world use cases.”

A Different Conclusion

The takeaway, then, is not that AI is doomed to fail. It’s that enterprises must demand a different kind of AI: one that is transparent about its uncertainty, tightly integrated into workflows, and capable of improving with every interaction.

The MIT report is right to highlight the GenAI Divide. But if we only focus on the 95% that failed, we miss the 5% that are actually scaling — and why.

The companies that build and adopt AI that admits when it doesn’t know are quietly rewriting the story. PromptQL is one of them.

And if their traction holds, the conclusion isn’t that enterprise AI is a bubble. It’s that a small handful of companies have already figured out how to burst it.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleDon’t sleep on Cohere: Command A Reasoning, its first reasoning model, is built for enterprise customer service and more
Next Article NASA and IBM Launch AI that Predicts Solar Activity
Advanced AI Editor
  • Website

Related Posts

MIT develops compact laser ‘comb’ to detect chemicals with extreme precision | Technology News

August 21, 2025

AI investments failing? 95 per cent of firms see no returns, says MIT | Technology News

August 21, 2025

MIT study: 95% GenAI projects fail to show returns

August 21, 2025

Comments are closed.

Latest Posts

White House Targets Specific Artworks at Smithsonian Museums

French Art Historian Trying to Block Bayeux Tapestry’s Move to London

Czech Man Sues Christie’s For Information on Nazi-Looted Artworks

Tanya Bonakdar Gallery to Close Los Angeles Space

Latest Posts

Nvidia Halts H20 AI Chip Production as China Cracks Down on Purchases

August 22, 2025

A Survey on Large Language Model Benchmarks – Takara TLDR

August 22, 2025

ET Soonicorns Summit 2025: Can India build its own ChatGPT or DeepSeek?

August 22, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • Nvidia Halts H20 AI Chip Production as China Cracks Down on Purchases
  • A Survey on Large Language Model Benchmarks – Takara TLDR
  • ET Soonicorns Summit 2025: Can India build its own ChatGPT or DeepSeek?
  • China issues new warning for Nvidia A…
  • AI FOMO Could Be Fueling a Risky Bubble in AI’s Hottest Companies

Recent Comments

  1. Grovervot on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  2. nadeem majdalany pandoras box austria on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  3. Grovervot on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  4. whiteboardcrm.com lừa đảo công an truy quét cấm người chơi tham gia on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  5. SteveDiale on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.