Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

MIT Vs Caltech for STEM: Which University to choose?

OpenAI gives content owners more control over Sora AI video app

Moveworks Showcases the Power of Its Next-Gen Copilot at Moveworks.global 2024

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Business AI
    • Advanced AI News Features
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
Gary Marcus

OpenAI’s Waterloo? [with corrections] – Marcus on AI

By Advanced AI EditorAugust 14, 2025No Comments8 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


For the last several years, OpenAI has received more press than God.

GPT-3, which most people hardly remember, “wrote” an oped in The Guardian in September 2020 (with human assistance behind the scenes) and ever since media coverage of large language models, generally focusing on OpenAI, has been nonstop. A zillion stories, often fawning, have been written about Sam Altman the boy genius and how ChatGPT would usher in some kind of amazing new era in science and medicine (spoiler alert: it hasn’t, at least not yet) , and how it would radically increase GDP and productivity, ushering in age of abundance (that hasn’t happened yet, either). For a while (mercifully finally over) every journalist and their cousin seem to think that the cleverest thing in the world was to open or close their essay with a quote from ChatGPT. Some people used to it write their wedding vows.

As anyone who reads this knows, I have never been quite so positive. Around the same time as the Guardian Oped, Ernest Davis I warned that GPT0-3 was “a fluent spouter of bullshit.”. In 2023, shortly after ChatGPT was launched, I doubled down, and told 60 Minutes that LLM output was “authoritative bullshit”. (In order to accommodate the delicate ears of network television the expletive was partly bleeped out.( I railed endlessly here and elsewhere about hallucinatioms and warned people to keep their expectations about GPT-4 and GPT-5 modest. From the first day of this newsletter, in May 2022, my theme was that scaling alone would not get us to AGI.

Without a doubt, that left me cast as hater (when in truth I love AI and want it to succeed) and a villain. I was, almost daily, mocked and ridiculed for my criticism of LLMs, including by some of the most powerful people in the tech, from Elon Musk to Yann LeCun (who to his credit, eventually saw for himself the limits of LLMs) to Altman himself, who recently called me a “troll” on X (only to turn tail when I responded in detail).

I endlessly challenged these people to debate, to discuss the facts at hand. None of them accepted. Not once. Nobody ever wanted to talk science.

And they didn’t need to, not with media folks that I won’t name often acting like cheerleaders, and even sometimes personally taking a shots against me. For a long time, the strategy of endless hype and only occasional engagement with science worked like a charm, commercially if not scientifically.

But media coverage is not science, and braggadacio alone cannot yield AGI. What is shocking is that suddenly, almost everywhere all at once, the veil has begun to lift. Over the last week the world woke up to the fact Altman wildly overpromised on GPT-5, with me no longer cast as villain but as hero, which is probably the stuff of Sam’s worst nighmares.

To regular readers of this newsletter, the underwhelming delivery of GPT-5 in its itself should not have been surprising. But what is startling (and frankly satisfying) is how rapidly and radically the narrative has changed.

You can see that narrative flip all over the place. CNN, for example,

Which read in part (highlighting add by the reader who passed the story to me):

Futurism had this to say:

writing in part, name checking me and with a supportive quote from Edinburgh researcher:

Though Marcus’s more realistic view of AI made him a pariah in the excitable AI community, he’s no longer standing alone against scalable AI. Yesterday, University of Edinburgh AI scholar Michael Rovatsos wrote that “it is possible that the release of GPT-5 marks a shift in the evolution of AI which… might usher in the end of creating ever more complicated models whose thought processes are impossible for anyone to understand.”

And my own essay on all this, GPT-5: Overdue, overhyped and underwhelming, went viral, with over 163,000 views.

Meanwhile, to my immense satisfaction, The New Yorker landed firmly on team Marcus, with computer scientist Cal Newport adding in car metaphor I wish I had coined myself:

In the aftermath of GPT-5’s launch, it has become more difficult to take bombastic predictions about A.I. at face value, and the views of critics like Marcus seem increasingly moderate… Post-training improvements don’t seem to be strengthening models as thoroughly as scaling once did. A lot of utility can come from souping up your Camry, but no amount of tweaking will turn it into a Ferrari.”

Even better Newport wound up with this, echoing a passage in my notorious Deep Learning is Hitting a Wall, in which I warned that scaling was not a physical law::

As if Sam’s week couldn’t get worse, columnist Steve Rosenbush at The Wall Street Journal gave me further props, on the topic of neurosymbolic AI, the alternative to LLMs that I long advocated, describing how Amazon was now putting neurosymbolic AI into practice and closing his essay with a pointer to my views and a link to another of this newsletter’s essays.

Gartner, meanwhile, as reported today in the NYT, is forecasting an “trough of disillusionment”

“Gary Marcus was right” memes, none of which reflect well on OpenAI, were everywhere. A reader of this Substack even went to so far to an invent a hilarious Gary Marcus Apology Form.

I have rarely seen a double reversal of fortune (OpenAI down, me up) so swift.

All that said, I was never in reality “standing alone”; I surely took more heat than anyone else but dozens if not hundreds of others spoke out against the scaling-über-alles hypothesis over the years. Indeed, although it did not get the press it deserved, recent AAAI survey showed that the vast majority of academic AI researchers doubted that scaling would get us all the way to AGI. They were right.

At last this is more widely known.

§

I would never count Sam Altman fully out, and increasingly skeptical mainstream media coverage is not the same as, for example a change in investor opinion dramatic enough to cause a bubble to deflate. Time will tell how this all plays out. But already some are wondering whether GPT-5’s disappointments could spark a new AI winter. There is a real sense in which GPT-5 is starting to look Altman’s Waterloo. (Or if it is Moby Dick you prefer, GPT-5 may be his white whale).

Three years of hype add up, and Altman simply could not deliver what he (over)promised. GPT-5 was his primary mission as a leader, and he couldn’t get there convincingly. In no way was it AGI. Some customers even wanted the old models back. Almost nobody felt satisfied.

This raises questions about the technology, about the company’s research prowess, and about Altman himself. Was he bullshitting all this time that he told us that the company knew how to build AGI?

Altman didn’t look a lot better telling CNBC that AGI “is not a super useful term” days after the disappointments of GPT-5, when he had been hyping AGI for years and — literally just a few days earlier—he said claimed that GPT-5 was a “significant step along our path toward AGI”. Two years ago, many treated Altman like an oracle; in the eyes of many, Altman now looks more like a snake oil salesman.

And of course, it is not just about Altman; others have been stalking the same whale; and nobody has yet delivered. Models like Grok-4 and Llama-4 have also underwhelmed.

§

What’s the moral of this story?

Science is not a popularity contest; you can’t bully your way to truth. And you can’t make AI better if you drown out the critics and keep pouring good money after bad. Science simply cannot advance without people sticking to the truth even in the face of opposition.

The good news here is that science is self-correcting; new approaches will rise again from the ashes. And AGI—hopefully safe, trustworthy AGI– will eventually come. Maybe in the next decade.

But between the disappointments of GPT-5 and a new study from METR that shows that LLMs do markedly better on coding benchmarks than in real-word practice, I think it is safe to say that LLMs won’t lead the way. And at last that fact is starting to become widely understood.

As people begin to recognize that truth (the bitter lesson about limits of Sutton’s The Bitter Lesson), scientists will start to charts new paths. One of the new paths may even give rise to what we so desperately need: AI that we can trust.

I can’t wait to see a whole new batch of discoveries, with minds at last wide open.

Gary Marcus appreciates the support of loyal readers of this newsletter, who stood behind him during darker times. Thank you!



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleServer Sales Now 41% Of Revenue
Next Article Top 5 free AI video generators to supercharge your content
Advanced AI Editor
  • Website

Related Posts

Game over for pure LLMs. Even Turing Award Winner Rich Sutton has gotten off the bus.

September 26, 2025

Peak bubble – by Gary Marcus

September 11, 2025

OpenAI’s future, foretold? – by Gary Marcus

September 7, 2025

Comments are closed.

Latest Posts

Former ARTnews Publisher Dies at 97

National Gallery of Art Closes as a Result of Government Shutdown

Almine Rech Closes London Gallery After More Than a Decade

Record Exec and Art Collector Gets Over 4 Years

Latest Posts

MIT Vs Caltech for STEM: Which University to choose?

October 5, 2025

OpenAI gives content owners more control over Sora AI video app

October 5, 2025

Moveworks Showcases the Power of Its Next-Gen Copilot at Moveworks.global 2024

October 5, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • MIT Vs Caltech for STEM: Which University to choose?
  • OpenAI gives content owners more control over Sora AI video app
  • Moveworks Showcases the Power of Its Next-Gen Copilot at Moveworks.global 2024
  • TOUCAN: Synthesizing 1.5M Tool-Agentic Data from Real-World MCP Environments – Takara TLDR
  • DeepSeek AI Models Are Unsafe and Unreliable, Finds NIST-Backed Study

Recent Comments

  1. fuzzypanda7Nalay on Ballet Tech Forms The Future Through Dance
  2. wildpickle3Nalay on Study: AI-Powered Research Prowess Now Outstrips Human Experts, Raising Bioweapon Risks
  3. animsaga.com on European Commission & AI: Guidelines on Prohibited Practices | Paul Hastings LLP
  4. whimsybubblecrab6Nalay on Foundation AI: Cisco launches AI model for integration in security applications
  5. EarnestJoize on Google DeepMind’s Demis Hassabis Wants to Build AI Email Assistant That Can Reply in Your Style: Report

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.