Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

OpenAI delays launch of open model again, cites safety concerns

A new paradigm for AI: How ‘thinking as optimization’ leads to better general-purpose models

OpenAI delays the release of its open model, again

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Industry AI
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
OpenAI

When ChatGPT got too friendly—why OpenAI rolled back its April update

By Advanced AI EditorMay 4, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


OpenAI has acknowledged a significant misstep in its April 25 update to ChatGPT’s GPT-4o model, which led to responses that were overly agreeable and, in some cases, reinforced users’ negative emotions or impulsive behaviour. The company began rolling back the update just three days later, on April 28, after identifying safety concerns related to the model’s tone and behaviour.

“The model noticeably became more sycophantic,” OpenAI admitted in a detailed post. “It aimed to please the user, not just as flattery, but also as validating doubts, fueling anger, urging impulsive actions, or reinforcing negative emotions in ways that were not intended.”

The rollback reinstated an earlier version of GPT-4o with what OpenAI described as “more balanced responses.” The company also shared technical details about how it trains and evaluates ChatGPT updates to explain how the issue went unnoticed.

ALSO READ: Here’s a step-by-step guide to use ChatGPT on WhatsApp

What happened and why

The April 25 update was designed to improve the model by integrating fresh data, better memory handling, and user feedback signals like thumbs-up/thumbs-down ratings. While these components were beneficial in isolation, OpenAI now believes that, combined, they inadvertently weakened the influence of the system’s core reward mechanisms—particularly those that had kept sycophancy in check.

“User feedback in particular can sometimes favor more agreeable responses, likely amplifying the shift we saw,” the company said. While some internal testers felt the model’s tone was slightly “off,” sycophancy was not explicitly flagged during evaluation.

Where the system failed

According to OpenAI, the model passed standard offline evaluations and A/B testing with early users, where two versions are shown to different user groups to see which performs better based on engagement and feedback.

These tests, while useful, didn’t fully capture the change in tone or its potential implications. The company admitted its evaluation pipeline lacked specific checks for sycophancy.

ALSO READ: AI Mode in Google Labs now available without waitlist: Here’s what it can do

“Our offline evals weren’t broad or deep enough to catch sycophantic behavior—something the Model Spec explicitly discourages—and our A/B tests didn’t have the right signals to show how the model was performing on that front with enough detail,” OpenAI said.

Despite some expert testers raising red flags about changes in tone, the update was pushed live, based on the positive metrics and feedback. “Unfortunately, this was the wrong call,” the company conceded. “We build these models for our users and while user feedback is critical to our decisions, it’s ultimately our responsibility to interpret that feedback correctly.”

What OpenAI did next

The company said it first noticed signs of concerning behaviour within two days of rollout. Immediate mitigation began late on Sunday, April 27, via updates to the system prompt, followed by a full rollback completed on Monday. OpenAI said it acted quickly to avoid introducing further instability during the rollback.

Lessons learned

In the wake of the incident, OpenAI is making several changes to its review and deployment process. Among the key steps:

Explicit behaviour approvals: All future launches will require explicit approval of model behaviour, weighing both qualitative and quantitative signals.

Opt-in alpha testing: Select users will be able to test upcoming versions and give feedback before broader rollouts.

Elevating human spot checks: Internal “vibe checks” and interactive testing will be given greater weight, not just in safety assessments but also in tone and helpfulness.

Improved evaluation tools: The company is working to strengthen offline evaluations and A/B test setups to better catch issues like sycophancy.

Better adherence checks: OpenAI plans to build stronger evaluations around its Model Spec—principles that guide ChatGPT’s intended behaviour.

Clearer communication: The company pledged to more proactively communicate about future updates, even subtle ones, and will include known limitations in its release notes.

“This launch taught us a number of lessons,” OpenAI said. “Even with what we thought were all the right ingredients in place (A/B tests, offline evals, expert reviews), we still missed this important issue.”

The company said it will treat model behaviour issues as seriously as safety risks: “We need to treat model behavior issues as launch-blocking like we do other safety risks.”

ALSO READ: Musk promises ‘dramatically better’ recommendations from Grok



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleAI news recap: New Meta AI app, ChatGPT’s bad model behavior [May 2025]
Next Article Apple partners with Anthropic to bring Claude AI to Xcode for AI-powered coding
Advanced AI Editor
  • Website

Related Posts

OpenAI delays launch of open model again, cites safety concerns

July 12, 2025

Windsurf’s CEO Is Headed to Google DeepMind — Not to OpenAI

July 11, 2025

When AI is the interface: why OpenAI, Google and Meta are racing to control the next browser

July 11, 2025
Leave A Reply

Latest Posts

Homeland Security Targets Chicago’s National Museum of Puerto Rican Arts & Culture

1,600-Year-Old Tomb of Mayan City’s Founding King Discovered in Belize

Centre Pompidou Cancels Caribbean Art Show, Raising Controversy

‘Night at the Museum’ Reboot in the Works

Latest Posts

OpenAI delays launch of open model again, cites safety concerns

July 12, 2025

A new paradigm for AI: How ‘thinking as optimization’ leads to better general-purpose models

July 12, 2025

OpenAI delays the release of its open model, again

July 12, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • OpenAI delays launch of open model again, cites safety concerns
  • A new paradigm for AI: How ‘thinking as optimization’ leads to better general-purpose models
  • OpenAI delays the release of its open model, again
  • TU Wien Rendering #10 – Camera models
  • D-Wave Slips as IBM and Google Threaten Its Quantum Edge: What Now? – July 10, 2025

Recent Comments

  1. Compte Binance on Anthropic’s Lawyers Apologize After its Claude AI Hallucinates Legal Citation in Copyright Lawsuit
  2. Index Home on Artists Through The Eyes Of Artists’ At Pallant House Gallery
  3. código binance on Five takeaways from IBM Think 2025
  4. Dang k'y binance on Qwen 2.5 Coder and Qwen 3 Lead in Open Source LLM Over DeepSeek and Meta
  5. "oppna binance-konto on Trump crypto czar Sacks stablecoin bill unlock trillions for Treasury

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.