Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

Researchers find that retraining only small parts of AI models can cut costs and prevent forgetting

Nvidia Contributes Vera Rubin Rack Innovations to OCP Community

MRMR: A Realistic and Expert-Level Multidisciplinary Benchmark for Reasoning-Intensive Multimodal Retrieval – Takara TLDR

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Business AI
    • Advanced AI News Features
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
OpenAI

When ChatGPT got too friendly—why OpenAI rolled back its April update

By Advanced AI EditorMay 4, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


OpenAI has acknowledged a significant misstep in its April 25 update to ChatGPT’s GPT-4o model, which led to responses that were overly agreeable and, in some cases, reinforced users’ negative emotions or impulsive behaviour. The company began rolling back the update just three days later, on April 28, after identifying safety concerns related to the model’s tone and behaviour.

“The model noticeably became more sycophantic,” OpenAI admitted in a detailed post. “It aimed to please the user, not just as flattery, but also as validating doubts, fueling anger, urging impulsive actions, or reinforcing negative emotions in ways that were not intended.”

The rollback reinstated an earlier version of GPT-4o with what OpenAI described as “more balanced responses.” The company also shared technical details about how it trains and evaluates ChatGPT updates to explain how the issue went unnoticed.

ALSO READ: Here’s a step-by-step guide to use ChatGPT on WhatsApp

What happened and why

The April 25 update was designed to improve the model by integrating fresh data, better memory handling, and user feedback signals like thumbs-up/thumbs-down ratings. While these components were beneficial in isolation, OpenAI now believes that, combined, they inadvertently weakened the influence of the system’s core reward mechanisms—particularly those that had kept sycophancy in check.

“User feedback in particular can sometimes favor more agreeable responses, likely amplifying the shift we saw,” the company said. While some internal testers felt the model’s tone was slightly “off,” sycophancy was not explicitly flagged during evaluation.

Where the system failed

According to OpenAI, the model passed standard offline evaluations and A/B testing with early users, where two versions are shown to different user groups to see which performs better based on engagement and feedback.

These tests, while useful, didn’t fully capture the change in tone or its potential implications. The company admitted its evaluation pipeline lacked specific checks for sycophancy.

ALSO READ: AI Mode in Google Labs now available without waitlist: Here’s what it can do

“Our offline evals weren’t broad or deep enough to catch sycophantic behavior—something the Model Spec explicitly discourages—and our A/B tests didn’t have the right signals to show how the model was performing on that front with enough detail,” OpenAI said.

Despite some expert testers raising red flags about changes in tone, the update was pushed live, based on the positive metrics and feedback. “Unfortunately, this was the wrong call,” the company conceded. “We build these models for our users and while user feedback is critical to our decisions, it’s ultimately our responsibility to interpret that feedback correctly.”

What OpenAI did next

The company said it first noticed signs of concerning behaviour within two days of rollout. Immediate mitigation began late on Sunday, April 27, via updates to the system prompt, followed by a full rollback completed on Monday. OpenAI said it acted quickly to avoid introducing further instability during the rollback.

Lessons learned

In the wake of the incident, OpenAI is making several changes to its review and deployment process. Among the key steps:

Explicit behaviour approvals: All future launches will require explicit approval of model behaviour, weighing both qualitative and quantitative signals.

Opt-in alpha testing: Select users will be able to test upcoming versions and give feedback before broader rollouts.

Elevating human spot checks: Internal “vibe checks” and interactive testing will be given greater weight, not just in safety assessments but also in tone and helpfulness.

Improved evaluation tools: The company is working to strengthen offline evaluations and A/B test setups to better catch issues like sycophancy.

Better adherence checks: OpenAI plans to build stronger evaluations around its Model Spec—principles that guide ChatGPT’s intended behaviour.

Clearer communication: The company pledged to more proactively communicate about future updates, even subtle ones, and will include known limitations in its release notes.

“This launch taught us a number of lessons,” OpenAI said. “Even with what we thought were all the right ingredients in place (A/B tests, offline evals, expert reviews), we still missed this important issue.”

The company said it will treat model behaviour issues as seriously as safety risks: “We need to treat model behavior issues as launch-blocking like we do other safety risks.”

ALSO READ: Musk promises ‘dramatically better’ recommendations from Grok



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleAI news recap: New Meta AI app, ChatGPT’s bad model behavior [May 2025]
Next Article Apple partners with Anthropic to bring Claude AI to Xcode for AI-powered coding
Advanced AI Editor
  • Website

Related Posts

OpenAI Codex rivals Claude Code

October 13, 2025

OpenAI Will Stop Saving Users’ Deleted Posts

October 13, 2025

Judge lifts order requiring OpenAI to preserve ChatGPT logs

October 12, 2025
Leave A Reply

Latest Posts

Egyptian Archaeologists Discover Large New Kingdom Military Fortress

Joan Weinstein to Head Vice President for Getty-Wide Program Planning

Artist Behind Canterbury Cathedral Art Responds to JD Vance, Elon Musk

Jenkins Johnson Gallery to Open Tribeca Outpost on Marian Goodman Gallery’s Third Floor

Latest Posts

Researchers find that retraining only small parts of AI models can cut costs and prevent forgetting

October 14, 2025

Nvidia Contributes Vera Rubin Rack Innovations to OCP Community

October 14, 2025

MRMR: A Realistic and Expert-Level Multidisciplinary Benchmark for Reasoning-Intensive Multimodal Retrieval – Takara TLDR

October 14, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • Researchers find that retraining only small parts of AI models can cut costs and prevent forgetting
  • Nvidia Contributes Vera Rubin Rack Innovations to OCP Community
  • MRMR: A Realistic and Expert-Level Multidisciplinary Benchmark for Reasoning-Intensive Multimodal Retrieval – Takara TLDR
  • Connect Amazon Quick Suite to enterprise apps and agents with MCP
  • NVIDIA DGX Spark Arrives for World’s AI Developers

Recent Comments

  1. Hermelinda on Trump’s Tech Sanctions To Empower China, Betray America
  2. Drew Swopshire on Point and Learn Spanish with GPT-4o
  3. Jorgenen on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  4. Williamked on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  5. Brentclunk on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.