Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

Dave Plummer: Programming, Autism, and Old-School Microsoft Stories | Lex Fridman Podcast #479

Cohere Health Funding: $90 Million Series C Fuels AI Expansion

How DeepSeek AI will reshape Nigeria’s business landscape

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Business AI
    • Advanced AI News Features
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
Google DeepMind

AI Giants Push for Transparency on Models’ Inner Monologue

By Advanced AI EditorJuly 17, 2025No Comments5 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Artificial Intelligence & Machine Learning
,
Next-Generation Technologies & Secure Development

Experts Aim to Probe How AI Models Reason, and Why It Matters

Rashmi Ramesh (rashmiramesh_) •
July 16, 2025    

AI Giants Push for Transparency on Models' Inner Monologue
AI researchers from OpenAI, Google DeepMind and Anthropic and others have urged deeper study into chain-of-thought monitoring. (Image: Shutterstock)

As an artificial intelligence model narrates what it claims is its own thoughts, it’s tempting to believe that we finally know what it is thinking. But researchers from AI enterprise giants that make these models caution that this glimpse into machine reasoning could be fleeting, and that much more needs to be understood before labelling it as true transparency.

See Also: OnDemand Webinar | Trends, Threats and Expert Takeaways: 2025 Global IR Report Insights

A coalition of scientists from OpenAI, Google DeepMind and Anthropic called for a systematic investigation into monitoring the so-called chains-of-thought, or CoTs, that underpin modern AI reasoning models. These models include OpenAI’s o3 and DeepSeek’s R1, which are designed to tackle complex tasks by breaking them down step by step, similar to how a human might jot notes to solve a problem.

The researchers in their paper described CoT monitoring as an additional safety measure for frontier AI, offering an unusual window into how AI agents make decisions. They also warned that the current level of visibility into these processes could diminish over time. The authors encouraged the research community and AI developers to make use of “CoT monitorability” while it exists and to study how it might be preserved as models advance.

Chains-of-thought have become a central feature of reasoning models, which are increasingly integral to the ambitions of companies building AI agents. By revealing the intermediate steps a model uses to produce an answer, CoT monitoring offers a potential means to assess whether a model is reasoning safely or drifting into unintended behavior. But it’s unclear what makes this transparency robust and what might undermine it, researchers said (see: A Peek Into How AI ‘Thinks’ – and Why It Hallucinates).

The paper asked developers to explore what factors influence CoT monitorability, including whether interventions, architecture changes or optimization techniques could reduce transparency or reliability. The authors warned that CoT monitoring might be fragile and advised against changes that could degrade the clarity of a model’s reasoning process.

Among those endorsing the call to action were OpenAI Chief Research Officer Mark Chen, Safe Superintelligence CEO Ilya Sutskever, Nobel laureate Geoffrey Hinton, Google DeepMind co-founder Shane Legg, xAI safety adviser Dan Hendrycks and Thinking Machines co-founder John Schulman. The first authors include contributors from the U.K. AI Safety Institute and Apollo Research, with additional signatures from researchers affiliated with Amazon, Meta and UC Berkeley.

The position paper comes at a time when leading labs are racing to outdo each other in building more capable AI agents, or models that can plan, reason and act autonomously across tasks. In September, OpenAI previewed its first AI reasoning model, o1. In the following months, Google, DeepMind, xAI and Anthropic introduced competitors that demonstrated similar or superior performance on several benchmarks.

But the rapid improvements in performance have not necessarily translated into a deeper understanding of how these systems arrive at their conclusions, the paper authors said.

Anthropic has particularly invested heavily in interpretability research. Earlier this year, CEO Dario Amodei announced a commitment to crack open the black box of AI models within the next few years and said the company would expand funding and research into interpretability. He also called on OpenAI and Google DeepMind to increase their efforts in the same area.

Earlier findings from Anthropic suggest that CoTs may not always be a fully reliable reflection of how models reach their answers. The position paper says that chains-of-thought could be influenced by prompting methods or external factors, potentially creating a misleading impression of transparency. OpenAI researchers have said that with further study, CoT monitoring could eventually serve as a practical way to track alignment and safety (see: AI Hijacked: New Jailbreak Exploits Chain-of-Thought).

There is fierce competition in the industry at the moment to recruit researchers capable of advancing AI reasoning models. Meta has reportedly been offering compensation packages in the million-dollar range to lure talent away from Anthropic, OpenAI and Google DeepMind. Many of the most sought-after researchers are those specializing in the systems that the paper seeks to make more transparent.

The stakes are high for companies signing the paper. As AI agents become more capable, the pressure to show that they behave predictably and safely will likely intensify, and without clear methods to monitor their reasoning, assurances about safety could remain just empty words.

The authors described their publication as an effort to raise awareness and draw more attention to research on CoT monitoring. They wrote that the purpose of the paper is to signal-boost the topic and encourage the field to prioritize it, while acknowledging that further work is essential.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleStanford HAI Welcomes Six Distinguished Scholars as Senior Fellows
Next Article Accenture scales video analysis with Amazon Nova and Amazon Bedrock Agents
Advanced AI Editor
  • Website

Related Posts

British lawmakers accuse Google DeepMind of ‘breach of trust’ over delayed Gemini 2.5 Pro safety report

August 30, 2025

Nano Banana: Google DeepMind upgrades Gemini app with advanced AI image editing model

August 29, 2025

Summit With OpenAI, Google DeepMind Reaches Bleak Agreement

August 29, 2025

Comments are closed.

Latest Posts

Woodmere Art Museum Sues Trump Administration Over Canceled IMLS Grant

Barbara Gladstone’s Chelsea Townhouse in NYC Sells for $13.1 M.

Trump Meets with Smithsonian Leader Amid Threats of Content Review

Australian School Faces Pushback over AI Art Course—and More Art News

Latest Posts

Dave Plummer: Programming, Autism, and Old-School Microsoft Stories | Lex Fridman Podcast #479

August 31, 2025

Cohere Health Funding: $90 Million Series C Fuels AI Expansion

August 31, 2025

How DeepSeek AI will reshape Nigeria’s business landscape

August 31, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • Dave Plummer: Programming, Autism, and Old-School Microsoft Stories | Lex Fridman Podcast #479
  • Cohere Health Funding: $90 Million Series C Fuels AI Expansion
  • How DeepSeek AI will reshape Nigeria’s business landscape
  • OpenAI and Anthropic evaluated each others’ models – which ones came out on top
  • Strategic Sector Positioning and Operational ROI for Long-Term Retail Dominance

Recent Comments

  1. Jorgevopsy on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  2. Thomasnuh on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  3. مهندسی صنایع یا مکانیک on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  4. Thomasnuh on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  5. stroycata1og-395 on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.