Tech giants warn window to monitor AI reasoning is closing, urge action

July 17, 2025 report

The GIST Tech giants warn window to monitor AI reasoning is closing, urge action

Paul Arnold

contributing writer

Gaby Clark

scientific editor

Andrew Zinin

lead editor

Editors' notes

This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

preprint

trusted source

proofread

Tech giants call for better AI safety monitoring
Long serial chains of cognition must pass through the chain of thought. Credit: arXiv (2025). DOI: 10.48550/arxiv.2507.11473

Artificial intelligence is advancing at a dizzying speed. Like many new technologies, it offers significant benefits but also poses safety risks. Recognizing the potential dangers, leading researchers from Google DeepMind, OpenAI, Meta, Anthropic and a coalition of companies and nonprofit groups have come together to call for more to be done to monitor how AI systems "think."

In a joint paper published earlier this week and endorsed by prominent industry figures, including Geoffrey Hinton (widely regarded as the "godfather of AI") and OpenAI co-founder Ilya Sutskever, the scientists argue that a brief window to monitor AI reasoning may soon close.

Improving AI monitoring

They are calling for more monitoring of chains-of-thought (CoTs), a technique that enables AI models to solve complex challenges by breaking them down into smaller steps, much like humans work through complicated tasks, such as a tricky math problem.

CoTs are key features of advanced AI models, including DeepSeek R1 and Language Learning Models (LLMs). However, as AI systems become more advanced, interpreting their decision-making processes will become even more challenging. This is a concern because existing AI oversight methods are imperfect and can miss misbehavior.

In the paper, the scientists have highlighted how CoT monitoring has already proved its worth by detecting examples of AI misbehavior, such as when models act in a misaligned way "by exploiting flaws in their reward functions during training" or "manipulating data to achieve an outcome."

The scientists believe that better monitoring of CoTs could be a valuable way to keep AI agents under control as they become more capable.

"Chain of thought monitoring presents a valuable addition to safety measures for frontier AI, offering a rare glimpse into how AI agents make decisions," said the researchers in their paper. "Yet, there is no guarantee that the current degree of visibility will persist. We encourage the research community and frontier AI developers to make the best use of CoT monitorability and study how it can be preserved."

One key request from the researchers is for AI developers to study what makes CoTs monitorable. In other words, how can we better understand how AI models arrive at their answers? They also want developers to study how CoT monitorability could be included as a safety measure.

The joint paper marks a rare moment of unity between fiercely competitive tech giants, highlighting just how concerned they are about safety. As AI systems become more powerful and integrated into society, ensuring their safety has never been more important or urgent.

Written for you by our author Paul Arnold, edited by Gaby Clark, and fact-checked and reviewed by Andrew Zinin—this article is the result of careful human work. We rely on readers like you to keep independent science journalism alive. If this reporting matters to you, please consider a donation (especially monthly). You'll get an ad-free account as a thank-you.

More information: Tomek Korbak et al, Chain of Thought Monitorability: A New and Fragile Opportunity for AI Safety, arXiv (2025). DOI: 10.48550/arxiv.2507.11473

Journal information: arXiv

© 2025 Science X Network

Citation: Tech giants warn window to monitor AI reasoning is closing, urge action (2025, July 17) retrieved 17 July 2025 from https://techxplore.com/news/2025-07-tech-giants-window-ai-urge.html This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

OpenAI's Altman says 'no plans' to sue China's DeepSeek 0 shares

Feedback to editors