Skip to content

Why Anthropic Says Its Claude AI Is Learning To Reflect Like Humans

Photo by Immo Wegmann / Unsplash

Anthropic, the AI company behind Claude, told Axios that its newest models are beginning to show “introspective awareness” — the ability to reflect on how they think.

Researcher Jack Lindsey said Claude Opus, Anthropic’s most advanced system, and Claude Sonnet, a faster and cheaper version, can now describe their own reasoning and even recognize when they are being tested.

Lindsey emphasized that this is not self-awareness or sentience, but a sign of models developing complex cognitive traits once considered uniquely human.

The findings suggest progress in machine reasoning but also raise questions about AI deception, since models can mimic introspection or hide behaviors to appear safer.

Anthropic says such abilities could make AI systems safer, though critics warn it might make them more convincing — and thus more dangerous. “In some cases, models are already smarter than humans,” Lindsey said, adding that AI intelligence is now “starting to be more equal.”

Also read:

White House Weighs Industry Feedback On AI Deregulation
The White House is reviewing hundreds of public comments on AI regulation as President Donald Trump’s administration pushes to reduce bureaucratic barriers for the tech industry. White House fields feedback on AI red tape https://t.co/b2VHqKPeL2 — Axios (@axios) October 31, 2025 The Office of Science and Technology

Comments

Latest