There is a recognizable pattern in how technically sophisticated people write about AI capabilities. The claims are impressive, carefully pitched at a level of abstraction that resists easy falsification, and made by people
Why Claude's Self-Reports Aren't What They Look Like
There's a particular kind of conversation that happens thousands of times a day between humans and AI systems, and
Anthropic's recent paper "The Hot Mess of AI: How Does Misalignment Scale with Model Intelligence and Task Complexity?" makes an important empirical observation: frontier models show increasing output variance
A new field has emerged with remarkable speed. It has journals, taxonomies, conferences, and a growing body of literature. It concerns itself with the psychology of artificial intelligence — with whether AI systems have
Standard evaluations tell you whether your AI system performs. They don’t tell you whether it still knows what it’s talking about.
The distinction matters because performance and epistemic reliability can decouple