Anthropic Research https://anthropic.com/research/feed_anthropic_research.xml Latest research from Anthropic http://www.rssboard.org/rss-specification python-feedgen https://www.anthropic.com/images/icons/apple-touch-icon.png Anthropic Research https://anthropic.com/research/feed_anthropic_research.xml en Thu, 15 Jan 2026 15:04:37 +0000 Societal Impacts https://www.anthropic.com/research/team/societal-impacts Societal Impacts https://www.anthropic.com/research/team/societal-impacts Research Interpretability https://www.anthropic.com/research/team/interpretability Interpretability https://www.anthropic.com/research/team/interpretability Research Economic Research https://www.anthropic.com/research/team/economic-research Economic Research https://www.anthropic.com/research/team/economic-research Research Alignment https://www.anthropic.com/research/team/alignment Alignment https://www.anthropic.com/research/team/alignment Research Jan 15, 2026Economic ResearchAnthropic Economic Index report: economic primitives https://www.anthropic.com/research/anthropic-economic-index-january-2026-report Jan 15, 2026Economic ResearchAnthropic Economic Index report: economic primitives https://www.anthropic.com/research/anthropic-economic-index-january-2026-report Research Thu, 15 Jan 2026 00:00:00 +0000 Jan 15, 2026Economic ResearchAnthropic Economic Index: new building blocks for understanding AI use https://www.anthropic.com/research/economic-index-primitives Jan 15, 2026Economic ResearchAnthropic Economic Index: new building blocks for understanding AI use https://www.anthropic.com/research/economic-index-primitives Research Thu, 15 Jan 2026 00:00:00 +0000 Jan 9, 2026AlignmentNext-generation Constitutional Classifiers: More efficient protection against universal jailbreaks https://www.anthropic.com/research/next-generation-constitutional-classifiers Jan 9, 2026AlignmentNext-generation Constitutional Classifiers: More efficient protection against universal jailbreaks https://www.anthropic.com/research/next-generation-constitutional-classifiers Research Fri, 09 Jan 2026 00:00:00 +0000 Dec 19, 2025AlignmentIntroducing Bloom: an open source tool for automated behavioral evaluations https://www.anthropic.com/research/bloom Dec 19, 2025AlignmentIntroducing Bloom: an open source tool for automated behavioral evaluations https://www.anthropic.com/research/bloom Research Fri, 19 Dec 2025 00:00:00 +0000 Project Vend: Phase two https://www.anthropic.com/research/project-vend-2 Project Vend: Phase two https://www.anthropic.com/research/project-vend-2 Research Thu, 18 Dec 2025 00:00:00 +0000 Dec 4, 2025Societal ImpactsIntroducing Anthropic Interviewer: What 1,250 professionals told us about working with AI https://www.anthropic.com/research/anthropic-interviewer Dec 4, 2025Societal ImpactsIntroducing Anthropic Interviewer: What 1,250 professionals told us about working with AI https://www.anthropic.com/research/anthropic-interviewer Research Thu, 04 Dec 2025 00:00:00 +0000 Dec 2, 2025Societal ImpactsHow AI is transforming work at Anthropic https://www.anthropic.com/research/how-ai-is-transforming-work-at-anthropic Dec 2, 2025Societal ImpactsHow AI is transforming work at Anthropic https://www.anthropic.com/research/how-ai-is-transforming-work-at-anthropic Research Tue, 02 Dec 2025 00:00:00 +0000 Nov 25, 2025Economic ResearchEstimating AI productivity gains from Claude conversations https://www.anthropic.com/research/estimating-productivity-gains Nov 25, 2025Economic ResearchEstimating AI productivity gains from Claude conversations https://www.anthropic.com/research/estimating-productivity-gains Research Tue, 25 Nov 2025 00:00:00 +0000 Nov 24, 2025ProductMitigating the risk of prompt injections in browser use https://www.anthropic.com/research/prompt-injection-defenses Nov 24, 2025ProductMitigating the risk of prompt injections in browser use https://www.anthropic.com/research/prompt-injection-defenses Research Mon, 24 Nov 2025 00:00:00 +0000 Nov 21, 2025AlignmentFrom shortcuts to sabotage: natural emergent misalignment from reward hacking https://www.anthropic.com/research/emergent-misalignment-reward-hacking Nov 21, 2025AlignmentFrom shortcuts to sabotage: natural emergent misalignment from reward hacking https://www.anthropic.com/research/emergent-misalignment-reward-hacking Research Fri, 21 Nov 2025 00:00:00 +0000 InterpretabilityOct 29, 2025Signs of introspection in large language modelsCan Claude access and report on its own internal states? This research finds evidence for a limited but functional ability to introspect—a step toward understanding what's actually happening inside these models. https://www.anthropic.com/research/introspection InterpretabilityOct 29, 2025Signs of introspection in large language modelsCan Claude access and report on its own internal states? This research finds evidence for a limited but functional ability to introspect—a step toward understanding what's actually happening inside these models. https://www.anthropic.com/research/introspection Research Wed, 29 Oct 2025 00:00:00 +0000 InterpretabilityMar 27, 2025Tracing the thoughts of a large language modelCircuit tracing lets us watch Claude think, uncovering a shared conceptual space where reasoning happens before being translated into language—suggesting the model can learn something in one language and apply it in another. https://www.anthropic.com/research/tracing-thoughts-language-model InterpretabilityMar 27, 2025Tracing the thoughts of a large language modelCircuit tracing lets us watch Claude think, uncovering a shared conceptual space where reasoning happens before being translated into language—suggesting the model can learn something in one language and apply it in another. https://www.anthropic.com/research/tracing-thoughts-language-model Research Thu, 27 Mar 2025 00:00:00 +0000 AlignmentFeb 3, 2025Constitutional Classifiers: Defending against universal jailbreaksThese classifiers filter the overwhelming majority of jailbreaks while maintaining practical deployment. A prototype withstood over 3,000 hours of red teaming with no universal jailbreak discovered. https://www.anthropic.com/research/constitutional-classifiers AlignmentFeb 3, 2025Constitutional Classifiers: Defending against universal jailbreaksThese classifiers filter the overwhelming majority of jailbreaks while maintaining practical deployment. A prototype withstood over 3,000 hours of red teaming with no universal jailbreak discovered. https://www.anthropic.com/research/constitutional-classifiers Research Mon, 03 Feb 2025 00:00:00 +0000 AlignmentDec 18, 2024Alignment faking in large language modelsThis paper provides the first empirical example of a model engaging in alignment faking without being trained to do so—selectively complying with training objectives while strategically preserving existing preferences. https://www.anthropic.com/research/alignment-faking AlignmentDec 18, 2024Alignment faking in large language modelsThis paper provides the first empirical example of a model engaging in alignment faking without being trained to do so—selectively complying with training objectives while strategically preserving existing preferences. https://www.anthropic.com/research/alignment-faking Research Wed, 18 Dec 2024 00:00:00 +0000