upcarta
  • Sign In
  • Sign Up
  • Explore
  • Search

Anthropic

2 Followers
community-curated profile

We're an AI research company that builds reliable, interpretable, and steerable AI systems. Our first product is Claude, an AI assistant for tasks at any scale.

Overview Content
Featured content
See All
  • Core Views on AI Safety: When, Why, What, and How
    by Anthropic
  • The fact that most individual neurons are uninterpretable presents a serious roadblock to a mechanistic understanding of language models. We demonstrate a method for decomposing groups of neurons into interpretable features with the potential to move
    by Anthropic
  • In this short note, we explore thinking of the traditional idea of "distributed representations" as two distinct phenomena: "composition" and "superposition". We walk through toy examples from Thorpe (1989), discussing them from this lens. twitter.co
    by Anthropic
  • upcarta ©2025
  • Home
  • About
  • Terms
  • Privacy
  • Cookies
  • @upcarta