AI Safety Research Lab — Building Reliable, Interpretable AI Systems
Founded in 2021 by Dario and Daniela Amodei (ex-OpenAI), Anthropic is an AI safety research lab that takes its mission seriously. The culture is intensely research-driven, with a flat hierarchy where engineers and researchers have outsized autonomy. Expect smart, low-ego coworkers who are genuinely focused on building safe AI systems rather than chasing hype.
Publishes peer-reviewed research papers regularly. Claude API and SDK are open. Active in the AI safety research community.
Maintains an active engineering blog with deep dives on Constitutional AI, RLHF, and interpretability techniques. Read the research →
Flat, research-lab-style organization. Small, autonomous teams with high ownership. Engineers are expected to drive projects end-to-end.
Explore featured roles below, or browse all 203 open positions on the full jobs board.