AI Interpretability Research Lab — Reverse-Engineering Neural Networks to Make AI Safe and Controllable
Frontier AI safety research with world-class pedigree. Choose Goodfire if you want to work on mechanistic interpretability alongside researchers from the top AI labs — but go in knowing it's early-stage with all that entails.
Goodfire is a San Francisco-based AI research lab and public benefit corporation focused on mechanistic interpretability — understanding how neural networks actually work at a fundamental level. Founded in 2023 by Eric Ho, Dan Balsam, and Tom McGrath (formerly of DeepMind's interpretability team), the company has raised $209M including a $150M Series B at a $1.25B valuation in February 2026.
The team of ~51 includes researchers from OpenAI, DeepMind, Harvard, and Stanford. Anthropic is both an investor and a validation of Goodfire's mission — making AI systems understandable and controllable. As a public benefit corporation, the mission isn't just marketing — it's legally embedded in the company structure.
Goodfire has no Glassdoor reviews yet. Ratings above are estimated based on company stage, funding, and comparable AI research labs. We'll update this section when real employee reviews become available.
Researchers from OpenAI, DeepMind, Harvard, and Stanford. Chief Scientist Tom McGrath led interpretability research at DeepMind before co-founding Goodfire.
Public benefit corporation — safety mission is legally embedded. Small team of ~51 with flat structure and high individual ownership across research and engineering.
$209M total funding: $7M seed (Lightspeed), $50M Series A (Menlo + Anthropic), $150M Series B (B Capital). Valued at $1.25B as of February 2026.