Cerebras Systems is not a typical AI company. While most of the industry trains and deploys models on clusters of NVIDIA GPUs, Cerebras took a radically different approach: build the largest chip ever made. A single processor the size of an entire silicon wafer — 46,225 square millimeters of compute containing 4 trillion transistors. It’s an engineering bet so audacious that most semiconductor veterans called it impossible when the company was founded in 2015.
Today, May 14, 2026, that bet goes public. Cerebras is listing on the Nasdaq under the ticker CBRS at a valuation of roughly $38 billion, after its IPO was oversubscribed by more than 20x. The company reported $510 million in revenue with 47% net income margins — numbers that would be impressive for any tech company, let alone a hardware startup that spent years in deep R&D before finding product-market fit.
But behind the IPO headlines, what is it actually like to work there? We analyzed 57 employee reviews, compensation data, and the company’s engineering culture to give you an honest picture of Cerebras as an employer — the good, the hard, and the trade-offs you should weigh before applying.
Cerebras at a Glance
| Founded | 2015 |
| Headquarters | Sunnyvale, CA |
| CEO | Andrew Feldman |
| Company Size | ~700 employees |
| Valuation | ~$38B (IPO day) |
| Revenue | $510M (76% growth YoY) |
| Glassdoor Rating | 4.0 / 5.0 (57 reviews) |
| Work-Life Balance | 3.8 / 5.0 |
| CEO Approval | 86% (Andrew Feldman) |
| Recommend to Friend | 82% |
| Culture Values | Eng-Driven, Learning, Product Impact |
Cerebras sits in a unique position in our Culture Directory. It’s one of only a handful of companies building custom silicon for AI — alongside Tenstorrent and a few stealth startups. At ~700 employees, it’s large enough to have real organizational structure but small enough that individual engineers still shape the product. The 4.0 Glassdoor rating places it in line with companies like Stripe and OpenAI, but with a very different work profile.
The Wafer-Scale Bet: Why Cerebras Exists
To understand the culture, you need to understand the product. Every semiconductor company in the world cuts silicon wafers into hundreds of individual chips. Cerebras doesn’t. They use the entire wafer as a single processor.
The latest WSE-3 (Wafer Scale Engine, third generation) has 900,000 AI-optimized cores, 4 trillion transistors, and delivers 125 petaflops of compute. For context, NVIDIA’s B200 — widely considered the best GPU on the market — has roughly 19x fewer transistors. Cerebras claims a 7,000x memory bandwidth advantage and inference speeds 10–70x faster than GPU clusters for large language models.
The engineering challenge is staggering. Manufacturing defects are inevitable on a piece of silicon this large, so Cerebras designed a “fail-in-place” architecture — redundant cores, redundant routing, and systems that detect and route around defective regions automatically. This is not a GPU competitor built on incremental improvements. It’s a fundamentally different approach to computing, and working on it requires engineers who are comfortable with first-principles thinking at scale.
Engineering Culture: First-Principles at Wafer Scale
Cerebras’ engineering culture is shaped by the audacity of its product. When you’re building something that the semiconductor industry considered impossible, you don’t have the luxury of following playbooks. Engineers at Cerebras work at the intersection of chip architecture, compiler design, systems software, and machine learning infrastructure — often within the same team.
The founding team — Andrew Feldman, Gary Lauterbach, Sean Lie, Michael James, and Jean-Philippe Fricker — previously built SeaMicro, a server company they sold to AMD for $334 million in 2012. That shared history matters. Cerebras isn’t a company of strangers who found each other through recruiting. The leadership team has worked together for over a decade, and that continuity permeates the culture with an unusual level of trust and directness.
Tech Stack & Engineering Domains
Engineering at Cerebras spans several distinct domains:
- Chip design. Hardware engineers work on the WSE architecture itself — core design, memory hierarchy, on-wafer networking, and the fail-in-place redundancy systems. This is deep RTL and physical design work.
- Compiler & systems software. A chip this different needs a custom software stack. Cerebras built its own compiler (based on LLVM) that maps ML workloads onto 900,000 cores, handling data flow, memory management, and communication across the wafer.
- ML infrastructure. Making the WSE accessible to researchers and enterprises means building frameworks, APIs, and integrations with PyTorch and other standard ML tooling. This team bridges the gap between custom hardware and the workflows data scientists actually use.
- Inference platform. Cerebras offers cloud-based inference services, and the team building this handles the full stack from hardware management to API design.
Multiple reviews highlight the caliber of colleagues as the standout perk. Employees describe teams where “most team members graduated from top CS universities and have Master+ degrees or years of work experience.” The mentorship culture is consistently praised — senior engineers invest real time in growing junior talent, which is unusual for a company operating at this pace.
What Employees Actually Say
We analyzed themes across Cerebras’ 57 Glassdoor reviews to surface the patterns beyond the aggregate numbers.
What employees love
The theme is clear: Cerebras attracts engineers who want to work on problems that don’t exist anywhere else. If building a wafer-scale compiler or designing a 4-trillion-transistor processor excites you at a gut level, this is one of the only places on Earth where you can do it. The mentorship culture and team autonomy add a collaborative dimension that distinguishes Cerebras from the “grind and ship” culture of many AI startups.
What could be better
The cons paint a picture of a company in growth mode that hasn’t fully matured its internal processes. The aggressive timelines are a natural consequence of competing against NVIDIA — a company with 30,000+ employees and decades of ecosystem lock-in. But several reviews note that the bias toward speed creates tech debt that compounds over time. The promotion transparency issue is common in pre-IPO companies and may improve now that Cerebras is public.
Glassdoor Ratings Breakdown
Cerebras’ overall 4.0 rating places it solidly among the best AI companies for engineers. Here’s how the sub-categories break down:
The 3.8 WLB score is honest for a hardware company racing to establish itself against NVIDIA. For comparison, CoreWeave — another AI infrastructure company in our directory — operates at a similar intensity. If work-life balance is your top priority, companies like Notion (4.2) or Linear (4.4) are better fits. But for engineers who thrive under intensity and want to work on generational hardware problems, the 3.8 is a reasonable trade-off.
Compensation & the IPO Effect
Compensation at Cerebras has been competitive but not at the top of the AI market — until now. Based on employee-reported data, here’s what the compensation landscape looks like:
Base salaries for Member of Technical Staff (MTS) roles — Cerebras’ primary engineering title — range from $164k to $196k, with total compensation reaching $200k–$350k+ when you include equity. The equity component is where things get interesting. With today’s IPO at a $38 billion valuation, early employees who received stock grants at earlier valuations could see life-changing returns.
Cash compensation alone puts Cerebras below frontier AI labs like Anthropic and OpenAI, where senior engineers can earn $300k–$500k+ in total comp. But the equity story is different. Cerebras has real revenue ($510M), real profitability (47% net margins), and 76% year-over-year growth. For engineers who joined in the last 2–3 years, the stock appreciation may more than make up for lower base comp.
Post-IPO, expect Cerebras to face pressure to raise cash compensation to compete with public company peers like NVIDIA and AMD, as well as well-funded AI startups. The IPO typically marks an inflection point where companies need to offer competitive RSU packages alongside salaries that match the broader market. For prospective candidates, this means the next 6–12 months may be an unusually good time to negotiate.
The IPO: What Changes for Employees
Going public transforms a company’s internal dynamics in predictable ways. Here’s what Cerebras employees should expect:
- Equity liquidity. The biggest immediate change. Pre-IPO stock grants were paper wealth — now they’re liquid. Employees who joined early at lower valuations may see significant paydays after lock-up periods expire (typically 90–180 days post-IPO).
- Increased scrutiny. Public companies face quarterly earnings pressure. This can shift internal culture from long-term R&D bets toward shorter-term revenue targets. Cerebras’ leadership will need to balance Wall Street expectations with the multi-year hardware development cycles their product requires.
- Process overhead. SOX compliance, insider trading windows, earnings blackout periods, and more formal financial controls are all part of being public. Engineers rarely feel this directly, but managers and leadership will.
- Hiring leverage. A successful IPO validates the company and makes it easier to recruit. Cerebras can now offer liquid RSUs instead of illiquid private stock, which is a meaningful recruiting advantage against startups.
The 20x oversubscription of the IPO suggests strong market confidence. For current employees, that’s validation. For prospective hires, it means Cerebras is likely to invest heavily in headcount growth — the 93 open positions on our platform today are probably just the beginning.
Who Thrives at Cerebras
Cerebras is not for everyone, and the company doesn’t try to be. Based on the culture signals, employee reviews, and the nature of the work, here’s who tends to excel:
- Hardware-software bridge engineers. The most impactful engineers at Cerebras understand both silicon and software. If you can reason about cache hierarchies, data flow across a wafer, and compiler optimization in the same conversation, you’re rare and valuable here.
- First-principles thinkers. There are no textbooks for wafer-scale computing. Engineers who default to “how has this been done before?” will struggle. Engineers who ask “what does the physics allow?” will thrive. This is a company where novel solutions are the norm, not the exception.
- People who want product impact at a hardware level. If you want to see your work change how AI models are trained and deployed at a fundamental level, Cerebras offers that. A single WSE-3 replaces entire racks of GPUs. Your engineering decisions ripple through the entire AI infrastructure stack.
- Engineers comfortable with ambiguity. At ~700 employees competing against NVIDIA’s 30,000+, Cerebras can’t plan everything meticulously. You’ll need to make decisions with incomplete information, prototype quickly, and course-correct. The pace is fast and the target moves.
Cerebras is not ideal for engineers who prioritize stability and predictability, prefer working in large teams with well-defined processes, or need strict work-life balance boundaries. The aggressive timelines and hardware development cycles mean crunch periods are real. If that’s a dealbreaker, consider companies like GitLab or HubSpot that offer more structured environments with better WLB scores.
How Cerebras Compares
Cerebras occupies a specific niche in the AI landscape. Here’s how it stacks up against comparable companies in our directory:
- vs. Tenstorrent: Also building custom AI chips, led by Jim Keller (legendary chip architect). Tenstorrent is smaller and pre-revenue. Cerebras has the advantage of proven product-market fit and a public offering. Both attract similar engineering profiles.
- vs. CoreWeave: CoreWeave builds GPU cloud infrastructure, primarily using NVIDIA hardware. Cerebras builds the chips themselves. Different technical challenges, similar intensity levels, similar compensation ranges.
- vs. Anthropic / OpenAI: The AI labs focus on model research and deployment. Cerebras builds the hardware those models run on. Compensation at the labs is higher (especially for research roles), but Cerebras offers more direct hardware engineering impact and IPO equity upside.
Use our comparison tool to see Cerebras side-by-side with any of the companies in our directory.
Open Positions at Cerebras
Cerebras currently has 93 open positions on our platform, spanning hardware engineering, software, ML infrastructure, and operations roles. Most positions are based in Sunnyvale, California, with some hybrid arrangements. Post-IPO, expect this number to grow as the company invests in scaling its team.
For the full list of roles, culture values, and side-by-side comparisons, visit the Cerebras culture profile page.
Frequently Asked Questions About Working at Cerebras
Explore Cerebras’ 93 open roles
Hardware engineering, ML infrastructure, compiler design — see all positions with culture context and employee insights.
View Cerebras Profile → Browse Cerebras Jobs →