xAI Co-Founders All Gone: Musk Says Grok 'Built Wrong'
All 11 xAI co-founders have left and Musk admits Grok was 'built wrong.' It ranks 8th on AI benchmarks — with 3M+ deepfakes traced to the platform.
When every single founding engineer quietly walks out the door, something has gone fundamentally wrong. At xAI — Elon Musk's AI company behind the Grok chatbot — reports indicate all 11 co-founders have now departed. That's not a restructuring. That's a rout. And Musk's own reported admission — that xAI was "built wrong" — makes the exits harder to explain away.
This isn't just a corporate drama story. Grok is actively used by millions of X (formerly Twitter) Premium subscribers. If the architects who built it have all left, users face a real question: can a product whose entire founding team has walked out still be trusted?
xAI: All 11 Co-Founders Gone, Zero Remaining
The xAI founding team was assembled in mid-2023 from the most competitive AI talent pools on the planet — researchers and engineers poached from Google DeepMind, OpenAI, Tesla's Autopilot division, and top academic institutions. These weren't filler hires. They were the specific individuals Musk recruited to challenge the very labs he had previously co-founded and then publicly criticized.
Founder departures at tech startups aren't rare. One or two leaving in year one is unremarkable. Three or four might signal a strategic disagreement. But when the reported count reaches eleven — every member of the founding cohort — the pattern is unmistakable. Industry observers typically point to several root causes when a full founding exodus occurs:
- Direction conflict: The company's actual trajectory diverges from what founders signed up to build
- Technical judgment overruled: A dominant founder repeatedly overrides engineering recommendations
- Safety or ethics disagreements: Researchers uncomfortable with deployment pace or product governance
- Equity or compensation disputes: Especially common during rapid organizational scaling
What makes the xAI situation unusual is Musk's own reported acknowledgment. The phrase "built wrong" — attributed to Musk in industry reporting — is an extraordinary admission from a founder. It's the kind of language that rarely reaches public discourse unless the internal situation is significantly worse than what's visible from the outside. It also reframes the co-founder departures: if the founding team was building something the company's own CEO now considers fundamentally flawed, the exits look less like defections and more like a logical response.
Grok AI Benchmark Reality vs. 'Lowest Hallucination' Claims
xAI has aggressively marketed Grok on a specific, high-stakes differentiator: trustworthiness. The company's "lowest hallucination rates" claim — meaning Grok supposedly invents fewer false facts than competitors — has been central to its pitch, particularly to enterprise users, journalists, and researchers who depend on accurate information.
Independent benchmark data tells a different story. On third-party intelligence evaluations — standardized tests that measure AI reasoning, math problem-solving, factual recall, and language understanding (think of them as comprehensive academic exams for AI systems) — Grok currently ranks 8th among major AI models. For a company that launched specifically to compete with OpenAI and Google, eighth place is a difficult position to defend against its own marketing claims.
The Problem With Grok's Self-Reported Hallucination Numbers
Hallucination measurement — tracking how often an AI confidently states something completely false — is notoriously inconsistent across the industry. Different labs test different knowledge domains, use different evaluation sets, and sometimes define "hallucination" in ways that favor their own models. A system can score well on the benchmark it was specifically trained against while performing significantly worse on independent, blind evaluations run by third parties.
The gap between xAI's internal hallucination claims and Grok's 8th-place ranking on independent benchmarks is precisely the kind of discrepancy that causes researcher frustration — and that a technically strong founding team would likely have flagged internally. The departures may reflect exactly that tension: engineers who built the system knowing what it actually does, versus marketing claims that describe something different.
For comparison, models currently ranked in the top positions by independent evaluators include offerings from Anthropic, OpenAI, and Google DeepMind — the three companies xAI positioned itself as a challenger to when it launched in 2023.
3 Million xAI Deepfakes: The Number That Demands Accountability
The most damaging figure associated with xAI isn't its benchmark ranking — it's the reported count of over 3 million AI-generated deepfakes traced to the Grok platform. A deepfake (synthetic photo or video designed to look real, typically of a real person without their consent) represents one of the most serious harms in consumer AI, enabling harassment campaigns, non-consensual intimate imagery, financial fraud, and political disinformation.
Unlike hallucinations — which are unintended errors — deepfakes are frequently intentional misuse. Platform responsibility means implementing guardrails (technical restrictions that prevent specific harmful outputs) before those tools reach the public. The industry standard has moved decisively in that direction:
- OpenAI's DALL-E 3 implemented identity protection filters and blocked recognizable-person synthesis before general release
- Midjourney banned all political figure imagery ahead of the 2024 election cycle
- Stability AI faced regulatory scrutiny and multiple lawsuits over deepfake-enabling capabilities
- Grok on xAI: 3+ million deepfakes generated — suggesting governance controls were either absent or insufficient at scale
Three million is not an edge case. It's a policy failure at operational scale. For context, that volume of harmful synthetic content is large enough to meaningfully affect real people's lives, reputations, and safety. With all 11 co-founders reportedly gone and no clear technical leadership to point to, the critical question — who owns this problem and is actively fixing it — currently has no obvious answer.
Should You Keep Using Grok? A Practical Guide
If Grok is part of your daily workflow, the evidence above suggests four concrete actions:
Treat every factual claim as a draft, not a source. An 8th-place benchmark ranking means Grok generates more errors than the 7 models currently ranked above it. For casual idea generation or brainstorming, that margin may be acceptable. For anything requiring accuracy — research, legal summaries, financial analysis, medical information — verify Grok's output against primary sources before relying on it.
Watch for governance policy changes. External pressure over the deepfake figures will force xAI to implement new restrictions on image generation. Those changes will affect workflows that currently depend on Grok's image capabilities. Expect policy announcements in the coming weeks.
Know your alternatives. Models currently ranking ahead of Grok on independent third-party benchmarks include Claude 3.5 Sonnet (Anthropic), GPT-4o and o3 (OpenAI), and Gemini 1.5 Pro (Google). All have free access tiers. If accuracy is your priority, our AI automation model comparison guide breaks down which tool fits which use case.
Watch what you generate. If you use Grok for image creation or AI automation workflows, be aware that platform rules may tighten significantly. Any workflow relying on current permissiveness in Grok's image generation could be disrupted if governance changes land quickly.
The xAI story is still unfolding — and may accelerate as regulatory pressure over deepfakes increases globally. What's already documented is stark: a company whose CEO reportedly admitted it was built wrong, with all 11 founding engineers departed, a benchmark ranking that contradicts its own marketing, and a deepfake count measured in millions. Whether xAI can rebuild credibly from that starting point is the AI story worth watching in 2026. For ongoing coverage of the AI industry, follow our AI automation news.
Related Content — Get Started | Guides | More News
Stay updated on AI news
Simple explanations of the latest AI developments