You want to talk edge? Let’s talk silicon. While everyone else was busy chasing ChatGPT integrations like it was Pokémon Go for enterprise, Groq Inc. built the AI version of a hypercar engine, then slipped it into the cloud and hit the gas.
Founded by Jonathan Ross, the engineer who kickstarted Google’s TPU project before leaving Google X with more ambition than budget, Groq is now raising up to $500 million at a $6 billion valuation, and doing it on the back of a $1.5 billion infrastructure deal with Saudi Arabia. That’s not hype. That’s geopolitical-grade validation.
But let’s not get lost in the zeros. What matters here is why they’re getting that check. Groq didn’t try to play Nvidia’s game; they built a new one. Their Language Processing Unit (LPU) isn’t a rebranded GPU; it’s a deterministic beast designed for real-time inference, not academic benchmarks. No caches. No branch prediction. No CUDA babysitting. Just raw, predictable speed at batch size one. You want 826 tokens per second on Gemma? Done. You want latency under 300 milliseconds on Mixtral? Already live.
And while competitors are still doing datacenter foreplay, Groq is dropping 108,000 LPUs into production by Q1 2025, from Mountain View to Dammam to Helsinki.
Jonathan Ross may be the visionary, but this isn’t a one-man symphony. Sunny Madra, the serial entrepreneur behind multiple exits, now runs operations, GTM, and GroqCloud like it’s his fourth company because it is. Mohsen Moazami is translating Groq’s silicon into global reach. Claire Hart keeps the legal lanes clear, and Allison Hopkins is building a talent engine that scales like their chips. Let’s not forget Chelsey Susin Kantor, taking Groq’s brand from engineering whisper to cloud-age anthem.
And behind the cap table, this isn’t your average startup term sheet roulette. BlackRock Private Equity Partners led the $640M Series D in 2024, backed by Cisco Investments, Samsung Catalyst Fund, Neuberger Berman, and a dozen others who know what it means when silicon meets scale and sovereign wealth.
This raise isn’t just fuel. It’s a marker. A signal that inference, the quiet, unsexy cousin of AI training, is now where the real money moves. Training may make headlines, but inference runs the world. Groq gets it. So do 360,000+ devs already building on GroqCloud.
From TPU to LPU, from Google X to Dammam, Jonathan Ross and team didn’t just imagine a new compute future; they’re wiring it into the fabric of global AI infrastructure.

