MatX Raises $500M Series B — Why Wall Street Is Betting on Custom LLM Chips
MatX just secured $500M in Series B funding to build custom chips for large language models. Led by Jane Street Capital and Situational Awareness, this massive round signals a fundamental shift in AI infrastructure — specialized hardware is becoming critical, and financial firms are leading the charge.

MatX just closed a $500M Series B round led by Jane Street Capital and Situational Awareness, two firms not typically known for early-stage hardware bets. The startup designs custom chips and hardware architectures built specifically for large language models — and that laser focus is why some of the smartest money in tech is betting big.
This isn't just another AI funding story. It's a signal that the infrastructure layer of AI is fragmenting. General-purpose GPUs dominated the first wave of deep learning. But LLMs have specific computational patterns — massive matrix multiplications, attention mechanisms, transformer layers — that don't map efficiently to Nvidia's architecture. MatX is building silicon that does one thing extremely well: run frontier-scale language models faster and cheaper than anything else.
Why Jane Street Is Involved
Jane Street Capital is a quantitative trading firm that runs some of the most optimized computational systems on the planet. They don't invest in buzzwords. They invest in fundamental efficiency gains.
Their bet on MatX tells you two things:
- LLM inference costs are a massive, unsolved problem — Running models like GPT-4 or Claude at scale costs millions per month. Dedicated hardware could cut that by 5-10x.
- Financial firms see AI infrastructure as strategic — Proprietary trading shops, hedge funds, and market makers are building internal AI systems. They need compute that's both fast and economical.

The Custom Chip Thesis
Nvidia's H100 and B200 GPUs are incredibly powerful, but they're designed for a wide range of workloads — training, inference, simulation, graphics. That flexibility comes at a cost: efficiency.
MatX is taking the opposite approach: specialize everything.
- Custom memory hierarchies optimized for attention layers
- Tensor cores designed for the specific operations transformers use
- Lower precision arithmetic where it doesn't hurt model quality
- Tighter integration between compute and networking for multi-chip inference
This matters because LLMs are hitting a hardware wall. Models are getting bigger, but GPU scaling isn't keeping pace. You can't just throw more H100s at the problem — at some point, interconnect bandwidth and memory become the bottleneck.
MatX is designing around that reality.
Who Else Is Playing This Game
MatX isn't alone. The custom AI chip market is heating up:
- Groq — Raised $640M to build deterministic inference chips (now shipping LPUs to customers)
- Cerebras — Wafer-scale AI chips for training and inference (went public via SPAC)
- SambaNova — Just raised $350M for enterprise AI chip systems
- d-Matrix — Focused on in-memory compute for transformers
But MatX has something most don't: backing from Jane Street and Situational Awareness, two firms that understand computational efficiency at the atomic level. That's not just capital — it's strategic guidance from people who've built some of the fastest systems in the world.
What This Means For Your Business
If you're running AI at scale, here's what the MatX round should tell you:
- LLM infrastructure is about to get cheaper — Custom chips will drive down inference costs over the next 18-24 months. Plan your budgets accordingly.
- Specialized beats general-purpose for production workloads — If you're running the same model architecture continuously (customer support, code generation, document processing), dedicated hardware makes financial sense.
- Hardware diversity is coming — The Nvidia monoculture is ending. Expect cloud providers to offer multiple chip types for different AI workloads.
The Bigger Picture
We're watching the AI stack mature. The first wave was about algorithms — transformers, attention, scaling laws. The second wave was about infrastructure — GPUs, cloud platforms, orchestration. The third wave is specialization.
Just like mobile phones moved from general-purpose ARM chips to custom Apple Silicon and Google Tensor, AI is moving from general-purpose GPUs to task-specific accelerators.
MatX is betting that the biggest task — running LLMs in production — is large enough to justify custom silicon. Given that OpenAI spends an estimated $700K per day on compute, and Anthropic is burning similar amounts, that bet looks pretty solid.
What To Watch Next
MatX hasn't publicly announced timelines, but $500M buys a lot of silicon engineers and fab capacity. Expect:
- First chips in 12-18 months — Hardware takes time, but well-funded teams can move fast
- Cloud partnerships — These chips won't sell to end customers directly; they'll show up in AWS, GCP, or Azure
- Benchmark wars — Expect MatX to publish aggressive performance comparisons vs. H100 for specific LLM workloads
The real test will be whether they can ship production silicon that actually delivers the promised efficiency gains. Hardware is hard. But with Jane Street's backing and a $500M runway, they've got a real shot at redefining what LLM inference looks like.
Build AI That Works For Your Business
At AI Agents Plus, we help companies move from AI experiments to production systems that deliver real ROI. Whether you need:
- Custom AI Agents — Autonomous systems that handle complex workflows, from customer service to operations
- Rapid AI Prototyping — Go from idea to working demo in days using vibe coding and modern AI frameworks
- Voice AI Solutions — Natural conversational interfaces for your products and services
We've built AI systems for startups and enterprises across Africa and beyond.
Ready to explore what AI can do for your business? Let's talk →
About AI Agents Plus Editorial
AI automation expert and thought leader in business transformation through artificial intelligence.



