Gimlet Labs just closed an $80 million Series A to tackle one of AI's thorniest infrastructure problems - getting models to run seamlessly across wildly different chip architectures. The startup's platform lets companies deploy AI workloads across Nvidia, AMD, Intel, ARM, Cerebras, and d-Matrix processors simultaneously, without rewriting code or dealing with vendor lock-in. It's a solution that couldn't come at a better time, as enterprises struggle with GPU shortages and skyrocketing inference costs.
The AI industry has a dirty secret - most companies are locked into Nvidia's ecosystem whether they like it or not. Training frameworks, inference engines, and deployment tools are built for CUDA, making it brutally expensive and time-consuming to switch providers. Gimlet Labs just raised $80 million to blow that bottleneck wide open.
The Series A round, reported by TechCrunch, backs technology that lets AI models run across Nvidia, AMD, Intel, ARM, Cerebras, and d-Matrix chips without developers needing to touch a single line of hardware-specific code. It's the kind of abstraction layer that sounds simple but solves a multi-billion dollar headache.
Here's why this matters now. Enterprise AI spending is exploding, but GPU availability isn't keeping pace. Companies are paying premium prices for Nvidia H100s when cheaper alternatives from AMD or specialized inference chips from Cerebras sit underutilized. The problem isn't performance - it's compatibility. Switching chips means rewriting inference pipelines, retraining operations teams, and risking production outages.
Gimlet's approach treats chip architecture as a backend detail rather than a frontend constraint. Developers write inference code once, and the platform handles translation across different silicon. Think of it as Kubernetes for AI chips - you declare what you need, and the system figures out where to run it. That flexibility becomes crucial as ramps MI300 production, pushes Gaudi accelerators, and startups like ship purpose-built inference silicon.












