Microsoft just lifted the curtain on the world's most powerful AI facility. The company's new Fairwater datacenter in Wisconsin houses hundreds of thousands of cutting-edge NVIDIA GPUs in a single massive supercomputer that delivers 10 times the performance of today's fastest systems. This isn't just another cloud facility - it's Microsoft's blueprint for competing in the AI arms race.
Microsoft just dropped the most detailed look yet at the infrastructure powering the AI revolution. The company's new Fairwater datacenter in Wisconsin represents a fundamental shift in cloud computing - from traditional servers handling diverse workloads to purpose-built AI factories optimized for training frontier models.
The scale is staggering. This single facility spans 315 acres, required 46.6 miles of foundation piles and 26.5 million pounds of structural steel to construct. But the real story is what's inside: hundreds of thousands of NVIDIA GPUs operating as one massive supercomputer that will "deliver 10X the performance of the world's fastest supercomputer today," according to Scott Guthrie's detailed blog post.
Azure was the first cloud provider to deploy NVIDIA's new GB200 servers at scale, and each rack packs 72 NVIDIA Blackwell GPUs into what Guthrie calls "a single, giant accelerator." These racks process an astonishing 865,000 tokens per second - the highest throughput of any cloud platform available today. The GPUs share 14 terabytes of pooled memory through NVIDIA's NVLink technology, essentially eliminating traditional memory bottlenecks.
But Microsoft isn't stopping at Wisconsin. The company revealed it's building "multiple identical Fairwater datacenters under construction in other locations across the US," while also announcing new AI facilities in Norway and the UK through partnerships with nScale. These datacenters will use NVIDIA's next-generation GB300 chips with even more memory per rack.
What makes these facilities different from traditional cloud datacenters is their architecture. Instead of running thousands of independent workloads, everything is designed around a single flat network connecting all GPUs. "From the outside, this architecture looks like many independent servers, but at scale it functions as a single supercomputer," Guthrie explained.
The networking challenge is immense. At the rack level, GPUs communicate at terabytes per second through NVLink and NVSwitch. Between racks, Microsoft uses both InfiniBand and Ethernet fabrics delivering 800 Gbps in what the company calls "a full fat tree non-blocking architecture." The Wisconsin facility even uses a two-story configuration to reduce physical distances between racks, minimizing latency.