Microsoft operates Azure, a global cloud platform providing large-scale GPU clusters for internal and customer AI workloads.
Overall, Microsoft’s accessible AI compute is roughly 1×10^20 dense INT8 operations per second.
This scale allows Microsoft to serve commercial customers and partners like OpenAI, but concentrates critical AI infrastructure under one corporation.
Observer, “Microsoft’s Massive GPU Purchases Fuel AI Race,” 2023. https://observer.com/microsoft-150000-h100-gpus/ ↩
Colfax, “NVIDIA H100 Tensor Core GPU,” 2023. https://colfaxresearch.com/nvidia-h100-performance/ ↩