bolt

AI Infrastructure

GPUs, accelerators, training clusters, inference serving, networking, cooling — the hardware and systems that power AI
Co-Created by Kiran Shirol and Claude
Topics GPUs & Accelerators Networking Training Clusters Inference Serving Power & Cooling Cloud & Orchestration
home Learning Portal play_arrow Start Learning summarize Key Insights dictionary Glossary 14 chapters · 5 sections
Section 1

Foundation — Why Special Hardware?

CPUs vs GPUs, architecture deep dive, and the accelerator landscape.
Section 2

Core Techniques — Connecting the Dots

Memory bottlenecks, interconnects, and network topologies for AI clusters.
Section 3

Training at Scale

Distributed training parallelism and the anatomy of an AI factory.
Section 4

Serving and Storing

Inference infrastructure, storage pipelines, and the energy question.
Section 5

Strategy — Making It Work

Cloud vs on-prem, orchestration, and the future of AI infrastructure.