The Environmental Cost
Training and running LLMs has a significant environmental footprint: Training energy — training GPT-3 consumed approximately 1,287 MWh of electricity and emitted ~552 tonnes of CO2 (equivalent to 123 gasoline cars driven for a year). Larger models like GPT-4 are estimated to be 10–100x more. Inference energy — a single ChatGPT query uses roughly 10x the energy of a Google search. With hundreds of millions of daily queries, inference costs dominate over time. Water usage — data centers use enormous amounts of water for cooling. Microsoft reported a 34% increase in water consumption (2023), largely attributed to AI workloads. Hardware lifecycle — GPU manufacturing requires rare earth minerals, and hardware is replaced every 2–3 years. The ethical question: is the benefit of AI worth its environmental cost?
Environmental Footprint
// Environmental cost of LLMs
Training:
GPT-3: ~1,287 MWh, ~552t CO2
// = 123 cars for one year
GPT-4: estimated 10-100x more
Llama 3 405B: ~30,000 GPU-hours
Inference:
ChatGPT query: ~10x Google search
Hundreds of millions of queries/day
Inference > training over time
Water:
Microsoft: +34% water use (2023)
Google: +20% water use (2023)
Data center cooling is thirsty
Mitigation:
Smaller models (distillation)
Efficient architectures (MoE)
Renewable energy data centers
Model sharing (open source)
Inference optimization (quantization)
Carbon-aware scheduling
Reporting:
Disclose training compute
Report carbon footprint
// Few companies do this today
Key insight: Inference energy dominates over time because training happens once but inference runs continuously. The most impactful environmental intervention is making models smaller and more efficient (distillation, quantization, MoE), not just using renewable energy for training.