Discussion about this post

User's avatar
Kenny Easwaran's avatar

This is very helpful, since I’m updating a lecture on the resource use of AI for an AI literacy class I’m teaching right now!

A few minor points - in my academic field (philosophy) there are some people who are complaining about the carbon emissions associated with conference travel, and arguing that we should have more online conferences and fewer in-person conferences. (See, eg, the comments here: https://dailynous.com/2026/01/15/apa-to-end-experiment-with-online-divisional-meetings/ ) I expect there are some people making similar arguments on computer science.

One thing that this really drives home for me is that there’s a difference between the general resource use of AI and specifically the electricity use. Unlike legos and CDs, which involve moving physical objects and making things out of refined petroleum, AI has its emissions almost entirely through electricity. It looks like there are individual data centers that use as much electricity as the entire city of San Diego! (For instance, the Amazon datacenter for Claude training in New Carlisle, IN: https://epoch.ai/data/data-centers/ .) But the carbon emissions and water use of this datacenter are more comparable to the emissions and water use of South Bend, or perhaps even New Carlisle. The water use change could be a significant effect locally, but it’s nowhere near the potential disruption of having to power a new San Diego in the middle of Indiana!

The issues of water and emissions will be handled by the same processes that are handling those issues for everything else in the world. But electricity use will be a significantly different type of issue to deal with, especially as we also electrify transportation, cooking, and climate control.

XP's avatar
40mEdited

Great work as always. My gut reaction to hearing any of these comparisons for the first time has always been: "But... that's _tiny_!"

And speaking of GTA V, I also have to wonder about the emissions of hundreds of employees driving to work for six days a week, for five long years, all using high-end workstations and servers...

There's a talking point that surfaced over the past few months, where people claim that the rapid iteration of new models is making the training energy consumption that much worse (and the profitability of any models unlikely), e.g. "GPT-5.0 was followed by GPT-5.1 just six weeks later!" Of course, the various versions of GPT-5.x don't involve the same massive pretraining of the entire base model, as they're mostly fine-tunes, merges, RLHF/SFT and whatever other arcane arts the AI labs employ.

No posts

Ready for more?