AI's energy footprint
This program posts news to thousands of machines throughout the entire civilized world. Your message will cost the net hundreds if not thousands of dollars to send everywhere. Please be sure you know what you are doing. Are you absolutely sure that you want to do this? [ny]
Maybe we meed something similar in LLM clients. Could be phrased in terms of how many pounds of atmospheric carbon the request will produce.
Likewise, I doubt that USENET warning was ever true beyond the first few years of the networks' lifetime. Certainly if everything was connected via dial-up, yes, a single message could incur hundreds of dollars of cost when you added the few seconds of line time it took to send up across the whole world. But that's accounting for a lot of Ma Bell markup. Most connections between sites and ISPs on USENET were done through private lines that ran at far faster speeds than what you could shove down copper phone wiring back then.
Are you saying all of that new capacity is needed to power non-LLM stuff like classifiers, adtech, etc? That seems unlikely.
Had you said that inference costs are tiny compared to the upfront cost of training the base model, I might have believed it. But even that isn't accurate -- there's a big upfront energy cost to train a model, but once it becomes popular like GPT-4, the inference energy cost over time is dramatically higher than the upfront training cost.
You mentioned batch computing as well, but how does that fit into the picture? I don't see how batching would reduce energy use. Does "doing lots of work at once" somehow reduce the total work / total energy expended?
If you want to know more about energy consumption, see this 2 part series that goes into tons of nitty-gritty details: https://blog.giovanh.com/blog/2024/08/18/is-ai-eating-all-th...
If what you're saying is true, why are we hearing about AI companies wanting to build nuclear power plants to power new data centers they think they need to build?
Something to temper this, lots of these AI datacenter projects are being cancelled or put on hiatus because the demand isnt there.
But if someone wants to build a nuke reactor to power their datacenter, awesome. No downsides? We are concerned about energy consumption only because of its impact on the earth in terms of carbon footprint. If its nuclear, the problem has already been solved.
Something to temper this, lots of these AI datacenter projects are being cancelled or put on hiatus because the demand isnt there.
Wait, any sources for that? Because everywhere I go, there seems to be this hype for more AI data centers. Some fresh air would be nice.
AI seems like it is speedrunning all the phases of the hype cycle.
TD Cowen analysts Michael Elias, Cooper Belanger, and Gregory Williams wrote in the latest research note: “We continue to believe the lease cancellations and deferrals of capacity points to data center oversupply relative to its current demand forecast.”
If what you're saying is true, why are we hearing about AI companies wanting to build nuclear power plants to power new data centers they think they need to build?
Well, partly because they (all but X, IIRC) have commitments to shift to carbon-neutral energy.
But also, from the article:
ChatGPT is now estimated to be the fifth-most visited website in the world
That's ChatGPT today. They're looking ahead to 100x-ing (or 1,000,000x-ing) the usage as AI replaces more and more existing work.
I can run Llama 3 on my laptop, and we can measure the energy usage of my laptop--it maxes out at around 0.1 toasters. o3 is presumably a bit more energy intensive, but the reason it's using a lot of power is the >100MM daily users, not that a single user uses a lot of energy for a simple chat.
Even driving your car around you at least are somewhat aware of the gas you are burning.
rubber on roads
Funny how this suddenly became a thing after electrification became a thing. Need to find a new way to wag the finger after all.
The oil industry is a conglomerate of degenerates spamming boomer logic all the way down to the workers. Their memes propagate throughout society and lead to the other boomer characteristic of rewriting personal and societal history.
The finger waggers now are being programmed to pretend they talked about tire particulates and the carheads are being programmed to pretend they never cared about 0-60. This another "We have always been at war with Eastasia", just like they all opposed the Iraq war from day 1 and didn't cancel the Dixie Chicks, et cetra.
This may have been discussed in specialist literature somewhere but even when I did ecology courses in university circa 2001ish, I never heard about tire particulates, while I did hear a lot about greenhouse gasses.
It's a concern but not a civilization ending concern like climate change. I low key resent these attempts to move goalposts to satisfy the writer's urge for negativity.
Consider that a bus has six to ten tires that each weigh around ten times more than a typical car tire. This is presented as the alternative to cars, is it even any different? Not implausible that it could actually be worse, especially if the bus isn't at full occupancy at all times.
Meanwhile the weight difference between EVs and petroleum cars is emphasized in the complaints, even though it isn't very large, while the much larger weight difference between any cars and buses is ignored. Because the point isn't to complain about tires, it's to complain about EVs.
And if the point actually was to complain about tires then you still wouldn't be talking about EVs, you would be talking about tires and how to make them shed less or construct them out of lower toxicity materials etc.
To make those sorts of calculations easy, you can ignore all the pressure/usage/etc nonsense and just do basic math on tire dimensions (including min/max tread depth and width, not just radius, though I typically ignore siping and whatnot) and typical longevity. Volume lost per mile driven is basic high-school arithmetic, and the only real questions are regarding data quality and whether the self-imposed constraints (e.g., examining real-world wear rather than wear given optimal driving or something) are reasonable.
The city bus uses tires with a harder rubber and dimensions such that the pressure at the road is less
Harder rubber seems like it could make a difference, but then you could also put tires with harder rubber on a car.
You can get a heavier vehicle to have the same pressure at the road by using more and bigger tires, but then the problem is that the tires are bigger and there are more of them.
plus its normal driving patterns have less wear than typical Tesla use.
Isn't a city bus constantly starting and stopping, both as a result of city traffic and picking up and dropping off passengers?
To make those sorts of calculations easy, you can ignore all the pressure/usage/etc nonsense and just do basic math on tire dimensions (including min/max tread depth and width, not just radius, though I typically ignore siping and whatnot) and typical longevity.
I tried plugging these in and it still comes out as a 6-wheel commercial bus has several times the tire wear as a 4-wheel light truck, rather than being the same.
And I expected the difference to be even more, but I guess that goes to show how much the weight argument is motivated reasoning if ~7x the weight is only ~3x the tire wear and then people are complaining about something which is only ~1.2x the weight.
Tire temperature also will play a big roll in tire wear, and I wouldn't expect bus tires to get very hot only rolling half the time and at a lower speed than the typical car.
And of course you also gotta factor in passenger count. Buses generally have more than just 1 or 2 people, while the vast majority of cars will have 1 or 2 people most of the time. And even if a bus tires were to wear out twice as fast as a car's tire, that is still less wear per person than a car.
I tried plugging these in and it still comes out as a 6-wheel commercial bus has several times the tire wear as a 4-wheel light truck, rather than being the same.
Pardon me if I ask the obvious question, but did you divide your result by the average number of people moved? Because that's the actual utility of mass vs. individual transport. I would find it rather surprising if tire wear was the one measure were buses didn't win out.
The biggest problem with tailpipe emissions used to be horrendous smog. That was mostly solved in many places, and now the biggest problem is the impact on the global climate.
The biggest issue with childhood mortality used to be disease. Now we (correctly) focus more on accidental deaths.
EVs solved tailpipe emissions, but they’re not perfect. Their biggest problem is just something else.
There’s been decades of lies about climate change. And once the truth got out society was already massively dependent on it. For cars specifically it was a deliberate policy to make e.g. the US car-dependent. And once the truth got undeniable the cope was switched to people’s “carbon footprint” (British Petroleum). In fact there are rumors that the cope echoes to this day.
Zoom out enough and it becomes obviously unproductive to make “mass ignorance” the locus of attention.
Or, even worse God forbid, think about how much carbon is produced to create a single bottle or carton of water. Then consider how casually people down bottles of water.
Emissions should be fixed on the production side (decarbonization) not on the demand side (guilt/austerity).
Scaling up battery production makes EVs more appealing on the demand side. How do you disincentivize fossil fuel production?
The problem with fossil fuels isn’t that they pollute, but that most of the negative impact of that pollution is borne by others. This results in an artificially low price which distorts the market and results in economically inefficient overuse.
Capture that cost by making producers pay a tax of the corresponding amount, and market forces will use the “right” amount of fossil fuels in ways that are a net benefit.
Do you really think the average person could within 2 orders of magnitude when estimating their carbon footprint for a year?
Why? AI isn't a human being. We have no obligation to be "fair" to it.
I can picture an Elizabeth Holmesian cartoon clutching her diamond necklace.
Oh, won't somebody think of the tech billionaires?!
If you don't freak out about running your shower or microwave for a couple seconds or driving a few hundred feet
The basic premise of the modern tech industry is scale. It's not one person running a microwave for a couple of seconds, it's a few billion people running a microwave for the equivalent of decades.
If memory serves Jet A is not taxed at all federally in the case of for-profit corporations (while non-commercial users DO pay a tax!) and many states also either do not tax it or tax it very litte.
It's completely insane that we do not tax fuel usage for probably the most energy-intensive way to move people and/or goods and often that movement of people is entirely frivelous.
Joe driving to work spends a larger fraction of his income on fuel and thus fuel tax than his rich counterpart.
This is true for all "let the consumer/poluter pay" taxes, they're all regressive. They say: it's fine to burn up the world as long as you're rich.
Personally I like the idea of setting the price for emitting 1 ton of CO2 equivalent emissions to the realistic cost of capturing 1 ton of CO2. At least, that seems like a reasonable end goal for a carbon tax, since that could fully account for the negative externality. This would of course be obscenely expensive, which would be a strong incentive to lower carbon emissions where possible, and for people to consume less of products that require large emissions to make or use.
The carbon tax would also have to apply to imported goods to be effective, and tracking how much tax should apply to imports would be even more difficult than doing so for domestic sources of pollution.
They're incredibly unpopular because the ultraweathly use massive amounts of fossil fuels and thus lobby very, very hard against them...and make sure the public is often told just how evil they are and how expensive they'd hurt Johnny Everyday Worker, even car ownership, especially in a city (where much of the US populative lives) is not affordable to a large segment of the population.
Eh. It's not Bill Gates and Alice Walton. Sometimes the obvious answer is the real one: It's the fossil fuel industry.
It's completely insane that we do not tax fuel usage for probably the most energy-intensive way to move people and/or goods and often that movement of people is entirely frivelous.
That one's just the arbitrage problem. Planes move around. If there is an international flight to a country that doesn't tax jet fuel (or taxes it less) then the plane is going to fly into LAX with enough fuel still in the tank to get back to the other jurisdiction and fill up again. Which actually increases fuel consumption because fuel is heavy and they otherwise wouldn't want to do that.
This is the same reason the EU doesn't tax jet fuel.
Are solar panels convenient? All polysilicon today is made with fossil fuels, and the R&D to make it with renewable energy is still in-progress. Not to mention that we ship them across the ocean with fossil fuel.
They've been implemented all over the world, because they're effective. They cover 31% of emissions in developed nations.
To whatever degree you could say they are unpopular, they're unpopular in regions where the government doing stuff about climate change (or just "the government doing stuff") is unpopular, which makes it odd to single out putting a price on carbon specifically
See where they are used here: https://carbonpricingdashboard.worldbank.org/
You can see traffic. It's easy to understand the dangers in a collision because when you drive into something unexpectedly your body takes a hit and you get frightened since you immediately realise that it might cost you a lot of money but you don't know for sure.
Being subtly manipulated by a disgustingly subservient fake conversationalist is another thing altogether.
The only purpose is to scapegoat the possible environmental or economic fallout. Might as well put it on individuals. Like what’s always done.
I’ve already seen it on the national broadcast. There some supposed experts were wagging their fingers about using AI for “fun”. Making silly images.
Meanwhile we’re gonna put AI to good use in arms races: more spam (automated applications, ads, ads, ads, abuse of services) and anti-spam. There’s gonna be so much economic activity. Disruptive.
So they used to send this message, but then it stopped I assume. Costs lowered a lot or the benefits outweighed all associated costs. Same can happen here.
Maybe that says the fees aren't yet high enough for high income people to change behavior, but I'm willing to bet they never truly will be due to the influence this subset of the population holds over politics.
Even if rich people don’t consume much more energy than poor people (I have no idea, just engaging with your idea as stated), they must be buying something with their money… carbon taxes should raise the price of goods with lots of embodied carbon.
If they aren’t consuming much energy and am they aren’t buying stuff with much embodied carbon… I dunno, I guess that’s the goal, right?
As for the stove, how much it uses is directly related to the kind of cooking you do, and for how long.
You could do it better than we are doing now, but you'll always have people saying: "that's unfair, why are you picking on me"
Mind you people won't like that since we're so used to using the atmosphere as a free sewer. The idea of having to pay for our pollution isn't palatable since the gasses are mostly invisible.
Though it's sad that we're talking about market solutions rather than outright bans for the majority of applications like we did for leaded gas.
Meanwhile the people with a 10 year old car they drive 5000 miles a year will keep it until it's a 20 year old car, at which point they'll buy another 10 year old car, but by then that one will run on electricity.
Then you could theoretically ban it, but by then do you even need to?
Even with things like orphaned natural gas that gets flared otherwise - rescuing the energy is great but we could use it for many things, not just LLMs or bitcoin mining!
the clean energy could displace other dirty generation.
If you would have built 10GW of solar or nuclear to replace other generation and instead the data center operators provide funding to build 20GW so that 10GW can go to data centers, the alternative wasn't replacing any of the other dirty generation. And the economies of scale may give the non-carbon alternatives a better cost advantage so you can build even more.
Increasing demand can lead to stimulus of green energy production.
I wouldn't be surprised if mankind will evolve similar to an organism and use 20% of all energy it produces on AI. Which is about 10x of what we use for software at the moment.
But then more AI also means more physical activity. When robots drive cars, we will have more cars driving around. When robots build houses, we will have more houses being built, etc. So energy usage will probably go up exponentially.
At the moment, the sun sends more energy to earth in an hour than humans use in a year. So the sun alone will be able to power this for the foreseeable future.
When robots drive cars, we will have more cars driving around
This doesn't seem true. In SF, waymo with 300 cars does more rides than lyft with 45k drivers. If self driving cars interleave different tasks based on their routes I imagine they would be much more efficient per mile.
With more than 700 vehicles in its fleet - 300 of which operate in San Francisco - Waymo is the only U.S. firm that runs uncrewed robotaxis that collect fares.
Those numbers are from April 2025.
We’ve also incrementally grown our commercial fleet as we’ve welcomed more riders, with over 1,500 vehicles across San Francisco, Los Angeles, Phoenix, and Austin.
https://waymo.com/blog/2025/05/scaling-our-fleet-through-us-...
This doesn't seem true.
Seems like we are way too early in the adoption curve to tell. Currently the average number of passengers per trip is >1.0 across the whole fleet. Some day, I'd expect that to dip below 1.0, as people send an empty car to pick up the dog from the vet, or circle the block to avoid having to pay for parking, etc.
A car driving from A to B will cost less than 50% of the current price. Which will unlock a huge amount of new rides.
The carbon intensity of electricity used by data centers was 48% higher than the US average.
If the people always talking about how cheap solar is want to fix this, find a way to make that cheapness actually make it into the customer's electric bill.
There are periodically news articles and such about data centers in Iceland, of course, but I get the impression it's mostly a fad, and the real build-outs are still in Northern Virginia as they've always been.
The typical answer I've seen is that Internet access and low latency matter more than cooling and power, but LLMs seem like they wouldn't care about that. I mean, you're literally interacting with them over text, and there's already plenty of latency - a few extra ms shouldn't matter?
I'd assume construction costs and costs of shipping in equipment also play a role, but Iceland and Canada aren't that far away.
I don't think there will be much carbon intensive energy creation in a few decades from now. It does not make sense economically.
Anyway I hope you're right, but so far global CO2 output is still growing. All the other energy has only come on top of carbon intensive energy, it hasn't replaced any of it. Every time we build more, we find new ways of spending that much energy and more.
I remember how me and my friends discovered email in 1999 and were like "Yay, in the future we'll all do this instead of sending letters!". And it took about 20 years until letters were largely replaced by email and the web. And when the first videos appeared on the web, it was quite clear to us that they would replace DVDs.
Similar with the advent of self driving cars and solar energy I think.
Not sure about the exact numbers, but I guess that at the moment normal roofs and solar panels absorb very roughly about the same percentage of sunlight.
So if in the future solar panels become more efficient, then yes, the amount of sunlight turned into heat could double.
Maybe that can be offset by covering other parts of earth with reflective materials or finding a way to send the heat back into the universe more effectively.
The discussion about nuclear vs solar remind me of the discussions about spinning HDs versus solid state drives when they were new.
GAFAM nuclear are mere announcements, intentions.
On the other front most already progress. https://www.reuters.com/sustainability/climate-energy/micros...
https://www.reuters.com/sustainability/climate-energy/micros...
energy usage will probably go up exponentially
kindof sounds like Jevons paradox? https://wiki.froth.zone/wiki/Jevons_paradox
The racks I am personally responsible for consume 17.2kW. That’s consistent across the year; sure things dip a bit when applications are shut down, but in general 17.2kW is the number. Presuming a national average of 1.2kW per home, each rack of equipment I oversee could potentially power 14 houses. I am responsible for hundreds of these racks, while my larger organization has many thousands of these racks in many locations worldwide.
I’ve found no other way to let the scale of this sink in. When put this way she is very clear: the price isn’t worth it to humanity. Being able to get, say, Door Dash, is pretty neat! But not at the cost of all our hoarded treasure and certainly not at the cost of the environment on the only planet we have access to.
The work done by AI will only ever benefit the people at the top. Because to be frank: they won’t share. Because the very wealthy have hoarding disorder.
Each!
We are in no meaningful sense torching the biosphere to get AI.
It seems like you are having an emotional response to not understanding the general energy picture.
This is condescending and rude. It also strikes me as obviously wrong. The 'response' isn't emotional in the slightest; it just explains the emotional and cognitive experience of acquiring understanding. It's a reasonable, well-reasoned explanation of the difficulty of intuitively grasping how much energy these data centers, and thus the companies that use them, consume -- and then of the shock many experience when it dawns on them.
For example, an A320 aloft uses the same energy as two thousand of your hypothetical racks (2.5 tons of kerosene per hour).Each!
We are in no meaningful sense torching the biosphere to get AI.
What exactly is the reasoning here? That airplanes use a lot of energy, just like data centers or compared to data centers, and therefore that AI isn't ecologically damaging -- or isn't "meaningfully" damaging, whatever that means? That's not just wrong. It's a nonsequitur.
There's a simpler, easier, better way to wrap our heads around the data, one that doesn't require false dichotomies (or something like that): humans are torching the biosphere both to get AI and to travel.
When you ask an AI model to write you a joke or generate a video of a puppy, that query comes with a small but measurable energy toll and an associated amount of emissions spewed into the atmosphere. Given that each individual request often uses less energy than running a kitchen appliance for a few moments, it may seem insignificant.But as more of us turn to AI tools, these impacts start to add up. And increasingly, you don’t need to go looking to use AI: It’s being integrated into every corner of our digital lives.
Forward looking, I imagine this will be the biggest factor in increasing energy demands for AI: companies shoving it into products that nobody wants or needs.
I think the bigger underrated concern is if LLMs fall into an unfortunate bucket where they are in fact generally useful, but not in ways that help us decarbonize our energy supply (or that do, but not enough to offset their own energy usage).
I’m sorry. I’m being blocked by some mysterious force from understanding what “actual human” means. And I don’t know how to get you in contact with your car manufacturer. Would you like me to repeat my 20 step suggestion on how to troubleshoot “why does my shitty car put the A/C on freezer mode whenever “Indian Summer” tops the charts in Bulgaria”, but with more festive emojis?
So yes, I'd like to disable this completely. Even if it's just a single birthday candle worth of energy usage.
On that note, what’s the energy footprint of the return to office initiatives that many companies have initiated?
a person who was doing tedium solved by AI and now can do something more productive that AI can’t necessarily do.
Like driving a uber or delivering food on a bicycle ? Amazing!
Of course that human will likely have another job, but what’s the math of a person who was doing tedium solved by AI and now can do something more productive that AI can’t necessarily do
That’s a lot of big assumptions - that the job getting replaced was tedious in the first place, that those other “more productive” job exists, that the thing AI can’t necessarily do will stay that way long enough for it not to be taken over by AI as well, that the tediousness was not part of the point (e.g. art)…
(cgroups, as per a sibbling comment, are addressed in this write-up as "not maximally satisfying")
I found this article to be a little too one sided. For instance, it didn’t talk about the 10x reductions in power achieved this past year — essentially how gpt4 can now run on a laptop.
Viz, via sama “The cost to use a given level of AI falls about 10x every 12 months, and lower prices lead to much more use. You can see this in the token cost from GPT-4 in early 2023 to GPT-4o in mid-2024, where the price per token dropped about 150x in that time period. Moore’s law changed the world at 2x every 18 months; this is unbelievably stronger.” https://blog.samaltman.com/three-observations
He’s one of the most knowledgeable persons in the world on the topic. It’s like saying that the ceo of BMW isn’t citable on conversations about expected cost decreases in electric cars.
Any environmental topic is susceptible to a huge amount of groupthink. The world isn’t so binary as people make it out to be. It is far from truth that LLMs=bad for environment, any more than computers=bad for the environment.
1/ did not exist before
2/ does not replace/reduce previous/other power (some, very much more critical and essential) usages.
3/ a LOT of tasks are still way more energy/time-efficiently done with regular existing methods (dedicated software, or even by hand), but are still asked/improperly routed to AI chatbots that ... statistically guess the answer.
It also leads to automation and efficiency, even if it isn’t a fully linear path.
AI isnt a waste. We can’t let environmental consciousness get in the way of rather natural human development. Especially CO2. (I have different opinions about biodiversity because of the irreversible loss. I also believe that we have the technology to pause and reverse climate change — but don’t pursue it because of degrowth ideologies)
In 2017, AI began to change everything. Data centers started getting built with energy-intensive hardware designed for AI, which led them to double their electricity consumption by 2023.
As we all know, the generative AI boom only really kicked into high gear in November 2022 with ChatGPT. That's five years of "AI" growth between 2017 and 2022 which presumably was mostly not generative AI.
"The carbon intensity of electricity used by data centers was 48% higher than the US average."
I'd be fine with as many data centers as they want if they stimulated production of clean energy to run them.
But that quote links to another article by the same author. Which says
"Notably, the sources for all this power are particularly “dirty.” Since so many data centers are located in coal-producing regions, like Virginia, the “carbon intensity” of the energy they use is 48% higher than the national average. The paper, which was published on arXiv and has not yet been peer-reviewed, found that 95% of data centers in the US are built in places with sources of electricity that are dirtier than the national average. "
Which in turn links to https://arxiv.org/abs/2411.09786
Which puts the bulk of that 48% higher claim on
"The average carbon intensity of the US data centers in our study (weighted by the energy they consumed) was 548 grams of CO2e per kilowatt hour (kWh), approximately 48% higher than the US national average of 369 gCO2e / kWh (26)."
Which points to https://ourworldindata.org/grapher/carbon-intensity-electric...
For the average of 369g/KWh. That's close enough to the figure in the table at https://www.epa.gov/system/files/documents/2024-01/egrid2022...
which shows 375g/KWh (after converting from lb/MWh)
But the table they compare against shows.
VA 576g/KWh
TX 509g/KWh
CA 374g/KWh
and the EPA table shows VA 268g/KWh
TX 372g/KWh
CA 207g/KWh
Which seem more likely to be true. The paper has California at only marginally better than the national average for renewables (Which I guess they needed to support their argument given the number of data centers there)I like arxiv, It's a great place to see new ideas, the fields I look at have things that I can test myself to see if the idea actually works. I would not recommend it as a source of truth. Peer review still has a place.
If they were gathering emissions data from states themselves, they should have caclulated the average from that data, not pulled the average from another potentially completely different measure. Then their conclusions would have been valid regardless what weird scaling factor they bought in to their state calculations. The numbers might have been wrong but the proportion would have been accurate, and it is the proportion that is being highlighted.
I am glad that something is making us upgrade the grid
A few big consumers in centralized locations isn't changing the grid as much as the energy transition from fuels to electricity is
If you don't want to go there, it doesn't really matter how much energy the human uses because the human will just use the same energy to do something else.
Human's got to exist and needs to work to eat. They don't really, necessarily, existentially need to be 10x productive with the help of AI.
But I'll be honest, that's not really a solid argument, because it could rapidly lead to the question of why they do this exact job in the first place, instead of e.g. farming or whatever else there might be that can be called a net positive for humanity without reservations.
Oil might be able to carry more heat but it's more expensive to use.
Oil immersion is something nerds like to think is amazing but it's just a pain in the ass for negligible benefits. Imagine the annoyance of doing maintenance.
https://dgtlinfra.com/data-center-water-usage/
https://datacenters.microsoft.com/sustainability/efficiency/
I expect rapid progress in both model efficiency and hardware specialization. Local inference on edge devices, using chips designed specifically for AI workloads, will drastically reduce energy consumption for the majority of tasks. This shift will free up large-scale compute resources to focus on truly complex scientific problems, which seems like a worthwhile goal to me.
I can imagine that doing some clever offloading to a normal programs and using the LLM as a sort of "fuzzy glue" for the rest could improve the efficiency on many common tasks.
...and many buildings use millions of gallons of water (often fresh, potable water) per day in their cooling operations.
This is outrageous. People still struggle to access fresh water (and power), but hey "sustainability is all to our company" is always promoted as if something nice is being done on from the behemoth's sides. BS. What a waste of resources.
I truly condemn all this. To this day I do still refuse to use any of this technology and hope that all this ends in the near future. It's madness. I see this as nothing more than next-gen restrictive lousy search engines, and as many have pointed out ads are going to roll soon. The more people adopt it the worse will be for everyone.
I always emphasize this: 10-15 years ago I could find everything through simple web searches. Everything. In many cases even landing on niche and unexpected but useful and interesting websites. Today that is a difficult/impossible task.
Perhaps there is still room for a well-done traditional search engine (haven't tried Kagi but people in general do say nice things about it) to surface and take the lead but I doubt it, when hype arrives especially in the tech industry people follow blindly. There are still flourishing "ai" startups and from night to day everyone has become a voice or expert on the subject. Again: BS.
Traditional web engines and searches were absolutely just fine and quite impressed with their outputs. I remember it. What the heck has happened?
I guess it becomes okay when the companies guzzling the energy are some of the biggest tech employers in the world, buttering your bread in some way.
VR is back in its niche.
3DTV .. I've not seen that marketed for quite a while now. Things that are fads will die out eventually.
Crypto meanwhile is in an odd space: everyone knows blockchains are too much of a hassle and that the volatility is too high, so they're using centralized exchanges and "stablecoins" (shadow dollars). There's still a huge amount of money there but not quite as much as the stadium ads / FTX peak.
I doubt this is going to change.
That said, the flip side of energy cost being not a big factor is that you could probably eat the increase of energy cost by a factor of say 2 and this could possibly enable installation of short term (say 12h) battery storage to enable you to use only intermittent clean energy AND drive 100% utilization.
you might think it’s like measuring a car’s fuel economy or a dishwasher’s energy rating: a knowable value with a shared methodology for calculating it. You’d be wrong.
But everyone knows fuel economy is everything but a knowable value. Everything from if it has rained in the past four hours to temperature to loading of the vehicle to the chemical composition of the fuel (HVO vs traditional), how worn are your tires? Are they installed the right way? Are your brakes lagging? The possibilities are endless. You could end up with twice the consumption.
By the way, copy-pasting from the website is terrible on desktop firefox, the site just lags every second, for a second.
The largest model we tested has 405 billion parameters, but others, such as DeepSeek, have gone much further, with over 600 billion parameters.
Very quickly skimming, I have some trouble taking this post seriously when it omits that the larger DeepSeek one is a mixture-of-experts that will only use 12.5% (iirc) of its components for each token.
The best summary of text energy use I've seen is this (seemingly more rigorous, although its estimates are consistent with the final numbers made by the present post): epoch.ai/gradient-updates/how-much-energy-does-chatgpt-use
Estimates for a given response widely for a "typical" query (0.3 Wh; 1080 joules) and a maximal-context query (40 Wh; 144k joules). Assuming most uses don't come close to maximizing the context, the energy use of text seems very small compared to the benefits. That being said, the energy use for video generation seems substantial
I would be interested in seeing the numbers corresponding to how LLMs are typically used for code generation
The new model uses [energy] equivalent to riding 38 miles on an e-bike... AI companies have defended these numbers saying that generative video has a smaller footprint than the film shoots and travel that go into typical video production. That claim is hard to test and doesn’t account for the surge in video generation that might follow if AI videos become cheap to produce.
"Hard to test", but very obviously true if you make any attempt at guessing based on making a few assumptions... like they seem comfortable doing for all the closed source models they don't have access to being run in conditions they're not testing for. Especially considering they're presenting their numbers as definitive, and then just a couple paragraphs down admit that, yeah, they're just guessing.
Regardless, I know for a fact that a typical commercial shoot uses way more energy than driving across the TMZ in an e-bike (considering they're definitely using cars to transport gear, which gives you less than 4 miles for the same energy).
Even were it matches sorta (the 400 feet e-bike thing) that only works out for me because I use an AMD card. An NVIDIA card can have several times the generation speed at the same power draw so it all falls down again.
And the parameters they tried to standardize their figures with (the 1024x1024 thing) is also a bit meh because the SAME amount of pixels in a different aspect ratio can have huge variations in gen speed and thus power usage. for instance for most illustrious type checkpoints the speed is about 60% higher at aspect ratios other than 1024x1024. Its all a bit of a mess.
This leaves even those whose job it is to predict energy demands forced to assemble a puzzle with countless missing pieces, making it nearly impossible to plan for AI’s future impact on energy grids and emissions. Worse, the deals that utility companies make with the data centers will likely transfer the costs of the AI revolution to the rest of us, in the form of higher electricity bills.
... So don't? Explicitly shift the cost to the customer.
If I want to hook up to the energy grid with 3-phase power, I pay the utility to do it.
If a business wants more power and it isn't available, then the business can pay for it.
Then only businesses that really need it will be willing to step up to the plate.
No amount of "accounting" or "energy needs prediction" will guard against regulatory capture.
There is a significant caveat to this math. These numbers cannot serve as a proxy for how much energy is required to power something like ChatGPT 4o.
Otherwise this is an excellent article critiquing the very real problem that is opacity of these companies regarding model sizes and deployments. Not having an honest accounting of computing deployed worldwide is a problem, and while it's true that we didn't really do this in the past (early versions of Google searches were undoubtedly inefficient!), it's not an excuse today.
I also wish this article talked about the compute trends. That is, compute per token is going significantly down, but that also means use of that compute can spread more. Where does that lead us?
It's from 2021 so won't cover the 2022-onwards generative AI boom.
From the Wikipedia summary it sounds like it's about machine learning algorithms like classification, AlphaGo and concerns about ethics of training and bias.
Training SOTA models will, like steel mills or other large industrial projects, require a lot of environmental footprint to produce. But my prediction is that over time the vast majority of use cases in the hands of users will be essentially run on device and be basically zero impact, both in monetary cost and environment.
Table A1 , PDF page 29:
Finding the balance between innovation and long-term responsibility feels more important than ever.
unprecedented and comprehensive look at how much energy the AI industry uses
Not sure about comprehensive claim here if end-to-end query chains were not considered.
For example the mobile wireless node (that're being used by the majority of the users) contribution to the energy consumption are totally ignored. The wireless power amplifier or PA for both sides of users and base-stations are notorious for their inefficiency being only less than than 50% in practice although in theory can be around 80%. Almost all of the current AI applications are cloud based not local-first thus the end users energy consumption and contribution are necessary.
Impressive how Big Tech refuses to share data with society for collective decisions.
I'd also recommend the Data Vampires podcast series:
https://techwontsave.us/episode/241_data_vampires_going_hype...
https://techwontsave.us/episode/243_data_vampires_opposing_d...
https://techwontsave.us/episode/245_data_vampires_sacrificin...
https://techwontsave.us/episode/247_data_vampires_fighting_f...