It's unfortunate that no math is ever done in these stories.
If you take the "350,000" H100s that Facebook wants by EOY, each of those can do 700W, which gives you almost 250 MW for just the GPUs. That sounds like a lot, until you realize that a single large power plant is measured in Gigawatts. All of Google's data centers combined are O(10 GW) which are matched with renewable power offsets [1].
Importantly, the world installed >500 Gigawatts of renewable energy in 2023 [2], mostly driven by PV Solar in China. The amount of potential solar and wind and other renewable-ish (hydro) outstrips even a 10x'ing of a lot of these numbers. But even for a single site, dams like Three Gorges are >20 GW.
There are real efficiency and scale challenges in doing AI in a single, large site. But existing power generation systems deliver plenty of power.
The big point of course is that there is massive asymmetry between training and inference and that even inference at scale is going to require massive amounts of energy and that likely OpenAI's business model isn't viable at scale. It works right now because they have capital to burn but when the music stops it may well turn out that their model isn't sustainable at all.
Efficiency gains should come first, long before they start looking at alternative energy sources.
Facebook is not alone, and there is growth. Also cooling is to be taken account of.
And third: renewables need to be associated with its backup like hydro/step or ... batteries which cost a lot. Gas can't be taken in as it's not CO2-free. All that unless training and inference happen when there's the corresponding wind and sun shining. And I'm not seeing that happening right now.
Energy production capacity in the US is relatively flat for the last couple of decades. The renewable installation is offsetting the decommission of coal. The capacity installed in China is not really accessible to Open AI due to recent security competition (both export restrictions on AI and a desire to import less energy). The capital costs of power are also quite high so I think he is pretty accurate considering the expectations of a startup to hockey stick.
> This report also analyzes prospective generation capacity in four categories — under construction, permitted, application pending, and proposed. More than 466,000 MW of new generation capacity is under development in the United States — a 13% increase over 2022. Sixty-one percent of capacity most likely to come online, permitted plants and plants that are under construction, are in solar.
China's growth in power capacity is non-trivially due to increasing demand. If the US or Europe or wherever suddenly wanted to build XXX GW per year, they could (modulo bureaucracy, which is very real).
This is my concern with AI in general. Cost, both real and monetary. Right now Microsoft and VCs are dumping money into AI operation to help with growth and adoption. What happens when AI's business focus moves from cost to grow to cost to serve? Will all these business who integrated in AI suddenly be saddled with huge bills? What if your product depends on AI, and suddenly is not profitable to operate? Anecdotally I have already seen people pull back AI features, because it turned out to be too expensive to serve in the long run.
I already pay for a GPT subscription, and its reliability is one of the worst of any product I pay for. The novelty keeps me paying, but I can't imagine building a business on it.
Ah we've now reached the "We're ready to go except for the magical development in another domain that's already resisted decades of concerted research and is absolutely necessary for what we do to actually make sense" stage. Good to know.
Record time compared to the crypto space, "all we need is infinite data storage with instant recall. Then we can store everything on the chain, and verify its authenticity!"
The size some of the large language models have is just insane. To put it into context GPT3 is a Large Language Model (LLM) but GPT4 has like (handwaveing) 5+ times as many parameters. This means the energy cost is also at least that much larger for inference.
And if we look at training instead of inference it's a quite a bit more complicated (each iteration is more costly, more parameter can also require more iterations, but at the same time you might need to spend less time in an area of increasingly diminishing returns per iteration), but if we go with GPT3 vs. GPT4 the 5x+ increase of parameters lead to a 10x+ increase in training cost (of which a huge part is energy cost, through also amortized hardware cost; ~10M$ to >100M$).
Additionally there are various analysis steps you might do when creating new models which can be hugely (energy) costly.
And that is with GPT4 and with OpenAI any major increase in version seem to come with a major increase in parameter size so with that trend we are looking at energy bills in the (potential many) hundreds of million US dollar for training alone.
Another thing wrt. inference cost is that with my limited understanding currently the best way to reach AGI and also a lot of other tasks is to run multiple models in tandem. Through this model might be domain adoptions of the same model, so not twice the training cost.
---
Domain Adaption == take a trained model and then train it a bit more to "specialize" it on a specific task, potentially adding additional knowledge etc. While you can try to do so just with prompt engineering there a larger prompt comes at a larger cost and a higher chance for unexpected failure, so by in a certain way "burning in" some additional behavior, knowledge etc. you can get nice benefit.
He's probably trying to build out a vertically integrated AI conglomerate spanning power, chip fabs, software, etc. And he's great at marketing in the meantime
I think both the OpenAI and the Sam Altman brand sustained some damage. He may be great at marketing but now that the gun is on the table and the cloth is withdrawn we'll see how much success he has in the future with trying to wrap people around his fingers.
Well, if the AI is good enough then ask it to design one.
Joking aside: the options are to turn it off or to charge what it should cost and given that no matter what it costs and no matter what the source if it uses too much energy the better solution would be to make it use less energy rather than to look for a breakthrough energy source because that also reduces the climate impact.
Well actually: "In a twist, Microsoft is experimenting with generative artificial intelligence to see if AI could help streamline the [nuclear power] approval process, according to Microsoft executives." [0]
One should be suspicious of ulterior motives when the CEO of an AI company makes a claim like this.
On one hand, LLMs do require significant amounts of compute to train. But the other hand, if you amortize training costs across all user sessions, is it really that big a deal? And that’s not even factoring in Moore’s law and incremental improvements to model training efficiency.
yes, at least as long as you constantly develop new AI models
and you still need to run the models, and e.g. for GPT4 that is alone already non trivial (energy cost/compute wise)
through for small LLMs if they are not run too much it might be not that bad
---
Generally I would always look for ulterior motives for any "relevant" public statement Sam Altman makes. As history has shown there often seems to be some (through in that usage "ulterior" has a bit too much of a "bad"/"evil" undertone).
To cut it short he seem to be invested in some Nuclear Fusion company, which is one of the potential ways to "solve" that problem. Another potential way is to use smaller LLMs but smaller LLMs can also be potentially a way how OpenAI loses their dominant position, as there is a much smaller barrier for training them.
Data centers cost 2% of electricity (Statistics from a few years ago).
AI inference is so costly at scale, one can easily see data centers start using 4% of total electricity, and in the next decade 8%. That will start to have severe effects on the power grid, basically require planning many years in advance to setup new power plants and such.
"Moore's law and incremental improvements' are irrelevant in the face of scaling laws. Since we aren't at AGI yet, every extra bit of compute will be dumped back into scaling the models and improving performance.
Is this the new PR playbook? Claim to be building something big and dangerous and ready to ruin everything if everyone doesn't drop everything and pay attention to me?
The AI winter is coming, and it's going to be bad. So many promises have been made, to the extent that the average person thinks that AGI is just around the corner.
first evidence that OpenAI is seriously lowballing it's tech bills and unless there's serious GPU efficiency gains, they're doing ye olde uber dances trying to get their consumers locked in before the price hikes start.
People and governments really need to start asking if the results of the current generation of FAANG-size LLMs are worth the costs, in additional chips, power, and cooling water -- especially in locations that are constrained for any of those resources. Are we getting ten times better Google results for ten times more energy? I don't think so.
This is why we need to focus on smaller, more efficient models.
A lot of the 13 and 30bn parameter models are actually quite good, maybe not GPT4 but it can run on my M1 MBP without an issue, and a bit faster than GPT4 anyway.
I don't need a chat interface that is amazing at all things at once, I need it to be great at 1 subject at a time.
That's the Big Tech playbook: go towards technology so expensive to run that competition is impossible. Tech that cannot be replicated. And of course the main fuel is our work, our content.
> we need better ways to store energy from solar power.
Why? I understand that for a normal website you always want to use the closest server possible, but for queries that take 30 seconds on average anyway, why does it matter if you're adding an extra 250ms to get it from the opposite side of the world?
If AI can legitimately replace human labor, why would anyone care if it takes an extra few milliseconds?
What kind of 'AI' is he talking about? The AGI kind? The Superintelligence kind? The chatbot which can't easily be ridiculed using esoteric word incantations kind?
I want one which can instantly calculate numberwang[0] and get it right every time
We have companies like Google and Microsoft who have the ability to build or create power generation for things like this. I just think it's interesting that Microsoft could be energy self sufficient by building generation next to a datacenter.
Amazon is a huge investor in green energy, both solar and wind. Google also buys solar. Colocated energy is great in theory, but doesn't always scale. That is why most companies take an offset approach.
at least out west they build near power generation already (usually next to hydro dams). At least that is what I remember them doing, I could be a few years out of date.
Bitcoin is still wasting gigawatts every year, but later blockchains are literally a million times more efficient. If the current generations of AI turn out to be useful, maybe someone will figure it out an efficiency optimization.
> Energy from nuclear fusion is available today – you just need solar panels to capture it
Nitpick: stellar fusion is fundamentally different from thermonuclear fusion.
Proton-proton fusion is mediated by the weak force [1][2]. A lump of the Sun radiates less power than your body. It works because there is a lot of Sun. Thermonuclear fusion is far more energetic because it's mediated by the strong force.
Yes, and there's also a stealth way to capture it if your planet has any type of gaseous atmosphere: take advantage of air currents that arise from the repeated heating and cooling of your atmosphere using a wind turbine.
It's unfortunate that no math is ever done in these stories.
If you take the "350,000" H100s that Facebook wants by EOY, each of those can do 700W, which gives you almost 250 MW for just the GPUs. That sounds like a lot, until you realize that a single large power plant is measured in Gigawatts. All of Google's data centers combined are O(10 GW) which are matched with renewable power offsets [1].
Importantly, the world installed >500 Gigawatts of renewable energy in 2023 [2], mostly driven by PV Solar in China. The amount of potential solar and wind and other renewable-ish (hydro) outstrips even a 10x'ing of a lot of these numbers. But even for a single site, dams like Three Gorges are >20 GW.
There are real efficiency and scale challenges in doing AI in a single, large site. But existing power generation systems deliver plenty of power.
[1] https://www.gstatic.com/gumdrop/sustainability/google-2023-e...
[2] https://www.iea.org/reports/renewables-2023/executive-summar...
The big point of course is that there is massive asymmetry between training and inference and that even inference at scale is going to require massive amounts of energy and that likely OpenAI's business model isn't viable at scale. It works right now because they have capital to burn but when the music stops it may well turn out that their model isn't sustainable at all.
Efficiency gains should come first, long before they start looking at alternative energy sources.
Facebook is not alone, and there is growth. Also cooling is to be taken account of.
And third: renewables need to be associated with its backup like hydro/step or ... batteries which cost a lot. Gas can't be taken in as it's not CO2-free. All that unless training and inference happen when there's the corresponding wind and sun shining. And I'm not seeing that happening right now.
Definitely the right way to look at things.
> If you take the "350,000" H100s that Facebook wants by EOY, each of those can do 700W
Plus ~500W for cooling
Less than a 1:1 ratio here is an extremely efficient datacenter. The heat has to get out of the building, not just off the chip.
2 replies →
Energy production capacity in the US is relatively flat for the last couple of decades. The renewable installation is offsetting the decommission of coal. The capacity installed in China is not really accessible to Open AI due to recent security competition (both export restrictions on AI and a desire to import less energy). The capital costs of power are also quite high so I think he is pretty accurate considering the expectations of a startup to hockey stick.
https://www.publicpower.org/system/files/documents/Americas_... via https://www.publicpower.org/resource/americas-electricity-ge... disagrees with you even in the last decade. And the future seems mostly solar:
> This report also analyzes prospective generation capacity in four categories — under construction, permitted, application pending, and proposed. More than 466,000 MW of new generation capacity is under development in the United States — a 13% increase over 2022. Sixty-one percent of capacity most likely to come online, permitted plants and plants that are under construction, are in solar.
China's growth in power capacity is non-trivially due to increasing demand. If the US or Europe or wherever suddenly wanted to build XXX GW per year, they could (modulo bureaucracy, which is very real).
1 reply →
This is my concern with AI in general. Cost, both real and monetary. Right now Microsoft and VCs are dumping money into AI operation to help with growth and adoption. What happens when AI's business focus moves from cost to grow to cost to serve? Will all these business who integrated in AI suddenly be saddled with huge bills? What if your product depends on AI, and suddenly is not profitable to operate? Anecdotally I have already seen people pull back AI features, because it turned out to be too expensive to serve in the long run.
I already pay for a GPT subscription, and its reliability is one of the worst of any product I pay for. The novelty keeps me paying, but I can't imagine building a business on it.
They will charge the salary of the worker they are replacing with an EC2 style AI simulacrum.
Employers will save health, logistics, HR, etc.
Governments will have to pay for unemployment
Just the same as always - privatize the gains
In other words, if the people providing the AI demand the same money as the workers it replaces, it doesn't seem like society actually benefits.
25 replies →
Ah we've now reached the "We're ready to go except for the magical development in another domain that's already resisted decades of concerted research and is absolutely necessary for what we do to actually make sense" stage. Good to know.
Record time compared to the crypto space, "all we need is infinite data storage with instant recall. Then we can store everything on the chain, and verify its authenticity!"
And throw in infinite intra-node bandwidth while you're at it.
Is this because of actual energy needs? Or because Sam is trying to drive investments into the fusion energy company he's backing?
probably both
The size some of the large language models have is just insane. To put it into context GPT3 is a Large Language Model (LLM) but GPT4 has like (handwaveing) 5+ times as many parameters. This means the energy cost is also at least that much larger for inference.
And if we look at training instead of inference it's a quite a bit more complicated (each iteration is more costly, more parameter can also require more iterations, but at the same time you might need to spend less time in an area of increasingly diminishing returns per iteration), but if we go with GPT3 vs. GPT4 the 5x+ increase of parameters lead to a 10x+ increase in training cost (of which a huge part is energy cost, through also amortized hardware cost; ~10M$ to >100M$).
Additionally there are various analysis steps you might do when creating new models which can be hugely (energy) costly.
And that is with GPT4 and with OpenAI any major increase in version seem to come with a major increase in parameter size so with that trend we are looking at energy bills in the (potential many) hundreds of million US dollar for training alone.
Another thing wrt. inference cost is that with my limited understanding currently the best way to reach AGI and also a lot of other tasks is to run multiple models in tandem. Through this model might be domain adoptions of the same model, so not twice the training cost.
---
Domain Adaption == take a trained model and then train it a bit more to "specialize" it on a specific task, potentially adding additional knowledge etc. While you can try to do so just with prompt engineering there a larger prompt comes at a larger cost and a higher chance for unexpected failure, so by in a certain way "burning in" some additional behavior, knowledge etc. you can get nice benefit.
He's probably trying to build out a vertically integrated AI conglomerate spanning power, chip fabs, software, etc. And he's great at marketing in the meantime
I think both the OpenAI and the Sam Altman brand sustained some damage. He may be great at marketing but now that the gun is on the table and the cloth is withdrawn we'll see how much success he has in the future with trying to wrap people around his fingers.
Well, if the AI is good enough then ask it to design one.
Joking aside: the options are to turn it off or to charge what it should cost and given that no matter what it costs and no matter what the source if it uses too much energy the better solution would be to make it use less energy rather than to look for a breakthrough energy source because that also reduces the climate impact.
Well actually: "In a twist, Microsoft is experimenting with generative artificial intelligence to see if AI could help streamline the [nuclear power] approval process, according to Microsoft executives." [0]
[0] https://www.wsj.com/tech/ai/microsoft-targets-nuclear-to-pow...
That's edging closer and closer to Douglas Adams 'Reason' software application.
1 reply →
One should be suspicious of ulterior motives when the CEO of an AI company makes a claim like this.
On one hand, LLMs do require significant amounts of compute to train. But the other hand, if you amortize training costs across all user sessions, is it really that big a deal? And that’s not even factoring in Moore’s law and incremental improvements to model training efficiency.
The inference cost is also huge in energy terms, not just the training. Remember that Altman's company builds gigantic models.
> is it really that big a deal
yes, at least as long as you constantly develop new AI models
and you still need to run the models, and e.g. for GPT4 that is alone already non trivial (energy cost/compute wise)
through for small LLMs if they are not run too much it might be not that bad
---
Generally I would always look for ulterior motives for any "relevant" public statement Sam Altman makes. As history has shown there often seems to be some (through in that usage "ulterior" has a bit too much of a "bad"/"evil" undertone).
To cut it short he seem to be invested in some Nuclear Fusion company, which is one of the potential ways to "solve" that problem. Another potential way is to use smaller LLMs but smaller LLMs can also be potentially a way how OpenAI loses their dominant position, as there is a much smaller barrier for training them.
Data centers cost 2% of electricity (Statistics from a few years ago).
AI inference is so costly at scale, one can easily see data centers start using 4% of total electricity, and in the next decade 8%. That will start to have severe effects on the power grid, basically require planning many years in advance to setup new power plants and such.
"Moore's law and incremental improvements' are irrelevant in the face of scaling laws. Since we aren't at AGI yet, every extra bit of compute will be dumped back into scaling the models and improving performance.
Is this the new PR playbook? Claim to be building something big and dangerous and ready to ruin everything if everyone doesn't drop everything and pay attention to me?
The AI winter is coming, and it's going to be bad. So many promises have been made, to the extent that the average person thinks that AGI is just around the corner.
first evidence that OpenAI is seriously lowballing it's tech bills and unless there's serious GPU efficiency gains, they're doing ye olde uber dances trying to get their consumers locked in before the price hikes start.
Hence, using OpenAI/perplexity everyday, as I am afraid they might start increasing subscriptions after the VC funds dry out.
People and governments really need to start asking if the results of the current generation of FAANG-size LLMs are worth the costs, in additional chips, power, and cooling water -- especially in locations that are constrained for any of those resources. Are we getting ten times better Google results for ten times more energy? I don't think so.
That question is for the investors to worry about, as long as the public and government aren't the ones footing the bill for compute.
This is why we need to focus on smaller, more efficient models.
A lot of the 13 and 30bn parameter models are actually quite good, maybe not GPT4 but it can run on my M1 MBP without an issue, and a bit faster than GPT4 anyway.
I don't need a chat interface that is amazing at all things at once, I need it to be great at 1 subject at a time.
That's the Big Tech playbook: go towards technology so expensive to run that competition is impossible. Tech that cannot be replicated. And of course the main fuel is our work, our content.
> we need better ways to store energy from solar power.
Why? I understand that for a normal website you always want to use the closest server possible, but for queries that take 30 seconds on average anyway, why does it matter if you're adding an extra 250ms to get it from the opposite side of the world?
If AI can legitimately replace human labor, why would anyone care if it takes an extra few milliseconds?
What kind of 'AI' is he talking about? The AGI kind? The Superintelligence kind? The chatbot which can't easily be ridiculed using esoteric word incantations kind?
I want one which can instantly calculate numberwang[0] and get it right every time
[0] https://www.youtube.com/watch?v=6ofnDdi-vfw
We have companies like Google and Microsoft who have the ability to build or create power generation for things like this. I just think it's interesting that Microsoft could be energy self sufficient by building generation next to a datacenter.
Amazon is a huge investor in green energy, both solar and wind. Google also buys solar. Colocated energy is great in theory, but doesn't always scale. That is why most companies take an offset approach.
https://www.aboutamazon.com/news/sustainability/amazon-renew...
at least out west they build near power generation already (usually next to hydro dams). At least that is what I remember them doing, I could be a few years out of date.
Bitcoin is still wasting gigawatts every year, but later blockchains are literally a million times more efficient. If the current generations of AI turn out to be useful, maybe someone will figure it out an efficiency optimization.
More energy... or differently designed chips?
Maybe Apple will have some ideas with the power/performance ratio of their chips compared to what's used in graphic cards.
Expect price increases for OpenAI, then. Maybe OpenAI's business model of cheap AI for banal purposes won't work.
Foremost it will require a nvidia substitute using photonics for matrix multiplications
How's analogue computing progressing as a more energy-efficient alternative?
Not. Noise would kill it. Your answers would essentially be nonsense.
Isn't ML moving towards low-precision floating point anyways?
1 reply →
So AGI will invent fusion, but we need fusion for AGI. Chicken-egg :)
Remember that the chicken and egg problem was eventually solved ;) (Don’t you ask how)
If there is a first chicken, there has to have been an egg with it inside it?
This is blockchain all over again isn't it.
I have found some legit use cases for generative AI at least.
this is just not true.
- lots of hype
- technically complex
- use cases uncertain
- except for crime/spam/other unpleasantness
- Nvidia is smiling
- cooking the planet
seems like it to me!
1 reply →
"Make the pie higher" ~ Sam Altman
[flagged]
Energy from nuclear fusion is available today – you just need solar panels to capture it.
> Energy from nuclear fusion is available today – you just need solar panels to capture it
Nitpick: stellar fusion is fundamentally different from thermonuclear fusion.
Proton-proton fusion is mediated by the weak force [1][2]. A lump of the Sun radiates less power than your body. It works because there is a lot of Sun. Thermonuclear fusion is far more energetic because it's mediated by the strong force.
[1] https://en.wikipedia.org/wiki/Proton–proton_chain
[2] https://en.wikipedia.org/wiki/Beta_plus_decay
Yes, and there's also a stealth way to capture it if your planet has any type of gaseous atmosphere: take advantage of air currents that arise from the repeated heating and cooling of your atmosphere using a wind turbine.
Isn't storage still a problem? I suppose there's batteries and pumped-storage hydro.
2 replies →
There actually was some graphene news recently.
I saw a graphene food cooker concept at CES