epoch.ai
The post was written by @ansonwhho, @benmcottier, and @YafahEdelman.
The post was written by @ansonwhho, @benmcottier, and @YafahEdelman.
So unlike the Manhattan Project, AI data center buildouts are hard to keep secret from the rest of the world.
epoch.ai/data/data-c...
So unlike the Manhattan Project, AI data center buildouts are hard to keep secret from the rest of the world.
epoch.ai/data/data-c...
So companies probably won’t need to decentralize AI training over the next 2 years.
However, in practice they might choose to anyway, e.g. to soak up excess power in the grid.
So companies probably won’t need to decentralize AI training over the next 2 years.
However, in practice they might choose to anyway, e.g. to soak up excess power in the grid.
1) AI’s climate impact has been small so far.
For example, AI data centers use 1% of US power vs 8% for lighting and air conditioning (12%).
But if trends continue, this could change in the next decade.
1) AI’s climate impact has been small so far.
For example, AI data centers use 1% of US power vs 8% for lighting and air conditioning (12%).
But if trends continue, this could change in the next decade.
This means a huge amount of heat in a small space. So you can’t cool these chips with fans — you need liquid coolants to efficiently soak up the heat.
This means a huge amount of heat in a small space. So you can’t cool these chips with fans — you need liquid coolants to efficiently soak up the heat.
Usually a mix of on-site fossil fuel generation and interconnection to the grid.
E.g. Stargate Abilene will start off with on-site natural gas, then connect to the grid to access Texas’ abundant renewable power.
Usually a mix of on-site fossil fuel generation and interconnection to the grid.
E.g. Stargate Abilene will start off with on-site natural gas, then connect to the grid to access Texas’ abundant renewable power.
Other factors like latency matter surprisingly little — it takes >100× longer to generate model responses than transmit data from Texas to Tokyo.
Even serving LLMs from the Moon may not be a big latency issue!
Other factors like latency matter surprisingly little — it takes >100× longer to generate model responses than transmit data from Texas to Tokyo.
Even serving LLMs from the Moon may not be a big latency issue!
E.g. 30 GW is ~5% of the US’ power, ~2.5% of China’s, but ~90% of the UK’s
Other countries can build some frontier data centers and grow their power capacity — but they need more time/money
E.g. 30 GW is ~5% of the US’ power, ~2.5% of China’s, but ~90% of the UK’s
Other countries can build some frontier data centers and grow their power capacity — but they need more time/money
That’s bigger than the Apollo Program (0.8%) and Manhattan Project (0.4%) at their peaks.
That’s bigger than the Apollo Program (0.8%) and Manhattan Project (0.4%) at their peaks.
e.g. OpenAI’s Stargate Abilene will need:
- As much power as Seattle (1 GW)
- >250× the compute of the GPT-4 cluster
- 450 soccer fields of land
- $32B
- Thousands of workers
- 2 years to build
e.g. OpenAI’s Stargate Abilene will need:
- As much power as Seattle (1 GW)
- >250× the compute of the GPT-4 cluster
- 450 soccer fields of land
- $32B
- Thousands of workers
- 2 years to build
You can read more details on their methodology at the following link:
epoch.ai/data-insigh...
You can read more details on their methodology at the following link:
epoch.ai/data-insigh...
How? By reusing existing industrial shells and generating its own power early using gas turbines and batteries, before full grid connection.
How? By reusing existing industrial shells and generating its own power early using gas turbines and batteries, before full grid connection.
No model has scored more than 29% on FrontierMath, so saturating it definitely requires more than the current SOTA of 150… but will it take an ECI of 175? 200? For now, it’s hard to be sure.
No model has scored more than 29% on FrontierMath, so saturating it definitely requires more than the current SOTA of 150… but will it take an ECI of 175? 200? For now, it’s hard to be sure.
For instance, GPT-5 underperforms in GPQA Diamond but overperforms in VPCT.
For instance, GPT-5 underperforms in GPQA Diamond but overperforms in VPCT.
1. Benchmarks vary in overall difficulty, and in slope. Steeper slopes imply a narrower range of difficulties at the question level and mean the benchmark saturates quickly once some progress is made.
1. Benchmarks vary in overall difficulty, and in slope. Steeper slopes imply a narrower range of difficulties at the question level and mean the benchmark saturates quickly once some progress is made.
You can find the article here: www.theinformation.com/articles/ant...?
You can find the article here: www.theinformation.com/articles/ant...?