epoch.ai
This means a huge amount of heat in a small space. So you can’t cool these chips with fans — you need liquid coolants to efficiently soak up the heat.
This means a huge amount of heat in a small space. So you can’t cool these chips with fans — you need liquid coolants to efficiently soak up the heat.
So we spent the last few months reading legal permits, staring at satellite images, and scouring news sources.
Here’s what you need to know. 🧵
So we spent the last few months reading legal permits, staring at satellite images, and scouring news sources.
Here’s what you need to know. 🧵
Some hyperscalers plan to do it in just 1-2 years from the start of construction.
If they succeed, we’ll see the first GW-scale data centers online in 2026, marking one of the fastest infrastructure build-outs in history. 🧵
Some hyperscalers plan to do it in just 1-2 years from the start of construction.
If they succeed, we’ll see the first GW-scale data centers online in 2026, marking one of the fastest infrastructure build-outs in history. 🧵
One way to read our new capability index is by plotting the benchmark performance you expect to see, for a range of ECI scores 🧵
One way to read our new capability index is by plotting the benchmark performance you expect to see, for a range of ECI scores 🧵
bsky.app/profile/epo...
bsky.app/profile/epo...
The most expensive data center we know about is Microsoft Fairwater, projected to exceed $100 billion in total capital cost upon completion.
The most expensive data center we know about is Microsoft Fairwater, projected to exceed $100 billion in total capital cost upon completion.
Even larger data centers are coming: Meta Hyperion and Microsoft Fairwater will each have 5M H100e when they reach full capacity in late 2027 to early 2028.
Even larger data centers are coming: Meta Hyperion and Microsoft Fairwater will each have 5M H100e when they reach full capacity in late 2027 to early 2028.
- Anthropic–Amazon New Carlisle (January)
- xAI Colossus 2 (February)
- Microsoft Fayetteville (March, borderline 1GW)
- Meta Prometheus (May)
- OpenAI Stargate Abilene (July)
- Anthropic–Amazon New Carlisle (January)
- xAI Colossus 2 (February)
- Microsoft Fayetteville (March, borderline 1GW)
- Meta Prometheus (May)
- OpenAI Stargate Abilene (July)
The world is about to see multiple 1 GW+ AI data centers.
We mapped their construction using satellite imagery, permits & public sources — releasing everything for free, including commissioned satellite images.
Highlights in thread!
The world is about to see multiple 1 GW+ AI data centers.
We mapped their construction using satellite imagery, permits & public sources — releasing everything for free, including commissioned satellite images.
Highlights in thread!
About 15% can be solved with only the terminal and a further 30% can rely heavily on Python scripts.
We even found cases of models downloading packages to manipulate spreadsheets.
About 15% can be solved with only the terminal and a further 30% can rely heavily on Python scripts.
We even found cases of models downloading packages to manipulate spreadsheets.
These tasks take humans only a few minutes to complete.
These tasks take humans only a few minutes to complete.
Models get an Ubuntu VM and task instructions, and write code to interact with the mouse and keyboard.
Here is one task's starting state. Instructions: "Make a duplicate of the last two slides for me, please."
Models get an Ubuntu VM and task instructions, and write code to interact with the mouse and keyboard.
Here is one task's starting state. Instructions: "Make a duplicate of the last two slides for me, please."
Our findings: tasks are simple, many don't require GUIs, and success often hinges on interpreting ambiguous instructions. The benchmark is also not stable over time.
See thread for details!
Our findings: tasks are simple, many don't require GUIs, and success often hinges on interpreting ambiguous instructions. The benchmark is also not stable over time.
See thread for details!
Corrected results: GPT-5 (high) scores slightly higher than GPT-5 (medium) on the benchmarks we run. They are also now tied on the Epoch Capabilities Index (ECI).
Corrected results: GPT-5 (high) scores slightly higher than GPT-5 (medium) on the benchmarks we run. They are also now tied on the Epoch Capabilities Index (ECI).
The result? This gap is smaller than previously estimated.
On average, it takes 3.5 months for an open-weight model to catch up with closed-source SOTA.
The result? This gap is smaller than previously estimated.
On average, it takes 3.5 months for an open-weight model to catch up with closed-source SOTA.
- more complex permitting processes
- additional engineering to manage a long-range network connection and reliability
- constraints on communication-heavy paradigms
- more complex permitting processes
- additional engineering to manage a long-range network connection and reliability
- constraints on communication-heavy paradigms
Our research suggests that conducting 10 GW training runs across two dozen sites—linked by a network spanning thousands of km long—is feasible.
Our research suggests that conducting 10 GW training runs across two dozen sites—linked by a network spanning thousands of km long—is feasible.
It currently covers models from 2023 on, and it allows us to track trends in capabilities as they emerge.
It currently covers models from 2023 on, and it allows us to track trends in capabilities as they emerge.
However, by combining scores from different benchmarks, we created a single scale that captures the full range of model performance over time.
However, by combining scores from different benchmarks, we created a single scale that captures the full range of model performance over time.