AI training cluster sizes increased by more than 20x since 2016
By Robi Rahman
From Google’s NASv3 RL network, trained on 800 GPUs in 2016, to Meta’s Llama 3.1 405B, using 16,384 H100 GPUs in 2024, the number of processors used increased by a factor of over 20. Gemini Ultra was trained with an even larger number of TPUs, but precise details were not reported.
Epoch's work is free to use, distribute, and reproduce provided the source and authors are credited under the Creative Commons BY
license.
Explore this data
Machine Learning Hardware
Key data on over 170 AI accelerators, such as graphics processing units (GPUs) and tensor processing units (TPUs).
Epoch AI’s work is free to use, distribute, and reproduce provided the source and authors are credited under the Creative Commons Attribution license.
Citation
Robi Rahman (2024), "AI training cluster sizes increased by more than 20x since 2016". Published online at epoch.ai. Retrieved from 'https://epoch.ai/data-insights/training-cluster-size' [online resource]. Accessed 2 Apr 2026.
BibTeX Citation
@misc{epoch2024trainingclustersize,
title={AI training cluster sizes increased by more than 20x since 2016},
author={Robi Rahman},
year={2024},
url={https://epoch.ai/data-insights/training-cluster-size},
note={Accessed: 2026-04-02}}
Feedback
Have a question? Noticed something wrong? Let us know.
AI training cluster sizes increased by more than 20x since 2016
Epoch AI is a research institute investigating key trends and questions that will shape the trajectory and governance of Artificial Intelligence.