This is the repo for the paper "Will LLMs Scaling Hit the Wall? Breaking Barriers via Distributed Resources on Massive Edge Devices".
In this position paper, we challenge the current trajectory of LLM scaling and propose a paradigm shift towards distributed edge computing. We believe that the future of AI lies not in centralized data centers, but in the collective power of billions of edge devices. Our vision is to democratize AI development by:
- 🌐 Unlocking Untapped Resources: Leveraging the vast computational power (2,758+ EFLOPS) and data (182+ ZB by 2025) available on edge devices worldwide
- 🤝 Democratizing AI: Enabling anyone to participate in LLM training using everyday devices
- 🔒 Preserving Privacy: Keeping data on user devices while contributing to model training
- 🌱 Environmental Sustainability: Utilizing existing computing resources instead of building more data centers
We open-source our analysis and findings to foster collaboration and accelerate the development of distributed AI systems. Join us in breaking down the barriers to AI scaling!
- [03/11/2025] 📝 Our paper is now available on arXiv.
- [03/01/2025] 📝 Our paper is submitted to arXiv.
Neural scaling laws show that model performance improves with increased data and compute, but this trajectory faces critical challenges:
- Data Wall: High-quality public text data is becoming scarce, with potential exhaustion by 2028
- Compute Wall: AI training demands growing at 13.4× annually since 2022, dominated by tech giants
- Solution: We propose leveraging distributed edge devices, revealing that:
- Global edge data volume will reach 182 ZB by 2025
- Collective smartphone computing power exceeds 2,758 EFLOPS
- Just 60,723 edge devices could train a DeepSeek-v3 scale model in one week
|
IoT Data Contribution Analysis: Growing from 33.2% to 43.6% of Global Data Volume |
Data Growth Trends in Edge Devices and Smartphones |
- Global data volume projection: 182 ZB by 2025
- IoT device contribution: 13.6 ZB (2019) → 79.4 ZB (2025)
- Smartphone data growth: 5 EB (2018) → 8 EB (2028)
- 5-year accumulated smartphone data: ~33.1 EB (pre-2025)
|
Edge Computing Power Growth Trend |
Smartphone Computing Power Evolution: Reaching 2,758 EFLOPS by 2024 |
- Smartphone collective computing power:
- 2020: 817 EFLOPS
- 2024: 2,758 EFLOPS
- 5-year cumulative: 9,278 EFLOPS
- Performance comparison:
- Single flagship device: >2 TFLOPS
- 30 smartphones ≈ 1 H100 GPU (59.30 TFLOPS)
-
Small Language Models at Edges
- Deploying compact language models on edge devices
- Model compression, knowledge distillation, and quantization
- Reduces computational and memory requirements
- Maintains acceptable performance
-
Collaborative Inference
- Distributing inference across multiple devices
- Enables more complex models than possible on individual devices
- Maintains low latency and reduces bandwidth requirements
- No single device handles entire computational load
-
Collaborative Training
- Federated learning across distributed devices without requiring data to leave device
- Preserves privacy while leveraging collective computational power
- Reduces inter-node communication costs
- Novel approaches for varying computational capabilities
- Creates more inclusive environment for diverse participants
- Reduces dependence on major tech companies
- Significantly lowers barriers to AI development participation
- Enables smaller organizations, academic institutions, and individual developers
- Data remains on user devices, reducing privacy risks
- Gives users greater control over their data
- Addresses concerns with stringent global privacy regulations
- Utilizes idle computing capacity of existing devices
- Reduces energy consumption and need for dedicated data centers
- Leverages billions of devices already in operation
- Lowers carbon footprint associated with AI training infrastructure
Looking ahead, we anticipate:
- Continuous enhancement of edge device hardware capabilities
- More efficient distributed learning algorithms that minimize communication overhead
- Specialized small language model architectures optimized for edge deployment
- Advanced frameworks supporting secure, privacy-preserving collaborative learning
The distributed capacity of edge devices will foster a democratized AI ecosystem where developers worldwide can participate in training and applying large language models, addressing broader societal needs and unlocking new possibilities for AI innovation.
@misc{shen2025llmsscalinghitwall,
title={Will LLMs Scaling Hit the Wall? Breaking Barriers via Distributed Resources on Massive Edge Devices},
author={Shen, Tao and Zhu, Didi and Zhao, Ziyu and Wu, Chao and Wu, Fei},
year={2025},
eprint={2503.08223},
archivePrefix={arXiv},
primaryClass={cs.DC},
url={https://arxiv.org/abs/2503.08223},
}This project is licensed under the MIT License. See LICENSE file for details.
We thank our colleagues and the research community for their valuable feedback and support.