Green scheduling for LLM workloads with model and data reuse across geo-distributed data centers✩
Hao Liu , Xiaonyu Hu , Ran Wang , Jie Hao , Qiang Wu , Hongke Zhang
›› 2026, Vol. 12 ›› Issue (2) : 236 -251.
The explosive proliferation of Large Language Models (LLMs) imposes significant energy and operational bur-dens on Geographically Distributed Data Centers (GDDCs), thereby demanding an efficient mechanism for LLMs task scheduling. While prior geo-distributed scheduling methods reduce cost and carbon emissions by exploiting regional heterogeneity, they largely overlook model and data reuse opportunities and the uncertainty of LLM execution times. In this paper, we introduce GCOS, to the best of our knowledge, the first green scheduling framework that incorporates a dual-cache system for both data and models, while jointly optimizing task assign-ment and cache migration. We firstly propose a dual-cache mechanism that decouples model and data caching to enable fine-grained reuse and minimize redundant transmissions. Subsequently, we propose the Multi-Agent Cache-aware Cooperative Scheduling (MACCS) algorithm, which leverages reinforcement learning to optimize task placement with a focus on minimizing both carbon emissions and cost. Additionally, we design a lightweight execution time predictor, DiPTree, to address the high variability in task execution times. Extensive experiments on real-world datasets demonstrate that GCOS reduces overall cost by up to 92.6 % and carbon emissions by 90.3 %, significantly outperforming existing baselines.
Large language model / Geographically distributed data center / Green communication / Task scheduling / Multi-agent reinforcement learning
| [1] |
|
| [2] |
|
| [3] |
|
| [4] |
|
| [5] |
|
| [6] |
Electricitymaps, Electricitymaps: get carbon electricity data for your research project. https://www.electricitymaps.com/research2025. (accessed 20 March 2025). |
| [7] |
|
| [8] |
|
| [9] |
|
| [10] |
|
| [11] |
|
| [12] |
|
| [13] |
|
| [14] |
|
| [15] |
|
| [16] |
|
| [17] |
|
| [18] |
|
| [19] |
|
| [20] |
|
| [21] |
|
| [22] |
|
| [23] |
|
| [24] |
|
| [25] |
|
| [26] |
|
| [27] |
|
| [28] |
|
| [29] |
|
| [30] |
|
| [31] |
|
| [32] |
|
| [33] |
|
| [34] |
|
| [35] |
|
| [36] |
|
| [37] |
|
| [38] |
|
| [39] |
|
| [40] |
|
| [41] |
|
| [42] |
|
| [43] |
|
| [44] |
|
| [45] |
|
| [46] |
|
| [47] |
|
| [48] |
|
| [49] |
|
| [50] |
Ollama, Ollama: GitHub Repository. 2025. (accessed: 05 February 2025). https://github.com/ollama/ollama/blob/main/README.md |
| [51] |
Zjh-819, LLMDataHub: Awesome Datasets for LLM Training. 2023. (accessed: 15 July 2025). https://github.com/Zjh-819/LLMDataHub |
| [52] |
|
| [53] |
|
| [54] |
|
/
| 〈 |
|
〉 |