Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Transforming Cooling Optimization for Green Data Center via Deep Reinforcement Learning (1709.05077v4)

Published 15 Sep 2017 in cs.AI and cs.SY

Abstract: Cooling system plays a critical role in a modern data center (DC). Developing an optimal control policy for DC cooling system is a challenging task. The prevailing approaches often rely on approximating system models that are built upon the knowledge of mechanical cooling, electrical and thermal management, which is difficult to design and may lead to sub-optimal or unstable performances. In this paper, we propose utilizing the large amount of monitoring data in DC to optimize the control policy. To do so, we cast the cooling control policy design into an energy cost minimization problem with temperature constraints, and tap it into the emerging deep reinforcement learning (DRL) framework. Specifically, we propose an end-to-end cooling control algorithm (CCA) that is based on the actor-critic framework and an off-policy offline version of the deep deterministic policy gradient (DDPG) algorithm. In the proposed CCA, an evaluation network is trained to predict an energy cost counter penalized by the cooling status of the DC room, and a policy network is trained to predict optimized control settings when gave the current load and weather information. The proposed algorithm is evaluated on the EnergyPlus simulation platform and on a real data trace collected from the National Super Computing Centre (NSCC) of Singapore. Our results show that the proposed CCA can achieve about 11% cooling cost saving on the simulation platform compared with a manually configured baseline control algorithm. In the trace-based study, we propose a de-underestimation validation mechanism as we cannot directly test the algorithm on a real DC. Even though with DUE the results are conservative, we can still achieve about 15% cooling energy saving on the NSCC data trace if we set the inlet temperature threshold at 26.6 degree Celsius.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Yuanlong Li (6 papers)
  2. Yonggang Wen (84 papers)
  3. Kyle Guan (8 papers)
  4. Dacheng Tao (829 papers)
Citations (163)

Summary

Transforming Cooling Optimization for Green Data Centers via Deep Reinforcement Learning

The paper authored by Yuanlong Li, Yonggang Wen, Kyle Guan, and Dacheng Tao offers a comprehensive approach to optimizing cooling systems in data centers (DCs) using deep reinforcement learning (DRL). DCs are crucial infrastructure components in modern technology ecosystems, yet their energy consumption, particularly for cooling, is significant. This research deploys an end-to-end cooling control algorithm (CCA) based on the actor-critic framework and the deep deterministic policy gradient (DDPG) to address inefficiencies in DC cooling mechanisms.

Research Objective and Methodology

The paper aims to develop an optimal cooling control policy that minimizes energy costs while adhering to specific temperature constraints within DC environments. Traditional methods depend heavily on modeling the system’s mechanical, electrical, and thermal principles, which can be inadequate and lead to suboptimal or unstable performances. In contrast, the proposed DRL-based solution leverages vast quantities of monitoring data to train a neural network that can directly predict control settings based on current load and weather information.

By integrating DRL into the cooling optimization process, the researchers devised a cooling control algorithm adaptable from DDPG and actor-critic architecture. Key to their approach is the formulation of the cooling control policy design as an energy cost minimization problem. The algorithm is trained offline using pre-collected data traces, enabling it to circumvent extensive computational demands typical of simulation-based methods.

Results and Evaluation

Assessments of the proposed algorithm involved simulations on the EnergyPlus platform and analysis of real data from the National Super Computing Centre (NSCC) of Singapore. The CCA demonstrated considerable efficiency, achieving approximately 11% cooling cost savings in simulation tests compared to manually configured baseline controls. Furthermore, a conservative validation approach in real-world trace analysis indicated potential cooling energy savings of about 15% with appropriate temperature threshold settings.

Significance and Implications

The implications of this research are profound, particularly with respect to minimizing manual interventions and the complexities of DC management. The application of DRL for cooling optimization represents an advance in DC operations, suggesting that AI-based methodologies can significantly enhance control efficiency without relying on detailed system models. The insights gained from this research can inform future developments within AI-focused DC management strategies.

Future Directions

While the method has demonstrated effectiveness in simulation and controlled paper trials, future research could focus on real-world applications across diverse DC settings and conditions. Enhancements to the DRL framework to accommodate dynamic systems with less predictable fluctuations in load and external temperature conditions would be beneficial.

Overall, this paper exemplifies the potential for integrating advanced AI techniques into conventional industry controls, highlighting DRL's applicability in energy management solutions for data centers. Such developments are pivotal as they provide pathways to achieve sustainable and economically viable operations within critical IT infrastructure.