Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Adaptive Control of Data Center Cooling using Deep Reinforcement Learning
Lund University, Sweden.
RISE Research Institutes of Sweden, Digital Systems, Data Science.
RISE Research Institutes of Sweden, Digital Systems, Data Science.ORCID iD: 0000-0003-4293-6408
RISE Research Institutes of Sweden, Digital Systems, Data Science.ORCID iD: 0000-0002-9759-5594
Show others and affiliations
2022 (English)In: Proceedings - 2022 IEEE International Conference on Autonomic Computing and Self-Organizing Systems Companion, ACSOS-C 2022, Institute of Electrical and Electronics Engineers Inc. , 2022Conference paper, Published paper (Refereed)
Abstract [en]

In this paper, we explore the use of Reinforcement Learning (RL) to improve the control of cooling equipment in Data Centers (DCs). DCs are inherently complex systems, and thus challenging to model from first principles. Machine learning offers a way to address this by instead training a model to capture the thermal dynamics of a DC. In RL, an agent learns to control a system through trial-and-error. However, for systems such as DCs, an interactive trial-and-error approach is not possible, and instead, a high-fidelity model is needed. In this paper, we develop a DC model using Computational Fluid Dynamics (CFD) based on the Lattice Boltzmann Method (LBM) Bhatnagar-Gross-Krook (BGK) algorithm. The model features transient boundary conditions for simulating the DC room, heat-generating servers, and Computer Room Air Handlers (CRAHs) as well as rejection components outside the server room such as heat exchangers, compressors, and dry coolers. This model is used to train an RL agent to control the cooling equipment. Evaluations show that the RL agent can outperform traditional controllers and also can adapt to changes in the environment, such as equipment breaking down. 

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers Inc. , 2022.
Keywords [en]
Adaptive Control, CFD Modeling, Data-center Cooling, Reinforcement Learning, Adaptive control systems, Computational fluid dynamics, Controllers, Cooling, Deep learning, Learning systems, Computational fluid dynamics modeling, Cooling equipment, Data center cooling, Datacenter, First principles, Machine-learning, Reinforcement learning agent, Reinforcement learnings, Thermal dynamics
National Category
Control Engineering
Identifiers
URN: urn:nbn:se:ri:diva-61598DOI: 10.1109/ACSOSC56246.2022.00018Scopus ID: 2-s2.0-85143075382ISBN: 9781665471374 (electronic)OAI: oai:DiVA.org:ri-61598DiVA, id: diva2:1721193
Conference
3rd IEEE International Conference on Autonomic Computing and Self-Organizing Systems Companion, ACSOS-C 2022, 19 September 2022 through 23 September 2022
Note

Funding details: VINNOVA, ITEA3-17002; Funding text 1: This work was supported by Vinnova grant ITEA3-17002 (AutoDC).

Available from: 2022-12-21 Created: 2022-12-21 Last updated: 2023-06-07Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Brännvall, RickardGustafsson, Jonas

Search in DiVA

By author/editor
Brännvall, RickardGustafsson, Jonas
By organisation
Data Science
Control Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 33 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf