TY - JOUR
T1 - Hierarchical power management of a system with autonomously power-managed components using reinforcement learning
AU - Triki, M.
AU - Wang, Y.
AU - Ammari, A. C.
AU - Pedram, M.
N1 - Publisher Copyright:
© 2014 Elsevier B.V. All rights reserved.
PY - 2015
Y1 - 2015
N2 - This paper presents a hierarchical dynamic power management (DPM) framework based on reinforcement learning (RL) technique, which aims at power savings in a computer system with multiple I/O devices running a number of heterogeneous applications. The proposed framework interacts with the CPU scheduler to perform effective application-level scheduling, thereby enabling further power savings. Moreover, it considers non-stationary workloads and differentiates between the service request generation rates of various software application. The online adaptive DPM technique consists of two layers: component-level local power manager and system-level global power manager. The component-level PM policy is pre-specified and fixed whereas the system-level PM employs temporal difference learning on semi-Markov decision process as the model-free RL technique, and it is specifically optimized for a heterogeneous application pool. Experiments show that the proposed approach considerably enhances power savings while maintaining good performance levels. In comparison with other reference systems, the proposed RL-based DPM approach, further enhances power savings, performs well under various workloads, can simultaneously consider power and performance, and achieves wide and deep power-performance tradeoff curves. Experiments conducted with multiple service providers confirm that up to 63% maximum energy saving per service provider can be achieved.
AB - This paper presents a hierarchical dynamic power management (DPM) framework based on reinforcement learning (RL) technique, which aims at power savings in a computer system with multiple I/O devices running a number of heterogeneous applications. The proposed framework interacts with the CPU scheduler to perform effective application-level scheduling, thereby enabling further power savings. Moreover, it considers non-stationary workloads and differentiates between the service request generation rates of various software application. The online adaptive DPM technique consists of two layers: component-level local power manager and system-level global power manager. The component-level PM policy is pre-specified and fixed whereas the system-level PM employs temporal difference learning on semi-Markov decision process as the model-free RL technique, and it is specifically optimized for a heterogeneous application pool. Experiments show that the proposed approach considerably enhances power savings while maintaining good performance levels. In comparison with other reference systems, the proposed RL-based DPM approach, further enhances power savings, performs well under various workloads, can simultaneously consider power and performance, and achieves wide and deep power-performance tradeoff curves. Experiments conducted with multiple service providers confirm that up to 63% maximum energy saving per service provider can be achieved.
KW - Power management
KW - Reinforcement learning
KW - Semi-Markov decision process
KW - Temporal difference learning
UR - http://www.scopus.com/inward/record.url?scp=84922936466&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84922936466&partnerID=8YFLogxK
U2 - 10.1016/j.vlsi.2014.06.001
DO - 10.1016/j.vlsi.2014.06.001
M3 - Article
AN - SCOPUS:84922936466
SN - 0167-9260
VL - 48
SP - 10
EP - 20
JO - Integration, the VLSI Journal
JF - Integration, the VLSI Journal
IS - 1
ER -