TY - GEN
T1 - Experience-driven Networking
T2 - 2018 IEEE Conference on Computer Communications, INFOCOM 2018
AU - Xu, Zhiyuan
AU - Tang, Jian
AU - Meng, Jingsong
AU - Zhang, Weiyi
AU - Wang, Yanzhi
AU - Liu, Chi Harold
AU - Yang, Dejun
N1 - Funding Information:
Zhiyuan Xu, Jian Tang, Jingsong Meng and Yanzhi Wang are with Department of Electrical Engineering and Computer Science, Syracuse University, Syracuse, NY 13244, USA. Email: {zxu105, jtang02, jmeng02, ywang393}@syr.edu. Weiyi Zhang is with AT&T Labs Research, Middle-town, NJ 07748 USA. Chi Harold Liu is with Beijing Institute of Technology, Beijing, China, 100081. Dejun Yang is with Department of Electrical Engineering and Computer Science, Colorado School of Mines, Golden, CO 80401, USA. This research was supported in part by NSF grants 1704662, 1525920 and 1443966. The information reported here does not reflect the position or the policy of the federal government.
Publisher Copyright:
© 2018 IEEE.
PY - 2018/10/8
Y1 - 2018/10/8
N2 - Modern communication networks have become very complicated and highly dynamic, which makes them hard to model, predict and control. In this paper, we develop a novel experience-driven approach that can learn to well control a communication network from its own experience rather than an accurate mathematical model, just as a human learns a new skill (such as driving, swimming, etc). Specifically, we, for the first time, propose to leverage emerging Deep Reinforcement Learning (DRL) for enabling model-free control in communication networks; and present a novel and highly effective DRL-based control framework, DRL-TE, for a fundamental networking problem: Traffic Engineering (TE). The proposed framework maximizes a widely-used utility function by jointly learning network environment and its dynamics, and making decisions under the guidance of powerful Deep Neural Networks (DNNs). We propose two new techniques, TE-aware exploration and actor-critic-based prioritized experience replay, to optimize the general DRL framework particularly for TE. To validate and evaluate the proposed framework, we implemented it in ns-3, and tested it comprehensively with both representative and randomly generated network topologies. Extensive packet-level simulation results show that 1) compared to several widely-used baseline methods, DRL-TE significantly reduces end-to-end delay and consistently improves the network utility, while offering better or comparable throughput; 2) DRL-TE is robust to network changes; and 3) DRL-TE consistently outperforms a state-of-the-art DRL method (for continuous control), Deep Deterministic Policy Gradient (DDPG), which, however, does not offer satisfying performance.
AB - Modern communication networks have become very complicated and highly dynamic, which makes them hard to model, predict and control. In this paper, we develop a novel experience-driven approach that can learn to well control a communication network from its own experience rather than an accurate mathematical model, just as a human learns a new skill (such as driving, swimming, etc). Specifically, we, for the first time, propose to leverage emerging Deep Reinforcement Learning (DRL) for enabling model-free control in communication networks; and present a novel and highly effective DRL-based control framework, DRL-TE, for a fundamental networking problem: Traffic Engineering (TE). The proposed framework maximizes a widely-used utility function by jointly learning network environment and its dynamics, and making decisions under the guidance of powerful Deep Neural Networks (DNNs). We propose two new techniques, TE-aware exploration and actor-critic-based prioritized experience replay, to optimize the general DRL framework particularly for TE. To validate and evaluate the proposed framework, we implemented it in ns-3, and tested it comprehensively with both representative and randomly generated network topologies. Extensive packet-level simulation results show that 1) compared to several widely-used baseline methods, DRL-TE significantly reduces end-to-end delay and consistently improves the network utility, while offering better or comparable throughput; 2) DRL-TE is robust to network changes; and 3) DRL-TE consistently outperforms a state-of-the-art DRL method (for continuous control), Deep Deterministic Policy Gradient (DDPG), which, however, does not offer satisfying performance.
KW - Deep Reinforcement Learning
KW - Experience-driven Networking
KW - Traffic Engineering
UR - http://www.scopus.com/inward/record.url?scp=85055701483&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85055701483&partnerID=8YFLogxK
U2 - 10.1109/INFOCOM.2018.8485853
DO - 10.1109/INFOCOM.2018.8485853
M3 - Conference contribution
AN - SCOPUS:85055701483
T3 - Proceedings - IEEE INFOCOM
SP - 1871
EP - 1879
BT - INFOCOM 2018 - IEEE Conference on Computer Communications
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 15 April 2018 through 19 April 2018
ER -