TY - JOUR
T1 - Experience-Driven Congestion Control
T2 - When Multi-Path TCP Meets Deep Reinforcement Learning
AU - Xu, Zhiyuan
AU - Tang, Jian
AU - Yin, Chengxiang
AU - Wang, Yanzhi
AU - Xue, Guoliang
N1 - Funding Information:
Manuscript received July 21, 2018; revised December 20, 2018; accepted February 27, 2019. Date of publication March 11, 2019; date of current version May 15, 2019. This work was supported by NSF under Grant 1704662 and Grant 1704092. (Corresponding author: Jian Tang.) Z. Xu, J. Tang, and C. Yin are with the Department of Electrical Engineering and Computer Science, Syracuse University, Syracuse, NY 13244 USA (e-mail: zxu105@syr.edu; jtang02@syr.edu; cyin02@syr.edu).
Publisher Copyright:
© 1983-2012 IEEE.
PY - 2019/6
Y1 - 2019/6
N2 - In this paper, we aim to study networking problems from a whole new perspective by leveraging emerging deep learning, to develop an experience-driven approach, which enables a network or a protocol to learn the best way to control itself from its own experience (e.g., runtime statistics data), just as a human learns a skill. We present design, implementation and evaluation of a deep reinforcement learning (DRL)-based control framework, DRL-CC (DRL for Congestion Control), which realizes our experience-driven design philosophy on multi-path TCP (MPTCP) congestion control. DRL-CC utilizes a single (instead of multiple independent) agent to dynamically and jointly perform congestion control for all active MPTCP flows on an end host with the objective of maximizing the overall utility. The novelty of our design is to utilize a flexible recurrent neural network, LSTM, under a DRL framework for learning a representation for all active flows and dealing with their dynamics. Moreover, we, for the first time, integrate the above LSTM-based representation network into an actor-critic framework for continuous (congestion) control, which leverages the emerging deterministic policy gradient to train critic, actor, and LSTM networks in an end-to-end manner. We implemented DRL-CC based on the MPTCP implementation in the Linux kernel. The experimental results show that 1) DRL-CC consistently and significantly outperforms a few well-known MPTCP congestion control algorithms in terms of goodput without sacrificing fairness, 2) it is flexible and robust to highly-dynamic network environments with time-varying flows, and 3) it is friendly to regular TCP.
AB - In this paper, we aim to study networking problems from a whole new perspective by leveraging emerging deep learning, to develop an experience-driven approach, which enables a network or a protocol to learn the best way to control itself from its own experience (e.g., runtime statistics data), just as a human learns a skill. We present design, implementation and evaluation of a deep reinforcement learning (DRL)-based control framework, DRL-CC (DRL for Congestion Control), which realizes our experience-driven design philosophy on multi-path TCP (MPTCP) congestion control. DRL-CC utilizes a single (instead of multiple independent) agent to dynamically and jointly perform congestion control for all active MPTCP flows on an end host with the objective of maximizing the overall utility. The novelty of our design is to utilize a flexible recurrent neural network, LSTM, under a DRL framework for learning a representation for all active flows and dealing with their dynamics. Moreover, we, for the first time, integrate the above LSTM-based representation network into an actor-critic framework for continuous (congestion) control, which leverages the emerging deterministic policy gradient to train critic, actor, and LSTM networks in an end-to-end manner. We implemented DRL-CC based on the MPTCP implementation in the Linux kernel. The experimental results show that 1) DRL-CC consistently and significantly outperforms a few well-known MPTCP congestion control algorithms in terms of goodput without sacrificing fairness, 2) it is flexible and robust to highly-dynamic network environments with time-varying flows, and 3) it is friendly to regular TCP.
KW - AI
KW - TCP
KW - congestion control
KW - deep learning
KW - experience-driven control
KW - multi-path TCP
UR - http://www.scopus.com/inward/record.url?scp=85065867897&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85065867897&partnerID=8YFLogxK
U2 - 10.1109/JSAC.2019.2904358
DO - 10.1109/JSAC.2019.2904358
M3 - Article
AN - SCOPUS:85065867897
SN - 0733-8716
VL - 37
SP - 1325
EP - 1336
JO - IEEE Journal on Selected Areas in Communications
JF - IEEE Journal on Selected Areas in Communications
IS - 6
M1 - 8664598
ER -