TY - GEN
T1 - A deep reinforcement learning-based framework for content caching
AU - Zhong, Chen
AU - Gursoy, M. Cenk
AU - Velipasalar, Senem
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2018/5/21
Y1 - 2018/5/21
N2 - Content caching at the edge nodes is a promising technique to reduce the data traffic in next-generation wireless networks. Inspired by the success of Deep Reinforcement Learning (DRL) in solving complicated control problems, this work presents a DRL-based framework with Wolpertinger architecture for content caching at the base station. The proposed framework is aimed at maximizing the long-term cache hit rate, and it requires no knowledge of the content popularity distribution. To evaluate the proposed framework, we compare the performance with other caching algorithms, including Least Recently Used (LRU), Least Frequently Used (LFU), and First-In First-Out (FIFO) caching strategies. Meanwhile, since the Wolpertinger architecture can effectively limit the action space size, we also compare the performance with Deep Q-Network to identify the impact of dropping a portion of the actions. Our results show that the proposed framework can achieve improved short-term cache hit rate and improved and stable long-term cache hit rate in comparison with LRU, LFU, and FIFO schemes. Additionally, the performance is shown to be competitive in comparison to Deep Q-learning, while the proposed framework can provide significant savings in runtime.
AB - Content caching at the edge nodes is a promising technique to reduce the data traffic in next-generation wireless networks. Inspired by the success of Deep Reinforcement Learning (DRL) in solving complicated control problems, this work presents a DRL-based framework with Wolpertinger architecture for content caching at the base station. The proposed framework is aimed at maximizing the long-term cache hit rate, and it requires no knowledge of the content popularity distribution. To evaluate the proposed framework, we compare the performance with other caching algorithms, including Least Recently Used (LRU), Least Frequently Used (LFU), and First-In First-Out (FIFO) caching strategies. Meanwhile, since the Wolpertinger architecture can effectively limit the action space size, we also compare the performance with Deep Q-Network to identify the impact of dropping a portion of the actions. Our results show that the proposed framework can achieve improved short-term cache hit rate and improved and stable long-term cache hit rate in comparison with LRU, LFU, and FIFO schemes. Additionally, the performance is shown to be competitive in comparison to Deep Q-learning, while the proposed framework can provide significant savings in runtime.
KW - Deep reinforcement learning
KW - Wolpertinger architecture
KW - content caching
UR - http://www.scopus.com/inward/record.url?scp=85048572871&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85048572871&partnerID=8YFLogxK
U2 - 10.1109/CISS.2018.8362276
DO - 10.1109/CISS.2018.8362276
M3 - Conference contribution
AN - SCOPUS:85048572871
T3 - 2018 52nd Annual Conference on Information Sciences and Systems, CISS 2018
SP - 1
EP - 6
BT - 2018 52nd Annual Conference on Information Sciences and Systems, CISS 2018
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 52nd Annual Conference on Information Sciences and Systems, CISS 2018
Y2 - 21 March 2018 through 23 March 2018
ER -