A Smart Cache Content Update Policy Based on Deep Reinforcement Learning

6Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This paper proposes a DRL-based cache content update policy in the cache-enabled network to improve the cache hit ratio and reduce the average latency. In contrast to the existing policies, a more practical cache scenario is considered in this work, in which the content requests vary by both time and location. Considering the constraint of the limited cache capacity, the dynamic content update problem is modeled as a Markov decision process (MDP). Besides that, the deep Q-learning network (DQN) algorithm is utilised to solve the MDP problem. Specifically, the neural network is optimised to approximate the Q value where the training data are chosen from the experience replay memory. The DQN agent derives the optimal policy for the cache decision. Compared with the existing policies, the simulation results show that our proposed policy is 56%-64% improved in terms of the cache hit ratio and 56%-59% decreased in terms of the average latency.

Cite

CITATION STYLE

APA

Li, L., Kwong, C. F., Liu, Q., & Wang, J. (2020). A Smart Cache Content Update Policy Based on Deep Reinforcement Learning. Wireless Communications and Mobile Computing, 2020. https://doi.org/10.1155/2020/8836592

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free