Abstract

Proactive caching is an effective way to alleviate peak-hour traffic congestion by prefetching popular contents at the wireless network edge. To maximize the caching efficiency requires the knowledge of content popularity profile, which however is often unavailable in advance. In this paper, we first propose a new linear prediction model, named grouped linear model (GLM) to estimate the future content requests based on historical data. Unlike many existing works that assumed the static content popularity profile, our model can adapt to the temporal variation of the content popularity in practical systems due to the arrival of new contents and dynamics of user preference. Based on the predicted content requests, we then propose a reinforcement learning approach with model-free acceleration (RLMA) for online cache replacement by taking into account both the cache hits and replacement cost. This approach accelerates the learning process in non-stationary environment by generating imaginary samples for Q-value updates. Numerical results based on real-world traces show that the proposed prediction and learning based online caching policy outperform all considered existing schemes.

Comment: 6 pages, 4 figures, ICC 2018 worksho


Original document

The different versions of the original document can be found in:

http://dx.doi.org/10.1109/iccw.2018.8403712
http://ui.adsabs.harvard.edu/abs/2018arXiv180304675Z/abstract,
https://academic.microsoft.com/#/detail/2964150361
Back to Top

Document information

Published on 01/01/2018

Volume 2018, 2018
DOI: 10.1109/iccw.2018.8403712
Licence: CC BY-NC-SA license

Document Score

0

Views 0
Recommendations 0

Share this document

claim authorship

Are you one of the authors of this document?