We consider a mobile user accessing contents in a dynamic environment, where new contents are generated over time (by the user's contacts) and remain relevant to the user for random lifetimes. The user, equipped with a finite-capacity cache memory, randomly accesses the system and requests all the relevant contents at the time of access. The system incurs an energy cost associated with the number of contents downloaded and the channel quality at that time. Assuming causal knowledge of the channel quality, the content profile, and the user-access behavior, we model the proactive caching problem as a Markov decision process with the goal of minimizing the long-term average energy cost. We first prove the optimality of a threshold-based proactive caching scheme, which dynamically caches or removes appropriate contents from the memory, prior to being requested by the user, depending on the channel state. The optimal threshold values depend on the system state and hence are computationally intractable. Therefore, we propose parametric representations for the threshold values and use reinforcement-learning algorithms to find near-optimal parameterizations. We demonstrate through simulations that the proposed schemes significantly outperform classical reactive downloading and perform very close to a genie-aided lower bound.

A Reinforcement-Learning Approach to Proactive Caching in Wireless Networks / Somuyiwa, S. O.; Gyorgy, A.; Gunduz, D.. - In: IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS. - ISSN 0733-8716. - 36:6(2018), pp. 1331-1344. [10.1109/JSAC.2018.2844985]

A Reinforcement-Learning Approach to Proactive Caching in Wireless Networks

D. Gunduz
2018

Abstract

We consider a mobile user accessing contents in a dynamic environment, where new contents are generated over time (by the user's contacts) and remain relevant to the user for random lifetimes. The user, equipped with a finite-capacity cache memory, randomly accesses the system and requests all the relevant contents at the time of access. The system incurs an energy cost associated with the number of contents downloaded and the channel quality at that time. Assuming causal knowledge of the channel quality, the content profile, and the user-access behavior, we model the proactive caching problem as a Markov decision process with the goal of minimizing the long-term average energy cost. We first prove the optimality of a threshold-based proactive caching scheme, which dynamically caches or removes appropriate contents from the memory, prior to being requested by the user, depending on the channel state. The optimal threshold values depend on the system state and hence are computationally intractable. Therefore, we propose parametric representations for the threshold values and use reinforcement-learning algorithms to find near-optimal parameterizations. We demonstrate through simulations that the proposed schemes significantly outperform classical reactive downloading and perform very close to a genie-aided lower bound.
2018
36
6
1331
1344
A Reinforcement-Learning Approach to Proactive Caching in Wireless Networks / Somuyiwa, S. O.; Gyorgy, A.; Gunduz, D.. - In: IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS. - ISSN 0733-8716. - 36:6(2018), pp. 1331-1344. [10.1109/JSAC.2018.2844985]
Somuyiwa, S. O.; Gyorgy, A.; Gunduz, D.
File in questo prodotto:
Non ci sono file associati a questo prodotto.
Pubblicazioni consigliate

Licenza Creative Commons
I metadati presenti in IRIS UNIMORE sono rilasciati con licenza Creative Commons CC0 1.0 Universal, mentre i file delle pubblicazioni sono rilasciati con licenza Attribuzione 4.0 Internazionale (CC BY 4.0), salvo diversa indicazione.
In caso di violazione di copyright, contattare Supporto Iris

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11380/1202571
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 98
  • ???jsp.display-item.citation.isi??? 85
social impact