Scheduling the transmission of time-sensitive data to multiple users over error-prone communication channels is studied with the goal of minimizing the longterm average age of information (AoI) at the users under a constraint on the average number of transmissions. The source can transmit only to a single user at each time slot, and after each transmission, it receives an instantaneous ACK/NACK feedback from the intended receiver, and decides on when and to which user to transmit the next update. The optimal scheduling policy is first studied under different feedback mechanisms when the channel statistics are known; in particular, the standard automatic repeat request (ARQ) and hybrid ARQ (HARQ) protocols are considered. Then a reinforcement learning (RL) approach is introduced, which does not assume any a priori information on the random processes governing the channel states. Different RL methods are applied and compared through numerical simulations.
A Reinforcement Learning Approach to Age of Information in Multi-User Networks / Ceran, E. T.; Gunduz, D.; Gyorgy, A.. - 2018-:(2018), pp. 1967-1971. (Intervento presentato al convegno 29th IEEE Annual International Symposium on Personal, Indoor and Mobile Radio Communications, PIMRC 2018 tenutosi a ita nel 2018) [10.1109/PIMRC.2018.8580701].
A Reinforcement Learning Approach to Age of Information in Multi-User Networks
D. Gunduz;
2018
Abstract
Scheduling the transmission of time-sensitive data to multiple users over error-prone communication channels is studied with the goal of minimizing the longterm average age of information (AoI) at the users under a constraint on the average number of transmissions. The source can transmit only to a single user at each time slot, and after each transmission, it receives an instantaneous ACK/NACK feedback from the intended receiver, and decides on when and to which user to transmit the next update. The optimal scheduling policy is first studied under different feedback mechanisms when the channel statistics are known; in particular, the standard automatic repeat request (ARQ) and hybrid ARQ (HARQ) protocols are considered. Then a reinforcement learning (RL) approach is introduced, which does not assume any a priori information on the random processes governing the channel states. Different RL methods are applied and compared through numerical simulations.Pubblicazioni consigliate
I metadati presenti in IRIS UNIMORE sono rilasciati con licenza Creative Commons CC0 1.0 Universal, mentre i file delle pubblicazioni sono rilasciati con licenza Attribuzione 4.0 Internazionale (CC BY 4.0), salvo diversa indicazione.
In caso di violazione di copyright, contattare Supporto Iris