TY - GEN
T1 - Deep Reinforcement Learning for IoT Networks
T2 - 2020 IEEE Global Communications Conference, GLOBECOM 2020
AU - Wu, Xiongwei
AU - Li, Xiuhua
AU - Li, Jun
AU - Ching, P. C.
AU - Poor, H. Vincent
N1 - Funding Information:
This work was supported in part by the Global Scholarship Programme for Research Excellence from CUHK, and in part by the U.S. National Science Foundation under Grant CCF-1908308.
Publisher Copyright:
© 2020 IEEE.
PY - 2020/12
Y1 - 2020/12
N2 - In most Internet of Things (IoT) networks, edge nodes are commonly used as to relays to cache sensing data generated by IoT sensors as well as provide communication services for data consumers. However, a critical issue of IoT sensing is that data are usually transient, which necessitates temporal updates of caching content items while frequent cache updates could lead to considerable energy cost and challenge the lifetime of IoT sensors. To address this issue, we adopt the Age of Information (AoI) to quantity data freshness and propose an online cache update scheme to obtain an effective tradeoff between the average AoI and energy cost. Specifically, we first develop a characterization of transmission energy consumption at IoT sensors by incorporating a successful transmission condition. Then, we model cache updating as a Markov decision process to minimize average weighted cost with judicious definitions of state, action, and reward. Since user preference towards content items is usually unknown and often temporally evolving, we therefore develop a deep reinforcement learning (DRL) algorithm to enable intelligent cache updates. Through trial-and-error explorations, an effective caching policy can be learned without requiring exact knowledge of content popularity. Simulation results demonstrate the superiority of the proposed framework.
AB - In most Internet of Things (IoT) networks, edge nodes are commonly used as to relays to cache sensing data generated by IoT sensors as well as provide communication services for data consumers. However, a critical issue of IoT sensing is that data are usually transient, which necessitates temporal updates of caching content items while frequent cache updates could lead to considerable energy cost and challenge the lifetime of IoT sensors. To address this issue, we adopt the Age of Information (AoI) to quantity data freshness and propose an online cache update scheme to obtain an effective tradeoff between the average AoI and energy cost. Specifically, we first develop a characterization of transmission energy consumption at IoT sensors by incorporating a successful transmission condition. Then, we model cache updating as a Markov decision process to minimize average weighted cost with judicious definitions of state, action, and reward. Since user preference towards content items is usually unknown and often temporally evolving, we therefore develop a deep reinforcement learning (DRL) algorithm to enable intelligent cache updates. Through trial-and-error explorations, an effective caching policy can be learned without requiring exact knowledge of content popularity. Simulation results demonstrate the superiority of the proposed framework.
UR - http://www.scopus.com/inward/record.url?scp=85097827665&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85097827665&partnerID=8YFLogxK
U2 - 10.1109/GLOBECOM42002.2020.9322415
DO - 10.1109/GLOBECOM42002.2020.9322415
M3 - Conference contribution
AN - SCOPUS:85097827665
T3 - 2020 IEEE Global Communications Conference, GLOBECOM 2020 - Proceedings
BT - 2020 IEEE Global Communications Conference, GLOBECOM 2020 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 7 December 2020 through 11 December 2020
ER -