TY - JOUR
T1 - Deep Reinforcement Learning-Based Optimization for IRS-Assisted Cognitive Radio Systems
AU - Zhong, Canwei
AU - Cui, Miao
AU - Zhang, Guangchi
AU - Wu, Qingqing
AU - Guan, Xinrong
AU - Chu, Xiaoli
AU - Vincent Poor, H.
N1 - Funding Information:
The work of Qingqing Wu was supported by the FDCT under Grant 0119/2020/A3, SKL-IOTSC(UM)-2021-2023, and the GDST under Grant 2021A1515011900 and Grant 2020B1212030003. The work of Xinrong Guan was supported by the National Natural Science Foundation of China under Grant 62171461 and the Natural Science Foundation on Frontier Leading Technology Basic Research Project of Jiangsu under Grant BK20212001.
Publisher Copyright:
© 1972-2012 IEEE.
PY - 2022/6/1
Y1 - 2022/6/1
N2 - In this paper, we consider an intelligent reflecting surface (IRS)-assisted cognitive radio system and maximize the secondary user (SU) rate by jointly optimizing the transmit power of secondary transmitter (ST) and the IRS's reflect beamforming, subject to the constraints of the minimum required signal-to-interference-plus-noise ratio at the primary receiver, the ST's maximum transmit power, and the unit modulus of the IRS reflect beamforming vector. This joint optimization problem can be solved suboptimally by the non-convex optimization techniques, which however usually require complicated mathematical transformations and are computationally intensive. To address this challenge, we propose an algorithm based on the deep deterministic policy gradient (DDPG) method. To achieve a higher learning efficiency and a lower reward variance, we propose another algorithm based on the soft actor-critic (SAC) method. In these proposed algorithms, a reward impact adjustment approach is proposed to improve their learning efficiency and stability. Simulation results show that the two proposed algorithms can achieve comparable SU rate performance with much shorter running time, as compared to the existing non-convex optimization-based benchmark algorithm, and that the proposed SAC-based algorithm learns faster and achieves a higher average reward with lower variance, as compared to the proposed DDPG-based algorithm.
AB - In this paper, we consider an intelligent reflecting surface (IRS)-assisted cognitive radio system and maximize the secondary user (SU) rate by jointly optimizing the transmit power of secondary transmitter (ST) and the IRS's reflect beamforming, subject to the constraints of the minimum required signal-to-interference-plus-noise ratio at the primary receiver, the ST's maximum transmit power, and the unit modulus of the IRS reflect beamforming vector. This joint optimization problem can be solved suboptimally by the non-convex optimization techniques, which however usually require complicated mathematical transformations and are computationally intensive. To address this challenge, we propose an algorithm based on the deep deterministic policy gradient (DDPG) method. To achieve a higher learning efficiency and a lower reward variance, we propose another algorithm based on the soft actor-critic (SAC) method. In these proposed algorithms, a reward impact adjustment approach is proposed to improve their learning efficiency and stability. Simulation results show that the two proposed algorithms can achieve comparable SU rate performance with much shorter running time, as compared to the existing non-convex optimization-based benchmark algorithm, and that the proposed SAC-based algorithm learns faster and achieves a higher average reward with lower variance, as compared to the proposed DDPG-based algorithm.
KW - Cognitive radio
KW - Deep reinforcement learning
KW - Intelligent reflecting surface
KW - Reflect beamforming
KW - Transmit power control
UR - http://www.scopus.com/inward/record.url?scp=85129685697&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85129685697&partnerID=8YFLogxK
U2 - 10.1109/TCOMM.2022.3171837
DO - 10.1109/TCOMM.2022.3171837
M3 - Article
AN - SCOPUS:85129685697
SN - 1558-0857
VL - 70
SP - 3849
EP - 3864
JO - IEEE Transactions on Communications
JF - IEEE Transactions on Communications
IS - 6
ER -