File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Hybrid Reinforcement Learning for STAR-RISs: A Coupled Phase-Shift Model Based Beamformer

TitleHybrid Reinforcement Learning for STAR-RISs: A Coupled Phase-Shift Model Based Beamformer
Authors
KeywordsBeamforming
deep reinforcement learning (DRL)
reconfigurable intelligent surfaces (RISs)
simultaneous transmitting and reflecting reconfigurable intelligent surfaces (STAR-RISs)
Issue Date2022
Citation
IEEE Journal on Selected Areas in Communications, 2022, v. 40, n. 9, p. 2556-2569 How to Cite?
AbstractA simultaneous transmitting and reflecting reconfigurable intelligent surface (STAR-RIS) assisted multi-user downlink multiple-input single-output (MISO) communication system is investigated. In contrast to the existing ideal STAR-RIS model assuming an independent transmission and reflection phase-shift control, a practical coupled phase-shift model is considered. Then, a joint active and passive beamforming optimization problem is formulated for minimizing the long-term transmission power consumption, subject to the coupled phase-shift constraint and the minimum data rate constraint. Despite the coupled nature of the phase-shift model, the formulated problem is solved by invoking a hybrid continuous and discrete phase-shift control policy. Inspired by this observation, a pair of hybrid reinforcement learning (RL) algorithms, namely the hybrid deep deterministic policy gradient (hybrid DDPG) algorithm and the joint DDPG & deep-Q network (DDPG-DQN) based algorithm are proposed. The hybrid DDPG algorithm controls the associated high-dimensional continuous and discrete actions by relying on the hybrid action mapping. By contrast, the joint DDPG-DQN algorithm constructs two Markov decision processes (MDPs) relying on an inner and an outer environment, thereby amalgamating the two agents to accomplish a joint hybrid control. Simulation results demonstrate that the STAR-RIS has superiority over other conventional RISs in terms of its energy consumption. Furthermore, both the proposed algorithms outperform the baseline DDPG algorithm, and the joint DDPG-DQN algorithm achieves a superior performance, albeit at an increased computational complexity.
Persistent Identifierhttp://hdl.handle.net/10722/349770
ISSN
2023 Impact Factor: 13.8
2023 SCImago Journal Rankings: 8.707

 

DC FieldValueLanguage
dc.contributor.authorZhong, Ruikang-
dc.contributor.authorLiu, Yuanwei-
dc.contributor.authorMu, Xidong-
dc.contributor.authorChen, Yue-
dc.contributor.authorWang, Xianbin-
dc.contributor.authorHanzo, Lajos-
dc.date.accessioned2024-10-17T07:00:42Z-
dc.date.available2024-10-17T07:00:42Z-
dc.date.issued2022-
dc.identifier.citationIEEE Journal on Selected Areas in Communications, 2022, v. 40, n. 9, p. 2556-2569-
dc.identifier.issn0733-8716-
dc.identifier.urihttp://hdl.handle.net/10722/349770-
dc.description.abstractA simultaneous transmitting and reflecting reconfigurable intelligent surface (STAR-RIS) assisted multi-user downlink multiple-input single-output (MISO) communication system is investigated. In contrast to the existing ideal STAR-RIS model assuming an independent transmission and reflection phase-shift control, a practical coupled phase-shift model is considered. Then, a joint active and passive beamforming optimization problem is formulated for minimizing the long-term transmission power consumption, subject to the coupled phase-shift constraint and the minimum data rate constraint. Despite the coupled nature of the phase-shift model, the formulated problem is solved by invoking a hybrid continuous and discrete phase-shift control policy. Inspired by this observation, a pair of hybrid reinforcement learning (RL) algorithms, namely the hybrid deep deterministic policy gradient (hybrid DDPG) algorithm and the joint DDPG & deep-Q network (DDPG-DQN) based algorithm are proposed. The hybrid DDPG algorithm controls the associated high-dimensional continuous and discrete actions by relying on the hybrid action mapping. By contrast, the joint DDPG-DQN algorithm constructs two Markov decision processes (MDPs) relying on an inner and an outer environment, thereby amalgamating the two agents to accomplish a joint hybrid control. Simulation results demonstrate that the STAR-RIS has superiority over other conventional RISs in terms of its energy consumption. Furthermore, both the proposed algorithms outperform the baseline DDPG algorithm, and the joint DDPG-DQN algorithm achieves a superior performance, albeit at an increased computational complexity.-
dc.languageeng-
dc.relation.ispartofIEEE Journal on Selected Areas in Communications-
dc.subjectBeamforming-
dc.subjectdeep reinforcement learning (DRL)-
dc.subjectreconfigurable intelligent surfaces (RISs)-
dc.subjectsimultaneous transmitting and reflecting reconfigurable intelligent surfaces (STAR-RISs)-
dc.titleHybrid Reinforcement Learning for STAR-RISs: A Coupled Phase-Shift Model Based Beamformer-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/JSAC.2022.3192053-
dc.identifier.scopuseid_2-s2.0-85135748311-
dc.identifier.volume40-
dc.identifier.issue9-
dc.identifier.spage2556-
dc.identifier.epage2569-
dc.identifier.eissn1558-0008-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats