File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: State-Aware Perturbation Optimization for Robust Deep Reinforcement Learning

TitleState-Aware Perturbation Optimization for Robust Deep Reinforcement Learning
Authors
KeywordsAdversarial Attack
Deep Reinforcement Learning
Markov Decision Process
Robotic Manipulation
Issue Date21-Aug-2025
PublisherInstitute of Electrical and Electronics Engineers
Citation
IEEE Transactions on Mobile Computing, 2025 How to Cite?
AbstractRecently, deep reinforcement learning (DRL) has emerged as a promising approach for robotic control. However, the deployment of DRL in real-world robots is hindered by its sensitivity to environmental perturbations. While existing white-box adversarial attacks rely on local gradient information and apply uniform perturbations across all states to evaluate DRL robustness, they fail to account for temporal dynamics and state-specific vulnerabilities. To combat the above challenge, we first conduct a theoretical analysis of white-box attacks in DRL by establishing the adversarial victim-dynamics Markov decision process (AVD-MDP), to derive the necessary and sufficient conditions for a successful attack. Based on this, we propose a selective state-aware reinforcement adversarial attack method, named STAR, to optimize perturbation stealthiness and state visitation dispersion. STAR first employs a soft mask-based state-targeting mechanism to minimize redundant perturbations, enhancing stealthiness and attack effectiveness. Then, it incorporates an information-theoretic optimization objective to maximize mutual information between perturbations, environmental states, and victim actions, ensuring a dispersed state-visitation distribution that steers the victim agent into vulnerable states for maximum return reduction. Extensive experiments demonstrate that STAR outperforms state-of-the-art benchmarks.
Persistent Identifierhttp://hdl.handle.net/10722/366578
ISSN
2023 Impact Factor: 7.7
2023 SCImago Journal Rankings: 2.755

 

DC FieldValueLanguage
dc.contributor.authorZhang, Zongyuan-
dc.contributor.authorDuan, Tianyang-
dc.contributor.authorLin, Zheng-
dc.contributor.authorHuang, Dong-
dc.contributor.authorFang, Zihan-
dc.contributor.authorSun, Zekai-
dc.contributor.authorXiong, Ling-
dc.contributor.authorLiang, Hongbin-
dc.contributor.authorCui, Heming-
dc.contributor.authorCui, Yong -
dc.date.accessioned2025-11-25T04:20:14Z-
dc.date.available2025-11-25T04:20:14Z-
dc.date.issued2025-08-21-
dc.identifier.citationIEEE Transactions on Mobile Computing, 2025-
dc.identifier.issn1536-1233-
dc.identifier.urihttp://hdl.handle.net/10722/366578-
dc.description.abstractRecently, deep reinforcement learning (DRL) has emerged as a promising approach for robotic control. However, the deployment of DRL in real-world robots is hindered by its sensitivity to environmental perturbations. While existing white-box adversarial attacks rely on local gradient information and apply uniform perturbations across all states to evaluate DRL robustness, they fail to account for temporal dynamics and state-specific vulnerabilities. To combat the above challenge, we first conduct a theoretical analysis of white-box attacks in DRL by establishing the adversarial victim-dynamics Markov decision process (AVD-MDP), to derive the necessary and sufficient conditions for a successful attack. Based on this, we propose a selective state-aware reinforcement adversarial attack method, named STAR, to optimize perturbation stealthiness and state visitation dispersion. STAR first employs a soft mask-based state-targeting mechanism to minimize redundant perturbations, enhancing stealthiness and attack effectiveness. Then, it incorporates an information-theoretic optimization objective to maximize mutual information between perturbations, environmental states, and victim actions, ensuring a dispersed state-visitation distribution that steers the victim agent into vulnerable states for maximum return reduction. Extensive experiments demonstrate that STAR outperforms state-of-the-art benchmarks.-
dc.languageeng-
dc.publisherInstitute of Electrical and Electronics Engineers-
dc.relation.ispartofIEEE Transactions on Mobile Computing-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectAdversarial Attack-
dc.subjectDeep Reinforcement Learning-
dc.subjectMarkov Decision Process-
dc.subjectRobotic Manipulation-
dc.titleState-Aware Perturbation Optimization for Robust Deep Reinforcement Learning-
dc.typeArticle-
dc.identifier.doi10.1109/TMC.2025.3601531-
dc.identifier.scopuseid_2-s2.0-105013840603-
dc.identifier.eissn1558-0660-
dc.identifier.issnl1536-1233-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats