File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/TMC.2025.3601531
- Scopus: eid_2-s2.0-105013840603
- Find via

Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Article: State-Aware Perturbation Optimization for Robust Deep Reinforcement Learning
| Title | State-Aware Perturbation Optimization for Robust Deep Reinforcement Learning |
|---|---|
| Authors | |
| Keywords | Adversarial Attack Deep Reinforcement Learning Markov Decision Process Robotic Manipulation |
| Issue Date | 21-Aug-2025 |
| Publisher | Institute of Electrical and Electronics Engineers |
| Citation | IEEE Transactions on Mobile Computing, 2025 How to Cite? |
| Abstract | Recently, deep reinforcement learning (DRL) has emerged as a promising approach for robotic control. However, the deployment of DRL in real-world robots is hindered by its sensitivity to environmental perturbations. While existing white-box adversarial attacks rely on local gradient information and apply uniform perturbations across all states to evaluate DRL robustness, they fail to account for temporal dynamics and state-specific vulnerabilities. To combat the above challenge, we first conduct a theoretical analysis of white-box attacks in DRL by establishing the adversarial victim-dynamics Markov decision process (AVD-MDP), to derive the necessary and sufficient conditions for a successful attack. Based on this, we propose a selective state-aware reinforcement adversarial attack method, named STAR, to optimize perturbation stealthiness and state visitation dispersion. STAR first employs a soft mask-based state-targeting mechanism to minimize redundant perturbations, enhancing stealthiness and attack effectiveness. Then, it incorporates an information-theoretic optimization objective to maximize mutual information between perturbations, environmental states, and victim actions, ensuring a dispersed state-visitation distribution that steers the victim agent into vulnerable states for maximum return reduction. Extensive experiments demonstrate that STAR outperforms state-of-the-art benchmarks. |
| Persistent Identifier | http://hdl.handle.net/10722/366578 |
| ISSN | 2023 Impact Factor: 7.7 2023 SCImago Journal Rankings: 2.755 |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Zhang, Zongyuan | - |
| dc.contributor.author | Duan, Tianyang | - |
| dc.contributor.author | Lin, Zheng | - |
| dc.contributor.author | Huang, Dong | - |
| dc.contributor.author | Fang, Zihan | - |
| dc.contributor.author | Sun, Zekai | - |
| dc.contributor.author | Xiong, Ling | - |
| dc.contributor.author | Liang, Hongbin | - |
| dc.contributor.author | Cui, Heming | - |
| dc.contributor.author | Cui, Yong | - |
| dc.date.accessioned | 2025-11-25T04:20:14Z | - |
| dc.date.available | 2025-11-25T04:20:14Z | - |
| dc.date.issued | 2025-08-21 | - |
| dc.identifier.citation | IEEE Transactions on Mobile Computing, 2025 | - |
| dc.identifier.issn | 1536-1233 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/366578 | - |
| dc.description.abstract | Recently, deep reinforcement learning (DRL) has emerged as a promising approach for robotic control. However, the deployment of DRL in real-world robots is hindered by its sensitivity to environmental perturbations. While existing white-box adversarial attacks rely on local gradient information and apply uniform perturbations across all states to evaluate DRL robustness, they fail to account for temporal dynamics and state-specific vulnerabilities. To combat the above challenge, we first conduct a theoretical analysis of white-box attacks in DRL by establishing the adversarial victim-dynamics Markov decision process (AVD-MDP), to derive the necessary and sufficient conditions for a successful attack. Based on this, we propose a selective state-aware reinforcement adversarial attack method, named STAR, to optimize perturbation stealthiness and state visitation dispersion. STAR first employs a soft mask-based state-targeting mechanism to minimize redundant perturbations, enhancing stealthiness and attack effectiveness. Then, it incorporates an information-theoretic optimization objective to maximize mutual information between perturbations, environmental states, and victim actions, ensuring a dispersed state-visitation distribution that steers the victim agent into vulnerable states for maximum return reduction. Extensive experiments demonstrate that STAR outperforms state-of-the-art benchmarks. | - |
| dc.language | eng | - |
| dc.publisher | Institute of Electrical and Electronics Engineers | - |
| dc.relation.ispartof | IEEE Transactions on Mobile Computing | - |
| dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
| dc.subject | Adversarial Attack | - |
| dc.subject | Deep Reinforcement Learning | - |
| dc.subject | Markov Decision Process | - |
| dc.subject | Robotic Manipulation | - |
| dc.title | State-Aware Perturbation Optimization for Robust Deep Reinforcement Learning | - |
| dc.type | Article | - |
| dc.identifier.doi | 10.1109/TMC.2025.3601531 | - |
| dc.identifier.scopus | eid_2-s2.0-105013840603 | - |
| dc.identifier.eissn | 1558-0660 | - |
| dc.identifier.issnl | 1536-1233 | - |
