File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Multi-Timescale Hierarchical Reinforcement Learning for Unified Behavior and Control of Autonomous Driving

TitleMulti-Timescale Hierarchical Reinforcement Learning for Unified Behavior and Control of Autonomous Driving
Authors
Keywordsautonomous driving
motion and path planning
multiple timescale
Reinforcement learning
Issue Date17-Oct-2025
PublisherInstitute of Electrical and Electronics Engineers
Citation
IEEE Robotics and Automation Letters, 2025, v. 10, n. 12, p. 12772-12779 How to Cite?
AbstractReinforcement Learning (RL) is increasingly used in autonomous driving (AD) and shows clear advantages. However, most RL-based AD methods overlook policy structure design. An RL policy that only outputs short-timescale vehicle control commands results in fluctuating driving behavior due to fluctuations in network outputs, while one that only outputs long-timescale driving goals cannot achieve unified optimality of driving behavior and control. Therefore, we propose a multi-timescale hierarchical reinforcement learning approach. Our approach adopts a hierarchical policy structure, where high- and low-level RL policies are unified-trained to produce long-timescale motion guidance and short-timescale control commands, respectively. Therein, motion guidance is explicitly represented by hybrid actions to capture multimodal driving behaviors on structured road and support incremental low-level extend-state updates. Additionally, a hierarchical safety mechanism is designed to ensure multi-timescale safety. Evaluation in simulator-based and HighD dataset-based highway multi-lane scenarios demonstrates that our approach significantly improves AD performance, effectively increasing driving efficiency, action consistency and safety.
Persistent Identifierhttp://hdl.handle.net/10722/366759

 

DC FieldValueLanguage
dc.contributor.authorJin, Guizhe-
dc.contributor.authorLi, Zhuoren-
dc.contributor.authorLeng, Bo-
dc.contributor.authorYu, Ran-
dc.contributor.authorXiong, Lu-
dc.contributor.authorSun, Chen-
dc.date.accessioned2025-11-25T04:21:41Z-
dc.date.available2025-11-25T04:21:41Z-
dc.date.issued2025-10-17-
dc.identifier.citationIEEE Robotics and Automation Letters, 2025, v. 10, n. 12, p. 12772-12779-
dc.identifier.urihttp://hdl.handle.net/10722/366759-
dc.description.abstractReinforcement Learning (RL) is increasingly used in autonomous driving (AD) and shows clear advantages. However, most RL-based AD methods overlook policy structure design. An RL policy that only outputs short-timescale vehicle control commands results in fluctuating driving behavior due to fluctuations in network outputs, while one that only outputs long-timescale driving goals cannot achieve unified optimality of driving behavior and control. Therefore, we propose a multi-timescale hierarchical reinforcement learning approach. Our approach adopts a hierarchical policy structure, where high- and low-level RL policies are unified-trained to produce long-timescale motion guidance and short-timescale control commands, respectively. Therein, motion guidance is explicitly represented by hybrid actions to capture multimodal driving behaviors on structured road and support incremental low-level extend-state updates. Additionally, a hierarchical safety mechanism is designed to ensure multi-timescale safety. Evaluation in simulator-based and HighD dataset-based highway multi-lane scenarios demonstrates that our approach significantly improves AD performance, effectively increasing driving efficiency, action consistency and safety.-
dc.languageeng-
dc.publisherInstitute of Electrical and Electronics Engineers-
dc.relation.ispartofIEEE Robotics and Automation Letters-
dc.subjectautonomous driving-
dc.subjectmotion and path planning-
dc.subjectmultiple timescale-
dc.subjectReinforcement learning-
dc.titleMulti-Timescale Hierarchical Reinforcement Learning for Unified Behavior and Control of Autonomous Driving-
dc.typeArticle-
dc.identifier.doi10.1109/LRA.2025.3623016-
dc.identifier.scopuseid_2-s2.0-105019696149-
dc.identifier.volume10-
dc.identifier.issue12-
dc.identifier.spage12772-
dc.identifier.epage12779-
dc.identifier.eissn2377-3766-
dc.identifier.issnl2377-3766-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats