File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Option-Aware Adversarial Inverse Reinforcement Learning for Robotic Control

TitleOption-Aware Adversarial Inverse Reinforcement Learning for Robotic Control
Authors
Issue Date2023
Citation
Proceedings IEEE International Conference on Robotics and Automation, 2023, v. 2023-May, p. 5902-5908 How to Cite?
AbstractHierarchical Imitation Learning (HIL) has been proposed to recover highly-complex behaviors in long-horizon tasks from expert demonstrations by modeling the task hierarchy with the option framework. Existing methods either overlook the causal relationship between the subtask and its corresponding policy or cannot learn the policy in an end-to-end fashion, which leads to suboptimality. In this work, we develop a novel HIL algorithm based on Adversarial Inverse Reinforcement Learning and adapt it with the Expectation-Maximization algorithm in order to directly recover a hierarchical policy from the unannotated demonstrations. Further, we introduce a directed information term to the objective function to enhance the causality and propose a Variational Autoencoder framework for learning with our objectives in an end-to-end fashion. Theoretical justifications and evaluations on challenging robotic control tasks are provided to show the superiority of our algorithm. The codes are available at https://github.com/LucasCJYSDL/HierAIRL.
Persistent Identifierhttp://hdl.handle.net/10722/361740
ISSN
2023 SCImago Journal Rankings: 1.620

 

DC FieldValueLanguage
dc.contributor.authorChen, Jiayu-
dc.contributor.authorLan, Tian-
dc.contributor.authorAggarwal, Vaneet-
dc.date.accessioned2025-09-16T04:19:38Z-
dc.date.available2025-09-16T04:19:38Z-
dc.date.issued2023-
dc.identifier.citationProceedings IEEE International Conference on Robotics and Automation, 2023, v. 2023-May, p. 5902-5908-
dc.identifier.issn1050-4729-
dc.identifier.urihttp://hdl.handle.net/10722/361740-
dc.description.abstractHierarchical Imitation Learning (HIL) has been proposed to recover highly-complex behaviors in long-horizon tasks from expert demonstrations by modeling the task hierarchy with the option framework. Existing methods either overlook the causal relationship between the subtask and its corresponding policy or cannot learn the policy in an end-to-end fashion, which leads to suboptimality. In this work, we develop a novel HIL algorithm based on Adversarial Inverse Reinforcement Learning and adapt it with the Expectation-Maximization algorithm in order to directly recover a hierarchical policy from the unannotated demonstrations. Further, we introduce a directed information term to the objective function to enhance the causality and propose a Variational Autoencoder framework for learning with our objectives in an end-to-end fashion. Theoretical justifications and evaluations on challenging robotic control tasks are provided to show the superiority of our algorithm. The codes are available at https://github.com/LucasCJYSDL/HierAIRL.-
dc.languageeng-
dc.relation.ispartofProceedings IEEE International Conference on Robotics and Automation-
dc.titleOption-Aware Adversarial Inverse Reinforcement Learning for Robotic Control-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/ICRA48891.2023.10160374-
dc.identifier.scopuseid_2-s2.0-85165216248-
dc.identifier.volume2023-May-
dc.identifier.spage5902-
dc.identifier.epage5908-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats