File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Analyzing sequential activity and travel decisions with interpretable deep inverse reinforcement learning

TitleAnalyzing sequential activity and travel decisions with interpretable deep inverse reinforcement learning
Authors
KeywordsActivity-based travel demand model
Deep learning
Explainable artificial intelligence
Inverse reinforcement learning
Issue Date1-Apr-2026
PublisherElsevier
Citation
Travel Behaviour and Society, 2026, v. 43 How to Cite?
Abstract

Travel demand modeling has shifted from aggregated trip-based models to behavior-oriented activity-based models because daily trips are essentially driven by human activities. To analyze the sequential activity-travel decisions, deep inverse reinforcement learning (DIRL) has proven effective in learning the decision mechanisms by approximating a reward function to represent preferences and a policy function to replicate observed behavior using deep neural networks (DNNs). However, most DIRL applications emphasize prediction accuracy and treat the learned functions as black boxes, offering limited behavioral insight. To address this gap, we propose an interpretable DIRL framework that adapts an adversarial IRL approach for modeling sequential activity-travel behavior. Interpretability is achieved in two ways: (1) we distill the learned policy into a surrogate interpretable Multinomial Logit (MNL) model, enabling the extraction of behavioral drivers from model parameters; and (2) we derive short-term rewards and long-term returns from the learned reward function, quantifying immediate preferences and overall decision outcomes across activity sequences. Applied to real-world travel survey data from Singapore, our framework uncovers meaningful behavioral patterns. The MNL-based surrogate model reveals that travel decisions are shaped by activity schedules, travel time, and socio-demographic attributes, particularly employment type. Reward and return analysis distinguish returners with regular patterns from explorers with irregular ones. Regular patterns yield higher long-term returns, while females and elderly individuals exhibit lower returns, indicating disparities in individual activity patterns. These findings bridge the gap between theory-driven behavioral models and data-driven machine learning, offering actionable insights for transport policy and urban planning.


Persistent Identifierhttp://hdl.handle.net/10722/368275
ISSN
2023 Impact Factor: 5.1
2023 SCImago Journal Rankings: 1.570

 

DC FieldValueLanguage
dc.contributor.authorLiang, Yuebing-
dc.contributor.authorWang, Shenhao-
dc.contributor.authorYu, Jiangbo-
dc.contributor.authorZhao, Zhan-
dc.contributor.authorZhao, Jinhua-
dc.contributor.authorPentland, Sandy-
dc.date.accessioned2025-12-24T00:37:13Z-
dc.date.available2025-12-24T00:37:13Z-
dc.date.issued2026-04-01-
dc.identifier.citationTravel Behaviour and Society, 2026, v. 43-
dc.identifier.issn2214-367X-
dc.identifier.urihttp://hdl.handle.net/10722/368275-
dc.description.abstract<p>Travel demand modeling has shifted from aggregated trip-based models to behavior-oriented activity-based models because daily trips are essentially driven by human activities. To analyze the sequential activity-travel decisions, deep inverse reinforcement learning (DIRL) has proven effective in learning the decision mechanisms by approximating a reward function to represent preferences and a policy function to replicate observed behavior using deep neural networks (DNNs). However, most DIRL applications emphasize prediction accuracy and treat the learned functions as black boxes, offering limited behavioral insight. To address this gap, we propose an interpretable DIRL framework that adapts an adversarial IRL approach for modeling sequential activity-travel behavior. Interpretability is achieved in two ways: (1) we distill the learned policy into a surrogate interpretable Multinomial Logit (MNL) model, enabling the extraction of behavioral drivers from model parameters; and (2) we derive short-term rewards and long-term returns from the learned reward function, quantifying immediate preferences and overall decision outcomes across activity sequences. Applied to real-world travel survey data from Singapore, our framework uncovers meaningful behavioral patterns. The MNL-based surrogate model reveals that travel decisions are shaped by activity schedules, travel time, and socio-demographic attributes, particularly employment type. Reward and return analysis distinguish returners with regular patterns from explorers with irregular ones. Regular patterns yield higher long-term returns, while females and elderly individuals exhibit lower returns, indicating disparities in individual activity patterns. These findings bridge the gap between theory-driven behavioral models and data-driven machine learning, offering actionable insights for transport policy and urban planning.</p>-
dc.languageeng-
dc.publisherElsevier-
dc.relation.ispartofTravel Behaviour and Society-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectActivity-based travel demand model-
dc.subjectDeep learning-
dc.subjectExplainable artificial intelligence-
dc.subjectInverse reinforcement learning-
dc.titleAnalyzing sequential activity and travel decisions with interpretable deep inverse reinforcement learning -
dc.typeArticle-
dc.identifier.doi10.1016/j.tbs.2025.101171-
dc.identifier.scopuseid_2-s2.0-105021667849-
dc.identifier.volume43-
dc.identifier.eissn2214-3688-
dc.identifier.issnl2214-367X-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats