Interpretable Machine Learning for Mode Choice Modeling on Tracking-Based Revealed Preference Data

Victoria Dahmen, S. Weikl, K. Bogenberger
{"title":"Interpretable Machine Learning for Mode Choice Modeling on Tracking-Based Revealed Preference Data","authors":"Victoria Dahmen, S. Weikl, K. Bogenberger","doi":"10.1177/03611981241246973","DOIUrl":null,"url":null,"abstract":"Mode choice modeling is imperative for predicting and understanding travel behavior. For this purpose, machine learning (ML) models have increasingly been applied to stated preference and traditional self-recorded revealed preference data with promising results, particularly for extreme gradient boosting (XGBoost) and random forest (RF) models. Because of the rise in the use of tracking-based smartphone applications for recording travel behavior, we address the important and unprecedented task of testing these ML models for mode choice modeling on such data. Furthermore, as ML approaches are still criticized for leading to results that are hard to understand, we consider it essential to provide an in-depth interpretability analysis of the best-performing model. Our results show that the XGBoost and RF models far outperform a conventional multinomial logit model, both overall and for each mode. The interpretability analysis using the Shapley additive explanations approach reveals that the XGBoost model can be explained well at the overall and mode level. In addition, we demonstrate how to analyze individual predictions. Lastly, a sensitivity analysis gives insight into the relative importance of different data sources, sample size, and user involvement. We conclude that the XGBoost model performs best, while also being explainable. Insights generated by such models can be used, for instance, to predict mode choice decisions for arbitrary origin–destination pairs to see which impacts infrastructural changes would have on the mode share.","PeriodicalId":309251,"journal":{"name":"Transportation Research Record: Journal of the Transportation Research Board","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Transportation Research Record: Journal of the Transportation Research Board","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1177/03611981241246973","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Mode choice modeling is imperative for predicting and understanding travel behavior. For this purpose, machine learning (ML) models have increasingly been applied to stated preference and traditional self-recorded revealed preference data with promising results, particularly for extreme gradient boosting (XGBoost) and random forest (RF) models. Because of the rise in the use of tracking-based smartphone applications for recording travel behavior, we address the important and unprecedented task of testing these ML models for mode choice modeling on such data. Furthermore, as ML approaches are still criticized for leading to results that are hard to understand, we consider it essential to provide an in-depth interpretability analysis of the best-performing model. Our results show that the XGBoost and RF models far outperform a conventional multinomial logit model, both overall and for each mode. The interpretability analysis using the Shapley additive explanations approach reveals that the XGBoost model can be explained well at the overall and mode level. In addition, we demonstrate how to analyze individual predictions. Lastly, a sensitivity analysis gives insight into the relative importance of different data sources, sample size, and user involvement. We conclude that the XGBoost model performs best, while also being explainable. Insights generated by such models can be used, for instance, to predict mode choice decisions for arbitrary origin–destination pairs to see which impacts infrastructural changes would have on the mode share.
基于跟踪揭示偏好数据的模式选择建模的可解释机器学习
模式选择建模对于预测和理解旅行行为至关重要。为此,越来越多的机器学习(ML)模型被应用于陈述偏好和传统的自我记录揭示偏好数据,尤其是极端梯度提升(XGBoost)和随机森林(RF)模型,并取得了可喜的成果。由于使用基于跟踪的智能手机应用程序来记录出行行为的情况日益增多,我们将在此类数据上测试这些用于模式选择建模的 ML 模型,这是一项前所未有的重要任务。此外,由于 ML 方法仍然被批评为导致难以理解的结果,我们认为对表现最佳的模型进行深入的可解释性分析是非常必要的。我们的结果表明,XGBoost 和 RF 模型无论在整体上还是在每种模式上都远远优于传统的多二项 logit 模型。使用 Shapley 加法解释方法进行的可解释性分析表明,XGBoost 模型在整体和模式层面都能得到很好的解释。此外,我们还演示了如何分析单个预测。最后,通过敏感性分析,我们了解了不同数据源、样本大小和用户参与的相对重要性。我们得出的结论是,XGBoost 模型表现最佳,同时也具有可解释性。例如,此类模型产生的洞察力可用于预测任意出发地-目的地对的模式选择决策,以了解基础设施的变化会对模式共享产生哪些影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信