Tackling the Accuracy-Interpretability Trade-off: Interpretable Deep Learning Models for Satellite Image-based Real Estate Appraisal

IF 2.5 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS
Jan-Peter Kucklick, Oliver Müller
{"title":"Tackling the Accuracy-Interpretability Trade-off: Interpretable Deep Learning Models for Satellite Image-based Real Estate Appraisal","authors":"Jan-Peter Kucklick, Oliver Müller","doi":"10.1145/3567430","DOIUrl":null,"url":null,"abstract":"Deep learning models fuel many modern decision support systems, because they typically provide high predictive performance. Among other domains, deep learning is used in real-estate appraisal, where it allows extending the analysis from hard facts only (e.g., size, age) to also consider more implicit information about the location or appearance of houses in the form of image data. However, one downside of deep learning models is their intransparent mechanic of decision making, which leads to a trade-off between accuracy and interpretability. This limits their applicability for tasks where a justification of the decision is necessary. Therefore, in this article, we first combine different perspectives on interpretability into a multi-dimensional framework for a socio-technical perspective on explainable artificial intelligence. Second, we measure the performance gains of using multi-view deep learning, which leverages additional image data (satellite images) for real estate appraisal. Third, we propose and test a novel post hoc explainability method called Grad-Ram. This modified version of Grad-Cam mitigates the intransparency of convolutional neural networks for predicting continuous outcome variables. With this, we try to reduce the accuracy-interpretability trade-off of multi-view deep learning models. Our proposed network architecture outperforms traditional hedonic regression models by 34% in terms of MAE. Furthermore, we find that the used satellite images are the second most important predictor after square feet in our model and that the network learns interpretable patterns about the neighborhood structure and density.","PeriodicalId":45274,"journal":{"name":"ACM Transactions on Management Information Systems","volume":"14 1","pages":"1 - 24"},"PeriodicalIF":2.5000,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM Transactions on Management Information Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3567430","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 1

Abstract

Deep learning models fuel many modern decision support systems, because they typically provide high predictive performance. Among other domains, deep learning is used in real-estate appraisal, where it allows extending the analysis from hard facts only (e.g., size, age) to also consider more implicit information about the location or appearance of houses in the form of image data. However, one downside of deep learning models is their intransparent mechanic of decision making, which leads to a trade-off between accuracy and interpretability. This limits their applicability for tasks where a justification of the decision is necessary. Therefore, in this article, we first combine different perspectives on interpretability into a multi-dimensional framework for a socio-technical perspective on explainable artificial intelligence. Second, we measure the performance gains of using multi-view deep learning, which leverages additional image data (satellite images) for real estate appraisal. Third, we propose and test a novel post hoc explainability method called Grad-Ram. This modified version of Grad-Cam mitigates the intransparency of convolutional neural networks for predicting continuous outcome variables. With this, we try to reduce the accuracy-interpretability trade-off of multi-view deep learning models. Our proposed network architecture outperforms traditional hedonic regression models by 34% in terms of MAE. Furthermore, we find that the used satellite images are the second most important predictor after square feet in our model and that the network learns interpretable patterns about the neighborhood structure and density.
解决精度-可解释性权衡:基于卫星图像的房地产评估的可解释深度学习模型
深度学习模型为许多现代决策支持系统提供动力,因为它们通常提供高预测性能。在其他领域中,深度学习被用于房地产评估,它允许将分析从硬事实(例如,大小,年龄)扩展到以图像数据的形式考虑更多关于房屋位置或外观的隐含信息。然而,深度学习模型的一个缺点是它们的决策机制不透明,这导致了准确性和可解释性之间的权衡。这限制了它们在需要对决策进行证明的任务中的适用性。因此,在本文中,我们首先将关于可解释性的不同观点结合到一个多维框架中,以社会技术视角来研究可解释性人工智能。其次,我们测量了使用多视图深度学习的性能增益,它利用额外的图像数据(卫星图像)进行房地产评估。第三,我们提出并测试了一种新的事后可解释性方法,称为Grad-Ram。这种改进版本的Grad-Cam减轻了卷积神经网络预测连续结果变量的不透明性。因此,我们试图减少多视图深度学习模型的准确性和可解释性之间的权衡。我们提出的网络架构在MAE方面优于传统的享乐回归模型34%。此外,我们发现使用的卫星图像是我们模型中仅次于平方英尺的第二个最重要的预测因子,并且网络学习了关于社区结构和密度的可解释模式。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
ACM Transactions on Management Information Systems
ACM Transactions on Management Information Systems COMPUTER SCIENCE, INFORMATION SYSTEMS-
CiteScore
6.30
自引率
20.00%
发文量
60
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信