Growing Deep Forests Efficiently with Soft Routing and Learned Connectivity

Jianghao Shen, Sicheng Wang, Zhangyang Wang
{"title":"Growing Deep Forests Efficiently with Soft Routing and Learned Connectivity","authors":"Jianghao Shen, Sicheng Wang, Zhangyang Wang","doi":"10.1109/ICDMW.2018.00065","DOIUrl":null,"url":null,"abstract":"Despite the latest prevailing success of deep neural networks (DNNs), several concerns have been raised against their usage, including the lack of intepretability the gap between DNNs and other well-established machine learning models, and the growingly expensive computational costs. A number of recent works [1],[2],[3] explored the alternative to sequentially stacking decision tree/random forest building blocks in a purely feed-forward way, with no need of back propagation. Since decision trees enjoy inherent reasoning transparency, such deep forest models can also facilitate the understanding of the internal decision making process. This paper further extends the deep forest idea in several important aspects. Firstly, we employ a probabilistic tree whose nodes make probabilistic routing decisions, a.k.a., \"soft routing\", rather than hard binary decisions. Besides enhancing the flexibility, it also enables non-greedy optimization for each tree. Second, we propose an innovative topology learning strategy: every node in the ree now maintains a new learnable hyperparameter indicating the probability that it will be a leaf node. In that way, the tree will jointly optimize both its parameters and the tree topology during training. Experiments on the MNIST dataset demonstrate that our empowered deep forests can achieve better or comparable performance than [1],[3], with dramatically reduced model complexity. For example, our model with only 1 layer of 15 trees can perform comparably with the model in [3] with 2 layers of 2000 trees each.","PeriodicalId":259600,"journal":{"name":"2018 IEEE International Conference on Data Mining Workshops (ICDMW)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE International Conference on Data Mining Workshops (ICDMW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDMW.2018.00065","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Despite the latest prevailing success of deep neural networks (DNNs), several concerns have been raised against their usage, including the lack of intepretability the gap between DNNs and other well-established machine learning models, and the growingly expensive computational costs. A number of recent works [1],[2],[3] explored the alternative to sequentially stacking decision tree/random forest building blocks in a purely feed-forward way, with no need of back propagation. Since decision trees enjoy inherent reasoning transparency, such deep forest models can also facilitate the understanding of the internal decision making process. This paper further extends the deep forest idea in several important aspects. Firstly, we employ a probabilistic tree whose nodes make probabilistic routing decisions, a.k.a., "soft routing", rather than hard binary decisions. Besides enhancing the flexibility, it also enables non-greedy optimization for each tree. Second, we propose an innovative topology learning strategy: every node in the ree now maintains a new learnable hyperparameter indicating the probability that it will be a leaf node. In that way, the tree will jointly optimize both its parameters and the tree topology during training. Experiments on the MNIST dataset demonstrate that our empowered deep forests can achieve better or comparable performance than [1],[3], with dramatically reduced model complexity. For example, our model with only 1 layer of 15 trees can perform comparably with the model in [3] with 2 layers of 2000 trees each.
用软路由和学习连接有效地生长深森林
尽管深度神经网络(dnn)最近取得了普遍的成功,但人们对其使用提出了一些担忧,包括缺乏可解释性,dnn与其他成熟的机器学习模型之间的差距,以及日益昂贵的计算成本。最近的一些研究[1],[2],[3]探讨了以纯前馈方式顺序堆叠决策树/随机森林构建块的替代方案,不需要反向传播。由于决策树具有固有的推理透明性,这种深度森林模型也可以促进对内部决策过程的理解。本文在几个重要方面进一步扩展了深林思想。首先,我们采用一个概率树,它的节点做出概率路由决策,也就是“软路由”,而不是硬二进制决策。除了增强灵活性外,它还支持对每棵树进行非贪婪优化。其次,我们提出了一种创新的拓扑学习策略:树中的每个节点现在都保持一个新的可学习超参数,表示它将成为叶节点的概率。这样,树将在训练过程中共同优化其参数和树的拓扑结构。在MNIST数据集上的实验表明,我们的授权深度森林可以获得比[1],[3]更好或相当的性能,同时显著降低了模型复杂性。例如,我们的模型只有1层15棵树,与[3]中的模型有2层,每层2000棵树的表现相当。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信