DONN: leveraging heterogeneous outer products for CTR prediction

Tae-Suk Kim
{"title":"DONN: leveraging heterogeneous outer products for CTR prediction","authors":"Tae-Suk Kim","doi":"10.1007/s00521-024-10296-x","DOIUrl":null,"url":null,"abstract":"<p>A primary strategy for constructing click-through rate models based on deep learning involves combining a multi-layer perceptron (MLP) with custom networks that can effectively capture the interactions between different features. This is due to the widespread recognition that relying solely on a vanilla MLP network is not effective in acquiring knowledge about multiplicative feature interactions. These custom networks often employ product methods, such as inner, Hadamard, and outer products, to construct dedicated architectures for this purpose. Among these methods, the outer product has shown superiority in capturing feature interactions. However, the resulting quadratic form from the outer product operation limits the conveyance of informative higher-order interactions to the MLP. Efforts to address this limitation have led to models attempting to increase interaction degrees to higher orders. However, utilizing matrix factorization techniques to reduce learning parameters has resulted in information loss and decreased performance. Furthermore, previous studies have constrained the MLP’s potential by providing it with inputs consisting of homogeneous outer products, thus limiting available information diversity. To overcome these challenges, we introduce DONN, a model that leverages a composite-wise bilinear module incorporating factorized bilinear pooling to mitigate information loss and facilitate higher-order interaction development. Additionally, DONN utilizes a feature-wise bilinear module for outer product computations between feature pairs, augmenting the MLP with combined information. By employing heterogeneous outer products, DONN enhances the MLP’s prediction capabilities, enabling the recognition of additional nonlinear interdependencies. Our evaluation on two benchmark datasets demonstrates that DONN surpasses state-of-the-art models in terms of performance.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Computing and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s00521-024-10296-x","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

A primary strategy for constructing click-through rate models based on deep learning involves combining a multi-layer perceptron (MLP) with custom networks that can effectively capture the interactions between different features. This is due to the widespread recognition that relying solely on a vanilla MLP network is not effective in acquiring knowledge about multiplicative feature interactions. These custom networks often employ product methods, such as inner, Hadamard, and outer products, to construct dedicated architectures for this purpose. Among these methods, the outer product has shown superiority in capturing feature interactions. However, the resulting quadratic form from the outer product operation limits the conveyance of informative higher-order interactions to the MLP. Efforts to address this limitation have led to models attempting to increase interaction degrees to higher orders. However, utilizing matrix factorization techniques to reduce learning parameters has resulted in information loss and decreased performance. Furthermore, previous studies have constrained the MLP’s potential by providing it with inputs consisting of homogeneous outer products, thus limiting available information diversity. To overcome these challenges, we introduce DONN, a model that leverages a composite-wise bilinear module incorporating factorized bilinear pooling to mitigate information loss and facilitate higher-order interaction development. Additionally, DONN utilizes a feature-wise bilinear module for outer product computations between feature pairs, augmenting the MLP with combined information. By employing heterogeneous outer products, DONN enhances the MLP’s prediction capabilities, enabling the recognition of additional nonlinear interdependencies. Our evaluation on two benchmark datasets demonstrates that DONN surpasses state-of-the-art models in terms of performance.

Abstract Image

DONN:利用异构外部产品进行点击率预测
基于深度学习构建点击率模型的一个主要策略是将多层感知器(MLP)与能有效捕捉不同特征之间相互作用的定制网络相结合。这是因为人们普遍认识到,仅仅依靠普通的 MLP 网络无法有效获取关于乘法特征交互的知识。这些定制网络通常采用内积、哈达玛积和外积等积方法来构建专用架构。在这些方法中,外积法在捕捉特征相互作用方面表现出了优越性。然而,外积运算产生的二次方形式限制了向 MLP 传递高阶交互信息。为解决这一局限性,一些模型试图将交互度提高到更高阶。然而,利用矩阵因式分解技术来减少学习参数会导致信息丢失和性能下降。此外,以往的研究还限制了 MLP 的潜能,为其提供了由同质外积组成的输入,从而限制了可用信息的多样性。为了克服这些挑战,我们引入了 DONN 模型,该模型利用复合双线性模块(composite-wise bilinear module)和因子化双线性集合(factorized bilinear pooling)来减少信息丢失,促进高阶交互的发展。此外,DONN 还利用特征双线性模块进行特征对之间的外积计算,用组合信息增强 MLP。通过使用异质外积,DONN 增强了 MLP 的预测能力,从而能够识别更多的非线性相互依存关系。我们在两个基准数据集上进行的评估表明,DONN 在性能方面超越了最先进的模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信