Prototype Analysis in Hopfield Networks with Hebbian Learning

Hayden McAlister, Anthony Robins, Lech Szymanski
{"title":"Prototype Analysis in Hopfield Networks with Hebbian Learning","authors":"Hayden McAlister, Anthony Robins, Lech Szymanski","doi":"arxiv-2407.03342","DOIUrl":null,"url":null,"abstract":"We discuss prototype formation in the Hopfield network. Typically, Hebbian\nlearning with highly correlated states leads to degraded memory performance. We\nshow this type of learning can lead to prototype formation, where unlearned\nstates emerge as representatives of large correlated subsets of states,\nalleviating capacity woes. This process has similarities to prototype learning\nin human cognition. We provide a substantial literature review of prototype\nlearning in associative memories, covering contributions from psychology,\nstatistical physics, and computer science. We analyze prototype formation from\na theoretical perspective and derive a stability condition for these states\nbased on the number of examples of the prototype presented for learning, the\nnoise in those examples, and the number of non-example states presented. The\nstability condition is used to construct a probability of stability for a\nprototype state as the factors of stability change. We also note similarities\nto traditional network analysis, allowing us to find a prototype capacity. We\ncorroborate these expectations of prototype formation with experiments using a\nsimple Hopfield network with standard Hebbian learning. We extend our\nexperiments to a Hopfield network trained on data with multiple prototypes and\nfind the network is capable of stabilizing multiple prototypes concurrently. We\nmeasure the basins of attraction of the multiple prototype states, finding\nattractor strength grows with the number of examples and the agreement of\nexamples. We link the stability and dominance of prototype states to the energy\nprofile of these states, particularly when comparing the profile shape to\ntarget states or other spurious states.","PeriodicalId":501066,"journal":{"name":"arXiv - PHYS - Disordered Systems and Neural Networks","volume":"364 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - PHYS - Disordered Systems and Neural Networks","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.03342","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

We discuss prototype formation in the Hopfield network. Typically, Hebbian learning with highly correlated states leads to degraded memory performance. We show this type of learning can lead to prototype formation, where unlearned states emerge as representatives of large correlated subsets of states, alleviating capacity woes. This process has similarities to prototype learning in human cognition. We provide a substantial literature review of prototype learning in associative memories, covering contributions from psychology, statistical physics, and computer science. We analyze prototype formation from a theoretical perspective and derive a stability condition for these states based on the number of examples of the prototype presented for learning, the noise in those examples, and the number of non-example states presented. The stability condition is used to construct a probability of stability for a prototype state as the factors of stability change. We also note similarities to traditional network analysis, allowing us to find a prototype capacity. We corroborate these expectations of prototype formation with experiments using a simple Hopfield network with standard Hebbian learning. We extend our experiments to a Hopfield network trained on data with multiple prototypes and find the network is capable of stabilizing multiple prototypes concurrently. We measure the basins of attraction of the multiple prototype states, finding attractor strength grows with the number of examples and the agreement of examples. We link the stability and dominance of prototype states to the energy profile of these states, particularly when comparing the profile shape to target states or other spurious states.
采用 Hebbian 学习的 Hopfield 网络中的原型分析
我们讨论了 Hopfield 网络中的原型形成。通常,具有高度相关状态的希比安学习会导致记忆性能下降。在这种情况下,未学习的状态会作为大相关状态子集的代表出现,从而缓解容量问题。这一过程与人类认知中的原型学习有相似之处。我们对联想记忆中的原型学习进行了大量的文献综述,涉及心理学、统计物理学和计算机科学等领域。我们从理论角度分析了原型的形成,并根据为学习而呈现的原型示例的数量、这些示例中的噪声以及呈现的非示例状态的数量,推导出了这些状态的稳定条件。随着稳定因素的变化,稳定条件被用来构建原型状态的稳定概率。我们还注意到与传统网络分析的相似性,这使我们能够找到原型容量。我们通过使用标准海比学习的简单 Hopfield 网络进行实验,证实了对原型形成的这些预期。我们将实验扩展到在具有多个原型的数据上训练的 Hopfield 网络,发现该网络能够同时稳定多个原型。我们测量了多个原型状态的吸引盆地,发现吸引器强度会随着示例数量和示例一致性的增加而增加。我们将原型态的稳定性和主导性与这些态的能量剖面联系起来,特别是在将剖面形状与目标态或其他虚假态进行比较时。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信