Silvio Kalaj, Clarissa Lauditi, Gabriele Perugini, Carlo Lucibello, Enrico M. Malatesta, Matteo Negri
{"title":"Random Features Hopfield Networks generalize retrieval to previously unseen examples","authors":"Silvio Kalaj, Clarissa Lauditi, Gabriele Perugini, Carlo Lucibello, Enrico M. Malatesta, Matteo Negri","doi":"arxiv-2407.05658","DOIUrl":null,"url":null,"abstract":"It has been recently shown that a learning transition happens when a Hopfield\nNetwork stores examples generated as superpositions of random features, where\nnew attractors corresponding to such features appear in the model. In this work\nwe reveal that the network also develops attractors corresponding to previously\nunseen examples generated with the same set of features. We explain this\nsurprising behaviour in terms of spurious states of the learned features: we\nargue that, increasing the number of stored examples beyond the learning\ntransition, the model also learns to mix the features to represent both stored\nand previously unseen examples. We support this claim with the computation of\nthe phase diagram of the model.","PeriodicalId":501066,"journal":{"name":"arXiv - PHYS - Disordered Systems and Neural Networks","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - PHYS - Disordered Systems and Neural Networks","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.05658","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
It has been recently shown that a learning transition happens when a Hopfield
Network stores examples generated as superpositions of random features, where
new attractors corresponding to such features appear in the model. In this work
we reveal that the network also develops attractors corresponding to previously
unseen examples generated with the same set of features. We explain this
surprising behaviour in terms of spurious states of the learned features: we
argue that, increasing the number of stored examples beyond the learning
transition, the model also learns to mix the features to represent both stored
and previously unseen examples. We support this claim with the computation of
the phase diagram of the model.