Hao Liao, Wei Zhang, Zhanyi Huang, Zexiao Long, Mingyang Zhou, Xiaoqun Wu, Rui Mao, Chi Ho Yeung
{"title":"Exploring Loss Landscapes through the Lens of Spin Glass Theory","authors":"Hao Liao, Wei Zhang, Zhanyi Huang, Zexiao Long, Mingyang Zhou, Xiaoqun Wu, Rui Mao, Chi Ho Yeung","doi":"arxiv-2407.20724","DOIUrl":null,"url":null,"abstract":"In the past decade, significant strides in deep learning have led to numerous\ngroundbreaking applications. Despite these advancements, the understanding of\nthe high generalizability of deep learning, especially in such an\nover-parametrized space, remains limited. Successful applications are often\nconsidered as empirical rather than scientific achievements. For instance, deep\nneural networks' (DNNs) internal representations, decision-making mechanism,\nabsence of overfitting in an over-parametrized space, high generalizability,\netc., remain less understood. This paper delves into the loss landscape of DNNs\nthrough the lens of spin glass in statistical physics, i.e. a system\ncharacterized by a complex energy landscape with numerous metastable states, to\nbetter understand how DNNs work. We investigated a single hidden layer\nRectified Linear Unit (ReLU) neural network model, and introduced several\nprotocols to examine the analogy between DNNs (trained with datasets including\nMNIST and CIFAR10) and spin glass. Specifically, we used (1) random walk in the\nparameter space of DNNs to unravel the structures in their loss landscape; (2)\na permutation-interpolation protocol to study the connection between copies of\nidentical regions in the loss landscape due to the permutation symmetry in the\nhidden layers; (3) hierarchical clustering to reveal the hierarchy among\ntrained solutions of DNNs, reminiscent of the so-called Replica Symmetry\nBreaking (RSB) phenomenon (i.e. the Parisi solution) in analogy to spin glass;\n(4) finally, we examine the relationship between the degree of the ruggedness\nof the loss landscape of the DNN and its generalizability, showing an\nimprovement of flattened minima.","PeriodicalId":501066,"journal":{"name":"arXiv - PHYS - Disordered Systems and Neural Networks","volume":"22 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - PHYS - Disordered Systems and Neural Networks","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.20724","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In the past decade, significant strides in deep learning have led to numerous
groundbreaking applications. Despite these advancements, the understanding of
the high generalizability of deep learning, especially in such an
over-parametrized space, remains limited. Successful applications are often
considered as empirical rather than scientific achievements. For instance, deep
neural networks' (DNNs) internal representations, decision-making mechanism,
absence of overfitting in an over-parametrized space, high generalizability,
etc., remain less understood. This paper delves into the loss landscape of DNNs
through the lens of spin glass in statistical physics, i.e. a system
characterized by a complex energy landscape with numerous metastable states, to
better understand how DNNs work. We investigated a single hidden layer
Rectified Linear Unit (ReLU) neural network model, and introduced several
protocols to examine the analogy between DNNs (trained with datasets including
MNIST and CIFAR10) and spin glass. Specifically, we used (1) random walk in the
parameter space of DNNs to unravel the structures in their loss landscape; (2)
a permutation-interpolation protocol to study the connection between copies of
identical regions in the loss landscape due to the permutation symmetry in the
hidden layers; (3) hierarchical clustering to reveal the hierarchy among
trained solutions of DNNs, reminiscent of the so-called Replica Symmetry
Breaking (RSB) phenomenon (i.e. the Parisi solution) in analogy to spin glass;
(4) finally, we examine the relationship between the degree of the ruggedness
of the loss landscape of the DNN and its generalizability, showing an
improvement of flattened minima.