Continual learning benefits from multiple sleep stages: NREM, REM, and Synaptic Downscaling

Brian S. Robinson, Clare W. Lau, Alexander New, S. Nichols, Erik C. Johnson, M. Wolmetz, W. Coon
{"title":"Continual learning benefits from multiple sleep stages: NREM, REM, and Synaptic Downscaling","authors":"Brian S. Robinson, Clare W. Lau, Alexander New, S. Nichols, Erik C. Johnson, M. Wolmetz, W. Coon","doi":"10.1109/IJCNN55064.2022.9891965","DOIUrl":null,"url":null,"abstract":"Learning new tasks and skills in succession without overwriting or interfering with prior learning (i.e., “catastrophic forgetting”) is a computational challenge for both artificial and biological neural networks, yet artificial systems struggle to achieve even rudimentary parity with the performance and functionality apparent in biology. One of the processes found in biology that can be adapted for use in artificial systems is sleep, in which the brain deploys numerous neural operations relevant to continual learning and ripe for artificial adaptation. Here, we investigate how modeling three distinct components of mammalian sleep together affects continual learning in artificial neural networks: (1) a veridical memory replay process observed during non-rapid eye movement (NREM) sleep; (2) a generative memory replay process linked to REM sleep; and (3) a synaptic downscaling process which has been proposed to tune signal-to-noise ratios and support neural upkeep. To create this tripartite artificial sleep, we modeled NREM veridical replay by training the network using intermediate representations of samples from the current task. We modeled REM by utilizing a generator network to create intermediate representations of samples from previous tasks for training. Synaptic downscaling, a novel con-tribution, is modeled utilizing a size-dependent downscaling of network weights. We find benefits from the inclusion of all three sleep components when evaluating performance on a continual learning CIFAR-100 image classification benchmark. Maximum accuracy improved during training and catastrophic forgetting was reduced during later tasks. While some catastrophic forget-ting persisted over the course of network training, higher levels of synaptic downscaling lead to better retention of early tasks and further facilitated the recovery of early task accuracy during subsequent training. One key takeaway is that there is a trade-off at hand when considering the level of synaptic downscaling to use - more aggressive downscaling better protects early tasks, but less downscaling enhances the ability to learn new tasks. Intermediate levels can strike a balance with the highest overall accuracies during training. Overall, our results both provide insight into how to adapt sleep components to enhance artificial continual learning systems and highlight areas for future neuroscientific sleep research to further such systems.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Joint Conference on Neural Networks (IJCNN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN55064.2022.9891965","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Learning new tasks and skills in succession without overwriting or interfering with prior learning (i.e., “catastrophic forgetting”) is a computational challenge for both artificial and biological neural networks, yet artificial systems struggle to achieve even rudimentary parity with the performance and functionality apparent in biology. One of the processes found in biology that can be adapted for use in artificial systems is sleep, in which the brain deploys numerous neural operations relevant to continual learning and ripe for artificial adaptation. Here, we investigate how modeling three distinct components of mammalian sleep together affects continual learning in artificial neural networks: (1) a veridical memory replay process observed during non-rapid eye movement (NREM) sleep; (2) a generative memory replay process linked to REM sleep; and (3) a synaptic downscaling process which has been proposed to tune signal-to-noise ratios and support neural upkeep. To create this tripartite artificial sleep, we modeled NREM veridical replay by training the network using intermediate representations of samples from the current task. We modeled REM by utilizing a generator network to create intermediate representations of samples from previous tasks for training. Synaptic downscaling, a novel con-tribution, is modeled utilizing a size-dependent downscaling of network weights. We find benefits from the inclusion of all three sleep components when evaluating performance on a continual learning CIFAR-100 image classification benchmark. Maximum accuracy improved during training and catastrophic forgetting was reduced during later tasks. While some catastrophic forget-ting persisted over the course of network training, higher levels of synaptic downscaling lead to better retention of early tasks and further facilitated the recovery of early task accuracy during subsequent training. One key takeaway is that there is a trade-off at hand when considering the level of synaptic downscaling to use - more aggressive downscaling better protects early tasks, but less downscaling enhances the ability to learn new tasks. Intermediate levels can strike a balance with the highest overall accuracies during training. Overall, our results both provide insight into how to adapt sleep components to enhance artificial continual learning systems and highlight areas for future neuroscientific sleep research to further such systems.
持续学习受益于多个睡眠阶段:非快速眼动、快速眼动和突触降阶
连续学习新的任务和技能而不覆盖或干扰先前的学习(即“灾难性遗忘”)对人工和生物神经网络来说都是一个计算挑战,然而人工系统很难达到与生物学中明显的性能和功能相当的基本水平。在生物学中发现的一个可以用于人工系统的过程是睡眠,在睡眠中,大脑部署了许多与持续学习相关的神经操作,并且为人工适应做好了准备。在这里,我们研究了如何将哺乳动物睡眠的三个不同组成部分建模在一起影响人工神经网络中的持续学习:(1)在非快速眼动(NREM)睡眠期间观察到的真实记忆重放过程;(2)与快速眼动睡眠相关的生成性记忆重放过程;(3)突触降尺度过程,该过程被提出用于调节信噪比和支持神经维持。为了创建这种三方人工睡眠,我们通过使用来自当前任务的样本的中间表示来训练网络,从而模拟了NREM的验证重放。我们通过使用生成器网络来创建来自先前训练任务的样本的中间表示来建模REM。突触降尺度是一种新的贡献,它利用网络权重的大小依赖降尺度来建模。在持续学习CIFAR-100图像分类基准上评估性能时,我们发现包含所有三个睡眠组件的好处。在训练过程中,最大准确度得到了提高,而在随后的任务中,灾难性遗忘则有所减少。虽然一些灾难性的遗忘在网络训练过程中持续存在,但较高水平的突触缩小导致对早期任务的更好保留,并进一步促进了在后续训练中早期任务准确性的恢复。一个关键的结论是,当考虑到突触缩小使用的水平时,有一个权衡——更积极的缩小可以更好地保护早期任务,但更少的缩小可以增强学习新任务的能力。中级水平可以在训练中达到最高的整体准确度。总的来说,我们的研究结果既提供了如何调整睡眠成分以增强人工持续学习系统的见解,也突出了未来神经科学睡眠研究的领域,以进一步发展此类系统。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信