Learning low-dimensional generalizable natural features from retina using a U-net.

Siwei Wang, Benjamin Hoshal, Elizabeth A de Laittre, Olivier Marre, Michael J Berry, Stephanie E Palmer
{"title":"Learning low-dimensional generalizable natural features from retina using a U-net.","authors":"Siwei Wang, Benjamin Hoshal, Elizabeth A de Laittre, Olivier Marre, Michael J Berry, Stephanie E Palmer","doi":"","DOIUrl":null,"url":null,"abstract":"<p><p>Much of sensory neuroscience focuses on presenting stimuli that are chosen by the experimenter because they are parametric and easy to sample and are thought to be behaviorally relevant to the organism. However, it is not generally known what these relevant features are in complex, natural scenes. This work focuses on using the retinal encoding of natural movies to determine the presumably behaviorally-relevant features that the brain represents. It is prohibitive to parameterize a natural movie and its respective retinal encoding fully. We use time within a natural movie as a proxy for the whole suite of features evolving across the scene. We then use a task-agnostic deep architecture, an encoder-decoder, to model the retinal encoding process and characterize its representation of \"time in the natural scene\" in a compressed latent space. In our end-to-end training, an encoder learns a compressed latent representation from a large population of salamander retinal ganglion cells responding to natural movies, while a decoder samples from this compressed latent space to generate the appropriate future movie frame. By comparing latent representations of retinal activity from three movies, we find that the retina has a generalizable encoding for time in the natural scene: the precise, low-dimensional representation of time learned from one movie can be used to represent time in a different movie, with up to 17 ms resolution. We then show that static textures and velocity features of a natural movie are synergistic. The retina simultaneously encodes both to establishes a generalizable, low-dimensional representation of time in the natural scene.</p>","PeriodicalId":72099,"journal":{"name":"Advances in neural information processing systems","volume":"35 ","pages":"11355-11368"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10289798/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advances in neural information processing systems","FirstCategoryId":"1085","ListUrlMain":"","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Much of sensory neuroscience focuses on presenting stimuli that are chosen by the experimenter because they are parametric and easy to sample and are thought to be behaviorally relevant to the organism. However, it is not generally known what these relevant features are in complex, natural scenes. This work focuses on using the retinal encoding of natural movies to determine the presumably behaviorally-relevant features that the brain represents. It is prohibitive to parameterize a natural movie and its respective retinal encoding fully. We use time within a natural movie as a proxy for the whole suite of features evolving across the scene. We then use a task-agnostic deep architecture, an encoder-decoder, to model the retinal encoding process and characterize its representation of "time in the natural scene" in a compressed latent space. In our end-to-end training, an encoder learns a compressed latent representation from a large population of salamander retinal ganglion cells responding to natural movies, while a decoder samples from this compressed latent space to generate the appropriate future movie frame. By comparing latent representations of retinal activity from three movies, we find that the retina has a generalizable encoding for time in the natural scene: the precise, low-dimensional representation of time learned from one movie can be used to represent time in a different movie, with up to 17 ms resolution. We then show that static textures and velocity features of a natural movie are synergistic. The retina simultaneously encodes both to establishes a generalizable, low-dimensional representation of time in the natural scene.

使用U-net从视网膜学习低维可概括的自然特征。
许多感觉神经科学的重点是呈现由实验者选择的刺激,因为它们是参数化的,易于采样,并且被认为与生物体的行为相关。然而,在复杂的自然场景中,人们通常不知道这些相关特征是什么。这项工作的重点是利用自然电影的视网膜编码来确定大脑所代表的可能与行为相关的特征。这是禁止参数化一个自然电影和其各自的视网膜编码完全。我们使用自然电影中的时间作为整个场景中演变的一整套功能的代理。然后,我们使用一个任务不可知的深度架构,一个编码器-解码器,来模拟视网膜编码过程,并在压缩的潜在空间中表征其“自然场景中的时间”的表示。在我们的端到端训练中,编码器从大量响应自然电影的蝾螈视网膜神经节细胞中学习压缩的潜在表示,而解码器从压缩的潜在空间中采样以生成适当的未来电影帧。通过比较三部电影中视网膜活动的潜在表征,我们发现视网膜对自然场景中的时间有一个可推广的编码:从一部电影中学习到的精确的低维时间表征可以用来表示不同电影中的时间,分辨率高达17毫秒。然后,我们证明了静态纹理和自然电影的速度特征是协同的。视网膜同时对两者进行编码,以在自然场景中建立一个可概括的、低维的时间表示。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信