{"title":"用于从时间序列中提取主成分的外积神经网络","authors":"L. E. Russo","doi":"10.1109/NNSP.1991.239525","DOIUrl":null,"url":null,"abstract":"An outer product neural network architecture has been developed based on subspace concepts. The network is trained by auto-encoding the input exemplars, and will represent the input signal by k-principal components, k being the number of neurons or processing elements in the network. The network is essentially a single linear layer. The weight matrix columns orthonormalize during training. The output signal converges to the projection of the input onto a k-principal component subspace, while the residual signal represents the novelty of the input. An application to extracting sinusoids from a noisy time series is given.<<ETX>>","PeriodicalId":354832,"journal":{"name":"Neural Networks for Signal Processing Proceedings of the 1991 IEEE Workshop","volume":"112 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1991-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"9","resultStr":"{\"title\":\"An outer product neural network for extracting principal components from a time series\",\"authors\":\"L. E. Russo\",\"doi\":\"10.1109/NNSP.1991.239525\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"An outer product neural network architecture has been developed based on subspace concepts. The network is trained by auto-encoding the input exemplars, and will represent the input signal by k-principal components, k being the number of neurons or processing elements in the network. The network is essentially a single linear layer. The weight matrix columns orthonormalize during training. The output signal converges to the projection of the input onto a k-principal component subspace, while the residual signal represents the novelty of the input. An application to extracting sinusoids from a noisy time series is given.<<ETX>>\",\"PeriodicalId\":354832,\"journal\":{\"name\":\"Neural Networks for Signal Processing Proceedings of the 1991 IEEE Workshop\",\"volume\":\"112 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1991-09-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"9\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neural Networks for Signal Processing Proceedings of the 1991 IEEE Workshop\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/NNSP.1991.239525\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks for Signal Processing Proceedings of the 1991 IEEE Workshop","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NNSP.1991.239525","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An outer product neural network for extracting principal components from a time series
An outer product neural network architecture has been developed based on subspace concepts. The network is trained by auto-encoding the input exemplars, and will represent the input signal by k-principal components, k being the number of neurons or processing elements in the network. The network is essentially a single linear layer. The weight matrix columns orthonormalize during training. The output signal converges to the projection of the input onto a k-principal component subspace, while the residual signal represents the novelty of the input. An application to extracting sinusoids from a noisy time series is given.<>