{"title":"序列协方差矩阵估计及其在减轻灾难性遗忘中的应用","authors":"Tomer Lancewicki, Benjamin Goodrich, I. Arel","doi":"10.1109/ICMLA.2015.109","DOIUrl":null,"url":null,"abstract":"Catastrophic forgetting is a problem encountered with neural networks as well as other learning systems whereby past representations are lost as new representations are learned. It has been shown that catastrophic forgetting can be mitigated in neural networks by using a neuron selection technique, dubbed \"cluster-select,\" which performs online clustering over the network inputs to partition the network such that only a subset of neurons are used for a given input vector. Cluster-select can benefit by using Mahalanobis distance which relies on an inverse covariance estimate. Unfortunately, covariance estimation is problematic when lacking a very large number of samples relative to the number of input dimensions. One way to tackle this problem is through the use of a shrinkage estimator that offers a compromise between the sample covariance matrix and a well-conditioned matrix with the aim of minimizing the mean-squared error (MSE). In online environments, such as those in which catastrophic forgetting can occur, data arrives sequentially, requiring the covariance matrix to be estimated sequentially. Therefore, in this work we derive sequential update rules for the shrinkage estimator and approximate it's related inverse. The online covariance estimator is applied to the cluster-select technique with results that demonstrate further improvements in terms of effectively mitigating catastrophic forgetting.","PeriodicalId":288427,"journal":{"name":"2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"Sequential Covariance-Matrix Estimation with Application to Mitigating Catastrophic Forgetting\",\"authors\":\"Tomer Lancewicki, Benjamin Goodrich, I. Arel\",\"doi\":\"10.1109/ICMLA.2015.109\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Catastrophic forgetting is a problem encountered with neural networks as well as other learning systems whereby past representations are lost as new representations are learned. It has been shown that catastrophic forgetting can be mitigated in neural networks by using a neuron selection technique, dubbed \\\"cluster-select,\\\" which performs online clustering over the network inputs to partition the network such that only a subset of neurons are used for a given input vector. Cluster-select can benefit by using Mahalanobis distance which relies on an inverse covariance estimate. Unfortunately, covariance estimation is problematic when lacking a very large number of samples relative to the number of input dimensions. One way to tackle this problem is through the use of a shrinkage estimator that offers a compromise between the sample covariance matrix and a well-conditioned matrix with the aim of minimizing the mean-squared error (MSE). In online environments, such as those in which catastrophic forgetting can occur, data arrives sequentially, requiring the covariance matrix to be estimated sequentially. Therefore, in this work we derive sequential update rules for the shrinkage estimator and approximate it's related inverse. The online covariance estimator is applied to the cluster-select technique with results that demonstrate further improvements in terms of effectively mitigating catastrophic forgetting.\",\"PeriodicalId\":288427,\"journal\":{\"name\":\"2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA)\",\"volume\":\"32 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2015-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICMLA.2015.109\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLA.2015.109","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Sequential Covariance-Matrix Estimation with Application to Mitigating Catastrophic Forgetting
Catastrophic forgetting is a problem encountered with neural networks as well as other learning systems whereby past representations are lost as new representations are learned. It has been shown that catastrophic forgetting can be mitigated in neural networks by using a neuron selection technique, dubbed "cluster-select," which performs online clustering over the network inputs to partition the network such that only a subset of neurons are used for a given input vector. Cluster-select can benefit by using Mahalanobis distance which relies on an inverse covariance estimate. Unfortunately, covariance estimation is problematic when lacking a very large number of samples relative to the number of input dimensions. One way to tackle this problem is through the use of a shrinkage estimator that offers a compromise between the sample covariance matrix and a well-conditioned matrix with the aim of minimizing the mean-squared error (MSE). In online environments, such as those in which catastrophic forgetting can occur, data arrives sequentially, requiring the covariance matrix to be estimated sequentially. Therefore, in this work we derive sequential update rules for the shrinkage estimator and approximate it's related inverse. The online covariance estimator is applied to the cluster-select technique with results that demonstrate further improvements in terms of effectively mitigating catastrophic forgetting.