{"title":"源识别和压缩混合数据从有限的观测","authors":"A. Abdi, F. Fekri","doi":"10.1109/ITWF.2015.7360728","DOIUrl":null,"url":null,"abstract":"We consider the problem of the identification of a mixture of ergodic stationary sources from a limited number of finite-length observations of a mixture. We propose an algorithm based on Bayesian Information Criterion and Expectation Maximization to identify the sources' models and estimate the mixture parameters. Based on this algorithm, the sources' distributions can be computed and used for nearly optimal memory-assisted coding of the sequences generated by the mixture. Further, we provide upper and lower bounds on the entropy of the mixture source and show that it converges to the upper bound as the length of the sequences increases and derive the convergence rate for the per-symbol entropy of the mixture of finite memory sources.","PeriodicalId":281890,"journal":{"name":"2015 IEEE Information Theory Workshop - Fall (ITW)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Source identification and compression of mixture data from finite observations\",\"authors\":\"A. Abdi, F. Fekri\",\"doi\":\"10.1109/ITWF.2015.7360728\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We consider the problem of the identification of a mixture of ergodic stationary sources from a limited number of finite-length observations of a mixture. We propose an algorithm based on Bayesian Information Criterion and Expectation Maximization to identify the sources' models and estimate the mixture parameters. Based on this algorithm, the sources' distributions can be computed and used for nearly optimal memory-assisted coding of the sequences generated by the mixture. Further, we provide upper and lower bounds on the entropy of the mixture source and show that it converges to the upper bound as the length of the sequences increases and derive the convergence rate for the per-symbol entropy of the mixture of finite memory sources.\",\"PeriodicalId\":281890,\"journal\":{\"name\":\"2015 IEEE Information Theory Workshop - Fall (ITW)\",\"volume\":\"43 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2015-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2015 IEEE Information Theory Workshop - Fall (ITW)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ITWF.2015.7360728\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 IEEE Information Theory Workshop - Fall (ITW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITWF.2015.7360728","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Source identification and compression of mixture data from finite observations
We consider the problem of the identification of a mixture of ergodic stationary sources from a limited number of finite-length observations of a mixture. We propose an algorithm based on Bayesian Information Criterion and Expectation Maximization to identify the sources' models and estimate the mixture parameters. Based on this algorithm, the sources' distributions can be computed and used for nearly optimal memory-assisted coding of the sequences generated by the mixture. Further, we provide upper and lower bounds on the entropy of the mixture source and show that it converges to the upper bound as the length of the sequences increases and derive the convergence rate for the per-symbol entropy of the mixture of finite memory sources.