{"title":"ISI信道容量的渐近逼近","authors":"G. Taricco, J. Boutros","doi":"10.1109/ITA.2014.6804211","DOIUrl":null,"url":null,"abstract":"An asymptotic method to calculate the information rate of an ISI channel is presented in this work. The method is based on an integral representation of the mutual information, which is then calculated by using a saddlepoint approximation along with an asymptotic expansion stemming from the Hubbard-Stratonovich transform. This asymptotic result is evaluated repeatedly to generate a large number of samples required for the Monte-Carlo approximation of the final result. The proposed method has the advantage of being manageable even when the channel memory becomes very large since the complexity grows with polynomial order in the memory length.","PeriodicalId":338302,"journal":{"name":"2014 Information Theory and Applications Workshop (ITA)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2014-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"An asymptotic approximation of the ISI channel capacity\",\"authors\":\"G. Taricco, J. Boutros\",\"doi\":\"10.1109/ITA.2014.6804211\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"An asymptotic method to calculate the information rate of an ISI channel is presented in this work. The method is based on an integral representation of the mutual information, which is then calculated by using a saddlepoint approximation along with an asymptotic expansion stemming from the Hubbard-Stratonovich transform. This asymptotic result is evaluated repeatedly to generate a large number of samples required for the Monte-Carlo approximation of the final result. The proposed method has the advantage of being manageable even when the channel memory becomes very large since the complexity grows with polynomial order in the memory length.\",\"PeriodicalId\":338302,\"journal\":{\"name\":\"2014 Information Theory and Applications Workshop (ITA)\",\"volume\":\"72 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2014-04-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2014 Information Theory and Applications Workshop (ITA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ITA.2014.6804211\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2014 Information Theory and Applications Workshop (ITA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITA.2014.6804211","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An asymptotic approximation of the ISI channel capacity
An asymptotic method to calculate the information rate of an ISI channel is presented in this work. The method is based on an integral representation of the mutual information, which is then calculated by using a saddlepoint approximation along with an asymptotic expansion stemming from the Hubbard-Stratonovich transform. This asymptotic result is evaluated repeatedly to generate a large number of samples required for the Monte-Carlo approximation of the final result. The proposed method has the advantage of being manageable even when the channel memory becomes very large since the complexity grows with polynomial order in the memory length.