{"title":"Optimal causal quantization of Markov Sources with distortion constraints","authors":"S. Yuksel, T. Başar, S. Meyn","doi":"10.1109/ITA.2008.4601015","DOIUrl":null,"url":null,"abstract":"For Markov sources, the structure of optimal causal encoders minimizing the total communication rate subject to a mean-square distortion constraint is studied. The class of sources considered lives in a continuous alphabet, and the encoder is allowed to be variable-rate. Both the finite-horizon and the infinite-horizon problems are considered. In the finite-horizon case, the problem is non-convex, whereas in the infinite-horizon case the problem can be convexified under certain assumptions. For a finite horizon problem, the optimal deterministic causal encoder for a kth-order Markov source uses only the most recent k source symbols and the information available at the receiver, whereas the optimal causal coder for a memoryless source is memoryless. For the infinite-horizon problem, a convex-analytic approach is adopted. Randomized stationary quantizers are suboptimal in the absence of common randomness between the encoder and the decoder. If there is common randomness, the optimal quantizer requires the randomization of at most two deterministic quantizers. In the absence of common randomness, the optimal quantizer is non-stationary and a recurrence-based time-sharing of two deterministic quantizers is optimal. A linear source driven by Gaussian noise is considered. If the process is stable, innovation coding is almost optimal at high-rates, whereas if the source is unstable, then even a high-rate time-invariant innovation coding scheme leads to an unstable estimation process.","PeriodicalId":345196,"journal":{"name":"2008 Information Theory and Applications Workshop","volume":"82 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 Information Theory and Applications Workshop","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITA.2008.4601015","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 11
Abstract
For Markov sources, the structure of optimal causal encoders minimizing the total communication rate subject to a mean-square distortion constraint is studied. The class of sources considered lives in a continuous alphabet, and the encoder is allowed to be variable-rate. Both the finite-horizon and the infinite-horizon problems are considered. In the finite-horizon case, the problem is non-convex, whereas in the infinite-horizon case the problem can be convexified under certain assumptions. For a finite horizon problem, the optimal deterministic causal encoder for a kth-order Markov source uses only the most recent k source symbols and the information available at the receiver, whereas the optimal causal coder for a memoryless source is memoryless. For the infinite-horizon problem, a convex-analytic approach is adopted. Randomized stationary quantizers are suboptimal in the absence of common randomness between the encoder and the decoder. If there is common randomness, the optimal quantizer requires the randomization of at most two deterministic quantizers. In the absence of common randomness, the optimal quantizer is non-stationary and a recurrence-based time-sharing of two deterministic quantizers is optimal. A linear source driven by Gaussian noise is considered. If the process is stable, innovation coding is almost optimal at high-rates, whereas if the source is unstable, then even a high-rate time-invariant innovation coding scheme leads to an unstable estimation process.