{"title":"分组算法的无损数据压缩","authors":"N. Tadayon, G. Feng, T. Rao, E. Hinds","doi":"10.1109/DCC.1998.672316","DOIUrl":null,"url":null,"abstract":"Summary form only given. There are in fact two main parts in this paper. One is a modification to context-tree weighting algorithm known as CTW, and the other is a new algorithm called grouping. In the CTW method, we consider a binary tree as a context-tree T/sub D/, where each node s in this tree has length l(s) with 0/spl ges/l(s)/spl ges/D for a source generating a sequence of binary digits. There are counts a/sub s/, and b/sub s/, for each node s of T/sub D/ denoting the number of zeros and ones respectively. Each internal node s of the tree has two children 0s and 1s. The root of the tree corresponds to memoryless model /spl lambda/ and each node corresponds to a prefix sequence. The second part of the paper, introduces a new algorithm that considers all different binary trees of length /spl ges/D as complete sets of alphabets for a binary source. By using the KT estimator for each of these alphabet models, we find the probability distribution gained by all different complete extended alphabets. By grouping these models, we define the coding distribution as the average of the probabilities for all the models. We have demonstrated a quick algorithm for this idea and call this approach a grouping algorithm. This approach also breaks the extended alphabet model probability distribution into the non-extended one. Note that the result of this algorithm will produce at most log M(D) more code words than the optimal selection of strings of length at most D, as letters of the alphabet.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Grouping algorithm for lossless data compression\",\"authors\":\"N. Tadayon, G. Feng, T. Rao, E. Hinds\",\"doi\":\"10.1109/DCC.1998.672316\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Summary form only given. There are in fact two main parts in this paper. One is a modification to context-tree weighting algorithm known as CTW, and the other is a new algorithm called grouping. In the CTW method, we consider a binary tree as a context-tree T/sub D/, where each node s in this tree has length l(s) with 0/spl ges/l(s)/spl ges/D for a source generating a sequence of binary digits. There are counts a/sub s/, and b/sub s/, for each node s of T/sub D/ denoting the number of zeros and ones respectively. Each internal node s of the tree has two children 0s and 1s. The root of the tree corresponds to memoryless model /spl lambda/ and each node corresponds to a prefix sequence. The second part of the paper, introduces a new algorithm that considers all different binary trees of length /spl ges/D as complete sets of alphabets for a binary source. By using the KT estimator for each of these alphabet models, we find the probability distribution gained by all different complete extended alphabets. By grouping these models, we define the coding distribution as the average of the probabilities for all the models. We have demonstrated a quick algorithm for this idea and call this approach a grouping algorithm. This approach also breaks the extended alphabet model probability distribution into the non-extended one. Note that the result of this algorithm will produce at most log M(D) more code words than the optimal selection of strings of length at most D, as letters of the alphabet.\",\"PeriodicalId\":191890,\"journal\":{\"name\":\"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)\",\"volume\":\"7 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1998-03-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/DCC.1998.672316\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DCC.1998.672316","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Summary form only given. There are in fact two main parts in this paper. One is a modification to context-tree weighting algorithm known as CTW, and the other is a new algorithm called grouping. In the CTW method, we consider a binary tree as a context-tree T/sub D/, where each node s in this tree has length l(s) with 0/spl ges/l(s)/spl ges/D for a source generating a sequence of binary digits. There are counts a/sub s/, and b/sub s/, for each node s of T/sub D/ denoting the number of zeros and ones respectively. Each internal node s of the tree has two children 0s and 1s. The root of the tree corresponds to memoryless model /spl lambda/ and each node corresponds to a prefix sequence. The second part of the paper, introduces a new algorithm that considers all different binary trees of length /spl ges/D as complete sets of alphabets for a binary source. By using the KT estimator for each of these alphabet models, we find the probability distribution gained by all different complete extended alphabets. By grouping these models, we define the coding distribution as the average of the probabilities for all the models. We have demonstrated a quick algorithm for this idea and call this approach a grouping algorithm. This approach also breaks the extended alphabet model probability distribution into the non-extended one. Note that the result of this algorithm will produce at most log M(D) more code words than the optimal selection of strings of length at most D, as letters of the alphabet.