{"title":"A character elimination algorithm for lossless data compression","authors":"Mark Hosang","doi":"10.1109/DCC.2002.1000000","DOIUrl":"https://doi.org/10.1109/DCC.2002.1000000","url":null,"abstract":"Summary form only given. We present a detailed description of a lossless compression algorithm intended for use on files with non-uniform character distributions. This algorithm takes advantage of the relatively small distances between character occurrences once we remove the less frequent characters. This allows it to create a compressed version of the file that, when decompressed, is an exact copy of the file that was compressed. We begin by performing a Burrows-Wheeler (1994) Transform (BWT) on the file. The algorithm scans this BWT file to create a character frequency model for the compression phase. To deal with the issue of bit encoding, we write every number as a byte or sequence of bytes to the compressed file and run an arithmetic encoder after the file has been compiled.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133772053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Rate control using conditional mean estimator","authors":"H. Kim, Hyung-Suk Kim, T. Acharya","doi":"10.1109/DCC.2002.1000001","DOIUrl":"https://doi.org/10.1109/DCC.2002.1000001","url":null,"abstract":"Summary form only given. This paper presents a simple, fast and accurate rate control algorithm using conditional mean estimator (nonlinear regression) that plays a central role in estimation theory. Central to nonlinear estimation and stochastic control problems is the determination of the probability density function of the state conditioned on the available data. If this a posteriori density function is known, then an estimate of the state for any performance can be determined. The proposed algorithm measures this conditional mean by estimating a joint probability density function (PDF) using Parzen's window by extending it to multivariate case. We use this window function to estimate a joint PDF using long training data. The training data pick up the joint PDF between the quantization parameter (QP) and the bits spent for each macroblock depending on the sum of absolute differences (SAD) value from motion estimation. Since the SAD information is obtained as by-product of motion estimation, the additional complexity is minimal. We increase the accuracy of this joint PDF by clustering the training data depending on the QP values within admissible ranges. This localization helps understand image characteristics more accurately. Then we apply the adaptive vector quantization to simplify the conditional mean estimation of the rate given the SAD and QP values. This information is stored into three look-up tables for I, P and B pictures. They contain the localized R-D function on macroblock basis. We use these tables to find the optimal QP values in least-mean-square (LMS) sense for a given bit budget of the current frame. We compared our proposed algorithm with the MPEG-4-rate control algorithm (Q2). Simulation results show that the proposed algorithm outperforms the informative MPEG-4 rate control algorithm in terms of reproduced image quality and coding efficiency while requiring much less implementation complexity.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133892333","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"String matching with stopper compression","authors":"J. Rautio, Jani Tanninen, J. Tarhio","doi":"10.1109/DCC.2002.1000012","DOIUrl":"https://doi.org/10.1109/DCC.2002.1000012","url":null,"abstract":"Summary form only given. We consider string searching in compressed texts. We utilize a compression method related to static Huffman compression. Characters are encoded as variable length sequences of base symbols, which consist of a fixed number of bits. Because the length of a code as base symbols varies, we divide base symbols into stoppers and continuers in order to be able to recognize where a new code starts. Stoppers can only be used as the last base symbol of a code. All other base symbols are continuers which can be used anywhere but as the last base symbol of a code. Our searching algorithm is a variation of the Boyer-Moore-Horspool algorithm. The shift function is based on several base symbols in order to achieve longer jumps than the ordinary occurrence heuristic. If four bits are used for base symbols, we apply bytes of eight bits for shift calculation.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"172 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123017794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"New techniques for bounding the channel capacity of read/write isolated memory","authors":"Xuerong Yong, M. Golin","doi":"10.1109/DCC.2002.1000025","DOIUrl":"https://doi.org/10.1109/DCC.2002.1000025","url":null,"abstract":"Summary form only given. A serial binary (0,1) memory is read isolated if no two consecutive positions in the memory may both store 1's; it is write isolated if no two consecutive positions in the memory can be changed during rewriting. Such restrictions have arisen in the contexts of asymmetric error-correcting ternary codes and of rewritable optical discs etc. A read/write isolated memory is a binary, linearly ordered, rewritable storage medium that obeys both the read and write constraints. We introduce new compressed matrix techniques. The new contribution of this paper is to show that it is possible to take advantage of the recursive structures of the transfer matrices to (i) build other matrices of the same size whose eigenvalues yield provably better bounds or (ii) build smaller matrices whose largest eigenvalues are the same as those of the transfer matrices. Thus, it is possible to get the same bounds with less computation. We call these approaches compressed matrix techniques. While technique (ii) was specific to this problem technique (i) is applicable to many other two-dimensional constraint problems.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122750361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Chang-Ming Tsai, Wen-Yan Chang, Chu-Song Chen, G. Tang
{"title":"Compression of 3D objects with multistage color-depth panoramic maps","authors":"Chang-Ming Tsai, Wen-Yan Chang, Chu-Song Chen, G. Tang","doi":"10.1109/DCC.2002.1000018","DOIUrl":"https://doi.org/10.1109/DCC.2002.1000018","url":null,"abstract":"Summary form only given. A new representation method, the multistage color-depth panoramic map (or panomap), is proposed for compressing 3D graphic objects. The idea of the proposed method is to transform a 3D graphic object, including both the shape and color information, into a single image. Existing image compression techniques can then be applied for compressing the panomap structure, which can achieve a highly efficient representation due to its regularity. In our experiments, compressing the color part of a CMP with a lossy method (JPEG) and the depth part with a lossless one (PNG) achieves good reconstruction quality with low bit rates.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129663814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fast motion detection for thin client compression","authors":"B. O. Christiansen, K. Schauser","doi":"10.1109/DCC.2002.999971","DOIUrl":"https://doi.org/10.1109/DCC.2002.999971","url":null,"abstract":"State-of-the-art lossless compression methods have enabled thin client computing across wide area networks which is rapidly gaining momentum. While these compression methods exploit redundancies typically found in synthetic images, none of them exploits large blocks often occurring verbatim in sequences of display updates because detecting movements of such blocks is computationally demanding. In this paper, we present a novel algorithm for detecting block movements in image sequences that is fast enough for interactive logins. We also integrate our algorithm into TCC, the best previously known codec for thin client computing. In the presence of motion, our new codec, TCC-M, typically compresses 50 to 950 times more efficiently than TCC. It also reduces the end-to-end latency of scrolling Web pages at DSL speeds by a factor of 5.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129685120","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Codecell contiguity in optimal fixed-rate and entropy-constrained network scalar quantizers","authors":"M. Effros, D. Muresan","doi":"10.1109/DCC.2002.999969","DOIUrl":"https://doi.org/10.1109/DCC.2002.999969","url":null,"abstract":"We consider the properties of optimal fixed-rate and entropy-constrained scalar quantizers for finite alphabet sources. In particular, we consider conditions under which the optimal scalar quantizer with contiguous codecells achieves performance no worse than the optimal scalar quantizer without the constraint of codecell contiguity. In addition to traditional scalar quantizers, we consider multi-resolution scalar quantizers and multiple description scalar quantizers and also look briefly at codes with decoder side information (Wyner-Ziv codes). While the conditions under which codecell contiguity is consistent with optimality in fixed-rate and entropy-constrained scalar quantization are quite broad, even with the squared error distortion measure, codecell contiguity in fixed-rate and entropy-constrained multi-resolution, multiple description, and Wyner-Ziv scalar quantization can preclude optimality for some sources.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128575638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A parallel algorithm for lossless image compression by block matching","authors":"L. Cinque, S. Agostino, F. Liberati","doi":"10.1109/DCC.2002.999993","DOIUrl":"https://doi.org/10.1109/DCC.2002.999993","url":null,"abstract":"Summary form only given. We show a parallel algorithm using a rectangle greedy matching technique which requires a linear number of processors and O(log(M)log(n)) time on the PRAM EREW model. The algorithm is suitable for practical parallel architectures as a mesh of trees, a pyramid or a multigrid. We implement a sequential procedure which simulates the compression performed by the parallel algorithm and it achieves 95 to 97 percent of the compression of a previous sequential heuristic. To achieve logarithmic time we partition an m/spl times/n image, I, in x/spl times/y rectangular areas where x and y are /spl Theta/(log/sup 1/2 / mn). In parallel for each area, one processor applies the sequential parsing algorithm, so that, in logarithmic time, each area is parsed in rectangles, some of which are monochromatic. Before encoding, we compute larger monochromatic rectangles by merging the ones adjacent on the horizontal boundaries and then on the vertical boundaries, doubling in this way the length and width of each area at each step.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124782555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Interpolation of bandlimited functions from quantized irregular samples","authors":"Z. Cvetković, B. Logan, I. Daubechies","doi":"10.1109/DCC.2002.999981","DOIUrl":"https://doi.org/10.1109/DCC.2002.999981","url":null,"abstract":"The problem of reconstructing a /spl pi/-bandlimited signal f from its quantized samples taken at an irregular sequence of points (t/sub k/)/sub k/spl isin//spl Zopf// arises in oversampled analog-to-digital conversion. The input signal can be reconstructed from the quantized samples (f(t/sub k/))/sub k/spl isin//spl Zopf// by estimating samples (f(n//spl lambda/))/sub n/spl isin//spl Zopf//, where /spl lambda/ is the average uniform density of the sequence (tk)/sub k/spl isin//spl Zopf//, assumed here to be greater than one, followed by linear low-pass filtering. We study three techniques for estimating samples (f(n//spl lambda/))/sub n/spl isin//spl Zopf// from quantized irregular samples (f(t/sub k/))/sub k/spl isin//spl Zopf//, including Lagrangian interpolation, and two other techniques which result in a better overall accuracy of oversampled A/D conversion.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127091102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"On optimal multi-resolution scalar quantization","authors":"Xiaolin Wu, S. Dumitrescu","doi":"10.1109/DCC.2002.999970","DOIUrl":"https://doi.org/10.1109/DCC.2002.999970","url":null,"abstract":"Any scalar quantizer of 2/sup h/ bins, where h is a positive integer, can be structured by a balanced binary quantizer tree T of h levels. Any pruned subtree /spl tau/ of T corresponds to an operational rate R(/spl tau/) and distortion D(/spl tau/) pair. Denote by S/sub n/ the set of all pruned subtrees of n leaf nodes, 1/spl les/n/spl les/2/sup h/. We consider the problem of designing a 2/sup h/-bin scalar quantizer that minimizes the weighted average distortion D~=/spl Sigma//sub n=1//sup 2(h)/ D(/spl tau/)W(n), where W(n) is a weighting function in the size of pruned subtrees (or the resolution of the underlying quantizer). We present an O(hN/sup 3/) algorithm to solve the underlying optimization problem (N is the number of points of the histogram that represents the source probability mass function), and call the resulting quantizer optimal multi-resolution scalar quantizer in the sense that it minimizes a global distortion measure averaged over all quantization resolutions of T. Interestingly, a similar quantizer design problem studied by Brunk et al. (1996) is a special case of our formulation, and can thus be solved exactly and efficiently using our algorithm. Furthermore, we present an algorithm to generate a sequence of 2/sup h/ nested pruned subtrees of T, from the root of T to the entire tree T itself, which minimizes an expected distortion over a range of operational rates. The resulting nested pruned subtree sequence generates an optimized embedded (rate-distortion scalable) code stream with maximum granularity of 2/sup h/ quantization stages, as opposed to existing successively refinable quantizers, such as the popular bit-plane coding scheme, which offer only h stages.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126057867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}