{"title":"Bit allocation in sub-linear time and the multiple-choice knapsack problem","authors":"A. Mohr","doi":"10.1109/DCC.2002.999973","DOIUrl":"https://doi.org/10.1109/DCC.2002.999973","url":null,"abstract":"We show that the problem of optimal bit allocation among a set of independent discrete quantizers given a budget constraint is equivalent to the multiple choice knapsack problem (MCKP). This result has three implications: first, it provides a trivial proof that the problem of optimal bit allocation is NP-hard and that its related decision problem is NP-complete; second, it unifies research into solving these problems that has to date been done independently in the data compression community and the operations research community; third, many practical algorithms for approximating the optimal solution to MCKP can be used for bit allocation. We implement the GBFOS, partition-search, and Dudzinski-Walukiewicz algorithms and compare their running times for a variety of problem sizes.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124343901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An analysis of perceptual artifacts in MPEG scalable audio coding","authors":"C. Creusere","doi":"10.1109/DCC.2002.999953","DOIUrl":"https://doi.org/10.1109/DCC.2002.999953","url":null,"abstract":"We study coding artifacts in MPEG-compressed scalable audio. Specifically, we consider the MPEG advanced audio coder (AAC) using bit slice scalable arithmetic coding (BSAC) as implemented in the MPEG 4 reference software. First, we perform human subjective testing using the comparison category rating (CCR) approach, quantitatively comparing the performance of scalable BSAC with the nonscalable TwinVQ and AAC algorithms. This testing indicates that scalable BSAC performs very poorly relative to TwinVQ at the lowest bitrate considered (16 kb/s), largely because of an annoying and seemingly random mid-range tonal signal that is superimposed onto the desired output. In order to understand better and quantify perceptually the various forms of distortion introduced into compressed audio at low bit rates, we apply two analysis techniques: Reng probing and time-frequency decomposition. The Reng probing technique is capable of separating the linear time-invariant component of a multirate system from its nonlinear and periodically time-varying components. Using this technique, we conclude that aliasing is probably not the cause of the annoying tonal signal; instead, time-frequency analysis indicates that its cause is most likely suboptimal bit allocation.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132494301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Extended Golomb codes for binary Markov sources","authors":"E. S. Hong, R. Ladner","doi":"10.1109/DCC.2002.999999","DOIUrl":"https://doi.org/10.1109/DCC.2002.999999","url":null,"abstract":"Summary form only given. Elementary Golomb codes have been widely used for compressing correlated binary sources. We study the theoretical bit-rate performance of two different Golomb coding methods on binary Markov sources: the sequential coding method, and the interleaved coding method. Although the theoretical bit-rate performance for these codes on on i.i.d. sources is known, to the best of our knowledge, theoretical performance results for elementary Golomb codes on correlated Markov sources have not been published.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128546934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
K. H. Randall, Raymie Stata, J. Wiener, Rajiv Wickremesinghe
{"title":"The Link Database: fast access to graphs of the Web","authors":"K. H. Randall, Raymie Stata, J. Wiener, Rajiv Wickremesinghe","doi":"10.1109/DCC.2002.999950","DOIUrl":"https://doi.org/10.1109/DCC.2002.999950","url":null,"abstract":"The Connectivity Server is a special-purpose database whose schema models the Web as a graph: a set of nodes (URL) connected by directed edges (hyperlinks). The Link Database provides fast access to the hyperlinks. To support easy implementation of a wide range of graph algorithms we have found it important to fit the Link Database into RAM. In the first version of the Link Database, we achieved this fit by using machines with lots of memory (8 GB), and storing each hyperlink in 32 bits. However, this approach was limited to roughly 100 million Web pages. This paper presents techniques to compress the links to accommodate larger graphs. Our techniques combine well-known compression methods with methods that depend on the properties of the Web graph. The first compression technique takes advantage of the fact that most hyperlinks on most Web pages point to other pages on the same host as the page itself. The second technique takes advantage of the fact that many pages on the same host share hyperlinks, that is, they tend to point to a common set of pages. Together, these techniques reduce space requirements to under 6 bits per link. While (de)compression adds latency to the hyperlink access time, we can still compute the strongly connected components of a 6 billion-edge graph in 22 minutes and run applications such as Kleinberg's HITS in real time. This paper describes our techniques for compressing the Link Database, and provides performance numbers for compression ratios and decompression speed.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115153502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Turbo source coding: a noise-robust approach to data compression","authors":"P. Mitran, J. Bajcsy","doi":"10.1109/DCC.2002.1000008","DOIUrl":"https://doi.org/10.1109/DCC.2002.1000008","url":null,"abstract":"Summary form only given. All traditional data compression techniques, such as Huffman coding, the Lempel-Ziv algorithm, run-length limited coding, Tunstall coding and arithmetic coding are highly susceptible to residual channel errors and noise. We have previously proposed the use of parallel concatenated codes and iterative decoding for fixed-length to fixed-length source coding, i.e., turbo coding for data compression purposes. The work presented here extends these results and also considers the case when decompression must be done from compressed data corrupted by additive white Gaussian noise (AWGN).","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123576091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Computing rate-distortion optimized policies for streaming media to wireless clients","authors":"Jacob Chakareski, B. Aazhang, P. Chou","doi":"10.1109/DCC.2002.999943","DOIUrl":"https://doi.org/10.1109/DCC.2002.999943","url":null,"abstract":"We consider the problem of streaming packetized media over the Internet from a server through a base station to a wireless client, in a rate-distortion optimized way. For error control, we employ an incremental redundancy scheme, in which the server can incrementally transmit parity packets in response to negative acknowledgements fed back from the client. Computing the optimal transmission policy for the server involves estimation of the probability that a single packet will be communicated in error as a function of the expected redundancy (or cost) used to communicate the packet. In this paper, we show how to compute this error-cost function, and thereby optimize the server's transmission policy, in this scenario.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123815038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Compression with side information using turbo codes","authors":"A. Aaron, B. Girod","doi":"10.1109/DCC.2002.999963","DOIUrl":"https://doi.org/10.1109/DCC.2002.999963","url":null,"abstract":"We show that turbo codes can come close to the Slepian-Wolf bound in lossless distributed source coding. In the asymmetric scenario considered, X and Y are statistically dependent signals and X is encoded with no knowledge of Y. However, Y is known as side information at the decoder. We use a system based on turbo codes to send X at a rate close to H(X|Y). We apply our system to binary sequences and simulations show performance close to the information-theoretic limit. For distributed source coding of Gaussian sequences, our results show significant improvement over previous work. The scheme also performs well for joint source-channel coding.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115037104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
N. Ponomarenko, K. Egiazarian, V. Lukin, J. Astola
{"title":"Compression of image block means for non-equal block size partition schemes using Delaunay triangulation and prediction","authors":"N. Ponomarenko, K. Egiazarian, V. Lukin, J. Astola","doi":"10.1109/DCC.2002.1000011","DOIUrl":"https://doi.org/10.1109/DCC.2002.1000011","url":null,"abstract":"Summary form only given. An approach based on applying Delaunay triangulation to compression of mean values of image blocks that have non-identical shape and size is proposed. It can be useful for image compression methods that require the use of image partition schemes with non-equal block size like fractal and DCT-based image coding. Several methods of block mean value coding are considered. In particular, the drawbacks of using quantization with further redundancy elimination by entropy coders are discussed. Another considered method is the forming of the block mean value image and its further compression by lossy coders. Finally, the motivations in favor of Delaunay triangulation application to block mean value image coding are presented.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132383995","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
W. Tham, Sandra I. Woolley, Sue Cribbs, Don Anderson
{"title":"Diagnostically lossless compression of pipeline inspection data","authors":"W. Tham, Sandra I. Woolley, Sue Cribbs, Don Anderson","doi":"10.1109/DCC.2002.1000016","DOIUrl":"https://doi.org/10.1109/DCC.2002.1000016","url":null,"abstract":"Summary form only given. All pipelines are subject to corrosion and require inspection in accordance with regulatory requirements to ensure human safety. Intelligent pipeline inspection gauges (PIGs) have provided reliable online inspection of pipelines, supplying operators with detailed information about pipeline condition. We present a method for the diagnostically lossless compression of pipeline inspection data and discuss important pipeline features, e.g. welds, cracks and erosion objects. The dataset, transverse field inspection (TFI) data, is a new type of pipeline inspection data in contrast to the traditional magnetic flux leakage (MFL) inspection data. The nature of the data makes feature preservation essential. TFI pipeline features have been collected, classified and analysed and examples are shown. Feature detection is desirable in order to identify regions of diagnostic interest. Incorporation of region-of-interest (ROI) into the SPIHT encoding scheme enables the allocation of a greater proportion of the total allowance of bits to the regions of the image identified as diagnostically significant. Our quality assessment is based on the preservation of important defect parameters to ensure diagnostically lossless performance. We present results comparing performance between ROI SPIHT and non-ROI SPIHT.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132589194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Quantizers with uniform encoders and channel optimized decoders","authors":"B. Farber, K. Zeger","doi":"10.1109/DCC.2002.999967","DOIUrl":"https://doi.org/10.1109/DCC.2002.999967","url":null,"abstract":"Scalar quantizers with uniform encoders and channel optimized decoders are studied for uniform sources and binary symmetric channels. It is shown that the natural binary code and folded binary code induce point density functions that are uniform on proper subintervals of the source support, whereas the Gray code does not induce a point density function. The mean squared errors for the natural binary code, folded binary code, Gray code, and for randomly chosen index assignments are calculated and the natural binary code is shown to be mean squared optimal among all possible index assignments. In contrast, it is shown that almost all index assignments perform poorly and have degenerate codebooks.","PeriodicalId":420897,"journal":{"name":"Proceedings DCC 2002. Data Compression Conference","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133463699","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}