Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)最新文献

筛选
英文 中文
Preprocessing text to improve compression ratios 预处理文本以提高压缩比
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672295
H. Kruse, A. Mukherjee
{"title":"Preprocessing text to improve compression ratios","authors":"H. Kruse, A. Mukherjee","doi":"10.1109/DCC.1998.672295","DOIUrl":"https://doi.org/10.1109/DCC.1998.672295","url":null,"abstract":"Summary form only given. We discuss the use of a text preprocessing algorithm that can improve the compression ratio of standard data compression algorithms, in particular 'bzip2', when used on text files, by up to 20%. The text preprocessing algorithm uses a static dictionary of the English language that is kept separately from the compressed file. The method in which the dictionary is used by the algorithm to transform the text is based on earlier work of Holger Kruse, Amar Mukherjee (see Proc. Data Comp. Conf., IEEE Comp. Society Press, p.447, 1997). The idea is to replace each word in the input text by a character sequence which encodes the position of the original word in the dictionary. The character sequences used for this encoding are chosen carefully in such a way that specific back-end compression algorithms can often compress these sequences more easily than the original words, increasing the overall compression ratio for the input text. In addition to the original method, this paper describes a variation of the method specifically for the 'bzip2' data compression algorithm. The new method yields an improvements in compression ratio of up to 20% over bzip2. We also describe methods how our algorithm can be used on wide area networks such as the Internet, and in particular how dictionaries can automatically be synchronized and kept up to date in a distributed environment, by using the existing system of URLs, caching and document types, and applying it to dictionaries and text files.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128853686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 46
A multimode context-based lossless wavelet image coder 基于上下文的多模无损小波图像编码器
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672300
Tetra Lindarto
{"title":"A multimode context-based lossless wavelet image coder","authors":"Tetra Lindarto","doi":"10.1109/DCC.1998.672300","DOIUrl":"https://doi.org/10.1109/DCC.1998.672300","url":null,"abstract":"Summary form only given. Currently, the most difficult part of a context-based lossless image compression system is how to determine the contexts. There is a tendency that the techniques for choosing contexts are based on practical experiences. For example, Wu et al. proposed the gradient of the current pixel as the base to choose the contexts for CALIC. Weinberger et al. proposed a similar strategy for LOGO-I. The main idea of this paper is to establish a general concept for the context selection mechanism. One way to do it is to employ an adaptive system that can switch between context schemes if there is enough evidence that switching will give us some performance gains. To be able to do this, some statistics such as the message length for each context scheme and the occurrence of each scheme-have to be collected along the way. Given several possible context selection schemes, the proposed system chooses the scheme that provides the best total message length in the immediate past. The first objective of this switching mechanism is to improve the compression even further. This new concept is combined with wavelet transformations to form a new lossless image compression system. Golomb-Rice code is used to encode the symbols, to reduce the overhead in updating the statistics and actually encoding the symbols. The current results of the experiments indicate that switching can improve the performance only slightly. This is due to the fact that each possible context performs almost the same. The overall results are slightly worse than CALIC or LOGO.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125015109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Breakpoint skeletal representation and compression of document images 断点骨架表示和文档图像的压缩
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672317
David Tam, W. Barrett, B. Morse, Eric N. Mortensen
{"title":"Breakpoint skeletal representation and compression of document images","authors":"David Tam, W. Barrett, B. Morse, Eric N. Mortensen","doi":"10.1109/DCC.1998.672317","DOIUrl":"https://doi.org/10.1109/DCC.1998.672317","url":null,"abstract":"Summary form only given. We present a new method for representation and (lossy) compression of bitonal document images. The technique extracts a skeletal medial axis from each object using a true Euclidean distance map of the image and then finds piecewise linear breakpoints in the skeleton to create a breakpoint skeletal representation, b.p.s. The b.p.s. is encoded for each object as a set of triples. The original binary object is reconstructed by first reconstructing the skeleton using linear interpolation between breakpoints and then fractionally dilating each point on the skeleton with the (linearly interpolated) radius, r/sub i/. For noninteger r/sub i/ fractional dilation provides a natural antialiasing in the reconstructed image. Breakpoints can be extracted to preserve fine detail or a more coarse representation by tightening or relaxing the pruning radius respectively. If, in extracting breakpoints, the pruning radius is set to zero, the reconstruction is almost lossless, but the compression is worse.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134115697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
The context trees of block sorting compression 块排序压缩的上下文树
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672147
N. Larsson
{"title":"The context trees of block sorting compression","authors":"N. Larsson","doi":"10.1109/DCC.1998.672147","DOIUrl":"https://doi.org/10.1109/DCC.1998.672147","url":null,"abstract":"The Burrows-Wheeler (1994) transform (BWT) and block sorting compression are closely related to the context trees of PPM. The usual approach of treating BWT as merely a permutation is not able to fully exploit this relation. We show that an explicit context tree for BWT can be efficiently generated by taking a subset of the corresponding suffix tree, identifying the central problems in exploiting its structure, and tracing the influence of the context tree on the common move-to-front schemes. We experimentally obtain limits for compression using the constructed trees, and, as an attempt at utilizing the full context tree, present a compression scheme that represents the context tree explicitly as part of the compressed data. We argue that a conscious treatment of the context tree should be able to achieve the full compression performance of PPM while maintaining the computational efficiency of BWT. Thus, BWT with explicit context trees is a strong candidate for powerful general compression, especially for large data files.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"247 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132689842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 52
Multiple pattern matching in LZW compressed text LZW压缩文本中的多模式匹配
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672136
T. Kida, M. Takeda, A. Shinohara, Masamichi Miyazaki, S. Arikawa
{"title":"Multiple pattern matching in LZW compressed text","authors":"T. Kida, M. Takeda, A. Shinohara, Masamichi Miyazaki, S. Arikawa","doi":"10.1109/DCC.1998.672136","DOIUrl":"https://doi.org/10.1109/DCC.1998.672136","url":null,"abstract":"We address the problem of searching in LZW compressed text directly, and present a new algorithm for finding multiple patterns by simulating the move of the Aho-Corasick (1975) pattern matching machine. The new algorithm finds all occurrences of multiple patterns whereas the algorithm proposed by Amir, Benson, and Farach (see Journal of Computer and System Sciences, vol.52, p.299-307, 1996) finds only the first occurrence of a single pattern. The new algorithm runs in O(n+m/sup 2/+r/sub a/) time using O(n+m/sup 2/) space, where n is the length of the compressed text, m is the length of the total length of the patterns, and r is the number of occurrences of the patterns. We implemented a simple version of the algorithm, and showed that it is approximately twice faster than a decompression followed by a search using the Aho-Corasick machine.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"301 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132136775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 69
Efficient algorithms for optimal video transmission 优化视频传输的高效算法
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672151
D. Kozen, Y. Minsky, B. Smith
{"title":"Efficient algorithms for optimal video transmission","authors":"D. Kozen, Y. Minsky, B. Smith","doi":"10.1109/DCC.1998.672151","DOIUrl":"https://doi.org/10.1109/DCC.1998.672151","url":null,"abstract":"This paper addresses the problem of sending an MPEG-encoded video stream over a channel of limited bandwidth. When there is insufficient bandwidth available for the rate at which the sequence was encoded, some data must be dropped. In this paper we give fast algorithms to determine a prioritization of the data that optimizes the visual quality of the received video sequence in the sense that the maximum gap of unplayable frames is minimized. Our results are obtained in a new model of encoded video data that is applicable to MPEG and other encoding technologies. The model identifies a certain key relationship between the play order and dependence order of frames that allows fast determination of optimal send orders by dynamic programming.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132335697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Fast wavelet packet image compression 快速小波包图像压缩
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672305
François G. Meyer, A. Averbuch, J. Strömberg, R. Coifman
{"title":"Fast wavelet packet image compression","authors":"François G. Meyer, A. Averbuch, J. Strömberg, R. Coifman","doi":"10.1109/DCC.1998.672305","DOIUrl":"https://doi.org/10.1109/DCC.1998.672305","url":null,"abstract":"Summary form only given. Presents a new fast wavelet packet compression algorithm that encodes very efficiently textured images. This fast wavelet packet compression technique relies on four stages: 1. Very fast convolution and decimation of the image with factorized filters. 2. Selection of a best basis in a large library of waveforms. The best basis is that basis which is best adapted to the content of the image. 3. Scanning of the wavelet packet coefficients by increasing frequency. This organization yields sequences of coefficients with a rapid decay. 4. Successive embedded approximation quantization, and entropy coding of the coefficients. We implemented the wavelet packet coder and decoder, and actual bit streams were created for each experiment. Our implementation used the 7-9 biorthogonal filters. We present the results of the algorithm, using the test image 512 x 512 Barbara. In order to evaluate the performance of the algorithm, we compared our algorithm to the SPIHT wavelet coder of Said and Pearlman (1996).","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129257810","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Pattern matching in text compressed with the ID heuristic 使用ID启发式压缩文本中的模式匹配
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672137
Piera Barcaccia, A. Cresti, S. Agostino
{"title":"Pattern matching in text compressed with the ID heuristic","authors":"Piera Barcaccia, A. Cresti, S. Agostino","doi":"10.1109/DCC.1998.672137","DOIUrl":"https://doi.org/10.1109/DCC.1998.672137","url":null,"abstract":"We show an O(m+t) space algorithm to find all the occurrences of a pattern in a text compressed with the ID heuristic that runs in time O(n(m+t)), where m is the pattern length, n is the size of the compressed text and 1 is the maximum target length.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131775011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
On optimality of variants of the block sorting compression 块排序压缩变体的最优性
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672312
K. Sadakane
{"title":"On optimality of variants of the block sorting compression","authors":"K. Sadakane","doi":"10.1109/DCC.1998.672312","DOIUrl":"https://doi.org/10.1109/DCC.1998.672312","url":null,"abstract":"Summary form only given. Block sorting uses the Burrows-Wheeler transformation (BWT) which permutes an input string. The permutation is defined by the lexicographic order of contexts of symbols. If we assume that symbol probability is defined by preceding k symbols called context, symbols whose contexts are the same are collected in consecutive regions after the BWT. Sadakane (1997) proposed a variant of the block sorting and it is asymptotically optimal for any finite-order Markov source if permutation of symbols whose contexts are the same is random. However, the variant encodes 1 symbols as a block and therefore it is not practical because 1 is large. We propose two compression schemes not using blocks but encoding symbols one by one by using arithmetic codes. The move-to-front transformation is not used. The former encodes symbols by different codes defined by symbol frequencies in contexts. It is asymptotically optimal for k-th order Markov sources. However, it is available only if the order k of the source is already known. The latter divides the permuted string into many parts and encodes symbols using different arithmetic codes by the parts. Each part, has symbols whose contexts are the same. If the permutation is random, the scheme is asymptotically optimal for any finite-order Markov source. The permutation in the BWT is not completely random. However, we conjecture that the permuted string is memoryless and our schemes work.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124895533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Simple pre-processors significantly improve LZ 1 compression 简单的预处理器显著改善了lz1压缩
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225) Pub Date : 1998-03-30 DOI: 10.1109/DCC.1998.672261
D. J. Craft
{"title":"Simple pre-processors significantly improve LZ 1 compression","authors":"D. J. Craft","doi":"10.1109/DCC.1998.672261","DOIUrl":"https://doi.org/10.1109/DCC.1998.672261","url":null,"abstract":"Summary form only given. The effectiveness of the LZ 1 class of lossless adaptive data compression algorithms can, for many different types of data, be significantly improved by employing a dual stage compression/decompression process. A pre-processing stage first re-codes the input data stream in such a way as to make it more amenable to subsequent LZ 1 compression. To decode the data, the inverse post-processing function is applied to the output from an LZ 1 decompressor, thus regenerating the original.","PeriodicalId":191890,"journal":{"name":"Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125409484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信