Results 1 
9 of
9
Data Compression
 ACM Computing Surveys
, 1987
"... This paper surveys a variety of data compression methods spanning almost forty years of research, from the work of Shannon, Fano and Huffman in the late 40's to a technique developed in 1986. The aim of data compression is to reduce redundancy in stored or communicated data, thus increasing effectiv ..."
Abstract

Cited by 87 (3 self)
 Add to MetaCart
This paper surveys a variety of data compression methods spanning almost forty years of research, from the work of Shannon, Fano and Huffman in the late 40's to a technique developed in 1986. The aim of data compression is to reduce redundancy in stored or communicated data, thus increasing effective data density. Data compression has important application in the areas of file storage and distributed systems. Concepts from information theory, as they relate to the goals and evaluation of data compression methods, are discussed briefly. A framework for evaluation and comparison of methods is constructed and applied to the algorithms presented. Comparisons of both theoretical and empirical natures are reported and possibilities for future research are suggested. INTRODUCTION Data compression is often referred to as coding, where coding is a very general term encompassing any special representation of data which satisfies a given need. Information theory is defined to be the study of eff...
Adding Compression to a FullText Retrieval System
, 1995
"... We describe the implementation of a data compression scheme as an integral and transparent layer within a fulltext... ..."
Abstract

Cited by 81 (25 self)
 Add to MetaCart
We describe the implementation of a data compression scheme as an integral and transparent layer within a fulltext...
Lossless Compression for Text and Images
 International Journal of High Speed Electronics and Systems
, 1995
"... Most data that is inherently discrete needs to be compressed in such a way that it can be recovered exactly, without any loss. Examples include text of all kinds, experimental results, and statistical databases. Other forms of data may need to be stored exactly, such as imagesparticularly bilevel ..."
Abstract

Cited by 6 (0 self)
 Add to MetaCart
Most data that is inherently discrete needs to be compressed in such a way that it can be recovered exactly, without any loss. Examples include text of all kinds, experimental results, and statistical databases. Other forms of data may need to be stored exactly, such as imagesparticularly bilevel ones, or ones arising in medical and remotesensing applications, or ones that may be required to be certified true for legal reasons. Moreover, during the process of lossy compression, many occasions for lossless compression of coefficients or other information arise. This paper surveys techniques for lossless compression. The process of compression can be broken down into modeling and coding. We provide an extensive discussion of coding techniques, and then introduce methods of modeling that are appropriate for text and images. Standard methods used in popular utilities (in the case of text) and international standards (in the case of images) are described. Keywords Text compression, ima...
Efficient Implementation of the WARMUP Algorithm for the Construction of LengthRestricted Prefix Codes
 in Proceedings of the ALENEX
, 1999
"... . Given an alphabet \Sigma = fa1 ; : : : ; ang with a corresponding list of positive weights fw1 ; : : : ; wng and a length restriction L, the lengthrestricted prefix code problem is to find, a prefix code that minimizes P n i=1 w i l i , where l i , the length of the codeword assigned to a i , ..."
Abstract

Cited by 5 (0 self)
 Add to MetaCart
. Given an alphabet \Sigma = fa1 ; : : : ; ang with a corresponding list of positive weights fw1 ; : : : ; wng and a length restriction L, the lengthrestricted prefix code problem is to find, a prefix code that minimizes P n i=1 w i l i , where l i , the length of the codeword assigned to a i , cannot be greater than L, for i = 1; : : : ; n. In this paper, we present an efficient implementation of the WARMUP algorithm, an approximative method for this problem. The worstcase time complexity of WARMUP is O(n log n +n log wn ), where wn is the greatest weight. However, some experiments with a previous implementation of WARMUP show that it runs in linear time for several practical cases, if the input weights are already sorted. In addition, it often produces optimal codes. The proposed implementation combines two new enhancements to reduce the space usage of WARMUP and to improve its execution time. As a result, it is about ten times faster than the previous implementat...
Comparative Study between Various Algorithms of Data Compression Techniques
"... The spread of computing has led to an explosion in the volume of data to be stored on hard disks and sent over the Internet. This growth has led to a need for "data compression", that is, the ability to reduce the amount of storage or Internet bandwidth required to handle this data. This p ..."
Abstract

Cited by 3 (0 self)
 Add to MetaCart
The spread of computing has led to an explosion in the volume of data to be stored on hard disks and sent over the Internet. This growth has led to a need for "data compression", that is, the ability to reduce the amount of storage or Internet bandwidth required to handle this data. This paper provides a survey of data compression techniques. The focus is on the most prominent data compression schemes, particularly popular.DOC,.TXT,.BMP,.TIF,.GIF, and.JPG files. By using different compression algorithms, we get some results and regarding to these results we suggest the efficient algorithm to be used with a certain type of file to be compressed taking into consideration both the compression ratio and compressed file size.
Two SpaceEconomical Algorithms for Calculating Minimum Redundancy Prefix Codes (Extended Abstract)
 In Proceedings of the DCC
, 1999
"... The minimum redundancy prefix code problem is to determine, for a given list W ={w 1 , ..., w n } of n positive symbol weights, a list L =#` 1 ;:::;` n # of n corresponding integer codeword lengths such that P i=1 2 ,` i # 1 and P i=1 w i ` i is minimized. Let us consider the case where W is already ..."
Abstract

Cited by 1 (1 self)
 Add to MetaCart
The minimum redundancy prefix code problem is to determine, for a given list W ={w 1 , ..., w n } of n positive symbol weights, a list L =#` 1 ;:::;` n # of n corresponding integer codeword lengths such that P i=1 2 ,` i # 1 and P i=1 w i ` i is minimized. Let us consider the case where W is already sorted. In this case, the output list L can be represented by a list M =#m 1 ;:::;m H #, where m ` , for ` = 1;:::;H, denotes the multiplicity of the codeword length ` in L and H is the length of the greatest codeword. Fortunately, H is proved to be O(min{log(1/p 1 ), n}), where p 1 is the smallest symbol probability, given by w 1 = P i=1 w i . In this paper, we present the FLazyHuff and the ELazyHuff algorithms. FLazyHuff runs in O(n) time but requires O(min{H², n}) additional space. On the other hand, ELazyHuff runs in O(n log(n/H)) time, requiring only O(H) additional space. Finally, since our two algorithms have the advantage of not writing at the input buffer during the code calculation, we discuss some applications where this feature is very useful.
LowMemory Adaptive Prefix Coding
, 811
"... In this paper we study the adaptive prefix coding problem in cases where the size of the input alphabet is large. We present an online prefix coding algorithm that uses O(σ 1/λ+ǫ) bits of space for any constants ε> 0, λ> 1, and encodes the string of symbols in O(log log σ) time per symbol in the wor ..."
Abstract
 Add to MetaCart
In this paper we study the adaptive prefix coding problem in cases where the size of the input alphabet is large. We present an online prefix coding algorithm that uses O(σ 1/λ+ǫ) bits of space for any constants ε> 0, λ> 1, and encodes the string of symbols in O(log log σ) time per symbol in the worst case, where σ is the size of the alphabet. The upper bound on the encoding length is λnH(s)+(λln 2+2+ǫ)n+O(σ 1/λ log 2 σ) bits. 1
Data Compression using Huffman based LZW Encoding Technique
"... Data compression is of interest in business data processing, both because of the cost savings it offers and because of the large volume of data manipulated in many business applications. A method and system for transmitting a digital image (i.e., an array of pixels) from a digital data source to a d ..."
Abstract
 Add to MetaCart
Data compression is of interest in business data processing, both because of the cost savings it offers and because of the large volume of data manipulated in many business applications. A method and system for transmitting a digital image (i.e., an array of pixels) from a digital data source to a digital data receiver. More the size of the data be smaller, it provides better transmission speed and saves time. In this communication we always want to transmit data efficiently and noise free. Both the LZW and Huffman data compression methods are lossless in manner. These methods or some versions of them are very common in use of compressing different types of data. Even though on average Huffman gives better compression results, it determines the case in which the LZW performs best and when the compression efficiency gap between the LZW algorithm and its Huffman counterpart is the largest. In the case of Hybrid compression it gives better compression ratio than in single compression. So, at first I wanted to compress original data by Huffman Encoding Technique then by the LZW Encoding Technique.But it did not give better compression ratio than in single LZW compression. At that time I have found that if we compress the data by Huffman first and then by LZW all the cases it gives better compression ratio. Then it named as “Data compression using Huffman based LZW Encoding”. Its compression ratio most of the cases above 2.55 and in some cases it becomes above 3.25 or more. It will provide cheap, reliable and efficient system for data compression in digital communication system.
An Approximation Algorithm for Constructing Error Detecting Prefix Codes
, 2006
"... A kbit Hamming prefix code is a binary code with the following property: for any codeword x and any prefix y of another codeword, both x and y having the same length, the Hamming distance between x and y is at least k. Given an alphabet A = [a1,...,an] with corresponding probabilities [p1,...,pn], ..."
Abstract
 Add to MetaCart
A kbit Hamming prefix code is a binary code with the following property: for any codeword x and any prefix y of another codeword, both x and y having the same length, the Hamming distance between x and y is at least k. Given an alphabet A = [a1,...,an] with corresponding probabilities [p1,...,pn], the kbit Hamming prefix code problem is to find a kbit Hamming prefix code for A with minimum average codeword length ∑n i=1 piℓi, where ℓi is the length of the codeword assigned to ai. In this paper, we propose an approximation algorithm for the 2bit Hamming prefix code problem. Our algorithm spends O(nlog 3 n) time to calculate a 2bit Hamming prefix code with an additive error of at most O(log log log n) bits with respect to the entropy H = − ∑n i=1 pi log2 pi. 1