You are on page 1of 5

International Journal of Computer Trends and Technology- March to April Issue 2011

An Analysis in various Compression Algorithms


A. Nagarajan, Assistant Professor. Dept of Master of Computer Application (MCA) PSNA College of Engineering & Technology Dindigul. Abstract: Image compression is the application of Data compression on digital images. It is the process of reducing the amount of data needed for storage or transmission of a given piece of information (text, graphics, video, sound, etc.). Data compression stands for compressing data or files containing data so that they can be stored in much less memory space than they had been stored in their original form. In this paper we analysis the various compression methods and their pros & cons. Keywords: JPEG, JPEG2000, RLE, zero trees, set partitioning in hierarchical trees Introduction: Compression is used everywhere in the network applications. All the images you get on the web are compressed, typically in the JPEG or GIF formats, most modems use compression, HDTV will be compressed using MPEG-2, and several file systems automatically compress files when stored, and the rest of us do it by hand., These algorithms used in the real world make heavy use of a wide set of algorithmic tools, including sorting, hash tables, tries, and FFTs. Furthermore, There are two different types of digital graphics system available one is vector and another one is bitmap. JPEG compression only works on bitmap images since vector graphics are not digital images, and cannot be made any smaller. Bitmaps, on the other hand, can be compressed and the process is called lossy compression because when the image is compressed some of the information. Bitmap images are m n matrices where every entry in the picture matrix corresponds to a small square in the picture. Bitmaps come in four main types the first is binary, where the image is an m n matrix with every element in the matrix being either a 0 for black, or a 1 for white, these images are very poor quality, and are almost never used for storing digital images. The second type is intensity images which store a number between 0 and 1, in every entry of the matrix. Intensity images, like binary images, have zero as complete black, and one as absolute white. However, unlike binary images, intensity images have many different shades of gray, corresponding to numbers between zero and one. The next type is indexed images, where there are 256 different colors, which are stored in the image file as the image index. Every element in image matrix has a number between 0-255 which corresponds to a certain color, in the index. The last type is true color, or RGB, where the image is composed of

ISSN:2231-2803

-1-

IJCTT

International Journal of Computer Trends and Technology- March to April Issue 2011 three layered m n matrices, one for each red, green, and blue. Today most pictures are True color, but there are some indexed and intensity images in use. JPEG Compression is enough to compress all three types of images. JPEG: The JPEG compression algorithm initiates to develop compression techniques for the transmission of color and grayscale images. Developed by the Joint Photographic Experts Group of the International Standards Organization, JPEG is a lossy technique that provides best compression rates with complex 24bit images. It achieves its effect by discarding image data that is imperceptible to the human eye, using a technique called Discrete Cosine Transform (DCT). It then applies Huffman encoding to achieve further compression. JPEG contains some basic functionalities, JPEG allows a JPEG decoder to build up and display an image progressively, rather than waiting for the entire image data to be received, and can be useful for applications which need to stream image data. Arithmetic encoding is an extension of baseline JPEG that offers higher compression rates, but is slower and subject to patent. Lossless JPEG uses a different algorithm to baseline JPEG to provide lossless compression. Support for lossless JPEG is very limited. The JPEG specification allows users to set the degree of compression, using an abstract Quality Setting. This provides a trade-off between compression rate and image quality, the higher the setting, the better the quality of the resultant image but at the cost of a larger file size. It is important to note that the Quality Setting is not an absolute value; different JPEG encoders use different scales, although 1 to 100 is typical. In addition, with some implementations, 100 would represent maximum compression rather than maximum image quality. It is also essential to be aware that even the maximum quality setting for baseline JPEG involves some degree of lossy compression. Repeated saving of an image will always lead to increasing degradation of the image quality. JPEG 2000 JPEG 2000 is alternative for the JPEG algorithm, developed by the ISO. It provides for lossy and lossless compression, and uses wavelet compression to achieve higher compression rates with a lower corresponding reduction in image quality. JPEG 2000 may utilize some patented technologies. The JPEG 2000 standard defines a minimum file interchange format in a similar manner to JFIF and SPIFF. Support for JPEG 2000 is now available in many commercial software applications. A given input image or part of the image is sent to a set of wavelet filters, which transform the pixel information into wavelet coefficients, which are then grouped into several sub bands Each sub band contains wavelet coefficients that describe a specific horizontal and vertical spatial frequency range of the entire original image. This means that lower-frequency; lessdetailed information is contained in the first transform level, while more-detailed and higher-frequency information is contained in higher transform levels. For simplicity, only two levels of transform are shown here. The first transform level

ISSN:2231-2803

-2-

IJCTT

International Journal of Computer Trends and Technology- March to April Issue 2011 results in sub bands LH1, HH1, HL1, and LL1. Only sub band LL1 is passed on for further filtering, generating the next transform level and creating sub bands LH2, HH2, HL2, and LL2. Equally sized code blocks, which are essentially bit streams of data, are generated within each sub band. This break-down is necessary for coefficient modeling and coding, and is done on a code-block-by-code-block basis. In essence, the actual compression is achieved by truncating and re-quantizing the bit streams contained in each code block. These bit streams are then optimally truncated using a technique knows as post-compression-rate-control. Code blocks can be accessed independently. Their bit streams are coded with three coding passes per bit plane. This process, called context modeling, is used to assign information about the importance of each individual coefficient bit. The code blocks can then be grouped according to their significance. On the decoding side it is then possible to extract information according to its significance, allowing the most significant information to be seen first. Applications CCTV Security When transmitting or storing picture information, compression must be employed to maintain picture resolution while making best use of limited channel bandwidth. Compression is defined as lossless if full recovery of the original is available from the channel without any loss of information; otherwise, it is lossy. Standards are required to ensure interoperability. JPEG 2000 is the only standard compression scheme that provides for both lossless and lossy compression. As such, it lends itself to applications that require highquality images despite limitations on storage or transmission bandwidths. High Definition At extreme a compression level, JPEG 2000 video starts to blur, but is still quite viewable. MPEG or JPEG artifacts are much more disturbing to the eye, with the picture visibly broken down into small blocks at high compression ratios. The high image quality at medium-to-high bit rates and contents that contain a lot of motion, lack of block artifacts, and high efficiency make JPEG 2000 ideal for high-definition applications, such as digital cinema, High Definition recording systems, and High Definition camera equipment. Internet Image Distribution Progressive coding is another feature of the JPEG 2000 standard, means that the bit stream can be coded in such a way as to contain less-detailed information at the beginning of the stream and more detailed information as the stream progresses. This makes it ideal for network applications especially with large images and low bandwidths as the image can be seen instantly on the decoding side, even with low-speed networks or image databases. The lower sub bands are shown first, and more detail is added as time progresses. The picture thus becomes sharper and more detailed over time, and the entire image does not have to be downloaded before it can be seen. Run Length Algorithm: RLE is a very simple form of data compression in sequences in which the same data value occurs in many consecutive data elements are stored as a single data value and count, rather than

ISSN:2231-2803

-3-

IJCTT

International Journal of Computer Trends and Technology- March to April Issue 2011 as the original run. This is useful on data that contains many such runs, for example, relatively simple graphic images such as icons, line drawings, and animations. It is not useful with files that don't have many runs as it could potentially file size is increase. In our example, let take a screen containing plain black text on a solid white background. There will be many long runs of white pixels in the blank space, and many short runs of black pixels within the text. Let us take a hypothetical single scan line, with B representing a black pixel and W representing white: WWWWWWWWWWWWBW WWWWWWWWWWWBBBWWWW WWWWWWWWWWWWWWWWW WWWBWWWWWWWWWWWWW W If we apply the run-length encoding data compression algorithm to the above hypothetical scan line, we get the following: 12W1B12W3B24W1B14W Interpret this as twelve W's, one B, twelve W's, three B's, etc. The run-length code represents the original 67 characters in only 18. Of course, the actual format used for the storage of images is generally binary rather than ASCII characters like this, but the principle remains the same. Even binary data files can be compressed with this method; file format specifications often dictate repeated bytes in files as padding space. Zero trees follow this data structure. A wavelet coefficient x is said to be insignificant with respect to a given threshold if |x|<T. The zero tree is based on the hypothesis that if a wavelet coefficient at a coarse scale is insignificant with respect to a threshold, then all wavelet coefficients of the same orientation in the same spatial location at the finer scale are likely to be insignificant with respect to the same threshold. More specifically, in a hierarchical sub band system, with the exception of the highest frequency sub bands, ever coefficient at a given scale can be related to a set of coefficients at the next finer scale of similar orientation. The coefficient at the coarse scale is called the parent, and all coefficients corresponding to the same spatial location at the next finer scale of similar orientation are called children. Embedded zero tree wavelet algorithms: The embedded zero tree wavelet algorithms is a simple, yet remarkable effective, image compression algorithm, having the property that the bits in the bit stream are generated in order of importance, yielding a fully embedded code. Using an embedded coding algorithm, an encoder can terminate the encoding at any point thereby allowing a target rate or target distortion metric to be met exactly. Also, given a bit stream, the decoder can cease decoding at any point in the bit stream and still produce exactly the same image that would have been encoded at the bit rate corresponding to the truncated stream. In addition to producing a fully embedded bit stream, EZW consistently produces compression results that are competitive with virtually all known compression algorithms. Set partitioning in hierarchical trees: The compression efficiency of EZW is to some extent due to the use of arithmetic coding. Said and Pearlman have introduced a variant of coding of wavelet coefficients by successive approximation, that even without

ISSN:2231-2803

-4-

IJCTT

International Journal of Computer Trends and Technology- March to April Issue 2011 arithmetic coding outperforms EZW. They call it set partitioning in hierarchical trees (SPIHT). Both EZW and SPIHT are spatial tree-based encoding techniques that exploit magnitude correlation across bands of the decomposition. Each generates a fidelity progressive bit stream by encoding, in turn, each bit plane of quantized dyadic sub band decomposition. Both use a significance test on sets of coefficients to efficiently isolate and encode high magnitude coefficients. However, the crucial parts in the SPIHT coding process are the way the subsets of the wavelet coefficients are partitioned and the significant information is conveyed. One of the main features of this scheme in transmitting the ordering data is that it is based on the fact that the execution path of an algorithm is defined by the results of the comparisons of its branching points. So, if the encoder and decoder have the same sorting algorithm, then the decoder can duplicate the encoder's execution path if it receives the results of the magnitude comparisons. The ordering information can be recovered from the execution path. Conclusion: In this paper we analysis the various compression methods and its advantages as well as drawbacks. Still there is lots of compression algorithm available but here we discussed only few of them. These compression algorithms have major role in networking application. Future Work: Here we discussed various algorithms and its their features. Every method has some drawbacks and peculiarity. In future we are going to develop some new compression technique, which will use in many platforms. References: [1] Guy E. Blelloch -Introduction to Data Compression Computer Science Department Carnegie Mellon University September 25, 2010. [2] Michael W. Marcellin1, Michael J. Gormish, Ali Bilgin, Martin P. BoliekAn Overview of JPEG-2000- IEEE Data Compression Conference. [3] Kamrul Hasan Talukder and Koichi Harada- Haar Wavelet Based Approach for Image Compression and Quality Assessment of Compressed Image. [4] James S. Walker-Wavelet-based Image Compression Sub-chapter of CRC Press book: Transforms and Data Compression. [5] Zixiang Xiong and Kannan Ramchandran- Texas A&M University, University of California [6] A new fast and efficient image codec based on set portioning in hierarchical Trees.- Amir said- IEEE transactions.

ISSN:2231-2803

-5-

IJCTT

You might also like