Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Data Compression

Download as ppt, pdf, or txt
Download as ppt, pdf, or txt
You are on page 1of 23

A SEMINAR ON DATA COMPRESSION

PRESENTED BY:PRANAV KUMAR ROLL NO:0901289347

CONTENTS
Introduction What ,when Some question Uses Major steps Type of data compression disadvantages conclusion

COMPRESSION: A BASIC IDEA


Compression is the technique in which the original data is converted to a format that requires fewer bits, so that the data can be stored or transmitted more efficiently . As name implies, makes your data smaller, saving space The shrinking of the original size of the data to a compressed form is known as the compression of data. The inverse of the compression process is decompression. If it produces an exact replica of the original data then the compression is lossless. Lossy compression, usually applied to image data, does not allow reproduction of an exact replica of the original image, but has a higher compression ratio. compression ratio = compressed form / original size. We are more repetitive than we think - text often compresses over 50% A device (software or hardware) that compresses data is often know as an encoder or coder, whereas a device that decompresses data is known as a decoder.

OBJECTIVES:
The objectives are: To store the data in a compressed form in which the quality of the data will not be degraded. To transfer the compressed data easily and efficiently and read the data easily by using the decompression technique

Why do we need compression?


The high bit rates that result from the various types of digital video make their transmission through their intended channels very difficult. Even entertainment video with modest frame rates and dimensions would require bandwidth and storage space far in excess of that available from CD-ROM.

Why do we need compression? contd..

Finally, even if the storage and transportation problems of digital video were overcome, the processing power needed to manage such volumes of data would make the receiver hardware very expensive.

USES OF DATA
COMPRESSION : More and more data is being stored electronically. Digital video libraries, for example, contain vast amounts of data, and compression allows costeffective storage of the data.

New technology has allowed the possibility of interactive digital television and the demand is for high-quality transmissions, a wide selection of programs to choose from and inexpensive hardware. But for digital television to be a success, it must use data compression . Data compression reduces the number of bits required to represent or transmit information.

MAJOR STEPS

Uncompress Preparation Quantization Entropy Compress dataData Encoding

o Preparation:-It include analog to digital conversion and generating appropriate digital representation of the information. An image is divided into blocks of 8/8 pixels, and represented by affix no. of bit per pixel. o Processing:-It is 1st stage of compression process which make use sophisticated algorithms.

o Quantization:-It is the result of previous step. It specifies the granularity of the mapping of real number into integer number. This process results in a reduction of precision. o Entropy encoding: - It is the last step. It compresses a sequential digital data stream without loss. For ex:compress sequence of zeroes specifying the no. of occurrence.

TYPES OF DATA COMPRESSION Entropy encoding -- lossless. Data considered a simple digital sequence and semantics of data are ignored.
: Source encoding -- lossy. Takes semantics of data into account. Amount of compression depends on data contents.

Hybrid encoding -- combination of entropy and source. Most multimedia systems use these.

Run Length Encoding:


Run Length Encoding is a compression technique that replaces consecutive occurrences of a symbol with the symbol followed by the number of times it is repeated. For example, the string 111110000003355 could be represented by 15063252.

Relative Encoding:
Relative encoding is a transmission technique that attempts to improve efficiency by transmitting the difference between each value and its predecessor, in place of the value itself. Thus the values 15106433003 would be transmitted as 1+4-4-1+6-2-1+03+0+3.

Huffman Coding:
The total number of bits required to transmit the data can be considerably less than the number required if the fixed length representation is used. Huffman coding is particularly effective where the data are dominated by a small number of symbols. Huffman coding is a popular compression technique that assigns variable length codes (VLC) to symbols, so that the most frequently occurring symbols have the shortest codes. On decompression the symbols are reassigned their original fixed length codes. When used to compress text, for example, variable length codes are used in place of ASCII codes, and the most common characters, usually space, e, and t are assigned the shortest codes.

Arithmetic Coding:
Arithmetic coding is usually more efficient than the more popular Huffman technique. Although more efficient than Huffman coding, arithmetic coding is more complex.

INTRAFRAME COMPRESSION:
Intraframe compression is compression applied to still images, such as photographs and diagrams. Intraframe compression techniques can be applied to individual frames of a video sequence.

Sub-sampling of images usually takes place in one of two ways. In


the first, the original image is copied but only a fraction of the pixels from the original are used.
Alternatively, sub-sampling can be implemented by calculating the average pixel value for each group of several pixels, and then substituting this average in the appropriate place in the approximated image.

LZ-77 ENCODING
Good as they are, Huffman and arithmetic coding are not perfect for encoding text because they don't capture the higherorder relationships between words and phrases. There is a simple, clever, and effective approach to compressing text known as "LZ-77", which uses the redundant nature of text to provide compression.

VECTOR QUANTIZATION: Vector quantization is a more complex form of quantization that first divides the input data stream into blocks. A pre-defined table contains a set of patterns for blocks and each block is coded using the pattern from the table that is most similar. If the number of quantization levels (i.e. blocks in the table) is very small, as in Figure, then the compression will be lossy. Because images often contain many repeated sections vector quantization can be quite successful for image compression

LOSSLESS DATA COMPRESSION


Run-Length Coding Runs (sequences) of data are stored as a single value and count, rather than the individual run. Example: This: WWWWWWWWWWWWBWWWWWWWWW WWWBBBWWWWWWWWWWWWWWWWW WWWWWWWBWWWWWWWWWWWWWW Becomes: 12WB12W3B24WB14W

COARSE QUANTIZATION:
It is similar to sub-sampling in that information is discarded, but the compression is accomplished by reducing the numbers of bits used to describe each pixel, rather than reducing the number of pixels. Each pixel is reassigned an alternative value and the number of alternate values is less than that in the original image. Quantization where the number of ranges is small is known as coarse quantization.

INTERFRAME COMPRESSION:
Interframe compression is compression applied to a sequence of video frames, rather than a single image. In general, relatively little changes from one video frame to the next. Interframe compression exploits the similarities between successive frames, to reduce the volume of data required to describe the sequence.

Difference Coding
Block Based Difference Coding
Block

Based Motion Compensation

LOSSY COMPRESSION
Data is lost, but not too much: Audio.Video. Still images, medical images, photographs. Compression ratios of 10:1 often yield quite Major techniques include: Vector Quantization ,Block transforms, Standards JPEG, JPEG 2000, MPEG (1, 2, 4, 7).

DISADVANTAGES
Some technique are there by which data can compress efficiently. But there is a chance of losses data.

CONCLUSION:
Improving television with larger screens and better resolution requires a huge increase in transmission bit rates. The bit rates are, however, limited by the available broadcast spectrum or network connection. The only resource is lossy image compression, most commonly JPEG, MPEG-2. Lossy by name and lossy by nature. The more the image is compressed, using lossy methods, the worse the image quality. Because random noise in the images is interpreted as movement any increase in noise increases the transmissions. The cleaner the images and the less noise, the lower the average transmission bandwidth. It is therefore cheaper to transmit high quality images.

THANK YOU !!!!

You might also like