Image Compression
Image Compression
Image Compression
2/19/01
12:00 PM
Page 19
Basics of image
representation
An image is essentially a 2-D signal
processed by the Human Visual System.
The signals representing images are usu-
FEBRUARY/MARCH 2001
The taxonomy
The image compression techniques
are broadly classified as either lossless or
lossy, depending, respectively, on
whether or not an exact replica of the
original image could be reconstructed
using the compressed image. Lossless
compression is also referred to as
entropy coding. In addition to using the
spatial and spectral redundancies, lossy
techniques also take advantage of the
way people see to discard data that are
perceptually insignificant.
Lossy schemes provide much higher
compression ratios than lossless
schemes. Lossless compression is used
only for a few applications with stringent
requirements such as medical imaging.
of
the
lossy coding scheme is
further compressed using a
lossless coding scheme
such as Huffman or
Arithmetic coding.
Digital Vision Ltd.
Lossless image
compression
In lossless compression techniques,
the original image can be perfectly recovered from the compressed (encoded)
image. These are also called noiseless
[since they do not add noise to the signal
Authorized licensed use limited to: UNIVERSITY OF NEVADA RENO. Downloaded on November 22, 2008 at 18:05 from IEEE Xplore. Restrictions apply.
19
019_23
2/19/01
12:00 PM
Page 20
Repetitive
Sequence
Encoding
RLE
Statistical
Encoding
Huffman
Arithmetic
LZW
Lossless
Predictive
Coding
Bitplane
Encoding
Block
Truncation
Coding
DPCM
Lossy
Predictive
Coding
DPCM
ADPCM
Delta
Modulation
Transform
Coding
Subband
Coding
DFT
DCT
Haar
Hadamard
Subbands
Wavelets
Fractal
Coding
20
IEEE POTENTIALS
Authorized licensed use limited to: UNIVERSITY OF NEVADA RENO. Downloaded on November 22, 2008 at 18:05 from IEEE Xplore. Restrictions apply.
019_23
2/19/01
12:00 PM
Page 21
1 N M
Ii, j Ii, j
NM i =1 j =1
the energy of the original data being concentrated in only a few of the significant
transform coefficients. This is the basis
of achieving the compression. Only
those few significant coefficients are
selected and the remaining are discarded.
The selected coefficients are considered
for further quantization and entropy
encoding. DCT coding has been the most
common approach to transform coding.
It is also adopted in the JPEG image
compression standard. The broad outline
of transform coding of images is shown
in Fig. 5.
Subband coding. In this scheme, the
image is analyzed to produce the components containing frequencies in welldefined bands, the subbands.
Subsequently, quantization and coding is
applied to each of the bands. The advantage of this scheme is that the quantization and coding well suited for each of
the subbands can be designed separately.
The broad outline of transform coding of
images is shown in Fig. 6.
Vector quantization. The basic idea in
this technique is to develop a dictionary
of fixed-size vectors, called code vectors.
A vector is usually a block of pixel values. A given image is then partitioned
into non-overlapping blocks (vectors)
called image vectors. Then for each
Encoder
Third-Order
Predictor
dm
xm
dm
Quantizer
qm
Entropy
Coder
Channel
B
pm
Reconstructor
sm
Predictor
rm
xm
pm
dm
rm
sm
Original Value
Predicted Value
Prediction Error
Reconstructed Value
Quantized, Reconstructed,
Prediction Error
A
Predicted Value of this Pixel:
a.A + b.B + c.C
Quantize
Each Block
Bitmap
of Block
Entropy Code:
Bitmap and
Reconstruction
Level
FEBRUARY/MARCH 2001
Authorized licensed use limited to: UNIVERSITY OF NEVADA RENO. Downloaded on November 22, 2008 at 18:05 from IEEE Xplore. Restrictions apply.
21
019_23
2/19/01
12:01 PM
Page 22
Quantization
Coefficient Selection:
Zonal/Threshold
Entropy
Coding
Bit Allocation
Divide Image Into
Non-Overlapping
Blocks
Apply Transform
to Each Block
Group 4 Fax
Image Components
at Different
Frequency Subbands
Original Image
Analysis
SubFilter
Sampling
Bank
Quantization
Entropy
Coding
Image compression
standards
Image compression standards have
been developed to facilitate the interoperability of compression and decompression of schemes across several hardware
platforms, operating systems and applications. Most standards are hybrid systems making use of a few of the basic
techniques already mentioned. The major
image compression standards are Group
3, Group 4, and JBIG (Joint Bi-level
Image Group) for bi-tonal images, and
JPEG (Joint Photographic Experts
Group) for continuous-tone images. The
most common application that uses compression of bi-tonal images is digital facsimile (FAX).
Group 3 Fax
The image is scanned left-to-right and
top-to-bottom and the runs of each
colorblack and whiteare determined. A run refers to a sequence of con-
JBIG
The Joint Bi-Level Image Group
(JBIG) standard was developed by the
Scalar quantization
Quantization is a process (function) that maps a very large (possibly infinite)
set of values to a much smaller (finite) set of values. In scalar quantization, the
values that are mapped are scalars (numbers). In the context of image coding
and decoding, the range of pixel values say N, is divided into L non-overlapping
intervals, also known as quantization levels.
Each interval i is defined by its decision boundaries (di, di+1). During encoding, the quantizer maps a given pixel value x to a quantization level
l: l = Q(x), such that dl x < d l+1. Each quantization level i has its associated
reconstruction level ri.
During decoding, the (de)quantizer maps a given level l to a reconstruction
pixel value rl = x, x = Q-1 (l). This introduces noise or error in the image (signal)
called quantization error. This is the root mean square value of the x - x.
The essential difference among various types of quantizers is in terms of how
the forward and inverse mappings are defined. These definitions are dictated
according to the number of quantization levels, the decision boundaries and
the reconstruction values. The basic design objective of a quantizer is to minimize
the quantization error, while being computationally simple. The quantizer has a
large impact on the compression ratio and image quality of a lossy scheme.
There are two broad types of scalar quantizersuniform and non-uniform. In
a uniform quantizer of k levels, the range of values is divided into k equally
spaced intervals. The reconstruction values are the mid-points of the intervals.
This is simple to implement but it does not attempt to minimize the quantization
error. A quantizer that takes into account the probability distributions of the pixels in images performs better. Such a quantizer is a non-uniform quantizer,
where the intervals are non-uniform. The most common non-uniform quantizer
is the Lloyd-Max quantizer. For it, the decision boundaries and the reconstruction levels are determined using the probability model of the image pixels such
that the quantization error is minimized.SRS
22
IEEE POTENTIALS
Authorized licensed use limited to: UNIVERSITY OF NEVADA RENO. Downloaded on November 22, 2008 at 18:05 from IEEE Xplore. Restrictions apply.
019_23
2/19/01
12:01 PM
Page 23
JPEG
The Joint Photographic Experts
Group (JPEG) is a standard developed
for compressing continuous-tone still
images. JPEG has been widely accepted
for still image compression throughout
the industry. JPEG can be used on both
gray-scale and color images. JPEG consists of four modes: lossless, sequential,
progressive and hierarchical. The first
one is a lossless mode and the other
three are lossy modes. The sequential
mode, also called baseline JPEG, is the
most commonly used scheme.
The lossless JPEG mode uses linear
predictive schemes. It provides seven
different predictors. Pixel values (except
those at the boundaries) are predicted
based on neighboring pixels. The residual, which is the difference between the
original and the predicted image, is
encoded using entropy (lossless) coding
such as Huffman or arithmetic coding.
In the baseline JPEG scheme, the
image is divided into non-overlapping
blocks of 8 x 8 pixels. DCT is applied to
each block to obtain the transform coefficients. The coefficients are then quantized using a table specified by the standard, which contains the quantizer step
sizes. The quantized coefficients are then
ordered using a zigzag ordering. The
ordered quantized values are then encoded using Huffman coding tables, specified by the standard.
Progressive JPEG compression is
similar to the sequential (baseline)
JPEG scheme in the formation of DCT
coefficients and quantization. The key
difference is that each coefficient
(image component) is coded in multiple scans instead of a single scan. Each
successive scan refines the image until
the quality determined the quantization
Summary
Image Processing
Original
Image
Segments
Color Separation
Edge Detection
Spectrum Analysis
Texture Analysis
Entropy
Coding
Library of Fractals
Lookup
Closest
IFS Code
of Original
Image
Affine Transformation:
Combination of Rotations,
Scalings, Translations
are
then described with well-defined image
models. They are expressible using
much less information than the original
data. The challenge is in devising good
models that achieve good compression
without loss of fidelity.
FEBRUARY/MARCH 2001
Authorized licensed use limited to: UNIVERSITY OF NEVADA RENO. Downloaded on November 22, 2008 at 18:05 from IEEE Xplore. Restrictions apply.
23