The Lecture Contains:: Lecture 41: Performance Measures, Intraframe Coding, Predictive and Transform Coding
The Lecture Contains:: Lecture 41: Performance Measures, Intraframe Coding, Predictive and Transform Coding
The Lecture Contains:: Lecture 41: Performance Measures, Intraframe Coding, Predictive and Transform Coding
Subjective measures are normally evaluated by showing the reconstructed video to a group of
subjects and asking for their views on the perceived quality. A number of subjective assessment
methodologies have been developed over the years. Examples are the double stimulus impairment
scale (DSIS) and the double and single stimulus continuous quality scales, (DSCQS) and (SSCQS),
respectively. Despite their reliability, subjective quality experiments are expensive and time
consuming.
Objective measures provide cheaper and faster alternatives. One commonly used objective measure
is the mean squared error (MSE), which is defined as
(8.10)
Where H and V are the horizontal and vertical dimensions of the frame, respectively, and
and are the pel values at location of the original and reconstructed frames,
respectively. Care should be taken to include color components and to take into account any chroma
subsampling. For example, the MSE of a reconstructed 4:2:0 color frame can be calculated as
(8.11)
At the decoder, the same prediction is produced using previously decoded pels, and the received
error signal is added to reconstruct the current pel. A block diagram of a predictive coding system is
depicted in Figure 8.4.
Predictive coding is commonly referred to as differential pulse code modulation (DPCM). A special
case of this method is delta modulation (DM), which quantizes the error signal using two quantization
levels only.
Predictive coding can take many forms, depending on the design of the predictor and the quantizer
blocks. The predictor can use a linear or a nonlinear function of the previously decoded pels, it can
be 1-D (using pels from the same line) or 2-D (using pels from the same line and from previous
lines), and it can be fixed or adaptive. The quantizer also can be uniform or nonuniform, and it can be
fixed or adaptive.
(a) Encoder
(b) Decoder
When choosing a transform, three main properties are desired: good energy compaction, data-
independent basis functions, and fast implementation.
The performance of many suboptimal transforms with data-independent basis functions has been
studied. Examples are the discrete Fourier transform (DFT), the discrete cosine transform (DCT), the
Walsh-Hadamard transform (WHT), and the Haar transform. It has been demonstrated that the DCT
has the closest energy-compaction performance to that of the optimum KLT. This has motivated the
development of a number of fast DCT algorithms Due to these attractive features, i.e., near-
optimum energy-compaction, data-independent basis functions and fast algorithms, the DCT
has become the "workhorse" of most image and video coding standards .
The transform coefficient F(0,0) at the top-left corner of the transformed block is called the DC
coefficient because it contains the lowest frequencies in both the horizontal and vertical dimensions.
The corresponding inverse DCT transform is given by
(8.15
)
It can be deduced from Equation (8.13) that the computational complexity of an 2-D DCT is of
the order . However, one of the advantages of the DCT is that it is separable. This means
that a 2-D DCT can be separated into a pair of 1-D DCTs. Thus, to obtain the 2-D DCT of an
block, a 1-D DCT is performed first on each of the N rows of the block and then on each of the N
columns of the resulting block (or vice versa). The same applies to the inverse DCT. This reduces
the complexity to . Further reductions in complexity can be achieved using a number of fast
DCT algorithms.
Compared to predictive coding, transform coding provides higher compression with less sensitivity to
errors and less dependence on the input data statistics. Its higher computational complexity and
storage requirements have been offset by advances in integrated circuit technology. One
disadvantage, however, is that when compression factors are pushed to the limit, three types of
artefacts start to occur: (i) "graininess" due to coarse quantization of some coefficients, (ii) "blurring"
due to the truncation of high-frequency coefficients, and (iii) "blocking artifacts," which refer to
artificial discontinuities appearing at the borders of neighboring blocks due to independent processing
of each block. Since blocking artefacts are the most disturbing, a number of methods have been
proposed to reduce them. Examples are overlapping blocks at the encoder, the use of the lapped
orthogonal transform , and postprocessing using filtering and image restoration techniques.