Topic - 1 Introduction To Image and Vision
Topic - 1 Introduction To Image and Vision
CoSc-6412
Dr. V. Anitha
1
Email: gamya21@gmail.com
Course Description
This course is designed to give students the
fundamentals in 2D digital image processing with
emphasis in image processing techniques, image
filtering design, segmentation, enhancement,
morphological processing of image, recognition of
objects in an image and the applications of image
processing.
2
Course Objective
On the successful completion of the course the students will be
able to:
➢ have a clear understanding of the principals the Digital Image
Processing
➢ understand the mathematical foundations for digital manipulation of
images
➢ learn and understand the Image Enhancement in the Spatial and
Frequency Domain.
➢ understand Image Restoration, Compression, Segmentation,
Recognition, Representation and Description.
3
Course Content
Chapter/Topic Sub-Topics
Chapter 1: ➢ Elements of visual perception
Introduction to Image ➢ Image sensing and acquisition
andVision ➢ Image sampling and quantization
➢ Digital image representation
➢ Linear and nonlinear representation
5
Reference
1. Gonzalez, R. C. andWoods, R. E. [2002/2008], Digital Image
Processing, 2nd/3rd ed., Prentice Hall
2. Sonka, M., Hlavac,V., Boyle, R. [1999]. Image Processing,Analysis and
MachineVision (2nd edition), PWS Publishing, or (3rd edition)Thompson
Engineering, 2007
3. Gonzalez, R. C.,Woods, R. E., and Eddins, S. L. [2009].Digital Image
Processing UsingMATLAB, 2nd ed., Gatesmark Publishing, Knoxville,TN.
4. Anil K. Jain [2001], Fundamentals of digital image processing (2nd
Edition), Prentice-Hall, NJ
5. Willian K. Pratt [2001], Digital Image Processing (3rd Edition), , John
Wiley & Sons, NY
6. 6. Burger,Willhelm and Burge, Mark J. [2008]. Digital Image Processing:
AnAlgorithmic Introduction Using Java, Springer
6
Course Evaluation
Journal Paper Review and Presentation---15%
MiniTest--------------------------------------25%
Group Mini-Project (Pair)/Presentation--10%
L a b a n d Class Engagement andActivities 10%
Final Exam-----------------------------------40%
7
Scientific Paper Review
IndividualTask
Select article of your interest
Make a Scientific Review
Presentation Date ??????? (To be decided)
8
Image Analysis and Computer Vision
CoSc-6412
Topic 1:
Introduction to Image and Vision
9
Topic Coverage
Chapter 1: ➢ Elements of visual perception
➢ Image sensing and acquisition
Introduction to ➢ Image sampling and quantization
Image andVision ➢ Linear and nonlinear representation
➢ Digitalimage representation
10
Every picture tells a story
Image carries vast amount of
information.
We, humans, are selective of what we
consume through visual sense.
The goal of computer vision is to
write computer programs that can
interpret images
Can computers match human
perception?
Yes and no (but mostly no!)
humans are much better at “hard” things
computers can be better at “easy” things
11
Overview: Computer Imaging
• Definition of computer imaging:
– Acquisition and processing of visual
information by computer.
• Why is it important?
– Human primary sense is visual sense.
– Information can be conveyed well through
images (one picture worth a thousand words).
– Computer is required because the amount of
data to be processed is huge.
Overview: Computer Imaging
Computer Image
Vision Processing
What is Computer Vision?
Deals with the development of the theoretical
and algorithmic basis by which useful
information about the 3D world can be
automatically extracted and analyzed from a
single or multiple o 2D images of the world.
27
Computer Vision
Recognition
Feature extraction
Segmentation of image parts
Detect and identify objects
Understanding
Giving context to image parts
Knowing what is happening in the scene?
13
Why is Computer Vision Difficult?
It is a many-to-one mapping
A variety of surfaces with different material and geometrical
properties, possibly under different lighting conditions, could
lead to identical images
Inverse mapping has non unique solution (a lot of information is
lost in the transformation from the 3D world to the 2D image)
It is computationally intensive
We do not understand the recognition problem
30
What is an Image?
The pattern is defined is a coordinate system whose origin is
conventionally defined as the upper-left corner of the image .
We can describe the pattern by a function f(x,y).
(0,0)
14
Image Processing and Related Fields
16
What Is Digital Image Processing?
Digital image processing helps us enhance images to make
them visually pleasing, or emphasize regions or features of an
image to better represent the content.
For example, we may wish to enhance the brightness and
contrast to make a better print of a photograph, similar to
popular photo-processing software.
In a magnetic resonance image (MRI) of the brain, we may
want to accentuate a certain region of image intensities to see
certain parts of the brain.
12
Image Processing
• Processed images are to be used by
human.
– Therefore, it requires some understanding on
how the human visual system operates.
• Among the major topics are:
– Image restoration.
– Image enhancement.
– Image compression.
Image Processing
• Image restoration:
– The process of taking an image with some
know, or estimated degradation, and restoring
it to its original appearance.
– Done by performing the reverse of the
degradation process to the image.
– Examples: correcting distortion in the optical
system of a telescope.
Image Processing
• Image enhancement:
– Improve an image visually by taking an
advantage of human visual system’s response.
– Example: improve contrast, image sharpening,
and image smoothing.
Image Processing
• Image compression:
– Remove the amount of data required to
represent an image by:
• Removing unnecessary data that are visually
unnecessary.
• Taking advantage of the redundancy that is
inherent in most images.
– Example: JPEG, MPEG, etc.
Elements ofVisual Perception
How do people perceive image?
How images are formed in the eye?
How human and electronic imaging compare in terms of
resolution and ability to adapt to changes in illumination
17
Structure of the Human Eye
The eye is nearly a sphere, with an average diameter of
approximately 20 mm.
Three membranes enclose the eye:
1. the cornea (transparent exterior portion of the eye covering the iris..color part
of the eye ) and sclera (protects the delicate structure inside) outer cover;
2. the choroid (another layer found underneath the sclera); and
3. the retina (is a collection of light sensitive tissues).
The cornea is a tough, transparent tissue that covers the
anterior surface of the eye.
Continuous with the cornea, the sclera is an opaque
membrane that encloses the remainder of the optic globe.
The choroid lies directly below the sclera.
18
Structure of the Human Eye
The lens is made up of concentric layers of fibrous cells and is
suspended by fibers that attach to the ciliary body.
The Lens contains 60 to 70%water, about 6%fat,and more
protein than any other tissue in the eye.
The innermost membrane of the eye is the retina, which lines
the inside of the wall’s entire posterior portion.
When the eye is properly focused, light from an object
outside the eye is imaged on the retina. Pattern vision is
afforded by the distribution of discrete light receptors over
the surface of the retina.
19
Structure of the Human Eye
20
Structure of the Human Eye
21
Image formation in the Eye
The principal difference between the lens of the eye and
an ordinary optical lens is that lens of the eye is
flexible.
The shape of the lens is controlled by tension in the
fibers of the ciliary body.
To focus on distant objects, the controlling muscles
cause the lens to be relatively flattened.
Similarly, these muscles allow the lens to become thicker
in order to focus on objects near the eye.
22
The Human Visual System
The Human Visual System
• This is how human visual system works:
– Light energy is focused by the lens of the eye
into sensors and retina.
– The sensors respond to the light by an
electrochemical reaction that sends an
electrical signal to the brain (through the optic
nerve).
– The brain uses the signals to create
neurological patterns that we perceive as
images.
The Human Visual System
• Cones:
– For daylight vision.
– Sensitive to color.
– Concentrated in the central region of eye.
– High resolution capability (differentiate small
changes).
The Human Visual System
• Blind spot:
– No sensors.
– Place for optic nerve.
– We do not perceive it as a blind spot because
the brain fills in the missing visual information.
• Why does an object should be in center
field of vision in order to perceive it in fine
detail?
– This is where the cones are concentrated.
The Human Visual System
• Cones have higher resolution than rods
because they have individual nerves tied to
each sensor.
• Rods have multiple sensors tied to each
nerve.
• Rods react even in low light but see only a
single spectral band. They cannot
distinguish color.
Image formation in the Eye
The distance between the center of the lens and the retina
(called the focal length) varies from approximately 17 mm
to about 14 mm, as the refractive power of the lens increases
from its minimum to its maximum.
If h is the height in mm of that object in the retinal image, the
geometry of visualization yields 15/100=h/17 or h=2.55
mm.
23
The Human Visual System
24
Issue of Illumination
Same objects and arrangement
Different angle of light – Many to one mapping
25
Perception—Illusions
The border of the
square is visible There seems to be
despite there is no a circle in the
border line middle
26
Image Processing
28
The Three Processing Levels
1. Low-level processing
Standard procedures are applied to improve image quality
Procedures are required to have no intelligent capabilities.
32
The Three Processing Levels (cont’d)
2. Intermediate-level processing
Extract and characterize components in the image
Some intelligent capabilities are required.
33
The Three Processing Levels (cont’d)
3. High-level processing
Recognition and interpretation.
Procedures require high intelligent capabilities.
34
35
36
37
38
39
40
41
Mathematics in Computer Vision
In the early days of computer vision, vision systems
employed simple heuristic methods.
Today, the domain is heavily inclined towards theoretically,
well-founded methods involving non-trivial mathematics.
Calculus
LinearAlgebra
Probabilities and Statistics
Signal Processing
Projective Geometry
Computational Geometry
OptimizationTheory
42
ControlTheory
Computer Vision Applications
Industrial inspection/quality control
Surveillance and security
Face recognition
Gesture recognition
Space applications
Medical image analysis
Autonomous vehicles
Virtual reality and much more …...
43
Face Detection Face Blurring
44
Medical Image Analysis
45
Surveillance and Tracking
46
Surveillance and Tracking
47
Smart cars
49
Self-driving cars
51
Optical character recognition
52
Sports Video Analysis
54
Critical issues
What information should be extracted?
55
Computer Imaging Systems
• Computer imaging systems comprises of
both hardware and software.
• The hardware components can be divided
into three subsystems:
– The computer
– Image acquisition: camera, scanner, video
recorder.
– Image display: monitor, printer, film, video
player.
Computer Imaging Systems
• The software is used for the following
tasks:
– Manipulate the image and perform any
desired processing on the image data.
– Control the image acquisition and storage
process.
• The computer system may be a general-
purpose computer with a frame grabber or
image digitizer board in it.
Computer Imaging Systems
What do these
numbers represent?
56
Image Sampling and Quantization
Objective of imaging is to generate digital images
(representation) from sensed data (observation)
In creating digital image, there is a need to convert the continuous
sensed data into digital form.This involves two processes:
sampling and quantization.
An image may be continuous with respect to the x- and y-
coordinates, and also in amplitude.
To convert it to digital form, we have to sample the function in
both coordinates and in amplitude.
1. Digitizing the coordinate values is called sampling.
2. Digitizing the amplitude(brightness) values is called
quantization.
61
Image Sampling and Quantization
62
Image Sampling and Quantization
Sampling and
quantization
63
Image Representation
64
Some Basic Relationships Between Pixels
65
Distance Measures
Euclidean distance
1
De ( p, q) [(x s)2 ( y t) 2
City-block distance ]2
D4 ( p, q) | (x s) | | (y t) |
Chessboard distance
D8 ( p, q) max(| (x s) |,| ( y t)
|)
66
Region/Boundary/Edge
Region
We call R a region of the image if R is a connected set
Boundary
The boundary of a region R is the set of pixels in the
region that have one or more neighbors that are not in R
Edge
Pixels with derivative values that exceed a preset
threshold
67
Image Representation
1. Image capture
2. Image quality measurements
3. Image resolution
4. Colour representation
5. Camera calibration
6. Parallels with human visual system
68
Image Capture
Many sources
Consider requirements of system
Resolution
69
Representation
Sampled data
Spatial.......Area based
Amplitude......light intensity
On a rectangular array
Multidimensional array
70
Image Resolution
How many pixels
Spatial resolution
How many shades of grey/colours
Amplitude resolution
How many frames per second
Temporal resolution (Motion)
71
Spatial Resolution
72
Spatial Frequency Resolution
• To understand the concept of spatial
frequency, we must first understand the
concept of resolution.
• Resolution: the ability to separate two
adjacent pixels.
– If we can see that two adjacent pixels as being
separate, then we can say that we can resolve
the two.
Spatial Frequency Resolution
73
Shades of Grey
74
End of Topic 1
75