Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Farooq CHAPTER ONE - Three

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 8

CHAPTER ONE

INTRODUCTION

1.1 Background of Study


Data are essential resources treasured by all organizations, hence, making its security
and access to its use a top priority. This is the reason why modern business organizations
improve on the security of their data and the security of who have access to any of what is
theirs.

Different security approaches have been utilized by different organizations, formal


approaches used are not limited to; biometrics access, web-based attendance systems,
attendance system software, and smart swipe cards. But, the loopholes detected in those
approaches are the reason why a better approach becomes paramount. This has necessitated
the need to design an Intelligent Based Recognition System using Matlab for the department
of computer engineering, in Olabisi Onabanjo University, Ibogun Campus.

Intelligent Based Recognition system makes it easy for the any organization utilizing it to
provide a more secure and authenticated access only to people they wish to give the access to
which will also allow them keep track of all records and to fully eliminate the problem of
staff impersonation in the organizations. It will operate in two phases: Face detection and
Face recognition. Each phase has its own approaches to achieving the aim and objectives of
the system.

1.2 Statement of the problem

The problems this project intends to solve are quite difficult. The major one is the
issue of data security; other problems facing this project are data access, impersonation, staff
attendance system and staff payroll. Inappropriate handling of important data can lead to the
end of any organization, which is the reason why an improved and advanced approached to
data security is being worked on.

1.3 Aim and Objectives

The main aim of this project is to design an Intelligent Based Recognition System
using Matlab for the department of Computer Engineering, Olabisi Onabanjo University, Ago
Iwoye, Ibogun Campus. The following are the objectives of this project:
1. To formulate the required procedures that will demonstrate a real time face detection
and recognition operation on Matlab.
2. To design an expansible database inside the Matlab software.
3. To train the neural network to achieve the detection and recognition processes to
secure access to data through simulation.
4. To analyze the result generated from simulation.

1.4 Scope of study


This project explains the state of data security in the existing approaches as well as
their loopholes, it further explains the processes involved in creating databases in Matlab
software, processes involved in training the Matlab’s neural network so it performs the
required functions. Thus it will establish further research on providing good mechanism for
detection of faces and their recognitions while providing more secure access to classified
data.

1.5 Significance of study


1. This project will improve the security and access to data in any organization.
2. This project will help organization explore the advantages of artificial intelligence.
3. This project will help reduce the loopholes detected in some of the existing
approaches.

1.6 Tools Used


1. Operating System: Windows 7 Ultimate
2. Integrated Design Environment: Matlab
CHAPTER TWO

LITERATURE REVIEW

Face detection is a computer technology that determines the location and size of human face
in arbitrary (digital) image. The facial features are detected and any other objects like trees,
buildings and bodies etc. are ignored from the digital image. It can be regarded as a specific
case of object-class detection, where the task is finding the location and sizes of all objects in
an image that belong to a given class. Face detection, can be regarded as a more general case
of face localization. In face localization, the task is to find the locations and sizes of a known
number of faces (usually one). Basically there are two types of approaches to detect facial
part in the given image i.e. feature base and image base approach. Feature base approach tries
to extract features of the image and match it against the knowledge of the face features.
While image base approach tries to get best match between training and testing images.
Beymer, D. and Poggio, T. (2015).

2.1 FEATURE BASE APPROCH:


Active Shape Model focuses on complex non-rigid features like actual physical and
higher level appearance of features means that Active Shape Models (ASMs) are aimed at
automatically locating landmark points that define the shape of any statistically modeled
approach.
a) Facial model from a training set containing images with manually annotated
landmarks. ASMs is classified into three groups i.e. snakes, PDM, Deformable
templates.
b) Object in an image. For any of facial features such as the eyes, lips, nose, mouth and
eyebrows. The training stage of an ASM involves the building of a statistical

2.1.1 Deformable Templates


Deformable templates were then introduced by Yuille et al. to take into account the
priority of facial features and to better the performance of formal approaches. Locating a
facial feature boundary is not an easy task because the local evidence of facial edges is
difficult to organize into a sensible global entity using generic contours. The low brightness
contrast around some of these features also makes the edge detection process. Yuille et al.
took the concept of snakes a step further by incorporating global information of the eye to
improve the reliability of the extraction process.
Deformable templates approaches are developed to solve this problem. Deformation is based
on local valley, edge, peak, and brightness.

2.2 Low Level Analysis


Based on low level visual features like color, intensity, edges, motion etc.Skin Color
Base Color is a vital feature of human faces. Using skin-color as a feature for tracking a face
has several advantages. Color processing is much faster than processing other facial features.
Under certain lighting conditions, color is orientation invariant. This property makes motion
estimation much easier because only a translation model is needed for motion estimation.
Tracking human faces using color as a feature has several problems like the color
representation of a face obtained by a camera is influenced by many factors (ambient light,
object movement, etc.
Majorly three different face detection algorithms are available based on RGB, YCbCr, and
HIS color space models. In the implementation of the algorithms there are three main steps
viz.
(1) Classify the skin region in the color space,
(2) Apply threshold to mask the skin region and
(3) Draw bounding box to extract the face image.
Crowley and Coutaz (2014) suggested simplest skin color algorithms for detecting skin
pixels. The perceived human color varies as a function of the relative direction to the
illumination. The pixels for skin region can be detected using a normalized color histogram,
and can be normalized for changes in intensity on dividing by luminance.

2.3 Motion Base


When use of video sequence is available, motion information can be used to locate
moving objects. Moving silhouettes like face and body parts can be extracted by simply
thresholding accumulated frame differences. Besides face regions, facial features can be
located by frame differences.

2.3.1 Gray Scale Base


Gray information within a face can also be treated as important features. Facial
features such as eyebrows, pupils, and lips appear generally darker than their surrounding
facial regions. Various recent feature extraction algorithms search for local gray minima
within segmented facial regions. In these algorithms, the input images are first enhanced by
contrast-stretching and gray-scale morphological routines to improve the quality of local dark
patches and thereby make detection easier. The extraction of dark patches is achieved by low-
level gray-scale thresholding. Based method and consist three levels. Heisele, B. and Poggio,
T. (2009) presented new approach i.e. faces gray scale behaviour in pyramid (mosaic)
images. This system utilizes hierarchical face location consist three levels. Higher two level
based on mosaic images at different resolution. In the lower level, edge detection method is
proposed. Moreover these algorithms give fine response in complex background where size
of the face is unknown.

2.3.2 Edge Base


Face detection based on edges was introduced by Sakai et al. This work was based on
analyzing line drawings of the faces from photographs, aiming to locate facial features. Than
later Craw et al. proposed a hierarchical framework based on Sakai et al‘s work to trace a
human head outline. Then after remarkable works were carried out by many researchers in
this specific area. Some other methods proposed frame work which consist three steps i.e.
initially the images are enhanced by applying median filter for noise removal and histogram
equalization for contrast adjustment. Then a novel edge tracking algorithm is applied to
extract the sub windows from the enhanced image based on edges. Further they used Back
Propagation Neural Network (BPN) algorithm to classify the sub-window as either face or
non-face.

2.4 Feature Analysis


These algorithms aim to find structural features that exist even when the pose,
viewpoint, or lighting conditions vary, and then use these to locate faces. These methods are
designed mainly for face localization.
2.4.1 Feature Searching
Viola Jones Method:
Paul Viola and Michael Jones presented an approach for object detection which
minimizes computation time while achieving high detection accuracy. Paul Viola and
Michael Jones (2007) proposed a fast and robust method for face detection which is 15 times
quicker than any technique at the time of release with 95% accuracy at around 17 fps. The
technique relies on the use of simple Haar-like features that are evaluated quickly through the
use of a new image representation. Based on the concept of an Integral Image it generates a
large set of features and uses the boosting algorithm to reduce the over complete set and the
introduction of a degenerative tree of the boosted classifiers provides for robust and fast
interferences. The detector is applied in a scanning fashion and used on gray-scale images,
the scanned window that is applied can also be scaled, as well as the features evaluated.
Gabor Feature Method
Sharif et al proposed an Elastic Bunch Graph Map (EBGM) algorithm that
successfully implements face detection using Gabor filters. The proposed system applies 40
different Gabor filters on an image. As a result of which 40 images with different angles and
orientations are received. The system reduces these points in accordance to distance between
them. The next step is calculating the distances between the reduced points using distance
formula. At last, the distances are compared with database. If match occurs, it means that the
faces in the image are detected. Baron, R. J (1981).

2.5 Constellation Method


All methods discussed so far are able to track faces but still some issue like locating
faces of various poses in complex background is truly difficult. To reduce this difficulty
investigator form a group of facial features in face-like constellations using more robust
modeling approaches such as statistical analysis. Various types of face constellations have
been proposed by Burl et al. They establish use of statistical shape theory on the features
detected from a multiscale Gaussian derivative filter. Brunelli, R. and Poggio, T. (2003).

2.5.1 Neural Network


Neural networks gaining much more attention in many pattern recognition problems
such as OCR, object recognition, and autonomous robot driving. Since face detection can be
treated as a two class pattern recognition problem, various neural network algorithms have
been proposed. The advantage of using neural networks for face detection is the feasibility of
training a system to capture the complex class conditional density of face patterns. However,
one demerit is that the network architecture has to be extensively tuned (number of layers,
number of nodes, learning rates, etc.) to get exceptional performance. In early days most
hierarchical neural network was proposed by Agui et al. The first stage having two parallel
sub networks in which the inputs are filtered intensity values from an original image.
CHAPTER THREE

RESEARCH METHODOLODY

This chapter describes the training of neural network towards achieving the detection and
recognition of faces for data access and security. The following tools will be used to achieve
our goals: Windows Operating System and Matlab.
REFERENCES

1. Beymer, D. and Poggio, T. (1995) Face Recognition From One Example View, A.I.
Memo No. 1536, C.B.C.L. Paper No. 121. MIT.

2. Crowley and Coutaz (2014) Computer and Robot Vision, Volume I. Addison-
Wesley.

3. Heisele, B. and Poggio, T. (1999) Face Detection. Artificial Intelligence Laboratory.


MIT.

4. Sakai et al. Mechanisms of human facial recognition. International Journal of Man


Machine Studies, 15:137-178.

5. Baron, R. J. (2001). Mechanisms of human facial recognition. International Journal of


Man Machine Studies, 15:137-178.

6. Brunelli, R. and Poggio, T. (1993), Face Recognition: Features versus Templates.


IEEE Transactions on Pattern Analysis and Machine Intelligence,
15(10):1042-1052.

You might also like