A Survey On Video-Based Face Recognition Approaches
A Survey On Video-Based Face Recognition Approaches
A Survey On Video-Based Face Recognition Approaches
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com Volume 3, Issue 2, February 2014 ISSN 2319 - 4847
Abstract
Face recognition is one of the most suitable applications of image analysis. Its a true challenge to build an automated system which equals human ability to recognize faces. While traditional face recognition is typically based on still images, face recognition from video sequences has become popular recently due to more abundant information than still images. This paper presents an overview of face recognition scenarios and video-based face recognition system architecture and various approaches are used in video-based face recognition system. This paper provides an up-to-date survey of video-based face recognition research.
1. INTRODUCTION
Face recognition is a biometric approach that employs automated method to verify or recognize the identity of a living person based on his/her physiological characteristics. It also used in wide range of commercial and law enforcement and interesting area in real time applications. Face recognition has several advantages over other biometric technologies: It is natural, nonintrusive, and easy to use [1]. Face recognition system can help in many ways: for example some applications are Checking for criminal records and Detection of a criminal at public place, Finding lost children's by using the images received from the cameras fitted at some public places and detection of thiefs at ATM machines, Knowing in advance if some unknown person is entering at the border checkpoints and so on. A face recognition system can operate in either or both of two modes: (1) face verification (or authentication), and (2) face identification (or recognition). Face verification involves a one to-one match that compares a query face image against a template face image. Face identification involves one-to-many matches that compare a query face image against all the template images in the database to determine the identity of the query face. The first automatic face recognition system was developed by Kanade[2], so the performance of face recognition systems has improved significantly. Face recognition in videos is an active topic in the field of image processing, computer vision and biometrics over many years. Compared with still face recognition videos contain more abundant information than a single image so video contain spatio-temporal information. To improve the accuracy of face recognition in videos to get more robust and stable recognition can be achieved by fusing information of multi frames and temporal information and multi poses of faces in videos make it possible to explore shape information of face and combined into the framework of face recognition. The video-based recognition has more advantages over the image-based recognition. First, the temporal information of faces can be utilized to facilitate the recognition task. Secondly, more effective representations, such as a 3D face model or super-resolution images, can be obtained from the video sequence and used to improve recognition results. Finally, videobased recognition allows learning or updating the subject model over time to improve recognition results for future frames. So video based face recognition is also a very challenging problem, which suffers from following nuisance factors such as low quality facial images, scale variations, illumination changes, pose variations, Motion blur, and occlusions and so on[3]. Face recognition can generally be categorized into one of the following three scenarios based on the characteristics of the image(s) to be matched. Such as Still-to-still recognition, Video-to-image face recognition, Video-to-video face recognition [4]. i) Research on still image face recognition has been done for nearly half a century. Still-to-still image matching is the most common process and is used in both constrained and unconstrained applications. but it suffers from several factors those are the need to constrain the face recognition problem, computational constraints, and the large amount of legacy still face images (e.g. id cards, mug shots). ii) Video-to-image face recognition can be seen as an extension of still image based face recognition. Video-to-still image matching occurs when a sequence of video frames is matched against a database of still images (e.g. mug shots or Identification photos). The input of the system is videos while the database is still face images. Compared to traditional still image based face recognition, how to explore the multi-frame information of the input video is the key to enhance the performance. In summary, image-video based methods make use of multi-frame information to improve the accuracy of face recognition, and improve the robustness to deal with pose variations, occlusions and illumination changes.iii) Video-to-video matching, or re-identification, is performed to find all occurrences of a subject within a collection of video data. Re-identification is generally a necessary
Page 208
2.1.1 Face detection Face detection is the first stage of a face recognition system. This module system takes a frame of a video sequence and performs some image processing techniques on it in order to find locates candidate face region. System can operate on static images, where this procedure is called face localization and dealing with videos procedure is called face tracking. The purpose of face localizing and extracting the face region from the background. Face detection can be performed based on several things those are skin texture, motion (for faces in videos), facial/head shape, facial appearance, or a combination of these parameters. An input image is scanned at all possible locations and scales by a sub window. Face detection is posed as classifying the pattern in the sub window as either face or non-face. Face detection techniques similar to those applied for still images are then employed to find the exact location of faces in the current frame, thus initiating face and facial feature tracking. Face tracking techniques include head tracking, where the head is viewed as a rigid object performing translations and rotations. While the tracking module finds the exact position of facial features in the current frame based on an estimate of face or feature locations in the previous frame(s). Face detection based on videos consists of generally three main processes. Firstly frame based detection, in this process, lots of traditional methods[6] for still images can be introduced such as statistical modeling method, Neural networkbased method, SVM-based method, HMM method, BOOST method and Color-based face detection, etc. Secondly combination of detection and tracking, this says that detecting face in the first frame and then tracking it through the whole sequence. Since detection and tracking are independent and information from one source is just in use at one time, loss of information is unavoidable. Finally, instead of detecting each frame, temporal approach exploits temporal relationships between the frames to detect multiple human faces in a video sequence.
Page 209
Fig. 2.
Page 210
Page 211
Page 212
Fig 3. A sequence of face images of single person arbitrarily rotating his head In [28], each faces in the database and input video are modeled with GMM, and Kullback-Leibler divergence is measured as the similarity measurement to achieve recognition. In [29],Arandjelovic et al. made use of kernel-based methods to map low-dimensional space to high dimensional space, and then use low dimensional space of linear methods (such as PCA) to solve complex nonlinear problems in the high-dimensional space. Zhou et al. [30] mapped the vector space into RKHS (Rep reducing Kernel Hilbert Space) by kernel based methods to calculate the distance between the probability distribution. In a multi-view dynamic face model is built to achieve face recognition. Firstly, dynamic face model are constructed including a 3D model, a texture model and an affine change model. Then a Kalman filter is adopted to obtain the shape and texture, which builds a segmented linear manifold for each single person with the face texture reduced by KDA (Kernel Discriminate Analysis). Face recognition is achieved in the following by trajectory matching. However, the 3D model estimation requires a lot of multi-angle images and a larger complexity computational. This kind of solution is much better than the feature vector based solution, which makes use of probability theory to enhance the performance. However, the dynamic change information of faces in videos is neglected, which has potential to improve the video based face recognition. 4.3 Dynamic changes of faces in videos The temporal information in video sequences enables the analysis of facial dynamic changes and its application as a biometric identifier for person recognition. we have utilize the human nature that human will have at least small amount of movements such as eyes blinking and/or mouth and face boundary movements. We can get this information easily because dealing with video sequence by which the whole sequence of the object's movements can be obtained. Taking that point in to account we can reduce the error that occurs due to false detection of a human face and minimize the time of simulation.
Fig 4. Example of a Hidden Markov Model applied to video Matta et al. proposed a multi-modal recognition system [31,32]. They successfully integrated the facial motion information with mouth motion and facial appearance by taking advantage of a unified probabilistic framework. In [33], Huang and Trivedi developed a face recognition system by employing HMMs for facial dynamic information modeling in videos. Each covariance matrix was gradually adapted from a global diagonal one by using its class-dependent data in training algorithms. Afterwards, Liu and Cheng [34] successfully applied HMMs for temporal video recognition (as illustrated in Fig. 4)by improving the basic implementation of Huang and Trivedi. Each test sequence was used to update the model parameters of the client in question by applying a maximum a posteriori (MAP) adaptation technique.
Page 213
REFERENCES
[1] ShailJa A Patil, Paramod j Deore, Video-based face recognition: a survey, Proceedings of Conference on Advances in Communication and Computing (NCACC'12) Held at R.C.Patel Institute of Technology, Shirpur, Dist. Dhule,Maharastra,India.April 21, 2012. [2] Kanade, Picture Processing by Computer Complex and Recognitionof Human Faces. Ph.D. thesis,Kyoto University, 1973 [3] Zhaoxiang Zhang, Chao Wang and Yunhong Wang, Video-Based Face Recognition: State of the Art ," [4] Lacey Best-Rowden , Brendan Klare , Video-to-Video Face Matching: Establishing a Baseline for Unconstrained Face Recognition,To appear in the Proc. IEEE BTAS, 2013. [5] Jeremiah r. Barr , kevin w. Bowyer , Face recognition from video: a review, International Journal of Pattern Recognition and Artificial Intelligence Vol. 26, No. 5 (2012) 1266002 (53 pages) [6] Yongzhong Lu, Jingli Zhou, Shengsheng Yu, a survey of face detection, extraction and recognition, Computing and Informatics, Vol. 22, 2003. [7] G. Shakhnarovich, J. W. Fisher, and T. Darrell, Face recognition from long-term observations., In Proc. European Conf. on Computer Vision,volume 3, pp: 851-865, 2002. [8] S. Zhou and R. Chellappa, Probabilistic human recognition from video, in Proceedings of the European Conference on Computer Vision, pp. 681697, Copenhagen, Denmark, 2002. [9] V. Krueger and S. Zhou.,Exemplar-based face recognition from video., In Proc. European Conf. on Computer Vision, volume 4, pp: 732-746. [10] S. Zhou, V. Krueger, R. Chellappa, Face recognition from video: A condensation approach , in IEEE Int. Conf. on Automatic Face and Gesture Recognition, 2002, pp. 221-228. [11] Abdenour Hadid; Matti Pietikinen; Combining appearance and motion for face and gender recognition from videos, Pattern Recognition, Vol:42(11),2009,pp: 2818-2827. [12] Chellappa, R. Kruger, V..Shaohua Zhou, Probabilistic recognition of human faces from video, 2002 International Conference on ImageProcessing, Vol 1, 2002, pp. 41-45. [13] Gregory Shakhnarovich, Baback Moghaddam Face Recognition in Subspaces, Handbook of Face Recognition, 2004. [14] N. Vaswani and R. Chellappa, Principal components null space analysis for image and video classification, IEEETransactions on Image Processing, vol. 15, no. 7, pp. 18161830, 2006. [15] S. Soatto, G. Doretto, and Y. Wu, Dynamic textures, in Proceedings of the International Conference on Computer Vision, vol. 2, pp. 439446, Vancouver, Canada, 2001. [16] M. Kim, S. Kumar, V. Pavlovic, and H. Rowley, Face tracking and recognition with visual constraints in real-world videos, in Proceedings of the 26th IEEE Conference on Computer Vision and Pattern Recognition (CVPR '08), June 2008. [17] C. Shan, S. Gong, P. Mcowan, Learning gender from human gaits and faces,IEEE International Conference on Advanced Video and Signal based Surveillance, 2007, pp:505-510. [18] Christian Micheloni , Sergio Canazza , Gian Luca Foresti; Audio-video biometric recognition for non-collaborative access granting; Visual Languages and Computing,2009. [19] M. Balasubramanian , S. Palanivela, and V. Ramalingama; Real time face and mouth recognition using radial basis function neural networks; Expert Systems with Applications, Vol:36(3), pp: 6879-6888. [20] Georghiades, A., Kriegman, D., & Belhumeur, P. From few to many:Generative models for recognition under variable pose and illumination. IEEE Transactions Pattern Analysis and Machine Intelligence, 40,643-660. 2001. [21] M. Everingham and A. Zisserman, Identifying individuals in video by combinig 'generative' and discriminative head models, in Proceedings of the 10th IEEE International Conference on Computer Vision (ICCV '05). [22] Dalong Jiang,Yuxiao Hu, Shuicheng Yan, Lei Zhang, Hongjiang Zhang, Wen Gao, Efficient 3D reconstruction for face recognition, Pattern Recognition 38 (2005), pp:787 798 [23] Y. Xu, A. Roy-Chowdhury and K. Patel, Integrating illumination, motion and shape models for robust face recognition in video, EURASIP J. Adv. Signal Process. (2008) [24] H. Eidenberger. Kalman filtering for pose-invariant face recognition. In Image Processing, 2006 IEEE International Conference on, pages 20372040. IEEE. [25] U. Park, A.K. Jain, and A. Ross. Face recognition in video: Adaptive fusion of multiple matchers, In 2007 IEEE Conference on Computer Vision and Pattern Recognition, pages 18. IEEE, 2007.
Page 214
J.SUNETHA received the B.Tech. Degree in Computer Science from Jawaharlal Nehru Technological University, Hyderabad and M.Tech. Degree in Computer Science from Sathyabhama University, Chennai.. She is Research Scholar at Sri padmavathi Mahila Viswa Vidyalayam,Tirupathi, Andhra Pradesh, India.
Page 215