Liu 2018
Liu 2018
Liu 2018
Abstract— Face recognition (FR) systems in real-world appli- occluded FR is twofold. Firstly, occlusions distort the discrim-
cations need to deal with a wide range of interferences, such as inative facial features and increases the distance between two
occlusions and disguises in face images. Compared with other face images of the same subject in the feature space. The
forms of interferences such as nonuniform illumination and pose
changes, face with occlusions has not attracted enough attention intra-class variations are larger than the inter-class variations,
yet. A novel approach, coined dynamic image-to-class warping which results in poorer recognition performance. Secondly,
(DICW), is proposed in this work to deal with this challenge when facial landmarks are occluded, large registration errors
in FR. The face consists of the forehead, eyes, nose, mouth, and usually occur and degrade the recognition rate [1].
chin in a natural order and this order does not change despite Note that there are two related but different problems to FR
occlusions. Thus, a face image is partitioned into patches, which
are then concatenated in the raster scan order to form an ordered with occlusions: occluded face detection and occluded face
sequence. Considering this order information, DICW computes recovery. The first task is to determine whether a face image
the image-to-class distance between a query face and those of an is occluded or not [2], which can be used for automatically
enrolled subject by finding the optimal alignment between the rejecting the occluded images in applications such as passport
query sequence and all sequences of that subject along both the image enrolment. This rejection mechanism is not always
time dimension and within-class dimension. Unlike most existing
methods, our method is able to deal with occlusions which exist in applicable in some scenarios (e.g., surveillance) where no
both gallery and probe images. Extensive experiments on public alternative image can be obtained due to the lack of user
face databases with various types of occlusions have confirmed cooperation. The second task is to restore the occluded region
the effectiveness of the proposed method. in face images [3], [4]. It can recover the occluded area but is
Index Terms— Face recognition, occlusion, image-to-class unable to directly contribute to recognition since the identity
distance, dynamic time warping, biometrics. information can be contaminated during inpainting.
An intuitive idea for handling occlusions in FR is to detect
I. I NTRODUCTION the occluded region first and then perform recognition using
only the unoccluded part. Min et al. [5] adopted a SVM
F ACE recognition (FR) is one of the most active research
topics in computer vision and patten recognition over the
past few decades. Nowadays, automatic FR system achieves
classifier to detect the occluded region in a face image then
used only the unoccluded area of a probe face (i.e., query
face) as well as the corresponding area of the gallery faces
significant progress in controlled conditions. However, the
(i.e., reference faces) for recognition. But note that the occlu-
performance in unconstrained conditions (e.g., large variations
sion types in the training images are the same as those in the
in illumination, pose, expression, etc.) is still unsatisfactory.
testing images. Jia and Martinez [6], [7] used a skin colour
In the real-world environments, faces are easily occluded by
based mask to remove the occluded area for recognition.
facial accessories (e.g., sunglasses, scarf, hat, veil), objects in
However, the types of occlusions are unpredictable in practical
front of the face (e.g., hand, food, mobile phone), extreme
scenarios. The location, size and shape of occlusions are
illumination (e.g., shadow), self-occlusion (e.g., non-frontal
unknown, hence increasing the difficulty in segmenting the
pose) or poor image quality (e.g., blurring). The difficulty of
occluded region from the face images. Currently most of the
Manuscript received March 15, 2014; revised July 11, 2014; accepted occlusion detectors are trained on faces with specific types
September 3, 2014. Date of publication September 22, 2014; date of current of occlusions (i.e., the training is data-dependent) and hence
version November 10, 2014. The work of Z. Lei was supported by the National
Natural Science Foundation of China under Project 61103156 and Project generalise poorly to various types of occlusions in the real-
61473291. The associate editor coordinating the review of this manuscript world environment.
and approving it for publication was Prof. Gérard Medioni. (Corresponding In this paper, we focus on performing recognition in the
author: Chang-Tsun Li.)
X. Wei and C.-T. Li are with the Department of Computer Science, Univer- presence of occlusions. There are two main categories of
sity of Warwick, Coventry CV4 7AL, U.K. (e-mail: x.wei@warwick.ac.uk; approaches in this direction. The first is the reconstruction
ct.li@warwick.ac.uk). based approaches which treat occluded FR as a reconstruction
Z. Lei, D. Yi, and S. Z. Li are with the Center for Biometrics and
Security Research & National Laboratory of Pattern Recognition, Institute of problem [6], [8]–[13]. The sparse representation based classi-
Automation, Chinese Academy of Sciences, Beijing 100190, China (e-mail: fication (SRC) [8] is a representative example. A clean image
zlei@cbsr.ia.ac.cn; dyi@cbsr.ia.ac.cn; szli@cbsr.ia.ac.cn). is reconstructed from an occluded probe image by a linear
Color versions of one or more of the figures in this paper are available
online at http://ieeexplore.ieee.org. combination of gallery images. Then the occluded image is
Digital Object Identifier 10.1109/TIFS.2014.2359632 assigned to the class with the minimal reconstruction error.
1556-6013 © 2014 IEEE. Translations and content mining are permitted for academic research only. Personal use is also permitted,
but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
2036 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 9, NO. 12, DECEMBER 2014
TABLE I
T HREE T YPICAL O CCLUSION C ASES
The reconstruction based approaches usually require a large finding the optimal alignment between the query sequence
number of samples per subject to represent a probe image. and all enrolled sequences of that subject. Our method allows
However, a sufficient number of samples are not always elastic match in both time and with-class directions.
available in practical scenarios. Most of the existing works that simply treat occluded FR as
The second category is the local matching based a signal recovery problem or just employ the framework for
approaches [14]–[17]. Facial features are extracted from local general object classification, neglect the inherent structure of
areas of a face, for example, overlapping or non-overlapping the face. Wang et al. proposed a Markov Random Field (MRF)
patches of an image, so the affected and unaffected parts of based method [23] for FR and confirmed that the contextual
the face can be analysed in isolation. In order to minimise information between facial features plays an important role
matching errors due to occluded parts, different strategies in recognition. In this paper, we propose a novel approach
such as local subspace learning [14], [15], partial distance that takes the facial order, which contains the geometry infor-
learning [16] and multi-task sparse representation learning [17] mation of the face, into account when recognising partially
are performed. Our method belongs to this category but does occluded faces. In uncontrolled environments with uncooper-
not require training. ative subjects, the occlusion preprocessing and the collection
In addition to the above approaches, which focus on of sufficient and representative training samples are generally
improving the robustness during the recognition stage, recently very difficult. Our method which performs recognition directly
many researchers [18], [19] also pay attention to the image in the presence of occlusions and does not require training, is
presentation stage and attempt to extract stable, occlusion- hence feasible for realistic FR applications.
insensitive features from face images. Since the forms of This paper is built upon our preliminary work reported
occlusions in real-world scenarios are unpredictable, it is still in [24] and [25]. The remainder of this paper is organised
difficult to find a suitable representation which is insensitive as follows. The proposed Dynamic Image-to-Class Warping
to the variations in occlusions. method, from image representation, modelling to implementa-
Most of the current methods assume that occlusions only tion, is described in Section II. Extensive experiments includ-
exist in the probe images and the gallery or training images ing discussions are presented in Section III. Further analysis
are clean. In practical scenarios, occlusions may occur in both about why the proposed method works; when and why it will
gallery and probe images [6], [7], [20]. When the number of fail and how to improve it is discussed in Section IV. Finally
gallery/training images is limited, excluding these occluded the work is concluded in Section V.
images would, on the one hand, lead to small sample size (SSS)
problem [21], and on the other hand, ignore useful information
for recognition [20]. We summarise three occlusion cases in II. DYNAMIC I MAGE - TO -C LASS WARPING
Table I, which a FR system may encounter in the real-world
A. Image Representation
applications. Most of the current methods rely on a clean
gallery or training set and only consider the first case. The An image is partitioned into J non-overlapping patches of
latter two cases would also occur in real environment but have d ×d pixels. Those patches are then concatenated in the raster
not yet received much attention. scan order (i.e., from left to right and top to bottom) to form a
We propose a local matching based method, Dynamic single sequence. The reason for doing so is that the forehead,
Image-to-Class Warping (DICW), for occluded FR. eyes, nose, mouth and chin are located in the face in a natural
DICW is motivated by the Dynamic Time Warping (DTW) order, which does not change despite occlusions or imprecise
algorithm [22] which allows elastic match of two time registration. This spatial facial order, which is contained in
sequences. It has been successfully applied to the area of the patch sequence, can be viewed as the temporal order in
speech recognition [22]. In our work, an image is partitioned the time sequence. In this way, a face image can be viewed
into patches, which are then concatenated in the raster scan as a time sequence so the image matching problem can be
order to form a sequence. In this way, a face is represented handled by the time series analysis technique like DTW [22].
by a patch sequence which contains the order information of Let f (x, y) be the intensity of the pixel at coordinates (x, y)
facial features. DICW calculates the Image-to-Class distance and f j (x, y) be the j -patch. A difference patch f j (x, y) is
between a query face and those of an enrolled subject by computed (Fig. 1) by subtracting f j (x, y) from its immediate
WEI et al.: DYNAMIC IMAGE-TO-CLASS WARPING FOR OCCLUDED FR 2037
Fig. 2. Distributions of face image distance of the same and different classes. Using the difference patch (b), the distance distribution of the same class and
that of the different classes are separated more widely compared with those using the original patch (a).
Fig. 4. An illustration of warping path in (a) DTW and the (b) proposed DICW. The arrows indicate the matching correspondence. The dashed line marks
the optimal warping path. Black blocks indicate the occluded patches.
Fig. 7. Uvs.O: identification results on the FRGC database with different number of gallery images per subject: a) K = 1, b) K = 2, c) K = 3 and d) K = 4.
Fig. 8. a) Ovs.U and b) Ovs.O: identification results on the FRGC database with occlusions in gallery or/and probe sets.
following settings: 1) 400 occluded images (four images per occluded images as the gallery set and 400 occluded images
subject) from the original set as the gallery set and 400 images as the probe set (Ovs.O). Note that the images in the gallery
from the unoccluded set as the probe set (Ovs.U) and 2) 400 set are different from those in the probe sets. Fig. 8 shows the
2042 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 9, NO. 12, DECEMBER 2014
Fig. 9. Uvs.O: identification results on the AR database with (a) sunglasses occlusion and (b) scarf occlusion.
TABLE III
U VS .O: C OMPARISON OF DICW AND THE -S TATE - OF - THE -A RT M ETHODS (K = 8)
scarves as the gallery set and two images with sunglasses as C. Face Identification With General Occlusions
the probe set; 2) vice versa. Note that with this setting, in each in Realistic Environment
test the occlusion type in the gallery set is different from that In this Section, we test our method on the The Face We
in the probe set, which is very challenging for recognition. Make (TFWM) [28] database captured under natural and arbi-
The results are shown in Fig. 10. On the gallery set which trary conditions. It has more than 2,000 images which contains
contains occluded faces, the results of HMM and SVM are frontal view faces of strangers on the streets with uncon-
much worse than others as expected. In the Ovs.O testing, trolled lighting. The sources of occlusions include glasses,
there are only two gallery images per subject. It is very difficult sunglasses, hat, hair and hand on the face. Besides occlusions,
for SRC to reconstruct an unoccluded probe image with these images also contain expression, pose and head rotation
such limited number of gallery images. Local matching based variations. In our experiments, we use images of 100 subjects
2044 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 9, NO. 12, DECEMBER 2014
Fig. 16. Sample images of the same subject from the AR database without
alignment (AR-VJ).
TABLE IV
I DENTIFICATION R ATE (%) ON THE AR-VJ D ATASET
Fig. 15. Identification rates (%) of using different image descriptors and the
difference patches.
Fig. 17. Sample images from the LFW database (six matched image pairs
for six subjects).
TABLE VI
C OMPARISON OF AVERAGE RUNTIME ( S )
IV. F URTHER A NALYSIS AND I MPROVEMENT Fig. 19. (a) The probe image from class 74. (b) Classification result (class 5)
In the previous sections, we evaluate DICW using exten- by NBNN. (c) Classification result (class 74) by DICW. Distance to each class
computed by (d) NBNN and by (e) DICW.
sive experiments with face images with various uncontrolled
variations. In this section we will further analysis why the
DICW works compared with similar methods, and when and
why it will fail. We also discuss the idea for improving the
performance of DICW.
patches within a proper range. This is guaranteed by the four
NBNN [34] presented in the previous sections is a similar
constraints mentioned in Section II-B. In addition, with the
method to ours. It also calculates the Image-to-Class distance
help of Dynamic Programming, DICW actually tries every
between a probe patch set and a gallery patch set from a given
possible combination of matching correspondence of patch
class. The difference is that it does not consider the spatial
pairs so the final matching is the global optimum for the
relationship between patches like ours and each probe patch
probe patch set and the gallery patch set. Compared with
can be matched to any patches from any location in the gallery
NBNN, DICW considers both the texture similarity and the
patch set. Fig. 19 is an illustration example. The occluded
geometry similarity of patches. The work in [23] points out
probe image is from class 74 but is incorrectly classified to
that the contextual information between facial features plays
the class 5 by NBNN. Actually the images from class 74
an important role in recognition. Our work confirms their
and class 5 are not alike. But the texture of sunglasses is
observation. Although a location weight can be adopted in
very similar to that of beard in class 5. Without the location
NBNN, the weight needs to be manually set for different
constraint, the beard patches are wrongly matched to the
testing dataset as analysed before, which is not suitable
sunglasses thus the distance is affected by this occlusion.
for practical applications. In DICW, the order constraint is
On the other hand, DICW keeps the order information and
naturally encoded during distance computation.
matches patches within a proper range which leads to correct
DICW represents a face image as a patch sequence which
classification.
maintains the facial order of a face. To some extent, the
NBNN calculates the distance between two patch sets and
geometric information of a face is reduced from 2D to 1D.
the overall distance is the sum of patch-pair distances. On the
However, the direct 2D image warping is an NP-complete
other hand, in DICW, the probe and gallery patch set are
problem [55]. P2DW-FOSE mentioned in Section III-D4 is a
ordered. The spatial relationship between patches is encoded.
pseudo 2D warping method but with a remarkably large com-
When a probe patch is matched to a gallery patch, the
putational cost (i.e., quadratic in the number of pixels) [45].
following probe patches will only be matched to the gallery
DICW incurs a lower computational cost due to its patch
3 We use the l1_ls package for implementation. http://www.stanford.edu/~ sequence representation. In addition, each patch still contains
boyd/l1_ls/ the local 2D information which is helpful for classification.
2048 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 9, NO. 12, DECEMBER 2014
TABLE VII
I DENTIFICATION R ATES (%) OF DICW AND THE I MPROVEMENT
S CHEME ON THE AR D ATABASE
Fig. 20. (a) A probe image from class 51. (b) The wrong class (class 72)
classified by DICW. (c) The gallery image from class 51.
[4] T. Hosoi, S. Nagashima, K. Kobayashi, K. Ito, and T. Aoki, [27] P. J. Phillips et al., “Overview of the face recognition grand challenge,”
“Restoring occluded regions using FW-PCA for face recognition,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit.
in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. (CVPR), vol. 1. Jun. 2005, pp. 947–954.
Workshops (CVPRW), Jun. 2012, pp. 23–30. [28] D. Miranda. The Face We Make. [Online]. Available:
[5] R. Min, A. Hadid, and J.-L. Dugelay, “Improving the recognition of http://www.thefacewemake.org, accessed 2014.
faces occluded by facial accessories,” in Proc. IEEE Int. Conf. Autom. [29] G. B. Huang, M. Ramesh, T. Berg, and E. Learned-Miller, “Labeled
Face Gesture Recognit. (FG), Mar. 2011, pp. 442–447. faces in the wild: A database for studying face recognition in
[6] H. Jia and A. M. Martinez, “Face recognition with occlusions in the unconstrained environments,” Dept. Comput. Sci., Univ. Massachusetts,
training and testing sets,” in Proc. IEEE Int. Conf. Autom. Face Gesture Amherst, MA, USA, Tech. Rep. 07-49, Oct. 2007.
Recognit. (FG), Sep. 2008, pp. 1–6. [30] S. U. Hussain, T. Napoléon, and F. Jurie, “Face recognition using local
[7] H. Jia and A. M. Martinez, “Support Vector Machines in face recog- quantized patterns,” in Proc. Brit. Mach. Vis. Conf. (BMVC), 2012,
nition with occlusions,” in Proc. IEEE Conf. Comput. Vis. Pattern pp. 99.1–99.11.
Recognit. (CVPR), Jun. 2009, pp. 136–141. [31] T. Ahonen, A. Hadid, and M. Pietikainen, “Face description with local
[8] J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, “Robust face binary patterns: Application to face recognition,” IEEE Trans. Pattern
recognition via sparse representation,” IEEE Trans. Pattern Anal. Mach. Anal. Mach. Intell., vol. 28, no. 12, pp. 2037–2041, Dec. 2006.
Intell., vol. 31, no. 2, pp. 210–227, Feb. 2009. [32] C.-C. Chang and C.-J. Lin, “LIBSVM: A library for sup-
[9] A. I. Naseem, B. R. Togneri, and C. M. Bennamoun, “Linear regression port vector machines,” ACM Trans. Intell. Syst. Technol., vol. 2,
for face recognition,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 32, no. 3, pp. 27:1–27:27, Apr. 2011. [Online]. Available: http://www.
no. 11, pp. 2106–2112, Nov. 2010. csie.ntu.edu.tw/~cjlin/libsvm
[10] M. Yang and L. Zhang, “Gabor feature based sparse represen- [33] M. A. Turk and A. P. Pentland, “Face recognition using eigenfaces,” in
tation for face recognition with Gabor occlusion dictionary,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 1991,
Proc. 11th Eur. Conf. Comput. Vis. (ECCV), vol. 6316. 2010, pp. 586–591.
pp. 448–461. [34] O. Boiman, E. Shechtman, and M. Irani, “In defense of nearest-neighbor
[11] D. Zhang, M. Yang, and X. Feng, “Sparse representation or based image classification,” in Proc. IEEE Conf. Comput. Vis. Pattern
collaborative representation: Which helps face recognition?” Recognit. (CVPR), Jun. 2008, pp. 1–8.
in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), Nov. 2011, [35] A. V. Nefian and M. H. Hayes, III, “Hidden Markov models for face
pp. 471–478. recognition,” in Proc. IEEE Int. Conf. Acoust., Speech Signal Process.,
[12] R. He, W.-S. Zheng, and B.-G. Hu, “Maximum correntropy criterion vol. 5. May 1998, pp. 2721–2724, doi: 10.1109/ICASSP.1998.678085.
for robust face recognition,” IEEE Trans. Pattern Anal. Mach. Intell., [36] K. Jia, T.-H. Chan, and Y. Ma, “Robust and practical face recognition
vol. 33, no. 8, pp. 1561–1576, Aug. 2011. via structured sparsity,” in Proc. 12th Eur. Conf. Comput. Vis. (ECCV),
[13] X. Wei, C.-T. Li, and Y. Hu, “Robust face recognition under vary- vol. 7575. 2012, pp. 331–344.
ing illumination and occlusion considering structured sparsity,” in [37] R. Weng, J. Lu, J. Hu, G. Yang, and Y.-P. Tan, “Robust feature set
Proc. Int. Conf. Digit. Image Comput. Techn. Appl. (DICTA), 2012, matching for partial face recognition,” in Proc. IEEE Int. Conf. Comput.
pp. 1–7. Vis. (ICCV), Dec. 2013, pp. 601–608.
[14] A. M. Martinez, “Recognizing imprecisely localized, partially occluded, [38] A. M. Martinez and A. C. Kak, “PCA versus LDA,” IEEE Trans. Pattern
and expression variant faces from a single sample per class,” IEEE Anal. Mach. Intell., vol. 23, no. 2, pp. 228–233, Feb. 2001.
Trans. Pattern Anal. Mach. Intell., vol. 24, no. 6, pp. 748–763, [39] D. G. Lowe, “Distinctive image features from scale-invariant keypoints,”
Jun. 2002. Int. J. Comput. Vis., vol. 60, no. 2, pp. 91–110, Nov. 2004.
[15] X. Tan, S. Chen, Z.-H. Zhou, and F. Zhang, “Recognizing partially [40] H. Bay, T. Tuytelaars, and L. Van Gool, “SURF: Speeded up robust
occluded, expression variant faces from single training image per person features,” in Proc. 9th Eur. Conf. Comput. Vis. (ECCV), vol. 3951. 2006,
with SOM and soft k-NN ensemble,” IEEE Trans. Neural Netw., vol. 16, pp. 404–417.
no. 4, pp. 875–886, Jul. 2005. [41] Z. Zhou, A. Wagner, H. Mobahi, J. Wright, and Y. Ma, “Face recognition
[16] X. Tan, S. Chen, Z.-H. Zhou, and J. Liu, “Face recognition under with contiguous occlusion using Markov random fields,” in Proc. IEEE
occlusions and variant expressions with partial similarity,” IEEE Trans. 12th Int. Conf. Comput. Vis. (ICCV), Sep./Oct. 2009, pp. 1050–1057.
Inf. Forensics Security, vol. 4, no. 2, pp. 217–230, Jun. 2009. [42] W.-H. Lai and C.-T. Li, “Skin colour-based face detection in
[17] S. Liao, A. K. Jain, and S. Z. Li, “Partial face recognition: Alignment- colour images,” in Proc. IEEE Int. Conf. Adv. Video Signal-Based
free approach,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 35, no. 5, Surveill. (AVSS), Nov. 2006, pp. 56–61.
pp. 1193–1205, May 2013. [43] C. Liu and H. Wechsler, “Gabor feature based classification using the
[18] G. Tzimiropoulos, S. Zafeiriou, and M. Pantic, “Subspace learning from enhanced fisher linear discriminant model for face recognition,” IEEE
image gradient orientations,” IEEE Trans. Pattern Anal. Mach. Intell., Trans. Image Process., vol. 11, no. 4, pp. 467–476, Apr. 2002.
vol. 34, no. 12, pp. 2454–2466, Dec. 2012. [44] S. Lazebnik, C. Schmid, and J. Ponce, “Beyond bags of features: Spatial
[19] J. Zhu, D. Cao, S. Liu, Z. Lei, and S. Z. Li, “Discriminant analysis with pyramid matching for recognizing natural scene categories,” in Proc.
Gabor phase for robust face recognition,” in Proc. 5th IAPR Int. Conf. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. (CVPR), vol. 2.
Biometrics (ICB), Mar./Apr. 2012, pp. 13–18. Jun. 2006, pp. 2169–2178.
[20] C.-F. Chen, C.-P. Wei, and Y.-C. F. Wang, “Low-rank matrix recov- [45] L. Pishchulin, T. Gass, P. Dreuw, and H. Ney, “Image warping for face
ery with structural incoherence for robust face recognition,” in Proc. recognition: From local optimality towards global optimization,” Pattern
IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2012, Recognit., vol. 45, no. 9, pp. 3131–3140, 2012.
pp. 2618–2625. [46] P. Viola and M. J. Jones, “Robust real-time face detection,” Int. J. Com-
[21] X. Tan, S. Chen, Z.-H. Zhou, and F. Zhang, “Face recognition from a put. Vis., vol. 57, no. 2, pp. 137–154, 2004.
single image per person: A survey,” Pattern Recognit., vol. 39, no. 9, [47] K. Tan and S. Chen, “Adaptively weighted sub-pattern PCA for face
pp. 1725–1745, 2006. recognition,” Neurocomputing, vol. 64, pp. 505–511, Mar. 2005.
[22] H. Sakoe and S. Chiba, “Dynamic programming algorithm optimization [48] P. Dreuw, P. Steingrube, H. Hanselmann, and H. Ney, “SURF-face:
for spoken word recognition,” IEEE Trans. Acoust., Speech, Signal Face recognition under viewpoint consistency constraints,” in Proc. Brit.
Process., vol. 26, no. 1, pp. 43–49, Feb. 1978. Mach. Vis. Conf. (BMVC), 2009, pp. 7.1–7.11.
[23] R. Wang, Z. Lei, M. Ao, and S. Z. Li, “Bayesian face recognition [49] D. Chen, X. Cao, F. Wen, and J. Sun, “Blessing of dimensionality: High-
based on Markov random field modeling,” in Proc. IAPR 3rd Int. Conf. dimensional feature and its efficient compression for face verification,”
Biometrics (ICB), vol. 5558. 2009, pp. 42–51, doi: 10.1007/978-3-642- in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2013,
01793-3_5. pp. 3025–3032.
[24] X. Wei, C.-T. Li, and Y. Hu, “Face recognition with occlusion using [50] T. F. Cootes, C. J. Taylor, D. H. Cooper, and J. Graham, “Active shape
dynamic image-to-class warping (DICW),” in Proc. IEEE Int. Conf. models-their training and application,” Comput. Vis. Image Understand.,
Autom. Face Gesture Recognit. (FG), Apr. 2013, pp. 1–6. vol. 61, no. 1, pp. 38–59, Jan. 1995.
[25] X. Wei, C.-T. Li, and Y. Hu, “Robust face recognition with occlusions [51] S. Z. Li, D. Yi, Z. Lei, and S. Liao, “The CASIA NIR-VIS 2.0
in both reference and query images,” in Proc. Int. Workshop Biometrics face database,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.
Forensics, Apr. 2013, pp. 1–4. Workshops (CVPRW), Jun. 2013, pp. 348–353.
[26] A. M. Martinez and R. Benavente, “The AR face database,” [52] H. J. Seo and P. Milanfar, “Face verification using the LARK
Autonomous Univ. Barcelona, Barcelona, Spain, Tech. Rep. 24, representation,” IEEE Trans. Inf. Forensics Security, vol. 6, no. 4,
1998. pp. 1275–1286, Dec. 2011.
2050 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 9, NO. 12, DECEMBER 2014
[53] G. Sharma, S. ul Hussain, and F. Jurie, “Local higher-order Zhen Lei received the B.S. degree in automation
statistics (LHS) for texture categorization and facial analysis,” in from the University of Science and Technology of
Proc. 12th Eur. Conf. Comput. Vis. (ECCV), vol. 7578. 2012, pp. 1–12. China, Hefei, China, in 2005, and the Ph.D. degree
[54] J. Ruiz-del Solar, R. Verschae, and M. Correa, “Recognition of faces in from the Institute of Automation, Chinese Academy
unconstrained environments: A comparative study,” EURASIP J. Adv. of Sciences, Beijing, China, in 2010, where he
Signal Process., vol. 2009, pp. 1:1–1:19, Jan. 2009. is currently an Associate Professor. His research
[55] D. Keysers and W. Unger, “Elastic image matching is NP-complete,” interests are in computer vision, pattern recognition,
Pattern Recognit. Lett., vol. 24, nos. 1–3, pp. 445–453, Jan. 2003. image processing, and in particular, face recognition.
[56] X. Wei and C.-T. Li, “Fixation and saccade based face recognition from He has authored over 80 papers in international
single image per person with various occlusions and expressions,” in journals and conferences. He serves as the Area
Proc. IEEE Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), Chair the of International Joint Conference on Bio-
Jun. 2013, pp. 70–75. metrics in 2014, the IAPR/IEEE International Conference on Biometric in
2015, and the IEEE International Conference on Automatic Face and Gesture
Recognition in 2015.