Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Optical Vapor Sensing on Single Wing Scales and on Whole Wings of the Albulina metallica Butterfly
Previous Article in Journal
Reference Phase Stabilizer for Distributed Underwater Sonar Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Highly Accurate and Fully Automatic 3D Head Pose Estimation and Eye Gaze Estimation Using RGB-D Sensors and 3D Morphable Models

by
Reza Shoja Ghiass
1,
Ognjen Arandjelovć
2 and
Denis Laurendeau
1,*
1
Computer Vision and Systems Laboratory, Laval University, 1665 Rue de l’Universite, Universite Laval, Quebec City, QC G1V 0A6, Canada
2
School of Computer Science, University of St Andrews, St Andrews, KY16 9SX Scotland, UK
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(12), 4280; https://doi.org/10.3390/s18124280
Submission received: 14 October 2018 / Revised: 27 November 2018 / Accepted: 30 November 2018 / Published: 5 December 2018
(This article belongs to the Section Physical Sensors)

Abstract

:
This work addresses the problem of automatic head pose estimation and its application in 3D gaze estimation using low quality RGB-D sensors without any subject cooperation or manual intervention. The previous works on 3D head pose estimation using RGB-D sensors require either an offline step for supervised learning or 3D head model construction, which may require manual intervention or subject cooperation for complete head model reconstruction. In this paper, we propose a 3D pose estimator based on low quality depth data, which is not limited by any of the aforementioned steps. Instead, the proposed technique relies on modeling the subject’s face in 3D rather than the complete head, which, in turn, relaxes all of the constraints in the previous works. The proposed method is robust, highly accurate and fully automatic. Moreover, it does not need any offline step. Unlike some of the previous works, the method only uses depth data for pose estimation. The experimental results on the Biwi head pose database confirm the efficiency of our algorithm in handling large pose variations and partial occlusion. We also evaluated the performance of our algorithm on IDIAP database for 3D head pose and eye gaze estimation.

1. Introduction

Head pose estimation is a key step in understanding human behavior and can have different interpretations depending on the context. From the computer vision point of view, head pose estimation is the task of inferring the direction of head from digital images or range data compared to the imaging sensor coordinate system. In the literature, the head is assumed to be a rigid object with three degrees of freedom, i.e., the head pose estimation is expressed in terms of yaw, roll and pitch. Generally, the previous works on head pose estimation can be divided into two categories: (i) the methods based on 2D images; and (ii) depth data [1]. The pose estimators based on 2D images generally require some pre-processing steps to translate the pixel-based representation of the head into some direction cues. Several challenges such as camera distortion, projective geometry, lighting or changes in facial expression exist in 2D image-based head pose estimators. A comprehensive study of pose estimation is given in [1] and the reader can refer to this reference for more details on the literature.
Unlike the 2D pose estimators, the systems based on 3D range data or their combination with 2D images have demonstrated very good performance in the literature [2,3,4,5,6,7]. While most of the work on 3D pose estimation in the literature is based on non-consumer level sensors [8,9,10], recent advances in production of consumer level RGB-D sensors such as the Microsoft Kinect or the Asus Xtion has facilitated the design and implementation of real-time facial performance capture systems such as consumer-level 3D pose estimators, 3D face tracking systems, 3D facial expression capture systems and 3D eye gaze estimators. In this paper, we focus on the recent 3D pose estimators and tracking systems and their application in appearance-based eye gaze estimation systems using consumer level RGB-D sensors.

1.1. Related Work on 3D Pose Estimation Using RGB-D Sensors

The 3D head pose estimation systems can be divided into three categories: (i) statistical approaches; (ii) model-based posed estimation methods; and (iii) facial feature-based pose estimation techniques [11]. Each of these approaches comes with their specific limits and advantages. Statistical methods may need a large database for training a regressor. However, they can estimate the subject head pose on air, i.e., the system can estimate the head pose for each frame even in a shuffled video sequence. In contrast, model-based approaches generally need an offline step for subject-specific head model reconstruction with significant subject cooperation. Next, a point cloud registration technique such as rigid/non-rigid ICP should be used to register the model with depth data. In other words, unlike the supervised learning based approaches, they are generally based on tracking. Thus, re-initialization becomes a challenge. Facial feature-based pose estimation techniques try to track facial features or patches, which, in turn, can help in calculation of pose using techniques such as PnP [10] or encoding the face 3D shape using view-invariant descriptors and infer head pose through matching [12].
To the best of our knowledge, one of the most important works on pose estimation using consumer level RGB-D sensors is the work of Fanelli et al. [2,3]. As the authors provided a ground truth data and a database for comparison, their work has become the gold standard for comparison in the literature. Their work falls in the category of statistical approaches. In their work, the authors proposed a pose estimation system based on Random Forests. For the evaluation of their system, they acquired a database of 20 subjects, which is called the Biwi head pose database. Next, they divided the database into a training and test sets. Afterwards, a commercial face tracker is used for annotation of the training set, i.e., a subject-specific head model is constructed using the commercial system to match each person’s identity and track the head in training depth frames. The commercial tracker measures a subject’s 3D head locations and orientations, which, in turn, are used to train their regression based system. Finally, some patches of fixed size from the region of the image containing the head as positives samples, and from outside the head region as negatives were randomly selected for training the system. A major limitation of this system is that it requires an offline training phase with subject cooperation. Moreover, the performance of the system in the testing phase is subject to the output of the commercial head tracker in the training phase. In [3], the authors continued their previous work [2] by creating a dataset of synthetic depth images of heads, and extracting the positive patches from the synthetic data, while using the original depth data to to extract negative patches. A drawback of this system is the limited number of synthetic models and negative patches for performing a regression task, without learning subject’s own head [3]. Ref. [13] proposed a system based on cascaded tree classifiers with higher accuracies than Fanelli et al. [8] proposes a 3D face tracker based on particle filters. The main idea in their system is the combination of depth and 2D image data in the observation model of the particle filter.

1.2. Related Work on 3D Gaze Estimation Using RGB-D Sensors

Based on the context, the term Gaze Estimation can be interpreted as one of the following closely related concepts: (i) 3D Line of Sight (LoS); (ii) 3D Line of Gaze (LoG); and (iii) Point of Regard (PoR). Within the eyeball coordinate system, the LoG is simply the optical axis, while the LoS is the ray pointing out from fovea and eyeball rotation center. The PoR is a 3D point in the scene to which the LoS points. Figure 1 demonstrates a simplified schematic of human eye with LoS, GoS and PoR. In the literature non-intrusive gaze estimation approaches generally fall into one of the following categories: (i) feature-based approaches; and (ii) appearance-based approaches [14].
Feature-based approaches extract some eye specific features such as eye corner, eye contour, limbus, iris, pupil, etc. These features may be aggregated with the reflection of external light setup on the eye (called glints or Purkinje images) to infer the gaze. These methods are generally divided into two categories: (i) model-based (geometric); and (ii) interpolation-based.
Model based methods rely on the geometry of the eye. These methods directly calculate the point of regard by calculating the gaze direction (LoS) first. Next, the intersection of the gaze direction and the nearest object of the scene (e.g., a monitor in many applications) generates the point of regard. Most of the model-based approaches require some prior knowledge such as the camera calibration or the global geometric model of the external lighting setup [15,16,17,18,19,20].
Unlike the model-based methods, the interpolation-based approaches do not perform an explicit calculation about the LoS. Instead, they rely on a training session based on interpolation (i.e., a regression problem in a supervised learning context). In these methods, the feature vector between pupil center and corneal glint is mapped to the corresponding gaze coordinates on a frontal screen. The interpolation problem is formalized using a parametric mapping function such as a polynomial transformation function. The function is used later to estimate the PoR on the screen during the testing session. A calibration board maybe used during the training session [21,22,23,24,25,26,27]. The main challenge with the interpolation-based approaches is that they can not handle the head pose movements [14]. Notice that feature-based approaches in general need high resolution images to precisely extract the eye specific features as well as the glints. Moreover, they may require external lighting setups which are not ubiquitous. This motivates the researchers to train appearance-based gaze estimators, which rely on low quality eye images (a holistic-based approach instead of feature-based). However, appearance-based approaches generally have less accuracy.
As opposed to feature-based approaches, appearance-based methods do not rely on eye-specific features. Instead, they learn a one-to-one mapping from the eye appearance (i.e., the entire eye image) to the gaze vector. Appearance-based methods do not require camera calibration or any prior knowledge on the geometry data. The reason is that the mapping is made directly on the eye image, which makes these methods suitable for gaze estimation from low resolution images, but with less accuracy. In this context, they share some similarities with interpolation-based approaches. Similar to the the interpolation-based methods, appearance-based methods do not handle the head pose.
Baluja and Pomerleau [28] first used the interpolation concept from image content to the screen coordinate. Their method is based on training a neural network. However, their method requires more than 2000 training samples. To reduce such a large number of training examples, Tan et al. [29] proposed using a linear interpolation and reconstructed a test sample from the local appearance manifold within the training data. By exploiting this topological information of eye appearance, the authors reduced the training samples to 252. Later, Lu et al. [30] proposed a similar approach to exploit topological information encoded in the two-dimensional space of gaze space. To further reduce the number of the training data, Williams et al. [31] proposed a semi-supervised sparse Gaussian process regression method S3GP. Note that most of these methods assumed a fixed head pose. Alternatively, some other researchers used head mounted setups, but these methods are no longer non-intrusive [32,33].
With the main intention of designing a gaze estimator robust to head pose, Funes and Odobez [7,34] proposed the first model-based pose estimator by building a subject-specific model-based face tracker using Iterative Closest Point (ICP) and 3D Morphable Models. Their system is not only able to estimate the pose, but is also able to track the face and stabilize it. A major limitation of their method is the offline step for subject specific 3D head model reconstruction. For this purpose, they manually placed landmarks (eye corners, eyebrows, mouth corners) on RGB image of the subject, and consequently added an extra term to the cost function in their ICP formulation. In other words, their ICP formulation is supported by a manual term. Moreover, the user has to cooperate with the system and turn their head from left to right. Recently, the authors proposed a more recent version of their system in the work of [11] without the need for manual intervention.

1.3. Contribution of the Proposed Approach

Unlike [2,3,7,34], our proposed system does not require any commercial system to learn a subject’s head or any offline step. A key contribution of our approach is to propose a method to automatically learn a subject’s 3D face rather than the entire 3D head. Consequently, we no longer need subject’s cooperation (i.e., turning their head from left to right), which is important in previous works for model-based pose estimation systems. In addition, unlike [7], our system does not require any manual intervention for model reconstruction. Instead, we rely on Haar features and boosting for facial feature detection, which, in turn, can be used for face model construction. Note that we use only one RGB frame for model reconstruction. The tracking step is based on depth frames only. After learning a subject’s face, the pose estimation task is performed by a fully automatic, user non-cooperative and generic ICP formulation without any manual term. Our ICP formulation is robustified with Tukey functions in tracking mode. Thanks to the Tukey functions, our method successfully tracks a subject face in challenging scenarios. The outline of the paper is as follows: The method details are explained in Section 2. Afterwards, the experimental results are discussed in Section 3. Finally, the conclusions are drawn in Section 4.

2. Method Details

Our method consists of four key steps: (i) geometry processing of a generic face model and the first depth frame; (ii) generic face model initialization (i.e., model positioning at the location of the head); (iii) subject-specific face model construction by morphing the initialized generic model; and (iv) tracking the face in the next depth frames using the subject-specific face model. In our proposed system, we only model the face of the subject rather than the entire head, which, in turn, helps us to design a very robust, accurate and non-cooperative head tracking and pose estimation system. To accomplish this goal, a generic model is positioned on the subject’s face in depth data. Next, it learns the subject’s face and finally starts to track it. Both positioning a generic model on the subject’s face and tracking it in depth data are accomplished using an ICP-based technique. However, the ICP registration technique which serves for positioning faces a major challenge: the generic model is a model of a complete head and not only the face. On the other hand, the depth data contain not only the face of the subject, but also other body parts such as the torso or the background (Figure 2). Note that a major difficulty with ICP is its sensitivity to outliers and missing data between two 3D point clouds. To tackle this problem in model initialization, we perform geometry processing, which is explained next.

2.1. Geometry Processing

In this step, the goal is trimming the depth data and the generic model to remove spurious data and outliers. Note that we perform this step on the first depth frame only. The reason is that we should initialize (position) the model at the position of the subject’s head, before tracking starts. For this purpose, we capture the entire environment using the Kinect. Next, we filter out the spurious point cloud and just keep the region of interest in the first depth frame, i.e., the facial surface. For this purpose, the first depth frame is automatically trimmed to discard the residual data. To this end, we need to automatically detect and localize the facial features (i.e., eyes, nose, and mouth) on the point cloud to determine the way the depth data should be trimmed. Detection of the facial features from a noisy depth frame directly is a challenge. Fortunately, the Kinect provide us the first RGB frame. Thus, the face and facial features are detected on the first RGB frame by first using Haar features and boosting [35]. Figure 3 demonstrates an example of the face and facial feature detection. As some false detections may occur, the next step is to reject them automatically. This is accomplished by utilizing the prior knowledge about the structure of a face and the relative positions of eyes, nose and mouth on a detected face.
After the features are detected on the first RGB frame, their 3D loci are determined on the first depth frame through back projection using Kinect calibration data. To trim the depth data, a 3D plane passing through the 3D coordinates of the eyes and mouth is defined and shifted by an offset equal to the distance between the left and right eye. The shifted plane is called the cropping plane. Next, the depth data beneath the plane are discarded. Figure 4 shows the 3D loci of the facial features on the corresponding depth data trimmed by the cropping plane.
Once the subject’s face is captured and trimmed in 3D, the next step is to construct a model which simulates the subjects face (rather than the complete head). The type of 3D model we use to simulate the subject’s identity is a family of Active Appearance Models (AAMs) called 3D Morphable models. Using these models, a subject’s 3D head scan can be reconstructed by adding a set of weighted principal components (PCs) to the mean shape (the mean shape is the mean of all of the 200 subject’s face scans in the database). For instance, we focus on the mean shape of the model. Similar to trimming the depth data, the mean shape of the 3D Morphable model is trimmed to facilitate the procedure of subject specific model construction through registration. In this context, a plane similar to that in Figure 4 is fitted to the model’s mean shape. Once the mean shape is trimmed, it should be scaled to the size of the subject’s face in 3D space. To this end, the model is scaled so the distance between the left and right eyes of the model and that of the subject’s face scan (i.e., the first depth frame) becomes equal. Figure 5 demonstrates the mean shape of model, m, before and after trimming.

2.2. Generic Model Positioning

After processing both model and depth data, the trimmed model is positioned on the face of the subject using rigid ICP. Figure 6 demonstrates this step for the model and first depth frame. After initializing the generic model on the subject’s face, the model is ready to morph and learn the subject’s face (Section 2.3) and track it afterwards (Section 2.4).

2.3. Learning and Modeling the Subject’s Face

Capturing the subject’s facial shape variations via morphing the mean shape is the main objective of this step. This problem can be considered as finding the weights of shape PCs in the 3D Morphable model, where each weight describes the contribution of its corresponding PC in simulating a subject’s face. Similar to the generic ICP problem, this part also can be described by minimization of a cost function. Thus, we can unify both ICP and PCA terms into a unique equation and reformulate a more generic ICP problem through minimizing the following energy function [36]:
E ( Z , d , R , t ) = ω 1 E m a t c h + ω 2 E r i g i d + + ω 3 E m o d e l E m a t c h = i = 1 n ( N i T ( z i C Y ( z i ) ) ) 2 E r i g i d = i = 1 n z i ( R x i + t ) 2 2 E m o d e l = i = 1 n z i ( P i d + m i ) 2 2
where Y is the target surface in R 3 , X is the source surface, and Z is a deformed version of X which should be aligned with Y. Notice also that C y ( z i ) is the closest point in the target surface to the point z i ( i = 1 , 2 , , n , where n is the number of points in source). In this equation, the first term is the point-to-plane matching error, the second term is the point-to-point matching error, while the third term is the model error (for more details about these error the reader is referred to [36]). The energy function can be minimized by linearizing Equation (1) and iteratively solving the following linear system:
argmin Z i t + 1 , d , R ˜ , t ˜ i = 1 n ω 1 ( n i T ( z i t + 1 C Y ( z i ) t ) ) 2 + ω 2 z i t + 1 ( R ˜ ( R x i + t ) + t ˜ ) 2 2 + ω 3 z i t + 1 ( P i d + m i ) 2 2
where t is the number of iterations, z i 0 = x i , d contains the weights of PCs, and R ˜ and t ˜ are the linear updates that we obtain for the rotation (R) and translation (t) matrices, respectively, at each iteration. Notice that n i is the normal to the surface at point C Y ( z i ) t , i.e., point to plane matching error. For more details, the reader is referred to the tutorial by Bouaziz and Pauly [36].

2.4. 3D Head Tracking and Pose Estimation

Once the model is constructed from the first depth frame, the pose (orientation alone) of the head can be calculated directly from the rotation matrix, R, in terms of roll, pitch and yaw [37]. In Section 2.3, the rotation matrix corresponding to the first depth frame of the subject is obtained during model construction. A question arises here: How can one obtain the rotation matrices for the next depth frames? Indeed, this question is addressed by 3D registration of the form of Equation (1) with some differences. The first difference is that we no longer need to capture the subject’s face variations, d, because it is calculated only once for the entire procedure. Thus, the E m o d e l term is dropped from Equation (1). The other difference is that we no longer need to trim the next depth frames. The reason is that the model is already fitted to the first depth frame during model construction (see Figure 6) and we expect the system to work in tracking mode. In tracking mode, head displacement in the next frame compared to the current frame is small and the model displacement should be very small compared to the initialization mode. Thus, instead of trimming the next depth frames, one can take advantage of registration using Tukey functions, which will filter out bad correspondences with large distances. The pose estimation procedure for the next depth frames is as follows: for the second depth frame, the model rotation and translation increments are calculated relative to that of the first depth frame. Next, the rotation and translation matrices for the second depth frame are obtained by applying the updates to the rotation and translation matrices in the first depth frame. This procedure is continued for the next frames. For each frame, the head pose can be directly calculated from the rotation matrix in terms of pitch, yaw and roll.
Robustness of Registration to Outliers
As mentioned, partial overlaps among source, target and outliers in the data are the most challenging problems in registration through ICP [38]. Two types of outliers exist: (i) outliers in the source point cloud; and (ii) outliers in the target point cloud. Discarding unreliable correspondences between the source and the target is the most common way to handle this problem. In Section 2.2, this goal was accomplished by trimming both model and depth data in the first depth frame. However, for the 3D face tracking mode, the same method cannot be used because the initialization modality is based on detection of facial features. Applying facial feature detection for each frame can decrease the frame rate at which the system operates. On the other hand, a limit of our method is that the system cannot start from an extreme pose, as the facial feature detection algorithms will fail. Fortunately, as the model is already positioned onto the face of the subject, we no longer need to trim the upcoming depth frames to perform tracking using ICP. Instead, we use Tukey functions to robustify the ICP. Tukey functions assign less weight to the bad correspondences and decrease or remove their effect on the energy function.
Robustness of Registration to Extreme Pose
A question may arise at this point: Can the method handle the case of a face with extreme pose where most facial parts cannot be sensed by the Kinect? In this case, the model should be registered with a partial point cloud of the subject’s face. This leads to increasing the number of points in the source (trimmed model) without good correspondences in the target (partial point cloud of face). As a result, such points will form bad correspondences with relatively large Euclidean distance values. Fortunately, we also address this problem by using robust functions in the tracking mode, as robust functions discard/decrease the effect of such bad correspondences in the energy function. To clarify this, notice that bad correspondences inherently produce large Euclidean distances, while this is not the case for good correspondences. On other other hand, narrow robust functions act as low pass filters and discard the bad correspondences.
Robustness of Registration to Facial Expression Changes
As we use rigid ICP, facial expression changes may be considered as a challenging factor. In this context, Funes and Odobez [7] used a mask and only considered the upper part of the face in the rigid registration part of their system. Notice that we do not use such a mask. The reason is that, most of the time, the subject may not show significant facial expression changes (such as laughing or opening the mouth). On the other hand, relying on more data of the face may result in a more robust registration task. The problem becomes more challenging if we consider that self-occlusion may occur on the upper part of the face. For these reasons, we prefer not to use a mask. Instead, we rely on the robust Tukey functions to improve the robustness in the case of facial expression changes.

2.4.1. Head Pose Stabilization

This step is a pre-processing step for gaze estimation. First, we assume that the extrinsic parameters from the camera coordinates to the 3D depth sensor coordinates are known. As soon as the head pose is calculated, the texture of the corresponding RGB frame can be back-projected to a pose free 3D head model. This pose free head model can be visualized from any direction (e.g., see Figure 7b,c), but the ideal direction is the frontal view. The next step is to crop the eye appearances from this frontal view (e.g., see Figure 7d).

2.5. Gaze Estimation: Point of Regard

In this paper, gaze estimation refers to estimation of the point of regard (PoR). Our gaze estimation system consists of two parts: (i) training; and (ii) testing. The goal of the training part is to learn the parameters of an interpolation function which maps the pose-free eye appearance to the gaze vectors. In the training phase, the gaze vectors are known, because a computer screen is used to serve as a calibration pattern. At each time step, a moving target is displayed on the screen, where the subject’s eye locations are known thanks to the model-based head tracker (Section 2.4).
On the other hand, the goal of the testing phase is estimating the PoR for a new eye–appearance. As mentioned previously, the PoR is the intersection of the line of sight with an object in the world coordinate system (WCS). The gaze estimation problem can be divided into four parts: (i) head pose estimation in the world coordinate system; (ii) line of sight estimation in the head coordinate system; (iii) line of sight calculation in the world coordinate system; and (iv) intersecting the line of sight with the object (i.e., a display in our case) in the WCS.

2.5.1. Training

In this section, we want to train an appearance-based gaze estimator to calculate a one to one mapping between the pose-free eye appearance (i.e., eye image) and its corresponding gaze vector in the head coordinate system. As an example, Figure 8 demonstrates a pose-free eye appearance which looks directly to the front (i.e., a 0 degree gaze vector).
Let us assume that [ ( I 1 , G 1 ) , ( I 2 , G 2 ) , , ( I N , G N ) ] are the training data, where I i and G i ( i = 1 , 2 , , N ) stand for the eye appearance image and their corresponding gaze vectors, respectively. It is possible to train an interpolation-based gaze estimator using these N training examples. For any new G t e s t , it is possible to estimate the gaze by interpolation. Note that, in appearance-based gaze estimators, the G i ( i = 1 , 2 , , N ) vectors are calculated using a calibration pattern [7,30,34]. In this paper, the same methods explained in [7,30,34] are for interpolation: (i) A K-NN based approach; and (ii) Adaptive Linear Regression (ALR) which is based on manifold learning. The reader is referred to [7,30,34] for further details.

2.5.2. Head Pose Estimation in the WCS

This step is explained in Section 2.4. As a subject-specific model is tracking the head in the depth frames, it is possible to precisely track the location of eyes in the WCS. On the other hand, in geometry, a line can be uniquely defined with a point and a vector. Thus, the eye locations given by the head tracker are can be used to calculate the line of sight passing through them. In general, two goals are accomplished in this part: (i) estimating the pose (i.e., both position and orientation) of the head and eyes; and (ii) head pose stabilization (See Section 2.4).

2.5.3. LoS Calculation in the WCS

The LoS is estimated in the head coordinate system. On the other hand, the head is tracked using the head tracker. Thus, we can calculate the rotation and translation of the head coordinate system in the WCS. Consequently, we can re-express the line of sight in the WCS.

2.5.4. PoR Calculation

Once the LoS is calculated in WCS, we can intersect it with the computer screen, which is a plane in the WCS (the position of the computer screen is known).

3. Experimental Evaluation

In this section, we report our empirical evaluation. We start by describing the datasets used in our experiments, follow this with an explanation of the evaluation protocol, and finish with a report of the results and their discussion.

3.1. Databases

3.1.1. 3D Basel Face Model (BFM)

The 3D Basel Face Model (BFM) is a Morphable model calculated from registered 3D scans of 100 male and 100 female faces. The model geometry consists of 53,490 3D vertices connected by 160,470 triangles. The model is given by the following:
  • The mean shape
  • The 199 principal components (PCs) of shape obtained by applying PCA on 200 subjects facial shape in the database
  • The variance of shape
  • The mesh topology
  • The mean texture
  • The 199 principal components (PCs) of texture obtained by applying PCA on 200 subjects facial texture in the database
  • The texture variance
Figure 9 and Figure 10 demonstrate the mean and the first, second and third principal components (visualized: ± 5 standard deviation) of the shape and texture model respectively.
Any unknown face can be explained as a linear combination of the principal components and the mean shape/texture. In this paper, we only use the shape dataset (i.e., shape principal components together with mean shape) for the construction of a subject’s specific face model (i.e., the head trackers).

3.1.2. Biwi Kinect Head Pose Database

We used the Biwi Kinect Head Pose Database [2,3] to evaluate the effectiveness of our method for the following reasons. Firstly, to the best of our knowledge, it is the only RGB-D database for pose estimation reported in the literature. Secondly, it provides ground truth data for comparison, and we wanted to make our results directly comparable to not only those of Fanelli et al. [2,3], but also to the recent works that have used this database. The dataset contains over 15000 depth frames and RGB image of 20 people, six females and fourteen males, where four people were recorded twice. The head pose ranges through about 75 degrees yaw and 60 degrees pitch. The ground truth for head rotation is also provided by a commercial software.

3.1.3. EYEDIAP Database

We used the EYEDIAP gaze database [39] to evaluate the effectiveness of gaze estimation part of our method for the following reasons. Firstly, to our knowledge, it is the only Kinect based database for gaze estimation in the literature. Secondly, it provides ground truth data for comparison of gaze estimation (but not pose) results. In addition, we wanted to make our results statistically comparable to the work of Funes and Odobez [7,34]. The dataset contains over 4450 depth frames and RGB image of 16 people, among them 14 subjects participated in a screen-based gaze estimation scenario. Each session itself is divided into two other sessions, where the subject was asked to keep the head stationary or moving.

3.2. Evaluation Methodology

3.2.1. Subject Specific Model Construction

We evaluated the proposed algorithm in a setting in which the first RGB frame and the first depth frame were used for learning in an unsupervised context, while the other depth frames were used for testing. Figure 11 demonstrates the registration procedure. In this figure, the blue point cloud is the (down sampled) mean shape of the BASEL data, while the red point cloud is the trimmed depth scan of the first subject in the Biwi database (the subject in Figure 3). Notice that trimming the model is not shown here, but it is considered in calculations. We wanted to register the two shapes with each other and, at the same time, capture the variation of the subject’s face by minimizing Equation (1).

3.2.2. Pose Estimation and Tracking

After the subject’s specific model was constructed from the first (trimmed) depth frame, we dropped the model term from the energy function and continued the registration of the model and depth data. The pose estimation for each frame could be directly calculated from the rotation matrix, R, obtained from registration. Figure 12a shows a sample where the model (red) was registered with the depth data (blue). The model was superimposed on the corresponding RGB frame through the Kinect calibration data (Figure 12b) for a better visualization.
Figure 13 shows the result of pose estimation in terms of yaw, depth, and roll for 10,000 frames from the Biwi head pose database compared to the ground truth.
A summary of the key evaluation results and method features of the proposed algorithm compared to two previous works is shown in Table 1. Three criteria were considered to compare the systems according to Pauly [40]:
  • Accuracy: The 3D head tracker should estimate the head pose with a high precision compared to a ground truth. Note that this ground truth was generated by applying a third-party commercial software on a public dataset, on which we performed our experiments. Thus, the term ground truth is used just to be consistent with the literature and be able to compare with the previous works. In theory, any pose estimator including the commercial software should have some inaccuracy.
  • Robustness: The 3D head tracker should be able to perform well under poor lighting conditions, fast motion of head and partial occlusion.
  • Usability in real scenarios: User-specific training, system calibration and manual intervention need to be kept to a minimum. The tracker should be non-invasive. This will make the tracker be widely accepted by the users. Ideally, the system should be calibration-free.
On the one hand, our system demonstrates better results than [2,3] in terms of average error and standard deviation. Both systems proposed by Fanelli et al. show slightly better performance than our system only in terms of standard deviation of roll. Moreover, our system is generic and the training phase is performed with a single RGB/depth frame. On the other hand, the systems of Fanelli et al. can work on a frame-by-frame basis, while our system can only work in tracking mode (i.e., the subject’s head motion in successive frames should be small). Notice that both systems of Fanelli et al. need a training phase supported by a commercial face tracker, while we propose a new face tracker in this work. As both systems of Fanelli et al. require a training phase based on positive and negative patches cropped from a database of 20 subjects, the generic aspect of their system is an issue. We also compared our system to other non-model-based approaches [8,10,13,41], and the results show the effectiveness of the proposed system. The only comparable system is the model-based work of [11], which shows very good precision too.

3.2.3. Gaze Estimation

Adaptive Linear Regression (ALR)

The results of testing the algorithm on the stationary head session of IDIAP database with 13 subjects are summarized in Table 2 and Table 3 for left eye and right eye, respectively, while Table 4 and Table 5 show the same results for the moving head session.

K-Nearest Neighbor

We used 40 training eye images from IDIAP database and represented the eye appearances as 15D feature vectors. These feature vectors are given to a K-NN regressor. Similar to Funes Mora and Odobez [7], we chose K = 5 . The steps of K-NN regression is as follows:
  • Given a test image, find the K closest set of sample images forming a neighborhood based on K-NN.
  • Find a set of weights: Inverse the distances of the K images from test image and normalize them.
  • Use the same weights to interpolate the parameters to obtain the estimated parameters for the test gazing point.
The results of testing the algorithm on the stationary head session of the IDIAP database with 13 subjects are summarized in Table 6 and Table 7 for the left eye and the right eye respectively, while Table 8 and Table 9 show the same results for the moving head session.
A summary of the key evaluation results and method features of the proposed algorithm compared to two previous works are shown in Table 10, Table 11, Table 12 and Table 13. Three criteria were considered to compare the systems:
  • Accuracy compared to the ground truth data
  • Robustness to occlusions, bad lighting and fast motions
  • Usability in real scenarios, i.e., user-specific training, system calibration and manual intervention need to be kept to a minimum
Our system demonstrates better results than Funes and Odobez in terms of average gaze estimation error. One possible reason for this can be the high precession of our pose estimation system, which performs similar to a commercial state-of-the-art pose estimator, while the pose estimator of Kenneth and Odobez shows slight deviation from our precise pose estimator, which results in an imprecise texture warping on the head model that in turn can affect the gaze estimation process.

4. Conclusions

This work addressed the problem of automatic facial pose and gaze estimation without subject cooperation or manual intervention using low quality depth data provided by the Microsoft Kinect. Previous works on pose estimation using the Kinect are based on supervised learning or require manual intervention. In this work, we proposed a 3D pose estimator based on low quality depth data. The proposed method is generic and fully automatic. The experimental results on the Biwi head pose database confirm the efficiency of our algorithm in handling large head pose variations and partial occlusion. Our results also confirm that model-based approaches outperform other approaches in terms of precision. We also evaluated the performance of our algorithm on the IDIAP database for 3D eye gaze estimation (i.e., point of regard) and we obtained promising results. Although the feature-based gaze estimators are the most accurate ones in the literature, they require high resolution images. As Kinect has a low resolution RGB camera, we designed two appearance-based gaze estimators (i.e., manifold based ALR and K-NN) that do not rely on the local eye features. Instead, our proposed systems depend on the entire eye image content. Our gaze estimators outperformed the other appearance-based gaze estimators from several aspects, thanks to the high precision of the head tracker. Moreover, the user can freely turn their head, while this is not the case for most of the appearance-based methods, where the user should use a chin rest.

Author Contributions

Conceptualization R.S.G., O.A., D.L.; Methodology R.S.G., O.A., D.L.; Software R.S.G.; Validation R.S.G.; Writing—Original Draft Preparation R.S.G.; Writing Review and Editing R.S.G., D.L.; Supervision D.L.; Funding Acquisition D.L.

Funding

The research presented in the paper was funded by grant F506-FSA of the Auto21 Networks of Centers of Excellence Program of Canada.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Murphy-Chutorian, E.; Trivedi, M.M. Head pose estimation in computer vision: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 607–626. [Google Scholar] [CrossRef] [PubMed]
  2. Fanelli, G.; Weise, T.; Gall, J.; Van Gool, L. Real time head pose estimation from consumer depth cameras. In Pattern Recognition; Springer: New York, NY, USA, 2011; pp. 101–110. [Google Scholar]
  3. Fanelli, G.; Dantone, M.; Gall, J.; Fossati, A.; Van Gool, L. Random Forests for Real Time 3D Face Analysis. Int. J. Comput. Vis. 2013, 101, 437–458. [Google Scholar] [CrossRef]
  4. Breitenstein, M.D.; Kuettel, D.; Weise, T.; Van Gool, L.; Pfister, H. Real-time face pose estimation from single range images. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  5. Fanelli, G.; Gall, J.; Van Gool, L. Real time head pose estimation with random regression forests. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 617–624. [Google Scholar]
  6. Seemann, E.; Nickel, K.; Stiefelhagen, R. Head pose estimation using stereo vision for human-robot interaction. In Proceedings of the Sixth IEEE International Conference on Automatic Face and Gesture Recognition, Seoul, Korea, 19 May 2004; pp. 626–631. [Google Scholar]
  7. Funes Mora, K.A.; Odobez, J. Gaze estimation from multimodal Kinect data. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; pp. 25–30. [Google Scholar]
  8. Rekik, A.; Ben-Hamadou, A.; Mahdi, W. 3D Face Pose Tracking using Low Quality Depth Cameras. In Proceedings of the International Conference on Computer Vision Theory and Applications—VISAPP 2013, Barcelona, Spain, 21–24 February 2013; Volume 2. [Google Scholar]
  9. Cai, Q.; Gallup, D.; Zhang, C.; Zhang, Z. 3D Deformable Face Tracking with a Commodity Depth Camera. In Proceedings of the 11th European Conference on Computer Vision Conference on Computer Vision: Part III, ECCV’10, Heraklion, Crete, Greece, 5–11 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 229–242. [Google Scholar]
  10. Baltrusaitis, T.; Robinson, P.; Morency, L.P. 3D Constrained Local Model for rigid and non-rigid facial tracking. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2610–2617. [Google Scholar] [CrossRef]
  11. Yu, Y.; Mora, K.A.F.; Odobez, J.M. Robust and Accurate 3D Head Pose Estimation through 3DMM and Online Head Model Reconstruction. In Proceedings of the 2017 12th IEEE International Conference on Automatic Face Gesture Recognition (FG 2017), Washington, DC, USA, 30 May–3 June 2017; pp. 711–718. [Google Scholar] [CrossRef]
  12. Papazov, C.; Marks, T.K.; Jones, M. Real-time 3D head pose and facial landmark estimation from depth images using triangular surface patch features. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 4722–4730. [Google Scholar] [CrossRef]
  13. Tulyakov, S.; Vieriu, R.L.; Semeniuta, S.; Sebe, N. Robust Real-Time Extreme Head Pose Estimation. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; pp. 2263–2268. [Google Scholar] [CrossRef]
  14. Hansen, D.; Ji, Q. In the Eye of the Beholder: A Survey of Models for Eyes and Gaze. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 478–500. [Google Scholar] [CrossRef] [PubMed]
  15. Villanueva, A.; Cabeza, R.; Porta, S. Eye tracking: Pupil orientation geometrical modeling. Image Vis. Comput. 2006, 24, 663–679. [Google Scholar] [CrossRef]
  16. Wang, J.G.; Sung, E.; Venkateswarlu, R. Estimating the eye gaze from one eye. Comput. Vis. Image Underst. 2005, 98, 83–103. [Google Scholar] [CrossRef]
  17. Villanueva, A.; Cabeza, R.; Porta, S. Gaze Tracking System Model Based on Physical Parameters. IJPRAI Int. J. Pattern Recognit. Artif. Intell. 2007, 21, 855–877. [Google Scholar] [CrossRef]
  18. White, K.P.; Hutchinson, T.E.; Carley, J.M. Spatially dynamic calibration of an eye-tracking system. IEEE Trans. Syst. Man Cybern. 1993, 23, 1162–1168. [Google Scholar] [CrossRef]
  19. Guestrin, E.D.; Eizenman, M. General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Trans. Biomed. Eng. 2006, 53, 1124–1133. [Google Scholar] [CrossRef] [PubMed]
  20. Morimoto, C.H.; Amir, A.; Flickner, M. Detecting eye position and gaze from a single camera and 2 light sources. In Proceedings of the Object Recognition Supported by User Interaction for Service Robots, Quebec City, QC, Canada, 11–15 August 2002; Volume 4, pp. 314–317. [Google Scholar] [CrossRef]
  21. Blignaut, P. Mapping the pupil-glint vector to gaze coordinates in a simple video-based eye tracker. J. Eye Mov. Res. 2013, 7. [Google Scholar] [CrossRef]
  22. Cherif, Z.R.; Nait-Ali, A.; Motsch, J.; Krebs, M. An adaptive calibration of an infrared light device used for gaze tracking. In Proceedings of the 19th IEEE Instrumentation and Measurement Technology Conference, IMTC/2002, Anchorage, AK, USA, 21–23 May 2002; IEEE: Piscataway, NJ, USA, 2002; Volume 2, pp. 1029–1033. [Google Scholar]
  23. Brolly, X.L.; Mulligan, J.B. Implicit calibration of a remote gaze tracker. In Proceedings of the 2004 Conference on Computer Vision and Pattern Recognition Workshop, CVPRW’04, Washington, DC, USA, 27 June–2 July 2004; IEEE: Piscataway, NJ, USA, 2004; p. 134. [Google Scholar]
  24. Cerrolaza, J.J.; Villanueva, A.; Cabeza, R. Taxonomic study of polynomial regressions applied to the calibration of video-oculographic systems. In Proceedings of the 2008 Symposium on Eye Tracking Research & Applications, Savannah, GA, USA, 26–28 March 2008; ACM: New York, NY, USA, 2008; pp. 259–266. [Google Scholar]
  25. Zhu, Z.; Ji, Q. Eye and gaze tracking for interactive graphic display. Mach. Vis. Appl. 2004, 15, 139–148. [Google Scholar] [CrossRef]
  26. Chi, J.; Zhang, C.; Yan, Y.; Liu, Y.; Zhang, H. Eye gaze calculation based on nonlinear polynomial and generalized regression neural network. In Proceedings of the Fifth International Conference on Natural Computation, ICNC’09, Tianjin, China, 14–16 August 2009; IEEE: Piscataway, NJ, USA, 2009; Volume 3, pp. 617–623. [Google Scholar]
  27. Ma, C.; Choi, K.A.; Choi, B.D.; Ko, S.J. Robust remote gaze estimation method based on multiple geometric transforms. Opt. Eng. 2015, 54, 083103. [Google Scholar] [CrossRef]
  28. Pomerleau, D.; Baluja, S. Non-intrusive gaze tracking using artificial neural networks. In Proceedings of the AAAI Fall Symposium on Machine Learning in Computer Vision, Raleigh, NC, USA, 22–24 October 1993; pp. 153–156. [Google Scholar]
  29. Tan, K.H.; Kriegman, D.J.; Ahuja, N. Appearance-based eye gaze estimation. In Proceedings of the Sixth IEEE Workshop on Applications of Computer Vision (WACV 2002), Orlando, FL, USA, 4 December 2002; IEEE: Piscataway, NJ, USA, 2002; pp. 191–195. [Google Scholar]
  30. Lu, F.; Sugano, Y.; Okabe, T.; Sato, Y. Inferring human gaze from appearance via adaptive linear regression. In Proceedings of the 2011 IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 153–160. [Google Scholar]
  31. Williams, O.; Blake, A.; Cipolla, R. Sparse and Semi-supervised Visual Mapping with the S^ 3GP. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006; IEEE: Piscataway, NJ, USA, 2006; Volume 1, pp. 230–237. [Google Scholar]
  32. Noris, B.; Keller, J.B.; Billard, A. A wearable gaze tracking system for children in unconstrained environments. Comput. Vis. Image Underst. 2011, 115, 476–486. [Google Scholar] [CrossRef]
  33. Choi, D.H.; Jang, I.H.; Kim, M.H.; Kim, N.C. Color image enhancement based on single-scale retinex with a JND-based nonlinear filter. In Proceedings of the IEEE International Symposium on Circuits and Systems, ISCAS 2007, New Orleans, LA, USA, 27–30 May 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 3948–3951. [Google Scholar]
  34. Funes-Mora, K.A.; Odobez, J.M. Gaze estimation in the 3d space using rgb-d sensors. Int. J. Comput. Vis. 2016, 118, 194–216. [Google Scholar] [CrossRef]
  35. Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, Kauai, HI, USA, 8–14 December 2001; Volume 1, pp. 1–511. [Google Scholar]
  36. Bouaziz, S.; Pauly, M. Dynamic 2d/3d Registration for the Kinect; ACM: New York, NY, USA, 2013; p. 21. [Google Scholar]
  37. LaValle, S.M. Planning Algorithms; Cambridge University Press: Cambridge, UK, 2006; Available online: http://planning.cs.uiuc.edu/node103.html (accessed on 5 December 2018).
  38. Bouaziz, S.; Tagliasacchi, A.; Pauly, M. Sparse iterative closest point. In Proceedings of the Eleventh Eurographics/ACMSIGGRAPH Symposium on Geometry Processing, Genova, Italy, 3–5 July 2013; Volume 32, pp. 113–123. [Google Scholar]
  39. Funes Mora, K.A.; Monay, F.; Odobez, J.M. EYEDIAP Database: Data Description and Gaze Tracking Evaluation Benchmarks; Idiap-RR Idiap-RR-08-2014; Idiap: Martigny, Switzerland, 2014. [Google Scholar]
  40. Pauly, M. Realtime Performance-Based Facial Avatars for Immersive Gameplay. In Proceedings of the Motion on Games, Dublin, Ireland, 6–8 November 2013; pp. 23–28. [Google Scholar] [CrossRef]
  41. Martin, M.; van de Camp, F.; Stiefelhagen, R. Real Time Head Model Creation and Head Pose Estimation on Consumer Depth Cameras. In Proceedings of the 2014 2nd International Conference on 3D Vision, Tokyo, Japan, 8–11 December 2014; Volume 1, pp. 641–648. [Google Scholar] [CrossRef]
Figure 1. A simplified Schematic of Human Eye with LoS, GoS and PoR.
Figure 1. A simplified Schematic of Human Eye with LoS, GoS and PoR.
Sensors 18 04280 g001
Figure 2. Depth data obtained from the first frame and visualized from profile. They consist of both the facial part and data from other body parts.
Figure 2. Depth data obtained from the first frame and visualized from profile. They consist of both the facial part and data from other body parts.
Sensors 18 04280 g002
Figure 3. Face and facial features detection from the first RGB frame of a subject.
Figure 3. Face and facial features detection from the first RGB frame of a subject.
Sensors 18 04280 g003
Figure 4. Trimming the first depth frame: (left) top view; and (right) profile view.
Figure 4. Trimming the first depth frame: (left) top view; and (right) profile view.
Sensors 18 04280 g004
Figure 5. Trimming the 3D Morphable model mean shape: (left) before trimming; and (right) after trimming.
Figure 5. Trimming the 3D Morphable model mean shape: (left) before trimming; and (right) after trimming.
Sensors 18 04280 g005
Figure 6. Positioning the model on the face. The blue object is the trimmed model, while the green object is the subject scanned by the Kinect.
Figure 6. Positioning the model on the face. The blue object is the trimmed model, while the green object is the subject scanned by the Kinect.
Sensors 18 04280 g006
Figure 7. Applying the face texture from: (a) an RGB frame to the subject’s specific model. (b) The model visualized from side view; and (c) the same model visualized from down view. The left and eye appearances (d) are some examples of pose-free eye appearance which can be used to train an interpolation function.
Figure 7. Applying the face texture from: (a) an RGB frame to the subject’s specific model. (b) The model visualized from side view; and (c) the same model visualized from down view. The left and eye appearances (d) are some examples of pose-free eye appearance which can be used to train an interpolation function.
Sensors 18 04280 g007
Figure 8. An exemplar pose-free eye appearance which looks forward.
Figure 8. An exemplar pose-free eye appearance which looks forward.
Sensors 18 04280 g008
Figure 9. (a) The mean; (b) the first; (c) the second; and (d) the third principal components (visualized: ± 5 standard deviation) of the shape model. The images are taken from the database website
Figure 9. (a) The mean; (b) the first; (c) the second; and (d) the third principal components (visualized: ± 5 standard deviation) of the shape model. The images are taken from the database website
Sensors 18 04280 g009
Figure 10. (a) The mean; (b) the first; (c) the second; and (d) the third principal components (visualized: ± 5 standard deviation) of the texture model. The images are taken from the database website.
Figure 10. (a) The mean; (b) the first; (c) the second; and (d) the third principal components (visualized: ± 5 standard deviation) of the texture model. The images are taken from the database website.
Sensors 18 04280 g010
Figure 11. Registration procedure to capture subject’s variations, d, together with transformation matrices R and t: (a) before model initialization; (b) after model initialization and before capturing subject’s face variations; and (c) after the initialization is accomplished (ready for tracking).
Figure 11. Registration procedure to capture subject’s variations, d, together with transformation matrices R and t: (a) before model initialization; (b) after model initialization and before capturing subject’s face variations; and (c) after the initialization is accomplished (ready for tracking).
Sensors 18 04280 g011
Figure 12. An instant of the tracking mode: the model (red) tracks the depth data (blue) and it is back-projected to the RGB frame
Figure 12. An instant of the tracking mode: the model (red) tracks the depth data (blue) and it is back-projected to the RGB frame
Sensors 18 04280 g012
Figure 13. The experimental results of the proposed facial pose estimator on Biwi head pose database compared to the ground truth.
Figure 13. The experimental results of the proposed facial pose estimator on Biwi head pose database compared to the ground truth.
Sensors 18 04280 g013
Table 1. A summary of the key evaluation results and method features of the proposed algorithm, and the two previous works based on supervised learning Fanelli et al. [2,3]. Notice that we do not compare the results with those of Funes and Odobez  [7] for face pose estimation due to lack of details in yaw, roll and pitch. Legend: Sensors 18 04280 i001, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Table 1. A summary of the key evaluation results and method features of the proposed algorithm, and the two previous works based on supervised learning Fanelli et al. [2,3]. Notice that we do not compare the results with those of Funes and Odobez  [7] for face pose estimation due to lack of details in yaw, roll and pitch. Legend: Sensors 18 04280 i001, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Pose Estimation ErrorSpecifications
PitchYawRollAccuracyRobustnessUsability
Our Proposed Method0.1 ± 6.7 0.25 ± 8.7 0.26 ± 9.3 Sensors 18 04280 i001 Sensors 18 04280 i001 Sensors 18 04280 i002
1st report [2]8.5 ± 9.9 8.9 ± 13.0 7.9 ± 8.3 Sensors 18 04280 i002 Sensors 18 04280 i001 Sensors 18 04280 i002
2nd report [3]5.2 ± 7.7 6.6 ± 12.6 6.0 ± 7.1 Sensors 18 04280 i002 Sensors 18 04280 i001 Sensors 18 04280 i002
[13]N/A3.18 ± 5.3 N/A Sensors 18 04280 i001 Sensors 18 04280 i001 Sensors 18 04280 i002
[10]5.10 6.29 11.29 Sensors 18 04280 i002 Sensors 18 04280 i001 Sensors 18 04280 i002
[8]4.32 ± 2.65 5.13 ± 3.33 5.24 ± 3.33 Sensors 18 04280 i002 Sensors 18 04280 i001 Sensors 18 04280 i002
[12]2.5 ± 7.4 3 ± 9.6 3.8 ± 16 Sensors 18 04280 i002 Sensors 18 04280 i001 Sensors 18 04280 i002
[41]2.54 2.57 3.62 Sensors 18 04280 i001 Sensors 18 04280 i001 Sensors 18 04280 i002
[11]1.7 2.5 2.3 Sensors 18 04280 i001 Sensors 18 04280 i001 Sensors 18 04280 i002
Table 2. Left eye, Adaptive Linear Regression, stationary head.
Table 2. Left eye, Adaptive Linear Regression, stationary head.
SubjectMeanMedianMinMaxStd. Dev.
1.00006.56685.52130.065628.57744.5051
2.00006.52045.21140.100930.33824.9035
3.00005.23084.18090.034332.32514.1530
4.00007.66066.23220.034333.41975.6697
5.00008.83907.43930.169036.20755.8689
6.00008.38176.94230.083932.30265.7648
8.00007.66425.69130.059334.59036.0913
9.00008.56177.72700.034328.08754.8624
10.00007.04226.00810.059333.94984.8673
11.00006.09044.94690.028030.91624.4417
14.00005.97175.42380.088526.01223.5437
15.000014.86584.95480.1009179.590738.0497
16.00004.78653.75440.074030.73444.0378
Table 3. Right eye, Adaptive Linear Regression, stationary head.
Table 3. Right eye, Adaptive Linear Regression, stationary head.
SubjectMeanMedianMinMaxStd. Dev.
1.00006.56265.11440.071332.55615.2122
2.00006.43495.06250.090731.71405.0184
3.00006.23394.22590.104733.98365.4509
4.00008.44046.69090.090732.58266.2366
5.00008.92057.41860.283935.77946.0221
6.00007.79855.99690.044235.70145.9868
8.00006.76695.49830.062637.19414.9395
9.00008.97768.13800.113630.36155.2164
10.00006.00154.85190.039632.23594.6771
11.00005.58463.93050.044230.45355.0931
14.00006.19535.69080.092826.01493.4915
15.00005.83074.87410.034328.70524.2100
16.00005.88754.45540.039632.42134.8225
Table 4. Left eye, Adaptive Linear Regression, moving head.
Table 4. Left eye, Adaptive Linear Regression, moving head.
SubjectMeanMedianMinMaxStd. Dev.
1.000014.920812.37440.118763.557710.5413
2.000010.04848.61270.068541.71076.5863
3.000010.62908.52830.019852.67258.1252
4.000013.612711.26500.173649.99729.6486
5.00008.65517.21480.056050.26606.0730
6.00009.10248.09210.034333.34125.6728
8.000013.642710.77720.126753.14649.8346
9.00009.64167.77820.076646.38177.1335
10.00009.22507.31120.059343.36087.0226
11.00007.82735.89340.106552.24616.7665
14.00005.17534.35530.102825.98973.7616
15.00007.70786.41440.118729.62655.2408
16.00006.89884.98010.179138.44385.7533
Table 5. Right eye, Adaptive Linear Regression, moving head.
Table 5. Right eye, Adaptive Linear Regression, moving head.
SubjectMeanMedianMinMaxStd. Dev.
1.000014.315411.93870.110158.825810.0281
2.00009.58038.13920.149446.82546.4423
3.00009.93087.88590.134248.77367.5894
4.000012.42699.97700.065650.12409.2286
5.00009.86388.47710.264748.74626.3292
6.00006.66625.53800.056033.52274.7564
8.000014.378410.44780.028053.726511.5794
9.00009.49427.59570.062644.82856.9502
10.00009.07226.60250.056046.52597.5245
11.00007.53775.84380.068543.72696.1082
14.00005.13214.12590.102824.89753.8079
15.00007.50456.57320.048532.91214.5594
16.00007.44445.1954036.11096.5095
Table 6. Left eye, KNN, stationary.
Table 6. Left eye, KNN, stationary.
SubjectMeanMedianMinMaxStd. Dev.
1.00006.26375.66060.048529.39133.8814
2.00006.74245.66520.034329.75004.5975
3.00005.44874.48400.065632.35944.0323
4.00008.06477.17610.034329.92594.9577
5.00008.71747.68010.281230.91905.1745
6.00007.97546.76320.092833.20635.4553
8.00006.87066.26720.056034.95994.1641
9.00008.32987.80830.184523.83394.1954
10.00006.74405.79910.108431.70584.4667
11.00005.96574.89000.106529.51534.2912
14.00005.72155.09240.129724.62033.4290
15.000032.12865.11320.0280179.433962.0019
16.00005.79705.07940.052330.85343.8672
Table 7. Right eye, KNN, stationary.
Table 7. Right eye, KNN, stationary.
SubjectMeanMedianMinMaxStd. Dev.
1.00006.10095.31650.081627.70143.9828
2.00006.60695.79640.044231.13154.2246
3.00004.94094.00030.039631.69543.8844
4.00007.86556.97780.019830.85934.7412
5.00008.00417.39850.203730.74984.5424
6.00007.37536.12970.079134.92565.0427
8.00006.66056.19250.157034.72943.9806
9.00008.44867.71630.071326.22164.6055
10.00005.95784.79600.253331.63154.4807
11.00005.51724.34710.056030.77134.1658
14.00006.02445.33240.096925.05513.7456
15.00005.02044.16620.068528.71283.8207
16.00005.83535.12760.028028.75203.7264
Table 8. Left eye, KNN, moving.
Table 8. Left eye, KNN, moving.
SubjectMeanMedianMinMaxStd. Dev.
1.000012.697210.63830.079162.49219.0053
2.00009.26817.93030.094946.04306.3844
3.000010.02208.47900.203750.29887.1276
4.000012.438210.74940.269847.88938.1049
5.00008.61437.17740.1824124.35366.0929
6.00008.67927.65990.193835.76265.0370
8.000010.26918.27490.164340.47637.3319
9.00009.48057.83640.056046.41326.6735
10.00008.78267.13640.113656.22276.7277
11.00007.08455.93440.065657.53335.1368
14.00005.62544.70950.028026.29633.8907
15.00007.83716.65400.149431.01814.9755
16.00006.57435.62390.100934.76074.3908
Table 9. Right eye, KNN, moving.
Table 9. Right eye, KNN, moving.
SubjectMeanMedianMinMaxStd. Dev.
1.000012.323710.42690.188745.33877.9340
2.00008.97637.77120.205638.00115.8696
3.00009.05027.92960.079140.73726.0890
4.000011.22509.14580.1413123.28898.1434
5.00009.32958.72520.214945.18315.4824
6.00007.01466.43000.134233.49724.0739
8.000012.15598.86430.071377.731512.3338
9.00009.42437.59410.219445.44376.9684
10.00008.70826.85010.111948.62727.0996
11.00007.24996.16850.048548.60695.3183
14.00005.33954.48470.155824.17233.5878
15.00007.56716.81550.132733.67824.4902
16.00006.85815.00580.132740.44605.8367
Table 10. A summary of the key evaluation results and method features of the proposed algorithm, and the previous work when Adaptive Linear Regression (ALR) is used and the subjects keep the head stationary: Sensors 18 04280 i001, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Table 10. A summary of the key evaluation results and method features of the proposed algorithm, and the previous work when Adaptive Linear Regression (ALR) is used and the subjects keep the head stationary: Sensors 18 04280 i001, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Gaze Estimation ErrorSpecifications
Left EyeRight EyeAccuracyRobustnessUsability
Our Proposed Method7.55 6.89 Sensors 18 04280 i002 Sensors 18 04280 i001 Sensors 18 04280 i001
Funes and Odobez Method9.73 10.5 Sensors 18 04280 i002 Sensors 18 04280 i002 Sensors 18 04280 i002
Table 11. A summary of the key evaluation results and method features of the proposed algorithm, and the previous work when K-NN is used and the subject keeps the head stationary: Sensors 18 04280 i001, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Table 11. A summary of the key evaluation results and method features of the proposed algorithm, and the previous work when K-NN is used and the subject keeps the head stationary: Sensors 18 04280 i001, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Gaze Estimation ErrorSpecifications
Left EyeRight EyeAccuracyRobustnessUsability
Our Proposed Method8.83 6.49 Sensors 18 04280 i002 Sensors 18 04280 i001 Sensors 18 04280 i001
Funes and Odobez Method10.23 9.56 Sensors 18 04280 i002 Sensors 18 04280 i002 Sensors 18 04280 i002
Table 12. A summary of the key evaluation results and method features of the proposed algorithm, and the previous work when Adaptive Linear Regression (ALR) is used and the subjects have free head motion: ⚫, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Table 12. A summary of the key evaluation results and method features of the proposed algorithm, and the previous work when Adaptive Linear Regression (ALR) is used and the subjects have free head motion: ⚫, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Gaze Estimation ErrorSpecifications
Left EyeRight EyeAccuracyRobustnessUsability
Our Proposed Method9.78 9.49 Sensors 18 04280 i002 Sensors 18 04280 i001 Sensors 18 04280 i001
Funes and Odobez Method15.57 14.2 Sensors 18 04280 i002 Sensors 18 04280 i002 Sensors 18 04280 i002
Table 13. A summary of the key evaluation results and method features of the proposed algorithm, and the previous work when K-NN is used and the subjects have free head motion: Sensors 18 04280 i001, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Table 13. A summary of the key evaluation results and method features of the proposed algorithm, and the previous work when K-NN is used and the subjects have free head motion: Sensors 18 04280 i001, very good; Sensors 18 04280 i002, good; Sensors 18 04280 i003, weak.
Gaze Estimation ErrorSpecifications
Left EyeRight EyeAccuracyRobustnessUsability
Our Proposed Method9.03 8.86 Sensors 18 04280 i002 Sensors 18 04280 i001 Sensors 18 04280 i001
Funes and Odobez Method17.97 14.63 Sensors 18 04280 i002 Sensors 18 04280 i002 Sensors 18 04280 i002

Share and Cite

MDPI and ACS Style

Shoja Ghiass, R.; Arandjelovć, O.; Laurendeau, D. Highly Accurate and Fully Automatic 3D Head Pose Estimation and Eye Gaze Estimation Using RGB-D Sensors and 3D Morphable Models. Sensors 2018, 18, 4280. https://doi.org/10.3390/s18124280

AMA Style

Shoja Ghiass R, Arandjelovć O, Laurendeau D. Highly Accurate and Fully Automatic 3D Head Pose Estimation and Eye Gaze Estimation Using RGB-D Sensors and 3D Morphable Models. Sensors. 2018; 18(12):4280. https://doi.org/10.3390/s18124280

Chicago/Turabian Style

Shoja Ghiass, Reza, Ognjen Arandjelovć, and Denis Laurendeau. 2018. "Highly Accurate and Fully Automatic 3D Head Pose Estimation and Eye Gaze Estimation Using RGB-D Sensors and 3D Morphable Models" Sensors 18, no. 12: 4280. https://doi.org/10.3390/s18124280

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop