A New Post-Processing Proposal for Improving Biometric Gait Recognition Using Wearable Devices
Abstract
:1. Introduction
- Fusion of features that can be implemented at distinct levels and used in biometrics to improve accuracy [24]. Here, we use score-level fusion. Two main approaches can be found at this level: fusing scores (outputs) of different classifiers over the same input (e.g., [25]) and fusing the scores of different characteristics, multibiometric system (e.g., [26,27]). We propose a different and simpler approach here, since the scores of a single classifier and characteristic (gait) are combined (Section 4).
- Exploiting context to enhance system accuracy [28]. This is habitual in problems where the context is relevant, e.g., those related with video (e.g., [29]), prosody recognition (e.g., [30]), or speech. Since, in gait, a step is related to the ones that go before and after, our proposal aims to exploit this dependence (Section 4).
2. Related Works
3. State-of-the-Art (Reference) System
- Biometric sample, or sample for short: the analog or digital representation (Figure 3) of biometric characteristics [51] (Figure 1), gait in our case. This is specified in more detail in Section 5.2.
3.1. Acquisition
3.2. Preprocessing
- Sample cleaning, which consists of eliminating the noisy parts of the signal and detecting and correcting acquisition errors. Not much work has been carried out into acquisition problems with real devices, so this has led us to propose our own alternative, which can be seen in Appendix B.
- Period Normalization. With real devices, it is not possible to ensure a fixed sampling rate. This can be seen in Figure 5, where the distribution of the time between two consecutive datum of our database is shown. To fix this, the sample must be resampled [13,21]. To perform this operation, the following must be set: (i) the interpolation method and (ii) the sampling rate. For the first, and following the literature [13], linear interpolation has been used. For the second, after analyzing the frequency components of the data, we saw that components bigger than 6 Hz were negligible; so, following the Nyquist–Shannon sampling theorem, a sampling rate of 12 Hz (a period of 83.3 ms) was fixed. This value is in accordance with that shown in [52], where it is demonstrated that the arm moves at a maximum of 8.6 Hz, making the movement as fast as possible. As our data are collected from walking, the sampling rate selected seems reasonable.
- Amplitude Normalization. The goal is to change the value of the data to a common scale [13]. The need to perform this operation is machine learning algorithm dependent, so the default option for each classifier in the software used for the experiments (RStudio) is used.
3.3. Feature Extraction
- The cycles of the sample are grouped into segments called windows, with 20% of overlap [45] (Figure 3). Therefore, a window, , is composed of m consecutive gait cycles. A cycle is a piece of signal that is too short to be representative of the user’s gait, so these are grouped into windows, which, from now on, will be the basic unit of information used to model and recognize the user.
- From each window, , a feature vector, , is extracted as shown in the next two sections.
3.3.1. Time Domain Feature Extraction
3.3.2. Frequency Domain Feature Extraction
3.4. Classification
4. Window Score Fusion Post-Processing Proposal
5. Experimental Methodology
5.1. Experiment Design
- With regard to the window size in the pre-processing stage, from the values in Table 1, the following number of cycles in the window were selected to be tested: . This set includes both small and high values, being representative of those used in the state of the art.
- With regard to the features extracted in the feature extraction stage, as shown in Section 3.3, typical features are extracted in both the time and frequency domains. In addition, this allows the performance of both approaches to be compared.
- With regard to the classifier in the classification stage, two criteria were fixed:
- Variety in the tested algorithms.
- Most used in the reference works.
The most used were shown in Section 3.4: SVM, MLP-ANN, K-NN, and Random Forest. These also fulfill the first criterion, since their theoretical bases are completely different. A deep study of each classifier is beyond the scope of this work, so a brief description of each one is included, focusing on the main differences between them:- –
- K-NN. Unlike the rest of the classifiers, this does not need to be trained to build a model. The training or enrollment sample(s) (see Section 5.3) is (are) used directly to create the user template, . More specifically, the user template is made up of the feature vectors extracted from the enrollment sample(s): , where N is the number of windows of the enrollment sample(s). The classifier output is based on distance; to be precise, given a test trial feature vector (see Section 5.3), its score is calculated as shown in Equation (2).Since the K-NN output is based on distance, the interpretation is as follows: the lower its value, the higher the degree of belonging of the biometric input to the user.
- –
- SVM [60]. This classifier is based on separating two classes by means of an optimal hyperplane . The parameters of the hyperplane are fixed using the so called “support vectors” (Figure 7a). To avoid overfitting, a soft margin solution is used in the training phase (calculation of hyperplane parameters), allowing “mistakes” in the training samples (Figure 7a); this is controlled by the regularization parameter C: a small C allows a large margin of mistakes, while a large C makes constraints hard to ignore. With the hyperplane set, the classification is performed as shown in Equation (3).The problem is that real-world data are rarely linearly separable. The solution is to increase the dimensionality of the feature space, aiming to map the input space into a linear separable one, where the linear classifier will be applicable. This is performed by means of the “kernel trick”; i.e., a kernel function (e.g., lineal: , radial: ) is used, allowing the mapping to be performed without increasing the complexity of the training algorithm.As can be seen in Equation (3), the sign of the output is used to classify the input. However, here we need a score, i.e., a level of belonging to each class. The Platt scaling [61] is used to accomplish this. Therefore, the score here is a probability, . Therefore, with a different interpretation regarding K-NN, the higher its value, the higher the degree of belonging of the biometric input to the user.
- –
- MLP [62]. This is a net composed of a set of neurons or units organized in layers (Figure 7b). The architecture of the net is defined by the number of layers and neurons in each layer. Each neuron in a layer is connected (its outputs are the inputs) with all the neurons of the following layer, except the last one, whose neurons will be the output(s) of the net. The first layer is the input of the net, which will be the feature vector. The operation performed for each neuron is that shown in Equation (4), where is the output of the neuron h for the input p, is the weight (real number) that connects the neuron h with the neuron j of the previous layer, is the output of this neuron j for the input p, and is the bias or offset of the neuron h. is a function that must be derivable; typical functions are the sigmoid or the hyperbolic tangent.During the learning or training stage, the weights of all the neurons are set using the backpropagation algorithm, so that the value of an error function, E, will be minimized. The most common error function is the squared error, , where is the desired output for the output neuron of the net o for the input p and is the output of the neuron.In our problem, the net has a single neuron in the output, being trained to obtain 0 (the desired output) for training examples of the impostor class and 1 for training examples of the subject (authentic class), using the sigmoid as the activation function, . Therefore, for the MLP, , with j being each neuron of the last hidden layer ( in Figure 7b). Although the output is not really a probability, due to the values of the desired outputs, it can be considered as such, so its interpretation is the same as that seen with SVM.
- –
- Random Forest [63]. This is an ensemble of relatively uncorrelated decision tree classifiers. A decision tree is a supervised classifier that has a flowchart-like tree structure (Figure 7c); each decision node represents a decision rule, finishing in the leaf nodes with the final decisions. This tree is constructed following the algorithm below:
- Using Attribute Selection Measures (ASM), select the best feature (attribute) to split the dataset.
- Create the decision node with the corresponding decision rule. If the node is the first, it is called the root node.
- Using the decision rule, divide the corpus into subsets.
- Repeat the previous steps recursively for each subset until the nodes cannot be further classified due to all of the subset belonging to the same feature value, due to there being no more features, or due to there being no more data.
Based on this classifier, Random Forest works as follows:- *
- Training stage:
- Split the training set randomly into subsets with replacement.
- Train a decision tree with each subset.
- *
- Prediction or test stage:
- Each tree predicts a class.
- Probabilities are calculated from these classes using the predictions.
Therefore, for this classifier, .
These classifiers were those selected for the tests. In addition to the above, as will be seen in the results, the performance of the selected classifiers is very different, which confirms the variety of the selections.For their configuration, we tried to use the default options of the software used (RStudio) as much as we could; the reason is to avoid possible bias in the results when optimizing the classifier, since our goal here is not to obtain the best results, but to test our proposal in the most objective way. Under this criterion, only the following particular configurations were posed:- –
- SVM: radial kernel. From previous experiments, this kernel showed the best performance. R library used: e1071.
- –
- MLP-ANN: JE_Weights initialization function was selected. Others were tested, but the system showed inconsistencies. R library used: RSNNS.
- –
- K-NN: was selected. As with the SVM, from previous experiments, this value showed the best performance, and it is the simplest configuration. R library used: FNN.
- –
- Random Forest: no particular configuration was used in this case. R library used: randomForest.
- The performance of the reference system (Section 3) was calculated in all of the proposed system configurations.
- The performance, when our proposal was used (Section 4), was also calculated in all of the proposed system configurations, and for the different values of n.
- Both results, under the same system configurations, were compared.
- For objective results, the experimental conditions were the same in all of the experiments performed.
5.2. Corpus Data Acquisition
5.3. Experimental Sets
- Short period authentication: the enrollment and testing samples belong to the same session.
- Long period authentication: the enrolment and testing samples belong to different sessions. Testing under this condition is critical as user behavior is different from day to day.
- Enrollment data (genuine class training set). “Enrollment” is, in biometrics, the step where a subject (claimant) C supplies the biometric data to build their biometric model or template, . In pattern recognition terminology, they are called training data. The samples used to build this model or template are called biometric enrolment data record (enrolment data in short from now on). The first sample captured is used for enrolment data, i.e., the first sample of the first session of each subject, as is usual in biometrics.
- Cohort set (impostor class training set), used to train the classifiers with examples of impostors. This set must be completely different from the impostor class test set in order to obtain objective results. Thus, we randomly split the individuals in the database different from the claimant into two different sets; one for training (cohort set), and the other for testing, as shown below. One sample is randomly selected from each individual in the cohort set. For objective comparisons, the cohort set so formed is always the same throughout all the experiments.
- First, for each subject in the database, we selected the test samples:
- Genuine test samples (for biometric mated comparison trials [51]). For these tests, we used the two samples of the second session of the claimant.
- Impostor test samples (for biometric non-mated comparison trials). For these tests, we used random forgeries, i.e., a set of individuals in the database different from the claimant playing the role of impostors (system attackers); this is common in most biometric characteristics for technology evaluation, including gait, e.g., in [42,47], to cite two recent ones. For impostors, as mentioned, we used the subjects of the database different from the claimant not used in the cohort set. From each of these individuals, one of their samples was randomly selected to form this set, a set that is the same throughout all the experiments in order to achieve objective comparisons.
- For both genuine and impostor tests, the corresponding mated and non-mated trials for each subject C are accomplished from each test sample as follows:
- (a)
- The test sample is windowed, i.e., their cycles are grouped as shown in Section 3.3.
- (b)
- From each window, , its corresponding feature vector, , is extracted.
- (c)
- The corresponding score (classifier output), , is calculated. This output is a comparison score [51].
Therefore, for each test sample, we have as many comparison scores or test trials as windows into which it is divided. - With these scores, two sets are created for each claimant C:
- One Test Set (TS) with genuine comparison scores, , achieved from the genuine test samples;
- Another test set with impostor comparison scores, , achieved from the impostors test samples.
As reference, Table 2 shows the total number of tests performed for each window size, joining the corresponding test sets of all claimants. - The system performance is calculated using these two sets, as shown in the next Section (Section 5.4).
- One test set with genuine comparison scores, but now, these scores will be . The set is achieved from the scores in . We call this set .
- Another test set, , with fused scores, but now achieved from .
5.4. Performance Measures
6. Results
- The title shows the feature extraction domain, the sensor coordinate (or fusing all by means of the module), and the classifier.
- The Y-axis shows the performance (EER in %). This axis has the same scale for each classifier to better compare results.
- The X-axis shows the results for each window size, measured by number of cycles.
- For each window size, four bars are shown. The first (brown) shows the result of the reference system (WSFPP is not used). The rest show the system performance when WSFPP is used, for (second bar, blue), (third bar, orange), and (fourth bar, purple).For each of these three last bars, the percentage of improvement or worsening with regard to the reference system (first bar) has been added; this calculation has been performed as shown in Equation (5), where is the performance of the reference system and is the performance when WSFPP is used.
7. Discussion
- With regard to the classifier. Although the use of WSFPP has improved the results with all, this improvement is higher the better the performance of the classifier. The classifier with the best performance with the reference system is SVM, achieving improvements with WSFPP up to 90% in a lot of cases, even reaching 100%, which allowed 0% of EER to be achieved, a result not shown in any previous work. The second best classifier is Random Forest, which also achieves important improvements (higher than 90% in some cases) when WSFPP is used. The other two classifiers show a worse performance, and although the improvements are lower, these have reached 36% with 1-NN and 57% with MLP.
- With regard to the feature extraction domain. The state of the art shows mainly feature extraction in the time domain (Table 1). However, the results show that the features in the frequency domain are an interesting alternative, since similar, and sometimes even better, results have been achieved with these features. Focusing on the case when WSFPP is used, the frequency domain shows higher improvements in general, which has allowed 0% of EER to be reached; except for 1-NN, the best results were achieved in the frequency domain: 0.2% with Random Forest, 9% with MLP and 0% with SVM.
- With regard to the window size. There is no clear tendency. Both with the reference system and with WSFPP, the performance is dependent on the rest of the system parameters (sensor coordinates, feature extraction domain, and classifier). An interesting result is that, although not always, very good performances have been achieved with a size of two cycles, which is very small. Even more, with SVM and frequency domain features, 0% error has been achieved with this size, and X coordinates; this implies that, with a signal of only about 8 s, it has been possible to recognize a person by means of their way of walking using WSFPP.
- With regard to the sensor coordinate. The improvements with WSFPP are similar in all of the sensor coordinates, including the module: once the rest of the parameters of system have been fixed (each row in Figure 8 and Figure 9), the figures of the improvement are, in general, similar for the same values of n. This implies that the performance of WSFPP is independent of this parameter. Focusing on an analysis of the performance, as with the window size, it is dependent on the rest of the parameters of the system. However, if one must be selected, the best alternative is the module; the performance with this is, in the worst case, similar or slightly worst than the best with the other options (X, Y, or Z sensor coordinates), which are almost always better.
- With regard to the value of n in the WSFPP proposal. The first important aspect to note is that, with all of the values, the system performance has improved. This improvement is higher the higher the number of fused scores, n, is. Nevertheless, when the reference system has a good performance, very good results have been achieved with low values of n, e.g., with module, a window size of 2 cycles, frequency domain features and Random Forest (EER = 0.2% for ), or with the X coordinate, a window size of 2 cycles, frequency domain features, and SVM (EER = 0.05% for ).
8. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Appendix A. Gait Cycle Calculation
Appendix B. Signal Cleaning
- Incorrect time stamps. The red line in Figure A1c shows the jumps in time. This happens in isolated data that are moved in time and was observed in a single sample of two users. We have no explanation, leaving their study for future acquisitions.
- Autocorrelation, R. The autocorrelation of the signal or a portion of it is defined as (Equation (A1)). This measure is between 0 and 1, , so that a value near 0 means that the signal does not follow a pattern and a value near 1 means the opposite ( implies a perfect periodic signal).This coefficient allows chaotic zones to be detected, but not the connection loss problem, since when the signal is plain, the values of R are also high. To detect these zones, the measures in the following items were used.
- Energy, E. Defined as , where X, Y and Z are the data of each accelerometer coordinate. This measure has also been used in other biometric characteristics to clean the signal and to isolate noise (low values) from the signal (high values).
- Zero-crossing, Zc. In a periodic signal, this is the point where the signal crosses the X axis. The number of these points has been used in other biometric characteristics (e.g., voice) under the same interpretation as E.
- Time between consecutive datum, T.Figure 5 shows a histogram of the time between two consecutive datum. A high value of T means problems in the acquisition.
- , with , where is the autocorrelation of the window j and N is the number of windows of the sample. In the following items, a similar notation is used.
- ms
References
- Hill, C. Wearables—The future of biometric technology? Biom. Technol. Today 2015, 2015, 5–9. [Google Scholar] [CrossRef]
- Vhaduri, S.; Poellabauer, C. Multi-Modal Biometric-Based Implicit Authentication of Wearable Device Users. IEEE Trans. Inf. Forensics Secur. 2019, 14, 3116–3125. [Google Scholar] [CrossRef]
- Wan, C.; Wang, L.; Phoha, V.V. A Survey on Gait Recognition. ACM Comput. Surv. 2018, 51, 1–35. [Google Scholar] [CrossRef] [Green Version]
- Findling, R.D.; Hölzl, M.; Mayrhofer, R. Mobile Match-on-Card Authentication Using Offline-Simplified Models with Gait and Face Biometrics. IEEE Trans. Mob. Comput. 2018, 17, 2578–2590. [Google Scholar] [CrossRef] [Green Version]
- Cao, Y.; Zhang, Q.; Li, F.; Yang, S.; Wang, Y. PPGPass: Nonintrusive and Secure Mobile Two-Factor Authentication via Wearables. In Proceedings of the 2020 IEEE International Conference on Computer Communications, Toronto, ON, Canada, 6–9 July 2020. [Google Scholar]
- Diab, M.; Seif, A.; Sabbah, M.; El-Abed, M.; Aloulou, N. A Review on ECG-Based Biometric Authentication Systems; Springer: Berlin/Heidelberg, Germany, 2020; pp. 17–44. [Google Scholar] [CrossRef]
- Singh, Y.N. Human recognition using Fisher’s discriminant analysis of heartbeat interval features and ECG morphology. Neurocomputing 2015, 167, 322–335. [Google Scholar] [CrossRef]
- Liu, R.; Cornelius, C.; Rawassizadeh, R.; Peterson, R.; Kotz, D. Vocal Resonance: Using Internal Body Voice for Wearable Authentication. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 1–23. [Google Scholar] [CrossRef]
- Peng, G.; Zhou, G.; Nguyen, D.T.; Qi, X.; Yang, Q.; Wang, S. Continuous Authentication With Touch Behavioral Biometrics and Voice on Wearable Glasses. IEEE Trans.-Hum.-Mach. Syst. 2017, 47, 404–416. [Google Scholar] [CrossRef]
- Acar, A.; Aksu, H.; Uluagac, A.S.; Akkaya, K. WACA: Wearable-Assisted Continuous Authentication. In Proceedings of the 2018 IEEE Security and Privacy Workshops (SPW), San Francisco, CA, USA, 24 May 2018; pp. 264–269. [Google Scholar] [CrossRef] [Green Version]
- Blasco, J.; Peris-Lopez, P. On the Feasibility of Low-Cost Wearable Sensors for Multi-Modal Biometric Verification. Sensors 2018, 18, 2782. [Google Scholar] [CrossRef]
- Lin, F.; Cho, K.W.; Song, C.; Jin, Z.; Xu, W. Exploring a Brain-Based Cancelable Biometrics for Smart Headwear: Concept, Implementation, and Evaluation. IEEE Trans. Mob. Comput. 2020, 19, 2774–2792. [Google Scholar] [CrossRef]
- Marsico, M.D.; Mecca, A. A Survey on Gait Recognition via Wearable Sensors. ACM Comput. Surv. 2019, 52. [Google Scholar] [CrossRef] [Green Version]
- Gafurov, D.; Helkala, K.; Soendrol, T. Gait recognition using acceleration from MEMS. In Proceedings of the First International Conference on Availability, Reliability and Security (ARES’06), Vienna, Austria, 20–22 April 2006; pp. 6–439. [Google Scholar] [CrossRef]
- Gafurov, D.; Snekkenes, E.; Bours, P. Gait Authentication and Identification Using Wearable Accelerometer Sensor. In Proceedings of the 2007 IEEE Workshop on Automatic Identification Advanced Technologies, Alghero, Italy, 7–8 June 2007; pp. 220–225. [Google Scholar] [CrossRef]
- Rong, L.; Jianzhong, Z.; Ming, L.; Xiangfeng, H. A Wearable Acceleration Sensor System for Gait Recognition. In Proceedings of the 2007 2nd IEEE Conference on Industrial Electronics and Applications, Harbin, China, 23–25 May 2007; pp. 2654–2659. [Google Scholar] [CrossRef]
- Nowlan, M.F. Human Identification via Gait Recognition Using Accelerometer Gyro Forces. Available online: https://pdfs.semanticscholar.org/a63e/04fefd2be621488646ae11bfe66c98d9649e.pdf (accessed on 23 May 2020).
- Weiss, G.M.; Yoneda, K.; Hayajneh, T. Smartphone and Smartwatch-Based Biometrics Using Activities of Daily Living. IEEE Access 2019, 7, 133190–133202. [Google Scholar] [CrossRef]
- Al-Naffakh, N.; Clarke, N.; Li, F. Continuous user authentication using smartwatch motion sensor data. In Proceedings of the IFIP International Conference on Trust Management, Toronto, ON, Canada, 29 June–1 July 2018; pp. 15–28. [Google Scholar]
- Ahmad, M.; Alqarni, M.A.; Khan, A.; Khan, A.; Chauhdary, S.H.; Mazzara, M.; Umer, T.; Distefano, S. Smartwatch-Based Legitimate User Identification for Cloud-Based Secure Services. Mobile Inf. Syst. 2018. [Google Scholar] [CrossRef]
- Johnston, A.H.; Weiss, G.M. Smartwatch-based biometric gait recognition. In Proceedings of the IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS), Arlington, VA, USA, 8–11 September 2015; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
- Mansfield, A.J.; Wayman, J.L. Best Practices in Testing and Reporting Performance of Biometric Devices; Technical Report; Centre for Mathematics and Scientific Computing, National Physical Laboratory: Middlesex, UK, 2002. [Google Scholar]
- Mandalapu, H.; N, A.R.P.; Ramachandra, R.; Rao, K.S.; Mitra, P.; Prasanna, S.R.M.; Busch, C. Audio-Visual Biometric Recognition and Presentation Attack Detection: A Comprehensive Survey. IEEE Access 2021, 9, 37431–37455. [Google Scholar] [CrossRef]
- Jindal, G.; Kaur, G. A Comprehensive Overview of Quality Enhancement Approach-Based Biometric Fusion System Using Artificial Intelligence Techniques. In Communication and Intelligent Systems; Sharma, H., Gupta, M.K., Tomar, G.S., Lipo, W., Eds.; Springer: Singapore, 2021; pp. 81–98. [Google Scholar]
- Villalba, J.; Chen, N.; Snyder, D.; Garcia-Romero, D.; McCree, A.; Sell, G.; Borgstrom, J.; Richardson, F.; Shon, S.; Grondin, F.; et al. State-of-the-art Speaker Recognition for Telephone and Video Speech: The JHU-MIT Submission for NIST SRE18. In Proceedings of the Interspeech 2019, ISCA, Graz, Austria, 15–19 September 2019; pp. 1488–1492. [Google Scholar]
- Stragapede, G.; Vera-Rodriguez, R.; Tolosana, R.; Morales, A.; Acien, A.; Le Lan, G. Mobile behavioral biometrics for passive authentication. Pattern Recognit. Lett. 2022, 157, 35–41. [Google Scholar] [CrossRef]
- Vhaduri, S.; Dibbo, S.V.; Cheung, W. HIAuth: A Hierarchical Implicit Authentication System for IoT Wearables Using Multiple Biometrics. IEEE Access 2021, 9, 116395–116406. [Google Scholar] [CrossRef]
- Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification; John Wiley & Sons: Hoboken, NJ, USA, 2001. [Google Scholar]
- Saitoh, T.; Zhou, Z.; Zhao, G.; Pietikäinen, M. Concatenated Frame Image Based CNN for Visual Speech Recognition. In Computer Vision—ACCV 2016 Workshops; Chen, C.S., Lu, J., Ma, K.K., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 277–289. [Google Scholar]
- Gonzalez-Ferreras, C.; Escudero-Mancebo, D.; Vivaracho-Pascual, C.; Cardenoso-Payo, V. Improving Automatic Classification of Prosodic Events by Pairwise Coupling. IEEE Trans. Audio Speech Lang. Process. 2012, 20, 2045–2058. [Google Scholar] [CrossRef]
- Tekampe, N.; Merle, A.; Bringer, J.; Gomez-Barrero, M.; Fierrez, J.; Galbally, J. D6.5: Towards the Common Criteria Evaluations of Biometric Systems; Technical Report, Biometrics Evaluation and Testing (BEAT) Project. 2016. Available online: https://konfidas.de/downloads/D6.5.pdf (accessed on 23 November 2022).
- Sprager, S.; Juric, M.B. Inertial Sensor-Based Gait Recognition: A Review. Sensors 2015, 15, 22089–22127. [Google Scholar] [CrossRef]
- Sun, H.; Yuao, T. Curve aligning approach for gait authentication based on a wearable accelerometer. Physiol. Meas. 2012, 33, 1111–1120. [Google Scholar] [CrossRef]
- Ngo, T.T.; Makihara, Y.; Nagahara, H.; Mukaigawa, Y.; Yagi, Y. The largest inertial sensor-based gait database and performance evaluation of gait-based personal authentication. Pattern Recognit. 2014, 47, 228–237. [Google Scholar] [CrossRef] [Green Version]
- Zhang, Y.; Pan, G.; Jia, K.; Lu, M.; Wang, Y.; Wu, Z. Accelerometer-Based Gait Recognition by Sparse Representation of Signature Points With Clusters. IEEE Trans. Cybern. 2015, 45, 1864–1875. [Google Scholar] [CrossRef]
- Chereshnev, R.; Kertész-Farkas, A. HuGaDB: Human Gait Database for Activity Recognition from Wearable Inertial Sensor Networks. In Analysis of Images, Social Networks and Texts; Springer International Publishing: Cham, Switzerland, 2017; pp. 131–141. [Google Scholar]
- Kork, S.K.A.; Gowthami, I.; Savatier, X.; Beyrouthy, T.; Korbane, J.A.; Roshdi, S. Biometric database for human gait recognition using wearable sensors and a smartphone. In Proceedings of the 2017 2nd International Conference on Bio-Engineering for Smart Technologies (BioSMART), Paris, France, 30 August–1 September 2017; pp. 1–4. [Google Scholar] [CrossRef]
- Verma, A.; Moghaddam, V.; Anwar, A. Data-Driven Behavioural Biometrics for Continuous and Adaptive User Verification Using Smartphone and Smartwatch. Sustainability 2022, 14, 7362. [Google Scholar] [CrossRef]
- Luo, F.; Khan, S.; Huang, Y.; Wu, K. Activity-based person identification using multimodal wearable sensor data. IEEE Internet Things J. 2022, 10, 1711–1723. [Google Scholar] [CrossRef]
- Vecchio, A.; Nocerino, R.; Cola, G. Gait-based Authentication: Evaluation of Energy Consumption on Commercial Devices. In Proceedings of the 2022 IEEE International Conference on Pervasive Computing and Communications Workshops and other Affiliated Events (PerCom Workshops), Atlanta, GA, USA, 21–25 March 2022; pp. 793–798. [Google Scholar] [CrossRef]
- Giorgi, G.; Saracino, A.; Martinelli, F. Using recurrent neural networks for continuous authentication through gait analysis. Pattern Recognit. Lett. 2021, 147, 157–163. [Google Scholar] [CrossRef]
- Cheung, W.; Vhaduri, S. Context-Dependent Implicit Authentication for Wearable Device User. arXiv 2020. [Google Scholar] [CrossRef]
- Watanabe, K.; Nagatomo, M.; Aburada, K.; Okazaki, N.; Park, M. Gait-Based Authentication for Smart Locks Using Accelerometers in Two Devices. In Advances in Networked-Based Information Systems; Barolli, L., Nishino, H., Enokido, T., Takizawa, M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 281–291. [Google Scholar]
- Musale, P.; Baek, D.; Werellagama, N.; Woo, S.S.; Choi, B.J. You Walk, We Authenticate: Lightweight Seamless Authentication Based on Gait in Wearable IoT Systems. IEEE Access 2019, 7, 37883–37895. [Google Scholar] [CrossRef]
- Xu, W.; Shen, Y.; Zhang, Y.; Bergmann, N.; Hu, W. Gait-Watch: A Context-Aware Authentication System for Smart Watch Based on Gait Recognition. In Proceedings of the 2017 IEEE/ACM Second International Conference on Internet-of-Things Design and Implementation (IoTDI), Pittsburgh, PA, USA, 18–21 April 2017; pp. 59–70. [Google Scholar]
- Davidson, S.; Smith, D.; Yang, C.; Cheah, S.C. Smartwatch User Identification as a Means of Authentication; University of California San Diego: San Diego, CA, USA, 2016. [Google Scholar]
- Cola, G.; Vecchio, A.; Avvenuti, M. Continuous authentication through gait analysis on a wrist-worn device. Pervasive Mob. Comput. 2021, 78, 101483. [Google Scholar] [CrossRef]
- Lee, S.; Lee, S.; Park, E.; Lee, J.; Kim, I.Y. Gait-Based Continuous Authentication Using a Novel Sensor Compensation Algorithm and Geometric Features Extracted From Wearable Sensors. IEEE Access 2022, 10, 120122–120135. [Google Scholar] [CrossRef]
- Kececi, A.; Yildirak, A.; Ozyazici, K.; Ayluctarhan, G.; Agbulut, O.; Zincir, I. Implementation of machine learning algorithms for gait recognition. Eng. Sci. Technol. Int. J. 2020. [Google Scholar] [CrossRef]
- Wu, G.; Wang, J.; Zhang, Y.; Jiang, S. Authentication Scheme Based on Physiological and Behavioral Characteristics. Sensors 2018, 18, 179. [Google Scholar] [CrossRef] [Green Version]
- ISO/IEC. ISO/IEC 2382-37:2022(en) Information technology—Vocabulary—Part 37: Biometrics. Available online: https://www.iso.org/standard/73514.html (accessed on 23 November 2022).
- Tveit, B. Analyzing Behavioral Biometrics of Handwriting Using Myo Gesture Control Armband. Master’s Thesis, Fakultet for Naturvitenskap og Teknologi, The Artic University of Norway, Tromsø, Norway, 2018. [Google Scholar]
- Said, S.; Al-kork, S.; Nair, V.; Gowthami, I.; Beyrouthy, T.; Savatier, X.; Abdrabbo, M.F. Experimental Investigation of Human Gait Recognition Database using Wearable Sensors. Adv. Sci. Technol. Eng. Syst. J. 2018, 3, 201–210. [Google Scholar] [CrossRef] [Green Version]
- Sun, F.; Mao, C.; Fan, X.; Li, Y. Accelerometer-Based Speed-Adaptive Gait Authentication Method for Wearable IoT Devices. IEEE Internet Things J. 2019, 6, 820–830. [Google Scholar] [CrossRef]
- Lu, H.; Huang, J.; Saha, T.; Nachman, L. Unobtrusive Gait Verification for Mobile Phones. In Proceedings of the 2014 ACM International Symposium on Wearable Computers, Seattle, WA, USA, 13–17 September 2014; Association for Computing Machinery: New York, NY, USA, 2014; pp. 91–98. [Google Scholar] [CrossRef]
- Pascual-Gaspar, J.M.; Faundez-Zanuy, M.; Vivaracho, C. Fast on-line signature recognition based on VQ with time modeling. Eng. Appl. Artif. Intell. 2011, 24, 368–377. [Google Scholar] [CrossRef]
- Navarro, C.F.; Perez, C.A. Color-Texture Pattern Classification Using Global-Local Feature Extraction, an SVM Classifier, with Bagging Ensemble Post-Processing. Appl. Sci. 2019, 9, 3130. [Google Scholar] [CrossRef] [Green Version]
- Vivaracho-Pascual, C.; Faundez-Zanuy, M.; Pascual, J.M. An efficient low cost approach for on-line signature recognition based on length normalization and fractional distances. Pattern Recognit. 2009, 42, 183–193. [Google Scholar] [CrossRef]
- Cola, G.; Avvenuti, M.; Musso, F.; Vecchio, A. Gait-Based Authentication Using a Wrist-Worn Device. In Proceedings of the 13th International Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services, Hiroshima, Japan, 28 November–1 December 2016; MOBIQUITOUS 2016. Association for Computing Machinery: New York, NY, USA, 2016; pp. 208–217. [Google Scholar] [CrossRef] [Green Version]
- Kowalczyk, A. Support Vector Machines Succinctly; Syncfusion: Morrisville, NC, USA, 2017. [Google Scholar]
- Platt, J.C. Probabilistic Outputs for Support Vector Machines and Comparisons to Regularized Likelihood Methods. In Advances in Large Margin Classifiers; MIT Press: Cambridge, MA, USA, 1999; pp. 61–74. [Google Scholar]
- Kröse, B.; Van der Smagt, P. An Introduction to Neural Networks, 8th ed.; University of Amsterdam: Amsterdam, The Netherlands, 1996; Available online: https://www.researchgate.net/publication/272832321_An_introduction_to_neural_networks (accessed on 23 November 2022).
- Breiman, L. Random Forests. Mach. Learn. 2001, 1, 5–32. [Google Scholar] [CrossRef]
Work | Device | Classifier | Features | Window Size | Performance |
---|---|---|---|---|---|
Verma_22 [38] | WISDM-Database | Random Forest | Time domain: statistical, max-min value, time between peaks | 10 s (10 cycles **, approximately) | EER *** = 11% |
Vecchio_22 [40] | TicWatch E2 | K-NN * | Based on [47] | Based on [47] | EER = 5% |
Lee_22 [48] | Own-built wrist device | SVM | 2D cyclogram features | Tested from 1 to 9 cycles | EER = 5.8% |
Cola_21 [47] | Shimmer3 | SVM * | Time domain: statistical and autocorrelation-based | Tested from 2 to 6 cycles (called gait segment) | EER = 3.5% |
Giorgi_21 [41] | WISDM-Database | RNN * | raw data | 2.56 s | EER = 2.4% |
Kececi_20 [49] | Own-built | Ripper, MLP *, Random Forest, Decision Tree, k-NN, Bagging, Linear Regression, Random Tree, Naive Bayes, Bayesnet | Not found | Not found | FNMR = 0.3% FMR = 0.01% |
Cheung_20 [42] | Smart-Watch | SVM | Time domain: statistical features | 10-sample | EER = 6% |
Weiss_19 [18] | Smart-watch | k-NN, Decision Tree, Random Forest | Time domain: statistical, max-min value, time between peaks | 10 s (10 cycles, approximately) | EER = 6.8% |
Musale_19 [44] | Smart-watch | Random Forest, K-NN, MLP | Time domain: statistical, correlation-based, physical (pitch, roll and yaw), force | Tested from 1 to 10 cycles | EER = 8.2% |
Al-Naffakh_18 [19] | Smart-band | MLP | Time domain: statistical, correlation-based, max-min value, peaks-based | 10 s (10 cycles, approximately) | EER = 0.05% |
Wu_18 [50] | Own-built | SVM, ANN *, k-NN | Time domain (statistical, correlation, power, max-min) + Frequency domain (mean frequency, Bandwidth, Entropy) + Wavelet-domain (FFT Coefficient, Wavelet Energy) | Tested from 2 to 11 s (cycles, approximately) | FNMR = 5.0% FMR = 4.7% |
Xu_17 [45] | Smart-watch | Sparse Fusion | Sparse Fusion Classification | Tested from 1 to 6 cycles for identification task and fixed to 8 cycles for verification task | EER = 3.1% |
Johnston_15 [21] | Smart-watch | MLP, Random Forest, Rotation Forest, Naive Bayes | Time domain: statistical, time between peaks, max-min | 10 s (10 cycles, approximately) | EER = 1.4% |
Window Size | #Genuine | #Impostor |
---|---|---|
2 cycles | 11,353 | 224,650 |
4 cycles | 5486 | 108,285 |
8 cycles | 2632 | 51,701 |
12 cycles | 1729 | 33,824 |
Classifier | WS | RF | WSFPP |
---|---|---|---|
1-NN | 12 cycles | 21.4% | 14% |
Random Forest | 2 cycles | 3.2% | 0.2% |
MLP | 12 cycles | 20.1% | 8.6% |
SVM | 12 cycles | 0.4% | 0% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Salvador-Ortega, I.; Vivaracho-Pascual, C.; Simon-Hurtado, A. A New Post-Processing Proposal for Improving Biometric Gait Recognition Using Wearable Devices. Sensors 2023, 23, 1054. https://doi.org/10.3390/s23031054
Salvador-Ortega I, Vivaracho-Pascual C, Simon-Hurtado A. A New Post-Processing Proposal for Improving Biometric Gait Recognition Using Wearable Devices. Sensors. 2023; 23(3):1054. https://doi.org/10.3390/s23031054
Chicago/Turabian StyleSalvador-Ortega, Irene, Carlos Vivaracho-Pascual, and Arancha Simon-Hurtado. 2023. "A New Post-Processing Proposal for Improving Biometric Gait Recognition Using Wearable Devices" Sensors 23, no. 3: 1054. https://doi.org/10.3390/s23031054