Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
An Energy Data-Driven Approach for Operating Status Recognition of Machine Tools Based on Deep Learning
Previous Article in Journal
Occluded Pedestrian-Attribute Recognition for Video Sensors Using Group Sparsity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

User Authentication Method Based on Keystroke Dynamics and Mouse Dynamics with Scene-Irrelated Features in Hybrid Scenes

1
Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
2
School of Cyberspace Security, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(17), 6627; https://doi.org/10.3390/s22176627
Submission received: 28 July 2022 / Revised: 21 August 2022 / Accepted: 31 August 2022 / Published: 1 September 2022
(This article belongs to the Section Intelligent Sensors)

Abstract

:
In order to improve user authentication accuracy based on keystroke dynamics and mouse dynamics in hybrid scenes and to consider the user operation changes in different scenes that aggravate user status changes and make it difficult to simulate user behaviors, we present a user authentication method entitled SIURUA. SIURUA uses scene-irrelated features and user-related features for user identification. First, features are extracted based on keystroke data and mouse movement data. Next, scene-irrelated features that have a low correlation with scenes are obtained. Finally, scene-irrelated features are fused with user-related features to ensure the integrity of the features. Experimental results show that the proposed method has the advantage of improving user authentication accuracy in hybrid scenes, with an accuracy of 84% obtained in the experiment.

1. Introduction

With the development of computer technology and the internet, increasingly important data and personal information are being stored on computers and on the internet. Therefore, ensuring the security of these data is a growing concern. In recent years, biometric technologies have been widely used. A biometric system is an access control system that can distinguish between legal users and illegal users. Legal users can be authenticated to use the system, while illegal users cannot. Biometric systems allow only legal users to access the system while forbidding access to illegal users, even if they pretend to be legal users. Biometric systems can identify users by the inherent physiological characteristics of the human body (such as fingerprints and the iris) and by behavioral characteristics (such as sound, keystroke habits, and mouse usage habits). Compared with traditional user authentication methods (such as key, username, and password), biometrics has many advantages in that it is difficult to forget, is not easily forged, and is excellent anti-counterfeiting technology. In addition, the successful commercial use of biometrics based on physiological and behavioral characteristics, such as fingerprints, iris, and voice, proves that keystroke dynamics and mouse dynamics have a long-term development prospect.
Biometrics based on keystroke dynamics was first proposed by Gaines et al. in 1980 [1]. Unlike passwords, this method authenticates a user’s identity by the way they type. Keystroke dynamics is an analysis of people’s typing habits, so the key issue is not what the user types but how they type, such as how long they hold down a key or the interval between two keystrokes, which can produce unique patterns and characteristics of an individual. In addition, typing habits are hard to intercept or steal, so keystroke dynamics is an excellent user authentication scheme that can be added to conventional ID/password authentication schemes. Biometrics based on mouse dynamics was first proposed by Ahmed et al. in 2007 [2]. As an analysis of a person’s mouse usage habits, it studied the characteristics of the average moving speed of a mouse in all directions and the time intervals between single clicks or double clicks. The method of combining keystroke dynamics and mouse dynamics was also proposed by Ahmed et al. [3]. Using keystroke and mouse features at the same time enables them to complement each other while ensuring their respective performances so that a better performance can be achieved. Biometric systems based on keystroke dynamics and mouse dynamics can authenticate users when using computers without additional operations and can therefore continue authenticating users after they log into the system. In addition, as both technologies are based on the existing external devices of computers, the cost of authentication methods based on keystroke dynamics and mouse dynamics is lower than for other authentication methods, thus it has stronger generality as well as better development prospects.
In the practical use of computers, users employ the keyboard and mouse for a period of time, and therefore, compared with the fusion of keystroke features and mouse features, detection based only on a single type of feature will reduce the security and stability of the authentication system. Moreover, both keystroke dynamics and mouse dynamics authentication methods are limited to single scenes (such as only focusing on a typing scene) and real scenes (data collected by computers in daily life) [4,5]; however, user authentication in a single scene cannot be applied to real life because the user authentication system cannot determine the computer usage scene, while user authentication in real scenes has low accuracy due to the severe variability of data [6]. Hence, we believe that using multi-scene hybrid data (namely, hybrid scenes), which are close to the real scene data, to train the model can result in the effective authentication of users without accurate scene information. However, as the authentication accuracy of hybrid scenes is lower than that of all single scenes, user authentication accuracy in hybrid scenes is severely reduced.
In order to overcome the above research limitations and to improve the security and stability of user authentication systems, this paper proposes a method based on scene-irrelated features and user-related features that are selected from keystroke dynamics and mouse dynamics features in hybrid scenes. The selected features that have low correlations with scenes are named scene-irrelated features, and those that have high correlations with users are named user-related features. Scene-irrelated features and user-related features are then fused to obtain user-scene features for user authentication. The proposed method is defined as user authentication based on scene-irrelated features and user-related features (SIURUA), and the main contributions of this paper are summarized as follows:
  • A user authentication method is proposed to filter scene-irrelated features in hybrid scenes to reduce the impact of scenes on user authentication;
  • A user authentication method is proposed to fuse scene-irrelated features with user-related features in hybrid scenes.
The remainder of the paper is organized as follows. Section 2 introduces related work. The proposed scene-irrelated features, user-related features, user authentication algorithm, and the evaluation indices of experiments are introduced in Section 3. We provide the experimental configurations and analyses of the experimental results in Section 4. Finally, Section 5 concludes the paper and outlines future work.

2. Related Research

2.1. Feature Selection

Feature selection is a common method used to improve the accuracy of classification models and can be divided into filter methods, wrapper methods, and embedded methods [7]. Filter methods are independent of machine learning algorithms and use evaluation criteria to enhance the correlations between features and classes and to reduce the correlations between features and features. Wrapper methods use the accuracy of learning algorithms to evaluate feature subsets. Embedded methods automatically select feature subsets in the training process, making up for the shortcomings of filter methods and wrapper methods.

2.1.1. Filter Methods

Filter methods select the features by calculating their weights before using the learning algorithm. The simplest filtering feature selection methods are mutual information [8], chi-square test, and F-test, which all select features based on the correlations between features and labels. In addition, some new filter methods have been proposed recently. Cai et al. [6] proposed a mouse dynamics feature dimension reduction method based on multidimensional scaling (MDS) and isometric feature mapping (ISOMAP). This method generated a weight matrix first, then the ratio of the sum of certain eigenvalues to the sum of total eigenvalues was compared with the threshold. If the ratio was greater than the threshold, the corresponding feature of the eigenvalue was selected as the feature subset. This method could effectively reduce the behavioral variability of mouse dynamics and improve the learning effect.

2.1.2. Wrapper Methods

Wrapper methods select the optimal subset based on an analysis of the pros and cons of feature subsets through models. The most common wrapper method is a feature subset search, such as a genetic algorithm [9] and hill-climbing [10]. Yang et al. proposed a multi-task feature selection method based on a Support Vector Machine (SVM) and a multi-task matrix [11]. This method could optimize the feature weight of a task by using the features of the remaining tasks in multi-task learning. Finally, the feature subset was selected according to the weight to achieve the purpose of improving the learning effect.

2.1.3. Embedded Methods

Embedded methods apply feature selection to the learning process, with the advantage of not needing to evaluate different feature subsets. Common embedded methods include 1 −regularization [12] and 2 −regularization [13]. Regularization is a method of adjusting feature coefficients through the value of a regularization term, and the feature selection is completed when some of the feature coefficients reduce to zero. Feiping et al. [14] proposed a feature selection method based on 2 , 1 −norm. In a multi-task experiment, the application of 2 , 1 −norm to the feature coefficient matrix composed of the feature coefficients of each task resulted in the matrix’s rows becoming dense and the columns becoming sparse through learning, so that multi-task feature selection was realized through intertask sharing of information.

2.2. Keystroke Dynamics and Mouse Dynamics

2.2.1. Keystroke Dynamics

Since being first proposed by Gaines et al. in 1980 [1], biometric technology based on keystroke dynamics has been developed for 40 years, and some researchers have made considerable progress. Research into keystroke dynamics can be divided into two categories: static authentication based on fixed text and dynamic authentication based on the free text [15].
In recent years, with the development of computer technologies, research into keystroke dynamics has made great advancements. In 2007, Azevedo et al. [16] proposed a hybrid system based on the combination of SVM and genetic algorithm (GA). The experiment obtained a 1.18% False Acceptance Rate (FAR) and 1.58% False Recognition Rate (FRR), while the feature dimension was reduced by 47.51%. Arwa et al. [17] suggested using the fusion method to improve the performance of user authentication based on free text keystroke dynamics. They proposed novel keystroke dynamics features called “semi-timing features”, which had been proved to appear in most users’ keystroke behaviors. The authors combined traditional keystroke dynamics features with “semi-timing features” and used SVM for classification, obtaining 1.56% FAR and 21.5% FRR. Epp et al. [18] applied keystroke dynamics to emotional recognition and achieved an accuracy of between 77% and 80%. Antal et al. [19] fused keystroke dynamics into smartphones and obtained a 15.3% Equal Error Rate (EER). Tsai and Huang [20] put forward a method to detect fraudulent messages through a voting-based statistical classifier to analyze users’ keystroke dynamics, obtaining 10.4% EER. Ayotte et al. [21] proposed a novel instance-based graph comparison algorithm called ITAD that could reduce the keystroke number for user authentication and ultimately obtained 7.8% EER for the Clarkson II dataset and 3.0% EER for the Buffalo dataset. With the development of the Artificial Neural Network (ANN), some neural network-based keystroke dynamics user authentication methods were proposed. For example, Tewari and Verma [22] combined keystroke dynamics data and image data with artificial image data, using AlexNet and ResNet to classify the artificial image, and achieving an accuracy of 98.57%. Lu et al. [23] combined a Convolution Neural Network (CNN) with Recurrent Neural Network (RNN) architecture to test the model using the sliding window extraction of n-gram features, achieving an EER of 2.67% in the best case.

2.2.2. Mouse Dynamics

User authentication based on mouse dynamics can be divided into continuous authentication and static authentication. Continuous authentication means authenticating the user at all times while the user is using the system. Static authentication means authenticating the user under certain circumstances.
Early mouse dynamics research focused on the recognition of user electronic signatures. Higashino et al. [24] used neural networks to study handwriting signatures. In 2003, Everitt et al. [25] conducted a study on signing with a mouse. In 2007, Ahmed et al. [2] verified the feasibility of using mouse movement data to authenticate identity. The proposed features extraction method has been used to date, and the authors obtained 2.46% FAR and 2.46% FRR through experiments. Since then, increasingly more research papers based on mouse dynamics have been published.
Fecher et al. [26] proposed new mouse dynamics features, such as jitters and straightness, and then input these features and other features proposed by Ahmed into a multi-layer classifier based on random forest, finally obtaining 7.5% EER. Kasprowski et al. [27] proposed a biometric method fusing mouse dynamics and eye movement biometrics, finally achieving 92.9% accuracy and 6.8% EER. Gao et al. [28] proposed a continuous authentication method using mouse dynamics based on decision-level fusion. Antal et al. [29] used a convolutional neural network to learn the mouse dynamics features directly, obtaining 0.94 AUC. Hu et al. [30] visualized mouse movements as images and used CNN to classify the images, ultimately attaining 2.94% FAR and 2.28 FRR.

2.2.3. Fusion of Keystroke Dynamics and Mouse Dynamics

A keyboard and a mouse are the main devices by which a user interacts with a computer. Over a period of time, a user will use a keyboard and mouse at the same time to operate a computer, providing the possibility for the fusion of keystroke dynamics and mouse dynamics. Ahmed et al. [3] verified the fusion of keystroke dynamics and mouse dynamics for the first time. They spliced keystroke dynamics and mouse dynamics, then input the features into a neural network, finally obtaining 1.312% FAR and 0.651% FRR. Bailey et al. [31] used the J48 algorithm to achieve decision-level fusion for keystroke dynamics features, mouse dynamics features, and GUI interaction features, finally achieving 2.1% FAR and 2.24% FRR. Mondal et al. [32] proposed a continuous authentication system based on the fusion of keystroke dynamics and mouse dynamics. The confidence of the users of this system depended on the deviation of the users’ operations, with legal users locked after 40,134 operations and illegal users locked after 419 operations. Some studies applied keystroke dynamics and mouse dynamics to soft biometric identification. For example, Earl et al. [33] proposed the use of keystroke and mouse dynamics features to identify users’ gender, handedness, or age, and their research demonstrates that biometric authentication technology can be used in many more areas.

2.3. Multiple Kernel Learning

In 1992, Boser et al. [34] introduced the concept of kernel function into machine learning when they researched the support vector machine algorithm. The kernel function maps the linearly inseparable eigenvector x in original feature space to the linearly separable eigenvectors ϕ ( x ) in high-dimensional space. In fact, choosing the correct kernel has an even stronger impact on the classification results compared with the classifier, but the data may come from different distributions, and it may be necessary to use different kernels for mapping. The objective of multiple kernel learning (MKL) is to combine kernels from multiple sources to improve the accuracy of the target kernel. The parameters used in MKL are generally positive, and the MKL formula is:
K = r = 1 R η r K r ,       η r 0
Based on this formula, MKL is used to combine multiple kernels into a large kernel to improve classification accuracy.
The simplest MKL algorithm is AverageMKL, proposed by Belanche et al. [35], in which the parameter of each kernel is the reciprocal of the total number of kernels. Hence, the kernel fusion formula is:
k μ ( x , z ) = r P μ r k r ( x r , z r ) ,       μ r = 1 P  
Kloft et al. [36] and Xu et al. [37] proposed a lasso-based MKL algorithm that uses 1 −norm to regularize kernel weights. Do et al. [38] found that kernel combination maximizes the decision boundary by fusing the kernel radius into the MKL, naming it Radius MKL (R-MKL). EasyMKL [39] is an improved version of AverageMKL that obtained a combination of kernel parameters through learning.
Multiple kernel learning can improve the accuracy of models by combining multiple types of features. Therefore, MKL has a broader application prospect in machine learning.

3. Proposed Approach

In this section, we will describe the proposed SIURUA in detail. In order to reduce the impact of different scenes on the hybrid scene features, we propose the selection of scene-irrelated features and user-related features and then propose fusing them to improve the authentication accuracy of the model in hybrid scenes. Figure 1 shows the block diagram of SIURUA. The steps include feature extraction, feature processing, and model training. In Figure 1, we can see the basic process of SIURUA in more detail:
  • First, features are extracted from the collected user operation data (details will be provided in Section 3.1);
  • Second, scene-irrelated features and user-related features are selected from the original features (details will be elaborated in Section 3.2);
  • Finally, scene-irrelated features and user-related features are fused and the model is trained (details will be presented in Section 3.3).

3.1. Feature Extraction

The features adopted in this experiment are keystroke dynamics features and mouse dynamics features. We extract features according to different time lengths. Each dimension of the keystroke dynamics feature and mouse dynamics feature is the average of a feature extracted by a user operating within the time length, so the numbers of user operations are varied at different time lengths, and the extracted feature vectors are different. We use the variable time window to represent the length of time used to extract the features; for example, time window = 60 s or time window = 120 s. In the feature extraction section, features are extracted according to the value of the time window, so the value of the time window is the decisive factor determining the amount of information contained in the features.
In the next feature selection section, we select features from the extracted original features based on the user operation, and we define the original features as O r i g F e a t u r e = ( f 1 , f 2 , , f l ) , where l is the total number of dimensions of the original features. The original user operation features consist of the directly extracted keystroke dynamics features and the mouse dynamics features. Therefore, we define the keystroke dynamics feature as K e y = ( k 1 , k 2 , , k n ) (as shown in Section 3.1.1) and the mouse dynamics feature as M o u s e = ( m 1 , m 2 , , m m ) (as shown in Section 3.1.2), where n is the dimension size of keystroke dynamics features and m is the dimension size of mouse dynamics features. Then, we directly splice the keystroke dynamics features and mouse dynamics features without additional operations to obtain the user operation original features, as illustrated above. Finally, we obtain the original features O r i g F e a t u r e = ( f 1 , f 2 , , f l ) = ( k 1 , k 2 , , k n , m 1 , m 2 , , m m ) , where l = n + m . The above features are extracted from the collected keystroke data and mouse data according to the time window, and the values of the time window are 10 s, 20 s, 30 s, 40 s, 50 s, 60 s, 120 s, 180 s, 240 s, 300 s, 360 s, 420 s, and 480 s. In the process of extracting the features, the original features are extracted from all data under the same value of time window, and empty feature components are filled with 0. Finally, we obtain the features shown in Table 1 ( i representing the number of features and j representing the dimension of features).
Next, we will elaborate on the extraction methods for keystroke dynamics features and mouse dynamics features.

3.1.1. Keystroke Features

We obtain the key pairs according to the time keys are pressed. After the key pairs are obtained, the keystroke features can be extracted [40]. According to the pressing time (down) and the release time (up) of each key, the single key features, and key pair features, as shown in Figure 2, can be obtained (in milliseconds) [41]:
  • Single key features: keystroke duration (KD). Each key pair has two single key features:
    • the keystroke duration of the first key (KD1);
    • the keystroke duration of the second key (KD2).
  • Key pair features: the diagram latency between two keys. Each key pair has six key pair features:
    • down-down diagram latency (DDDL);
    • down-up diagram latency (DUDL);
    • up-down diagram latency (UDDL);
    • up-up diagram latency (UUDL).
The above keystroke features can be applied to every key and every key pair. For a full-size keyboard with 110 keys, the KD features of a single key are 110 dimensions, and the DDDL, DUDL, UDDL, and UUDL features of the key pairs are 12,100 dimensions; thus, there are 48,400 dimensions in total. Hence, the keystroke features in each time window have 48,510 dimensions. Therefore, the keystroke dynamics feature dimension n (defined in Section 3.1) is equal to 48,510, and the keystroke dynamics features are expressed as K e y = ( k 1 , k 2 , , k 48510 ) .

3.1.2. Mouse Features

We extract mouse dynamics features according to the operation types [2]. On the basis of the recorded mouse movement data—left-click data, right-click data, and eight movement directions are shown in Figure 3—we can extract seven types of mouse features:
  • Movement speed compared with traveled distance (MSD): The average speed of all moves within different moving distances. The range of moving distance is every interval of 100 pixels within 1–800 pixels (such as 1–100 pixels, 101–200 pixels). The length of the MSD feature vector is 8;
  • Average movement speed per movement direction (MDA): The average speed of all movements in different moving directions. The moving direction is divided into eight equal parts. The length of the MDA feature vector is 8;
  • Average speed for different types of actions (ATA): The average mouse movement speed of different operation types. There are three operations, namely, mouse-move, drag-drop, and point-click. The length of the ATA feature vector is 3;
  • Traveled distance histogram (TDH): The ratio of the number of mouse operations in different moving distances to the total number of mouse operations. The length of the TDH feature vector is 8;
  • Movement direction histogram (MDH): The ratio of the number of mouse operations in different moving directions to the total number of mouse operations. The length of the MDH feature vector is 8;
  • Actions types histogram (ATH): The ratio of the number of mouse operations of different operation types to the total number of mouse operations. The length of the ATH feature vector is 3;
  • Movement elapsed time histogram (MTH): The ratio of the number of mouse operations in different operation durations to the total number of mouse operations. The time range is 5 time periods separated by 200 milliseconds within 1–1000 milliseconds. The length of the MTH feature vector is 5.
We combine the above features into a feature vector, finally obtaining 43-dimensional mouse dynamics features. Therefore, the mouse dynamics feature dimension m (defined in Section 3.1) is equal to 43, and the mouse dynamics features are expressed as M o u s e = ( m 1 , m 2 , , m 43 ) .

3.2. Feature Processing

3.2.1. Scene-Irrelated Features and User-Related Features

As single scene data are collected in restricted environments, we consider that there will be some changes in users’ keystrokes and mouse operations in different scenes, and we call the factors that lead to those changes “scene information”. Due to the scene information, we consider that features that are highly correlated with scenes will affect user authentication accuracy. Therefore, scene-irrelated features can be selected to effectively distinguish users and reduce the correlation between features and scenes. We calculate the correlation C O R s c e n e ( f i ) between each dimension of features f i and scenes. Based on these data, we obtain a sequence of scene correlation degree:
S c e n e C O R = { C O R s c e n e ( f 1 ) ,   C O R s c e n e ( f 2 ) ,   , C O R s c e n e ( f j ) } ,
and then select n dimensions of features according to the inequality:
C O R s c e n e ( f i ) < = C O R s c e n e ( f n ) ,
where C O R s c e n e ( f n ) is the n - t h lowest correlation and C O R s c e n e ( f i ) is any correlation less than or equal to C O R s c e n e ( f n ) in S c e n e C O R . Finally, we obtain n -dimensional features that have the lowest correlation with scenes. We name them “scene-irrelated features” and define them as S c e n e i r r e l a t e d = ( S i 1 ,   S i 2 ,   , S i n ) .
Contrary to the scene-irrelated features, some features will have great differences between different users and will have few differences for the same user. These features are more distinguishable to users than other features, therefore we call these features “user-related features”. Using user-related features to classify can achieve excellent results. We calculate the correlation C O R u s e r ( f i ) between each dimension of features f i and users. Based on these data, we obtain a sequence of user correlation degree:
User COR = { C O R u s e r ( f 1 ) ,   C O R u s e r ( f 2 ) ,   , C O R u s e r ( f j ) } ,
and then select m dimensions of features according to the inequality:
C O R u s e r ( f i ) > = C O R u s e r ( f m ) ,
where C O R u s e r ( f m ) is the m - t h lowest correlation and C O R u s e r ( f i ) is either correlation in U s e r C O R . Finally, we obtain m -dimensional features that have the highest correlation with users. We name them as user-related features and define them as U s e r r e l a t e d = ( U r 1 ,   U r 2 ,   , U r m ) .
We consider that user-related features still contain some scene information in hybrid scenes; the ratio of scene information contained in user-related features can be reduced by fusing scene-irrelated features (we call the generated features “user-scene features”), which can improve the user authentication accuracy in hybrid scenes. We define user-scene features as U s e r S c e n e = K ( U r , S i ) , where K ( · ) is the kernel fusion function (which will be introduced in Section 3.3).

3.2.2. Feature Selection Method

The SIURUA algorithm searches for scene-irrelated features and user-related features, so it is necessary to measure the correlations between features and users and the correlations between features and scenes. As mutual information cannot effectively reflect the correlation between two datasets when the features have many values, adjusted mutual information (AMI) is used to calculate the user correlation and scene correlation of our SIURUA algorithm [42]. First, we number the user sequence U s e r = ( u 1 , u 2 , , u i ) according to 41 users, and the scene sequence S c e n e = ( s 1 , s 2 , , s i ) according to 4 scenes, and separate each dimension of the original features O r i g F e a t u r e to obtain the sequence f j = ( f 1 , j , f 2 , j , , f i , j ) . We subsequently calculate C O R u s e r ( f j ) = A M I ( f j , U s e r ) and C O R s c e n e ( f j ) = A M I ( f j , S c e n e ) , where A M I ( f j , U s e r ) and A M I ( f j , S c e n e ) are calculated using the AMI calculation formula introduced in [42]:
A M I ( f j , U s e r ) = M I ( f j , U s e r ) E { M I ( f j , U s e r ) } a v g { H ( f j ) , H ( U s e r ) } E { M I ( f j , U s e r ) }
and
A M I ( f j , S c e n e ) = M I ( f j , S c e n e ) E { M I ( f j , S c e n e ) } a v g { H ( f j ) , H ( S c e n e ) } E { M I ( f j , S c e n e ) } ,
where E { M I ( f j , U s e r ) } and E { M I ( f j , S c e n e ) } represent the expectations of mutual information M I ( f j , U s e r ) and M I ( f j , S c e n e ) ; and H ( f j ) , H ( U s e r ) , and H ( S c e n e ) are the entropies of f j = ( f 1 , j , f 2 , j , , f i , j ) , U s e r = ( u 1 , , u i ) , and S c e n e = ( s 1 , s 2 , , s i ) , respectively.
The calculation formulas of the entropies mentioned above are H ( f j ) = m = 1 i P ( f m , j ) l o g P ( f m , j ) , H ( U s e r ) = m = 1 i P ( u m ) l o g P ( u m ) , and H ( S c e n e ) = m = 1 i P ( s m ) l o g P ( s m ) , where log denotes the logarithm with a base of two. P ( · ) is the occurrence with a probability of f m , j , u m , and s m , where 1 m i .
In addition, the expectation of mutual information is introduced in [42], and therefore M I ( f j , U s e r ) and M I ( f j , S c e n e ) are calculated as follows:
E { M I ( f j , U s e r ) } = k = 1 M j l = 1 41 n u k l = m a x ( a k , j + b l N U , 0 ) m i n ( a k , j , b l ) n u k l N U l o g ( N U · n u k l a k , j b l ) × a k , j ! b l ! ( N U a k , j ) ! ( N U b l ) ! N U ! n u k l ! ( a k , j n u k l ) ! ( b l n u k l ) ! ( N U a k , j b l + n u k l ) !
and
E { M I ( f j , S c e n e ) } = k = 1 M j l = 1 4 n s k l = m a x ( a k , j + c l N S , 0 ) m i n ( a k , j , c l ) n s k l N S l o g ( N S · n s k l a k , j c l ) × a k , j ! c l ! ( N S a k , j ) ! ( N S c l ) ! N S ! n s k l ! ( a k , j n s k l ) ! ( c l n s k l ) ! ( N S a k , j c l + n s k l ) !
where M j is the number of clusters of clustering f j = { F 1 , j ,   F 2 , j ,   ,   F M j , j } . Similarly, the number of clusters of clustering U s e r = { U 1 ,   U 2 , ,   U 41 } and S c e n e = { S 1 ,   S 2 , S 3 , S 4 } are 41 and 4, respectively. In addition, a k , j = | F k , j | , b l = | U l | , c l = | S l | , n u k l = | F k , j U l | , and n s k l = | F k , j S l | . Finally, N U = k l | F k , j U l | and N S = k l | F k , j S l | .
For example, we have a set of data, as shown in Table 2, which is a feature component, and each dimension of features has its corresponding S c e n e label.
First, calculate the information entropy and mutual information of f 1 and S c e n e to obtain H ( f 1 ) = 1.33218 and H ( S c e n e ) = 1.05492 , so a v g { H ( f 1 ) , H ( S c e n e ) } = 1.19355 . Second, calculate the mutual information and the expectation of mutual information of f 1 and S c e n e to obtain M I ( f 1 , S c e n e ) = 1.05492 and E { M I ( f 1 , S c e n e ) } = 0.83311 . Finally, calculate the adjusted mutual information A M I ( f 1 , S c e n e ) = 0.61539 .
The relationships between these variables are shown in Table 3 and Table 4.
After obtaining the A M I ( f j , U s e r ) and A M I ( f j , S c e n e ) of each feature dimension, all values are classified into two sequences of correlation degree: sequence of user correlation degree User COR = { C O R u s e r ( f 1 ) ,   C O R u s e r ( f 2 ) ,   , C O R u s e r ( f 48553 ) } = { A M I ( f 1 , U s e r ) , A M I ( f 2 , U s e r ) , , A M I ( f 48553 , U s e r ) } and sequence of scene correlation degree S c e n e C O R = { C O R s c e n e ( f 1 ) ,   C O R s c e n e ( f 2 ) ,   , C O R s c e n e ( f 48553 ) } = { A M I ( f 1 , S c e n e ) , A M I ( f 2 , S c e n e ) , , A M I ( f 48553 , S c e n e ) } . The values of U s e r C O R are sorted from largest to smallest, and the values of S c e n e C O R are sorted from smallest to largest, then the user-related features U s e r r e l a t e d and scene-irrelated features S c e n e i r r e l a t e d are selected as the features used for the authentication model.

3.3. Feature Fusion Based on MKL

The scene-irrelated features and user-related features that are obtained by the method described in Section 3.2 can improve authentication accuracy. The authentication accuracy can be further improved by fusion. As the two sets of features are composed of different feature components, the two sets of features have different contributions to user authentication. Therefore, feature fusion is essential for combining the advantages of the two groups of features to improve the user authentication accuracy of SIURUA. In this process, EasyMKL is adopted to fuse scene-irrelated features and user-related features. Support vector machine is used for classification. EasyMKL is an improved multiple kernel learning algorithm compared with other multiple kernel learning algorithms represented by AverageMKL. EasyMKL can obtain the optimal parameter combination through learning, and different parameters can provide different weights to each kernel function. This characteristic of EasyMKL is applicable to the proposed fusion of scene-irrelated features and user-related features, as user-related features are helpful in classifying users, and scene-irrelated features can contribute to diluting the scene information in user-related features while classifying users. The two features have different contributions to user authentication, so EasyMKL is chosen for kernel fusion to ensure that the two features can obtain the optimal weight.
EasyMKL can be used to fuse scene-irrelated features and user-related features through two kinds of kernel changes and weights and then combine them for classification. For example, the RBF kernel is used for the scene-irrelated features, and the linear kernel is used for the user-related features. Here is an example of the formula, and kernel function selection is described in detail in Section 4.2.4. In this case, the corresponding multiple kernel learning formula proposed in [39] becomes:
K ( S c e n e i r r e l a t e d , U s e r r e l a t e d ) = μ 1 k 1 ( S i i , S i j ) + μ 2 k 2 ( U r i , U r j ) ,     μ 1 0 μ 2 0   μ 1 + μ 2 = 1
where k 1 ( S i i , S i j ) = e x p ( | | S i i S i j | | 1 σ 2 ) is the RBF kernel, and k 2 ( U r i , U r j ) = U r i T U r j represents the linear kernels, S i i , S i j , U r i , and U r j , which, as described in Section 3.2, are the components of S c e n e i r r e l a t e d and U s e r r e l a t e d . μ 1 and μ 2 are the weights of k 1 ( · ) and k 2 ( · ) . Taking Formula (11) as an example, the U s e r S c e n e features are obtained.

4. Experiments and Result Analysis

This section aims to validate the feasibility of SIURUA through a series of experiments and compare it with existing algorithms to analyze the experiment results.

4.1. Experiment

4.1.1. Data Collection

Forty-five students from the Beijing University of Technology participated in the data collection work. They were asked to perform four tasks (typing, Taobao, Weibo, and gaming) on the same laptop, with each task taking one hour. User data of those that did not complete all four tasks were filtered. Finally, the data of 41 users remained; that is, the total number of users in the experiment is N = 41 .
The four tasks involved are the most common ones performed by users of computers. Typing required users to type the same article on Microsoft Office Word; Taobao asked users to browse Taobao; in the Weibo data collection process, users browsed Weibo; and in the gaming data collection process, users played a home-made game. Among the four tasks, the typing task was mainly used to collect keystroke data, the gaming task mainly collected mouse movement data, and the remaining two tasks mainly required a mouse to operate, so we were able to collect a large amount of mouse movement data and a small amount of keystroke data.
The home-made gaming task tested the user’s ability to control the mouse. The game randomly displayed circles on the screen, with the circles gradually shrinking until they disappeared. The users needed to click on the circle before the circles disappeared. If users missed any circles, the game would end. The data collection program collected the user’s keyboard and mouse operation details. The keyboard data recorded the pressing and releasing time for each key, and the mouse data recorded the mouse movement, click, release, and scroll times, along with the coordinates on the screen. The collected data were saved in text files.
The final collected data format samples are shown in Figure 4 and Figure 5. Figure 4 shows an example of the collected keystroke data. The first column represents the time in milliseconds; in the second column, “key dn” means that the key is pressed and “key up” means that the key is released, and the third column shows the data regarding which keys were recorded. Therefore, Figure 4 corresponds to the process of the user inputting the word “IS”. Figure 5 shows a sample of mouse movement data. The first column represents the type of mouse operation (e.g., 512 represents mouse movement), the second column shows the operation time in milliseconds, the third column displays the x-axis coordinate of the screen position of the mouse, and the fourth column represents the mouse y-axis coordinates.
The equipment used in the data collection process of this experiment is Apple MC968CH/A, the processor of this equipment is Intel [email protected] GHz, the memory size is 4.00 GB, and the system is 64-bit Windows 10 Professional Edition.

4.1.2. Process Description

After obtaining keystroke data and mouse movement data, we extracted keystroke dynamics features and mouse dynamics features from the collected data, with the feature extraction performed within a time window. If there was more than one same operation in a time window, the average operation value was calculated as the feature value (as described in Section 3.1). The following are the detailed steps of the experiment process:
  • If user   i   ( 1     i     N )   is marked as a legal user, the remaining 40 ( N 1 ) users are marked as illegal users. The legal user data are taken as positive samples, and the illegal user data are taken as negative samples;
  • In order to prevent data imbalance, negative samples are under-sampled, and the same number of negative samples as positive samples are randomly obtained;
  • After combining the positive and negative samples, the scene-irrelated features and user-related features are selected and fused (according to the method described in Section 3.2) to obtain the user-scene features U s e r S c e n e used for classification;
  • The classification models based on a support vector machine and MKL are trained and tested by 5-fold cross-validation using the obtained U s e r S c e n e features.
Steps 1–4 are repeated 41 times until each user acts as a positive sample, and 41 models are trained and tested with each user. Finally, the experiment results of 41 users are averaged to evaluate the performance of SIURUA.

4.1.3. Evaluation

After obtaining the authentication model of a user through the method described in Section 4.1.2, we can calculate the indicators of true positive (TP), false positive (FP), true negative (TN), and false negative (FN). After obtaining the above four basic indicators, we can calculate the well-known evaluation indicators in this field, including accuracy, precision, true positive rate (TPR), false positive rate (FPR), false accept rate (FAR), and false reject rate (FRR), and the F1 scores are used to estimate the classification quality of the model. The calculation methods for FPR and FAR are the same, so they have the same evaluation index. We calculate the above indicators of the model of user i through Formulas (12)–(17):
A c c i = T P i + T N i T P i + T N i + F P i + F N i
P r e c i = T P i T P i + F P i
T P R i = T P i T P i + F N i
F A R i ( F P R i ) = F P i F P i + T N i
F R R i = F N i F N i + T P i  
F 1 S c o r e i = 2 T P i 2 T P i + F P i + F N i
Since this experiment trains a model for each user, the average of the above evaluation indices is calculated in the experiment to evaluate the classification quality. We named these evaluation indices: average accuracy (aAcc), average precision (aPrec), average true positive rate (aTPR), average false positive rate (aFPR), average false accept rate (aFAR), average false reject rate (aFRR), and average F1 score (aF1), and their definitions are shown in Formulas (18)–(23):
a A c c = i = 1 N A c c i N
a P r e c = i = 1 N P r e c i N
a T P R = i = 1 N T P R i N
a F A R ( a F P R ) = i = 1 N F A R i N
a F R R = i = 1 N F R R i N
a F 1 = i = 1 N F 1 S c o r e i N

4.2. Analysis of Experimental Results

4.2.1. Illustrate the Reduction in User Authentication Accuracy of Hybrid Scenes

To verify the viewpoint that “the authentication accuracy of a hybrid scene is lower than that of all single scenes”, we trained the user authentication model based on a support vector machine with a linear kernel in a single scene and a hybrid scene consisting of four scenes. Figure 6 provides the aAcc of the authentication models. It can be seen that in each time window, the authentication aAcc of the hybrid scenes is lower than that of the signal scenes. Therefore, improving the accuracy of user authentication in hybrid scenes is valuable.

4.2.2. Determine Suitable Feature Combination

In the process of using machine learning algorithms to classify, the number of selected features has an impact on the classification quality. This research has two steps for feature selection: selecting the scene-irrelated features and selecting the user-related features. The selected feature number of these two feature selection processes will affect the classification quality; therefore, this experiment is to determine the best-selected feature number. Figure 7 shows 18 feature combinations; for example, 300_200 represents the combination of 300 user-related features and 200 scene-irrelated features. It can be seen that the optimal aAcc (80.86%) is obtained when the features consist of 200 user-related features and 200 scene-irrelated features. Therefore, the suitable feature combination chosen for the SIURUA algorithm combines 200 user-related features and 200 scene-irrelated features.

4.2.3. Determining Suitable Time window

The different values of the time window selected during feature extraction will have a certain impact on the classification algorithm; therefore, this experiment is introduced to determine a suitable time window. Specifically, the time windows selected in this experiment are 10 s, 20 s, 30 s, 40 s, 50 s, 60 s, 120 s, 180 s, 240 s, 300 s, 360 s, 420 s, and 480 s. We combined 200 user-related features and 200 scene-irrelated features based on the experience described in Section 4.2.2. As shown in Figure 8, we used SIURUA, SVM with linear kernel, decision tree (DT), logistic regression (LR), and naive Bayes (NB) to classify the hybrid scene data. SIURUA uses a combination of 200 user-related features and 200 scenario-irrelated features, while the other algorithms use all features. It can be seen from Figure 8 that SIURUA achieved the maximum aAcc (82.5%) at 300 s. SVM and LR achieved the maximum aAcc (75.2% and 75.8%, respectively) at 360 s. DT gained the maximum aAcc (71.2%) at 120 s. NB achieved the maximum aAcc (59.8%) when the time window was 420 s. It can be seen that SIURUA can achieve better classification results than other algorithms when the time window = 300 s. It is worth noting that, although the time window is longer compared with DT, the best aAcc of SIURUA has a 16% improvement compared with the best aAcc of DT. When compared with SVM, NB, and LR, SIURUA improves the classification quality while shortening the time window required to obtain the best aAcc. Therefore, the suitable value of the time window for the SIURUA algorithm is chosen to be 300 s (5 min), and in our hybrid scene data, there are 146 keystroke operations and 314 mouse operations on average in 300 s.
It can be seen from Figure 8 that the authentication accuracy fluctuates. The authentication accuracy of SIURUA decreases slightly after 300 s, the authentication accuracy of DT decreases after 180 s, and the authentication accuracy of LR and SVM decreases after 360 s. The time window becomes longer, and the user’s operation within a time window changes. For example, in the Taobao scene, the user changes from browsing the product details page to browsing the product list page. The uncertainty brought by the switching of applications in the same scene affects the accuracy of the model authentication, and different models have different abilities to resist such interference.
As shown in Table 5, we compare the computational cost of SVM, DT, NB, LR, and SIURUA. The comparison is divided into two parts: building and authentication. The time complexity of building a classifier is decided by the machine learning algorithm [20]. For example, if we use SVM, the time complexity is O ( n 3 ) , where n is the number of training data. The building time complexity using DT, NB, LR, and SIURUA is O ( n m d ) , O ( n m c ) , O ( n m ) , and O ( m + n 3 ) , where n is the number of training data, m is the feature dimension, d is the depth of decision tree, and c is the number of categories.
The authentication time of each algorithm is very rapid. Similar to the building process, the authentication time complexity using SVM, DT, NB, LR, and SIURUA is O ( m ) ,   O ( d ) , O ( m c ) , O ( m ) , and O ( s m ) , respectively, where s is the number of support vectors.

4.2.4. Determine Appropriate Kernel

The proposed SIURUA algorithm is based on EasyMKL, and the kernel of the EasyMKL algorithm is to use different kernel functions and weights to fuse data. The kernel functions need to be specified in advance, and the weights are obtained in the learning process. Therefore, this experience is to decide the appropriate kernels that are used to map scene-irrelated features and user-related features. As we know, linear kernels and RBF kernels have four combinations:
  • Linear plus Linear (expressed as linear_linear) use linear kernels to map user-related features and linear kernels to map scene-irrelated features;
  • Linear plus RBF (expressed as linear_rbf) use linear kernels to map user-related features and RBF kernels to map scene-irrelated features;
  • RBF plus Linear (expressed as rbf_linear) use RBF kernels to map user-related features and Linear kernels to map scene-irrelated features;
  • RBF plus RBF (expressed as rbf_rbf) use RBF kernels to map user-related features and RBF kernels to map scene-irrelated features.
This experiment selects the optimal kernel combination from the above four kernel combinations. The aAccs of the four kernel combinations with each value of the time window are shown in Figure 9, the number of feature combinations is 200 plus 200, and the time window = 300 s. It can be seen that in the case of RBF plus RBF and Linear plus RBF, the highest aAccs are 84% and 82.5% obtained in 300 s. In the case of Linear plus Linear and RBF plus Linear, the highest aAccs are 79.7% and 80.9% obtained in 420 s. We can see that RBF plus RBF gains the maximum aAcc in the smallest value of the time window. Therefore, the appropriate kernel combination for SIURUA is RBF plus RBF when the time window = 300 s and the number of feature combinations is 200 plus 200.

4.2.5. Verify Feasibility of SIURUA

In order to further illustrate the feasibility of this algorithm, we use the features without feature selection and the features selected by mutual information as baselines. The number of features selected by SIURUA is 200 plus 200; in order to ensure the balance of the features, Mutual information needs to select the 400 features that have the highest correlation with users. Figure 10 shows the classification accuracy of different features with different values of the time window in hybrid scenes, where “All features” represents using whole feature sets to classify, SelectKBest (K = 400) indicates selecting 400 user-related features, and SVM with RBF kernel was selected as the classifier. We can see that the classification accuracy (75.2%) without feature selection is the worst. The classification accuracy (78.7%) with 400 user-related features has improved, and the SIURUA classification accuracy (84.0%) in the case of 200 plus 200 is the best. The 200 plus 200 features improve the accuracy of the model when using the same number of features as SelectKBest (K = 400), and this result proves that we successfully reduce the impacts of scene information on user-related features through fusing scene-irrelated features. Therefore, the result proves the feasibility of SIURUA.
For further proof, we employed this algorithm on four sets of single scene features, and the results are shown in Figure 11a–d. Figure 11a represents a typing scene, Figure 11b shows a Taobao scene, a Weibo scene is illustrated in Figure 11c, and a gaming scene is represented in Figure 11d. We can see that SIURUA can also improve the classification accuracy in the four single scenes. Although single-scene user authentication cannot be applied to the biometric system, the increasing accuracy can further verify the feasibility and versatility of SIURUA.
Combining the results of Figure 10 and Figure 11a–d, we can prove that fusing scene-irrelated features and user-related features can greatly improve the accuracy of the model, with the results verifying the feasibility of SIURUA.

4.2.6. Determine Fill Values for Empty Features

After determining all the parameters, we tested the time window = 300 s, the kernel combination rbf_rbf, and the feature combination 200_200 by selecting zero, the median, and the mean value to fill the empty features, as shown in Table 1. The experimental results are illustrated in Figure 12, which shows that the user authentication accuracy is highest for filling the empty features with zero and the lowest for filling with the median values. Therefore, we chose to fill the empty features in Table 1 with 0.

4.2.7. Proven Advantages of SIURUA

This section will compare SIURUA with some proposed keystroke dynamics and mouse dynamics algorithms. The selected algorithms are as follows:
  • MPCA [43]: Piantari et al. proposed a mouse dynamics user authentication method based on Principal Components Analysis (PCA) and SVM;
  • UIKDMM [44]: Panasiuk et al. proposed a multimodal biometric user authentication system based on keystroke dynamics and mouse movements, which authenticates users through K-Nearest Neighbor (KNN) and by fusing keystroke dynamics and mouse dynamics;
  • UAMKL [45]: Wang et al. proposed UAMKL, which is an AverageMKL-based keystroke dynamics and mouse dynamics fusion user authentication method;
  • TEM [46]: Chen et al. proposed a multimodal biometric user authentication system based on keystroke dynamics and mouse dynamics with Context Information. The user authentication model in the system is a comparison, which fuses the SVM based on keystroke dynamics features and the NB based on mouse dynamics features by using the majority voting mechanism.
Due to the particularity of a hybrid scene user keystroke dynamics and mouse dynamics dataset, we reproduced the above algorithms and experimented on multiple time window values. Figure 13 shows the authentication accuracy of SIURUA and the above four algorithms. It can be seen that the accuracy of SIURUA with all-time window values is better than that of some of the existing methods. The maximum aAcc of SIURUA is 84.0% at 300 s. UIKDMM and UAMKL achieve the maximum aAcc, 67.4%, and 77.3%, respectively, at 420 s. MPCA achieves the maximum aAcc 73.9% at 300 s. TEM achieves the maximum aAcc 72.2% at 480 s. Therefore, it is necessary to reduce the impacts of scene information in hybrid scenes, and SIURUA is superior to some of the existing algorithms.

4.2.8. Comprehensive Comparison of Above Experiments

This section will summarize the various algorithms mentioned in Section 4.2.2, Section 4.2.3, Section 4.2.4, Section 4.2.5, Section 4.2.6 and Section 4.2.7. According to Section 4.2.2, Section 4.2.3, Section 4.2.4, Section 4.2.5, Section 4.2.6 and Section 4.2.7, when the time window = 300 s, the feature selection number is 200 plus 200, and the kernel combination is RBF plus RBF, SIURUA achieves the best accuracy, which is 84.0%. As shown in Table 6, it can be seen that SIURUA has obtained the highest aAcc, aPrec, aTPR, and aF1, as well as the lowest aFPR (aFAR) and aFRR of all the methods. Table 6 shows that SIURUA, as proposed in this paper, has excellent performance. It can obtain 0.840 aAcc, 0.841 aPrec, 0.85 aTPR, 0.169 aFPR (aFAR), 0.15 aFRR, and 0.839 aF1 with the conditions of time window = 300 s, 200 plus 200 features, and an RBF plus RBF kernel combination.

5. Conclusions and Future work

We summarize our results and discuss future research directions in the following sections.

5.1. Summaries and Discussion

The purpose of biometric technology based on keystroke dynamics and mouse dynamics is to simulate users’ behaviors and to find the distinguishing factors determining users’ identities. This paper not only fuses keystroke dynamics features and mouse dynamics features but also proposes a method to select the scene-irrelated features and user-related features for the fusion experiment in hybrid scenes for the first time. In the experiments, we tested the SIURUA algorithm on the collected hybrid scene data to verify the necessity of fusing scene-irrelated features and considered the possibility of SIURUA as a means of user authentication. Through the elaboration of SIURUA and the results of a comparison with other existing algorithms, it was found that the authentication performance of SIURUA in hybrid scenes with noise is better than other user authentication algorithms based on keystroke dynamics and mouse dynamics.
These results are encouraging and indicate that the proposed hybrid scene feature selection method and fusing the selected scene-irrelated features and user-related features can effectively improve the performance of the user authentication system.

5.2. Future Work

Although this paper verifies the feasibility of user authentication with hybrid scene features, we only considered four known scenes without verifying the application effect in more scenes or even unknown scenes. In the future, more hybrid scene data needs to be collected to further restrain the impact of scene information on the data. On the other hand, we do not consider the correlation between the scenes of the hybrid scene used in this paper. In addition, the feature selection method of hybrid scenes can be extended to be used not only in user authentication but also in other applications.

Author Contributions

Conceptualization, X.W. and Y.S.; methodology, X.W. and Y.S.; software, Y.S., Y.Z. and W.H.; validation, Y.S.; formal analysis, Y.S.; investigation, Y.S.; resources, Y.S.; data curation, Y.S. and S.C.; writing—original draft preparation, Y.S.; writing—review and editing, Y.S.; visualization, Y.S.; supervision, X.W.; project administration, X.W.; funding acquisition, X.W. and K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key R&D Program of China, grant number 2017YFB0802803 and Beijing Natural Science Foundation, grant number 4202002.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gaines, R.S.; Lisowski, W.; Press, S.J.; Shapiro, N. Authentication by Keystroke Timing: Some Preliminary Results; RAND Corporation: Santa Monica, CA, USA, 1980. [Google Scholar]
  2. Ahmed, A.A.E.; Traore, I. A New Biometric Technology Based on Mouse Dynamics. IEEE Trans. Dependable Secur. Comput. 2007, 4, 165–179. [Google Scholar] [CrossRef]
  3. Ahmed, A.A.E.; Traore, I. Anomaly Intrusion Detection Based on Biometrics. In Proceedings of the Sixth Annual IEEE Systems, Man and Cybernetics (SMC) Information Assurance Workshop, West Point, NY, USA, 15–17 June 2005; IEEE: West Point, NY, USA, 2005; pp. 452–453. [Google Scholar]
  4. Biometric personal authentication using keystroke dynamics: A review. Appl. Soft Comput. 2011, 11, 1565–1573. [CrossRef]
  5. Peng, J.; Choo, K.-K.R.; Ashman, H. User Profiling in Intrusion Detection: A Review. J. Netw. Comput. Appl. 2016, 72, 14–27. [Google Scholar] [CrossRef]
  6. Cai, Z.; Shen, C.; Guan, X. Mitigating Behavioral Variability for Mouse Dynamics: A Dimensionality-Reduction-Based Approach. IEEE Trans. Hum.-Mach. Syst. 2014, 44, 244–255. [Google Scholar] [CrossRef]
  7. Quafafou, M.; Boussouf, M. Induction of Strong Feature Subsets. In Principles of Data Mining and Knowledge Discovery; Komorowski, J., Zytkow, J., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 1997; Volume 1263, pp. 384–392. ISBN 978-3-540-63223-8. [Google Scholar]
  8. Peng, H.; Long, F.; Ding, C. Feature Selection Based on Mutual Information Criteria of Max-Dependency, Max-Relevance, and Min-Redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  9. Grefenstette, J.J. Genetic Algorithms and Machine Learning. In Proceedings of the Sixth Annual Conference on Computational Learning Theory, New York, NY, USA, 26–28 July 1993; pp. 3–4. [Google Scholar] [CrossRef]
  10. Kohavi, R.; John, G.H. Wrappers for Feature Subset Selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef]
  11. Yang, Y.; Ma, Z.; Hauptmann, A.G.; Sebe, N. Feature Selection for Multimedia Analysis by Sharing Information Among Multiple Tasks. IEEE Trans. Multimedia 2013, 15, 661–669. [Google Scholar] [CrossRef]
  12. Tibshirani, R. Regression Shrinkage and Selection Via the Lasso. J. R. Stat. Soc. Ser. B (Methodol.) 1996, 58, 267–288. [Google Scholar] [CrossRef]
  13. Hoerl, E.; Kennard, W. Ridge Regression: Biased Estimation for Nonorthogonal Problems. Technometrics 2000, 42, 7. [Google Scholar] [CrossRef]
  14. Nie, F.; Cai, X.; Huang, H.; Ding, C. Efficient and Robust Feature Selection via Joint 2,1-Norms Minimization. In Proceedings of the 23rd International Conference on Neural Information Processing Systems, Vancouver British Columbia, Canada, 6–9 December 2010; pp. 1813–1821. [Google Scholar]
  15. Pisani, P.H.; Lorena, A.C. A Systematic Review on Keystroke Dynamics. J. Braz. Comput. Soc. 2013, 19, 573–587. [Google Scholar] [CrossRef]
  16. Azevedo, G.L.F.B.G.; Cavalcanti, G.D.C.; Carvalho Filho, E.C.B. An Approach to Feature Selection for Keystroke Dynamics Systems Based on PSO and Feature Weighting. In Proceedings of the 2007 IEEE Congress on Evolutionary Computation, Singapore, 25–28 September 2007; pp. 3577–3584. [Google Scholar]
  17. Alsultan, A.; Warwick, K.; Wei, H. Improving the Performance of Free-Text Keystroke Dynamics Authentication by Fusion. Appl. Soft Comput. 2018, 70, 1024–1033. [Google Scholar] [CrossRef]
  18. Epp, C.; Lippold, M.; Mandryk, R.L. Identifying Emotional States Using Keystroke Dynamics. In Proceedings of the Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Vancouver, BC, Canada, 7–12 May 2011; ACM: Vancouver, BC, Canada; pp. 715–724. [Google Scholar]
  19. Antal, M.; Szabó, L.Z.; László, I. Keystroke Dynamics on Android Platform. Procedia Technol. 2015, 19, 820–826. [Google Scholar] [CrossRef]
  20. Tsai, C.-J.; Huang, P.-H. Keyword-Based Approach for Recognizing Fraudulent Messages by Keystroke Dynamics. Pattern Recognit. 2020, 98, 107067. [Google Scholar] [CrossRef]
  21. Ayotte, B.; Banavar, M.; Hou, D.; Schuckers, S. Fast Free-Text Authentication via Instance-Based Keystroke Dynamics. IEEE Trans. Biom. Behav. Identity Sci. 2020, 2, 377–387. [Google Scholar] [CrossRef]
  22. Tewari, A.; Verma, P. An Improved User Identification Based on Keystroke-Dynamics and Transfer Learning. WEB 2022, 19, 5369–5387. [Google Scholar] [CrossRef]
  23. Lu, X.; Zhang, S.; Hui, P.; Lio, P. Continuous Authentication by Free-Text Keystroke Based on CNN and RNN. Comput. Secur. 2020, 96, 101861. [Google Scholar] [CrossRef]
  24. Higashino, J. Signature Verification System on Neuro-Computer. In Proceedings of the 11th IAPR International Conference on Pattern Recognition. Vol. IV. Conference D: Architectures for Vision and Pattern Recognition, The Hague, Netherlands, 30 August–1 September 1992; pp. 517–521. [Google Scholar]
  25. Everitt, R.A.J.; McOwan, P.W. Java-Based Internet Biometric Authentication System. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1166–1172. [Google Scholar] [CrossRef]
  26. Feher, C.; Elovici, Y.; Moskovitch, R.; Rokach, L.; Schclar, A. User Identity Verification via Mouse Dynamics. Inf. Sci. 2012, 201, 19–36. [Google Scholar] [CrossRef]
  27. Kasprowski, P.; Harezlak, K. Fusion of Eye Movement and Mouse Dynamics for Reliable Behavioral Biometrics. Pattern Anal. Appl. 2018, 21, 91–103. [Google Scholar] [CrossRef] [Green Version]
  28. Gao, L.; Lian, Y.; Yang, H.; Xin, R.; Yu, Z.; Chen, W.; Liu, W.; Zhang, Y.; Zhu, Y.; Xu, S.; et al. Continuous Authentication of Mouse Dynamics Based on Decision Level Fusion. In Proceedings of the 2020 International Wireless Communications and Mobile Computing (IWCMC), Limassol, Cyprus, 15–19 June 2020; pp. 210–214. [Google Scholar]
  29. Antal, M.; Fejer, N.; Buza, K. SapiMouse: Mouse Dynamics-Based User Authentication Using Deep Feature Learning. In Proceedings of the 2021 IEEE 15th International Symposium on Applied Computational Intelligence and Informatics (SACI), Timisoara, Romania, 19–21 May 2021; pp. 61–66. [Google Scholar]
  30. Hu, T.; Niu, W.; Zhang, X.; Liu, X.; Lu, J.; Liu, Y. An Insider Threat Detection Approach Based on Mouse Dynamics and Deep Learning. Secur. Commun. Netw. 2019, 2019, 1–12. [Google Scholar] [CrossRef]
  31. Bailey, K.O.; Okolica, J.S.; Peterson, G.L. User Identification and Authentication Using Multi-Modal Behavioral Biometrics. Comput. Secur. 2014, 43, 77–89. [Google Scholar] [CrossRef]
  32. Mondal, S.; Bours, P. A Study on Continuous Authentication Using a Combination of Keystroke and Mouse Biometrics. Neurocomputing 2016, 230, 1–22. [Google Scholar] [CrossRef]
  33. Earl, S.; Campbell, J.; Buckley, O. Identifying Soft Biometric Features from a Combination of Keystroke and Mouse Dynamics. In Proceedings of the Advances in Human Factors in Robots, Unmanned Systems and Cybersecurity, Washington DC, USA, 25–29 July 2021; Zallio, M., Raymundo Ibañez, C., Hernandez, J.H., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 184–190. [Google Scholar]
  34. Boser, B.E.; Guyon, I.M.; Vapnik, V.N. A Training Algorithm for Optimal Margin Classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory-COLT ’92, Pittsburgh, PA, USA, 27–29 July 1992; ACM Press: Pittsburgh, PA, USA, 1992; pp. 144–152. [Google Scholar]
  35. Belanche, L.A.; Tosi, A. Averaging of Kernel Functions. Neurocomputing 2013, 112, 19–25. [Google Scholar] [CrossRef]
  36. Kloft, M.; Brefeld, U.; Sonnenburg, S.; Zien, A. Non-Sparse Regularization and Efficient Training with Multiple Kernels; EECS Department, University of California: Berkeley, CA, USA, 2010. [Google Scholar]
  37. Xu, Z.; Jin, R.; Yang, H.; King, I.; Lyu, M.R. Simple and Efficient Multiple Kernel Learning by Group Lasso. In Proceedings of the 27th International Conference on International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 1175–1182. [Google Scholar]
  38. Do, H.; Kalousis, A.; Woznica, A.; Hilario, M. Margin and Radius Based Multiple Kernel Learning. In Machine Learning and Knowledge Discovery in Databases; Buntine, W., Grobelnik, M., Mladenić, D., Shawe-Taylor, J., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2009; Volume 5781, pp. 330–343. ISBN 978-3-642-04179-2. [Google Scholar]
  39. Aiolli, F.; Donini, M. EasyMKL: A Scalable Multiple Kernel Learning Algorithm. Neurocomputing 2015, 169, 215–224. [Google Scholar] [CrossRef]
  40. Teh, P.S.; Teoh, A.B.J.; Ong, T.S.; Neo, H.F. Statistical Fusion Approach on Keystroke Dynamics. In Proceedings of the 2007 Third International IEEE Conference on Signal-Image Technologies and Internet-Based System, Shanghai, China, 16–18 December 2022; IEEE: Shanghai, China, 2007; pp. 918–923. [Google Scholar]
  41. Tsimperidis, I.; Yucel, C.; Katos, V. Age and Gender as Cyber Attribution Features in Keystroke Dynamic-Based User Classification Processes. Electronics 2021, 10, 835. [Google Scholar] [CrossRef]
  42. Vinh, N.X.; Epps, J.; Bailey, J. Information Theoretic Measures for Clusterings Comparison: Is a Correction for Chance Necessary? In Proceedings of the 26th Annual International Conference on Machine Learning-ICML ’09, Montreal, QC, Canada, 14–18 June 2009; ACM Press: Montreal, QC, Canada, 2009; pp. 1–8. [Google Scholar]
  43. Munir; Piantari, E.; Firman, F.N. Dimensional Reduction in Behavioral Biometrics Authentication System. In Proceedings of the ICETIT 2019, Delhi, India, 21–22 June 2019; Singh, P.K., Panigrahi, B.K., Suryadevara, N.K., Sharma, S.K., Singh, A.P., Eds.; Lecture Notes in Electrical Engineering. Springer International Publishing: Cham, Switzerland, 2020; Volume 605, pp. 984–992, ISBN 978-3-030-30576-5. [Google Scholar]
  44. Panasiuk, P.; Szymkowski, M.; Dąbrowski, M.; Saeed, K. A Multimodal Biometric User Identification System Based on Keystroke Dynamics and Mouse Movements. In Computer Information Systems and Industrial Management; Saeed, K., Homenda, W., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2016; Volume 9842, pp. 672–681. ISBN 978-3-319-45377-4. [Google Scholar]
  45. Wang, X.; Zheng, Q.; Zheng, K.; Wu, T. User Authentication Method Based on MKL for Keystroke and Mouse Behavioral Feature Fusion. Secur. Commun. Netw. 2020, 2020, 1–14. [Google Scholar] [CrossRef]
  46. Chen, L.; Zhong, Y.; Ai, W.; Zhang, D. Continuous Authentication Based on User Interaction Behavior. In Proceedings of the 2019 7th International Symposium on Digital Forensics and Security (ISDFS), Barcelos, Portugal, 10–12 June 2019; IEEE: Barcelos, Portugal, 2019; pp. 1–6. [Google Scholar]
Figure 1. Block diagram of SIURUA.
Figure 1. Block diagram of SIURUA.
Sensors 22 06627 g001
Figure 2. Keystroke dynamics.
Figure 2. Keystroke dynamics.
Sensors 22 06627 g002
Figure 3. Eight directions for mouse dynamics feature extraction.
Figure 3. Eight directions for mouse dynamics feature extraction.
Sensors 22 06627 g003
Figure 4. Example of collected keystroke data.
Figure 4. Example of collected keystroke data.
Sensors 22 06627 g004
Figure 5. Example of collected mouse data.
Figure 5. Example of collected mouse data.
Sensors 22 06627 g005
Figure 6. Comparing the user authentication accuracy of hybrid scenes and single scenes.
Figure 6. Comparing the user authentication accuracy of hybrid scenes and single scenes.
Sensors 22 06627 g006
Figure 7. Accuracy of SIURUA in each feature combination.
Figure 7. Accuracy of SIURUA in each feature combination.
Sensors 22 06627 g007
Figure 8. The accuracy of each algorithm in different value of time windows.
Figure 8. The accuracy of each algorithm in different value of time windows.
Sensors 22 06627 g008
Figure 9. Comparison of the accuracy of SIURUA with different kernel functions.
Figure 9. Comparison of the accuracy of SIURUA with different kernel functions.
Sensors 22 06627 g009
Figure 10. Comparison of accuracy between SIURUA and selecting 400 best features in a hybrid scene.
Figure 10. Comparison of accuracy between SIURUA and selecting 400 best features in a hybrid scene.
Sensors 22 06627 g010
Figure 11. (a) Comparison of accuracy between SIURUA and the selection of 400 best features in a typing scene; (b) comparison of accuracy between SIURUA and the selection of 400 best features in a Taobao scene; (c) comparison of accuracy between SIURUA and the selection of 400 best features in a Weibo scene; (d) comparison of accuracy between SIURUA and the selection of 400 best features in a gaming scene.
Figure 11. (a) Comparison of accuracy between SIURUA and the selection of 400 best features in a typing scene; (b) comparison of accuracy between SIURUA and the selection of 400 best features in a Taobao scene; (c) comparison of accuracy between SIURUA and the selection of 400 best features in a Weibo scene; (d) comparison of accuracy between SIURUA and the selection of 400 best features in a gaming scene.
Sensors 22 06627 g011
Figure 12. Comparison of accuracy between different fill values for empty features.
Figure 12. Comparison of accuracy between different fill values for empty features.
Sensors 22 06627 g012
Figure 13. Comparison of accuracy between SIURUA and the proposed user authentication method.
Figure 13. Comparison of accuracy between SIURUA and the proposed user authentication method.
Sensors 22 06627 g013
Table 1. Examples of keystroke mouse dynamics features.
Table 1. Examples of keystroke mouse dynamics features.
Feature f 1 f 2 f 3 f j
K M F e a t u r e 1 f 1 , 1 f 1 , 2 f 1 , 3 f 1 , j
K M F e a t u r e 2 f 2 , 1 f 2 , 2 f 2 , 3 f 2 , j
K M F e a t u r e 3 f 3 , 1 f 3 , 2 f 3 , 3 f 3 , j
K M F e a t u r e i f i , 1 f i , 2 f i , 3 f i , j
Table 2. Examples features of calculating AMI.
Table 2. Examples features of calculating AMI.
Feature F e a t u r e 1 F e a t u r e 2 F e a t u r e 3 F e a t u r e 4 F e a t u r e 5
f 1 384 485 433 498 458
S c e n e 0 3 0 1 3
Table 3. Contingency table of f j and U s e r .
Table 3. Contingency table of f j and U s e r .
f j \ U s e r   U 1   U 2   U 41   Sums
F 1 , j n u 1 , 1 n u 1 , 2 n u 1 , 41 a 1 , j
F 2 , j n u 2 , 1 n u 2 , 2 n u 2 , 41 a 2 , j
F M j , j n u M j , 1 n u M j , 2 n u M j , 41 a M j , j
Sums b 1 b 2 b 41 k l n u k l = N U
Table 4. Contingency table of f j and S c e n e .
Table 4. Contingency table of f j and S c e n e .
f j \ S c e n e   S 1   S 2   S 4   Sums
F 1 , j n s 1 , 1 n s 1 , 2 n s 1 , 41 a 1 , j
F 2 , j n s 2 , 1 n s 2 , 2 n s 2 , 41 a 2 , j
F M j , j n s M j , 1 n s M j , 2 n s M j , 41 a M j , j
Sums c 1 c 2 c 4 k l n s k l = N S
Table 5. Time complexity comparison of machine learning algorithms and SIURUA.
Table 5. Time complexity comparison of machine learning algorithms and SIURUA.
AlgorithmSVMDTNBLRSIURUA
Building O ( n 3 ) O ( n m d ) O ( n m c ) O ( n m ) O ( m + n 3 )
Authentication O ( m ) O ( d ) O ( m c ) O ( m ) O ( s m )
Table 6. Comprehensive comparison of the above methods.
Table 6. Comprehensive comparison of the above methods.
AlgorithmaAccaPrecaTPRaFPR(aFAR)aFRR
SVM0.7520.7540.7940.2900.206
DT0.7120.7140.7160.2910.284
NB0.5980.6320.5610.3520.439
LR0.7580.7420.7900.2820.210
SVM_4000.7660.7620.7980.2660.202
MPCA [43]0.7390.7350.7700.2900.230
UIKDMM [44]0.6740.6740.7050.3350.295
UAMKL [45]0.7730.7790.7880.2430.212
TEM [46]0.7220.7130.7820.3330.217
SIURUA0.8400.8410.8500.1690.150
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, X.; Shi, Y.; Zheng, K.; Zhang, Y.; Hong, W.; Cao, S. User Authentication Method Based on Keystroke Dynamics and Mouse Dynamics with Scene-Irrelated Features in Hybrid Scenes. Sensors 2022, 22, 6627. https://doi.org/10.3390/s22176627

AMA Style

Wang X, Shi Y, Zheng K, Zhang Y, Hong W, Cao S. User Authentication Method Based on Keystroke Dynamics and Mouse Dynamics with Scene-Irrelated Features in Hybrid Scenes. Sensors. 2022; 22(17):6627. https://doi.org/10.3390/s22176627

Chicago/Turabian Style

Wang, Xiujuan, Yutong Shi, Kangfeng Zheng, Yuyang Zhang, Weijie Hong, and Siwei Cao. 2022. "User Authentication Method Based on Keystroke Dynamics and Mouse Dynamics with Scene-Irrelated Features in Hybrid Scenes" Sensors 22, no. 17: 6627. https://doi.org/10.3390/s22176627

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop