Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Non-Invasive Determination of the Mass Flow Rate for Particulate Solids Using Microwaves
Previous Article in Journal
Efficient Filter Design to Compensate Fabrication Imperfections in 6G Communication Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Supervised Manifold Learning Based on Multi-Feature Information Discriminative Fusion within an Adaptive Nearest Neighbor Strategy Applied to Rolling Bearing Fault Diagnosis

1
Center of Shanxi Engineering Research for Coal Mine Intelligent Equipment, Taiyuan University of Technology, Taiyuan 030024, China
2
College of Mechanical and Vehicle Engineering, Taiyuan University of Technology, Taiyuan 030024, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(24), 9820; https://doi.org/10.3390/s23249820
Submission received: 22 November 2023 / Revised: 10 December 2023 / Accepted: 12 December 2023 / Published: 14 December 2023
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
Rolling bearings are a key component for ensuring the safe and smooth operation of rotating machinery and are very prone to failure. Therefore, intelligent fault diagnosis research on rolling bearings has become a crucial task in the field of mechanical fault diagnosis. This paper proposes research on the fault diagnosis of rolling bearings based on an adaptive nearest neighbor strategy and the discriminative fusion of multi-feature information using supervised manifold learning (AN-MFIDFS-Isomap). Firstly, an adaptive nearest neighbor strategy is proposed using the Euclidean distance and cosine similarity to optimize the selection of neighboring points. Secondly, three feature space transformation and feature information extraction methods are proposed, among which an innovative exponential linear kernel function is introduced to provide new feature information descriptions for the data, enhancing feature sensitivity. Finally, under the adaptive nearest neighbor strategy, a novel AN-MFIDFS-Isomap algorithm is proposed for rolling bearing fault diagnosis by fusing various feature information and classifiers through discriminative fusion with label information. The proposed AN-MFIDFS-Isomap algorithm is validated on the CWRU open dataset and our experimental dataset. The experiments show that the proposed method outperforms other traditional manifold learning methods in terms of data clustering and fault diagnosis.

1. Introduction

With the rapid development of industrial intelligence, intelligent fault diagnosis technology is playing an increasingly important role in maintaining the health of mechanical equipment and ensuring its safe and stable operation [1]. Rolling bearings, as critical components of rotating machinery, often operate in harsh and variable environments, such as at high speed and under heavy loads, making them prone to wear and failure and leading to severe mechanical accidents [2]. To avoid significant economic losses and casualties, research on intelligent fault diagnosis methods for rolling bearings is particularly important [3]. Fault diagnosis is fundamentally a pattern recognition problem, and analyzing and processing vibration signals during the operation of rolling bearings is an effective approach for diagnosing faults in rotating machinery. One of the most crucial aims is to break the “curse of dimensionality” and extract low-dimensional features with high sensitivity [4,5,6].
To enhance the quality of feature extraction and address the severe “curse of dimensionality” issue at the current stage, manifold learning algorithms have emerged. Classical manifold learning algorithms mainly include Principal Component Analysis (PCA), Independent Component Analysis (ICA), Linear Discriminant Analysis (LDA), and Multidimensional Scaling (MDS). However, these algorithms are primarily used for linear dimensionality reduction and may not be suitable for high-dimensional nonlinear vibration data from rolling bearings [7,8].
In the year 2000, Joshua B. Tenenbaum and Sam T. Roweis [9,10] proposed two classic nonlinear manifold learning dimensionality reduction algorithms, Isometric Feature Mapping (Isomap) and Locally Linear Embedding (LLE), in science. Since then, manifold learning algorithms have been extensively researched by researchers and have gradually become a research hotspot in the fields of dimensionality reduction and pattern recognition [9,10,11]. Based on their different mathematical assumptions, manifold learning algorithms are divided into two major categories: locally preserving embedding methods and globally preserving embedding methods. Laplacian Eigenmaps (LE) [12], LLE [10], Hessian-based Locally Linear Embedding (HLLE) [13], and Local Tangent Space Alignment (LTSA) [14] are considered local preserving embedding methods in manifold learning, while Isomap [8], Diffusion Maps (DM) [2], and t-Stochastic Neighbor Embedding (t-SNE) [15] are regarded as globally preserving embedding methods. However, regardless of the manifold learning algorithm’s mathematical assumptions, there is a bottleneck in selecting neighboring points [6,11].
To address the sensitivity of neighbor point selection in manifold learning algorithms, Zhenyue Zhang et al. conducted research from two perspectives: adaptive neighbor selection and the interaction between manifold curvature and sampling density [16]. They explored methods for constructing nonlinear low-dimensional manifolds from high-dimensional space samples, providing directions for subsequent researchers. Chuang Sun et al. conducted research from the perspective of adaptive neighbors and used the kernel sparse representation method to select sample neighbors and reconstruct the weights of the neighbor graph for the LLE algorithm [17]. Yan Zhang et al. integrated nonnegative matrix factorization with sparsity constraints based on the work in reference [17] and applied it to the LLE algorithm to jointly minimize the neighborhood reconstruction error on the weight matrix [18]. All of these methods use sparsity constraints to select neighbor points, but they perform relatively averagely when the data contain noise points and outliers.
To address this issue, Yunlong Gao et al. proposed a discriminant analysis based on the reliability of local neighborhoods, enhancing the performance of effective samples in low-dimensional space and filtering the interference of outliers, thereby improving the dimensionality reduction ability [19]. Jing An et al. introduced an adaptive neighborhood-preserving discriminant projection model [20]. By updating sparse reconstruction coefficients, the adverse effects of noise and outliers on the dimensionality reduction were mitigated, enhancing sample clustering. Jiaqi Xue et al. proposed a locally linear embedding method by applying an adaptive neighbor strategy, preserving more original information when embedding high-dimensional data manifolds into low-dimensional space and achieving better clustering results [11]. It can be observed that various discrimination methods have been widely applied in manifold learning models. Most of these adaptive neighbor strategies and discrimination methods are applied in locally preserving embedding methods of manifold learning algorithms, while their application in globally preserving embedding manifold learning algorithms is limited, especially in unsupervised learning models.
To address the aforementioned issue, incorporating label information into the algorithm’s supervised learning mode can further enhance its clustering capability. Ratthachat Chatpatanasiri et al. proposed a general framework for manifold learning semi-supervised dimensionality reduction, providing research directions for subsequent researchers [21]. Jing Wang et al. proposed a semi-supervised manifold alignment algorithm that utilizes sample points and their corresponding relationships to construct connections between different manifolds [22]. Zohre Karimi et al. introduced a novel hierarchical spatial semi-supervised metric learning approach, integrating local constraints and information-theoretic nonlocal constraints to better represent the smoothness assumption of multiple manifolds using the metric matrix [23]. Mingxia Chen et al. proposed a robust semi-supervised manifold learning framework applied in locally preserving embedding manifold learning algorithms to eliminate adverse effects caused by noise points [24]. Ding Li et al. derived an extension of a semi-supervised manifold regularization algorithm for classification tasks, optimizing the algorithm’s performance on multi-class problems using weighted strategies [25]. Jun Ma et al. proposed a secure semi-supervised learning framework, using both manifold and discriminant regularization to mitigate the influence of unlabeled points and boundary points in the pattern recognition process [26]. However, the impact of unlabeled points and boundary points in the semi-supervised learning mode on the model’s clustering and classification capabilities remains unresolved.
Therefore, researchers have applied supervised learning modes to manifold learning models, which, compared to manifold learning models under the semi-supervised learning mode, demonstrate stronger robustness in handling classification problems [27,28,29]. However, current research methods are limited to dimensionality reduction and fault diagnosis tasks on a single feature space within the manifold learning model. The feature information in the data is singular and incomprehensive.
In summary, manifold learning methods have been widely applied in the fields of dimensionality reduction and fault diagnosis. However, they still have limitations. The issues of neighbor point selection in manifold global preservation embedding, the influence of data outliers on clustering effectiveness, and the singular and incomplete feature information contained in the data have not been fully addressed. To address these problems and build upon existing research, this paper proposes a supervised manifold learning approach for rolling bearing fault diagnosis based on the discriminative fusion of multiple pieces of feature information using an adaptive nearest neighbor strategy.
The main contributions of this paper are summarized as follows:
  • Propose an adaptive neighbor selection strategy that amalgamates the Euclidean distance and cosine similarity measures. This strategy systematically computes both the distance and angular information among neighboring points, utilizing the metric mean as the discriminant criterion. By configuring the preset neighboring points as the criterion object, it dynamically adjusts the proximity graph to refine the local structure of the manifold. This process is aimed at enhancing the precision of the manifold space depiction and local feature representation and reducing the adverse effects of data outliers on clustering performance.
  • Propose three methods for transforming feature spaces and extracting spatial feature information and space information. Notably, this paper proposes a unique form of the kernel function, the exponential linear kernel function, which serves to project data into a novel kernel Hilbert space. Concurrently, this function is employed as the nonlinear discriminant mapping function in the Supervised Version of the Isometric Feature Mapping (S-Isomap) algorithm, thus providing a distinct representation of data in the manifold space. The extracted feature information, originating from diverse kernel Hilbert spaces and manifold spaces, ensures the intricate and sensitive nature of the features.
  • Propose a fault diagnosis algorithm model for rolling bearings by employing a supervised learning paradigm. Under the adaptive neighbor selection strategy, features from different spaces are merged which are both sensitive and complex to form a multi-space metric matrix. This matrix is designed to encapsulate substantial multi-space feature information, enabling its fusion with machine learning classifier algorithms to facilitate fault diagnosis.
The structure of this paper is organized as follows: Section 2 introduces the foundational manifold learning algorithms Isomap and S-Isomap along with their relevant theories. Section 3 presents the proposed supervised manifold learning method involving an adaptive neighbor strategy, the extraction of multi-space feature information, and the discriminative fusion of multiple pieces of feature information. Section 4 conducts an evaluation of the model’s clustering and classification capabilities, analyzing and comparing the proposed approach in this paper with traditional manifold learning methods from both qualitative and quantitative perspectives. Finally, Section 5 provides a comprehensive summary of the entire paper.

2. Related Work

The core idea of manifold learning is based on the manifold assumption, which posits that data are distributed on a smooth low-dimensional manifold embedded in a high-dimensional space. Traditional manifold learning algorithms such as Isomap, LLE, and LTSA aim to find the embedded low-dimensional manifold within a high-dimensional space [30]. This can be described mathematically as finding the mapping process, f , such that f :   X = x 1 , x 2 , , x N T N × D Y = y 1 , y 2 , , y N T N × d d D , where x i denotes the sample in a high-dimensional space, X; y i is the mapping of sample x i in a low-dimensional space, Y; N is the number of data points; D is the number of high-dimensional features; and d is the number of low-dimensional features.

2.1. Isometric Feature Mapping (Isomap)

Isomap, as one of the most traditional manifold learning algorithms, operates on the principle of preserving the global geometric properties of the intrinsic low-dimensional manifold to obtain a low-dimensional representation of the data. Isomap modifies the measurement method used in MDS, where the Euclidean distance describes the relationship between two data points, to a method based on the geodesic distance on the manifold [9]. The Isomap algorithm is as follows:
(1)
Calculate the Euclidean distance, d E x i , x j , between any two data points. Then, use k-nearest neighbors (k-NNs) from the number information of samples or ε-nearest neighbors (ε-NNs) from the distance information of samples to construct a simple undirected nearest neighbor graph, G. If x i and x j are neighbors, then connect x i and x j in G and assign a weight, d E x i , x j , to the edge.
(2)
Based on the edges of the simple undirected nearest neighbor graph, G, use Dijkstra’s algorithm or Floyd’s algorithm to calculate the geodesic distances, d G x i , x j .
(3)
To establish the low-dimensional embedded manifold coordinates, Y, based on an objective function, the typical objective function for Isomap can be expressed as follows:
L I s o = min Y i , j d G x i , x j d E y i , y j 2
where d E y i , y j represents the Euclidean distance of the mapping of the Euclidean distance, d E x i , x j , in a high-dimension space.
(4)
To compute the low-dimensional embedded manifold coordinates, Y, use MDS. To be specific, let D G 2 = d G 2 x i , x j and set τ D G 2 = H D G 2 H / 2 , where H = I e e T / N is the centered matrix and e is the column vector of all the matrices. Perform an eigenvalue decomposition on τ D G 2 to obtain the low-dimensional embedded manifold coordinates Y = y 1 , y 2 , , y N T = λ 1 u 1 , λ 2 u 2 , , λ d u d T , where λ p and u p denote the p t h dominant eigenvalue and its eigenvector, respectively [30].

2.2. Supervised Version of Isometric Feature Mapping (S-Isomap)

The S-Isomap algorithm incorporates label information from the data as prior knowledge to guide the dimensionality reduction process. Utilizing data labels as discriminant information, the data are initially divided into a true nearest neighbor set, S + , and a pseudo nearest neighbor set, S , which guides the discriminant manifold learning process. S + and S are defined as follows:
S + = x i , x j L x i = L x j , x j N x i
S = x i , x j L x i L x j , x j N x i
where L x i denotes the class label of x i and N x i are the neighborhood sets of x i . Next, based on S + and S , we construct the true nearest neighbor graph, G + , and pseudo nearest neighbor graph, G . Then, the objective of the S-Isomap algorithm is to preserve the intrinsic geometric structure of the data within the same class and to separate different classes by optimizing the following objective function:
L S I s o S + = min Y ( x i , x j ) S + d E y i , y j d G x i , x j 2
L S I s o S = max Y ( x i , x j ) S d E y i , y j d G x i , x j 2
To find the optimal solution for the objective function mentioned earlier, the S-Isomap algorithm rescales the metric (Euclidean distance) between two data points as follows:
D E x i , x j = 1 e d E 2 x i , x j β L x i = L x j e d E 2 x i , x j β α L x i L x j
where β is a parameter with relation to the scale of a dataset and α is a parameter which can adjust the intra-class dissimilarity sensitively. The two values α and β are usually determined empirically. Finally, we compute the low-dimensional embedded manifold coordinates, Y, using MDS [30].

3. AN-MFIDFS-Isomap Reduction

3.1. Adaptive Nearest Neighbor Strategy

The value of the nearest neighbor count, k, serves as a hyperparameter in manifold learning algorithms, determining the size of local regions within the manifold. A larger k value may smooth or eliminate small-scale structures within the manifold. Conversely, a smaller k value might incorrectly partition a continuous manifold into disjoint submanifolds, thereby affecting the accuracy of the algorithm in approximating the global geometric structure and computing the metric matrix during the process. The appropriate choice of k is crucial to balance between capturing local details and preserving the overall manifold structure accurately.
In this study, we introduce an adaptive nearest neighbor strategy by integrating both distance and angular information among data points. This strategy incorporates the concepts of initial nearest neighbors, pseudo nearest neighbors, and true nearest neighbors. By combining the data distance and angular information, we calculate the edge weights between data points, which are used to construct a simple undirected nearest neighbor graph for true nearest neighbors. The adaptive nearest neighbor strategy can effectively suppress the adverse effect of data outliers on clustering performance. The algorithm’s workflow is illustrated in Figure 1, and detailed steps of the algorithm are outlined as follows:
(1)
Define the initial nearest neighbor count, k, and construct a simple undirected nearest neighbor graph, G;
(2)
Construct the cosine similarity matrix for the data. Compute the cosine similarity between two data points using the following equation to obtain angular information:
S i j cos ˜ = s i j cos ˜ N × k = x i , x j x i 2 · x j 2 x j N x i
where · , · denotes the vector’s inner product, · 2 denotes the L2-norm of the vector, and · N × k denotes the shape of the matrix, which is N × k .
(3)
Compute the normalized cosine similarity matrix by normalizing the angular information using the following equation:
S i j cos = s i j cos N × k = log 2 1.5 0.5 s i j cos ˜
where 1 1.5 0.5 s i j cos ˜ 2 .
(4)
Construct the Euclidean distance matrix for the data. Compute the Euclidean distance between two data points using the following equation to obtain distance information:
D E ˜ = d i j E ˜ N × k = x i x j 2 x j N x i
(5)
Compute the normalized Euclidean distance matrix by normalizing the distance information using the following equation:
D E = d i j E N × k = d i j E ˜ min d i · E ˜ max d i · E ˜ min d i · E ˜
where d i · E ˜ represents all elements of row i, d E ˜ .
(6)
Construct the weight matrix by integrating the angular and distance information. Reassign the edge weights in the simple undirected nearest neighbor graph, G, using the following equation to fuse the angular and distance information:
W i j E cos = ω i j E cos N × k = s i j cos + d i j E
(7)
Compute the weight discrimination criterion. Calculate the average weight for each data point using the following equation to serve as the discrimination criterion:
W i j E cos ¯ = ω i j E cos ¯ N = 1 k + 1 j = 1 k ω i j E cos
If ω i j E cos ω i j E cos ¯ , then x j is considered a true nearest neighbor of x i . Conversely, if ω i j E cos > ω i j E cos ¯ , then x j is considered a pseudo nearest neighbor of x i . This process leads to dynamically obtaining the true nearest neighbor values, k + , a simple undirected nearest neighbor graph, G + for true nearest neighbors, and a new set of rules for calculating the metric matrix.
The adaptive neighbor strategy proposed in this section offers a novel approach for dynamically selecting neighboring points in manifold learning. Essentially, it seamlessly combines k-NNs and ε-NNs by integrating the angular and distance information to derive weights. These weights are compared with the average weight, ultimately resulting in the construction of a simple undirected nearest neighbor graph, G + , for true nearest neighbors.

3.2. Multi-Space Transformation and Feature Extraction

3.2.1. Kernel Trick

The data matrix in the original space, denoted as X = x 1 , x 2 , , x N T , consists of N observation vectors of the data [31]. The kernel trick is a mathematical technique that utilizes a kernel function to map the data from the original space to a higher-dimensional Hilbert space, as follows:
Φ : X H   x i ϕ x i
where ϕ · denotes the mapping of the kernel function.
The data matrix in the higher-dimensional Hilbert space is computed through the inner product of the observation vectors using the kernel function. This can be seen as the metric matrix in the higher-dimensional Hilbert space, as follows:
k x i , x j = ϕ x i · ϕ x j = ϕ x i T ϕ x j

3.2.2. The First Spatial Transformation and Feature Extraction

In this section, we use K-Isomap as the first spatial transformation method. This method considers the double centering process in Isomap as a kernel trick mapping process [32]. Therefore, the objective function is as follows:
L K I s o = min Y i , j k x i , x j d E y i , y j 2
To obtain the optimal solution for the objective function mentioned above, the problem is transformed into constructing a space metric matrix that better adheres to the manifold assumption. K-Isomap considers the geodesic distance matrix of the original data as the metric matrix during the double centering process. To ensure the semi-positive definiteness of the double-centered matrix, the Mercer kernel matrix method [33] is utilized to construct the first spatial metric matrix. The spectral radius of this matrix is computed to represent the first piece of spatial feature information.
Compared to the classical Isomap algorithm, the K-Isomap algorithm exhibits stronger robustness. The specific process of the algorithm is as follows:
(1)
Using the traditional Isomap algorithm, construct a simple undirected nearest neighbor graph, G. Then, calculate the geodesic distance, d G x i , x j , between data points x i and x j using the shortest path algorithm and define the matrix D G 2 = d G 2 x i , x j and τ D G 2 = H D G 2 H / 2 ;
(2)
Construct the following block matrix:
0 2 τ D G 2 I 4 τ D G
where 0 denotes the full-zero matrix and I denotes the unit matrix. Next, calculate its spectral radius, ρ 1 = max λ 1 , λ 2 , , λ 2 N , where λ i are all the eigenvalues of the above block matrix;
(3)
Construct the Mercer kernel matrix as the first spatial metric matrix using the following equation:
K 1 = τ D G 2 + 2 ρ 1 m τ D G + 1 2 ρ 1 m 2 H
where the parameter ρ 1 m ρ 1 * ;
(4)
Perform eigenvalue decomposition and select the top λ 1 eigenvalues’ corresponding to the eigenvectors Γ 1 as the first piece of spatial feature information for K 1 . Calculate the spectral radius ρ 1 of the first spatial metric matrix, K 1 , to represent the first piece spatial information.

3.2.3. The Second Spatial Transformation and Feature Extraction

In the current manifold learning algorithms, widely used kernel functions include the Gaussian kernel function and the linear kernel function. The Gaussian kernel function is suitable for addressing complex nonlinear problems; however, it is highly affected by hyperparameter noise. On the other hand, the linear kernel function is suitable for addressing issues related to stability and poor generalization capabilities but does not effectively solve complex nonlinear problems [34]. In this section, we propose an exponential linear kernel function that combines the advantages of the above two kernel functions. The expression for the exponential linear kernel function is as follows:
k E l i n e a r x i , x j = e ϖ d E x i , x j + b
where the parameter ϖ > 0 denotes the linear weight and the parameter b denotes the offset coefficient. A detailed proof of the positive semi-definite property of the constructed kernel matrix for the exponential linear kernel function is provided in Section 6, Appendix A.
In this section, we employ the exponential linear kernel function to construct the second space. We utilize the Mercer kernel matrix method to construct the metric matrix for the second space. The spectral radius of this matrix is then calculated to represent the second piece of spatial feature information. The specific steps of the algorithm are as follows:
(1)
Apply the traditional Isomap algorithm to construct a simple undirected nearest neighbor graph, G, and calculate the Euclidean distance, d E x i , x j ;
(2)
Construct the exponential linear kernel matrix as follows, based on Equation (17):
K E l i n e a r = k E l i n e a r x i , x j = e ϖ d E x i , x j + b
Then, perform an eigenvalue decomposition and select the top λ 2 eigenvalues’ corresponding to the eigenvectors Γ 2 as the second piece of spatial feature information for K E l i n e a r . Let the spectral radius ρ 2 represent the second piece of spatial information.

3.2.4. The Third Spatial Transformation and Feature Extraction

In this section, we propose a KS-Isomap manifold learning algorithm as the third space transformation method. The KS-Isomap algorithm utilizes data label information to guide the construction of the manifold and employs kernel tricks to map the metric matrix to a more abstract feature space. The KS-Isomap algorithm combines the robustness of the K-Isomap algorithm with the discriminative power of the S-Isomap algorithm.
The KS-Isomap algorithm follows the structure of the S-Isomap algorithm. It utilizes label information to guide the construction of the manifold. Simultaneously, it replaces the original metric matrix with the geodesic distance matrix and maps the geodesic distance matrix to the kernel space. Therefore, the objective function of the algorithm is represented by the following equation:
L K S I s o S + = min Y ( x i , x j ) S + d G y i , y j k x i , x j 2
L K S I s o S = max Y ( x i , x j ) S d G y i , y j k x i , x j 2
To obtain the optimal solution for the above objective function, the problem is transformed to construct a more sensitive discriminative distance matrix. The specific steps of the KS-Isomap algorithm are as follows:
(1)
Using the traditional Isomap algorithm, construct a simple undirected nearest neighbor graph, G. Then, calculate the geodesic distance, d G x i , x j , between data points x i and x j using the shortest path algorithm;
(2)
Construct the discriminative distance matrix based on Equations (6) and (18):
D H = d H x i , x j = 1 k E l i n e a r x i , x j f o r   L x i = L x j k E l i n e a r x i , x j φ f o r   L x i L x j
And define
D H 2 = 1 k E l i n e a r 2 x i , x j f o r   L x i = L x j k E l i n e a r 2 x i , x j φ f o r   L x i L x j
where the parameter φ denotes the bias coefficient;
(3)
Construct the following block matrix:
0 2 D H 2 I 4 D H
And calculate its spectral radius, ρ 3 ;
(4)
Construct the Mercer kernel matrix as the metric matrix for the third space based on the following equation:
K 3 = K D 2 + 2 ρ 3 m K D + 1 2 ρ 3 m 2 H
where the parameter ρ 3 m ρ 3 * ;
(5)
Perform an eigenvalue decomposition and select the top λ 3 eigenvalues’ corresponding to the eigenvectors Γ 3 as the third piece of spatial feature information for K 3 . Calculate the spectral radius ρ 3 of the first spatial metric matrix, K 3 , to represent the first piece of spatial information.

3.3. AN-MFIDFS-Isomap Fault Diagnosis Arithmetric

Currently, research on manifold learning algorithms is primarily confined to a high-dimensional Euclidean space or kernel Hilbert space, and the potential of leveraging feature information from multiple spaces corresponding to the same data has been overlooked. Therefore, this study, by utilizing the adaptive neighbor selection strategy proposed in Section 3.1 to obtain the metric matrix and employing the method proposed in Section 3.2 for feature space transformation and feature information extraction, presents a fault diagnosis algorithm called AN-MFIDFS-Isomap. The algorithm’s flowchart is illustrated in Figure 2, and the detailed steps of the algorithm are as follows:
(1)
Divide the data into a training set, X t r a i n , and a testing set, X t e s t , preprocess the data, and construct the corresponding label information for each set;
(2)
For the training set, utilize the adaptive nearest neighbor strategy proposed in Section 3.1. Compute the geodesic distances under the rule of true nearest neighbors for the simple undirected neighbor graph, G + , constructing the geodesic distance matrix, i.e., the metric matrix in the original space;
(3)
Employ the first spatial transformation method proposed in Section 3.2.2 to obtain the first piece of spatial feature information, Γ 1 , and the first piece of spatial information, ρ 1 ;
(4)
Let the first piece of spatial information, ρ 1 , be the linear weights, ϖ , from Equation (19), incorporating the first piece of spatial information into the second spatial transformation method to obtain the second spatial feature information, Γ 2 , and the second spatial information, ρ 2 ;
(5)
Let the second piece of spatial information, ρ 2 , be the linear weights, ϖ of, the exponential linear kernel function from Equation (24), incorporating the first and second pieces of spatial feature information into the third spatial transformation method to obtain the third piece of spatial feature information, Γ 3 . Linearly combine the feature information from the three spaces to obtain Γ = Γ 1 , Γ 2 , Γ 3 N * λ 1 + λ 2 + λ 3 , λ i = D , i = 1 , 2 , 3 and define each element in the matrix Γ 2 as the square of the corresponding element in the matrix Γ .
(6)
Construct the block matrix as follows:
0 2 Γ 2 I 4 Γ
And calculate its spectral radius, ρ . Then, construct the Mercer kernel matrix according to the following equation, which serves as the metric matrix for the fused space:
K F = Γ 2 + 2 ρ m Γ + 1 2 ρ m 2 H
where the parameter ρ m ρ ;
(7)
Apply classical MDS to compute the low-dimensional embedded manifold coordinates Y t r a i n for the training set using the metric matrix K F ;
(8)
Train a classifier using the low-dimensional embedded coordinates, Y, and their corresponding label information. Simultaneously, use a Multi-Layer Perceptron (MLP) to iteratively obtain the mapping matrix from the original training set, X, to the low-dimensional embedded manifold coordinates, Y;
(9)
Multiply the test set X t e s t by the mapping matrix obtained through the MLP iterations to obtain the low-dimensional embedded manifold coordinates for the test set, denoted as Y t e s t ;
(10)
Perform fault diagnosis on the low-dimensional embedded manifold coordinates for the test set Y t e s t using the trained classifier.
The three indicators in Fisher’s discriminant criterion, the inter-class scatter, S inter , the intra-class scatter, S intra , and Fisher’s metric, F, help calculate the distances and separation level between data points in the low-dimensional space, describing the clustering effectiveness of the dimensionality reduction algorithm. The definitions of these three indicators are as follows:
S inter = i = 1 L x i ¯ x ¯ x i ¯ x ¯ T
S intra = i = 1 L j = 1 N i x i j x i ¯ x i j x i ¯ T
F = w T S inter w w T S i ntra w
where w denotes all the one-column vectors, x i ¯ represents the mean vector for the i-th class, x ¯ represents the overall mean vector for all the sample points, x i j represents the j-th sample point in the i-th class, and L represents the number of classes. The confusion matrix records the complete results of the fault diagnosis, where the rows represent predicted labels and the columns represent true labels.
For the classification algorithm performance, we record the diagnostic process in detail using a confusion matrix and evaluate it using evaluation indexes such as accuracy, precision, recall, and the F1 score with the following expressions:
A c c u r a c y = T P + T N T P + F P + T N + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 _ s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
where TP denotes a true positive and TN denotes a true negative. FP denotes the false positive, and FN denotes a false negative.

4. Experiment and Analysis

In this section, the effectiveness of the proposed algorithm is evaluated using two cases in the context of rolling bearings, which are essential components of rotating machinery. The diagnostic results are compared with eight other commonly used manifold learning algorithms. The experiments were conducted using Python 3 and PyCharm 2021 software, PyCharm Community Edition 2021.3.2, running on hardware equipped with an 11-th Gen Intel(R) Core(TM) i7-11850H CPU @ 2.50GHz.

4.1. Case 1: CWRU

4.1.1. Data Description

The vibration data used in this case study are sourced from the Case Western Reserve University (CWRU) Bearing Data Center, as depicted in Figure 3. The experimental setup comprises components such as a motor, accelerometer, torque sensor, bearing, and dynamometer. The data collection involved a deep groove ball bearing SFK 6205, with bearing health states including Normal (NO), Rolling Element Fault (RF), Inner Race Fault (IF), and Outer Race Fault (OF), totaling four distinct modes [35].
The data in this case study were collected using an accelerometer, with the measurement point located at the motor drive end. The sampling frequency was set at 12 kHz. Detailed information regarding the rolling bearings is presented in Table 1. Additionally, the original vibration signals are depicted in Figure 4.
The experimental data were resampled to select 160 samples for each operational state of the bearing as the training set. The sampling window had a length of 1024, and the sliding window size was set at 512. For each sample, 17 time-domain features and 12 frequency-domain features were extracted. The same methodology was applied to select 40 samples for each operational state of the bearing as for the testing set.

4.1.2. Model Parameters Setting and Implementation

To comprehensively evaluate the proposed method, it is compared with eight other algorithms: PCA, MDS, Isomap, LE, LLE, HLLE, LTSA, and T-SNE. Among them, PCA and MDS are two commonly used linear manifold learning algorithms that have shown good performance in fault diagnosis applications. Isomap is a classic algorithm for preserving the global embedding in manifold learning, and LLE is a classic algorithm for preserving the local embedding. LE, HLLE, and LTSA have been improved by researchers based on LLE and have shown good performance in dimensionality reduction.
The detail parameter descriptions of all the methods are listed in Table 2.
To ensure the completeness of the experiments, we used six different classifiers: a logistic regression, a decision tree, a random forest, a plain Bayes, k-nearest neighbors (KNNs), and a support vector machine (SVM). These classifiers were trained using the dimensionality-reduced data obtained from various methods. Subsequently, the trained classifiers were tested using the test set.
For each method, we applied the dimensionality reduction technique and then fed the reduced data into each of the six classifiers mentioned above. The classifiers were trained using the training data, and their performance was evaluated using the test data.

4.1.3. Diagnosis Results and Discussion

We conducted two comprehensive experiments to compare our proposed method with others.
In the first experiment, we applied all the experimental methods to perform a dimensionality reduction on the data and conducted a quantitative analysis of clustering capability using Fisher’s discriminant criterion. In the second experiment, we separately trained different classifiers using the reduced data obtained from each method and conducted a quantitative analysis of the model’s diagnostic accuracy.
In the first experiment, to intuitively demonstrate the superiority of our method in terms of its clustering capability, we applied our proposed method and eight other methods to perform a dimensionality reduction on the vibration data from different states of the bearing. In Figure 5, we present 3D scatter plots of the abstract feature space of the bearing data after the dimensionality reduction.
From Figure 5a–h, it can be observed that after the dimensionality reduction using the traditional manifold learning algorithms, there are varying degrees of overlap between samples of different categories in the 3D feature space. Conversely, after the dimensionality reduction, using our proposed method, the data in the 3D feature space exhibit a better separation. This demonstrates the strong clustering capability achieved through the fusion of multi-feature information.
To provide a more precise description of our method’s clustering capability, we calculated Fisher’s discriminant criterion for the dimensionality-reduced data from all the methods. Table 3 presents detailed indicators of the Fisher criterion for all the methods. From Table 3, it can be observed that our method has the largest between-class scatter and the smallest within-class scatter. Although our method does not have the maximum class separability (Fisher’s criterion), it does have the largest Fisher measure (F), indicating that using our method for dimensionality reduction makes it easier to distinguish between data from different categories. This once again demonstrates the strong clustering capability of our method.
In the second experiment, to demonstrate the better diagnostic accuracy of our algorithm, we applied our proposed method and eight other methods to diagnose the vibration data from different states of the bearing. Figure 6 illustrates the diagnostic accuracy of all the methods when using different classifiers. It can be observed that our proposed method exhibits high accuracy across various classifiers, reaching a maximum classification accuracy of 100%.
We used a confusion matrix to document in detail the diagnostic process with the highest accuracy for the different fault diagnostic methods. Figure 7 presents the detailed contents of the confusion matrices. Evidently, most misclassified samples have overlapping fault characteristics, thus reducing the distinctiveness between the different categories of samples. Table 4 records the evaluation indexes at the best accuracy of each type of fault diagnosis method. It can be seen that from the data in Case 1 that our proposed method achieves good performance on all evaluation indexes.

4.2. Case 2: Laboratory-Built Bearing Experimental Rig

4.2.1. Data Description

The bearing vibration dataset used in this case study was obtained from a laboratory-built experimental rig. The laboratory-built bearing experimental rig and its theory graph are illustrated in Figure 8 and Figure 9, respectively. The setup comprises components such as a motor, DC driver, healthy bearing, experimental bearing, accelerometer, and loading system. The data collection was performed on a tapered roller bearing, model 33007. The bearing’s health conditions encompass five modes, Normal (NO), Rolling Element Fault (RF), Inner Race Fault (IF), Outer Race Fault (OF), and Cage Fault (CF), as depicted in Figure 10.
The data used in this case were collected with an accelerometer, located at the testing bearing position, with a sampling frequency of 12 kHz. The motor speed was 600 rpm, and the axial and radial loads were 1 kN. This case focuses on diagnosing the vibration data of rolling bearings under bi-directional loads. Detailed information about the rolling bearing is provided in Table 5, and the original vibration signals are depicted in Figure 11. The experimental data were preprocessed using the same methods as in Case 1.

4.2.2. Model Parameters Setting and Implementation

To comprehensively evaluate the performance of the proposed method on the laboratory-built experimental rig, we continue to utilize a grid search to select the optimal hyperparameters for each algorithm. The detailed parameter descriptions for all the methods are provided in Table 6. This is to substantiate what was mentioned in Section 3.2.3.

4.2.3. Diagnosis Results and Discussion

We conduct two experiments, similar to Case 1, to comprehensively compare the performance of our proposed method and other methods on the laboratory-built experimental rig.
In the first experiment, we apply all experimental methods to perform a dimensionality reduction on the data and quantitatively analyze the clustering ability using Fisher’s discriminant criterion. In the second experiment, we train different classifiers separately using the dimensionally reduced data obtained by each method and analyze the model’s diagnostic accuracy. Additionally, in the third experiment, we use a Gaussian kernel function, a linear kernel function, and the proposed exponential linear kernel function to construct the second spatial matrix for the dimensionality reduction. We then quantify the model’s diagnostic accuracy using the best-performing classifier.
In the first experiment, Figure 12 visually demonstrates the superiority of our method in terms of its clustering ability. As observed in Figure 12a–h, the traditional manifold learning algorithms still exhibit varying degrees of overlap in the case of Case 2. On the other hand, our proposed method continues to perform well, showcasing its robust clustering ability when applied to different datasets.
To provide a more precise description of our method’s clustering ability, we calculate Fisher’s discriminant criterion for the dimensionally reduced data obtained by all the methods. Table 6 presents the detailed metrics based on the Fisher discriminant criterion for each method. A comparison between Table 3 and Table 7 reveals an overall improvement in the F values for all the methods. Our proposed method maintains the highest F value, indicating excellent discriminability between different categories of data in Case 2 after using our method for dimensionality reduction. This once again strongly validates the robust clustering ability of our proposed method.
In the second experiment, to validate the superior performance of our algorithm in terms of its diagnostic accuracy, we applied our proposed method as well as eight other methods to diagnose faults in the Case 2 bearing vibration data. Figure 13 illustrates the diagnostic accuracy of all the methods when using different classifiers. It is evident that our proposed method exhibits higher stability in terms of diagnostic accuracy across various classifiers, with the highest accuracy reaching 100%.
For the fault diagnosis, we utilized the classifier that demonstrated the best accuracy for each method and recorded the diagnostic process using confusion matrices. Figure 14 provides detailed insights into the confusion matrices. Table 8 records the evaluation indexes at the best accuracy of each type of fault diagnosis method. It can be seen that for the data in Case 2, our proposed method achieves better performance on all the evaluation indexes than any other method.
In the third experiment, to showcase the superiority of our proposed kernel function over the traditional ones, we employed a linear kernel, a Gaussian kernel, and our proposed exponential linear kernel to extract features in the second space. Figure 15 illustrates confusion matrices detailing fault diagnosis using different kernel functions for feature extraction in the second space. It is evident that our proposed exponential linear kernel can extract more effective feature information, resulting in higher diagnostic accuracy.

5. Conclusions

Addressing the issue of dimensionality is a challenge in the current field of intelligent fault diagnosis of rolling bearings, so this paper proposes a supervised manifold learning method that integrates multiple pieces of feature information for diagnosing rolling bearing faults. Firstly, an adaptive nearest neighbor strategy is employed to reconstruct the manifold neighbor graph. Subsequently, multiple spatial transformation techniques are introduced to acquire feature information in different spaces. Notably, an innovative exponential linear kernel function and the KS-Isomap algorithm are presented to enrich the feature space with novel information. The multi-space feature information is then fused with discriminative information derived from the data labels, leading to the development of a supervised manifold learning method for feature extraction. Finally, this method is employed in collaboration with classifiers to conduct fault diagnosis on rolling bearings. The experimental validation using the CWRU open dataset and our laboratory-built experimental data demonstrates that the proposed AN-MSDIS-Isomap algorithm outperforms traditional manifold learning methods in clustering, dimension reduction, and fault diagnosis. It exhibits consistently good classification accuracy across various classifiers, with the highest classification accuracy reaching 100%.
The proposed method addresses the challenge of dimensionality and effectively extracts significant features representing the data. When combined with a classifier, it performs well in fault diagnosis tasks. Future research will focus on fault diagnosis tasks for bearing vibration signals in strong noise environments and the optimization of fault diagnosis performance between algorithms and different classifiers. Meanwhile, aiming to solve the problem of the high computational complexity of machine learning algorithms, research on methods to reduce computational costs and improve computational efficiency should be conducted.

6. Patents

There is a patent named fault diagnosis methods for mechanical equipment (patent number: ZL 2023 1 0839479.7) resulting from the work reported in this manuscript.

Author Contributions

Funding acquisition, H.W. (Hongwei Wang), H.W. (Haoran Wang) and L.T.; Conceptualization, L.Y., Y.L. and Z.L.; Supervision, H.W. (Hongwei Wang), D.W. and R.H. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was supported in part by the Key R&D Program of Shanxi Province (grant number: 202102100401017), the National Key Research and Development Program of China (grant number: 2020YFB1314004), the Bidding Project of Shanxi Province (grant number: 20201101008), the National Key Research and Development Program of Shanxi Province (grant number: 202102100401015), and the Central Guidance for Local Science and Technology Development Projects (grant number: YDZJSX2022A015).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to the grant program has not yet been completed.

Conflicts of Interest

The authors declare that they have no known competing financial interest or personal relationships that could have appeared to influence the work reported in this paper.

Appendix A

Proof of Semi-Positive Definite Property to Exponential Linear Kernel Function Matrix

According to the definition of the exponential linear kernel function in Section 3.2.3, it is evident that it incorporates distance information, rendering the exponential linear kernel function matrix a symmetric matrix. Subsequently, we only need to demonstrate its positive semi-definite property to establish the exponential linear kernel function as a valid kernel function. The expression for the exponential linear kernel function matrix is as follows:
K E l i n e a r = k E l i n e a r x i , x j = e ϖ d E x i , x j + b
The proof of the positive semi-definite property is presented as follows:
e ϖ d E x i , x j + b
= e ϖ x i x j 2 + b
= e b · e ϖ x i x j 2
= e b · e ϖ k = 1 D x i k 2 · e ϖ k = 1 D x j k 2 · k = 1 D e 2 ϖ x i k x j k
By performing a Taylor expansion on the fourth term of Equation (A4), we obtain:
= e b · e ϖ k = 1 D x i k 2 · e ϖ k = 1 D x j k 2 · k = 1 D m = 0 2 ϖ m m ! x i k m x j k m
= m = 0 e ϖ k = 1 D x i k 2 · e b / 2 · k = 1 D 2 ϖ m m ! x i k m · m = 0 e ϖ k = 1 D x j k 2 · e b / 2 · k = 1 D 2 ϖ m m ! x j k m
= i = 1 N Φ x i 2 0
Φ x i = m = 0 e ϖ k = 1 D x i k 2 · e b / 2 · k = 1 D 2 ϖ m m ! x i k m .
Therefore, it is evident that the exponential linear kernel function satisfies the necessary and sufficient conditions for being a kernel function.

References

  1. Zhu, Z.; Lei, Y.; Qi, G.; Chai, Y.; Mazur, N.; An, Y.; Huang, X. A Review of the Application of Deep Learning in Intelligent Fault Diagnosis of Rotating Machinery. Measurement 2023, 206, 112346. [Google Scholar] [CrossRef]
  2. Trstanova, Z.; Leimkuhler, B.; Lelièvre, T. Local and Global Perspectives on Diffusion Maps in the Analysis of Molecular Systems. Proc. R. Soc. Math. Phys. Eng. Sci. 2020, 476, 20190036. [Google Scholar] [CrossRef]
  3. Wang, H.; Fang, Z.; Wang, H.; Li, Y.; Geng, Y.; Chen, L.; Chang, X. A Novel Time-Frequency Analysis Method for Fault Diagnosis Based on Generalized S-Transform and Synchroextracting Transform. Meas. Sci. Technol. 2023, 35, 036101. [Google Scholar] [CrossRef]
  4. Peng, C.; Gao, H.; Liu, X.; Liu, B. A Visual Vibration Characterization Method for Intelligent Fault Diagnosis of Rotating Machinery. Mech. Syst. Signal Process. 2023, 192, 110229. [Google Scholar] [CrossRef]
  5. Su, Z.; Tang, B.; Deng, L.; Liu, Z. Fault Diagnosis Method Using Supervised Extended Local Tangent Space Alignment for Dimension Reduction. Measurement 2015, 62, 1–14. [Google Scholar] [CrossRef]
  6. Wang, Z.; Yao, L.; Cai, Y. Rolling Bearing Fault Diagnosis Using Generalized Refined Composite Multiscale Sample Entropy and Optimized Support Vector Machine. Measurement 2020, 156, 107574. [Google Scholar] [CrossRef]
  7. Anowar, F.; Sadaoui, S.; Selim, B. Conceptual and Empirical Comparison of Dimensionality Reduction Algorithms (PCA, KPCA, LDA, MDS, SVD, LLE, ISOMAP, LE, ICA, t-SNE). Comput. Sci. Rev. 2021, 40, 100378. [Google Scholar] [CrossRef]
  8. Ray, P.; Reddy, S.S.; Banerjee, T. Various Dimension Reduction Techniques for High Dimensional Data Analysis: A Review. Artif. Intell. Rev. 2021, 54, 3473–3515. [Google Scholar] [CrossRef]
  9. Tenenbaum, J.B.; Silva, V.D.; Langford, J.C. A Global Geometric Framework for Nonlinear Dimensionality Reduction. Science 2000, 290, 2319–2323. [Google Scholar] [CrossRef]
  10. Roweis, S.T.; Saul, L.K. Nonlinear Dimensionality Reduction by Locally Linear Embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef]
  11. Xue, J.; Zhang, B.; Qiang, Q. Local Linear Embedding with Adaptive Neighbors. Pattern Recognit. 2023, 136, 109205. [Google Scholar] [CrossRef]
  12. Cao, Y.; Chen, D.-R. On the Regularized Laplacian Eigenmaps. J. Stat. Plan. Inference 2012, 142, 1627–1643. [Google Scholar] [CrossRef]
  13. Donoho, D.L.; Grimes, C. Hessian Eigenmaps: Locally Linear Embedding Techniques for High-Dimensional Data. Proc. Natl. Acad. Sci. USA 2003, 100, 5591–5596. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, T.; Yang, J.; Zhao, D.; Ge, X. Linear Local Tangent Space Alignment and Application to Face Recognition. Neurocomputing 2007, 70, 1547–1553. [Google Scholar] [CrossRef]
  15. Zheng, J.; Jiang, Z.; Pan, H. Sigmoid-Based Refined Composite Multiscale Fuzzy Entropy and t-SNE Based Fault Diagnosis Approach for Rolling Bearing. Measurement 2018, 129, 332–342. [Google Scholar] [CrossRef]
  16. Zhang, Z.; Wang, J.; Zha, H. Adaptive Manifold Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 253–265. [Google Scholar] [CrossRef] [PubMed]
  17. Sun, C.; Wang, P.; Yan, R.; Gao, R.X.; Chen, X. Machine Health Monitoring Based on Locally Linear Embedding with Kernel Sparse Representation for Neighborhood Optimization. Mech. Syst. Signal Process. 2019, 114, 25–34. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Zhang, Z.; Li, S.; Qin, J.; Liu, G.; Wang, M.; Yan, S. Unsupervised Nonnegative Adaptive Feature Extraction for Data Representation. IEEE Trans. Knowl. Data Eng. 2019, 31, 2423–2440. [Google Scholar] [CrossRef]
  19. Gao, Y.; Zhang, Y.; Pan, J.; Luo, S.; Yang, C. Discriminant Analysis Based on Reliability of Local Neighborhood. Expert Syst. Appl. 2021, 175, 114790. [Google Scholar] [CrossRef]
  20. An, J.; Zhao, X.; Shi, M.; Liu, X.; Guo, J. Joint Neighborhood Preserving and Projected Clustering for Feature Extraction. Neurocomputing 2022, 488, 572–580. [Google Scholar] [CrossRef]
  21. Chatpatanasiri, R.; Kijsirikul, B. A Unified Semi-Supervised Dimensionality Reduction Framework for Manifold Learning. Neurocomputing 2010, 73, 1631–1640. [Google Scholar] [CrossRef]
  22. Wang, J.; Zhang, X.; Li, X.; Du, J. Semi-Supervised Manifold Alignment with Few Correspondences. Neurocomputing 2017, 230, 322–331. [Google Scholar] [CrossRef]
  23. Karimi, Z.; Shiry Ghidary, S. Semi-Supervised Metric Learning in Stratified Spaces via Intergrating Local Constraints and Information-Theoretic Non-Local Constraints. Neurocomputing 2018, 312, 165–176. [Google Scholar] [CrossRef]
  24. Chen, M.; Wang, J.; Li, X.; Sun, X. Robust Semi-Supervised Manifold Learning Algorithm for Classification. Math. Probl. Eng. 2018, 2018, 2382803. [Google Scholar] [CrossRef]
  25. Li, D.; Dick, S. Semi-Supervised Multi-Label Classification Using an Extended Graph-Based Manifold Regularization. Complex Intell. Syst. 2022, 8, 1561–1577. [Google Scholar] [CrossRef] [PubMed]
  26. Ma, J.; Yu, G.; Xiong, W.; Zhu, X. Safe Semi-Supervised Learning for Pattern Classification. Eng. Appl. Artif. Intell. 2023, 121, 106021. [Google Scholar] [CrossRef]
  27. Ornek, C.; Vural, E. Nonlinear Supervised Dimensionality Reduction via Smooth Regular Embeddings. Pattern Recognit. 2019, 87, 55–66. [Google Scholar] [CrossRef]
  28. Hajderanj, L.; Chen, D.; Weheliye, I. The Impact of Supervised Manifold Learning on Structure Preserving and Classification Error: A Theoretical Study. IEEE Access 2021, 9, 43909–43922. [Google Scholar] [CrossRef]
  29. Qu, H.; Li, L.; Li, Z.; Zheng, J. Supervised Discriminant Isomap with Maximum Margin Graph Regularization for Dimensionality Reduction. Expert Syst. Appl. 2021, 180, 115055. [Google Scholar] [CrossRef]
  30. Faigenbaum-Golovin, S.; Levin, D. Manifold Reconstruction and Denoising from Scattered Data in High Dimension. J. Comput. Appl. Math. 2023, 421, 114818. [Google Scholar] [CrossRef]
  31. Apsemidis, A.; Psarakis, S.; Moguerza, J.M. A Review of Machine Learning Kernel Methods in Statistical Process Monitoring. Comput. Ind. Eng. 2020, 142, 106376. [Google Scholar] [CrossRef]
  32. Choi, H.; Choi, S. Robust Kernel Isomap. Pattern Recognit. 2007, 40, 853–862. [Google Scholar] [CrossRef]
  33. Hofmann, T.; Schölkopf, B.; Smola, A.J. Kernel Methods in Machine Learning. Ann. Stat. 2008, 36, 1171–1220. [Google Scholar] [CrossRef]
  34. Pilario, K.E.; Shafiee, M.; Cao, Y.; Lao, L.; Yang, S.-H. A Review of Kernel Methods for Feature Extraction in Nonlinear Process Monitoring. Processes 2019, 8, 24. [Google Scholar] [CrossRef]
  35. Smith, W.A.; Randall, R.B. Rolling Element Bearing Diagnostics Using the Case Western Reserve University Data: A Benchmark Study. Mech. Syst. Signal Process. 2015, 64–65, 100–131. [Google Scholar] [CrossRef]
Figure 1. Flowchart of adaptive neatest neighbor strategy.
Figure 1. Flowchart of adaptive neatest neighbor strategy.
Sensors 23 09820 g001
Figure 2. Flowchart of AN-MFIDFS-Isomap fault diagnosis arithmetic.
Figure 2. Flowchart of AN-MFIDFS-Isomap fault diagnosis arithmetic.
Sensors 23 09820 g002
Figure 3. The bearing test rig of CWRU.
Figure 3. The bearing test rig of CWRU.
Sensors 23 09820 g003
Figure 4. Vibration signals of four work conditions for Case 1: (a) NO; (b) IF; (c) RF; (d) OF.
Figure 4. Vibration signals of four work conditions for Case 1: (a) NO; (b) IF; (c) RF; (d) OF.
Sensors 23 09820 g004
Figure 5. Scatter plots of low-dimensional features in 3D feature space for Case 1: (a) PCA; (b) MDS; (c) Isomap; (d) LE; (e) LLE; (f) HLLE; (g) LTSA; (h) T-SNE; (i) proposed method.
Figure 5. Scatter plots of low-dimensional features in 3D feature space for Case 1: (a) PCA; (b) MDS; (c) Isomap; (d) LE; (e) LLE; (f) HLLE; (g) LTSA; (h) T-SNE; (i) proposed method.
Sensors 23 09820 g005
Figure 6. Accuracy of all methods with different classifiers for Case 1.
Figure 6. Accuracy of all methods with different classifiers for Case 1.
Sensors 23 09820 g006
Figure 7. Confusion matrices of all method for Case 1: (a) PCA–Random Forest; (b) MDS–Decision Tree; (c) Isomap–Plain Bayes; (d) LE–Plain Bayes; (e) LLE–Plain Bayes; (f) HLLE–Plain Bayes; (g) LTSA–Plain Bayes; (h) T-SNE–SVM; (i) proposed method–SVM.
Figure 7. Confusion matrices of all method for Case 1: (a) PCA–Random Forest; (b) MDS–Decision Tree; (c) Isomap–Plain Bayes; (d) LE–Plain Bayes; (e) LLE–Plain Bayes; (f) HLLE–Plain Bayes; (g) LTSA–Plain Bayes; (h) T-SNE–SVM; (i) proposed method–SVM.
Sensors 23 09820 g007
Figure 8. The laboratory-built bearing experimental rig.
Figure 8. The laboratory-built bearing experimental rig.
Sensors 23 09820 g008
Figure 9. The working schematic of laboratory-built bearing experimental rig.
Figure 9. The working schematic of laboratory-built bearing experimental rig.
Sensors 23 09820 g009
Figure 10. Bearing health status: (a) Rolling Element Fault; (b) Inner Race Fault; (c) Outer Race Fault; (d) Cage Fault; (e) Healthy Bearing. The red box shows the exact shape of the bearing fault in detail.
Figure 10. Bearing health status: (a) Rolling Element Fault; (b) Inner Race Fault; (c) Outer Race Fault; (d) Cage Fault; (e) Healthy Bearing. The red box shows the exact shape of the bearing fault in detail.
Sensors 23 09820 g010
Figure 11. Vibration signals of five work conditions for Case 2: (a) NO; (b) IF; (c) RF; (d) OF; (e) CF.
Figure 11. Vibration signals of five work conditions for Case 2: (a) NO; (b) IF; (c) RF; (d) OF; (e) CF.
Sensors 23 09820 g011
Figure 12. Scatter plots of low-dimensional features in 3D feature space for Case 2: (a) PCA; (b) MDS; (c) Isomap; (d) LE; (e) LLE; (f) HLLE; (g) LTSA; (h) T-SNE; (i) proposed method.
Figure 12. Scatter plots of low-dimensional features in 3D feature space for Case 2: (a) PCA; (b) MDS; (c) Isomap; (d) LE; (e) LLE; (f) HLLE; (g) LTSA; (h) T-SNE; (i) proposed method.
Sensors 23 09820 g012
Figure 13. Accuracy of all methods with different classifiers for Case 2.
Figure 13. Accuracy of all methods with different classifiers for Case 2.
Sensors 23 09820 g013
Figure 14. Confusion matrices of all method for Case 2: (a) PCA–Decision Tree; (b) MDS–Decision Tree; (c) Isomap–SVM; (d) LE–Random Forest; (e) LLE–Logistic Regression; (f) HLLE–Plain Bayes; (g) LTSA–Plain Bayes; (h) T-SNE–SVM; (i) proposed method–SVM.
Figure 14. Confusion matrices of all method for Case 2: (a) PCA–Decision Tree; (b) MDS–Decision Tree; (c) Isomap–SVM; (d) LE–Random Forest; (e) LLE–Logistic Regression; (f) HLLE–Plain Bayes; (g) LTSA–Plain Bayes; (h) T-SNE–SVM; (i) proposed method–SVM.
Sensors 23 09820 g014
Figure 15. Confusion matrix of different kernel functions for Case 2: (a) linear kernel function; (b) gaussian kernel function; (c) exponential linear kernel function.
Figure 15. Confusion matrix of different kernel functions for Case 2: (a) linear kernel function; (b) gaussian kernel function; (c) exponential linear kernel function.
Sensors 23 09820 g015
Table 1. Different work conditions of drive end bearings details.
Table 1. Different work conditions of drive end bearings details.
PatternsLabels
NO0
IF1
RF2
OF3
Table 2. Parameter settings of all methods in Case 1.
Table 2. Parameter settings of all methods in Case 1.
MethodsParameters Settings
Proposed methodIntrinsic dimension d = 4, the number of nearest neighbors k = 10, the offset coefficient b = 0.5, the bias coefficient φ = 0.5. The structure of MPL is 29-128-256-128-4 for the training data. The learning rate = 0.001; the iteration number of pre-training is 1000. The optimizer is Adam.
PCAIntrinsic dimension d = 4.
MDSIntrinsic dimension d = 4.
IsomapIntrinsic dimension d = 5; the number of nearest neighbor k = 5.
LEIntrinsic dimension d = 4; the number of nearest neighbor k = 5.
LLEIntrinsic dimension d = 2; the number of nearest neighbor k = 30.
HLLEIntrinsic dimension d = 3; the number of nearest neighbor k = 25.
LTSAIntrinsic dimension d = 9; the number of nearest neighbor k = 30.
T-SNEIntrinsic dimension d = 2; perplexity p = 30.
Table 3. The detailed Fisher statistical quantity of all methods for Case 1.
Table 3. The detailed Fisher statistical quantity of all methods for Case 1.
Methods S inter S intra F
Proposed method2.5009 × 1043.6600 × 106.8331 × 102
PCA6.3453 × 10103.0215 × 10102.1000
MDS6.3283 × 10103.0406 × 10102.0812
Isomap8.9909 × 10103.2154 × 10102.7962
LE1.6468 × 10−11.12566 × 10−11.3106
LLE2.55002.0001 × 10−11.2749 × 10
HLLE8.9230 × 10−11.43636.2125 × 10−1
LTSA7.86899 × 10−11.86334.2231 × 10−1
T-SNE7.2028 × 1041.3775 × 1055.2287 × 10−1
Table 4. The detailed evaluation indexes of all methods with the best accuracy for Case 1.
Table 4. The detailed evaluation indexes of all methods with the best accuracy for Case 1.
MethodsAccuracyPrecisionRecallF1_Score
Proposed method100%100%100%1
PCA99.375%97.5%100%0.9873
MDS86.25%96.5035%89.0323%0.9261
Isomap68.125%83.8462%78.4173%0.8104
LE49.375%76.6990%57.2464%0.6556
LLE67.5%90.7563%72.4832%0.8060
HLLE64.375%69.5946%89.5652%0.7833
LTSA64.375%67.3203%93.6364%0.7833
T-SNE44.375%68.9320%56.3492%0.6201
Table 5. Different work conditions of laboratory-built bearing experimental rig details.
Table 5. Different work conditions of laboratory-built bearing experimental rig details.
PatternsFault Diameter
(mm)
Labels
NO00
IF13.8 × 4.81
RF13.5 × 3.22
OF16.6 × 4.73
CF16.7 × 4.34
Table 6. Parameter settings of all methods in Case 2.
Table 6. Parameter settings of all methods in Case 2.
MethodsParameters Settings
Proposed methodIntrinsic dimension d = 3, the number of nearest neighbor k = 10, the offset coefficient b = 0.5, the bias coefficient φ = 0.5. The structure of MPL is 29-128-256-128-5 for the training data. The learning rate = 0.001; the iteration number of pre-training is 1000. The optimizer is Adam.
PCAIntrinsic dimension d = 4.
MDSIntrinsic dimension d = 4.
IsomapIntrinsic dimension d = 4; the number of nearest neighbor k = 15.
LEIntrinsic dimension d = 3; the number of nearest neighbor k = 25.
LLEIntrinsic dimension d = 4; the number of nearest neighbor k = 15.
HLLEIntrinsic dimension d = 3; the number of nearest neighbor k = 25.
LTSAIntrinsic dimension d = 7; the number of nearest neighbor k = 25.
T-SNEIntrinsic dimension d = 4; perplexity p = 15.
Table 7. The detail Fisher statistical quantity of all methods for Case 2.
Table 7. The detail Fisher statistical quantity of all methods for Case 2.
Methods S inter S intra F
Proposed method4.2742 × 1043.8155 × 101.1202 × 103
PCA1.2840 × 10115.7755 × 10102.2232
MDS1.2842 × 10115.7817 × 10102.2211
Isomap2.0071 × 10116.5189 × 10103.0789
LE2.7861 × 10−14.0982 × 10−26.7985
LLE1.98528.1480 × 10−12.4364 × 10
HLLE1.58335.0543 × 10−13.1325
LTSA1.11781.31318.5121 × 10−1
T-SNE1.2986 × 1051.0083 × 1051.2879
Table 8. The detailed evaluation indexes of all methods with the best accuracy for Case 2.
Table 8. The detailed evaluation indexes of all methods with the best accuracy for Case 2.
MethodsAccuracyPrecisionRecallF1_Score
Proposed method100%100%100%1
PCA98%98.9899%98.9899%0.9899
Isomap92%98.3957%93.4010%0.9583
LE61%75.3086%76.25%0.7578
LLE60%70.5882%80%0.75
MDS33%39.0533%69.4737%0.5
HLLE39.5%39.5%100%0.5663
LTSA39.5%39.5%100%0.5663
T-SNE21%21%100%0.3471
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, H.; Yao, L.; Wang, H.; Liu, Y.; Li, Z.; Wang, D.; Hu, R.; Tao, L. Supervised Manifold Learning Based on Multi-Feature Information Discriminative Fusion within an Adaptive Nearest Neighbor Strategy Applied to Rolling Bearing Fault Diagnosis. Sensors 2023, 23, 9820. https://doi.org/10.3390/s23249820

AMA Style

Wang H, Yao L, Wang H, Liu Y, Li Z, Wang D, Hu R, Tao L. Supervised Manifold Learning Based on Multi-Feature Information Discriminative Fusion within an Adaptive Nearest Neighbor Strategy Applied to Rolling Bearing Fault Diagnosis. Sensors. 2023; 23(24):9820. https://doi.org/10.3390/s23249820

Chicago/Turabian Style

Wang, Hongwei, Linhu Yao, Haoran Wang, Yu Liu, Zhiyuan Li, Di Wang, Ren Hu, and Lei Tao. 2023. "Supervised Manifold Learning Based on Multi-Feature Information Discriminative Fusion within an Adaptive Nearest Neighbor Strategy Applied to Rolling Bearing Fault Diagnosis" Sensors 23, no. 24: 9820. https://doi.org/10.3390/s23249820

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop