Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Blind Separation of Skin Chromophores from Multispectral Dermatological Images
Previous Article in Journal
Early Cervical Cancer Diagnosis with SWIN-Transformer and Convolutional Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Leveraging Classifier Performance Using Heuristic Optimization for Detecting Cardiovascular Disease from PPG Signals

by
Sivamani Palanisamy
1 and
Harikumar Rajaguru
2,*
1
Department of Electronics and Communication Engineering, Jansons Institute of Technology, Coimbatore 641659, India
2
Department of Electronics and Communication Engineering, Bannari Amman Institute of Technology, Sathyamangalam 638401, India
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(20), 2287; https://doi.org/10.3390/diagnostics14202287
Submission received: 26 August 2024 / Revised: 24 September 2024 / Accepted: 10 October 2024 / Published: 14 October 2024
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Background/Objectives: Photoplethysmography (PPG) signals, which measure blood volume changes through light absorption, are increasingly used for non-invasive cardiovascular disease (CVD) detection. Analyzing PPG signals can help identify irregular heart patterns and other indicators of CVD. Methods: This research involves a total of 41 subjects sourced from the CapnoBase database, consisting of 21 normal subjects and 20 CVD cases. In the initial stage, heuristic optimization algorithms, such as ABC-PSO, the Cuckoo Search algorithm (CSA), and the Dragonfly algorithm (DFA), were applied to reduce the dimension of the PPG data. Next, these Dimensionally Reduced (DR) PPG data are then fed into various classifiers such as Linear Regression (LR), Linear Regression with Bayesian Linear Discriminant Classifier (LR-BLDC), K-Nearest Neighbors (KNN), PCA-Firefly, Linear Discriminant Analysis (LDA), Kernel LDA (KLDA), Probabilistic LDA (ProbLDA), SVM-Linear, SVM-Polynomial, and SVM-RBF, to identify CVD. Classifier performance is evaluated using Accuracy, Kappa, MCC, F1 Score, Good Detection Rate (GDR), Error rate, and Jaccard Index (JI). Results: The SVM-RBF classifier for ABC PSO dimensionality reduced values outperforms other classifiers, achieving the highest accuracy of 95.12% along with the minimum error rate of 4.88%. In addition to that, it provides an MCC and kappa value of 0.90, a GDR and F1 score of 95%, and a Jaccard Index of 90.48%. Conclusions: This study demonstrated that heuristic-based optimization and machine learning classification of PPG signals are highly effective for the non-invasive detection of cardiovascular disease.

1. Introduction

Cardiovascular disease continues to be the leading cause of death globally, significantly burdening healthcare systems and economies. It accounts for approximately 17.9 million deaths annually, necessitating urgent advancements in prevention, diagnosis, and treatment strategies. Globally, life expectancy is on the rise due to significant advancements in healthcare, medicine, and a heightened awareness of personal hygiene and environmental hygiene [1]. Optimizing computer-aided diagnosis holds immense potential for healthcare. It fosters a more objective and consistent diagnostic approach, ultimately benefiting patient outcomes. Photoplethysmography (PPG) technology has the potential to revolutionize early detection of cardiovascular issues, aimed at increasing life expectancy and reducing healthcare costs [2,3]. It can be easily incorporated into wearable medical technology to measure various health-related metrics. Due to its diagnostic capabilities, PPG is widely utilized in clinical practices. Over the past decades, there has been a swift increase in the implementation of Wireless Patient Monitoring (WPM) and Internet of Things (IoT) models worldwide [4]. Cutting-edge smart home automation and e-healthcare system technologies now enable in-home medical services, reducing the need for hospital visits [5]. Digital health has increasingly been integrated into daily life through digital health tools like smartwatches and health apps, facilitating real-time monitoring and diagnostics [6]. High blood pressure is a significant threat to developing cardiovascular diseases (CVDs) [7]. The World Health Organization (WHO) identifies cardiovascular disease as the leading chronic illness globally, significantly contributing to the overall disease burden and responsible for 31% of deaths worldwide [8]. PPG technology, highlighted as versatile and cost-effective [9], utilizes a photoelectric sensor to detect changes in light transmitted or reflected by blood vessels in the skin [10]. PPG does not require specific placement of sensors at predetermined locations on the body [11]; Measurements can be easily taken from the finger, wrist, or even earlobe [12]. This user-friendly approach makes PPG a popular bio signal for wearables, especially for monitoring heart rate during exercise and physical activity.
One of the most widely used technologies for monitoring a patient’s physiological conditions is PPG. It is popular due to its non-invasive nature and low cost. Additionally, PPG’s ability to provide continuous readings makes it ideal for use in pulse oximetry [13]. PPG signals require minimal hardware compared to traditional ECG monitoring systems, making them more accessible [14]. Shintomi et al. [15] investigated the effectiveness of compensating for heartbeat errors in mobile and wearable sensor data to improve heart rate variability (HRV) analysis. PPG sensors do not require a reference signal, making them ideal for integration into wristbands. This enhances their application, utility, and clinical applicability, making them highly effective for various research purposes in the analysis and diagnosis of CVD [16]. Moshawrab et al. [17] explored the use of smart wearable devices for detecting and predicting CVDs. The authors also reviewed the development and use of these wearables, which demonstrated their high effectiveness in managing CVDs. Reisner et al. [18] developed the use of the PPG for circulatory monitoring.
PPG signals are mostly preferred for CVD analysis because they are non-invasive and inexpensive while still offering valuable insights into blood flow variations that might indicate cardiovascular issues. In recent years, machine learning algorithms have gained significant attention in medical diagnostics, providing a lot of potential for accuracy and efficiency in the detection of diseases [19]. When paired with effective dimensionality reduction techniques, they can greatly improve the accuracy of CVD diagnosis. This research seeks to address these challenges by exploring the effectiveness of dimensionality reduction techniques in enhancing the accuracy of CVD detection using PPG data. We explore three heuristic DR techniques: ABC-PSO, the Cuckoo Search algorithm (CSA), and the Dragonfly algorithm (DFA). In this work, these techniques are integrated with various classification methods to create a robust framework for early and accurate CVD detection.
Researchers have explored a diverse range of techniques for diagnosing CVD from PPG signals. Several DR techniques can be applied to the PPG dataset to reduce both the dimensionality and computational complexity of PPG signals for CVD classification. The methods used for various types of CVD classification problems, along with their limitations, are presented in Table 1. In light of these limitations, this paper explores advanced techniques such as heuristic optimization algorithms for reducing the dimensionality of PPG data.
Compared to the feature extraction techniques mentioned in Table 1, the selected techniques (ABC-PSO, Cuckoo Search, and Dragonfly) are simpler and provide enhanced interpretability. These techniques allow researchers to identify which features are most significant for classification. This is essential in a medical context, where understanding the reasoning behind a model’s decisions is crucial. Additionally, the selected techniques are typically less computationally intensive than other models. This enables quicker training and analysis, which can be advantageous in a clinical environment.
This study focuses on comparing optimization techniques and classification methods rather than aiming for the absolute highest accuracy. The selected techniques offer a solid basis for this comparison while ensuring interpretability and efficiency. However, it is still important to consider interpretability and data size limitations when applying machine learning algorithms for CVD classification. In summary, the chosen techniques emphasize interpretability, efficiency, and the ability to facilitate comparisons between optimization technique and classification methods, making them well-suited for this particular study.
This research work makes the following key contributions:
The study proposes an early detection and intervention method for cardiovascular diseases using PPG signals.
Three metaheuristic optimization algorithms are used as DR techniques to reduce the dimension of the high-dimensional PPG data.
The dimensionality-reduced PPG data were then analyzed using ten different classification algorithms to detect the presence of CVD. The classifiers’ performance is evaluated using parameters such as accuracy, GDR, MCC, Kappa, error rate, F1 score, and Jaccard index.
In summary, to the best of our knowledge, no study in the literature has exclusively analyzed heuristic-based DR techniques to enhance classifier performance for detecting CVD from PPG data.
The structure of this article is as follows: Section 1 presents the introduction, followed by the methodology in Section 2. Heuristic dimensionality reduction techniques are discussed in Section 3. Section 4 covers the ten different classifiers based on learning and selection parameters used to differentiate between normal and CVD-related segments within the PPG signal data. The findings of the work are thoroughly examined and interpreted in Section 5, and the conclusion of the work is provided in Section 6.

2. Materials and Methods

2.1. Materials

PPG data recordings with diverse wave shapes from the CapnoBase database have been used in this work. This database is a publicly available online resource that adheres to the IEEE TMBE pulse oximeter standard [30]. This article explores the CapnoBase dataset, utilizing the complete IEEE benchmark (41 records) for the experiment, with 20 records representing CVD and 21 representing normal conditions. The PPG signals are digitized at a rate of 200 samples per second. For analysis purposes, each one-second interval of the PPG signal is defined as a segment. Therefore, each patient has 720 individual segments for further examination. Consequently, each patient has 144,000 samples (720 segments × 200 samples per segment). The total number of CVD segments is [20 × 720 = 14,400], and in normal cases, the total number of segments is [21 × 720 = 15,120]. Therefore, in total, 29,520 one-second segments are available for analysis from the 41 cases. The PPG signals are analyzed based on signal segments across the patients. Beat-to-beat analysis is not included in this study. Noise components from the PPG signals were removed by utilizing Independent Component Analysis (ICA). This investigation is conducted using MATLAB R2020a software. Figure 1 shows a normal PPG signal, and Figure 2 depicts the PPG signal obtained from a CVD person.

2.2. Methods

CVD analysis using PPG signals is crucial due to the rising prevalence of cardiovascular diseases globally. PPG technology offers a non-invasive, cost-effective, and easily deployable method for continuous heart health monitoring, facilitating early detection and timely intervention. This can reduce the burden on healthcare systems and improve patient outcomes. The integration of PPG analysis with wearable devices and advanced machine learning algorithms enhances diagnostic accuracy and efficiency. As healthcare shifts towards personalized and preventive models, PPG-based CVD analysis becomes a pivotal tool in the fight against CVDs. Therefore, CVD detection from PPG signals is considered in this research work. The main objective of this work is to enhance CVD diagnosis through more precise classification systems. Effectively categorizing CVD data not only ensures patients receive appropriate care at reduced costs but also lowers their risk of developing the disease. Classifier accuracy tends to decrease when unimportant and noisy signals are present in recorded signals. To enhance the quality of the recorded PPG signals, an efficient filtering technique is implemented to eliminate unwanted noise and artifacts.
This research utilizes heuristic dimensionality reduction (DR) techniques as the initial step to reduce the dimensionality of the PPG data. Specifically, the research incorporates methods such as ABC-PSO (Artificial Bee Colony-Particle Swarm Optimization), the Cuckoo Search algorithm, and the Dragonfly algorithm. In the second stage, the optimized PPG data were input into ten different classifiers to detect cardiovascular disease from PPG signals. The classifiers’ performance is evaluated and differentiated using parameter metrics for heuristically optimized PPG values. Figure 3 provides a detailed illustration of the workflow.

3. Dimensionality Reduction Techniques

Now, each person has 144,000 samples of PPG signals (720 × 200). The objective of dimensionality reduction in PPG signals is to decrease the number of variables, thereby improving computational efficiency and reducing the risk of overfitting. High-dimensional data can be noisy and redundant, making it difficult for classifiers to accurately identify patterns associated with CVD. By extracting the most relevant features, dimensionality reduction techniques such as ABC-PSO, Cuckoo Search, and Dragonfly are employed to enhance the performance of machine learning models.

3.1. ABC-PSO (Artificial Bee Colony-Particle Swarm Optimization)

ABC-PSO is a hybrid dimensionality reduction technique that combines the global search capabilities of the ABC algorithm with the local exploitation abilities of Particle Swarm Optimization (PSO). ABC emphasizes exploration through employed and onlooker bees, where solutions are iteratively improved. PSO, inspired by social behavior, optimizes by updating particle velocities based on personal best and global best solutions. By integrating these approaches, the hybrid algorithm aims to enhance exploration-exploitation balance, leveraging ABC’s local search capability and PSO’s global search efficiency. Figure 4 illustrates the flowchart of the ABC-PSO algorithm. This synergy enhances the efficiency of handling large datasets and improving the performance of data analysis [31].
ABC-PSO Algorithm Steps:
Initialization: Initialize bee and particle populations with random solutions. Set the number of employed bees, onlooker bees, and scout bees, as well as particle positions x i and velocities v i .
Employed Bee Phase: Each employed bee explores new food sources (solutions) using the following:
V i j = X i j + i j X i j X k j
where i j is a random number and X k j is a neighboring solution.
Onlooker Bee Phase: Onlooker bees in the algorithm choose their food sources probabilistically:
P i   = f i i = 1 N f i
where f i is the fitness of solution x i .
Scout Bee Phase: Abandon poor solutions and have scout bees search for new random solutions.
PSO Update: Update particle velocities and positions using the following equations [32]:
v i t + 1 = ω v i t + c 1 r 1 p i x i + c 2 r 2 g x i
x i t + 1 = x i t + v i t + 1
where p i is the personal best position, g is the global best position, ω is the inertia weight, r 1 and r 2 are random numbers, c 1 and c 2 are acceleration coefficients.
Evaluation and Selection: Evaluate new solutions and select the best ones based on fitness.
Convergence Check: Repeat the steps until stopping criteria, such as maximum iterations or a convergence threshold are met.

3.2. Cuckoo Search Algorithm (CSA)

It is a nature-inspired metaheuristic optimization technique established by Xin-She Yang and Suash Deb in 2009 [33]. Figure 5 illustrates the methodology of the CSA. It is effective for dimensionality reduction by finding optimal feature subsets. The feature subset is chosen that optimizes a performance metric, specifically aiming for the lowest MSE. This algorithm draws inspiration from the reproductive strategy of cuckoo birds. It simulates the unique behavior of cuckoos, which lay their eggs in the nests of other bird species. The algorithm employs both randomization and local search techniques to guide the cuckoos towards the optimal solution. A notable characteristic of CSA is its utilization of Levy flights, which are random walks characterized by a heavy-tailed distribution. This method regulates the movement of cuckoos, enabling efficient exploration of the feature space while swiftly navigating towards promising regions. Therefore, CSA is selected as a DR technique for its capability to balance exploration and exploitation and effectively manage high-dimensional data. The CSA commences its process by generating an initial population of cuckoos with random positions in the feature space. The position of each cuckoo in the population is then updated using the following equation [34]:
z k n e w = z k o l d + α L e v y ( λ )
where the cuckoo’s previous position is represented by z k o l d , while its updated position is denoted as z k n e w , where α is a scaling factor, and Levy(λ) represents a step drawn from a Levy distribution.
The cuckoos’ random search patterns are modeled by a concept called the Lévy flight, described by the following equation:
s = τ r 1 μ
Here, ‘ s ’ represents the step size, τ is a parameter that dictates the scale of the Levy flight, and shape of the Levy flight is determined by the parameter μ . A random number r chosen uniformly between 0 and 1. The optimal nest in the population is chosen based on the fitness value. MSE is used to derive the fitness value, and the best nest serves as the starting point for the next iteration of the algorithm. Also, abandon a fraction P a of the worst nests and generate new solutions to replace them. This process helps keep the search space varied, allowing the algorithm to explore new potential solutions. By using P a to periodically introduce new random solutions, the algorithm maintains a balance between exploration and exploitation, which is crucial for effective dimensionality reduction and optimization.

3.3. Dragonfly Algorithm

The Dragonfly Algorithm (DFA) is a heuristic optimization technique inspired by the fascinating static and dynamic swarming behaviors of dragonflies. DFA was developed by Seyedali Mirjalili in 2015 [35], and this algorithm mimics the way dragonflies search for food and avoid predators. Figure 6 presents the DFA flowchart. Dragonflies display two main behaviors: static, where they gather in small groups and hover around a target, and dynamic, where they collectively move towards a distant target. These behaviors are translated into mathematical models that help in exploring and exploiting the search space efficiently. It is effective for dimensionality reduction by finding optimal feature subsets through exploration and exploitation of the search space. As represented by Rahman et al. [36], separation, alignment, cohesion, attraction to food, and distraction from the enemy are important features of the DFA. In the following equations, q   represents the current position of an individual dragonfly, q j denotes the position of the j t h dragonfly, and ‘ M ’ indicates the total number of neighboring dragonflies.
Separation: It prevents dragonflies from crowding together by maintaining a minimum distance from each other.
S k = j = 1 M q q j
The symbol S k represents the motion of separation exhibited by the k t h individual.
Alignment: It refers to the tendency of dragonflies to match their velocity with that of their neighbors.
A k = j = 1 M V j M
where A k represents the alignment motion for the k t h individual, V j indicates j t h neighboring individual dragon velocity.
Cohesion: is the behavior that drives dragonflies to move towards the center of their neighboring individuals.
C k = j = 1 M q j M q  
Attraction: It is the tendency of dragonflies to move towards food sources
F k = q + q
where F k represents the attraction of nutrition source for the k t h individual flyand the position of the food source is denoted as q + .
Distraction: It is the tendency of dragonflies to move away from enemies.
E k = q + q
where E k represents the distraction motion caused by the enemy for the k t h individual and the location of the enemy is denoted as q .
The positions of artificial dragonflies inside the designated search area are revised using the current position vector ( q ) and the step vector (∆ q ). The direction of their movement is determined by the step vector (∆ q ), and it is calculated as,
q t + 1 = s S k + a A k + c C k + f F k + e E k + ω q t  
where iteration number is denoted as ‘ t ’, ‘ ω ’ denotes the inertia weight and weights assigned to separation, alignment, cohesion, attraction, and enemy, are denoted as   s ,   a ,   c , f , and   e respectively. The exploitation and exploration phases can be achieved by modifying the weights, once the step vector calculation is completed, calculation of the position vectors commences as below:
q t + 1 = q t + q t + 1
Table 2 demonstrates the selection of the optimal parameters for heuristic algorithms.
Statistical metrics such as Pearson Correlation Coefficient (PCC), Sample Entropy, and Canonical Correlation Analysis (CCA), mean, variance, kurtosis, and skewness were employed to identify distinct characteristics in the PPG signals between the different classes, allowing for faster analysis as presented in Table 3.
It is perceived from Table 3 that the calculated mean values are notably lower for both normal and CVD cases under ABC PSO DR technique and also for cuckoo search normal category. For cuckoo search CVD class, a higher mean value is attained. Under dragon fly optimization, negative mean values are achieved for both Normal and CVD. Skewness and kurtosis values are considerably skewed for both normal and CVD for all three DR techniques. Table 3 indicates that the sample entropy values are same across classes with the exception of the cuckoo search DR technique in the CVD case. Table 3 further illustrates that the low PCC values suggest that the optimized features exhibit nonlinearity and lack correlation across different classes. If CCA values exceed 0.5, significant correlation between the classes can be expected. Table 3 shows that the dragon fly optimization is highly related to all of the classes compare to other DR techniques. It also shows that the ABC PSO DR technique has a lower correlation with the other classes.
Figure 7 depicts the histogram plot of ABC PSO DR values for a normal person and it gives that the histogram with higher peaks in the middle and lower at either end. Figure 8 displays the histogram plot of ABC PSO DR values for a CVD patient, indicating that the histogram has a normal and Gaussian distribution. The scatter plot of ABC PSO DR values of normal and CVD data is depicted in Figure 9, which clearly indicates that the DR features are highly merged among the classes at the center and lesser overlapping at the end.The scatter plot of Dragon Fly DR values is portrayed in Figure 10. It shows that the lesser overlapping is present among the normal and CVD classes.

4. Classifiers for Classification of CVD from Dimensionality Reduced Values

4.1. Linear Regression as a Classifier

Linear regression (LR) falls under the category of supervised learning techniques. It is primarily employed to forecast continuous numerical outcomes. This algorithm establishes a linear relationship between input features and the target variable, allowing it to make predictions on new, unseen data based on the patterns learned from the training set. The linear equation is defined by a set of coefficients that are estimated using the training data. Linear regression is is primarily designed for predicting continuous numerical outcomes, it demonstrates versatility in its application. Through appropriate modifications, this algorithm can be effectively repurposed to address classification problems. The basic linear regression model is defined as follows [37]:
q = b 0 + b 1 x 1 + b 2 x 2 + + b n x n + ϵ
where x 1 , x 2 , …, x n are independent variables, q is the dependent variable, b 0 is the intercept, the coefficients are b 1 , b 2 , …, b n , and the error term is ϵ . For classification, the predicted y values can be thresholded to determine class membership. In this research, we set a threshold of 0.5 to classify PPG data as either normal or CVD. Predictions greater than 0.5 are classified as CVD, while those less than 0.5 are considered normal. This approach is simple but can be limited by linear regression’s assumptions and the nature of classification problems.

4.2. Linear Regression with BLDC

LR and LR-BLDC models use the same linear relationship. However, LR is used for regression tasks with continuous outputs, whereas LR-BLDC adapts this relationship for binary classification by applying a threshold to the predicted values. The Bayesian Linear Discriminant Classifier operates as a probabilistic generative model specifically tailored for classification challenges. It involves estimating the class-conditional probability distribution of the input variables for each class and using Bayes’ rule to calculate the posterior Probability of each class based on the input variables [38]. To combine linear regression and BLDC, the predicted output of the LR model is used as an input to the BLDC model. The linear regression classifier output is used to estimate the mean of the class-conditional probability distribution for each class in the BLDC model.

4.3. K-Nearest Neighbor as a Classifier

KNN (K-Nearest Neighbors) is a machine learning algorithm that classifies data based on the nearest neighbors. The parameter “K” represents the number of neighbors considered. The core idea is to classify a data point by finding the closest training points to it, based on a similarity measure [39]. After identifying the K number of nearest neighbors, it allots the input point to the class with the maximum frequency among its K-nearest neighbors. It does not need a different phase for training. Instead, the algorithm retains the complete training dataset and employs it to classify new data points. Weighted KNN is a classification algorithm that assigns weights to the neighbors based on their distance to the query point. In the KNN algorithm, for a given point z , the algorithm finds the K data points closest to a new point by measuring their distance using a metric like Euclidean distance [40]:
d z , z i = j = 1 m z j z i j 2
In weighted KNN, each neighbor z i is assigned a weight w i inversely proportional to its distance from z :
w i = 1 d z , z i + ε
where ε is a small constant to avoid division by zero. The algorithm predicts the class of the query point by considering the most frequent class among its K closest neighbors:
b ^ = arg m a x c i M K z w i . I ( b i = c )
where M K z denotes the set of K nearest neighbors, the class label of neighbor z i is b i and the indicator function is I (⋅). This method improves classification accuracy by giving more influence to closer neighbors. Figure 11 depicts the flow diagram of the KNN algorithm.

4.4. PCA-Firefly

The hybrid PCA firefly technique is used to choose the most relevant features while removing irrelevant ones, thereby optimizing accuracy to its fullest extent. The PPG dataset undergoes dimensionality reduction using the PCA algorithm, which effectively reduces the quantity of stochastic variables to a concise group of primary variables. This approach greatly enhances the accuracy of the prediction results [41]. To further optimize the process, the firefly optimization algorithm is utilized to select the most appropriate attributes from this refined reduced dataset.
PCA simplifies data by finding a new set of features that are not related to each other by solving the following eigenvalue problem:
Y T Y v = λ v
where Y is the data matrix, v   are the eigenvectors (principal components), and λ are the eigenvalues.
The Firefly Algorithm is a clever problem-solving method that mimics how fireflies communicate with their flashes, optimizes the selection of principal components. Each firefly represents a potential solution with an intensity I proportional to its fitness. The attractiveness α between fireflies i and j is given by the following [42]:
α i j = α 0 e γ p i j 2
where α 0 is the attractiveness at p = 0, γ   is the light absorption coefficient, and p i j is the distance between fireflies i and j . Fireflies are drawn to brighter ones, constantly adjusting their location:
x i = x i + α i j x j x i + β ϵ i
where β   is a randomization parameter, and ϵ i is a random vector. This process iteratively refines the feature selection, enhancing classification performance. Figure 12 presents the workflow of the PCA-Firefly algorithm.

4.5. Linear Discriminant Analysis as a Classifier

Linear Discriminant Analysis (LDA) is a dimensionality reduction technique used for classification, aiming to find the best way to combine features to distinguish between different groups. LDA assumes that each class follows a Gaussian distribution with a shared covariance matrix. The steps of LDA are as follows [43]:
1. Calculate the mean vector for each class
μ z = 1 M z i C z x i
2. Calculate the within-class scatter matrix
S w = z = 1 M i C z x i μ z x i μ z T
3. Calculate the between-class scatter matrix
S b = z = 1 M M z μ z μ μ z μ T
where μ is the overall mean of the dataset.
4. Solve the generalized eigenvalue problem
S w 1 S b v = λ v
The eigenvectors v corresponding to the largest eigenvalues λ form the transformation matrix. Project the data onto this lower-dimensional space to maximize class separability. In classification, the new data point x is projected and assigned to the class with the closest mean in this reduced space.

4.6. Kernel LDA as a Classifier

Kernel Linear Discriminant Analysis (KLDA) is an extension of LDA that incorporates the kernel trick to handle nonlinear data. The objective of this technique is to identify an optimal projection of the data in a high-dimensional feature space, which is created through the application of a kernel function. This projection is designed to simultaneously maximize the separation between different classes (between-class scatter) and minimize the spread within each individual class (within-class scatter).
The objective function of KLDA is formulated as follows [44]:
m a x v v T S b v v T S w v
where S b is the between-class scatter matrix and S w is the within-class scatter matrix in the kernel space. The solution v corresponds to the eigenvector associated with the largest eigenvalue of the generalized eigenvalue problem S w v = λ S b v . KLDA is effective for nonlinear data leveraging a kernel function to implicitly project the original input space into a higher-dimensional feature space, where data separation is potentially easier to achieve. It has found applications in various fields where nonlinear relationships among data features are prevalent.

4.7. Probabilistic LDA as a Classifier

Probabilistic Linear Discriminant Analysis (ProbLDA) is an extension of Linear Discriminant Analysis (LDA) that models class distributions probabilistically. ProbLDA assumes that each class follows a Gaussian distribution and utilizes Bayesian principles for classification [45]. The class mean, within-class scatter matrix and between-class scatter matrix are computed as per the Equations (21)–(23). Then, for each class z Gaussian distributions can be assumed as follows:
p x C z = 1 2 π d 2 Ʃ 1 2 e x p 1 2 x μ z T Ʃ 1 x μ z
where   Ʃ is the Shared covariance matrix estimated as S w .
To calculate the posterior probabilities for classification, Bayes’ theorem can be used as per the below equation:
p C z x = p x C z   p C z p ( x )
Here p ( C z ) prior probability of class z and p ( x ) is the marginal probability of x . Assign the new data point x to the class with the highest posterior probability.
g ^ = a r g   m a x z   p C z x
ProbLDA integrates the strengths of LDA with probabilistic modeling, providing a robust framework for classification.

4.8. Support Vector Machine as a Classifier

The Support Vector Machine (SVM) is a sophisticated supervised learning technique primarily employed for classification purposes.
SVM treats the input data as an n-dimensional feature vector space and seeks to find an (n − 1) dimensional hyperplane that divides the space into two regions. This hyperplane is positioned to maximize the minimum distance between any data point and the boundary [46].The n -dimensional input data x j   (where j = 1, 2, ….., N, with N representing the number of samples) is assigned labels y j = 1   for class 1 and y j = 1   for class 2 using the y j matrix. For linearly separable data, a hyperplane can be defined as f x = 0 [47].
f x = w . x + d = j = 1 n w j x j + d = 0
w is an   n -dimensional vector, and d is a scalar. These parameters define the position of the hyperplane, establishing clear boundaries for classification. In SVM, input data are mapped to a higher-dimensional feature space using kernel functions to address multiclass classification problems. SVM aims to achieve high generalization by effectively separating classes based on the training data.
In this research, SVM with three different kernels are used: Radial Basis Function (RBF), linear, and polynomial [48].
SVM-Linear:
ker z i , z j = z i T z j
SVM-Polynomial:
Ker z i , z j = 1 + ϒ z i T . z j q ,   ϒ > 0
where the degree of polynomial kernel is denoted as q and ‘ϒ’ denotes the gamma term in the kernel function.
SVM-RBF:
ker z i , z j = exp ϒ z i z j 2 , ϒ > 0
where, z i z j is the Euclidean distance between two input vectors   z i and   z j .
These kernel functions operate by projecting the input data into a space of increased dimensionality. This transformation facilitates the identification of an optimal hyperplane that effectively delineates between different classes.

5. Results and Discussion

5.1. Training and Testing of the Classifiers

This research work employed a data partitioning strategy wherein 90% of the available dataset was dedicated to the model’s training process, while the remaining 10% was set aside for subsequent testing and validation. To assess the classifier’s efficacy and reliability, we implemented a ten-fold cross-validation approach. A True Positive (TP) occurs when the classifier correctly identifies a positive sample, while a True Negative (TN) is when it accurately labels a negative sample. A False Positive (FP) occurs when the model erroneously classifies a negative instance as positive, whereas a False Negative (FN) happens when the model incorrectly identifies a positive instance as negative. The Mean Squared Error ( M S E ) is defined by the following mathematical formula:
M S E = 1 M k = 1 M O k T i 2
where M denotes the complete set of data points within the PPG dataset, and it is assumed as 1000. The target value of model i is designated by T i , where the range of i varies from 1 to 15; at a specific time, O k represents the observed value. The training was conducted in a manner that significantly reduced the classifier’s Mean square error to a minimal value.
Table 4 presents the testing and training MSE for classifiers using three different DR techniques. The training MSE ranges from 10−5 to 10−10, in contrast, the testing phase yielded MSE values ranging from 10−3 to 10−9. The SVM-RBF classifier under the ABC-PSO dimensionality reduction method achieved the lowest training of 1.92 × 10−10 and testing MSE value of 2.45 × 10−9. The DFA DR method yields somewhat reduced MSE values for both training and testing phases across the classifiers, when compared to the alternative DR approaches evaluated in this work. Similarly, for Cuckoo Search DR features the LR classifier yields the minimal testing MSE value of 1.37 × 10−8, due the correct labeling of PPG signal for both CVD and normal subjects. The PCA Firefly classifier is plugged with a high number of FP and FN cases, resulting in the highest testing MSE of 6.08 × 10−3. Similarly, for Dragon Fly DR values, the SVM-RBF classifier achieves the minimum testing Mean square error of 3.62 × 10−9. This superior performance is attributed to its accurate labeling of PPG signals for both CVD and normal subjects. Conversely, the SVM-Linear classifier shows poorer performance, resulting in a much higher testing MSE of 9.03 × 10−3. This higher error rate is due to the SVM-Linear misclassifying a significant number of cases, producing both false negatives and false positives.
Figure 13 displays the training performance of the ten classifiers using the ABC-PSO dimensionally reduced PPG signals as input. As shown in Figure 13, the maximum number of iterations is set to 500. The Logistic Regression, SVM-Polynomial and SVM- RBF Classifiers reached the minimum MSE value at 50 iterations itself. All the other classifiers are settled at minimum MSE values after 250 iterations. The LDA, SVM-linear and PLDA classifiers depict a distinctive performance apart from all the other classifiers up to 350 iterations.
Figure 14 displays the training performance of the ten classifiers for the cuckoo search dimensionally reduced PPG signals as the input. As exhibited in Figure 14, the maximum number of iterations is fixed at 500. The Linear Regression, LR-BLDC and KLDA Classifiers reached the minimum MSE value at 50 iterations itself. All the other classifiers are settled at minimum MSE values after 300 iterations. The LDA, PLDA and PCA-Firefly classifiers depict a peculiar performance apart from all the other classifiers up to 350 iterations.
Figure 15 explores the training MSE performance of the ten classifiers for the dragonfly dimensionally reduced PPG signals as the input. As mentioned in Figure 15, the maximum number of iterations is fixed at 500. The LR-BLDC and SVM-Linear classifiers reached the minimum MSE value at 50 iterations itself. All the other classifiers are settled at minimum MSE values after 300 iterations. The PLDA, SVM-linear and SVM-Polynomial classifiers are depicting a unique performance apart from all the other classifiers up to 350 iterations.

5.2. Optimal Parameters Selection for Classifiers

When selecting the target values for the binary classification of PPG dataset (CVD and Normal class), a deliberate choice is made to assign the target ( T C V D )   values towards the upper end of the 0 to 1 range. The criteria used to select T C V D is as follows:
1 Y p = 1 Y μ p T C V D
The complete set of CVD PPG data features, denoted as ( Y ), underwent a normalization process with mean μ p . For the target T N o r m a l values, a deliberate choice is made to assign them towards the lower end of the 0 to 1 range. The selection criteria for determining the value of T N o r m a l are governed by the following parameters:
1 X q = 1 X μ q T N o r m a l
The complete set of normal PPG data features, denoted as ( X ), underwent a normalization process with mean μ p . For optimal categorization, the following equation is used:
T C V D T N o r m a l 0.5
Based on the condition provided in (36), in this research work, the targets have been set at 0.1 for normal and 0.85 for CVD. Table 5 details the iterative process of selecting optimal parameters for the classifier during its training phase. A maximum of 1000 iterations is allowed to control the convergence criteria.

5.3. Performance Analysis of the Classifier

The classifiers’ effectiveness is assessed using a comprehensive set of metrics, such as Accuracy, Good Detection Rate (GDR), F1 Score, Kappa, Matthews Correlation Coefficient (MCC), Error rate, and Jaccard Index. The following formula [49,50] used for evaluating the overall effectiveness of the classification method.
A c c u r a c y = T N + T P T P + T N + F P + F N × 100 %
G o o d   d e t e c t i o n   r a t e   ( G D R ) = T N + T P F N ( T N + T P ) + F P × 100 %
E r r o r   R a t e = F P + F N T P + T N + F P + F N × 100 %
K a p p a = T P + T N 100 E a c c 1 E a c c
where, E a c c = ( ( ( T P + F P ) / 100 ) ( T P + F N ) / 100 + ( ( ( T N + F P ) / 100 ) ( ( T N + F N ) / 100 )
M C C = T N . T P ( F N . F P ) T N + F N . T N + F P T P + F N . T P + F P × 100 %
F 1   S c o r e = 2 T P 2 T P + F P + F N × 100 %
J a c c a r d   I n d e x = T P T P + F P + F N × 100 %
Table 6 presents the results of the performance analysis for classifiers with three dimensionality reduction techniques. The results in Table 6 reveal that the classifier SVM-RBF for ABC PSO dimensionality reduction technique achieved very high scores on all benchmark metrics, including a superior accuracy of 95.12%, a F1 score and GDR of 95%, and a lowest error rate of 4.88%. Furthermore, the high Kappa and MCC are 0.90, with a Jaccard Index of 98.48%. The result demonstrates that the SVM-RBF classifier is outperforming for the ABC-PSO dimensionality reduction technique. Conversely, the KLDA classifier for the ABC-PSO dimensionality reduction method and the PCA-Firefly classifier for the Cuckoo Search dimensionality reduction technique demonstrated poor performance across all parameter values. This was evident from the lowest accuracy and F1 score of 53.66%, a GDR of 38.71%, and the highest error rate of 46.34%. Also, the low Kappa and MCC are 0.07, with a Jaccard Index of 36.67%. Upon analyzing individual dimensionality reduction techniques, the SVM (RBF) classifier performs better than other classifiers for the ABC-PSO DR technique, with a F1 score and GDR of 95% and a Jaccard Index of 98.48%. On the other hand, KLDA is the lowest-performing classifier due to a higher error rate of 46.34% and reduced Kappa and MCC values of 0.07. Similarly, for the CSA DR method, the linear regression classifier achieved better accuracy at 90.24%, with a strong F1 score of 90% and a lower error rate of 9.76%. Meanwhile, for the Dragon Fly DR technique, the SVM-RBF classifier attained a higher accuracy of 92.68%, a GDR of 92.50%, and a high Matthews correlation coefficient of 0.85 due to a lower error rate of 7.32%.
Figure 16 illustrates the comparison of accuracy performance among classifiers using ABC PSO, Cuckoo Search, and Dragon Fly DR techniques.
Figure 16 presents the performance evaluation of classifiers across various DR techniques based on accuracy. Figure 16 reveals that the SVM-RBF classifier reigns supreme with an accuracy of 95.12%. Conversely, the KLDA classifier yields the lowest accuracy, at 53.66% for the ABC PSO DR method. Similarly, with CSA dimensionality reduced values, the LR classifier achieves the highest accuracy of 90.24%, while PCA-Firefly exhibits the lowest accuracy at 53.66%. The SVM-RBF classifier exhibited robust performance, achieving a notable accuracy of 92.68% when paired with the Dragonfly DR method. In contrast, the SVM with linear kernel struggled comparatively, yielding a considerably lower accuracy of 58.54%. Additionally, Figure 16 shows that the LR classifier performed well across all three DR techniques, achieving an accuracy of 90.24%.
Figure 17 shows that all classifiers using the three different DR techniques achieve an F1 score of around 60% or higher, indicating they are performing significantly better than random guessing. This indicates a notable alignment between the model’s forecasts and the true class labels. It is also perceived that the maximum error rate observed is 46%, while the highest F1 score is 95%.
Figure 18 presents the histogram of accuracy and the Jaccard index for the classifiers. It is perceived that the maximum accuracy reaches 95%, while the highest Jaccard index is 90%. The accuracy histogram is left-skewed, indicating that the classifier’s accuracy does not drop below 50% for any of the DR techniques. The Jaccard index histogram spans the entire spectrum, with values below 50% attributed to the classifiers’ high false-positive rates.

5.4. Analysis of the Computational Complexity of Classifiers

Computational complexity is an important performance metric for classifiers. It is evaluated by considering the input size, denoted as m . The computational complexity remains very low when the input size is O ( 1 ) . However, the computational complexity increases with the number of inputs. In this research, the computational complexity is independent of input size, which is a highly desirable characteristic for any algorithm. The term O ( log m ) denotes the logarithmic rise in computational complexity with respect to m .
The computational complexity of classifiers across different DR techniques is presented in Table 7. From Table 7, the results clearly demonstrate that the SVM-RBF classifier has the highest computational complexity when used with the ABC-PSO and dragonfly DR techniques. Consequently, SVM-RBF achieves the highest accuracy, with 95.12% for the ABC-PSO dimensionality reduction technique and 92.68% for the dragonfly DR technique. Conversely, the KLDA classifier for the ABC-PSO dimensionality reduction technique and the PCA-Firefly classifier for the cuckoo search DR values produce the lowest accuracy of 53.66% while also having high computational complexity. This is attributed to their high false-positive rate and low Jaccard index.

5.5. Limitations

Table 8 compares the results of various machine learning methods used for CVD classification from PPG data. It explores the potential of using PPG data for the early identification of CVD and the prediction of associated diseases. The analysis highlighted promising classification techniques that could be beneficial for screening and identifying CVD patients. The main limitation of this work is that the PCA Firefly classifier provides the lowest average classification accuracy across all three optimization techniques, leading to an increased likelihood of false alarms in second-to-second detection of PPG signals. At the same time, 30 s segmented epochs of PPG signals were used to improve the classification accuracy of the classifiers. However, this could lead to overfitting during training, resulting in artificially high accuracy. To address this, a compromise was made by using one-minute segments of raw PPG signals to achieve more reliable classification accuracy.
As seen in Table 8, various machine learning classifiers, such as LR, NB, RBF NN, DCNN, KNN, DNN, ELM, ANN, and SVM (RBF) have been utilized for classification of CVD from clinical database. The performance spectrum of these classifiers spans from a moderate 65% to an impressive 95% in terms of accuracy. However, this study specifically targets CVD detection using Capnobase dataset, with SVM (RBF) achieving the highest accuracy of 95.12%.

6. Conclusions

The main objective of this work is to classify PPG signals as either normal or indicative of cardiovascular disease. High-quality PPG samples were obtained by extracting useful features using heuristic-based DR methods such as ABC-PSO, Cuckoo Search, and Dragonfly techniques. Ten classifiers were used for this purpose: LR, LR-BLDC, KNN, PCA-Firefly, LDA, KLDA, ProbLDA, SVM-Linear, SVM-Polynomial, and SVM-RBF. The SVM-RBF classifier, combined with the hybrid ABC-PSO dimensionality reduction technique, demonstrated superior performance. This approach achieved a remarkable accuracy of 95.12% while maintaining a minimal error rate of just 4.88%. Furthermore, it exhibited robust reliability, as evidenced by its high Matthews correlation coefficient and Kappa values, both reaching 0.90. Additionally, the second highest accuracy of 92.62% was achieved by the SVM-RBF classifier for Dragon Fly optimized values. The third highest accuracy of 90.24% was obtained by the LR classifier across all three DR techniques. Future research is focusing on convolutional neural networks (CNNs) and deep neural networks (DNNs) to swiftly detect cardiovascular disease.

Author Contributions

Conceptualization, S.P.; methodology, S.P.; software, S.P.; validation, H.R. and S.P.; formal analysis, H.R.; investigation, H.R. and S.P.; resources, H.R. and S.P.; data curation, H.R.; writing—original draft preparation, S.P.; writing—review and editing, H.R.; visualization, H.R. and S.P.; supervision, H.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Centers for Disease Control and Prevention. The State of Aging and Health in America 2013; Centers for Disease Control and Prevention, Ed.; US Department of Health and Human Services: Atlanta, GA, USA, 2013.
  2. Kranjec, J.; Beguš, S.; Geršak, G.; Drnovšek, J. Non-contact heart rate and heart rate variability measurements: A review. Biomed. Signal Process. Control 2014, 13, 102–112. [Google Scholar] [CrossRef]
  3. Miranda, E.; Irwansyah, E.; Amelga, A.Y.; Maribondang, M.M.; Salim, M. Detection of cardiovascular disease risk’s level for adults using naive Bayes classifier. Healthc. Inform. Res. 2016, 22, 196. [Google Scholar] [CrossRef]
  4. Majumder, S.; Mondal, T.; Deen, M.J. Wearable sensors for remote health monitoring. Sensors 2017, 17, 130. [Google Scholar] [CrossRef]
  5. Manogaran, G.; Shakeel, P.M.; Fouad, H.; Nam, Y.; Baskar, S.; Chilamkurti, N.; Sundarasekar, R. Wearable IoT smart-log patch: An edge computing-based bayesian deep learning network system for multi access physical monitoring system. Sensors 2019, 19, 3030. [Google Scholar] [CrossRef] [PubMed]
  6. Norgeot, B.; Glicksberg, B.S.; Butte, A.J. A call for deep-learning healthcare. Nat. Med. 2019, 25, 14–15. [Google Scholar] [CrossRef]
  7. Al-Zaben, A.; Fora, M.; Obaidat, A. Detection of premature ventricular beats from arterial blood pressure signal. In Proceedings of the 2018 IEEE 4th Middle East Conference on Biomedical Engineering (MECBME), IEEE, Tunis, Tunisia, 28–30 March 2018; pp. 17–19. [Google Scholar]
  8. Hindia, M.N.; Rahman, T.A.; Ojukwu, H.; Hanafi, E.B.; Fattouh, A. Enabling remote health-caring utilizing iot concept over LTE-femtocell networks. PLoS ONE 2016, 11, e0155077. [Google Scholar] [CrossRef]
  9. Savkar, A.; Khatate, P.; Patil, C. Study on techniques involved in tourniqueteless blood pressure measurement using PPG. In Proceedings of the 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS), IEEE, Madurai, India, 14–15 June 2018; pp. 170–172. [Google Scholar]
  10. Elgendi, M.; Fletcher, R.; Liang, Y.; Howard, N.; Lovell, N.H.; Abbott, D.; Lim, K.; Ward, R. The use of photoplethysmography for assessing hypertension. npj Digit. Med. 2019, 2, 60. [Google Scholar] [CrossRef]
  11. Allen, J. Photoplethysmography and its application in clinical physiological measurement. Physiol. Meas. 2007, 28, R1–R39. [Google Scholar] [CrossRef]
  12. Gu-Young, J.; Yu, K.-H.; Nam-Gyun, K. Continuous blood pressure monitoring using pulse wave transit time. In Proceedings of the ICCAS 2005 International Conference on Control, Automation and Systems, Gyeonggi-Do, Republic of Korea, 2–5 June 2005; pp. 834–837. [Google Scholar]
  13. Shin, H.; Min, S. Feasibility study for the non-invasive blood pressure estimation based on ppg morphology: Normotensive subject study. Biomed. Eng. Online 2017, 16, 10. [Google Scholar] [CrossRef]
  14. Moraes, J.L.; Rocha, M.X.; Vasconcelos, G.G.; Filho, J.E.; de Albuquerque, V.H.C.; Alexandria, A.R. Advances in photopletysmography signal analysis for biomedical applications. Sensors 2018, 18, 1894. [Google Scholar] [CrossRef]
  15. Shintomi, A.; Izumi, S.; Yoshimoto, M.; Kawaguchi, H. Effectiveness of the heartbeat interval error and compensation method on heart rate variability analysis. Healthc. Technol. Lett. 2022, 9, 9–15. [Google Scholar] [CrossRef]
  16. Rubins, U. Finger and ear photoplethysmogram waveform analysis by fitting with Gaussians. Med. Biol. Eng. 2008, 46, 1271–1276. [Google Scholar] [CrossRef] [PubMed]
  17. Moshawrab, M.; Adda, M.; Bouzouane, A.; Ibrahim, H.; Raad, A. Smart Wearables for the Detection of Cardiovascular Diseases: A Systematic Literature Review. Sensors 2023, 23, 828. [Google Scholar] [CrossRef] [PubMed]
  18. Reisner, A.; Shaltis, P.A.; McCombie, D.; Asada, H.H. Utility of the photoplethysmogram in circulatory monitoring. Anesthesiology 2008, 108, 950–958. [Google Scholar] [CrossRef] [PubMed]
  19. Shivahare, B.D.; Singh, J.; Ravi, V.; Chandan, R.R.; Alahmadi, T.J.; Singh, P.; Diwakar, M. Delving into Machine Learning’s Influence on Disease Diagnosis and Prediction. Open Public Health J. 2024, 17, e18749445297804. [Google Scholar] [CrossRef]
  20. Ihsan, M.F.; Mandala, S.; Pramudyo, M. Study of Feature Extraction Algorithms on Photoplethysmography (PPG) Signals to Detect Coronary Heart Disease. In Proceedings of the 2022 International Conference on Data Science and Its Applications (ICoDSA), Bandung, Indonesia, 6 July 2022; pp. 300–304. [Google Scholar] [CrossRef]
  21. Pal, P.; Ghosh, S.; Chattopadhyay, B.P.; Saha, K.K.; Mahadevappa, M. Screening of ischemic heart disease based on PPG signals using machine learning techniques. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), IEEE, Montreal, QC, Canada, 20–24 July 2020; pp. 5980–5983. [Google Scholar]
  22. Kanawade, R.; Tewary, S.; Sardana, H.K. Photoplethysmography based arrhythmia detection and classification. In Proceedings of the 6th International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 7–8 March 2019; pp. 944–948. [Google Scholar] [CrossRef]
  23. Paradkar, N.; Chowdhury, S.R. Coronary artery disease detection using photoplethysmography. In Proceedings of the 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju, Republic of Korea, 11–15 July 2017; pp. 100–103. [Google Scholar] [CrossRef]
  24. Banerjee, R.; Vempada, R.; Mandana, K.M.; Choudhury, A.D.; Pal, A. Identifying coronary artery disease from photoplethysmogram. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Adjunct, Heidelberg, Germany, 12–16 September 2016; pp. 1084–1088. [Google Scholar]
  25. Xing, X.; Sun, M. Optical blood pressure estimation with photoplethysmography and FFT-based neural networks. Biomed. Opt. Express 2016, 7, 3007–3020. [Google Scholar] [CrossRef] [PubMed]
  26. Chakraborty, A.; Sadhukhan, D.; Pal, S.; Mitra, M. Automated myocardial infarction identification based on interbeat variability analysis of the photoplethysmographic data. Biomed. Signal Process. Control 2020, 57, 101747. [Google Scholar] [CrossRef]
  27. Mangathayaru, N.; Padmaja Rani, B.; Janaki, V.; Patel, S.A.; Sai Mohan, G.; Lalith Bharadwaj, B. An imperative diagnostic framework for PPG signal classification using GRU. In Proceedings of the International Conference on Advanced Informatics for Computing Research, Springer, Singapore, 26–27 December 2020; pp. 606–621. [Google Scholar]
  28. Prabhakar, S.K.; Rajaguru, H.; Lee, S.W. Metaheuristic-based dimensionality reduction and classification analysis of PPG signals for interpreting cardiovascular disease. IEEE Access 2019, 7, 165181–165206. [Google Scholar] [CrossRef]
  29. Tjahjadi, H.; Ramli, K.; Murfi, H. Noninvasive classification of blood pressure based on photoplethysmography signals using bidirectional long short-term memory and time-frequency analysis. IEEE Access 2020, 8, 20735–20748. [Google Scholar] [CrossRef]
  30. Karlen, W.; Turner, M.; Cooke, E.; Dumont, G.; Ansermino, J.M. CapnoBase: Signal database and tools to collect, share and annotate respiratory signals. In Proceedings of the 2010 Annual Meeting of the Society for Technology in Anesthesia, West Palm Beach, FL, USA, 13–16 January 2010; Society for Technology in Anesthesia: Milwaukee, WI, USA, 2010; p. 27. [Google Scholar]
  31. Rajaguru, H.; Prabhakar, S.K. Bayesian linear discriminant analysis with hybrid ABC-PSO classifier for classifying epilepsy from EEG signals. In Proceedings of the 2017 International Conference on Computing Methodologies and Communication (ICCMC), IEEE, Erode, India, 18–19 July 2017; pp. 977–980. [Google Scholar]
  32. Bharanidharan, N.; Rajaguru, H. Performance analysis of particle swarm optimization technique in classification of dementia using MRI images. In Proceedings of the 2017 2nd International Conference on Communication and Electronics Systems (ICCES), IEEE, Coimbatore, India, 19–20 October 2017; pp. 311–315. [Google Scholar]
  33. Yang, X.-S.; Suash, D. Cuckoo search via Lévy flights. In Proceedings of the 2009 World Congress on Nature & Biologically Inspired Computing (NaBIC), IEEE, Coimbatore, India, 9–11 December 2009; pp. 210–214. [Google Scholar]
  34. Sannasi Chakravarthy, S.R.; Rajaguru, H. Comparison analysis of linear discriminant analysis and cuckoo-search algorithm in the classification of breast cancer from digital mammograms. Asian Pac. J. Cancer Prev. APJCP 2019, 20, 2333. [Google Scholar]
  35. Mirjalili, S. Dragonfly algorithm: A new meta-heuristic optimization technique for solving single-objective, discrete, and multi-objective problems. Neural Comput. Appl. 2016, 27, 1053–1073. [Google Scholar] [CrossRef]
  36. Rahman, C.M.; Rashid, T.A. Dragonfly algorithm and its applications in applied science survey. Comput. Intell. Neurosci. 2019, 2019, 9293617. [Google Scholar] [CrossRef] [PubMed]
  37. Hasan, M.A.; Hasan, M.K.; Mottalib, M.A. Linear regression–based feature selection for microarray data classification. Int. J. Data Min. Bioinform. 2015, 11, 167–179. [Google Scholar] [CrossRef] [PubMed]
  38. Zhou, W.; Liu, Y.; Yuan, Q.; Li, X. Epileptic seizure detection using lacunarity and Bayesian linear discriminant analysis in intracranial EEG. IEEE Trans. Biomed. Eng. 2013, 60, 3375–3381. [Google Scholar] [CrossRef] [PubMed]
  39. Sunnetci, K.M.; Ulukaya, S.; Alkan, A. Periodontal bone loss detection based on hybrid deep learning and machine learning models with a user-friendly application. Biomed. Signal Process. Control 2022, 77, 103844. [Google Scholar]
  40. Pandey, A.; Jain, A. Comparative analysis of KNN algorithm using various normalization techniques. International Journal of Computer Network and Information Security 2017, 10, 36. [Google Scholar] [CrossRef]
  41. Nirmalakumari, K.; Rajaguru, H.; Rajkumar, P. Performance analysis of classifiers for colon cancer detection from dimensionality reduced microarray gene data. Int. J. Imaging Syst. Technol. 2020, 30, 1012–1032. [Google Scholar] [CrossRef]
  42. Moazenzadeh, R.; Mohammadi, B.; Shamshirband, S.; Chau, K.W. Coupling a firefly algorithm with support vector regression to predict evaporation in northern Iran. Eng. Appl. Comput. Fluid Mech. 2018, 12, 584–597. [Google Scholar] [CrossRef]
  43. Zhao, H.; Lai, Z.; Leung, H.; Zhang, X. Linear Discriminant Analysis. In Feature Learning and Understanding. Information Fusion and Data Science; Springer: Cham, Germany, 2020. [Google Scholar] [CrossRef]
  44. Li, S.; Zhang, H.; Ma, R.; Zhou, J.; Wen, J.; Zhang, B. Linear discriminant analysis with generalized kernel constraint for robust image classification. Pattern Recognit. 2023, 136, 109196. [Google Scholar] [CrossRef]
  45. Ioffe, S. Probabilistic Linear Discriminant Analysis. Computer Vision—ECCV 2006. In Lecture Notes in Computer Science; Leonardis, A., Bischof, H., Pinz, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; Volume 3954. [Google Scholar] [CrossRef]
  46. Günay, M.; Alkan, A. Classification of EMG signals by spectral methods and SVM classifier. Kahramanmaraş Sütçü İmam Üniversitesi Mühendislik Bilim. Dergisi 2016, 13, 63–80. [Google Scholar] [CrossRef]
  47. Alkan, A. Analysis of knee osteoarthritis by using fuzzy c-means clustering and SVM classification. Sci. Res. Essays 2011, 6, 4213–4219. [Google Scholar]
  48. Ramamoorthy, K.; Rajaguru, H. Exploitation of Bio-Inspired Classifiers for Performance Enhancement in Liver Cirrhosis Detection from Ultrasonic Images. Biomimetics 2024, 9, 356. [Google Scholar] [CrossRef] [PubMed]
  49. Palanisamy, S.; Rajaguru, H. Machine learning techniques for the performance enhancement of multiple classifiers in the detection of cardiovascular disease from PPG signals. Bioengineering 2023, 10, 678. [Google Scholar] [CrossRef]
  50. Chellappan, D.; Rajaguru, H. Machine Learning Meets Meta-Heuristics: Bald Eagle Search Optimization and Red Deer Optimization for Feature Selection in Type II Diabetes Diagnosis. Bioengineering 2024, 11, 766. [Google Scholar] [CrossRef] [PubMed]
  51. Rajaguru, H.; Shankar, M.G.; Nanthakumar, S.P.; Murugan, I.A. Performance analysis of classifiers in detection of CVD using PPG signals. In AIP Conference Proceedings; AIP Publishing LLC: Melville, NY, USA, 2023; Volume 2725, p. 020002. [Google Scholar] [CrossRef]
  52. Al Fahoum, A.S.; Abu Al-Haija, A.O.; Alshraideh, H.A. Identification of Coronary Artery Diseases Using Photoplethysmography Signals and Practical Feature Selection Process. Bioengineering 2023, 10, 249. [Google Scholar] [CrossRef] [PubMed]
  53. Prabhakar, S.K.; Rajaguru, H.; Kim, S.H. Fuzzy-inspired photoplethysmography signal classification with bioinspired optimization for analyzing cardiovascular disorders. Diagnostics 2020, 10, 763. [Google Scholar] [CrossRef]
  54. Liu, Z.; Zhou, B.; Jiang, Z.; Chen, X.; Li, Y.; Tang, M.; Miao, F. Multiclass Arrhythmia Detection and Classification from Photoplethysmography Signals Using a Deep Convolutional Neural Network. J. Am. Heart Assoc. 2022, 11, e023555. [Google Scholar] [CrossRef]
  55. Hosseini, Z.S.; Zahedi, E.; Attar, H.M.; Fakhrzadeh, H.; Parsafar, M.H. Discrimination between different degrees of coronary artery disease using time-domain features of the finger photoplethysmogram in response to reactive hyperemia. Biomed. Signal Process. Control 2015, 18, 282–292. [Google Scholar] [CrossRef]
  56. Miao, K.H.; Miao, J.H. Coronary heart disease diagnosis using deep neural networks. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 1–8. [Google Scholar] [CrossRef]
  57. Shobitha, S.; Sandhya, R.; Ali, M.A. Recognizing cardiovascular risk from photoplethysmogram signals using ELM. In Proceedings of the Second International Conference on Cognitive Computing and Information Processing (CCIP), Mysore, India, 12–13 August 2016; pp. 1–5. [Google Scholar] [CrossRef]
  58. Soltane, M.; Ismail, M.; Rashid, Z.A. Artificial Neural Networks (ANN) approach to PPG signal classification. Int. J. Comput. Inf. Sci. 2004, 2, 58–65. [Google Scholar]
Figure 1. PPG signals for Normal Subject.
Figure 1. PPG signals for Normal Subject.
Diagnostics 14 02287 g001
Figure 2. PPG signals for CVD Subject.
Figure 2. PPG signals for CVD Subject.
Diagnostics 14 02287 g002
Figure 3. Detailed illustration of the workflow.
Figure 3. Detailed illustration of the workflow.
Diagnostics 14 02287 g003
Figure 4. Flowchart of the ABC-PSO Algorithm.
Figure 4. Flowchart of the ABC-PSO Algorithm.
Diagnostics 14 02287 g004
Figure 5. Flowchart of the Cuckoo Search Algorithm.
Figure 5. Flowchart of the Cuckoo Search Algorithm.
Diagnostics 14 02287 g005
Figure 6. Flow Chart representation of the Dragonfly Algorithm.
Figure 6. Flow Chart representation of the Dragonfly Algorithm.
Diagnostics 14 02287 g006
Figure 7. Histogram of ABC PSO DR PPG signals for Normal Person.
Figure 7. Histogram of ABC PSO DR PPG signals for Normal Person.
Diagnostics 14 02287 g007
Figure 8. Histogram of ABC PSO DR PPG signals for CVD Patient.
Figure 8. Histogram of ABC PSO DR PPG signals for CVD Patient.
Diagnostics 14 02287 g008
Figure 9. Scatter Plot of ABC PSO based DR values of PPG signals.
Figure 9. Scatter Plot of ABC PSO based DR values of PPG signals.
Diagnostics 14 02287 g009
Figure 10. Scatter Plot of Dragon Fly based DR values of PPG signals.
Figure 10. Scatter Plot of Dragon Fly based DR values of PPG signals.
Diagnostics 14 02287 g010
Figure 11. Flow diagram of the KNN algorithm.
Figure 11. Flow diagram of the KNN algorithm.
Diagnostics 14 02287 g011
Figure 12. Workflow of the PCA-Firefly algorithm.
Figure 12. Workflow of the PCA-Firefly algorithm.
Diagnostics 14 02287 g012
Figure 13. Training MSE of the Classifiers for ABC-PSO DR Techniques.
Figure 13. Training MSE of the Classifiers for ABC-PSO DR Techniques.
Diagnostics 14 02287 g013
Figure 14. Training MSE of the Classifiers for Cuckoo Search DR Techniques.
Figure 14. Training MSE of the Classifiers for Cuckoo Search DR Techniques.
Diagnostics 14 02287 g014
Figure 15. Training MSE of the Classifiers for Dragonfly DR Techniques.
Figure 15. Training MSE of the Classifiers for Dragonfly DR Techniques.
Diagnostics 14 02287 g015
Figure 16. Accuracy performance across various classifiers with DR Techniques.
Figure 16. Accuracy performance across various classifiers with DR Techniques.
Diagnostics 14 02287 g016
Figure 17. Significance of Error Rate and F1 Score Performance of Classifiers for different DR Techniques.
Figure 17. Significance of Error Rate and F1 Score Performance of Classifiers for different DR Techniques.
Diagnostics 14 02287 g017
Figure 18. Significance of Accuracy and Jaccard Index Performance of Classifiers.
Figure 18. Significance of Accuracy and Jaccard Index Performance of Classifiers.
Diagnostics 14 02287 g018
Table 1. Feature extraction and its limitations in previous related work.
Table 1. Feature extraction and its limitations in previous related work.
S.
No
Author and
Year
DatabaseFeature ExtractionClassifiers
Used
Evaluation
Metric
Limitations
1Ihsan et al.
(2022) [20]
MIMIC IIWavelet transform and time-domain featuresDecision tree (DT)Accuracy: 94.4%
sensitivity: 100%
specificity: 90.9%
Lack of model interpretability.
2Pal et al.
(2020) [21]
Medical College and Hospital, Kolkata.Time domain features-Double differentiation for peak and trough identification in cardiac cyclesDT, Discriminant Analysis, Logistic Regression (LogR), SVM, KNN, Boosted Trees (BT)For Boosted tree
Accuracy = 94%
Sensitivity = 95%
Precision = 97%
Requires the tuning of a regularization parameter, which can be challenging, and may arbitrarily select correlated features.
3Kanawade et al.
(2019) [22]
PhysioNet MIMIC IICrest to crest interval feature, valley to valley interval feature, transit time and beats per minute featureANN, SVM,
LogR, DT, Random forest
SVM
Accuracy = 97.67%
High computational cost for handling large datasets and a lack of a unique solution make it highly sensitive to hyperparameter settings.
4Paradkar et al. (2017) [23]PhysioNet MIMIC IITemporal features obtained through wavelet TransformSVMAccuracy = 85%
Specificity = 78%
Insufficient discussion regarding the possible biases in the dataset employed for predictive analytics.
5Banerjee et al.
(2016) [24]
MIMIC II and
Inhouse datasets
Time domain features, Frequency Domain Features from HRVSVM-RBFAccuracy: 85% for MIMIC II and 80% for inhouse datasetSelection of tuning parameters are challenging and
Can be utilized only in clinics with limited infrastructure.
6Xing et al.
(2016) [25]
MIMIC IIFeatures obtained through Fast Fourier TransformANNAverage Accuracy: 82.33%Resource-intensive for large feature sets and may fail to capture interactions between feature.
7Chakraborty et al. (2020) [26]PPG data from actual subjects using
BIOPAC MP 45
Different time-plane parametersDT, LogR, KNN
SVM-linear,
SVM-Nonlinear
For SVM-linear
Sensitivity: 92.70%
Accuracy: 95.4%
Absence of large-scale validation, the actual capabilities and limitations of the proposed system for diagnosing myocardial infarction remain certain.
8Mangathayaru et al. (2020) [27] BIDMC-PGG Features obtained through dual-tree complex wavelet transform (DT-CWT)Neural network architecture composed of successive GRU layersAccuracy: 98.82%Highly reliant on the selection of nearest neighbors, and may perform poorly when handling imbalanced datasets.
9Prabhakar et al.(2019) [28]Capnobase dataset Chi square PDF, Density Peaks, Chi square CDF, Harmonic search, Elephant search, Particle swarm, Chicken swarm and Cat swarm optimization LR, SVM-linear, SVM-Polynomial,
SVM-Gaussian, KNN, ANN, NBC, DT, GMM, ELM
Accuracy: 99.48%May not investigate the effects of varying hyperparameters on the performance of the proposed models, which might have impacted the overall results.
10Tjahjadi et al. (2020) [29]PPG-BP figshare databaseFeatures obtained through Short-Time Fourier TransformBidirectional long short-term memory (BLSTM) networkAverage accuracy: 96.20%A larger and more sample size is required to further validate and refine the classification performance.
Table 2. Selection of parameters for heuristic algorithms.
Table 2. Selection of parameters for heuristic algorithms.
ParametersHeuristic Algorithms
ABC-PSOCSADFA
Population Size200200200
Control parametersInertia weight ω : 0.45
Acceleration coefficients c 1 = 1 and c 2 = 1.1
Probability P a = 0.4
Step Size α = 1.5
Separation   s = 0.05
Alignment   a = 0.06
Cohesion c = 0.06
Attraction   f = 0.05
Distraction e = 0.2
AlgorithmSwarm intelligence with Hybrid Levy flightSwarm intelligence
Stopping CriteriaTraining MSE of 10−5Training MSE of 10−5Training MSE of 10−5
Number of iteration200200200
Local Minima ProblemAvailable in ABC. With proper selection of c 1 and c 2 in the PSO algorithm through trial and error method. The local minima problem will be solved.No local minima problemNo local minima problem
Over fittingOver fitting is available due to α and β values of ABC. This can be overcome with the proper selection of Weight ( ω ) of PSO AlgorithmOver fitting is not presentedOver fitting is not presented
Table 3. Average Statistical Metrics of ABC-PSO, Cuckoo Search, and Dragonfly Dimensionality Reduction Methods for Normal and CVD Patients.
Table 3. Average Statistical Metrics of ABC-PSO, Cuckoo Search, and Dragonfly Dimensionality Reduction Methods for Normal and CVD Patients.
Dimensionality Reduction TechniquesCategoryStatistical Metrics
MeanVarianceSkewnessKurtosisPCCSample EntropyCCA
ABC-PSONormal0.07320.0063−0.11650.2713−0.05979.94940.1066
CVD0.78720.3353−0.10000.14350.01339.9473
Cuckoo searchNormal0.52360.04750.1575−0.45560.33939.94940.3674
CVD7.893134.5391−0.0901−1.72900.22944.9919
DragonflyNormal−1.5850378.4756−0.0243−0.9585−0.21459.94990.4621
CVD−3.4728271.97350.0381−0.69190.10449.9522
Table 4. Comparative Analysis of Training and Testing Mean Square Error Across Various Dimensionality Reduction Approaches.
Table 4. Comparative Analysis of Training and Testing Mean Square Error Across Various Dimensionality Reduction Approaches.
ClassifiersABC PSOCuckoo SearchDragon Fly
Training MSETesting MSETraining MSETesting MSETraining MSETesting MSE
Linear Regression3.52 × 10−92.92 × 10−75.69 × 10−91.37 × 10−85.99 × 10−91.44 × 10−6
Linear Regression with BDLC2.32 × 10−61.10 × 10−49.69 × 10−88.65 × 10−66.03 × 10−82.72 × 10−6
KNN (weighted)5.72 × 10−81.44 × 10−64.60 × 10−62.81 × 10−37.02 × 10−83.24 × 10−6
PCA firefly6.65 × 10−73.80 × 10−58.45 × 10−66.08 × 10−38.69 × 10−66.25 × 10−5
LDA6.69 × 10−55.48 × 10−35.05 × 10−61.44 × 10−55.54 × 10−62.70 × 10−5
KLDA7.34 × 10−64.84 × 10−34.63 × 10−81.69 × 10−66.63 × 10−61.22 × 10−5
ProbLDA5.83 × 10−63.06 × 10−57.97 × 10−86.76 × 10−65.99 × 10−71.68 × 10−5
SVM (Linear)4.05 × 10−61.69 × 10−44.85 × 10−81.82 × 10−67.89 × 10−69.03 × 10−3
SVM (Polynomial)8.29 × 10−86.76 × 10−66.74 × 10−81.44 × 10−68.20 × 10−77.29 × 10−6
SVM(RBF)1.92 × 10−62.45 × 10−95.38 × 10−71.85 × 10−52.45 × 10−103.62 × 10−9
Table 5. Optimal Parameter Selection for Classifiers.
Table 5. Optimal Parameter Selection for Classifiers.
ClassifiersOptimal Parameters of the Classifiers
Linear Regression (LR)Uniform weight w = 0.451, bias: 0.003,
Criterion: MSE
LR with BLDCThe cascading configuration of LR with the following BLDC parameters: Class mean μ p = 0.8   and μ q = 0.1 ,
Prior probability P(x): 0.5
K-Nearest Neighbors (KNN)Number of clusters = 2
PCA FireflyPCA: A threshold value of 0.72 and decorrelated Eigen vector w k , using a trial and error training approach
Firefly: Initial conditions of α s 0 = 0.65, γ = 0.1
For both PCA and firefly, consider MSE of (10)−5 or reaching a maximum of 1000 iterations, whichever comes earliest.
Criterion: MSE
Linear Discriminant Analysis (LDA)Weight w = 0.56, bias: 0.0018
Kernel LDA (KLDA)Number of clusters: 2, w1: 0.38, w2: 0.642, bias: 0.0026 ± 0.0001
Probabilistic LDA(ProbLDA)Weight w = 0.56, bias: 0.0018, Assigned Probability > 0.5
SVM-LinearClass weights: 0.4
Parameter for Regularization [C]: 0.85
Criteria for Convergence: MSE
SVM-PolynomialParameter for Regularization [C]: 0.76
Class weights: 0.5
Kernel Function Coefficient [Gamma]: 10
Criteria for Convergence: MSE
SVM-RBFParameter for Regularization [C]: 1
Class weights: 0.86
Kernel Function Coefficient [Gamma]: 100
Criteria for Convergence: MSE
Table 6. Comparative Analysis of Classifiers on Dimensionality Reduced Features.
Table 6. Comparative Analysis of Classifiers on Dimensionality Reduced Features.
DR TechniquesClassifiers Accuracy
(%)
GDR (%)Error Rate (%)KappaMCCF1 Score
(%)
JI
(%)
ABC-PSOLinear Regression 90.2489.749.760.800.8090.0081.82
LR-BLDC78.0572.7321.950.560.6080.8567.86
K-Nearest Neighbors 78.0572.7321.950.560.6080.8567.86
PCA Firefly65.8557.5834.150.320.3266.6750.00
Linear Discriminant Analysis 58.5448.4841.460.170.1756.4139.29
Kernel LDA 53.6638.7146.340.070.0753.6636.67
Probabilistic LDA68.2959.3831.710.370.3871.1155.17
SVM-Linear65.8557.5834.150.320.3266.6750.00
SVM-Polynomial82.9381.0817.070.660.6682.9370.83
SVM-RBF95.1295.004.880.900.9095.0090.48
Cuckoo SearchLinear Regression 90.2489.749.760.800.8090.0081.82
LR-BLDC75.6170.5924.390.510.5277.2762.96
K-Nearest Neighbors 63.4153.1336.590.270.2765.1248.28
PCA Firefly53.6638.7146.340.070.0753.6636.67
Linear Discriminant Analysis 75.6174.3624.390.510.5370.5954.55
Kernel LDA 75.6170.5924.390.510.5277.2762.96
Probabilistic LDA85.3785.0014.630.710.7283.3371.43
SVM-Linear75.6175.0024.390.510.5568.7552.38
SVM-Polynomial78.0576.9221.950.560.5874.2959.09
SVM-RBF85.3783.7814.630.710.7185.7175.00
Dragon FlyLinear Regression 90.2489.749.760.800.8090.0081.82
LR-BLDC85.3785.0014.630.710.7283.3371.43
K-Nearest Neighbors 70.7362.5029.270.420.4473.9158.62
PCA Firefly68.2958.0631.710.370.3972.3456.67
Linear Discriminant Analysis 68.2958.0631.710.370.3972.3456.67
Kernel LDA 82.9381.5817.070.660.6682.0569.57
Probabilistic LDA68.2962.8631.710.360.3766.6750.00
SVM-Linear58.5446.8841.460.170.1758.5441.38
SVM-Polynomial63.4153.1336.590.270.2765.1248.28
SVM-RBF92.6892.507.320.850.8592.6886.36
Table 7. Computational complexity for all classifiers across different DR methods.
Table 7. Computational complexity for all classifiers across different DR methods.
ClassifiersHeuristic Dimensionality Reduction Techniques
ABC-PSOCSADFA
Linear Regression (LR) O ( m 5 ) O ( 2 m 2 log m ) O ( 4 m 2 log m )
LR with BLDC O ( m 7 ) O ( 2 m 4 log m ) O ( 4 m 4 log m )
K-Nearest Neighbors (KNN) O ( m 5 ) O ( 2 m 2 log m ) O ( 4 m 2 log m )
PCA Firefly O ( m 9 log m ) O ( 2 m 3 log 2 m ) O ( 3 m 3 log 2 m )
Linear Discriminant Analysis (LDA) O ( m 5 ) O ( 2 m 2 log m ) O ( 4 m 2 log m )
Kernel LDA (KLDA) O ( m 6 ) O ( 2 m 3 log m ) O ( 4 m 3 log m )
Probabilistic LDA(ProbLDA) O ( m 6 l o g m ) O ( 2 m 3 log 2 m ) O ( 4 m 3 log 2 m )
SVM-Linear O ( 2 m 4 log m ) O ( 4 m log m ) O ( 8 m log 2 m )
SVM-Polynomial O ( 2 m 5 log m ) O ( 4 m 2 log 2 m ) O ( 8 m 2 log 2 m )
SVM-RBF O ( 2 m 6 log 4 m ) O ( 2 m 3 log 5 m ) O ( 8 m 3 log 5 m )
Table 8. Comparison of previous works on CVD classification using PPG signals.
Table 8. Comparison of previous works on CVD classification using PPG signals.
SL.NOAuthorsDatasetNumber of SubjectsClassifiersClassesAccuracy (%)
1Rajaguru et al. [51]
2023
Capnobase datasetSingle patientLRCVD, Normal65.85%
2Al Fahoum et al. [52]
2023
Internal medicine clinic of Princess Basma Hospital200 healthy and 160 with CVDNBNormal and
abnormal
89.37%
3Prabhakar et al. [53]
2020
Capnobase dataset28 CVD
14 Normal
SVM–RBF
RBF NN
CVD, Normal95.05%
94.79%
4Liu et al. [54]
2022
GitHub
https://github.com/zdzdliu/PPGArrhythmiaDetection (accessed on 9 October 2024)
45 SubjectsDCNNCVD, Normal85%
5Hosseini et al. [55]
2015
Tehran Heart Center18 Normal
30 CVD
KNNLow risk
High risk
81.5%
6Miao and Miao [56]
2018
Cleveland Clinic Foundation303 patientsDNNCVD, Normal83.67%
7Shobita et al. [57]
2016
Biomedical Research Lab30 healthy 30 pathologicalELMHealthy
Risk of CVD
89.33%
8Soltane et al. [58]
2005
Seremban Hospital114 healthy 56 pathologicalANNCVD, Normal94.70%
9This researchCapnobase dataset21 Normal
20 CVD
SVM-RBFCVD, Normal95.12%
LR-Linear regression; NB- Naive Bayes; DCNN-Deep Convolutional Neural Network; SVM-RBF- Support Vector Machine-Radial Basis Function; RBF NN-Radial Basis Function Neural Network; KNN- K-Nearest Neighbor; ELM- Extreme learning machine; DNN-Deep Neural Network; ANN-Artificial Neural Network.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Palanisamy, S.; Rajaguru, H. Leveraging Classifier Performance Using Heuristic Optimization for Detecting Cardiovascular Disease from PPG Signals. Diagnostics 2024, 14, 2287. https://doi.org/10.3390/diagnostics14202287

AMA Style

Palanisamy S, Rajaguru H. Leveraging Classifier Performance Using Heuristic Optimization for Detecting Cardiovascular Disease from PPG Signals. Diagnostics. 2024; 14(20):2287. https://doi.org/10.3390/diagnostics14202287

Chicago/Turabian Style

Palanisamy, Sivamani, and Harikumar Rajaguru. 2024. "Leveraging Classifier Performance Using Heuristic Optimization for Detecting Cardiovascular Disease from PPG Signals" Diagnostics 14, no. 20: 2287. https://doi.org/10.3390/diagnostics14202287

APA Style

Palanisamy, S., & Rajaguru, H. (2024). Leveraging Classifier Performance Using Heuristic Optimization for Detecting Cardiovascular Disease from PPG Signals. Diagnostics, 14(20), 2287. https://doi.org/10.3390/diagnostics14202287

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop